@roo-code/types 1.91.0 → 1.93.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.js CHANGED
@@ -571,6 +571,7 @@ var anthropicModels = {
571
571
  supportsImages: true,
572
572
  supportsPromptCache: true,
573
573
  supportsNativeTools: true,
574
+ defaultToolProtocol: "native",
574
575
  inputPrice: 3,
575
576
  // $3 per million input tokens (≤200K context)
576
577
  outputPrice: 15,
@@ -604,6 +605,7 @@ var anthropicModels = {
604
605
  supportsImages: true,
605
606
  supportsPromptCache: true,
606
607
  supportsNativeTools: true,
608
+ defaultToolProtocol: "native",
607
609
  inputPrice: 3,
608
610
  // $3 per million input tokens (≤200K context)
609
611
  outputPrice: 15,
@@ -636,6 +638,7 @@ var anthropicModels = {
636
638
  supportsImages: true,
637
639
  supportsPromptCache: true,
638
640
  supportsNativeTools: true,
641
+ defaultToolProtocol: "native",
639
642
  inputPrice: 5,
640
643
  // $5 per million input tokens
641
644
  outputPrice: 25,
@@ -653,6 +656,7 @@ var anthropicModels = {
653
656
  supportsImages: true,
654
657
  supportsPromptCache: true,
655
658
  supportsNativeTools: true,
659
+ defaultToolProtocol: "native",
656
660
  inputPrice: 15,
657
661
  // $15 per million input tokens
658
662
  outputPrice: 75,
@@ -670,6 +674,7 @@ var anthropicModels = {
670
674
  supportsImages: true,
671
675
  supportsPromptCache: true,
672
676
  supportsNativeTools: true,
677
+ defaultToolProtocol: "native",
673
678
  inputPrice: 15,
674
679
  // $15 per million input tokens
675
680
  outputPrice: 75,
@@ -687,6 +692,7 @@ var anthropicModels = {
687
692
  supportsImages: true,
688
693
  supportsPromptCache: true,
689
694
  supportsNativeTools: true,
695
+ defaultToolProtocol: "native",
690
696
  inputPrice: 3,
691
697
  // $3 per million input tokens
692
698
  outputPrice: 15,
@@ -705,6 +711,7 @@ var anthropicModels = {
705
711
  supportsImages: true,
706
712
  supportsPromptCache: true,
707
713
  supportsNativeTools: true,
714
+ defaultToolProtocol: "native",
708
715
  inputPrice: 3,
709
716
  // $3 per million input tokens
710
717
  outputPrice: 15,
@@ -720,6 +727,7 @@ var anthropicModels = {
720
727
  supportsImages: true,
721
728
  supportsPromptCache: true,
722
729
  supportsNativeTools: true,
730
+ defaultToolProtocol: "native",
723
731
  inputPrice: 3,
724
732
  // $3 per million input tokens
725
733
  outputPrice: 15,
@@ -735,6 +743,7 @@ var anthropicModels = {
735
743
  supportsImages: false,
736
744
  supportsPromptCache: true,
737
745
  supportsNativeTools: true,
746
+ defaultToolProtocol: "native",
738
747
  inputPrice: 1,
739
748
  outputPrice: 5,
740
749
  cacheWritesPrice: 1.25,
@@ -746,6 +755,7 @@ var anthropicModels = {
746
755
  supportsImages: true,
747
756
  supportsPromptCache: true,
748
757
  supportsNativeTools: true,
758
+ defaultToolProtocol: "native",
749
759
  inputPrice: 15,
750
760
  outputPrice: 75,
751
761
  cacheWritesPrice: 18.75,
@@ -757,6 +767,7 @@ var anthropicModels = {
757
767
  supportsImages: true,
758
768
  supportsPromptCache: true,
759
769
  supportsNativeTools: true,
770
+ defaultToolProtocol: "native",
760
771
  inputPrice: 0.25,
761
772
  outputPrice: 1.25,
762
773
  cacheWritesPrice: 0.3,
@@ -768,6 +779,7 @@ var anthropicModels = {
768
779
  supportsImages: true,
769
780
  supportsPromptCache: true,
770
781
  supportsNativeTools: true,
782
+ defaultToolProtocol: "native",
771
783
  inputPrice: 1,
772
784
  outputPrice: 5,
773
785
  cacheWritesPrice: 1.25,
@@ -920,6 +932,7 @@ var bedrockModels = {
920
932
  supportsPromptCache: true,
921
933
  supportsReasoningBudget: true,
922
934
  supportsNativeTools: true,
935
+ defaultToolProtocol: "native",
923
936
  inputPrice: 3,
924
937
  outputPrice: 15,
925
938
  cacheWritesPrice: 3.75,
@@ -974,6 +987,22 @@ var bedrockModels = {
974
987
  maxCachePoints: 1,
975
988
  cachableFields: ["system"]
976
989
  },
990
+ "amazon.nova-2-lite-v1:0": {
991
+ maxTokens: 65535,
992
+ contextWindow: 1e6,
993
+ supportsImages: true,
994
+ supportsPromptCache: true,
995
+ supportsNativeTools: true,
996
+ inputPrice: 0.33,
997
+ outputPrice: 2.75,
998
+ cacheWritesPrice: 0,
999
+ cacheReadsPrice: 0.0825,
1000
+ // 75% less than input price
1001
+ minTokensPerCachePoint: 1,
1002
+ maxCachePoints: 1,
1003
+ cachableFields: ["system"],
1004
+ description: "Amazon Nova 2 Lite - Comparable to Claude Haiku 4.5"
1005
+ },
977
1006
  "amazon.nova-micro-v1:0": {
978
1007
  maxTokens: 5e3,
979
1008
  contextWindow: 128e3,
@@ -997,6 +1026,7 @@ var bedrockModels = {
997
1026
  supportsPromptCache: true,
998
1027
  supportsReasoningBudget: true,
999
1028
  supportsNativeTools: true,
1029
+ defaultToolProtocol: "native",
1000
1030
  inputPrice: 3,
1001
1031
  outputPrice: 15,
1002
1032
  cacheWritesPrice: 3.75,
@@ -1012,6 +1042,7 @@ var bedrockModels = {
1012
1042
  supportsPromptCache: true,
1013
1043
  supportsReasoningBudget: true,
1014
1044
  supportsNativeTools: true,
1045
+ defaultToolProtocol: "native",
1015
1046
  inputPrice: 15,
1016
1047
  outputPrice: 75,
1017
1048
  cacheWritesPrice: 18.75,
@@ -1027,6 +1058,7 @@ var bedrockModels = {
1027
1058
  supportsPromptCache: true,
1028
1059
  supportsReasoningBudget: true,
1029
1060
  supportsNativeTools: true,
1061
+ defaultToolProtocol: "native",
1030
1062
  inputPrice: 5,
1031
1063
  outputPrice: 25,
1032
1064
  cacheWritesPrice: 6.25,
@@ -1042,6 +1074,7 @@ var bedrockModels = {
1042
1074
  supportsPromptCache: true,
1043
1075
  supportsReasoningBudget: true,
1044
1076
  supportsNativeTools: true,
1077
+ defaultToolProtocol: "native",
1045
1078
  inputPrice: 15,
1046
1079
  outputPrice: 75,
1047
1080
  cacheWritesPrice: 18.75,
@@ -1057,6 +1090,7 @@ var bedrockModels = {
1057
1090
  supportsPromptCache: true,
1058
1091
  supportsReasoningBudget: true,
1059
1092
  supportsNativeTools: true,
1093
+ defaultToolProtocol: "native",
1060
1094
  inputPrice: 3,
1061
1095
  outputPrice: 15,
1062
1096
  cacheWritesPrice: 3.75,
@@ -1071,6 +1105,7 @@ var bedrockModels = {
1071
1105
  supportsImages: true,
1072
1106
  supportsPromptCache: true,
1073
1107
  supportsNativeTools: true,
1108
+ defaultToolProtocol: "native",
1074
1109
  inputPrice: 3,
1075
1110
  outputPrice: 15,
1076
1111
  cacheWritesPrice: 3.75,
@@ -1085,6 +1120,7 @@ var bedrockModels = {
1085
1120
  supportsImages: false,
1086
1121
  supportsPromptCache: true,
1087
1122
  supportsNativeTools: true,
1123
+ defaultToolProtocol: "native",
1088
1124
  inputPrice: 0.8,
1089
1125
  outputPrice: 4,
1090
1126
  cacheWritesPrice: 1,
@@ -1100,6 +1136,7 @@ var bedrockModels = {
1100
1136
  supportsPromptCache: true,
1101
1137
  supportsReasoningBudget: true,
1102
1138
  supportsNativeTools: true,
1139
+ defaultToolProtocol: "native",
1103
1140
  inputPrice: 1,
1104
1141
  outputPrice: 5,
1105
1142
  cacheWritesPrice: 1.25,
@@ -1116,6 +1153,7 @@ var bedrockModels = {
1116
1153
  supportsImages: true,
1117
1154
  supportsPromptCache: false,
1118
1155
  supportsNativeTools: true,
1156
+ defaultToolProtocol: "native",
1119
1157
  inputPrice: 3,
1120
1158
  outputPrice: 15
1121
1159
  },
@@ -1125,6 +1163,7 @@ var bedrockModels = {
1125
1163
  supportsImages: true,
1126
1164
  supportsPromptCache: false,
1127
1165
  supportsNativeTools: true,
1166
+ defaultToolProtocol: "native",
1128
1167
  inputPrice: 15,
1129
1168
  outputPrice: 75
1130
1169
  },
@@ -1134,6 +1173,7 @@ var bedrockModels = {
1134
1173
  supportsImages: true,
1135
1174
  supportsPromptCache: false,
1136
1175
  supportsNativeTools: true,
1176
+ defaultToolProtocol: "native",
1137
1177
  inputPrice: 3,
1138
1178
  outputPrice: 15
1139
1179
  },
@@ -1143,6 +1183,7 @@ var bedrockModels = {
1143
1183
  supportsImages: true,
1144
1184
  supportsPromptCache: false,
1145
1185
  supportsNativeTools: true,
1186
+ defaultToolProtocol: "native",
1146
1187
  inputPrice: 0.25,
1147
1188
  outputPrice: 1.25
1148
1189
  },
@@ -1152,6 +1193,7 @@ var bedrockModels = {
1152
1193
  supportsImages: false,
1153
1194
  supportsPromptCache: false,
1154
1195
  supportsNativeTools: true,
1196
+ defaultToolProtocol: "native",
1155
1197
  inputPrice: 8,
1156
1198
  outputPrice: 24,
1157
1199
  description: "Claude 2.1"
@@ -1162,6 +1204,7 @@ var bedrockModels = {
1162
1204
  supportsImages: false,
1163
1205
  supportsPromptCache: false,
1164
1206
  supportsNativeTools: true,
1207
+ defaultToolProtocol: "native",
1165
1208
  inputPrice: 8,
1166
1209
  outputPrice: 24,
1167
1210
  description: "Claude 2.0"
@@ -1172,6 +1215,7 @@ var bedrockModels = {
1172
1215
  supportsImages: false,
1173
1216
  supportsPromptCache: false,
1174
1217
  supportsNativeTools: true,
1218
+ defaultToolProtocol: "native",
1175
1219
  inputPrice: 0.8,
1176
1220
  outputPrice: 2.4,
1177
1221
  description: "Claude Instant"
@@ -1454,6 +1498,29 @@ var BEDROCK_GLOBAL_INFERENCE_MODEL_IDS = [
1454
1498
  "anthropic.claude-haiku-4-5-20251001-v1:0",
1455
1499
  "anthropic.claude-opus-4-5-20251101-v1:0"
1456
1500
  ];
1501
+ var BEDROCK_SERVICE_TIER_MODEL_IDS = [
1502
+ // Amazon Nova models
1503
+ "amazon.nova-lite-v1:0",
1504
+ "amazon.nova-2-lite-v1:0",
1505
+ "amazon.nova-pro-v1:0",
1506
+ "amazon.nova-pro-latency-optimized-v1:0",
1507
+ // DeepSeek models
1508
+ "deepseek.r1-v1:0",
1509
+ // Qwen models
1510
+ "qwen.qwen3-next-80b-a3b",
1511
+ "qwen.qwen3-coder-480b-a35b-v1:0",
1512
+ // OpenAI GPT-OSS models
1513
+ "openai.gpt-oss-20b-1:0",
1514
+ "openai.gpt-oss-120b-1:0"
1515
+ ];
1516
+ var BEDROCK_SERVICE_TIER_PRICING = {
1517
+ STANDARD: 1,
1518
+ // Base price
1519
+ FLEX: 0.5,
1520
+ // 50% discount from standard
1521
+ PRIORITY: 1.75
1522
+ // 75% premium over standard
1523
+ };
1457
1524
 
1458
1525
  // src/providers/cerebras.ts
1459
1526
  var cerebrasDefaultModelId = "gpt-oss-120b";
@@ -1465,6 +1532,7 @@ var cerebrasModels = {
1465
1532
  supportsImages: false,
1466
1533
  supportsPromptCache: false,
1467
1534
  supportsNativeTools: true,
1535
+ defaultToolProtocol: "native",
1468
1536
  inputPrice: 0,
1469
1537
  outputPrice: 0,
1470
1538
  description: "Highly intelligent general purpose model with up to 1,000 tokens/s"
@@ -1476,6 +1544,7 @@ var cerebrasModels = {
1476
1544
  supportsImages: false,
1477
1545
  supportsPromptCache: false,
1478
1546
  supportsNativeTools: true,
1547
+ defaultToolProtocol: "native",
1479
1548
  inputPrice: 0,
1480
1549
  outputPrice: 0,
1481
1550
  description: "Intelligent model with ~1400 tokens/s"
@@ -1487,6 +1556,7 @@ var cerebrasModels = {
1487
1556
  supportsImages: false,
1488
1557
  supportsPromptCache: false,
1489
1558
  supportsNativeTools: true,
1559
+ defaultToolProtocol: "native",
1490
1560
  inputPrice: 0,
1491
1561
  outputPrice: 0,
1492
1562
  description: "Powerful model with ~2600 tokens/s"
@@ -1498,6 +1568,7 @@ var cerebrasModels = {
1498
1568
  supportsImages: false,
1499
1569
  supportsPromptCache: false,
1500
1570
  supportsNativeTools: true,
1571
+ defaultToolProtocol: "native",
1501
1572
  inputPrice: 0,
1502
1573
  outputPrice: 0,
1503
1574
  description: "SOTA coding performance with ~2500 tokens/s"
@@ -1509,6 +1580,7 @@ var cerebrasModels = {
1509
1580
  supportsImages: false,
1510
1581
  supportsPromptCache: false,
1511
1582
  supportsNativeTools: true,
1583
+ defaultToolProtocol: "native",
1512
1584
  inputPrice: 0,
1513
1585
  outputPrice: 0,
1514
1586
  description: "OpenAI GPT OSS model with ~2800 tokens/s\n\n\u2022 64K context window\n\u2022 Excels at efficient reasoning across science, math, and coding"
@@ -1523,6 +1595,8 @@ var chutesModels = {
1523
1595
  contextWindow: 163840,
1524
1596
  supportsImages: false,
1525
1597
  supportsPromptCache: false,
1598
+ supportsNativeTools: true,
1599
+ defaultToolProtocol: "native",
1526
1600
  inputPrice: 0,
1527
1601
  outputPrice: 0,
1528
1602
  description: "DeepSeek R1 0528 model."
@@ -1532,6 +1606,8 @@ var chutesModels = {
1532
1606
  contextWindow: 163840,
1533
1607
  supportsImages: false,
1534
1608
  supportsPromptCache: false,
1609
+ supportsNativeTools: true,
1610
+ defaultToolProtocol: "native",
1535
1611
  inputPrice: 0,
1536
1612
  outputPrice: 0,
1537
1613
  description: "DeepSeek R1 model."
@@ -1541,6 +1617,8 @@ var chutesModels = {
1541
1617
  contextWindow: 163840,
1542
1618
  supportsImages: false,
1543
1619
  supportsPromptCache: false,
1620
+ supportsNativeTools: true,
1621
+ defaultToolProtocol: "native",
1544
1622
  inputPrice: 0,
1545
1623
  outputPrice: 0,
1546
1624
  description: "DeepSeek V3 model."
@@ -1550,6 +1628,8 @@ var chutesModels = {
1550
1628
  contextWindow: 163840,
1551
1629
  supportsImages: false,
1552
1630
  supportsPromptCache: false,
1631
+ supportsNativeTools: true,
1632
+ defaultToolProtocol: "native",
1553
1633
  inputPrice: 0,
1554
1634
  outputPrice: 0,
1555
1635
  description: "DeepSeek V3.1 model."
@@ -1559,6 +1639,8 @@ var chutesModels = {
1559
1639
  contextWindow: 163840,
1560
1640
  supportsImages: false,
1561
1641
  supportsPromptCache: false,
1642
+ supportsNativeTools: true,
1643
+ defaultToolProtocol: "native",
1562
1644
  inputPrice: 0.23,
1563
1645
  outputPrice: 0.9,
1564
1646
  description: "DeepSeek\u2011V3.1\u2011Terminus is an update to V3.1 that improves language consistency by reducing CN/EN mix\u2011ups and eliminating random characters, while strengthening agent capabilities with notably better Code Agent and Search Agent performance."
@@ -1568,6 +1650,8 @@ var chutesModels = {
1568
1650
  contextWindow: 163840,
1569
1651
  supportsImages: false,
1570
1652
  supportsPromptCache: false,
1653
+ supportsNativeTools: true,
1654
+ defaultToolProtocol: "native",
1571
1655
  inputPrice: 1,
1572
1656
  outputPrice: 3,
1573
1657
  description: "DeepSeek-V3.1-turbo is an FP8, speculative-decoding turbo variant optimized for ultra-fast single-shot queries (~200 TPS), with outputs close to the originals and solid function calling/reasoning/structured output, priced at $1/M input and $3/M output tokens, using 2\xD7 quota per request and not intended for bulk workloads."
@@ -1577,6 +1661,8 @@ var chutesModels = {
1577
1661
  contextWindow: 163840,
1578
1662
  supportsImages: false,
1579
1663
  supportsPromptCache: false,
1664
+ supportsNativeTools: true,
1665
+ defaultToolProtocol: "native",
1580
1666
  inputPrice: 0.25,
1581
1667
  outputPrice: 0.35,
1582
1668
  description: "DeepSeek-V3.2-Exp is an experimental LLM that introduces DeepSeek Sparse Attention to improve long\u2011context training and inference efficiency while maintaining performance comparable to V3.1\u2011Terminus."
@@ -1588,6 +1674,8 @@ var chutesModels = {
1588
1674
  // From Groq
1589
1675
  supportsImages: false,
1590
1676
  supportsPromptCache: false,
1677
+ supportsNativeTools: true,
1678
+ defaultToolProtocol: "native",
1591
1679
  inputPrice: 0,
1592
1680
  outputPrice: 0,
1593
1681
  description: "Unsloth Llama 3.3 70B Instruct model."
@@ -1597,6 +1685,8 @@ var chutesModels = {
1597
1685
  contextWindow: 512e3,
1598
1686
  supportsImages: false,
1599
1687
  supportsPromptCache: false,
1688
+ supportsNativeTools: true,
1689
+ defaultToolProtocol: "native",
1600
1690
  inputPrice: 0,
1601
1691
  outputPrice: 0,
1602
1692
  description: "ChutesAI Llama 4 Scout 17B Instruct model, 512K context."
@@ -1606,6 +1696,8 @@ var chutesModels = {
1606
1696
  contextWindow: 128e3,
1607
1697
  supportsImages: false,
1608
1698
  supportsPromptCache: false,
1699
+ supportsNativeTools: true,
1700
+ defaultToolProtocol: "native",
1609
1701
  inputPrice: 0,
1610
1702
  outputPrice: 0,
1611
1703
  description: "Unsloth Mistral Nemo Instruct model."
@@ -1615,6 +1707,8 @@ var chutesModels = {
1615
1707
  contextWindow: 131072,
1616
1708
  supportsImages: false,
1617
1709
  supportsPromptCache: false,
1710
+ supportsNativeTools: true,
1711
+ defaultToolProtocol: "native",
1618
1712
  inputPrice: 0,
1619
1713
  outputPrice: 0,
1620
1714
  description: "Unsloth Gemma 3 12B IT model."
@@ -1624,6 +1718,8 @@ var chutesModels = {
1624
1718
  contextWindow: 131072,
1625
1719
  supportsImages: false,
1626
1720
  supportsPromptCache: false,
1721
+ supportsNativeTools: true,
1722
+ defaultToolProtocol: "native",
1627
1723
  inputPrice: 0,
1628
1724
  outputPrice: 0,
1629
1725
  description: "Nous DeepHermes 3 Llama 3 8B Preview model."
@@ -1633,6 +1729,8 @@ var chutesModels = {
1633
1729
  contextWindow: 131072,
1634
1730
  supportsImages: false,
1635
1731
  supportsPromptCache: false,
1732
+ supportsNativeTools: true,
1733
+ defaultToolProtocol: "native",
1636
1734
  inputPrice: 0,
1637
1735
  outputPrice: 0,
1638
1736
  description: "Unsloth Gemma 3 4B IT model."
@@ -1642,6 +1740,8 @@ var chutesModels = {
1642
1740
  contextWindow: 131072,
1643
1741
  supportsImages: false,
1644
1742
  supportsPromptCache: false,
1743
+ supportsNativeTools: true,
1744
+ defaultToolProtocol: "native",
1645
1745
  inputPrice: 0,
1646
1746
  outputPrice: 0,
1647
1747
  description: "Nvidia Llama 3.3 Nemotron Super 49B model."
@@ -1651,6 +1751,8 @@ var chutesModels = {
1651
1751
  contextWindow: 131072,
1652
1752
  supportsImages: false,
1653
1753
  supportsPromptCache: false,
1754
+ supportsNativeTools: true,
1755
+ defaultToolProtocol: "native",
1654
1756
  inputPrice: 0,
1655
1757
  outputPrice: 0,
1656
1758
  description: "Nvidia Llama 3.1 Nemotron Ultra 253B model."
@@ -1660,6 +1762,8 @@ var chutesModels = {
1660
1762
  contextWindow: 256e3,
1661
1763
  supportsImages: false,
1662
1764
  supportsPromptCache: false,
1765
+ supportsNativeTools: true,
1766
+ defaultToolProtocol: "native",
1663
1767
  inputPrice: 0,
1664
1768
  outputPrice: 0,
1665
1769
  description: "ChutesAI Llama 4 Maverick 17B Instruct FP8 model."
@@ -1669,6 +1773,8 @@ var chutesModels = {
1669
1773
  contextWindow: 163840,
1670
1774
  supportsImages: false,
1671
1775
  supportsPromptCache: false,
1776
+ supportsNativeTools: true,
1777
+ defaultToolProtocol: "native",
1672
1778
  inputPrice: 0,
1673
1779
  outputPrice: 0,
1674
1780
  description: "DeepSeek V3 Base model."
@@ -1678,6 +1784,8 @@ var chutesModels = {
1678
1784
  contextWindow: 163840,
1679
1785
  supportsImages: false,
1680
1786
  supportsPromptCache: false,
1787
+ supportsNativeTools: true,
1788
+ defaultToolProtocol: "native",
1681
1789
  inputPrice: 0,
1682
1790
  outputPrice: 0,
1683
1791
  description: "DeepSeek R1 Zero model."
@@ -1687,6 +1795,8 @@ var chutesModels = {
1687
1795
  contextWindow: 163840,
1688
1796
  supportsImages: false,
1689
1797
  supportsPromptCache: false,
1798
+ supportsNativeTools: true,
1799
+ defaultToolProtocol: "native",
1690
1800
  inputPrice: 0,
1691
1801
  outputPrice: 0,
1692
1802
  description: "DeepSeek V3 (0324) model."
@@ -1696,6 +1806,8 @@ var chutesModels = {
1696
1806
  contextWindow: 262144,
1697
1807
  supportsImages: false,
1698
1808
  supportsPromptCache: false,
1809
+ supportsNativeTools: true,
1810
+ defaultToolProtocol: "native",
1699
1811
  inputPrice: 0,
1700
1812
  outputPrice: 0,
1701
1813
  description: "Qwen3 235B A22B Instruct 2507 model with 262K context window."
@@ -1705,6 +1817,8 @@ var chutesModels = {
1705
1817
  contextWindow: 40960,
1706
1818
  supportsImages: false,
1707
1819
  supportsPromptCache: false,
1820
+ supportsNativeTools: true,
1821
+ defaultToolProtocol: "native",
1708
1822
  inputPrice: 0,
1709
1823
  outputPrice: 0,
1710
1824
  description: "Qwen3 235B A22B model."
@@ -1714,6 +1828,8 @@ var chutesModels = {
1714
1828
  contextWindow: 40960,
1715
1829
  supportsImages: false,
1716
1830
  supportsPromptCache: false,
1831
+ supportsNativeTools: true,
1832
+ defaultToolProtocol: "native",
1717
1833
  inputPrice: 0,
1718
1834
  outputPrice: 0,
1719
1835
  description: "Qwen3 32B model."
@@ -1723,6 +1839,8 @@ var chutesModels = {
1723
1839
  contextWindow: 40960,
1724
1840
  supportsImages: false,
1725
1841
  supportsPromptCache: false,
1842
+ supportsNativeTools: true,
1843
+ defaultToolProtocol: "native",
1726
1844
  inputPrice: 0,
1727
1845
  outputPrice: 0,
1728
1846
  description: "Qwen3 30B A3B model."
@@ -1732,6 +1850,8 @@ var chutesModels = {
1732
1850
  contextWindow: 40960,
1733
1851
  supportsImages: false,
1734
1852
  supportsPromptCache: false,
1853
+ supportsNativeTools: true,
1854
+ defaultToolProtocol: "native",
1735
1855
  inputPrice: 0,
1736
1856
  outputPrice: 0,
1737
1857
  description: "Qwen3 14B model."
@@ -1741,6 +1861,8 @@ var chutesModels = {
1741
1861
  contextWindow: 40960,
1742
1862
  supportsImages: false,
1743
1863
  supportsPromptCache: false,
1864
+ supportsNativeTools: true,
1865
+ defaultToolProtocol: "native",
1744
1866
  inputPrice: 0,
1745
1867
  outputPrice: 0,
1746
1868
  description: "Qwen3 8B model."
@@ -1750,6 +1872,8 @@ var chutesModels = {
1750
1872
  contextWindow: 163840,
1751
1873
  supportsImages: false,
1752
1874
  supportsPromptCache: false,
1875
+ supportsNativeTools: true,
1876
+ defaultToolProtocol: "native",
1753
1877
  inputPrice: 0,
1754
1878
  outputPrice: 0,
1755
1879
  description: "Microsoft MAI-DS-R1 FP8 model."
@@ -1759,6 +1883,8 @@ var chutesModels = {
1759
1883
  contextWindow: 163840,
1760
1884
  supportsImages: false,
1761
1885
  supportsPromptCache: false,
1886
+ supportsNativeTools: true,
1887
+ defaultToolProtocol: "native",
1762
1888
  inputPrice: 0,
1763
1889
  outputPrice: 0,
1764
1890
  description: "TNGTech DeepSeek R1T Chimera model."
@@ -1768,6 +1894,8 @@ var chutesModels = {
1768
1894
  contextWindow: 151329,
1769
1895
  supportsImages: false,
1770
1896
  supportsPromptCache: false,
1897
+ supportsNativeTools: true,
1898
+ defaultToolProtocol: "native",
1771
1899
  inputPrice: 0,
1772
1900
  outputPrice: 0,
1773
1901
  description: "GLM-4.5-Air model with 151,329 token context window and 106B total parameters with 12B activated."
@@ -1777,6 +1905,8 @@ var chutesModels = {
1777
1905
  contextWindow: 131072,
1778
1906
  supportsImages: false,
1779
1907
  supportsPromptCache: false,
1908
+ supportsNativeTools: true,
1909
+ defaultToolProtocol: "native",
1780
1910
  inputPrice: 0,
1781
1911
  outputPrice: 0,
1782
1912
  description: "GLM-4.5-FP8 model with 128k token context window, optimized for agent-based applications with MoE architecture."
@@ -1786,6 +1916,8 @@ var chutesModels = {
1786
1916
  contextWindow: 131072,
1787
1917
  supportsImages: false,
1788
1918
  supportsPromptCache: false,
1919
+ supportsNativeTools: true,
1920
+ defaultToolProtocol: "native",
1789
1921
  inputPrice: 1,
1790
1922
  outputPrice: 3,
1791
1923
  description: "GLM-4.5-turbo model with 128K token context window, optimized for fast inference."
@@ -1795,6 +1927,8 @@ var chutesModels = {
1795
1927
  contextWindow: 202752,
1796
1928
  supportsImages: false,
1797
1929
  supportsPromptCache: false,
1930
+ supportsNativeTools: true,
1931
+ defaultToolProtocol: "native",
1798
1932
  inputPrice: 0,
1799
1933
  outputPrice: 0,
1800
1934
  description: "GLM-4.6 introduces major upgrades over GLM-4.5, including a longer 200K-token context window for complex tasks, stronger coding performance in benchmarks and real-world tools (such as Claude Code, Cline, Roo Code, and Kilo Code), improved reasoning with tool use during inference, more capable and efficient agent integration, and refined writing that better matches human style, readability, and natural role-play scenarios."
@@ -1805,6 +1939,8 @@ var chutesModels = {
1805
1939
  contextWindow: 202752,
1806
1940
  supportsImages: false,
1807
1941
  supportsPromptCache: false,
1942
+ supportsNativeTools: true,
1943
+ defaultToolProtocol: "native",
1808
1944
  inputPrice: 1.15,
1809
1945
  outputPrice: 3.25,
1810
1946
  description: "GLM-4.6-turbo model with 200K-token context window, optimized for fast inference."
@@ -1814,6 +1950,8 @@ var chutesModels = {
1814
1950
  contextWindow: 128e3,
1815
1951
  supportsImages: false,
1816
1952
  supportsPromptCache: false,
1953
+ supportsNativeTools: true,
1954
+ defaultToolProtocol: "native",
1817
1955
  inputPrice: 0,
1818
1956
  outputPrice: 0,
1819
1957
  description: "LongCat Flash Thinking FP8 model with 128K context window, optimized for complex reasoning and coding tasks."
@@ -1823,6 +1961,8 @@ var chutesModels = {
1823
1961
  contextWindow: 262144,
1824
1962
  supportsImages: false,
1825
1963
  supportsPromptCache: false,
1964
+ supportsNativeTools: true,
1965
+ defaultToolProtocol: "native",
1826
1966
  inputPrice: 0,
1827
1967
  outputPrice: 0,
1828
1968
  description: "Qwen3 Coder 480B A35B Instruct FP8 model, optimized for coding tasks."
@@ -1832,6 +1972,8 @@ var chutesModels = {
1832
1972
  contextWindow: 75e3,
1833
1973
  supportsImages: false,
1834
1974
  supportsPromptCache: false,
1975
+ supportsNativeTools: true,
1976
+ defaultToolProtocol: "native",
1835
1977
  inputPrice: 0.1481,
1836
1978
  outputPrice: 0.5926,
1837
1979
  description: "Moonshot AI Kimi K2 Instruct model with 75k context window."
@@ -1841,6 +1983,8 @@ var chutesModels = {
1841
1983
  contextWindow: 262144,
1842
1984
  supportsImages: false,
1843
1985
  supportsPromptCache: false,
1986
+ supportsNativeTools: true,
1987
+ defaultToolProtocol: "native",
1844
1988
  inputPrice: 0.1999,
1845
1989
  outputPrice: 0.8001,
1846
1990
  description: "Moonshot AI Kimi K2 Instruct 0905 model with 256k context window."
@@ -1850,6 +1994,8 @@ var chutesModels = {
1850
1994
  contextWindow: 262144,
1851
1995
  supportsImages: false,
1852
1996
  supportsPromptCache: false,
1997
+ supportsNativeTools: true,
1998
+ defaultToolProtocol: "native",
1853
1999
  inputPrice: 0.077968332,
1854
2000
  outputPrice: 0.31202496,
1855
2001
  description: "Qwen3 235B A22B Thinking 2507 model with 262K context window."
@@ -1859,6 +2005,8 @@ var chutesModels = {
1859
2005
  contextWindow: 131072,
1860
2006
  supportsImages: false,
1861
2007
  supportsPromptCache: false,
2008
+ supportsNativeTools: true,
2009
+ defaultToolProtocol: "native",
1862
2010
  inputPrice: 0,
1863
2011
  outputPrice: 0,
1864
2012
  description: "Fast, stable instruction-tuned model optimized for complex tasks, RAG, and tool use without thinking traces."
@@ -1868,6 +2016,8 @@ var chutesModels = {
1868
2016
  contextWindow: 131072,
1869
2017
  supportsImages: false,
1870
2018
  supportsPromptCache: false,
2019
+ supportsNativeTools: true,
2020
+ defaultToolProtocol: "native",
1871
2021
  inputPrice: 0,
1872
2022
  outputPrice: 0,
1873
2023
  description: "Reasoning-first model with structured thinking traces for multi-step problems, math proofs, and code synthesis."
@@ -1877,6 +2027,8 @@ var chutesModels = {
1877
2027
  contextWindow: 262144,
1878
2028
  supportsImages: true,
1879
2029
  supportsPromptCache: false,
2030
+ supportsNativeTools: true,
2031
+ defaultToolProtocol: "native",
1880
2032
  inputPrice: 0.16,
1881
2033
  outputPrice: 0.65,
1882
2034
  description: "Qwen3\u2011VL\u2011235B\u2011A22B\u2011Thinking is an open\u2011weight MoE vision\u2011language model (235B total, ~22B activated) optimized for deliberate multi\u2011step reasoning with strong text\u2011image\u2011video understanding and long\u2011context capabilities."
@@ -1885,139 +2037,71 @@ var chutesModels = {
1885
2037
  var chutesDefaultModelInfo = chutesModels[chutesDefaultModelId];
1886
2038
 
1887
2039
  // src/providers/claude-code.ts
1888
- var VERTEX_DATE_PATTERN = /-(\d{8})$/;
1889
- function convertModelNameForVertex(modelName) {
1890
- return modelName.replace(VERTEX_DATE_PATTERN, "@$1");
1891
- }
1892
- var claudeCodeDefaultModelId = "claude-sonnet-4-5";
1893
- var CLAUDE_CODE_DEFAULT_MAX_OUTPUT_TOKENS = 16e3;
1894
- function getClaudeCodeModelId(baseModelId, useVertex = false) {
1895
- return useVertex ? convertModelNameForVertex(baseModelId) : baseModelId;
1896
- }
2040
+ var DATE_SUFFIX_PATTERN = /-\d{8}$/;
1897
2041
  var claudeCodeModels = {
1898
- "claude-sonnet-4-5": {
1899
- ...anthropicModels["claude-sonnet-4-5"],
1900
- supportsImages: false,
1901
- supportsPromptCache: true,
1902
- // Claude Code does report cache tokens
1903
- supportsReasoningEffort: false,
1904
- supportsReasoningBudget: false,
1905
- requiredReasoningBudget: false,
1906
- // Claude Code manages its own tools and temperature via the CLI
1907
- supportsNativeTools: false,
1908
- supportsTemperature: false
1909
- },
1910
- "claude-sonnet-4-5-20250929[1m]": {
1911
- ...anthropicModels["claude-sonnet-4-5"],
1912
- contextWindow: 1e6,
1913
- // 1M token context window (requires [1m] suffix)
1914
- supportsImages: false,
1915
- supportsPromptCache: true,
1916
- // Claude Code does report cache tokens
1917
- supportsReasoningEffort: false,
1918
- supportsReasoningBudget: false,
1919
- requiredReasoningBudget: false,
1920
- // Claude Code manages its own tools and temperature via the CLI
1921
- supportsNativeTools: false,
1922
- supportsTemperature: false
1923
- },
1924
- "claude-sonnet-4-20250514": {
1925
- ...anthropicModels["claude-sonnet-4-20250514"],
1926
- supportsImages: false,
1927
- supportsPromptCache: true,
1928
- // Claude Code does report cache tokens
1929
- supportsReasoningEffort: false,
1930
- supportsReasoningBudget: false,
1931
- requiredReasoningBudget: false,
1932
- // Claude Code manages its own tools and temperature via the CLI
1933
- supportsNativeTools: false,
1934
- supportsTemperature: false
1935
- },
1936
- "claude-opus-4-5-20251101": {
1937
- ...anthropicModels["claude-opus-4-5-20251101"],
1938
- supportsImages: false,
1939
- supportsPromptCache: true,
1940
- // Claude Code does report cache tokens
1941
- supportsReasoningEffort: false,
1942
- supportsReasoningBudget: false,
1943
- requiredReasoningBudget: false,
1944
- // Claude Code manages its own tools and temperature via the CLI
1945
- supportsNativeTools: false,
1946
- supportsTemperature: false
1947
- },
1948
- "claude-opus-4-1-20250805": {
1949
- ...anthropicModels["claude-opus-4-1-20250805"],
1950
- supportsImages: false,
1951
- supportsPromptCache: true,
1952
- // Claude Code does report cache tokens
1953
- supportsReasoningEffort: false,
1954
- supportsReasoningBudget: false,
1955
- requiredReasoningBudget: false,
1956
- // Claude Code manages its own tools and temperature via the CLI
1957
- supportsNativeTools: false,
1958
- supportsTemperature: false
1959
- },
1960
- "claude-opus-4-20250514": {
1961
- ...anthropicModels["claude-opus-4-20250514"],
1962
- supportsImages: false,
1963
- supportsPromptCache: true,
1964
- // Claude Code does report cache tokens
1965
- supportsReasoningEffort: false,
1966
- supportsReasoningBudget: false,
1967
- requiredReasoningBudget: false,
1968
- // Claude Code manages its own tools and temperature via the CLI
1969
- supportsNativeTools: false,
1970
- supportsTemperature: false
1971
- },
1972
- "claude-3-7-sonnet-20250219": {
1973
- ...anthropicModels["claude-3-7-sonnet-20250219"],
1974
- supportsImages: false,
1975
- supportsPromptCache: true,
1976
- // Claude Code does report cache tokens
1977
- supportsReasoningEffort: false,
1978
- supportsReasoningBudget: false,
1979
- requiredReasoningBudget: false,
1980
- // Claude Code manages its own tools and temperature via the CLI
1981
- supportsNativeTools: false,
1982
- supportsTemperature: false
1983
- },
1984
- "claude-3-5-sonnet-20241022": {
1985
- ...anthropicModels["claude-3-5-sonnet-20241022"],
1986
- supportsImages: false,
2042
+ "claude-haiku-4-5": {
2043
+ maxTokens: 32768,
2044
+ contextWindow: 2e5,
2045
+ supportsImages: true,
1987
2046
  supportsPromptCache: true,
1988
- // Claude Code does report cache tokens
1989
- supportsReasoningEffort: false,
1990
- supportsReasoningBudget: false,
1991
- requiredReasoningBudget: false,
1992
- // Claude Code manages its own tools and temperature via the CLI
1993
- supportsNativeTools: false,
1994
- supportsTemperature: false
2047
+ supportsNativeTools: true,
2048
+ defaultToolProtocol: "native",
2049
+ supportsReasoningEffort: ["disable", "low", "medium", "high"],
2050
+ reasoningEffort: "medium",
2051
+ description: "Claude Haiku 4.5 - Fast and efficient with thinking"
1995
2052
  },
1996
- "claude-3-5-haiku-20241022": {
1997
- ...anthropicModels["claude-3-5-haiku-20241022"],
1998
- supportsImages: false,
2053
+ "claude-sonnet-4-5": {
2054
+ maxTokens: 32768,
2055
+ contextWindow: 2e5,
2056
+ supportsImages: true,
1999
2057
  supportsPromptCache: true,
2000
- // Claude Code does report cache tokens
2001
- supportsReasoningEffort: false,
2002
- supportsReasoningBudget: false,
2003
- requiredReasoningBudget: false,
2004
- // Claude Code manages its own tools and temperature via the CLI
2005
- supportsNativeTools: false,
2006
- supportsTemperature: false
2058
+ supportsNativeTools: true,
2059
+ defaultToolProtocol: "native",
2060
+ supportsReasoningEffort: ["disable", "low", "medium", "high"],
2061
+ reasoningEffort: "medium",
2062
+ description: "Claude Sonnet 4.5 - Balanced performance with thinking"
2007
2063
  },
2008
- "claude-haiku-4-5-20251001": {
2009
- ...anthropicModels["claude-haiku-4-5-20251001"],
2010
- supportsImages: false,
2064
+ "claude-opus-4-5": {
2065
+ maxTokens: 32768,
2066
+ contextWindow: 2e5,
2067
+ supportsImages: true,
2011
2068
  supportsPromptCache: true,
2012
- // Claude Code does report cache tokens
2013
- supportsReasoningEffort: false,
2014
- supportsReasoningBudget: false,
2015
- requiredReasoningBudget: false,
2016
- // Claude Code manages its own tools and temperature via the CLI
2017
- supportsNativeTools: false,
2018
- supportsTemperature: false
2069
+ supportsNativeTools: true,
2070
+ defaultToolProtocol: "native",
2071
+ supportsReasoningEffort: ["disable", "low", "medium", "high"],
2072
+ reasoningEffort: "medium",
2073
+ description: "Claude Opus 4.5 - Most capable with thinking"
2019
2074
  }
2020
2075
  };
2076
+ var claudeCodeDefaultModelId = "claude-sonnet-4-5";
2077
+ var MODEL_FAMILY_PATTERNS = [
2078
+ // Opus models (any version) → claude-opus-4-5
2079
+ { pattern: /opus/i, target: "claude-opus-4-5" },
2080
+ // Haiku models (any version) → claude-haiku-4-5
2081
+ { pattern: /haiku/i, target: "claude-haiku-4-5" },
2082
+ // Sonnet models (any version) → claude-sonnet-4-5
2083
+ { pattern: /sonnet/i, target: "claude-sonnet-4-5" }
2084
+ ];
2085
+ function normalizeClaudeCodeModelId(modelId) {
2086
+ if (Object.hasOwn(claudeCodeModels, modelId)) {
2087
+ return modelId;
2088
+ }
2089
+ const withoutDate = modelId.replace(DATE_SUFFIX_PATTERN, "");
2090
+ if (Object.hasOwn(claudeCodeModels, withoutDate)) {
2091
+ return withoutDate;
2092
+ }
2093
+ for (const { pattern, target } of MODEL_FAMILY_PATTERNS) {
2094
+ if (pattern.test(modelId)) {
2095
+ return target;
2096
+ }
2097
+ }
2098
+ return claudeCodeDefaultModelId;
2099
+ }
2100
+ var claudeCodeReasoningConfig = {
2101
+ low: { budgetTokens: 16e3 },
2102
+ medium: { budgetTokens: 32e3 },
2103
+ high: { budgetTokens: 64e3 }
2104
+ };
2021
2105
 
2022
2106
  // src/providers/deepseek.ts
2023
2107
  var deepSeekDefaultModelId = "deepseek-chat";
@@ -2048,6 +2132,7 @@ var deepSeekModels = {
2048
2132
  supportsPromptCache: true,
2049
2133
  supportsNativeTools: true,
2050
2134
  defaultToolProtocol: "native",
2135
+ preserveReasoning: true,
2051
2136
  inputPrice: 0.28,
2052
2137
  // $0.28 per million tokens (cache miss) - Updated Dec 9, 2025
2053
2138
  outputPrice: 0.42,
@@ -2059,7 +2144,7 @@ var deepSeekModels = {
2059
2144
  description: `DeepSeek-V3.2 (Thinking Mode) achieves performance comparable to OpenAI-o1 across math, code, and reasoning tasks. Supports Chain of Thought reasoning with up to 8K output tokens. Supports JSON output, tool calls, and chat prefix completion (beta).`
2060
2145
  }
2061
2146
  };
2062
- var DEEP_SEEK_DEFAULT_TEMPERATURE = 0;
2147
+ var DEEP_SEEK_DEFAULT_TEMPERATURE = 0.3;
2063
2148
 
2064
2149
  // src/providers/doubao.ts
2065
2150
  var doubaoDefaultModelId = "doubao-seed-1-6-250615";
@@ -2070,6 +2155,7 @@ var doubaoModels = {
2070
2155
  supportsImages: true,
2071
2156
  supportsPromptCache: true,
2072
2157
  supportsNativeTools: true,
2158
+ defaultToolProtocol: "native",
2073
2159
  inputPrice: 1e-4,
2074
2160
  // $0.0001 per million tokens (cache miss)
2075
2161
  outputPrice: 4e-4,
@@ -2086,6 +2172,7 @@ var doubaoModels = {
2086
2172
  supportsImages: true,
2087
2173
  supportsPromptCache: true,
2088
2174
  supportsNativeTools: true,
2175
+ defaultToolProtocol: "native",
2089
2176
  inputPrice: 2e-4,
2090
2177
  // $0.0002 per million tokens
2091
2178
  outputPrice: 8e-4,
@@ -2102,6 +2189,7 @@ var doubaoModels = {
2102
2189
  supportsImages: true,
2103
2190
  supportsPromptCache: true,
2104
2191
  supportsNativeTools: true,
2192
+ defaultToolProtocol: "native",
2105
2193
  inputPrice: 15e-5,
2106
2194
  // $0.00015 per million tokens
2107
2195
  outputPrice: 6e-4,
@@ -2178,6 +2266,7 @@ var fireworksModels = {
2178
2266
  supportsImages: false,
2179
2267
  supportsPromptCache: true,
2180
2268
  supportsNativeTools: true,
2269
+ defaultToolProtocol: "native",
2181
2270
  inputPrice: 0.6,
2182
2271
  outputPrice: 2.5,
2183
2272
  cacheReadsPrice: 0.15,
@@ -2189,6 +2278,7 @@ var fireworksModels = {
2189
2278
  supportsImages: false,
2190
2279
  supportsPromptCache: false,
2191
2280
  supportsNativeTools: true,
2281
+ defaultToolProtocol: "native",
2192
2282
  inputPrice: 0.6,
2193
2283
  outputPrice: 2.5,
2194
2284
  description: "Kimi K2 is a state-of-the-art mixture-of-experts (MoE) language model with 32 billion activated parameters and 1 trillion total parameters. Trained with the Muon optimizer, Kimi K2 achieves exceptional performance across frontier knowledge, reasoning, and coding tasks while being meticulously optimized for agentic capabilities."
@@ -2199,6 +2289,7 @@ var fireworksModels = {
2199
2289
  supportsImages: false,
2200
2290
  supportsPromptCache: false,
2201
2291
  supportsNativeTools: true,
2292
+ defaultToolProtocol: "native",
2202
2293
  inputPrice: 0.3,
2203
2294
  outputPrice: 1.2,
2204
2295
  description: "MiniMax M2 is a high-performance language model with 204.8K context window, optimized for long-context understanding and generation tasks."
@@ -2209,6 +2300,7 @@ var fireworksModels = {
2209
2300
  supportsImages: false,
2210
2301
  supportsPromptCache: false,
2211
2302
  supportsNativeTools: true,
2303
+ defaultToolProtocol: "native",
2212
2304
  inputPrice: 0.22,
2213
2305
  outputPrice: 0.88,
2214
2306
  description: "Latest Qwen3 thinking model, competitive against the best closed source models in Jul 2025."
@@ -2219,6 +2311,7 @@ var fireworksModels = {
2219
2311
  supportsImages: false,
2220
2312
  supportsPromptCache: false,
2221
2313
  supportsNativeTools: true,
2314
+ defaultToolProtocol: "native",
2222
2315
  inputPrice: 0.45,
2223
2316
  outputPrice: 1.8,
2224
2317
  description: "Qwen3's most agentic code model to date."
@@ -2229,6 +2322,7 @@ var fireworksModels = {
2229
2322
  supportsImages: false,
2230
2323
  supportsPromptCache: false,
2231
2324
  supportsNativeTools: true,
2325
+ defaultToolProtocol: "native",
2232
2326
  inputPrice: 3,
2233
2327
  outputPrice: 8,
2234
2328
  description: "05/28 updated checkpoint of Deepseek R1. Its overall performance is now approaching that of leading models, such as O3 and Gemini 2.5 Pro. Compared to the previous version, the upgraded model shows significant improvements in handling complex reasoning tasks, and this version also offers a reduced hallucination rate, enhanced support for function calling, and better experience for vibe coding. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
@@ -2239,6 +2333,7 @@ var fireworksModels = {
2239
2333
  supportsImages: false,
2240
2334
  supportsPromptCache: false,
2241
2335
  supportsNativeTools: true,
2336
+ defaultToolProtocol: "native",
2242
2337
  inputPrice: 0.9,
2243
2338
  outputPrice: 0.9,
2244
2339
  description: "A strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token from Deepseek. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
@@ -2249,6 +2344,7 @@ var fireworksModels = {
2249
2344
  supportsImages: false,
2250
2345
  supportsPromptCache: false,
2251
2346
  supportsNativeTools: true,
2347
+ defaultToolProtocol: "native",
2252
2348
  inputPrice: 0.56,
2253
2349
  outputPrice: 1.68,
2254
2350
  description: "DeepSeek v3.1 is an improved version of the v3 model with enhanced performance, better reasoning capabilities, and improved code generation. This Mixture-of-Experts (MoE) model maintains the same 671B total parameters with 37B activated per token."
@@ -2259,6 +2355,7 @@ var fireworksModels = {
2259
2355
  supportsImages: false,
2260
2356
  supportsPromptCache: false,
2261
2357
  supportsNativeTools: true,
2358
+ defaultToolProtocol: "native",
2262
2359
  inputPrice: 0.55,
2263
2360
  outputPrice: 2.19,
2264
2361
  description: "Z.ai GLM-4.5 with 355B total parameters and 32B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
@@ -2269,6 +2366,7 @@ var fireworksModels = {
2269
2366
  supportsImages: false,
2270
2367
  supportsPromptCache: false,
2271
2368
  supportsNativeTools: true,
2369
+ defaultToolProtocol: "native",
2272
2370
  inputPrice: 0.55,
2273
2371
  outputPrice: 2.19,
2274
2372
  description: "Z.ai GLM-4.5-Air with 106B total parameters and 12B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
@@ -2279,6 +2377,7 @@ var fireworksModels = {
2279
2377
  supportsImages: false,
2280
2378
  supportsPromptCache: false,
2281
2379
  supportsNativeTools: true,
2380
+ defaultToolProtocol: "native",
2282
2381
  inputPrice: 0.55,
2283
2382
  outputPrice: 2.19,
2284
2383
  description: "Z.ai GLM-4.6 is an advanced coding model with exceptional performance on complex programming tasks. Features improved reasoning capabilities and enhanced code generation quality, making it ideal for software development workflows."
@@ -2289,6 +2388,7 @@ var fireworksModels = {
2289
2388
  supportsImages: false,
2290
2389
  supportsPromptCache: false,
2291
2390
  supportsNativeTools: true,
2391
+ defaultToolProtocol: "native",
2292
2392
  inputPrice: 0.07,
2293
2393
  outputPrice: 0.3,
2294
2394
  description: "OpenAI gpt-oss-20b: Compact model for local/edge deployments. Optimized for low-latency and resource-constrained environments with chain-of-thought output, adjustable reasoning, and agentic workflows."
@@ -2299,6 +2399,7 @@ var fireworksModels = {
2299
2399
  supportsImages: false,
2300
2400
  supportsPromptCache: false,
2301
2401
  supportsNativeTools: true,
2402
+ defaultToolProtocol: "native",
2302
2403
  inputPrice: 0.15,
2303
2404
  outputPrice: 0.6,
2304
2405
  description: "OpenAI gpt-oss-120b: Production-grade, general-purpose model that fits on a single H100 GPU. Features complex reasoning, configurable effort, full chain-of-thought transparency, and supports function calling, tool use, and structured outputs."
@@ -2306,16 +2407,18 @@ var fireworksModels = {
2306
2407
  };
2307
2408
 
2308
2409
  // src/providers/gemini.ts
2309
- var geminiDefaultModelId = "gemini-2.5-pro";
2410
+ var geminiDefaultModelId = "gemini-3-pro-preview";
2310
2411
  var geminiModels = {
2311
2412
  "gemini-3-pro-preview": {
2312
2413
  maxTokens: 65536,
2313
2414
  contextWindow: 1048576,
2314
2415
  supportsImages: true,
2315
2416
  supportsNativeTools: true,
2417
+ defaultToolProtocol: "native",
2316
2418
  supportsPromptCache: true,
2317
2419
  supportsReasoningEffort: ["low", "high"],
2318
2420
  reasoningEffort: "low",
2421
+ includedTools: ["write_file", "edit_file"],
2319
2422
  supportsTemperature: true,
2320
2423
  defaultTemperature: 1,
2321
2424
  inputPrice: 4,
@@ -2333,13 +2436,32 @@ var geminiModels = {
2333
2436
  }
2334
2437
  ]
2335
2438
  },
2439
+ "gemini-3-flash-preview": {
2440
+ maxTokens: 65536,
2441
+ contextWindow: 1048576,
2442
+ supportsImages: true,
2443
+ supportsNativeTools: true,
2444
+ defaultToolProtocol: "native",
2445
+ supportsPromptCache: true,
2446
+ supportsReasoningEffort: ["minimal", "low", "medium", "high"],
2447
+ reasoningEffort: "medium",
2448
+ includedTools: ["write_file", "edit_file"],
2449
+ supportsTemperature: true,
2450
+ defaultTemperature: 1,
2451
+ inputPrice: 0.3,
2452
+ outputPrice: 2.5,
2453
+ cacheReadsPrice: 0.075,
2454
+ cacheWritesPrice: 1
2455
+ },
2336
2456
  // 2.5 Pro models
2337
2457
  "gemini-2.5-pro": {
2338
2458
  maxTokens: 64e3,
2339
2459
  contextWindow: 1048576,
2340
2460
  supportsImages: true,
2341
2461
  supportsNativeTools: true,
2462
+ defaultToolProtocol: "native",
2342
2463
  supportsPromptCache: true,
2464
+ includedTools: ["write_file", "edit_file"],
2343
2465
  inputPrice: 2.5,
2344
2466
  // This is the pricing for prompts above 200k tokens.
2345
2467
  outputPrice: 15,
@@ -2368,7 +2490,9 @@ var geminiModels = {
2368
2490
  contextWindow: 1048576,
2369
2491
  supportsImages: true,
2370
2492
  supportsNativeTools: true,
2493
+ defaultToolProtocol: "native",
2371
2494
  supportsPromptCache: true,
2495
+ includedTools: ["write_file", "edit_file"],
2372
2496
  inputPrice: 2.5,
2373
2497
  // This is the pricing for prompts above 200k tokens.
2374
2498
  outputPrice: 15,
@@ -2396,7 +2520,9 @@ var geminiModels = {
2396
2520
  contextWindow: 1048576,
2397
2521
  supportsImages: true,
2398
2522
  supportsNativeTools: true,
2523
+ defaultToolProtocol: "native",
2399
2524
  supportsPromptCache: true,
2525
+ includedTools: ["write_file", "edit_file"],
2400
2526
  inputPrice: 2.5,
2401
2527
  // This is the pricing for prompts above 200k tokens.
2402
2528
  outputPrice: 15,
@@ -2422,7 +2548,9 @@ var geminiModels = {
2422
2548
  contextWindow: 1048576,
2423
2549
  supportsImages: true,
2424
2550
  supportsNativeTools: true,
2551
+ defaultToolProtocol: "native",
2425
2552
  supportsPromptCache: true,
2553
+ includedTools: ["write_file", "edit_file"],
2426
2554
  inputPrice: 2.5,
2427
2555
  // This is the pricing for prompts above 200k tokens.
2428
2556
  outputPrice: 15,
@@ -2451,7 +2579,9 @@ var geminiModels = {
2451
2579
  contextWindow: 1048576,
2452
2580
  supportsImages: true,
2453
2581
  supportsNativeTools: true,
2582
+ defaultToolProtocol: "native",
2454
2583
  supportsPromptCache: true,
2584
+ includedTools: ["write_file", "edit_file"],
2455
2585
  inputPrice: 0.3,
2456
2586
  outputPrice: 2.5,
2457
2587
  cacheReadsPrice: 0.075,
@@ -2464,7 +2594,9 @@ var geminiModels = {
2464
2594
  contextWindow: 1048576,
2465
2595
  supportsImages: true,
2466
2596
  supportsNativeTools: true,
2597
+ defaultToolProtocol: "native",
2467
2598
  supportsPromptCache: true,
2599
+ includedTools: ["write_file", "edit_file"],
2468
2600
  inputPrice: 0.3,
2469
2601
  outputPrice: 2.5,
2470
2602
  cacheReadsPrice: 0.075,
@@ -2477,7 +2609,9 @@ var geminiModels = {
2477
2609
  contextWindow: 1048576,
2478
2610
  supportsImages: true,
2479
2611
  supportsNativeTools: true,
2612
+ defaultToolProtocol: "native",
2480
2613
  supportsPromptCache: true,
2614
+ includedTools: ["write_file", "edit_file"],
2481
2615
  inputPrice: 0.3,
2482
2616
  outputPrice: 2.5,
2483
2617
  cacheReadsPrice: 0.075,
@@ -2491,7 +2625,9 @@ var geminiModels = {
2491
2625
  contextWindow: 1048576,
2492
2626
  supportsImages: true,
2493
2627
  supportsNativeTools: true,
2628
+ defaultToolProtocol: "native",
2494
2629
  supportsPromptCache: true,
2630
+ includedTools: ["write_file", "edit_file"],
2495
2631
  inputPrice: 0.1,
2496
2632
  outputPrice: 0.4,
2497
2633
  cacheReadsPrice: 0.025,
@@ -2504,7 +2640,9 @@ var geminiModels = {
2504
2640
  contextWindow: 1048576,
2505
2641
  supportsImages: true,
2506
2642
  supportsNativeTools: true,
2643
+ defaultToolProtocol: "native",
2507
2644
  supportsPromptCache: true,
2645
+ includedTools: ["write_file", "edit_file"],
2508
2646
  inputPrice: 0.1,
2509
2647
  outputPrice: 0.4,
2510
2648
  cacheReadsPrice: 0.025,
@@ -2524,6 +2662,7 @@ var groqModels = {
2524
2662
  supportsImages: false,
2525
2663
  supportsPromptCache: false,
2526
2664
  supportsNativeTools: true,
2665
+ defaultToolProtocol: "native",
2527
2666
  inputPrice: 0.05,
2528
2667
  outputPrice: 0.08,
2529
2668
  description: "Meta Llama 3.1 8B Instant model, 128K context."
@@ -2534,6 +2673,7 @@ var groqModels = {
2534
2673
  supportsImages: false,
2535
2674
  supportsPromptCache: false,
2536
2675
  supportsNativeTools: true,
2676
+ defaultToolProtocol: "native",
2537
2677
  inputPrice: 0.59,
2538
2678
  outputPrice: 0.79,
2539
2679
  description: "Meta Llama 3.3 70B Versatile model, 128K context."
@@ -2544,6 +2684,7 @@ var groqModels = {
2544
2684
  supportsImages: false,
2545
2685
  supportsPromptCache: false,
2546
2686
  supportsNativeTools: true,
2687
+ defaultToolProtocol: "native",
2547
2688
  inputPrice: 0.11,
2548
2689
  outputPrice: 0.34,
2549
2690
  description: "Meta Llama 4 Scout 17B Instruct model, 128K context."
@@ -2581,6 +2722,7 @@ var groqModels = {
2581
2722
  supportsImages: false,
2582
2723
  supportsPromptCache: false,
2583
2724
  supportsNativeTools: true,
2725
+ defaultToolProtocol: "native",
2584
2726
  inputPrice: 0.29,
2585
2727
  outputPrice: 0.59,
2586
2728
  description: "Alibaba Qwen 3 32B model, 128K context."
@@ -2611,6 +2753,7 @@ var groqModels = {
2611
2753
  supportsImages: false,
2612
2754
  supportsPromptCache: true,
2613
2755
  supportsNativeTools: true,
2756
+ defaultToolProtocol: "native",
2614
2757
  inputPrice: 0.6,
2615
2758
  outputPrice: 2.5,
2616
2759
  cacheReadsPrice: 0.15,
@@ -2622,6 +2765,7 @@ var groqModels = {
2622
2765
  supportsImages: false,
2623
2766
  supportsPromptCache: false,
2624
2767
  supportsNativeTools: true,
2768
+ defaultToolProtocol: "native",
2625
2769
  inputPrice: 0.15,
2626
2770
  outputPrice: 0.75,
2627
2771
  description: "GPT-OSS 120B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 128 experts."
@@ -2632,6 +2776,7 @@ var groqModels = {
2632
2776
  supportsImages: false,
2633
2777
  supportsPromptCache: false,
2634
2778
  supportsNativeTools: true,
2779
+ defaultToolProtocol: "native",
2635
2780
  inputPrice: 0.1,
2636
2781
  outputPrice: 0.5,
2637
2782
  description: "GPT-OSS 20B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 32 experts."
@@ -2695,6 +2840,7 @@ var litellmDefaultModelInfo = {
2695
2840
  supportsImages: true,
2696
2841
  supportsPromptCache: true,
2697
2842
  supportsNativeTools: true,
2843
+ defaultToolProtocol: "native",
2698
2844
  inputPrice: 3,
2699
2845
  outputPrice: 15,
2700
2846
  cacheWritesPrice: 3.75,
@@ -2725,6 +2871,7 @@ var mistralModels = {
2725
2871
  supportsImages: true,
2726
2872
  supportsPromptCache: false,
2727
2873
  supportsNativeTools: true,
2874
+ defaultToolProtocol: "native",
2728
2875
  inputPrice: 2,
2729
2876
  outputPrice: 5
2730
2877
  },
@@ -2734,6 +2881,7 @@ var mistralModels = {
2734
2881
  supportsImages: true,
2735
2882
  supportsPromptCache: false,
2736
2883
  supportsNativeTools: true,
2884
+ defaultToolProtocol: "native",
2737
2885
  inputPrice: 0.4,
2738
2886
  outputPrice: 2
2739
2887
  },
@@ -2743,6 +2891,7 @@ var mistralModels = {
2743
2891
  supportsImages: true,
2744
2892
  supportsPromptCache: false,
2745
2893
  supportsNativeTools: true,
2894
+ defaultToolProtocol: "native",
2746
2895
  inputPrice: 0.4,
2747
2896
  outputPrice: 2
2748
2897
  },
@@ -2752,6 +2901,7 @@ var mistralModels = {
2752
2901
  supportsImages: false,
2753
2902
  supportsPromptCache: false,
2754
2903
  supportsNativeTools: true,
2904
+ defaultToolProtocol: "native",
2755
2905
  inputPrice: 0.3,
2756
2906
  outputPrice: 0.9
2757
2907
  },
@@ -2761,6 +2911,7 @@ var mistralModels = {
2761
2911
  supportsImages: false,
2762
2912
  supportsPromptCache: false,
2763
2913
  supportsNativeTools: true,
2914
+ defaultToolProtocol: "native",
2764
2915
  inputPrice: 2,
2765
2916
  outputPrice: 6
2766
2917
  },
@@ -2770,6 +2921,7 @@ var mistralModels = {
2770
2921
  supportsImages: false,
2771
2922
  supportsPromptCache: false,
2772
2923
  supportsNativeTools: true,
2924
+ defaultToolProtocol: "native",
2773
2925
  inputPrice: 0.1,
2774
2926
  outputPrice: 0.1
2775
2927
  },
@@ -2779,6 +2931,7 @@ var mistralModels = {
2779
2931
  supportsImages: false,
2780
2932
  supportsPromptCache: false,
2781
2933
  supportsNativeTools: true,
2934
+ defaultToolProtocol: "native",
2782
2935
  inputPrice: 0.04,
2783
2936
  outputPrice: 0.04
2784
2937
  },
@@ -2788,6 +2941,7 @@ var mistralModels = {
2788
2941
  supportsImages: false,
2789
2942
  supportsPromptCache: false,
2790
2943
  supportsNativeTools: true,
2944
+ defaultToolProtocol: "native",
2791
2945
  inputPrice: 0.2,
2792
2946
  outputPrice: 0.6
2793
2947
  },
@@ -2797,6 +2951,7 @@ var mistralModels = {
2797
2951
  supportsImages: true,
2798
2952
  supportsPromptCache: false,
2799
2953
  supportsNativeTools: true,
2954
+ defaultToolProtocol: "native",
2800
2955
  inputPrice: 2,
2801
2956
  outputPrice: 6
2802
2957
  }
@@ -2812,6 +2967,7 @@ var moonshotModels = {
2812
2967
  supportsImages: false,
2813
2968
  supportsPromptCache: true,
2814
2969
  supportsNativeTools: true,
2970
+ defaultToolProtocol: "native",
2815
2971
  inputPrice: 0.6,
2816
2972
  // $0.60 per million tokens (cache miss)
2817
2973
  outputPrice: 2.5,
@@ -2828,6 +2984,7 @@ var moonshotModels = {
2828
2984
  supportsImages: false,
2829
2985
  supportsPromptCache: true,
2830
2986
  supportsNativeTools: true,
2987
+ defaultToolProtocol: "native",
2831
2988
  inputPrice: 0.6,
2832
2989
  outputPrice: 2.5,
2833
2990
  cacheReadsPrice: 0.15,
@@ -2839,6 +2996,7 @@ var moonshotModels = {
2839
2996
  supportsImages: false,
2840
2997
  supportsPromptCache: true,
2841
2998
  supportsNativeTools: true,
2999
+ defaultToolProtocol: "native",
2842
3000
  inputPrice: 2.4,
2843
3001
  // $2.40 per million tokens (cache miss)
2844
3002
  outputPrice: 10,
@@ -2858,6 +3016,7 @@ var moonshotModels = {
2858
3016
  // Text-only (no image/vision support)
2859
3017
  supportsPromptCache: true,
2860
3018
  supportsNativeTools: true,
3019
+ defaultToolProtocol: "native",
2861
3020
  inputPrice: 0.6,
2862
3021
  // $0.60 per million tokens (cache miss)
2863
3022
  outputPrice: 2.5,
@@ -2897,6 +3056,7 @@ var openAiNativeModels = {
2897
3056
  maxTokens: 128e3,
2898
3057
  contextWindow: 4e5,
2899
3058
  supportsNativeTools: true,
3059
+ defaultToolProtocol: "native",
2900
3060
  includedTools: ["apply_patch"],
2901
3061
  excludedTools: ["apply_diff", "write_to_file"],
2902
3062
  supportsImages: true,
@@ -2915,6 +3075,7 @@ var openAiNativeModels = {
2915
3075
  maxTokens: 128e3,
2916
3076
  contextWindow: 4e5,
2917
3077
  supportsNativeTools: true,
3078
+ defaultToolProtocol: "native",
2918
3079
  includedTools: ["apply_patch"],
2919
3080
  excludedTools: ["apply_diff", "write_to_file"],
2920
3081
  supportsImages: true,
@@ -2937,6 +3098,7 @@ var openAiNativeModels = {
2937
3098
  maxTokens: 16384,
2938
3099
  contextWindow: 128e3,
2939
3100
  supportsNativeTools: true,
3101
+ defaultToolProtocol: "native",
2940
3102
  includedTools: ["apply_patch"],
2941
3103
  excludedTools: ["apply_diff", "write_to_file"],
2942
3104
  supportsImages: true,
@@ -2950,6 +3112,7 @@ var openAiNativeModels = {
2950
3112
  maxTokens: 128e3,
2951
3113
  contextWindow: 4e5,
2952
3114
  supportsNativeTools: true,
3115
+ defaultToolProtocol: "native",
2953
3116
  includedTools: ["apply_patch"],
2954
3117
  excludedTools: ["apply_diff", "write_to_file"],
2955
3118
  supportsImages: true,
@@ -2972,6 +3135,7 @@ var openAiNativeModels = {
2972
3135
  maxTokens: 128e3,
2973
3136
  contextWindow: 4e5,
2974
3137
  supportsNativeTools: true,
3138
+ defaultToolProtocol: "native",
2975
3139
  includedTools: ["apply_patch"],
2976
3140
  excludedTools: ["apply_diff", "write_to_file"],
2977
3141
  supportsImages: true,
@@ -2990,6 +3154,7 @@ var openAiNativeModels = {
2990
3154
  maxTokens: 128e3,
2991
3155
  contextWindow: 4e5,
2992
3156
  supportsNativeTools: true,
3157
+ defaultToolProtocol: "native",
2993
3158
  includedTools: ["apply_patch"],
2994
3159
  excludedTools: ["apply_diff", "write_to_file"],
2995
3160
  supportsImages: true,
@@ -3007,6 +3172,7 @@ var openAiNativeModels = {
3007
3172
  maxTokens: 128e3,
3008
3173
  contextWindow: 4e5,
3009
3174
  supportsNativeTools: true,
3175
+ defaultToolProtocol: "native",
3010
3176
  includedTools: ["apply_patch"],
3011
3177
  excludedTools: ["apply_diff", "write_to_file"],
3012
3178
  supportsImages: true,
@@ -3028,6 +3194,7 @@ var openAiNativeModels = {
3028
3194
  maxTokens: 128e3,
3029
3195
  contextWindow: 4e5,
3030
3196
  supportsNativeTools: true,
3197
+ defaultToolProtocol: "native",
3031
3198
  includedTools: ["apply_patch"],
3032
3199
  excludedTools: ["apply_diff", "write_to_file"],
3033
3200
  supportsImages: true,
@@ -3049,6 +3216,7 @@ var openAiNativeModels = {
3049
3216
  maxTokens: 128e3,
3050
3217
  contextWindow: 4e5,
3051
3218
  supportsNativeTools: true,
3219
+ defaultToolProtocol: "native",
3052
3220
  includedTools: ["apply_patch"],
3053
3221
  excludedTools: ["apply_diff", "write_to_file"],
3054
3222
  supportsImages: true,
@@ -3066,6 +3234,7 @@ var openAiNativeModels = {
3066
3234
  maxTokens: 128e3,
3067
3235
  contextWindow: 4e5,
3068
3236
  supportsNativeTools: true,
3237
+ defaultToolProtocol: "native",
3069
3238
  includedTools: ["apply_patch"],
3070
3239
  excludedTools: ["apply_diff", "write_to_file"],
3071
3240
  supportsImages: true,
@@ -3084,6 +3253,7 @@ var openAiNativeModels = {
3084
3253
  maxTokens: 128e3,
3085
3254
  contextWindow: 4e5,
3086
3255
  supportsNativeTools: true,
3256
+ defaultToolProtocol: "native",
3087
3257
  includedTools: ["apply_patch"],
3088
3258
  excludedTools: ["apply_diff", "write_to_file"],
3089
3259
  supportsImages: true,
@@ -3097,6 +3267,7 @@ var openAiNativeModels = {
3097
3267
  maxTokens: 32768,
3098
3268
  contextWindow: 1047576,
3099
3269
  supportsNativeTools: true,
3270
+ defaultToolProtocol: "native",
3100
3271
  includedTools: ["apply_patch"],
3101
3272
  excludedTools: ["apply_diff", "write_to_file"],
3102
3273
  supportsImages: true,
@@ -3113,6 +3284,7 @@ var openAiNativeModels = {
3113
3284
  maxTokens: 32768,
3114
3285
  contextWindow: 1047576,
3115
3286
  supportsNativeTools: true,
3287
+ defaultToolProtocol: "native",
3116
3288
  includedTools: ["apply_patch"],
3117
3289
  excludedTools: ["apply_diff", "write_to_file"],
3118
3290
  supportsImages: true,
@@ -3129,6 +3301,7 @@ var openAiNativeModels = {
3129
3301
  maxTokens: 32768,
3130
3302
  contextWindow: 1047576,
3131
3303
  supportsNativeTools: true,
3304
+ defaultToolProtocol: "native",
3132
3305
  includedTools: ["apply_patch"],
3133
3306
  excludedTools: ["apply_diff", "write_to_file"],
3134
3307
  supportsImages: true,
@@ -3145,6 +3318,7 @@ var openAiNativeModels = {
3145
3318
  maxTokens: 1e5,
3146
3319
  contextWindow: 2e5,
3147
3320
  supportsNativeTools: true,
3321
+ defaultToolProtocol: "native",
3148
3322
  supportsImages: true,
3149
3323
  supportsPromptCache: true,
3150
3324
  inputPrice: 2,
@@ -3162,6 +3336,7 @@ var openAiNativeModels = {
3162
3336
  maxTokens: 1e5,
3163
3337
  contextWindow: 2e5,
3164
3338
  supportsNativeTools: true,
3339
+ defaultToolProtocol: "native",
3165
3340
  supportsImages: true,
3166
3341
  supportsPromptCache: true,
3167
3342
  inputPrice: 2,
@@ -3174,6 +3349,7 @@ var openAiNativeModels = {
3174
3349
  maxTokens: 1e5,
3175
3350
  contextWindow: 2e5,
3176
3351
  supportsNativeTools: true,
3352
+ defaultToolProtocol: "native",
3177
3353
  supportsImages: true,
3178
3354
  supportsPromptCache: true,
3179
3355
  inputPrice: 2,
@@ -3186,6 +3362,7 @@ var openAiNativeModels = {
3186
3362
  maxTokens: 1e5,
3187
3363
  contextWindow: 2e5,
3188
3364
  supportsNativeTools: true,
3365
+ defaultToolProtocol: "native",
3189
3366
  supportsImages: true,
3190
3367
  supportsPromptCache: true,
3191
3368
  inputPrice: 1.1,
@@ -3203,6 +3380,7 @@ var openAiNativeModels = {
3203
3380
  maxTokens: 1e5,
3204
3381
  contextWindow: 2e5,
3205
3382
  supportsNativeTools: true,
3383
+ defaultToolProtocol: "native",
3206
3384
  supportsImages: true,
3207
3385
  supportsPromptCache: true,
3208
3386
  inputPrice: 1.1,
@@ -3215,6 +3393,7 @@ var openAiNativeModels = {
3215
3393
  maxTokens: 1e5,
3216
3394
  contextWindow: 2e5,
3217
3395
  supportsNativeTools: true,
3396
+ defaultToolProtocol: "native",
3218
3397
  supportsImages: true,
3219
3398
  supportsPromptCache: true,
3220
3399
  inputPrice: 1.1,
@@ -3227,6 +3406,7 @@ var openAiNativeModels = {
3227
3406
  maxTokens: 1e5,
3228
3407
  contextWindow: 2e5,
3229
3408
  supportsNativeTools: true,
3409
+ defaultToolProtocol: "native",
3230
3410
  supportsImages: false,
3231
3411
  supportsPromptCache: true,
3232
3412
  inputPrice: 1.1,
@@ -3240,6 +3420,7 @@ var openAiNativeModels = {
3240
3420
  maxTokens: 1e5,
3241
3421
  contextWindow: 2e5,
3242
3422
  supportsNativeTools: true,
3423
+ defaultToolProtocol: "native",
3243
3424
  supportsImages: false,
3244
3425
  supportsPromptCache: true,
3245
3426
  inputPrice: 1.1,
@@ -3252,6 +3433,7 @@ var openAiNativeModels = {
3252
3433
  maxTokens: 1e5,
3253
3434
  contextWindow: 2e5,
3254
3435
  supportsNativeTools: true,
3436
+ defaultToolProtocol: "native",
3255
3437
  supportsImages: false,
3256
3438
  supportsPromptCache: true,
3257
3439
  inputPrice: 1.1,
@@ -3264,6 +3446,7 @@ var openAiNativeModels = {
3264
3446
  maxTokens: 1e5,
3265
3447
  contextWindow: 2e5,
3266
3448
  supportsNativeTools: true,
3449
+ defaultToolProtocol: "native",
3267
3450
  supportsImages: true,
3268
3451
  supportsPromptCache: true,
3269
3452
  inputPrice: 15,
@@ -3275,6 +3458,7 @@ var openAiNativeModels = {
3275
3458
  maxTokens: 32768,
3276
3459
  contextWindow: 128e3,
3277
3460
  supportsNativeTools: true,
3461
+ defaultToolProtocol: "native",
3278
3462
  supportsImages: true,
3279
3463
  supportsPromptCache: true,
3280
3464
  inputPrice: 15,
@@ -3286,6 +3470,7 @@ var openAiNativeModels = {
3286
3470
  maxTokens: 65536,
3287
3471
  contextWindow: 128e3,
3288
3472
  supportsNativeTools: true,
3473
+ defaultToolProtocol: "native",
3289
3474
  supportsImages: true,
3290
3475
  supportsPromptCache: true,
3291
3476
  inputPrice: 1.1,
@@ -3297,6 +3482,7 @@ var openAiNativeModels = {
3297
3482
  maxTokens: 16384,
3298
3483
  contextWindow: 128e3,
3299
3484
  supportsNativeTools: true,
3485
+ defaultToolProtocol: "native",
3300
3486
  supportsImages: true,
3301
3487
  supportsPromptCache: true,
3302
3488
  inputPrice: 2.5,
@@ -3311,6 +3497,7 @@ var openAiNativeModels = {
3311
3497
  maxTokens: 16384,
3312
3498
  contextWindow: 128e3,
3313
3499
  supportsNativeTools: true,
3500
+ defaultToolProtocol: "native",
3314
3501
  supportsImages: true,
3315
3502
  supportsPromptCache: true,
3316
3503
  inputPrice: 0.15,
@@ -3325,6 +3512,7 @@ var openAiNativeModels = {
3325
3512
  maxTokens: 16384,
3326
3513
  contextWindow: 2e5,
3327
3514
  supportsNativeTools: true,
3515
+ defaultToolProtocol: "native",
3328
3516
  supportsImages: false,
3329
3517
  supportsPromptCache: false,
3330
3518
  inputPrice: 1.5,
@@ -3338,6 +3526,7 @@ var openAiNativeModels = {
3338
3526
  maxTokens: 128e3,
3339
3527
  contextWindow: 4e5,
3340
3528
  supportsNativeTools: true,
3529
+ defaultToolProtocol: "native",
3341
3530
  includedTools: ["apply_patch"],
3342
3531
  excludedTools: ["apply_diff", "write_to_file"],
3343
3532
  supportsImages: true,
@@ -3359,6 +3548,7 @@ var openAiNativeModels = {
3359
3548
  maxTokens: 128e3,
3360
3549
  contextWindow: 4e5,
3361
3550
  supportsNativeTools: true,
3551
+ defaultToolProtocol: "native",
3362
3552
  includedTools: ["apply_patch"],
3363
3553
  excludedTools: ["apply_diff", "write_to_file"],
3364
3554
  supportsImages: true,
@@ -3380,6 +3570,7 @@ var openAiNativeModels = {
3380
3570
  maxTokens: 128e3,
3381
3571
  contextWindow: 4e5,
3382
3572
  supportsNativeTools: true,
3573
+ defaultToolProtocol: "native",
3383
3574
  includedTools: ["apply_patch"],
3384
3575
  excludedTools: ["apply_diff", "write_to_file"],
3385
3576
  supportsImages: true,
@@ -3402,7 +3593,8 @@ var openAiModelInfoSaneDefaults = {
3402
3593
  supportsPromptCache: false,
3403
3594
  inputPrice: 0,
3404
3595
  outputPrice: 0,
3405
- supportsNativeTools: true
3596
+ supportsNativeTools: true,
3597
+ defaultToolProtocol: "native"
3406
3598
  };
3407
3599
  var azureOpenAiDefaultApiVersion = "2024-08-01-preview";
3408
3600
  var OPENAI_NATIVE_DEFAULT_TEMPERATURE = 0;
@@ -3515,6 +3707,8 @@ var requestyDefaultModelInfo = {
3515
3707
  contextWindow: 2e5,
3516
3708
  supportsImages: true,
3517
3709
  supportsPromptCache: true,
3710
+ supportsNativeTools: true,
3711
+ defaultToolProtocol: "native",
3518
3712
  inputPrice: 3,
3519
3713
  outputPrice: 15,
3520
3714
  cacheWritesPrice: 3.75,
@@ -3571,6 +3765,7 @@ var sambaNovaModels = {
3571
3765
  supportsImages: false,
3572
3766
  supportsPromptCache: false,
3573
3767
  supportsNativeTools: true,
3768
+ defaultToolProtocol: "native",
3574
3769
  inputPrice: 0.1,
3575
3770
  outputPrice: 0.2,
3576
3771
  description: "Meta Llama 3.1 8B Instruct model with 16K context window."
@@ -3581,6 +3776,7 @@ var sambaNovaModels = {
3581
3776
  supportsImages: false,
3582
3777
  supportsPromptCache: false,
3583
3778
  supportsNativeTools: true,
3779
+ defaultToolProtocol: "native",
3584
3780
  inputPrice: 0.6,
3585
3781
  outputPrice: 1.2,
3586
3782
  description: "Meta Llama 3.3 70B Instruct model with 128K context window."
@@ -3592,6 +3788,7 @@ var sambaNovaModels = {
3592
3788
  supportsPromptCache: false,
3593
3789
  supportsReasoningBudget: true,
3594
3790
  supportsNativeTools: true,
3791
+ defaultToolProtocol: "native",
3595
3792
  inputPrice: 5,
3596
3793
  outputPrice: 7,
3597
3794
  description: "DeepSeek R1 reasoning model with 32K context window."
@@ -3602,6 +3799,7 @@ var sambaNovaModels = {
3602
3799
  supportsImages: false,
3603
3800
  supportsPromptCache: false,
3604
3801
  supportsNativeTools: true,
3802
+ defaultToolProtocol: "native",
3605
3803
  inputPrice: 3,
3606
3804
  outputPrice: 4.5,
3607
3805
  description: "DeepSeek V3 model with 32K context window."
@@ -3612,6 +3810,7 @@ var sambaNovaModels = {
3612
3810
  supportsImages: false,
3613
3811
  supportsPromptCache: false,
3614
3812
  supportsNativeTools: true,
3813
+ defaultToolProtocol: "native",
3615
3814
  inputPrice: 3,
3616
3815
  outputPrice: 4.5,
3617
3816
  description: "DeepSeek V3.1 model with 32K context window."
@@ -3631,6 +3830,7 @@ var sambaNovaModels = {
3631
3830
  supportsImages: true,
3632
3831
  supportsPromptCache: false,
3633
3832
  supportsNativeTools: true,
3833
+ defaultToolProtocol: "native",
3634
3834
  inputPrice: 0.63,
3635
3835
  outputPrice: 1.8,
3636
3836
  description: "Meta Llama 4 Maverick 17B 128E Instruct model with 128K context window."
@@ -3650,6 +3850,7 @@ var sambaNovaModels = {
3650
3850
  supportsImages: false,
3651
3851
  supportsPromptCache: false,
3652
3852
  supportsNativeTools: true,
3853
+ defaultToolProtocol: "native",
3653
3854
  inputPrice: 0.4,
3654
3855
  outputPrice: 0.8,
3655
3856
  description: "Alibaba Qwen 3 32B model with 8K context window."
@@ -3660,6 +3861,7 @@ var sambaNovaModels = {
3660
3861
  supportsImages: false,
3661
3862
  supportsPromptCache: false,
3662
3863
  supportsNativeTools: true,
3864
+ defaultToolProtocol: "native",
3663
3865
  inputPrice: 0.22,
3664
3866
  outputPrice: 0.59,
3665
3867
  description: "OpenAI gpt oss 120b model with 128k context window."
@@ -3688,9 +3890,11 @@ var vertexModels = {
3688
3890
  contextWindow: 1048576,
3689
3891
  supportsImages: true,
3690
3892
  supportsNativeTools: true,
3893
+ defaultToolProtocol: "native",
3691
3894
  supportsPromptCache: true,
3692
3895
  supportsReasoningEffort: ["low", "high"],
3693
3896
  reasoningEffort: "low",
3897
+ includedTools: ["write_file", "edit_file"],
3694
3898
  supportsTemperature: true,
3695
3899
  defaultTemperature: 1,
3696
3900
  inputPrice: 4,
@@ -3708,12 +3912,31 @@ var vertexModels = {
3708
3912
  }
3709
3913
  ]
3710
3914
  },
3915
+ "gemini-3-flash-preview": {
3916
+ maxTokens: 65536,
3917
+ contextWindow: 1048576,
3918
+ supportsImages: true,
3919
+ supportsNativeTools: true,
3920
+ defaultToolProtocol: "native",
3921
+ supportsPromptCache: true,
3922
+ supportsReasoningEffort: ["minimal", "low", "medium", "high"],
3923
+ reasoningEffort: "medium",
3924
+ includedTools: ["write_file", "edit_file"],
3925
+ supportsTemperature: true,
3926
+ defaultTemperature: 1,
3927
+ inputPrice: 0.3,
3928
+ outputPrice: 2.5,
3929
+ cacheReadsPrice: 0.075,
3930
+ cacheWritesPrice: 1
3931
+ },
3711
3932
  "gemini-2.5-flash-preview-05-20:thinking": {
3712
3933
  maxTokens: 65535,
3713
3934
  contextWindow: 1048576,
3714
3935
  supportsImages: true,
3715
3936
  supportsNativeTools: true,
3937
+ defaultToolProtocol: "native",
3716
3938
  supportsPromptCache: true,
3939
+ includedTools: ["write_file", "edit_file"],
3717
3940
  inputPrice: 0.15,
3718
3941
  outputPrice: 3.5,
3719
3942
  maxThinkingTokens: 24576,
@@ -3725,7 +3948,9 @@ var vertexModels = {
3725
3948
  contextWindow: 1048576,
3726
3949
  supportsImages: true,
3727
3950
  supportsNativeTools: true,
3951
+ defaultToolProtocol: "native",
3728
3952
  supportsPromptCache: true,
3953
+ includedTools: ["write_file", "edit_file"],
3729
3954
  inputPrice: 0.15,
3730
3955
  outputPrice: 0.6
3731
3956
  },
@@ -3734,7 +3959,9 @@ var vertexModels = {
3734
3959
  contextWindow: 1048576,
3735
3960
  supportsImages: true,
3736
3961
  supportsNativeTools: true,
3962
+ defaultToolProtocol: "native",
3737
3963
  supportsPromptCache: true,
3964
+ includedTools: ["write_file", "edit_file"],
3738
3965
  inputPrice: 0.3,
3739
3966
  outputPrice: 2.5,
3740
3967
  cacheReadsPrice: 0.075,
@@ -3747,7 +3974,9 @@ var vertexModels = {
3747
3974
  contextWindow: 1048576,
3748
3975
  supportsImages: true,
3749
3976
  supportsNativeTools: true,
3977
+ defaultToolProtocol: "native",
3750
3978
  supportsPromptCache: false,
3979
+ includedTools: ["write_file", "edit_file"],
3751
3980
  inputPrice: 0.15,
3752
3981
  outputPrice: 3.5,
3753
3982
  maxThinkingTokens: 24576,
@@ -3759,7 +3988,9 @@ var vertexModels = {
3759
3988
  contextWindow: 1048576,
3760
3989
  supportsImages: true,
3761
3990
  supportsNativeTools: true,
3991
+ defaultToolProtocol: "native",
3762
3992
  supportsPromptCache: false,
3993
+ includedTools: ["write_file", "edit_file"],
3763
3994
  inputPrice: 0.15,
3764
3995
  outputPrice: 0.6
3765
3996
  },
@@ -3768,7 +3999,9 @@ var vertexModels = {
3768
3999
  contextWindow: 1048576,
3769
4000
  supportsImages: true,
3770
4001
  supportsNativeTools: true,
4002
+ defaultToolProtocol: "native",
3771
4003
  supportsPromptCache: true,
4004
+ includedTools: ["write_file", "edit_file"],
3772
4005
  inputPrice: 2.5,
3773
4006
  outputPrice: 15
3774
4007
  },
@@ -3777,7 +4010,9 @@ var vertexModels = {
3777
4010
  contextWindow: 1048576,
3778
4011
  supportsImages: true,
3779
4012
  supportsNativeTools: true,
4013
+ defaultToolProtocol: "native",
3780
4014
  supportsPromptCache: true,
4015
+ includedTools: ["write_file", "edit_file"],
3781
4016
  inputPrice: 2.5,
3782
4017
  outputPrice: 15
3783
4018
  },
@@ -3786,7 +4021,9 @@ var vertexModels = {
3786
4021
  contextWindow: 1048576,
3787
4022
  supportsImages: true,
3788
4023
  supportsNativeTools: true,
4024
+ defaultToolProtocol: "native",
3789
4025
  supportsPromptCache: true,
4026
+ includedTools: ["write_file", "edit_file"],
3790
4027
  inputPrice: 2.5,
3791
4028
  outputPrice: 15,
3792
4029
  maxThinkingTokens: 32768,
@@ -3797,7 +4034,9 @@ var vertexModels = {
3797
4034
  contextWindow: 1048576,
3798
4035
  supportsImages: true,
3799
4036
  supportsNativeTools: true,
4037
+ defaultToolProtocol: "native",
3800
4038
  supportsPromptCache: true,
4039
+ includedTools: ["write_file", "edit_file"],
3801
4040
  inputPrice: 2.5,
3802
4041
  outputPrice: 15,
3803
4042
  maxThinkingTokens: 32768,
@@ -3823,7 +4062,9 @@ var vertexModels = {
3823
4062
  contextWindow: 1048576,
3824
4063
  supportsImages: true,
3825
4064
  supportsNativeTools: true,
4065
+ defaultToolProtocol: "native",
3826
4066
  supportsPromptCache: false,
4067
+ includedTools: ["write_file", "edit_file"],
3827
4068
  inputPrice: 0,
3828
4069
  outputPrice: 0
3829
4070
  },
@@ -3832,7 +4073,9 @@ var vertexModels = {
3832
4073
  contextWindow: 2097152,
3833
4074
  supportsImages: true,
3834
4075
  supportsNativeTools: true,
4076
+ defaultToolProtocol: "native",
3835
4077
  supportsPromptCache: false,
4078
+ includedTools: ["write_file", "edit_file"],
3836
4079
  inputPrice: 0,
3837
4080
  outputPrice: 0
3838
4081
  },
@@ -3841,7 +4084,9 @@ var vertexModels = {
3841
4084
  contextWindow: 1048576,
3842
4085
  supportsImages: true,
3843
4086
  supportsNativeTools: true,
4087
+ defaultToolProtocol: "native",
3844
4088
  supportsPromptCache: true,
4089
+ includedTools: ["write_file", "edit_file"],
3845
4090
  inputPrice: 0.15,
3846
4091
  outputPrice: 0.6
3847
4092
  },
@@ -3850,7 +4095,9 @@ var vertexModels = {
3850
4095
  contextWindow: 1048576,
3851
4096
  supportsImages: true,
3852
4097
  supportsNativeTools: true,
4098
+ defaultToolProtocol: "native",
3853
4099
  supportsPromptCache: false,
4100
+ includedTools: ["write_file", "edit_file"],
3854
4101
  inputPrice: 0.075,
3855
4102
  outputPrice: 0.3
3856
4103
  },
@@ -3859,7 +4106,9 @@ var vertexModels = {
3859
4106
  contextWindow: 32768,
3860
4107
  supportsImages: true,
3861
4108
  supportsNativeTools: true,
4109
+ defaultToolProtocol: "native",
3862
4110
  supportsPromptCache: false,
4111
+ includedTools: ["write_file", "edit_file"],
3863
4112
  inputPrice: 0,
3864
4113
  outputPrice: 0
3865
4114
  },
@@ -3868,7 +4117,9 @@ var vertexModels = {
3868
4117
  contextWindow: 1048576,
3869
4118
  supportsImages: true,
3870
4119
  supportsNativeTools: true,
4120
+ defaultToolProtocol: "native",
3871
4121
  supportsPromptCache: true,
4122
+ includedTools: ["write_file", "edit_file"],
3872
4123
  inputPrice: 0.075,
3873
4124
  outputPrice: 0.3
3874
4125
  },
@@ -3877,7 +4128,9 @@ var vertexModels = {
3877
4128
  contextWindow: 2097152,
3878
4129
  supportsImages: true,
3879
4130
  supportsNativeTools: true,
4131
+ defaultToolProtocol: "native",
3880
4132
  supportsPromptCache: false,
4133
+ includedTools: ["write_file", "edit_file"],
3881
4134
  inputPrice: 1.25,
3882
4135
  outputPrice: 5
3883
4136
  },
@@ -4023,7 +4276,9 @@ var vertexModels = {
4023
4276
  contextWindow: 1048576,
4024
4277
  supportsImages: true,
4025
4278
  supportsNativeTools: true,
4279
+ defaultToolProtocol: "native",
4026
4280
  supportsPromptCache: true,
4281
+ includedTools: ["write_file", "edit_file"],
4027
4282
  inputPrice: 0.1,
4028
4283
  outputPrice: 0.4,
4029
4284
  cacheReadsPrice: 0.025,
@@ -4325,6 +4580,7 @@ var xaiModels = {
4325
4580
  supportsImages: true,
4326
4581
  supportsPromptCache: true,
4327
4582
  supportsNativeTools: true,
4583
+ defaultToolProtocol: "native",
4328
4584
  inputPrice: 0.2,
4329
4585
  outputPrice: 1.5,
4330
4586
  cacheWritesPrice: 0.02,
@@ -4339,6 +4595,7 @@ var xaiModels = {
4339
4595
  supportsImages: true,
4340
4596
  supportsPromptCache: true,
4341
4597
  supportsNativeTools: true,
4598
+ defaultToolProtocol: "native",
4342
4599
  inputPrice: 0.2,
4343
4600
  outputPrice: 0.5,
4344
4601
  cacheWritesPrice: 0.05,
@@ -4353,6 +4610,7 @@ var xaiModels = {
4353
4610
  supportsImages: true,
4354
4611
  supportsPromptCache: true,
4355
4612
  supportsNativeTools: true,
4613
+ defaultToolProtocol: "native",
4356
4614
  inputPrice: 0.2,
4357
4615
  outputPrice: 0.5,
4358
4616
  cacheWritesPrice: 0.05,
@@ -4367,6 +4625,7 @@ var xaiModels = {
4367
4625
  supportsImages: true,
4368
4626
  supportsPromptCache: true,
4369
4627
  supportsNativeTools: true,
4628
+ defaultToolProtocol: "native",
4370
4629
  inputPrice: 0.2,
4371
4630
  outputPrice: 0.5,
4372
4631
  cacheWritesPrice: 0.05,
@@ -4381,6 +4640,7 @@ var xaiModels = {
4381
4640
  supportsImages: true,
4382
4641
  supportsPromptCache: true,
4383
4642
  supportsNativeTools: true,
4643
+ defaultToolProtocol: "native",
4384
4644
  inputPrice: 0.2,
4385
4645
  outputPrice: 0.5,
4386
4646
  cacheWritesPrice: 0.05,
@@ -4395,6 +4655,7 @@ var xaiModels = {
4395
4655
  supportsImages: true,
4396
4656
  supportsPromptCache: true,
4397
4657
  supportsNativeTools: true,
4658
+ defaultToolProtocol: "native",
4398
4659
  inputPrice: 3,
4399
4660
  outputPrice: 15,
4400
4661
  cacheWritesPrice: 0.75,
@@ -4409,6 +4670,7 @@ var xaiModels = {
4409
4670
  supportsImages: true,
4410
4671
  supportsPromptCache: true,
4411
4672
  supportsNativeTools: true,
4673
+ defaultToolProtocol: "native",
4412
4674
  inputPrice: 0.3,
4413
4675
  outputPrice: 0.5,
4414
4676
  cacheWritesPrice: 0.07,
@@ -4425,6 +4687,7 @@ var xaiModels = {
4425
4687
  supportsImages: true,
4426
4688
  supportsPromptCache: true,
4427
4689
  supportsNativeTools: true,
4690
+ defaultToolProtocol: "native",
4428
4691
  inputPrice: 3,
4429
4692
  outputPrice: 15,
4430
4693
  cacheWritesPrice: 0.75,
@@ -4538,6 +4801,7 @@ var internationalZAiModels = {
4538
4801
  supportsImages: false,
4539
4802
  supportsPromptCache: true,
4540
4803
  supportsNativeTools: true,
4804
+ defaultToolProtocol: "native",
4541
4805
  inputPrice: 0.6,
4542
4806
  outputPrice: 2.2,
4543
4807
  cacheWritesPrice: 0,
@@ -4550,6 +4814,7 @@ var internationalZAiModels = {
4550
4814
  supportsImages: false,
4551
4815
  supportsPromptCache: true,
4552
4816
  supportsNativeTools: true,
4817
+ defaultToolProtocol: "native",
4553
4818
  inputPrice: 0.2,
4554
4819
  outputPrice: 1.1,
4555
4820
  cacheWritesPrice: 0,
@@ -4562,6 +4827,7 @@ var internationalZAiModels = {
4562
4827
  supportsImages: false,
4563
4828
  supportsPromptCache: true,
4564
4829
  supportsNativeTools: true,
4830
+ defaultToolProtocol: "native",
4565
4831
  inputPrice: 2.2,
4566
4832
  outputPrice: 8.9,
4567
4833
  cacheWritesPrice: 0,
@@ -4574,6 +4840,7 @@ var internationalZAiModels = {
4574
4840
  supportsImages: false,
4575
4841
  supportsPromptCache: true,
4576
4842
  supportsNativeTools: true,
4843
+ defaultToolProtocol: "native",
4577
4844
  inputPrice: 1.1,
4578
4845
  outputPrice: 4.5,
4579
4846
  cacheWritesPrice: 0,
@@ -4586,6 +4853,7 @@ var internationalZAiModels = {
4586
4853
  supportsImages: false,
4587
4854
  supportsPromptCache: true,
4588
4855
  supportsNativeTools: true,
4856
+ defaultToolProtocol: "native",
4589
4857
  inputPrice: 0,
4590
4858
  outputPrice: 0,
4591
4859
  cacheWritesPrice: 0,
@@ -4598,6 +4866,7 @@ var internationalZAiModels = {
4598
4866
  supportsImages: true,
4599
4867
  supportsPromptCache: true,
4600
4868
  supportsNativeTools: true,
4869
+ defaultToolProtocol: "native",
4601
4870
  inputPrice: 0.6,
4602
4871
  outputPrice: 1.8,
4603
4872
  cacheWritesPrice: 0,
@@ -4610,6 +4879,7 @@ var internationalZAiModels = {
4610
4879
  supportsImages: false,
4611
4880
  supportsPromptCache: true,
4612
4881
  supportsNativeTools: true,
4882
+ defaultToolProtocol: "native",
4613
4883
  inputPrice: 0.6,
4614
4884
  outputPrice: 2.2,
4615
4885
  cacheWritesPrice: 0,
@@ -4622,6 +4892,7 @@ var internationalZAiModels = {
4622
4892
  supportsImages: false,
4623
4893
  supportsPromptCache: false,
4624
4894
  supportsNativeTools: true,
4895
+ defaultToolProtocol: "native",
4625
4896
  inputPrice: 0.1,
4626
4897
  outputPrice: 0.1,
4627
4898
  cacheWritesPrice: 0,
@@ -4637,6 +4908,7 @@ var mainlandZAiModels = {
4637
4908
  supportsImages: false,
4638
4909
  supportsPromptCache: true,
4639
4910
  supportsNativeTools: true,
4911
+ defaultToolProtocol: "native",
4640
4912
  inputPrice: 0.29,
4641
4913
  outputPrice: 1.14,
4642
4914
  cacheWritesPrice: 0,
@@ -4649,6 +4921,7 @@ var mainlandZAiModels = {
4649
4921
  supportsImages: false,
4650
4922
  supportsPromptCache: true,
4651
4923
  supportsNativeTools: true,
4924
+ defaultToolProtocol: "native",
4652
4925
  inputPrice: 0.1,
4653
4926
  outputPrice: 0.6,
4654
4927
  cacheWritesPrice: 0,
@@ -4661,6 +4934,7 @@ var mainlandZAiModels = {
4661
4934
  supportsImages: false,
4662
4935
  supportsPromptCache: true,
4663
4936
  supportsNativeTools: true,
4937
+ defaultToolProtocol: "native",
4664
4938
  inputPrice: 0.29,
4665
4939
  outputPrice: 1.14,
4666
4940
  cacheWritesPrice: 0,
@@ -4673,6 +4947,7 @@ var mainlandZAiModels = {
4673
4947
  supportsImages: false,
4674
4948
  supportsPromptCache: true,
4675
4949
  supportsNativeTools: true,
4950
+ defaultToolProtocol: "native",
4676
4951
  inputPrice: 0.1,
4677
4952
  outputPrice: 0.6,
4678
4953
  cacheWritesPrice: 0,
@@ -4685,6 +4960,7 @@ var mainlandZAiModels = {
4685
4960
  supportsImages: false,
4686
4961
  supportsPromptCache: true,
4687
4962
  supportsNativeTools: true,
4963
+ defaultToolProtocol: "native",
4688
4964
  inputPrice: 0,
4689
4965
  outputPrice: 0,
4690
4966
  cacheWritesPrice: 0,
@@ -4697,6 +4973,7 @@ var mainlandZAiModels = {
4697
4973
  supportsImages: true,
4698
4974
  supportsPromptCache: true,
4699
4975
  supportsNativeTools: true,
4976
+ defaultToolProtocol: "native",
4700
4977
  inputPrice: 0.29,
4701
4978
  outputPrice: 0.93,
4702
4979
  cacheWritesPrice: 0,
@@ -4709,6 +4986,7 @@ var mainlandZAiModels = {
4709
4986
  supportsImages: false,
4710
4987
  supportsPromptCache: true,
4711
4988
  supportsNativeTools: true,
4989
+ defaultToolProtocol: "native",
4712
4990
  inputPrice: 0.29,
4713
4991
  outputPrice: 1.14,
4714
4992
  cacheWritesPrice: 0,
@@ -4762,6 +5040,7 @@ var minimaxModels = {
4762
5040
  supportsImages: false,
4763
5041
  supportsPromptCache: true,
4764
5042
  supportsNativeTools: true,
5043
+ defaultToolProtocol: "native",
4765
5044
  preserveReasoning: true,
4766
5045
  inputPrice: 0.3,
4767
5046
  outputPrice: 1.2,
@@ -4775,6 +5054,7 @@ var minimaxModels = {
4775
5054
  supportsImages: false,
4776
5055
  supportsPromptCache: true,
4777
5056
  supportsNativeTools: true,
5057
+ defaultToolProtocol: "native",
4778
5058
  preserveReasoning: true,
4779
5059
  inputPrice: 0.3,
4780
5060
  outputPrice: 1.2,
@@ -4957,10 +5237,7 @@ var anthropicSchema = apiModelIdProviderModelSchema.extend({
4957
5237
  anthropicBeta1MContext: z8.boolean().optional()
4958
5238
  // Enable 'context-1m-2025-08-07' beta for 1M context window.
4959
5239
  });
4960
- var claudeCodeSchema = apiModelIdProviderModelSchema.extend({
4961
- claudeCodePath: z8.string().optional(),
4962
- claudeCodeMaxOutputTokens: z8.number().int().min(1).max(2e5).optional()
4963
- });
5240
+ var claudeCodeSchema = apiModelIdProviderModelSchema.extend({});
4964
5241
  var openRouterSchema = baseProviderSettingsSchema.extend({
4965
5242
  openRouterApiKey: z8.string().optional(),
4966
5243
  openRouterModelId: z8.string().optional(),
@@ -4985,8 +5262,10 @@ var bedrockSchema = apiModelIdProviderModelSchema.extend({
4985
5262
  awsModelContextWindow: z8.number().optional(),
4986
5263
  awsBedrockEndpointEnabled: z8.boolean().optional(),
4987
5264
  awsBedrockEndpoint: z8.string().optional(),
4988
- awsBedrock1MContext: z8.boolean().optional()
5265
+ awsBedrock1MContext: z8.boolean().optional(),
4989
5266
  // Enable 'context-1m-2025-08-07' beta for 1M context window.
5267
+ awsBedrockServiceTier: z8.enum(["STANDARD", "FLEX", "PRIORITY"]).optional()
5268
+ // AWS Bedrock service tier selection
4990
5269
  });
4991
5270
  var vertexSchema = apiModelIdProviderModelSchema.extend({
4992
5271
  vertexKeyFile: z8.string().optional(),
@@ -5633,11 +5912,44 @@ function getErrorStatusCode(error) {
5633
5912
  }
5634
5913
  return void 0;
5635
5914
  }
5915
+ function extractMessageFromJsonPayload(message) {
5916
+ const jsonStartIndex = message.indexOf("{");
5917
+ if (jsonStartIndex === -1) {
5918
+ return void 0;
5919
+ }
5920
+ const potentialJson = message.slice(jsonStartIndex);
5921
+ try {
5922
+ const parsed = JSON.parse(potentialJson);
5923
+ if (parsed?.error?.message && typeof parsed.error.message === "string") {
5924
+ return parsed.error.message;
5925
+ }
5926
+ if (parsed?.message && typeof parsed.message === "string") {
5927
+ return parsed.message;
5928
+ }
5929
+ } catch {
5930
+ }
5931
+ return void 0;
5932
+ }
5636
5933
  function getErrorMessage(error) {
5934
+ let message;
5637
5935
  if (isOpenAISdkError(error)) {
5638
- return error.error?.metadata?.raw || error.error?.message || error.message;
5936
+ message = error.error?.metadata?.raw || error.error?.message || error.message;
5937
+ } else if (error instanceof Error) {
5938
+ message = error.message;
5939
+ } else if (typeof error === "object" && error !== null && "message" in error) {
5940
+ const msgValue = error.message;
5941
+ if (typeof msgValue === "string") {
5942
+ message = msgValue;
5943
+ }
5639
5944
  }
5640
- return void 0;
5945
+ if (!message) {
5946
+ return void 0;
5947
+ }
5948
+ const extractedMessage = extractMessageFromJsonPayload(message);
5949
+ if (extractedMessage) {
5950
+ return extractedMessage;
5951
+ }
5952
+ return message;
5641
5953
  }
5642
5954
  function shouldReportApiErrorToTelemetry(errorCode, errorMessage) {
5643
5955
  if (errorCode !== void 0 && EXPECTED_API_ERROR_CODES.has(errorCode)) {
@@ -5866,7 +6178,6 @@ var globalSettingsSchema = z14.object({
5866
6178
  alwaysAllowWriteProtected: z14.boolean().optional(),
5867
6179
  writeDelayMs: z14.number().min(0).optional(),
5868
6180
  alwaysAllowBrowser: z14.boolean().optional(),
5869
- alwaysApproveResubmit: z14.boolean().optional(),
5870
6181
  requestDelaySeconds: z14.number().optional(),
5871
6182
  alwaysAllowMcp: z14.boolean().optional(),
5872
6183
  alwaysAllowModeSwitch: z14.boolean().optional(),
@@ -5874,7 +6185,6 @@ var globalSettingsSchema = z14.object({
5874
6185
  alwaysAllowExecute: z14.boolean().optional(),
5875
6186
  alwaysAllowFollowupQuestions: z14.boolean().optional(),
5876
6187
  followupAutoApproveTimeoutMs: z14.number().optional(),
5877
- alwaysAllowUpdateTodoList: z14.boolean().optional(),
5878
6188
  allowedCommands: z14.array(z14.string()).optional(),
5879
6189
  deniedCommands: z14.array(z14.string()).optional(),
5880
6190
  commandExecutionTimeout: z14.number().optional(),
@@ -6036,14 +6346,12 @@ var EVALS_SETTINGS = {
6036
6346
  alwaysAllowWriteProtected: false,
6037
6347
  writeDelayMs: 1e3,
6038
6348
  alwaysAllowBrowser: true,
6039
- alwaysApproveResubmit: true,
6040
6349
  requestDelaySeconds: 10,
6041
6350
  alwaysAllowMcp: true,
6042
6351
  alwaysAllowModeSwitch: true,
6043
6352
  alwaysAllowSubtasks: true,
6044
6353
  alwaysAllowExecute: true,
6045
6354
  alwaysAllowFollowupQuestions: true,
6046
- alwaysAllowUpdateTodoList: true,
6047
6355
  followupAutoApproveTimeoutMs: 0,
6048
6356
  allowedCommands: ["*"],
6049
6357
  commandExecutionTimeout: 20,
@@ -6173,10 +6481,11 @@ var organizationDefaultSettingsSchema = globalSettingsSchema.pick({
6173
6481
  terminalShellIntegrationTimeout: z16.number().int().nonnegative().optional()
6174
6482
  })
6175
6483
  );
6176
- var workspaceTaskVisibilitySchema = z16.enum(["all", "list-only", "full-lockdown"]);
6484
+ var workspaceTaskVisibilitySchema = z16.enum(["all", "list-only", "admins-and-creator", "creator", "full-lockdown"]);
6177
6485
  var organizationCloudSettingsSchema = z16.object({
6178
6486
  recordTaskMessages: z16.boolean().optional(),
6179
6487
  enableTaskSharing: z16.boolean().optional(),
6488
+ allowPublicTaskSharing: z16.boolean().optional(),
6180
6489
  taskShareExpirationDays: z16.number().int().positive().optional(),
6181
6490
  allowMembersViewAllTasks: z16.boolean().optional(),
6182
6491
  workspaceTaskVisibility: workspaceTaskVisibilitySchema.optional()
@@ -6216,6 +6525,7 @@ var ORGANIZATION_DEFAULT = {
6216
6525
  cloudSettings: {
6217
6526
  recordTaskMessages: true,
6218
6527
  enableTaskSharing: true,
6528
+ allowPublicTaskSharing: true,
6219
6529
  taskShareExpirationDays: 30,
6220
6530
  allowMembersViewAllTasks: true
6221
6531
  },
@@ -6751,7 +7061,8 @@ export {
6751
7061
  BEDROCK_GLOBAL_INFERENCE_MODEL_IDS,
6752
7062
  BEDROCK_MAX_TOKENS,
6753
7063
  BEDROCK_REGIONS,
6754
- CLAUDE_CODE_DEFAULT_MAX_OUTPUT_TOKENS,
7064
+ BEDROCK_SERVICE_TIER_MODEL_IDS,
7065
+ BEDROCK_SERVICE_TIER_PRICING,
6755
7066
  CODEBASE_INDEX_DEFAULTS,
6756
7067
  CONSENT_COOKIE_NAME,
6757
7068
  CONTEXT_MANAGEMENT_EVENTS,
@@ -6841,6 +7152,7 @@ export {
6841
7152
  chutesModels,
6842
7153
  claudeCodeDefaultModelId,
6843
7154
  claudeCodeModels,
7155
+ claudeCodeReasoningConfig,
6844
7156
  clineAskSchema,
6845
7157
  clineAsks,
6846
7158
  clineMessageSchema,
@@ -6855,7 +7167,6 @@ export {
6855
7167
  commandIds,
6856
7168
  contextCondenseSchema,
6857
7169
  contextTruncationSchema,
6858
- convertModelNameForVertex,
6859
7170
  customModePromptsSchema,
6860
7171
  customModesSettingsSchema,
6861
7172
  customProviders,
@@ -6877,6 +7188,7 @@ export {
6877
7188
  extensionBridgeEventSchema,
6878
7189
  extensionInstanceSchema,
6879
7190
  extractApiProviderErrorProperties,
7191
+ extractMessageFromJsonPayload,
6880
7192
  fauxProviders,
6881
7193
  featherlessDefaultModelId,
6882
7194
  featherlessModels,
@@ -6886,7 +7198,6 @@ export {
6886
7198
  geminiDefaultModelId,
6887
7199
  geminiModels,
6888
7200
  getApiProtocol,
6889
- getClaudeCodeModelId,
6890
7201
  getEffectiveProtocol,
6891
7202
  getErrorMessage,
6892
7203
  getErrorStatusCode,
@@ -6958,6 +7269,7 @@ export {
6958
7269
  moonshotDefaultModelId,
6959
7270
  moonshotModels,
6960
7271
  nonBlockingAsks,
7272
+ normalizeClaudeCodeModelId,
6961
7273
  ollamaDefaultModelId,
6962
7274
  ollamaDefaultModelInfo,
6963
7275
  openAiModelInfoSaneDefaults,