@roo-code/types 1.92.0 → 1.93.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.js CHANGED
@@ -571,6 +571,7 @@ var anthropicModels = {
571
571
  supportsImages: true,
572
572
  supportsPromptCache: true,
573
573
  supportsNativeTools: true,
574
+ defaultToolProtocol: "native",
574
575
  inputPrice: 3,
575
576
  // $3 per million input tokens (≤200K context)
576
577
  outputPrice: 15,
@@ -604,6 +605,7 @@ var anthropicModels = {
604
605
  supportsImages: true,
605
606
  supportsPromptCache: true,
606
607
  supportsNativeTools: true,
608
+ defaultToolProtocol: "native",
607
609
  inputPrice: 3,
608
610
  // $3 per million input tokens (≤200K context)
609
611
  outputPrice: 15,
@@ -636,6 +638,7 @@ var anthropicModels = {
636
638
  supportsImages: true,
637
639
  supportsPromptCache: true,
638
640
  supportsNativeTools: true,
641
+ defaultToolProtocol: "native",
639
642
  inputPrice: 5,
640
643
  // $5 per million input tokens
641
644
  outputPrice: 25,
@@ -653,6 +656,7 @@ var anthropicModels = {
653
656
  supportsImages: true,
654
657
  supportsPromptCache: true,
655
658
  supportsNativeTools: true,
659
+ defaultToolProtocol: "native",
656
660
  inputPrice: 15,
657
661
  // $15 per million input tokens
658
662
  outputPrice: 75,
@@ -670,6 +674,7 @@ var anthropicModels = {
670
674
  supportsImages: true,
671
675
  supportsPromptCache: true,
672
676
  supportsNativeTools: true,
677
+ defaultToolProtocol: "native",
673
678
  inputPrice: 15,
674
679
  // $15 per million input tokens
675
680
  outputPrice: 75,
@@ -687,6 +692,7 @@ var anthropicModels = {
687
692
  supportsImages: true,
688
693
  supportsPromptCache: true,
689
694
  supportsNativeTools: true,
695
+ defaultToolProtocol: "native",
690
696
  inputPrice: 3,
691
697
  // $3 per million input tokens
692
698
  outputPrice: 15,
@@ -705,6 +711,7 @@ var anthropicModels = {
705
711
  supportsImages: true,
706
712
  supportsPromptCache: true,
707
713
  supportsNativeTools: true,
714
+ defaultToolProtocol: "native",
708
715
  inputPrice: 3,
709
716
  // $3 per million input tokens
710
717
  outputPrice: 15,
@@ -720,6 +727,7 @@ var anthropicModels = {
720
727
  supportsImages: true,
721
728
  supportsPromptCache: true,
722
729
  supportsNativeTools: true,
730
+ defaultToolProtocol: "native",
723
731
  inputPrice: 3,
724
732
  // $3 per million input tokens
725
733
  outputPrice: 15,
@@ -735,6 +743,7 @@ var anthropicModels = {
735
743
  supportsImages: false,
736
744
  supportsPromptCache: true,
737
745
  supportsNativeTools: true,
746
+ defaultToolProtocol: "native",
738
747
  inputPrice: 1,
739
748
  outputPrice: 5,
740
749
  cacheWritesPrice: 1.25,
@@ -746,6 +755,7 @@ var anthropicModels = {
746
755
  supportsImages: true,
747
756
  supportsPromptCache: true,
748
757
  supportsNativeTools: true,
758
+ defaultToolProtocol: "native",
749
759
  inputPrice: 15,
750
760
  outputPrice: 75,
751
761
  cacheWritesPrice: 18.75,
@@ -757,6 +767,7 @@ var anthropicModels = {
757
767
  supportsImages: true,
758
768
  supportsPromptCache: true,
759
769
  supportsNativeTools: true,
770
+ defaultToolProtocol: "native",
760
771
  inputPrice: 0.25,
761
772
  outputPrice: 1.25,
762
773
  cacheWritesPrice: 0.3,
@@ -768,6 +779,7 @@ var anthropicModels = {
768
779
  supportsImages: true,
769
780
  supportsPromptCache: true,
770
781
  supportsNativeTools: true,
782
+ defaultToolProtocol: "native",
771
783
  inputPrice: 1,
772
784
  outputPrice: 5,
773
785
  cacheWritesPrice: 1.25,
@@ -920,6 +932,7 @@ var bedrockModels = {
920
932
  supportsPromptCache: true,
921
933
  supportsReasoningBudget: true,
922
934
  supportsNativeTools: true,
935
+ defaultToolProtocol: "native",
923
936
  inputPrice: 3,
924
937
  outputPrice: 15,
925
938
  cacheWritesPrice: 3.75,
@@ -1013,6 +1026,7 @@ var bedrockModels = {
1013
1026
  supportsPromptCache: true,
1014
1027
  supportsReasoningBudget: true,
1015
1028
  supportsNativeTools: true,
1029
+ defaultToolProtocol: "native",
1016
1030
  inputPrice: 3,
1017
1031
  outputPrice: 15,
1018
1032
  cacheWritesPrice: 3.75,
@@ -1028,6 +1042,7 @@ var bedrockModels = {
1028
1042
  supportsPromptCache: true,
1029
1043
  supportsReasoningBudget: true,
1030
1044
  supportsNativeTools: true,
1045
+ defaultToolProtocol: "native",
1031
1046
  inputPrice: 15,
1032
1047
  outputPrice: 75,
1033
1048
  cacheWritesPrice: 18.75,
@@ -1043,6 +1058,7 @@ var bedrockModels = {
1043
1058
  supportsPromptCache: true,
1044
1059
  supportsReasoningBudget: true,
1045
1060
  supportsNativeTools: true,
1061
+ defaultToolProtocol: "native",
1046
1062
  inputPrice: 5,
1047
1063
  outputPrice: 25,
1048
1064
  cacheWritesPrice: 6.25,
@@ -1058,6 +1074,7 @@ var bedrockModels = {
1058
1074
  supportsPromptCache: true,
1059
1075
  supportsReasoningBudget: true,
1060
1076
  supportsNativeTools: true,
1077
+ defaultToolProtocol: "native",
1061
1078
  inputPrice: 15,
1062
1079
  outputPrice: 75,
1063
1080
  cacheWritesPrice: 18.75,
@@ -1073,6 +1090,7 @@ var bedrockModels = {
1073
1090
  supportsPromptCache: true,
1074
1091
  supportsReasoningBudget: true,
1075
1092
  supportsNativeTools: true,
1093
+ defaultToolProtocol: "native",
1076
1094
  inputPrice: 3,
1077
1095
  outputPrice: 15,
1078
1096
  cacheWritesPrice: 3.75,
@@ -1087,6 +1105,7 @@ var bedrockModels = {
1087
1105
  supportsImages: true,
1088
1106
  supportsPromptCache: true,
1089
1107
  supportsNativeTools: true,
1108
+ defaultToolProtocol: "native",
1090
1109
  inputPrice: 3,
1091
1110
  outputPrice: 15,
1092
1111
  cacheWritesPrice: 3.75,
@@ -1101,6 +1120,7 @@ var bedrockModels = {
1101
1120
  supportsImages: false,
1102
1121
  supportsPromptCache: true,
1103
1122
  supportsNativeTools: true,
1123
+ defaultToolProtocol: "native",
1104
1124
  inputPrice: 0.8,
1105
1125
  outputPrice: 4,
1106
1126
  cacheWritesPrice: 1,
@@ -1116,6 +1136,7 @@ var bedrockModels = {
1116
1136
  supportsPromptCache: true,
1117
1137
  supportsReasoningBudget: true,
1118
1138
  supportsNativeTools: true,
1139
+ defaultToolProtocol: "native",
1119
1140
  inputPrice: 1,
1120
1141
  outputPrice: 5,
1121
1142
  cacheWritesPrice: 1.25,
@@ -1132,6 +1153,7 @@ var bedrockModels = {
1132
1153
  supportsImages: true,
1133
1154
  supportsPromptCache: false,
1134
1155
  supportsNativeTools: true,
1156
+ defaultToolProtocol: "native",
1135
1157
  inputPrice: 3,
1136
1158
  outputPrice: 15
1137
1159
  },
@@ -1141,6 +1163,7 @@ var bedrockModels = {
1141
1163
  supportsImages: true,
1142
1164
  supportsPromptCache: false,
1143
1165
  supportsNativeTools: true,
1166
+ defaultToolProtocol: "native",
1144
1167
  inputPrice: 15,
1145
1168
  outputPrice: 75
1146
1169
  },
@@ -1150,6 +1173,7 @@ var bedrockModels = {
1150
1173
  supportsImages: true,
1151
1174
  supportsPromptCache: false,
1152
1175
  supportsNativeTools: true,
1176
+ defaultToolProtocol: "native",
1153
1177
  inputPrice: 3,
1154
1178
  outputPrice: 15
1155
1179
  },
@@ -1159,6 +1183,7 @@ var bedrockModels = {
1159
1183
  supportsImages: true,
1160
1184
  supportsPromptCache: false,
1161
1185
  supportsNativeTools: true,
1186
+ defaultToolProtocol: "native",
1162
1187
  inputPrice: 0.25,
1163
1188
  outputPrice: 1.25
1164
1189
  },
@@ -1168,6 +1193,7 @@ var bedrockModels = {
1168
1193
  supportsImages: false,
1169
1194
  supportsPromptCache: false,
1170
1195
  supportsNativeTools: true,
1196
+ defaultToolProtocol: "native",
1171
1197
  inputPrice: 8,
1172
1198
  outputPrice: 24,
1173
1199
  description: "Claude 2.1"
@@ -1178,6 +1204,7 @@ var bedrockModels = {
1178
1204
  supportsImages: false,
1179
1205
  supportsPromptCache: false,
1180
1206
  supportsNativeTools: true,
1207
+ defaultToolProtocol: "native",
1181
1208
  inputPrice: 8,
1182
1209
  outputPrice: 24,
1183
1210
  description: "Claude 2.0"
@@ -1188,6 +1215,7 @@ var bedrockModels = {
1188
1215
  supportsImages: false,
1189
1216
  supportsPromptCache: false,
1190
1217
  supportsNativeTools: true,
1218
+ defaultToolProtocol: "native",
1191
1219
  inputPrice: 0.8,
1192
1220
  outputPrice: 2.4,
1193
1221
  description: "Claude Instant"
@@ -1504,6 +1532,7 @@ var cerebrasModels = {
1504
1532
  supportsImages: false,
1505
1533
  supportsPromptCache: false,
1506
1534
  supportsNativeTools: true,
1535
+ defaultToolProtocol: "native",
1507
1536
  inputPrice: 0,
1508
1537
  outputPrice: 0,
1509
1538
  description: "Highly intelligent general purpose model with up to 1,000 tokens/s"
@@ -1515,6 +1544,7 @@ var cerebrasModels = {
1515
1544
  supportsImages: false,
1516
1545
  supportsPromptCache: false,
1517
1546
  supportsNativeTools: true,
1547
+ defaultToolProtocol: "native",
1518
1548
  inputPrice: 0,
1519
1549
  outputPrice: 0,
1520
1550
  description: "Intelligent model with ~1400 tokens/s"
@@ -1526,6 +1556,7 @@ var cerebrasModels = {
1526
1556
  supportsImages: false,
1527
1557
  supportsPromptCache: false,
1528
1558
  supportsNativeTools: true,
1559
+ defaultToolProtocol: "native",
1529
1560
  inputPrice: 0,
1530
1561
  outputPrice: 0,
1531
1562
  description: "Powerful model with ~2600 tokens/s"
@@ -1537,6 +1568,7 @@ var cerebrasModels = {
1537
1568
  supportsImages: false,
1538
1569
  supportsPromptCache: false,
1539
1570
  supportsNativeTools: true,
1571
+ defaultToolProtocol: "native",
1540
1572
  inputPrice: 0,
1541
1573
  outputPrice: 0,
1542
1574
  description: "SOTA coding performance with ~2500 tokens/s"
@@ -1548,6 +1580,7 @@ var cerebrasModels = {
1548
1580
  supportsImages: false,
1549
1581
  supportsPromptCache: false,
1550
1582
  supportsNativeTools: true,
1583
+ defaultToolProtocol: "native",
1551
1584
  inputPrice: 0,
1552
1585
  outputPrice: 0,
1553
1586
  description: "OpenAI GPT OSS model with ~2800 tokens/s\n\n\u2022 64K context window\n\u2022 Excels at efficient reasoning across science, math, and coding"
@@ -1562,6 +1595,8 @@ var chutesModels = {
1562
1595
  contextWindow: 163840,
1563
1596
  supportsImages: false,
1564
1597
  supportsPromptCache: false,
1598
+ supportsNativeTools: true,
1599
+ defaultToolProtocol: "native",
1565
1600
  inputPrice: 0,
1566
1601
  outputPrice: 0,
1567
1602
  description: "DeepSeek R1 0528 model."
@@ -1571,6 +1606,8 @@ var chutesModels = {
1571
1606
  contextWindow: 163840,
1572
1607
  supportsImages: false,
1573
1608
  supportsPromptCache: false,
1609
+ supportsNativeTools: true,
1610
+ defaultToolProtocol: "native",
1574
1611
  inputPrice: 0,
1575
1612
  outputPrice: 0,
1576
1613
  description: "DeepSeek R1 model."
@@ -1580,6 +1617,8 @@ var chutesModels = {
1580
1617
  contextWindow: 163840,
1581
1618
  supportsImages: false,
1582
1619
  supportsPromptCache: false,
1620
+ supportsNativeTools: true,
1621
+ defaultToolProtocol: "native",
1583
1622
  inputPrice: 0,
1584
1623
  outputPrice: 0,
1585
1624
  description: "DeepSeek V3 model."
@@ -1589,6 +1628,8 @@ var chutesModels = {
1589
1628
  contextWindow: 163840,
1590
1629
  supportsImages: false,
1591
1630
  supportsPromptCache: false,
1631
+ supportsNativeTools: true,
1632
+ defaultToolProtocol: "native",
1592
1633
  inputPrice: 0,
1593
1634
  outputPrice: 0,
1594
1635
  description: "DeepSeek V3.1 model."
@@ -1598,6 +1639,8 @@ var chutesModels = {
1598
1639
  contextWindow: 163840,
1599
1640
  supportsImages: false,
1600
1641
  supportsPromptCache: false,
1642
+ supportsNativeTools: true,
1643
+ defaultToolProtocol: "native",
1601
1644
  inputPrice: 0.23,
1602
1645
  outputPrice: 0.9,
1603
1646
  description: "DeepSeek\u2011V3.1\u2011Terminus is an update to V3.1 that improves language consistency by reducing CN/EN mix\u2011ups and eliminating random characters, while strengthening agent capabilities with notably better Code Agent and Search Agent performance."
@@ -1607,6 +1650,8 @@ var chutesModels = {
1607
1650
  contextWindow: 163840,
1608
1651
  supportsImages: false,
1609
1652
  supportsPromptCache: false,
1653
+ supportsNativeTools: true,
1654
+ defaultToolProtocol: "native",
1610
1655
  inputPrice: 1,
1611
1656
  outputPrice: 3,
1612
1657
  description: "DeepSeek-V3.1-turbo is an FP8, speculative-decoding turbo variant optimized for ultra-fast single-shot queries (~200 TPS), with outputs close to the originals and solid function calling/reasoning/structured output, priced at $1/M input and $3/M output tokens, using 2\xD7 quota per request and not intended for bulk workloads."
@@ -1616,6 +1661,8 @@ var chutesModels = {
1616
1661
  contextWindow: 163840,
1617
1662
  supportsImages: false,
1618
1663
  supportsPromptCache: false,
1664
+ supportsNativeTools: true,
1665
+ defaultToolProtocol: "native",
1619
1666
  inputPrice: 0.25,
1620
1667
  outputPrice: 0.35,
1621
1668
  description: "DeepSeek-V3.2-Exp is an experimental LLM that introduces DeepSeek Sparse Attention to improve long\u2011context training and inference efficiency while maintaining performance comparable to V3.1\u2011Terminus."
@@ -1627,6 +1674,8 @@ var chutesModels = {
1627
1674
  // From Groq
1628
1675
  supportsImages: false,
1629
1676
  supportsPromptCache: false,
1677
+ supportsNativeTools: true,
1678
+ defaultToolProtocol: "native",
1630
1679
  inputPrice: 0,
1631
1680
  outputPrice: 0,
1632
1681
  description: "Unsloth Llama 3.3 70B Instruct model."
@@ -1636,6 +1685,8 @@ var chutesModels = {
1636
1685
  contextWindow: 512e3,
1637
1686
  supportsImages: false,
1638
1687
  supportsPromptCache: false,
1688
+ supportsNativeTools: true,
1689
+ defaultToolProtocol: "native",
1639
1690
  inputPrice: 0,
1640
1691
  outputPrice: 0,
1641
1692
  description: "ChutesAI Llama 4 Scout 17B Instruct model, 512K context."
@@ -1645,6 +1696,8 @@ var chutesModels = {
1645
1696
  contextWindow: 128e3,
1646
1697
  supportsImages: false,
1647
1698
  supportsPromptCache: false,
1699
+ supportsNativeTools: true,
1700
+ defaultToolProtocol: "native",
1648
1701
  inputPrice: 0,
1649
1702
  outputPrice: 0,
1650
1703
  description: "Unsloth Mistral Nemo Instruct model."
@@ -1654,6 +1707,8 @@ var chutesModels = {
1654
1707
  contextWindow: 131072,
1655
1708
  supportsImages: false,
1656
1709
  supportsPromptCache: false,
1710
+ supportsNativeTools: true,
1711
+ defaultToolProtocol: "native",
1657
1712
  inputPrice: 0,
1658
1713
  outputPrice: 0,
1659
1714
  description: "Unsloth Gemma 3 12B IT model."
@@ -1663,6 +1718,8 @@ var chutesModels = {
1663
1718
  contextWindow: 131072,
1664
1719
  supportsImages: false,
1665
1720
  supportsPromptCache: false,
1721
+ supportsNativeTools: true,
1722
+ defaultToolProtocol: "native",
1666
1723
  inputPrice: 0,
1667
1724
  outputPrice: 0,
1668
1725
  description: "Nous DeepHermes 3 Llama 3 8B Preview model."
@@ -1672,6 +1729,8 @@ var chutesModels = {
1672
1729
  contextWindow: 131072,
1673
1730
  supportsImages: false,
1674
1731
  supportsPromptCache: false,
1732
+ supportsNativeTools: true,
1733
+ defaultToolProtocol: "native",
1675
1734
  inputPrice: 0,
1676
1735
  outputPrice: 0,
1677
1736
  description: "Unsloth Gemma 3 4B IT model."
@@ -1681,6 +1740,8 @@ var chutesModels = {
1681
1740
  contextWindow: 131072,
1682
1741
  supportsImages: false,
1683
1742
  supportsPromptCache: false,
1743
+ supportsNativeTools: true,
1744
+ defaultToolProtocol: "native",
1684
1745
  inputPrice: 0,
1685
1746
  outputPrice: 0,
1686
1747
  description: "Nvidia Llama 3.3 Nemotron Super 49B model."
@@ -1690,6 +1751,8 @@ var chutesModels = {
1690
1751
  contextWindow: 131072,
1691
1752
  supportsImages: false,
1692
1753
  supportsPromptCache: false,
1754
+ supportsNativeTools: true,
1755
+ defaultToolProtocol: "native",
1693
1756
  inputPrice: 0,
1694
1757
  outputPrice: 0,
1695
1758
  description: "Nvidia Llama 3.1 Nemotron Ultra 253B model."
@@ -1699,6 +1762,8 @@ var chutesModels = {
1699
1762
  contextWindow: 256e3,
1700
1763
  supportsImages: false,
1701
1764
  supportsPromptCache: false,
1765
+ supportsNativeTools: true,
1766
+ defaultToolProtocol: "native",
1702
1767
  inputPrice: 0,
1703
1768
  outputPrice: 0,
1704
1769
  description: "ChutesAI Llama 4 Maverick 17B Instruct FP8 model."
@@ -1708,6 +1773,8 @@ var chutesModels = {
1708
1773
  contextWindow: 163840,
1709
1774
  supportsImages: false,
1710
1775
  supportsPromptCache: false,
1776
+ supportsNativeTools: true,
1777
+ defaultToolProtocol: "native",
1711
1778
  inputPrice: 0,
1712
1779
  outputPrice: 0,
1713
1780
  description: "DeepSeek V3 Base model."
@@ -1717,6 +1784,8 @@ var chutesModels = {
1717
1784
  contextWindow: 163840,
1718
1785
  supportsImages: false,
1719
1786
  supportsPromptCache: false,
1787
+ supportsNativeTools: true,
1788
+ defaultToolProtocol: "native",
1720
1789
  inputPrice: 0,
1721
1790
  outputPrice: 0,
1722
1791
  description: "DeepSeek R1 Zero model."
@@ -1726,6 +1795,8 @@ var chutesModels = {
1726
1795
  contextWindow: 163840,
1727
1796
  supportsImages: false,
1728
1797
  supportsPromptCache: false,
1798
+ supportsNativeTools: true,
1799
+ defaultToolProtocol: "native",
1729
1800
  inputPrice: 0,
1730
1801
  outputPrice: 0,
1731
1802
  description: "DeepSeek V3 (0324) model."
@@ -1735,6 +1806,8 @@ var chutesModels = {
1735
1806
  contextWindow: 262144,
1736
1807
  supportsImages: false,
1737
1808
  supportsPromptCache: false,
1809
+ supportsNativeTools: true,
1810
+ defaultToolProtocol: "native",
1738
1811
  inputPrice: 0,
1739
1812
  outputPrice: 0,
1740
1813
  description: "Qwen3 235B A22B Instruct 2507 model with 262K context window."
@@ -1744,6 +1817,8 @@ var chutesModels = {
1744
1817
  contextWindow: 40960,
1745
1818
  supportsImages: false,
1746
1819
  supportsPromptCache: false,
1820
+ supportsNativeTools: true,
1821
+ defaultToolProtocol: "native",
1747
1822
  inputPrice: 0,
1748
1823
  outputPrice: 0,
1749
1824
  description: "Qwen3 235B A22B model."
@@ -1753,6 +1828,8 @@ var chutesModels = {
1753
1828
  contextWindow: 40960,
1754
1829
  supportsImages: false,
1755
1830
  supportsPromptCache: false,
1831
+ supportsNativeTools: true,
1832
+ defaultToolProtocol: "native",
1756
1833
  inputPrice: 0,
1757
1834
  outputPrice: 0,
1758
1835
  description: "Qwen3 32B model."
@@ -1762,6 +1839,8 @@ var chutesModels = {
1762
1839
  contextWindow: 40960,
1763
1840
  supportsImages: false,
1764
1841
  supportsPromptCache: false,
1842
+ supportsNativeTools: true,
1843
+ defaultToolProtocol: "native",
1765
1844
  inputPrice: 0,
1766
1845
  outputPrice: 0,
1767
1846
  description: "Qwen3 30B A3B model."
@@ -1771,6 +1850,8 @@ var chutesModels = {
1771
1850
  contextWindow: 40960,
1772
1851
  supportsImages: false,
1773
1852
  supportsPromptCache: false,
1853
+ supportsNativeTools: true,
1854
+ defaultToolProtocol: "native",
1774
1855
  inputPrice: 0,
1775
1856
  outputPrice: 0,
1776
1857
  description: "Qwen3 14B model."
@@ -1780,6 +1861,8 @@ var chutesModels = {
1780
1861
  contextWindow: 40960,
1781
1862
  supportsImages: false,
1782
1863
  supportsPromptCache: false,
1864
+ supportsNativeTools: true,
1865
+ defaultToolProtocol: "native",
1783
1866
  inputPrice: 0,
1784
1867
  outputPrice: 0,
1785
1868
  description: "Qwen3 8B model."
@@ -1789,6 +1872,8 @@ var chutesModels = {
1789
1872
  contextWindow: 163840,
1790
1873
  supportsImages: false,
1791
1874
  supportsPromptCache: false,
1875
+ supportsNativeTools: true,
1876
+ defaultToolProtocol: "native",
1792
1877
  inputPrice: 0,
1793
1878
  outputPrice: 0,
1794
1879
  description: "Microsoft MAI-DS-R1 FP8 model."
@@ -1798,6 +1883,8 @@ var chutesModels = {
1798
1883
  contextWindow: 163840,
1799
1884
  supportsImages: false,
1800
1885
  supportsPromptCache: false,
1886
+ supportsNativeTools: true,
1887
+ defaultToolProtocol: "native",
1801
1888
  inputPrice: 0,
1802
1889
  outputPrice: 0,
1803
1890
  description: "TNGTech DeepSeek R1T Chimera model."
@@ -1807,6 +1894,8 @@ var chutesModels = {
1807
1894
  contextWindow: 151329,
1808
1895
  supportsImages: false,
1809
1896
  supportsPromptCache: false,
1897
+ supportsNativeTools: true,
1898
+ defaultToolProtocol: "native",
1810
1899
  inputPrice: 0,
1811
1900
  outputPrice: 0,
1812
1901
  description: "GLM-4.5-Air model with 151,329 token context window and 106B total parameters with 12B activated."
@@ -1816,6 +1905,8 @@ var chutesModels = {
1816
1905
  contextWindow: 131072,
1817
1906
  supportsImages: false,
1818
1907
  supportsPromptCache: false,
1908
+ supportsNativeTools: true,
1909
+ defaultToolProtocol: "native",
1819
1910
  inputPrice: 0,
1820
1911
  outputPrice: 0,
1821
1912
  description: "GLM-4.5-FP8 model with 128k token context window, optimized for agent-based applications with MoE architecture."
@@ -1825,6 +1916,8 @@ var chutesModels = {
1825
1916
  contextWindow: 131072,
1826
1917
  supportsImages: false,
1827
1918
  supportsPromptCache: false,
1919
+ supportsNativeTools: true,
1920
+ defaultToolProtocol: "native",
1828
1921
  inputPrice: 1,
1829
1922
  outputPrice: 3,
1830
1923
  description: "GLM-4.5-turbo model with 128K token context window, optimized for fast inference."
@@ -1834,6 +1927,8 @@ var chutesModels = {
1834
1927
  contextWindow: 202752,
1835
1928
  supportsImages: false,
1836
1929
  supportsPromptCache: false,
1930
+ supportsNativeTools: true,
1931
+ defaultToolProtocol: "native",
1837
1932
  inputPrice: 0,
1838
1933
  outputPrice: 0,
1839
1934
  description: "GLM-4.6 introduces major upgrades over GLM-4.5, including a longer 200K-token context window for complex tasks, stronger coding performance in benchmarks and real-world tools (such as Claude Code, Cline, Roo Code, and Kilo Code), improved reasoning with tool use during inference, more capable and efficient agent integration, and refined writing that better matches human style, readability, and natural role-play scenarios."
@@ -1844,6 +1939,8 @@ var chutesModels = {
1844
1939
  contextWindow: 202752,
1845
1940
  supportsImages: false,
1846
1941
  supportsPromptCache: false,
1942
+ supportsNativeTools: true,
1943
+ defaultToolProtocol: "native",
1847
1944
  inputPrice: 1.15,
1848
1945
  outputPrice: 3.25,
1849
1946
  description: "GLM-4.6-turbo model with 200K-token context window, optimized for fast inference."
@@ -1853,6 +1950,8 @@ var chutesModels = {
1853
1950
  contextWindow: 128e3,
1854
1951
  supportsImages: false,
1855
1952
  supportsPromptCache: false,
1953
+ supportsNativeTools: true,
1954
+ defaultToolProtocol: "native",
1856
1955
  inputPrice: 0,
1857
1956
  outputPrice: 0,
1858
1957
  description: "LongCat Flash Thinking FP8 model with 128K context window, optimized for complex reasoning and coding tasks."
@@ -1862,6 +1961,8 @@ var chutesModels = {
1862
1961
  contextWindow: 262144,
1863
1962
  supportsImages: false,
1864
1963
  supportsPromptCache: false,
1964
+ supportsNativeTools: true,
1965
+ defaultToolProtocol: "native",
1865
1966
  inputPrice: 0,
1866
1967
  outputPrice: 0,
1867
1968
  description: "Qwen3 Coder 480B A35B Instruct FP8 model, optimized for coding tasks."
@@ -1871,6 +1972,8 @@ var chutesModels = {
1871
1972
  contextWindow: 75e3,
1872
1973
  supportsImages: false,
1873
1974
  supportsPromptCache: false,
1975
+ supportsNativeTools: true,
1976
+ defaultToolProtocol: "native",
1874
1977
  inputPrice: 0.1481,
1875
1978
  outputPrice: 0.5926,
1876
1979
  description: "Moonshot AI Kimi K2 Instruct model with 75k context window."
@@ -1880,6 +1983,8 @@ var chutesModels = {
1880
1983
  contextWindow: 262144,
1881
1984
  supportsImages: false,
1882
1985
  supportsPromptCache: false,
1986
+ supportsNativeTools: true,
1987
+ defaultToolProtocol: "native",
1883
1988
  inputPrice: 0.1999,
1884
1989
  outputPrice: 0.8001,
1885
1990
  description: "Moonshot AI Kimi K2 Instruct 0905 model with 256k context window."
@@ -1889,6 +1994,8 @@ var chutesModels = {
1889
1994
  contextWindow: 262144,
1890
1995
  supportsImages: false,
1891
1996
  supportsPromptCache: false,
1997
+ supportsNativeTools: true,
1998
+ defaultToolProtocol: "native",
1892
1999
  inputPrice: 0.077968332,
1893
2000
  outputPrice: 0.31202496,
1894
2001
  description: "Qwen3 235B A22B Thinking 2507 model with 262K context window."
@@ -1898,6 +2005,8 @@ var chutesModels = {
1898
2005
  contextWindow: 131072,
1899
2006
  supportsImages: false,
1900
2007
  supportsPromptCache: false,
2008
+ supportsNativeTools: true,
2009
+ defaultToolProtocol: "native",
1901
2010
  inputPrice: 0,
1902
2011
  outputPrice: 0,
1903
2012
  description: "Fast, stable instruction-tuned model optimized for complex tasks, RAG, and tool use without thinking traces."
@@ -1907,6 +2016,8 @@ var chutesModels = {
1907
2016
  contextWindow: 131072,
1908
2017
  supportsImages: false,
1909
2018
  supportsPromptCache: false,
2019
+ supportsNativeTools: true,
2020
+ defaultToolProtocol: "native",
1910
2021
  inputPrice: 0,
1911
2022
  outputPrice: 0,
1912
2023
  description: "Reasoning-first model with structured thinking traces for multi-step problems, math proofs, and code synthesis."
@@ -1916,6 +2027,8 @@ var chutesModels = {
1916
2027
  contextWindow: 262144,
1917
2028
  supportsImages: true,
1918
2029
  supportsPromptCache: false,
2030
+ supportsNativeTools: true,
2031
+ defaultToolProtocol: "native",
1919
2032
  inputPrice: 0.16,
1920
2033
  outputPrice: 0.65,
1921
2034
  description: "Qwen3\u2011VL\u2011235B\u2011A22B\u2011Thinking is an open\u2011weight MoE vision\u2011language model (235B total, ~22B activated) optimized for deliberate multi\u2011step reasoning with strong text\u2011image\u2011video understanding and long\u2011context capabilities."
@@ -1924,139 +2037,71 @@ var chutesModels = {
1924
2037
  var chutesDefaultModelInfo = chutesModels[chutesDefaultModelId];
1925
2038
 
1926
2039
  // src/providers/claude-code.ts
1927
- var VERTEX_DATE_PATTERN = /-(\d{8})$/;
1928
- function convertModelNameForVertex(modelName) {
1929
- return modelName.replace(VERTEX_DATE_PATTERN, "@$1");
1930
- }
1931
- var claudeCodeDefaultModelId = "claude-sonnet-4-5";
1932
- var CLAUDE_CODE_DEFAULT_MAX_OUTPUT_TOKENS = 16e3;
1933
- function getClaudeCodeModelId(baseModelId, useVertex = false) {
1934
- return useVertex ? convertModelNameForVertex(baseModelId) : baseModelId;
1935
- }
2040
+ var DATE_SUFFIX_PATTERN = /-\d{8}$/;
1936
2041
  var claudeCodeModels = {
1937
- "claude-sonnet-4-5": {
1938
- ...anthropicModels["claude-sonnet-4-5"],
1939
- supportsImages: false,
1940
- supportsPromptCache: true,
1941
- // Claude Code does report cache tokens
1942
- supportsReasoningEffort: false,
1943
- supportsReasoningBudget: false,
1944
- requiredReasoningBudget: false,
1945
- // Claude Code manages its own tools and temperature via the CLI
1946
- supportsNativeTools: false,
1947
- supportsTemperature: false
1948
- },
1949
- "claude-sonnet-4-5-20250929[1m]": {
1950
- ...anthropicModels["claude-sonnet-4-5"],
1951
- contextWindow: 1e6,
1952
- // 1M token context window (requires [1m] suffix)
1953
- supportsImages: false,
1954
- supportsPromptCache: true,
1955
- // Claude Code does report cache tokens
1956
- supportsReasoningEffort: false,
1957
- supportsReasoningBudget: false,
1958
- requiredReasoningBudget: false,
1959
- // Claude Code manages its own tools and temperature via the CLI
1960
- supportsNativeTools: false,
1961
- supportsTemperature: false
1962
- },
1963
- "claude-sonnet-4-20250514": {
1964
- ...anthropicModels["claude-sonnet-4-20250514"],
1965
- supportsImages: false,
1966
- supportsPromptCache: true,
1967
- // Claude Code does report cache tokens
1968
- supportsReasoningEffort: false,
1969
- supportsReasoningBudget: false,
1970
- requiredReasoningBudget: false,
1971
- // Claude Code manages its own tools and temperature via the CLI
1972
- supportsNativeTools: false,
1973
- supportsTemperature: false
1974
- },
1975
- "claude-opus-4-5-20251101": {
1976
- ...anthropicModels["claude-opus-4-5-20251101"],
1977
- supportsImages: false,
1978
- supportsPromptCache: true,
1979
- // Claude Code does report cache tokens
1980
- supportsReasoningEffort: false,
1981
- supportsReasoningBudget: false,
1982
- requiredReasoningBudget: false,
1983
- // Claude Code manages its own tools and temperature via the CLI
1984
- supportsNativeTools: false,
1985
- supportsTemperature: false
1986
- },
1987
- "claude-opus-4-1-20250805": {
1988
- ...anthropicModels["claude-opus-4-1-20250805"],
1989
- supportsImages: false,
1990
- supportsPromptCache: true,
1991
- // Claude Code does report cache tokens
1992
- supportsReasoningEffort: false,
1993
- supportsReasoningBudget: false,
1994
- requiredReasoningBudget: false,
1995
- // Claude Code manages its own tools and temperature via the CLI
1996
- supportsNativeTools: false,
1997
- supportsTemperature: false
1998
- },
1999
- "claude-opus-4-20250514": {
2000
- ...anthropicModels["claude-opus-4-20250514"],
2001
- supportsImages: false,
2002
- supportsPromptCache: true,
2003
- // Claude Code does report cache tokens
2004
- supportsReasoningEffort: false,
2005
- supportsReasoningBudget: false,
2006
- requiredReasoningBudget: false,
2007
- // Claude Code manages its own tools and temperature via the CLI
2008
- supportsNativeTools: false,
2009
- supportsTemperature: false
2010
- },
2011
- "claude-3-7-sonnet-20250219": {
2012
- ...anthropicModels["claude-3-7-sonnet-20250219"],
2013
- supportsImages: false,
2014
- supportsPromptCache: true,
2015
- // Claude Code does report cache tokens
2016
- supportsReasoningEffort: false,
2017
- supportsReasoningBudget: false,
2018
- requiredReasoningBudget: false,
2019
- // Claude Code manages its own tools and temperature via the CLI
2020
- supportsNativeTools: false,
2021
- supportsTemperature: false
2022
- },
2023
- "claude-3-5-sonnet-20241022": {
2024
- ...anthropicModels["claude-3-5-sonnet-20241022"],
2025
- supportsImages: false,
2042
+ "claude-haiku-4-5": {
2043
+ maxTokens: 32768,
2044
+ contextWindow: 2e5,
2045
+ supportsImages: true,
2026
2046
  supportsPromptCache: true,
2027
- // Claude Code does report cache tokens
2028
- supportsReasoningEffort: false,
2029
- supportsReasoningBudget: false,
2030
- requiredReasoningBudget: false,
2031
- // Claude Code manages its own tools and temperature via the CLI
2032
- supportsNativeTools: false,
2033
- supportsTemperature: false
2047
+ supportsNativeTools: true,
2048
+ defaultToolProtocol: "native",
2049
+ supportsReasoningEffort: ["disable", "low", "medium", "high"],
2050
+ reasoningEffort: "medium",
2051
+ description: "Claude Haiku 4.5 - Fast and efficient with thinking"
2034
2052
  },
2035
- "claude-3-5-haiku-20241022": {
2036
- ...anthropicModels["claude-3-5-haiku-20241022"],
2037
- supportsImages: false,
2053
+ "claude-sonnet-4-5": {
2054
+ maxTokens: 32768,
2055
+ contextWindow: 2e5,
2056
+ supportsImages: true,
2038
2057
  supportsPromptCache: true,
2039
- // Claude Code does report cache tokens
2040
- supportsReasoningEffort: false,
2041
- supportsReasoningBudget: false,
2042
- requiredReasoningBudget: false,
2043
- // Claude Code manages its own tools and temperature via the CLI
2044
- supportsNativeTools: false,
2045
- supportsTemperature: false
2058
+ supportsNativeTools: true,
2059
+ defaultToolProtocol: "native",
2060
+ supportsReasoningEffort: ["disable", "low", "medium", "high"],
2061
+ reasoningEffort: "medium",
2062
+ description: "Claude Sonnet 4.5 - Balanced performance with thinking"
2046
2063
  },
2047
- "claude-haiku-4-5-20251001": {
2048
- ...anthropicModels["claude-haiku-4-5-20251001"],
2049
- supportsImages: false,
2064
+ "claude-opus-4-5": {
2065
+ maxTokens: 32768,
2066
+ contextWindow: 2e5,
2067
+ supportsImages: true,
2050
2068
  supportsPromptCache: true,
2051
- // Claude Code does report cache tokens
2052
- supportsReasoningEffort: false,
2053
- supportsReasoningBudget: false,
2054
- requiredReasoningBudget: false,
2055
- // Claude Code manages its own tools and temperature via the CLI
2056
- supportsNativeTools: false,
2057
- supportsTemperature: false
2069
+ supportsNativeTools: true,
2070
+ defaultToolProtocol: "native",
2071
+ supportsReasoningEffort: ["disable", "low", "medium", "high"],
2072
+ reasoningEffort: "medium",
2073
+ description: "Claude Opus 4.5 - Most capable with thinking"
2058
2074
  }
2059
2075
  };
2076
+ var claudeCodeDefaultModelId = "claude-sonnet-4-5";
2077
+ var MODEL_FAMILY_PATTERNS = [
2078
+ // Opus models (any version) → claude-opus-4-5
2079
+ { pattern: /opus/i, target: "claude-opus-4-5" },
2080
+ // Haiku models (any version) → claude-haiku-4-5
2081
+ { pattern: /haiku/i, target: "claude-haiku-4-5" },
2082
+ // Sonnet models (any version) → claude-sonnet-4-5
2083
+ { pattern: /sonnet/i, target: "claude-sonnet-4-5" }
2084
+ ];
2085
+ function normalizeClaudeCodeModelId(modelId) {
2086
+ if (Object.hasOwn(claudeCodeModels, modelId)) {
2087
+ return modelId;
2088
+ }
2089
+ const withoutDate = modelId.replace(DATE_SUFFIX_PATTERN, "");
2090
+ if (Object.hasOwn(claudeCodeModels, withoutDate)) {
2091
+ return withoutDate;
2092
+ }
2093
+ for (const { pattern, target } of MODEL_FAMILY_PATTERNS) {
2094
+ if (pattern.test(modelId)) {
2095
+ return target;
2096
+ }
2097
+ }
2098
+ return claudeCodeDefaultModelId;
2099
+ }
2100
+ var claudeCodeReasoningConfig = {
2101
+ low: { budgetTokens: 16e3 },
2102
+ medium: { budgetTokens: 32e3 },
2103
+ high: { budgetTokens: 64e3 }
2104
+ };
2060
2105
 
2061
2106
  // src/providers/deepseek.ts
2062
2107
  var deepSeekDefaultModelId = "deepseek-chat";
@@ -2087,6 +2132,7 @@ var deepSeekModels = {
2087
2132
  supportsPromptCache: true,
2088
2133
  supportsNativeTools: true,
2089
2134
  defaultToolProtocol: "native",
2135
+ preserveReasoning: true,
2090
2136
  inputPrice: 0.28,
2091
2137
  // $0.28 per million tokens (cache miss) - Updated Dec 9, 2025
2092
2138
  outputPrice: 0.42,
@@ -2098,7 +2144,7 @@ var deepSeekModels = {
2098
2144
  description: `DeepSeek-V3.2 (Thinking Mode) achieves performance comparable to OpenAI-o1 across math, code, and reasoning tasks. Supports Chain of Thought reasoning with up to 8K output tokens. Supports JSON output, tool calls, and chat prefix completion (beta).`
2099
2145
  }
2100
2146
  };
2101
- var DEEP_SEEK_DEFAULT_TEMPERATURE = 0;
2147
+ var DEEP_SEEK_DEFAULT_TEMPERATURE = 0.3;
2102
2148
 
2103
2149
  // src/providers/doubao.ts
2104
2150
  var doubaoDefaultModelId = "doubao-seed-1-6-250615";
@@ -2109,6 +2155,7 @@ var doubaoModels = {
2109
2155
  supportsImages: true,
2110
2156
  supportsPromptCache: true,
2111
2157
  supportsNativeTools: true,
2158
+ defaultToolProtocol: "native",
2112
2159
  inputPrice: 1e-4,
2113
2160
  // $0.0001 per million tokens (cache miss)
2114
2161
  outputPrice: 4e-4,
@@ -2125,6 +2172,7 @@ var doubaoModels = {
2125
2172
  supportsImages: true,
2126
2173
  supportsPromptCache: true,
2127
2174
  supportsNativeTools: true,
2175
+ defaultToolProtocol: "native",
2128
2176
  inputPrice: 2e-4,
2129
2177
  // $0.0002 per million tokens
2130
2178
  outputPrice: 8e-4,
@@ -2141,6 +2189,7 @@ var doubaoModels = {
2141
2189
  supportsImages: true,
2142
2190
  supportsPromptCache: true,
2143
2191
  supportsNativeTools: true,
2192
+ defaultToolProtocol: "native",
2144
2193
  inputPrice: 15e-5,
2145
2194
  // $0.00015 per million tokens
2146
2195
  outputPrice: 6e-4,
@@ -2217,6 +2266,7 @@ var fireworksModels = {
2217
2266
  supportsImages: false,
2218
2267
  supportsPromptCache: true,
2219
2268
  supportsNativeTools: true,
2269
+ defaultToolProtocol: "native",
2220
2270
  inputPrice: 0.6,
2221
2271
  outputPrice: 2.5,
2222
2272
  cacheReadsPrice: 0.15,
@@ -2228,6 +2278,7 @@ var fireworksModels = {
2228
2278
  supportsImages: false,
2229
2279
  supportsPromptCache: false,
2230
2280
  supportsNativeTools: true,
2281
+ defaultToolProtocol: "native",
2231
2282
  inputPrice: 0.6,
2232
2283
  outputPrice: 2.5,
2233
2284
  description: "Kimi K2 is a state-of-the-art mixture-of-experts (MoE) language model with 32 billion activated parameters and 1 trillion total parameters. Trained with the Muon optimizer, Kimi K2 achieves exceptional performance across frontier knowledge, reasoning, and coding tasks while being meticulously optimized for agentic capabilities."
@@ -2238,6 +2289,7 @@ var fireworksModels = {
2238
2289
  supportsImages: false,
2239
2290
  supportsPromptCache: false,
2240
2291
  supportsNativeTools: true,
2292
+ defaultToolProtocol: "native",
2241
2293
  inputPrice: 0.3,
2242
2294
  outputPrice: 1.2,
2243
2295
  description: "MiniMax M2 is a high-performance language model with 204.8K context window, optimized for long-context understanding and generation tasks."
@@ -2248,6 +2300,7 @@ var fireworksModels = {
2248
2300
  supportsImages: false,
2249
2301
  supportsPromptCache: false,
2250
2302
  supportsNativeTools: true,
2303
+ defaultToolProtocol: "native",
2251
2304
  inputPrice: 0.22,
2252
2305
  outputPrice: 0.88,
2253
2306
  description: "Latest Qwen3 thinking model, competitive against the best closed source models in Jul 2025."
@@ -2258,6 +2311,7 @@ var fireworksModels = {
2258
2311
  supportsImages: false,
2259
2312
  supportsPromptCache: false,
2260
2313
  supportsNativeTools: true,
2314
+ defaultToolProtocol: "native",
2261
2315
  inputPrice: 0.45,
2262
2316
  outputPrice: 1.8,
2263
2317
  description: "Qwen3's most agentic code model to date."
@@ -2268,6 +2322,7 @@ var fireworksModels = {
2268
2322
  supportsImages: false,
2269
2323
  supportsPromptCache: false,
2270
2324
  supportsNativeTools: true,
2325
+ defaultToolProtocol: "native",
2271
2326
  inputPrice: 3,
2272
2327
  outputPrice: 8,
2273
2328
  description: "05/28 updated checkpoint of Deepseek R1. Its overall performance is now approaching that of leading models, such as O3 and Gemini 2.5 Pro. Compared to the previous version, the upgraded model shows significant improvements in handling complex reasoning tasks, and this version also offers a reduced hallucination rate, enhanced support for function calling, and better experience for vibe coding. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
@@ -2278,6 +2333,7 @@ var fireworksModels = {
2278
2333
  supportsImages: false,
2279
2334
  supportsPromptCache: false,
2280
2335
  supportsNativeTools: true,
2336
+ defaultToolProtocol: "native",
2281
2337
  inputPrice: 0.9,
2282
2338
  outputPrice: 0.9,
2283
2339
  description: "A strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token from Deepseek. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
@@ -2288,6 +2344,7 @@ var fireworksModels = {
2288
2344
  supportsImages: false,
2289
2345
  supportsPromptCache: false,
2290
2346
  supportsNativeTools: true,
2347
+ defaultToolProtocol: "native",
2291
2348
  inputPrice: 0.56,
2292
2349
  outputPrice: 1.68,
2293
2350
  description: "DeepSeek v3.1 is an improved version of the v3 model with enhanced performance, better reasoning capabilities, and improved code generation. This Mixture-of-Experts (MoE) model maintains the same 671B total parameters with 37B activated per token."
@@ -2298,6 +2355,7 @@ var fireworksModels = {
2298
2355
  supportsImages: false,
2299
2356
  supportsPromptCache: false,
2300
2357
  supportsNativeTools: true,
2358
+ defaultToolProtocol: "native",
2301
2359
  inputPrice: 0.55,
2302
2360
  outputPrice: 2.19,
2303
2361
  description: "Z.ai GLM-4.5 with 355B total parameters and 32B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
@@ -2308,6 +2366,7 @@ var fireworksModels = {
2308
2366
  supportsImages: false,
2309
2367
  supportsPromptCache: false,
2310
2368
  supportsNativeTools: true,
2369
+ defaultToolProtocol: "native",
2311
2370
  inputPrice: 0.55,
2312
2371
  outputPrice: 2.19,
2313
2372
  description: "Z.ai GLM-4.5-Air with 106B total parameters and 12B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
@@ -2318,6 +2377,7 @@ var fireworksModels = {
2318
2377
  supportsImages: false,
2319
2378
  supportsPromptCache: false,
2320
2379
  supportsNativeTools: true,
2380
+ defaultToolProtocol: "native",
2321
2381
  inputPrice: 0.55,
2322
2382
  outputPrice: 2.19,
2323
2383
  description: "Z.ai GLM-4.6 is an advanced coding model with exceptional performance on complex programming tasks. Features improved reasoning capabilities and enhanced code generation quality, making it ideal for software development workflows."
@@ -2328,6 +2388,7 @@ var fireworksModels = {
2328
2388
  supportsImages: false,
2329
2389
  supportsPromptCache: false,
2330
2390
  supportsNativeTools: true,
2391
+ defaultToolProtocol: "native",
2331
2392
  inputPrice: 0.07,
2332
2393
  outputPrice: 0.3,
2333
2394
  description: "OpenAI gpt-oss-20b: Compact model for local/edge deployments. Optimized for low-latency and resource-constrained environments with chain-of-thought output, adjustable reasoning, and agentic workflows."
@@ -2338,6 +2399,7 @@ var fireworksModels = {
2338
2399
  supportsImages: false,
2339
2400
  supportsPromptCache: false,
2340
2401
  supportsNativeTools: true,
2402
+ defaultToolProtocol: "native",
2341
2403
  inputPrice: 0.15,
2342
2404
  outputPrice: 0.6,
2343
2405
  description: "OpenAI gpt-oss-120b: Production-grade, general-purpose model that fits on a single H100 GPU. Features complex reasoning, configurable effort, full chain-of-thought transparency, and supports function calling, tool use, and structured outputs."
@@ -2345,16 +2407,18 @@ var fireworksModels = {
2345
2407
  };
2346
2408
 
2347
2409
  // src/providers/gemini.ts
2348
- var geminiDefaultModelId = "gemini-2.5-pro";
2410
+ var geminiDefaultModelId = "gemini-3-pro-preview";
2349
2411
  var geminiModels = {
2350
2412
  "gemini-3-pro-preview": {
2351
2413
  maxTokens: 65536,
2352
2414
  contextWindow: 1048576,
2353
2415
  supportsImages: true,
2354
2416
  supportsNativeTools: true,
2417
+ defaultToolProtocol: "native",
2355
2418
  supportsPromptCache: true,
2356
2419
  supportsReasoningEffort: ["low", "high"],
2357
2420
  reasoningEffort: "low",
2421
+ includedTools: ["write_file", "edit_file"],
2358
2422
  supportsTemperature: true,
2359
2423
  defaultTemperature: 1,
2360
2424
  inputPrice: 4,
@@ -2372,13 +2436,32 @@ var geminiModels = {
2372
2436
  }
2373
2437
  ]
2374
2438
  },
2439
+ "gemini-3-flash-preview": {
2440
+ maxTokens: 65536,
2441
+ contextWindow: 1048576,
2442
+ supportsImages: true,
2443
+ supportsNativeTools: true,
2444
+ defaultToolProtocol: "native",
2445
+ supportsPromptCache: true,
2446
+ supportsReasoningEffort: ["minimal", "low", "medium", "high"],
2447
+ reasoningEffort: "medium",
2448
+ includedTools: ["write_file", "edit_file"],
2449
+ supportsTemperature: true,
2450
+ defaultTemperature: 1,
2451
+ inputPrice: 0.3,
2452
+ outputPrice: 2.5,
2453
+ cacheReadsPrice: 0.075,
2454
+ cacheWritesPrice: 1
2455
+ },
2375
2456
  // 2.5 Pro models
2376
2457
  "gemini-2.5-pro": {
2377
2458
  maxTokens: 64e3,
2378
2459
  contextWindow: 1048576,
2379
2460
  supportsImages: true,
2380
2461
  supportsNativeTools: true,
2462
+ defaultToolProtocol: "native",
2381
2463
  supportsPromptCache: true,
2464
+ includedTools: ["write_file", "edit_file"],
2382
2465
  inputPrice: 2.5,
2383
2466
  // This is the pricing for prompts above 200k tokens.
2384
2467
  outputPrice: 15,
@@ -2407,7 +2490,9 @@ var geminiModels = {
2407
2490
  contextWindow: 1048576,
2408
2491
  supportsImages: true,
2409
2492
  supportsNativeTools: true,
2493
+ defaultToolProtocol: "native",
2410
2494
  supportsPromptCache: true,
2495
+ includedTools: ["write_file", "edit_file"],
2411
2496
  inputPrice: 2.5,
2412
2497
  // This is the pricing for prompts above 200k tokens.
2413
2498
  outputPrice: 15,
@@ -2435,7 +2520,9 @@ var geminiModels = {
2435
2520
  contextWindow: 1048576,
2436
2521
  supportsImages: true,
2437
2522
  supportsNativeTools: true,
2523
+ defaultToolProtocol: "native",
2438
2524
  supportsPromptCache: true,
2525
+ includedTools: ["write_file", "edit_file"],
2439
2526
  inputPrice: 2.5,
2440
2527
  // This is the pricing for prompts above 200k tokens.
2441
2528
  outputPrice: 15,
@@ -2461,7 +2548,9 @@ var geminiModels = {
2461
2548
  contextWindow: 1048576,
2462
2549
  supportsImages: true,
2463
2550
  supportsNativeTools: true,
2551
+ defaultToolProtocol: "native",
2464
2552
  supportsPromptCache: true,
2553
+ includedTools: ["write_file", "edit_file"],
2465
2554
  inputPrice: 2.5,
2466
2555
  // This is the pricing for prompts above 200k tokens.
2467
2556
  outputPrice: 15,
@@ -2490,7 +2579,9 @@ var geminiModels = {
2490
2579
  contextWindow: 1048576,
2491
2580
  supportsImages: true,
2492
2581
  supportsNativeTools: true,
2582
+ defaultToolProtocol: "native",
2493
2583
  supportsPromptCache: true,
2584
+ includedTools: ["write_file", "edit_file"],
2494
2585
  inputPrice: 0.3,
2495
2586
  outputPrice: 2.5,
2496
2587
  cacheReadsPrice: 0.075,
@@ -2503,7 +2594,9 @@ var geminiModels = {
2503
2594
  contextWindow: 1048576,
2504
2595
  supportsImages: true,
2505
2596
  supportsNativeTools: true,
2597
+ defaultToolProtocol: "native",
2506
2598
  supportsPromptCache: true,
2599
+ includedTools: ["write_file", "edit_file"],
2507
2600
  inputPrice: 0.3,
2508
2601
  outputPrice: 2.5,
2509
2602
  cacheReadsPrice: 0.075,
@@ -2516,7 +2609,9 @@ var geminiModels = {
2516
2609
  contextWindow: 1048576,
2517
2610
  supportsImages: true,
2518
2611
  supportsNativeTools: true,
2612
+ defaultToolProtocol: "native",
2519
2613
  supportsPromptCache: true,
2614
+ includedTools: ["write_file", "edit_file"],
2520
2615
  inputPrice: 0.3,
2521
2616
  outputPrice: 2.5,
2522
2617
  cacheReadsPrice: 0.075,
@@ -2530,7 +2625,9 @@ var geminiModels = {
2530
2625
  contextWindow: 1048576,
2531
2626
  supportsImages: true,
2532
2627
  supportsNativeTools: true,
2628
+ defaultToolProtocol: "native",
2533
2629
  supportsPromptCache: true,
2630
+ includedTools: ["write_file", "edit_file"],
2534
2631
  inputPrice: 0.1,
2535
2632
  outputPrice: 0.4,
2536
2633
  cacheReadsPrice: 0.025,
@@ -2543,7 +2640,9 @@ var geminiModels = {
2543
2640
  contextWindow: 1048576,
2544
2641
  supportsImages: true,
2545
2642
  supportsNativeTools: true,
2643
+ defaultToolProtocol: "native",
2546
2644
  supportsPromptCache: true,
2645
+ includedTools: ["write_file", "edit_file"],
2547
2646
  inputPrice: 0.1,
2548
2647
  outputPrice: 0.4,
2549
2648
  cacheReadsPrice: 0.025,
@@ -2563,6 +2662,7 @@ var groqModels = {
2563
2662
  supportsImages: false,
2564
2663
  supportsPromptCache: false,
2565
2664
  supportsNativeTools: true,
2665
+ defaultToolProtocol: "native",
2566
2666
  inputPrice: 0.05,
2567
2667
  outputPrice: 0.08,
2568
2668
  description: "Meta Llama 3.1 8B Instant model, 128K context."
@@ -2573,6 +2673,7 @@ var groqModels = {
2573
2673
  supportsImages: false,
2574
2674
  supportsPromptCache: false,
2575
2675
  supportsNativeTools: true,
2676
+ defaultToolProtocol: "native",
2576
2677
  inputPrice: 0.59,
2577
2678
  outputPrice: 0.79,
2578
2679
  description: "Meta Llama 3.3 70B Versatile model, 128K context."
@@ -2583,6 +2684,7 @@ var groqModels = {
2583
2684
  supportsImages: false,
2584
2685
  supportsPromptCache: false,
2585
2686
  supportsNativeTools: true,
2687
+ defaultToolProtocol: "native",
2586
2688
  inputPrice: 0.11,
2587
2689
  outputPrice: 0.34,
2588
2690
  description: "Meta Llama 4 Scout 17B Instruct model, 128K context."
@@ -2620,6 +2722,7 @@ var groqModels = {
2620
2722
  supportsImages: false,
2621
2723
  supportsPromptCache: false,
2622
2724
  supportsNativeTools: true,
2725
+ defaultToolProtocol: "native",
2623
2726
  inputPrice: 0.29,
2624
2727
  outputPrice: 0.59,
2625
2728
  description: "Alibaba Qwen 3 32B model, 128K context."
@@ -2650,6 +2753,7 @@ var groqModels = {
2650
2753
  supportsImages: false,
2651
2754
  supportsPromptCache: true,
2652
2755
  supportsNativeTools: true,
2756
+ defaultToolProtocol: "native",
2653
2757
  inputPrice: 0.6,
2654
2758
  outputPrice: 2.5,
2655
2759
  cacheReadsPrice: 0.15,
@@ -2661,6 +2765,7 @@ var groqModels = {
2661
2765
  supportsImages: false,
2662
2766
  supportsPromptCache: false,
2663
2767
  supportsNativeTools: true,
2768
+ defaultToolProtocol: "native",
2664
2769
  inputPrice: 0.15,
2665
2770
  outputPrice: 0.75,
2666
2771
  description: "GPT-OSS 120B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 128 experts."
@@ -2671,6 +2776,7 @@ var groqModels = {
2671
2776
  supportsImages: false,
2672
2777
  supportsPromptCache: false,
2673
2778
  supportsNativeTools: true,
2779
+ defaultToolProtocol: "native",
2674
2780
  inputPrice: 0.1,
2675
2781
  outputPrice: 0.5,
2676
2782
  description: "GPT-OSS 20B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 32 experts."
@@ -2734,6 +2840,7 @@ var litellmDefaultModelInfo = {
2734
2840
  supportsImages: true,
2735
2841
  supportsPromptCache: true,
2736
2842
  supportsNativeTools: true,
2843
+ defaultToolProtocol: "native",
2737
2844
  inputPrice: 3,
2738
2845
  outputPrice: 15,
2739
2846
  cacheWritesPrice: 3.75,
@@ -2764,6 +2871,7 @@ var mistralModels = {
2764
2871
  supportsImages: true,
2765
2872
  supportsPromptCache: false,
2766
2873
  supportsNativeTools: true,
2874
+ defaultToolProtocol: "native",
2767
2875
  inputPrice: 2,
2768
2876
  outputPrice: 5
2769
2877
  },
@@ -2773,6 +2881,7 @@ var mistralModels = {
2773
2881
  supportsImages: true,
2774
2882
  supportsPromptCache: false,
2775
2883
  supportsNativeTools: true,
2884
+ defaultToolProtocol: "native",
2776
2885
  inputPrice: 0.4,
2777
2886
  outputPrice: 2
2778
2887
  },
@@ -2782,6 +2891,7 @@ var mistralModels = {
2782
2891
  supportsImages: true,
2783
2892
  supportsPromptCache: false,
2784
2893
  supportsNativeTools: true,
2894
+ defaultToolProtocol: "native",
2785
2895
  inputPrice: 0.4,
2786
2896
  outputPrice: 2
2787
2897
  },
@@ -2791,6 +2901,7 @@ var mistralModels = {
2791
2901
  supportsImages: false,
2792
2902
  supportsPromptCache: false,
2793
2903
  supportsNativeTools: true,
2904
+ defaultToolProtocol: "native",
2794
2905
  inputPrice: 0.3,
2795
2906
  outputPrice: 0.9
2796
2907
  },
@@ -2800,6 +2911,7 @@ var mistralModels = {
2800
2911
  supportsImages: false,
2801
2912
  supportsPromptCache: false,
2802
2913
  supportsNativeTools: true,
2914
+ defaultToolProtocol: "native",
2803
2915
  inputPrice: 2,
2804
2916
  outputPrice: 6
2805
2917
  },
@@ -2809,6 +2921,7 @@ var mistralModels = {
2809
2921
  supportsImages: false,
2810
2922
  supportsPromptCache: false,
2811
2923
  supportsNativeTools: true,
2924
+ defaultToolProtocol: "native",
2812
2925
  inputPrice: 0.1,
2813
2926
  outputPrice: 0.1
2814
2927
  },
@@ -2818,6 +2931,7 @@ var mistralModels = {
2818
2931
  supportsImages: false,
2819
2932
  supportsPromptCache: false,
2820
2933
  supportsNativeTools: true,
2934
+ defaultToolProtocol: "native",
2821
2935
  inputPrice: 0.04,
2822
2936
  outputPrice: 0.04
2823
2937
  },
@@ -2827,6 +2941,7 @@ var mistralModels = {
2827
2941
  supportsImages: false,
2828
2942
  supportsPromptCache: false,
2829
2943
  supportsNativeTools: true,
2944
+ defaultToolProtocol: "native",
2830
2945
  inputPrice: 0.2,
2831
2946
  outputPrice: 0.6
2832
2947
  },
@@ -2836,6 +2951,7 @@ var mistralModels = {
2836
2951
  supportsImages: true,
2837
2952
  supportsPromptCache: false,
2838
2953
  supportsNativeTools: true,
2954
+ defaultToolProtocol: "native",
2839
2955
  inputPrice: 2,
2840
2956
  outputPrice: 6
2841
2957
  }
@@ -2851,6 +2967,7 @@ var moonshotModels = {
2851
2967
  supportsImages: false,
2852
2968
  supportsPromptCache: true,
2853
2969
  supportsNativeTools: true,
2970
+ defaultToolProtocol: "native",
2854
2971
  inputPrice: 0.6,
2855
2972
  // $0.60 per million tokens (cache miss)
2856
2973
  outputPrice: 2.5,
@@ -2867,6 +2984,7 @@ var moonshotModels = {
2867
2984
  supportsImages: false,
2868
2985
  supportsPromptCache: true,
2869
2986
  supportsNativeTools: true,
2987
+ defaultToolProtocol: "native",
2870
2988
  inputPrice: 0.6,
2871
2989
  outputPrice: 2.5,
2872
2990
  cacheReadsPrice: 0.15,
@@ -2878,6 +2996,7 @@ var moonshotModels = {
2878
2996
  supportsImages: false,
2879
2997
  supportsPromptCache: true,
2880
2998
  supportsNativeTools: true,
2999
+ defaultToolProtocol: "native",
2881
3000
  inputPrice: 2.4,
2882
3001
  // $2.40 per million tokens (cache miss)
2883
3002
  outputPrice: 10,
@@ -2897,6 +3016,7 @@ var moonshotModels = {
2897
3016
  // Text-only (no image/vision support)
2898
3017
  supportsPromptCache: true,
2899
3018
  supportsNativeTools: true,
3019
+ defaultToolProtocol: "native",
2900
3020
  inputPrice: 0.6,
2901
3021
  // $0.60 per million tokens (cache miss)
2902
3022
  outputPrice: 2.5,
@@ -2936,6 +3056,7 @@ var openAiNativeModels = {
2936
3056
  maxTokens: 128e3,
2937
3057
  contextWindow: 4e5,
2938
3058
  supportsNativeTools: true,
3059
+ defaultToolProtocol: "native",
2939
3060
  includedTools: ["apply_patch"],
2940
3061
  excludedTools: ["apply_diff", "write_to_file"],
2941
3062
  supportsImages: true,
@@ -2954,6 +3075,7 @@ var openAiNativeModels = {
2954
3075
  maxTokens: 128e3,
2955
3076
  contextWindow: 4e5,
2956
3077
  supportsNativeTools: true,
3078
+ defaultToolProtocol: "native",
2957
3079
  includedTools: ["apply_patch"],
2958
3080
  excludedTools: ["apply_diff", "write_to_file"],
2959
3081
  supportsImages: true,
@@ -2976,6 +3098,7 @@ var openAiNativeModels = {
2976
3098
  maxTokens: 16384,
2977
3099
  contextWindow: 128e3,
2978
3100
  supportsNativeTools: true,
3101
+ defaultToolProtocol: "native",
2979
3102
  includedTools: ["apply_patch"],
2980
3103
  excludedTools: ["apply_diff", "write_to_file"],
2981
3104
  supportsImages: true,
@@ -2989,6 +3112,7 @@ var openAiNativeModels = {
2989
3112
  maxTokens: 128e3,
2990
3113
  contextWindow: 4e5,
2991
3114
  supportsNativeTools: true,
3115
+ defaultToolProtocol: "native",
2992
3116
  includedTools: ["apply_patch"],
2993
3117
  excludedTools: ["apply_diff", "write_to_file"],
2994
3118
  supportsImages: true,
@@ -3011,6 +3135,7 @@ var openAiNativeModels = {
3011
3135
  maxTokens: 128e3,
3012
3136
  contextWindow: 4e5,
3013
3137
  supportsNativeTools: true,
3138
+ defaultToolProtocol: "native",
3014
3139
  includedTools: ["apply_patch"],
3015
3140
  excludedTools: ["apply_diff", "write_to_file"],
3016
3141
  supportsImages: true,
@@ -3029,6 +3154,7 @@ var openAiNativeModels = {
3029
3154
  maxTokens: 128e3,
3030
3155
  contextWindow: 4e5,
3031
3156
  supportsNativeTools: true,
3157
+ defaultToolProtocol: "native",
3032
3158
  includedTools: ["apply_patch"],
3033
3159
  excludedTools: ["apply_diff", "write_to_file"],
3034
3160
  supportsImages: true,
@@ -3046,6 +3172,7 @@ var openAiNativeModels = {
3046
3172
  maxTokens: 128e3,
3047
3173
  contextWindow: 4e5,
3048
3174
  supportsNativeTools: true,
3175
+ defaultToolProtocol: "native",
3049
3176
  includedTools: ["apply_patch"],
3050
3177
  excludedTools: ["apply_diff", "write_to_file"],
3051
3178
  supportsImages: true,
@@ -3067,6 +3194,7 @@ var openAiNativeModels = {
3067
3194
  maxTokens: 128e3,
3068
3195
  contextWindow: 4e5,
3069
3196
  supportsNativeTools: true,
3197
+ defaultToolProtocol: "native",
3070
3198
  includedTools: ["apply_patch"],
3071
3199
  excludedTools: ["apply_diff", "write_to_file"],
3072
3200
  supportsImages: true,
@@ -3088,6 +3216,7 @@ var openAiNativeModels = {
3088
3216
  maxTokens: 128e3,
3089
3217
  contextWindow: 4e5,
3090
3218
  supportsNativeTools: true,
3219
+ defaultToolProtocol: "native",
3091
3220
  includedTools: ["apply_patch"],
3092
3221
  excludedTools: ["apply_diff", "write_to_file"],
3093
3222
  supportsImages: true,
@@ -3105,6 +3234,7 @@ var openAiNativeModels = {
3105
3234
  maxTokens: 128e3,
3106
3235
  contextWindow: 4e5,
3107
3236
  supportsNativeTools: true,
3237
+ defaultToolProtocol: "native",
3108
3238
  includedTools: ["apply_patch"],
3109
3239
  excludedTools: ["apply_diff", "write_to_file"],
3110
3240
  supportsImages: true,
@@ -3123,6 +3253,7 @@ var openAiNativeModels = {
3123
3253
  maxTokens: 128e3,
3124
3254
  contextWindow: 4e5,
3125
3255
  supportsNativeTools: true,
3256
+ defaultToolProtocol: "native",
3126
3257
  includedTools: ["apply_patch"],
3127
3258
  excludedTools: ["apply_diff", "write_to_file"],
3128
3259
  supportsImages: true,
@@ -3136,6 +3267,7 @@ var openAiNativeModels = {
3136
3267
  maxTokens: 32768,
3137
3268
  contextWindow: 1047576,
3138
3269
  supportsNativeTools: true,
3270
+ defaultToolProtocol: "native",
3139
3271
  includedTools: ["apply_patch"],
3140
3272
  excludedTools: ["apply_diff", "write_to_file"],
3141
3273
  supportsImages: true,
@@ -3152,6 +3284,7 @@ var openAiNativeModels = {
3152
3284
  maxTokens: 32768,
3153
3285
  contextWindow: 1047576,
3154
3286
  supportsNativeTools: true,
3287
+ defaultToolProtocol: "native",
3155
3288
  includedTools: ["apply_patch"],
3156
3289
  excludedTools: ["apply_diff", "write_to_file"],
3157
3290
  supportsImages: true,
@@ -3168,6 +3301,7 @@ var openAiNativeModels = {
3168
3301
  maxTokens: 32768,
3169
3302
  contextWindow: 1047576,
3170
3303
  supportsNativeTools: true,
3304
+ defaultToolProtocol: "native",
3171
3305
  includedTools: ["apply_patch"],
3172
3306
  excludedTools: ["apply_diff", "write_to_file"],
3173
3307
  supportsImages: true,
@@ -3184,6 +3318,7 @@ var openAiNativeModels = {
3184
3318
  maxTokens: 1e5,
3185
3319
  contextWindow: 2e5,
3186
3320
  supportsNativeTools: true,
3321
+ defaultToolProtocol: "native",
3187
3322
  supportsImages: true,
3188
3323
  supportsPromptCache: true,
3189
3324
  inputPrice: 2,
@@ -3201,6 +3336,7 @@ var openAiNativeModels = {
3201
3336
  maxTokens: 1e5,
3202
3337
  contextWindow: 2e5,
3203
3338
  supportsNativeTools: true,
3339
+ defaultToolProtocol: "native",
3204
3340
  supportsImages: true,
3205
3341
  supportsPromptCache: true,
3206
3342
  inputPrice: 2,
@@ -3213,6 +3349,7 @@ var openAiNativeModels = {
3213
3349
  maxTokens: 1e5,
3214
3350
  contextWindow: 2e5,
3215
3351
  supportsNativeTools: true,
3352
+ defaultToolProtocol: "native",
3216
3353
  supportsImages: true,
3217
3354
  supportsPromptCache: true,
3218
3355
  inputPrice: 2,
@@ -3225,6 +3362,7 @@ var openAiNativeModels = {
3225
3362
  maxTokens: 1e5,
3226
3363
  contextWindow: 2e5,
3227
3364
  supportsNativeTools: true,
3365
+ defaultToolProtocol: "native",
3228
3366
  supportsImages: true,
3229
3367
  supportsPromptCache: true,
3230
3368
  inputPrice: 1.1,
@@ -3242,6 +3380,7 @@ var openAiNativeModels = {
3242
3380
  maxTokens: 1e5,
3243
3381
  contextWindow: 2e5,
3244
3382
  supportsNativeTools: true,
3383
+ defaultToolProtocol: "native",
3245
3384
  supportsImages: true,
3246
3385
  supportsPromptCache: true,
3247
3386
  inputPrice: 1.1,
@@ -3254,6 +3393,7 @@ var openAiNativeModels = {
3254
3393
  maxTokens: 1e5,
3255
3394
  contextWindow: 2e5,
3256
3395
  supportsNativeTools: true,
3396
+ defaultToolProtocol: "native",
3257
3397
  supportsImages: true,
3258
3398
  supportsPromptCache: true,
3259
3399
  inputPrice: 1.1,
@@ -3266,6 +3406,7 @@ var openAiNativeModels = {
3266
3406
  maxTokens: 1e5,
3267
3407
  contextWindow: 2e5,
3268
3408
  supportsNativeTools: true,
3409
+ defaultToolProtocol: "native",
3269
3410
  supportsImages: false,
3270
3411
  supportsPromptCache: true,
3271
3412
  inputPrice: 1.1,
@@ -3279,6 +3420,7 @@ var openAiNativeModels = {
3279
3420
  maxTokens: 1e5,
3280
3421
  contextWindow: 2e5,
3281
3422
  supportsNativeTools: true,
3423
+ defaultToolProtocol: "native",
3282
3424
  supportsImages: false,
3283
3425
  supportsPromptCache: true,
3284
3426
  inputPrice: 1.1,
@@ -3291,6 +3433,7 @@ var openAiNativeModels = {
3291
3433
  maxTokens: 1e5,
3292
3434
  contextWindow: 2e5,
3293
3435
  supportsNativeTools: true,
3436
+ defaultToolProtocol: "native",
3294
3437
  supportsImages: false,
3295
3438
  supportsPromptCache: true,
3296
3439
  inputPrice: 1.1,
@@ -3303,6 +3446,7 @@ var openAiNativeModels = {
3303
3446
  maxTokens: 1e5,
3304
3447
  contextWindow: 2e5,
3305
3448
  supportsNativeTools: true,
3449
+ defaultToolProtocol: "native",
3306
3450
  supportsImages: true,
3307
3451
  supportsPromptCache: true,
3308
3452
  inputPrice: 15,
@@ -3314,6 +3458,7 @@ var openAiNativeModels = {
3314
3458
  maxTokens: 32768,
3315
3459
  contextWindow: 128e3,
3316
3460
  supportsNativeTools: true,
3461
+ defaultToolProtocol: "native",
3317
3462
  supportsImages: true,
3318
3463
  supportsPromptCache: true,
3319
3464
  inputPrice: 15,
@@ -3325,6 +3470,7 @@ var openAiNativeModels = {
3325
3470
  maxTokens: 65536,
3326
3471
  contextWindow: 128e3,
3327
3472
  supportsNativeTools: true,
3473
+ defaultToolProtocol: "native",
3328
3474
  supportsImages: true,
3329
3475
  supportsPromptCache: true,
3330
3476
  inputPrice: 1.1,
@@ -3336,6 +3482,7 @@ var openAiNativeModels = {
3336
3482
  maxTokens: 16384,
3337
3483
  contextWindow: 128e3,
3338
3484
  supportsNativeTools: true,
3485
+ defaultToolProtocol: "native",
3339
3486
  supportsImages: true,
3340
3487
  supportsPromptCache: true,
3341
3488
  inputPrice: 2.5,
@@ -3350,6 +3497,7 @@ var openAiNativeModels = {
3350
3497
  maxTokens: 16384,
3351
3498
  contextWindow: 128e3,
3352
3499
  supportsNativeTools: true,
3500
+ defaultToolProtocol: "native",
3353
3501
  supportsImages: true,
3354
3502
  supportsPromptCache: true,
3355
3503
  inputPrice: 0.15,
@@ -3364,6 +3512,7 @@ var openAiNativeModels = {
3364
3512
  maxTokens: 16384,
3365
3513
  contextWindow: 2e5,
3366
3514
  supportsNativeTools: true,
3515
+ defaultToolProtocol: "native",
3367
3516
  supportsImages: false,
3368
3517
  supportsPromptCache: false,
3369
3518
  inputPrice: 1.5,
@@ -3377,6 +3526,7 @@ var openAiNativeModels = {
3377
3526
  maxTokens: 128e3,
3378
3527
  contextWindow: 4e5,
3379
3528
  supportsNativeTools: true,
3529
+ defaultToolProtocol: "native",
3380
3530
  includedTools: ["apply_patch"],
3381
3531
  excludedTools: ["apply_diff", "write_to_file"],
3382
3532
  supportsImages: true,
@@ -3398,6 +3548,7 @@ var openAiNativeModels = {
3398
3548
  maxTokens: 128e3,
3399
3549
  contextWindow: 4e5,
3400
3550
  supportsNativeTools: true,
3551
+ defaultToolProtocol: "native",
3401
3552
  includedTools: ["apply_patch"],
3402
3553
  excludedTools: ["apply_diff", "write_to_file"],
3403
3554
  supportsImages: true,
@@ -3419,6 +3570,7 @@ var openAiNativeModels = {
3419
3570
  maxTokens: 128e3,
3420
3571
  contextWindow: 4e5,
3421
3572
  supportsNativeTools: true,
3573
+ defaultToolProtocol: "native",
3422
3574
  includedTools: ["apply_patch"],
3423
3575
  excludedTools: ["apply_diff", "write_to_file"],
3424
3576
  supportsImages: true,
@@ -3441,7 +3593,8 @@ var openAiModelInfoSaneDefaults = {
3441
3593
  supportsPromptCache: false,
3442
3594
  inputPrice: 0,
3443
3595
  outputPrice: 0,
3444
- supportsNativeTools: true
3596
+ supportsNativeTools: true,
3597
+ defaultToolProtocol: "native"
3445
3598
  };
3446
3599
  var azureOpenAiDefaultApiVersion = "2024-08-01-preview";
3447
3600
  var OPENAI_NATIVE_DEFAULT_TEMPERATURE = 0;
@@ -3554,6 +3707,8 @@ var requestyDefaultModelInfo = {
3554
3707
  contextWindow: 2e5,
3555
3708
  supportsImages: true,
3556
3709
  supportsPromptCache: true,
3710
+ supportsNativeTools: true,
3711
+ defaultToolProtocol: "native",
3557
3712
  inputPrice: 3,
3558
3713
  outputPrice: 15,
3559
3714
  cacheWritesPrice: 3.75,
@@ -3610,6 +3765,7 @@ var sambaNovaModels = {
3610
3765
  supportsImages: false,
3611
3766
  supportsPromptCache: false,
3612
3767
  supportsNativeTools: true,
3768
+ defaultToolProtocol: "native",
3613
3769
  inputPrice: 0.1,
3614
3770
  outputPrice: 0.2,
3615
3771
  description: "Meta Llama 3.1 8B Instruct model with 16K context window."
@@ -3620,6 +3776,7 @@ var sambaNovaModels = {
3620
3776
  supportsImages: false,
3621
3777
  supportsPromptCache: false,
3622
3778
  supportsNativeTools: true,
3779
+ defaultToolProtocol: "native",
3623
3780
  inputPrice: 0.6,
3624
3781
  outputPrice: 1.2,
3625
3782
  description: "Meta Llama 3.3 70B Instruct model with 128K context window."
@@ -3631,6 +3788,7 @@ var sambaNovaModels = {
3631
3788
  supportsPromptCache: false,
3632
3789
  supportsReasoningBudget: true,
3633
3790
  supportsNativeTools: true,
3791
+ defaultToolProtocol: "native",
3634
3792
  inputPrice: 5,
3635
3793
  outputPrice: 7,
3636
3794
  description: "DeepSeek R1 reasoning model with 32K context window."
@@ -3641,6 +3799,7 @@ var sambaNovaModels = {
3641
3799
  supportsImages: false,
3642
3800
  supportsPromptCache: false,
3643
3801
  supportsNativeTools: true,
3802
+ defaultToolProtocol: "native",
3644
3803
  inputPrice: 3,
3645
3804
  outputPrice: 4.5,
3646
3805
  description: "DeepSeek V3 model with 32K context window."
@@ -3651,6 +3810,7 @@ var sambaNovaModels = {
3651
3810
  supportsImages: false,
3652
3811
  supportsPromptCache: false,
3653
3812
  supportsNativeTools: true,
3813
+ defaultToolProtocol: "native",
3654
3814
  inputPrice: 3,
3655
3815
  outputPrice: 4.5,
3656
3816
  description: "DeepSeek V3.1 model with 32K context window."
@@ -3670,6 +3830,7 @@ var sambaNovaModels = {
3670
3830
  supportsImages: true,
3671
3831
  supportsPromptCache: false,
3672
3832
  supportsNativeTools: true,
3833
+ defaultToolProtocol: "native",
3673
3834
  inputPrice: 0.63,
3674
3835
  outputPrice: 1.8,
3675
3836
  description: "Meta Llama 4 Maverick 17B 128E Instruct model with 128K context window."
@@ -3689,6 +3850,7 @@ var sambaNovaModels = {
3689
3850
  supportsImages: false,
3690
3851
  supportsPromptCache: false,
3691
3852
  supportsNativeTools: true,
3853
+ defaultToolProtocol: "native",
3692
3854
  inputPrice: 0.4,
3693
3855
  outputPrice: 0.8,
3694
3856
  description: "Alibaba Qwen 3 32B model with 8K context window."
@@ -3699,6 +3861,7 @@ var sambaNovaModels = {
3699
3861
  supportsImages: false,
3700
3862
  supportsPromptCache: false,
3701
3863
  supportsNativeTools: true,
3864
+ defaultToolProtocol: "native",
3702
3865
  inputPrice: 0.22,
3703
3866
  outputPrice: 0.59,
3704
3867
  description: "OpenAI gpt oss 120b model with 128k context window."
@@ -3727,9 +3890,11 @@ var vertexModels = {
3727
3890
  contextWindow: 1048576,
3728
3891
  supportsImages: true,
3729
3892
  supportsNativeTools: true,
3893
+ defaultToolProtocol: "native",
3730
3894
  supportsPromptCache: true,
3731
3895
  supportsReasoningEffort: ["low", "high"],
3732
3896
  reasoningEffort: "low",
3897
+ includedTools: ["write_file", "edit_file"],
3733
3898
  supportsTemperature: true,
3734
3899
  defaultTemperature: 1,
3735
3900
  inputPrice: 4,
@@ -3747,12 +3912,31 @@ var vertexModels = {
3747
3912
  }
3748
3913
  ]
3749
3914
  },
3915
+ "gemini-3-flash-preview": {
3916
+ maxTokens: 65536,
3917
+ contextWindow: 1048576,
3918
+ supportsImages: true,
3919
+ supportsNativeTools: true,
3920
+ defaultToolProtocol: "native",
3921
+ supportsPromptCache: true,
3922
+ supportsReasoningEffort: ["minimal", "low", "medium", "high"],
3923
+ reasoningEffort: "medium",
3924
+ includedTools: ["write_file", "edit_file"],
3925
+ supportsTemperature: true,
3926
+ defaultTemperature: 1,
3927
+ inputPrice: 0.3,
3928
+ outputPrice: 2.5,
3929
+ cacheReadsPrice: 0.075,
3930
+ cacheWritesPrice: 1
3931
+ },
3750
3932
  "gemini-2.5-flash-preview-05-20:thinking": {
3751
3933
  maxTokens: 65535,
3752
3934
  contextWindow: 1048576,
3753
3935
  supportsImages: true,
3754
3936
  supportsNativeTools: true,
3937
+ defaultToolProtocol: "native",
3755
3938
  supportsPromptCache: true,
3939
+ includedTools: ["write_file", "edit_file"],
3756
3940
  inputPrice: 0.15,
3757
3941
  outputPrice: 3.5,
3758
3942
  maxThinkingTokens: 24576,
@@ -3764,7 +3948,9 @@ var vertexModels = {
3764
3948
  contextWindow: 1048576,
3765
3949
  supportsImages: true,
3766
3950
  supportsNativeTools: true,
3951
+ defaultToolProtocol: "native",
3767
3952
  supportsPromptCache: true,
3953
+ includedTools: ["write_file", "edit_file"],
3768
3954
  inputPrice: 0.15,
3769
3955
  outputPrice: 0.6
3770
3956
  },
@@ -3773,7 +3959,9 @@ var vertexModels = {
3773
3959
  contextWindow: 1048576,
3774
3960
  supportsImages: true,
3775
3961
  supportsNativeTools: true,
3962
+ defaultToolProtocol: "native",
3776
3963
  supportsPromptCache: true,
3964
+ includedTools: ["write_file", "edit_file"],
3777
3965
  inputPrice: 0.3,
3778
3966
  outputPrice: 2.5,
3779
3967
  cacheReadsPrice: 0.075,
@@ -3786,7 +3974,9 @@ var vertexModels = {
3786
3974
  contextWindow: 1048576,
3787
3975
  supportsImages: true,
3788
3976
  supportsNativeTools: true,
3977
+ defaultToolProtocol: "native",
3789
3978
  supportsPromptCache: false,
3979
+ includedTools: ["write_file", "edit_file"],
3790
3980
  inputPrice: 0.15,
3791
3981
  outputPrice: 3.5,
3792
3982
  maxThinkingTokens: 24576,
@@ -3798,7 +3988,9 @@ var vertexModels = {
3798
3988
  contextWindow: 1048576,
3799
3989
  supportsImages: true,
3800
3990
  supportsNativeTools: true,
3991
+ defaultToolProtocol: "native",
3801
3992
  supportsPromptCache: false,
3993
+ includedTools: ["write_file", "edit_file"],
3802
3994
  inputPrice: 0.15,
3803
3995
  outputPrice: 0.6
3804
3996
  },
@@ -3807,7 +3999,9 @@ var vertexModels = {
3807
3999
  contextWindow: 1048576,
3808
4000
  supportsImages: true,
3809
4001
  supportsNativeTools: true,
4002
+ defaultToolProtocol: "native",
3810
4003
  supportsPromptCache: true,
4004
+ includedTools: ["write_file", "edit_file"],
3811
4005
  inputPrice: 2.5,
3812
4006
  outputPrice: 15
3813
4007
  },
@@ -3816,7 +4010,9 @@ var vertexModels = {
3816
4010
  contextWindow: 1048576,
3817
4011
  supportsImages: true,
3818
4012
  supportsNativeTools: true,
4013
+ defaultToolProtocol: "native",
3819
4014
  supportsPromptCache: true,
4015
+ includedTools: ["write_file", "edit_file"],
3820
4016
  inputPrice: 2.5,
3821
4017
  outputPrice: 15
3822
4018
  },
@@ -3825,7 +4021,9 @@ var vertexModels = {
3825
4021
  contextWindow: 1048576,
3826
4022
  supportsImages: true,
3827
4023
  supportsNativeTools: true,
4024
+ defaultToolProtocol: "native",
3828
4025
  supportsPromptCache: true,
4026
+ includedTools: ["write_file", "edit_file"],
3829
4027
  inputPrice: 2.5,
3830
4028
  outputPrice: 15,
3831
4029
  maxThinkingTokens: 32768,
@@ -3836,7 +4034,9 @@ var vertexModels = {
3836
4034
  contextWindow: 1048576,
3837
4035
  supportsImages: true,
3838
4036
  supportsNativeTools: true,
4037
+ defaultToolProtocol: "native",
3839
4038
  supportsPromptCache: true,
4039
+ includedTools: ["write_file", "edit_file"],
3840
4040
  inputPrice: 2.5,
3841
4041
  outputPrice: 15,
3842
4042
  maxThinkingTokens: 32768,
@@ -3862,7 +4062,9 @@ var vertexModels = {
3862
4062
  contextWindow: 1048576,
3863
4063
  supportsImages: true,
3864
4064
  supportsNativeTools: true,
4065
+ defaultToolProtocol: "native",
3865
4066
  supportsPromptCache: false,
4067
+ includedTools: ["write_file", "edit_file"],
3866
4068
  inputPrice: 0,
3867
4069
  outputPrice: 0
3868
4070
  },
@@ -3871,7 +4073,9 @@ var vertexModels = {
3871
4073
  contextWindow: 2097152,
3872
4074
  supportsImages: true,
3873
4075
  supportsNativeTools: true,
4076
+ defaultToolProtocol: "native",
3874
4077
  supportsPromptCache: false,
4078
+ includedTools: ["write_file", "edit_file"],
3875
4079
  inputPrice: 0,
3876
4080
  outputPrice: 0
3877
4081
  },
@@ -3880,7 +4084,9 @@ var vertexModels = {
3880
4084
  contextWindow: 1048576,
3881
4085
  supportsImages: true,
3882
4086
  supportsNativeTools: true,
4087
+ defaultToolProtocol: "native",
3883
4088
  supportsPromptCache: true,
4089
+ includedTools: ["write_file", "edit_file"],
3884
4090
  inputPrice: 0.15,
3885
4091
  outputPrice: 0.6
3886
4092
  },
@@ -3889,7 +4095,9 @@ var vertexModels = {
3889
4095
  contextWindow: 1048576,
3890
4096
  supportsImages: true,
3891
4097
  supportsNativeTools: true,
4098
+ defaultToolProtocol: "native",
3892
4099
  supportsPromptCache: false,
4100
+ includedTools: ["write_file", "edit_file"],
3893
4101
  inputPrice: 0.075,
3894
4102
  outputPrice: 0.3
3895
4103
  },
@@ -3898,7 +4106,9 @@ var vertexModels = {
3898
4106
  contextWindow: 32768,
3899
4107
  supportsImages: true,
3900
4108
  supportsNativeTools: true,
4109
+ defaultToolProtocol: "native",
3901
4110
  supportsPromptCache: false,
4111
+ includedTools: ["write_file", "edit_file"],
3902
4112
  inputPrice: 0,
3903
4113
  outputPrice: 0
3904
4114
  },
@@ -3907,7 +4117,9 @@ var vertexModels = {
3907
4117
  contextWindow: 1048576,
3908
4118
  supportsImages: true,
3909
4119
  supportsNativeTools: true,
4120
+ defaultToolProtocol: "native",
3910
4121
  supportsPromptCache: true,
4122
+ includedTools: ["write_file", "edit_file"],
3911
4123
  inputPrice: 0.075,
3912
4124
  outputPrice: 0.3
3913
4125
  },
@@ -3916,7 +4128,9 @@ var vertexModels = {
3916
4128
  contextWindow: 2097152,
3917
4129
  supportsImages: true,
3918
4130
  supportsNativeTools: true,
4131
+ defaultToolProtocol: "native",
3919
4132
  supportsPromptCache: false,
4133
+ includedTools: ["write_file", "edit_file"],
3920
4134
  inputPrice: 1.25,
3921
4135
  outputPrice: 5
3922
4136
  },
@@ -4062,7 +4276,9 @@ var vertexModels = {
4062
4276
  contextWindow: 1048576,
4063
4277
  supportsImages: true,
4064
4278
  supportsNativeTools: true,
4279
+ defaultToolProtocol: "native",
4065
4280
  supportsPromptCache: true,
4281
+ includedTools: ["write_file", "edit_file"],
4066
4282
  inputPrice: 0.1,
4067
4283
  outputPrice: 0.4,
4068
4284
  cacheReadsPrice: 0.025,
@@ -4364,6 +4580,7 @@ var xaiModels = {
4364
4580
  supportsImages: true,
4365
4581
  supportsPromptCache: true,
4366
4582
  supportsNativeTools: true,
4583
+ defaultToolProtocol: "native",
4367
4584
  inputPrice: 0.2,
4368
4585
  outputPrice: 1.5,
4369
4586
  cacheWritesPrice: 0.02,
@@ -4378,6 +4595,7 @@ var xaiModels = {
4378
4595
  supportsImages: true,
4379
4596
  supportsPromptCache: true,
4380
4597
  supportsNativeTools: true,
4598
+ defaultToolProtocol: "native",
4381
4599
  inputPrice: 0.2,
4382
4600
  outputPrice: 0.5,
4383
4601
  cacheWritesPrice: 0.05,
@@ -4392,6 +4610,7 @@ var xaiModels = {
4392
4610
  supportsImages: true,
4393
4611
  supportsPromptCache: true,
4394
4612
  supportsNativeTools: true,
4613
+ defaultToolProtocol: "native",
4395
4614
  inputPrice: 0.2,
4396
4615
  outputPrice: 0.5,
4397
4616
  cacheWritesPrice: 0.05,
@@ -4406,6 +4625,7 @@ var xaiModels = {
4406
4625
  supportsImages: true,
4407
4626
  supportsPromptCache: true,
4408
4627
  supportsNativeTools: true,
4628
+ defaultToolProtocol: "native",
4409
4629
  inputPrice: 0.2,
4410
4630
  outputPrice: 0.5,
4411
4631
  cacheWritesPrice: 0.05,
@@ -4420,6 +4640,7 @@ var xaiModels = {
4420
4640
  supportsImages: true,
4421
4641
  supportsPromptCache: true,
4422
4642
  supportsNativeTools: true,
4643
+ defaultToolProtocol: "native",
4423
4644
  inputPrice: 0.2,
4424
4645
  outputPrice: 0.5,
4425
4646
  cacheWritesPrice: 0.05,
@@ -4434,6 +4655,7 @@ var xaiModels = {
4434
4655
  supportsImages: true,
4435
4656
  supportsPromptCache: true,
4436
4657
  supportsNativeTools: true,
4658
+ defaultToolProtocol: "native",
4437
4659
  inputPrice: 3,
4438
4660
  outputPrice: 15,
4439
4661
  cacheWritesPrice: 0.75,
@@ -4448,6 +4670,7 @@ var xaiModels = {
4448
4670
  supportsImages: true,
4449
4671
  supportsPromptCache: true,
4450
4672
  supportsNativeTools: true,
4673
+ defaultToolProtocol: "native",
4451
4674
  inputPrice: 0.3,
4452
4675
  outputPrice: 0.5,
4453
4676
  cacheWritesPrice: 0.07,
@@ -4464,6 +4687,7 @@ var xaiModels = {
4464
4687
  supportsImages: true,
4465
4688
  supportsPromptCache: true,
4466
4689
  supportsNativeTools: true,
4690
+ defaultToolProtocol: "native",
4467
4691
  inputPrice: 3,
4468
4692
  outputPrice: 15,
4469
4693
  cacheWritesPrice: 0.75,
@@ -4577,6 +4801,7 @@ var internationalZAiModels = {
4577
4801
  supportsImages: false,
4578
4802
  supportsPromptCache: true,
4579
4803
  supportsNativeTools: true,
4804
+ defaultToolProtocol: "native",
4580
4805
  inputPrice: 0.6,
4581
4806
  outputPrice: 2.2,
4582
4807
  cacheWritesPrice: 0,
@@ -4589,6 +4814,7 @@ var internationalZAiModels = {
4589
4814
  supportsImages: false,
4590
4815
  supportsPromptCache: true,
4591
4816
  supportsNativeTools: true,
4817
+ defaultToolProtocol: "native",
4592
4818
  inputPrice: 0.2,
4593
4819
  outputPrice: 1.1,
4594
4820
  cacheWritesPrice: 0,
@@ -4601,6 +4827,7 @@ var internationalZAiModels = {
4601
4827
  supportsImages: false,
4602
4828
  supportsPromptCache: true,
4603
4829
  supportsNativeTools: true,
4830
+ defaultToolProtocol: "native",
4604
4831
  inputPrice: 2.2,
4605
4832
  outputPrice: 8.9,
4606
4833
  cacheWritesPrice: 0,
@@ -4613,6 +4840,7 @@ var internationalZAiModels = {
4613
4840
  supportsImages: false,
4614
4841
  supportsPromptCache: true,
4615
4842
  supportsNativeTools: true,
4843
+ defaultToolProtocol: "native",
4616
4844
  inputPrice: 1.1,
4617
4845
  outputPrice: 4.5,
4618
4846
  cacheWritesPrice: 0,
@@ -4625,6 +4853,7 @@ var internationalZAiModels = {
4625
4853
  supportsImages: false,
4626
4854
  supportsPromptCache: true,
4627
4855
  supportsNativeTools: true,
4856
+ defaultToolProtocol: "native",
4628
4857
  inputPrice: 0,
4629
4858
  outputPrice: 0,
4630
4859
  cacheWritesPrice: 0,
@@ -4637,6 +4866,7 @@ var internationalZAiModels = {
4637
4866
  supportsImages: true,
4638
4867
  supportsPromptCache: true,
4639
4868
  supportsNativeTools: true,
4869
+ defaultToolProtocol: "native",
4640
4870
  inputPrice: 0.6,
4641
4871
  outputPrice: 1.8,
4642
4872
  cacheWritesPrice: 0,
@@ -4649,6 +4879,7 @@ var internationalZAiModels = {
4649
4879
  supportsImages: false,
4650
4880
  supportsPromptCache: true,
4651
4881
  supportsNativeTools: true,
4882
+ defaultToolProtocol: "native",
4652
4883
  inputPrice: 0.6,
4653
4884
  outputPrice: 2.2,
4654
4885
  cacheWritesPrice: 0,
@@ -4661,6 +4892,7 @@ var internationalZAiModels = {
4661
4892
  supportsImages: false,
4662
4893
  supportsPromptCache: false,
4663
4894
  supportsNativeTools: true,
4895
+ defaultToolProtocol: "native",
4664
4896
  inputPrice: 0.1,
4665
4897
  outputPrice: 0.1,
4666
4898
  cacheWritesPrice: 0,
@@ -4676,6 +4908,7 @@ var mainlandZAiModels = {
4676
4908
  supportsImages: false,
4677
4909
  supportsPromptCache: true,
4678
4910
  supportsNativeTools: true,
4911
+ defaultToolProtocol: "native",
4679
4912
  inputPrice: 0.29,
4680
4913
  outputPrice: 1.14,
4681
4914
  cacheWritesPrice: 0,
@@ -4688,6 +4921,7 @@ var mainlandZAiModels = {
4688
4921
  supportsImages: false,
4689
4922
  supportsPromptCache: true,
4690
4923
  supportsNativeTools: true,
4924
+ defaultToolProtocol: "native",
4691
4925
  inputPrice: 0.1,
4692
4926
  outputPrice: 0.6,
4693
4927
  cacheWritesPrice: 0,
@@ -4700,6 +4934,7 @@ var mainlandZAiModels = {
4700
4934
  supportsImages: false,
4701
4935
  supportsPromptCache: true,
4702
4936
  supportsNativeTools: true,
4937
+ defaultToolProtocol: "native",
4703
4938
  inputPrice: 0.29,
4704
4939
  outputPrice: 1.14,
4705
4940
  cacheWritesPrice: 0,
@@ -4712,6 +4947,7 @@ var mainlandZAiModels = {
4712
4947
  supportsImages: false,
4713
4948
  supportsPromptCache: true,
4714
4949
  supportsNativeTools: true,
4950
+ defaultToolProtocol: "native",
4715
4951
  inputPrice: 0.1,
4716
4952
  outputPrice: 0.6,
4717
4953
  cacheWritesPrice: 0,
@@ -4724,6 +4960,7 @@ var mainlandZAiModels = {
4724
4960
  supportsImages: false,
4725
4961
  supportsPromptCache: true,
4726
4962
  supportsNativeTools: true,
4963
+ defaultToolProtocol: "native",
4727
4964
  inputPrice: 0,
4728
4965
  outputPrice: 0,
4729
4966
  cacheWritesPrice: 0,
@@ -4736,6 +4973,7 @@ var mainlandZAiModels = {
4736
4973
  supportsImages: true,
4737
4974
  supportsPromptCache: true,
4738
4975
  supportsNativeTools: true,
4976
+ defaultToolProtocol: "native",
4739
4977
  inputPrice: 0.29,
4740
4978
  outputPrice: 0.93,
4741
4979
  cacheWritesPrice: 0,
@@ -4748,6 +4986,7 @@ var mainlandZAiModels = {
4748
4986
  supportsImages: false,
4749
4987
  supportsPromptCache: true,
4750
4988
  supportsNativeTools: true,
4989
+ defaultToolProtocol: "native",
4751
4990
  inputPrice: 0.29,
4752
4991
  outputPrice: 1.14,
4753
4992
  cacheWritesPrice: 0,
@@ -4801,6 +5040,7 @@ var minimaxModels = {
4801
5040
  supportsImages: false,
4802
5041
  supportsPromptCache: true,
4803
5042
  supportsNativeTools: true,
5043
+ defaultToolProtocol: "native",
4804
5044
  preserveReasoning: true,
4805
5045
  inputPrice: 0.3,
4806
5046
  outputPrice: 1.2,
@@ -4814,6 +5054,7 @@ var minimaxModels = {
4814
5054
  supportsImages: false,
4815
5055
  supportsPromptCache: true,
4816
5056
  supportsNativeTools: true,
5057
+ defaultToolProtocol: "native",
4817
5058
  preserveReasoning: true,
4818
5059
  inputPrice: 0.3,
4819
5060
  outputPrice: 1.2,
@@ -4996,10 +5237,7 @@ var anthropicSchema = apiModelIdProviderModelSchema.extend({
4996
5237
  anthropicBeta1MContext: z8.boolean().optional()
4997
5238
  // Enable 'context-1m-2025-08-07' beta for 1M context window.
4998
5239
  });
4999
- var claudeCodeSchema = apiModelIdProviderModelSchema.extend({
5000
- claudeCodePath: z8.string().optional(),
5001
- claudeCodeMaxOutputTokens: z8.number().int().min(1).max(2e5).optional()
5002
- });
5240
+ var claudeCodeSchema = apiModelIdProviderModelSchema.extend({});
5003
5241
  var openRouterSchema = baseProviderSettingsSchema.extend({
5004
5242
  openRouterApiKey: z8.string().optional(),
5005
5243
  openRouterModelId: z8.string().optional(),
@@ -5674,11 +5912,44 @@ function getErrorStatusCode(error) {
5674
5912
  }
5675
5913
  return void 0;
5676
5914
  }
5915
+ function extractMessageFromJsonPayload(message) {
5916
+ const jsonStartIndex = message.indexOf("{");
5917
+ if (jsonStartIndex === -1) {
5918
+ return void 0;
5919
+ }
5920
+ const potentialJson = message.slice(jsonStartIndex);
5921
+ try {
5922
+ const parsed = JSON.parse(potentialJson);
5923
+ if (parsed?.error?.message && typeof parsed.error.message === "string") {
5924
+ return parsed.error.message;
5925
+ }
5926
+ if (parsed?.message && typeof parsed.message === "string") {
5927
+ return parsed.message;
5928
+ }
5929
+ } catch {
5930
+ }
5931
+ return void 0;
5932
+ }
5677
5933
  function getErrorMessage(error) {
5934
+ let message;
5678
5935
  if (isOpenAISdkError(error)) {
5679
- return error.error?.metadata?.raw || error.error?.message || error.message;
5936
+ message = error.error?.metadata?.raw || error.error?.message || error.message;
5937
+ } else if (error instanceof Error) {
5938
+ message = error.message;
5939
+ } else if (typeof error === "object" && error !== null && "message" in error) {
5940
+ const msgValue = error.message;
5941
+ if (typeof msgValue === "string") {
5942
+ message = msgValue;
5943
+ }
5680
5944
  }
5681
- return void 0;
5945
+ if (!message) {
5946
+ return void 0;
5947
+ }
5948
+ const extractedMessage = extractMessageFromJsonPayload(message);
5949
+ if (extractedMessage) {
5950
+ return extractedMessage;
5951
+ }
5952
+ return message;
5682
5953
  }
5683
5954
  function shouldReportApiErrorToTelemetry(errorCode, errorMessage) {
5684
5955
  if (errorCode !== void 0 && EXPECTED_API_ERROR_CODES.has(errorCode)) {
@@ -6210,7 +6481,7 @@ var organizationDefaultSettingsSchema = globalSettingsSchema.pick({
6210
6481
  terminalShellIntegrationTimeout: z16.number().int().nonnegative().optional()
6211
6482
  })
6212
6483
  );
6213
- var workspaceTaskVisibilitySchema = z16.enum(["all", "list-only", "full-lockdown"]);
6484
+ var workspaceTaskVisibilitySchema = z16.enum(["all", "list-only", "admins-and-creator", "creator", "full-lockdown"]);
6214
6485
  var organizationCloudSettingsSchema = z16.object({
6215
6486
  recordTaskMessages: z16.boolean().optional(),
6216
6487
  enableTaskSharing: z16.boolean().optional(),
@@ -6792,7 +7063,6 @@ export {
6792
7063
  BEDROCK_REGIONS,
6793
7064
  BEDROCK_SERVICE_TIER_MODEL_IDS,
6794
7065
  BEDROCK_SERVICE_TIER_PRICING,
6795
- CLAUDE_CODE_DEFAULT_MAX_OUTPUT_TOKENS,
6796
7066
  CODEBASE_INDEX_DEFAULTS,
6797
7067
  CONSENT_COOKIE_NAME,
6798
7068
  CONTEXT_MANAGEMENT_EVENTS,
@@ -6882,6 +7152,7 @@ export {
6882
7152
  chutesModels,
6883
7153
  claudeCodeDefaultModelId,
6884
7154
  claudeCodeModels,
7155
+ claudeCodeReasoningConfig,
6885
7156
  clineAskSchema,
6886
7157
  clineAsks,
6887
7158
  clineMessageSchema,
@@ -6896,7 +7167,6 @@ export {
6896
7167
  commandIds,
6897
7168
  contextCondenseSchema,
6898
7169
  contextTruncationSchema,
6899
- convertModelNameForVertex,
6900
7170
  customModePromptsSchema,
6901
7171
  customModesSettingsSchema,
6902
7172
  customProviders,
@@ -6918,6 +7188,7 @@ export {
6918
7188
  extensionBridgeEventSchema,
6919
7189
  extensionInstanceSchema,
6920
7190
  extractApiProviderErrorProperties,
7191
+ extractMessageFromJsonPayload,
6921
7192
  fauxProviders,
6922
7193
  featherlessDefaultModelId,
6923
7194
  featherlessModels,
@@ -6927,7 +7198,6 @@ export {
6927
7198
  geminiDefaultModelId,
6928
7199
  geminiModels,
6929
7200
  getApiProtocol,
6930
- getClaudeCodeModelId,
6931
7201
  getEffectiveProtocol,
6932
7202
  getErrorMessage,
6933
7203
  getErrorStatusCode,
@@ -6999,6 +7269,7 @@ export {
6999
7269
  moonshotDefaultModelId,
7000
7270
  moonshotModels,
7001
7271
  nonBlockingAsks,
7272
+ normalizeClaudeCodeModelId,
7002
7273
  ollamaDefaultModelId,
7003
7274
  ollamaDefaultModelInfo,
7004
7275
  openAiModelInfoSaneDefaults,