@roo-code/types 1.92.0 → 1.94.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.js CHANGED
@@ -149,6 +149,7 @@ var toolNames = [
149
149
  "apply_diff",
150
150
  "search_and_replace",
151
151
  "search_replace",
152
+ "edit_file",
152
153
  "apply_patch",
153
154
  "search_files",
154
155
  "list_files",
@@ -177,6 +178,10 @@ var TOOL_PROTOCOL = {
177
178
  XML: "xml",
178
179
  NATIVE: "native"
179
180
  };
181
+ var NATIVE_TOOL_DEFAULTS = {
182
+ supportsNativeTools: true,
183
+ defaultToolProtocol: TOOL_PROTOCOL.NATIVE
184
+ };
180
185
  function isNativeProtocol(protocol) {
181
186
  return protocol === TOOL_PROTOCOL.NATIVE;
182
187
  }
@@ -571,6 +576,7 @@ var anthropicModels = {
571
576
  supportsImages: true,
572
577
  supportsPromptCache: true,
573
578
  supportsNativeTools: true,
579
+ defaultToolProtocol: "native",
574
580
  inputPrice: 3,
575
581
  // $3 per million input tokens (≤200K context)
576
582
  outputPrice: 15,
@@ -604,6 +610,7 @@ var anthropicModels = {
604
610
  supportsImages: true,
605
611
  supportsPromptCache: true,
606
612
  supportsNativeTools: true,
613
+ defaultToolProtocol: "native",
607
614
  inputPrice: 3,
608
615
  // $3 per million input tokens (≤200K context)
609
616
  outputPrice: 15,
@@ -636,6 +643,7 @@ var anthropicModels = {
636
643
  supportsImages: true,
637
644
  supportsPromptCache: true,
638
645
  supportsNativeTools: true,
646
+ defaultToolProtocol: "native",
639
647
  inputPrice: 5,
640
648
  // $5 per million input tokens
641
649
  outputPrice: 25,
@@ -653,6 +661,7 @@ var anthropicModels = {
653
661
  supportsImages: true,
654
662
  supportsPromptCache: true,
655
663
  supportsNativeTools: true,
664
+ defaultToolProtocol: "native",
656
665
  inputPrice: 15,
657
666
  // $15 per million input tokens
658
667
  outputPrice: 75,
@@ -670,6 +679,7 @@ var anthropicModels = {
670
679
  supportsImages: true,
671
680
  supportsPromptCache: true,
672
681
  supportsNativeTools: true,
682
+ defaultToolProtocol: "native",
673
683
  inputPrice: 15,
674
684
  // $15 per million input tokens
675
685
  outputPrice: 75,
@@ -687,6 +697,7 @@ var anthropicModels = {
687
697
  supportsImages: true,
688
698
  supportsPromptCache: true,
689
699
  supportsNativeTools: true,
700
+ defaultToolProtocol: "native",
690
701
  inputPrice: 3,
691
702
  // $3 per million input tokens
692
703
  outputPrice: 15,
@@ -705,6 +716,7 @@ var anthropicModels = {
705
716
  supportsImages: true,
706
717
  supportsPromptCache: true,
707
718
  supportsNativeTools: true,
719
+ defaultToolProtocol: "native",
708
720
  inputPrice: 3,
709
721
  // $3 per million input tokens
710
722
  outputPrice: 15,
@@ -720,6 +732,7 @@ var anthropicModels = {
720
732
  supportsImages: true,
721
733
  supportsPromptCache: true,
722
734
  supportsNativeTools: true,
735
+ defaultToolProtocol: "native",
723
736
  inputPrice: 3,
724
737
  // $3 per million input tokens
725
738
  outputPrice: 15,
@@ -735,6 +748,7 @@ var anthropicModels = {
735
748
  supportsImages: false,
736
749
  supportsPromptCache: true,
737
750
  supportsNativeTools: true,
751
+ defaultToolProtocol: "native",
738
752
  inputPrice: 1,
739
753
  outputPrice: 5,
740
754
  cacheWritesPrice: 1.25,
@@ -746,6 +760,7 @@ var anthropicModels = {
746
760
  supportsImages: true,
747
761
  supportsPromptCache: true,
748
762
  supportsNativeTools: true,
763
+ defaultToolProtocol: "native",
749
764
  inputPrice: 15,
750
765
  outputPrice: 75,
751
766
  cacheWritesPrice: 18.75,
@@ -757,6 +772,7 @@ var anthropicModels = {
757
772
  supportsImages: true,
758
773
  supportsPromptCache: true,
759
774
  supportsNativeTools: true,
775
+ defaultToolProtocol: "native",
760
776
  inputPrice: 0.25,
761
777
  outputPrice: 1.25,
762
778
  cacheWritesPrice: 0.3,
@@ -768,6 +784,7 @@ var anthropicModels = {
768
784
  supportsImages: true,
769
785
  supportsPromptCache: true,
770
786
  supportsNativeTools: true,
787
+ defaultToolProtocol: "native",
771
788
  inputPrice: 1,
772
789
  outputPrice: 5,
773
790
  cacheWritesPrice: 1.25,
@@ -920,6 +937,7 @@ var bedrockModels = {
920
937
  supportsPromptCache: true,
921
938
  supportsReasoningBudget: true,
922
939
  supportsNativeTools: true,
940
+ defaultToolProtocol: "native",
923
941
  inputPrice: 3,
924
942
  outputPrice: 15,
925
943
  cacheWritesPrice: 3.75,
@@ -1013,6 +1031,7 @@ var bedrockModels = {
1013
1031
  supportsPromptCache: true,
1014
1032
  supportsReasoningBudget: true,
1015
1033
  supportsNativeTools: true,
1034
+ defaultToolProtocol: "native",
1016
1035
  inputPrice: 3,
1017
1036
  outputPrice: 15,
1018
1037
  cacheWritesPrice: 3.75,
@@ -1028,6 +1047,7 @@ var bedrockModels = {
1028
1047
  supportsPromptCache: true,
1029
1048
  supportsReasoningBudget: true,
1030
1049
  supportsNativeTools: true,
1050
+ defaultToolProtocol: "native",
1031
1051
  inputPrice: 15,
1032
1052
  outputPrice: 75,
1033
1053
  cacheWritesPrice: 18.75,
@@ -1043,6 +1063,7 @@ var bedrockModels = {
1043
1063
  supportsPromptCache: true,
1044
1064
  supportsReasoningBudget: true,
1045
1065
  supportsNativeTools: true,
1066
+ defaultToolProtocol: "native",
1046
1067
  inputPrice: 5,
1047
1068
  outputPrice: 25,
1048
1069
  cacheWritesPrice: 6.25,
@@ -1058,6 +1079,7 @@ var bedrockModels = {
1058
1079
  supportsPromptCache: true,
1059
1080
  supportsReasoningBudget: true,
1060
1081
  supportsNativeTools: true,
1082
+ defaultToolProtocol: "native",
1061
1083
  inputPrice: 15,
1062
1084
  outputPrice: 75,
1063
1085
  cacheWritesPrice: 18.75,
@@ -1073,6 +1095,7 @@ var bedrockModels = {
1073
1095
  supportsPromptCache: true,
1074
1096
  supportsReasoningBudget: true,
1075
1097
  supportsNativeTools: true,
1098
+ defaultToolProtocol: "native",
1076
1099
  inputPrice: 3,
1077
1100
  outputPrice: 15,
1078
1101
  cacheWritesPrice: 3.75,
@@ -1087,6 +1110,7 @@ var bedrockModels = {
1087
1110
  supportsImages: true,
1088
1111
  supportsPromptCache: true,
1089
1112
  supportsNativeTools: true,
1113
+ defaultToolProtocol: "native",
1090
1114
  inputPrice: 3,
1091
1115
  outputPrice: 15,
1092
1116
  cacheWritesPrice: 3.75,
@@ -1101,6 +1125,7 @@ var bedrockModels = {
1101
1125
  supportsImages: false,
1102
1126
  supportsPromptCache: true,
1103
1127
  supportsNativeTools: true,
1128
+ defaultToolProtocol: "native",
1104
1129
  inputPrice: 0.8,
1105
1130
  outputPrice: 4,
1106
1131
  cacheWritesPrice: 1,
@@ -1116,6 +1141,7 @@ var bedrockModels = {
1116
1141
  supportsPromptCache: true,
1117
1142
  supportsReasoningBudget: true,
1118
1143
  supportsNativeTools: true,
1144
+ defaultToolProtocol: "native",
1119
1145
  inputPrice: 1,
1120
1146
  outputPrice: 5,
1121
1147
  cacheWritesPrice: 1.25,
@@ -1132,6 +1158,7 @@ var bedrockModels = {
1132
1158
  supportsImages: true,
1133
1159
  supportsPromptCache: false,
1134
1160
  supportsNativeTools: true,
1161
+ defaultToolProtocol: "native",
1135
1162
  inputPrice: 3,
1136
1163
  outputPrice: 15
1137
1164
  },
@@ -1141,6 +1168,7 @@ var bedrockModels = {
1141
1168
  supportsImages: true,
1142
1169
  supportsPromptCache: false,
1143
1170
  supportsNativeTools: true,
1171
+ defaultToolProtocol: "native",
1144
1172
  inputPrice: 15,
1145
1173
  outputPrice: 75
1146
1174
  },
@@ -1150,6 +1178,7 @@ var bedrockModels = {
1150
1178
  supportsImages: true,
1151
1179
  supportsPromptCache: false,
1152
1180
  supportsNativeTools: true,
1181
+ defaultToolProtocol: "native",
1153
1182
  inputPrice: 3,
1154
1183
  outputPrice: 15
1155
1184
  },
@@ -1159,6 +1188,7 @@ var bedrockModels = {
1159
1188
  supportsImages: true,
1160
1189
  supportsPromptCache: false,
1161
1190
  supportsNativeTools: true,
1191
+ defaultToolProtocol: "native",
1162
1192
  inputPrice: 0.25,
1163
1193
  outputPrice: 1.25
1164
1194
  },
@@ -1168,6 +1198,7 @@ var bedrockModels = {
1168
1198
  supportsImages: false,
1169
1199
  supportsPromptCache: false,
1170
1200
  supportsNativeTools: true,
1201
+ defaultToolProtocol: "native",
1171
1202
  inputPrice: 8,
1172
1203
  outputPrice: 24,
1173
1204
  description: "Claude 2.1"
@@ -1178,6 +1209,7 @@ var bedrockModels = {
1178
1209
  supportsImages: false,
1179
1210
  supportsPromptCache: false,
1180
1211
  supportsNativeTools: true,
1212
+ defaultToolProtocol: "native",
1181
1213
  inputPrice: 8,
1182
1214
  outputPrice: 24,
1183
1215
  description: "Claude 2.0"
@@ -1188,6 +1220,7 @@ var bedrockModels = {
1188
1220
  supportsImages: false,
1189
1221
  supportsPromptCache: false,
1190
1222
  supportsNativeTools: true,
1223
+ defaultToolProtocol: "native",
1191
1224
  inputPrice: 0.8,
1192
1225
  outputPrice: 2.4,
1193
1226
  description: "Claude Instant"
@@ -1504,6 +1537,7 @@ var cerebrasModels = {
1504
1537
  supportsImages: false,
1505
1538
  supportsPromptCache: false,
1506
1539
  supportsNativeTools: true,
1540
+ defaultToolProtocol: "native",
1507
1541
  inputPrice: 0,
1508
1542
  outputPrice: 0,
1509
1543
  description: "Highly intelligent general purpose model with up to 1,000 tokens/s"
@@ -1515,6 +1549,7 @@ var cerebrasModels = {
1515
1549
  supportsImages: false,
1516
1550
  supportsPromptCache: false,
1517
1551
  supportsNativeTools: true,
1552
+ defaultToolProtocol: "native",
1518
1553
  inputPrice: 0,
1519
1554
  outputPrice: 0,
1520
1555
  description: "Intelligent model with ~1400 tokens/s"
@@ -1526,6 +1561,7 @@ var cerebrasModels = {
1526
1561
  supportsImages: false,
1527
1562
  supportsPromptCache: false,
1528
1563
  supportsNativeTools: true,
1564
+ defaultToolProtocol: "native",
1529
1565
  inputPrice: 0,
1530
1566
  outputPrice: 0,
1531
1567
  description: "Powerful model with ~2600 tokens/s"
@@ -1537,6 +1573,7 @@ var cerebrasModels = {
1537
1573
  supportsImages: false,
1538
1574
  supportsPromptCache: false,
1539
1575
  supportsNativeTools: true,
1576
+ defaultToolProtocol: "native",
1540
1577
  inputPrice: 0,
1541
1578
  outputPrice: 0,
1542
1579
  description: "SOTA coding performance with ~2500 tokens/s"
@@ -1548,6 +1585,7 @@ var cerebrasModels = {
1548
1585
  supportsImages: false,
1549
1586
  supportsPromptCache: false,
1550
1587
  supportsNativeTools: true,
1588
+ defaultToolProtocol: "native",
1551
1589
  inputPrice: 0,
1552
1590
  outputPrice: 0,
1553
1591
  description: "OpenAI GPT OSS model with ~2800 tokens/s\n\n\u2022 64K context window\n\u2022 Excels at efficient reasoning across science, math, and coding"
@@ -1562,6 +1600,8 @@ var chutesModels = {
1562
1600
  contextWindow: 163840,
1563
1601
  supportsImages: false,
1564
1602
  supportsPromptCache: false,
1603
+ supportsNativeTools: true,
1604
+ defaultToolProtocol: "native",
1565
1605
  inputPrice: 0,
1566
1606
  outputPrice: 0,
1567
1607
  description: "DeepSeek R1 0528 model."
@@ -1571,6 +1611,8 @@ var chutesModels = {
1571
1611
  contextWindow: 163840,
1572
1612
  supportsImages: false,
1573
1613
  supportsPromptCache: false,
1614
+ supportsNativeTools: true,
1615
+ defaultToolProtocol: "native",
1574
1616
  inputPrice: 0,
1575
1617
  outputPrice: 0,
1576
1618
  description: "DeepSeek R1 model."
@@ -1580,6 +1622,8 @@ var chutesModels = {
1580
1622
  contextWindow: 163840,
1581
1623
  supportsImages: false,
1582
1624
  supportsPromptCache: false,
1625
+ supportsNativeTools: true,
1626
+ defaultToolProtocol: "native",
1583
1627
  inputPrice: 0,
1584
1628
  outputPrice: 0,
1585
1629
  description: "DeepSeek V3 model."
@@ -1589,6 +1633,8 @@ var chutesModels = {
1589
1633
  contextWindow: 163840,
1590
1634
  supportsImages: false,
1591
1635
  supportsPromptCache: false,
1636
+ supportsNativeTools: true,
1637
+ defaultToolProtocol: "native",
1592
1638
  inputPrice: 0,
1593
1639
  outputPrice: 0,
1594
1640
  description: "DeepSeek V3.1 model."
@@ -1598,6 +1644,8 @@ var chutesModels = {
1598
1644
  contextWindow: 163840,
1599
1645
  supportsImages: false,
1600
1646
  supportsPromptCache: false,
1647
+ supportsNativeTools: true,
1648
+ defaultToolProtocol: "native",
1601
1649
  inputPrice: 0.23,
1602
1650
  outputPrice: 0.9,
1603
1651
  description: "DeepSeek\u2011V3.1\u2011Terminus is an update to V3.1 that improves language consistency by reducing CN/EN mix\u2011ups and eliminating random characters, while strengthening agent capabilities with notably better Code Agent and Search Agent performance."
@@ -1607,6 +1655,8 @@ var chutesModels = {
1607
1655
  contextWindow: 163840,
1608
1656
  supportsImages: false,
1609
1657
  supportsPromptCache: false,
1658
+ supportsNativeTools: true,
1659
+ defaultToolProtocol: "native",
1610
1660
  inputPrice: 1,
1611
1661
  outputPrice: 3,
1612
1662
  description: "DeepSeek-V3.1-turbo is an FP8, speculative-decoding turbo variant optimized for ultra-fast single-shot queries (~200 TPS), with outputs close to the originals and solid function calling/reasoning/structured output, priced at $1/M input and $3/M output tokens, using 2\xD7 quota per request and not intended for bulk workloads."
@@ -1616,6 +1666,8 @@ var chutesModels = {
1616
1666
  contextWindow: 163840,
1617
1667
  supportsImages: false,
1618
1668
  supportsPromptCache: false,
1669
+ supportsNativeTools: true,
1670
+ defaultToolProtocol: "native",
1619
1671
  inputPrice: 0.25,
1620
1672
  outputPrice: 0.35,
1621
1673
  description: "DeepSeek-V3.2-Exp is an experimental LLM that introduces DeepSeek Sparse Attention to improve long\u2011context training and inference efficiency while maintaining performance comparable to V3.1\u2011Terminus."
@@ -1627,6 +1679,8 @@ var chutesModels = {
1627
1679
  // From Groq
1628
1680
  supportsImages: false,
1629
1681
  supportsPromptCache: false,
1682
+ supportsNativeTools: true,
1683
+ defaultToolProtocol: "native",
1630
1684
  inputPrice: 0,
1631
1685
  outputPrice: 0,
1632
1686
  description: "Unsloth Llama 3.3 70B Instruct model."
@@ -1636,6 +1690,8 @@ var chutesModels = {
1636
1690
  contextWindow: 512e3,
1637
1691
  supportsImages: false,
1638
1692
  supportsPromptCache: false,
1693
+ supportsNativeTools: true,
1694
+ defaultToolProtocol: "native",
1639
1695
  inputPrice: 0,
1640
1696
  outputPrice: 0,
1641
1697
  description: "ChutesAI Llama 4 Scout 17B Instruct model, 512K context."
@@ -1645,6 +1701,8 @@ var chutesModels = {
1645
1701
  contextWindow: 128e3,
1646
1702
  supportsImages: false,
1647
1703
  supportsPromptCache: false,
1704
+ supportsNativeTools: true,
1705
+ defaultToolProtocol: "native",
1648
1706
  inputPrice: 0,
1649
1707
  outputPrice: 0,
1650
1708
  description: "Unsloth Mistral Nemo Instruct model."
@@ -1654,6 +1712,8 @@ var chutesModels = {
1654
1712
  contextWindow: 131072,
1655
1713
  supportsImages: false,
1656
1714
  supportsPromptCache: false,
1715
+ supportsNativeTools: true,
1716
+ defaultToolProtocol: "native",
1657
1717
  inputPrice: 0,
1658
1718
  outputPrice: 0,
1659
1719
  description: "Unsloth Gemma 3 12B IT model."
@@ -1663,6 +1723,8 @@ var chutesModels = {
1663
1723
  contextWindow: 131072,
1664
1724
  supportsImages: false,
1665
1725
  supportsPromptCache: false,
1726
+ supportsNativeTools: true,
1727
+ defaultToolProtocol: "native",
1666
1728
  inputPrice: 0,
1667
1729
  outputPrice: 0,
1668
1730
  description: "Nous DeepHermes 3 Llama 3 8B Preview model."
@@ -1672,6 +1734,8 @@ var chutesModels = {
1672
1734
  contextWindow: 131072,
1673
1735
  supportsImages: false,
1674
1736
  supportsPromptCache: false,
1737
+ supportsNativeTools: true,
1738
+ defaultToolProtocol: "native",
1675
1739
  inputPrice: 0,
1676
1740
  outputPrice: 0,
1677
1741
  description: "Unsloth Gemma 3 4B IT model."
@@ -1681,6 +1745,8 @@ var chutesModels = {
1681
1745
  contextWindow: 131072,
1682
1746
  supportsImages: false,
1683
1747
  supportsPromptCache: false,
1748
+ supportsNativeTools: true,
1749
+ defaultToolProtocol: "native",
1684
1750
  inputPrice: 0,
1685
1751
  outputPrice: 0,
1686
1752
  description: "Nvidia Llama 3.3 Nemotron Super 49B model."
@@ -1690,6 +1756,8 @@ var chutesModels = {
1690
1756
  contextWindow: 131072,
1691
1757
  supportsImages: false,
1692
1758
  supportsPromptCache: false,
1759
+ supportsNativeTools: true,
1760
+ defaultToolProtocol: "native",
1693
1761
  inputPrice: 0,
1694
1762
  outputPrice: 0,
1695
1763
  description: "Nvidia Llama 3.1 Nemotron Ultra 253B model."
@@ -1699,6 +1767,8 @@ var chutesModels = {
1699
1767
  contextWindow: 256e3,
1700
1768
  supportsImages: false,
1701
1769
  supportsPromptCache: false,
1770
+ supportsNativeTools: true,
1771
+ defaultToolProtocol: "native",
1702
1772
  inputPrice: 0,
1703
1773
  outputPrice: 0,
1704
1774
  description: "ChutesAI Llama 4 Maverick 17B Instruct FP8 model."
@@ -1708,6 +1778,8 @@ var chutesModels = {
1708
1778
  contextWindow: 163840,
1709
1779
  supportsImages: false,
1710
1780
  supportsPromptCache: false,
1781
+ supportsNativeTools: true,
1782
+ defaultToolProtocol: "native",
1711
1783
  inputPrice: 0,
1712
1784
  outputPrice: 0,
1713
1785
  description: "DeepSeek V3 Base model."
@@ -1717,6 +1789,8 @@ var chutesModels = {
1717
1789
  contextWindow: 163840,
1718
1790
  supportsImages: false,
1719
1791
  supportsPromptCache: false,
1792
+ supportsNativeTools: true,
1793
+ defaultToolProtocol: "native",
1720
1794
  inputPrice: 0,
1721
1795
  outputPrice: 0,
1722
1796
  description: "DeepSeek R1 Zero model."
@@ -1726,6 +1800,8 @@ var chutesModels = {
1726
1800
  contextWindow: 163840,
1727
1801
  supportsImages: false,
1728
1802
  supportsPromptCache: false,
1803
+ supportsNativeTools: true,
1804
+ defaultToolProtocol: "native",
1729
1805
  inputPrice: 0,
1730
1806
  outputPrice: 0,
1731
1807
  description: "DeepSeek V3 (0324) model."
@@ -1735,6 +1811,8 @@ var chutesModels = {
1735
1811
  contextWindow: 262144,
1736
1812
  supportsImages: false,
1737
1813
  supportsPromptCache: false,
1814
+ supportsNativeTools: true,
1815
+ defaultToolProtocol: "native",
1738
1816
  inputPrice: 0,
1739
1817
  outputPrice: 0,
1740
1818
  description: "Qwen3 235B A22B Instruct 2507 model with 262K context window."
@@ -1744,6 +1822,8 @@ var chutesModels = {
1744
1822
  contextWindow: 40960,
1745
1823
  supportsImages: false,
1746
1824
  supportsPromptCache: false,
1825
+ supportsNativeTools: true,
1826
+ defaultToolProtocol: "native",
1747
1827
  inputPrice: 0,
1748
1828
  outputPrice: 0,
1749
1829
  description: "Qwen3 235B A22B model."
@@ -1753,6 +1833,8 @@ var chutesModels = {
1753
1833
  contextWindow: 40960,
1754
1834
  supportsImages: false,
1755
1835
  supportsPromptCache: false,
1836
+ supportsNativeTools: true,
1837
+ defaultToolProtocol: "native",
1756
1838
  inputPrice: 0,
1757
1839
  outputPrice: 0,
1758
1840
  description: "Qwen3 32B model."
@@ -1762,6 +1844,8 @@ var chutesModels = {
1762
1844
  contextWindow: 40960,
1763
1845
  supportsImages: false,
1764
1846
  supportsPromptCache: false,
1847
+ supportsNativeTools: true,
1848
+ defaultToolProtocol: "native",
1765
1849
  inputPrice: 0,
1766
1850
  outputPrice: 0,
1767
1851
  description: "Qwen3 30B A3B model."
@@ -1771,6 +1855,8 @@ var chutesModels = {
1771
1855
  contextWindow: 40960,
1772
1856
  supportsImages: false,
1773
1857
  supportsPromptCache: false,
1858
+ supportsNativeTools: true,
1859
+ defaultToolProtocol: "native",
1774
1860
  inputPrice: 0,
1775
1861
  outputPrice: 0,
1776
1862
  description: "Qwen3 14B model."
@@ -1780,6 +1866,8 @@ var chutesModels = {
1780
1866
  contextWindow: 40960,
1781
1867
  supportsImages: false,
1782
1868
  supportsPromptCache: false,
1869
+ supportsNativeTools: true,
1870
+ defaultToolProtocol: "native",
1783
1871
  inputPrice: 0,
1784
1872
  outputPrice: 0,
1785
1873
  description: "Qwen3 8B model."
@@ -1789,6 +1877,8 @@ var chutesModels = {
1789
1877
  contextWindow: 163840,
1790
1878
  supportsImages: false,
1791
1879
  supportsPromptCache: false,
1880
+ supportsNativeTools: true,
1881
+ defaultToolProtocol: "native",
1792
1882
  inputPrice: 0,
1793
1883
  outputPrice: 0,
1794
1884
  description: "Microsoft MAI-DS-R1 FP8 model."
@@ -1798,6 +1888,8 @@ var chutesModels = {
1798
1888
  contextWindow: 163840,
1799
1889
  supportsImages: false,
1800
1890
  supportsPromptCache: false,
1891
+ supportsNativeTools: true,
1892
+ defaultToolProtocol: "native",
1801
1893
  inputPrice: 0,
1802
1894
  outputPrice: 0,
1803
1895
  description: "TNGTech DeepSeek R1T Chimera model."
@@ -1807,6 +1899,8 @@ var chutesModels = {
1807
1899
  contextWindow: 151329,
1808
1900
  supportsImages: false,
1809
1901
  supportsPromptCache: false,
1902
+ supportsNativeTools: true,
1903
+ defaultToolProtocol: "native",
1810
1904
  inputPrice: 0,
1811
1905
  outputPrice: 0,
1812
1906
  description: "GLM-4.5-Air model with 151,329 token context window and 106B total parameters with 12B activated."
@@ -1816,6 +1910,8 @@ var chutesModels = {
1816
1910
  contextWindow: 131072,
1817
1911
  supportsImages: false,
1818
1912
  supportsPromptCache: false,
1913
+ supportsNativeTools: true,
1914
+ defaultToolProtocol: "native",
1819
1915
  inputPrice: 0,
1820
1916
  outputPrice: 0,
1821
1917
  description: "GLM-4.5-FP8 model with 128k token context window, optimized for agent-based applications with MoE architecture."
@@ -1825,6 +1921,8 @@ var chutesModels = {
1825
1921
  contextWindow: 131072,
1826
1922
  supportsImages: false,
1827
1923
  supportsPromptCache: false,
1924
+ supportsNativeTools: true,
1925
+ defaultToolProtocol: "native",
1828
1926
  inputPrice: 1,
1829
1927
  outputPrice: 3,
1830
1928
  description: "GLM-4.5-turbo model with 128K token context window, optimized for fast inference."
@@ -1834,6 +1932,8 @@ var chutesModels = {
1834
1932
  contextWindow: 202752,
1835
1933
  supportsImages: false,
1836
1934
  supportsPromptCache: false,
1935
+ supportsNativeTools: true,
1936
+ defaultToolProtocol: "native",
1837
1937
  inputPrice: 0,
1838
1938
  outputPrice: 0,
1839
1939
  description: "GLM-4.6 introduces major upgrades over GLM-4.5, including a longer 200K-token context window for complex tasks, stronger coding performance in benchmarks and real-world tools (such as Claude Code, Cline, Roo Code, and Kilo Code), improved reasoning with tool use during inference, more capable and efficient agent integration, and refined writing that better matches human style, readability, and natural role-play scenarios."
@@ -1844,6 +1944,8 @@ var chutesModels = {
1844
1944
  contextWindow: 202752,
1845
1945
  supportsImages: false,
1846
1946
  supportsPromptCache: false,
1947
+ supportsNativeTools: true,
1948
+ defaultToolProtocol: "native",
1847
1949
  inputPrice: 1.15,
1848
1950
  outputPrice: 3.25,
1849
1951
  description: "GLM-4.6-turbo model with 200K-token context window, optimized for fast inference."
@@ -1853,6 +1955,8 @@ var chutesModels = {
1853
1955
  contextWindow: 128e3,
1854
1956
  supportsImages: false,
1855
1957
  supportsPromptCache: false,
1958
+ supportsNativeTools: true,
1959
+ defaultToolProtocol: "native",
1856
1960
  inputPrice: 0,
1857
1961
  outputPrice: 0,
1858
1962
  description: "LongCat Flash Thinking FP8 model with 128K context window, optimized for complex reasoning and coding tasks."
@@ -1862,6 +1966,8 @@ var chutesModels = {
1862
1966
  contextWindow: 262144,
1863
1967
  supportsImages: false,
1864
1968
  supportsPromptCache: false,
1969
+ supportsNativeTools: true,
1970
+ defaultToolProtocol: "native",
1865
1971
  inputPrice: 0,
1866
1972
  outputPrice: 0,
1867
1973
  description: "Qwen3 Coder 480B A35B Instruct FP8 model, optimized for coding tasks."
@@ -1871,6 +1977,8 @@ var chutesModels = {
1871
1977
  contextWindow: 75e3,
1872
1978
  supportsImages: false,
1873
1979
  supportsPromptCache: false,
1980
+ supportsNativeTools: true,
1981
+ defaultToolProtocol: "native",
1874
1982
  inputPrice: 0.1481,
1875
1983
  outputPrice: 0.5926,
1876
1984
  description: "Moonshot AI Kimi K2 Instruct model with 75k context window."
@@ -1880,6 +1988,8 @@ var chutesModels = {
1880
1988
  contextWindow: 262144,
1881
1989
  supportsImages: false,
1882
1990
  supportsPromptCache: false,
1991
+ supportsNativeTools: true,
1992
+ defaultToolProtocol: "native",
1883
1993
  inputPrice: 0.1999,
1884
1994
  outputPrice: 0.8001,
1885
1995
  description: "Moonshot AI Kimi K2 Instruct 0905 model with 256k context window."
@@ -1889,6 +1999,8 @@ var chutesModels = {
1889
1999
  contextWindow: 262144,
1890
2000
  supportsImages: false,
1891
2001
  supportsPromptCache: false,
2002
+ supportsNativeTools: true,
2003
+ defaultToolProtocol: "native",
1892
2004
  inputPrice: 0.077968332,
1893
2005
  outputPrice: 0.31202496,
1894
2006
  description: "Qwen3 235B A22B Thinking 2507 model with 262K context window."
@@ -1898,6 +2010,8 @@ var chutesModels = {
1898
2010
  contextWindow: 131072,
1899
2011
  supportsImages: false,
1900
2012
  supportsPromptCache: false,
2013
+ supportsNativeTools: true,
2014
+ defaultToolProtocol: "native",
1901
2015
  inputPrice: 0,
1902
2016
  outputPrice: 0,
1903
2017
  description: "Fast, stable instruction-tuned model optimized for complex tasks, RAG, and tool use without thinking traces."
@@ -1907,6 +2021,8 @@ var chutesModels = {
1907
2021
  contextWindow: 131072,
1908
2022
  supportsImages: false,
1909
2023
  supportsPromptCache: false,
2024
+ supportsNativeTools: true,
2025
+ defaultToolProtocol: "native",
1910
2026
  inputPrice: 0,
1911
2027
  outputPrice: 0,
1912
2028
  description: "Reasoning-first model with structured thinking traces for multi-step problems, math proofs, and code synthesis."
@@ -1916,6 +2032,8 @@ var chutesModels = {
1916
2032
  contextWindow: 262144,
1917
2033
  supportsImages: true,
1918
2034
  supportsPromptCache: false,
2035
+ supportsNativeTools: true,
2036
+ defaultToolProtocol: "native",
1919
2037
  inputPrice: 0.16,
1920
2038
  outputPrice: 0.65,
1921
2039
  description: "Qwen3\u2011VL\u2011235B\u2011A22B\u2011Thinking is an open\u2011weight MoE vision\u2011language model (235B total, ~22B activated) optimized for deliberate multi\u2011step reasoning with strong text\u2011image\u2011video understanding and long\u2011context capabilities."
@@ -1924,139 +2042,71 @@ var chutesModels = {
1924
2042
  var chutesDefaultModelInfo = chutesModels[chutesDefaultModelId];
1925
2043
 
1926
2044
  // src/providers/claude-code.ts
1927
- var VERTEX_DATE_PATTERN = /-(\d{8})$/;
1928
- function convertModelNameForVertex(modelName) {
1929
- return modelName.replace(VERTEX_DATE_PATTERN, "@$1");
1930
- }
1931
- var claudeCodeDefaultModelId = "claude-sonnet-4-5";
1932
- var CLAUDE_CODE_DEFAULT_MAX_OUTPUT_TOKENS = 16e3;
1933
- function getClaudeCodeModelId(baseModelId, useVertex = false) {
1934
- return useVertex ? convertModelNameForVertex(baseModelId) : baseModelId;
1935
- }
2045
+ var DATE_SUFFIX_PATTERN = /-\d{8}$/;
1936
2046
  var claudeCodeModels = {
1937
- "claude-sonnet-4-5": {
1938
- ...anthropicModels["claude-sonnet-4-5"],
1939
- supportsImages: false,
1940
- supportsPromptCache: true,
1941
- // Claude Code does report cache tokens
1942
- supportsReasoningEffort: false,
1943
- supportsReasoningBudget: false,
1944
- requiredReasoningBudget: false,
1945
- // Claude Code manages its own tools and temperature via the CLI
1946
- supportsNativeTools: false,
1947
- supportsTemperature: false
1948
- },
1949
- "claude-sonnet-4-5-20250929[1m]": {
1950
- ...anthropicModels["claude-sonnet-4-5"],
1951
- contextWindow: 1e6,
1952
- // 1M token context window (requires [1m] suffix)
1953
- supportsImages: false,
1954
- supportsPromptCache: true,
1955
- // Claude Code does report cache tokens
1956
- supportsReasoningEffort: false,
1957
- supportsReasoningBudget: false,
1958
- requiredReasoningBudget: false,
1959
- // Claude Code manages its own tools and temperature via the CLI
1960
- supportsNativeTools: false,
1961
- supportsTemperature: false
1962
- },
1963
- "claude-sonnet-4-20250514": {
1964
- ...anthropicModels["claude-sonnet-4-20250514"],
1965
- supportsImages: false,
1966
- supportsPromptCache: true,
1967
- // Claude Code does report cache tokens
1968
- supportsReasoningEffort: false,
1969
- supportsReasoningBudget: false,
1970
- requiredReasoningBudget: false,
1971
- // Claude Code manages its own tools and temperature via the CLI
1972
- supportsNativeTools: false,
1973
- supportsTemperature: false
1974
- },
1975
- "claude-opus-4-5-20251101": {
1976
- ...anthropicModels["claude-opus-4-5-20251101"],
1977
- supportsImages: false,
1978
- supportsPromptCache: true,
1979
- // Claude Code does report cache tokens
1980
- supportsReasoningEffort: false,
1981
- supportsReasoningBudget: false,
1982
- requiredReasoningBudget: false,
1983
- // Claude Code manages its own tools and temperature via the CLI
1984
- supportsNativeTools: false,
1985
- supportsTemperature: false
1986
- },
1987
- "claude-opus-4-1-20250805": {
1988
- ...anthropicModels["claude-opus-4-1-20250805"],
1989
- supportsImages: false,
1990
- supportsPromptCache: true,
1991
- // Claude Code does report cache tokens
1992
- supportsReasoningEffort: false,
1993
- supportsReasoningBudget: false,
1994
- requiredReasoningBudget: false,
1995
- // Claude Code manages its own tools and temperature via the CLI
1996
- supportsNativeTools: false,
1997
- supportsTemperature: false
1998
- },
1999
- "claude-opus-4-20250514": {
2000
- ...anthropicModels["claude-opus-4-20250514"],
2001
- supportsImages: false,
2002
- supportsPromptCache: true,
2003
- // Claude Code does report cache tokens
2004
- supportsReasoningEffort: false,
2005
- supportsReasoningBudget: false,
2006
- requiredReasoningBudget: false,
2007
- // Claude Code manages its own tools and temperature via the CLI
2008
- supportsNativeTools: false,
2009
- supportsTemperature: false
2010
- },
2011
- "claude-3-7-sonnet-20250219": {
2012
- ...anthropicModels["claude-3-7-sonnet-20250219"],
2013
- supportsImages: false,
2014
- supportsPromptCache: true,
2015
- // Claude Code does report cache tokens
2016
- supportsReasoningEffort: false,
2017
- supportsReasoningBudget: false,
2018
- requiredReasoningBudget: false,
2019
- // Claude Code manages its own tools and temperature via the CLI
2020
- supportsNativeTools: false,
2021
- supportsTemperature: false
2022
- },
2023
- "claude-3-5-sonnet-20241022": {
2024
- ...anthropicModels["claude-3-5-sonnet-20241022"],
2025
- supportsImages: false,
2047
+ "claude-haiku-4-5": {
2048
+ maxTokens: 32768,
2049
+ contextWindow: 2e5,
2050
+ supportsImages: true,
2026
2051
  supportsPromptCache: true,
2027
- // Claude Code does report cache tokens
2028
- supportsReasoningEffort: false,
2029
- supportsReasoningBudget: false,
2030
- requiredReasoningBudget: false,
2031
- // Claude Code manages its own tools and temperature via the CLI
2032
- supportsNativeTools: false,
2033
- supportsTemperature: false
2052
+ supportsNativeTools: true,
2053
+ defaultToolProtocol: "native",
2054
+ supportsReasoningEffort: ["disable", "low", "medium", "high"],
2055
+ reasoningEffort: "medium",
2056
+ description: "Claude Haiku 4.5 - Fast and efficient with thinking"
2034
2057
  },
2035
- "claude-3-5-haiku-20241022": {
2036
- ...anthropicModels["claude-3-5-haiku-20241022"],
2037
- supportsImages: false,
2058
+ "claude-sonnet-4-5": {
2059
+ maxTokens: 32768,
2060
+ contextWindow: 2e5,
2061
+ supportsImages: true,
2038
2062
  supportsPromptCache: true,
2039
- // Claude Code does report cache tokens
2040
- supportsReasoningEffort: false,
2041
- supportsReasoningBudget: false,
2042
- requiredReasoningBudget: false,
2043
- // Claude Code manages its own tools and temperature via the CLI
2044
- supportsNativeTools: false,
2045
- supportsTemperature: false
2063
+ supportsNativeTools: true,
2064
+ defaultToolProtocol: "native",
2065
+ supportsReasoningEffort: ["disable", "low", "medium", "high"],
2066
+ reasoningEffort: "medium",
2067
+ description: "Claude Sonnet 4.5 - Balanced performance with thinking"
2046
2068
  },
2047
- "claude-haiku-4-5-20251001": {
2048
- ...anthropicModels["claude-haiku-4-5-20251001"],
2049
- supportsImages: false,
2069
+ "claude-opus-4-5": {
2070
+ maxTokens: 32768,
2071
+ contextWindow: 2e5,
2072
+ supportsImages: true,
2050
2073
  supportsPromptCache: true,
2051
- // Claude Code does report cache tokens
2052
- supportsReasoningEffort: false,
2053
- supportsReasoningBudget: false,
2054
- requiredReasoningBudget: false,
2055
- // Claude Code manages its own tools and temperature via the CLI
2056
- supportsNativeTools: false,
2057
- supportsTemperature: false
2074
+ supportsNativeTools: true,
2075
+ defaultToolProtocol: "native",
2076
+ supportsReasoningEffort: ["disable", "low", "medium", "high"],
2077
+ reasoningEffort: "medium",
2078
+ description: "Claude Opus 4.5 - Most capable with thinking"
2058
2079
  }
2059
2080
  };
2081
+ var claudeCodeDefaultModelId = "claude-sonnet-4-5";
2082
+ var MODEL_FAMILY_PATTERNS = [
2083
+ // Opus models (any version) → claude-opus-4-5
2084
+ { pattern: /opus/i, target: "claude-opus-4-5" },
2085
+ // Haiku models (any version) → claude-haiku-4-5
2086
+ { pattern: /haiku/i, target: "claude-haiku-4-5" },
2087
+ // Sonnet models (any version) → claude-sonnet-4-5
2088
+ { pattern: /sonnet/i, target: "claude-sonnet-4-5" }
2089
+ ];
2090
+ function normalizeClaudeCodeModelId(modelId) {
2091
+ if (Object.hasOwn(claudeCodeModels, modelId)) {
2092
+ return modelId;
2093
+ }
2094
+ const withoutDate = modelId.replace(DATE_SUFFIX_PATTERN, "");
2095
+ if (Object.hasOwn(claudeCodeModels, withoutDate)) {
2096
+ return withoutDate;
2097
+ }
2098
+ for (const { pattern, target } of MODEL_FAMILY_PATTERNS) {
2099
+ if (pattern.test(modelId)) {
2100
+ return target;
2101
+ }
2102
+ }
2103
+ return claudeCodeDefaultModelId;
2104
+ }
2105
+ var claudeCodeReasoningConfig = {
2106
+ low: { budgetTokens: 16e3 },
2107
+ medium: { budgetTokens: 32e3 },
2108
+ high: { budgetTokens: 64e3 }
2109
+ };
2060
2110
 
2061
2111
  // src/providers/deepseek.ts
2062
2112
  var deepSeekDefaultModelId = "deepseek-chat";
@@ -2087,6 +2137,7 @@ var deepSeekModels = {
2087
2137
  supportsPromptCache: true,
2088
2138
  supportsNativeTools: true,
2089
2139
  defaultToolProtocol: "native",
2140
+ preserveReasoning: true,
2090
2141
  inputPrice: 0.28,
2091
2142
  // $0.28 per million tokens (cache miss) - Updated Dec 9, 2025
2092
2143
  outputPrice: 0.42,
@@ -2098,7 +2149,7 @@ var deepSeekModels = {
2098
2149
  description: `DeepSeek-V3.2 (Thinking Mode) achieves performance comparable to OpenAI-o1 across math, code, and reasoning tasks. Supports Chain of Thought reasoning with up to 8K output tokens. Supports JSON output, tool calls, and chat prefix completion (beta).`
2099
2150
  }
2100
2151
  };
2101
- var DEEP_SEEK_DEFAULT_TEMPERATURE = 0;
2152
+ var DEEP_SEEK_DEFAULT_TEMPERATURE = 0.3;
2102
2153
 
2103
2154
  // src/providers/doubao.ts
2104
2155
  var doubaoDefaultModelId = "doubao-seed-1-6-250615";
@@ -2109,6 +2160,7 @@ var doubaoModels = {
2109
2160
  supportsImages: true,
2110
2161
  supportsPromptCache: true,
2111
2162
  supportsNativeTools: true,
2163
+ defaultToolProtocol: "native",
2112
2164
  inputPrice: 1e-4,
2113
2165
  // $0.0001 per million tokens (cache miss)
2114
2166
  outputPrice: 4e-4,
@@ -2125,6 +2177,7 @@ var doubaoModels = {
2125
2177
  supportsImages: true,
2126
2178
  supportsPromptCache: true,
2127
2179
  supportsNativeTools: true,
2180
+ defaultToolProtocol: "native",
2128
2181
  inputPrice: 2e-4,
2129
2182
  // $0.0002 per million tokens
2130
2183
  outputPrice: 8e-4,
@@ -2141,6 +2194,7 @@ var doubaoModels = {
2141
2194
  supportsImages: true,
2142
2195
  supportsPromptCache: true,
2143
2196
  supportsNativeTools: true,
2197
+ defaultToolProtocol: "native",
2144
2198
  inputPrice: 15e-5,
2145
2199
  // $0.00015 per million tokens
2146
2200
  outputPrice: 6e-4,
@@ -2217,6 +2271,7 @@ var fireworksModels = {
2217
2271
  supportsImages: false,
2218
2272
  supportsPromptCache: true,
2219
2273
  supportsNativeTools: true,
2274
+ defaultToolProtocol: "native",
2220
2275
  inputPrice: 0.6,
2221
2276
  outputPrice: 2.5,
2222
2277
  cacheReadsPrice: 0.15,
@@ -2228,6 +2283,7 @@ var fireworksModels = {
2228
2283
  supportsImages: false,
2229
2284
  supportsPromptCache: false,
2230
2285
  supportsNativeTools: true,
2286
+ defaultToolProtocol: "native",
2231
2287
  inputPrice: 0.6,
2232
2288
  outputPrice: 2.5,
2233
2289
  description: "Kimi K2 is a state-of-the-art mixture-of-experts (MoE) language model with 32 billion activated parameters and 1 trillion total parameters. Trained with the Muon optimizer, Kimi K2 achieves exceptional performance across frontier knowledge, reasoning, and coding tasks while being meticulously optimized for agentic capabilities."
@@ -2238,6 +2294,7 @@ var fireworksModels = {
2238
2294
  supportsImages: false,
2239
2295
  supportsPromptCache: false,
2240
2296
  supportsNativeTools: true,
2297
+ defaultToolProtocol: "native",
2241
2298
  inputPrice: 0.3,
2242
2299
  outputPrice: 1.2,
2243
2300
  description: "MiniMax M2 is a high-performance language model with 204.8K context window, optimized for long-context understanding and generation tasks."
@@ -2248,6 +2305,7 @@ var fireworksModels = {
2248
2305
  supportsImages: false,
2249
2306
  supportsPromptCache: false,
2250
2307
  supportsNativeTools: true,
2308
+ defaultToolProtocol: "native",
2251
2309
  inputPrice: 0.22,
2252
2310
  outputPrice: 0.88,
2253
2311
  description: "Latest Qwen3 thinking model, competitive against the best closed source models in Jul 2025."
@@ -2258,6 +2316,7 @@ var fireworksModels = {
2258
2316
  supportsImages: false,
2259
2317
  supportsPromptCache: false,
2260
2318
  supportsNativeTools: true,
2319
+ defaultToolProtocol: "native",
2261
2320
  inputPrice: 0.45,
2262
2321
  outputPrice: 1.8,
2263
2322
  description: "Qwen3's most agentic code model to date."
@@ -2268,6 +2327,7 @@ var fireworksModels = {
2268
2327
  supportsImages: false,
2269
2328
  supportsPromptCache: false,
2270
2329
  supportsNativeTools: true,
2330
+ defaultToolProtocol: "native",
2271
2331
  inputPrice: 3,
2272
2332
  outputPrice: 8,
2273
2333
  description: "05/28 updated checkpoint of Deepseek R1. Its overall performance is now approaching that of leading models, such as O3 and Gemini 2.5 Pro. Compared to the previous version, the upgraded model shows significant improvements in handling complex reasoning tasks, and this version also offers a reduced hallucination rate, enhanced support for function calling, and better experience for vibe coding. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
@@ -2278,6 +2338,7 @@ var fireworksModels = {
2278
2338
  supportsImages: false,
2279
2339
  supportsPromptCache: false,
2280
2340
  supportsNativeTools: true,
2341
+ defaultToolProtocol: "native",
2281
2342
  inputPrice: 0.9,
2282
2343
  outputPrice: 0.9,
2283
2344
  description: "A strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token from Deepseek. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
@@ -2288,6 +2349,7 @@ var fireworksModels = {
2288
2349
  supportsImages: false,
2289
2350
  supportsPromptCache: false,
2290
2351
  supportsNativeTools: true,
2352
+ defaultToolProtocol: "native",
2291
2353
  inputPrice: 0.56,
2292
2354
  outputPrice: 1.68,
2293
2355
  description: "DeepSeek v3.1 is an improved version of the v3 model with enhanced performance, better reasoning capabilities, and improved code generation. This Mixture-of-Experts (MoE) model maintains the same 671B total parameters with 37B activated per token."
@@ -2298,6 +2360,7 @@ var fireworksModels = {
2298
2360
  supportsImages: false,
2299
2361
  supportsPromptCache: false,
2300
2362
  supportsNativeTools: true,
2363
+ defaultToolProtocol: "native",
2301
2364
  inputPrice: 0.55,
2302
2365
  outputPrice: 2.19,
2303
2366
  description: "Z.ai GLM-4.5 with 355B total parameters and 32B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
@@ -2308,6 +2371,7 @@ var fireworksModels = {
2308
2371
  supportsImages: false,
2309
2372
  supportsPromptCache: false,
2310
2373
  supportsNativeTools: true,
2374
+ defaultToolProtocol: "native",
2311
2375
  inputPrice: 0.55,
2312
2376
  outputPrice: 2.19,
2313
2377
  description: "Z.ai GLM-4.5-Air with 106B total parameters and 12B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
@@ -2318,6 +2382,7 @@ var fireworksModels = {
2318
2382
  supportsImages: false,
2319
2383
  supportsPromptCache: false,
2320
2384
  supportsNativeTools: true,
2385
+ defaultToolProtocol: "native",
2321
2386
  inputPrice: 0.55,
2322
2387
  outputPrice: 2.19,
2323
2388
  description: "Z.ai GLM-4.6 is an advanced coding model with exceptional performance on complex programming tasks. Features improved reasoning capabilities and enhanced code generation quality, making it ideal for software development workflows."
@@ -2328,6 +2393,7 @@ var fireworksModels = {
2328
2393
  supportsImages: false,
2329
2394
  supportsPromptCache: false,
2330
2395
  supportsNativeTools: true,
2396
+ defaultToolProtocol: "native",
2331
2397
  inputPrice: 0.07,
2332
2398
  outputPrice: 0.3,
2333
2399
  description: "OpenAI gpt-oss-20b: Compact model for local/edge deployments. Optimized for low-latency and resource-constrained environments with chain-of-thought output, adjustable reasoning, and agentic workflows."
@@ -2338,6 +2404,7 @@ var fireworksModels = {
2338
2404
  supportsImages: false,
2339
2405
  supportsPromptCache: false,
2340
2406
  supportsNativeTools: true,
2407
+ defaultToolProtocol: "native",
2341
2408
  inputPrice: 0.15,
2342
2409
  outputPrice: 0.6,
2343
2410
  description: "OpenAI gpt-oss-120b: Production-grade, general-purpose model that fits on a single H100 GPU. Features complex reasoning, configurable effort, full chain-of-thought transparency, and supports function calling, tool use, and structured outputs."
@@ -2345,16 +2412,19 @@ var fireworksModels = {
2345
2412
  };
2346
2413
 
2347
2414
  // src/providers/gemini.ts
2348
- var geminiDefaultModelId = "gemini-2.5-pro";
2415
+ var geminiDefaultModelId = "gemini-3-pro-preview";
2349
2416
  var geminiModels = {
2350
2417
  "gemini-3-pro-preview": {
2351
2418
  maxTokens: 65536,
2352
2419
  contextWindow: 1048576,
2353
2420
  supportsImages: true,
2354
2421
  supportsNativeTools: true,
2422
+ defaultToolProtocol: "native",
2355
2423
  supportsPromptCache: true,
2356
2424
  supportsReasoningEffort: ["low", "high"],
2357
2425
  reasoningEffort: "low",
2426
+ includedTools: ["write_file", "edit_file"],
2427
+ excludedTools: ["apply_diff"],
2358
2428
  supportsTemperature: true,
2359
2429
  defaultTemperature: 1,
2360
2430
  inputPrice: 4,
@@ -2372,13 +2442,34 @@ var geminiModels = {
2372
2442
  }
2373
2443
  ]
2374
2444
  },
2445
+ "gemini-3-flash-preview": {
2446
+ maxTokens: 65536,
2447
+ contextWindow: 1048576,
2448
+ supportsImages: true,
2449
+ supportsNativeTools: true,
2450
+ defaultToolProtocol: "native",
2451
+ supportsPromptCache: true,
2452
+ supportsReasoningEffort: ["minimal", "low", "medium", "high"],
2453
+ reasoningEffort: "medium",
2454
+ includedTools: ["write_file", "edit_file"],
2455
+ excludedTools: ["apply_diff"],
2456
+ supportsTemperature: true,
2457
+ defaultTemperature: 1,
2458
+ inputPrice: 0.3,
2459
+ outputPrice: 2.5,
2460
+ cacheReadsPrice: 0.075,
2461
+ cacheWritesPrice: 1
2462
+ },
2375
2463
  // 2.5 Pro models
2376
2464
  "gemini-2.5-pro": {
2377
2465
  maxTokens: 64e3,
2378
2466
  contextWindow: 1048576,
2379
2467
  supportsImages: true,
2380
2468
  supportsNativeTools: true,
2469
+ defaultToolProtocol: "native",
2381
2470
  supportsPromptCache: true,
2471
+ includedTools: ["write_file", "edit_file"],
2472
+ excludedTools: ["apply_diff"],
2382
2473
  inputPrice: 2.5,
2383
2474
  // This is the pricing for prompts above 200k tokens.
2384
2475
  outputPrice: 15,
@@ -2407,7 +2498,10 @@ var geminiModels = {
2407
2498
  contextWindow: 1048576,
2408
2499
  supportsImages: true,
2409
2500
  supportsNativeTools: true,
2501
+ defaultToolProtocol: "native",
2410
2502
  supportsPromptCache: true,
2503
+ includedTools: ["write_file", "edit_file"],
2504
+ excludedTools: ["apply_diff"],
2411
2505
  inputPrice: 2.5,
2412
2506
  // This is the pricing for prompts above 200k tokens.
2413
2507
  outputPrice: 15,
@@ -2435,7 +2529,10 @@ var geminiModels = {
2435
2529
  contextWindow: 1048576,
2436
2530
  supportsImages: true,
2437
2531
  supportsNativeTools: true,
2532
+ defaultToolProtocol: "native",
2438
2533
  supportsPromptCache: true,
2534
+ includedTools: ["write_file", "edit_file"],
2535
+ excludedTools: ["apply_diff"],
2439
2536
  inputPrice: 2.5,
2440
2537
  // This is the pricing for prompts above 200k tokens.
2441
2538
  outputPrice: 15,
@@ -2461,7 +2558,10 @@ var geminiModels = {
2461
2558
  contextWindow: 1048576,
2462
2559
  supportsImages: true,
2463
2560
  supportsNativeTools: true,
2561
+ defaultToolProtocol: "native",
2464
2562
  supportsPromptCache: true,
2563
+ includedTools: ["write_file", "edit_file"],
2564
+ excludedTools: ["apply_diff"],
2465
2565
  inputPrice: 2.5,
2466
2566
  // This is the pricing for prompts above 200k tokens.
2467
2567
  outputPrice: 15,
@@ -2490,7 +2590,10 @@ var geminiModels = {
2490
2590
  contextWindow: 1048576,
2491
2591
  supportsImages: true,
2492
2592
  supportsNativeTools: true,
2593
+ defaultToolProtocol: "native",
2493
2594
  supportsPromptCache: true,
2595
+ includedTools: ["write_file", "edit_file"],
2596
+ excludedTools: ["apply_diff"],
2494
2597
  inputPrice: 0.3,
2495
2598
  outputPrice: 2.5,
2496
2599
  cacheReadsPrice: 0.075,
@@ -2503,7 +2606,10 @@ var geminiModels = {
2503
2606
  contextWindow: 1048576,
2504
2607
  supportsImages: true,
2505
2608
  supportsNativeTools: true,
2609
+ defaultToolProtocol: "native",
2506
2610
  supportsPromptCache: true,
2611
+ includedTools: ["write_file", "edit_file"],
2612
+ excludedTools: ["apply_diff"],
2507
2613
  inputPrice: 0.3,
2508
2614
  outputPrice: 2.5,
2509
2615
  cacheReadsPrice: 0.075,
@@ -2516,7 +2622,10 @@ var geminiModels = {
2516
2622
  contextWindow: 1048576,
2517
2623
  supportsImages: true,
2518
2624
  supportsNativeTools: true,
2625
+ defaultToolProtocol: "native",
2519
2626
  supportsPromptCache: true,
2627
+ includedTools: ["write_file", "edit_file"],
2628
+ excludedTools: ["apply_diff"],
2520
2629
  inputPrice: 0.3,
2521
2630
  outputPrice: 2.5,
2522
2631
  cacheReadsPrice: 0.075,
@@ -2530,7 +2639,10 @@ var geminiModels = {
2530
2639
  contextWindow: 1048576,
2531
2640
  supportsImages: true,
2532
2641
  supportsNativeTools: true,
2642
+ defaultToolProtocol: "native",
2533
2643
  supportsPromptCache: true,
2644
+ includedTools: ["write_file", "edit_file"],
2645
+ excludedTools: ["apply_diff"],
2534
2646
  inputPrice: 0.1,
2535
2647
  outputPrice: 0.4,
2536
2648
  cacheReadsPrice: 0.025,
@@ -2543,7 +2655,10 @@ var geminiModels = {
2543
2655
  contextWindow: 1048576,
2544
2656
  supportsImages: true,
2545
2657
  supportsNativeTools: true,
2658
+ defaultToolProtocol: "native",
2546
2659
  supportsPromptCache: true,
2660
+ includedTools: ["write_file", "edit_file"],
2661
+ excludedTools: ["apply_diff"],
2547
2662
  inputPrice: 0.1,
2548
2663
  outputPrice: 0.4,
2549
2664
  cacheReadsPrice: 0.025,
@@ -2563,6 +2678,7 @@ var groqModels = {
2563
2678
  supportsImages: false,
2564
2679
  supportsPromptCache: false,
2565
2680
  supportsNativeTools: true,
2681
+ defaultToolProtocol: "native",
2566
2682
  inputPrice: 0.05,
2567
2683
  outputPrice: 0.08,
2568
2684
  description: "Meta Llama 3.1 8B Instant model, 128K context."
@@ -2573,6 +2689,7 @@ var groqModels = {
2573
2689
  supportsImages: false,
2574
2690
  supportsPromptCache: false,
2575
2691
  supportsNativeTools: true,
2692
+ defaultToolProtocol: "native",
2576
2693
  inputPrice: 0.59,
2577
2694
  outputPrice: 0.79,
2578
2695
  description: "Meta Llama 3.3 70B Versatile model, 128K context."
@@ -2583,6 +2700,7 @@ var groqModels = {
2583
2700
  supportsImages: false,
2584
2701
  supportsPromptCache: false,
2585
2702
  supportsNativeTools: true,
2703
+ defaultToolProtocol: "native",
2586
2704
  inputPrice: 0.11,
2587
2705
  outputPrice: 0.34,
2588
2706
  description: "Meta Llama 4 Scout 17B Instruct model, 128K context."
@@ -2620,6 +2738,7 @@ var groqModels = {
2620
2738
  supportsImages: false,
2621
2739
  supportsPromptCache: false,
2622
2740
  supportsNativeTools: true,
2741
+ defaultToolProtocol: "native",
2623
2742
  inputPrice: 0.29,
2624
2743
  outputPrice: 0.59,
2625
2744
  description: "Alibaba Qwen 3 32B model, 128K context."
@@ -2650,6 +2769,7 @@ var groqModels = {
2650
2769
  supportsImages: false,
2651
2770
  supportsPromptCache: true,
2652
2771
  supportsNativeTools: true,
2772
+ defaultToolProtocol: "native",
2653
2773
  inputPrice: 0.6,
2654
2774
  outputPrice: 2.5,
2655
2775
  cacheReadsPrice: 0.15,
@@ -2661,6 +2781,7 @@ var groqModels = {
2661
2781
  supportsImages: false,
2662
2782
  supportsPromptCache: false,
2663
2783
  supportsNativeTools: true,
2784
+ defaultToolProtocol: "native",
2664
2785
  inputPrice: 0.15,
2665
2786
  outputPrice: 0.75,
2666
2787
  description: "GPT-OSS 120B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 128 experts."
@@ -2671,6 +2792,7 @@ var groqModels = {
2671
2792
  supportsImages: false,
2672
2793
  supportsPromptCache: false,
2673
2794
  supportsNativeTools: true,
2795
+ defaultToolProtocol: "native",
2674
2796
  inputPrice: 0.1,
2675
2797
  outputPrice: 0.5,
2676
2798
  description: "GPT-OSS 20B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 32 experts."
@@ -2734,6 +2856,7 @@ var litellmDefaultModelInfo = {
2734
2856
  supportsImages: true,
2735
2857
  supportsPromptCache: true,
2736
2858
  supportsNativeTools: true,
2859
+ defaultToolProtocol: "native",
2737
2860
  inputPrice: 3,
2738
2861
  outputPrice: 15,
2739
2862
  cacheWritesPrice: 3.75,
@@ -2748,6 +2871,8 @@ var lMStudioDefaultModelInfo = {
2748
2871
  contextWindow: 2e5,
2749
2872
  supportsImages: true,
2750
2873
  supportsPromptCache: true,
2874
+ supportsNativeTools: true,
2875
+ defaultToolProtocol: "native",
2751
2876
  inputPrice: 0,
2752
2877
  outputPrice: 0,
2753
2878
  cacheWritesPrice: 0,
@@ -2764,6 +2889,7 @@ var mistralModels = {
2764
2889
  supportsImages: true,
2765
2890
  supportsPromptCache: false,
2766
2891
  supportsNativeTools: true,
2892
+ defaultToolProtocol: "native",
2767
2893
  inputPrice: 2,
2768
2894
  outputPrice: 5
2769
2895
  },
@@ -2773,6 +2899,7 @@ var mistralModels = {
2773
2899
  supportsImages: true,
2774
2900
  supportsPromptCache: false,
2775
2901
  supportsNativeTools: true,
2902
+ defaultToolProtocol: "native",
2776
2903
  inputPrice: 0.4,
2777
2904
  outputPrice: 2
2778
2905
  },
@@ -2782,6 +2909,7 @@ var mistralModels = {
2782
2909
  supportsImages: true,
2783
2910
  supportsPromptCache: false,
2784
2911
  supportsNativeTools: true,
2912
+ defaultToolProtocol: "native",
2785
2913
  inputPrice: 0.4,
2786
2914
  outputPrice: 2
2787
2915
  },
@@ -2791,6 +2919,7 @@ var mistralModels = {
2791
2919
  supportsImages: false,
2792
2920
  supportsPromptCache: false,
2793
2921
  supportsNativeTools: true,
2922
+ defaultToolProtocol: "native",
2794
2923
  inputPrice: 0.3,
2795
2924
  outputPrice: 0.9
2796
2925
  },
@@ -2800,6 +2929,7 @@ var mistralModels = {
2800
2929
  supportsImages: false,
2801
2930
  supportsPromptCache: false,
2802
2931
  supportsNativeTools: true,
2932
+ defaultToolProtocol: "native",
2803
2933
  inputPrice: 2,
2804
2934
  outputPrice: 6
2805
2935
  },
@@ -2809,6 +2939,7 @@ var mistralModels = {
2809
2939
  supportsImages: false,
2810
2940
  supportsPromptCache: false,
2811
2941
  supportsNativeTools: true,
2942
+ defaultToolProtocol: "native",
2812
2943
  inputPrice: 0.1,
2813
2944
  outputPrice: 0.1
2814
2945
  },
@@ -2818,6 +2949,7 @@ var mistralModels = {
2818
2949
  supportsImages: false,
2819
2950
  supportsPromptCache: false,
2820
2951
  supportsNativeTools: true,
2952
+ defaultToolProtocol: "native",
2821
2953
  inputPrice: 0.04,
2822
2954
  outputPrice: 0.04
2823
2955
  },
@@ -2827,6 +2959,7 @@ var mistralModels = {
2827
2959
  supportsImages: false,
2828
2960
  supportsPromptCache: false,
2829
2961
  supportsNativeTools: true,
2962
+ defaultToolProtocol: "native",
2830
2963
  inputPrice: 0.2,
2831
2964
  outputPrice: 0.6
2832
2965
  },
@@ -2836,6 +2969,7 @@ var mistralModels = {
2836
2969
  supportsImages: true,
2837
2970
  supportsPromptCache: false,
2838
2971
  supportsNativeTools: true,
2972
+ defaultToolProtocol: "native",
2839
2973
  inputPrice: 2,
2840
2974
  outputPrice: 6
2841
2975
  }
@@ -2851,6 +2985,7 @@ var moonshotModels = {
2851
2985
  supportsImages: false,
2852
2986
  supportsPromptCache: true,
2853
2987
  supportsNativeTools: true,
2988
+ defaultToolProtocol: "native",
2854
2989
  inputPrice: 0.6,
2855
2990
  // $0.60 per million tokens (cache miss)
2856
2991
  outputPrice: 2.5,
@@ -2867,6 +3002,7 @@ var moonshotModels = {
2867
3002
  supportsImages: false,
2868
3003
  supportsPromptCache: true,
2869
3004
  supportsNativeTools: true,
3005
+ defaultToolProtocol: "native",
2870
3006
  inputPrice: 0.6,
2871
3007
  outputPrice: 2.5,
2872
3008
  cacheReadsPrice: 0.15,
@@ -2878,6 +3014,7 @@ var moonshotModels = {
2878
3014
  supportsImages: false,
2879
3015
  supportsPromptCache: true,
2880
3016
  supportsNativeTools: true,
3017
+ defaultToolProtocol: "native",
2881
3018
  inputPrice: 2.4,
2882
3019
  // $2.40 per million tokens (cache miss)
2883
3020
  outputPrice: 10,
@@ -2897,6 +3034,7 @@ var moonshotModels = {
2897
3034
  // Text-only (no image/vision support)
2898
3035
  supportsPromptCache: true,
2899
3036
  supportsNativeTools: true,
3037
+ defaultToolProtocol: "native",
2900
3038
  inputPrice: 0.6,
2901
3039
  // $0.60 per million tokens (cache miss)
2902
3040
  outputPrice: 2.5,
@@ -2936,6 +3074,7 @@ var openAiNativeModels = {
2936
3074
  maxTokens: 128e3,
2937
3075
  contextWindow: 4e5,
2938
3076
  supportsNativeTools: true,
3077
+ defaultToolProtocol: "native",
2939
3078
  includedTools: ["apply_patch"],
2940
3079
  excludedTools: ["apply_diff", "write_to_file"],
2941
3080
  supportsImages: true,
@@ -2954,6 +3093,7 @@ var openAiNativeModels = {
2954
3093
  maxTokens: 128e3,
2955
3094
  contextWindow: 4e5,
2956
3095
  supportsNativeTools: true,
3096
+ defaultToolProtocol: "native",
2957
3097
  includedTools: ["apply_patch"],
2958
3098
  excludedTools: ["apply_diff", "write_to_file"],
2959
3099
  supportsImages: true,
@@ -2976,6 +3116,7 @@ var openAiNativeModels = {
2976
3116
  maxTokens: 16384,
2977
3117
  contextWindow: 128e3,
2978
3118
  supportsNativeTools: true,
3119
+ defaultToolProtocol: "native",
2979
3120
  includedTools: ["apply_patch"],
2980
3121
  excludedTools: ["apply_diff", "write_to_file"],
2981
3122
  supportsImages: true,
@@ -2989,6 +3130,7 @@ var openAiNativeModels = {
2989
3130
  maxTokens: 128e3,
2990
3131
  contextWindow: 4e5,
2991
3132
  supportsNativeTools: true,
3133
+ defaultToolProtocol: "native",
2992
3134
  includedTools: ["apply_patch"],
2993
3135
  excludedTools: ["apply_diff", "write_to_file"],
2994
3136
  supportsImages: true,
@@ -3011,6 +3153,7 @@ var openAiNativeModels = {
3011
3153
  maxTokens: 128e3,
3012
3154
  contextWindow: 4e5,
3013
3155
  supportsNativeTools: true,
3156
+ defaultToolProtocol: "native",
3014
3157
  includedTools: ["apply_patch"],
3015
3158
  excludedTools: ["apply_diff", "write_to_file"],
3016
3159
  supportsImages: true,
@@ -3029,6 +3172,7 @@ var openAiNativeModels = {
3029
3172
  maxTokens: 128e3,
3030
3173
  contextWindow: 4e5,
3031
3174
  supportsNativeTools: true,
3175
+ defaultToolProtocol: "native",
3032
3176
  includedTools: ["apply_patch"],
3033
3177
  excludedTools: ["apply_diff", "write_to_file"],
3034
3178
  supportsImages: true,
@@ -3046,6 +3190,7 @@ var openAiNativeModels = {
3046
3190
  maxTokens: 128e3,
3047
3191
  contextWindow: 4e5,
3048
3192
  supportsNativeTools: true,
3193
+ defaultToolProtocol: "native",
3049
3194
  includedTools: ["apply_patch"],
3050
3195
  excludedTools: ["apply_diff", "write_to_file"],
3051
3196
  supportsImages: true,
@@ -3067,6 +3212,7 @@ var openAiNativeModels = {
3067
3212
  maxTokens: 128e3,
3068
3213
  contextWindow: 4e5,
3069
3214
  supportsNativeTools: true,
3215
+ defaultToolProtocol: "native",
3070
3216
  includedTools: ["apply_patch"],
3071
3217
  excludedTools: ["apply_diff", "write_to_file"],
3072
3218
  supportsImages: true,
@@ -3088,6 +3234,7 @@ var openAiNativeModels = {
3088
3234
  maxTokens: 128e3,
3089
3235
  contextWindow: 4e5,
3090
3236
  supportsNativeTools: true,
3237
+ defaultToolProtocol: "native",
3091
3238
  includedTools: ["apply_patch"],
3092
3239
  excludedTools: ["apply_diff", "write_to_file"],
3093
3240
  supportsImages: true,
@@ -3105,6 +3252,7 @@ var openAiNativeModels = {
3105
3252
  maxTokens: 128e3,
3106
3253
  contextWindow: 4e5,
3107
3254
  supportsNativeTools: true,
3255
+ defaultToolProtocol: "native",
3108
3256
  includedTools: ["apply_patch"],
3109
3257
  excludedTools: ["apply_diff", "write_to_file"],
3110
3258
  supportsImages: true,
@@ -3123,6 +3271,7 @@ var openAiNativeModels = {
3123
3271
  maxTokens: 128e3,
3124
3272
  contextWindow: 4e5,
3125
3273
  supportsNativeTools: true,
3274
+ defaultToolProtocol: "native",
3126
3275
  includedTools: ["apply_patch"],
3127
3276
  excludedTools: ["apply_diff", "write_to_file"],
3128
3277
  supportsImages: true,
@@ -3136,6 +3285,7 @@ var openAiNativeModels = {
3136
3285
  maxTokens: 32768,
3137
3286
  contextWindow: 1047576,
3138
3287
  supportsNativeTools: true,
3288
+ defaultToolProtocol: "native",
3139
3289
  includedTools: ["apply_patch"],
3140
3290
  excludedTools: ["apply_diff", "write_to_file"],
3141
3291
  supportsImages: true,
@@ -3152,6 +3302,7 @@ var openAiNativeModels = {
3152
3302
  maxTokens: 32768,
3153
3303
  contextWindow: 1047576,
3154
3304
  supportsNativeTools: true,
3305
+ defaultToolProtocol: "native",
3155
3306
  includedTools: ["apply_patch"],
3156
3307
  excludedTools: ["apply_diff", "write_to_file"],
3157
3308
  supportsImages: true,
@@ -3168,6 +3319,7 @@ var openAiNativeModels = {
3168
3319
  maxTokens: 32768,
3169
3320
  contextWindow: 1047576,
3170
3321
  supportsNativeTools: true,
3322
+ defaultToolProtocol: "native",
3171
3323
  includedTools: ["apply_patch"],
3172
3324
  excludedTools: ["apply_diff", "write_to_file"],
3173
3325
  supportsImages: true,
@@ -3184,6 +3336,7 @@ var openAiNativeModels = {
3184
3336
  maxTokens: 1e5,
3185
3337
  contextWindow: 2e5,
3186
3338
  supportsNativeTools: true,
3339
+ defaultToolProtocol: "native",
3187
3340
  supportsImages: true,
3188
3341
  supportsPromptCache: true,
3189
3342
  inputPrice: 2,
@@ -3201,6 +3354,7 @@ var openAiNativeModels = {
3201
3354
  maxTokens: 1e5,
3202
3355
  contextWindow: 2e5,
3203
3356
  supportsNativeTools: true,
3357
+ defaultToolProtocol: "native",
3204
3358
  supportsImages: true,
3205
3359
  supportsPromptCache: true,
3206
3360
  inputPrice: 2,
@@ -3213,6 +3367,7 @@ var openAiNativeModels = {
3213
3367
  maxTokens: 1e5,
3214
3368
  contextWindow: 2e5,
3215
3369
  supportsNativeTools: true,
3370
+ defaultToolProtocol: "native",
3216
3371
  supportsImages: true,
3217
3372
  supportsPromptCache: true,
3218
3373
  inputPrice: 2,
@@ -3225,6 +3380,7 @@ var openAiNativeModels = {
3225
3380
  maxTokens: 1e5,
3226
3381
  contextWindow: 2e5,
3227
3382
  supportsNativeTools: true,
3383
+ defaultToolProtocol: "native",
3228
3384
  supportsImages: true,
3229
3385
  supportsPromptCache: true,
3230
3386
  inputPrice: 1.1,
@@ -3242,6 +3398,7 @@ var openAiNativeModels = {
3242
3398
  maxTokens: 1e5,
3243
3399
  contextWindow: 2e5,
3244
3400
  supportsNativeTools: true,
3401
+ defaultToolProtocol: "native",
3245
3402
  supportsImages: true,
3246
3403
  supportsPromptCache: true,
3247
3404
  inputPrice: 1.1,
@@ -3254,6 +3411,7 @@ var openAiNativeModels = {
3254
3411
  maxTokens: 1e5,
3255
3412
  contextWindow: 2e5,
3256
3413
  supportsNativeTools: true,
3414
+ defaultToolProtocol: "native",
3257
3415
  supportsImages: true,
3258
3416
  supportsPromptCache: true,
3259
3417
  inputPrice: 1.1,
@@ -3266,6 +3424,7 @@ var openAiNativeModels = {
3266
3424
  maxTokens: 1e5,
3267
3425
  contextWindow: 2e5,
3268
3426
  supportsNativeTools: true,
3427
+ defaultToolProtocol: "native",
3269
3428
  supportsImages: false,
3270
3429
  supportsPromptCache: true,
3271
3430
  inputPrice: 1.1,
@@ -3279,6 +3438,7 @@ var openAiNativeModels = {
3279
3438
  maxTokens: 1e5,
3280
3439
  contextWindow: 2e5,
3281
3440
  supportsNativeTools: true,
3441
+ defaultToolProtocol: "native",
3282
3442
  supportsImages: false,
3283
3443
  supportsPromptCache: true,
3284
3444
  inputPrice: 1.1,
@@ -3291,6 +3451,7 @@ var openAiNativeModels = {
3291
3451
  maxTokens: 1e5,
3292
3452
  contextWindow: 2e5,
3293
3453
  supportsNativeTools: true,
3454
+ defaultToolProtocol: "native",
3294
3455
  supportsImages: false,
3295
3456
  supportsPromptCache: true,
3296
3457
  inputPrice: 1.1,
@@ -3303,6 +3464,7 @@ var openAiNativeModels = {
3303
3464
  maxTokens: 1e5,
3304
3465
  contextWindow: 2e5,
3305
3466
  supportsNativeTools: true,
3467
+ defaultToolProtocol: "native",
3306
3468
  supportsImages: true,
3307
3469
  supportsPromptCache: true,
3308
3470
  inputPrice: 15,
@@ -3314,6 +3476,7 @@ var openAiNativeModels = {
3314
3476
  maxTokens: 32768,
3315
3477
  contextWindow: 128e3,
3316
3478
  supportsNativeTools: true,
3479
+ defaultToolProtocol: "native",
3317
3480
  supportsImages: true,
3318
3481
  supportsPromptCache: true,
3319
3482
  inputPrice: 15,
@@ -3325,6 +3488,7 @@ var openAiNativeModels = {
3325
3488
  maxTokens: 65536,
3326
3489
  contextWindow: 128e3,
3327
3490
  supportsNativeTools: true,
3491
+ defaultToolProtocol: "native",
3328
3492
  supportsImages: true,
3329
3493
  supportsPromptCache: true,
3330
3494
  inputPrice: 1.1,
@@ -3336,6 +3500,7 @@ var openAiNativeModels = {
3336
3500
  maxTokens: 16384,
3337
3501
  contextWindow: 128e3,
3338
3502
  supportsNativeTools: true,
3503
+ defaultToolProtocol: "native",
3339
3504
  supportsImages: true,
3340
3505
  supportsPromptCache: true,
3341
3506
  inputPrice: 2.5,
@@ -3350,6 +3515,7 @@ var openAiNativeModels = {
3350
3515
  maxTokens: 16384,
3351
3516
  contextWindow: 128e3,
3352
3517
  supportsNativeTools: true,
3518
+ defaultToolProtocol: "native",
3353
3519
  supportsImages: true,
3354
3520
  supportsPromptCache: true,
3355
3521
  inputPrice: 0.15,
@@ -3364,6 +3530,7 @@ var openAiNativeModels = {
3364
3530
  maxTokens: 16384,
3365
3531
  contextWindow: 2e5,
3366
3532
  supportsNativeTools: true,
3533
+ defaultToolProtocol: "native",
3367
3534
  supportsImages: false,
3368
3535
  supportsPromptCache: false,
3369
3536
  inputPrice: 1.5,
@@ -3377,6 +3544,7 @@ var openAiNativeModels = {
3377
3544
  maxTokens: 128e3,
3378
3545
  contextWindow: 4e5,
3379
3546
  supportsNativeTools: true,
3547
+ defaultToolProtocol: "native",
3380
3548
  includedTools: ["apply_patch"],
3381
3549
  excludedTools: ["apply_diff", "write_to_file"],
3382
3550
  supportsImages: true,
@@ -3398,6 +3566,7 @@ var openAiNativeModels = {
3398
3566
  maxTokens: 128e3,
3399
3567
  contextWindow: 4e5,
3400
3568
  supportsNativeTools: true,
3569
+ defaultToolProtocol: "native",
3401
3570
  includedTools: ["apply_patch"],
3402
3571
  excludedTools: ["apply_diff", "write_to_file"],
3403
3572
  supportsImages: true,
@@ -3419,6 +3588,7 @@ var openAiNativeModels = {
3419
3588
  maxTokens: 128e3,
3420
3589
  contextWindow: 4e5,
3421
3590
  supportsNativeTools: true,
3591
+ defaultToolProtocol: "native",
3422
3592
  includedTools: ["apply_patch"],
3423
3593
  excludedTools: ["apply_diff", "write_to_file"],
3424
3594
  supportsImages: true,
@@ -3441,7 +3611,8 @@ var openAiModelInfoSaneDefaults = {
3441
3611
  supportsPromptCache: false,
3442
3612
  inputPrice: 0,
3443
3613
  outputPrice: 0,
3444
- supportsNativeTools: true
3614
+ supportsNativeTools: true,
3615
+ defaultToolProtocol: "native"
3445
3616
  };
3446
3617
  var azureOpenAiDefaultApiVersion = "2024-08-01-preview";
3447
3618
  var OPENAI_NATIVE_DEFAULT_TEMPERATURE = 0;
@@ -3528,6 +3699,8 @@ var qwenCodeModels = {
3528
3699
  contextWindow: 1e6,
3529
3700
  supportsImages: false,
3530
3701
  supportsPromptCache: false,
3702
+ supportsNativeTools: true,
3703
+ defaultToolProtocol: "native",
3531
3704
  inputPrice: 0,
3532
3705
  outputPrice: 0,
3533
3706
  cacheWritesPrice: 0,
@@ -3539,6 +3712,8 @@ var qwenCodeModels = {
3539
3712
  contextWindow: 1e6,
3540
3713
  supportsImages: false,
3541
3714
  supportsPromptCache: false,
3715
+ supportsNativeTools: true,
3716
+ defaultToolProtocol: "native",
3542
3717
  inputPrice: 0,
3543
3718
  outputPrice: 0,
3544
3719
  cacheWritesPrice: 0,
@@ -3554,6 +3729,8 @@ var requestyDefaultModelInfo = {
3554
3729
  contextWindow: 2e5,
3555
3730
  supportsImages: true,
3556
3731
  supportsPromptCache: true,
3732
+ supportsNativeTools: true,
3733
+ defaultToolProtocol: "native",
3557
3734
  inputPrice: 3,
3558
3735
  outputPrice: 15,
3559
3736
  cacheWritesPrice: 3.75,
@@ -3610,6 +3787,7 @@ var sambaNovaModels = {
3610
3787
  supportsImages: false,
3611
3788
  supportsPromptCache: false,
3612
3789
  supportsNativeTools: true,
3790
+ defaultToolProtocol: "native",
3613
3791
  inputPrice: 0.1,
3614
3792
  outputPrice: 0.2,
3615
3793
  description: "Meta Llama 3.1 8B Instruct model with 16K context window."
@@ -3620,6 +3798,7 @@ var sambaNovaModels = {
3620
3798
  supportsImages: false,
3621
3799
  supportsPromptCache: false,
3622
3800
  supportsNativeTools: true,
3801
+ defaultToolProtocol: "native",
3623
3802
  inputPrice: 0.6,
3624
3803
  outputPrice: 1.2,
3625
3804
  description: "Meta Llama 3.3 70B Instruct model with 128K context window."
@@ -3631,6 +3810,7 @@ var sambaNovaModels = {
3631
3810
  supportsPromptCache: false,
3632
3811
  supportsReasoningBudget: true,
3633
3812
  supportsNativeTools: true,
3813
+ defaultToolProtocol: "native",
3634
3814
  inputPrice: 5,
3635
3815
  outputPrice: 7,
3636
3816
  description: "DeepSeek R1 reasoning model with 32K context window."
@@ -3641,6 +3821,7 @@ var sambaNovaModels = {
3641
3821
  supportsImages: false,
3642
3822
  supportsPromptCache: false,
3643
3823
  supportsNativeTools: true,
3824
+ defaultToolProtocol: "native",
3644
3825
  inputPrice: 3,
3645
3826
  outputPrice: 4.5,
3646
3827
  description: "DeepSeek V3 model with 32K context window."
@@ -3651,6 +3832,7 @@ var sambaNovaModels = {
3651
3832
  supportsImages: false,
3652
3833
  supportsPromptCache: false,
3653
3834
  supportsNativeTools: true,
3835
+ defaultToolProtocol: "native",
3654
3836
  inputPrice: 3,
3655
3837
  outputPrice: 4.5,
3656
3838
  description: "DeepSeek V3.1 model with 32K context window."
@@ -3670,6 +3852,7 @@ var sambaNovaModels = {
3670
3852
  supportsImages: true,
3671
3853
  supportsPromptCache: false,
3672
3854
  supportsNativeTools: true,
3855
+ defaultToolProtocol: "native",
3673
3856
  inputPrice: 0.63,
3674
3857
  outputPrice: 1.8,
3675
3858
  description: "Meta Llama 4 Maverick 17B 128E Instruct model with 128K context window."
@@ -3689,6 +3872,7 @@ var sambaNovaModels = {
3689
3872
  supportsImages: false,
3690
3873
  supportsPromptCache: false,
3691
3874
  supportsNativeTools: true,
3875
+ defaultToolProtocol: "native",
3692
3876
  inputPrice: 0.4,
3693
3877
  outputPrice: 0.8,
3694
3878
  description: "Alibaba Qwen 3 32B model with 8K context window."
@@ -3699,6 +3883,7 @@ var sambaNovaModels = {
3699
3883
  supportsImages: false,
3700
3884
  supportsPromptCache: false,
3701
3885
  supportsNativeTools: true,
3886
+ defaultToolProtocol: "native",
3702
3887
  inputPrice: 0.22,
3703
3888
  outputPrice: 0.59,
3704
3889
  description: "OpenAI gpt oss 120b model with 128k context window."
@@ -3727,9 +3912,12 @@ var vertexModels = {
3727
3912
  contextWindow: 1048576,
3728
3913
  supportsImages: true,
3729
3914
  supportsNativeTools: true,
3915
+ defaultToolProtocol: "native",
3730
3916
  supportsPromptCache: true,
3731
3917
  supportsReasoningEffort: ["low", "high"],
3732
3918
  reasoningEffort: "low",
3919
+ includedTools: ["write_file", "edit_file"],
3920
+ excludedTools: ["apply_diff"],
3733
3921
  supportsTemperature: true,
3734
3922
  defaultTemperature: 1,
3735
3923
  inputPrice: 4,
@@ -3747,12 +3935,33 @@ var vertexModels = {
3747
3935
  }
3748
3936
  ]
3749
3937
  },
3938
+ "gemini-3-flash-preview": {
3939
+ maxTokens: 65536,
3940
+ contextWindow: 1048576,
3941
+ supportsImages: true,
3942
+ supportsNativeTools: true,
3943
+ defaultToolProtocol: "native",
3944
+ supportsPromptCache: true,
3945
+ supportsReasoningEffort: ["minimal", "low", "medium", "high"],
3946
+ reasoningEffort: "medium",
3947
+ includedTools: ["write_file", "edit_file"],
3948
+ excludedTools: ["apply_diff"],
3949
+ supportsTemperature: true,
3950
+ defaultTemperature: 1,
3951
+ inputPrice: 0.3,
3952
+ outputPrice: 2.5,
3953
+ cacheReadsPrice: 0.075,
3954
+ cacheWritesPrice: 1
3955
+ },
3750
3956
  "gemini-2.5-flash-preview-05-20:thinking": {
3751
3957
  maxTokens: 65535,
3752
3958
  contextWindow: 1048576,
3753
3959
  supportsImages: true,
3754
3960
  supportsNativeTools: true,
3961
+ defaultToolProtocol: "native",
3755
3962
  supportsPromptCache: true,
3963
+ includedTools: ["write_file", "edit_file"],
3964
+ excludedTools: ["apply_diff"],
3756
3965
  inputPrice: 0.15,
3757
3966
  outputPrice: 3.5,
3758
3967
  maxThinkingTokens: 24576,
@@ -3764,7 +3973,10 @@ var vertexModels = {
3764
3973
  contextWindow: 1048576,
3765
3974
  supportsImages: true,
3766
3975
  supportsNativeTools: true,
3976
+ defaultToolProtocol: "native",
3767
3977
  supportsPromptCache: true,
3978
+ includedTools: ["write_file", "edit_file"],
3979
+ excludedTools: ["apply_diff"],
3768
3980
  inputPrice: 0.15,
3769
3981
  outputPrice: 0.6
3770
3982
  },
@@ -3773,7 +3985,10 @@ var vertexModels = {
3773
3985
  contextWindow: 1048576,
3774
3986
  supportsImages: true,
3775
3987
  supportsNativeTools: true,
3988
+ defaultToolProtocol: "native",
3776
3989
  supportsPromptCache: true,
3990
+ includedTools: ["write_file", "edit_file"],
3991
+ excludedTools: ["apply_diff"],
3777
3992
  inputPrice: 0.3,
3778
3993
  outputPrice: 2.5,
3779
3994
  cacheReadsPrice: 0.075,
@@ -3786,7 +4001,10 @@ var vertexModels = {
3786
4001
  contextWindow: 1048576,
3787
4002
  supportsImages: true,
3788
4003
  supportsNativeTools: true,
4004
+ defaultToolProtocol: "native",
3789
4005
  supportsPromptCache: false,
4006
+ includedTools: ["write_file", "edit_file"],
4007
+ excludedTools: ["apply_diff"],
3790
4008
  inputPrice: 0.15,
3791
4009
  outputPrice: 3.5,
3792
4010
  maxThinkingTokens: 24576,
@@ -3798,7 +4016,10 @@ var vertexModels = {
3798
4016
  contextWindow: 1048576,
3799
4017
  supportsImages: true,
3800
4018
  supportsNativeTools: true,
4019
+ defaultToolProtocol: "native",
3801
4020
  supportsPromptCache: false,
4021
+ includedTools: ["write_file", "edit_file"],
4022
+ excludedTools: ["apply_diff"],
3802
4023
  inputPrice: 0.15,
3803
4024
  outputPrice: 0.6
3804
4025
  },
@@ -3807,7 +4028,10 @@ var vertexModels = {
3807
4028
  contextWindow: 1048576,
3808
4029
  supportsImages: true,
3809
4030
  supportsNativeTools: true,
4031
+ defaultToolProtocol: "native",
3810
4032
  supportsPromptCache: true,
4033
+ includedTools: ["write_file", "edit_file"],
4034
+ excludedTools: ["apply_diff"],
3811
4035
  inputPrice: 2.5,
3812
4036
  outputPrice: 15
3813
4037
  },
@@ -3816,7 +4040,10 @@ var vertexModels = {
3816
4040
  contextWindow: 1048576,
3817
4041
  supportsImages: true,
3818
4042
  supportsNativeTools: true,
4043
+ defaultToolProtocol: "native",
3819
4044
  supportsPromptCache: true,
4045
+ includedTools: ["write_file", "edit_file"],
4046
+ excludedTools: ["apply_diff"],
3820
4047
  inputPrice: 2.5,
3821
4048
  outputPrice: 15
3822
4049
  },
@@ -3825,7 +4052,10 @@ var vertexModels = {
3825
4052
  contextWindow: 1048576,
3826
4053
  supportsImages: true,
3827
4054
  supportsNativeTools: true,
4055
+ defaultToolProtocol: "native",
3828
4056
  supportsPromptCache: true,
4057
+ includedTools: ["write_file", "edit_file"],
4058
+ excludedTools: ["apply_diff"],
3829
4059
  inputPrice: 2.5,
3830
4060
  outputPrice: 15,
3831
4061
  maxThinkingTokens: 32768,
@@ -3836,7 +4066,10 @@ var vertexModels = {
3836
4066
  contextWindow: 1048576,
3837
4067
  supportsImages: true,
3838
4068
  supportsNativeTools: true,
4069
+ defaultToolProtocol: "native",
3839
4070
  supportsPromptCache: true,
4071
+ includedTools: ["write_file", "edit_file"],
4072
+ excludedTools: ["apply_diff"],
3840
4073
  inputPrice: 2.5,
3841
4074
  outputPrice: 15,
3842
4075
  maxThinkingTokens: 32768,
@@ -3862,7 +4095,10 @@ var vertexModels = {
3862
4095
  contextWindow: 1048576,
3863
4096
  supportsImages: true,
3864
4097
  supportsNativeTools: true,
4098
+ defaultToolProtocol: "native",
3865
4099
  supportsPromptCache: false,
4100
+ includedTools: ["write_file", "edit_file"],
4101
+ excludedTools: ["apply_diff"],
3866
4102
  inputPrice: 0,
3867
4103
  outputPrice: 0
3868
4104
  },
@@ -3871,7 +4107,10 @@ var vertexModels = {
3871
4107
  contextWindow: 2097152,
3872
4108
  supportsImages: true,
3873
4109
  supportsNativeTools: true,
4110
+ defaultToolProtocol: "native",
3874
4111
  supportsPromptCache: false,
4112
+ includedTools: ["write_file", "edit_file"],
4113
+ excludedTools: ["apply_diff"],
3875
4114
  inputPrice: 0,
3876
4115
  outputPrice: 0
3877
4116
  },
@@ -3880,7 +4119,10 @@ var vertexModels = {
3880
4119
  contextWindow: 1048576,
3881
4120
  supportsImages: true,
3882
4121
  supportsNativeTools: true,
4122
+ defaultToolProtocol: "native",
3883
4123
  supportsPromptCache: true,
4124
+ includedTools: ["write_file", "edit_file"],
4125
+ excludedTools: ["apply_diff"],
3884
4126
  inputPrice: 0.15,
3885
4127
  outputPrice: 0.6
3886
4128
  },
@@ -3889,7 +4131,10 @@ var vertexModels = {
3889
4131
  contextWindow: 1048576,
3890
4132
  supportsImages: true,
3891
4133
  supportsNativeTools: true,
4134
+ defaultToolProtocol: "native",
3892
4135
  supportsPromptCache: false,
4136
+ includedTools: ["write_file", "edit_file"],
4137
+ excludedTools: ["apply_diff"],
3893
4138
  inputPrice: 0.075,
3894
4139
  outputPrice: 0.3
3895
4140
  },
@@ -3898,7 +4143,10 @@ var vertexModels = {
3898
4143
  contextWindow: 32768,
3899
4144
  supportsImages: true,
3900
4145
  supportsNativeTools: true,
4146
+ defaultToolProtocol: "native",
3901
4147
  supportsPromptCache: false,
4148
+ includedTools: ["write_file", "edit_file"],
4149
+ excludedTools: ["apply_diff"],
3902
4150
  inputPrice: 0,
3903
4151
  outputPrice: 0
3904
4152
  },
@@ -3907,7 +4155,10 @@ var vertexModels = {
3907
4155
  contextWindow: 1048576,
3908
4156
  supportsImages: true,
3909
4157
  supportsNativeTools: true,
4158
+ defaultToolProtocol: "native",
3910
4159
  supportsPromptCache: true,
4160
+ includedTools: ["write_file", "edit_file"],
4161
+ excludedTools: ["apply_diff"],
3911
4162
  inputPrice: 0.075,
3912
4163
  outputPrice: 0.3
3913
4164
  },
@@ -3916,37 +4167,86 @@ var vertexModels = {
3916
4167
  contextWindow: 2097152,
3917
4168
  supportsImages: true,
3918
4169
  supportsNativeTools: true,
4170
+ defaultToolProtocol: "native",
3919
4171
  supportsPromptCache: false,
4172
+ includedTools: ["write_file", "edit_file"],
4173
+ excludedTools: ["apply_diff"],
3920
4174
  inputPrice: 1.25,
3921
4175
  outputPrice: 5
3922
4176
  },
3923
4177
  "claude-sonnet-4@20250514": {
3924
4178
  maxTokens: 8192,
3925
4179
  contextWindow: 2e5,
4180
+ // Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
3926
4181
  supportsImages: true,
3927
4182
  supportsPromptCache: true,
4183
+ supportsNativeTools: true,
4184
+ defaultToolProtocol: "native",
3928
4185
  inputPrice: 3,
4186
+ // $3 per million input tokens (≤200K context)
3929
4187
  outputPrice: 15,
4188
+ // $15 per million output tokens (≤200K context)
3930
4189
  cacheWritesPrice: 3.75,
4190
+ // $3.75 per million tokens
3931
4191
  cacheReadsPrice: 0.3,
3932
- supportsReasoningBudget: true
4192
+ // $0.30 per million tokens
4193
+ supportsReasoningBudget: true,
4194
+ // Tiered pricing for extended context (requires beta flag 'context-1m-2025-08-07')
4195
+ tiers: [
4196
+ {
4197
+ contextWindow: 1e6,
4198
+ // 1M tokens with beta flag
4199
+ inputPrice: 6,
4200
+ // $6 per million input tokens (>200K context)
4201
+ outputPrice: 22.5,
4202
+ // $22.50 per million output tokens (>200K context)
4203
+ cacheWritesPrice: 7.5,
4204
+ // $7.50 per million tokens (>200K context)
4205
+ cacheReadsPrice: 0.6
4206
+ // $0.60 per million tokens (>200K context)
4207
+ }
4208
+ ]
3933
4209
  },
3934
4210
  "claude-sonnet-4-5@20250929": {
3935
4211
  maxTokens: 8192,
3936
4212
  contextWindow: 2e5,
4213
+ // Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
3937
4214
  supportsImages: true,
3938
4215
  supportsPromptCache: true,
4216
+ supportsNativeTools: true,
4217
+ defaultToolProtocol: "native",
3939
4218
  inputPrice: 3,
4219
+ // $3 per million input tokens (≤200K context)
3940
4220
  outputPrice: 15,
4221
+ // $15 per million output tokens (≤200K context)
3941
4222
  cacheWritesPrice: 3.75,
4223
+ // $3.75 per million tokens
3942
4224
  cacheReadsPrice: 0.3,
3943
- supportsReasoningBudget: true
4225
+ // $0.30 per million tokens
4226
+ supportsReasoningBudget: true,
4227
+ // Tiered pricing for extended context (requires beta flag 'context-1m-2025-08-07')
4228
+ tiers: [
4229
+ {
4230
+ contextWindow: 1e6,
4231
+ // 1M tokens with beta flag
4232
+ inputPrice: 6,
4233
+ // $6 per million input tokens (>200K context)
4234
+ outputPrice: 22.5,
4235
+ // $22.50 per million output tokens (>200K context)
4236
+ cacheWritesPrice: 7.5,
4237
+ // $7.50 per million tokens (>200K context)
4238
+ cacheReadsPrice: 0.6
4239
+ // $0.60 per million tokens (>200K context)
4240
+ }
4241
+ ]
3944
4242
  },
3945
4243
  "claude-haiku-4-5@20251001": {
3946
4244
  maxTokens: 8192,
3947
4245
  contextWindow: 2e5,
3948
4246
  supportsImages: true,
3949
4247
  supportsPromptCache: true,
4248
+ supportsNativeTools: true,
4249
+ defaultToolProtocol: "native",
3950
4250
  inputPrice: 1,
3951
4251
  outputPrice: 5,
3952
4252
  cacheWritesPrice: 1.25,
@@ -3958,6 +4258,8 @@ var vertexModels = {
3958
4258
  contextWindow: 2e5,
3959
4259
  supportsImages: true,
3960
4260
  supportsPromptCache: true,
4261
+ supportsNativeTools: true,
4262
+ defaultToolProtocol: "native",
3961
4263
  inputPrice: 5,
3962
4264
  outputPrice: 25,
3963
4265
  cacheWritesPrice: 6.25,
@@ -3969,6 +4271,8 @@ var vertexModels = {
3969
4271
  contextWindow: 2e5,
3970
4272
  supportsImages: true,
3971
4273
  supportsPromptCache: true,
4274
+ supportsNativeTools: true,
4275
+ defaultToolProtocol: "native",
3972
4276
  inputPrice: 15,
3973
4277
  outputPrice: 75,
3974
4278
  cacheWritesPrice: 18.75,
@@ -3980,6 +4284,8 @@ var vertexModels = {
3980
4284
  contextWindow: 2e5,
3981
4285
  supportsImages: true,
3982
4286
  supportsPromptCache: true,
4287
+ supportsNativeTools: true,
4288
+ defaultToolProtocol: "native",
3983
4289
  inputPrice: 15,
3984
4290
  outputPrice: 75,
3985
4291
  cacheWritesPrice: 18.75,
@@ -3990,6 +4296,8 @@ var vertexModels = {
3990
4296
  contextWindow: 2e5,
3991
4297
  supportsImages: true,
3992
4298
  supportsPromptCache: true,
4299
+ supportsNativeTools: true,
4300
+ defaultToolProtocol: "native",
3993
4301
  inputPrice: 3,
3994
4302
  outputPrice: 15,
3995
4303
  cacheWritesPrice: 3.75,
@@ -4002,6 +4310,8 @@ var vertexModels = {
4002
4310
  contextWindow: 2e5,
4003
4311
  supportsImages: true,
4004
4312
  supportsPromptCache: true,
4313
+ supportsNativeTools: true,
4314
+ defaultToolProtocol: "native",
4005
4315
  inputPrice: 3,
4006
4316
  outputPrice: 15,
4007
4317
  cacheWritesPrice: 3.75,
@@ -4012,6 +4322,8 @@ var vertexModels = {
4012
4322
  contextWindow: 2e5,
4013
4323
  supportsImages: true,
4014
4324
  supportsPromptCache: true,
4325
+ supportsNativeTools: true,
4326
+ defaultToolProtocol: "native",
4015
4327
  inputPrice: 3,
4016
4328
  outputPrice: 15,
4017
4329
  cacheWritesPrice: 3.75,
@@ -4022,6 +4334,8 @@ var vertexModels = {
4022
4334
  contextWindow: 2e5,
4023
4335
  supportsImages: true,
4024
4336
  supportsPromptCache: true,
4337
+ supportsNativeTools: true,
4338
+ defaultToolProtocol: "native",
4025
4339
  inputPrice: 3,
4026
4340
  outputPrice: 15,
4027
4341
  cacheWritesPrice: 3.75,
@@ -4032,6 +4346,8 @@ var vertexModels = {
4032
4346
  contextWindow: 2e5,
4033
4347
  supportsImages: false,
4034
4348
  supportsPromptCache: true,
4349
+ supportsNativeTools: true,
4350
+ defaultToolProtocol: "native",
4035
4351
  inputPrice: 1,
4036
4352
  outputPrice: 5,
4037
4353
  cacheWritesPrice: 1.25,
@@ -4042,6 +4358,8 @@ var vertexModels = {
4042
4358
  contextWindow: 2e5,
4043
4359
  supportsImages: true,
4044
4360
  supportsPromptCache: true,
4361
+ supportsNativeTools: true,
4362
+ defaultToolProtocol: "native",
4045
4363
  inputPrice: 15,
4046
4364
  outputPrice: 75,
4047
4365
  cacheWritesPrice: 18.75,
@@ -4052,6 +4370,8 @@ var vertexModels = {
4052
4370
  contextWindow: 2e5,
4053
4371
  supportsImages: true,
4054
4372
  supportsPromptCache: true,
4373
+ supportsNativeTools: true,
4374
+ defaultToolProtocol: "native",
4055
4375
  inputPrice: 0.25,
4056
4376
  outputPrice: 1.25,
4057
4377
  cacheWritesPrice: 0.3,
@@ -4062,7 +4382,10 @@ var vertexModels = {
4062
4382
  contextWindow: 1048576,
4063
4383
  supportsImages: true,
4064
4384
  supportsNativeTools: true,
4385
+ defaultToolProtocol: "native",
4065
4386
  supportsPromptCache: true,
4387
+ includedTools: ["write_file", "edit_file"],
4388
+ excludedTools: ["apply_diff"],
4066
4389
  inputPrice: 0.1,
4067
4390
  outputPrice: 0.4,
4068
4391
  cacheReadsPrice: 0.025,
@@ -4134,6 +4457,7 @@ var vertexModels = {
4134
4457
  description: "Qwen3 235B A22B Instruct. Available in us-south1"
4135
4458
  }
4136
4459
  };
4460
+ var VERTEX_1M_CONTEXT_MODEL_IDS = ["claude-sonnet-4@20250514", "claude-sonnet-4-5@20250929"];
4137
4461
  var VERTEX_REGIONS = [
4138
4462
  { value: "global", label: "global" },
4139
4463
  { value: "us-central1", label: "us-central1" },
@@ -4364,6 +4688,7 @@ var xaiModels = {
4364
4688
  supportsImages: true,
4365
4689
  supportsPromptCache: true,
4366
4690
  supportsNativeTools: true,
4691
+ defaultToolProtocol: "native",
4367
4692
  inputPrice: 0.2,
4368
4693
  outputPrice: 1.5,
4369
4694
  cacheWritesPrice: 0.02,
@@ -4378,6 +4703,7 @@ var xaiModels = {
4378
4703
  supportsImages: true,
4379
4704
  supportsPromptCache: true,
4380
4705
  supportsNativeTools: true,
4706
+ defaultToolProtocol: "native",
4381
4707
  inputPrice: 0.2,
4382
4708
  outputPrice: 0.5,
4383
4709
  cacheWritesPrice: 0.05,
@@ -4392,6 +4718,7 @@ var xaiModels = {
4392
4718
  supportsImages: true,
4393
4719
  supportsPromptCache: true,
4394
4720
  supportsNativeTools: true,
4721
+ defaultToolProtocol: "native",
4395
4722
  inputPrice: 0.2,
4396
4723
  outputPrice: 0.5,
4397
4724
  cacheWritesPrice: 0.05,
@@ -4406,6 +4733,7 @@ var xaiModels = {
4406
4733
  supportsImages: true,
4407
4734
  supportsPromptCache: true,
4408
4735
  supportsNativeTools: true,
4736
+ defaultToolProtocol: "native",
4409
4737
  inputPrice: 0.2,
4410
4738
  outputPrice: 0.5,
4411
4739
  cacheWritesPrice: 0.05,
@@ -4420,6 +4748,7 @@ var xaiModels = {
4420
4748
  supportsImages: true,
4421
4749
  supportsPromptCache: true,
4422
4750
  supportsNativeTools: true,
4751
+ defaultToolProtocol: "native",
4423
4752
  inputPrice: 0.2,
4424
4753
  outputPrice: 0.5,
4425
4754
  cacheWritesPrice: 0.05,
@@ -4434,6 +4763,7 @@ var xaiModels = {
4434
4763
  supportsImages: true,
4435
4764
  supportsPromptCache: true,
4436
4765
  supportsNativeTools: true,
4766
+ defaultToolProtocol: "native",
4437
4767
  inputPrice: 3,
4438
4768
  outputPrice: 15,
4439
4769
  cacheWritesPrice: 0.75,
@@ -4448,6 +4778,7 @@ var xaiModels = {
4448
4778
  supportsImages: true,
4449
4779
  supportsPromptCache: true,
4450
4780
  supportsNativeTools: true,
4781
+ defaultToolProtocol: "native",
4451
4782
  inputPrice: 0.3,
4452
4783
  outputPrice: 0.5,
4453
4784
  cacheWritesPrice: 0.07,
@@ -4464,6 +4795,7 @@ var xaiModels = {
4464
4795
  supportsImages: true,
4465
4796
  supportsPromptCache: true,
4466
4797
  supportsNativeTools: true,
4798
+ defaultToolProtocol: "native",
4467
4799
  inputPrice: 3,
4468
4800
  outputPrice: 15,
4469
4801
  cacheWritesPrice: 0.75,
@@ -4577,6 +4909,7 @@ var internationalZAiModels = {
4577
4909
  supportsImages: false,
4578
4910
  supportsPromptCache: true,
4579
4911
  supportsNativeTools: true,
4912
+ defaultToolProtocol: "native",
4580
4913
  inputPrice: 0.6,
4581
4914
  outputPrice: 2.2,
4582
4915
  cacheWritesPrice: 0,
@@ -4589,6 +4922,7 @@ var internationalZAiModels = {
4589
4922
  supportsImages: false,
4590
4923
  supportsPromptCache: true,
4591
4924
  supportsNativeTools: true,
4925
+ defaultToolProtocol: "native",
4592
4926
  inputPrice: 0.2,
4593
4927
  outputPrice: 1.1,
4594
4928
  cacheWritesPrice: 0,
@@ -4601,6 +4935,7 @@ var internationalZAiModels = {
4601
4935
  supportsImages: false,
4602
4936
  supportsPromptCache: true,
4603
4937
  supportsNativeTools: true,
4938
+ defaultToolProtocol: "native",
4604
4939
  inputPrice: 2.2,
4605
4940
  outputPrice: 8.9,
4606
4941
  cacheWritesPrice: 0,
@@ -4613,6 +4948,7 @@ var internationalZAiModels = {
4613
4948
  supportsImages: false,
4614
4949
  supportsPromptCache: true,
4615
4950
  supportsNativeTools: true,
4951
+ defaultToolProtocol: "native",
4616
4952
  inputPrice: 1.1,
4617
4953
  outputPrice: 4.5,
4618
4954
  cacheWritesPrice: 0,
@@ -4625,6 +4961,7 @@ var internationalZAiModels = {
4625
4961
  supportsImages: false,
4626
4962
  supportsPromptCache: true,
4627
4963
  supportsNativeTools: true,
4964
+ defaultToolProtocol: "native",
4628
4965
  inputPrice: 0,
4629
4966
  outputPrice: 0,
4630
4967
  cacheWritesPrice: 0,
@@ -4637,6 +4974,7 @@ var internationalZAiModels = {
4637
4974
  supportsImages: true,
4638
4975
  supportsPromptCache: true,
4639
4976
  supportsNativeTools: true,
4977
+ defaultToolProtocol: "native",
4640
4978
  inputPrice: 0.6,
4641
4979
  outputPrice: 1.8,
4642
4980
  cacheWritesPrice: 0,
@@ -4649,6 +4987,7 @@ var internationalZAiModels = {
4649
4987
  supportsImages: false,
4650
4988
  supportsPromptCache: true,
4651
4989
  supportsNativeTools: true,
4990
+ defaultToolProtocol: "native",
4652
4991
  inputPrice: 0.6,
4653
4992
  outputPrice: 2.2,
4654
4993
  cacheWritesPrice: 0,
@@ -4661,6 +5000,7 @@ var internationalZAiModels = {
4661
5000
  supportsImages: false,
4662
5001
  supportsPromptCache: false,
4663
5002
  supportsNativeTools: true,
5003
+ defaultToolProtocol: "native",
4664
5004
  inputPrice: 0.1,
4665
5005
  outputPrice: 0.1,
4666
5006
  cacheWritesPrice: 0,
@@ -4676,6 +5016,7 @@ var mainlandZAiModels = {
4676
5016
  supportsImages: false,
4677
5017
  supportsPromptCache: true,
4678
5018
  supportsNativeTools: true,
5019
+ defaultToolProtocol: "native",
4679
5020
  inputPrice: 0.29,
4680
5021
  outputPrice: 1.14,
4681
5022
  cacheWritesPrice: 0,
@@ -4688,6 +5029,7 @@ var mainlandZAiModels = {
4688
5029
  supportsImages: false,
4689
5030
  supportsPromptCache: true,
4690
5031
  supportsNativeTools: true,
5032
+ defaultToolProtocol: "native",
4691
5033
  inputPrice: 0.1,
4692
5034
  outputPrice: 0.6,
4693
5035
  cacheWritesPrice: 0,
@@ -4700,6 +5042,7 @@ var mainlandZAiModels = {
4700
5042
  supportsImages: false,
4701
5043
  supportsPromptCache: true,
4702
5044
  supportsNativeTools: true,
5045
+ defaultToolProtocol: "native",
4703
5046
  inputPrice: 0.29,
4704
5047
  outputPrice: 1.14,
4705
5048
  cacheWritesPrice: 0,
@@ -4712,6 +5055,7 @@ var mainlandZAiModels = {
4712
5055
  supportsImages: false,
4713
5056
  supportsPromptCache: true,
4714
5057
  supportsNativeTools: true,
5058
+ defaultToolProtocol: "native",
4715
5059
  inputPrice: 0.1,
4716
5060
  outputPrice: 0.6,
4717
5061
  cacheWritesPrice: 0,
@@ -4724,6 +5068,7 @@ var mainlandZAiModels = {
4724
5068
  supportsImages: false,
4725
5069
  supportsPromptCache: true,
4726
5070
  supportsNativeTools: true,
5071
+ defaultToolProtocol: "native",
4727
5072
  inputPrice: 0,
4728
5073
  outputPrice: 0,
4729
5074
  cacheWritesPrice: 0,
@@ -4736,6 +5081,7 @@ var mainlandZAiModels = {
4736
5081
  supportsImages: true,
4737
5082
  supportsPromptCache: true,
4738
5083
  supportsNativeTools: true,
5084
+ defaultToolProtocol: "native",
4739
5085
  inputPrice: 0.29,
4740
5086
  outputPrice: 0.93,
4741
5087
  cacheWritesPrice: 0,
@@ -4748,6 +5094,7 @@ var mainlandZAiModels = {
4748
5094
  supportsImages: false,
4749
5095
  supportsPromptCache: true,
4750
5096
  supportsNativeTools: true,
5097
+ defaultToolProtocol: "native",
4751
5098
  inputPrice: 0.29,
4752
5099
  outputPrice: 1.14,
4753
5100
  cacheWritesPrice: 0,
@@ -4801,6 +5148,7 @@ var minimaxModels = {
4801
5148
  supportsImages: false,
4802
5149
  supportsPromptCache: true,
4803
5150
  supportsNativeTools: true,
5151
+ defaultToolProtocol: "native",
4804
5152
  preserveReasoning: true,
4805
5153
  inputPrice: 0.3,
4806
5154
  outputPrice: 1.2,
@@ -4814,6 +5162,7 @@ var minimaxModels = {
4814
5162
  supportsImages: false,
4815
5163
  supportsPromptCache: true,
4816
5164
  supportsNativeTools: true,
5165
+ defaultToolProtocol: "native",
4817
5166
  preserveReasoning: true,
4818
5167
  inputPrice: 0.3,
4819
5168
  outputPrice: 1.2,
@@ -4996,10 +5345,7 @@ var anthropicSchema = apiModelIdProviderModelSchema.extend({
4996
5345
  anthropicBeta1MContext: z8.boolean().optional()
4997
5346
  // Enable 'context-1m-2025-08-07' beta for 1M context window.
4998
5347
  });
4999
- var claudeCodeSchema = apiModelIdProviderModelSchema.extend({
5000
- claudeCodePath: z8.string().optional(),
5001
- claudeCodeMaxOutputTokens: z8.number().int().min(1).max(2e5).optional()
5002
- });
5348
+ var claudeCodeSchema = apiModelIdProviderModelSchema.extend({});
5003
5349
  var openRouterSchema = baseProviderSettingsSchema.extend({
5004
5350
  openRouterApiKey: z8.string().optional(),
5005
5351
  openRouterModelId: z8.string().optional(),
@@ -5035,7 +5381,9 @@ var vertexSchema = apiModelIdProviderModelSchema.extend({
5035
5381
  vertexProjectId: z8.string().optional(),
5036
5382
  vertexRegion: z8.string().optional(),
5037
5383
  enableUrlContext: z8.boolean().optional(),
5038
- enableGrounding: z8.boolean().optional()
5384
+ enableGrounding: z8.boolean().optional(),
5385
+ vertex1MContext: z8.boolean().optional()
5386
+ // Enable 'context-1m-2025-08-07' beta for 1M context window.
5039
5387
  });
5040
5388
  var openAiSchema = baseProviderSettingsSchema.extend({
5041
5389
  openAiBaseUrl: z8.string().optional(),
@@ -5452,6 +5800,16 @@ var historyItemSchema = z9.object({
5452
5800
  size: z9.number().optional(),
5453
5801
  workspace: z9.string().optional(),
5454
5802
  mode: z9.string().optional(),
5803
+ /**
5804
+ * The tool protocol used by this task. Once a task uses tools with a specific
5805
+ * protocol (XML or Native), it is permanently locked to that protocol.
5806
+ *
5807
+ * - "xml": Tool calls are parsed from XML text (no tool IDs)
5808
+ * - "native": Tool calls come as tool_call chunks with IDs
5809
+ *
5810
+ * This ensures task resumption works correctly even when NTC settings change.
5811
+ */
5812
+ toolProtocol: z9.enum(["xml", "native"]).optional(),
5455
5813
  status: z9.enum(["active", "completed", "delegated"]).optional(),
5456
5814
  delegatedToId: z9.string().optional(),
5457
5815
  // Last child this parent delegated to
@@ -5674,11 +6032,44 @@ function getErrorStatusCode(error) {
5674
6032
  }
5675
6033
  return void 0;
5676
6034
  }
6035
+ function extractMessageFromJsonPayload(message) {
6036
+ const jsonStartIndex = message.indexOf("{");
6037
+ if (jsonStartIndex === -1) {
6038
+ return void 0;
6039
+ }
6040
+ const potentialJson = message.slice(jsonStartIndex);
6041
+ try {
6042
+ const parsed = JSON.parse(potentialJson);
6043
+ if (parsed?.error?.message && typeof parsed.error.message === "string") {
6044
+ return parsed.error.message;
6045
+ }
6046
+ if (parsed?.message && typeof parsed.message === "string") {
6047
+ return parsed.message;
6048
+ }
6049
+ } catch {
6050
+ }
6051
+ return void 0;
6052
+ }
5677
6053
  function getErrorMessage(error) {
6054
+ let message;
5678
6055
  if (isOpenAISdkError(error)) {
5679
- return error.error?.metadata?.raw || error.error?.message || error.message;
6056
+ message = error.error?.metadata?.raw || error.error?.message || error.message;
6057
+ } else if (error instanceof Error) {
6058
+ message = error.message;
6059
+ } else if (typeof error === "object" && error !== null && "message" in error) {
6060
+ const msgValue = error.message;
6061
+ if (typeof msgValue === "string") {
6062
+ message = msgValue;
6063
+ }
5680
6064
  }
5681
- return void 0;
6065
+ if (!message) {
6066
+ return void 0;
6067
+ }
6068
+ const extractedMessage = extractMessageFromJsonPayload(message);
6069
+ if (extractedMessage) {
6070
+ return extractedMessage;
6071
+ }
6072
+ return message;
5682
6073
  }
5683
6074
  function shouldReportApiErrorToTelemetry(errorCode, errorMessage) {
5684
6075
  if (errorCode !== void 0 && EXPECTED_API_ERROR_CODES.has(errorCode)) {
@@ -5714,6 +6105,31 @@ function extractApiProviderErrorProperties(error) {
5714
6105
  ...error.errorCode !== void 0 && { errorCode: error.errorCode }
5715
6106
  };
5716
6107
  }
6108
+ var ConsecutiveMistakeError = class extends Error {
6109
+ constructor(message, taskId, consecutiveMistakeCount, consecutiveMistakeLimit, reason = "unknown", provider, modelId) {
6110
+ super(message);
6111
+ this.taskId = taskId;
6112
+ this.consecutiveMistakeCount = consecutiveMistakeCount;
6113
+ this.consecutiveMistakeLimit = consecutiveMistakeLimit;
6114
+ this.reason = reason;
6115
+ this.provider = provider;
6116
+ this.modelId = modelId;
6117
+ this.name = "ConsecutiveMistakeError";
6118
+ }
6119
+ };
6120
+ function isConsecutiveMistakeError(error) {
6121
+ return error instanceof Error && error.name === "ConsecutiveMistakeError" && "taskId" in error && "consecutiveMistakeCount" in error && "consecutiveMistakeLimit" in error;
6122
+ }
6123
+ function extractConsecutiveMistakeErrorProperties(error) {
6124
+ return {
6125
+ taskId: error.taskId,
6126
+ consecutiveMistakeCount: error.consecutiveMistakeCount,
6127
+ consecutiveMistakeLimit: error.consecutiveMistakeLimit,
6128
+ reason: error.reason,
6129
+ ...error.provider !== void 0 && { provider: error.provider },
6130
+ ...error.modelId !== void 0 && { modelId: error.modelId }
6131
+ };
6132
+ }
5717
6133
 
5718
6134
  // src/mode.ts
5719
6135
  import { z as z12 } from "zod";
@@ -6210,7 +6626,7 @@ var organizationDefaultSettingsSchema = globalSettingsSchema.pick({
6210
6626
  terminalShellIntegrationTimeout: z16.number().int().nonnegative().optional()
6211
6627
  })
6212
6628
  );
6213
- var workspaceTaskVisibilitySchema = z16.enum(["all", "list-only", "full-lockdown"]);
6629
+ var workspaceTaskVisibilitySchema = z16.enum(["all", "list-only", "admins-and-creator", "creator", "full-lockdown"]);
6214
6630
  var organizationCloudSettingsSchema = z16.object({
6215
6631
  recordTaskMessages: z16.boolean().optional(),
6216
6632
  enableTaskSharing: z16.boolean().optional(),
@@ -6598,15 +7014,21 @@ var COOKIE_CONSENT_EVENTS = {
6598
7014
  CHANGED: "cookieConsentChanged"
6599
7015
  };
6600
7016
 
7017
+ // src/custom-tool.ts
7018
+ import { z as z17 } from "zod/v4";
7019
+ function defineCustomTool(definition) {
7020
+ return definition;
7021
+ }
7022
+
6601
7023
  // src/followup.ts
6602
- import { z as z17 } from "zod";
6603
- var suggestionItemSchema = z17.object({
6604
- answer: z17.string(),
6605
- mode: z17.string().optional()
7024
+ import { z as z18 } from "zod";
7025
+ var suggestionItemSchema = z18.object({
7026
+ answer: z18.string(),
7027
+ mode: z18.string().optional()
6606
7028
  });
6607
- var followUpDataSchema = z17.object({
6608
- question: z17.string().optional(),
6609
- suggest: z17.array(suggestionItemSchema).optional()
7029
+ var followUpDataSchema = z18.object({
7030
+ question: z18.string().optional(),
7031
+ suggest: z18.array(suggestionItemSchema).optional()
6610
7032
  });
6611
7033
 
6612
7034
  // src/image-generation.ts
@@ -6634,7 +7056,7 @@ function getImageGenerationProvider(explicitProvider, hasExistingModel) {
6634
7056
  }
6635
7057
 
6636
7058
  // src/ipc.ts
6637
- import { z as z18 } from "zod";
7059
+ import { z as z19 } from "zod";
6638
7060
  var IpcMessageType = /* @__PURE__ */ ((IpcMessageType2) => {
6639
7061
  IpcMessageType2["Connect"] = "Connect";
6640
7062
  IpcMessageType2["Disconnect"] = "Disconnect";
@@ -6648,10 +7070,10 @@ var IpcOrigin = /* @__PURE__ */ ((IpcOrigin2) => {
6648
7070
  IpcOrigin2["Server"] = "server";
6649
7071
  return IpcOrigin2;
6650
7072
  })(IpcOrigin || {});
6651
- var ackSchema = z18.object({
6652
- clientId: z18.string(),
6653
- pid: z18.number(),
6654
- ppid: z18.number()
7073
+ var ackSchema = z19.object({
7074
+ clientId: z19.string(),
7075
+ pid: z19.number(),
7076
+ ppid: z19.number()
6655
7077
  });
6656
7078
  var TaskCommandName = /* @__PURE__ */ ((TaskCommandName2) => {
6657
7079
  TaskCommandName2["StartNewTask"] = "StartNewTask";
@@ -6661,79 +7083,79 @@ var TaskCommandName = /* @__PURE__ */ ((TaskCommandName2) => {
6661
7083
  TaskCommandName2["SendMessage"] = "SendMessage";
6662
7084
  return TaskCommandName2;
6663
7085
  })(TaskCommandName || {});
6664
- var taskCommandSchema = z18.discriminatedUnion("commandName", [
6665
- z18.object({
6666
- commandName: z18.literal("StartNewTask" /* StartNewTask */),
6667
- data: z18.object({
7086
+ var taskCommandSchema = z19.discriminatedUnion("commandName", [
7087
+ z19.object({
7088
+ commandName: z19.literal("StartNewTask" /* StartNewTask */),
7089
+ data: z19.object({
6668
7090
  configuration: rooCodeSettingsSchema,
6669
- text: z18.string(),
6670
- images: z18.array(z18.string()).optional(),
6671
- newTab: z18.boolean().optional()
7091
+ text: z19.string(),
7092
+ images: z19.array(z19.string()).optional(),
7093
+ newTab: z19.boolean().optional()
6672
7094
  })
6673
7095
  }),
6674
- z18.object({
6675
- commandName: z18.literal("CancelTask" /* CancelTask */),
6676
- data: z18.string()
7096
+ z19.object({
7097
+ commandName: z19.literal("CancelTask" /* CancelTask */),
7098
+ data: z19.string()
6677
7099
  }),
6678
- z18.object({
6679
- commandName: z18.literal("CloseTask" /* CloseTask */),
6680
- data: z18.string()
7100
+ z19.object({
7101
+ commandName: z19.literal("CloseTask" /* CloseTask */),
7102
+ data: z19.string()
6681
7103
  }),
6682
- z18.object({
6683
- commandName: z18.literal("ResumeTask" /* ResumeTask */),
6684
- data: z18.string()
7104
+ z19.object({
7105
+ commandName: z19.literal("ResumeTask" /* ResumeTask */),
7106
+ data: z19.string()
6685
7107
  }),
6686
- z18.object({
6687
- commandName: z18.literal("SendMessage" /* SendMessage */),
6688
- data: z18.object({
6689
- text: z18.string().optional(),
6690
- images: z18.array(z18.string()).optional()
7108
+ z19.object({
7109
+ commandName: z19.literal("SendMessage" /* SendMessage */),
7110
+ data: z19.object({
7111
+ text: z19.string().optional(),
7112
+ images: z19.array(z19.string()).optional()
6691
7113
  })
6692
7114
  })
6693
7115
  ]);
6694
- var ipcMessageSchema = z18.discriminatedUnion("type", [
6695
- z18.object({
6696
- type: z18.literal("Ack" /* Ack */),
6697
- origin: z18.literal("server" /* Server */),
7116
+ var ipcMessageSchema = z19.discriminatedUnion("type", [
7117
+ z19.object({
7118
+ type: z19.literal("Ack" /* Ack */),
7119
+ origin: z19.literal("server" /* Server */),
6698
7120
  data: ackSchema
6699
7121
  }),
6700
- z18.object({
6701
- type: z18.literal("TaskCommand" /* TaskCommand */),
6702
- origin: z18.literal("client" /* Client */),
6703
- clientId: z18.string(),
7122
+ z19.object({
7123
+ type: z19.literal("TaskCommand" /* TaskCommand */),
7124
+ origin: z19.literal("client" /* Client */),
7125
+ clientId: z19.string(),
6704
7126
  data: taskCommandSchema
6705
7127
  }),
6706
- z18.object({
6707
- type: z18.literal("TaskEvent" /* TaskEvent */),
6708
- origin: z18.literal("server" /* Server */),
6709
- relayClientId: z18.string().optional(),
7128
+ z19.object({
7129
+ type: z19.literal("TaskEvent" /* TaskEvent */),
7130
+ origin: z19.literal("server" /* Server */),
7131
+ relayClientId: z19.string().optional(),
6710
7132
  data: taskEventSchema
6711
7133
  })
6712
7134
  ]);
6713
7135
 
6714
7136
  // src/mcp.ts
6715
- import { z as z19 } from "zod";
6716
- var mcpExecutionStatusSchema = z19.discriminatedUnion("status", [
6717
- z19.object({
6718
- executionId: z19.string(),
6719
- status: z19.literal("started"),
6720
- serverName: z19.string(),
6721
- toolName: z19.string()
7137
+ import { z as z20 } from "zod";
7138
+ var mcpExecutionStatusSchema = z20.discriminatedUnion("status", [
7139
+ z20.object({
7140
+ executionId: z20.string(),
7141
+ status: z20.literal("started"),
7142
+ serverName: z20.string(),
7143
+ toolName: z20.string()
6722
7144
  }),
6723
- z19.object({
6724
- executionId: z19.string(),
6725
- status: z19.literal("output"),
6726
- response: z19.string()
7145
+ z20.object({
7146
+ executionId: z20.string(),
7147
+ status: z20.literal("output"),
7148
+ response: z20.string()
6727
7149
  }),
6728
- z19.object({
6729
- executionId: z19.string(),
6730
- status: z19.literal("completed"),
6731
- response: z19.string().optional()
7150
+ z20.object({
7151
+ executionId: z20.string(),
7152
+ status: z20.literal("completed"),
7153
+ response: z20.string().optional()
6732
7154
  }),
6733
- z19.object({
6734
- executionId: z19.string(),
6735
- status: z19.literal("error"),
6736
- error: z19.string().optional()
7155
+ z20.object({
7156
+ executionId: z20.string(),
7157
+ status: z20.literal("error"),
7158
+ error: z20.string().optional()
6737
7159
  })
6738
7160
  ]);
6739
7161
 
@@ -6743,40 +7165,40 @@ function shouldUseSingleFileRead(modelId) {
6743
7165
  }
6744
7166
 
6745
7167
  // src/todo.ts
6746
- import { z as z20 } from "zod";
6747
- var todoStatusSchema = z20.enum(["pending", "in_progress", "completed"]);
6748
- var todoItemSchema = z20.object({
6749
- id: z20.string(),
6750
- content: z20.string(),
7168
+ import { z as z21 } from "zod";
7169
+ var todoStatusSchema = z21.enum(["pending", "in_progress", "completed"]);
7170
+ var todoItemSchema = z21.object({
7171
+ id: z21.string(),
7172
+ content: z21.string(),
6751
7173
  status: todoStatusSchema
6752
7174
  });
6753
7175
 
6754
7176
  // src/terminal.ts
6755
- import { z as z21 } from "zod";
6756
- var commandExecutionStatusSchema = z21.discriminatedUnion("status", [
6757
- z21.object({
6758
- executionId: z21.string(),
6759
- status: z21.literal("started"),
6760
- pid: z21.number().optional(),
6761
- command: z21.string()
7177
+ import { z as z22 } from "zod";
7178
+ var commandExecutionStatusSchema = z22.discriminatedUnion("status", [
7179
+ z22.object({
7180
+ executionId: z22.string(),
7181
+ status: z22.literal("started"),
7182
+ pid: z22.number().optional(),
7183
+ command: z22.string()
6762
7184
  }),
6763
- z21.object({
6764
- executionId: z21.string(),
6765
- status: z21.literal("output"),
6766
- output: z21.string()
7185
+ z22.object({
7186
+ executionId: z22.string(),
7187
+ status: z22.literal("output"),
7188
+ output: z22.string()
6767
7189
  }),
6768
- z21.object({
6769
- executionId: z21.string(),
6770
- status: z21.literal("exited"),
6771
- exitCode: z21.number().optional()
7190
+ z22.object({
7191
+ executionId: z22.string(),
7192
+ status: z22.literal("exited"),
7193
+ exitCode: z22.number().optional()
6772
7194
  }),
6773
- z21.object({
6774
- executionId: z21.string(),
6775
- status: z21.literal("fallback")
7195
+ z22.object({
7196
+ executionId: z22.string(),
7197
+ status: z22.literal("fallback")
6776
7198
  }),
6777
- z21.object({
6778
- executionId: z21.string(),
6779
- status: z21.literal("timeout")
7199
+ z22.object({
7200
+ executionId: z22.string(),
7201
+ status: z22.literal("timeout")
6780
7202
  })
6781
7203
  ]);
6782
7204
  export {
@@ -6792,12 +7214,12 @@ export {
6792
7214
  BEDROCK_REGIONS,
6793
7215
  BEDROCK_SERVICE_TIER_MODEL_IDS,
6794
7216
  BEDROCK_SERVICE_TIER_PRICING,
6795
- CLAUDE_CODE_DEFAULT_MAX_OUTPUT_TOKENS,
6796
7217
  CODEBASE_INDEX_DEFAULTS,
6797
7218
  CONSENT_COOKIE_NAME,
6798
7219
  CONTEXT_MANAGEMENT_EVENTS,
6799
7220
  COOKIE_CONSENT_EVENTS,
6800
7221
  ConnectionState,
7222
+ ConsecutiveMistakeError,
6801
7223
  DEEP_SEEK_DEFAULT_TEMPERATURE,
6802
7224
  DEFAULT_CHECKPOINT_TIMEOUT_SECONDS,
6803
7225
  DEFAULT_CONSECUTIVE_MISTAKE_LIMIT,
@@ -6838,6 +7260,7 @@ export {
6838
7260
  MISTRAL_DEFAULT_TEMPERATURE,
6839
7261
  MODELS_BY_PROVIDER,
6840
7262
  MOONSHOT_DEFAULT_TEMPERATURE,
7263
+ NATIVE_TOOL_DEFAULTS,
6841
7264
  OPENAI_AZURE_AI_INFERENCE_PATH,
6842
7265
  OPENAI_NATIVE_DEFAULT_TEMPERATURE,
6843
7266
  OPENROUTER_DEFAULT_PROVIDER_NAME,
@@ -6863,6 +7286,7 @@ export {
6863
7286
  VERCEL_AI_GATEWAY_PROMPT_CACHING_MODELS,
6864
7287
  VERCEL_AI_GATEWAY_VISION_AND_TOOLS_MODELS,
6865
7288
  VERCEL_AI_GATEWAY_VISION_ONLY_MODELS,
7289
+ VERTEX_1M_CONTEXT_MODEL_IDS,
6866
7290
  VERTEX_REGIONS,
6867
7291
  ZAI_DEFAULT_TEMPERATURE,
6868
7292
  ackSchema,
@@ -6882,6 +7306,7 @@ export {
6882
7306
  chutesModels,
6883
7307
  claudeCodeDefaultModelId,
6884
7308
  claudeCodeModels,
7309
+ claudeCodeReasoningConfig,
6885
7310
  clineAskSchema,
6886
7311
  clineAsks,
6887
7312
  clineMessageSchema,
@@ -6896,7 +7321,6 @@ export {
6896
7321
  commandIds,
6897
7322
  contextCondenseSchema,
6898
7323
  contextTruncationSchema,
6899
- convertModelNameForVertex,
6900
7324
  customModePromptsSchema,
6901
7325
  customModesSettingsSchema,
6902
7326
  customProviders,
@@ -6905,6 +7329,7 @@ export {
6905
7329
  deepInfraDefaultModelInfo,
6906
7330
  deepSeekDefaultModelId,
6907
7331
  deepSeekModels,
7332
+ defineCustomTool,
6908
7333
  discriminatedProviderSettingsWithIdSchema,
6909
7334
  doubaoDefaultModelId,
6910
7335
  doubaoDefaultModelInfo,
@@ -6918,6 +7343,8 @@ export {
6918
7343
  extensionBridgeEventSchema,
6919
7344
  extensionInstanceSchema,
6920
7345
  extractApiProviderErrorProperties,
7346
+ extractConsecutiveMistakeErrorProperties,
7347
+ extractMessageFromJsonPayload,
6921
7348
  fauxProviders,
6922
7349
  featherlessDefaultModelId,
6923
7350
  featherlessModels,
@@ -6927,7 +7354,6 @@ export {
6927
7354
  geminiDefaultModelId,
6928
7355
  geminiModels,
6929
7356
  getApiProtocol,
6930
- getClaudeCodeModelId,
6931
7357
  getEffectiveProtocol,
6932
7358
  getErrorMessage,
6933
7359
  getErrorStatusCode,
@@ -6952,6 +7378,7 @@ export {
6952
7378
  ioIntelligenceModels,
6953
7379
  ipcMessageSchema,
6954
7380
  isApiProviderError,
7381
+ isConsecutiveMistakeError,
6955
7382
  isContextManagementEvent,
6956
7383
  isCustomProvider,
6957
7384
  isDynamicProvider,
@@ -6999,6 +7426,7 @@ export {
6999
7426
  moonshotDefaultModelId,
7000
7427
  moonshotModels,
7001
7428
  nonBlockingAsks,
7429
+ normalizeClaudeCodeModelId,
7002
7430
  ollamaDefaultModelId,
7003
7431
  ollamaDefaultModelInfo,
7004
7432
  openAiModelInfoSaneDefaults,
@@ -7011,6 +7439,7 @@ export {
7011
7439
  organizationDefaultSettingsSchema,
7012
7440
  organizationFeaturesSchema,
7013
7441
  organizationSettingsSchema,
7442
+ z17 as parametersSchema,
7014
7443
  promptComponentSchema,
7015
7444
  providerNames,
7016
7445
  providerNamesSchema,