@roo-code/types 1.106.0 → 1.108.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.js CHANGED
@@ -78,7 +78,9 @@ var clineSays = [
78
78
  "condense_context_error",
79
79
  "sliding_window_truncation",
80
80
  "codebase_search_result",
81
- "user_edit_todos"
81
+ "user_edit_todos",
82
+ "too_many_tools_warning",
83
+ "tool"
82
84
  ];
83
85
  var clineSaySchema = z.enum(clineSays);
84
86
  var toolProgressStatusSchema = z.object({
@@ -146,6 +148,7 @@ var toolGroupsSchema = z2.enum(toolGroups);
146
148
  var toolNames = [
147
149
  "execute_command",
148
150
  "read_file",
151
+ "read_command_output",
149
152
  "write_to_file",
150
153
  "apply_diff",
151
154
  "search_and_replace",
@@ -161,10 +164,10 @@ var toolNames = [
161
164
  "attempt_completion",
162
165
  "switch_mode",
163
166
  "new_task",
164
- "fetch_instructions",
165
167
  "codebase_search",
166
168
  "update_todo_list",
167
169
  "run_slash_command",
170
+ "skill",
168
171
  "generate_image",
169
172
  "custom_tool"
170
173
  ];
@@ -176,20 +179,6 @@ var toolUsageSchema = z2.record(
176
179
  failures: z2.number()
177
180
  })
178
181
  );
179
- var TOOL_PROTOCOL = {
180
- XML: "xml",
181
- NATIVE: "native"
182
- };
183
- var NATIVE_TOOL_DEFAULTS = {
184
- supportsNativeTools: true,
185
- defaultToolProtocol: TOOL_PROTOCOL.NATIVE
186
- };
187
- function isNativeProtocol(protocol) {
188
- return protocol === TOOL_PROTOCOL.NATIVE;
189
- }
190
- function getEffectiveProtocol(toolProtocol) {
191
- return toolProtocol || TOOL_PROTOCOL.XML;
192
- }
193
182
 
194
183
  // src/events.ts
195
184
  var RooCodeEventName = /* @__PURE__ */ ((RooCodeEventName2) => {
@@ -477,10 +466,6 @@ var modelInfoSchema = z5.object({
477
466
  isStealthModel: z5.boolean().optional(),
478
467
  // Flag to indicate if the model is free (no cost)
479
468
  isFree: z5.boolean().optional(),
480
- // Flag to indicate if the model supports native tool calling (OpenAI-style function calling)
481
- supportsNativeTools: z5.boolean().optional(),
482
- // Default tool protocol preferred by this model (if not specified, falls back to capability/provider defaults)
483
- defaultToolProtocol: z5.enum(["xml", "native"]).optional(),
484
469
  // Exclude specific native tools from being available (only applies to native protocol)
485
470
  // These tools will be removed from the set of tools available to the model
486
471
  excludedTools: z5.array(z5.string()).optional(),
@@ -577,8 +562,6 @@ var anthropicModels = {
577
562
  // Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
578
563
  supportsImages: true,
579
564
  supportsPromptCache: true,
580
- supportsNativeTools: true,
581
- defaultToolProtocol: "native",
582
565
  inputPrice: 3,
583
566
  // $3 per million input tokens (≤200K context)
584
567
  outputPrice: 15,
@@ -611,8 +594,6 @@ var anthropicModels = {
611
594
  // Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
612
595
  supportsImages: true,
613
596
  supportsPromptCache: true,
614
- supportsNativeTools: true,
615
- defaultToolProtocol: "native",
616
597
  inputPrice: 3,
617
598
  // $3 per million input tokens (≤200K context)
618
599
  outputPrice: 15,
@@ -644,8 +625,6 @@ var anthropicModels = {
644
625
  contextWindow: 2e5,
645
626
  supportsImages: true,
646
627
  supportsPromptCache: true,
647
- supportsNativeTools: true,
648
- defaultToolProtocol: "native",
649
628
  inputPrice: 5,
650
629
  // $5 per million input tokens
651
630
  outputPrice: 25,
@@ -662,8 +641,6 @@ var anthropicModels = {
662
641
  contextWindow: 2e5,
663
642
  supportsImages: true,
664
643
  supportsPromptCache: true,
665
- supportsNativeTools: true,
666
- defaultToolProtocol: "native",
667
644
  inputPrice: 15,
668
645
  // $15 per million input tokens
669
646
  outputPrice: 75,
@@ -680,8 +657,6 @@ var anthropicModels = {
680
657
  contextWindow: 2e5,
681
658
  supportsImages: true,
682
659
  supportsPromptCache: true,
683
- supportsNativeTools: true,
684
- defaultToolProtocol: "native",
685
660
  inputPrice: 15,
686
661
  // $15 per million input tokens
687
662
  outputPrice: 75,
@@ -698,8 +673,6 @@ var anthropicModels = {
698
673
  contextWindow: 2e5,
699
674
  supportsImages: true,
700
675
  supportsPromptCache: true,
701
- supportsNativeTools: true,
702
- defaultToolProtocol: "native",
703
676
  inputPrice: 3,
704
677
  // $3 per million input tokens
705
678
  outputPrice: 15,
@@ -717,8 +690,6 @@ var anthropicModels = {
717
690
  contextWindow: 2e5,
718
691
  supportsImages: true,
719
692
  supportsPromptCache: true,
720
- supportsNativeTools: true,
721
- defaultToolProtocol: "native",
722
693
  inputPrice: 3,
723
694
  // $3 per million input tokens
724
695
  outputPrice: 15,
@@ -733,8 +704,6 @@ var anthropicModels = {
733
704
  contextWindow: 2e5,
734
705
  supportsImages: true,
735
706
  supportsPromptCache: true,
736
- supportsNativeTools: true,
737
- defaultToolProtocol: "native",
738
707
  inputPrice: 3,
739
708
  // $3 per million input tokens
740
709
  outputPrice: 15,
@@ -749,8 +718,6 @@ var anthropicModels = {
749
718
  contextWindow: 2e5,
750
719
  supportsImages: false,
751
720
  supportsPromptCache: true,
752
- supportsNativeTools: true,
753
- defaultToolProtocol: "native",
754
721
  inputPrice: 1,
755
722
  outputPrice: 5,
756
723
  cacheWritesPrice: 1.25,
@@ -761,8 +728,6 @@ var anthropicModels = {
761
728
  contextWindow: 2e5,
762
729
  supportsImages: true,
763
730
  supportsPromptCache: true,
764
- supportsNativeTools: true,
765
- defaultToolProtocol: "native",
766
731
  inputPrice: 15,
767
732
  outputPrice: 75,
768
733
  cacheWritesPrice: 18.75,
@@ -773,8 +738,6 @@ var anthropicModels = {
773
738
  contextWindow: 2e5,
774
739
  supportsImages: true,
775
740
  supportsPromptCache: true,
776
- supportsNativeTools: true,
777
- defaultToolProtocol: "native",
778
741
  inputPrice: 0.25,
779
742
  outputPrice: 1.25,
780
743
  cacheWritesPrice: 0.3,
@@ -785,8 +748,6 @@ var anthropicModels = {
785
748
  contextWindow: 2e5,
786
749
  supportsImages: true,
787
750
  supportsPromptCache: true,
788
- supportsNativeTools: true,
789
- defaultToolProtocol: "native",
790
751
  inputPrice: 1,
791
752
  outputPrice: 5,
792
753
  cacheWritesPrice: 1.25,
@@ -804,7 +765,6 @@ var basetenModels = {
804
765
  contextWindow: 262e3,
805
766
  supportsImages: false,
806
767
  supportsPromptCache: false,
807
- supportsNativeTools: true,
808
768
  inputPrice: 0.6,
809
769
  outputPrice: 2.5,
810
770
  cacheWritesPrice: 0,
@@ -816,7 +776,6 @@ var basetenModels = {
816
776
  contextWindow: 2e5,
817
777
  supportsImages: false,
818
778
  supportsPromptCache: false,
819
- supportsNativeTools: true,
820
779
  inputPrice: 0.6,
821
780
  outputPrice: 2.2,
822
781
  cacheWritesPrice: 0,
@@ -828,7 +787,6 @@ var basetenModels = {
828
787
  contextWindow: 163840,
829
788
  supportsImages: false,
830
789
  supportsPromptCache: false,
831
- supportsNativeTools: true,
832
790
  inputPrice: 2.55,
833
791
  outputPrice: 5.95,
834
792
  cacheWritesPrice: 0,
@@ -840,7 +798,6 @@ var basetenModels = {
840
798
  contextWindow: 163840,
841
799
  supportsImages: false,
842
800
  supportsPromptCache: false,
843
- supportsNativeTools: true,
844
801
  inputPrice: 2.55,
845
802
  outputPrice: 5.95,
846
803
  cacheWritesPrice: 0,
@@ -852,7 +809,6 @@ var basetenModels = {
852
809
  contextWindow: 163840,
853
810
  supportsImages: false,
854
811
  supportsPromptCache: false,
855
- supportsNativeTools: true,
856
812
  inputPrice: 0.77,
857
813
  outputPrice: 0.77,
858
814
  cacheWritesPrice: 0,
@@ -864,7 +820,6 @@ var basetenModels = {
864
820
  contextWindow: 163840,
865
821
  supportsImages: false,
866
822
  supportsPromptCache: false,
867
- supportsNativeTools: true,
868
823
  inputPrice: 0.5,
869
824
  outputPrice: 1.5,
870
825
  cacheWritesPrice: 0,
@@ -876,7 +831,6 @@ var basetenModels = {
876
831
  contextWindow: 163840,
877
832
  supportsImages: false,
878
833
  supportsPromptCache: false,
879
- supportsNativeTools: true,
880
834
  inputPrice: 0.3,
881
835
  outputPrice: 0.45,
882
836
  cacheWritesPrice: 0,
@@ -888,7 +842,6 @@ var basetenModels = {
888
842
  contextWindow: 128072,
889
843
  supportsImages: false,
890
844
  supportsPromptCache: false,
891
- supportsNativeTools: true,
892
845
  inputPrice: 0.1,
893
846
  outputPrice: 0.5,
894
847
  cacheWritesPrice: 0,
@@ -900,7 +853,6 @@ var basetenModels = {
900
853
  contextWindow: 262144,
901
854
  supportsImages: false,
902
855
  supportsPromptCache: false,
903
- supportsNativeTools: true,
904
856
  inputPrice: 0.22,
905
857
  outputPrice: 0.8,
906
858
  cacheWritesPrice: 0,
@@ -912,7 +864,6 @@ var basetenModels = {
912
864
  contextWindow: 262144,
913
865
  supportsImages: false,
914
866
  supportsPromptCache: false,
915
- supportsNativeTools: true,
916
867
  inputPrice: 0.38,
917
868
  outputPrice: 1.53,
918
869
  cacheWritesPrice: 0,
@@ -924,7 +875,6 @@ var basetenModels = {
924
875
  contextWindow: 262e3,
925
876
  supportsImages: false,
926
877
  supportsPromptCache: false,
927
- supportsNativeTools: true,
928
878
  inputPrice: 0.6,
929
879
  outputPrice: 2.5,
930
880
  cacheWritesPrice: 0,
@@ -944,8 +894,6 @@ var bedrockModels = {
944
894
  supportsImages: true,
945
895
  supportsPromptCache: true,
946
896
  supportsReasoningBudget: true,
947
- supportsNativeTools: true,
948
- defaultToolProtocol: "native",
949
897
  inputPrice: 3,
950
898
  outputPrice: 15,
951
899
  cacheWritesPrice: 3.75,
@@ -959,7 +907,6 @@ var bedrockModels = {
959
907
  contextWindow: 3e5,
960
908
  supportsImages: true,
961
909
  supportsPromptCache: true,
962
- supportsNativeTools: true,
963
910
  inputPrice: 0.8,
964
911
  outputPrice: 3.2,
965
912
  cacheWritesPrice: 0.8,
@@ -975,7 +922,6 @@ var bedrockModels = {
975
922
  contextWindow: 3e5,
976
923
  supportsImages: true,
977
924
  supportsPromptCache: false,
978
- supportsNativeTools: true,
979
925
  inputPrice: 1,
980
926
  outputPrice: 4,
981
927
  cacheWritesPrice: 1,
@@ -989,7 +935,6 @@ var bedrockModels = {
989
935
  contextWindow: 3e5,
990
936
  supportsImages: true,
991
937
  supportsPromptCache: true,
992
- supportsNativeTools: true,
993
938
  inputPrice: 0.06,
994
939
  outputPrice: 0.24,
995
940
  cacheWritesPrice: 0.06,
@@ -1005,7 +950,6 @@ var bedrockModels = {
1005
950
  contextWindow: 1e6,
1006
951
  supportsImages: true,
1007
952
  supportsPromptCache: true,
1008
- supportsNativeTools: true,
1009
953
  inputPrice: 0.33,
1010
954
  outputPrice: 2.75,
1011
955
  cacheWritesPrice: 0,
@@ -1021,7 +965,6 @@ var bedrockModels = {
1021
965
  contextWindow: 128e3,
1022
966
  supportsImages: false,
1023
967
  supportsPromptCache: true,
1024
- supportsNativeTools: true,
1025
968
  inputPrice: 0.035,
1026
969
  outputPrice: 0.14,
1027
970
  cacheWritesPrice: 0.035,
@@ -1038,8 +981,6 @@ var bedrockModels = {
1038
981
  supportsImages: true,
1039
982
  supportsPromptCache: true,
1040
983
  supportsReasoningBudget: true,
1041
- supportsNativeTools: true,
1042
- defaultToolProtocol: "native",
1043
984
  inputPrice: 3,
1044
985
  outputPrice: 15,
1045
986
  cacheWritesPrice: 3.75,
@@ -1054,8 +995,6 @@ var bedrockModels = {
1054
995
  supportsImages: true,
1055
996
  supportsPromptCache: true,
1056
997
  supportsReasoningBudget: true,
1057
- supportsNativeTools: true,
1058
- defaultToolProtocol: "native",
1059
998
  inputPrice: 15,
1060
999
  outputPrice: 75,
1061
1000
  cacheWritesPrice: 18.75,
@@ -1070,8 +1009,6 @@ var bedrockModels = {
1070
1009
  supportsImages: true,
1071
1010
  supportsPromptCache: true,
1072
1011
  supportsReasoningBudget: true,
1073
- supportsNativeTools: true,
1074
- defaultToolProtocol: "native",
1075
1012
  inputPrice: 5,
1076
1013
  outputPrice: 25,
1077
1014
  cacheWritesPrice: 6.25,
@@ -1086,8 +1023,6 @@ var bedrockModels = {
1086
1023
  supportsImages: true,
1087
1024
  supportsPromptCache: true,
1088
1025
  supportsReasoningBudget: true,
1089
- supportsNativeTools: true,
1090
- defaultToolProtocol: "native",
1091
1026
  inputPrice: 15,
1092
1027
  outputPrice: 75,
1093
1028
  cacheWritesPrice: 18.75,
@@ -1102,8 +1037,6 @@ var bedrockModels = {
1102
1037
  supportsImages: true,
1103
1038
  supportsPromptCache: true,
1104
1039
  supportsReasoningBudget: true,
1105
- supportsNativeTools: true,
1106
- defaultToolProtocol: "native",
1107
1040
  inputPrice: 3,
1108
1041
  outputPrice: 15,
1109
1042
  cacheWritesPrice: 3.75,
@@ -1117,8 +1050,6 @@ var bedrockModels = {
1117
1050
  contextWindow: 2e5,
1118
1051
  supportsImages: true,
1119
1052
  supportsPromptCache: true,
1120
- supportsNativeTools: true,
1121
- defaultToolProtocol: "native",
1122
1053
  inputPrice: 3,
1123
1054
  outputPrice: 15,
1124
1055
  cacheWritesPrice: 3.75,
@@ -1132,8 +1063,6 @@ var bedrockModels = {
1132
1063
  contextWindow: 2e5,
1133
1064
  supportsImages: false,
1134
1065
  supportsPromptCache: true,
1135
- supportsNativeTools: true,
1136
- defaultToolProtocol: "native",
1137
1066
  inputPrice: 0.8,
1138
1067
  outputPrice: 4,
1139
1068
  cacheWritesPrice: 1,
@@ -1148,8 +1077,6 @@ var bedrockModels = {
1148
1077
  supportsImages: true,
1149
1078
  supportsPromptCache: true,
1150
1079
  supportsReasoningBudget: true,
1151
- supportsNativeTools: true,
1152
- defaultToolProtocol: "native",
1153
1080
  inputPrice: 1,
1154
1081
  outputPrice: 5,
1155
1082
  cacheWritesPrice: 1.25,
@@ -1165,8 +1092,6 @@ var bedrockModels = {
1165
1092
  contextWindow: 2e5,
1166
1093
  supportsImages: true,
1167
1094
  supportsPromptCache: false,
1168
- supportsNativeTools: true,
1169
- defaultToolProtocol: "native",
1170
1095
  inputPrice: 3,
1171
1096
  outputPrice: 15
1172
1097
  },
@@ -1175,8 +1100,6 @@ var bedrockModels = {
1175
1100
  contextWindow: 2e5,
1176
1101
  supportsImages: true,
1177
1102
  supportsPromptCache: false,
1178
- supportsNativeTools: true,
1179
- defaultToolProtocol: "native",
1180
1103
  inputPrice: 15,
1181
1104
  outputPrice: 75
1182
1105
  },
@@ -1185,8 +1108,6 @@ var bedrockModels = {
1185
1108
  contextWindow: 2e5,
1186
1109
  supportsImages: true,
1187
1110
  supportsPromptCache: false,
1188
- supportsNativeTools: true,
1189
- defaultToolProtocol: "native",
1190
1111
  inputPrice: 3,
1191
1112
  outputPrice: 15
1192
1113
  },
@@ -1195,8 +1116,6 @@ var bedrockModels = {
1195
1116
  contextWindow: 2e5,
1196
1117
  supportsImages: true,
1197
1118
  supportsPromptCache: false,
1198
- supportsNativeTools: true,
1199
- defaultToolProtocol: "native",
1200
1119
  inputPrice: 0.25,
1201
1120
  outputPrice: 1.25
1202
1121
  },
@@ -1205,7 +1124,6 @@ var bedrockModels = {
1205
1124
  contextWindow: 128e3,
1206
1125
  supportsImages: false,
1207
1126
  supportsPromptCache: false,
1208
- supportsNativeTools: true,
1209
1127
  inputPrice: 1.35,
1210
1128
  outputPrice: 5.4
1211
1129
  },
@@ -1214,7 +1132,6 @@ var bedrockModels = {
1214
1132
  contextWindow: 128e3,
1215
1133
  supportsImages: false,
1216
1134
  supportsPromptCache: false,
1217
- supportsNativeTools: true,
1218
1135
  inputPrice: 0.5,
1219
1136
  outputPrice: 1.5,
1220
1137
  description: "GPT-OSS 20B - Optimized for low latency and local/specialized use cases"
@@ -1224,7 +1141,6 @@ var bedrockModels = {
1224
1141
  contextWindow: 128e3,
1225
1142
  supportsImages: false,
1226
1143
  supportsPromptCache: false,
1227
- supportsNativeTools: true,
1228
1144
  inputPrice: 2,
1229
1145
  outputPrice: 6,
1230
1146
  description: "GPT-OSS 120B - Production-ready, general-purpose, high-reasoning model"
@@ -1234,7 +1150,6 @@ var bedrockModels = {
1234
1150
  contextWindow: 128e3,
1235
1151
  supportsImages: false,
1236
1152
  supportsPromptCache: false,
1237
- supportsNativeTools: true,
1238
1153
  inputPrice: 0.72,
1239
1154
  outputPrice: 0.72,
1240
1155
  description: "Llama 3.3 Instruct (70B)"
@@ -1244,7 +1159,6 @@ var bedrockModels = {
1244
1159
  contextWindow: 128e3,
1245
1160
  supportsImages: true,
1246
1161
  supportsPromptCache: false,
1247
- supportsNativeTools: true,
1248
1162
  inputPrice: 0.72,
1249
1163
  outputPrice: 0.72,
1250
1164
  description: "Llama 3.2 Instruct (90B)"
@@ -1254,7 +1168,6 @@ var bedrockModels = {
1254
1168
  contextWindow: 128e3,
1255
1169
  supportsImages: true,
1256
1170
  supportsPromptCache: false,
1257
- supportsNativeTools: true,
1258
1171
  inputPrice: 0.16,
1259
1172
  outputPrice: 0.16,
1260
1173
  description: "Llama 3.2 Instruct (11B)"
@@ -1264,7 +1177,6 @@ var bedrockModels = {
1264
1177
  contextWindow: 128e3,
1265
1178
  supportsImages: false,
1266
1179
  supportsPromptCache: false,
1267
- supportsNativeTools: true,
1268
1180
  inputPrice: 0.15,
1269
1181
  outputPrice: 0.15,
1270
1182
  description: "Llama 3.2 Instruct (3B)"
@@ -1274,7 +1186,6 @@ var bedrockModels = {
1274
1186
  contextWindow: 128e3,
1275
1187
  supportsImages: false,
1276
1188
  supportsPromptCache: false,
1277
- supportsNativeTools: true,
1278
1189
  inputPrice: 0.1,
1279
1190
  outputPrice: 0.1,
1280
1191
  description: "Llama 3.2 Instruct (1B)"
@@ -1284,7 +1195,6 @@ var bedrockModels = {
1284
1195
  contextWindow: 128e3,
1285
1196
  supportsImages: false,
1286
1197
  supportsPromptCache: false,
1287
- supportsNativeTools: true,
1288
1198
  inputPrice: 2.4,
1289
1199
  outputPrice: 2.4,
1290
1200
  description: "Llama 3.1 Instruct (405B)"
@@ -1294,7 +1204,6 @@ var bedrockModels = {
1294
1204
  contextWindow: 128e3,
1295
1205
  supportsImages: false,
1296
1206
  supportsPromptCache: false,
1297
- supportsNativeTools: true,
1298
1207
  inputPrice: 0.72,
1299
1208
  outputPrice: 0.72,
1300
1209
  description: "Llama 3.1 Instruct (70B)"
@@ -1304,7 +1213,6 @@ var bedrockModels = {
1304
1213
  contextWindow: 128e3,
1305
1214
  supportsImages: false,
1306
1215
  supportsPromptCache: false,
1307
- supportsNativeTools: true,
1308
1216
  inputPrice: 0.9,
1309
1217
  outputPrice: 0.9,
1310
1218
  description: "Llama 3.1 Instruct (70B) (w/ latency optimized inference)"
@@ -1314,7 +1222,6 @@ var bedrockModels = {
1314
1222
  contextWindow: 8e3,
1315
1223
  supportsImages: false,
1316
1224
  supportsPromptCache: false,
1317
- supportsNativeTools: true,
1318
1225
  inputPrice: 0.22,
1319
1226
  outputPrice: 0.22,
1320
1227
  description: "Llama 3.1 Instruct (8B)"
@@ -1324,7 +1231,6 @@ var bedrockModels = {
1324
1231
  contextWindow: 8e3,
1325
1232
  supportsImages: false,
1326
1233
  supportsPromptCache: false,
1327
- supportsNativeTools: true,
1328
1234
  inputPrice: 2.65,
1329
1235
  outputPrice: 3.5
1330
1236
  },
@@ -1333,7 +1239,6 @@ var bedrockModels = {
1333
1239
  contextWindow: 4e3,
1334
1240
  supportsImages: false,
1335
1241
  supportsPromptCache: false,
1336
- supportsNativeTools: true,
1337
1242
  inputPrice: 0.3,
1338
1243
  outputPrice: 0.6
1339
1244
  },
@@ -1342,7 +1247,6 @@ var bedrockModels = {
1342
1247
  contextWindow: 8e3,
1343
1248
  supportsImages: false,
1344
1249
  supportsPromptCache: false,
1345
- supportsNativeTools: true,
1346
1250
  inputPrice: 0.15,
1347
1251
  outputPrice: 0.2,
1348
1252
  description: "Amazon Titan Text Lite"
@@ -1352,7 +1256,6 @@ var bedrockModels = {
1352
1256
  contextWindow: 8e3,
1353
1257
  supportsImages: false,
1354
1258
  supportsPromptCache: false,
1355
- supportsNativeTools: true,
1356
1259
  inputPrice: 0.2,
1357
1260
  outputPrice: 0.6,
1358
1261
  description: "Amazon Titan Text Express"
@@ -1362,8 +1265,6 @@ var bedrockModels = {
1362
1265
  contextWindow: 262144,
1363
1266
  supportsImages: false,
1364
1267
  supportsPromptCache: false,
1365
- supportsNativeTools: true,
1366
- defaultToolProtocol: "native",
1367
1268
  preserveReasoning: true,
1368
1269
  inputPrice: 0.6,
1369
1270
  outputPrice: 2.5,
@@ -1374,8 +1275,6 @@ var bedrockModels = {
1374
1275
  contextWindow: 196608,
1375
1276
  supportsImages: false,
1376
1277
  supportsPromptCache: false,
1377
- supportsNativeTools: true,
1378
- defaultToolProtocol: "native",
1379
1278
  preserveReasoning: true,
1380
1279
  inputPrice: 0.3,
1381
1280
  outputPrice: 1.2,
@@ -1386,8 +1285,6 @@ var bedrockModels = {
1386
1285
  contextWindow: 262144,
1387
1286
  supportsImages: false,
1388
1287
  supportsPromptCache: false,
1389
- supportsNativeTools: true,
1390
- defaultToolProtocol: "native",
1391
1288
  inputPrice: 0.15,
1392
1289
  outputPrice: 1.2,
1393
1290
  description: "Qwen3 Next 80B (MoE model with 3B active parameters)"
@@ -1397,8 +1294,6 @@ var bedrockModels = {
1397
1294
  contextWindow: 262144,
1398
1295
  supportsImages: false,
1399
1296
  supportsPromptCache: false,
1400
- supportsNativeTools: true,
1401
- defaultToolProtocol: "native",
1402
1297
  inputPrice: 0.45,
1403
1298
  outputPrice: 1.8,
1404
1299
  description: "Qwen3 Coder 480B (MoE model with 35B active parameters)"
@@ -1489,26 +1384,14 @@ var BEDROCK_SERVICE_TIER_PRICING = {
1489
1384
  // src/providers/cerebras.ts
1490
1385
  var cerebrasDefaultModelId = "gpt-oss-120b";
1491
1386
  var cerebrasModels = {
1492
- "zai-glm-4.6": {
1493
- maxTokens: 16384,
1494
- // Conservative default to avoid premature rate limiting (Cerebras reserves quota upfront)
1495
- contextWindow: 131072,
1496
- supportsImages: false,
1497
- supportsPromptCache: false,
1498
- supportsNativeTools: true,
1499
- defaultToolProtocol: "native",
1500
- inputPrice: 0,
1501
- outputPrice: 0,
1502
- description: "Fast general-purpose model on Cerebras (up to 1,000 tokens/s). To be deprecated soon."
1503
- },
1504
1387
  "zai-glm-4.7": {
1505
1388
  maxTokens: 16384,
1506
1389
  // Conservative default to avoid premature rate limiting (Cerebras reserves quota upfront)
1507
1390
  contextWindow: 131072,
1508
1391
  supportsImages: false,
1509
- supportsPromptCache: false,
1510
- supportsNativeTools: true,
1511
- defaultToolProtocol: "native",
1392
+ supportsPromptCache: true,
1393
+ supportsTemperature: true,
1394
+ defaultTemperature: 1,
1512
1395
  inputPrice: 0,
1513
1396
  outputPrice: 0,
1514
1397
  description: "Highly capable general-purpose model on Cerebras (up to 1,000 tokens/s), competitive with leading proprietary models on coding tasks."
@@ -1519,8 +1402,6 @@ var cerebrasModels = {
1519
1402
  contextWindow: 64e3,
1520
1403
  supportsImages: false,
1521
1404
  supportsPromptCache: false,
1522
- supportsNativeTools: true,
1523
- defaultToolProtocol: "native",
1524
1405
  inputPrice: 0,
1525
1406
  outputPrice: 0,
1526
1407
  description: "Intelligent model with ~1400 tokens/s"
@@ -1531,8 +1412,6 @@ var cerebrasModels = {
1531
1412
  contextWindow: 64e3,
1532
1413
  supportsImages: false,
1533
1414
  supportsPromptCache: false,
1534
- supportsNativeTools: true,
1535
- defaultToolProtocol: "native",
1536
1415
  inputPrice: 0,
1537
1416
  outputPrice: 0,
1538
1417
  description: "Powerful model with ~2600 tokens/s"
@@ -1543,8 +1422,6 @@ var cerebrasModels = {
1543
1422
  contextWindow: 64e3,
1544
1423
  supportsImages: false,
1545
1424
  supportsPromptCache: false,
1546
- supportsNativeTools: true,
1547
- defaultToolProtocol: "native",
1548
1425
  inputPrice: 0,
1549
1426
  outputPrice: 0,
1550
1427
  description: "SOTA coding performance with ~2500 tokens/s"
@@ -1555,8 +1432,6 @@ var cerebrasModels = {
1555
1432
  contextWindow: 64e3,
1556
1433
  supportsImages: false,
1557
1434
  supportsPromptCache: false,
1558
- supportsNativeTools: true,
1559
- defaultToolProtocol: "native",
1560
1435
  inputPrice: 0,
1561
1436
  outputPrice: 0,
1562
1437
  description: "OpenAI GPT OSS model with ~2800 tokens/s\n\n\u2022 64K context window\n\u2022 Excels at efficient reasoning across science, math, and coding"
@@ -1571,8 +1446,6 @@ var chutesModels = {
1571
1446
  contextWindow: 163840,
1572
1447
  supportsImages: false,
1573
1448
  supportsPromptCache: false,
1574
- supportsNativeTools: true,
1575
- defaultToolProtocol: "native",
1576
1449
  inputPrice: 0,
1577
1450
  outputPrice: 0,
1578
1451
  description: "DeepSeek R1 0528 model."
@@ -1582,8 +1455,6 @@ var chutesModels = {
1582
1455
  contextWindow: 163840,
1583
1456
  supportsImages: false,
1584
1457
  supportsPromptCache: false,
1585
- supportsNativeTools: true,
1586
- defaultToolProtocol: "native",
1587
1458
  inputPrice: 0,
1588
1459
  outputPrice: 0,
1589
1460
  description: "DeepSeek R1 model."
@@ -1593,8 +1464,6 @@ var chutesModels = {
1593
1464
  contextWindow: 163840,
1594
1465
  supportsImages: false,
1595
1466
  supportsPromptCache: false,
1596
- supportsNativeTools: true,
1597
- defaultToolProtocol: "native",
1598
1467
  inputPrice: 0,
1599
1468
  outputPrice: 0,
1600
1469
  description: "DeepSeek V3 model."
@@ -1604,8 +1473,6 @@ var chutesModels = {
1604
1473
  contextWindow: 163840,
1605
1474
  supportsImages: false,
1606
1475
  supportsPromptCache: false,
1607
- supportsNativeTools: true,
1608
- defaultToolProtocol: "native",
1609
1476
  inputPrice: 0,
1610
1477
  outputPrice: 0,
1611
1478
  description: "DeepSeek V3.1 model."
@@ -1615,8 +1482,6 @@ var chutesModels = {
1615
1482
  contextWindow: 163840,
1616
1483
  supportsImages: false,
1617
1484
  supportsPromptCache: false,
1618
- supportsNativeTools: true,
1619
- defaultToolProtocol: "native",
1620
1485
  inputPrice: 0.23,
1621
1486
  outputPrice: 0.9,
1622
1487
  description: "DeepSeek\u2011V3.1\u2011Terminus is an update to V3.1 that improves language consistency by reducing CN/EN mix\u2011ups and eliminating random characters, while strengthening agent capabilities with notably better Code Agent and Search Agent performance."
@@ -1626,8 +1491,6 @@ var chutesModels = {
1626
1491
  contextWindow: 163840,
1627
1492
  supportsImages: false,
1628
1493
  supportsPromptCache: false,
1629
- supportsNativeTools: true,
1630
- defaultToolProtocol: "native",
1631
1494
  inputPrice: 1,
1632
1495
  outputPrice: 3,
1633
1496
  description: "DeepSeek-V3.1-turbo is an FP8, speculative-decoding turbo variant optimized for ultra-fast single-shot queries (~200 TPS), with outputs close to the originals and solid function calling/reasoning/structured output, priced at $1/M input and $3/M output tokens, using 2\xD7 quota per request and not intended for bulk workloads."
@@ -1637,8 +1500,6 @@ var chutesModels = {
1637
1500
  contextWindow: 163840,
1638
1501
  supportsImages: false,
1639
1502
  supportsPromptCache: false,
1640
- supportsNativeTools: true,
1641
- defaultToolProtocol: "native",
1642
1503
  inputPrice: 0.25,
1643
1504
  outputPrice: 0.35,
1644
1505
  description: "DeepSeek-V3.2-Exp is an experimental LLM that introduces DeepSeek Sparse Attention to improve long\u2011context training and inference efficiency while maintaining performance comparable to V3.1\u2011Terminus."
@@ -1650,8 +1511,6 @@ var chutesModels = {
1650
1511
  // From Groq
1651
1512
  supportsImages: false,
1652
1513
  supportsPromptCache: false,
1653
- supportsNativeTools: true,
1654
- defaultToolProtocol: "native",
1655
1514
  inputPrice: 0,
1656
1515
  outputPrice: 0,
1657
1516
  description: "Unsloth Llama 3.3 70B Instruct model."
@@ -1661,8 +1520,6 @@ var chutesModels = {
1661
1520
  contextWindow: 512e3,
1662
1521
  supportsImages: false,
1663
1522
  supportsPromptCache: false,
1664
- supportsNativeTools: true,
1665
- defaultToolProtocol: "native",
1666
1523
  inputPrice: 0,
1667
1524
  outputPrice: 0,
1668
1525
  description: "ChutesAI Llama 4 Scout 17B Instruct model, 512K context."
@@ -1672,8 +1529,6 @@ var chutesModels = {
1672
1529
  contextWindow: 128e3,
1673
1530
  supportsImages: false,
1674
1531
  supportsPromptCache: false,
1675
- supportsNativeTools: true,
1676
- defaultToolProtocol: "native",
1677
1532
  inputPrice: 0,
1678
1533
  outputPrice: 0,
1679
1534
  description: "Unsloth Mistral Nemo Instruct model."
@@ -1683,8 +1538,6 @@ var chutesModels = {
1683
1538
  contextWindow: 131072,
1684
1539
  supportsImages: false,
1685
1540
  supportsPromptCache: false,
1686
- supportsNativeTools: true,
1687
- defaultToolProtocol: "native",
1688
1541
  inputPrice: 0,
1689
1542
  outputPrice: 0,
1690
1543
  description: "Unsloth Gemma 3 12B IT model."
@@ -1694,8 +1547,6 @@ var chutesModels = {
1694
1547
  contextWindow: 131072,
1695
1548
  supportsImages: false,
1696
1549
  supportsPromptCache: false,
1697
- supportsNativeTools: true,
1698
- defaultToolProtocol: "native",
1699
1550
  inputPrice: 0,
1700
1551
  outputPrice: 0,
1701
1552
  description: "Nous DeepHermes 3 Llama 3 8B Preview model."
@@ -1705,8 +1556,6 @@ var chutesModels = {
1705
1556
  contextWindow: 131072,
1706
1557
  supportsImages: false,
1707
1558
  supportsPromptCache: false,
1708
- supportsNativeTools: true,
1709
- defaultToolProtocol: "native",
1710
1559
  inputPrice: 0,
1711
1560
  outputPrice: 0,
1712
1561
  description: "Unsloth Gemma 3 4B IT model."
@@ -1716,8 +1565,6 @@ var chutesModels = {
1716
1565
  contextWindow: 131072,
1717
1566
  supportsImages: false,
1718
1567
  supportsPromptCache: false,
1719
- supportsNativeTools: true,
1720
- defaultToolProtocol: "native",
1721
1568
  inputPrice: 0,
1722
1569
  outputPrice: 0,
1723
1570
  description: "Nvidia Llama 3.3 Nemotron Super 49B model."
@@ -1727,8 +1574,6 @@ var chutesModels = {
1727
1574
  contextWindow: 131072,
1728
1575
  supportsImages: false,
1729
1576
  supportsPromptCache: false,
1730
- supportsNativeTools: true,
1731
- defaultToolProtocol: "native",
1732
1577
  inputPrice: 0,
1733
1578
  outputPrice: 0,
1734
1579
  description: "Nvidia Llama 3.1 Nemotron Ultra 253B model."
@@ -1738,8 +1583,6 @@ var chutesModels = {
1738
1583
  contextWindow: 256e3,
1739
1584
  supportsImages: false,
1740
1585
  supportsPromptCache: false,
1741
- supportsNativeTools: true,
1742
- defaultToolProtocol: "native",
1743
1586
  inputPrice: 0,
1744
1587
  outputPrice: 0,
1745
1588
  description: "ChutesAI Llama 4 Maverick 17B Instruct FP8 model."
@@ -1749,8 +1592,6 @@ var chutesModels = {
1749
1592
  contextWindow: 163840,
1750
1593
  supportsImages: false,
1751
1594
  supportsPromptCache: false,
1752
- supportsNativeTools: true,
1753
- defaultToolProtocol: "native",
1754
1595
  inputPrice: 0,
1755
1596
  outputPrice: 0,
1756
1597
  description: "DeepSeek V3 Base model."
@@ -1760,8 +1601,6 @@ var chutesModels = {
1760
1601
  contextWindow: 163840,
1761
1602
  supportsImages: false,
1762
1603
  supportsPromptCache: false,
1763
- supportsNativeTools: true,
1764
- defaultToolProtocol: "native",
1765
1604
  inputPrice: 0,
1766
1605
  outputPrice: 0,
1767
1606
  description: "DeepSeek R1 Zero model."
@@ -1771,8 +1610,6 @@ var chutesModels = {
1771
1610
  contextWindow: 163840,
1772
1611
  supportsImages: false,
1773
1612
  supportsPromptCache: false,
1774
- supportsNativeTools: true,
1775
- defaultToolProtocol: "native",
1776
1613
  inputPrice: 0,
1777
1614
  outputPrice: 0,
1778
1615
  description: "DeepSeek V3 (0324) model."
@@ -1782,8 +1619,6 @@ var chutesModels = {
1782
1619
  contextWindow: 262144,
1783
1620
  supportsImages: false,
1784
1621
  supportsPromptCache: false,
1785
- supportsNativeTools: true,
1786
- defaultToolProtocol: "native",
1787
1622
  inputPrice: 0,
1788
1623
  outputPrice: 0,
1789
1624
  description: "Qwen3 235B A22B Instruct 2507 model with 262K context window."
@@ -1793,8 +1628,6 @@ var chutesModels = {
1793
1628
  contextWindow: 40960,
1794
1629
  supportsImages: false,
1795
1630
  supportsPromptCache: false,
1796
- supportsNativeTools: true,
1797
- defaultToolProtocol: "native",
1798
1631
  inputPrice: 0,
1799
1632
  outputPrice: 0,
1800
1633
  description: "Qwen3 235B A22B model."
@@ -1804,8 +1637,6 @@ var chutesModels = {
1804
1637
  contextWindow: 40960,
1805
1638
  supportsImages: false,
1806
1639
  supportsPromptCache: false,
1807
- supportsNativeTools: true,
1808
- defaultToolProtocol: "native",
1809
1640
  inputPrice: 0,
1810
1641
  outputPrice: 0,
1811
1642
  description: "Qwen3 32B model."
@@ -1815,8 +1646,6 @@ var chutesModels = {
1815
1646
  contextWindow: 40960,
1816
1647
  supportsImages: false,
1817
1648
  supportsPromptCache: false,
1818
- supportsNativeTools: true,
1819
- defaultToolProtocol: "native",
1820
1649
  inputPrice: 0,
1821
1650
  outputPrice: 0,
1822
1651
  description: "Qwen3 30B A3B model."
@@ -1826,8 +1655,6 @@ var chutesModels = {
1826
1655
  contextWindow: 40960,
1827
1656
  supportsImages: false,
1828
1657
  supportsPromptCache: false,
1829
- supportsNativeTools: true,
1830
- defaultToolProtocol: "native",
1831
1658
  inputPrice: 0,
1832
1659
  outputPrice: 0,
1833
1660
  description: "Qwen3 14B model."
@@ -1837,8 +1664,6 @@ var chutesModels = {
1837
1664
  contextWindow: 40960,
1838
1665
  supportsImages: false,
1839
1666
  supportsPromptCache: false,
1840
- supportsNativeTools: true,
1841
- defaultToolProtocol: "native",
1842
1667
  inputPrice: 0,
1843
1668
  outputPrice: 0,
1844
1669
  description: "Qwen3 8B model."
@@ -1848,8 +1673,6 @@ var chutesModels = {
1848
1673
  contextWindow: 163840,
1849
1674
  supportsImages: false,
1850
1675
  supportsPromptCache: false,
1851
- supportsNativeTools: true,
1852
- defaultToolProtocol: "native",
1853
1676
  inputPrice: 0,
1854
1677
  outputPrice: 0,
1855
1678
  description: "Microsoft MAI-DS-R1 FP8 model."
@@ -1859,8 +1682,6 @@ var chutesModels = {
1859
1682
  contextWindow: 163840,
1860
1683
  supportsImages: false,
1861
1684
  supportsPromptCache: false,
1862
- supportsNativeTools: true,
1863
- defaultToolProtocol: "native",
1864
1685
  inputPrice: 0,
1865
1686
  outputPrice: 0,
1866
1687
  description: "TNGTech DeepSeek R1T Chimera model."
@@ -1870,8 +1691,6 @@ var chutesModels = {
1870
1691
  contextWindow: 151329,
1871
1692
  supportsImages: false,
1872
1693
  supportsPromptCache: false,
1873
- supportsNativeTools: true,
1874
- defaultToolProtocol: "native",
1875
1694
  inputPrice: 0,
1876
1695
  outputPrice: 0,
1877
1696
  description: "GLM-4.5-Air model with 151,329 token context window and 106B total parameters with 12B activated."
@@ -1881,8 +1700,6 @@ var chutesModels = {
1881
1700
  contextWindow: 131072,
1882
1701
  supportsImages: false,
1883
1702
  supportsPromptCache: false,
1884
- supportsNativeTools: true,
1885
- defaultToolProtocol: "native",
1886
1703
  inputPrice: 0,
1887
1704
  outputPrice: 0,
1888
1705
  description: "GLM-4.5-FP8 model with 128k token context window, optimized for agent-based applications with MoE architecture."
@@ -1892,8 +1709,6 @@ var chutesModels = {
1892
1709
  contextWindow: 131072,
1893
1710
  supportsImages: false,
1894
1711
  supportsPromptCache: false,
1895
- supportsNativeTools: true,
1896
- defaultToolProtocol: "native",
1897
1712
  inputPrice: 1,
1898
1713
  outputPrice: 3,
1899
1714
  description: "GLM-4.5-turbo model with 128K token context window, optimized for fast inference."
@@ -1903,8 +1718,6 @@ var chutesModels = {
1903
1718
  contextWindow: 202752,
1904
1719
  supportsImages: false,
1905
1720
  supportsPromptCache: false,
1906
- supportsNativeTools: true,
1907
- defaultToolProtocol: "native",
1908
1721
  inputPrice: 0,
1909
1722
  outputPrice: 0,
1910
1723
  description: "GLM-4.6 introduces major upgrades over GLM-4.5, including a longer 200K-token context window for complex tasks, stronger coding performance in benchmarks and real-world tools (such as Claude Code, Cline, Roo Code, and Kilo Code), improved reasoning with tool use during inference, more capable and efficient agent integration, and refined writing that better matches human style, readability, and natural role-play scenarios."
@@ -1915,8 +1728,6 @@ var chutesModels = {
1915
1728
  contextWindow: 202752,
1916
1729
  supportsImages: false,
1917
1730
  supportsPromptCache: false,
1918
- supportsNativeTools: true,
1919
- defaultToolProtocol: "native",
1920
1731
  inputPrice: 1.15,
1921
1732
  outputPrice: 3.25,
1922
1733
  description: "GLM-4.6-turbo model with 200K-token context window, optimized for fast inference."
@@ -1926,8 +1737,6 @@ var chutesModels = {
1926
1737
  contextWindow: 128e3,
1927
1738
  supportsImages: false,
1928
1739
  supportsPromptCache: false,
1929
- supportsNativeTools: true,
1930
- defaultToolProtocol: "native",
1931
1740
  inputPrice: 0,
1932
1741
  outputPrice: 0,
1933
1742
  description: "LongCat Flash Thinking FP8 model with 128K context window, optimized for complex reasoning and coding tasks."
@@ -1937,8 +1746,6 @@ var chutesModels = {
1937
1746
  contextWindow: 262144,
1938
1747
  supportsImages: false,
1939
1748
  supportsPromptCache: false,
1940
- supportsNativeTools: true,
1941
- defaultToolProtocol: "native",
1942
1749
  inputPrice: 0,
1943
1750
  outputPrice: 0,
1944
1751
  description: "Qwen3 Coder 480B A35B Instruct FP8 model, optimized for coding tasks."
@@ -1948,8 +1755,6 @@ var chutesModels = {
1948
1755
  contextWindow: 75e3,
1949
1756
  supportsImages: false,
1950
1757
  supportsPromptCache: false,
1951
- supportsNativeTools: true,
1952
- defaultToolProtocol: "native",
1953
1758
  inputPrice: 0.1481,
1954
1759
  outputPrice: 0.5926,
1955
1760
  description: "Moonshot AI Kimi K2 Instruct model with 75k context window."
@@ -1959,8 +1764,6 @@ var chutesModels = {
1959
1764
  contextWindow: 262144,
1960
1765
  supportsImages: false,
1961
1766
  supportsPromptCache: false,
1962
- supportsNativeTools: true,
1963
- defaultToolProtocol: "native",
1964
1767
  inputPrice: 0.1999,
1965
1768
  outputPrice: 0.8001,
1966
1769
  description: "Moonshot AI Kimi K2 Instruct 0905 model with 256k context window."
@@ -1970,8 +1773,6 @@ var chutesModels = {
1970
1773
  contextWindow: 262144,
1971
1774
  supportsImages: false,
1972
1775
  supportsPromptCache: false,
1973
- supportsNativeTools: true,
1974
- defaultToolProtocol: "native",
1975
1776
  inputPrice: 0.077968332,
1976
1777
  outputPrice: 0.31202496,
1977
1778
  description: "Qwen3 235B A22B Thinking 2507 model with 262K context window."
@@ -1981,8 +1782,6 @@ var chutesModels = {
1981
1782
  contextWindow: 131072,
1982
1783
  supportsImages: false,
1983
1784
  supportsPromptCache: false,
1984
- supportsNativeTools: true,
1985
- defaultToolProtocol: "native",
1986
1785
  inputPrice: 0,
1987
1786
  outputPrice: 0,
1988
1787
  description: "Fast, stable instruction-tuned model optimized for complex tasks, RAG, and tool use without thinking traces."
@@ -1992,8 +1791,6 @@ var chutesModels = {
1992
1791
  contextWindow: 131072,
1993
1792
  supportsImages: false,
1994
1793
  supportsPromptCache: false,
1995
- supportsNativeTools: true,
1996
- defaultToolProtocol: "native",
1997
1794
  inputPrice: 0,
1998
1795
  outputPrice: 0,
1999
1796
  description: "Reasoning-first model with structured thinking traces for multi-step problems, math proofs, and code synthesis."
@@ -2003,8 +1800,6 @@ var chutesModels = {
2003
1800
  contextWindow: 262144,
2004
1801
  supportsImages: true,
2005
1802
  supportsPromptCache: false,
2006
- supportsNativeTools: true,
2007
- defaultToolProtocol: "native",
2008
1803
  inputPrice: 0.16,
2009
1804
  outputPrice: 0.65,
2010
1805
  description: "Qwen3\u2011VL\u2011235B\u2011A22B\u2011Thinking is an open\u2011weight MoE vision\u2011language model (235B total, ~22B activated) optimized for deliberate multi\u2011step reasoning with strong text\u2011image\u2011video understanding and long\u2011context capabilities."
@@ -2012,73 +1807,6 @@ var chutesModels = {
2012
1807
  };
2013
1808
  var chutesDefaultModelInfo = chutesModels[chutesDefaultModelId];
2014
1809
 
2015
- // src/providers/claude-code.ts
2016
- var DATE_SUFFIX_PATTERN = /-\d{8}$/;
2017
- var claudeCodeModels = {
2018
- "claude-haiku-4-5": {
2019
- maxTokens: 32768,
2020
- contextWindow: 2e5,
2021
- supportsImages: true,
2022
- supportsPromptCache: true,
2023
- supportsNativeTools: true,
2024
- defaultToolProtocol: "native",
2025
- supportsReasoningEffort: ["disable", "low", "medium", "high"],
2026
- reasoningEffort: "medium",
2027
- description: "Claude Haiku 4.5 - Fast and efficient with thinking"
2028
- },
2029
- "claude-sonnet-4-5": {
2030
- maxTokens: 32768,
2031
- contextWindow: 2e5,
2032
- supportsImages: true,
2033
- supportsPromptCache: true,
2034
- supportsNativeTools: true,
2035
- defaultToolProtocol: "native",
2036
- supportsReasoningEffort: ["disable", "low", "medium", "high"],
2037
- reasoningEffort: "medium",
2038
- description: "Claude Sonnet 4.5 - Balanced performance with thinking"
2039
- },
2040
- "claude-opus-4-5": {
2041
- maxTokens: 32768,
2042
- contextWindow: 2e5,
2043
- supportsImages: true,
2044
- supportsPromptCache: true,
2045
- supportsNativeTools: true,
2046
- defaultToolProtocol: "native",
2047
- supportsReasoningEffort: ["disable", "low", "medium", "high"],
2048
- reasoningEffort: "medium",
2049
- description: "Claude Opus 4.5 - Most capable with thinking"
2050
- }
2051
- };
2052
- var claudeCodeDefaultModelId = "claude-sonnet-4-5";
2053
- var MODEL_FAMILY_PATTERNS = [
2054
- // Opus models (any version) → claude-opus-4-5
2055
- { pattern: /opus/i, target: "claude-opus-4-5" },
2056
- // Haiku models (any version) → claude-haiku-4-5
2057
- { pattern: /haiku/i, target: "claude-haiku-4-5" },
2058
- // Sonnet models (any version) → claude-sonnet-4-5
2059
- { pattern: /sonnet/i, target: "claude-sonnet-4-5" }
2060
- ];
2061
- function normalizeClaudeCodeModelId(modelId) {
2062
- if (Object.hasOwn(claudeCodeModels, modelId)) {
2063
- return modelId;
2064
- }
2065
- const withoutDate = modelId.replace(DATE_SUFFIX_PATTERN, "");
2066
- if (Object.hasOwn(claudeCodeModels, withoutDate)) {
2067
- return withoutDate;
2068
- }
2069
- for (const { pattern, target } of MODEL_FAMILY_PATTERNS) {
2070
- if (pattern.test(modelId)) {
2071
- return target;
2072
- }
2073
- }
2074
- return claudeCodeDefaultModelId;
2075
- }
2076
- var claudeCodeReasoningConfig = {
2077
- low: { budgetTokens: 16e3 },
2078
- medium: { budgetTokens: 32e3 },
2079
- high: { budgetTokens: 64e3 }
2080
- };
2081
-
2082
1810
  // src/providers/deepseek.ts
2083
1811
  var deepSeekDefaultModelId = "deepseek-chat";
2084
1812
  var deepSeekModels = {
@@ -2088,8 +1816,6 @@ var deepSeekModels = {
2088
1816
  contextWindow: 128e3,
2089
1817
  supportsImages: false,
2090
1818
  supportsPromptCache: true,
2091
- supportsNativeTools: true,
2092
- defaultToolProtocol: "native",
2093
1819
  inputPrice: 0.28,
2094
1820
  // $0.28 per million tokens (cache miss) - Updated Dec 9, 2025
2095
1821
  outputPrice: 0.42,
@@ -2106,8 +1832,6 @@ var deepSeekModels = {
2106
1832
  contextWindow: 128e3,
2107
1833
  supportsImages: false,
2108
1834
  supportsPromptCache: true,
2109
- supportsNativeTools: true,
2110
- defaultToolProtocol: "native",
2111
1835
  preserveReasoning: true,
2112
1836
  inputPrice: 0.28,
2113
1837
  // $0.28 per million tokens (cache miss) - Updated Dec 9, 2025
@@ -2130,8 +1854,6 @@ var doubaoModels = {
2130
1854
  contextWindow: 128e3,
2131
1855
  supportsImages: true,
2132
1856
  supportsPromptCache: true,
2133
- supportsNativeTools: true,
2134
- defaultToolProtocol: "native",
2135
1857
  inputPrice: 1e-4,
2136
1858
  // $0.0001 per million tokens (cache miss)
2137
1859
  outputPrice: 4e-4,
@@ -2147,8 +1869,6 @@ var doubaoModels = {
2147
1869
  contextWindow: 128e3,
2148
1870
  supportsImages: true,
2149
1871
  supportsPromptCache: true,
2150
- supportsNativeTools: true,
2151
- defaultToolProtocol: "native",
2152
1872
  inputPrice: 2e-4,
2153
1873
  // $0.0002 per million tokens
2154
1874
  outputPrice: 8e-4,
@@ -2164,8 +1884,6 @@ var doubaoModels = {
2164
1884
  contextWindow: 128e3,
2165
1885
  supportsImages: true,
2166
1886
  supportsPromptCache: true,
2167
- supportsNativeTools: true,
2168
- defaultToolProtocol: "native",
2169
1887
  inputPrice: 15e-5,
2170
1888
  // $0.00015 per million tokens
2171
1889
  outputPrice: 6e-4,
@@ -2188,7 +1906,6 @@ var featherlessModels = {
2188
1906
  contextWindow: 32678,
2189
1907
  supportsImages: false,
2190
1908
  supportsPromptCache: false,
2191
- supportsNativeTools: true,
2192
1909
  inputPrice: 0,
2193
1910
  outputPrice: 0,
2194
1911
  description: "DeepSeek V3 0324 model."
@@ -2198,7 +1915,6 @@ var featherlessModels = {
2198
1915
  contextWindow: 32678,
2199
1916
  supportsImages: false,
2200
1917
  supportsPromptCache: false,
2201
- supportsNativeTools: true,
2202
1918
  inputPrice: 0,
2203
1919
  outputPrice: 0,
2204
1920
  description: "DeepSeek R1 0528 model."
@@ -2208,7 +1924,6 @@ var featherlessModels = {
2208
1924
  contextWindow: 32678,
2209
1925
  supportsImages: false,
2210
1926
  supportsPromptCache: false,
2211
- supportsNativeTools: true,
2212
1927
  inputPrice: 0,
2213
1928
  outputPrice: 0,
2214
1929
  description: "Kimi K2 Instruct model."
@@ -2218,7 +1933,6 @@ var featherlessModels = {
2218
1933
  contextWindow: 32678,
2219
1934
  supportsImages: false,
2220
1935
  supportsPromptCache: false,
2221
- supportsNativeTools: true,
2222
1936
  inputPrice: 0,
2223
1937
  outputPrice: 0,
2224
1938
  description: "GPT-OSS 120B model."
@@ -2228,7 +1942,6 @@ var featherlessModels = {
2228
1942
  contextWindow: 32678,
2229
1943
  supportsImages: false,
2230
1944
  supportsPromptCache: false,
2231
- supportsNativeTools: true,
2232
1945
  inputPrice: 0,
2233
1946
  outputPrice: 0,
2234
1947
  description: "Qwen3 Coder 480B A35B Instruct model."
@@ -2244,8 +1957,6 @@ var fireworksModels = {
2244
1957
  contextWindow: 262144,
2245
1958
  supportsImages: false,
2246
1959
  supportsPromptCache: true,
2247
- supportsNativeTools: true,
2248
- defaultToolProtocol: "native",
2249
1960
  inputPrice: 0.6,
2250
1961
  outputPrice: 2.5,
2251
1962
  cacheReadsPrice: 0.15,
@@ -2256,8 +1967,6 @@ var fireworksModels = {
2256
1967
  contextWindow: 128e3,
2257
1968
  supportsImages: false,
2258
1969
  supportsPromptCache: false,
2259
- supportsNativeTools: true,
2260
- defaultToolProtocol: "native",
2261
1970
  inputPrice: 0.6,
2262
1971
  outputPrice: 2.5,
2263
1972
  description: "Kimi K2 is a state-of-the-art mixture-of-experts (MoE) language model with 32 billion activated parameters and 1 trillion total parameters. Trained with the Muon optimizer, Kimi K2 achieves exceptional performance across frontier knowledge, reasoning, and coding tasks while being meticulously optimized for agentic capabilities."
@@ -2267,7 +1976,6 @@ var fireworksModels = {
2267
1976
  contextWindow: 256e3,
2268
1977
  supportsImages: false,
2269
1978
  supportsPromptCache: true,
2270
- supportsNativeTools: true,
2271
1979
  supportsTemperature: true,
2272
1980
  preserveReasoning: true,
2273
1981
  defaultTemperature: 1,
@@ -2281,8 +1989,6 @@ var fireworksModels = {
2281
1989
  contextWindow: 204800,
2282
1990
  supportsImages: false,
2283
1991
  supportsPromptCache: false,
2284
- supportsNativeTools: true,
2285
- defaultToolProtocol: "native",
2286
1992
  inputPrice: 0.3,
2287
1993
  outputPrice: 1.2,
2288
1994
  description: "MiniMax M2 is a high-performance language model with 204.8K context window, optimized for long-context understanding and generation tasks."
@@ -2292,8 +1998,6 @@ var fireworksModels = {
2292
1998
  contextWindow: 256e3,
2293
1999
  supportsImages: false,
2294
2000
  supportsPromptCache: false,
2295
- supportsNativeTools: true,
2296
- defaultToolProtocol: "native",
2297
2001
  inputPrice: 0.22,
2298
2002
  outputPrice: 0.88,
2299
2003
  description: "Latest Qwen3 thinking model, competitive against the best closed source models in Jul 2025."
@@ -2303,8 +2007,6 @@ var fireworksModels = {
2303
2007
  contextWindow: 256e3,
2304
2008
  supportsImages: false,
2305
2009
  supportsPromptCache: false,
2306
- supportsNativeTools: true,
2307
- defaultToolProtocol: "native",
2308
2010
  inputPrice: 0.45,
2309
2011
  outputPrice: 1.8,
2310
2012
  description: "Qwen3's most agentic code model to date."
@@ -2314,8 +2016,6 @@ var fireworksModels = {
2314
2016
  contextWindow: 16e4,
2315
2017
  supportsImages: false,
2316
2018
  supportsPromptCache: false,
2317
- supportsNativeTools: true,
2318
- defaultToolProtocol: "native",
2319
2019
  inputPrice: 3,
2320
2020
  outputPrice: 8,
2321
2021
  description: "05/28 updated checkpoint of Deepseek R1. Its overall performance is now approaching that of leading models, such as O3 and Gemini 2.5 Pro. Compared to the previous version, the upgraded model shows significant improvements in handling complex reasoning tasks, and this version also offers a reduced hallucination rate, enhanced support for function calling, and better experience for vibe coding. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
@@ -2325,8 +2025,6 @@ var fireworksModels = {
2325
2025
  contextWindow: 128e3,
2326
2026
  supportsImages: false,
2327
2027
  supportsPromptCache: false,
2328
- supportsNativeTools: true,
2329
- defaultToolProtocol: "native",
2330
2028
  inputPrice: 0.9,
2331
2029
  outputPrice: 0.9,
2332
2030
  description: "A strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token from Deepseek. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
@@ -2336,8 +2034,6 @@ var fireworksModels = {
2336
2034
  contextWindow: 163840,
2337
2035
  supportsImages: false,
2338
2036
  supportsPromptCache: false,
2339
- supportsNativeTools: true,
2340
- defaultToolProtocol: "native",
2341
2037
  inputPrice: 0.56,
2342
2038
  outputPrice: 1.68,
2343
2039
  description: "DeepSeek v3.1 is an improved version of the v3 model with enhanced performance, better reasoning capabilities, and improved code generation. This Mixture-of-Experts (MoE) model maintains the same 671B total parameters with 37B activated per token."
@@ -2347,8 +2043,6 @@ var fireworksModels = {
2347
2043
  contextWindow: 128e3,
2348
2044
  supportsImages: false,
2349
2045
  supportsPromptCache: false,
2350
- supportsNativeTools: true,
2351
- defaultToolProtocol: "native",
2352
2046
  inputPrice: 0.55,
2353
2047
  outputPrice: 2.19,
2354
2048
  description: "Z.ai GLM-4.5 with 355B total parameters and 32B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
@@ -2358,8 +2052,6 @@ var fireworksModels = {
2358
2052
  contextWindow: 128e3,
2359
2053
  supportsImages: false,
2360
2054
  supportsPromptCache: false,
2361
- supportsNativeTools: true,
2362
- defaultToolProtocol: "native",
2363
2055
  inputPrice: 0.55,
2364
2056
  outputPrice: 2.19,
2365
2057
  description: "Z.ai GLM-4.5-Air with 106B total parameters and 12B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
@@ -2369,8 +2061,6 @@ var fireworksModels = {
2369
2061
  contextWindow: 198e3,
2370
2062
  supportsImages: false,
2371
2063
  supportsPromptCache: false,
2372
- supportsNativeTools: true,
2373
- defaultToolProtocol: "native",
2374
2064
  inputPrice: 0.55,
2375
2065
  outputPrice: 2.19,
2376
2066
  description: "Z.ai GLM-4.6 is an advanced coding model with exceptional performance on complex programming tasks. Features improved reasoning capabilities and enhanced code generation quality, making it ideal for software development workflows."
@@ -2380,8 +2070,6 @@ var fireworksModels = {
2380
2070
  contextWindow: 128e3,
2381
2071
  supportsImages: false,
2382
2072
  supportsPromptCache: false,
2383
- supportsNativeTools: true,
2384
- defaultToolProtocol: "native",
2385
2073
  inputPrice: 0.07,
2386
2074
  outputPrice: 0.3,
2387
2075
  description: "OpenAI gpt-oss-20b: Compact model for local/edge deployments. Optimized for low-latency and resource-constrained environments with chain-of-thought output, adjustable reasoning, and agentic workflows."
@@ -2391,11 +2079,63 @@ var fireworksModels = {
2391
2079
  contextWindow: 128e3,
2392
2080
  supportsImages: false,
2393
2081
  supportsPromptCache: false,
2394
- supportsNativeTools: true,
2395
- defaultToolProtocol: "native",
2396
2082
  inputPrice: 0.15,
2397
2083
  outputPrice: 0.6,
2398
2084
  description: "OpenAI gpt-oss-120b: Production-grade, general-purpose model that fits on a single H100 GPU. Features complex reasoning, configurable effort, full chain-of-thought transparency, and supports function calling, tool use, and structured outputs."
2085
+ },
2086
+ "accounts/fireworks/models/minimax-m2p1": {
2087
+ maxTokens: 4096,
2088
+ contextWindow: 204800,
2089
+ supportsImages: false,
2090
+ supportsPromptCache: false,
2091
+ inputPrice: 0.3,
2092
+ outputPrice: 1.2,
2093
+ description: "MiniMax M2.1 is an upgraded version of M2 with improved performance on complex reasoning, coding, and long-context understanding tasks."
2094
+ },
2095
+ "accounts/fireworks/models/deepseek-v3p2": {
2096
+ maxTokens: 16384,
2097
+ contextWindow: 163840,
2098
+ supportsImages: false,
2099
+ supportsPromptCache: false,
2100
+ inputPrice: 0.56,
2101
+ outputPrice: 1.68,
2102
+ description: "DeepSeek V3.2 is the latest iteration of the V3 model family with enhanced reasoning capabilities, improved code generation, and better instruction following."
2103
+ },
2104
+ "accounts/fireworks/models/glm-4p7": {
2105
+ maxTokens: 25344,
2106
+ contextWindow: 198e3,
2107
+ supportsImages: false,
2108
+ supportsPromptCache: false,
2109
+ inputPrice: 0.55,
2110
+ outputPrice: 2.19,
2111
+ description: "Z.ai GLM-4.7 is the latest coding model with exceptional performance on complex programming tasks. Features improved reasoning capabilities and enhanced code generation quality."
2112
+ },
2113
+ "accounts/fireworks/models/llama-v3p3-70b-instruct": {
2114
+ maxTokens: 16384,
2115
+ contextWindow: 131072,
2116
+ supportsImages: false,
2117
+ supportsPromptCache: false,
2118
+ inputPrice: 0.9,
2119
+ outputPrice: 0.9,
2120
+ description: "Meta Llama 3.3 70B Instruct is a highly capable instruction-tuned model with strong reasoning, coding, and general task performance."
2121
+ },
2122
+ "accounts/fireworks/models/llama4-maverick-instruct-basic": {
2123
+ maxTokens: 16384,
2124
+ contextWindow: 131072,
2125
+ supportsImages: true,
2126
+ supportsPromptCache: false,
2127
+ inputPrice: 0.22,
2128
+ outputPrice: 0.88,
2129
+ description: "Llama 4 Maverick is Meta's latest multimodal model with vision capabilities, optimized for instruction following and coding tasks."
2130
+ },
2131
+ "accounts/fireworks/models/llama4-scout-instruct-basic": {
2132
+ maxTokens: 16384,
2133
+ contextWindow: 131072,
2134
+ supportsImages: true,
2135
+ supportsPromptCache: false,
2136
+ inputPrice: 0.15,
2137
+ outputPrice: 0.6,
2138
+ description: "Llama 4 Scout is a smaller, faster variant of Llama 4 with multimodal capabilities, ideal for quick iterations and cost-effective deployments."
2399
2139
  }
2400
2140
  };
2401
2141
 
@@ -2406,8 +2146,6 @@ var geminiModels = {
2406
2146
  maxTokens: 65536,
2407
2147
  contextWindow: 1048576,
2408
2148
  supportsImages: true,
2409
- supportsNativeTools: true,
2410
- defaultToolProtocol: "native",
2411
2149
  supportsPromptCache: true,
2412
2150
  supportsReasoningEffort: ["low", "high"],
2413
2151
  reasoningEffort: "low",
@@ -2415,16 +2153,19 @@ var geminiModels = {
2415
2153
  defaultTemperature: 1,
2416
2154
  inputPrice: 4,
2417
2155
  outputPrice: 18,
2156
+ cacheReadsPrice: 0.4,
2418
2157
  tiers: [
2419
2158
  {
2420
2159
  contextWindow: 2e5,
2421
2160
  inputPrice: 2,
2422
- outputPrice: 12
2161
+ outputPrice: 12,
2162
+ cacheReadsPrice: 0.2
2423
2163
  },
2424
2164
  {
2425
2165
  contextWindow: Infinity,
2426
2166
  inputPrice: 4,
2427
- outputPrice: 18
2167
+ outputPrice: 18,
2168
+ cacheReadsPrice: 0.4
2428
2169
  }
2429
2170
  ]
2430
2171
  },
@@ -2432,25 +2173,20 @@ var geminiModels = {
2432
2173
  maxTokens: 65536,
2433
2174
  contextWindow: 1048576,
2434
2175
  supportsImages: true,
2435
- supportsNativeTools: true,
2436
- defaultToolProtocol: "native",
2437
2176
  supportsPromptCache: true,
2438
2177
  supportsReasoningEffort: ["minimal", "low", "medium", "high"],
2439
2178
  reasoningEffort: "medium",
2440
2179
  supportsTemperature: true,
2441
2180
  defaultTemperature: 1,
2442
- inputPrice: 0.3,
2443
- outputPrice: 2.5,
2444
- cacheReadsPrice: 0.075,
2445
- cacheWritesPrice: 1
2181
+ inputPrice: 0.5,
2182
+ outputPrice: 3,
2183
+ cacheReadsPrice: 0.05
2446
2184
  },
2447
2185
  // 2.5 Pro models
2448
2186
  "gemini-2.5-pro": {
2449
2187
  maxTokens: 64e3,
2450
2188
  contextWindow: 1048576,
2451
2189
  supportsImages: true,
2452
- supportsNativeTools: true,
2453
- defaultToolProtocol: "native",
2454
2190
  supportsPromptCache: true,
2455
2191
  inputPrice: 2.5,
2456
2192
  // This is the pricing for prompts above 200k tokens.
@@ -2479,8 +2215,6 @@ var geminiModels = {
2479
2215
  maxTokens: 65535,
2480
2216
  contextWindow: 1048576,
2481
2217
  supportsImages: true,
2482
- supportsNativeTools: true,
2483
- defaultToolProtocol: "native",
2484
2218
  supportsPromptCache: true,
2485
2219
  inputPrice: 2.5,
2486
2220
  // This is the pricing for prompts above 200k tokens.
@@ -2508,8 +2242,6 @@ var geminiModels = {
2508
2242
  maxTokens: 65535,
2509
2243
  contextWindow: 1048576,
2510
2244
  supportsImages: true,
2511
- supportsNativeTools: true,
2512
- defaultToolProtocol: "native",
2513
2245
  supportsPromptCache: true,
2514
2246
  inputPrice: 2.5,
2515
2247
  // This is the pricing for prompts above 200k tokens.
@@ -2535,8 +2267,6 @@ var geminiModels = {
2535
2267
  maxTokens: 65535,
2536
2268
  contextWindow: 1048576,
2537
2269
  supportsImages: true,
2538
- supportsNativeTools: true,
2539
- defaultToolProtocol: "native",
2540
2270
  supportsPromptCache: true,
2541
2271
  inputPrice: 2.5,
2542
2272
  // This is the pricing for prompts above 200k tokens.
@@ -2565,8 +2295,6 @@ var geminiModels = {
2565
2295
  maxTokens: 65536,
2566
2296
  contextWindow: 1048576,
2567
2297
  supportsImages: true,
2568
- supportsNativeTools: true,
2569
- defaultToolProtocol: "native",
2570
2298
  supportsPromptCache: true,
2571
2299
  inputPrice: 0.3,
2572
2300
  outputPrice: 2.5,
@@ -2579,8 +2307,6 @@ var geminiModels = {
2579
2307
  maxTokens: 65536,
2580
2308
  contextWindow: 1048576,
2581
2309
  supportsImages: true,
2582
- supportsNativeTools: true,
2583
- defaultToolProtocol: "native",
2584
2310
  supportsPromptCache: true,
2585
2311
  inputPrice: 0.3,
2586
2312
  outputPrice: 2.5,
@@ -2593,8 +2319,6 @@ var geminiModels = {
2593
2319
  maxTokens: 64e3,
2594
2320
  contextWindow: 1048576,
2595
2321
  supportsImages: true,
2596
- supportsNativeTools: true,
2597
- defaultToolProtocol: "native",
2598
2322
  supportsPromptCache: true,
2599
2323
  inputPrice: 0.3,
2600
2324
  outputPrice: 2.5,
@@ -2608,8 +2332,6 @@ var geminiModels = {
2608
2332
  maxTokens: 65536,
2609
2333
  contextWindow: 1048576,
2610
2334
  supportsImages: true,
2611
- supportsNativeTools: true,
2612
- defaultToolProtocol: "native",
2613
2335
  supportsPromptCache: true,
2614
2336
  inputPrice: 0.1,
2615
2337
  outputPrice: 0.4,
@@ -2622,8 +2344,6 @@ var geminiModels = {
2622
2344
  maxTokens: 65536,
2623
2345
  contextWindow: 1048576,
2624
2346
  supportsImages: true,
2625
- supportsNativeTools: true,
2626
- defaultToolProtocol: "native",
2627
2347
  supportsPromptCache: true,
2628
2348
  inputPrice: 0.1,
2629
2349
  outputPrice: 0.4,
@@ -2643,8 +2363,6 @@ var groqModels = {
2643
2363
  contextWindow: 131072,
2644
2364
  supportsImages: false,
2645
2365
  supportsPromptCache: false,
2646
- supportsNativeTools: true,
2647
- defaultToolProtocol: "native",
2648
2366
  inputPrice: 0.05,
2649
2367
  outputPrice: 0.08,
2650
2368
  description: "Meta Llama 3.1 8B Instant model, 128K context."
@@ -2654,8 +2372,6 @@ var groqModels = {
2654
2372
  contextWindow: 131072,
2655
2373
  supportsImages: false,
2656
2374
  supportsPromptCache: false,
2657
- supportsNativeTools: true,
2658
- defaultToolProtocol: "native",
2659
2375
  inputPrice: 0.59,
2660
2376
  outputPrice: 0.79,
2661
2377
  description: "Meta Llama 3.3 70B Versatile model, 128K context."
@@ -2665,8 +2381,6 @@ var groqModels = {
2665
2381
  contextWindow: 131072,
2666
2382
  supportsImages: false,
2667
2383
  supportsPromptCache: false,
2668
- supportsNativeTools: true,
2669
- defaultToolProtocol: "native",
2670
2384
  inputPrice: 0.11,
2671
2385
  outputPrice: 0.34,
2672
2386
  description: "Meta Llama 4 Scout 17B Instruct model, 128K context."
@@ -2676,8 +2390,6 @@ var groqModels = {
2676
2390
  contextWindow: 131072,
2677
2391
  supportsImages: false,
2678
2392
  supportsPromptCache: false,
2679
- supportsNativeTools: true,
2680
- defaultToolProtocol: "native",
2681
2393
  inputPrice: 0.29,
2682
2394
  outputPrice: 0.59,
2683
2395
  description: "Alibaba Qwen 3 32B model, 128K context."
@@ -2687,8 +2399,6 @@ var groqModels = {
2687
2399
  contextWindow: 262144,
2688
2400
  supportsImages: false,
2689
2401
  supportsPromptCache: true,
2690
- supportsNativeTools: true,
2691
- defaultToolProtocol: "native",
2692
2402
  inputPrice: 0.6,
2693
2403
  outputPrice: 2.5,
2694
2404
  cacheReadsPrice: 0.15,
@@ -2699,8 +2409,6 @@ var groqModels = {
2699
2409
  contextWindow: 131072,
2700
2410
  supportsImages: false,
2701
2411
  supportsPromptCache: false,
2702
- supportsNativeTools: true,
2703
- defaultToolProtocol: "native",
2704
2412
  inputPrice: 0.15,
2705
2413
  outputPrice: 0.75,
2706
2414
  description: "GPT-OSS 120B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 128 experts."
@@ -2710,8 +2418,6 @@ var groqModels = {
2710
2418
  contextWindow: 131072,
2711
2419
  supportsImages: false,
2712
2420
  supportsPromptCache: false,
2713
- supportsNativeTools: true,
2714
- defaultToolProtocol: "native",
2715
2421
  inputPrice: 0.1,
2716
2422
  outputPrice: 0.5,
2717
2423
  description: "GPT-OSS 20B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 32 experts."
@@ -2738,7 +2444,6 @@ var ioIntelligenceModels = {
2738
2444
  contextWindow: 128e3,
2739
2445
  supportsImages: false,
2740
2446
  supportsPromptCache: false,
2741
- supportsNativeTools: true,
2742
2447
  description: "DeepSeek R1 reasoning model"
2743
2448
  },
2744
2449
  "meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8": {
@@ -2746,7 +2451,6 @@ var ioIntelligenceModels = {
2746
2451
  contextWindow: 43e4,
2747
2452
  supportsImages: true,
2748
2453
  supportsPromptCache: false,
2749
- supportsNativeTools: true,
2750
2454
  description: "Llama 4 Maverick 17B model"
2751
2455
  },
2752
2456
  "Intel/Qwen3-Coder-480B-A35B-Instruct-int4-mixed-ar": {
@@ -2754,7 +2458,6 @@ var ioIntelligenceModels = {
2754
2458
  contextWindow: 106e3,
2755
2459
  supportsImages: false,
2756
2460
  supportsPromptCache: false,
2757
- supportsNativeTools: true,
2758
2461
  description: "Qwen3 Coder 480B specialized for coding"
2759
2462
  },
2760
2463
  "openai/gpt-oss-120b": {
@@ -2762,7 +2465,6 @@ var ioIntelligenceModels = {
2762
2465
  contextWindow: 131072,
2763
2466
  supportsImages: false,
2764
2467
  supportsPromptCache: false,
2765
- supportsNativeTools: true,
2766
2468
  description: "OpenAI GPT-OSS 120B model"
2767
2469
  }
2768
2470
  };
@@ -2774,8 +2476,6 @@ var litellmDefaultModelInfo = {
2774
2476
  contextWindow: 2e5,
2775
2477
  supportsImages: true,
2776
2478
  supportsPromptCache: true,
2777
- supportsNativeTools: true,
2778
- defaultToolProtocol: "native",
2779
2479
  inputPrice: 3,
2780
2480
  outputPrice: 15,
2781
2481
  cacheWritesPrice: 3.75,
@@ -2790,8 +2490,6 @@ var lMStudioDefaultModelInfo = {
2790
2490
  contextWindow: 2e5,
2791
2491
  supportsImages: true,
2792
2492
  supportsPromptCache: true,
2793
- supportsNativeTools: true,
2794
- defaultToolProtocol: "native",
2795
2493
  inputPrice: 0,
2796
2494
  outputPrice: 0,
2797
2495
  cacheWritesPrice: 0,
@@ -2807,8 +2505,6 @@ var mistralModels = {
2807
2505
  contextWindow: 128e3,
2808
2506
  supportsImages: true,
2809
2507
  supportsPromptCache: false,
2810
- supportsNativeTools: true,
2811
- defaultToolProtocol: "native",
2812
2508
  inputPrice: 2,
2813
2509
  outputPrice: 5
2814
2510
  },
@@ -2817,8 +2513,6 @@ var mistralModels = {
2817
2513
  contextWindow: 131e3,
2818
2514
  supportsImages: true,
2819
2515
  supportsPromptCache: false,
2820
- supportsNativeTools: true,
2821
- defaultToolProtocol: "native",
2822
2516
  inputPrice: 0.4,
2823
2517
  outputPrice: 2
2824
2518
  },
@@ -2827,8 +2521,6 @@ var mistralModels = {
2827
2521
  contextWindow: 131e3,
2828
2522
  supportsImages: true,
2829
2523
  supportsPromptCache: false,
2830
- supportsNativeTools: true,
2831
- defaultToolProtocol: "native",
2832
2524
  inputPrice: 0.4,
2833
2525
  outputPrice: 2
2834
2526
  },
@@ -2837,8 +2529,6 @@ var mistralModels = {
2837
2529
  contextWindow: 256e3,
2838
2530
  supportsImages: false,
2839
2531
  supportsPromptCache: false,
2840
- supportsNativeTools: true,
2841
- defaultToolProtocol: "native",
2842
2532
  inputPrice: 0.3,
2843
2533
  outputPrice: 0.9
2844
2534
  },
@@ -2847,8 +2537,6 @@ var mistralModels = {
2847
2537
  contextWindow: 131e3,
2848
2538
  supportsImages: false,
2849
2539
  supportsPromptCache: false,
2850
- supportsNativeTools: true,
2851
- defaultToolProtocol: "native",
2852
2540
  inputPrice: 2,
2853
2541
  outputPrice: 6
2854
2542
  },
@@ -2857,8 +2545,6 @@ var mistralModels = {
2857
2545
  contextWindow: 131e3,
2858
2546
  supportsImages: false,
2859
2547
  supportsPromptCache: false,
2860
- supportsNativeTools: true,
2861
- defaultToolProtocol: "native",
2862
2548
  inputPrice: 0.1,
2863
2549
  outputPrice: 0.1
2864
2550
  },
@@ -2867,8 +2553,6 @@ var mistralModels = {
2867
2553
  contextWindow: 131e3,
2868
2554
  supportsImages: false,
2869
2555
  supportsPromptCache: false,
2870
- supportsNativeTools: true,
2871
- defaultToolProtocol: "native",
2872
2556
  inputPrice: 0.04,
2873
2557
  outputPrice: 0.04
2874
2558
  },
@@ -2877,8 +2561,6 @@ var mistralModels = {
2877
2561
  contextWindow: 32e3,
2878
2562
  supportsImages: false,
2879
2563
  supportsPromptCache: false,
2880
- supportsNativeTools: true,
2881
- defaultToolProtocol: "native",
2882
2564
  inputPrice: 0.2,
2883
2565
  outputPrice: 0.6
2884
2566
  },
@@ -2887,8 +2569,6 @@ var mistralModels = {
2887
2569
  contextWindow: 131e3,
2888
2570
  supportsImages: true,
2889
2571
  supportsPromptCache: false,
2890
- supportsNativeTools: true,
2891
- defaultToolProtocol: "native",
2892
2572
  inputPrice: 2,
2893
2573
  outputPrice: 6
2894
2574
  }
@@ -2903,8 +2583,6 @@ var moonshotModels = {
2903
2583
  contextWindow: 131072,
2904
2584
  supportsImages: false,
2905
2585
  supportsPromptCache: true,
2906
- supportsNativeTools: true,
2907
- defaultToolProtocol: "native",
2908
2586
  inputPrice: 0.6,
2909
2587
  // $0.60 per million tokens (cache miss)
2910
2588
  outputPrice: 2.5,
@@ -2920,8 +2598,6 @@ var moonshotModels = {
2920
2598
  contextWindow: 262144,
2921
2599
  supportsImages: false,
2922
2600
  supportsPromptCache: true,
2923
- supportsNativeTools: true,
2924
- defaultToolProtocol: "native",
2925
2601
  inputPrice: 0.6,
2926
2602
  outputPrice: 2.5,
2927
2603
  cacheReadsPrice: 0.15,
@@ -2932,8 +2608,6 @@ var moonshotModels = {
2932
2608
  contextWindow: 262144,
2933
2609
  supportsImages: false,
2934
2610
  supportsPromptCache: true,
2935
- supportsNativeTools: true,
2936
- defaultToolProtocol: "native",
2937
2611
  inputPrice: 2.4,
2938
2612
  // $2.40 per million tokens (cache miss)
2939
2613
  outputPrice: 10,
@@ -2952,8 +2626,6 @@ var moonshotModels = {
2952
2626
  supportsImages: false,
2953
2627
  // Text-only (no image/vision support)
2954
2628
  supportsPromptCache: true,
2955
- supportsNativeTools: true,
2956
- defaultToolProtocol: "native",
2957
2629
  inputPrice: 0.6,
2958
2630
  // $0.60 per million tokens (cache miss)
2959
2631
  outputPrice: 2.5,
@@ -2967,6 +2639,21 @@ var moonshotModels = {
2967
2639
  preserveReasoning: true,
2968
2640
  defaultTemperature: 1,
2969
2641
  description: `The kimi-k2-thinking model is a general-purpose agentic reasoning model developed by Moonshot AI. Thanks to its strength in deep reasoning and multi-turn tool use, it can solve even the hardest problems.`
2642
+ },
2643
+ "kimi-k2.5": {
2644
+ maxTokens: 16384,
2645
+ contextWindow: 262144,
2646
+ supportsImages: false,
2647
+ supportsPromptCache: true,
2648
+ inputPrice: 0.6,
2649
+ // $0.60 per million tokens (cache miss)
2650
+ outputPrice: 3,
2651
+ // $3.00 per million tokens
2652
+ cacheReadsPrice: 0.1,
2653
+ // $0.10 per million tokens (cache hit)
2654
+ supportsTemperature: true,
2655
+ defaultTemperature: 1,
2656
+ description: "Kimi K2.5 is the latest generation of Moonshot AI's Kimi series, featuring improved reasoning capabilities and enhanced performance across diverse tasks."
2970
2657
  }
2971
2658
  };
2972
2659
  var MOONSHOT_DEFAULT_TEMPERATURE = 0.6;
@@ -2978,7 +2665,6 @@ var ollamaDefaultModelInfo = {
2978
2665
  contextWindow: 2e5,
2979
2666
  supportsImages: true,
2980
2667
  supportsPromptCache: true,
2981
- supportsNativeTools: true,
2982
2668
  inputPrice: 0,
2983
2669
  outputPrice: 0,
2984
2670
  cacheWritesPrice: 0,
@@ -2992,8 +2678,6 @@ var openAiNativeModels = {
2992
2678
  "gpt-5.1-codex-max": {
2993
2679
  maxTokens: 128e3,
2994
2680
  contextWindow: 4e5,
2995
- supportsNativeTools: true,
2996
- defaultToolProtocol: "native",
2997
2681
  includedTools: ["apply_patch"],
2998
2682
  excludedTools: ["apply_diff", "write_to_file"],
2999
2683
  supportsImages: true,
@@ -3011,8 +2695,6 @@ var openAiNativeModels = {
3011
2695
  "gpt-5.2": {
3012
2696
  maxTokens: 128e3,
3013
2697
  contextWindow: 4e5,
3014
- supportsNativeTools: true,
3015
- defaultToolProtocol: "native",
3016
2698
  includedTools: ["apply_patch"],
3017
2699
  excludedTools: ["apply_diff", "write_to_file"],
3018
2700
  supportsImages: true,
@@ -3034,8 +2716,6 @@ var openAiNativeModels = {
3034
2716
  "gpt-5.2-codex": {
3035
2717
  maxTokens: 128e3,
3036
2718
  contextWindow: 4e5,
3037
- supportsNativeTools: true,
3038
- defaultToolProtocol: "native",
3039
2719
  includedTools: ["apply_patch"],
3040
2720
  excludedTools: ["apply_diff", "write_to_file"],
3041
2721
  supportsImages: true,
@@ -3053,8 +2733,6 @@ var openAiNativeModels = {
3053
2733
  "gpt-5.2-chat-latest": {
3054
2734
  maxTokens: 16384,
3055
2735
  contextWindow: 128e3,
3056
- supportsNativeTools: true,
3057
- defaultToolProtocol: "native",
3058
2736
  includedTools: ["apply_patch"],
3059
2737
  excludedTools: ["apply_diff", "write_to_file"],
3060
2738
  supportsImages: true,
@@ -3067,8 +2745,6 @@ var openAiNativeModels = {
3067
2745
  "gpt-5.1": {
3068
2746
  maxTokens: 128e3,
3069
2747
  contextWindow: 4e5,
3070
- supportsNativeTools: true,
3071
- defaultToolProtocol: "native",
3072
2748
  includedTools: ["apply_patch"],
3073
2749
  excludedTools: ["apply_diff", "write_to_file"],
3074
2750
  supportsImages: true,
@@ -3090,8 +2766,6 @@ var openAiNativeModels = {
3090
2766
  "gpt-5.1-codex": {
3091
2767
  maxTokens: 128e3,
3092
2768
  contextWindow: 4e5,
3093
- supportsNativeTools: true,
3094
- defaultToolProtocol: "native",
3095
2769
  includedTools: ["apply_patch"],
3096
2770
  excludedTools: ["apply_diff", "write_to_file"],
3097
2771
  supportsImages: true,
@@ -3109,8 +2783,6 @@ var openAiNativeModels = {
3109
2783
  "gpt-5.1-codex-mini": {
3110
2784
  maxTokens: 128e3,
3111
2785
  contextWindow: 4e5,
3112
- supportsNativeTools: true,
3113
- defaultToolProtocol: "native",
3114
2786
  includedTools: ["apply_patch"],
3115
2787
  excludedTools: ["apply_diff", "write_to_file"],
3116
2788
  supportsImages: true,
@@ -3127,8 +2799,6 @@ var openAiNativeModels = {
3127
2799
  "gpt-5": {
3128
2800
  maxTokens: 128e3,
3129
2801
  contextWindow: 4e5,
3130
- supportsNativeTools: true,
3131
- defaultToolProtocol: "native",
3132
2802
  includedTools: ["apply_patch"],
3133
2803
  excludedTools: ["apply_diff", "write_to_file"],
3134
2804
  supportsImages: true,
@@ -3149,8 +2819,6 @@ var openAiNativeModels = {
3149
2819
  "gpt-5-mini": {
3150
2820
  maxTokens: 128e3,
3151
2821
  contextWindow: 4e5,
3152
- supportsNativeTools: true,
3153
- defaultToolProtocol: "native",
3154
2822
  includedTools: ["apply_patch"],
3155
2823
  excludedTools: ["apply_diff", "write_to_file"],
3156
2824
  supportsImages: true,
@@ -3171,8 +2839,6 @@ var openAiNativeModels = {
3171
2839
  "gpt-5-codex": {
3172
2840
  maxTokens: 128e3,
3173
2841
  contextWindow: 4e5,
3174
- supportsNativeTools: true,
3175
- defaultToolProtocol: "native",
3176
2842
  includedTools: ["apply_patch"],
3177
2843
  excludedTools: ["apply_diff", "write_to_file"],
3178
2844
  supportsImages: true,
@@ -3189,8 +2855,6 @@ var openAiNativeModels = {
3189
2855
  "gpt-5-nano": {
3190
2856
  maxTokens: 128e3,
3191
2857
  contextWindow: 4e5,
3192
- supportsNativeTools: true,
3193
- defaultToolProtocol: "native",
3194
2858
  includedTools: ["apply_patch"],
3195
2859
  excludedTools: ["apply_diff", "write_to_file"],
3196
2860
  supportsImages: true,
@@ -3208,8 +2872,6 @@ var openAiNativeModels = {
3208
2872
  "gpt-5-chat-latest": {
3209
2873
  maxTokens: 128e3,
3210
2874
  contextWindow: 4e5,
3211
- supportsNativeTools: true,
3212
- defaultToolProtocol: "native",
3213
2875
  includedTools: ["apply_patch"],
3214
2876
  excludedTools: ["apply_diff", "write_to_file"],
3215
2877
  supportsImages: true,
@@ -3222,8 +2884,6 @@ var openAiNativeModels = {
3222
2884
  "gpt-4.1": {
3223
2885
  maxTokens: 32768,
3224
2886
  contextWindow: 1047576,
3225
- supportsNativeTools: true,
3226
- defaultToolProtocol: "native",
3227
2887
  includedTools: ["apply_patch"],
3228
2888
  excludedTools: ["apply_diff", "write_to_file"],
3229
2889
  supportsImages: true,
@@ -3239,8 +2899,6 @@ var openAiNativeModels = {
3239
2899
  "gpt-4.1-mini": {
3240
2900
  maxTokens: 32768,
3241
2901
  contextWindow: 1047576,
3242
- supportsNativeTools: true,
3243
- defaultToolProtocol: "native",
3244
2902
  includedTools: ["apply_patch"],
3245
2903
  excludedTools: ["apply_diff", "write_to_file"],
3246
2904
  supportsImages: true,
@@ -3256,8 +2914,6 @@ var openAiNativeModels = {
3256
2914
  "gpt-4.1-nano": {
3257
2915
  maxTokens: 32768,
3258
2916
  contextWindow: 1047576,
3259
- supportsNativeTools: true,
3260
- defaultToolProtocol: "native",
3261
2917
  includedTools: ["apply_patch"],
3262
2918
  excludedTools: ["apply_diff", "write_to_file"],
3263
2919
  supportsImages: true,
@@ -3273,8 +2929,6 @@ var openAiNativeModels = {
3273
2929
  o3: {
3274
2930
  maxTokens: 1e5,
3275
2931
  contextWindow: 2e5,
3276
- supportsNativeTools: true,
3277
- defaultToolProtocol: "native",
3278
2932
  supportsImages: true,
3279
2933
  supportsPromptCache: true,
3280
2934
  inputPrice: 2,
@@ -3291,8 +2945,6 @@ var openAiNativeModels = {
3291
2945
  "o3-high": {
3292
2946
  maxTokens: 1e5,
3293
2947
  contextWindow: 2e5,
3294
- supportsNativeTools: true,
3295
- defaultToolProtocol: "native",
3296
2948
  supportsImages: true,
3297
2949
  supportsPromptCache: true,
3298
2950
  inputPrice: 2,
@@ -3304,8 +2956,6 @@ var openAiNativeModels = {
3304
2956
  "o3-low": {
3305
2957
  maxTokens: 1e5,
3306
2958
  contextWindow: 2e5,
3307
- supportsNativeTools: true,
3308
- defaultToolProtocol: "native",
3309
2959
  supportsImages: true,
3310
2960
  supportsPromptCache: true,
3311
2961
  inputPrice: 2,
@@ -3317,8 +2967,6 @@ var openAiNativeModels = {
3317
2967
  "o4-mini": {
3318
2968
  maxTokens: 1e5,
3319
2969
  contextWindow: 2e5,
3320
- supportsNativeTools: true,
3321
- defaultToolProtocol: "native",
3322
2970
  supportsImages: true,
3323
2971
  supportsPromptCache: true,
3324
2972
  inputPrice: 1.1,
@@ -3335,8 +2983,6 @@ var openAiNativeModels = {
3335
2983
  "o4-mini-high": {
3336
2984
  maxTokens: 1e5,
3337
2985
  contextWindow: 2e5,
3338
- supportsNativeTools: true,
3339
- defaultToolProtocol: "native",
3340
2986
  supportsImages: true,
3341
2987
  supportsPromptCache: true,
3342
2988
  inputPrice: 1.1,
@@ -3348,8 +2994,6 @@ var openAiNativeModels = {
3348
2994
  "o4-mini-low": {
3349
2995
  maxTokens: 1e5,
3350
2996
  contextWindow: 2e5,
3351
- supportsNativeTools: true,
3352
- defaultToolProtocol: "native",
3353
2997
  supportsImages: true,
3354
2998
  supportsPromptCache: true,
3355
2999
  inputPrice: 1.1,
@@ -3361,8 +3005,6 @@ var openAiNativeModels = {
3361
3005
  "o3-mini": {
3362
3006
  maxTokens: 1e5,
3363
3007
  contextWindow: 2e5,
3364
- supportsNativeTools: true,
3365
- defaultToolProtocol: "native",
3366
3008
  supportsImages: false,
3367
3009
  supportsPromptCache: true,
3368
3010
  inputPrice: 1.1,
@@ -3375,8 +3017,6 @@ var openAiNativeModels = {
3375
3017
  "o3-mini-high": {
3376
3018
  maxTokens: 1e5,
3377
3019
  contextWindow: 2e5,
3378
- supportsNativeTools: true,
3379
- defaultToolProtocol: "native",
3380
3020
  supportsImages: false,
3381
3021
  supportsPromptCache: true,
3382
3022
  inputPrice: 1.1,
@@ -3388,8 +3028,6 @@ var openAiNativeModels = {
3388
3028
  "o3-mini-low": {
3389
3029
  maxTokens: 1e5,
3390
3030
  contextWindow: 2e5,
3391
- supportsNativeTools: true,
3392
- defaultToolProtocol: "native",
3393
3031
  supportsImages: false,
3394
3032
  supportsPromptCache: true,
3395
3033
  inputPrice: 1.1,
@@ -3401,8 +3039,6 @@ var openAiNativeModels = {
3401
3039
  o1: {
3402
3040
  maxTokens: 1e5,
3403
3041
  contextWindow: 2e5,
3404
- supportsNativeTools: true,
3405
- defaultToolProtocol: "native",
3406
3042
  supportsImages: true,
3407
3043
  supportsPromptCache: true,
3408
3044
  inputPrice: 15,
@@ -3413,8 +3049,6 @@ var openAiNativeModels = {
3413
3049
  "o1-preview": {
3414
3050
  maxTokens: 32768,
3415
3051
  contextWindow: 128e3,
3416
- supportsNativeTools: true,
3417
- defaultToolProtocol: "native",
3418
3052
  supportsImages: true,
3419
3053
  supportsPromptCache: true,
3420
3054
  inputPrice: 15,
@@ -3425,8 +3059,6 @@ var openAiNativeModels = {
3425
3059
  "o1-mini": {
3426
3060
  maxTokens: 65536,
3427
3061
  contextWindow: 128e3,
3428
- supportsNativeTools: true,
3429
- defaultToolProtocol: "native",
3430
3062
  supportsImages: true,
3431
3063
  supportsPromptCache: true,
3432
3064
  inputPrice: 1.1,
@@ -3437,8 +3069,6 @@ var openAiNativeModels = {
3437
3069
  "gpt-4o": {
3438
3070
  maxTokens: 16384,
3439
3071
  contextWindow: 128e3,
3440
- supportsNativeTools: true,
3441
- defaultToolProtocol: "native",
3442
3072
  supportsImages: true,
3443
3073
  supportsPromptCache: true,
3444
3074
  inputPrice: 2.5,
@@ -3452,8 +3082,6 @@ var openAiNativeModels = {
3452
3082
  "gpt-4o-mini": {
3453
3083
  maxTokens: 16384,
3454
3084
  contextWindow: 128e3,
3455
- supportsNativeTools: true,
3456
- defaultToolProtocol: "native",
3457
3085
  supportsImages: true,
3458
3086
  supportsPromptCache: true,
3459
3087
  inputPrice: 0.15,
@@ -3467,8 +3095,6 @@ var openAiNativeModels = {
3467
3095
  "codex-mini-latest": {
3468
3096
  maxTokens: 16384,
3469
3097
  contextWindow: 2e5,
3470
- supportsNativeTools: true,
3471
- defaultToolProtocol: "native",
3472
3098
  supportsImages: false,
3473
3099
  supportsPromptCache: false,
3474
3100
  inputPrice: 1.5,
@@ -3481,8 +3107,6 @@ var openAiNativeModels = {
3481
3107
  "gpt-5-2025-08-07": {
3482
3108
  maxTokens: 128e3,
3483
3109
  contextWindow: 4e5,
3484
- supportsNativeTools: true,
3485
- defaultToolProtocol: "native",
3486
3110
  includedTools: ["apply_patch"],
3487
3111
  excludedTools: ["apply_diff", "write_to_file"],
3488
3112
  supportsImages: true,
@@ -3503,8 +3127,6 @@ var openAiNativeModels = {
3503
3127
  "gpt-5-mini-2025-08-07": {
3504
3128
  maxTokens: 128e3,
3505
3129
  contextWindow: 4e5,
3506
- supportsNativeTools: true,
3507
- defaultToolProtocol: "native",
3508
3130
  includedTools: ["apply_patch"],
3509
3131
  excludedTools: ["apply_diff", "write_to_file"],
3510
3132
  supportsImages: true,
@@ -3525,8 +3147,6 @@ var openAiNativeModels = {
3525
3147
  "gpt-5-nano-2025-08-07": {
3526
3148
  maxTokens: 128e3,
3527
3149
  contextWindow: 4e5,
3528
- supportsNativeTools: true,
3529
- defaultToolProtocol: "native",
3530
3150
  includedTools: ["apply_patch"],
3531
3151
  excludedTools: ["apply_diff", "write_to_file"],
3532
3152
  supportsImages: true,
@@ -3548,9 +3168,7 @@ var openAiModelInfoSaneDefaults = {
3548
3168
  supportsImages: true,
3549
3169
  supportsPromptCache: false,
3550
3170
  inputPrice: 0,
3551
- outputPrice: 0,
3552
- supportsNativeTools: true,
3553
- defaultToolProtocol: "native"
3171
+ outputPrice: 0
3554
3172
  };
3555
3173
  var azureOpenAiDefaultApiVersion = "2024-08-01-preview";
3556
3174
  var OPENAI_NATIVE_DEFAULT_TEMPERATURE = 0;
@@ -3562,8 +3180,6 @@ var openAiCodexModels = {
3562
3180
  "gpt-5.1-codex-max": {
3563
3181
  maxTokens: 128e3,
3564
3182
  contextWindow: 4e5,
3565
- supportsNativeTools: true,
3566
- defaultToolProtocol: "native",
3567
3183
  includedTools: ["apply_patch"],
3568
3184
  excludedTools: ["apply_diff", "write_to_file"],
3569
3185
  supportsImages: true,
@@ -3576,11 +3192,24 @@ var openAiCodexModels = {
3576
3192
  supportsTemperature: false,
3577
3193
  description: "GPT-5.1 Codex Max: Maximum capability coding model via ChatGPT subscription"
3578
3194
  },
3195
+ "gpt-5.1-codex": {
3196
+ maxTokens: 128e3,
3197
+ contextWindow: 4e5,
3198
+ includedTools: ["apply_patch"],
3199
+ excludedTools: ["apply_diff", "write_to_file"],
3200
+ supportsImages: true,
3201
+ supportsPromptCache: true,
3202
+ supportsReasoningEffort: ["low", "medium", "high"],
3203
+ reasoningEffort: "medium",
3204
+ // Subscription-based: no per-token costs
3205
+ inputPrice: 0,
3206
+ outputPrice: 0,
3207
+ supportsTemperature: false,
3208
+ description: "GPT-5.1 Codex: GPT-5.1 optimized for agentic coding via ChatGPT subscription"
3209
+ },
3579
3210
  "gpt-5.2-codex": {
3580
3211
  maxTokens: 128e3,
3581
3212
  contextWindow: 4e5,
3582
- supportsNativeTools: true,
3583
- defaultToolProtocol: "native",
3584
3213
  includedTools: ["apply_patch"],
3585
3214
  excludedTools: ["apply_diff", "write_to_file"],
3586
3215
  supportsImages: true,
@@ -3592,11 +3221,71 @@ var openAiCodexModels = {
3592
3221
  supportsTemperature: false,
3593
3222
  description: "GPT-5.2 Codex: OpenAI's flagship coding model via ChatGPT subscription"
3594
3223
  },
3224
+ "gpt-5.1": {
3225
+ maxTokens: 128e3,
3226
+ contextWindow: 4e5,
3227
+ includedTools: ["apply_patch"],
3228
+ excludedTools: ["apply_diff", "write_to_file"],
3229
+ supportsImages: true,
3230
+ supportsPromptCache: true,
3231
+ supportsReasoningEffort: ["none", "low", "medium", "high"],
3232
+ reasoningEffort: "medium",
3233
+ // Subscription-based: no per-token costs
3234
+ inputPrice: 0,
3235
+ outputPrice: 0,
3236
+ supportsVerbosity: true,
3237
+ supportsTemperature: false,
3238
+ description: "GPT-5.1: General GPT-5.1 model via ChatGPT subscription"
3239
+ },
3240
+ "gpt-5": {
3241
+ maxTokens: 128e3,
3242
+ contextWindow: 4e5,
3243
+ includedTools: ["apply_patch"],
3244
+ excludedTools: ["apply_diff", "write_to_file"],
3245
+ supportsImages: true,
3246
+ supportsPromptCache: true,
3247
+ supportsReasoningEffort: ["minimal", "low", "medium", "high"],
3248
+ reasoningEffort: "medium",
3249
+ // Subscription-based: no per-token costs
3250
+ inputPrice: 0,
3251
+ outputPrice: 0,
3252
+ supportsVerbosity: true,
3253
+ supportsTemperature: false,
3254
+ description: "GPT-5: General GPT-5 model via ChatGPT subscription"
3255
+ },
3256
+ "gpt-5-codex": {
3257
+ maxTokens: 128e3,
3258
+ contextWindow: 4e5,
3259
+ includedTools: ["apply_patch"],
3260
+ excludedTools: ["apply_diff", "write_to_file"],
3261
+ supportsImages: true,
3262
+ supportsPromptCache: true,
3263
+ supportsReasoningEffort: ["low", "medium", "high"],
3264
+ reasoningEffort: "medium",
3265
+ // Subscription-based: no per-token costs
3266
+ inputPrice: 0,
3267
+ outputPrice: 0,
3268
+ supportsTemperature: false,
3269
+ description: "GPT-5 Codex: GPT-5 optimized for agentic coding via ChatGPT subscription"
3270
+ },
3271
+ "gpt-5-codex-mini": {
3272
+ maxTokens: 128e3,
3273
+ contextWindow: 4e5,
3274
+ includedTools: ["apply_patch"],
3275
+ excludedTools: ["apply_diff", "write_to_file"],
3276
+ supportsImages: true,
3277
+ supportsPromptCache: true,
3278
+ supportsReasoningEffort: ["low", "medium", "high"],
3279
+ reasoningEffort: "medium",
3280
+ // Subscription-based: no per-token costs
3281
+ inputPrice: 0,
3282
+ outputPrice: 0,
3283
+ supportsTemperature: false,
3284
+ description: "GPT-5 Codex Mini: Faster coding model via ChatGPT subscription"
3285
+ },
3595
3286
  "gpt-5.1-codex-mini": {
3596
3287
  maxTokens: 128e3,
3597
3288
  contextWindow: 4e5,
3598
- supportsNativeTools: true,
3599
- defaultToolProtocol: "native",
3600
3289
  includedTools: ["apply_patch"],
3601
3290
  excludedTools: ["apply_diff", "write_to_file"],
3602
3291
  supportsImages: true,
@@ -3611,8 +3300,6 @@ var openAiCodexModels = {
3611
3300
  "gpt-5.2": {
3612
3301
  maxTokens: 128e3,
3613
3302
  contextWindow: 4e5,
3614
- supportsNativeTools: true,
3615
- defaultToolProtocol: "native",
3616
3303
  includedTools: ["apply_patch"],
3617
3304
  excludedTools: ["apply_diff", "write_to_file"],
3618
3305
  supportsImages: true,
@@ -3633,7 +3320,6 @@ var openRouterDefaultModelInfo = {
3633
3320
  contextWindow: 2e5,
3634
3321
  supportsImages: true,
3635
3322
  supportsPromptCache: true,
3636
- supportsNativeTools: true,
3637
3323
  inputPrice: 3,
3638
3324
  outputPrice: 15,
3639
3325
  cacheWritesPrice: 3.75,
@@ -3707,8 +3393,6 @@ var qwenCodeModels = {
3707
3393
  contextWindow: 1e6,
3708
3394
  supportsImages: false,
3709
3395
  supportsPromptCache: false,
3710
- supportsNativeTools: true,
3711
- defaultToolProtocol: "native",
3712
3396
  inputPrice: 0,
3713
3397
  outputPrice: 0,
3714
3398
  cacheWritesPrice: 0,
@@ -3720,8 +3404,6 @@ var qwenCodeModels = {
3720
3404
  contextWindow: 1e6,
3721
3405
  supportsImages: false,
3722
3406
  supportsPromptCache: false,
3723
- supportsNativeTools: true,
3724
- defaultToolProtocol: "native",
3725
3407
  inputPrice: 0,
3726
3408
  outputPrice: 0,
3727
3409
  cacheWritesPrice: 0,
@@ -3737,8 +3419,6 @@ var requestyDefaultModelInfo = {
3737
3419
  contextWindow: 2e5,
3738
3420
  supportsImages: true,
3739
3421
  supportsPromptCache: true,
3740
- supportsNativeTools: true,
3741
- defaultToolProtocol: "native",
3742
3422
  inputPrice: 3,
3743
3423
  outputPrice: 15,
3744
3424
  cacheWritesPrice: 3.75,
@@ -3794,8 +3474,6 @@ var sambaNovaModels = {
3794
3474
  contextWindow: 16384,
3795
3475
  supportsImages: false,
3796
3476
  supportsPromptCache: false,
3797
- supportsNativeTools: true,
3798
- defaultToolProtocol: "native",
3799
3477
  inputPrice: 0.1,
3800
3478
  outputPrice: 0.2,
3801
3479
  description: "Meta Llama 3.1 8B Instruct model with 16K context window."
@@ -3805,8 +3483,6 @@ var sambaNovaModels = {
3805
3483
  contextWindow: 131072,
3806
3484
  supportsImages: false,
3807
3485
  supportsPromptCache: false,
3808
- supportsNativeTools: true,
3809
- defaultToolProtocol: "native",
3810
3486
  inputPrice: 0.6,
3811
3487
  outputPrice: 1.2,
3812
3488
  description: "Meta Llama 3.3 70B Instruct model with 128K context window."
@@ -3817,8 +3493,6 @@ var sambaNovaModels = {
3817
3493
  supportsImages: false,
3818
3494
  supportsPromptCache: false,
3819
3495
  supportsReasoningBudget: true,
3820
- supportsNativeTools: true,
3821
- defaultToolProtocol: "native",
3822
3496
  inputPrice: 5,
3823
3497
  outputPrice: 7,
3824
3498
  description: "DeepSeek R1 reasoning model with 32K context window."
@@ -3828,8 +3502,6 @@ var sambaNovaModels = {
3828
3502
  contextWindow: 32768,
3829
3503
  supportsImages: false,
3830
3504
  supportsPromptCache: false,
3831
- supportsNativeTools: true,
3832
- defaultToolProtocol: "native",
3833
3505
  inputPrice: 3,
3834
3506
  outputPrice: 4.5,
3835
3507
  description: "DeepSeek V3 model with 32K context window."
@@ -3839,8 +3511,6 @@ var sambaNovaModels = {
3839
3511
  contextWindow: 32768,
3840
3512
  supportsImages: false,
3841
3513
  supportsPromptCache: false,
3842
- supportsNativeTools: true,
3843
- defaultToolProtocol: "native",
3844
3514
  inputPrice: 3,
3845
3515
  outputPrice: 4.5,
3846
3516
  description: "DeepSeek V3.1 model with 32K context window."
@@ -3850,8 +3520,6 @@ var sambaNovaModels = {
3850
3520
  contextWindow: 131072,
3851
3521
  supportsImages: true,
3852
3522
  supportsPromptCache: false,
3853
- supportsNativeTools: true,
3854
- defaultToolProtocol: "native",
3855
3523
  inputPrice: 0.63,
3856
3524
  outputPrice: 1.8,
3857
3525
  description: "Meta Llama 4 Maverick 17B 128E Instruct model with 128K context window."
@@ -3861,8 +3529,6 @@ var sambaNovaModels = {
3861
3529
  contextWindow: 8192,
3862
3530
  supportsImages: false,
3863
3531
  supportsPromptCache: false,
3864
- supportsNativeTools: true,
3865
- defaultToolProtocol: "native",
3866
3532
  inputPrice: 0.4,
3867
3533
  outputPrice: 0.8,
3868
3534
  description: "Alibaba Qwen 3 32B model with 8K context window."
@@ -3872,8 +3538,6 @@ var sambaNovaModels = {
3872
3538
  contextWindow: 131072,
3873
3539
  supportsImages: false,
3874
3540
  supportsPromptCache: false,
3875
- supportsNativeTools: true,
3876
- defaultToolProtocol: "native",
3877
3541
  inputPrice: 0.22,
3878
3542
  outputPrice: 0.59,
3879
3543
  description: "OpenAI gpt oss 120b model with 128k context window."
@@ -3887,7 +3551,6 @@ var unboundDefaultModelInfo = {
3887
3551
  contextWindow: 2e5,
3888
3552
  supportsImages: true,
3889
3553
  supportsPromptCache: true,
3890
- supportsNativeTools: true,
3891
3554
  inputPrice: 3,
3892
3555
  outputPrice: 15,
3893
3556
  cacheWritesPrice: 3.75,
@@ -3901,8 +3564,6 @@ var vertexModels = {
3901
3564
  maxTokens: 65536,
3902
3565
  contextWindow: 1048576,
3903
3566
  supportsImages: true,
3904
- supportsNativeTools: true,
3905
- defaultToolProtocol: "native",
3906
3567
  supportsPromptCache: true,
3907
3568
  supportsReasoningEffort: ["low", "high"],
3908
3569
  reasoningEffort: "low",
@@ -3910,16 +3571,19 @@ var vertexModels = {
3910
3571
  defaultTemperature: 1,
3911
3572
  inputPrice: 4,
3912
3573
  outputPrice: 18,
3574
+ cacheReadsPrice: 0.4,
3913
3575
  tiers: [
3914
3576
  {
3915
3577
  contextWindow: 2e5,
3916
3578
  inputPrice: 2,
3917
- outputPrice: 12
3579
+ outputPrice: 12,
3580
+ cacheReadsPrice: 0.2
3918
3581
  },
3919
3582
  {
3920
3583
  contextWindow: Infinity,
3921
3584
  inputPrice: 4,
3922
- outputPrice: 18
3585
+ outputPrice: 18,
3586
+ cacheReadsPrice: 0.4
3923
3587
  }
3924
3588
  ]
3925
3589
  },
@@ -3927,24 +3591,19 @@ var vertexModels = {
3927
3591
  maxTokens: 65536,
3928
3592
  contextWindow: 1048576,
3929
3593
  supportsImages: true,
3930
- supportsNativeTools: true,
3931
- defaultToolProtocol: "native",
3932
3594
  supportsPromptCache: true,
3933
3595
  supportsReasoningEffort: ["minimal", "low", "medium", "high"],
3934
3596
  reasoningEffort: "medium",
3935
3597
  supportsTemperature: true,
3936
3598
  defaultTemperature: 1,
3937
- inputPrice: 0.3,
3938
- outputPrice: 2.5,
3939
- cacheReadsPrice: 0.075,
3940
- cacheWritesPrice: 1
3599
+ inputPrice: 0.5,
3600
+ outputPrice: 3,
3601
+ cacheReadsPrice: 0.05
3941
3602
  },
3942
3603
  "gemini-2.5-flash-preview-05-20:thinking": {
3943
3604
  maxTokens: 65535,
3944
3605
  contextWindow: 1048576,
3945
3606
  supportsImages: true,
3946
- supportsNativeTools: true,
3947
- defaultToolProtocol: "native",
3948
3607
  supportsPromptCache: true,
3949
3608
  inputPrice: 0.15,
3950
3609
  outputPrice: 3.5,
@@ -3956,8 +3615,6 @@ var vertexModels = {
3956
3615
  maxTokens: 65535,
3957
3616
  contextWindow: 1048576,
3958
3617
  supportsImages: true,
3959
- supportsNativeTools: true,
3960
- defaultToolProtocol: "native",
3961
3618
  supportsPromptCache: true,
3962
3619
  inputPrice: 0.15,
3963
3620
  outputPrice: 0.6
@@ -3966,8 +3623,6 @@ var vertexModels = {
3966
3623
  maxTokens: 64e3,
3967
3624
  contextWindow: 1048576,
3968
3625
  supportsImages: true,
3969
- supportsNativeTools: true,
3970
- defaultToolProtocol: "native",
3971
3626
  supportsPromptCache: true,
3972
3627
  inputPrice: 0.3,
3973
3628
  outputPrice: 2.5,
@@ -3980,8 +3635,6 @@ var vertexModels = {
3980
3635
  maxTokens: 65535,
3981
3636
  contextWindow: 1048576,
3982
3637
  supportsImages: true,
3983
- supportsNativeTools: true,
3984
- defaultToolProtocol: "native",
3985
3638
  supportsPromptCache: false,
3986
3639
  inputPrice: 0.15,
3987
3640
  outputPrice: 3.5,
@@ -3993,8 +3646,6 @@ var vertexModels = {
3993
3646
  maxTokens: 65535,
3994
3647
  contextWindow: 1048576,
3995
3648
  supportsImages: true,
3996
- supportsNativeTools: true,
3997
- defaultToolProtocol: "native",
3998
3649
  supportsPromptCache: false,
3999
3650
  inputPrice: 0.15,
4000
3651
  outputPrice: 0.6
@@ -4003,8 +3654,6 @@ var vertexModels = {
4003
3654
  maxTokens: 65535,
4004
3655
  contextWindow: 1048576,
4005
3656
  supportsImages: true,
4006
- supportsNativeTools: true,
4007
- defaultToolProtocol: "native",
4008
3657
  supportsPromptCache: true,
4009
3658
  inputPrice: 2.5,
4010
3659
  outputPrice: 15
@@ -4013,8 +3662,6 @@ var vertexModels = {
4013
3662
  maxTokens: 65535,
4014
3663
  contextWindow: 1048576,
4015
3664
  supportsImages: true,
4016
- supportsNativeTools: true,
4017
- defaultToolProtocol: "native",
4018
3665
  supportsPromptCache: true,
4019
3666
  inputPrice: 2.5,
4020
3667
  outputPrice: 15
@@ -4023,8 +3670,6 @@ var vertexModels = {
4023
3670
  maxTokens: 65535,
4024
3671
  contextWindow: 1048576,
4025
3672
  supportsImages: true,
4026
- supportsNativeTools: true,
4027
- defaultToolProtocol: "native",
4028
3673
  supportsPromptCache: true,
4029
3674
  inputPrice: 2.5,
4030
3675
  outputPrice: 15,
@@ -4035,8 +3680,6 @@ var vertexModels = {
4035
3680
  maxTokens: 64e3,
4036
3681
  contextWindow: 1048576,
4037
3682
  supportsImages: true,
4038
- supportsNativeTools: true,
4039
- defaultToolProtocol: "native",
4040
3683
  supportsPromptCache: true,
4041
3684
  inputPrice: 2.5,
4042
3685
  outputPrice: 15,
@@ -4062,8 +3705,6 @@ var vertexModels = {
4062
3705
  maxTokens: 65535,
4063
3706
  contextWindow: 1048576,
4064
3707
  supportsImages: true,
4065
- supportsNativeTools: true,
4066
- defaultToolProtocol: "native",
4067
3708
  supportsPromptCache: false,
4068
3709
  inputPrice: 0,
4069
3710
  outputPrice: 0
@@ -4072,8 +3713,6 @@ var vertexModels = {
4072
3713
  maxTokens: 8192,
4073
3714
  contextWindow: 2097152,
4074
3715
  supportsImages: true,
4075
- supportsNativeTools: true,
4076
- defaultToolProtocol: "native",
4077
3716
  supportsPromptCache: false,
4078
3717
  inputPrice: 0,
4079
3718
  outputPrice: 0
@@ -4082,8 +3721,6 @@ var vertexModels = {
4082
3721
  maxTokens: 8192,
4083
3722
  contextWindow: 1048576,
4084
3723
  supportsImages: true,
4085
- supportsNativeTools: true,
4086
- defaultToolProtocol: "native",
4087
3724
  supportsPromptCache: true,
4088
3725
  inputPrice: 0.15,
4089
3726
  outputPrice: 0.6
@@ -4092,8 +3729,6 @@ var vertexModels = {
4092
3729
  maxTokens: 8192,
4093
3730
  contextWindow: 1048576,
4094
3731
  supportsImages: true,
4095
- supportsNativeTools: true,
4096
- defaultToolProtocol: "native",
4097
3732
  supportsPromptCache: false,
4098
3733
  inputPrice: 0.075,
4099
3734
  outputPrice: 0.3
@@ -4102,8 +3737,6 @@ var vertexModels = {
4102
3737
  maxTokens: 8192,
4103
3738
  contextWindow: 32768,
4104
3739
  supportsImages: true,
4105
- supportsNativeTools: true,
4106
- defaultToolProtocol: "native",
4107
3740
  supportsPromptCache: false,
4108
3741
  inputPrice: 0,
4109
3742
  outputPrice: 0
@@ -4112,8 +3745,6 @@ var vertexModels = {
4112
3745
  maxTokens: 8192,
4113
3746
  contextWindow: 1048576,
4114
3747
  supportsImages: true,
4115
- supportsNativeTools: true,
4116
- defaultToolProtocol: "native",
4117
3748
  supportsPromptCache: true,
4118
3749
  inputPrice: 0.075,
4119
3750
  outputPrice: 0.3
@@ -4122,8 +3753,6 @@ var vertexModels = {
4122
3753
  maxTokens: 8192,
4123
3754
  contextWindow: 2097152,
4124
3755
  supportsImages: true,
4125
- supportsNativeTools: true,
4126
- defaultToolProtocol: "native",
4127
3756
  supportsPromptCache: false,
4128
3757
  inputPrice: 1.25,
4129
3758
  outputPrice: 5
@@ -4134,8 +3763,6 @@ var vertexModels = {
4134
3763
  // Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
4135
3764
  supportsImages: true,
4136
3765
  supportsPromptCache: true,
4137
- supportsNativeTools: true,
4138
- defaultToolProtocol: "native",
4139
3766
  inputPrice: 3,
4140
3767
  // $3 per million input tokens (≤200K context)
4141
3768
  outputPrice: 15,
@@ -4167,8 +3794,6 @@ var vertexModels = {
4167
3794
  // Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
4168
3795
  supportsImages: true,
4169
3796
  supportsPromptCache: true,
4170
- supportsNativeTools: true,
4171
- defaultToolProtocol: "native",
4172
3797
  inputPrice: 3,
4173
3798
  // $3 per million input tokens (≤200K context)
4174
3799
  outputPrice: 15,
@@ -4199,8 +3824,6 @@ var vertexModels = {
4199
3824
  contextWindow: 2e5,
4200
3825
  supportsImages: true,
4201
3826
  supportsPromptCache: true,
4202
- supportsNativeTools: true,
4203
- defaultToolProtocol: "native",
4204
3827
  inputPrice: 1,
4205
3828
  outputPrice: 5,
4206
3829
  cacheWritesPrice: 1.25,
@@ -4212,8 +3835,6 @@ var vertexModels = {
4212
3835
  contextWindow: 2e5,
4213
3836
  supportsImages: true,
4214
3837
  supportsPromptCache: true,
4215
- supportsNativeTools: true,
4216
- defaultToolProtocol: "native",
4217
3838
  inputPrice: 5,
4218
3839
  outputPrice: 25,
4219
3840
  cacheWritesPrice: 6.25,
@@ -4225,8 +3846,6 @@ var vertexModels = {
4225
3846
  contextWindow: 2e5,
4226
3847
  supportsImages: true,
4227
3848
  supportsPromptCache: true,
4228
- supportsNativeTools: true,
4229
- defaultToolProtocol: "native",
4230
3849
  inputPrice: 15,
4231
3850
  outputPrice: 75,
4232
3851
  cacheWritesPrice: 18.75,
@@ -4238,8 +3857,6 @@ var vertexModels = {
4238
3857
  contextWindow: 2e5,
4239
3858
  supportsImages: true,
4240
3859
  supportsPromptCache: true,
4241
- supportsNativeTools: true,
4242
- defaultToolProtocol: "native",
4243
3860
  inputPrice: 15,
4244
3861
  outputPrice: 75,
4245
3862
  cacheWritesPrice: 18.75,
@@ -4250,8 +3867,6 @@ var vertexModels = {
4250
3867
  contextWindow: 2e5,
4251
3868
  supportsImages: true,
4252
3869
  supportsPromptCache: true,
4253
- supportsNativeTools: true,
4254
- defaultToolProtocol: "native",
4255
3870
  inputPrice: 3,
4256
3871
  outputPrice: 15,
4257
3872
  cacheWritesPrice: 3.75,
@@ -4264,8 +3879,6 @@ var vertexModels = {
4264
3879
  contextWindow: 2e5,
4265
3880
  supportsImages: true,
4266
3881
  supportsPromptCache: true,
4267
- supportsNativeTools: true,
4268
- defaultToolProtocol: "native",
4269
3882
  inputPrice: 3,
4270
3883
  outputPrice: 15,
4271
3884
  cacheWritesPrice: 3.75,
@@ -4276,8 +3889,6 @@ var vertexModels = {
4276
3889
  contextWindow: 2e5,
4277
3890
  supportsImages: true,
4278
3891
  supportsPromptCache: true,
4279
- supportsNativeTools: true,
4280
- defaultToolProtocol: "native",
4281
3892
  inputPrice: 3,
4282
3893
  outputPrice: 15,
4283
3894
  cacheWritesPrice: 3.75,
@@ -4288,8 +3899,6 @@ var vertexModels = {
4288
3899
  contextWindow: 2e5,
4289
3900
  supportsImages: true,
4290
3901
  supportsPromptCache: true,
4291
- supportsNativeTools: true,
4292
- defaultToolProtocol: "native",
4293
3902
  inputPrice: 3,
4294
3903
  outputPrice: 15,
4295
3904
  cacheWritesPrice: 3.75,
@@ -4300,8 +3909,6 @@ var vertexModels = {
4300
3909
  contextWindow: 2e5,
4301
3910
  supportsImages: false,
4302
3911
  supportsPromptCache: true,
4303
- supportsNativeTools: true,
4304
- defaultToolProtocol: "native",
4305
3912
  inputPrice: 1,
4306
3913
  outputPrice: 5,
4307
3914
  cacheWritesPrice: 1.25,
@@ -4312,8 +3919,6 @@ var vertexModels = {
4312
3919
  contextWindow: 2e5,
4313
3920
  supportsImages: true,
4314
3921
  supportsPromptCache: true,
4315
- supportsNativeTools: true,
4316
- defaultToolProtocol: "native",
4317
3922
  inputPrice: 15,
4318
3923
  outputPrice: 75,
4319
3924
  cacheWritesPrice: 18.75,
@@ -4324,8 +3929,6 @@ var vertexModels = {
4324
3929
  contextWindow: 2e5,
4325
3930
  supportsImages: true,
4326
3931
  supportsPromptCache: true,
4327
- supportsNativeTools: true,
4328
- defaultToolProtocol: "native",
4329
3932
  inputPrice: 0.25,
4330
3933
  outputPrice: 1.25,
4331
3934
  cacheWritesPrice: 0.3,
@@ -4335,8 +3938,6 @@ var vertexModels = {
4335
3938
  maxTokens: 64e3,
4336
3939
  contextWindow: 1048576,
4337
3940
  supportsImages: true,
4338
- supportsNativeTools: true,
4339
- defaultToolProtocol: "native",
4340
3941
  supportsPromptCache: true,
4341
3942
  inputPrice: 0.1,
4342
3943
  outputPrice: 0.4,
@@ -4350,7 +3951,6 @@ var vertexModels = {
4350
3951
  contextWindow: 131072,
4351
3952
  supportsImages: false,
4352
3953
  supportsPromptCache: false,
4353
- supportsNativeTools: true,
4354
3954
  inputPrice: 0.35,
4355
3955
  outputPrice: 1.15,
4356
3956
  description: "Meta Llama 4 Maverick 17B Instruct model, 128K context."
@@ -4360,7 +3960,6 @@ var vertexModels = {
4360
3960
  contextWindow: 163840,
4361
3961
  supportsImages: false,
4362
3962
  supportsPromptCache: false,
4363
- supportsNativeTools: true,
4364
3963
  inputPrice: 1.35,
4365
3964
  outputPrice: 5.4,
4366
3965
  description: "DeepSeek R1 (0528). Available in us-central1"
@@ -4370,7 +3969,6 @@ var vertexModels = {
4370
3969
  contextWindow: 163840,
4371
3970
  supportsImages: false,
4372
3971
  supportsPromptCache: false,
4373
- supportsNativeTools: true,
4374
3972
  inputPrice: 0.6,
4375
3973
  outputPrice: 1.7,
4376
3974
  description: "DeepSeek V3.1. Available in us-west2"
@@ -4380,7 +3978,6 @@ var vertexModels = {
4380
3978
  contextWindow: 131072,
4381
3979
  supportsImages: false,
4382
3980
  supportsPromptCache: false,
4383
- supportsNativeTools: true,
4384
3981
  inputPrice: 0.15,
4385
3982
  outputPrice: 0.6,
4386
3983
  description: "OpenAI gpt-oss 120B. Available in us-central1"
@@ -4390,7 +3987,6 @@ var vertexModels = {
4390
3987
  contextWindow: 131072,
4391
3988
  supportsImages: false,
4392
3989
  supportsPromptCache: false,
4393
- supportsNativeTools: true,
4394
3990
  inputPrice: 0.075,
4395
3991
  outputPrice: 0.3,
4396
3992
  description: "OpenAI gpt-oss 20B. Available in us-central1"
@@ -4400,7 +3996,6 @@ var vertexModels = {
4400
3996
  contextWindow: 262144,
4401
3997
  supportsImages: false,
4402
3998
  supportsPromptCache: false,
4403
- supportsNativeTools: true,
4404
3999
  inputPrice: 1,
4405
4000
  outputPrice: 4,
4406
4001
  description: "Qwen3 Coder 480B A35B Instruct. Available in us-south1"
@@ -4410,10 +4005,18 @@ var vertexModels = {
4410
4005
  contextWindow: 262144,
4411
4006
  supportsImages: false,
4412
4007
  supportsPromptCache: false,
4413
- supportsNativeTools: true,
4414
4008
  inputPrice: 0.25,
4415
4009
  outputPrice: 1,
4416
4010
  description: "Qwen3 235B A22B Instruct. Available in us-south1"
4011
+ },
4012
+ "moonshotai/kimi-k2-thinking-maas": {
4013
+ maxTokens: 16384,
4014
+ contextWindow: 262144,
4015
+ supportsPromptCache: false,
4016
+ supportsImages: false,
4017
+ inputPrice: 0.6,
4018
+ outputPrice: 2.5,
4019
+ description: "Kimi K2 Thinking Model with 256K context window."
4417
4020
  }
4418
4021
  };
4419
4022
  var VERTEX_1M_CONTEXT_MODEL_IDS = ["claude-sonnet-4@20250514", "claude-sonnet-4-5@20250929"];
@@ -4646,8 +4249,6 @@ var xaiModels = {
4646
4249
  contextWindow: 256e3,
4647
4250
  supportsImages: true,
4648
4251
  supportsPromptCache: true,
4649
- supportsNativeTools: true,
4650
- defaultToolProtocol: "native",
4651
4252
  inputPrice: 0.2,
4652
4253
  outputPrice: 1.5,
4653
4254
  cacheWritesPrice: 0.02,
@@ -4661,8 +4262,6 @@ var xaiModels = {
4661
4262
  contextWindow: 2e6,
4662
4263
  supportsImages: true,
4663
4264
  supportsPromptCache: true,
4664
- supportsNativeTools: true,
4665
- defaultToolProtocol: "native",
4666
4265
  inputPrice: 0.2,
4667
4266
  outputPrice: 0.5,
4668
4267
  cacheWritesPrice: 0.05,
@@ -4676,8 +4275,6 @@ var xaiModels = {
4676
4275
  contextWindow: 2e6,
4677
4276
  supportsImages: true,
4678
4277
  supportsPromptCache: true,
4679
- supportsNativeTools: true,
4680
- defaultToolProtocol: "native",
4681
4278
  inputPrice: 0.2,
4682
4279
  outputPrice: 0.5,
4683
4280
  cacheWritesPrice: 0.05,
@@ -4691,8 +4288,6 @@ var xaiModels = {
4691
4288
  contextWindow: 2e6,
4692
4289
  supportsImages: true,
4693
4290
  supportsPromptCache: true,
4694
- supportsNativeTools: true,
4695
- defaultToolProtocol: "native",
4696
4291
  inputPrice: 0.2,
4697
4292
  outputPrice: 0.5,
4698
4293
  cacheWritesPrice: 0.05,
@@ -4706,8 +4301,6 @@ var xaiModels = {
4706
4301
  contextWindow: 2e6,
4707
4302
  supportsImages: true,
4708
4303
  supportsPromptCache: true,
4709
- supportsNativeTools: true,
4710
- defaultToolProtocol: "native",
4711
4304
  inputPrice: 0.2,
4712
4305
  outputPrice: 0.5,
4713
4306
  cacheWritesPrice: 0.05,
@@ -4721,8 +4314,6 @@ var xaiModels = {
4721
4314
  contextWindow: 256e3,
4722
4315
  supportsImages: true,
4723
4316
  supportsPromptCache: true,
4724
- supportsNativeTools: true,
4725
- defaultToolProtocol: "native",
4726
4317
  inputPrice: 3,
4727
4318
  outputPrice: 15,
4728
4319
  cacheWritesPrice: 0.75,
@@ -4736,8 +4327,6 @@ var xaiModels = {
4736
4327
  contextWindow: 131072,
4737
4328
  supportsImages: true,
4738
4329
  supportsPromptCache: true,
4739
- supportsNativeTools: true,
4740
- defaultToolProtocol: "native",
4741
4330
  inputPrice: 0.3,
4742
4331
  outputPrice: 0.5,
4743
4332
  cacheWritesPrice: 0.07,
@@ -4753,8 +4342,6 @@ var xaiModels = {
4753
4342
  contextWindow: 131072,
4754
4343
  supportsImages: true,
4755
4344
  supportsPromptCache: true,
4756
- supportsNativeTools: true,
4757
- defaultToolProtocol: "native",
4758
4345
  inputPrice: 3,
4759
4346
  outputPrice: 15,
4760
4347
  cacheWritesPrice: 0.75,
@@ -4851,7 +4438,6 @@ var vercelAiGatewayDefaultModelInfo = {
4851
4438
  contextWindow: 2e5,
4852
4439
  supportsImages: true,
4853
4440
  supportsPromptCache: true,
4854
- supportsNativeTools: true,
4855
4441
  inputPrice: 3,
4856
4442
  outputPrice: 15,
4857
4443
  cacheWritesPrice: 3.75,
@@ -4868,8 +4454,6 @@ var internationalZAiModels = {
4868
4454
  contextWindow: 131072,
4869
4455
  supportsImages: false,
4870
4456
  supportsPromptCache: true,
4871
- supportsNativeTools: true,
4872
- defaultToolProtocol: "native",
4873
4457
  inputPrice: 0.6,
4874
4458
  outputPrice: 2.2,
4875
4459
  cacheWritesPrice: 0,
@@ -4881,8 +4465,6 @@ var internationalZAiModels = {
4881
4465
  contextWindow: 131072,
4882
4466
  supportsImages: false,
4883
4467
  supportsPromptCache: true,
4884
- supportsNativeTools: true,
4885
- defaultToolProtocol: "native",
4886
4468
  inputPrice: 0.2,
4887
4469
  outputPrice: 1.1,
4888
4470
  cacheWritesPrice: 0,
@@ -4894,8 +4476,6 @@ var internationalZAiModels = {
4894
4476
  contextWindow: 131072,
4895
4477
  supportsImages: false,
4896
4478
  supportsPromptCache: true,
4897
- supportsNativeTools: true,
4898
- defaultToolProtocol: "native",
4899
4479
  inputPrice: 2.2,
4900
4480
  outputPrice: 8.9,
4901
4481
  cacheWritesPrice: 0,
@@ -4907,8 +4487,6 @@ var internationalZAiModels = {
4907
4487
  contextWindow: 131072,
4908
4488
  supportsImages: false,
4909
4489
  supportsPromptCache: true,
4910
- supportsNativeTools: true,
4911
- defaultToolProtocol: "native",
4912
4490
  inputPrice: 1.1,
4913
4491
  outputPrice: 4.5,
4914
4492
  cacheWritesPrice: 0,
@@ -4920,8 +4498,6 @@ var internationalZAiModels = {
4920
4498
  contextWindow: 131072,
4921
4499
  supportsImages: false,
4922
4500
  supportsPromptCache: true,
4923
- supportsNativeTools: true,
4924
- defaultToolProtocol: "native",
4925
4501
  inputPrice: 0,
4926
4502
  outputPrice: 0,
4927
4503
  cacheWritesPrice: 0,
@@ -4933,21 +4509,28 @@ var internationalZAiModels = {
4933
4509
  contextWindow: 131072,
4934
4510
  supportsImages: true,
4935
4511
  supportsPromptCache: true,
4936
- supportsNativeTools: true,
4937
- defaultToolProtocol: "native",
4938
4512
  inputPrice: 0.6,
4939
4513
  outputPrice: 1.8,
4940
4514
  cacheWritesPrice: 0,
4941
4515
  cacheReadsPrice: 0.11,
4942
4516
  description: "GLM-4.5V is Z.AI's multimodal visual reasoning model (image/video/text/file input), optimized for GUI tasks, grounding, and document/video understanding."
4943
4517
  },
4518
+ "glm-4.6v": {
4519
+ maxTokens: 16384,
4520
+ contextWindow: 131072,
4521
+ supportsImages: true,
4522
+ supportsPromptCache: true,
4523
+ inputPrice: 0.3,
4524
+ outputPrice: 0.9,
4525
+ cacheWritesPrice: 0,
4526
+ cacheReadsPrice: 0.05,
4527
+ description: "GLM-4.6V is an advanced multimodal vision model with improved performance and cost-efficiency for visual understanding tasks."
4528
+ },
4944
4529
  "glm-4.6": {
4945
4530
  maxTokens: 16384,
4946
4531
  contextWindow: 2e5,
4947
4532
  supportsImages: false,
4948
4533
  supportsPromptCache: true,
4949
- supportsNativeTools: true,
4950
- defaultToolProtocol: "native",
4951
4534
  inputPrice: 0.6,
4952
4535
  outputPrice: 2.2,
4953
4536
  cacheWritesPrice: 0,
@@ -4959,8 +4542,6 @@ var internationalZAiModels = {
4959
4542
  contextWindow: 2e5,
4960
4543
  supportsImages: false,
4961
4544
  supportsPromptCache: true,
4962
- supportsNativeTools: true,
4963
- defaultToolProtocol: "native",
4964
4545
  supportsReasoningEffort: ["disable", "medium"],
4965
4546
  reasoningEffort: "medium",
4966
4547
  preserveReasoning: true,
@@ -4970,13 +4551,55 @@ var internationalZAiModels = {
4970
4551
  cacheReadsPrice: 0.11,
4971
4552
  description: "GLM-4.7 is Zhipu's latest model with built-in thinking capabilities enabled by default. It provides enhanced reasoning for complex tasks while maintaining fast response times."
4972
4553
  },
4554
+ "glm-4.7-flash": {
4555
+ maxTokens: 16384,
4556
+ contextWindow: 2e5,
4557
+ supportsImages: false,
4558
+ supportsPromptCache: true,
4559
+ inputPrice: 0,
4560
+ outputPrice: 0,
4561
+ cacheWritesPrice: 0,
4562
+ cacheReadsPrice: 0,
4563
+ description: "GLM-4.7-Flash is a free, high-speed variant of GLM-4.7 offering fast responses for reasoning and coding tasks."
4564
+ },
4565
+ "glm-4.7-flashx": {
4566
+ maxTokens: 16384,
4567
+ contextWindow: 2e5,
4568
+ supportsImages: false,
4569
+ supportsPromptCache: true,
4570
+ inputPrice: 0.07,
4571
+ outputPrice: 0.4,
4572
+ cacheWritesPrice: 0,
4573
+ cacheReadsPrice: 0.01,
4574
+ description: "GLM-4.7-FlashX is an ultra-fast variant of GLM-4.7 with exceptional speed and cost-effectiveness for high-throughput applications."
4575
+ },
4576
+ "glm-4.6v-flash": {
4577
+ maxTokens: 16384,
4578
+ contextWindow: 131072,
4579
+ supportsImages: true,
4580
+ supportsPromptCache: true,
4581
+ inputPrice: 0,
4582
+ outputPrice: 0,
4583
+ cacheWritesPrice: 0,
4584
+ cacheReadsPrice: 0,
4585
+ description: "GLM-4.6V-Flash is a free, high-speed multimodal vision model for rapid image understanding and visual reasoning tasks."
4586
+ },
4587
+ "glm-4.6v-flashx": {
4588
+ maxTokens: 16384,
4589
+ contextWindow: 131072,
4590
+ supportsImages: true,
4591
+ supportsPromptCache: true,
4592
+ inputPrice: 0.04,
4593
+ outputPrice: 0.4,
4594
+ cacheWritesPrice: 0,
4595
+ cacheReadsPrice: 4e-3,
4596
+ description: "GLM-4.6V-FlashX is an ultra-fast multimodal vision model optimized for high-speed visual processing at low cost."
4597
+ },
4973
4598
  "glm-4-32b-0414-128k": {
4974
4599
  maxTokens: 16384,
4975
4600
  contextWindow: 131072,
4976
4601
  supportsImages: false,
4977
4602
  supportsPromptCache: false,
4978
- supportsNativeTools: true,
4979
- defaultToolProtocol: "native",
4980
4603
  inputPrice: 0.1,
4981
4604
  outputPrice: 0.1,
4982
4605
  cacheWritesPrice: 0,
@@ -4991,8 +4614,6 @@ var mainlandZAiModels = {
4991
4614
  contextWindow: 131072,
4992
4615
  supportsImages: false,
4993
4616
  supportsPromptCache: true,
4994
- supportsNativeTools: true,
4995
- defaultToolProtocol: "native",
4996
4617
  inputPrice: 0.29,
4997
4618
  outputPrice: 1.14,
4998
4619
  cacheWritesPrice: 0,
@@ -5004,8 +4625,6 @@ var mainlandZAiModels = {
5004
4625
  contextWindow: 131072,
5005
4626
  supportsImages: false,
5006
4627
  supportsPromptCache: true,
5007
- supportsNativeTools: true,
5008
- defaultToolProtocol: "native",
5009
4628
  inputPrice: 0.1,
5010
4629
  outputPrice: 0.6,
5011
4630
  cacheWritesPrice: 0,
@@ -5017,8 +4636,6 @@ var mainlandZAiModels = {
5017
4636
  contextWindow: 131072,
5018
4637
  supportsImages: false,
5019
4638
  supportsPromptCache: true,
5020
- supportsNativeTools: true,
5021
- defaultToolProtocol: "native",
5022
4639
  inputPrice: 0.29,
5023
4640
  outputPrice: 1.14,
5024
4641
  cacheWritesPrice: 0,
@@ -5030,8 +4647,6 @@ var mainlandZAiModels = {
5030
4647
  contextWindow: 131072,
5031
4648
  supportsImages: false,
5032
4649
  supportsPromptCache: true,
5033
- supportsNativeTools: true,
5034
- defaultToolProtocol: "native",
5035
4650
  inputPrice: 0.1,
5036
4651
  outputPrice: 0.6,
5037
4652
  cacheWritesPrice: 0,
@@ -5043,8 +4658,6 @@ var mainlandZAiModels = {
5043
4658
  contextWindow: 131072,
5044
4659
  supportsImages: false,
5045
4660
  supportsPromptCache: true,
5046
- supportsNativeTools: true,
5047
- defaultToolProtocol: "native",
5048
4661
  inputPrice: 0,
5049
4662
  outputPrice: 0,
5050
4663
  cacheWritesPrice: 0,
@@ -5056,8 +4669,6 @@ var mainlandZAiModels = {
5056
4669
  contextWindow: 131072,
5057
4670
  supportsImages: true,
5058
4671
  supportsPromptCache: true,
5059
- supportsNativeTools: true,
5060
- defaultToolProtocol: "native",
5061
4672
  inputPrice: 0.29,
5062
4673
  outputPrice: 0.93,
5063
4674
  cacheWritesPrice: 0,
@@ -5069,8 +4680,6 @@ var mainlandZAiModels = {
5069
4680
  contextWindow: 204800,
5070
4681
  supportsImages: false,
5071
4682
  supportsPromptCache: true,
5072
- supportsNativeTools: true,
5073
- defaultToolProtocol: "native",
5074
4683
  inputPrice: 0.29,
5075
4684
  outputPrice: 1.14,
5076
4685
  cacheWritesPrice: 0,
@@ -5082,8 +4691,6 @@ var mainlandZAiModels = {
5082
4691
  contextWindow: 204800,
5083
4692
  supportsImages: false,
5084
4693
  supportsPromptCache: true,
5085
- supportsNativeTools: true,
5086
- defaultToolProtocol: "native",
5087
4694
  supportsReasoningEffort: ["disable", "medium"],
5088
4695
  reasoningEffort: "medium",
5089
4696
  preserveReasoning: true,
@@ -5092,6 +4699,61 @@ var mainlandZAiModels = {
5092
4699
  cacheWritesPrice: 0,
5093
4700
  cacheReadsPrice: 0.057,
5094
4701
  description: "GLM-4.7 is Zhipu's latest model with built-in thinking capabilities enabled by default. It provides enhanced reasoning for complex tasks while maintaining fast response times."
4702
+ },
4703
+ "glm-4.7-flash": {
4704
+ maxTokens: 16384,
4705
+ contextWindow: 204800,
4706
+ supportsImages: false,
4707
+ supportsPromptCache: true,
4708
+ inputPrice: 0,
4709
+ outputPrice: 0,
4710
+ cacheWritesPrice: 0,
4711
+ cacheReadsPrice: 0,
4712
+ description: "GLM-4.7-Flash is a free, high-speed variant of GLM-4.7 offering fast responses for reasoning and coding tasks."
4713
+ },
4714
+ "glm-4.7-flashx": {
4715
+ maxTokens: 16384,
4716
+ contextWindow: 204800,
4717
+ supportsImages: false,
4718
+ supportsPromptCache: true,
4719
+ inputPrice: 0.035,
4720
+ outputPrice: 0.2,
4721
+ cacheWritesPrice: 0,
4722
+ cacheReadsPrice: 5e-3,
4723
+ description: "GLM-4.7-FlashX is an ultra-fast variant of GLM-4.7 with exceptional speed and cost-effectiveness for high-throughput applications."
4724
+ },
4725
+ "glm-4.6v": {
4726
+ maxTokens: 16384,
4727
+ contextWindow: 131072,
4728
+ supportsImages: true,
4729
+ supportsPromptCache: true,
4730
+ inputPrice: 0.15,
4731
+ outputPrice: 0.45,
4732
+ cacheWritesPrice: 0,
4733
+ cacheReadsPrice: 0.025,
4734
+ description: "GLM-4.6V is an advanced multimodal vision model with improved performance and cost-efficiency for visual understanding tasks."
4735
+ },
4736
+ "glm-4.6v-flash": {
4737
+ maxTokens: 16384,
4738
+ contextWindow: 131072,
4739
+ supportsImages: true,
4740
+ supportsPromptCache: true,
4741
+ inputPrice: 0,
4742
+ outputPrice: 0,
4743
+ cacheWritesPrice: 0,
4744
+ cacheReadsPrice: 0,
4745
+ description: "GLM-4.6V-Flash is a free, high-speed multimodal vision model for rapid image understanding and visual reasoning tasks."
4746
+ },
4747
+ "glm-4.6v-flashx": {
4748
+ maxTokens: 16384,
4749
+ contextWindow: 131072,
4750
+ supportsImages: true,
4751
+ supportsPromptCache: true,
4752
+ inputPrice: 0.02,
4753
+ outputPrice: 0.2,
4754
+ cacheWritesPrice: 0,
4755
+ cacheReadsPrice: 2e-3,
4756
+ description: "GLM-4.6V-FlashX is an ultra-fast multimodal vision model optimized for high-speed visual processing at low cost."
5095
4757
  }
5096
4758
  };
5097
4759
  var ZAI_DEFAULT_TEMPERATURE = 0.6;
@@ -5125,7 +4787,6 @@ var deepInfraDefaultModelInfo = {
5125
4787
  contextWindow: 262144,
5126
4788
  supportsImages: false,
5127
4789
  supportsPromptCache: false,
5128
- supportsNativeTools: true,
5129
4790
  inputPrice: 0.3,
5130
4791
  outputPrice: 1.2,
5131
4792
  description: "Qwen 3 Coder 480B A35B Instruct Turbo model, 256K context."
@@ -5139,8 +4800,6 @@ var minimaxModels = {
5139
4800
  contextWindow: 192e3,
5140
4801
  supportsImages: false,
5141
4802
  supportsPromptCache: true,
5142
- supportsNativeTools: true,
5143
- defaultToolProtocol: "native",
5144
4803
  includedTools: ["search_and_replace"],
5145
4804
  excludedTools: ["apply_diff"],
5146
4805
  preserveReasoning: true,
@@ -5155,8 +4814,6 @@ var minimaxModels = {
5155
4814
  contextWindow: 192e3,
5156
4815
  supportsImages: false,
5157
4816
  supportsPromptCache: true,
5158
- supportsNativeTools: true,
5159
- defaultToolProtocol: "native",
5160
4817
  includedTools: ["search_and_replace"],
5161
4818
  excludedTools: ["apply_diff"],
5162
4819
  preserveReasoning: true,
@@ -5171,8 +4828,6 @@ var minimaxModels = {
5171
4828
  contextWindow: 192e3,
5172
4829
  supportsImages: false,
5173
4830
  supportsPromptCache: true,
5174
- supportsNativeTools: true,
5175
- defaultToolProtocol: "native",
5176
4831
  includedTools: ["search_and_replace"],
5177
4832
  excludedTools: ["apply_diff"],
5178
4833
  preserveReasoning: true,
@@ -5244,8 +4899,6 @@ function getProviderDefaultModelId(provider, options = { isChina: false }) {
5244
4899
  return deepInfraDefaultModelId;
5245
4900
  case "vscode-lm":
5246
4901
  return vscodeLlmDefaultModelId;
5247
- case "claude-code":
5248
- return claudeCodeDefaultModelId;
5249
4902
  case "cerebras":
5250
4903
  return cerebrasDefaultModelId;
5251
4904
  case "sambanova":
@@ -5303,7 +4956,6 @@ var providerNames = [
5303
4956
  "bedrock",
5304
4957
  "baseten",
5305
4958
  "cerebras",
5306
- "claude-code",
5307
4959
  "doubao",
5308
4960
  "deepseek",
5309
4961
  "featherless",
@@ -5333,9 +4985,7 @@ var providerSettingsEntrySchema = z8.object({
5333
4985
  });
5334
4986
  var baseProviderSettingsSchema = z8.object({
5335
4987
  includeMaxTokens: z8.boolean().optional(),
5336
- diffEnabled: z8.boolean().optional(),
5337
4988
  todoListEnabled: z8.boolean().optional(),
5338
- fuzzyMatchThreshold: z8.number().optional(),
5339
4989
  modelTemperature: z8.number().nullish(),
5340
4990
  rateLimitSeconds: z8.number().optional(),
5341
4991
  consecutiveMistakeLimit: z8.number().min(0).optional(),
@@ -5345,9 +4995,7 @@ var baseProviderSettingsSchema = z8.object({
5345
4995
  modelMaxTokens: z8.number().optional(),
5346
4996
  modelMaxThinkingTokens: z8.number().optional(),
5347
4997
  // Model verbosity.
5348
- verbosity: verbosityLevelsSchema.optional(),
5349
- // Tool protocol override for this profile.
5350
- toolProtocol: z8.enum(["xml", "native"]).optional()
4998
+ verbosity: verbosityLevelsSchema.optional()
5351
4999
  });
5352
5000
  var apiModelIdProviderModelSchema = baseProviderSettingsSchema.extend({
5353
5001
  apiModelId: z8.string().optional()
@@ -5359,7 +5007,6 @@ var anthropicSchema = apiModelIdProviderModelSchema.extend({
5359
5007
  anthropicBeta1MContext: z8.boolean().optional()
5360
5008
  // Enable 'context-1m-2025-08-07' beta for 1M context window.
5361
5009
  });
5362
- var claudeCodeSchema = apiModelIdProviderModelSchema.extend({});
5363
5010
  var openRouterSchema = baseProviderSettingsSchema.extend({
5364
5011
  openRouterApiKey: z8.string().optional(),
5365
5012
  openRouterModelId: z8.string().optional(),
@@ -5548,7 +5195,6 @@ var defaultSchema = z8.object({
5548
5195
  });
5549
5196
  var providerSettingsSchemaDiscriminated = z8.discriminatedUnion("apiProvider", [
5550
5197
  anthropicSchema.merge(z8.object({ apiProvider: z8.literal("anthropic") })),
5551
- claudeCodeSchema.merge(z8.object({ apiProvider: z8.literal("claude-code") })),
5552
5198
  openRouterSchema.merge(z8.object({ apiProvider: z8.literal("openrouter") })),
5553
5199
  bedrockSchema.merge(z8.object({ apiProvider: z8.literal("bedrock") })),
5554
5200
  vertexSchema.merge(z8.object({ apiProvider: z8.literal("vertex") })),
@@ -5589,7 +5235,6 @@ var providerSettingsSchemaDiscriminated = z8.discriminatedUnion("apiProvider", [
5589
5235
  var providerSettingsSchema = z8.object({
5590
5236
  apiProvider: providerNamesSchema.optional(),
5591
5237
  ...anthropicSchema.shape,
5592
- ...claudeCodeSchema.shape,
5593
5238
  ...openRouterSchema.shape,
5594
5239
  ...bedrockSchema.shape,
5595
5240
  ...vertexSchema.shape,
@@ -5654,7 +5299,6 @@ var getModelId = (settings) => {
5654
5299
  var isTypicalProvider = (key) => isProviderName(key) && !isInternalProvider(key) && !isCustomProvider(key) && !isFauxProvider(key);
5655
5300
  var modelIdKeysByProvider = {
5656
5301
  anthropic: "apiModelId",
5657
- "claude-code": "apiModelId",
5658
5302
  openrouter: "openRouterModelId",
5659
5303
  bedrock: "apiModelId",
5660
5304
  vertex: "apiModelId",
@@ -5688,7 +5332,7 @@ var modelIdKeysByProvider = {
5688
5332
  roo: "apiModelId",
5689
5333
  "vercel-ai-gateway": "vercelAiGatewayModelId"
5690
5334
  };
5691
- var ANTHROPIC_STYLE_PROVIDERS = ["anthropic", "claude-code", "bedrock", "minimax"];
5335
+ var ANTHROPIC_STYLE_PROVIDERS = ["anthropic", "bedrock", "minimax"];
5692
5336
  var getApiProtocol = (provider, modelId) => {
5693
5337
  if (provider && ANTHROPIC_STYLE_PROVIDERS.includes(provider)) {
5694
5338
  return "anthropic";
@@ -5717,7 +5361,6 @@ var MODELS_BY_PROVIDER = {
5717
5361
  label: "Cerebras",
5718
5362
  models: Object.keys(cerebrasModels)
5719
5363
  },
5720
- "claude-code": { id: "claude-code", label: "Claude Code", models: Object.keys(claudeCodeModels) },
5721
5364
  deepseek: {
5722
5365
  id: "deepseek",
5723
5366
  label: "DeepSeek",
@@ -5821,16 +5464,6 @@ var historyItemSchema = z9.object({
5821
5464
  size: z9.number().optional(),
5822
5465
  workspace: z9.string().optional(),
5823
5466
  mode: z9.string().optional(),
5824
- /**
5825
- * The tool protocol used by this task. Once a task uses tools with a specific
5826
- * protocol (XML or Native), it is permanently locked to that protocol.
5827
- *
5828
- * - "xml": Tool calls are parsed from XML text (no tool IDs)
5829
- * - "native": Tool calls come as tool_call chunks with IDs
5830
- *
5831
- * This ensures task resumption works correctly even when NTC settings change.
5832
- */
5833
- toolProtocol: z9.enum(["xml", "native"]).optional(),
5834
5467
  apiConfigName: z9.string().optional(),
5835
5468
  // Provider profile name for sticky profile feature
5836
5469
  status: z9.enum(["active", "completed", "delegated"]).optional(),
@@ -5848,23 +5481,12 @@ var historyItemSchema = z9.object({
5848
5481
 
5849
5482
  // src/experiment.ts
5850
5483
  import { z as z10 } from "zod";
5851
- var experimentIds = [
5852
- "powerSteering",
5853
- "multiFileApplyDiff",
5854
- "preventFocusDisruption",
5855
- "imageGeneration",
5856
- "runSlashCommand",
5857
- "multipleNativeToolCalls",
5858
- "customTools"
5859
- ];
5484
+ var experimentIds = ["preventFocusDisruption", "imageGeneration", "runSlashCommand", "customTools"];
5860
5485
  var experimentIdsSchema = z10.enum(experimentIds);
5861
5486
  var experimentsSchema = z10.object({
5862
- powerSteering: z10.boolean().optional(),
5863
- multiFileApplyDiff: z10.boolean().optional(),
5864
5487
  preventFocusDisruption: z10.boolean().optional(),
5865
5488
  imageGeneration: z10.boolean().optional(),
5866
5489
  runSlashCommand: z10.boolean().optional(),
5867
- multipleNativeToolCalls: z10.boolean().optional(),
5868
5490
  customTools: z10.boolean().optional()
5869
5491
  });
5870
5492
 
@@ -5916,6 +5538,7 @@ var TelemetryEventName = /* @__PURE__ */ ((TelemetryEventName2) => {
5916
5538
  TelemetryEventName2["CODE_INDEX_ERROR"] = "Code Index Error";
5917
5539
  TelemetryEventName2["TELEMETRY_SETTINGS_CHANGED"] = "Telemetry Settings Changed";
5918
5540
  TelemetryEventName2["MODEL_CACHE_EMPTY_RESPONSE"] = "Model Cache Empty Response";
5541
+ TelemetryEventName2["READ_FILE_LEGACY_FORMAT_USED"] = "Read File Legacy Format Used";
5919
5542
  return TelemetryEventName2;
5920
5543
  })(TelemetryEventName || {});
5921
5544
  var staticAppPropertiesSchema = z11.object({
@@ -6004,7 +5627,8 @@ var rooCodeTelemetryEventSchema = z11.discriminatedUnion("type", [
6004
5627
  "Sliding Window Truncation" /* SLIDING_WINDOW_TRUNCATION */,
6005
5628
  "Tab Shown" /* TAB_SHOWN */,
6006
5629
  "Mode Setting Changed" /* MODE_SETTINGS_CHANGED */,
6007
- "Custom Mode Created" /* CUSTOM_MODE_CREATED */
5630
+ "Custom Mode Created" /* CUSTOM_MODE_CREATED */,
5631
+ "Read File Legacy Format Used" /* READ_FILE_LEGACY_FORMAT_USED */
6008
5632
  ]),
6009
5633
  properties: telemetryPropertiesSchema
6010
5634
  }),
@@ -6318,7 +5942,15 @@ var isLanguage = (value) => languages.includes(value);
6318
5942
 
6319
5943
  // src/global-settings.ts
6320
5944
  var DEFAULT_WRITE_DELAY_MS = 1e3;
6321
- var DEFAULT_TERMINAL_OUTPUT_CHARACTER_LIMIT = 5e4;
5945
+ var TERMINAL_PREVIEW_BYTES = {
5946
+ small: 5 * 1024,
5947
+ // 5KB
5948
+ medium: 10 * 1024,
5949
+ // 10KB
5950
+ large: 20 * 1024
5951
+ // 20KB
5952
+ };
5953
+ var DEFAULT_TERMINAL_OUTPUT_PREVIEW_SIZE = "medium";
6322
5954
  var MIN_CHECKPOINT_TIMEOUT_SECONDS = 10;
6323
5955
  var MAX_CHECKPOINT_TIMEOUT_SECONDS = 60;
6324
5956
  var DEFAULT_CHECKPOINT_TIMEOUT_SECONDS = 15;
@@ -6334,7 +5966,6 @@ var globalSettingsSchema = z14.object({
6334
5966
  imageGenerationProvider: z14.enum(["openrouter", "roo"]).optional(),
6335
5967
  openRouterImageApiKey: z14.string().optional(),
6336
5968
  openRouterImageGenerationSelectedModel: z14.string().optional(),
6337
- condensingApiConfigId: z14.string().optional(),
6338
5969
  customCondensingPrompt: z14.string().optional(),
6339
5970
  autoApprovalEnabled: z14.boolean().optional(),
6340
5971
  alwaysAllowReadOnly: z14.boolean().optional(),
@@ -6360,7 +5991,6 @@ var globalSettingsSchema = z14.object({
6360
5991
  allowedMaxCost: z14.number().nullish(),
6361
5992
  autoCondenseContext: z14.boolean().optional(),
6362
5993
  autoCondenseContextPercent: z14.number().optional(),
6363
- maxConcurrentFileReads: z14.number().optional(),
6364
5994
  /**
6365
5995
  * Whether to include current time in the environment details
6366
5996
  * @default true
@@ -6403,11 +6033,9 @@ var globalSettingsSchema = z14.object({
6403
6033
  maxWorkspaceFiles: z14.number().optional(),
6404
6034
  showRooIgnoredFiles: z14.boolean().optional(),
6405
6035
  enableSubfolderRules: z14.boolean().optional(),
6406
- maxReadFileLine: z14.number().optional(),
6407
6036
  maxImageFileSize: z14.number().optional(),
6408
6037
  maxTotalImageSize: z14.number().optional(),
6409
- terminalOutputLineLimit: z14.number().optional(),
6410
- terminalOutputCharacterLimit: z14.number().optional(),
6038
+ terminalOutputPreviewSize: z14.enum(["small", "medium", "large"]).optional(),
6411
6039
  terminalShellIntegrationTimeout: z14.number().optional(),
6412
6040
  terminalShellIntegrationDisabled: z14.boolean().optional(),
6413
6041
  terminalCommandDelay: z14.number().optional(),
@@ -6416,18 +6044,14 @@ var globalSettingsSchema = z14.object({
6416
6044
  terminalZshOhMy: z14.boolean().optional(),
6417
6045
  terminalZshP10k: z14.boolean().optional(),
6418
6046
  terminalZdotdir: z14.boolean().optional(),
6419
- terminalCompressProgressBar: z14.boolean().optional(),
6420
6047
  diagnosticsEnabled: z14.boolean().optional(),
6421
6048
  rateLimitSeconds: z14.number().optional(),
6422
- diffEnabled: z14.boolean().optional(),
6423
- fuzzyMatchThreshold: z14.number().optional(),
6424
6049
  experiments: experimentsSchema.optional(),
6425
6050
  codebaseIndexModels: codebaseIndexModelsSchema.optional(),
6426
6051
  codebaseIndexConfig: codebaseIndexConfigSchema.optional(),
6427
6052
  language: languagesSchema.optional(),
6428
6053
  telemetrySetting: telemetrySettingsSchema.optional(),
6429
6054
  mcpEnabled: z14.boolean().optional(),
6430
- enableMcpServerCreation: z14.boolean().optional(),
6431
6055
  mode: z14.string().optional(),
6432
6056
  modeApiConfigs: z14.record(z14.string(), z14.string()).optional(),
6433
6057
  customModes: z14.array(modeConfigSchema).optional(),
@@ -6447,7 +6071,20 @@ var globalSettingsSchema = z14.object({
6447
6071
  profileThresholds: z14.record(z14.string(), z14.number()).optional(),
6448
6072
  hasOpenedModeSelector: z14.boolean().optional(),
6449
6073
  lastModeExportPath: z14.string().optional(),
6450
- lastModeImportPath: z14.string().optional()
6074
+ lastModeImportPath: z14.string().optional(),
6075
+ lastSettingsExportPath: z14.string().optional(),
6076
+ lastTaskExportPath: z14.string().optional(),
6077
+ lastImageSavePath: z14.string().optional(),
6078
+ /**
6079
+ * Path to worktree to auto-open after switching workspaces.
6080
+ * Used by the worktree feature to open the Roo Code sidebar in a new window.
6081
+ */
6082
+ worktreeAutoOpenPath: z14.string().optional(),
6083
+ /**
6084
+ * Whether to show the worktree selector in the home screen.
6085
+ * @default true
6086
+ */
6087
+ showWorktreesInHomeScreen: z14.boolean().optional()
6451
6088
  });
6452
6089
  var GLOBAL_SETTINGS_KEYS = globalSettingsSchema.keyof().options;
6453
6090
  var rooCodeSettingsSchema = providerSettingsSchema.merge(globalSettingsSchema);
@@ -6531,8 +6168,6 @@ var EVALS_SETTINGS = {
6531
6168
  ttsSpeed: 1,
6532
6169
  soundEnabled: false,
6533
6170
  soundVolume: 0.5,
6534
- terminalOutputLineLimit: 500,
6535
- terminalOutputCharacterLimit: DEFAULT_TERMINAL_OUTPUT_CHARACTER_LIMIT,
6536
6171
  terminalShellIntegrationTimeout: 3e4,
6537
6172
  terminalCommandDelay: 0,
6538
6173
  terminalPowershellCounter: false,
@@ -6540,19 +6175,14 @@ var EVALS_SETTINGS = {
6540
6175
  terminalZshClearEolMark: true,
6541
6176
  terminalZshP10k: false,
6542
6177
  terminalZdotdir: true,
6543
- terminalCompressProgressBar: true,
6544
6178
  terminalShellIntegrationDisabled: true,
6545
6179
  diagnosticsEnabled: true,
6546
- diffEnabled: true,
6547
- fuzzyMatchThreshold: 1,
6548
6180
  enableCheckpoints: false,
6549
6181
  rateLimitSeconds: 0,
6550
6182
  maxOpenTabsContext: 20,
6551
6183
  maxWorkspaceFiles: 200,
6552
6184
  maxGitStatusFiles: 20,
6553
6185
  showRooIgnoredFiles: true,
6554
- maxReadFileLine: -1,
6555
- // -1 to enable full file reading.
6556
6186
  includeDiagnosticMessages: true,
6557
6187
  maxDiagnosticMessages: 50,
6558
6188
  language: "en",
@@ -6626,24 +6256,18 @@ var organizationAllowListSchema = z16.object({
6626
6256
  });
6627
6257
  var organizationDefaultSettingsSchema = globalSettingsSchema.pick({
6628
6258
  enableCheckpoints: true,
6629
- fuzzyMatchThreshold: true,
6630
6259
  maxOpenTabsContext: true,
6631
- maxReadFileLine: true,
6632
6260
  maxWorkspaceFiles: true,
6633
6261
  showRooIgnoredFiles: true,
6634
6262
  terminalCommandDelay: true,
6635
- terminalCompressProgressBar: true,
6636
- terminalOutputLineLimit: true,
6637
6263
  terminalShellIntegrationDisabled: true,
6638
6264
  terminalShellIntegrationTimeout: true,
6639
6265
  terminalZshClearEolMark: true
6640
6266
  }).merge(
6641
6267
  z16.object({
6642
6268
  maxOpenTabsContext: z16.number().int().nonnegative().optional(),
6643
- maxReadFileLine: z16.number().int().gte(-1).optional(),
6644
6269
  maxWorkspaceFiles: z16.number().int().nonnegative().optional(),
6645
6270
  terminalCommandDelay: z16.number().int().nonnegative().optional(),
6646
- terminalOutputLineLimit: z16.number().int().nonnegative().optional(),
6647
6271
  terminalShellIntegrationTimeout: z16.number().int().nonnegative().optional()
6648
6272
  })
6649
6273
  );
@@ -7159,6 +6783,7 @@ var ipcMessageSchema = z19.discriminatedUnion("type", [
7159
6783
 
7160
6784
  // src/mcp.ts
7161
6785
  import { z as z20 } from "zod";
6786
+ var MAX_MCP_TOOLS_THRESHOLD = 60;
7162
6787
  var mcpExecutionStatusSchema = z20.discriminatedUnion("status", [
7163
6788
  z20.object({
7164
6789
  executionId: z20.string(),
@@ -7182,6 +6807,46 @@ var mcpExecutionStatusSchema = z20.discriminatedUnion("status", [
7182
6807
  error: z20.string().optional()
7183
6808
  })
7184
6809
  ]);
6810
+ function countEnabledMcpTools(servers) {
6811
+ let serverCount = 0;
6812
+ let toolCount = 0;
6813
+ for (const server of servers) {
6814
+ if (server.disabled) continue;
6815
+ if (server.status !== "connected") continue;
6816
+ serverCount++;
6817
+ if (server.tools) {
6818
+ for (const tool of server.tools) {
6819
+ if (tool.enabledForPrompt !== false) {
6820
+ toolCount++;
6821
+ }
6822
+ }
6823
+ }
6824
+ }
6825
+ return { enabledToolCount: toolCount, enabledServerCount: serverCount };
6826
+ }
6827
+
6828
+ // src/skills.ts
6829
+ var SKILL_NAME_MIN_LENGTH = 1;
6830
+ var SKILL_NAME_MAX_LENGTH = 64;
6831
+ var SKILL_NAME_REGEX = /^[a-z0-9]+(?:-[a-z0-9]+)*$/;
6832
+ var SkillNameValidationError = /* @__PURE__ */ ((SkillNameValidationError2) => {
6833
+ SkillNameValidationError2["Empty"] = "empty";
6834
+ SkillNameValidationError2["TooLong"] = "too_long";
6835
+ SkillNameValidationError2["InvalidFormat"] = "invalid_format";
6836
+ return SkillNameValidationError2;
6837
+ })(SkillNameValidationError || {});
6838
+ function validateSkillName(name) {
6839
+ if (!name || name.length < SKILL_NAME_MIN_LENGTH) {
6840
+ return { valid: false, error: "empty" /* Empty */ };
6841
+ }
6842
+ if (name.length > SKILL_NAME_MAX_LENGTH) {
6843
+ return { valid: false, error: "too_long" /* TooLong */ };
6844
+ }
6845
+ if (!SKILL_NAME_REGEX.test(name)) {
6846
+ return { valid: false, error: "invalid_format" /* InvalidFormat */ };
6847
+ }
6848
+ return { valid: true };
6849
+ }
7185
6850
 
7186
6851
  // src/todo.ts
7187
6852
  import { z as z21 } from "zod";
@@ -7221,6 +6886,11 @@ var commandExecutionStatusSchema = z22.discriminatedUnion("status", [
7221
6886
  })
7222
6887
  ]);
7223
6888
 
6889
+ // src/tool-params.ts
6890
+ function isLegacyReadFileParams(params) {
6891
+ return "_legacyFormat" in params && params._legacyFormat === true;
6892
+ }
6893
+
7224
6894
  // src/vscode-extension-host.ts
7225
6895
  import { z as z23 } from "zod";
7226
6896
  var checkoutDiffPayloadSchema = z23.object({
@@ -7273,7 +6943,7 @@ export {
7273
6943
  DEFAULT_CHECKPOINT_TIMEOUT_SECONDS,
7274
6944
  DEFAULT_CONSECUTIVE_MISTAKE_LIMIT,
7275
6945
  DEFAULT_MODES,
7276
- DEFAULT_TERMINAL_OUTPUT_CHARACTER_LIMIT,
6946
+ DEFAULT_TERMINAL_OUTPUT_PREVIEW_SIZE,
7277
6947
  DEFAULT_WRITE_DELAY_MS,
7278
6948
  DOUBAO_API_BASE_URL,
7279
6949
  DOUBAO_API_CHAT_PATH,
@@ -7303,13 +6973,13 @@ export {
7303
6973
  IpcOrigin,
7304
6974
  LMSTUDIO_DEFAULT_TEMPERATURE,
7305
6975
  MAX_CHECKPOINT_TIMEOUT_SECONDS,
6976
+ MAX_MCP_TOOLS_THRESHOLD,
7306
6977
  MINIMAX_DEFAULT_MAX_TOKENS,
7307
6978
  MINIMAX_DEFAULT_TEMPERATURE,
7308
6979
  MIN_CHECKPOINT_TIMEOUT_SECONDS,
7309
6980
  MISTRAL_DEFAULT_TEMPERATURE,
7310
6981
  MODELS_BY_PROVIDER,
7311
6982
  MOONSHOT_DEFAULT_TEMPERATURE,
7312
- NATIVE_TOOL_DEFAULTS,
7313
6983
  OPENAI_AZURE_AI_INFERENCE_PATH,
7314
6984
  OPENAI_NATIVE_DEFAULT_TEMPERATURE,
7315
6985
  OPENROUTER_DEFAULT_PROVIDER_NAME,
@@ -7324,7 +6994,11 @@ export {
7324
6994
  RooModelsResponseSchema,
7325
6995
  RooPricingSchema,
7326
6996
  SECRET_STATE_KEYS,
7327
- TOOL_PROTOCOL,
6997
+ SKILL_NAME_MAX_LENGTH,
6998
+ SKILL_NAME_MIN_LENGTH,
6999
+ SKILL_NAME_REGEX,
7000
+ SkillNameValidationError,
7001
+ TERMINAL_PREVIEW_BYTES,
7328
7002
  TaskBridgeCommandName,
7329
7003
  TaskBridgeEventName,
7330
7004
  TaskCommandName,
@@ -7356,9 +7030,6 @@ export {
7356
7030
  chutesDefaultModelId,
7357
7031
  chutesDefaultModelInfo,
7358
7032
  chutesModels,
7359
- claudeCodeDefaultModelId,
7360
- claudeCodeModels,
7361
- claudeCodeReasoningConfig,
7362
7033
  clineAskSchema,
7363
7034
  clineAsks,
7364
7035
  clineMessageSchema,
@@ -7373,6 +7044,7 @@ export {
7373
7044
  commandIds,
7374
7045
  contextCondenseSchema,
7375
7046
  contextTruncationSchema,
7047
+ countEnabledMcpTools,
7376
7048
  customModePromptsSchema,
7377
7049
  customModesSettingsSchema,
7378
7050
  customProviders,
@@ -7406,7 +7078,6 @@ export {
7406
7078
  geminiDefaultModelId,
7407
7079
  geminiModels,
7408
7080
  getApiProtocol,
7409
- getEffectiveProtocol,
7410
7081
  getErrorMessage,
7411
7082
  getErrorStatusCode,
7412
7083
  getImageGenerationProvider,
@@ -7441,9 +7112,9 @@ export {
7441
7112
  isInteractiveAsk,
7442
7113
  isInternalProvider,
7443
7114
  isLanguage,
7115
+ isLegacyReadFileParams,
7444
7116
  isLocalProvider,
7445
7117
  isModelParameter,
7446
- isNativeProtocol,
7447
7118
  isNonBlockingAsk,
7448
7119
  isProviderName,
7449
7120
  isResumableAsk,
@@ -7479,7 +7150,6 @@ export {
7479
7150
  moonshotDefaultModelId,
7480
7151
  moonshotModels,
7481
7152
  nonBlockingAsks,
7482
- normalizeClaudeCodeModelId,
7483
7153
  ollamaDefaultModelId,
7484
7154
  ollamaDefaultModelInfo,
7485
7155
  openAiCodexDefaultModelId,
@@ -7553,6 +7223,7 @@ export {
7553
7223
  userFeaturesSchema,
7554
7224
  userSettingsConfigSchema,
7555
7225
  userSettingsDataSchema,
7226
+ validateSkillName,
7556
7227
  verbosityLevels,
7557
7228
  verbosityLevelsSchema,
7558
7229
  vercelAiGatewayDefaultModelId,