@roo-code/types 1.106.0 → 1.107.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/index.cjs +395 -718
- package/dist/index.cjs.map +1 -1
- package/dist/index.d.cts +1553 -2546
- package/dist/index.d.ts +1553 -2546
- package/dist/index.js +385 -709
- package/dist/index.js.map +1 -1
- package/package.json +2 -2
package/dist/index.js
CHANGED
|
@@ -78,7 +78,9 @@ var clineSays = [
|
|
|
78
78
|
"condense_context_error",
|
|
79
79
|
"sliding_window_truncation",
|
|
80
80
|
"codebase_search_result",
|
|
81
|
-
"user_edit_todos"
|
|
81
|
+
"user_edit_todos",
|
|
82
|
+
"too_many_tools_warning",
|
|
83
|
+
"tool"
|
|
82
84
|
];
|
|
83
85
|
var clineSaySchema = z.enum(clineSays);
|
|
84
86
|
var toolProgressStatusSchema = z.object({
|
|
@@ -146,6 +148,7 @@ var toolGroupsSchema = z2.enum(toolGroups);
|
|
|
146
148
|
var toolNames = [
|
|
147
149
|
"execute_command",
|
|
148
150
|
"read_file",
|
|
151
|
+
"read_command_output",
|
|
149
152
|
"write_to_file",
|
|
150
153
|
"apply_diff",
|
|
151
154
|
"search_and_replace",
|
|
@@ -161,10 +164,10 @@ var toolNames = [
|
|
|
161
164
|
"attempt_completion",
|
|
162
165
|
"switch_mode",
|
|
163
166
|
"new_task",
|
|
164
|
-
"fetch_instructions",
|
|
165
167
|
"codebase_search",
|
|
166
168
|
"update_todo_list",
|
|
167
169
|
"run_slash_command",
|
|
170
|
+
"skill",
|
|
168
171
|
"generate_image",
|
|
169
172
|
"custom_tool"
|
|
170
173
|
];
|
|
@@ -176,20 +179,6 @@ var toolUsageSchema = z2.record(
|
|
|
176
179
|
failures: z2.number()
|
|
177
180
|
})
|
|
178
181
|
);
|
|
179
|
-
var TOOL_PROTOCOL = {
|
|
180
|
-
XML: "xml",
|
|
181
|
-
NATIVE: "native"
|
|
182
|
-
};
|
|
183
|
-
var NATIVE_TOOL_DEFAULTS = {
|
|
184
|
-
supportsNativeTools: true,
|
|
185
|
-
defaultToolProtocol: TOOL_PROTOCOL.NATIVE
|
|
186
|
-
};
|
|
187
|
-
function isNativeProtocol(protocol) {
|
|
188
|
-
return protocol === TOOL_PROTOCOL.NATIVE;
|
|
189
|
-
}
|
|
190
|
-
function getEffectiveProtocol(toolProtocol) {
|
|
191
|
-
return toolProtocol || TOOL_PROTOCOL.XML;
|
|
192
|
-
}
|
|
193
182
|
|
|
194
183
|
// src/events.ts
|
|
195
184
|
var RooCodeEventName = /* @__PURE__ */ ((RooCodeEventName2) => {
|
|
@@ -213,6 +202,7 @@ var RooCodeEventName = /* @__PURE__ */ ((RooCodeEventName2) => {
|
|
|
213
202
|
RooCodeEventName2["TaskModeSwitched"] = "taskModeSwitched";
|
|
214
203
|
RooCodeEventName2["TaskAskResponded"] = "taskAskResponded";
|
|
215
204
|
RooCodeEventName2["TaskUserMessage"] = "taskUserMessage";
|
|
205
|
+
RooCodeEventName2["QueuedMessagesUpdated"] = "queuedMessagesUpdated";
|
|
216
206
|
RooCodeEventName2["TaskTokenUsageUpdated"] = "taskTokenUsageUpdated";
|
|
217
207
|
RooCodeEventName2["TaskToolFailed"] = "taskToolFailed";
|
|
218
208
|
RooCodeEventName2["ModeChanged"] = "modeChanged";
|
|
@@ -272,6 +262,7 @@ var rooCodeEventsSchema = z3.object({
|
|
|
272
262
|
["taskModeSwitched" /* TaskModeSwitched */]: z3.tuple([z3.string(), z3.string()]),
|
|
273
263
|
["taskAskResponded" /* TaskAskResponded */]: z3.tuple([z3.string()]),
|
|
274
264
|
["taskUserMessage" /* TaskUserMessage */]: z3.tuple([z3.string()]),
|
|
265
|
+
["queuedMessagesUpdated" /* QueuedMessagesUpdated */]: z3.tuple([z3.string(), z3.array(queuedMessageSchema)]),
|
|
275
266
|
["taskToolFailed" /* TaskToolFailed */]: z3.tuple([z3.string(), toolNamesSchema, z3.string()]),
|
|
276
267
|
["taskTokenUsageUpdated" /* TaskTokenUsageUpdated */]: z3.tuple([z3.string(), tokenUsageSchema, toolUsageSchema]),
|
|
277
268
|
["modeChanged" /* ModeChanged */]: z3.tuple([z3.string()]),
|
|
@@ -377,6 +368,11 @@ var taskEventSchema = z3.discriminatedUnion("eventName", [
|
|
|
377
368
|
payload: rooCodeEventsSchema.shape["taskAskResponded" /* TaskAskResponded */],
|
|
378
369
|
taskId: z3.number().optional()
|
|
379
370
|
}),
|
|
371
|
+
z3.object({
|
|
372
|
+
eventName: z3.literal("queuedMessagesUpdated" /* QueuedMessagesUpdated */),
|
|
373
|
+
payload: rooCodeEventsSchema.shape["queuedMessagesUpdated" /* QueuedMessagesUpdated */],
|
|
374
|
+
taskId: z3.number().optional()
|
|
375
|
+
}),
|
|
380
376
|
// Task Analytics
|
|
381
377
|
z3.object({
|
|
382
378
|
eventName: z3.literal("taskToolFailed" /* TaskToolFailed */),
|
|
@@ -477,10 +473,6 @@ var modelInfoSchema = z5.object({
|
|
|
477
473
|
isStealthModel: z5.boolean().optional(),
|
|
478
474
|
// Flag to indicate if the model is free (no cost)
|
|
479
475
|
isFree: z5.boolean().optional(),
|
|
480
|
-
// Flag to indicate if the model supports native tool calling (OpenAI-style function calling)
|
|
481
|
-
supportsNativeTools: z5.boolean().optional(),
|
|
482
|
-
// Default tool protocol preferred by this model (if not specified, falls back to capability/provider defaults)
|
|
483
|
-
defaultToolProtocol: z5.enum(["xml", "native"]).optional(),
|
|
484
476
|
// Exclude specific native tools from being available (only applies to native protocol)
|
|
485
477
|
// These tools will be removed from the set of tools available to the model
|
|
486
478
|
excludedTools: z5.array(z5.string()).optional(),
|
|
@@ -577,8 +569,6 @@ var anthropicModels = {
|
|
|
577
569
|
// Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
|
|
578
570
|
supportsImages: true,
|
|
579
571
|
supportsPromptCache: true,
|
|
580
|
-
supportsNativeTools: true,
|
|
581
|
-
defaultToolProtocol: "native",
|
|
582
572
|
inputPrice: 3,
|
|
583
573
|
// $3 per million input tokens (≤200K context)
|
|
584
574
|
outputPrice: 15,
|
|
@@ -611,8 +601,6 @@ var anthropicModels = {
|
|
|
611
601
|
// Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
|
|
612
602
|
supportsImages: true,
|
|
613
603
|
supportsPromptCache: true,
|
|
614
|
-
supportsNativeTools: true,
|
|
615
|
-
defaultToolProtocol: "native",
|
|
616
604
|
inputPrice: 3,
|
|
617
605
|
// $3 per million input tokens (≤200K context)
|
|
618
606
|
outputPrice: 15,
|
|
@@ -644,8 +632,6 @@ var anthropicModels = {
|
|
|
644
632
|
contextWindow: 2e5,
|
|
645
633
|
supportsImages: true,
|
|
646
634
|
supportsPromptCache: true,
|
|
647
|
-
supportsNativeTools: true,
|
|
648
|
-
defaultToolProtocol: "native",
|
|
649
635
|
inputPrice: 5,
|
|
650
636
|
// $5 per million input tokens
|
|
651
637
|
outputPrice: 25,
|
|
@@ -662,8 +648,6 @@ var anthropicModels = {
|
|
|
662
648
|
contextWindow: 2e5,
|
|
663
649
|
supportsImages: true,
|
|
664
650
|
supportsPromptCache: true,
|
|
665
|
-
supportsNativeTools: true,
|
|
666
|
-
defaultToolProtocol: "native",
|
|
667
651
|
inputPrice: 15,
|
|
668
652
|
// $15 per million input tokens
|
|
669
653
|
outputPrice: 75,
|
|
@@ -680,8 +664,6 @@ var anthropicModels = {
|
|
|
680
664
|
contextWindow: 2e5,
|
|
681
665
|
supportsImages: true,
|
|
682
666
|
supportsPromptCache: true,
|
|
683
|
-
supportsNativeTools: true,
|
|
684
|
-
defaultToolProtocol: "native",
|
|
685
667
|
inputPrice: 15,
|
|
686
668
|
// $15 per million input tokens
|
|
687
669
|
outputPrice: 75,
|
|
@@ -698,8 +680,6 @@ var anthropicModels = {
|
|
|
698
680
|
contextWindow: 2e5,
|
|
699
681
|
supportsImages: true,
|
|
700
682
|
supportsPromptCache: true,
|
|
701
|
-
supportsNativeTools: true,
|
|
702
|
-
defaultToolProtocol: "native",
|
|
703
683
|
inputPrice: 3,
|
|
704
684
|
// $3 per million input tokens
|
|
705
685
|
outputPrice: 15,
|
|
@@ -717,8 +697,6 @@ var anthropicModels = {
|
|
|
717
697
|
contextWindow: 2e5,
|
|
718
698
|
supportsImages: true,
|
|
719
699
|
supportsPromptCache: true,
|
|
720
|
-
supportsNativeTools: true,
|
|
721
|
-
defaultToolProtocol: "native",
|
|
722
700
|
inputPrice: 3,
|
|
723
701
|
// $3 per million input tokens
|
|
724
702
|
outputPrice: 15,
|
|
@@ -733,8 +711,6 @@ var anthropicModels = {
|
|
|
733
711
|
contextWindow: 2e5,
|
|
734
712
|
supportsImages: true,
|
|
735
713
|
supportsPromptCache: true,
|
|
736
|
-
supportsNativeTools: true,
|
|
737
|
-
defaultToolProtocol: "native",
|
|
738
714
|
inputPrice: 3,
|
|
739
715
|
// $3 per million input tokens
|
|
740
716
|
outputPrice: 15,
|
|
@@ -749,8 +725,6 @@ var anthropicModels = {
|
|
|
749
725
|
contextWindow: 2e5,
|
|
750
726
|
supportsImages: false,
|
|
751
727
|
supportsPromptCache: true,
|
|
752
|
-
supportsNativeTools: true,
|
|
753
|
-
defaultToolProtocol: "native",
|
|
754
728
|
inputPrice: 1,
|
|
755
729
|
outputPrice: 5,
|
|
756
730
|
cacheWritesPrice: 1.25,
|
|
@@ -761,8 +735,6 @@ var anthropicModels = {
|
|
|
761
735
|
contextWindow: 2e5,
|
|
762
736
|
supportsImages: true,
|
|
763
737
|
supportsPromptCache: true,
|
|
764
|
-
supportsNativeTools: true,
|
|
765
|
-
defaultToolProtocol: "native",
|
|
766
738
|
inputPrice: 15,
|
|
767
739
|
outputPrice: 75,
|
|
768
740
|
cacheWritesPrice: 18.75,
|
|
@@ -773,8 +745,6 @@ var anthropicModels = {
|
|
|
773
745
|
contextWindow: 2e5,
|
|
774
746
|
supportsImages: true,
|
|
775
747
|
supportsPromptCache: true,
|
|
776
|
-
supportsNativeTools: true,
|
|
777
|
-
defaultToolProtocol: "native",
|
|
778
748
|
inputPrice: 0.25,
|
|
779
749
|
outputPrice: 1.25,
|
|
780
750
|
cacheWritesPrice: 0.3,
|
|
@@ -785,8 +755,6 @@ var anthropicModels = {
|
|
|
785
755
|
contextWindow: 2e5,
|
|
786
756
|
supportsImages: true,
|
|
787
757
|
supportsPromptCache: true,
|
|
788
|
-
supportsNativeTools: true,
|
|
789
|
-
defaultToolProtocol: "native",
|
|
790
758
|
inputPrice: 1,
|
|
791
759
|
outputPrice: 5,
|
|
792
760
|
cacheWritesPrice: 1.25,
|
|
@@ -804,7 +772,6 @@ var basetenModels = {
|
|
|
804
772
|
contextWindow: 262e3,
|
|
805
773
|
supportsImages: false,
|
|
806
774
|
supportsPromptCache: false,
|
|
807
|
-
supportsNativeTools: true,
|
|
808
775
|
inputPrice: 0.6,
|
|
809
776
|
outputPrice: 2.5,
|
|
810
777
|
cacheWritesPrice: 0,
|
|
@@ -816,7 +783,6 @@ var basetenModels = {
|
|
|
816
783
|
contextWindow: 2e5,
|
|
817
784
|
supportsImages: false,
|
|
818
785
|
supportsPromptCache: false,
|
|
819
|
-
supportsNativeTools: true,
|
|
820
786
|
inputPrice: 0.6,
|
|
821
787
|
outputPrice: 2.2,
|
|
822
788
|
cacheWritesPrice: 0,
|
|
@@ -828,7 +794,6 @@ var basetenModels = {
|
|
|
828
794
|
contextWindow: 163840,
|
|
829
795
|
supportsImages: false,
|
|
830
796
|
supportsPromptCache: false,
|
|
831
|
-
supportsNativeTools: true,
|
|
832
797
|
inputPrice: 2.55,
|
|
833
798
|
outputPrice: 5.95,
|
|
834
799
|
cacheWritesPrice: 0,
|
|
@@ -840,7 +805,6 @@ var basetenModels = {
|
|
|
840
805
|
contextWindow: 163840,
|
|
841
806
|
supportsImages: false,
|
|
842
807
|
supportsPromptCache: false,
|
|
843
|
-
supportsNativeTools: true,
|
|
844
808
|
inputPrice: 2.55,
|
|
845
809
|
outputPrice: 5.95,
|
|
846
810
|
cacheWritesPrice: 0,
|
|
@@ -852,7 +816,6 @@ var basetenModels = {
|
|
|
852
816
|
contextWindow: 163840,
|
|
853
817
|
supportsImages: false,
|
|
854
818
|
supportsPromptCache: false,
|
|
855
|
-
supportsNativeTools: true,
|
|
856
819
|
inputPrice: 0.77,
|
|
857
820
|
outputPrice: 0.77,
|
|
858
821
|
cacheWritesPrice: 0,
|
|
@@ -864,7 +827,6 @@ var basetenModels = {
|
|
|
864
827
|
contextWindow: 163840,
|
|
865
828
|
supportsImages: false,
|
|
866
829
|
supportsPromptCache: false,
|
|
867
|
-
supportsNativeTools: true,
|
|
868
830
|
inputPrice: 0.5,
|
|
869
831
|
outputPrice: 1.5,
|
|
870
832
|
cacheWritesPrice: 0,
|
|
@@ -876,7 +838,6 @@ var basetenModels = {
|
|
|
876
838
|
contextWindow: 163840,
|
|
877
839
|
supportsImages: false,
|
|
878
840
|
supportsPromptCache: false,
|
|
879
|
-
supportsNativeTools: true,
|
|
880
841
|
inputPrice: 0.3,
|
|
881
842
|
outputPrice: 0.45,
|
|
882
843
|
cacheWritesPrice: 0,
|
|
@@ -888,7 +849,6 @@ var basetenModels = {
|
|
|
888
849
|
contextWindow: 128072,
|
|
889
850
|
supportsImages: false,
|
|
890
851
|
supportsPromptCache: false,
|
|
891
|
-
supportsNativeTools: true,
|
|
892
852
|
inputPrice: 0.1,
|
|
893
853
|
outputPrice: 0.5,
|
|
894
854
|
cacheWritesPrice: 0,
|
|
@@ -900,7 +860,6 @@ var basetenModels = {
|
|
|
900
860
|
contextWindow: 262144,
|
|
901
861
|
supportsImages: false,
|
|
902
862
|
supportsPromptCache: false,
|
|
903
|
-
supportsNativeTools: true,
|
|
904
863
|
inputPrice: 0.22,
|
|
905
864
|
outputPrice: 0.8,
|
|
906
865
|
cacheWritesPrice: 0,
|
|
@@ -912,7 +871,6 @@ var basetenModels = {
|
|
|
912
871
|
contextWindow: 262144,
|
|
913
872
|
supportsImages: false,
|
|
914
873
|
supportsPromptCache: false,
|
|
915
|
-
supportsNativeTools: true,
|
|
916
874
|
inputPrice: 0.38,
|
|
917
875
|
outputPrice: 1.53,
|
|
918
876
|
cacheWritesPrice: 0,
|
|
@@ -924,7 +882,6 @@ var basetenModels = {
|
|
|
924
882
|
contextWindow: 262e3,
|
|
925
883
|
supportsImages: false,
|
|
926
884
|
supportsPromptCache: false,
|
|
927
|
-
supportsNativeTools: true,
|
|
928
885
|
inputPrice: 0.6,
|
|
929
886
|
outputPrice: 2.5,
|
|
930
887
|
cacheWritesPrice: 0,
|
|
@@ -944,8 +901,6 @@ var bedrockModels = {
|
|
|
944
901
|
supportsImages: true,
|
|
945
902
|
supportsPromptCache: true,
|
|
946
903
|
supportsReasoningBudget: true,
|
|
947
|
-
supportsNativeTools: true,
|
|
948
|
-
defaultToolProtocol: "native",
|
|
949
904
|
inputPrice: 3,
|
|
950
905
|
outputPrice: 15,
|
|
951
906
|
cacheWritesPrice: 3.75,
|
|
@@ -959,7 +914,6 @@ var bedrockModels = {
|
|
|
959
914
|
contextWindow: 3e5,
|
|
960
915
|
supportsImages: true,
|
|
961
916
|
supportsPromptCache: true,
|
|
962
|
-
supportsNativeTools: true,
|
|
963
917
|
inputPrice: 0.8,
|
|
964
918
|
outputPrice: 3.2,
|
|
965
919
|
cacheWritesPrice: 0.8,
|
|
@@ -975,7 +929,6 @@ var bedrockModels = {
|
|
|
975
929
|
contextWindow: 3e5,
|
|
976
930
|
supportsImages: true,
|
|
977
931
|
supportsPromptCache: false,
|
|
978
|
-
supportsNativeTools: true,
|
|
979
932
|
inputPrice: 1,
|
|
980
933
|
outputPrice: 4,
|
|
981
934
|
cacheWritesPrice: 1,
|
|
@@ -989,7 +942,6 @@ var bedrockModels = {
|
|
|
989
942
|
contextWindow: 3e5,
|
|
990
943
|
supportsImages: true,
|
|
991
944
|
supportsPromptCache: true,
|
|
992
|
-
supportsNativeTools: true,
|
|
993
945
|
inputPrice: 0.06,
|
|
994
946
|
outputPrice: 0.24,
|
|
995
947
|
cacheWritesPrice: 0.06,
|
|
@@ -1005,7 +957,6 @@ var bedrockModels = {
|
|
|
1005
957
|
contextWindow: 1e6,
|
|
1006
958
|
supportsImages: true,
|
|
1007
959
|
supportsPromptCache: true,
|
|
1008
|
-
supportsNativeTools: true,
|
|
1009
960
|
inputPrice: 0.33,
|
|
1010
961
|
outputPrice: 2.75,
|
|
1011
962
|
cacheWritesPrice: 0,
|
|
@@ -1021,7 +972,6 @@ var bedrockModels = {
|
|
|
1021
972
|
contextWindow: 128e3,
|
|
1022
973
|
supportsImages: false,
|
|
1023
974
|
supportsPromptCache: true,
|
|
1024
|
-
supportsNativeTools: true,
|
|
1025
975
|
inputPrice: 0.035,
|
|
1026
976
|
outputPrice: 0.14,
|
|
1027
977
|
cacheWritesPrice: 0.035,
|
|
@@ -1038,8 +988,6 @@ var bedrockModels = {
|
|
|
1038
988
|
supportsImages: true,
|
|
1039
989
|
supportsPromptCache: true,
|
|
1040
990
|
supportsReasoningBudget: true,
|
|
1041
|
-
supportsNativeTools: true,
|
|
1042
|
-
defaultToolProtocol: "native",
|
|
1043
991
|
inputPrice: 3,
|
|
1044
992
|
outputPrice: 15,
|
|
1045
993
|
cacheWritesPrice: 3.75,
|
|
@@ -1054,8 +1002,6 @@ var bedrockModels = {
|
|
|
1054
1002
|
supportsImages: true,
|
|
1055
1003
|
supportsPromptCache: true,
|
|
1056
1004
|
supportsReasoningBudget: true,
|
|
1057
|
-
supportsNativeTools: true,
|
|
1058
|
-
defaultToolProtocol: "native",
|
|
1059
1005
|
inputPrice: 15,
|
|
1060
1006
|
outputPrice: 75,
|
|
1061
1007
|
cacheWritesPrice: 18.75,
|
|
@@ -1070,8 +1016,6 @@ var bedrockModels = {
|
|
|
1070
1016
|
supportsImages: true,
|
|
1071
1017
|
supportsPromptCache: true,
|
|
1072
1018
|
supportsReasoningBudget: true,
|
|
1073
|
-
supportsNativeTools: true,
|
|
1074
|
-
defaultToolProtocol: "native",
|
|
1075
1019
|
inputPrice: 5,
|
|
1076
1020
|
outputPrice: 25,
|
|
1077
1021
|
cacheWritesPrice: 6.25,
|
|
@@ -1086,8 +1030,6 @@ var bedrockModels = {
|
|
|
1086
1030
|
supportsImages: true,
|
|
1087
1031
|
supportsPromptCache: true,
|
|
1088
1032
|
supportsReasoningBudget: true,
|
|
1089
|
-
supportsNativeTools: true,
|
|
1090
|
-
defaultToolProtocol: "native",
|
|
1091
1033
|
inputPrice: 15,
|
|
1092
1034
|
outputPrice: 75,
|
|
1093
1035
|
cacheWritesPrice: 18.75,
|
|
@@ -1102,8 +1044,6 @@ var bedrockModels = {
|
|
|
1102
1044
|
supportsImages: true,
|
|
1103
1045
|
supportsPromptCache: true,
|
|
1104
1046
|
supportsReasoningBudget: true,
|
|
1105
|
-
supportsNativeTools: true,
|
|
1106
|
-
defaultToolProtocol: "native",
|
|
1107
1047
|
inputPrice: 3,
|
|
1108
1048
|
outputPrice: 15,
|
|
1109
1049
|
cacheWritesPrice: 3.75,
|
|
@@ -1117,8 +1057,6 @@ var bedrockModels = {
|
|
|
1117
1057
|
contextWindow: 2e5,
|
|
1118
1058
|
supportsImages: true,
|
|
1119
1059
|
supportsPromptCache: true,
|
|
1120
|
-
supportsNativeTools: true,
|
|
1121
|
-
defaultToolProtocol: "native",
|
|
1122
1060
|
inputPrice: 3,
|
|
1123
1061
|
outputPrice: 15,
|
|
1124
1062
|
cacheWritesPrice: 3.75,
|
|
@@ -1132,8 +1070,6 @@ var bedrockModels = {
|
|
|
1132
1070
|
contextWindow: 2e5,
|
|
1133
1071
|
supportsImages: false,
|
|
1134
1072
|
supportsPromptCache: true,
|
|
1135
|
-
supportsNativeTools: true,
|
|
1136
|
-
defaultToolProtocol: "native",
|
|
1137
1073
|
inputPrice: 0.8,
|
|
1138
1074
|
outputPrice: 4,
|
|
1139
1075
|
cacheWritesPrice: 1,
|
|
@@ -1148,8 +1084,6 @@ var bedrockModels = {
|
|
|
1148
1084
|
supportsImages: true,
|
|
1149
1085
|
supportsPromptCache: true,
|
|
1150
1086
|
supportsReasoningBudget: true,
|
|
1151
|
-
supportsNativeTools: true,
|
|
1152
|
-
defaultToolProtocol: "native",
|
|
1153
1087
|
inputPrice: 1,
|
|
1154
1088
|
outputPrice: 5,
|
|
1155
1089
|
cacheWritesPrice: 1.25,
|
|
@@ -1165,8 +1099,6 @@ var bedrockModels = {
|
|
|
1165
1099
|
contextWindow: 2e5,
|
|
1166
1100
|
supportsImages: true,
|
|
1167
1101
|
supportsPromptCache: false,
|
|
1168
|
-
supportsNativeTools: true,
|
|
1169
|
-
defaultToolProtocol: "native",
|
|
1170
1102
|
inputPrice: 3,
|
|
1171
1103
|
outputPrice: 15
|
|
1172
1104
|
},
|
|
@@ -1175,8 +1107,6 @@ var bedrockModels = {
|
|
|
1175
1107
|
contextWindow: 2e5,
|
|
1176
1108
|
supportsImages: true,
|
|
1177
1109
|
supportsPromptCache: false,
|
|
1178
|
-
supportsNativeTools: true,
|
|
1179
|
-
defaultToolProtocol: "native",
|
|
1180
1110
|
inputPrice: 15,
|
|
1181
1111
|
outputPrice: 75
|
|
1182
1112
|
},
|
|
@@ -1185,8 +1115,6 @@ var bedrockModels = {
|
|
|
1185
1115
|
contextWindow: 2e5,
|
|
1186
1116
|
supportsImages: true,
|
|
1187
1117
|
supportsPromptCache: false,
|
|
1188
|
-
supportsNativeTools: true,
|
|
1189
|
-
defaultToolProtocol: "native",
|
|
1190
1118
|
inputPrice: 3,
|
|
1191
1119
|
outputPrice: 15
|
|
1192
1120
|
},
|
|
@@ -1195,8 +1123,6 @@ var bedrockModels = {
|
|
|
1195
1123
|
contextWindow: 2e5,
|
|
1196
1124
|
supportsImages: true,
|
|
1197
1125
|
supportsPromptCache: false,
|
|
1198
|
-
supportsNativeTools: true,
|
|
1199
|
-
defaultToolProtocol: "native",
|
|
1200
1126
|
inputPrice: 0.25,
|
|
1201
1127
|
outputPrice: 1.25
|
|
1202
1128
|
},
|
|
@@ -1205,7 +1131,6 @@ var bedrockModels = {
|
|
|
1205
1131
|
contextWindow: 128e3,
|
|
1206
1132
|
supportsImages: false,
|
|
1207
1133
|
supportsPromptCache: false,
|
|
1208
|
-
supportsNativeTools: true,
|
|
1209
1134
|
inputPrice: 1.35,
|
|
1210
1135
|
outputPrice: 5.4
|
|
1211
1136
|
},
|
|
@@ -1214,7 +1139,6 @@ var bedrockModels = {
|
|
|
1214
1139
|
contextWindow: 128e3,
|
|
1215
1140
|
supportsImages: false,
|
|
1216
1141
|
supportsPromptCache: false,
|
|
1217
|
-
supportsNativeTools: true,
|
|
1218
1142
|
inputPrice: 0.5,
|
|
1219
1143
|
outputPrice: 1.5,
|
|
1220
1144
|
description: "GPT-OSS 20B - Optimized for low latency and local/specialized use cases"
|
|
@@ -1224,7 +1148,6 @@ var bedrockModels = {
|
|
|
1224
1148
|
contextWindow: 128e3,
|
|
1225
1149
|
supportsImages: false,
|
|
1226
1150
|
supportsPromptCache: false,
|
|
1227
|
-
supportsNativeTools: true,
|
|
1228
1151
|
inputPrice: 2,
|
|
1229
1152
|
outputPrice: 6,
|
|
1230
1153
|
description: "GPT-OSS 120B - Production-ready, general-purpose, high-reasoning model"
|
|
@@ -1234,7 +1157,6 @@ var bedrockModels = {
|
|
|
1234
1157
|
contextWindow: 128e3,
|
|
1235
1158
|
supportsImages: false,
|
|
1236
1159
|
supportsPromptCache: false,
|
|
1237
|
-
supportsNativeTools: true,
|
|
1238
1160
|
inputPrice: 0.72,
|
|
1239
1161
|
outputPrice: 0.72,
|
|
1240
1162
|
description: "Llama 3.3 Instruct (70B)"
|
|
@@ -1244,7 +1166,6 @@ var bedrockModels = {
|
|
|
1244
1166
|
contextWindow: 128e3,
|
|
1245
1167
|
supportsImages: true,
|
|
1246
1168
|
supportsPromptCache: false,
|
|
1247
|
-
supportsNativeTools: true,
|
|
1248
1169
|
inputPrice: 0.72,
|
|
1249
1170
|
outputPrice: 0.72,
|
|
1250
1171
|
description: "Llama 3.2 Instruct (90B)"
|
|
@@ -1254,7 +1175,6 @@ var bedrockModels = {
|
|
|
1254
1175
|
contextWindow: 128e3,
|
|
1255
1176
|
supportsImages: true,
|
|
1256
1177
|
supportsPromptCache: false,
|
|
1257
|
-
supportsNativeTools: true,
|
|
1258
1178
|
inputPrice: 0.16,
|
|
1259
1179
|
outputPrice: 0.16,
|
|
1260
1180
|
description: "Llama 3.2 Instruct (11B)"
|
|
@@ -1264,7 +1184,6 @@ var bedrockModels = {
|
|
|
1264
1184
|
contextWindow: 128e3,
|
|
1265
1185
|
supportsImages: false,
|
|
1266
1186
|
supportsPromptCache: false,
|
|
1267
|
-
supportsNativeTools: true,
|
|
1268
1187
|
inputPrice: 0.15,
|
|
1269
1188
|
outputPrice: 0.15,
|
|
1270
1189
|
description: "Llama 3.2 Instruct (3B)"
|
|
@@ -1274,7 +1193,6 @@ var bedrockModels = {
|
|
|
1274
1193
|
contextWindow: 128e3,
|
|
1275
1194
|
supportsImages: false,
|
|
1276
1195
|
supportsPromptCache: false,
|
|
1277
|
-
supportsNativeTools: true,
|
|
1278
1196
|
inputPrice: 0.1,
|
|
1279
1197
|
outputPrice: 0.1,
|
|
1280
1198
|
description: "Llama 3.2 Instruct (1B)"
|
|
@@ -1284,7 +1202,6 @@ var bedrockModels = {
|
|
|
1284
1202
|
contextWindow: 128e3,
|
|
1285
1203
|
supportsImages: false,
|
|
1286
1204
|
supportsPromptCache: false,
|
|
1287
|
-
supportsNativeTools: true,
|
|
1288
1205
|
inputPrice: 2.4,
|
|
1289
1206
|
outputPrice: 2.4,
|
|
1290
1207
|
description: "Llama 3.1 Instruct (405B)"
|
|
@@ -1294,7 +1211,6 @@ var bedrockModels = {
|
|
|
1294
1211
|
contextWindow: 128e3,
|
|
1295
1212
|
supportsImages: false,
|
|
1296
1213
|
supportsPromptCache: false,
|
|
1297
|
-
supportsNativeTools: true,
|
|
1298
1214
|
inputPrice: 0.72,
|
|
1299
1215
|
outputPrice: 0.72,
|
|
1300
1216
|
description: "Llama 3.1 Instruct (70B)"
|
|
@@ -1304,7 +1220,6 @@ var bedrockModels = {
|
|
|
1304
1220
|
contextWindow: 128e3,
|
|
1305
1221
|
supportsImages: false,
|
|
1306
1222
|
supportsPromptCache: false,
|
|
1307
|
-
supportsNativeTools: true,
|
|
1308
1223
|
inputPrice: 0.9,
|
|
1309
1224
|
outputPrice: 0.9,
|
|
1310
1225
|
description: "Llama 3.1 Instruct (70B) (w/ latency optimized inference)"
|
|
@@ -1314,7 +1229,6 @@ var bedrockModels = {
|
|
|
1314
1229
|
contextWindow: 8e3,
|
|
1315
1230
|
supportsImages: false,
|
|
1316
1231
|
supportsPromptCache: false,
|
|
1317
|
-
supportsNativeTools: true,
|
|
1318
1232
|
inputPrice: 0.22,
|
|
1319
1233
|
outputPrice: 0.22,
|
|
1320
1234
|
description: "Llama 3.1 Instruct (8B)"
|
|
@@ -1324,7 +1238,6 @@ var bedrockModels = {
|
|
|
1324
1238
|
contextWindow: 8e3,
|
|
1325
1239
|
supportsImages: false,
|
|
1326
1240
|
supportsPromptCache: false,
|
|
1327
|
-
supportsNativeTools: true,
|
|
1328
1241
|
inputPrice: 2.65,
|
|
1329
1242
|
outputPrice: 3.5
|
|
1330
1243
|
},
|
|
@@ -1333,7 +1246,6 @@ var bedrockModels = {
|
|
|
1333
1246
|
contextWindow: 4e3,
|
|
1334
1247
|
supportsImages: false,
|
|
1335
1248
|
supportsPromptCache: false,
|
|
1336
|
-
supportsNativeTools: true,
|
|
1337
1249
|
inputPrice: 0.3,
|
|
1338
1250
|
outputPrice: 0.6
|
|
1339
1251
|
},
|
|
@@ -1342,7 +1254,6 @@ var bedrockModels = {
|
|
|
1342
1254
|
contextWindow: 8e3,
|
|
1343
1255
|
supportsImages: false,
|
|
1344
1256
|
supportsPromptCache: false,
|
|
1345
|
-
supportsNativeTools: true,
|
|
1346
1257
|
inputPrice: 0.15,
|
|
1347
1258
|
outputPrice: 0.2,
|
|
1348
1259
|
description: "Amazon Titan Text Lite"
|
|
@@ -1352,7 +1263,6 @@ var bedrockModels = {
|
|
|
1352
1263
|
contextWindow: 8e3,
|
|
1353
1264
|
supportsImages: false,
|
|
1354
1265
|
supportsPromptCache: false,
|
|
1355
|
-
supportsNativeTools: true,
|
|
1356
1266
|
inputPrice: 0.2,
|
|
1357
1267
|
outputPrice: 0.6,
|
|
1358
1268
|
description: "Amazon Titan Text Express"
|
|
@@ -1362,8 +1272,6 @@ var bedrockModels = {
|
|
|
1362
1272
|
contextWindow: 262144,
|
|
1363
1273
|
supportsImages: false,
|
|
1364
1274
|
supportsPromptCache: false,
|
|
1365
|
-
supportsNativeTools: true,
|
|
1366
|
-
defaultToolProtocol: "native",
|
|
1367
1275
|
preserveReasoning: true,
|
|
1368
1276
|
inputPrice: 0.6,
|
|
1369
1277
|
outputPrice: 2.5,
|
|
@@ -1374,8 +1282,6 @@ var bedrockModels = {
|
|
|
1374
1282
|
contextWindow: 196608,
|
|
1375
1283
|
supportsImages: false,
|
|
1376
1284
|
supportsPromptCache: false,
|
|
1377
|
-
supportsNativeTools: true,
|
|
1378
|
-
defaultToolProtocol: "native",
|
|
1379
1285
|
preserveReasoning: true,
|
|
1380
1286
|
inputPrice: 0.3,
|
|
1381
1287
|
outputPrice: 1.2,
|
|
@@ -1386,8 +1292,6 @@ var bedrockModels = {
|
|
|
1386
1292
|
contextWindow: 262144,
|
|
1387
1293
|
supportsImages: false,
|
|
1388
1294
|
supportsPromptCache: false,
|
|
1389
|
-
supportsNativeTools: true,
|
|
1390
|
-
defaultToolProtocol: "native",
|
|
1391
1295
|
inputPrice: 0.15,
|
|
1392
1296
|
outputPrice: 1.2,
|
|
1393
1297
|
description: "Qwen3 Next 80B (MoE model with 3B active parameters)"
|
|
@@ -1397,8 +1301,6 @@ var bedrockModels = {
|
|
|
1397
1301
|
contextWindow: 262144,
|
|
1398
1302
|
supportsImages: false,
|
|
1399
1303
|
supportsPromptCache: false,
|
|
1400
|
-
supportsNativeTools: true,
|
|
1401
|
-
defaultToolProtocol: "native",
|
|
1402
1304
|
inputPrice: 0.45,
|
|
1403
1305
|
outputPrice: 1.8,
|
|
1404
1306
|
description: "Qwen3 Coder 480B (MoE model with 35B active parameters)"
|
|
@@ -1489,26 +1391,14 @@ var BEDROCK_SERVICE_TIER_PRICING = {
|
|
|
1489
1391
|
// src/providers/cerebras.ts
|
|
1490
1392
|
var cerebrasDefaultModelId = "gpt-oss-120b";
|
|
1491
1393
|
var cerebrasModels = {
|
|
1492
|
-
"zai-glm-4.6": {
|
|
1493
|
-
maxTokens: 16384,
|
|
1494
|
-
// Conservative default to avoid premature rate limiting (Cerebras reserves quota upfront)
|
|
1495
|
-
contextWindow: 131072,
|
|
1496
|
-
supportsImages: false,
|
|
1497
|
-
supportsPromptCache: false,
|
|
1498
|
-
supportsNativeTools: true,
|
|
1499
|
-
defaultToolProtocol: "native",
|
|
1500
|
-
inputPrice: 0,
|
|
1501
|
-
outputPrice: 0,
|
|
1502
|
-
description: "Fast general-purpose model on Cerebras (up to 1,000 tokens/s). To be deprecated soon."
|
|
1503
|
-
},
|
|
1504
1394
|
"zai-glm-4.7": {
|
|
1505
1395
|
maxTokens: 16384,
|
|
1506
1396
|
// Conservative default to avoid premature rate limiting (Cerebras reserves quota upfront)
|
|
1507
1397
|
contextWindow: 131072,
|
|
1508
1398
|
supportsImages: false,
|
|
1509
|
-
supportsPromptCache:
|
|
1510
|
-
|
|
1511
|
-
|
|
1399
|
+
supportsPromptCache: true,
|
|
1400
|
+
supportsTemperature: true,
|
|
1401
|
+
defaultTemperature: 1,
|
|
1512
1402
|
inputPrice: 0,
|
|
1513
1403
|
outputPrice: 0,
|
|
1514
1404
|
description: "Highly capable general-purpose model on Cerebras (up to 1,000 tokens/s), competitive with leading proprietary models on coding tasks."
|
|
@@ -1519,8 +1409,6 @@ var cerebrasModels = {
|
|
|
1519
1409
|
contextWindow: 64e3,
|
|
1520
1410
|
supportsImages: false,
|
|
1521
1411
|
supportsPromptCache: false,
|
|
1522
|
-
supportsNativeTools: true,
|
|
1523
|
-
defaultToolProtocol: "native",
|
|
1524
1412
|
inputPrice: 0,
|
|
1525
1413
|
outputPrice: 0,
|
|
1526
1414
|
description: "Intelligent model with ~1400 tokens/s"
|
|
@@ -1531,8 +1419,6 @@ var cerebrasModels = {
|
|
|
1531
1419
|
contextWindow: 64e3,
|
|
1532
1420
|
supportsImages: false,
|
|
1533
1421
|
supportsPromptCache: false,
|
|
1534
|
-
supportsNativeTools: true,
|
|
1535
|
-
defaultToolProtocol: "native",
|
|
1536
1422
|
inputPrice: 0,
|
|
1537
1423
|
outputPrice: 0,
|
|
1538
1424
|
description: "Powerful model with ~2600 tokens/s"
|
|
@@ -1543,8 +1429,6 @@ var cerebrasModels = {
|
|
|
1543
1429
|
contextWindow: 64e3,
|
|
1544
1430
|
supportsImages: false,
|
|
1545
1431
|
supportsPromptCache: false,
|
|
1546
|
-
supportsNativeTools: true,
|
|
1547
|
-
defaultToolProtocol: "native",
|
|
1548
1432
|
inputPrice: 0,
|
|
1549
1433
|
outputPrice: 0,
|
|
1550
1434
|
description: "SOTA coding performance with ~2500 tokens/s"
|
|
@@ -1555,8 +1439,6 @@ var cerebrasModels = {
|
|
|
1555
1439
|
contextWindow: 64e3,
|
|
1556
1440
|
supportsImages: false,
|
|
1557
1441
|
supportsPromptCache: false,
|
|
1558
|
-
supportsNativeTools: true,
|
|
1559
|
-
defaultToolProtocol: "native",
|
|
1560
1442
|
inputPrice: 0,
|
|
1561
1443
|
outputPrice: 0,
|
|
1562
1444
|
description: "OpenAI GPT OSS model with ~2800 tokens/s\n\n\u2022 64K context window\n\u2022 Excels at efficient reasoning across science, math, and coding"
|
|
@@ -1571,8 +1453,6 @@ var chutesModels = {
|
|
|
1571
1453
|
contextWindow: 163840,
|
|
1572
1454
|
supportsImages: false,
|
|
1573
1455
|
supportsPromptCache: false,
|
|
1574
|
-
supportsNativeTools: true,
|
|
1575
|
-
defaultToolProtocol: "native",
|
|
1576
1456
|
inputPrice: 0,
|
|
1577
1457
|
outputPrice: 0,
|
|
1578
1458
|
description: "DeepSeek R1 0528 model."
|
|
@@ -1582,8 +1462,6 @@ var chutesModels = {
|
|
|
1582
1462
|
contextWindow: 163840,
|
|
1583
1463
|
supportsImages: false,
|
|
1584
1464
|
supportsPromptCache: false,
|
|
1585
|
-
supportsNativeTools: true,
|
|
1586
|
-
defaultToolProtocol: "native",
|
|
1587
1465
|
inputPrice: 0,
|
|
1588
1466
|
outputPrice: 0,
|
|
1589
1467
|
description: "DeepSeek R1 model."
|
|
@@ -1593,8 +1471,6 @@ var chutesModels = {
|
|
|
1593
1471
|
contextWindow: 163840,
|
|
1594
1472
|
supportsImages: false,
|
|
1595
1473
|
supportsPromptCache: false,
|
|
1596
|
-
supportsNativeTools: true,
|
|
1597
|
-
defaultToolProtocol: "native",
|
|
1598
1474
|
inputPrice: 0,
|
|
1599
1475
|
outputPrice: 0,
|
|
1600
1476
|
description: "DeepSeek V3 model."
|
|
@@ -1604,8 +1480,6 @@ var chutesModels = {
|
|
|
1604
1480
|
contextWindow: 163840,
|
|
1605
1481
|
supportsImages: false,
|
|
1606
1482
|
supportsPromptCache: false,
|
|
1607
|
-
supportsNativeTools: true,
|
|
1608
|
-
defaultToolProtocol: "native",
|
|
1609
1483
|
inputPrice: 0,
|
|
1610
1484
|
outputPrice: 0,
|
|
1611
1485
|
description: "DeepSeek V3.1 model."
|
|
@@ -1615,8 +1489,6 @@ var chutesModels = {
|
|
|
1615
1489
|
contextWindow: 163840,
|
|
1616
1490
|
supportsImages: false,
|
|
1617
1491
|
supportsPromptCache: false,
|
|
1618
|
-
supportsNativeTools: true,
|
|
1619
|
-
defaultToolProtocol: "native",
|
|
1620
1492
|
inputPrice: 0.23,
|
|
1621
1493
|
outputPrice: 0.9,
|
|
1622
1494
|
description: "DeepSeek\u2011V3.1\u2011Terminus is an update to V3.1 that improves language consistency by reducing CN/EN mix\u2011ups and eliminating random characters, while strengthening agent capabilities with notably better Code Agent and Search Agent performance."
|
|
@@ -1626,8 +1498,6 @@ var chutesModels = {
|
|
|
1626
1498
|
contextWindow: 163840,
|
|
1627
1499
|
supportsImages: false,
|
|
1628
1500
|
supportsPromptCache: false,
|
|
1629
|
-
supportsNativeTools: true,
|
|
1630
|
-
defaultToolProtocol: "native",
|
|
1631
1501
|
inputPrice: 1,
|
|
1632
1502
|
outputPrice: 3,
|
|
1633
1503
|
description: "DeepSeek-V3.1-turbo is an FP8, speculative-decoding turbo variant optimized for ultra-fast single-shot queries (~200 TPS), with outputs close to the originals and solid function calling/reasoning/structured output, priced at $1/M input and $3/M output tokens, using 2\xD7 quota per request and not intended for bulk workloads."
|
|
@@ -1637,8 +1507,6 @@ var chutesModels = {
|
|
|
1637
1507
|
contextWindow: 163840,
|
|
1638
1508
|
supportsImages: false,
|
|
1639
1509
|
supportsPromptCache: false,
|
|
1640
|
-
supportsNativeTools: true,
|
|
1641
|
-
defaultToolProtocol: "native",
|
|
1642
1510
|
inputPrice: 0.25,
|
|
1643
1511
|
outputPrice: 0.35,
|
|
1644
1512
|
description: "DeepSeek-V3.2-Exp is an experimental LLM that introduces DeepSeek Sparse Attention to improve long\u2011context training and inference efficiency while maintaining performance comparable to V3.1\u2011Terminus."
|
|
@@ -1650,8 +1518,6 @@ var chutesModels = {
|
|
|
1650
1518
|
// From Groq
|
|
1651
1519
|
supportsImages: false,
|
|
1652
1520
|
supportsPromptCache: false,
|
|
1653
|
-
supportsNativeTools: true,
|
|
1654
|
-
defaultToolProtocol: "native",
|
|
1655
1521
|
inputPrice: 0,
|
|
1656
1522
|
outputPrice: 0,
|
|
1657
1523
|
description: "Unsloth Llama 3.3 70B Instruct model."
|
|
@@ -1661,8 +1527,6 @@ var chutesModels = {
|
|
|
1661
1527
|
contextWindow: 512e3,
|
|
1662
1528
|
supportsImages: false,
|
|
1663
1529
|
supportsPromptCache: false,
|
|
1664
|
-
supportsNativeTools: true,
|
|
1665
|
-
defaultToolProtocol: "native",
|
|
1666
1530
|
inputPrice: 0,
|
|
1667
1531
|
outputPrice: 0,
|
|
1668
1532
|
description: "ChutesAI Llama 4 Scout 17B Instruct model, 512K context."
|
|
@@ -1672,8 +1536,6 @@ var chutesModels = {
|
|
|
1672
1536
|
contextWindow: 128e3,
|
|
1673
1537
|
supportsImages: false,
|
|
1674
1538
|
supportsPromptCache: false,
|
|
1675
|
-
supportsNativeTools: true,
|
|
1676
|
-
defaultToolProtocol: "native",
|
|
1677
1539
|
inputPrice: 0,
|
|
1678
1540
|
outputPrice: 0,
|
|
1679
1541
|
description: "Unsloth Mistral Nemo Instruct model."
|
|
@@ -1683,8 +1545,6 @@ var chutesModels = {
|
|
|
1683
1545
|
contextWindow: 131072,
|
|
1684
1546
|
supportsImages: false,
|
|
1685
1547
|
supportsPromptCache: false,
|
|
1686
|
-
supportsNativeTools: true,
|
|
1687
|
-
defaultToolProtocol: "native",
|
|
1688
1548
|
inputPrice: 0,
|
|
1689
1549
|
outputPrice: 0,
|
|
1690
1550
|
description: "Unsloth Gemma 3 12B IT model."
|
|
@@ -1694,8 +1554,6 @@ var chutesModels = {
|
|
|
1694
1554
|
contextWindow: 131072,
|
|
1695
1555
|
supportsImages: false,
|
|
1696
1556
|
supportsPromptCache: false,
|
|
1697
|
-
supportsNativeTools: true,
|
|
1698
|
-
defaultToolProtocol: "native",
|
|
1699
1557
|
inputPrice: 0,
|
|
1700
1558
|
outputPrice: 0,
|
|
1701
1559
|
description: "Nous DeepHermes 3 Llama 3 8B Preview model."
|
|
@@ -1705,8 +1563,6 @@ var chutesModels = {
|
|
|
1705
1563
|
contextWindow: 131072,
|
|
1706
1564
|
supportsImages: false,
|
|
1707
1565
|
supportsPromptCache: false,
|
|
1708
|
-
supportsNativeTools: true,
|
|
1709
|
-
defaultToolProtocol: "native",
|
|
1710
1566
|
inputPrice: 0,
|
|
1711
1567
|
outputPrice: 0,
|
|
1712
1568
|
description: "Unsloth Gemma 3 4B IT model."
|
|
@@ -1716,8 +1572,6 @@ var chutesModels = {
|
|
|
1716
1572
|
contextWindow: 131072,
|
|
1717
1573
|
supportsImages: false,
|
|
1718
1574
|
supportsPromptCache: false,
|
|
1719
|
-
supportsNativeTools: true,
|
|
1720
|
-
defaultToolProtocol: "native",
|
|
1721
1575
|
inputPrice: 0,
|
|
1722
1576
|
outputPrice: 0,
|
|
1723
1577
|
description: "Nvidia Llama 3.3 Nemotron Super 49B model."
|
|
@@ -1727,8 +1581,6 @@ var chutesModels = {
|
|
|
1727
1581
|
contextWindow: 131072,
|
|
1728
1582
|
supportsImages: false,
|
|
1729
1583
|
supportsPromptCache: false,
|
|
1730
|
-
supportsNativeTools: true,
|
|
1731
|
-
defaultToolProtocol: "native",
|
|
1732
1584
|
inputPrice: 0,
|
|
1733
1585
|
outputPrice: 0,
|
|
1734
1586
|
description: "Nvidia Llama 3.1 Nemotron Ultra 253B model."
|
|
@@ -1738,8 +1590,6 @@ var chutesModels = {
|
|
|
1738
1590
|
contextWindow: 256e3,
|
|
1739
1591
|
supportsImages: false,
|
|
1740
1592
|
supportsPromptCache: false,
|
|
1741
|
-
supportsNativeTools: true,
|
|
1742
|
-
defaultToolProtocol: "native",
|
|
1743
1593
|
inputPrice: 0,
|
|
1744
1594
|
outputPrice: 0,
|
|
1745
1595
|
description: "ChutesAI Llama 4 Maverick 17B Instruct FP8 model."
|
|
@@ -1749,8 +1599,6 @@ var chutesModels = {
|
|
|
1749
1599
|
contextWindow: 163840,
|
|
1750
1600
|
supportsImages: false,
|
|
1751
1601
|
supportsPromptCache: false,
|
|
1752
|
-
supportsNativeTools: true,
|
|
1753
|
-
defaultToolProtocol: "native",
|
|
1754
1602
|
inputPrice: 0,
|
|
1755
1603
|
outputPrice: 0,
|
|
1756
1604
|
description: "DeepSeek V3 Base model."
|
|
@@ -1760,8 +1608,6 @@ var chutesModels = {
|
|
|
1760
1608
|
contextWindow: 163840,
|
|
1761
1609
|
supportsImages: false,
|
|
1762
1610
|
supportsPromptCache: false,
|
|
1763
|
-
supportsNativeTools: true,
|
|
1764
|
-
defaultToolProtocol: "native",
|
|
1765
1611
|
inputPrice: 0,
|
|
1766
1612
|
outputPrice: 0,
|
|
1767
1613
|
description: "DeepSeek R1 Zero model."
|
|
@@ -1771,8 +1617,6 @@ var chutesModels = {
|
|
|
1771
1617
|
contextWindow: 163840,
|
|
1772
1618
|
supportsImages: false,
|
|
1773
1619
|
supportsPromptCache: false,
|
|
1774
|
-
supportsNativeTools: true,
|
|
1775
|
-
defaultToolProtocol: "native",
|
|
1776
1620
|
inputPrice: 0,
|
|
1777
1621
|
outputPrice: 0,
|
|
1778
1622
|
description: "DeepSeek V3 (0324) model."
|
|
@@ -1782,8 +1626,6 @@ var chutesModels = {
|
|
|
1782
1626
|
contextWindow: 262144,
|
|
1783
1627
|
supportsImages: false,
|
|
1784
1628
|
supportsPromptCache: false,
|
|
1785
|
-
supportsNativeTools: true,
|
|
1786
|
-
defaultToolProtocol: "native",
|
|
1787
1629
|
inputPrice: 0,
|
|
1788
1630
|
outputPrice: 0,
|
|
1789
1631
|
description: "Qwen3 235B A22B Instruct 2507 model with 262K context window."
|
|
@@ -1793,8 +1635,6 @@ var chutesModels = {
|
|
|
1793
1635
|
contextWindow: 40960,
|
|
1794
1636
|
supportsImages: false,
|
|
1795
1637
|
supportsPromptCache: false,
|
|
1796
|
-
supportsNativeTools: true,
|
|
1797
|
-
defaultToolProtocol: "native",
|
|
1798
1638
|
inputPrice: 0,
|
|
1799
1639
|
outputPrice: 0,
|
|
1800
1640
|
description: "Qwen3 235B A22B model."
|
|
@@ -1804,8 +1644,6 @@ var chutesModels = {
|
|
|
1804
1644
|
contextWindow: 40960,
|
|
1805
1645
|
supportsImages: false,
|
|
1806
1646
|
supportsPromptCache: false,
|
|
1807
|
-
supportsNativeTools: true,
|
|
1808
|
-
defaultToolProtocol: "native",
|
|
1809
1647
|
inputPrice: 0,
|
|
1810
1648
|
outputPrice: 0,
|
|
1811
1649
|
description: "Qwen3 32B model."
|
|
@@ -1815,8 +1653,6 @@ var chutesModels = {
|
|
|
1815
1653
|
contextWindow: 40960,
|
|
1816
1654
|
supportsImages: false,
|
|
1817
1655
|
supportsPromptCache: false,
|
|
1818
|
-
supportsNativeTools: true,
|
|
1819
|
-
defaultToolProtocol: "native",
|
|
1820
1656
|
inputPrice: 0,
|
|
1821
1657
|
outputPrice: 0,
|
|
1822
1658
|
description: "Qwen3 30B A3B model."
|
|
@@ -1826,8 +1662,6 @@ var chutesModels = {
|
|
|
1826
1662
|
contextWindow: 40960,
|
|
1827
1663
|
supportsImages: false,
|
|
1828
1664
|
supportsPromptCache: false,
|
|
1829
|
-
supportsNativeTools: true,
|
|
1830
|
-
defaultToolProtocol: "native",
|
|
1831
1665
|
inputPrice: 0,
|
|
1832
1666
|
outputPrice: 0,
|
|
1833
1667
|
description: "Qwen3 14B model."
|
|
@@ -1837,8 +1671,6 @@ var chutesModels = {
|
|
|
1837
1671
|
contextWindow: 40960,
|
|
1838
1672
|
supportsImages: false,
|
|
1839
1673
|
supportsPromptCache: false,
|
|
1840
|
-
supportsNativeTools: true,
|
|
1841
|
-
defaultToolProtocol: "native",
|
|
1842
1674
|
inputPrice: 0,
|
|
1843
1675
|
outputPrice: 0,
|
|
1844
1676
|
description: "Qwen3 8B model."
|
|
@@ -1848,8 +1680,6 @@ var chutesModels = {
|
|
|
1848
1680
|
contextWindow: 163840,
|
|
1849
1681
|
supportsImages: false,
|
|
1850
1682
|
supportsPromptCache: false,
|
|
1851
|
-
supportsNativeTools: true,
|
|
1852
|
-
defaultToolProtocol: "native",
|
|
1853
1683
|
inputPrice: 0,
|
|
1854
1684
|
outputPrice: 0,
|
|
1855
1685
|
description: "Microsoft MAI-DS-R1 FP8 model."
|
|
@@ -1859,8 +1689,6 @@ var chutesModels = {
|
|
|
1859
1689
|
contextWindow: 163840,
|
|
1860
1690
|
supportsImages: false,
|
|
1861
1691
|
supportsPromptCache: false,
|
|
1862
|
-
supportsNativeTools: true,
|
|
1863
|
-
defaultToolProtocol: "native",
|
|
1864
1692
|
inputPrice: 0,
|
|
1865
1693
|
outputPrice: 0,
|
|
1866
1694
|
description: "TNGTech DeepSeek R1T Chimera model."
|
|
@@ -1870,8 +1698,6 @@ var chutesModels = {
|
|
|
1870
1698
|
contextWindow: 151329,
|
|
1871
1699
|
supportsImages: false,
|
|
1872
1700
|
supportsPromptCache: false,
|
|
1873
|
-
supportsNativeTools: true,
|
|
1874
|
-
defaultToolProtocol: "native",
|
|
1875
1701
|
inputPrice: 0,
|
|
1876
1702
|
outputPrice: 0,
|
|
1877
1703
|
description: "GLM-4.5-Air model with 151,329 token context window and 106B total parameters with 12B activated."
|
|
@@ -1881,8 +1707,6 @@ var chutesModels = {
|
|
|
1881
1707
|
contextWindow: 131072,
|
|
1882
1708
|
supportsImages: false,
|
|
1883
1709
|
supportsPromptCache: false,
|
|
1884
|
-
supportsNativeTools: true,
|
|
1885
|
-
defaultToolProtocol: "native",
|
|
1886
1710
|
inputPrice: 0,
|
|
1887
1711
|
outputPrice: 0,
|
|
1888
1712
|
description: "GLM-4.5-FP8 model with 128k token context window, optimized for agent-based applications with MoE architecture."
|
|
@@ -1892,8 +1716,6 @@ var chutesModels = {
|
|
|
1892
1716
|
contextWindow: 131072,
|
|
1893
1717
|
supportsImages: false,
|
|
1894
1718
|
supportsPromptCache: false,
|
|
1895
|
-
supportsNativeTools: true,
|
|
1896
|
-
defaultToolProtocol: "native",
|
|
1897
1719
|
inputPrice: 1,
|
|
1898
1720
|
outputPrice: 3,
|
|
1899
1721
|
description: "GLM-4.5-turbo model with 128K token context window, optimized for fast inference."
|
|
@@ -1903,8 +1725,6 @@ var chutesModels = {
|
|
|
1903
1725
|
contextWindow: 202752,
|
|
1904
1726
|
supportsImages: false,
|
|
1905
1727
|
supportsPromptCache: false,
|
|
1906
|
-
supportsNativeTools: true,
|
|
1907
|
-
defaultToolProtocol: "native",
|
|
1908
1728
|
inputPrice: 0,
|
|
1909
1729
|
outputPrice: 0,
|
|
1910
1730
|
description: "GLM-4.6 introduces major upgrades over GLM-4.5, including a longer 200K-token context window for complex tasks, stronger coding performance in benchmarks and real-world tools (such as Claude Code, Cline, Roo Code, and Kilo Code), improved reasoning with tool use during inference, more capable and efficient agent integration, and refined writing that better matches human style, readability, and natural role-play scenarios."
|
|
@@ -1915,8 +1735,6 @@ var chutesModels = {
|
|
|
1915
1735
|
contextWindow: 202752,
|
|
1916
1736
|
supportsImages: false,
|
|
1917
1737
|
supportsPromptCache: false,
|
|
1918
|
-
supportsNativeTools: true,
|
|
1919
|
-
defaultToolProtocol: "native",
|
|
1920
1738
|
inputPrice: 1.15,
|
|
1921
1739
|
outputPrice: 3.25,
|
|
1922
1740
|
description: "GLM-4.6-turbo model with 200K-token context window, optimized for fast inference."
|
|
@@ -1926,8 +1744,6 @@ var chutesModels = {
|
|
|
1926
1744
|
contextWindow: 128e3,
|
|
1927
1745
|
supportsImages: false,
|
|
1928
1746
|
supportsPromptCache: false,
|
|
1929
|
-
supportsNativeTools: true,
|
|
1930
|
-
defaultToolProtocol: "native",
|
|
1931
1747
|
inputPrice: 0,
|
|
1932
1748
|
outputPrice: 0,
|
|
1933
1749
|
description: "LongCat Flash Thinking FP8 model with 128K context window, optimized for complex reasoning and coding tasks."
|
|
@@ -1937,8 +1753,6 @@ var chutesModels = {
|
|
|
1937
1753
|
contextWindow: 262144,
|
|
1938
1754
|
supportsImages: false,
|
|
1939
1755
|
supportsPromptCache: false,
|
|
1940
|
-
supportsNativeTools: true,
|
|
1941
|
-
defaultToolProtocol: "native",
|
|
1942
1756
|
inputPrice: 0,
|
|
1943
1757
|
outputPrice: 0,
|
|
1944
1758
|
description: "Qwen3 Coder 480B A35B Instruct FP8 model, optimized for coding tasks."
|
|
@@ -1948,8 +1762,6 @@ var chutesModels = {
|
|
|
1948
1762
|
contextWindow: 75e3,
|
|
1949
1763
|
supportsImages: false,
|
|
1950
1764
|
supportsPromptCache: false,
|
|
1951
|
-
supportsNativeTools: true,
|
|
1952
|
-
defaultToolProtocol: "native",
|
|
1953
1765
|
inputPrice: 0.1481,
|
|
1954
1766
|
outputPrice: 0.5926,
|
|
1955
1767
|
description: "Moonshot AI Kimi K2 Instruct model with 75k context window."
|
|
@@ -1959,8 +1771,6 @@ var chutesModels = {
|
|
|
1959
1771
|
contextWindow: 262144,
|
|
1960
1772
|
supportsImages: false,
|
|
1961
1773
|
supportsPromptCache: false,
|
|
1962
|
-
supportsNativeTools: true,
|
|
1963
|
-
defaultToolProtocol: "native",
|
|
1964
1774
|
inputPrice: 0.1999,
|
|
1965
1775
|
outputPrice: 0.8001,
|
|
1966
1776
|
description: "Moonshot AI Kimi K2 Instruct 0905 model with 256k context window."
|
|
@@ -1970,8 +1780,6 @@ var chutesModels = {
|
|
|
1970
1780
|
contextWindow: 262144,
|
|
1971
1781
|
supportsImages: false,
|
|
1972
1782
|
supportsPromptCache: false,
|
|
1973
|
-
supportsNativeTools: true,
|
|
1974
|
-
defaultToolProtocol: "native",
|
|
1975
1783
|
inputPrice: 0.077968332,
|
|
1976
1784
|
outputPrice: 0.31202496,
|
|
1977
1785
|
description: "Qwen3 235B A22B Thinking 2507 model with 262K context window."
|
|
@@ -1981,8 +1789,6 @@ var chutesModels = {
|
|
|
1981
1789
|
contextWindow: 131072,
|
|
1982
1790
|
supportsImages: false,
|
|
1983
1791
|
supportsPromptCache: false,
|
|
1984
|
-
supportsNativeTools: true,
|
|
1985
|
-
defaultToolProtocol: "native",
|
|
1986
1792
|
inputPrice: 0,
|
|
1987
1793
|
outputPrice: 0,
|
|
1988
1794
|
description: "Fast, stable instruction-tuned model optimized for complex tasks, RAG, and tool use without thinking traces."
|
|
@@ -1992,8 +1798,6 @@ var chutesModels = {
|
|
|
1992
1798
|
contextWindow: 131072,
|
|
1993
1799
|
supportsImages: false,
|
|
1994
1800
|
supportsPromptCache: false,
|
|
1995
|
-
supportsNativeTools: true,
|
|
1996
|
-
defaultToolProtocol: "native",
|
|
1997
1801
|
inputPrice: 0,
|
|
1998
1802
|
outputPrice: 0,
|
|
1999
1803
|
description: "Reasoning-first model with structured thinking traces for multi-step problems, math proofs, and code synthesis."
|
|
@@ -2003,8 +1807,6 @@ var chutesModels = {
|
|
|
2003
1807
|
contextWindow: 262144,
|
|
2004
1808
|
supportsImages: true,
|
|
2005
1809
|
supportsPromptCache: false,
|
|
2006
|
-
supportsNativeTools: true,
|
|
2007
|
-
defaultToolProtocol: "native",
|
|
2008
1810
|
inputPrice: 0.16,
|
|
2009
1811
|
outputPrice: 0.65,
|
|
2010
1812
|
description: "Qwen3\u2011VL\u2011235B\u2011A22B\u2011Thinking is an open\u2011weight MoE vision\u2011language model (235B total, ~22B activated) optimized for deliberate multi\u2011step reasoning with strong text\u2011image\u2011video understanding and long\u2011context capabilities."
|
|
@@ -2012,73 +1814,6 @@ var chutesModels = {
|
|
|
2012
1814
|
};
|
|
2013
1815
|
var chutesDefaultModelInfo = chutesModels[chutesDefaultModelId];
|
|
2014
1816
|
|
|
2015
|
-
// src/providers/claude-code.ts
|
|
2016
|
-
var DATE_SUFFIX_PATTERN = /-\d{8}$/;
|
|
2017
|
-
var claudeCodeModels = {
|
|
2018
|
-
"claude-haiku-4-5": {
|
|
2019
|
-
maxTokens: 32768,
|
|
2020
|
-
contextWindow: 2e5,
|
|
2021
|
-
supportsImages: true,
|
|
2022
|
-
supportsPromptCache: true,
|
|
2023
|
-
supportsNativeTools: true,
|
|
2024
|
-
defaultToolProtocol: "native",
|
|
2025
|
-
supportsReasoningEffort: ["disable", "low", "medium", "high"],
|
|
2026
|
-
reasoningEffort: "medium",
|
|
2027
|
-
description: "Claude Haiku 4.5 - Fast and efficient with thinking"
|
|
2028
|
-
},
|
|
2029
|
-
"claude-sonnet-4-5": {
|
|
2030
|
-
maxTokens: 32768,
|
|
2031
|
-
contextWindow: 2e5,
|
|
2032
|
-
supportsImages: true,
|
|
2033
|
-
supportsPromptCache: true,
|
|
2034
|
-
supportsNativeTools: true,
|
|
2035
|
-
defaultToolProtocol: "native",
|
|
2036
|
-
supportsReasoningEffort: ["disable", "low", "medium", "high"],
|
|
2037
|
-
reasoningEffort: "medium",
|
|
2038
|
-
description: "Claude Sonnet 4.5 - Balanced performance with thinking"
|
|
2039
|
-
},
|
|
2040
|
-
"claude-opus-4-5": {
|
|
2041
|
-
maxTokens: 32768,
|
|
2042
|
-
contextWindow: 2e5,
|
|
2043
|
-
supportsImages: true,
|
|
2044
|
-
supportsPromptCache: true,
|
|
2045
|
-
supportsNativeTools: true,
|
|
2046
|
-
defaultToolProtocol: "native",
|
|
2047
|
-
supportsReasoningEffort: ["disable", "low", "medium", "high"],
|
|
2048
|
-
reasoningEffort: "medium",
|
|
2049
|
-
description: "Claude Opus 4.5 - Most capable with thinking"
|
|
2050
|
-
}
|
|
2051
|
-
};
|
|
2052
|
-
var claudeCodeDefaultModelId = "claude-sonnet-4-5";
|
|
2053
|
-
var MODEL_FAMILY_PATTERNS = [
|
|
2054
|
-
// Opus models (any version) → claude-opus-4-5
|
|
2055
|
-
{ pattern: /opus/i, target: "claude-opus-4-5" },
|
|
2056
|
-
// Haiku models (any version) → claude-haiku-4-5
|
|
2057
|
-
{ pattern: /haiku/i, target: "claude-haiku-4-5" },
|
|
2058
|
-
// Sonnet models (any version) → claude-sonnet-4-5
|
|
2059
|
-
{ pattern: /sonnet/i, target: "claude-sonnet-4-5" }
|
|
2060
|
-
];
|
|
2061
|
-
function normalizeClaudeCodeModelId(modelId) {
|
|
2062
|
-
if (Object.hasOwn(claudeCodeModels, modelId)) {
|
|
2063
|
-
return modelId;
|
|
2064
|
-
}
|
|
2065
|
-
const withoutDate = modelId.replace(DATE_SUFFIX_PATTERN, "");
|
|
2066
|
-
if (Object.hasOwn(claudeCodeModels, withoutDate)) {
|
|
2067
|
-
return withoutDate;
|
|
2068
|
-
}
|
|
2069
|
-
for (const { pattern, target } of MODEL_FAMILY_PATTERNS) {
|
|
2070
|
-
if (pattern.test(modelId)) {
|
|
2071
|
-
return target;
|
|
2072
|
-
}
|
|
2073
|
-
}
|
|
2074
|
-
return claudeCodeDefaultModelId;
|
|
2075
|
-
}
|
|
2076
|
-
var claudeCodeReasoningConfig = {
|
|
2077
|
-
low: { budgetTokens: 16e3 },
|
|
2078
|
-
medium: { budgetTokens: 32e3 },
|
|
2079
|
-
high: { budgetTokens: 64e3 }
|
|
2080
|
-
};
|
|
2081
|
-
|
|
2082
1817
|
// src/providers/deepseek.ts
|
|
2083
1818
|
var deepSeekDefaultModelId = "deepseek-chat";
|
|
2084
1819
|
var deepSeekModels = {
|
|
@@ -2088,8 +1823,6 @@ var deepSeekModels = {
|
|
|
2088
1823
|
contextWindow: 128e3,
|
|
2089
1824
|
supportsImages: false,
|
|
2090
1825
|
supportsPromptCache: true,
|
|
2091
|
-
supportsNativeTools: true,
|
|
2092
|
-
defaultToolProtocol: "native",
|
|
2093
1826
|
inputPrice: 0.28,
|
|
2094
1827
|
// $0.28 per million tokens (cache miss) - Updated Dec 9, 2025
|
|
2095
1828
|
outputPrice: 0.42,
|
|
@@ -2106,8 +1839,6 @@ var deepSeekModels = {
|
|
|
2106
1839
|
contextWindow: 128e3,
|
|
2107
1840
|
supportsImages: false,
|
|
2108
1841
|
supportsPromptCache: true,
|
|
2109
|
-
supportsNativeTools: true,
|
|
2110
|
-
defaultToolProtocol: "native",
|
|
2111
1842
|
preserveReasoning: true,
|
|
2112
1843
|
inputPrice: 0.28,
|
|
2113
1844
|
// $0.28 per million tokens (cache miss) - Updated Dec 9, 2025
|
|
@@ -2130,8 +1861,6 @@ var doubaoModels = {
|
|
|
2130
1861
|
contextWindow: 128e3,
|
|
2131
1862
|
supportsImages: true,
|
|
2132
1863
|
supportsPromptCache: true,
|
|
2133
|
-
supportsNativeTools: true,
|
|
2134
|
-
defaultToolProtocol: "native",
|
|
2135
1864
|
inputPrice: 1e-4,
|
|
2136
1865
|
// $0.0001 per million tokens (cache miss)
|
|
2137
1866
|
outputPrice: 4e-4,
|
|
@@ -2147,8 +1876,6 @@ var doubaoModels = {
|
|
|
2147
1876
|
contextWindow: 128e3,
|
|
2148
1877
|
supportsImages: true,
|
|
2149
1878
|
supportsPromptCache: true,
|
|
2150
|
-
supportsNativeTools: true,
|
|
2151
|
-
defaultToolProtocol: "native",
|
|
2152
1879
|
inputPrice: 2e-4,
|
|
2153
1880
|
// $0.0002 per million tokens
|
|
2154
1881
|
outputPrice: 8e-4,
|
|
@@ -2164,8 +1891,6 @@ var doubaoModels = {
|
|
|
2164
1891
|
contextWindow: 128e3,
|
|
2165
1892
|
supportsImages: true,
|
|
2166
1893
|
supportsPromptCache: true,
|
|
2167
|
-
supportsNativeTools: true,
|
|
2168
|
-
defaultToolProtocol: "native",
|
|
2169
1894
|
inputPrice: 15e-5,
|
|
2170
1895
|
// $0.00015 per million tokens
|
|
2171
1896
|
outputPrice: 6e-4,
|
|
@@ -2188,7 +1913,6 @@ var featherlessModels = {
|
|
|
2188
1913
|
contextWindow: 32678,
|
|
2189
1914
|
supportsImages: false,
|
|
2190
1915
|
supportsPromptCache: false,
|
|
2191
|
-
supportsNativeTools: true,
|
|
2192
1916
|
inputPrice: 0,
|
|
2193
1917
|
outputPrice: 0,
|
|
2194
1918
|
description: "DeepSeek V3 0324 model."
|
|
@@ -2198,7 +1922,6 @@ var featherlessModels = {
|
|
|
2198
1922
|
contextWindow: 32678,
|
|
2199
1923
|
supportsImages: false,
|
|
2200
1924
|
supportsPromptCache: false,
|
|
2201
|
-
supportsNativeTools: true,
|
|
2202
1925
|
inputPrice: 0,
|
|
2203
1926
|
outputPrice: 0,
|
|
2204
1927
|
description: "DeepSeek R1 0528 model."
|
|
@@ -2208,7 +1931,6 @@ var featherlessModels = {
|
|
|
2208
1931
|
contextWindow: 32678,
|
|
2209
1932
|
supportsImages: false,
|
|
2210
1933
|
supportsPromptCache: false,
|
|
2211
|
-
supportsNativeTools: true,
|
|
2212
1934
|
inputPrice: 0,
|
|
2213
1935
|
outputPrice: 0,
|
|
2214
1936
|
description: "Kimi K2 Instruct model."
|
|
@@ -2218,7 +1940,6 @@ var featherlessModels = {
|
|
|
2218
1940
|
contextWindow: 32678,
|
|
2219
1941
|
supportsImages: false,
|
|
2220
1942
|
supportsPromptCache: false,
|
|
2221
|
-
supportsNativeTools: true,
|
|
2222
1943
|
inputPrice: 0,
|
|
2223
1944
|
outputPrice: 0,
|
|
2224
1945
|
description: "GPT-OSS 120B model."
|
|
@@ -2228,7 +1949,6 @@ var featherlessModels = {
|
|
|
2228
1949
|
contextWindow: 32678,
|
|
2229
1950
|
supportsImages: false,
|
|
2230
1951
|
supportsPromptCache: false,
|
|
2231
|
-
supportsNativeTools: true,
|
|
2232
1952
|
inputPrice: 0,
|
|
2233
1953
|
outputPrice: 0,
|
|
2234
1954
|
description: "Qwen3 Coder 480B A35B Instruct model."
|
|
@@ -2244,8 +1964,6 @@ var fireworksModels = {
|
|
|
2244
1964
|
contextWindow: 262144,
|
|
2245
1965
|
supportsImages: false,
|
|
2246
1966
|
supportsPromptCache: true,
|
|
2247
|
-
supportsNativeTools: true,
|
|
2248
|
-
defaultToolProtocol: "native",
|
|
2249
1967
|
inputPrice: 0.6,
|
|
2250
1968
|
outputPrice: 2.5,
|
|
2251
1969
|
cacheReadsPrice: 0.15,
|
|
@@ -2256,8 +1974,6 @@ var fireworksModels = {
|
|
|
2256
1974
|
contextWindow: 128e3,
|
|
2257
1975
|
supportsImages: false,
|
|
2258
1976
|
supportsPromptCache: false,
|
|
2259
|
-
supportsNativeTools: true,
|
|
2260
|
-
defaultToolProtocol: "native",
|
|
2261
1977
|
inputPrice: 0.6,
|
|
2262
1978
|
outputPrice: 2.5,
|
|
2263
1979
|
description: "Kimi K2 is a state-of-the-art mixture-of-experts (MoE) language model with 32 billion activated parameters and 1 trillion total parameters. Trained with the Muon optimizer, Kimi K2 achieves exceptional performance across frontier knowledge, reasoning, and coding tasks while being meticulously optimized for agentic capabilities."
|
|
@@ -2267,7 +1983,6 @@ var fireworksModels = {
|
|
|
2267
1983
|
contextWindow: 256e3,
|
|
2268
1984
|
supportsImages: false,
|
|
2269
1985
|
supportsPromptCache: true,
|
|
2270
|
-
supportsNativeTools: true,
|
|
2271
1986
|
supportsTemperature: true,
|
|
2272
1987
|
preserveReasoning: true,
|
|
2273
1988
|
defaultTemperature: 1,
|
|
@@ -2281,8 +1996,6 @@ var fireworksModels = {
|
|
|
2281
1996
|
contextWindow: 204800,
|
|
2282
1997
|
supportsImages: false,
|
|
2283
1998
|
supportsPromptCache: false,
|
|
2284
|
-
supportsNativeTools: true,
|
|
2285
|
-
defaultToolProtocol: "native",
|
|
2286
1999
|
inputPrice: 0.3,
|
|
2287
2000
|
outputPrice: 1.2,
|
|
2288
2001
|
description: "MiniMax M2 is a high-performance language model with 204.8K context window, optimized for long-context understanding and generation tasks."
|
|
@@ -2292,8 +2005,6 @@ var fireworksModels = {
|
|
|
2292
2005
|
contextWindow: 256e3,
|
|
2293
2006
|
supportsImages: false,
|
|
2294
2007
|
supportsPromptCache: false,
|
|
2295
|
-
supportsNativeTools: true,
|
|
2296
|
-
defaultToolProtocol: "native",
|
|
2297
2008
|
inputPrice: 0.22,
|
|
2298
2009
|
outputPrice: 0.88,
|
|
2299
2010
|
description: "Latest Qwen3 thinking model, competitive against the best closed source models in Jul 2025."
|
|
@@ -2303,8 +2014,6 @@ var fireworksModels = {
|
|
|
2303
2014
|
contextWindow: 256e3,
|
|
2304
2015
|
supportsImages: false,
|
|
2305
2016
|
supportsPromptCache: false,
|
|
2306
|
-
supportsNativeTools: true,
|
|
2307
|
-
defaultToolProtocol: "native",
|
|
2308
2017
|
inputPrice: 0.45,
|
|
2309
2018
|
outputPrice: 1.8,
|
|
2310
2019
|
description: "Qwen3's most agentic code model to date."
|
|
@@ -2314,8 +2023,6 @@ var fireworksModels = {
|
|
|
2314
2023
|
contextWindow: 16e4,
|
|
2315
2024
|
supportsImages: false,
|
|
2316
2025
|
supportsPromptCache: false,
|
|
2317
|
-
supportsNativeTools: true,
|
|
2318
|
-
defaultToolProtocol: "native",
|
|
2319
2026
|
inputPrice: 3,
|
|
2320
2027
|
outputPrice: 8,
|
|
2321
2028
|
description: "05/28 updated checkpoint of Deepseek R1. Its overall performance is now approaching that of leading models, such as O3 and Gemini 2.5 Pro. Compared to the previous version, the upgraded model shows significant improvements in handling complex reasoning tasks, and this version also offers a reduced hallucination rate, enhanced support for function calling, and better experience for vibe coding. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
|
|
@@ -2325,8 +2032,6 @@ var fireworksModels = {
|
|
|
2325
2032
|
contextWindow: 128e3,
|
|
2326
2033
|
supportsImages: false,
|
|
2327
2034
|
supportsPromptCache: false,
|
|
2328
|
-
supportsNativeTools: true,
|
|
2329
|
-
defaultToolProtocol: "native",
|
|
2330
2035
|
inputPrice: 0.9,
|
|
2331
2036
|
outputPrice: 0.9,
|
|
2332
2037
|
description: "A strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token from Deepseek. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
|
|
@@ -2336,8 +2041,6 @@ var fireworksModels = {
|
|
|
2336
2041
|
contextWindow: 163840,
|
|
2337
2042
|
supportsImages: false,
|
|
2338
2043
|
supportsPromptCache: false,
|
|
2339
|
-
supportsNativeTools: true,
|
|
2340
|
-
defaultToolProtocol: "native",
|
|
2341
2044
|
inputPrice: 0.56,
|
|
2342
2045
|
outputPrice: 1.68,
|
|
2343
2046
|
description: "DeepSeek v3.1 is an improved version of the v3 model with enhanced performance, better reasoning capabilities, and improved code generation. This Mixture-of-Experts (MoE) model maintains the same 671B total parameters with 37B activated per token."
|
|
@@ -2347,8 +2050,6 @@ var fireworksModels = {
|
|
|
2347
2050
|
contextWindow: 128e3,
|
|
2348
2051
|
supportsImages: false,
|
|
2349
2052
|
supportsPromptCache: false,
|
|
2350
|
-
supportsNativeTools: true,
|
|
2351
|
-
defaultToolProtocol: "native",
|
|
2352
2053
|
inputPrice: 0.55,
|
|
2353
2054
|
outputPrice: 2.19,
|
|
2354
2055
|
description: "Z.ai GLM-4.5 with 355B total parameters and 32B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
|
|
@@ -2358,8 +2059,6 @@ var fireworksModels = {
|
|
|
2358
2059
|
contextWindow: 128e3,
|
|
2359
2060
|
supportsImages: false,
|
|
2360
2061
|
supportsPromptCache: false,
|
|
2361
|
-
supportsNativeTools: true,
|
|
2362
|
-
defaultToolProtocol: "native",
|
|
2363
2062
|
inputPrice: 0.55,
|
|
2364
2063
|
outputPrice: 2.19,
|
|
2365
2064
|
description: "Z.ai GLM-4.5-Air with 106B total parameters and 12B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
|
|
@@ -2369,8 +2068,6 @@ var fireworksModels = {
|
|
|
2369
2068
|
contextWindow: 198e3,
|
|
2370
2069
|
supportsImages: false,
|
|
2371
2070
|
supportsPromptCache: false,
|
|
2372
|
-
supportsNativeTools: true,
|
|
2373
|
-
defaultToolProtocol: "native",
|
|
2374
2071
|
inputPrice: 0.55,
|
|
2375
2072
|
outputPrice: 2.19,
|
|
2376
2073
|
description: "Z.ai GLM-4.6 is an advanced coding model with exceptional performance on complex programming tasks. Features improved reasoning capabilities and enhanced code generation quality, making it ideal for software development workflows."
|
|
@@ -2380,8 +2077,6 @@ var fireworksModels = {
|
|
|
2380
2077
|
contextWindow: 128e3,
|
|
2381
2078
|
supportsImages: false,
|
|
2382
2079
|
supportsPromptCache: false,
|
|
2383
|
-
supportsNativeTools: true,
|
|
2384
|
-
defaultToolProtocol: "native",
|
|
2385
2080
|
inputPrice: 0.07,
|
|
2386
2081
|
outputPrice: 0.3,
|
|
2387
2082
|
description: "OpenAI gpt-oss-20b: Compact model for local/edge deployments. Optimized for low-latency and resource-constrained environments with chain-of-thought output, adjustable reasoning, and agentic workflows."
|
|
@@ -2391,11 +2086,63 @@ var fireworksModels = {
|
|
|
2391
2086
|
contextWindow: 128e3,
|
|
2392
2087
|
supportsImages: false,
|
|
2393
2088
|
supportsPromptCache: false,
|
|
2394
|
-
supportsNativeTools: true,
|
|
2395
|
-
defaultToolProtocol: "native",
|
|
2396
2089
|
inputPrice: 0.15,
|
|
2397
2090
|
outputPrice: 0.6,
|
|
2398
2091
|
description: "OpenAI gpt-oss-120b: Production-grade, general-purpose model that fits on a single H100 GPU. Features complex reasoning, configurable effort, full chain-of-thought transparency, and supports function calling, tool use, and structured outputs."
|
|
2092
|
+
},
|
|
2093
|
+
"accounts/fireworks/models/minimax-m2p1": {
|
|
2094
|
+
maxTokens: 4096,
|
|
2095
|
+
contextWindow: 204800,
|
|
2096
|
+
supportsImages: false,
|
|
2097
|
+
supportsPromptCache: false,
|
|
2098
|
+
inputPrice: 0.3,
|
|
2099
|
+
outputPrice: 1.2,
|
|
2100
|
+
description: "MiniMax M2.1 is an upgraded version of M2 with improved performance on complex reasoning, coding, and long-context understanding tasks."
|
|
2101
|
+
},
|
|
2102
|
+
"accounts/fireworks/models/deepseek-v3p2": {
|
|
2103
|
+
maxTokens: 16384,
|
|
2104
|
+
contextWindow: 163840,
|
|
2105
|
+
supportsImages: false,
|
|
2106
|
+
supportsPromptCache: false,
|
|
2107
|
+
inputPrice: 0.56,
|
|
2108
|
+
outputPrice: 1.68,
|
|
2109
|
+
description: "DeepSeek V3.2 is the latest iteration of the V3 model family with enhanced reasoning capabilities, improved code generation, and better instruction following."
|
|
2110
|
+
},
|
|
2111
|
+
"accounts/fireworks/models/glm-4p7": {
|
|
2112
|
+
maxTokens: 25344,
|
|
2113
|
+
contextWindow: 198e3,
|
|
2114
|
+
supportsImages: false,
|
|
2115
|
+
supportsPromptCache: false,
|
|
2116
|
+
inputPrice: 0.55,
|
|
2117
|
+
outputPrice: 2.19,
|
|
2118
|
+
description: "Z.ai GLM-4.7 is the latest coding model with exceptional performance on complex programming tasks. Features improved reasoning capabilities and enhanced code generation quality."
|
|
2119
|
+
},
|
|
2120
|
+
"accounts/fireworks/models/llama-v3p3-70b-instruct": {
|
|
2121
|
+
maxTokens: 16384,
|
|
2122
|
+
contextWindow: 131072,
|
|
2123
|
+
supportsImages: false,
|
|
2124
|
+
supportsPromptCache: false,
|
|
2125
|
+
inputPrice: 0.9,
|
|
2126
|
+
outputPrice: 0.9,
|
|
2127
|
+
description: "Meta Llama 3.3 70B Instruct is a highly capable instruction-tuned model with strong reasoning, coding, and general task performance."
|
|
2128
|
+
},
|
|
2129
|
+
"accounts/fireworks/models/llama4-maverick-instruct-basic": {
|
|
2130
|
+
maxTokens: 16384,
|
|
2131
|
+
contextWindow: 131072,
|
|
2132
|
+
supportsImages: true,
|
|
2133
|
+
supportsPromptCache: false,
|
|
2134
|
+
inputPrice: 0.22,
|
|
2135
|
+
outputPrice: 0.88,
|
|
2136
|
+
description: "Llama 4 Maverick is Meta's latest multimodal model with vision capabilities, optimized for instruction following and coding tasks."
|
|
2137
|
+
},
|
|
2138
|
+
"accounts/fireworks/models/llama4-scout-instruct-basic": {
|
|
2139
|
+
maxTokens: 16384,
|
|
2140
|
+
contextWindow: 131072,
|
|
2141
|
+
supportsImages: true,
|
|
2142
|
+
supportsPromptCache: false,
|
|
2143
|
+
inputPrice: 0.15,
|
|
2144
|
+
outputPrice: 0.6,
|
|
2145
|
+
description: "Llama 4 Scout is a smaller, faster variant of Llama 4 with multimodal capabilities, ideal for quick iterations and cost-effective deployments."
|
|
2399
2146
|
}
|
|
2400
2147
|
};
|
|
2401
2148
|
|
|
@@ -2406,8 +2153,6 @@ var geminiModels = {
|
|
|
2406
2153
|
maxTokens: 65536,
|
|
2407
2154
|
contextWindow: 1048576,
|
|
2408
2155
|
supportsImages: true,
|
|
2409
|
-
supportsNativeTools: true,
|
|
2410
|
-
defaultToolProtocol: "native",
|
|
2411
2156
|
supportsPromptCache: true,
|
|
2412
2157
|
supportsReasoningEffort: ["low", "high"],
|
|
2413
2158
|
reasoningEffort: "low",
|
|
@@ -2415,16 +2160,19 @@ var geminiModels = {
|
|
|
2415
2160
|
defaultTemperature: 1,
|
|
2416
2161
|
inputPrice: 4,
|
|
2417
2162
|
outputPrice: 18,
|
|
2163
|
+
cacheReadsPrice: 0.4,
|
|
2418
2164
|
tiers: [
|
|
2419
2165
|
{
|
|
2420
2166
|
contextWindow: 2e5,
|
|
2421
2167
|
inputPrice: 2,
|
|
2422
|
-
outputPrice: 12
|
|
2168
|
+
outputPrice: 12,
|
|
2169
|
+
cacheReadsPrice: 0.2
|
|
2423
2170
|
},
|
|
2424
2171
|
{
|
|
2425
2172
|
contextWindow: Infinity,
|
|
2426
2173
|
inputPrice: 4,
|
|
2427
|
-
outputPrice: 18
|
|
2174
|
+
outputPrice: 18,
|
|
2175
|
+
cacheReadsPrice: 0.4
|
|
2428
2176
|
}
|
|
2429
2177
|
]
|
|
2430
2178
|
},
|
|
@@ -2432,25 +2180,20 @@ var geminiModels = {
|
|
|
2432
2180
|
maxTokens: 65536,
|
|
2433
2181
|
contextWindow: 1048576,
|
|
2434
2182
|
supportsImages: true,
|
|
2435
|
-
supportsNativeTools: true,
|
|
2436
|
-
defaultToolProtocol: "native",
|
|
2437
2183
|
supportsPromptCache: true,
|
|
2438
2184
|
supportsReasoningEffort: ["minimal", "low", "medium", "high"],
|
|
2439
2185
|
reasoningEffort: "medium",
|
|
2440
2186
|
supportsTemperature: true,
|
|
2441
2187
|
defaultTemperature: 1,
|
|
2442
|
-
inputPrice: 0.
|
|
2443
|
-
outputPrice:
|
|
2444
|
-
cacheReadsPrice: 0.
|
|
2445
|
-
cacheWritesPrice: 1
|
|
2188
|
+
inputPrice: 0.5,
|
|
2189
|
+
outputPrice: 3,
|
|
2190
|
+
cacheReadsPrice: 0.05
|
|
2446
2191
|
},
|
|
2447
2192
|
// 2.5 Pro models
|
|
2448
2193
|
"gemini-2.5-pro": {
|
|
2449
2194
|
maxTokens: 64e3,
|
|
2450
2195
|
contextWindow: 1048576,
|
|
2451
2196
|
supportsImages: true,
|
|
2452
|
-
supportsNativeTools: true,
|
|
2453
|
-
defaultToolProtocol: "native",
|
|
2454
2197
|
supportsPromptCache: true,
|
|
2455
2198
|
inputPrice: 2.5,
|
|
2456
2199
|
// This is the pricing for prompts above 200k tokens.
|
|
@@ -2479,8 +2222,6 @@ var geminiModels = {
|
|
|
2479
2222
|
maxTokens: 65535,
|
|
2480
2223
|
contextWindow: 1048576,
|
|
2481
2224
|
supportsImages: true,
|
|
2482
|
-
supportsNativeTools: true,
|
|
2483
|
-
defaultToolProtocol: "native",
|
|
2484
2225
|
supportsPromptCache: true,
|
|
2485
2226
|
inputPrice: 2.5,
|
|
2486
2227
|
// This is the pricing for prompts above 200k tokens.
|
|
@@ -2508,8 +2249,6 @@ var geminiModels = {
|
|
|
2508
2249
|
maxTokens: 65535,
|
|
2509
2250
|
contextWindow: 1048576,
|
|
2510
2251
|
supportsImages: true,
|
|
2511
|
-
supportsNativeTools: true,
|
|
2512
|
-
defaultToolProtocol: "native",
|
|
2513
2252
|
supportsPromptCache: true,
|
|
2514
2253
|
inputPrice: 2.5,
|
|
2515
2254
|
// This is the pricing for prompts above 200k tokens.
|
|
@@ -2535,8 +2274,6 @@ var geminiModels = {
|
|
|
2535
2274
|
maxTokens: 65535,
|
|
2536
2275
|
contextWindow: 1048576,
|
|
2537
2276
|
supportsImages: true,
|
|
2538
|
-
supportsNativeTools: true,
|
|
2539
|
-
defaultToolProtocol: "native",
|
|
2540
2277
|
supportsPromptCache: true,
|
|
2541
2278
|
inputPrice: 2.5,
|
|
2542
2279
|
// This is the pricing for prompts above 200k tokens.
|
|
@@ -2565,8 +2302,6 @@ var geminiModels = {
|
|
|
2565
2302
|
maxTokens: 65536,
|
|
2566
2303
|
contextWindow: 1048576,
|
|
2567
2304
|
supportsImages: true,
|
|
2568
|
-
supportsNativeTools: true,
|
|
2569
|
-
defaultToolProtocol: "native",
|
|
2570
2305
|
supportsPromptCache: true,
|
|
2571
2306
|
inputPrice: 0.3,
|
|
2572
2307
|
outputPrice: 2.5,
|
|
@@ -2579,8 +2314,6 @@ var geminiModels = {
|
|
|
2579
2314
|
maxTokens: 65536,
|
|
2580
2315
|
contextWindow: 1048576,
|
|
2581
2316
|
supportsImages: true,
|
|
2582
|
-
supportsNativeTools: true,
|
|
2583
|
-
defaultToolProtocol: "native",
|
|
2584
2317
|
supportsPromptCache: true,
|
|
2585
2318
|
inputPrice: 0.3,
|
|
2586
2319
|
outputPrice: 2.5,
|
|
@@ -2593,8 +2326,6 @@ var geminiModels = {
|
|
|
2593
2326
|
maxTokens: 64e3,
|
|
2594
2327
|
contextWindow: 1048576,
|
|
2595
2328
|
supportsImages: true,
|
|
2596
|
-
supportsNativeTools: true,
|
|
2597
|
-
defaultToolProtocol: "native",
|
|
2598
2329
|
supportsPromptCache: true,
|
|
2599
2330
|
inputPrice: 0.3,
|
|
2600
2331
|
outputPrice: 2.5,
|
|
@@ -2608,8 +2339,6 @@ var geminiModels = {
|
|
|
2608
2339
|
maxTokens: 65536,
|
|
2609
2340
|
contextWindow: 1048576,
|
|
2610
2341
|
supportsImages: true,
|
|
2611
|
-
supportsNativeTools: true,
|
|
2612
|
-
defaultToolProtocol: "native",
|
|
2613
2342
|
supportsPromptCache: true,
|
|
2614
2343
|
inputPrice: 0.1,
|
|
2615
2344
|
outputPrice: 0.4,
|
|
@@ -2622,8 +2351,6 @@ var geminiModels = {
|
|
|
2622
2351
|
maxTokens: 65536,
|
|
2623
2352
|
contextWindow: 1048576,
|
|
2624
2353
|
supportsImages: true,
|
|
2625
|
-
supportsNativeTools: true,
|
|
2626
|
-
defaultToolProtocol: "native",
|
|
2627
2354
|
supportsPromptCache: true,
|
|
2628
2355
|
inputPrice: 0.1,
|
|
2629
2356
|
outputPrice: 0.4,
|
|
@@ -2643,8 +2370,6 @@ var groqModels = {
|
|
|
2643
2370
|
contextWindow: 131072,
|
|
2644
2371
|
supportsImages: false,
|
|
2645
2372
|
supportsPromptCache: false,
|
|
2646
|
-
supportsNativeTools: true,
|
|
2647
|
-
defaultToolProtocol: "native",
|
|
2648
2373
|
inputPrice: 0.05,
|
|
2649
2374
|
outputPrice: 0.08,
|
|
2650
2375
|
description: "Meta Llama 3.1 8B Instant model, 128K context."
|
|
@@ -2654,8 +2379,6 @@ var groqModels = {
|
|
|
2654
2379
|
contextWindow: 131072,
|
|
2655
2380
|
supportsImages: false,
|
|
2656
2381
|
supportsPromptCache: false,
|
|
2657
|
-
supportsNativeTools: true,
|
|
2658
|
-
defaultToolProtocol: "native",
|
|
2659
2382
|
inputPrice: 0.59,
|
|
2660
2383
|
outputPrice: 0.79,
|
|
2661
2384
|
description: "Meta Llama 3.3 70B Versatile model, 128K context."
|
|
@@ -2665,8 +2388,6 @@ var groqModels = {
|
|
|
2665
2388
|
contextWindow: 131072,
|
|
2666
2389
|
supportsImages: false,
|
|
2667
2390
|
supportsPromptCache: false,
|
|
2668
|
-
supportsNativeTools: true,
|
|
2669
|
-
defaultToolProtocol: "native",
|
|
2670
2391
|
inputPrice: 0.11,
|
|
2671
2392
|
outputPrice: 0.34,
|
|
2672
2393
|
description: "Meta Llama 4 Scout 17B Instruct model, 128K context."
|
|
@@ -2676,8 +2397,6 @@ var groqModels = {
|
|
|
2676
2397
|
contextWindow: 131072,
|
|
2677
2398
|
supportsImages: false,
|
|
2678
2399
|
supportsPromptCache: false,
|
|
2679
|
-
supportsNativeTools: true,
|
|
2680
|
-
defaultToolProtocol: "native",
|
|
2681
2400
|
inputPrice: 0.29,
|
|
2682
2401
|
outputPrice: 0.59,
|
|
2683
2402
|
description: "Alibaba Qwen 3 32B model, 128K context."
|
|
@@ -2687,8 +2406,6 @@ var groqModels = {
|
|
|
2687
2406
|
contextWindow: 262144,
|
|
2688
2407
|
supportsImages: false,
|
|
2689
2408
|
supportsPromptCache: true,
|
|
2690
|
-
supportsNativeTools: true,
|
|
2691
|
-
defaultToolProtocol: "native",
|
|
2692
2409
|
inputPrice: 0.6,
|
|
2693
2410
|
outputPrice: 2.5,
|
|
2694
2411
|
cacheReadsPrice: 0.15,
|
|
@@ -2699,8 +2416,6 @@ var groqModels = {
|
|
|
2699
2416
|
contextWindow: 131072,
|
|
2700
2417
|
supportsImages: false,
|
|
2701
2418
|
supportsPromptCache: false,
|
|
2702
|
-
supportsNativeTools: true,
|
|
2703
|
-
defaultToolProtocol: "native",
|
|
2704
2419
|
inputPrice: 0.15,
|
|
2705
2420
|
outputPrice: 0.75,
|
|
2706
2421
|
description: "GPT-OSS 120B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 128 experts."
|
|
@@ -2710,8 +2425,6 @@ var groqModels = {
|
|
|
2710
2425
|
contextWindow: 131072,
|
|
2711
2426
|
supportsImages: false,
|
|
2712
2427
|
supportsPromptCache: false,
|
|
2713
|
-
supportsNativeTools: true,
|
|
2714
|
-
defaultToolProtocol: "native",
|
|
2715
2428
|
inputPrice: 0.1,
|
|
2716
2429
|
outputPrice: 0.5,
|
|
2717
2430
|
description: "GPT-OSS 20B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 32 experts."
|
|
@@ -2738,7 +2451,6 @@ var ioIntelligenceModels = {
|
|
|
2738
2451
|
contextWindow: 128e3,
|
|
2739
2452
|
supportsImages: false,
|
|
2740
2453
|
supportsPromptCache: false,
|
|
2741
|
-
supportsNativeTools: true,
|
|
2742
2454
|
description: "DeepSeek R1 reasoning model"
|
|
2743
2455
|
},
|
|
2744
2456
|
"meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8": {
|
|
@@ -2746,7 +2458,6 @@ var ioIntelligenceModels = {
|
|
|
2746
2458
|
contextWindow: 43e4,
|
|
2747
2459
|
supportsImages: true,
|
|
2748
2460
|
supportsPromptCache: false,
|
|
2749
|
-
supportsNativeTools: true,
|
|
2750
2461
|
description: "Llama 4 Maverick 17B model"
|
|
2751
2462
|
},
|
|
2752
2463
|
"Intel/Qwen3-Coder-480B-A35B-Instruct-int4-mixed-ar": {
|
|
@@ -2754,7 +2465,6 @@ var ioIntelligenceModels = {
|
|
|
2754
2465
|
contextWindow: 106e3,
|
|
2755
2466
|
supportsImages: false,
|
|
2756
2467
|
supportsPromptCache: false,
|
|
2757
|
-
supportsNativeTools: true,
|
|
2758
2468
|
description: "Qwen3 Coder 480B specialized for coding"
|
|
2759
2469
|
},
|
|
2760
2470
|
"openai/gpt-oss-120b": {
|
|
@@ -2762,7 +2472,6 @@ var ioIntelligenceModels = {
|
|
|
2762
2472
|
contextWindow: 131072,
|
|
2763
2473
|
supportsImages: false,
|
|
2764
2474
|
supportsPromptCache: false,
|
|
2765
|
-
supportsNativeTools: true,
|
|
2766
2475
|
description: "OpenAI GPT-OSS 120B model"
|
|
2767
2476
|
}
|
|
2768
2477
|
};
|
|
@@ -2774,8 +2483,6 @@ var litellmDefaultModelInfo = {
|
|
|
2774
2483
|
contextWindow: 2e5,
|
|
2775
2484
|
supportsImages: true,
|
|
2776
2485
|
supportsPromptCache: true,
|
|
2777
|
-
supportsNativeTools: true,
|
|
2778
|
-
defaultToolProtocol: "native",
|
|
2779
2486
|
inputPrice: 3,
|
|
2780
2487
|
outputPrice: 15,
|
|
2781
2488
|
cacheWritesPrice: 3.75,
|
|
@@ -2790,8 +2497,6 @@ var lMStudioDefaultModelInfo = {
|
|
|
2790
2497
|
contextWindow: 2e5,
|
|
2791
2498
|
supportsImages: true,
|
|
2792
2499
|
supportsPromptCache: true,
|
|
2793
|
-
supportsNativeTools: true,
|
|
2794
|
-
defaultToolProtocol: "native",
|
|
2795
2500
|
inputPrice: 0,
|
|
2796
2501
|
outputPrice: 0,
|
|
2797
2502
|
cacheWritesPrice: 0,
|
|
@@ -2807,8 +2512,6 @@ var mistralModels = {
|
|
|
2807
2512
|
contextWindow: 128e3,
|
|
2808
2513
|
supportsImages: true,
|
|
2809
2514
|
supportsPromptCache: false,
|
|
2810
|
-
supportsNativeTools: true,
|
|
2811
|
-
defaultToolProtocol: "native",
|
|
2812
2515
|
inputPrice: 2,
|
|
2813
2516
|
outputPrice: 5
|
|
2814
2517
|
},
|
|
@@ -2817,8 +2520,6 @@ var mistralModels = {
|
|
|
2817
2520
|
contextWindow: 131e3,
|
|
2818
2521
|
supportsImages: true,
|
|
2819
2522
|
supportsPromptCache: false,
|
|
2820
|
-
supportsNativeTools: true,
|
|
2821
|
-
defaultToolProtocol: "native",
|
|
2822
2523
|
inputPrice: 0.4,
|
|
2823
2524
|
outputPrice: 2
|
|
2824
2525
|
},
|
|
@@ -2827,8 +2528,6 @@ var mistralModels = {
|
|
|
2827
2528
|
contextWindow: 131e3,
|
|
2828
2529
|
supportsImages: true,
|
|
2829
2530
|
supportsPromptCache: false,
|
|
2830
|
-
supportsNativeTools: true,
|
|
2831
|
-
defaultToolProtocol: "native",
|
|
2832
2531
|
inputPrice: 0.4,
|
|
2833
2532
|
outputPrice: 2
|
|
2834
2533
|
},
|
|
@@ -2837,8 +2536,6 @@ var mistralModels = {
|
|
|
2837
2536
|
contextWindow: 256e3,
|
|
2838
2537
|
supportsImages: false,
|
|
2839
2538
|
supportsPromptCache: false,
|
|
2840
|
-
supportsNativeTools: true,
|
|
2841
|
-
defaultToolProtocol: "native",
|
|
2842
2539
|
inputPrice: 0.3,
|
|
2843
2540
|
outputPrice: 0.9
|
|
2844
2541
|
},
|
|
@@ -2847,8 +2544,6 @@ var mistralModels = {
|
|
|
2847
2544
|
contextWindow: 131e3,
|
|
2848
2545
|
supportsImages: false,
|
|
2849
2546
|
supportsPromptCache: false,
|
|
2850
|
-
supportsNativeTools: true,
|
|
2851
|
-
defaultToolProtocol: "native",
|
|
2852
2547
|
inputPrice: 2,
|
|
2853
2548
|
outputPrice: 6
|
|
2854
2549
|
},
|
|
@@ -2857,8 +2552,6 @@ var mistralModels = {
|
|
|
2857
2552
|
contextWindow: 131e3,
|
|
2858
2553
|
supportsImages: false,
|
|
2859
2554
|
supportsPromptCache: false,
|
|
2860
|
-
supportsNativeTools: true,
|
|
2861
|
-
defaultToolProtocol: "native",
|
|
2862
2555
|
inputPrice: 0.1,
|
|
2863
2556
|
outputPrice: 0.1
|
|
2864
2557
|
},
|
|
@@ -2867,8 +2560,6 @@ var mistralModels = {
|
|
|
2867
2560
|
contextWindow: 131e3,
|
|
2868
2561
|
supportsImages: false,
|
|
2869
2562
|
supportsPromptCache: false,
|
|
2870
|
-
supportsNativeTools: true,
|
|
2871
|
-
defaultToolProtocol: "native",
|
|
2872
2563
|
inputPrice: 0.04,
|
|
2873
2564
|
outputPrice: 0.04
|
|
2874
2565
|
},
|
|
@@ -2877,8 +2568,6 @@ var mistralModels = {
|
|
|
2877
2568
|
contextWindow: 32e3,
|
|
2878
2569
|
supportsImages: false,
|
|
2879
2570
|
supportsPromptCache: false,
|
|
2880
|
-
supportsNativeTools: true,
|
|
2881
|
-
defaultToolProtocol: "native",
|
|
2882
2571
|
inputPrice: 0.2,
|
|
2883
2572
|
outputPrice: 0.6
|
|
2884
2573
|
},
|
|
@@ -2887,8 +2576,6 @@ var mistralModels = {
|
|
|
2887
2576
|
contextWindow: 131e3,
|
|
2888
2577
|
supportsImages: true,
|
|
2889
2578
|
supportsPromptCache: false,
|
|
2890
|
-
supportsNativeTools: true,
|
|
2891
|
-
defaultToolProtocol: "native",
|
|
2892
2579
|
inputPrice: 2,
|
|
2893
2580
|
outputPrice: 6
|
|
2894
2581
|
}
|
|
@@ -2903,8 +2590,6 @@ var moonshotModels = {
|
|
|
2903
2590
|
contextWindow: 131072,
|
|
2904
2591
|
supportsImages: false,
|
|
2905
2592
|
supportsPromptCache: true,
|
|
2906
|
-
supportsNativeTools: true,
|
|
2907
|
-
defaultToolProtocol: "native",
|
|
2908
2593
|
inputPrice: 0.6,
|
|
2909
2594
|
// $0.60 per million tokens (cache miss)
|
|
2910
2595
|
outputPrice: 2.5,
|
|
@@ -2920,8 +2605,6 @@ var moonshotModels = {
|
|
|
2920
2605
|
contextWindow: 262144,
|
|
2921
2606
|
supportsImages: false,
|
|
2922
2607
|
supportsPromptCache: true,
|
|
2923
|
-
supportsNativeTools: true,
|
|
2924
|
-
defaultToolProtocol: "native",
|
|
2925
2608
|
inputPrice: 0.6,
|
|
2926
2609
|
outputPrice: 2.5,
|
|
2927
2610
|
cacheReadsPrice: 0.15,
|
|
@@ -2932,8 +2615,6 @@ var moonshotModels = {
|
|
|
2932
2615
|
contextWindow: 262144,
|
|
2933
2616
|
supportsImages: false,
|
|
2934
2617
|
supportsPromptCache: true,
|
|
2935
|
-
supportsNativeTools: true,
|
|
2936
|
-
defaultToolProtocol: "native",
|
|
2937
2618
|
inputPrice: 2.4,
|
|
2938
2619
|
// $2.40 per million tokens (cache miss)
|
|
2939
2620
|
outputPrice: 10,
|
|
@@ -2952,8 +2633,6 @@ var moonshotModels = {
|
|
|
2952
2633
|
supportsImages: false,
|
|
2953
2634
|
// Text-only (no image/vision support)
|
|
2954
2635
|
supportsPromptCache: true,
|
|
2955
|
-
supportsNativeTools: true,
|
|
2956
|
-
defaultToolProtocol: "native",
|
|
2957
2636
|
inputPrice: 0.6,
|
|
2958
2637
|
// $0.60 per million tokens (cache miss)
|
|
2959
2638
|
outputPrice: 2.5,
|
|
@@ -2967,6 +2646,21 @@ var moonshotModels = {
|
|
|
2967
2646
|
preserveReasoning: true,
|
|
2968
2647
|
defaultTemperature: 1,
|
|
2969
2648
|
description: `The kimi-k2-thinking model is a general-purpose agentic reasoning model developed by Moonshot AI. Thanks to its strength in deep reasoning and multi-turn tool use, it can solve even the hardest problems.`
|
|
2649
|
+
},
|
|
2650
|
+
"kimi-k2.5": {
|
|
2651
|
+
maxTokens: 16384,
|
|
2652
|
+
contextWindow: 262144,
|
|
2653
|
+
supportsImages: false,
|
|
2654
|
+
supportsPromptCache: true,
|
|
2655
|
+
inputPrice: 0.6,
|
|
2656
|
+
// $0.60 per million tokens (cache miss)
|
|
2657
|
+
outputPrice: 3,
|
|
2658
|
+
// $3.00 per million tokens
|
|
2659
|
+
cacheReadsPrice: 0.1,
|
|
2660
|
+
// $0.10 per million tokens (cache hit)
|
|
2661
|
+
supportsTemperature: true,
|
|
2662
|
+
defaultTemperature: 1,
|
|
2663
|
+
description: "Kimi K2.5 is the latest generation of Moonshot AI's Kimi series, featuring improved reasoning capabilities and enhanced performance across diverse tasks."
|
|
2970
2664
|
}
|
|
2971
2665
|
};
|
|
2972
2666
|
var MOONSHOT_DEFAULT_TEMPERATURE = 0.6;
|
|
@@ -2978,7 +2672,6 @@ var ollamaDefaultModelInfo = {
|
|
|
2978
2672
|
contextWindow: 2e5,
|
|
2979
2673
|
supportsImages: true,
|
|
2980
2674
|
supportsPromptCache: true,
|
|
2981
|
-
supportsNativeTools: true,
|
|
2982
2675
|
inputPrice: 0,
|
|
2983
2676
|
outputPrice: 0,
|
|
2984
2677
|
cacheWritesPrice: 0,
|
|
@@ -2992,8 +2685,6 @@ var openAiNativeModels = {
|
|
|
2992
2685
|
"gpt-5.1-codex-max": {
|
|
2993
2686
|
maxTokens: 128e3,
|
|
2994
2687
|
contextWindow: 4e5,
|
|
2995
|
-
supportsNativeTools: true,
|
|
2996
|
-
defaultToolProtocol: "native",
|
|
2997
2688
|
includedTools: ["apply_patch"],
|
|
2998
2689
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
2999
2690
|
supportsImages: true,
|
|
@@ -3011,8 +2702,6 @@ var openAiNativeModels = {
|
|
|
3011
2702
|
"gpt-5.2": {
|
|
3012
2703
|
maxTokens: 128e3,
|
|
3013
2704
|
contextWindow: 4e5,
|
|
3014
|
-
supportsNativeTools: true,
|
|
3015
|
-
defaultToolProtocol: "native",
|
|
3016
2705
|
includedTools: ["apply_patch"],
|
|
3017
2706
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3018
2707
|
supportsImages: true,
|
|
@@ -3034,8 +2723,6 @@ var openAiNativeModels = {
|
|
|
3034
2723
|
"gpt-5.2-codex": {
|
|
3035
2724
|
maxTokens: 128e3,
|
|
3036
2725
|
contextWindow: 4e5,
|
|
3037
|
-
supportsNativeTools: true,
|
|
3038
|
-
defaultToolProtocol: "native",
|
|
3039
2726
|
includedTools: ["apply_patch"],
|
|
3040
2727
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3041
2728
|
supportsImages: true,
|
|
@@ -3053,8 +2740,6 @@ var openAiNativeModels = {
|
|
|
3053
2740
|
"gpt-5.2-chat-latest": {
|
|
3054
2741
|
maxTokens: 16384,
|
|
3055
2742
|
contextWindow: 128e3,
|
|
3056
|
-
supportsNativeTools: true,
|
|
3057
|
-
defaultToolProtocol: "native",
|
|
3058
2743
|
includedTools: ["apply_patch"],
|
|
3059
2744
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3060
2745
|
supportsImages: true,
|
|
@@ -3067,8 +2752,6 @@ var openAiNativeModels = {
|
|
|
3067
2752
|
"gpt-5.1": {
|
|
3068
2753
|
maxTokens: 128e3,
|
|
3069
2754
|
contextWindow: 4e5,
|
|
3070
|
-
supportsNativeTools: true,
|
|
3071
|
-
defaultToolProtocol: "native",
|
|
3072
2755
|
includedTools: ["apply_patch"],
|
|
3073
2756
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3074
2757
|
supportsImages: true,
|
|
@@ -3090,8 +2773,6 @@ var openAiNativeModels = {
|
|
|
3090
2773
|
"gpt-5.1-codex": {
|
|
3091
2774
|
maxTokens: 128e3,
|
|
3092
2775
|
contextWindow: 4e5,
|
|
3093
|
-
supportsNativeTools: true,
|
|
3094
|
-
defaultToolProtocol: "native",
|
|
3095
2776
|
includedTools: ["apply_patch"],
|
|
3096
2777
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3097
2778
|
supportsImages: true,
|
|
@@ -3109,8 +2790,6 @@ var openAiNativeModels = {
|
|
|
3109
2790
|
"gpt-5.1-codex-mini": {
|
|
3110
2791
|
maxTokens: 128e3,
|
|
3111
2792
|
contextWindow: 4e5,
|
|
3112
|
-
supportsNativeTools: true,
|
|
3113
|
-
defaultToolProtocol: "native",
|
|
3114
2793
|
includedTools: ["apply_patch"],
|
|
3115
2794
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3116
2795
|
supportsImages: true,
|
|
@@ -3127,8 +2806,6 @@ var openAiNativeModels = {
|
|
|
3127
2806
|
"gpt-5": {
|
|
3128
2807
|
maxTokens: 128e3,
|
|
3129
2808
|
contextWindow: 4e5,
|
|
3130
|
-
supportsNativeTools: true,
|
|
3131
|
-
defaultToolProtocol: "native",
|
|
3132
2809
|
includedTools: ["apply_patch"],
|
|
3133
2810
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3134
2811
|
supportsImages: true,
|
|
@@ -3149,8 +2826,6 @@ var openAiNativeModels = {
|
|
|
3149
2826
|
"gpt-5-mini": {
|
|
3150
2827
|
maxTokens: 128e3,
|
|
3151
2828
|
contextWindow: 4e5,
|
|
3152
|
-
supportsNativeTools: true,
|
|
3153
|
-
defaultToolProtocol: "native",
|
|
3154
2829
|
includedTools: ["apply_patch"],
|
|
3155
2830
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3156
2831
|
supportsImages: true,
|
|
@@ -3171,8 +2846,6 @@ var openAiNativeModels = {
|
|
|
3171
2846
|
"gpt-5-codex": {
|
|
3172
2847
|
maxTokens: 128e3,
|
|
3173
2848
|
contextWindow: 4e5,
|
|
3174
|
-
supportsNativeTools: true,
|
|
3175
|
-
defaultToolProtocol: "native",
|
|
3176
2849
|
includedTools: ["apply_patch"],
|
|
3177
2850
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3178
2851
|
supportsImages: true,
|
|
@@ -3189,8 +2862,6 @@ var openAiNativeModels = {
|
|
|
3189
2862
|
"gpt-5-nano": {
|
|
3190
2863
|
maxTokens: 128e3,
|
|
3191
2864
|
contextWindow: 4e5,
|
|
3192
|
-
supportsNativeTools: true,
|
|
3193
|
-
defaultToolProtocol: "native",
|
|
3194
2865
|
includedTools: ["apply_patch"],
|
|
3195
2866
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3196
2867
|
supportsImages: true,
|
|
@@ -3208,8 +2879,6 @@ var openAiNativeModels = {
|
|
|
3208
2879
|
"gpt-5-chat-latest": {
|
|
3209
2880
|
maxTokens: 128e3,
|
|
3210
2881
|
contextWindow: 4e5,
|
|
3211
|
-
supportsNativeTools: true,
|
|
3212
|
-
defaultToolProtocol: "native",
|
|
3213
2882
|
includedTools: ["apply_patch"],
|
|
3214
2883
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3215
2884
|
supportsImages: true,
|
|
@@ -3222,8 +2891,6 @@ var openAiNativeModels = {
|
|
|
3222
2891
|
"gpt-4.1": {
|
|
3223
2892
|
maxTokens: 32768,
|
|
3224
2893
|
contextWindow: 1047576,
|
|
3225
|
-
supportsNativeTools: true,
|
|
3226
|
-
defaultToolProtocol: "native",
|
|
3227
2894
|
includedTools: ["apply_patch"],
|
|
3228
2895
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3229
2896
|
supportsImages: true,
|
|
@@ -3239,8 +2906,6 @@ var openAiNativeModels = {
|
|
|
3239
2906
|
"gpt-4.1-mini": {
|
|
3240
2907
|
maxTokens: 32768,
|
|
3241
2908
|
contextWindow: 1047576,
|
|
3242
|
-
supportsNativeTools: true,
|
|
3243
|
-
defaultToolProtocol: "native",
|
|
3244
2909
|
includedTools: ["apply_patch"],
|
|
3245
2910
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3246
2911
|
supportsImages: true,
|
|
@@ -3256,8 +2921,6 @@ var openAiNativeModels = {
|
|
|
3256
2921
|
"gpt-4.1-nano": {
|
|
3257
2922
|
maxTokens: 32768,
|
|
3258
2923
|
contextWindow: 1047576,
|
|
3259
|
-
supportsNativeTools: true,
|
|
3260
|
-
defaultToolProtocol: "native",
|
|
3261
2924
|
includedTools: ["apply_patch"],
|
|
3262
2925
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3263
2926
|
supportsImages: true,
|
|
@@ -3273,8 +2936,6 @@ var openAiNativeModels = {
|
|
|
3273
2936
|
o3: {
|
|
3274
2937
|
maxTokens: 1e5,
|
|
3275
2938
|
contextWindow: 2e5,
|
|
3276
|
-
supportsNativeTools: true,
|
|
3277
|
-
defaultToolProtocol: "native",
|
|
3278
2939
|
supportsImages: true,
|
|
3279
2940
|
supportsPromptCache: true,
|
|
3280
2941
|
inputPrice: 2,
|
|
@@ -3291,8 +2952,6 @@ var openAiNativeModels = {
|
|
|
3291
2952
|
"o3-high": {
|
|
3292
2953
|
maxTokens: 1e5,
|
|
3293
2954
|
contextWindow: 2e5,
|
|
3294
|
-
supportsNativeTools: true,
|
|
3295
|
-
defaultToolProtocol: "native",
|
|
3296
2955
|
supportsImages: true,
|
|
3297
2956
|
supportsPromptCache: true,
|
|
3298
2957
|
inputPrice: 2,
|
|
@@ -3304,8 +2963,6 @@ var openAiNativeModels = {
|
|
|
3304
2963
|
"o3-low": {
|
|
3305
2964
|
maxTokens: 1e5,
|
|
3306
2965
|
contextWindow: 2e5,
|
|
3307
|
-
supportsNativeTools: true,
|
|
3308
|
-
defaultToolProtocol: "native",
|
|
3309
2966
|
supportsImages: true,
|
|
3310
2967
|
supportsPromptCache: true,
|
|
3311
2968
|
inputPrice: 2,
|
|
@@ -3317,8 +2974,6 @@ var openAiNativeModels = {
|
|
|
3317
2974
|
"o4-mini": {
|
|
3318
2975
|
maxTokens: 1e5,
|
|
3319
2976
|
contextWindow: 2e5,
|
|
3320
|
-
supportsNativeTools: true,
|
|
3321
|
-
defaultToolProtocol: "native",
|
|
3322
2977
|
supportsImages: true,
|
|
3323
2978
|
supportsPromptCache: true,
|
|
3324
2979
|
inputPrice: 1.1,
|
|
@@ -3335,8 +2990,6 @@ var openAiNativeModels = {
|
|
|
3335
2990
|
"o4-mini-high": {
|
|
3336
2991
|
maxTokens: 1e5,
|
|
3337
2992
|
contextWindow: 2e5,
|
|
3338
|
-
supportsNativeTools: true,
|
|
3339
|
-
defaultToolProtocol: "native",
|
|
3340
2993
|
supportsImages: true,
|
|
3341
2994
|
supportsPromptCache: true,
|
|
3342
2995
|
inputPrice: 1.1,
|
|
@@ -3348,8 +3001,6 @@ var openAiNativeModels = {
|
|
|
3348
3001
|
"o4-mini-low": {
|
|
3349
3002
|
maxTokens: 1e5,
|
|
3350
3003
|
contextWindow: 2e5,
|
|
3351
|
-
supportsNativeTools: true,
|
|
3352
|
-
defaultToolProtocol: "native",
|
|
3353
3004
|
supportsImages: true,
|
|
3354
3005
|
supportsPromptCache: true,
|
|
3355
3006
|
inputPrice: 1.1,
|
|
@@ -3361,8 +3012,6 @@ var openAiNativeModels = {
|
|
|
3361
3012
|
"o3-mini": {
|
|
3362
3013
|
maxTokens: 1e5,
|
|
3363
3014
|
contextWindow: 2e5,
|
|
3364
|
-
supportsNativeTools: true,
|
|
3365
|
-
defaultToolProtocol: "native",
|
|
3366
3015
|
supportsImages: false,
|
|
3367
3016
|
supportsPromptCache: true,
|
|
3368
3017
|
inputPrice: 1.1,
|
|
@@ -3375,8 +3024,6 @@ var openAiNativeModels = {
|
|
|
3375
3024
|
"o3-mini-high": {
|
|
3376
3025
|
maxTokens: 1e5,
|
|
3377
3026
|
contextWindow: 2e5,
|
|
3378
|
-
supportsNativeTools: true,
|
|
3379
|
-
defaultToolProtocol: "native",
|
|
3380
3027
|
supportsImages: false,
|
|
3381
3028
|
supportsPromptCache: true,
|
|
3382
3029
|
inputPrice: 1.1,
|
|
@@ -3388,8 +3035,6 @@ var openAiNativeModels = {
|
|
|
3388
3035
|
"o3-mini-low": {
|
|
3389
3036
|
maxTokens: 1e5,
|
|
3390
3037
|
contextWindow: 2e5,
|
|
3391
|
-
supportsNativeTools: true,
|
|
3392
|
-
defaultToolProtocol: "native",
|
|
3393
3038
|
supportsImages: false,
|
|
3394
3039
|
supportsPromptCache: true,
|
|
3395
3040
|
inputPrice: 1.1,
|
|
@@ -3401,8 +3046,6 @@ var openAiNativeModels = {
|
|
|
3401
3046
|
o1: {
|
|
3402
3047
|
maxTokens: 1e5,
|
|
3403
3048
|
contextWindow: 2e5,
|
|
3404
|
-
supportsNativeTools: true,
|
|
3405
|
-
defaultToolProtocol: "native",
|
|
3406
3049
|
supportsImages: true,
|
|
3407
3050
|
supportsPromptCache: true,
|
|
3408
3051
|
inputPrice: 15,
|
|
@@ -3413,8 +3056,6 @@ var openAiNativeModels = {
|
|
|
3413
3056
|
"o1-preview": {
|
|
3414
3057
|
maxTokens: 32768,
|
|
3415
3058
|
contextWindow: 128e3,
|
|
3416
|
-
supportsNativeTools: true,
|
|
3417
|
-
defaultToolProtocol: "native",
|
|
3418
3059
|
supportsImages: true,
|
|
3419
3060
|
supportsPromptCache: true,
|
|
3420
3061
|
inputPrice: 15,
|
|
@@ -3425,8 +3066,6 @@ var openAiNativeModels = {
|
|
|
3425
3066
|
"o1-mini": {
|
|
3426
3067
|
maxTokens: 65536,
|
|
3427
3068
|
contextWindow: 128e3,
|
|
3428
|
-
supportsNativeTools: true,
|
|
3429
|
-
defaultToolProtocol: "native",
|
|
3430
3069
|
supportsImages: true,
|
|
3431
3070
|
supportsPromptCache: true,
|
|
3432
3071
|
inputPrice: 1.1,
|
|
@@ -3437,8 +3076,6 @@ var openAiNativeModels = {
|
|
|
3437
3076
|
"gpt-4o": {
|
|
3438
3077
|
maxTokens: 16384,
|
|
3439
3078
|
contextWindow: 128e3,
|
|
3440
|
-
supportsNativeTools: true,
|
|
3441
|
-
defaultToolProtocol: "native",
|
|
3442
3079
|
supportsImages: true,
|
|
3443
3080
|
supportsPromptCache: true,
|
|
3444
3081
|
inputPrice: 2.5,
|
|
@@ -3452,8 +3089,6 @@ var openAiNativeModels = {
|
|
|
3452
3089
|
"gpt-4o-mini": {
|
|
3453
3090
|
maxTokens: 16384,
|
|
3454
3091
|
contextWindow: 128e3,
|
|
3455
|
-
supportsNativeTools: true,
|
|
3456
|
-
defaultToolProtocol: "native",
|
|
3457
3092
|
supportsImages: true,
|
|
3458
3093
|
supportsPromptCache: true,
|
|
3459
3094
|
inputPrice: 0.15,
|
|
@@ -3467,8 +3102,6 @@ var openAiNativeModels = {
|
|
|
3467
3102
|
"codex-mini-latest": {
|
|
3468
3103
|
maxTokens: 16384,
|
|
3469
3104
|
contextWindow: 2e5,
|
|
3470
|
-
supportsNativeTools: true,
|
|
3471
|
-
defaultToolProtocol: "native",
|
|
3472
3105
|
supportsImages: false,
|
|
3473
3106
|
supportsPromptCache: false,
|
|
3474
3107
|
inputPrice: 1.5,
|
|
@@ -3481,8 +3114,6 @@ var openAiNativeModels = {
|
|
|
3481
3114
|
"gpt-5-2025-08-07": {
|
|
3482
3115
|
maxTokens: 128e3,
|
|
3483
3116
|
contextWindow: 4e5,
|
|
3484
|
-
supportsNativeTools: true,
|
|
3485
|
-
defaultToolProtocol: "native",
|
|
3486
3117
|
includedTools: ["apply_patch"],
|
|
3487
3118
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3488
3119
|
supportsImages: true,
|
|
@@ -3503,8 +3134,6 @@ var openAiNativeModels = {
|
|
|
3503
3134
|
"gpt-5-mini-2025-08-07": {
|
|
3504
3135
|
maxTokens: 128e3,
|
|
3505
3136
|
contextWindow: 4e5,
|
|
3506
|
-
supportsNativeTools: true,
|
|
3507
|
-
defaultToolProtocol: "native",
|
|
3508
3137
|
includedTools: ["apply_patch"],
|
|
3509
3138
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3510
3139
|
supportsImages: true,
|
|
@@ -3525,8 +3154,6 @@ var openAiNativeModels = {
|
|
|
3525
3154
|
"gpt-5-nano-2025-08-07": {
|
|
3526
3155
|
maxTokens: 128e3,
|
|
3527
3156
|
contextWindow: 4e5,
|
|
3528
|
-
supportsNativeTools: true,
|
|
3529
|
-
defaultToolProtocol: "native",
|
|
3530
3157
|
includedTools: ["apply_patch"],
|
|
3531
3158
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3532
3159
|
supportsImages: true,
|
|
@@ -3548,9 +3175,7 @@ var openAiModelInfoSaneDefaults = {
|
|
|
3548
3175
|
supportsImages: true,
|
|
3549
3176
|
supportsPromptCache: false,
|
|
3550
3177
|
inputPrice: 0,
|
|
3551
|
-
outputPrice: 0
|
|
3552
|
-
supportsNativeTools: true,
|
|
3553
|
-
defaultToolProtocol: "native"
|
|
3178
|
+
outputPrice: 0
|
|
3554
3179
|
};
|
|
3555
3180
|
var azureOpenAiDefaultApiVersion = "2024-08-01-preview";
|
|
3556
3181
|
var OPENAI_NATIVE_DEFAULT_TEMPERATURE = 0;
|
|
@@ -3562,8 +3187,6 @@ var openAiCodexModels = {
|
|
|
3562
3187
|
"gpt-5.1-codex-max": {
|
|
3563
3188
|
maxTokens: 128e3,
|
|
3564
3189
|
contextWindow: 4e5,
|
|
3565
|
-
supportsNativeTools: true,
|
|
3566
|
-
defaultToolProtocol: "native",
|
|
3567
3190
|
includedTools: ["apply_patch"],
|
|
3568
3191
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3569
3192
|
supportsImages: true,
|
|
@@ -3576,11 +3199,24 @@ var openAiCodexModels = {
|
|
|
3576
3199
|
supportsTemperature: false,
|
|
3577
3200
|
description: "GPT-5.1 Codex Max: Maximum capability coding model via ChatGPT subscription"
|
|
3578
3201
|
},
|
|
3202
|
+
"gpt-5.1-codex": {
|
|
3203
|
+
maxTokens: 128e3,
|
|
3204
|
+
contextWindow: 4e5,
|
|
3205
|
+
includedTools: ["apply_patch"],
|
|
3206
|
+
excludedTools: ["apply_diff", "write_to_file"],
|
|
3207
|
+
supportsImages: true,
|
|
3208
|
+
supportsPromptCache: true,
|
|
3209
|
+
supportsReasoningEffort: ["low", "medium", "high"],
|
|
3210
|
+
reasoningEffort: "medium",
|
|
3211
|
+
// Subscription-based: no per-token costs
|
|
3212
|
+
inputPrice: 0,
|
|
3213
|
+
outputPrice: 0,
|
|
3214
|
+
supportsTemperature: false,
|
|
3215
|
+
description: "GPT-5.1 Codex: GPT-5.1 optimized for agentic coding via ChatGPT subscription"
|
|
3216
|
+
},
|
|
3579
3217
|
"gpt-5.2-codex": {
|
|
3580
3218
|
maxTokens: 128e3,
|
|
3581
3219
|
contextWindow: 4e5,
|
|
3582
|
-
supportsNativeTools: true,
|
|
3583
|
-
defaultToolProtocol: "native",
|
|
3584
3220
|
includedTools: ["apply_patch"],
|
|
3585
3221
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3586
3222
|
supportsImages: true,
|
|
@@ -3592,11 +3228,71 @@ var openAiCodexModels = {
|
|
|
3592
3228
|
supportsTemperature: false,
|
|
3593
3229
|
description: "GPT-5.2 Codex: OpenAI's flagship coding model via ChatGPT subscription"
|
|
3594
3230
|
},
|
|
3231
|
+
"gpt-5.1": {
|
|
3232
|
+
maxTokens: 128e3,
|
|
3233
|
+
contextWindow: 4e5,
|
|
3234
|
+
includedTools: ["apply_patch"],
|
|
3235
|
+
excludedTools: ["apply_diff", "write_to_file"],
|
|
3236
|
+
supportsImages: true,
|
|
3237
|
+
supportsPromptCache: true,
|
|
3238
|
+
supportsReasoningEffort: ["none", "low", "medium", "high"],
|
|
3239
|
+
reasoningEffort: "medium",
|
|
3240
|
+
// Subscription-based: no per-token costs
|
|
3241
|
+
inputPrice: 0,
|
|
3242
|
+
outputPrice: 0,
|
|
3243
|
+
supportsVerbosity: true,
|
|
3244
|
+
supportsTemperature: false,
|
|
3245
|
+
description: "GPT-5.1: General GPT-5.1 model via ChatGPT subscription"
|
|
3246
|
+
},
|
|
3247
|
+
"gpt-5": {
|
|
3248
|
+
maxTokens: 128e3,
|
|
3249
|
+
contextWindow: 4e5,
|
|
3250
|
+
includedTools: ["apply_patch"],
|
|
3251
|
+
excludedTools: ["apply_diff", "write_to_file"],
|
|
3252
|
+
supportsImages: true,
|
|
3253
|
+
supportsPromptCache: true,
|
|
3254
|
+
supportsReasoningEffort: ["minimal", "low", "medium", "high"],
|
|
3255
|
+
reasoningEffort: "medium",
|
|
3256
|
+
// Subscription-based: no per-token costs
|
|
3257
|
+
inputPrice: 0,
|
|
3258
|
+
outputPrice: 0,
|
|
3259
|
+
supportsVerbosity: true,
|
|
3260
|
+
supportsTemperature: false,
|
|
3261
|
+
description: "GPT-5: General GPT-5 model via ChatGPT subscription"
|
|
3262
|
+
},
|
|
3263
|
+
"gpt-5-codex": {
|
|
3264
|
+
maxTokens: 128e3,
|
|
3265
|
+
contextWindow: 4e5,
|
|
3266
|
+
includedTools: ["apply_patch"],
|
|
3267
|
+
excludedTools: ["apply_diff", "write_to_file"],
|
|
3268
|
+
supportsImages: true,
|
|
3269
|
+
supportsPromptCache: true,
|
|
3270
|
+
supportsReasoningEffort: ["low", "medium", "high"],
|
|
3271
|
+
reasoningEffort: "medium",
|
|
3272
|
+
// Subscription-based: no per-token costs
|
|
3273
|
+
inputPrice: 0,
|
|
3274
|
+
outputPrice: 0,
|
|
3275
|
+
supportsTemperature: false,
|
|
3276
|
+
description: "GPT-5 Codex: GPT-5 optimized for agentic coding via ChatGPT subscription"
|
|
3277
|
+
},
|
|
3278
|
+
"gpt-5-codex-mini": {
|
|
3279
|
+
maxTokens: 128e3,
|
|
3280
|
+
contextWindow: 4e5,
|
|
3281
|
+
includedTools: ["apply_patch"],
|
|
3282
|
+
excludedTools: ["apply_diff", "write_to_file"],
|
|
3283
|
+
supportsImages: true,
|
|
3284
|
+
supportsPromptCache: true,
|
|
3285
|
+
supportsReasoningEffort: ["low", "medium", "high"],
|
|
3286
|
+
reasoningEffort: "medium",
|
|
3287
|
+
// Subscription-based: no per-token costs
|
|
3288
|
+
inputPrice: 0,
|
|
3289
|
+
outputPrice: 0,
|
|
3290
|
+
supportsTemperature: false,
|
|
3291
|
+
description: "GPT-5 Codex Mini: Faster coding model via ChatGPT subscription"
|
|
3292
|
+
},
|
|
3595
3293
|
"gpt-5.1-codex-mini": {
|
|
3596
3294
|
maxTokens: 128e3,
|
|
3597
3295
|
contextWindow: 4e5,
|
|
3598
|
-
supportsNativeTools: true,
|
|
3599
|
-
defaultToolProtocol: "native",
|
|
3600
3296
|
includedTools: ["apply_patch"],
|
|
3601
3297
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3602
3298
|
supportsImages: true,
|
|
@@ -3611,8 +3307,6 @@ var openAiCodexModels = {
|
|
|
3611
3307
|
"gpt-5.2": {
|
|
3612
3308
|
maxTokens: 128e3,
|
|
3613
3309
|
contextWindow: 4e5,
|
|
3614
|
-
supportsNativeTools: true,
|
|
3615
|
-
defaultToolProtocol: "native",
|
|
3616
3310
|
includedTools: ["apply_patch"],
|
|
3617
3311
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3618
3312
|
supportsImages: true,
|
|
@@ -3633,7 +3327,6 @@ var openRouterDefaultModelInfo = {
|
|
|
3633
3327
|
contextWindow: 2e5,
|
|
3634
3328
|
supportsImages: true,
|
|
3635
3329
|
supportsPromptCache: true,
|
|
3636
|
-
supportsNativeTools: true,
|
|
3637
3330
|
inputPrice: 3,
|
|
3638
3331
|
outputPrice: 15,
|
|
3639
3332
|
cacheWritesPrice: 3.75,
|
|
@@ -3707,8 +3400,6 @@ var qwenCodeModels = {
|
|
|
3707
3400
|
contextWindow: 1e6,
|
|
3708
3401
|
supportsImages: false,
|
|
3709
3402
|
supportsPromptCache: false,
|
|
3710
|
-
supportsNativeTools: true,
|
|
3711
|
-
defaultToolProtocol: "native",
|
|
3712
3403
|
inputPrice: 0,
|
|
3713
3404
|
outputPrice: 0,
|
|
3714
3405
|
cacheWritesPrice: 0,
|
|
@@ -3720,8 +3411,6 @@ var qwenCodeModels = {
|
|
|
3720
3411
|
contextWindow: 1e6,
|
|
3721
3412
|
supportsImages: false,
|
|
3722
3413
|
supportsPromptCache: false,
|
|
3723
|
-
supportsNativeTools: true,
|
|
3724
|
-
defaultToolProtocol: "native",
|
|
3725
3414
|
inputPrice: 0,
|
|
3726
3415
|
outputPrice: 0,
|
|
3727
3416
|
cacheWritesPrice: 0,
|
|
@@ -3737,8 +3426,6 @@ var requestyDefaultModelInfo = {
|
|
|
3737
3426
|
contextWindow: 2e5,
|
|
3738
3427
|
supportsImages: true,
|
|
3739
3428
|
supportsPromptCache: true,
|
|
3740
|
-
supportsNativeTools: true,
|
|
3741
|
-
defaultToolProtocol: "native",
|
|
3742
3429
|
inputPrice: 3,
|
|
3743
3430
|
outputPrice: 15,
|
|
3744
3431
|
cacheWritesPrice: 3.75,
|
|
@@ -3794,8 +3481,6 @@ var sambaNovaModels = {
|
|
|
3794
3481
|
contextWindow: 16384,
|
|
3795
3482
|
supportsImages: false,
|
|
3796
3483
|
supportsPromptCache: false,
|
|
3797
|
-
supportsNativeTools: true,
|
|
3798
|
-
defaultToolProtocol: "native",
|
|
3799
3484
|
inputPrice: 0.1,
|
|
3800
3485
|
outputPrice: 0.2,
|
|
3801
3486
|
description: "Meta Llama 3.1 8B Instruct model with 16K context window."
|
|
@@ -3805,8 +3490,6 @@ var sambaNovaModels = {
|
|
|
3805
3490
|
contextWindow: 131072,
|
|
3806
3491
|
supportsImages: false,
|
|
3807
3492
|
supportsPromptCache: false,
|
|
3808
|
-
supportsNativeTools: true,
|
|
3809
|
-
defaultToolProtocol: "native",
|
|
3810
3493
|
inputPrice: 0.6,
|
|
3811
3494
|
outputPrice: 1.2,
|
|
3812
3495
|
description: "Meta Llama 3.3 70B Instruct model with 128K context window."
|
|
@@ -3817,8 +3500,6 @@ var sambaNovaModels = {
|
|
|
3817
3500
|
supportsImages: false,
|
|
3818
3501
|
supportsPromptCache: false,
|
|
3819
3502
|
supportsReasoningBudget: true,
|
|
3820
|
-
supportsNativeTools: true,
|
|
3821
|
-
defaultToolProtocol: "native",
|
|
3822
3503
|
inputPrice: 5,
|
|
3823
3504
|
outputPrice: 7,
|
|
3824
3505
|
description: "DeepSeek R1 reasoning model with 32K context window."
|
|
@@ -3828,8 +3509,6 @@ var sambaNovaModels = {
|
|
|
3828
3509
|
contextWindow: 32768,
|
|
3829
3510
|
supportsImages: false,
|
|
3830
3511
|
supportsPromptCache: false,
|
|
3831
|
-
supportsNativeTools: true,
|
|
3832
|
-
defaultToolProtocol: "native",
|
|
3833
3512
|
inputPrice: 3,
|
|
3834
3513
|
outputPrice: 4.5,
|
|
3835
3514
|
description: "DeepSeek V3 model with 32K context window."
|
|
@@ -3839,8 +3518,6 @@ var sambaNovaModels = {
|
|
|
3839
3518
|
contextWindow: 32768,
|
|
3840
3519
|
supportsImages: false,
|
|
3841
3520
|
supportsPromptCache: false,
|
|
3842
|
-
supportsNativeTools: true,
|
|
3843
|
-
defaultToolProtocol: "native",
|
|
3844
3521
|
inputPrice: 3,
|
|
3845
3522
|
outputPrice: 4.5,
|
|
3846
3523
|
description: "DeepSeek V3.1 model with 32K context window."
|
|
@@ -3850,8 +3527,6 @@ var sambaNovaModels = {
|
|
|
3850
3527
|
contextWindow: 131072,
|
|
3851
3528
|
supportsImages: true,
|
|
3852
3529
|
supportsPromptCache: false,
|
|
3853
|
-
supportsNativeTools: true,
|
|
3854
|
-
defaultToolProtocol: "native",
|
|
3855
3530
|
inputPrice: 0.63,
|
|
3856
3531
|
outputPrice: 1.8,
|
|
3857
3532
|
description: "Meta Llama 4 Maverick 17B 128E Instruct model with 128K context window."
|
|
@@ -3861,8 +3536,6 @@ var sambaNovaModels = {
|
|
|
3861
3536
|
contextWindow: 8192,
|
|
3862
3537
|
supportsImages: false,
|
|
3863
3538
|
supportsPromptCache: false,
|
|
3864
|
-
supportsNativeTools: true,
|
|
3865
|
-
defaultToolProtocol: "native",
|
|
3866
3539
|
inputPrice: 0.4,
|
|
3867
3540
|
outputPrice: 0.8,
|
|
3868
3541
|
description: "Alibaba Qwen 3 32B model with 8K context window."
|
|
@@ -3872,8 +3545,6 @@ var sambaNovaModels = {
|
|
|
3872
3545
|
contextWindow: 131072,
|
|
3873
3546
|
supportsImages: false,
|
|
3874
3547
|
supportsPromptCache: false,
|
|
3875
|
-
supportsNativeTools: true,
|
|
3876
|
-
defaultToolProtocol: "native",
|
|
3877
3548
|
inputPrice: 0.22,
|
|
3878
3549
|
outputPrice: 0.59,
|
|
3879
3550
|
description: "OpenAI gpt oss 120b model with 128k context window."
|
|
@@ -3887,7 +3558,6 @@ var unboundDefaultModelInfo = {
|
|
|
3887
3558
|
contextWindow: 2e5,
|
|
3888
3559
|
supportsImages: true,
|
|
3889
3560
|
supportsPromptCache: true,
|
|
3890
|
-
supportsNativeTools: true,
|
|
3891
3561
|
inputPrice: 3,
|
|
3892
3562
|
outputPrice: 15,
|
|
3893
3563
|
cacheWritesPrice: 3.75,
|
|
@@ -3901,8 +3571,6 @@ var vertexModels = {
|
|
|
3901
3571
|
maxTokens: 65536,
|
|
3902
3572
|
contextWindow: 1048576,
|
|
3903
3573
|
supportsImages: true,
|
|
3904
|
-
supportsNativeTools: true,
|
|
3905
|
-
defaultToolProtocol: "native",
|
|
3906
3574
|
supportsPromptCache: true,
|
|
3907
3575
|
supportsReasoningEffort: ["low", "high"],
|
|
3908
3576
|
reasoningEffort: "low",
|
|
@@ -3910,16 +3578,19 @@ var vertexModels = {
|
|
|
3910
3578
|
defaultTemperature: 1,
|
|
3911
3579
|
inputPrice: 4,
|
|
3912
3580
|
outputPrice: 18,
|
|
3581
|
+
cacheReadsPrice: 0.4,
|
|
3913
3582
|
tiers: [
|
|
3914
3583
|
{
|
|
3915
3584
|
contextWindow: 2e5,
|
|
3916
3585
|
inputPrice: 2,
|
|
3917
|
-
outputPrice: 12
|
|
3586
|
+
outputPrice: 12,
|
|
3587
|
+
cacheReadsPrice: 0.2
|
|
3918
3588
|
},
|
|
3919
3589
|
{
|
|
3920
3590
|
contextWindow: Infinity,
|
|
3921
3591
|
inputPrice: 4,
|
|
3922
|
-
outputPrice: 18
|
|
3592
|
+
outputPrice: 18,
|
|
3593
|
+
cacheReadsPrice: 0.4
|
|
3923
3594
|
}
|
|
3924
3595
|
]
|
|
3925
3596
|
},
|
|
@@ -3927,24 +3598,19 @@ var vertexModels = {
|
|
|
3927
3598
|
maxTokens: 65536,
|
|
3928
3599
|
contextWindow: 1048576,
|
|
3929
3600
|
supportsImages: true,
|
|
3930
|
-
supportsNativeTools: true,
|
|
3931
|
-
defaultToolProtocol: "native",
|
|
3932
3601
|
supportsPromptCache: true,
|
|
3933
3602
|
supportsReasoningEffort: ["minimal", "low", "medium", "high"],
|
|
3934
3603
|
reasoningEffort: "medium",
|
|
3935
3604
|
supportsTemperature: true,
|
|
3936
3605
|
defaultTemperature: 1,
|
|
3937
|
-
inputPrice: 0.
|
|
3938
|
-
outputPrice:
|
|
3939
|
-
cacheReadsPrice: 0.
|
|
3940
|
-
cacheWritesPrice: 1
|
|
3606
|
+
inputPrice: 0.5,
|
|
3607
|
+
outputPrice: 3,
|
|
3608
|
+
cacheReadsPrice: 0.05
|
|
3941
3609
|
},
|
|
3942
3610
|
"gemini-2.5-flash-preview-05-20:thinking": {
|
|
3943
3611
|
maxTokens: 65535,
|
|
3944
3612
|
contextWindow: 1048576,
|
|
3945
3613
|
supportsImages: true,
|
|
3946
|
-
supportsNativeTools: true,
|
|
3947
|
-
defaultToolProtocol: "native",
|
|
3948
3614
|
supportsPromptCache: true,
|
|
3949
3615
|
inputPrice: 0.15,
|
|
3950
3616
|
outputPrice: 3.5,
|
|
@@ -3956,8 +3622,6 @@ var vertexModels = {
|
|
|
3956
3622
|
maxTokens: 65535,
|
|
3957
3623
|
contextWindow: 1048576,
|
|
3958
3624
|
supportsImages: true,
|
|
3959
|
-
supportsNativeTools: true,
|
|
3960
|
-
defaultToolProtocol: "native",
|
|
3961
3625
|
supportsPromptCache: true,
|
|
3962
3626
|
inputPrice: 0.15,
|
|
3963
3627
|
outputPrice: 0.6
|
|
@@ -3966,8 +3630,6 @@ var vertexModels = {
|
|
|
3966
3630
|
maxTokens: 64e3,
|
|
3967
3631
|
contextWindow: 1048576,
|
|
3968
3632
|
supportsImages: true,
|
|
3969
|
-
supportsNativeTools: true,
|
|
3970
|
-
defaultToolProtocol: "native",
|
|
3971
3633
|
supportsPromptCache: true,
|
|
3972
3634
|
inputPrice: 0.3,
|
|
3973
3635
|
outputPrice: 2.5,
|
|
@@ -3980,8 +3642,6 @@ var vertexModels = {
|
|
|
3980
3642
|
maxTokens: 65535,
|
|
3981
3643
|
contextWindow: 1048576,
|
|
3982
3644
|
supportsImages: true,
|
|
3983
|
-
supportsNativeTools: true,
|
|
3984
|
-
defaultToolProtocol: "native",
|
|
3985
3645
|
supportsPromptCache: false,
|
|
3986
3646
|
inputPrice: 0.15,
|
|
3987
3647
|
outputPrice: 3.5,
|
|
@@ -3993,8 +3653,6 @@ var vertexModels = {
|
|
|
3993
3653
|
maxTokens: 65535,
|
|
3994
3654
|
contextWindow: 1048576,
|
|
3995
3655
|
supportsImages: true,
|
|
3996
|
-
supportsNativeTools: true,
|
|
3997
|
-
defaultToolProtocol: "native",
|
|
3998
3656
|
supportsPromptCache: false,
|
|
3999
3657
|
inputPrice: 0.15,
|
|
4000
3658
|
outputPrice: 0.6
|
|
@@ -4003,8 +3661,6 @@ var vertexModels = {
|
|
|
4003
3661
|
maxTokens: 65535,
|
|
4004
3662
|
contextWindow: 1048576,
|
|
4005
3663
|
supportsImages: true,
|
|
4006
|
-
supportsNativeTools: true,
|
|
4007
|
-
defaultToolProtocol: "native",
|
|
4008
3664
|
supportsPromptCache: true,
|
|
4009
3665
|
inputPrice: 2.5,
|
|
4010
3666
|
outputPrice: 15
|
|
@@ -4013,8 +3669,6 @@ var vertexModels = {
|
|
|
4013
3669
|
maxTokens: 65535,
|
|
4014
3670
|
contextWindow: 1048576,
|
|
4015
3671
|
supportsImages: true,
|
|
4016
|
-
supportsNativeTools: true,
|
|
4017
|
-
defaultToolProtocol: "native",
|
|
4018
3672
|
supportsPromptCache: true,
|
|
4019
3673
|
inputPrice: 2.5,
|
|
4020
3674
|
outputPrice: 15
|
|
@@ -4023,8 +3677,6 @@ var vertexModels = {
|
|
|
4023
3677
|
maxTokens: 65535,
|
|
4024
3678
|
contextWindow: 1048576,
|
|
4025
3679
|
supportsImages: true,
|
|
4026
|
-
supportsNativeTools: true,
|
|
4027
|
-
defaultToolProtocol: "native",
|
|
4028
3680
|
supportsPromptCache: true,
|
|
4029
3681
|
inputPrice: 2.5,
|
|
4030
3682
|
outputPrice: 15,
|
|
@@ -4035,8 +3687,6 @@ var vertexModels = {
|
|
|
4035
3687
|
maxTokens: 64e3,
|
|
4036
3688
|
contextWindow: 1048576,
|
|
4037
3689
|
supportsImages: true,
|
|
4038
|
-
supportsNativeTools: true,
|
|
4039
|
-
defaultToolProtocol: "native",
|
|
4040
3690
|
supportsPromptCache: true,
|
|
4041
3691
|
inputPrice: 2.5,
|
|
4042
3692
|
outputPrice: 15,
|
|
@@ -4062,8 +3712,6 @@ var vertexModels = {
|
|
|
4062
3712
|
maxTokens: 65535,
|
|
4063
3713
|
contextWindow: 1048576,
|
|
4064
3714
|
supportsImages: true,
|
|
4065
|
-
supportsNativeTools: true,
|
|
4066
|
-
defaultToolProtocol: "native",
|
|
4067
3715
|
supportsPromptCache: false,
|
|
4068
3716
|
inputPrice: 0,
|
|
4069
3717
|
outputPrice: 0
|
|
@@ -4072,8 +3720,6 @@ var vertexModels = {
|
|
|
4072
3720
|
maxTokens: 8192,
|
|
4073
3721
|
contextWindow: 2097152,
|
|
4074
3722
|
supportsImages: true,
|
|
4075
|
-
supportsNativeTools: true,
|
|
4076
|
-
defaultToolProtocol: "native",
|
|
4077
3723
|
supportsPromptCache: false,
|
|
4078
3724
|
inputPrice: 0,
|
|
4079
3725
|
outputPrice: 0
|
|
@@ -4082,8 +3728,6 @@ var vertexModels = {
|
|
|
4082
3728
|
maxTokens: 8192,
|
|
4083
3729
|
contextWindow: 1048576,
|
|
4084
3730
|
supportsImages: true,
|
|
4085
|
-
supportsNativeTools: true,
|
|
4086
|
-
defaultToolProtocol: "native",
|
|
4087
3731
|
supportsPromptCache: true,
|
|
4088
3732
|
inputPrice: 0.15,
|
|
4089
3733
|
outputPrice: 0.6
|
|
@@ -4092,8 +3736,6 @@ var vertexModels = {
|
|
|
4092
3736
|
maxTokens: 8192,
|
|
4093
3737
|
contextWindow: 1048576,
|
|
4094
3738
|
supportsImages: true,
|
|
4095
|
-
supportsNativeTools: true,
|
|
4096
|
-
defaultToolProtocol: "native",
|
|
4097
3739
|
supportsPromptCache: false,
|
|
4098
3740
|
inputPrice: 0.075,
|
|
4099
3741
|
outputPrice: 0.3
|
|
@@ -4102,8 +3744,6 @@ var vertexModels = {
|
|
|
4102
3744
|
maxTokens: 8192,
|
|
4103
3745
|
contextWindow: 32768,
|
|
4104
3746
|
supportsImages: true,
|
|
4105
|
-
supportsNativeTools: true,
|
|
4106
|
-
defaultToolProtocol: "native",
|
|
4107
3747
|
supportsPromptCache: false,
|
|
4108
3748
|
inputPrice: 0,
|
|
4109
3749
|
outputPrice: 0
|
|
@@ -4112,8 +3752,6 @@ var vertexModels = {
|
|
|
4112
3752
|
maxTokens: 8192,
|
|
4113
3753
|
contextWindow: 1048576,
|
|
4114
3754
|
supportsImages: true,
|
|
4115
|
-
supportsNativeTools: true,
|
|
4116
|
-
defaultToolProtocol: "native",
|
|
4117
3755
|
supportsPromptCache: true,
|
|
4118
3756
|
inputPrice: 0.075,
|
|
4119
3757
|
outputPrice: 0.3
|
|
@@ -4122,8 +3760,6 @@ var vertexModels = {
|
|
|
4122
3760
|
maxTokens: 8192,
|
|
4123
3761
|
contextWindow: 2097152,
|
|
4124
3762
|
supportsImages: true,
|
|
4125
|
-
supportsNativeTools: true,
|
|
4126
|
-
defaultToolProtocol: "native",
|
|
4127
3763
|
supportsPromptCache: false,
|
|
4128
3764
|
inputPrice: 1.25,
|
|
4129
3765
|
outputPrice: 5
|
|
@@ -4134,8 +3770,6 @@ var vertexModels = {
|
|
|
4134
3770
|
// Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
|
|
4135
3771
|
supportsImages: true,
|
|
4136
3772
|
supportsPromptCache: true,
|
|
4137
|
-
supportsNativeTools: true,
|
|
4138
|
-
defaultToolProtocol: "native",
|
|
4139
3773
|
inputPrice: 3,
|
|
4140
3774
|
// $3 per million input tokens (≤200K context)
|
|
4141
3775
|
outputPrice: 15,
|
|
@@ -4167,8 +3801,6 @@ var vertexModels = {
|
|
|
4167
3801
|
// Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
|
|
4168
3802
|
supportsImages: true,
|
|
4169
3803
|
supportsPromptCache: true,
|
|
4170
|
-
supportsNativeTools: true,
|
|
4171
|
-
defaultToolProtocol: "native",
|
|
4172
3804
|
inputPrice: 3,
|
|
4173
3805
|
// $3 per million input tokens (≤200K context)
|
|
4174
3806
|
outputPrice: 15,
|
|
@@ -4199,8 +3831,6 @@ var vertexModels = {
|
|
|
4199
3831
|
contextWindow: 2e5,
|
|
4200
3832
|
supportsImages: true,
|
|
4201
3833
|
supportsPromptCache: true,
|
|
4202
|
-
supportsNativeTools: true,
|
|
4203
|
-
defaultToolProtocol: "native",
|
|
4204
3834
|
inputPrice: 1,
|
|
4205
3835
|
outputPrice: 5,
|
|
4206
3836
|
cacheWritesPrice: 1.25,
|
|
@@ -4212,8 +3842,6 @@ var vertexModels = {
|
|
|
4212
3842
|
contextWindow: 2e5,
|
|
4213
3843
|
supportsImages: true,
|
|
4214
3844
|
supportsPromptCache: true,
|
|
4215
|
-
supportsNativeTools: true,
|
|
4216
|
-
defaultToolProtocol: "native",
|
|
4217
3845
|
inputPrice: 5,
|
|
4218
3846
|
outputPrice: 25,
|
|
4219
3847
|
cacheWritesPrice: 6.25,
|
|
@@ -4225,8 +3853,6 @@ var vertexModels = {
|
|
|
4225
3853
|
contextWindow: 2e5,
|
|
4226
3854
|
supportsImages: true,
|
|
4227
3855
|
supportsPromptCache: true,
|
|
4228
|
-
supportsNativeTools: true,
|
|
4229
|
-
defaultToolProtocol: "native",
|
|
4230
3856
|
inputPrice: 15,
|
|
4231
3857
|
outputPrice: 75,
|
|
4232
3858
|
cacheWritesPrice: 18.75,
|
|
@@ -4238,8 +3864,6 @@ var vertexModels = {
|
|
|
4238
3864
|
contextWindow: 2e5,
|
|
4239
3865
|
supportsImages: true,
|
|
4240
3866
|
supportsPromptCache: true,
|
|
4241
|
-
supportsNativeTools: true,
|
|
4242
|
-
defaultToolProtocol: "native",
|
|
4243
3867
|
inputPrice: 15,
|
|
4244
3868
|
outputPrice: 75,
|
|
4245
3869
|
cacheWritesPrice: 18.75,
|
|
@@ -4250,8 +3874,6 @@ var vertexModels = {
|
|
|
4250
3874
|
contextWindow: 2e5,
|
|
4251
3875
|
supportsImages: true,
|
|
4252
3876
|
supportsPromptCache: true,
|
|
4253
|
-
supportsNativeTools: true,
|
|
4254
|
-
defaultToolProtocol: "native",
|
|
4255
3877
|
inputPrice: 3,
|
|
4256
3878
|
outputPrice: 15,
|
|
4257
3879
|
cacheWritesPrice: 3.75,
|
|
@@ -4264,8 +3886,6 @@ var vertexModels = {
|
|
|
4264
3886
|
contextWindow: 2e5,
|
|
4265
3887
|
supportsImages: true,
|
|
4266
3888
|
supportsPromptCache: true,
|
|
4267
|
-
supportsNativeTools: true,
|
|
4268
|
-
defaultToolProtocol: "native",
|
|
4269
3889
|
inputPrice: 3,
|
|
4270
3890
|
outputPrice: 15,
|
|
4271
3891
|
cacheWritesPrice: 3.75,
|
|
@@ -4276,8 +3896,6 @@ var vertexModels = {
|
|
|
4276
3896
|
contextWindow: 2e5,
|
|
4277
3897
|
supportsImages: true,
|
|
4278
3898
|
supportsPromptCache: true,
|
|
4279
|
-
supportsNativeTools: true,
|
|
4280
|
-
defaultToolProtocol: "native",
|
|
4281
3899
|
inputPrice: 3,
|
|
4282
3900
|
outputPrice: 15,
|
|
4283
3901
|
cacheWritesPrice: 3.75,
|
|
@@ -4288,8 +3906,6 @@ var vertexModels = {
|
|
|
4288
3906
|
contextWindow: 2e5,
|
|
4289
3907
|
supportsImages: true,
|
|
4290
3908
|
supportsPromptCache: true,
|
|
4291
|
-
supportsNativeTools: true,
|
|
4292
|
-
defaultToolProtocol: "native",
|
|
4293
3909
|
inputPrice: 3,
|
|
4294
3910
|
outputPrice: 15,
|
|
4295
3911
|
cacheWritesPrice: 3.75,
|
|
@@ -4300,8 +3916,6 @@ var vertexModels = {
|
|
|
4300
3916
|
contextWindow: 2e5,
|
|
4301
3917
|
supportsImages: false,
|
|
4302
3918
|
supportsPromptCache: true,
|
|
4303
|
-
supportsNativeTools: true,
|
|
4304
|
-
defaultToolProtocol: "native",
|
|
4305
3919
|
inputPrice: 1,
|
|
4306
3920
|
outputPrice: 5,
|
|
4307
3921
|
cacheWritesPrice: 1.25,
|
|
@@ -4312,8 +3926,6 @@ var vertexModels = {
|
|
|
4312
3926
|
contextWindow: 2e5,
|
|
4313
3927
|
supportsImages: true,
|
|
4314
3928
|
supportsPromptCache: true,
|
|
4315
|
-
supportsNativeTools: true,
|
|
4316
|
-
defaultToolProtocol: "native",
|
|
4317
3929
|
inputPrice: 15,
|
|
4318
3930
|
outputPrice: 75,
|
|
4319
3931
|
cacheWritesPrice: 18.75,
|
|
@@ -4324,8 +3936,6 @@ var vertexModels = {
|
|
|
4324
3936
|
contextWindow: 2e5,
|
|
4325
3937
|
supportsImages: true,
|
|
4326
3938
|
supportsPromptCache: true,
|
|
4327
|
-
supportsNativeTools: true,
|
|
4328
|
-
defaultToolProtocol: "native",
|
|
4329
3939
|
inputPrice: 0.25,
|
|
4330
3940
|
outputPrice: 1.25,
|
|
4331
3941
|
cacheWritesPrice: 0.3,
|
|
@@ -4335,8 +3945,6 @@ var vertexModels = {
|
|
|
4335
3945
|
maxTokens: 64e3,
|
|
4336
3946
|
contextWindow: 1048576,
|
|
4337
3947
|
supportsImages: true,
|
|
4338
|
-
supportsNativeTools: true,
|
|
4339
|
-
defaultToolProtocol: "native",
|
|
4340
3948
|
supportsPromptCache: true,
|
|
4341
3949
|
inputPrice: 0.1,
|
|
4342
3950
|
outputPrice: 0.4,
|
|
@@ -4350,7 +3958,6 @@ var vertexModels = {
|
|
|
4350
3958
|
contextWindow: 131072,
|
|
4351
3959
|
supportsImages: false,
|
|
4352
3960
|
supportsPromptCache: false,
|
|
4353
|
-
supportsNativeTools: true,
|
|
4354
3961
|
inputPrice: 0.35,
|
|
4355
3962
|
outputPrice: 1.15,
|
|
4356
3963
|
description: "Meta Llama 4 Maverick 17B Instruct model, 128K context."
|
|
@@ -4360,7 +3967,6 @@ var vertexModels = {
|
|
|
4360
3967
|
contextWindow: 163840,
|
|
4361
3968
|
supportsImages: false,
|
|
4362
3969
|
supportsPromptCache: false,
|
|
4363
|
-
supportsNativeTools: true,
|
|
4364
3970
|
inputPrice: 1.35,
|
|
4365
3971
|
outputPrice: 5.4,
|
|
4366
3972
|
description: "DeepSeek R1 (0528). Available in us-central1"
|
|
@@ -4370,7 +3976,6 @@ var vertexModels = {
|
|
|
4370
3976
|
contextWindow: 163840,
|
|
4371
3977
|
supportsImages: false,
|
|
4372
3978
|
supportsPromptCache: false,
|
|
4373
|
-
supportsNativeTools: true,
|
|
4374
3979
|
inputPrice: 0.6,
|
|
4375
3980
|
outputPrice: 1.7,
|
|
4376
3981
|
description: "DeepSeek V3.1. Available in us-west2"
|
|
@@ -4380,7 +3985,6 @@ var vertexModels = {
|
|
|
4380
3985
|
contextWindow: 131072,
|
|
4381
3986
|
supportsImages: false,
|
|
4382
3987
|
supportsPromptCache: false,
|
|
4383
|
-
supportsNativeTools: true,
|
|
4384
3988
|
inputPrice: 0.15,
|
|
4385
3989
|
outputPrice: 0.6,
|
|
4386
3990
|
description: "OpenAI gpt-oss 120B. Available in us-central1"
|
|
@@ -4390,7 +3994,6 @@ var vertexModels = {
|
|
|
4390
3994
|
contextWindow: 131072,
|
|
4391
3995
|
supportsImages: false,
|
|
4392
3996
|
supportsPromptCache: false,
|
|
4393
|
-
supportsNativeTools: true,
|
|
4394
3997
|
inputPrice: 0.075,
|
|
4395
3998
|
outputPrice: 0.3,
|
|
4396
3999
|
description: "OpenAI gpt-oss 20B. Available in us-central1"
|
|
@@ -4400,7 +4003,6 @@ var vertexModels = {
|
|
|
4400
4003
|
contextWindow: 262144,
|
|
4401
4004
|
supportsImages: false,
|
|
4402
4005
|
supportsPromptCache: false,
|
|
4403
|
-
supportsNativeTools: true,
|
|
4404
4006
|
inputPrice: 1,
|
|
4405
4007
|
outputPrice: 4,
|
|
4406
4008
|
description: "Qwen3 Coder 480B A35B Instruct. Available in us-south1"
|
|
@@ -4410,10 +4012,18 @@ var vertexModels = {
|
|
|
4410
4012
|
contextWindow: 262144,
|
|
4411
4013
|
supportsImages: false,
|
|
4412
4014
|
supportsPromptCache: false,
|
|
4413
|
-
supportsNativeTools: true,
|
|
4414
4015
|
inputPrice: 0.25,
|
|
4415
4016
|
outputPrice: 1,
|
|
4416
4017
|
description: "Qwen3 235B A22B Instruct. Available in us-south1"
|
|
4018
|
+
},
|
|
4019
|
+
"moonshotai/kimi-k2-thinking-maas": {
|
|
4020
|
+
maxTokens: 16384,
|
|
4021
|
+
contextWindow: 262144,
|
|
4022
|
+
supportsPromptCache: false,
|
|
4023
|
+
supportsImages: false,
|
|
4024
|
+
inputPrice: 0.6,
|
|
4025
|
+
outputPrice: 2.5,
|
|
4026
|
+
description: "Kimi K2 Thinking Model with 256K context window."
|
|
4417
4027
|
}
|
|
4418
4028
|
};
|
|
4419
4029
|
var VERTEX_1M_CONTEXT_MODEL_IDS = ["claude-sonnet-4@20250514", "claude-sonnet-4-5@20250929"];
|
|
@@ -4646,8 +4256,6 @@ var xaiModels = {
|
|
|
4646
4256
|
contextWindow: 256e3,
|
|
4647
4257
|
supportsImages: true,
|
|
4648
4258
|
supportsPromptCache: true,
|
|
4649
|
-
supportsNativeTools: true,
|
|
4650
|
-
defaultToolProtocol: "native",
|
|
4651
4259
|
inputPrice: 0.2,
|
|
4652
4260
|
outputPrice: 1.5,
|
|
4653
4261
|
cacheWritesPrice: 0.02,
|
|
@@ -4661,8 +4269,6 @@ var xaiModels = {
|
|
|
4661
4269
|
contextWindow: 2e6,
|
|
4662
4270
|
supportsImages: true,
|
|
4663
4271
|
supportsPromptCache: true,
|
|
4664
|
-
supportsNativeTools: true,
|
|
4665
|
-
defaultToolProtocol: "native",
|
|
4666
4272
|
inputPrice: 0.2,
|
|
4667
4273
|
outputPrice: 0.5,
|
|
4668
4274
|
cacheWritesPrice: 0.05,
|
|
@@ -4676,8 +4282,6 @@ var xaiModels = {
|
|
|
4676
4282
|
contextWindow: 2e6,
|
|
4677
4283
|
supportsImages: true,
|
|
4678
4284
|
supportsPromptCache: true,
|
|
4679
|
-
supportsNativeTools: true,
|
|
4680
|
-
defaultToolProtocol: "native",
|
|
4681
4285
|
inputPrice: 0.2,
|
|
4682
4286
|
outputPrice: 0.5,
|
|
4683
4287
|
cacheWritesPrice: 0.05,
|
|
@@ -4691,8 +4295,6 @@ var xaiModels = {
|
|
|
4691
4295
|
contextWindow: 2e6,
|
|
4692
4296
|
supportsImages: true,
|
|
4693
4297
|
supportsPromptCache: true,
|
|
4694
|
-
supportsNativeTools: true,
|
|
4695
|
-
defaultToolProtocol: "native",
|
|
4696
4298
|
inputPrice: 0.2,
|
|
4697
4299
|
outputPrice: 0.5,
|
|
4698
4300
|
cacheWritesPrice: 0.05,
|
|
@@ -4706,8 +4308,6 @@ var xaiModels = {
|
|
|
4706
4308
|
contextWindow: 2e6,
|
|
4707
4309
|
supportsImages: true,
|
|
4708
4310
|
supportsPromptCache: true,
|
|
4709
|
-
supportsNativeTools: true,
|
|
4710
|
-
defaultToolProtocol: "native",
|
|
4711
4311
|
inputPrice: 0.2,
|
|
4712
4312
|
outputPrice: 0.5,
|
|
4713
4313
|
cacheWritesPrice: 0.05,
|
|
@@ -4721,8 +4321,6 @@ var xaiModels = {
|
|
|
4721
4321
|
contextWindow: 256e3,
|
|
4722
4322
|
supportsImages: true,
|
|
4723
4323
|
supportsPromptCache: true,
|
|
4724
|
-
supportsNativeTools: true,
|
|
4725
|
-
defaultToolProtocol: "native",
|
|
4726
4324
|
inputPrice: 3,
|
|
4727
4325
|
outputPrice: 15,
|
|
4728
4326
|
cacheWritesPrice: 0.75,
|
|
@@ -4736,8 +4334,6 @@ var xaiModels = {
|
|
|
4736
4334
|
contextWindow: 131072,
|
|
4737
4335
|
supportsImages: true,
|
|
4738
4336
|
supportsPromptCache: true,
|
|
4739
|
-
supportsNativeTools: true,
|
|
4740
|
-
defaultToolProtocol: "native",
|
|
4741
4337
|
inputPrice: 0.3,
|
|
4742
4338
|
outputPrice: 0.5,
|
|
4743
4339
|
cacheWritesPrice: 0.07,
|
|
@@ -4753,8 +4349,6 @@ var xaiModels = {
|
|
|
4753
4349
|
contextWindow: 131072,
|
|
4754
4350
|
supportsImages: true,
|
|
4755
4351
|
supportsPromptCache: true,
|
|
4756
|
-
supportsNativeTools: true,
|
|
4757
|
-
defaultToolProtocol: "native",
|
|
4758
4352
|
inputPrice: 3,
|
|
4759
4353
|
outputPrice: 15,
|
|
4760
4354
|
cacheWritesPrice: 0.75,
|
|
@@ -4851,7 +4445,6 @@ var vercelAiGatewayDefaultModelInfo = {
|
|
|
4851
4445
|
contextWindow: 2e5,
|
|
4852
4446
|
supportsImages: true,
|
|
4853
4447
|
supportsPromptCache: true,
|
|
4854
|
-
supportsNativeTools: true,
|
|
4855
4448
|
inputPrice: 3,
|
|
4856
4449
|
outputPrice: 15,
|
|
4857
4450
|
cacheWritesPrice: 3.75,
|
|
@@ -4868,8 +4461,6 @@ var internationalZAiModels = {
|
|
|
4868
4461
|
contextWindow: 131072,
|
|
4869
4462
|
supportsImages: false,
|
|
4870
4463
|
supportsPromptCache: true,
|
|
4871
|
-
supportsNativeTools: true,
|
|
4872
|
-
defaultToolProtocol: "native",
|
|
4873
4464
|
inputPrice: 0.6,
|
|
4874
4465
|
outputPrice: 2.2,
|
|
4875
4466
|
cacheWritesPrice: 0,
|
|
@@ -4881,8 +4472,6 @@ var internationalZAiModels = {
|
|
|
4881
4472
|
contextWindow: 131072,
|
|
4882
4473
|
supportsImages: false,
|
|
4883
4474
|
supportsPromptCache: true,
|
|
4884
|
-
supportsNativeTools: true,
|
|
4885
|
-
defaultToolProtocol: "native",
|
|
4886
4475
|
inputPrice: 0.2,
|
|
4887
4476
|
outputPrice: 1.1,
|
|
4888
4477
|
cacheWritesPrice: 0,
|
|
@@ -4894,8 +4483,6 @@ var internationalZAiModels = {
|
|
|
4894
4483
|
contextWindow: 131072,
|
|
4895
4484
|
supportsImages: false,
|
|
4896
4485
|
supportsPromptCache: true,
|
|
4897
|
-
supportsNativeTools: true,
|
|
4898
|
-
defaultToolProtocol: "native",
|
|
4899
4486
|
inputPrice: 2.2,
|
|
4900
4487
|
outputPrice: 8.9,
|
|
4901
4488
|
cacheWritesPrice: 0,
|
|
@@ -4907,8 +4494,6 @@ var internationalZAiModels = {
|
|
|
4907
4494
|
contextWindow: 131072,
|
|
4908
4495
|
supportsImages: false,
|
|
4909
4496
|
supportsPromptCache: true,
|
|
4910
|
-
supportsNativeTools: true,
|
|
4911
|
-
defaultToolProtocol: "native",
|
|
4912
4497
|
inputPrice: 1.1,
|
|
4913
4498
|
outputPrice: 4.5,
|
|
4914
4499
|
cacheWritesPrice: 0,
|
|
@@ -4920,8 +4505,6 @@ var internationalZAiModels = {
|
|
|
4920
4505
|
contextWindow: 131072,
|
|
4921
4506
|
supportsImages: false,
|
|
4922
4507
|
supportsPromptCache: true,
|
|
4923
|
-
supportsNativeTools: true,
|
|
4924
|
-
defaultToolProtocol: "native",
|
|
4925
4508
|
inputPrice: 0,
|
|
4926
4509
|
outputPrice: 0,
|
|
4927
4510
|
cacheWritesPrice: 0,
|
|
@@ -4933,21 +4516,28 @@ var internationalZAiModels = {
|
|
|
4933
4516
|
contextWindow: 131072,
|
|
4934
4517
|
supportsImages: true,
|
|
4935
4518
|
supportsPromptCache: true,
|
|
4936
|
-
supportsNativeTools: true,
|
|
4937
|
-
defaultToolProtocol: "native",
|
|
4938
4519
|
inputPrice: 0.6,
|
|
4939
4520
|
outputPrice: 1.8,
|
|
4940
4521
|
cacheWritesPrice: 0,
|
|
4941
4522
|
cacheReadsPrice: 0.11,
|
|
4942
4523
|
description: "GLM-4.5V is Z.AI's multimodal visual reasoning model (image/video/text/file input), optimized for GUI tasks, grounding, and document/video understanding."
|
|
4943
4524
|
},
|
|
4525
|
+
"glm-4.6v": {
|
|
4526
|
+
maxTokens: 16384,
|
|
4527
|
+
contextWindow: 131072,
|
|
4528
|
+
supportsImages: true,
|
|
4529
|
+
supportsPromptCache: true,
|
|
4530
|
+
inputPrice: 0.3,
|
|
4531
|
+
outputPrice: 0.9,
|
|
4532
|
+
cacheWritesPrice: 0,
|
|
4533
|
+
cacheReadsPrice: 0.05,
|
|
4534
|
+
description: "GLM-4.6V is an advanced multimodal vision model with improved performance and cost-efficiency for visual understanding tasks."
|
|
4535
|
+
},
|
|
4944
4536
|
"glm-4.6": {
|
|
4945
4537
|
maxTokens: 16384,
|
|
4946
4538
|
contextWindow: 2e5,
|
|
4947
4539
|
supportsImages: false,
|
|
4948
4540
|
supportsPromptCache: true,
|
|
4949
|
-
supportsNativeTools: true,
|
|
4950
|
-
defaultToolProtocol: "native",
|
|
4951
4541
|
inputPrice: 0.6,
|
|
4952
4542
|
outputPrice: 2.2,
|
|
4953
4543
|
cacheWritesPrice: 0,
|
|
@@ -4959,8 +4549,6 @@ var internationalZAiModels = {
|
|
|
4959
4549
|
contextWindow: 2e5,
|
|
4960
4550
|
supportsImages: false,
|
|
4961
4551
|
supportsPromptCache: true,
|
|
4962
|
-
supportsNativeTools: true,
|
|
4963
|
-
defaultToolProtocol: "native",
|
|
4964
4552
|
supportsReasoningEffort: ["disable", "medium"],
|
|
4965
4553
|
reasoningEffort: "medium",
|
|
4966
4554
|
preserveReasoning: true,
|
|
@@ -4970,13 +4558,55 @@ var internationalZAiModels = {
|
|
|
4970
4558
|
cacheReadsPrice: 0.11,
|
|
4971
4559
|
description: "GLM-4.7 is Zhipu's latest model with built-in thinking capabilities enabled by default. It provides enhanced reasoning for complex tasks while maintaining fast response times."
|
|
4972
4560
|
},
|
|
4561
|
+
"glm-4.7-flash": {
|
|
4562
|
+
maxTokens: 16384,
|
|
4563
|
+
contextWindow: 2e5,
|
|
4564
|
+
supportsImages: false,
|
|
4565
|
+
supportsPromptCache: true,
|
|
4566
|
+
inputPrice: 0,
|
|
4567
|
+
outputPrice: 0,
|
|
4568
|
+
cacheWritesPrice: 0,
|
|
4569
|
+
cacheReadsPrice: 0,
|
|
4570
|
+
description: "GLM-4.7-Flash is a free, high-speed variant of GLM-4.7 offering fast responses for reasoning and coding tasks."
|
|
4571
|
+
},
|
|
4572
|
+
"glm-4.7-flashx": {
|
|
4573
|
+
maxTokens: 16384,
|
|
4574
|
+
contextWindow: 2e5,
|
|
4575
|
+
supportsImages: false,
|
|
4576
|
+
supportsPromptCache: true,
|
|
4577
|
+
inputPrice: 0.07,
|
|
4578
|
+
outputPrice: 0.4,
|
|
4579
|
+
cacheWritesPrice: 0,
|
|
4580
|
+
cacheReadsPrice: 0.01,
|
|
4581
|
+
description: "GLM-4.7-FlashX is an ultra-fast variant of GLM-4.7 with exceptional speed and cost-effectiveness for high-throughput applications."
|
|
4582
|
+
},
|
|
4583
|
+
"glm-4.6v-flash": {
|
|
4584
|
+
maxTokens: 16384,
|
|
4585
|
+
contextWindow: 131072,
|
|
4586
|
+
supportsImages: true,
|
|
4587
|
+
supportsPromptCache: true,
|
|
4588
|
+
inputPrice: 0,
|
|
4589
|
+
outputPrice: 0,
|
|
4590
|
+
cacheWritesPrice: 0,
|
|
4591
|
+
cacheReadsPrice: 0,
|
|
4592
|
+
description: "GLM-4.6V-Flash is a free, high-speed multimodal vision model for rapid image understanding and visual reasoning tasks."
|
|
4593
|
+
},
|
|
4594
|
+
"glm-4.6v-flashx": {
|
|
4595
|
+
maxTokens: 16384,
|
|
4596
|
+
contextWindow: 131072,
|
|
4597
|
+
supportsImages: true,
|
|
4598
|
+
supportsPromptCache: true,
|
|
4599
|
+
inputPrice: 0.04,
|
|
4600
|
+
outputPrice: 0.4,
|
|
4601
|
+
cacheWritesPrice: 0,
|
|
4602
|
+
cacheReadsPrice: 4e-3,
|
|
4603
|
+
description: "GLM-4.6V-FlashX is an ultra-fast multimodal vision model optimized for high-speed visual processing at low cost."
|
|
4604
|
+
},
|
|
4973
4605
|
"glm-4-32b-0414-128k": {
|
|
4974
4606
|
maxTokens: 16384,
|
|
4975
4607
|
contextWindow: 131072,
|
|
4976
4608
|
supportsImages: false,
|
|
4977
4609
|
supportsPromptCache: false,
|
|
4978
|
-
supportsNativeTools: true,
|
|
4979
|
-
defaultToolProtocol: "native",
|
|
4980
4610
|
inputPrice: 0.1,
|
|
4981
4611
|
outputPrice: 0.1,
|
|
4982
4612
|
cacheWritesPrice: 0,
|
|
@@ -4991,8 +4621,6 @@ var mainlandZAiModels = {
|
|
|
4991
4621
|
contextWindow: 131072,
|
|
4992
4622
|
supportsImages: false,
|
|
4993
4623
|
supportsPromptCache: true,
|
|
4994
|
-
supportsNativeTools: true,
|
|
4995
|
-
defaultToolProtocol: "native",
|
|
4996
4624
|
inputPrice: 0.29,
|
|
4997
4625
|
outputPrice: 1.14,
|
|
4998
4626
|
cacheWritesPrice: 0,
|
|
@@ -5004,8 +4632,6 @@ var mainlandZAiModels = {
|
|
|
5004
4632
|
contextWindow: 131072,
|
|
5005
4633
|
supportsImages: false,
|
|
5006
4634
|
supportsPromptCache: true,
|
|
5007
|
-
supportsNativeTools: true,
|
|
5008
|
-
defaultToolProtocol: "native",
|
|
5009
4635
|
inputPrice: 0.1,
|
|
5010
4636
|
outputPrice: 0.6,
|
|
5011
4637
|
cacheWritesPrice: 0,
|
|
@@ -5017,8 +4643,6 @@ var mainlandZAiModels = {
|
|
|
5017
4643
|
contextWindow: 131072,
|
|
5018
4644
|
supportsImages: false,
|
|
5019
4645
|
supportsPromptCache: true,
|
|
5020
|
-
supportsNativeTools: true,
|
|
5021
|
-
defaultToolProtocol: "native",
|
|
5022
4646
|
inputPrice: 0.29,
|
|
5023
4647
|
outputPrice: 1.14,
|
|
5024
4648
|
cacheWritesPrice: 0,
|
|
@@ -5030,8 +4654,6 @@ var mainlandZAiModels = {
|
|
|
5030
4654
|
contextWindow: 131072,
|
|
5031
4655
|
supportsImages: false,
|
|
5032
4656
|
supportsPromptCache: true,
|
|
5033
|
-
supportsNativeTools: true,
|
|
5034
|
-
defaultToolProtocol: "native",
|
|
5035
4657
|
inputPrice: 0.1,
|
|
5036
4658
|
outputPrice: 0.6,
|
|
5037
4659
|
cacheWritesPrice: 0,
|
|
@@ -5043,8 +4665,6 @@ var mainlandZAiModels = {
|
|
|
5043
4665
|
contextWindow: 131072,
|
|
5044
4666
|
supportsImages: false,
|
|
5045
4667
|
supportsPromptCache: true,
|
|
5046
|
-
supportsNativeTools: true,
|
|
5047
|
-
defaultToolProtocol: "native",
|
|
5048
4668
|
inputPrice: 0,
|
|
5049
4669
|
outputPrice: 0,
|
|
5050
4670
|
cacheWritesPrice: 0,
|
|
@@ -5056,8 +4676,6 @@ var mainlandZAiModels = {
|
|
|
5056
4676
|
contextWindow: 131072,
|
|
5057
4677
|
supportsImages: true,
|
|
5058
4678
|
supportsPromptCache: true,
|
|
5059
|
-
supportsNativeTools: true,
|
|
5060
|
-
defaultToolProtocol: "native",
|
|
5061
4679
|
inputPrice: 0.29,
|
|
5062
4680
|
outputPrice: 0.93,
|
|
5063
4681
|
cacheWritesPrice: 0,
|
|
@@ -5069,8 +4687,6 @@ var mainlandZAiModels = {
|
|
|
5069
4687
|
contextWindow: 204800,
|
|
5070
4688
|
supportsImages: false,
|
|
5071
4689
|
supportsPromptCache: true,
|
|
5072
|
-
supportsNativeTools: true,
|
|
5073
|
-
defaultToolProtocol: "native",
|
|
5074
4690
|
inputPrice: 0.29,
|
|
5075
4691
|
outputPrice: 1.14,
|
|
5076
4692
|
cacheWritesPrice: 0,
|
|
@@ -5082,8 +4698,6 @@ var mainlandZAiModels = {
|
|
|
5082
4698
|
contextWindow: 204800,
|
|
5083
4699
|
supportsImages: false,
|
|
5084
4700
|
supportsPromptCache: true,
|
|
5085
|
-
supportsNativeTools: true,
|
|
5086
|
-
defaultToolProtocol: "native",
|
|
5087
4701
|
supportsReasoningEffort: ["disable", "medium"],
|
|
5088
4702
|
reasoningEffort: "medium",
|
|
5089
4703
|
preserveReasoning: true,
|
|
@@ -5092,6 +4706,61 @@ var mainlandZAiModels = {
|
|
|
5092
4706
|
cacheWritesPrice: 0,
|
|
5093
4707
|
cacheReadsPrice: 0.057,
|
|
5094
4708
|
description: "GLM-4.7 is Zhipu's latest model with built-in thinking capabilities enabled by default. It provides enhanced reasoning for complex tasks while maintaining fast response times."
|
|
4709
|
+
},
|
|
4710
|
+
"glm-4.7-flash": {
|
|
4711
|
+
maxTokens: 16384,
|
|
4712
|
+
contextWindow: 204800,
|
|
4713
|
+
supportsImages: false,
|
|
4714
|
+
supportsPromptCache: true,
|
|
4715
|
+
inputPrice: 0,
|
|
4716
|
+
outputPrice: 0,
|
|
4717
|
+
cacheWritesPrice: 0,
|
|
4718
|
+
cacheReadsPrice: 0,
|
|
4719
|
+
description: "GLM-4.7-Flash is a free, high-speed variant of GLM-4.7 offering fast responses for reasoning and coding tasks."
|
|
4720
|
+
},
|
|
4721
|
+
"glm-4.7-flashx": {
|
|
4722
|
+
maxTokens: 16384,
|
|
4723
|
+
contextWindow: 204800,
|
|
4724
|
+
supportsImages: false,
|
|
4725
|
+
supportsPromptCache: true,
|
|
4726
|
+
inputPrice: 0.035,
|
|
4727
|
+
outputPrice: 0.2,
|
|
4728
|
+
cacheWritesPrice: 0,
|
|
4729
|
+
cacheReadsPrice: 5e-3,
|
|
4730
|
+
description: "GLM-4.7-FlashX is an ultra-fast variant of GLM-4.7 with exceptional speed and cost-effectiveness for high-throughput applications."
|
|
4731
|
+
},
|
|
4732
|
+
"glm-4.6v": {
|
|
4733
|
+
maxTokens: 16384,
|
|
4734
|
+
contextWindow: 131072,
|
|
4735
|
+
supportsImages: true,
|
|
4736
|
+
supportsPromptCache: true,
|
|
4737
|
+
inputPrice: 0.15,
|
|
4738
|
+
outputPrice: 0.45,
|
|
4739
|
+
cacheWritesPrice: 0,
|
|
4740
|
+
cacheReadsPrice: 0.025,
|
|
4741
|
+
description: "GLM-4.6V is an advanced multimodal vision model with improved performance and cost-efficiency for visual understanding tasks."
|
|
4742
|
+
},
|
|
4743
|
+
"glm-4.6v-flash": {
|
|
4744
|
+
maxTokens: 16384,
|
|
4745
|
+
contextWindow: 131072,
|
|
4746
|
+
supportsImages: true,
|
|
4747
|
+
supportsPromptCache: true,
|
|
4748
|
+
inputPrice: 0,
|
|
4749
|
+
outputPrice: 0,
|
|
4750
|
+
cacheWritesPrice: 0,
|
|
4751
|
+
cacheReadsPrice: 0,
|
|
4752
|
+
description: "GLM-4.6V-Flash is a free, high-speed multimodal vision model for rapid image understanding and visual reasoning tasks."
|
|
4753
|
+
},
|
|
4754
|
+
"glm-4.6v-flashx": {
|
|
4755
|
+
maxTokens: 16384,
|
|
4756
|
+
contextWindow: 131072,
|
|
4757
|
+
supportsImages: true,
|
|
4758
|
+
supportsPromptCache: true,
|
|
4759
|
+
inputPrice: 0.02,
|
|
4760
|
+
outputPrice: 0.2,
|
|
4761
|
+
cacheWritesPrice: 0,
|
|
4762
|
+
cacheReadsPrice: 2e-3,
|
|
4763
|
+
description: "GLM-4.6V-FlashX is an ultra-fast multimodal vision model optimized for high-speed visual processing at low cost."
|
|
5095
4764
|
}
|
|
5096
4765
|
};
|
|
5097
4766
|
var ZAI_DEFAULT_TEMPERATURE = 0.6;
|
|
@@ -5125,7 +4794,6 @@ var deepInfraDefaultModelInfo = {
|
|
|
5125
4794
|
contextWindow: 262144,
|
|
5126
4795
|
supportsImages: false,
|
|
5127
4796
|
supportsPromptCache: false,
|
|
5128
|
-
supportsNativeTools: true,
|
|
5129
4797
|
inputPrice: 0.3,
|
|
5130
4798
|
outputPrice: 1.2,
|
|
5131
4799
|
description: "Qwen 3 Coder 480B A35B Instruct Turbo model, 256K context."
|
|
@@ -5139,8 +4807,6 @@ var minimaxModels = {
|
|
|
5139
4807
|
contextWindow: 192e3,
|
|
5140
4808
|
supportsImages: false,
|
|
5141
4809
|
supportsPromptCache: true,
|
|
5142
|
-
supportsNativeTools: true,
|
|
5143
|
-
defaultToolProtocol: "native",
|
|
5144
4810
|
includedTools: ["search_and_replace"],
|
|
5145
4811
|
excludedTools: ["apply_diff"],
|
|
5146
4812
|
preserveReasoning: true,
|
|
@@ -5155,8 +4821,6 @@ var minimaxModels = {
|
|
|
5155
4821
|
contextWindow: 192e3,
|
|
5156
4822
|
supportsImages: false,
|
|
5157
4823
|
supportsPromptCache: true,
|
|
5158
|
-
supportsNativeTools: true,
|
|
5159
|
-
defaultToolProtocol: "native",
|
|
5160
4824
|
includedTools: ["search_and_replace"],
|
|
5161
4825
|
excludedTools: ["apply_diff"],
|
|
5162
4826
|
preserveReasoning: true,
|
|
@@ -5171,8 +4835,6 @@ var minimaxModels = {
|
|
|
5171
4835
|
contextWindow: 192e3,
|
|
5172
4836
|
supportsImages: false,
|
|
5173
4837
|
supportsPromptCache: true,
|
|
5174
|
-
supportsNativeTools: true,
|
|
5175
|
-
defaultToolProtocol: "native",
|
|
5176
4838
|
includedTools: ["search_and_replace"],
|
|
5177
4839
|
excludedTools: ["apply_diff"],
|
|
5178
4840
|
preserveReasoning: true,
|
|
@@ -5244,8 +4906,6 @@ function getProviderDefaultModelId(provider, options = { isChina: false }) {
|
|
|
5244
4906
|
return deepInfraDefaultModelId;
|
|
5245
4907
|
case "vscode-lm":
|
|
5246
4908
|
return vscodeLlmDefaultModelId;
|
|
5247
|
-
case "claude-code":
|
|
5248
|
-
return claudeCodeDefaultModelId;
|
|
5249
4909
|
case "cerebras":
|
|
5250
4910
|
return cerebrasDefaultModelId;
|
|
5251
4911
|
case "sambanova":
|
|
@@ -5303,7 +4963,6 @@ var providerNames = [
|
|
|
5303
4963
|
"bedrock",
|
|
5304
4964
|
"baseten",
|
|
5305
4965
|
"cerebras",
|
|
5306
|
-
"claude-code",
|
|
5307
4966
|
"doubao",
|
|
5308
4967
|
"deepseek",
|
|
5309
4968
|
"featherless",
|
|
@@ -5333,9 +4992,7 @@ var providerSettingsEntrySchema = z8.object({
|
|
|
5333
4992
|
});
|
|
5334
4993
|
var baseProviderSettingsSchema = z8.object({
|
|
5335
4994
|
includeMaxTokens: z8.boolean().optional(),
|
|
5336
|
-
diffEnabled: z8.boolean().optional(),
|
|
5337
4995
|
todoListEnabled: z8.boolean().optional(),
|
|
5338
|
-
fuzzyMatchThreshold: z8.number().optional(),
|
|
5339
4996
|
modelTemperature: z8.number().nullish(),
|
|
5340
4997
|
rateLimitSeconds: z8.number().optional(),
|
|
5341
4998
|
consecutiveMistakeLimit: z8.number().min(0).optional(),
|
|
@@ -5345,9 +5002,7 @@ var baseProviderSettingsSchema = z8.object({
|
|
|
5345
5002
|
modelMaxTokens: z8.number().optional(),
|
|
5346
5003
|
modelMaxThinkingTokens: z8.number().optional(),
|
|
5347
5004
|
// Model verbosity.
|
|
5348
|
-
verbosity: verbosityLevelsSchema.optional()
|
|
5349
|
-
// Tool protocol override for this profile.
|
|
5350
|
-
toolProtocol: z8.enum(["xml", "native"]).optional()
|
|
5005
|
+
verbosity: verbosityLevelsSchema.optional()
|
|
5351
5006
|
});
|
|
5352
5007
|
var apiModelIdProviderModelSchema = baseProviderSettingsSchema.extend({
|
|
5353
5008
|
apiModelId: z8.string().optional()
|
|
@@ -5359,7 +5014,6 @@ var anthropicSchema = apiModelIdProviderModelSchema.extend({
|
|
|
5359
5014
|
anthropicBeta1MContext: z8.boolean().optional()
|
|
5360
5015
|
// Enable 'context-1m-2025-08-07' beta for 1M context window.
|
|
5361
5016
|
});
|
|
5362
|
-
var claudeCodeSchema = apiModelIdProviderModelSchema.extend({});
|
|
5363
5017
|
var openRouterSchema = baseProviderSettingsSchema.extend({
|
|
5364
5018
|
openRouterApiKey: z8.string().optional(),
|
|
5365
5019
|
openRouterModelId: z8.string().optional(),
|
|
@@ -5548,7 +5202,6 @@ var defaultSchema = z8.object({
|
|
|
5548
5202
|
});
|
|
5549
5203
|
var providerSettingsSchemaDiscriminated = z8.discriminatedUnion("apiProvider", [
|
|
5550
5204
|
anthropicSchema.merge(z8.object({ apiProvider: z8.literal("anthropic") })),
|
|
5551
|
-
claudeCodeSchema.merge(z8.object({ apiProvider: z8.literal("claude-code") })),
|
|
5552
5205
|
openRouterSchema.merge(z8.object({ apiProvider: z8.literal("openrouter") })),
|
|
5553
5206
|
bedrockSchema.merge(z8.object({ apiProvider: z8.literal("bedrock") })),
|
|
5554
5207
|
vertexSchema.merge(z8.object({ apiProvider: z8.literal("vertex") })),
|
|
@@ -5589,7 +5242,6 @@ var providerSettingsSchemaDiscriminated = z8.discriminatedUnion("apiProvider", [
|
|
|
5589
5242
|
var providerSettingsSchema = z8.object({
|
|
5590
5243
|
apiProvider: providerNamesSchema.optional(),
|
|
5591
5244
|
...anthropicSchema.shape,
|
|
5592
|
-
...claudeCodeSchema.shape,
|
|
5593
5245
|
...openRouterSchema.shape,
|
|
5594
5246
|
...bedrockSchema.shape,
|
|
5595
5247
|
...vertexSchema.shape,
|
|
@@ -5654,7 +5306,6 @@ var getModelId = (settings) => {
|
|
|
5654
5306
|
var isTypicalProvider = (key) => isProviderName(key) && !isInternalProvider(key) && !isCustomProvider(key) && !isFauxProvider(key);
|
|
5655
5307
|
var modelIdKeysByProvider = {
|
|
5656
5308
|
anthropic: "apiModelId",
|
|
5657
|
-
"claude-code": "apiModelId",
|
|
5658
5309
|
openrouter: "openRouterModelId",
|
|
5659
5310
|
bedrock: "apiModelId",
|
|
5660
5311
|
vertex: "apiModelId",
|
|
@@ -5688,7 +5339,7 @@ var modelIdKeysByProvider = {
|
|
|
5688
5339
|
roo: "apiModelId",
|
|
5689
5340
|
"vercel-ai-gateway": "vercelAiGatewayModelId"
|
|
5690
5341
|
};
|
|
5691
|
-
var ANTHROPIC_STYLE_PROVIDERS = ["anthropic", "
|
|
5342
|
+
var ANTHROPIC_STYLE_PROVIDERS = ["anthropic", "bedrock", "minimax"];
|
|
5692
5343
|
var getApiProtocol = (provider, modelId) => {
|
|
5693
5344
|
if (provider && ANTHROPIC_STYLE_PROVIDERS.includes(provider)) {
|
|
5694
5345
|
return "anthropic";
|
|
@@ -5717,7 +5368,6 @@ var MODELS_BY_PROVIDER = {
|
|
|
5717
5368
|
label: "Cerebras",
|
|
5718
5369
|
models: Object.keys(cerebrasModels)
|
|
5719
5370
|
},
|
|
5720
|
-
"claude-code": { id: "claude-code", label: "Claude Code", models: Object.keys(claudeCodeModels) },
|
|
5721
5371
|
deepseek: {
|
|
5722
5372
|
id: "deepseek",
|
|
5723
5373
|
label: "DeepSeek",
|
|
@@ -5821,16 +5471,6 @@ var historyItemSchema = z9.object({
|
|
|
5821
5471
|
size: z9.number().optional(),
|
|
5822
5472
|
workspace: z9.string().optional(),
|
|
5823
5473
|
mode: z9.string().optional(),
|
|
5824
|
-
/**
|
|
5825
|
-
* The tool protocol used by this task. Once a task uses tools with a specific
|
|
5826
|
-
* protocol (XML or Native), it is permanently locked to that protocol.
|
|
5827
|
-
*
|
|
5828
|
-
* - "xml": Tool calls are parsed from XML text (no tool IDs)
|
|
5829
|
-
* - "native": Tool calls come as tool_call chunks with IDs
|
|
5830
|
-
*
|
|
5831
|
-
* This ensures task resumption works correctly even when NTC settings change.
|
|
5832
|
-
*/
|
|
5833
|
-
toolProtocol: z9.enum(["xml", "native"]).optional(),
|
|
5834
5474
|
apiConfigName: z9.string().optional(),
|
|
5835
5475
|
// Provider profile name for sticky profile feature
|
|
5836
5476
|
status: z9.enum(["active", "completed", "delegated"]).optional(),
|
|
@@ -5848,23 +5488,12 @@ var historyItemSchema = z9.object({
|
|
|
5848
5488
|
|
|
5849
5489
|
// src/experiment.ts
|
|
5850
5490
|
import { z as z10 } from "zod";
|
|
5851
|
-
var experimentIds = [
|
|
5852
|
-
"powerSteering",
|
|
5853
|
-
"multiFileApplyDiff",
|
|
5854
|
-
"preventFocusDisruption",
|
|
5855
|
-
"imageGeneration",
|
|
5856
|
-
"runSlashCommand",
|
|
5857
|
-
"multipleNativeToolCalls",
|
|
5858
|
-
"customTools"
|
|
5859
|
-
];
|
|
5491
|
+
var experimentIds = ["preventFocusDisruption", "imageGeneration", "runSlashCommand", "customTools"];
|
|
5860
5492
|
var experimentIdsSchema = z10.enum(experimentIds);
|
|
5861
5493
|
var experimentsSchema = z10.object({
|
|
5862
|
-
powerSteering: z10.boolean().optional(),
|
|
5863
|
-
multiFileApplyDiff: z10.boolean().optional(),
|
|
5864
5494
|
preventFocusDisruption: z10.boolean().optional(),
|
|
5865
5495
|
imageGeneration: z10.boolean().optional(),
|
|
5866
5496
|
runSlashCommand: z10.boolean().optional(),
|
|
5867
|
-
multipleNativeToolCalls: z10.boolean().optional(),
|
|
5868
5497
|
customTools: z10.boolean().optional()
|
|
5869
5498
|
});
|
|
5870
5499
|
|
|
@@ -5916,6 +5545,7 @@ var TelemetryEventName = /* @__PURE__ */ ((TelemetryEventName2) => {
|
|
|
5916
5545
|
TelemetryEventName2["CODE_INDEX_ERROR"] = "Code Index Error";
|
|
5917
5546
|
TelemetryEventName2["TELEMETRY_SETTINGS_CHANGED"] = "Telemetry Settings Changed";
|
|
5918
5547
|
TelemetryEventName2["MODEL_CACHE_EMPTY_RESPONSE"] = "Model Cache Empty Response";
|
|
5548
|
+
TelemetryEventName2["READ_FILE_LEGACY_FORMAT_USED"] = "Read File Legacy Format Used";
|
|
5919
5549
|
return TelemetryEventName2;
|
|
5920
5550
|
})(TelemetryEventName || {});
|
|
5921
5551
|
var staticAppPropertiesSchema = z11.object({
|
|
@@ -6004,7 +5634,8 @@ var rooCodeTelemetryEventSchema = z11.discriminatedUnion("type", [
|
|
|
6004
5634
|
"Sliding Window Truncation" /* SLIDING_WINDOW_TRUNCATION */,
|
|
6005
5635
|
"Tab Shown" /* TAB_SHOWN */,
|
|
6006
5636
|
"Mode Setting Changed" /* MODE_SETTINGS_CHANGED */,
|
|
6007
|
-
"Custom Mode Created" /* CUSTOM_MODE_CREATED
|
|
5637
|
+
"Custom Mode Created" /* CUSTOM_MODE_CREATED */,
|
|
5638
|
+
"Read File Legacy Format Used" /* READ_FILE_LEGACY_FORMAT_USED */
|
|
6008
5639
|
]),
|
|
6009
5640
|
properties: telemetryPropertiesSchema
|
|
6010
5641
|
}),
|
|
@@ -6318,7 +5949,15 @@ var isLanguage = (value) => languages.includes(value);
|
|
|
6318
5949
|
|
|
6319
5950
|
// src/global-settings.ts
|
|
6320
5951
|
var DEFAULT_WRITE_DELAY_MS = 1e3;
|
|
6321
|
-
var
|
|
5952
|
+
var TERMINAL_PREVIEW_BYTES = {
|
|
5953
|
+
small: 5 * 1024,
|
|
5954
|
+
// 5KB
|
|
5955
|
+
medium: 10 * 1024,
|
|
5956
|
+
// 10KB
|
|
5957
|
+
large: 20 * 1024
|
|
5958
|
+
// 20KB
|
|
5959
|
+
};
|
|
5960
|
+
var DEFAULT_TERMINAL_OUTPUT_PREVIEW_SIZE = "medium";
|
|
6322
5961
|
var MIN_CHECKPOINT_TIMEOUT_SECONDS = 10;
|
|
6323
5962
|
var MAX_CHECKPOINT_TIMEOUT_SECONDS = 60;
|
|
6324
5963
|
var DEFAULT_CHECKPOINT_TIMEOUT_SECONDS = 15;
|
|
@@ -6334,7 +5973,6 @@ var globalSettingsSchema = z14.object({
|
|
|
6334
5973
|
imageGenerationProvider: z14.enum(["openrouter", "roo"]).optional(),
|
|
6335
5974
|
openRouterImageApiKey: z14.string().optional(),
|
|
6336
5975
|
openRouterImageGenerationSelectedModel: z14.string().optional(),
|
|
6337
|
-
condensingApiConfigId: z14.string().optional(),
|
|
6338
5976
|
customCondensingPrompt: z14.string().optional(),
|
|
6339
5977
|
autoApprovalEnabled: z14.boolean().optional(),
|
|
6340
5978
|
alwaysAllowReadOnly: z14.boolean().optional(),
|
|
@@ -6360,7 +5998,6 @@ var globalSettingsSchema = z14.object({
|
|
|
6360
5998
|
allowedMaxCost: z14.number().nullish(),
|
|
6361
5999
|
autoCondenseContext: z14.boolean().optional(),
|
|
6362
6000
|
autoCondenseContextPercent: z14.number().optional(),
|
|
6363
|
-
maxConcurrentFileReads: z14.number().optional(),
|
|
6364
6001
|
/**
|
|
6365
6002
|
* Whether to include current time in the environment details
|
|
6366
6003
|
* @default true
|
|
@@ -6403,11 +6040,9 @@ var globalSettingsSchema = z14.object({
|
|
|
6403
6040
|
maxWorkspaceFiles: z14.number().optional(),
|
|
6404
6041
|
showRooIgnoredFiles: z14.boolean().optional(),
|
|
6405
6042
|
enableSubfolderRules: z14.boolean().optional(),
|
|
6406
|
-
maxReadFileLine: z14.number().optional(),
|
|
6407
6043
|
maxImageFileSize: z14.number().optional(),
|
|
6408
6044
|
maxTotalImageSize: z14.number().optional(),
|
|
6409
|
-
|
|
6410
|
-
terminalOutputCharacterLimit: z14.number().optional(),
|
|
6045
|
+
terminalOutputPreviewSize: z14.enum(["small", "medium", "large"]).optional(),
|
|
6411
6046
|
terminalShellIntegrationTimeout: z14.number().optional(),
|
|
6412
6047
|
terminalShellIntegrationDisabled: z14.boolean().optional(),
|
|
6413
6048
|
terminalCommandDelay: z14.number().optional(),
|
|
@@ -6416,18 +6051,14 @@ var globalSettingsSchema = z14.object({
|
|
|
6416
6051
|
terminalZshOhMy: z14.boolean().optional(),
|
|
6417
6052
|
terminalZshP10k: z14.boolean().optional(),
|
|
6418
6053
|
terminalZdotdir: z14.boolean().optional(),
|
|
6419
|
-
terminalCompressProgressBar: z14.boolean().optional(),
|
|
6420
6054
|
diagnosticsEnabled: z14.boolean().optional(),
|
|
6421
6055
|
rateLimitSeconds: z14.number().optional(),
|
|
6422
|
-
diffEnabled: z14.boolean().optional(),
|
|
6423
|
-
fuzzyMatchThreshold: z14.number().optional(),
|
|
6424
6056
|
experiments: experimentsSchema.optional(),
|
|
6425
6057
|
codebaseIndexModels: codebaseIndexModelsSchema.optional(),
|
|
6426
6058
|
codebaseIndexConfig: codebaseIndexConfigSchema.optional(),
|
|
6427
6059
|
language: languagesSchema.optional(),
|
|
6428
6060
|
telemetrySetting: telemetrySettingsSchema.optional(),
|
|
6429
6061
|
mcpEnabled: z14.boolean().optional(),
|
|
6430
|
-
enableMcpServerCreation: z14.boolean().optional(),
|
|
6431
6062
|
mode: z14.string().optional(),
|
|
6432
6063
|
modeApiConfigs: z14.record(z14.string(), z14.string()).optional(),
|
|
6433
6064
|
customModes: z14.array(modeConfigSchema).optional(),
|
|
@@ -6447,7 +6078,20 @@ var globalSettingsSchema = z14.object({
|
|
|
6447
6078
|
profileThresholds: z14.record(z14.string(), z14.number()).optional(),
|
|
6448
6079
|
hasOpenedModeSelector: z14.boolean().optional(),
|
|
6449
6080
|
lastModeExportPath: z14.string().optional(),
|
|
6450
|
-
lastModeImportPath: z14.string().optional()
|
|
6081
|
+
lastModeImportPath: z14.string().optional(),
|
|
6082
|
+
lastSettingsExportPath: z14.string().optional(),
|
|
6083
|
+
lastTaskExportPath: z14.string().optional(),
|
|
6084
|
+
lastImageSavePath: z14.string().optional(),
|
|
6085
|
+
/**
|
|
6086
|
+
* Path to worktree to auto-open after switching workspaces.
|
|
6087
|
+
* Used by the worktree feature to open the Roo Code sidebar in a new window.
|
|
6088
|
+
*/
|
|
6089
|
+
worktreeAutoOpenPath: z14.string().optional(),
|
|
6090
|
+
/**
|
|
6091
|
+
* Whether to show the worktree selector in the home screen.
|
|
6092
|
+
* @default true
|
|
6093
|
+
*/
|
|
6094
|
+
showWorktreesInHomeScreen: z14.boolean().optional()
|
|
6451
6095
|
});
|
|
6452
6096
|
var GLOBAL_SETTINGS_KEYS = globalSettingsSchema.keyof().options;
|
|
6453
6097
|
var rooCodeSettingsSchema = providerSettingsSchema.merge(globalSettingsSchema);
|
|
@@ -6531,8 +6175,6 @@ var EVALS_SETTINGS = {
|
|
|
6531
6175
|
ttsSpeed: 1,
|
|
6532
6176
|
soundEnabled: false,
|
|
6533
6177
|
soundVolume: 0.5,
|
|
6534
|
-
terminalOutputLineLimit: 500,
|
|
6535
|
-
terminalOutputCharacterLimit: DEFAULT_TERMINAL_OUTPUT_CHARACTER_LIMIT,
|
|
6536
6178
|
terminalShellIntegrationTimeout: 3e4,
|
|
6537
6179
|
terminalCommandDelay: 0,
|
|
6538
6180
|
terminalPowershellCounter: false,
|
|
@@ -6540,19 +6182,14 @@ var EVALS_SETTINGS = {
|
|
|
6540
6182
|
terminalZshClearEolMark: true,
|
|
6541
6183
|
terminalZshP10k: false,
|
|
6542
6184
|
terminalZdotdir: true,
|
|
6543
|
-
terminalCompressProgressBar: true,
|
|
6544
6185
|
terminalShellIntegrationDisabled: true,
|
|
6545
6186
|
diagnosticsEnabled: true,
|
|
6546
|
-
diffEnabled: true,
|
|
6547
|
-
fuzzyMatchThreshold: 1,
|
|
6548
6187
|
enableCheckpoints: false,
|
|
6549
6188
|
rateLimitSeconds: 0,
|
|
6550
6189
|
maxOpenTabsContext: 20,
|
|
6551
6190
|
maxWorkspaceFiles: 200,
|
|
6552
6191
|
maxGitStatusFiles: 20,
|
|
6553
6192
|
showRooIgnoredFiles: true,
|
|
6554
|
-
maxReadFileLine: -1,
|
|
6555
|
-
// -1 to enable full file reading.
|
|
6556
6193
|
includeDiagnosticMessages: true,
|
|
6557
6194
|
maxDiagnosticMessages: 50,
|
|
6558
6195
|
language: "en",
|
|
@@ -6626,24 +6263,18 @@ var organizationAllowListSchema = z16.object({
|
|
|
6626
6263
|
});
|
|
6627
6264
|
var organizationDefaultSettingsSchema = globalSettingsSchema.pick({
|
|
6628
6265
|
enableCheckpoints: true,
|
|
6629
|
-
fuzzyMatchThreshold: true,
|
|
6630
6266
|
maxOpenTabsContext: true,
|
|
6631
|
-
maxReadFileLine: true,
|
|
6632
6267
|
maxWorkspaceFiles: true,
|
|
6633
6268
|
showRooIgnoredFiles: true,
|
|
6634
6269
|
terminalCommandDelay: true,
|
|
6635
|
-
terminalCompressProgressBar: true,
|
|
6636
|
-
terminalOutputLineLimit: true,
|
|
6637
6270
|
terminalShellIntegrationDisabled: true,
|
|
6638
6271
|
terminalShellIntegrationTimeout: true,
|
|
6639
6272
|
terminalZshClearEolMark: true
|
|
6640
6273
|
}).merge(
|
|
6641
6274
|
z16.object({
|
|
6642
6275
|
maxOpenTabsContext: z16.number().int().nonnegative().optional(),
|
|
6643
|
-
maxReadFileLine: z16.number().int().gte(-1).optional(),
|
|
6644
6276
|
maxWorkspaceFiles: z16.number().int().nonnegative().optional(),
|
|
6645
6277
|
terminalCommandDelay: z16.number().int().nonnegative().optional(),
|
|
6646
|
-
terminalOutputLineLimit: z16.number().int().nonnegative().optional(),
|
|
6647
6278
|
terminalShellIntegrationTimeout: z16.number().int().nonnegative().optional()
|
|
6648
6279
|
})
|
|
6649
6280
|
);
|
|
@@ -7118,12 +6749,10 @@ var taskCommandSchema = z19.discriminatedUnion("commandName", [
|
|
|
7118
6749
|
})
|
|
7119
6750
|
}),
|
|
7120
6751
|
z19.object({
|
|
7121
|
-
commandName: z19.literal("CancelTask" /* CancelTask */)
|
|
7122
|
-
data: z19.string()
|
|
6752
|
+
commandName: z19.literal("CancelTask" /* CancelTask */)
|
|
7123
6753
|
}),
|
|
7124
6754
|
z19.object({
|
|
7125
|
-
commandName: z19.literal("CloseTask" /* CloseTask */)
|
|
7126
|
-
data: z19.string()
|
|
6755
|
+
commandName: z19.literal("CloseTask" /* CloseTask */)
|
|
7127
6756
|
}),
|
|
7128
6757
|
z19.object({
|
|
7129
6758
|
commandName: z19.literal("ResumeTask" /* ResumeTask */),
|
|
@@ -7159,6 +6788,7 @@ var ipcMessageSchema = z19.discriminatedUnion("type", [
|
|
|
7159
6788
|
|
|
7160
6789
|
// src/mcp.ts
|
|
7161
6790
|
import { z as z20 } from "zod";
|
|
6791
|
+
var MAX_MCP_TOOLS_THRESHOLD = 60;
|
|
7162
6792
|
var mcpExecutionStatusSchema = z20.discriminatedUnion("status", [
|
|
7163
6793
|
z20.object({
|
|
7164
6794
|
executionId: z20.string(),
|
|
@@ -7182,6 +6812,46 @@ var mcpExecutionStatusSchema = z20.discriminatedUnion("status", [
|
|
|
7182
6812
|
error: z20.string().optional()
|
|
7183
6813
|
})
|
|
7184
6814
|
]);
|
|
6815
|
+
function countEnabledMcpTools(servers) {
|
|
6816
|
+
let serverCount = 0;
|
|
6817
|
+
let toolCount = 0;
|
|
6818
|
+
for (const server of servers) {
|
|
6819
|
+
if (server.disabled) continue;
|
|
6820
|
+
if (server.status !== "connected") continue;
|
|
6821
|
+
serverCount++;
|
|
6822
|
+
if (server.tools) {
|
|
6823
|
+
for (const tool of server.tools) {
|
|
6824
|
+
if (tool.enabledForPrompt !== false) {
|
|
6825
|
+
toolCount++;
|
|
6826
|
+
}
|
|
6827
|
+
}
|
|
6828
|
+
}
|
|
6829
|
+
}
|
|
6830
|
+
return { enabledToolCount: toolCount, enabledServerCount: serverCount };
|
|
6831
|
+
}
|
|
6832
|
+
|
|
6833
|
+
// src/skills.ts
|
|
6834
|
+
var SKILL_NAME_MIN_LENGTH = 1;
|
|
6835
|
+
var SKILL_NAME_MAX_LENGTH = 64;
|
|
6836
|
+
var SKILL_NAME_REGEX = /^[a-z0-9]+(?:-[a-z0-9]+)*$/;
|
|
6837
|
+
var SkillNameValidationError = /* @__PURE__ */ ((SkillNameValidationError2) => {
|
|
6838
|
+
SkillNameValidationError2["Empty"] = "empty";
|
|
6839
|
+
SkillNameValidationError2["TooLong"] = "too_long";
|
|
6840
|
+
SkillNameValidationError2["InvalidFormat"] = "invalid_format";
|
|
6841
|
+
return SkillNameValidationError2;
|
|
6842
|
+
})(SkillNameValidationError || {});
|
|
6843
|
+
function validateSkillName(name) {
|
|
6844
|
+
if (!name || name.length < SKILL_NAME_MIN_LENGTH) {
|
|
6845
|
+
return { valid: false, error: "empty" /* Empty */ };
|
|
6846
|
+
}
|
|
6847
|
+
if (name.length > SKILL_NAME_MAX_LENGTH) {
|
|
6848
|
+
return { valid: false, error: "too_long" /* TooLong */ };
|
|
6849
|
+
}
|
|
6850
|
+
if (!SKILL_NAME_REGEX.test(name)) {
|
|
6851
|
+
return { valid: false, error: "invalid_format" /* InvalidFormat */ };
|
|
6852
|
+
}
|
|
6853
|
+
return { valid: true };
|
|
6854
|
+
}
|
|
7185
6855
|
|
|
7186
6856
|
// src/todo.ts
|
|
7187
6857
|
import { z as z21 } from "zod";
|
|
@@ -7221,6 +6891,11 @@ var commandExecutionStatusSchema = z22.discriminatedUnion("status", [
|
|
|
7221
6891
|
})
|
|
7222
6892
|
]);
|
|
7223
6893
|
|
|
6894
|
+
// src/tool-params.ts
|
|
6895
|
+
function isLegacyReadFileParams(params) {
|
|
6896
|
+
return "_legacyFormat" in params && params._legacyFormat === true;
|
|
6897
|
+
}
|
|
6898
|
+
|
|
7224
6899
|
// src/vscode-extension-host.ts
|
|
7225
6900
|
import { z as z23 } from "zod";
|
|
7226
6901
|
var checkoutDiffPayloadSchema = z23.object({
|
|
@@ -7273,7 +6948,7 @@ export {
|
|
|
7273
6948
|
DEFAULT_CHECKPOINT_TIMEOUT_SECONDS,
|
|
7274
6949
|
DEFAULT_CONSECUTIVE_MISTAKE_LIMIT,
|
|
7275
6950
|
DEFAULT_MODES,
|
|
7276
|
-
|
|
6951
|
+
DEFAULT_TERMINAL_OUTPUT_PREVIEW_SIZE,
|
|
7277
6952
|
DEFAULT_WRITE_DELAY_MS,
|
|
7278
6953
|
DOUBAO_API_BASE_URL,
|
|
7279
6954
|
DOUBAO_API_CHAT_PATH,
|
|
@@ -7303,13 +6978,13 @@ export {
|
|
|
7303
6978
|
IpcOrigin,
|
|
7304
6979
|
LMSTUDIO_DEFAULT_TEMPERATURE,
|
|
7305
6980
|
MAX_CHECKPOINT_TIMEOUT_SECONDS,
|
|
6981
|
+
MAX_MCP_TOOLS_THRESHOLD,
|
|
7306
6982
|
MINIMAX_DEFAULT_MAX_TOKENS,
|
|
7307
6983
|
MINIMAX_DEFAULT_TEMPERATURE,
|
|
7308
6984
|
MIN_CHECKPOINT_TIMEOUT_SECONDS,
|
|
7309
6985
|
MISTRAL_DEFAULT_TEMPERATURE,
|
|
7310
6986
|
MODELS_BY_PROVIDER,
|
|
7311
6987
|
MOONSHOT_DEFAULT_TEMPERATURE,
|
|
7312
|
-
NATIVE_TOOL_DEFAULTS,
|
|
7313
6988
|
OPENAI_AZURE_AI_INFERENCE_PATH,
|
|
7314
6989
|
OPENAI_NATIVE_DEFAULT_TEMPERATURE,
|
|
7315
6990
|
OPENROUTER_DEFAULT_PROVIDER_NAME,
|
|
@@ -7324,7 +6999,11 @@ export {
|
|
|
7324
6999
|
RooModelsResponseSchema,
|
|
7325
7000
|
RooPricingSchema,
|
|
7326
7001
|
SECRET_STATE_KEYS,
|
|
7327
|
-
|
|
7002
|
+
SKILL_NAME_MAX_LENGTH,
|
|
7003
|
+
SKILL_NAME_MIN_LENGTH,
|
|
7004
|
+
SKILL_NAME_REGEX,
|
|
7005
|
+
SkillNameValidationError,
|
|
7006
|
+
TERMINAL_PREVIEW_BYTES,
|
|
7328
7007
|
TaskBridgeCommandName,
|
|
7329
7008
|
TaskBridgeEventName,
|
|
7330
7009
|
TaskCommandName,
|
|
@@ -7356,9 +7035,6 @@ export {
|
|
|
7356
7035
|
chutesDefaultModelId,
|
|
7357
7036
|
chutesDefaultModelInfo,
|
|
7358
7037
|
chutesModels,
|
|
7359
|
-
claudeCodeDefaultModelId,
|
|
7360
|
-
claudeCodeModels,
|
|
7361
|
-
claudeCodeReasoningConfig,
|
|
7362
7038
|
clineAskSchema,
|
|
7363
7039
|
clineAsks,
|
|
7364
7040
|
clineMessageSchema,
|
|
@@ -7373,6 +7049,7 @@ export {
|
|
|
7373
7049
|
commandIds,
|
|
7374
7050
|
contextCondenseSchema,
|
|
7375
7051
|
contextTruncationSchema,
|
|
7052
|
+
countEnabledMcpTools,
|
|
7376
7053
|
customModePromptsSchema,
|
|
7377
7054
|
customModesSettingsSchema,
|
|
7378
7055
|
customProviders,
|
|
@@ -7406,7 +7083,6 @@ export {
|
|
|
7406
7083
|
geminiDefaultModelId,
|
|
7407
7084
|
geminiModels,
|
|
7408
7085
|
getApiProtocol,
|
|
7409
|
-
getEffectiveProtocol,
|
|
7410
7086
|
getErrorMessage,
|
|
7411
7087
|
getErrorStatusCode,
|
|
7412
7088
|
getImageGenerationProvider,
|
|
@@ -7441,9 +7117,9 @@ export {
|
|
|
7441
7117
|
isInteractiveAsk,
|
|
7442
7118
|
isInternalProvider,
|
|
7443
7119
|
isLanguage,
|
|
7120
|
+
isLegacyReadFileParams,
|
|
7444
7121
|
isLocalProvider,
|
|
7445
7122
|
isModelParameter,
|
|
7446
|
-
isNativeProtocol,
|
|
7447
7123
|
isNonBlockingAsk,
|
|
7448
7124
|
isProviderName,
|
|
7449
7125
|
isResumableAsk,
|
|
@@ -7479,7 +7155,6 @@ export {
|
|
|
7479
7155
|
moonshotDefaultModelId,
|
|
7480
7156
|
moonshotModels,
|
|
7481
7157
|
nonBlockingAsks,
|
|
7482
|
-
normalizeClaudeCodeModelId,
|
|
7483
7158
|
ollamaDefaultModelId,
|
|
7484
7159
|
ollamaDefaultModelInfo,
|
|
7485
7160
|
openAiCodexDefaultModelId,
|
|
@@ -7553,6 +7228,7 @@ export {
|
|
|
7553
7228
|
userFeaturesSchema,
|
|
7554
7229
|
userSettingsConfigSchema,
|
|
7555
7230
|
userSettingsDataSchema,
|
|
7231
|
+
validateSkillName,
|
|
7556
7232
|
verbosityLevels,
|
|
7557
7233
|
verbosityLevelsSchema,
|
|
7558
7234
|
vercelAiGatewayDefaultModelId,
|