@roo-code/types 1.105.0 → 1.107.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/index.cjs +494 -709
- package/dist/index.cjs.map +1 -1
- package/dist/index.d.cts +1746 -2518
- package/dist/index.d.ts +1746 -2518
- package/dist/index.js +482 -700
- package/dist/index.js.map +1 -1
- package/package.json +2 -2
package/dist/index.js
CHANGED
|
@@ -78,7 +78,9 @@ var clineSays = [
|
|
|
78
78
|
"condense_context_error",
|
|
79
79
|
"sliding_window_truncation",
|
|
80
80
|
"codebase_search_result",
|
|
81
|
-
"user_edit_todos"
|
|
81
|
+
"user_edit_todos",
|
|
82
|
+
"too_many_tools_warning",
|
|
83
|
+
"tool"
|
|
82
84
|
];
|
|
83
85
|
var clineSaySchema = z.enum(clineSays);
|
|
84
86
|
var toolProgressStatusSchema = z.object({
|
|
@@ -146,6 +148,7 @@ var toolGroupsSchema = z2.enum(toolGroups);
|
|
|
146
148
|
var toolNames = [
|
|
147
149
|
"execute_command",
|
|
148
150
|
"read_file",
|
|
151
|
+
"read_command_output",
|
|
149
152
|
"write_to_file",
|
|
150
153
|
"apply_diff",
|
|
151
154
|
"search_and_replace",
|
|
@@ -161,10 +164,10 @@ var toolNames = [
|
|
|
161
164
|
"attempt_completion",
|
|
162
165
|
"switch_mode",
|
|
163
166
|
"new_task",
|
|
164
|
-
"fetch_instructions",
|
|
165
167
|
"codebase_search",
|
|
166
168
|
"update_todo_list",
|
|
167
169
|
"run_slash_command",
|
|
170
|
+
"skill",
|
|
168
171
|
"generate_image",
|
|
169
172
|
"custom_tool"
|
|
170
173
|
];
|
|
@@ -176,20 +179,6 @@ var toolUsageSchema = z2.record(
|
|
|
176
179
|
failures: z2.number()
|
|
177
180
|
})
|
|
178
181
|
);
|
|
179
|
-
var TOOL_PROTOCOL = {
|
|
180
|
-
XML: "xml",
|
|
181
|
-
NATIVE: "native"
|
|
182
|
-
};
|
|
183
|
-
var NATIVE_TOOL_DEFAULTS = {
|
|
184
|
-
supportsNativeTools: true,
|
|
185
|
-
defaultToolProtocol: TOOL_PROTOCOL.NATIVE
|
|
186
|
-
};
|
|
187
|
-
function isNativeProtocol(protocol) {
|
|
188
|
-
return protocol === TOOL_PROTOCOL.NATIVE;
|
|
189
|
-
}
|
|
190
|
-
function getEffectiveProtocol(toolProtocol) {
|
|
191
|
-
return toolProtocol || TOOL_PROTOCOL.XML;
|
|
192
|
-
}
|
|
193
182
|
|
|
194
183
|
// src/events.ts
|
|
195
184
|
var RooCodeEventName = /* @__PURE__ */ ((RooCodeEventName2) => {
|
|
@@ -213,6 +202,7 @@ var RooCodeEventName = /* @__PURE__ */ ((RooCodeEventName2) => {
|
|
|
213
202
|
RooCodeEventName2["TaskModeSwitched"] = "taskModeSwitched";
|
|
214
203
|
RooCodeEventName2["TaskAskResponded"] = "taskAskResponded";
|
|
215
204
|
RooCodeEventName2["TaskUserMessage"] = "taskUserMessage";
|
|
205
|
+
RooCodeEventName2["QueuedMessagesUpdated"] = "queuedMessagesUpdated";
|
|
216
206
|
RooCodeEventName2["TaskTokenUsageUpdated"] = "taskTokenUsageUpdated";
|
|
217
207
|
RooCodeEventName2["TaskToolFailed"] = "taskToolFailed";
|
|
218
208
|
RooCodeEventName2["ModeChanged"] = "modeChanged";
|
|
@@ -272,6 +262,7 @@ var rooCodeEventsSchema = z3.object({
|
|
|
272
262
|
["taskModeSwitched" /* TaskModeSwitched */]: z3.tuple([z3.string(), z3.string()]),
|
|
273
263
|
["taskAskResponded" /* TaskAskResponded */]: z3.tuple([z3.string()]),
|
|
274
264
|
["taskUserMessage" /* TaskUserMessage */]: z3.tuple([z3.string()]),
|
|
265
|
+
["queuedMessagesUpdated" /* QueuedMessagesUpdated */]: z3.tuple([z3.string(), z3.array(queuedMessageSchema)]),
|
|
275
266
|
["taskToolFailed" /* TaskToolFailed */]: z3.tuple([z3.string(), toolNamesSchema, z3.string()]),
|
|
276
267
|
["taskTokenUsageUpdated" /* TaskTokenUsageUpdated */]: z3.tuple([z3.string(), tokenUsageSchema, toolUsageSchema]),
|
|
277
268
|
["modeChanged" /* ModeChanged */]: z3.tuple([z3.string()]),
|
|
@@ -377,6 +368,11 @@ var taskEventSchema = z3.discriminatedUnion("eventName", [
|
|
|
377
368
|
payload: rooCodeEventsSchema.shape["taskAskResponded" /* TaskAskResponded */],
|
|
378
369
|
taskId: z3.number().optional()
|
|
379
370
|
}),
|
|
371
|
+
z3.object({
|
|
372
|
+
eventName: z3.literal("queuedMessagesUpdated" /* QueuedMessagesUpdated */),
|
|
373
|
+
payload: rooCodeEventsSchema.shape["queuedMessagesUpdated" /* QueuedMessagesUpdated */],
|
|
374
|
+
taskId: z3.number().optional()
|
|
375
|
+
}),
|
|
380
376
|
// Task Analytics
|
|
381
377
|
z3.object({
|
|
382
378
|
eventName: z3.literal("taskToolFailed" /* TaskToolFailed */),
|
|
@@ -477,10 +473,6 @@ var modelInfoSchema = z5.object({
|
|
|
477
473
|
isStealthModel: z5.boolean().optional(),
|
|
478
474
|
// Flag to indicate if the model is free (no cost)
|
|
479
475
|
isFree: z5.boolean().optional(),
|
|
480
|
-
// Flag to indicate if the model supports native tool calling (OpenAI-style function calling)
|
|
481
|
-
supportsNativeTools: z5.boolean().optional(),
|
|
482
|
-
// Default tool protocol preferred by this model (if not specified, falls back to capability/provider defaults)
|
|
483
|
-
defaultToolProtocol: z5.enum(["xml", "native"]).optional(),
|
|
484
476
|
// Exclude specific native tools from being available (only applies to native protocol)
|
|
485
477
|
// These tools will be removed from the set of tools available to the model
|
|
486
478
|
excludedTools: z5.array(z5.string()).optional(),
|
|
@@ -577,8 +569,6 @@ var anthropicModels = {
|
|
|
577
569
|
// Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
|
|
578
570
|
supportsImages: true,
|
|
579
571
|
supportsPromptCache: true,
|
|
580
|
-
supportsNativeTools: true,
|
|
581
|
-
defaultToolProtocol: "native",
|
|
582
572
|
inputPrice: 3,
|
|
583
573
|
// $3 per million input tokens (≤200K context)
|
|
584
574
|
outputPrice: 15,
|
|
@@ -611,8 +601,6 @@ var anthropicModels = {
|
|
|
611
601
|
// Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
|
|
612
602
|
supportsImages: true,
|
|
613
603
|
supportsPromptCache: true,
|
|
614
|
-
supportsNativeTools: true,
|
|
615
|
-
defaultToolProtocol: "native",
|
|
616
604
|
inputPrice: 3,
|
|
617
605
|
// $3 per million input tokens (≤200K context)
|
|
618
606
|
outputPrice: 15,
|
|
@@ -644,8 +632,6 @@ var anthropicModels = {
|
|
|
644
632
|
contextWindow: 2e5,
|
|
645
633
|
supportsImages: true,
|
|
646
634
|
supportsPromptCache: true,
|
|
647
|
-
supportsNativeTools: true,
|
|
648
|
-
defaultToolProtocol: "native",
|
|
649
635
|
inputPrice: 5,
|
|
650
636
|
// $5 per million input tokens
|
|
651
637
|
outputPrice: 25,
|
|
@@ -662,8 +648,6 @@ var anthropicModels = {
|
|
|
662
648
|
contextWindow: 2e5,
|
|
663
649
|
supportsImages: true,
|
|
664
650
|
supportsPromptCache: true,
|
|
665
|
-
supportsNativeTools: true,
|
|
666
|
-
defaultToolProtocol: "native",
|
|
667
651
|
inputPrice: 15,
|
|
668
652
|
// $15 per million input tokens
|
|
669
653
|
outputPrice: 75,
|
|
@@ -680,8 +664,6 @@ var anthropicModels = {
|
|
|
680
664
|
contextWindow: 2e5,
|
|
681
665
|
supportsImages: true,
|
|
682
666
|
supportsPromptCache: true,
|
|
683
|
-
supportsNativeTools: true,
|
|
684
|
-
defaultToolProtocol: "native",
|
|
685
667
|
inputPrice: 15,
|
|
686
668
|
// $15 per million input tokens
|
|
687
669
|
outputPrice: 75,
|
|
@@ -698,8 +680,6 @@ var anthropicModels = {
|
|
|
698
680
|
contextWindow: 2e5,
|
|
699
681
|
supportsImages: true,
|
|
700
682
|
supportsPromptCache: true,
|
|
701
|
-
supportsNativeTools: true,
|
|
702
|
-
defaultToolProtocol: "native",
|
|
703
683
|
inputPrice: 3,
|
|
704
684
|
// $3 per million input tokens
|
|
705
685
|
outputPrice: 15,
|
|
@@ -717,8 +697,6 @@ var anthropicModels = {
|
|
|
717
697
|
contextWindow: 2e5,
|
|
718
698
|
supportsImages: true,
|
|
719
699
|
supportsPromptCache: true,
|
|
720
|
-
supportsNativeTools: true,
|
|
721
|
-
defaultToolProtocol: "native",
|
|
722
700
|
inputPrice: 3,
|
|
723
701
|
// $3 per million input tokens
|
|
724
702
|
outputPrice: 15,
|
|
@@ -733,8 +711,6 @@ var anthropicModels = {
|
|
|
733
711
|
contextWindow: 2e5,
|
|
734
712
|
supportsImages: true,
|
|
735
713
|
supportsPromptCache: true,
|
|
736
|
-
supportsNativeTools: true,
|
|
737
|
-
defaultToolProtocol: "native",
|
|
738
714
|
inputPrice: 3,
|
|
739
715
|
// $3 per million input tokens
|
|
740
716
|
outputPrice: 15,
|
|
@@ -749,8 +725,6 @@ var anthropicModels = {
|
|
|
749
725
|
contextWindow: 2e5,
|
|
750
726
|
supportsImages: false,
|
|
751
727
|
supportsPromptCache: true,
|
|
752
|
-
supportsNativeTools: true,
|
|
753
|
-
defaultToolProtocol: "native",
|
|
754
728
|
inputPrice: 1,
|
|
755
729
|
outputPrice: 5,
|
|
756
730
|
cacheWritesPrice: 1.25,
|
|
@@ -761,8 +735,6 @@ var anthropicModels = {
|
|
|
761
735
|
contextWindow: 2e5,
|
|
762
736
|
supportsImages: true,
|
|
763
737
|
supportsPromptCache: true,
|
|
764
|
-
supportsNativeTools: true,
|
|
765
|
-
defaultToolProtocol: "native",
|
|
766
738
|
inputPrice: 15,
|
|
767
739
|
outputPrice: 75,
|
|
768
740
|
cacheWritesPrice: 18.75,
|
|
@@ -773,8 +745,6 @@ var anthropicModels = {
|
|
|
773
745
|
contextWindow: 2e5,
|
|
774
746
|
supportsImages: true,
|
|
775
747
|
supportsPromptCache: true,
|
|
776
|
-
supportsNativeTools: true,
|
|
777
|
-
defaultToolProtocol: "native",
|
|
778
748
|
inputPrice: 0.25,
|
|
779
749
|
outputPrice: 1.25,
|
|
780
750
|
cacheWritesPrice: 0.3,
|
|
@@ -785,8 +755,6 @@ var anthropicModels = {
|
|
|
785
755
|
contextWindow: 2e5,
|
|
786
756
|
supportsImages: true,
|
|
787
757
|
supportsPromptCache: true,
|
|
788
|
-
supportsNativeTools: true,
|
|
789
|
-
defaultToolProtocol: "native",
|
|
790
758
|
inputPrice: 1,
|
|
791
759
|
outputPrice: 5,
|
|
792
760
|
cacheWritesPrice: 1.25,
|
|
@@ -804,7 +772,6 @@ var basetenModels = {
|
|
|
804
772
|
contextWindow: 262e3,
|
|
805
773
|
supportsImages: false,
|
|
806
774
|
supportsPromptCache: false,
|
|
807
|
-
supportsNativeTools: true,
|
|
808
775
|
inputPrice: 0.6,
|
|
809
776
|
outputPrice: 2.5,
|
|
810
777
|
cacheWritesPrice: 0,
|
|
@@ -816,7 +783,6 @@ var basetenModels = {
|
|
|
816
783
|
contextWindow: 2e5,
|
|
817
784
|
supportsImages: false,
|
|
818
785
|
supportsPromptCache: false,
|
|
819
|
-
supportsNativeTools: true,
|
|
820
786
|
inputPrice: 0.6,
|
|
821
787
|
outputPrice: 2.2,
|
|
822
788
|
cacheWritesPrice: 0,
|
|
@@ -828,7 +794,6 @@ var basetenModels = {
|
|
|
828
794
|
contextWindow: 163840,
|
|
829
795
|
supportsImages: false,
|
|
830
796
|
supportsPromptCache: false,
|
|
831
|
-
supportsNativeTools: true,
|
|
832
797
|
inputPrice: 2.55,
|
|
833
798
|
outputPrice: 5.95,
|
|
834
799
|
cacheWritesPrice: 0,
|
|
@@ -840,7 +805,6 @@ var basetenModels = {
|
|
|
840
805
|
contextWindow: 163840,
|
|
841
806
|
supportsImages: false,
|
|
842
807
|
supportsPromptCache: false,
|
|
843
|
-
supportsNativeTools: true,
|
|
844
808
|
inputPrice: 2.55,
|
|
845
809
|
outputPrice: 5.95,
|
|
846
810
|
cacheWritesPrice: 0,
|
|
@@ -852,7 +816,6 @@ var basetenModels = {
|
|
|
852
816
|
contextWindow: 163840,
|
|
853
817
|
supportsImages: false,
|
|
854
818
|
supportsPromptCache: false,
|
|
855
|
-
supportsNativeTools: true,
|
|
856
819
|
inputPrice: 0.77,
|
|
857
820
|
outputPrice: 0.77,
|
|
858
821
|
cacheWritesPrice: 0,
|
|
@@ -864,7 +827,6 @@ var basetenModels = {
|
|
|
864
827
|
contextWindow: 163840,
|
|
865
828
|
supportsImages: false,
|
|
866
829
|
supportsPromptCache: false,
|
|
867
|
-
supportsNativeTools: true,
|
|
868
830
|
inputPrice: 0.5,
|
|
869
831
|
outputPrice: 1.5,
|
|
870
832
|
cacheWritesPrice: 0,
|
|
@@ -876,7 +838,6 @@ var basetenModels = {
|
|
|
876
838
|
contextWindow: 163840,
|
|
877
839
|
supportsImages: false,
|
|
878
840
|
supportsPromptCache: false,
|
|
879
|
-
supportsNativeTools: true,
|
|
880
841
|
inputPrice: 0.3,
|
|
881
842
|
outputPrice: 0.45,
|
|
882
843
|
cacheWritesPrice: 0,
|
|
@@ -888,7 +849,6 @@ var basetenModels = {
|
|
|
888
849
|
contextWindow: 128072,
|
|
889
850
|
supportsImages: false,
|
|
890
851
|
supportsPromptCache: false,
|
|
891
|
-
supportsNativeTools: true,
|
|
892
852
|
inputPrice: 0.1,
|
|
893
853
|
outputPrice: 0.5,
|
|
894
854
|
cacheWritesPrice: 0,
|
|
@@ -900,7 +860,6 @@ var basetenModels = {
|
|
|
900
860
|
contextWindow: 262144,
|
|
901
861
|
supportsImages: false,
|
|
902
862
|
supportsPromptCache: false,
|
|
903
|
-
supportsNativeTools: true,
|
|
904
863
|
inputPrice: 0.22,
|
|
905
864
|
outputPrice: 0.8,
|
|
906
865
|
cacheWritesPrice: 0,
|
|
@@ -912,7 +871,6 @@ var basetenModels = {
|
|
|
912
871
|
contextWindow: 262144,
|
|
913
872
|
supportsImages: false,
|
|
914
873
|
supportsPromptCache: false,
|
|
915
|
-
supportsNativeTools: true,
|
|
916
874
|
inputPrice: 0.38,
|
|
917
875
|
outputPrice: 1.53,
|
|
918
876
|
cacheWritesPrice: 0,
|
|
@@ -924,7 +882,6 @@ var basetenModels = {
|
|
|
924
882
|
contextWindow: 262e3,
|
|
925
883
|
supportsImages: false,
|
|
926
884
|
supportsPromptCache: false,
|
|
927
|
-
supportsNativeTools: true,
|
|
928
885
|
inputPrice: 0.6,
|
|
929
886
|
outputPrice: 2.5,
|
|
930
887
|
cacheWritesPrice: 0,
|
|
@@ -944,8 +901,6 @@ var bedrockModels = {
|
|
|
944
901
|
supportsImages: true,
|
|
945
902
|
supportsPromptCache: true,
|
|
946
903
|
supportsReasoningBudget: true,
|
|
947
|
-
supportsNativeTools: true,
|
|
948
|
-
defaultToolProtocol: "native",
|
|
949
904
|
inputPrice: 3,
|
|
950
905
|
outputPrice: 15,
|
|
951
906
|
cacheWritesPrice: 3.75,
|
|
@@ -959,7 +914,6 @@ var bedrockModels = {
|
|
|
959
914
|
contextWindow: 3e5,
|
|
960
915
|
supportsImages: true,
|
|
961
916
|
supportsPromptCache: true,
|
|
962
|
-
supportsNativeTools: true,
|
|
963
917
|
inputPrice: 0.8,
|
|
964
918
|
outputPrice: 3.2,
|
|
965
919
|
cacheWritesPrice: 0.8,
|
|
@@ -975,7 +929,6 @@ var bedrockModels = {
|
|
|
975
929
|
contextWindow: 3e5,
|
|
976
930
|
supportsImages: true,
|
|
977
931
|
supportsPromptCache: false,
|
|
978
|
-
supportsNativeTools: true,
|
|
979
932
|
inputPrice: 1,
|
|
980
933
|
outputPrice: 4,
|
|
981
934
|
cacheWritesPrice: 1,
|
|
@@ -989,7 +942,6 @@ var bedrockModels = {
|
|
|
989
942
|
contextWindow: 3e5,
|
|
990
943
|
supportsImages: true,
|
|
991
944
|
supportsPromptCache: true,
|
|
992
|
-
supportsNativeTools: true,
|
|
993
945
|
inputPrice: 0.06,
|
|
994
946
|
outputPrice: 0.24,
|
|
995
947
|
cacheWritesPrice: 0.06,
|
|
@@ -1005,7 +957,6 @@ var bedrockModels = {
|
|
|
1005
957
|
contextWindow: 1e6,
|
|
1006
958
|
supportsImages: true,
|
|
1007
959
|
supportsPromptCache: true,
|
|
1008
|
-
supportsNativeTools: true,
|
|
1009
960
|
inputPrice: 0.33,
|
|
1010
961
|
outputPrice: 2.75,
|
|
1011
962
|
cacheWritesPrice: 0,
|
|
@@ -1021,7 +972,6 @@ var bedrockModels = {
|
|
|
1021
972
|
contextWindow: 128e3,
|
|
1022
973
|
supportsImages: false,
|
|
1023
974
|
supportsPromptCache: true,
|
|
1024
|
-
supportsNativeTools: true,
|
|
1025
975
|
inputPrice: 0.035,
|
|
1026
976
|
outputPrice: 0.14,
|
|
1027
977
|
cacheWritesPrice: 0.035,
|
|
@@ -1038,8 +988,6 @@ var bedrockModels = {
|
|
|
1038
988
|
supportsImages: true,
|
|
1039
989
|
supportsPromptCache: true,
|
|
1040
990
|
supportsReasoningBudget: true,
|
|
1041
|
-
supportsNativeTools: true,
|
|
1042
|
-
defaultToolProtocol: "native",
|
|
1043
991
|
inputPrice: 3,
|
|
1044
992
|
outputPrice: 15,
|
|
1045
993
|
cacheWritesPrice: 3.75,
|
|
@@ -1054,8 +1002,6 @@ var bedrockModels = {
|
|
|
1054
1002
|
supportsImages: true,
|
|
1055
1003
|
supportsPromptCache: true,
|
|
1056
1004
|
supportsReasoningBudget: true,
|
|
1057
|
-
supportsNativeTools: true,
|
|
1058
|
-
defaultToolProtocol: "native",
|
|
1059
1005
|
inputPrice: 15,
|
|
1060
1006
|
outputPrice: 75,
|
|
1061
1007
|
cacheWritesPrice: 18.75,
|
|
@@ -1070,8 +1016,6 @@ var bedrockModels = {
|
|
|
1070
1016
|
supportsImages: true,
|
|
1071
1017
|
supportsPromptCache: true,
|
|
1072
1018
|
supportsReasoningBudget: true,
|
|
1073
|
-
supportsNativeTools: true,
|
|
1074
|
-
defaultToolProtocol: "native",
|
|
1075
1019
|
inputPrice: 5,
|
|
1076
1020
|
outputPrice: 25,
|
|
1077
1021
|
cacheWritesPrice: 6.25,
|
|
@@ -1086,8 +1030,6 @@ var bedrockModels = {
|
|
|
1086
1030
|
supportsImages: true,
|
|
1087
1031
|
supportsPromptCache: true,
|
|
1088
1032
|
supportsReasoningBudget: true,
|
|
1089
|
-
supportsNativeTools: true,
|
|
1090
|
-
defaultToolProtocol: "native",
|
|
1091
1033
|
inputPrice: 15,
|
|
1092
1034
|
outputPrice: 75,
|
|
1093
1035
|
cacheWritesPrice: 18.75,
|
|
@@ -1102,8 +1044,6 @@ var bedrockModels = {
|
|
|
1102
1044
|
supportsImages: true,
|
|
1103
1045
|
supportsPromptCache: true,
|
|
1104
1046
|
supportsReasoningBudget: true,
|
|
1105
|
-
supportsNativeTools: true,
|
|
1106
|
-
defaultToolProtocol: "native",
|
|
1107
1047
|
inputPrice: 3,
|
|
1108
1048
|
outputPrice: 15,
|
|
1109
1049
|
cacheWritesPrice: 3.75,
|
|
@@ -1117,8 +1057,6 @@ var bedrockModels = {
|
|
|
1117
1057
|
contextWindow: 2e5,
|
|
1118
1058
|
supportsImages: true,
|
|
1119
1059
|
supportsPromptCache: true,
|
|
1120
|
-
supportsNativeTools: true,
|
|
1121
|
-
defaultToolProtocol: "native",
|
|
1122
1060
|
inputPrice: 3,
|
|
1123
1061
|
outputPrice: 15,
|
|
1124
1062
|
cacheWritesPrice: 3.75,
|
|
@@ -1132,8 +1070,6 @@ var bedrockModels = {
|
|
|
1132
1070
|
contextWindow: 2e5,
|
|
1133
1071
|
supportsImages: false,
|
|
1134
1072
|
supportsPromptCache: true,
|
|
1135
|
-
supportsNativeTools: true,
|
|
1136
|
-
defaultToolProtocol: "native",
|
|
1137
1073
|
inputPrice: 0.8,
|
|
1138
1074
|
outputPrice: 4,
|
|
1139
1075
|
cacheWritesPrice: 1,
|
|
@@ -1148,8 +1084,6 @@ var bedrockModels = {
|
|
|
1148
1084
|
supportsImages: true,
|
|
1149
1085
|
supportsPromptCache: true,
|
|
1150
1086
|
supportsReasoningBudget: true,
|
|
1151
|
-
supportsNativeTools: true,
|
|
1152
|
-
defaultToolProtocol: "native",
|
|
1153
1087
|
inputPrice: 1,
|
|
1154
1088
|
outputPrice: 5,
|
|
1155
1089
|
cacheWritesPrice: 1.25,
|
|
@@ -1165,8 +1099,6 @@ var bedrockModels = {
|
|
|
1165
1099
|
contextWindow: 2e5,
|
|
1166
1100
|
supportsImages: true,
|
|
1167
1101
|
supportsPromptCache: false,
|
|
1168
|
-
supportsNativeTools: true,
|
|
1169
|
-
defaultToolProtocol: "native",
|
|
1170
1102
|
inputPrice: 3,
|
|
1171
1103
|
outputPrice: 15
|
|
1172
1104
|
},
|
|
@@ -1175,8 +1107,6 @@ var bedrockModels = {
|
|
|
1175
1107
|
contextWindow: 2e5,
|
|
1176
1108
|
supportsImages: true,
|
|
1177
1109
|
supportsPromptCache: false,
|
|
1178
|
-
supportsNativeTools: true,
|
|
1179
|
-
defaultToolProtocol: "native",
|
|
1180
1110
|
inputPrice: 15,
|
|
1181
1111
|
outputPrice: 75
|
|
1182
1112
|
},
|
|
@@ -1185,8 +1115,6 @@ var bedrockModels = {
|
|
|
1185
1115
|
contextWindow: 2e5,
|
|
1186
1116
|
supportsImages: true,
|
|
1187
1117
|
supportsPromptCache: false,
|
|
1188
|
-
supportsNativeTools: true,
|
|
1189
|
-
defaultToolProtocol: "native",
|
|
1190
1118
|
inputPrice: 3,
|
|
1191
1119
|
outputPrice: 15
|
|
1192
1120
|
},
|
|
@@ -1195,8 +1123,6 @@ var bedrockModels = {
|
|
|
1195
1123
|
contextWindow: 2e5,
|
|
1196
1124
|
supportsImages: true,
|
|
1197
1125
|
supportsPromptCache: false,
|
|
1198
|
-
supportsNativeTools: true,
|
|
1199
|
-
defaultToolProtocol: "native",
|
|
1200
1126
|
inputPrice: 0.25,
|
|
1201
1127
|
outputPrice: 1.25
|
|
1202
1128
|
},
|
|
@@ -1205,7 +1131,6 @@ var bedrockModels = {
|
|
|
1205
1131
|
contextWindow: 128e3,
|
|
1206
1132
|
supportsImages: false,
|
|
1207
1133
|
supportsPromptCache: false,
|
|
1208
|
-
supportsNativeTools: true,
|
|
1209
1134
|
inputPrice: 1.35,
|
|
1210
1135
|
outputPrice: 5.4
|
|
1211
1136
|
},
|
|
@@ -1214,7 +1139,6 @@ var bedrockModels = {
|
|
|
1214
1139
|
contextWindow: 128e3,
|
|
1215
1140
|
supportsImages: false,
|
|
1216
1141
|
supportsPromptCache: false,
|
|
1217
|
-
supportsNativeTools: true,
|
|
1218
1142
|
inputPrice: 0.5,
|
|
1219
1143
|
outputPrice: 1.5,
|
|
1220
1144
|
description: "GPT-OSS 20B - Optimized for low latency and local/specialized use cases"
|
|
@@ -1224,7 +1148,6 @@ var bedrockModels = {
|
|
|
1224
1148
|
contextWindow: 128e3,
|
|
1225
1149
|
supportsImages: false,
|
|
1226
1150
|
supportsPromptCache: false,
|
|
1227
|
-
supportsNativeTools: true,
|
|
1228
1151
|
inputPrice: 2,
|
|
1229
1152
|
outputPrice: 6,
|
|
1230
1153
|
description: "GPT-OSS 120B - Production-ready, general-purpose, high-reasoning model"
|
|
@@ -1234,7 +1157,6 @@ var bedrockModels = {
|
|
|
1234
1157
|
contextWindow: 128e3,
|
|
1235
1158
|
supportsImages: false,
|
|
1236
1159
|
supportsPromptCache: false,
|
|
1237
|
-
supportsNativeTools: true,
|
|
1238
1160
|
inputPrice: 0.72,
|
|
1239
1161
|
outputPrice: 0.72,
|
|
1240
1162
|
description: "Llama 3.3 Instruct (70B)"
|
|
@@ -1244,7 +1166,6 @@ var bedrockModels = {
|
|
|
1244
1166
|
contextWindow: 128e3,
|
|
1245
1167
|
supportsImages: true,
|
|
1246
1168
|
supportsPromptCache: false,
|
|
1247
|
-
supportsNativeTools: true,
|
|
1248
1169
|
inputPrice: 0.72,
|
|
1249
1170
|
outputPrice: 0.72,
|
|
1250
1171
|
description: "Llama 3.2 Instruct (90B)"
|
|
@@ -1254,7 +1175,6 @@ var bedrockModels = {
|
|
|
1254
1175
|
contextWindow: 128e3,
|
|
1255
1176
|
supportsImages: true,
|
|
1256
1177
|
supportsPromptCache: false,
|
|
1257
|
-
supportsNativeTools: true,
|
|
1258
1178
|
inputPrice: 0.16,
|
|
1259
1179
|
outputPrice: 0.16,
|
|
1260
1180
|
description: "Llama 3.2 Instruct (11B)"
|
|
@@ -1264,7 +1184,6 @@ var bedrockModels = {
|
|
|
1264
1184
|
contextWindow: 128e3,
|
|
1265
1185
|
supportsImages: false,
|
|
1266
1186
|
supportsPromptCache: false,
|
|
1267
|
-
supportsNativeTools: true,
|
|
1268
1187
|
inputPrice: 0.15,
|
|
1269
1188
|
outputPrice: 0.15,
|
|
1270
1189
|
description: "Llama 3.2 Instruct (3B)"
|
|
@@ -1274,7 +1193,6 @@ var bedrockModels = {
|
|
|
1274
1193
|
contextWindow: 128e3,
|
|
1275
1194
|
supportsImages: false,
|
|
1276
1195
|
supportsPromptCache: false,
|
|
1277
|
-
supportsNativeTools: true,
|
|
1278
1196
|
inputPrice: 0.1,
|
|
1279
1197
|
outputPrice: 0.1,
|
|
1280
1198
|
description: "Llama 3.2 Instruct (1B)"
|
|
@@ -1284,7 +1202,6 @@ var bedrockModels = {
|
|
|
1284
1202
|
contextWindow: 128e3,
|
|
1285
1203
|
supportsImages: false,
|
|
1286
1204
|
supportsPromptCache: false,
|
|
1287
|
-
supportsNativeTools: true,
|
|
1288
1205
|
inputPrice: 2.4,
|
|
1289
1206
|
outputPrice: 2.4,
|
|
1290
1207
|
description: "Llama 3.1 Instruct (405B)"
|
|
@@ -1294,7 +1211,6 @@ var bedrockModels = {
|
|
|
1294
1211
|
contextWindow: 128e3,
|
|
1295
1212
|
supportsImages: false,
|
|
1296
1213
|
supportsPromptCache: false,
|
|
1297
|
-
supportsNativeTools: true,
|
|
1298
1214
|
inputPrice: 0.72,
|
|
1299
1215
|
outputPrice: 0.72,
|
|
1300
1216
|
description: "Llama 3.1 Instruct (70B)"
|
|
@@ -1304,7 +1220,6 @@ var bedrockModels = {
|
|
|
1304
1220
|
contextWindow: 128e3,
|
|
1305
1221
|
supportsImages: false,
|
|
1306
1222
|
supportsPromptCache: false,
|
|
1307
|
-
supportsNativeTools: true,
|
|
1308
1223
|
inputPrice: 0.9,
|
|
1309
1224
|
outputPrice: 0.9,
|
|
1310
1225
|
description: "Llama 3.1 Instruct (70B) (w/ latency optimized inference)"
|
|
@@ -1314,7 +1229,6 @@ var bedrockModels = {
|
|
|
1314
1229
|
contextWindow: 8e3,
|
|
1315
1230
|
supportsImages: false,
|
|
1316
1231
|
supportsPromptCache: false,
|
|
1317
|
-
supportsNativeTools: true,
|
|
1318
1232
|
inputPrice: 0.22,
|
|
1319
1233
|
outputPrice: 0.22,
|
|
1320
1234
|
description: "Llama 3.1 Instruct (8B)"
|
|
@@ -1324,7 +1238,6 @@ var bedrockModels = {
|
|
|
1324
1238
|
contextWindow: 8e3,
|
|
1325
1239
|
supportsImages: false,
|
|
1326
1240
|
supportsPromptCache: false,
|
|
1327
|
-
supportsNativeTools: true,
|
|
1328
1241
|
inputPrice: 2.65,
|
|
1329
1242
|
outputPrice: 3.5
|
|
1330
1243
|
},
|
|
@@ -1333,7 +1246,6 @@ var bedrockModels = {
|
|
|
1333
1246
|
contextWindow: 4e3,
|
|
1334
1247
|
supportsImages: false,
|
|
1335
1248
|
supportsPromptCache: false,
|
|
1336
|
-
supportsNativeTools: true,
|
|
1337
1249
|
inputPrice: 0.3,
|
|
1338
1250
|
outputPrice: 0.6
|
|
1339
1251
|
},
|
|
@@ -1342,7 +1254,6 @@ var bedrockModels = {
|
|
|
1342
1254
|
contextWindow: 8e3,
|
|
1343
1255
|
supportsImages: false,
|
|
1344
1256
|
supportsPromptCache: false,
|
|
1345
|
-
supportsNativeTools: true,
|
|
1346
1257
|
inputPrice: 0.15,
|
|
1347
1258
|
outputPrice: 0.2,
|
|
1348
1259
|
description: "Amazon Titan Text Lite"
|
|
@@ -1352,7 +1263,6 @@ var bedrockModels = {
|
|
|
1352
1263
|
contextWindow: 8e3,
|
|
1353
1264
|
supportsImages: false,
|
|
1354
1265
|
supportsPromptCache: false,
|
|
1355
|
-
supportsNativeTools: true,
|
|
1356
1266
|
inputPrice: 0.2,
|
|
1357
1267
|
outputPrice: 0.6,
|
|
1358
1268
|
description: "Amazon Titan Text Express"
|
|
@@ -1362,8 +1272,6 @@ var bedrockModels = {
|
|
|
1362
1272
|
contextWindow: 262144,
|
|
1363
1273
|
supportsImages: false,
|
|
1364
1274
|
supportsPromptCache: false,
|
|
1365
|
-
supportsNativeTools: true,
|
|
1366
|
-
defaultToolProtocol: "native",
|
|
1367
1275
|
preserveReasoning: true,
|
|
1368
1276
|
inputPrice: 0.6,
|
|
1369
1277
|
outputPrice: 2.5,
|
|
@@ -1374,8 +1282,6 @@ var bedrockModels = {
|
|
|
1374
1282
|
contextWindow: 196608,
|
|
1375
1283
|
supportsImages: false,
|
|
1376
1284
|
supportsPromptCache: false,
|
|
1377
|
-
supportsNativeTools: true,
|
|
1378
|
-
defaultToolProtocol: "native",
|
|
1379
1285
|
preserveReasoning: true,
|
|
1380
1286
|
inputPrice: 0.3,
|
|
1381
1287
|
outputPrice: 1.2,
|
|
@@ -1386,8 +1292,6 @@ var bedrockModels = {
|
|
|
1386
1292
|
contextWindow: 262144,
|
|
1387
1293
|
supportsImages: false,
|
|
1388
1294
|
supportsPromptCache: false,
|
|
1389
|
-
supportsNativeTools: true,
|
|
1390
|
-
defaultToolProtocol: "native",
|
|
1391
1295
|
inputPrice: 0.15,
|
|
1392
1296
|
outputPrice: 1.2,
|
|
1393
1297
|
description: "Qwen3 Next 80B (MoE model with 3B active parameters)"
|
|
@@ -1397,8 +1301,6 @@ var bedrockModels = {
|
|
|
1397
1301
|
contextWindow: 262144,
|
|
1398
1302
|
supportsImages: false,
|
|
1399
1303
|
supportsPromptCache: false,
|
|
1400
|
-
supportsNativeTools: true,
|
|
1401
|
-
defaultToolProtocol: "native",
|
|
1402
1304
|
inputPrice: 0.45,
|
|
1403
1305
|
outputPrice: 1.8,
|
|
1404
1306
|
description: "Qwen3 Coder 480B (MoE model with 35B active parameters)"
|
|
@@ -1489,26 +1391,14 @@ var BEDROCK_SERVICE_TIER_PRICING = {
|
|
|
1489
1391
|
// src/providers/cerebras.ts
|
|
1490
1392
|
var cerebrasDefaultModelId = "gpt-oss-120b";
|
|
1491
1393
|
var cerebrasModels = {
|
|
1492
|
-
"zai-glm-4.6": {
|
|
1493
|
-
maxTokens: 16384,
|
|
1494
|
-
// Conservative default to avoid premature rate limiting (Cerebras reserves quota upfront)
|
|
1495
|
-
contextWindow: 131072,
|
|
1496
|
-
supportsImages: false,
|
|
1497
|
-
supportsPromptCache: false,
|
|
1498
|
-
supportsNativeTools: true,
|
|
1499
|
-
defaultToolProtocol: "native",
|
|
1500
|
-
inputPrice: 0,
|
|
1501
|
-
outputPrice: 0,
|
|
1502
|
-
description: "Fast general-purpose model on Cerebras (up to 1,000 tokens/s). To be deprecated soon."
|
|
1503
|
-
},
|
|
1504
1394
|
"zai-glm-4.7": {
|
|
1505
1395
|
maxTokens: 16384,
|
|
1506
1396
|
// Conservative default to avoid premature rate limiting (Cerebras reserves quota upfront)
|
|
1507
1397
|
contextWindow: 131072,
|
|
1508
1398
|
supportsImages: false,
|
|
1509
|
-
supportsPromptCache:
|
|
1510
|
-
|
|
1511
|
-
|
|
1399
|
+
supportsPromptCache: true,
|
|
1400
|
+
supportsTemperature: true,
|
|
1401
|
+
defaultTemperature: 1,
|
|
1512
1402
|
inputPrice: 0,
|
|
1513
1403
|
outputPrice: 0,
|
|
1514
1404
|
description: "Highly capable general-purpose model on Cerebras (up to 1,000 tokens/s), competitive with leading proprietary models on coding tasks."
|
|
@@ -1519,8 +1409,6 @@ var cerebrasModels = {
|
|
|
1519
1409
|
contextWindow: 64e3,
|
|
1520
1410
|
supportsImages: false,
|
|
1521
1411
|
supportsPromptCache: false,
|
|
1522
|
-
supportsNativeTools: true,
|
|
1523
|
-
defaultToolProtocol: "native",
|
|
1524
1412
|
inputPrice: 0,
|
|
1525
1413
|
outputPrice: 0,
|
|
1526
1414
|
description: "Intelligent model with ~1400 tokens/s"
|
|
@@ -1531,8 +1419,6 @@ var cerebrasModels = {
|
|
|
1531
1419
|
contextWindow: 64e3,
|
|
1532
1420
|
supportsImages: false,
|
|
1533
1421
|
supportsPromptCache: false,
|
|
1534
|
-
supportsNativeTools: true,
|
|
1535
|
-
defaultToolProtocol: "native",
|
|
1536
1422
|
inputPrice: 0,
|
|
1537
1423
|
outputPrice: 0,
|
|
1538
1424
|
description: "Powerful model with ~2600 tokens/s"
|
|
@@ -1543,8 +1429,6 @@ var cerebrasModels = {
|
|
|
1543
1429
|
contextWindow: 64e3,
|
|
1544
1430
|
supportsImages: false,
|
|
1545
1431
|
supportsPromptCache: false,
|
|
1546
|
-
supportsNativeTools: true,
|
|
1547
|
-
defaultToolProtocol: "native",
|
|
1548
1432
|
inputPrice: 0,
|
|
1549
1433
|
outputPrice: 0,
|
|
1550
1434
|
description: "SOTA coding performance with ~2500 tokens/s"
|
|
@@ -1555,8 +1439,6 @@ var cerebrasModels = {
|
|
|
1555
1439
|
contextWindow: 64e3,
|
|
1556
1440
|
supportsImages: false,
|
|
1557
1441
|
supportsPromptCache: false,
|
|
1558
|
-
supportsNativeTools: true,
|
|
1559
|
-
defaultToolProtocol: "native",
|
|
1560
1442
|
inputPrice: 0,
|
|
1561
1443
|
outputPrice: 0,
|
|
1562
1444
|
description: "OpenAI GPT OSS model with ~2800 tokens/s\n\n\u2022 64K context window\n\u2022 Excels at efficient reasoning across science, math, and coding"
|
|
@@ -1571,8 +1453,6 @@ var chutesModels = {
|
|
|
1571
1453
|
contextWindow: 163840,
|
|
1572
1454
|
supportsImages: false,
|
|
1573
1455
|
supportsPromptCache: false,
|
|
1574
|
-
supportsNativeTools: true,
|
|
1575
|
-
defaultToolProtocol: "native",
|
|
1576
1456
|
inputPrice: 0,
|
|
1577
1457
|
outputPrice: 0,
|
|
1578
1458
|
description: "DeepSeek R1 0528 model."
|
|
@@ -1582,8 +1462,6 @@ var chutesModels = {
|
|
|
1582
1462
|
contextWindow: 163840,
|
|
1583
1463
|
supportsImages: false,
|
|
1584
1464
|
supportsPromptCache: false,
|
|
1585
|
-
supportsNativeTools: true,
|
|
1586
|
-
defaultToolProtocol: "native",
|
|
1587
1465
|
inputPrice: 0,
|
|
1588
1466
|
outputPrice: 0,
|
|
1589
1467
|
description: "DeepSeek R1 model."
|
|
@@ -1593,8 +1471,6 @@ var chutesModels = {
|
|
|
1593
1471
|
contextWindow: 163840,
|
|
1594
1472
|
supportsImages: false,
|
|
1595
1473
|
supportsPromptCache: false,
|
|
1596
|
-
supportsNativeTools: true,
|
|
1597
|
-
defaultToolProtocol: "native",
|
|
1598
1474
|
inputPrice: 0,
|
|
1599
1475
|
outputPrice: 0,
|
|
1600
1476
|
description: "DeepSeek V3 model."
|
|
@@ -1604,8 +1480,6 @@ var chutesModels = {
|
|
|
1604
1480
|
contextWindow: 163840,
|
|
1605
1481
|
supportsImages: false,
|
|
1606
1482
|
supportsPromptCache: false,
|
|
1607
|
-
supportsNativeTools: true,
|
|
1608
|
-
defaultToolProtocol: "native",
|
|
1609
1483
|
inputPrice: 0,
|
|
1610
1484
|
outputPrice: 0,
|
|
1611
1485
|
description: "DeepSeek V3.1 model."
|
|
@@ -1615,8 +1489,6 @@ var chutesModels = {
|
|
|
1615
1489
|
contextWindow: 163840,
|
|
1616
1490
|
supportsImages: false,
|
|
1617
1491
|
supportsPromptCache: false,
|
|
1618
|
-
supportsNativeTools: true,
|
|
1619
|
-
defaultToolProtocol: "native",
|
|
1620
1492
|
inputPrice: 0.23,
|
|
1621
1493
|
outputPrice: 0.9,
|
|
1622
1494
|
description: "DeepSeek\u2011V3.1\u2011Terminus is an update to V3.1 that improves language consistency by reducing CN/EN mix\u2011ups and eliminating random characters, while strengthening agent capabilities with notably better Code Agent and Search Agent performance."
|
|
@@ -1626,8 +1498,6 @@ var chutesModels = {
|
|
|
1626
1498
|
contextWindow: 163840,
|
|
1627
1499
|
supportsImages: false,
|
|
1628
1500
|
supportsPromptCache: false,
|
|
1629
|
-
supportsNativeTools: true,
|
|
1630
|
-
defaultToolProtocol: "native",
|
|
1631
1501
|
inputPrice: 1,
|
|
1632
1502
|
outputPrice: 3,
|
|
1633
1503
|
description: "DeepSeek-V3.1-turbo is an FP8, speculative-decoding turbo variant optimized for ultra-fast single-shot queries (~200 TPS), with outputs close to the originals and solid function calling/reasoning/structured output, priced at $1/M input and $3/M output tokens, using 2\xD7 quota per request and not intended for bulk workloads."
|
|
@@ -1637,8 +1507,6 @@ var chutesModels = {
|
|
|
1637
1507
|
contextWindow: 163840,
|
|
1638
1508
|
supportsImages: false,
|
|
1639
1509
|
supportsPromptCache: false,
|
|
1640
|
-
supportsNativeTools: true,
|
|
1641
|
-
defaultToolProtocol: "native",
|
|
1642
1510
|
inputPrice: 0.25,
|
|
1643
1511
|
outputPrice: 0.35,
|
|
1644
1512
|
description: "DeepSeek-V3.2-Exp is an experimental LLM that introduces DeepSeek Sparse Attention to improve long\u2011context training and inference efficiency while maintaining performance comparable to V3.1\u2011Terminus."
|
|
@@ -1650,8 +1518,6 @@ var chutesModels = {
|
|
|
1650
1518
|
// From Groq
|
|
1651
1519
|
supportsImages: false,
|
|
1652
1520
|
supportsPromptCache: false,
|
|
1653
|
-
supportsNativeTools: true,
|
|
1654
|
-
defaultToolProtocol: "native",
|
|
1655
1521
|
inputPrice: 0,
|
|
1656
1522
|
outputPrice: 0,
|
|
1657
1523
|
description: "Unsloth Llama 3.3 70B Instruct model."
|
|
@@ -1661,8 +1527,6 @@ var chutesModels = {
|
|
|
1661
1527
|
contextWindow: 512e3,
|
|
1662
1528
|
supportsImages: false,
|
|
1663
1529
|
supportsPromptCache: false,
|
|
1664
|
-
supportsNativeTools: true,
|
|
1665
|
-
defaultToolProtocol: "native",
|
|
1666
1530
|
inputPrice: 0,
|
|
1667
1531
|
outputPrice: 0,
|
|
1668
1532
|
description: "ChutesAI Llama 4 Scout 17B Instruct model, 512K context."
|
|
@@ -1672,8 +1536,6 @@ var chutesModels = {
|
|
|
1672
1536
|
contextWindow: 128e3,
|
|
1673
1537
|
supportsImages: false,
|
|
1674
1538
|
supportsPromptCache: false,
|
|
1675
|
-
supportsNativeTools: true,
|
|
1676
|
-
defaultToolProtocol: "native",
|
|
1677
1539
|
inputPrice: 0,
|
|
1678
1540
|
outputPrice: 0,
|
|
1679
1541
|
description: "Unsloth Mistral Nemo Instruct model."
|
|
@@ -1683,8 +1545,6 @@ var chutesModels = {
|
|
|
1683
1545
|
contextWindow: 131072,
|
|
1684
1546
|
supportsImages: false,
|
|
1685
1547
|
supportsPromptCache: false,
|
|
1686
|
-
supportsNativeTools: true,
|
|
1687
|
-
defaultToolProtocol: "native",
|
|
1688
1548
|
inputPrice: 0,
|
|
1689
1549
|
outputPrice: 0,
|
|
1690
1550
|
description: "Unsloth Gemma 3 12B IT model."
|
|
@@ -1694,8 +1554,6 @@ var chutesModels = {
|
|
|
1694
1554
|
contextWindow: 131072,
|
|
1695
1555
|
supportsImages: false,
|
|
1696
1556
|
supportsPromptCache: false,
|
|
1697
|
-
supportsNativeTools: true,
|
|
1698
|
-
defaultToolProtocol: "native",
|
|
1699
1557
|
inputPrice: 0,
|
|
1700
1558
|
outputPrice: 0,
|
|
1701
1559
|
description: "Nous DeepHermes 3 Llama 3 8B Preview model."
|
|
@@ -1705,8 +1563,6 @@ var chutesModels = {
|
|
|
1705
1563
|
contextWindow: 131072,
|
|
1706
1564
|
supportsImages: false,
|
|
1707
1565
|
supportsPromptCache: false,
|
|
1708
|
-
supportsNativeTools: true,
|
|
1709
|
-
defaultToolProtocol: "native",
|
|
1710
1566
|
inputPrice: 0,
|
|
1711
1567
|
outputPrice: 0,
|
|
1712
1568
|
description: "Unsloth Gemma 3 4B IT model."
|
|
@@ -1716,8 +1572,6 @@ var chutesModels = {
|
|
|
1716
1572
|
contextWindow: 131072,
|
|
1717
1573
|
supportsImages: false,
|
|
1718
1574
|
supportsPromptCache: false,
|
|
1719
|
-
supportsNativeTools: true,
|
|
1720
|
-
defaultToolProtocol: "native",
|
|
1721
1575
|
inputPrice: 0,
|
|
1722
1576
|
outputPrice: 0,
|
|
1723
1577
|
description: "Nvidia Llama 3.3 Nemotron Super 49B model."
|
|
@@ -1727,8 +1581,6 @@ var chutesModels = {
|
|
|
1727
1581
|
contextWindow: 131072,
|
|
1728
1582
|
supportsImages: false,
|
|
1729
1583
|
supportsPromptCache: false,
|
|
1730
|
-
supportsNativeTools: true,
|
|
1731
|
-
defaultToolProtocol: "native",
|
|
1732
1584
|
inputPrice: 0,
|
|
1733
1585
|
outputPrice: 0,
|
|
1734
1586
|
description: "Nvidia Llama 3.1 Nemotron Ultra 253B model."
|
|
@@ -1738,8 +1590,6 @@ var chutesModels = {
|
|
|
1738
1590
|
contextWindow: 256e3,
|
|
1739
1591
|
supportsImages: false,
|
|
1740
1592
|
supportsPromptCache: false,
|
|
1741
|
-
supportsNativeTools: true,
|
|
1742
|
-
defaultToolProtocol: "native",
|
|
1743
1593
|
inputPrice: 0,
|
|
1744
1594
|
outputPrice: 0,
|
|
1745
1595
|
description: "ChutesAI Llama 4 Maverick 17B Instruct FP8 model."
|
|
@@ -1749,8 +1599,6 @@ var chutesModels = {
|
|
|
1749
1599
|
contextWindow: 163840,
|
|
1750
1600
|
supportsImages: false,
|
|
1751
1601
|
supportsPromptCache: false,
|
|
1752
|
-
supportsNativeTools: true,
|
|
1753
|
-
defaultToolProtocol: "native",
|
|
1754
1602
|
inputPrice: 0,
|
|
1755
1603
|
outputPrice: 0,
|
|
1756
1604
|
description: "DeepSeek V3 Base model."
|
|
@@ -1760,8 +1608,6 @@ var chutesModels = {
|
|
|
1760
1608
|
contextWindow: 163840,
|
|
1761
1609
|
supportsImages: false,
|
|
1762
1610
|
supportsPromptCache: false,
|
|
1763
|
-
supportsNativeTools: true,
|
|
1764
|
-
defaultToolProtocol: "native",
|
|
1765
1611
|
inputPrice: 0,
|
|
1766
1612
|
outputPrice: 0,
|
|
1767
1613
|
description: "DeepSeek R1 Zero model."
|
|
@@ -1771,8 +1617,6 @@ var chutesModels = {
|
|
|
1771
1617
|
contextWindow: 163840,
|
|
1772
1618
|
supportsImages: false,
|
|
1773
1619
|
supportsPromptCache: false,
|
|
1774
|
-
supportsNativeTools: true,
|
|
1775
|
-
defaultToolProtocol: "native",
|
|
1776
1620
|
inputPrice: 0,
|
|
1777
1621
|
outputPrice: 0,
|
|
1778
1622
|
description: "DeepSeek V3 (0324) model."
|
|
@@ -1782,8 +1626,6 @@ var chutesModels = {
|
|
|
1782
1626
|
contextWindow: 262144,
|
|
1783
1627
|
supportsImages: false,
|
|
1784
1628
|
supportsPromptCache: false,
|
|
1785
|
-
supportsNativeTools: true,
|
|
1786
|
-
defaultToolProtocol: "native",
|
|
1787
1629
|
inputPrice: 0,
|
|
1788
1630
|
outputPrice: 0,
|
|
1789
1631
|
description: "Qwen3 235B A22B Instruct 2507 model with 262K context window."
|
|
@@ -1793,8 +1635,6 @@ var chutesModels = {
|
|
|
1793
1635
|
contextWindow: 40960,
|
|
1794
1636
|
supportsImages: false,
|
|
1795
1637
|
supportsPromptCache: false,
|
|
1796
|
-
supportsNativeTools: true,
|
|
1797
|
-
defaultToolProtocol: "native",
|
|
1798
1638
|
inputPrice: 0,
|
|
1799
1639
|
outputPrice: 0,
|
|
1800
1640
|
description: "Qwen3 235B A22B model."
|
|
@@ -1804,8 +1644,6 @@ var chutesModels = {
|
|
|
1804
1644
|
contextWindow: 40960,
|
|
1805
1645
|
supportsImages: false,
|
|
1806
1646
|
supportsPromptCache: false,
|
|
1807
|
-
supportsNativeTools: true,
|
|
1808
|
-
defaultToolProtocol: "native",
|
|
1809
1647
|
inputPrice: 0,
|
|
1810
1648
|
outputPrice: 0,
|
|
1811
1649
|
description: "Qwen3 32B model."
|
|
@@ -1815,8 +1653,6 @@ var chutesModels = {
|
|
|
1815
1653
|
contextWindow: 40960,
|
|
1816
1654
|
supportsImages: false,
|
|
1817
1655
|
supportsPromptCache: false,
|
|
1818
|
-
supportsNativeTools: true,
|
|
1819
|
-
defaultToolProtocol: "native",
|
|
1820
1656
|
inputPrice: 0,
|
|
1821
1657
|
outputPrice: 0,
|
|
1822
1658
|
description: "Qwen3 30B A3B model."
|
|
@@ -1826,8 +1662,6 @@ var chutesModels = {
|
|
|
1826
1662
|
contextWindow: 40960,
|
|
1827
1663
|
supportsImages: false,
|
|
1828
1664
|
supportsPromptCache: false,
|
|
1829
|
-
supportsNativeTools: true,
|
|
1830
|
-
defaultToolProtocol: "native",
|
|
1831
1665
|
inputPrice: 0,
|
|
1832
1666
|
outputPrice: 0,
|
|
1833
1667
|
description: "Qwen3 14B model."
|
|
@@ -1837,8 +1671,6 @@ var chutesModels = {
|
|
|
1837
1671
|
contextWindow: 40960,
|
|
1838
1672
|
supportsImages: false,
|
|
1839
1673
|
supportsPromptCache: false,
|
|
1840
|
-
supportsNativeTools: true,
|
|
1841
|
-
defaultToolProtocol: "native",
|
|
1842
1674
|
inputPrice: 0,
|
|
1843
1675
|
outputPrice: 0,
|
|
1844
1676
|
description: "Qwen3 8B model."
|
|
@@ -1848,8 +1680,6 @@ var chutesModels = {
|
|
|
1848
1680
|
contextWindow: 163840,
|
|
1849
1681
|
supportsImages: false,
|
|
1850
1682
|
supportsPromptCache: false,
|
|
1851
|
-
supportsNativeTools: true,
|
|
1852
|
-
defaultToolProtocol: "native",
|
|
1853
1683
|
inputPrice: 0,
|
|
1854
1684
|
outputPrice: 0,
|
|
1855
1685
|
description: "Microsoft MAI-DS-R1 FP8 model."
|
|
@@ -1859,8 +1689,6 @@ var chutesModels = {
|
|
|
1859
1689
|
contextWindow: 163840,
|
|
1860
1690
|
supportsImages: false,
|
|
1861
1691
|
supportsPromptCache: false,
|
|
1862
|
-
supportsNativeTools: true,
|
|
1863
|
-
defaultToolProtocol: "native",
|
|
1864
1692
|
inputPrice: 0,
|
|
1865
1693
|
outputPrice: 0,
|
|
1866
1694
|
description: "TNGTech DeepSeek R1T Chimera model."
|
|
@@ -1870,8 +1698,6 @@ var chutesModels = {
|
|
|
1870
1698
|
contextWindow: 151329,
|
|
1871
1699
|
supportsImages: false,
|
|
1872
1700
|
supportsPromptCache: false,
|
|
1873
|
-
supportsNativeTools: true,
|
|
1874
|
-
defaultToolProtocol: "native",
|
|
1875
1701
|
inputPrice: 0,
|
|
1876
1702
|
outputPrice: 0,
|
|
1877
1703
|
description: "GLM-4.5-Air model with 151,329 token context window and 106B total parameters with 12B activated."
|
|
@@ -1881,8 +1707,6 @@ var chutesModels = {
|
|
|
1881
1707
|
contextWindow: 131072,
|
|
1882
1708
|
supportsImages: false,
|
|
1883
1709
|
supportsPromptCache: false,
|
|
1884
|
-
supportsNativeTools: true,
|
|
1885
|
-
defaultToolProtocol: "native",
|
|
1886
1710
|
inputPrice: 0,
|
|
1887
1711
|
outputPrice: 0,
|
|
1888
1712
|
description: "GLM-4.5-FP8 model with 128k token context window, optimized for agent-based applications with MoE architecture."
|
|
@@ -1892,8 +1716,6 @@ var chutesModels = {
|
|
|
1892
1716
|
contextWindow: 131072,
|
|
1893
1717
|
supportsImages: false,
|
|
1894
1718
|
supportsPromptCache: false,
|
|
1895
|
-
supportsNativeTools: true,
|
|
1896
|
-
defaultToolProtocol: "native",
|
|
1897
1719
|
inputPrice: 1,
|
|
1898
1720
|
outputPrice: 3,
|
|
1899
1721
|
description: "GLM-4.5-turbo model with 128K token context window, optimized for fast inference."
|
|
@@ -1903,8 +1725,6 @@ var chutesModels = {
|
|
|
1903
1725
|
contextWindow: 202752,
|
|
1904
1726
|
supportsImages: false,
|
|
1905
1727
|
supportsPromptCache: false,
|
|
1906
|
-
supportsNativeTools: true,
|
|
1907
|
-
defaultToolProtocol: "native",
|
|
1908
1728
|
inputPrice: 0,
|
|
1909
1729
|
outputPrice: 0,
|
|
1910
1730
|
description: "GLM-4.6 introduces major upgrades over GLM-4.5, including a longer 200K-token context window for complex tasks, stronger coding performance in benchmarks and real-world tools (such as Claude Code, Cline, Roo Code, and Kilo Code), improved reasoning with tool use during inference, more capable and efficient agent integration, and refined writing that better matches human style, readability, and natural role-play scenarios."
|
|
@@ -1915,8 +1735,6 @@ var chutesModels = {
|
|
|
1915
1735
|
contextWindow: 202752,
|
|
1916
1736
|
supportsImages: false,
|
|
1917
1737
|
supportsPromptCache: false,
|
|
1918
|
-
supportsNativeTools: true,
|
|
1919
|
-
defaultToolProtocol: "native",
|
|
1920
1738
|
inputPrice: 1.15,
|
|
1921
1739
|
outputPrice: 3.25,
|
|
1922
1740
|
description: "GLM-4.6-turbo model with 200K-token context window, optimized for fast inference."
|
|
@@ -1926,8 +1744,6 @@ var chutesModels = {
|
|
|
1926
1744
|
contextWindow: 128e3,
|
|
1927
1745
|
supportsImages: false,
|
|
1928
1746
|
supportsPromptCache: false,
|
|
1929
|
-
supportsNativeTools: true,
|
|
1930
|
-
defaultToolProtocol: "native",
|
|
1931
1747
|
inputPrice: 0,
|
|
1932
1748
|
outputPrice: 0,
|
|
1933
1749
|
description: "LongCat Flash Thinking FP8 model with 128K context window, optimized for complex reasoning and coding tasks."
|
|
@@ -1937,8 +1753,6 @@ var chutesModels = {
|
|
|
1937
1753
|
contextWindow: 262144,
|
|
1938
1754
|
supportsImages: false,
|
|
1939
1755
|
supportsPromptCache: false,
|
|
1940
|
-
supportsNativeTools: true,
|
|
1941
|
-
defaultToolProtocol: "native",
|
|
1942
1756
|
inputPrice: 0,
|
|
1943
1757
|
outputPrice: 0,
|
|
1944
1758
|
description: "Qwen3 Coder 480B A35B Instruct FP8 model, optimized for coding tasks."
|
|
@@ -1948,8 +1762,6 @@ var chutesModels = {
|
|
|
1948
1762
|
contextWindow: 75e3,
|
|
1949
1763
|
supportsImages: false,
|
|
1950
1764
|
supportsPromptCache: false,
|
|
1951
|
-
supportsNativeTools: true,
|
|
1952
|
-
defaultToolProtocol: "native",
|
|
1953
1765
|
inputPrice: 0.1481,
|
|
1954
1766
|
outputPrice: 0.5926,
|
|
1955
1767
|
description: "Moonshot AI Kimi K2 Instruct model with 75k context window."
|
|
@@ -1959,8 +1771,6 @@ var chutesModels = {
|
|
|
1959
1771
|
contextWindow: 262144,
|
|
1960
1772
|
supportsImages: false,
|
|
1961
1773
|
supportsPromptCache: false,
|
|
1962
|
-
supportsNativeTools: true,
|
|
1963
|
-
defaultToolProtocol: "native",
|
|
1964
1774
|
inputPrice: 0.1999,
|
|
1965
1775
|
outputPrice: 0.8001,
|
|
1966
1776
|
description: "Moonshot AI Kimi K2 Instruct 0905 model with 256k context window."
|
|
@@ -1970,8 +1780,6 @@ var chutesModels = {
|
|
|
1970
1780
|
contextWindow: 262144,
|
|
1971
1781
|
supportsImages: false,
|
|
1972
1782
|
supportsPromptCache: false,
|
|
1973
|
-
supportsNativeTools: true,
|
|
1974
|
-
defaultToolProtocol: "native",
|
|
1975
1783
|
inputPrice: 0.077968332,
|
|
1976
1784
|
outputPrice: 0.31202496,
|
|
1977
1785
|
description: "Qwen3 235B A22B Thinking 2507 model with 262K context window."
|
|
@@ -1981,8 +1789,6 @@ var chutesModels = {
|
|
|
1981
1789
|
contextWindow: 131072,
|
|
1982
1790
|
supportsImages: false,
|
|
1983
1791
|
supportsPromptCache: false,
|
|
1984
|
-
supportsNativeTools: true,
|
|
1985
|
-
defaultToolProtocol: "native",
|
|
1986
1792
|
inputPrice: 0,
|
|
1987
1793
|
outputPrice: 0,
|
|
1988
1794
|
description: "Fast, stable instruction-tuned model optimized for complex tasks, RAG, and tool use without thinking traces."
|
|
@@ -1992,8 +1798,6 @@ var chutesModels = {
|
|
|
1992
1798
|
contextWindow: 131072,
|
|
1993
1799
|
supportsImages: false,
|
|
1994
1800
|
supportsPromptCache: false,
|
|
1995
|
-
supportsNativeTools: true,
|
|
1996
|
-
defaultToolProtocol: "native",
|
|
1997
1801
|
inputPrice: 0,
|
|
1998
1802
|
outputPrice: 0,
|
|
1999
1803
|
description: "Reasoning-first model with structured thinking traces for multi-step problems, math proofs, and code synthesis."
|
|
@@ -2003,8 +1807,6 @@ var chutesModels = {
|
|
|
2003
1807
|
contextWindow: 262144,
|
|
2004
1808
|
supportsImages: true,
|
|
2005
1809
|
supportsPromptCache: false,
|
|
2006
|
-
supportsNativeTools: true,
|
|
2007
|
-
defaultToolProtocol: "native",
|
|
2008
1810
|
inputPrice: 0.16,
|
|
2009
1811
|
outputPrice: 0.65,
|
|
2010
1812
|
description: "Qwen3\u2011VL\u2011235B\u2011A22B\u2011Thinking is an open\u2011weight MoE vision\u2011language model (235B total, ~22B activated) optimized for deliberate multi\u2011step reasoning with strong text\u2011image\u2011video understanding and long\u2011context capabilities."
|
|
@@ -2012,73 +1814,6 @@ var chutesModels = {
|
|
|
2012
1814
|
};
|
|
2013
1815
|
var chutesDefaultModelInfo = chutesModels[chutesDefaultModelId];
|
|
2014
1816
|
|
|
2015
|
-
// src/providers/claude-code.ts
|
|
2016
|
-
var DATE_SUFFIX_PATTERN = /-\d{8}$/;
|
|
2017
|
-
var claudeCodeModels = {
|
|
2018
|
-
"claude-haiku-4-5": {
|
|
2019
|
-
maxTokens: 32768,
|
|
2020
|
-
contextWindow: 2e5,
|
|
2021
|
-
supportsImages: true,
|
|
2022
|
-
supportsPromptCache: true,
|
|
2023
|
-
supportsNativeTools: true,
|
|
2024
|
-
defaultToolProtocol: "native",
|
|
2025
|
-
supportsReasoningEffort: ["disable", "low", "medium", "high"],
|
|
2026
|
-
reasoningEffort: "medium",
|
|
2027
|
-
description: "Claude Haiku 4.5 - Fast and efficient with thinking"
|
|
2028
|
-
},
|
|
2029
|
-
"claude-sonnet-4-5": {
|
|
2030
|
-
maxTokens: 32768,
|
|
2031
|
-
contextWindow: 2e5,
|
|
2032
|
-
supportsImages: true,
|
|
2033
|
-
supportsPromptCache: true,
|
|
2034
|
-
supportsNativeTools: true,
|
|
2035
|
-
defaultToolProtocol: "native",
|
|
2036
|
-
supportsReasoningEffort: ["disable", "low", "medium", "high"],
|
|
2037
|
-
reasoningEffort: "medium",
|
|
2038
|
-
description: "Claude Sonnet 4.5 - Balanced performance with thinking"
|
|
2039
|
-
},
|
|
2040
|
-
"claude-opus-4-5": {
|
|
2041
|
-
maxTokens: 32768,
|
|
2042
|
-
contextWindow: 2e5,
|
|
2043
|
-
supportsImages: true,
|
|
2044
|
-
supportsPromptCache: true,
|
|
2045
|
-
supportsNativeTools: true,
|
|
2046
|
-
defaultToolProtocol: "native",
|
|
2047
|
-
supportsReasoningEffort: ["disable", "low", "medium", "high"],
|
|
2048
|
-
reasoningEffort: "medium",
|
|
2049
|
-
description: "Claude Opus 4.5 - Most capable with thinking"
|
|
2050
|
-
}
|
|
2051
|
-
};
|
|
2052
|
-
var claudeCodeDefaultModelId = "claude-sonnet-4-5";
|
|
2053
|
-
var MODEL_FAMILY_PATTERNS = [
|
|
2054
|
-
// Opus models (any version) → claude-opus-4-5
|
|
2055
|
-
{ pattern: /opus/i, target: "claude-opus-4-5" },
|
|
2056
|
-
// Haiku models (any version) → claude-haiku-4-5
|
|
2057
|
-
{ pattern: /haiku/i, target: "claude-haiku-4-5" },
|
|
2058
|
-
// Sonnet models (any version) → claude-sonnet-4-5
|
|
2059
|
-
{ pattern: /sonnet/i, target: "claude-sonnet-4-5" }
|
|
2060
|
-
];
|
|
2061
|
-
function normalizeClaudeCodeModelId(modelId) {
|
|
2062
|
-
if (Object.hasOwn(claudeCodeModels, modelId)) {
|
|
2063
|
-
return modelId;
|
|
2064
|
-
}
|
|
2065
|
-
const withoutDate = modelId.replace(DATE_SUFFIX_PATTERN, "");
|
|
2066
|
-
if (Object.hasOwn(claudeCodeModels, withoutDate)) {
|
|
2067
|
-
return withoutDate;
|
|
2068
|
-
}
|
|
2069
|
-
for (const { pattern, target } of MODEL_FAMILY_PATTERNS) {
|
|
2070
|
-
if (pattern.test(modelId)) {
|
|
2071
|
-
return target;
|
|
2072
|
-
}
|
|
2073
|
-
}
|
|
2074
|
-
return claudeCodeDefaultModelId;
|
|
2075
|
-
}
|
|
2076
|
-
var claudeCodeReasoningConfig = {
|
|
2077
|
-
low: { budgetTokens: 16e3 },
|
|
2078
|
-
medium: { budgetTokens: 32e3 },
|
|
2079
|
-
high: { budgetTokens: 64e3 }
|
|
2080
|
-
};
|
|
2081
|
-
|
|
2082
1817
|
// src/providers/deepseek.ts
|
|
2083
1818
|
var deepSeekDefaultModelId = "deepseek-chat";
|
|
2084
1819
|
var deepSeekModels = {
|
|
@@ -2088,8 +1823,6 @@ var deepSeekModels = {
|
|
|
2088
1823
|
contextWindow: 128e3,
|
|
2089
1824
|
supportsImages: false,
|
|
2090
1825
|
supportsPromptCache: true,
|
|
2091
|
-
supportsNativeTools: true,
|
|
2092
|
-
defaultToolProtocol: "native",
|
|
2093
1826
|
inputPrice: 0.28,
|
|
2094
1827
|
// $0.28 per million tokens (cache miss) - Updated Dec 9, 2025
|
|
2095
1828
|
outputPrice: 0.42,
|
|
@@ -2106,8 +1839,6 @@ var deepSeekModels = {
|
|
|
2106
1839
|
contextWindow: 128e3,
|
|
2107
1840
|
supportsImages: false,
|
|
2108
1841
|
supportsPromptCache: true,
|
|
2109
|
-
supportsNativeTools: true,
|
|
2110
|
-
defaultToolProtocol: "native",
|
|
2111
1842
|
preserveReasoning: true,
|
|
2112
1843
|
inputPrice: 0.28,
|
|
2113
1844
|
// $0.28 per million tokens (cache miss) - Updated Dec 9, 2025
|
|
@@ -2130,8 +1861,6 @@ var doubaoModels = {
|
|
|
2130
1861
|
contextWindow: 128e3,
|
|
2131
1862
|
supportsImages: true,
|
|
2132
1863
|
supportsPromptCache: true,
|
|
2133
|
-
supportsNativeTools: true,
|
|
2134
|
-
defaultToolProtocol: "native",
|
|
2135
1864
|
inputPrice: 1e-4,
|
|
2136
1865
|
// $0.0001 per million tokens (cache miss)
|
|
2137
1866
|
outputPrice: 4e-4,
|
|
@@ -2147,8 +1876,6 @@ var doubaoModels = {
|
|
|
2147
1876
|
contextWindow: 128e3,
|
|
2148
1877
|
supportsImages: true,
|
|
2149
1878
|
supportsPromptCache: true,
|
|
2150
|
-
supportsNativeTools: true,
|
|
2151
|
-
defaultToolProtocol: "native",
|
|
2152
1879
|
inputPrice: 2e-4,
|
|
2153
1880
|
// $0.0002 per million tokens
|
|
2154
1881
|
outputPrice: 8e-4,
|
|
@@ -2164,8 +1891,6 @@ var doubaoModels = {
|
|
|
2164
1891
|
contextWindow: 128e3,
|
|
2165
1892
|
supportsImages: true,
|
|
2166
1893
|
supportsPromptCache: true,
|
|
2167
|
-
supportsNativeTools: true,
|
|
2168
|
-
defaultToolProtocol: "native",
|
|
2169
1894
|
inputPrice: 15e-5,
|
|
2170
1895
|
// $0.00015 per million tokens
|
|
2171
1896
|
outputPrice: 6e-4,
|
|
@@ -2188,7 +1913,6 @@ var featherlessModels = {
|
|
|
2188
1913
|
contextWindow: 32678,
|
|
2189
1914
|
supportsImages: false,
|
|
2190
1915
|
supportsPromptCache: false,
|
|
2191
|
-
supportsNativeTools: true,
|
|
2192
1916
|
inputPrice: 0,
|
|
2193
1917
|
outputPrice: 0,
|
|
2194
1918
|
description: "DeepSeek V3 0324 model."
|
|
@@ -2198,7 +1922,6 @@ var featherlessModels = {
|
|
|
2198
1922
|
contextWindow: 32678,
|
|
2199
1923
|
supportsImages: false,
|
|
2200
1924
|
supportsPromptCache: false,
|
|
2201
|
-
supportsNativeTools: true,
|
|
2202
1925
|
inputPrice: 0,
|
|
2203
1926
|
outputPrice: 0,
|
|
2204
1927
|
description: "DeepSeek R1 0528 model."
|
|
@@ -2208,7 +1931,6 @@ var featherlessModels = {
|
|
|
2208
1931
|
contextWindow: 32678,
|
|
2209
1932
|
supportsImages: false,
|
|
2210
1933
|
supportsPromptCache: false,
|
|
2211
|
-
supportsNativeTools: true,
|
|
2212
1934
|
inputPrice: 0,
|
|
2213
1935
|
outputPrice: 0,
|
|
2214
1936
|
description: "Kimi K2 Instruct model."
|
|
@@ -2218,7 +1940,6 @@ var featherlessModels = {
|
|
|
2218
1940
|
contextWindow: 32678,
|
|
2219
1941
|
supportsImages: false,
|
|
2220
1942
|
supportsPromptCache: false,
|
|
2221
|
-
supportsNativeTools: true,
|
|
2222
1943
|
inputPrice: 0,
|
|
2223
1944
|
outputPrice: 0,
|
|
2224
1945
|
description: "GPT-OSS 120B model."
|
|
@@ -2228,7 +1949,6 @@ var featherlessModels = {
|
|
|
2228
1949
|
contextWindow: 32678,
|
|
2229
1950
|
supportsImages: false,
|
|
2230
1951
|
supportsPromptCache: false,
|
|
2231
|
-
supportsNativeTools: true,
|
|
2232
1952
|
inputPrice: 0,
|
|
2233
1953
|
outputPrice: 0,
|
|
2234
1954
|
description: "Qwen3 Coder 480B A35B Instruct model."
|
|
@@ -2244,8 +1964,6 @@ var fireworksModels = {
|
|
|
2244
1964
|
contextWindow: 262144,
|
|
2245
1965
|
supportsImages: false,
|
|
2246
1966
|
supportsPromptCache: true,
|
|
2247
|
-
supportsNativeTools: true,
|
|
2248
|
-
defaultToolProtocol: "native",
|
|
2249
1967
|
inputPrice: 0.6,
|
|
2250
1968
|
outputPrice: 2.5,
|
|
2251
1969
|
cacheReadsPrice: 0.15,
|
|
@@ -2256,8 +1974,6 @@ var fireworksModels = {
|
|
|
2256
1974
|
contextWindow: 128e3,
|
|
2257
1975
|
supportsImages: false,
|
|
2258
1976
|
supportsPromptCache: false,
|
|
2259
|
-
supportsNativeTools: true,
|
|
2260
|
-
defaultToolProtocol: "native",
|
|
2261
1977
|
inputPrice: 0.6,
|
|
2262
1978
|
outputPrice: 2.5,
|
|
2263
1979
|
description: "Kimi K2 is a state-of-the-art mixture-of-experts (MoE) language model with 32 billion activated parameters and 1 trillion total parameters. Trained with the Muon optimizer, Kimi K2 achieves exceptional performance across frontier knowledge, reasoning, and coding tasks while being meticulously optimized for agentic capabilities."
|
|
@@ -2267,7 +1983,6 @@ var fireworksModels = {
|
|
|
2267
1983
|
contextWindow: 256e3,
|
|
2268
1984
|
supportsImages: false,
|
|
2269
1985
|
supportsPromptCache: true,
|
|
2270
|
-
supportsNativeTools: true,
|
|
2271
1986
|
supportsTemperature: true,
|
|
2272
1987
|
preserveReasoning: true,
|
|
2273
1988
|
defaultTemperature: 1,
|
|
@@ -2281,8 +1996,6 @@ var fireworksModels = {
|
|
|
2281
1996
|
contextWindow: 204800,
|
|
2282
1997
|
supportsImages: false,
|
|
2283
1998
|
supportsPromptCache: false,
|
|
2284
|
-
supportsNativeTools: true,
|
|
2285
|
-
defaultToolProtocol: "native",
|
|
2286
1999
|
inputPrice: 0.3,
|
|
2287
2000
|
outputPrice: 1.2,
|
|
2288
2001
|
description: "MiniMax M2 is a high-performance language model with 204.8K context window, optimized for long-context understanding and generation tasks."
|
|
@@ -2292,8 +2005,6 @@ var fireworksModels = {
|
|
|
2292
2005
|
contextWindow: 256e3,
|
|
2293
2006
|
supportsImages: false,
|
|
2294
2007
|
supportsPromptCache: false,
|
|
2295
|
-
supportsNativeTools: true,
|
|
2296
|
-
defaultToolProtocol: "native",
|
|
2297
2008
|
inputPrice: 0.22,
|
|
2298
2009
|
outputPrice: 0.88,
|
|
2299
2010
|
description: "Latest Qwen3 thinking model, competitive against the best closed source models in Jul 2025."
|
|
@@ -2303,8 +2014,6 @@ var fireworksModels = {
|
|
|
2303
2014
|
contextWindow: 256e3,
|
|
2304
2015
|
supportsImages: false,
|
|
2305
2016
|
supportsPromptCache: false,
|
|
2306
|
-
supportsNativeTools: true,
|
|
2307
|
-
defaultToolProtocol: "native",
|
|
2308
2017
|
inputPrice: 0.45,
|
|
2309
2018
|
outputPrice: 1.8,
|
|
2310
2019
|
description: "Qwen3's most agentic code model to date."
|
|
@@ -2314,8 +2023,6 @@ var fireworksModels = {
|
|
|
2314
2023
|
contextWindow: 16e4,
|
|
2315
2024
|
supportsImages: false,
|
|
2316
2025
|
supportsPromptCache: false,
|
|
2317
|
-
supportsNativeTools: true,
|
|
2318
|
-
defaultToolProtocol: "native",
|
|
2319
2026
|
inputPrice: 3,
|
|
2320
2027
|
outputPrice: 8,
|
|
2321
2028
|
description: "05/28 updated checkpoint of Deepseek R1. Its overall performance is now approaching that of leading models, such as O3 and Gemini 2.5 Pro. Compared to the previous version, the upgraded model shows significant improvements in handling complex reasoning tasks, and this version also offers a reduced hallucination rate, enhanced support for function calling, and better experience for vibe coding. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
|
|
@@ -2325,8 +2032,6 @@ var fireworksModels = {
|
|
|
2325
2032
|
contextWindow: 128e3,
|
|
2326
2033
|
supportsImages: false,
|
|
2327
2034
|
supportsPromptCache: false,
|
|
2328
|
-
supportsNativeTools: true,
|
|
2329
|
-
defaultToolProtocol: "native",
|
|
2330
2035
|
inputPrice: 0.9,
|
|
2331
2036
|
outputPrice: 0.9,
|
|
2332
2037
|
description: "A strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token from Deepseek. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
|
|
@@ -2336,8 +2041,6 @@ var fireworksModels = {
|
|
|
2336
2041
|
contextWindow: 163840,
|
|
2337
2042
|
supportsImages: false,
|
|
2338
2043
|
supportsPromptCache: false,
|
|
2339
|
-
supportsNativeTools: true,
|
|
2340
|
-
defaultToolProtocol: "native",
|
|
2341
2044
|
inputPrice: 0.56,
|
|
2342
2045
|
outputPrice: 1.68,
|
|
2343
2046
|
description: "DeepSeek v3.1 is an improved version of the v3 model with enhanced performance, better reasoning capabilities, and improved code generation. This Mixture-of-Experts (MoE) model maintains the same 671B total parameters with 37B activated per token."
|
|
@@ -2347,8 +2050,6 @@ var fireworksModels = {
|
|
|
2347
2050
|
contextWindow: 128e3,
|
|
2348
2051
|
supportsImages: false,
|
|
2349
2052
|
supportsPromptCache: false,
|
|
2350
|
-
supportsNativeTools: true,
|
|
2351
|
-
defaultToolProtocol: "native",
|
|
2352
2053
|
inputPrice: 0.55,
|
|
2353
2054
|
outputPrice: 2.19,
|
|
2354
2055
|
description: "Z.ai GLM-4.5 with 355B total parameters and 32B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
|
|
@@ -2358,8 +2059,6 @@ var fireworksModels = {
|
|
|
2358
2059
|
contextWindow: 128e3,
|
|
2359
2060
|
supportsImages: false,
|
|
2360
2061
|
supportsPromptCache: false,
|
|
2361
|
-
supportsNativeTools: true,
|
|
2362
|
-
defaultToolProtocol: "native",
|
|
2363
2062
|
inputPrice: 0.55,
|
|
2364
2063
|
outputPrice: 2.19,
|
|
2365
2064
|
description: "Z.ai GLM-4.5-Air with 106B total parameters and 12B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
|
|
@@ -2369,8 +2068,6 @@ var fireworksModels = {
|
|
|
2369
2068
|
contextWindow: 198e3,
|
|
2370
2069
|
supportsImages: false,
|
|
2371
2070
|
supportsPromptCache: false,
|
|
2372
|
-
supportsNativeTools: true,
|
|
2373
|
-
defaultToolProtocol: "native",
|
|
2374
2071
|
inputPrice: 0.55,
|
|
2375
2072
|
outputPrice: 2.19,
|
|
2376
2073
|
description: "Z.ai GLM-4.6 is an advanced coding model with exceptional performance on complex programming tasks. Features improved reasoning capabilities and enhanced code generation quality, making it ideal for software development workflows."
|
|
@@ -2380,8 +2077,6 @@ var fireworksModels = {
|
|
|
2380
2077
|
contextWindow: 128e3,
|
|
2381
2078
|
supportsImages: false,
|
|
2382
2079
|
supportsPromptCache: false,
|
|
2383
|
-
supportsNativeTools: true,
|
|
2384
|
-
defaultToolProtocol: "native",
|
|
2385
2080
|
inputPrice: 0.07,
|
|
2386
2081
|
outputPrice: 0.3,
|
|
2387
2082
|
description: "OpenAI gpt-oss-20b: Compact model for local/edge deployments. Optimized for low-latency and resource-constrained environments with chain-of-thought output, adjustable reasoning, and agentic workflows."
|
|
@@ -2391,11 +2086,63 @@ var fireworksModels = {
|
|
|
2391
2086
|
contextWindow: 128e3,
|
|
2392
2087
|
supportsImages: false,
|
|
2393
2088
|
supportsPromptCache: false,
|
|
2394
|
-
supportsNativeTools: true,
|
|
2395
|
-
defaultToolProtocol: "native",
|
|
2396
2089
|
inputPrice: 0.15,
|
|
2397
2090
|
outputPrice: 0.6,
|
|
2398
2091
|
description: "OpenAI gpt-oss-120b: Production-grade, general-purpose model that fits on a single H100 GPU. Features complex reasoning, configurable effort, full chain-of-thought transparency, and supports function calling, tool use, and structured outputs."
|
|
2092
|
+
},
|
|
2093
|
+
"accounts/fireworks/models/minimax-m2p1": {
|
|
2094
|
+
maxTokens: 4096,
|
|
2095
|
+
contextWindow: 204800,
|
|
2096
|
+
supportsImages: false,
|
|
2097
|
+
supportsPromptCache: false,
|
|
2098
|
+
inputPrice: 0.3,
|
|
2099
|
+
outputPrice: 1.2,
|
|
2100
|
+
description: "MiniMax M2.1 is an upgraded version of M2 with improved performance on complex reasoning, coding, and long-context understanding tasks."
|
|
2101
|
+
},
|
|
2102
|
+
"accounts/fireworks/models/deepseek-v3p2": {
|
|
2103
|
+
maxTokens: 16384,
|
|
2104
|
+
contextWindow: 163840,
|
|
2105
|
+
supportsImages: false,
|
|
2106
|
+
supportsPromptCache: false,
|
|
2107
|
+
inputPrice: 0.56,
|
|
2108
|
+
outputPrice: 1.68,
|
|
2109
|
+
description: "DeepSeek V3.2 is the latest iteration of the V3 model family with enhanced reasoning capabilities, improved code generation, and better instruction following."
|
|
2110
|
+
},
|
|
2111
|
+
"accounts/fireworks/models/glm-4p7": {
|
|
2112
|
+
maxTokens: 25344,
|
|
2113
|
+
contextWindow: 198e3,
|
|
2114
|
+
supportsImages: false,
|
|
2115
|
+
supportsPromptCache: false,
|
|
2116
|
+
inputPrice: 0.55,
|
|
2117
|
+
outputPrice: 2.19,
|
|
2118
|
+
description: "Z.ai GLM-4.7 is the latest coding model with exceptional performance on complex programming tasks. Features improved reasoning capabilities and enhanced code generation quality."
|
|
2119
|
+
},
|
|
2120
|
+
"accounts/fireworks/models/llama-v3p3-70b-instruct": {
|
|
2121
|
+
maxTokens: 16384,
|
|
2122
|
+
contextWindow: 131072,
|
|
2123
|
+
supportsImages: false,
|
|
2124
|
+
supportsPromptCache: false,
|
|
2125
|
+
inputPrice: 0.9,
|
|
2126
|
+
outputPrice: 0.9,
|
|
2127
|
+
description: "Meta Llama 3.3 70B Instruct is a highly capable instruction-tuned model with strong reasoning, coding, and general task performance."
|
|
2128
|
+
},
|
|
2129
|
+
"accounts/fireworks/models/llama4-maverick-instruct-basic": {
|
|
2130
|
+
maxTokens: 16384,
|
|
2131
|
+
contextWindow: 131072,
|
|
2132
|
+
supportsImages: true,
|
|
2133
|
+
supportsPromptCache: false,
|
|
2134
|
+
inputPrice: 0.22,
|
|
2135
|
+
outputPrice: 0.88,
|
|
2136
|
+
description: "Llama 4 Maverick is Meta's latest multimodal model with vision capabilities, optimized for instruction following and coding tasks."
|
|
2137
|
+
},
|
|
2138
|
+
"accounts/fireworks/models/llama4-scout-instruct-basic": {
|
|
2139
|
+
maxTokens: 16384,
|
|
2140
|
+
contextWindow: 131072,
|
|
2141
|
+
supportsImages: true,
|
|
2142
|
+
supportsPromptCache: false,
|
|
2143
|
+
inputPrice: 0.15,
|
|
2144
|
+
outputPrice: 0.6,
|
|
2145
|
+
description: "Llama 4 Scout is a smaller, faster variant of Llama 4 with multimodal capabilities, ideal for quick iterations and cost-effective deployments."
|
|
2399
2146
|
}
|
|
2400
2147
|
};
|
|
2401
2148
|
|
|
@@ -2406,8 +2153,6 @@ var geminiModels = {
|
|
|
2406
2153
|
maxTokens: 65536,
|
|
2407
2154
|
contextWindow: 1048576,
|
|
2408
2155
|
supportsImages: true,
|
|
2409
|
-
supportsNativeTools: true,
|
|
2410
|
-
defaultToolProtocol: "native",
|
|
2411
2156
|
supportsPromptCache: true,
|
|
2412
2157
|
supportsReasoningEffort: ["low", "high"],
|
|
2413
2158
|
reasoningEffort: "low",
|
|
@@ -2415,16 +2160,19 @@ var geminiModels = {
|
|
|
2415
2160
|
defaultTemperature: 1,
|
|
2416
2161
|
inputPrice: 4,
|
|
2417
2162
|
outputPrice: 18,
|
|
2163
|
+
cacheReadsPrice: 0.4,
|
|
2418
2164
|
tiers: [
|
|
2419
2165
|
{
|
|
2420
2166
|
contextWindow: 2e5,
|
|
2421
2167
|
inputPrice: 2,
|
|
2422
|
-
outputPrice: 12
|
|
2168
|
+
outputPrice: 12,
|
|
2169
|
+
cacheReadsPrice: 0.2
|
|
2423
2170
|
},
|
|
2424
2171
|
{
|
|
2425
2172
|
contextWindow: Infinity,
|
|
2426
2173
|
inputPrice: 4,
|
|
2427
|
-
outputPrice: 18
|
|
2174
|
+
outputPrice: 18,
|
|
2175
|
+
cacheReadsPrice: 0.4
|
|
2428
2176
|
}
|
|
2429
2177
|
]
|
|
2430
2178
|
},
|
|
@@ -2432,25 +2180,20 @@ var geminiModels = {
|
|
|
2432
2180
|
maxTokens: 65536,
|
|
2433
2181
|
contextWindow: 1048576,
|
|
2434
2182
|
supportsImages: true,
|
|
2435
|
-
supportsNativeTools: true,
|
|
2436
|
-
defaultToolProtocol: "native",
|
|
2437
2183
|
supportsPromptCache: true,
|
|
2438
2184
|
supportsReasoningEffort: ["minimal", "low", "medium", "high"],
|
|
2439
2185
|
reasoningEffort: "medium",
|
|
2440
2186
|
supportsTemperature: true,
|
|
2441
2187
|
defaultTemperature: 1,
|
|
2442
|
-
inputPrice: 0.
|
|
2443
|
-
outputPrice:
|
|
2444
|
-
cacheReadsPrice: 0.
|
|
2445
|
-
cacheWritesPrice: 1
|
|
2188
|
+
inputPrice: 0.5,
|
|
2189
|
+
outputPrice: 3,
|
|
2190
|
+
cacheReadsPrice: 0.05
|
|
2446
2191
|
},
|
|
2447
2192
|
// 2.5 Pro models
|
|
2448
2193
|
"gemini-2.5-pro": {
|
|
2449
2194
|
maxTokens: 64e3,
|
|
2450
2195
|
contextWindow: 1048576,
|
|
2451
2196
|
supportsImages: true,
|
|
2452
|
-
supportsNativeTools: true,
|
|
2453
|
-
defaultToolProtocol: "native",
|
|
2454
2197
|
supportsPromptCache: true,
|
|
2455
2198
|
inputPrice: 2.5,
|
|
2456
2199
|
// This is the pricing for prompts above 200k tokens.
|
|
@@ -2479,8 +2222,6 @@ var geminiModels = {
|
|
|
2479
2222
|
maxTokens: 65535,
|
|
2480
2223
|
contextWindow: 1048576,
|
|
2481
2224
|
supportsImages: true,
|
|
2482
|
-
supportsNativeTools: true,
|
|
2483
|
-
defaultToolProtocol: "native",
|
|
2484
2225
|
supportsPromptCache: true,
|
|
2485
2226
|
inputPrice: 2.5,
|
|
2486
2227
|
// This is the pricing for prompts above 200k tokens.
|
|
@@ -2508,8 +2249,6 @@ var geminiModels = {
|
|
|
2508
2249
|
maxTokens: 65535,
|
|
2509
2250
|
contextWindow: 1048576,
|
|
2510
2251
|
supportsImages: true,
|
|
2511
|
-
supportsNativeTools: true,
|
|
2512
|
-
defaultToolProtocol: "native",
|
|
2513
2252
|
supportsPromptCache: true,
|
|
2514
2253
|
inputPrice: 2.5,
|
|
2515
2254
|
// This is the pricing for prompts above 200k tokens.
|
|
@@ -2535,8 +2274,6 @@ var geminiModels = {
|
|
|
2535
2274
|
maxTokens: 65535,
|
|
2536
2275
|
contextWindow: 1048576,
|
|
2537
2276
|
supportsImages: true,
|
|
2538
|
-
supportsNativeTools: true,
|
|
2539
|
-
defaultToolProtocol: "native",
|
|
2540
2277
|
supportsPromptCache: true,
|
|
2541
2278
|
inputPrice: 2.5,
|
|
2542
2279
|
// This is the pricing for prompts above 200k tokens.
|
|
@@ -2565,8 +2302,6 @@ var geminiModels = {
|
|
|
2565
2302
|
maxTokens: 65536,
|
|
2566
2303
|
contextWindow: 1048576,
|
|
2567
2304
|
supportsImages: true,
|
|
2568
|
-
supportsNativeTools: true,
|
|
2569
|
-
defaultToolProtocol: "native",
|
|
2570
2305
|
supportsPromptCache: true,
|
|
2571
2306
|
inputPrice: 0.3,
|
|
2572
2307
|
outputPrice: 2.5,
|
|
@@ -2579,8 +2314,6 @@ var geminiModels = {
|
|
|
2579
2314
|
maxTokens: 65536,
|
|
2580
2315
|
contextWindow: 1048576,
|
|
2581
2316
|
supportsImages: true,
|
|
2582
|
-
supportsNativeTools: true,
|
|
2583
|
-
defaultToolProtocol: "native",
|
|
2584
2317
|
supportsPromptCache: true,
|
|
2585
2318
|
inputPrice: 0.3,
|
|
2586
2319
|
outputPrice: 2.5,
|
|
@@ -2593,8 +2326,6 @@ var geminiModels = {
|
|
|
2593
2326
|
maxTokens: 64e3,
|
|
2594
2327
|
contextWindow: 1048576,
|
|
2595
2328
|
supportsImages: true,
|
|
2596
|
-
supportsNativeTools: true,
|
|
2597
|
-
defaultToolProtocol: "native",
|
|
2598
2329
|
supportsPromptCache: true,
|
|
2599
2330
|
inputPrice: 0.3,
|
|
2600
2331
|
outputPrice: 2.5,
|
|
@@ -2608,8 +2339,6 @@ var geminiModels = {
|
|
|
2608
2339
|
maxTokens: 65536,
|
|
2609
2340
|
contextWindow: 1048576,
|
|
2610
2341
|
supportsImages: true,
|
|
2611
|
-
supportsNativeTools: true,
|
|
2612
|
-
defaultToolProtocol: "native",
|
|
2613
2342
|
supportsPromptCache: true,
|
|
2614
2343
|
inputPrice: 0.1,
|
|
2615
2344
|
outputPrice: 0.4,
|
|
@@ -2622,8 +2351,6 @@ var geminiModels = {
|
|
|
2622
2351
|
maxTokens: 65536,
|
|
2623
2352
|
contextWindow: 1048576,
|
|
2624
2353
|
supportsImages: true,
|
|
2625
|
-
supportsNativeTools: true,
|
|
2626
|
-
defaultToolProtocol: "native",
|
|
2627
2354
|
supportsPromptCache: true,
|
|
2628
2355
|
inputPrice: 0.1,
|
|
2629
2356
|
outputPrice: 0.4,
|
|
@@ -2643,8 +2370,6 @@ var groqModels = {
|
|
|
2643
2370
|
contextWindow: 131072,
|
|
2644
2371
|
supportsImages: false,
|
|
2645
2372
|
supportsPromptCache: false,
|
|
2646
|
-
supportsNativeTools: true,
|
|
2647
|
-
defaultToolProtocol: "native",
|
|
2648
2373
|
inputPrice: 0.05,
|
|
2649
2374
|
outputPrice: 0.08,
|
|
2650
2375
|
description: "Meta Llama 3.1 8B Instant model, 128K context."
|
|
@@ -2654,8 +2379,6 @@ var groqModels = {
|
|
|
2654
2379
|
contextWindow: 131072,
|
|
2655
2380
|
supportsImages: false,
|
|
2656
2381
|
supportsPromptCache: false,
|
|
2657
|
-
supportsNativeTools: true,
|
|
2658
|
-
defaultToolProtocol: "native",
|
|
2659
2382
|
inputPrice: 0.59,
|
|
2660
2383
|
outputPrice: 0.79,
|
|
2661
2384
|
description: "Meta Llama 3.3 70B Versatile model, 128K context."
|
|
@@ -2665,8 +2388,6 @@ var groqModels = {
|
|
|
2665
2388
|
contextWindow: 131072,
|
|
2666
2389
|
supportsImages: false,
|
|
2667
2390
|
supportsPromptCache: false,
|
|
2668
|
-
supportsNativeTools: true,
|
|
2669
|
-
defaultToolProtocol: "native",
|
|
2670
2391
|
inputPrice: 0.11,
|
|
2671
2392
|
outputPrice: 0.34,
|
|
2672
2393
|
description: "Meta Llama 4 Scout 17B Instruct model, 128K context."
|
|
@@ -2676,8 +2397,6 @@ var groqModels = {
|
|
|
2676
2397
|
contextWindow: 131072,
|
|
2677
2398
|
supportsImages: false,
|
|
2678
2399
|
supportsPromptCache: false,
|
|
2679
|
-
supportsNativeTools: true,
|
|
2680
|
-
defaultToolProtocol: "native",
|
|
2681
2400
|
inputPrice: 0.29,
|
|
2682
2401
|
outputPrice: 0.59,
|
|
2683
2402
|
description: "Alibaba Qwen 3 32B model, 128K context."
|
|
@@ -2687,8 +2406,6 @@ var groqModels = {
|
|
|
2687
2406
|
contextWindow: 262144,
|
|
2688
2407
|
supportsImages: false,
|
|
2689
2408
|
supportsPromptCache: true,
|
|
2690
|
-
supportsNativeTools: true,
|
|
2691
|
-
defaultToolProtocol: "native",
|
|
2692
2409
|
inputPrice: 0.6,
|
|
2693
2410
|
outputPrice: 2.5,
|
|
2694
2411
|
cacheReadsPrice: 0.15,
|
|
@@ -2699,8 +2416,6 @@ var groqModels = {
|
|
|
2699
2416
|
contextWindow: 131072,
|
|
2700
2417
|
supportsImages: false,
|
|
2701
2418
|
supportsPromptCache: false,
|
|
2702
|
-
supportsNativeTools: true,
|
|
2703
|
-
defaultToolProtocol: "native",
|
|
2704
2419
|
inputPrice: 0.15,
|
|
2705
2420
|
outputPrice: 0.75,
|
|
2706
2421
|
description: "GPT-OSS 120B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 128 experts."
|
|
@@ -2710,8 +2425,6 @@ var groqModels = {
|
|
|
2710
2425
|
contextWindow: 131072,
|
|
2711
2426
|
supportsImages: false,
|
|
2712
2427
|
supportsPromptCache: false,
|
|
2713
|
-
supportsNativeTools: true,
|
|
2714
|
-
defaultToolProtocol: "native",
|
|
2715
2428
|
inputPrice: 0.1,
|
|
2716
2429
|
outputPrice: 0.5,
|
|
2717
2430
|
description: "GPT-OSS 20B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 32 experts."
|
|
@@ -2738,7 +2451,6 @@ var ioIntelligenceModels = {
|
|
|
2738
2451
|
contextWindow: 128e3,
|
|
2739
2452
|
supportsImages: false,
|
|
2740
2453
|
supportsPromptCache: false,
|
|
2741
|
-
supportsNativeTools: true,
|
|
2742
2454
|
description: "DeepSeek R1 reasoning model"
|
|
2743
2455
|
},
|
|
2744
2456
|
"meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8": {
|
|
@@ -2746,7 +2458,6 @@ var ioIntelligenceModels = {
|
|
|
2746
2458
|
contextWindow: 43e4,
|
|
2747
2459
|
supportsImages: true,
|
|
2748
2460
|
supportsPromptCache: false,
|
|
2749
|
-
supportsNativeTools: true,
|
|
2750
2461
|
description: "Llama 4 Maverick 17B model"
|
|
2751
2462
|
},
|
|
2752
2463
|
"Intel/Qwen3-Coder-480B-A35B-Instruct-int4-mixed-ar": {
|
|
@@ -2754,7 +2465,6 @@ var ioIntelligenceModels = {
|
|
|
2754
2465
|
contextWindow: 106e3,
|
|
2755
2466
|
supportsImages: false,
|
|
2756
2467
|
supportsPromptCache: false,
|
|
2757
|
-
supportsNativeTools: true,
|
|
2758
2468
|
description: "Qwen3 Coder 480B specialized for coding"
|
|
2759
2469
|
},
|
|
2760
2470
|
"openai/gpt-oss-120b": {
|
|
@@ -2762,7 +2472,6 @@ var ioIntelligenceModels = {
|
|
|
2762
2472
|
contextWindow: 131072,
|
|
2763
2473
|
supportsImages: false,
|
|
2764
2474
|
supportsPromptCache: false,
|
|
2765
|
-
supportsNativeTools: true,
|
|
2766
2475
|
description: "OpenAI GPT-OSS 120B model"
|
|
2767
2476
|
}
|
|
2768
2477
|
};
|
|
@@ -2774,8 +2483,6 @@ var litellmDefaultModelInfo = {
|
|
|
2774
2483
|
contextWindow: 2e5,
|
|
2775
2484
|
supportsImages: true,
|
|
2776
2485
|
supportsPromptCache: true,
|
|
2777
|
-
supportsNativeTools: true,
|
|
2778
|
-
defaultToolProtocol: "native",
|
|
2779
2486
|
inputPrice: 3,
|
|
2780
2487
|
outputPrice: 15,
|
|
2781
2488
|
cacheWritesPrice: 3.75,
|
|
@@ -2790,8 +2497,6 @@ var lMStudioDefaultModelInfo = {
|
|
|
2790
2497
|
contextWindow: 2e5,
|
|
2791
2498
|
supportsImages: true,
|
|
2792
2499
|
supportsPromptCache: true,
|
|
2793
|
-
supportsNativeTools: true,
|
|
2794
|
-
defaultToolProtocol: "native",
|
|
2795
2500
|
inputPrice: 0,
|
|
2796
2501
|
outputPrice: 0,
|
|
2797
2502
|
cacheWritesPrice: 0,
|
|
@@ -2807,8 +2512,6 @@ var mistralModels = {
|
|
|
2807
2512
|
contextWindow: 128e3,
|
|
2808
2513
|
supportsImages: true,
|
|
2809
2514
|
supportsPromptCache: false,
|
|
2810
|
-
supportsNativeTools: true,
|
|
2811
|
-
defaultToolProtocol: "native",
|
|
2812
2515
|
inputPrice: 2,
|
|
2813
2516
|
outputPrice: 5
|
|
2814
2517
|
},
|
|
@@ -2817,8 +2520,6 @@ var mistralModels = {
|
|
|
2817
2520
|
contextWindow: 131e3,
|
|
2818
2521
|
supportsImages: true,
|
|
2819
2522
|
supportsPromptCache: false,
|
|
2820
|
-
supportsNativeTools: true,
|
|
2821
|
-
defaultToolProtocol: "native",
|
|
2822
2523
|
inputPrice: 0.4,
|
|
2823
2524
|
outputPrice: 2
|
|
2824
2525
|
},
|
|
@@ -2827,8 +2528,6 @@ var mistralModels = {
|
|
|
2827
2528
|
contextWindow: 131e3,
|
|
2828
2529
|
supportsImages: true,
|
|
2829
2530
|
supportsPromptCache: false,
|
|
2830
|
-
supportsNativeTools: true,
|
|
2831
|
-
defaultToolProtocol: "native",
|
|
2832
2531
|
inputPrice: 0.4,
|
|
2833
2532
|
outputPrice: 2
|
|
2834
2533
|
},
|
|
@@ -2837,8 +2536,6 @@ var mistralModels = {
|
|
|
2837
2536
|
contextWindow: 256e3,
|
|
2838
2537
|
supportsImages: false,
|
|
2839
2538
|
supportsPromptCache: false,
|
|
2840
|
-
supportsNativeTools: true,
|
|
2841
|
-
defaultToolProtocol: "native",
|
|
2842
2539
|
inputPrice: 0.3,
|
|
2843
2540
|
outputPrice: 0.9
|
|
2844
2541
|
},
|
|
@@ -2847,8 +2544,6 @@ var mistralModels = {
|
|
|
2847
2544
|
contextWindow: 131e3,
|
|
2848
2545
|
supportsImages: false,
|
|
2849
2546
|
supportsPromptCache: false,
|
|
2850
|
-
supportsNativeTools: true,
|
|
2851
|
-
defaultToolProtocol: "native",
|
|
2852
2547
|
inputPrice: 2,
|
|
2853
2548
|
outputPrice: 6
|
|
2854
2549
|
},
|
|
@@ -2857,8 +2552,6 @@ var mistralModels = {
|
|
|
2857
2552
|
contextWindow: 131e3,
|
|
2858
2553
|
supportsImages: false,
|
|
2859
2554
|
supportsPromptCache: false,
|
|
2860
|
-
supportsNativeTools: true,
|
|
2861
|
-
defaultToolProtocol: "native",
|
|
2862
2555
|
inputPrice: 0.1,
|
|
2863
2556
|
outputPrice: 0.1
|
|
2864
2557
|
},
|
|
@@ -2867,8 +2560,6 @@ var mistralModels = {
|
|
|
2867
2560
|
contextWindow: 131e3,
|
|
2868
2561
|
supportsImages: false,
|
|
2869
2562
|
supportsPromptCache: false,
|
|
2870
|
-
supportsNativeTools: true,
|
|
2871
|
-
defaultToolProtocol: "native",
|
|
2872
2563
|
inputPrice: 0.04,
|
|
2873
2564
|
outputPrice: 0.04
|
|
2874
2565
|
},
|
|
@@ -2877,8 +2568,6 @@ var mistralModels = {
|
|
|
2877
2568
|
contextWindow: 32e3,
|
|
2878
2569
|
supportsImages: false,
|
|
2879
2570
|
supportsPromptCache: false,
|
|
2880
|
-
supportsNativeTools: true,
|
|
2881
|
-
defaultToolProtocol: "native",
|
|
2882
2571
|
inputPrice: 0.2,
|
|
2883
2572
|
outputPrice: 0.6
|
|
2884
2573
|
},
|
|
@@ -2887,8 +2576,6 @@ var mistralModels = {
|
|
|
2887
2576
|
contextWindow: 131e3,
|
|
2888
2577
|
supportsImages: true,
|
|
2889
2578
|
supportsPromptCache: false,
|
|
2890
|
-
supportsNativeTools: true,
|
|
2891
|
-
defaultToolProtocol: "native",
|
|
2892
2579
|
inputPrice: 2,
|
|
2893
2580
|
outputPrice: 6
|
|
2894
2581
|
}
|
|
@@ -2903,8 +2590,6 @@ var moonshotModels = {
|
|
|
2903
2590
|
contextWindow: 131072,
|
|
2904
2591
|
supportsImages: false,
|
|
2905
2592
|
supportsPromptCache: true,
|
|
2906
|
-
supportsNativeTools: true,
|
|
2907
|
-
defaultToolProtocol: "native",
|
|
2908
2593
|
inputPrice: 0.6,
|
|
2909
2594
|
// $0.60 per million tokens (cache miss)
|
|
2910
2595
|
outputPrice: 2.5,
|
|
@@ -2920,8 +2605,6 @@ var moonshotModels = {
|
|
|
2920
2605
|
contextWindow: 262144,
|
|
2921
2606
|
supportsImages: false,
|
|
2922
2607
|
supportsPromptCache: true,
|
|
2923
|
-
supportsNativeTools: true,
|
|
2924
|
-
defaultToolProtocol: "native",
|
|
2925
2608
|
inputPrice: 0.6,
|
|
2926
2609
|
outputPrice: 2.5,
|
|
2927
2610
|
cacheReadsPrice: 0.15,
|
|
@@ -2932,8 +2615,6 @@ var moonshotModels = {
|
|
|
2932
2615
|
contextWindow: 262144,
|
|
2933
2616
|
supportsImages: false,
|
|
2934
2617
|
supportsPromptCache: true,
|
|
2935
|
-
supportsNativeTools: true,
|
|
2936
|
-
defaultToolProtocol: "native",
|
|
2937
2618
|
inputPrice: 2.4,
|
|
2938
2619
|
// $2.40 per million tokens (cache miss)
|
|
2939
2620
|
outputPrice: 10,
|
|
@@ -2952,8 +2633,6 @@ var moonshotModels = {
|
|
|
2952
2633
|
supportsImages: false,
|
|
2953
2634
|
// Text-only (no image/vision support)
|
|
2954
2635
|
supportsPromptCache: true,
|
|
2955
|
-
supportsNativeTools: true,
|
|
2956
|
-
defaultToolProtocol: "native",
|
|
2957
2636
|
inputPrice: 0.6,
|
|
2958
2637
|
// $0.60 per million tokens (cache miss)
|
|
2959
2638
|
outputPrice: 2.5,
|
|
@@ -2967,6 +2646,21 @@ var moonshotModels = {
|
|
|
2967
2646
|
preserveReasoning: true,
|
|
2968
2647
|
defaultTemperature: 1,
|
|
2969
2648
|
description: `The kimi-k2-thinking model is a general-purpose agentic reasoning model developed by Moonshot AI. Thanks to its strength in deep reasoning and multi-turn tool use, it can solve even the hardest problems.`
|
|
2649
|
+
},
|
|
2650
|
+
"kimi-k2.5": {
|
|
2651
|
+
maxTokens: 16384,
|
|
2652
|
+
contextWindow: 262144,
|
|
2653
|
+
supportsImages: false,
|
|
2654
|
+
supportsPromptCache: true,
|
|
2655
|
+
inputPrice: 0.6,
|
|
2656
|
+
// $0.60 per million tokens (cache miss)
|
|
2657
|
+
outputPrice: 3,
|
|
2658
|
+
// $3.00 per million tokens
|
|
2659
|
+
cacheReadsPrice: 0.1,
|
|
2660
|
+
// $0.10 per million tokens (cache hit)
|
|
2661
|
+
supportsTemperature: true,
|
|
2662
|
+
defaultTemperature: 1,
|
|
2663
|
+
description: "Kimi K2.5 is the latest generation of Moonshot AI's Kimi series, featuring improved reasoning capabilities and enhanced performance across diverse tasks."
|
|
2970
2664
|
}
|
|
2971
2665
|
};
|
|
2972
2666
|
var MOONSHOT_DEFAULT_TEMPERATURE = 0.6;
|
|
@@ -2978,7 +2672,6 @@ var ollamaDefaultModelInfo = {
|
|
|
2978
2672
|
contextWindow: 2e5,
|
|
2979
2673
|
supportsImages: true,
|
|
2980
2674
|
supportsPromptCache: true,
|
|
2981
|
-
supportsNativeTools: true,
|
|
2982
2675
|
inputPrice: 0,
|
|
2983
2676
|
outputPrice: 0,
|
|
2984
2677
|
cacheWritesPrice: 0,
|
|
@@ -2992,8 +2685,6 @@ var openAiNativeModels = {
|
|
|
2992
2685
|
"gpt-5.1-codex-max": {
|
|
2993
2686
|
maxTokens: 128e3,
|
|
2994
2687
|
contextWindow: 4e5,
|
|
2995
|
-
supportsNativeTools: true,
|
|
2996
|
-
defaultToolProtocol: "native",
|
|
2997
2688
|
includedTools: ["apply_patch"],
|
|
2998
2689
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
2999
2690
|
supportsImages: true,
|
|
@@ -3011,8 +2702,6 @@ var openAiNativeModels = {
|
|
|
3011
2702
|
"gpt-5.2": {
|
|
3012
2703
|
maxTokens: 128e3,
|
|
3013
2704
|
contextWindow: 4e5,
|
|
3014
|
-
supportsNativeTools: true,
|
|
3015
|
-
defaultToolProtocol: "native",
|
|
3016
2705
|
includedTools: ["apply_patch"],
|
|
3017
2706
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3018
2707
|
supportsImages: true,
|
|
@@ -3031,11 +2720,26 @@ var openAiNativeModels = {
|
|
|
3031
2720
|
],
|
|
3032
2721
|
description: "GPT-5.2: Our flagship model for coding and agentic tasks across industries"
|
|
3033
2722
|
},
|
|
2723
|
+
"gpt-5.2-codex": {
|
|
2724
|
+
maxTokens: 128e3,
|
|
2725
|
+
contextWindow: 4e5,
|
|
2726
|
+
includedTools: ["apply_patch"],
|
|
2727
|
+
excludedTools: ["apply_diff", "write_to_file"],
|
|
2728
|
+
supportsImages: true,
|
|
2729
|
+
supportsPromptCache: true,
|
|
2730
|
+
promptCacheRetention: "24h",
|
|
2731
|
+
supportsReasoningEffort: ["low", "medium", "high", "xhigh"],
|
|
2732
|
+
reasoningEffort: "medium",
|
|
2733
|
+
inputPrice: 1.75,
|
|
2734
|
+
outputPrice: 14,
|
|
2735
|
+
cacheReadsPrice: 0.175,
|
|
2736
|
+
supportsTemperature: false,
|
|
2737
|
+
tiers: [{ name: "priority", contextWindow: 4e5, inputPrice: 3.5, outputPrice: 28, cacheReadsPrice: 0.35 }],
|
|
2738
|
+
description: "GPT-5.2 Codex: Our most intelligent coding model optimized for long-horizon, agentic coding tasks"
|
|
2739
|
+
},
|
|
3034
2740
|
"gpt-5.2-chat-latest": {
|
|
3035
2741
|
maxTokens: 16384,
|
|
3036
2742
|
contextWindow: 128e3,
|
|
3037
|
-
supportsNativeTools: true,
|
|
3038
|
-
defaultToolProtocol: "native",
|
|
3039
2743
|
includedTools: ["apply_patch"],
|
|
3040
2744
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3041
2745
|
supportsImages: true,
|
|
@@ -3048,8 +2752,6 @@ var openAiNativeModels = {
|
|
|
3048
2752
|
"gpt-5.1": {
|
|
3049
2753
|
maxTokens: 128e3,
|
|
3050
2754
|
contextWindow: 4e5,
|
|
3051
|
-
supportsNativeTools: true,
|
|
3052
|
-
defaultToolProtocol: "native",
|
|
3053
2755
|
includedTools: ["apply_patch"],
|
|
3054
2756
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3055
2757
|
supportsImages: true,
|
|
@@ -3071,8 +2773,6 @@ var openAiNativeModels = {
|
|
|
3071
2773
|
"gpt-5.1-codex": {
|
|
3072
2774
|
maxTokens: 128e3,
|
|
3073
2775
|
contextWindow: 4e5,
|
|
3074
|
-
supportsNativeTools: true,
|
|
3075
|
-
defaultToolProtocol: "native",
|
|
3076
2776
|
includedTools: ["apply_patch"],
|
|
3077
2777
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3078
2778
|
supportsImages: true,
|
|
@@ -3090,8 +2790,6 @@ var openAiNativeModels = {
|
|
|
3090
2790
|
"gpt-5.1-codex-mini": {
|
|
3091
2791
|
maxTokens: 128e3,
|
|
3092
2792
|
contextWindow: 4e5,
|
|
3093
|
-
supportsNativeTools: true,
|
|
3094
|
-
defaultToolProtocol: "native",
|
|
3095
2793
|
includedTools: ["apply_patch"],
|
|
3096
2794
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3097
2795
|
supportsImages: true,
|
|
@@ -3108,8 +2806,6 @@ var openAiNativeModels = {
|
|
|
3108
2806
|
"gpt-5": {
|
|
3109
2807
|
maxTokens: 128e3,
|
|
3110
2808
|
contextWindow: 4e5,
|
|
3111
|
-
supportsNativeTools: true,
|
|
3112
|
-
defaultToolProtocol: "native",
|
|
3113
2809
|
includedTools: ["apply_patch"],
|
|
3114
2810
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3115
2811
|
supportsImages: true,
|
|
@@ -3130,8 +2826,6 @@ var openAiNativeModels = {
|
|
|
3130
2826
|
"gpt-5-mini": {
|
|
3131
2827
|
maxTokens: 128e3,
|
|
3132
2828
|
contextWindow: 4e5,
|
|
3133
|
-
supportsNativeTools: true,
|
|
3134
|
-
defaultToolProtocol: "native",
|
|
3135
2829
|
includedTools: ["apply_patch"],
|
|
3136
2830
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3137
2831
|
supportsImages: true,
|
|
@@ -3152,8 +2846,6 @@ var openAiNativeModels = {
|
|
|
3152
2846
|
"gpt-5-codex": {
|
|
3153
2847
|
maxTokens: 128e3,
|
|
3154
2848
|
contextWindow: 4e5,
|
|
3155
|
-
supportsNativeTools: true,
|
|
3156
|
-
defaultToolProtocol: "native",
|
|
3157
2849
|
includedTools: ["apply_patch"],
|
|
3158
2850
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3159
2851
|
supportsImages: true,
|
|
@@ -3170,8 +2862,6 @@ var openAiNativeModels = {
|
|
|
3170
2862
|
"gpt-5-nano": {
|
|
3171
2863
|
maxTokens: 128e3,
|
|
3172
2864
|
contextWindow: 4e5,
|
|
3173
|
-
supportsNativeTools: true,
|
|
3174
|
-
defaultToolProtocol: "native",
|
|
3175
2865
|
includedTools: ["apply_patch"],
|
|
3176
2866
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3177
2867
|
supportsImages: true,
|
|
@@ -3189,8 +2879,6 @@ var openAiNativeModels = {
|
|
|
3189
2879
|
"gpt-5-chat-latest": {
|
|
3190
2880
|
maxTokens: 128e3,
|
|
3191
2881
|
contextWindow: 4e5,
|
|
3192
|
-
supportsNativeTools: true,
|
|
3193
|
-
defaultToolProtocol: "native",
|
|
3194
2882
|
includedTools: ["apply_patch"],
|
|
3195
2883
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3196
2884
|
supportsImages: true,
|
|
@@ -3203,8 +2891,6 @@ var openAiNativeModels = {
|
|
|
3203
2891
|
"gpt-4.1": {
|
|
3204
2892
|
maxTokens: 32768,
|
|
3205
2893
|
contextWindow: 1047576,
|
|
3206
|
-
supportsNativeTools: true,
|
|
3207
|
-
defaultToolProtocol: "native",
|
|
3208
2894
|
includedTools: ["apply_patch"],
|
|
3209
2895
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3210
2896
|
supportsImages: true,
|
|
@@ -3220,8 +2906,6 @@ var openAiNativeModels = {
|
|
|
3220
2906
|
"gpt-4.1-mini": {
|
|
3221
2907
|
maxTokens: 32768,
|
|
3222
2908
|
contextWindow: 1047576,
|
|
3223
|
-
supportsNativeTools: true,
|
|
3224
|
-
defaultToolProtocol: "native",
|
|
3225
2909
|
includedTools: ["apply_patch"],
|
|
3226
2910
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3227
2911
|
supportsImages: true,
|
|
@@ -3237,8 +2921,6 @@ var openAiNativeModels = {
|
|
|
3237
2921
|
"gpt-4.1-nano": {
|
|
3238
2922
|
maxTokens: 32768,
|
|
3239
2923
|
contextWindow: 1047576,
|
|
3240
|
-
supportsNativeTools: true,
|
|
3241
|
-
defaultToolProtocol: "native",
|
|
3242
2924
|
includedTools: ["apply_patch"],
|
|
3243
2925
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3244
2926
|
supportsImages: true,
|
|
@@ -3254,8 +2936,6 @@ var openAiNativeModels = {
|
|
|
3254
2936
|
o3: {
|
|
3255
2937
|
maxTokens: 1e5,
|
|
3256
2938
|
contextWindow: 2e5,
|
|
3257
|
-
supportsNativeTools: true,
|
|
3258
|
-
defaultToolProtocol: "native",
|
|
3259
2939
|
supportsImages: true,
|
|
3260
2940
|
supportsPromptCache: true,
|
|
3261
2941
|
inputPrice: 2,
|
|
@@ -3272,8 +2952,6 @@ var openAiNativeModels = {
|
|
|
3272
2952
|
"o3-high": {
|
|
3273
2953
|
maxTokens: 1e5,
|
|
3274
2954
|
contextWindow: 2e5,
|
|
3275
|
-
supportsNativeTools: true,
|
|
3276
|
-
defaultToolProtocol: "native",
|
|
3277
2955
|
supportsImages: true,
|
|
3278
2956
|
supportsPromptCache: true,
|
|
3279
2957
|
inputPrice: 2,
|
|
@@ -3285,8 +2963,6 @@ var openAiNativeModels = {
|
|
|
3285
2963
|
"o3-low": {
|
|
3286
2964
|
maxTokens: 1e5,
|
|
3287
2965
|
contextWindow: 2e5,
|
|
3288
|
-
supportsNativeTools: true,
|
|
3289
|
-
defaultToolProtocol: "native",
|
|
3290
2966
|
supportsImages: true,
|
|
3291
2967
|
supportsPromptCache: true,
|
|
3292
2968
|
inputPrice: 2,
|
|
@@ -3298,8 +2974,6 @@ var openAiNativeModels = {
|
|
|
3298
2974
|
"o4-mini": {
|
|
3299
2975
|
maxTokens: 1e5,
|
|
3300
2976
|
contextWindow: 2e5,
|
|
3301
|
-
supportsNativeTools: true,
|
|
3302
|
-
defaultToolProtocol: "native",
|
|
3303
2977
|
supportsImages: true,
|
|
3304
2978
|
supportsPromptCache: true,
|
|
3305
2979
|
inputPrice: 1.1,
|
|
@@ -3316,8 +2990,6 @@ var openAiNativeModels = {
|
|
|
3316
2990
|
"o4-mini-high": {
|
|
3317
2991
|
maxTokens: 1e5,
|
|
3318
2992
|
contextWindow: 2e5,
|
|
3319
|
-
supportsNativeTools: true,
|
|
3320
|
-
defaultToolProtocol: "native",
|
|
3321
2993
|
supportsImages: true,
|
|
3322
2994
|
supportsPromptCache: true,
|
|
3323
2995
|
inputPrice: 1.1,
|
|
@@ -3329,8 +3001,6 @@ var openAiNativeModels = {
|
|
|
3329
3001
|
"o4-mini-low": {
|
|
3330
3002
|
maxTokens: 1e5,
|
|
3331
3003
|
contextWindow: 2e5,
|
|
3332
|
-
supportsNativeTools: true,
|
|
3333
|
-
defaultToolProtocol: "native",
|
|
3334
3004
|
supportsImages: true,
|
|
3335
3005
|
supportsPromptCache: true,
|
|
3336
3006
|
inputPrice: 1.1,
|
|
@@ -3342,8 +3012,6 @@ var openAiNativeModels = {
|
|
|
3342
3012
|
"o3-mini": {
|
|
3343
3013
|
maxTokens: 1e5,
|
|
3344
3014
|
contextWindow: 2e5,
|
|
3345
|
-
supportsNativeTools: true,
|
|
3346
|
-
defaultToolProtocol: "native",
|
|
3347
3015
|
supportsImages: false,
|
|
3348
3016
|
supportsPromptCache: true,
|
|
3349
3017
|
inputPrice: 1.1,
|
|
@@ -3356,8 +3024,6 @@ var openAiNativeModels = {
|
|
|
3356
3024
|
"o3-mini-high": {
|
|
3357
3025
|
maxTokens: 1e5,
|
|
3358
3026
|
contextWindow: 2e5,
|
|
3359
|
-
supportsNativeTools: true,
|
|
3360
|
-
defaultToolProtocol: "native",
|
|
3361
3027
|
supportsImages: false,
|
|
3362
3028
|
supportsPromptCache: true,
|
|
3363
3029
|
inputPrice: 1.1,
|
|
@@ -3369,8 +3035,6 @@ var openAiNativeModels = {
|
|
|
3369
3035
|
"o3-mini-low": {
|
|
3370
3036
|
maxTokens: 1e5,
|
|
3371
3037
|
contextWindow: 2e5,
|
|
3372
|
-
supportsNativeTools: true,
|
|
3373
|
-
defaultToolProtocol: "native",
|
|
3374
3038
|
supportsImages: false,
|
|
3375
3039
|
supportsPromptCache: true,
|
|
3376
3040
|
inputPrice: 1.1,
|
|
@@ -3382,8 +3046,6 @@ var openAiNativeModels = {
|
|
|
3382
3046
|
o1: {
|
|
3383
3047
|
maxTokens: 1e5,
|
|
3384
3048
|
contextWindow: 2e5,
|
|
3385
|
-
supportsNativeTools: true,
|
|
3386
|
-
defaultToolProtocol: "native",
|
|
3387
3049
|
supportsImages: true,
|
|
3388
3050
|
supportsPromptCache: true,
|
|
3389
3051
|
inputPrice: 15,
|
|
@@ -3394,8 +3056,6 @@ var openAiNativeModels = {
|
|
|
3394
3056
|
"o1-preview": {
|
|
3395
3057
|
maxTokens: 32768,
|
|
3396
3058
|
contextWindow: 128e3,
|
|
3397
|
-
supportsNativeTools: true,
|
|
3398
|
-
defaultToolProtocol: "native",
|
|
3399
3059
|
supportsImages: true,
|
|
3400
3060
|
supportsPromptCache: true,
|
|
3401
3061
|
inputPrice: 15,
|
|
@@ -3406,8 +3066,6 @@ var openAiNativeModels = {
|
|
|
3406
3066
|
"o1-mini": {
|
|
3407
3067
|
maxTokens: 65536,
|
|
3408
3068
|
contextWindow: 128e3,
|
|
3409
|
-
supportsNativeTools: true,
|
|
3410
|
-
defaultToolProtocol: "native",
|
|
3411
3069
|
supportsImages: true,
|
|
3412
3070
|
supportsPromptCache: true,
|
|
3413
3071
|
inputPrice: 1.1,
|
|
@@ -3418,8 +3076,6 @@ var openAiNativeModels = {
|
|
|
3418
3076
|
"gpt-4o": {
|
|
3419
3077
|
maxTokens: 16384,
|
|
3420
3078
|
contextWindow: 128e3,
|
|
3421
|
-
supportsNativeTools: true,
|
|
3422
|
-
defaultToolProtocol: "native",
|
|
3423
3079
|
supportsImages: true,
|
|
3424
3080
|
supportsPromptCache: true,
|
|
3425
3081
|
inputPrice: 2.5,
|
|
@@ -3433,8 +3089,6 @@ var openAiNativeModels = {
|
|
|
3433
3089
|
"gpt-4o-mini": {
|
|
3434
3090
|
maxTokens: 16384,
|
|
3435
3091
|
contextWindow: 128e3,
|
|
3436
|
-
supportsNativeTools: true,
|
|
3437
|
-
defaultToolProtocol: "native",
|
|
3438
3092
|
supportsImages: true,
|
|
3439
3093
|
supportsPromptCache: true,
|
|
3440
3094
|
inputPrice: 0.15,
|
|
@@ -3448,8 +3102,6 @@ var openAiNativeModels = {
|
|
|
3448
3102
|
"codex-mini-latest": {
|
|
3449
3103
|
maxTokens: 16384,
|
|
3450
3104
|
contextWindow: 2e5,
|
|
3451
|
-
supportsNativeTools: true,
|
|
3452
|
-
defaultToolProtocol: "native",
|
|
3453
3105
|
supportsImages: false,
|
|
3454
3106
|
supportsPromptCache: false,
|
|
3455
3107
|
inputPrice: 1.5,
|
|
@@ -3462,8 +3114,6 @@ var openAiNativeModels = {
|
|
|
3462
3114
|
"gpt-5-2025-08-07": {
|
|
3463
3115
|
maxTokens: 128e3,
|
|
3464
3116
|
contextWindow: 4e5,
|
|
3465
|
-
supportsNativeTools: true,
|
|
3466
|
-
defaultToolProtocol: "native",
|
|
3467
3117
|
includedTools: ["apply_patch"],
|
|
3468
3118
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3469
3119
|
supportsImages: true,
|
|
@@ -3484,8 +3134,6 @@ var openAiNativeModels = {
|
|
|
3484
3134
|
"gpt-5-mini-2025-08-07": {
|
|
3485
3135
|
maxTokens: 128e3,
|
|
3486
3136
|
contextWindow: 4e5,
|
|
3487
|
-
supportsNativeTools: true,
|
|
3488
|
-
defaultToolProtocol: "native",
|
|
3489
3137
|
includedTools: ["apply_patch"],
|
|
3490
3138
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3491
3139
|
supportsImages: true,
|
|
@@ -3506,8 +3154,6 @@ var openAiNativeModels = {
|
|
|
3506
3154
|
"gpt-5-nano-2025-08-07": {
|
|
3507
3155
|
maxTokens: 128e3,
|
|
3508
3156
|
contextWindow: 4e5,
|
|
3509
|
-
supportsNativeTools: true,
|
|
3510
|
-
defaultToolProtocol: "native",
|
|
3511
3157
|
includedTools: ["apply_patch"],
|
|
3512
3158
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3513
3159
|
supportsImages: true,
|
|
@@ -3529,14 +3175,151 @@ var openAiModelInfoSaneDefaults = {
|
|
|
3529
3175
|
supportsImages: true,
|
|
3530
3176
|
supportsPromptCache: false,
|
|
3531
3177
|
inputPrice: 0,
|
|
3532
|
-
outputPrice: 0
|
|
3533
|
-
supportsNativeTools: true,
|
|
3534
|
-
defaultToolProtocol: "native"
|
|
3178
|
+
outputPrice: 0
|
|
3535
3179
|
};
|
|
3536
3180
|
var azureOpenAiDefaultApiVersion = "2024-08-01-preview";
|
|
3537
3181
|
var OPENAI_NATIVE_DEFAULT_TEMPERATURE = 0;
|
|
3538
3182
|
var OPENAI_AZURE_AI_INFERENCE_PATH = "/models/chat/completions";
|
|
3539
3183
|
|
|
3184
|
+
// src/providers/openai-codex.ts
|
|
3185
|
+
var openAiCodexDefaultModelId = "gpt-5.2-codex";
|
|
3186
|
+
var openAiCodexModels = {
|
|
3187
|
+
"gpt-5.1-codex-max": {
|
|
3188
|
+
maxTokens: 128e3,
|
|
3189
|
+
contextWindow: 4e5,
|
|
3190
|
+
includedTools: ["apply_patch"],
|
|
3191
|
+
excludedTools: ["apply_diff", "write_to_file"],
|
|
3192
|
+
supportsImages: true,
|
|
3193
|
+
supportsPromptCache: true,
|
|
3194
|
+
supportsReasoningEffort: ["low", "medium", "high", "xhigh"],
|
|
3195
|
+
reasoningEffort: "xhigh",
|
|
3196
|
+
// Subscription-based: no per-token costs
|
|
3197
|
+
inputPrice: 0,
|
|
3198
|
+
outputPrice: 0,
|
|
3199
|
+
supportsTemperature: false,
|
|
3200
|
+
description: "GPT-5.1 Codex Max: Maximum capability coding model via ChatGPT subscription"
|
|
3201
|
+
},
|
|
3202
|
+
"gpt-5.1-codex": {
|
|
3203
|
+
maxTokens: 128e3,
|
|
3204
|
+
contextWindow: 4e5,
|
|
3205
|
+
includedTools: ["apply_patch"],
|
|
3206
|
+
excludedTools: ["apply_diff", "write_to_file"],
|
|
3207
|
+
supportsImages: true,
|
|
3208
|
+
supportsPromptCache: true,
|
|
3209
|
+
supportsReasoningEffort: ["low", "medium", "high"],
|
|
3210
|
+
reasoningEffort: "medium",
|
|
3211
|
+
// Subscription-based: no per-token costs
|
|
3212
|
+
inputPrice: 0,
|
|
3213
|
+
outputPrice: 0,
|
|
3214
|
+
supportsTemperature: false,
|
|
3215
|
+
description: "GPT-5.1 Codex: GPT-5.1 optimized for agentic coding via ChatGPT subscription"
|
|
3216
|
+
},
|
|
3217
|
+
"gpt-5.2-codex": {
|
|
3218
|
+
maxTokens: 128e3,
|
|
3219
|
+
contextWindow: 4e5,
|
|
3220
|
+
includedTools: ["apply_patch"],
|
|
3221
|
+
excludedTools: ["apply_diff", "write_to_file"],
|
|
3222
|
+
supportsImages: true,
|
|
3223
|
+
supportsPromptCache: true,
|
|
3224
|
+
supportsReasoningEffort: ["low", "medium", "high", "xhigh"],
|
|
3225
|
+
reasoningEffort: "medium",
|
|
3226
|
+
inputPrice: 0,
|
|
3227
|
+
outputPrice: 0,
|
|
3228
|
+
supportsTemperature: false,
|
|
3229
|
+
description: "GPT-5.2 Codex: OpenAI's flagship coding model via ChatGPT subscription"
|
|
3230
|
+
},
|
|
3231
|
+
"gpt-5.1": {
|
|
3232
|
+
maxTokens: 128e3,
|
|
3233
|
+
contextWindow: 4e5,
|
|
3234
|
+
includedTools: ["apply_patch"],
|
|
3235
|
+
excludedTools: ["apply_diff", "write_to_file"],
|
|
3236
|
+
supportsImages: true,
|
|
3237
|
+
supportsPromptCache: true,
|
|
3238
|
+
supportsReasoningEffort: ["none", "low", "medium", "high"],
|
|
3239
|
+
reasoningEffort: "medium",
|
|
3240
|
+
// Subscription-based: no per-token costs
|
|
3241
|
+
inputPrice: 0,
|
|
3242
|
+
outputPrice: 0,
|
|
3243
|
+
supportsVerbosity: true,
|
|
3244
|
+
supportsTemperature: false,
|
|
3245
|
+
description: "GPT-5.1: General GPT-5.1 model via ChatGPT subscription"
|
|
3246
|
+
},
|
|
3247
|
+
"gpt-5": {
|
|
3248
|
+
maxTokens: 128e3,
|
|
3249
|
+
contextWindow: 4e5,
|
|
3250
|
+
includedTools: ["apply_patch"],
|
|
3251
|
+
excludedTools: ["apply_diff", "write_to_file"],
|
|
3252
|
+
supportsImages: true,
|
|
3253
|
+
supportsPromptCache: true,
|
|
3254
|
+
supportsReasoningEffort: ["minimal", "low", "medium", "high"],
|
|
3255
|
+
reasoningEffort: "medium",
|
|
3256
|
+
// Subscription-based: no per-token costs
|
|
3257
|
+
inputPrice: 0,
|
|
3258
|
+
outputPrice: 0,
|
|
3259
|
+
supportsVerbosity: true,
|
|
3260
|
+
supportsTemperature: false,
|
|
3261
|
+
description: "GPT-5: General GPT-5 model via ChatGPT subscription"
|
|
3262
|
+
},
|
|
3263
|
+
"gpt-5-codex": {
|
|
3264
|
+
maxTokens: 128e3,
|
|
3265
|
+
contextWindow: 4e5,
|
|
3266
|
+
includedTools: ["apply_patch"],
|
|
3267
|
+
excludedTools: ["apply_diff", "write_to_file"],
|
|
3268
|
+
supportsImages: true,
|
|
3269
|
+
supportsPromptCache: true,
|
|
3270
|
+
supportsReasoningEffort: ["low", "medium", "high"],
|
|
3271
|
+
reasoningEffort: "medium",
|
|
3272
|
+
// Subscription-based: no per-token costs
|
|
3273
|
+
inputPrice: 0,
|
|
3274
|
+
outputPrice: 0,
|
|
3275
|
+
supportsTemperature: false,
|
|
3276
|
+
description: "GPT-5 Codex: GPT-5 optimized for agentic coding via ChatGPT subscription"
|
|
3277
|
+
},
|
|
3278
|
+
"gpt-5-codex-mini": {
|
|
3279
|
+
maxTokens: 128e3,
|
|
3280
|
+
contextWindow: 4e5,
|
|
3281
|
+
includedTools: ["apply_patch"],
|
|
3282
|
+
excludedTools: ["apply_diff", "write_to_file"],
|
|
3283
|
+
supportsImages: true,
|
|
3284
|
+
supportsPromptCache: true,
|
|
3285
|
+
supportsReasoningEffort: ["low", "medium", "high"],
|
|
3286
|
+
reasoningEffort: "medium",
|
|
3287
|
+
// Subscription-based: no per-token costs
|
|
3288
|
+
inputPrice: 0,
|
|
3289
|
+
outputPrice: 0,
|
|
3290
|
+
supportsTemperature: false,
|
|
3291
|
+
description: "GPT-5 Codex Mini: Faster coding model via ChatGPT subscription"
|
|
3292
|
+
},
|
|
3293
|
+
"gpt-5.1-codex-mini": {
|
|
3294
|
+
maxTokens: 128e3,
|
|
3295
|
+
contextWindow: 4e5,
|
|
3296
|
+
includedTools: ["apply_patch"],
|
|
3297
|
+
excludedTools: ["apply_diff", "write_to_file"],
|
|
3298
|
+
supportsImages: true,
|
|
3299
|
+
supportsPromptCache: true,
|
|
3300
|
+
supportsReasoningEffort: ["low", "medium", "high"],
|
|
3301
|
+
reasoningEffort: "medium",
|
|
3302
|
+
inputPrice: 0,
|
|
3303
|
+
outputPrice: 0,
|
|
3304
|
+
supportsTemperature: false,
|
|
3305
|
+
description: "GPT-5.1 Codex Mini: Faster version for coding tasks via ChatGPT subscription"
|
|
3306
|
+
},
|
|
3307
|
+
"gpt-5.2": {
|
|
3308
|
+
maxTokens: 128e3,
|
|
3309
|
+
contextWindow: 4e5,
|
|
3310
|
+
includedTools: ["apply_patch"],
|
|
3311
|
+
excludedTools: ["apply_diff", "write_to_file"],
|
|
3312
|
+
supportsImages: true,
|
|
3313
|
+
supportsPromptCache: true,
|
|
3314
|
+
supportsReasoningEffort: ["none", "low", "medium", "high", "xhigh"],
|
|
3315
|
+
reasoningEffort: "medium",
|
|
3316
|
+
inputPrice: 0,
|
|
3317
|
+
outputPrice: 0,
|
|
3318
|
+
supportsTemperature: false,
|
|
3319
|
+
description: "GPT-5.2: Latest GPT model via ChatGPT subscription"
|
|
3320
|
+
}
|
|
3321
|
+
};
|
|
3322
|
+
|
|
3540
3323
|
// src/providers/openrouter.ts
|
|
3541
3324
|
var openRouterDefaultModelId = "anthropic/claude-sonnet-4.5";
|
|
3542
3325
|
var openRouterDefaultModelInfo = {
|
|
@@ -3544,7 +3327,6 @@ var openRouterDefaultModelInfo = {
|
|
|
3544
3327
|
contextWindow: 2e5,
|
|
3545
3328
|
supportsImages: true,
|
|
3546
3329
|
supportsPromptCache: true,
|
|
3547
|
-
supportsNativeTools: true,
|
|
3548
3330
|
inputPrice: 3,
|
|
3549
3331
|
outputPrice: 15,
|
|
3550
3332
|
cacheWritesPrice: 3.75,
|
|
@@ -3618,8 +3400,6 @@ var qwenCodeModels = {
|
|
|
3618
3400
|
contextWindow: 1e6,
|
|
3619
3401
|
supportsImages: false,
|
|
3620
3402
|
supportsPromptCache: false,
|
|
3621
|
-
supportsNativeTools: true,
|
|
3622
|
-
defaultToolProtocol: "native",
|
|
3623
3403
|
inputPrice: 0,
|
|
3624
3404
|
outputPrice: 0,
|
|
3625
3405
|
cacheWritesPrice: 0,
|
|
@@ -3631,8 +3411,6 @@ var qwenCodeModels = {
|
|
|
3631
3411
|
contextWindow: 1e6,
|
|
3632
3412
|
supportsImages: false,
|
|
3633
3413
|
supportsPromptCache: false,
|
|
3634
|
-
supportsNativeTools: true,
|
|
3635
|
-
defaultToolProtocol: "native",
|
|
3636
3414
|
inputPrice: 0,
|
|
3637
3415
|
outputPrice: 0,
|
|
3638
3416
|
cacheWritesPrice: 0,
|
|
@@ -3648,8 +3426,6 @@ var requestyDefaultModelInfo = {
|
|
|
3648
3426
|
contextWindow: 2e5,
|
|
3649
3427
|
supportsImages: true,
|
|
3650
3428
|
supportsPromptCache: true,
|
|
3651
|
-
supportsNativeTools: true,
|
|
3652
|
-
defaultToolProtocol: "native",
|
|
3653
3429
|
inputPrice: 3,
|
|
3654
3430
|
outputPrice: 15,
|
|
3655
3431
|
cacheWritesPrice: 3.75,
|
|
@@ -3705,8 +3481,6 @@ var sambaNovaModels = {
|
|
|
3705
3481
|
contextWindow: 16384,
|
|
3706
3482
|
supportsImages: false,
|
|
3707
3483
|
supportsPromptCache: false,
|
|
3708
|
-
supportsNativeTools: true,
|
|
3709
|
-
defaultToolProtocol: "native",
|
|
3710
3484
|
inputPrice: 0.1,
|
|
3711
3485
|
outputPrice: 0.2,
|
|
3712
3486
|
description: "Meta Llama 3.1 8B Instruct model with 16K context window."
|
|
@@ -3716,8 +3490,6 @@ var sambaNovaModels = {
|
|
|
3716
3490
|
contextWindow: 131072,
|
|
3717
3491
|
supportsImages: false,
|
|
3718
3492
|
supportsPromptCache: false,
|
|
3719
|
-
supportsNativeTools: true,
|
|
3720
|
-
defaultToolProtocol: "native",
|
|
3721
3493
|
inputPrice: 0.6,
|
|
3722
3494
|
outputPrice: 1.2,
|
|
3723
3495
|
description: "Meta Llama 3.3 70B Instruct model with 128K context window."
|
|
@@ -3728,8 +3500,6 @@ var sambaNovaModels = {
|
|
|
3728
3500
|
supportsImages: false,
|
|
3729
3501
|
supportsPromptCache: false,
|
|
3730
3502
|
supportsReasoningBudget: true,
|
|
3731
|
-
supportsNativeTools: true,
|
|
3732
|
-
defaultToolProtocol: "native",
|
|
3733
3503
|
inputPrice: 5,
|
|
3734
3504
|
outputPrice: 7,
|
|
3735
3505
|
description: "DeepSeek R1 reasoning model with 32K context window."
|
|
@@ -3739,8 +3509,6 @@ var sambaNovaModels = {
|
|
|
3739
3509
|
contextWindow: 32768,
|
|
3740
3510
|
supportsImages: false,
|
|
3741
3511
|
supportsPromptCache: false,
|
|
3742
|
-
supportsNativeTools: true,
|
|
3743
|
-
defaultToolProtocol: "native",
|
|
3744
3512
|
inputPrice: 3,
|
|
3745
3513
|
outputPrice: 4.5,
|
|
3746
3514
|
description: "DeepSeek V3 model with 32K context window."
|
|
@@ -3750,8 +3518,6 @@ var sambaNovaModels = {
|
|
|
3750
3518
|
contextWindow: 32768,
|
|
3751
3519
|
supportsImages: false,
|
|
3752
3520
|
supportsPromptCache: false,
|
|
3753
|
-
supportsNativeTools: true,
|
|
3754
|
-
defaultToolProtocol: "native",
|
|
3755
3521
|
inputPrice: 3,
|
|
3756
3522
|
outputPrice: 4.5,
|
|
3757
3523
|
description: "DeepSeek V3.1 model with 32K context window."
|
|
@@ -3761,8 +3527,6 @@ var sambaNovaModels = {
|
|
|
3761
3527
|
contextWindow: 131072,
|
|
3762
3528
|
supportsImages: true,
|
|
3763
3529
|
supportsPromptCache: false,
|
|
3764
|
-
supportsNativeTools: true,
|
|
3765
|
-
defaultToolProtocol: "native",
|
|
3766
3530
|
inputPrice: 0.63,
|
|
3767
3531
|
outputPrice: 1.8,
|
|
3768
3532
|
description: "Meta Llama 4 Maverick 17B 128E Instruct model with 128K context window."
|
|
@@ -3772,8 +3536,6 @@ var sambaNovaModels = {
|
|
|
3772
3536
|
contextWindow: 8192,
|
|
3773
3537
|
supportsImages: false,
|
|
3774
3538
|
supportsPromptCache: false,
|
|
3775
|
-
supportsNativeTools: true,
|
|
3776
|
-
defaultToolProtocol: "native",
|
|
3777
3539
|
inputPrice: 0.4,
|
|
3778
3540
|
outputPrice: 0.8,
|
|
3779
3541
|
description: "Alibaba Qwen 3 32B model with 8K context window."
|
|
@@ -3783,8 +3545,6 @@ var sambaNovaModels = {
|
|
|
3783
3545
|
contextWindow: 131072,
|
|
3784
3546
|
supportsImages: false,
|
|
3785
3547
|
supportsPromptCache: false,
|
|
3786
|
-
supportsNativeTools: true,
|
|
3787
|
-
defaultToolProtocol: "native",
|
|
3788
3548
|
inputPrice: 0.22,
|
|
3789
3549
|
outputPrice: 0.59,
|
|
3790
3550
|
description: "OpenAI gpt oss 120b model with 128k context window."
|
|
@@ -3798,7 +3558,6 @@ var unboundDefaultModelInfo = {
|
|
|
3798
3558
|
contextWindow: 2e5,
|
|
3799
3559
|
supportsImages: true,
|
|
3800
3560
|
supportsPromptCache: true,
|
|
3801
|
-
supportsNativeTools: true,
|
|
3802
3561
|
inputPrice: 3,
|
|
3803
3562
|
outputPrice: 15,
|
|
3804
3563
|
cacheWritesPrice: 3.75,
|
|
@@ -3812,8 +3571,6 @@ var vertexModels = {
|
|
|
3812
3571
|
maxTokens: 65536,
|
|
3813
3572
|
contextWindow: 1048576,
|
|
3814
3573
|
supportsImages: true,
|
|
3815
|
-
supportsNativeTools: true,
|
|
3816
|
-
defaultToolProtocol: "native",
|
|
3817
3574
|
supportsPromptCache: true,
|
|
3818
3575
|
supportsReasoningEffort: ["low", "high"],
|
|
3819
3576
|
reasoningEffort: "low",
|
|
@@ -3821,16 +3578,19 @@ var vertexModels = {
|
|
|
3821
3578
|
defaultTemperature: 1,
|
|
3822
3579
|
inputPrice: 4,
|
|
3823
3580
|
outputPrice: 18,
|
|
3581
|
+
cacheReadsPrice: 0.4,
|
|
3824
3582
|
tiers: [
|
|
3825
3583
|
{
|
|
3826
3584
|
contextWindow: 2e5,
|
|
3827
3585
|
inputPrice: 2,
|
|
3828
|
-
outputPrice: 12
|
|
3586
|
+
outputPrice: 12,
|
|
3587
|
+
cacheReadsPrice: 0.2
|
|
3829
3588
|
},
|
|
3830
3589
|
{
|
|
3831
3590
|
contextWindow: Infinity,
|
|
3832
3591
|
inputPrice: 4,
|
|
3833
|
-
outputPrice: 18
|
|
3592
|
+
outputPrice: 18,
|
|
3593
|
+
cacheReadsPrice: 0.4
|
|
3834
3594
|
}
|
|
3835
3595
|
]
|
|
3836
3596
|
},
|
|
@@ -3838,24 +3598,19 @@ var vertexModels = {
|
|
|
3838
3598
|
maxTokens: 65536,
|
|
3839
3599
|
contextWindow: 1048576,
|
|
3840
3600
|
supportsImages: true,
|
|
3841
|
-
supportsNativeTools: true,
|
|
3842
|
-
defaultToolProtocol: "native",
|
|
3843
3601
|
supportsPromptCache: true,
|
|
3844
3602
|
supportsReasoningEffort: ["minimal", "low", "medium", "high"],
|
|
3845
3603
|
reasoningEffort: "medium",
|
|
3846
3604
|
supportsTemperature: true,
|
|
3847
3605
|
defaultTemperature: 1,
|
|
3848
|
-
inputPrice: 0.
|
|
3849
|
-
outputPrice:
|
|
3850
|
-
cacheReadsPrice: 0.
|
|
3851
|
-
cacheWritesPrice: 1
|
|
3606
|
+
inputPrice: 0.5,
|
|
3607
|
+
outputPrice: 3,
|
|
3608
|
+
cacheReadsPrice: 0.05
|
|
3852
3609
|
},
|
|
3853
3610
|
"gemini-2.5-flash-preview-05-20:thinking": {
|
|
3854
3611
|
maxTokens: 65535,
|
|
3855
3612
|
contextWindow: 1048576,
|
|
3856
3613
|
supportsImages: true,
|
|
3857
|
-
supportsNativeTools: true,
|
|
3858
|
-
defaultToolProtocol: "native",
|
|
3859
3614
|
supportsPromptCache: true,
|
|
3860
3615
|
inputPrice: 0.15,
|
|
3861
3616
|
outputPrice: 3.5,
|
|
@@ -3867,8 +3622,6 @@ var vertexModels = {
|
|
|
3867
3622
|
maxTokens: 65535,
|
|
3868
3623
|
contextWindow: 1048576,
|
|
3869
3624
|
supportsImages: true,
|
|
3870
|
-
supportsNativeTools: true,
|
|
3871
|
-
defaultToolProtocol: "native",
|
|
3872
3625
|
supportsPromptCache: true,
|
|
3873
3626
|
inputPrice: 0.15,
|
|
3874
3627
|
outputPrice: 0.6
|
|
@@ -3877,8 +3630,6 @@ var vertexModels = {
|
|
|
3877
3630
|
maxTokens: 64e3,
|
|
3878
3631
|
contextWindow: 1048576,
|
|
3879
3632
|
supportsImages: true,
|
|
3880
|
-
supportsNativeTools: true,
|
|
3881
|
-
defaultToolProtocol: "native",
|
|
3882
3633
|
supportsPromptCache: true,
|
|
3883
3634
|
inputPrice: 0.3,
|
|
3884
3635
|
outputPrice: 2.5,
|
|
@@ -3891,8 +3642,6 @@ var vertexModels = {
|
|
|
3891
3642
|
maxTokens: 65535,
|
|
3892
3643
|
contextWindow: 1048576,
|
|
3893
3644
|
supportsImages: true,
|
|
3894
|
-
supportsNativeTools: true,
|
|
3895
|
-
defaultToolProtocol: "native",
|
|
3896
3645
|
supportsPromptCache: false,
|
|
3897
3646
|
inputPrice: 0.15,
|
|
3898
3647
|
outputPrice: 3.5,
|
|
@@ -3904,8 +3653,6 @@ var vertexModels = {
|
|
|
3904
3653
|
maxTokens: 65535,
|
|
3905
3654
|
contextWindow: 1048576,
|
|
3906
3655
|
supportsImages: true,
|
|
3907
|
-
supportsNativeTools: true,
|
|
3908
|
-
defaultToolProtocol: "native",
|
|
3909
3656
|
supportsPromptCache: false,
|
|
3910
3657
|
inputPrice: 0.15,
|
|
3911
3658
|
outputPrice: 0.6
|
|
@@ -3914,8 +3661,6 @@ var vertexModels = {
|
|
|
3914
3661
|
maxTokens: 65535,
|
|
3915
3662
|
contextWindow: 1048576,
|
|
3916
3663
|
supportsImages: true,
|
|
3917
|
-
supportsNativeTools: true,
|
|
3918
|
-
defaultToolProtocol: "native",
|
|
3919
3664
|
supportsPromptCache: true,
|
|
3920
3665
|
inputPrice: 2.5,
|
|
3921
3666
|
outputPrice: 15
|
|
@@ -3924,8 +3669,6 @@ var vertexModels = {
|
|
|
3924
3669
|
maxTokens: 65535,
|
|
3925
3670
|
contextWindow: 1048576,
|
|
3926
3671
|
supportsImages: true,
|
|
3927
|
-
supportsNativeTools: true,
|
|
3928
|
-
defaultToolProtocol: "native",
|
|
3929
3672
|
supportsPromptCache: true,
|
|
3930
3673
|
inputPrice: 2.5,
|
|
3931
3674
|
outputPrice: 15
|
|
@@ -3934,8 +3677,6 @@ var vertexModels = {
|
|
|
3934
3677
|
maxTokens: 65535,
|
|
3935
3678
|
contextWindow: 1048576,
|
|
3936
3679
|
supportsImages: true,
|
|
3937
|
-
supportsNativeTools: true,
|
|
3938
|
-
defaultToolProtocol: "native",
|
|
3939
3680
|
supportsPromptCache: true,
|
|
3940
3681
|
inputPrice: 2.5,
|
|
3941
3682
|
outputPrice: 15,
|
|
@@ -3946,8 +3687,6 @@ var vertexModels = {
|
|
|
3946
3687
|
maxTokens: 64e3,
|
|
3947
3688
|
contextWindow: 1048576,
|
|
3948
3689
|
supportsImages: true,
|
|
3949
|
-
supportsNativeTools: true,
|
|
3950
|
-
defaultToolProtocol: "native",
|
|
3951
3690
|
supportsPromptCache: true,
|
|
3952
3691
|
inputPrice: 2.5,
|
|
3953
3692
|
outputPrice: 15,
|
|
@@ -3973,8 +3712,6 @@ var vertexModels = {
|
|
|
3973
3712
|
maxTokens: 65535,
|
|
3974
3713
|
contextWindow: 1048576,
|
|
3975
3714
|
supportsImages: true,
|
|
3976
|
-
supportsNativeTools: true,
|
|
3977
|
-
defaultToolProtocol: "native",
|
|
3978
3715
|
supportsPromptCache: false,
|
|
3979
3716
|
inputPrice: 0,
|
|
3980
3717
|
outputPrice: 0
|
|
@@ -3983,8 +3720,6 @@ var vertexModels = {
|
|
|
3983
3720
|
maxTokens: 8192,
|
|
3984
3721
|
contextWindow: 2097152,
|
|
3985
3722
|
supportsImages: true,
|
|
3986
|
-
supportsNativeTools: true,
|
|
3987
|
-
defaultToolProtocol: "native",
|
|
3988
3723
|
supportsPromptCache: false,
|
|
3989
3724
|
inputPrice: 0,
|
|
3990
3725
|
outputPrice: 0
|
|
@@ -3993,8 +3728,6 @@ var vertexModels = {
|
|
|
3993
3728
|
maxTokens: 8192,
|
|
3994
3729
|
contextWindow: 1048576,
|
|
3995
3730
|
supportsImages: true,
|
|
3996
|
-
supportsNativeTools: true,
|
|
3997
|
-
defaultToolProtocol: "native",
|
|
3998
3731
|
supportsPromptCache: true,
|
|
3999
3732
|
inputPrice: 0.15,
|
|
4000
3733
|
outputPrice: 0.6
|
|
@@ -4003,8 +3736,6 @@ var vertexModels = {
|
|
|
4003
3736
|
maxTokens: 8192,
|
|
4004
3737
|
contextWindow: 1048576,
|
|
4005
3738
|
supportsImages: true,
|
|
4006
|
-
supportsNativeTools: true,
|
|
4007
|
-
defaultToolProtocol: "native",
|
|
4008
3739
|
supportsPromptCache: false,
|
|
4009
3740
|
inputPrice: 0.075,
|
|
4010
3741
|
outputPrice: 0.3
|
|
@@ -4013,8 +3744,6 @@ var vertexModels = {
|
|
|
4013
3744
|
maxTokens: 8192,
|
|
4014
3745
|
contextWindow: 32768,
|
|
4015
3746
|
supportsImages: true,
|
|
4016
|
-
supportsNativeTools: true,
|
|
4017
|
-
defaultToolProtocol: "native",
|
|
4018
3747
|
supportsPromptCache: false,
|
|
4019
3748
|
inputPrice: 0,
|
|
4020
3749
|
outputPrice: 0
|
|
@@ -4023,8 +3752,6 @@ var vertexModels = {
|
|
|
4023
3752
|
maxTokens: 8192,
|
|
4024
3753
|
contextWindow: 1048576,
|
|
4025
3754
|
supportsImages: true,
|
|
4026
|
-
supportsNativeTools: true,
|
|
4027
|
-
defaultToolProtocol: "native",
|
|
4028
3755
|
supportsPromptCache: true,
|
|
4029
3756
|
inputPrice: 0.075,
|
|
4030
3757
|
outputPrice: 0.3
|
|
@@ -4033,8 +3760,6 @@ var vertexModels = {
|
|
|
4033
3760
|
maxTokens: 8192,
|
|
4034
3761
|
contextWindow: 2097152,
|
|
4035
3762
|
supportsImages: true,
|
|
4036
|
-
supportsNativeTools: true,
|
|
4037
|
-
defaultToolProtocol: "native",
|
|
4038
3763
|
supportsPromptCache: false,
|
|
4039
3764
|
inputPrice: 1.25,
|
|
4040
3765
|
outputPrice: 5
|
|
@@ -4045,8 +3770,6 @@ var vertexModels = {
|
|
|
4045
3770
|
// Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
|
|
4046
3771
|
supportsImages: true,
|
|
4047
3772
|
supportsPromptCache: true,
|
|
4048
|
-
supportsNativeTools: true,
|
|
4049
|
-
defaultToolProtocol: "native",
|
|
4050
3773
|
inputPrice: 3,
|
|
4051
3774
|
// $3 per million input tokens (≤200K context)
|
|
4052
3775
|
outputPrice: 15,
|
|
@@ -4078,8 +3801,6 @@ var vertexModels = {
|
|
|
4078
3801
|
// Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
|
|
4079
3802
|
supportsImages: true,
|
|
4080
3803
|
supportsPromptCache: true,
|
|
4081
|
-
supportsNativeTools: true,
|
|
4082
|
-
defaultToolProtocol: "native",
|
|
4083
3804
|
inputPrice: 3,
|
|
4084
3805
|
// $3 per million input tokens (≤200K context)
|
|
4085
3806
|
outputPrice: 15,
|
|
@@ -4110,8 +3831,6 @@ var vertexModels = {
|
|
|
4110
3831
|
contextWindow: 2e5,
|
|
4111
3832
|
supportsImages: true,
|
|
4112
3833
|
supportsPromptCache: true,
|
|
4113
|
-
supportsNativeTools: true,
|
|
4114
|
-
defaultToolProtocol: "native",
|
|
4115
3834
|
inputPrice: 1,
|
|
4116
3835
|
outputPrice: 5,
|
|
4117
3836
|
cacheWritesPrice: 1.25,
|
|
@@ -4123,8 +3842,6 @@ var vertexModels = {
|
|
|
4123
3842
|
contextWindow: 2e5,
|
|
4124
3843
|
supportsImages: true,
|
|
4125
3844
|
supportsPromptCache: true,
|
|
4126
|
-
supportsNativeTools: true,
|
|
4127
|
-
defaultToolProtocol: "native",
|
|
4128
3845
|
inputPrice: 5,
|
|
4129
3846
|
outputPrice: 25,
|
|
4130
3847
|
cacheWritesPrice: 6.25,
|
|
@@ -4136,8 +3853,6 @@ var vertexModels = {
|
|
|
4136
3853
|
contextWindow: 2e5,
|
|
4137
3854
|
supportsImages: true,
|
|
4138
3855
|
supportsPromptCache: true,
|
|
4139
|
-
supportsNativeTools: true,
|
|
4140
|
-
defaultToolProtocol: "native",
|
|
4141
3856
|
inputPrice: 15,
|
|
4142
3857
|
outputPrice: 75,
|
|
4143
3858
|
cacheWritesPrice: 18.75,
|
|
@@ -4149,8 +3864,6 @@ var vertexModels = {
|
|
|
4149
3864
|
contextWindow: 2e5,
|
|
4150
3865
|
supportsImages: true,
|
|
4151
3866
|
supportsPromptCache: true,
|
|
4152
|
-
supportsNativeTools: true,
|
|
4153
|
-
defaultToolProtocol: "native",
|
|
4154
3867
|
inputPrice: 15,
|
|
4155
3868
|
outputPrice: 75,
|
|
4156
3869
|
cacheWritesPrice: 18.75,
|
|
@@ -4161,8 +3874,6 @@ var vertexModels = {
|
|
|
4161
3874
|
contextWindow: 2e5,
|
|
4162
3875
|
supportsImages: true,
|
|
4163
3876
|
supportsPromptCache: true,
|
|
4164
|
-
supportsNativeTools: true,
|
|
4165
|
-
defaultToolProtocol: "native",
|
|
4166
3877
|
inputPrice: 3,
|
|
4167
3878
|
outputPrice: 15,
|
|
4168
3879
|
cacheWritesPrice: 3.75,
|
|
@@ -4175,8 +3886,6 @@ var vertexModels = {
|
|
|
4175
3886
|
contextWindow: 2e5,
|
|
4176
3887
|
supportsImages: true,
|
|
4177
3888
|
supportsPromptCache: true,
|
|
4178
|
-
supportsNativeTools: true,
|
|
4179
|
-
defaultToolProtocol: "native",
|
|
4180
3889
|
inputPrice: 3,
|
|
4181
3890
|
outputPrice: 15,
|
|
4182
3891
|
cacheWritesPrice: 3.75,
|
|
@@ -4187,8 +3896,6 @@ var vertexModels = {
|
|
|
4187
3896
|
contextWindow: 2e5,
|
|
4188
3897
|
supportsImages: true,
|
|
4189
3898
|
supportsPromptCache: true,
|
|
4190
|
-
supportsNativeTools: true,
|
|
4191
|
-
defaultToolProtocol: "native",
|
|
4192
3899
|
inputPrice: 3,
|
|
4193
3900
|
outputPrice: 15,
|
|
4194
3901
|
cacheWritesPrice: 3.75,
|
|
@@ -4199,8 +3906,6 @@ var vertexModels = {
|
|
|
4199
3906
|
contextWindow: 2e5,
|
|
4200
3907
|
supportsImages: true,
|
|
4201
3908
|
supportsPromptCache: true,
|
|
4202
|
-
supportsNativeTools: true,
|
|
4203
|
-
defaultToolProtocol: "native",
|
|
4204
3909
|
inputPrice: 3,
|
|
4205
3910
|
outputPrice: 15,
|
|
4206
3911
|
cacheWritesPrice: 3.75,
|
|
@@ -4211,8 +3916,6 @@ var vertexModels = {
|
|
|
4211
3916
|
contextWindow: 2e5,
|
|
4212
3917
|
supportsImages: false,
|
|
4213
3918
|
supportsPromptCache: true,
|
|
4214
|
-
supportsNativeTools: true,
|
|
4215
|
-
defaultToolProtocol: "native",
|
|
4216
3919
|
inputPrice: 1,
|
|
4217
3920
|
outputPrice: 5,
|
|
4218
3921
|
cacheWritesPrice: 1.25,
|
|
@@ -4223,8 +3926,6 @@ var vertexModels = {
|
|
|
4223
3926
|
contextWindow: 2e5,
|
|
4224
3927
|
supportsImages: true,
|
|
4225
3928
|
supportsPromptCache: true,
|
|
4226
|
-
supportsNativeTools: true,
|
|
4227
|
-
defaultToolProtocol: "native",
|
|
4228
3929
|
inputPrice: 15,
|
|
4229
3930
|
outputPrice: 75,
|
|
4230
3931
|
cacheWritesPrice: 18.75,
|
|
@@ -4235,8 +3936,6 @@ var vertexModels = {
|
|
|
4235
3936
|
contextWindow: 2e5,
|
|
4236
3937
|
supportsImages: true,
|
|
4237
3938
|
supportsPromptCache: true,
|
|
4238
|
-
supportsNativeTools: true,
|
|
4239
|
-
defaultToolProtocol: "native",
|
|
4240
3939
|
inputPrice: 0.25,
|
|
4241
3940
|
outputPrice: 1.25,
|
|
4242
3941
|
cacheWritesPrice: 0.3,
|
|
@@ -4246,8 +3945,6 @@ var vertexModels = {
|
|
|
4246
3945
|
maxTokens: 64e3,
|
|
4247
3946
|
contextWindow: 1048576,
|
|
4248
3947
|
supportsImages: true,
|
|
4249
|
-
supportsNativeTools: true,
|
|
4250
|
-
defaultToolProtocol: "native",
|
|
4251
3948
|
supportsPromptCache: true,
|
|
4252
3949
|
inputPrice: 0.1,
|
|
4253
3950
|
outputPrice: 0.4,
|
|
@@ -4261,7 +3958,6 @@ var vertexModels = {
|
|
|
4261
3958
|
contextWindow: 131072,
|
|
4262
3959
|
supportsImages: false,
|
|
4263
3960
|
supportsPromptCache: false,
|
|
4264
|
-
supportsNativeTools: true,
|
|
4265
3961
|
inputPrice: 0.35,
|
|
4266
3962
|
outputPrice: 1.15,
|
|
4267
3963
|
description: "Meta Llama 4 Maverick 17B Instruct model, 128K context."
|
|
@@ -4271,7 +3967,6 @@ var vertexModels = {
|
|
|
4271
3967
|
contextWindow: 163840,
|
|
4272
3968
|
supportsImages: false,
|
|
4273
3969
|
supportsPromptCache: false,
|
|
4274
|
-
supportsNativeTools: true,
|
|
4275
3970
|
inputPrice: 1.35,
|
|
4276
3971
|
outputPrice: 5.4,
|
|
4277
3972
|
description: "DeepSeek R1 (0528). Available in us-central1"
|
|
@@ -4281,7 +3976,6 @@ var vertexModels = {
|
|
|
4281
3976
|
contextWindow: 163840,
|
|
4282
3977
|
supportsImages: false,
|
|
4283
3978
|
supportsPromptCache: false,
|
|
4284
|
-
supportsNativeTools: true,
|
|
4285
3979
|
inputPrice: 0.6,
|
|
4286
3980
|
outputPrice: 1.7,
|
|
4287
3981
|
description: "DeepSeek V3.1. Available in us-west2"
|
|
@@ -4291,7 +3985,6 @@ var vertexModels = {
|
|
|
4291
3985
|
contextWindow: 131072,
|
|
4292
3986
|
supportsImages: false,
|
|
4293
3987
|
supportsPromptCache: false,
|
|
4294
|
-
supportsNativeTools: true,
|
|
4295
3988
|
inputPrice: 0.15,
|
|
4296
3989
|
outputPrice: 0.6,
|
|
4297
3990
|
description: "OpenAI gpt-oss 120B. Available in us-central1"
|
|
@@ -4301,7 +3994,6 @@ var vertexModels = {
|
|
|
4301
3994
|
contextWindow: 131072,
|
|
4302
3995
|
supportsImages: false,
|
|
4303
3996
|
supportsPromptCache: false,
|
|
4304
|
-
supportsNativeTools: true,
|
|
4305
3997
|
inputPrice: 0.075,
|
|
4306
3998
|
outputPrice: 0.3,
|
|
4307
3999
|
description: "OpenAI gpt-oss 20B. Available in us-central1"
|
|
@@ -4311,7 +4003,6 @@ var vertexModels = {
|
|
|
4311
4003
|
contextWindow: 262144,
|
|
4312
4004
|
supportsImages: false,
|
|
4313
4005
|
supportsPromptCache: false,
|
|
4314
|
-
supportsNativeTools: true,
|
|
4315
4006
|
inputPrice: 1,
|
|
4316
4007
|
outputPrice: 4,
|
|
4317
4008
|
description: "Qwen3 Coder 480B A35B Instruct. Available in us-south1"
|
|
@@ -4321,10 +4012,18 @@ var vertexModels = {
|
|
|
4321
4012
|
contextWindow: 262144,
|
|
4322
4013
|
supportsImages: false,
|
|
4323
4014
|
supportsPromptCache: false,
|
|
4324
|
-
supportsNativeTools: true,
|
|
4325
4015
|
inputPrice: 0.25,
|
|
4326
4016
|
outputPrice: 1,
|
|
4327
4017
|
description: "Qwen3 235B A22B Instruct. Available in us-south1"
|
|
4018
|
+
},
|
|
4019
|
+
"moonshotai/kimi-k2-thinking-maas": {
|
|
4020
|
+
maxTokens: 16384,
|
|
4021
|
+
contextWindow: 262144,
|
|
4022
|
+
supportsPromptCache: false,
|
|
4023
|
+
supportsImages: false,
|
|
4024
|
+
inputPrice: 0.6,
|
|
4025
|
+
outputPrice: 2.5,
|
|
4026
|
+
description: "Kimi K2 Thinking Model with 256K context window."
|
|
4328
4027
|
}
|
|
4329
4028
|
};
|
|
4330
4029
|
var VERTEX_1M_CONTEXT_MODEL_IDS = ["claude-sonnet-4@20250514", "claude-sonnet-4-5@20250929"];
|
|
@@ -4557,8 +4256,6 @@ var xaiModels = {
|
|
|
4557
4256
|
contextWindow: 256e3,
|
|
4558
4257
|
supportsImages: true,
|
|
4559
4258
|
supportsPromptCache: true,
|
|
4560
|
-
supportsNativeTools: true,
|
|
4561
|
-
defaultToolProtocol: "native",
|
|
4562
4259
|
inputPrice: 0.2,
|
|
4563
4260
|
outputPrice: 1.5,
|
|
4564
4261
|
cacheWritesPrice: 0.02,
|
|
@@ -4572,8 +4269,6 @@ var xaiModels = {
|
|
|
4572
4269
|
contextWindow: 2e6,
|
|
4573
4270
|
supportsImages: true,
|
|
4574
4271
|
supportsPromptCache: true,
|
|
4575
|
-
supportsNativeTools: true,
|
|
4576
|
-
defaultToolProtocol: "native",
|
|
4577
4272
|
inputPrice: 0.2,
|
|
4578
4273
|
outputPrice: 0.5,
|
|
4579
4274
|
cacheWritesPrice: 0.05,
|
|
@@ -4587,8 +4282,6 @@ var xaiModels = {
|
|
|
4587
4282
|
contextWindow: 2e6,
|
|
4588
4283
|
supportsImages: true,
|
|
4589
4284
|
supportsPromptCache: true,
|
|
4590
|
-
supportsNativeTools: true,
|
|
4591
|
-
defaultToolProtocol: "native",
|
|
4592
4285
|
inputPrice: 0.2,
|
|
4593
4286
|
outputPrice: 0.5,
|
|
4594
4287
|
cacheWritesPrice: 0.05,
|
|
@@ -4602,8 +4295,6 @@ var xaiModels = {
|
|
|
4602
4295
|
contextWindow: 2e6,
|
|
4603
4296
|
supportsImages: true,
|
|
4604
4297
|
supportsPromptCache: true,
|
|
4605
|
-
supportsNativeTools: true,
|
|
4606
|
-
defaultToolProtocol: "native",
|
|
4607
4298
|
inputPrice: 0.2,
|
|
4608
4299
|
outputPrice: 0.5,
|
|
4609
4300
|
cacheWritesPrice: 0.05,
|
|
@@ -4617,8 +4308,6 @@ var xaiModels = {
|
|
|
4617
4308
|
contextWindow: 2e6,
|
|
4618
4309
|
supportsImages: true,
|
|
4619
4310
|
supportsPromptCache: true,
|
|
4620
|
-
supportsNativeTools: true,
|
|
4621
|
-
defaultToolProtocol: "native",
|
|
4622
4311
|
inputPrice: 0.2,
|
|
4623
4312
|
outputPrice: 0.5,
|
|
4624
4313
|
cacheWritesPrice: 0.05,
|
|
@@ -4632,8 +4321,6 @@ var xaiModels = {
|
|
|
4632
4321
|
contextWindow: 256e3,
|
|
4633
4322
|
supportsImages: true,
|
|
4634
4323
|
supportsPromptCache: true,
|
|
4635
|
-
supportsNativeTools: true,
|
|
4636
|
-
defaultToolProtocol: "native",
|
|
4637
4324
|
inputPrice: 3,
|
|
4638
4325
|
outputPrice: 15,
|
|
4639
4326
|
cacheWritesPrice: 0.75,
|
|
@@ -4647,8 +4334,6 @@ var xaiModels = {
|
|
|
4647
4334
|
contextWindow: 131072,
|
|
4648
4335
|
supportsImages: true,
|
|
4649
4336
|
supportsPromptCache: true,
|
|
4650
|
-
supportsNativeTools: true,
|
|
4651
|
-
defaultToolProtocol: "native",
|
|
4652
4337
|
inputPrice: 0.3,
|
|
4653
4338
|
outputPrice: 0.5,
|
|
4654
4339
|
cacheWritesPrice: 0.07,
|
|
@@ -4664,8 +4349,6 @@ var xaiModels = {
|
|
|
4664
4349
|
contextWindow: 131072,
|
|
4665
4350
|
supportsImages: true,
|
|
4666
4351
|
supportsPromptCache: true,
|
|
4667
|
-
supportsNativeTools: true,
|
|
4668
|
-
defaultToolProtocol: "native",
|
|
4669
4352
|
inputPrice: 3,
|
|
4670
4353
|
outputPrice: 15,
|
|
4671
4354
|
cacheWritesPrice: 0.75,
|
|
@@ -4762,7 +4445,6 @@ var vercelAiGatewayDefaultModelInfo = {
|
|
|
4762
4445
|
contextWindow: 2e5,
|
|
4763
4446
|
supportsImages: true,
|
|
4764
4447
|
supportsPromptCache: true,
|
|
4765
|
-
supportsNativeTools: true,
|
|
4766
4448
|
inputPrice: 3,
|
|
4767
4449
|
outputPrice: 15,
|
|
4768
4450
|
cacheWritesPrice: 3.75,
|
|
@@ -4779,8 +4461,6 @@ var internationalZAiModels = {
|
|
|
4779
4461
|
contextWindow: 131072,
|
|
4780
4462
|
supportsImages: false,
|
|
4781
4463
|
supportsPromptCache: true,
|
|
4782
|
-
supportsNativeTools: true,
|
|
4783
|
-
defaultToolProtocol: "native",
|
|
4784
4464
|
inputPrice: 0.6,
|
|
4785
4465
|
outputPrice: 2.2,
|
|
4786
4466
|
cacheWritesPrice: 0,
|
|
@@ -4792,8 +4472,6 @@ var internationalZAiModels = {
|
|
|
4792
4472
|
contextWindow: 131072,
|
|
4793
4473
|
supportsImages: false,
|
|
4794
4474
|
supportsPromptCache: true,
|
|
4795
|
-
supportsNativeTools: true,
|
|
4796
|
-
defaultToolProtocol: "native",
|
|
4797
4475
|
inputPrice: 0.2,
|
|
4798
4476
|
outputPrice: 1.1,
|
|
4799
4477
|
cacheWritesPrice: 0,
|
|
@@ -4805,8 +4483,6 @@ var internationalZAiModels = {
|
|
|
4805
4483
|
contextWindow: 131072,
|
|
4806
4484
|
supportsImages: false,
|
|
4807
4485
|
supportsPromptCache: true,
|
|
4808
|
-
supportsNativeTools: true,
|
|
4809
|
-
defaultToolProtocol: "native",
|
|
4810
4486
|
inputPrice: 2.2,
|
|
4811
4487
|
outputPrice: 8.9,
|
|
4812
4488
|
cacheWritesPrice: 0,
|
|
@@ -4818,8 +4494,6 @@ var internationalZAiModels = {
|
|
|
4818
4494
|
contextWindow: 131072,
|
|
4819
4495
|
supportsImages: false,
|
|
4820
4496
|
supportsPromptCache: true,
|
|
4821
|
-
supportsNativeTools: true,
|
|
4822
|
-
defaultToolProtocol: "native",
|
|
4823
4497
|
inputPrice: 1.1,
|
|
4824
4498
|
outputPrice: 4.5,
|
|
4825
4499
|
cacheWritesPrice: 0,
|
|
@@ -4831,8 +4505,6 @@ var internationalZAiModels = {
|
|
|
4831
4505
|
contextWindow: 131072,
|
|
4832
4506
|
supportsImages: false,
|
|
4833
4507
|
supportsPromptCache: true,
|
|
4834
|
-
supportsNativeTools: true,
|
|
4835
|
-
defaultToolProtocol: "native",
|
|
4836
4508
|
inputPrice: 0,
|
|
4837
4509
|
outputPrice: 0,
|
|
4838
4510
|
cacheWritesPrice: 0,
|
|
@@ -4844,21 +4516,28 @@ var internationalZAiModels = {
|
|
|
4844
4516
|
contextWindow: 131072,
|
|
4845
4517
|
supportsImages: true,
|
|
4846
4518
|
supportsPromptCache: true,
|
|
4847
|
-
supportsNativeTools: true,
|
|
4848
|
-
defaultToolProtocol: "native",
|
|
4849
4519
|
inputPrice: 0.6,
|
|
4850
4520
|
outputPrice: 1.8,
|
|
4851
4521
|
cacheWritesPrice: 0,
|
|
4852
4522
|
cacheReadsPrice: 0.11,
|
|
4853
4523
|
description: "GLM-4.5V is Z.AI's multimodal visual reasoning model (image/video/text/file input), optimized for GUI tasks, grounding, and document/video understanding."
|
|
4854
4524
|
},
|
|
4525
|
+
"glm-4.6v": {
|
|
4526
|
+
maxTokens: 16384,
|
|
4527
|
+
contextWindow: 131072,
|
|
4528
|
+
supportsImages: true,
|
|
4529
|
+
supportsPromptCache: true,
|
|
4530
|
+
inputPrice: 0.3,
|
|
4531
|
+
outputPrice: 0.9,
|
|
4532
|
+
cacheWritesPrice: 0,
|
|
4533
|
+
cacheReadsPrice: 0.05,
|
|
4534
|
+
description: "GLM-4.6V is an advanced multimodal vision model with improved performance and cost-efficiency for visual understanding tasks."
|
|
4535
|
+
},
|
|
4855
4536
|
"glm-4.6": {
|
|
4856
4537
|
maxTokens: 16384,
|
|
4857
4538
|
contextWindow: 2e5,
|
|
4858
4539
|
supportsImages: false,
|
|
4859
4540
|
supportsPromptCache: true,
|
|
4860
|
-
supportsNativeTools: true,
|
|
4861
|
-
defaultToolProtocol: "native",
|
|
4862
4541
|
inputPrice: 0.6,
|
|
4863
4542
|
outputPrice: 2.2,
|
|
4864
4543
|
cacheWritesPrice: 0,
|
|
@@ -4870,8 +4549,6 @@ var internationalZAiModels = {
|
|
|
4870
4549
|
contextWindow: 2e5,
|
|
4871
4550
|
supportsImages: false,
|
|
4872
4551
|
supportsPromptCache: true,
|
|
4873
|
-
supportsNativeTools: true,
|
|
4874
|
-
defaultToolProtocol: "native",
|
|
4875
4552
|
supportsReasoningEffort: ["disable", "medium"],
|
|
4876
4553
|
reasoningEffort: "medium",
|
|
4877
4554
|
preserveReasoning: true,
|
|
@@ -4881,13 +4558,55 @@ var internationalZAiModels = {
|
|
|
4881
4558
|
cacheReadsPrice: 0.11,
|
|
4882
4559
|
description: "GLM-4.7 is Zhipu's latest model with built-in thinking capabilities enabled by default. It provides enhanced reasoning for complex tasks while maintaining fast response times."
|
|
4883
4560
|
},
|
|
4561
|
+
"glm-4.7-flash": {
|
|
4562
|
+
maxTokens: 16384,
|
|
4563
|
+
contextWindow: 2e5,
|
|
4564
|
+
supportsImages: false,
|
|
4565
|
+
supportsPromptCache: true,
|
|
4566
|
+
inputPrice: 0,
|
|
4567
|
+
outputPrice: 0,
|
|
4568
|
+
cacheWritesPrice: 0,
|
|
4569
|
+
cacheReadsPrice: 0,
|
|
4570
|
+
description: "GLM-4.7-Flash is a free, high-speed variant of GLM-4.7 offering fast responses for reasoning and coding tasks."
|
|
4571
|
+
},
|
|
4572
|
+
"glm-4.7-flashx": {
|
|
4573
|
+
maxTokens: 16384,
|
|
4574
|
+
contextWindow: 2e5,
|
|
4575
|
+
supportsImages: false,
|
|
4576
|
+
supportsPromptCache: true,
|
|
4577
|
+
inputPrice: 0.07,
|
|
4578
|
+
outputPrice: 0.4,
|
|
4579
|
+
cacheWritesPrice: 0,
|
|
4580
|
+
cacheReadsPrice: 0.01,
|
|
4581
|
+
description: "GLM-4.7-FlashX is an ultra-fast variant of GLM-4.7 with exceptional speed and cost-effectiveness for high-throughput applications."
|
|
4582
|
+
},
|
|
4583
|
+
"glm-4.6v-flash": {
|
|
4584
|
+
maxTokens: 16384,
|
|
4585
|
+
contextWindow: 131072,
|
|
4586
|
+
supportsImages: true,
|
|
4587
|
+
supportsPromptCache: true,
|
|
4588
|
+
inputPrice: 0,
|
|
4589
|
+
outputPrice: 0,
|
|
4590
|
+
cacheWritesPrice: 0,
|
|
4591
|
+
cacheReadsPrice: 0,
|
|
4592
|
+
description: "GLM-4.6V-Flash is a free, high-speed multimodal vision model for rapid image understanding and visual reasoning tasks."
|
|
4593
|
+
},
|
|
4594
|
+
"glm-4.6v-flashx": {
|
|
4595
|
+
maxTokens: 16384,
|
|
4596
|
+
contextWindow: 131072,
|
|
4597
|
+
supportsImages: true,
|
|
4598
|
+
supportsPromptCache: true,
|
|
4599
|
+
inputPrice: 0.04,
|
|
4600
|
+
outputPrice: 0.4,
|
|
4601
|
+
cacheWritesPrice: 0,
|
|
4602
|
+
cacheReadsPrice: 4e-3,
|
|
4603
|
+
description: "GLM-4.6V-FlashX is an ultra-fast multimodal vision model optimized for high-speed visual processing at low cost."
|
|
4604
|
+
},
|
|
4884
4605
|
"glm-4-32b-0414-128k": {
|
|
4885
4606
|
maxTokens: 16384,
|
|
4886
4607
|
contextWindow: 131072,
|
|
4887
4608
|
supportsImages: false,
|
|
4888
4609
|
supportsPromptCache: false,
|
|
4889
|
-
supportsNativeTools: true,
|
|
4890
|
-
defaultToolProtocol: "native",
|
|
4891
4610
|
inputPrice: 0.1,
|
|
4892
4611
|
outputPrice: 0.1,
|
|
4893
4612
|
cacheWritesPrice: 0,
|
|
@@ -4902,8 +4621,6 @@ var mainlandZAiModels = {
|
|
|
4902
4621
|
contextWindow: 131072,
|
|
4903
4622
|
supportsImages: false,
|
|
4904
4623
|
supportsPromptCache: true,
|
|
4905
|
-
supportsNativeTools: true,
|
|
4906
|
-
defaultToolProtocol: "native",
|
|
4907
4624
|
inputPrice: 0.29,
|
|
4908
4625
|
outputPrice: 1.14,
|
|
4909
4626
|
cacheWritesPrice: 0,
|
|
@@ -4915,8 +4632,6 @@ var mainlandZAiModels = {
|
|
|
4915
4632
|
contextWindow: 131072,
|
|
4916
4633
|
supportsImages: false,
|
|
4917
4634
|
supportsPromptCache: true,
|
|
4918
|
-
supportsNativeTools: true,
|
|
4919
|
-
defaultToolProtocol: "native",
|
|
4920
4635
|
inputPrice: 0.1,
|
|
4921
4636
|
outputPrice: 0.6,
|
|
4922
4637
|
cacheWritesPrice: 0,
|
|
@@ -4928,8 +4643,6 @@ var mainlandZAiModels = {
|
|
|
4928
4643
|
contextWindow: 131072,
|
|
4929
4644
|
supportsImages: false,
|
|
4930
4645
|
supportsPromptCache: true,
|
|
4931
|
-
supportsNativeTools: true,
|
|
4932
|
-
defaultToolProtocol: "native",
|
|
4933
4646
|
inputPrice: 0.29,
|
|
4934
4647
|
outputPrice: 1.14,
|
|
4935
4648
|
cacheWritesPrice: 0,
|
|
@@ -4941,8 +4654,6 @@ var mainlandZAiModels = {
|
|
|
4941
4654
|
contextWindow: 131072,
|
|
4942
4655
|
supportsImages: false,
|
|
4943
4656
|
supportsPromptCache: true,
|
|
4944
|
-
supportsNativeTools: true,
|
|
4945
|
-
defaultToolProtocol: "native",
|
|
4946
4657
|
inputPrice: 0.1,
|
|
4947
4658
|
outputPrice: 0.6,
|
|
4948
4659
|
cacheWritesPrice: 0,
|
|
@@ -4954,8 +4665,6 @@ var mainlandZAiModels = {
|
|
|
4954
4665
|
contextWindow: 131072,
|
|
4955
4666
|
supportsImages: false,
|
|
4956
4667
|
supportsPromptCache: true,
|
|
4957
|
-
supportsNativeTools: true,
|
|
4958
|
-
defaultToolProtocol: "native",
|
|
4959
4668
|
inputPrice: 0,
|
|
4960
4669
|
outputPrice: 0,
|
|
4961
4670
|
cacheWritesPrice: 0,
|
|
@@ -4967,8 +4676,6 @@ var mainlandZAiModels = {
|
|
|
4967
4676
|
contextWindow: 131072,
|
|
4968
4677
|
supportsImages: true,
|
|
4969
4678
|
supportsPromptCache: true,
|
|
4970
|
-
supportsNativeTools: true,
|
|
4971
|
-
defaultToolProtocol: "native",
|
|
4972
4679
|
inputPrice: 0.29,
|
|
4973
4680
|
outputPrice: 0.93,
|
|
4974
4681
|
cacheWritesPrice: 0,
|
|
@@ -4980,8 +4687,6 @@ var mainlandZAiModels = {
|
|
|
4980
4687
|
contextWindow: 204800,
|
|
4981
4688
|
supportsImages: false,
|
|
4982
4689
|
supportsPromptCache: true,
|
|
4983
|
-
supportsNativeTools: true,
|
|
4984
|
-
defaultToolProtocol: "native",
|
|
4985
4690
|
inputPrice: 0.29,
|
|
4986
4691
|
outputPrice: 1.14,
|
|
4987
4692
|
cacheWritesPrice: 0,
|
|
@@ -4993,8 +4698,6 @@ var mainlandZAiModels = {
|
|
|
4993
4698
|
contextWindow: 204800,
|
|
4994
4699
|
supportsImages: false,
|
|
4995
4700
|
supportsPromptCache: true,
|
|
4996
|
-
supportsNativeTools: true,
|
|
4997
|
-
defaultToolProtocol: "native",
|
|
4998
4701
|
supportsReasoningEffort: ["disable", "medium"],
|
|
4999
4702
|
reasoningEffort: "medium",
|
|
5000
4703
|
preserveReasoning: true,
|
|
@@ -5003,6 +4706,61 @@ var mainlandZAiModels = {
|
|
|
5003
4706
|
cacheWritesPrice: 0,
|
|
5004
4707
|
cacheReadsPrice: 0.057,
|
|
5005
4708
|
description: "GLM-4.7 is Zhipu's latest model with built-in thinking capabilities enabled by default. It provides enhanced reasoning for complex tasks while maintaining fast response times."
|
|
4709
|
+
},
|
|
4710
|
+
"glm-4.7-flash": {
|
|
4711
|
+
maxTokens: 16384,
|
|
4712
|
+
contextWindow: 204800,
|
|
4713
|
+
supportsImages: false,
|
|
4714
|
+
supportsPromptCache: true,
|
|
4715
|
+
inputPrice: 0,
|
|
4716
|
+
outputPrice: 0,
|
|
4717
|
+
cacheWritesPrice: 0,
|
|
4718
|
+
cacheReadsPrice: 0,
|
|
4719
|
+
description: "GLM-4.7-Flash is a free, high-speed variant of GLM-4.7 offering fast responses for reasoning and coding tasks."
|
|
4720
|
+
},
|
|
4721
|
+
"glm-4.7-flashx": {
|
|
4722
|
+
maxTokens: 16384,
|
|
4723
|
+
contextWindow: 204800,
|
|
4724
|
+
supportsImages: false,
|
|
4725
|
+
supportsPromptCache: true,
|
|
4726
|
+
inputPrice: 0.035,
|
|
4727
|
+
outputPrice: 0.2,
|
|
4728
|
+
cacheWritesPrice: 0,
|
|
4729
|
+
cacheReadsPrice: 5e-3,
|
|
4730
|
+
description: "GLM-4.7-FlashX is an ultra-fast variant of GLM-4.7 with exceptional speed and cost-effectiveness for high-throughput applications."
|
|
4731
|
+
},
|
|
4732
|
+
"glm-4.6v": {
|
|
4733
|
+
maxTokens: 16384,
|
|
4734
|
+
contextWindow: 131072,
|
|
4735
|
+
supportsImages: true,
|
|
4736
|
+
supportsPromptCache: true,
|
|
4737
|
+
inputPrice: 0.15,
|
|
4738
|
+
outputPrice: 0.45,
|
|
4739
|
+
cacheWritesPrice: 0,
|
|
4740
|
+
cacheReadsPrice: 0.025,
|
|
4741
|
+
description: "GLM-4.6V is an advanced multimodal vision model with improved performance and cost-efficiency for visual understanding tasks."
|
|
4742
|
+
},
|
|
4743
|
+
"glm-4.6v-flash": {
|
|
4744
|
+
maxTokens: 16384,
|
|
4745
|
+
contextWindow: 131072,
|
|
4746
|
+
supportsImages: true,
|
|
4747
|
+
supportsPromptCache: true,
|
|
4748
|
+
inputPrice: 0,
|
|
4749
|
+
outputPrice: 0,
|
|
4750
|
+
cacheWritesPrice: 0,
|
|
4751
|
+
cacheReadsPrice: 0,
|
|
4752
|
+
description: "GLM-4.6V-Flash is a free, high-speed multimodal vision model for rapid image understanding and visual reasoning tasks."
|
|
4753
|
+
},
|
|
4754
|
+
"glm-4.6v-flashx": {
|
|
4755
|
+
maxTokens: 16384,
|
|
4756
|
+
contextWindow: 131072,
|
|
4757
|
+
supportsImages: true,
|
|
4758
|
+
supportsPromptCache: true,
|
|
4759
|
+
inputPrice: 0.02,
|
|
4760
|
+
outputPrice: 0.2,
|
|
4761
|
+
cacheWritesPrice: 0,
|
|
4762
|
+
cacheReadsPrice: 2e-3,
|
|
4763
|
+
description: "GLM-4.6V-FlashX is an ultra-fast multimodal vision model optimized for high-speed visual processing at low cost."
|
|
5006
4764
|
}
|
|
5007
4765
|
};
|
|
5008
4766
|
var ZAI_DEFAULT_TEMPERATURE = 0.6;
|
|
@@ -5036,7 +4794,6 @@ var deepInfraDefaultModelInfo = {
|
|
|
5036
4794
|
contextWindow: 262144,
|
|
5037
4795
|
supportsImages: false,
|
|
5038
4796
|
supportsPromptCache: false,
|
|
5039
|
-
supportsNativeTools: true,
|
|
5040
4797
|
inputPrice: 0.3,
|
|
5041
4798
|
outputPrice: 1.2,
|
|
5042
4799
|
description: "Qwen 3 Coder 480B A35B Instruct Turbo model, 256K context."
|
|
@@ -5050,8 +4807,6 @@ var minimaxModels = {
|
|
|
5050
4807
|
contextWindow: 192e3,
|
|
5051
4808
|
supportsImages: false,
|
|
5052
4809
|
supportsPromptCache: true,
|
|
5053
|
-
supportsNativeTools: true,
|
|
5054
|
-
defaultToolProtocol: "native",
|
|
5055
4810
|
includedTools: ["search_and_replace"],
|
|
5056
4811
|
excludedTools: ["apply_diff"],
|
|
5057
4812
|
preserveReasoning: true,
|
|
@@ -5066,8 +4821,6 @@ var minimaxModels = {
|
|
|
5066
4821
|
contextWindow: 192e3,
|
|
5067
4822
|
supportsImages: false,
|
|
5068
4823
|
supportsPromptCache: true,
|
|
5069
|
-
supportsNativeTools: true,
|
|
5070
|
-
defaultToolProtocol: "native",
|
|
5071
4824
|
includedTools: ["search_and_replace"],
|
|
5072
4825
|
excludedTools: ["apply_diff"],
|
|
5073
4826
|
preserveReasoning: true,
|
|
@@ -5082,8 +4835,6 @@ var minimaxModels = {
|
|
|
5082
4835
|
contextWindow: 192e3,
|
|
5083
4836
|
supportsImages: false,
|
|
5084
4837
|
supportsPromptCache: true,
|
|
5085
|
-
supportsNativeTools: true,
|
|
5086
|
-
defaultToolProtocol: "native",
|
|
5087
4838
|
includedTools: ["search_and_replace"],
|
|
5088
4839
|
excludedTools: ["apply_diff"],
|
|
5089
4840
|
preserveReasoning: true,
|
|
@@ -5138,6 +4889,8 @@ function getProviderDefaultModelId(provider, options = { isChina: false }) {
|
|
|
5138
4889
|
case "openai-native":
|
|
5139
4890
|
return "gpt-4o";
|
|
5140
4891
|
// Based on openai-native patterns
|
|
4892
|
+
case "openai-codex":
|
|
4893
|
+
return openAiCodexDefaultModelId;
|
|
5141
4894
|
case "mistral":
|
|
5142
4895
|
return mistralDefaultModelId;
|
|
5143
4896
|
case "openai":
|
|
@@ -5153,8 +4906,6 @@ function getProviderDefaultModelId(provider, options = { isChina: false }) {
|
|
|
5153
4906
|
return deepInfraDefaultModelId;
|
|
5154
4907
|
case "vscode-lm":
|
|
5155
4908
|
return vscodeLlmDefaultModelId;
|
|
5156
|
-
case "claude-code":
|
|
5157
|
-
return claudeCodeDefaultModelId;
|
|
5158
4909
|
case "cerebras":
|
|
5159
4910
|
return cerebrasDefaultModelId;
|
|
5160
4911
|
case "sambanova":
|
|
@@ -5212,7 +4963,6 @@ var providerNames = [
|
|
|
5212
4963
|
"bedrock",
|
|
5213
4964
|
"baseten",
|
|
5214
4965
|
"cerebras",
|
|
5215
|
-
"claude-code",
|
|
5216
4966
|
"doubao",
|
|
5217
4967
|
"deepseek",
|
|
5218
4968
|
"featherless",
|
|
@@ -5223,6 +4973,7 @@ var providerNames = [
|
|
|
5223
4973
|
"mistral",
|
|
5224
4974
|
"moonshot",
|
|
5225
4975
|
"minimax",
|
|
4976
|
+
"openai-codex",
|
|
5226
4977
|
"openai-native",
|
|
5227
4978
|
"qwen-code",
|
|
5228
4979
|
"roo",
|
|
@@ -5241,9 +4992,7 @@ var providerSettingsEntrySchema = z8.object({
|
|
|
5241
4992
|
});
|
|
5242
4993
|
var baseProviderSettingsSchema = z8.object({
|
|
5243
4994
|
includeMaxTokens: z8.boolean().optional(),
|
|
5244
|
-
diffEnabled: z8.boolean().optional(),
|
|
5245
4995
|
todoListEnabled: z8.boolean().optional(),
|
|
5246
|
-
fuzzyMatchThreshold: z8.number().optional(),
|
|
5247
4996
|
modelTemperature: z8.number().nullish(),
|
|
5248
4997
|
rateLimitSeconds: z8.number().optional(),
|
|
5249
4998
|
consecutiveMistakeLimit: z8.number().min(0).optional(),
|
|
@@ -5253,9 +5002,7 @@ var baseProviderSettingsSchema = z8.object({
|
|
|
5253
5002
|
modelMaxTokens: z8.number().optional(),
|
|
5254
5003
|
modelMaxThinkingTokens: z8.number().optional(),
|
|
5255
5004
|
// Model verbosity.
|
|
5256
|
-
verbosity: verbosityLevelsSchema.optional()
|
|
5257
|
-
// Tool protocol override for this profile.
|
|
5258
|
-
toolProtocol: z8.enum(["xml", "native"]).optional()
|
|
5005
|
+
verbosity: verbosityLevelsSchema.optional()
|
|
5259
5006
|
});
|
|
5260
5007
|
var apiModelIdProviderModelSchema = baseProviderSettingsSchema.extend({
|
|
5261
5008
|
apiModelId: z8.string().optional()
|
|
@@ -5267,7 +5014,6 @@ var anthropicSchema = apiModelIdProviderModelSchema.extend({
|
|
|
5267
5014
|
anthropicBeta1MContext: z8.boolean().optional()
|
|
5268
5015
|
// Enable 'context-1m-2025-08-07' beta for 1M context window.
|
|
5269
5016
|
});
|
|
5270
|
-
var claudeCodeSchema = apiModelIdProviderModelSchema.extend({});
|
|
5271
5017
|
var openRouterSchema = baseProviderSettingsSchema.extend({
|
|
5272
5018
|
openRouterApiKey: z8.string().optional(),
|
|
5273
5019
|
openRouterModelId: z8.string().optional(),
|
|
@@ -5349,6 +5095,9 @@ var geminiCliSchema = apiModelIdProviderModelSchema.extend({
|
|
|
5349
5095
|
geminiCliOAuthPath: z8.string().optional(),
|
|
5350
5096
|
geminiCliProjectId: z8.string().optional()
|
|
5351
5097
|
});
|
|
5098
|
+
var openAiCodexSchema = apiModelIdProviderModelSchema.extend({
|
|
5099
|
+
// No additional settings needed - uses OAuth authentication
|
|
5100
|
+
});
|
|
5352
5101
|
var openAiNativeSchema = apiModelIdProviderModelSchema.extend({
|
|
5353
5102
|
openAiNativeApiKey: z8.string().optional(),
|
|
5354
5103
|
openAiNativeBaseUrl: z8.string().optional(),
|
|
@@ -5453,7 +5202,6 @@ var defaultSchema = z8.object({
|
|
|
5453
5202
|
});
|
|
5454
5203
|
var providerSettingsSchemaDiscriminated = z8.discriminatedUnion("apiProvider", [
|
|
5455
5204
|
anthropicSchema.merge(z8.object({ apiProvider: z8.literal("anthropic") })),
|
|
5456
|
-
claudeCodeSchema.merge(z8.object({ apiProvider: z8.literal("claude-code") })),
|
|
5457
5205
|
openRouterSchema.merge(z8.object({ apiProvider: z8.literal("openrouter") })),
|
|
5458
5206
|
bedrockSchema.merge(z8.object({ apiProvider: z8.literal("bedrock") })),
|
|
5459
5207
|
vertexSchema.merge(z8.object({ apiProvider: z8.literal("vertex") })),
|
|
@@ -5463,6 +5211,7 @@ var providerSettingsSchemaDiscriminated = z8.discriminatedUnion("apiProvider", [
|
|
|
5463
5211
|
lmStudioSchema.merge(z8.object({ apiProvider: z8.literal("lmstudio") })),
|
|
5464
5212
|
geminiSchema.merge(z8.object({ apiProvider: z8.literal("gemini") })),
|
|
5465
5213
|
geminiCliSchema.merge(z8.object({ apiProvider: z8.literal("gemini-cli") })),
|
|
5214
|
+
openAiCodexSchema.merge(z8.object({ apiProvider: z8.literal("openai-codex") })),
|
|
5466
5215
|
openAiNativeSchema.merge(z8.object({ apiProvider: z8.literal("openai-native") })),
|
|
5467
5216
|
mistralSchema.merge(z8.object({ apiProvider: z8.literal("mistral") })),
|
|
5468
5217
|
deepSeekSchema.merge(z8.object({ apiProvider: z8.literal("deepseek") })),
|
|
@@ -5493,7 +5242,6 @@ var providerSettingsSchemaDiscriminated = z8.discriminatedUnion("apiProvider", [
|
|
|
5493
5242
|
var providerSettingsSchema = z8.object({
|
|
5494
5243
|
apiProvider: providerNamesSchema.optional(),
|
|
5495
5244
|
...anthropicSchema.shape,
|
|
5496
|
-
...claudeCodeSchema.shape,
|
|
5497
5245
|
...openRouterSchema.shape,
|
|
5498
5246
|
...bedrockSchema.shape,
|
|
5499
5247
|
...vertexSchema.shape,
|
|
@@ -5503,6 +5251,7 @@ var providerSettingsSchema = z8.object({
|
|
|
5503
5251
|
...lmStudioSchema.shape,
|
|
5504
5252
|
...geminiSchema.shape,
|
|
5505
5253
|
...geminiCliSchema.shape,
|
|
5254
|
+
...openAiCodexSchema.shape,
|
|
5506
5255
|
...openAiNativeSchema.shape,
|
|
5507
5256
|
...mistralSchema.shape,
|
|
5508
5257
|
...deepSeekSchema.shape,
|
|
@@ -5557,10 +5306,10 @@ var getModelId = (settings) => {
|
|
|
5557
5306
|
var isTypicalProvider = (key) => isProviderName(key) && !isInternalProvider(key) && !isCustomProvider(key) && !isFauxProvider(key);
|
|
5558
5307
|
var modelIdKeysByProvider = {
|
|
5559
5308
|
anthropic: "apiModelId",
|
|
5560
|
-
"claude-code": "apiModelId",
|
|
5561
5309
|
openrouter: "openRouterModelId",
|
|
5562
5310
|
bedrock: "apiModelId",
|
|
5563
5311
|
vertex: "apiModelId",
|
|
5312
|
+
"openai-codex": "apiModelId",
|
|
5564
5313
|
"openai-native": "openAiModelId",
|
|
5565
5314
|
ollama: "ollamaModelId",
|
|
5566
5315
|
lmstudio: "lmStudioModelId",
|
|
@@ -5590,7 +5339,7 @@ var modelIdKeysByProvider = {
|
|
|
5590
5339
|
roo: "apiModelId",
|
|
5591
5340
|
"vercel-ai-gateway": "vercelAiGatewayModelId"
|
|
5592
5341
|
};
|
|
5593
|
-
var ANTHROPIC_STYLE_PROVIDERS = ["anthropic", "
|
|
5342
|
+
var ANTHROPIC_STYLE_PROVIDERS = ["anthropic", "bedrock", "minimax"];
|
|
5594
5343
|
var getApiProtocol = (provider, modelId) => {
|
|
5595
5344
|
if (provider && ANTHROPIC_STYLE_PROVIDERS.includes(provider)) {
|
|
5596
5345
|
return "anthropic";
|
|
@@ -5619,7 +5368,6 @@ var MODELS_BY_PROVIDER = {
|
|
|
5619
5368
|
label: "Cerebras",
|
|
5620
5369
|
models: Object.keys(cerebrasModels)
|
|
5621
5370
|
},
|
|
5622
|
-
"claude-code": { id: "claude-code", label: "Claude Code", models: Object.keys(claudeCodeModels) },
|
|
5623
5371
|
deepseek: {
|
|
5624
5372
|
id: "deepseek",
|
|
5625
5373
|
label: "DeepSeek",
|
|
@@ -5662,6 +5410,11 @@ var MODELS_BY_PROVIDER = {
|
|
|
5662
5410
|
label: "MiniMax",
|
|
5663
5411
|
models: Object.keys(minimaxModels)
|
|
5664
5412
|
},
|
|
5413
|
+
"openai-codex": {
|
|
5414
|
+
id: "openai-codex",
|
|
5415
|
+
label: "OpenAI - ChatGPT Plus/Pro",
|
|
5416
|
+
models: Object.keys(openAiCodexModels)
|
|
5417
|
+
},
|
|
5665
5418
|
"openai-native": {
|
|
5666
5419
|
id: "openai-native",
|
|
5667
5420
|
label: "OpenAI",
|
|
@@ -5718,16 +5471,6 @@ var historyItemSchema = z9.object({
|
|
|
5718
5471
|
size: z9.number().optional(),
|
|
5719
5472
|
workspace: z9.string().optional(),
|
|
5720
5473
|
mode: z9.string().optional(),
|
|
5721
|
-
/**
|
|
5722
|
-
* The tool protocol used by this task. Once a task uses tools with a specific
|
|
5723
|
-
* protocol (XML or Native), it is permanently locked to that protocol.
|
|
5724
|
-
*
|
|
5725
|
-
* - "xml": Tool calls are parsed from XML text (no tool IDs)
|
|
5726
|
-
* - "native": Tool calls come as tool_call chunks with IDs
|
|
5727
|
-
*
|
|
5728
|
-
* This ensures task resumption works correctly even when NTC settings change.
|
|
5729
|
-
*/
|
|
5730
|
-
toolProtocol: z9.enum(["xml", "native"]).optional(),
|
|
5731
5474
|
apiConfigName: z9.string().optional(),
|
|
5732
5475
|
// Provider profile name for sticky profile feature
|
|
5733
5476
|
status: z9.enum(["active", "completed", "delegated"]).optional(),
|
|
@@ -5745,23 +5488,12 @@ var historyItemSchema = z9.object({
|
|
|
5745
5488
|
|
|
5746
5489
|
// src/experiment.ts
|
|
5747
5490
|
import { z as z10 } from "zod";
|
|
5748
|
-
var experimentIds = [
|
|
5749
|
-
"powerSteering",
|
|
5750
|
-
"multiFileApplyDiff",
|
|
5751
|
-
"preventFocusDisruption",
|
|
5752
|
-
"imageGeneration",
|
|
5753
|
-
"runSlashCommand",
|
|
5754
|
-
"multipleNativeToolCalls",
|
|
5755
|
-
"customTools"
|
|
5756
|
-
];
|
|
5491
|
+
var experimentIds = ["preventFocusDisruption", "imageGeneration", "runSlashCommand", "customTools"];
|
|
5757
5492
|
var experimentIdsSchema = z10.enum(experimentIds);
|
|
5758
5493
|
var experimentsSchema = z10.object({
|
|
5759
|
-
powerSteering: z10.boolean().optional(),
|
|
5760
|
-
multiFileApplyDiff: z10.boolean().optional(),
|
|
5761
5494
|
preventFocusDisruption: z10.boolean().optional(),
|
|
5762
5495
|
imageGeneration: z10.boolean().optional(),
|
|
5763
5496
|
runSlashCommand: z10.boolean().optional(),
|
|
5764
|
-
multipleNativeToolCalls: z10.boolean().optional(),
|
|
5765
5497
|
customTools: z10.boolean().optional()
|
|
5766
5498
|
});
|
|
5767
5499
|
|
|
@@ -5813,6 +5545,7 @@ var TelemetryEventName = /* @__PURE__ */ ((TelemetryEventName2) => {
|
|
|
5813
5545
|
TelemetryEventName2["CODE_INDEX_ERROR"] = "Code Index Error";
|
|
5814
5546
|
TelemetryEventName2["TELEMETRY_SETTINGS_CHANGED"] = "Telemetry Settings Changed";
|
|
5815
5547
|
TelemetryEventName2["MODEL_CACHE_EMPTY_RESPONSE"] = "Model Cache Empty Response";
|
|
5548
|
+
TelemetryEventName2["READ_FILE_LEGACY_FORMAT_USED"] = "Read File Legacy Format Used";
|
|
5816
5549
|
return TelemetryEventName2;
|
|
5817
5550
|
})(TelemetryEventName || {});
|
|
5818
5551
|
var staticAppPropertiesSchema = z11.object({
|
|
@@ -5901,7 +5634,8 @@ var rooCodeTelemetryEventSchema = z11.discriminatedUnion("type", [
|
|
|
5901
5634
|
"Sliding Window Truncation" /* SLIDING_WINDOW_TRUNCATION */,
|
|
5902
5635
|
"Tab Shown" /* TAB_SHOWN */,
|
|
5903
5636
|
"Mode Setting Changed" /* MODE_SETTINGS_CHANGED */,
|
|
5904
|
-
"Custom Mode Created" /* CUSTOM_MODE_CREATED
|
|
5637
|
+
"Custom Mode Created" /* CUSTOM_MODE_CREATED */,
|
|
5638
|
+
"Read File Legacy Format Used" /* READ_FILE_LEGACY_FORMAT_USED */
|
|
5905
5639
|
]),
|
|
5906
5640
|
properties: telemetryPropertiesSchema
|
|
5907
5641
|
}),
|
|
@@ -6215,7 +5949,15 @@ var isLanguage = (value) => languages.includes(value);
|
|
|
6215
5949
|
|
|
6216
5950
|
// src/global-settings.ts
|
|
6217
5951
|
var DEFAULT_WRITE_DELAY_MS = 1e3;
|
|
6218
|
-
var
|
|
5952
|
+
var TERMINAL_PREVIEW_BYTES = {
|
|
5953
|
+
small: 5 * 1024,
|
|
5954
|
+
// 5KB
|
|
5955
|
+
medium: 10 * 1024,
|
|
5956
|
+
// 10KB
|
|
5957
|
+
large: 20 * 1024
|
|
5958
|
+
// 20KB
|
|
5959
|
+
};
|
|
5960
|
+
var DEFAULT_TERMINAL_OUTPUT_PREVIEW_SIZE = "medium";
|
|
6219
5961
|
var MIN_CHECKPOINT_TIMEOUT_SECONDS = 10;
|
|
6220
5962
|
var MAX_CHECKPOINT_TIMEOUT_SECONDS = 60;
|
|
6221
5963
|
var DEFAULT_CHECKPOINT_TIMEOUT_SECONDS = 15;
|
|
@@ -6231,7 +5973,6 @@ var globalSettingsSchema = z14.object({
|
|
|
6231
5973
|
imageGenerationProvider: z14.enum(["openrouter", "roo"]).optional(),
|
|
6232
5974
|
openRouterImageApiKey: z14.string().optional(),
|
|
6233
5975
|
openRouterImageGenerationSelectedModel: z14.string().optional(),
|
|
6234
|
-
condensingApiConfigId: z14.string().optional(),
|
|
6235
5976
|
customCondensingPrompt: z14.string().optional(),
|
|
6236
5977
|
autoApprovalEnabled: z14.boolean().optional(),
|
|
6237
5978
|
alwaysAllowReadOnly: z14.boolean().optional(),
|
|
@@ -6257,7 +5998,6 @@ var globalSettingsSchema = z14.object({
|
|
|
6257
5998
|
allowedMaxCost: z14.number().nullish(),
|
|
6258
5999
|
autoCondenseContext: z14.boolean().optional(),
|
|
6259
6000
|
autoCondenseContextPercent: z14.number().optional(),
|
|
6260
|
-
maxConcurrentFileReads: z14.number().optional(),
|
|
6261
6001
|
/**
|
|
6262
6002
|
* Whether to include current time in the environment details
|
|
6263
6003
|
* @default true
|
|
@@ -6300,11 +6040,9 @@ var globalSettingsSchema = z14.object({
|
|
|
6300
6040
|
maxWorkspaceFiles: z14.number().optional(),
|
|
6301
6041
|
showRooIgnoredFiles: z14.boolean().optional(),
|
|
6302
6042
|
enableSubfolderRules: z14.boolean().optional(),
|
|
6303
|
-
maxReadFileLine: z14.number().optional(),
|
|
6304
6043
|
maxImageFileSize: z14.number().optional(),
|
|
6305
6044
|
maxTotalImageSize: z14.number().optional(),
|
|
6306
|
-
|
|
6307
|
-
terminalOutputCharacterLimit: z14.number().optional(),
|
|
6045
|
+
terminalOutputPreviewSize: z14.enum(["small", "medium", "large"]).optional(),
|
|
6308
6046
|
terminalShellIntegrationTimeout: z14.number().optional(),
|
|
6309
6047
|
terminalShellIntegrationDisabled: z14.boolean().optional(),
|
|
6310
6048
|
terminalCommandDelay: z14.number().optional(),
|
|
@@ -6313,18 +6051,14 @@ var globalSettingsSchema = z14.object({
|
|
|
6313
6051
|
terminalZshOhMy: z14.boolean().optional(),
|
|
6314
6052
|
terminalZshP10k: z14.boolean().optional(),
|
|
6315
6053
|
terminalZdotdir: z14.boolean().optional(),
|
|
6316
|
-
terminalCompressProgressBar: z14.boolean().optional(),
|
|
6317
6054
|
diagnosticsEnabled: z14.boolean().optional(),
|
|
6318
6055
|
rateLimitSeconds: z14.number().optional(),
|
|
6319
|
-
diffEnabled: z14.boolean().optional(),
|
|
6320
|
-
fuzzyMatchThreshold: z14.number().optional(),
|
|
6321
6056
|
experiments: experimentsSchema.optional(),
|
|
6322
6057
|
codebaseIndexModels: codebaseIndexModelsSchema.optional(),
|
|
6323
6058
|
codebaseIndexConfig: codebaseIndexConfigSchema.optional(),
|
|
6324
6059
|
language: languagesSchema.optional(),
|
|
6325
6060
|
telemetrySetting: telemetrySettingsSchema.optional(),
|
|
6326
6061
|
mcpEnabled: z14.boolean().optional(),
|
|
6327
|
-
enableMcpServerCreation: z14.boolean().optional(),
|
|
6328
6062
|
mode: z14.string().optional(),
|
|
6329
6063
|
modeApiConfigs: z14.record(z14.string(), z14.string()).optional(),
|
|
6330
6064
|
customModes: z14.array(modeConfigSchema).optional(),
|
|
@@ -6344,7 +6078,20 @@ var globalSettingsSchema = z14.object({
|
|
|
6344
6078
|
profileThresholds: z14.record(z14.string(), z14.number()).optional(),
|
|
6345
6079
|
hasOpenedModeSelector: z14.boolean().optional(),
|
|
6346
6080
|
lastModeExportPath: z14.string().optional(),
|
|
6347
|
-
lastModeImportPath: z14.string().optional()
|
|
6081
|
+
lastModeImportPath: z14.string().optional(),
|
|
6082
|
+
lastSettingsExportPath: z14.string().optional(),
|
|
6083
|
+
lastTaskExportPath: z14.string().optional(),
|
|
6084
|
+
lastImageSavePath: z14.string().optional(),
|
|
6085
|
+
/**
|
|
6086
|
+
* Path to worktree to auto-open after switching workspaces.
|
|
6087
|
+
* Used by the worktree feature to open the Roo Code sidebar in a new window.
|
|
6088
|
+
*/
|
|
6089
|
+
worktreeAutoOpenPath: z14.string().optional(),
|
|
6090
|
+
/**
|
|
6091
|
+
* Whether to show the worktree selector in the home screen.
|
|
6092
|
+
* @default true
|
|
6093
|
+
*/
|
|
6094
|
+
showWorktreesInHomeScreen: z14.boolean().optional()
|
|
6348
6095
|
});
|
|
6349
6096
|
var GLOBAL_SETTINGS_KEYS = globalSettingsSchema.keyof().options;
|
|
6350
6097
|
var rooCodeSettingsSchema = providerSettingsSchema.merge(globalSettingsSchema);
|
|
@@ -6428,8 +6175,6 @@ var EVALS_SETTINGS = {
|
|
|
6428
6175
|
ttsSpeed: 1,
|
|
6429
6176
|
soundEnabled: false,
|
|
6430
6177
|
soundVolume: 0.5,
|
|
6431
|
-
terminalOutputLineLimit: 500,
|
|
6432
|
-
terminalOutputCharacterLimit: DEFAULT_TERMINAL_OUTPUT_CHARACTER_LIMIT,
|
|
6433
6178
|
terminalShellIntegrationTimeout: 3e4,
|
|
6434
6179
|
terminalCommandDelay: 0,
|
|
6435
6180
|
terminalPowershellCounter: false,
|
|
@@ -6437,19 +6182,14 @@ var EVALS_SETTINGS = {
|
|
|
6437
6182
|
terminalZshClearEolMark: true,
|
|
6438
6183
|
terminalZshP10k: false,
|
|
6439
6184
|
terminalZdotdir: true,
|
|
6440
|
-
terminalCompressProgressBar: true,
|
|
6441
6185
|
terminalShellIntegrationDisabled: true,
|
|
6442
6186
|
diagnosticsEnabled: true,
|
|
6443
|
-
diffEnabled: true,
|
|
6444
|
-
fuzzyMatchThreshold: 1,
|
|
6445
6187
|
enableCheckpoints: false,
|
|
6446
6188
|
rateLimitSeconds: 0,
|
|
6447
6189
|
maxOpenTabsContext: 20,
|
|
6448
6190
|
maxWorkspaceFiles: 200,
|
|
6449
6191
|
maxGitStatusFiles: 20,
|
|
6450
6192
|
showRooIgnoredFiles: true,
|
|
6451
|
-
maxReadFileLine: -1,
|
|
6452
|
-
// -1 to enable full file reading.
|
|
6453
6193
|
includeDiagnosticMessages: true,
|
|
6454
6194
|
maxDiagnosticMessages: 50,
|
|
6455
6195
|
language: "en",
|
|
@@ -6523,24 +6263,18 @@ var organizationAllowListSchema = z16.object({
|
|
|
6523
6263
|
});
|
|
6524
6264
|
var organizationDefaultSettingsSchema = globalSettingsSchema.pick({
|
|
6525
6265
|
enableCheckpoints: true,
|
|
6526
|
-
fuzzyMatchThreshold: true,
|
|
6527
6266
|
maxOpenTabsContext: true,
|
|
6528
|
-
maxReadFileLine: true,
|
|
6529
6267
|
maxWorkspaceFiles: true,
|
|
6530
6268
|
showRooIgnoredFiles: true,
|
|
6531
6269
|
terminalCommandDelay: true,
|
|
6532
|
-
terminalCompressProgressBar: true,
|
|
6533
|
-
terminalOutputLineLimit: true,
|
|
6534
6270
|
terminalShellIntegrationDisabled: true,
|
|
6535
6271
|
terminalShellIntegrationTimeout: true,
|
|
6536
6272
|
terminalZshClearEolMark: true
|
|
6537
6273
|
}).merge(
|
|
6538
6274
|
z16.object({
|
|
6539
6275
|
maxOpenTabsContext: z16.number().int().nonnegative().optional(),
|
|
6540
|
-
maxReadFileLine: z16.number().int().gte(-1).optional(),
|
|
6541
6276
|
maxWorkspaceFiles: z16.number().int().nonnegative().optional(),
|
|
6542
6277
|
terminalCommandDelay: z16.number().int().nonnegative().optional(),
|
|
6543
|
-
terminalOutputLineLimit: z16.number().int().nonnegative().optional(),
|
|
6544
6278
|
terminalShellIntegrationTimeout: z16.number().int().nonnegative().optional()
|
|
6545
6279
|
})
|
|
6546
6280
|
);
|
|
@@ -6573,7 +6307,8 @@ var userFeaturesSchema = z16.object({
|
|
|
6573
6307
|
});
|
|
6574
6308
|
var userSettingsConfigSchema = z16.object({
|
|
6575
6309
|
extensionBridgeEnabled: z16.boolean().optional(),
|
|
6576
|
-
taskSyncEnabled: z16.boolean().optional()
|
|
6310
|
+
taskSyncEnabled: z16.boolean().optional(),
|
|
6311
|
+
llmEnhancedFeaturesEnabled: z16.boolean().optional()
|
|
6577
6312
|
});
|
|
6578
6313
|
var userSettingsDataSchema = z16.object({
|
|
6579
6314
|
features: userFeaturesSchema,
|
|
@@ -7014,12 +6749,10 @@ var taskCommandSchema = z19.discriminatedUnion("commandName", [
|
|
|
7014
6749
|
})
|
|
7015
6750
|
}),
|
|
7016
6751
|
z19.object({
|
|
7017
|
-
commandName: z19.literal("CancelTask" /* CancelTask */)
|
|
7018
|
-
data: z19.string()
|
|
6752
|
+
commandName: z19.literal("CancelTask" /* CancelTask */)
|
|
7019
6753
|
}),
|
|
7020
6754
|
z19.object({
|
|
7021
|
-
commandName: z19.literal("CloseTask" /* CloseTask */)
|
|
7022
|
-
data: z19.string()
|
|
6755
|
+
commandName: z19.literal("CloseTask" /* CloseTask */)
|
|
7023
6756
|
}),
|
|
7024
6757
|
z19.object({
|
|
7025
6758
|
commandName: z19.literal("ResumeTask" /* ResumeTask */),
|
|
@@ -7055,6 +6788,7 @@ var ipcMessageSchema = z19.discriminatedUnion("type", [
|
|
|
7055
6788
|
|
|
7056
6789
|
// src/mcp.ts
|
|
7057
6790
|
import { z as z20 } from "zod";
|
|
6791
|
+
var MAX_MCP_TOOLS_THRESHOLD = 60;
|
|
7058
6792
|
var mcpExecutionStatusSchema = z20.discriminatedUnion("status", [
|
|
7059
6793
|
z20.object({
|
|
7060
6794
|
executionId: z20.string(),
|
|
@@ -7078,6 +6812,46 @@ var mcpExecutionStatusSchema = z20.discriminatedUnion("status", [
|
|
|
7078
6812
|
error: z20.string().optional()
|
|
7079
6813
|
})
|
|
7080
6814
|
]);
|
|
6815
|
+
function countEnabledMcpTools(servers) {
|
|
6816
|
+
let serverCount = 0;
|
|
6817
|
+
let toolCount = 0;
|
|
6818
|
+
for (const server of servers) {
|
|
6819
|
+
if (server.disabled) continue;
|
|
6820
|
+
if (server.status !== "connected") continue;
|
|
6821
|
+
serverCount++;
|
|
6822
|
+
if (server.tools) {
|
|
6823
|
+
for (const tool of server.tools) {
|
|
6824
|
+
if (tool.enabledForPrompt !== false) {
|
|
6825
|
+
toolCount++;
|
|
6826
|
+
}
|
|
6827
|
+
}
|
|
6828
|
+
}
|
|
6829
|
+
}
|
|
6830
|
+
return { enabledToolCount: toolCount, enabledServerCount: serverCount };
|
|
6831
|
+
}
|
|
6832
|
+
|
|
6833
|
+
// src/skills.ts
|
|
6834
|
+
var SKILL_NAME_MIN_LENGTH = 1;
|
|
6835
|
+
var SKILL_NAME_MAX_LENGTH = 64;
|
|
6836
|
+
var SKILL_NAME_REGEX = /^[a-z0-9]+(?:-[a-z0-9]+)*$/;
|
|
6837
|
+
var SkillNameValidationError = /* @__PURE__ */ ((SkillNameValidationError2) => {
|
|
6838
|
+
SkillNameValidationError2["Empty"] = "empty";
|
|
6839
|
+
SkillNameValidationError2["TooLong"] = "too_long";
|
|
6840
|
+
SkillNameValidationError2["InvalidFormat"] = "invalid_format";
|
|
6841
|
+
return SkillNameValidationError2;
|
|
6842
|
+
})(SkillNameValidationError || {});
|
|
6843
|
+
function validateSkillName(name) {
|
|
6844
|
+
if (!name || name.length < SKILL_NAME_MIN_LENGTH) {
|
|
6845
|
+
return { valid: false, error: "empty" /* Empty */ };
|
|
6846
|
+
}
|
|
6847
|
+
if (name.length > SKILL_NAME_MAX_LENGTH) {
|
|
6848
|
+
return { valid: false, error: "too_long" /* TooLong */ };
|
|
6849
|
+
}
|
|
6850
|
+
if (!SKILL_NAME_REGEX.test(name)) {
|
|
6851
|
+
return { valid: false, error: "invalid_format" /* InvalidFormat */ };
|
|
6852
|
+
}
|
|
6853
|
+
return { valid: true };
|
|
6854
|
+
}
|
|
7081
6855
|
|
|
7082
6856
|
// src/todo.ts
|
|
7083
6857
|
import { z as z21 } from "zod";
|
|
@@ -7117,6 +6891,11 @@ var commandExecutionStatusSchema = z22.discriminatedUnion("status", [
|
|
|
7117
6891
|
})
|
|
7118
6892
|
]);
|
|
7119
6893
|
|
|
6894
|
+
// src/tool-params.ts
|
|
6895
|
+
function isLegacyReadFileParams(params) {
|
|
6896
|
+
return "_legacyFormat" in params && params._legacyFormat === true;
|
|
6897
|
+
}
|
|
6898
|
+
|
|
7120
6899
|
// src/vscode-extension-host.ts
|
|
7121
6900
|
import { z as z23 } from "zod";
|
|
7122
6901
|
var checkoutDiffPayloadSchema = z23.object({
|
|
@@ -7169,7 +6948,7 @@ export {
|
|
|
7169
6948
|
DEFAULT_CHECKPOINT_TIMEOUT_SECONDS,
|
|
7170
6949
|
DEFAULT_CONSECUTIVE_MISTAKE_LIMIT,
|
|
7171
6950
|
DEFAULT_MODES,
|
|
7172
|
-
|
|
6951
|
+
DEFAULT_TERMINAL_OUTPUT_PREVIEW_SIZE,
|
|
7173
6952
|
DEFAULT_WRITE_DELAY_MS,
|
|
7174
6953
|
DOUBAO_API_BASE_URL,
|
|
7175
6954
|
DOUBAO_API_CHAT_PATH,
|
|
@@ -7199,13 +6978,13 @@ export {
|
|
|
7199
6978
|
IpcOrigin,
|
|
7200
6979
|
LMSTUDIO_DEFAULT_TEMPERATURE,
|
|
7201
6980
|
MAX_CHECKPOINT_TIMEOUT_SECONDS,
|
|
6981
|
+
MAX_MCP_TOOLS_THRESHOLD,
|
|
7202
6982
|
MINIMAX_DEFAULT_MAX_TOKENS,
|
|
7203
6983
|
MINIMAX_DEFAULT_TEMPERATURE,
|
|
7204
6984
|
MIN_CHECKPOINT_TIMEOUT_SECONDS,
|
|
7205
6985
|
MISTRAL_DEFAULT_TEMPERATURE,
|
|
7206
6986
|
MODELS_BY_PROVIDER,
|
|
7207
6987
|
MOONSHOT_DEFAULT_TEMPERATURE,
|
|
7208
|
-
NATIVE_TOOL_DEFAULTS,
|
|
7209
6988
|
OPENAI_AZURE_AI_INFERENCE_PATH,
|
|
7210
6989
|
OPENAI_NATIVE_DEFAULT_TEMPERATURE,
|
|
7211
6990
|
OPENROUTER_DEFAULT_PROVIDER_NAME,
|
|
@@ -7220,7 +6999,11 @@ export {
|
|
|
7220
6999
|
RooModelsResponseSchema,
|
|
7221
7000
|
RooPricingSchema,
|
|
7222
7001
|
SECRET_STATE_KEYS,
|
|
7223
|
-
|
|
7002
|
+
SKILL_NAME_MAX_LENGTH,
|
|
7003
|
+
SKILL_NAME_MIN_LENGTH,
|
|
7004
|
+
SKILL_NAME_REGEX,
|
|
7005
|
+
SkillNameValidationError,
|
|
7006
|
+
TERMINAL_PREVIEW_BYTES,
|
|
7224
7007
|
TaskBridgeCommandName,
|
|
7225
7008
|
TaskBridgeEventName,
|
|
7226
7009
|
TaskCommandName,
|
|
@@ -7252,9 +7035,6 @@ export {
|
|
|
7252
7035
|
chutesDefaultModelId,
|
|
7253
7036
|
chutesDefaultModelInfo,
|
|
7254
7037
|
chutesModels,
|
|
7255
|
-
claudeCodeDefaultModelId,
|
|
7256
|
-
claudeCodeModels,
|
|
7257
|
-
claudeCodeReasoningConfig,
|
|
7258
7038
|
clineAskSchema,
|
|
7259
7039
|
clineAsks,
|
|
7260
7040
|
clineMessageSchema,
|
|
@@ -7269,6 +7049,7 @@ export {
|
|
|
7269
7049
|
commandIds,
|
|
7270
7050
|
contextCondenseSchema,
|
|
7271
7051
|
contextTruncationSchema,
|
|
7052
|
+
countEnabledMcpTools,
|
|
7272
7053
|
customModePromptsSchema,
|
|
7273
7054
|
customModesSettingsSchema,
|
|
7274
7055
|
customProviders,
|
|
@@ -7302,7 +7083,6 @@ export {
|
|
|
7302
7083
|
geminiDefaultModelId,
|
|
7303
7084
|
geminiModels,
|
|
7304
7085
|
getApiProtocol,
|
|
7305
|
-
getEffectiveProtocol,
|
|
7306
7086
|
getErrorMessage,
|
|
7307
7087
|
getErrorStatusCode,
|
|
7308
7088
|
getImageGenerationProvider,
|
|
@@ -7337,9 +7117,9 @@ export {
|
|
|
7337
7117
|
isInteractiveAsk,
|
|
7338
7118
|
isInternalProvider,
|
|
7339
7119
|
isLanguage,
|
|
7120
|
+
isLegacyReadFileParams,
|
|
7340
7121
|
isLocalProvider,
|
|
7341
7122
|
isModelParameter,
|
|
7342
|
-
isNativeProtocol,
|
|
7343
7123
|
isNonBlockingAsk,
|
|
7344
7124
|
isProviderName,
|
|
7345
7125
|
isResumableAsk,
|
|
@@ -7375,9 +7155,10 @@ export {
|
|
|
7375
7155
|
moonshotDefaultModelId,
|
|
7376
7156
|
moonshotModels,
|
|
7377
7157
|
nonBlockingAsks,
|
|
7378
|
-
normalizeClaudeCodeModelId,
|
|
7379
7158
|
ollamaDefaultModelId,
|
|
7380
7159
|
ollamaDefaultModelInfo,
|
|
7160
|
+
openAiCodexDefaultModelId,
|
|
7161
|
+
openAiCodexModels,
|
|
7381
7162
|
openAiModelInfoSaneDefaults,
|
|
7382
7163
|
openAiNativeDefaultModelId,
|
|
7383
7164
|
openAiNativeModels,
|
|
@@ -7447,6 +7228,7 @@ export {
|
|
|
7447
7228
|
userFeaturesSchema,
|
|
7448
7229
|
userSettingsConfigSchema,
|
|
7449
7230
|
userSettingsDataSchema,
|
|
7231
|
+
validateSkillName,
|
|
7450
7232
|
verbosityLevels,
|
|
7451
7233
|
verbosityLevelsSchema,
|
|
7452
7234
|
vercelAiGatewayDefaultModelId,
|