@roo-code/types 1.106.0 → 1.108.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/index.cjs +386 -714
- package/dist/index.cjs.map +1 -1
- package/dist/index.d.cts +1430 -2530
- package/dist/index.d.ts +1430 -2530
- package/dist/index.js +376 -705
- package/dist/index.js.map +1 -1
- package/package.json +2 -2
package/dist/index.js
CHANGED
|
@@ -78,7 +78,9 @@ var clineSays = [
|
|
|
78
78
|
"condense_context_error",
|
|
79
79
|
"sliding_window_truncation",
|
|
80
80
|
"codebase_search_result",
|
|
81
|
-
"user_edit_todos"
|
|
81
|
+
"user_edit_todos",
|
|
82
|
+
"too_many_tools_warning",
|
|
83
|
+
"tool"
|
|
82
84
|
];
|
|
83
85
|
var clineSaySchema = z.enum(clineSays);
|
|
84
86
|
var toolProgressStatusSchema = z.object({
|
|
@@ -146,6 +148,7 @@ var toolGroupsSchema = z2.enum(toolGroups);
|
|
|
146
148
|
var toolNames = [
|
|
147
149
|
"execute_command",
|
|
148
150
|
"read_file",
|
|
151
|
+
"read_command_output",
|
|
149
152
|
"write_to_file",
|
|
150
153
|
"apply_diff",
|
|
151
154
|
"search_and_replace",
|
|
@@ -161,10 +164,10 @@ var toolNames = [
|
|
|
161
164
|
"attempt_completion",
|
|
162
165
|
"switch_mode",
|
|
163
166
|
"new_task",
|
|
164
|
-
"fetch_instructions",
|
|
165
167
|
"codebase_search",
|
|
166
168
|
"update_todo_list",
|
|
167
169
|
"run_slash_command",
|
|
170
|
+
"skill",
|
|
168
171
|
"generate_image",
|
|
169
172
|
"custom_tool"
|
|
170
173
|
];
|
|
@@ -176,20 +179,6 @@ var toolUsageSchema = z2.record(
|
|
|
176
179
|
failures: z2.number()
|
|
177
180
|
})
|
|
178
181
|
);
|
|
179
|
-
var TOOL_PROTOCOL = {
|
|
180
|
-
XML: "xml",
|
|
181
|
-
NATIVE: "native"
|
|
182
|
-
};
|
|
183
|
-
var NATIVE_TOOL_DEFAULTS = {
|
|
184
|
-
supportsNativeTools: true,
|
|
185
|
-
defaultToolProtocol: TOOL_PROTOCOL.NATIVE
|
|
186
|
-
};
|
|
187
|
-
function isNativeProtocol(protocol) {
|
|
188
|
-
return protocol === TOOL_PROTOCOL.NATIVE;
|
|
189
|
-
}
|
|
190
|
-
function getEffectiveProtocol(toolProtocol) {
|
|
191
|
-
return toolProtocol || TOOL_PROTOCOL.XML;
|
|
192
|
-
}
|
|
193
182
|
|
|
194
183
|
// src/events.ts
|
|
195
184
|
var RooCodeEventName = /* @__PURE__ */ ((RooCodeEventName2) => {
|
|
@@ -477,10 +466,6 @@ var modelInfoSchema = z5.object({
|
|
|
477
466
|
isStealthModel: z5.boolean().optional(),
|
|
478
467
|
// Flag to indicate if the model is free (no cost)
|
|
479
468
|
isFree: z5.boolean().optional(),
|
|
480
|
-
// Flag to indicate if the model supports native tool calling (OpenAI-style function calling)
|
|
481
|
-
supportsNativeTools: z5.boolean().optional(),
|
|
482
|
-
// Default tool protocol preferred by this model (if not specified, falls back to capability/provider defaults)
|
|
483
|
-
defaultToolProtocol: z5.enum(["xml", "native"]).optional(),
|
|
484
469
|
// Exclude specific native tools from being available (only applies to native protocol)
|
|
485
470
|
// These tools will be removed from the set of tools available to the model
|
|
486
471
|
excludedTools: z5.array(z5.string()).optional(),
|
|
@@ -577,8 +562,6 @@ var anthropicModels = {
|
|
|
577
562
|
// Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
|
|
578
563
|
supportsImages: true,
|
|
579
564
|
supportsPromptCache: true,
|
|
580
|
-
supportsNativeTools: true,
|
|
581
|
-
defaultToolProtocol: "native",
|
|
582
565
|
inputPrice: 3,
|
|
583
566
|
// $3 per million input tokens (≤200K context)
|
|
584
567
|
outputPrice: 15,
|
|
@@ -611,8 +594,6 @@ var anthropicModels = {
|
|
|
611
594
|
// Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
|
|
612
595
|
supportsImages: true,
|
|
613
596
|
supportsPromptCache: true,
|
|
614
|
-
supportsNativeTools: true,
|
|
615
|
-
defaultToolProtocol: "native",
|
|
616
597
|
inputPrice: 3,
|
|
617
598
|
// $3 per million input tokens (≤200K context)
|
|
618
599
|
outputPrice: 15,
|
|
@@ -644,8 +625,6 @@ var anthropicModels = {
|
|
|
644
625
|
contextWindow: 2e5,
|
|
645
626
|
supportsImages: true,
|
|
646
627
|
supportsPromptCache: true,
|
|
647
|
-
supportsNativeTools: true,
|
|
648
|
-
defaultToolProtocol: "native",
|
|
649
628
|
inputPrice: 5,
|
|
650
629
|
// $5 per million input tokens
|
|
651
630
|
outputPrice: 25,
|
|
@@ -662,8 +641,6 @@ var anthropicModels = {
|
|
|
662
641
|
contextWindow: 2e5,
|
|
663
642
|
supportsImages: true,
|
|
664
643
|
supportsPromptCache: true,
|
|
665
|
-
supportsNativeTools: true,
|
|
666
|
-
defaultToolProtocol: "native",
|
|
667
644
|
inputPrice: 15,
|
|
668
645
|
// $15 per million input tokens
|
|
669
646
|
outputPrice: 75,
|
|
@@ -680,8 +657,6 @@ var anthropicModels = {
|
|
|
680
657
|
contextWindow: 2e5,
|
|
681
658
|
supportsImages: true,
|
|
682
659
|
supportsPromptCache: true,
|
|
683
|
-
supportsNativeTools: true,
|
|
684
|
-
defaultToolProtocol: "native",
|
|
685
660
|
inputPrice: 15,
|
|
686
661
|
// $15 per million input tokens
|
|
687
662
|
outputPrice: 75,
|
|
@@ -698,8 +673,6 @@ var anthropicModels = {
|
|
|
698
673
|
contextWindow: 2e5,
|
|
699
674
|
supportsImages: true,
|
|
700
675
|
supportsPromptCache: true,
|
|
701
|
-
supportsNativeTools: true,
|
|
702
|
-
defaultToolProtocol: "native",
|
|
703
676
|
inputPrice: 3,
|
|
704
677
|
// $3 per million input tokens
|
|
705
678
|
outputPrice: 15,
|
|
@@ -717,8 +690,6 @@ var anthropicModels = {
|
|
|
717
690
|
contextWindow: 2e5,
|
|
718
691
|
supportsImages: true,
|
|
719
692
|
supportsPromptCache: true,
|
|
720
|
-
supportsNativeTools: true,
|
|
721
|
-
defaultToolProtocol: "native",
|
|
722
693
|
inputPrice: 3,
|
|
723
694
|
// $3 per million input tokens
|
|
724
695
|
outputPrice: 15,
|
|
@@ -733,8 +704,6 @@ var anthropicModels = {
|
|
|
733
704
|
contextWindow: 2e5,
|
|
734
705
|
supportsImages: true,
|
|
735
706
|
supportsPromptCache: true,
|
|
736
|
-
supportsNativeTools: true,
|
|
737
|
-
defaultToolProtocol: "native",
|
|
738
707
|
inputPrice: 3,
|
|
739
708
|
// $3 per million input tokens
|
|
740
709
|
outputPrice: 15,
|
|
@@ -749,8 +718,6 @@ var anthropicModels = {
|
|
|
749
718
|
contextWindow: 2e5,
|
|
750
719
|
supportsImages: false,
|
|
751
720
|
supportsPromptCache: true,
|
|
752
|
-
supportsNativeTools: true,
|
|
753
|
-
defaultToolProtocol: "native",
|
|
754
721
|
inputPrice: 1,
|
|
755
722
|
outputPrice: 5,
|
|
756
723
|
cacheWritesPrice: 1.25,
|
|
@@ -761,8 +728,6 @@ var anthropicModels = {
|
|
|
761
728
|
contextWindow: 2e5,
|
|
762
729
|
supportsImages: true,
|
|
763
730
|
supportsPromptCache: true,
|
|
764
|
-
supportsNativeTools: true,
|
|
765
|
-
defaultToolProtocol: "native",
|
|
766
731
|
inputPrice: 15,
|
|
767
732
|
outputPrice: 75,
|
|
768
733
|
cacheWritesPrice: 18.75,
|
|
@@ -773,8 +738,6 @@ var anthropicModels = {
|
|
|
773
738
|
contextWindow: 2e5,
|
|
774
739
|
supportsImages: true,
|
|
775
740
|
supportsPromptCache: true,
|
|
776
|
-
supportsNativeTools: true,
|
|
777
|
-
defaultToolProtocol: "native",
|
|
778
741
|
inputPrice: 0.25,
|
|
779
742
|
outputPrice: 1.25,
|
|
780
743
|
cacheWritesPrice: 0.3,
|
|
@@ -785,8 +748,6 @@ var anthropicModels = {
|
|
|
785
748
|
contextWindow: 2e5,
|
|
786
749
|
supportsImages: true,
|
|
787
750
|
supportsPromptCache: true,
|
|
788
|
-
supportsNativeTools: true,
|
|
789
|
-
defaultToolProtocol: "native",
|
|
790
751
|
inputPrice: 1,
|
|
791
752
|
outputPrice: 5,
|
|
792
753
|
cacheWritesPrice: 1.25,
|
|
@@ -804,7 +765,6 @@ var basetenModels = {
|
|
|
804
765
|
contextWindow: 262e3,
|
|
805
766
|
supportsImages: false,
|
|
806
767
|
supportsPromptCache: false,
|
|
807
|
-
supportsNativeTools: true,
|
|
808
768
|
inputPrice: 0.6,
|
|
809
769
|
outputPrice: 2.5,
|
|
810
770
|
cacheWritesPrice: 0,
|
|
@@ -816,7 +776,6 @@ var basetenModels = {
|
|
|
816
776
|
contextWindow: 2e5,
|
|
817
777
|
supportsImages: false,
|
|
818
778
|
supportsPromptCache: false,
|
|
819
|
-
supportsNativeTools: true,
|
|
820
779
|
inputPrice: 0.6,
|
|
821
780
|
outputPrice: 2.2,
|
|
822
781
|
cacheWritesPrice: 0,
|
|
@@ -828,7 +787,6 @@ var basetenModels = {
|
|
|
828
787
|
contextWindow: 163840,
|
|
829
788
|
supportsImages: false,
|
|
830
789
|
supportsPromptCache: false,
|
|
831
|
-
supportsNativeTools: true,
|
|
832
790
|
inputPrice: 2.55,
|
|
833
791
|
outputPrice: 5.95,
|
|
834
792
|
cacheWritesPrice: 0,
|
|
@@ -840,7 +798,6 @@ var basetenModels = {
|
|
|
840
798
|
contextWindow: 163840,
|
|
841
799
|
supportsImages: false,
|
|
842
800
|
supportsPromptCache: false,
|
|
843
|
-
supportsNativeTools: true,
|
|
844
801
|
inputPrice: 2.55,
|
|
845
802
|
outputPrice: 5.95,
|
|
846
803
|
cacheWritesPrice: 0,
|
|
@@ -852,7 +809,6 @@ var basetenModels = {
|
|
|
852
809
|
contextWindow: 163840,
|
|
853
810
|
supportsImages: false,
|
|
854
811
|
supportsPromptCache: false,
|
|
855
|
-
supportsNativeTools: true,
|
|
856
812
|
inputPrice: 0.77,
|
|
857
813
|
outputPrice: 0.77,
|
|
858
814
|
cacheWritesPrice: 0,
|
|
@@ -864,7 +820,6 @@ var basetenModels = {
|
|
|
864
820
|
contextWindow: 163840,
|
|
865
821
|
supportsImages: false,
|
|
866
822
|
supportsPromptCache: false,
|
|
867
|
-
supportsNativeTools: true,
|
|
868
823
|
inputPrice: 0.5,
|
|
869
824
|
outputPrice: 1.5,
|
|
870
825
|
cacheWritesPrice: 0,
|
|
@@ -876,7 +831,6 @@ var basetenModels = {
|
|
|
876
831
|
contextWindow: 163840,
|
|
877
832
|
supportsImages: false,
|
|
878
833
|
supportsPromptCache: false,
|
|
879
|
-
supportsNativeTools: true,
|
|
880
834
|
inputPrice: 0.3,
|
|
881
835
|
outputPrice: 0.45,
|
|
882
836
|
cacheWritesPrice: 0,
|
|
@@ -888,7 +842,6 @@ var basetenModels = {
|
|
|
888
842
|
contextWindow: 128072,
|
|
889
843
|
supportsImages: false,
|
|
890
844
|
supportsPromptCache: false,
|
|
891
|
-
supportsNativeTools: true,
|
|
892
845
|
inputPrice: 0.1,
|
|
893
846
|
outputPrice: 0.5,
|
|
894
847
|
cacheWritesPrice: 0,
|
|
@@ -900,7 +853,6 @@ var basetenModels = {
|
|
|
900
853
|
contextWindow: 262144,
|
|
901
854
|
supportsImages: false,
|
|
902
855
|
supportsPromptCache: false,
|
|
903
|
-
supportsNativeTools: true,
|
|
904
856
|
inputPrice: 0.22,
|
|
905
857
|
outputPrice: 0.8,
|
|
906
858
|
cacheWritesPrice: 0,
|
|
@@ -912,7 +864,6 @@ var basetenModels = {
|
|
|
912
864
|
contextWindow: 262144,
|
|
913
865
|
supportsImages: false,
|
|
914
866
|
supportsPromptCache: false,
|
|
915
|
-
supportsNativeTools: true,
|
|
916
867
|
inputPrice: 0.38,
|
|
917
868
|
outputPrice: 1.53,
|
|
918
869
|
cacheWritesPrice: 0,
|
|
@@ -924,7 +875,6 @@ var basetenModels = {
|
|
|
924
875
|
contextWindow: 262e3,
|
|
925
876
|
supportsImages: false,
|
|
926
877
|
supportsPromptCache: false,
|
|
927
|
-
supportsNativeTools: true,
|
|
928
878
|
inputPrice: 0.6,
|
|
929
879
|
outputPrice: 2.5,
|
|
930
880
|
cacheWritesPrice: 0,
|
|
@@ -944,8 +894,6 @@ var bedrockModels = {
|
|
|
944
894
|
supportsImages: true,
|
|
945
895
|
supportsPromptCache: true,
|
|
946
896
|
supportsReasoningBudget: true,
|
|
947
|
-
supportsNativeTools: true,
|
|
948
|
-
defaultToolProtocol: "native",
|
|
949
897
|
inputPrice: 3,
|
|
950
898
|
outputPrice: 15,
|
|
951
899
|
cacheWritesPrice: 3.75,
|
|
@@ -959,7 +907,6 @@ var bedrockModels = {
|
|
|
959
907
|
contextWindow: 3e5,
|
|
960
908
|
supportsImages: true,
|
|
961
909
|
supportsPromptCache: true,
|
|
962
|
-
supportsNativeTools: true,
|
|
963
910
|
inputPrice: 0.8,
|
|
964
911
|
outputPrice: 3.2,
|
|
965
912
|
cacheWritesPrice: 0.8,
|
|
@@ -975,7 +922,6 @@ var bedrockModels = {
|
|
|
975
922
|
contextWindow: 3e5,
|
|
976
923
|
supportsImages: true,
|
|
977
924
|
supportsPromptCache: false,
|
|
978
|
-
supportsNativeTools: true,
|
|
979
925
|
inputPrice: 1,
|
|
980
926
|
outputPrice: 4,
|
|
981
927
|
cacheWritesPrice: 1,
|
|
@@ -989,7 +935,6 @@ var bedrockModels = {
|
|
|
989
935
|
contextWindow: 3e5,
|
|
990
936
|
supportsImages: true,
|
|
991
937
|
supportsPromptCache: true,
|
|
992
|
-
supportsNativeTools: true,
|
|
993
938
|
inputPrice: 0.06,
|
|
994
939
|
outputPrice: 0.24,
|
|
995
940
|
cacheWritesPrice: 0.06,
|
|
@@ -1005,7 +950,6 @@ var bedrockModels = {
|
|
|
1005
950
|
contextWindow: 1e6,
|
|
1006
951
|
supportsImages: true,
|
|
1007
952
|
supportsPromptCache: true,
|
|
1008
|
-
supportsNativeTools: true,
|
|
1009
953
|
inputPrice: 0.33,
|
|
1010
954
|
outputPrice: 2.75,
|
|
1011
955
|
cacheWritesPrice: 0,
|
|
@@ -1021,7 +965,6 @@ var bedrockModels = {
|
|
|
1021
965
|
contextWindow: 128e3,
|
|
1022
966
|
supportsImages: false,
|
|
1023
967
|
supportsPromptCache: true,
|
|
1024
|
-
supportsNativeTools: true,
|
|
1025
968
|
inputPrice: 0.035,
|
|
1026
969
|
outputPrice: 0.14,
|
|
1027
970
|
cacheWritesPrice: 0.035,
|
|
@@ -1038,8 +981,6 @@ var bedrockModels = {
|
|
|
1038
981
|
supportsImages: true,
|
|
1039
982
|
supportsPromptCache: true,
|
|
1040
983
|
supportsReasoningBudget: true,
|
|
1041
|
-
supportsNativeTools: true,
|
|
1042
|
-
defaultToolProtocol: "native",
|
|
1043
984
|
inputPrice: 3,
|
|
1044
985
|
outputPrice: 15,
|
|
1045
986
|
cacheWritesPrice: 3.75,
|
|
@@ -1054,8 +995,6 @@ var bedrockModels = {
|
|
|
1054
995
|
supportsImages: true,
|
|
1055
996
|
supportsPromptCache: true,
|
|
1056
997
|
supportsReasoningBudget: true,
|
|
1057
|
-
supportsNativeTools: true,
|
|
1058
|
-
defaultToolProtocol: "native",
|
|
1059
998
|
inputPrice: 15,
|
|
1060
999
|
outputPrice: 75,
|
|
1061
1000
|
cacheWritesPrice: 18.75,
|
|
@@ -1070,8 +1009,6 @@ var bedrockModels = {
|
|
|
1070
1009
|
supportsImages: true,
|
|
1071
1010
|
supportsPromptCache: true,
|
|
1072
1011
|
supportsReasoningBudget: true,
|
|
1073
|
-
supportsNativeTools: true,
|
|
1074
|
-
defaultToolProtocol: "native",
|
|
1075
1012
|
inputPrice: 5,
|
|
1076
1013
|
outputPrice: 25,
|
|
1077
1014
|
cacheWritesPrice: 6.25,
|
|
@@ -1086,8 +1023,6 @@ var bedrockModels = {
|
|
|
1086
1023
|
supportsImages: true,
|
|
1087
1024
|
supportsPromptCache: true,
|
|
1088
1025
|
supportsReasoningBudget: true,
|
|
1089
|
-
supportsNativeTools: true,
|
|
1090
|
-
defaultToolProtocol: "native",
|
|
1091
1026
|
inputPrice: 15,
|
|
1092
1027
|
outputPrice: 75,
|
|
1093
1028
|
cacheWritesPrice: 18.75,
|
|
@@ -1102,8 +1037,6 @@ var bedrockModels = {
|
|
|
1102
1037
|
supportsImages: true,
|
|
1103
1038
|
supportsPromptCache: true,
|
|
1104
1039
|
supportsReasoningBudget: true,
|
|
1105
|
-
supportsNativeTools: true,
|
|
1106
|
-
defaultToolProtocol: "native",
|
|
1107
1040
|
inputPrice: 3,
|
|
1108
1041
|
outputPrice: 15,
|
|
1109
1042
|
cacheWritesPrice: 3.75,
|
|
@@ -1117,8 +1050,6 @@ var bedrockModels = {
|
|
|
1117
1050
|
contextWindow: 2e5,
|
|
1118
1051
|
supportsImages: true,
|
|
1119
1052
|
supportsPromptCache: true,
|
|
1120
|
-
supportsNativeTools: true,
|
|
1121
|
-
defaultToolProtocol: "native",
|
|
1122
1053
|
inputPrice: 3,
|
|
1123
1054
|
outputPrice: 15,
|
|
1124
1055
|
cacheWritesPrice: 3.75,
|
|
@@ -1132,8 +1063,6 @@ var bedrockModels = {
|
|
|
1132
1063
|
contextWindow: 2e5,
|
|
1133
1064
|
supportsImages: false,
|
|
1134
1065
|
supportsPromptCache: true,
|
|
1135
|
-
supportsNativeTools: true,
|
|
1136
|
-
defaultToolProtocol: "native",
|
|
1137
1066
|
inputPrice: 0.8,
|
|
1138
1067
|
outputPrice: 4,
|
|
1139
1068
|
cacheWritesPrice: 1,
|
|
@@ -1148,8 +1077,6 @@ var bedrockModels = {
|
|
|
1148
1077
|
supportsImages: true,
|
|
1149
1078
|
supportsPromptCache: true,
|
|
1150
1079
|
supportsReasoningBudget: true,
|
|
1151
|
-
supportsNativeTools: true,
|
|
1152
|
-
defaultToolProtocol: "native",
|
|
1153
1080
|
inputPrice: 1,
|
|
1154
1081
|
outputPrice: 5,
|
|
1155
1082
|
cacheWritesPrice: 1.25,
|
|
@@ -1165,8 +1092,6 @@ var bedrockModels = {
|
|
|
1165
1092
|
contextWindow: 2e5,
|
|
1166
1093
|
supportsImages: true,
|
|
1167
1094
|
supportsPromptCache: false,
|
|
1168
|
-
supportsNativeTools: true,
|
|
1169
|
-
defaultToolProtocol: "native",
|
|
1170
1095
|
inputPrice: 3,
|
|
1171
1096
|
outputPrice: 15
|
|
1172
1097
|
},
|
|
@@ -1175,8 +1100,6 @@ var bedrockModels = {
|
|
|
1175
1100
|
contextWindow: 2e5,
|
|
1176
1101
|
supportsImages: true,
|
|
1177
1102
|
supportsPromptCache: false,
|
|
1178
|
-
supportsNativeTools: true,
|
|
1179
|
-
defaultToolProtocol: "native",
|
|
1180
1103
|
inputPrice: 15,
|
|
1181
1104
|
outputPrice: 75
|
|
1182
1105
|
},
|
|
@@ -1185,8 +1108,6 @@ var bedrockModels = {
|
|
|
1185
1108
|
contextWindow: 2e5,
|
|
1186
1109
|
supportsImages: true,
|
|
1187
1110
|
supportsPromptCache: false,
|
|
1188
|
-
supportsNativeTools: true,
|
|
1189
|
-
defaultToolProtocol: "native",
|
|
1190
1111
|
inputPrice: 3,
|
|
1191
1112
|
outputPrice: 15
|
|
1192
1113
|
},
|
|
@@ -1195,8 +1116,6 @@ var bedrockModels = {
|
|
|
1195
1116
|
contextWindow: 2e5,
|
|
1196
1117
|
supportsImages: true,
|
|
1197
1118
|
supportsPromptCache: false,
|
|
1198
|
-
supportsNativeTools: true,
|
|
1199
|
-
defaultToolProtocol: "native",
|
|
1200
1119
|
inputPrice: 0.25,
|
|
1201
1120
|
outputPrice: 1.25
|
|
1202
1121
|
},
|
|
@@ -1205,7 +1124,6 @@ var bedrockModels = {
|
|
|
1205
1124
|
contextWindow: 128e3,
|
|
1206
1125
|
supportsImages: false,
|
|
1207
1126
|
supportsPromptCache: false,
|
|
1208
|
-
supportsNativeTools: true,
|
|
1209
1127
|
inputPrice: 1.35,
|
|
1210
1128
|
outputPrice: 5.4
|
|
1211
1129
|
},
|
|
@@ -1214,7 +1132,6 @@ var bedrockModels = {
|
|
|
1214
1132
|
contextWindow: 128e3,
|
|
1215
1133
|
supportsImages: false,
|
|
1216
1134
|
supportsPromptCache: false,
|
|
1217
|
-
supportsNativeTools: true,
|
|
1218
1135
|
inputPrice: 0.5,
|
|
1219
1136
|
outputPrice: 1.5,
|
|
1220
1137
|
description: "GPT-OSS 20B - Optimized for low latency and local/specialized use cases"
|
|
@@ -1224,7 +1141,6 @@ var bedrockModels = {
|
|
|
1224
1141
|
contextWindow: 128e3,
|
|
1225
1142
|
supportsImages: false,
|
|
1226
1143
|
supportsPromptCache: false,
|
|
1227
|
-
supportsNativeTools: true,
|
|
1228
1144
|
inputPrice: 2,
|
|
1229
1145
|
outputPrice: 6,
|
|
1230
1146
|
description: "GPT-OSS 120B - Production-ready, general-purpose, high-reasoning model"
|
|
@@ -1234,7 +1150,6 @@ var bedrockModels = {
|
|
|
1234
1150
|
contextWindow: 128e3,
|
|
1235
1151
|
supportsImages: false,
|
|
1236
1152
|
supportsPromptCache: false,
|
|
1237
|
-
supportsNativeTools: true,
|
|
1238
1153
|
inputPrice: 0.72,
|
|
1239
1154
|
outputPrice: 0.72,
|
|
1240
1155
|
description: "Llama 3.3 Instruct (70B)"
|
|
@@ -1244,7 +1159,6 @@ var bedrockModels = {
|
|
|
1244
1159
|
contextWindow: 128e3,
|
|
1245
1160
|
supportsImages: true,
|
|
1246
1161
|
supportsPromptCache: false,
|
|
1247
|
-
supportsNativeTools: true,
|
|
1248
1162
|
inputPrice: 0.72,
|
|
1249
1163
|
outputPrice: 0.72,
|
|
1250
1164
|
description: "Llama 3.2 Instruct (90B)"
|
|
@@ -1254,7 +1168,6 @@ var bedrockModels = {
|
|
|
1254
1168
|
contextWindow: 128e3,
|
|
1255
1169
|
supportsImages: true,
|
|
1256
1170
|
supportsPromptCache: false,
|
|
1257
|
-
supportsNativeTools: true,
|
|
1258
1171
|
inputPrice: 0.16,
|
|
1259
1172
|
outputPrice: 0.16,
|
|
1260
1173
|
description: "Llama 3.2 Instruct (11B)"
|
|
@@ -1264,7 +1177,6 @@ var bedrockModels = {
|
|
|
1264
1177
|
contextWindow: 128e3,
|
|
1265
1178
|
supportsImages: false,
|
|
1266
1179
|
supportsPromptCache: false,
|
|
1267
|
-
supportsNativeTools: true,
|
|
1268
1180
|
inputPrice: 0.15,
|
|
1269
1181
|
outputPrice: 0.15,
|
|
1270
1182
|
description: "Llama 3.2 Instruct (3B)"
|
|
@@ -1274,7 +1186,6 @@ var bedrockModels = {
|
|
|
1274
1186
|
contextWindow: 128e3,
|
|
1275
1187
|
supportsImages: false,
|
|
1276
1188
|
supportsPromptCache: false,
|
|
1277
|
-
supportsNativeTools: true,
|
|
1278
1189
|
inputPrice: 0.1,
|
|
1279
1190
|
outputPrice: 0.1,
|
|
1280
1191
|
description: "Llama 3.2 Instruct (1B)"
|
|
@@ -1284,7 +1195,6 @@ var bedrockModels = {
|
|
|
1284
1195
|
contextWindow: 128e3,
|
|
1285
1196
|
supportsImages: false,
|
|
1286
1197
|
supportsPromptCache: false,
|
|
1287
|
-
supportsNativeTools: true,
|
|
1288
1198
|
inputPrice: 2.4,
|
|
1289
1199
|
outputPrice: 2.4,
|
|
1290
1200
|
description: "Llama 3.1 Instruct (405B)"
|
|
@@ -1294,7 +1204,6 @@ var bedrockModels = {
|
|
|
1294
1204
|
contextWindow: 128e3,
|
|
1295
1205
|
supportsImages: false,
|
|
1296
1206
|
supportsPromptCache: false,
|
|
1297
|
-
supportsNativeTools: true,
|
|
1298
1207
|
inputPrice: 0.72,
|
|
1299
1208
|
outputPrice: 0.72,
|
|
1300
1209
|
description: "Llama 3.1 Instruct (70B)"
|
|
@@ -1304,7 +1213,6 @@ var bedrockModels = {
|
|
|
1304
1213
|
contextWindow: 128e3,
|
|
1305
1214
|
supportsImages: false,
|
|
1306
1215
|
supportsPromptCache: false,
|
|
1307
|
-
supportsNativeTools: true,
|
|
1308
1216
|
inputPrice: 0.9,
|
|
1309
1217
|
outputPrice: 0.9,
|
|
1310
1218
|
description: "Llama 3.1 Instruct (70B) (w/ latency optimized inference)"
|
|
@@ -1314,7 +1222,6 @@ var bedrockModels = {
|
|
|
1314
1222
|
contextWindow: 8e3,
|
|
1315
1223
|
supportsImages: false,
|
|
1316
1224
|
supportsPromptCache: false,
|
|
1317
|
-
supportsNativeTools: true,
|
|
1318
1225
|
inputPrice: 0.22,
|
|
1319
1226
|
outputPrice: 0.22,
|
|
1320
1227
|
description: "Llama 3.1 Instruct (8B)"
|
|
@@ -1324,7 +1231,6 @@ var bedrockModels = {
|
|
|
1324
1231
|
contextWindow: 8e3,
|
|
1325
1232
|
supportsImages: false,
|
|
1326
1233
|
supportsPromptCache: false,
|
|
1327
|
-
supportsNativeTools: true,
|
|
1328
1234
|
inputPrice: 2.65,
|
|
1329
1235
|
outputPrice: 3.5
|
|
1330
1236
|
},
|
|
@@ -1333,7 +1239,6 @@ var bedrockModels = {
|
|
|
1333
1239
|
contextWindow: 4e3,
|
|
1334
1240
|
supportsImages: false,
|
|
1335
1241
|
supportsPromptCache: false,
|
|
1336
|
-
supportsNativeTools: true,
|
|
1337
1242
|
inputPrice: 0.3,
|
|
1338
1243
|
outputPrice: 0.6
|
|
1339
1244
|
},
|
|
@@ -1342,7 +1247,6 @@ var bedrockModels = {
|
|
|
1342
1247
|
contextWindow: 8e3,
|
|
1343
1248
|
supportsImages: false,
|
|
1344
1249
|
supportsPromptCache: false,
|
|
1345
|
-
supportsNativeTools: true,
|
|
1346
1250
|
inputPrice: 0.15,
|
|
1347
1251
|
outputPrice: 0.2,
|
|
1348
1252
|
description: "Amazon Titan Text Lite"
|
|
@@ -1352,7 +1256,6 @@ var bedrockModels = {
|
|
|
1352
1256
|
contextWindow: 8e3,
|
|
1353
1257
|
supportsImages: false,
|
|
1354
1258
|
supportsPromptCache: false,
|
|
1355
|
-
supportsNativeTools: true,
|
|
1356
1259
|
inputPrice: 0.2,
|
|
1357
1260
|
outputPrice: 0.6,
|
|
1358
1261
|
description: "Amazon Titan Text Express"
|
|
@@ -1362,8 +1265,6 @@ var bedrockModels = {
|
|
|
1362
1265
|
contextWindow: 262144,
|
|
1363
1266
|
supportsImages: false,
|
|
1364
1267
|
supportsPromptCache: false,
|
|
1365
|
-
supportsNativeTools: true,
|
|
1366
|
-
defaultToolProtocol: "native",
|
|
1367
1268
|
preserveReasoning: true,
|
|
1368
1269
|
inputPrice: 0.6,
|
|
1369
1270
|
outputPrice: 2.5,
|
|
@@ -1374,8 +1275,6 @@ var bedrockModels = {
|
|
|
1374
1275
|
contextWindow: 196608,
|
|
1375
1276
|
supportsImages: false,
|
|
1376
1277
|
supportsPromptCache: false,
|
|
1377
|
-
supportsNativeTools: true,
|
|
1378
|
-
defaultToolProtocol: "native",
|
|
1379
1278
|
preserveReasoning: true,
|
|
1380
1279
|
inputPrice: 0.3,
|
|
1381
1280
|
outputPrice: 1.2,
|
|
@@ -1386,8 +1285,6 @@ var bedrockModels = {
|
|
|
1386
1285
|
contextWindow: 262144,
|
|
1387
1286
|
supportsImages: false,
|
|
1388
1287
|
supportsPromptCache: false,
|
|
1389
|
-
supportsNativeTools: true,
|
|
1390
|
-
defaultToolProtocol: "native",
|
|
1391
1288
|
inputPrice: 0.15,
|
|
1392
1289
|
outputPrice: 1.2,
|
|
1393
1290
|
description: "Qwen3 Next 80B (MoE model with 3B active parameters)"
|
|
@@ -1397,8 +1294,6 @@ var bedrockModels = {
|
|
|
1397
1294
|
contextWindow: 262144,
|
|
1398
1295
|
supportsImages: false,
|
|
1399
1296
|
supportsPromptCache: false,
|
|
1400
|
-
supportsNativeTools: true,
|
|
1401
|
-
defaultToolProtocol: "native",
|
|
1402
1297
|
inputPrice: 0.45,
|
|
1403
1298
|
outputPrice: 1.8,
|
|
1404
1299
|
description: "Qwen3 Coder 480B (MoE model with 35B active parameters)"
|
|
@@ -1489,26 +1384,14 @@ var BEDROCK_SERVICE_TIER_PRICING = {
|
|
|
1489
1384
|
// src/providers/cerebras.ts
|
|
1490
1385
|
var cerebrasDefaultModelId = "gpt-oss-120b";
|
|
1491
1386
|
var cerebrasModels = {
|
|
1492
|
-
"zai-glm-4.6": {
|
|
1493
|
-
maxTokens: 16384,
|
|
1494
|
-
// Conservative default to avoid premature rate limiting (Cerebras reserves quota upfront)
|
|
1495
|
-
contextWindow: 131072,
|
|
1496
|
-
supportsImages: false,
|
|
1497
|
-
supportsPromptCache: false,
|
|
1498
|
-
supportsNativeTools: true,
|
|
1499
|
-
defaultToolProtocol: "native",
|
|
1500
|
-
inputPrice: 0,
|
|
1501
|
-
outputPrice: 0,
|
|
1502
|
-
description: "Fast general-purpose model on Cerebras (up to 1,000 tokens/s). To be deprecated soon."
|
|
1503
|
-
},
|
|
1504
1387
|
"zai-glm-4.7": {
|
|
1505
1388
|
maxTokens: 16384,
|
|
1506
1389
|
// Conservative default to avoid premature rate limiting (Cerebras reserves quota upfront)
|
|
1507
1390
|
contextWindow: 131072,
|
|
1508
1391
|
supportsImages: false,
|
|
1509
|
-
supportsPromptCache:
|
|
1510
|
-
|
|
1511
|
-
|
|
1392
|
+
supportsPromptCache: true,
|
|
1393
|
+
supportsTemperature: true,
|
|
1394
|
+
defaultTemperature: 1,
|
|
1512
1395
|
inputPrice: 0,
|
|
1513
1396
|
outputPrice: 0,
|
|
1514
1397
|
description: "Highly capable general-purpose model on Cerebras (up to 1,000 tokens/s), competitive with leading proprietary models on coding tasks."
|
|
@@ -1519,8 +1402,6 @@ var cerebrasModels = {
|
|
|
1519
1402
|
contextWindow: 64e3,
|
|
1520
1403
|
supportsImages: false,
|
|
1521
1404
|
supportsPromptCache: false,
|
|
1522
|
-
supportsNativeTools: true,
|
|
1523
|
-
defaultToolProtocol: "native",
|
|
1524
1405
|
inputPrice: 0,
|
|
1525
1406
|
outputPrice: 0,
|
|
1526
1407
|
description: "Intelligent model with ~1400 tokens/s"
|
|
@@ -1531,8 +1412,6 @@ var cerebrasModels = {
|
|
|
1531
1412
|
contextWindow: 64e3,
|
|
1532
1413
|
supportsImages: false,
|
|
1533
1414
|
supportsPromptCache: false,
|
|
1534
|
-
supportsNativeTools: true,
|
|
1535
|
-
defaultToolProtocol: "native",
|
|
1536
1415
|
inputPrice: 0,
|
|
1537
1416
|
outputPrice: 0,
|
|
1538
1417
|
description: "Powerful model with ~2600 tokens/s"
|
|
@@ -1543,8 +1422,6 @@ var cerebrasModels = {
|
|
|
1543
1422
|
contextWindow: 64e3,
|
|
1544
1423
|
supportsImages: false,
|
|
1545
1424
|
supportsPromptCache: false,
|
|
1546
|
-
supportsNativeTools: true,
|
|
1547
|
-
defaultToolProtocol: "native",
|
|
1548
1425
|
inputPrice: 0,
|
|
1549
1426
|
outputPrice: 0,
|
|
1550
1427
|
description: "SOTA coding performance with ~2500 tokens/s"
|
|
@@ -1555,8 +1432,6 @@ var cerebrasModels = {
|
|
|
1555
1432
|
contextWindow: 64e3,
|
|
1556
1433
|
supportsImages: false,
|
|
1557
1434
|
supportsPromptCache: false,
|
|
1558
|
-
supportsNativeTools: true,
|
|
1559
|
-
defaultToolProtocol: "native",
|
|
1560
1435
|
inputPrice: 0,
|
|
1561
1436
|
outputPrice: 0,
|
|
1562
1437
|
description: "OpenAI GPT OSS model with ~2800 tokens/s\n\n\u2022 64K context window\n\u2022 Excels at efficient reasoning across science, math, and coding"
|
|
@@ -1571,8 +1446,6 @@ var chutesModels = {
|
|
|
1571
1446
|
contextWindow: 163840,
|
|
1572
1447
|
supportsImages: false,
|
|
1573
1448
|
supportsPromptCache: false,
|
|
1574
|
-
supportsNativeTools: true,
|
|
1575
|
-
defaultToolProtocol: "native",
|
|
1576
1449
|
inputPrice: 0,
|
|
1577
1450
|
outputPrice: 0,
|
|
1578
1451
|
description: "DeepSeek R1 0528 model."
|
|
@@ -1582,8 +1455,6 @@ var chutesModels = {
|
|
|
1582
1455
|
contextWindow: 163840,
|
|
1583
1456
|
supportsImages: false,
|
|
1584
1457
|
supportsPromptCache: false,
|
|
1585
|
-
supportsNativeTools: true,
|
|
1586
|
-
defaultToolProtocol: "native",
|
|
1587
1458
|
inputPrice: 0,
|
|
1588
1459
|
outputPrice: 0,
|
|
1589
1460
|
description: "DeepSeek R1 model."
|
|
@@ -1593,8 +1464,6 @@ var chutesModels = {
|
|
|
1593
1464
|
contextWindow: 163840,
|
|
1594
1465
|
supportsImages: false,
|
|
1595
1466
|
supportsPromptCache: false,
|
|
1596
|
-
supportsNativeTools: true,
|
|
1597
|
-
defaultToolProtocol: "native",
|
|
1598
1467
|
inputPrice: 0,
|
|
1599
1468
|
outputPrice: 0,
|
|
1600
1469
|
description: "DeepSeek V3 model."
|
|
@@ -1604,8 +1473,6 @@ var chutesModels = {
|
|
|
1604
1473
|
contextWindow: 163840,
|
|
1605
1474
|
supportsImages: false,
|
|
1606
1475
|
supportsPromptCache: false,
|
|
1607
|
-
supportsNativeTools: true,
|
|
1608
|
-
defaultToolProtocol: "native",
|
|
1609
1476
|
inputPrice: 0,
|
|
1610
1477
|
outputPrice: 0,
|
|
1611
1478
|
description: "DeepSeek V3.1 model."
|
|
@@ -1615,8 +1482,6 @@ var chutesModels = {
|
|
|
1615
1482
|
contextWindow: 163840,
|
|
1616
1483
|
supportsImages: false,
|
|
1617
1484
|
supportsPromptCache: false,
|
|
1618
|
-
supportsNativeTools: true,
|
|
1619
|
-
defaultToolProtocol: "native",
|
|
1620
1485
|
inputPrice: 0.23,
|
|
1621
1486
|
outputPrice: 0.9,
|
|
1622
1487
|
description: "DeepSeek\u2011V3.1\u2011Terminus is an update to V3.1 that improves language consistency by reducing CN/EN mix\u2011ups and eliminating random characters, while strengthening agent capabilities with notably better Code Agent and Search Agent performance."
|
|
@@ -1626,8 +1491,6 @@ var chutesModels = {
|
|
|
1626
1491
|
contextWindow: 163840,
|
|
1627
1492
|
supportsImages: false,
|
|
1628
1493
|
supportsPromptCache: false,
|
|
1629
|
-
supportsNativeTools: true,
|
|
1630
|
-
defaultToolProtocol: "native",
|
|
1631
1494
|
inputPrice: 1,
|
|
1632
1495
|
outputPrice: 3,
|
|
1633
1496
|
description: "DeepSeek-V3.1-turbo is an FP8, speculative-decoding turbo variant optimized for ultra-fast single-shot queries (~200 TPS), with outputs close to the originals and solid function calling/reasoning/structured output, priced at $1/M input and $3/M output tokens, using 2\xD7 quota per request and not intended for bulk workloads."
|
|
@@ -1637,8 +1500,6 @@ var chutesModels = {
|
|
|
1637
1500
|
contextWindow: 163840,
|
|
1638
1501
|
supportsImages: false,
|
|
1639
1502
|
supportsPromptCache: false,
|
|
1640
|
-
supportsNativeTools: true,
|
|
1641
|
-
defaultToolProtocol: "native",
|
|
1642
1503
|
inputPrice: 0.25,
|
|
1643
1504
|
outputPrice: 0.35,
|
|
1644
1505
|
description: "DeepSeek-V3.2-Exp is an experimental LLM that introduces DeepSeek Sparse Attention to improve long\u2011context training and inference efficiency while maintaining performance comparable to V3.1\u2011Terminus."
|
|
@@ -1650,8 +1511,6 @@ var chutesModels = {
|
|
|
1650
1511
|
// From Groq
|
|
1651
1512
|
supportsImages: false,
|
|
1652
1513
|
supportsPromptCache: false,
|
|
1653
|
-
supportsNativeTools: true,
|
|
1654
|
-
defaultToolProtocol: "native",
|
|
1655
1514
|
inputPrice: 0,
|
|
1656
1515
|
outputPrice: 0,
|
|
1657
1516
|
description: "Unsloth Llama 3.3 70B Instruct model."
|
|
@@ -1661,8 +1520,6 @@ var chutesModels = {
|
|
|
1661
1520
|
contextWindow: 512e3,
|
|
1662
1521
|
supportsImages: false,
|
|
1663
1522
|
supportsPromptCache: false,
|
|
1664
|
-
supportsNativeTools: true,
|
|
1665
|
-
defaultToolProtocol: "native",
|
|
1666
1523
|
inputPrice: 0,
|
|
1667
1524
|
outputPrice: 0,
|
|
1668
1525
|
description: "ChutesAI Llama 4 Scout 17B Instruct model, 512K context."
|
|
@@ -1672,8 +1529,6 @@ var chutesModels = {
|
|
|
1672
1529
|
contextWindow: 128e3,
|
|
1673
1530
|
supportsImages: false,
|
|
1674
1531
|
supportsPromptCache: false,
|
|
1675
|
-
supportsNativeTools: true,
|
|
1676
|
-
defaultToolProtocol: "native",
|
|
1677
1532
|
inputPrice: 0,
|
|
1678
1533
|
outputPrice: 0,
|
|
1679
1534
|
description: "Unsloth Mistral Nemo Instruct model."
|
|
@@ -1683,8 +1538,6 @@ var chutesModels = {
|
|
|
1683
1538
|
contextWindow: 131072,
|
|
1684
1539
|
supportsImages: false,
|
|
1685
1540
|
supportsPromptCache: false,
|
|
1686
|
-
supportsNativeTools: true,
|
|
1687
|
-
defaultToolProtocol: "native",
|
|
1688
1541
|
inputPrice: 0,
|
|
1689
1542
|
outputPrice: 0,
|
|
1690
1543
|
description: "Unsloth Gemma 3 12B IT model."
|
|
@@ -1694,8 +1547,6 @@ var chutesModels = {
|
|
|
1694
1547
|
contextWindow: 131072,
|
|
1695
1548
|
supportsImages: false,
|
|
1696
1549
|
supportsPromptCache: false,
|
|
1697
|
-
supportsNativeTools: true,
|
|
1698
|
-
defaultToolProtocol: "native",
|
|
1699
1550
|
inputPrice: 0,
|
|
1700
1551
|
outputPrice: 0,
|
|
1701
1552
|
description: "Nous DeepHermes 3 Llama 3 8B Preview model."
|
|
@@ -1705,8 +1556,6 @@ var chutesModels = {
|
|
|
1705
1556
|
contextWindow: 131072,
|
|
1706
1557
|
supportsImages: false,
|
|
1707
1558
|
supportsPromptCache: false,
|
|
1708
|
-
supportsNativeTools: true,
|
|
1709
|
-
defaultToolProtocol: "native",
|
|
1710
1559
|
inputPrice: 0,
|
|
1711
1560
|
outputPrice: 0,
|
|
1712
1561
|
description: "Unsloth Gemma 3 4B IT model."
|
|
@@ -1716,8 +1565,6 @@ var chutesModels = {
|
|
|
1716
1565
|
contextWindow: 131072,
|
|
1717
1566
|
supportsImages: false,
|
|
1718
1567
|
supportsPromptCache: false,
|
|
1719
|
-
supportsNativeTools: true,
|
|
1720
|
-
defaultToolProtocol: "native",
|
|
1721
1568
|
inputPrice: 0,
|
|
1722
1569
|
outputPrice: 0,
|
|
1723
1570
|
description: "Nvidia Llama 3.3 Nemotron Super 49B model."
|
|
@@ -1727,8 +1574,6 @@ var chutesModels = {
|
|
|
1727
1574
|
contextWindow: 131072,
|
|
1728
1575
|
supportsImages: false,
|
|
1729
1576
|
supportsPromptCache: false,
|
|
1730
|
-
supportsNativeTools: true,
|
|
1731
|
-
defaultToolProtocol: "native",
|
|
1732
1577
|
inputPrice: 0,
|
|
1733
1578
|
outputPrice: 0,
|
|
1734
1579
|
description: "Nvidia Llama 3.1 Nemotron Ultra 253B model."
|
|
@@ -1738,8 +1583,6 @@ var chutesModels = {
|
|
|
1738
1583
|
contextWindow: 256e3,
|
|
1739
1584
|
supportsImages: false,
|
|
1740
1585
|
supportsPromptCache: false,
|
|
1741
|
-
supportsNativeTools: true,
|
|
1742
|
-
defaultToolProtocol: "native",
|
|
1743
1586
|
inputPrice: 0,
|
|
1744
1587
|
outputPrice: 0,
|
|
1745
1588
|
description: "ChutesAI Llama 4 Maverick 17B Instruct FP8 model."
|
|
@@ -1749,8 +1592,6 @@ var chutesModels = {
|
|
|
1749
1592
|
contextWindow: 163840,
|
|
1750
1593
|
supportsImages: false,
|
|
1751
1594
|
supportsPromptCache: false,
|
|
1752
|
-
supportsNativeTools: true,
|
|
1753
|
-
defaultToolProtocol: "native",
|
|
1754
1595
|
inputPrice: 0,
|
|
1755
1596
|
outputPrice: 0,
|
|
1756
1597
|
description: "DeepSeek V3 Base model."
|
|
@@ -1760,8 +1601,6 @@ var chutesModels = {
|
|
|
1760
1601
|
contextWindow: 163840,
|
|
1761
1602
|
supportsImages: false,
|
|
1762
1603
|
supportsPromptCache: false,
|
|
1763
|
-
supportsNativeTools: true,
|
|
1764
|
-
defaultToolProtocol: "native",
|
|
1765
1604
|
inputPrice: 0,
|
|
1766
1605
|
outputPrice: 0,
|
|
1767
1606
|
description: "DeepSeek R1 Zero model."
|
|
@@ -1771,8 +1610,6 @@ var chutesModels = {
|
|
|
1771
1610
|
contextWindow: 163840,
|
|
1772
1611
|
supportsImages: false,
|
|
1773
1612
|
supportsPromptCache: false,
|
|
1774
|
-
supportsNativeTools: true,
|
|
1775
|
-
defaultToolProtocol: "native",
|
|
1776
1613
|
inputPrice: 0,
|
|
1777
1614
|
outputPrice: 0,
|
|
1778
1615
|
description: "DeepSeek V3 (0324) model."
|
|
@@ -1782,8 +1619,6 @@ var chutesModels = {
|
|
|
1782
1619
|
contextWindow: 262144,
|
|
1783
1620
|
supportsImages: false,
|
|
1784
1621
|
supportsPromptCache: false,
|
|
1785
|
-
supportsNativeTools: true,
|
|
1786
|
-
defaultToolProtocol: "native",
|
|
1787
1622
|
inputPrice: 0,
|
|
1788
1623
|
outputPrice: 0,
|
|
1789
1624
|
description: "Qwen3 235B A22B Instruct 2507 model with 262K context window."
|
|
@@ -1793,8 +1628,6 @@ var chutesModels = {
|
|
|
1793
1628
|
contextWindow: 40960,
|
|
1794
1629
|
supportsImages: false,
|
|
1795
1630
|
supportsPromptCache: false,
|
|
1796
|
-
supportsNativeTools: true,
|
|
1797
|
-
defaultToolProtocol: "native",
|
|
1798
1631
|
inputPrice: 0,
|
|
1799
1632
|
outputPrice: 0,
|
|
1800
1633
|
description: "Qwen3 235B A22B model."
|
|
@@ -1804,8 +1637,6 @@ var chutesModels = {
|
|
|
1804
1637
|
contextWindow: 40960,
|
|
1805
1638
|
supportsImages: false,
|
|
1806
1639
|
supportsPromptCache: false,
|
|
1807
|
-
supportsNativeTools: true,
|
|
1808
|
-
defaultToolProtocol: "native",
|
|
1809
1640
|
inputPrice: 0,
|
|
1810
1641
|
outputPrice: 0,
|
|
1811
1642
|
description: "Qwen3 32B model."
|
|
@@ -1815,8 +1646,6 @@ var chutesModels = {
|
|
|
1815
1646
|
contextWindow: 40960,
|
|
1816
1647
|
supportsImages: false,
|
|
1817
1648
|
supportsPromptCache: false,
|
|
1818
|
-
supportsNativeTools: true,
|
|
1819
|
-
defaultToolProtocol: "native",
|
|
1820
1649
|
inputPrice: 0,
|
|
1821
1650
|
outputPrice: 0,
|
|
1822
1651
|
description: "Qwen3 30B A3B model."
|
|
@@ -1826,8 +1655,6 @@ var chutesModels = {
|
|
|
1826
1655
|
contextWindow: 40960,
|
|
1827
1656
|
supportsImages: false,
|
|
1828
1657
|
supportsPromptCache: false,
|
|
1829
|
-
supportsNativeTools: true,
|
|
1830
|
-
defaultToolProtocol: "native",
|
|
1831
1658
|
inputPrice: 0,
|
|
1832
1659
|
outputPrice: 0,
|
|
1833
1660
|
description: "Qwen3 14B model."
|
|
@@ -1837,8 +1664,6 @@ var chutesModels = {
|
|
|
1837
1664
|
contextWindow: 40960,
|
|
1838
1665
|
supportsImages: false,
|
|
1839
1666
|
supportsPromptCache: false,
|
|
1840
|
-
supportsNativeTools: true,
|
|
1841
|
-
defaultToolProtocol: "native",
|
|
1842
1667
|
inputPrice: 0,
|
|
1843
1668
|
outputPrice: 0,
|
|
1844
1669
|
description: "Qwen3 8B model."
|
|
@@ -1848,8 +1673,6 @@ var chutesModels = {
|
|
|
1848
1673
|
contextWindow: 163840,
|
|
1849
1674
|
supportsImages: false,
|
|
1850
1675
|
supportsPromptCache: false,
|
|
1851
|
-
supportsNativeTools: true,
|
|
1852
|
-
defaultToolProtocol: "native",
|
|
1853
1676
|
inputPrice: 0,
|
|
1854
1677
|
outputPrice: 0,
|
|
1855
1678
|
description: "Microsoft MAI-DS-R1 FP8 model."
|
|
@@ -1859,8 +1682,6 @@ var chutesModels = {
|
|
|
1859
1682
|
contextWindow: 163840,
|
|
1860
1683
|
supportsImages: false,
|
|
1861
1684
|
supportsPromptCache: false,
|
|
1862
|
-
supportsNativeTools: true,
|
|
1863
|
-
defaultToolProtocol: "native",
|
|
1864
1685
|
inputPrice: 0,
|
|
1865
1686
|
outputPrice: 0,
|
|
1866
1687
|
description: "TNGTech DeepSeek R1T Chimera model."
|
|
@@ -1870,8 +1691,6 @@ var chutesModels = {
|
|
|
1870
1691
|
contextWindow: 151329,
|
|
1871
1692
|
supportsImages: false,
|
|
1872
1693
|
supportsPromptCache: false,
|
|
1873
|
-
supportsNativeTools: true,
|
|
1874
|
-
defaultToolProtocol: "native",
|
|
1875
1694
|
inputPrice: 0,
|
|
1876
1695
|
outputPrice: 0,
|
|
1877
1696
|
description: "GLM-4.5-Air model with 151,329 token context window and 106B total parameters with 12B activated."
|
|
@@ -1881,8 +1700,6 @@ var chutesModels = {
|
|
|
1881
1700
|
contextWindow: 131072,
|
|
1882
1701
|
supportsImages: false,
|
|
1883
1702
|
supportsPromptCache: false,
|
|
1884
|
-
supportsNativeTools: true,
|
|
1885
|
-
defaultToolProtocol: "native",
|
|
1886
1703
|
inputPrice: 0,
|
|
1887
1704
|
outputPrice: 0,
|
|
1888
1705
|
description: "GLM-4.5-FP8 model with 128k token context window, optimized for agent-based applications with MoE architecture."
|
|
@@ -1892,8 +1709,6 @@ var chutesModels = {
|
|
|
1892
1709
|
contextWindow: 131072,
|
|
1893
1710
|
supportsImages: false,
|
|
1894
1711
|
supportsPromptCache: false,
|
|
1895
|
-
supportsNativeTools: true,
|
|
1896
|
-
defaultToolProtocol: "native",
|
|
1897
1712
|
inputPrice: 1,
|
|
1898
1713
|
outputPrice: 3,
|
|
1899
1714
|
description: "GLM-4.5-turbo model with 128K token context window, optimized for fast inference."
|
|
@@ -1903,8 +1718,6 @@ var chutesModels = {
|
|
|
1903
1718
|
contextWindow: 202752,
|
|
1904
1719
|
supportsImages: false,
|
|
1905
1720
|
supportsPromptCache: false,
|
|
1906
|
-
supportsNativeTools: true,
|
|
1907
|
-
defaultToolProtocol: "native",
|
|
1908
1721
|
inputPrice: 0,
|
|
1909
1722
|
outputPrice: 0,
|
|
1910
1723
|
description: "GLM-4.6 introduces major upgrades over GLM-4.5, including a longer 200K-token context window for complex tasks, stronger coding performance in benchmarks and real-world tools (such as Claude Code, Cline, Roo Code, and Kilo Code), improved reasoning with tool use during inference, more capable and efficient agent integration, and refined writing that better matches human style, readability, and natural role-play scenarios."
|
|
@@ -1915,8 +1728,6 @@ var chutesModels = {
|
|
|
1915
1728
|
contextWindow: 202752,
|
|
1916
1729
|
supportsImages: false,
|
|
1917
1730
|
supportsPromptCache: false,
|
|
1918
|
-
supportsNativeTools: true,
|
|
1919
|
-
defaultToolProtocol: "native",
|
|
1920
1731
|
inputPrice: 1.15,
|
|
1921
1732
|
outputPrice: 3.25,
|
|
1922
1733
|
description: "GLM-4.6-turbo model with 200K-token context window, optimized for fast inference."
|
|
@@ -1926,8 +1737,6 @@ var chutesModels = {
|
|
|
1926
1737
|
contextWindow: 128e3,
|
|
1927
1738
|
supportsImages: false,
|
|
1928
1739
|
supportsPromptCache: false,
|
|
1929
|
-
supportsNativeTools: true,
|
|
1930
|
-
defaultToolProtocol: "native",
|
|
1931
1740
|
inputPrice: 0,
|
|
1932
1741
|
outputPrice: 0,
|
|
1933
1742
|
description: "LongCat Flash Thinking FP8 model with 128K context window, optimized for complex reasoning and coding tasks."
|
|
@@ -1937,8 +1746,6 @@ var chutesModels = {
|
|
|
1937
1746
|
contextWindow: 262144,
|
|
1938
1747
|
supportsImages: false,
|
|
1939
1748
|
supportsPromptCache: false,
|
|
1940
|
-
supportsNativeTools: true,
|
|
1941
|
-
defaultToolProtocol: "native",
|
|
1942
1749
|
inputPrice: 0,
|
|
1943
1750
|
outputPrice: 0,
|
|
1944
1751
|
description: "Qwen3 Coder 480B A35B Instruct FP8 model, optimized for coding tasks."
|
|
@@ -1948,8 +1755,6 @@ var chutesModels = {
|
|
|
1948
1755
|
contextWindow: 75e3,
|
|
1949
1756
|
supportsImages: false,
|
|
1950
1757
|
supportsPromptCache: false,
|
|
1951
|
-
supportsNativeTools: true,
|
|
1952
|
-
defaultToolProtocol: "native",
|
|
1953
1758
|
inputPrice: 0.1481,
|
|
1954
1759
|
outputPrice: 0.5926,
|
|
1955
1760
|
description: "Moonshot AI Kimi K2 Instruct model with 75k context window."
|
|
@@ -1959,8 +1764,6 @@ var chutesModels = {
|
|
|
1959
1764
|
contextWindow: 262144,
|
|
1960
1765
|
supportsImages: false,
|
|
1961
1766
|
supportsPromptCache: false,
|
|
1962
|
-
supportsNativeTools: true,
|
|
1963
|
-
defaultToolProtocol: "native",
|
|
1964
1767
|
inputPrice: 0.1999,
|
|
1965
1768
|
outputPrice: 0.8001,
|
|
1966
1769
|
description: "Moonshot AI Kimi K2 Instruct 0905 model with 256k context window."
|
|
@@ -1970,8 +1773,6 @@ var chutesModels = {
|
|
|
1970
1773
|
contextWindow: 262144,
|
|
1971
1774
|
supportsImages: false,
|
|
1972
1775
|
supportsPromptCache: false,
|
|
1973
|
-
supportsNativeTools: true,
|
|
1974
|
-
defaultToolProtocol: "native",
|
|
1975
1776
|
inputPrice: 0.077968332,
|
|
1976
1777
|
outputPrice: 0.31202496,
|
|
1977
1778
|
description: "Qwen3 235B A22B Thinking 2507 model with 262K context window."
|
|
@@ -1981,8 +1782,6 @@ var chutesModels = {
|
|
|
1981
1782
|
contextWindow: 131072,
|
|
1982
1783
|
supportsImages: false,
|
|
1983
1784
|
supportsPromptCache: false,
|
|
1984
|
-
supportsNativeTools: true,
|
|
1985
|
-
defaultToolProtocol: "native",
|
|
1986
1785
|
inputPrice: 0,
|
|
1987
1786
|
outputPrice: 0,
|
|
1988
1787
|
description: "Fast, stable instruction-tuned model optimized for complex tasks, RAG, and tool use without thinking traces."
|
|
@@ -1992,8 +1791,6 @@ var chutesModels = {
|
|
|
1992
1791
|
contextWindow: 131072,
|
|
1993
1792
|
supportsImages: false,
|
|
1994
1793
|
supportsPromptCache: false,
|
|
1995
|
-
supportsNativeTools: true,
|
|
1996
|
-
defaultToolProtocol: "native",
|
|
1997
1794
|
inputPrice: 0,
|
|
1998
1795
|
outputPrice: 0,
|
|
1999
1796
|
description: "Reasoning-first model with structured thinking traces for multi-step problems, math proofs, and code synthesis."
|
|
@@ -2003,8 +1800,6 @@ var chutesModels = {
|
|
|
2003
1800
|
contextWindow: 262144,
|
|
2004
1801
|
supportsImages: true,
|
|
2005
1802
|
supportsPromptCache: false,
|
|
2006
|
-
supportsNativeTools: true,
|
|
2007
|
-
defaultToolProtocol: "native",
|
|
2008
1803
|
inputPrice: 0.16,
|
|
2009
1804
|
outputPrice: 0.65,
|
|
2010
1805
|
description: "Qwen3\u2011VL\u2011235B\u2011A22B\u2011Thinking is an open\u2011weight MoE vision\u2011language model (235B total, ~22B activated) optimized for deliberate multi\u2011step reasoning with strong text\u2011image\u2011video understanding and long\u2011context capabilities."
|
|
@@ -2012,73 +1807,6 @@ var chutesModels = {
|
|
|
2012
1807
|
};
|
|
2013
1808
|
var chutesDefaultModelInfo = chutesModels[chutesDefaultModelId];
|
|
2014
1809
|
|
|
2015
|
-
// src/providers/claude-code.ts
|
|
2016
|
-
var DATE_SUFFIX_PATTERN = /-\d{8}$/;
|
|
2017
|
-
var claudeCodeModels = {
|
|
2018
|
-
"claude-haiku-4-5": {
|
|
2019
|
-
maxTokens: 32768,
|
|
2020
|
-
contextWindow: 2e5,
|
|
2021
|
-
supportsImages: true,
|
|
2022
|
-
supportsPromptCache: true,
|
|
2023
|
-
supportsNativeTools: true,
|
|
2024
|
-
defaultToolProtocol: "native",
|
|
2025
|
-
supportsReasoningEffort: ["disable", "low", "medium", "high"],
|
|
2026
|
-
reasoningEffort: "medium",
|
|
2027
|
-
description: "Claude Haiku 4.5 - Fast and efficient with thinking"
|
|
2028
|
-
},
|
|
2029
|
-
"claude-sonnet-4-5": {
|
|
2030
|
-
maxTokens: 32768,
|
|
2031
|
-
contextWindow: 2e5,
|
|
2032
|
-
supportsImages: true,
|
|
2033
|
-
supportsPromptCache: true,
|
|
2034
|
-
supportsNativeTools: true,
|
|
2035
|
-
defaultToolProtocol: "native",
|
|
2036
|
-
supportsReasoningEffort: ["disable", "low", "medium", "high"],
|
|
2037
|
-
reasoningEffort: "medium",
|
|
2038
|
-
description: "Claude Sonnet 4.5 - Balanced performance with thinking"
|
|
2039
|
-
},
|
|
2040
|
-
"claude-opus-4-5": {
|
|
2041
|
-
maxTokens: 32768,
|
|
2042
|
-
contextWindow: 2e5,
|
|
2043
|
-
supportsImages: true,
|
|
2044
|
-
supportsPromptCache: true,
|
|
2045
|
-
supportsNativeTools: true,
|
|
2046
|
-
defaultToolProtocol: "native",
|
|
2047
|
-
supportsReasoningEffort: ["disable", "low", "medium", "high"],
|
|
2048
|
-
reasoningEffort: "medium",
|
|
2049
|
-
description: "Claude Opus 4.5 - Most capable with thinking"
|
|
2050
|
-
}
|
|
2051
|
-
};
|
|
2052
|
-
var claudeCodeDefaultModelId = "claude-sonnet-4-5";
|
|
2053
|
-
var MODEL_FAMILY_PATTERNS = [
|
|
2054
|
-
// Opus models (any version) → claude-opus-4-5
|
|
2055
|
-
{ pattern: /opus/i, target: "claude-opus-4-5" },
|
|
2056
|
-
// Haiku models (any version) → claude-haiku-4-5
|
|
2057
|
-
{ pattern: /haiku/i, target: "claude-haiku-4-5" },
|
|
2058
|
-
// Sonnet models (any version) → claude-sonnet-4-5
|
|
2059
|
-
{ pattern: /sonnet/i, target: "claude-sonnet-4-5" }
|
|
2060
|
-
];
|
|
2061
|
-
function normalizeClaudeCodeModelId(modelId) {
|
|
2062
|
-
if (Object.hasOwn(claudeCodeModels, modelId)) {
|
|
2063
|
-
return modelId;
|
|
2064
|
-
}
|
|
2065
|
-
const withoutDate = modelId.replace(DATE_SUFFIX_PATTERN, "");
|
|
2066
|
-
if (Object.hasOwn(claudeCodeModels, withoutDate)) {
|
|
2067
|
-
return withoutDate;
|
|
2068
|
-
}
|
|
2069
|
-
for (const { pattern, target } of MODEL_FAMILY_PATTERNS) {
|
|
2070
|
-
if (pattern.test(modelId)) {
|
|
2071
|
-
return target;
|
|
2072
|
-
}
|
|
2073
|
-
}
|
|
2074
|
-
return claudeCodeDefaultModelId;
|
|
2075
|
-
}
|
|
2076
|
-
var claudeCodeReasoningConfig = {
|
|
2077
|
-
low: { budgetTokens: 16e3 },
|
|
2078
|
-
medium: { budgetTokens: 32e3 },
|
|
2079
|
-
high: { budgetTokens: 64e3 }
|
|
2080
|
-
};
|
|
2081
|
-
|
|
2082
1810
|
// src/providers/deepseek.ts
|
|
2083
1811
|
var deepSeekDefaultModelId = "deepseek-chat";
|
|
2084
1812
|
var deepSeekModels = {
|
|
@@ -2088,8 +1816,6 @@ var deepSeekModels = {
|
|
|
2088
1816
|
contextWindow: 128e3,
|
|
2089
1817
|
supportsImages: false,
|
|
2090
1818
|
supportsPromptCache: true,
|
|
2091
|
-
supportsNativeTools: true,
|
|
2092
|
-
defaultToolProtocol: "native",
|
|
2093
1819
|
inputPrice: 0.28,
|
|
2094
1820
|
// $0.28 per million tokens (cache miss) - Updated Dec 9, 2025
|
|
2095
1821
|
outputPrice: 0.42,
|
|
@@ -2106,8 +1832,6 @@ var deepSeekModels = {
|
|
|
2106
1832
|
contextWindow: 128e3,
|
|
2107
1833
|
supportsImages: false,
|
|
2108
1834
|
supportsPromptCache: true,
|
|
2109
|
-
supportsNativeTools: true,
|
|
2110
|
-
defaultToolProtocol: "native",
|
|
2111
1835
|
preserveReasoning: true,
|
|
2112
1836
|
inputPrice: 0.28,
|
|
2113
1837
|
// $0.28 per million tokens (cache miss) - Updated Dec 9, 2025
|
|
@@ -2130,8 +1854,6 @@ var doubaoModels = {
|
|
|
2130
1854
|
contextWindow: 128e3,
|
|
2131
1855
|
supportsImages: true,
|
|
2132
1856
|
supportsPromptCache: true,
|
|
2133
|
-
supportsNativeTools: true,
|
|
2134
|
-
defaultToolProtocol: "native",
|
|
2135
1857
|
inputPrice: 1e-4,
|
|
2136
1858
|
// $0.0001 per million tokens (cache miss)
|
|
2137
1859
|
outputPrice: 4e-4,
|
|
@@ -2147,8 +1869,6 @@ var doubaoModels = {
|
|
|
2147
1869
|
contextWindow: 128e3,
|
|
2148
1870
|
supportsImages: true,
|
|
2149
1871
|
supportsPromptCache: true,
|
|
2150
|
-
supportsNativeTools: true,
|
|
2151
|
-
defaultToolProtocol: "native",
|
|
2152
1872
|
inputPrice: 2e-4,
|
|
2153
1873
|
// $0.0002 per million tokens
|
|
2154
1874
|
outputPrice: 8e-4,
|
|
@@ -2164,8 +1884,6 @@ var doubaoModels = {
|
|
|
2164
1884
|
contextWindow: 128e3,
|
|
2165
1885
|
supportsImages: true,
|
|
2166
1886
|
supportsPromptCache: true,
|
|
2167
|
-
supportsNativeTools: true,
|
|
2168
|
-
defaultToolProtocol: "native",
|
|
2169
1887
|
inputPrice: 15e-5,
|
|
2170
1888
|
// $0.00015 per million tokens
|
|
2171
1889
|
outputPrice: 6e-4,
|
|
@@ -2188,7 +1906,6 @@ var featherlessModels = {
|
|
|
2188
1906
|
contextWindow: 32678,
|
|
2189
1907
|
supportsImages: false,
|
|
2190
1908
|
supportsPromptCache: false,
|
|
2191
|
-
supportsNativeTools: true,
|
|
2192
1909
|
inputPrice: 0,
|
|
2193
1910
|
outputPrice: 0,
|
|
2194
1911
|
description: "DeepSeek V3 0324 model."
|
|
@@ -2198,7 +1915,6 @@ var featherlessModels = {
|
|
|
2198
1915
|
contextWindow: 32678,
|
|
2199
1916
|
supportsImages: false,
|
|
2200
1917
|
supportsPromptCache: false,
|
|
2201
|
-
supportsNativeTools: true,
|
|
2202
1918
|
inputPrice: 0,
|
|
2203
1919
|
outputPrice: 0,
|
|
2204
1920
|
description: "DeepSeek R1 0528 model."
|
|
@@ -2208,7 +1924,6 @@ var featherlessModels = {
|
|
|
2208
1924
|
contextWindow: 32678,
|
|
2209
1925
|
supportsImages: false,
|
|
2210
1926
|
supportsPromptCache: false,
|
|
2211
|
-
supportsNativeTools: true,
|
|
2212
1927
|
inputPrice: 0,
|
|
2213
1928
|
outputPrice: 0,
|
|
2214
1929
|
description: "Kimi K2 Instruct model."
|
|
@@ -2218,7 +1933,6 @@ var featherlessModels = {
|
|
|
2218
1933
|
contextWindow: 32678,
|
|
2219
1934
|
supportsImages: false,
|
|
2220
1935
|
supportsPromptCache: false,
|
|
2221
|
-
supportsNativeTools: true,
|
|
2222
1936
|
inputPrice: 0,
|
|
2223
1937
|
outputPrice: 0,
|
|
2224
1938
|
description: "GPT-OSS 120B model."
|
|
@@ -2228,7 +1942,6 @@ var featherlessModels = {
|
|
|
2228
1942
|
contextWindow: 32678,
|
|
2229
1943
|
supportsImages: false,
|
|
2230
1944
|
supportsPromptCache: false,
|
|
2231
|
-
supportsNativeTools: true,
|
|
2232
1945
|
inputPrice: 0,
|
|
2233
1946
|
outputPrice: 0,
|
|
2234
1947
|
description: "Qwen3 Coder 480B A35B Instruct model."
|
|
@@ -2244,8 +1957,6 @@ var fireworksModels = {
|
|
|
2244
1957
|
contextWindow: 262144,
|
|
2245
1958
|
supportsImages: false,
|
|
2246
1959
|
supportsPromptCache: true,
|
|
2247
|
-
supportsNativeTools: true,
|
|
2248
|
-
defaultToolProtocol: "native",
|
|
2249
1960
|
inputPrice: 0.6,
|
|
2250
1961
|
outputPrice: 2.5,
|
|
2251
1962
|
cacheReadsPrice: 0.15,
|
|
@@ -2256,8 +1967,6 @@ var fireworksModels = {
|
|
|
2256
1967
|
contextWindow: 128e3,
|
|
2257
1968
|
supportsImages: false,
|
|
2258
1969
|
supportsPromptCache: false,
|
|
2259
|
-
supportsNativeTools: true,
|
|
2260
|
-
defaultToolProtocol: "native",
|
|
2261
1970
|
inputPrice: 0.6,
|
|
2262
1971
|
outputPrice: 2.5,
|
|
2263
1972
|
description: "Kimi K2 is a state-of-the-art mixture-of-experts (MoE) language model with 32 billion activated parameters and 1 trillion total parameters. Trained with the Muon optimizer, Kimi K2 achieves exceptional performance across frontier knowledge, reasoning, and coding tasks while being meticulously optimized for agentic capabilities."
|
|
@@ -2267,7 +1976,6 @@ var fireworksModels = {
|
|
|
2267
1976
|
contextWindow: 256e3,
|
|
2268
1977
|
supportsImages: false,
|
|
2269
1978
|
supportsPromptCache: true,
|
|
2270
|
-
supportsNativeTools: true,
|
|
2271
1979
|
supportsTemperature: true,
|
|
2272
1980
|
preserveReasoning: true,
|
|
2273
1981
|
defaultTemperature: 1,
|
|
@@ -2281,8 +1989,6 @@ var fireworksModels = {
|
|
|
2281
1989
|
contextWindow: 204800,
|
|
2282
1990
|
supportsImages: false,
|
|
2283
1991
|
supportsPromptCache: false,
|
|
2284
|
-
supportsNativeTools: true,
|
|
2285
|
-
defaultToolProtocol: "native",
|
|
2286
1992
|
inputPrice: 0.3,
|
|
2287
1993
|
outputPrice: 1.2,
|
|
2288
1994
|
description: "MiniMax M2 is a high-performance language model with 204.8K context window, optimized for long-context understanding and generation tasks."
|
|
@@ -2292,8 +1998,6 @@ var fireworksModels = {
|
|
|
2292
1998
|
contextWindow: 256e3,
|
|
2293
1999
|
supportsImages: false,
|
|
2294
2000
|
supportsPromptCache: false,
|
|
2295
|
-
supportsNativeTools: true,
|
|
2296
|
-
defaultToolProtocol: "native",
|
|
2297
2001
|
inputPrice: 0.22,
|
|
2298
2002
|
outputPrice: 0.88,
|
|
2299
2003
|
description: "Latest Qwen3 thinking model, competitive against the best closed source models in Jul 2025."
|
|
@@ -2303,8 +2007,6 @@ var fireworksModels = {
|
|
|
2303
2007
|
contextWindow: 256e3,
|
|
2304
2008
|
supportsImages: false,
|
|
2305
2009
|
supportsPromptCache: false,
|
|
2306
|
-
supportsNativeTools: true,
|
|
2307
|
-
defaultToolProtocol: "native",
|
|
2308
2010
|
inputPrice: 0.45,
|
|
2309
2011
|
outputPrice: 1.8,
|
|
2310
2012
|
description: "Qwen3's most agentic code model to date."
|
|
@@ -2314,8 +2016,6 @@ var fireworksModels = {
|
|
|
2314
2016
|
contextWindow: 16e4,
|
|
2315
2017
|
supportsImages: false,
|
|
2316
2018
|
supportsPromptCache: false,
|
|
2317
|
-
supportsNativeTools: true,
|
|
2318
|
-
defaultToolProtocol: "native",
|
|
2319
2019
|
inputPrice: 3,
|
|
2320
2020
|
outputPrice: 8,
|
|
2321
2021
|
description: "05/28 updated checkpoint of Deepseek R1. Its overall performance is now approaching that of leading models, such as O3 and Gemini 2.5 Pro. Compared to the previous version, the upgraded model shows significant improvements in handling complex reasoning tasks, and this version also offers a reduced hallucination rate, enhanced support for function calling, and better experience for vibe coding. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
|
|
@@ -2325,8 +2025,6 @@ var fireworksModels = {
|
|
|
2325
2025
|
contextWindow: 128e3,
|
|
2326
2026
|
supportsImages: false,
|
|
2327
2027
|
supportsPromptCache: false,
|
|
2328
|
-
supportsNativeTools: true,
|
|
2329
|
-
defaultToolProtocol: "native",
|
|
2330
2028
|
inputPrice: 0.9,
|
|
2331
2029
|
outputPrice: 0.9,
|
|
2332
2030
|
description: "A strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token from Deepseek. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
|
|
@@ -2336,8 +2034,6 @@ var fireworksModels = {
|
|
|
2336
2034
|
contextWindow: 163840,
|
|
2337
2035
|
supportsImages: false,
|
|
2338
2036
|
supportsPromptCache: false,
|
|
2339
|
-
supportsNativeTools: true,
|
|
2340
|
-
defaultToolProtocol: "native",
|
|
2341
2037
|
inputPrice: 0.56,
|
|
2342
2038
|
outputPrice: 1.68,
|
|
2343
2039
|
description: "DeepSeek v3.1 is an improved version of the v3 model with enhanced performance, better reasoning capabilities, and improved code generation. This Mixture-of-Experts (MoE) model maintains the same 671B total parameters with 37B activated per token."
|
|
@@ -2347,8 +2043,6 @@ var fireworksModels = {
|
|
|
2347
2043
|
contextWindow: 128e3,
|
|
2348
2044
|
supportsImages: false,
|
|
2349
2045
|
supportsPromptCache: false,
|
|
2350
|
-
supportsNativeTools: true,
|
|
2351
|
-
defaultToolProtocol: "native",
|
|
2352
2046
|
inputPrice: 0.55,
|
|
2353
2047
|
outputPrice: 2.19,
|
|
2354
2048
|
description: "Z.ai GLM-4.5 with 355B total parameters and 32B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
|
|
@@ -2358,8 +2052,6 @@ var fireworksModels = {
|
|
|
2358
2052
|
contextWindow: 128e3,
|
|
2359
2053
|
supportsImages: false,
|
|
2360
2054
|
supportsPromptCache: false,
|
|
2361
|
-
supportsNativeTools: true,
|
|
2362
|
-
defaultToolProtocol: "native",
|
|
2363
2055
|
inputPrice: 0.55,
|
|
2364
2056
|
outputPrice: 2.19,
|
|
2365
2057
|
description: "Z.ai GLM-4.5-Air with 106B total parameters and 12B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
|
|
@@ -2369,8 +2061,6 @@ var fireworksModels = {
|
|
|
2369
2061
|
contextWindow: 198e3,
|
|
2370
2062
|
supportsImages: false,
|
|
2371
2063
|
supportsPromptCache: false,
|
|
2372
|
-
supportsNativeTools: true,
|
|
2373
|
-
defaultToolProtocol: "native",
|
|
2374
2064
|
inputPrice: 0.55,
|
|
2375
2065
|
outputPrice: 2.19,
|
|
2376
2066
|
description: "Z.ai GLM-4.6 is an advanced coding model with exceptional performance on complex programming tasks. Features improved reasoning capabilities and enhanced code generation quality, making it ideal for software development workflows."
|
|
@@ -2380,8 +2070,6 @@ var fireworksModels = {
|
|
|
2380
2070
|
contextWindow: 128e3,
|
|
2381
2071
|
supportsImages: false,
|
|
2382
2072
|
supportsPromptCache: false,
|
|
2383
|
-
supportsNativeTools: true,
|
|
2384
|
-
defaultToolProtocol: "native",
|
|
2385
2073
|
inputPrice: 0.07,
|
|
2386
2074
|
outputPrice: 0.3,
|
|
2387
2075
|
description: "OpenAI gpt-oss-20b: Compact model for local/edge deployments. Optimized for low-latency and resource-constrained environments with chain-of-thought output, adjustable reasoning, and agentic workflows."
|
|
@@ -2391,11 +2079,63 @@ var fireworksModels = {
|
|
|
2391
2079
|
contextWindow: 128e3,
|
|
2392
2080
|
supportsImages: false,
|
|
2393
2081
|
supportsPromptCache: false,
|
|
2394
|
-
supportsNativeTools: true,
|
|
2395
|
-
defaultToolProtocol: "native",
|
|
2396
2082
|
inputPrice: 0.15,
|
|
2397
2083
|
outputPrice: 0.6,
|
|
2398
2084
|
description: "OpenAI gpt-oss-120b: Production-grade, general-purpose model that fits on a single H100 GPU. Features complex reasoning, configurable effort, full chain-of-thought transparency, and supports function calling, tool use, and structured outputs."
|
|
2085
|
+
},
|
|
2086
|
+
"accounts/fireworks/models/minimax-m2p1": {
|
|
2087
|
+
maxTokens: 4096,
|
|
2088
|
+
contextWindow: 204800,
|
|
2089
|
+
supportsImages: false,
|
|
2090
|
+
supportsPromptCache: false,
|
|
2091
|
+
inputPrice: 0.3,
|
|
2092
|
+
outputPrice: 1.2,
|
|
2093
|
+
description: "MiniMax M2.1 is an upgraded version of M2 with improved performance on complex reasoning, coding, and long-context understanding tasks."
|
|
2094
|
+
},
|
|
2095
|
+
"accounts/fireworks/models/deepseek-v3p2": {
|
|
2096
|
+
maxTokens: 16384,
|
|
2097
|
+
contextWindow: 163840,
|
|
2098
|
+
supportsImages: false,
|
|
2099
|
+
supportsPromptCache: false,
|
|
2100
|
+
inputPrice: 0.56,
|
|
2101
|
+
outputPrice: 1.68,
|
|
2102
|
+
description: "DeepSeek V3.2 is the latest iteration of the V3 model family with enhanced reasoning capabilities, improved code generation, and better instruction following."
|
|
2103
|
+
},
|
|
2104
|
+
"accounts/fireworks/models/glm-4p7": {
|
|
2105
|
+
maxTokens: 25344,
|
|
2106
|
+
contextWindow: 198e3,
|
|
2107
|
+
supportsImages: false,
|
|
2108
|
+
supportsPromptCache: false,
|
|
2109
|
+
inputPrice: 0.55,
|
|
2110
|
+
outputPrice: 2.19,
|
|
2111
|
+
description: "Z.ai GLM-4.7 is the latest coding model with exceptional performance on complex programming tasks. Features improved reasoning capabilities and enhanced code generation quality."
|
|
2112
|
+
},
|
|
2113
|
+
"accounts/fireworks/models/llama-v3p3-70b-instruct": {
|
|
2114
|
+
maxTokens: 16384,
|
|
2115
|
+
contextWindow: 131072,
|
|
2116
|
+
supportsImages: false,
|
|
2117
|
+
supportsPromptCache: false,
|
|
2118
|
+
inputPrice: 0.9,
|
|
2119
|
+
outputPrice: 0.9,
|
|
2120
|
+
description: "Meta Llama 3.3 70B Instruct is a highly capable instruction-tuned model with strong reasoning, coding, and general task performance."
|
|
2121
|
+
},
|
|
2122
|
+
"accounts/fireworks/models/llama4-maverick-instruct-basic": {
|
|
2123
|
+
maxTokens: 16384,
|
|
2124
|
+
contextWindow: 131072,
|
|
2125
|
+
supportsImages: true,
|
|
2126
|
+
supportsPromptCache: false,
|
|
2127
|
+
inputPrice: 0.22,
|
|
2128
|
+
outputPrice: 0.88,
|
|
2129
|
+
description: "Llama 4 Maverick is Meta's latest multimodal model with vision capabilities, optimized for instruction following and coding tasks."
|
|
2130
|
+
},
|
|
2131
|
+
"accounts/fireworks/models/llama4-scout-instruct-basic": {
|
|
2132
|
+
maxTokens: 16384,
|
|
2133
|
+
contextWindow: 131072,
|
|
2134
|
+
supportsImages: true,
|
|
2135
|
+
supportsPromptCache: false,
|
|
2136
|
+
inputPrice: 0.15,
|
|
2137
|
+
outputPrice: 0.6,
|
|
2138
|
+
description: "Llama 4 Scout is a smaller, faster variant of Llama 4 with multimodal capabilities, ideal for quick iterations and cost-effective deployments."
|
|
2399
2139
|
}
|
|
2400
2140
|
};
|
|
2401
2141
|
|
|
@@ -2406,8 +2146,6 @@ var geminiModels = {
|
|
|
2406
2146
|
maxTokens: 65536,
|
|
2407
2147
|
contextWindow: 1048576,
|
|
2408
2148
|
supportsImages: true,
|
|
2409
|
-
supportsNativeTools: true,
|
|
2410
|
-
defaultToolProtocol: "native",
|
|
2411
2149
|
supportsPromptCache: true,
|
|
2412
2150
|
supportsReasoningEffort: ["low", "high"],
|
|
2413
2151
|
reasoningEffort: "low",
|
|
@@ -2415,16 +2153,19 @@ var geminiModels = {
|
|
|
2415
2153
|
defaultTemperature: 1,
|
|
2416
2154
|
inputPrice: 4,
|
|
2417
2155
|
outputPrice: 18,
|
|
2156
|
+
cacheReadsPrice: 0.4,
|
|
2418
2157
|
tiers: [
|
|
2419
2158
|
{
|
|
2420
2159
|
contextWindow: 2e5,
|
|
2421
2160
|
inputPrice: 2,
|
|
2422
|
-
outputPrice: 12
|
|
2161
|
+
outputPrice: 12,
|
|
2162
|
+
cacheReadsPrice: 0.2
|
|
2423
2163
|
},
|
|
2424
2164
|
{
|
|
2425
2165
|
contextWindow: Infinity,
|
|
2426
2166
|
inputPrice: 4,
|
|
2427
|
-
outputPrice: 18
|
|
2167
|
+
outputPrice: 18,
|
|
2168
|
+
cacheReadsPrice: 0.4
|
|
2428
2169
|
}
|
|
2429
2170
|
]
|
|
2430
2171
|
},
|
|
@@ -2432,25 +2173,20 @@ var geminiModels = {
|
|
|
2432
2173
|
maxTokens: 65536,
|
|
2433
2174
|
contextWindow: 1048576,
|
|
2434
2175
|
supportsImages: true,
|
|
2435
|
-
supportsNativeTools: true,
|
|
2436
|
-
defaultToolProtocol: "native",
|
|
2437
2176
|
supportsPromptCache: true,
|
|
2438
2177
|
supportsReasoningEffort: ["minimal", "low", "medium", "high"],
|
|
2439
2178
|
reasoningEffort: "medium",
|
|
2440
2179
|
supportsTemperature: true,
|
|
2441
2180
|
defaultTemperature: 1,
|
|
2442
|
-
inputPrice: 0.
|
|
2443
|
-
outputPrice:
|
|
2444
|
-
cacheReadsPrice: 0.
|
|
2445
|
-
cacheWritesPrice: 1
|
|
2181
|
+
inputPrice: 0.5,
|
|
2182
|
+
outputPrice: 3,
|
|
2183
|
+
cacheReadsPrice: 0.05
|
|
2446
2184
|
},
|
|
2447
2185
|
// 2.5 Pro models
|
|
2448
2186
|
"gemini-2.5-pro": {
|
|
2449
2187
|
maxTokens: 64e3,
|
|
2450
2188
|
contextWindow: 1048576,
|
|
2451
2189
|
supportsImages: true,
|
|
2452
|
-
supportsNativeTools: true,
|
|
2453
|
-
defaultToolProtocol: "native",
|
|
2454
2190
|
supportsPromptCache: true,
|
|
2455
2191
|
inputPrice: 2.5,
|
|
2456
2192
|
// This is the pricing for prompts above 200k tokens.
|
|
@@ -2479,8 +2215,6 @@ var geminiModels = {
|
|
|
2479
2215
|
maxTokens: 65535,
|
|
2480
2216
|
contextWindow: 1048576,
|
|
2481
2217
|
supportsImages: true,
|
|
2482
|
-
supportsNativeTools: true,
|
|
2483
|
-
defaultToolProtocol: "native",
|
|
2484
2218
|
supportsPromptCache: true,
|
|
2485
2219
|
inputPrice: 2.5,
|
|
2486
2220
|
// This is the pricing for prompts above 200k tokens.
|
|
@@ -2508,8 +2242,6 @@ var geminiModels = {
|
|
|
2508
2242
|
maxTokens: 65535,
|
|
2509
2243
|
contextWindow: 1048576,
|
|
2510
2244
|
supportsImages: true,
|
|
2511
|
-
supportsNativeTools: true,
|
|
2512
|
-
defaultToolProtocol: "native",
|
|
2513
2245
|
supportsPromptCache: true,
|
|
2514
2246
|
inputPrice: 2.5,
|
|
2515
2247
|
// This is the pricing for prompts above 200k tokens.
|
|
@@ -2535,8 +2267,6 @@ var geminiModels = {
|
|
|
2535
2267
|
maxTokens: 65535,
|
|
2536
2268
|
contextWindow: 1048576,
|
|
2537
2269
|
supportsImages: true,
|
|
2538
|
-
supportsNativeTools: true,
|
|
2539
|
-
defaultToolProtocol: "native",
|
|
2540
2270
|
supportsPromptCache: true,
|
|
2541
2271
|
inputPrice: 2.5,
|
|
2542
2272
|
// This is the pricing for prompts above 200k tokens.
|
|
@@ -2565,8 +2295,6 @@ var geminiModels = {
|
|
|
2565
2295
|
maxTokens: 65536,
|
|
2566
2296
|
contextWindow: 1048576,
|
|
2567
2297
|
supportsImages: true,
|
|
2568
|
-
supportsNativeTools: true,
|
|
2569
|
-
defaultToolProtocol: "native",
|
|
2570
2298
|
supportsPromptCache: true,
|
|
2571
2299
|
inputPrice: 0.3,
|
|
2572
2300
|
outputPrice: 2.5,
|
|
@@ -2579,8 +2307,6 @@ var geminiModels = {
|
|
|
2579
2307
|
maxTokens: 65536,
|
|
2580
2308
|
contextWindow: 1048576,
|
|
2581
2309
|
supportsImages: true,
|
|
2582
|
-
supportsNativeTools: true,
|
|
2583
|
-
defaultToolProtocol: "native",
|
|
2584
2310
|
supportsPromptCache: true,
|
|
2585
2311
|
inputPrice: 0.3,
|
|
2586
2312
|
outputPrice: 2.5,
|
|
@@ -2593,8 +2319,6 @@ var geminiModels = {
|
|
|
2593
2319
|
maxTokens: 64e3,
|
|
2594
2320
|
contextWindow: 1048576,
|
|
2595
2321
|
supportsImages: true,
|
|
2596
|
-
supportsNativeTools: true,
|
|
2597
|
-
defaultToolProtocol: "native",
|
|
2598
2322
|
supportsPromptCache: true,
|
|
2599
2323
|
inputPrice: 0.3,
|
|
2600
2324
|
outputPrice: 2.5,
|
|
@@ -2608,8 +2332,6 @@ var geminiModels = {
|
|
|
2608
2332
|
maxTokens: 65536,
|
|
2609
2333
|
contextWindow: 1048576,
|
|
2610
2334
|
supportsImages: true,
|
|
2611
|
-
supportsNativeTools: true,
|
|
2612
|
-
defaultToolProtocol: "native",
|
|
2613
2335
|
supportsPromptCache: true,
|
|
2614
2336
|
inputPrice: 0.1,
|
|
2615
2337
|
outputPrice: 0.4,
|
|
@@ -2622,8 +2344,6 @@ var geminiModels = {
|
|
|
2622
2344
|
maxTokens: 65536,
|
|
2623
2345
|
contextWindow: 1048576,
|
|
2624
2346
|
supportsImages: true,
|
|
2625
|
-
supportsNativeTools: true,
|
|
2626
|
-
defaultToolProtocol: "native",
|
|
2627
2347
|
supportsPromptCache: true,
|
|
2628
2348
|
inputPrice: 0.1,
|
|
2629
2349
|
outputPrice: 0.4,
|
|
@@ -2643,8 +2363,6 @@ var groqModels = {
|
|
|
2643
2363
|
contextWindow: 131072,
|
|
2644
2364
|
supportsImages: false,
|
|
2645
2365
|
supportsPromptCache: false,
|
|
2646
|
-
supportsNativeTools: true,
|
|
2647
|
-
defaultToolProtocol: "native",
|
|
2648
2366
|
inputPrice: 0.05,
|
|
2649
2367
|
outputPrice: 0.08,
|
|
2650
2368
|
description: "Meta Llama 3.1 8B Instant model, 128K context."
|
|
@@ -2654,8 +2372,6 @@ var groqModels = {
|
|
|
2654
2372
|
contextWindow: 131072,
|
|
2655
2373
|
supportsImages: false,
|
|
2656
2374
|
supportsPromptCache: false,
|
|
2657
|
-
supportsNativeTools: true,
|
|
2658
|
-
defaultToolProtocol: "native",
|
|
2659
2375
|
inputPrice: 0.59,
|
|
2660
2376
|
outputPrice: 0.79,
|
|
2661
2377
|
description: "Meta Llama 3.3 70B Versatile model, 128K context."
|
|
@@ -2665,8 +2381,6 @@ var groqModels = {
|
|
|
2665
2381
|
contextWindow: 131072,
|
|
2666
2382
|
supportsImages: false,
|
|
2667
2383
|
supportsPromptCache: false,
|
|
2668
|
-
supportsNativeTools: true,
|
|
2669
|
-
defaultToolProtocol: "native",
|
|
2670
2384
|
inputPrice: 0.11,
|
|
2671
2385
|
outputPrice: 0.34,
|
|
2672
2386
|
description: "Meta Llama 4 Scout 17B Instruct model, 128K context."
|
|
@@ -2676,8 +2390,6 @@ var groqModels = {
|
|
|
2676
2390
|
contextWindow: 131072,
|
|
2677
2391
|
supportsImages: false,
|
|
2678
2392
|
supportsPromptCache: false,
|
|
2679
|
-
supportsNativeTools: true,
|
|
2680
|
-
defaultToolProtocol: "native",
|
|
2681
2393
|
inputPrice: 0.29,
|
|
2682
2394
|
outputPrice: 0.59,
|
|
2683
2395
|
description: "Alibaba Qwen 3 32B model, 128K context."
|
|
@@ -2687,8 +2399,6 @@ var groqModels = {
|
|
|
2687
2399
|
contextWindow: 262144,
|
|
2688
2400
|
supportsImages: false,
|
|
2689
2401
|
supportsPromptCache: true,
|
|
2690
|
-
supportsNativeTools: true,
|
|
2691
|
-
defaultToolProtocol: "native",
|
|
2692
2402
|
inputPrice: 0.6,
|
|
2693
2403
|
outputPrice: 2.5,
|
|
2694
2404
|
cacheReadsPrice: 0.15,
|
|
@@ -2699,8 +2409,6 @@ var groqModels = {
|
|
|
2699
2409
|
contextWindow: 131072,
|
|
2700
2410
|
supportsImages: false,
|
|
2701
2411
|
supportsPromptCache: false,
|
|
2702
|
-
supportsNativeTools: true,
|
|
2703
|
-
defaultToolProtocol: "native",
|
|
2704
2412
|
inputPrice: 0.15,
|
|
2705
2413
|
outputPrice: 0.75,
|
|
2706
2414
|
description: "GPT-OSS 120B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 128 experts."
|
|
@@ -2710,8 +2418,6 @@ var groqModels = {
|
|
|
2710
2418
|
contextWindow: 131072,
|
|
2711
2419
|
supportsImages: false,
|
|
2712
2420
|
supportsPromptCache: false,
|
|
2713
|
-
supportsNativeTools: true,
|
|
2714
|
-
defaultToolProtocol: "native",
|
|
2715
2421
|
inputPrice: 0.1,
|
|
2716
2422
|
outputPrice: 0.5,
|
|
2717
2423
|
description: "GPT-OSS 20B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 32 experts."
|
|
@@ -2738,7 +2444,6 @@ var ioIntelligenceModels = {
|
|
|
2738
2444
|
contextWindow: 128e3,
|
|
2739
2445
|
supportsImages: false,
|
|
2740
2446
|
supportsPromptCache: false,
|
|
2741
|
-
supportsNativeTools: true,
|
|
2742
2447
|
description: "DeepSeek R1 reasoning model"
|
|
2743
2448
|
},
|
|
2744
2449
|
"meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8": {
|
|
@@ -2746,7 +2451,6 @@ var ioIntelligenceModels = {
|
|
|
2746
2451
|
contextWindow: 43e4,
|
|
2747
2452
|
supportsImages: true,
|
|
2748
2453
|
supportsPromptCache: false,
|
|
2749
|
-
supportsNativeTools: true,
|
|
2750
2454
|
description: "Llama 4 Maverick 17B model"
|
|
2751
2455
|
},
|
|
2752
2456
|
"Intel/Qwen3-Coder-480B-A35B-Instruct-int4-mixed-ar": {
|
|
@@ -2754,7 +2458,6 @@ var ioIntelligenceModels = {
|
|
|
2754
2458
|
contextWindow: 106e3,
|
|
2755
2459
|
supportsImages: false,
|
|
2756
2460
|
supportsPromptCache: false,
|
|
2757
|
-
supportsNativeTools: true,
|
|
2758
2461
|
description: "Qwen3 Coder 480B specialized for coding"
|
|
2759
2462
|
},
|
|
2760
2463
|
"openai/gpt-oss-120b": {
|
|
@@ -2762,7 +2465,6 @@ var ioIntelligenceModels = {
|
|
|
2762
2465
|
contextWindow: 131072,
|
|
2763
2466
|
supportsImages: false,
|
|
2764
2467
|
supportsPromptCache: false,
|
|
2765
|
-
supportsNativeTools: true,
|
|
2766
2468
|
description: "OpenAI GPT-OSS 120B model"
|
|
2767
2469
|
}
|
|
2768
2470
|
};
|
|
@@ -2774,8 +2476,6 @@ var litellmDefaultModelInfo = {
|
|
|
2774
2476
|
contextWindow: 2e5,
|
|
2775
2477
|
supportsImages: true,
|
|
2776
2478
|
supportsPromptCache: true,
|
|
2777
|
-
supportsNativeTools: true,
|
|
2778
|
-
defaultToolProtocol: "native",
|
|
2779
2479
|
inputPrice: 3,
|
|
2780
2480
|
outputPrice: 15,
|
|
2781
2481
|
cacheWritesPrice: 3.75,
|
|
@@ -2790,8 +2490,6 @@ var lMStudioDefaultModelInfo = {
|
|
|
2790
2490
|
contextWindow: 2e5,
|
|
2791
2491
|
supportsImages: true,
|
|
2792
2492
|
supportsPromptCache: true,
|
|
2793
|
-
supportsNativeTools: true,
|
|
2794
|
-
defaultToolProtocol: "native",
|
|
2795
2493
|
inputPrice: 0,
|
|
2796
2494
|
outputPrice: 0,
|
|
2797
2495
|
cacheWritesPrice: 0,
|
|
@@ -2807,8 +2505,6 @@ var mistralModels = {
|
|
|
2807
2505
|
contextWindow: 128e3,
|
|
2808
2506
|
supportsImages: true,
|
|
2809
2507
|
supportsPromptCache: false,
|
|
2810
|
-
supportsNativeTools: true,
|
|
2811
|
-
defaultToolProtocol: "native",
|
|
2812
2508
|
inputPrice: 2,
|
|
2813
2509
|
outputPrice: 5
|
|
2814
2510
|
},
|
|
@@ -2817,8 +2513,6 @@ var mistralModels = {
|
|
|
2817
2513
|
contextWindow: 131e3,
|
|
2818
2514
|
supportsImages: true,
|
|
2819
2515
|
supportsPromptCache: false,
|
|
2820
|
-
supportsNativeTools: true,
|
|
2821
|
-
defaultToolProtocol: "native",
|
|
2822
2516
|
inputPrice: 0.4,
|
|
2823
2517
|
outputPrice: 2
|
|
2824
2518
|
},
|
|
@@ -2827,8 +2521,6 @@ var mistralModels = {
|
|
|
2827
2521
|
contextWindow: 131e3,
|
|
2828
2522
|
supportsImages: true,
|
|
2829
2523
|
supportsPromptCache: false,
|
|
2830
|
-
supportsNativeTools: true,
|
|
2831
|
-
defaultToolProtocol: "native",
|
|
2832
2524
|
inputPrice: 0.4,
|
|
2833
2525
|
outputPrice: 2
|
|
2834
2526
|
},
|
|
@@ -2837,8 +2529,6 @@ var mistralModels = {
|
|
|
2837
2529
|
contextWindow: 256e3,
|
|
2838
2530
|
supportsImages: false,
|
|
2839
2531
|
supportsPromptCache: false,
|
|
2840
|
-
supportsNativeTools: true,
|
|
2841
|
-
defaultToolProtocol: "native",
|
|
2842
2532
|
inputPrice: 0.3,
|
|
2843
2533
|
outputPrice: 0.9
|
|
2844
2534
|
},
|
|
@@ -2847,8 +2537,6 @@ var mistralModels = {
|
|
|
2847
2537
|
contextWindow: 131e3,
|
|
2848
2538
|
supportsImages: false,
|
|
2849
2539
|
supportsPromptCache: false,
|
|
2850
|
-
supportsNativeTools: true,
|
|
2851
|
-
defaultToolProtocol: "native",
|
|
2852
2540
|
inputPrice: 2,
|
|
2853
2541
|
outputPrice: 6
|
|
2854
2542
|
},
|
|
@@ -2857,8 +2545,6 @@ var mistralModels = {
|
|
|
2857
2545
|
contextWindow: 131e3,
|
|
2858
2546
|
supportsImages: false,
|
|
2859
2547
|
supportsPromptCache: false,
|
|
2860
|
-
supportsNativeTools: true,
|
|
2861
|
-
defaultToolProtocol: "native",
|
|
2862
2548
|
inputPrice: 0.1,
|
|
2863
2549
|
outputPrice: 0.1
|
|
2864
2550
|
},
|
|
@@ -2867,8 +2553,6 @@ var mistralModels = {
|
|
|
2867
2553
|
contextWindow: 131e3,
|
|
2868
2554
|
supportsImages: false,
|
|
2869
2555
|
supportsPromptCache: false,
|
|
2870
|
-
supportsNativeTools: true,
|
|
2871
|
-
defaultToolProtocol: "native",
|
|
2872
2556
|
inputPrice: 0.04,
|
|
2873
2557
|
outputPrice: 0.04
|
|
2874
2558
|
},
|
|
@@ -2877,8 +2561,6 @@ var mistralModels = {
|
|
|
2877
2561
|
contextWindow: 32e3,
|
|
2878
2562
|
supportsImages: false,
|
|
2879
2563
|
supportsPromptCache: false,
|
|
2880
|
-
supportsNativeTools: true,
|
|
2881
|
-
defaultToolProtocol: "native",
|
|
2882
2564
|
inputPrice: 0.2,
|
|
2883
2565
|
outputPrice: 0.6
|
|
2884
2566
|
},
|
|
@@ -2887,8 +2569,6 @@ var mistralModels = {
|
|
|
2887
2569
|
contextWindow: 131e3,
|
|
2888
2570
|
supportsImages: true,
|
|
2889
2571
|
supportsPromptCache: false,
|
|
2890
|
-
supportsNativeTools: true,
|
|
2891
|
-
defaultToolProtocol: "native",
|
|
2892
2572
|
inputPrice: 2,
|
|
2893
2573
|
outputPrice: 6
|
|
2894
2574
|
}
|
|
@@ -2903,8 +2583,6 @@ var moonshotModels = {
|
|
|
2903
2583
|
contextWindow: 131072,
|
|
2904
2584
|
supportsImages: false,
|
|
2905
2585
|
supportsPromptCache: true,
|
|
2906
|
-
supportsNativeTools: true,
|
|
2907
|
-
defaultToolProtocol: "native",
|
|
2908
2586
|
inputPrice: 0.6,
|
|
2909
2587
|
// $0.60 per million tokens (cache miss)
|
|
2910
2588
|
outputPrice: 2.5,
|
|
@@ -2920,8 +2598,6 @@ var moonshotModels = {
|
|
|
2920
2598
|
contextWindow: 262144,
|
|
2921
2599
|
supportsImages: false,
|
|
2922
2600
|
supportsPromptCache: true,
|
|
2923
|
-
supportsNativeTools: true,
|
|
2924
|
-
defaultToolProtocol: "native",
|
|
2925
2601
|
inputPrice: 0.6,
|
|
2926
2602
|
outputPrice: 2.5,
|
|
2927
2603
|
cacheReadsPrice: 0.15,
|
|
@@ -2932,8 +2608,6 @@ var moonshotModels = {
|
|
|
2932
2608
|
contextWindow: 262144,
|
|
2933
2609
|
supportsImages: false,
|
|
2934
2610
|
supportsPromptCache: true,
|
|
2935
|
-
supportsNativeTools: true,
|
|
2936
|
-
defaultToolProtocol: "native",
|
|
2937
2611
|
inputPrice: 2.4,
|
|
2938
2612
|
// $2.40 per million tokens (cache miss)
|
|
2939
2613
|
outputPrice: 10,
|
|
@@ -2952,8 +2626,6 @@ var moonshotModels = {
|
|
|
2952
2626
|
supportsImages: false,
|
|
2953
2627
|
// Text-only (no image/vision support)
|
|
2954
2628
|
supportsPromptCache: true,
|
|
2955
|
-
supportsNativeTools: true,
|
|
2956
|
-
defaultToolProtocol: "native",
|
|
2957
2629
|
inputPrice: 0.6,
|
|
2958
2630
|
// $0.60 per million tokens (cache miss)
|
|
2959
2631
|
outputPrice: 2.5,
|
|
@@ -2967,6 +2639,21 @@ var moonshotModels = {
|
|
|
2967
2639
|
preserveReasoning: true,
|
|
2968
2640
|
defaultTemperature: 1,
|
|
2969
2641
|
description: `The kimi-k2-thinking model is a general-purpose agentic reasoning model developed by Moonshot AI. Thanks to its strength in deep reasoning and multi-turn tool use, it can solve even the hardest problems.`
|
|
2642
|
+
},
|
|
2643
|
+
"kimi-k2.5": {
|
|
2644
|
+
maxTokens: 16384,
|
|
2645
|
+
contextWindow: 262144,
|
|
2646
|
+
supportsImages: false,
|
|
2647
|
+
supportsPromptCache: true,
|
|
2648
|
+
inputPrice: 0.6,
|
|
2649
|
+
// $0.60 per million tokens (cache miss)
|
|
2650
|
+
outputPrice: 3,
|
|
2651
|
+
// $3.00 per million tokens
|
|
2652
|
+
cacheReadsPrice: 0.1,
|
|
2653
|
+
// $0.10 per million tokens (cache hit)
|
|
2654
|
+
supportsTemperature: true,
|
|
2655
|
+
defaultTemperature: 1,
|
|
2656
|
+
description: "Kimi K2.5 is the latest generation of Moonshot AI's Kimi series, featuring improved reasoning capabilities and enhanced performance across diverse tasks."
|
|
2970
2657
|
}
|
|
2971
2658
|
};
|
|
2972
2659
|
var MOONSHOT_DEFAULT_TEMPERATURE = 0.6;
|
|
@@ -2978,7 +2665,6 @@ var ollamaDefaultModelInfo = {
|
|
|
2978
2665
|
contextWindow: 2e5,
|
|
2979
2666
|
supportsImages: true,
|
|
2980
2667
|
supportsPromptCache: true,
|
|
2981
|
-
supportsNativeTools: true,
|
|
2982
2668
|
inputPrice: 0,
|
|
2983
2669
|
outputPrice: 0,
|
|
2984
2670
|
cacheWritesPrice: 0,
|
|
@@ -2992,8 +2678,6 @@ var openAiNativeModels = {
|
|
|
2992
2678
|
"gpt-5.1-codex-max": {
|
|
2993
2679
|
maxTokens: 128e3,
|
|
2994
2680
|
contextWindow: 4e5,
|
|
2995
|
-
supportsNativeTools: true,
|
|
2996
|
-
defaultToolProtocol: "native",
|
|
2997
2681
|
includedTools: ["apply_patch"],
|
|
2998
2682
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
2999
2683
|
supportsImages: true,
|
|
@@ -3011,8 +2695,6 @@ var openAiNativeModels = {
|
|
|
3011
2695
|
"gpt-5.2": {
|
|
3012
2696
|
maxTokens: 128e3,
|
|
3013
2697
|
contextWindow: 4e5,
|
|
3014
|
-
supportsNativeTools: true,
|
|
3015
|
-
defaultToolProtocol: "native",
|
|
3016
2698
|
includedTools: ["apply_patch"],
|
|
3017
2699
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3018
2700
|
supportsImages: true,
|
|
@@ -3034,8 +2716,6 @@ var openAiNativeModels = {
|
|
|
3034
2716
|
"gpt-5.2-codex": {
|
|
3035
2717
|
maxTokens: 128e3,
|
|
3036
2718
|
contextWindow: 4e5,
|
|
3037
|
-
supportsNativeTools: true,
|
|
3038
|
-
defaultToolProtocol: "native",
|
|
3039
2719
|
includedTools: ["apply_patch"],
|
|
3040
2720
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3041
2721
|
supportsImages: true,
|
|
@@ -3053,8 +2733,6 @@ var openAiNativeModels = {
|
|
|
3053
2733
|
"gpt-5.2-chat-latest": {
|
|
3054
2734
|
maxTokens: 16384,
|
|
3055
2735
|
contextWindow: 128e3,
|
|
3056
|
-
supportsNativeTools: true,
|
|
3057
|
-
defaultToolProtocol: "native",
|
|
3058
2736
|
includedTools: ["apply_patch"],
|
|
3059
2737
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3060
2738
|
supportsImages: true,
|
|
@@ -3067,8 +2745,6 @@ var openAiNativeModels = {
|
|
|
3067
2745
|
"gpt-5.1": {
|
|
3068
2746
|
maxTokens: 128e3,
|
|
3069
2747
|
contextWindow: 4e5,
|
|
3070
|
-
supportsNativeTools: true,
|
|
3071
|
-
defaultToolProtocol: "native",
|
|
3072
2748
|
includedTools: ["apply_patch"],
|
|
3073
2749
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3074
2750
|
supportsImages: true,
|
|
@@ -3090,8 +2766,6 @@ var openAiNativeModels = {
|
|
|
3090
2766
|
"gpt-5.1-codex": {
|
|
3091
2767
|
maxTokens: 128e3,
|
|
3092
2768
|
contextWindow: 4e5,
|
|
3093
|
-
supportsNativeTools: true,
|
|
3094
|
-
defaultToolProtocol: "native",
|
|
3095
2769
|
includedTools: ["apply_patch"],
|
|
3096
2770
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3097
2771
|
supportsImages: true,
|
|
@@ -3109,8 +2783,6 @@ var openAiNativeModels = {
|
|
|
3109
2783
|
"gpt-5.1-codex-mini": {
|
|
3110
2784
|
maxTokens: 128e3,
|
|
3111
2785
|
contextWindow: 4e5,
|
|
3112
|
-
supportsNativeTools: true,
|
|
3113
|
-
defaultToolProtocol: "native",
|
|
3114
2786
|
includedTools: ["apply_patch"],
|
|
3115
2787
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3116
2788
|
supportsImages: true,
|
|
@@ -3127,8 +2799,6 @@ var openAiNativeModels = {
|
|
|
3127
2799
|
"gpt-5": {
|
|
3128
2800
|
maxTokens: 128e3,
|
|
3129
2801
|
contextWindow: 4e5,
|
|
3130
|
-
supportsNativeTools: true,
|
|
3131
|
-
defaultToolProtocol: "native",
|
|
3132
2802
|
includedTools: ["apply_patch"],
|
|
3133
2803
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3134
2804
|
supportsImages: true,
|
|
@@ -3149,8 +2819,6 @@ var openAiNativeModels = {
|
|
|
3149
2819
|
"gpt-5-mini": {
|
|
3150
2820
|
maxTokens: 128e3,
|
|
3151
2821
|
contextWindow: 4e5,
|
|
3152
|
-
supportsNativeTools: true,
|
|
3153
|
-
defaultToolProtocol: "native",
|
|
3154
2822
|
includedTools: ["apply_patch"],
|
|
3155
2823
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3156
2824
|
supportsImages: true,
|
|
@@ -3171,8 +2839,6 @@ var openAiNativeModels = {
|
|
|
3171
2839
|
"gpt-5-codex": {
|
|
3172
2840
|
maxTokens: 128e3,
|
|
3173
2841
|
contextWindow: 4e5,
|
|
3174
|
-
supportsNativeTools: true,
|
|
3175
|
-
defaultToolProtocol: "native",
|
|
3176
2842
|
includedTools: ["apply_patch"],
|
|
3177
2843
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3178
2844
|
supportsImages: true,
|
|
@@ -3189,8 +2855,6 @@ var openAiNativeModels = {
|
|
|
3189
2855
|
"gpt-5-nano": {
|
|
3190
2856
|
maxTokens: 128e3,
|
|
3191
2857
|
contextWindow: 4e5,
|
|
3192
|
-
supportsNativeTools: true,
|
|
3193
|
-
defaultToolProtocol: "native",
|
|
3194
2858
|
includedTools: ["apply_patch"],
|
|
3195
2859
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3196
2860
|
supportsImages: true,
|
|
@@ -3208,8 +2872,6 @@ var openAiNativeModels = {
|
|
|
3208
2872
|
"gpt-5-chat-latest": {
|
|
3209
2873
|
maxTokens: 128e3,
|
|
3210
2874
|
contextWindow: 4e5,
|
|
3211
|
-
supportsNativeTools: true,
|
|
3212
|
-
defaultToolProtocol: "native",
|
|
3213
2875
|
includedTools: ["apply_patch"],
|
|
3214
2876
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3215
2877
|
supportsImages: true,
|
|
@@ -3222,8 +2884,6 @@ var openAiNativeModels = {
|
|
|
3222
2884
|
"gpt-4.1": {
|
|
3223
2885
|
maxTokens: 32768,
|
|
3224
2886
|
contextWindow: 1047576,
|
|
3225
|
-
supportsNativeTools: true,
|
|
3226
|
-
defaultToolProtocol: "native",
|
|
3227
2887
|
includedTools: ["apply_patch"],
|
|
3228
2888
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3229
2889
|
supportsImages: true,
|
|
@@ -3239,8 +2899,6 @@ var openAiNativeModels = {
|
|
|
3239
2899
|
"gpt-4.1-mini": {
|
|
3240
2900
|
maxTokens: 32768,
|
|
3241
2901
|
contextWindow: 1047576,
|
|
3242
|
-
supportsNativeTools: true,
|
|
3243
|
-
defaultToolProtocol: "native",
|
|
3244
2902
|
includedTools: ["apply_patch"],
|
|
3245
2903
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3246
2904
|
supportsImages: true,
|
|
@@ -3256,8 +2914,6 @@ var openAiNativeModels = {
|
|
|
3256
2914
|
"gpt-4.1-nano": {
|
|
3257
2915
|
maxTokens: 32768,
|
|
3258
2916
|
contextWindow: 1047576,
|
|
3259
|
-
supportsNativeTools: true,
|
|
3260
|
-
defaultToolProtocol: "native",
|
|
3261
2917
|
includedTools: ["apply_patch"],
|
|
3262
2918
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3263
2919
|
supportsImages: true,
|
|
@@ -3273,8 +2929,6 @@ var openAiNativeModels = {
|
|
|
3273
2929
|
o3: {
|
|
3274
2930
|
maxTokens: 1e5,
|
|
3275
2931
|
contextWindow: 2e5,
|
|
3276
|
-
supportsNativeTools: true,
|
|
3277
|
-
defaultToolProtocol: "native",
|
|
3278
2932
|
supportsImages: true,
|
|
3279
2933
|
supportsPromptCache: true,
|
|
3280
2934
|
inputPrice: 2,
|
|
@@ -3291,8 +2945,6 @@ var openAiNativeModels = {
|
|
|
3291
2945
|
"o3-high": {
|
|
3292
2946
|
maxTokens: 1e5,
|
|
3293
2947
|
contextWindow: 2e5,
|
|
3294
|
-
supportsNativeTools: true,
|
|
3295
|
-
defaultToolProtocol: "native",
|
|
3296
2948
|
supportsImages: true,
|
|
3297
2949
|
supportsPromptCache: true,
|
|
3298
2950
|
inputPrice: 2,
|
|
@@ -3304,8 +2956,6 @@ var openAiNativeModels = {
|
|
|
3304
2956
|
"o3-low": {
|
|
3305
2957
|
maxTokens: 1e5,
|
|
3306
2958
|
contextWindow: 2e5,
|
|
3307
|
-
supportsNativeTools: true,
|
|
3308
|
-
defaultToolProtocol: "native",
|
|
3309
2959
|
supportsImages: true,
|
|
3310
2960
|
supportsPromptCache: true,
|
|
3311
2961
|
inputPrice: 2,
|
|
@@ -3317,8 +2967,6 @@ var openAiNativeModels = {
|
|
|
3317
2967
|
"o4-mini": {
|
|
3318
2968
|
maxTokens: 1e5,
|
|
3319
2969
|
contextWindow: 2e5,
|
|
3320
|
-
supportsNativeTools: true,
|
|
3321
|
-
defaultToolProtocol: "native",
|
|
3322
2970
|
supportsImages: true,
|
|
3323
2971
|
supportsPromptCache: true,
|
|
3324
2972
|
inputPrice: 1.1,
|
|
@@ -3335,8 +2983,6 @@ var openAiNativeModels = {
|
|
|
3335
2983
|
"o4-mini-high": {
|
|
3336
2984
|
maxTokens: 1e5,
|
|
3337
2985
|
contextWindow: 2e5,
|
|
3338
|
-
supportsNativeTools: true,
|
|
3339
|
-
defaultToolProtocol: "native",
|
|
3340
2986
|
supportsImages: true,
|
|
3341
2987
|
supportsPromptCache: true,
|
|
3342
2988
|
inputPrice: 1.1,
|
|
@@ -3348,8 +2994,6 @@ var openAiNativeModels = {
|
|
|
3348
2994
|
"o4-mini-low": {
|
|
3349
2995
|
maxTokens: 1e5,
|
|
3350
2996
|
contextWindow: 2e5,
|
|
3351
|
-
supportsNativeTools: true,
|
|
3352
|
-
defaultToolProtocol: "native",
|
|
3353
2997
|
supportsImages: true,
|
|
3354
2998
|
supportsPromptCache: true,
|
|
3355
2999
|
inputPrice: 1.1,
|
|
@@ -3361,8 +3005,6 @@ var openAiNativeModels = {
|
|
|
3361
3005
|
"o3-mini": {
|
|
3362
3006
|
maxTokens: 1e5,
|
|
3363
3007
|
contextWindow: 2e5,
|
|
3364
|
-
supportsNativeTools: true,
|
|
3365
|
-
defaultToolProtocol: "native",
|
|
3366
3008
|
supportsImages: false,
|
|
3367
3009
|
supportsPromptCache: true,
|
|
3368
3010
|
inputPrice: 1.1,
|
|
@@ -3375,8 +3017,6 @@ var openAiNativeModels = {
|
|
|
3375
3017
|
"o3-mini-high": {
|
|
3376
3018
|
maxTokens: 1e5,
|
|
3377
3019
|
contextWindow: 2e5,
|
|
3378
|
-
supportsNativeTools: true,
|
|
3379
|
-
defaultToolProtocol: "native",
|
|
3380
3020
|
supportsImages: false,
|
|
3381
3021
|
supportsPromptCache: true,
|
|
3382
3022
|
inputPrice: 1.1,
|
|
@@ -3388,8 +3028,6 @@ var openAiNativeModels = {
|
|
|
3388
3028
|
"o3-mini-low": {
|
|
3389
3029
|
maxTokens: 1e5,
|
|
3390
3030
|
contextWindow: 2e5,
|
|
3391
|
-
supportsNativeTools: true,
|
|
3392
|
-
defaultToolProtocol: "native",
|
|
3393
3031
|
supportsImages: false,
|
|
3394
3032
|
supportsPromptCache: true,
|
|
3395
3033
|
inputPrice: 1.1,
|
|
@@ -3401,8 +3039,6 @@ var openAiNativeModels = {
|
|
|
3401
3039
|
o1: {
|
|
3402
3040
|
maxTokens: 1e5,
|
|
3403
3041
|
contextWindow: 2e5,
|
|
3404
|
-
supportsNativeTools: true,
|
|
3405
|
-
defaultToolProtocol: "native",
|
|
3406
3042
|
supportsImages: true,
|
|
3407
3043
|
supportsPromptCache: true,
|
|
3408
3044
|
inputPrice: 15,
|
|
@@ -3413,8 +3049,6 @@ var openAiNativeModels = {
|
|
|
3413
3049
|
"o1-preview": {
|
|
3414
3050
|
maxTokens: 32768,
|
|
3415
3051
|
contextWindow: 128e3,
|
|
3416
|
-
supportsNativeTools: true,
|
|
3417
|
-
defaultToolProtocol: "native",
|
|
3418
3052
|
supportsImages: true,
|
|
3419
3053
|
supportsPromptCache: true,
|
|
3420
3054
|
inputPrice: 15,
|
|
@@ -3425,8 +3059,6 @@ var openAiNativeModels = {
|
|
|
3425
3059
|
"o1-mini": {
|
|
3426
3060
|
maxTokens: 65536,
|
|
3427
3061
|
contextWindow: 128e3,
|
|
3428
|
-
supportsNativeTools: true,
|
|
3429
|
-
defaultToolProtocol: "native",
|
|
3430
3062
|
supportsImages: true,
|
|
3431
3063
|
supportsPromptCache: true,
|
|
3432
3064
|
inputPrice: 1.1,
|
|
@@ -3437,8 +3069,6 @@ var openAiNativeModels = {
|
|
|
3437
3069
|
"gpt-4o": {
|
|
3438
3070
|
maxTokens: 16384,
|
|
3439
3071
|
contextWindow: 128e3,
|
|
3440
|
-
supportsNativeTools: true,
|
|
3441
|
-
defaultToolProtocol: "native",
|
|
3442
3072
|
supportsImages: true,
|
|
3443
3073
|
supportsPromptCache: true,
|
|
3444
3074
|
inputPrice: 2.5,
|
|
@@ -3452,8 +3082,6 @@ var openAiNativeModels = {
|
|
|
3452
3082
|
"gpt-4o-mini": {
|
|
3453
3083
|
maxTokens: 16384,
|
|
3454
3084
|
contextWindow: 128e3,
|
|
3455
|
-
supportsNativeTools: true,
|
|
3456
|
-
defaultToolProtocol: "native",
|
|
3457
3085
|
supportsImages: true,
|
|
3458
3086
|
supportsPromptCache: true,
|
|
3459
3087
|
inputPrice: 0.15,
|
|
@@ -3467,8 +3095,6 @@ var openAiNativeModels = {
|
|
|
3467
3095
|
"codex-mini-latest": {
|
|
3468
3096
|
maxTokens: 16384,
|
|
3469
3097
|
contextWindow: 2e5,
|
|
3470
|
-
supportsNativeTools: true,
|
|
3471
|
-
defaultToolProtocol: "native",
|
|
3472
3098
|
supportsImages: false,
|
|
3473
3099
|
supportsPromptCache: false,
|
|
3474
3100
|
inputPrice: 1.5,
|
|
@@ -3481,8 +3107,6 @@ var openAiNativeModels = {
|
|
|
3481
3107
|
"gpt-5-2025-08-07": {
|
|
3482
3108
|
maxTokens: 128e3,
|
|
3483
3109
|
contextWindow: 4e5,
|
|
3484
|
-
supportsNativeTools: true,
|
|
3485
|
-
defaultToolProtocol: "native",
|
|
3486
3110
|
includedTools: ["apply_patch"],
|
|
3487
3111
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3488
3112
|
supportsImages: true,
|
|
@@ -3503,8 +3127,6 @@ var openAiNativeModels = {
|
|
|
3503
3127
|
"gpt-5-mini-2025-08-07": {
|
|
3504
3128
|
maxTokens: 128e3,
|
|
3505
3129
|
contextWindow: 4e5,
|
|
3506
|
-
supportsNativeTools: true,
|
|
3507
|
-
defaultToolProtocol: "native",
|
|
3508
3130
|
includedTools: ["apply_patch"],
|
|
3509
3131
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3510
3132
|
supportsImages: true,
|
|
@@ -3525,8 +3147,6 @@ var openAiNativeModels = {
|
|
|
3525
3147
|
"gpt-5-nano-2025-08-07": {
|
|
3526
3148
|
maxTokens: 128e3,
|
|
3527
3149
|
contextWindow: 4e5,
|
|
3528
|
-
supportsNativeTools: true,
|
|
3529
|
-
defaultToolProtocol: "native",
|
|
3530
3150
|
includedTools: ["apply_patch"],
|
|
3531
3151
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3532
3152
|
supportsImages: true,
|
|
@@ -3548,9 +3168,7 @@ var openAiModelInfoSaneDefaults = {
|
|
|
3548
3168
|
supportsImages: true,
|
|
3549
3169
|
supportsPromptCache: false,
|
|
3550
3170
|
inputPrice: 0,
|
|
3551
|
-
outputPrice: 0
|
|
3552
|
-
supportsNativeTools: true,
|
|
3553
|
-
defaultToolProtocol: "native"
|
|
3171
|
+
outputPrice: 0
|
|
3554
3172
|
};
|
|
3555
3173
|
var azureOpenAiDefaultApiVersion = "2024-08-01-preview";
|
|
3556
3174
|
var OPENAI_NATIVE_DEFAULT_TEMPERATURE = 0;
|
|
@@ -3562,8 +3180,6 @@ var openAiCodexModels = {
|
|
|
3562
3180
|
"gpt-5.1-codex-max": {
|
|
3563
3181
|
maxTokens: 128e3,
|
|
3564
3182
|
contextWindow: 4e5,
|
|
3565
|
-
supportsNativeTools: true,
|
|
3566
|
-
defaultToolProtocol: "native",
|
|
3567
3183
|
includedTools: ["apply_patch"],
|
|
3568
3184
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3569
3185
|
supportsImages: true,
|
|
@@ -3576,11 +3192,24 @@ var openAiCodexModels = {
|
|
|
3576
3192
|
supportsTemperature: false,
|
|
3577
3193
|
description: "GPT-5.1 Codex Max: Maximum capability coding model via ChatGPT subscription"
|
|
3578
3194
|
},
|
|
3195
|
+
"gpt-5.1-codex": {
|
|
3196
|
+
maxTokens: 128e3,
|
|
3197
|
+
contextWindow: 4e5,
|
|
3198
|
+
includedTools: ["apply_patch"],
|
|
3199
|
+
excludedTools: ["apply_diff", "write_to_file"],
|
|
3200
|
+
supportsImages: true,
|
|
3201
|
+
supportsPromptCache: true,
|
|
3202
|
+
supportsReasoningEffort: ["low", "medium", "high"],
|
|
3203
|
+
reasoningEffort: "medium",
|
|
3204
|
+
// Subscription-based: no per-token costs
|
|
3205
|
+
inputPrice: 0,
|
|
3206
|
+
outputPrice: 0,
|
|
3207
|
+
supportsTemperature: false,
|
|
3208
|
+
description: "GPT-5.1 Codex: GPT-5.1 optimized for agentic coding via ChatGPT subscription"
|
|
3209
|
+
},
|
|
3579
3210
|
"gpt-5.2-codex": {
|
|
3580
3211
|
maxTokens: 128e3,
|
|
3581
3212
|
contextWindow: 4e5,
|
|
3582
|
-
supportsNativeTools: true,
|
|
3583
|
-
defaultToolProtocol: "native",
|
|
3584
3213
|
includedTools: ["apply_patch"],
|
|
3585
3214
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3586
3215
|
supportsImages: true,
|
|
@@ -3592,11 +3221,71 @@ var openAiCodexModels = {
|
|
|
3592
3221
|
supportsTemperature: false,
|
|
3593
3222
|
description: "GPT-5.2 Codex: OpenAI's flagship coding model via ChatGPT subscription"
|
|
3594
3223
|
},
|
|
3224
|
+
"gpt-5.1": {
|
|
3225
|
+
maxTokens: 128e3,
|
|
3226
|
+
contextWindow: 4e5,
|
|
3227
|
+
includedTools: ["apply_patch"],
|
|
3228
|
+
excludedTools: ["apply_diff", "write_to_file"],
|
|
3229
|
+
supportsImages: true,
|
|
3230
|
+
supportsPromptCache: true,
|
|
3231
|
+
supportsReasoningEffort: ["none", "low", "medium", "high"],
|
|
3232
|
+
reasoningEffort: "medium",
|
|
3233
|
+
// Subscription-based: no per-token costs
|
|
3234
|
+
inputPrice: 0,
|
|
3235
|
+
outputPrice: 0,
|
|
3236
|
+
supportsVerbosity: true,
|
|
3237
|
+
supportsTemperature: false,
|
|
3238
|
+
description: "GPT-5.1: General GPT-5.1 model via ChatGPT subscription"
|
|
3239
|
+
},
|
|
3240
|
+
"gpt-5": {
|
|
3241
|
+
maxTokens: 128e3,
|
|
3242
|
+
contextWindow: 4e5,
|
|
3243
|
+
includedTools: ["apply_patch"],
|
|
3244
|
+
excludedTools: ["apply_diff", "write_to_file"],
|
|
3245
|
+
supportsImages: true,
|
|
3246
|
+
supportsPromptCache: true,
|
|
3247
|
+
supportsReasoningEffort: ["minimal", "low", "medium", "high"],
|
|
3248
|
+
reasoningEffort: "medium",
|
|
3249
|
+
// Subscription-based: no per-token costs
|
|
3250
|
+
inputPrice: 0,
|
|
3251
|
+
outputPrice: 0,
|
|
3252
|
+
supportsVerbosity: true,
|
|
3253
|
+
supportsTemperature: false,
|
|
3254
|
+
description: "GPT-5: General GPT-5 model via ChatGPT subscription"
|
|
3255
|
+
},
|
|
3256
|
+
"gpt-5-codex": {
|
|
3257
|
+
maxTokens: 128e3,
|
|
3258
|
+
contextWindow: 4e5,
|
|
3259
|
+
includedTools: ["apply_patch"],
|
|
3260
|
+
excludedTools: ["apply_diff", "write_to_file"],
|
|
3261
|
+
supportsImages: true,
|
|
3262
|
+
supportsPromptCache: true,
|
|
3263
|
+
supportsReasoningEffort: ["low", "medium", "high"],
|
|
3264
|
+
reasoningEffort: "medium",
|
|
3265
|
+
// Subscription-based: no per-token costs
|
|
3266
|
+
inputPrice: 0,
|
|
3267
|
+
outputPrice: 0,
|
|
3268
|
+
supportsTemperature: false,
|
|
3269
|
+
description: "GPT-5 Codex: GPT-5 optimized for agentic coding via ChatGPT subscription"
|
|
3270
|
+
},
|
|
3271
|
+
"gpt-5-codex-mini": {
|
|
3272
|
+
maxTokens: 128e3,
|
|
3273
|
+
contextWindow: 4e5,
|
|
3274
|
+
includedTools: ["apply_patch"],
|
|
3275
|
+
excludedTools: ["apply_diff", "write_to_file"],
|
|
3276
|
+
supportsImages: true,
|
|
3277
|
+
supportsPromptCache: true,
|
|
3278
|
+
supportsReasoningEffort: ["low", "medium", "high"],
|
|
3279
|
+
reasoningEffort: "medium",
|
|
3280
|
+
// Subscription-based: no per-token costs
|
|
3281
|
+
inputPrice: 0,
|
|
3282
|
+
outputPrice: 0,
|
|
3283
|
+
supportsTemperature: false,
|
|
3284
|
+
description: "GPT-5 Codex Mini: Faster coding model via ChatGPT subscription"
|
|
3285
|
+
},
|
|
3595
3286
|
"gpt-5.1-codex-mini": {
|
|
3596
3287
|
maxTokens: 128e3,
|
|
3597
3288
|
contextWindow: 4e5,
|
|
3598
|
-
supportsNativeTools: true,
|
|
3599
|
-
defaultToolProtocol: "native",
|
|
3600
3289
|
includedTools: ["apply_patch"],
|
|
3601
3290
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3602
3291
|
supportsImages: true,
|
|
@@ -3611,8 +3300,6 @@ var openAiCodexModels = {
|
|
|
3611
3300
|
"gpt-5.2": {
|
|
3612
3301
|
maxTokens: 128e3,
|
|
3613
3302
|
contextWindow: 4e5,
|
|
3614
|
-
supportsNativeTools: true,
|
|
3615
|
-
defaultToolProtocol: "native",
|
|
3616
3303
|
includedTools: ["apply_patch"],
|
|
3617
3304
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3618
3305
|
supportsImages: true,
|
|
@@ -3633,7 +3320,6 @@ var openRouterDefaultModelInfo = {
|
|
|
3633
3320
|
contextWindow: 2e5,
|
|
3634
3321
|
supportsImages: true,
|
|
3635
3322
|
supportsPromptCache: true,
|
|
3636
|
-
supportsNativeTools: true,
|
|
3637
3323
|
inputPrice: 3,
|
|
3638
3324
|
outputPrice: 15,
|
|
3639
3325
|
cacheWritesPrice: 3.75,
|
|
@@ -3707,8 +3393,6 @@ var qwenCodeModels = {
|
|
|
3707
3393
|
contextWindow: 1e6,
|
|
3708
3394
|
supportsImages: false,
|
|
3709
3395
|
supportsPromptCache: false,
|
|
3710
|
-
supportsNativeTools: true,
|
|
3711
|
-
defaultToolProtocol: "native",
|
|
3712
3396
|
inputPrice: 0,
|
|
3713
3397
|
outputPrice: 0,
|
|
3714
3398
|
cacheWritesPrice: 0,
|
|
@@ -3720,8 +3404,6 @@ var qwenCodeModels = {
|
|
|
3720
3404
|
contextWindow: 1e6,
|
|
3721
3405
|
supportsImages: false,
|
|
3722
3406
|
supportsPromptCache: false,
|
|
3723
|
-
supportsNativeTools: true,
|
|
3724
|
-
defaultToolProtocol: "native",
|
|
3725
3407
|
inputPrice: 0,
|
|
3726
3408
|
outputPrice: 0,
|
|
3727
3409
|
cacheWritesPrice: 0,
|
|
@@ -3737,8 +3419,6 @@ var requestyDefaultModelInfo = {
|
|
|
3737
3419
|
contextWindow: 2e5,
|
|
3738
3420
|
supportsImages: true,
|
|
3739
3421
|
supportsPromptCache: true,
|
|
3740
|
-
supportsNativeTools: true,
|
|
3741
|
-
defaultToolProtocol: "native",
|
|
3742
3422
|
inputPrice: 3,
|
|
3743
3423
|
outputPrice: 15,
|
|
3744
3424
|
cacheWritesPrice: 3.75,
|
|
@@ -3794,8 +3474,6 @@ var sambaNovaModels = {
|
|
|
3794
3474
|
contextWindow: 16384,
|
|
3795
3475
|
supportsImages: false,
|
|
3796
3476
|
supportsPromptCache: false,
|
|
3797
|
-
supportsNativeTools: true,
|
|
3798
|
-
defaultToolProtocol: "native",
|
|
3799
3477
|
inputPrice: 0.1,
|
|
3800
3478
|
outputPrice: 0.2,
|
|
3801
3479
|
description: "Meta Llama 3.1 8B Instruct model with 16K context window."
|
|
@@ -3805,8 +3483,6 @@ var sambaNovaModels = {
|
|
|
3805
3483
|
contextWindow: 131072,
|
|
3806
3484
|
supportsImages: false,
|
|
3807
3485
|
supportsPromptCache: false,
|
|
3808
|
-
supportsNativeTools: true,
|
|
3809
|
-
defaultToolProtocol: "native",
|
|
3810
3486
|
inputPrice: 0.6,
|
|
3811
3487
|
outputPrice: 1.2,
|
|
3812
3488
|
description: "Meta Llama 3.3 70B Instruct model with 128K context window."
|
|
@@ -3817,8 +3493,6 @@ var sambaNovaModels = {
|
|
|
3817
3493
|
supportsImages: false,
|
|
3818
3494
|
supportsPromptCache: false,
|
|
3819
3495
|
supportsReasoningBudget: true,
|
|
3820
|
-
supportsNativeTools: true,
|
|
3821
|
-
defaultToolProtocol: "native",
|
|
3822
3496
|
inputPrice: 5,
|
|
3823
3497
|
outputPrice: 7,
|
|
3824
3498
|
description: "DeepSeek R1 reasoning model with 32K context window."
|
|
@@ -3828,8 +3502,6 @@ var sambaNovaModels = {
|
|
|
3828
3502
|
contextWindow: 32768,
|
|
3829
3503
|
supportsImages: false,
|
|
3830
3504
|
supportsPromptCache: false,
|
|
3831
|
-
supportsNativeTools: true,
|
|
3832
|
-
defaultToolProtocol: "native",
|
|
3833
3505
|
inputPrice: 3,
|
|
3834
3506
|
outputPrice: 4.5,
|
|
3835
3507
|
description: "DeepSeek V3 model with 32K context window."
|
|
@@ -3839,8 +3511,6 @@ var sambaNovaModels = {
|
|
|
3839
3511
|
contextWindow: 32768,
|
|
3840
3512
|
supportsImages: false,
|
|
3841
3513
|
supportsPromptCache: false,
|
|
3842
|
-
supportsNativeTools: true,
|
|
3843
|
-
defaultToolProtocol: "native",
|
|
3844
3514
|
inputPrice: 3,
|
|
3845
3515
|
outputPrice: 4.5,
|
|
3846
3516
|
description: "DeepSeek V3.1 model with 32K context window."
|
|
@@ -3850,8 +3520,6 @@ var sambaNovaModels = {
|
|
|
3850
3520
|
contextWindow: 131072,
|
|
3851
3521
|
supportsImages: true,
|
|
3852
3522
|
supportsPromptCache: false,
|
|
3853
|
-
supportsNativeTools: true,
|
|
3854
|
-
defaultToolProtocol: "native",
|
|
3855
3523
|
inputPrice: 0.63,
|
|
3856
3524
|
outputPrice: 1.8,
|
|
3857
3525
|
description: "Meta Llama 4 Maverick 17B 128E Instruct model with 128K context window."
|
|
@@ -3861,8 +3529,6 @@ var sambaNovaModels = {
|
|
|
3861
3529
|
contextWindow: 8192,
|
|
3862
3530
|
supportsImages: false,
|
|
3863
3531
|
supportsPromptCache: false,
|
|
3864
|
-
supportsNativeTools: true,
|
|
3865
|
-
defaultToolProtocol: "native",
|
|
3866
3532
|
inputPrice: 0.4,
|
|
3867
3533
|
outputPrice: 0.8,
|
|
3868
3534
|
description: "Alibaba Qwen 3 32B model with 8K context window."
|
|
@@ -3872,8 +3538,6 @@ var sambaNovaModels = {
|
|
|
3872
3538
|
contextWindow: 131072,
|
|
3873
3539
|
supportsImages: false,
|
|
3874
3540
|
supportsPromptCache: false,
|
|
3875
|
-
supportsNativeTools: true,
|
|
3876
|
-
defaultToolProtocol: "native",
|
|
3877
3541
|
inputPrice: 0.22,
|
|
3878
3542
|
outputPrice: 0.59,
|
|
3879
3543
|
description: "OpenAI gpt oss 120b model with 128k context window."
|
|
@@ -3887,7 +3551,6 @@ var unboundDefaultModelInfo = {
|
|
|
3887
3551
|
contextWindow: 2e5,
|
|
3888
3552
|
supportsImages: true,
|
|
3889
3553
|
supportsPromptCache: true,
|
|
3890
|
-
supportsNativeTools: true,
|
|
3891
3554
|
inputPrice: 3,
|
|
3892
3555
|
outputPrice: 15,
|
|
3893
3556
|
cacheWritesPrice: 3.75,
|
|
@@ -3901,8 +3564,6 @@ var vertexModels = {
|
|
|
3901
3564
|
maxTokens: 65536,
|
|
3902
3565
|
contextWindow: 1048576,
|
|
3903
3566
|
supportsImages: true,
|
|
3904
|
-
supportsNativeTools: true,
|
|
3905
|
-
defaultToolProtocol: "native",
|
|
3906
3567
|
supportsPromptCache: true,
|
|
3907
3568
|
supportsReasoningEffort: ["low", "high"],
|
|
3908
3569
|
reasoningEffort: "low",
|
|
@@ -3910,16 +3571,19 @@ var vertexModels = {
|
|
|
3910
3571
|
defaultTemperature: 1,
|
|
3911
3572
|
inputPrice: 4,
|
|
3912
3573
|
outputPrice: 18,
|
|
3574
|
+
cacheReadsPrice: 0.4,
|
|
3913
3575
|
tiers: [
|
|
3914
3576
|
{
|
|
3915
3577
|
contextWindow: 2e5,
|
|
3916
3578
|
inputPrice: 2,
|
|
3917
|
-
outputPrice: 12
|
|
3579
|
+
outputPrice: 12,
|
|
3580
|
+
cacheReadsPrice: 0.2
|
|
3918
3581
|
},
|
|
3919
3582
|
{
|
|
3920
3583
|
contextWindow: Infinity,
|
|
3921
3584
|
inputPrice: 4,
|
|
3922
|
-
outputPrice: 18
|
|
3585
|
+
outputPrice: 18,
|
|
3586
|
+
cacheReadsPrice: 0.4
|
|
3923
3587
|
}
|
|
3924
3588
|
]
|
|
3925
3589
|
},
|
|
@@ -3927,24 +3591,19 @@ var vertexModels = {
|
|
|
3927
3591
|
maxTokens: 65536,
|
|
3928
3592
|
contextWindow: 1048576,
|
|
3929
3593
|
supportsImages: true,
|
|
3930
|
-
supportsNativeTools: true,
|
|
3931
|
-
defaultToolProtocol: "native",
|
|
3932
3594
|
supportsPromptCache: true,
|
|
3933
3595
|
supportsReasoningEffort: ["minimal", "low", "medium", "high"],
|
|
3934
3596
|
reasoningEffort: "medium",
|
|
3935
3597
|
supportsTemperature: true,
|
|
3936
3598
|
defaultTemperature: 1,
|
|
3937
|
-
inputPrice: 0.
|
|
3938
|
-
outputPrice:
|
|
3939
|
-
cacheReadsPrice: 0.
|
|
3940
|
-
cacheWritesPrice: 1
|
|
3599
|
+
inputPrice: 0.5,
|
|
3600
|
+
outputPrice: 3,
|
|
3601
|
+
cacheReadsPrice: 0.05
|
|
3941
3602
|
},
|
|
3942
3603
|
"gemini-2.5-flash-preview-05-20:thinking": {
|
|
3943
3604
|
maxTokens: 65535,
|
|
3944
3605
|
contextWindow: 1048576,
|
|
3945
3606
|
supportsImages: true,
|
|
3946
|
-
supportsNativeTools: true,
|
|
3947
|
-
defaultToolProtocol: "native",
|
|
3948
3607
|
supportsPromptCache: true,
|
|
3949
3608
|
inputPrice: 0.15,
|
|
3950
3609
|
outputPrice: 3.5,
|
|
@@ -3956,8 +3615,6 @@ var vertexModels = {
|
|
|
3956
3615
|
maxTokens: 65535,
|
|
3957
3616
|
contextWindow: 1048576,
|
|
3958
3617
|
supportsImages: true,
|
|
3959
|
-
supportsNativeTools: true,
|
|
3960
|
-
defaultToolProtocol: "native",
|
|
3961
3618
|
supportsPromptCache: true,
|
|
3962
3619
|
inputPrice: 0.15,
|
|
3963
3620
|
outputPrice: 0.6
|
|
@@ -3966,8 +3623,6 @@ var vertexModels = {
|
|
|
3966
3623
|
maxTokens: 64e3,
|
|
3967
3624
|
contextWindow: 1048576,
|
|
3968
3625
|
supportsImages: true,
|
|
3969
|
-
supportsNativeTools: true,
|
|
3970
|
-
defaultToolProtocol: "native",
|
|
3971
3626
|
supportsPromptCache: true,
|
|
3972
3627
|
inputPrice: 0.3,
|
|
3973
3628
|
outputPrice: 2.5,
|
|
@@ -3980,8 +3635,6 @@ var vertexModels = {
|
|
|
3980
3635
|
maxTokens: 65535,
|
|
3981
3636
|
contextWindow: 1048576,
|
|
3982
3637
|
supportsImages: true,
|
|
3983
|
-
supportsNativeTools: true,
|
|
3984
|
-
defaultToolProtocol: "native",
|
|
3985
3638
|
supportsPromptCache: false,
|
|
3986
3639
|
inputPrice: 0.15,
|
|
3987
3640
|
outputPrice: 3.5,
|
|
@@ -3993,8 +3646,6 @@ var vertexModels = {
|
|
|
3993
3646
|
maxTokens: 65535,
|
|
3994
3647
|
contextWindow: 1048576,
|
|
3995
3648
|
supportsImages: true,
|
|
3996
|
-
supportsNativeTools: true,
|
|
3997
|
-
defaultToolProtocol: "native",
|
|
3998
3649
|
supportsPromptCache: false,
|
|
3999
3650
|
inputPrice: 0.15,
|
|
4000
3651
|
outputPrice: 0.6
|
|
@@ -4003,8 +3654,6 @@ var vertexModels = {
|
|
|
4003
3654
|
maxTokens: 65535,
|
|
4004
3655
|
contextWindow: 1048576,
|
|
4005
3656
|
supportsImages: true,
|
|
4006
|
-
supportsNativeTools: true,
|
|
4007
|
-
defaultToolProtocol: "native",
|
|
4008
3657
|
supportsPromptCache: true,
|
|
4009
3658
|
inputPrice: 2.5,
|
|
4010
3659
|
outputPrice: 15
|
|
@@ -4013,8 +3662,6 @@ var vertexModels = {
|
|
|
4013
3662
|
maxTokens: 65535,
|
|
4014
3663
|
contextWindow: 1048576,
|
|
4015
3664
|
supportsImages: true,
|
|
4016
|
-
supportsNativeTools: true,
|
|
4017
|
-
defaultToolProtocol: "native",
|
|
4018
3665
|
supportsPromptCache: true,
|
|
4019
3666
|
inputPrice: 2.5,
|
|
4020
3667
|
outputPrice: 15
|
|
@@ -4023,8 +3670,6 @@ var vertexModels = {
|
|
|
4023
3670
|
maxTokens: 65535,
|
|
4024
3671
|
contextWindow: 1048576,
|
|
4025
3672
|
supportsImages: true,
|
|
4026
|
-
supportsNativeTools: true,
|
|
4027
|
-
defaultToolProtocol: "native",
|
|
4028
3673
|
supportsPromptCache: true,
|
|
4029
3674
|
inputPrice: 2.5,
|
|
4030
3675
|
outputPrice: 15,
|
|
@@ -4035,8 +3680,6 @@ var vertexModels = {
|
|
|
4035
3680
|
maxTokens: 64e3,
|
|
4036
3681
|
contextWindow: 1048576,
|
|
4037
3682
|
supportsImages: true,
|
|
4038
|
-
supportsNativeTools: true,
|
|
4039
|
-
defaultToolProtocol: "native",
|
|
4040
3683
|
supportsPromptCache: true,
|
|
4041
3684
|
inputPrice: 2.5,
|
|
4042
3685
|
outputPrice: 15,
|
|
@@ -4062,8 +3705,6 @@ var vertexModels = {
|
|
|
4062
3705
|
maxTokens: 65535,
|
|
4063
3706
|
contextWindow: 1048576,
|
|
4064
3707
|
supportsImages: true,
|
|
4065
|
-
supportsNativeTools: true,
|
|
4066
|
-
defaultToolProtocol: "native",
|
|
4067
3708
|
supportsPromptCache: false,
|
|
4068
3709
|
inputPrice: 0,
|
|
4069
3710
|
outputPrice: 0
|
|
@@ -4072,8 +3713,6 @@ var vertexModels = {
|
|
|
4072
3713
|
maxTokens: 8192,
|
|
4073
3714
|
contextWindow: 2097152,
|
|
4074
3715
|
supportsImages: true,
|
|
4075
|
-
supportsNativeTools: true,
|
|
4076
|
-
defaultToolProtocol: "native",
|
|
4077
3716
|
supportsPromptCache: false,
|
|
4078
3717
|
inputPrice: 0,
|
|
4079
3718
|
outputPrice: 0
|
|
@@ -4082,8 +3721,6 @@ var vertexModels = {
|
|
|
4082
3721
|
maxTokens: 8192,
|
|
4083
3722
|
contextWindow: 1048576,
|
|
4084
3723
|
supportsImages: true,
|
|
4085
|
-
supportsNativeTools: true,
|
|
4086
|
-
defaultToolProtocol: "native",
|
|
4087
3724
|
supportsPromptCache: true,
|
|
4088
3725
|
inputPrice: 0.15,
|
|
4089
3726
|
outputPrice: 0.6
|
|
@@ -4092,8 +3729,6 @@ var vertexModels = {
|
|
|
4092
3729
|
maxTokens: 8192,
|
|
4093
3730
|
contextWindow: 1048576,
|
|
4094
3731
|
supportsImages: true,
|
|
4095
|
-
supportsNativeTools: true,
|
|
4096
|
-
defaultToolProtocol: "native",
|
|
4097
3732
|
supportsPromptCache: false,
|
|
4098
3733
|
inputPrice: 0.075,
|
|
4099
3734
|
outputPrice: 0.3
|
|
@@ -4102,8 +3737,6 @@ var vertexModels = {
|
|
|
4102
3737
|
maxTokens: 8192,
|
|
4103
3738
|
contextWindow: 32768,
|
|
4104
3739
|
supportsImages: true,
|
|
4105
|
-
supportsNativeTools: true,
|
|
4106
|
-
defaultToolProtocol: "native",
|
|
4107
3740
|
supportsPromptCache: false,
|
|
4108
3741
|
inputPrice: 0,
|
|
4109
3742
|
outputPrice: 0
|
|
@@ -4112,8 +3745,6 @@ var vertexModels = {
|
|
|
4112
3745
|
maxTokens: 8192,
|
|
4113
3746
|
contextWindow: 1048576,
|
|
4114
3747
|
supportsImages: true,
|
|
4115
|
-
supportsNativeTools: true,
|
|
4116
|
-
defaultToolProtocol: "native",
|
|
4117
3748
|
supportsPromptCache: true,
|
|
4118
3749
|
inputPrice: 0.075,
|
|
4119
3750
|
outputPrice: 0.3
|
|
@@ -4122,8 +3753,6 @@ var vertexModels = {
|
|
|
4122
3753
|
maxTokens: 8192,
|
|
4123
3754
|
contextWindow: 2097152,
|
|
4124
3755
|
supportsImages: true,
|
|
4125
|
-
supportsNativeTools: true,
|
|
4126
|
-
defaultToolProtocol: "native",
|
|
4127
3756
|
supportsPromptCache: false,
|
|
4128
3757
|
inputPrice: 1.25,
|
|
4129
3758
|
outputPrice: 5
|
|
@@ -4134,8 +3763,6 @@ var vertexModels = {
|
|
|
4134
3763
|
// Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
|
|
4135
3764
|
supportsImages: true,
|
|
4136
3765
|
supportsPromptCache: true,
|
|
4137
|
-
supportsNativeTools: true,
|
|
4138
|
-
defaultToolProtocol: "native",
|
|
4139
3766
|
inputPrice: 3,
|
|
4140
3767
|
// $3 per million input tokens (≤200K context)
|
|
4141
3768
|
outputPrice: 15,
|
|
@@ -4167,8 +3794,6 @@ var vertexModels = {
|
|
|
4167
3794
|
// Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
|
|
4168
3795
|
supportsImages: true,
|
|
4169
3796
|
supportsPromptCache: true,
|
|
4170
|
-
supportsNativeTools: true,
|
|
4171
|
-
defaultToolProtocol: "native",
|
|
4172
3797
|
inputPrice: 3,
|
|
4173
3798
|
// $3 per million input tokens (≤200K context)
|
|
4174
3799
|
outputPrice: 15,
|
|
@@ -4199,8 +3824,6 @@ var vertexModels = {
|
|
|
4199
3824
|
contextWindow: 2e5,
|
|
4200
3825
|
supportsImages: true,
|
|
4201
3826
|
supportsPromptCache: true,
|
|
4202
|
-
supportsNativeTools: true,
|
|
4203
|
-
defaultToolProtocol: "native",
|
|
4204
3827
|
inputPrice: 1,
|
|
4205
3828
|
outputPrice: 5,
|
|
4206
3829
|
cacheWritesPrice: 1.25,
|
|
@@ -4212,8 +3835,6 @@ var vertexModels = {
|
|
|
4212
3835
|
contextWindow: 2e5,
|
|
4213
3836
|
supportsImages: true,
|
|
4214
3837
|
supportsPromptCache: true,
|
|
4215
|
-
supportsNativeTools: true,
|
|
4216
|
-
defaultToolProtocol: "native",
|
|
4217
3838
|
inputPrice: 5,
|
|
4218
3839
|
outputPrice: 25,
|
|
4219
3840
|
cacheWritesPrice: 6.25,
|
|
@@ -4225,8 +3846,6 @@ var vertexModels = {
|
|
|
4225
3846
|
contextWindow: 2e5,
|
|
4226
3847
|
supportsImages: true,
|
|
4227
3848
|
supportsPromptCache: true,
|
|
4228
|
-
supportsNativeTools: true,
|
|
4229
|
-
defaultToolProtocol: "native",
|
|
4230
3849
|
inputPrice: 15,
|
|
4231
3850
|
outputPrice: 75,
|
|
4232
3851
|
cacheWritesPrice: 18.75,
|
|
@@ -4238,8 +3857,6 @@ var vertexModels = {
|
|
|
4238
3857
|
contextWindow: 2e5,
|
|
4239
3858
|
supportsImages: true,
|
|
4240
3859
|
supportsPromptCache: true,
|
|
4241
|
-
supportsNativeTools: true,
|
|
4242
|
-
defaultToolProtocol: "native",
|
|
4243
3860
|
inputPrice: 15,
|
|
4244
3861
|
outputPrice: 75,
|
|
4245
3862
|
cacheWritesPrice: 18.75,
|
|
@@ -4250,8 +3867,6 @@ var vertexModels = {
|
|
|
4250
3867
|
contextWindow: 2e5,
|
|
4251
3868
|
supportsImages: true,
|
|
4252
3869
|
supportsPromptCache: true,
|
|
4253
|
-
supportsNativeTools: true,
|
|
4254
|
-
defaultToolProtocol: "native",
|
|
4255
3870
|
inputPrice: 3,
|
|
4256
3871
|
outputPrice: 15,
|
|
4257
3872
|
cacheWritesPrice: 3.75,
|
|
@@ -4264,8 +3879,6 @@ var vertexModels = {
|
|
|
4264
3879
|
contextWindow: 2e5,
|
|
4265
3880
|
supportsImages: true,
|
|
4266
3881
|
supportsPromptCache: true,
|
|
4267
|
-
supportsNativeTools: true,
|
|
4268
|
-
defaultToolProtocol: "native",
|
|
4269
3882
|
inputPrice: 3,
|
|
4270
3883
|
outputPrice: 15,
|
|
4271
3884
|
cacheWritesPrice: 3.75,
|
|
@@ -4276,8 +3889,6 @@ var vertexModels = {
|
|
|
4276
3889
|
contextWindow: 2e5,
|
|
4277
3890
|
supportsImages: true,
|
|
4278
3891
|
supportsPromptCache: true,
|
|
4279
|
-
supportsNativeTools: true,
|
|
4280
|
-
defaultToolProtocol: "native",
|
|
4281
3892
|
inputPrice: 3,
|
|
4282
3893
|
outputPrice: 15,
|
|
4283
3894
|
cacheWritesPrice: 3.75,
|
|
@@ -4288,8 +3899,6 @@ var vertexModels = {
|
|
|
4288
3899
|
contextWindow: 2e5,
|
|
4289
3900
|
supportsImages: true,
|
|
4290
3901
|
supportsPromptCache: true,
|
|
4291
|
-
supportsNativeTools: true,
|
|
4292
|
-
defaultToolProtocol: "native",
|
|
4293
3902
|
inputPrice: 3,
|
|
4294
3903
|
outputPrice: 15,
|
|
4295
3904
|
cacheWritesPrice: 3.75,
|
|
@@ -4300,8 +3909,6 @@ var vertexModels = {
|
|
|
4300
3909
|
contextWindow: 2e5,
|
|
4301
3910
|
supportsImages: false,
|
|
4302
3911
|
supportsPromptCache: true,
|
|
4303
|
-
supportsNativeTools: true,
|
|
4304
|
-
defaultToolProtocol: "native",
|
|
4305
3912
|
inputPrice: 1,
|
|
4306
3913
|
outputPrice: 5,
|
|
4307
3914
|
cacheWritesPrice: 1.25,
|
|
@@ -4312,8 +3919,6 @@ var vertexModels = {
|
|
|
4312
3919
|
contextWindow: 2e5,
|
|
4313
3920
|
supportsImages: true,
|
|
4314
3921
|
supportsPromptCache: true,
|
|
4315
|
-
supportsNativeTools: true,
|
|
4316
|
-
defaultToolProtocol: "native",
|
|
4317
3922
|
inputPrice: 15,
|
|
4318
3923
|
outputPrice: 75,
|
|
4319
3924
|
cacheWritesPrice: 18.75,
|
|
@@ -4324,8 +3929,6 @@ var vertexModels = {
|
|
|
4324
3929
|
contextWindow: 2e5,
|
|
4325
3930
|
supportsImages: true,
|
|
4326
3931
|
supportsPromptCache: true,
|
|
4327
|
-
supportsNativeTools: true,
|
|
4328
|
-
defaultToolProtocol: "native",
|
|
4329
3932
|
inputPrice: 0.25,
|
|
4330
3933
|
outputPrice: 1.25,
|
|
4331
3934
|
cacheWritesPrice: 0.3,
|
|
@@ -4335,8 +3938,6 @@ var vertexModels = {
|
|
|
4335
3938
|
maxTokens: 64e3,
|
|
4336
3939
|
contextWindow: 1048576,
|
|
4337
3940
|
supportsImages: true,
|
|
4338
|
-
supportsNativeTools: true,
|
|
4339
|
-
defaultToolProtocol: "native",
|
|
4340
3941
|
supportsPromptCache: true,
|
|
4341
3942
|
inputPrice: 0.1,
|
|
4342
3943
|
outputPrice: 0.4,
|
|
@@ -4350,7 +3951,6 @@ var vertexModels = {
|
|
|
4350
3951
|
contextWindow: 131072,
|
|
4351
3952
|
supportsImages: false,
|
|
4352
3953
|
supportsPromptCache: false,
|
|
4353
|
-
supportsNativeTools: true,
|
|
4354
3954
|
inputPrice: 0.35,
|
|
4355
3955
|
outputPrice: 1.15,
|
|
4356
3956
|
description: "Meta Llama 4 Maverick 17B Instruct model, 128K context."
|
|
@@ -4360,7 +3960,6 @@ var vertexModels = {
|
|
|
4360
3960
|
contextWindow: 163840,
|
|
4361
3961
|
supportsImages: false,
|
|
4362
3962
|
supportsPromptCache: false,
|
|
4363
|
-
supportsNativeTools: true,
|
|
4364
3963
|
inputPrice: 1.35,
|
|
4365
3964
|
outputPrice: 5.4,
|
|
4366
3965
|
description: "DeepSeek R1 (0528). Available in us-central1"
|
|
@@ -4370,7 +3969,6 @@ var vertexModels = {
|
|
|
4370
3969
|
contextWindow: 163840,
|
|
4371
3970
|
supportsImages: false,
|
|
4372
3971
|
supportsPromptCache: false,
|
|
4373
|
-
supportsNativeTools: true,
|
|
4374
3972
|
inputPrice: 0.6,
|
|
4375
3973
|
outputPrice: 1.7,
|
|
4376
3974
|
description: "DeepSeek V3.1. Available in us-west2"
|
|
@@ -4380,7 +3978,6 @@ var vertexModels = {
|
|
|
4380
3978
|
contextWindow: 131072,
|
|
4381
3979
|
supportsImages: false,
|
|
4382
3980
|
supportsPromptCache: false,
|
|
4383
|
-
supportsNativeTools: true,
|
|
4384
3981
|
inputPrice: 0.15,
|
|
4385
3982
|
outputPrice: 0.6,
|
|
4386
3983
|
description: "OpenAI gpt-oss 120B. Available in us-central1"
|
|
@@ -4390,7 +3987,6 @@ var vertexModels = {
|
|
|
4390
3987
|
contextWindow: 131072,
|
|
4391
3988
|
supportsImages: false,
|
|
4392
3989
|
supportsPromptCache: false,
|
|
4393
|
-
supportsNativeTools: true,
|
|
4394
3990
|
inputPrice: 0.075,
|
|
4395
3991
|
outputPrice: 0.3,
|
|
4396
3992
|
description: "OpenAI gpt-oss 20B. Available in us-central1"
|
|
@@ -4400,7 +3996,6 @@ var vertexModels = {
|
|
|
4400
3996
|
contextWindow: 262144,
|
|
4401
3997
|
supportsImages: false,
|
|
4402
3998
|
supportsPromptCache: false,
|
|
4403
|
-
supportsNativeTools: true,
|
|
4404
3999
|
inputPrice: 1,
|
|
4405
4000
|
outputPrice: 4,
|
|
4406
4001
|
description: "Qwen3 Coder 480B A35B Instruct. Available in us-south1"
|
|
@@ -4410,10 +4005,18 @@ var vertexModels = {
|
|
|
4410
4005
|
contextWindow: 262144,
|
|
4411
4006
|
supportsImages: false,
|
|
4412
4007
|
supportsPromptCache: false,
|
|
4413
|
-
supportsNativeTools: true,
|
|
4414
4008
|
inputPrice: 0.25,
|
|
4415
4009
|
outputPrice: 1,
|
|
4416
4010
|
description: "Qwen3 235B A22B Instruct. Available in us-south1"
|
|
4011
|
+
},
|
|
4012
|
+
"moonshotai/kimi-k2-thinking-maas": {
|
|
4013
|
+
maxTokens: 16384,
|
|
4014
|
+
contextWindow: 262144,
|
|
4015
|
+
supportsPromptCache: false,
|
|
4016
|
+
supportsImages: false,
|
|
4017
|
+
inputPrice: 0.6,
|
|
4018
|
+
outputPrice: 2.5,
|
|
4019
|
+
description: "Kimi K2 Thinking Model with 256K context window."
|
|
4417
4020
|
}
|
|
4418
4021
|
};
|
|
4419
4022
|
var VERTEX_1M_CONTEXT_MODEL_IDS = ["claude-sonnet-4@20250514", "claude-sonnet-4-5@20250929"];
|
|
@@ -4646,8 +4249,6 @@ var xaiModels = {
|
|
|
4646
4249
|
contextWindow: 256e3,
|
|
4647
4250
|
supportsImages: true,
|
|
4648
4251
|
supportsPromptCache: true,
|
|
4649
|
-
supportsNativeTools: true,
|
|
4650
|
-
defaultToolProtocol: "native",
|
|
4651
4252
|
inputPrice: 0.2,
|
|
4652
4253
|
outputPrice: 1.5,
|
|
4653
4254
|
cacheWritesPrice: 0.02,
|
|
@@ -4661,8 +4262,6 @@ var xaiModels = {
|
|
|
4661
4262
|
contextWindow: 2e6,
|
|
4662
4263
|
supportsImages: true,
|
|
4663
4264
|
supportsPromptCache: true,
|
|
4664
|
-
supportsNativeTools: true,
|
|
4665
|
-
defaultToolProtocol: "native",
|
|
4666
4265
|
inputPrice: 0.2,
|
|
4667
4266
|
outputPrice: 0.5,
|
|
4668
4267
|
cacheWritesPrice: 0.05,
|
|
@@ -4676,8 +4275,6 @@ var xaiModels = {
|
|
|
4676
4275
|
contextWindow: 2e6,
|
|
4677
4276
|
supportsImages: true,
|
|
4678
4277
|
supportsPromptCache: true,
|
|
4679
|
-
supportsNativeTools: true,
|
|
4680
|
-
defaultToolProtocol: "native",
|
|
4681
4278
|
inputPrice: 0.2,
|
|
4682
4279
|
outputPrice: 0.5,
|
|
4683
4280
|
cacheWritesPrice: 0.05,
|
|
@@ -4691,8 +4288,6 @@ var xaiModels = {
|
|
|
4691
4288
|
contextWindow: 2e6,
|
|
4692
4289
|
supportsImages: true,
|
|
4693
4290
|
supportsPromptCache: true,
|
|
4694
|
-
supportsNativeTools: true,
|
|
4695
|
-
defaultToolProtocol: "native",
|
|
4696
4291
|
inputPrice: 0.2,
|
|
4697
4292
|
outputPrice: 0.5,
|
|
4698
4293
|
cacheWritesPrice: 0.05,
|
|
@@ -4706,8 +4301,6 @@ var xaiModels = {
|
|
|
4706
4301
|
contextWindow: 2e6,
|
|
4707
4302
|
supportsImages: true,
|
|
4708
4303
|
supportsPromptCache: true,
|
|
4709
|
-
supportsNativeTools: true,
|
|
4710
|
-
defaultToolProtocol: "native",
|
|
4711
4304
|
inputPrice: 0.2,
|
|
4712
4305
|
outputPrice: 0.5,
|
|
4713
4306
|
cacheWritesPrice: 0.05,
|
|
@@ -4721,8 +4314,6 @@ var xaiModels = {
|
|
|
4721
4314
|
contextWindow: 256e3,
|
|
4722
4315
|
supportsImages: true,
|
|
4723
4316
|
supportsPromptCache: true,
|
|
4724
|
-
supportsNativeTools: true,
|
|
4725
|
-
defaultToolProtocol: "native",
|
|
4726
4317
|
inputPrice: 3,
|
|
4727
4318
|
outputPrice: 15,
|
|
4728
4319
|
cacheWritesPrice: 0.75,
|
|
@@ -4736,8 +4327,6 @@ var xaiModels = {
|
|
|
4736
4327
|
contextWindow: 131072,
|
|
4737
4328
|
supportsImages: true,
|
|
4738
4329
|
supportsPromptCache: true,
|
|
4739
|
-
supportsNativeTools: true,
|
|
4740
|
-
defaultToolProtocol: "native",
|
|
4741
4330
|
inputPrice: 0.3,
|
|
4742
4331
|
outputPrice: 0.5,
|
|
4743
4332
|
cacheWritesPrice: 0.07,
|
|
@@ -4753,8 +4342,6 @@ var xaiModels = {
|
|
|
4753
4342
|
contextWindow: 131072,
|
|
4754
4343
|
supportsImages: true,
|
|
4755
4344
|
supportsPromptCache: true,
|
|
4756
|
-
supportsNativeTools: true,
|
|
4757
|
-
defaultToolProtocol: "native",
|
|
4758
4345
|
inputPrice: 3,
|
|
4759
4346
|
outputPrice: 15,
|
|
4760
4347
|
cacheWritesPrice: 0.75,
|
|
@@ -4851,7 +4438,6 @@ var vercelAiGatewayDefaultModelInfo = {
|
|
|
4851
4438
|
contextWindow: 2e5,
|
|
4852
4439
|
supportsImages: true,
|
|
4853
4440
|
supportsPromptCache: true,
|
|
4854
|
-
supportsNativeTools: true,
|
|
4855
4441
|
inputPrice: 3,
|
|
4856
4442
|
outputPrice: 15,
|
|
4857
4443
|
cacheWritesPrice: 3.75,
|
|
@@ -4868,8 +4454,6 @@ var internationalZAiModels = {
|
|
|
4868
4454
|
contextWindow: 131072,
|
|
4869
4455
|
supportsImages: false,
|
|
4870
4456
|
supportsPromptCache: true,
|
|
4871
|
-
supportsNativeTools: true,
|
|
4872
|
-
defaultToolProtocol: "native",
|
|
4873
4457
|
inputPrice: 0.6,
|
|
4874
4458
|
outputPrice: 2.2,
|
|
4875
4459
|
cacheWritesPrice: 0,
|
|
@@ -4881,8 +4465,6 @@ var internationalZAiModels = {
|
|
|
4881
4465
|
contextWindow: 131072,
|
|
4882
4466
|
supportsImages: false,
|
|
4883
4467
|
supportsPromptCache: true,
|
|
4884
|
-
supportsNativeTools: true,
|
|
4885
|
-
defaultToolProtocol: "native",
|
|
4886
4468
|
inputPrice: 0.2,
|
|
4887
4469
|
outputPrice: 1.1,
|
|
4888
4470
|
cacheWritesPrice: 0,
|
|
@@ -4894,8 +4476,6 @@ var internationalZAiModels = {
|
|
|
4894
4476
|
contextWindow: 131072,
|
|
4895
4477
|
supportsImages: false,
|
|
4896
4478
|
supportsPromptCache: true,
|
|
4897
|
-
supportsNativeTools: true,
|
|
4898
|
-
defaultToolProtocol: "native",
|
|
4899
4479
|
inputPrice: 2.2,
|
|
4900
4480
|
outputPrice: 8.9,
|
|
4901
4481
|
cacheWritesPrice: 0,
|
|
@@ -4907,8 +4487,6 @@ var internationalZAiModels = {
|
|
|
4907
4487
|
contextWindow: 131072,
|
|
4908
4488
|
supportsImages: false,
|
|
4909
4489
|
supportsPromptCache: true,
|
|
4910
|
-
supportsNativeTools: true,
|
|
4911
|
-
defaultToolProtocol: "native",
|
|
4912
4490
|
inputPrice: 1.1,
|
|
4913
4491
|
outputPrice: 4.5,
|
|
4914
4492
|
cacheWritesPrice: 0,
|
|
@@ -4920,8 +4498,6 @@ var internationalZAiModels = {
|
|
|
4920
4498
|
contextWindow: 131072,
|
|
4921
4499
|
supportsImages: false,
|
|
4922
4500
|
supportsPromptCache: true,
|
|
4923
|
-
supportsNativeTools: true,
|
|
4924
|
-
defaultToolProtocol: "native",
|
|
4925
4501
|
inputPrice: 0,
|
|
4926
4502
|
outputPrice: 0,
|
|
4927
4503
|
cacheWritesPrice: 0,
|
|
@@ -4933,21 +4509,28 @@ var internationalZAiModels = {
|
|
|
4933
4509
|
contextWindow: 131072,
|
|
4934
4510
|
supportsImages: true,
|
|
4935
4511
|
supportsPromptCache: true,
|
|
4936
|
-
supportsNativeTools: true,
|
|
4937
|
-
defaultToolProtocol: "native",
|
|
4938
4512
|
inputPrice: 0.6,
|
|
4939
4513
|
outputPrice: 1.8,
|
|
4940
4514
|
cacheWritesPrice: 0,
|
|
4941
4515
|
cacheReadsPrice: 0.11,
|
|
4942
4516
|
description: "GLM-4.5V is Z.AI's multimodal visual reasoning model (image/video/text/file input), optimized for GUI tasks, grounding, and document/video understanding."
|
|
4943
4517
|
},
|
|
4518
|
+
"glm-4.6v": {
|
|
4519
|
+
maxTokens: 16384,
|
|
4520
|
+
contextWindow: 131072,
|
|
4521
|
+
supportsImages: true,
|
|
4522
|
+
supportsPromptCache: true,
|
|
4523
|
+
inputPrice: 0.3,
|
|
4524
|
+
outputPrice: 0.9,
|
|
4525
|
+
cacheWritesPrice: 0,
|
|
4526
|
+
cacheReadsPrice: 0.05,
|
|
4527
|
+
description: "GLM-4.6V is an advanced multimodal vision model with improved performance and cost-efficiency for visual understanding tasks."
|
|
4528
|
+
},
|
|
4944
4529
|
"glm-4.6": {
|
|
4945
4530
|
maxTokens: 16384,
|
|
4946
4531
|
contextWindow: 2e5,
|
|
4947
4532
|
supportsImages: false,
|
|
4948
4533
|
supportsPromptCache: true,
|
|
4949
|
-
supportsNativeTools: true,
|
|
4950
|
-
defaultToolProtocol: "native",
|
|
4951
4534
|
inputPrice: 0.6,
|
|
4952
4535
|
outputPrice: 2.2,
|
|
4953
4536
|
cacheWritesPrice: 0,
|
|
@@ -4959,8 +4542,6 @@ var internationalZAiModels = {
|
|
|
4959
4542
|
contextWindow: 2e5,
|
|
4960
4543
|
supportsImages: false,
|
|
4961
4544
|
supportsPromptCache: true,
|
|
4962
|
-
supportsNativeTools: true,
|
|
4963
|
-
defaultToolProtocol: "native",
|
|
4964
4545
|
supportsReasoningEffort: ["disable", "medium"],
|
|
4965
4546
|
reasoningEffort: "medium",
|
|
4966
4547
|
preserveReasoning: true,
|
|
@@ -4970,13 +4551,55 @@ var internationalZAiModels = {
|
|
|
4970
4551
|
cacheReadsPrice: 0.11,
|
|
4971
4552
|
description: "GLM-4.7 is Zhipu's latest model with built-in thinking capabilities enabled by default. It provides enhanced reasoning for complex tasks while maintaining fast response times."
|
|
4972
4553
|
},
|
|
4554
|
+
"glm-4.7-flash": {
|
|
4555
|
+
maxTokens: 16384,
|
|
4556
|
+
contextWindow: 2e5,
|
|
4557
|
+
supportsImages: false,
|
|
4558
|
+
supportsPromptCache: true,
|
|
4559
|
+
inputPrice: 0,
|
|
4560
|
+
outputPrice: 0,
|
|
4561
|
+
cacheWritesPrice: 0,
|
|
4562
|
+
cacheReadsPrice: 0,
|
|
4563
|
+
description: "GLM-4.7-Flash is a free, high-speed variant of GLM-4.7 offering fast responses for reasoning and coding tasks."
|
|
4564
|
+
},
|
|
4565
|
+
"glm-4.7-flashx": {
|
|
4566
|
+
maxTokens: 16384,
|
|
4567
|
+
contextWindow: 2e5,
|
|
4568
|
+
supportsImages: false,
|
|
4569
|
+
supportsPromptCache: true,
|
|
4570
|
+
inputPrice: 0.07,
|
|
4571
|
+
outputPrice: 0.4,
|
|
4572
|
+
cacheWritesPrice: 0,
|
|
4573
|
+
cacheReadsPrice: 0.01,
|
|
4574
|
+
description: "GLM-4.7-FlashX is an ultra-fast variant of GLM-4.7 with exceptional speed and cost-effectiveness for high-throughput applications."
|
|
4575
|
+
},
|
|
4576
|
+
"glm-4.6v-flash": {
|
|
4577
|
+
maxTokens: 16384,
|
|
4578
|
+
contextWindow: 131072,
|
|
4579
|
+
supportsImages: true,
|
|
4580
|
+
supportsPromptCache: true,
|
|
4581
|
+
inputPrice: 0,
|
|
4582
|
+
outputPrice: 0,
|
|
4583
|
+
cacheWritesPrice: 0,
|
|
4584
|
+
cacheReadsPrice: 0,
|
|
4585
|
+
description: "GLM-4.6V-Flash is a free, high-speed multimodal vision model for rapid image understanding and visual reasoning tasks."
|
|
4586
|
+
},
|
|
4587
|
+
"glm-4.6v-flashx": {
|
|
4588
|
+
maxTokens: 16384,
|
|
4589
|
+
contextWindow: 131072,
|
|
4590
|
+
supportsImages: true,
|
|
4591
|
+
supportsPromptCache: true,
|
|
4592
|
+
inputPrice: 0.04,
|
|
4593
|
+
outputPrice: 0.4,
|
|
4594
|
+
cacheWritesPrice: 0,
|
|
4595
|
+
cacheReadsPrice: 4e-3,
|
|
4596
|
+
description: "GLM-4.6V-FlashX is an ultra-fast multimodal vision model optimized for high-speed visual processing at low cost."
|
|
4597
|
+
},
|
|
4973
4598
|
"glm-4-32b-0414-128k": {
|
|
4974
4599
|
maxTokens: 16384,
|
|
4975
4600
|
contextWindow: 131072,
|
|
4976
4601
|
supportsImages: false,
|
|
4977
4602
|
supportsPromptCache: false,
|
|
4978
|
-
supportsNativeTools: true,
|
|
4979
|
-
defaultToolProtocol: "native",
|
|
4980
4603
|
inputPrice: 0.1,
|
|
4981
4604
|
outputPrice: 0.1,
|
|
4982
4605
|
cacheWritesPrice: 0,
|
|
@@ -4991,8 +4614,6 @@ var mainlandZAiModels = {
|
|
|
4991
4614
|
contextWindow: 131072,
|
|
4992
4615
|
supportsImages: false,
|
|
4993
4616
|
supportsPromptCache: true,
|
|
4994
|
-
supportsNativeTools: true,
|
|
4995
|
-
defaultToolProtocol: "native",
|
|
4996
4617
|
inputPrice: 0.29,
|
|
4997
4618
|
outputPrice: 1.14,
|
|
4998
4619
|
cacheWritesPrice: 0,
|
|
@@ -5004,8 +4625,6 @@ var mainlandZAiModels = {
|
|
|
5004
4625
|
contextWindow: 131072,
|
|
5005
4626
|
supportsImages: false,
|
|
5006
4627
|
supportsPromptCache: true,
|
|
5007
|
-
supportsNativeTools: true,
|
|
5008
|
-
defaultToolProtocol: "native",
|
|
5009
4628
|
inputPrice: 0.1,
|
|
5010
4629
|
outputPrice: 0.6,
|
|
5011
4630
|
cacheWritesPrice: 0,
|
|
@@ -5017,8 +4636,6 @@ var mainlandZAiModels = {
|
|
|
5017
4636
|
contextWindow: 131072,
|
|
5018
4637
|
supportsImages: false,
|
|
5019
4638
|
supportsPromptCache: true,
|
|
5020
|
-
supportsNativeTools: true,
|
|
5021
|
-
defaultToolProtocol: "native",
|
|
5022
4639
|
inputPrice: 0.29,
|
|
5023
4640
|
outputPrice: 1.14,
|
|
5024
4641
|
cacheWritesPrice: 0,
|
|
@@ -5030,8 +4647,6 @@ var mainlandZAiModels = {
|
|
|
5030
4647
|
contextWindow: 131072,
|
|
5031
4648
|
supportsImages: false,
|
|
5032
4649
|
supportsPromptCache: true,
|
|
5033
|
-
supportsNativeTools: true,
|
|
5034
|
-
defaultToolProtocol: "native",
|
|
5035
4650
|
inputPrice: 0.1,
|
|
5036
4651
|
outputPrice: 0.6,
|
|
5037
4652
|
cacheWritesPrice: 0,
|
|
@@ -5043,8 +4658,6 @@ var mainlandZAiModels = {
|
|
|
5043
4658
|
contextWindow: 131072,
|
|
5044
4659
|
supportsImages: false,
|
|
5045
4660
|
supportsPromptCache: true,
|
|
5046
|
-
supportsNativeTools: true,
|
|
5047
|
-
defaultToolProtocol: "native",
|
|
5048
4661
|
inputPrice: 0,
|
|
5049
4662
|
outputPrice: 0,
|
|
5050
4663
|
cacheWritesPrice: 0,
|
|
@@ -5056,8 +4669,6 @@ var mainlandZAiModels = {
|
|
|
5056
4669
|
contextWindow: 131072,
|
|
5057
4670
|
supportsImages: true,
|
|
5058
4671
|
supportsPromptCache: true,
|
|
5059
|
-
supportsNativeTools: true,
|
|
5060
|
-
defaultToolProtocol: "native",
|
|
5061
4672
|
inputPrice: 0.29,
|
|
5062
4673
|
outputPrice: 0.93,
|
|
5063
4674
|
cacheWritesPrice: 0,
|
|
@@ -5069,8 +4680,6 @@ var mainlandZAiModels = {
|
|
|
5069
4680
|
contextWindow: 204800,
|
|
5070
4681
|
supportsImages: false,
|
|
5071
4682
|
supportsPromptCache: true,
|
|
5072
|
-
supportsNativeTools: true,
|
|
5073
|
-
defaultToolProtocol: "native",
|
|
5074
4683
|
inputPrice: 0.29,
|
|
5075
4684
|
outputPrice: 1.14,
|
|
5076
4685
|
cacheWritesPrice: 0,
|
|
@@ -5082,8 +4691,6 @@ var mainlandZAiModels = {
|
|
|
5082
4691
|
contextWindow: 204800,
|
|
5083
4692
|
supportsImages: false,
|
|
5084
4693
|
supportsPromptCache: true,
|
|
5085
|
-
supportsNativeTools: true,
|
|
5086
|
-
defaultToolProtocol: "native",
|
|
5087
4694
|
supportsReasoningEffort: ["disable", "medium"],
|
|
5088
4695
|
reasoningEffort: "medium",
|
|
5089
4696
|
preserveReasoning: true,
|
|
@@ -5092,6 +4699,61 @@ var mainlandZAiModels = {
|
|
|
5092
4699
|
cacheWritesPrice: 0,
|
|
5093
4700
|
cacheReadsPrice: 0.057,
|
|
5094
4701
|
description: "GLM-4.7 is Zhipu's latest model with built-in thinking capabilities enabled by default. It provides enhanced reasoning for complex tasks while maintaining fast response times."
|
|
4702
|
+
},
|
|
4703
|
+
"glm-4.7-flash": {
|
|
4704
|
+
maxTokens: 16384,
|
|
4705
|
+
contextWindow: 204800,
|
|
4706
|
+
supportsImages: false,
|
|
4707
|
+
supportsPromptCache: true,
|
|
4708
|
+
inputPrice: 0,
|
|
4709
|
+
outputPrice: 0,
|
|
4710
|
+
cacheWritesPrice: 0,
|
|
4711
|
+
cacheReadsPrice: 0,
|
|
4712
|
+
description: "GLM-4.7-Flash is a free, high-speed variant of GLM-4.7 offering fast responses for reasoning and coding tasks."
|
|
4713
|
+
},
|
|
4714
|
+
"glm-4.7-flashx": {
|
|
4715
|
+
maxTokens: 16384,
|
|
4716
|
+
contextWindow: 204800,
|
|
4717
|
+
supportsImages: false,
|
|
4718
|
+
supportsPromptCache: true,
|
|
4719
|
+
inputPrice: 0.035,
|
|
4720
|
+
outputPrice: 0.2,
|
|
4721
|
+
cacheWritesPrice: 0,
|
|
4722
|
+
cacheReadsPrice: 5e-3,
|
|
4723
|
+
description: "GLM-4.7-FlashX is an ultra-fast variant of GLM-4.7 with exceptional speed and cost-effectiveness for high-throughput applications."
|
|
4724
|
+
},
|
|
4725
|
+
"glm-4.6v": {
|
|
4726
|
+
maxTokens: 16384,
|
|
4727
|
+
contextWindow: 131072,
|
|
4728
|
+
supportsImages: true,
|
|
4729
|
+
supportsPromptCache: true,
|
|
4730
|
+
inputPrice: 0.15,
|
|
4731
|
+
outputPrice: 0.45,
|
|
4732
|
+
cacheWritesPrice: 0,
|
|
4733
|
+
cacheReadsPrice: 0.025,
|
|
4734
|
+
description: "GLM-4.6V is an advanced multimodal vision model with improved performance and cost-efficiency for visual understanding tasks."
|
|
4735
|
+
},
|
|
4736
|
+
"glm-4.6v-flash": {
|
|
4737
|
+
maxTokens: 16384,
|
|
4738
|
+
contextWindow: 131072,
|
|
4739
|
+
supportsImages: true,
|
|
4740
|
+
supportsPromptCache: true,
|
|
4741
|
+
inputPrice: 0,
|
|
4742
|
+
outputPrice: 0,
|
|
4743
|
+
cacheWritesPrice: 0,
|
|
4744
|
+
cacheReadsPrice: 0,
|
|
4745
|
+
description: "GLM-4.6V-Flash is a free, high-speed multimodal vision model for rapid image understanding and visual reasoning tasks."
|
|
4746
|
+
},
|
|
4747
|
+
"glm-4.6v-flashx": {
|
|
4748
|
+
maxTokens: 16384,
|
|
4749
|
+
contextWindow: 131072,
|
|
4750
|
+
supportsImages: true,
|
|
4751
|
+
supportsPromptCache: true,
|
|
4752
|
+
inputPrice: 0.02,
|
|
4753
|
+
outputPrice: 0.2,
|
|
4754
|
+
cacheWritesPrice: 0,
|
|
4755
|
+
cacheReadsPrice: 2e-3,
|
|
4756
|
+
description: "GLM-4.6V-FlashX is an ultra-fast multimodal vision model optimized for high-speed visual processing at low cost."
|
|
5095
4757
|
}
|
|
5096
4758
|
};
|
|
5097
4759
|
var ZAI_DEFAULT_TEMPERATURE = 0.6;
|
|
@@ -5125,7 +4787,6 @@ var deepInfraDefaultModelInfo = {
|
|
|
5125
4787
|
contextWindow: 262144,
|
|
5126
4788
|
supportsImages: false,
|
|
5127
4789
|
supportsPromptCache: false,
|
|
5128
|
-
supportsNativeTools: true,
|
|
5129
4790
|
inputPrice: 0.3,
|
|
5130
4791
|
outputPrice: 1.2,
|
|
5131
4792
|
description: "Qwen 3 Coder 480B A35B Instruct Turbo model, 256K context."
|
|
@@ -5139,8 +4800,6 @@ var minimaxModels = {
|
|
|
5139
4800
|
contextWindow: 192e3,
|
|
5140
4801
|
supportsImages: false,
|
|
5141
4802
|
supportsPromptCache: true,
|
|
5142
|
-
supportsNativeTools: true,
|
|
5143
|
-
defaultToolProtocol: "native",
|
|
5144
4803
|
includedTools: ["search_and_replace"],
|
|
5145
4804
|
excludedTools: ["apply_diff"],
|
|
5146
4805
|
preserveReasoning: true,
|
|
@@ -5155,8 +4814,6 @@ var minimaxModels = {
|
|
|
5155
4814
|
contextWindow: 192e3,
|
|
5156
4815
|
supportsImages: false,
|
|
5157
4816
|
supportsPromptCache: true,
|
|
5158
|
-
supportsNativeTools: true,
|
|
5159
|
-
defaultToolProtocol: "native",
|
|
5160
4817
|
includedTools: ["search_and_replace"],
|
|
5161
4818
|
excludedTools: ["apply_diff"],
|
|
5162
4819
|
preserveReasoning: true,
|
|
@@ -5171,8 +4828,6 @@ var minimaxModels = {
|
|
|
5171
4828
|
contextWindow: 192e3,
|
|
5172
4829
|
supportsImages: false,
|
|
5173
4830
|
supportsPromptCache: true,
|
|
5174
|
-
supportsNativeTools: true,
|
|
5175
|
-
defaultToolProtocol: "native",
|
|
5176
4831
|
includedTools: ["search_and_replace"],
|
|
5177
4832
|
excludedTools: ["apply_diff"],
|
|
5178
4833
|
preserveReasoning: true,
|
|
@@ -5244,8 +4899,6 @@ function getProviderDefaultModelId(provider, options = { isChina: false }) {
|
|
|
5244
4899
|
return deepInfraDefaultModelId;
|
|
5245
4900
|
case "vscode-lm":
|
|
5246
4901
|
return vscodeLlmDefaultModelId;
|
|
5247
|
-
case "claude-code":
|
|
5248
|
-
return claudeCodeDefaultModelId;
|
|
5249
4902
|
case "cerebras":
|
|
5250
4903
|
return cerebrasDefaultModelId;
|
|
5251
4904
|
case "sambanova":
|
|
@@ -5303,7 +4956,6 @@ var providerNames = [
|
|
|
5303
4956
|
"bedrock",
|
|
5304
4957
|
"baseten",
|
|
5305
4958
|
"cerebras",
|
|
5306
|
-
"claude-code",
|
|
5307
4959
|
"doubao",
|
|
5308
4960
|
"deepseek",
|
|
5309
4961
|
"featherless",
|
|
@@ -5333,9 +4985,7 @@ var providerSettingsEntrySchema = z8.object({
|
|
|
5333
4985
|
});
|
|
5334
4986
|
var baseProviderSettingsSchema = z8.object({
|
|
5335
4987
|
includeMaxTokens: z8.boolean().optional(),
|
|
5336
|
-
diffEnabled: z8.boolean().optional(),
|
|
5337
4988
|
todoListEnabled: z8.boolean().optional(),
|
|
5338
|
-
fuzzyMatchThreshold: z8.number().optional(),
|
|
5339
4989
|
modelTemperature: z8.number().nullish(),
|
|
5340
4990
|
rateLimitSeconds: z8.number().optional(),
|
|
5341
4991
|
consecutiveMistakeLimit: z8.number().min(0).optional(),
|
|
@@ -5345,9 +4995,7 @@ var baseProviderSettingsSchema = z8.object({
|
|
|
5345
4995
|
modelMaxTokens: z8.number().optional(),
|
|
5346
4996
|
modelMaxThinkingTokens: z8.number().optional(),
|
|
5347
4997
|
// Model verbosity.
|
|
5348
|
-
verbosity: verbosityLevelsSchema.optional()
|
|
5349
|
-
// Tool protocol override for this profile.
|
|
5350
|
-
toolProtocol: z8.enum(["xml", "native"]).optional()
|
|
4998
|
+
verbosity: verbosityLevelsSchema.optional()
|
|
5351
4999
|
});
|
|
5352
5000
|
var apiModelIdProviderModelSchema = baseProviderSettingsSchema.extend({
|
|
5353
5001
|
apiModelId: z8.string().optional()
|
|
@@ -5359,7 +5007,6 @@ var anthropicSchema = apiModelIdProviderModelSchema.extend({
|
|
|
5359
5007
|
anthropicBeta1MContext: z8.boolean().optional()
|
|
5360
5008
|
// Enable 'context-1m-2025-08-07' beta for 1M context window.
|
|
5361
5009
|
});
|
|
5362
|
-
var claudeCodeSchema = apiModelIdProviderModelSchema.extend({});
|
|
5363
5010
|
var openRouterSchema = baseProviderSettingsSchema.extend({
|
|
5364
5011
|
openRouterApiKey: z8.string().optional(),
|
|
5365
5012
|
openRouterModelId: z8.string().optional(),
|
|
@@ -5548,7 +5195,6 @@ var defaultSchema = z8.object({
|
|
|
5548
5195
|
});
|
|
5549
5196
|
var providerSettingsSchemaDiscriminated = z8.discriminatedUnion("apiProvider", [
|
|
5550
5197
|
anthropicSchema.merge(z8.object({ apiProvider: z8.literal("anthropic") })),
|
|
5551
|
-
claudeCodeSchema.merge(z8.object({ apiProvider: z8.literal("claude-code") })),
|
|
5552
5198
|
openRouterSchema.merge(z8.object({ apiProvider: z8.literal("openrouter") })),
|
|
5553
5199
|
bedrockSchema.merge(z8.object({ apiProvider: z8.literal("bedrock") })),
|
|
5554
5200
|
vertexSchema.merge(z8.object({ apiProvider: z8.literal("vertex") })),
|
|
@@ -5589,7 +5235,6 @@ var providerSettingsSchemaDiscriminated = z8.discriminatedUnion("apiProvider", [
|
|
|
5589
5235
|
var providerSettingsSchema = z8.object({
|
|
5590
5236
|
apiProvider: providerNamesSchema.optional(),
|
|
5591
5237
|
...anthropicSchema.shape,
|
|
5592
|
-
...claudeCodeSchema.shape,
|
|
5593
5238
|
...openRouterSchema.shape,
|
|
5594
5239
|
...bedrockSchema.shape,
|
|
5595
5240
|
...vertexSchema.shape,
|
|
@@ -5654,7 +5299,6 @@ var getModelId = (settings) => {
|
|
|
5654
5299
|
var isTypicalProvider = (key) => isProviderName(key) && !isInternalProvider(key) && !isCustomProvider(key) && !isFauxProvider(key);
|
|
5655
5300
|
var modelIdKeysByProvider = {
|
|
5656
5301
|
anthropic: "apiModelId",
|
|
5657
|
-
"claude-code": "apiModelId",
|
|
5658
5302
|
openrouter: "openRouterModelId",
|
|
5659
5303
|
bedrock: "apiModelId",
|
|
5660
5304
|
vertex: "apiModelId",
|
|
@@ -5688,7 +5332,7 @@ var modelIdKeysByProvider = {
|
|
|
5688
5332
|
roo: "apiModelId",
|
|
5689
5333
|
"vercel-ai-gateway": "vercelAiGatewayModelId"
|
|
5690
5334
|
};
|
|
5691
|
-
var ANTHROPIC_STYLE_PROVIDERS = ["anthropic", "
|
|
5335
|
+
var ANTHROPIC_STYLE_PROVIDERS = ["anthropic", "bedrock", "minimax"];
|
|
5692
5336
|
var getApiProtocol = (provider, modelId) => {
|
|
5693
5337
|
if (provider && ANTHROPIC_STYLE_PROVIDERS.includes(provider)) {
|
|
5694
5338
|
return "anthropic";
|
|
@@ -5717,7 +5361,6 @@ var MODELS_BY_PROVIDER = {
|
|
|
5717
5361
|
label: "Cerebras",
|
|
5718
5362
|
models: Object.keys(cerebrasModels)
|
|
5719
5363
|
},
|
|
5720
|
-
"claude-code": { id: "claude-code", label: "Claude Code", models: Object.keys(claudeCodeModels) },
|
|
5721
5364
|
deepseek: {
|
|
5722
5365
|
id: "deepseek",
|
|
5723
5366
|
label: "DeepSeek",
|
|
@@ -5821,16 +5464,6 @@ var historyItemSchema = z9.object({
|
|
|
5821
5464
|
size: z9.number().optional(),
|
|
5822
5465
|
workspace: z9.string().optional(),
|
|
5823
5466
|
mode: z9.string().optional(),
|
|
5824
|
-
/**
|
|
5825
|
-
* The tool protocol used by this task. Once a task uses tools with a specific
|
|
5826
|
-
* protocol (XML or Native), it is permanently locked to that protocol.
|
|
5827
|
-
*
|
|
5828
|
-
* - "xml": Tool calls are parsed from XML text (no tool IDs)
|
|
5829
|
-
* - "native": Tool calls come as tool_call chunks with IDs
|
|
5830
|
-
*
|
|
5831
|
-
* This ensures task resumption works correctly even when NTC settings change.
|
|
5832
|
-
*/
|
|
5833
|
-
toolProtocol: z9.enum(["xml", "native"]).optional(),
|
|
5834
5467
|
apiConfigName: z9.string().optional(),
|
|
5835
5468
|
// Provider profile name for sticky profile feature
|
|
5836
5469
|
status: z9.enum(["active", "completed", "delegated"]).optional(),
|
|
@@ -5848,23 +5481,12 @@ var historyItemSchema = z9.object({
|
|
|
5848
5481
|
|
|
5849
5482
|
// src/experiment.ts
|
|
5850
5483
|
import { z as z10 } from "zod";
|
|
5851
|
-
var experimentIds = [
|
|
5852
|
-
"powerSteering",
|
|
5853
|
-
"multiFileApplyDiff",
|
|
5854
|
-
"preventFocusDisruption",
|
|
5855
|
-
"imageGeneration",
|
|
5856
|
-
"runSlashCommand",
|
|
5857
|
-
"multipleNativeToolCalls",
|
|
5858
|
-
"customTools"
|
|
5859
|
-
];
|
|
5484
|
+
var experimentIds = ["preventFocusDisruption", "imageGeneration", "runSlashCommand", "customTools"];
|
|
5860
5485
|
var experimentIdsSchema = z10.enum(experimentIds);
|
|
5861
5486
|
var experimentsSchema = z10.object({
|
|
5862
|
-
powerSteering: z10.boolean().optional(),
|
|
5863
|
-
multiFileApplyDiff: z10.boolean().optional(),
|
|
5864
5487
|
preventFocusDisruption: z10.boolean().optional(),
|
|
5865
5488
|
imageGeneration: z10.boolean().optional(),
|
|
5866
5489
|
runSlashCommand: z10.boolean().optional(),
|
|
5867
|
-
multipleNativeToolCalls: z10.boolean().optional(),
|
|
5868
5490
|
customTools: z10.boolean().optional()
|
|
5869
5491
|
});
|
|
5870
5492
|
|
|
@@ -5916,6 +5538,7 @@ var TelemetryEventName = /* @__PURE__ */ ((TelemetryEventName2) => {
|
|
|
5916
5538
|
TelemetryEventName2["CODE_INDEX_ERROR"] = "Code Index Error";
|
|
5917
5539
|
TelemetryEventName2["TELEMETRY_SETTINGS_CHANGED"] = "Telemetry Settings Changed";
|
|
5918
5540
|
TelemetryEventName2["MODEL_CACHE_EMPTY_RESPONSE"] = "Model Cache Empty Response";
|
|
5541
|
+
TelemetryEventName2["READ_FILE_LEGACY_FORMAT_USED"] = "Read File Legacy Format Used";
|
|
5919
5542
|
return TelemetryEventName2;
|
|
5920
5543
|
})(TelemetryEventName || {});
|
|
5921
5544
|
var staticAppPropertiesSchema = z11.object({
|
|
@@ -6004,7 +5627,8 @@ var rooCodeTelemetryEventSchema = z11.discriminatedUnion("type", [
|
|
|
6004
5627
|
"Sliding Window Truncation" /* SLIDING_WINDOW_TRUNCATION */,
|
|
6005
5628
|
"Tab Shown" /* TAB_SHOWN */,
|
|
6006
5629
|
"Mode Setting Changed" /* MODE_SETTINGS_CHANGED */,
|
|
6007
|
-
"Custom Mode Created" /* CUSTOM_MODE_CREATED
|
|
5630
|
+
"Custom Mode Created" /* CUSTOM_MODE_CREATED */,
|
|
5631
|
+
"Read File Legacy Format Used" /* READ_FILE_LEGACY_FORMAT_USED */
|
|
6008
5632
|
]),
|
|
6009
5633
|
properties: telemetryPropertiesSchema
|
|
6010
5634
|
}),
|
|
@@ -6318,7 +5942,15 @@ var isLanguage = (value) => languages.includes(value);
|
|
|
6318
5942
|
|
|
6319
5943
|
// src/global-settings.ts
|
|
6320
5944
|
var DEFAULT_WRITE_DELAY_MS = 1e3;
|
|
6321
|
-
var
|
|
5945
|
+
var TERMINAL_PREVIEW_BYTES = {
|
|
5946
|
+
small: 5 * 1024,
|
|
5947
|
+
// 5KB
|
|
5948
|
+
medium: 10 * 1024,
|
|
5949
|
+
// 10KB
|
|
5950
|
+
large: 20 * 1024
|
|
5951
|
+
// 20KB
|
|
5952
|
+
};
|
|
5953
|
+
var DEFAULT_TERMINAL_OUTPUT_PREVIEW_SIZE = "medium";
|
|
6322
5954
|
var MIN_CHECKPOINT_TIMEOUT_SECONDS = 10;
|
|
6323
5955
|
var MAX_CHECKPOINT_TIMEOUT_SECONDS = 60;
|
|
6324
5956
|
var DEFAULT_CHECKPOINT_TIMEOUT_SECONDS = 15;
|
|
@@ -6334,7 +5966,6 @@ var globalSettingsSchema = z14.object({
|
|
|
6334
5966
|
imageGenerationProvider: z14.enum(["openrouter", "roo"]).optional(),
|
|
6335
5967
|
openRouterImageApiKey: z14.string().optional(),
|
|
6336
5968
|
openRouterImageGenerationSelectedModel: z14.string().optional(),
|
|
6337
|
-
condensingApiConfigId: z14.string().optional(),
|
|
6338
5969
|
customCondensingPrompt: z14.string().optional(),
|
|
6339
5970
|
autoApprovalEnabled: z14.boolean().optional(),
|
|
6340
5971
|
alwaysAllowReadOnly: z14.boolean().optional(),
|
|
@@ -6360,7 +5991,6 @@ var globalSettingsSchema = z14.object({
|
|
|
6360
5991
|
allowedMaxCost: z14.number().nullish(),
|
|
6361
5992
|
autoCondenseContext: z14.boolean().optional(),
|
|
6362
5993
|
autoCondenseContextPercent: z14.number().optional(),
|
|
6363
|
-
maxConcurrentFileReads: z14.number().optional(),
|
|
6364
5994
|
/**
|
|
6365
5995
|
* Whether to include current time in the environment details
|
|
6366
5996
|
* @default true
|
|
@@ -6403,11 +6033,9 @@ var globalSettingsSchema = z14.object({
|
|
|
6403
6033
|
maxWorkspaceFiles: z14.number().optional(),
|
|
6404
6034
|
showRooIgnoredFiles: z14.boolean().optional(),
|
|
6405
6035
|
enableSubfolderRules: z14.boolean().optional(),
|
|
6406
|
-
maxReadFileLine: z14.number().optional(),
|
|
6407
6036
|
maxImageFileSize: z14.number().optional(),
|
|
6408
6037
|
maxTotalImageSize: z14.number().optional(),
|
|
6409
|
-
|
|
6410
|
-
terminalOutputCharacterLimit: z14.number().optional(),
|
|
6038
|
+
terminalOutputPreviewSize: z14.enum(["small", "medium", "large"]).optional(),
|
|
6411
6039
|
terminalShellIntegrationTimeout: z14.number().optional(),
|
|
6412
6040
|
terminalShellIntegrationDisabled: z14.boolean().optional(),
|
|
6413
6041
|
terminalCommandDelay: z14.number().optional(),
|
|
@@ -6416,18 +6044,14 @@ var globalSettingsSchema = z14.object({
|
|
|
6416
6044
|
terminalZshOhMy: z14.boolean().optional(),
|
|
6417
6045
|
terminalZshP10k: z14.boolean().optional(),
|
|
6418
6046
|
terminalZdotdir: z14.boolean().optional(),
|
|
6419
|
-
terminalCompressProgressBar: z14.boolean().optional(),
|
|
6420
6047
|
diagnosticsEnabled: z14.boolean().optional(),
|
|
6421
6048
|
rateLimitSeconds: z14.number().optional(),
|
|
6422
|
-
diffEnabled: z14.boolean().optional(),
|
|
6423
|
-
fuzzyMatchThreshold: z14.number().optional(),
|
|
6424
6049
|
experiments: experimentsSchema.optional(),
|
|
6425
6050
|
codebaseIndexModels: codebaseIndexModelsSchema.optional(),
|
|
6426
6051
|
codebaseIndexConfig: codebaseIndexConfigSchema.optional(),
|
|
6427
6052
|
language: languagesSchema.optional(),
|
|
6428
6053
|
telemetrySetting: telemetrySettingsSchema.optional(),
|
|
6429
6054
|
mcpEnabled: z14.boolean().optional(),
|
|
6430
|
-
enableMcpServerCreation: z14.boolean().optional(),
|
|
6431
6055
|
mode: z14.string().optional(),
|
|
6432
6056
|
modeApiConfigs: z14.record(z14.string(), z14.string()).optional(),
|
|
6433
6057
|
customModes: z14.array(modeConfigSchema).optional(),
|
|
@@ -6447,7 +6071,20 @@ var globalSettingsSchema = z14.object({
|
|
|
6447
6071
|
profileThresholds: z14.record(z14.string(), z14.number()).optional(),
|
|
6448
6072
|
hasOpenedModeSelector: z14.boolean().optional(),
|
|
6449
6073
|
lastModeExportPath: z14.string().optional(),
|
|
6450
|
-
lastModeImportPath: z14.string().optional()
|
|
6074
|
+
lastModeImportPath: z14.string().optional(),
|
|
6075
|
+
lastSettingsExportPath: z14.string().optional(),
|
|
6076
|
+
lastTaskExportPath: z14.string().optional(),
|
|
6077
|
+
lastImageSavePath: z14.string().optional(),
|
|
6078
|
+
/**
|
|
6079
|
+
* Path to worktree to auto-open after switching workspaces.
|
|
6080
|
+
* Used by the worktree feature to open the Roo Code sidebar in a new window.
|
|
6081
|
+
*/
|
|
6082
|
+
worktreeAutoOpenPath: z14.string().optional(),
|
|
6083
|
+
/**
|
|
6084
|
+
* Whether to show the worktree selector in the home screen.
|
|
6085
|
+
* @default true
|
|
6086
|
+
*/
|
|
6087
|
+
showWorktreesInHomeScreen: z14.boolean().optional()
|
|
6451
6088
|
});
|
|
6452
6089
|
var GLOBAL_SETTINGS_KEYS = globalSettingsSchema.keyof().options;
|
|
6453
6090
|
var rooCodeSettingsSchema = providerSettingsSchema.merge(globalSettingsSchema);
|
|
@@ -6531,8 +6168,6 @@ var EVALS_SETTINGS = {
|
|
|
6531
6168
|
ttsSpeed: 1,
|
|
6532
6169
|
soundEnabled: false,
|
|
6533
6170
|
soundVolume: 0.5,
|
|
6534
|
-
terminalOutputLineLimit: 500,
|
|
6535
|
-
terminalOutputCharacterLimit: DEFAULT_TERMINAL_OUTPUT_CHARACTER_LIMIT,
|
|
6536
6171
|
terminalShellIntegrationTimeout: 3e4,
|
|
6537
6172
|
terminalCommandDelay: 0,
|
|
6538
6173
|
terminalPowershellCounter: false,
|
|
@@ -6540,19 +6175,14 @@ var EVALS_SETTINGS = {
|
|
|
6540
6175
|
terminalZshClearEolMark: true,
|
|
6541
6176
|
terminalZshP10k: false,
|
|
6542
6177
|
terminalZdotdir: true,
|
|
6543
|
-
terminalCompressProgressBar: true,
|
|
6544
6178
|
terminalShellIntegrationDisabled: true,
|
|
6545
6179
|
diagnosticsEnabled: true,
|
|
6546
|
-
diffEnabled: true,
|
|
6547
|
-
fuzzyMatchThreshold: 1,
|
|
6548
6180
|
enableCheckpoints: false,
|
|
6549
6181
|
rateLimitSeconds: 0,
|
|
6550
6182
|
maxOpenTabsContext: 20,
|
|
6551
6183
|
maxWorkspaceFiles: 200,
|
|
6552
6184
|
maxGitStatusFiles: 20,
|
|
6553
6185
|
showRooIgnoredFiles: true,
|
|
6554
|
-
maxReadFileLine: -1,
|
|
6555
|
-
// -1 to enable full file reading.
|
|
6556
6186
|
includeDiagnosticMessages: true,
|
|
6557
6187
|
maxDiagnosticMessages: 50,
|
|
6558
6188
|
language: "en",
|
|
@@ -6626,24 +6256,18 @@ var organizationAllowListSchema = z16.object({
|
|
|
6626
6256
|
});
|
|
6627
6257
|
var organizationDefaultSettingsSchema = globalSettingsSchema.pick({
|
|
6628
6258
|
enableCheckpoints: true,
|
|
6629
|
-
fuzzyMatchThreshold: true,
|
|
6630
6259
|
maxOpenTabsContext: true,
|
|
6631
|
-
maxReadFileLine: true,
|
|
6632
6260
|
maxWorkspaceFiles: true,
|
|
6633
6261
|
showRooIgnoredFiles: true,
|
|
6634
6262
|
terminalCommandDelay: true,
|
|
6635
|
-
terminalCompressProgressBar: true,
|
|
6636
|
-
terminalOutputLineLimit: true,
|
|
6637
6263
|
terminalShellIntegrationDisabled: true,
|
|
6638
6264
|
terminalShellIntegrationTimeout: true,
|
|
6639
6265
|
terminalZshClearEolMark: true
|
|
6640
6266
|
}).merge(
|
|
6641
6267
|
z16.object({
|
|
6642
6268
|
maxOpenTabsContext: z16.number().int().nonnegative().optional(),
|
|
6643
|
-
maxReadFileLine: z16.number().int().gte(-1).optional(),
|
|
6644
6269
|
maxWorkspaceFiles: z16.number().int().nonnegative().optional(),
|
|
6645
6270
|
terminalCommandDelay: z16.number().int().nonnegative().optional(),
|
|
6646
|
-
terminalOutputLineLimit: z16.number().int().nonnegative().optional(),
|
|
6647
6271
|
terminalShellIntegrationTimeout: z16.number().int().nonnegative().optional()
|
|
6648
6272
|
})
|
|
6649
6273
|
);
|
|
@@ -7159,6 +6783,7 @@ var ipcMessageSchema = z19.discriminatedUnion("type", [
|
|
|
7159
6783
|
|
|
7160
6784
|
// src/mcp.ts
|
|
7161
6785
|
import { z as z20 } from "zod";
|
|
6786
|
+
var MAX_MCP_TOOLS_THRESHOLD = 60;
|
|
7162
6787
|
var mcpExecutionStatusSchema = z20.discriminatedUnion("status", [
|
|
7163
6788
|
z20.object({
|
|
7164
6789
|
executionId: z20.string(),
|
|
@@ -7182,6 +6807,46 @@ var mcpExecutionStatusSchema = z20.discriminatedUnion("status", [
|
|
|
7182
6807
|
error: z20.string().optional()
|
|
7183
6808
|
})
|
|
7184
6809
|
]);
|
|
6810
|
+
function countEnabledMcpTools(servers) {
|
|
6811
|
+
let serverCount = 0;
|
|
6812
|
+
let toolCount = 0;
|
|
6813
|
+
for (const server of servers) {
|
|
6814
|
+
if (server.disabled) continue;
|
|
6815
|
+
if (server.status !== "connected") continue;
|
|
6816
|
+
serverCount++;
|
|
6817
|
+
if (server.tools) {
|
|
6818
|
+
for (const tool of server.tools) {
|
|
6819
|
+
if (tool.enabledForPrompt !== false) {
|
|
6820
|
+
toolCount++;
|
|
6821
|
+
}
|
|
6822
|
+
}
|
|
6823
|
+
}
|
|
6824
|
+
}
|
|
6825
|
+
return { enabledToolCount: toolCount, enabledServerCount: serverCount };
|
|
6826
|
+
}
|
|
6827
|
+
|
|
6828
|
+
// src/skills.ts
|
|
6829
|
+
var SKILL_NAME_MIN_LENGTH = 1;
|
|
6830
|
+
var SKILL_NAME_MAX_LENGTH = 64;
|
|
6831
|
+
var SKILL_NAME_REGEX = /^[a-z0-9]+(?:-[a-z0-9]+)*$/;
|
|
6832
|
+
var SkillNameValidationError = /* @__PURE__ */ ((SkillNameValidationError2) => {
|
|
6833
|
+
SkillNameValidationError2["Empty"] = "empty";
|
|
6834
|
+
SkillNameValidationError2["TooLong"] = "too_long";
|
|
6835
|
+
SkillNameValidationError2["InvalidFormat"] = "invalid_format";
|
|
6836
|
+
return SkillNameValidationError2;
|
|
6837
|
+
})(SkillNameValidationError || {});
|
|
6838
|
+
function validateSkillName(name) {
|
|
6839
|
+
if (!name || name.length < SKILL_NAME_MIN_LENGTH) {
|
|
6840
|
+
return { valid: false, error: "empty" /* Empty */ };
|
|
6841
|
+
}
|
|
6842
|
+
if (name.length > SKILL_NAME_MAX_LENGTH) {
|
|
6843
|
+
return { valid: false, error: "too_long" /* TooLong */ };
|
|
6844
|
+
}
|
|
6845
|
+
if (!SKILL_NAME_REGEX.test(name)) {
|
|
6846
|
+
return { valid: false, error: "invalid_format" /* InvalidFormat */ };
|
|
6847
|
+
}
|
|
6848
|
+
return { valid: true };
|
|
6849
|
+
}
|
|
7185
6850
|
|
|
7186
6851
|
// src/todo.ts
|
|
7187
6852
|
import { z as z21 } from "zod";
|
|
@@ -7221,6 +6886,11 @@ var commandExecutionStatusSchema = z22.discriminatedUnion("status", [
|
|
|
7221
6886
|
})
|
|
7222
6887
|
]);
|
|
7223
6888
|
|
|
6889
|
+
// src/tool-params.ts
|
|
6890
|
+
function isLegacyReadFileParams(params) {
|
|
6891
|
+
return "_legacyFormat" in params && params._legacyFormat === true;
|
|
6892
|
+
}
|
|
6893
|
+
|
|
7224
6894
|
// src/vscode-extension-host.ts
|
|
7225
6895
|
import { z as z23 } from "zod";
|
|
7226
6896
|
var checkoutDiffPayloadSchema = z23.object({
|
|
@@ -7273,7 +6943,7 @@ export {
|
|
|
7273
6943
|
DEFAULT_CHECKPOINT_TIMEOUT_SECONDS,
|
|
7274
6944
|
DEFAULT_CONSECUTIVE_MISTAKE_LIMIT,
|
|
7275
6945
|
DEFAULT_MODES,
|
|
7276
|
-
|
|
6946
|
+
DEFAULT_TERMINAL_OUTPUT_PREVIEW_SIZE,
|
|
7277
6947
|
DEFAULT_WRITE_DELAY_MS,
|
|
7278
6948
|
DOUBAO_API_BASE_URL,
|
|
7279
6949
|
DOUBAO_API_CHAT_PATH,
|
|
@@ -7303,13 +6973,13 @@ export {
|
|
|
7303
6973
|
IpcOrigin,
|
|
7304
6974
|
LMSTUDIO_DEFAULT_TEMPERATURE,
|
|
7305
6975
|
MAX_CHECKPOINT_TIMEOUT_SECONDS,
|
|
6976
|
+
MAX_MCP_TOOLS_THRESHOLD,
|
|
7306
6977
|
MINIMAX_DEFAULT_MAX_TOKENS,
|
|
7307
6978
|
MINIMAX_DEFAULT_TEMPERATURE,
|
|
7308
6979
|
MIN_CHECKPOINT_TIMEOUT_SECONDS,
|
|
7309
6980
|
MISTRAL_DEFAULT_TEMPERATURE,
|
|
7310
6981
|
MODELS_BY_PROVIDER,
|
|
7311
6982
|
MOONSHOT_DEFAULT_TEMPERATURE,
|
|
7312
|
-
NATIVE_TOOL_DEFAULTS,
|
|
7313
6983
|
OPENAI_AZURE_AI_INFERENCE_PATH,
|
|
7314
6984
|
OPENAI_NATIVE_DEFAULT_TEMPERATURE,
|
|
7315
6985
|
OPENROUTER_DEFAULT_PROVIDER_NAME,
|
|
@@ -7324,7 +6994,11 @@ export {
|
|
|
7324
6994
|
RooModelsResponseSchema,
|
|
7325
6995
|
RooPricingSchema,
|
|
7326
6996
|
SECRET_STATE_KEYS,
|
|
7327
|
-
|
|
6997
|
+
SKILL_NAME_MAX_LENGTH,
|
|
6998
|
+
SKILL_NAME_MIN_LENGTH,
|
|
6999
|
+
SKILL_NAME_REGEX,
|
|
7000
|
+
SkillNameValidationError,
|
|
7001
|
+
TERMINAL_PREVIEW_BYTES,
|
|
7328
7002
|
TaskBridgeCommandName,
|
|
7329
7003
|
TaskBridgeEventName,
|
|
7330
7004
|
TaskCommandName,
|
|
@@ -7356,9 +7030,6 @@ export {
|
|
|
7356
7030
|
chutesDefaultModelId,
|
|
7357
7031
|
chutesDefaultModelInfo,
|
|
7358
7032
|
chutesModels,
|
|
7359
|
-
claudeCodeDefaultModelId,
|
|
7360
|
-
claudeCodeModels,
|
|
7361
|
-
claudeCodeReasoningConfig,
|
|
7362
7033
|
clineAskSchema,
|
|
7363
7034
|
clineAsks,
|
|
7364
7035
|
clineMessageSchema,
|
|
@@ -7373,6 +7044,7 @@ export {
|
|
|
7373
7044
|
commandIds,
|
|
7374
7045
|
contextCondenseSchema,
|
|
7375
7046
|
contextTruncationSchema,
|
|
7047
|
+
countEnabledMcpTools,
|
|
7376
7048
|
customModePromptsSchema,
|
|
7377
7049
|
customModesSettingsSchema,
|
|
7378
7050
|
customProviders,
|
|
@@ -7406,7 +7078,6 @@ export {
|
|
|
7406
7078
|
geminiDefaultModelId,
|
|
7407
7079
|
geminiModels,
|
|
7408
7080
|
getApiProtocol,
|
|
7409
|
-
getEffectiveProtocol,
|
|
7410
7081
|
getErrorMessage,
|
|
7411
7082
|
getErrorStatusCode,
|
|
7412
7083
|
getImageGenerationProvider,
|
|
@@ -7441,9 +7112,9 @@ export {
|
|
|
7441
7112
|
isInteractiveAsk,
|
|
7442
7113
|
isInternalProvider,
|
|
7443
7114
|
isLanguage,
|
|
7115
|
+
isLegacyReadFileParams,
|
|
7444
7116
|
isLocalProvider,
|
|
7445
7117
|
isModelParameter,
|
|
7446
|
-
isNativeProtocol,
|
|
7447
7118
|
isNonBlockingAsk,
|
|
7448
7119
|
isProviderName,
|
|
7449
7120
|
isResumableAsk,
|
|
@@ -7479,7 +7150,6 @@ export {
|
|
|
7479
7150
|
moonshotDefaultModelId,
|
|
7480
7151
|
moonshotModels,
|
|
7481
7152
|
nonBlockingAsks,
|
|
7482
|
-
normalizeClaudeCodeModelId,
|
|
7483
7153
|
ollamaDefaultModelId,
|
|
7484
7154
|
ollamaDefaultModelInfo,
|
|
7485
7155
|
openAiCodexDefaultModelId,
|
|
@@ -7553,6 +7223,7 @@ export {
|
|
|
7553
7223
|
userFeaturesSchema,
|
|
7554
7224
|
userSettingsConfigSchema,
|
|
7555
7225
|
userSettingsDataSchema,
|
|
7226
|
+
validateSkillName,
|
|
7556
7227
|
verbosityLevels,
|
|
7557
7228
|
verbosityLevelsSchema,
|
|
7558
7229
|
vercelAiGatewayDefaultModelId,
|