@roo-code/types 1.105.0 → 1.107.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/index.cjs +494 -709
- package/dist/index.cjs.map +1 -1
- package/dist/index.d.cts +1746 -2518
- package/dist/index.d.ts +1746 -2518
- package/dist/index.js +482 -700
- package/dist/index.js.map +1 -1
- package/package.json +2 -2
package/dist/index.cjs
CHANGED
|
@@ -42,7 +42,7 @@ __export(index_exports, {
|
|
|
42
42
|
DEFAULT_CHECKPOINT_TIMEOUT_SECONDS: () => DEFAULT_CHECKPOINT_TIMEOUT_SECONDS,
|
|
43
43
|
DEFAULT_CONSECUTIVE_MISTAKE_LIMIT: () => DEFAULT_CONSECUTIVE_MISTAKE_LIMIT,
|
|
44
44
|
DEFAULT_MODES: () => DEFAULT_MODES,
|
|
45
|
-
|
|
45
|
+
DEFAULT_TERMINAL_OUTPUT_PREVIEW_SIZE: () => DEFAULT_TERMINAL_OUTPUT_PREVIEW_SIZE,
|
|
46
46
|
DEFAULT_WRITE_DELAY_MS: () => DEFAULT_WRITE_DELAY_MS,
|
|
47
47
|
DOUBAO_API_BASE_URL: () => DOUBAO_API_BASE_URL,
|
|
48
48
|
DOUBAO_API_CHAT_PATH: () => DOUBAO_API_CHAT_PATH,
|
|
@@ -72,13 +72,13 @@ __export(index_exports, {
|
|
|
72
72
|
IpcOrigin: () => IpcOrigin,
|
|
73
73
|
LMSTUDIO_DEFAULT_TEMPERATURE: () => LMSTUDIO_DEFAULT_TEMPERATURE,
|
|
74
74
|
MAX_CHECKPOINT_TIMEOUT_SECONDS: () => MAX_CHECKPOINT_TIMEOUT_SECONDS,
|
|
75
|
+
MAX_MCP_TOOLS_THRESHOLD: () => MAX_MCP_TOOLS_THRESHOLD,
|
|
75
76
|
MINIMAX_DEFAULT_MAX_TOKENS: () => MINIMAX_DEFAULT_MAX_TOKENS,
|
|
76
77
|
MINIMAX_DEFAULT_TEMPERATURE: () => MINIMAX_DEFAULT_TEMPERATURE,
|
|
77
78
|
MIN_CHECKPOINT_TIMEOUT_SECONDS: () => MIN_CHECKPOINT_TIMEOUT_SECONDS,
|
|
78
79
|
MISTRAL_DEFAULT_TEMPERATURE: () => MISTRAL_DEFAULT_TEMPERATURE,
|
|
79
80
|
MODELS_BY_PROVIDER: () => MODELS_BY_PROVIDER,
|
|
80
81
|
MOONSHOT_DEFAULT_TEMPERATURE: () => MOONSHOT_DEFAULT_TEMPERATURE,
|
|
81
|
-
NATIVE_TOOL_DEFAULTS: () => NATIVE_TOOL_DEFAULTS,
|
|
82
82
|
OPENAI_AZURE_AI_INFERENCE_PATH: () => OPENAI_AZURE_AI_INFERENCE_PATH,
|
|
83
83
|
OPENAI_NATIVE_DEFAULT_TEMPERATURE: () => OPENAI_NATIVE_DEFAULT_TEMPERATURE,
|
|
84
84
|
OPENROUTER_DEFAULT_PROVIDER_NAME: () => OPENROUTER_DEFAULT_PROVIDER_NAME,
|
|
@@ -93,7 +93,11 @@ __export(index_exports, {
|
|
|
93
93
|
RooModelsResponseSchema: () => RooModelsResponseSchema,
|
|
94
94
|
RooPricingSchema: () => RooPricingSchema,
|
|
95
95
|
SECRET_STATE_KEYS: () => SECRET_STATE_KEYS,
|
|
96
|
-
|
|
96
|
+
SKILL_NAME_MAX_LENGTH: () => SKILL_NAME_MAX_LENGTH,
|
|
97
|
+
SKILL_NAME_MIN_LENGTH: () => SKILL_NAME_MIN_LENGTH,
|
|
98
|
+
SKILL_NAME_REGEX: () => SKILL_NAME_REGEX,
|
|
99
|
+
SkillNameValidationError: () => SkillNameValidationError,
|
|
100
|
+
TERMINAL_PREVIEW_BYTES: () => TERMINAL_PREVIEW_BYTES,
|
|
97
101
|
TaskBridgeCommandName: () => TaskBridgeCommandName,
|
|
98
102
|
TaskBridgeEventName: () => TaskBridgeEventName,
|
|
99
103
|
TaskCommandName: () => TaskCommandName,
|
|
@@ -125,9 +129,6 @@ __export(index_exports, {
|
|
|
125
129
|
chutesDefaultModelId: () => chutesDefaultModelId,
|
|
126
130
|
chutesDefaultModelInfo: () => chutesDefaultModelInfo,
|
|
127
131
|
chutesModels: () => chutesModels,
|
|
128
|
-
claudeCodeDefaultModelId: () => claudeCodeDefaultModelId,
|
|
129
|
-
claudeCodeModels: () => claudeCodeModels,
|
|
130
|
-
claudeCodeReasoningConfig: () => claudeCodeReasoningConfig,
|
|
131
132
|
clineAskSchema: () => clineAskSchema,
|
|
132
133
|
clineAsks: () => clineAsks,
|
|
133
134
|
clineMessageSchema: () => clineMessageSchema,
|
|
@@ -142,6 +143,7 @@ __export(index_exports, {
|
|
|
142
143
|
commandIds: () => commandIds,
|
|
143
144
|
contextCondenseSchema: () => contextCondenseSchema,
|
|
144
145
|
contextTruncationSchema: () => contextTruncationSchema,
|
|
146
|
+
countEnabledMcpTools: () => countEnabledMcpTools,
|
|
145
147
|
customModePromptsSchema: () => customModePromptsSchema,
|
|
146
148
|
customModesSettingsSchema: () => customModesSettingsSchema,
|
|
147
149
|
customProviders: () => customProviders,
|
|
@@ -175,7 +177,6 @@ __export(index_exports, {
|
|
|
175
177
|
geminiDefaultModelId: () => geminiDefaultModelId,
|
|
176
178
|
geminiModels: () => geminiModels,
|
|
177
179
|
getApiProtocol: () => getApiProtocol,
|
|
178
|
-
getEffectiveProtocol: () => getEffectiveProtocol,
|
|
179
180
|
getErrorMessage: () => getErrorMessage,
|
|
180
181
|
getErrorStatusCode: () => getErrorStatusCode,
|
|
181
182
|
getImageGenerationProvider: () => getImageGenerationProvider,
|
|
@@ -210,9 +211,9 @@ __export(index_exports, {
|
|
|
210
211
|
isInteractiveAsk: () => isInteractiveAsk,
|
|
211
212
|
isInternalProvider: () => isInternalProvider,
|
|
212
213
|
isLanguage: () => isLanguage,
|
|
214
|
+
isLegacyReadFileParams: () => isLegacyReadFileParams,
|
|
213
215
|
isLocalProvider: () => isLocalProvider,
|
|
214
216
|
isModelParameter: () => isModelParameter,
|
|
215
|
-
isNativeProtocol: () => isNativeProtocol,
|
|
216
217
|
isNonBlockingAsk: () => isNonBlockingAsk,
|
|
217
218
|
isProviderName: () => isProviderName,
|
|
218
219
|
isResumableAsk: () => isResumableAsk,
|
|
@@ -248,9 +249,10 @@ __export(index_exports, {
|
|
|
248
249
|
moonshotDefaultModelId: () => moonshotDefaultModelId,
|
|
249
250
|
moonshotModels: () => moonshotModels,
|
|
250
251
|
nonBlockingAsks: () => nonBlockingAsks,
|
|
251
|
-
normalizeClaudeCodeModelId: () => normalizeClaudeCodeModelId,
|
|
252
252
|
ollamaDefaultModelId: () => ollamaDefaultModelId,
|
|
253
253
|
ollamaDefaultModelInfo: () => ollamaDefaultModelInfo,
|
|
254
|
+
openAiCodexDefaultModelId: () => openAiCodexDefaultModelId,
|
|
255
|
+
openAiCodexModels: () => openAiCodexModels,
|
|
254
256
|
openAiModelInfoSaneDefaults: () => openAiModelInfoSaneDefaults,
|
|
255
257
|
openAiNativeDefaultModelId: () => openAiNativeDefaultModelId,
|
|
256
258
|
openAiNativeModels: () => openAiNativeModels,
|
|
@@ -320,6 +322,7 @@ __export(index_exports, {
|
|
|
320
322
|
userFeaturesSchema: () => userFeaturesSchema,
|
|
321
323
|
userSettingsConfigSchema: () => userSettingsConfigSchema,
|
|
322
324
|
userSettingsDataSchema: () => userSettingsDataSchema,
|
|
325
|
+
validateSkillName: () => validateSkillName,
|
|
323
326
|
verbosityLevels: () => verbosityLevels,
|
|
324
327
|
verbosityLevelsSchema: () => verbosityLevelsSchema,
|
|
325
328
|
vercelAiGatewayDefaultModelId: () => vercelAiGatewayDefaultModelId,
|
|
@@ -415,7 +418,9 @@ var clineSays = [
|
|
|
415
418
|
"condense_context_error",
|
|
416
419
|
"sliding_window_truncation",
|
|
417
420
|
"codebase_search_result",
|
|
418
|
-
"user_edit_todos"
|
|
421
|
+
"user_edit_todos",
|
|
422
|
+
"too_many_tools_warning",
|
|
423
|
+
"tool"
|
|
419
424
|
];
|
|
420
425
|
var clineSaySchema = import_zod.z.enum(clineSays);
|
|
421
426
|
var toolProgressStatusSchema = import_zod.z.object({
|
|
@@ -483,6 +488,7 @@ var toolGroupsSchema = import_zod2.z.enum(toolGroups);
|
|
|
483
488
|
var toolNames = [
|
|
484
489
|
"execute_command",
|
|
485
490
|
"read_file",
|
|
491
|
+
"read_command_output",
|
|
486
492
|
"write_to_file",
|
|
487
493
|
"apply_diff",
|
|
488
494
|
"search_and_replace",
|
|
@@ -498,10 +504,10 @@ var toolNames = [
|
|
|
498
504
|
"attempt_completion",
|
|
499
505
|
"switch_mode",
|
|
500
506
|
"new_task",
|
|
501
|
-
"fetch_instructions",
|
|
502
507
|
"codebase_search",
|
|
503
508
|
"update_todo_list",
|
|
504
509
|
"run_slash_command",
|
|
510
|
+
"skill",
|
|
505
511
|
"generate_image",
|
|
506
512
|
"custom_tool"
|
|
507
513
|
];
|
|
@@ -513,20 +519,6 @@ var toolUsageSchema = import_zod2.z.record(
|
|
|
513
519
|
failures: import_zod2.z.number()
|
|
514
520
|
})
|
|
515
521
|
);
|
|
516
|
-
var TOOL_PROTOCOL = {
|
|
517
|
-
XML: "xml",
|
|
518
|
-
NATIVE: "native"
|
|
519
|
-
};
|
|
520
|
-
var NATIVE_TOOL_DEFAULTS = {
|
|
521
|
-
supportsNativeTools: true,
|
|
522
|
-
defaultToolProtocol: TOOL_PROTOCOL.NATIVE
|
|
523
|
-
};
|
|
524
|
-
function isNativeProtocol(protocol) {
|
|
525
|
-
return protocol === TOOL_PROTOCOL.NATIVE;
|
|
526
|
-
}
|
|
527
|
-
function getEffectiveProtocol(toolProtocol) {
|
|
528
|
-
return toolProtocol || TOOL_PROTOCOL.XML;
|
|
529
|
-
}
|
|
530
522
|
|
|
531
523
|
// src/events.ts
|
|
532
524
|
var RooCodeEventName = /* @__PURE__ */ ((RooCodeEventName2) => {
|
|
@@ -550,6 +542,7 @@ var RooCodeEventName = /* @__PURE__ */ ((RooCodeEventName2) => {
|
|
|
550
542
|
RooCodeEventName2["TaskModeSwitched"] = "taskModeSwitched";
|
|
551
543
|
RooCodeEventName2["TaskAskResponded"] = "taskAskResponded";
|
|
552
544
|
RooCodeEventName2["TaskUserMessage"] = "taskUserMessage";
|
|
545
|
+
RooCodeEventName2["QueuedMessagesUpdated"] = "queuedMessagesUpdated";
|
|
553
546
|
RooCodeEventName2["TaskTokenUsageUpdated"] = "taskTokenUsageUpdated";
|
|
554
547
|
RooCodeEventName2["TaskToolFailed"] = "taskToolFailed";
|
|
555
548
|
RooCodeEventName2["ModeChanged"] = "modeChanged";
|
|
@@ -609,6 +602,7 @@ var rooCodeEventsSchema = import_zod3.z.object({
|
|
|
609
602
|
["taskModeSwitched" /* TaskModeSwitched */]: import_zod3.z.tuple([import_zod3.z.string(), import_zod3.z.string()]),
|
|
610
603
|
["taskAskResponded" /* TaskAskResponded */]: import_zod3.z.tuple([import_zod3.z.string()]),
|
|
611
604
|
["taskUserMessage" /* TaskUserMessage */]: import_zod3.z.tuple([import_zod3.z.string()]),
|
|
605
|
+
["queuedMessagesUpdated" /* QueuedMessagesUpdated */]: import_zod3.z.tuple([import_zod3.z.string(), import_zod3.z.array(queuedMessageSchema)]),
|
|
612
606
|
["taskToolFailed" /* TaskToolFailed */]: import_zod3.z.tuple([import_zod3.z.string(), toolNamesSchema, import_zod3.z.string()]),
|
|
613
607
|
["taskTokenUsageUpdated" /* TaskTokenUsageUpdated */]: import_zod3.z.tuple([import_zod3.z.string(), tokenUsageSchema, toolUsageSchema]),
|
|
614
608
|
["modeChanged" /* ModeChanged */]: import_zod3.z.tuple([import_zod3.z.string()]),
|
|
@@ -714,6 +708,11 @@ var taskEventSchema = import_zod3.z.discriminatedUnion("eventName", [
|
|
|
714
708
|
payload: rooCodeEventsSchema.shape["taskAskResponded" /* TaskAskResponded */],
|
|
715
709
|
taskId: import_zod3.z.number().optional()
|
|
716
710
|
}),
|
|
711
|
+
import_zod3.z.object({
|
|
712
|
+
eventName: import_zod3.z.literal("queuedMessagesUpdated" /* QueuedMessagesUpdated */),
|
|
713
|
+
payload: rooCodeEventsSchema.shape["queuedMessagesUpdated" /* QueuedMessagesUpdated */],
|
|
714
|
+
taskId: import_zod3.z.number().optional()
|
|
715
|
+
}),
|
|
717
716
|
// Task Analytics
|
|
718
717
|
import_zod3.z.object({
|
|
719
718
|
eventName: import_zod3.z.literal("taskToolFailed" /* TaskToolFailed */),
|
|
@@ -814,10 +813,6 @@ var modelInfoSchema = import_zod5.z.object({
|
|
|
814
813
|
isStealthModel: import_zod5.z.boolean().optional(),
|
|
815
814
|
// Flag to indicate if the model is free (no cost)
|
|
816
815
|
isFree: import_zod5.z.boolean().optional(),
|
|
817
|
-
// Flag to indicate if the model supports native tool calling (OpenAI-style function calling)
|
|
818
|
-
supportsNativeTools: import_zod5.z.boolean().optional(),
|
|
819
|
-
// Default tool protocol preferred by this model (if not specified, falls back to capability/provider defaults)
|
|
820
|
-
defaultToolProtocol: import_zod5.z.enum(["xml", "native"]).optional(),
|
|
821
816
|
// Exclude specific native tools from being available (only applies to native protocol)
|
|
822
817
|
// These tools will be removed from the set of tools available to the model
|
|
823
818
|
excludedTools: import_zod5.z.array(import_zod5.z.string()).optional(),
|
|
@@ -914,8 +909,6 @@ var anthropicModels = {
|
|
|
914
909
|
// Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
|
|
915
910
|
supportsImages: true,
|
|
916
911
|
supportsPromptCache: true,
|
|
917
|
-
supportsNativeTools: true,
|
|
918
|
-
defaultToolProtocol: "native",
|
|
919
912
|
inputPrice: 3,
|
|
920
913
|
// $3 per million input tokens (≤200K context)
|
|
921
914
|
outputPrice: 15,
|
|
@@ -948,8 +941,6 @@ var anthropicModels = {
|
|
|
948
941
|
// Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
|
|
949
942
|
supportsImages: true,
|
|
950
943
|
supportsPromptCache: true,
|
|
951
|
-
supportsNativeTools: true,
|
|
952
|
-
defaultToolProtocol: "native",
|
|
953
944
|
inputPrice: 3,
|
|
954
945
|
// $3 per million input tokens (≤200K context)
|
|
955
946
|
outputPrice: 15,
|
|
@@ -981,8 +972,6 @@ var anthropicModels = {
|
|
|
981
972
|
contextWindow: 2e5,
|
|
982
973
|
supportsImages: true,
|
|
983
974
|
supportsPromptCache: true,
|
|
984
|
-
supportsNativeTools: true,
|
|
985
|
-
defaultToolProtocol: "native",
|
|
986
975
|
inputPrice: 5,
|
|
987
976
|
// $5 per million input tokens
|
|
988
977
|
outputPrice: 25,
|
|
@@ -999,8 +988,6 @@ var anthropicModels = {
|
|
|
999
988
|
contextWindow: 2e5,
|
|
1000
989
|
supportsImages: true,
|
|
1001
990
|
supportsPromptCache: true,
|
|
1002
|
-
supportsNativeTools: true,
|
|
1003
|
-
defaultToolProtocol: "native",
|
|
1004
991
|
inputPrice: 15,
|
|
1005
992
|
// $15 per million input tokens
|
|
1006
993
|
outputPrice: 75,
|
|
@@ -1017,8 +1004,6 @@ var anthropicModels = {
|
|
|
1017
1004
|
contextWindow: 2e5,
|
|
1018
1005
|
supportsImages: true,
|
|
1019
1006
|
supportsPromptCache: true,
|
|
1020
|
-
supportsNativeTools: true,
|
|
1021
|
-
defaultToolProtocol: "native",
|
|
1022
1007
|
inputPrice: 15,
|
|
1023
1008
|
// $15 per million input tokens
|
|
1024
1009
|
outputPrice: 75,
|
|
@@ -1035,8 +1020,6 @@ var anthropicModels = {
|
|
|
1035
1020
|
contextWindow: 2e5,
|
|
1036
1021
|
supportsImages: true,
|
|
1037
1022
|
supportsPromptCache: true,
|
|
1038
|
-
supportsNativeTools: true,
|
|
1039
|
-
defaultToolProtocol: "native",
|
|
1040
1023
|
inputPrice: 3,
|
|
1041
1024
|
// $3 per million input tokens
|
|
1042
1025
|
outputPrice: 15,
|
|
@@ -1054,8 +1037,6 @@ var anthropicModels = {
|
|
|
1054
1037
|
contextWindow: 2e5,
|
|
1055
1038
|
supportsImages: true,
|
|
1056
1039
|
supportsPromptCache: true,
|
|
1057
|
-
supportsNativeTools: true,
|
|
1058
|
-
defaultToolProtocol: "native",
|
|
1059
1040
|
inputPrice: 3,
|
|
1060
1041
|
// $3 per million input tokens
|
|
1061
1042
|
outputPrice: 15,
|
|
@@ -1070,8 +1051,6 @@ var anthropicModels = {
|
|
|
1070
1051
|
contextWindow: 2e5,
|
|
1071
1052
|
supportsImages: true,
|
|
1072
1053
|
supportsPromptCache: true,
|
|
1073
|
-
supportsNativeTools: true,
|
|
1074
|
-
defaultToolProtocol: "native",
|
|
1075
1054
|
inputPrice: 3,
|
|
1076
1055
|
// $3 per million input tokens
|
|
1077
1056
|
outputPrice: 15,
|
|
@@ -1086,8 +1065,6 @@ var anthropicModels = {
|
|
|
1086
1065
|
contextWindow: 2e5,
|
|
1087
1066
|
supportsImages: false,
|
|
1088
1067
|
supportsPromptCache: true,
|
|
1089
|
-
supportsNativeTools: true,
|
|
1090
|
-
defaultToolProtocol: "native",
|
|
1091
1068
|
inputPrice: 1,
|
|
1092
1069
|
outputPrice: 5,
|
|
1093
1070
|
cacheWritesPrice: 1.25,
|
|
@@ -1098,8 +1075,6 @@ var anthropicModels = {
|
|
|
1098
1075
|
contextWindow: 2e5,
|
|
1099
1076
|
supportsImages: true,
|
|
1100
1077
|
supportsPromptCache: true,
|
|
1101
|
-
supportsNativeTools: true,
|
|
1102
|
-
defaultToolProtocol: "native",
|
|
1103
1078
|
inputPrice: 15,
|
|
1104
1079
|
outputPrice: 75,
|
|
1105
1080
|
cacheWritesPrice: 18.75,
|
|
@@ -1110,8 +1085,6 @@ var anthropicModels = {
|
|
|
1110
1085
|
contextWindow: 2e5,
|
|
1111
1086
|
supportsImages: true,
|
|
1112
1087
|
supportsPromptCache: true,
|
|
1113
|
-
supportsNativeTools: true,
|
|
1114
|
-
defaultToolProtocol: "native",
|
|
1115
1088
|
inputPrice: 0.25,
|
|
1116
1089
|
outputPrice: 1.25,
|
|
1117
1090
|
cacheWritesPrice: 0.3,
|
|
@@ -1122,8 +1095,6 @@ var anthropicModels = {
|
|
|
1122
1095
|
contextWindow: 2e5,
|
|
1123
1096
|
supportsImages: true,
|
|
1124
1097
|
supportsPromptCache: true,
|
|
1125
|
-
supportsNativeTools: true,
|
|
1126
|
-
defaultToolProtocol: "native",
|
|
1127
1098
|
inputPrice: 1,
|
|
1128
1099
|
outputPrice: 5,
|
|
1129
1100
|
cacheWritesPrice: 1.25,
|
|
@@ -1141,7 +1112,6 @@ var basetenModels = {
|
|
|
1141
1112
|
contextWindow: 262e3,
|
|
1142
1113
|
supportsImages: false,
|
|
1143
1114
|
supportsPromptCache: false,
|
|
1144
|
-
supportsNativeTools: true,
|
|
1145
1115
|
inputPrice: 0.6,
|
|
1146
1116
|
outputPrice: 2.5,
|
|
1147
1117
|
cacheWritesPrice: 0,
|
|
@@ -1153,7 +1123,6 @@ var basetenModels = {
|
|
|
1153
1123
|
contextWindow: 2e5,
|
|
1154
1124
|
supportsImages: false,
|
|
1155
1125
|
supportsPromptCache: false,
|
|
1156
|
-
supportsNativeTools: true,
|
|
1157
1126
|
inputPrice: 0.6,
|
|
1158
1127
|
outputPrice: 2.2,
|
|
1159
1128
|
cacheWritesPrice: 0,
|
|
@@ -1165,7 +1134,6 @@ var basetenModels = {
|
|
|
1165
1134
|
contextWindow: 163840,
|
|
1166
1135
|
supportsImages: false,
|
|
1167
1136
|
supportsPromptCache: false,
|
|
1168
|
-
supportsNativeTools: true,
|
|
1169
1137
|
inputPrice: 2.55,
|
|
1170
1138
|
outputPrice: 5.95,
|
|
1171
1139
|
cacheWritesPrice: 0,
|
|
@@ -1177,7 +1145,6 @@ var basetenModels = {
|
|
|
1177
1145
|
contextWindow: 163840,
|
|
1178
1146
|
supportsImages: false,
|
|
1179
1147
|
supportsPromptCache: false,
|
|
1180
|
-
supportsNativeTools: true,
|
|
1181
1148
|
inputPrice: 2.55,
|
|
1182
1149
|
outputPrice: 5.95,
|
|
1183
1150
|
cacheWritesPrice: 0,
|
|
@@ -1189,7 +1156,6 @@ var basetenModels = {
|
|
|
1189
1156
|
contextWindow: 163840,
|
|
1190
1157
|
supportsImages: false,
|
|
1191
1158
|
supportsPromptCache: false,
|
|
1192
|
-
supportsNativeTools: true,
|
|
1193
1159
|
inputPrice: 0.77,
|
|
1194
1160
|
outputPrice: 0.77,
|
|
1195
1161
|
cacheWritesPrice: 0,
|
|
@@ -1201,7 +1167,6 @@ var basetenModels = {
|
|
|
1201
1167
|
contextWindow: 163840,
|
|
1202
1168
|
supportsImages: false,
|
|
1203
1169
|
supportsPromptCache: false,
|
|
1204
|
-
supportsNativeTools: true,
|
|
1205
1170
|
inputPrice: 0.5,
|
|
1206
1171
|
outputPrice: 1.5,
|
|
1207
1172
|
cacheWritesPrice: 0,
|
|
@@ -1213,7 +1178,6 @@ var basetenModels = {
|
|
|
1213
1178
|
contextWindow: 163840,
|
|
1214
1179
|
supportsImages: false,
|
|
1215
1180
|
supportsPromptCache: false,
|
|
1216
|
-
supportsNativeTools: true,
|
|
1217
1181
|
inputPrice: 0.3,
|
|
1218
1182
|
outputPrice: 0.45,
|
|
1219
1183
|
cacheWritesPrice: 0,
|
|
@@ -1225,7 +1189,6 @@ var basetenModels = {
|
|
|
1225
1189
|
contextWindow: 128072,
|
|
1226
1190
|
supportsImages: false,
|
|
1227
1191
|
supportsPromptCache: false,
|
|
1228
|
-
supportsNativeTools: true,
|
|
1229
1192
|
inputPrice: 0.1,
|
|
1230
1193
|
outputPrice: 0.5,
|
|
1231
1194
|
cacheWritesPrice: 0,
|
|
@@ -1237,7 +1200,6 @@ var basetenModels = {
|
|
|
1237
1200
|
contextWindow: 262144,
|
|
1238
1201
|
supportsImages: false,
|
|
1239
1202
|
supportsPromptCache: false,
|
|
1240
|
-
supportsNativeTools: true,
|
|
1241
1203
|
inputPrice: 0.22,
|
|
1242
1204
|
outputPrice: 0.8,
|
|
1243
1205
|
cacheWritesPrice: 0,
|
|
@@ -1249,7 +1211,6 @@ var basetenModels = {
|
|
|
1249
1211
|
contextWindow: 262144,
|
|
1250
1212
|
supportsImages: false,
|
|
1251
1213
|
supportsPromptCache: false,
|
|
1252
|
-
supportsNativeTools: true,
|
|
1253
1214
|
inputPrice: 0.38,
|
|
1254
1215
|
outputPrice: 1.53,
|
|
1255
1216
|
cacheWritesPrice: 0,
|
|
@@ -1261,7 +1222,6 @@ var basetenModels = {
|
|
|
1261
1222
|
contextWindow: 262e3,
|
|
1262
1223
|
supportsImages: false,
|
|
1263
1224
|
supportsPromptCache: false,
|
|
1264
|
-
supportsNativeTools: true,
|
|
1265
1225
|
inputPrice: 0.6,
|
|
1266
1226
|
outputPrice: 2.5,
|
|
1267
1227
|
cacheWritesPrice: 0,
|
|
@@ -1281,8 +1241,6 @@ var bedrockModels = {
|
|
|
1281
1241
|
supportsImages: true,
|
|
1282
1242
|
supportsPromptCache: true,
|
|
1283
1243
|
supportsReasoningBudget: true,
|
|
1284
|
-
supportsNativeTools: true,
|
|
1285
|
-
defaultToolProtocol: "native",
|
|
1286
1244
|
inputPrice: 3,
|
|
1287
1245
|
outputPrice: 15,
|
|
1288
1246
|
cacheWritesPrice: 3.75,
|
|
@@ -1296,7 +1254,6 @@ var bedrockModels = {
|
|
|
1296
1254
|
contextWindow: 3e5,
|
|
1297
1255
|
supportsImages: true,
|
|
1298
1256
|
supportsPromptCache: true,
|
|
1299
|
-
supportsNativeTools: true,
|
|
1300
1257
|
inputPrice: 0.8,
|
|
1301
1258
|
outputPrice: 3.2,
|
|
1302
1259
|
cacheWritesPrice: 0.8,
|
|
@@ -1312,7 +1269,6 @@ var bedrockModels = {
|
|
|
1312
1269
|
contextWindow: 3e5,
|
|
1313
1270
|
supportsImages: true,
|
|
1314
1271
|
supportsPromptCache: false,
|
|
1315
|
-
supportsNativeTools: true,
|
|
1316
1272
|
inputPrice: 1,
|
|
1317
1273
|
outputPrice: 4,
|
|
1318
1274
|
cacheWritesPrice: 1,
|
|
@@ -1326,7 +1282,6 @@ var bedrockModels = {
|
|
|
1326
1282
|
contextWindow: 3e5,
|
|
1327
1283
|
supportsImages: true,
|
|
1328
1284
|
supportsPromptCache: true,
|
|
1329
|
-
supportsNativeTools: true,
|
|
1330
1285
|
inputPrice: 0.06,
|
|
1331
1286
|
outputPrice: 0.24,
|
|
1332
1287
|
cacheWritesPrice: 0.06,
|
|
@@ -1342,7 +1297,6 @@ var bedrockModels = {
|
|
|
1342
1297
|
contextWindow: 1e6,
|
|
1343
1298
|
supportsImages: true,
|
|
1344
1299
|
supportsPromptCache: true,
|
|
1345
|
-
supportsNativeTools: true,
|
|
1346
1300
|
inputPrice: 0.33,
|
|
1347
1301
|
outputPrice: 2.75,
|
|
1348
1302
|
cacheWritesPrice: 0,
|
|
@@ -1358,7 +1312,6 @@ var bedrockModels = {
|
|
|
1358
1312
|
contextWindow: 128e3,
|
|
1359
1313
|
supportsImages: false,
|
|
1360
1314
|
supportsPromptCache: true,
|
|
1361
|
-
supportsNativeTools: true,
|
|
1362
1315
|
inputPrice: 0.035,
|
|
1363
1316
|
outputPrice: 0.14,
|
|
1364
1317
|
cacheWritesPrice: 0.035,
|
|
@@ -1375,8 +1328,6 @@ var bedrockModels = {
|
|
|
1375
1328
|
supportsImages: true,
|
|
1376
1329
|
supportsPromptCache: true,
|
|
1377
1330
|
supportsReasoningBudget: true,
|
|
1378
|
-
supportsNativeTools: true,
|
|
1379
|
-
defaultToolProtocol: "native",
|
|
1380
1331
|
inputPrice: 3,
|
|
1381
1332
|
outputPrice: 15,
|
|
1382
1333
|
cacheWritesPrice: 3.75,
|
|
@@ -1391,8 +1342,6 @@ var bedrockModels = {
|
|
|
1391
1342
|
supportsImages: true,
|
|
1392
1343
|
supportsPromptCache: true,
|
|
1393
1344
|
supportsReasoningBudget: true,
|
|
1394
|
-
supportsNativeTools: true,
|
|
1395
|
-
defaultToolProtocol: "native",
|
|
1396
1345
|
inputPrice: 15,
|
|
1397
1346
|
outputPrice: 75,
|
|
1398
1347
|
cacheWritesPrice: 18.75,
|
|
@@ -1407,8 +1356,6 @@ var bedrockModels = {
|
|
|
1407
1356
|
supportsImages: true,
|
|
1408
1357
|
supportsPromptCache: true,
|
|
1409
1358
|
supportsReasoningBudget: true,
|
|
1410
|
-
supportsNativeTools: true,
|
|
1411
|
-
defaultToolProtocol: "native",
|
|
1412
1359
|
inputPrice: 5,
|
|
1413
1360
|
outputPrice: 25,
|
|
1414
1361
|
cacheWritesPrice: 6.25,
|
|
@@ -1423,8 +1370,6 @@ var bedrockModels = {
|
|
|
1423
1370
|
supportsImages: true,
|
|
1424
1371
|
supportsPromptCache: true,
|
|
1425
1372
|
supportsReasoningBudget: true,
|
|
1426
|
-
supportsNativeTools: true,
|
|
1427
|
-
defaultToolProtocol: "native",
|
|
1428
1373
|
inputPrice: 15,
|
|
1429
1374
|
outputPrice: 75,
|
|
1430
1375
|
cacheWritesPrice: 18.75,
|
|
@@ -1439,8 +1384,6 @@ var bedrockModels = {
|
|
|
1439
1384
|
supportsImages: true,
|
|
1440
1385
|
supportsPromptCache: true,
|
|
1441
1386
|
supportsReasoningBudget: true,
|
|
1442
|
-
supportsNativeTools: true,
|
|
1443
|
-
defaultToolProtocol: "native",
|
|
1444
1387
|
inputPrice: 3,
|
|
1445
1388
|
outputPrice: 15,
|
|
1446
1389
|
cacheWritesPrice: 3.75,
|
|
@@ -1454,8 +1397,6 @@ var bedrockModels = {
|
|
|
1454
1397
|
contextWindow: 2e5,
|
|
1455
1398
|
supportsImages: true,
|
|
1456
1399
|
supportsPromptCache: true,
|
|
1457
|
-
supportsNativeTools: true,
|
|
1458
|
-
defaultToolProtocol: "native",
|
|
1459
1400
|
inputPrice: 3,
|
|
1460
1401
|
outputPrice: 15,
|
|
1461
1402
|
cacheWritesPrice: 3.75,
|
|
@@ -1469,8 +1410,6 @@ var bedrockModels = {
|
|
|
1469
1410
|
contextWindow: 2e5,
|
|
1470
1411
|
supportsImages: false,
|
|
1471
1412
|
supportsPromptCache: true,
|
|
1472
|
-
supportsNativeTools: true,
|
|
1473
|
-
defaultToolProtocol: "native",
|
|
1474
1413
|
inputPrice: 0.8,
|
|
1475
1414
|
outputPrice: 4,
|
|
1476
1415
|
cacheWritesPrice: 1,
|
|
@@ -1485,8 +1424,6 @@ var bedrockModels = {
|
|
|
1485
1424
|
supportsImages: true,
|
|
1486
1425
|
supportsPromptCache: true,
|
|
1487
1426
|
supportsReasoningBudget: true,
|
|
1488
|
-
supportsNativeTools: true,
|
|
1489
|
-
defaultToolProtocol: "native",
|
|
1490
1427
|
inputPrice: 1,
|
|
1491
1428
|
outputPrice: 5,
|
|
1492
1429
|
cacheWritesPrice: 1.25,
|
|
@@ -1502,8 +1439,6 @@ var bedrockModels = {
|
|
|
1502
1439
|
contextWindow: 2e5,
|
|
1503
1440
|
supportsImages: true,
|
|
1504
1441
|
supportsPromptCache: false,
|
|
1505
|
-
supportsNativeTools: true,
|
|
1506
|
-
defaultToolProtocol: "native",
|
|
1507
1442
|
inputPrice: 3,
|
|
1508
1443
|
outputPrice: 15
|
|
1509
1444
|
},
|
|
@@ -1512,8 +1447,6 @@ var bedrockModels = {
|
|
|
1512
1447
|
contextWindow: 2e5,
|
|
1513
1448
|
supportsImages: true,
|
|
1514
1449
|
supportsPromptCache: false,
|
|
1515
|
-
supportsNativeTools: true,
|
|
1516
|
-
defaultToolProtocol: "native",
|
|
1517
1450
|
inputPrice: 15,
|
|
1518
1451
|
outputPrice: 75
|
|
1519
1452
|
},
|
|
@@ -1522,8 +1455,6 @@ var bedrockModels = {
|
|
|
1522
1455
|
contextWindow: 2e5,
|
|
1523
1456
|
supportsImages: true,
|
|
1524
1457
|
supportsPromptCache: false,
|
|
1525
|
-
supportsNativeTools: true,
|
|
1526
|
-
defaultToolProtocol: "native",
|
|
1527
1458
|
inputPrice: 3,
|
|
1528
1459
|
outputPrice: 15
|
|
1529
1460
|
},
|
|
@@ -1532,8 +1463,6 @@ var bedrockModels = {
|
|
|
1532
1463
|
contextWindow: 2e5,
|
|
1533
1464
|
supportsImages: true,
|
|
1534
1465
|
supportsPromptCache: false,
|
|
1535
|
-
supportsNativeTools: true,
|
|
1536
|
-
defaultToolProtocol: "native",
|
|
1537
1466
|
inputPrice: 0.25,
|
|
1538
1467
|
outputPrice: 1.25
|
|
1539
1468
|
},
|
|
@@ -1542,7 +1471,6 @@ var bedrockModels = {
|
|
|
1542
1471
|
contextWindow: 128e3,
|
|
1543
1472
|
supportsImages: false,
|
|
1544
1473
|
supportsPromptCache: false,
|
|
1545
|
-
supportsNativeTools: true,
|
|
1546
1474
|
inputPrice: 1.35,
|
|
1547
1475
|
outputPrice: 5.4
|
|
1548
1476
|
},
|
|
@@ -1551,7 +1479,6 @@ var bedrockModels = {
|
|
|
1551
1479
|
contextWindow: 128e3,
|
|
1552
1480
|
supportsImages: false,
|
|
1553
1481
|
supportsPromptCache: false,
|
|
1554
|
-
supportsNativeTools: true,
|
|
1555
1482
|
inputPrice: 0.5,
|
|
1556
1483
|
outputPrice: 1.5,
|
|
1557
1484
|
description: "GPT-OSS 20B - Optimized for low latency and local/specialized use cases"
|
|
@@ -1561,7 +1488,6 @@ var bedrockModels = {
|
|
|
1561
1488
|
contextWindow: 128e3,
|
|
1562
1489
|
supportsImages: false,
|
|
1563
1490
|
supportsPromptCache: false,
|
|
1564
|
-
supportsNativeTools: true,
|
|
1565
1491
|
inputPrice: 2,
|
|
1566
1492
|
outputPrice: 6,
|
|
1567
1493
|
description: "GPT-OSS 120B - Production-ready, general-purpose, high-reasoning model"
|
|
@@ -1571,7 +1497,6 @@ var bedrockModels = {
|
|
|
1571
1497
|
contextWindow: 128e3,
|
|
1572
1498
|
supportsImages: false,
|
|
1573
1499
|
supportsPromptCache: false,
|
|
1574
|
-
supportsNativeTools: true,
|
|
1575
1500
|
inputPrice: 0.72,
|
|
1576
1501
|
outputPrice: 0.72,
|
|
1577
1502
|
description: "Llama 3.3 Instruct (70B)"
|
|
@@ -1581,7 +1506,6 @@ var bedrockModels = {
|
|
|
1581
1506
|
contextWindow: 128e3,
|
|
1582
1507
|
supportsImages: true,
|
|
1583
1508
|
supportsPromptCache: false,
|
|
1584
|
-
supportsNativeTools: true,
|
|
1585
1509
|
inputPrice: 0.72,
|
|
1586
1510
|
outputPrice: 0.72,
|
|
1587
1511
|
description: "Llama 3.2 Instruct (90B)"
|
|
@@ -1591,7 +1515,6 @@ var bedrockModels = {
|
|
|
1591
1515
|
contextWindow: 128e3,
|
|
1592
1516
|
supportsImages: true,
|
|
1593
1517
|
supportsPromptCache: false,
|
|
1594
|
-
supportsNativeTools: true,
|
|
1595
1518
|
inputPrice: 0.16,
|
|
1596
1519
|
outputPrice: 0.16,
|
|
1597
1520
|
description: "Llama 3.2 Instruct (11B)"
|
|
@@ -1601,7 +1524,6 @@ var bedrockModels = {
|
|
|
1601
1524
|
contextWindow: 128e3,
|
|
1602
1525
|
supportsImages: false,
|
|
1603
1526
|
supportsPromptCache: false,
|
|
1604
|
-
supportsNativeTools: true,
|
|
1605
1527
|
inputPrice: 0.15,
|
|
1606
1528
|
outputPrice: 0.15,
|
|
1607
1529
|
description: "Llama 3.2 Instruct (3B)"
|
|
@@ -1611,7 +1533,6 @@ var bedrockModels = {
|
|
|
1611
1533
|
contextWindow: 128e3,
|
|
1612
1534
|
supportsImages: false,
|
|
1613
1535
|
supportsPromptCache: false,
|
|
1614
|
-
supportsNativeTools: true,
|
|
1615
1536
|
inputPrice: 0.1,
|
|
1616
1537
|
outputPrice: 0.1,
|
|
1617
1538
|
description: "Llama 3.2 Instruct (1B)"
|
|
@@ -1621,7 +1542,6 @@ var bedrockModels = {
|
|
|
1621
1542
|
contextWindow: 128e3,
|
|
1622
1543
|
supportsImages: false,
|
|
1623
1544
|
supportsPromptCache: false,
|
|
1624
|
-
supportsNativeTools: true,
|
|
1625
1545
|
inputPrice: 2.4,
|
|
1626
1546
|
outputPrice: 2.4,
|
|
1627
1547
|
description: "Llama 3.1 Instruct (405B)"
|
|
@@ -1631,7 +1551,6 @@ var bedrockModels = {
|
|
|
1631
1551
|
contextWindow: 128e3,
|
|
1632
1552
|
supportsImages: false,
|
|
1633
1553
|
supportsPromptCache: false,
|
|
1634
|
-
supportsNativeTools: true,
|
|
1635
1554
|
inputPrice: 0.72,
|
|
1636
1555
|
outputPrice: 0.72,
|
|
1637
1556
|
description: "Llama 3.1 Instruct (70B)"
|
|
@@ -1641,7 +1560,6 @@ var bedrockModels = {
|
|
|
1641
1560
|
contextWindow: 128e3,
|
|
1642
1561
|
supportsImages: false,
|
|
1643
1562
|
supportsPromptCache: false,
|
|
1644
|
-
supportsNativeTools: true,
|
|
1645
1563
|
inputPrice: 0.9,
|
|
1646
1564
|
outputPrice: 0.9,
|
|
1647
1565
|
description: "Llama 3.1 Instruct (70B) (w/ latency optimized inference)"
|
|
@@ -1651,7 +1569,6 @@ var bedrockModels = {
|
|
|
1651
1569
|
contextWindow: 8e3,
|
|
1652
1570
|
supportsImages: false,
|
|
1653
1571
|
supportsPromptCache: false,
|
|
1654
|
-
supportsNativeTools: true,
|
|
1655
1572
|
inputPrice: 0.22,
|
|
1656
1573
|
outputPrice: 0.22,
|
|
1657
1574
|
description: "Llama 3.1 Instruct (8B)"
|
|
@@ -1661,7 +1578,6 @@ var bedrockModels = {
|
|
|
1661
1578
|
contextWindow: 8e3,
|
|
1662
1579
|
supportsImages: false,
|
|
1663
1580
|
supportsPromptCache: false,
|
|
1664
|
-
supportsNativeTools: true,
|
|
1665
1581
|
inputPrice: 2.65,
|
|
1666
1582
|
outputPrice: 3.5
|
|
1667
1583
|
},
|
|
@@ -1670,7 +1586,6 @@ var bedrockModels = {
|
|
|
1670
1586
|
contextWindow: 4e3,
|
|
1671
1587
|
supportsImages: false,
|
|
1672
1588
|
supportsPromptCache: false,
|
|
1673
|
-
supportsNativeTools: true,
|
|
1674
1589
|
inputPrice: 0.3,
|
|
1675
1590
|
outputPrice: 0.6
|
|
1676
1591
|
},
|
|
@@ -1679,7 +1594,6 @@ var bedrockModels = {
|
|
|
1679
1594
|
contextWindow: 8e3,
|
|
1680
1595
|
supportsImages: false,
|
|
1681
1596
|
supportsPromptCache: false,
|
|
1682
|
-
supportsNativeTools: true,
|
|
1683
1597
|
inputPrice: 0.15,
|
|
1684
1598
|
outputPrice: 0.2,
|
|
1685
1599
|
description: "Amazon Titan Text Lite"
|
|
@@ -1689,7 +1603,6 @@ var bedrockModels = {
|
|
|
1689
1603
|
contextWindow: 8e3,
|
|
1690
1604
|
supportsImages: false,
|
|
1691
1605
|
supportsPromptCache: false,
|
|
1692
|
-
supportsNativeTools: true,
|
|
1693
1606
|
inputPrice: 0.2,
|
|
1694
1607
|
outputPrice: 0.6,
|
|
1695
1608
|
description: "Amazon Titan Text Express"
|
|
@@ -1699,8 +1612,6 @@ var bedrockModels = {
|
|
|
1699
1612
|
contextWindow: 262144,
|
|
1700
1613
|
supportsImages: false,
|
|
1701
1614
|
supportsPromptCache: false,
|
|
1702
|
-
supportsNativeTools: true,
|
|
1703
|
-
defaultToolProtocol: "native",
|
|
1704
1615
|
preserveReasoning: true,
|
|
1705
1616
|
inputPrice: 0.6,
|
|
1706
1617
|
outputPrice: 2.5,
|
|
@@ -1711,8 +1622,6 @@ var bedrockModels = {
|
|
|
1711
1622
|
contextWindow: 196608,
|
|
1712
1623
|
supportsImages: false,
|
|
1713
1624
|
supportsPromptCache: false,
|
|
1714
|
-
supportsNativeTools: true,
|
|
1715
|
-
defaultToolProtocol: "native",
|
|
1716
1625
|
preserveReasoning: true,
|
|
1717
1626
|
inputPrice: 0.3,
|
|
1718
1627
|
outputPrice: 1.2,
|
|
@@ -1723,8 +1632,6 @@ var bedrockModels = {
|
|
|
1723
1632
|
contextWindow: 262144,
|
|
1724
1633
|
supportsImages: false,
|
|
1725
1634
|
supportsPromptCache: false,
|
|
1726
|
-
supportsNativeTools: true,
|
|
1727
|
-
defaultToolProtocol: "native",
|
|
1728
1635
|
inputPrice: 0.15,
|
|
1729
1636
|
outputPrice: 1.2,
|
|
1730
1637
|
description: "Qwen3 Next 80B (MoE model with 3B active parameters)"
|
|
@@ -1734,8 +1641,6 @@ var bedrockModels = {
|
|
|
1734
1641
|
contextWindow: 262144,
|
|
1735
1642
|
supportsImages: false,
|
|
1736
1643
|
supportsPromptCache: false,
|
|
1737
|
-
supportsNativeTools: true,
|
|
1738
|
-
defaultToolProtocol: "native",
|
|
1739
1644
|
inputPrice: 0.45,
|
|
1740
1645
|
outputPrice: 1.8,
|
|
1741
1646
|
description: "Qwen3 Coder 480B (MoE model with 35B active parameters)"
|
|
@@ -1826,26 +1731,14 @@ var BEDROCK_SERVICE_TIER_PRICING = {
|
|
|
1826
1731
|
// src/providers/cerebras.ts
|
|
1827
1732
|
var cerebrasDefaultModelId = "gpt-oss-120b";
|
|
1828
1733
|
var cerebrasModels = {
|
|
1829
|
-
"zai-glm-4.6": {
|
|
1830
|
-
maxTokens: 16384,
|
|
1831
|
-
// Conservative default to avoid premature rate limiting (Cerebras reserves quota upfront)
|
|
1832
|
-
contextWindow: 131072,
|
|
1833
|
-
supportsImages: false,
|
|
1834
|
-
supportsPromptCache: false,
|
|
1835
|
-
supportsNativeTools: true,
|
|
1836
|
-
defaultToolProtocol: "native",
|
|
1837
|
-
inputPrice: 0,
|
|
1838
|
-
outputPrice: 0,
|
|
1839
|
-
description: "Fast general-purpose model on Cerebras (up to 1,000 tokens/s). To be deprecated soon."
|
|
1840
|
-
},
|
|
1841
1734
|
"zai-glm-4.7": {
|
|
1842
1735
|
maxTokens: 16384,
|
|
1843
1736
|
// Conservative default to avoid premature rate limiting (Cerebras reserves quota upfront)
|
|
1844
1737
|
contextWindow: 131072,
|
|
1845
1738
|
supportsImages: false,
|
|
1846
|
-
supportsPromptCache:
|
|
1847
|
-
|
|
1848
|
-
|
|
1739
|
+
supportsPromptCache: true,
|
|
1740
|
+
supportsTemperature: true,
|
|
1741
|
+
defaultTemperature: 1,
|
|
1849
1742
|
inputPrice: 0,
|
|
1850
1743
|
outputPrice: 0,
|
|
1851
1744
|
description: "Highly capable general-purpose model on Cerebras (up to 1,000 tokens/s), competitive with leading proprietary models on coding tasks."
|
|
@@ -1856,8 +1749,6 @@ var cerebrasModels = {
|
|
|
1856
1749
|
contextWindow: 64e3,
|
|
1857
1750
|
supportsImages: false,
|
|
1858
1751
|
supportsPromptCache: false,
|
|
1859
|
-
supportsNativeTools: true,
|
|
1860
|
-
defaultToolProtocol: "native",
|
|
1861
1752
|
inputPrice: 0,
|
|
1862
1753
|
outputPrice: 0,
|
|
1863
1754
|
description: "Intelligent model with ~1400 tokens/s"
|
|
@@ -1868,8 +1759,6 @@ var cerebrasModels = {
|
|
|
1868
1759
|
contextWindow: 64e3,
|
|
1869
1760
|
supportsImages: false,
|
|
1870
1761
|
supportsPromptCache: false,
|
|
1871
|
-
supportsNativeTools: true,
|
|
1872
|
-
defaultToolProtocol: "native",
|
|
1873
1762
|
inputPrice: 0,
|
|
1874
1763
|
outputPrice: 0,
|
|
1875
1764
|
description: "Powerful model with ~2600 tokens/s"
|
|
@@ -1880,8 +1769,6 @@ var cerebrasModels = {
|
|
|
1880
1769
|
contextWindow: 64e3,
|
|
1881
1770
|
supportsImages: false,
|
|
1882
1771
|
supportsPromptCache: false,
|
|
1883
|
-
supportsNativeTools: true,
|
|
1884
|
-
defaultToolProtocol: "native",
|
|
1885
1772
|
inputPrice: 0,
|
|
1886
1773
|
outputPrice: 0,
|
|
1887
1774
|
description: "SOTA coding performance with ~2500 tokens/s"
|
|
@@ -1892,8 +1779,6 @@ var cerebrasModels = {
|
|
|
1892
1779
|
contextWindow: 64e3,
|
|
1893
1780
|
supportsImages: false,
|
|
1894
1781
|
supportsPromptCache: false,
|
|
1895
|
-
supportsNativeTools: true,
|
|
1896
|
-
defaultToolProtocol: "native",
|
|
1897
1782
|
inputPrice: 0,
|
|
1898
1783
|
outputPrice: 0,
|
|
1899
1784
|
description: "OpenAI GPT OSS model with ~2800 tokens/s\n\n\u2022 64K context window\n\u2022 Excels at efficient reasoning across science, math, and coding"
|
|
@@ -1908,8 +1793,6 @@ var chutesModels = {
|
|
|
1908
1793
|
contextWindow: 163840,
|
|
1909
1794
|
supportsImages: false,
|
|
1910
1795
|
supportsPromptCache: false,
|
|
1911
|
-
supportsNativeTools: true,
|
|
1912
|
-
defaultToolProtocol: "native",
|
|
1913
1796
|
inputPrice: 0,
|
|
1914
1797
|
outputPrice: 0,
|
|
1915
1798
|
description: "DeepSeek R1 0528 model."
|
|
@@ -1919,8 +1802,6 @@ var chutesModels = {
|
|
|
1919
1802
|
contextWindow: 163840,
|
|
1920
1803
|
supportsImages: false,
|
|
1921
1804
|
supportsPromptCache: false,
|
|
1922
|
-
supportsNativeTools: true,
|
|
1923
|
-
defaultToolProtocol: "native",
|
|
1924
1805
|
inputPrice: 0,
|
|
1925
1806
|
outputPrice: 0,
|
|
1926
1807
|
description: "DeepSeek R1 model."
|
|
@@ -1930,8 +1811,6 @@ var chutesModels = {
|
|
|
1930
1811
|
contextWindow: 163840,
|
|
1931
1812
|
supportsImages: false,
|
|
1932
1813
|
supportsPromptCache: false,
|
|
1933
|
-
supportsNativeTools: true,
|
|
1934
|
-
defaultToolProtocol: "native",
|
|
1935
1814
|
inputPrice: 0,
|
|
1936
1815
|
outputPrice: 0,
|
|
1937
1816
|
description: "DeepSeek V3 model."
|
|
@@ -1941,8 +1820,6 @@ var chutesModels = {
|
|
|
1941
1820
|
contextWindow: 163840,
|
|
1942
1821
|
supportsImages: false,
|
|
1943
1822
|
supportsPromptCache: false,
|
|
1944
|
-
supportsNativeTools: true,
|
|
1945
|
-
defaultToolProtocol: "native",
|
|
1946
1823
|
inputPrice: 0,
|
|
1947
1824
|
outputPrice: 0,
|
|
1948
1825
|
description: "DeepSeek V3.1 model."
|
|
@@ -1952,8 +1829,6 @@ var chutesModels = {
|
|
|
1952
1829
|
contextWindow: 163840,
|
|
1953
1830
|
supportsImages: false,
|
|
1954
1831
|
supportsPromptCache: false,
|
|
1955
|
-
supportsNativeTools: true,
|
|
1956
|
-
defaultToolProtocol: "native",
|
|
1957
1832
|
inputPrice: 0.23,
|
|
1958
1833
|
outputPrice: 0.9,
|
|
1959
1834
|
description: "DeepSeek\u2011V3.1\u2011Terminus is an update to V3.1 that improves language consistency by reducing CN/EN mix\u2011ups and eliminating random characters, while strengthening agent capabilities with notably better Code Agent and Search Agent performance."
|
|
@@ -1963,8 +1838,6 @@ var chutesModels = {
|
|
|
1963
1838
|
contextWindow: 163840,
|
|
1964
1839
|
supportsImages: false,
|
|
1965
1840
|
supportsPromptCache: false,
|
|
1966
|
-
supportsNativeTools: true,
|
|
1967
|
-
defaultToolProtocol: "native",
|
|
1968
1841
|
inputPrice: 1,
|
|
1969
1842
|
outputPrice: 3,
|
|
1970
1843
|
description: "DeepSeek-V3.1-turbo is an FP8, speculative-decoding turbo variant optimized for ultra-fast single-shot queries (~200 TPS), with outputs close to the originals and solid function calling/reasoning/structured output, priced at $1/M input and $3/M output tokens, using 2\xD7 quota per request and not intended for bulk workloads."
|
|
@@ -1974,8 +1847,6 @@ var chutesModels = {
|
|
|
1974
1847
|
contextWindow: 163840,
|
|
1975
1848
|
supportsImages: false,
|
|
1976
1849
|
supportsPromptCache: false,
|
|
1977
|
-
supportsNativeTools: true,
|
|
1978
|
-
defaultToolProtocol: "native",
|
|
1979
1850
|
inputPrice: 0.25,
|
|
1980
1851
|
outputPrice: 0.35,
|
|
1981
1852
|
description: "DeepSeek-V3.2-Exp is an experimental LLM that introduces DeepSeek Sparse Attention to improve long\u2011context training and inference efficiency while maintaining performance comparable to V3.1\u2011Terminus."
|
|
@@ -1987,8 +1858,6 @@ var chutesModels = {
|
|
|
1987
1858
|
// From Groq
|
|
1988
1859
|
supportsImages: false,
|
|
1989
1860
|
supportsPromptCache: false,
|
|
1990
|
-
supportsNativeTools: true,
|
|
1991
|
-
defaultToolProtocol: "native",
|
|
1992
1861
|
inputPrice: 0,
|
|
1993
1862
|
outputPrice: 0,
|
|
1994
1863
|
description: "Unsloth Llama 3.3 70B Instruct model."
|
|
@@ -1998,8 +1867,6 @@ var chutesModels = {
|
|
|
1998
1867
|
contextWindow: 512e3,
|
|
1999
1868
|
supportsImages: false,
|
|
2000
1869
|
supportsPromptCache: false,
|
|
2001
|
-
supportsNativeTools: true,
|
|
2002
|
-
defaultToolProtocol: "native",
|
|
2003
1870
|
inputPrice: 0,
|
|
2004
1871
|
outputPrice: 0,
|
|
2005
1872
|
description: "ChutesAI Llama 4 Scout 17B Instruct model, 512K context."
|
|
@@ -2009,8 +1876,6 @@ var chutesModels = {
|
|
|
2009
1876
|
contextWindow: 128e3,
|
|
2010
1877
|
supportsImages: false,
|
|
2011
1878
|
supportsPromptCache: false,
|
|
2012
|
-
supportsNativeTools: true,
|
|
2013
|
-
defaultToolProtocol: "native",
|
|
2014
1879
|
inputPrice: 0,
|
|
2015
1880
|
outputPrice: 0,
|
|
2016
1881
|
description: "Unsloth Mistral Nemo Instruct model."
|
|
@@ -2020,8 +1885,6 @@ var chutesModels = {
|
|
|
2020
1885
|
contextWindow: 131072,
|
|
2021
1886
|
supportsImages: false,
|
|
2022
1887
|
supportsPromptCache: false,
|
|
2023
|
-
supportsNativeTools: true,
|
|
2024
|
-
defaultToolProtocol: "native",
|
|
2025
1888
|
inputPrice: 0,
|
|
2026
1889
|
outputPrice: 0,
|
|
2027
1890
|
description: "Unsloth Gemma 3 12B IT model."
|
|
@@ -2031,8 +1894,6 @@ var chutesModels = {
|
|
|
2031
1894
|
contextWindow: 131072,
|
|
2032
1895
|
supportsImages: false,
|
|
2033
1896
|
supportsPromptCache: false,
|
|
2034
|
-
supportsNativeTools: true,
|
|
2035
|
-
defaultToolProtocol: "native",
|
|
2036
1897
|
inputPrice: 0,
|
|
2037
1898
|
outputPrice: 0,
|
|
2038
1899
|
description: "Nous DeepHermes 3 Llama 3 8B Preview model."
|
|
@@ -2042,8 +1903,6 @@ var chutesModels = {
|
|
|
2042
1903
|
contextWindow: 131072,
|
|
2043
1904
|
supportsImages: false,
|
|
2044
1905
|
supportsPromptCache: false,
|
|
2045
|
-
supportsNativeTools: true,
|
|
2046
|
-
defaultToolProtocol: "native",
|
|
2047
1906
|
inputPrice: 0,
|
|
2048
1907
|
outputPrice: 0,
|
|
2049
1908
|
description: "Unsloth Gemma 3 4B IT model."
|
|
@@ -2053,8 +1912,6 @@ var chutesModels = {
|
|
|
2053
1912
|
contextWindow: 131072,
|
|
2054
1913
|
supportsImages: false,
|
|
2055
1914
|
supportsPromptCache: false,
|
|
2056
|
-
supportsNativeTools: true,
|
|
2057
|
-
defaultToolProtocol: "native",
|
|
2058
1915
|
inputPrice: 0,
|
|
2059
1916
|
outputPrice: 0,
|
|
2060
1917
|
description: "Nvidia Llama 3.3 Nemotron Super 49B model."
|
|
@@ -2064,8 +1921,6 @@ var chutesModels = {
|
|
|
2064
1921
|
contextWindow: 131072,
|
|
2065
1922
|
supportsImages: false,
|
|
2066
1923
|
supportsPromptCache: false,
|
|
2067
|
-
supportsNativeTools: true,
|
|
2068
|
-
defaultToolProtocol: "native",
|
|
2069
1924
|
inputPrice: 0,
|
|
2070
1925
|
outputPrice: 0,
|
|
2071
1926
|
description: "Nvidia Llama 3.1 Nemotron Ultra 253B model."
|
|
@@ -2075,8 +1930,6 @@ var chutesModels = {
|
|
|
2075
1930
|
contextWindow: 256e3,
|
|
2076
1931
|
supportsImages: false,
|
|
2077
1932
|
supportsPromptCache: false,
|
|
2078
|
-
supportsNativeTools: true,
|
|
2079
|
-
defaultToolProtocol: "native",
|
|
2080
1933
|
inputPrice: 0,
|
|
2081
1934
|
outputPrice: 0,
|
|
2082
1935
|
description: "ChutesAI Llama 4 Maverick 17B Instruct FP8 model."
|
|
@@ -2086,8 +1939,6 @@ var chutesModels = {
|
|
|
2086
1939
|
contextWindow: 163840,
|
|
2087
1940
|
supportsImages: false,
|
|
2088
1941
|
supportsPromptCache: false,
|
|
2089
|
-
supportsNativeTools: true,
|
|
2090
|
-
defaultToolProtocol: "native",
|
|
2091
1942
|
inputPrice: 0,
|
|
2092
1943
|
outputPrice: 0,
|
|
2093
1944
|
description: "DeepSeek V3 Base model."
|
|
@@ -2097,8 +1948,6 @@ var chutesModels = {
|
|
|
2097
1948
|
contextWindow: 163840,
|
|
2098
1949
|
supportsImages: false,
|
|
2099
1950
|
supportsPromptCache: false,
|
|
2100
|
-
supportsNativeTools: true,
|
|
2101
|
-
defaultToolProtocol: "native",
|
|
2102
1951
|
inputPrice: 0,
|
|
2103
1952
|
outputPrice: 0,
|
|
2104
1953
|
description: "DeepSeek R1 Zero model."
|
|
@@ -2108,8 +1957,6 @@ var chutesModels = {
|
|
|
2108
1957
|
contextWindow: 163840,
|
|
2109
1958
|
supportsImages: false,
|
|
2110
1959
|
supportsPromptCache: false,
|
|
2111
|
-
supportsNativeTools: true,
|
|
2112
|
-
defaultToolProtocol: "native",
|
|
2113
1960
|
inputPrice: 0,
|
|
2114
1961
|
outputPrice: 0,
|
|
2115
1962
|
description: "DeepSeek V3 (0324) model."
|
|
@@ -2119,8 +1966,6 @@ var chutesModels = {
|
|
|
2119
1966
|
contextWindow: 262144,
|
|
2120
1967
|
supportsImages: false,
|
|
2121
1968
|
supportsPromptCache: false,
|
|
2122
|
-
supportsNativeTools: true,
|
|
2123
|
-
defaultToolProtocol: "native",
|
|
2124
1969
|
inputPrice: 0,
|
|
2125
1970
|
outputPrice: 0,
|
|
2126
1971
|
description: "Qwen3 235B A22B Instruct 2507 model with 262K context window."
|
|
@@ -2130,8 +1975,6 @@ var chutesModels = {
|
|
|
2130
1975
|
contextWindow: 40960,
|
|
2131
1976
|
supportsImages: false,
|
|
2132
1977
|
supportsPromptCache: false,
|
|
2133
|
-
supportsNativeTools: true,
|
|
2134
|
-
defaultToolProtocol: "native",
|
|
2135
1978
|
inputPrice: 0,
|
|
2136
1979
|
outputPrice: 0,
|
|
2137
1980
|
description: "Qwen3 235B A22B model."
|
|
@@ -2141,8 +1984,6 @@ var chutesModels = {
|
|
|
2141
1984
|
contextWindow: 40960,
|
|
2142
1985
|
supportsImages: false,
|
|
2143
1986
|
supportsPromptCache: false,
|
|
2144
|
-
supportsNativeTools: true,
|
|
2145
|
-
defaultToolProtocol: "native",
|
|
2146
1987
|
inputPrice: 0,
|
|
2147
1988
|
outputPrice: 0,
|
|
2148
1989
|
description: "Qwen3 32B model."
|
|
@@ -2152,8 +1993,6 @@ var chutesModels = {
|
|
|
2152
1993
|
contextWindow: 40960,
|
|
2153
1994
|
supportsImages: false,
|
|
2154
1995
|
supportsPromptCache: false,
|
|
2155
|
-
supportsNativeTools: true,
|
|
2156
|
-
defaultToolProtocol: "native",
|
|
2157
1996
|
inputPrice: 0,
|
|
2158
1997
|
outputPrice: 0,
|
|
2159
1998
|
description: "Qwen3 30B A3B model."
|
|
@@ -2163,8 +2002,6 @@ var chutesModels = {
|
|
|
2163
2002
|
contextWindow: 40960,
|
|
2164
2003
|
supportsImages: false,
|
|
2165
2004
|
supportsPromptCache: false,
|
|
2166
|
-
supportsNativeTools: true,
|
|
2167
|
-
defaultToolProtocol: "native",
|
|
2168
2005
|
inputPrice: 0,
|
|
2169
2006
|
outputPrice: 0,
|
|
2170
2007
|
description: "Qwen3 14B model."
|
|
@@ -2174,8 +2011,6 @@ var chutesModels = {
|
|
|
2174
2011
|
contextWindow: 40960,
|
|
2175
2012
|
supportsImages: false,
|
|
2176
2013
|
supportsPromptCache: false,
|
|
2177
|
-
supportsNativeTools: true,
|
|
2178
|
-
defaultToolProtocol: "native",
|
|
2179
2014
|
inputPrice: 0,
|
|
2180
2015
|
outputPrice: 0,
|
|
2181
2016
|
description: "Qwen3 8B model."
|
|
@@ -2185,8 +2020,6 @@ var chutesModels = {
|
|
|
2185
2020
|
contextWindow: 163840,
|
|
2186
2021
|
supportsImages: false,
|
|
2187
2022
|
supportsPromptCache: false,
|
|
2188
|
-
supportsNativeTools: true,
|
|
2189
|
-
defaultToolProtocol: "native",
|
|
2190
2023
|
inputPrice: 0,
|
|
2191
2024
|
outputPrice: 0,
|
|
2192
2025
|
description: "Microsoft MAI-DS-R1 FP8 model."
|
|
@@ -2196,8 +2029,6 @@ var chutesModels = {
|
|
|
2196
2029
|
contextWindow: 163840,
|
|
2197
2030
|
supportsImages: false,
|
|
2198
2031
|
supportsPromptCache: false,
|
|
2199
|
-
supportsNativeTools: true,
|
|
2200
|
-
defaultToolProtocol: "native",
|
|
2201
2032
|
inputPrice: 0,
|
|
2202
2033
|
outputPrice: 0,
|
|
2203
2034
|
description: "TNGTech DeepSeek R1T Chimera model."
|
|
@@ -2207,8 +2038,6 @@ var chutesModels = {
|
|
|
2207
2038
|
contextWindow: 151329,
|
|
2208
2039
|
supportsImages: false,
|
|
2209
2040
|
supportsPromptCache: false,
|
|
2210
|
-
supportsNativeTools: true,
|
|
2211
|
-
defaultToolProtocol: "native",
|
|
2212
2041
|
inputPrice: 0,
|
|
2213
2042
|
outputPrice: 0,
|
|
2214
2043
|
description: "GLM-4.5-Air model with 151,329 token context window and 106B total parameters with 12B activated."
|
|
@@ -2218,8 +2047,6 @@ var chutesModels = {
|
|
|
2218
2047
|
contextWindow: 131072,
|
|
2219
2048
|
supportsImages: false,
|
|
2220
2049
|
supportsPromptCache: false,
|
|
2221
|
-
supportsNativeTools: true,
|
|
2222
|
-
defaultToolProtocol: "native",
|
|
2223
2050
|
inputPrice: 0,
|
|
2224
2051
|
outputPrice: 0,
|
|
2225
2052
|
description: "GLM-4.5-FP8 model with 128k token context window, optimized for agent-based applications with MoE architecture."
|
|
@@ -2229,8 +2056,6 @@ var chutesModels = {
|
|
|
2229
2056
|
contextWindow: 131072,
|
|
2230
2057
|
supportsImages: false,
|
|
2231
2058
|
supportsPromptCache: false,
|
|
2232
|
-
supportsNativeTools: true,
|
|
2233
|
-
defaultToolProtocol: "native",
|
|
2234
2059
|
inputPrice: 1,
|
|
2235
2060
|
outputPrice: 3,
|
|
2236
2061
|
description: "GLM-4.5-turbo model with 128K token context window, optimized for fast inference."
|
|
@@ -2240,8 +2065,6 @@ var chutesModels = {
|
|
|
2240
2065
|
contextWindow: 202752,
|
|
2241
2066
|
supportsImages: false,
|
|
2242
2067
|
supportsPromptCache: false,
|
|
2243
|
-
supportsNativeTools: true,
|
|
2244
|
-
defaultToolProtocol: "native",
|
|
2245
2068
|
inputPrice: 0,
|
|
2246
2069
|
outputPrice: 0,
|
|
2247
2070
|
description: "GLM-4.6 introduces major upgrades over GLM-4.5, including a longer 200K-token context window for complex tasks, stronger coding performance in benchmarks and real-world tools (such as Claude Code, Cline, Roo Code, and Kilo Code), improved reasoning with tool use during inference, more capable and efficient agent integration, and refined writing that better matches human style, readability, and natural role-play scenarios."
|
|
@@ -2252,8 +2075,6 @@ var chutesModels = {
|
|
|
2252
2075
|
contextWindow: 202752,
|
|
2253
2076
|
supportsImages: false,
|
|
2254
2077
|
supportsPromptCache: false,
|
|
2255
|
-
supportsNativeTools: true,
|
|
2256
|
-
defaultToolProtocol: "native",
|
|
2257
2078
|
inputPrice: 1.15,
|
|
2258
2079
|
outputPrice: 3.25,
|
|
2259
2080
|
description: "GLM-4.6-turbo model with 200K-token context window, optimized for fast inference."
|
|
@@ -2263,8 +2084,6 @@ var chutesModels = {
|
|
|
2263
2084
|
contextWindow: 128e3,
|
|
2264
2085
|
supportsImages: false,
|
|
2265
2086
|
supportsPromptCache: false,
|
|
2266
|
-
supportsNativeTools: true,
|
|
2267
|
-
defaultToolProtocol: "native",
|
|
2268
2087
|
inputPrice: 0,
|
|
2269
2088
|
outputPrice: 0,
|
|
2270
2089
|
description: "LongCat Flash Thinking FP8 model with 128K context window, optimized for complex reasoning and coding tasks."
|
|
@@ -2274,8 +2093,6 @@ var chutesModels = {
|
|
|
2274
2093
|
contextWindow: 262144,
|
|
2275
2094
|
supportsImages: false,
|
|
2276
2095
|
supportsPromptCache: false,
|
|
2277
|
-
supportsNativeTools: true,
|
|
2278
|
-
defaultToolProtocol: "native",
|
|
2279
2096
|
inputPrice: 0,
|
|
2280
2097
|
outputPrice: 0,
|
|
2281
2098
|
description: "Qwen3 Coder 480B A35B Instruct FP8 model, optimized for coding tasks."
|
|
@@ -2285,8 +2102,6 @@ var chutesModels = {
|
|
|
2285
2102
|
contextWindow: 75e3,
|
|
2286
2103
|
supportsImages: false,
|
|
2287
2104
|
supportsPromptCache: false,
|
|
2288
|
-
supportsNativeTools: true,
|
|
2289
|
-
defaultToolProtocol: "native",
|
|
2290
2105
|
inputPrice: 0.1481,
|
|
2291
2106
|
outputPrice: 0.5926,
|
|
2292
2107
|
description: "Moonshot AI Kimi K2 Instruct model with 75k context window."
|
|
@@ -2296,8 +2111,6 @@ var chutesModels = {
|
|
|
2296
2111
|
contextWindow: 262144,
|
|
2297
2112
|
supportsImages: false,
|
|
2298
2113
|
supportsPromptCache: false,
|
|
2299
|
-
supportsNativeTools: true,
|
|
2300
|
-
defaultToolProtocol: "native",
|
|
2301
2114
|
inputPrice: 0.1999,
|
|
2302
2115
|
outputPrice: 0.8001,
|
|
2303
2116
|
description: "Moonshot AI Kimi K2 Instruct 0905 model with 256k context window."
|
|
@@ -2307,8 +2120,6 @@ var chutesModels = {
|
|
|
2307
2120
|
contextWindow: 262144,
|
|
2308
2121
|
supportsImages: false,
|
|
2309
2122
|
supportsPromptCache: false,
|
|
2310
|
-
supportsNativeTools: true,
|
|
2311
|
-
defaultToolProtocol: "native",
|
|
2312
2123
|
inputPrice: 0.077968332,
|
|
2313
2124
|
outputPrice: 0.31202496,
|
|
2314
2125
|
description: "Qwen3 235B A22B Thinking 2507 model with 262K context window."
|
|
@@ -2318,8 +2129,6 @@ var chutesModels = {
|
|
|
2318
2129
|
contextWindow: 131072,
|
|
2319
2130
|
supportsImages: false,
|
|
2320
2131
|
supportsPromptCache: false,
|
|
2321
|
-
supportsNativeTools: true,
|
|
2322
|
-
defaultToolProtocol: "native",
|
|
2323
2132
|
inputPrice: 0,
|
|
2324
2133
|
outputPrice: 0,
|
|
2325
2134
|
description: "Fast, stable instruction-tuned model optimized for complex tasks, RAG, and tool use without thinking traces."
|
|
@@ -2329,8 +2138,6 @@ var chutesModels = {
|
|
|
2329
2138
|
contextWindow: 131072,
|
|
2330
2139
|
supportsImages: false,
|
|
2331
2140
|
supportsPromptCache: false,
|
|
2332
|
-
supportsNativeTools: true,
|
|
2333
|
-
defaultToolProtocol: "native",
|
|
2334
2141
|
inputPrice: 0,
|
|
2335
2142
|
outputPrice: 0,
|
|
2336
2143
|
description: "Reasoning-first model with structured thinking traces for multi-step problems, math proofs, and code synthesis."
|
|
@@ -2340,8 +2147,6 @@ var chutesModels = {
|
|
|
2340
2147
|
contextWindow: 262144,
|
|
2341
2148
|
supportsImages: true,
|
|
2342
2149
|
supportsPromptCache: false,
|
|
2343
|
-
supportsNativeTools: true,
|
|
2344
|
-
defaultToolProtocol: "native",
|
|
2345
2150
|
inputPrice: 0.16,
|
|
2346
2151
|
outputPrice: 0.65,
|
|
2347
2152
|
description: "Qwen3\u2011VL\u2011235B\u2011A22B\u2011Thinking is an open\u2011weight MoE vision\u2011language model (235B total, ~22B activated) optimized for deliberate multi\u2011step reasoning with strong text\u2011image\u2011video understanding and long\u2011context capabilities."
|
|
@@ -2349,73 +2154,6 @@ var chutesModels = {
|
|
|
2349
2154
|
};
|
|
2350
2155
|
var chutesDefaultModelInfo = chutesModels[chutesDefaultModelId];
|
|
2351
2156
|
|
|
2352
|
-
// src/providers/claude-code.ts
|
|
2353
|
-
var DATE_SUFFIX_PATTERN = /-\d{8}$/;
|
|
2354
|
-
var claudeCodeModels = {
|
|
2355
|
-
"claude-haiku-4-5": {
|
|
2356
|
-
maxTokens: 32768,
|
|
2357
|
-
contextWindow: 2e5,
|
|
2358
|
-
supportsImages: true,
|
|
2359
|
-
supportsPromptCache: true,
|
|
2360
|
-
supportsNativeTools: true,
|
|
2361
|
-
defaultToolProtocol: "native",
|
|
2362
|
-
supportsReasoningEffort: ["disable", "low", "medium", "high"],
|
|
2363
|
-
reasoningEffort: "medium",
|
|
2364
|
-
description: "Claude Haiku 4.5 - Fast and efficient with thinking"
|
|
2365
|
-
},
|
|
2366
|
-
"claude-sonnet-4-5": {
|
|
2367
|
-
maxTokens: 32768,
|
|
2368
|
-
contextWindow: 2e5,
|
|
2369
|
-
supportsImages: true,
|
|
2370
|
-
supportsPromptCache: true,
|
|
2371
|
-
supportsNativeTools: true,
|
|
2372
|
-
defaultToolProtocol: "native",
|
|
2373
|
-
supportsReasoningEffort: ["disable", "low", "medium", "high"],
|
|
2374
|
-
reasoningEffort: "medium",
|
|
2375
|
-
description: "Claude Sonnet 4.5 - Balanced performance with thinking"
|
|
2376
|
-
},
|
|
2377
|
-
"claude-opus-4-5": {
|
|
2378
|
-
maxTokens: 32768,
|
|
2379
|
-
contextWindow: 2e5,
|
|
2380
|
-
supportsImages: true,
|
|
2381
|
-
supportsPromptCache: true,
|
|
2382
|
-
supportsNativeTools: true,
|
|
2383
|
-
defaultToolProtocol: "native",
|
|
2384
|
-
supportsReasoningEffort: ["disable", "low", "medium", "high"],
|
|
2385
|
-
reasoningEffort: "medium",
|
|
2386
|
-
description: "Claude Opus 4.5 - Most capable with thinking"
|
|
2387
|
-
}
|
|
2388
|
-
};
|
|
2389
|
-
var claudeCodeDefaultModelId = "claude-sonnet-4-5";
|
|
2390
|
-
var MODEL_FAMILY_PATTERNS = [
|
|
2391
|
-
// Opus models (any version) → claude-opus-4-5
|
|
2392
|
-
{ pattern: /opus/i, target: "claude-opus-4-5" },
|
|
2393
|
-
// Haiku models (any version) → claude-haiku-4-5
|
|
2394
|
-
{ pattern: /haiku/i, target: "claude-haiku-4-5" },
|
|
2395
|
-
// Sonnet models (any version) → claude-sonnet-4-5
|
|
2396
|
-
{ pattern: /sonnet/i, target: "claude-sonnet-4-5" }
|
|
2397
|
-
];
|
|
2398
|
-
function normalizeClaudeCodeModelId(modelId) {
|
|
2399
|
-
if (Object.hasOwn(claudeCodeModels, modelId)) {
|
|
2400
|
-
return modelId;
|
|
2401
|
-
}
|
|
2402
|
-
const withoutDate = modelId.replace(DATE_SUFFIX_PATTERN, "");
|
|
2403
|
-
if (Object.hasOwn(claudeCodeModels, withoutDate)) {
|
|
2404
|
-
return withoutDate;
|
|
2405
|
-
}
|
|
2406
|
-
for (const { pattern, target } of MODEL_FAMILY_PATTERNS) {
|
|
2407
|
-
if (pattern.test(modelId)) {
|
|
2408
|
-
return target;
|
|
2409
|
-
}
|
|
2410
|
-
}
|
|
2411
|
-
return claudeCodeDefaultModelId;
|
|
2412
|
-
}
|
|
2413
|
-
var claudeCodeReasoningConfig = {
|
|
2414
|
-
low: { budgetTokens: 16e3 },
|
|
2415
|
-
medium: { budgetTokens: 32e3 },
|
|
2416
|
-
high: { budgetTokens: 64e3 }
|
|
2417
|
-
};
|
|
2418
|
-
|
|
2419
2157
|
// src/providers/deepseek.ts
|
|
2420
2158
|
var deepSeekDefaultModelId = "deepseek-chat";
|
|
2421
2159
|
var deepSeekModels = {
|
|
@@ -2425,8 +2163,6 @@ var deepSeekModels = {
|
|
|
2425
2163
|
contextWindow: 128e3,
|
|
2426
2164
|
supportsImages: false,
|
|
2427
2165
|
supportsPromptCache: true,
|
|
2428
|
-
supportsNativeTools: true,
|
|
2429
|
-
defaultToolProtocol: "native",
|
|
2430
2166
|
inputPrice: 0.28,
|
|
2431
2167
|
// $0.28 per million tokens (cache miss) - Updated Dec 9, 2025
|
|
2432
2168
|
outputPrice: 0.42,
|
|
@@ -2443,8 +2179,6 @@ var deepSeekModels = {
|
|
|
2443
2179
|
contextWindow: 128e3,
|
|
2444
2180
|
supportsImages: false,
|
|
2445
2181
|
supportsPromptCache: true,
|
|
2446
|
-
supportsNativeTools: true,
|
|
2447
|
-
defaultToolProtocol: "native",
|
|
2448
2182
|
preserveReasoning: true,
|
|
2449
2183
|
inputPrice: 0.28,
|
|
2450
2184
|
// $0.28 per million tokens (cache miss) - Updated Dec 9, 2025
|
|
@@ -2467,8 +2201,6 @@ var doubaoModels = {
|
|
|
2467
2201
|
contextWindow: 128e3,
|
|
2468
2202
|
supportsImages: true,
|
|
2469
2203
|
supportsPromptCache: true,
|
|
2470
|
-
supportsNativeTools: true,
|
|
2471
|
-
defaultToolProtocol: "native",
|
|
2472
2204
|
inputPrice: 1e-4,
|
|
2473
2205
|
// $0.0001 per million tokens (cache miss)
|
|
2474
2206
|
outputPrice: 4e-4,
|
|
@@ -2484,8 +2216,6 @@ var doubaoModels = {
|
|
|
2484
2216
|
contextWindow: 128e3,
|
|
2485
2217
|
supportsImages: true,
|
|
2486
2218
|
supportsPromptCache: true,
|
|
2487
|
-
supportsNativeTools: true,
|
|
2488
|
-
defaultToolProtocol: "native",
|
|
2489
2219
|
inputPrice: 2e-4,
|
|
2490
2220
|
// $0.0002 per million tokens
|
|
2491
2221
|
outputPrice: 8e-4,
|
|
@@ -2501,8 +2231,6 @@ var doubaoModels = {
|
|
|
2501
2231
|
contextWindow: 128e3,
|
|
2502
2232
|
supportsImages: true,
|
|
2503
2233
|
supportsPromptCache: true,
|
|
2504
|
-
supportsNativeTools: true,
|
|
2505
|
-
defaultToolProtocol: "native",
|
|
2506
2234
|
inputPrice: 15e-5,
|
|
2507
2235
|
// $0.00015 per million tokens
|
|
2508
2236
|
outputPrice: 6e-4,
|
|
@@ -2525,7 +2253,6 @@ var featherlessModels = {
|
|
|
2525
2253
|
contextWindow: 32678,
|
|
2526
2254
|
supportsImages: false,
|
|
2527
2255
|
supportsPromptCache: false,
|
|
2528
|
-
supportsNativeTools: true,
|
|
2529
2256
|
inputPrice: 0,
|
|
2530
2257
|
outputPrice: 0,
|
|
2531
2258
|
description: "DeepSeek V3 0324 model."
|
|
@@ -2535,7 +2262,6 @@ var featherlessModels = {
|
|
|
2535
2262
|
contextWindow: 32678,
|
|
2536
2263
|
supportsImages: false,
|
|
2537
2264
|
supportsPromptCache: false,
|
|
2538
|
-
supportsNativeTools: true,
|
|
2539
2265
|
inputPrice: 0,
|
|
2540
2266
|
outputPrice: 0,
|
|
2541
2267
|
description: "DeepSeek R1 0528 model."
|
|
@@ -2545,7 +2271,6 @@ var featherlessModels = {
|
|
|
2545
2271
|
contextWindow: 32678,
|
|
2546
2272
|
supportsImages: false,
|
|
2547
2273
|
supportsPromptCache: false,
|
|
2548
|
-
supportsNativeTools: true,
|
|
2549
2274
|
inputPrice: 0,
|
|
2550
2275
|
outputPrice: 0,
|
|
2551
2276
|
description: "Kimi K2 Instruct model."
|
|
@@ -2555,7 +2280,6 @@ var featherlessModels = {
|
|
|
2555
2280
|
contextWindow: 32678,
|
|
2556
2281
|
supportsImages: false,
|
|
2557
2282
|
supportsPromptCache: false,
|
|
2558
|
-
supportsNativeTools: true,
|
|
2559
2283
|
inputPrice: 0,
|
|
2560
2284
|
outputPrice: 0,
|
|
2561
2285
|
description: "GPT-OSS 120B model."
|
|
@@ -2565,7 +2289,6 @@ var featherlessModels = {
|
|
|
2565
2289
|
contextWindow: 32678,
|
|
2566
2290
|
supportsImages: false,
|
|
2567
2291
|
supportsPromptCache: false,
|
|
2568
|
-
supportsNativeTools: true,
|
|
2569
2292
|
inputPrice: 0,
|
|
2570
2293
|
outputPrice: 0,
|
|
2571
2294
|
description: "Qwen3 Coder 480B A35B Instruct model."
|
|
@@ -2581,8 +2304,6 @@ var fireworksModels = {
|
|
|
2581
2304
|
contextWindow: 262144,
|
|
2582
2305
|
supportsImages: false,
|
|
2583
2306
|
supportsPromptCache: true,
|
|
2584
|
-
supportsNativeTools: true,
|
|
2585
|
-
defaultToolProtocol: "native",
|
|
2586
2307
|
inputPrice: 0.6,
|
|
2587
2308
|
outputPrice: 2.5,
|
|
2588
2309
|
cacheReadsPrice: 0.15,
|
|
@@ -2593,8 +2314,6 @@ var fireworksModels = {
|
|
|
2593
2314
|
contextWindow: 128e3,
|
|
2594
2315
|
supportsImages: false,
|
|
2595
2316
|
supportsPromptCache: false,
|
|
2596
|
-
supportsNativeTools: true,
|
|
2597
|
-
defaultToolProtocol: "native",
|
|
2598
2317
|
inputPrice: 0.6,
|
|
2599
2318
|
outputPrice: 2.5,
|
|
2600
2319
|
description: "Kimi K2 is a state-of-the-art mixture-of-experts (MoE) language model with 32 billion activated parameters and 1 trillion total parameters. Trained with the Muon optimizer, Kimi K2 achieves exceptional performance across frontier knowledge, reasoning, and coding tasks while being meticulously optimized for agentic capabilities."
|
|
@@ -2604,7 +2323,6 @@ var fireworksModels = {
|
|
|
2604
2323
|
contextWindow: 256e3,
|
|
2605
2324
|
supportsImages: false,
|
|
2606
2325
|
supportsPromptCache: true,
|
|
2607
|
-
supportsNativeTools: true,
|
|
2608
2326
|
supportsTemperature: true,
|
|
2609
2327
|
preserveReasoning: true,
|
|
2610
2328
|
defaultTemperature: 1,
|
|
@@ -2618,8 +2336,6 @@ var fireworksModels = {
|
|
|
2618
2336
|
contextWindow: 204800,
|
|
2619
2337
|
supportsImages: false,
|
|
2620
2338
|
supportsPromptCache: false,
|
|
2621
|
-
supportsNativeTools: true,
|
|
2622
|
-
defaultToolProtocol: "native",
|
|
2623
2339
|
inputPrice: 0.3,
|
|
2624
2340
|
outputPrice: 1.2,
|
|
2625
2341
|
description: "MiniMax M2 is a high-performance language model with 204.8K context window, optimized for long-context understanding and generation tasks."
|
|
@@ -2629,8 +2345,6 @@ var fireworksModels = {
|
|
|
2629
2345
|
contextWindow: 256e3,
|
|
2630
2346
|
supportsImages: false,
|
|
2631
2347
|
supportsPromptCache: false,
|
|
2632
|
-
supportsNativeTools: true,
|
|
2633
|
-
defaultToolProtocol: "native",
|
|
2634
2348
|
inputPrice: 0.22,
|
|
2635
2349
|
outputPrice: 0.88,
|
|
2636
2350
|
description: "Latest Qwen3 thinking model, competitive against the best closed source models in Jul 2025."
|
|
@@ -2640,8 +2354,6 @@ var fireworksModels = {
|
|
|
2640
2354
|
contextWindow: 256e3,
|
|
2641
2355
|
supportsImages: false,
|
|
2642
2356
|
supportsPromptCache: false,
|
|
2643
|
-
supportsNativeTools: true,
|
|
2644
|
-
defaultToolProtocol: "native",
|
|
2645
2357
|
inputPrice: 0.45,
|
|
2646
2358
|
outputPrice: 1.8,
|
|
2647
2359
|
description: "Qwen3's most agentic code model to date."
|
|
@@ -2651,8 +2363,6 @@ var fireworksModels = {
|
|
|
2651
2363
|
contextWindow: 16e4,
|
|
2652
2364
|
supportsImages: false,
|
|
2653
2365
|
supportsPromptCache: false,
|
|
2654
|
-
supportsNativeTools: true,
|
|
2655
|
-
defaultToolProtocol: "native",
|
|
2656
2366
|
inputPrice: 3,
|
|
2657
2367
|
outputPrice: 8,
|
|
2658
2368
|
description: "05/28 updated checkpoint of Deepseek R1. Its overall performance is now approaching that of leading models, such as O3 and Gemini 2.5 Pro. Compared to the previous version, the upgraded model shows significant improvements in handling complex reasoning tasks, and this version also offers a reduced hallucination rate, enhanced support for function calling, and better experience for vibe coding. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
|
|
@@ -2662,8 +2372,6 @@ var fireworksModels = {
|
|
|
2662
2372
|
contextWindow: 128e3,
|
|
2663
2373
|
supportsImages: false,
|
|
2664
2374
|
supportsPromptCache: false,
|
|
2665
|
-
supportsNativeTools: true,
|
|
2666
|
-
defaultToolProtocol: "native",
|
|
2667
2375
|
inputPrice: 0.9,
|
|
2668
2376
|
outputPrice: 0.9,
|
|
2669
2377
|
description: "A strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token from Deepseek. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
|
|
@@ -2673,8 +2381,6 @@ var fireworksModels = {
|
|
|
2673
2381
|
contextWindow: 163840,
|
|
2674
2382
|
supportsImages: false,
|
|
2675
2383
|
supportsPromptCache: false,
|
|
2676
|
-
supportsNativeTools: true,
|
|
2677
|
-
defaultToolProtocol: "native",
|
|
2678
2384
|
inputPrice: 0.56,
|
|
2679
2385
|
outputPrice: 1.68,
|
|
2680
2386
|
description: "DeepSeek v3.1 is an improved version of the v3 model with enhanced performance, better reasoning capabilities, and improved code generation. This Mixture-of-Experts (MoE) model maintains the same 671B total parameters with 37B activated per token."
|
|
@@ -2684,8 +2390,6 @@ var fireworksModels = {
|
|
|
2684
2390
|
contextWindow: 128e3,
|
|
2685
2391
|
supportsImages: false,
|
|
2686
2392
|
supportsPromptCache: false,
|
|
2687
|
-
supportsNativeTools: true,
|
|
2688
|
-
defaultToolProtocol: "native",
|
|
2689
2393
|
inputPrice: 0.55,
|
|
2690
2394
|
outputPrice: 2.19,
|
|
2691
2395
|
description: "Z.ai GLM-4.5 with 355B total parameters and 32B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
|
|
@@ -2695,8 +2399,6 @@ var fireworksModels = {
|
|
|
2695
2399
|
contextWindow: 128e3,
|
|
2696
2400
|
supportsImages: false,
|
|
2697
2401
|
supportsPromptCache: false,
|
|
2698
|
-
supportsNativeTools: true,
|
|
2699
|
-
defaultToolProtocol: "native",
|
|
2700
2402
|
inputPrice: 0.55,
|
|
2701
2403
|
outputPrice: 2.19,
|
|
2702
2404
|
description: "Z.ai GLM-4.5-Air with 106B total parameters and 12B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
|
|
@@ -2706,8 +2408,6 @@ var fireworksModels = {
|
|
|
2706
2408
|
contextWindow: 198e3,
|
|
2707
2409
|
supportsImages: false,
|
|
2708
2410
|
supportsPromptCache: false,
|
|
2709
|
-
supportsNativeTools: true,
|
|
2710
|
-
defaultToolProtocol: "native",
|
|
2711
2411
|
inputPrice: 0.55,
|
|
2712
2412
|
outputPrice: 2.19,
|
|
2713
2413
|
description: "Z.ai GLM-4.6 is an advanced coding model with exceptional performance on complex programming tasks. Features improved reasoning capabilities and enhanced code generation quality, making it ideal for software development workflows."
|
|
@@ -2717,8 +2417,6 @@ var fireworksModels = {
|
|
|
2717
2417
|
contextWindow: 128e3,
|
|
2718
2418
|
supportsImages: false,
|
|
2719
2419
|
supportsPromptCache: false,
|
|
2720
|
-
supportsNativeTools: true,
|
|
2721
|
-
defaultToolProtocol: "native",
|
|
2722
2420
|
inputPrice: 0.07,
|
|
2723
2421
|
outputPrice: 0.3,
|
|
2724
2422
|
description: "OpenAI gpt-oss-20b: Compact model for local/edge deployments. Optimized for low-latency and resource-constrained environments with chain-of-thought output, adjustable reasoning, and agentic workflows."
|
|
@@ -2728,11 +2426,63 @@ var fireworksModels = {
|
|
|
2728
2426
|
contextWindow: 128e3,
|
|
2729
2427
|
supportsImages: false,
|
|
2730
2428
|
supportsPromptCache: false,
|
|
2731
|
-
supportsNativeTools: true,
|
|
2732
|
-
defaultToolProtocol: "native",
|
|
2733
2429
|
inputPrice: 0.15,
|
|
2734
2430
|
outputPrice: 0.6,
|
|
2735
2431
|
description: "OpenAI gpt-oss-120b: Production-grade, general-purpose model that fits on a single H100 GPU. Features complex reasoning, configurable effort, full chain-of-thought transparency, and supports function calling, tool use, and structured outputs."
|
|
2432
|
+
},
|
|
2433
|
+
"accounts/fireworks/models/minimax-m2p1": {
|
|
2434
|
+
maxTokens: 4096,
|
|
2435
|
+
contextWindow: 204800,
|
|
2436
|
+
supportsImages: false,
|
|
2437
|
+
supportsPromptCache: false,
|
|
2438
|
+
inputPrice: 0.3,
|
|
2439
|
+
outputPrice: 1.2,
|
|
2440
|
+
description: "MiniMax M2.1 is an upgraded version of M2 with improved performance on complex reasoning, coding, and long-context understanding tasks."
|
|
2441
|
+
},
|
|
2442
|
+
"accounts/fireworks/models/deepseek-v3p2": {
|
|
2443
|
+
maxTokens: 16384,
|
|
2444
|
+
contextWindow: 163840,
|
|
2445
|
+
supportsImages: false,
|
|
2446
|
+
supportsPromptCache: false,
|
|
2447
|
+
inputPrice: 0.56,
|
|
2448
|
+
outputPrice: 1.68,
|
|
2449
|
+
description: "DeepSeek V3.2 is the latest iteration of the V3 model family with enhanced reasoning capabilities, improved code generation, and better instruction following."
|
|
2450
|
+
},
|
|
2451
|
+
"accounts/fireworks/models/glm-4p7": {
|
|
2452
|
+
maxTokens: 25344,
|
|
2453
|
+
contextWindow: 198e3,
|
|
2454
|
+
supportsImages: false,
|
|
2455
|
+
supportsPromptCache: false,
|
|
2456
|
+
inputPrice: 0.55,
|
|
2457
|
+
outputPrice: 2.19,
|
|
2458
|
+
description: "Z.ai GLM-4.7 is the latest coding model with exceptional performance on complex programming tasks. Features improved reasoning capabilities and enhanced code generation quality."
|
|
2459
|
+
},
|
|
2460
|
+
"accounts/fireworks/models/llama-v3p3-70b-instruct": {
|
|
2461
|
+
maxTokens: 16384,
|
|
2462
|
+
contextWindow: 131072,
|
|
2463
|
+
supportsImages: false,
|
|
2464
|
+
supportsPromptCache: false,
|
|
2465
|
+
inputPrice: 0.9,
|
|
2466
|
+
outputPrice: 0.9,
|
|
2467
|
+
description: "Meta Llama 3.3 70B Instruct is a highly capable instruction-tuned model with strong reasoning, coding, and general task performance."
|
|
2468
|
+
},
|
|
2469
|
+
"accounts/fireworks/models/llama4-maverick-instruct-basic": {
|
|
2470
|
+
maxTokens: 16384,
|
|
2471
|
+
contextWindow: 131072,
|
|
2472
|
+
supportsImages: true,
|
|
2473
|
+
supportsPromptCache: false,
|
|
2474
|
+
inputPrice: 0.22,
|
|
2475
|
+
outputPrice: 0.88,
|
|
2476
|
+
description: "Llama 4 Maverick is Meta's latest multimodal model with vision capabilities, optimized for instruction following and coding tasks."
|
|
2477
|
+
},
|
|
2478
|
+
"accounts/fireworks/models/llama4-scout-instruct-basic": {
|
|
2479
|
+
maxTokens: 16384,
|
|
2480
|
+
contextWindow: 131072,
|
|
2481
|
+
supportsImages: true,
|
|
2482
|
+
supportsPromptCache: false,
|
|
2483
|
+
inputPrice: 0.15,
|
|
2484
|
+
outputPrice: 0.6,
|
|
2485
|
+
description: "Llama 4 Scout is a smaller, faster variant of Llama 4 with multimodal capabilities, ideal for quick iterations and cost-effective deployments."
|
|
2736
2486
|
}
|
|
2737
2487
|
};
|
|
2738
2488
|
|
|
@@ -2743,8 +2493,6 @@ var geminiModels = {
|
|
|
2743
2493
|
maxTokens: 65536,
|
|
2744
2494
|
contextWindow: 1048576,
|
|
2745
2495
|
supportsImages: true,
|
|
2746
|
-
supportsNativeTools: true,
|
|
2747
|
-
defaultToolProtocol: "native",
|
|
2748
2496
|
supportsPromptCache: true,
|
|
2749
2497
|
supportsReasoningEffort: ["low", "high"],
|
|
2750
2498
|
reasoningEffort: "low",
|
|
@@ -2752,16 +2500,19 @@ var geminiModels = {
|
|
|
2752
2500
|
defaultTemperature: 1,
|
|
2753
2501
|
inputPrice: 4,
|
|
2754
2502
|
outputPrice: 18,
|
|
2503
|
+
cacheReadsPrice: 0.4,
|
|
2755
2504
|
tiers: [
|
|
2756
2505
|
{
|
|
2757
2506
|
contextWindow: 2e5,
|
|
2758
2507
|
inputPrice: 2,
|
|
2759
|
-
outputPrice: 12
|
|
2508
|
+
outputPrice: 12,
|
|
2509
|
+
cacheReadsPrice: 0.2
|
|
2760
2510
|
},
|
|
2761
2511
|
{
|
|
2762
2512
|
contextWindow: Infinity,
|
|
2763
2513
|
inputPrice: 4,
|
|
2764
|
-
outputPrice: 18
|
|
2514
|
+
outputPrice: 18,
|
|
2515
|
+
cacheReadsPrice: 0.4
|
|
2765
2516
|
}
|
|
2766
2517
|
]
|
|
2767
2518
|
},
|
|
@@ -2769,25 +2520,20 @@ var geminiModels = {
|
|
|
2769
2520
|
maxTokens: 65536,
|
|
2770
2521
|
contextWindow: 1048576,
|
|
2771
2522
|
supportsImages: true,
|
|
2772
|
-
supportsNativeTools: true,
|
|
2773
|
-
defaultToolProtocol: "native",
|
|
2774
2523
|
supportsPromptCache: true,
|
|
2775
2524
|
supportsReasoningEffort: ["minimal", "low", "medium", "high"],
|
|
2776
2525
|
reasoningEffort: "medium",
|
|
2777
2526
|
supportsTemperature: true,
|
|
2778
2527
|
defaultTemperature: 1,
|
|
2779
|
-
inputPrice: 0.
|
|
2780
|
-
outputPrice:
|
|
2781
|
-
cacheReadsPrice: 0.
|
|
2782
|
-
cacheWritesPrice: 1
|
|
2528
|
+
inputPrice: 0.5,
|
|
2529
|
+
outputPrice: 3,
|
|
2530
|
+
cacheReadsPrice: 0.05
|
|
2783
2531
|
},
|
|
2784
2532
|
// 2.5 Pro models
|
|
2785
2533
|
"gemini-2.5-pro": {
|
|
2786
2534
|
maxTokens: 64e3,
|
|
2787
2535
|
contextWindow: 1048576,
|
|
2788
2536
|
supportsImages: true,
|
|
2789
|
-
supportsNativeTools: true,
|
|
2790
|
-
defaultToolProtocol: "native",
|
|
2791
2537
|
supportsPromptCache: true,
|
|
2792
2538
|
inputPrice: 2.5,
|
|
2793
2539
|
// This is the pricing for prompts above 200k tokens.
|
|
@@ -2816,8 +2562,6 @@ var geminiModels = {
|
|
|
2816
2562
|
maxTokens: 65535,
|
|
2817
2563
|
contextWindow: 1048576,
|
|
2818
2564
|
supportsImages: true,
|
|
2819
|
-
supportsNativeTools: true,
|
|
2820
|
-
defaultToolProtocol: "native",
|
|
2821
2565
|
supportsPromptCache: true,
|
|
2822
2566
|
inputPrice: 2.5,
|
|
2823
2567
|
// This is the pricing for prompts above 200k tokens.
|
|
@@ -2845,8 +2589,6 @@ var geminiModels = {
|
|
|
2845
2589
|
maxTokens: 65535,
|
|
2846
2590
|
contextWindow: 1048576,
|
|
2847
2591
|
supportsImages: true,
|
|
2848
|
-
supportsNativeTools: true,
|
|
2849
|
-
defaultToolProtocol: "native",
|
|
2850
2592
|
supportsPromptCache: true,
|
|
2851
2593
|
inputPrice: 2.5,
|
|
2852
2594
|
// This is the pricing for prompts above 200k tokens.
|
|
@@ -2872,8 +2614,6 @@ var geminiModels = {
|
|
|
2872
2614
|
maxTokens: 65535,
|
|
2873
2615
|
contextWindow: 1048576,
|
|
2874
2616
|
supportsImages: true,
|
|
2875
|
-
supportsNativeTools: true,
|
|
2876
|
-
defaultToolProtocol: "native",
|
|
2877
2617
|
supportsPromptCache: true,
|
|
2878
2618
|
inputPrice: 2.5,
|
|
2879
2619
|
// This is the pricing for prompts above 200k tokens.
|
|
@@ -2902,8 +2642,6 @@ var geminiModels = {
|
|
|
2902
2642
|
maxTokens: 65536,
|
|
2903
2643
|
contextWindow: 1048576,
|
|
2904
2644
|
supportsImages: true,
|
|
2905
|
-
supportsNativeTools: true,
|
|
2906
|
-
defaultToolProtocol: "native",
|
|
2907
2645
|
supportsPromptCache: true,
|
|
2908
2646
|
inputPrice: 0.3,
|
|
2909
2647
|
outputPrice: 2.5,
|
|
@@ -2916,8 +2654,6 @@ var geminiModels = {
|
|
|
2916
2654
|
maxTokens: 65536,
|
|
2917
2655
|
contextWindow: 1048576,
|
|
2918
2656
|
supportsImages: true,
|
|
2919
|
-
supportsNativeTools: true,
|
|
2920
|
-
defaultToolProtocol: "native",
|
|
2921
2657
|
supportsPromptCache: true,
|
|
2922
2658
|
inputPrice: 0.3,
|
|
2923
2659
|
outputPrice: 2.5,
|
|
@@ -2930,8 +2666,6 @@ var geminiModels = {
|
|
|
2930
2666
|
maxTokens: 64e3,
|
|
2931
2667
|
contextWindow: 1048576,
|
|
2932
2668
|
supportsImages: true,
|
|
2933
|
-
supportsNativeTools: true,
|
|
2934
|
-
defaultToolProtocol: "native",
|
|
2935
2669
|
supportsPromptCache: true,
|
|
2936
2670
|
inputPrice: 0.3,
|
|
2937
2671
|
outputPrice: 2.5,
|
|
@@ -2945,8 +2679,6 @@ var geminiModels = {
|
|
|
2945
2679
|
maxTokens: 65536,
|
|
2946
2680
|
contextWindow: 1048576,
|
|
2947
2681
|
supportsImages: true,
|
|
2948
|
-
supportsNativeTools: true,
|
|
2949
|
-
defaultToolProtocol: "native",
|
|
2950
2682
|
supportsPromptCache: true,
|
|
2951
2683
|
inputPrice: 0.1,
|
|
2952
2684
|
outputPrice: 0.4,
|
|
@@ -2959,8 +2691,6 @@ var geminiModels = {
|
|
|
2959
2691
|
maxTokens: 65536,
|
|
2960
2692
|
contextWindow: 1048576,
|
|
2961
2693
|
supportsImages: true,
|
|
2962
|
-
supportsNativeTools: true,
|
|
2963
|
-
defaultToolProtocol: "native",
|
|
2964
2694
|
supportsPromptCache: true,
|
|
2965
2695
|
inputPrice: 0.1,
|
|
2966
2696
|
outputPrice: 0.4,
|
|
@@ -2980,8 +2710,6 @@ var groqModels = {
|
|
|
2980
2710
|
contextWindow: 131072,
|
|
2981
2711
|
supportsImages: false,
|
|
2982
2712
|
supportsPromptCache: false,
|
|
2983
|
-
supportsNativeTools: true,
|
|
2984
|
-
defaultToolProtocol: "native",
|
|
2985
2713
|
inputPrice: 0.05,
|
|
2986
2714
|
outputPrice: 0.08,
|
|
2987
2715
|
description: "Meta Llama 3.1 8B Instant model, 128K context."
|
|
@@ -2991,8 +2719,6 @@ var groqModels = {
|
|
|
2991
2719
|
contextWindow: 131072,
|
|
2992
2720
|
supportsImages: false,
|
|
2993
2721
|
supportsPromptCache: false,
|
|
2994
|
-
supportsNativeTools: true,
|
|
2995
|
-
defaultToolProtocol: "native",
|
|
2996
2722
|
inputPrice: 0.59,
|
|
2997
2723
|
outputPrice: 0.79,
|
|
2998
2724
|
description: "Meta Llama 3.3 70B Versatile model, 128K context."
|
|
@@ -3002,8 +2728,6 @@ var groqModels = {
|
|
|
3002
2728
|
contextWindow: 131072,
|
|
3003
2729
|
supportsImages: false,
|
|
3004
2730
|
supportsPromptCache: false,
|
|
3005
|
-
supportsNativeTools: true,
|
|
3006
|
-
defaultToolProtocol: "native",
|
|
3007
2731
|
inputPrice: 0.11,
|
|
3008
2732
|
outputPrice: 0.34,
|
|
3009
2733
|
description: "Meta Llama 4 Scout 17B Instruct model, 128K context."
|
|
@@ -3013,8 +2737,6 @@ var groqModels = {
|
|
|
3013
2737
|
contextWindow: 131072,
|
|
3014
2738
|
supportsImages: false,
|
|
3015
2739
|
supportsPromptCache: false,
|
|
3016
|
-
supportsNativeTools: true,
|
|
3017
|
-
defaultToolProtocol: "native",
|
|
3018
2740
|
inputPrice: 0.29,
|
|
3019
2741
|
outputPrice: 0.59,
|
|
3020
2742
|
description: "Alibaba Qwen 3 32B model, 128K context."
|
|
@@ -3024,8 +2746,6 @@ var groqModels = {
|
|
|
3024
2746
|
contextWindow: 262144,
|
|
3025
2747
|
supportsImages: false,
|
|
3026
2748
|
supportsPromptCache: true,
|
|
3027
|
-
supportsNativeTools: true,
|
|
3028
|
-
defaultToolProtocol: "native",
|
|
3029
2749
|
inputPrice: 0.6,
|
|
3030
2750
|
outputPrice: 2.5,
|
|
3031
2751
|
cacheReadsPrice: 0.15,
|
|
@@ -3036,8 +2756,6 @@ var groqModels = {
|
|
|
3036
2756
|
contextWindow: 131072,
|
|
3037
2757
|
supportsImages: false,
|
|
3038
2758
|
supportsPromptCache: false,
|
|
3039
|
-
supportsNativeTools: true,
|
|
3040
|
-
defaultToolProtocol: "native",
|
|
3041
2759
|
inputPrice: 0.15,
|
|
3042
2760
|
outputPrice: 0.75,
|
|
3043
2761
|
description: "GPT-OSS 120B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 128 experts."
|
|
@@ -3047,8 +2765,6 @@ var groqModels = {
|
|
|
3047
2765
|
contextWindow: 131072,
|
|
3048
2766
|
supportsImages: false,
|
|
3049
2767
|
supportsPromptCache: false,
|
|
3050
|
-
supportsNativeTools: true,
|
|
3051
|
-
defaultToolProtocol: "native",
|
|
3052
2768
|
inputPrice: 0.1,
|
|
3053
2769
|
outputPrice: 0.5,
|
|
3054
2770
|
description: "GPT-OSS 20B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 32 experts."
|
|
@@ -3075,7 +2791,6 @@ var ioIntelligenceModels = {
|
|
|
3075
2791
|
contextWindow: 128e3,
|
|
3076
2792
|
supportsImages: false,
|
|
3077
2793
|
supportsPromptCache: false,
|
|
3078
|
-
supportsNativeTools: true,
|
|
3079
2794
|
description: "DeepSeek R1 reasoning model"
|
|
3080
2795
|
},
|
|
3081
2796
|
"meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8": {
|
|
@@ -3083,7 +2798,6 @@ var ioIntelligenceModels = {
|
|
|
3083
2798
|
contextWindow: 43e4,
|
|
3084
2799
|
supportsImages: true,
|
|
3085
2800
|
supportsPromptCache: false,
|
|
3086
|
-
supportsNativeTools: true,
|
|
3087
2801
|
description: "Llama 4 Maverick 17B model"
|
|
3088
2802
|
},
|
|
3089
2803
|
"Intel/Qwen3-Coder-480B-A35B-Instruct-int4-mixed-ar": {
|
|
@@ -3091,7 +2805,6 @@ var ioIntelligenceModels = {
|
|
|
3091
2805
|
contextWindow: 106e3,
|
|
3092
2806
|
supportsImages: false,
|
|
3093
2807
|
supportsPromptCache: false,
|
|
3094
|
-
supportsNativeTools: true,
|
|
3095
2808
|
description: "Qwen3 Coder 480B specialized for coding"
|
|
3096
2809
|
},
|
|
3097
2810
|
"openai/gpt-oss-120b": {
|
|
@@ -3099,7 +2812,6 @@ var ioIntelligenceModels = {
|
|
|
3099
2812
|
contextWindow: 131072,
|
|
3100
2813
|
supportsImages: false,
|
|
3101
2814
|
supportsPromptCache: false,
|
|
3102
|
-
supportsNativeTools: true,
|
|
3103
2815
|
description: "OpenAI GPT-OSS 120B model"
|
|
3104
2816
|
}
|
|
3105
2817
|
};
|
|
@@ -3111,8 +2823,6 @@ var litellmDefaultModelInfo = {
|
|
|
3111
2823
|
contextWindow: 2e5,
|
|
3112
2824
|
supportsImages: true,
|
|
3113
2825
|
supportsPromptCache: true,
|
|
3114
|
-
supportsNativeTools: true,
|
|
3115
|
-
defaultToolProtocol: "native",
|
|
3116
2826
|
inputPrice: 3,
|
|
3117
2827
|
outputPrice: 15,
|
|
3118
2828
|
cacheWritesPrice: 3.75,
|
|
@@ -3127,8 +2837,6 @@ var lMStudioDefaultModelInfo = {
|
|
|
3127
2837
|
contextWindow: 2e5,
|
|
3128
2838
|
supportsImages: true,
|
|
3129
2839
|
supportsPromptCache: true,
|
|
3130
|
-
supportsNativeTools: true,
|
|
3131
|
-
defaultToolProtocol: "native",
|
|
3132
2840
|
inputPrice: 0,
|
|
3133
2841
|
outputPrice: 0,
|
|
3134
2842
|
cacheWritesPrice: 0,
|
|
@@ -3144,8 +2852,6 @@ var mistralModels = {
|
|
|
3144
2852
|
contextWindow: 128e3,
|
|
3145
2853
|
supportsImages: true,
|
|
3146
2854
|
supportsPromptCache: false,
|
|
3147
|
-
supportsNativeTools: true,
|
|
3148
|
-
defaultToolProtocol: "native",
|
|
3149
2855
|
inputPrice: 2,
|
|
3150
2856
|
outputPrice: 5
|
|
3151
2857
|
},
|
|
@@ -3154,8 +2860,6 @@ var mistralModels = {
|
|
|
3154
2860
|
contextWindow: 131e3,
|
|
3155
2861
|
supportsImages: true,
|
|
3156
2862
|
supportsPromptCache: false,
|
|
3157
|
-
supportsNativeTools: true,
|
|
3158
|
-
defaultToolProtocol: "native",
|
|
3159
2863
|
inputPrice: 0.4,
|
|
3160
2864
|
outputPrice: 2
|
|
3161
2865
|
},
|
|
@@ -3164,8 +2868,6 @@ var mistralModels = {
|
|
|
3164
2868
|
contextWindow: 131e3,
|
|
3165
2869
|
supportsImages: true,
|
|
3166
2870
|
supportsPromptCache: false,
|
|
3167
|
-
supportsNativeTools: true,
|
|
3168
|
-
defaultToolProtocol: "native",
|
|
3169
2871
|
inputPrice: 0.4,
|
|
3170
2872
|
outputPrice: 2
|
|
3171
2873
|
},
|
|
@@ -3174,8 +2876,6 @@ var mistralModels = {
|
|
|
3174
2876
|
contextWindow: 256e3,
|
|
3175
2877
|
supportsImages: false,
|
|
3176
2878
|
supportsPromptCache: false,
|
|
3177
|
-
supportsNativeTools: true,
|
|
3178
|
-
defaultToolProtocol: "native",
|
|
3179
2879
|
inputPrice: 0.3,
|
|
3180
2880
|
outputPrice: 0.9
|
|
3181
2881
|
},
|
|
@@ -3184,8 +2884,6 @@ var mistralModels = {
|
|
|
3184
2884
|
contextWindow: 131e3,
|
|
3185
2885
|
supportsImages: false,
|
|
3186
2886
|
supportsPromptCache: false,
|
|
3187
|
-
supportsNativeTools: true,
|
|
3188
|
-
defaultToolProtocol: "native",
|
|
3189
2887
|
inputPrice: 2,
|
|
3190
2888
|
outputPrice: 6
|
|
3191
2889
|
},
|
|
@@ -3194,8 +2892,6 @@ var mistralModels = {
|
|
|
3194
2892
|
contextWindow: 131e3,
|
|
3195
2893
|
supportsImages: false,
|
|
3196
2894
|
supportsPromptCache: false,
|
|
3197
|
-
supportsNativeTools: true,
|
|
3198
|
-
defaultToolProtocol: "native",
|
|
3199
2895
|
inputPrice: 0.1,
|
|
3200
2896
|
outputPrice: 0.1
|
|
3201
2897
|
},
|
|
@@ -3204,8 +2900,6 @@ var mistralModels = {
|
|
|
3204
2900
|
contextWindow: 131e3,
|
|
3205
2901
|
supportsImages: false,
|
|
3206
2902
|
supportsPromptCache: false,
|
|
3207
|
-
supportsNativeTools: true,
|
|
3208
|
-
defaultToolProtocol: "native",
|
|
3209
2903
|
inputPrice: 0.04,
|
|
3210
2904
|
outputPrice: 0.04
|
|
3211
2905
|
},
|
|
@@ -3214,8 +2908,6 @@ var mistralModels = {
|
|
|
3214
2908
|
contextWindow: 32e3,
|
|
3215
2909
|
supportsImages: false,
|
|
3216
2910
|
supportsPromptCache: false,
|
|
3217
|
-
supportsNativeTools: true,
|
|
3218
|
-
defaultToolProtocol: "native",
|
|
3219
2911
|
inputPrice: 0.2,
|
|
3220
2912
|
outputPrice: 0.6
|
|
3221
2913
|
},
|
|
@@ -3224,8 +2916,6 @@ var mistralModels = {
|
|
|
3224
2916
|
contextWindow: 131e3,
|
|
3225
2917
|
supportsImages: true,
|
|
3226
2918
|
supportsPromptCache: false,
|
|
3227
|
-
supportsNativeTools: true,
|
|
3228
|
-
defaultToolProtocol: "native",
|
|
3229
2919
|
inputPrice: 2,
|
|
3230
2920
|
outputPrice: 6
|
|
3231
2921
|
}
|
|
@@ -3240,8 +2930,6 @@ var moonshotModels = {
|
|
|
3240
2930
|
contextWindow: 131072,
|
|
3241
2931
|
supportsImages: false,
|
|
3242
2932
|
supportsPromptCache: true,
|
|
3243
|
-
supportsNativeTools: true,
|
|
3244
|
-
defaultToolProtocol: "native",
|
|
3245
2933
|
inputPrice: 0.6,
|
|
3246
2934
|
// $0.60 per million tokens (cache miss)
|
|
3247
2935
|
outputPrice: 2.5,
|
|
@@ -3257,8 +2945,6 @@ var moonshotModels = {
|
|
|
3257
2945
|
contextWindow: 262144,
|
|
3258
2946
|
supportsImages: false,
|
|
3259
2947
|
supportsPromptCache: true,
|
|
3260
|
-
supportsNativeTools: true,
|
|
3261
|
-
defaultToolProtocol: "native",
|
|
3262
2948
|
inputPrice: 0.6,
|
|
3263
2949
|
outputPrice: 2.5,
|
|
3264
2950
|
cacheReadsPrice: 0.15,
|
|
@@ -3269,8 +2955,6 @@ var moonshotModels = {
|
|
|
3269
2955
|
contextWindow: 262144,
|
|
3270
2956
|
supportsImages: false,
|
|
3271
2957
|
supportsPromptCache: true,
|
|
3272
|
-
supportsNativeTools: true,
|
|
3273
|
-
defaultToolProtocol: "native",
|
|
3274
2958
|
inputPrice: 2.4,
|
|
3275
2959
|
// $2.40 per million tokens (cache miss)
|
|
3276
2960
|
outputPrice: 10,
|
|
@@ -3289,8 +2973,6 @@ var moonshotModels = {
|
|
|
3289
2973
|
supportsImages: false,
|
|
3290
2974
|
// Text-only (no image/vision support)
|
|
3291
2975
|
supportsPromptCache: true,
|
|
3292
|
-
supportsNativeTools: true,
|
|
3293
|
-
defaultToolProtocol: "native",
|
|
3294
2976
|
inputPrice: 0.6,
|
|
3295
2977
|
// $0.60 per million tokens (cache miss)
|
|
3296
2978
|
outputPrice: 2.5,
|
|
@@ -3304,6 +2986,21 @@ var moonshotModels = {
|
|
|
3304
2986
|
preserveReasoning: true,
|
|
3305
2987
|
defaultTemperature: 1,
|
|
3306
2988
|
description: `The kimi-k2-thinking model is a general-purpose agentic reasoning model developed by Moonshot AI. Thanks to its strength in deep reasoning and multi-turn tool use, it can solve even the hardest problems.`
|
|
2989
|
+
},
|
|
2990
|
+
"kimi-k2.5": {
|
|
2991
|
+
maxTokens: 16384,
|
|
2992
|
+
contextWindow: 262144,
|
|
2993
|
+
supportsImages: false,
|
|
2994
|
+
supportsPromptCache: true,
|
|
2995
|
+
inputPrice: 0.6,
|
|
2996
|
+
// $0.60 per million tokens (cache miss)
|
|
2997
|
+
outputPrice: 3,
|
|
2998
|
+
// $3.00 per million tokens
|
|
2999
|
+
cacheReadsPrice: 0.1,
|
|
3000
|
+
// $0.10 per million tokens (cache hit)
|
|
3001
|
+
supportsTemperature: true,
|
|
3002
|
+
defaultTemperature: 1,
|
|
3003
|
+
description: "Kimi K2.5 is the latest generation of Moonshot AI's Kimi series, featuring improved reasoning capabilities and enhanced performance across diverse tasks."
|
|
3307
3004
|
}
|
|
3308
3005
|
};
|
|
3309
3006
|
var MOONSHOT_DEFAULT_TEMPERATURE = 0.6;
|
|
@@ -3315,7 +3012,6 @@ var ollamaDefaultModelInfo = {
|
|
|
3315
3012
|
contextWindow: 2e5,
|
|
3316
3013
|
supportsImages: true,
|
|
3317
3014
|
supportsPromptCache: true,
|
|
3318
|
-
supportsNativeTools: true,
|
|
3319
3015
|
inputPrice: 0,
|
|
3320
3016
|
outputPrice: 0,
|
|
3321
3017
|
cacheWritesPrice: 0,
|
|
@@ -3329,8 +3025,6 @@ var openAiNativeModels = {
|
|
|
3329
3025
|
"gpt-5.1-codex-max": {
|
|
3330
3026
|
maxTokens: 128e3,
|
|
3331
3027
|
contextWindow: 4e5,
|
|
3332
|
-
supportsNativeTools: true,
|
|
3333
|
-
defaultToolProtocol: "native",
|
|
3334
3028
|
includedTools: ["apply_patch"],
|
|
3335
3029
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3336
3030
|
supportsImages: true,
|
|
@@ -3348,8 +3042,6 @@ var openAiNativeModels = {
|
|
|
3348
3042
|
"gpt-5.2": {
|
|
3349
3043
|
maxTokens: 128e3,
|
|
3350
3044
|
contextWindow: 4e5,
|
|
3351
|
-
supportsNativeTools: true,
|
|
3352
|
-
defaultToolProtocol: "native",
|
|
3353
3045
|
includedTools: ["apply_patch"],
|
|
3354
3046
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3355
3047
|
supportsImages: true,
|
|
@@ -3368,11 +3060,26 @@ var openAiNativeModels = {
|
|
|
3368
3060
|
],
|
|
3369
3061
|
description: "GPT-5.2: Our flagship model for coding and agentic tasks across industries"
|
|
3370
3062
|
},
|
|
3063
|
+
"gpt-5.2-codex": {
|
|
3064
|
+
maxTokens: 128e3,
|
|
3065
|
+
contextWindow: 4e5,
|
|
3066
|
+
includedTools: ["apply_patch"],
|
|
3067
|
+
excludedTools: ["apply_diff", "write_to_file"],
|
|
3068
|
+
supportsImages: true,
|
|
3069
|
+
supportsPromptCache: true,
|
|
3070
|
+
promptCacheRetention: "24h",
|
|
3071
|
+
supportsReasoningEffort: ["low", "medium", "high", "xhigh"],
|
|
3072
|
+
reasoningEffort: "medium",
|
|
3073
|
+
inputPrice: 1.75,
|
|
3074
|
+
outputPrice: 14,
|
|
3075
|
+
cacheReadsPrice: 0.175,
|
|
3076
|
+
supportsTemperature: false,
|
|
3077
|
+
tiers: [{ name: "priority", contextWindow: 4e5, inputPrice: 3.5, outputPrice: 28, cacheReadsPrice: 0.35 }],
|
|
3078
|
+
description: "GPT-5.2 Codex: Our most intelligent coding model optimized for long-horizon, agentic coding tasks"
|
|
3079
|
+
},
|
|
3371
3080
|
"gpt-5.2-chat-latest": {
|
|
3372
3081
|
maxTokens: 16384,
|
|
3373
3082
|
contextWindow: 128e3,
|
|
3374
|
-
supportsNativeTools: true,
|
|
3375
|
-
defaultToolProtocol: "native",
|
|
3376
3083
|
includedTools: ["apply_patch"],
|
|
3377
3084
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3378
3085
|
supportsImages: true,
|
|
@@ -3385,8 +3092,6 @@ var openAiNativeModels = {
|
|
|
3385
3092
|
"gpt-5.1": {
|
|
3386
3093
|
maxTokens: 128e3,
|
|
3387
3094
|
contextWindow: 4e5,
|
|
3388
|
-
supportsNativeTools: true,
|
|
3389
|
-
defaultToolProtocol: "native",
|
|
3390
3095
|
includedTools: ["apply_patch"],
|
|
3391
3096
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3392
3097
|
supportsImages: true,
|
|
@@ -3408,8 +3113,6 @@ var openAiNativeModels = {
|
|
|
3408
3113
|
"gpt-5.1-codex": {
|
|
3409
3114
|
maxTokens: 128e3,
|
|
3410
3115
|
contextWindow: 4e5,
|
|
3411
|
-
supportsNativeTools: true,
|
|
3412
|
-
defaultToolProtocol: "native",
|
|
3413
3116
|
includedTools: ["apply_patch"],
|
|
3414
3117
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3415
3118
|
supportsImages: true,
|
|
@@ -3427,8 +3130,6 @@ var openAiNativeModels = {
|
|
|
3427
3130
|
"gpt-5.1-codex-mini": {
|
|
3428
3131
|
maxTokens: 128e3,
|
|
3429
3132
|
contextWindow: 4e5,
|
|
3430
|
-
supportsNativeTools: true,
|
|
3431
|
-
defaultToolProtocol: "native",
|
|
3432
3133
|
includedTools: ["apply_patch"],
|
|
3433
3134
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3434
3135
|
supportsImages: true,
|
|
@@ -3445,8 +3146,6 @@ var openAiNativeModels = {
|
|
|
3445
3146
|
"gpt-5": {
|
|
3446
3147
|
maxTokens: 128e3,
|
|
3447
3148
|
contextWindow: 4e5,
|
|
3448
|
-
supportsNativeTools: true,
|
|
3449
|
-
defaultToolProtocol: "native",
|
|
3450
3149
|
includedTools: ["apply_patch"],
|
|
3451
3150
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3452
3151
|
supportsImages: true,
|
|
@@ -3467,8 +3166,6 @@ var openAiNativeModels = {
|
|
|
3467
3166
|
"gpt-5-mini": {
|
|
3468
3167
|
maxTokens: 128e3,
|
|
3469
3168
|
contextWindow: 4e5,
|
|
3470
|
-
supportsNativeTools: true,
|
|
3471
|
-
defaultToolProtocol: "native",
|
|
3472
3169
|
includedTools: ["apply_patch"],
|
|
3473
3170
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3474
3171
|
supportsImages: true,
|
|
@@ -3489,8 +3186,6 @@ var openAiNativeModels = {
|
|
|
3489
3186
|
"gpt-5-codex": {
|
|
3490
3187
|
maxTokens: 128e3,
|
|
3491
3188
|
contextWindow: 4e5,
|
|
3492
|
-
supportsNativeTools: true,
|
|
3493
|
-
defaultToolProtocol: "native",
|
|
3494
3189
|
includedTools: ["apply_patch"],
|
|
3495
3190
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3496
3191
|
supportsImages: true,
|
|
@@ -3507,8 +3202,6 @@ var openAiNativeModels = {
|
|
|
3507
3202
|
"gpt-5-nano": {
|
|
3508
3203
|
maxTokens: 128e3,
|
|
3509
3204
|
contextWindow: 4e5,
|
|
3510
|
-
supportsNativeTools: true,
|
|
3511
|
-
defaultToolProtocol: "native",
|
|
3512
3205
|
includedTools: ["apply_patch"],
|
|
3513
3206
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3514
3207
|
supportsImages: true,
|
|
@@ -3526,8 +3219,6 @@ var openAiNativeModels = {
|
|
|
3526
3219
|
"gpt-5-chat-latest": {
|
|
3527
3220
|
maxTokens: 128e3,
|
|
3528
3221
|
contextWindow: 4e5,
|
|
3529
|
-
supportsNativeTools: true,
|
|
3530
|
-
defaultToolProtocol: "native",
|
|
3531
3222
|
includedTools: ["apply_patch"],
|
|
3532
3223
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3533
3224
|
supportsImages: true,
|
|
@@ -3540,8 +3231,6 @@ var openAiNativeModels = {
|
|
|
3540
3231
|
"gpt-4.1": {
|
|
3541
3232
|
maxTokens: 32768,
|
|
3542
3233
|
contextWindow: 1047576,
|
|
3543
|
-
supportsNativeTools: true,
|
|
3544
|
-
defaultToolProtocol: "native",
|
|
3545
3234
|
includedTools: ["apply_patch"],
|
|
3546
3235
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3547
3236
|
supportsImages: true,
|
|
@@ -3557,8 +3246,6 @@ var openAiNativeModels = {
|
|
|
3557
3246
|
"gpt-4.1-mini": {
|
|
3558
3247
|
maxTokens: 32768,
|
|
3559
3248
|
contextWindow: 1047576,
|
|
3560
|
-
supportsNativeTools: true,
|
|
3561
|
-
defaultToolProtocol: "native",
|
|
3562
3249
|
includedTools: ["apply_patch"],
|
|
3563
3250
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3564
3251
|
supportsImages: true,
|
|
@@ -3574,8 +3261,6 @@ var openAiNativeModels = {
|
|
|
3574
3261
|
"gpt-4.1-nano": {
|
|
3575
3262
|
maxTokens: 32768,
|
|
3576
3263
|
contextWindow: 1047576,
|
|
3577
|
-
supportsNativeTools: true,
|
|
3578
|
-
defaultToolProtocol: "native",
|
|
3579
3264
|
includedTools: ["apply_patch"],
|
|
3580
3265
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3581
3266
|
supportsImages: true,
|
|
@@ -3591,8 +3276,6 @@ var openAiNativeModels = {
|
|
|
3591
3276
|
o3: {
|
|
3592
3277
|
maxTokens: 1e5,
|
|
3593
3278
|
contextWindow: 2e5,
|
|
3594
|
-
supportsNativeTools: true,
|
|
3595
|
-
defaultToolProtocol: "native",
|
|
3596
3279
|
supportsImages: true,
|
|
3597
3280
|
supportsPromptCache: true,
|
|
3598
3281
|
inputPrice: 2,
|
|
@@ -3609,8 +3292,6 @@ var openAiNativeModels = {
|
|
|
3609
3292
|
"o3-high": {
|
|
3610
3293
|
maxTokens: 1e5,
|
|
3611
3294
|
contextWindow: 2e5,
|
|
3612
|
-
supportsNativeTools: true,
|
|
3613
|
-
defaultToolProtocol: "native",
|
|
3614
3295
|
supportsImages: true,
|
|
3615
3296
|
supportsPromptCache: true,
|
|
3616
3297
|
inputPrice: 2,
|
|
@@ -3622,8 +3303,6 @@ var openAiNativeModels = {
|
|
|
3622
3303
|
"o3-low": {
|
|
3623
3304
|
maxTokens: 1e5,
|
|
3624
3305
|
contextWindow: 2e5,
|
|
3625
|
-
supportsNativeTools: true,
|
|
3626
|
-
defaultToolProtocol: "native",
|
|
3627
3306
|
supportsImages: true,
|
|
3628
3307
|
supportsPromptCache: true,
|
|
3629
3308
|
inputPrice: 2,
|
|
@@ -3635,8 +3314,6 @@ var openAiNativeModels = {
|
|
|
3635
3314
|
"o4-mini": {
|
|
3636
3315
|
maxTokens: 1e5,
|
|
3637
3316
|
contextWindow: 2e5,
|
|
3638
|
-
supportsNativeTools: true,
|
|
3639
|
-
defaultToolProtocol: "native",
|
|
3640
3317
|
supportsImages: true,
|
|
3641
3318
|
supportsPromptCache: true,
|
|
3642
3319
|
inputPrice: 1.1,
|
|
@@ -3653,8 +3330,6 @@ var openAiNativeModels = {
|
|
|
3653
3330
|
"o4-mini-high": {
|
|
3654
3331
|
maxTokens: 1e5,
|
|
3655
3332
|
contextWindow: 2e5,
|
|
3656
|
-
supportsNativeTools: true,
|
|
3657
|
-
defaultToolProtocol: "native",
|
|
3658
3333
|
supportsImages: true,
|
|
3659
3334
|
supportsPromptCache: true,
|
|
3660
3335
|
inputPrice: 1.1,
|
|
@@ -3666,8 +3341,6 @@ var openAiNativeModels = {
|
|
|
3666
3341
|
"o4-mini-low": {
|
|
3667
3342
|
maxTokens: 1e5,
|
|
3668
3343
|
contextWindow: 2e5,
|
|
3669
|
-
supportsNativeTools: true,
|
|
3670
|
-
defaultToolProtocol: "native",
|
|
3671
3344
|
supportsImages: true,
|
|
3672
3345
|
supportsPromptCache: true,
|
|
3673
3346
|
inputPrice: 1.1,
|
|
@@ -3679,8 +3352,6 @@ var openAiNativeModels = {
|
|
|
3679
3352
|
"o3-mini": {
|
|
3680
3353
|
maxTokens: 1e5,
|
|
3681
3354
|
contextWindow: 2e5,
|
|
3682
|
-
supportsNativeTools: true,
|
|
3683
|
-
defaultToolProtocol: "native",
|
|
3684
3355
|
supportsImages: false,
|
|
3685
3356
|
supportsPromptCache: true,
|
|
3686
3357
|
inputPrice: 1.1,
|
|
@@ -3693,8 +3364,6 @@ var openAiNativeModels = {
|
|
|
3693
3364
|
"o3-mini-high": {
|
|
3694
3365
|
maxTokens: 1e5,
|
|
3695
3366
|
contextWindow: 2e5,
|
|
3696
|
-
supportsNativeTools: true,
|
|
3697
|
-
defaultToolProtocol: "native",
|
|
3698
3367
|
supportsImages: false,
|
|
3699
3368
|
supportsPromptCache: true,
|
|
3700
3369
|
inputPrice: 1.1,
|
|
@@ -3706,8 +3375,6 @@ var openAiNativeModels = {
|
|
|
3706
3375
|
"o3-mini-low": {
|
|
3707
3376
|
maxTokens: 1e5,
|
|
3708
3377
|
contextWindow: 2e5,
|
|
3709
|
-
supportsNativeTools: true,
|
|
3710
|
-
defaultToolProtocol: "native",
|
|
3711
3378
|
supportsImages: false,
|
|
3712
3379
|
supportsPromptCache: true,
|
|
3713
3380
|
inputPrice: 1.1,
|
|
@@ -3719,8 +3386,6 @@ var openAiNativeModels = {
|
|
|
3719
3386
|
o1: {
|
|
3720
3387
|
maxTokens: 1e5,
|
|
3721
3388
|
contextWindow: 2e5,
|
|
3722
|
-
supportsNativeTools: true,
|
|
3723
|
-
defaultToolProtocol: "native",
|
|
3724
3389
|
supportsImages: true,
|
|
3725
3390
|
supportsPromptCache: true,
|
|
3726
3391
|
inputPrice: 15,
|
|
@@ -3731,8 +3396,6 @@ var openAiNativeModels = {
|
|
|
3731
3396
|
"o1-preview": {
|
|
3732
3397
|
maxTokens: 32768,
|
|
3733
3398
|
contextWindow: 128e3,
|
|
3734
|
-
supportsNativeTools: true,
|
|
3735
|
-
defaultToolProtocol: "native",
|
|
3736
3399
|
supportsImages: true,
|
|
3737
3400
|
supportsPromptCache: true,
|
|
3738
3401
|
inputPrice: 15,
|
|
@@ -3743,8 +3406,6 @@ var openAiNativeModels = {
|
|
|
3743
3406
|
"o1-mini": {
|
|
3744
3407
|
maxTokens: 65536,
|
|
3745
3408
|
contextWindow: 128e3,
|
|
3746
|
-
supportsNativeTools: true,
|
|
3747
|
-
defaultToolProtocol: "native",
|
|
3748
3409
|
supportsImages: true,
|
|
3749
3410
|
supportsPromptCache: true,
|
|
3750
3411
|
inputPrice: 1.1,
|
|
@@ -3755,8 +3416,6 @@ var openAiNativeModels = {
|
|
|
3755
3416
|
"gpt-4o": {
|
|
3756
3417
|
maxTokens: 16384,
|
|
3757
3418
|
contextWindow: 128e3,
|
|
3758
|
-
supportsNativeTools: true,
|
|
3759
|
-
defaultToolProtocol: "native",
|
|
3760
3419
|
supportsImages: true,
|
|
3761
3420
|
supportsPromptCache: true,
|
|
3762
3421
|
inputPrice: 2.5,
|
|
@@ -3770,8 +3429,6 @@ var openAiNativeModels = {
|
|
|
3770
3429
|
"gpt-4o-mini": {
|
|
3771
3430
|
maxTokens: 16384,
|
|
3772
3431
|
contextWindow: 128e3,
|
|
3773
|
-
supportsNativeTools: true,
|
|
3774
|
-
defaultToolProtocol: "native",
|
|
3775
3432
|
supportsImages: true,
|
|
3776
3433
|
supportsPromptCache: true,
|
|
3777
3434
|
inputPrice: 0.15,
|
|
@@ -3785,8 +3442,6 @@ var openAiNativeModels = {
|
|
|
3785
3442
|
"codex-mini-latest": {
|
|
3786
3443
|
maxTokens: 16384,
|
|
3787
3444
|
contextWindow: 2e5,
|
|
3788
|
-
supportsNativeTools: true,
|
|
3789
|
-
defaultToolProtocol: "native",
|
|
3790
3445
|
supportsImages: false,
|
|
3791
3446
|
supportsPromptCache: false,
|
|
3792
3447
|
inputPrice: 1.5,
|
|
@@ -3799,8 +3454,6 @@ var openAiNativeModels = {
|
|
|
3799
3454
|
"gpt-5-2025-08-07": {
|
|
3800
3455
|
maxTokens: 128e3,
|
|
3801
3456
|
contextWindow: 4e5,
|
|
3802
|
-
supportsNativeTools: true,
|
|
3803
|
-
defaultToolProtocol: "native",
|
|
3804
3457
|
includedTools: ["apply_patch"],
|
|
3805
3458
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3806
3459
|
supportsImages: true,
|
|
@@ -3821,8 +3474,6 @@ var openAiNativeModels = {
|
|
|
3821
3474
|
"gpt-5-mini-2025-08-07": {
|
|
3822
3475
|
maxTokens: 128e3,
|
|
3823
3476
|
contextWindow: 4e5,
|
|
3824
|
-
supportsNativeTools: true,
|
|
3825
|
-
defaultToolProtocol: "native",
|
|
3826
3477
|
includedTools: ["apply_patch"],
|
|
3827
3478
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3828
3479
|
supportsImages: true,
|
|
@@ -3843,8 +3494,6 @@ var openAiNativeModels = {
|
|
|
3843
3494
|
"gpt-5-nano-2025-08-07": {
|
|
3844
3495
|
maxTokens: 128e3,
|
|
3845
3496
|
contextWindow: 4e5,
|
|
3846
|
-
supportsNativeTools: true,
|
|
3847
|
-
defaultToolProtocol: "native",
|
|
3848
3497
|
includedTools: ["apply_patch"],
|
|
3849
3498
|
excludedTools: ["apply_diff", "write_to_file"],
|
|
3850
3499
|
supportsImages: true,
|
|
@@ -3866,14 +3515,151 @@ var openAiModelInfoSaneDefaults = {
|
|
|
3866
3515
|
supportsImages: true,
|
|
3867
3516
|
supportsPromptCache: false,
|
|
3868
3517
|
inputPrice: 0,
|
|
3869
|
-
outputPrice: 0
|
|
3870
|
-
supportsNativeTools: true,
|
|
3871
|
-
defaultToolProtocol: "native"
|
|
3518
|
+
outputPrice: 0
|
|
3872
3519
|
};
|
|
3873
3520
|
var azureOpenAiDefaultApiVersion = "2024-08-01-preview";
|
|
3874
3521
|
var OPENAI_NATIVE_DEFAULT_TEMPERATURE = 0;
|
|
3875
3522
|
var OPENAI_AZURE_AI_INFERENCE_PATH = "/models/chat/completions";
|
|
3876
3523
|
|
|
3524
|
+
// src/providers/openai-codex.ts
|
|
3525
|
+
var openAiCodexDefaultModelId = "gpt-5.2-codex";
|
|
3526
|
+
var openAiCodexModels = {
|
|
3527
|
+
"gpt-5.1-codex-max": {
|
|
3528
|
+
maxTokens: 128e3,
|
|
3529
|
+
contextWindow: 4e5,
|
|
3530
|
+
includedTools: ["apply_patch"],
|
|
3531
|
+
excludedTools: ["apply_diff", "write_to_file"],
|
|
3532
|
+
supportsImages: true,
|
|
3533
|
+
supportsPromptCache: true,
|
|
3534
|
+
supportsReasoningEffort: ["low", "medium", "high", "xhigh"],
|
|
3535
|
+
reasoningEffort: "xhigh",
|
|
3536
|
+
// Subscription-based: no per-token costs
|
|
3537
|
+
inputPrice: 0,
|
|
3538
|
+
outputPrice: 0,
|
|
3539
|
+
supportsTemperature: false,
|
|
3540
|
+
description: "GPT-5.1 Codex Max: Maximum capability coding model via ChatGPT subscription"
|
|
3541
|
+
},
|
|
3542
|
+
"gpt-5.1-codex": {
|
|
3543
|
+
maxTokens: 128e3,
|
|
3544
|
+
contextWindow: 4e5,
|
|
3545
|
+
includedTools: ["apply_patch"],
|
|
3546
|
+
excludedTools: ["apply_diff", "write_to_file"],
|
|
3547
|
+
supportsImages: true,
|
|
3548
|
+
supportsPromptCache: true,
|
|
3549
|
+
supportsReasoningEffort: ["low", "medium", "high"],
|
|
3550
|
+
reasoningEffort: "medium",
|
|
3551
|
+
// Subscription-based: no per-token costs
|
|
3552
|
+
inputPrice: 0,
|
|
3553
|
+
outputPrice: 0,
|
|
3554
|
+
supportsTemperature: false,
|
|
3555
|
+
description: "GPT-5.1 Codex: GPT-5.1 optimized for agentic coding via ChatGPT subscription"
|
|
3556
|
+
},
|
|
3557
|
+
"gpt-5.2-codex": {
|
|
3558
|
+
maxTokens: 128e3,
|
|
3559
|
+
contextWindow: 4e5,
|
|
3560
|
+
includedTools: ["apply_patch"],
|
|
3561
|
+
excludedTools: ["apply_diff", "write_to_file"],
|
|
3562
|
+
supportsImages: true,
|
|
3563
|
+
supportsPromptCache: true,
|
|
3564
|
+
supportsReasoningEffort: ["low", "medium", "high", "xhigh"],
|
|
3565
|
+
reasoningEffort: "medium",
|
|
3566
|
+
inputPrice: 0,
|
|
3567
|
+
outputPrice: 0,
|
|
3568
|
+
supportsTemperature: false,
|
|
3569
|
+
description: "GPT-5.2 Codex: OpenAI's flagship coding model via ChatGPT subscription"
|
|
3570
|
+
},
|
|
3571
|
+
"gpt-5.1": {
|
|
3572
|
+
maxTokens: 128e3,
|
|
3573
|
+
contextWindow: 4e5,
|
|
3574
|
+
includedTools: ["apply_patch"],
|
|
3575
|
+
excludedTools: ["apply_diff", "write_to_file"],
|
|
3576
|
+
supportsImages: true,
|
|
3577
|
+
supportsPromptCache: true,
|
|
3578
|
+
supportsReasoningEffort: ["none", "low", "medium", "high"],
|
|
3579
|
+
reasoningEffort: "medium",
|
|
3580
|
+
// Subscription-based: no per-token costs
|
|
3581
|
+
inputPrice: 0,
|
|
3582
|
+
outputPrice: 0,
|
|
3583
|
+
supportsVerbosity: true,
|
|
3584
|
+
supportsTemperature: false,
|
|
3585
|
+
description: "GPT-5.1: General GPT-5.1 model via ChatGPT subscription"
|
|
3586
|
+
},
|
|
3587
|
+
"gpt-5": {
|
|
3588
|
+
maxTokens: 128e3,
|
|
3589
|
+
contextWindow: 4e5,
|
|
3590
|
+
includedTools: ["apply_patch"],
|
|
3591
|
+
excludedTools: ["apply_diff", "write_to_file"],
|
|
3592
|
+
supportsImages: true,
|
|
3593
|
+
supportsPromptCache: true,
|
|
3594
|
+
supportsReasoningEffort: ["minimal", "low", "medium", "high"],
|
|
3595
|
+
reasoningEffort: "medium",
|
|
3596
|
+
// Subscription-based: no per-token costs
|
|
3597
|
+
inputPrice: 0,
|
|
3598
|
+
outputPrice: 0,
|
|
3599
|
+
supportsVerbosity: true,
|
|
3600
|
+
supportsTemperature: false,
|
|
3601
|
+
description: "GPT-5: General GPT-5 model via ChatGPT subscription"
|
|
3602
|
+
},
|
|
3603
|
+
"gpt-5-codex": {
|
|
3604
|
+
maxTokens: 128e3,
|
|
3605
|
+
contextWindow: 4e5,
|
|
3606
|
+
includedTools: ["apply_patch"],
|
|
3607
|
+
excludedTools: ["apply_diff", "write_to_file"],
|
|
3608
|
+
supportsImages: true,
|
|
3609
|
+
supportsPromptCache: true,
|
|
3610
|
+
supportsReasoningEffort: ["low", "medium", "high"],
|
|
3611
|
+
reasoningEffort: "medium",
|
|
3612
|
+
// Subscription-based: no per-token costs
|
|
3613
|
+
inputPrice: 0,
|
|
3614
|
+
outputPrice: 0,
|
|
3615
|
+
supportsTemperature: false,
|
|
3616
|
+
description: "GPT-5 Codex: GPT-5 optimized for agentic coding via ChatGPT subscription"
|
|
3617
|
+
},
|
|
3618
|
+
"gpt-5-codex-mini": {
|
|
3619
|
+
maxTokens: 128e3,
|
|
3620
|
+
contextWindow: 4e5,
|
|
3621
|
+
includedTools: ["apply_patch"],
|
|
3622
|
+
excludedTools: ["apply_diff", "write_to_file"],
|
|
3623
|
+
supportsImages: true,
|
|
3624
|
+
supportsPromptCache: true,
|
|
3625
|
+
supportsReasoningEffort: ["low", "medium", "high"],
|
|
3626
|
+
reasoningEffort: "medium",
|
|
3627
|
+
// Subscription-based: no per-token costs
|
|
3628
|
+
inputPrice: 0,
|
|
3629
|
+
outputPrice: 0,
|
|
3630
|
+
supportsTemperature: false,
|
|
3631
|
+
description: "GPT-5 Codex Mini: Faster coding model via ChatGPT subscription"
|
|
3632
|
+
},
|
|
3633
|
+
"gpt-5.1-codex-mini": {
|
|
3634
|
+
maxTokens: 128e3,
|
|
3635
|
+
contextWindow: 4e5,
|
|
3636
|
+
includedTools: ["apply_patch"],
|
|
3637
|
+
excludedTools: ["apply_diff", "write_to_file"],
|
|
3638
|
+
supportsImages: true,
|
|
3639
|
+
supportsPromptCache: true,
|
|
3640
|
+
supportsReasoningEffort: ["low", "medium", "high"],
|
|
3641
|
+
reasoningEffort: "medium",
|
|
3642
|
+
inputPrice: 0,
|
|
3643
|
+
outputPrice: 0,
|
|
3644
|
+
supportsTemperature: false,
|
|
3645
|
+
description: "GPT-5.1 Codex Mini: Faster version for coding tasks via ChatGPT subscription"
|
|
3646
|
+
},
|
|
3647
|
+
"gpt-5.2": {
|
|
3648
|
+
maxTokens: 128e3,
|
|
3649
|
+
contextWindow: 4e5,
|
|
3650
|
+
includedTools: ["apply_patch"],
|
|
3651
|
+
excludedTools: ["apply_diff", "write_to_file"],
|
|
3652
|
+
supportsImages: true,
|
|
3653
|
+
supportsPromptCache: true,
|
|
3654
|
+
supportsReasoningEffort: ["none", "low", "medium", "high", "xhigh"],
|
|
3655
|
+
reasoningEffort: "medium",
|
|
3656
|
+
inputPrice: 0,
|
|
3657
|
+
outputPrice: 0,
|
|
3658
|
+
supportsTemperature: false,
|
|
3659
|
+
description: "GPT-5.2: Latest GPT model via ChatGPT subscription"
|
|
3660
|
+
}
|
|
3661
|
+
};
|
|
3662
|
+
|
|
3877
3663
|
// src/providers/openrouter.ts
|
|
3878
3664
|
var openRouterDefaultModelId = "anthropic/claude-sonnet-4.5";
|
|
3879
3665
|
var openRouterDefaultModelInfo = {
|
|
@@ -3881,7 +3667,6 @@ var openRouterDefaultModelInfo = {
|
|
|
3881
3667
|
contextWindow: 2e5,
|
|
3882
3668
|
supportsImages: true,
|
|
3883
3669
|
supportsPromptCache: true,
|
|
3884
|
-
supportsNativeTools: true,
|
|
3885
3670
|
inputPrice: 3,
|
|
3886
3671
|
outputPrice: 15,
|
|
3887
3672
|
cacheWritesPrice: 3.75,
|
|
@@ -3955,8 +3740,6 @@ var qwenCodeModels = {
|
|
|
3955
3740
|
contextWindow: 1e6,
|
|
3956
3741
|
supportsImages: false,
|
|
3957
3742
|
supportsPromptCache: false,
|
|
3958
|
-
supportsNativeTools: true,
|
|
3959
|
-
defaultToolProtocol: "native",
|
|
3960
3743
|
inputPrice: 0,
|
|
3961
3744
|
outputPrice: 0,
|
|
3962
3745
|
cacheWritesPrice: 0,
|
|
@@ -3968,8 +3751,6 @@ var qwenCodeModels = {
|
|
|
3968
3751
|
contextWindow: 1e6,
|
|
3969
3752
|
supportsImages: false,
|
|
3970
3753
|
supportsPromptCache: false,
|
|
3971
|
-
supportsNativeTools: true,
|
|
3972
|
-
defaultToolProtocol: "native",
|
|
3973
3754
|
inputPrice: 0,
|
|
3974
3755
|
outputPrice: 0,
|
|
3975
3756
|
cacheWritesPrice: 0,
|
|
@@ -3985,8 +3766,6 @@ var requestyDefaultModelInfo = {
|
|
|
3985
3766
|
contextWindow: 2e5,
|
|
3986
3767
|
supportsImages: true,
|
|
3987
3768
|
supportsPromptCache: true,
|
|
3988
|
-
supportsNativeTools: true,
|
|
3989
|
-
defaultToolProtocol: "native",
|
|
3990
3769
|
inputPrice: 3,
|
|
3991
3770
|
outputPrice: 15,
|
|
3992
3771
|
cacheWritesPrice: 3.75,
|
|
@@ -4042,8 +3821,6 @@ var sambaNovaModels = {
|
|
|
4042
3821
|
contextWindow: 16384,
|
|
4043
3822
|
supportsImages: false,
|
|
4044
3823
|
supportsPromptCache: false,
|
|
4045
|
-
supportsNativeTools: true,
|
|
4046
|
-
defaultToolProtocol: "native",
|
|
4047
3824
|
inputPrice: 0.1,
|
|
4048
3825
|
outputPrice: 0.2,
|
|
4049
3826
|
description: "Meta Llama 3.1 8B Instruct model with 16K context window."
|
|
@@ -4053,8 +3830,6 @@ var sambaNovaModels = {
|
|
|
4053
3830
|
contextWindow: 131072,
|
|
4054
3831
|
supportsImages: false,
|
|
4055
3832
|
supportsPromptCache: false,
|
|
4056
|
-
supportsNativeTools: true,
|
|
4057
|
-
defaultToolProtocol: "native",
|
|
4058
3833
|
inputPrice: 0.6,
|
|
4059
3834
|
outputPrice: 1.2,
|
|
4060
3835
|
description: "Meta Llama 3.3 70B Instruct model with 128K context window."
|
|
@@ -4065,8 +3840,6 @@ var sambaNovaModels = {
|
|
|
4065
3840
|
supportsImages: false,
|
|
4066
3841
|
supportsPromptCache: false,
|
|
4067
3842
|
supportsReasoningBudget: true,
|
|
4068
|
-
supportsNativeTools: true,
|
|
4069
|
-
defaultToolProtocol: "native",
|
|
4070
3843
|
inputPrice: 5,
|
|
4071
3844
|
outputPrice: 7,
|
|
4072
3845
|
description: "DeepSeek R1 reasoning model with 32K context window."
|
|
@@ -4076,8 +3849,6 @@ var sambaNovaModels = {
|
|
|
4076
3849
|
contextWindow: 32768,
|
|
4077
3850
|
supportsImages: false,
|
|
4078
3851
|
supportsPromptCache: false,
|
|
4079
|
-
supportsNativeTools: true,
|
|
4080
|
-
defaultToolProtocol: "native",
|
|
4081
3852
|
inputPrice: 3,
|
|
4082
3853
|
outputPrice: 4.5,
|
|
4083
3854
|
description: "DeepSeek V3 model with 32K context window."
|
|
@@ -4087,8 +3858,6 @@ var sambaNovaModels = {
|
|
|
4087
3858
|
contextWindow: 32768,
|
|
4088
3859
|
supportsImages: false,
|
|
4089
3860
|
supportsPromptCache: false,
|
|
4090
|
-
supportsNativeTools: true,
|
|
4091
|
-
defaultToolProtocol: "native",
|
|
4092
3861
|
inputPrice: 3,
|
|
4093
3862
|
outputPrice: 4.5,
|
|
4094
3863
|
description: "DeepSeek V3.1 model with 32K context window."
|
|
@@ -4098,8 +3867,6 @@ var sambaNovaModels = {
|
|
|
4098
3867
|
contextWindow: 131072,
|
|
4099
3868
|
supportsImages: true,
|
|
4100
3869
|
supportsPromptCache: false,
|
|
4101
|
-
supportsNativeTools: true,
|
|
4102
|
-
defaultToolProtocol: "native",
|
|
4103
3870
|
inputPrice: 0.63,
|
|
4104
3871
|
outputPrice: 1.8,
|
|
4105
3872
|
description: "Meta Llama 4 Maverick 17B 128E Instruct model with 128K context window."
|
|
@@ -4109,8 +3876,6 @@ var sambaNovaModels = {
|
|
|
4109
3876
|
contextWindow: 8192,
|
|
4110
3877
|
supportsImages: false,
|
|
4111
3878
|
supportsPromptCache: false,
|
|
4112
|
-
supportsNativeTools: true,
|
|
4113
|
-
defaultToolProtocol: "native",
|
|
4114
3879
|
inputPrice: 0.4,
|
|
4115
3880
|
outputPrice: 0.8,
|
|
4116
3881
|
description: "Alibaba Qwen 3 32B model with 8K context window."
|
|
@@ -4120,8 +3885,6 @@ var sambaNovaModels = {
|
|
|
4120
3885
|
contextWindow: 131072,
|
|
4121
3886
|
supportsImages: false,
|
|
4122
3887
|
supportsPromptCache: false,
|
|
4123
|
-
supportsNativeTools: true,
|
|
4124
|
-
defaultToolProtocol: "native",
|
|
4125
3888
|
inputPrice: 0.22,
|
|
4126
3889
|
outputPrice: 0.59,
|
|
4127
3890
|
description: "OpenAI gpt oss 120b model with 128k context window."
|
|
@@ -4135,7 +3898,6 @@ var unboundDefaultModelInfo = {
|
|
|
4135
3898
|
contextWindow: 2e5,
|
|
4136
3899
|
supportsImages: true,
|
|
4137
3900
|
supportsPromptCache: true,
|
|
4138
|
-
supportsNativeTools: true,
|
|
4139
3901
|
inputPrice: 3,
|
|
4140
3902
|
outputPrice: 15,
|
|
4141
3903
|
cacheWritesPrice: 3.75,
|
|
@@ -4149,8 +3911,6 @@ var vertexModels = {
|
|
|
4149
3911
|
maxTokens: 65536,
|
|
4150
3912
|
contextWindow: 1048576,
|
|
4151
3913
|
supportsImages: true,
|
|
4152
|
-
supportsNativeTools: true,
|
|
4153
|
-
defaultToolProtocol: "native",
|
|
4154
3914
|
supportsPromptCache: true,
|
|
4155
3915
|
supportsReasoningEffort: ["low", "high"],
|
|
4156
3916
|
reasoningEffort: "low",
|
|
@@ -4158,16 +3918,19 @@ var vertexModels = {
|
|
|
4158
3918
|
defaultTemperature: 1,
|
|
4159
3919
|
inputPrice: 4,
|
|
4160
3920
|
outputPrice: 18,
|
|
3921
|
+
cacheReadsPrice: 0.4,
|
|
4161
3922
|
tiers: [
|
|
4162
3923
|
{
|
|
4163
3924
|
contextWindow: 2e5,
|
|
4164
3925
|
inputPrice: 2,
|
|
4165
|
-
outputPrice: 12
|
|
3926
|
+
outputPrice: 12,
|
|
3927
|
+
cacheReadsPrice: 0.2
|
|
4166
3928
|
},
|
|
4167
3929
|
{
|
|
4168
3930
|
contextWindow: Infinity,
|
|
4169
3931
|
inputPrice: 4,
|
|
4170
|
-
outputPrice: 18
|
|
3932
|
+
outputPrice: 18,
|
|
3933
|
+
cacheReadsPrice: 0.4
|
|
4171
3934
|
}
|
|
4172
3935
|
]
|
|
4173
3936
|
},
|
|
@@ -4175,24 +3938,19 @@ var vertexModels = {
|
|
|
4175
3938
|
maxTokens: 65536,
|
|
4176
3939
|
contextWindow: 1048576,
|
|
4177
3940
|
supportsImages: true,
|
|
4178
|
-
supportsNativeTools: true,
|
|
4179
|
-
defaultToolProtocol: "native",
|
|
4180
3941
|
supportsPromptCache: true,
|
|
4181
3942
|
supportsReasoningEffort: ["minimal", "low", "medium", "high"],
|
|
4182
3943
|
reasoningEffort: "medium",
|
|
4183
3944
|
supportsTemperature: true,
|
|
4184
3945
|
defaultTemperature: 1,
|
|
4185
|
-
inputPrice: 0.
|
|
4186
|
-
outputPrice:
|
|
4187
|
-
cacheReadsPrice: 0.
|
|
4188
|
-
cacheWritesPrice: 1
|
|
3946
|
+
inputPrice: 0.5,
|
|
3947
|
+
outputPrice: 3,
|
|
3948
|
+
cacheReadsPrice: 0.05
|
|
4189
3949
|
},
|
|
4190
3950
|
"gemini-2.5-flash-preview-05-20:thinking": {
|
|
4191
3951
|
maxTokens: 65535,
|
|
4192
3952
|
contextWindow: 1048576,
|
|
4193
3953
|
supportsImages: true,
|
|
4194
|
-
supportsNativeTools: true,
|
|
4195
|
-
defaultToolProtocol: "native",
|
|
4196
3954
|
supportsPromptCache: true,
|
|
4197
3955
|
inputPrice: 0.15,
|
|
4198
3956
|
outputPrice: 3.5,
|
|
@@ -4204,8 +3962,6 @@ var vertexModels = {
|
|
|
4204
3962
|
maxTokens: 65535,
|
|
4205
3963
|
contextWindow: 1048576,
|
|
4206
3964
|
supportsImages: true,
|
|
4207
|
-
supportsNativeTools: true,
|
|
4208
|
-
defaultToolProtocol: "native",
|
|
4209
3965
|
supportsPromptCache: true,
|
|
4210
3966
|
inputPrice: 0.15,
|
|
4211
3967
|
outputPrice: 0.6
|
|
@@ -4214,8 +3970,6 @@ var vertexModels = {
|
|
|
4214
3970
|
maxTokens: 64e3,
|
|
4215
3971
|
contextWindow: 1048576,
|
|
4216
3972
|
supportsImages: true,
|
|
4217
|
-
supportsNativeTools: true,
|
|
4218
|
-
defaultToolProtocol: "native",
|
|
4219
3973
|
supportsPromptCache: true,
|
|
4220
3974
|
inputPrice: 0.3,
|
|
4221
3975
|
outputPrice: 2.5,
|
|
@@ -4228,8 +3982,6 @@ var vertexModels = {
|
|
|
4228
3982
|
maxTokens: 65535,
|
|
4229
3983
|
contextWindow: 1048576,
|
|
4230
3984
|
supportsImages: true,
|
|
4231
|
-
supportsNativeTools: true,
|
|
4232
|
-
defaultToolProtocol: "native",
|
|
4233
3985
|
supportsPromptCache: false,
|
|
4234
3986
|
inputPrice: 0.15,
|
|
4235
3987
|
outputPrice: 3.5,
|
|
@@ -4241,8 +3993,6 @@ var vertexModels = {
|
|
|
4241
3993
|
maxTokens: 65535,
|
|
4242
3994
|
contextWindow: 1048576,
|
|
4243
3995
|
supportsImages: true,
|
|
4244
|
-
supportsNativeTools: true,
|
|
4245
|
-
defaultToolProtocol: "native",
|
|
4246
3996
|
supportsPromptCache: false,
|
|
4247
3997
|
inputPrice: 0.15,
|
|
4248
3998
|
outputPrice: 0.6
|
|
@@ -4251,8 +4001,6 @@ var vertexModels = {
|
|
|
4251
4001
|
maxTokens: 65535,
|
|
4252
4002
|
contextWindow: 1048576,
|
|
4253
4003
|
supportsImages: true,
|
|
4254
|
-
supportsNativeTools: true,
|
|
4255
|
-
defaultToolProtocol: "native",
|
|
4256
4004
|
supportsPromptCache: true,
|
|
4257
4005
|
inputPrice: 2.5,
|
|
4258
4006
|
outputPrice: 15
|
|
@@ -4261,8 +4009,6 @@ var vertexModels = {
|
|
|
4261
4009
|
maxTokens: 65535,
|
|
4262
4010
|
contextWindow: 1048576,
|
|
4263
4011
|
supportsImages: true,
|
|
4264
|
-
supportsNativeTools: true,
|
|
4265
|
-
defaultToolProtocol: "native",
|
|
4266
4012
|
supportsPromptCache: true,
|
|
4267
4013
|
inputPrice: 2.5,
|
|
4268
4014
|
outputPrice: 15
|
|
@@ -4271,8 +4017,6 @@ var vertexModels = {
|
|
|
4271
4017
|
maxTokens: 65535,
|
|
4272
4018
|
contextWindow: 1048576,
|
|
4273
4019
|
supportsImages: true,
|
|
4274
|
-
supportsNativeTools: true,
|
|
4275
|
-
defaultToolProtocol: "native",
|
|
4276
4020
|
supportsPromptCache: true,
|
|
4277
4021
|
inputPrice: 2.5,
|
|
4278
4022
|
outputPrice: 15,
|
|
@@ -4283,8 +4027,6 @@ var vertexModels = {
|
|
|
4283
4027
|
maxTokens: 64e3,
|
|
4284
4028
|
contextWindow: 1048576,
|
|
4285
4029
|
supportsImages: true,
|
|
4286
|
-
supportsNativeTools: true,
|
|
4287
|
-
defaultToolProtocol: "native",
|
|
4288
4030
|
supportsPromptCache: true,
|
|
4289
4031
|
inputPrice: 2.5,
|
|
4290
4032
|
outputPrice: 15,
|
|
@@ -4310,8 +4052,6 @@ var vertexModels = {
|
|
|
4310
4052
|
maxTokens: 65535,
|
|
4311
4053
|
contextWindow: 1048576,
|
|
4312
4054
|
supportsImages: true,
|
|
4313
|
-
supportsNativeTools: true,
|
|
4314
|
-
defaultToolProtocol: "native",
|
|
4315
4055
|
supportsPromptCache: false,
|
|
4316
4056
|
inputPrice: 0,
|
|
4317
4057
|
outputPrice: 0
|
|
@@ -4320,8 +4060,6 @@ var vertexModels = {
|
|
|
4320
4060
|
maxTokens: 8192,
|
|
4321
4061
|
contextWindow: 2097152,
|
|
4322
4062
|
supportsImages: true,
|
|
4323
|
-
supportsNativeTools: true,
|
|
4324
|
-
defaultToolProtocol: "native",
|
|
4325
4063
|
supportsPromptCache: false,
|
|
4326
4064
|
inputPrice: 0,
|
|
4327
4065
|
outputPrice: 0
|
|
@@ -4330,8 +4068,6 @@ var vertexModels = {
|
|
|
4330
4068
|
maxTokens: 8192,
|
|
4331
4069
|
contextWindow: 1048576,
|
|
4332
4070
|
supportsImages: true,
|
|
4333
|
-
supportsNativeTools: true,
|
|
4334
|
-
defaultToolProtocol: "native",
|
|
4335
4071
|
supportsPromptCache: true,
|
|
4336
4072
|
inputPrice: 0.15,
|
|
4337
4073
|
outputPrice: 0.6
|
|
@@ -4340,8 +4076,6 @@ var vertexModels = {
|
|
|
4340
4076
|
maxTokens: 8192,
|
|
4341
4077
|
contextWindow: 1048576,
|
|
4342
4078
|
supportsImages: true,
|
|
4343
|
-
supportsNativeTools: true,
|
|
4344
|
-
defaultToolProtocol: "native",
|
|
4345
4079
|
supportsPromptCache: false,
|
|
4346
4080
|
inputPrice: 0.075,
|
|
4347
4081
|
outputPrice: 0.3
|
|
@@ -4350,8 +4084,6 @@ var vertexModels = {
|
|
|
4350
4084
|
maxTokens: 8192,
|
|
4351
4085
|
contextWindow: 32768,
|
|
4352
4086
|
supportsImages: true,
|
|
4353
|
-
supportsNativeTools: true,
|
|
4354
|
-
defaultToolProtocol: "native",
|
|
4355
4087
|
supportsPromptCache: false,
|
|
4356
4088
|
inputPrice: 0,
|
|
4357
4089
|
outputPrice: 0
|
|
@@ -4360,8 +4092,6 @@ var vertexModels = {
|
|
|
4360
4092
|
maxTokens: 8192,
|
|
4361
4093
|
contextWindow: 1048576,
|
|
4362
4094
|
supportsImages: true,
|
|
4363
|
-
supportsNativeTools: true,
|
|
4364
|
-
defaultToolProtocol: "native",
|
|
4365
4095
|
supportsPromptCache: true,
|
|
4366
4096
|
inputPrice: 0.075,
|
|
4367
4097
|
outputPrice: 0.3
|
|
@@ -4370,8 +4100,6 @@ var vertexModels = {
|
|
|
4370
4100
|
maxTokens: 8192,
|
|
4371
4101
|
contextWindow: 2097152,
|
|
4372
4102
|
supportsImages: true,
|
|
4373
|
-
supportsNativeTools: true,
|
|
4374
|
-
defaultToolProtocol: "native",
|
|
4375
4103
|
supportsPromptCache: false,
|
|
4376
4104
|
inputPrice: 1.25,
|
|
4377
4105
|
outputPrice: 5
|
|
@@ -4382,8 +4110,6 @@ var vertexModels = {
|
|
|
4382
4110
|
// Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
|
|
4383
4111
|
supportsImages: true,
|
|
4384
4112
|
supportsPromptCache: true,
|
|
4385
|
-
supportsNativeTools: true,
|
|
4386
|
-
defaultToolProtocol: "native",
|
|
4387
4113
|
inputPrice: 3,
|
|
4388
4114
|
// $3 per million input tokens (≤200K context)
|
|
4389
4115
|
outputPrice: 15,
|
|
@@ -4415,8 +4141,6 @@ var vertexModels = {
|
|
|
4415
4141
|
// Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
|
|
4416
4142
|
supportsImages: true,
|
|
4417
4143
|
supportsPromptCache: true,
|
|
4418
|
-
supportsNativeTools: true,
|
|
4419
|
-
defaultToolProtocol: "native",
|
|
4420
4144
|
inputPrice: 3,
|
|
4421
4145
|
// $3 per million input tokens (≤200K context)
|
|
4422
4146
|
outputPrice: 15,
|
|
@@ -4447,8 +4171,6 @@ var vertexModels = {
|
|
|
4447
4171
|
contextWindow: 2e5,
|
|
4448
4172
|
supportsImages: true,
|
|
4449
4173
|
supportsPromptCache: true,
|
|
4450
|
-
supportsNativeTools: true,
|
|
4451
|
-
defaultToolProtocol: "native",
|
|
4452
4174
|
inputPrice: 1,
|
|
4453
4175
|
outputPrice: 5,
|
|
4454
4176
|
cacheWritesPrice: 1.25,
|
|
@@ -4460,8 +4182,6 @@ var vertexModels = {
|
|
|
4460
4182
|
contextWindow: 2e5,
|
|
4461
4183
|
supportsImages: true,
|
|
4462
4184
|
supportsPromptCache: true,
|
|
4463
|
-
supportsNativeTools: true,
|
|
4464
|
-
defaultToolProtocol: "native",
|
|
4465
4185
|
inputPrice: 5,
|
|
4466
4186
|
outputPrice: 25,
|
|
4467
4187
|
cacheWritesPrice: 6.25,
|
|
@@ -4473,8 +4193,6 @@ var vertexModels = {
|
|
|
4473
4193
|
contextWindow: 2e5,
|
|
4474
4194
|
supportsImages: true,
|
|
4475
4195
|
supportsPromptCache: true,
|
|
4476
|
-
supportsNativeTools: true,
|
|
4477
|
-
defaultToolProtocol: "native",
|
|
4478
4196
|
inputPrice: 15,
|
|
4479
4197
|
outputPrice: 75,
|
|
4480
4198
|
cacheWritesPrice: 18.75,
|
|
@@ -4486,8 +4204,6 @@ var vertexModels = {
|
|
|
4486
4204
|
contextWindow: 2e5,
|
|
4487
4205
|
supportsImages: true,
|
|
4488
4206
|
supportsPromptCache: true,
|
|
4489
|
-
supportsNativeTools: true,
|
|
4490
|
-
defaultToolProtocol: "native",
|
|
4491
4207
|
inputPrice: 15,
|
|
4492
4208
|
outputPrice: 75,
|
|
4493
4209
|
cacheWritesPrice: 18.75,
|
|
@@ -4498,8 +4214,6 @@ var vertexModels = {
|
|
|
4498
4214
|
contextWindow: 2e5,
|
|
4499
4215
|
supportsImages: true,
|
|
4500
4216
|
supportsPromptCache: true,
|
|
4501
|
-
supportsNativeTools: true,
|
|
4502
|
-
defaultToolProtocol: "native",
|
|
4503
4217
|
inputPrice: 3,
|
|
4504
4218
|
outputPrice: 15,
|
|
4505
4219
|
cacheWritesPrice: 3.75,
|
|
@@ -4512,8 +4226,6 @@ var vertexModels = {
|
|
|
4512
4226
|
contextWindow: 2e5,
|
|
4513
4227
|
supportsImages: true,
|
|
4514
4228
|
supportsPromptCache: true,
|
|
4515
|
-
supportsNativeTools: true,
|
|
4516
|
-
defaultToolProtocol: "native",
|
|
4517
4229
|
inputPrice: 3,
|
|
4518
4230
|
outputPrice: 15,
|
|
4519
4231
|
cacheWritesPrice: 3.75,
|
|
@@ -4524,8 +4236,6 @@ var vertexModels = {
|
|
|
4524
4236
|
contextWindow: 2e5,
|
|
4525
4237
|
supportsImages: true,
|
|
4526
4238
|
supportsPromptCache: true,
|
|
4527
|
-
supportsNativeTools: true,
|
|
4528
|
-
defaultToolProtocol: "native",
|
|
4529
4239
|
inputPrice: 3,
|
|
4530
4240
|
outputPrice: 15,
|
|
4531
4241
|
cacheWritesPrice: 3.75,
|
|
@@ -4536,8 +4246,6 @@ var vertexModels = {
|
|
|
4536
4246
|
contextWindow: 2e5,
|
|
4537
4247
|
supportsImages: true,
|
|
4538
4248
|
supportsPromptCache: true,
|
|
4539
|
-
supportsNativeTools: true,
|
|
4540
|
-
defaultToolProtocol: "native",
|
|
4541
4249
|
inputPrice: 3,
|
|
4542
4250
|
outputPrice: 15,
|
|
4543
4251
|
cacheWritesPrice: 3.75,
|
|
@@ -4548,8 +4256,6 @@ var vertexModels = {
|
|
|
4548
4256
|
contextWindow: 2e5,
|
|
4549
4257
|
supportsImages: false,
|
|
4550
4258
|
supportsPromptCache: true,
|
|
4551
|
-
supportsNativeTools: true,
|
|
4552
|
-
defaultToolProtocol: "native",
|
|
4553
4259
|
inputPrice: 1,
|
|
4554
4260
|
outputPrice: 5,
|
|
4555
4261
|
cacheWritesPrice: 1.25,
|
|
@@ -4560,8 +4266,6 @@ var vertexModels = {
|
|
|
4560
4266
|
contextWindow: 2e5,
|
|
4561
4267
|
supportsImages: true,
|
|
4562
4268
|
supportsPromptCache: true,
|
|
4563
|
-
supportsNativeTools: true,
|
|
4564
|
-
defaultToolProtocol: "native",
|
|
4565
4269
|
inputPrice: 15,
|
|
4566
4270
|
outputPrice: 75,
|
|
4567
4271
|
cacheWritesPrice: 18.75,
|
|
@@ -4572,8 +4276,6 @@ var vertexModels = {
|
|
|
4572
4276
|
contextWindow: 2e5,
|
|
4573
4277
|
supportsImages: true,
|
|
4574
4278
|
supportsPromptCache: true,
|
|
4575
|
-
supportsNativeTools: true,
|
|
4576
|
-
defaultToolProtocol: "native",
|
|
4577
4279
|
inputPrice: 0.25,
|
|
4578
4280
|
outputPrice: 1.25,
|
|
4579
4281
|
cacheWritesPrice: 0.3,
|
|
@@ -4583,8 +4285,6 @@ var vertexModels = {
|
|
|
4583
4285
|
maxTokens: 64e3,
|
|
4584
4286
|
contextWindow: 1048576,
|
|
4585
4287
|
supportsImages: true,
|
|
4586
|
-
supportsNativeTools: true,
|
|
4587
|
-
defaultToolProtocol: "native",
|
|
4588
4288
|
supportsPromptCache: true,
|
|
4589
4289
|
inputPrice: 0.1,
|
|
4590
4290
|
outputPrice: 0.4,
|
|
@@ -4598,7 +4298,6 @@ var vertexModels = {
|
|
|
4598
4298
|
contextWindow: 131072,
|
|
4599
4299
|
supportsImages: false,
|
|
4600
4300
|
supportsPromptCache: false,
|
|
4601
|
-
supportsNativeTools: true,
|
|
4602
4301
|
inputPrice: 0.35,
|
|
4603
4302
|
outputPrice: 1.15,
|
|
4604
4303
|
description: "Meta Llama 4 Maverick 17B Instruct model, 128K context."
|
|
@@ -4608,7 +4307,6 @@ var vertexModels = {
|
|
|
4608
4307
|
contextWindow: 163840,
|
|
4609
4308
|
supportsImages: false,
|
|
4610
4309
|
supportsPromptCache: false,
|
|
4611
|
-
supportsNativeTools: true,
|
|
4612
4310
|
inputPrice: 1.35,
|
|
4613
4311
|
outputPrice: 5.4,
|
|
4614
4312
|
description: "DeepSeek R1 (0528). Available in us-central1"
|
|
@@ -4618,7 +4316,6 @@ var vertexModels = {
|
|
|
4618
4316
|
contextWindow: 163840,
|
|
4619
4317
|
supportsImages: false,
|
|
4620
4318
|
supportsPromptCache: false,
|
|
4621
|
-
supportsNativeTools: true,
|
|
4622
4319
|
inputPrice: 0.6,
|
|
4623
4320
|
outputPrice: 1.7,
|
|
4624
4321
|
description: "DeepSeek V3.1. Available in us-west2"
|
|
@@ -4628,7 +4325,6 @@ var vertexModels = {
|
|
|
4628
4325
|
contextWindow: 131072,
|
|
4629
4326
|
supportsImages: false,
|
|
4630
4327
|
supportsPromptCache: false,
|
|
4631
|
-
supportsNativeTools: true,
|
|
4632
4328
|
inputPrice: 0.15,
|
|
4633
4329
|
outputPrice: 0.6,
|
|
4634
4330
|
description: "OpenAI gpt-oss 120B. Available in us-central1"
|
|
@@ -4638,7 +4334,6 @@ var vertexModels = {
|
|
|
4638
4334
|
contextWindow: 131072,
|
|
4639
4335
|
supportsImages: false,
|
|
4640
4336
|
supportsPromptCache: false,
|
|
4641
|
-
supportsNativeTools: true,
|
|
4642
4337
|
inputPrice: 0.075,
|
|
4643
4338
|
outputPrice: 0.3,
|
|
4644
4339
|
description: "OpenAI gpt-oss 20B. Available in us-central1"
|
|
@@ -4648,7 +4343,6 @@ var vertexModels = {
|
|
|
4648
4343
|
contextWindow: 262144,
|
|
4649
4344
|
supportsImages: false,
|
|
4650
4345
|
supportsPromptCache: false,
|
|
4651
|
-
supportsNativeTools: true,
|
|
4652
4346
|
inputPrice: 1,
|
|
4653
4347
|
outputPrice: 4,
|
|
4654
4348
|
description: "Qwen3 Coder 480B A35B Instruct. Available in us-south1"
|
|
@@ -4658,10 +4352,18 @@ var vertexModels = {
|
|
|
4658
4352
|
contextWindow: 262144,
|
|
4659
4353
|
supportsImages: false,
|
|
4660
4354
|
supportsPromptCache: false,
|
|
4661
|
-
supportsNativeTools: true,
|
|
4662
4355
|
inputPrice: 0.25,
|
|
4663
4356
|
outputPrice: 1,
|
|
4664
4357
|
description: "Qwen3 235B A22B Instruct. Available in us-south1"
|
|
4358
|
+
},
|
|
4359
|
+
"moonshotai/kimi-k2-thinking-maas": {
|
|
4360
|
+
maxTokens: 16384,
|
|
4361
|
+
contextWindow: 262144,
|
|
4362
|
+
supportsPromptCache: false,
|
|
4363
|
+
supportsImages: false,
|
|
4364
|
+
inputPrice: 0.6,
|
|
4365
|
+
outputPrice: 2.5,
|
|
4366
|
+
description: "Kimi K2 Thinking Model with 256K context window."
|
|
4665
4367
|
}
|
|
4666
4368
|
};
|
|
4667
4369
|
var VERTEX_1M_CONTEXT_MODEL_IDS = ["claude-sonnet-4@20250514", "claude-sonnet-4-5@20250929"];
|
|
@@ -4894,8 +4596,6 @@ var xaiModels = {
|
|
|
4894
4596
|
contextWindow: 256e3,
|
|
4895
4597
|
supportsImages: true,
|
|
4896
4598
|
supportsPromptCache: true,
|
|
4897
|
-
supportsNativeTools: true,
|
|
4898
|
-
defaultToolProtocol: "native",
|
|
4899
4599
|
inputPrice: 0.2,
|
|
4900
4600
|
outputPrice: 1.5,
|
|
4901
4601
|
cacheWritesPrice: 0.02,
|
|
@@ -4909,8 +4609,6 @@ var xaiModels = {
|
|
|
4909
4609
|
contextWindow: 2e6,
|
|
4910
4610
|
supportsImages: true,
|
|
4911
4611
|
supportsPromptCache: true,
|
|
4912
|
-
supportsNativeTools: true,
|
|
4913
|
-
defaultToolProtocol: "native",
|
|
4914
4612
|
inputPrice: 0.2,
|
|
4915
4613
|
outputPrice: 0.5,
|
|
4916
4614
|
cacheWritesPrice: 0.05,
|
|
@@ -4924,8 +4622,6 @@ var xaiModels = {
|
|
|
4924
4622
|
contextWindow: 2e6,
|
|
4925
4623
|
supportsImages: true,
|
|
4926
4624
|
supportsPromptCache: true,
|
|
4927
|
-
supportsNativeTools: true,
|
|
4928
|
-
defaultToolProtocol: "native",
|
|
4929
4625
|
inputPrice: 0.2,
|
|
4930
4626
|
outputPrice: 0.5,
|
|
4931
4627
|
cacheWritesPrice: 0.05,
|
|
@@ -4939,8 +4635,6 @@ var xaiModels = {
|
|
|
4939
4635
|
contextWindow: 2e6,
|
|
4940
4636
|
supportsImages: true,
|
|
4941
4637
|
supportsPromptCache: true,
|
|
4942
|
-
supportsNativeTools: true,
|
|
4943
|
-
defaultToolProtocol: "native",
|
|
4944
4638
|
inputPrice: 0.2,
|
|
4945
4639
|
outputPrice: 0.5,
|
|
4946
4640
|
cacheWritesPrice: 0.05,
|
|
@@ -4954,8 +4648,6 @@ var xaiModels = {
|
|
|
4954
4648
|
contextWindow: 2e6,
|
|
4955
4649
|
supportsImages: true,
|
|
4956
4650
|
supportsPromptCache: true,
|
|
4957
|
-
supportsNativeTools: true,
|
|
4958
|
-
defaultToolProtocol: "native",
|
|
4959
4651
|
inputPrice: 0.2,
|
|
4960
4652
|
outputPrice: 0.5,
|
|
4961
4653
|
cacheWritesPrice: 0.05,
|
|
@@ -4969,8 +4661,6 @@ var xaiModels = {
|
|
|
4969
4661
|
contextWindow: 256e3,
|
|
4970
4662
|
supportsImages: true,
|
|
4971
4663
|
supportsPromptCache: true,
|
|
4972
|
-
supportsNativeTools: true,
|
|
4973
|
-
defaultToolProtocol: "native",
|
|
4974
4664
|
inputPrice: 3,
|
|
4975
4665
|
outputPrice: 15,
|
|
4976
4666
|
cacheWritesPrice: 0.75,
|
|
@@ -4984,8 +4674,6 @@ var xaiModels = {
|
|
|
4984
4674
|
contextWindow: 131072,
|
|
4985
4675
|
supportsImages: true,
|
|
4986
4676
|
supportsPromptCache: true,
|
|
4987
|
-
supportsNativeTools: true,
|
|
4988
|
-
defaultToolProtocol: "native",
|
|
4989
4677
|
inputPrice: 0.3,
|
|
4990
4678
|
outputPrice: 0.5,
|
|
4991
4679
|
cacheWritesPrice: 0.07,
|
|
@@ -5001,8 +4689,6 @@ var xaiModels = {
|
|
|
5001
4689
|
contextWindow: 131072,
|
|
5002
4690
|
supportsImages: true,
|
|
5003
4691
|
supportsPromptCache: true,
|
|
5004
|
-
supportsNativeTools: true,
|
|
5005
|
-
defaultToolProtocol: "native",
|
|
5006
4692
|
inputPrice: 3,
|
|
5007
4693
|
outputPrice: 15,
|
|
5008
4694
|
cacheWritesPrice: 0.75,
|
|
@@ -5099,7 +4785,6 @@ var vercelAiGatewayDefaultModelInfo = {
|
|
|
5099
4785
|
contextWindow: 2e5,
|
|
5100
4786
|
supportsImages: true,
|
|
5101
4787
|
supportsPromptCache: true,
|
|
5102
|
-
supportsNativeTools: true,
|
|
5103
4788
|
inputPrice: 3,
|
|
5104
4789
|
outputPrice: 15,
|
|
5105
4790
|
cacheWritesPrice: 3.75,
|
|
@@ -5116,8 +4801,6 @@ var internationalZAiModels = {
|
|
|
5116
4801
|
contextWindow: 131072,
|
|
5117
4802
|
supportsImages: false,
|
|
5118
4803
|
supportsPromptCache: true,
|
|
5119
|
-
supportsNativeTools: true,
|
|
5120
|
-
defaultToolProtocol: "native",
|
|
5121
4804
|
inputPrice: 0.6,
|
|
5122
4805
|
outputPrice: 2.2,
|
|
5123
4806
|
cacheWritesPrice: 0,
|
|
@@ -5129,8 +4812,6 @@ var internationalZAiModels = {
|
|
|
5129
4812
|
contextWindow: 131072,
|
|
5130
4813
|
supportsImages: false,
|
|
5131
4814
|
supportsPromptCache: true,
|
|
5132
|
-
supportsNativeTools: true,
|
|
5133
|
-
defaultToolProtocol: "native",
|
|
5134
4815
|
inputPrice: 0.2,
|
|
5135
4816
|
outputPrice: 1.1,
|
|
5136
4817
|
cacheWritesPrice: 0,
|
|
@@ -5142,8 +4823,6 @@ var internationalZAiModels = {
|
|
|
5142
4823
|
contextWindow: 131072,
|
|
5143
4824
|
supportsImages: false,
|
|
5144
4825
|
supportsPromptCache: true,
|
|
5145
|
-
supportsNativeTools: true,
|
|
5146
|
-
defaultToolProtocol: "native",
|
|
5147
4826
|
inputPrice: 2.2,
|
|
5148
4827
|
outputPrice: 8.9,
|
|
5149
4828
|
cacheWritesPrice: 0,
|
|
@@ -5155,8 +4834,6 @@ var internationalZAiModels = {
|
|
|
5155
4834
|
contextWindow: 131072,
|
|
5156
4835
|
supportsImages: false,
|
|
5157
4836
|
supportsPromptCache: true,
|
|
5158
|
-
supportsNativeTools: true,
|
|
5159
|
-
defaultToolProtocol: "native",
|
|
5160
4837
|
inputPrice: 1.1,
|
|
5161
4838
|
outputPrice: 4.5,
|
|
5162
4839
|
cacheWritesPrice: 0,
|
|
@@ -5168,8 +4845,6 @@ var internationalZAiModels = {
|
|
|
5168
4845
|
contextWindow: 131072,
|
|
5169
4846
|
supportsImages: false,
|
|
5170
4847
|
supportsPromptCache: true,
|
|
5171
|
-
supportsNativeTools: true,
|
|
5172
|
-
defaultToolProtocol: "native",
|
|
5173
4848
|
inputPrice: 0,
|
|
5174
4849
|
outputPrice: 0,
|
|
5175
4850
|
cacheWritesPrice: 0,
|
|
@@ -5181,21 +4856,28 @@ var internationalZAiModels = {
|
|
|
5181
4856
|
contextWindow: 131072,
|
|
5182
4857
|
supportsImages: true,
|
|
5183
4858
|
supportsPromptCache: true,
|
|
5184
|
-
supportsNativeTools: true,
|
|
5185
|
-
defaultToolProtocol: "native",
|
|
5186
4859
|
inputPrice: 0.6,
|
|
5187
4860
|
outputPrice: 1.8,
|
|
5188
4861
|
cacheWritesPrice: 0,
|
|
5189
4862
|
cacheReadsPrice: 0.11,
|
|
5190
4863
|
description: "GLM-4.5V is Z.AI's multimodal visual reasoning model (image/video/text/file input), optimized for GUI tasks, grounding, and document/video understanding."
|
|
5191
4864
|
},
|
|
4865
|
+
"glm-4.6v": {
|
|
4866
|
+
maxTokens: 16384,
|
|
4867
|
+
contextWindow: 131072,
|
|
4868
|
+
supportsImages: true,
|
|
4869
|
+
supportsPromptCache: true,
|
|
4870
|
+
inputPrice: 0.3,
|
|
4871
|
+
outputPrice: 0.9,
|
|
4872
|
+
cacheWritesPrice: 0,
|
|
4873
|
+
cacheReadsPrice: 0.05,
|
|
4874
|
+
description: "GLM-4.6V is an advanced multimodal vision model with improved performance and cost-efficiency for visual understanding tasks."
|
|
4875
|
+
},
|
|
5192
4876
|
"glm-4.6": {
|
|
5193
4877
|
maxTokens: 16384,
|
|
5194
4878
|
contextWindow: 2e5,
|
|
5195
4879
|
supportsImages: false,
|
|
5196
4880
|
supportsPromptCache: true,
|
|
5197
|
-
supportsNativeTools: true,
|
|
5198
|
-
defaultToolProtocol: "native",
|
|
5199
4881
|
inputPrice: 0.6,
|
|
5200
4882
|
outputPrice: 2.2,
|
|
5201
4883
|
cacheWritesPrice: 0,
|
|
@@ -5207,8 +4889,6 @@ var internationalZAiModels = {
|
|
|
5207
4889
|
contextWindow: 2e5,
|
|
5208
4890
|
supportsImages: false,
|
|
5209
4891
|
supportsPromptCache: true,
|
|
5210
|
-
supportsNativeTools: true,
|
|
5211
|
-
defaultToolProtocol: "native",
|
|
5212
4892
|
supportsReasoningEffort: ["disable", "medium"],
|
|
5213
4893
|
reasoningEffort: "medium",
|
|
5214
4894
|
preserveReasoning: true,
|
|
@@ -5218,13 +4898,55 @@ var internationalZAiModels = {
|
|
|
5218
4898
|
cacheReadsPrice: 0.11,
|
|
5219
4899
|
description: "GLM-4.7 is Zhipu's latest model with built-in thinking capabilities enabled by default. It provides enhanced reasoning for complex tasks while maintaining fast response times."
|
|
5220
4900
|
},
|
|
4901
|
+
"glm-4.7-flash": {
|
|
4902
|
+
maxTokens: 16384,
|
|
4903
|
+
contextWindow: 2e5,
|
|
4904
|
+
supportsImages: false,
|
|
4905
|
+
supportsPromptCache: true,
|
|
4906
|
+
inputPrice: 0,
|
|
4907
|
+
outputPrice: 0,
|
|
4908
|
+
cacheWritesPrice: 0,
|
|
4909
|
+
cacheReadsPrice: 0,
|
|
4910
|
+
description: "GLM-4.7-Flash is a free, high-speed variant of GLM-4.7 offering fast responses for reasoning and coding tasks."
|
|
4911
|
+
},
|
|
4912
|
+
"glm-4.7-flashx": {
|
|
4913
|
+
maxTokens: 16384,
|
|
4914
|
+
contextWindow: 2e5,
|
|
4915
|
+
supportsImages: false,
|
|
4916
|
+
supportsPromptCache: true,
|
|
4917
|
+
inputPrice: 0.07,
|
|
4918
|
+
outputPrice: 0.4,
|
|
4919
|
+
cacheWritesPrice: 0,
|
|
4920
|
+
cacheReadsPrice: 0.01,
|
|
4921
|
+
description: "GLM-4.7-FlashX is an ultra-fast variant of GLM-4.7 with exceptional speed and cost-effectiveness for high-throughput applications."
|
|
4922
|
+
},
|
|
4923
|
+
"glm-4.6v-flash": {
|
|
4924
|
+
maxTokens: 16384,
|
|
4925
|
+
contextWindow: 131072,
|
|
4926
|
+
supportsImages: true,
|
|
4927
|
+
supportsPromptCache: true,
|
|
4928
|
+
inputPrice: 0,
|
|
4929
|
+
outputPrice: 0,
|
|
4930
|
+
cacheWritesPrice: 0,
|
|
4931
|
+
cacheReadsPrice: 0,
|
|
4932
|
+
description: "GLM-4.6V-Flash is a free, high-speed multimodal vision model for rapid image understanding and visual reasoning tasks."
|
|
4933
|
+
},
|
|
4934
|
+
"glm-4.6v-flashx": {
|
|
4935
|
+
maxTokens: 16384,
|
|
4936
|
+
contextWindow: 131072,
|
|
4937
|
+
supportsImages: true,
|
|
4938
|
+
supportsPromptCache: true,
|
|
4939
|
+
inputPrice: 0.04,
|
|
4940
|
+
outputPrice: 0.4,
|
|
4941
|
+
cacheWritesPrice: 0,
|
|
4942
|
+
cacheReadsPrice: 4e-3,
|
|
4943
|
+
description: "GLM-4.6V-FlashX is an ultra-fast multimodal vision model optimized for high-speed visual processing at low cost."
|
|
4944
|
+
},
|
|
5221
4945
|
"glm-4-32b-0414-128k": {
|
|
5222
4946
|
maxTokens: 16384,
|
|
5223
4947
|
contextWindow: 131072,
|
|
5224
4948
|
supportsImages: false,
|
|
5225
4949
|
supportsPromptCache: false,
|
|
5226
|
-
supportsNativeTools: true,
|
|
5227
|
-
defaultToolProtocol: "native",
|
|
5228
4950
|
inputPrice: 0.1,
|
|
5229
4951
|
outputPrice: 0.1,
|
|
5230
4952
|
cacheWritesPrice: 0,
|
|
@@ -5239,8 +4961,6 @@ var mainlandZAiModels = {
|
|
|
5239
4961
|
contextWindow: 131072,
|
|
5240
4962
|
supportsImages: false,
|
|
5241
4963
|
supportsPromptCache: true,
|
|
5242
|
-
supportsNativeTools: true,
|
|
5243
|
-
defaultToolProtocol: "native",
|
|
5244
4964
|
inputPrice: 0.29,
|
|
5245
4965
|
outputPrice: 1.14,
|
|
5246
4966
|
cacheWritesPrice: 0,
|
|
@@ -5252,8 +4972,6 @@ var mainlandZAiModels = {
|
|
|
5252
4972
|
contextWindow: 131072,
|
|
5253
4973
|
supportsImages: false,
|
|
5254
4974
|
supportsPromptCache: true,
|
|
5255
|
-
supportsNativeTools: true,
|
|
5256
|
-
defaultToolProtocol: "native",
|
|
5257
4975
|
inputPrice: 0.1,
|
|
5258
4976
|
outputPrice: 0.6,
|
|
5259
4977
|
cacheWritesPrice: 0,
|
|
@@ -5265,8 +4983,6 @@ var mainlandZAiModels = {
|
|
|
5265
4983
|
contextWindow: 131072,
|
|
5266
4984
|
supportsImages: false,
|
|
5267
4985
|
supportsPromptCache: true,
|
|
5268
|
-
supportsNativeTools: true,
|
|
5269
|
-
defaultToolProtocol: "native",
|
|
5270
4986
|
inputPrice: 0.29,
|
|
5271
4987
|
outputPrice: 1.14,
|
|
5272
4988
|
cacheWritesPrice: 0,
|
|
@@ -5278,8 +4994,6 @@ var mainlandZAiModels = {
|
|
|
5278
4994
|
contextWindow: 131072,
|
|
5279
4995
|
supportsImages: false,
|
|
5280
4996
|
supportsPromptCache: true,
|
|
5281
|
-
supportsNativeTools: true,
|
|
5282
|
-
defaultToolProtocol: "native",
|
|
5283
4997
|
inputPrice: 0.1,
|
|
5284
4998
|
outputPrice: 0.6,
|
|
5285
4999
|
cacheWritesPrice: 0,
|
|
@@ -5291,8 +5005,6 @@ var mainlandZAiModels = {
|
|
|
5291
5005
|
contextWindow: 131072,
|
|
5292
5006
|
supportsImages: false,
|
|
5293
5007
|
supportsPromptCache: true,
|
|
5294
|
-
supportsNativeTools: true,
|
|
5295
|
-
defaultToolProtocol: "native",
|
|
5296
5008
|
inputPrice: 0,
|
|
5297
5009
|
outputPrice: 0,
|
|
5298
5010
|
cacheWritesPrice: 0,
|
|
@@ -5304,8 +5016,6 @@ var mainlandZAiModels = {
|
|
|
5304
5016
|
contextWindow: 131072,
|
|
5305
5017
|
supportsImages: true,
|
|
5306
5018
|
supportsPromptCache: true,
|
|
5307
|
-
supportsNativeTools: true,
|
|
5308
|
-
defaultToolProtocol: "native",
|
|
5309
5019
|
inputPrice: 0.29,
|
|
5310
5020
|
outputPrice: 0.93,
|
|
5311
5021
|
cacheWritesPrice: 0,
|
|
@@ -5317,8 +5027,6 @@ var mainlandZAiModels = {
|
|
|
5317
5027
|
contextWindow: 204800,
|
|
5318
5028
|
supportsImages: false,
|
|
5319
5029
|
supportsPromptCache: true,
|
|
5320
|
-
supportsNativeTools: true,
|
|
5321
|
-
defaultToolProtocol: "native",
|
|
5322
5030
|
inputPrice: 0.29,
|
|
5323
5031
|
outputPrice: 1.14,
|
|
5324
5032
|
cacheWritesPrice: 0,
|
|
@@ -5330,8 +5038,6 @@ var mainlandZAiModels = {
|
|
|
5330
5038
|
contextWindow: 204800,
|
|
5331
5039
|
supportsImages: false,
|
|
5332
5040
|
supportsPromptCache: true,
|
|
5333
|
-
supportsNativeTools: true,
|
|
5334
|
-
defaultToolProtocol: "native",
|
|
5335
5041
|
supportsReasoningEffort: ["disable", "medium"],
|
|
5336
5042
|
reasoningEffort: "medium",
|
|
5337
5043
|
preserveReasoning: true,
|
|
@@ -5340,6 +5046,61 @@ var mainlandZAiModels = {
|
|
|
5340
5046
|
cacheWritesPrice: 0,
|
|
5341
5047
|
cacheReadsPrice: 0.057,
|
|
5342
5048
|
description: "GLM-4.7 is Zhipu's latest model with built-in thinking capabilities enabled by default. It provides enhanced reasoning for complex tasks while maintaining fast response times."
|
|
5049
|
+
},
|
|
5050
|
+
"glm-4.7-flash": {
|
|
5051
|
+
maxTokens: 16384,
|
|
5052
|
+
contextWindow: 204800,
|
|
5053
|
+
supportsImages: false,
|
|
5054
|
+
supportsPromptCache: true,
|
|
5055
|
+
inputPrice: 0,
|
|
5056
|
+
outputPrice: 0,
|
|
5057
|
+
cacheWritesPrice: 0,
|
|
5058
|
+
cacheReadsPrice: 0,
|
|
5059
|
+
description: "GLM-4.7-Flash is a free, high-speed variant of GLM-4.7 offering fast responses for reasoning and coding tasks."
|
|
5060
|
+
},
|
|
5061
|
+
"glm-4.7-flashx": {
|
|
5062
|
+
maxTokens: 16384,
|
|
5063
|
+
contextWindow: 204800,
|
|
5064
|
+
supportsImages: false,
|
|
5065
|
+
supportsPromptCache: true,
|
|
5066
|
+
inputPrice: 0.035,
|
|
5067
|
+
outputPrice: 0.2,
|
|
5068
|
+
cacheWritesPrice: 0,
|
|
5069
|
+
cacheReadsPrice: 5e-3,
|
|
5070
|
+
description: "GLM-4.7-FlashX is an ultra-fast variant of GLM-4.7 with exceptional speed and cost-effectiveness for high-throughput applications."
|
|
5071
|
+
},
|
|
5072
|
+
"glm-4.6v": {
|
|
5073
|
+
maxTokens: 16384,
|
|
5074
|
+
contextWindow: 131072,
|
|
5075
|
+
supportsImages: true,
|
|
5076
|
+
supportsPromptCache: true,
|
|
5077
|
+
inputPrice: 0.15,
|
|
5078
|
+
outputPrice: 0.45,
|
|
5079
|
+
cacheWritesPrice: 0,
|
|
5080
|
+
cacheReadsPrice: 0.025,
|
|
5081
|
+
description: "GLM-4.6V is an advanced multimodal vision model with improved performance and cost-efficiency for visual understanding tasks."
|
|
5082
|
+
},
|
|
5083
|
+
"glm-4.6v-flash": {
|
|
5084
|
+
maxTokens: 16384,
|
|
5085
|
+
contextWindow: 131072,
|
|
5086
|
+
supportsImages: true,
|
|
5087
|
+
supportsPromptCache: true,
|
|
5088
|
+
inputPrice: 0,
|
|
5089
|
+
outputPrice: 0,
|
|
5090
|
+
cacheWritesPrice: 0,
|
|
5091
|
+
cacheReadsPrice: 0,
|
|
5092
|
+
description: "GLM-4.6V-Flash is a free, high-speed multimodal vision model for rapid image understanding and visual reasoning tasks."
|
|
5093
|
+
},
|
|
5094
|
+
"glm-4.6v-flashx": {
|
|
5095
|
+
maxTokens: 16384,
|
|
5096
|
+
contextWindow: 131072,
|
|
5097
|
+
supportsImages: true,
|
|
5098
|
+
supportsPromptCache: true,
|
|
5099
|
+
inputPrice: 0.02,
|
|
5100
|
+
outputPrice: 0.2,
|
|
5101
|
+
cacheWritesPrice: 0,
|
|
5102
|
+
cacheReadsPrice: 2e-3,
|
|
5103
|
+
description: "GLM-4.6V-FlashX is an ultra-fast multimodal vision model optimized for high-speed visual processing at low cost."
|
|
5343
5104
|
}
|
|
5344
5105
|
};
|
|
5345
5106
|
var ZAI_DEFAULT_TEMPERATURE = 0.6;
|
|
@@ -5373,7 +5134,6 @@ var deepInfraDefaultModelInfo = {
|
|
|
5373
5134
|
contextWindow: 262144,
|
|
5374
5135
|
supportsImages: false,
|
|
5375
5136
|
supportsPromptCache: false,
|
|
5376
|
-
supportsNativeTools: true,
|
|
5377
5137
|
inputPrice: 0.3,
|
|
5378
5138
|
outputPrice: 1.2,
|
|
5379
5139
|
description: "Qwen 3 Coder 480B A35B Instruct Turbo model, 256K context."
|
|
@@ -5387,8 +5147,6 @@ var minimaxModels = {
|
|
|
5387
5147
|
contextWindow: 192e3,
|
|
5388
5148
|
supportsImages: false,
|
|
5389
5149
|
supportsPromptCache: true,
|
|
5390
|
-
supportsNativeTools: true,
|
|
5391
|
-
defaultToolProtocol: "native",
|
|
5392
5150
|
includedTools: ["search_and_replace"],
|
|
5393
5151
|
excludedTools: ["apply_diff"],
|
|
5394
5152
|
preserveReasoning: true,
|
|
@@ -5403,8 +5161,6 @@ var minimaxModels = {
|
|
|
5403
5161
|
contextWindow: 192e3,
|
|
5404
5162
|
supportsImages: false,
|
|
5405
5163
|
supportsPromptCache: true,
|
|
5406
|
-
supportsNativeTools: true,
|
|
5407
|
-
defaultToolProtocol: "native",
|
|
5408
5164
|
includedTools: ["search_and_replace"],
|
|
5409
5165
|
excludedTools: ["apply_diff"],
|
|
5410
5166
|
preserveReasoning: true,
|
|
@@ -5419,8 +5175,6 @@ var minimaxModels = {
|
|
|
5419
5175
|
contextWindow: 192e3,
|
|
5420
5176
|
supportsImages: false,
|
|
5421
5177
|
supportsPromptCache: true,
|
|
5422
|
-
supportsNativeTools: true,
|
|
5423
|
-
defaultToolProtocol: "native",
|
|
5424
5178
|
includedTools: ["search_and_replace"],
|
|
5425
5179
|
excludedTools: ["apply_diff"],
|
|
5426
5180
|
preserveReasoning: true,
|
|
@@ -5475,6 +5229,8 @@ function getProviderDefaultModelId(provider, options = { isChina: false }) {
|
|
|
5475
5229
|
case "openai-native":
|
|
5476
5230
|
return "gpt-4o";
|
|
5477
5231
|
// Based on openai-native patterns
|
|
5232
|
+
case "openai-codex":
|
|
5233
|
+
return openAiCodexDefaultModelId;
|
|
5478
5234
|
case "mistral":
|
|
5479
5235
|
return mistralDefaultModelId;
|
|
5480
5236
|
case "openai":
|
|
@@ -5490,8 +5246,6 @@ function getProviderDefaultModelId(provider, options = { isChina: false }) {
|
|
|
5490
5246
|
return deepInfraDefaultModelId;
|
|
5491
5247
|
case "vscode-lm":
|
|
5492
5248
|
return vscodeLlmDefaultModelId;
|
|
5493
|
-
case "claude-code":
|
|
5494
|
-
return claudeCodeDefaultModelId;
|
|
5495
5249
|
case "cerebras":
|
|
5496
5250
|
return cerebrasDefaultModelId;
|
|
5497
5251
|
case "sambanova":
|
|
@@ -5549,7 +5303,6 @@ var providerNames = [
|
|
|
5549
5303
|
"bedrock",
|
|
5550
5304
|
"baseten",
|
|
5551
5305
|
"cerebras",
|
|
5552
|
-
"claude-code",
|
|
5553
5306
|
"doubao",
|
|
5554
5307
|
"deepseek",
|
|
5555
5308
|
"featherless",
|
|
@@ -5560,6 +5313,7 @@ var providerNames = [
|
|
|
5560
5313
|
"mistral",
|
|
5561
5314
|
"moonshot",
|
|
5562
5315
|
"minimax",
|
|
5316
|
+
"openai-codex",
|
|
5563
5317
|
"openai-native",
|
|
5564
5318
|
"qwen-code",
|
|
5565
5319
|
"roo",
|
|
@@ -5578,9 +5332,7 @@ var providerSettingsEntrySchema = import_zod8.z.object({
|
|
|
5578
5332
|
});
|
|
5579
5333
|
var baseProviderSettingsSchema = import_zod8.z.object({
|
|
5580
5334
|
includeMaxTokens: import_zod8.z.boolean().optional(),
|
|
5581
|
-
diffEnabled: import_zod8.z.boolean().optional(),
|
|
5582
5335
|
todoListEnabled: import_zod8.z.boolean().optional(),
|
|
5583
|
-
fuzzyMatchThreshold: import_zod8.z.number().optional(),
|
|
5584
5336
|
modelTemperature: import_zod8.z.number().nullish(),
|
|
5585
5337
|
rateLimitSeconds: import_zod8.z.number().optional(),
|
|
5586
5338
|
consecutiveMistakeLimit: import_zod8.z.number().min(0).optional(),
|
|
@@ -5590,9 +5342,7 @@ var baseProviderSettingsSchema = import_zod8.z.object({
|
|
|
5590
5342
|
modelMaxTokens: import_zod8.z.number().optional(),
|
|
5591
5343
|
modelMaxThinkingTokens: import_zod8.z.number().optional(),
|
|
5592
5344
|
// Model verbosity.
|
|
5593
|
-
verbosity: verbosityLevelsSchema.optional()
|
|
5594
|
-
// Tool protocol override for this profile.
|
|
5595
|
-
toolProtocol: import_zod8.z.enum(["xml", "native"]).optional()
|
|
5345
|
+
verbosity: verbosityLevelsSchema.optional()
|
|
5596
5346
|
});
|
|
5597
5347
|
var apiModelIdProviderModelSchema = baseProviderSettingsSchema.extend({
|
|
5598
5348
|
apiModelId: import_zod8.z.string().optional()
|
|
@@ -5604,7 +5354,6 @@ var anthropicSchema = apiModelIdProviderModelSchema.extend({
|
|
|
5604
5354
|
anthropicBeta1MContext: import_zod8.z.boolean().optional()
|
|
5605
5355
|
// Enable 'context-1m-2025-08-07' beta for 1M context window.
|
|
5606
5356
|
});
|
|
5607
|
-
var claudeCodeSchema = apiModelIdProviderModelSchema.extend({});
|
|
5608
5357
|
var openRouterSchema = baseProviderSettingsSchema.extend({
|
|
5609
5358
|
openRouterApiKey: import_zod8.z.string().optional(),
|
|
5610
5359
|
openRouterModelId: import_zod8.z.string().optional(),
|
|
@@ -5686,6 +5435,9 @@ var geminiCliSchema = apiModelIdProviderModelSchema.extend({
|
|
|
5686
5435
|
geminiCliOAuthPath: import_zod8.z.string().optional(),
|
|
5687
5436
|
geminiCliProjectId: import_zod8.z.string().optional()
|
|
5688
5437
|
});
|
|
5438
|
+
var openAiCodexSchema = apiModelIdProviderModelSchema.extend({
|
|
5439
|
+
// No additional settings needed - uses OAuth authentication
|
|
5440
|
+
});
|
|
5689
5441
|
var openAiNativeSchema = apiModelIdProviderModelSchema.extend({
|
|
5690
5442
|
openAiNativeApiKey: import_zod8.z.string().optional(),
|
|
5691
5443
|
openAiNativeBaseUrl: import_zod8.z.string().optional(),
|
|
@@ -5790,7 +5542,6 @@ var defaultSchema = import_zod8.z.object({
|
|
|
5790
5542
|
});
|
|
5791
5543
|
var providerSettingsSchemaDiscriminated = import_zod8.z.discriminatedUnion("apiProvider", [
|
|
5792
5544
|
anthropicSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("anthropic") })),
|
|
5793
|
-
claudeCodeSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("claude-code") })),
|
|
5794
5545
|
openRouterSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("openrouter") })),
|
|
5795
5546
|
bedrockSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("bedrock") })),
|
|
5796
5547
|
vertexSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("vertex") })),
|
|
@@ -5800,6 +5551,7 @@ var providerSettingsSchemaDiscriminated = import_zod8.z.discriminatedUnion("apiP
|
|
|
5800
5551
|
lmStudioSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("lmstudio") })),
|
|
5801
5552
|
geminiSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("gemini") })),
|
|
5802
5553
|
geminiCliSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("gemini-cli") })),
|
|
5554
|
+
openAiCodexSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("openai-codex") })),
|
|
5803
5555
|
openAiNativeSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("openai-native") })),
|
|
5804
5556
|
mistralSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("mistral") })),
|
|
5805
5557
|
deepSeekSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("deepseek") })),
|
|
@@ -5830,7 +5582,6 @@ var providerSettingsSchemaDiscriminated = import_zod8.z.discriminatedUnion("apiP
|
|
|
5830
5582
|
var providerSettingsSchema = import_zod8.z.object({
|
|
5831
5583
|
apiProvider: providerNamesSchema.optional(),
|
|
5832
5584
|
...anthropicSchema.shape,
|
|
5833
|
-
...claudeCodeSchema.shape,
|
|
5834
5585
|
...openRouterSchema.shape,
|
|
5835
5586
|
...bedrockSchema.shape,
|
|
5836
5587
|
...vertexSchema.shape,
|
|
@@ -5840,6 +5591,7 @@ var providerSettingsSchema = import_zod8.z.object({
|
|
|
5840
5591
|
...lmStudioSchema.shape,
|
|
5841
5592
|
...geminiSchema.shape,
|
|
5842
5593
|
...geminiCliSchema.shape,
|
|
5594
|
+
...openAiCodexSchema.shape,
|
|
5843
5595
|
...openAiNativeSchema.shape,
|
|
5844
5596
|
...mistralSchema.shape,
|
|
5845
5597
|
...deepSeekSchema.shape,
|
|
@@ -5894,10 +5646,10 @@ var getModelId = (settings) => {
|
|
|
5894
5646
|
var isTypicalProvider = (key) => isProviderName(key) && !isInternalProvider(key) && !isCustomProvider(key) && !isFauxProvider(key);
|
|
5895
5647
|
var modelIdKeysByProvider = {
|
|
5896
5648
|
anthropic: "apiModelId",
|
|
5897
|
-
"claude-code": "apiModelId",
|
|
5898
5649
|
openrouter: "openRouterModelId",
|
|
5899
5650
|
bedrock: "apiModelId",
|
|
5900
5651
|
vertex: "apiModelId",
|
|
5652
|
+
"openai-codex": "apiModelId",
|
|
5901
5653
|
"openai-native": "openAiModelId",
|
|
5902
5654
|
ollama: "ollamaModelId",
|
|
5903
5655
|
lmstudio: "lmStudioModelId",
|
|
@@ -5927,7 +5679,7 @@ var modelIdKeysByProvider = {
|
|
|
5927
5679
|
roo: "apiModelId",
|
|
5928
5680
|
"vercel-ai-gateway": "vercelAiGatewayModelId"
|
|
5929
5681
|
};
|
|
5930
|
-
var ANTHROPIC_STYLE_PROVIDERS = ["anthropic", "
|
|
5682
|
+
var ANTHROPIC_STYLE_PROVIDERS = ["anthropic", "bedrock", "minimax"];
|
|
5931
5683
|
var getApiProtocol = (provider, modelId) => {
|
|
5932
5684
|
if (provider && ANTHROPIC_STYLE_PROVIDERS.includes(provider)) {
|
|
5933
5685
|
return "anthropic";
|
|
@@ -5956,7 +5708,6 @@ var MODELS_BY_PROVIDER = {
|
|
|
5956
5708
|
label: "Cerebras",
|
|
5957
5709
|
models: Object.keys(cerebrasModels)
|
|
5958
5710
|
},
|
|
5959
|
-
"claude-code": { id: "claude-code", label: "Claude Code", models: Object.keys(claudeCodeModels) },
|
|
5960
5711
|
deepseek: {
|
|
5961
5712
|
id: "deepseek",
|
|
5962
5713
|
label: "DeepSeek",
|
|
@@ -5999,6 +5750,11 @@ var MODELS_BY_PROVIDER = {
|
|
|
5999
5750
|
label: "MiniMax",
|
|
6000
5751
|
models: Object.keys(minimaxModels)
|
|
6001
5752
|
},
|
|
5753
|
+
"openai-codex": {
|
|
5754
|
+
id: "openai-codex",
|
|
5755
|
+
label: "OpenAI - ChatGPT Plus/Pro",
|
|
5756
|
+
models: Object.keys(openAiCodexModels)
|
|
5757
|
+
},
|
|
6002
5758
|
"openai-native": {
|
|
6003
5759
|
id: "openai-native",
|
|
6004
5760
|
label: "OpenAI",
|
|
@@ -6055,16 +5811,6 @@ var historyItemSchema = import_zod9.z.object({
|
|
|
6055
5811
|
size: import_zod9.z.number().optional(),
|
|
6056
5812
|
workspace: import_zod9.z.string().optional(),
|
|
6057
5813
|
mode: import_zod9.z.string().optional(),
|
|
6058
|
-
/**
|
|
6059
|
-
* The tool protocol used by this task. Once a task uses tools with a specific
|
|
6060
|
-
* protocol (XML or Native), it is permanently locked to that protocol.
|
|
6061
|
-
*
|
|
6062
|
-
* - "xml": Tool calls are parsed from XML text (no tool IDs)
|
|
6063
|
-
* - "native": Tool calls come as tool_call chunks with IDs
|
|
6064
|
-
*
|
|
6065
|
-
* This ensures task resumption works correctly even when NTC settings change.
|
|
6066
|
-
*/
|
|
6067
|
-
toolProtocol: import_zod9.z.enum(["xml", "native"]).optional(),
|
|
6068
5814
|
apiConfigName: import_zod9.z.string().optional(),
|
|
6069
5815
|
// Provider profile name for sticky profile feature
|
|
6070
5816
|
status: import_zod9.z.enum(["active", "completed", "delegated"]).optional(),
|
|
@@ -6082,23 +5828,12 @@ var historyItemSchema = import_zod9.z.object({
|
|
|
6082
5828
|
|
|
6083
5829
|
// src/experiment.ts
|
|
6084
5830
|
var import_zod10 = require("zod");
|
|
6085
|
-
var experimentIds = [
|
|
6086
|
-
"powerSteering",
|
|
6087
|
-
"multiFileApplyDiff",
|
|
6088
|
-
"preventFocusDisruption",
|
|
6089
|
-
"imageGeneration",
|
|
6090
|
-
"runSlashCommand",
|
|
6091
|
-
"multipleNativeToolCalls",
|
|
6092
|
-
"customTools"
|
|
6093
|
-
];
|
|
5831
|
+
var experimentIds = ["preventFocusDisruption", "imageGeneration", "runSlashCommand", "customTools"];
|
|
6094
5832
|
var experimentIdsSchema = import_zod10.z.enum(experimentIds);
|
|
6095
5833
|
var experimentsSchema = import_zod10.z.object({
|
|
6096
|
-
powerSteering: import_zod10.z.boolean().optional(),
|
|
6097
|
-
multiFileApplyDiff: import_zod10.z.boolean().optional(),
|
|
6098
5834
|
preventFocusDisruption: import_zod10.z.boolean().optional(),
|
|
6099
5835
|
imageGeneration: import_zod10.z.boolean().optional(),
|
|
6100
5836
|
runSlashCommand: import_zod10.z.boolean().optional(),
|
|
6101
|
-
multipleNativeToolCalls: import_zod10.z.boolean().optional(),
|
|
6102
5837
|
customTools: import_zod10.z.boolean().optional()
|
|
6103
5838
|
});
|
|
6104
5839
|
|
|
@@ -6150,6 +5885,7 @@ var TelemetryEventName = /* @__PURE__ */ ((TelemetryEventName2) => {
|
|
|
6150
5885
|
TelemetryEventName2["CODE_INDEX_ERROR"] = "Code Index Error";
|
|
6151
5886
|
TelemetryEventName2["TELEMETRY_SETTINGS_CHANGED"] = "Telemetry Settings Changed";
|
|
6152
5887
|
TelemetryEventName2["MODEL_CACHE_EMPTY_RESPONSE"] = "Model Cache Empty Response";
|
|
5888
|
+
TelemetryEventName2["READ_FILE_LEGACY_FORMAT_USED"] = "Read File Legacy Format Used";
|
|
6153
5889
|
return TelemetryEventName2;
|
|
6154
5890
|
})(TelemetryEventName || {});
|
|
6155
5891
|
var staticAppPropertiesSchema = import_zod11.z.object({
|
|
@@ -6238,7 +5974,8 @@ var rooCodeTelemetryEventSchema = import_zod11.z.discriminatedUnion("type", [
|
|
|
6238
5974
|
"Sliding Window Truncation" /* SLIDING_WINDOW_TRUNCATION */,
|
|
6239
5975
|
"Tab Shown" /* TAB_SHOWN */,
|
|
6240
5976
|
"Mode Setting Changed" /* MODE_SETTINGS_CHANGED */,
|
|
6241
|
-
"Custom Mode Created" /* CUSTOM_MODE_CREATED
|
|
5977
|
+
"Custom Mode Created" /* CUSTOM_MODE_CREATED */,
|
|
5978
|
+
"Read File Legacy Format Used" /* READ_FILE_LEGACY_FORMAT_USED */
|
|
6242
5979
|
]),
|
|
6243
5980
|
properties: telemetryPropertiesSchema
|
|
6244
5981
|
}),
|
|
@@ -6552,7 +6289,15 @@ var isLanguage = (value) => languages.includes(value);
|
|
|
6552
6289
|
|
|
6553
6290
|
// src/global-settings.ts
|
|
6554
6291
|
var DEFAULT_WRITE_DELAY_MS = 1e3;
|
|
6555
|
-
var
|
|
6292
|
+
var TERMINAL_PREVIEW_BYTES = {
|
|
6293
|
+
small: 5 * 1024,
|
|
6294
|
+
// 5KB
|
|
6295
|
+
medium: 10 * 1024,
|
|
6296
|
+
// 10KB
|
|
6297
|
+
large: 20 * 1024
|
|
6298
|
+
// 20KB
|
|
6299
|
+
};
|
|
6300
|
+
var DEFAULT_TERMINAL_OUTPUT_PREVIEW_SIZE = "medium";
|
|
6556
6301
|
var MIN_CHECKPOINT_TIMEOUT_SECONDS = 10;
|
|
6557
6302
|
var MAX_CHECKPOINT_TIMEOUT_SECONDS = 60;
|
|
6558
6303
|
var DEFAULT_CHECKPOINT_TIMEOUT_SECONDS = 15;
|
|
@@ -6568,7 +6313,6 @@ var globalSettingsSchema = import_zod14.z.object({
|
|
|
6568
6313
|
imageGenerationProvider: import_zod14.z.enum(["openrouter", "roo"]).optional(),
|
|
6569
6314
|
openRouterImageApiKey: import_zod14.z.string().optional(),
|
|
6570
6315
|
openRouterImageGenerationSelectedModel: import_zod14.z.string().optional(),
|
|
6571
|
-
condensingApiConfigId: import_zod14.z.string().optional(),
|
|
6572
6316
|
customCondensingPrompt: import_zod14.z.string().optional(),
|
|
6573
6317
|
autoApprovalEnabled: import_zod14.z.boolean().optional(),
|
|
6574
6318
|
alwaysAllowReadOnly: import_zod14.z.boolean().optional(),
|
|
@@ -6594,7 +6338,6 @@ var globalSettingsSchema = import_zod14.z.object({
|
|
|
6594
6338
|
allowedMaxCost: import_zod14.z.number().nullish(),
|
|
6595
6339
|
autoCondenseContext: import_zod14.z.boolean().optional(),
|
|
6596
6340
|
autoCondenseContextPercent: import_zod14.z.number().optional(),
|
|
6597
|
-
maxConcurrentFileReads: import_zod14.z.number().optional(),
|
|
6598
6341
|
/**
|
|
6599
6342
|
* Whether to include current time in the environment details
|
|
6600
6343
|
* @default true
|
|
@@ -6637,11 +6380,9 @@ var globalSettingsSchema = import_zod14.z.object({
|
|
|
6637
6380
|
maxWorkspaceFiles: import_zod14.z.number().optional(),
|
|
6638
6381
|
showRooIgnoredFiles: import_zod14.z.boolean().optional(),
|
|
6639
6382
|
enableSubfolderRules: import_zod14.z.boolean().optional(),
|
|
6640
|
-
maxReadFileLine: import_zod14.z.number().optional(),
|
|
6641
6383
|
maxImageFileSize: import_zod14.z.number().optional(),
|
|
6642
6384
|
maxTotalImageSize: import_zod14.z.number().optional(),
|
|
6643
|
-
|
|
6644
|
-
terminalOutputCharacterLimit: import_zod14.z.number().optional(),
|
|
6385
|
+
terminalOutputPreviewSize: import_zod14.z.enum(["small", "medium", "large"]).optional(),
|
|
6645
6386
|
terminalShellIntegrationTimeout: import_zod14.z.number().optional(),
|
|
6646
6387
|
terminalShellIntegrationDisabled: import_zod14.z.boolean().optional(),
|
|
6647
6388
|
terminalCommandDelay: import_zod14.z.number().optional(),
|
|
@@ -6650,18 +6391,14 @@ var globalSettingsSchema = import_zod14.z.object({
|
|
|
6650
6391
|
terminalZshOhMy: import_zod14.z.boolean().optional(),
|
|
6651
6392
|
terminalZshP10k: import_zod14.z.boolean().optional(),
|
|
6652
6393
|
terminalZdotdir: import_zod14.z.boolean().optional(),
|
|
6653
|
-
terminalCompressProgressBar: import_zod14.z.boolean().optional(),
|
|
6654
6394
|
diagnosticsEnabled: import_zod14.z.boolean().optional(),
|
|
6655
6395
|
rateLimitSeconds: import_zod14.z.number().optional(),
|
|
6656
|
-
diffEnabled: import_zod14.z.boolean().optional(),
|
|
6657
|
-
fuzzyMatchThreshold: import_zod14.z.number().optional(),
|
|
6658
6396
|
experiments: experimentsSchema.optional(),
|
|
6659
6397
|
codebaseIndexModels: codebaseIndexModelsSchema.optional(),
|
|
6660
6398
|
codebaseIndexConfig: codebaseIndexConfigSchema.optional(),
|
|
6661
6399
|
language: languagesSchema.optional(),
|
|
6662
6400
|
telemetrySetting: telemetrySettingsSchema.optional(),
|
|
6663
6401
|
mcpEnabled: import_zod14.z.boolean().optional(),
|
|
6664
|
-
enableMcpServerCreation: import_zod14.z.boolean().optional(),
|
|
6665
6402
|
mode: import_zod14.z.string().optional(),
|
|
6666
6403
|
modeApiConfigs: import_zod14.z.record(import_zod14.z.string(), import_zod14.z.string()).optional(),
|
|
6667
6404
|
customModes: import_zod14.z.array(modeConfigSchema).optional(),
|
|
@@ -6681,7 +6418,20 @@ var globalSettingsSchema = import_zod14.z.object({
|
|
|
6681
6418
|
profileThresholds: import_zod14.z.record(import_zod14.z.string(), import_zod14.z.number()).optional(),
|
|
6682
6419
|
hasOpenedModeSelector: import_zod14.z.boolean().optional(),
|
|
6683
6420
|
lastModeExportPath: import_zod14.z.string().optional(),
|
|
6684
|
-
lastModeImportPath: import_zod14.z.string().optional()
|
|
6421
|
+
lastModeImportPath: import_zod14.z.string().optional(),
|
|
6422
|
+
lastSettingsExportPath: import_zod14.z.string().optional(),
|
|
6423
|
+
lastTaskExportPath: import_zod14.z.string().optional(),
|
|
6424
|
+
lastImageSavePath: import_zod14.z.string().optional(),
|
|
6425
|
+
/**
|
|
6426
|
+
* Path to worktree to auto-open after switching workspaces.
|
|
6427
|
+
* Used by the worktree feature to open the Roo Code sidebar in a new window.
|
|
6428
|
+
*/
|
|
6429
|
+
worktreeAutoOpenPath: import_zod14.z.string().optional(),
|
|
6430
|
+
/**
|
|
6431
|
+
* Whether to show the worktree selector in the home screen.
|
|
6432
|
+
* @default true
|
|
6433
|
+
*/
|
|
6434
|
+
showWorktreesInHomeScreen: import_zod14.z.boolean().optional()
|
|
6685
6435
|
});
|
|
6686
6436
|
var GLOBAL_SETTINGS_KEYS = globalSettingsSchema.keyof().options;
|
|
6687
6437
|
var rooCodeSettingsSchema = providerSettingsSchema.merge(globalSettingsSchema);
|
|
@@ -6765,8 +6515,6 @@ var EVALS_SETTINGS = {
|
|
|
6765
6515
|
ttsSpeed: 1,
|
|
6766
6516
|
soundEnabled: false,
|
|
6767
6517
|
soundVolume: 0.5,
|
|
6768
|
-
terminalOutputLineLimit: 500,
|
|
6769
|
-
terminalOutputCharacterLimit: DEFAULT_TERMINAL_OUTPUT_CHARACTER_LIMIT,
|
|
6770
6518
|
terminalShellIntegrationTimeout: 3e4,
|
|
6771
6519
|
terminalCommandDelay: 0,
|
|
6772
6520
|
terminalPowershellCounter: false,
|
|
@@ -6774,19 +6522,14 @@ var EVALS_SETTINGS = {
|
|
|
6774
6522
|
terminalZshClearEolMark: true,
|
|
6775
6523
|
terminalZshP10k: false,
|
|
6776
6524
|
terminalZdotdir: true,
|
|
6777
|
-
terminalCompressProgressBar: true,
|
|
6778
6525
|
terminalShellIntegrationDisabled: true,
|
|
6779
6526
|
diagnosticsEnabled: true,
|
|
6780
|
-
diffEnabled: true,
|
|
6781
|
-
fuzzyMatchThreshold: 1,
|
|
6782
6527
|
enableCheckpoints: false,
|
|
6783
6528
|
rateLimitSeconds: 0,
|
|
6784
6529
|
maxOpenTabsContext: 20,
|
|
6785
6530
|
maxWorkspaceFiles: 200,
|
|
6786
6531
|
maxGitStatusFiles: 20,
|
|
6787
6532
|
showRooIgnoredFiles: true,
|
|
6788
|
-
maxReadFileLine: -1,
|
|
6789
|
-
// -1 to enable full file reading.
|
|
6790
6533
|
includeDiagnosticMessages: true,
|
|
6791
6534
|
maxDiagnosticMessages: 50,
|
|
6792
6535
|
language: "en",
|
|
@@ -6860,24 +6603,18 @@ var organizationAllowListSchema = import_zod16.z.object({
|
|
|
6860
6603
|
});
|
|
6861
6604
|
var organizationDefaultSettingsSchema = globalSettingsSchema.pick({
|
|
6862
6605
|
enableCheckpoints: true,
|
|
6863
|
-
fuzzyMatchThreshold: true,
|
|
6864
6606
|
maxOpenTabsContext: true,
|
|
6865
|
-
maxReadFileLine: true,
|
|
6866
6607
|
maxWorkspaceFiles: true,
|
|
6867
6608
|
showRooIgnoredFiles: true,
|
|
6868
6609
|
terminalCommandDelay: true,
|
|
6869
|
-
terminalCompressProgressBar: true,
|
|
6870
|
-
terminalOutputLineLimit: true,
|
|
6871
6610
|
terminalShellIntegrationDisabled: true,
|
|
6872
6611
|
terminalShellIntegrationTimeout: true,
|
|
6873
6612
|
terminalZshClearEolMark: true
|
|
6874
6613
|
}).merge(
|
|
6875
6614
|
import_zod16.z.object({
|
|
6876
6615
|
maxOpenTabsContext: import_zod16.z.number().int().nonnegative().optional(),
|
|
6877
|
-
maxReadFileLine: import_zod16.z.number().int().gte(-1).optional(),
|
|
6878
6616
|
maxWorkspaceFiles: import_zod16.z.number().int().nonnegative().optional(),
|
|
6879
6617
|
terminalCommandDelay: import_zod16.z.number().int().nonnegative().optional(),
|
|
6880
|
-
terminalOutputLineLimit: import_zod16.z.number().int().nonnegative().optional(),
|
|
6881
6618
|
terminalShellIntegrationTimeout: import_zod16.z.number().int().nonnegative().optional()
|
|
6882
6619
|
})
|
|
6883
6620
|
);
|
|
@@ -6910,7 +6647,8 @@ var userFeaturesSchema = import_zod16.z.object({
|
|
|
6910
6647
|
});
|
|
6911
6648
|
var userSettingsConfigSchema = import_zod16.z.object({
|
|
6912
6649
|
extensionBridgeEnabled: import_zod16.z.boolean().optional(),
|
|
6913
|
-
taskSyncEnabled: import_zod16.z.boolean().optional()
|
|
6650
|
+
taskSyncEnabled: import_zod16.z.boolean().optional(),
|
|
6651
|
+
llmEnhancedFeaturesEnabled: import_zod16.z.boolean().optional()
|
|
6914
6652
|
});
|
|
6915
6653
|
var userSettingsDataSchema = import_zod16.z.object({
|
|
6916
6654
|
features: userFeaturesSchema,
|
|
@@ -7351,12 +7089,10 @@ var taskCommandSchema = import_zod18.z.discriminatedUnion("commandName", [
|
|
|
7351
7089
|
})
|
|
7352
7090
|
}),
|
|
7353
7091
|
import_zod18.z.object({
|
|
7354
|
-
commandName: import_zod18.z.literal("CancelTask" /* CancelTask */)
|
|
7355
|
-
data: import_zod18.z.string()
|
|
7092
|
+
commandName: import_zod18.z.literal("CancelTask" /* CancelTask */)
|
|
7356
7093
|
}),
|
|
7357
7094
|
import_zod18.z.object({
|
|
7358
|
-
commandName: import_zod18.z.literal("CloseTask" /* CloseTask */)
|
|
7359
|
-
data: import_zod18.z.string()
|
|
7095
|
+
commandName: import_zod18.z.literal("CloseTask" /* CloseTask */)
|
|
7360
7096
|
}),
|
|
7361
7097
|
import_zod18.z.object({
|
|
7362
7098
|
commandName: import_zod18.z.literal("ResumeTask" /* ResumeTask */),
|
|
@@ -7392,6 +7128,7 @@ var ipcMessageSchema = import_zod18.z.discriminatedUnion("type", [
|
|
|
7392
7128
|
|
|
7393
7129
|
// src/mcp.ts
|
|
7394
7130
|
var import_zod19 = require("zod");
|
|
7131
|
+
var MAX_MCP_TOOLS_THRESHOLD = 60;
|
|
7395
7132
|
var mcpExecutionStatusSchema = import_zod19.z.discriminatedUnion("status", [
|
|
7396
7133
|
import_zod19.z.object({
|
|
7397
7134
|
executionId: import_zod19.z.string(),
|
|
@@ -7415,6 +7152,46 @@ var mcpExecutionStatusSchema = import_zod19.z.discriminatedUnion("status", [
|
|
|
7415
7152
|
error: import_zod19.z.string().optional()
|
|
7416
7153
|
})
|
|
7417
7154
|
]);
|
|
7155
|
+
function countEnabledMcpTools(servers) {
|
|
7156
|
+
let serverCount = 0;
|
|
7157
|
+
let toolCount = 0;
|
|
7158
|
+
for (const server of servers) {
|
|
7159
|
+
if (server.disabled) continue;
|
|
7160
|
+
if (server.status !== "connected") continue;
|
|
7161
|
+
serverCount++;
|
|
7162
|
+
if (server.tools) {
|
|
7163
|
+
for (const tool of server.tools) {
|
|
7164
|
+
if (tool.enabledForPrompt !== false) {
|
|
7165
|
+
toolCount++;
|
|
7166
|
+
}
|
|
7167
|
+
}
|
|
7168
|
+
}
|
|
7169
|
+
}
|
|
7170
|
+
return { enabledToolCount: toolCount, enabledServerCount: serverCount };
|
|
7171
|
+
}
|
|
7172
|
+
|
|
7173
|
+
// src/skills.ts
|
|
7174
|
+
var SKILL_NAME_MIN_LENGTH = 1;
|
|
7175
|
+
var SKILL_NAME_MAX_LENGTH = 64;
|
|
7176
|
+
var SKILL_NAME_REGEX = /^[a-z0-9]+(?:-[a-z0-9]+)*$/;
|
|
7177
|
+
var SkillNameValidationError = /* @__PURE__ */ ((SkillNameValidationError2) => {
|
|
7178
|
+
SkillNameValidationError2["Empty"] = "empty";
|
|
7179
|
+
SkillNameValidationError2["TooLong"] = "too_long";
|
|
7180
|
+
SkillNameValidationError2["InvalidFormat"] = "invalid_format";
|
|
7181
|
+
return SkillNameValidationError2;
|
|
7182
|
+
})(SkillNameValidationError || {});
|
|
7183
|
+
function validateSkillName(name) {
|
|
7184
|
+
if (!name || name.length < SKILL_NAME_MIN_LENGTH) {
|
|
7185
|
+
return { valid: false, error: "empty" /* Empty */ };
|
|
7186
|
+
}
|
|
7187
|
+
if (name.length > SKILL_NAME_MAX_LENGTH) {
|
|
7188
|
+
return { valid: false, error: "too_long" /* TooLong */ };
|
|
7189
|
+
}
|
|
7190
|
+
if (!SKILL_NAME_REGEX.test(name)) {
|
|
7191
|
+
return { valid: false, error: "invalid_format" /* InvalidFormat */ };
|
|
7192
|
+
}
|
|
7193
|
+
return { valid: true };
|
|
7194
|
+
}
|
|
7418
7195
|
|
|
7419
7196
|
// src/todo.ts
|
|
7420
7197
|
var import_zod20 = require("zod");
|
|
@@ -7454,6 +7231,11 @@ var commandExecutionStatusSchema = import_zod21.z.discriminatedUnion("status", [
|
|
|
7454
7231
|
})
|
|
7455
7232
|
]);
|
|
7456
7233
|
|
|
7234
|
+
// src/tool-params.ts
|
|
7235
|
+
function isLegacyReadFileParams(params) {
|
|
7236
|
+
return "_legacyFormat" in params && params._legacyFormat === true;
|
|
7237
|
+
}
|
|
7238
|
+
|
|
7457
7239
|
// src/vscode-extension-host.ts
|
|
7458
7240
|
var import_zod22 = require("zod");
|
|
7459
7241
|
var checkoutDiffPayloadSchema = import_zod22.z.object({
|
|
@@ -7507,7 +7289,7 @@ var browserActions = [
|
|
|
7507
7289
|
DEFAULT_CHECKPOINT_TIMEOUT_SECONDS,
|
|
7508
7290
|
DEFAULT_CONSECUTIVE_MISTAKE_LIMIT,
|
|
7509
7291
|
DEFAULT_MODES,
|
|
7510
|
-
|
|
7292
|
+
DEFAULT_TERMINAL_OUTPUT_PREVIEW_SIZE,
|
|
7511
7293
|
DEFAULT_WRITE_DELAY_MS,
|
|
7512
7294
|
DOUBAO_API_BASE_URL,
|
|
7513
7295
|
DOUBAO_API_CHAT_PATH,
|
|
@@ -7537,13 +7319,13 @@ var browserActions = [
|
|
|
7537
7319
|
IpcOrigin,
|
|
7538
7320
|
LMSTUDIO_DEFAULT_TEMPERATURE,
|
|
7539
7321
|
MAX_CHECKPOINT_TIMEOUT_SECONDS,
|
|
7322
|
+
MAX_MCP_TOOLS_THRESHOLD,
|
|
7540
7323
|
MINIMAX_DEFAULT_MAX_TOKENS,
|
|
7541
7324
|
MINIMAX_DEFAULT_TEMPERATURE,
|
|
7542
7325
|
MIN_CHECKPOINT_TIMEOUT_SECONDS,
|
|
7543
7326
|
MISTRAL_DEFAULT_TEMPERATURE,
|
|
7544
7327
|
MODELS_BY_PROVIDER,
|
|
7545
7328
|
MOONSHOT_DEFAULT_TEMPERATURE,
|
|
7546
|
-
NATIVE_TOOL_DEFAULTS,
|
|
7547
7329
|
OPENAI_AZURE_AI_INFERENCE_PATH,
|
|
7548
7330
|
OPENAI_NATIVE_DEFAULT_TEMPERATURE,
|
|
7549
7331
|
OPENROUTER_DEFAULT_PROVIDER_NAME,
|
|
@@ -7558,7 +7340,11 @@ var browserActions = [
|
|
|
7558
7340
|
RooModelsResponseSchema,
|
|
7559
7341
|
RooPricingSchema,
|
|
7560
7342
|
SECRET_STATE_KEYS,
|
|
7561
|
-
|
|
7343
|
+
SKILL_NAME_MAX_LENGTH,
|
|
7344
|
+
SKILL_NAME_MIN_LENGTH,
|
|
7345
|
+
SKILL_NAME_REGEX,
|
|
7346
|
+
SkillNameValidationError,
|
|
7347
|
+
TERMINAL_PREVIEW_BYTES,
|
|
7562
7348
|
TaskBridgeCommandName,
|
|
7563
7349
|
TaskBridgeEventName,
|
|
7564
7350
|
TaskCommandName,
|
|
@@ -7590,9 +7376,6 @@ var browserActions = [
|
|
|
7590
7376
|
chutesDefaultModelId,
|
|
7591
7377
|
chutesDefaultModelInfo,
|
|
7592
7378
|
chutesModels,
|
|
7593
|
-
claudeCodeDefaultModelId,
|
|
7594
|
-
claudeCodeModels,
|
|
7595
|
-
claudeCodeReasoningConfig,
|
|
7596
7379
|
clineAskSchema,
|
|
7597
7380
|
clineAsks,
|
|
7598
7381
|
clineMessageSchema,
|
|
@@ -7607,6 +7390,7 @@ var browserActions = [
|
|
|
7607
7390
|
commandIds,
|
|
7608
7391
|
contextCondenseSchema,
|
|
7609
7392
|
contextTruncationSchema,
|
|
7393
|
+
countEnabledMcpTools,
|
|
7610
7394
|
customModePromptsSchema,
|
|
7611
7395
|
customModesSettingsSchema,
|
|
7612
7396
|
customProviders,
|
|
@@ -7640,7 +7424,6 @@ var browserActions = [
|
|
|
7640
7424
|
geminiDefaultModelId,
|
|
7641
7425
|
geminiModels,
|
|
7642
7426
|
getApiProtocol,
|
|
7643
|
-
getEffectiveProtocol,
|
|
7644
7427
|
getErrorMessage,
|
|
7645
7428
|
getErrorStatusCode,
|
|
7646
7429
|
getImageGenerationProvider,
|
|
@@ -7675,9 +7458,9 @@ var browserActions = [
|
|
|
7675
7458
|
isInteractiveAsk,
|
|
7676
7459
|
isInternalProvider,
|
|
7677
7460
|
isLanguage,
|
|
7461
|
+
isLegacyReadFileParams,
|
|
7678
7462
|
isLocalProvider,
|
|
7679
7463
|
isModelParameter,
|
|
7680
|
-
isNativeProtocol,
|
|
7681
7464
|
isNonBlockingAsk,
|
|
7682
7465
|
isProviderName,
|
|
7683
7466
|
isResumableAsk,
|
|
@@ -7713,9 +7496,10 @@ var browserActions = [
|
|
|
7713
7496
|
moonshotDefaultModelId,
|
|
7714
7497
|
moonshotModels,
|
|
7715
7498
|
nonBlockingAsks,
|
|
7716
|
-
normalizeClaudeCodeModelId,
|
|
7717
7499
|
ollamaDefaultModelId,
|
|
7718
7500
|
ollamaDefaultModelInfo,
|
|
7501
|
+
openAiCodexDefaultModelId,
|
|
7502
|
+
openAiCodexModels,
|
|
7719
7503
|
openAiModelInfoSaneDefaults,
|
|
7720
7504
|
openAiNativeDefaultModelId,
|
|
7721
7505
|
openAiNativeModels,
|
|
@@ -7785,6 +7569,7 @@ var browserActions = [
|
|
|
7785
7569
|
userFeaturesSchema,
|
|
7786
7570
|
userSettingsConfigSchema,
|
|
7787
7571
|
userSettingsDataSchema,
|
|
7572
|
+
validateSkillName,
|
|
7788
7573
|
verbosityLevels,
|
|
7789
7574
|
verbosityLevelsSchema,
|
|
7790
7575
|
vercelAiGatewayDefaultModelId,
|