@mariozechner/pi-ai 0.5.43 → 0.5.44
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/models.generated.d.ts +456 -235
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +990 -769
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/anthropic.js +24 -1
- package/dist/providers/anthropic.js.map +1 -1
- package/dist/providers/google.js +1 -1
- package/dist/providers/google.js.map +1 -1
- package/dist/providers/transorm-messages.d.ts.map +1 -1
- package/dist/providers/transorm-messages.js +46 -1
- package/dist/providers/transorm-messages.js.map +1 -1
- package/dist/utils/event-stream.js +3 -7
- package/dist/utils/event-stream.js.map +1 -1
- package/dist/utils/validation.d.ts.map +1 -1
- package/dist/utils/validation.js +25 -3
- package/dist/utils/validation.js.map +1 -1
- package/package.json +3 -3
package/dist/models.generated.js
CHANGED
|
@@ -2,13 +2,13 @@
|
|
|
2
2
|
// Do not edit manually - run 'npm run generate-models' to update
|
|
3
3
|
export const MODELS = {
|
|
4
4
|
anthropic: {
|
|
5
|
-
"claude-3-
|
|
6
|
-
id: "claude-3-
|
|
7
|
-
name: "Claude Sonnet 3.
|
|
5
|
+
"claude-3-5-sonnet-20241022": {
|
|
6
|
+
id: "claude-3-5-sonnet-20241022",
|
|
7
|
+
name: "Claude Sonnet 3.5 v2",
|
|
8
8
|
api: "anthropic-messages",
|
|
9
9
|
provider: "anthropic",
|
|
10
10
|
baseUrl: "https://api.anthropic.com",
|
|
11
|
-
reasoning:
|
|
11
|
+
reasoning: false,
|
|
12
12
|
input: ["text", "image"],
|
|
13
13
|
cost: {
|
|
14
14
|
input: 3,
|
|
@@ -17,15 +17,32 @@ export const MODELS = {
|
|
|
17
17
|
cacheWrite: 3.75,
|
|
18
18
|
},
|
|
19
19
|
contextWindow: 200000,
|
|
20
|
-
maxTokens:
|
|
20
|
+
maxTokens: 8192,
|
|
21
21
|
},
|
|
22
|
-
"claude-
|
|
23
|
-
id: "claude-
|
|
24
|
-
name: "Claude
|
|
22
|
+
"claude-3-5-sonnet-20240620": {
|
|
23
|
+
id: "claude-3-5-sonnet-20240620",
|
|
24
|
+
name: "Claude Sonnet 3.5",
|
|
25
25
|
api: "anthropic-messages",
|
|
26
26
|
provider: "anthropic",
|
|
27
27
|
baseUrl: "https://api.anthropic.com",
|
|
28
|
-
reasoning:
|
|
28
|
+
reasoning: false,
|
|
29
|
+
input: ["text", "image"],
|
|
30
|
+
cost: {
|
|
31
|
+
input: 3,
|
|
32
|
+
output: 15,
|
|
33
|
+
cacheRead: 0.3,
|
|
34
|
+
cacheWrite: 3.75,
|
|
35
|
+
},
|
|
36
|
+
contextWindow: 200000,
|
|
37
|
+
maxTokens: 8192,
|
|
38
|
+
},
|
|
39
|
+
"claude-3-opus-20240229": {
|
|
40
|
+
id: "claude-3-opus-20240229",
|
|
41
|
+
name: "Claude Opus 3",
|
|
42
|
+
api: "anthropic-messages",
|
|
43
|
+
provider: "anthropic",
|
|
44
|
+
baseUrl: "https://api.anthropic.com",
|
|
45
|
+
reasoning: false,
|
|
29
46
|
input: ["text", "image"],
|
|
30
47
|
cost: {
|
|
31
48
|
input: 15,
|
|
@@ -34,41 +51,41 @@ export const MODELS = {
|
|
|
34
51
|
cacheWrite: 18.75,
|
|
35
52
|
},
|
|
36
53
|
contextWindow: 200000,
|
|
37
|
-
maxTokens:
|
|
54
|
+
maxTokens: 4096,
|
|
38
55
|
},
|
|
39
|
-
"claude-
|
|
40
|
-
id: "claude-
|
|
41
|
-
name: "Claude
|
|
56
|
+
"claude-sonnet-4-5-20250929": {
|
|
57
|
+
id: "claude-sonnet-4-5-20250929",
|
|
58
|
+
name: "Claude Sonnet 4.5",
|
|
42
59
|
api: "anthropic-messages",
|
|
43
60
|
provider: "anthropic",
|
|
44
61
|
baseUrl: "https://api.anthropic.com",
|
|
45
|
-
reasoning:
|
|
62
|
+
reasoning: true,
|
|
46
63
|
input: ["text", "image"],
|
|
47
64
|
cost: {
|
|
48
|
-
input:
|
|
49
|
-
output:
|
|
50
|
-
cacheRead: 0.
|
|
51
|
-
cacheWrite:
|
|
65
|
+
input: 3,
|
|
66
|
+
output: 15,
|
|
67
|
+
cacheRead: 0.3,
|
|
68
|
+
cacheWrite: 3.75,
|
|
52
69
|
},
|
|
53
70
|
contextWindow: 200000,
|
|
54
|
-
maxTokens:
|
|
71
|
+
maxTokens: 64000,
|
|
55
72
|
},
|
|
56
|
-
"claude-
|
|
57
|
-
id: "claude-
|
|
58
|
-
name: "Claude
|
|
73
|
+
"claude-sonnet-4-20250514": {
|
|
74
|
+
id: "claude-sonnet-4-20250514",
|
|
75
|
+
name: "Claude Sonnet 4",
|
|
59
76
|
api: "anthropic-messages",
|
|
60
77
|
provider: "anthropic",
|
|
61
78
|
baseUrl: "https://api.anthropic.com",
|
|
62
|
-
reasoning:
|
|
79
|
+
reasoning: true,
|
|
63
80
|
input: ["text", "image"],
|
|
64
81
|
cost: {
|
|
65
|
-
input:
|
|
66
|
-
output:
|
|
67
|
-
cacheRead: 0.
|
|
68
|
-
cacheWrite:
|
|
82
|
+
input: 3,
|
|
83
|
+
output: 15,
|
|
84
|
+
cacheRead: 0.3,
|
|
85
|
+
cacheWrite: 3.75,
|
|
69
86
|
},
|
|
70
87
|
contextWindow: 200000,
|
|
71
|
-
maxTokens:
|
|
88
|
+
maxTokens: 64000,
|
|
72
89
|
},
|
|
73
90
|
"claude-opus-4-20250514": {
|
|
74
91
|
id: "claude-opus-4-20250514",
|
|
@@ -87,87 +104,87 @@ export const MODELS = {
|
|
|
87
104
|
contextWindow: 200000,
|
|
88
105
|
maxTokens: 32000,
|
|
89
106
|
},
|
|
90
|
-
"claude-3-5-
|
|
91
|
-
id: "claude-3-5-
|
|
92
|
-
name: "Claude
|
|
107
|
+
"claude-3-5-haiku-20241022": {
|
|
108
|
+
id: "claude-3-5-haiku-20241022",
|
|
109
|
+
name: "Claude Haiku 3.5",
|
|
93
110
|
api: "anthropic-messages",
|
|
94
111
|
provider: "anthropic",
|
|
95
112
|
baseUrl: "https://api.anthropic.com",
|
|
96
113
|
reasoning: false,
|
|
97
114
|
input: ["text", "image"],
|
|
98
115
|
cost: {
|
|
99
|
-
input:
|
|
100
|
-
output:
|
|
101
|
-
cacheRead: 0.
|
|
102
|
-
cacheWrite:
|
|
116
|
+
input: 0.8,
|
|
117
|
+
output: 4,
|
|
118
|
+
cacheRead: 0.08,
|
|
119
|
+
cacheWrite: 1,
|
|
103
120
|
},
|
|
104
121
|
contextWindow: 200000,
|
|
105
122
|
maxTokens: 8192,
|
|
106
123
|
},
|
|
107
|
-
"claude-3-
|
|
108
|
-
id: "claude-3-
|
|
109
|
-
name: "Claude
|
|
124
|
+
"claude-3-haiku-20240307": {
|
|
125
|
+
id: "claude-3-haiku-20240307",
|
|
126
|
+
name: "Claude Haiku 3",
|
|
110
127
|
api: "anthropic-messages",
|
|
111
128
|
provider: "anthropic",
|
|
112
129
|
baseUrl: "https://api.anthropic.com",
|
|
113
130
|
reasoning: false,
|
|
114
131
|
input: ["text", "image"],
|
|
115
132
|
cost: {
|
|
116
|
-
input:
|
|
117
|
-
output:
|
|
118
|
-
cacheRead: 0.
|
|
119
|
-
cacheWrite: 3
|
|
133
|
+
input: 0.25,
|
|
134
|
+
output: 1.25,
|
|
135
|
+
cacheRead: 0.03,
|
|
136
|
+
cacheWrite: 0.3,
|
|
120
137
|
},
|
|
121
138
|
contextWindow: 200000,
|
|
122
|
-
maxTokens:
|
|
139
|
+
maxTokens: 4096,
|
|
123
140
|
},
|
|
124
|
-
"claude-3-sonnet-
|
|
125
|
-
id: "claude-3-sonnet-
|
|
126
|
-
name: "Claude Sonnet 3",
|
|
141
|
+
"claude-3-7-sonnet-20250219": {
|
|
142
|
+
id: "claude-3-7-sonnet-20250219",
|
|
143
|
+
name: "Claude Sonnet 3.7",
|
|
127
144
|
api: "anthropic-messages",
|
|
128
145
|
provider: "anthropic",
|
|
129
146
|
baseUrl: "https://api.anthropic.com",
|
|
130
|
-
reasoning:
|
|
147
|
+
reasoning: true,
|
|
131
148
|
input: ["text", "image"],
|
|
132
149
|
cost: {
|
|
133
150
|
input: 3,
|
|
134
151
|
output: 15,
|
|
135
152
|
cacheRead: 0.3,
|
|
136
|
-
cacheWrite:
|
|
153
|
+
cacheWrite: 3.75,
|
|
137
154
|
},
|
|
138
155
|
contextWindow: 200000,
|
|
139
|
-
maxTokens:
|
|
156
|
+
maxTokens: 64000,
|
|
140
157
|
},
|
|
141
|
-
"claude-
|
|
142
|
-
id: "claude-
|
|
143
|
-
name: "Claude
|
|
158
|
+
"claude-opus-4-1-20250805": {
|
|
159
|
+
id: "claude-opus-4-1-20250805",
|
|
160
|
+
name: "Claude Opus 4.1",
|
|
144
161
|
api: "anthropic-messages",
|
|
145
162
|
provider: "anthropic",
|
|
146
163
|
baseUrl: "https://api.anthropic.com",
|
|
147
164
|
reasoning: true,
|
|
148
165
|
input: ["text", "image"],
|
|
149
166
|
cost: {
|
|
150
|
-
input:
|
|
151
|
-
output:
|
|
152
|
-
cacheRead:
|
|
153
|
-
cacheWrite:
|
|
167
|
+
input: 15,
|
|
168
|
+
output: 75,
|
|
169
|
+
cacheRead: 1.5,
|
|
170
|
+
cacheWrite: 18.75,
|
|
154
171
|
},
|
|
155
172
|
contextWindow: 200000,
|
|
156
|
-
maxTokens:
|
|
173
|
+
maxTokens: 32000,
|
|
157
174
|
},
|
|
158
|
-
"claude-3-
|
|
159
|
-
id: "claude-3-
|
|
160
|
-
name: "Claude
|
|
175
|
+
"claude-3-sonnet-20240229": {
|
|
176
|
+
id: "claude-3-sonnet-20240229",
|
|
177
|
+
name: "Claude Sonnet 3",
|
|
161
178
|
api: "anthropic-messages",
|
|
162
179
|
provider: "anthropic",
|
|
163
180
|
baseUrl: "https://api.anthropic.com",
|
|
164
181
|
reasoning: false,
|
|
165
182
|
input: ["text", "image"],
|
|
166
183
|
cost: {
|
|
167
|
-
input:
|
|
168
|
-
output:
|
|
169
|
-
cacheRead:
|
|
170
|
-
cacheWrite:
|
|
184
|
+
input: 3,
|
|
185
|
+
output: 15,
|
|
186
|
+
cacheRead: 0.3,
|
|
187
|
+
cacheWrite: 0.3,
|
|
171
188
|
},
|
|
172
189
|
contextWindow: 200000,
|
|
173
190
|
maxTokens: 4096,
|
|
@@ -191,39 +208,73 @@ export const MODELS = {
|
|
|
191
208
|
contextWindow: 1048576,
|
|
192
209
|
maxTokens: 65536,
|
|
193
210
|
},
|
|
194
|
-
"gemini-
|
|
195
|
-
id: "gemini-
|
|
196
|
-
name: "Gemini
|
|
211
|
+
"gemini-flash-lite-latest": {
|
|
212
|
+
id: "gemini-flash-lite-latest",
|
|
213
|
+
name: "Gemini Flash-Lite Latest",
|
|
197
214
|
api: "google-generative-ai",
|
|
198
215
|
provider: "google",
|
|
199
216
|
baseUrl: "https://generativelanguage.googleapis.com/v1beta",
|
|
200
217
|
reasoning: true,
|
|
201
218
|
input: ["text", "image"],
|
|
202
219
|
cost: {
|
|
203
|
-
input: 1
|
|
204
|
-
output:
|
|
205
|
-
cacheRead: 0.
|
|
220
|
+
input: 0.1,
|
|
221
|
+
output: 0.4,
|
|
222
|
+
cacheRead: 0.025,
|
|
206
223
|
cacheWrite: 0,
|
|
207
224
|
},
|
|
208
225
|
contextWindow: 1048576,
|
|
209
226
|
maxTokens: 65536,
|
|
210
227
|
},
|
|
211
|
-
"gemini-
|
|
212
|
-
id: "gemini-
|
|
213
|
-
name: "Gemini
|
|
228
|
+
"gemini-2.5-flash": {
|
|
229
|
+
id: "gemini-2.5-flash",
|
|
230
|
+
name: "Gemini 2.5 Flash",
|
|
214
231
|
api: "google-generative-ai",
|
|
215
232
|
provider: "google",
|
|
216
233
|
baseUrl: "https://generativelanguage.googleapis.com/v1beta",
|
|
217
|
-
reasoning:
|
|
234
|
+
reasoning: true,
|
|
218
235
|
input: ["text", "image"],
|
|
219
236
|
cost: {
|
|
220
|
-
input: 0.
|
|
221
|
-
output:
|
|
222
|
-
cacheRead: 0.
|
|
237
|
+
input: 0.3,
|
|
238
|
+
output: 2.5,
|
|
239
|
+
cacheRead: 0.075,
|
|
223
240
|
cacheWrite: 0,
|
|
224
241
|
},
|
|
225
|
-
contextWindow:
|
|
226
|
-
maxTokens:
|
|
242
|
+
contextWindow: 1048576,
|
|
243
|
+
maxTokens: 65536,
|
|
244
|
+
},
|
|
245
|
+
"gemini-flash-latest": {
|
|
246
|
+
id: "gemini-flash-latest",
|
|
247
|
+
name: "Gemini Flash Latest",
|
|
248
|
+
api: "google-generative-ai",
|
|
249
|
+
provider: "google",
|
|
250
|
+
baseUrl: "https://generativelanguage.googleapis.com/v1beta",
|
|
251
|
+
reasoning: true,
|
|
252
|
+
input: ["text", "image"],
|
|
253
|
+
cost: {
|
|
254
|
+
input: 0.15,
|
|
255
|
+
output: 0.6,
|
|
256
|
+
cacheRead: 0.0375,
|
|
257
|
+
cacheWrite: 0,
|
|
258
|
+
},
|
|
259
|
+
contextWindow: 1048576,
|
|
260
|
+
maxTokens: 65536,
|
|
261
|
+
},
|
|
262
|
+
"gemini-2.5-pro-preview-05-06": {
|
|
263
|
+
id: "gemini-2.5-pro-preview-05-06",
|
|
264
|
+
name: "Gemini 2.5 Pro Preview 05-06",
|
|
265
|
+
api: "google-generative-ai",
|
|
266
|
+
provider: "google",
|
|
267
|
+
baseUrl: "https://generativelanguage.googleapis.com/v1beta",
|
|
268
|
+
reasoning: true,
|
|
269
|
+
input: ["text", "image"],
|
|
270
|
+
cost: {
|
|
271
|
+
input: 1.25,
|
|
272
|
+
output: 10,
|
|
273
|
+
cacheRead: 0.31,
|
|
274
|
+
cacheWrite: 0,
|
|
275
|
+
},
|
|
276
|
+
contextWindow: 1048576,
|
|
277
|
+
maxTokens: 65536,
|
|
227
278
|
},
|
|
228
279
|
"gemini-2.0-flash-lite": {
|
|
229
280
|
id: "gemini-2.0-flash-lite",
|
|
@@ -242,52 +293,52 @@ export const MODELS = {
|
|
|
242
293
|
contextWindow: 1048576,
|
|
243
294
|
maxTokens: 8192,
|
|
244
295
|
},
|
|
245
|
-
"gemini-
|
|
246
|
-
id: "gemini-
|
|
247
|
-
name: "Gemini
|
|
296
|
+
"gemini-live-2.5-flash-preview-native-audio": {
|
|
297
|
+
id: "gemini-live-2.5-flash-preview-native-audio",
|
|
298
|
+
name: "Gemini Live 2.5 Flash Preview Native Audio",
|
|
248
299
|
api: "google-generative-ai",
|
|
249
300
|
provider: "google",
|
|
250
301
|
baseUrl: "https://generativelanguage.googleapis.com/v1beta",
|
|
251
|
-
reasoning:
|
|
252
|
-
input: ["text"
|
|
302
|
+
reasoning: true,
|
|
303
|
+
input: ["text"],
|
|
253
304
|
cost: {
|
|
254
|
-
input:
|
|
255
|
-
output:
|
|
256
|
-
cacheRead: 0
|
|
305
|
+
input: 0.5,
|
|
306
|
+
output: 2,
|
|
307
|
+
cacheRead: 0,
|
|
257
308
|
cacheWrite: 0,
|
|
258
309
|
},
|
|
259
|
-
contextWindow:
|
|
260
|
-
maxTokens:
|
|
310
|
+
contextWindow: 131072,
|
|
311
|
+
maxTokens: 65536,
|
|
261
312
|
},
|
|
262
|
-
"gemini-
|
|
263
|
-
id: "gemini-
|
|
264
|
-
name: "Gemini
|
|
313
|
+
"gemini-2.0-flash": {
|
|
314
|
+
id: "gemini-2.0-flash",
|
|
315
|
+
name: "Gemini 2.0 Flash",
|
|
265
316
|
api: "google-generative-ai",
|
|
266
317
|
provider: "google",
|
|
267
318
|
baseUrl: "https://generativelanguage.googleapis.com/v1beta",
|
|
268
319
|
reasoning: false,
|
|
269
320
|
input: ["text", "image"],
|
|
270
321
|
cost: {
|
|
271
|
-
input: 0.
|
|
272
|
-
output: 0.
|
|
273
|
-
cacheRead: 0.
|
|
322
|
+
input: 0.1,
|
|
323
|
+
output: 0.4,
|
|
324
|
+
cacheRead: 0.025,
|
|
274
325
|
cacheWrite: 0,
|
|
275
326
|
},
|
|
276
|
-
contextWindow:
|
|
327
|
+
contextWindow: 1048576,
|
|
277
328
|
maxTokens: 8192,
|
|
278
329
|
},
|
|
279
|
-
"gemini-2.5-flash": {
|
|
280
|
-
id: "gemini-2.5-flash",
|
|
281
|
-
name: "Gemini 2.5 Flash",
|
|
330
|
+
"gemini-2.5-flash-lite": {
|
|
331
|
+
id: "gemini-2.5-flash-lite",
|
|
332
|
+
name: "Gemini 2.5 Flash Lite",
|
|
282
333
|
api: "google-generative-ai",
|
|
283
334
|
provider: "google",
|
|
284
335
|
baseUrl: "https://generativelanguage.googleapis.com/v1beta",
|
|
285
336
|
reasoning: true,
|
|
286
337
|
input: ["text", "image"],
|
|
287
338
|
cost: {
|
|
288
|
-
input: 0.
|
|
289
|
-
output:
|
|
290
|
-
cacheRead: 0.
|
|
339
|
+
input: 0.1,
|
|
340
|
+
output: 0.4,
|
|
341
|
+
cacheRead: 0.025,
|
|
291
342
|
cacheWrite: 0,
|
|
292
343
|
},
|
|
293
344
|
contextWindow: 1048576,
|
|
@@ -310,43 +361,111 @@ export const MODELS = {
|
|
|
310
361
|
contextWindow: 1048576,
|
|
311
362
|
maxTokens: 65536,
|
|
312
363
|
},
|
|
313
|
-
"gemini-2.5-
|
|
314
|
-
id: "gemini-2.5-
|
|
315
|
-
name: "Gemini 2.5
|
|
364
|
+
"gemini-2.5-flash-lite-preview-06-17": {
|
|
365
|
+
id: "gemini-2.5-flash-lite-preview-06-17",
|
|
366
|
+
name: "Gemini 2.5 Flash Lite Preview 06-17",
|
|
316
367
|
api: "google-generative-ai",
|
|
317
368
|
provider: "google",
|
|
318
369
|
baseUrl: "https://generativelanguage.googleapis.com/v1beta",
|
|
319
370
|
reasoning: true,
|
|
320
371
|
input: ["text", "image"],
|
|
321
372
|
cost: {
|
|
322
|
-
input: 1
|
|
323
|
-
output:
|
|
324
|
-
cacheRead: 0.
|
|
373
|
+
input: 0.1,
|
|
374
|
+
output: 0.4,
|
|
375
|
+
cacheRead: 0.025,
|
|
325
376
|
cacheWrite: 0,
|
|
326
377
|
},
|
|
327
378
|
contextWindow: 1048576,
|
|
328
379
|
maxTokens: 65536,
|
|
329
380
|
},
|
|
330
|
-
"gemini-2.
|
|
331
|
-
id: "gemini-2.
|
|
332
|
-
name: "Gemini 2.
|
|
381
|
+
"gemini-2.5-flash-preview-09-2025": {
|
|
382
|
+
id: "gemini-2.5-flash-preview-09-2025",
|
|
383
|
+
name: "Gemini 2.5 Flash Preview 09-25",
|
|
333
384
|
api: "google-generative-ai",
|
|
334
385
|
provider: "google",
|
|
335
386
|
baseUrl: "https://generativelanguage.googleapis.com/v1beta",
|
|
336
|
-
reasoning:
|
|
387
|
+
reasoning: true,
|
|
337
388
|
input: ["text", "image"],
|
|
338
389
|
cost: {
|
|
339
|
-
input: 0.
|
|
340
|
-
output: 0.
|
|
341
|
-
cacheRead: 0.
|
|
390
|
+
input: 0.15,
|
|
391
|
+
output: 0.6,
|
|
392
|
+
cacheRead: 0.0375,
|
|
393
|
+
cacheWrite: 0,
|
|
394
|
+
},
|
|
395
|
+
contextWindow: 1048576,
|
|
396
|
+
maxTokens: 65536,
|
|
397
|
+
},
|
|
398
|
+
"gemini-2.5-flash-preview-04-17": {
|
|
399
|
+
id: "gemini-2.5-flash-preview-04-17",
|
|
400
|
+
name: "Gemini 2.5 Flash Preview 04-17",
|
|
401
|
+
api: "google-generative-ai",
|
|
402
|
+
provider: "google",
|
|
403
|
+
baseUrl: "https://generativelanguage.googleapis.com/v1beta",
|
|
404
|
+
reasoning: true,
|
|
405
|
+
input: ["text", "image"],
|
|
406
|
+
cost: {
|
|
407
|
+
input: 0.15,
|
|
408
|
+
output: 0.6,
|
|
409
|
+
cacheRead: 0.0375,
|
|
410
|
+
cacheWrite: 0,
|
|
411
|
+
},
|
|
412
|
+
contextWindow: 1048576,
|
|
413
|
+
maxTokens: 65536,
|
|
414
|
+
},
|
|
415
|
+
"gemini-2.5-pro": {
|
|
416
|
+
id: "gemini-2.5-pro",
|
|
417
|
+
name: "Gemini 2.5 Pro",
|
|
418
|
+
api: "google-generative-ai",
|
|
419
|
+
provider: "google",
|
|
420
|
+
baseUrl: "https://generativelanguage.googleapis.com/v1beta",
|
|
421
|
+
reasoning: true,
|
|
422
|
+
input: ["text", "image"],
|
|
423
|
+
cost: {
|
|
424
|
+
input: 1.25,
|
|
425
|
+
output: 10,
|
|
426
|
+
cacheRead: 0.31,
|
|
342
427
|
cacheWrite: 0,
|
|
343
428
|
},
|
|
344
429
|
contextWindow: 1048576,
|
|
430
|
+
maxTokens: 65536,
|
|
431
|
+
},
|
|
432
|
+
"gemini-1.5-flash": {
|
|
433
|
+
id: "gemini-1.5-flash",
|
|
434
|
+
name: "Gemini 1.5 Flash",
|
|
435
|
+
api: "google-generative-ai",
|
|
436
|
+
provider: "google",
|
|
437
|
+
baseUrl: "https://generativelanguage.googleapis.com/v1beta",
|
|
438
|
+
reasoning: false,
|
|
439
|
+
input: ["text", "image"],
|
|
440
|
+
cost: {
|
|
441
|
+
input: 0.075,
|
|
442
|
+
output: 0.3,
|
|
443
|
+
cacheRead: 0.01875,
|
|
444
|
+
cacheWrite: 0,
|
|
445
|
+
},
|
|
446
|
+
contextWindow: 1000000,
|
|
345
447
|
maxTokens: 8192,
|
|
346
448
|
},
|
|
347
|
-
"gemini-
|
|
348
|
-
id: "gemini-
|
|
349
|
-
name: "Gemini
|
|
449
|
+
"gemini-1.5-flash-8b": {
|
|
450
|
+
id: "gemini-1.5-flash-8b",
|
|
451
|
+
name: "Gemini 1.5 Flash-8B",
|
|
452
|
+
api: "google-generative-ai",
|
|
453
|
+
provider: "google",
|
|
454
|
+
baseUrl: "https://generativelanguage.googleapis.com/v1beta",
|
|
455
|
+
reasoning: false,
|
|
456
|
+
input: ["text", "image"],
|
|
457
|
+
cost: {
|
|
458
|
+
input: 0.0375,
|
|
459
|
+
output: 0.15,
|
|
460
|
+
cacheRead: 0.01,
|
|
461
|
+
cacheWrite: 0,
|
|
462
|
+
},
|
|
463
|
+
contextWindow: 1000000,
|
|
464
|
+
maxTokens: 8192,
|
|
465
|
+
},
|
|
466
|
+
"gemini-2.5-flash-lite-preview-09-2025": {
|
|
467
|
+
id: "gemini-2.5-flash-lite-preview-09-2025",
|
|
468
|
+
name: "Gemini 2.5 Flash Lite Preview 09-25",
|
|
350
469
|
api: "google-generative-ai",
|
|
351
470
|
provider: "google",
|
|
352
471
|
baseUrl: "https://generativelanguage.googleapis.com/v1beta",
|
|
@@ -358,129 +477,129 @@ export const MODELS = {
|
|
|
358
477
|
cacheRead: 0.025,
|
|
359
478
|
cacheWrite: 0,
|
|
360
479
|
},
|
|
361
|
-
contextWindow:
|
|
480
|
+
contextWindow: 1048576,
|
|
362
481
|
maxTokens: 65536,
|
|
363
482
|
},
|
|
364
|
-
"gemini-
|
|
365
|
-
id: "gemini-
|
|
366
|
-
name: "Gemini
|
|
483
|
+
"gemini-1.5-pro": {
|
|
484
|
+
id: "gemini-1.5-pro",
|
|
485
|
+
name: "Gemini 1.5 Pro",
|
|
367
486
|
api: "google-generative-ai",
|
|
368
487
|
provider: "google",
|
|
369
488
|
baseUrl: "https://generativelanguage.googleapis.com/v1beta",
|
|
370
|
-
reasoning:
|
|
489
|
+
reasoning: false,
|
|
371
490
|
input: ["text", "image"],
|
|
372
491
|
cost: {
|
|
373
|
-
input:
|
|
374
|
-
output:
|
|
375
|
-
cacheRead: 0.
|
|
492
|
+
input: 1.25,
|
|
493
|
+
output: 5,
|
|
494
|
+
cacheRead: 0.3125,
|
|
376
495
|
cacheWrite: 0,
|
|
377
496
|
},
|
|
378
|
-
contextWindow:
|
|
379
|
-
maxTokens:
|
|
497
|
+
contextWindow: 1000000,
|
|
498
|
+
maxTokens: 8192,
|
|
380
499
|
},
|
|
381
500
|
},
|
|
382
501
|
openai: {
|
|
383
|
-
"gpt-
|
|
384
|
-
id: "gpt-
|
|
385
|
-
name: "GPT-
|
|
502
|
+
"gpt-4.1-nano": {
|
|
503
|
+
id: "gpt-4.1-nano",
|
|
504
|
+
name: "GPT-4.1 nano",
|
|
386
505
|
api: "openai-responses",
|
|
387
506
|
provider: "openai",
|
|
388
507
|
baseUrl: "https://api.openai.com/v1",
|
|
389
|
-
reasoning:
|
|
508
|
+
reasoning: false,
|
|
390
509
|
input: ["text", "image"],
|
|
391
510
|
cost: {
|
|
392
|
-
input: 0.
|
|
511
|
+
input: 0.1,
|
|
393
512
|
output: 0.4,
|
|
394
|
-
cacheRead: 0.
|
|
513
|
+
cacheRead: 0.03,
|
|
395
514
|
cacheWrite: 0,
|
|
396
515
|
},
|
|
397
|
-
contextWindow:
|
|
398
|
-
maxTokens:
|
|
516
|
+
contextWindow: 1047576,
|
|
517
|
+
maxTokens: 32768,
|
|
399
518
|
},
|
|
400
|
-
"
|
|
401
|
-
id: "
|
|
402
|
-
name: "
|
|
519
|
+
"gpt-4": {
|
|
520
|
+
id: "gpt-4",
|
|
521
|
+
name: "GPT-4",
|
|
403
522
|
api: "openai-responses",
|
|
404
523
|
provider: "openai",
|
|
405
524
|
baseUrl: "https://api.openai.com/v1",
|
|
406
|
-
reasoning:
|
|
407
|
-
input: ["text"
|
|
525
|
+
reasoning: false,
|
|
526
|
+
input: ["text"],
|
|
408
527
|
cost: {
|
|
409
|
-
input:
|
|
410
|
-
output:
|
|
528
|
+
input: 30,
|
|
529
|
+
output: 60,
|
|
411
530
|
cacheRead: 0,
|
|
412
531
|
cacheWrite: 0,
|
|
413
532
|
},
|
|
414
|
-
contextWindow:
|
|
415
|
-
maxTokens:
|
|
533
|
+
contextWindow: 8192,
|
|
534
|
+
maxTokens: 8192,
|
|
416
535
|
},
|
|
417
|
-
"
|
|
418
|
-
id: "
|
|
419
|
-
name: "
|
|
536
|
+
"o1-pro": {
|
|
537
|
+
id: "o1-pro",
|
|
538
|
+
name: "o1-pro",
|
|
420
539
|
api: "openai-responses",
|
|
421
540
|
provider: "openai",
|
|
422
541
|
baseUrl: "https://api.openai.com/v1",
|
|
423
542
|
reasoning: true,
|
|
424
|
-
input: ["text"],
|
|
543
|
+
input: ["text", "image"],
|
|
425
544
|
cost: {
|
|
426
|
-
input:
|
|
427
|
-
output:
|
|
428
|
-
cacheRead: 0
|
|
545
|
+
input: 150,
|
|
546
|
+
output: 600,
|
|
547
|
+
cacheRead: 0,
|
|
429
548
|
cacheWrite: 0,
|
|
430
549
|
},
|
|
431
550
|
contextWindow: 200000,
|
|
432
551
|
maxTokens: 100000,
|
|
433
552
|
},
|
|
434
|
-
"gpt-
|
|
435
|
-
id: "gpt-
|
|
436
|
-
name: "GPT-
|
|
553
|
+
"gpt-4o-2024-05-13": {
|
|
554
|
+
id: "gpt-4o-2024-05-13",
|
|
555
|
+
name: "GPT-4o (2024-05-13)",
|
|
437
556
|
api: "openai-responses",
|
|
438
557
|
provider: "openai",
|
|
439
558
|
baseUrl: "https://api.openai.com/v1",
|
|
440
559
|
reasoning: false,
|
|
441
560
|
input: ["text", "image"],
|
|
442
561
|
cost: {
|
|
443
|
-
input:
|
|
444
|
-
output:
|
|
445
|
-
cacheRead: 0
|
|
562
|
+
input: 5,
|
|
563
|
+
output: 15,
|
|
564
|
+
cacheRead: 0,
|
|
446
565
|
cacheWrite: 0,
|
|
447
566
|
},
|
|
448
|
-
contextWindow:
|
|
449
|
-
maxTokens:
|
|
567
|
+
contextWindow: 128000,
|
|
568
|
+
maxTokens: 4096,
|
|
450
569
|
},
|
|
451
|
-
"gpt-
|
|
452
|
-
id: "gpt-
|
|
453
|
-
name: "GPT-
|
|
570
|
+
"gpt-4o-2024-08-06": {
|
|
571
|
+
id: "gpt-4o-2024-08-06",
|
|
572
|
+
name: "GPT-4o (2024-08-06)",
|
|
454
573
|
api: "openai-responses",
|
|
455
574
|
provider: "openai",
|
|
456
575
|
baseUrl: "https://api.openai.com/v1",
|
|
457
576
|
reasoning: false,
|
|
458
577
|
input: ["text", "image"],
|
|
459
578
|
cost: {
|
|
460
|
-
input:
|
|
461
|
-
output:
|
|
462
|
-
cacheRead:
|
|
579
|
+
input: 2.5,
|
|
580
|
+
output: 10,
|
|
581
|
+
cacheRead: 1.25,
|
|
463
582
|
cacheWrite: 0,
|
|
464
583
|
},
|
|
465
584
|
contextWindow: 128000,
|
|
466
|
-
maxTokens:
|
|
585
|
+
maxTokens: 16384,
|
|
467
586
|
},
|
|
468
|
-
"
|
|
469
|
-
id: "
|
|
470
|
-
name: "
|
|
587
|
+
"gpt-4.1-mini": {
|
|
588
|
+
id: "gpt-4.1-mini",
|
|
589
|
+
name: "GPT-4.1 mini",
|
|
471
590
|
api: "openai-responses",
|
|
472
591
|
provider: "openai",
|
|
473
592
|
baseUrl: "https://api.openai.com/v1",
|
|
474
|
-
reasoning:
|
|
593
|
+
reasoning: false,
|
|
475
594
|
input: ["text", "image"],
|
|
476
595
|
cost: {
|
|
477
|
-
input:
|
|
478
|
-
output:
|
|
479
|
-
cacheRead:
|
|
596
|
+
input: 0.4,
|
|
597
|
+
output: 1.6,
|
|
598
|
+
cacheRead: 0.1,
|
|
480
599
|
cacheWrite: 0,
|
|
481
600
|
},
|
|
482
|
-
contextWindow:
|
|
483
|
-
maxTokens:
|
|
601
|
+
contextWindow: 1047576,
|
|
602
|
+
maxTokens: 32768,
|
|
484
603
|
},
|
|
485
604
|
"o3-deep-research": {
|
|
486
605
|
id: "o3-deep-research",
|
|
@@ -499,145 +618,196 @@ export const MODELS = {
|
|
|
499
618
|
contextWindow: 200000,
|
|
500
619
|
maxTokens: 100000,
|
|
501
620
|
},
|
|
502
|
-
"gpt-
|
|
503
|
-
id: "gpt-
|
|
504
|
-
name: "GPT-
|
|
621
|
+
"gpt-4-turbo": {
|
|
622
|
+
id: "gpt-4-turbo",
|
|
623
|
+
name: "GPT-4 Turbo",
|
|
505
624
|
api: "openai-responses",
|
|
506
625
|
provider: "openai",
|
|
507
626
|
baseUrl: "https://api.openai.com/v1",
|
|
508
|
-
reasoning:
|
|
627
|
+
reasoning: false,
|
|
509
628
|
input: ["text", "image"],
|
|
510
629
|
cost: {
|
|
511
|
-
input:
|
|
512
|
-
output:
|
|
513
|
-
cacheRead: 0
|
|
630
|
+
input: 10,
|
|
631
|
+
output: 30,
|
|
632
|
+
cacheRead: 0,
|
|
514
633
|
cacheWrite: 0,
|
|
515
634
|
},
|
|
516
|
-
contextWindow:
|
|
517
|
-
maxTokens:
|
|
635
|
+
contextWindow: 128000,
|
|
636
|
+
maxTokens: 4096,
|
|
518
637
|
},
|
|
519
|
-
"
|
|
520
|
-
id: "
|
|
521
|
-
name: "
|
|
638
|
+
"o3-mini": {
|
|
639
|
+
id: "o3-mini",
|
|
640
|
+
name: "o3-mini",
|
|
522
641
|
api: "openai-responses",
|
|
523
642
|
provider: "openai",
|
|
524
643
|
baseUrl: "https://api.openai.com/v1",
|
|
525
644
|
reasoning: true,
|
|
526
|
-
input: ["text"
|
|
645
|
+
input: ["text"],
|
|
527
646
|
cost: {
|
|
528
|
-
input:
|
|
529
|
-
output:
|
|
530
|
-
cacheRead: 0,
|
|
647
|
+
input: 1.1,
|
|
648
|
+
output: 4.4,
|
|
649
|
+
cacheRead: 0.55,
|
|
531
650
|
cacheWrite: 0,
|
|
532
651
|
},
|
|
533
652
|
contextWindow: 200000,
|
|
534
653
|
maxTokens: 100000,
|
|
535
654
|
},
|
|
536
|
-
"
|
|
537
|
-
id: "
|
|
538
|
-
name: "
|
|
655
|
+
"codex-mini-latest": {
|
|
656
|
+
id: "codex-mini-latest",
|
|
657
|
+
name: "Codex Mini",
|
|
539
658
|
api: "openai-responses",
|
|
540
659
|
provider: "openai",
|
|
541
660
|
baseUrl: "https://api.openai.com/v1",
|
|
542
661
|
reasoning: true,
|
|
543
|
-
input: ["text"
|
|
662
|
+
input: ["text"],
|
|
544
663
|
cost: {
|
|
545
|
-
input:
|
|
546
|
-
output:
|
|
547
|
-
cacheRead: 0.
|
|
664
|
+
input: 1.5,
|
|
665
|
+
output: 6,
|
|
666
|
+
cacheRead: 0.375,
|
|
548
667
|
cacheWrite: 0,
|
|
549
668
|
},
|
|
550
669
|
contextWindow: 200000,
|
|
551
670
|
maxTokens: 100000,
|
|
552
671
|
},
|
|
553
|
-
"gpt-5-
|
|
554
|
-
id: "gpt-5-
|
|
555
|
-
name: "GPT-5
|
|
672
|
+
"gpt-5-nano": {
|
|
673
|
+
id: "gpt-5-nano",
|
|
674
|
+
name: "GPT-5 Nano",
|
|
556
675
|
api: "openai-responses",
|
|
557
676
|
provider: "openai",
|
|
558
677
|
baseUrl: "https://api.openai.com/v1",
|
|
559
678
|
reasoning: true,
|
|
560
679
|
input: ["text", "image"],
|
|
561
680
|
cost: {
|
|
562
|
-
input: 0.
|
|
563
|
-
output:
|
|
564
|
-
cacheRead: 0.
|
|
681
|
+
input: 0.05,
|
|
682
|
+
output: 0.4,
|
|
683
|
+
cacheRead: 0.01,
|
|
565
684
|
cacheWrite: 0,
|
|
566
685
|
},
|
|
567
686
|
contextWindow: 400000,
|
|
568
687
|
maxTokens: 128000,
|
|
569
688
|
},
|
|
570
|
-
"
|
|
571
|
-
id: "
|
|
572
|
-
name: "
|
|
689
|
+
"gpt-5-codex": {
|
|
690
|
+
id: "gpt-5-codex",
|
|
691
|
+
name: "GPT-5-Codex",
|
|
573
692
|
api: "openai-responses",
|
|
574
693
|
provider: "openai",
|
|
575
694
|
baseUrl: "https://api.openai.com/v1",
|
|
576
695
|
reasoning: true,
|
|
577
696
|
input: ["text", "image"],
|
|
578
697
|
cost: {
|
|
579
|
-
input:
|
|
580
|
-
output:
|
|
581
|
-
cacheRead: 0
|
|
698
|
+
input: 0,
|
|
699
|
+
output: 0,
|
|
700
|
+
cacheRead: 0,
|
|
582
701
|
cacheWrite: 0,
|
|
583
702
|
},
|
|
584
|
-
contextWindow:
|
|
585
|
-
maxTokens:
|
|
703
|
+
contextWindow: 400000,
|
|
704
|
+
maxTokens: 128000,
|
|
586
705
|
},
|
|
587
|
-
"gpt-4o
|
|
588
|
-
id: "gpt-4o
|
|
589
|
-
name: "GPT-4o
|
|
706
|
+
"gpt-4o": {
|
|
707
|
+
id: "gpt-4o",
|
|
708
|
+
name: "GPT-4o",
|
|
590
709
|
api: "openai-responses",
|
|
591
710
|
provider: "openai",
|
|
592
711
|
baseUrl: "https://api.openai.com/v1",
|
|
593
712
|
reasoning: false,
|
|
594
713
|
input: ["text", "image"],
|
|
595
714
|
cost: {
|
|
596
|
-
input:
|
|
597
|
-
output:
|
|
598
|
-
cacheRead:
|
|
715
|
+
input: 2.5,
|
|
716
|
+
output: 10,
|
|
717
|
+
cacheRead: 1.25,
|
|
599
718
|
cacheWrite: 0,
|
|
600
719
|
},
|
|
601
720
|
contextWindow: 128000,
|
|
602
721
|
maxTokens: 16384,
|
|
603
722
|
},
|
|
604
|
-
"gpt-4.1
|
|
605
|
-
id: "gpt-4.1
|
|
606
|
-
name: "GPT-4.1
|
|
723
|
+
"gpt-4.1": {
|
|
724
|
+
id: "gpt-4.1",
|
|
725
|
+
name: "GPT-4.1",
|
|
607
726
|
api: "openai-responses",
|
|
608
727
|
provider: "openai",
|
|
609
728
|
baseUrl: "https://api.openai.com/v1",
|
|
610
729
|
reasoning: false,
|
|
611
730
|
input: ["text", "image"],
|
|
612
731
|
cost: {
|
|
613
|
-
input:
|
|
614
|
-
output:
|
|
615
|
-
cacheRead: 0.
|
|
732
|
+
input: 2,
|
|
733
|
+
output: 8,
|
|
734
|
+
cacheRead: 0.5,
|
|
616
735
|
cacheWrite: 0,
|
|
617
736
|
},
|
|
618
737
|
contextWindow: 1047576,
|
|
619
738
|
maxTokens: 32768,
|
|
620
739
|
},
|
|
621
|
-
"
|
|
622
|
-
id: "
|
|
623
|
-
name: "
|
|
740
|
+
"o4-mini": {
|
|
741
|
+
id: "o4-mini",
|
|
742
|
+
name: "o4-mini",
|
|
624
743
|
api: "openai-responses",
|
|
625
744
|
provider: "openai",
|
|
626
745
|
baseUrl: "https://api.openai.com/v1",
|
|
627
|
-
reasoning:
|
|
746
|
+
reasoning: true,
|
|
628
747
|
input: ["text", "image"],
|
|
629
748
|
cost: {
|
|
630
|
-
input:
|
|
631
|
-
output:
|
|
632
|
-
cacheRead: 0.
|
|
749
|
+
input: 1.1,
|
|
750
|
+
output: 4.4,
|
|
751
|
+
cacheRead: 0.28,
|
|
752
|
+
cacheWrite: 0,
|
|
753
|
+
},
|
|
754
|
+
contextWindow: 200000,
|
|
755
|
+
maxTokens: 100000,
|
|
756
|
+
},
|
|
757
|
+
"o1": {
|
|
758
|
+
id: "o1",
|
|
759
|
+
name: "o1",
|
|
760
|
+
api: "openai-responses",
|
|
761
|
+
provider: "openai",
|
|
762
|
+
baseUrl: "https://api.openai.com/v1",
|
|
763
|
+
reasoning: true,
|
|
764
|
+
input: ["text", "image"],
|
|
765
|
+
cost: {
|
|
766
|
+
input: 15,
|
|
767
|
+
output: 60,
|
|
768
|
+
cacheRead: 7.5,
|
|
769
|
+
cacheWrite: 0,
|
|
770
|
+
},
|
|
771
|
+
contextWindow: 200000,
|
|
772
|
+
maxTokens: 100000,
|
|
773
|
+
},
|
|
774
|
+
"gpt-5-mini": {
|
|
775
|
+
id: "gpt-5-mini",
|
|
776
|
+
name: "GPT-5 Mini",
|
|
777
|
+
api: "openai-responses",
|
|
778
|
+
provider: "openai",
|
|
779
|
+
baseUrl: "https://api.openai.com/v1",
|
|
780
|
+
reasoning: true,
|
|
781
|
+
input: ["text", "image"],
|
|
782
|
+
cost: {
|
|
783
|
+
input: 0.25,
|
|
784
|
+
output: 2,
|
|
785
|
+
cacheRead: 0.03,
|
|
786
|
+
cacheWrite: 0,
|
|
787
|
+
},
|
|
788
|
+
contextWindow: 400000,
|
|
789
|
+
maxTokens: 128000,
|
|
790
|
+
},
|
|
791
|
+
"o3-pro": {
|
|
792
|
+
id: "o3-pro",
|
|
793
|
+
name: "o3-pro",
|
|
794
|
+
api: "openai-responses",
|
|
795
|
+
provider: "openai",
|
|
796
|
+
baseUrl: "https://api.openai.com/v1",
|
|
797
|
+
reasoning: true,
|
|
798
|
+
input: ["text", "image"],
|
|
799
|
+
cost: {
|
|
800
|
+
input: 20,
|
|
801
|
+
output: 80,
|
|
802
|
+
cacheRead: 0,
|
|
633
803
|
cacheWrite: 0,
|
|
634
804
|
},
|
|
635
|
-
contextWindow:
|
|
636
|
-
maxTokens:
|
|
805
|
+
contextWindow: 200000,
|
|
806
|
+
maxTokens: 100000,
|
|
637
807
|
},
|
|
638
|
-
"gpt-4o": {
|
|
639
|
-
id: "gpt-4o",
|
|
640
|
-
name: "GPT-4o",
|
|
808
|
+
"gpt-4o-2024-11-20": {
|
|
809
|
+
id: "gpt-4o-2024-11-20",
|
|
810
|
+
name: "GPT-4o (2024-11-20)",
|
|
641
811
|
api: "openai-responses",
|
|
642
812
|
provider: "openai",
|
|
643
813
|
baseUrl: "https://api.openai.com/v1",
|
|
@@ -652,56 +822,73 @@ export const MODELS = {
|
|
|
652
822
|
contextWindow: 128000,
|
|
653
823
|
maxTokens: 16384,
|
|
654
824
|
},
|
|
655
|
-
"
|
|
656
|
-
id: "
|
|
657
|
-
name: "
|
|
825
|
+
"o3": {
|
|
826
|
+
id: "o3",
|
|
827
|
+
name: "o3",
|
|
658
828
|
api: "openai-responses",
|
|
659
829
|
provider: "openai",
|
|
660
830
|
baseUrl: "https://api.openai.com/v1",
|
|
661
|
-
reasoning:
|
|
662
|
-
input: ["text"],
|
|
831
|
+
reasoning: true,
|
|
832
|
+
input: ["text", "image"],
|
|
663
833
|
cost: {
|
|
664
|
-
input:
|
|
665
|
-
output:
|
|
666
|
-
cacheRead: 0,
|
|
834
|
+
input: 2,
|
|
835
|
+
output: 8,
|
|
836
|
+
cacheRead: 0.5,
|
|
667
837
|
cacheWrite: 0,
|
|
668
838
|
},
|
|
669
|
-
contextWindow:
|
|
670
|
-
maxTokens:
|
|
839
|
+
contextWindow: 200000,
|
|
840
|
+
maxTokens: 100000,
|
|
671
841
|
},
|
|
672
|
-
"o4-mini": {
|
|
673
|
-
id: "o4-mini",
|
|
674
|
-
name: "o4-mini",
|
|
842
|
+
"o4-mini-deep-research": {
|
|
843
|
+
id: "o4-mini-deep-research",
|
|
844
|
+
name: "o4-mini-deep-research",
|
|
675
845
|
api: "openai-responses",
|
|
676
846
|
provider: "openai",
|
|
677
847
|
baseUrl: "https://api.openai.com/v1",
|
|
678
848
|
reasoning: true,
|
|
679
849
|
input: ["text", "image"],
|
|
680
850
|
cost: {
|
|
681
|
-
input:
|
|
682
|
-
output:
|
|
683
|
-
cacheRead: 0.
|
|
851
|
+
input: 2,
|
|
852
|
+
output: 8,
|
|
853
|
+
cacheRead: 0.5,
|
|
684
854
|
cacheWrite: 0,
|
|
685
855
|
},
|
|
686
856
|
contextWindow: 200000,
|
|
687
857
|
maxTokens: 100000,
|
|
688
858
|
},
|
|
689
|
-
"
|
|
690
|
-
id: "
|
|
691
|
-
name: "
|
|
859
|
+
"gpt-4o-mini": {
|
|
860
|
+
id: "gpt-4o-mini",
|
|
861
|
+
name: "GPT-4o mini",
|
|
862
|
+
api: "openai-responses",
|
|
863
|
+
provider: "openai",
|
|
864
|
+
baseUrl: "https://api.openai.com/v1",
|
|
865
|
+
reasoning: false,
|
|
866
|
+
input: ["text", "image"],
|
|
867
|
+
cost: {
|
|
868
|
+
input: 0.15,
|
|
869
|
+
output: 0.6,
|
|
870
|
+
cacheRead: 0.08,
|
|
871
|
+
cacheWrite: 0,
|
|
872
|
+
},
|
|
873
|
+
contextWindow: 128000,
|
|
874
|
+
maxTokens: 16384,
|
|
875
|
+
},
|
|
876
|
+
"gpt-5": {
|
|
877
|
+
id: "gpt-5",
|
|
878
|
+
name: "GPT-5",
|
|
692
879
|
api: "openai-responses",
|
|
693
880
|
provider: "openai",
|
|
694
881
|
baseUrl: "https://api.openai.com/v1",
|
|
695
882
|
reasoning: true,
|
|
696
|
-
input: ["text"],
|
|
883
|
+
input: ["text", "image"],
|
|
697
884
|
cost: {
|
|
698
|
-
input: 1.
|
|
699
|
-
output:
|
|
700
|
-
cacheRead: 0.
|
|
885
|
+
input: 1.25,
|
|
886
|
+
output: 10,
|
|
887
|
+
cacheRead: 0.13,
|
|
701
888
|
cacheWrite: 0,
|
|
702
889
|
},
|
|
703
|
-
contextWindow:
|
|
704
|
-
maxTokens:
|
|
890
|
+
contextWindow: 400000,
|
|
891
|
+
maxTokens: 128000,
|
|
705
892
|
},
|
|
706
893
|
"gpt-5-chat-latest": {
|
|
707
894
|
id: "gpt-5-chat-latest",
|
|
@@ -739,6 +926,40 @@ export const MODELS = {
|
|
|
739
926
|
contextWindow: 131072,
|
|
740
927
|
maxTokens: 8192,
|
|
741
928
|
},
|
|
929
|
+
"mistral-saba-24b": {
|
|
930
|
+
id: "mistral-saba-24b",
|
|
931
|
+
name: "Mistral Saba 24B",
|
|
932
|
+
api: "openai-completions",
|
|
933
|
+
provider: "groq",
|
|
934
|
+
baseUrl: "https://api.groq.com/openai/v1",
|
|
935
|
+
reasoning: false,
|
|
936
|
+
input: ["text"],
|
|
937
|
+
cost: {
|
|
938
|
+
input: 0.79,
|
|
939
|
+
output: 0.79,
|
|
940
|
+
cacheRead: 0,
|
|
941
|
+
cacheWrite: 0,
|
|
942
|
+
},
|
|
943
|
+
contextWindow: 32768,
|
|
944
|
+
maxTokens: 32768,
|
|
945
|
+
},
|
|
946
|
+
"llama3-8b-8192": {
|
|
947
|
+
id: "llama3-8b-8192",
|
|
948
|
+
name: "Llama 3 8B",
|
|
949
|
+
api: "openai-completions",
|
|
950
|
+
provider: "groq",
|
|
951
|
+
baseUrl: "https://api.groq.com/openai/v1",
|
|
952
|
+
reasoning: false,
|
|
953
|
+
input: ["text"],
|
|
954
|
+
cost: {
|
|
955
|
+
input: 0.05,
|
|
956
|
+
output: 0.08,
|
|
957
|
+
cacheRead: 0,
|
|
958
|
+
cacheWrite: 0,
|
|
959
|
+
},
|
|
960
|
+
contextWindow: 8192,
|
|
961
|
+
maxTokens: 8192,
|
|
962
|
+
},
|
|
742
963
|
"qwen-qwq-32b": {
|
|
743
964
|
id: "qwen-qwq-32b",
|
|
744
965
|
name: "Qwen QwQ 32B",
|
|
@@ -790,23 +1011,6 @@ export const MODELS = {
|
|
|
790
1011
|
contextWindow: 131072,
|
|
791
1012
|
maxTokens: 8192,
|
|
792
1013
|
},
|
|
793
|
-
"llama3-8b-8192": {
|
|
794
|
-
id: "llama3-8b-8192",
|
|
795
|
-
name: "Llama 3 8B",
|
|
796
|
-
api: "openai-completions",
|
|
797
|
-
provider: "groq",
|
|
798
|
-
baseUrl: "https://api.groq.com/openai/v1",
|
|
799
|
-
reasoning: false,
|
|
800
|
-
input: ["text"],
|
|
801
|
-
cost: {
|
|
802
|
-
input: 0.05,
|
|
803
|
-
output: 0.08,
|
|
804
|
-
cacheRead: 0,
|
|
805
|
-
cacheWrite: 0,
|
|
806
|
-
},
|
|
807
|
-
contextWindow: 8192,
|
|
808
|
-
maxTokens: 8192,
|
|
809
|
-
},
|
|
810
1014
|
"gemma2-9b-it": {
|
|
811
1015
|
id: "gemma2-9b-it",
|
|
812
1016
|
name: "Gemma 2 9B",
|
|
@@ -841,22 +1045,39 @@ export const MODELS = {
|
|
|
841
1045
|
contextWindow: 131072,
|
|
842
1046
|
maxTokens: 32768,
|
|
843
1047
|
},
|
|
844
|
-
"
|
|
845
|
-
id: "
|
|
846
|
-
name: "
|
|
1048
|
+
"moonshotai/kimi-k2-instruct-0905": {
|
|
1049
|
+
id: "moonshotai/kimi-k2-instruct-0905",
|
|
1050
|
+
name: "Kimi K2 Instruct 0905",
|
|
847
1051
|
api: "openai-completions",
|
|
848
1052
|
provider: "groq",
|
|
849
1053
|
baseUrl: "https://api.groq.com/openai/v1",
|
|
850
1054
|
reasoning: false,
|
|
851
1055
|
input: ["text"],
|
|
852
1056
|
cost: {
|
|
853
|
-
input:
|
|
854
|
-
output:
|
|
1057
|
+
input: 1,
|
|
1058
|
+
output: 3,
|
|
855
1059
|
cacheRead: 0,
|
|
856
1060
|
cacheWrite: 0,
|
|
857
1061
|
},
|
|
858
|
-
contextWindow:
|
|
859
|
-
maxTokens:
|
|
1062
|
+
contextWindow: 262144,
|
|
1063
|
+
maxTokens: 16384,
|
|
1064
|
+
},
|
|
1065
|
+
"moonshotai/kimi-k2-instruct": {
|
|
1066
|
+
id: "moonshotai/kimi-k2-instruct",
|
|
1067
|
+
name: "Kimi K2 Instruct",
|
|
1068
|
+
api: "openai-completions",
|
|
1069
|
+
provider: "groq",
|
|
1070
|
+
baseUrl: "https://api.groq.com/openai/v1",
|
|
1071
|
+
reasoning: false,
|
|
1072
|
+
input: ["text"],
|
|
1073
|
+
cost: {
|
|
1074
|
+
input: 1,
|
|
1075
|
+
output: 3,
|
|
1076
|
+
cacheRead: 0,
|
|
1077
|
+
cacheWrite: 0,
|
|
1078
|
+
},
|
|
1079
|
+
contextWindow: 131072,
|
|
1080
|
+
maxTokens: 16384,
|
|
860
1081
|
},
|
|
861
1082
|
"openai/gpt-oss-20b": {
|
|
862
1083
|
id: "openai/gpt-oss-20b",
|
|
@@ -892,40 +1113,6 @@ export const MODELS = {
|
|
|
892
1113
|
contextWindow: 131072,
|
|
893
1114
|
maxTokens: 32768,
|
|
894
1115
|
},
|
|
895
|
-
"meta-llama/llama-4-maverick-17b-128e-instruct": {
|
|
896
|
-
id: "meta-llama/llama-4-maverick-17b-128e-instruct",
|
|
897
|
-
name: "Llama 4 Maverick 17B",
|
|
898
|
-
api: "openai-completions",
|
|
899
|
-
provider: "groq",
|
|
900
|
-
baseUrl: "https://api.groq.com/openai/v1",
|
|
901
|
-
reasoning: false,
|
|
902
|
-
input: ["text", "image"],
|
|
903
|
-
cost: {
|
|
904
|
-
input: 0.2,
|
|
905
|
-
output: 0.6,
|
|
906
|
-
cacheRead: 0,
|
|
907
|
-
cacheWrite: 0,
|
|
908
|
-
},
|
|
909
|
-
contextWindow: 131072,
|
|
910
|
-
maxTokens: 8192,
|
|
911
|
-
},
|
|
912
|
-
"meta-llama/llama-4-scout-17b-16e-instruct": {
|
|
913
|
-
id: "meta-llama/llama-4-scout-17b-16e-instruct",
|
|
914
|
-
name: "Llama 4 Scout 17B",
|
|
915
|
-
api: "openai-completions",
|
|
916
|
-
provider: "groq",
|
|
917
|
-
baseUrl: "https://api.groq.com/openai/v1",
|
|
918
|
-
reasoning: false,
|
|
919
|
-
input: ["text", "image"],
|
|
920
|
-
cost: {
|
|
921
|
-
input: 0.11,
|
|
922
|
-
output: 0.34,
|
|
923
|
-
cacheRead: 0,
|
|
924
|
-
cacheWrite: 0,
|
|
925
|
-
},
|
|
926
|
-
contextWindow: 131072,
|
|
927
|
-
maxTokens: 8192,
|
|
928
|
-
},
|
|
929
1116
|
"qwen/qwen3-32b": {
|
|
930
1117
|
id: "qwen/qwen3-32b",
|
|
931
1118
|
name: "Qwen3 32B",
|
|
@@ -943,39 +1130,39 @@ export const MODELS = {
|
|
|
943
1130
|
contextWindow: 131072,
|
|
944
1131
|
maxTokens: 16384,
|
|
945
1132
|
},
|
|
946
|
-
"
|
|
947
|
-
id: "
|
|
948
|
-
name: "
|
|
1133
|
+
"meta-llama/llama-4-scout-17b-16e-instruct": {
|
|
1134
|
+
id: "meta-llama/llama-4-scout-17b-16e-instruct",
|
|
1135
|
+
name: "Llama 4 Scout 17B",
|
|
949
1136
|
api: "openai-completions",
|
|
950
1137
|
provider: "groq",
|
|
951
1138
|
baseUrl: "https://api.groq.com/openai/v1",
|
|
952
1139
|
reasoning: false,
|
|
953
|
-
input: ["text"],
|
|
1140
|
+
input: ["text", "image"],
|
|
954
1141
|
cost: {
|
|
955
|
-
input:
|
|
956
|
-
output:
|
|
1142
|
+
input: 0.11,
|
|
1143
|
+
output: 0.34,
|
|
957
1144
|
cacheRead: 0,
|
|
958
1145
|
cacheWrite: 0,
|
|
959
1146
|
},
|
|
960
|
-
contextWindow:
|
|
961
|
-
maxTokens:
|
|
1147
|
+
contextWindow: 131072,
|
|
1148
|
+
maxTokens: 8192,
|
|
962
1149
|
},
|
|
963
|
-
"
|
|
964
|
-
id: "
|
|
965
|
-
name: "
|
|
1150
|
+
"meta-llama/llama-4-maverick-17b-128e-instruct": {
|
|
1151
|
+
id: "meta-llama/llama-4-maverick-17b-128e-instruct",
|
|
1152
|
+
name: "Llama 4 Maverick 17B",
|
|
966
1153
|
api: "openai-completions",
|
|
967
1154
|
provider: "groq",
|
|
968
1155
|
baseUrl: "https://api.groq.com/openai/v1",
|
|
969
1156
|
reasoning: false,
|
|
970
|
-
input: ["text"],
|
|
1157
|
+
input: ["text", "image"],
|
|
971
1158
|
cost: {
|
|
972
|
-
input:
|
|
973
|
-
output:
|
|
1159
|
+
input: 0.2,
|
|
1160
|
+
output: 0.6,
|
|
974
1161
|
cacheRead: 0,
|
|
975
1162
|
cacheWrite: 0,
|
|
976
1163
|
},
|
|
977
1164
|
contextWindow: 131072,
|
|
978
|
-
maxTokens:
|
|
1165
|
+
maxTokens: 8192,
|
|
979
1166
|
},
|
|
980
1167
|
},
|
|
981
1168
|
cerebras: {
|
|
@@ -996,23 +1183,6 @@ export const MODELS = {
|
|
|
996
1183
|
contextWindow: 131000,
|
|
997
1184
|
maxTokens: 32000,
|
|
998
1185
|
},
|
|
999
|
-
"gpt-oss-120b": {
|
|
1000
|
-
id: "gpt-oss-120b",
|
|
1001
|
-
name: "GPT OSS 120B",
|
|
1002
|
-
api: "openai-completions",
|
|
1003
|
-
provider: "cerebras",
|
|
1004
|
-
baseUrl: "https://api.cerebras.ai/v1",
|
|
1005
|
-
reasoning: true,
|
|
1006
|
-
input: ["text"],
|
|
1007
|
-
cost: {
|
|
1008
|
-
input: 0.25,
|
|
1009
|
-
output: 0.69,
|
|
1010
|
-
cacheRead: 0,
|
|
1011
|
-
cacheWrite: 0,
|
|
1012
|
-
},
|
|
1013
|
-
contextWindow: 131072,
|
|
1014
|
-
maxTokens: 32768,
|
|
1015
|
-
},
|
|
1016
1186
|
"qwen-3-coder-480b": {
|
|
1017
1187
|
id: "qwen-3-coder-480b",
|
|
1018
1188
|
name: "Qwen 3 Coder 480B",
|
|
@@ -1022,70 +1192,53 @@ export const MODELS = {
|
|
|
1022
1192
|
reasoning: false,
|
|
1023
1193
|
input: ["text"],
|
|
1024
1194
|
cost: {
|
|
1025
|
-
input: 2,
|
|
1026
|
-
output: 2,
|
|
1027
|
-
cacheRead: 0,
|
|
1028
|
-
cacheWrite: 0,
|
|
1029
|
-
},
|
|
1030
|
-
contextWindow: 131000,
|
|
1031
|
-
maxTokens: 32000,
|
|
1032
|
-
},
|
|
1033
|
-
},
|
|
1034
|
-
xai: {
|
|
1035
|
-
"grok-3-mini-fast-latest": {
|
|
1036
|
-
id: "grok-3-mini-fast-latest",
|
|
1037
|
-
name: "Grok 3 Mini Fast Latest",
|
|
1038
|
-
api: "openai-completions",
|
|
1039
|
-
provider: "xai",
|
|
1040
|
-
baseUrl: "https://api.x.ai/v1",
|
|
1041
|
-
reasoning: true,
|
|
1042
|
-
input: ["text"],
|
|
1043
|
-
cost: {
|
|
1044
|
-
input: 0.6,
|
|
1045
|
-
output: 4,
|
|
1046
|
-
cacheRead: 0.15,
|
|
1195
|
+
input: 2,
|
|
1196
|
+
output: 2,
|
|
1197
|
+
cacheRead: 0,
|
|
1047
1198
|
cacheWrite: 0,
|
|
1048
1199
|
},
|
|
1049
|
-
contextWindow:
|
|
1050
|
-
maxTokens:
|
|
1200
|
+
contextWindow: 131000,
|
|
1201
|
+
maxTokens: 32000,
|
|
1051
1202
|
},
|
|
1052
|
-
"
|
|
1053
|
-
id: "
|
|
1054
|
-
name: "
|
|
1203
|
+
"gpt-oss-120b": {
|
|
1204
|
+
id: "gpt-oss-120b",
|
|
1205
|
+
name: "GPT OSS 120B",
|
|
1055
1206
|
api: "openai-completions",
|
|
1056
|
-
provider: "
|
|
1057
|
-
baseUrl: "https://api.
|
|
1207
|
+
provider: "cerebras",
|
|
1208
|
+
baseUrl: "https://api.cerebras.ai/v1",
|
|
1058
1209
|
reasoning: true,
|
|
1059
1210
|
input: ["text"],
|
|
1060
1211
|
cost: {
|
|
1061
|
-
input: 0.
|
|
1062
|
-
output: 0.
|
|
1063
|
-
cacheRead: 0
|
|
1212
|
+
input: 0.25,
|
|
1213
|
+
output: 0.69,
|
|
1214
|
+
cacheRead: 0,
|
|
1064
1215
|
cacheWrite: 0,
|
|
1065
1216
|
},
|
|
1066
1217
|
contextWindow: 131072,
|
|
1067
|
-
maxTokens:
|
|
1218
|
+
maxTokens: 32768,
|
|
1068
1219
|
},
|
|
1069
|
-
|
|
1070
|
-
|
|
1071
|
-
|
|
1220
|
+
},
|
|
1221
|
+
xai: {
|
|
1222
|
+
"grok-4-fast-non-reasoning": {
|
|
1223
|
+
id: "grok-4-fast-non-reasoning",
|
|
1224
|
+
name: "Grok 4 Fast (Non-Reasoning)",
|
|
1072
1225
|
api: "openai-completions",
|
|
1073
1226
|
provider: "xai",
|
|
1074
1227
|
baseUrl: "https://api.x.ai/v1",
|
|
1075
1228
|
reasoning: false,
|
|
1076
|
-
input: ["text"],
|
|
1229
|
+
input: ["text", "image"],
|
|
1077
1230
|
cost: {
|
|
1078
|
-
input:
|
|
1079
|
-
output:
|
|
1080
|
-
cacheRead:
|
|
1231
|
+
input: 0.2,
|
|
1232
|
+
output: 0.5,
|
|
1233
|
+
cacheRead: 0.05,
|
|
1081
1234
|
cacheWrite: 0,
|
|
1082
1235
|
},
|
|
1083
|
-
contextWindow:
|
|
1084
|
-
maxTokens:
|
|
1236
|
+
contextWindow: 2000000,
|
|
1237
|
+
maxTokens: 30000,
|
|
1085
1238
|
},
|
|
1086
|
-
"grok-3-fast
|
|
1087
|
-
id: "grok-3-fast
|
|
1088
|
-
name: "Grok 3 Fast
|
|
1239
|
+
"grok-3-fast": {
|
|
1240
|
+
id: "grok-3-fast",
|
|
1241
|
+
name: "Grok 3 Fast",
|
|
1089
1242
|
api: "openai-completions",
|
|
1090
1243
|
provider: "xai",
|
|
1091
1244
|
baseUrl: "https://api.x.ai/v1",
|
|
@@ -1100,13 +1253,13 @@ export const MODELS = {
|
|
|
1100
1253
|
contextWindow: 131072,
|
|
1101
1254
|
maxTokens: 8192,
|
|
1102
1255
|
},
|
|
1103
|
-
"grok-
|
|
1104
|
-
id: "grok-
|
|
1105
|
-
name: "Grok
|
|
1256
|
+
"grok-4": {
|
|
1257
|
+
id: "grok-4",
|
|
1258
|
+
name: "Grok 4",
|
|
1106
1259
|
api: "openai-completions",
|
|
1107
1260
|
provider: "xai",
|
|
1108
1261
|
baseUrl: "https://api.x.ai/v1",
|
|
1109
|
-
reasoning:
|
|
1262
|
+
reasoning: true,
|
|
1110
1263
|
input: ["text"],
|
|
1111
1264
|
cost: {
|
|
1112
1265
|
input: 3,
|
|
@@ -1114,29 +1267,12 @@ export const MODELS = {
|
|
|
1114
1267
|
cacheRead: 0.75,
|
|
1115
1268
|
cacheWrite: 0,
|
|
1116
1269
|
},
|
|
1117
|
-
contextWindow:
|
|
1118
|
-
maxTokens:
|
|
1119
|
-
},
|
|
1120
|
-
"grok-3-mini": {
|
|
1121
|
-
id: "grok-3-mini",
|
|
1122
|
-
name: "Grok 3 Mini",
|
|
1123
|
-
api: "openai-completions",
|
|
1124
|
-
provider: "xai",
|
|
1125
|
-
baseUrl: "https://api.x.ai/v1",
|
|
1126
|
-
reasoning: true,
|
|
1127
|
-
input: ["text"],
|
|
1128
|
-
cost: {
|
|
1129
|
-
input: 0.3,
|
|
1130
|
-
output: 0.5,
|
|
1131
|
-
cacheRead: 0.075,
|
|
1132
|
-
cacheWrite: 0,
|
|
1133
|
-
},
|
|
1134
|
-
contextWindow: 131072,
|
|
1135
|
-
maxTokens: 8192,
|
|
1270
|
+
contextWindow: 256000,
|
|
1271
|
+
maxTokens: 64000,
|
|
1136
1272
|
},
|
|
1137
|
-
"grok-2-vision
|
|
1138
|
-
id: "grok-2-vision
|
|
1139
|
-
name: "Grok 2 Vision
|
|
1273
|
+
"grok-2-vision": {
|
|
1274
|
+
id: "grok-2-vision",
|
|
1275
|
+
name: "Grok 2 Vision",
|
|
1140
1276
|
api: "openai-completions",
|
|
1141
1277
|
provider: "xai",
|
|
1142
1278
|
baseUrl: "https://api.x.ai/v1",
|
|
@@ -1151,6 +1287,23 @@ export const MODELS = {
|
|
|
1151
1287
|
contextWindow: 8192,
|
|
1152
1288
|
maxTokens: 4096,
|
|
1153
1289
|
},
|
|
1290
|
+
"grok-code-fast-1": {
|
|
1291
|
+
id: "grok-code-fast-1",
|
|
1292
|
+
name: "Grok Code Fast 1",
|
|
1293
|
+
api: "openai-completions",
|
|
1294
|
+
provider: "xai",
|
|
1295
|
+
baseUrl: "https://api.x.ai/v1",
|
|
1296
|
+
reasoning: true,
|
|
1297
|
+
input: ["text"],
|
|
1298
|
+
cost: {
|
|
1299
|
+
input: 0.2,
|
|
1300
|
+
output: 1.5,
|
|
1301
|
+
cacheRead: 0.02,
|
|
1302
|
+
cacheWrite: 0,
|
|
1303
|
+
},
|
|
1304
|
+
contextWindow: 256000,
|
|
1305
|
+
maxTokens: 10000,
|
|
1306
|
+
},
|
|
1154
1307
|
"grok-2": {
|
|
1155
1308
|
id: "grok-2",
|
|
1156
1309
|
name: "Grok 2",
|
|
@@ -1168,9 +1321,26 @@ export const MODELS = {
|
|
|
1168
1321
|
contextWindow: 131072,
|
|
1169
1322
|
maxTokens: 8192,
|
|
1170
1323
|
},
|
|
1171
|
-
"grok-
|
|
1172
|
-
id: "grok-
|
|
1173
|
-
name: "Grok
|
|
1324
|
+
"grok-3-mini-fast-latest": {
|
|
1325
|
+
id: "grok-3-mini-fast-latest",
|
|
1326
|
+
name: "Grok 3 Mini Fast Latest",
|
|
1327
|
+
api: "openai-completions",
|
|
1328
|
+
provider: "xai",
|
|
1329
|
+
baseUrl: "https://api.x.ai/v1",
|
|
1330
|
+
reasoning: true,
|
|
1331
|
+
input: ["text"],
|
|
1332
|
+
cost: {
|
|
1333
|
+
input: 0.6,
|
|
1334
|
+
output: 4,
|
|
1335
|
+
cacheRead: 0.15,
|
|
1336
|
+
cacheWrite: 0,
|
|
1337
|
+
},
|
|
1338
|
+
contextWindow: 131072,
|
|
1339
|
+
maxTokens: 8192,
|
|
1340
|
+
},
|
|
1341
|
+
"grok-2-vision-1212": {
|
|
1342
|
+
id: "grok-2-vision-1212",
|
|
1343
|
+
name: "Grok 2 Vision (1212)",
|
|
1174
1344
|
api: "openai-completions",
|
|
1175
1345
|
provider: "xai",
|
|
1176
1346
|
baseUrl: "https://api.x.ai/v1",
|
|
@@ -1185,9 +1355,9 @@ export const MODELS = {
|
|
|
1185
1355
|
contextWindow: 8192,
|
|
1186
1356
|
maxTokens: 4096,
|
|
1187
1357
|
},
|
|
1188
|
-
"grok-3
|
|
1189
|
-
id: "grok-3
|
|
1190
|
-
name: "Grok 3
|
|
1358
|
+
"grok-3": {
|
|
1359
|
+
id: "grok-3",
|
|
1360
|
+
name: "Grok 3",
|
|
1191
1361
|
api: "openai-completions",
|
|
1192
1362
|
provider: "xai",
|
|
1193
1363
|
baseUrl: "https://api.x.ai/v1",
|
|
@@ -1202,22 +1372,22 @@ export const MODELS = {
|
|
|
1202
1372
|
contextWindow: 131072,
|
|
1203
1373
|
maxTokens: 8192,
|
|
1204
1374
|
},
|
|
1205
|
-
"grok-
|
|
1206
|
-
id: "grok-
|
|
1207
|
-
name: "Grok
|
|
1375
|
+
"grok-4-fast": {
|
|
1376
|
+
id: "grok-4-fast",
|
|
1377
|
+
name: "Grok 4 Fast",
|
|
1208
1378
|
api: "openai-completions",
|
|
1209
1379
|
provider: "xai",
|
|
1210
1380
|
baseUrl: "https://api.x.ai/v1",
|
|
1211
|
-
reasoning:
|
|
1381
|
+
reasoning: true,
|
|
1212
1382
|
input: ["text", "image"],
|
|
1213
1383
|
cost: {
|
|
1214
|
-
input: 2,
|
|
1215
|
-
output:
|
|
1216
|
-
cacheRead:
|
|
1384
|
+
input: 0.2,
|
|
1385
|
+
output: 0.5,
|
|
1386
|
+
cacheRead: 0.05,
|
|
1217
1387
|
cacheWrite: 0,
|
|
1218
1388
|
},
|
|
1219
|
-
contextWindow:
|
|
1220
|
-
maxTokens:
|
|
1389
|
+
contextWindow: 2000000,
|
|
1390
|
+
maxTokens: 30000,
|
|
1221
1391
|
},
|
|
1222
1392
|
"grok-2-latest": {
|
|
1223
1393
|
id: "grok-2-latest",
|
|
@@ -1236,47 +1406,47 @@ export const MODELS = {
|
|
|
1236
1406
|
contextWindow: 131072,
|
|
1237
1407
|
maxTokens: 8192,
|
|
1238
1408
|
},
|
|
1239
|
-
"grok-
|
|
1240
|
-
id: "grok-
|
|
1241
|
-
name: "Grok
|
|
1409
|
+
"grok-2-1212": {
|
|
1410
|
+
id: "grok-2-1212",
|
|
1411
|
+
name: "Grok 2 (1212)",
|
|
1242
1412
|
api: "openai-completions",
|
|
1243
1413
|
provider: "xai",
|
|
1244
1414
|
baseUrl: "https://api.x.ai/v1",
|
|
1245
1415
|
reasoning: false,
|
|
1246
1416
|
input: ["text"],
|
|
1247
1417
|
cost: {
|
|
1248
|
-
input:
|
|
1249
|
-
output:
|
|
1250
|
-
cacheRead:
|
|
1418
|
+
input: 2,
|
|
1419
|
+
output: 10,
|
|
1420
|
+
cacheRead: 2,
|
|
1251
1421
|
cacheWrite: 0,
|
|
1252
1422
|
},
|
|
1253
1423
|
contextWindow: 131072,
|
|
1254
1424
|
maxTokens: 8192,
|
|
1255
1425
|
},
|
|
1256
|
-
"grok-
|
|
1257
|
-
id: "grok-
|
|
1258
|
-
name: "Grok
|
|
1426
|
+
"grok-3-fast-latest": {
|
|
1427
|
+
id: "grok-3-fast-latest",
|
|
1428
|
+
name: "Grok 3 Fast Latest",
|
|
1259
1429
|
api: "openai-completions",
|
|
1260
1430
|
provider: "xai",
|
|
1261
1431
|
baseUrl: "https://api.x.ai/v1",
|
|
1262
1432
|
reasoning: false,
|
|
1263
1433
|
input: ["text"],
|
|
1264
1434
|
cost: {
|
|
1265
|
-
input:
|
|
1266
|
-
output:
|
|
1267
|
-
cacheRead:
|
|
1435
|
+
input: 5,
|
|
1436
|
+
output: 25,
|
|
1437
|
+
cacheRead: 1.25,
|
|
1268
1438
|
cacheWrite: 0,
|
|
1269
1439
|
},
|
|
1270
1440
|
contextWindow: 131072,
|
|
1271
1441
|
maxTokens: 8192,
|
|
1272
1442
|
},
|
|
1273
|
-
"grok-
|
|
1274
|
-
id: "grok-
|
|
1275
|
-
name: "Grok
|
|
1443
|
+
"grok-3-latest": {
|
|
1444
|
+
id: "grok-3-latest",
|
|
1445
|
+
name: "Grok 3 Latest",
|
|
1276
1446
|
api: "openai-completions",
|
|
1277
1447
|
provider: "xai",
|
|
1278
1448
|
baseUrl: "https://api.x.ai/v1",
|
|
1279
|
-
reasoning:
|
|
1449
|
+
reasoning: false,
|
|
1280
1450
|
input: ["text"],
|
|
1281
1451
|
cost: {
|
|
1282
1452
|
input: 3,
|
|
@@ -1284,8 +1454,25 @@ export const MODELS = {
|
|
|
1284
1454
|
cacheRead: 0.75,
|
|
1285
1455
|
cacheWrite: 0,
|
|
1286
1456
|
},
|
|
1287
|
-
contextWindow:
|
|
1288
|
-
maxTokens:
|
|
1457
|
+
contextWindow: 131072,
|
|
1458
|
+
maxTokens: 8192,
|
|
1459
|
+
},
|
|
1460
|
+
"grok-2-vision-latest": {
|
|
1461
|
+
id: "grok-2-vision-latest",
|
|
1462
|
+
name: "Grok 2 Vision Latest",
|
|
1463
|
+
api: "openai-completions",
|
|
1464
|
+
provider: "xai",
|
|
1465
|
+
baseUrl: "https://api.x.ai/v1",
|
|
1466
|
+
reasoning: false,
|
|
1467
|
+
input: ["text", "image"],
|
|
1468
|
+
cost: {
|
|
1469
|
+
input: 2,
|
|
1470
|
+
output: 10,
|
|
1471
|
+
cacheRead: 2,
|
|
1472
|
+
cacheWrite: 0,
|
|
1473
|
+
},
|
|
1474
|
+
contextWindow: 8192,
|
|
1475
|
+
maxTokens: 4096,
|
|
1289
1476
|
},
|
|
1290
1477
|
"grok-vision-beta": {
|
|
1291
1478
|
id: "grok-vision-beta",
|
|
@@ -1304,6 +1491,57 @@ export const MODELS = {
|
|
|
1304
1491
|
contextWindow: 8192,
|
|
1305
1492
|
maxTokens: 4096,
|
|
1306
1493
|
},
|
|
1494
|
+
"grok-3-mini": {
|
|
1495
|
+
id: "grok-3-mini",
|
|
1496
|
+
name: "Grok 3 Mini",
|
|
1497
|
+
api: "openai-completions",
|
|
1498
|
+
provider: "xai",
|
|
1499
|
+
baseUrl: "https://api.x.ai/v1",
|
|
1500
|
+
reasoning: true,
|
|
1501
|
+
input: ["text"],
|
|
1502
|
+
cost: {
|
|
1503
|
+
input: 0.3,
|
|
1504
|
+
output: 0.5,
|
|
1505
|
+
cacheRead: 0.075,
|
|
1506
|
+
cacheWrite: 0,
|
|
1507
|
+
},
|
|
1508
|
+
contextWindow: 131072,
|
|
1509
|
+
maxTokens: 8192,
|
|
1510
|
+
},
|
|
1511
|
+
"grok-beta": {
|
|
1512
|
+
id: "grok-beta",
|
|
1513
|
+
name: "Grok Beta",
|
|
1514
|
+
api: "openai-completions",
|
|
1515
|
+
provider: "xai",
|
|
1516
|
+
baseUrl: "https://api.x.ai/v1",
|
|
1517
|
+
reasoning: false,
|
|
1518
|
+
input: ["text"],
|
|
1519
|
+
cost: {
|
|
1520
|
+
input: 5,
|
|
1521
|
+
output: 15,
|
|
1522
|
+
cacheRead: 5,
|
|
1523
|
+
cacheWrite: 0,
|
|
1524
|
+
},
|
|
1525
|
+
contextWindow: 131072,
|
|
1526
|
+
maxTokens: 4096,
|
|
1527
|
+
},
|
|
1528
|
+
"grok-3-mini-latest": {
|
|
1529
|
+
id: "grok-3-mini-latest",
|
|
1530
|
+
name: "Grok 3 Mini Latest",
|
|
1531
|
+
api: "openai-completions",
|
|
1532
|
+
provider: "xai",
|
|
1533
|
+
baseUrl: "https://api.x.ai/v1",
|
|
1534
|
+
reasoning: true,
|
|
1535
|
+
input: ["text"],
|
|
1536
|
+
cost: {
|
|
1537
|
+
input: 0.3,
|
|
1538
|
+
output: 0.5,
|
|
1539
|
+
cacheRead: 0.075,
|
|
1540
|
+
cacheWrite: 0,
|
|
1541
|
+
},
|
|
1542
|
+
contextWindow: 131072,
|
|
1543
|
+
maxTokens: 8192,
|
|
1544
|
+
},
|
|
1307
1545
|
"grok-3-mini-fast": {
|
|
1308
1546
|
id: "grok-3-mini-fast",
|
|
1309
1547
|
name: "Grok 3 Mini Fast",
|
|
@@ -1313,33 +1551,50 @@ export const MODELS = {
|
|
|
1313
1551
|
reasoning: true,
|
|
1314
1552
|
input: ["text"],
|
|
1315
1553
|
cost: {
|
|
1316
|
-
input: 0.6,
|
|
1317
|
-
output: 4,
|
|
1318
|
-
cacheRead: 0.15,
|
|
1554
|
+
input: 0.6,
|
|
1555
|
+
output: 4,
|
|
1556
|
+
cacheRead: 0.15,
|
|
1557
|
+
cacheWrite: 0,
|
|
1558
|
+
},
|
|
1559
|
+
contextWindow: 131072,
|
|
1560
|
+
maxTokens: 8192,
|
|
1561
|
+
},
|
|
1562
|
+
},
|
|
1563
|
+
zai: {
|
|
1564
|
+
"glm-4.5-flash": {
|
|
1565
|
+
id: "glm-4.5-flash",
|
|
1566
|
+
name: "GLM-4.5-Flash",
|
|
1567
|
+
api: "anthropic-messages",
|
|
1568
|
+
provider: "zai",
|
|
1569
|
+
baseUrl: "https://api.z.ai/api/anthropic",
|
|
1570
|
+
reasoning: true,
|
|
1571
|
+
input: ["text"],
|
|
1572
|
+
cost: {
|
|
1573
|
+
input: 0,
|
|
1574
|
+
output: 0,
|
|
1575
|
+
cacheRead: 0,
|
|
1319
1576
|
cacheWrite: 0,
|
|
1320
1577
|
},
|
|
1321
1578
|
contextWindow: 131072,
|
|
1322
|
-
maxTokens:
|
|
1579
|
+
maxTokens: 98304,
|
|
1323
1580
|
},
|
|
1324
|
-
"
|
|
1325
|
-
id: "
|
|
1326
|
-
name: "
|
|
1327
|
-
api: "
|
|
1328
|
-
provider: "
|
|
1329
|
-
baseUrl: "https://api.
|
|
1330
|
-
reasoning:
|
|
1581
|
+
"glm-4.5": {
|
|
1582
|
+
id: "glm-4.5",
|
|
1583
|
+
name: "GLM-4.5",
|
|
1584
|
+
api: "anthropic-messages",
|
|
1585
|
+
provider: "zai",
|
|
1586
|
+
baseUrl: "https://api.z.ai/api/anthropic",
|
|
1587
|
+
reasoning: true,
|
|
1331
1588
|
input: ["text"],
|
|
1332
1589
|
cost: {
|
|
1333
|
-
input: 0.
|
|
1334
|
-
output:
|
|
1335
|
-
cacheRead: 0.
|
|
1590
|
+
input: 0.6,
|
|
1591
|
+
output: 2.2,
|
|
1592
|
+
cacheRead: 0.11,
|
|
1336
1593
|
cacheWrite: 0,
|
|
1337
1594
|
},
|
|
1338
|
-
contextWindow:
|
|
1339
|
-
maxTokens:
|
|
1595
|
+
contextWindow: 131072,
|
|
1596
|
+
maxTokens: 98304,
|
|
1340
1597
|
},
|
|
1341
|
-
},
|
|
1342
|
-
zai: {
|
|
1343
1598
|
"glm-4.5-air": {
|
|
1344
1599
|
id: "glm-4.5-air",
|
|
1345
1600
|
name: "GLM-4.5-Air",
|
|
@@ -1374,75 +1629,109 @@ export const MODELS = {
|
|
|
1374
1629
|
contextWindow: 64000,
|
|
1375
1630
|
maxTokens: 16384,
|
|
1376
1631
|
},
|
|
1377
|
-
"glm-4.
|
|
1378
|
-
id: "glm-4.
|
|
1379
|
-
name: "GLM-4.
|
|
1632
|
+
"glm-4.6": {
|
|
1633
|
+
id: "glm-4.6",
|
|
1634
|
+
name: "GLM-4.6",
|
|
1380
1635
|
api: "anthropic-messages",
|
|
1381
1636
|
provider: "zai",
|
|
1382
1637
|
baseUrl: "https://api.z.ai/api/anthropic",
|
|
1383
1638
|
reasoning: true,
|
|
1384
1639
|
input: ["text"],
|
|
1385
1640
|
cost: {
|
|
1386
|
-
input: 0,
|
|
1387
|
-
output:
|
|
1388
|
-
cacheRead: 0,
|
|
1641
|
+
input: 0.6,
|
|
1642
|
+
output: 2.2,
|
|
1643
|
+
cacheRead: 0.11,
|
|
1389
1644
|
cacheWrite: 0,
|
|
1390
1645
|
},
|
|
1391
|
-
contextWindow:
|
|
1392
|
-
maxTokens:
|
|
1646
|
+
contextWindow: 204800,
|
|
1647
|
+
maxTokens: 131072,
|
|
1393
1648
|
},
|
|
1394
|
-
|
|
1395
|
-
|
|
1396
|
-
|
|
1397
|
-
|
|
1398
|
-
|
|
1399
|
-
|
|
1649
|
+
},
|
|
1650
|
+
openrouter: {
|
|
1651
|
+
"z-ai/glm-4.6": {
|
|
1652
|
+
id: "z-ai/glm-4.6",
|
|
1653
|
+
name: "Z.AI: GLM 4.6",
|
|
1654
|
+
api: "openai-completions",
|
|
1655
|
+
provider: "openrouter",
|
|
1656
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
1400
1657
|
reasoning: true,
|
|
1401
1658
|
input: ["text"],
|
|
1402
1659
|
cost: {
|
|
1403
|
-
input: 0.
|
|
1404
|
-
output:
|
|
1405
|
-
cacheRead: 0
|
|
1660
|
+
input: 0.5,
|
|
1661
|
+
output: 1.75,
|
|
1662
|
+
cacheRead: 0,
|
|
1406
1663
|
cacheWrite: 0,
|
|
1407
1664
|
},
|
|
1408
|
-
contextWindow:
|
|
1409
|
-
maxTokens:
|
|
1665
|
+
contextWindow: 202752,
|
|
1666
|
+
maxTokens: 202752,
|
|
1410
1667
|
},
|
|
1411
|
-
|
|
1412
|
-
|
|
1413
|
-
|
|
1414
|
-
id: "alibaba/tongyi-deepresearch-30b-a3b",
|
|
1415
|
-
name: "Tongyi DeepResearch 30B A3B",
|
|
1668
|
+
"deepseek/deepseek-v3.2-exp": {
|
|
1669
|
+
id: "deepseek/deepseek-v3.2-exp",
|
|
1670
|
+
name: "DeepSeek: DeepSeek V3.2 Exp",
|
|
1416
1671
|
api: "openai-completions",
|
|
1417
1672
|
provider: "openrouter",
|
|
1418
1673
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
1419
1674
|
reasoning: true,
|
|
1420
1675
|
input: ["text"],
|
|
1421
1676
|
cost: {
|
|
1422
|
-
input: 0.
|
|
1423
|
-
output: 0.
|
|
1677
|
+
input: 0.27,
|
|
1678
|
+
output: 0.39999999999999997,
|
|
1679
|
+
cacheRead: 0,
|
|
1680
|
+
cacheWrite: 0,
|
|
1681
|
+
},
|
|
1682
|
+
contextWindow: 163840,
|
|
1683
|
+
maxTokens: 4096,
|
|
1684
|
+
},
|
|
1685
|
+
"qwen/qwen3-vl-235b-a22b-thinking": {
|
|
1686
|
+
id: "qwen/qwen3-vl-235b-a22b-thinking",
|
|
1687
|
+
name: "Qwen: Qwen3 VL 235B A22B Thinking",
|
|
1688
|
+
api: "openai-completions",
|
|
1689
|
+
provider: "openrouter",
|
|
1690
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
1691
|
+
reasoning: true,
|
|
1692
|
+
input: ["text", "image"],
|
|
1693
|
+
cost: {
|
|
1694
|
+
input: 0.5,
|
|
1695
|
+
output: 3.5,
|
|
1696
|
+
cacheRead: 0,
|
|
1697
|
+
cacheWrite: 0,
|
|
1698
|
+
},
|
|
1699
|
+
contextWindow: 65536,
|
|
1700
|
+
maxTokens: 65536,
|
|
1701
|
+
},
|
|
1702
|
+
"qwen/qwen3-vl-235b-a22b-instruct": {
|
|
1703
|
+
id: "qwen/qwen3-vl-235b-a22b-instruct",
|
|
1704
|
+
name: "Qwen: Qwen3 VL 235B A22B Instruct",
|
|
1705
|
+
api: "openai-completions",
|
|
1706
|
+
provider: "openrouter",
|
|
1707
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
1708
|
+
reasoning: true,
|
|
1709
|
+
input: ["text", "image"],
|
|
1710
|
+
cost: {
|
|
1711
|
+
input: 0.3,
|
|
1712
|
+
output: 1.5,
|
|
1424
1713
|
cacheRead: 0,
|
|
1425
1714
|
cacheWrite: 0,
|
|
1426
1715
|
},
|
|
1427
1716
|
contextWindow: 131072,
|
|
1428
|
-
maxTokens:
|
|
1717
|
+
maxTokens: 32768,
|
|
1429
1718
|
},
|
|
1430
|
-
"qwen/qwen3-
|
|
1431
|
-
id: "qwen/qwen3-
|
|
1432
|
-
name: "Qwen: Qwen3
|
|
1719
|
+
"qwen/qwen3-max": {
|
|
1720
|
+
id: "qwen/qwen3-max",
|
|
1721
|
+
name: "Qwen: Qwen3 Max",
|
|
1433
1722
|
api: "openai-completions",
|
|
1434
1723
|
provider: "openrouter",
|
|
1435
1724
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
1436
1725
|
reasoning: false,
|
|
1437
1726
|
input: ["text"],
|
|
1438
1727
|
cost: {
|
|
1439
|
-
input:
|
|
1440
|
-
output:
|
|
1441
|
-
cacheRead: 0.
|
|
1728
|
+
input: 1.2,
|
|
1729
|
+
output: 6,
|
|
1730
|
+
cacheRead: 0.24,
|
|
1442
1731
|
cacheWrite: 0,
|
|
1443
1732
|
},
|
|
1444
|
-
contextWindow:
|
|
1445
|
-
maxTokens:
|
|
1733
|
+
contextWindow: 256000,
|
|
1734
|
+
maxTokens: 32768,
|
|
1446
1735
|
},
|
|
1447
1736
|
"qwen/qwen3-coder-plus": {
|
|
1448
1737
|
id: "qwen/qwen3-coder-plus",
|
|
@@ -1461,6 +1750,74 @@ export const MODELS = {
|
|
|
1461
1750
|
contextWindow: 128000,
|
|
1462
1751
|
maxTokens: 65536,
|
|
1463
1752
|
},
|
|
1753
|
+
"deepseek/deepseek-v3.1-terminus": {
|
|
1754
|
+
id: "deepseek/deepseek-v3.1-terminus",
|
|
1755
|
+
name: "DeepSeek: DeepSeek V3.1 Terminus",
|
|
1756
|
+
api: "openai-completions",
|
|
1757
|
+
provider: "openrouter",
|
|
1758
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
1759
|
+
reasoning: true,
|
|
1760
|
+
input: ["text"],
|
|
1761
|
+
cost: {
|
|
1762
|
+
input: 0.22999999999999998,
|
|
1763
|
+
output: 0.8999999999999999,
|
|
1764
|
+
cacheRead: 0,
|
|
1765
|
+
cacheWrite: 0,
|
|
1766
|
+
},
|
|
1767
|
+
contextWindow: 163840,
|
|
1768
|
+
maxTokens: 163840,
|
|
1769
|
+
},
|
|
1770
|
+
"alibaba/tongyi-deepresearch-30b-a3b:free": {
|
|
1771
|
+
id: "alibaba/tongyi-deepresearch-30b-a3b:free",
|
|
1772
|
+
name: "Tongyi DeepResearch 30B A3B (free)",
|
|
1773
|
+
api: "openai-completions",
|
|
1774
|
+
provider: "openrouter",
|
|
1775
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
1776
|
+
reasoning: false,
|
|
1777
|
+
input: ["text"],
|
|
1778
|
+
cost: {
|
|
1779
|
+
input: 0,
|
|
1780
|
+
output: 0,
|
|
1781
|
+
cacheRead: 0,
|
|
1782
|
+
cacheWrite: 0,
|
|
1783
|
+
},
|
|
1784
|
+
contextWindow: 131072,
|
|
1785
|
+
maxTokens: 131072,
|
|
1786
|
+
},
|
|
1787
|
+
"alibaba/tongyi-deepresearch-30b-a3b": {
|
|
1788
|
+
id: "alibaba/tongyi-deepresearch-30b-a3b",
|
|
1789
|
+
name: "Tongyi DeepResearch 30B A3B",
|
|
1790
|
+
api: "openai-completions",
|
|
1791
|
+
provider: "openrouter",
|
|
1792
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
1793
|
+
reasoning: true,
|
|
1794
|
+
input: ["text"],
|
|
1795
|
+
cost: {
|
|
1796
|
+
input: 0.09,
|
|
1797
|
+
output: 0.39999999999999997,
|
|
1798
|
+
cacheRead: 0,
|
|
1799
|
+
cacheWrite: 0,
|
|
1800
|
+
},
|
|
1801
|
+
contextWindow: 131072,
|
|
1802
|
+
maxTokens: 131072,
|
|
1803
|
+
},
|
|
1804
|
+
"qwen/qwen3-coder-flash": {
|
|
1805
|
+
id: "qwen/qwen3-coder-flash",
|
|
1806
|
+
name: "Qwen: Qwen3 Coder Flash",
|
|
1807
|
+
api: "openai-completions",
|
|
1808
|
+
provider: "openrouter",
|
|
1809
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
1810
|
+
reasoning: false,
|
|
1811
|
+
input: ["text"],
|
|
1812
|
+
cost: {
|
|
1813
|
+
input: 0.3,
|
|
1814
|
+
output: 1.5,
|
|
1815
|
+
cacheRead: 0.08,
|
|
1816
|
+
cacheWrite: 0,
|
|
1817
|
+
},
|
|
1818
|
+
contextWindow: 128000,
|
|
1819
|
+
maxTokens: 65536,
|
|
1820
|
+
},
|
|
1464
1821
|
"qwen/qwen3-next-80b-a3b-thinking": {
|
|
1465
1822
|
id: "qwen/qwen3-next-80b-a3b-thinking",
|
|
1466
1823
|
name: "Qwen: Qwen3 Next 80B A3B Thinking",
|
|
@@ -1476,7 +1833,7 @@ export const MODELS = {
|
|
|
1476
1833
|
cacheWrite: 0,
|
|
1477
1834
|
},
|
|
1478
1835
|
contextWindow: 262144,
|
|
1479
|
-
maxTokens:
|
|
1836
|
+
maxTokens: 262144,
|
|
1480
1837
|
},
|
|
1481
1838
|
"qwen/qwen3-next-80b-a3b-instruct": {
|
|
1482
1839
|
id: "qwen/qwen3-next-80b-a3b-instruct",
|
|
@@ -1493,24 +1850,24 @@ export const MODELS = {
|
|
|
1493
1850
|
cacheWrite: 0,
|
|
1494
1851
|
},
|
|
1495
1852
|
contextWindow: 262144,
|
|
1496
|
-
maxTokens:
|
|
1853
|
+
maxTokens: 262144,
|
|
1497
1854
|
},
|
|
1498
|
-
"meituan/longcat-flash-chat": {
|
|
1499
|
-
id: "meituan/longcat-flash-chat",
|
|
1500
|
-
name: "Meituan: LongCat Flash Chat",
|
|
1855
|
+
"meituan/longcat-flash-chat:free": {
|
|
1856
|
+
id: "meituan/longcat-flash-chat:free",
|
|
1857
|
+
name: "Meituan: LongCat Flash Chat (free)",
|
|
1501
1858
|
api: "openai-completions",
|
|
1502
1859
|
provider: "openrouter",
|
|
1503
1860
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
1504
1861
|
reasoning: false,
|
|
1505
1862
|
input: ["text"],
|
|
1506
1863
|
cost: {
|
|
1507
|
-
input: 0
|
|
1508
|
-
output: 0
|
|
1864
|
+
input: 0,
|
|
1865
|
+
output: 0,
|
|
1509
1866
|
cacheRead: 0,
|
|
1510
1867
|
cacheWrite: 0,
|
|
1511
1868
|
},
|
|
1512
1869
|
contextWindow: 131072,
|
|
1513
|
-
maxTokens:
|
|
1870
|
+
maxTokens: 131072,
|
|
1514
1871
|
},
|
|
1515
1872
|
"qwen/qwen-plus-2025-07-28": {
|
|
1516
1873
|
id: "qwen/qwen-plus-2025-07-28",
|
|
@@ -1580,57 +1937,6 @@ export const MODELS = {
|
|
|
1580
1937
|
contextWindow: 131072,
|
|
1581
1938
|
maxTokens: 4096,
|
|
1582
1939
|
},
|
|
1583
|
-
"openrouter/sonoma-dusk-alpha": {
|
|
1584
|
-
id: "openrouter/sonoma-dusk-alpha",
|
|
1585
|
-
name: "Sonoma Dusk Alpha",
|
|
1586
|
-
api: "openai-completions",
|
|
1587
|
-
provider: "openrouter",
|
|
1588
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
1589
|
-
reasoning: false,
|
|
1590
|
-
input: ["text", "image"],
|
|
1591
|
-
cost: {
|
|
1592
|
-
input: 0,
|
|
1593
|
-
output: 0,
|
|
1594
|
-
cacheRead: 0,
|
|
1595
|
-
cacheWrite: 0,
|
|
1596
|
-
},
|
|
1597
|
-
contextWindow: 2000000,
|
|
1598
|
-
maxTokens: 4096,
|
|
1599
|
-
},
|
|
1600
|
-
"openrouter/sonoma-sky-alpha": {
|
|
1601
|
-
id: "openrouter/sonoma-sky-alpha",
|
|
1602
|
-
name: "Sonoma Sky Alpha",
|
|
1603
|
-
api: "openai-completions",
|
|
1604
|
-
provider: "openrouter",
|
|
1605
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
1606
|
-
reasoning: true,
|
|
1607
|
-
input: ["text", "image"],
|
|
1608
|
-
cost: {
|
|
1609
|
-
input: 0,
|
|
1610
|
-
output: 0,
|
|
1611
|
-
cacheRead: 0,
|
|
1612
|
-
cacheWrite: 0,
|
|
1613
|
-
},
|
|
1614
|
-
contextWindow: 2000000,
|
|
1615
|
-
maxTokens: 4096,
|
|
1616
|
-
},
|
|
1617
|
-
"qwen/qwen3-max": {
|
|
1618
|
-
id: "qwen/qwen3-max",
|
|
1619
|
-
name: "Qwen: Qwen3 Max",
|
|
1620
|
-
api: "openai-completions",
|
|
1621
|
-
provider: "openrouter",
|
|
1622
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
1623
|
-
reasoning: false,
|
|
1624
|
-
input: ["text"],
|
|
1625
|
-
cost: {
|
|
1626
|
-
input: 1.2,
|
|
1627
|
-
output: 6,
|
|
1628
|
-
cacheRead: 0.24,
|
|
1629
|
-
cacheWrite: 0,
|
|
1630
|
-
},
|
|
1631
|
-
contextWindow: 256000,
|
|
1632
|
-
maxTokens: 32768,
|
|
1633
|
-
},
|
|
1634
1940
|
"moonshotai/kimi-k2-0905": {
|
|
1635
1941
|
id: "moonshotai/kimi-k2-0905",
|
|
1636
1942
|
name: "MoonshotAI: Kimi K2 0905",
|
|
@@ -1640,13 +1946,13 @@ export const MODELS = {
|
|
|
1640
1946
|
reasoning: false,
|
|
1641
1947
|
input: ["text"],
|
|
1642
1948
|
cost: {
|
|
1643
|
-
input: 0.
|
|
1644
|
-
output: 1.
|
|
1949
|
+
input: 0.39,
|
|
1950
|
+
output: 1.9,
|
|
1645
1951
|
cacheRead: 0,
|
|
1646
1952
|
cacheWrite: 0,
|
|
1647
1953
|
},
|
|
1648
1954
|
contextWindow: 262144,
|
|
1649
|
-
maxTokens:
|
|
1955
|
+
maxTokens: 262144,
|
|
1650
1956
|
},
|
|
1651
1957
|
"deepcogito/cogito-v2-preview-llama-109b-moe": {
|
|
1652
1958
|
id: "deepcogito/cogito-v2-preview-llama-109b-moe",
|
|
@@ -1714,7 +2020,7 @@ export const MODELS = {
|
|
|
1714
2020
|
cacheWrite: 0,
|
|
1715
2021
|
},
|
|
1716
2022
|
contextWindow: 131072,
|
|
1717
|
-
maxTokens:
|
|
2023
|
+
maxTokens: 131072,
|
|
1718
2024
|
},
|
|
1719
2025
|
"nousresearch/hermes-4-405b": {
|
|
1720
2026
|
id: "nousresearch/hermes-4-405b",
|
|
@@ -1725,13 +2031,13 @@ export const MODELS = {
|
|
|
1725
2031
|
reasoning: true,
|
|
1726
2032
|
input: ["text"],
|
|
1727
2033
|
cost: {
|
|
1728
|
-
input: 0.
|
|
1729
|
-
output:
|
|
2034
|
+
input: 0.3,
|
|
2035
|
+
output: 1.2,
|
|
1730
2036
|
cacheRead: 0,
|
|
1731
2037
|
cacheWrite: 0,
|
|
1732
2038
|
},
|
|
1733
2039
|
contextWindow: 131072,
|
|
1734
|
-
maxTokens:
|
|
2040
|
+
maxTokens: 131072,
|
|
1735
2041
|
},
|
|
1736
2042
|
"deepseek/deepseek-chat-v3.1:free": {
|
|
1737
2043
|
id: "deepseek/deepseek-chat-v3.1:free",
|
|
@@ -1747,7 +2053,7 @@ export const MODELS = {
|
|
|
1747
2053
|
cacheRead: 0,
|
|
1748
2054
|
cacheWrite: 0,
|
|
1749
2055
|
},
|
|
1750
|
-
contextWindow:
|
|
2056
|
+
contextWindow: 163800,
|
|
1751
2057
|
maxTokens: 4096,
|
|
1752
2058
|
},
|
|
1753
2059
|
"deepseek/deepseek-chat-v3.1": {
|
|
@@ -1759,13 +2065,13 @@ export const MODELS = {
|
|
|
1759
2065
|
reasoning: true,
|
|
1760
2066
|
input: ["text"],
|
|
1761
2067
|
cost: {
|
|
1762
|
-
input: 0.
|
|
1763
|
-
output: 0.
|
|
2068
|
+
input: 0.19999999999999998,
|
|
2069
|
+
output: 0.7999999999999999,
|
|
1764
2070
|
cacheRead: 0,
|
|
1765
2071
|
cacheWrite: 0,
|
|
1766
2072
|
},
|
|
1767
2073
|
contextWindow: 163840,
|
|
1768
|
-
maxTokens:
|
|
2074
|
+
maxTokens: 163840,
|
|
1769
2075
|
},
|
|
1770
2076
|
"mistralai/mistral-medium-3.1": {
|
|
1771
2077
|
id: "mistralai/mistral-medium-3.1",
|
|
@@ -1793,13 +2099,13 @@ export const MODELS = {
|
|
|
1793
2099
|
reasoning: true,
|
|
1794
2100
|
input: ["text", "image"],
|
|
1795
2101
|
cost: {
|
|
1796
|
-
input: 0.
|
|
2102
|
+
input: 0.6,
|
|
1797
2103
|
output: 1.7999999999999998,
|
|
1798
2104
|
cacheRead: 0,
|
|
1799
2105
|
cacheWrite: 0,
|
|
1800
2106
|
},
|
|
1801
2107
|
contextWindow: 65536,
|
|
1802
|
-
maxTokens:
|
|
2108
|
+
maxTokens: 16384,
|
|
1803
2109
|
},
|
|
1804
2110
|
"ai21/jamba-mini-1.7": {
|
|
1805
2111
|
id: "ai21/jamba-mini-1.7",
|
|
@@ -1861,13 +2167,13 @@ export const MODELS = {
|
|
|
1861
2167
|
reasoning: false,
|
|
1862
2168
|
input: ["text"],
|
|
1863
2169
|
cost: {
|
|
1864
|
-
input: 0.
|
|
1865
|
-
output: 0.
|
|
2170
|
+
input: 0.06,
|
|
2171
|
+
output: 0.25,
|
|
1866
2172
|
cacheRead: 0,
|
|
1867
2173
|
cacheWrite: 0,
|
|
1868
2174
|
},
|
|
1869
2175
|
contextWindow: 262144,
|
|
1870
|
-
maxTokens:
|
|
2176
|
+
maxTokens: 262144,
|
|
1871
2177
|
},
|
|
1872
2178
|
"qwen/qwen3-30b-a3b-instruct-2507": {
|
|
1873
2179
|
id: "qwen/qwen3-30b-a3b-instruct-2507",
|
|
@@ -1878,13 +2184,13 @@ export const MODELS = {
|
|
|
1878
2184
|
reasoning: false,
|
|
1879
2185
|
input: ["text"],
|
|
1880
2186
|
cost: {
|
|
1881
|
-
input: 0.
|
|
1882
|
-
output: 0.
|
|
2187
|
+
input: 0.08,
|
|
2188
|
+
output: 0.33,
|
|
1883
2189
|
cacheRead: 0,
|
|
1884
2190
|
cacheWrite: 0,
|
|
1885
2191
|
},
|
|
1886
2192
|
contextWindow: 262144,
|
|
1887
|
-
maxTokens:
|
|
2193
|
+
maxTokens: 262144,
|
|
1888
2194
|
},
|
|
1889
2195
|
"z-ai/glm-4.5": {
|
|
1890
2196
|
id: "z-ai/glm-4.5",
|
|
@@ -1895,13 +2201,13 @@ export const MODELS = {
|
|
|
1895
2201
|
reasoning: true,
|
|
1896
2202
|
input: ["text"],
|
|
1897
2203
|
cost: {
|
|
1898
|
-
input: 0.
|
|
1899
|
-
output: 1.
|
|
2204
|
+
input: 0.35,
|
|
2205
|
+
output: 1.55,
|
|
1900
2206
|
cacheRead: 0,
|
|
1901
2207
|
cacheWrite: 0,
|
|
1902
2208
|
},
|
|
1903
2209
|
contextWindow: 131072,
|
|
1904
|
-
maxTokens:
|
|
2210
|
+
maxTokens: 131072,
|
|
1905
2211
|
},
|
|
1906
2212
|
"z-ai/glm-4.5-air:free": {
|
|
1907
2213
|
id: "z-ai/glm-4.5-air:free",
|
|
@@ -1918,7 +2224,7 @@ export const MODELS = {
|
|
|
1918
2224
|
cacheWrite: 0,
|
|
1919
2225
|
},
|
|
1920
2226
|
contextWindow: 131072,
|
|
1921
|
-
maxTokens:
|
|
2227
|
+
maxTokens: 131072,
|
|
1922
2228
|
},
|
|
1923
2229
|
"z-ai/glm-4.5-air": {
|
|
1924
2230
|
id: "z-ai/glm-4.5-air",
|
|
@@ -1946,13 +2252,13 @@ export const MODELS = {
|
|
|
1946
2252
|
reasoning: true,
|
|
1947
2253
|
input: ["text"],
|
|
1948
2254
|
cost: {
|
|
1949
|
-
input: 0.
|
|
1950
|
-
output: 0.
|
|
2255
|
+
input: 0.11,
|
|
2256
|
+
output: 0.6,
|
|
1951
2257
|
cacheRead: 0,
|
|
1952
2258
|
cacheWrite: 0,
|
|
1953
2259
|
},
|
|
1954
2260
|
contextWindow: 262144,
|
|
1955
|
-
maxTokens:
|
|
2261
|
+
maxTokens: 262144,
|
|
1956
2262
|
},
|
|
1957
2263
|
"z-ai/glm-4-32b": {
|
|
1958
2264
|
id: "z-ai/glm-4-32b",
|
|
@@ -2003,7 +2309,7 @@ export const MODELS = {
|
|
|
2003
2309
|
cacheWrite: 0,
|
|
2004
2310
|
},
|
|
2005
2311
|
contextWindow: 262144,
|
|
2006
|
-
maxTokens:
|
|
2312
|
+
maxTokens: 262144,
|
|
2007
2313
|
},
|
|
2008
2314
|
"qwen/qwen3-235b-a22b-2507": {
|
|
2009
2315
|
id: "qwen/qwen3-235b-a22b-2507",
|
|
@@ -2014,31 +2320,14 @@ export const MODELS = {
|
|
|
2014
2320
|
reasoning: false,
|
|
2015
2321
|
input: ["text"],
|
|
2016
2322
|
cost: {
|
|
2017
|
-
input: 0.
|
|
2018
|
-
output: 0.
|
|
2323
|
+
input: 0.08,
|
|
2324
|
+
output: 0.55,
|
|
2019
2325
|
cacheRead: 0,
|
|
2020
2326
|
cacheWrite: 0,
|
|
2021
2327
|
},
|
|
2022
2328
|
contextWindow: 262144,
|
|
2023
2329
|
maxTokens: 262144,
|
|
2024
2330
|
},
|
|
2025
|
-
"moonshotai/kimi-k2:free": {
|
|
2026
|
-
id: "moonshotai/kimi-k2:free",
|
|
2027
|
-
name: "MoonshotAI: Kimi K2 0711 (free)",
|
|
2028
|
-
api: "openai-completions",
|
|
2029
|
-
provider: "openrouter",
|
|
2030
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
2031
|
-
reasoning: false,
|
|
2032
|
-
input: ["text"],
|
|
2033
|
-
cost: {
|
|
2034
|
-
input: 0,
|
|
2035
|
-
output: 0,
|
|
2036
|
-
cacheRead: 0,
|
|
2037
|
-
cacheWrite: 0,
|
|
2038
|
-
},
|
|
2039
|
-
contextWindow: 32768,
|
|
2040
|
-
maxTokens: 4096,
|
|
2041
|
-
},
|
|
2042
2331
|
"moonshotai/kimi-k2": {
|
|
2043
2332
|
id: "moonshotai/kimi-k2",
|
|
2044
2333
|
name: "MoonshotAI: Kimi K2 0711",
|
|
@@ -2133,13 +2422,13 @@ export const MODELS = {
|
|
|
2133
2422
|
reasoning: false,
|
|
2134
2423
|
input: ["text", "image"],
|
|
2135
2424
|
cost: {
|
|
2136
|
-
input: 0.
|
|
2137
|
-
output: 0.
|
|
2425
|
+
input: 0.06,
|
|
2426
|
+
output: 0.18,
|
|
2138
2427
|
cacheRead: 0,
|
|
2139
2428
|
cacheWrite: 0,
|
|
2140
2429
|
},
|
|
2141
|
-
contextWindow:
|
|
2142
|
-
maxTokens:
|
|
2430
|
+
contextWindow: 131072,
|
|
2431
|
+
maxTokens: 131072,
|
|
2143
2432
|
},
|
|
2144
2433
|
"minimax/minimax-m1": {
|
|
2145
2434
|
id: "minimax/minimax-m1",
|
|
@@ -2224,7 +2513,7 @@ export const MODELS = {
|
|
|
2224
2513
|
cacheWrite: 0,
|
|
2225
2514
|
},
|
|
2226
2515
|
contextWindow: 163840,
|
|
2227
|
-
maxTokens:
|
|
2516
|
+
maxTokens: 163840,
|
|
2228
2517
|
},
|
|
2229
2518
|
"mistralai/devstral-small-2505:free": {
|
|
2230
2519
|
id: "mistralai/devstral-small-2505:free",
|
|
@@ -2252,13 +2541,13 @@ export const MODELS = {
|
|
|
2252
2541
|
reasoning: false,
|
|
2253
2542
|
input: ["text"],
|
|
2254
2543
|
cost: {
|
|
2255
|
-
input: 0.
|
|
2256
|
-
output: 0.
|
|
2544
|
+
input: 0.049999999999999996,
|
|
2545
|
+
output: 0.22,
|
|
2257
2546
|
cacheRead: 0,
|
|
2258
2547
|
cacheWrite: 0,
|
|
2259
2548
|
},
|
|
2260
2549
|
contextWindow: 131072,
|
|
2261
|
-
maxTokens:
|
|
2550
|
+
maxTokens: 131072,
|
|
2262
2551
|
},
|
|
2263
2552
|
"meta-llama/llama-3.3-8b-instruct:free": {
|
|
2264
2553
|
id: "meta-llama/llama-3.3-8b-instruct:free",
|
|
@@ -2360,7 +2649,7 @@ export const MODELS = {
|
|
|
2360
2649
|
cacheWrite: 0,
|
|
2361
2650
|
},
|
|
2362
2651
|
contextWindow: 40960,
|
|
2363
|
-
maxTokens:
|
|
2652
|
+
maxTokens: 40960,
|
|
2364
2653
|
},
|
|
2365
2654
|
"qwen/qwen3-14b": {
|
|
2366
2655
|
id: "qwen/qwen3-14b",
|
|
@@ -2371,8 +2660,8 @@ export const MODELS = {
|
|
|
2371
2660
|
reasoning: true,
|
|
2372
2661
|
input: ["text"],
|
|
2373
2662
|
cost: {
|
|
2374
|
-
input: 0.
|
|
2375
|
-
output: 0.
|
|
2663
|
+
input: 0.049999999999999996,
|
|
2664
|
+
output: 0.22,
|
|
2376
2665
|
cacheRead: 0,
|
|
2377
2666
|
cacheWrite: 0,
|
|
2378
2667
|
},
|
|
@@ -2388,13 +2677,13 @@ export const MODELS = {
|
|
|
2388
2677
|
reasoning: true,
|
|
2389
2678
|
input: ["text"],
|
|
2390
2679
|
cost: {
|
|
2391
|
-
input: 0.
|
|
2392
|
-
output: 0.
|
|
2680
|
+
input: 0.049999999999999996,
|
|
2681
|
+
output: 0.19999999999999998,
|
|
2393
2682
|
cacheRead: 0,
|
|
2394
2683
|
cacheWrite: 0,
|
|
2395
2684
|
},
|
|
2396
2685
|
contextWindow: 40960,
|
|
2397
|
-
maxTokens:
|
|
2686
|
+
maxTokens: 40960,
|
|
2398
2687
|
},
|
|
2399
2688
|
"qwen/qwen3-235b-a22b:free": {
|
|
2400
2689
|
id: "qwen/qwen3-235b-a22b:free",
|
|
@@ -2495,8 +2784,8 @@ export const MODELS = {
|
|
|
2495
2784
|
cacheRead: 0,
|
|
2496
2785
|
cacheWrite: 0,
|
|
2497
2786
|
},
|
|
2498
|
-
contextWindow:
|
|
2499
|
-
maxTokens:
|
|
2787
|
+
contextWindow: 327680,
|
|
2788
|
+
maxTokens: 16384,
|
|
2500
2789
|
},
|
|
2501
2790
|
"deepseek/deepseek-chat-v3-0324:free": {
|
|
2502
2791
|
id: "deepseek/deepseek-chat-v3-0324:free",
|
|
@@ -2524,13 +2813,13 @@ export const MODELS = {
|
|
|
2524
2813
|
reasoning: false,
|
|
2525
2814
|
input: ["text"],
|
|
2526
2815
|
cost: {
|
|
2527
|
-
input: 0.
|
|
2528
|
-
output: 0.
|
|
2816
|
+
input: 0.24,
|
|
2817
|
+
output: 0.84,
|
|
2529
2818
|
cacheRead: 0,
|
|
2530
2819
|
cacheWrite: 0,
|
|
2531
2820
|
},
|
|
2532
2821
|
contextWindow: 163840,
|
|
2533
|
-
maxTokens:
|
|
2822
|
+
maxTokens: 163840,
|
|
2534
2823
|
},
|
|
2535
2824
|
"mistralai/mistral-small-3.1-24b-instruct:free": {
|
|
2536
2825
|
id: "mistralai/mistral-small-3.1-24b-instruct:free",
|
|
@@ -2558,13 +2847,13 @@ export const MODELS = {
|
|
|
2558
2847
|
reasoning: false,
|
|
2559
2848
|
input: ["text", "image"],
|
|
2560
2849
|
cost: {
|
|
2561
|
-
input: 0.
|
|
2562
|
-
output: 0.
|
|
2850
|
+
input: 0.049999999999999996,
|
|
2851
|
+
output: 0.09999999999999999,
|
|
2563
2852
|
cacheRead: 0,
|
|
2564
2853
|
cacheWrite: 0,
|
|
2565
2854
|
},
|
|
2566
|
-
contextWindow:
|
|
2567
|
-
maxTokens:
|
|
2855
|
+
contextWindow: 128000,
|
|
2856
|
+
maxTokens: 4096,
|
|
2568
2857
|
},
|
|
2569
2858
|
"microsoft/phi-4-multimodal-instruct": {
|
|
2570
2859
|
id: "microsoft/phi-4-multimodal-instruct",
|
|
@@ -2677,13 +2966,13 @@ export const MODELS = {
|
|
|
2677
2966
|
reasoning: false,
|
|
2678
2967
|
input: ["text"],
|
|
2679
2968
|
cost: {
|
|
2680
|
-
input: 0.
|
|
2681
|
-
output: 0.
|
|
2969
|
+
input: 0.049999999999999996,
|
|
2970
|
+
output: 0.08,
|
|
2682
2971
|
cacheRead: 0,
|
|
2683
2972
|
cacheWrite: 0,
|
|
2684
2973
|
},
|
|
2685
2974
|
contextWindow: 32768,
|
|
2686
|
-
maxTokens:
|
|
2975
|
+
maxTokens: 16384,
|
|
2687
2976
|
},
|
|
2688
2977
|
"deepseek/deepseek-r1-distill-llama-70b": {
|
|
2689
2978
|
id: "deepseek/deepseek-r1-distill-llama-70b",
|
|
@@ -2700,7 +2989,7 @@ export const MODELS = {
|
|
|
2700
2989
|
cacheWrite: 0,
|
|
2701
2990
|
},
|
|
2702
2991
|
contextWindow: 131072,
|
|
2703
|
-
maxTokens:
|
|
2992
|
+
maxTokens: 131072,
|
|
2704
2993
|
},
|
|
2705
2994
|
"deepseek/deepseek-r1": {
|
|
2706
2995
|
id: "deepseek/deepseek-r1",
|
|
@@ -2745,13 +3034,13 @@ export const MODELS = {
|
|
|
2745
3034
|
reasoning: false,
|
|
2746
3035
|
input: ["text"],
|
|
2747
3036
|
cost: {
|
|
2748
|
-
input: 0.
|
|
2749
|
-
output: 0.
|
|
3037
|
+
input: 0.3,
|
|
3038
|
+
output: 0.85,
|
|
2750
3039
|
cacheRead: 0,
|
|
2751
3040
|
cacheWrite: 0,
|
|
2752
3041
|
},
|
|
2753
3042
|
contextWindow: 163840,
|
|
2754
|
-
maxTokens:
|
|
3043
|
+
maxTokens: 163840,
|
|
2755
3044
|
},
|
|
2756
3045
|
"meta-llama/llama-3.3-70b-instruct:free": {
|
|
2757
3046
|
id: "meta-llama/llama-3.3-70b-instruct:free",
|
|
@@ -2779,13 +3068,13 @@ export const MODELS = {
|
|
|
2779
3068
|
reasoning: false,
|
|
2780
3069
|
input: ["text"],
|
|
2781
3070
|
cost: {
|
|
2782
|
-
input: 0.
|
|
2783
|
-
output: 0.
|
|
3071
|
+
input: 0.04,
|
|
3072
|
+
output: 0.12,
|
|
2784
3073
|
cacheRead: 0,
|
|
2785
3074
|
cacheWrite: 0,
|
|
2786
3075
|
},
|
|
2787
3076
|
contextWindow: 131072,
|
|
2788
|
-
maxTokens:
|
|
3077
|
+
maxTokens: 131072,
|
|
2789
3078
|
},
|
|
2790
3079
|
"amazon/nova-lite-v1": {
|
|
2791
3080
|
id: "amazon/nova-lite-v1",
|
|
@@ -2989,7 +3278,7 @@ export const MODELS = {
|
|
|
2989
3278
|
cacheWrite: 0,
|
|
2990
3279
|
},
|
|
2991
3280
|
contextWindow: 32768,
|
|
2992
|
-
maxTokens:
|
|
3281
|
+
maxTokens: 32768,
|
|
2993
3282
|
},
|
|
2994
3283
|
"mistralai/pixtral-12b": {
|
|
2995
3284
|
id: "mistralai/pixtral-12b",
|
|
@@ -3008,34 +3297,34 @@ export const MODELS = {
|
|
|
3008
3297
|
contextWindow: 32768,
|
|
3009
3298
|
maxTokens: 4096,
|
|
3010
3299
|
},
|
|
3011
|
-
"cohere/command-r-08-2024": {
|
|
3012
|
-
id: "cohere/command-r-08-2024",
|
|
3013
|
-
name: "Cohere: Command R (08-2024)",
|
|
3300
|
+
"cohere/command-r-plus-08-2024": {
|
|
3301
|
+
id: "cohere/command-r-plus-08-2024",
|
|
3302
|
+
name: "Cohere: Command R+ (08-2024)",
|
|
3014
3303
|
api: "openai-completions",
|
|
3015
3304
|
provider: "openrouter",
|
|
3016
3305
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
3017
3306
|
reasoning: false,
|
|
3018
3307
|
input: ["text"],
|
|
3019
3308
|
cost: {
|
|
3020
|
-
input:
|
|
3021
|
-
output:
|
|
3309
|
+
input: 2.5,
|
|
3310
|
+
output: 10,
|
|
3022
3311
|
cacheRead: 0,
|
|
3023
3312
|
cacheWrite: 0,
|
|
3024
3313
|
},
|
|
3025
3314
|
contextWindow: 128000,
|
|
3026
3315
|
maxTokens: 4000,
|
|
3027
3316
|
},
|
|
3028
|
-
"cohere/command-r-
|
|
3029
|
-
id: "cohere/command-r-
|
|
3030
|
-
name: "Cohere: Command R
|
|
3317
|
+
"cohere/command-r-08-2024": {
|
|
3318
|
+
id: "cohere/command-r-08-2024",
|
|
3319
|
+
name: "Cohere: Command R (08-2024)",
|
|
3031
3320
|
api: "openai-completions",
|
|
3032
3321
|
provider: "openrouter",
|
|
3033
3322
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
3034
3323
|
reasoning: false,
|
|
3035
3324
|
input: ["text"],
|
|
3036
3325
|
cost: {
|
|
3037
|
-
input:
|
|
3038
|
-
output:
|
|
3326
|
+
input: 0.15,
|
|
3327
|
+
output: 0.6,
|
|
3039
3328
|
cacheRead: 0,
|
|
3040
3329
|
cacheWrite: 0,
|
|
3041
3330
|
},
|
|
@@ -3068,46 +3357,46 @@ export const MODELS = {
|
|
|
3068
3357
|
reasoning: false,
|
|
3069
3358
|
input: ["text"],
|
|
3070
3359
|
cost: {
|
|
3071
|
-
input: 0.
|
|
3360
|
+
input: 0.3,
|
|
3072
3361
|
output: 0.3,
|
|
3073
3362
|
cacheRead: 0,
|
|
3074
3363
|
cacheWrite: 0,
|
|
3075
3364
|
},
|
|
3076
|
-
contextWindow:
|
|
3077
|
-
maxTokens:
|
|
3365
|
+
contextWindow: 65000,
|
|
3366
|
+
maxTokens: 4096,
|
|
3078
3367
|
},
|
|
3079
|
-
"meta-llama/llama-3.1-
|
|
3080
|
-
id: "meta-llama/llama-3.1-
|
|
3081
|
-
name: "Meta: Llama 3.1
|
|
3368
|
+
"meta-llama/llama-3.1-405b-instruct": {
|
|
3369
|
+
id: "meta-llama/llama-3.1-405b-instruct",
|
|
3370
|
+
name: "Meta: Llama 3.1 405B Instruct",
|
|
3082
3371
|
api: "openai-completions",
|
|
3083
3372
|
provider: "openrouter",
|
|
3084
3373
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
3085
3374
|
reasoning: false,
|
|
3086
3375
|
input: ["text"],
|
|
3087
3376
|
cost: {
|
|
3088
|
-
input: 0.
|
|
3089
|
-
output: 0.
|
|
3377
|
+
input: 0.7999999999999999,
|
|
3378
|
+
output: 0.7999999999999999,
|
|
3090
3379
|
cacheRead: 0,
|
|
3091
3380
|
cacheWrite: 0,
|
|
3092
3381
|
},
|
|
3093
|
-
contextWindow:
|
|
3382
|
+
contextWindow: 32768,
|
|
3094
3383
|
maxTokens: 16384,
|
|
3095
3384
|
},
|
|
3096
|
-
"meta-llama/llama-3.1-
|
|
3097
|
-
id: "meta-llama/llama-3.1-
|
|
3098
|
-
name: "Meta: Llama 3.1
|
|
3385
|
+
"meta-llama/llama-3.1-8b-instruct": {
|
|
3386
|
+
id: "meta-llama/llama-3.1-8b-instruct",
|
|
3387
|
+
name: "Meta: Llama 3.1 8B Instruct",
|
|
3099
3388
|
api: "openai-completions",
|
|
3100
3389
|
provider: "openrouter",
|
|
3101
3390
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
3102
3391
|
reasoning: false,
|
|
3103
3392
|
input: ["text"],
|
|
3104
3393
|
cost: {
|
|
3105
|
-
input: 0.
|
|
3106
|
-
output: 0.
|
|
3394
|
+
input: 0.02,
|
|
3395
|
+
output: 0.03,
|
|
3107
3396
|
cacheRead: 0,
|
|
3108
3397
|
cacheWrite: 0,
|
|
3109
3398
|
},
|
|
3110
|
-
contextWindow:
|
|
3399
|
+
contextWindow: 16384,
|
|
3111
3400
|
maxTokens: 16384,
|
|
3112
3401
|
},
|
|
3113
3402
|
"meta-llama/llama-3.1-70b-instruct": {
|
|
@@ -3119,13 +3408,13 @@ export const MODELS = {
|
|
|
3119
3408
|
reasoning: false,
|
|
3120
3409
|
input: ["text"],
|
|
3121
3410
|
cost: {
|
|
3122
|
-
input: 0.
|
|
3123
|
-
output: 0.
|
|
3411
|
+
input: 0.39999999999999997,
|
|
3412
|
+
output: 0.39999999999999997,
|
|
3124
3413
|
cacheRead: 0,
|
|
3125
3414
|
cacheWrite: 0,
|
|
3126
3415
|
},
|
|
3127
3416
|
contextWindow: 131072,
|
|
3128
|
-
maxTokens:
|
|
3417
|
+
maxTokens: 4096,
|
|
3129
3418
|
},
|
|
3130
3419
|
"mistralai/mistral-nemo": {
|
|
3131
3420
|
id: "mistralai/mistral-nemo",
|
|
@@ -3144,43 +3433,43 @@ export const MODELS = {
|
|
|
3144
3433
|
contextWindow: 131072,
|
|
3145
3434
|
maxTokens: 16384,
|
|
3146
3435
|
},
|
|
3147
|
-
"mistralai/mistral-7b-instruct
|
|
3148
|
-
id: "mistralai/mistral-7b-instruct
|
|
3149
|
-
name: "Mistral: Mistral 7B Instruct
|
|
3436
|
+
"mistralai/mistral-7b-instruct:free": {
|
|
3437
|
+
id: "mistralai/mistral-7b-instruct:free",
|
|
3438
|
+
name: "Mistral: Mistral 7B Instruct (free)",
|
|
3150
3439
|
api: "openai-completions",
|
|
3151
3440
|
provider: "openrouter",
|
|
3152
3441
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
3153
3442
|
reasoning: false,
|
|
3154
3443
|
input: ["text"],
|
|
3155
3444
|
cost: {
|
|
3156
|
-
input: 0
|
|
3157
|
-
output: 0
|
|
3445
|
+
input: 0,
|
|
3446
|
+
output: 0,
|
|
3158
3447
|
cacheRead: 0,
|
|
3159
3448
|
cacheWrite: 0,
|
|
3160
3449
|
},
|
|
3161
3450
|
contextWindow: 32768,
|
|
3162
3451
|
maxTokens: 16384,
|
|
3163
3452
|
},
|
|
3164
|
-
"mistralai/mistral-7b-instruct
|
|
3165
|
-
id: "mistralai/mistral-7b-instruct
|
|
3166
|
-
name: "Mistral: Mistral 7B Instruct
|
|
3453
|
+
"mistralai/mistral-7b-instruct": {
|
|
3454
|
+
id: "mistralai/mistral-7b-instruct",
|
|
3455
|
+
name: "Mistral: Mistral 7B Instruct",
|
|
3167
3456
|
api: "openai-completions",
|
|
3168
3457
|
provider: "openrouter",
|
|
3169
3458
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
3170
3459
|
reasoning: false,
|
|
3171
3460
|
input: ["text"],
|
|
3172
3461
|
cost: {
|
|
3173
|
-
input: 0,
|
|
3174
|
-
output: 0,
|
|
3462
|
+
input: 0.028,
|
|
3463
|
+
output: 0.054,
|
|
3175
3464
|
cacheRead: 0,
|
|
3176
3465
|
cacheWrite: 0,
|
|
3177
3466
|
},
|
|
3178
3467
|
contextWindow: 32768,
|
|
3179
3468
|
maxTokens: 16384,
|
|
3180
3469
|
},
|
|
3181
|
-
"mistralai/mistral-7b-instruct": {
|
|
3182
|
-
id: "mistralai/mistral-7b-instruct",
|
|
3183
|
-
name: "Mistral: Mistral 7B Instruct",
|
|
3470
|
+
"mistralai/mistral-7b-instruct-v0.3": {
|
|
3471
|
+
id: "mistralai/mistral-7b-instruct-v0.3",
|
|
3472
|
+
name: "Mistral: Mistral 7B Instruct v0.3",
|
|
3184
3473
|
api: "openai-completions",
|
|
3185
3474
|
provider: "openrouter",
|
|
3186
3475
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -3280,74 +3569,6 @@ export const MODELS = {
|
|
|
3280
3569
|
contextWindow: 65536,
|
|
3281
3570
|
maxTokens: 4096,
|
|
3282
3571
|
},
|
|
3283
|
-
"cohere/command-r-plus": {
|
|
3284
|
-
id: "cohere/command-r-plus",
|
|
3285
|
-
name: "Cohere: Command R+",
|
|
3286
|
-
api: "openai-completions",
|
|
3287
|
-
provider: "openrouter",
|
|
3288
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
3289
|
-
reasoning: false,
|
|
3290
|
-
input: ["text"],
|
|
3291
|
-
cost: {
|
|
3292
|
-
input: 3,
|
|
3293
|
-
output: 15,
|
|
3294
|
-
cacheRead: 0,
|
|
3295
|
-
cacheWrite: 0,
|
|
3296
|
-
},
|
|
3297
|
-
contextWindow: 128000,
|
|
3298
|
-
maxTokens: 4000,
|
|
3299
|
-
},
|
|
3300
|
-
"cohere/command-r-plus-04-2024": {
|
|
3301
|
-
id: "cohere/command-r-plus-04-2024",
|
|
3302
|
-
name: "Cohere: Command R+ (04-2024)",
|
|
3303
|
-
api: "openai-completions",
|
|
3304
|
-
provider: "openrouter",
|
|
3305
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
3306
|
-
reasoning: false,
|
|
3307
|
-
input: ["text"],
|
|
3308
|
-
cost: {
|
|
3309
|
-
input: 3,
|
|
3310
|
-
output: 15,
|
|
3311
|
-
cacheRead: 0,
|
|
3312
|
-
cacheWrite: 0,
|
|
3313
|
-
},
|
|
3314
|
-
contextWindow: 128000,
|
|
3315
|
-
maxTokens: 4000,
|
|
3316
|
-
},
|
|
3317
|
-
"cohere/command-r": {
|
|
3318
|
-
id: "cohere/command-r",
|
|
3319
|
-
name: "Cohere: Command R",
|
|
3320
|
-
api: "openai-completions",
|
|
3321
|
-
provider: "openrouter",
|
|
3322
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
3323
|
-
reasoning: false,
|
|
3324
|
-
input: ["text"],
|
|
3325
|
-
cost: {
|
|
3326
|
-
input: 0.5,
|
|
3327
|
-
output: 1.5,
|
|
3328
|
-
cacheRead: 0,
|
|
3329
|
-
cacheWrite: 0,
|
|
3330
|
-
},
|
|
3331
|
-
contextWindow: 128000,
|
|
3332
|
-
maxTokens: 4000,
|
|
3333
|
-
},
|
|
3334
|
-
"cohere/command-r-03-2024": {
|
|
3335
|
-
id: "cohere/command-r-03-2024",
|
|
3336
|
-
name: "Cohere: Command R (03-2024)",
|
|
3337
|
-
api: "openai-completions",
|
|
3338
|
-
provider: "openrouter",
|
|
3339
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
3340
|
-
reasoning: false,
|
|
3341
|
-
input: ["text"],
|
|
3342
|
-
cost: {
|
|
3343
|
-
input: 0.5,
|
|
3344
|
-
output: 1.5,
|
|
3345
|
-
cacheRead: 0,
|
|
3346
|
-
cacheWrite: 0,
|
|
3347
|
-
},
|
|
3348
|
-
contextWindow: 128000,
|
|
3349
|
-
maxTokens: 4000,
|
|
3350
|
-
},
|
|
3351
3572
|
"mistralai/mistral-large": {
|
|
3352
3573
|
id: "mistralai/mistral-large",
|
|
3353
3574
|
name: "Mistral Large",
|
|
@@ -3365,34 +3586,34 @@ export const MODELS = {
|
|
|
3365
3586
|
contextWindow: 128000,
|
|
3366
3587
|
maxTokens: 4096,
|
|
3367
3588
|
},
|
|
3368
|
-
"mistralai/mistral-
|
|
3369
|
-
id: "mistralai/mistral-
|
|
3370
|
-
name: "Mistral
|
|
3589
|
+
"mistralai/mistral-small": {
|
|
3590
|
+
id: "mistralai/mistral-small",
|
|
3591
|
+
name: "Mistral Small",
|
|
3371
3592
|
api: "openai-completions",
|
|
3372
3593
|
provider: "openrouter",
|
|
3373
3594
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
3374
3595
|
reasoning: false,
|
|
3375
3596
|
input: ["text"],
|
|
3376
3597
|
cost: {
|
|
3377
|
-
input: 0.
|
|
3378
|
-
output: 0.
|
|
3598
|
+
input: 0.19999999999999998,
|
|
3599
|
+
output: 0.6,
|
|
3379
3600
|
cacheRead: 0,
|
|
3380
3601
|
cacheWrite: 0,
|
|
3381
3602
|
},
|
|
3382
3603
|
contextWindow: 32768,
|
|
3383
3604
|
maxTokens: 4096,
|
|
3384
3605
|
},
|
|
3385
|
-
"mistralai/mistral-
|
|
3386
|
-
id: "mistralai/mistral-
|
|
3387
|
-
name: "Mistral
|
|
3606
|
+
"mistralai/mistral-tiny": {
|
|
3607
|
+
id: "mistralai/mistral-tiny",
|
|
3608
|
+
name: "Mistral Tiny",
|
|
3388
3609
|
api: "openai-completions",
|
|
3389
3610
|
provider: "openrouter",
|
|
3390
3611
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
3391
3612
|
reasoning: false,
|
|
3392
3613
|
input: ["text"],
|
|
3393
3614
|
cost: {
|
|
3394
|
-
input: 0.
|
|
3395
|
-
output: 0.
|
|
3615
|
+
input: 0.25,
|
|
3616
|
+
output: 0.25,
|
|
3396
3617
|
cacheRead: 0,
|
|
3397
3618
|
cacheWrite: 0,
|
|
3398
3619
|
},
|
|
@@ -3408,8 +3629,8 @@ export const MODELS = {
|
|
|
3408
3629
|
reasoning: false,
|
|
3409
3630
|
input: ["text"],
|
|
3410
3631
|
cost: {
|
|
3411
|
-
input: 0.
|
|
3412
|
-
output: 0.
|
|
3632
|
+
input: 0.54,
|
|
3633
|
+
output: 0.54,
|
|
3413
3634
|
cacheRead: 0,
|
|
3414
3635
|
cacheWrite: 0,
|
|
3415
3636
|
},
|