@mariozechner/pi-ai 0.5.43 → 0.5.45
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/agent/agent-loop.d.ts +2 -2
- package/dist/agent/agent-loop.d.ts.map +1 -1
- package/dist/agent/agent-loop.js +18 -2
- package/dist/agent/agent-loop.js.map +1 -1
- package/dist/agent/index.d.ts +1 -1
- package/dist/agent/index.d.ts.map +1 -1
- package/dist/agent/index.js.map +1 -1
- package/dist/agent/types.d.ts +6 -1
- package/dist/agent/types.d.ts.map +1 -1
- package/dist/agent/types.js.map +1 -1
- package/dist/models.generated.d.ts +933 -270
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +1500 -837
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/anthropic.d.ts.map +1 -1
- package/dist/providers/anthropic.js +32 -8
- package/dist/providers/anthropic.js.map +1 -1
- package/dist/providers/google.d.ts.map +1 -1
- package/dist/providers/google.js +8 -7
- package/dist/providers/google.js.map +1 -1
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +6 -5
- package/dist/providers/openai-completions.js.map +1 -1
- package/dist/providers/openai-responses.d.ts.map +1 -1
- package/dist/providers/openai-responses.js +6 -5
- package/dist/providers/openai-responses.js.map +1 -1
- package/dist/providers/transorm-messages.d.ts.map +1 -1
- package/dist/providers/transorm-messages.js +46 -1
- package/dist/providers/transorm-messages.js.map +1 -1
- package/dist/utils/event-stream.js +3 -7
- package/dist/utils/event-stream.js.map +1 -1
- package/dist/utils/sanitize-unicode.d.ts +22 -0
- package/dist/utils/sanitize-unicode.d.ts.map +1 -0
- package/dist/utils/sanitize-unicode.js +26 -0
- package/dist/utils/sanitize-unicode.js.map +1 -0
- package/dist/utils/validation.d.ts.map +1 -1
- package/dist/utils/validation.js +25 -3
- package/dist/utils/validation.js.map +1 -1
- package/package.json +3 -3
|
@@ -1,12 +1,12 @@
|
|
|
1
1
|
export declare const MODELS: {
|
|
2
2
|
readonly anthropic: {
|
|
3
|
-
readonly "claude-3-
|
|
3
|
+
readonly "claude-3-5-sonnet-20241022": {
|
|
4
4
|
id: string;
|
|
5
5
|
name: string;
|
|
6
6
|
api: "anthropic-messages";
|
|
7
7
|
provider: string;
|
|
8
8
|
baseUrl: string;
|
|
9
|
-
reasoning:
|
|
9
|
+
reasoning: false;
|
|
10
10
|
input: ("text" | "image")[];
|
|
11
11
|
cost: {
|
|
12
12
|
input: number;
|
|
@@ -17,13 +17,13 @@ export declare const MODELS: {
|
|
|
17
17
|
contextWindow: number;
|
|
18
18
|
maxTokens: number;
|
|
19
19
|
};
|
|
20
|
-
readonly "claude-
|
|
20
|
+
readonly "claude-3-5-sonnet-20240620": {
|
|
21
21
|
id: string;
|
|
22
22
|
name: string;
|
|
23
23
|
api: "anthropic-messages";
|
|
24
24
|
provider: string;
|
|
25
25
|
baseUrl: string;
|
|
26
|
-
reasoning:
|
|
26
|
+
reasoning: false;
|
|
27
27
|
input: ("text" | "image")[];
|
|
28
28
|
cost: {
|
|
29
29
|
input: number;
|
|
@@ -34,7 +34,7 @@ export declare const MODELS: {
|
|
|
34
34
|
contextWindow: number;
|
|
35
35
|
maxTokens: number;
|
|
36
36
|
};
|
|
37
|
-
readonly "claude-3-
|
|
37
|
+
readonly "claude-3-opus-20240229": {
|
|
38
38
|
id: string;
|
|
39
39
|
name: string;
|
|
40
40
|
api: "anthropic-messages";
|
|
@@ -51,13 +51,30 @@ export declare const MODELS: {
|
|
|
51
51
|
contextWindow: number;
|
|
52
52
|
maxTokens: number;
|
|
53
53
|
};
|
|
54
|
-
readonly "claude-
|
|
54
|
+
readonly "claude-sonnet-4-5-20250929": {
|
|
55
55
|
id: string;
|
|
56
56
|
name: string;
|
|
57
57
|
api: "anthropic-messages";
|
|
58
58
|
provider: string;
|
|
59
59
|
baseUrl: string;
|
|
60
|
-
reasoning:
|
|
60
|
+
reasoning: true;
|
|
61
|
+
input: ("text" | "image")[];
|
|
62
|
+
cost: {
|
|
63
|
+
input: number;
|
|
64
|
+
output: number;
|
|
65
|
+
cacheRead: number;
|
|
66
|
+
cacheWrite: number;
|
|
67
|
+
};
|
|
68
|
+
contextWindow: number;
|
|
69
|
+
maxTokens: number;
|
|
70
|
+
};
|
|
71
|
+
readonly "claude-sonnet-4-20250514": {
|
|
72
|
+
id: string;
|
|
73
|
+
name: string;
|
|
74
|
+
api: "anthropic-messages";
|
|
75
|
+
provider: string;
|
|
76
|
+
baseUrl: string;
|
|
77
|
+
reasoning: true;
|
|
61
78
|
input: ("text" | "image")[];
|
|
62
79
|
cost: {
|
|
63
80
|
input: number;
|
|
@@ -85,7 +102,7 @@ export declare const MODELS: {
|
|
|
85
102
|
contextWindow: number;
|
|
86
103
|
maxTokens: number;
|
|
87
104
|
};
|
|
88
|
-
readonly "claude-3-5-
|
|
105
|
+
readonly "claude-3-5-haiku-20241022": {
|
|
89
106
|
id: string;
|
|
90
107
|
name: string;
|
|
91
108
|
api: "anthropic-messages";
|
|
@@ -102,7 +119,7 @@ export declare const MODELS: {
|
|
|
102
119
|
contextWindow: number;
|
|
103
120
|
maxTokens: number;
|
|
104
121
|
};
|
|
105
|
-
readonly "claude-3-
|
|
122
|
+
readonly "claude-3-haiku-20240307": {
|
|
106
123
|
id: string;
|
|
107
124
|
name: string;
|
|
108
125
|
api: "anthropic-messages";
|
|
@@ -119,13 +136,13 @@ export declare const MODELS: {
|
|
|
119
136
|
contextWindow: number;
|
|
120
137
|
maxTokens: number;
|
|
121
138
|
};
|
|
122
|
-
readonly "claude-3-sonnet-
|
|
139
|
+
readonly "claude-3-7-sonnet-20250219": {
|
|
123
140
|
id: string;
|
|
124
141
|
name: string;
|
|
125
142
|
api: "anthropic-messages";
|
|
126
143
|
provider: string;
|
|
127
144
|
baseUrl: string;
|
|
128
|
-
reasoning:
|
|
145
|
+
reasoning: true;
|
|
129
146
|
input: ("text" | "image")[];
|
|
130
147
|
cost: {
|
|
131
148
|
input: number;
|
|
@@ -136,7 +153,7 @@ export declare const MODELS: {
|
|
|
136
153
|
contextWindow: number;
|
|
137
154
|
maxTokens: number;
|
|
138
155
|
};
|
|
139
|
-
readonly "claude-
|
|
156
|
+
readonly "claude-opus-4-1-20250805": {
|
|
140
157
|
id: string;
|
|
141
158
|
name: string;
|
|
142
159
|
api: "anthropic-messages";
|
|
@@ -153,7 +170,7 @@ export declare const MODELS: {
|
|
|
153
170
|
contextWindow: number;
|
|
154
171
|
maxTokens: number;
|
|
155
172
|
};
|
|
156
|
-
readonly "claude-3-
|
|
173
|
+
readonly "claude-3-sonnet-20240229": {
|
|
157
174
|
id: string;
|
|
158
175
|
name: string;
|
|
159
176
|
api: "anthropic-messages";
|
|
@@ -170,12 +187,10 @@ export declare const MODELS: {
|
|
|
170
187
|
contextWindow: number;
|
|
171
188
|
maxTokens: number;
|
|
172
189
|
};
|
|
173
|
-
|
|
174
|
-
readonly google: {
|
|
175
|
-
readonly "gemini-2.5-flash-preview-05-20": {
|
|
190
|
+
readonly "claude-haiku-4-5-20251001": {
|
|
176
191
|
id: string;
|
|
177
192
|
name: string;
|
|
178
|
-
api: "
|
|
193
|
+
api: "anthropic-messages";
|
|
179
194
|
provider: string;
|
|
180
195
|
baseUrl: string;
|
|
181
196
|
reasoning: true;
|
|
@@ -189,10 +204,10 @@ export declare const MODELS: {
|
|
|
189
204
|
contextWindow: number;
|
|
190
205
|
maxTokens: number;
|
|
191
206
|
};
|
|
192
|
-
readonly "
|
|
207
|
+
readonly "claude-sonnet-4-0": {
|
|
193
208
|
id: string;
|
|
194
209
|
name: string;
|
|
195
|
-
api: "
|
|
210
|
+
api: "anthropic-messages";
|
|
196
211
|
provider: string;
|
|
197
212
|
baseUrl: string;
|
|
198
213
|
reasoning: true;
|
|
@@ -206,13 +221,13 @@ export declare const MODELS: {
|
|
|
206
221
|
contextWindow: number;
|
|
207
222
|
maxTokens: number;
|
|
208
223
|
};
|
|
209
|
-
readonly "
|
|
224
|
+
readonly "claude-3-7-sonnet-latest": {
|
|
210
225
|
id: string;
|
|
211
226
|
name: string;
|
|
212
|
-
api: "
|
|
227
|
+
api: "anthropic-messages";
|
|
213
228
|
provider: string;
|
|
214
229
|
baseUrl: string;
|
|
215
|
-
reasoning:
|
|
230
|
+
reasoning: true;
|
|
216
231
|
input: ("text" | "image")[];
|
|
217
232
|
cost: {
|
|
218
233
|
input: number;
|
|
@@ -223,13 +238,13 @@ export declare const MODELS: {
|
|
|
223
238
|
contextWindow: number;
|
|
224
239
|
maxTokens: number;
|
|
225
240
|
};
|
|
226
|
-
readonly "
|
|
241
|
+
readonly "claude-sonnet-4-5": {
|
|
227
242
|
id: string;
|
|
228
243
|
name: string;
|
|
229
|
-
api: "
|
|
244
|
+
api: "anthropic-messages";
|
|
230
245
|
provider: string;
|
|
231
246
|
baseUrl: string;
|
|
232
|
-
reasoning:
|
|
247
|
+
reasoning: true;
|
|
233
248
|
input: ("text" | "image")[];
|
|
234
249
|
cost: {
|
|
235
250
|
input: number;
|
|
@@ -240,10 +255,10 @@ export declare const MODELS: {
|
|
|
240
255
|
contextWindow: number;
|
|
241
256
|
maxTokens: number;
|
|
242
257
|
};
|
|
243
|
-
readonly "
|
|
258
|
+
readonly "claude-3-5-haiku-latest": {
|
|
244
259
|
id: string;
|
|
245
260
|
name: string;
|
|
246
|
-
api: "
|
|
261
|
+
api: "anthropic-messages";
|
|
247
262
|
provider: string;
|
|
248
263
|
baseUrl: string;
|
|
249
264
|
reasoning: false;
|
|
@@ -257,13 +272,13 @@ export declare const MODELS: {
|
|
|
257
272
|
contextWindow: number;
|
|
258
273
|
maxTokens: number;
|
|
259
274
|
};
|
|
260
|
-
readonly "
|
|
275
|
+
readonly "claude-haiku-4-5": {
|
|
261
276
|
id: string;
|
|
262
277
|
name: string;
|
|
263
|
-
api: "
|
|
278
|
+
api: "anthropic-messages";
|
|
264
279
|
provider: string;
|
|
265
280
|
baseUrl: string;
|
|
266
|
-
reasoning:
|
|
281
|
+
reasoning: true;
|
|
267
282
|
input: ("text" | "image")[];
|
|
268
283
|
cost: {
|
|
269
284
|
input: number;
|
|
@@ -274,10 +289,10 @@ export declare const MODELS: {
|
|
|
274
289
|
contextWindow: number;
|
|
275
290
|
maxTokens: number;
|
|
276
291
|
};
|
|
277
|
-
readonly "
|
|
292
|
+
readonly "claude-opus-4-1": {
|
|
278
293
|
id: string;
|
|
279
294
|
name: string;
|
|
280
|
-
api: "
|
|
295
|
+
api: "anthropic-messages";
|
|
281
296
|
provider: string;
|
|
282
297
|
baseUrl: string;
|
|
283
298
|
reasoning: true;
|
|
@@ -291,10 +306,10 @@ export declare const MODELS: {
|
|
|
291
306
|
contextWindow: number;
|
|
292
307
|
maxTokens: number;
|
|
293
308
|
};
|
|
294
|
-
readonly "
|
|
309
|
+
readonly "claude-opus-4-0": {
|
|
295
310
|
id: string;
|
|
296
311
|
name: string;
|
|
297
|
-
api: "
|
|
312
|
+
api: "anthropic-messages";
|
|
298
313
|
provider: string;
|
|
299
314
|
baseUrl: string;
|
|
300
315
|
reasoning: true;
|
|
@@ -308,7 +323,9 @@ export declare const MODELS: {
|
|
|
308
323
|
contextWindow: number;
|
|
309
324
|
maxTokens: number;
|
|
310
325
|
};
|
|
311
|
-
|
|
326
|
+
};
|
|
327
|
+
readonly google: {
|
|
328
|
+
readonly "gemini-2.5-flash-preview-05-20": {
|
|
312
329
|
id: string;
|
|
313
330
|
name: string;
|
|
314
331
|
api: "google-generative-ai";
|
|
@@ -325,13 +342,13 @@ export declare const MODELS: {
|
|
|
325
342
|
contextWindow: number;
|
|
326
343
|
maxTokens: number;
|
|
327
344
|
};
|
|
328
|
-
readonly "gemini-
|
|
345
|
+
readonly "gemini-flash-lite-latest": {
|
|
329
346
|
id: string;
|
|
330
347
|
name: string;
|
|
331
348
|
api: "google-generative-ai";
|
|
332
349
|
provider: string;
|
|
333
350
|
baseUrl: string;
|
|
334
|
-
reasoning:
|
|
351
|
+
reasoning: true;
|
|
335
352
|
input: ("text" | "image")[];
|
|
336
353
|
cost: {
|
|
337
354
|
input: number;
|
|
@@ -342,7 +359,7 @@ export declare const MODELS: {
|
|
|
342
359
|
contextWindow: number;
|
|
343
360
|
maxTokens: number;
|
|
344
361
|
};
|
|
345
|
-
readonly "gemini-2.5-flash
|
|
362
|
+
readonly "gemini-2.5-flash": {
|
|
346
363
|
id: string;
|
|
347
364
|
name: string;
|
|
348
365
|
api: "google-generative-ai";
|
|
@@ -359,7 +376,7 @@ export declare const MODELS: {
|
|
|
359
376
|
contextWindow: number;
|
|
360
377
|
maxTokens: number;
|
|
361
378
|
};
|
|
362
|
-
readonly "gemini-
|
|
379
|
+
readonly "gemini-flash-latest": {
|
|
363
380
|
id: string;
|
|
364
381
|
name: string;
|
|
365
382
|
api: "google-generative-ai";
|
|
@@ -376,12 +393,10 @@ export declare const MODELS: {
|
|
|
376
393
|
contextWindow: number;
|
|
377
394
|
maxTokens: number;
|
|
378
395
|
};
|
|
379
|
-
|
|
380
|
-
readonly openai: {
|
|
381
|
-
readonly "gpt-5-nano": {
|
|
396
|
+
readonly "gemini-2.5-pro-preview-05-06": {
|
|
382
397
|
id: string;
|
|
383
398
|
name: string;
|
|
384
|
-
api: "
|
|
399
|
+
api: "google-generative-ai";
|
|
385
400
|
provider: string;
|
|
386
401
|
baseUrl: string;
|
|
387
402
|
reasoning: true;
|
|
@@ -395,13 +410,13 @@ export declare const MODELS: {
|
|
|
395
410
|
contextWindow: number;
|
|
396
411
|
maxTokens: number;
|
|
397
412
|
};
|
|
398
|
-
readonly "
|
|
413
|
+
readonly "gemini-2.0-flash-lite": {
|
|
399
414
|
id: string;
|
|
400
415
|
name: string;
|
|
401
|
-
api: "
|
|
416
|
+
api: "google-generative-ai";
|
|
402
417
|
provider: string;
|
|
403
418
|
baseUrl: string;
|
|
404
|
-
reasoning:
|
|
419
|
+
reasoning: false;
|
|
405
420
|
input: ("text" | "image")[];
|
|
406
421
|
cost: {
|
|
407
422
|
input: number;
|
|
@@ -412,10 +427,10 @@ export declare const MODELS: {
|
|
|
412
427
|
contextWindow: number;
|
|
413
428
|
maxTokens: number;
|
|
414
429
|
};
|
|
415
|
-
readonly "
|
|
430
|
+
readonly "gemini-live-2.5-flash-preview-native-audio": {
|
|
416
431
|
id: string;
|
|
417
432
|
name: string;
|
|
418
|
-
api: "
|
|
433
|
+
api: "google-generative-ai";
|
|
419
434
|
provider: string;
|
|
420
435
|
baseUrl: string;
|
|
421
436
|
reasoning: true;
|
|
@@ -429,10 +444,10 @@ export declare const MODELS: {
|
|
|
429
444
|
contextWindow: number;
|
|
430
445
|
maxTokens: number;
|
|
431
446
|
};
|
|
432
|
-
readonly "
|
|
447
|
+
readonly "gemini-2.0-flash": {
|
|
433
448
|
id: string;
|
|
434
449
|
name: string;
|
|
435
|
-
api: "
|
|
450
|
+
api: "google-generative-ai";
|
|
436
451
|
provider: string;
|
|
437
452
|
baseUrl: string;
|
|
438
453
|
reasoning: false;
|
|
@@ -446,13 +461,13 @@ export declare const MODELS: {
|
|
|
446
461
|
contextWindow: number;
|
|
447
462
|
maxTokens: number;
|
|
448
463
|
};
|
|
449
|
-
readonly "
|
|
464
|
+
readonly "gemini-2.5-flash-lite": {
|
|
450
465
|
id: string;
|
|
451
466
|
name: string;
|
|
452
|
-
api: "
|
|
467
|
+
api: "google-generative-ai";
|
|
453
468
|
provider: string;
|
|
454
469
|
baseUrl: string;
|
|
455
|
-
reasoning:
|
|
470
|
+
reasoning: true;
|
|
456
471
|
input: ("text" | "image")[];
|
|
457
472
|
cost: {
|
|
458
473
|
input: number;
|
|
@@ -463,10 +478,10 @@ export declare const MODELS: {
|
|
|
463
478
|
contextWindow: number;
|
|
464
479
|
maxTokens: number;
|
|
465
480
|
};
|
|
466
|
-
readonly
|
|
481
|
+
readonly "gemini-2.5-pro-preview-06-05": {
|
|
467
482
|
id: string;
|
|
468
483
|
name: string;
|
|
469
|
-
api: "
|
|
484
|
+
api: "google-generative-ai";
|
|
470
485
|
provider: string;
|
|
471
486
|
baseUrl: string;
|
|
472
487
|
reasoning: true;
|
|
@@ -480,10 +495,10 @@ export declare const MODELS: {
|
|
|
480
495
|
contextWindow: number;
|
|
481
496
|
maxTokens: number;
|
|
482
497
|
};
|
|
483
|
-
readonly "
|
|
498
|
+
readonly "gemini-live-2.5-flash": {
|
|
484
499
|
id: string;
|
|
485
500
|
name: string;
|
|
486
|
-
api: "
|
|
501
|
+
api: "google-generative-ai";
|
|
487
502
|
provider: string;
|
|
488
503
|
baseUrl: string;
|
|
489
504
|
reasoning: true;
|
|
@@ -497,10 +512,10 @@ export declare const MODELS: {
|
|
|
497
512
|
contextWindow: number;
|
|
498
513
|
maxTokens: number;
|
|
499
514
|
};
|
|
500
|
-
readonly "
|
|
515
|
+
readonly "gemini-2.5-flash-lite-preview-06-17": {
|
|
501
516
|
id: string;
|
|
502
517
|
name: string;
|
|
503
|
-
api: "
|
|
518
|
+
api: "google-generative-ai";
|
|
504
519
|
provider: string;
|
|
505
520
|
baseUrl: string;
|
|
506
521
|
reasoning: true;
|
|
@@ -514,10 +529,10 @@ export declare const MODELS: {
|
|
|
514
529
|
contextWindow: number;
|
|
515
530
|
maxTokens: number;
|
|
516
531
|
};
|
|
517
|
-
readonly "
|
|
532
|
+
readonly "gemini-2.5-flash-preview-09-2025": {
|
|
518
533
|
id: string;
|
|
519
534
|
name: string;
|
|
520
|
-
api: "
|
|
535
|
+
api: "google-generative-ai";
|
|
521
536
|
provider: string;
|
|
522
537
|
baseUrl: string;
|
|
523
538
|
reasoning: true;
|
|
@@ -531,10 +546,10 @@ export declare const MODELS: {
|
|
|
531
546
|
contextWindow: number;
|
|
532
547
|
maxTokens: number;
|
|
533
548
|
};
|
|
534
|
-
readonly
|
|
549
|
+
readonly "gemini-2.5-flash-preview-04-17": {
|
|
535
550
|
id: string;
|
|
536
551
|
name: string;
|
|
537
|
-
api: "
|
|
552
|
+
api: "google-generative-ai";
|
|
538
553
|
provider: string;
|
|
539
554
|
baseUrl: string;
|
|
540
555
|
reasoning: true;
|
|
@@ -548,10 +563,10 @@ export declare const MODELS: {
|
|
|
548
563
|
contextWindow: number;
|
|
549
564
|
maxTokens: number;
|
|
550
565
|
};
|
|
551
|
-
readonly "
|
|
566
|
+
readonly "gemini-2.5-pro": {
|
|
552
567
|
id: string;
|
|
553
568
|
name: string;
|
|
554
|
-
api: "
|
|
569
|
+
api: "google-generative-ai";
|
|
555
570
|
provider: string;
|
|
556
571
|
baseUrl: string;
|
|
557
572
|
reasoning: true;
|
|
@@ -565,13 +580,13 @@ export declare const MODELS: {
|
|
|
565
580
|
contextWindow: number;
|
|
566
581
|
maxTokens: number;
|
|
567
582
|
};
|
|
568
|
-
readonly "
|
|
583
|
+
readonly "gemini-1.5-flash": {
|
|
569
584
|
id: string;
|
|
570
585
|
name: string;
|
|
571
|
-
api: "
|
|
586
|
+
api: "google-generative-ai";
|
|
572
587
|
provider: string;
|
|
573
588
|
baseUrl: string;
|
|
574
|
-
reasoning:
|
|
589
|
+
reasoning: false;
|
|
575
590
|
input: ("text" | "image")[];
|
|
576
591
|
cost: {
|
|
577
592
|
input: number;
|
|
@@ -582,10 +597,10 @@ export declare const MODELS: {
|
|
|
582
597
|
contextWindow: number;
|
|
583
598
|
maxTokens: number;
|
|
584
599
|
};
|
|
585
|
-
readonly "
|
|
600
|
+
readonly "gemini-1.5-flash-8b": {
|
|
586
601
|
id: string;
|
|
587
602
|
name: string;
|
|
588
|
-
api: "
|
|
603
|
+
api: "google-generative-ai";
|
|
589
604
|
provider: string;
|
|
590
605
|
baseUrl: string;
|
|
591
606
|
reasoning: false;
|
|
@@ -599,13 +614,13 @@ export declare const MODELS: {
|
|
|
599
614
|
contextWindow: number;
|
|
600
615
|
maxTokens: number;
|
|
601
616
|
};
|
|
602
|
-
readonly "
|
|
617
|
+
readonly "gemini-2.5-flash-lite-preview-09-2025": {
|
|
603
618
|
id: string;
|
|
604
619
|
name: string;
|
|
605
|
-
api: "
|
|
620
|
+
api: "google-generative-ai";
|
|
606
621
|
provider: string;
|
|
607
622
|
baseUrl: string;
|
|
608
|
-
reasoning:
|
|
623
|
+
reasoning: true;
|
|
609
624
|
input: ("text" | "image")[];
|
|
610
625
|
cost: {
|
|
611
626
|
input: number;
|
|
@@ -616,10 +631,10 @@ export declare const MODELS: {
|
|
|
616
631
|
contextWindow: number;
|
|
617
632
|
maxTokens: number;
|
|
618
633
|
};
|
|
619
|
-
readonly "
|
|
634
|
+
readonly "gemini-1.5-pro": {
|
|
620
635
|
id: string;
|
|
621
636
|
name: string;
|
|
622
|
-
api: "
|
|
637
|
+
api: "google-generative-ai";
|
|
623
638
|
provider: string;
|
|
624
639
|
baseUrl: string;
|
|
625
640
|
reasoning: false;
|
|
@@ -633,7 +648,9 @@ export declare const MODELS: {
|
|
|
633
648
|
contextWindow: number;
|
|
634
649
|
maxTokens: number;
|
|
635
650
|
};
|
|
636
|
-
|
|
651
|
+
};
|
|
652
|
+
readonly openai: {
|
|
653
|
+
readonly "gpt-4.1-nano": {
|
|
637
654
|
id: string;
|
|
638
655
|
name: string;
|
|
639
656
|
api: "openai-responses";
|
|
@@ -667,7 +684,7 @@ export declare const MODELS: {
|
|
|
667
684
|
contextWindow: number;
|
|
668
685
|
maxTokens: number;
|
|
669
686
|
};
|
|
670
|
-
readonly "
|
|
687
|
+
readonly "o1-pro": {
|
|
671
688
|
id: string;
|
|
672
689
|
name: string;
|
|
673
690
|
api: "openai-responses";
|
|
@@ -684,14 +701,14 @@ export declare const MODELS: {
|
|
|
684
701
|
contextWindow: number;
|
|
685
702
|
maxTokens: number;
|
|
686
703
|
};
|
|
687
|
-
readonly "
|
|
704
|
+
readonly "gpt-4o-2024-05-13": {
|
|
688
705
|
id: string;
|
|
689
706
|
name: string;
|
|
690
707
|
api: "openai-responses";
|
|
691
708
|
provider: string;
|
|
692
709
|
baseUrl: string;
|
|
693
|
-
reasoning:
|
|
694
|
-
input: "text"[];
|
|
710
|
+
reasoning: false;
|
|
711
|
+
input: ("text" | "image")[];
|
|
695
712
|
cost: {
|
|
696
713
|
input: number;
|
|
697
714
|
output: number;
|
|
@@ -701,7 +718,7 @@ export declare const MODELS: {
|
|
|
701
718
|
contextWindow: number;
|
|
702
719
|
maxTokens: number;
|
|
703
720
|
};
|
|
704
|
-
readonly "gpt-
|
|
721
|
+
readonly "gpt-4o-2024-08-06": {
|
|
705
722
|
id: string;
|
|
706
723
|
name: string;
|
|
707
724
|
api: "openai-responses";
|
|
@@ -718,16 +735,14 @@ export declare const MODELS: {
|
|
|
718
735
|
contextWindow: number;
|
|
719
736
|
maxTokens: number;
|
|
720
737
|
};
|
|
721
|
-
|
|
722
|
-
readonly groq: {
|
|
723
|
-
readonly "llama-3.1-8b-instant": {
|
|
738
|
+
readonly "gpt-4.1-mini": {
|
|
724
739
|
id: string;
|
|
725
740
|
name: string;
|
|
726
|
-
api: "openai-
|
|
741
|
+
api: "openai-responses";
|
|
727
742
|
provider: string;
|
|
728
743
|
baseUrl: string;
|
|
729
744
|
reasoning: false;
|
|
730
|
-
input: "text"[];
|
|
745
|
+
input: ("text" | "image")[];
|
|
731
746
|
cost: {
|
|
732
747
|
input: number;
|
|
733
748
|
output: number;
|
|
@@ -737,14 +752,14 @@ export declare const MODELS: {
|
|
|
737
752
|
contextWindow: number;
|
|
738
753
|
maxTokens: number;
|
|
739
754
|
};
|
|
740
|
-
readonly "
|
|
755
|
+
readonly "o3-deep-research": {
|
|
741
756
|
id: string;
|
|
742
757
|
name: string;
|
|
743
|
-
api: "openai-
|
|
758
|
+
api: "openai-responses";
|
|
744
759
|
provider: string;
|
|
745
760
|
baseUrl: string;
|
|
746
761
|
reasoning: true;
|
|
747
|
-
input: "text"[];
|
|
762
|
+
input: ("text" | "image")[];
|
|
748
763
|
cost: {
|
|
749
764
|
input: number;
|
|
750
765
|
output: number;
|
|
@@ -754,14 +769,14 @@ export declare const MODELS: {
|
|
|
754
769
|
contextWindow: number;
|
|
755
770
|
maxTokens: number;
|
|
756
771
|
};
|
|
757
|
-
readonly "
|
|
772
|
+
readonly "gpt-4-turbo": {
|
|
758
773
|
id: string;
|
|
759
774
|
name: string;
|
|
760
|
-
api: "openai-
|
|
775
|
+
api: "openai-responses";
|
|
761
776
|
provider: string;
|
|
762
777
|
baseUrl: string;
|
|
763
778
|
reasoning: false;
|
|
764
|
-
input: "text"[];
|
|
779
|
+
input: ("text" | "image")[];
|
|
765
780
|
cost: {
|
|
766
781
|
input: number;
|
|
767
782
|
output: number;
|
|
@@ -771,10 +786,10 @@ export declare const MODELS: {
|
|
|
771
786
|
contextWindow: number;
|
|
772
787
|
maxTokens: number;
|
|
773
788
|
};
|
|
774
|
-
readonly "
|
|
789
|
+
readonly "o3-mini": {
|
|
775
790
|
id: string;
|
|
776
791
|
name: string;
|
|
777
|
-
api: "openai-
|
|
792
|
+
api: "openai-responses";
|
|
778
793
|
provider: string;
|
|
779
794
|
baseUrl: string;
|
|
780
795
|
reasoning: true;
|
|
@@ -788,13 +803,13 @@ export declare const MODELS: {
|
|
|
788
803
|
contextWindow: number;
|
|
789
804
|
maxTokens: number;
|
|
790
805
|
};
|
|
791
|
-
readonly "
|
|
806
|
+
readonly "codex-mini-latest": {
|
|
792
807
|
id: string;
|
|
793
808
|
name: string;
|
|
794
|
-
api: "openai-
|
|
809
|
+
api: "openai-responses";
|
|
795
810
|
provider: string;
|
|
796
811
|
baseUrl: string;
|
|
797
|
-
reasoning:
|
|
812
|
+
reasoning: true;
|
|
798
813
|
input: "text"[];
|
|
799
814
|
cost: {
|
|
800
815
|
input: number;
|
|
@@ -805,14 +820,14 @@ export declare const MODELS: {
|
|
|
805
820
|
contextWindow: number;
|
|
806
821
|
maxTokens: number;
|
|
807
822
|
};
|
|
808
|
-
readonly "
|
|
823
|
+
readonly "gpt-5-nano": {
|
|
809
824
|
id: string;
|
|
810
825
|
name: string;
|
|
811
|
-
api: "openai-
|
|
826
|
+
api: "openai-responses";
|
|
812
827
|
provider: string;
|
|
813
828
|
baseUrl: string;
|
|
814
|
-
reasoning:
|
|
815
|
-
input: "text"[];
|
|
829
|
+
reasoning: true;
|
|
830
|
+
input: ("text" | "image")[];
|
|
816
831
|
cost: {
|
|
817
832
|
input: number;
|
|
818
833
|
output: number;
|
|
@@ -822,14 +837,14 @@ export declare const MODELS: {
|
|
|
822
837
|
contextWindow: number;
|
|
823
838
|
maxTokens: number;
|
|
824
839
|
};
|
|
825
|
-
readonly "
|
|
840
|
+
readonly "gpt-5-codex": {
|
|
826
841
|
id: string;
|
|
827
842
|
name: string;
|
|
828
|
-
api: "openai-
|
|
843
|
+
api: "openai-responses";
|
|
829
844
|
provider: string;
|
|
830
845
|
baseUrl: string;
|
|
831
|
-
reasoning:
|
|
832
|
-
input: "text"[];
|
|
846
|
+
reasoning: true;
|
|
847
|
+
input: ("text" | "image")[];
|
|
833
848
|
cost: {
|
|
834
849
|
input: number;
|
|
835
850
|
output: number;
|
|
@@ -839,14 +854,14 @@ export declare const MODELS: {
|
|
|
839
854
|
contextWindow: number;
|
|
840
855
|
maxTokens: number;
|
|
841
856
|
};
|
|
842
|
-
readonly "
|
|
857
|
+
readonly "gpt-4o": {
|
|
843
858
|
id: string;
|
|
844
859
|
name: string;
|
|
845
|
-
api: "openai-
|
|
860
|
+
api: "openai-responses";
|
|
846
861
|
provider: string;
|
|
847
862
|
baseUrl: string;
|
|
848
863
|
reasoning: false;
|
|
849
|
-
input: "text"[];
|
|
864
|
+
input: ("text" | "image")[];
|
|
850
865
|
cost: {
|
|
851
866
|
input: number;
|
|
852
867
|
output: number;
|
|
@@ -856,14 +871,14 @@ export declare const MODELS: {
|
|
|
856
871
|
contextWindow: number;
|
|
857
872
|
maxTokens: number;
|
|
858
873
|
};
|
|
859
|
-
readonly "
|
|
874
|
+
readonly "gpt-4.1": {
|
|
860
875
|
id: string;
|
|
861
876
|
name: string;
|
|
862
|
-
api: "openai-
|
|
877
|
+
api: "openai-responses";
|
|
863
878
|
provider: string;
|
|
864
879
|
baseUrl: string;
|
|
865
|
-
reasoning:
|
|
866
|
-
input: "text"[];
|
|
880
|
+
reasoning: false;
|
|
881
|
+
input: ("text" | "image")[];
|
|
867
882
|
cost: {
|
|
868
883
|
input: number;
|
|
869
884
|
output: number;
|
|
@@ -873,14 +888,14 @@ export declare const MODELS: {
|
|
|
873
888
|
contextWindow: number;
|
|
874
889
|
maxTokens: number;
|
|
875
890
|
};
|
|
876
|
-
readonly "
|
|
891
|
+
readonly "o4-mini": {
|
|
877
892
|
id: string;
|
|
878
893
|
name: string;
|
|
879
|
-
api: "openai-
|
|
894
|
+
api: "openai-responses";
|
|
880
895
|
provider: string;
|
|
881
896
|
baseUrl: string;
|
|
882
897
|
reasoning: true;
|
|
883
|
-
input: "text"[];
|
|
898
|
+
input: ("text" | "image")[];
|
|
884
899
|
cost: {
|
|
885
900
|
input: number;
|
|
886
901
|
output: number;
|
|
@@ -890,13 +905,13 @@ export declare const MODELS: {
|
|
|
890
905
|
contextWindow: number;
|
|
891
906
|
maxTokens: number;
|
|
892
907
|
};
|
|
893
|
-
readonly
|
|
908
|
+
readonly o1: {
|
|
894
909
|
id: string;
|
|
895
910
|
name: string;
|
|
896
|
-
api: "openai-
|
|
911
|
+
api: "openai-responses";
|
|
897
912
|
provider: string;
|
|
898
913
|
baseUrl: string;
|
|
899
|
-
reasoning:
|
|
914
|
+
reasoning: true;
|
|
900
915
|
input: ("text" | "image")[];
|
|
901
916
|
cost: {
|
|
902
917
|
input: number;
|
|
@@ -907,13 +922,13 @@ export declare const MODELS: {
|
|
|
907
922
|
contextWindow: number;
|
|
908
923
|
maxTokens: number;
|
|
909
924
|
};
|
|
910
|
-
readonly "
|
|
925
|
+
readonly "gpt-5-mini": {
|
|
911
926
|
id: string;
|
|
912
927
|
name: string;
|
|
913
|
-
api: "openai-
|
|
928
|
+
api: "openai-responses";
|
|
914
929
|
provider: string;
|
|
915
930
|
baseUrl: string;
|
|
916
|
-
reasoning:
|
|
931
|
+
reasoning: true;
|
|
917
932
|
input: ("text" | "image")[];
|
|
918
933
|
cost: {
|
|
919
934
|
input: number;
|
|
@@ -924,14 +939,14 @@ export declare const MODELS: {
|
|
|
924
939
|
contextWindow: number;
|
|
925
940
|
maxTokens: number;
|
|
926
941
|
};
|
|
927
|
-
readonly "
|
|
942
|
+
readonly "o3-pro": {
|
|
928
943
|
id: string;
|
|
929
944
|
name: string;
|
|
930
|
-
api: "openai-
|
|
945
|
+
api: "openai-responses";
|
|
931
946
|
provider: string;
|
|
932
947
|
baseUrl: string;
|
|
933
948
|
reasoning: true;
|
|
934
|
-
input: "text"[];
|
|
949
|
+
input: ("text" | "image")[];
|
|
935
950
|
cost: {
|
|
936
951
|
input: number;
|
|
937
952
|
output: number;
|
|
@@ -941,14 +956,14 @@ export declare const MODELS: {
|
|
|
941
956
|
contextWindow: number;
|
|
942
957
|
maxTokens: number;
|
|
943
958
|
};
|
|
944
|
-
readonly "
|
|
959
|
+
readonly "gpt-4o-2024-11-20": {
|
|
945
960
|
id: string;
|
|
946
961
|
name: string;
|
|
947
|
-
api: "openai-
|
|
962
|
+
api: "openai-responses";
|
|
948
963
|
provider: string;
|
|
949
964
|
baseUrl: string;
|
|
950
965
|
reasoning: false;
|
|
951
|
-
input: "text"[];
|
|
966
|
+
input: ("text" | "image")[];
|
|
952
967
|
cost: {
|
|
953
968
|
input: number;
|
|
954
969
|
output: number;
|
|
@@ -958,15 +973,393 @@ export declare const MODELS: {
|
|
|
958
973
|
contextWindow: number;
|
|
959
974
|
maxTokens: number;
|
|
960
975
|
};
|
|
961
|
-
readonly
|
|
976
|
+
readonly o3: {
|
|
962
977
|
id: string;
|
|
963
978
|
name: string;
|
|
964
|
-
api: "openai-
|
|
979
|
+
api: "openai-responses";
|
|
965
980
|
provider: string;
|
|
966
981
|
baseUrl: string;
|
|
967
|
-
reasoning:
|
|
968
|
-
input: "text"[];
|
|
969
|
-
cost: {
|
|
982
|
+
reasoning: true;
|
|
983
|
+
input: ("text" | "image")[];
|
|
984
|
+
cost: {
|
|
985
|
+
input: number;
|
|
986
|
+
output: number;
|
|
987
|
+
cacheRead: number;
|
|
988
|
+
cacheWrite: number;
|
|
989
|
+
};
|
|
990
|
+
contextWindow: number;
|
|
991
|
+
maxTokens: number;
|
|
992
|
+
};
|
|
993
|
+
readonly "o4-mini-deep-research": {
|
|
994
|
+
id: string;
|
|
995
|
+
name: string;
|
|
996
|
+
api: "openai-responses";
|
|
997
|
+
provider: string;
|
|
998
|
+
baseUrl: string;
|
|
999
|
+
reasoning: true;
|
|
1000
|
+
input: ("text" | "image")[];
|
|
1001
|
+
cost: {
|
|
1002
|
+
input: number;
|
|
1003
|
+
output: number;
|
|
1004
|
+
cacheRead: number;
|
|
1005
|
+
cacheWrite: number;
|
|
1006
|
+
};
|
|
1007
|
+
contextWindow: number;
|
|
1008
|
+
maxTokens: number;
|
|
1009
|
+
};
|
|
1010
|
+
readonly "gpt-4o-mini": {
|
|
1011
|
+
id: string;
|
|
1012
|
+
name: string;
|
|
1013
|
+
api: "openai-responses";
|
|
1014
|
+
provider: string;
|
|
1015
|
+
baseUrl: string;
|
|
1016
|
+
reasoning: false;
|
|
1017
|
+
input: ("text" | "image")[];
|
|
1018
|
+
cost: {
|
|
1019
|
+
input: number;
|
|
1020
|
+
output: number;
|
|
1021
|
+
cacheRead: number;
|
|
1022
|
+
cacheWrite: number;
|
|
1023
|
+
};
|
|
1024
|
+
contextWindow: number;
|
|
1025
|
+
maxTokens: number;
|
|
1026
|
+
};
|
|
1027
|
+
readonly "gpt-5": {
|
|
1028
|
+
id: string;
|
|
1029
|
+
name: string;
|
|
1030
|
+
api: "openai-responses";
|
|
1031
|
+
provider: string;
|
|
1032
|
+
baseUrl: string;
|
|
1033
|
+
reasoning: true;
|
|
1034
|
+
input: ("text" | "image")[];
|
|
1035
|
+
cost: {
|
|
1036
|
+
input: number;
|
|
1037
|
+
output: number;
|
|
1038
|
+
cacheRead: number;
|
|
1039
|
+
cacheWrite: number;
|
|
1040
|
+
};
|
|
1041
|
+
contextWindow: number;
|
|
1042
|
+
maxTokens: number;
|
|
1043
|
+
};
|
|
1044
|
+
readonly "gpt-5-chat-latest": {
|
|
1045
|
+
id: string;
|
|
1046
|
+
name: string;
|
|
1047
|
+
api: "openai-responses";
|
|
1048
|
+
provider: string;
|
|
1049
|
+
baseUrl: string;
|
|
1050
|
+
reasoning: false;
|
|
1051
|
+
input: ("text" | "image")[];
|
|
1052
|
+
cost: {
|
|
1053
|
+
input: number;
|
|
1054
|
+
output: number;
|
|
1055
|
+
cacheRead: number;
|
|
1056
|
+
cacheWrite: number;
|
|
1057
|
+
};
|
|
1058
|
+
contextWindow: number;
|
|
1059
|
+
maxTokens: number;
|
|
1060
|
+
};
|
|
1061
|
+
};
|
|
1062
|
+
readonly groq: {
|
|
1063
|
+
readonly "llama-3.1-8b-instant": {
|
|
1064
|
+
id: string;
|
|
1065
|
+
name: string;
|
|
1066
|
+
api: "openai-completions";
|
|
1067
|
+
provider: string;
|
|
1068
|
+
baseUrl: string;
|
|
1069
|
+
reasoning: false;
|
|
1070
|
+
input: "text"[];
|
|
1071
|
+
cost: {
|
|
1072
|
+
input: number;
|
|
1073
|
+
output: number;
|
|
1074
|
+
cacheRead: number;
|
|
1075
|
+
cacheWrite: number;
|
|
1076
|
+
};
|
|
1077
|
+
contextWindow: number;
|
|
1078
|
+
maxTokens: number;
|
|
1079
|
+
};
|
|
1080
|
+
readonly "mistral-saba-24b": {
|
|
1081
|
+
id: string;
|
|
1082
|
+
name: string;
|
|
1083
|
+
api: "openai-completions";
|
|
1084
|
+
provider: string;
|
|
1085
|
+
baseUrl: string;
|
|
1086
|
+
reasoning: false;
|
|
1087
|
+
input: "text"[];
|
|
1088
|
+
cost: {
|
|
1089
|
+
input: number;
|
|
1090
|
+
output: number;
|
|
1091
|
+
cacheRead: number;
|
|
1092
|
+
cacheWrite: number;
|
|
1093
|
+
};
|
|
1094
|
+
contextWindow: number;
|
|
1095
|
+
maxTokens: number;
|
|
1096
|
+
};
|
|
1097
|
+
readonly "llama3-8b-8192": {
|
|
1098
|
+
id: string;
|
|
1099
|
+
name: string;
|
|
1100
|
+
api: "openai-completions";
|
|
1101
|
+
provider: string;
|
|
1102
|
+
baseUrl: string;
|
|
1103
|
+
reasoning: false;
|
|
1104
|
+
input: "text"[];
|
|
1105
|
+
cost: {
|
|
1106
|
+
input: number;
|
|
1107
|
+
output: number;
|
|
1108
|
+
cacheRead: number;
|
|
1109
|
+
cacheWrite: number;
|
|
1110
|
+
};
|
|
1111
|
+
contextWindow: number;
|
|
1112
|
+
maxTokens: number;
|
|
1113
|
+
};
|
|
1114
|
+
readonly "qwen-qwq-32b": {
|
|
1115
|
+
id: string;
|
|
1116
|
+
name: string;
|
|
1117
|
+
api: "openai-completions";
|
|
1118
|
+
provider: string;
|
|
1119
|
+
baseUrl: string;
|
|
1120
|
+
reasoning: true;
|
|
1121
|
+
input: "text"[];
|
|
1122
|
+
cost: {
|
|
1123
|
+
input: number;
|
|
1124
|
+
output: number;
|
|
1125
|
+
cacheRead: number;
|
|
1126
|
+
cacheWrite: number;
|
|
1127
|
+
};
|
|
1128
|
+
contextWindow: number;
|
|
1129
|
+
maxTokens: number;
|
|
1130
|
+
};
|
|
1131
|
+
readonly "llama3-70b-8192": {
|
|
1132
|
+
id: string;
|
|
1133
|
+
name: string;
|
|
1134
|
+
api: "openai-completions";
|
|
1135
|
+
provider: string;
|
|
1136
|
+
baseUrl: string;
|
|
1137
|
+
reasoning: false;
|
|
1138
|
+
input: "text"[];
|
|
1139
|
+
cost: {
|
|
1140
|
+
input: number;
|
|
1141
|
+
output: number;
|
|
1142
|
+
cacheRead: number;
|
|
1143
|
+
cacheWrite: number;
|
|
1144
|
+
};
|
|
1145
|
+
contextWindow: number;
|
|
1146
|
+
maxTokens: number;
|
|
1147
|
+
};
|
|
1148
|
+
readonly "deepseek-r1-distill-llama-70b": {
|
|
1149
|
+
id: string;
|
|
1150
|
+
name: string;
|
|
1151
|
+
api: "openai-completions";
|
|
1152
|
+
provider: string;
|
|
1153
|
+
baseUrl: string;
|
|
1154
|
+
reasoning: true;
|
|
1155
|
+
input: "text"[];
|
|
1156
|
+
cost: {
|
|
1157
|
+
input: number;
|
|
1158
|
+
output: number;
|
|
1159
|
+
cacheRead: number;
|
|
1160
|
+
cacheWrite: number;
|
|
1161
|
+
};
|
|
1162
|
+
contextWindow: number;
|
|
1163
|
+
maxTokens: number;
|
|
1164
|
+
};
|
|
1165
|
+
readonly "gemma2-9b-it": {
|
|
1166
|
+
id: string;
|
|
1167
|
+
name: string;
|
|
1168
|
+
api: "openai-completions";
|
|
1169
|
+
provider: string;
|
|
1170
|
+
baseUrl: string;
|
|
1171
|
+
reasoning: false;
|
|
1172
|
+
input: "text"[];
|
|
1173
|
+
cost: {
|
|
1174
|
+
input: number;
|
|
1175
|
+
output: number;
|
|
1176
|
+
cacheRead: number;
|
|
1177
|
+
cacheWrite: number;
|
|
1178
|
+
};
|
|
1179
|
+
contextWindow: number;
|
|
1180
|
+
maxTokens: number;
|
|
1181
|
+
};
|
|
1182
|
+
readonly "llama-3.3-70b-versatile": {
|
|
1183
|
+
id: string;
|
|
1184
|
+
name: string;
|
|
1185
|
+
api: "openai-completions";
|
|
1186
|
+
provider: string;
|
|
1187
|
+
baseUrl: string;
|
|
1188
|
+
reasoning: false;
|
|
1189
|
+
input: "text"[];
|
|
1190
|
+
cost: {
|
|
1191
|
+
input: number;
|
|
1192
|
+
output: number;
|
|
1193
|
+
cacheRead: number;
|
|
1194
|
+
cacheWrite: number;
|
|
1195
|
+
};
|
|
1196
|
+
contextWindow: number;
|
|
1197
|
+
maxTokens: number;
|
|
1198
|
+
};
|
|
1199
|
+
readonly "moonshotai/kimi-k2-instruct-0905": {
|
|
1200
|
+
id: string;
|
|
1201
|
+
name: string;
|
|
1202
|
+
api: "openai-completions";
|
|
1203
|
+
provider: string;
|
|
1204
|
+
baseUrl: string;
|
|
1205
|
+
reasoning: false;
|
|
1206
|
+
input: "text"[];
|
|
1207
|
+
cost: {
|
|
1208
|
+
input: number;
|
|
1209
|
+
output: number;
|
|
1210
|
+
cacheRead: number;
|
|
1211
|
+
cacheWrite: number;
|
|
1212
|
+
};
|
|
1213
|
+
contextWindow: number;
|
|
1214
|
+
maxTokens: number;
|
|
1215
|
+
};
|
|
1216
|
+
readonly "moonshotai/kimi-k2-instruct": {
|
|
1217
|
+
id: string;
|
|
1218
|
+
name: string;
|
|
1219
|
+
api: "openai-completions";
|
|
1220
|
+
provider: string;
|
|
1221
|
+
baseUrl: string;
|
|
1222
|
+
reasoning: false;
|
|
1223
|
+
input: "text"[];
|
|
1224
|
+
cost: {
|
|
1225
|
+
input: number;
|
|
1226
|
+
output: number;
|
|
1227
|
+
cacheRead: number;
|
|
1228
|
+
cacheWrite: number;
|
|
1229
|
+
};
|
|
1230
|
+
contextWindow: number;
|
|
1231
|
+
maxTokens: number;
|
|
1232
|
+
};
|
|
1233
|
+
readonly "openai/gpt-oss-20b": {
|
|
1234
|
+
id: string;
|
|
1235
|
+
name: string;
|
|
1236
|
+
api: "openai-completions";
|
|
1237
|
+
provider: string;
|
|
1238
|
+
baseUrl: string;
|
|
1239
|
+
reasoning: true;
|
|
1240
|
+
input: "text"[];
|
|
1241
|
+
cost: {
|
|
1242
|
+
input: number;
|
|
1243
|
+
output: number;
|
|
1244
|
+
cacheRead: number;
|
|
1245
|
+
cacheWrite: number;
|
|
1246
|
+
};
|
|
1247
|
+
contextWindow: number;
|
|
1248
|
+
maxTokens: number;
|
|
1249
|
+
};
|
|
1250
|
+
readonly "openai/gpt-oss-120b": {
|
|
1251
|
+
id: string;
|
|
1252
|
+
name: string;
|
|
1253
|
+
api: "openai-completions";
|
|
1254
|
+
provider: string;
|
|
1255
|
+
baseUrl: string;
|
|
1256
|
+
reasoning: true;
|
|
1257
|
+
input: "text"[];
|
|
1258
|
+
cost: {
|
|
1259
|
+
input: number;
|
|
1260
|
+
output: number;
|
|
1261
|
+
cacheRead: number;
|
|
1262
|
+
cacheWrite: number;
|
|
1263
|
+
};
|
|
1264
|
+
contextWindow: number;
|
|
1265
|
+
maxTokens: number;
|
|
1266
|
+
};
|
|
1267
|
+
readonly "qwen/qwen3-32b": {
|
|
1268
|
+
id: string;
|
|
1269
|
+
name: string;
|
|
1270
|
+
api: "openai-completions";
|
|
1271
|
+
provider: string;
|
|
1272
|
+
baseUrl: string;
|
|
1273
|
+
reasoning: true;
|
|
1274
|
+
input: "text"[];
|
|
1275
|
+
cost: {
|
|
1276
|
+
input: number;
|
|
1277
|
+
output: number;
|
|
1278
|
+
cacheRead: number;
|
|
1279
|
+
cacheWrite: number;
|
|
1280
|
+
};
|
|
1281
|
+
contextWindow: number;
|
|
1282
|
+
maxTokens: number;
|
|
1283
|
+
};
|
|
1284
|
+
readonly "meta-llama/llama-4-scout-17b-16e-instruct": {
|
|
1285
|
+
id: string;
|
|
1286
|
+
name: string;
|
|
1287
|
+
api: "openai-completions";
|
|
1288
|
+
provider: string;
|
|
1289
|
+
baseUrl: string;
|
|
1290
|
+
reasoning: false;
|
|
1291
|
+
input: ("text" | "image")[];
|
|
1292
|
+
cost: {
|
|
1293
|
+
input: number;
|
|
1294
|
+
output: number;
|
|
1295
|
+
cacheRead: number;
|
|
1296
|
+
cacheWrite: number;
|
|
1297
|
+
};
|
|
1298
|
+
contextWindow: number;
|
|
1299
|
+
maxTokens: number;
|
|
1300
|
+
};
|
|
1301
|
+
readonly "meta-llama/llama-4-maverick-17b-128e-instruct": {
|
|
1302
|
+
id: string;
|
|
1303
|
+
name: string;
|
|
1304
|
+
api: "openai-completions";
|
|
1305
|
+
provider: string;
|
|
1306
|
+
baseUrl: string;
|
|
1307
|
+
reasoning: false;
|
|
1308
|
+
input: ("text" | "image")[];
|
|
1309
|
+
cost: {
|
|
1310
|
+
input: number;
|
|
1311
|
+
output: number;
|
|
1312
|
+
cacheRead: number;
|
|
1313
|
+
cacheWrite: number;
|
|
1314
|
+
};
|
|
1315
|
+
contextWindow: number;
|
|
1316
|
+
maxTokens: number;
|
|
1317
|
+
};
|
|
1318
|
+
};
|
|
1319
|
+
readonly cerebras: {
|
|
1320
|
+
readonly "qwen-3-235b-a22b-instruct-2507": {
|
|
1321
|
+
id: string;
|
|
1322
|
+
name: string;
|
|
1323
|
+
api: "openai-completions";
|
|
1324
|
+
provider: string;
|
|
1325
|
+
baseUrl: string;
|
|
1326
|
+
reasoning: false;
|
|
1327
|
+
input: "text"[];
|
|
1328
|
+
cost: {
|
|
1329
|
+
input: number;
|
|
1330
|
+
output: number;
|
|
1331
|
+
cacheRead: number;
|
|
1332
|
+
cacheWrite: number;
|
|
1333
|
+
};
|
|
1334
|
+
contextWindow: number;
|
|
1335
|
+
maxTokens: number;
|
|
1336
|
+
};
|
|
1337
|
+
readonly "qwen-3-coder-480b": {
|
|
1338
|
+
id: string;
|
|
1339
|
+
name: string;
|
|
1340
|
+
api: "openai-completions";
|
|
1341
|
+
provider: string;
|
|
1342
|
+
baseUrl: string;
|
|
1343
|
+
reasoning: false;
|
|
1344
|
+
input: "text"[];
|
|
1345
|
+
cost: {
|
|
1346
|
+
input: number;
|
|
1347
|
+
output: number;
|
|
1348
|
+
cacheRead: number;
|
|
1349
|
+
cacheWrite: number;
|
|
1350
|
+
};
|
|
1351
|
+
contextWindow: number;
|
|
1352
|
+
maxTokens: number;
|
|
1353
|
+
};
|
|
1354
|
+
readonly "gpt-oss-120b": {
|
|
1355
|
+
id: string;
|
|
1356
|
+
name: string;
|
|
1357
|
+
api: "openai-completions";
|
|
1358
|
+
provider: string;
|
|
1359
|
+
baseUrl: string;
|
|
1360
|
+
reasoning: true;
|
|
1361
|
+
input: "text"[];
|
|
1362
|
+
cost: {
|
|
970
1363
|
input: number;
|
|
971
1364
|
output: number;
|
|
972
1365
|
cacheRead: number;
|
|
@@ -976,8 +1369,25 @@ export declare const MODELS: {
|
|
|
976
1369
|
maxTokens: number;
|
|
977
1370
|
};
|
|
978
1371
|
};
|
|
979
|
-
readonly
|
|
980
|
-
readonly "
|
|
1372
|
+
readonly xai: {
|
|
1373
|
+
readonly "grok-4-fast-non-reasoning": {
|
|
1374
|
+
id: string;
|
|
1375
|
+
name: string;
|
|
1376
|
+
api: "openai-completions";
|
|
1377
|
+
provider: string;
|
|
1378
|
+
baseUrl: string;
|
|
1379
|
+
reasoning: false;
|
|
1380
|
+
input: ("text" | "image")[];
|
|
1381
|
+
cost: {
|
|
1382
|
+
input: number;
|
|
1383
|
+
output: number;
|
|
1384
|
+
cacheRead: number;
|
|
1385
|
+
cacheWrite: number;
|
|
1386
|
+
};
|
|
1387
|
+
contextWindow: number;
|
|
1388
|
+
maxTokens: number;
|
|
1389
|
+
};
|
|
1390
|
+
readonly "grok-3-fast": {
|
|
981
1391
|
id: string;
|
|
982
1392
|
name: string;
|
|
983
1393
|
api: "openai-completions";
|
|
@@ -994,7 +1404,7 @@ export declare const MODELS: {
|
|
|
994
1404
|
contextWindow: number;
|
|
995
1405
|
maxTokens: number;
|
|
996
1406
|
};
|
|
997
|
-
readonly "
|
|
1407
|
+
readonly "grok-4": {
|
|
998
1408
|
id: string;
|
|
999
1409
|
name: string;
|
|
1000
1410
|
api: "openai-completions";
|
|
@@ -1011,7 +1421,41 @@ export declare const MODELS: {
|
|
|
1011
1421
|
contextWindow: number;
|
|
1012
1422
|
maxTokens: number;
|
|
1013
1423
|
};
|
|
1014
|
-
readonly "
|
|
1424
|
+
readonly "grok-2-vision": {
|
|
1425
|
+
id: string;
|
|
1426
|
+
name: string;
|
|
1427
|
+
api: "openai-completions";
|
|
1428
|
+
provider: string;
|
|
1429
|
+
baseUrl: string;
|
|
1430
|
+
reasoning: false;
|
|
1431
|
+
input: ("text" | "image")[];
|
|
1432
|
+
cost: {
|
|
1433
|
+
input: number;
|
|
1434
|
+
output: number;
|
|
1435
|
+
cacheRead: number;
|
|
1436
|
+
cacheWrite: number;
|
|
1437
|
+
};
|
|
1438
|
+
contextWindow: number;
|
|
1439
|
+
maxTokens: number;
|
|
1440
|
+
};
|
|
1441
|
+
readonly "grok-code-fast-1": {
|
|
1442
|
+
id: string;
|
|
1443
|
+
name: string;
|
|
1444
|
+
api: "openai-completions";
|
|
1445
|
+
provider: string;
|
|
1446
|
+
baseUrl: string;
|
|
1447
|
+
reasoning: true;
|
|
1448
|
+
input: "text"[];
|
|
1449
|
+
cost: {
|
|
1450
|
+
input: number;
|
|
1451
|
+
output: number;
|
|
1452
|
+
cacheRead: number;
|
|
1453
|
+
cacheWrite: number;
|
|
1454
|
+
};
|
|
1455
|
+
contextWindow: number;
|
|
1456
|
+
maxTokens: number;
|
|
1457
|
+
};
|
|
1458
|
+
readonly "grok-2": {
|
|
1015
1459
|
id: string;
|
|
1016
1460
|
name: string;
|
|
1017
1461
|
api: "openai-completions";
|
|
@@ -1028,8 +1472,6 @@ export declare const MODELS: {
|
|
|
1028
1472
|
contextWindow: number;
|
|
1029
1473
|
maxTokens: number;
|
|
1030
1474
|
};
|
|
1031
|
-
};
|
|
1032
|
-
readonly xai: {
|
|
1033
1475
|
readonly "grok-3-mini-fast-latest": {
|
|
1034
1476
|
id: string;
|
|
1035
1477
|
name: string;
|
|
@@ -1047,13 +1489,64 @@ export declare const MODELS: {
|
|
|
1047
1489
|
contextWindow: number;
|
|
1048
1490
|
maxTokens: number;
|
|
1049
1491
|
};
|
|
1050
|
-
readonly "grok-
|
|
1492
|
+
readonly "grok-2-vision-1212": {
|
|
1493
|
+
id: string;
|
|
1494
|
+
name: string;
|
|
1495
|
+
api: "openai-completions";
|
|
1496
|
+
provider: string;
|
|
1497
|
+
baseUrl: string;
|
|
1498
|
+
reasoning: false;
|
|
1499
|
+
input: ("text" | "image")[];
|
|
1500
|
+
cost: {
|
|
1501
|
+
input: number;
|
|
1502
|
+
output: number;
|
|
1503
|
+
cacheRead: number;
|
|
1504
|
+
cacheWrite: number;
|
|
1505
|
+
};
|
|
1506
|
+
contextWindow: number;
|
|
1507
|
+
maxTokens: number;
|
|
1508
|
+
};
|
|
1509
|
+
readonly "grok-3": {
|
|
1510
|
+
id: string;
|
|
1511
|
+
name: string;
|
|
1512
|
+
api: "openai-completions";
|
|
1513
|
+
provider: string;
|
|
1514
|
+
baseUrl: string;
|
|
1515
|
+
reasoning: false;
|
|
1516
|
+
input: "text"[];
|
|
1517
|
+
cost: {
|
|
1518
|
+
input: number;
|
|
1519
|
+
output: number;
|
|
1520
|
+
cacheRead: number;
|
|
1521
|
+
cacheWrite: number;
|
|
1522
|
+
};
|
|
1523
|
+
contextWindow: number;
|
|
1524
|
+
maxTokens: number;
|
|
1525
|
+
};
|
|
1526
|
+
readonly "grok-4-fast": {
|
|
1051
1527
|
id: string;
|
|
1052
1528
|
name: string;
|
|
1053
1529
|
api: "openai-completions";
|
|
1054
1530
|
provider: string;
|
|
1055
1531
|
baseUrl: string;
|
|
1056
1532
|
reasoning: true;
|
|
1533
|
+
input: ("text" | "image")[];
|
|
1534
|
+
cost: {
|
|
1535
|
+
input: number;
|
|
1536
|
+
output: number;
|
|
1537
|
+
cacheRead: number;
|
|
1538
|
+
cacheWrite: number;
|
|
1539
|
+
};
|
|
1540
|
+
contextWindow: number;
|
|
1541
|
+
maxTokens: number;
|
|
1542
|
+
};
|
|
1543
|
+
readonly "grok-2-latest": {
|
|
1544
|
+
id: string;
|
|
1545
|
+
name: string;
|
|
1546
|
+
api: "openai-completions";
|
|
1547
|
+
provider: string;
|
|
1548
|
+
baseUrl: string;
|
|
1549
|
+
reasoning: false;
|
|
1057
1550
|
input: "text"[];
|
|
1058
1551
|
cost: {
|
|
1059
1552
|
input: number;
|
|
@@ -1064,7 +1557,7 @@ export declare const MODELS: {
|
|
|
1064
1557
|
contextWindow: number;
|
|
1065
1558
|
maxTokens: number;
|
|
1066
1559
|
};
|
|
1067
|
-
readonly "grok-
|
|
1560
|
+
readonly "grok-2-1212": {
|
|
1068
1561
|
id: string;
|
|
1069
1562
|
name: string;
|
|
1070
1563
|
api: "openai-completions";
|
|
@@ -1098,13 +1591,151 @@ export declare const MODELS: {
|
|
|
1098
1591
|
contextWindow: number;
|
|
1099
1592
|
maxTokens: number;
|
|
1100
1593
|
};
|
|
1101
|
-
readonly "grok-3": {
|
|
1594
|
+
readonly "grok-3-latest": {
|
|
1595
|
+
id: string;
|
|
1596
|
+
name: string;
|
|
1597
|
+
api: "openai-completions";
|
|
1598
|
+
provider: string;
|
|
1599
|
+
baseUrl: string;
|
|
1600
|
+
reasoning: false;
|
|
1601
|
+
input: "text"[];
|
|
1602
|
+
cost: {
|
|
1603
|
+
input: number;
|
|
1604
|
+
output: number;
|
|
1605
|
+
cacheRead: number;
|
|
1606
|
+
cacheWrite: number;
|
|
1607
|
+
};
|
|
1608
|
+
contextWindow: number;
|
|
1609
|
+
maxTokens: number;
|
|
1610
|
+
};
|
|
1611
|
+
readonly "grok-2-vision-latest": {
|
|
1612
|
+
id: string;
|
|
1613
|
+
name: string;
|
|
1614
|
+
api: "openai-completions";
|
|
1615
|
+
provider: string;
|
|
1616
|
+
baseUrl: string;
|
|
1617
|
+
reasoning: false;
|
|
1618
|
+
input: ("text" | "image")[];
|
|
1619
|
+
cost: {
|
|
1620
|
+
input: number;
|
|
1621
|
+
output: number;
|
|
1622
|
+
cacheRead: number;
|
|
1623
|
+
cacheWrite: number;
|
|
1624
|
+
};
|
|
1625
|
+
contextWindow: number;
|
|
1626
|
+
maxTokens: number;
|
|
1627
|
+
};
|
|
1628
|
+
readonly "grok-vision-beta": {
|
|
1629
|
+
id: string;
|
|
1630
|
+
name: string;
|
|
1631
|
+
api: "openai-completions";
|
|
1632
|
+
provider: string;
|
|
1633
|
+
baseUrl: string;
|
|
1634
|
+
reasoning: false;
|
|
1635
|
+
input: ("text" | "image")[];
|
|
1636
|
+
cost: {
|
|
1637
|
+
input: number;
|
|
1638
|
+
output: number;
|
|
1639
|
+
cacheRead: number;
|
|
1640
|
+
cacheWrite: number;
|
|
1641
|
+
};
|
|
1642
|
+
contextWindow: number;
|
|
1643
|
+
maxTokens: number;
|
|
1644
|
+
};
|
|
1645
|
+
readonly "grok-3-mini": {
|
|
1646
|
+
id: string;
|
|
1647
|
+
name: string;
|
|
1648
|
+
api: "openai-completions";
|
|
1649
|
+
provider: string;
|
|
1650
|
+
baseUrl: string;
|
|
1651
|
+
reasoning: true;
|
|
1652
|
+
input: "text"[];
|
|
1653
|
+
cost: {
|
|
1654
|
+
input: number;
|
|
1655
|
+
output: number;
|
|
1656
|
+
cacheRead: number;
|
|
1657
|
+
cacheWrite: number;
|
|
1658
|
+
};
|
|
1659
|
+
contextWindow: number;
|
|
1660
|
+
maxTokens: number;
|
|
1661
|
+
};
|
|
1662
|
+
readonly "grok-beta": {
|
|
1663
|
+
id: string;
|
|
1664
|
+
name: string;
|
|
1665
|
+
api: "openai-completions";
|
|
1666
|
+
provider: string;
|
|
1667
|
+
baseUrl: string;
|
|
1668
|
+
reasoning: false;
|
|
1669
|
+
input: "text"[];
|
|
1670
|
+
cost: {
|
|
1671
|
+
input: number;
|
|
1672
|
+
output: number;
|
|
1673
|
+
cacheRead: number;
|
|
1674
|
+
cacheWrite: number;
|
|
1675
|
+
};
|
|
1676
|
+
contextWindow: number;
|
|
1677
|
+
maxTokens: number;
|
|
1678
|
+
};
|
|
1679
|
+
readonly "grok-3-mini-latest": {
|
|
1680
|
+
id: string;
|
|
1681
|
+
name: string;
|
|
1682
|
+
api: "openai-completions";
|
|
1683
|
+
provider: string;
|
|
1684
|
+
baseUrl: string;
|
|
1685
|
+
reasoning: true;
|
|
1686
|
+
input: "text"[];
|
|
1687
|
+
cost: {
|
|
1688
|
+
input: number;
|
|
1689
|
+
output: number;
|
|
1690
|
+
cacheRead: number;
|
|
1691
|
+
cacheWrite: number;
|
|
1692
|
+
};
|
|
1693
|
+
contextWindow: number;
|
|
1694
|
+
maxTokens: number;
|
|
1695
|
+
};
|
|
1696
|
+
readonly "grok-3-mini-fast": {
|
|
1697
|
+
id: string;
|
|
1698
|
+
name: string;
|
|
1699
|
+
api: "openai-completions";
|
|
1700
|
+
provider: string;
|
|
1701
|
+
baseUrl: string;
|
|
1702
|
+
reasoning: true;
|
|
1703
|
+
input: "text"[];
|
|
1704
|
+
cost: {
|
|
1705
|
+
input: number;
|
|
1706
|
+
output: number;
|
|
1707
|
+
cacheRead: number;
|
|
1708
|
+
cacheWrite: number;
|
|
1709
|
+
};
|
|
1710
|
+
contextWindow: number;
|
|
1711
|
+
maxTokens: number;
|
|
1712
|
+
};
|
|
1713
|
+
};
|
|
1714
|
+
readonly zai: {
|
|
1715
|
+
readonly "glm-4.5-flash": {
|
|
1716
|
+
id: string;
|
|
1717
|
+
name: string;
|
|
1718
|
+
api: "anthropic-messages";
|
|
1719
|
+
provider: string;
|
|
1720
|
+
baseUrl: string;
|
|
1721
|
+
reasoning: true;
|
|
1722
|
+
input: "text"[];
|
|
1723
|
+
cost: {
|
|
1724
|
+
input: number;
|
|
1725
|
+
output: number;
|
|
1726
|
+
cacheRead: number;
|
|
1727
|
+
cacheWrite: number;
|
|
1728
|
+
};
|
|
1729
|
+
contextWindow: number;
|
|
1730
|
+
maxTokens: number;
|
|
1731
|
+
};
|
|
1732
|
+
readonly "glm-4.5": {
|
|
1102
1733
|
id: string;
|
|
1103
1734
|
name: string;
|
|
1104
|
-
api: "
|
|
1735
|
+
api: "anthropic-messages";
|
|
1105
1736
|
provider: string;
|
|
1106
1737
|
baseUrl: string;
|
|
1107
|
-
reasoning:
|
|
1738
|
+
reasoning: true;
|
|
1108
1739
|
input: "text"[];
|
|
1109
1740
|
cost: {
|
|
1110
1741
|
input: number;
|
|
@@ -1115,10 +1746,10 @@ export declare const MODELS: {
|
|
|
1115
1746
|
contextWindow: number;
|
|
1116
1747
|
maxTokens: number;
|
|
1117
1748
|
};
|
|
1118
|
-
readonly "
|
|
1749
|
+
readonly "glm-4.5-air": {
|
|
1119
1750
|
id: string;
|
|
1120
1751
|
name: string;
|
|
1121
|
-
api: "
|
|
1752
|
+
api: "anthropic-messages";
|
|
1122
1753
|
provider: string;
|
|
1123
1754
|
baseUrl: string;
|
|
1124
1755
|
reasoning: true;
|
|
@@ -1132,13 +1763,13 @@ export declare const MODELS: {
|
|
|
1132
1763
|
contextWindow: number;
|
|
1133
1764
|
maxTokens: number;
|
|
1134
1765
|
};
|
|
1135
|
-
readonly "
|
|
1766
|
+
readonly "glm-4.5v": {
|
|
1136
1767
|
id: string;
|
|
1137
1768
|
name: string;
|
|
1138
|
-
api: "
|
|
1769
|
+
api: "anthropic-messages";
|
|
1139
1770
|
provider: string;
|
|
1140
1771
|
baseUrl: string;
|
|
1141
|
-
reasoning:
|
|
1772
|
+
reasoning: true;
|
|
1142
1773
|
input: ("text" | "image")[];
|
|
1143
1774
|
cost: {
|
|
1144
1775
|
input: number;
|
|
@@ -1149,13 +1780,13 @@ export declare const MODELS: {
|
|
|
1149
1780
|
contextWindow: number;
|
|
1150
1781
|
maxTokens: number;
|
|
1151
1782
|
};
|
|
1152
|
-
readonly "
|
|
1783
|
+
readonly "glm-4.6": {
|
|
1153
1784
|
id: string;
|
|
1154
1785
|
name: string;
|
|
1155
|
-
api: "
|
|
1786
|
+
api: "anthropic-messages";
|
|
1156
1787
|
provider: string;
|
|
1157
1788
|
baseUrl: string;
|
|
1158
|
-
reasoning:
|
|
1789
|
+
reasoning: true;
|
|
1159
1790
|
input: "text"[];
|
|
1160
1791
|
cost: {
|
|
1161
1792
|
input: number;
|
|
@@ -1166,13 +1797,15 @@ export declare const MODELS: {
|
|
|
1166
1797
|
contextWindow: number;
|
|
1167
1798
|
maxTokens: number;
|
|
1168
1799
|
};
|
|
1169
|
-
|
|
1800
|
+
};
|
|
1801
|
+
readonly openrouter: {
|
|
1802
|
+
readonly "openrouter/andromeda-alpha": {
|
|
1170
1803
|
id: string;
|
|
1171
1804
|
name: string;
|
|
1172
1805
|
api: "openai-completions";
|
|
1173
1806
|
provider: string;
|
|
1174
1807
|
baseUrl: string;
|
|
1175
|
-
reasoning:
|
|
1808
|
+
reasoning: true;
|
|
1176
1809
|
input: ("text" | "image")[];
|
|
1177
1810
|
cost: {
|
|
1178
1811
|
input: number;
|
|
@@ -1183,13 +1816,13 @@ export declare const MODELS: {
|
|
|
1183
1816
|
contextWindow: number;
|
|
1184
1817
|
maxTokens: number;
|
|
1185
1818
|
};
|
|
1186
|
-
readonly "
|
|
1819
|
+
readonly "deepcogito/cogito-v2-preview-llama-405b": {
|
|
1187
1820
|
id: string;
|
|
1188
1821
|
name: string;
|
|
1189
1822
|
api: "openai-completions";
|
|
1190
1823
|
provider: string;
|
|
1191
1824
|
baseUrl: string;
|
|
1192
|
-
reasoning:
|
|
1825
|
+
reasoning: true;
|
|
1193
1826
|
input: "text"[];
|
|
1194
1827
|
cost: {
|
|
1195
1828
|
input: number;
|
|
@@ -1200,13 +1833,13 @@ export declare const MODELS: {
|
|
|
1200
1833
|
contextWindow: number;
|
|
1201
1834
|
maxTokens: number;
|
|
1202
1835
|
};
|
|
1203
|
-
readonly "
|
|
1836
|
+
readonly "qwen/qwen3-vl-8b-thinking": {
|
|
1204
1837
|
id: string;
|
|
1205
1838
|
name: string;
|
|
1206
1839
|
api: "openai-completions";
|
|
1207
1840
|
provider: string;
|
|
1208
1841
|
baseUrl: string;
|
|
1209
|
-
reasoning:
|
|
1842
|
+
reasoning: true;
|
|
1210
1843
|
input: ("text" | "image")[];
|
|
1211
1844
|
cost: {
|
|
1212
1845
|
input: number;
|
|
@@ -1217,14 +1850,14 @@ export declare const MODELS: {
|
|
|
1217
1850
|
contextWindow: number;
|
|
1218
1851
|
maxTokens: number;
|
|
1219
1852
|
};
|
|
1220
|
-
readonly "
|
|
1853
|
+
readonly "qwen/qwen3-vl-8b-instruct": {
|
|
1221
1854
|
id: string;
|
|
1222
1855
|
name: string;
|
|
1223
1856
|
api: "openai-completions";
|
|
1224
1857
|
provider: string;
|
|
1225
1858
|
baseUrl: string;
|
|
1226
1859
|
reasoning: false;
|
|
1227
|
-
input: "text"[];
|
|
1860
|
+
input: ("text" | "image")[];
|
|
1228
1861
|
cost: {
|
|
1229
1862
|
input: number;
|
|
1230
1863
|
output: number;
|
|
@@ -1234,13 +1867,13 @@ export declare const MODELS: {
|
|
|
1234
1867
|
contextWindow: number;
|
|
1235
1868
|
maxTokens: number;
|
|
1236
1869
|
};
|
|
1237
|
-
readonly "
|
|
1870
|
+
readonly "inclusionai/ring-1t": {
|
|
1238
1871
|
id: string;
|
|
1239
1872
|
name: string;
|
|
1240
1873
|
api: "openai-completions";
|
|
1241
1874
|
provider: string;
|
|
1242
1875
|
baseUrl: string;
|
|
1243
|
-
reasoning:
|
|
1876
|
+
reasoning: true;
|
|
1244
1877
|
input: "text"[];
|
|
1245
1878
|
cost: {
|
|
1246
1879
|
input: number;
|
|
@@ -1251,7 +1884,7 @@ export declare const MODELS: {
|
|
|
1251
1884
|
contextWindow: number;
|
|
1252
1885
|
maxTokens: number;
|
|
1253
1886
|
};
|
|
1254
|
-
readonly "
|
|
1887
|
+
readonly "inclusionai/ling-1t": {
|
|
1255
1888
|
id: string;
|
|
1256
1889
|
name: string;
|
|
1257
1890
|
api: "openai-completions";
|
|
@@ -1268,7 +1901,7 @@ export declare const MODELS: {
|
|
|
1268
1901
|
contextWindow: number;
|
|
1269
1902
|
maxTokens: number;
|
|
1270
1903
|
};
|
|
1271
|
-
readonly "
|
|
1904
|
+
readonly "nvidia/llama-3.3-nemotron-super-49b-v1.5": {
|
|
1272
1905
|
id: string;
|
|
1273
1906
|
name: string;
|
|
1274
1907
|
api: "openai-completions";
|
|
@@ -1285,13 +1918,13 @@ export declare const MODELS: {
|
|
|
1285
1918
|
contextWindow: number;
|
|
1286
1919
|
maxTokens: number;
|
|
1287
1920
|
};
|
|
1288
|
-
readonly "
|
|
1921
|
+
readonly "qwen/qwen3-vl-30b-a3b-thinking": {
|
|
1289
1922
|
id: string;
|
|
1290
1923
|
name: string;
|
|
1291
1924
|
api: "openai-completions";
|
|
1292
1925
|
provider: string;
|
|
1293
1926
|
baseUrl: string;
|
|
1294
|
-
reasoning:
|
|
1927
|
+
reasoning: true;
|
|
1295
1928
|
input: ("text" | "image")[];
|
|
1296
1929
|
cost: {
|
|
1297
1930
|
input: number;
|
|
@@ -1302,14 +1935,14 @@ export declare const MODELS: {
|
|
|
1302
1935
|
contextWindow: number;
|
|
1303
1936
|
maxTokens: number;
|
|
1304
1937
|
};
|
|
1305
|
-
readonly "
|
|
1938
|
+
readonly "qwen/qwen3-vl-30b-a3b-instruct": {
|
|
1306
1939
|
id: string;
|
|
1307
1940
|
name: string;
|
|
1308
1941
|
api: "openai-completions";
|
|
1309
1942
|
provider: string;
|
|
1310
1943
|
baseUrl: string;
|
|
1311
|
-
reasoning:
|
|
1312
|
-
input: "text"[];
|
|
1944
|
+
reasoning: false;
|
|
1945
|
+
input: ("text" | "image")[];
|
|
1313
1946
|
cost: {
|
|
1314
1947
|
input: number;
|
|
1315
1948
|
output: number;
|
|
@@ -1319,13 +1952,13 @@ export declare const MODELS: {
|
|
|
1319
1952
|
contextWindow: number;
|
|
1320
1953
|
maxTokens: number;
|
|
1321
1954
|
};
|
|
1322
|
-
readonly "
|
|
1955
|
+
readonly "z-ai/glm-4.6": {
|
|
1323
1956
|
id: string;
|
|
1324
1957
|
name: string;
|
|
1325
1958
|
api: "openai-completions";
|
|
1326
1959
|
provider: string;
|
|
1327
1960
|
baseUrl: string;
|
|
1328
|
-
reasoning:
|
|
1961
|
+
reasoning: true;
|
|
1329
1962
|
input: "text"[];
|
|
1330
1963
|
cost: {
|
|
1331
1964
|
input: number;
|
|
@@ -1336,12 +1969,10 @@ export declare const MODELS: {
|
|
|
1336
1969
|
contextWindow: number;
|
|
1337
1970
|
maxTokens: number;
|
|
1338
1971
|
};
|
|
1339
|
-
|
|
1340
|
-
readonly zai: {
|
|
1341
|
-
readonly "glm-4.5-air": {
|
|
1972
|
+
readonly "deepseek/deepseek-v3.2-exp": {
|
|
1342
1973
|
id: string;
|
|
1343
1974
|
name: string;
|
|
1344
|
-
api: "
|
|
1975
|
+
api: "openai-completions";
|
|
1345
1976
|
provider: string;
|
|
1346
1977
|
baseUrl: string;
|
|
1347
1978
|
reasoning: true;
|
|
@@ -1355,10 +1986,10 @@ export declare const MODELS: {
|
|
|
1355
1986
|
contextWindow: number;
|
|
1356
1987
|
maxTokens: number;
|
|
1357
1988
|
};
|
|
1358
|
-
readonly "
|
|
1989
|
+
readonly "qwen/qwen3-vl-235b-a22b-thinking": {
|
|
1359
1990
|
id: string;
|
|
1360
1991
|
name: string;
|
|
1361
|
-
api: "
|
|
1992
|
+
api: "openai-completions";
|
|
1362
1993
|
provider: string;
|
|
1363
1994
|
baseUrl: string;
|
|
1364
1995
|
reasoning: true;
|
|
@@ -1372,14 +2003,14 @@ export declare const MODELS: {
|
|
|
1372
2003
|
contextWindow: number;
|
|
1373
2004
|
maxTokens: number;
|
|
1374
2005
|
};
|
|
1375
|
-
readonly "
|
|
2006
|
+
readonly "qwen/qwen3-vl-235b-a22b-instruct": {
|
|
1376
2007
|
id: string;
|
|
1377
2008
|
name: string;
|
|
1378
|
-
api: "
|
|
2009
|
+
api: "openai-completions";
|
|
1379
2010
|
provider: string;
|
|
1380
2011
|
baseUrl: string;
|
|
1381
2012
|
reasoning: true;
|
|
1382
|
-
input: "text"[];
|
|
2013
|
+
input: ("text" | "image")[];
|
|
1383
2014
|
cost: {
|
|
1384
2015
|
input: number;
|
|
1385
2016
|
output: number;
|
|
@@ -1389,13 +2020,13 @@ export declare const MODELS: {
|
|
|
1389
2020
|
contextWindow: number;
|
|
1390
2021
|
maxTokens: number;
|
|
1391
2022
|
};
|
|
1392
|
-
readonly "
|
|
2023
|
+
readonly "qwen/qwen3-max": {
|
|
1393
2024
|
id: string;
|
|
1394
2025
|
name: string;
|
|
1395
|
-
api: "
|
|
2026
|
+
api: "openai-completions";
|
|
1396
2027
|
provider: string;
|
|
1397
2028
|
baseUrl: string;
|
|
1398
|
-
reasoning:
|
|
2029
|
+
reasoning: false;
|
|
1399
2030
|
input: "text"[];
|
|
1400
2031
|
cost: {
|
|
1401
2032
|
input: number;
|
|
@@ -1406,15 +2037,13 @@ export declare const MODELS: {
|
|
|
1406
2037
|
contextWindow: number;
|
|
1407
2038
|
maxTokens: number;
|
|
1408
2039
|
};
|
|
1409
|
-
|
|
1410
|
-
readonly openrouter: {
|
|
1411
|
-
readonly "alibaba/tongyi-deepresearch-30b-a3b": {
|
|
2040
|
+
readonly "qwen/qwen3-coder-plus": {
|
|
1412
2041
|
id: string;
|
|
1413
2042
|
name: string;
|
|
1414
2043
|
api: "openai-completions";
|
|
1415
2044
|
provider: string;
|
|
1416
2045
|
baseUrl: string;
|
|
1417
|
-
reasoning:
|
|
2046
|
+
reasoning: false;
|
|
1418
2047
|
input: "text"[];
|
|
1419
2048
|
cost: {
|
|
1420
2049
|
input: number;
|
|
@@ -1425,13 +2054,13 @@ export declare const MODELS: {
|
|
|
1425
2054
|
contextWindow: number;
|
|
1426
2055
|
maxTokens: number;
|
|
1427
2056
|
};
|
|
1428
|
-
readonly "
|
|
2057
|
+
readonly "deepseek/deepseek-v3.1-terminus": {
|
|
1429
2058
|
id: string;
|
|
1430
2059
|
name: string;
|
|
1431
2060
|
api: "openai-completions";
|
|
1432
2061
|
provider: string;
|
|
1433
2062
|
baseUrl: string;
|
|
1434
|
-
reasoning:
|
|
2063
|
+
reasoning: true;
|
|
1435
2064
|
input: "text"[];
|
|
1436
2065
|
cost: {
|
|
1437
2066
|
input: number;
|
|
@@ -1442,13 +2071,13 @@ export declare const MODELS: {
|
|
|
1442
2071
|
contextWindow: number;
|
|
1443
2072
|
maxTokens: number;
|
|
1444
2073
|
};
|
|
1445
|
-
readonly "
|
|
2074
|
+
readonly "alibaba/tongyi-deepresearch-30b-a3b:free": {
|
|
1446
2075
|
id: string;
|
|
1447
2076
|
name: string;
|
|
1448
2077
|
api: "openai-completions";
|
|
1449
2078
|
provider: string;
|
|
1450
2079
|
baseUrl: string;
|
|
1451
|
-
reasoning:
|
|
2080
|
+
reasoning: true;
|
|
1452
2081
|
input: "text"[];
|
|
1453
2082
|
cost: {
|
|
1454
2083
|
input: number;
|
|
@@ -1459,7 +2088,7 @@ export declare const MODELS: {
|
|
|
1459
2088
|
contextWindow: number;
|
|
1460
2089
|
maxTokens: number;
|
|
1461
2090
|
};
|
|
1462
|
-
readonly "
|
|
2091
|
+
readonly "alibaba/tongyi-deepresearch-30b-a3b": {
|
|
1463
2092
|
id: string;
|
|
1464
2093
|
name: string;
|
|
1465
2094
|
api: "openai-completions";
|
|
@@ -1476,7 +2105,7 @@ export declare const MODELS: {
|
|
|
1476
2105
|
contextWindow: number;
|
|
1477
2106
|
maxTokens: number;
|
|
1478
2107
|
};
|
|
1479
|
-
readonly "qwen/qwen3-
|
|
2108
|
+
readonly "qwen/qwen3-coder-flash": {
|
|
1480
2109
|
id: string;
|
|
1481
2110
|
name: string;
|
|
1482
2111
|
api: "openai-completions";
|
|
@@ -1493,13 +2122,13 @@ export declare const MODELS: {
|
|
|
1493
2122
|
contextWindow: number;
|
|
1494
2123
|
maxTokens: number;
|
|
1495
2124
|
};
|
|
1496
|
-
readonly "
|
|
2125
|
+
readonly "qwen/qwen3-next-80b-a3b-thinking": {
|
|
1497
2126
|
id: string;
|
|
1498
2127
|
name: string;
|
|
1499
2128
|
api: "openai-completions";
|
|
1500
2129
|
provider: string;
|
|
1501
2130
|
baseUrl: string;
|
|
1502
|
-
reasoning:
|
|
2131
|
+
reasoning: true;
|
|
1503
2132
|
input: "text"[];
|
|
1504
2133
|
cost: {
|
|
1505
2134
|
input: number;
|
|
@@ -1510,7 +2139,7 @@ export declare const MODELS: {
|
|
|
1510
2139
|
contextWindow: number;
|
|
1511
2140
|
maxTokens: number;
|
|
1512
2141
|
};
|
|
1513
|
-
readonly "qwen/
|
|
2142
|
+
readonly "qwen/qwen3-next-80b-a3b-instruct": {
|
|
1514
2143
|
id: string;
|
|
1515
2144
|
name: string;
|
|
1516
2145
|
api: "openai-completions";
|
|
@@ -1527,13 +2156,13 @@ export declare const MODELS: {
|
|
|
1527
2156
|
contextWindow: number;
|
|
1528
2157
|
maxTokens: number;
|
|
1529
2158
|
};
|
|
1530
|
-
readonly "
|
|
2159
|
+
readonly "meituan/longcat-flash-chat:free": {
|
|
1531
2160
|
id: string;
|
|
1532
2161
|
name: string;
|
|
1533
2162
|
api: "openai-completions";
|
|
1534
2163
|
provider: string;
|
|
1535
2164
|
baseUrl: string;
|
|
1536
|
-
reasoning:
|
|
2165
|
+
reasoning: false;
|
|
1537
2166
|
input: "text"[];
|
|
1538
2167
|
cost: {
|
|
1539
2168
|
input: number;
|
|
@@ -1544,13 +2173,13 @@ export declare const MODELS: {
|
|
|
1544
2173
|
contextWindow: number;
|
|
1545
2174
|
maxTokens: number;
|
|
1546
2175
|
};
|
|
1547
|
-
readonly "
|
|
2176
|
+
readonly "qwen/qwen-plus-2025-07-28": {
|
|
1548
2177
|
id: string;
|
|
1549
2178
|
name: string;
|
|
1550
2179
|
api: "openai-completions";
|
|
1551
2180
|
provider: string;
|
|
1552
2181
|
baseUrl: string;
|
|
1553
|
-
reasoning:
|
|
2182
|
+
reasoning: false;
|
|
1554
2183
|
input: "text"[];
|
|
1555
2184
|
cost: {
|
|
1556
2185
|
input: number;
|
|
@@ -1561,7 +2190,7 @@ export declare const MODELS: {
|
|
|
1561
2190
|
contextWindow: number;
|
|
1562
2191
|
maxTokens: number;
|
|
1563
2192
|
};
|
|
1564
|
-
readonly "
|
|
2193
|
+
readonly "qwen/qwen-plus-2025-07-28:thinking": {
|
|
1565
2194
|
id: string;
|
|
1566
2195
|
name: string;
|
|
1567
2196
|
api: "openai-completions";
|
|
@@ -1578,14 +2207,14 @@ export declare const MODELS: {
|
|
|
1578
2207
|
contextWindow: number;
|
|
1579
2208
|
maxTokens: number;
|
|
1580
2209
|
};
|
|
1581
|
-
readonly "
|
|
2210
|
+
readonly "nvidia/nemotron-nano-9b-v2:free": {
|
|
1582
2211
|
id: string;
|
|
1583
2212
|
name: string;
|
|
1584
2213
|
api: "openai-completions";
|
|
1585
2214
|
provider: string;
|
|
1586
2215
|
baseUrl: string;
|
|
1587
|
-
reasoning:
|
|
1588
|
-
input:
|
|
2216
|
+
reasoning: true;
|
|
2217
|
+
input: "text"[];
|
|
1589
2218
|
cost: {
|
|
1590
2219
|
input: number;
|
|
1591
2220
|
output: number;
|
|
@@ -1595,14 +2224,14 @@ export declare const MODELS: {
|
|
|
1595
2224
|
contextWindow: number;
|
|
1596
2225
|
maxTokens: number;
|
|
1597
2226
|
};
|
|
1598
|
-
readonly "
|
|
2227
|
+
readonly "nvidia/nemotron-nano-9b-v2": {
|
|
1599
2228
|
id: string;
|
|
1600
2229
|
name: string;
|
|
1601
2230
|
api: "openai-completions";
|
|
1602
2231
|
provider: string;
|
|
1603
2232
|
baseUrl: string;
|
|
1604
2233
|
reasoning: true;
|
|
1605
|
-
input:
|
|
2234
|
+
input: "text"[];
|
|
1606
2235
|
cost: {
|
|
1607
2236
|
input: number;
|
|
1608
2237
|
output: number;
|
|
@@ -1612,7 +2241,7 @@ export declare const MODELS: {
|
|
|
1612
2241
|
contextWindow: number;
|
|
1613
2242
|
maxTokens: number;
|
|
1614
2243
|
};
|
|
1615
|
-
readonly "
|
|
2244
|
+
readonly "moonshotai/kimi-k2-0905": {
|
|
1616
2245
|
id: string;
|
|
1617
2246
|
name: string;
|
|
1618
2247
|
api: "openai-completions";
|
|
@@ -1629,13 +2258,13 @@ export declare const MODELS: {
|
|
|
1629
2258
|
contextWindow: number;
|
|
1630
2259
|
maxTokens: number;
|
|
1631
2260
|
};
|
|
1632
|
-
readonly "
|
|
2261
|
+
readonly "deepcogito/cogito-v2-preview-llama-70b": {
|
|
1633
2262
|
id: string;
|
|
1634
2263
|
name: string;
|
|
1635
2264
|
api: "openai-completions";
|
|
1636
2265
|
provider: string;
|
|
1637
2266
|
baseUrl: string;
|
|
1638
|
-
reasoning:
|
|
2267
|
+
reasoning: true;
|
|
1639
2268
|
input: "text"[];
|
|
1640
2269
|
cost: {
|
|
1641
2270
|
input: number;
|
|
@@ -1731,7 +2360,7 @@ export declare const MODELS: {
|
|
|
1731
2360
|
contextWindow: number;
|
|
1732
2361
|
maxTokens: number;
|
|
1733
2362
|
};
|
|
1734
|
-
readonly "deepseek/deepseek-chat-v3.1
|
|
2363
|
+
readonly "deepseek/deepseek-chat-v3.1": {
|
|
1735
2364
|
id: string;
|
|
1736
2365
|
name: string;
|
|
1737
2366
|
api: "openai-completions";
|
|
@@ -1748,14 +2377,14 @@ export declare const MODELS: {
|
|
|
1748
2377
|
contextWindow: number;
|
|
1749
2378
|
maxTokens: number;
|
|
1750
2379
|
};
|
|
1751
|
-
readonly "
|
|
2380
|
+
readonly "mistralai/mistral-medium-3.1": {
|
|
1752
2381
|
id: string;
|
|
1753
2382
|
name: string;
|
|
1754
2383
|
api: "openai-completions";
|
|
1755
2384
|
provider: string;
|
|
1756
2385
|
baseUrl: string;
|
|
1757
|
-
reasoning:
|
|
1758
|
-
input: "text"[];
|
|
2386
|
+
reasoning: false;
|
|
2387
|
+
input: ("text" | "image")[];
|
|
1759
2388
|
cost: {
|
|
1760
2389
|
input: number;
|
|
1761
2390
|
output: number;
|
|
@@ -1765,13 +2394,30 @@ export declare const MODELS: {
|
|
|
1765
2394
|
contextWindow: number;
|
|
1766
2395
|
maxTokens: number;
|
|
1767
2396
|
};
|
|
1768
|
-
readonly "
|
|
2397
|
+
readonly "baidu/ernie-4.5-21b-a3b": {
|
|
1769
2398
|
id: string;
|
|
1770
2399
|
name: string;
|
|
1771
2400
|
api: "openai-completions";
|
|
1772
2401
|
provider: string;
|
|
1773
2402
|
baseUrl: string;
|
|
1774
2403
|
reasoning: false;
|
|
2404
|
+
input: "text"[];
|
|
2405
|
+
cost: {
|
|
2406
|
+
input: number;
|
|
2407
|
+
output: number;
|
|
2408
|
+
cacheRead: number;
|
|
2409
|
+
cacheWrite: number;
|
|
2410
|
+
};
|
|
2411
|
+
contextWindow: number;
|
|
2412
|
+
maxTokens: number;
|
|
2413
|
+
};
|
|
2414
|
+
readonly "baidu/ernie-4.5-vl-28b-a3b": {
|
|
2415
|
+
id: string;
|
|
2416
|
+
name: string;
|
|
2417
|
+
api: "openai-completions";
|
|
2418
|
+
provider: string;
|
|
2419
|
+
baseUrl: string;
|
|
2420
|
+
reasoning: true;
|
|
1775
2421
|
input: ("text" | "image")[];
|
|
1776
2422
|
cost: {
|
|
1777
2423
|
input: number;
|
|
@@ -1992,7 +2638,7 @@ export declare const MODELS: {
|
|
|
1992
2638
|
api: "openai-completions";
|
|
1993
2639
|
provider: string;
|
|
1994
2640
|
baseUrl: string;
|
|
1995
|
-
reasoning:
|
|
2641
|
+
reasoning: true;
|
|
1996
2642
|
input: "text"[];
|
|
1997
2643
|
cost: {
|
|
1998
2644
|
input: number;
|
|
@@ -2009,7 +2655,7 @@ export declare const MODELS: {
|
|
|
2009
2655
|
api: "openai-completions";
|
|
2010
2656
|
provider: string;
|
|
2011
2657
|
baseUrl: string;
|
|
2012
|
-
reasoning:
|
|
2658
|
+
reasoning: true;
|
|
2013
2659
|
input: "text"[];
|
|
2014
2660
|
cost: {
|
|
2015
2661
|
input: number;
|
|
@@ -2020,7 +2666,7 @@ export declare const MODELS: {
|
|
|
2020
2666
|
contextWindow: number;
|
|
2021
2667
|
maxTokens: number;
|
|
2022
2668
|
};
|
|
2023
|
-
readonly "moonshotai/kimi-k2
|
|
2669
|
+
readonly "moonshotai/kimi-k2": {
|
|
2024
2670
|
id: string;
|
|
2025
2671
|
name: string;
|
|
2026
2672
|
api: "openai-completions";
|
|
@@ -2037,7 +2683,7 @@ export declare const MODELS: {
|
|
|
2037
2683
|
contextWindow: number;
|
|
2038
2684
|
maxTokens: number;
|
|
2039
2685
|
};
|
|
2040
|
-
readonly "
|
|
2686
|
+
readonly "mistralai/devstral-medium": {
|
|
2041
2687
|
id: string;
|
|
2042
2688
|
name: string;
|
|
2043
2689
|
api: "openai-completions";
|
|
@@ -2054,7 +2700,7 @@ export declare const MODELS: {
|
|
|
2054
2700
|
contextWindow: number;
|
|
2055
2701
|
maxTokens: number;
|
|
2056
2702
|
};
|
|
2057
|
-
readonly "mistralai/devstral-
|
|
2703
|
+
readonly "mistralai/devstral-small": {
|
|
2058
2704
|
id: string;
|
|
2059
2705
|
name: string;
|
|
2060
2706
|
api: "openai-completions";
|
|
@@ -2071,13 +2717,13 @@ export declare const MODELS: {
|
|
|
2071
2717
|
contextWindow: number;
|
|
2072
2718
|
maxTokens: number;
|
|
2073
2719
|
};
|
|
2074
|
-
readonly "
|
|
2720
|
+
readonly "tngtech/deepseek-r1t2-chimera": {
|
|
2075
2721
|
id: string;
|
|
2076
2722
|
name: string;
|
|
2077
2723
|
api: "openai-completions";
|
|
2078
2724
|
provider: string;
|
|
2079
2725
|
baseUrl: string;
|
|
2080
|
-
reasoning:
|
|
2726
|
+
reasoning: true;
|
|
2081
2727
|
input: "text"[];
|
|
2082
2728
|
cost: {
|
|
2083
2729
|
input: number;
|
|
@@ -2275,6 +2921,23 @@ export declare const MODELS: {
|
|
|
2275
2921
|
contextWindow: number;
|
|
2276
2922
|
maxTokens: number;
|
|
2277
2923
|
};
|
|
2924
|
+
readonly "nousresearch/deephermes-3-mistral-24b-preview": {
|
|
2925
|
+
id: string;
|
|
2926
|
+
name: string;
|
|
2927
|
+
api: "openai-completions";
|
|
2928
|
+
provider: string;
|
|
2929
|
+
baseUrl: string;
|
|
2930
|
+
reasoning: true;
|
|
2931
|
+
input: "text"[];
|
|
2932
|
+
cost: {
|
|
2933
|
+
input: number;
|
|
2934
|
+
output: number;
|
|
2935
|
+
cacheRead: number;
|
|
2936
|
+
cacheWrite: number;
|
|
2937
|
+
};
|
|
2938
|
+
contextWindow: number;
|
|
2939
|
+
maxTokens: number;
|
|
2940
|
+
};
|
|
2278
2941
|
readonly "mistralai/mistral-medium-3": {
|
|
2279
2942
|
id: string;
|
|
2280
2943
|
name: string;
|
|
@@ -2519,7 +3182,7 @@ export declare const MODELS: {
|
|
|
2519
3182
|
api: "openai-completions";
|
|
2520
3183
|
provider: string;
|
|
2521
3184
|
baseUrl: string;
|
|
2522
|
-
reasoning:
|
|
3185
|
+
reasoning: true;
|
|
2523
3186
|
input: "text"[];
|
|
2524
3187
|
cost: {
|
|
2525
3188
|
input: number;
|
|
@@ -2564,14 +3227,14 @@ export declare const MODELS: {
|
|
|
2564
3227
|
contextWindow: number;
|
|
2565
3228
|
maxTokens: number;
|
|
2566
3229
|
};
|
|
2567
|
-
readonly "
|
|
3230
|
+
readonly "qwen/qwq-32b": {
|
|
2568
3231
|
id: string;
|
|
2569
3232
|
name: string;
|
|
2570
3233
|
api: "openai-completions";
|
|
2571
3234
|
provider: string;
|
|
2572
3235
|
baseUrl: string;
|
|
2573
|
-
reasoning:
|
|
2574
|
-
input:
|
|
3236
|
+
reasoning: true;
|
|
3237
|
+
input: "text"[];
|
|
2575
3238
|
cost: {
|
|
2576
3239
|
input: number;
|
|
2577
3240
|
output: number;
|
|
@@ -2581,13 +3244,13 @@ export declare const MODELS: {
|
|
|
2581
3244
|
contextWindow: number;
|
|
2582
3245
|
maxTokens: number;
|
|
2583
3246
|
};
|
|
2584
|
-
readonly "
|
|
3247
|
+
readonly "nousresearch/deephermes-3-llama-3-8b-preview": {
|
|
2585
3248
|
id: string;
|
|
2586
3249
|
name: string;
|
|
2587
3250
|
api: "openai-completions";
|
|
2588
3251
|
provider: string;
|
|
2589
3252
|
baseUrl: string;
|
|
2590
|
-
reasoning:
|
|
3253
|
+
reasoning: false;
|
|
2591
3254
|
input: "text"[];
|
|
2592
3255
|
cost: {
|
|
2593
3256
|
input: number;
|
|
@@ -2615,6 +3278,23 @@ export declare const MODELS: {
|
|
|
2615
3278
|
contextWindow: number;
|
|
2616
3279
|
maxTokens: number;
|
|
2617
3280
|
};
|
|
3281
|
+
readonly "qwen/qwen-vl-max": {
|
|
3282
|
+
id: string;
|
|
3283
|
+
name: string;
|
|
3284
|
+
api: "openai-completions";
|
|
3285
|
+
provider: string;
|
|
3286
|
+
baseUrl: string;
|
|
3287
|
+
reasoning: false;
|
|
3288
|
+
input: ("text" | "image")[];
|
|
3289
|
+
cost: {
|
|
3290
|
+
input: number;
|
|
3291
|
+
output: number;
|
|
3292
|
+
cacheRead: number;
|
|
3293
|
+
cacheWrite: number;
|
|
3294
|
+
};
|
|
3295
|
+
contextWindow: number;
|
|
3296
|
+
maxTokens: number;
|
|
3297
|
+
};
|
|
2618
3298
|
readonly "qwen/qwen-turbo": {
|
|
2619
3299
|
id: string;
|
|
2620
3300
|
name: string;
|
|
@@ -2921,7 +3601,7 @@ export declare const MODELS: {
|
|
|
2921
3601
|
contextWindow: number;
|
|
2922
3602
|
maxTokens: number;
|
|
2923
3603
|
};
|
|
2924
|
-
readonly "
|
|
3604
|
+
readonly "qwen/qwen-2.5-7b-instruct": {
|
|
2925
3605
|
id: string;
|
|
2926
3606
|
name: string;
|
|
2927
3607
|
api: "openai-completions";
|
|
@@ -2938,7 +3618,7 @@ export declare const MODELS: {
|
|
|
2938
3618
|
contextWindow: number;
|
|
2939
3619
|
maxTokens: number;
|
|
2940
3620
|
};
|
|
2941
|
-
readonly "
|
|
3621
|
+
readonly "nvidia/llama-3.1-nemotron-70b-instruct": {
|
|
2942
3622
|
id: string;
|
|
2943
3623
|
name: string;
|
|
2944
3624
|
api: "openai-completions";
|
|
@@ -2955,7 +3635,7 @@ export declare const MODELS: {
|
|
|
2955
3635
|
contextWindow: number;
|
|
2956
3636
|
maxTokens: number;
|
|
2957
3637
|
};
|
|
2958
|
-
readonly "
|
|
3638
|
+
readonly "thedrummer/rocinante-12b": {
|
|
2959
3639
|
id: string;
|
|
2960
3640
|
name: string;
|
|
2961
3641
|
api: "openai-completions";
|
|
@@ -2972,7 +3652,7 @@ export declare const MODELS: {
|
|
|
2972
3652
|
contextWindow: number;
|
|
2973
3653
|
maxTokens: number;
|
|
2974
3654
|
};
|
|
2975
|
-
readonly "
|
|
3655
|
+
readonly "meta-llama/llama-3.2-3b-instruct": {
|
|
2976
3656
|
id: string;
|
|
2977
3657
|
name: string;
|
|
2978
3658
|
api: "openai-completions";
|
|
@@ -2989,24 +3669,7 @@ export declare const MODELS: {
|
|
|
2989
3669
|
contextWindow: number;
|
|
2990
3670
|
maxTokens: number;
|
|
2991
3671
|
};
|
|
2992
|
-
readonly "
|
|
2993
|
-
id: string;
|
|
2994
|
-
name: string;
|
|
2995
|
-
api: "openai-completions";
|
|
2996
|
-
provider: string;
|
|
2997
|
-
baseUrl: string;
|
|
2998
|
-
reasoning: false;
|
|
2999
|
-
input: ("text" | "image")[];
|
|
3000
|
-
cost: {
|
|
3001
|
-
input: number;
|
|
3002
|
-
output: number;
|
|
3003
|
-
cacheRead: number;
|
|
3004
|
-
cacheWrite: number;
|
|
3005
|
-
};
|
|
3006
|
-
contextWindow: number;
|
|
3007
|
-
maxTokens: number;
|
|
3008
|
-
};
|
|
3009
|
-
readonly "cohere/command-r-08-2024": {
|
|
3672
|
+
readonly "qwen/qwen-2.5-72b-instruct": {
|
|
3010
3673
|
id: string;
|
|
3011
3674
|
name: string;
|
|
3012
3675
|
api: "openai-completions";
|
|
@@ -3023,14 +3686,14 @@ export declare const MODELS: {
|
|
|
3023
3686
|
contextWindow: number;
|
|
3024
3687
|
maxTokens: number;
|
|
3025
3688
|
};
|
|
3026
|
-
readonly "
|
|
3689
|
+
readonly "mistralai/pixtral-12b": {
|
|
3027
3690
|
id: string;
|
|
3028
3691
|
name: string;
|
|
3029
3692
|
api: "openai-completions";
|
|
3030
3693
|
provider: string;
|
|
3031
3694
|
baseUrl: string;
|
|
3032
3695
|
reasoning: false;
|
|
3033
|
-
input: "text"[];
|
|
3696
|
+
input: ("text" | "image")[];
|
|
3034
3697
|
cost: {
|
|
3035
3698
|
input: number;
|
|
3036
3699
|
output: number;
|
|
@@ -3040,7 +3703,7 @@ export declare const MODELS: {
|
|
|
3040
3703
|
contextWindow: number;
|
|
3041
3704
|
maxTokens: number;
|
|
3042
3705
|
};
|
|
3043
|
-
readonly "
|
|
3706
|
+
readonly "cohere/command-r-plus-08-2024": {
|
|
3044
3707
|
id: string;
|
|
3045
3708
|
name: string;
|
|
3046
3709
|
api: "openai-completions";
|
|
@@ -3057,7 +3720,7 @@ export declare const MODELS: {
|
|
|
3057
3720
|
contextWindow: number;
|
|
3058
3721
|
maxTokens: number;
|
|
3059
3722
|
};
|
|
3060
|
-
readonly "
|
|
3723
|
+
readonly "cohere/command-r-08-2024": {
|
|
3061
3724
|
id: string;
|
|
3062
3725
|
name: string;
|
|
3063
3726
|
api: "openai-completions";
|
|
@@ -3074,7 +3737,7 @@ export declare const MODELS: {
|
|
|
3074
3737
|
contextWindow: number;
|
|
3075
3738
|
maxTokens: number;
|
|
3076
3739
|
};
|
|
3077
|
-
readonly "
|
|
3740
|
+
readonly "sao10k/l3.1-euryale-70b": {
|
|
3078
3741
|
id: string;
|
|
3079
3742
|
name: string;
|
|
3080
3743
|
api: "openai-completions";
|
|
@@ -3091,7 +3754,7 @@ export declare const MODELS: {
|
|
|
3091
3754
|
contextWindow: number;
|
|
3092
3755
|
maxTokens: number;
|
|
3093
3756
|
};
|
|
3094
|
-
readonly "
|
|
3757
|
+
readonly "microsoft/phi-3.5-mini-128k-instruct": {
|
|
3095
3758
|
id: string;
|
|
3096
3759
|
name: string;
|
|
3097
3760
|
api: "openai-completions";
|
|
@@ -3108,7 +3771,7 @@ export declare const MODELS: {
|
|
|
3108
3771
|
contextWindow: number;
|
|
3109
3772
|
maxTokens: number;
|
|
3110
3773
|
};
|
|
3111
|
-
readonly "
|
|
3774
|
+
readonly "nousresearch/hermes-3-llama-3.1-70b": {
|
|
3112
3775
|
id: string;
|
|
3113
3776
|
name: string;
|
|
3114
3777
|
api: "openai-completions";
|
|
@@ -3125,7 +3788,7 @@ export declare const MODELS: {
|
|
|
3125
3788
|
contextWindow: number;
|
|
3126
3789
|
maxTokens: number;
|
|
3127
3790
|
};
|
|
3128
|
-
readonly "
|
|
3791
|
+
readonly "meta-llama/llama-3.1-8b-instruct": {
|
|
3129
3792
|
id: string;
|
|
3130
3793
|
name: string;
|
|
3131
3794
|
api: "openai-completions";
|
|
@@ -3142,7 +3805,7 @@ export declare const MODELS: {
|
|
|
3142
3805
|
contextWindow: number;
|
|
3143
3806
|
maxTokens: number;
|
|
3144
3807
|
};
|
|
3145
|
-
readonly "
|
|
3808
|
+
readonly "meta-llama/llama-3.1-405b-instruct": {
|
|
3146
3809
|
id: string;
|
|
3147
3810
|
name: string;
|
|
3148
3811
|
api: "openai-completions";
|
|
@@ -3159,7 +3822,7 @@ export declare const MODELS: {
|
|
|
3159
3822
|
contextWindow: number;
|
|
3160
3823
|
maxTokens: number;
|
|
3161
3824
|
};
|
|
3162
|
-
readonly "
|
|
3825
|
+
readonly "meta-llama/llama-3.1-70b-instruct": {
|
|
3163
3826
|
id: string;
|
|
3164
3827
|
name: string;
|
|
3165
3828
|
api: "openai-completions";
|
|
@@ -3176,7 +3839,7 @@ export declare const MODELS: {
|
|
|
3176
3839
|
contextWindow: number;
|
|
3177
3840
|
maxTokens: number;
|
|
3178
3841
|
};
|
|
3179
|
-
readonly "mistralai/mistral-
|
|
3842
|
+
readonly "mistralai/mistral-nemo": {
|
|
3180
3843
|
id: string;
|
|
3181
3844
|
name: string;
|
|
3182
3845
|
api: "openai-completions";
|
|
@@ -3193,7 +3856,7 @@ export declare const MODELS: {
|
|
|
3193
3856
|
contextWindow: number;
|
|
3194
3857
|
maxTokens: number;
|
|
3195
3858
|
};
|
|
3196
|
-
readonly "
|
|
3859
|
+
readonly "sao10k/l3-euryale-70b": {
|
|
3197
3860
|
id: string;
|
|
3198
3861
|
name: string;
|
|
3199
3862
|
api: "openai-completions";
|
|
@@ -3210,7 +3873,7 @@ export declare const MODELS: {
|
|
|
3210
3873
|
contextWindow: number;
|
|
3211
3874
|
maxTokens: number;
|
|
3212
3875
|
};
|
|
3213
|
-
readonly "
|
|
3876
|
+
readonly "mistralai/mistral-7b-instruct:free": {
|
|
3214
3877
|
id: string;
|
|
3215
3878
|
name: string;
|
|
3216
3879
|
api: "openai-completions";
|
|
@@ -3227,7 +3890,7 @@ export declare const MODELS: {
|
|
|
3227
3890
|
contextWindow: number;
|
|
3228
3891
|
maxTokens: number;
|
|
3229
3892
|
};
|
|
3230
|
-
readonly "
|
|
3893
|
+
readonly "mistralai/mistral-7b-instruct": {
|
|
3231
3894
|
id: string;
|
|
3232
3895
|
name: string;
|
|
3233
3896
|
api: "openai-completions";
|
|
@@ -3244,7 +3907,7 @@ export declare const MODELS: {
|
|
|
3244
3907
|
contextWindow: number;
|
|
3245
3908
|
maxTokens: number;
|
|
3246
3909
|
};
|
|
3247
|
-
readonly "
|
|
3910
|
+
readonly "mistralai/mistral-7b-instruct-v0.3": {
|
|
3248
3911
|
id: string;
|
|
3249
3912
|
name: string;
|
|
3250
3913
|
api: "openai-completions";
|
|
@@ -3261,7 +3924,7 @@ export declare const MODELS: {
|
|
|
3261
3924
|
contextWindow: number;
|
|
3262
3925
|
maxTokens: number;
|
|
3263
3926
|
};
|
|
3264
|
-
readonly "
|
|
3927
|
+
readonly "microsoft/phi-3-mini-128k-instruct": {
|
|
3265
3928
|
id: string;
|
|
3266
3929
|
name: string;
|
|
3267
3930
|
api: "openai-completions";
|
|
@@ -3278,7 +3941,7 @@ export declare const MODELS: {
|
|
|
3278
3941
|
contextWindow: number;
|
|
3279
3942
|
maxTokens: number;
|
|
3280
3943
|
};
|
|
3281
|
-
readonly "
|
|
3944
|
+
readonly "microsoft/phi-3-medium-128k-instruct": {
|
|
3282
3945
|
id: string;
|
|
3283
3946
|
name: string;
|
|
3284
3947
|
api: "openai-completions";
|
|
@@ -3295,7 +3958,7 @@ export declare const MODELS: {
|
|
|
3295
3958
|
contextWindow: number;
|
|
3296
3959
|
maxTokens: number;
|
|
3297
3960
|
};
|
|
3298
|
-
readonly "
|
|
3961
|
+
readonly "meta-llama/llama-3-8b-instruct": {
|
|
3299
3962
|
id: string;
|
|
3300
3963
|
name: string;
|
|
3301
3964
|
api: "openai-completions";
|
|
@@ -3312,7 +3975,7 @@ export declare const MODELS: {
|
|
|
3312
3975
|
contextWindow: number;
|
|
3313
3976
|
maxTokens: number;
|
|
3314
3977
|
};
|
|
3315
|
-
readonly "
|
|
3978
|
+
readonly "meta-llama/llama-3-70b-instruct": {
|
|
3316
3979
|
id: string;
|
|
3317
3980
|
name: string;
|
|
3318
3981
|
api: "openai-completions";
|
|
@@ -3329,7 +3992,7 @@ export declare const MODELS: {
|
|
|
3329
3992
|
contextWindow: number;
|
|
3330
3993
|
maxTokens: number;
|
|
3331
3994
|
};
|
|
3332
|
-
readonly "
|
|
3995
|
+
readonly "mistralai/mixtral-8x22b-instruct": {
|
|
3333
3996
|
id: string;
|
|
3334
3997
|
name: string;
|
|
3335
3998
|
api: "openai-completions";
|
|
@@ -3363,7 +4026,7 @@ export declare const MODELS: {
|
|
|
3363
4026
|
contextWindow: number;
|
|
3364
4027
|
maxTokens: number;
|
|
3365
4028
|
};
|
|
3366
|
-
readonly "mistralai/mistral-
|
|
4029
|
+
readonly "mistralai/mistral-small": {
|
|
3367
4030
|
id: string;
|
|
3368
4031
|
name: string;
|
|
3369
4032
|
api: "openai-completions";
|
|
@@ -3380,7 +4043,7 @@ export declare const MODELS: {
|
|
|
3380
4043
|
contextWindow: number;
|
|
3381
4044
|
maxTokens: number;
|
|
3382
4045
|
};
|
|
3383
|
-
readonly "mistralai/mistral-
|
|
4046
|
+
readonly "mistralai/mistral-tiny": {
|
|
3384
4047
|
id: string;
|
|
3385
4048
|
name: string;
|
|
3386
4049
|
api: "openai-completions";
|