@mariozechner/pi-ai 0.5.42 → 0.5.44
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/models.generated.d.ts +453 -232
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +960 -739
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/anthropic.js +24 -1
- package/dist/providers/anthropic.js.map +1 -1
- package/dist/providers/google.js +1 -1
- package/dist/providers/google.js.map +1 -1
- package/dist/providers/openai-responses.js +0 -4
- package/dist/providers/openai-responses.js.map +1 -1
- package/dist/providers/transorm-messages.d.ts.map +1 -1
- package/dist/providers/transorm-messages.js +46 -1
- package/dist/providers/transorm-messages.js.map +1 -1
- package/dist/utils/event-stream.js +3 -7
- package/dist/utils/event-stream.js.map +1 -1
- package/dist/utils/validation.d.ts.map +1 -1
- package/dist/utils/validation.js +25 -3
- package/dist/utils/validation.js.map +1 -1
- package/package.json +3 -3
|
@@ -1,12 +1,12 @@
|
|
|
1
1
|
export declare const MODELS: {
|
|
2
2
|
readonly anthropic: {
|
|
3
|
-
readonly "claude-3-
|
|
3
|
+
readonly "claude-3-5-sonnet-20241022": {
|
|
4
4
|
id: string;
|
|
5
5
|
name: string;
|
|
6
6
|
api: "anthropic-messages";
|
|
7
7
|
provider: string;
|
|
8
8
|
baseUrl: string;
|
|
9
|
-
reasoning:
|
|
9
|
+
reasoning: false;
|
|
10
10
|
input: ("text" | "image")[];
|
|
11
11
|
cost: {
|
|
12
12
|
input: number;
|
|
@@ -17,13 +17,13 @@ export declare const MODELS: {
|
|
|
17
17
|
contextWindow: number;
|
|
18
18
|
maxTokens: number;
|
|
19
19
|
};
|
|
20
|
-
readonly "claude-
|
|
20
|
+
readonly "claude-3-5-sonnet-20240620": {
|
|
21
21
|
id: string;
|
|
22
22
|
name: string;
|
|
23
23
|
api: "anthropic-messages";
|
|
24
24
|
provider: string;
|
|
25
25
|
baseUrl: string;
|
|
26
|
-
reasoning:
|
|
26
|
+
reasoning: false;
|
|
27
27
|
input: ("text" | "image")[];
|
|
28
28
|
cost: {
|
|
29
29
|
input: number;
|
|
@@ -34,7 +34,7 @@ export declare const MODELS: {
|
|
|
34
34
|
contextWindow: number;
|
|
35
35
|
maxTokens: number;
|
|
36
36
|
};
|
|
37
|
-
readonly "claude-3-
|
|
37
|
+
readonly "claude-3-opus-20240229": {
|
|
38
38
|
id: string;
|
|
39
39
|
name: string;
|
|
40
40
|
api: "anthropic-messages";
|
|
@@ -51,13 +51,30 @@ export declare const MODELS: {
|
|
|
51
51
|
contextWindow: number;
|
|
52
52
|
maxTokens: number;
|
|
53
53
|
};
|
|
54
|
-
readonly "claude-
|
|
54
|
+
readonly "claude-sonnet-4-5-20250929": {
|
|
55
55
|
id: string;
|
|
56
56
|
name: string;
|
|
57
57
|
api: "anthropic-messages";
|
|
58
58
|
provider: string;
|
|
59
59
|
baseUrl: string;
|
|
60
|
-
reasoning:
|
|
60
|
+
reasoning: true;
|
|
61
|
+
input: ("text" | "image")[];
|
|
62
|
+
cost: {
|
|
63
|
+
input: number;
|
|
64
|
+
output: number;
|
|
65
|
+
cacheRead: number;
|
|
66
|
+
cacheWrite: number;
|
|
67
|
+
};
|
|
68
|
+
contextWindow: number;
|
|
69
|
+
maxTokens: number;
|
|
70
|
+
};
|
|
71
|
+
readonly "claude-sonnet-4-20250514": {
|
|
72
|
+
id: string;
|
|
73
|
+
name: string;
|
|
74
|
+
api: "anthropic-messages";
|
|
75
|
+
provider: string;
|
|
76
|
+
baseUrl: string;
|
|
77
|
+
reasoning: true;
|
|
61
78
|
input: ("text" | "image")[];
|
|
62
79
|
cost: {
|
|
63
80
|
input: number;
|
|
@@ -85,7 +102,7 @@ export declare const MODELS: {
|
|
|
85
102
|
contextWindow: number;
|
|
86
103
|
maxTokens: number;
|
|
87
104
|
};
|
|
88
|
-
readonly "claude-3-5-
|
|
105
|
+
readonly "claude-3-5-haiku-20241022": {
|
|
89
106
|
id: string;
|
|
90
107
|
name: string;
|
|
91
108
|
api: "anthropic-messages";
|
|
@@ -102,7 +119,7 @@ export declare const MODELS: {
|
|
|
102
119
|
contextWindow: number;
|
|
103
120
|
maxTokens: number;
|
|
104
121
|
};
|
|
105
|
-
readonly "claude-3-
|
|
122
|
+
readonly "claude-3-haiku-20240307": {
|
|
106
123
|
id: string;
|
|
107
124
|
name: string;
|
|
108
125
|
api: "anthropic-messages";
|
|
@@ -119,13 +136,13 @@ export declare const MODELS: {
|
|
|
119
136
|
contextWindow: number;
|
|
120
137
|
maxTokens: number;
|
|
121
138
|
};
|
|
122
|
-
readonly "claude-3-sonnet-
|
|
139
|
+
readonly "claude-3-7-sonnet-20250219": {
|
|
123
140
|
id: string;
|
|
124
141
|
name: string;
|
|
125
142
|
api: "anthropic-messages";
|
|
126
143
|
provider: string;
|
|
127
144
|
baseUrl: string;
|
|
128
|
-
reasoning:
|
|
145
|
+
reasoning: true;
|
|
129
146
|
input: ("text" | "image")[];
|
|
130
147
|
cost: {
|
|
131
148
|
input: number;
|
|
@@ -136,7 +153,7 @@ export declare const MODELS: {
|
|
|
136
153
|
contextWindow: number;
|
|
137
154
|
maxTokens: number;
|
|
138
155
|
};
|
|
139
|
-
readonly "claude-
|
|
156
|
+
readonly "claude-opus-4-1-20250805": {
|
|
140
157
|
id: string;
|
|
141
158
|
name: string;
|
|
142
159
|
api: "anthropic-messages";
|
|
@@ -153,7 +170,7 @@ export declare const MODELS: {
|
|
|
153
170
|
contextWindow: number;
|
|
154
171
|
maxTokens: number;
|
|
155
172
|
};
|
|
156
|
-
readonly "claude-3-
|
|
173
|
+
readonly "claude-3-sonnet-20240229": {
|
|
157
174
|
id: string;
|
|
158
175
|
name: string;
|
|
159
176
|
api: "anthropic-messages";
|
|
@@ -189,7 +206,7 @@ export declare const MODELS: {
|
|
|
189
206
|
contextWindow: number;
|
|
190
207
|
maxTokens: number;
|
|
191
208
|
};
|
|
192
|
-
readonly "gemini-
|
|
209
|
+
readonly "gemini-flash-lite-latest": {
|
|
193
210
|
id: string;
|
|
194
211
|
name: string;
|
|
195
212
|
api: "google-generative-ai";
|
|
@@ -206,13 +223,13 @@ export declare const MODELS: {
|
|
|
206
223
|
contextWindow: number;
|
|
207
224
|
maxTokens: number;
|
|
208
225
|
};
|
|
209
|
-
readonly "gemini-
|
|
226
|
+
readonly "gemini-2.5-flash": {
|
|
210
227
|
id: string;
|
|
211
228
|
name: string;
|
|
212
229
|
api: "google-generative-ai";
|
|
213
230
|
provider: string;
|
|
214
231
|
baseUrl: string;
|
|
215
|
-
reasoning:
|
|
232
|
+
reasoning: true;
|
|
216
233
|
input: ("text" | "image")[];
|
|
217
234
|
cost: {
|
|
218
235
|
input: number;
|
|
@@ -223,13 +240,13 @@ export declare const MODELS: {
|
|
|
223
240
|
contextWindow: number;
|
|
224
241
|
maxTokens: number;
|
|
225
242
|
};
|
|
226
|
-
readonly "gemini-
|
|
243
|
+
readonly "gemini-flash-latest": {
|
|
227
244
|
id: string;
|
|
228
245
|
name: string;
|
|
229
246
|
api: "google-generative-ai";
|
|
230
247
|
provider: string;
|
|
231
248
|
baseUrl: string;
|
|
232
|
-
reasoning:
|
|
249
|
+
reasoning: true;
|
|
233
250
|
input: ("text" | "image")[];
|
|
234
251
|
cost: {
|
|
235
252
|
input: number;
|
|
@@ -240,7 +257,24 @@ export declare const MODELS: {
|
|
|
240
257
|
contextWindow: number;
|
|
241
258
|
maxTokens: number;
|
|
242
259
|
};
|
|
243
|
-
readonly "gemini-
|
|
260
|
+
readonly "gemini-2.5-pro-preview-05-06": {
|
|
261
|
+
id: string;
|
|
262
|
+
name: string;
|
|
263
|
+
api: "google-generative-ai";
|
|
264
|
+
provider: string;
|
|
265
|
+
baseUrl: string;
|
|
266
|
+
reasoning: true;
|
|
267
|
+
input: ("text" | "image")[];
|
|
268
|
+
cost: {
|
|
269
|
+
input: number;
|
|
270
|
+
output: number;
|
|
271
|
+
cacheRead: number;
|
|
272
|
+
cacheWrite: number;
|
|
273
|
+
};
|
|
274
|
+
contextWindow: number;
|
|
275
|
+
maxTokens: number;
|
|
276
|
+
};
|
|
277
|
+
readonly "gemini-2.0-flash-lite": {
|
|
244
278
|
id: string;
|
|
245
279
|
name: string;
|
|
246
280
|
api: "google-generative-ai";
|
|
@@ -257,7 +291,24 @@ export declare const MODELS: {
|
|
|
257
291
|
contextWindow: number;
|
|
258
292
|
maxTokens: number;
|
|
259
293
|
};
|
|
260
|
-
readonly "gemini-
|
|
294
|
+
readonly "gemini-live-2.5-flash-preview-native-audio": {
|
|
295
|
+
id: string;
|
|
296
|
+
name: string;
|
|
297
|
+
api: "google-generative-ai";
|
|
298
|
+
provider: string;
|
|
299
|
+
baseUrl: string;
|
|
300
|
+
reasoning: true;
|
|
301
|
+
input: "text"[];
|
|
302
|
+
cost: {
|
|
303
|
+
input: number;
|
|
304
|
+
output: number;
|
|
305
|
+
cacheRead: number;
|
|
306
|
+
cacheWrite: number;
|
|
307
|
+
};
|
|
308
|
+
contextWindow: number;
|
|
309
|
+
maxTokens: number;
|
|
310
|
+
};
|
|
311
|
+
readonly "gemini-2.0-flash": {
|
|
261
312
|
id: string;
|
|
262
313
|
name: string;
|
|
263
314
|
api: "google-generative-ai";
|
|
@@ -274,7 +325,7 @@ export declare const MODELS: {
|
|
|
274
325
|
contextWindow: number;
|
|
275
326
|
maxTokens: number;
|
|
276
327
|
};
|
|
277
|
-
readonly "gemini-2.5-flash": {
|
|
328
|
+
readonly "gemini-2.5-flash-lite": {
|
|
278
329
|
id: string;
|
|
279
330
|
name: string;
|
|
280
331
|
api: "google-generative-ai";
|
|
@@ -308,7 +359,7 @@ export declare const MODELS: {
|
|
|
308
359
|
contextWindow: number;
|
|
309
360
|
maxTokens: number;
|
|
310
361
|
};
|
|
311
|
-
readonly "gemini-2.5-
|
|
362
|
+
readonly "gemini-2.5-flash-lite-preview-06-17": {
|
|
312
363
|
id: string;
|
|
313
364
|
name: string;
|
|
314
365
|
api: "google-generative-ai";
|
|
@@ -325,13 +376,13 @@ export declare const MODELS: {
|
|
|
325
376
|
contextWindow: number;
|
|
326
377
|
maxTokens: number;
|
|
327
378
|
};
|
|
328
|
-
readonly "gemini-2.
|
|
379
|
+
readonly "gemini-2.5-flash-preview-09-2025": {
|
|
329
380
|
id: string;
|
|
330
381
|
name: string;
|
|
331
382
|
api: "google-generative-ai";
|
|
332
383
|
provider: string;
|
|
333
384
|
baseUrl: string;
|
|
334
|
-
reasoning:
|
|
385
|
+
reasoning: true;
|
|
335
386
|
input: ("text" | "image")[];
|
|
336
387
|
cost: {
|
|
337
388
|
input: number;
|
|
@@ -342,7 +393,7 @@ export declare const MODELS: {
|
|
|
342
393
|
contextWindow: number;
|
|
343
394
|
maxTokens: number;
|
|
344
395
|
};
|
|
345
|
-
readonly "gemini-2.5-flash-
|
|
396
|
+
readonly "gemini-2.5-flash-preview-04-17": {
|
|
346
397
|
id: string;
|
|
347
398
|
name: string;
|
|
348
399
|
api: "google-generative-ai";
|
|
@@ -359,7 +410,58 @@ export declare const MODELS: {
|
|
|
359
410
|
contextWindow: number;
|
|
360
411
|
maxTokens: number;
|
|
361
412
|
};
|
|
362
|
-
readonly "gemini-2.5-
|
|
413
|
+
readonly "gemini-2.5-pro": {
|
|
414
|
+
id: string;
|
|
415
|
+
name: string;
|
|
416
|
+
api: "google-generative-ai";
|
|
417
|
+
provider: string;
|
|
418
|
+
baseUrl: string;
|
|
419
|
+
reasoning: true;
|
|
420
|
+
input: ("text" | "image")[];
|
|
421
|
+
cost: {
|
|
422
|
+
input: number;
|
|
423
|
+
output: number;
|
|
424
|
+
cacheRead: number;
|
|
425
|
+
cacheWrite: number;
|
|
426
|
+
};
|
|
427
|
+
contextWindow: number;
|
|
428
|
+
maxTokens: number;
|
|
429
|
+
};
|
|
430
|
+
readonly "gemini-1.5-flash": {
|
|
431
|
+
id: string;
|
|
432
|
+
name: string;
|
|
433
|
+
api: "google-generative-ai";
|
|
434
|
+
provider: string;
|
|
435
|
+
baseUrl: string;
|
|
436
|
+
reasoning: false;
|
|
437
|
+
input: ("text" | "image")[];
|
|
438
|
+
cost: {
|
|
439
|
+
input: number;
|
|
440
|
+
output: number;
|
|
441
|
+
cacheRead: number;
|
|
442
|
+
cacheWrite: number;
|
|
443
|
+
};
|
|
444
|
+
contextWindow: number;
|
|
445
|
+
maxTokens: number;
|
|
446
|
+
};
|
|
447
|
+
readonly "gemini-1.5-flash-8b": {
|
|
448
|
+
id: string;
|
|
449
|
+
name: string;
|
|
450
|
+
api: "google-generative-ai";
|
|
451
|
+
provider: string;
|
|
452
|
+
baseUrl: string;
|
|
453
|
+
reasoning: false;
|
|
454
|
+
input: ("text" | "image")[];
|
|
455
|
+
cost: {
|
|
456
|
+
input: number;
|
|
457
|
+
output: number;
|
|
458
|
+
cacheRead: number;
|
|
459
|
+
cacheWrite: number;
|
|
460
|
+
};
|
|
461
|
+
contextWindow: number;
|
|
462
|
+
maxTokens: number;
|
|
463
|
+
};
|
|
464
|
+
readonly "gemini-2.5-flash-lite-preview-09-2025": {
|
|
363
465
|
id: string;
|
|
364
466
|
name: string;
|
|
365
467
|
api: "google-generative-ai";
|
|
@@ -376,15 +478,32 @@ export declare const MODELS: {
|
|
|
376
478
|
contextWindow: number;
|
|
377
479
|
maxTokens: number;
|
|
378
480
|
};
|
|
481
|
+
readonly "gemini-1.5-pro": {
|
|
482
|
+
id: string;
|
|
483
|
+
name: string;
|
|
484
|
+
api: "google-generative-ai";
|
|
485
|
+
provider: string;
|
|
486
|
+
baseUrl: string;
|
|
487
|
+
reasoning: false;
|
|
488
|
+
input: ("text" | "image")[];
|
|
489
|
+
cost: {
|
|
490
|
+
input: number;
|
|
491
|
+
output: number;
|
|
492
|
+
cacheRead: number;
|
|
493
|
+
cacheWrite: number;
|
|
494
|
+
};
|
|
495
|
+
contextWindow: number;
|
|
496
|
+
maxTokens: number;
|
|
497
|
+
};
|
|
379
498
|
};
|
|
380
499
|
readonly openai: {
|
|
381
|
-
readonly "gpt-
|
|
500
|
+
readonly "gpt-4.1-nano": {
|
|
382
501
|
id: string;
|
|
383
502
|
name: string;
|
|
384
503
|
api: "openai-responses";
|
|
385
504
|
provider: string;
|
|
386
505
|
baseUrl: string;
|
|
387
|
-
reasoning:
|
|
506
|
+
reasoning: false;
|
|
388
507
|
input: ("text" | "image")[];
|
|
389
508
|
cost: {
|
|
390
509
|
input: number;
|
|
@@ -395,14 +514,14 @@ export declare const MODELS: {
|
|
|
395
514
|
contextWindow: number;
|
|
396
515
|
maxTokens: number;
|
|
397
516
|
};
|
|
398
|
-
readonly "
|
|
517
|
+
readonly "gpt-4": {
|
|
399
518
|
id: string;
|
|
400
519
|
name: string;
|
|
401
520
|
api: "openai-responses";
|
|
402
521
|
provider: string;
|
|
403
522
|
baseUrl: string;
|
|
404
|
-
reasoning:
|
|
405
|
-
input:
|
|
523
|
+
reasoning: false;
|
|
524
|
+
input: "text"[];
|
|
406
525
|
cost: {
|
|
407
526
|
input: number;
|
|
408
527
|
output: number;
|
|
@@ -412,14 +531,14 @@ export declare const MODELS: {
|
|
|
412
531
|
contextWindow: number;
|
|
413
532
|
maxTokens: number;
|
|
414
533
|
};
|
|
415
|
-
readonly "
|
|
534
|
+
readonly "o1-pro": {
|
|
416
535
|
id: string;
|
|
417
536
|
name: string;
|
|
418
537
|
api: "openai-responses";
|
|
419
538
|
provider: string;
|
|
420
539
|
baseUrl: string;
|
|
421
540
|
reasoning: true;
|
|
422
|
-
input: "text"[];
|
|
541
|
+
input: ("text" | "image")[];
|
|
423
542
|
cost: {
|
|
424
543
|
input: number;
|
|
425
544
|
output: number;
|
|
@@ -429,7 +548,7 @@ export declare const MODELS: {
|
|
|
429
548
|
contextWindow: number;
|
|
430
549
|
maxTokens: number;
|
|
431
550
|
};
|
|
432
|
-
readonly "gpt-
|
|
551
|
+
readonly "gpt-4o-2024-05-13": {
|
|
433
552
|
id: string;
|
|
434
553
|
name: string;
|
|
435
554
|
api: "openai-responses";
|
|
@@ -446,7 +565,7 @@ export declare const MODELS: {
|
|
|
446
565
|
contextWindow: number;
|
|
447
566
|
maxTokens: number;
|
|
448
567
|
};
|
|
449
|
-
readonly "gpt-
|
|
568
|
+
readonly "gpt-4o-2024-08-06": {
|
|
450
569
|
id: string;
|
|
451
570
|
name: string;
|
|
452
571
|
api: "openai-responses";
|
|
@@ -463,13 +582,13 @@ export declare const MODELS: {
|
|
|
463
582
|
contextWindow: number;
|
|
464
583
|
maxTokens: number;
|
|
465
584
|
};
|
|
466
|
-
readonly
|
|
585
|
+
readonly "gpt-4.1-mini": {
|
|
467
586
|
id: string;
|
|
468
587
|
name: string;
|
|
469
588
|
api: "openai-responses";
|
|
470
589
|
provider: string;
|
|
471
590
|
baseUrl: string;
|
|
472
|
-
reasoning:
|
|
591
|
+
reasoning: false;
|
|
473
592
|
input: ("text" | "image")[];
|
|
474
593
|
cost: {
|
|
475
594
|
input: number;
|
|
@@ -497,13 +616,13 @@ export declare const MODELS: {
|
|
|
497
616
|
contextWindow: number;
|
|
498
617
|
maxTokens: number;
|
|
499
618
|
};
|
|
500
|
-
readonly "gpt-
|
|
619
|
+
readonly "gpt-4-turbo": {
|
|
501
620
|
id: string;
|
|
502
621
|
name: string;
|
|
503
622
|
api: "openai-responses";
|
|
504
623
|
provider: string;
|
|
505
624
|
baseUrl: string;
|
|
506
|
-
reasoning:
|
|
625
|
+
reasoning: false;
|
|
507
626
|
input: ("text" | "image")[];
|
|
508
627
|
cost: {
|
|
509
628
|
input: number;
|
|
@@ -514,14 +633,14 @@ export declare const MODELS: {
|
|
|
514
633
|
contextWindow: number;
|
|
515
634
|
maxTokens: number;
|
|
516
635
|
};
|
|
517
|
-
readonly "
|
|
636
|
+
readonly "o3-mini": {
|
|
518
637
|
id: string;
|
|
519
638
|
name: string;
|
|
520
639
|
api: "openai-responses";
|
|
521
640
|
provider: string;
|
|
522
641
|
baseUrl: string;
|
|
523
642
|
reasoning: true;
|
|
524
|
-
input:
|
|
643
|
+
input: "text"[];
|
|
525
644
|
cost: {
|
|
526
645
|
input: number;
|
|
527
646
|
output: number;
|
|
@@ -531,14 +650,14 @@ export declare const MODELS: {
|
|
|
531
650
|
contextWindow: number;
|
|
532
651
|
maxTokens: number;
|
|
533
652
|
};
|
|
534
|
-
readonly
|
|
653
|
+
readonly "codex-mini-latest": {
|
|
535
654
|
id: string;
|
|
536
655
|
name: string;
|
|
537
656
|
api: "openai-responses";
|
|
538
657
|
provider: string;
|
|
539
658
|
baseUrl: string;
|
|
540
659
|
reasoning: true;
|
|
541
|
-
input:
|
|
660
|
+
input: "text"[];
|
|
542
661
|
cost: {
|
|
543
662
|
input: number;
|
|
544
663
|
output: number;
|
|
@@ -548,7 +667,7 @@ export declare const MODELS: {
|
|
|
548
667
|
contextWindow: number;
|
|
549
668
|
maxTokens: number;
|
|
550
669
|
};
|
|
551
|
-
readonly "gpt-5-
|
|
670
|
+
readonly "gpt-5-nano": {
|
|
552
671
|
id: string;
|
|
553
672
|
name: string;
|
|
554
673
|
api: "openai-responses";
|
|
@@ -565,7 +684,7 @@ export declare const MODELS: {
|
|
|
565
684
|
contextWindow: number;
|
|
566
685
|
maxTokens: number;
|
|
567
686
|
};
|
|
568
|
-
readonly "
|
|
687
|
+
readonly "gpt-5-codex": {
|
|
569
688
|
id: string;
|
|
570
689
|
name: string;
|
|
571
690
|
api: "openai-responses";
|
|
@@ -582,7 +701,7 @@ export declare const MODELS: {
|
|
|
582
701
|
contextWindow: number;
|
|
583
702
|
maxTokens: number;
|
|
584
703
|
};
|
|
585
|
-
readonly "gpt-4o
|
|
704
|
+
readonly "gpt-4o": {
|
|
586
705
|
id: string;
|
|
587
706
|
name: string;
|
|
588
707
|
api: "openai-responses";
|
|
@@ -599,7 +718,7 @@ export declare const MODELS: {
|
|
|
599
718
|
contextWindow: number;
|
|
600
719
|
maxTokens: number;
|
|
601
720
|
};
|
|
602
|
-
readonly "gpt-4.1
|
|
721
|
+
readonly "gpt-4.1": {
|
|
603
722
|
id: string;
|
|
604
723
|
name: string;
|
|
605
724
|
api: "openai-responses";
|
|
@@ -616,13 +735,13 @@ export declare const MODELS: {
|
|
|
616
735
|
contextWindow: number;
|
|
617
736
|
maxTokens: number;
|
|
618
737
|
};
|
|
619
|
-
readonly "
|
|
738
|
+
readonly "o4-mini": {
|
|
620
739
|
id: string;
|
|
621
740
|
name: string;
|
|
622
741
|
api: "openai-responses";
|
|
623
742
|
provider: string;
|
|
624
743
|
baseUrl: string;
|
|
625
|
-
reasoning:
|
|
744
|
+
reasoning: true;
|
|
626
745
|
input: ("text" | "image")[];
|
|
627
746
|
cost: {
|
|
628
747
|
input: number;
|
|
@@ -633,13 +752,13 @@ export declare const MODELS: {
|
|
|
633
752
|
contextWindow: number;
|
|
634
753
|
maxTokens: number;
|
|
635
754
|
};
|
|
636
|
-
readonly
|
|
755
|
+
readonly o1: {
|
|
637
756
|
id: string;
|
|
638
757
|
name: string;
|
|
639
758
|
api: "openai-responses";
|
|
640
759
|
provider: string;
|
|
641
760
|
baseUrl: string;
|
|
642
|
-
reasoning:
|
|
761
|
+
reasoning: true;
|
|
643
762
|
input: ("text" | "image")[];
|
|
644
763
|
cost: {
|
|
645
764
|
input: number;
|
|
@@ -650,14 +769,48 @@ export declare const MODELS: {
|
|
|
650
769
|
contextWindow: number;
|
|
651
770
|
maxTokens: number;
|
|
652
771
|
};
|
|
653
|
-
readonly "gpt-
|
|
772
|
+
readonly "gpt-5-mini": {
|
|
773
|
+
id: string;
|
|
774
|
+
name: string;
|
|
775
|
+
api: "openai-responses";
|
|
776
|
+
provider: string;
|
|
777
|
+
baseUrl: string;
|
|
778
|
+
reasoning: true;
|
|
779
|
+
input: ("text" | "image")[];
|
|
780
|
+
cost: {
|
|
781
|
+
input: number;
|
|
782
|
+
output: number;
|
|
783
|
+
cacheRead: number;
|
|
784
|
+
cacheWrite: number;
|
|
785
|
+
};
|
|
786
|
+
contextWindow: number;
|
|
787
|
+
maxTokens: number;
|
|
788
|
+
};
|
|
789
|
+
readonly "o3-pro": {
|
|
790
|
+
id: string;
|
|
791
|
+
name: string;
|
|
792
|
+
api: "openai-responses";
|
|
793
|
+
provider: string;
|
|
794
|
+
baseUrl: string;
|
|
795
|
+
reasoning: true;
|
|
796
|
+
input: ("text" | "image")[];
|
|
797
|
+
cost: {
|
|
798
|
+
input: number;
|
|
799
|
+
output: number;
|
|
800
|
+
cacheRead: number;
|
|
801
|
+
cacheWrite: number;
|
|
802
|
+
};
|
|
803
|
+
contextWindow: number;
|
|
804
|
+
maxTokens: number;
|
|
805
|
+
};
|
|
806
|
+
readonly "gpt-4o-2024-11-20": {
|
|
654
807
|
id: string;
|
|
655
808
|
name: string;
|
|
656
809
|
api: "openai-responses";
|
|
657
810
|
provider: string;
|
|
658
811
|
baseUrl: string;
|
|
659
812
|
reasoning: false;
|
|
660
|
-
input: "text"[];
|
|
813
|
+
input: ("text" | "image")[];
|
|
661
814
|
cost: {
|
|
662
815
|
input: number;
|
|
663
816
|
output: number;
|
|
@@ -667,7 +820,7 @@ export declare const MODELS: {
|
|
|
667
820
|
contextWindow: number;
|
|
668
821
|
maxTokens: number;
|
|
669
822
|
};
|
|
670
|
-
readonly
|
|
823
|
+
readonly o3: {
|
|
671
824
|
id: string;
|
|
672
825
|
name: string;
|
|
673
826
|
api: "openai-responses";
|
|
@@ -684,14 +837,48 @@ export declare const MODELS: {
|
|
|
684
837
|
contextWindow: number;
|
|
685
838
|
maxTokens: number;
|
|
686
839
|
};
|
|
687
|
-
readonly "
|
|
840
|
+
readonly "o4-mini-deep-research": {
|
|
688
841
|
id: string;
|
|
689
842
|
name: string;
|
|
690
843
|
api: "openai-responses";
|
|
691
844
|
provider: string;
|
|
692
845
|
baseUrl: string;
|
|
693
846
|
reasoning: true;
|
|
694
|
-
input: "text"[];
|
|
847
|
+
input: ("text" | "image")[];
|
|
848
|
+
cost: {
|
|
849
|
+
input: number;
|
|
850
|
+
output: number;
|
|
851
|
+
cacheRead: number;
|
|
852
|
+
cacheWrite: number;
|
|
853
|
+
};
|
|
854
|
+
contextWindow: number;
|
|
855
|
+
maxTokens: number;
|
|
856
|
+
};
|
|
857
|
+
readonly "gpt-4o-mini": {
|
|
858
|
+
id: string;
|
|
859
|
+
name: string;
|
|
860
|
+
api: "openai-responses";
|
|
861
|
+
provider: string;
|
|
862
|
+
baseUrl: string;
|
|
863
|
+
reasoning: false;
|
|
864
|
+
input: ("text" | "image")[];
|
|
865
|
+
cost: {
|
|
866
|
+
input: number;
|
|
867
|
+
output: number;
|
|
868
|
+
cacheRead: number;
|
|
869
|
+
cacheWrite: number;
|
|
870
|
+
};
|
|
871
|
+
contextWindow: number;
|
|
872
|
+
maxTokens: number;
|
|
873
|
+
};
|
|
874
|
+
readonly "gpt-5": {
|
|
875
|
+
id: string;
|
|
876
|
+
name: string;
|
|
877
|
+
api: "openai-responses";
|
|
878
|
+
provider: string;
|
|
879
|
+
baseUrl: string;
|
|
880
|
+
reasoning: true;
|
|
881
|
+
input: ("text" | "image")[];
|
|
695
882
|
cost: {
|
|
696
883
|
input: number;
|
|
697
884
|
output: number;
|
|
@@ -737,6 +924,40 @@ export declare const MODELS: {
|
|
|
737
924
|
contextWindow: number;
|
|
738
925
|
maxTokens: number;
|
|
739
926
|
};
|
|
927
|
+
readonly "mistral-saba-24b": {
|
|
928
|
+
id: string;
|
|
929
|
+
name: string;
|
|
930
|
+
api: "openai-completions";
|
|
931
|
+
provider: string;
|
|
932
|
+
baseUrl: string;
|
|
933
|
+
reasoning: false;
|
|
934
|
+
input: "text"[];
|
|
935
|
+
cost: {
|
|
936
|
+
input: number;
|
|
937
|
+
output: number;
|
|
938
|
+
cacheRead: number;
|
|
939
|
+
cacheWrite: number;
|
|
940
|
+
};
|
|
941
|
+
contextWindow: number;
|
|
942
|
+
maxTokens: number;
|
|
943
|
+
};
|
|
944
|
+
readonly "llama3-8b-8192": {
|
|
945
|
+
id: string;
|
|
946
|
+
name: string;
|
|
947
|
+
api: "openai-completions";
|
|
948
|
+
provider: string;
|
|
949
|
+
baseUrl: string;
|
|
950
|
+
reasoning: false;
|
|
951
|
+
input: "text"[];
|
|
952
|
+
cost: {
|
|
953
|
+
input: number;
|
|
954
|
+
output: number;
|
|
955
|
+
cacheRead: number;
|
|
956
|
+
cacheWrite: number;
|
|
957
|
+
};
|
|
958
|
+
contextWindow: number;
|
|
959
|
+
maxTokens: number;
|
|
960
|
+
};
|
|
740
961
|
readonly "qwen-qwq-32b": {
|
|
741
962
|
id: string;
|
|
742
963
|
name: string;
|
|
@@ -788,7 +1009,7 @@ export declare const MODELS: {
|
|
|
788
1009
|
contextWindow: number;
|
|
789
1010
|
maxTokens: number;
|
|
790
1011
|
};
|
|
791
|
-
readonly "
|
|
1012
|
+
readonly "gemma2-9b-it": {
|
|
792
1013
|
id: string;
|
|
793
1014
|
name: string;
|
|
794
1015
|
api: "openai-completions";
|
|
@@ -805,7 +1026,7 @@ export declare const MODELS: {
|
|
|
805
1026
|
contextWindow: number;
|
|
806
1027
|
maxTokens: number;
|
|
807
1028
|
};
|
|
808
|
-
readonly "
|
|
1029
|
+
readonly "llama-3.3-70b-versatile": {
|
|
809
1030
|
id: string;
|
|
810
1031
|
name: string;
|
|
811
1032
|
api: "openai-completions";
|
|
@@ -822,7 +1043,7 @@ export declare const MODELS: {
|
|
|
822
1043
|
contextWindow: number;
|
|
823
1044
|
maxTokens: number;
|
|
824
1045
|
};
|
|
825
|
-
readonly "
|
|
1046
|
+
readonly "moonshotai/kimi-k2-instruct-0905": {
|
|
826
1047
|
id: string;
|
|
827
1048
|
name: string;
|
|
828
1049
|
api: "openai-completions";
|
|
@@ -839,7 +1060,7 @@ export declare const MODELS: {
|
|
|
839
1060
|
contextWindow: number;
|
|
840
1061
|
maxTokens: number;
|
|
841
1062
|
};
|
|
842
|
-
readonly "
|
|
1063
|
+
readonly "moonshotai/kimi-k2-instruct": {
|
|
843
1064
|
id: string;
|
|
844
1065
|
name: string;
|
|
845
1066
|
api: "openai-completions";
|
|
@@ -890,6 +1111,40 @@ export declare const MODELS: {
|
|
|
890
1111
|
contextWindow: number;
|
|
891
1112
|
maxTokens: number;
|
|
892
1113
|
};
|
|
1114
|
+
readonly "qwen/qwen3-32b": {
|
|
1115
|
+
id: string;
|
|
1116
|
+
name: string;
|
|
1117
|
+
api: "openai-completions";
|
|
1118
|
+
provider: string;
|
|
1119
|
+
baseUrl: string;
|
|
1120
|
+
reasoning: true;
|
|
1121
|
+
input: "text"[];
|
|
1122
|
+
cost: {
|
|
1123
|
+
input: number;
|
|
1124
|
+
output: number;
|
|
1125
|
+
cacheRead: number;
|
|
1126
|
+
cacheWrite: number;
|
|
1127
|
+
};
|
|
1128
|
+
contextWindow: number;
|
|
1129
|
+
maxTokens: number;
|
|
1130
|
+
};
|
|
1131
|
+
readonly "meta-llama/llama-4-scout-17b-16e-instruct": {
|
|
1132
|
+
id: string;
|
|
1133
|
+
name: string;
|
|
1134
|
+
api: "openai-completions";
|
|
1135
|
+
provider: string;
|
|
1136
|
+
baseUrl: string;
|
|
1137
|
+
reasoning: false;
|
|
1138
|
+
input: ("text" | "image")[];
|
|
1139
|
+
cost: {
|
|
1140
|
+
input: number;
|
|
1141
|
+
output: number;
|
|
1142
|
+
cacheRead: number;
|
|
1143
|
+
cacheWrite: number;
|
|
1144
|
+
};
|
|
1145
|
+
contextWindow: number;
|
|
1146
|
+
maxTokens: number;
|
|
1147
|
+
};
|
|
893
1148
|
readonly "meta-llama/llama-4-maverick-17b-128e-instruct": {
|
|
894
1149
|
id: string;
|
|
895
1150
|
name: string;
|
|
@@ -897,7 +1152,43 @@ export declare const MODELS: {
|
|
|
897
1152
|
provider: string;
|
|
898
1153
|
baseUrl: string;
|
|
899
1154
|
reasoning: false;
|
|
900
|
-
input: ("text" | "image")[];
|
|
1155
|
+
input: ("text" | "image")[];
|
|
1156
|
+
cost: {
|
|
1157
|
+
input: number;
|
|
1158
|
+
output: number;
|
|
1159
|
+
cacheRead: number;
|
|
1160
|
+
cacheWrite: number;
|
|
1161
|
+
};
|
|
1162
|
+
contextWindow: number;
|
|
1163
|
+
maxTokens: number;
|
|
1164
|
+
};
|
|
1165
|
+
};
|
|
1166
|
+
readonly cerebras: {
|
|
1167
|
+
readonly "qwen-3-235b-a22b-instruct-2507": {
|
|
1168
|
+
id: string;
|
|
1169
|
+
name: string;
|
|
1170
|
+
api: "openai-completions";
|
|
1171
|
+
provider: string;
|
|
1172
|
+
baseUrl: string;
|
|
1173
|
+
reasoning: false;
|
|
1174
|
+
input: "text"[];
|
|
1175
|
+
cost: {
|
|
1176
|
+
input: number;
|
|
1177
|
+
output: number;
|
|
1178
|
+
cacheRead: number;
|
|
1179
|
+
cacheWrite: number;
|
|
1180
|
+
};
|
|
1181
|
+
contextWindow: number;
|
|
1182
|
+
maxTokens: number;
|
|
1183
|
+
};
|
|
1184
|
+
readonly "qwen-3-coder-480b": {
|
|
1185
|
+
id: string;
|
|
1186
|
+
name: string;
|
|
1187
|
+
api: "openai-completions";
|
|
1188
|
+
provider: string;
|
|
1189
|
+
baseUrl: string;
|
|
1190
|
+
reasoning: false;
|
|
1191
|
+
input: "text"[];
|
|
901
1192
|
cost: {
|
|
902
1193
|
input: number;
|
|
903
1194
|
output: number;
|
|
@@ -907,14 +1198,14 @@ export declare const MODELS: {
|
|
|
907
1198
|
contextWindow: number;
|
|
908
1199
|
maxTokens: number;
|
|
909
1200
|
};
|
|
910
|
-
readonly "
|
|
1201
|
+
readonly "gpt-oss-120b": {
|
|
911
1202
|
id: string;
|
|
912
1203
|
name: string;
|
|
913
1204
|
api: "openai-completions";
|
|
914
1205
|
provider: string;
|
|
915
1206
|
baseUrl: string;
|
|
916
|
-
reasoning:
|
|
917
|
-
input:
|
|
1207
|
+
reasoning: true;
|
|
1208
|
+
input: "text"[];
|
|
918
1209
|
cost: {
|
|
919
1210
|
input: number;
|
|
920
1211
|
output: number;
|
|
@@ -924,14 +1215,16 @@ export declare const MODELS: {
|
|
|
924
1215
|
contextWindow: number;
|
|
925
1216
|
maxTokens: number;
|
|
926
1217
|
};
|
|
927
|
-
|
|
1218
|
+
};
|
|
1219
|
+
readonly xai: {
|
|
1220
|
+
readonly "grok-4-fast-non-reasoning": {
|
|
928
1221
|
id: string;
|
|
929
1222
|
name: string;
|
|
930
1223
|
api: "openai-completions";
|
|
931
1224
|
provider: string;
|
|
932
1225
|
baseUrl: string;
|
|
933
|
-
reasoning:
|
|
934
|
-
input: "text"[];
|
|
1226
|
+
reasoning: false;
|
|
1227
|
+
input: ("text" | "image")[];
|
|
935
1228
|
cost: {
|
|
936
1229
|
input: number;
|
|
937
1230
|
output: number;
|
|
@@ -941,7 +1234,7 @@ export declare const MODELS: {
|
|
|
941
1234
|
contextWindow: number;
|
|
942
1235
|
maxTokens: number;
|
|
943
1236
|
};
|
|
944
|
-
readonly "
|
|
1237
|
+
readonly "grok-3-fast": {
|
|
945
1238
|
id: string;
|
|
946
1239
|
name: string;
|
|
947
1240
|
api: "openai-completions";
|
|
@@ -958,13 +1251,13 @@ export declare const MODELS: {
|
|
|
958
1251
|
contextWindow: number;
|
|
959
1252
|
maxTokens: number;
|
|
960
1253
|
};
|
|
961
|
-
readonly "
|
|
1254
|
+
readonly "grok-4": {
|
|
962
1255
|
id: string;
|
|
963
1256
|
name: string;
|
|
964
1257
|
api: "openai-completions";
|
|
965
1258
|
provider: string;
|
|
966
1259
|
baseUrl: string;
|
|
967
|
-
reasoning:
|
|
1260
|
+
reasoning: true;
|
|
968
1261
|
input: "text"[];
|
|
969
1262
|
cost: {
|
|
970
1263
|
input: number;
|
|
@@ -975,16 +1268,14 @@ export declare const MODELS: {
|
|
|
975
1268
|
contextWindow: number;
|
|
976
1269
|
maxTokens: number;
|
|
977
1270
|
};
|
|
978
|
-
|
|
979
|
-
readonly cerebras: {
|
|
980
|
-
readonly "qwen-3-235b-a22b-instruct-2507": {
|
|
1271
|
+
readonly "grok-2-vision": {
|
|
981
1272
|
id: string;
|
|
982
1273
|
name: string;
|
|
983
1274
|
api: "openai-completions";
|
|
984
1275
|
provider: string;
|
|
985
1276
|
baseUrl: string;
|
|
986
1277
|
reasoning: false;
|
|
987
|
-
input: "text"[];
|
|
1278
|
+
input: ("text" | "image")[];
|
|
988
1279
|
cost: {
|
|
989
1280
|
input: number;
|
|
990
1281
|
output: number;
|
|
@@ -994,7 +1285,7 @@ export declare const MODELS: {
|
|
|
994
1285
|
contextWindow: number;
|
|
995
1286
|
maxTokens: number;
|
|
996
1287
|
};
|
|
997
|
-
readonly "
|
|
1288
|
+
readonly "grok-code-fast-1": {
|
|
998
1289
|
id: string;
|
|
999
1290
|
name: string;
|
|
1000
1291
|
api: "openai-completions";
|
|
@@ -1011,7 +1302,7 @@ export declare const MODELS: {
|
|
|
1011
1302
|
contextWindow: number;
|
|
1012
1303
|
maxTokens: number;
|
|
1013
1304
|
};
|
|
1014
|
-
readonly "
|
|
1305
|
+
readonly "grok-2": {
|
|
1015
1306
|
id: string;
|
|
1016
1307
|
name: string;
|
|
1017
1308
|
api: "openai-completions";
|
|
@@ -1028,8 +1319,6 @@ export declare const MODELS: {
|
|
|
1028
1319
|
contextWindow: number;
|
|
1029
1320
|
maxTokens: number;
|
|
1030
1321
|
};
|
|
1031
|
-
};
|
|
1032
|
-
readonly xai: {
|
|
1033
1322
|
readonly "grok-3-mini-fast-latest": {
|
|
1034
1323
|
id: string;
|
|
1035
1324
|
name: string;
|
|
@@ -1047,14 +1336,14 @@ export declare const MODELS: {
|
|
|
1047
1336
|
contextWindow: number;
|
|
1048
1337
|
maxTokens: number;
|
|
1049
1338
|
};
|
|
1050
|
-
readonly "grok-
|
|
1339
|
+
readonly "grok-2-vision-1212": {
|
|
1051
1340
|
id: string;
|
|
1052
1341
|
name: string;
|
|
1053
1342
|
api: "openai-completions";
|
|
1054
1343
|
provider: string;
|
|
1055
1344
|
baseUrl: string;
|
|
1056
|
-
reasoning:
|
|
1057
|
-
input: "text"[];
|
|
1345
|
+
reasoning: false;
|
|
1346
|
+
input: ("text" | "image")[];
|
|
1058
1347
|
cost: {
|
|
1059
1348
|
input: number;
|
|
1060
1349
|
output: number;
|
|
@@ -1064,7 +1353,7 @@ export declare const MODELS: {
|
|
|
1064
1353
|
contextWindow: number;
|
|
1065
1354
|
maxTokens: number;
|
|
1066
1355
|
};
|
|
1067
|
-
readonly "grok-
|
|
1356
|
+
readonly "grok-3": {
|
|
1068
1357
|
id: string;
|
|
1069
1358
|
name: string;
|
|
1070
1359
|
api: "openai-completions";
|
|
@@ -1081,14 +1370,14 @@ export declare const MODELS: {
|
|
|
1081
1370
|
contextWindow: number;
|
|
1082
1371
|
maxTokens: number;
|
|
1083
1372
|
};
|
|
1084
|
-
readonly "grok-
|
|
1373
|
+
readonly "grok-4-fast": {
|
|
1085
1374
|
id: string;
|
|
1086
1375
|
name: string;
|
|
1087
1376
|
api: "openai-completions";
|
|
1088
1377
|
provider: string;
|
|
1089
1378
|
baseUrl: string;
|
|
1090
|
-
reasoning:
|
|
1091
|
-
input: "text"[];
|
|
1379
|
+
reasoning: true;
|
|
1380
|
+
input: ("text" | "image")[];
|
|
1092
1381
|
cost: {
|
|
1093
1382
|
input: number;
|
|
1094
1383
|
output: number;
|
|
@@ -1098,7 +1387,7 @@ export declare const MODELS: {
|
|
|
1098
1387
|
contextWindow: number;
|
|
1099
1388
|
maxTokens: number;
|
|
1100
1389
|
};
|
|
1101
|
-
readonly "grok-
|
|
1390
|
+
readonly "grok-2-latest": {
|
|
1102
1391
|
id: string;
|
|
1103
1392
|
name: string;
|
|
1104
1393
|
api: "openai-completions";
|
|
@@ -1115,13 +1404,13 @@ export declare const MODELS: {
|
|
|
1115
1404
|
contextWindow: number;
|
|
1116
1405
|
maxTokens: number;
|
|
1117
1406
|
};
|
|
1118
|
-
readonly "grok-
|
|
1407
|
+
readonly "grok-2-1212": {
|
|
1119
1408
|
id: string;
|
|
1120
1409
|
name: string;
|
|
1121
1410
|
api: "openai-completions";
|
|
1122
1411
|
provider: string;
|
|
1123
1412
|
baseUrl: string;
|
|
1124
|
-
reasoning:
|
|
1413
|
+
reasoning: false;
|
|
1125
1414
|
input: "text"[];
|
|
1126
1415
|
cost: {
|
|
1127
1416
|
input: number;
|
|
@@ -1132,14 +1421,14 @@ export declare const MODELS: {
|
|
|
1132
1421
|
contextWindow: number;
|
|
1133
1422
|
maxTokens: number;
|
|
1134
1423
|
};
|
|
1135
|
-
readonly "grok-
|
|
1424
|
+
readonly "grok-3-fast-latest": {
|
|
1136
1425
|
id: string;
|
|
1137
1426
|
name: string;
|
|
1138
1427
|
api: "openai-completions";
|
|
1139
1428
|
provider: string;
|
|
1140
1429
|
baseUrl: string;
|
|
1141
1430
|
reasoning: false;
|
|
1142
|
-
input:
|
|
1431
|
+
input: "text"[];
|
|
1143
1432
|
cost: {
|
|
1144
1433
|
input: number;
|
|
1145
1434
|
output: number;
|
|
@@ -1149,7 +1438,7 @@ export declare const MODELS: {
|
|
|
1149
1438
|
contextWindow: number;
|
|
1150
1439
|
maxTokens: number;
|
|
1151
1440
|
};
|
|
1152
|
-
readonly "grok-
|
|
1441
|
+
readonly "grok-3-latest": {
|
|
1153
1442
|
id: string;
|
|
1154
1443
|
name: string;
|
|
1155
1444
|
api: "openai-completions";
|
|
@@ -1183,14 +1472,14 @@ export declare const MODELS: {
|
|
|
1183
1472
|
contextWindow: number;
|
|
1184
1473
|
maxTokens: number;
|
|
1185
1474
|
};
|
|
1186
|
-
readonly "grok-
|
|
1475
|
+
readonly "grok-vision-beta": {
|
|
1187
1476
|
id: string;
|
|
1188
1477
|
name: string;
|
|
1189
1478
|
api: "openai-completions";
|
|
1190
1479
|
provider: string;
|
|
1191
1480
|
baseUrl: string;
|
|
1192
1481
|
reasoning: false;
|
|
1193
|
-
input: "text"[];
|
|
1482
|
+
input: ("text" | "image")[];
|
|
1194
1483
|
cost: {
|
|
1195
1484
|
input: number;
|
|
1196
1485
|
output: number;
|
|
@@ -1200,14 +1489,14 @@ export declare const MODELS: {
|
|
|
1200
1489
|
contextWindow: number;
|
|
1201
1490
|
maxTokens: number;
|
|
1202
1491
|
};
|
|
1203
|
-
readonly "grok-
|
|
1492
|
+
readonly "grok-3-mini": {
|
|
1204
1493
|
id: string;
|
|
1205
1494
|
name: string;
|
|
1206
1495
|
api: "openai-completions";
|
|
1207
1496
|
provider: string;
|
|
1208
1497
|
baseUrl: string;
|
|
1209
|
-
reasoning:
|
|
1210
|
-
input:
|
|
1498
|
+
reasoning: true;
|
|
1499
|
+
input: "text"[];
|
|
1211
1500
|
cost: {
|
|
1212
1501
|
input: number;
|
|
1213
1502
|
output: number;
|
|
@@ -1217,7 +1506,7 @@ export declare const MODELS: {
|
|
|
1217
1506
|
contextWindow: number;
|
|
1218
1507
|
maxTokens: number;
|
|
1219
1508
|
};
|
|
1220
|
-
readonly "grok-
|
|
1509
|
+
readonly "grok-beta": {
|
|
1221
1510
|
id: string;
|
|
1222
1511
|
name: string;
|
|
1223
1512
|
api: "openai-completions";
|
|
@@ -1234,13 +1523,13 @@ export declare const MODELS: {
|
|
|
1234
1523
|
contextWindow: number;
|
|
1235
1524
|
maxTokens: number;
|
|
1236
1525
|
};
|
|
1237
|
-
readonly "grok-3-
|
|
1526
|
+
readonly "grok-3-mini-latest": {
|
|
1238
1527
|
id: string;
|
|
1239
1528
|
name: string;
|
|
1240
1529
|
api: "openai-completions";
|
|
1241
1530
|
provider: string;
|
|
1242
1531
|
baseUrl: string;
|
|
1243
|
-
reasoning:
|
|
1532
|
+
reasoning: true;
|
|
1244
1533
|
input: "text"[];
|
|
1245
1534
|
cost: {
|
|
1246
1535
|
input: number;
|
|
@@ -1251,13 +1540,13 @@ export declare const MODELS: {
|
|
|
1251
1540
|
contextWindow: number;
|
|
1252
1541
|
maxTokens: number;
|
|
1253
1542
|
};
|
|
1254
|
-
readonly "grok-
|
|
1543
|
+
readonly "grok-3-mini-fast": {
|
|
1255
1544
|
id: string;
|
|
1256
1545
|
name: string;
|
|
1257
1546
|
api: "openai-completions";
|
|
1258
1547
|
provider: string;
|
|
1259
1548
|
baseUrl: string;
|
|
1260
|
-
reasoning:
|
|
1549
|
+
reasoning: true;
|
|
1261
1550
|
input: "text"[];
|
|
1262
1551
|
cost: {
|
|
1263
1552
|
input: number;
|
|
@@ -1268,10 +1557,12 @@ export declare const MODELS: {
|
|
|
1268
1557
|
contextWindow: number;
|
|
1269
1558
|
maxTokens: number;
|
|
1270
1559
|
};
|
|
1271
|
-
|
|
1560
|
+
};
|
|
1561
|
+
readonly zai: {
|
|
1562
|
+
readonly "glm-4.5-flash": {
|
|
1272
1563
|
id: string;
|
|
1273
1564
|
name: string;
|
|
1274
|
-
api: "
|
|
1565
|
+
api: "anthropic-messages";
|
|
1275
1566
|
provider: string;
|
|
1276
1567
|
baseUrl: string;
|
|
1277
1568
|
reasoning: true;
|
|
@@ -1285,14 +1576,14 @@ export declare const MODELS: {
|
|
|
1285
1576
|
contextWindow: number;
|
|
1286
1577
|
maxTokens: number;
|
|
1287
1578
|
};
|
|
1288
|
-
readonly "
|
|
1579
|
+
readonly "glm-4.5": {
|
|
1289
1580
|
id: string;
|
|
1290
1581
|
name: string;
|
|
1291
|
-
api: "
|
|
1582
|
+
api: "anthropic-messages";
|
|
1292
1583
|
provider: string;
|
|
1293
1584
|
baseUrl: string;
|
|
1294
|
-
reasoning:
|
|
1295
|
-
input:
|
|
1585
|
+
reasoning: true;
|
|
1586
|
+
input: "text"[];
|
|
1296
1587
|
cost: {
|
|
1297
1588
|
input: number;
|
|
1298
1589
|
output: number;
|
|
@@ -1302,10 +1593,10 @@ export declare const MODELS: {
|
|
|
1302
1593
|
contextWindow: number;
|
|
1303
1594
|
maxTokens: number;
|
|
1304
1595
|
};
|
|
1305
|
-
readonly "
|
|
1596
|
+
readonly "glm-4.5-air": {
|
|
1306
1597
|
id: string;
|
|
1307
1598
|
name: string;
|
|
1308
|
-
api: "
|
|
1599
|
+
api: "anthropic-messages";
|
|
1309
1600
|
provider: string;
|
|
1310
1601
|
baseUrl: string;
|
|
1311
1602
|
reasoning: true;
|
|
@@ -1319,14 +1610,14 @@ export declare const MODELS: {
|
|
|
1319
1610
|
contextWindow: number;
|
|
1320
1611
|
maxTokens: number;
|
|
1321
1612
|
};
|
|
1322
|
-
readonly "
|
|
1613
|
+
readonly "glm-4.5v": {
|
|
1323
1614
|
id: string;
|
|
1324
1615
|
name: string;
|
|
1325
|
-
api: "
|
|
1616
|
+
api: "anthropic-messages";
|
|
1326
1617
|
provider: string;
|
|
1327
1618
|
baseUrl: string;
|
|
1328
|
-
reasoning:
|
|
1329
|
-
input: "text"[];
|
|
1619
|
+
reasoning: true;
|
|
1620
|
+
input: ("text" | "image")[];
|
|
1330
1621
|
cost: {
|
|
1331
1622
|
input: number;
|
|
1332
1623
|
output: number;
|
|
@@ -1336,9 +1627,7 @@ export declare const MODELS: {
|
|
|
1336
1627
|
contextWindow: number;
|
|
1337
1628
|
maxTokens: number;
|
|
1338
1629
|
};
|
|
1339
|
-
|
|
1340
|
-
readonly zai: {
|
|
1341
|
-
readonly "glm-4.5-air": {
|
|
1630
|
+
readonly "glm-4.6": {
|
|
1342
1631
|
id: string;
|
|
1343
1632
|
name: string;
|
|
1344
1633
|
api: "anthropic-messages";
|
|
@@ -1355,14 +1644,16 @@ export declare const MODELS: {
|
|
|
1355
1644
|
contextWindow: number;
|
|
1356
1645
|
maxTokens: number;
|
|
1357
1646
|
};
|
|
1358
|
-
|
|
1647
|
+
};
|
|
1648
|
+
readonly openrouter: {
|
|
1649
|
+
readonly "z-ai/glm-4.6": {
|
|
1359
1650
|
id: string;
|
|
1360
1651
|
name: string;
|
|
1361
|
-
api: "
|
|
1652
|
+
api: "openai-completions";
|
|
1362
1653
|
provider: string;
|
|
1363
1654
|
baseUrl: string;
|
|
1364
1655
|
reasoning: true;
|
|
1365
|
-
input:
|
|
1656
|
+
input: "text"[];
|
|
1366
1657
|
cost: {
|
|
1367
1658
|
input: number;
|
|
1368
1659
|
output: number;
|
|
@@ -1372,10 +1663,10 @@ export declare const MODELS: {
|
|
|
1372
1663
|
contextWindow: number;
|
|
1373
1664
|
maxTokens: number;
|
|
1374
1665
|
};
|
|
1375
|
-
readonly "
|
|
1666
|
+
readonly "deepseek/deepseek-v3.2-exp": {
|
|
1376
1667
|
id: string;
|
|
1377
1668
|
name: string;
|
|
1378
|
-
api: "
|
|
1669
|
+
api: "openai-completions";
|
|
1379
1670
|
provider: string;
|
|
1380
1671
|
baseUrl: string;
|
|
1381
1672
|
reasoning: true;
|
|
@@ -1389,14 +1680,14 @@ export declare const MODELS: {
|
|
|
1389
1680
|
contextWindow: number;
|
|
1390
1681
|
maxTokens: number;
|
|
1391
1682
|
};
|
|
1392
|
-
readonly "
|
|
1683
|
+
readonly "qwen/qwen3-vl-235b-a22b-thinking": {
|
|
1393
1684
|
id: string;
|
|
1394
1685
|
name: string;
|
|
1395
|
-
api: "
|
|
1686
|
+
api: "openai-completions";
|
|
1396
1687
|
provider: string;
|
|
1397
1688
|
baseUrl: string;
|
|
1398
1689
|
reasoning: true;
|
|
1399
|
-
input: "text"[];
|
|
1690
|
+
input: ("text" | "image")[];
|
|
1400
1691
|
cost: {
|
|
1401
1692
|
input: number;
|
|
1402
1693
|
output: number;
|
|
@@ -1406,16 +1697,14 @@ export declare const MODELS: {
|
|
|
1406
1697
|
contextWindow: number;
|
|
1407
1698
|
maxTokens: number;
|
|
1408
1699
|
};
|
|
1409
|
-
|
|
1410
|
-
readonly openrouter: {
|
|
1411
|
-
readonly "alibaba/tongyi-deepresearch-30b-a3b": {
|
|
1700
|
+
readonly "qwen/qwen3-vl-235b-a22b-instruct": {
|
|
1412
1701
|
id: string;
|
|
1413
1702
|
name: string;
|
|
1414
1703
|
api: "openai-completions";
|
|
1415
1704
|
provider: string;
|
|
1416
1705
|
baseUrl: string;
|
|
1417
1706
|
reasoning: true;
|
|
1418
|
-
input: "text"[];
|
|
1707
|
+
input: ("text" | "image")[];
|
|
1419
1708
|
cost: {
|
|
1420
1709
|
input: number;
|
|
1421
1710
|
output: number;
|
|
@@ -1425,7 +1714,7 @@ export declare const MODELS: {
|
|
|
1425
1714
|
contextWindow: number;
|
|
1426
1715
|
maxTokens: number;
|
|
1427
1716
|
};
|
|
1428
|
-
readonly "qwen/qwen3-
|
|
1717
|
+
readonly "qwen/qwen3-max": {
|
|
1429
1718
|
id: string;
|
|
1430
1719
|
name: string;
|
|
1431
1720
|
api: "openai-completions";
|
|
@@ -1459,7 +1748,7 @@ export declare const MODELS: {
|
|
|
1459
1748
|
contextWindow: number;
|
|
1460
1749
|
maxTokens: number;
|
|
1461
1750
|
};
|
|
1462
|
-
readonly "
|
|
1751
|
+
readonly "deepseek/deepseek-v3.1-terminus": {
|
|
1463
1752
|
id: string;
|
|
1464
1753
|
name: string;
|
|
1465
1754
|
api: "openai-completions";
|
|
@@ -1476,7 +1765,7 @@ export declare const MODELS: {
|
|
|
1476
1765
|
contextWindow: number;
|
|
1477
1766
|
maxTokens: number;
|
|
1478
1767
|
};
|
|
1479
|
-
readonly "
|
|
1768
|
+
readonly "alibaba/tongyi-deepresearch-30b-a3b:free": {
|
|
1480
1769
|
id: string;
|
|
1481
1770
|
name: string;
|
|
1482
1771
|
api: "openai-completions";
|
|
@@ -1493,13 +1782,13 @@ export declare const MODELS: {
|
|
|
1493
1782
|
contextWindow: number;
|
|
1494
1783
|
maxTokens: number;
|
|
1495
1784
|
};
|
|
1496
|
-
readonly "
|
|
1785
|
+
readonly "alibaba/tongyi-deepresearch-30b-a3b": {
|
|
1497
1786
|
id: string;
|
|
1498
1787
|
name: string;
|
|
1499
1788
|
api: "openai-completions";
|
|
1500
1789
|
provider: string;
|
|
1501
1790
|
baseUrl: string;
|
|
1502
|
-
reasoning:
|
|
1791
|
+
reasoning: true;
|
|
1503
1792
|
input: "text"[];
|
|
1504
1793
|
cost: {
|
|
1505
1794
|
input: number;
|
|
@@ -1510,7 +1799,7 @@ export declare const MODELS: {
|
|
|
1510
1799
|
contextWindow: number;
|
|
1511
1800
|
maxTokens: number;
|
|
1512
1801
|
};
|
|
1513
|
-
readonly "qwen/
|
|
1802
|
+
readonly "qwen/qwen3-coder-flash": {
|
|
1514
1803
|
id: string;
|
|
1515
1804
|
name: string;
|
|
1516
1805
|
api: "openai-completions";
|
|
@@ -1527,7 +1816,7 @@ export declare const MODELS: {
|
|
|
1527
1816
|
contextWindow: number;
|
|
1528
1817
|
maxTokens: number;
|
|
1529
1818
|
};
|
|
1530
|
-
readonly "qwen/
|
|
1819
|
+
readonly "qwen/qwen3-next-80b-a3b-thinking": {
|
|
1531
1820
|
id: string;
|
|
1532
1821
|
name: string;
|
|
1533
1822
|
api: "openai-completions";
|
|
@@ -1544,13 +1833,13 @@ export declare const MODELS: {
|
|
|
1544
1833
|
contextWindow: number;
|
|
1545
1834
|
maxTokens: number;
|
|
1546
1835
|
};
|
|
1547
|
-
readonly "
|
|
1836
|
+
readonly "qwen/qwen3-next-80b-a3b-instruct": {
|
|
1548
1837
|
id: string;
|
|
1549
1838
|
name: string;
|
|
1550
1839
|
api: "openai-completions";
|
|
1551
1840
|
provider: string;
|
|
1552
1841
|
baseUrl: string;
|
|
1553
|
-
reasoning:
|
|
1842
|
+
reasoning: false;
|
|
1554
1843
|
input: "text"[];
|
|
1555
1844
|
cost: {
|
|
1556
1845
|
input: number;
|
|
@@ -1561,13 +1850,13 @@ export declare const MODELS: {
|
|
|
1561
1850
|
contextWindow: number;
|
|
1562
1851
|
maxTokens: number;
|
|
1563
1852
|
};
|
|
1564
|
-
readonly "
|
|
1853
|
+
readonly "meituan/longcat-flash-chat:free": {
|
|
1565
1854
|
id: string;
|
|
1566
1855
|
name: string;
|
|
1567
1856
|
api: "openai-completions";
|
|
1568
1857
|
provider: string;
|
|
1569
1858
|
baseUrl: string;
|
|
1570
|
-
reasoning:
|
|
1859
|
+
reasoning: false;
|
|
1571
1860
|
input: "text"[];
|
|
1572
1861
|
cost: {
|
|
1573
1862
|
input: number;
|
|
@@ -1578,14 +1867,14 @@ export declare const MODELS: {
|
|
|
1578
1867
|
contextWindow: number;
|
|
1579
1868
|
maxTokens: number;
|
|
1580
1869
|
};
|
|
1581
|
-
readonly "
|
|
1870
|
+
readonly "qwen/qwen-plus-2025-07-28": {
|
|
1582
1871
|
id: string;
|
|
1583
1872
|
name: string;
|
|
1584
1873
|
api: "openai-completions";
|
|
1585
1874
|
provider: string;
|
|
1586
1875
|
baseUrl: string;
|
|
1587
1876
|
reasoning: false;
|
|
1588
|
-
input:
|
|
1877
|
+
input: "text"[];
|
|
1589
1878
|
cost: {
|
|
1590
1879
|
input: number;
|
|
1591
1880
|
output: number;
|
|
@@ -1595,14 +1884,14 @@ export declare const MODELS: {
|
|
|
1595
1884
|
contextWindow: number;
|
|
1596
1885
|
maxTokens: number;
|
|
1597
1886
|
};
|
|
1598
|
-
readonly "
|
|
1887
|
+
readonly "qwen/qwen-plus-2025-07-28:thinking": {
|
|
1599
1888
|
id: string;
|
|
1600
1889
|
name: string;
|
|
1601
1890
|
api: "openai-completions";
|
|
1602
1891
|
provider: string;
|
|
1603
1892
|
baseUrl: string;
|
|
1604
1893
|
reasoning: true;
|
|
1605
|
-
input:
|
|
1894
|
+
input: "text"[];
|
|
1606
1895
|
cost: {
|
|
1607
1896
|
input: number;
|
|
1608
1897
|
output: number;
|
|
@@ -1612,13 +1901,30 @@ export declare const MODELS: {
|
|
|
1612
1901
|
contextWindow: number;
|
|
1613
1902
|
maxTokens: number;
|
|
1614
1903
|
};
|
|
1615
|
-
readonly "
|
|
1904
|
+
readonly "nvidia/nemotron-nano-9b-v2:free": {
|
|
1616
1905
|
id: string;
|
|
1617
1906
|
name: string;
|
|
1618
1907
|
api: "openai-completions";
|
|
1619
1908
|
provider: string;
|
|
1620
1909
|
baseUrl: string;
|
|
1621
|
-
reasoning:
|
|
1910
|
+
reasoning: true;
|
|
1911
|
+
input: "text"[];
|
|
1912
|
+
cost: {
|
|
1913
|
+
input: number;
|
|
1914
|
+
output: number;
|
|
1915
|
+
cacheRead: number;
|
|
1916
|
+
cacheWrite: number;
|
|
1917
|
+
};
|
|
1918
|
+
contextWindow: number;
|
|
1919
|
+
maxTokens: number;
|
|
1920
|
+
};
|
|
1921
|
+
readonly "nvidia/nemotron-nano-9b-v2": {
|
|
1922
|
+
id: string;
|
|
1923
|
+
name: string;
|
|
1924
|
+
api: "openai-completions";
|
|
1925
|
+
provider: string;
|
|
1926
|
+
baseUrl: string;
|
|
1927
|
+
reasoning: true;
|
|
1622
1928
|
input: "text"[];
|
|
1623
1929
|
cost: {
|
|
1624
1930
|
input: number;
|
|
@@ -2020,23 +2326,6 @@ export declare const MODELS: {
|
|
|
2020
2326
|
contextWindow: number;
|
|
2021
2327
|
maxTokens: number;
|
|
2022
2328
|
};
|
|
2023
|
-
readonly "moonshotai/kimi-k2:free": {
|
|
2024
|
-
id: string;
|
|
2025
|
-
name: string;
|
|
2026
|
-
api: "openai-completions";
|
|
2027
|
-
provider: string;
|
|
2028
|
-
baseUrl: string;
|
|
2029
|
-
reasoning: false;
|
|
2030
|
-
input: "text"[];
|
|
2031
|
-
cost: {
|
|
2032
|
-
input: number;
|
|
2033
|
-
output: number;
|
|
2034
|
-
cacheRead: number;
|
|
2035
|
-
cacheWrite: number;
|
|
2036
|
-
};
|
|
2037
|
-
contextWindow: number;
|
|
2038
|
-
maxTokens: number;
|
|
2039
|
-
};
|
|
2040
2329
|
readonly "moonshotai/kimi-k2": {
|
|
2041
2330
|
id: string;
|
|
2042
2331
|
name: string;
|
|
@@ -3074,7 +3363,7 @@ export declare const MODELS: {
|
|
|
3074
3363
|
contextWindow: number;
|
|
3075
3364
|
maxTokens: number;
|
|
3076
3365
|
};
|
|
3077
|
-
readonly "meta-llama/llama-3.1-
|
|
3366
|
+
readonly "meta-llama/llama-3.1-405b-instruct": {
|
|
3078
3367
|
id: string;
|
|
3079
3368
|
name: string;
|
|
3080
3369
|
api: "openai-completions";
|
|
@@ -3091,7 +3380,7 @@ export declare const MODELS: {
|
|
|
3091
3380
|
contextWindow: number;
|
|
3092
3381
|
maxTokens: number;
|
|
3093
3382
|
};
|
|
3094
|
-
readonly "meta-llama/llama-3.1-
|
|
3383
|
+
readonly "meta-llama/llama-3.1-8b-instruct": {
|
|
3095
3384
|
id: string;
|
|
3096
3385
|
name: string;
|
|
3097
3386
|
api: "openai-completions";
|
|
@@ -3278,74 +3567,6 @@ export declare const MODELS: {
|
|
|
3278
3567
|
contextWindow: number;
|
|
3279
3568
|
maxTokens: number;
|
|
3280
3569
|
};
|
|
3281
|
-
readonly "cohere/command-r-plus": {
|
|
3282
|
-
id: string;
|
|
3283
|
-
name: string;
|
|
3284
|
-
api: "openai-completions";
|
|
3285
|
-
provider: string;
|
|
3286
|
-
baseUrl: string;
|
|
3287
|
-
reasoning: false;
|
|
3288
|
-
input: "text"[];
|
|
3289
|
-
cost: {
|
|
3290
|
-
input: number;
|
|
3291
|
-
output: number;
|
|
3292
|
-
cacheRead: number;
|
|
3293
|
-
cacheWrite: number;
|
|
3294
|
-
};
|
|
3295
|
-
contextWindow: number;
|
|
3296
|
-
maxTokens: number;
|
|
3297
|
-
};
|
|
3298
|
-
readonly "cohere/command-r-plus-04-2024": {
|
|
3299
|
-
id: string;
|
|
3300
|
-
name: string;
|
|
3301
|
-
api: "openai-completions";
|
|
3302
|
-
provider: string;
|
|
3303
|
-
baseUrl: string;
|
|
3304
|
-
reasoning: false;
|
|
3305
|
-
input: "text"[];
|
|
3306
|
-
cost: {
|
|
3307
|
-
input: number;
|
|
3308
|
-
output: number;
|
|
3309
|
-
cacheRead: number;
|
|
3310
|
-
cacheWrite: number;
|
|
3311
|
-
};
|
|
3312
|
-
contextWindow: number;
|
|
3313
|
-
maxTokens: number;
|
|
3314
|
-
};
|
|
3315
|
-
readonly "cohere/command-r": {
|
|
3316
|
-
id: string;
|
|
3317
|
-
name: string;
|
|
3318
|
-
api: "openai-completions";
|
|
3319
|
-
provider: string;
|
|
3320
|
-
baseUrl: string;
|
|
3321
|
-
reasoning: false;
|
|
3322
|
-
input: "text"[];
|
|
3323
|
-
cost: {
|
|
3324
|
-
input: number;
|
|
3325
|
-
output: number;
|
|
3326
|
-
cacheRead: number;
|
|
3327
|
-
cacheWrite: number;
|
|
3328
|
-
};
|
|
3329
|
-
contextWindow: number;
|
|
3330
|
-
maxTokens: number;
|
|
3331
|
-
};
|
|
3332
|
-
readonly "cohere/command-r-03-2024": {
|
|
3333
|
-
id: string;
|
|
3334
|
-
name: string;
|
|
3335
|
-
api: "openai-completions";
|
|
3336
|
-
provider: string;
|
|
3337
|
-
baseUrl: string;
|
|
3338
|
-
reasoning: false;
|
|
3339
|
-
input: "text"[];
|
|
3340
|
-
cost: {
|
|
3341
|
-
input: number;
|
|
3342
|
-
output: number;
|
|
3343
|
-
cacheRead: number;
|
|
3344
|
-
cacheWrite: number;
|
|
3345
|
-
};
|
|
3346
|
-
contextWindow: number;
|
|
3347
|
-
maxTokens: number;
|
|
3348
|
-
};
|
|
3349
3570
|
readonly "mistralai/mistral-large": {
|
|
3350
3571
|
id: string;
|
|
3351
3572
|
name: string;
|