@mariozechner/pi-ai 0.5.43 → 0.5.45

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (39) hide show
  1. package/dist/agent/agent-loop.d.ts +2 -2
  2. package/dist/agent/agent-loop.d.ts.map +1 -1
  3. package/dist/agent/agent-loop.js +18 -2
  4. package/dist/agent/agent-loop.js.map +1 -1
  5. package/dist/agent/index.d.ts +1 -1
  6. package/dist/agent/index.d.ts.map +1 -1
  7. package/dist/agent/index.js.map +1 -1
  8. package/dist/agent/types.d.ts +6 -1
  9. package/dist/agent/types.d.ts.map +1 -1
  10. package/dist/agent/types.js.map +1 -1
  11. package/dist/models.generated.d.ts +933 -270
  12. package/dist/models.generated.d.ts.map +1 -1
  13. package/dist/models.generated.js +1500 -837
  14. package/dist/models.generated.js.map +1 -1
  15. package/dist/providers/anthropic.d.ts.map +1 -1
  16. package/dist/providers/anthropic.js +32 -8
  17. package/dist/providers/anthropic.js.map +1 -1
  18. package/dist/providers/google.d.ts.map +1 -1
  19. package/dist/providers/google.js +8 -7
  20. package/dist/providers/google.js.map +1 -1
  21. package/dist/providers/openai-completions.d.ts.map +1 -1
  22. package/dist/providers/openai-completions.js +6 -5
  23. package/dist/providers/openai-completions.js.map +1 -1
  24. package/dist/providers/openai-responses.d.ts.map +1 -1
  25. package/dist/providers/openai-responses.js +6 -5
  26. package/dist/providers/openai-responses.js.map +1 -1
  27. package/dist/providers/transorm-messages.d.ts.map +1 -1
  28. package/dist/providers/transorm-messages.js +46 -1
  29. package/dist/providers/transorm-messages.js.map +1 -1
  30. package/dist/utils/event-stream.js +3 -7
  31. package/dist/utils/event-stream.js.map +1 -1
  32. package/dist/utils/sanitize-unicode.d.ts +22 -0
  33. package/dist/utils/sanitize-unicode.d.ts.map +1 -0
  34. package/dist/utils/sanitize-unicode.js +26 -0
  35. package/dist/utils/sanitize-unicode.js.map +1 -0
  36. package/dist/utils/validation.d.ts.map +1 -1
  37. package/dist/utils/validation.js +25 -3
  38. package/dist/utils/validation.js.map +1 -1
  39. package/package.json +3 -3
@@ -1,12 +1,12 @@
1
1
  export declare const MODELS: {
2
2
  readonly anthropic: {
3
- readonly "claude-3-7-sonnet-20250219": {
3
+ readonly "claude-3-5-sonnet-20241022": {
4
4
  id: string;
5
5
  name: string;
6
6
  api: "anthropic-messages";
7
7
  provider: string;
8
8
  baseUrl: string;
9
- reasoning: true;
9
+ reasoning: false;
10
10
  input: ("text" | "image")[];
11
11
  cost: {
12
12
  input: number;
@@ -17,13 +17,13 @@ export declare const MODELS: {
17
17
  contextWindow: number;
18
18
  maxTokens: number;
19
19
  };
20
- readonly "claude-opus-4-1-20250805": {
20
+ readonly "claude-3-5-sonnet-20240620": {
21
21
  id: string;
22
22
  name: string;
23
23
  api: "anthropic-messages";
24
24
  provider: string;
25
25
  baseUrl: string;
26
- reasoning: true;
26
+ reasoning: false;
27
27
  input: ("text" | "image")[];
28
28
  cost: {
29
29
  input: number;
@@ -34,7 +34,7 @@ export declare const MODELS: {
34
34
  contextWindow: number;
35
35
  maxTokens: number;
36
36
  };
37
- readonly "claude-3-haiku-20240307": {
37
+ readonly "claude-3-opus-20240229": {
38
38
  id: string;
39
39
  name: string;
40
40
  api: "anthropic-messages";
@@ -51,13 +51,30 @@ export declare const MODELS: {
51
51
  contextWindow: number;
52
52
  maxTokens: number;
53
53
  };
54
- readonly "claude-3-5-haiku-20241022": {
54
+ readonly "claude-sonnet-4-5-20250929": {
55
55
  id: string;
56
56
  name: string;
57
57
  api: "anthropic-messages";
58
58
  provider: string;
59
59
  baseUrl: string;
60
- reasoning: false;
60
+ reasoning: true;
61
+ input: ("text" | "image")[];
62
+ cost: {
63
+ input: number;
64
+ output: number;
65
+ cacheRead: number;
66
+ cacheWrite: number;
67
+ };
68
+ contextWindow: number;
69
+ maxTokens: number;
70
+ };
71
+ readonly "claude-sonnet-4-20250514": {
72
+ id: string;
73
+ name: string;
74
+ api: "anthropic-messages";
75
+ provider: string;
76
+ baseUrl: string;
77
+ reasoning: true;
61
78
  input: ("text" | "image")[];
62
79
  cost: {
63
80
  input: number;
@@ -85,7 +102,7 @@ export declare const MODELS: {
85
102
  contextWindow: number;
86
103
  maxTokens: number;
87
104
  };
88
- readonly "claude-3-5-sonnet-20241022": {
105
+ readonly "claude-3-5-haiku-20241022": {
89
106
  id: string;
90
107
  name: string;
91
108
  api: "anthropic-messages";
@@ -102,7 +119,7 @@ export declare const MODELS: {
102
119
  contextWindow: number;
103
120
  maxTokens: number;
104
121
  };
105
- readonly "claude-3-5-sonnet-20240620": {
122
+ readonly "claude-3-haiku-20240307": {
106
123
  id: string;
107
124
  name: string;
108
125
  api: "anthropic-messages";
@@ -119,13 +136,13 @@ export declare const MODELS: {
119
136
  contextWindow: number;
120
137
  maxTokens: number;
121
138
  };
122
- readonly "claude-3-sonnet-20240229": {
139
+ readonly "claude-3-7-sonnet-20250219": {
123
140
  id: string;
124
141
  name: string;
125
142
  api: "anthropic-messages";
126
143
  provider: string;
127
144
  baseUrl: string;
128
- reasoning: false;
145
+ reasoning: true;
129
146
  input: ("text" | "image")[];
130
147
  cost: {
131
148
  input: number;
@@ -136,7 +153,7 @@ export declare const MODELS: {
136
153
  contextWindow: number;
137
154
  maxTokens: number;
138
155
  };
139
- readonly "claude-sonnet-4-20250514": {
156
+ readonly "claude-opus-4-1-20250805": {
140
157
  id: string;
141
158
  name: string;
142
159
  api: "anthropic-messages";
@@ -153,7 +170,7 @@ export declare const MODELS: {
153
170
  contextWindow: number;
154
171
  maxTokens: number;
155
172
  };
156
- readonly "claude-3-opus-20240229": {
173
+ readonly "claude-3-sonnet-20240229": {
157
174
  id: string;
158
175
  name: string;
159
176
  api: "anthropic-messages";
@@ -170,12 +187,10 @@ export declare const MODELS: {
170
187
  contextWindow: number;
171
188
  maxTokens: number;
172
189
  };
173
- };
174
- readonly google: {
175
- readonly "gemini-2.5-flash-preview-05-20": {
190
+ readonly "claude-haiku-4-5-20251001": {
176
191
  id: string;
177
192
  name: string;
178
- api: "google-generative-ai";
193
+ api: "anthropic-messages";
179
194
  provider: string;
180
195
  baseUrl: string;
181
196
  reasoning: true;
@@ -189,10 +204,10 @@ export declare const MODELS: {
189
204
  contextWindow: number;
190
205
  maxTokens: number;
191
206
  };
192
- readonly "gemini-2.5-pro": {
207
+ readonly "claude-sonnet-4-0": {
193
208
  id: string;
194
209
  name: string;
195
- api: "google-generative-ai";
210
+ api: "anthropic-messages";
196
211
  provider: string;
197
212
  baseUrl: string;
198
213
  reasoning: true;
@@ -206,13 +221,13 @@ export declare const MODELS: {
206
221
  contextWindow: number;
207
222
  maxTokens: number;
208
223
  };
209
- readonly "gemini-1.5-flash": {
224
+ readonly "claude-3-7-sonnet-latest": {
210
225
  id: string;
211
226
  name: string;
212
- api: "google-generative-ai";
227
+ api: "anthropic-messages";
213
228
  provider: string;
214
229
  baseUrl: string;
215
- reasoning: false;
230
+ reasoning: true;
216
231
  input: ("text" | "image")[];
217
232
  cost: {
218
233
  input: number;
@@ -223,13 +238,13 @@ export declare const MODELS: {
223
238
  contextWindow: number;
224
239
  maxTokens: number;
225
240
  };
226
- readonly "gemini-2.0-flash-lite": {
241
+ readonly "claude-sonnet-4-5": {
227
242
  id: string;
228
243
  name: string;
229
- api: "google-generative-ai";
244
+ api: "anthropic-messages";
230
245
  provider: string;
231
246
  baseUrl: string;
232
- reasoning: false;
247
+ reasoning: true;
233
248
  input: ("text" | "image")[];
234
249
  cost: {
235
250
  input: number;
@@ -240,10 +255,10 @@ export declare const MODELS: {
240
255
  contextWindow: number;
241
256
  maxTokens: number;
242
257
  };
243
- readonly "gemini-1.5-pro": {
258
+ readonly "claude-3-5-haiku-latest": {
244
259
  id: string;
245
260
  name: string;
246
- api: "google-generative-ai";
261
+ api: "anthropic-messages";
247
262
  provider: string;
248
263
  baseUrl: string;
249
264
  reasoning: false;
@@ -257,13 +272,13 @@ export declare const MODELS: {
257
272
  contextWindow: number;
258
273
  maxTokens: number;
259
274
  };
260
- readonly "gemini-1.5-flash-8b": {
275
+ readonly "claude-haiku-4-5": {
261
276
  id: string;
262
277
  name: string;
263
- api: "google-generative-ai";
278
+ api: "anthropic-messages";
264
279
  provider: string;
265
280
  baseUrl: string;
266
- reasoning: false;
281
+ reasoning: true;
267
282
  input: ("text" | "image")[];
268
283
  cost: {
269
284
  input: number;
@@ -274,10 +289,10 @@ export declare const MODELS: {
274
289
  contextWindow: number;
275
290
  maxTokens: number;
276
291
  };
277
- readonly "gemini-2.5-flash": {
292
+ readonly "claude-opus-4-1": {
278
293
  id: string;
279
294
  name: string;
280
- api: "google-generative-ai";
295
+ api: "anthropic-messages";
281
296
  provider: string;
282
297
  baseUrl: string;
283
298
  reasoning: true;
@@ -291,10 +306,10 @@ export declare const MODELS: {
291
306
  contextWindow: number;
292
307
  maxTokens: number;
293
308
  };
294
- readonly "gemini-2.5-pro-preview-06-05": {
309
+ readonly "claude-opus-4-0": {
295
310
  id: string;
296
311
  name: string;
297
- api: "google-generative-ai";
312
+ api: "anthropic-messages";
298
313
  provider: string;
299
314
  baseUrl: string;
300
315
  reasoning: true;
@@ -308,7 +323,9 @@ export declare const MODELS: {
308
323
  contextWindow: number;
309
324
  maxTokens: number;
310
325
  };
311
- readonly "gemini-2.5-pro-preview-05-06": {
326
+ };
327
+ readonly google: {
328
+ readonly "gemini-2.5-flash-preview-05-20": {
312
329
  id: string;
313
330
  name: string;
314
331
  api: "google-generative-ai";
@@ -325,13 +342,13 @@ export declare const MODELS: {
325
342
  contextWindow: number;
326
343
  maxTokens: number;
327
344
  };
328
- readonly "gemini-2.0-flash": {
345
+ readonly "gemini-flash-lite-latest": {
329
346
  id: string;
330
347
  name: string;
331
348
  api: "google-generative-ai";
332
349
  provider: string;
333
350
  baseUrl: string;
334
- reasoning: false;
351
+ reasoning: true;
335
352
  input: ("text" | "image")[];
336
353
  cost: {
337
354
  input: number;
@@ -342,7 +359,7 @@ export declare const MODELS: {
342
359
  contextWindow: number;
343
360
  maxTokens: number;
344
361
  };
345
- readonly "gemini-2.5-flash-lite-preview-06-17": {
362
+ readonly "gemini-2.5-flash": {
346
363
  id: string;
347
364
  name: string;
348
365
  api: "google-generative-ai";
@@ -359,7 +376,7 @@ export declare const MODELS: {
359
376
  contextWindow: number;
360
377
  maxTokens: number;
361
378
  };
362
- readonly "gemini-2.5-flash-preview-04-17": {
379
+ readonly "gemini-flash-latest": {
363
380
  id: string;
364
381
  name: string;
365
382
  api: "google-generative-ai";
@@ -376,12 +393,10 @@ export declare const MODELS: {
376
393
  contextWindow: number;
377
394
  maxTokens: number;
378
395
  };
379
- };
380
- readonly openai: {
381
- readonly "gpt-5-nano": {
396
+ readonly "gemini-2.5-pro-preview-05-06": {
382
397
  id: string;
383
398
  name: string;
384
- api: "openai-responses";
399
+ api: "google-generative-ai";
385
400
  provider: string;
386
401
  baseUrl: string;
387
402
  reasoning: true;
@@ -395,13 +410,13 @@ export declare const MODELS: {
395
410
  contextWindow: number;
396
411
  maxTokens: number;
397
412
  };
398
- readonly "o3-pro": {
413
+ readonly "gemini-2.0-flash-lite": {
399
414
  id: string;
400
415
  name: string;
401
- api: "openai-responses";
416
+ api: "google-generative-ai";
402
417
  provider: string;
403
418
  baseUrl: string;
404
- reasoning: true;
419
+ reasoning: false;
405
420
  input: ("text" | "image")[];
406
421
  cost: {
407
422
  input: number;
@@ -412,10 +427,10 @@ export declare const MODELS: {
412
427
  contextWindow: number;
413
428
  maxTokens: number;
414
429
  };
415
- readonly "codex-mini-latest": {
430
+ readonly "gemini-live-2.5-flash-preview-native-audio": {
416
431
  id: string;
417
432
  name: string;
418
- api: "openai-responses";
433
+ api: "google-generative-ai";
419
434
  provider: string;
420
435
  baseUrl: string;
421
436
  reasoning: true;
@@ -429,10 +444,10 @@ export declare const MODELS: {
429
444
  contextWindow: number;
430
445
  maxTokens: number;
431
446
  };
432
- readonly "gpt-4.1": {
447
+ readonly "gemini-2.0-flash": {
433
448
  id: string;
434
449
  name: string;
435
- api: "openai-responses";
450
+ api: "google-generative-ai";
436
451
  provider: string;
437
452
  baseUrl: string;
438
453
  reasoning: false;
@@ -446,13 +461,13 @@ export declare const MODELS: {
446
461
  contextWindow: number;
447
462
  maxTokens: number;
448
463
  };
449
- readonly "gpt-4-turbo": {
464
+ readonly "gemini-2.5-flash-lite": {
450
465
  id: string;
451
466
  name: string;
452
- api: "openai-responses";
467
+ api: "google-generative-ai";
453
468
  provider: string;
454
469
  baseUrl: string;
455
- reasoning: false;
470
+ reasoning: true;
456
471
  input: ("text" | "image")[];
457
472
  cost: {
458
473
  input: number;
@@ -463,10 +478,10 @@ export declare const MODELS: {
463
478
  contextWindow: number;
464
479
  maxTokens: number;
465
480
  };
466
- readonly o1: {
481
+ readonly "gemini-2.5-pro-preview-06-05": {
467
482
  id: string;
468
483
  name: string;
469
- api: "openai-responses";
484
+ api: "google-generative-ai";
470
485
  provider: string;
471
486
  baseUrl: string;
472
487
  reasoning: true;
@@ -480,10 +495,10 @@ export declare const MODELS: {
480
495
  contextWindow: number;
481
496
  maxTokens: number;
482
497
  };
483
- readonly "o3-deep-research": {
498
+ readonly "gemini-live-2.5-flash": {
484
499
  id: string;
485
500
  name: string;
486
- api: "openai-responses";
501
+ api: "google-generative-ai";
487
502
  provider: string;
488
503
  baseUrl: string;
489
504
  reasoning: true;
@@ -497,10 +512,10 @@ export declare const MODELS: {
497
512
  contextWindow: number;
498
513
  maxTokens: number;
499
514
  };
500
- readonly "gpt-5": {
515
+ readonly "gemini-2.5-flash-lite-preview-06-17": {
501
516
  id: string;
502
517
  name: string;
503
- api: "openai-responses";
518
+ api: "google-generative-ai";
504
519
  provider: string;
505
520
  baseUrl: string;
506
521
  reasoning: true;
@@ -514,10 +529,10 @@ export declare const MODELS: {
514
529
  contextWindow: number;
515
530
  maxTokens: number;
516
531
  };
517
- readonly "o1-pro": {
532
+ readonly "gemini-2.5-flash-preview-09-2025": {
518
533
  id: string;
519
534
  name: string;
520
- api: "openai-responses";
535
+ api: "google-generative-ai";
521
536
  provider: string;
522
537
  baseUrl: string;
523
538
  reasoning: true;
@@ -531,10 +546,10 @@ export declare const MODELS: {
531
546
  contextWindow: number;
532
547
  maxTokens: number;
533
548
  };
534
- readonly o3: {
549
+ readonly "gemini-2.5-flash-preview-04-17": {
535
550
  id: string;
536
551
  name: string;
537
- api: "openai-responses";
552
+ api: "google-generative-ai";
538
553
  provider: string;
539
554
  baseUrl: string;
540
555
  reasoning: true;
@@ -548,10 +563,10 @@ export declare const MODELS: {
548
563
  contextWindow: number;
549
564
  maxTokens: number;
550
565
  };
551
- readonly "gpt-5-mini": {
566
+ readonly "gemini-2.5-pro": {
552
567
  id: string;
553
568
  name: string;
554
- api: "openai-responses";
569
+ api: "google-generative-ai";
555
570
  provider: string;
556
571
  baseUrl: string;
557
572
  reasoning: true;
@@ -565,13 +580,13 @@ export declare const MODELS: {
565
580
  contextWindow: number;
566
581
  maxTokens: number;
567
582
  };
568
- readonly "o4-mini-deep-research": {
583
+ readonly "gemini-1.5-flash": {
569
584
  id: string;
570
585
  name: string;
571
- api: "openai-responses";
586
+ api: "google-generative-ai";
572
587
  provider: string;
573
588
  baseUrl: string;
574
- reasoning: true;
589
+ reasoning: false;
575
590
  input: ("text" | "image")[];
576
591
  cost: {
577
592
  input: number;
@@ -582,10 +597,10 @@ export declare const MODELS: {
582
597
  contextWindow: number;
583
598
  maxTokens: number;
584
599
  };
585
- readonly "gpt-4o-mini": {
600
+ readonly "gemini-1.5-flash-8b": {
586
601
  id: string;
587
602
  name: string;
588
- api: "openai-responses";
603
+ api: "google-generative-ai";
589
604
  provider: string;
590
605
  baseUrl: string;
591
606
  reasoning: false;
@@ -599,13 +614,13 @@ export declare const MODELS: {
599
614
  contextWindow: number;
600
615
  maxTokens: number;
601
616
  };
602
- readonly "gpt-4.1-nano": {
617
+ readonly "gemini-2.5-flash-lite-preview-09-2025": {
603
618
  id: string;
604
619
  name: string;
605
- api: "openai-responses";
620
+ api: "google-generative-ai";
606
621
  provider: string;
607
622
  baseUrl: string;
608
- reasoning: false;
623
+ reasoning: true;
609
624
  input: ("text" | "image")[];
610
625
  cost: {
611
626
  input: number;
@@ -616,10 +631,10 @@ export declare const MODELS: {
616
631
  contextWindow: number;
617
632
  maxTokens: number;
618
633
  };
619
- readonly "gpt-4.1-mini": {
634
+ readonly "gemini-1.5-pro": {
620
635
  id: string;
621
636
  name: string;
622
- api: "openai-responses";
637
+ api: "google-generative-ai";
623
638
  provider: string;
624
639
  baseUrl: string;
625
640
  reasoning: false;
@@ -633,7 +648,9 @@ export declare const MODELS: {
633
648
  contextWindow: number;
634
649
  maxTokens: number;
635
650
  };
636
- readonly "gpt-4o": {
651
+ };
652
+ readonly openai: {
653
+ readonly "gpt-4.1-nano": {
637
654
  id: string;
638
655
  name: string;
639
656
  api: "openai-responses";
@@ -667,7 +684,7 @@ export declare const MODELS: {
667
684
  contextWindow: number;
668
685
  maxTokens: number;
669
686
  };
670
- readonly "o4-mini": {
687
+ readonly "o1-pro": {
671
688
  id: string;
672
689
  name: string;
673
690
  api: "openai-responses";
@@ -684,14 +701,14 @@ export declare const MODELS: {
684
701
  contextWindow: number;
685
702
  maxTokens: number;
686
703
  };
687
- readonly "o3-mini": {
704
+ readonly "gpt-4o-2024-05-13": {
688
705
  id: string;
689
706
  name: string;
690
707
  api: "openai-responses";
691
708
  provider: string;
692
709
  baseUrl: string;
693
- reasoning: true;
694
- input: "text"[];
710
+ reasoning: false;
711
+ input: ("text" | "image")[];
695
712
  cost: {
696
713
  input: number;
697
714
  output: number;
@@ -701,7 +718,7 @@ export declare const MODELS: {
701
718
  contextWindow: number;
702
719
  maxTokens: number;
703
720
  };
704
- readonly "gpt-5-chat-latest": {
721
+ readonly "gpt-4o-2024-08-06": {
705
722
  id: string;
706
723
  name: string;
707
724
  api: "openai-responses";
@@ -718,16 +735,14 @@ export declare const MODELS: {
718
735
  contextWindow: number;
719
736
  maxTokens: number;
720
737
  };
721
- };
722
- readonly groq: {
723
- readonly "llama-3.1-8b-instant": {
738
+ readonly "gpt-4.1-mini": {
724
739
  id: string;
725
740
  name: string;
726
- api: "openai-completions";
741
+ api: "openai-responses";
727
742
  provider: string;
728
743
  baseUrl: string;
729
744
  reasoning: false;
730
- input: "text"[];
745
+ input: ("text" | "image")[];
731
746
  cost: {
732
747
  input: number;
733
748
  output: number;
@@ -737,14 +752,14 @@ export declare const MODELS: {
737
752
  contextWindow: number;
738
753
  maxTokens: number;
739
754
  };
740
- readonly "qwen-qwq-32b": {
755
+ readonly "o3-deep-research": {
741
756
  id: string;
742
757
  name: string;
743
- api: "openai-completions";
758
+ api: "openai-responses";
744
759
  provider: string;
745
760
  baseUrl: string;
746
761
  reasoning: true;
747
- input: "text"[];
762
+ input: ("text" | "image")[];
748
763
  cost: {
749
764
  input: number;
750
765
  output: number;
@@ -754,14 +769,14 @@ export declare const MODELS: {
754
769
  contextWindow: number;
755
770
  maxTokens: number;
756
771
  };
757
- readonly "llama3-70b-8192": {
772
+ readonly "gpt-4-turbo": {
758
773
  id: string;
759
774
  name: string;
760
- api: "openai-completions";
775
+ api: "openai-responses";
761
776
  provider: string;
762
777
  baseUrl: string;
763
778
  reasoning: false;
764
- input: "text"[];
779
+ input: ("text" | "image")[];
765
780
  cost: {
766
781
  input: number;
767
782
  output: number;
@@ -771,10 +786,10 @@ export declare const MODELS: {
771
786
  contextWindow: number;
772
787
  maxTokens: number;
773
788
  };
774
- readonly "deepseek-r1-distill-llama-70b": {
789
+ readonly "o3-mini": {
775
790
  id: string;
776
791
  name: string;
777
- api: "openai-completions";
792
+ api: "openai-responses";
778
793
  provider: string;
779
794
  baseUrl: string;
780
795
  reasoning: true;
@@ -788,13 +803,13 @@ export declare const MODELS: {
788
803
  contextWindow: number;
789
804
  maxTokens: number;
790
805
  };
791
- readonly "llama3-8b-8192": {
806
+ readonly "codex-mini-latest": {
792
807
  id: string;
793
808
  name: string;
794
- api: "openai-completions";
809
+ api: "openai-responses";
795
810
  provider: string;
796
811
  baseUrl: string;
797
- reasoning: false;
812
+ reasoning: true;
798
813
  input: "text"[];
799
814
  cost: {
800
815
  input: number;
@@ -805,14 +820,14 @@ export declare const MODELS: {
805
820
  contextWindow: number;
806
821
  maxTokens: number;
807
822
  };
808
- readonly "gemma2-9b-it": {
823
+ readonly "gpt-5-nano": {
809
824
  id: string;
810
825
  name: string;
811
- api: "openai-completions";
826
+ api: "openai-responses";
812
827
  provider: string;
813
828
  baseUrl: string;
814
- reasoning: false;
815
- input: "text"[];
829
+ reasoning: true;
830
+ input: ("text" | "image")[];
816
831
  cost: {
817
832
  input: number;
818
833
  output: number;
@@ -822,14 +837,14 @@ export declare const MODELS: {
822
837
  contextWindow: number;
823
838
  maxTokens: number;
824
839
  };
825
- readonly "llama-3.3-70b-versatile": {
840
+ readonly "gpt-5-codex": {
826
841
  id: string;
827
842
  name: string;
828
- api: "openai-completions";
843
+ api: "openai-responses";
829
844
  provider: string;
830
845
  baseUrl: string;
831
- reasoning: false;
832
- input: "text"[];
846
+ reasoning: true;
847
+ input: ("text" | "image")[];
833
848
  cost: {
834
849
  input: number;
835
850
  output: number;
@@ -839,14 +854,14 @@ export declare const MODELS: {
839
854
  contextWindow: number;
840
855
  maxTokens: number;
841
856
  };
842
- readonly "mistral-saba-24b": {
857
+ readonly "gpt-4o": {
843
858
  id: string;
844
859
  name: string;
845
- api: "openai-completions";
860
+ api: "openai-responses";
846
861
  provider: string;
847
862
  baseUrl: string;
848
863
  reasoning: false;
849
- input: "text"[];
864
+ input: ("text" | "image")[];
850
865
  cost: {
851
866
  input: number;
852
867
  output: number;
@@ -856,14 +871,14 @@ export declare const MODELS: {
856
871
  contextWindow: number;
857
872
  maxTokens: number;
858
873
  };
859
- readonly "openai/gpt-oss-20b": {
874
+ readonly "gpt-4.1": {
860
875
  id: string;
861
876
  name: string;
862
- api: "openai-completions";
877
+ api: "openai-responses";
863
878
  provider: string;
864
879
  baseUrl: string;
865
- reasoning: true;
866
- input: "text"[];
880
+ reasoning: false;
881
+ input: ("text" | "image")[];
867
882
  cost: {
868
883
  input: number;
869
884
  output: number;
@@ -873,14 +888,14 @@ export declare const MODELS: {
873
888
  contextWindow: number;
874
889
  maxTokens: number;
875
890
  };
876
- readonly "openai/gpt-oss-120b": {
891
+ readonly "o4-mini": {
877
892
  id: string;
878
893
  name: string;
879
- api: "openai-completions";
894
+ api: "openai-responses";
880
895
  provider: string;
881
896
  baseUrl: string;
882
897
  reasoning: true;
883
- input: "text"[];
898
+ input: ("text" | "image")[];
884
899
  cost: {
885
900
  input: number;
886
901
  output: number;
@@ -890,13 +905,13 @@ export declare const MODELS: {
890
905
  contextWindow: number;
891
906
  maxTokens: number;
892
907
  };
893
- readonly "meta-llama/llama-4-maverick-17b-128e-instruct": {
908
+ readonly o1: {
894
909
  id: string;
895
910
  name: string;
896
- api: "openai-completions";
911
+ api: "openai-responses";
897
912
  provider: string;
898
913
  baseUrl: string;
899
- reasoning: false;
914
+ reasoning: true;
900
915
  input: ("text" | "image")[];
901
916
  cost: {
902
917
  input: number;
@@ -907,13 +922,13 @@ export declare const MODELS: {
907
922
  contextWindow: number;
908
923
  maxTokens: number;
909
924
  };
910
- readonly "meta-llama/llama-4-scout-17b-16e-instruct": {
925
+ readonly "gpt-5-mini": {
911
926
  id: string;
912
927
  name: string;
913
- api: "openai-completions";
928
+ api: "openai-responses";
914
929
  provider: string;
915
930
  baseUrl: string;
916
- reasoning: false;
931
+ reasoning: true;
917
932
  input: ("text" | "image")[];
918
933
  cost: {
919
934
  input: number;
@@ -924,14 +939,14 @@ export declare const MODELS: {
924
939
  contextWindow: number;
925
940
  maxTokens: number;
926
941
  };
927
- readonly "qwen/qwen3-32b": {
942
+ readonly "o3-pro": {
928
943
  id: string;
929
944
  name: string;
930
- api: "openai-completions";
945
+ api: "openai-responses";
931
946
  provider: string;
932
947
  baseUrl: string;
933
948
  reasoning: true;
934
- input: "text"[];
949
+ input: ("text" | "image")[];
935
950
  cost: {
936
951
  input: number;
937
952
  output: number;
@@ -941,14 +956,14 @@ export declare const MODELS: {
941
956
  contextWindow: number;
942
957
  maxTokens: number;
943
958
  };
944
- readonly "moonshotai/kimi-k2-instruct-0905": {
959
+ readonly "gpt-4o-2024-11-20": {
945
960
  id: string;
946
961
  name: string;
947
- api: "openai-completions";
962
+ api: "openai-responses";
948
963
  provider: string;
949
964
  baseUrl: string;
950
965
  reasoning: false;
951
- input: "text"[];
966
+ input: ("text" | "image")[];
952
967
  cost: {
953
968
  input: number;
954
969
  output: number;
@@ -958,15 +973,393 @@ export declare const MODELS: {
958
973
  contextWindow: number;
959
974
  maxTokens: number;
960
975
  };
961
- readonly "moonshotai/kimi-k2-instruct": {
976
+ readonly o3: {
962
977
  id: string;
963
978
  name: string;
964
- api: "openai-completions";
979
+ api: "openai-responses";
965
980
  provider: string;
966
981
  baseUrl: string;
967
- reasoning: false;
968
- input: "text"[];
969
- cost: {
982
+ reasoning: true;
983
+ input: ("text" | "image")[];
984
+ cost: {
985
+ input: number;
986
+ output: number;
987
+ cacheRead: number;
988
+ cacheWrite: number;
989
+ };
990
+ contextWindow: number;
991
+ maxTokens: number;
992
+ };
993
+ readonly "o4-mini-deep-research": {
994
+ id: string;
995
+ name: string;
996
+ api: "openai-responses";
997
+ provider: string;
998
+ baseUrl: string;
999
+ reasoning: true;
1000
+ input: ("text" | "image")[];
1001
+ cost: {
1002
+ input: number;
1003
+ output: number;
1004
+ cacheRead: number;
1005
+ cacheWrite: number;
1006
+ };
1007
+ contextWindow: number;
1008
+ maxTokens: number;
1009
+ };
1010
+ readonly "gpt-4o-mini": {
1011
+ id: string;
1012
+ name: string;
1013
+ api: "openai-responses";
1014
+ provider: string;
1015
+ baseUrl: string;
1016
+ reasoning: false;
1017
+ input: ("text" | "image")[];
1018
+ cost: {
1019
+ input: number;
1020
+ output: number;
1021
+ cacheRead: number;
1022
+ cacheWrite: number;
1023
+ };
1024
+ contextWindow: number;
1025
+ maxTokens: number;
1026
+ };
1027
+ readonly "gpt-5": {
1028
+ id: string;
1029
+ name: string;
1030
+ api: "openai-responses";
1031
+ provider: string;
1032
+ baseUrl: string;
1033
+ reasoning: true;
1034
+ input: ("text" | "image")[];
1035
+ cost: {
1036
+ input: number;
1037
+ output: number;
1038
+ cacheRead: number;
1039
+ cacheWrite: number;
1040
+ };
1041
+ contextWindow: number;
1042
+ maxTokens: number;
1043
+ };
1044
+ readonly "gpt-5-chat-latest": {
1045
+ id: string;
1046
+ name: string;
1047
+ api: "openai-responses";
1048
+ provider: string;
1049
+ baseUrl: string;
1050
+ reasoning: false;
1051
+ input: ("text" | "image")[];
1052
+ cost: {
1053
+ input: number;
1054
+ output: number;
1055
+ cacheRead: number;
1056
+ cacheWrite: number;
1057
+ };
1058
+ contextWindow: number;
1059
+ maxTokens: number;
1060
+ };
1061
+ };
1062
+ readonly groq: {
1063
+ readonly "llama-3.1-8b-instant": {
1064
+ id: string;
1065
+ name: string;
1066
+ api: "openai-completions";
1067
+ provider: string;
1068
+ baseUrl: string;
1069
+ reasoning: false;
1070
+ input: "text"[];
1071
+ cost: {
1072
+ input: number;
1073
+ output: number;
1074
+ cacheRead: number;
1075
+ cacheWrite: number;
1076
+ };
1077
+ contextWindow: number;
1078
+ maxTokens: number;
1079
+ };
1080
+ readonly "mistral-saba-24b": {
1081
+ id: string;
1082
+ name: string;
1083
+ api: "openai-completions";
1084
+ provider: string;
1085
+ baseUrl: string;
1086
+ reasoning: false;
1087
+ input: "text"[];
1088
+ cost: {
1089
+ input: number;
1090
+ output: number;
1091
+ cacheRead: number;
1092
+ cacheWrite: number;
1093
+ };
1094
+ contextWindow: number;
1095
+ maxTokens: number;
1096
+ };
1097
+ readonly "llama3-8b-8192": {
1098
+ id: string;
1099
+ name: string;
1100
+ api: "openai-completions";
1101
+ provider: string;
1102
+ baseUrl: string;
1103
+ reasoning: false;
1104
+ input: "text"[];
1105
+ cost: {
1106
+ input: number;
1107
+ output: number;
1108
+ cacheRead: number;
1109
+ cacheWrite: number;
1110
+ };
1111
+ contextWindow: number;
1112
+ maxTokens: number;
1113
+ };
1114
+ readonly "qwen-qwq-32b": {
1115
+ id: string;
1116
+ name: string;
1117
+ api: "openai-completions";
1118
+ provider: string;
1119
+ baseUrl: string;
1120
+ reasoning: true;
1121
+ input: "text"[];
1122
+ cost: {
1123
+ input: number;
1124
+ output: number;
1125
+ cacheRead: number;
1126
+ cacheWrite: number;
1127
+ };
1128
+ contextWindow: number;
1129
+ maxTokens: number;
1130
+ };
1131
+ readonly "llama3-70b-8192": {
1132
+ id: string;
1133
+ name: string;
1134
+ api: "openai-completions";
1135
+ provider: string;
1136
+ baseUrl: string;
1137
+ reasoning: false;
1138
+ input: "text"[];
1139
+ cost: {
1140
+ input: number;
1141
+ output: number;
1142
+ cacheRead: number;
1143
+ cacheWrite: number;
1144
+ };
1145
+ contextWindow: number;
1146
+ maxTokens: number;
1147
+ };
1148
+ readonly "deepseek-r1-distill-llama-70b": {
1149
+ id: string;
1150
+ name: string;
1151
+ api: "openai-completions";
1152
+ provider: string;
1153
+ baseUrl: string;
1154
+ reasoning: true;
1155
+ input: "text"[];
1156
+ cost: {
1157
+ input: number;
1158
+ output: number;
1159
+ cacheRead: number;
1160
+ cacheWrite: number;
1161
+ };
1162
+ contextWindow: number;
1163
+ maxTokens: number;
1164
+ };
1165
+ readonly "gemma2-9b-it": {
1166
+ id: string;
1167
+ name: string;
1168
+ api: "openai-completions";
1169
+ provider: string;
1170
+ baseUrl: string;
1171
+ reasoning: false;
1172
+ input: "text"[];
1173
+ cost: {
1174
+ input: number;
1175
+ output: number;
1176
+ cacheRead: number;
1177
+ cacheWrite: number;
1178
+ };
1179
+ contextWindow: number;
1180
+ maxTokens: number;
1181
+ };
1182
+ readonly "llama-3.3-70b-versatile": {
1183
+ id: string;
1184
+ name: string;
1185
+ api: "openai-completions";
1186
+ provider: string;
1187
+ baseUrl: string;
1188
+ reasoning: false;
1189
+ input: "text"[];
1190
+ cost: {
1191
+ input: number;
1192
+ output: number;
1193
+ cacheRead: number;
1194
+ cacheWrite: number;
1195
+ };
1196
+ contextWindow: number;
1197
+ maxTokens: number;
1198
+ };
1199
+ readonly "moonshotai/kimi-k2-instruct-0905": {
1200
+ id: string;
1201
+ name: string;
1202
+ api: "openai-completions";
1203
+ provider: string;
1204
+ baseUrl: string;
1205
+ reasoning: false;
1206
+ input: "text"[];
1207
+ cost: {
1208
+ input: number;
1209
+ output: number;
1210
+ cacheRead: number;
1211
+ cacheWrite: number;
1212
+ };
1213
+ contextWindow: number;
1214
+ maxTokens: number;
1215
+ };
1216
+ readonly "moonshotai/kimi-k2-instruct": {
1217
+ id: string;
1218
+ name: string;
1219
+ api: "openai-completions";
1220
+ provider: string;
1221
+ baseUrl: string;
1222
+ reasoning: false;
1223
+ input: "text"[];
1224
+ cost: {
1225
+ input: number;
1226
+ output: number;
1227
+ cacheRead: number;
1228
+ cacheWrite: number;
1229
+ };
1230
+ contextWindow: number;
1231
+ maxTokens: number;
1232
+ };
1233
+ readonly "openai/gpt-oss-20b": {
1234
+ id: string;
1235
+ name: string;
1236
+ api: "openai-completions";
1237
+ provider: string;
1238
+ baseUrl: string;
1239
+ reasoning: true;
1240
+ input: "text"[];
1241
+ cost: {
1242
+ input: number;
1243
+ output: number;
1244
+ cacheRead: number;
1245
+ cacheWrite: number;
1246
+ };
1247
+ contextWindow: number;
1248
+ maxTokens: number;
1249
+ };
1250
+ readonly "openai/gpt-oss-120b": {
1251
+ id: string;
1252
+ name: string;
1253
+ api: "openai-completions";
1254
+ provider: string;
1255
+ baseUrl: string;
1256
+ reasoning: true;
1257
+ input: "text"[];
1258
+ cost: {
1259
+ input: number;
1260
+ output: number;
1261
+ cacheRead: number;
1262
+ cacheWrite: number;
1263
+ };
1264
+ contextWindow: number;
1265
+ maxTokens: number;
1266
+ };
1267
+ readonly "qwen/qwen3-32b": {
1268
+ id: string;
1269
+ name: string;
1270
+ api: "openai-completions";
1271
+ provider: string;
1272
+ baseUrl: string;
1273
+ reasoning: true;
1274
+ input: "text"[];
1275
+ cost: {
1276
+ input: number;
1277
+ output: number;
1278
+ cacheRead: number;
1279
+ cacheWrite: number;
1280
+ };
1281
+ contextWindow: number;
1282
+ maxTokens: number;
1283
+ };
1284
+ readonly "meta-llama/llama-4-scout-17b-16e-instruct": {
1285
+ id: string;
1286
+ name: string;
1287
+ api: "openai-completions";
1288
+ provider: string;
1289
+ baseUrl: string;
1290
+ reasoning: false;
1291
+ input: ("text" | "image")[];
1292
+ cost: {
1293
+ input: number;
1294
+ output: number;
1295
+ cacheRead: number;
1296
+ cacheWrite: number;
1297
+ };
1298
+ contextWindow: number;
1299
+ maxTokens: number;
1300
+ };
1301
+ readonly "meta-llama/llama-4-maverick-17b-128e-instruct": {
1302
+ id: string;
1303
+ name: string;
1304
+ api: "openai-completions";
1305
+ provider: string;
1306
+ baseUrl: string;
1307
+ reasoning: false;
1308
+ input: ("text" | "image")[];
1309
+ cost: {
1310
+ input: number;
1311
+ output: number;
1312
+ cacheRead: number;
1313
+ cacheWrite: number;
1314
+ };
1315
+ contextWindow: number;
1316
+ maxTokens: number;
1317
+ };
1318
+ };
1319
+ readonly cerebras: {
1320
+ readonly "qwen-3-235b-a22b-instruct-2507": {
1321
+ id: string;
1322
+ name: string;
1323
+ api: "openai-completions";
1324
+ provider: string;
1325
+ baseUrl: string;
1326
+ reasoning: false;
1327
+ input: "text"[];
1328
+ cost: {
1329
+ input: number;
1330
+ output: number;
1331
+ cacheRead: number;
1332
+ cacheWrite: number;
1333
+ };
1334
+ contextWindow: number;
1335
+ maxTokens: number;
1336
+ };
1337
+ readonly "qwen-3-coder-480b": {
1338
+ id: string;
1339
+ name: string;
1340
+ api: "openai-completions";
1341
+ provider: string;
1342
+ baseUrl: string;
1343
+ reasoning: false;
1344
+ input: "text"[];
1345
+ cost: {
1346
+ input: number;
1347
+ output: number;
1348
+ cacheRead: number;
1349
+ cacheWrite: number;
1350
+ };
1351
+ contextWindow: number;
1352
+ maxTokens: number;
1353
+ };
1354
+ readonly "gpt-oss-120b": {
1355
+ id: string;
1356
+ name: string;
1357
+ api: "openai-completions";
1358
+ provider: string;
1359
+ baseUrl: string;
1360
+ reasoning: true;
1361
+ input: "text"[];
1362
+ cost: {
970
1363
  input: number;
971
1364
  output: number;
972
1365
  cacheRead: number;
@@ -976,8 +1369,25 @@ export declare const MODELS: {
976
1369
  maxTokens: number;
977
1370
  };
978
1371
  };
979
- readonly cerebras: {
980
- readonly "qwen-3-235b-a22b-instruct-2507": {
1372
+ readonly xai: {
1373
+ readonly "grok-4-fast-non-reasoning": {
1374
+ id: string;
1375
+ name: string;
1376
+ api: "openai-completions";
1377
+ provider: string;
1378
+ baseUrl: string;
1379
+ reasoning: false;
1380
+ input: ("text" | "image")[];
1381
+ cost: {
1382
+ input: number;
1383
+ output: number;
1384
+ cacheRead: number;
1385
+ cacheWrite: number;
1386
+ };
1387
+ contextWindow: number;
1388
+ maxTokens: number;
1389
+ };
1390
+ readonly "grok-3-fast": {
981
1391
  id: string;
982
1392
  name: string;
983
1393
  api: "openai-completions";
@@ -994,7 +1404,7 @@ export declare const MODELS: {
994
1404
  contextWindow: number;
995
1405
  maxTokens: number;
996
1406
  };
997
- readonly "gpt-oss-120b": {
1407
+ readonly "grok-4": {
998
1408
  id: string;
999
1409
  name: string;
1000
1410
  api: "openai-completions";
@@ -1011,7 +1421,41 @@ export declare const MODELS: {
1011
1421
  contextWindow: number;
1012
1422
  maxTokens: number;
1013
1423
  };
1014
- readonly "qwen-3-coder-480b": {
1424
+ readonly "grok-2-vision": {
1425
+ id: string;
1426
+ name: string;
1427
+ api: "openai-completions";
1428
+ provider: string;
1429
+ baseUrl: string;
1430
+ reasoning: false;
1431
+ input: ("text" | "image")[];
1432
+ cost: {
1433
+ input: number;
1434
+ output: number;
1435
+ cacheRead: number;
1436
+ cacheWrite: number;
1437
+ };
1438
+ contextWindow: number;
1439
+ maxTokens: number;
1440
+ };
1441
+ readonly "grok-code-fast-1": {
1442
+ id: string;
1443
+ name: string;
1444
+ api: "openai-completions";
1445
+ provider: string;
1446
+ baseUrl: string;
1447
+ reasoning: true;
1448
+ input: "text"[];
1449
+ cost: {
1450
+ input: number;
1451
+ output: number;
1452
+ cacheRead: number;
1453
+ cacheWrite: number;
1454
+ };
1455
+ contextWindow: number;
1456
+ maxTokens: number;
1457
+ };
1458
+ readonly "grok-2": {
1015
1459
  id: string;
1016
1460
  name: string;
1017
1461
  api: "openai-completions";
@@ -1028,8 +1472,6 @@ export declare const MODELS: {
1028
1472
  contextWindow: number;
1029
1473
  maxTokens: number;
1030
1474
  };
1031
- };
1032
- readonly xai: {
1033
1475
  readonly "grok-3-mini-fast-latest": {
1034
1476
  id: string;
1035
1477
  name: string;
@@ -1047,13 +1489,64 @@ export declare const MODELS: {
1047
1489
  contextWindow: number;
1048
1490
  maxTokens: number;
1049
1491
  };
1050
- readonly "grok-3-mini-latest": {
1492
+ readonly "grok-2-vision-1212": {
1493
+ id: string;
1494
+ name: string;
1495
+ api: "openai-completions";
1496
+ provider: string;
1497
+ baseUrl: string;
1498
+ reasoning: false;
1499
+ input: ("text" | "image")[];
1500
+ cost: {
1501
+ input: number;
1502
+ output: number;
1503
+ cacheRead: number;
1504
+ cacheWrite: number;
1505
+ };
1506
+ contextWindow: number;
1507
+ maxTokens: number;
1508
+ };
1509
+ readonly "grok-3": {
1510
+ id: string;
1511
+ name: string;
1512
+ api: "openai-completions";
1513
+ provider: string;
1514
+ baseUrl: string;
1515
+ reasoning: false;
1516
+ input: "text"[];
1517
+ cost: {
1518
+ input: number;
1519
+ output: number;
1520
+ cacheRead: number;
1521
+ cacheWrite: number;
1522
+ };
1523
+ contextWindow: number;
1524
+ maxTokens: number;
1525
+ };
1526
+ readonly "grok-4-fast": {
1051
1527
  id: string;
1052
1528
  name: string;
1053
1529
  api: "openai-completions";
1054
1530
  provider: string;
1055
1531
  baseUrl: string;
1056
1532
  reasoning: true;
1533
+ input: ("text" | "image")[];
1534
+ cost: {
1535
+ input: number;
1536
+ output: number;
1537
+ cacheRead: number;
1538
+ cacheWrite: number;
1539
+ };
1540
+ contextWindow: number;
1541
+ maxTokens: number;
1542
+ };
1543
+ readonly "grok-2-latest": {
1544
+ id: string;
1545
+ name: string;
1546
+ api: "openai-completions";
1547
+ provider: string;
1548
+ baseUrl: string;
1549
+ reasoning: false;
1057
1550
  input: "text"[];
1058
1551
  cost: {
1059
1552
  input: number;
@@ -1064,7 +1557,7 @@ export declare const MODELS: {
1064
1557
  contextWindow: number;
1065
1558
  maxTokens: number;
1066
1559
  };
1067
- readonly "grok-beta": {
1560
+ readonly "grok-2-1212": {
1068
1561
  id: string;
1069
1562
  name: string;
1070
1563
  api: "openai-completions";
@@ -1098,13 +1591,151 @@ export declare const MODELS: {
1098
1591
  contextWindow: number;
1099
1592
  maxTokens: number;
1100
1593
  };
1101
- readonly "grok-3": {
1594
+ readonly "grok-3-latest": {
1595
+ id: string;
1596
+ name: string;
1597
+ api: "openai-completions";
1598
+ provider: string;
1599
+ baseUrl: string;
1600
+ reasoning: false;
1601
+ input: "text"[];
1602
+ cost: {
1603
+ input: number;
1604
+ output: number;
1605
+ cacheRead: number;
1606
+ cacheWrite: number;
1607
+ };
1608
+ contextWindow: number;
1609
+ maxTokens: number;
1610
+ };
1611
+ readonly "grok-2-vision-latest": {
1612
+ id: string;
1613
+ name: string;
1614
+ api: "openai-completions";
1615
+ provider: string;
1616
+ baseUrl: string;
1617
+ reasoning: false;
1618
+ input: ("text" | "image")[];
1619
+ cost: {
1620
+ input: number;
1621
+ output: number;
1622
+ cacheRead: number;
1623
+ cacheWrite: number;
1624
+ };
1625
+ contextWindow: number;
1626
+ maxTokens: number;
1627
+ };
1628
+ readonly "grok-vision-beta": {
1629
+ id: string;
1630
+ name: string;
1631
+ api: "openai-completions";
1632
+ provider: string;
1633
+ baseUrl: string;
1634
+ reasoning: false;
1635
+ input: ("text" | "image")[];
1636
+ cost: {
1637
+ input: number;
1638
+ output: number;
1639
+ cacheRead: number;
1640
+ cacheWrite: number;
1641
+ };
1642
+ contextWindow: number;
1643
+ maxTokens: number;
1644
+ };
1645
+ readonly "grok-3-mini": {
1646
+ id: string;
1647
+ name: string;
1648
+ api: "openai-completions";
1649
+ provider: string;
1650
+ baseUrl: string;
1651
+ reasoning: true;
1652
+ input: "text"[];
1653
+ cost: {
1654
+ input: number;
1655
+ output: number;
1656
+ cacheRead: number;
1657
+ cacheWrite: number;
1658
+ };
1659
+ contextWindow: number;
1660
+ maxTokens: number;
1661
+ };
1662
+ readonly "grok-beta": {
1663
+ id: string;
1664
+ name: string;
1665
+ api: "openai-completions";
1666
+ provider: string;
1667
+ baseUrl: string;
1668
+ reasoning: false;
1669
+ input: "text"[];
1670
+ cost: {
1671
+ input: number;
1672
+ output: number;
1673
+ cacheRead: number;
1674
+ cacheWrite: number;
1675
+ };
1676
+ contextWindow: number;
1677
+ maxTokens: number;
1678
+ };
1679
+ readonly "grok-3-mini-latest": {
1680
+ id: string;
1681
+ name: string;
1682
+ api: "openai-completions";
1683
+ provider: string;
1684
+ baseUrl: string;
1685
+ reasoning: true;
1686
+ input: "text"[];
1687
+ cost: {
1688
+ input: number;
1689
+ output: number;
1690
+ cacheRead: number;
1691
+ cacheWrite: number;
1692
+ };
1693
+ contextWindow: number;
1694
+ maxTokens: number;
1695
+ };
1696
+ readonly "grok-3-mini-fast": {
1697
+ id: string;
1698
+ name: string;
1699
+ api: "openai-completions";
1700
+ provider: string;
1701
+ baseUrl: string;
1702
+ reasoning: true;
1703
+ input: "text"[];
1704
+ cost: {
1705
+ input: number;
1706
+ output: number;
1707
+ cacheRead: number;
1708
+ cacheWrite: number;
1709
+ };
1710
+ contextWindow: number;
1711
+ maxTokens: number;
1712
+ };
1713
+ };
1714
+ readonly zai: {
1715
+ readonly "glm-4.5-flash": {
1716
+ id: string;
1717
+ name: string;
1718
+ api: "anthropic-messages";
1719
+ provider: string;
1720
+ baseUrl: string;
1721
+ reasoning: true;
1722
+ input: "text"[];
1723
+ cost: {
1724
+ input: number;
1725
+ output: number;
1726
+ cacheRead: number;
1727
+ cacheWrite: number;
1728
+ };
1729
+ contextWindow: number;
1730
+ maxTokens: number;
1731
+ };
1732
+ readonly "glm-4.5": {
1102
1733
  id: string;
1103
1734
  name: string;
1104
- api: "openai-completions";
1735
+ api: "anthropic-messages";
1105
1736
  provider: string;
1106
1737
  baseUrl: string;
1107
- reasoning: false;
1738
+ reasoning: true;
1108
1739
  input: "text"[];
1109
1740
  cost: {
1110
1741
  input: number;
@@ -1115,10 +1746,10 @@ export declare const MODELS: {
1115
1746
  contextWindow: number;
1116
1747
  maxTokens: number;
1117
1748
  };
1118
- readonly "grok-3-mini": {
1749
+ readonly "glm-4.5-air": {
1119
1750
  id: string;
1120
1751
  name: string;
1121
- api: "openai-completions";
1752
+ api: "anthropic-messages";
1122
1753
  provider: string;
1123
1754
  baseUrl: string;
1124
1755
  reasoning: true;
@@ -1132,13 +1763,13 @@ export declare const MODELS: {
1132
1763
  contextWindow: number;
1133
1764
  maxTokens: number;
1134
1765
  };
1135
- readonly "grok-2-vision-1212": {
1766
+ readonly "glm-4.5v": {
1136
1767
  id: string;
1137
1768
  name: string;
1138
- api: "openai-completions";
1769
+ api: "anthropic-messages";
1139
1770
  provider: string;
1140
1771
  baseUrl: string;
1141
- reasoning: false;
1772
+ reasoning: true;
1142
1773
  input: ("text" | "image")[];
1143
1774
  cost: {
1144
1775
  input: number;
@@ -1149,13 +1780,13 @@ export declare const MODELS: {
1149
1780
  contextWindow: number;
1150
1781
  maxTokens: number;
1151
1782
  };
1152
- readonly "grok-2": {
1783
+ readonly "glm-4.6": {
1153
1784
  id: string;
1154
1785
  name: string;
1155
- api: "openai-completions";
1786
+ api: "anthropic-messages";
1156
1787
  provider: string;
1157
1788
  baseUrl: string;
1158
- reasoning: false;
1789
+ reasoning: true;
1159
1790
  input: "text"[];
1160
1791
  cost: {
1161
1792
  input: number;
@@ -1166,13 +1797,15 @@ export declare const MODELS: {
1166
1797
  contextWindow: number;
1167
1798
  maxTokens: number;
1168
1799
  };
1169
- readonly "grok-2-vision-latest": {
1800
+ };
1801
+ readonly openrouter: {
1802
+ readonly "openrouter/andromeda-alpha": {
1170
1803
  id: string;
1171
1804
  name: string;
1172
1805
  api: "openai-completions";
1173
1806
  provider: string;
1174
1807
  baseUrl: string;
1175
- reasoning: false;
1808
+ reasoning: true;
1176
1809
  input: ("text" | "image")[];
1177
1810
  cost: {
1178
1811
  input: number;
@@ -1183,13 +1816,13 @@ export declare const MODELS: {
1183
1816
  contextWindow: number;
1184
1817
  maxTokens: number;
1185
1818
  };
1186
- readonly "grok-3-latest": {
1819
+ readonly "deepcogito/cogito-v2-preview-llama-405b": {
1187
1820
  id: string;
1188
1821
  name: string;
1189
1822
  api: "openai-completions";
1190
1823
  provider: string;
1191
1824
  baseUrl: string;
1192
- reasoning: false;
1825
+ reasoning: true;
1193
1826
  input: "text"[];
1194
1827
  cost: {
1195
1828
  input: number;
@@ -1200,13 +1833,13 @@ export declare const MODELS: {
1200
1833
  contextWindow: number;
1201
1834
  maxTokens: number;
1202
1835
  };
1203
- readonly "grok-2-vision": {
1836
+ readonly "qwen/qwen3-vl-8b-thinking": {
1204
1837
  id: string;
1205
1838
  name: string;
1206
1839
  api: "openai-completions";
1207
1840
  provider: string;
1208
1841
  baseUrl: string;
1209
- reasoning: false;
1842
+ reasoning: true;
1210
1843
  input: ("text" | "image")[];
1211
1844
  cost: {
1212
1845
  input: number;
@@ -1217,14 +1850,14 @@ export declare const MODELS: {
1217
1850
  contextWindow: number;
1218
1851
  maxTokens: number;
1219
1852
  };
1220
- readonly "grok-2-latest": {
1853
+ readonly "qwen/qwen3-vl-8b-instruct": {
1221
1854
  id: string;
1222
1855
  name: string;
1223
1856
  api: "openai-completions";
1224
1857
  provider: string;
1225
1858
  baseUrl: string;
1226
1859
  reasoning: false;
1227
- input: "text"[];
1860
+ input: ("text" | "image")[];
1228
1861
  cost: {
1229
1862
  input: number;
1230
1863
  output: number;
@@ -1234,13 +1867,13 @@ export declare const MODELS: {
1234
1867
  contextWindow: number;
1235
1868
  maxTokens: number;
1236
1869
  };
1237
- readonly "grok-3-fast": {
1870
+ readonly "inclusionai/ring-1t": {
1238
1871
  id: string;
1239
1872
  name: string;
1240
1873
  api: "openai-completions";
1241
1874
  provider: string;
1242
1875
  baseUrl: string;
1243
- reasoning: false;
1876
+ reasoning: true;
1244
1877
  input: "text"[];
1245
1878
  cost: {
1246
1879
  input: number;
@@ -1251,7 +1884,7 @@ export declare const MODELS: {
1251
1884
  contextWindow: number;
1252
1885
  maxTokens: number;
1253
1886
  };
1254
- readonly "grok-2-1212": {
1887
+ readonly "inclusionai/ling-1t": {
1255
1888
  id: string;
1256
1889
  name: string;
1257
1890
  api: "openai-completions";
@@ -1268,7 +1901,7 @@ export declare const MODELS: {
1268
1901
  contextWindow: number;
1269
1902
  maxTokens: number;
1270
1903
  };
1271
- readonly "grok-4": {
1904
+ readonly "nvidia/llama-3.3-nemotron-super-49b-v1.5": {
1272
1905
  id: string;
1273
1906
  name: string;
1274
1907
  api: "openai-completions";
@@ -1285,13 +1918,13 @@ export declare const MODELS: {
1285
1918
  contextWindow: number;
1286
1919
  maxTokens: number;
1287
1920
  };
1288
- readonly "grok-vision-beta": {
1921
+ readonly "qwen/qwen3-vl-30b-a3b-thinking": {
1289
1922
  id: string;
1290
1923
  name: string;
1291
1924
  api: "openai-completions";
1292
1925
  provider: string;
1293
1926
  baseUrl: string;
1294
- reasoning: false;
1927
+ reasoning: true;
1295
1928
  input: ("text" | "image")[];
1296
1929
  cost: {
1297
1930
  input: number;
@@ -1302,14 +1935,14 @@ export declare const MODELS: {
1302
1935
  contextWindow: number;
1303
1936
  maxTokens: number;
1304
1937
  };
1305
- readonly "grok-3-mini-fast": {
1938
+ readonly "qwen/qwen3-vl-30b-a3b-instruct": {
1306
1939
  id: string;
1307
1940
  name: string;
1308
1941
  api: "openai-completions";
1309
1942
  provider: string;
1310
1943
  baseUrl: string;
1311
- reasoning: true;
1312
- input: "text"[];
1944
+ reasoning: false;
1945
+ input: ("text" | "image")[];
1313
1946
  cost: {
1314
1947
  input: number;
1315
1948
  output: number;
@@ -1319,13 +1952,13 @@ export declare const MODELS: {
1319
1952
  contextWindow: number;
1320
1953
  maxTokens: number;
1321
1954
  };
1322
- readonly "grok-code-fast-1": {
1955
+ readonly "z-ai/glm-4.6": {
1323
1956
  id: string;
1324
1957
  name: string;
1325
1958
  api: "openai-completions";
1326
1959
  provider: string;
1327
1960
  baseUrl: string;
1328
- reasoning: false;
1961
+ reasoning: true;
1329
1962
  input: "text"[];
1330
1963
  cost: {
1331
1964
  input: number;
@@ -1336,12 +1969,10 @@ export declare const MODELS: {
1336
1969
  contextWindow: number;
1337
1970
  maxTokens: number;
1338
1971
  };
1339
- };
1340
- readonly zai: {
1341
- readonly "glm-4.5-air": {
1972
+ readonly "deepseek/deepseek-v3.2-exp": {
1342
1973
  id: string;
1343
1974
  name: string;
1344
- api: "anthropic-messages";
1975
+ api: "openai-completions";
1345
1976
  provider: string;
1346
1977
  baseUrl: string;
1347
1978
  reasoning: true;
@@ -1355,10 +1986,10 @@ export declare const MODELS: {
1355
1986
  contextWindow: number;
1356
1987
  maxTokens: number;
1357
1988
  };
1358
- readonly "glm-4.5v": {
1989
+ readonly "qwen/qwen3-vl-235b-a22b-thinking": {
1359
1990
  id: string;
1360
1991
  name: string;
1361
- api: "anthropic-messages";
1992
+ api: "openai-completions";
1362
1993
  provider: string;
1363
1994
  baseUrl: string;
1364
1995
  reasoning: true;
@@ -1372,14 +2003,14 @@ export declare const MODELS: {
1372
2003
  contextWindow: number;
1373
2004
  maxTokens: number;
1374
2005
  };
1375
- readonly "glm-4.5-flash": {
2006
+ readonly "qwen/qwen3-vl-235b-a22b-instruct": {
1376
2007
  id: string;
1377
2008
  name: string;
1378
- api: "anthropic-messages";
2009
+ api: "openai-completions";
1379
2010
  provider: string;
1380
2011
  baseUrl: string;
1381
2012
  reasoning: true;
1382
- input: "text"[];
2013
+ input: ("text" | "image")[];
1383
2014
  cost: {
1384
2015
  input: number;
1385
2016
  output: number;
@@ -1389,13 +2020,13 @@ export declare const MODELS: {
1389
2020
  contextWindow: number;
1390
2021
  maxTokens: number;
1391
2022
  };
1392
- readonly "glm-4.5": {
2023
+ readonly "qwen/qwen3-max": {
1393
2024
  id: string;
1394
2025
  name: string;
1395
- api: "anthropic-messages";
2026
+ api: "openai-completions";
1396
2027
  provider: string;
1397
2028
  baseUrl: string;
1398
- reasoning: true;
2029
+ reasoning: false;
1399
2030
  input: "text"[];
1400
2031
  cost: {
1401
2032
  input: number;
@@ -1406,15 +2037,13 @@ export declare const MODELS: {
1406
2037
  contextWindow: number;
1407
2038
  maxTokens: number;
1408
2039
  };
1409
- };
1410
- readonly openrouter: {
1411
- readonly "alibaba/tongyi-deepresearch-30b-a3b": {
2040
+ readonly "qwen/qwen3-coder-plus": {
1412
2041
  id: string;
1413
2042
  name: string;
1414
2043
  api: "openai-completions";
1415
2044
  provider: string;
1416
2045
  baseUrl: string;
1417
- reasoning: true;
2046
+ reasoning: false;
1418
2047
  input: "text"[];
1419
2048
  cost: {
1420
2049
  input: number;
@@ -1425,13 +2054,13 @@ export declare const MODELS: {
1425
2054
  contextWindow: number;
1426
2055
  maxTokens: number;
1427
2056
  };
1428
- readonly "qwen/qwen3-coder-flash": {
2057
+ readonly "deepseek/deepseek-v3.1-terminus": {
1429
2058
  id: string;
1430
2059
  name: string;
1431
2060
  api: "openai-completions";
1432
2061
  provider: string;
1433
2062
  baseUrl: string;
1434
- reasoning: false;
2063
+ reasoning: true;
1435
2064
  input: "text"[];
1436
2065
  cost: {
1437
2066
  input: number;
@@ -1442,13 +2071,13 @@ export declare const MODELS: {
1442
2071
  contextWindow: number;
1443
2072
  maxTokens: number;
1444
2073
  };
1445
- readonly "qwen/qwen3-coder-plus": {
2074
+ readonly "alibaba/tongyi-deepresearch-30b-a3b:free": {
1446
2075
  id: string;
1447
2076
  name: string;
1448
2077
  api: "openai-completions";
1449
2078
  provider: string;
1450
2079
  baseUrl: string;
1451
- reasoning: false;
2080
+ reasoning: true;
1452
2081
  input: "text"[];
1453
2082
  cost: {
1454
2083
  input: number;
@@ -1459,7 +2088,7 @@ export declare const MODELS: {
1459
2088
  contextWindow: number;
1460
2089
  maxTokens: number;
1461
2090
  };
1462
- readonly "qwen/qwen3-next-80b-a3b-thinking": {
2091
+ readonly "alibaba/tongyi-deepresearch-30b-a3b": {
1463
2092
  id: string;
1464
2093
  name: string;
1465
2094
  api: "openai-completions";
@@ -1476,7 +2105,7 @@ export declare const MODELS: {
1476
2105
  contextWindow: number;
1477
2106
  maxTokens: number;
1478
2107
  };
1479
- readonly "qwen/qwen3-next-80b-a3b-instruct": {
2108
+ readonly "qwen/qwen3-coder-flash": {
1480
2109
  id: string;
1481
2110
  name: string;
1482
2111
  api: "openai-completions";
@@ -1493,13 +2122,13 @@ export declare const MODELS: {
1493
2122
  contextWindow: number;
1494
2123
  maxTokens: number;
1495
2124
  };
1496
- readonly "meituan/longcat-flash-chat": {
2125
+ readonly "qwen/qwen3-next-80b-a3b-thinking": {
1497
2126
  id: string;
1498
2127
  name: string;
1499
2128
  api: "openai-completions";
1500
2129
  provider: string;
1501
2130
  baseUrl: string;
1502
- reasoning: false;
2131
+ reasoning: true;
1503
2132
  input: "text"[];
1504
2133
  cost: {
1505
2134
  input: number;
@@ -1510,7 +2139,7 @@ export declare const MODELS: {
1510
2139
  contextWindow: number;
1511
2140
  maxTokens: number;
1512
2141
  };
1513
- readonly "qwen/qwen-plus-2025-07-28": {
2142
+ readonly "qwen/qwen3-next-80b-a3b-instruct": {
1514
2143
  id: string;
1515
2144
  name: string;
1516
2145
  api: "openai-completions";
@@ -1527,13 +2156,13 @@ export declare const MODELS: {
1527
2156
  contextWindow: number;
1528
2157
  maxTokens: number;
1529
2158
  };
1530
- readonly "qwen/qwen-plus-2025-07-28:thinking": {
2159
+ readonly "meituan/longcat-flash-chat:free": {
1531
2160
  id: string;
1532
2161
  name: string;
1533
2162
  api: "openai-completions";
1534
2163
  provider: string;
1535
2164
  baseUrl: string;
1536
- reasoning: true;
2165
+ reasoning: false;
1537
2166
  input: "text"[];
1538
2167
  cost: {
1539
2168
  input: number;
@@ -1544,13 +2173,13 @@ export declare const MODELS: {
1544
2173
  contextWindow: number;
1545
2174
  maxTokens: number;
1546
2175
  };
1547
- readonly "nvidia/nemotron-nano-9b-v2:free": {
2176
+ readonly "qwen/qwen-plus-2025-07-28": {
1548
2177
  id: string;
1549
2178
  name: string;
1550
2179
  api: "openai-completions";
1551
2180
  provider: string;
1552
2181
  baseUrl: string;
1553
- reasoning: true;
2182
+ reasoning: false;
1554
2183
  input: "text"[];
1555
2184
  cost: {
1556
2185
  input: number;
@@ -1561,7 +2190,7 @@ export declare const MODELS: {
1561
2190
  contextWindow: number;
1562
2191
  maxTokens: number;
1563
2192
  };
1564
- readonly "nvidia/nemotron-nano-9b-v2": {
2193
+ readonly "qwen/qwen-plus-2025-07-28:thinking": {
1565
2194
  id: string;
1566
2195
  name: string;
1567
2196
  api: "openai-completions";
@@ -1578,14 +2207,14 @@ export declare const MODELS: {
1578
2207
  contextWindow: number;
1579
2208
  maxTokens: number;
1580
2209
  };
1581
- readonly "openrouter/sonoma-dusk-alpha": {
2210
+ readonly "nvidia/nemotron-nano-9b-v2:free": {
1582
2211
  id: string;
1583
2212
  name: string;
1584
2213
  api: "openai-completions";
1585
2214
  provider: string;
1586
2215
  baseUrl: string;
1587
- reasoning: false;
1588
- input: ("text" | "image")[];
2216
+ reasoning: true;
2217
+ input: "text"[];
1589
2218
  cost: {
1590
2219
  input: number;
1591
2220
  output: number;
@@ -1595,14 +2224,14 @@ export declare const MODELS: {
1595
2224
  contextWindow: number;
1596
2225
  maxTokens: number;
1597
2226
  };
1598
- readonly "openrouter/sonoma-sky-alpha": {
2227
+ readonly "nvidia/nemotron-nano-9b-v2": {
1599
2228
  id: string;
1600
2229
  name: string;
1601
2230
  api: "openai-completions";
1602
2231
  provider: string;
1603
2232
  baseUrl: string;
1604
2233
  reasoning: true;
1605
- input: ("text" | "image")[];
2234
+ input: "text"[];
1606
2235
  cost: {
1607
2236
  input: number;
1608
2237
  output: number;
@@ -1612,7 +2241,7 @@ export declare const MODELS: {
1612
2241
  contextWindow: number;
1613
2242
  maxTokens: number;
1614
2243
  };
1615
- readonly "qwen/qwen3-max": {
2244
+ readonly "moonshotai/kimi-k2-0905": {
1616
2245
  id: string;
1617
2246
  name: string;
1618
2247
  api: "openai-completions";
@@ -1629,13 +2258,13 @@ export declare const MODELS: {
1629
2258
  contextWindow: number;
1630
2259
  maxTokens: number;
1631
2260
  };
1632
- readonly "moonshotai/kimi-k2-0905": {
2261
+ readonly "deepcogito/cogito-v2-preview-llama-70b": {
1633
2262
  id: string;
1634
2263
  name: string;
1635
2264
  api: "openai-completions";
1636
2265
  provider: string;
1637
2266
  baseUrl: string;
1638
- reasoning: false;
2267
+ reasoning: true;
1639
2268
  input: "text"[];
1640
2269
  cost: {
1641
2270
  input: number;
@@ -1731,7 +2360,7 @@ export declare const MODELS: {
1731
2360
  contextWindow: number;
1732
2361
  maxTokens: number;
1733
2362
  };
1734
- readonly "deepseek/deepseek-chat-v3.1:free": {
2363
+ readonly "deepseek/deepseek-chat-v3.1": {
1735
2364
  id: string;
1736
2365
  name: string;
1737
2366
  api: "openai-completions";
@@ -1748,14 +2377,14 @@ export declare const MODELS: {
1748
2377
  contextWindow: number;
1749
2378
  maxTokens: number;
1750
2379
  };
1751
- readonly "deepseek/deepseek-chat-v3.1": {
2380
+ readonly "mistralai/mistral-medium-3.1": {
1752
2381
  id: string;
1753
2382
  name: string;
1754
2383
  api: "openai-completions";
1755
2384
  provider: string;
1756
2385
  baseUrl: string;
1757
- reasoning: true;
1758
- input: "text"[];
2386
+ reasoning: false;
2387
+ input: ("text" | "image")[];
1759
2388
  cost: {
1760
2389
  input: number;
1761
2390
  output: number;
@@ -1765,13 +2394,30 @@ export declare const MODELS: {
1765
2394
  contextWindow: number;
1766
2395
  maxTokens: number;
1767
2396
  };
1768
- readonly "mistralai/mistral-medium-3.1": {
2397
+ readonly "baidu/ernie-4.5-21b-a3b": {
1769
2398
  id: string;
1770
2399
  name: string;
1771
2400
  api: "openai-completions";
1772
2401
  provider: string;
1773
2402
  baseUrl: string;
1774
2403
  reasoning: false;
2404
+ input: "text"[];
2405
+ cost: {
2406
+ input: number;
2407
+ output: number;
2408
+ cacheRead: number;
2409
+ cacheWrite: number;
2410
+ };
2411
+ contextWindow: number;
2412
+ maxTokens: number;
2413
+ };
2414
+ readonly "baidu/ernie-4.5-vl-28b-a3b": {
2415
+ id: string;
2416
+ name: string;
2417
+ api: "openai-completions";
2418
+ provider: string;
2419
+ baseUrl: string;
2420
+ reasoning: true;
1775
2421
  input: ("text" | "image")[];
1776
2422
  cost: {
1777
2423
  input: number;
@@ -1992,7 +2638,7 @@ export declare const MODELS: {
1992
2638
  api: "openai-completions";
1993
2639
  provider: string;
1994
2640
  baseUrl: string;
1995
- reasoning: false;
2641
+ reasoning: true;
1996
2642
  input: "text"[];
1997
2643
  cost: {
1998
2644
  input: number;
@@ -2009,7 +2655,7 @@ export declare const MODELS: {
2009
2655
  api: "openai-completions";
2010
2656
  provider: string;
2011
2657
  baseUrl: string;
2012
- reasoning: false;
2658
+ reasoning: true;
2013
2659
  input: "text"[];
2014
2660
  cost: {
2015
2661
  input: number;
@@ -2020,7 +2666,7 @@ export declare const MODELS: {
2020
2666
  contextWindow: number;
2021
2667
  maxTokens: number;
2022
2668
  };
2023
- readonly "moonshotai/kimi-k2:free": {
2669
+ readonly "moonshotai/kimi-k2": {
2024
2670
  id: string;
2025
2671
  name: string;
2026
2672
  api: "openai-completions";
@@ -2037,7 +2683,7 @@ export declare const MODELS: {
2037
2683
  contextWindow: number;
2038
2684
  maxTokens: number;
2039
2685
  };
2040
- readonly "moonshotai/kimi-k2": {
2686
+ readonly "mistralai/devstral-medium": {
2041
2687
  id: string;
2042
2688
  name: string;
2043
2689
  api: "openai-completions";
@@ -2054,7 +2700,7 @@ export declare const MODELS: {
2054
2700
  contextWindow: number;
2055
2701
  maxTokens: number;
2056
2702
  };
2057
- readonly "mistralai/devstral-medium": {
2703
+ readonly "mistralai/devstral-small": {
2058
2704
  id: string;
2059
2705
  name: string;
2060
2706
  api: "openai-completions";
@@ -2071,13 +2717,13 @@ export declare const MODELS: {
2071
2717
  contextWindow: number;
2072
2718
  maxTokens: number;
2073
2719
  };
2074
- readonly "mistralai/devstral-small": {
2720
+ readonly "tngtech/deepseek-r1t2-chimera": {
2075
2721
  id: string;
2076
2722
  name: string;
2077
2723
  api: "openai-completions";
2078
2724
  provider: string;
2079
2725
  baseUrl: string;
2080
- reasoning: false;
2726
+ reasoning: true;
2081
2727
  input: "text"[];
2082
2728
  cost: {
2083
2729
  input: number;
@@ -2275,6 +2921,23 @@ export declare const MODELS: {
2275
2921
  contextWindow: number;
2276
2922
  maxTokens: number;
2277
2923
  };
2924
+ readonly "nousresearch/deephermes-3-mistral-24b-preview": {
2925
+ id: string;
2926
+ name: string;
2927
+ api: "openai-completions";
2928
+ provider: string;
2929
+ baseUrl: string;
2930
+ reasoning: true;
2931
+ input: "text"[];
2932
+ cost: {
2933
+ input: number;
2934
+ output: number;
2935
+ cacheRead: number;
2936
+ cacheWrite: number;
2937
+ };
2938
+ contextWindow: number;
2939
+ maxTokens: number;
2940
+ };
2278
2941
  readonly "mistralai/mistral-medium-3": {
2279
2942
  id: string;
2280
2943
  name: string;
@@ -2519,7 +3182,7 @@ export declare const MODELS: {
2519
3182
  api: "openai-completions";
2520
3183
  provider: string;
2521
3184
  baseUrl: string;
2522
- reasoning: false;
3185
+ reasoning: true;
2523
3186
  input: "text"[];
2524
3187
  cost: {
2525
3188
  input: number;
@@ -2564,14 +3227,14 @@ export declare const MODELS: {
2564
3227
  contextWindow: number;
2565
3228
  maxTokens: number;
2566
3229
  };
2567
- readonly "microsoft/phi-4-multimodal-instruct": {
3230
+ readonly "qwen/qwq-32b": {
2568
3231
  id: string;
2569
3232
  name: string;
2570
3233
  api: "openai-completions";
2571
3234
  provider: string;
2572
3235
  baseUrl: string;
2573
- reasoning: false;
2574
- input: ("text" | "image")[];
3236
+ reasoning: true;
3237
+ input: "text"[];
2575
3238
  cost: {
2576
3239
  input: number;
2577
3240
  output: number;
@@ -2581,13 +3244,13 @@ export declare const MODELS: {
2581
3244
  contextWindow: number;
2582
3245
  maxTokens: number;
2583
3246
  };
2584
- readonly "qwen/qwq-32b": {
3247
+ readonly "nousresearch/deephermes-3-llama-3-8b-preview": {
2585
3248
  id: string;
2586
3249
  name: string;
2587
3250
  api: "openai-completions";
2588
3251
  provider: string;
2589
3252
  baseUrl: string;
2590
- reasoning: true;
3253
+ reasoning: false;
2591
3254
  input: "text"[];
2592
3255
  cost: {
2593
3256
  input: number;
@@ -2615,6 +3278,23 @@ export declare const MODELS: {
2615
3278
  contextWindow: number;
2616
3279
  maxTokens: number;
2617
3280
  };
3281
+ readonly "qwen/qwen-vl-max": {
3282
+ id: string;
3283
+ name: string;
3284
+ api: "openai-completions";
3285
+ provider: string;
3286
+ baseUrl: string;
3287
+ reasoning: false;
3288
+ input: ("text" | "image")[];
3289
+ cost: {
3290
+ input: number;
3291
+ output: number;
3292
+ cacheRead: number;
3293
+ cacheWrite: number;
3294
+ };
3295
+ contextWindow: number;
3296
+ maxTokens: number;
3297
+ };
2618
3298
  readonly "qwen/qwen-turbo": {
2619
3299
  id: string;
2620
3300
  name: string;
@@ -2921,7 +3601,7 @@ export declare const MODELS: {
2921
3601
  contextWindow: number;
2922
3602
  maxTokens: number;
2923
3603
  };
2924
- readonly "nvidia/llama-3.1-nemotron-70b-instruct": {
3604
+ readonly "qwen/qwen-2.5-7b-instruct": {
2925
3605
  id: string;
2926
3606
  name: string;
2927
3607
  api: "openai-completions";
@@ -2938,7 +3618,7 @@ export declare const MODELS: {
2938
3618
  contextWindow: number;
2939
3619
  maxTokens: number;
2940
3620
  };
2941
- readonly "thedrummer/rocinante-12b": {
3621
+ readonly "nvidia/llama-3.1-nemotron-70b-instruct": {
2942
3622
  id: string;
2943
3623
  name: string;
2944
3624
  api: "openai-completions";
@@ -2955,7 +3635,7 @@ export declare const MODELS: {
2955
3635
  contextWindow: number;
2956
3636
  maxTokens: number;
2957
3637
  };
2958
- readonly "meta-llama/llama-3.2-3b-instruct": {
3638
+ readonly "thedrummer/rocinante-12b": {
2959
3639
  id: string;
2960
3640
  name: string;
2961
3641
  api: "openai-completions";
@@ -2972,7 +3652,7 @@ export declare const MODELS: {
2972
3652
  contextWindow: number;
2973
3653
  maxTokens: number;
2974
3654
  };
2975
- readonly "qwen/qwen-2.5-72b-instruct": {
3655
+ readonly "meta-llama/llama-3.2-3b-instruct": {
2976
3656
  id: string;
2977
3657
  name: string;
2978
3658
  api: "openai-completions";
@@ -2989,24 +3669,7 @@ export declare const MODELS: {
2989
3669
  contextWindow: number;
2990
3670
  maxTokens: number;
2991
3671
  };
2992
- readonly "mistralai/pixtral-12b": {
2993
- id: string;
2994
- name: string;
2995
- api: "openai-completions";
2996
- provider: string;
2997
- baseUrl: string;
2998
- reasoning: false;
2999
- input: ("text" | "image")[];
3000
- cost: {
3001
- input: number;
3002
- output: number;
3003
- cacheRead: number;
3004
- cacheWrite: number;
3005
- };
3006
- contextWindow: number;
3007
- maxTokens: number;
3008
- };
3009
- readonly "cohere/command-r-08-2024": {
3672
+ readonly "qwen/qwen-2.5-72b-instruct": {
3010
3673
  id: string;
3011
3674
  name: string;
3012
3675
  api: "openai-completions";
@@ -3023,14 +3686,14 @@ export declare const MODELS: {
3023
3686
  contextWindow: number;
3024
3687
  maxTokens: number;
3025
3688
  };
3026
- readonly "cohere/command-r-plus-08-2024": {
3689
+ readonly "mistralai/pixtral-12b": {
3027
3690
  id: string;
3028
3691
  name: string;
3029
3692
  api: "openai-completions";
3030
3693
  provider: string;
3031
3694
  baseUrl: string;
3032
3695
  reasoning: false;
3033
- input: "text"[];
3696
+ input: ("text" | "image")[];
3034
3697
  cost: {
3035
3698
  input: number;
3036
3699
  output: number;
@@ -3040,7 +3703,7 @@ export declare const MODELS: {
3040
3703
  contextWindow: number;
3041
3704
  maxTokens: number;
3042
3705
  };
3043
- readonly "microsoft/phi-3.5-mini-128k-instruct": {
3706
+ readonly "cohere/command-r-plus-08-2024": {
3044
3707
  id: string;
3045
3708
  name: string;
3046
3709
  api: "openai-completions";
@@ -3057,7 +3720,7 @@ export declare const MODELS: {
3057
3720
  contextWindow: number;
3058
3721
  maxTokens: number;
3059
3722
  };
3060
- readonly "nousresearch/hermes-3-llama-3.1-70b": {
3723
+ readonly "cohere/command-r-08-2024": {
3061
3724
  id: string;
3062
3725
  name: string;
3063
3726
  api: "openai-completions";
@@ -3074,7 +3737,7 @@ export declare const MODELS: {
3074
3737
  contextWindow: number;
3075
3738
  maxTokens: number;
3076
3739
  };
3077
- readonly "meta-llama/llama-3.1-8b-instruct": {
3740
+ readonly "sao10k/l3.1-euryale-70b": {
3078
3741
  id: string;
3079
3742
  name: string;
3080
3743
  api: "openai-completions";
@@ -3091,7 +3754,7 @@ export declare const MODELS: {
3091
3754
  contextWindow: number;
3092
3755
  maxTokens: number;
3093
3756
  };
3094
- readonly "meta-llama/llama-3.1-405b-instruct": {
3757
+ readonly "microsoft/phi-3.5-mini-128k-instruct": {
3095
3758
  id: string;
3096
3759
  name: string;
3097
3760
  api: "openai-completions";
@@ -3108,7 +3771,7 @@ export declare const MODELS: {
3108
3771
  contextWindow: number;
3109
3772
  maxTokens: number;
3110
3773
  };
3111
- readonly "meta-llama/llama-3.1-70b-instruct": {
3774
+ readonly "nousresearch/hermes-3-llama-3.1-70b": {
3112
3775
  id: string;
3113
3776
  name: string;
3114
3777
  api: "openai-completions";
@@ -3125,7 +3788,7 @@ export declare const MODELS: {
3125
3788
  contextWindow: number;
3126
3789
  maxTokens: number;
3127
3790
  };
3128
- readonly "mistralai/mistral-nemo": {
3791
+ readonly "meta-llama/llama-3.1-8b-instruct": {
3129
3792
  id: string;
3130
3793
  name: string;
3131
3794
  api: "openai-completions";
@@ -3142,7 +3805,7 @@ export declare const MODELS: {
3142
3805
  contextWindow: number;
3143
3806
  maxTokens: number;
3144
3807
  };
3145
- readonly "mistralai/mistral-7b-instruct-v0.3": {
3808
+ readonly "meta-llama/llama-3.1-405b-instruct": {
3146
3809
  id: string;
3147
3810
  name: string;
3148
3811
  api: "openai-completions";
@@ -3159,7 +3822,7 @@ export declare const MODELS: {
3159
3822
  contextWindow: number;
3160
3823
  maxTokens: number;
3161
3824
  };
3162
- readonly "mistralai/mistral-7b-instruct:free": {
3825
+ readonly "meta-llama/llama-3.1-70b-instruct": {
3163
3826
  id: string;
3164
3827
  name: string;
3165
3828
  api: "openai-completions";
@@ -3176,7 +3839,7 @@ export declare const MODELS: {
3176
3839
  contextWindow: number;
3177
3840
  maxTokens: number;
3178
3841
  };
3179
- readonly "mistralai/mistral-7b-instruct": {
3842
+ readonly "mistralai/mistral-nemo": {
3180
3843
  id: string;
3181
3844
  name: string;
3182
3845
  api: "openai-completions";
@@ -3193,7 +3856,7 @@ export declare const MODELS: {
3193
3856
  contextWindow: number;
3194
3857
  maxTokens: number;
3195
3858
  };
3196
- readonly "microsoft/phi-3-mini-128k-instruct": {
3859
+ readonly "sao10k/l3-euryale-70b": {
3197
3860
  id: string;
3198
3861
  name: string;
3199
3862
  api: "openai-completions";
@@ -3210,7 +3873,7 @@ export declare const MODELS: {
3210
3873
  contextWindow: number;
3211
3874
  maxTokens: number;
3212
3875
  };
3213
- readonly "microsoft/phi-3-medium-128k-instruct": {
3876
+ readonly "mistralai/mistral-7b-instruct:free": {
3214
3877
  id: string;
3215
3878
  name: string;
3216
3879
  api: "openai-completions";
@@ -3227,7 +3890,7 @@ export declare const MODELS: {
3227
3890
  contextWindow: number;
3228
3891
  maxTokens: number;
3229
3892
  };
3230
- readonly "meta-llama/llama-3-8b-instruct": {
3893
+ readonly "mistralai/mistral-7b-instruct": {
3231
3894
  id: string;
3232
3895
  name: string;
3233
3896
  api: "openai-completions";
@@ -3244,7 +3907,7 @@ export declare const MODELS: {
3244
3907
  contextWindow: number;
3245
3908
  maxTokens: number;
3246
3909
  };
3247
- readonly "meta-llama/llama-3-70b-instruct": {
3910
+ readonly "mistralai/mistral-7b-instruct-v0.3": {
3248
3911
  id: string;
3249
3912
  name: string;
3250
3913
  api: "openai-completions";
@@ -3261,7 +3924,7 @@ export declare const MODELS: {
3261
3924
  contextWindow: number;
3262
3925
  maxTokens: number;
3263
3926
  };
3264
- readonly "mistralai/mixtral-8x22b-instruct": {
3927
+ readonly "microsoft/phi-3-mini-128k-instruct": {
3265
3928
  id: string;
3266
3929
  name: string;
3267
3930
  api: "openai-completions";
@@ -3278,7 +3941,7 @@ export declare const MODELS: {
3278
3941
  contextWindow: number;
3279
3942
  maxTokens: number;
3280
3943
  };
3281
- readonly "cohere/command-r-plus": {
3944
+ readonly "microsoft/phi-3-medium-128k-instruct": {
3282
3945
  id: string;
3283
3946
  name: string;
3284
3947
  api: "openai-completions";
@@ -3295,7 +3958,7 @@ export declare const MODELS: {
3295
3958
  contextWindow: number;
3296
3959
  maxTokens: number;
3297
3960
  };
3298
- readonly "cohere/command-r-plus-04-2024": {
3961
+ readonly "meta-llama/llama-3-8b-instruct": {
3299
3962
  id: string;
3300
3963
  name: string;
3301
3964
  api: "openai-completions";
@@ -3312,7 +3975,7 @@ export declare const MODELS: {
3312
3975
  contextWindow: number;
3313
3976
  maxTokens: number;
3314
3977
  };
3315
- readonly "cohere/command-r": {
3978
+ readonly "meta-llama/llama-3-70b-instruct": {
3316
3979
  id: string;
3317
3980
  name: string;
3318
3981
  api: "openai-completions";
@@ -3329,7 +3992,7 @@ export declare const MODELS: {
3329
3992
  contextWindow: number;
3330
3993
  maxTokens: number;
3331
3994
  };
3332
- readonly "cohere/command-r-03-2024": {
3995
+ readonly "mistralai/mixtral-8x22b-instruct": {
3333
3996
  id: string;
3334
3997
  name: string;
3335
3998
  api: "openai-completions";
@@ -3363,7 +4026,7 @@ export declare const MODELS: {
3363
4026
  contextWindow: number;
3364
4027
  maxTokens: number;
3365
4028
  };
3366
- readonly "mistralai/mistral-tiny": {
4029
+ readonly "mistralai/mistral-small": {
3367
4030
  id: string;
3368
4031
  name: string;
3369
4032
  api: "openai-completions";
@@ -3380,7 +4043,7 @@ export declare const MODELS: {
3380
4043
  contextWindow: number;
3381
4044
  maxTokens: number;
3382
4045
  };
3383
- readonly "mistralai/mistral-small": {
4046
+ readonly "mistralai/mistral-tiny": {
3384
4047
  id: string;
3385
4048
  name: string;
3386
4049
  api: "openai-completions";