@mariozechner/pi-ai 0.5.43 → 0.5.44

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1,12 +1,12 @@
1
1
  export declare const MODELS: {
2
2
  readonly anthropic: {
3
- readonly "claude-3-7-sonnet-20250219": {
3
+ readonly "claude-3-5-sonnet-20241022": {
4
4
  id: string;
5
5
  name: string;
6
6
  api: "anthropic-messages";
7
7
  provider: string;
8
8
  baseUrl: string;
9
- reasoning: true;
9
+ reasoning: false;
10
10
  input: ("text" | "image")[];
11
11
  cost: {
12
12
  input: number;
@@ -17,13 +17,13 @@ export declare const MODELS: {
17
17
  contextWindow: number;
18
18
  maxTokens: number;
19
19
  };
20
- readonly "claude-opus-4-1-20250805": {
20
+ readonly "claude-3-5-sonnet-20240620": {
21
21
  id: string;
22
22
  name: string;
23
23
  api: "anthropic-messages";
24
24
  provider: string;
25
25
  baseUrl: string;
26
- reasoning: true;
26
+ reasoning: false;
27
27
  input: ("text" | "image")[];
28
28
  cost: {
29
29
  input: number;
@@ -34,7 +34,7 @@ export declare const MODELS: {
34
34
  contextWindow: number;
35
35
  maxTokens: number;
36
36
  };
37
- readonly "claude-3-haiku-20240307": {
37
+ readonly "claude-3-opus-20240229": {
38
38
  id: string;
39
39
  name: string;
40
40
  api: "anthropic-messages";
@@ -51,13 +51,30 @@ export declare const MODELS: {
51
51
  contextWindow: number;
52
52
  maxTokens: number;
53
53
  };
54
- readonly "claude-3-5-haiku-20241022": {
54
+ readonly "claude-sonnet-4-5-20250929": {
55
55
  id: string;
56
56
  name: string;
57
57
  api: "anthropic-messages";
58
58
  provider: string;
59
59
  baseUrl: string;
60
- reasoning: false;
60
+ reasoning: true;
61
+ input: ("text" | "image")[];
62
+ cost: {
63
+ input: number;
64
+ output: number;
65
+ cacheRead: number;
66
+ cacheWrite: number;
67
+ };
68
+ contextWindow: number;
69
+ maxTokens: number;
70
+ };
71
+ readonly "claude-sonnet-4-20250514": {
72
+ id: string;
73
+ name: string;
74
+ api: "anthropic-messages";
75
+ provider: string;
76
+ baseUrl: string;
77
+ reasoning: true;
61
78
  input: ("text" | "image")[];
62
79
  cost: {
63
80
  input: number;
@@ -85,7 +102,7 @@ export declare const MODELS: {
85
102
  contextWindow: number;
86
103
  maxTokens: number;
87
104
  };
88
- readonly "claude-3-5-sonnet-20241022": {
105
+ readonly "claude-3-5-haiku-20241022": {
89
106
  id: string;
90
107
  name: string;
91
108
  api: "anthropic-messages";
@@ -102,7 +119,7 @@ export declare const MODELS: {
102
119
  contextWindow: number;
103
120
  maxTokens: number;
104
121
  };
105
- readonly "claude-3-5-sonnet-20240620": {
122
+ readonly "claude-3-haiku-20240307": {
106
123
  id: string;
107
124
  name: string;
108
125
  api: "anthropic-messages";
@@ -119,13 +136,13 @@ export declare const MODELS: {
119
136
  contextWindow: number;
120
137
  maxTokens: number;
121
138
  };
122
- readonly "claude-3-sonnet-20240229": {
139
+ readonly "claude-3-7-sonnet-20250219": {
123
140
  id: string;
124
141
  name: string;
125
142
  api: "anthropic-messages";
126
143
  provider: string;
127
144
  baseUrl: string;
128
- reasoning: false;
145
+ reasoning: true;
129
146
  input: ("text" | "image")[];
130
147
  cost: {
131
148
  input: number;
@@ -136,7 +153,7 @@ export declare const MODELS: {
136
153
  contextWindow: number;
137
154
  maxTokens: number;
138
155
  };
139
- readonly "claude-sonnet-4-20250514": {
156
+ readonly "claude-opus-4-1-20250805": {
140
157
  id: string;
141
158
  name: string;
142
159
  api: "anthropic-messages";
@@ -153,7 +170,7 @@ export declare const MODELS: {
153
170
  contextWindow: number;
154
171
  maxTokens: number;
155
172
  };
156
- readonly "claude-3-opus-20240229": {
173
+ readonly "claude-3-sonnet-20240229": {
157
174
  id: string;
158
175
  name: string;
159
176
  api: "anthropic-messages";
@@ -189,7 +206,7 @@ export declare const MODELS: {
189
206
  contextWindow: number;
190
207
  maxTokens: number;
191
208
  };
192
- readonly "gemini-2.5-pro": {
209
+ readonly "gemini-flash-lite-latest": {
193
210
  id: string;
194
211
  name: string;
195
212
  api: "google-generative-ai";
@@ -206,13 +223,13 @@ export declare const MODELS: {
206
223
  contextWindow: number;
207
224
  maxTokens: number;
208
225
  };
209
- readonly "gemini-1.5-flash": {
226
+ readonly "gemini-2.5-flash": {
210
227
  id: string;
211
228
  name: string;
212
229
  api: "google-generative-ai";
213
230
  provider: string;
214
231
  baseUrl: string;
215
- reasoning: false;
232
+ reasoning: true;
216
233
  input: ("text" | "image")[];
217
234
  cost: {
218
235
  input: number;
@@ -223,13 +240,13 @@ export declare const MODELS: {
223
240
  contextWindow: number;
224
241
  maxTokens: number;
225
242
  };
226
- readonly "gemini-2.0-flash-lite": {
243
+ readonly "gemini-flash-latest": {
227
244
  id: string;
228
245
  name: string;
229
246
  api: "google-generative-ai";
230
247
  provider: string;
231
248
  baseUrl: string;
232
- reasoning: false;
249
+ reasoning: true;
233
250
  input: ("text" | "image")[];
234
251
  cost: {
235
252
  input: number;
@@ -240,7 +257,24 @@ export declare const MODELS: {
240
257
  contextWindow: number;
241
258
  maxTokens: number;
242
259
  };
243
- readonly "gemini-1.5-pro": {
260
+ readonly "gemini-2.5-pro-preview-05-06": {
261
+ id: string;
262
+ name: string;
263
+ api: "google-generative-ai";
264
+ provider: string;
265
+ baseUrl: string;
266
+ reasoning: true;
267
+ input: ("text" | "image")[];
268
+ cost: {
269
+ input: number;
270
+ output: number;
271
+ cacheRead: number;
272
+ cacheWrite: number;
273
+ };
274
+ contextWindow: number;
275
+ maxTokens: number;
276
+ };
277
+ readonly "gemini-2.0-flash-lite": {
244
278
  id: string;
245
279
  name: string;
246
280
  api: "google-generative-ai";
@@ -257,7 +291,24 @@ export declare const MODELS: {
257
291
  contextWindow: number;
258
292
  maxTokens: number;
259
293
  };
260
- readonly "gemini-1.5-flash-8b": {
294
+ readonly "gemini-live-2.5-flash-preview-native-audio": {
295
+ id: string;
296
+ name: string;
297
+ api: "google-generative-ai";
298
+ provider: string;
299
+ baseUrl: string;
300
+ reasoning: true;
301
+ input: "text"[];
302
+ cost: {
303
+ input: number;
304
+ output: number;
305
+ cacheRead: number;
306
+ cacheWrite: number;
307
+ };
308
+ contextWindow: number;
309
+ maxTokens: number;
310
+ };
311
+ readonly "gemini-2.0-flash": {
261
312
  id: string;
262
313
  name: string;
263
314
  api: "google-generative-ai";
@@ -274,7 +325,7 @@ export declare const MODELS: {
274
325
  contextWindow: number;
275
326
  maxTokens: number;
276
327
  };
277
- readonly "gemini-2.5-flash": {
328
+ readonly "gemini-2.5-flash-lite": {
278
329
  id: string;
279
330
  name: string;
280
331
  api: "google-generative-ai";
@@ -308,7 +359,7 @@ export declare const MODELS: {
308
359
  contextWindow: number;
309
360
  maxTokens: number;
310
361
  };
311
- readonly "gemini-2.5-pro-preview-05-06": {
362
+ readonly "gemini-2.5-flash-lite-preview-06-17": {
312
363
  id: string;
313
364
  name: string;
314
365
  api: "google-generative-ai";
@@ -325,13 +376,13 @@ export declare const MODELS: {
325
376
  contextWindow: number;
326
377
  maxTokens: number;
327
378
  };
328
- readonly "gemini-2.0-flash": {
379
+ readonly "gemini-2.5-flash-preview-09-2025": {
329
380
  id: string;
330
381
  name: string;
331
382
  api: "google-generative-ai";
332
383
  provider: string;
333
384
  baseUrl: string;
334
- reasoning: false;
385
+ reasoning: true;
335
386
  input: ("text" | "image")[];
336
387
  cost: {
337
388
  input: number;
@@ -342,7 +393,7 @@ export declare const MODELS: {
342
393
  contextWindow: number;
343
394
  maxTokens: number;
344
395
  };
345
- readonly "gemini-2.5-flash-lite-preview-06-17": {
396
+ readonly "gemini-2.5-flash-preview-04-17": {
346
397
  id: string;
347
398
  name: string;
348
399
  api: "google-generative-ai";
@@ -359,7 +410,7 @@ export declare const MODELS: {
359
410
  contextWindow: number;
360
411
  maxTokens: number;
361
412
  };
362
- readonly "gemini-2.5-flash-preview-04-17": {
413
+ readonly "gemini-2.5-pro": {
363
414
  id: string;
364
415
  name: string;
365
416
  api: "google-generative-ai";
@@ -376,15 +427,83 @@ export declare const MODELS: {
376
427
  contextWindow: number;
377
428
  maxTokens: number;
378
429
  };
430
+ readonly "gemini-1.5-flash": {
431
+ id: string;
432
+ name: string;
433
+ api: "google-generative-ai";
434
+ provider: string;
435
+ baseUrl: string;
436
+ reasoning: false;
437
+ input: ("text" | "image")[];
438
+ cost: {
439
+ input: number;
440
+ output: number;
441
+ cacheRead: number;
442
+ cacheWrite: number;
443
+ };
444
+ contextWindow: number;
445
+ maxTokens: number;
446
+ };
447
+ readonly "gemini-1.5-flash-8b": {
448
+ id: string;
449
+ name: string;
450
+ api: "google-generative-ai";
451
+ provider: string;
452
+ baseUrl: string;
453
+ reasoning: false;
454
+ input: ("text" | "image")[];
455
+ cost: {
456
+ input: number;
457
+ output: number;
458
+ cacheRead: number;
459
+ cacheWrite: number;
460
+ };
461
+ contextWindow: number;
462
+ maxTokens: number;
463
+ };
464
+ readonly "gemini-2.5-flash-lite-preview-09-2025": {
465
+ id: string;
466
+ name: string;
467
+ api: "google-generative-ai";
468
+ provider: string;
469
+ baseUrl: string;
470
+ reasoning: true;
471
+ input: ("text" | "image")[];
472
+ cost: {
473
+ input: number;
474
+ output: number;
475
+ cacheRead: number;
476
+ cacheWrite: number;
477
+ };
478
+ contextWindow: number;
479
+ maxTokens: number;
480
+ };
481
+ readonly "gemini-1.5-pro": {
482
+ id: string;
483
+ name: string;
484
+ api: "google-generative-ai";
485
+ provider: string;
486
+ baseUrl: string;
487
+ reasoning: false;
488
+ input: ("text" | "image")[];
489
+ cost: {
490
+ input: number;
491
+ output: number;
492
+ cacheRead: number;
493
+ cacheWrite: number;
494
+ };
495
+ contextWindow: number;
496
+ maxTokens: number;
497
+ };
379
498
  };
380
499
  readonly openai: {
381
- readonly "gpt-5-nano": {
500
+ readonly "gpt-4.1-nano": {
382
501
  id: string;
383
502
  name: string;
384
503
  api: "openai-responses";
385
504
  provider: string;
386
505
  baseUrl: string;
387
- reasoning: true;
506
+ reasoning: false;
388
507
  input: ("text" | "image")[];
389
508
  cost: {
390
509
  input: number;
@@ -395,14 +514,14 @@ export declare const MODELS: {
395
514
  contextWindow: number;
396
515
  maxTokens: number;
397
516
  };
398
- readonly "o3-pro": {
517
+ readonly "gpt-4": {
399
518
  id: string;
400
519
  name: string;
401
520
  api: "openai-responses";
402
521
  provider: string;
403
522
  baseUrl: string;
404
- reasoning: true;
405
- input: ("text" | "image")[];
523
+ reasoning: false;
524
+ input: "text"[];
406
525
  cost: {
407
526
  input: number;
408
527
  output: number;
@@ -412,14 +531,14 @@ export declare const MODELS: {
412
531
  contextWindow: number;
413
532
  maxTokens: number;
414
533
  };
415
- readonly "codex-mini-latest": {
534
+ readonly "o1-pro": {
416
535
  id: string;
417
536
  name: string;
418
537
  api: "openai-responses";
419
538
  provider: string;
420
539
  baseUrl: string;
421
540
  reasoning: true;
422
- input: "text"[];
541
+ input: ("text" | "image")[];
423
542
  cost: {
424
543
  input: number;
425
544
  output: number;
@@ -429,7 +548,7 @@ export declare const MODELS: {
429
548
  contextWindow: number;
430
549
  maxTokens: number;
431
550
  };
432
- readonly "gpt-4.1": {
551
+ readonly "gpt-4o-2024-05-13": {
433
552
  id: string;
434
553
  name: string;
435
554
  api: "openai-responses";
@@ -446,7 +565,7 @@ export declare const MODELS: {
446
565
  contextWindow: number;
447
566
  maxTokens: number;
448
567
  };
449
- readonly "gpt-4-turbo": {
568
+ readonly "gpt-4o-2024-08-06": {
450
569
  id: string;
451
570
  name: string;
452
571
  api: "openai-responses";
@@ -463,13 +582,13 @@ export declare const MODELS: {
463
582
  contextWindow: number;
464
583
  maxTokens: number;
465
584
  };
466
- readonly o1: {
585
+ readonly "gpt-4.1-mini": {
467
586
  id: string;
468
587
  name: string;
469
588
  api: "openai-responses";
470
589
  provider: string;
471
590
  baseUrl: string;
472
- reasoning: true;
591
+ reasoning: false;
473
592
  input: ("text" | "image")[];
474
593
  cost: {
475
594
  input: number;
@@ -497,13 +616,13 @@ export declare const MODELS: {
497
616
  contextWindow: number;
498
617
  maxTokens: number;
499
618
  };
500
- readonly "gpt-5": {
619
+ readonly "gpt-4-turbo": {
501
620
  id: string;
502
621
  name: string;
503
622
  api: "openai-responses";
504
623
  provider: string;
505
624
  baseUrl: string;
506
- reasoning: true;
625
+ reasoning: false;
507
626
  input: ("text" | "image")[];
508
627
  cost: {
509
628
  input: number;
@@ -514,14 +633,14 @@ export declare const MODELS: {
514
633
  contextWindow: number;
515
634
  maxTokens: number;
516
635
  };
517
- readonly "o1-pro": {
636
+ readonly "o3-mini": {
518
637
  id: string;
519
638
  name: string;
520
639
  api: "openai-responses";
521
640
  provider: string;
522
641
  baseUrl: string;
523
642
  reasoning: true;
524
- input: ("text" | "image")[];
643
+ input: "text"[];
525
644
  cost: {
526
645
  input: number;
527
646
  output: number;
@@ -531,14 +650,14 @@ export declare const MODELS: {
531
650
  contextWindow: number;
532
651
  maxTokens: number;
533
652
  };
534
- readonly o3: {
653
+ readonly "codex-mini-latest": {
535
654
  id: string;
536
655
  name: string;
537
656
  api: "openai-responses";
538
657
  provider: string;
539
658
  baseUrl: string;
540
659
  reasoning: true;
541
- input: ("text" | "image")[];
660
+ input: "text"[];
542
661
  cost: {
543
662
  input: number;
544
663
  output: number;
@@ -548,7 +667,7 @@ export declare const MODELS: {
548
667
  contextWindow: number;
549
668
  maxTokens: number;
550
669
  };
551
- readonly "gpt-5-mini": {
670
+ readonly "gpt-5-nano": {
552
671
  id: string;
553
672
  name: string;
554
673
  api: "openai-responses";
@@ -565,7 +684,7 @@ export declare const MODELS: {
565
684
  contextWindow: number;
566
685
  maxTokens: number;
567
686
  };
568
- readonly "o4-mini-deep-research": {
687
+ readonly "gpt-5-codex": {
569
688
  id: string;
570
689
  name: string;
571
690
  api: "openai-responses";
@@ -582,7 +701,7 @@ export declare const MODELS: {
582
701
  contextWindow: number;
583
702
  maxTokens: number;
584
703
  };
585
- readonly "gpt-4o-mini": {
704
+ readonly "gpt-4o": {
586
705
  id: string;
587
706
  name: string;
588
707
  api: "openai-responses";
@@ -599,7 +718,7 @@ export declare const MODELS: {
599
718
  contextWindow: number;
600
719
  maxTokens: number;
601
720
  };
602
- readonly "gpt-4.1-nano": {
721
+ readonly "gpt-4.1": {
603
722
  id: string;
604
723
  name: string;
605
724
  api: "openai-responses";
@@ -616,13 +735,13 @@ export declare const MODELS: {
616
735
  contextWindow: number;
617
736
  maxTokens: number;
618
737
  };
619
- readonly "gpt-4.1-mini": {
738
+ readonly "o4-mini": {
620
739
  id: string;
621
740
  name: string;
622
741
  api: "openai-responses";
623
742
  provider: string;
624
743
  baseUrl: string;
625
- reasoning: false;
744
+ reasoning: true;
626
745
  input: ("text" | "image")[];
627
746
  cost: {
628
747
  input: number;
@@ -633,13 +752,13 @@ export declare const MODELS: {
633
752
  contextWindow: number;
634
753
  maxTokens: number;
635
754
  };
636
- readonly "gpt-4o": {
755
+ readonly o1: {
637
756
  id: string;
638
757
  name: string;
639
758
  api: "openai-responses";
640
759
  provider: string;
641
760
  baseUrl: string;
642
- reasoning: false;
761
+ reasoning: true;
643
762
  input: ("text" | "image")[];
644
763
  cost: {
645
764
  input: number;
@@ -650,14 +769,48 @@ export declare const MODELS: {
650
769
  contextWindow: number;
651
770
  maxTokens: number;
652
771
  };
653
- readonly "gpt-4": {
772
+ readonly "gpt-5-mini": {
773
+ id: string;
774
+ name: string;
775
+ api: "openai-responses";
776
+ provider: string;
777
+ baseUrl: string;
778
+ reasoning: true;
779
+ input: ("text" | "image")[];
780
+ cost: {
781
+ input: number;
782
+ output: number;
783
+ cacheRead: number;
784
+ cacheWrite: number;
785
+ };
786
+ contextWindow: number;
787
+ maxTokens: number;
788
+ };
789
+ readonly "o3-pro": {
790
+ id: string;
791
+ name: string;
792
+ api: "openai-responses";
793
+ provider: string;
794
+ baseUrl: string;
795
+ reasoning: true;
796
+ input: ("text" | "image")[];
797
+ cost: {
798
+ input: number;
799
+ output: number;
800
+ cacheRead: number;
801
+ cacheWrite: number;
802
+ };
803
+ contextWindow: number;
804
+ maxTokens: number;
805
+ };
806
+ readonly "gpt-4o-2024-11-20": {
654
807
  id: string;
655
808
  name: string;
656
809
  api: "openai-responses";
657
810
  provider: string;
658
811
  baseUrl: string;
659
812
  reasoning: false;
660
- input: "text"[];
813
+ input: ("text" | "image")[];
661
814
  cost: {
662
815
  input: number;
663
816
  output: number;
@@ -667,7 +820,7 @@ export declare const MODELS: {
667
820
  contextWindow: number;
668
821
  maxTokens: number;
669
822
  };
670
- readonly "o4-mini": {
823
+ readonly o3: {
671
824
  id: string;
672
825
  name: string;
673
826
  api: "openai-responses";
@@ -684,14 +837,48 @@ export declare const MODELS: {
684
837
  contextWindow: number;
685
838
  maxTokens: number;
686
839
  };
687
- readonly "o3-mini": {
840
+ readonly "o4-mini-deep-research": {
688
841
  id: string;
689
842
  name: string;
690
843
  api: "openai-responses";
691
844
  provider: string;
692
845
  baseUrl: string;
693
846
  reasoning: true;
694
- input: "text"[];
847
+ input: ("text" | "image")[];
848
+ cost: {
849
+ input: number;
850
+ output: number;
851
+ cacheRead: number;
852
+ cacheWrite: number;
853
+ };
854
+ contextWindow: number;
855
+ maxTokens: number;
856
+ };
857
+ readonly "gpt-4o-mini": {
858
+ id: string;
859
+ name: string;
860
+ api: "openai-responses";
861
+ provider: string;
862
+ baseUrl: string;
863
+ reasoning: false;
864
+ input: ("text" | "image")[];
865
+ cost: {
866
+ input: number;
867
+ output: number;
868
+ cacheRead: number;
869
+ cacheWrite: number;
870
+ };
871
+ contextWindow: number;
872
+ maxTokens: number;
873
+ };
874
+ readonly "gpt-5": {
875
+ id: string;
876
+ name: string;
877
+ api: "openai-responses";
878
+ provider: string;
879
+ baseUrl: string;
880
+ reasoning: true;
881
+ input: ("text" | "image")[];
695
882
  cost: {
696
883
  input: number;
697
884
  output: number;
@@ -737,6 +924,40 @@ export declare const MODELS: {
737
924
  contextWindow: number;
738
925
  maxTokens: number;
739
926
  };
927
+ readonly "mistral-saba-24b": {
928
+ id: string;
929
+ name: string;
930
+ api: "openai-completions";
931
+ provider: string;
932
+ baseUrl: string;
933
+ reasoning: false;
934
+ input: "text"[];
935
+ cost: {
936
+ input: number;
937
+ output: number;
938
+ cacheRead: number;
939
+ cacheWrite: number;
940
+ };
941
+ contextWindow: number;
942
+ maxTokens: number;
943
+ };
944
+ readonly "llama3-8b-8192": {
945
+ id: string;
946
+ name: string;
947
+ api: "openai-completions";
948
+ provider: string;
949
+ baseUrl: string;
950
+ reasoning: false;
951
+ input: "text"[];
952
+ cost: {
953
+ input: number;
954
+ output: number;
955
+ cacheRead: number;
956
+ cacheWrite: number;
957
+ };
958
+ contextWindow: number;
959
+ maxTokens: number;
960
+ };
740
961
  readonly "qwen-qwq-32b": {
741
962
  id: string;
742
963
  name: string;
@@ -788,7 +1009,7 @@ export declare const MODELS: {
788
1009
  contextWindow: number;
789
1010
  maxTokens: number;
790
1011
  };
791
- readonly "llama3-8b-8192": {
1012
+ readonly "gemma2-9b-it": {
792
1013
  id: string;
793
1014
  name: string;
794
1015
  api: "openai-completions";
@@ -805,7 +1026,7 @@ export declare const MODELS: {
805
1026
  contextWindow: number;
806
1027
  maxTokens: number;
807
1028
  };
808
- readonly "gemma2-9b-it": {
1029
+ readonly "llama-3.3-70b-versatile": {
809
1030
  id: string;
810
1031
  name: string;
811
1032
  api: "openai-completions";
@@ -822,7 +1043,7 @@ export declare const MODELS: {
822
1043
  contextWindow: number;
823
1044
  maxTokens: number;
824
1045
  };
825
- readonly "llama-3.3-70b-versatile": {
1046
+ readonly "moonshotai/kimi-k2-instruct-0905": {
826
1047
  id: string;
827
1048
  name: string;
828
1049
  api: "openai-completions";
@@ -839,7 +1060,7 @@ export declare const MODELS: {
839
1060
  contextWindow: number;
840
1061
  maxTokens: number;
841
1062
  };
842
- readonly "mistral-saba-24b": {
1063
+ readonly "moonshotai/kimi-k2-instruct": {
843
1064
  id: string;
844
1065
  name: string;
845
1066
  api: "openai-completions";
@@ -890,7 +1111,24 @@ export declare const MODELS: {
890
1111
  contextWindow: number;
891
1112
  maxTokens: number;
892
1113
  };
893
- readonly "meta-llama/llama-4-maverick-17b-128e-instruct": {
1114
+ readonly "qwen/qwen3-32b": {
1115
+ id: string;
1116
+ name: string;
1117
+ api: "openai-completions";
1118
+ provider: string;
1119
+ baseUrl: string;
1120
+ reasoning: true;
1121
+ input: "text"[];
1122
+ cost: {
1123
+ input: number;
1124
+ output: number;
1125
+ cacheRead: number;
1126
+ cacheWrite: number;
1127
+ };
1128
+ contextWindow: number;
1129
+ maxTokens: number;
1130
+ };
1131
+ readonly "meta-llama/llama-4-scout-17b-16e-instruct": {
894
1132
  id: string;
895
1133
  name: string;
896
1134
  api: "openai-completions";
@@ -907,7 +1145,7 @@ export declare const MODELS: {
907
1145
  contextWindow: number;
908
1146
  maxTokens: number;
909
1147
  };
910
- readonly "meta-llama/llama-4-scout-17b-16e-instruct": {
1148
+ readonly "meta-llama/llama-4-maverick-17b-128e-instruct": {
911
1149
  id: string;
912
1150
  name: string;
913
1151
  api: "openai-completions";
@@ -924,7 +1162,43 @@ export declare const MODELS: {
924
1162
  contextWindow: number;
925
1163
  maxTokens: number;
926
1164
  };
927
- readonly "qwen/qwen3-32b": {
1165
+ };
1166
+ readonly cerebras: {
1167
+ readonly "qwen-3-235b-a22b-instruct-2507": {
1168
+ id: string;
1169
+ name: string;
1170
+ api: "openai-completions";
1171
+ provider: string;
1172
+ baseUrl: string;
1173
+ reasoning: false;
1174
+ input: "text"[];
1175
+ cost: {
1176
+ input: number;
1177
+ output: number;
1178
+ cacheRead: number;
1179
+ cacheWrite: number;
1180
+ };
1181
+ contextWindow: number;
1182
+ maxTokens: number;
1183
+ };
1184
+ readonly "qwen-3-coder-480b": {
1185
+ id: string;
1186
+ name: string;
1187
+ api: "openai-completions";
1188
+ provider: string;
1189
+ baseUrl: string;
1190
+ reasoning: false;
1191
+ input: "text"[];
1192
+ cost: {
1193
+ input: number;
1194
+ output: number;
1195
+ cacheRead: number;
1196
+ cacheWrite: number;
1197
+ };
1198
+ contextWindow: number;
1199
+ maxTokens: number;
1200
+ };
1201
+ readonly "gpt-oss-120b": {
928
1202
  id: string;
929
1203
  name: string;
930
1204
  api: "openai-completions";
@@ -941,7 +1215,26 @@ export declare const MODELS: {
941
1215
  contextWindow: number;
942
1216
  maxTokens: number;
943
1217
  };
944
- readonly "moonshotai/kimi-k2-instruct-0905": {
1218
+ };
1219
+ readonly xai: {
1220
+ readonly "grok-4-fast-non-reasoning": {
1221
+ id: string;
1222
+ name: string;
1223
+ api: "openai-completions";
1224
+ provider: string;
1225
+ baseUrl: string;
1226
+ reasoning: false;
1227
+ input: ("text" | "image")[];
1228
+ cost: {
1229
+ input: number;
1230
+ output: number;
1231
+ cacheRead: number;
1232
+ cacheWrite: number;
1233
+ };
1234
+ contextWindow: number;
1235
+ maxTokens: number;
1236
+ };
1237
+ readonly "grok-3-fast": {
945
1238
  id: string;
946
1239
  name: string;
947
1240
  api: "openai-completions";
@@ -958,13 +1251,13 @@ export declare const MODELS: {
958
1251
  contextWindow: number;
959
1252
  maxTokens: number;
960
1253
  };
961
- readonly "moonshotai/kimi-k2-instruct": {
1254
+ readonly "grok-4": {
962
1255
  id: string;
963
1256
  name: string;
964
1257
  api: "openai-completions";
965
1258
  provider: string;
966
1259
  baseUrl: string;
967
- reasoning: false;
1260
+ reasoning: true;
968
1261
  input: "text"[];
969
1262
  cost: {
970
1263
  input: number;
@@ -975,16 +1268,14 @@ export declare const MODELS: {
975
1268
  contextWindow: number;
976
1269
  maxTokens: number;
977
1270
  };
978
- };
979
- readonly cerebras: {
980
- readonly "qwen-3-235b-a22b-instruct-2507": {
1271
+ readonly "grok-2-vision": {
981
1272
  id: string;
982
1273
  name: string;
983
1274
  api: "openai-completions";
984
1275
  provider: string;
985
1276
  baseUrl: string;
986
1277
  reasoning: false;
987
- input: "text"[];
1278
+ input: ("text" | "image")[];
988
1279
  cost: {
989
1280
  input: number;
990
1281
  output: number;
@@ -994,7 +1285,7 @@ export declare const MODELS: {
994
1285
  contextWindow: number;
995
1286
  maxTokens: number;
996
1287
  };
997
- readonly "gpt-oss-120b": {
1288
+ readonly "grok-code-fast-1": {
998
1289
  id: string;
999
1290
  name: string;
1000
1291
  api: "openai-completions";
@@ -1011,7 +1302,7 @@ export declare const MODELS: {
1011
1302
  contextWindow: number;
1012
1303
  maxTokens: number;
1013
1304
  };
1014
- readonly "qwen-3-coder-480b": {
1305
+ readonly "grok-2": {
1015
1306
  id: string;
1016
1307
  name: string;
1017
1308
  api: "openai-completions";
@@ -1028,8 +1319,6 @@ export declare const MODELS: {
1028
1319
  contextWindow: number;
1029
1320
  maxTokens: number;
1030
1321
  };
1031
- };
1032
- readonly xai: {
1033
1322
  readonly "grok-3-mini-fast-latest": {
1034
1323
  id: string;
1035
1324
  name: string;
@@ -1047,14 +1336,14 @@ export declare const MODELS: {
1047
1336
  contextWindow: number;
1048
1337
  maxTokens: number;
1049
1338
  };
1050
- readonly "grok-3-mini-latest": {
1339
+ readonly "grok-2-vision-1212": {
1051
1340
  id: string;
1052
1341
  name: string;
1053
1342
  api: "openai-completions";
1054
1343
  provider: string;
1055
1344
  baseUrl: string;
1056
- reasoning: true;
1057
- input: "text"[];
1345
+ reasoning: false;
1346
+ input: ("text" | "image")[];
1058
1347
  cost: {
1059
1348
  input: number;
1060
1349
  output: number;
@@ -1064,7 +1353,7 @@ export declare const MODELS: {
1064
1353
  contextWindow: number;
1065
1354
  maxTokens: number;
1066
1355
  };
1067
- readonly "grok-beta": {
1356
+ readonly "grok-3": {
1068
1357
  id: string;
1069
1358
  name: string;
1070
1359
  api: "openai-completions";
@@ -1081,14 +1370,14 @@ export declare const MODELS: {
1081
1370
  contextWindow: number;
1082
1371
  maxTokens: number;
1083
1372
  };
1084
- readonly "grok-3-fast-latest": {
1373
+ readonly "grok-4-fast": {
1085
1374
  id: string;
1086
1375
  name: string;
1087
1376
  api: "openai-completions";
1088
1377
  provider: string;
1089
1378
  baseUrl: string;
1090
- reasoning: false;
1091
- input: "text"[];
1379
+ reasoning: true;
1380
+ input: ("text" | "image")[];
1092
1381
  cost: {
1093
1382
  input: number;
1094
1383
  output: number;
@@ -1098,7 +1387,7 @@ export declare const MODELS: {
1098
1387
  contextWindow: number;
1099
1388
  maxTokens: number;
1100
1389
  };
1101
- readonly "grok-3": {
1390
+ readonly "grok-2-latest": {
1102
1391
  id: string;
1103
1392
  name: string;
1104
1393
  api: "openai-completions";
@@ -1115,13 +1404,13 @@ export declare const MODELS: {
1115
1404
  contextWindow: number;
1116
1405
  maxTokens: number;
1117
1406
  };
1118
- readonly "grok-3-mini": {
1407
+ readonly "grok-2-1212": {
1119
1408
  id: string;
1120
1409
  name: string;
1121
1410
  api: "openai-completions";
1122
1411
  provider: string;
1123
1412
  baseUrl: string;
1124
- reasoning: true;
1413
+ reasoning: false;
1125
1414
  input: "text"[];
1126
1415
  cost: {
1127
1416
  input: number;
@@ -1132,14 +1421,14 @@ export declare const MODELS: {
1132
1421
  contextWindow: number;
1133
1422
  maxTokens: number;
1134
1423
  };
1135
- readonly "grok-2-vision-1212": {
1424
+ readonly "grok-3-fast-latest": {
1136
1425
  id: string;
1137
1426
  name: string;
1138
1427
  api: "openai-completions";
1139
1428
  provider: string;
1140
1429
  baseUrl: string;
1141
1430
  reasoning: false;
1142
- input: ("text" | "image")[];
1431
+ input: "text"[];
1143
1432
  cost: {
1144
1433
  input: number;
1145
1434
  output: number;
@@ -1149,7 +1438,7 @@ export declare const MODELS: {
1149
1438
  contextWindow: number;
1150
1439
  maxTokens: number;
1151
1440
  };
1152
- readonly "grok-2": {
1441
+ readonly "grok-3-latest": {
1153
1442
  id: string;
1154
1443
  name: string;
1155
1444
  api: "openai-completions";
@@ -1183,14 +1472,14 @@ export declare const MODELS: {
1183
1472
  contextWindow: number;
1184
1473
  maxTokens: number;
1185
1474
  };
1186
- readonly "grok-3-latest": {
1475
+ readonly "grok-vision-beta": {
1187
1476
  id: string;
1188
1477
  name: string;
1189
1478
  api: "openai-completions";
1190
1479
  provider: string;
1191
1480
  baseUrl: string;
1192
1481
  reasoning: false;
1193
- input: "text"[];
1482
+ input: ("text" | "image")[];
1194
1483
  cost: {
1195
1484
  input: number;
1196
1485
  output: number;
@@ -1200,14 +1489,14 @@ export declare const MODELS: {
1200
1489
  contextWindow: number;
1201
1490
  maxTokens: number;
1202
1491
  };
1203
- readonly "grok-2-vision": {
1492
+ readonly "grok-3-mini": {
1204
1493
  id: string;
1205
1494
  name: string;
1206
1495
  api: "openai-completions";
1207
1496
  provider: string;
1208
1497
  baseUrl: string;
1209
- reasoning: false;
1210
- input: ("text" | "image")[];
1498
+ reasoning: true;
1499
+ input: "text"[];
1211
1500
  cost: {
1212
1501
  input: number;
1213
1502
  output: number;
@@ -1217,7 +1506,7 @@ export declare const MODELS: {
1217
1506
  contextWindow: number;
1218
1507
  maxTokens: number;
1219
1508
  };
1220
- readonly "grok-2-latest": {
1509
+ readonly "grok-beta": {
1221
1510
  id: string;
1222
1511
  name: string;
1223
1512
  api: "openai-completions";
@@ -1234,13 +1523,13 @@ export declare const MODELS: {
1234
1523
  contextWindow: number;
1235
1524
  maxTokens: number;
1236
1525
  };
1237
- readonly "grok-3-fast": {
1526
+ readonly "grok-3-mini-latest": {
1238
1527
  id: string;
1239
1528
  name: string;
1240
1529
  api: "openai-completions";
1241
1530
  provider: string;
1242
1531
  baseUrl: string;
1243
- reasoning: false;
1532
+ reasoning: true;
1244
1533
  input: "text"[];
1245
1534
  cost: {
1246
1535
  input: number;
@@ -1251,13 +1540,13 @@ export declare const MODELS: {
1251
1540
  contextWindow: number;
1252
1541
  maxTokens: number;
1253
1542
  };
1254
- readonly "grok-2-1212": {
1543
+ readonly "grok-3-mini-fast": {
1255
1544
  id: string;
1256
1545
  name: string;
1257
1546
  api: "openai-completions";
1258
1547
  provider: string;
1259
1548
  baseUrl: string;
1260
- reasoning: false;
1549
+ reasoning: true;
1261
1550
  input: "text"[];
1262
1551
  cost: {
1263
1552
  input: number;
@@ -1268,10 +1557,12 @@ export declare const MODELS: {
1268
1557
  contextWindow: number;
1269
1558
  maxTokens: number;
1270
1559
  };
1271
- readonly "grok-4": {
1560
+ };
1561
+ readonly zai: {
1562
+ readonly "glm-4.5-flash": {
1272
1563
  id: string;
1273
1564
  name: string;
1274
- api: "openai-completions";
1565
+ api: "anthropic-messages";
1275
1566
  provider: string;
1276
1567
  baseUrl: string;
1277
1568
  reasoning: true;
@@ -1285,14 +1576,14 @@ export declare const MODELS: {
1285
1576
  contextWindow: number;
1286
1577
  maxTokens: number;
1287
1578
  };
1288
- readonly "grok-vision-beta": {
1579
+ readonly "glm-4.5": {
1289
1580
  id: string;
1290
1581
  name: string;
1291
- api: "openai-completions";
1582
+ api: "anthropic-messages";
1292
1583
  provider: string;
1293
1584
  baseUrl: string;
1294
- reasoning: false;
1295
- input: ("text" | "image")[];
1585
+ reasoning: true;
1586
+ input: "text"[];
1296
1587
  cost: {
1297
1588
  input: number;
1298
1589
  output: number;
@@ -1302,10 +1593,10 @@ export declare const MODELS: {
1302
1593
  contextWindow: number;
1303
1594
  maxTokens: number;
1304
1595
  };
1305
- readonly "grok-3-mini-fast": {
1596
+ readonly "glm-4.5-air": {
1306
1597
  id: string;
1307
1598
  name: string;
1308
- api: "openai-completions";
1599
+ api: "anthropic-messages";
1309
1600
  provider: string;
1310
1601
  baseUrl: string;
1311
1602
  reasoning: true;
@@ -1319,14 +1610,14 @@ export declare const MODELS: {
1319
1610
  contextWindow: number;
1320
1611
  maxTokens: number;
1321
1612
  };
1322
- readonly "grok-code-fast-1": {
1613
+ readonly "glm-4.5v": {
1323
1614
  id: string;
1324
1615
  name: string;
1325
- api: "openai-completions";
1616
+ api: "anthropic-messages";
1326
1617
  provider: string;
1327
1618
  baseUrl: string;
1328
- reasoning: false;
1329
- input: "text"[];
1619
+ reasoning: true;
1620
+ input: ("text" | "image")[];
1330
1621
  cost: {
1331
1622
  input: number;
1332
1623
  output: number;
@@ -1336,9 +1627,7 @@ export declare const MODELS: {
1336
1627
  contextWindow: number;
1337
1628
  maxTokens: number;
1338
1629
  };
1339
- };
1340
- readonly zai: {
1341
- readonly "glm-4.5-air": {
1630
+ readonly "glm-4.6": {
1342
1631
  id: string;
1343
1632
  name: string;
1344
1633
  api: "anthropic-messages";
@@ -1355,14 +1644,16 @@ export declare const MODELS: {
1355
1644
  contextWindow: number;
1356
1645
  maxTokens: number;
1357
1646
  };
1358
- readonly "glm-4.5v": {
1647
+ };
1648
+ readonly openrouter: {
1649
+ readonly "z-ai/glm-4.6": {
1359
1650
  id: string;
1360
1651
  name: string;
1361
- api: "anthropic-messages";
1652
+ api: "openai-completions";
1362
1653
  provider: string;
1363
1654
  baseUrl: string;
1364
1655
  reasoning: true;
1365
- input: ("text" | "image")[];
1656
+ input: "text"[];
1366
1657
  cost: {
1367
1658
  input: number;
1368
1659
  output: number;
@@ -1372,10 +1663,10 @@ export declare const MODELS: {
1372
1663
  contextWindow: number;
1373
1664
  maxTokens: number;
1374
1665
  };
1375
- readonly "glm-4.5-flash": {
1666
+ readonly "deepseek/deepseek-v3.2-exp": {
1376
1667
  id: string;
1377
1668
  name: string;
1378
- api: "anthropic-messages";
1669
+ api: "openai-completions";
1379
1670
  provider: string;
1380
1671
  baseUrl: string;
1381
1672
  reasoning: true;
@@ -1389,14 +1680,14 @@ export declare const MODELS: {
1389
1680
  contextWindow: number;
1390
1681
  maxTokens: number;
1391
1682
  };
1392
- readonly "glm-4.5": {
1683
+ readonly "qwen/qwen3-vl-235b-a22b-thinking": {
1393
1684
  id: string;
1394
1685
  name: string;
1395
- api: "anthropic-messages";
1686
+ api: "openai-completions";
1396
1687
  provider: string;
1397
1688
  baseUrl: string;
1398
1689
  reasoning: true;
1399
- input: "text"[];
1690
+ input: ("text" | "image")[];
1400
1691
  cost: {
1401
1692
  input: number;
1402
1693
  output: number;
@@ -1406,16 +1697,14 @@ export declare const MODELS: {
1406
1697
  contextWindow: number;
1407
1698
  maxTokens: number;
1408
1699
  };
1409
- };
1410
- readonly openrouter: {
1411
- readonly "alibaba/tongyi-deepresearch-30b-a3b": {
1700
+ readonly "qwen/qwen3-vl-235b-a22b-instruct": {
1412
1701
  id: string;
1413
1702
  name: string;
1414
1703
  api: "openai-completions";
1415
1704
  provider: string;
1416
1705
  baseUrl: string;
1417
1706
  reasoning: true;
1418
- input: "text"[];
1707
+ input: ("text" | "image")[];
1419
1708
  cost: {
1420
1709
  input: number;
1421
1710
  output: number;
@@ -1425,7 +1714,7 @@ export declare const MODELS: {
1425
1714
  contextWindow: number;
1426
1715
  maxTokens: number;
1427
1716
  };
1428
- readonly "qwen/qwen3-coder-flash": {
1717
+ readonly "qwen/qwen3-max": {
1429
1718
  id: string;
1430
1719
  name: string;
1431
1720
  api: "openai-completions";
@@ -1459,7 +1748,7 @@ export declare const MODELS: {
1459
1748
  contextWindow: number;
1460
1749
  maxTokens: number;
1461
1750
  };
1462
- readonly "qwen/qwen3-next-80b-a3b-thinking": {
1751
+ readonly "deepseek/deepseek-v3.1-terminus": {
1463
1752
  id: string;
1464
1753
  name: string;
1465
1754
  api: "openai-completions";
@@ -1476,7 +1765,7 @@ export declare const MODELS: {
1476
1765
  contextWindow: number;
1477
1766
  maxTokens: number;
1478
1767
  };
1479
- readonly "qwen/qwen3-next-80b-a3b-instruct": {
1768
+ readonly "alibaba/tongyi-deepresearch-30b-a3b:free": {
1480
1769
  id: string;
1481
1770
  name: string;
1482
1771
  api: "openai-completions";
@@ -1493,13 +1782,13 @@ export declare const MODELS: {
1493
1782
  contextWindow: number;
1494
1783
  maxTokens: number;
1495
1784
  };
1496
- readonly "meituan/longcat-flash-chat": {
1785
+ readonly "alibaba/tongyi-deepresearch-30b-a3b": {
1497
1786
  id: string;
1498
1787
  name: string;
1499
1788
  api: "openai-completions";
1500
1789
  provider: string;
1501
1790
  baseUrl: string;
1502
- reasoning: false;
1791
+ reasoning: true;
1503
1792
  input: "text"[];
1504
1793
  cost: {
1505
1794
  input: number;
@@ -1510,7 +1799,7 @@ export declare const MODELS: {
1510
1799
  contextWindow: number;
1511
1800
  maxTokens: number;
1512
1801
  };
1513
- readonly "qwen/qwen-plus-2025-07-28": {
1802
+ readonly "qwen/qwen3-coder-flash": {
1514
1803
  id: string;
1515
1804
  name: string;
1516
1805
  api: "openai-completions";
@@ -1527,7 +1816,7 @@ export declare const MODELS: {
1527
1816
  contextWindow: number;
1528
1817
  maxTokens: number;
1529
1818
  };
1530
- readonly "qwen/qwen-plus-2025-07-28:thinking": {
1819
+ readonly "qwen/qwen3-next-80b-a3b-thinking": {
1531
1820
  id: string;
1532
1821
  name: string;
1533
1822
  api: "openai-completions";
@@ -1544,13 +1833,13 @@ export declare const MODELS: {
1544
1833
  contextWindow: number;
1545
1834
  maxTokens: number;
1546
1835
  };
1547
- readonly "nvidia/nemotron-nano-9b-v2:free": {
1836
+ readonly "qwen/qwen3-next-80b-a3b-instruct": {
1548
1837
  id: string;
1549
1838
  name: string;
1550
1839
  api: "openai-completions";
1551
1840
  provider: string;
1552
1841
  baseUrl: string;
1553
- reasoning: true;
1842
+ reasoning: false;
1554
1843
  input: "text"[];
1555
1844
  cost: {
1556
1845
  input: number;
@@ -1561,13 +1850,13 @@ export declare const MODELS: {
1561
1850
  contextWindow: number;
1562
1851
  maxTokens: number;
1563
1852
  };
1564
- readonly "nvidia/nemotron-nano-9b-v2": {
1853
+ readonly "meituan/longcat-flash-chat:free": {
1565
1854
  id: string;
1566
1855
  name: string;
1567
1856
  api: "openai-completions";
1568
1857
  provider: string;
1569
1858
  baseUrl: string;
1570
- reasoning: true;
1859
+ reasoning: false;
1571
1860
  input: "text"[];
1572
1861
  cost: {
1573
1862
  input: number;
@@ -1578,14 +1867,14 @@ export declare const MODELS: {
1578
1867
  contextWindow: number;
1579
1868
  maxTokens: number;
1580
1869
  };
1581
- readonly "openrouter/sonoma-dusk-alpha": {
1870
+ readonly "qwen/qwen-plus-2025-07-28": {
1582
1871
  id: string;
1583
1872
  name: string;
1584
1873
  api: "openai-completions";
1585
1874
  provider: string;
1586
1875
  baseUrl: string;
1587
1876
  reasoning: false;
1588
- input: ("text" | "image")[];
1877
+ input: "text"[];
1589
1878
  cost: {
1590
1879
  input: number;
1591
1880
  output: number;
@@ -1595,14 +1884,14 @@ export declare const MODELS: {
1595
1884
  contextWindow: number;
1596
1885
  maxTokens: number;
1597
1886
  };
1598
- readonly "openrouter/sonoma-sky-alpha": {
1887
+ readonly "qwen/qwen-plus-2025-07-28:thinking": {
1599
1888
  id: string;
1600
1889
  name: string;
1601
1890
  api: "openai-completions";
1602
1891
  provider: string;
1603
1892
  baseUrl: string;
1604
1893
  reasoning: true;
1605
- input: ("text" | "image")[];
1894
+ input: "text"[];
1606
1895
  cost: {
1607
1896
  input: number;
1608
1897
  output: number;
@@ -1612,13 +1901,30 @@ export declare const MODELS: {
1612
1901
  contextWindow: number;
1613
1902
  maxTokens: number;
1614
1903
  };
1615
- readonly "qwen/qwen3-max": {
1904
+ readonly "nvidia/nemotron-nano-9b-v2:free": {
1616
1905
  id: string;
1617
1906
  name: string;
1618
1907
  api: "openai-completions";
1619
1908
  provider: string;
1620
1909
  baseUrl: string;
1621
- reasoning: false;
1910
+ reasoning: true;
1911
+ input: "text"[];
1912
+ cost: {
1913
+ input: number;
1914
+ output: number;
1915
+ cacheRead: number;
1916
+ cacheWrite: number;
1917
+ };
1918
+ contextWindow: number;
1919
+ maxTokens: number;
1920
+ };
1921
+ readonly "nvidia/nemotron-nano-9b-v2": {
1922
+ id: string;
1923
+ name: string;
1924
+ api: "openai-completions";
1925
+ provider: string;
1926
+ baseUrl: string;
1927
+ reasoning: true;
1622
1928
  input: "text"[];
1623
1929
  cost: {
1624
1930
  input: number;
@@ -2020,23 +2326,6 @@ export declare const MODELS: {
2020
2326
  contextWindow: number;
2021
2327
  maxTokens: number;
2022
2328
  };
2023
- readonly "moonshotai/kimi-k2:free": {
2024
- id: string;
2025
- name: string;
2026
- api: "openai-completions";
2027
- provider: string;
2028
- baseUrl: string;
2029
- reasoning: false;
2030
- input: "text"[];
2031
- cost: {
2032
- input: number;
2033
- output: number;
2034
- cacheRead: number;
2035
- cacheWrite: number;
2036
- };
2037
- contextWindow: number;
2038
- maxTokens: number;
2039
- };
2040
2329
  readonly "moonshotai/kimi-k2": {
2041
2330
  id: string;
2042
2331
  name: string;
@@ -3006,7 +3295,7 @@ export declare const MODELS: {
3006
3295
  contextWindow: number;
3007
3296
  maxTokens: number;
3008
3297
  };
3009
- readonly "cohere/command-r-08-2024": {
3298
+ readonly "cohere/command-r-plus-08-2024": {
3010
3299
  id: string;
3011
3300
  name: string;
3012
3301
  api: "openai-completions";
@@ -3023,7 +3312,7 @@ export declare const MODELS: {
3023
3312
  contextWindow: number;
3024
3313
  maxTokens: number;
3025
3314
  };
3026
- readonly "cohere/command-r-plus-08-2024": {
3315
+ readonly "cohere/command-r-08-2024": {
3027
3316
  id: string;
3028
3317
  name: string;
3029
3318
  api: "openai-completions";
@@ -3074,7 +3363,7 @@ export declare const MODELS: {
3074
3363
  contextWindow: number;
3075
3364
  maxTokens: number;
3076
3365
  };
3077
- readonly "meta-llama/llama-3.1-8b-instruct": {
3366
+ readonly "meta-llama/llama-3.1-405b-instruct": {
3078
3367
  id: string;
3079
3368
  name: string;
3080
3369
  api: "openai-completions";
@@ -3091,7 +3380,7 @@ export declare const MODELS: {
3091
3380
  contextWindow: number;
3092
3381
  maxTokens: number;
3093
3382
  };
3094
- readonly "meta-llama/llama-3.1-405b-instruct": {
3383
+ readonly "meta-llama/llama-3.1-8b-instruct": {
3095
3384
  id: string;
3096
3385
  name: string;
3097
3386
  api: "openai-completions";
@@ -3142,7 +3431,7 @@ export declare const MODELS: {
3142
3431
  contextWindow: number;
3143
3432
  maxTokens: number;
3144
3433
  };
3145
- readonly "mistralai/mistral-7b-instruct-v0.3": {
3434
+ readonly "mistralai/mistral-7b-instruct:free": {
3146
3435
  id: string;
3147
3436
  name: string;
3148
3437
  api: "openai-completions";
@@ -3159,7 +3448,7 @@ export declare const MODELS: {
3159
3448
  contextWindow: number;
3160
3449
  maxTokens: number;
3161
3450
  };
3162
- readonly "mistralai/mistral-7b-instruct:free": {
3451
+ readonly "mistralai/mistral-7b-instruct": {
3163
3452
  id: string;
3164
3453
  name: string;
3165
3454
  api: "openai-completions";
@@ -3176,7 +3465,7 @@ export declare const MODELS: {
3176
3465
  contextWindow: number;
3177
3466
  maxTokens: number;
3178
3467
  };
3179
- readonly "mistralai/mistral-7b-instruct": {
3468
+ readonly "mistralai/mistral-7b-instruct-v0.3": {
3180
3469
  id: string;
3181
3470
  name: string;
3182
3471
  api: "openai-completions";
@@ -3278,74 +3567,6 @@ export declare const MODELS: {
3278
3567
  contextWindow: number;
3279
3568
  maxTokens: number;
3280
3569
  };
3281
- readonly "cohere/command-r-plus": {
3282
- id: string;
3283
- name: string;
3284
- api: "openai-completions";
3285
- provider: string;
3286
- baseUrl: string;
3287
- reasoning: false;
3288
- input: "text"[];
3289
- cost: {
3290
- input: number;
3291
- output: number;
3292
- cacheRead: number;
3293
- cacheWrite: number;
3294
- };
3295
- contextWindow: number;
3296
- maxTokens: number;
3297
- };
3298
- readonly "cohere/command-r-plus-04-2024": {
3299
- id: string;
3300
- name: string;
3301
- api: "openai-completions";
3302
- provider: string;
3303
- baseUrl: string;
3304
- reasoning: false;
3305
- input: "text"[];
3306
- cost: {
3307
- input: number;
3308
- output: number;
3309
- cacheRead: number;
3310
- cacheWrite: number;
3311
- };
3312
- contextWindow: number;
3313
- maxTokens: number;
3314
- };
3315
- readonly "cohere/command-r": {
3316
- id: string;
3317
- name: string;
3318
- api: "openai-completions";
3319
- provider: string;
3320
- baseUrl: string;
3321
- reasoning: false;
3322
- input: "text"[];
3323
- cost: {
3324
- input: number;
3325
- output: number;
3326
- cacheRead: number;
3327
- cacheWrite: number;
3328
- };
3329
- contextWindow: number;
3330
- maxTokens: number;
3331
- };
3332
- readonly "cohere/command-r-03-2024": {
3333
- id: string;
3334
- name: string;
3335
- api: "openai-completions";
3336
- provider: string;
3337
- baseUrl: string;
3338
- reasoning: false;
3339
- input: "text"[];
3340
- cost: {
3341
- input: number;
3342
- output: number;
3343
- cacheRead: number;
3344
- cacheWrite: number;
3345
- };
3346
- contextWindow: number;
3347
- maxTokens: number;
3348
- };
3349
3570
  readonly "mistralai/mistral-large": {
3350
3571
  id: string;
3351
3572
  name: string;
@@ -3363,7 +3584,7 @@ export declare const MODELS: {
3363
3584
  contextWindow: number;
3364
3585
  maxTokens: number;
3365
3586
  };
3366
- readonly "mistralai/mistral-tiny": {
3587
+ readonly "mistralai/mistral-small": {
3367
3588
  id: string;
3368
3589
  name: string;
3369
3590
  api: "openai-completions";
@@ -3380,7 +3601,7 @@ export declare const MODELS: {
3380
3601
  contextWindow: number;
3381
3602
  maxTokens: number;
3382
3603
  };
3383
- readonly "mistralai/mistral-small": {
3604
+ readonly "mistralai/mistral-tiny": {
3384
3605
  id: string;
3385
3606
  name: string;
3386
3607
  api: "openai-completions";