@mariozechner/pi-ai 0.57.1 → 0.58.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -85,14 +85,14 @@ export declare const MODELS: {
85
85
  contextWindow: number;
86
86
  maxTokens: number;
87
87
  };
88
- readonly "amazon.titan-text-express-v1": {
88
+ readonly "anthropic.claude-3-5-haiku-20241022-v1:0": {
89
89
  id: string;
90
90
  name: string;
91
91
  api: "bedrock-converse-stream";
92
92
  provider: string;
93
93
  baseUrl: string;
94
94
  reasoning: false;
95
- input: "text"[];
95
+ input: ("image" | "text")[];
96
96
  cost: {
97
97
  input: number;
98
98
  output: number;
@@ -102,14 +102,14 @@ export declare const MODELS: {
102
102
  contextWindow: number;
103
103
  maxTokens: number;
104
104
  };
105
- readonly "amazon.titan-text-express-v1:0:8k": {
105
+ readonly "anthropic.claude-3-5-sonnet-20240620-v1:0": {
106
106
  id: string;
107
107
  name: string;
108
108
  api: "bedrock-converse-stream";
109
109
  provider: string;
110
110
  baseUrl: string;
111
111
  reasoning: false;
112
- input: "text"[];
112
+ input: ("image" | "text")[];
113
113
  cost: {
114
114
  input: number;
115
115
  output: number;
@@ -119,7 +119,7 @@ export declare const MODELS: {
119
119
  contextWindow: number;
120
120
  maxTokens: number;
121
121
  };
122
- readonly "anthropic.claude-3-5-haiku-20241022-v1:0": {
122
+ readonly "anthropic.claude-3-5-sonnet-20241022-v2:0": {
123
123
  id: string;
124
124
  name: string;
125
125
  api: "bedrock-converse-stream";
@@ -136,7 +136,7 @@ export declare const MODELS: {
136
136
  contextWindow: number;
137
137
  maxTokens: number;
138
138
  };
139
- readonly "anthropic.claude-3-5-sonnet-20240620-v1:0": {
139
+ readonly "anthropic.claude-3-7-sonnet-20250219-v1:0": {
140
140
  id: string;
141
141
  name: string;
142
142
  api: "bedrock-converse-stream";
@@ -153,7 +153,7 @@ export declare const MODELS: {
153
153
  contextWindow: number;
154
154
  maxTokens: number;
155
155
  };
156
- readonly "anthropic.claude-3-5-sonnet-20241022-v2:0": {
156
+ readonly "anthropic.claude-3-haiku-20240307-v1:0": {
157
157
  id: string;
158
158
  name: string;
159
159
  api: "bedrock-converse-stream";
@@ -170,13 +170,13 @@ export declare const MODELS: {
170
170
  contextWindow: number;
171
171
  maxTokens: number;
172
172
  };
173
- readonly "anthropic.claude-3-7-sonnet-20250219-v1:0": {
173
+ readonly "anthropic.claude-haiku-4-5-20251001-v1:0": {
174
174
  id: string;
175
175
  name: string;
176
176
  api: "bedrock-converse-stream";
177
177
  provider: string;
178
178
  baseUrl: string;
179
- reasoning: false;
179
+ reasoning: true;
180
180
  input: ("image" | "text")[];
181
181
  cost: {
182
182
  input: number;
@@ -187,13 +187,13 @@ export declare const MODELS: {
187
187
  contextWindow: number;
188
188
  maxTokens: number;
189
189
  };
190
- readonly "anthropic.claude-3-haiku-20240307-v1:0": {
190
+ readonly "anthropic.claude-opus-4-1-20250805-v1:0": {
191
191
  id: string;
192
192
  name: string;
193
193
  api: "bedrock-converse-stream";
194
194
  provider: string;
195
195
  baseUrl: string;
196
- reasoning: false;
196
+ reasoning: true;
197
197
  input: ("image" | "text")[];
198
198
  cost: {
199
199
  input: number;
@@ -204,13 +204,13 @@ export declare const MODELS: {
204
204
  contextWindow: number;
205
205
  maxTokens: number;
206
206
  };
207
- readonly "anthropic.claude-3-opus-20240229-v1:0": {
207
+ readonly "anthropic.claude-opus-4-20250514-v1:0": {
208
208
  id: string;
209
209
  name: string;
210
210
  api: "bedrock-converse-stream";
211
211
  provider: string;
212
212
  baseUrl: string;
213
- reasoning: false;
213
+ reasoning: true;
214
214
  input: ("image" | "text")[];
215
215
  cost: {
216
216
  input: number;
@@ -221,13 +221,13 @@ export declare const MODELS: {
221
221
  contextWindow: number;
222
222
  maxTokens: number;
223
223
  };
224
- readonly "anthropic.claude-3-sonnet-20240229-v1:0": {
224
+ readonly "anthropic.claude-opus-4-5-20251101-v1:0": {
225
225
  id: string;
226
226
  name: string;
227
227
  api: "bedrock-converse-stream";
228
228
  provider: string;
229
229
  baseUrl: string;
230
- reasoning: false;
230
+ reasoning: true;
231
231
  input: ("image" | "text")[];
232
232
  cost: {
233
233
  input: number;
@@ -238,7 +238,7 @@ export declare const MODELS: {
238
238
  contextWindow: number;
239
239
  maxTokens: number;
240
240
  };
241
- readonly "anthropic.claude-haiku-4-5-20251001-v1:0": {
241
+ readonly "anthropic.claude-opus-4-6-v1": {
242
242
  id: string;
243
243
  name: string;
244
244
  api: "bedrock-converse-stream";
@@ -255,7 +255,7 @@ export declare const MODELS: {
255
255
  contextWindow: number;
256
256
  maxTokens: number;
257
257
  };
258
- readonly "anthropic.claude-opus-4-1-20250805-v1:0": {
258
+ readonly "anthropic.claude-sonnet-4-20250514-v1:0": {
259
259
  id: string;
260
260
  name: string;
261
261
  api: "bedrock-converse-stream";
@@ -272,7 +272,7 @@ export declare const MODELS: {
272
272
  contextWindow: number;
273
273
  maxTokens: number;
274
274
  };
275
- readonly "anthropic.claude-opus-4-20250514-v1:0": {
275
+ readonly "anthropic.claude-sonnet-4-5-20250929-v1:0": {
276
276
  id: string;
277
277
  name: string;
278
278
  api: "bedrock-converse-stream";
@@ -289,7 +289,7 @@ export declare const MODELS: {
289
289
  contextWindow: number;
290
290
  maxTokens: number;
291
291
  };
292
- readonly "anthropic.claude-opus-4-5-20251101-v1:0": {
292
+ readonly "anthropic.claude-sonnet-4-6": {
293
293
  id: string;
294
294
  name: string;
295
295
  api: "bedrock-converse-stream";
@@ -306,14 +306,14 @@ export declare const MODELS: {
306
306
  contextWindow: number;
307
307
  maxTokens: number;
308
308
  };
309
- readonly "anthropic.claude-opus-4-6-v1": {
309
+ readonly "deepseek.r1-v1:0": {
310
310
  id: string;
311
311
  name: string;
312
312
  api: "bedrock-converse-stream";
313
313
  provider: string;
314
314
  baseUrl: string;
315
315
  reasoning: true;
316
- input: ("image" | "text")[];
316
+ input: "text"[];
317
317
  cost: {
318
318
  input: number;
319
319
  output: number;
@@ -323,14 +323,14 @@ export declare const MODELS: {
323
323
  contextWindow: number;
324
324
  maxTokens: number;
325
325
  };
326
- readonly "anthropic.claude-sonnet-4-20250514-v1:0": {
326
+ readonly "deepseek.v3-v1:0": {
327
327
  id: string;
328
328
  name: string;
329
329
  api: "bedrock-converse-stream";
330
330
  provider: string;
331
331
  baseUrl: string;
332
332
  reasoning: true;
333
- input: ("image" | "text")[];
333
+ input: "text"[];
334
334
  cost: {
335
335
  input: number;
336
336
  output: number;
@@ -340,14 +340,14 @@ export declare const MODELS: {
340
340
  contextWindow: number;
341
341
  maxTokens: number;
342
342
  };
343
- readonly "anthropic.claude-sonnet-4-5-20250929-v1:0": {
343
+ readonly "deepseek.v3.2": {
344
344
  id: string;
345
345
  name: string;
346
346
  api: "bedrock-converse-stream";
347
347
  provider: string;
348
348
  baseUrl: string;
349
349
  reasoning: true;
350
- input: ("image" | "text")[];
350
+ input: "text"[];
351
351
  cost: {
352
352
  input: number;
353
353
  output: number;
@@ -357,7 +357,7 @@ export declare const MODELS: {
357
357
  contextWindow: number;
358
358
  maxTokens: number;
359
359
  };
360
- readonly "anthropic.claude-sonnet-4-6": {
360
+ readonly "eu.anthropic.claude-haiku-4-5-20251001-v1:0": {
361
361
  id: string;
362
362
  name: string;
363
363
  api: "bedrock-converse-stream";
@@ -374,14 +374,14 @@ export declare const MODELS: {
374
374
  contextWindow: number;
375
375
  maxTokens: number;
376
376
  };
377
- readonly "cohere.command-r-plus-v1:0": {
377
+ readonly "eu.anthropic.claude-opus-4-5-20251101-v1:0": {
378
378
  id: string;
379
379
  name: string;
380
380
  api: "bedrock-converse-stream";
381
381
  provider: string;
382
382
  baseUrl: string;
383
- reasoning: false;
384
- input: "text"[];
383
+ reasoning: true;
384
+ input: ("image" | "text")[];
385
385
  cost: {
386
386
  input: number;
387
387
  output: number;
@@ -391,14 +391,14 @@ export declare const MODELS: {
391
391
  contextWindow: number;
392
392
  maxTokens: number;
393
393
  };
394
- readonly "cohere.command-r-v1:0": {
394
+ readonly "eu.anthropic.claude-opus-4-6-v1": {
395
395
  id: string;
396
396
  name: string;
397
397
  api: "bedrock-converse-stream";
398
398
  provider: string;
399
399
  baseUrl: string;
400
- reasoning: false;
401
- input: "text"[];
400
+ reasoning: true;
401
+ input: ("image" | "text")[];
402
402
  cost: {
403
403
  input: number;
404
404
  output: number;
@@ -408,14 +408,14 @@ export declare const MODELS: {
408
408
  contextWindow: number;
409
409
  maxTokens: number;
410
410
  };
411
- readonly "deepseek.r1-v1:0": {
411
+ readonly "eu.anthropic.claude-sonnet-4-20250514-v1:0": {
412
412
  id: string;
413
413
  name: string;
414
414
  api: "bedrock-converse-stream";
415
415
  provider: string;
416
416
  baseUrl: string;
417
417
  reasoning: true;
418
- input: "text"[];
418
+ input: ("image" | "text")[];
419
419
  cost: {
420
420
  input: number;
421
421
  output: number;
@@ -425,14 +425,14 @@ export declare const MODELS: {
425
425
  contextWindow: number;
426
426
  maxTokens: number;
427
427
  };
428
- readonly "deepseek.v3-v1:0": {
428
+ readonly "eu.anthropic.claude-sonnet-4-5-20250929-v1:0": {
429
429
  id: string;
430
430
  name: string;
431
431
  api: "bedrock-converse-stream";
432
432
  provider: string;
433
433
  baseUrl: string;
434
434
  reasoning: true;
435
- input: "text"[];
435
+ input: ("image" | "text")[];
436
436
  cost: {
437
437
  input: number;
438
438
  output: number;
@@ -442,14 +442,14 @@ export declare const MODELS: {
442
442
  contextWindow: number;
443
443
  maxTokens: number;
444
444
  };
445
- readonly "deepseek.v3.2-v1:0": {
445
+ readonly "eu.anthropic.claude-sonnet-4-6": {
446
446
  id: string;
447
447
  name: string;
448
448
  api: "bedrock-converse-stream";
449
449
  provider: string;
450
450
  baseUrl: string;
451
451
  reasoning: true;
452
- input: "text"[];
452
+ input: ("image" | "text")[];
453
453
  cost: {
454
454
  input: number;
455
455
  output: number;
@@ -459,7 +459,7 @@ export declare const MODELS: {
459
459
  contextWindow: number;
460
460
  maxTokens: number;
461
461
  };
462
- readonly "eu.anthropic.claude-haiku-4-5-20251001-v1:0": {
462
+ readonly "global.anthropic.claude-haiku-4-5-20251001-v1:0": {
463
463
  id: string;
464
464
  name: string;
465
465
  api: "bedrock-converse-stream";
@@ -476,7 +476,7 @@ export declare const MODELS: {
476
476
  contextWindow: number;
477
477
  maxTokens: number;
478
478
  };
479
- readonly "eu.anthropic.claude-opus-4-5-20251101-v1:0": {
479
+ readonly "global.anthropic.claude-opus-4-5-20251101-v1:0": {
480
480
  id: string;
481
481
  name: string;
482
482
  api: "bedrock-converse-stream";
@@ -493,7 +493,7 @@ export declare const MODELS: {
493
493
  contextWindow: number;
494
494
  maxTokens: number;
495
495
  };
496
- readonly "eu.anthropic.claude-opus-4-6-v1": {
496
+ readonly "global.anthropic.claude-opus-4-6-v1": {
497
497
  id: string;
498
498
  name: string;
499
499
  api: "bedrock-converse-stream";
@@ -510,7 +510,7 @@ export declare const MODELS: {
510
510
  contextWindow: number;
511
511
  maxTokens: number;
512
512
  };
513
- readonly "eu.anthropic.claude-sonnet-4-20250514-v1:0": {
513
+ readonly "global.anthropic.claude-sonnet-4-20250514-v1:0": {
514
514
  id: string;
515
515
  name: string;
516
516
  api: "bedrock-converse-stream";
@@ -527,7 +527,7 @@ export declare const MODELS: {
527
527
  contextWindow: number;
528
528
  maxTokens: number;
529
529
  };
530
- readonly "eu.anthropic.claude-sonnet-4-5-20250929-v1:0": {
530
+ readonly "global.anthropic.claude-sonnet-4-5-20250929-v1:0": {
531
531
  id: string;
532
532
  name: string;
533
533
  api: "bedrock-converse-stream";
@@ -544,7 +544,7 @@ export declare const MODELS: {
544
544
  contextWindow: number;
545
545
  maxTokens: number;
546
546
  };
547
- readonly "eu.anthropic.claude-sonnet-4-6": {
547
+ readonly "global.anthropic.claude-sonnet-4-6": {
548
548
  id: string;
549
549
  name: string;
550
550
  api: "bedrock-converse-stream";
@@ -561,13 +561,13 @@ export declare const MODELS: {
561
561
  contextWindow: number;
562
562
  maxTokens: number;
563
563
  };
564
- readonly "global.anthropic.claude-haiku-4-5-20251001-v1:0": {
564
+ readonly "google.gemma-3-27b-it": {
565
565
  id: string;
566
566
  name: string;
567
567
  api: "bedrock-converse-stream";
568
568
  provider: string;
569
569
  baseUrl: string;
570
- reasoning: true;
570
+ reasoning: false;
571
571
  input: ("image" | "text")[];
572
572
  cost: {
573
573
  input: number;
@@ -578,13 +578,13 @@ export declare const MODELS: {
578
578
  contextWindow: number;
579
579
  maxTokens: number;
580
580
  };
581
- readonly "global.anthropic.claude-opus-4-5-20251101-v1:0": {
581
+ readonly "google.gemma-3-4b-it": {
582
582
  id: string;
583
583
  name: string;
584
584
  api: "bedrock-converse-stream";
585
585
  provider: string;
586
586
  baseUrl: string;
587
- reasoning: true;
587
+ reasoning: false;
588
588
  input: ("image" | "text")[];
589
589
  cost: {
590
590
  input: number;
@@ -595,14 +595,14 @@ export declare const MODELS: {
595
595
  contextWindow: number;
596
596
  maxTokens: number;
597
597
  };
598
- readonly "global.anthropic.claude-opus-4-6-v1": {
598
+ readonly "meta.llama3-1-405b-instruct-v1:0": {
599
599
  id: string;
600
600
  name: string;
601
601
  api: "bedrock-converse-stream";
602
602
  provider: string;
603
603
  baseUrl: string;
604
- reasoning: true;
605
- input: ("image" | "text")[];
604
+ reasoning: false;
605
+ input: "text"[];
606
606
  cost: {
607
607
  input: number;
608
608
  output: number;
@@ -612,14 +612,14 @@ export declare const MODELS: {
612
612
  contextWindow: number;
613
613
  maxTokens: number;
614
614
  };
615
- readonly "global.anthropic.claude-sonnet-4-20250514-v1:0": {
615
+ readonly "meta.llama3-1-70b-instruct-v1:0": {
616
616
  id: string;
617
617
  name: string;
618
618
  api: "bedrock-converse-stream";
619
619
  provider: string;
620
620
  baseUrl: string;
621
- reasoning: true;
622
- input: ("image" | "text")[];
621
+ reasoning: false;
622
+ input: "text"[];
623
623
  cost: {
624
624
  input: number;
625
625
  output: number;
@@ -629,14 +629,14 @@ export declare const MODELS: {
629
629
  contextWindow: number;
630
630
  maxTokens: number;
631
631
  };
632
- readonly "global.anthropic.claude-sonnet-4-5-20250929-v1:0": {
632
+ readonly "meta.llama3-1-8b-instruct-v1:0": {
633
633
  id: string;
634
634
  name: string;
635
635
  api: "bedrock-converse-stream";
636
636
  provider: string;
637
637
  baseUrl: string;
638
- reasoning: true;
639
- input: ("image" | "text")[];
638
+ reasoning: false;
639
+ input: "text"[];
640
640
  cost: {
641
641
  input: number;
642
642
  output: number;
@@ -646,13 +646,13 @@ export declare const MODELS: {
646
646
  contextWindow: number;
647
647
  maxTokens: number;
648
648
  };
649
- readonly "global.anthropic.claude-sonnet-4-6": {
649
+ readonly "meta.llama3-2-11b-instruct-v1:0": {
650
650
  id: string;
651
651
  name: string;
652
652
  api: "bedrock-converse-stream";
653
653
  provider: string;
654
654
  baseUrl: string;
655
- reasoning: true;
655
+ reasoning: false;
656
656
  input: ("image" | "text")[];
657
657
  cost: {
658
658
  input: number;
@@ -663,14 +663,14 @@ export declare const MODELS: {
663
663
  contextWindow: number;
664
664
  maxTokens: number;
665
665
  };
666
- readonly "google.gemma-3-27b-it": {
666
+ readonly "meta.llama3-2-1b-instruct-v1:0": {
667
667
  id: string;
668
668
  name: string;
669
669
  api: "bedrock-converse-stream";
670
670
  provider: string;
671
671
  baseUrl: string;
672
672
  reasoning: false;
673
- input: ("image" | "text")[];
673
+ input: "text"[];
674
674
  cost: {
675
675
  input: number;
676
676
  output: number;
@@ -680,14 +680,14 @@ export declare const MODELS: {
680
680
  contextWindow: number;
681
681
  maxTokens: number;
682
682
  };
683
- readonly "google.gemma-3-4b-it": {
683
+ readonly "meta.llama3-2-3b-instruct-v1:0": {
684
684
  id: string;
685
685
  name: string;
686
686
  api: "bedrock-converse-stream";
687
687
  provider: string;
688
688
  baseUrl: string;
689
689
  reasoning: false;
690
- input: ("image" | "text")[];
690
+ input: "text"[];
691
691
  cost: {
692
692
  input: number;
693
693
  output: number;
@@ -697,14 +697,14 @@ export declare const MODELS: {
697
697
  contextWindow: number;
698
698
  maxTokens: number;
699
699
  };
700
- readonly "meta.llama3-1-70b-instruct-v1:0": {
700
+ readonly "meta.llama3-2-90b-instruct-v1:0": {
701
701
  id: string;
702
702
  name: string;
703
703
  api: "bedrock-converse-stream";
704
704
  provider: string;
705
705
  baseUrl: string;
706
706
  reasoning: false;
707
- input: "text"[];
707
+ input: ("image" | "text")[];
708
708
  cost: {
709
709
  input: number;
710
710
  output: number;
@@ -714,7 +714,7 @@ export declare const MODELS: {
714
714
  contextWindow: number;
715
715
  maxTokens: number;
716
716
  };
717
- readonly "meta.llama3-1-8b-instruct-v1:0": {
717
+ readonly "meta.llama3-3-70b-instruct-v1:0": {
718
718
  id: string;
719
719
  name: string;
720
720
  api: "bedrock-converse-stream";
@@ -731,7 +731,7 @@ export declare const MODELS: {
731
731
  contextWindow: number;
732
732
  maxTokens: number;
733
733
  };
734
- readonly "meta.llama3-2-11b-instruct-v1:0": {
734
+ readonly "meta.llama4-maverick-17b-instruct-v1:0": {
735
735
  id: string;
736
736
  name: string;
737
737
  api: "bedrock-converse-stream";
@@ -748,14 +748,14 @@ export declare const MODELS: {
748
748
  contextWindow: number;
749
749
  maxTokens: number;
750
750
  };
751
- readonly "meta.llama3-2-1b-instruct-v1:0": {
751
+ readonly "meta.llama4-scout-17b-instruct-v1:0": {
752
752
  id: string;
753
753
  name: string;
754
754
  api: "bedrock-converse-stream";
755
755
  provider: string;
756
756
  baseUrl: string;
757
757
  reasoning: false;
758
- input: "text"[];
758
+ input: ("image" | "text")[];
759
759
  cost: {
760
760
  input: number;
761
761
  output: number;
@@ -765,13 +765,13 @@ export declare const MODELS: {
765
765
  contextWindow: number;
766
766
  maxTokens: number;
767
767
  };
768
- readonly "meta.llama3-2-3b-instruct-v1:0": {
768
+ readonly "minimax.minimax-m2": {
769
769
  id: string;
770
770
  name: string;
771
771
  api: "bedrock-converse-stream";
772
772
  provider: string;
773
773
  baseUrl: string;
774
- reasoning: false;
774
+ reasoning: true;
775
775
  input: "text"[];
776
776
  cost: {
777
777
  input: number;
@@ -782,14 +782,14 @@ export declare const MODELS: {
782
782
  contextWindow: number;
783
783
  maxTokens: number;
784
784
  };
785
- readonly "meta.llama3-2-90b-instruct-v1:0": {
785
+ readonly "minimax.minimax-m2.1": {
786
786
  id: string;
787
787
  name: string;
788
788
  api: "bedrock-converse-stream";
789
789
  provider: string;
790
790
  baseUrl: string;
791
- reasoning: false;
792
- input: ("image" | "text")[];
791
+ reasoning: true;
792
+ input: "text"[];
793
793
  cost: {
794
794
  input: number;
795
795
  output: number;
@@ -799,7 +799,7 @@ export declare const MODELS: {
799
799
  contextWindow: number;
800
800
  maxTokens: number;
801
801
  };
802
- readonly "meta.llama3-3-70b-instruct-v1:0": {
802
+ readonly "mistral.devstral-2-123b": {
803
803
  id: string;
804
804
  name: string;
805
805
  api: "bedrock-converse-stream";
@@ -816,13 +816,13 @@ export declare const MODELS: {
816
816
  contextWindow: number;
817
817
  maxTokens: number;
818
818
  };
819
- readonly "meta.llama4-maverick-17b-instruct-v1:0": {
819
+ readonly "mistral.magistral-small-2509": {
820
820
  id: string;
821
821
  name: string;
822
822
  api: "bedrock-converse-stream";
823
823
  provider: string;
824
824
  baseUrl: string;
825
- reasoning: false;
825
+ reasoning: true;
826
826
  input: ("image" | "text")[];
827
827
  cost: {
828
828
  input: number;
@@ -833,14 +833,14 @@ export declare const MODELS: {
833
833
  contextWindow: number;
834
834
  maxTokens: number;
835
835
  };
836
- readonly "meta.llama4-scout-17b-instruct-v1:0": {
836
+ readonly "mistral.ministral-3-14b-instruct": {
837
837
  id: string;
838
838
  name: string;
839
839
  api: "bedrock-converse-stream";
840
840
  provider: string;
841
841
  baseUrl: string;
842
842
  reasoning: false;
843
- input: ("image" | "text")[];
843
+ input: "text"[];
844
844
  cost: {
845
845
  input: number;
846
846
  output: number;
@@ -850,14 +850,14 @@ export declare const MODELS: {
850
850
  contextWindow: number;
851
851
  maxTokens: number;
852
852
  };
853
- readonly "minimax.minimax-m2": {
853
+ readonly "mistral.ministral-3-3b-instruct": {
854
854
  id: string;
855
855
  name: string;
856
856
  api: "bedrock-converse-stream";
857
857
  provider: string;
858
858
  baseUrl: string;
859
- reasoning: true;
860
- input: "text"[];
859
+ reasoning: false;
860
+ input: ("image" | "text")[];
861
861
  cost: {
862
862
  input: number;
863
863
  output: number;
@@ -867,13 +867,13 @@ export declare const MODELS: {
867
867
  contextWindow: number;
868
868
  maxTokens: number;
869
869
  };
870
- readonly "minimax.minimax-m2.1": {
870
+ readonly "mistral.ministral-3-8b-instruct": {
871
871
  id: string;
872
872
  name: string;
873
873
  api: "bedrock-converse-stream";
874
874
  provider: string;
875
875
  baseUrl: string;
876
- reasoning: true;
876
+ reasoning: false;
877
877
  input: "text"[];
878
878
  cost: {
879
879
  input: number;
@@ -884,14 +884,14 @@ export declare const MODELS: {
884
884
  contextWindow: number;
885
885
  maxTokens: number;
886
886
  };
887
- readonly "mistral.ministral-3-14b-instruct": {
887
+ readonly "mistral.mistral-large-3-675b-instruct": {
888
888
  id: string;
889
889
  name: string;
890
890
  api: "bedrock-converse-stream";
891
891
  provider: string;
892
892
  baseUrl: string;
893
893
  reasoning: false;
894
- input: "text"[];
894
+ input: ("image" | "text")[];
895
895
  cost: {
896
896
  input: number;
897
897
  output: number;
@@ -901,14 +901,14 @@ export declare const MODELS: {
901
901
  contextWindow: number;
902
902
  maxTokens: number;
903
903
  };
904
- readonly "mistral.ministral-3-8b-instruct": {
904
+ readonly "mistral.pixtral-large-2502-v1:0": {
905
905
  id: string;
906
906
  name: string;
907
907
  api: "bedrock-converse-stream";
908
908
  provider: string;
909
909
  baseUrl: string;
910
910
  reasoning: false;
911
- input: "text"[];
911
+ input: ("image" | "text")[];
912
912
  cost: {
913
913
  input: number;
914
914
  output: number;
@@ -918,7 +918,7 @@ export declare const MODELS: {
918
918
  contextWindow: number;
919
919
  maxTokens: number;
920
920
  };
921
- readonly "mistral.mistral-large-2402-v1:0": {
921
+ readonly "mistral.voxtral-mini-3b-2507": {
922
922
  id: string;
923
923
  name: string;
924
924
  api: "bedrock-converse-stream";
@@ -935,7 +935,7 @@ export declare const MODELS: {
935
935
  contextWindow: number;
936
936
  maxTokens: number;
937
937
  };
938
- readonly "mistral.voxtral-mini-3b-2507": {
938
+ readonly "mistral.voxtral-small-24b-2507": {
939
939
  id: string;
940
940
  name: string;
941
941
  api: "bedrock-converse-stream";
@@ -952,13 +952,13 @@ export declare const MODELS: {
952
952
  contextWindow: number;
953
953
  maxTokens: number;
954
954
  };
955
- readonly "mistral.voxtral-small-24b-2507": {
955
+ readonly "moonshot.kimi-k2-thinking": {
956
956
  id: string;
957
957
  name: string;
958
958
  api: "bedrock-converse-stream";
959
959
  provider: string;
960
960
  baseUrl: string;
961
- reasoning: false;
961
+ reasoning: true;
962
962
  input: "text"[];
963
963
  cost: {
964
964
  input: number;
@@ -969,14 +969,14 @@ export declare const MODELS: {
969
969
  contextWindow: number;
970
970
  maxTokens: number;
971
971
  };
972
- readonly "moonshot.kimi-k2-thinking": {
972
+ readonly "moonshotai.kimi-k2.5": {
973
973
  id: string;
974
974
  name: string;
975
975
  api: "bedrock-converse-stream";
976
976
  provider: string;
977
977
  baseUrl: string;
978
978
  reasoning: true;
979
- input: "text"[];
979
+ input: ("image" | "text")[];
980
980
  cost: {
981
981
  input: number;
982
982
  output: number;
@@ -986,13 +986,13 @@ export declare const MODELS: {
986
986
  contextWindow: number;
987
987
  maxTokens: number;
988
988
  };
989
- readonly "moonshotai.kimi-k2.5": {
989
+ readonly "nvidia.nemotron-nano-12b-v2": {
990
990
  id: string;
991
991
  name: string;
992
992
  api: "bedrock-converse-stream";
993
993
  provider: string;
994
994
  baseUrl: string;
995
- reasoning: true;
995
+ reasoning: false;
996
996
  input: ("image" | "text")[];
997
997
  cost: {
998
998
  input: number;
@@ -1003,14 +1003,14 @@ export declare const MODELS: {
1003
1003
  contextWindow: number;
1004
1004
  maxTokens: number;
1005
1005
  };
1006
- readonly "nvidia.nemotron-nano-12b-v2": {
1006
+ readonly "nvidia.nemotron-nano-3-30b": {
1007
1007
  id: string;
1008
1008
  name: string;
1009
1009
  api: "bedrock-converse-stream";
1010
1010
  provider: string;
1011
1011
  baseUrl: string;
1012
- reasoning: false;
1013
- input: ("image" | "text")[];
1012
+ reasoning: true;
1013
+ input: "text"[];
1014
1014
  cost: {
1015
1015
  input: number;
1016
1016
  output: number;
@@ -6449,6 +6449,23 @@ export declare const MODELS: {
6449
6449
  contextWindow: number;
6450
6450
  maxTokens: number;
6451
6451
  };
6452
+ readonly "mimo-v2-flash-free": {
6453
+ id: string;
6454
+ name: string;
6455
+ api: "openai-completions";
6456
+ provider: string;
6457
+ baseUrl: string;
6458
+ reasoning: true;
6459
+ input: "text"[];
6460
+ cost: {
6461
+ input: number;
6462
+ output: number;
6463
+ cacheRead: number;
6464
+ cacheWrite: number;
6465
+ };
6466
+ contextWindow: number;
6467
+ maxTokens: number;
6468
+ };
6452
6469
  readonly "minimax-m2.1": {
6453
6470
  id: string;
6454
6471
  name: string;
@@ -6500,6 +6517,23 @@ export declare const MODELS: {
6500
6517
  contextWindow: number;
6501
6518
  maxTokens: number;
6502
6519
  };
6520
+ readonly "nemotron-3-super-free": {
6521
+ id: string;
6522
+ name: string;
6523
+ api: "openai-completions";
6524
+ provider: string;
6525
+ baseUrl: string;
6526
+ reasoning: true;
6527
+ input: "text"[];
6528
+ cost: {
6529
+ input: number;
6530
+ output: number;
6531
+ cacheRead: number;
6532
+ cacheWrite: number;
6533
+ };
6534
+ contextWindow: number;
6535
+ maxTokens: number;
6536
+ };
6503
6537
  };
6504
6538
  readonly "opencode-go": {
6505
6539
  readonly "glm-5": {
@@ -7065,7 +7099,7 @@ export declare const MODELS: {
7065
7099
  contextWindow: number;
7066
7100
  maxTokens: number;
7067
7101
  };
7068
- readonly "bytedance-seed/seed-2.0-mini": {
7102
+ readonly "bytedance-seed/seed-2.0-lite": {
7069
7103
  id: string;
7070
7104
  name: string;
7071
7105
  api: "openai-completions";
@@ -7082,14 +7116,14 @@ export declare const MODELS: {
7082
7116
  contextWindow: number;
7083
7117
  maxTokens: number;
7084
7118
  };
7085
- readonly "cohere/command-r-08-2024": {
7119
+ readonly "bytedance-seed/seed-2.0-mini": {
7086
7120
  id: string;
7087
7121
  name: string;
7088
7122
  api: "openai-completions";
7089
7123
  provider: string;
7090
7124
  baseUrl: string;
7091
- reasoning: false;
7092
- input: "text"[];
7125
+ reasoning: true;
7126
+ input: ("image" | "text")[];
7093
7127
  cost: {
7094
7128
  input: number;
7095
7129
  output: number;
@@ -7099,7 +7133,7 @@ export declare const MODELS: {
7099
7133
  contextWindow: number;
7100
7134
  maxTokens: number;
7101
7135
  };
7102
- readonly "cohere/command-r-plus-08-2024": {
7136
+ readonly "cohere/command-r-08-2024": {
7103
7137
  id: string;
7104
7138
  name: string;
7105
7139
  api: "openai-completions";
@@ -7116,7 +7150,7 @@ export declare const MODELS: {
7116
7150
  contextWindow: number;
7117
7151
  maxTokens: number;
7118
7152
  };
7119
- readonly "deepseek/deepseek-chat": {
7153
+ readonly "cohere/command-r-plus-08-2024": {
7120
7154
  id: string;
7121
7155
  name: string;
7122
7156
  api: "openai-completions";
@@ -7133,13 +7167,13 @@ export declare const MODELS: {
7133
7167
  contextWindow: number;
7134
7168
  maxTokens: number;
7135
7169
  };
7136
- readonly "deepseek/deepseek-chat-v3-0324": {
7170
+ readonly "deepseek/deepseek-chat": {
7137
7171
  id: string;
7138
7172
  name: string;
7139
7173
  api: "openai-completions";
7140
7174
  provider: string;
7141
7175
  baseUrl: string;
7142
- reasoning: true;
7176
+ reasoning: false;
7143
7177
  input: "text"[];
7144
7178
  cost: {
7145
7179
  input: number;
@@ -7150,7 +7184,7 @@ export declare const MODELS: {
7150
7184
  contextWindow: number;
7151
7185
  maxTokens: number;
7152
7186
  };
7153
- readonly "deepseek/deepseek-chat-v3.1": {
7187
+ readonly "deepseek/deepseek-chat-v3-0324": {
7154
7188
  id: string;
7155
7189
  name: string;
7156
7190
  api: "openai-completions";
@@ -7167,7 +7201,7 @@ export declare const MODELS: {
7167
7201
  contextWindow: number;
7168
7202
  maxTokens: number;
7169
7203
  };
7170
- readonly "deepseek/deepseek-r1": {
7204
+ readonly "deepseek/deepseek-chat-v3.1": {
7171
7205
  id: string;
7172
7206
  name: string;
7173
7207
  api: "openai-completions";
@@ -7184,7 +7218,7 @@ export declare const MODELS: {
7184
7218
  contextWindow: number;
7185
7219
  maxTokens: number;
7186
7220
  };
7187
- readonly "deepseek/deepseek-r1-0528": {
7221
+ readonly "deepseek/deepseek-r1": {
7188
7222
  id: string;
7189
7223
  name: string;
7190
7224
  api: "openai-completions";
@@ -7201,7 +7235,7 @@ export declare const MODELS: {
7201
7235
  contextWindow: number;
7202
7236
  maxTokens: number;
7203
7237
  };
7204
- readonly "deepseek/deepseek-v3.1-terminus": {
7238
+ readonly "deepseek/deepseek-r1-0528": {
7205
7239
  id: string;
7206
7240
  name: string;
7207
7241
  api: "openai-completions";
@@ -7218,7 +7252,7 @@ export declare const MODELS: {
7218
7252
  contextWindow: number;
7219
7253
  maxTokens: number;
7220
7254
  };
7221
- readonly "deepseek/deepseek-v3.1-terminus:exacto": {
7255
+ readonly "deepseek/deepseek-v3.1-terminus": {
7222
7256
  id: string;
7223
7257
  name: string;
7224
7258
  api: "openai-completions";
@@ -7643,23 +7677,6 @@ export declare const MODELS: {
7643
7677
  contextWindow: number;
7644
7678
  maxTokens: number;
7645
7679
  };
7646
- readonly "meta-llama/llama-3.1-405b-instruct": {
7647
- id: string;
7648
- name: string;
7649
- api: "openai-completions";
7650
- provider: string;
7651
- baseUrl: string;
7652
- reasoning: false;
7653
- input: "text"[];
7654
- cost: {
7655
- input: number;
7656
- output: number;
7657
- cacheRead: number;
7658
- cacheWrite: number;
7659
- };
7660
- contextWindow: number;
7661
- maxTokens: number;
7662
- };
7663
7680
  readonly "meta-llama/llama-3.1-70b-instruct": {
7664
7681
  id: string;
7665
7682
  name: string;
@@ -8255,13 +8272,13 @@ export declare const MODELS: {
8255
8272
  contextWindow: number;
8256
8273
  maxTokens: number;
8257
8274
  };
8258
- readonly "moonshotai/kimi-k2-0905:exacto": {
8275
+ readonly "moonshotai/kimi-k2-thinking": {
8259
8276
  id: string;
8260
8277
  name: string;
8261
8278
  api: "openai-completions";
8262
8279
  provider: string;
8263
8280
  baseUrl: string;
8264
- reasoning: false;
8281
+ reasoning: true;
8265
8282
  input: "text"[];
8266
8283
  cost: {
8267
8284
  input: number;
@@ -8272,14 +8289,14 @@ export declare const MODELS: {
8272
8289
  contextWindow: number;
8273
8290
  maxTokens: number;
8274
8291
  };
8275
- readonly "moonshotai/kimi-k2-thinking": {
8292
+ readonly "moonshotai/kimi-k2.5": {
8276
8293
  id: string;
8277
8294
  name: string;
8278
8295
  api: "openai-completions";
8279
8296
  provider: string;
8280
8297
  baseUrl: string;
8281
8298
  reasoning: true;
8282
- input: "text"[];
8299
+ input: ("image" | "text")[];
8283
8300
  cost: {
8284
8301
  input: number;
8285
8302
  output: number;
@@ -8289,14 +8306,14 @@ export declare const MODELS: {
8289
8306
  contextWindow: number;
8290
8307
  maxTokens: number;
8291
8308
  };
8292
- readonly "moonshotai/kimi-k2.5": {
8309
+ readonly "nex-agi/deepseek-v3.1-nex-n1": {
8293
8310
  id: string;
8294
8311
  name: string;
8295
8312
  api: "openai-completions";
8296
8313
  provider: string;
8297
8314
  baseUrl: string;
8298
- reasoning: true;
8299
- input: ("image" | "text")[];
8315
+ reasoning: false;
8316
+ input: "text"[];
8300
8317
  cost: {
8301
8318
  input: number;
8302
8319
  output: number;
@@ -8306,7 +8323,7 @@ export declare const MODELS: {
8306
8323
  contextWindow: number;
8307
8324
  maxTokens: number;
8308
8325
  };
8309
- readonly "nex-agi/deepseek-v3.1-nex-n1": {
8326
+ readonly "nvidia/llama-3.1-nemotron-70b-instruct": {
8310
8327
  id: string;
8311
8328
  name: string;
8312
8329
  api: "openai-completions";
@@ -8323,13 +8340,13 @@ export declare const MODELS: {
8323
8340
  contextWindow: number;
8324
8341
  maxTokens: number;
8325
8342
  };
8326
- readonly "nvidia/llama-3.1-nemotron-70b-instruct": {
8343
+ readonly "nvidia/llama-3.3-nemotron-super-49b-v1.5": {
8327
8344
  id: string;
8328
8345
  name: string;
8329
8346
  api: "openai-completions";
8330
8347
  provider: string;
8331
8348
  baseUrl: string;
8332
- reasoning: false;
8349
+ reasoning: true;
8333
8350
  input: "text"[];
8334
8351
  cost: {
8335
8352
  input: number;
@@ -8340,7 +8357,7 @@ export declare const MODELS: {
8340
8357
  contextWindow: number;
8341
8358
  maxTokens: number;
8342
8359
  };
8343
- readonly "nvidia/llama-3.3-nemotron-super-49b-v1.5": {
8360
+ readonly "nvidia/nemotron-3-nano-30b-a3b": {
8344
8361
  id: string;
8345
8362
  name: string;
8346
8363
  api: "openai-completions";
@@ -8357,7 +8374,7 @@ export declare const MODELS: {
8357
8374
  contextWindow: number;
8358
8375
  maxTokens: number;
8359
8376
  };
8360
- readonly "nvidia/nemotron-3-nano-30b-a3b": {
8377
+ readonly "nvidia/nemotron-3-nano-30b-a3b:free": {
8361
8378
  id: string;
8362
8379
  name: string;
8363
8380
  api: "openai-completions";
@@ -8374,7 +8391,7 @@ export declare const MODELS: {
8374
8391
  contextWindow: number;
8375
8392
  maxTokens: number;
8376
8393
  };
8377
- readonly "nvidia/nemotron-3-nano-30b-a3b:free": {
8394
+ readonly "nvidia/nemotron-3-super-120b-a12b:free": {
8378
8395
  id: string;
8379
8396
  name: string;
8380
8397
  api: "openai-completions";
@@ -9122,23 +9139,6 @@ export declare const MODELS: {
9122
9139
  contextWindow: number;
9123
9140
  maxTokens: number;
9124
9141
  };
9125
- readonly "openai/gpt-oss-120b:exacto": {
9126
- id: string;
9127
- name: string;
9128
- api: "openai-completions";
9129
- provider: string;
9130
- baseUrl: string;
9131
- reasoning: true;
9132
- input: "text"[];
9133
- cost: {
9134
- input: number;
9135
- output: number;
9136
- cacheRead: number;
9137
- cacheWrite: number;
9138
- };
9139
- contextWindow: number;
9140
- maxTokens: number;
9141
- };
9142
9142
  readonly "openai/gpt-oss-120b:free": {
9143
9143
  id: string;
9144
9144
  name: string;
@@ -9394,6 +9394,40 @@ export declare const MODELS: {
9394
9394
  contextWindow: number;
9395
9395
  maxTokens: number;
9396
9396
  };
9397
+ readonly "openrouter/healer-alpha": {
9398
+ id: string;
9399
+ name: string;
9400
+ api: "openai-completions";
9401
+ provider: string;
9402
+ baseUrl: string;
9403
+ reasoning: true;
9404
+ input: ("image" | "text")[];
9405
+ cost: {
9406
+ input: number;
9407
+ output: number;
9408
+ cacheRead: number;
9409
+ cacheWrite: number;
9410
+ };
9411
+ contextWindow: number;
9412
+ maxTokens: number;
9413
+ };
9414
+ readonly "openrouter/hunter-alpha": {
9415
+ id: string;
9416
+ name: string;
9417
+ api: "openai-completions";
9418
+ provider: string;
9419
+ baseUrl: string;
9420
+ reasoning: true;
9421
+ input: "text"[];
9422
+ cost: {
9423
+ input: number;
9424
+ output: number;
9425
+ cacheRead: number;
9426
+ cacheWrite: number;
9427
+ };
9428
+ contextWindow: number;
9429
+ maxTokens: number;
9430
+ };
9397
9431
  readonly "prime-intellect/intellect-3": {
9398
9432
  id: string;
9399
9433
  name: string;
@@ -9802,23 +9836,6 @@ export declare const MODELS: {
9802
9836
  contextWindow: number;
9803
9837
  maxTokens: number;
9804
9838
  };
9805
- readonly "qwen/qwen3-coder:exacto": {
9806
- id: string;
9807
- name: string;
9808
- api: "openai-completions";
9809
- provider: string;
9810
- baseUrl: string;
9811
- reasoning: false;
9812
- input: "text"[];
9813
- cost: {
9814
- input: number;
9815
- output: number;
9816
- cacheRead: number;
9817
- cacheWrite: number;
9818
- };
9819
- contextWindow: number;
9820
- maxTokens: number;
9821
- };
9822
9839
  readonly "qwen/qwen3-coder:free": {
9823
9840
  id: string;
9824
9841
  name: string;
@@ -10108,6 +10125,23 @@ export declare const MODELS: {
10108
10125
  contextWindow: number;
10109
10126
  maxTokens: number;
10110
10127
  };
10128
+ readonly "qwen/qwen3.5-9b": {
10129
+ id: string;
10130
+ name: string;
10131
+ api: "openai-completions";
10132
+ provider: string;
10133
+ baseUrl: string;
10134
+ reasoning: true;
10135
+ input: ("image" | "text")[];
10136
+ cost: {
10137
+ input: number;
10138
+ output: number;
10139
+ cacheRead: number;
10140
+ cacheWrite: number;
10141
+ };
10142
+ contextWindow: number;
10143
+ maxTokens: number;
10144
+ };
10111
10145
  readonly "qwen/qwen3.5-flash-02-23": {
10112
10146
  id: string;
10113
10147
  name: string;
@@ -10431,14 +10465,14 @@ export declare const MODELS: {
10431
10465
  contextWindow: number;
10432
10466
  maxTokens: number;
10433
10467
  };
10434
- readonly "x-ai/grok-code-fast-1": {
10468
+ readonly "x-ai/grok-4.20-beta": {
10435
10469
  id: string;
10436
10470
  name: string;
10437
10471
  api: "openai-completions";
10438
10472
  provider: string;
10439
10473
  baseUrl: string;
10440
10474
  reasoning: true;
10441
- input: "text"[];
10475
+ input: ("image" | "text")[];
10442
10476
  cost: {
10443
10477
  input: number;
10444
10478
  output: number;
@@ -10448,7 +10482,7 @@ export declare const MODELS: {
10448
10482
  contextWindow: number;
10449
10483
  maxTokens: number;
10450
10484
  };
10451
- readonly "xiaomi/mimo-v2-flash": {
10485
+ readonly "x-ai/grok-code-fast-1": {
10452
10486
  id: string;
10453
10487
  name: string;
10454
10488
  api: "openai-completions";
@@ -10465,13 +10499,13 @@ export declare const MODELS: {
10465
10499
  contextWindow: number;
10466
10500
  maxTokens: number;
10467
10501
  };
10468
- readonly "z-ai/glm-4-32b": {
10502
+ readonly "xiaomi/mimo-v2-flash": {
10469
10503
  id: string;
10470
10504
  name: string;
10471
10505
  api: "openai-completions";
10472
10506
  provider: string;
10473
10507
  baseUrl: string;
10474
- reasoning: false;
10508
+ reasoning: true;
10475
10509
  input: "text"[];
10476
10510
  cost: {
10477
10511
  input: number;
@@ -10482,13 +10516,13 @@ export declare const MODELS: {
10482
10516
  contextWindow: number;
10483
10517
  maxTokens: number;
10484
10518
  };
10485
- readonly "z-ai/glm-4.5": {
10519
+ readonly "z-ai/glm-4-32b": {
10486
10520
  id: string;
10487
10521
  name: string;
10488
10522
  api: "openai-completions";
10489
10523
  provider: string;
10490
10524
  baseUrl: string;
10491
- reasoning: true;
10525
+ reasoning: false;
10492
10526
  input: "text"[];
10493
10527
  cost: {
10494
10528
  input: number;
@@ -10499,7 +10533,7 @@ export declare const MODELS: {
10499
10533
  contextWindow: number;
10500
10534
  maxTokens: number;
10501
10535
  };
10502
- readonly "z-ai/glm-4.5-air": {
10536
+ readonly "z-ai/glm-4.5": {
10503
10537
  id: string;
10504
10538
  name: string;
10505
10539
  api: "openai-completions";
@@ -10516,7 +10550,7 @@ export declare const MODELS: {
10516
10550
  contextWindow: number;
10517
10551
  maxTokens: number;
10518
10552
  };
10519
- readonly "z-ai/glm-4.5-air:free": {
10553
+ readonly "z-ai/glm-4.5-air": {
10520
10554
  id: string;
10521
10555
  name: string;
10522
10556
  api: "openai-completions";
@@ -10533,14 +10567,14 @@ export declare const MODELS: {
10533
10567
  contextWindow: number;
10534
10568
  maxTokens: number;
10535
10569
  };
10536
- readonly "z-ai/glm-4.5v": {
10570
+ readonly "z-ai/glm-4.5-air:free": {
10537
10571
  id: string;
10538
10572
  name: string;
10539
10573
  api: "openai-completions";
10540
10574
  provider: string;
10541
10575
  baseUrl: string;
10542
10576
  reasoning: true;
10543
- input: ("image" | "text")[];
10577
+ input: "text"[];
10544
10578
  cost: {
10545
10579
  input: number;
10546
10580
  output: number;
@@ -10550,14 +10584,14 @@ export declare const MODELS: {
10550
10584
  contextWindow: number;
10551
10585
  maxTokens: number;
10552
10586
  };
10553
- readonly "z-ai/glm-4.6": {
10587
+ readonly "z-ai/glm-4.5v": {
10554
10588
  id: string;
10555
10589
  name: string;
10556
10590
  api: "openai-completions";
10557
10591
  provider: string;
10558
10592
  baseUrl: string;
10559
10593
  reasoning: true;
10560
- input: "text"[];
10594
+ input: ("image" | "text")[];
10561
10595
  cost: {
10562
10596
  input: number;
10563
10597
  output: number;
@@ -10567,7 +10601,7 @@ export declare const MODELS: {
10567
10601
  contextWindow: number;
10568
10602
  maxTokens: number;
10569
10603
  };
10570
- readonly "z-ai/glm-4.6:exacto": {
10604
+ readonly "z-ai/glm-4.6": {
10571
10605
  id: string;
10572
10606
  name: string;
10573
10607
  api: "openai-completions";
@@ -10779,7 +10813,7 @@ export declare const MODELS: {
10779
10813
  api: "anthropic-messages";
10780
10814
  provider: string;
10781
10815
  baseUrl: string;
10782
- reasoning: true;
10816
+ reasoning: false;
10783
10817
  input: "text"[];
10784
10818
  cost: {
10785
10819
  input: number;
@@ -10807,6 +10841,23 @@ export declare const MODELS: {
10807
10841
  contextWindow: number;
10808
10842
  maxTokens: number;
10809
10843
  };
10844
+ readonly "alibaba/qwen3-max": {
10845
+ id: string;
10846
+ name: string;
10847
+ api: "anthropic-messages";
10848
+ provider: string;
10849
+ baseUrl: string;
10850
+ reasoning: false;
10851
+ input: "text"[];
10852
+ cost: {
10853
+ input: number;
10854
+ output: number;
10855
+ cacheRead: number;
10856
+ cacheWrite: number;
10857
+ };
10858
+ contextWindow: number;
10859
+ maxTokens: number;
10860
+ };
10810
10861
  readonly "alibaba/qwen3-max-preview": {
10811
10862
  id: string;
10812
10863
  name: string;
@@ -11164,6 +11215,23 @@ export declare const MODELS: {
11164
11215
  contextWindow: number;
11165
11216
  maxTokens: number;
11166
11217
  };
11218
+ readonly "deepseek/deepseek-r1": {
11219
+ id: string;
11220
+ name: string;
11221
+ api: "anthropic-messages";
11222
+ provider: string;
11223
+ baseUrl: string;
11224
+ reasoning: true;
11225
+ input: "text"[];
11226
+ cost: {
11227
+ input: number;
11228
+ output: number;
11229
+ cacheRead: number;
11230
+ cacheWrite: number;
11231
+ };
11232
+ contextWindow: number;
11233
+ maxTokens: number;
11234
+ };
11167
11235
  readonly "deepseek/deepseek-v3": {
11168
11236
  id: string;
11169
11237
  name: string;
@@ -11249,6 +11317,40 @@ export declare const MODELS: {
11249
11317
  contextWindow: number;
11250
11318
  maxTokens: number;
11251
11319
  };
11320
+ readonly "google/gemini-2.0-flash": {
11321
+ id: string;
11322
+ name: string;
11323
+ api: "anthropic-messages";
11324
+ provider: string;
11325
+ baseUrl: string;
11326
+ reasoning: false;
11327
+ input: ("image" | "text")[];
11328
+ cost: {
11329
+ input: number;
11330
+ output: number;
11331
+ cacheRead: number;
11332
+ cacheWrite: number;
11333
+ };
11334
+ contextWindow: number;
11335
+ maxTokens: number;
11336
+ };
11337
+ readonly "google/gemini-2.0-flash-lite": {
11338
+ id: string;
11339
+ name: string;
11340
+ api: "anthropic-messages";
11341
+ provider: string;
11342
+ baseUrl: string;
11343
+ reasoning: false;
11344
+ input: ("image" | "text")[];
11345
+ cost: {
11346
+ input: number;
11347
+ output: number;
11348
+ cacheRead: number;
11349
+ cacheWrite: number;
11350
+ };
11351
+ contextWindow: number;
11352
+ maxTokens: number;
11353
+ };
11252
11354
  readonly "google/gemini-2.5-flash": {
11253
11355
  id: string;
11254
11356
  name: string;
@@ -11256,7 +11358,7 @@ export declare const MODELS: {
11256
11358
  provider: string;
11257
11359
  baseUrl: string;
11258
11360
  reasoning: true;
11259
- input: "text"[];
11361
+ input: ("image" | "text")[];
11260
11362
  cost: {
11261
11363
  input: number;
11262
11364
  output: number;
@@ -11324,7 +11426,7 @@ export declare const MODELS: {
11324
11426
  provider: string;
11325
11427
  baseUrl: string;
11326
11428
  reasoning: true;
11327
- input: "text"[];
11429
+ input: ("image" | "text")[];
11328
11430
  cost: {
11329
11431
  input: number;
11330
11432
  output: number;
@@ -11861,13 +11963,13 @@ export declare const MODELS: {
11861
11963
  contextWindow: number;
11862
11964
  maxTokens: number;
11863
11965
  };
11864
- readonly "moonshotai/kimi-k2-thinking": {
11966
+ readonly "moonshotai/kimi-k2-0905": {
11865
11967
  id: string;
11866
11968
  name: string;
11867
11969
  api: "anthropic-messages";
11868
11970
  provider: string;
11869
11971
  baseUrl: string;
11870
- reasoning: true;
11972
+ reasoning: false;
11871
11973
  input: "text"[];
11872
11974
  cost: {
11873
11975
  input: number;
@@ -11878,7 +11980,7 @@ export declare const MODELS: {
11878
11980
  contextWindow: number;
11879
11981
  maxTokens: number;
11880
11982
  };
11881
- readonly "moonshotai/kimi-k2-thinking-turbo": {
11983
+ readonly "moonshotai/kimi-k2-thinking": {
11882
11984
  id: string;
11883
11985
  name: string;
11884
11986
  api: "anthropic-messages";
@@ -11895,13 +11997,13 @@ export declare const MODELS: {
11895
11997
  contextWindow: number;
11896
11998
  maxTokens: number;
11897
11999
  };
11898
- readonly "moonshotai/kimi-k2-turbo": {
12000
+ readonly "moonshotai/kimi-k2-thinking-turbo": {
11899
12001
  id: string;
11900
12002
  name: string;
11901
12003
  api: "anthropic-messages";
11902
12004
  provider: string;
11903
12005
  baseUrl: string;
11904
- reasoning: false;
12006
+ reasoning: true;
11905
12007
  input: "text"[];
11906
12008
  cost: {
11907
12009
  input: number;
@@ -11912,14 +12014,14 @@ export declare const MODELS: {
11912
12014
  contextWindow: number;
11913
12015
  maxTokens: number;
11914
12016
  };
11915
- readonly "moonshotai/kimi-k2.5": {
12017
+ readonly "moonshotai/kimi-k2-turbo": {
11916
12018
  id: string;
11917
12019
  name: string;
11918
12020
  api: "anthropic-messages";
11919
12021
  provider: string;
11920
12022
  baseUrl: string;
11921
- reasoning: true;
11922
- input: ("image" | "text")[];
12023
+ reasoning: false;
12024
+ input: "text"[];
11923
12025
  cost: {
11924
12026
  input: number;
11925
12027
  output: number;
@@ -11929,7 +12031,7 @@ export declare const MODELS: {
11929
12031
  contextWindow: number;
11930
12032
  maxTokens: number;
11931
12033
  };
11932
- readonly "nvidia/nemotron-nano-12b-v2-vl": {
12034
+ readonly "moonshotai/kimi-k2.5": {
11933
12035
  id: string;
11934
12036
  name: string;
11935
12037
  api: "anthropic-messages";
@@ -11946,14 +12048,14 @@ export declare const MODELS: {
11946
12048
  contextWindow: number;
11947
12049
  maxTokens: number;
11948
12050
  };
11949
- readonly "nvidia/nemotron-nano-9b-v2": {
12051
+ readonly "nvidia/nemotron-nano-12b-v2-vl": {
11950
12052
  id: string;
11951
12053
  name: string;
11952
12054
  api: "anthropic-messages";
11953
12055
  provider: string;
11954
12056
  baseUrl: string;
11955
12057
  reasoning: true;
11956
- input: "text"[];
12058
+ input: ("image" | "text")[];
11957
12059
  cost: {
11958
12060
  input: number;
11959
12061
  output: number;
@@ -11963,14 +12065,14 @@ export declare const MODELS: {
11963
12065
  contextWindow: number;
11964
12066
  maxTokens: number;
11965
12067
  };
11966
- readonly "openai/codex-mini": {
12068
+ readonly "nvidia/nemotron-nano-9b-v2": {
11967
12069
  id: string;
11968
12070
  name: string;
11969
12071
  api: "anthropic-messages";
11970
12072
  provider: string;
11971
12073
  baseUrl: string;
11972
12074
  reasoning: true;
11973
- input: ("image" | "text")[];
12075
+ input: "text"[];
11974
12076
  cost: {
11975
12077
  input: number;
11976
12078
  output: number;
@@ -12123,7 +12225,7 @@ export declare const MODELS: {
12123
12225
  provider: string;
12124
12226
  baseUrl: string;
12125
12227
  reasoning: true;
12126
- input: ("image" | "text")[];
12228
+ input: "text"[];
12127
12229
  cost: {
12128
12230
  input: number;
12129
12231
  output: number;
@@ -12405,23 +12507,6 @@ export declare const MODELS: {
12405
12507
  contextWindow: number;
12406
12508
  maxTokens: number;
12407
12509
  };
12408
- readonly "openai/gpt-oss-120b": {
12409
- id: string;
12410
- name: string;
12411
- api: "anthropic-messages";
12412
- provider: string;
12413
- baseUrl: string;
12414
- reasoning: true;
12415
- input: "text"[];
12416
- cost: {
12417
- input: number;
12418
- output: number;
12419
- cacheRead: number;
12420
- cacheWrite: number;
12421
- };
12422
- contextWindow: number;
12423
- maxTokens: number;
12424
- };
12425
12510
  readonly "openai/gpt-oss-20b": {
12426
12511
  id: string;
12427
12512
  name: string;
@@ -12813,6 +12898,57 @@ export declare const MODELS: {
12813
12898
  contextWindow: number;
12814
12899
  maxTokens: number;
12815
12900
  };
12901
+ readonly "xai/grok-4.20-multi-agent-beta": {
12902
+ id: string;
12903
+ name: string;
12904
+ api: "anthropic-messages";
12905
+ provider: string;
12906
+ baseUrl: string;
12907
+ reasoning: true;
12908
+ input: "text"[];
12909
+ cost: {
12910
+ input: number;
12911
+ output: number;
12912
+ cacheRead: number;
12913
+ cacheWrite: number;
12914
+ };
12915
+ contextWindow: number;
12916
+ maxTokens: number;
12917
+ };
12918
+ readonly "xai/grok-4.20-non-reasoning-beta": {
12919
+ id: string;
12920
+ name: string;
12921
+ api: "anthropic-messages";
12922
+ provider: string;
12923
+ baseUrl: string;
12924
+ reasoning: false;
12925
+ input: ("image" | "text")[];
12926
+ cost: {
12927
+ input: number;
12928
+ output: number;
12929
+ cacheRead: number;
12930
+ cacheWrite: number;
12931
+ };
12932
+ contextWindow: number;
12933
+ maxTokens: number;
12934
+ };
12935
+ readonly "xai/grok-4.20-reasoning-beta": {
12936
+ id: string;
12937
+ name: string;
12938
+ api: "anthropic-messages";
12939
+ provider: string;
12940
+ baseUrl: string;
12941
+ reasoning: true;
12942
+ input: ("image" | "text")[];
12943
+ cost: {
12944
+ input: number;
12945
+ output: number;
12946
+ cacheRead: number;
12947
+ cacheWrite: number;
12948
+ };
12949
+ contextWindow: number;
12950
+ maxTokens: number;
12951
+ };
12816
12952
  readonly "xai/grok-code-fast-1": {
12817
12953
  id: string;
12818
12954
  name: string;
@@ -12887,7 +13023,7 @@ export declare const MODELS: {
12887
13023
  api: "anthropic-messages";
12888
13024
  provider: string;
12889
13025
  baseUrl: string;
12890
- reasoning: true;
13026
+ reasoning: false;
12891
13027
  input: ("image" | "text")[];
12892
13028
  cost: {
12893
13029
  input: number;
@@ -12966,6 +13102,23 @@ export declare const MODELS: {
12966
13102
  contextWindow: number;
12967
13103
  maxTokens: number;
12968
13104
  };
13105
+ readonly "zai/glm-4.7-flash": {
13106
+ id: string;
13107
+ name: string;
13108
+ api: "anthropic-messages";
13109
+ provider: string;
13110
+ baseUrl: string;
13111
+ reasoning: true;
13112
+ input: "text"[];
13113
+ cost: {
13114
+ input: number;
13115
+ output: number;
13116
+ cacheRead: number;
13117
+ cacheWrite: number;
13118
+ };
13119
+ contextWindow: number;
13120
+ maxTokens: number;
13121
+ };
12969
13122
  readonly "zai/glm-4.7-flashx": {
12970
13123
  id: string;
12971
13124
  name: string;
@@ -13325,6 +13478,57 @@ export declare const MODELS: {
13325
13478
  contextWindow: number;
13326
13479
  maxTokens: number;
13327
13480
  };
13481
+ readonly "grok-4.20-beta-latest-non-reasoning": {
13482
+ id: string;
13483
+ name: string;
13484
+ api: "openai-completions";
13485
+ provider: string;
13486
+ baseUrl: string;
13487
+ reasoning: false;
13488
+ input: ("image" | "text")[];
13489
+ cost: {
13490
+ input: number;
13491
+ output: number;
13492
+ cacheRead: number;
13493
+ cacheWrite: number;
13494
+ };
13495
+ contextWindow: number;
13496
+ maxTokens: number;
13497
+ };
13498
+ readonly "grok-4.20-beta-latest-reasoning": {
13499
+ id: string;
13500
+ name: string;
13501
+ api: "openai-completions";
13502
+ provider: string;
13503
+ baseUrl: string;
13504
+ reasoning: true;
13505
+ input: ("image" | "text")[];
13506
+ cost: {
13507
+ input: number;
13508
+ output: number;
13509
+ cacheRead: number;
13510
+ cacheWrite: number;
13511
+ };
13512
+ contextWindow: number;
13513
+ maxTokens: number;
13514
+ };
13515
+ readonly "grok-4.20-multi-agent-beta-latest": {
13516
+ id: string;
13517
+ name: string;
13518
+ api: "openai-completions";
13519
+ provider: string;
13520
+ baseUrl: string;
13521
+ reasoning: true;
13522
+ input: ("image" | "text")[];
13523
+ cost: {
13524
+ input: number;
13525
+ output: number;
13526
+ cacheRead: number;
13527
+ cacheWrite: number;
13528
+ };
13529
+ contextWindow: number;
13530
+ maxTokens: number;
13531
+ };
13328
13532
  readonly "grok-beta": {
13329
13533
  id: string;
13330
13534
  name: string;