@mariozechner/pi-ai 0.5.48 → 0.7.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (55) hide show
  1. package/README.md +47 -11
  2. package/dist/agent/agent-loop.d.ts.map +1 -1
  3. package/dist/agent/agent-loop.js +3 -1
  4. package/dist/agent/agent-loop.js.map +1 -1
  5. package/dist/agent/index.d.ts.map +1 -1
  6. package/dist/agent/tools/calculate.d.ts +6 -3
  7. package/dist/agent/tools/calculate.d.ts.map +1 -1
  8. package/dist/agent/tools/calculate.js +1 -1
  9. package/dist/agent/tools/calculate.js.map +1 -1
  10. package/dist/agent/tools/get-current-time.d.ts.map +1 -1
  11. package/dist/agent/tools/get-current-time.js +8 -6
  12. package/dist/agent/tools/get-current-time.js.map +1 -1
  13. package/dist/agent/tools/index.d.ts.map +1 -1
  14. package/dist/agent/types.d.ts +2 -2
  15. package/dist/agent/types.d.ts.map +1 -1
  16. package/dist/agent/types.js.map +1 -1
  17. package/dist/index.d.ts.map +1 -1
  18. package/dist/models.d.ts.map +1 -1
  19. package/dist/models.generated.d.ts +1628 -302
  20. package/dist/models.generated.d.ts.map +1 -1
  21. package/dist/models.generated.js +1515 -189
  22. package/dist/models.generated.js.map +1 -1
  23. package/dist/models.js.map +1 -1
  24. package/dist/providers/anthropic.d.ts.map +1 -1
  25. package/dist/providers/anthropic.js +42 -4
  26. package/dist/providers/anthropic.js.map +1 -1
  27. package/dist/providers/google.d.ts.map +1 -1
  28. package/dist/providers/google.js +32 -12
  29. package/dist/providers/google.js.map +1 -1
  30. package/dist/providers/openai-completions.d.ts.map +1 -1
  31. package/dist/providers/openai-completions.js +33 -1
  32. package/dist/providers/openai-completions.js.map +1 -1
  33. package/dist/providers/openai-responses.d.ts.map +1 -1
  34. package/dist/providers/openai-responses.js +32 -1
  35. package/dist/providers/openai-responses.js.map +1 -1
  36. package/dist/providers/transorm-messages.d.ts.map +1 -1
  37. package/dist/providers/transorm-messages.js.map +1 -1
  38. package/dist/stream.d.ts.map +1 -1
  39. package/dist/stream.js +2 -2
  40. package/dist/stream.js.map +1 -1
  41. package/dist/types.d.ts +1 -1
  42. package/dist/types.d.ts.map +1 -1
  43. package/dist/types.js.map +1 -1
  44. package/dist/utils/event-stream.d.ts.map +1 -1
  45. package/dist/utils/event-stream.js +7 -3
  46. package/dist/utils/event-stream.js.map +1 -1
  47. package/dist/utils/json-parse.d.ts.map +1 -1
  48. package/dist/utils/json-parse.js.map +1 -1
  49. package/dist/utils/sanitize-unicode.d.ts.map +1 -1
  50. package/dist/utils/sanitize-unicode.js.map +1 -1
  51. package/dist/utils/typebox-helpers.d.ts.map +1 -1
  52. package/dist/utils/typebox-helpers.js.map +1 -1
  53. package/dist/utils/validation.d.ts.map +1 -1
  54. package/dist/utils/validation.js.map +1 -1
  55. package/package.json +5 -4
@@ -7,7 +7,7 @@ export declare const MODELS: {
7
7
  provider: string;
8
8
  baseUrl: string;
9
9
  reasoning: true;
10
- input: ("text" | "image")[];
10
+ input: ("image" | "text")[];
11
11
  cost: {
12
12
  input: number;
13
13
  output: number;
@@ -24,7 +24,7 @@ export declare const MODELS: {
24
24
  provider: string;
25
25
  baseUrl: string;
26
26
  reasoning: false;
27
- input: ("text" | "image")[];
27
+ input: ("image" | "text")[];
28
28
  cost: {
29
29
  input: number;
30
30
  output: number;
@@ -41,7 +41,7 @@ export declare const MODELS: {
41
41
  provider: string;
42
42
  baseUrl: string;
43
43
  reasoning: true;
44
- input: ("text" | "image")[];
44
+ input: ("image" | "text")[];
45
45
  cost: {
46
46
  input: number;
47
47
  output: number;
@@ -58,7 +58,7 @@ export declare const MODELS: {
58
58
  provider: string;
59
59
  baseUrl: string;
60
60
  reasoning: true;
61
- input: ("text" | "image")[];
61
+ input: ("image" | "text")[];
62
62
  cost: {
63
63
  input: number;
64
64
  output: number;
@@ -75,7 +75,7 @@ export declare const MODELS: {
75
75
  provider: string;
76
76
  baseUrl: string;
77
77
  reasoning: false;
78
- input: ("text" | "image")[];
78
+ input: ("image" | "text")[];
79
79
  cost: {
80
80
  input: number;
81
81
  output: number;
@@ -92,7 +92,7 @@ export declare const MODELS: {
92
92
  provider: string;
93
93
  baseUrl: string;
94
94
  reasoning: false;
95
- input: ("text" | "image")[];
95
+ input: ("image" | "text")[];
96
96
  cost: {
97
97
  input: number;
98
98
  output: number;
@@ -109,7 +109,7 @@ export declare const MODELS: {
109
109
  provider: string;
110
110
  baseUrl: string;
111
111
  reasoning: false;
112
- input: ("text" | "image")[];
112
+ input: ("image" | "text")[];
113
113
  cost: {
114
114
  input: number;
115
115
  output: number;
@@ -126,7 +126,7 @@ export declare const MODELS: {
126
126
  provider: string;
127
127
  baseUrl: string;
128
128
  reasoning: true;
129
- input: ("text" | "image")[];
129
+ input: ("image" | "text")[];
130
130
  cost: {
131
131
  input: number;
132
132
  output: number;
@@ -143,7 +143,7 @@ export declare const MODELS: {
143
143
  provider: string;
144
144
  baseUrl: string;
145
145
  reasoning: true;
146
- input: ("text" | "image")[];
146
+ input: ("image" | "text")[];
147
147
  cost: {
148
148
  input: number;
149
149
  output: number;
@@ -160,7 +160,7 @@ export declare const MODELS: {
160
160
  provider: string;
161
161
  baseUrl: string;
162
162
  reasoning: true;
163
- input: ("text" | "image")[];
163
+ input: ("image" | "text")[];
164
164
  cost: {
165
165
  input: number;
166
166
  output: number;
@@ -177,7 +177,7 @@ export declare const MODELS: {
177
177
  provider: string;
178
178
  baseUrl: string;
179
179
  reasoning: true;
180
- input: ("text" | "image")[];
180
+ input: ("image" | "text")[];
181
181
  cost: {
182
182
  input: number;
183
183
  output: number;
@@ -194,7 +194,7 @@ export declare const MODELS: {
194
194
  provider: string;
195
195
  baseUrl: string;
196
196
  reasoning: false;
197
- input: ("text" | "image")[];
197
+ input: ("image" | "text")[];
198
198
  cost: {
199
199
  input: number;
200
200
  output: number;
@@ -211,7 +211,7 @@ export declare const MODELS: {
211
211
  provider: string;
212
212
  baseUrl: string;
213
213
  reasoning: false;
214
- input: ("text" | "image")[];
214
+ input: ("image" | "text")[];
215
215
  cost: {
216
216
  input: number;
217
217
  output: number;
@@ -228,7 +228,7 @@ export declare const MODELS: {
228
228
  provider: string;
229
229
  baseUrl: string;
230
230
  reasoning: true;
231
- input: ("text" | "image")[];
231
+ input: ("image" | "text")[];
232
232
  cost: {
233
233
  input: number;
234
234
  output: number;
@@ -245,7 +245,7 @@ export declare const MODELS: {
245
245
  provider: string;
246
246
  baseUrl: string;
247
247
  reasoning: true;
248
- input: ("text" | "image")[];
248
+ input: ("image" | "text")[];
249
249
  cost: {
250
250
  input: number;
251
251
  output: number;
@@ -262,7 +262,7 @@ export declare const MODELS: {
262
262
  provider: string;
263
263
  baseUrl: string;
264
264
  reasoning: true;
265
- input: ("text" | "image")[];
265
+ input: ("image" | "text")[];
266
266
  cost: {
267
267
  input: number;
268
268
  output: number;
@@ -279,7 +279,7 @@ export declare const MODELS: {
279
279
  provider: string;
280
280
  baseUrl: string;
281
281
  reasoning: true;
282
- input: ("text" | "image")[];
282
+ input: ("image" | "text")[];
283
283
  cost: {
284
284
  input: number;
285
285
  output: number;
@@ -296,7 +296,7 @@ export declare const MODELS: {
296
296
  provider: string;
297
297
  baseUrl: string;
298
298
  reasoning: false;
299
- input: ("text" | "image")[];
299
+ input: ("image" | "text")[];
300
300
  cost: {
301
301
  input: number;
302
302
  output: number;
@@ -313,7 +313,7 @@ export declare const MODELS: {
313
313
  provider: string;
314
314
  baseUrl: string;
315
315
  reasoning: true;
316
- input: ("text" | "image")[];
316
+ input: ("image" | "text")[];
317
317
  cost: {
318
318
  input: number;
319
319
  output: number;
@@ -332,7 +332,7 @@ export declare const MODELS: {
332
332
  provider: string;
333
333
  baseUrl: string;
334
334
  reasoning: true;
335
- input: ("text" | "image")[];
335
+ input: ("image" | "text")[];
336
336
  cost: {
337
337
  input: number;
338
338
  output: number;
@@ -349,7 +349,7 @@ export declare const MODELS: {
349
349
  provider: string;
350
350
  baseUrl: string;
351
351
  reasoning: true;
352
- input: ("text" | "image")[];
352
+ input: ("image" | "text")[];
353
353
  cost: {
354
354
  input: number;
355
355
  output: number;
@@ -366,7 +366,7 @@ export declare const MODELS: {
366
366
  provider: string;
367
367
  baseUrl: string;
368
368
  reasoning: true;
369
- input: ("text" | "image")[];
369
+ input: ("image" | "text")[];
370
370
  cost: {
371
371
  input: number;
372
372
  output: number;
@@ -383,7 +383,7 @@ export declare const MODELS: {
383
383
  provider: string;
384
384
  baseUrl: string;
385
385
  reasoning: true;
386
- input: ("text" | "image")[];
386
+ input: ("image" | "text")[];
387
387
  cost: {
388
388
  input: number;
389
389
  output: number;
@@ -400,7 +400,7 @@ export declare const MODELS: {
400
400
  provider: string;
401
401
  baseUrl: string;
402
402
  reasoning: true;
403
- input: ("text" | "image")[];
403
+ input: ("image" | "text")[];
404
404
  cost: {
405
405
  input: number;
406
406
  output: number;
@@ -417,7 +417,7 @@ export declare const MODELS: {
417
417
  provider: string;
418
418
  baseUrl: string;
419
419
  reasoning: false;
420
- input: ("text" | "image")[];
420
+ input: ("image" | "text")[];
421
421
  cost: {
422
422
  input: number;
423
423
  output: number;
@@ -451,7 +451,7 @@ export declare const MODELS: {
451
451
  provider: string;
452
452
  baseUrl: string;
453
453
  reasoning: false;
454
- input: ("text" | "image")[];
454
+ input: ("image" | "text")[];
455
455
  cost: {
456
456
  input: number;
457
457
  output: number;
@@ -468,7 +468,7 @@ export declare const MODELS: {
468
468
  provider: string;
469
469
  baseUrl: string;
470
470
  reasoning: true;
471
- input: ("text" | "image")[];
471
+ input: ("image" | "text")[];
472
472
  cost: {
473
473
  input: number;
474
474
  output: number;
@@ -485,7 +485,7 @@ export declare const MODELS: {
485
485
  provider: string;
486
486
  baseUrl: string;
487
487
  reasoning: true;
488
- input: ("text" | "image")[];
488
+ input: ("image" | "text")[];
489
489
  cost: {
490
490
  input: number;
491
491
  output: number;
@@ -502,7 +502,7 @@ export declare const MODELS: {
502
502
  provider: string;
503
503
  baseUrl: string;
504
504
  reasoning: true;
505
- input: ("text" | "image")[];
505
+ input: ("image" | "text")[];
506
506
  cost: {
507
507
  input: number;
508
508
  output: number;
@@ -519,7 +519,7 @@ export declare const MODELS: {
519
519
  provider: string;
520
520
  baseUrl: string;
521
521
  reasoning: true;
522
- input: ("text" | "image")[];
522
+ input: ("image" | "text")[];
523
523
  cost: {
524
524
  input: number;
525
525
  output: number;
@@ -536,7 +536,7 @@ export declare const MODELS: {
536
536
  provider: string;
537
537
  baseUrl: string;
538
538
  reasoning: true;
539
- input: ("text" | "image")[];
539
+ input: ("image" | "text")[];
540
540
  cost: {
541
541
  input: number;
542
542
  output: number;
@@ -553,7 +553,7 @@ export declare const MODELS: {
553
553
  provider: string;
554
554
  baseUrl: string;
555
555
  reasoning: true;
556
- input: ("text" | "image")[];
556
+ input: ("image" | "text")[];
557
557
  cost: {
558
558
  input: number;
559
559
  output: number;
@@ -570,7 +570,7 @@ export declare const MODELS: {
570
570
  provider: string;
571
571
  baseUrl: string;
572
572
  reasoning: true;
573
- input: ("text" | "image")[];
573
+ input: ("image" | "text")[];
574
574
  cost: {
575
575
  input: number;
576
576
  output: number;
@@ -587,7 +587,7 @@ export declare const MODELS: {
587
587
  provider: string;
588
588
  baseUrl: string;
589
589
  reasoning: false;
590
- input: ("text" | "image")[];
590
+ input: ("image" | "text")[];
591
591
  cost: {
592
592
  input: number;
593
593
  output: number;
@@ -604,7 +604,7 @@ export declare const MODELS: {
604
604
  provider: string;
605
605
  baseUrl: string;
606
606
  reasoning: false;
607
- input: ("text" | "image")[];
607
+ input: ("image" | "text")[];
608
608
  cost: {
609
609
  input: number;
610
610
  output: number;
@@ -621,7 +621,7 @@ export declare const MODELS: {
621
621
  provider: string;
622
622
  baseUrl: string;
623
623
  reasoning: true;
624
- input: ("text" | "image")[];
624
+ input: ("image" | "text")[];
625
625
  cost: {
626
626
  input: number;
627
627
  output: number;
@@ -638,7 +638,7 @@ export declare const MODELS: {
638
638
  provider: string;
639
639
  baseUrl: string;
640
640
  reasoning: false;
641
- input: ("text" | "image")[];
641
+ input: ("image" | "text")[];
642
642
  cost: {
643
643
  input: number;
644
644
  output: number;
@@ -657,7 +657,7 @@ export declare const MODELS: {
657
657
  provider: string;
658
658
  baseUrl: string;
659
659
  reasoning: false;
660
- input: ("text" | "image")[];
660
+ input: ("image" | "text")[];
661
661
  cost: {
662
662
  input: number;
663
663
  output: number;
@@ -691,7 +691,7 @@ export declare const MODELS: {
691
691
  provider: string;
692
692
  baseUrl: string;
693
693
  reasoning: true;
694
- input: ("text" | "image")[];
694
+ input: ("image" | "text")[];
695
695
  cost: {
696
696
  input: number;
697
697
  output: number;
@@ -708,7 +708,7 @@ export declare const MODELS: {
708
708
  provider: string;
709
709
  baseUrl: string;
710
710
  reasoning: false;
711
- input: ("text" | "image")[];
711
+ input: ("image" | "text")[];
712
712
  cost: {
713
713
  input: number;
714
714
  output: number;
@@ -725,7 +725,7 @@ export declare const MODELS: {
725
725
  provider: string;
726
726
  baseUrl: string;
727
727
  reasoning: false;
728
- input: ("text" | "image")[];
728
+ input: ("image" | "text")[];
729
729
  cost: {
730
730
  input: number;
731
731
  output: number;
@@ -742,7 +742,7 @@ export declare const MODELS: {
742
742
  provider: string;
743
743
  baseUrl: string;
744
744
  reasoning: false;
745
- input: ("text" | "image")[];
745
+ input: ("image" | "text")[];
746
746
  cost: {
747
747
  input: number;
748
748
  output: number;
@@ -759,7 +759,7 @@ export declare const MODELS: {
759
759
  provider: string;
760
760
  baseUrl: string;
761
761
  reasoning: true;
762
- input: ("text" | "image")[];
762
+ input: ("image" | "text")[];
763
763
  cost: {
764
764
  input: number;
765
765
  output: number;
@@ -776,7 +776,7 @@ export declare const MODELS: {
776
776
  provider: string;
777
777
  baseUrl: string;
778
778
  reasoning: false;
779
- input: ("text" | "image")[];
779
+ input: ("image" | "text")[];
780
780
  cost: {
781
781
  input: number;
782
782
  output: number;
@@ -827,7 +827,7 @@ export declare const MODELS: {
827
827
  provider: string;
828
828
  baseUrl: string;
829
829
  reasoning: true;
830
- input: ("text" | "image")[];
830
+ input: ("image" | "text")[];
831
831
  cost: {
832
832
  input: number;
833
833
  output: number;
@@ -844,7 +844,7 @@ export declare const MODELS: {
844
844
  provider: string;
845
845
  baseUrl: string;
846
846
  reasoning: true;
847
- input: ("text" | "image")[];
847
+ input: ("image" | "text")[];
848
848
  cost: {
849
849
  input: number;
850
850
  output: number;
@@ -861,7 +861,7 @@ export declare const MODELS: {
861
861
  provider: string;
862
862
  baseUrl: string;
863
863
  reasoning: false;
864
- input: ("text" | "image")[];
864
+ input: ("image" | "text")[];
865
865
  cost: {
866
866
  input: number;
867
867
  output: number;
@@ -878,7 +878,7 @@ export declare const MODELS: {
878
878
  provider: string;
879
879
  baseUrl: string;
880
880
  reasoning: false;
881
- input: ("text" | "image")[];
881
+ input: ("image" | "text")[];
882
882
  cost: {
883
883
  input: number;
884
884
  output: number;
@@ -895,7 +895,7 @@ export declare const MODELS: {
895
895
  provider: string;
896
896
  baseUrl: string;
897
897
  reasoning: true;
898
- input: ("text" | "image")[];
898
+ input: ("image" | "text")[];
899
899
  cost: {
900
900
  input: number;
901
901
  output: number;
@@ -912,7 +912,7 @@ export declare const MODELS: {
912
912
  provider: string;
913
913
  baseUrl: string;
914
914
  reasoning: true;
915
- input: ("text" | "image")[];
915
+ input: ("image" | "text")[];
916
916
  cost: {
917
917
  input: number;
918
918
  output: number;
@@ -929,7 +929,7 @@ export declare const MODELS: {
929
929
  provider: string;
930
930
  baseUrl: string;
931
931
  reasoning: true;
932
- input: ("text" | "image")[];
932
+ input: ("image" | "text")[];
933
933
  cost: {
934
934
  input: number;
935
935
  output: number;
@@ -946,7 +946,7 @@ export declare const MODELS: {
946
946
  provider: string;
947
947
  baseUrl: string;
948
948
  reasoning: true;
949
- input: ("text" | "image")[];
949
+ input: ("image" | "text")[];
950
950
  cost: {
951
951
  input: number;
952
952
  output: number;
@@ -963,7 +963,7 @@ export declare const MODELS: {
963
963
  provider: string;
964
964
  baseUrl: string;
965
965
  reasoning: false;
966
- input: ("text" | "image")[];
966
+ input: ("image" | "text")[];
967
967
  cost: {
968
968
  input: number;
969
969
  output: number;
@@ -980,7 +980,7 @@ export declare const MODELS: {
980
980
  provider: string;
981
981
  baseUrl: string;
982
982
  reasoning: true;
983
- input: ("text" | "image")[];
983
+ input: ("image" | "text")[];
984
984
  cost: {
985
985
  input: number;
986
986
  output: number;
@@ -997,7 +997,7 @@ export declare const MODELS: {
997
997
  provider: string;
998
998
  baseUrl: string;
999
999
  reasoning: true;
1000
- input: ("text" | "image")[];
1000
+ input: ("image" | "text")[];
1001
1001
  cost: {
1002
1002
  input: number;
1003
1003
  output: number;
@@ -1014,7 +1014,7 @@ export declare const MODELS: {
1014
1014
  provider: string;
1015
1015
  baseUrl: string;
1016
1016
  reasoning: false;
1017
- input: ("text" | "image")[];
1017
+ input: ("image" | "text")[];
1018
1018
  cost: {
1019
1019
  input: number;
1020
1020
  output: number;
@@ -1031,7 +1031,24 @@ export declare const MODELS: {
1031
1031
  provider: string;
1032
1032
  baseUrl: string;
1033
1033
  reasoning: true;
1034
- input: ("text" | "image")[];
1034
+ input: ("image" | "text")[];
1035
+ cost: {
1036
+ input: number;
1037
+ output: number;
1038
+ cacheRead: number;
1039
+ cacheWrite: number;
1040
+ };
1041
+ contextWindow: number;
1042
+ maxTokens: number;
1043
+ };
1044
+ readonly "gpt-5-pro": {
1045
+ id: string;
1046
+ name: string;
1047
+ api: "openai-responses";
1048
+ provider: string;
1049
+ baseUrl: string;
1050
+ reasoning: true;
1051
+ input: ("image" | "text")[];
1035
1052
  cost: {
1036
1053
  input: number;
1037
1054
  output: number;
@@ -1048,7 +1065,7 @@ export declare const MODELS: {
1048
1065
  provider: string;
1049
1066
  baseUrl: string;
1050
1067
  reasoning: false;
1051
- input: ("text" | "image")[];
1068
+ input: ("image" | "text")[];
1052
1069
  cost: {
1053
1070
  input: number;
1054
1071
  output: number;
@@ -1288,7 +1305,7 @@ export declare const MODELS: {
1288
1305
  provider: string;
1289
1306
  baseUrl: string;
1290
1307
  reasoning: false;
1291
- input: ("text" | "image")[];
1308
+ input: ("image" | "text")[];
1292
1309
  cost: {
1293
1310
  input: number;
1294
1311
  output: number;
@@ -1305,7 +1322,7 @@ export declare const MODELS: {
1305
1322
  provider: string;
1306
1323
  baseUrl: string;
1307
1324
  reasoning: false;
1308
- input: ("text" | "image")[];
1325
+ input: ("image" | "text")[];
1309
1326
  cost: {
1310
1327
  input: number;
1311
1328
  output: number;
@@ -1334,6 +1351,23 @@ export declare const MODELS: {
1334
1351
  contextWindow: number;
1335
1352
  maxTokens: number;
1336
1353
  };
1354
+ readonly "zai-glm-4.6": {
1355
+ id: string;
1356
+ name: string;
1357
+ api: "openai-completions";
1358
+ provider: string;
1359
+ baseUrl: string;
1360
+ reasoning: false;
1361
+ input: "text"[];
1362
+ cost: {
1363
+ input: number;
1364
+ output: number;
1365
+ cacheRead: number;
1366
+ cacheWrite: number;
1367
+ };
1368
+ contextWindow: number;
1369
+ maxTokens: number;
1370
+ };
1337
1371
  readonly "qwen-3-coder-480b": {
1338
1372
  id: string;
1339
1373
  name: string;
@@ -1377,7 +1411,7 @@ export declare const MODELS: {
1377
1411
  provider: string;
1378
1412
  baseUrl: string;
1379
1413
  reasoning: false;
1380
- input: ("text" | "image")[];
1414
+ input: ("image" | "text")[];
1381
1415
  cost: {
1382
1416
  input: number;
1383
1417
  output: number;
@@ -1428,7 +1462,7 @@ export declare const MODELS: {
1428
1462
  provider: string;
1429
1463
  baseUrl: string;
1430
1464
  reasoning: false;
1431
- input: ("text" | "image")[];
1465
+ input: ("image" | "text")[];
1432
1466
  cost: {
1433
1467
  input: number;
1434
1468
  output: number;
@@ -1496,7 +1530,7 @@ export declare const MODELS: {
1496
1530
  provider: string;
1497
1531
  baseUrl: string;
1498
1532
  reasoning: false;
1499
- input: ("text" | "image")[];
1533
+ input: ("image" | "text")[];
1500
1534
  cost: {
1501
1535
  input: number;
1502
1536
  output: number;
@@ -1530,7 +1564,7 @@ export declare const MODELS: {
1530
1564
  provider: string;
1531
1565
  baseUrl: string;
1532
1566
  reasoning: true;
1533
- input: ("text" | "image")[];
1567
+ input: ("image" | "text")[];
1534
1568
  cost: {
1535
1569
  input: number;
1536
1570
  output: number;
@@ -1615,7 +1649,7 @@ export declare const MODELS: {
1615
1649
  provider: string;
1616
1650
  baseUrl: string;
1617
1651
  reasoning: false;
1618
- input: ("text" | "image")[];
1652
+ input: ("image" | "text")[];
1619
1653
  cost: {
1620
1654
  input: number;
1621
1655
  output: number;
@@ -1632,7 +1666,7 @@ export declare const MODELS: {
1632
1666
  provider: string;
1633
1667
  baseUrl: string;
1634
1668
  reasoning: false;
1635
- input: ("text" | "image")[];
1669
+ input: ("image" | "text")[];
1636
1670
  cost: {
1637
1671
  input: number;
1638
1672
  output: number;
@@ -1770,7 +1804,7 @@ export declare const MODELS: {
1770
1804
  provider: string;
1771
1805
  baseUrl: string;
1772
1806
  reasoning: true;
1773
- input: ("text" | "image")[];
1807
+ input: ("image" | "text")[];
1774
1808
  cost: {
1775
1809
  input: number;
1776
1810
  output: number;
@@ -1799,13 +1833,13 @@ export declare const MODELS: {
1799
1833
  };
1800
1834
  };
1801
1835
  readonly openrouter: {
1802
- readonly "minimax/minimax-m2:free": {
1836
+ readonly "kwaipilot/kat-coder-pro:free": {
1803
1837
  id: string;
1804
1838
  name: string;
1805
1839
  api: "openai-completions";
1806
1840
  provider: string;
1807
1841
  baseUrl: string;
1808
- reasoning: true;
1842
+ reasoning: false;
1809
1843
  input: "text"[];
1810
1844
  cost: {
1811
1845
  input: number;
@@ -1816,14 +1850,14 @@ export declare const MODELS: {
1816
1850
  contextWindow: number;
1817
1851
  maxTokens: number;
1818
1852
  };
1819
- readonly "openrouter/andromeda-alpha": {
1853
+ readonly "openrouter/polaris-alpha": {
1820
1854
  id: string;
1821
1855
  name: string;
1822
1856
  api: "openai-completions";
1823
1857
  provider: string;
1824
1858
  baseUrl: string;
1825
- reasoning: true;
1826
- input: ("text" | "image")[];
1859
+ reasoning: false;
1860
+ input: ("image" | "text")[];
1827
1861
  cost: {
1828
1862
  input: number;
1829
1863
  output: number;
@@ -1833,7 +1867,7 @@ export declare const MODELS: {
1833
1867
  contextWindow: number;
1834
1868
  maxTokens: number;
1835
1869
  };
1836
- readonly "deepcogito/cogito-v2-preview-llama-405b": {
1870
+ readonly "moonshotai/kimi-k2-thinking": {
1837
1871
  id: string;
1838
1872
  name: string;
1839
1873
  api: "openai-completions";
@@ -1850,14 +1884,14 @@ export declare const MODELS: {
1850
1884
  contextWindow: number;
1851
1885
  maxTokens: number;
1852
1886
  };
1853
- readonly "qwen/qwen3-vl-8b-thinking": {
1887
+ readonly "amazon/nova-premier-v1": {
1854
1888
  id: string;
1855
1889
  name: string;
1856
1890
  api: "openai-completions";
1857
1891
  provider: string;
1858
1892
  baseUrl: string;
1859
- reasoning: true;
1860
- input: ("text" | "image")[];
1893
+ reasoning: false;
1894
+ input: ("image" | "text")[];
1861
1895
  cost: {
1862
1896
  input: number;
1863
1897
  output: number;
@@ -1867,14 +1901,14 @@ export declare const MODELS: {
1867
1901
  contextWindow: number;
1868
1902
  maxTokens: number;
1869
1903
  };
1870
- readonly "qwen/qwen3-vl-8b-instruct": {
1904
+ readonly "mistralai/voxtral-small-24b-2507": {
1871
1905
  id: string;
1872
1906
  name: string;
1873
1907
  api: "openai-completions";
1874
1908
  provider: string;
1875
1909
  baseUrl: string;
1876
1910
  reasoning: false;
1877
- input: ("text" | "image")[];
1911
+ input: "text"[];
1878
1912
  cost: {
1879
1913
  input: number;
1880
1914
  output: number;
@@ -1884,7 +1918,7 @@ export declare const MODELS: {
1884
1918
  contextWindow: number;
1885
1919
  maxTokens: number;
1886
1920
  };
1887
- readonly "inclusionai/ring-1t": {
1921
+ readonly "openai/gpt-oss-safeguard-20b": {
1888
1922
  id: string;
1889
1923
  name: string;
1890
1924
  api: "openai-completions";
@@ -1901,14 +1935,14 @@ export declare const MODELS: {
1901
1935
  contextWindow: number;
1902
1936
  maxTokens: number;
1903
1937
  };
1904
- readonly "inclusionai/ling-1t": {
1938
+ readonly "nvidia/nemotron-nano-12b-v2-vl:free": {
1905
1939
  id: string;
1906
1940
  name: string;
1907
1941
  api: "openai-completions";
1908
1942
  provider: string;
1909
1943
  baseUrl: string;
1910
- reasoning: false;
1911
- input: "text"[];
1944
+ reasoning: true;
1945
+ input: ("image" | "text")[];
1912
1946
  cost: {
1913
1947
  input: number;
1914
1948
  output: number;
@@ -1918,7 +1952,7 @@ export declare const MODELS: {
1918
1952
  contextWindow: number;
1919
1953
  maxTokens: number;
1920
1954
  };
1921
- readonly "nvidia/llama-3.3-nemotron-super-49b-v1.5": {
1955
+ readonly "minimax/minimax-m2": {
1922
1956
  id: string;
1923
1957
  name: string;
1924
1958
  api: "openai-completions";
@@ -1935,14 +1969,14 @@ export declare const MODELS: {
1935
1969
  contextWindow: number;
1936
1970
  maxTokens: number;
1937
1971
  };
1938
- readonly "qwen/qwen3-vl-30b-a3b-thinking": {
1972
+ readonly "deepcogito/cogito-v2-preview-llama-405b": {
1939
1973
  id: string;
1940
1974
  name: string;
1941
1975
  api: "openai-completions";
1942
1976
  provider: string;
1943
1977
  baseUrl: string;
1944
1978
  reasoning: true;
1945
- input: ("text" | "image")[];
1979
+ input: "text"[];
1946
1980
  cost: {
1947
1981
  input: number;
1948
1982
  output: number;
@@ -1952,14 +1986,14 @@ export declare const MODELS: {
1952
1986
  contextWindow: number;
1953
1987
  maxTokens: number;
1954
1988
  };
1955
- readonly "qwen/qwen3-vl-30b-a3b-instruct": {
1989
+ readonly "openai/gpt-5-image-mini": {
1956
1990
  id: string;
1957
1991
  name: string;
1958
1992
  api: "openai-completions";
1959
1993
  provider: string;
1960
1994
  baseUrl: string;
1961
- reasoning: false;
1962
- input: ("text" | "image")[];
1995
+ reasoning: true;
1996
+ input: ("image" | "text")[];
1963
1997
  cost: {
1964
1998
  input: number;
1965
1999
  output: number;
@@ -1969,14 +2003,14 @@ export declare const MODELS: {
1969
2003
  contextWindow: number;
1970
2004
  maxTokens: number;
1971
2005
  };
1972
- readonly "z-ai/glm-4.6": {
2006
+ readonly "anthropic/claude-haiku-4.5": {
1973
2007
  id: string;
1974
2008
  name: string;
1975
2009
  api: "openai-completions";
1976
2010
  provider: string;
1977
2011
  baseUrl: string;
1978
2012
  reasoning: true;
1979
- input: "text"[];
2013
+ input: ("image" | "text")[];
1980
2014
  cost: {
1981
2015
  input: number;
1982
2016
  output: number;
@@ -1986,14 +2020,14 @@ export declare const MODELS: {
1986
2020
  contextWindow: number;
1987
2021
  maxTokens: number;
1988
2022
  };
1989
- readonly "z-ai/glm-4.6:exacto": {
2023
+ readonly "qwen/qwen3-vl-8b-thinking": {
1990
2024
  id: string;
1991
2025
  name: string;
1992
2026
  api: "openai-completions";
1993
2027
  provider: string;
1994
2028
  baseUrl: string;
1995
2029
  reasoning: true;
1996
- input: "text"[];
2030
+ input: ("image" | "text")[];
1997
2031
  cost: {
1998
2032
  input: number;
1999
2033
  output: number;
@@ -2003,14 +2037,14 @@ export declare const MODELS: {
2003
2037
  contextWindow: number;
2004
2038
  maxTokens: number;
2005
2039
  };
2006
- readonly "deepseek/deepseek-v3.2-exp": {
2040
+ readonly "qwen/qwen3-vl-8b-instruct": {
2007
2041
  id: string;
2008
2042
  name: string;
2009
2043
  api: "openai-completions";
2010
2044
  provider: string;
2011
2045
  baseUrl: string;
2012
- reasoning: true;
2013
- input: "text"[];
2046
+ reasoning: false;
2047
+ input: ("image" | "text")[];
2014
2048
  cost: {
2015
2049
  input: number;
2016
2050
  output: number;
@@ -2020,14 +2054,14 @@ export declare const MODELS: {
2020
2054
  contextWindow: number;
2021
2055
  maxTokens: number;
2022
2056
  };
2023
- readonly "qwen/qwen3-vl-235b-a22b-thinking": {
2057
+ readonly "openai/gpt-5-image": {
2024
2058
  id: string;
2025
2059
  name: string;
2026
2060
  api: "openai-completions";
2027
2061
  provider: string;
2028
2062
  baseUrl: string;
2029
2063
  reasoning: true;
2030
- input: ("text" | "image")[];
2064
+ input: ("image" | "text")[];
2031
2065
  cost: {
2032
2066
  input: number;
2033
2067
  output: number;
@@ -2037,14 +2071,14 @@ export declare const MODELS: {
2037
2071
  contextWindow: number;
2038
2072
  maxTokens: number;
2039
2073
  };
2040
- readonly "qwen/qwen3-vl-235b-a22b-instruct": {
2074
+ readonly "inclusionai/ring-1t": {
2041
2075
  id: string;
2042
2076
  name: string;
2043
2077
  api: "openai-completions";
2044
2078
  provider: string;
2045
2079
  baseUrl: string;
2046
- reasoning: false;
2047
- input: ("text" | "image")[];
2080
+ reasoning: true;
2081
+ input: "text"[];
2048
2082
  cost: {
2049
2083
  input: number;
2050
2084
  output: number;
@@ -2054,7 +2088,7 @@ export declare const MODELS: {
2054
2088
  contextWindow: number;
2055
2089
  maxTokens: number;
2056
2090
  };
2057
- readonly "qwen/qwen3-max": {
2091
+ readonly "inclusionai/ling-1t": {
2058
2092
  id: string;
2059
2093
  name: string;
2060
2094
  api: "openai-completions";
@@ -2071,14 +2105,14 @@ export declare const MODELS: {
2071
2105
  contextWindow: number;
2072
2106
  maxTokens: number;
2073
2107
  };
2074
- readonly "qwen/qwen3-coder-plus": {
2108
+ readonly "openai/o3-deep-research": {
2075
2109
  id: string;
2076
2110
  name: string;
2077
2111
  api: "openai-completions";
2078
2112
  provider: string;
2079
2113
  baseUrl: string;
2080
- reasoning: false;
2081
- input: "text"[];
2114
+ reasoning: true;
2115
+ input: ("image" | "text")[];
2082
2116
  cost: {
2083
2117
  input: number;
2084
2118
  output: number;
@@ -2088,14 +2122,14 @@ export declare const MODELS: {
2088
2122
  contextWindow: number;
2089
2123
  maxTokens: number;
2090
2124
  };
2091
- readonly "deepseek/deepseek-v3.1-terminus": {
2125
+ readonly "openai/o4-mini-deep-research": {
2092
2126
  id: string;
2093
2127
  name: string;
2094
2128
  api: "openai-completions";
2095
2129
  provider: string;
2096
2130
  baseUrl: string;
2097
2131
  reasoning: true;
2098
- input: "text"[];
2132
+ input: ("image" | "text")[];
2099
2133
  cost: {
2100
2134
  input: number;
2101
2135
  output: number;
@@ -2105,7 +2139,7 @@ export declare const MODELS: {
2105
2139
  contextWindow: number;
2106
2140
  maxTokens: number;
2107
2141
  };
2108
- readonly "deepseek/deepseek-v3.1-terminus:exacto": {
2142
+ readonly "nvidia/llama-3.3-nemotron-super-49b-v1.5": {
2109
2143
  id: string;
2110
2144
  name: string;
2111
2145
  api: "openai-completions";
@@ -2122,14 +2156,14 @@ export declare const MODELS: {
2122
2156
  contextWindow: number;
2123
2157
  maxTokens: number;
2124
2158
  };
2125
- readonly "alibaba/tongyi-deepresearch-30b-a3b:free": {
2159
+ readonly "qwen/qwen3-vl-30b-a3b-thinking": {
2126
2160
  id: string;
2127
2161
  name: string;
2128
2162
  api: "openai-completions";
2129
2163
  provider: string;
2130
2164
  baseUrl: string;
2131
2165
  reasoning: true;
2132
- input: "text"[];
2166
+ input: ("image" | "text")[];
2133
2167
  cost: {
2134
2168
  input: number;
2135
2169
  output: number;
@@ -2139,14 +2173,14 @@ export declare const MODELS: {
2139
2173
  contextWindow: number;
2140
2174
  maxTokens: number;
2141
2175
  };
2142
- readonly "alibaba/tongyi-deepresearch-30b-a3b": {
2176
+ readonly "qwen/qwen3-vl-30b-a3b-instruct": {
2143
2177
  id: string;
2144
2178
  name: string;
2145
2179
  api: "openai-completions";
2146
2180
  provider: string;
2147
2181
  baseUrl: string;
2148
- reasoning: true;
2149
- input: "text"[];
2182
+ reasoning: false;
2183
+ input: ("image" | "text")[];
2150
2184
  cost: {
2151
2185
  input: number;
2152
2186
  output: number;
@@ -2156,14 +2190,14 @@ export declare const MODELS: {
2156
2190
  contextWindow: number;
2157
2191
  maxTokens: number;
2158
2192
  };
2159
- readonly "qwen/qwen3-coder-flash": {
2193
+ readonly "openai/gpt-5-pro": {
2160
2194
  id: string;
2161
2195
  name: string;
2162
2196
  api: "openai-completions";
2163
2197
  provider: string;
2164
2198
  baseUrl: string;
2165
- reasoning: false;
2166
- input: "text"[];
2199
+ reasoning: true;
2200
+ input: ("image" | "text")[];
2167
2201
  cost: {
2168
2202
  input: number;
2169
2203
  output: number;
@@ -2173,7 +2207,7 @@ export declare const MODELS: {
2173
2207
  contextWindow: number;
2174
2208
  maxTokens: number;
2175
2209
  };
2176
- readonly "qwen/qwen3-next-80b-a3b-thinking": {
2210
+ readonly "z-ai/glm-4.6": {
2177
2211
  id: string;
2178
2212
  name: string;
2179
2213
  api: "openai-completions";
@@ -2190,13 +2224,13 @@ export declare const MODELS: {
2190
2224
  contextWindow: number;
2191
2225
  maxTokens: number;
2192
2226
  };
2193
- readonly "qwen/qwen3-next-80b-a3b-instruct": {
2227
+ readonly "z-ai/glm-4.6:exacto": {
2194
2228
  id: string;
2195
2229
  name: string;
2196
2230
  api: "openai-completions";
2197
2231
  provider: string;
2198
2232
  baseUrl: string;
2199
- reasoning: false;
2233
+ reasoning: true;
2200
2234
  input: "text"[];
2201
2235
  cost: {
2202
2236
  input: number;
@@ -2207,14 +2241,14 @@ export declare const MODELS: {
2207
2241
  contextWindow: number;
2208
2242
  maxTokens: number;
2209
2243
  };
2210
- readonly "meituan/longcat-flash-chat:free": {
2244
+ readonly "anthropic/claude-sonnet-4.5": {
2211
2245
  id: string;
2212
2246
  name: string;
2213
2247
  api: "openai-completions";
2214
2248
  provider: string;
2215
2249
  baseUrl: string;
2216
- reasoning: false;
2217
- input: "text"[];
2250
+ reasoning: true;
2251
+ input: ("image" | "text")[];
2218
2252
  cost: {
2219
2253
  input: number;
2220
2254
  output: number;
@@ -2224,13 +2258,13 @@ export declare const MODELS: {
2224
2258
  contextWindow: number;
2225
2259
  maxTokens: number;
2226
2260
  };
2227
- readonly "qwen/qwen-plus-2025-07-28": {
2261
+ readonly "deepseek/deepseek-v3.2-exp": {
2228
2262
  id: string;
2229
2263
  name: string;
2230
2264
  api: "openai-completions";
2231
2265
  provider: string;
2232
2266
  baseUrl: string;
2233
- reasoning: false;
2267
+ reasoning: true;
2234
2268
  input: "text"[];
2235
2269
  cost: {
2236
2270
  input: number;
@@ -2241,14 +2275,14 @@ export declare const MODELS: {
2241
2275
  contextWindow: number;
2242
2276
  maxTokens: number;
2243
2277
  };
2244
- readonly "qwen/qwen-plus-2025-07-28:thinking": {
2278
+ readonly "google/gemini-2.5-flash-preview-09-2025": {
2245
2279
  id: string;
2246
2280
  name: string;
2247
2281
  api: "openai-completions";
2248
2282
  provider: string;
2249
2283
  baseUrl: string;
2250
2284
  reasoning: true;
2251
- input: "text"[];
2285
+ input: ("image" | "text")[];
2252
2286
  cost: {
2253
2287
  input: number;
2254
2288
  output: number;
@@ -2258,14 +2292,14 @@ export declare const MODELS: {
2258
2292
  contextWindow: number;
2259
2293
  maxTokens: number;
2260
2294
  };
2261
- readonly "nvidia/nemotron-nano-9b-v2:free": {
2295
+ readonly "google/gemini-2.5-flash-lite-preview-09-2025": {
2262
2296
  id: string;
2263
2297
  name: string;
2264
2298
  api: "openai-completions";
2265
2299
  provider: string;
2266
2300
  baseUrl: string;
2267
2301
  reasoning: true;
2268
- input: "text"[];
2302
+ input: ("image" | "text")[];
2269
2303
  cost: {
2270
2304
  input: number;
2271
2305
  output: number;
@@ -2275,14 +2309,14 @@ export declare const MODELS: {
2275
2309
  contextWindow: number;
2276
2310
  maxTokens: number;
2277
2311
  };
2278
- readonly "nvidia/nemotron-nano-9b-v2": {
2312
+ readonly "qwen/qwen3-vl-235b-a22b-thinking": {
2279
2313
  id: string;
2280
2314
  name: string;
2281
2315
  api: "openai-completions";
2282
2316
  provider: string;
2283
2317
  baseUrl: string;
2284
2318
  reasoning: true;
2285
- input: "text"[];
2319
+ input: ("image" | "text")[];
2286
2320
  cost: {
2287
2321
  input: number;
2288
2322
  output: number;
@@ -2292,14 +2326,14 @@ export declare const MODELS: {
2292
2326
  contextWindow: number;
2293
2327
  maxTokens: number;
2294
2328
  };
2295
- readonly "moonshotai/kimi-k2-0905": {
2329
+ readonly "qwen/qwen3-vl-235b-a22b-instruct": {
2296
2330
  id: string;
2297
2331
  name: string;
2298
2332
  api: "openai-completions";
2299
2333
  provider: string;
2300
2334
  baseUrl: string;
2301
2335
  reasoning: false;
2302
- input: "text"[];
2336
+ input: ("image" | "text")[];
2303
2337
  cost: {
2304
2338
  input: number;
2305
2339
  output: number;
@@ -2309,7 +2343,7 @@ export declare const MODELS: {
2309
2343
  contextWindow: number;
2310
2344
  maxTokens: number;
2311
2345
  };
2312
- readonly "moonshotai/kimi-k2-0905:exacto": {
2346
+ readonly "qwen/qwen3-max": {
2313
2347
  id: string;
2314
2348
  name: string;
2315
2349
  api: "openai-completions";
@@ -2326,13 +2360,13 @@ export declare const MODELS: {
2326
2360
  contextWindow: number;
2327
2361
  maxTokens: number;
2328
2362
  };
2329
- readonly "deepcogito/cogito-v2-preview-llama-70b": {
2363
+ readonly "qwen/qwen3-coder-plus": {
2330
2364
  id: string;
2331
2365
  name: string;
2332
2366
  api: "openai-completions";
2333
2367
  provider: string;
2334
2368
  baseUrl: string;
2335
- reasoning: true;
2369
+ reasoning: false;
2336
2370
  input: "text"[];
2337
2371
  cost: {
2338
2372
  input: number;
@@ -2343,14 +2377,14 @@ export declare const MODELS: {
2343
2377
  contextWindow: number;
2344
2378
  maxTokens: number;
2345
2379
  };
2346
- readonly "deepcogito/cogito-v2-preview-llama-109b-moe": {
2380
+ readonly "openai/gpt-5-codex": {
2347
2381
  id: string;
2348
2382
  name: string;
2349
2383
  api: "openai-completions";
2350
2384
  provider: string;
2351
2385
  baseUrl: string;
2352
2386
  reasoning: true;
2353
- input: ("text" | "image")[];
2387
+ input: ("image" | "text")[];
2354
2388
  cost: {
2355
2389
  input: number;
2356
2390
  output: number;
@@ -2360,14 +2394,14 @@ export declare const MODELS: {
2360
2394
  contextWindow: number;
2361
2395
  maxTokens: number;
2362
2396
  };
2363
- readonly "stepfun-ai/step3": {
2397
+ readonly "deepseek/deepseek-v3.1-terminus": {
2364
2398
  id: string;
2365
2399
  name: string;
2366
2400
  api: "openai-completions";
2367
2401
  provider: string;
2368
2402
  baseUrl: string;
2369
2403
  reasoning: true;
2370
- input: ("text" | "image")[];
2404
+ input: "text"[];
2371
2405
  cost: {
2372
2406
  input: number;
2373
2407
  output: number;
@@ -2377,7 +2411,7 @@ export declare const MODELS: {
2377
2411
  contextWindow: number;
2378
2412
  maxTokens: number;
2379
2413
  };
2380
- readonly "qwen/qwen3-30b-a3b-thinking-2507": {
2414
+ readonly "deepseek/deepseek-v3.1-terminus:exacto": {
2381
2415
  id: string;
2382
2416
  name: string;
2383
2417
  api: "openai-completions";
@@ -2394,14 +2428,14 @@ export declare const MODELS: {
2394
2428
  contextWindow: number;
2395
2429
  maxTokens: number;
2396
2430
  };
2397
- readonly "nousresearch/hermes-4-70b": {
2431
+ readonly "x-ai/grok-4-fast": {
2398
2432
  id: string;
2399
2433
  name: string;
2400
2434
  api: "openai-completions";
2401
2435
  provider: string;
2402
2436
  baseUrl: string;
2403
2437
  reasoning: true;
2404
- input: "text"[];
2438
+ input: ("image" | "text")[];
2405
2439
  cost: {
2406
2440
  input: number;
2407
2441
  output: number;
@@ -2411,7 +2445,7 @@ export declare const MODELS: {
2411
2445
  contextWindow: number;
2412
2446
  maxTokens: number;
2413
2447
  };
2414
- readonly "nousresearch/hermes-4-405b": {
2448
+ readonly "alibaba/tongyi-deepresearch-30b-a3b:free": {
2415
2449
  id: string;
2416
2450
  name: string;
2417
2451
  api: "openai-completions";
@@ -2428,7 +2462,7 @@ export declare const MODELS: {
2428
2462
  contextWindow: number;
2429
2463
  maxTokens: number;
2430
2464
  };
2431
- readonly "deepseek/deepseek-chat-v3.1": {
2465
+ readonly "alibaba/tongyi-deepresearch-30b-a3b": {
2432
2466
  id: string;
2433
2467
  name: string;
2434
2468
  api: "openai-completions";
@@ -2445,14 +2479,14 @@ export declare const MODELS: {
2445
2479
  contextWindow: number;
2446
2480
  maxTokens: number;
2447
2481
  };
2448
- readonly "mistralai/mistral-medium-3.1": {
2482
+ readonly "qwen/qwen3-coder-flash": {
2449
2483
  id: string;
2450
2484
  name: string;
2451
2485
  api: "openai-completions";
2452
2486
  provider: string;
2453
2487
  baseUrl: string;
2454
2488
  reasoning: false;
2455
- input: ("text" | "image")[];
2489
+ input: "text"[];
2456
2490
  cost: {
2457
2491
  input: number;
2458
2492
  output: number;
@@ -2462,13 +2496,13 @@ export declare const MODELS: {
2462
2496
  contextWindow: number;
2463
2497
  maxTokens: number;
2464
2498
  };
2465
- readonly "baidu/ernie-4.5-21b-a3b": {
2499
+ readonly "qwen/qwen3-next-80b-a3b-thinking": {
2466
2500
  id: string;
2467
2501
  name: string;
2468
2502
  api: "openai-completions";
2469
2503
  provider: string;
2470
2504
  baseUrl: string;
2471
- reasoning: false;
2505
+ reasoning: true;
2472
2506
  input: "text"[];
2473
2507
  cost: {
2474
2508
  input: number;
@@ -2479,14 +2513,14 @@ export declare const MODELS: {
2479
2513
  contextWindow: number;
2480
2514
  maxTokens: number;
2481
2515
  };
2482
- readonly "baidu/ernie-4.5-vl-28b-a3b": {
2516
+ readonly "qwen/qwen3-next-80b-a3b-instruct": {
2483
2517
  id: string;
2484
2518
  name: string;
2485
2519
  api: "openai-completions";
2486
2520
  provider: string;
2487
2521
  baseUrl: string;
2488
- reasoning: true;
2489
- input: ("text" | "image")[];
2522
+ reasoning: false;
2523
+ input: "text"[];
2490
2524
  cost: {
2491
2525
  input: number;
2492
2526
  output: number;
@@ -2496,14 +2530,14 @@ export declare const MODELS: {
2496
2530
  contextWindow: number;
2497
2531
  maxTokens: number;
2498
2532
  };
2499
- readonly "z-ai/glm-4.5v": {
2533
+ readonly "meituan/longcat-flash-chat:free": {
2500
2534
  id: string;
2501
2535
  name: string;
2502
2536
  api: "openai-completions";
2503
2537
  provider: string;
2504
2538
  baseUrl: string;
2505
- reasoning: true;
2506
- input: ("text" | "image")[];
2539
+ reasoning: false;
2540
+ input: "text"[];
2507
2541
  cost: {
2508
2542
  input: number;
2509
2543
  output: number;
@@ -2513,7 +2547,7 @@ export declare const MODELS: {
2513
2547
  contextWindow: number;
2514
2548
  maxTokens: number;
2515
2549
  };
2516
- readonly "ai21/jamba-mini-1.7": {
2550
+ readonly "qwen/qwen-plus-2025-07-28": {
2517
2551
  id: string;
2518
2552
  name: string;
2519
2553
  api: "openai-completions";
@@ -2530,13 +2564,13 @@ export declare const MODELS: {
2530
2564
  contextWindow: number;
2531
2565
  maxTokens: number;
2532
2566
  };
2533
- readonly "ai21/jamba-large-1.7": {
2567
+ readonly "qwen/qwen-plus-2025-07-28:thinking": {
2534
2568
  id: string;
2535
2569
  name: string;
2536
2570
  api: "openai-completions";
2537
2571
  provider: string;
2538
2572
  baseUrl: string;
2539
- reasoning: false;
2573
+ reasoning: true;
2540
2574
  input: "text"[];
2541
2575
  cost: {
2542
2576
  input: number;
@@ -2547,13 +2581,13 @@ export declare const MODELS: {
2547
2581
  contextWindow: number;
2548
2582
  maxTokens: number;
2549
2583
  };
2550
- readonly "mistralai/codestral-2508": {
2584
+ readonly "nvidia/nemotron-nano-9b-v2:free": {
2551
2585
  id: string;
2552
2586
  name: string;
2553
2587
  api: "openai-completions";
2554
2588
  provider: string;
2555
2589
  baseUrl: string;
2556
- reasoning: false;
2590
+ reasoning: true;
2557
2591
  input: "text"[];
2558
2592
  cost: {
2559
2593
  input: number;
@@ -2564,13 +2598,13 @@ export declare const MODELS: {
2564
2598
  contextWindow: number;
2565
2599
  maxTokens: number;
2566
2600
  };
2567
- readonly "qwen/qwen3-coder-30b-a3b-instruct": {
2601
+ readonly "nvidia/nemotron-nano-9b-v2": {
2568
2602
  id: string;
2569
2603
  name: string;
2570
2604
  api: "openai-completions";
2571
2605
  provider: string;
2572
2606
  baseUrl: string;
2573
- reasoning: false;
2607
+ reasoning: true;
2574
2608
  input: "text"[];
2575
2609
  cost: {
2576
2610
  input: number;
@@ -2581,7 +2615,7 @@ export declare const MODELS: {
2581
2615
  contextWindow: number;
2582
2616
  maxTokens: number;
2583
2617
  };
2584
- readonly "qwen/qwen3-30b-a3b-instruct-2507": {
2618
+ readonly "moonshotai/kimi-k2-0905": {
2585
2619
  id: string;
2586
2620
  name: string;
2587
2621
  api: "openai-completions";
@@ -2598,13 +2632,13 @@ export declare const MODELS: {
2598
2632
  contextWindow: number;
2599
2633
  maxTokens: number;
2600
2634
  };
2601
- readonly "z-ai/glm-4.5": {
2635
+ readonly "moonshotai/kimi-k2-0905:exacto": {
2602
2636
  id: string;
2603
2637
  name: string;
2604
2638
  api: "openai-completions";
2605
2639
  provider: string;
2606
2640
  baseUrl: string;
2607
- reasoning: true;
2641
+ reasoning: false;
2608
2642
  input: "text"[];
2609
2643
  cost: {
2610
2644
  input: number;
@@ -2615,7 +2649,7 @@ export declare const MODELS: {
2615
2649
  contextWindow: number;
2616
2650
  maxTokens: number;
2617
2651
  };
2618
- readonly "z-ai/glm-4.5-air:free": {
2652
+ readonly "deepcogito/cogito-v2-preview-llama-70b": {
2619
2653
  id: string;
2620
2654
  name: string;
2621
2655
  api: "openai-completions";
@@ -2632,14 +2666,14 @@ export declare const MODELS: {
2632
2666
  contextWindow: number;
2633
2667
  maxTokens: number;
2634
2668
  };
2635
- readonly "z-ai/glm-4.5-air": {
2669
+ readonly "deepcogito/cogito-v2-preview-llama-109b-moe": {
2636
2670
  id: string;
2637
2671
  name: string;
2638
2672
  api: "openai-completions";
2639
2673
  provider: string;
2640
2674
  baseUrl: string;
2641
2675
  reasoning: true;
2642
- input: "text"[];
2676
+ input: ("image" | "text")[];
2643
2677
  cost: {
2644
2678
  input: number;
2645
2679
  output: number;
@@ -2649,14 +2683,14 @@ export declare const MODELS: {
2649
2683
  contextWindow: number;
2650
2684
  maxTokens: number;
2651
2685
  };
2652
- readonly "qwen/qwen3-235b-a22b-thinking-2507": {
2686
+ readonly "stepfun-ai/step3": {
2653
2687
  id: string;
2654
2688
  name: string;
2655
2689
  api: "openai-completions";
2656
2690
  provider: string;
2657
2691
  baseUrl: string;
2658
2692
  reasoning: true;
2659
- input: "text"[];
2693
+ input: ("image" | "text")[];
2660
2694
  cost: {
2661
2695
  input: number;
2662
2696
  output: number;
@@ -2666,13 +2700,13 @@ export declare const MODELS: {
2666
2700
  contextWindow: number;
2667
2701
  maxTokens: number;
2668
2702
  };
2669
- readonly "z-ai/glm-4-32b": {
2703
+ readonly "qwen/qwen3-30b-a3b-thinking-2507": {
2670
2704
  id: string;
2671
2705
  name: string;
2672
2706
  api: "openai-completions";
2673
2707
  provider: string;
2674
2708
  baseUrl: string;
2675
- reasoning: false;
2709
+ reasoning: true;
2676
2710
  input: "text"[];
2677
2711
  cost: {
2678
2712
  input: number;
@@ -2683,13 +2717,13 @@ export declare const MODELS: {
2683
2717
  contextWindow: number;
2684
2718
  maxTokens: number;
2685
2719
  };
2686
- readonly "qwen/qwen3-coder:free": {
2720
+ readonly "x-ai/grok-code-fast-1": {
2687
2721
  id: string;
2688
2722
  name: string;
2689
2723
  api: "openai-completions";
2690
2724
  provider: string;
2691
2725
  baseUrl: string;
2692
- reasoning: false;
2726
+ reasoning: true;
2693
2727
  input: "text"[];
2694
2728
  cost: {
2695
2729
  input: number;
@@ -2700,7 +2734,7 @@ export declare const MODELS: {
2700
2734
  contextWindow: number;
2701
2735
  maxTokens: number;
2702
2736
  };
2703
- readonly "qwen/qwen3-coder": {
2737
+ readonly "nousresearch/hermes-4-70b": {
2704
2738
  id: string;
2705
2739
  name: string;
2706
2740
  api: "openai-completions";
@@ -2717,7 +2751,7 @@ export declare const MODELS: {
2717
2751
  contextWindow: number;
2718
2752
  maxTokens: number;
2719
2753
  };
2720
- readonly "qwen/qwen3-coder:exacto": {
2754
+ readonly "nousresearch/hermes-4-405b": {
2721
2755
  id: string;
2722
2756
  name: string;
2723
2757
  api: "openai-completions";
@@ -2734,7 +2768,7 @@ export declare const MODELS: {
2734
2768
  contextWindow: number;
2735
2769
  maxTokens: number;
2736
2770
  };
2737
- readonly "qwen/qwen3-235b-a22b-2507": {
2771
+ readonly "deepseek/deepseek-chat-v3.1": {
2738
2772
  id: string;
2739
2773
  name: string;
2740
2774
  api: "openai-completions";
@@ -2751,7 +2785,7 @@ export declare const MODELS: {
2751
2785
  contextWindow: number;
2752
2786
  maxTokens: number;
2753
2787
  };
2754
- readonly "moonshotai/kimi-k2": {
2788
+ readonly "openai/gpt-4o-audio-preview": {
2755
2789
  id: string;
2756
2790
  name: string;
2757
2791
  api: "openai-completions";
@@ -2768,14 +2802,14 @@ export declare const MODELS: {
2768
2802
  contextWindow: number;
2769
2803
  maxTokens: number;
2770
2804
  };
2771
- readonly "mistralai/devstral-medium": {
2805
+ readonly "mistralai/mistral-medium-3.1": {
2772
2806
  id: string;
2773
2807
  name: string;
2774
2808
  api: "openai-completions";
2775
2809
  provider: string;
2776
2810
  baseUrl: string;
2777
2811
  reasoning: false;
2778
- input: "text"[];
2812
+ input: ("image" | "text")[];
2779
2813
  cost: {
2780
2814
  input: number;
2781
2815
  output: number;
@@ -2785,7 +2819,7 @@ export declare const MODELS: {
2785
2819
  contextWindow: number;
2786
2820
  maxTokens: number;
2787
2821
  };
2788
- readonly "mistralai/devstral-small": {
2822
+ readonly "baidu/ernie-4.5-21b-a3b": {
2789
2823
  id: string;
2790
2824
  name: string;
2791
2825
  api: "openai-completions";
@@ -2802,14 +2836,14 @@ export declare const MODELS: {
2802
2836
  contextWindow: number;
2803
2837
  maxTokens: number;
2804
2838
  };
2805
- readonly "tngtech/deepseek-r1t2-chimera": {
2839
+ readonly "baidu/ernie-4.5-vl-28b-a3b": {
2806
2840
  id: string;
2807
2841
  name: string;
2808
2842
  api: "openai-completions";
2809
2843
  provider: string;
2810
2844
  baseUrl: string;
2811
2845
  reasoning: true;
2812
- input: "text"[];
2846
+ input: ("image" | "text")[];
2813
2847
  cost: {
2814
2848
  input: number;
2815
2849
  output: number;
@@ -2819,14 +2853,14 @@ export declare const MODELS: {
2819
2853
  contextWindow: number;
2820
2854
  maxTokens: number;
2821
2855
  };
2822
- readonly "inception/mercury": {
2856
+ readonly "z-ai/glm-4.5v": {
2823
2857
  id: string;
2824
2858
  name: string;
2825
2859
  api: "openai-completions";
2826
2860
  provider: string;
2827
2861
  baseUrl: string;
2828
- reasoning: false;
2829
- input: "text"[];
2862
+ reasoning: true;
2863
+ input: ("image" | "text")[];
2830
2864
  cost: {
2831
2865
  input: number;
2832
2866
  output: number;
@@ -2836,14 +2870,14 @@ export declare const MODELS: {
2836
2870
  contextWindow: number;
2837
2871
  maxTokens: number;
2838
2872
  };
2839
- readonly "mistralai/mistral-small-3.2-24b-instruct:free": {
2873
+ readonly "ai21/jamba-mini-1.7": {
2840
2874
  id: string;
2841
2875
  name: string;
2842
2876
  api: "openai-completions";
2843
2877
  provider: string;
2844
2878
  baseUrl: string;
2845
2879
  reasoning: false;
2846
- input: ("text" | "image")[];
2880
+ input: "text"[];
2847
2881
  cost: {
2848
2882
  input: number;
2849
2883
  output: number;
@@ -2853,14 +2887,14 @@ export declare const MODELS: {
2853
2887
  contextWindow: number;
2854
2888
  maxTokens: number;
2855
2889
  };
2856
- readonly "mistralai/mistral-small-3.2-24b-instruct": {
2890
+ readonly "ai21/jamba-large-1.7": {
2857
2891
  id: string;
2858
2892
  name: string;
2859
2893
  api: "openai-completions";
2860
2894
  provider: string;
2861
2895
  baseUrl: string;
2862
2896
  reasoning: false;
2863
- input: ("text" | "image")[];
2897
+ input: "text"[];
2864
2898
  cost: {
2865
2899
  input: number;
2866
2900
  output: number;
@@ -2870,14 +2904,14 @@ export declare const MODELS: {
2870
2904
  contextWindow: number;
2871
2905
  maxTokens: number;
2872
2906
  };
2873
- readonly "minimax/minimax-m1": {
2907
+ readonly "openai/gpt-5": {
2874
2908
  id: string;
2875
2909
  name: string;
2876
2910
  api: "openai-completions";
2877
2911
  provider: string;
2878
2912
  baseUrl: string;
2879
2913
  reasoning: true;
2880
- input: "text"[];
2914
+ input: ("image" | "text")[];
2881
2915
  cost: {
2882
2916
  input: number;
2883
2917
  output: number;
@@ -2887,14 +2921,14 @@ export declare const MODELS: {
2887
2921
  contextWindow: number;
2888
2922
  maxTokens: number;
2889
2923
  };
2890
- readonly "mistralai/magistral-small-2506": {
2924
+ readonly "openai/gpt-5-mini": {
2891
2925
  id: string;
2892
2926
  name: string;
2893
2927
  api: "openai-completions";
2894
2928
  provider: string;
2895
2929
  baseUrl: string;
2896
2930
  reasoning: true;
2897
- input: "text"[];
2931
+ input: ("image" | "text")[];
2898
2932
  cost: {
2899
2933
  input: number;
2900
2934
  output: number;
@@ -2904,14 +2938,14 @@ export declare const MODELS: {
2904
2938
  contextWindow: number;
2905
2939
  maxTokens: number;
2906
2940
  };
2907
- readonly "mistralai/magistral-medium-2506:thinking": {
2941
+ readonly "openai/gpt-5-nano": {
2908
2942
  id: string;
2909
2943
  name: string;
2910
2944
  api: "openai-completions";
2911
2945
  provider: string;
2912
2946
  baseUrl: string;
2913
2947
  reasoning: true;
2914
- input: "text"[];
2948
+ input: ("image" | "text")[];
2915
2949
  cost: {
2916
2950
  input: number;
2917
2951
  output: number;
@@ -2921,7 +2955,7 @@ export declare const MODELS: {
2921
2955
  contextWindow: number;
2922
2956
  maxTokens: number;
2923
2957
  };
2924
- readonly "mistralai/magistral-medium-2506": {
2958
+ readonly "openai/gpt-oss-120b": {
2925
2959
  id: string;
2926
2960
  name: string;
2927
2961
  api: "openai-completions";
@@ -2938,7 +2972,7 @@ export declare const MODELS: {
2938
2972
  contextWindow: number;
2939
2973
  maxTokens: number;
2940
2974
  };
2941
- readonly "deepseek/deepseek-r1-0528": {
2975
+ readonly "openai/gpt-oss-120b:exacto": {
2942
2976
  id: string;
2943
2977
  name: string;
2944
2978
  api: "openai-completions";
@@ -2955,13 +2989,13 @@ export declare const MODELS: {
2955
2989
  contextWindow: number;
2956
2990
  maxTokens: number;
2957
2991
  };
2958
- readonly "mistralai/devstral-small-2505:free": {
2992
+ readonly "openai/gpt-oss-20b:free": {
2959
2993
  id: string;
2960
2994
  name: string;
2961
2995
  api: "openai-completions";
2962
2996
  provider: string;
2963
2997
  baseUrl: string;
2964
- reasoning: false;
2998
+ reasoning: true;
2965
2999
  input: "text"[];
2966
3000
  cost: {
2967
3001
  input: number;
@@ -2972,13 +3006,13 @@ export declare const MODELS: {
2972
3006
  contextWindow: number;
2973
3007
  maxTokens: number;
2974
3008
  };
2975
- readonly "mistralai/devstral-small-2505": {
3009
+ readonly "openai/gpt-oss-20b": {
2976
3010
  id: string;
2977
3011
  name: string;
2978
3012
  api: "openai-completions";
2979
3013
  provider: string;
2980
3014
  baseUrl: string;
2981
- reasoning: false;
3015
+ reasoning: true;
2982
3016
  input: "text"[];
2983
3017
  cost: {
2984
3018
  input: number;
@@ -2989,14 +3023,14 @@ export declare const MODELS: {
2989
3023
  contextWindow: number;
2990
3024
  maxTokens: number;
2991
3025
  };
2992
- readonly "meta-llama/llama-3.3-8b-instruct:free": {
3026
+ readonly "anthropic/claude-opus-4.1": {
2993
3027
  id: string;
2994
3028
  name: string;
2995
3029
  api: "openai-completions";
2996
3030
  provider: string;
2997
3031
  baseUrl: string;
2998
- reasoning: false;
2999
- input: "text"[];
3032
+ reasoning: true;
3033
+ input: ("image" | "text")[];
3000
3034
  cost: {
3001
3035
  input: number;
3002
3036
  output: number;
@@ -3006,13 +3040,13 @@ export declare const MODELS: {
3006
3040
  contextWindow: number;
3007
3041
  maxTokens: number;
3008
3042
  };
3009
- readonly "nousresearch/deephermes-3-mistral-24b-preview": {
3043
+ readonly "mistralai/codestral-2508": {
3010
3044
  id: string;
3011
3045
  name: string;
3012
3046
  api: "openai-completions";
3013
3047
  provider: string;
3014
3048
  baseUrl: string;
3015
- reasoning: true;
3049
+ reasoning: false;
3016
3050
  input: "text"[];
3017
3051
  cost: {
3018
3052
  input: number;
@@ -3023,14 +3057,14 @@ export declare const MODELS: {
3023
3057
  contextWindow: number;
3024
3058
  maxTokens: number;
3025
3059
  };
3026
- readonly "mistralai/mistral-medium-3": {
3060
+ readonly "qwen/qwen3-coder-30b-a3b-instruct": {
3027
3061
  id: string;
3028
3062
  name: string;
3029
3063
  api: "openai-completions";
3030
3064
  provider: string;
3031
3065
  baseUrl: string;
3032
3066
  reasoning: false;
3033
- input: ("text" | "image")[];
3067
+ input: "text"[];
3034
3068
  cost: {
3035
3069
  input: number;
3036
3070
  output: number;
@@ -3040,7 +3074,7 @@ export declare const MODELS: {
3040
3074
  contextWindow: number;
3041
3075
  maxTokens: number;
3042
3076
  };
3043
- readonly "arcee-ai/virtuoso-large": {
3077
+ readonly "qwen/qwen3-30b-a3b-instruct-2507": {
3044
3078
  id: string;
3045
3079
  name: string;
3046
3080
  api: "openai-completions";
@@ -3057,13 +3091,13 @@ export declare const MODELS: {
3057
3091
  contextWindow: number;
3058
3092
  maxTokens: number;
3059
3093
  };
3060
- readonly "inception/mercury-coder": {
3094
+ readonly "z-ai/glm-4.5": {
3061
3095
  id: string;
3062
3096
  name: string;
3063
3097
  api: "openai-completions";
3064
3098
  provider: string;
3065
3099
  baseUrl: string;
3066
- reasoning: false;
3100
+ reasoning: true;
3067
3101
  input: "text"[];
3068
3102
  cost: {
3069
3103
  input: number;
@@ -3074,7 +3108,7 @@ export declare const MODELS: {
3074
3108
  contextWindow: number;
3075
3109
  maxTokens: number;
3076
3110
  };
3077
- readonly "qwen/qwen3-4b:free": {
3111
+ readonly "z-ai/glm-4.5-air:free": {
3078
3112
  id: string;
3079
3113
  name: string;
3080
3114
  api: "openai-completions";
@@ -3091,7 +3125,7 @@ export declare const MODELS: {
3091
3125
  contextWindow: number;
3092
3126
  maxTokens: number;
3093
3127
  };
3094
- readonly "qwen/qwen3-30b-a3b": {
3128
+ readonly "z-ai/glm-4.5-air": {
3095
3129
  id: string;
3096
3130
  name: string;
3097
3131
  api: "openai-completions";
@@ -3108,7 +3142,7 @@ export declare const MODELS: {
3108
3142
  contextWindow: number;
3109
3143
  maxTokens: number;
3110
3144
  };
3111
- readonly "qwen/qwen3-8b": {
3145
+ readonly "qwen/qwen3-235b-a22b-thinking-2507": {
3112
3146
  id: string;
3113
3147
  name: string;
3114
3148
  api: "openai-completions";
@@ -3125,13 +3159,13 @@ export declare const MODELS: {
3125
3159
  contextWindow: number;
3126
3160
  maxTokens: number;
3127
3161
  };
3128
- readonly "qwen/qwen3-14b": {
3162
+ readonly "z-ai/glm-4-32b": {
3129
3163
  id: string;
3130
3164
  name: string;
3131
3165
  api: "openai-completions";
3132
3166
  provider: string;
3133
3167
  baseUrl: string;
3134
- reasoning: true;
3168
+ reasoning: false;
3135
3169
  input: "text"[];
3136
3170
  cost: {
3137
3171
  input: number;
@@ -3142,13 +3176,1169 @@ export declare const MODELS: {
3142
3176
  contextWindow: number;
3143
3177
  maxTokens: number;
3144
3178
  };
3145
- readonly "qwen/qwen3-32b": {
3179
+ readonly "qwen/qwen3-coder:free": {
3146
3180
  id: string;
3147
3181
  name: string;
3148
3182
  api: "openai-completions";
3149
3183
  provider: string;
3150
3184
  baseUrl: string;
3151
- reasoning: true;
3185
+ reasoning: false;
3186
+ input: "text"[];
3187
+ cost: {
3188
+ input: number;
3189
+ output: number;
3190
+ cacheRead: number;
3191
+ cacheWrite: number;
3192
+ };
3193
+ contextWindow: number;
3194
+ maxTokens: number;
3195
+ };
3196
+ readonly "qwen/qwen3-coder": {
3197
+ id: string;
3198
+ name: string;
3199
+ api: "openai-completions";
3200
+ provider: string;
3201
+ baseUrl: string;
3202
+ reasoning: true;
3203
+ input: "text"[];
3204
+ cost: {
3205
+ input: number;
3206
+ output: number;
3207
+ cacheRead: number;
3208
+ cacheWrite: number;
3209
+ };
3210
+ contextWindow: number;
3211
+ maxTokens: number;
3212
+ };
3213
+ readonly "qwen/qwen3-coder:exacto": {
3214
+ id: string;
3215
+ name: string;
3216
+ api: "openai-completions";
3217
+ provider: string;
3218
+ baseUrl: string;
3219
+ reasoning: true;
3220
+ input: "text"[];
3221
+ cost: {
3222
+ input: number;
3223
+ output: number;
3224
+ cacheRead: number;
3225
+ cacheWrite: number;
3226
+ };
3227
+ contextWindow: number;
3228
+ maxTokens: number;
3229
+ };
3230
+ readonly "google/gemini-2.5-flash-lite": {
3231
+ id: string;
3232
+ name: string;
3233
+ api: "openai-completions";
3234
+ provider: string;
3235
+ baseUrl: string;
3236
+ reasoning: true;
3237
+ input: ("image" | "text")[];
3238
+ cost: {
3239
+ input: number;
3240
+ output: number;
3241
+ cacheRead: number;
3242
+ cacheWrite: number;
3243
+ };
3244
+ contextWindow: number;
3245
+ maxTokens: number;
3246
+ };
3247
+ readonly "qwen/qwen3-235b-a22b-2507": {
3248
+ id: string;
3249
+ name: string;
3250
+ api: "openai-completions";
3251
+ provider: string;
3252
+ baseUrl: string;
3253
+ reasoning: true;
3254
+ input: "text"[];
3255
+ cost: {
3256
+ input: number;
3257
+ output: number;
3258
+ cacheRead: number;
3259
+ cacheWrite: number;
3260
+ };
3261
+ contextWindow: number;
3262
+ maxTokens: number;
3263
+ };
3264
+ readonly "moonshotai/kimi-k2": {
3265
+ id: string;
3266
+ name: string;
3267
+ api: "openai-completions";
3268
+ provider: string;
3269
+ baseUrl: string;
3270
+ reasoning: false;
3271
+ input: "text"[];
3272
+ cost: {
3273
+ input: number;
3274
+ output: number;
3275
+ cacheRead: number;
3276
+ cacheWrite: number;
3277
+ };
3278
+ contextWindow: number;
3279
+ maxTokens: number;
3280
+ };
3281
+ readonly "mistralai/devstral-medium": {
3282
+ id: string;
3283
+ name: string;
3284
+ api: "openai-completions";
3285
+ provider: string;
3286
+ baseUrl: string;
3287
+ reasoning: false;
3288
+ input: "text"[];
3289
+ cost: {
3290
+ input: number;
3291
+ output: number;
3292
+ cacheRead: number;
3293
+ cacheWrite: number;
3294
+ };
3295
+ contextWindow: number;
3296
+ maxTokens: number;
3297
+ };
3298
+ readonly "mistralai/devstral-small": {
3299
+ id: string;
3300
+ name: string;
3301
+ api: "openai-completions";
3302
+ provider: string;
3303
+ baseUrl: string;
3304
+ reasoning: false;
3305
+ input: "text"[];
3306
+ cost: {
3307
+ input: number;
3308
+ output: number;
3309
+ cacheRead: number;
3310
+ cacheWrite: number;
3311
+ };
3312
+ contextWindow: number;
3313
+ maxTokens: number;
3314
+ };
3315
+ readonly "x-ai/grok-4": {
3316
+ id: string;
3317
+ name: string;
3318
+ api: "openai-completions";
3319
+ provider: string;
3320
+ baseUrl: string;
3321
+ reasoning: true;
3322
+ input: ("image" | "text")[];
3323
+ cost: {
3324
+ input: number;
3325
+ output: number;
3326
+ cacheRead: number;
3327
+ cacheWrite: number;
3328
+ };
3329
+ contextWindow: number;
3330
+ maxTokens: number;
3331
+ };
3332
+ readonly "tngtech/deepseek-r1t2-chimera": {
3333
+ id: string;
3334
+ name: string;
3335
+ api: "openai-completions";
3336
+ provider: string;
3337
+ baseUrl: string;
3338
+ reasoning: true;
3339
+ input: "text"[];
3340
+ cost: {
3341
+ input: number;
3342
+ output: number;
3343
+ cacheRead: number;
3344
+ cacheWrite: number;
3345
+ };
3346
+ contextWindow: number;
3347
+ maxTokens: number;
3348
+ };
3349
+ readonly "inception/mercury": {
3350
+ id: string;
3351
+ name: string;
3352
+ api: "openai-completions";
3353
+ provider: string;
3354
+ baseUrl: string;
3355
+ reasoning: false;
3356
+ input: "text"[];
3357
+ cost: {
3358
+ input: number;
3359
+ output: number;
3360
+ cacheRead: number;
3361
+ cacheWrite: number;
3362
+ };
3363
+ contextWindow: number;
3364
+ maxTokens: number;
3365
+ };
3366
+ readonly "mistralai/mistral-small-3.2-24b-instruct:free": {
3367
+ id: string;
3368
+ name: string;
3369
+ api: "openai-completions";
3370
+ provider: string;
3371
+ baseUrl: string;
3372
+ reasoning: false;
3373
+ input: ("image" | "text")[];
3374
+ cost: {
3375
+ input: number;
3376
+ output: number;
3377
+ cacheRead: number;
3378
+ cacheWrite: number;
3379
+ };
3380
+ contextWindow: number;
3381
+ maxTokens: number;
3382
+ };
3383
+ readonly "mistralai/mistral-small-3.2-24b-instruct": {
3384
+ id: string;
3385
+ name: string;
3386
+ api: "openai-completions";
3387
+ provider: string;
3388
+ baseUrl: string;
3389
+ reasoning: false;
3390
+ input: ("image" | "text")[];
3391
+ cost: {
3392
+ input: number;
3393
+ output: number;
3394
+ cacheRead: number;
3395
+ cacheWrite: number;
3396
+ };
3397
+ contextWindow: number;
3398
+ maxTokens: number;
3399
+ };
3400
+ readonly "minimax/minimax-m1": {
3401
+ id: string;
3402
+ name: string;
3403
+ api: "openai-completions";
3404
+ provider: string;
3405
+ baseUrl: string;
3406
+ reasoning: true;
3407
+ input: "text"[];
3408
+ cost: {
3409
+ input: number;
3410
+ output: number;
3411
+ cacheRead: number;
3412
+ cacheWrite: number;
3413
+ };
3414
+ contextWindow: number;
3415
+ maxTokens: number;
3416
+ };
3417
+ readonly "google/gemini-2.5-flash-lite-preview-06-17": {
3418
+ id: string;
3419
+ name: string;
3420
+ api: "openai-completions";
3421
+ provider: string;
3422
+ baseUrl: string;
3423
+ reasoning: true;
3424
+ input: ("image" | "text")[];
3425
+ cost: {
3426
+ input: number;
3427
+ output: number;
3428
+ cacheRead: number;
3429
+ cacheWrite: number;
3430
+ };
3431
+ contextWindow: number;
3432
+ maxTokens: number;
3433
+ };
3434
+ readonly "google/gemini-2.5-flash": {
3435
+ id: string;
3436
+ name: string;
3437
+ api: "openai-completions";
3438
+ provider: string;
3439
+ baseUrl: string;
3440
+ reasoning: true;
3441
+ input: ("image" | "text")[];
3442
+ cost: {
3443
+ input: number;
3444
+ output: number;
3445
+ cacheRead: number;
3446
+ cacheWrite: number;
3447
+ };
3448
+ contextWindow: number;
3449
+ maxTokens: number;
3450
+ };
3451
+ readonly "google/gemini-2.5-pro": {
3452
+ id: string;
3453
+ name: string;
3454
+ api: "openai-completions";
3455
+ provider: string;
3456
+ baseUrl: string;
3457
+ reasoning: true;
3458
+ input: ("image" | "text")[];
3459
+ cost: {
3460
+ input: number;
3461
+ output: number;
3462
+ cacheRead: number;
3463
+ cacheWrite: number;
3464
+ };
3465
+ contextWindow: number;
3466
+ maxTokens: number;
3467
+ };
3468
+ readonly "openai/o3-pro": {
3469
+ id: string;
3470
+ name: string;
3471
+ api: "openai-completions";
3472
+ provider: string;
3473
+ baseUrl: string;
3474
+ reasoning: true;
3475
+ input: ("image" | "text")[];
3476
+ cost: {
3477
+ input: number;
3478
+ output: number;
3479
+ cacheRead: number;
3480
+ cacheWrite: number;
3481
+ };
3482
+ contextWindow: number;
3483
+ maxTokens: number;
3484
+ };
3485
+ readonly "x-ai/grok-3-mini": {
3486
+ id: string;
3487
+ name: string;
3488
+ api: "openai-completions";
3489
+ provider: string;
3490
+ baseUrl: string;
3491
+ reasoning: true;
3492
+ input: "text"[];
3493
+ cost: {
3494
+ input: number;
3495
+ output: number;
3496
+ cacheRead: number;
3497
+ cacheWrite: number;
3498
+ };
3499
+ contextWindow: number;
3500
+ maxTokens: number;
3501
+ };
3502
+ readonly "x-ai/grok-3": {
3503
+ id: string;
3504
+ name: string;
3505
+ api: "openai-completions";
3506
+ provider: string;
3507
+ baseUrl: string;
3508
+ reasoning: false;
3509
+ input: "text"[];
3510
+ cost: {
3511
+ input: number;
3512
+ output: number;
3513
+ cacheRead: number;
3514
+ cacheWrite: number;
3515
+ };
3516
+ contextWindow: number;
3517
+ maxTokens: number;
3518
+ };
3519
+ readonly "mistralai/magistral-small-2506": {
3520
+ id: string;
3521
+ name: string;
3522
+ api: "openai-completions";
3523
+ provider: string;
3524
+ baseUrl: string;
3525
+ reasoning: true;
3526
+ input: "text"[];
3527
+ cost: {
3528
+ input: number;
3529
+ output: number;
3530
+ cacheRead: number;
3531
+ cacheWrite: number;
3532
+ };
3533
+ contextWindow: number;
3534
+ maxTokens: number;
3535
+ };
3536
+ readonly "mistralai/magistral-medium-2506:thinking": {
3537
+ id: string;
3538
+ name: string;
3539
+ api: "openai-completions";
3540
+ provider: string;
3541
+ baseUrl: string;
3542
+ reasoning: true;
3543
+ input: "text"[];
3544
+ cost: {
3545
+ input: number;
3546
+ output: number;
3547
+ cacheRead: number;
3548
+ cacheWrite: number;
3549
+ };
3550
+ contextWindow: number;
3551
+ maxTokens: number;
3552
+ };
3553
+ readonly "mistralai/magistral-medium-2506": {
3554
+ id: string;
3555
+ name: string;
3556
+ api: "openai-completions";
3557
+ provider: string;
3558
+ baseUrl: string;
3559
+ reasoning: true;
3560
+ input: "text"[];
3561
+ cost: {
3562
+ input: number;
3563
+ output: number;
3564
+ cacheRead: number;
3565
+ cacheWrite: number;
3566
+ };
3567
+ contextWindow: number;
3568
+ maxTokens: number;
3569
+ };
3570
+ readonly "google/gemini-2.5-pro-preview": {
3571
+ id: string;
3572
+ name: string;
3573
+ api: "openai-completions";
3574
+ provider: string;
3575
+ baseUrl: string;
3576
+ reasoning: true;
3577
+ input: ("image" | "text")[];
3578
+ cost: {
3579
+ input: number;
3580
+ output: number;
3581
+ cacheRead: number;
3582
+ cacheWrite: number;
3583
+ };
3584
+ contextWindow: number;
3585
+ maxTokens: number;
3586
+ };
3587
+ readonly "deepseek/deepseek-r1-0528": {
3588
+ id: string;
3589
+ name: string;
3590
+ api: "openai-completions";
3591
+ provider: string;
3592
+ baseUrl: string;
3593
+ reasoning: true;
3594
+ input: "text"[];
3595
+ cost: {
3596
+ input: number;
3597
+ output: number;
3598
+ cacheRead: number;
3599
+ cacheWrite: number;
3600
+ };
3601
+ contextWindow: number;
3602
+ maxTokens: number;
3603
+ };
3604
+ readonly "anthropic/claude-opus-4": {
3605
+ id: string;
3606
+ name: string;
3607
+ api: "openai-completions";
3608
+ provider: string;
3609
+ baseUrl: string;
3610
+ reasoning: true;
3611
+ input: ("image" | "text")[];
3612
+ cost: {
3613
+ input: number;
3614
+ output: number;
3615
+ cacheRead: number;
3616
+ cacheWrite: number;
3617
+ };
3618
+ contextWindow: number;
3619
+ maxTokens: number;
3620
+ };
3621
+ readonly "anthropic/claude-sonnet-4": {
3622
+ id: string;
3623
+ name: string;
3624
+ api: "openai-completions";
3625
+ provider: string;
3626
+ baseUrl: string;
3627
+ reasoning: true;
3628
+ input: ("image" | "text")[];
3629
+ cost: {
3630
+ input: number;
3631
+ output: number;
3632
+ cacheRead: number;
3633
+ cacheWrite: number;
3634
+ };
3635
+ contextWindow: number;
3636
+ maxTokens: number;
3637
+ };
3638
+ readonly "mistralai/devstral-small-2505": {
3639
+ id: string;
3640
+ name: string;
3641
+ api: "openai-completions";
3642
+ provider: string;
3643
+ baseUrl: string;
3644
+ reasoning: false;
3645
+ input: "text"[];
3646
+ cost: {
3647
+ input: number;
3648
+ output: number;
3649
+ cacheRead: number;
3650
+ cacheWrite: number;
3651
+ };
3652
+ contextWindow: number;
3653
+ maxTokens: number;
3654
+ };
3655
+ readonly "openai/codex-mini": {
3656
+ id: string;
3657
+ name: string;
3658
+ api: "openai-completions";
3659
+ provider: string;
3660
+ baseUrl: string;
3661
+ reasoning: true;
3662
+ input: ("image" | "text")[];
3663
+ cost: {
3664
+ input: number;
3665
+ output: number;
3666
+ cacheRead: number;
3667
+ cacheWrite: number;
3668
+ };
3669
+ contextWindow: number;
3670
+ maxTokens: number;
3671
+ };
3672
+ readonly "meta-llama/llama-3.3-8b-instruct:free": {
3673
+ id: string;
3674
+ name: string;
3675
+ api: "openai-completions";
3676
+ provider: string;
3677
+ baseUrl: string;
3678
+ reasoning: false;
3679
+ input: "text"[];
3680
+ cost: {
3681
+ input: number;
3682
+ output: number;
3683
+ cacheRead: number;
3684
+ cacheWrite: number;
3685
+ };
3686
+ contextWindow: number;
3687
+ maxTokens: number;
3688
+ };
3689
+ readonly "nousresearch/deephermes-3-mistral-24b-preview": {
3690
+ id: string;
3691
+ name: string;
3692
+ api: "openai-completions";
3693
+ provider: string;
3694
+ baseUrl: string;
3695
+ reasoning: true;
3696
+ input: "text"[];
3697
+ cost: {
3698
+ input: number;
3699
+ output: number;
3700
+ cacheRead: number;
3701
+ cacheWrite: number;
3702
+ };
3703
+ contextWindow: number;
3704
+ maxTokens: number;
3705
+ };
3706
+ readonly "mistralai/mistral-medium-3": {
3707
+ id: string;
3708
+ name: string;
3709
+ api: "openai-completions";
3710
+ provider: string;
3711
+ baseUrl: string;
3712
+ reasoning: false;
3713
+ input: ("image" | "text")[];
3714
+ cost: {
3715
+ input: number;
3716
+ output: number;
3717
+ cacheRead: number;
3718
+ cacheWrite: number;
3719
+ };
3720
+ contextWindow: number;
3721
+ maxTokens: number;
3722
+ };
3723
+ readonly "google/gemini-2.5-pro-preview-05-06": {
3724
+ id: string;
3725
+ name: string;
3726
+ api: "openai-completions";
3727
+ provider: string;
3728
+ baseUrl: string;
3729
+ reasoning: true;
3730
+ input: ("image" | "text")[];
3731
+ cost: {
3732
+ input: number;
3733
+ output: number;
3734
+ cacheRead: number;
3735
+ cacheWrite: number;
3736
+ };
3737
+ contextWindow: number;
3738
+ maxTokens: number;
3739
+ };
3740
+ readonly "arcee-ai/virtuoso-large": {
3741
+ id: string;
3742
+ name: string;
3743
+ api: "openai-completions";
3744
+ provider: string;
3745
+ baseUrl: string;
3746
+ reasoning: false;
3747
+ input: "text"[];
3748
+ cost: {
3749
+ input: number;
3750
+ output: number;
3751
+ cacheRead: number;
3752
+ cacheWrite: number;
3753
+ };
3754
+ contextWindow: number;
3755
+ maxTokens: number;
3756
+ };
3757
+ readonly "inception/mercury-coder": {
3758
+ id: string;
3759
+ name: string;
3760
+ api: "openai-completions";
3761
+ provider: string;
3762
+ baseUrl: string;
3763
+ reasoning: false;
3764
+ input: "text"[];
3765
+ cost: {
3766
+ input: number;
3767
+ output: number;
3768
+ cacheRead: number;
3769
+ cacheWrite: number;
3770
+ };
3771
+ contextWindow: number;
3772
+ maxTokens: number;
3773
+ };
3774
+ readonly "qwen/qwen3-4b:free": {
3775
+ id: string;
3776
+ name: string;
3777
+ api: "openai-completions";
3778
+ provider: string;
3779
+ baseUrl: string;
3780
+ reasoning: true;
3781
+ input: "text"[];
3782
+ cost: {
3783
+ input: number;
3784
+ output: number;
3785
+ cacheRead: number;
3786
+ cacheWrite: number;
3787
+ };
3788
+ contextWindow: number;
3789
+ maxTokens: number;
3790
+ };
3791
+ readonly "qwen/qwen3-30b-a3b": {
3792
+ id: string;
3793
+ name: string;
3794
+ api: "openai-completions";
3795
+ provider: string;
3796
+ baseUrl: string;
3797
+ reasoning: true;
3798
+ input: "text"[];
3799
+ cost: {
3800
+ input: number;
3801
+ output: number;
3802
+ cacheRead: number;
3803
+ cacheWrite: number;
3804
+ };
3805
+ contextWindow: number;
3806
+ maxTokens: number;
3807
+ };
3808
+ readonly "qwen/qwen3-8b": {
3809
+ id: string;
3810
+ name: string;
3811
+ api: "openai-completions";
3812
+ provider: string;
3813
+ baseUrl: string;
3814
+ reasoning: true;
3815
+ input: "text"[];
3816
+ cost: {
3817
+ input: number;
3818
+ output: number;
3819
+ cacheRead: number;
3820
+ cacheWrite: number;
3821
+ };
3822
+ contextWindow: number;
3823
+ maxTokens: number;
3824
+ };
3825
+ readonly "qwen/qwen3-14b": {
3826
+ id: string;
3827
+ name: string;
3828
+ api: "openai-completions";
3829
+ provider: string;
3830
+ baseUrl: string;
3831
+ reasoning: true;
3832
+ input: "text"[];
3833
+ cost: {
3834
+ input: number;
3835
+ output: number;
3836
+ cacheRead: number;
3837
+ cacheWrite: number;
3838
+ };
3839
+ contextWindow: number;
3840
+ maxTokens: number;
3841
+ };
3842
+ readonly "qwen/qwen3-32b": {
3843
+ id: string;
3844
+ name: string;
3845
+ api: "openai-completions";
3846
+ provider: string;
3847
+ baseUrl: string;
3848
+ reasoning: true;
3849
+ input: "text"[];
3850
+ cost: {
3851
+ input: number;
3852
+ output: number;
3853
+ cacheRead: number;
3854
+ cacheWrite: number;
3855
+ };
3856
+ contextWindow: number;
3857
+ maxTokens: number;
3858
+ };
3859
+ readonly "qwen/qwen3-235b-a22b:free": {
3860
+ id: string;
3861
+ name: string;
3862
+ api: "openai-completions";
3863
+ provider: string;
3864
+ baseUrl: string;
3865
+ reasoning: true;
3866
+ input: "text"[];
3867
+ cost: {
3868
+ input: number;
3869
+ output: number;
3870
+ cacheRead: number;
3871
+ cacheWrite: number;
3872
+ };
3873
+ contextWindow: number;
3874
+ maxTokens: number;
3875
+ };
3876
+ readonly "qwen/qwen3-235b-a22b": {
3877
+ id: string;
3878
+ name: string;
3879
+ api: "openai-completions";
3880
+ provider: string;
3881
+ baseUrl: string;
3882
+ reasoning: true;
3883
+ input: "text"[];
3884
+ cost: {
3885
+ input: number;
3886
+ output: number;
3887
+ cacheRead: number;
3888
+ cacheWrite: number;
3889
+ };
3890
+ contextWindow: number;
3891
+ maxTokens: number;
3892
+ };
3893
+ readonly "openai/o4-mini-high": {
3894
+ id: string;
3895
+ name: string;
3896
+ api: "openai-completions";
3897
+ provider: string;
3898
+ baseUrl: string;
3899
+ reasoning: true;
3900
+ input: ("image" | "text")[];
3901
+ cost: {
3902
+ input: number;
3903
+ output: number;
3904
+ cacheRead: number;
3905
+ cacheWrite: number;
3906
+ };
3907
+ contextWindow: number;
3908
+ maxTokens: number;
3909
+ };
3910
+ readonly "openai/o3": {
3911
+ id: string;
3912
+ name: string;
3913
+ api: "openai-completions";
3914
+ provider: string;
3915
+ baseUrl: string;
3916
+ reasoning: true;
3917
+ input: ("image" | "text")[];
3918
+ cost: {
3919
+ input: number;
3920
+ output: number;
3921
+ cacheRead: number;
3922
+ cacheWrite: number;
3923
+ };
3924
+ contextWindow: number;
3925
+ maxTokens: number;
3926
+ };
3927
+ readonly "openai/o4-mini": {
3928
+ id: string;
3929
+ name: string;
3930
+ api: "openai-completions";
3931
+ provider: string;
3932
+ baseUrl: string;
3933
+ reasoning: true;
3934
+ input: ("image" | "text")[];
3935
+ cost: {
3936
+ input: number;
3937
+ output: number;
3938
+ cacheRead: number;
3939
+ cacheWrite: number;
3940
+ };
3941
+ contextWindow: number;
3942
+ maxTokens: number;
3943
+ };
3944
+ readonly "openai/gpt-4.1": {
3945
+ id: string;
3946
+ name: string;
3947
+ api: "openai-completions";
3948
+ provider: string;
3949
+ baseUrl: string;
3950
+ reasoning: false;
3951
+ input: ("image" | "text")[];
3952
+ cost: {
3953
+ input: number;
3954
+ output: number;
3955
+ cacheRead: number;
3956
+ cacheWrite: number;
3957
+ };
3958
+ contextWindow: number;
3959
+ maxTokens: number;
3960
+ };
3961
+ readonly "openai/gpt-4.1-mini": {
3962
+ id: string;
3963
+ name: string;
3964
+ api: "openai-completions";
3965
+ provider: string;
3966
+ baseUrl: string;
3967
+ reasoning: false;
3968
+ input: ("image" | "text")[];
3969
+ cost: {
3970
+ input: number;
3971
+ output: number;
3972
+ cacheRead: number;
3973
+ cacheWrite: number;
3974
+ };
3975
+ contextWindow: number;
3976
+ maxTokens: number;
3977
+ };
3978
+ readonly "openai/gpt-4.1-nano": {
3979
+ id: string;
3980
+ name: string;
3981
+ api: "openai-completions";
3982
+ provider: string;
3983
+ baseUrl: string;
3984
+ reasoning: false;
3985
+ input: ("image" | "text")[];
3986
+ cost: {
3987
+ input: number;
3988
+ output: number;
3989
+ cacheRead: number;
3990
+ cacheWrite: number;
3991
+ };
3992
+ contextWindow: number;
3993
+ maxTokens: number;
3994
+ };
3995
+ readonly "x-ai/grok-3-mini-beta": {
3996
+ id: string;
3997
+ name: string;
3998
+ api: "openai-completions";
3999
+ provider: string;
4000
+ baseUrl: string;
4001
+ reasoning: true;
4002
+ input: "text"[];
4003
+ cost: {
4004
+ input: number;
4005
+ output: number;
4006
+ cacheRead: number;
4007
+ cacheWrite: number;
4008
+ };
4009
+ contextWindow: number;
4010
+ maxTokens: number;
4011
+ };
4012
+ readonly "x-ai/grok-3-beta": {
4013
+ id: string;
4014
+ name: string;
4015
+ api: "openai-completions";
4016
+ provider: string;
4017
+ baseUrl: string;
4018
+ reasoning: false;
4019
+ input: "text"[];
4020
+ cost: {
4021
+ input: number;
4022
+ output: number;
4023
+ cacheRead: number;
4024
+ cacheWrite: number;
4025
+ };
4026
+ contextWindow: number;
4027
+ maxTokens: number;
4028
+ };
4029
+ readonly "meta-llama/llama-4-maverick:free": {
4030
+ id: string;
4031
+ name: string;
4032
+ api: "openai-completions";
4033
+ provider: string;
4034
+ baseUrl: string;
4035
+ reasoning: false;
4036
+ input: ("image" | "text")[];
4037
+ cost: {
4038
+ input: number;
4039
+ output: number;
4040
+ cacheRead: number;
4041
+ cacheWrite: number;
4042
+ };
4043
+ contextWindow: number;
4044
+ maxTokens: number;
4045
+ };
4046
+ readonly "meta-llama/llama-4-maverick": {
4047
+ id: string;
4048
+ name: string;
4049
+ api: "openai-completions";
4050
+ provider: string;
4051
+ baseUrl: string;
4052
+ reasoning: false;
4053
+ input: ("image" | "text")[];
4054
+ cost: {
4055
+ input: number;
4056
+ output: number;
4057
+ cacheRead: number;
4058
+ cacheWrite: number;
4059
+ };
4060
+ contextWindow: number;
4061
+ maxTokens: number;
4062
+ };
4063
+ readonly "meta-llama/llama-4-scout:free": {
4064
+ id: string;
4065
+ name: string;
4066
+ api: "openai-completions";
4067
+ provider: string;
4068
+ baseUrl: string;
4069
+ reasoning: false;
4070
+ input: ("image" | "text")[];
4071
+ cost: {
4072
+ input: number;
4073
+ output: number;
4074
+ cacheRead: number;
4075
+ cacheWrite: number;
4076
+ };
4077
+ contextWindow: number;
4078
+ maxTokens: number;
4079
+ };
4080
+ readonly "meta-llama/llama-4-scout": {
4081
+ id: string;
4082
+ name: string;
4083
+ api: "openai-completions";
4084
+ provider: string;
4085
+ baseUrl: string;
4086
+ reasoning: false;
4087
+ input: ("image" | "text")[];
4088
+ cost: {
4089
+ input: number;
4090
+ output: number;
4091
+ cacheRead: number;
4092
+ cacheWrite: number;
4093
+ };
4094
+ contextWindow: number;
4095
+ maxTokens: number;
4096
+ };
4097
+ readonly "deepseek/deepseek-chat-v3-0324:free": {
4098
+ id: string;
4099
+ name: string;
4100
+ api: "openai-completions";
4101
+ provider: string;
4102
+ baseUrl: string;
4103
+ reasoning: false;
4104
+ input: "text"[];
4105
+ cost: {
4106
+ input: number;
4107
+ output: number;
4108
+ cacheRead: number;
4109
+ cacheWrite: number;
4110
+ };
4111
+ contextWindow: number;
4112
+ maxTokens: number;
4113
+ };
4114
+ readonly "deepseek/deepseek-chat-v3-0324": {
4115
+ id: string;
4116
+ name: string;
4117
+ api: "openai-completions";
4118
+ provider: string;
4119
+ baseUrl: string;
4120
+ reasoning: true;
4121
+ input: "text"[];
4122
+ cost: {
4123
+ input: number;
4124
+ output: number;
4125
+ cacheRead: number;
4126
+ cacheWrite: number;
4127
+ };
4128
+ contextWindow: number;
4129
+ maxTokens: number;
4130
+ };
4131
+ readonly "mistralai/mistral-small-3.1-24b-instruct:free": {
4132
+ id: string;
4133
+ name: string;
4134
+ api: "openai-completions";
4135
+ provider: string;
4136
+ baseUrl: string;
4137
+ reasoning: false;
4138
+ input: ("image" | "text")[];
4139
+ cost: {
4140
+ input: number;
4141
+ output: number;
4142
+ cacheRead: number;
4143
+ cacheWrite: number;
4144
+ };
4145
+ contextWindow: number;
4146
+ maxTokens: number;
4147
+ };
4148
+ readonly "mistralai/mistral-small-3.1-24b-instruct": {
4149
+ id: string;
4150
+ name: string;
4151
+ api: "openai-completions";
4152
+ provider: string;
4153
+ baseUrl: string;
4154
+ reasoning: false;
4155
+ input: ("image" | "text")[];
4156
+ cost: {
4157
+ input: number;
4158
+ output: number;
4159
+ cacheRead: number;
4160
+ cacheWrite: number;
4161
+ };
4162
+ contextWindow: number;
4163
+ maxTokens: number;
4164
+ };
4165
+ readonly "google/gemma-3-27b-it": {
4166
+ id: string;
4167
+ name: string;
4168
+ api: "openai-completions";
4169
+ provider: string;
4170
+ baseUrl: string;
4171
+ reasoning: false;
4172
+ input: ("image" | "text")[];
4173
+ cost: {
4174
+ input: number;
4175
+ output: number;
4176
+ cacheRead: number;
4177
+ cacheWrite: number;
4178
+ };
4179
+ contextWindow: number;
4180
+ maxTokens: number;
4181
+ };
4182
+ readonly "qwen/qwq-32b": {
4183
+ id: string;
4184
+ name: string;
4185
+ api: "openai-completions";
4186
+ provider: string;
4187
+ baseUrl: string;
4188
+ reasoning: true;
4189
+ input: "text"[];
4190
+ cost: {
4191
+ input: number;
4192
+ output: number;
4193
+ cacheRead: number;
4194
+ cacheWrite: number;
4195
+ };
4196
+ contextWindow: number;
4197
+ maxTokens: number;
4198
+ };
4199
+ readonly "google/gemini-2.0-flash-lite-001": {
4200
+ id: string;
4201
+ name: string;
4202
+ api: "openai-completions";
4203
+ provider: string;
4204
+ baseUrl: string;
4205
+ reasoning: false;
4206
+ input: ("image" | "text")[];
4207
+ cost: {
4208
+ input: number;
4209
+ output: number;
4210
+ cacheRead: number;
4211
+ cacheWrite: number;
4212
+ };
4213
+ contextWindow: number;
4214
+ maxTokens: number;
4215
+ };
4216
+ readonly "anthropic/claude-3.7-sonnet:thinking": {
4217
+ id: string;
4218
+ name: string;
4219
+ api: "openai-completions";
4220
+ provider: string;
4221
+ baseUrl: string;
4222
+ reasoning: true;
4223
+ input: ("image" | "text")[];
4224
+ cost: {
4225
+ input: number;
4226
+ output: number;
4227
+ cacheRead: number;
4228
+ cacheWrite: number;
4229
+ };
4230
+ contextWindow: number;
4231
+ maxTokens: number;
4232
+ };
4233
+ readonly "anthropic/claude-3.7-sonnet": {
4234
+ id: string;
4235
+ name: string;
4236
+ api: "openai-completions";
4237
+ provider: string;
4238
+ baseUrl: string;
4239
+ reasoning: true;
4240
+ input: ("image" | "text")[];
4241
+ cost: {
4242
+ input: number;
4243
+ output: number;
4244
+ cacheRead: number;
4245
+ cacheWrite: number;
4246
+ };
4247
+ contextWindow: number;
4248
+ maxTokens: number;
4249
+ };
4250
+ readonly "mistralai/mistral-saba": {
4251
+ id: string;
4252
+ name: string;
4253
+ api: "openai-completions";
4254
+ provider: string;
4255
+ baseUrl: string;
4256
+ reasoning: false;
4257
+ input: "text"[];
4258
+ cost: {
4259
+ input: number;
4260
+ output: number;
4261
+ cacheRead: number;
4262
+ cacheWrite: number;
4263
+ };
4264
+ contextWindow: number;
4265
+ maxTokens: number;
4266
+ };
4267
+ readonly "openai/o3-mini-high": {
4268
+ id: string;
4269
+ name: string;
4270
+ api: "openai-completions";
4271
+ provider: string;
4272
+ baseUrl: string;
4273
+ reasoning: false;
4274
+ input: "text"[];
4275
+ cost: {
4276
+ input: number;
4277
+ output: number;
4278
+ cacheRead: number;
4279
+ cacheWrite: number;
4280
+ };
4281
+ contextWindow: number;
4282
+ maxTokens: number;
4283
+ };
4284
+ readonly "google/gemini-2.0-flash-001": {
4285
+ id: string;
4286
+ name: string;
4287
+ api: "openai-completions";
4288
+ provider: string;
4289
+ baseUrl: string;
4290
+ reasoning: false;
4291
+ input: ("image" | "text")[];
4292
+ cost: {
4293
+ input: number;
4294
+ output: number;
4295
+ cacheRead: number;
4296
+ cacheWrite: number;
4297
+ };
4298
+ contextWindow: number;
4299
+ maxTokens: number;
4300
+ };
4301
+ readonly "qwen/qwen-vl-max": {
4302
+ id: string;
4303
+ name: string;
4304
+ api: "openai-completions";
4305
+ provider: string;
4306
+ baseUrl: string;
4307
+ reasoning: false;
4308
+ input: ("image" | "text")[];
4309
+ cost: {
4310
+ input: number;
4311
+ output: number;
4312
+ cacheRead: number;
4313
+ cacheWrite: number;
4314
+ };
4315
+ contextWindow: number;
4316
+ maxTokens: number;
4317
+ };
4318
+ readonly "qwen/qwen-turbo": {
4319
+ id: string;
4320
+ name: string;
4321
+ api: "openai-completions";
4322
+ provider: string;
4323
+ baseUrl: string;
4324
+ reasoning: false;
4325
+ input: "text"[];
4326
+ cost: {
4327
+ input: number;
4328
+ output: number;
4329
+ cacheRead: number;
4330
+ cacheWrite: number;
4331
+ };
4332
+ contextWindow: number;
4333
+ maxTokens: number;
4334
+ };
4335
+ readonly "qwen/qwen-plus": {
4336
+ id: string;
4337
+ name: string;
4338
+ api: "openai-completions";
4339
+ provider: string;
4340
+ baseUrl: string;
4341
+ reasoning: false;
3152
4342
  input: "text"[];
3153
4343
  cost: {
3154
4344
  input: number;
@@ -3159,7 +4349,58 @@ export declare const MODELS: {
3159
4349
  contextWindow: number;
3160
4350
  maxTokens: number;
3161
4351
  };
3162
- readonly "qwen/qwen3-235b-a22b:free": {
4352
+ readonly "qwen/qwen-max": {
4353
+ id: string;
4354
+ name: string;
4355
+ api: "openai-completions";
4356
+ provider: string;
4357
+ baseUrl: string;
4358
+ reasoning: false;
4359
+ input: "text"[];
4360
+ cost: {
4361
+ input: number;
4362
+ output: number;
4363
+ cacheRead: number;
4364
+ cacheWrite: number;
4365
+ };
4366
+ contextWindow: number;
4367
+ maxTokens: number;
4368
+ };
4369
+ readonly "openai/o3-mini": {
4370
+ id: string;
4371
+ name: string;
4372
+ api: "openai-completions";
4373
+ provider: string;
4374
+ baseUrl: string;
4375
+ reasoning: false;
4376
+ input: "text"[];
4377
+ cost: {
4378
+ input: number;
4379
+ output: number;
4380
+ cacheRead: number;
4381
+ cacheWrite: number;
4382
+ };
4383
+ contextWindow: number;
4384
+ maxTokens: number;
4385
+ };
4386
+ readonly "mistralai/mistral-small-24b-instruct-2501": {
4387
+ id: string;
4388
+ name: string;
4389
+ api: "openai-completions";
4390
+ provider: string;
4391
+ baseUrl: string;
4392
+ reasoning: false;
4393
+ input: "text"[];
4394
+ cost: {
4395
+ input: number;
4396
+ output: number;
4397
+ cacheRead: number;
4398
+ cacheWrite: number;
4399
+ };
4400
+ contextWindow: number;
4401
+ maxTokens: number;
4402
+ };
4403
+ readonly "deepseek/deepseek-r1-distill-llama-70b": {
3163
4404
  id: string;
3164
4405
  name: string;
3165
4406
  api: "openai-completions";
@@ -3176,7 +4417,7 @@ export declare const MODELS: {
3176
4417
  contextWindow: number;
3177
4418
  maxTokens: number;
3178
4419
  };
3179
- readonly "qwen/qwen3-235b-a22b": {
4420
+ readonly "deepseek/deepseek-r1": {
3180
4421
  id: string;
3181
4422
  name: string;
3182
4423
  api: "openai-completions";
@@ -3193,14 +4434,14 @@ export declare const MODELS: {
3193
4434
  contextWindow: number;
3194
4435
  maxTokens: number;
3195
4436
  };
3196
- readonly "meta-llama/llama-4-maverick:free": {
4437
+ readonly "mistralai/codestral-2501": {
3197
4438
  id: string;
3198
4439
  name: string;
3199
4440
  api: "openai-completions";
3200
4441
  provider: string;
3201
4442
  baseUrl: string;
3202
4443
  reasoning: false;
3203
- input: ("text" | "image")[];
4444
+ input: "text"[];
3204
4445
  cost: {
3205
4446
  input: number;
3206
4447
  output: number;
@@ -3210,14 +4451,14 @@ export declare const MODELS: {
3210
4451
  contextWindow: number;
3211
4452
  maxTokens: number;
3212
4453
  };
3213
- readonly "meta-llama/llama-4-maverick": {
4454
+ readonly "deepseek/deepseek-chat": {
3214
4455
  id: string;
3215
4456
  name: string;
3216
4457
  api: "openai-completions";
3217
4458
  provider: string;
3218
4459
  baseUrl: string;
3219
4460
  reasoning: false;
3220
- input: ("text" | "image")[];
4461
+ input: "text"[];
3221
4462
  cost: {
3222
4463
  input: number;
3223
4464
  output: number;
@@ -3227,14 +4468,14 @@ export declare const MODELS: {
3227
4468
  contextWindow: number;
3228
4469
  maxTokens: number;
3229
4470
  };
3230
- readonly "meta-llama/llama-4-scout:free": {
4471
+ readonly "openai/o1": {
3231
4472
  id: string;
3232
4473
  name: string;
3233
4474
  api: "openai-completions";
3234
4475
  provider: string;
3235
4476
  baseUrl: string;
3236
4477
  reasoning: false;
3237
- input: ("text" | "image")[];
4478
+ input: ("image" | "text")[];
3238
4479
  cost: {
3239
4480
  input: number;
3240
4481
  output: number;
@@ -3244,14 +4485,14 @@ export declare const MODELS: {
3244
4485
  contextWindow: number;
3245
4486
  maxTokens: number;
3246
4487
  };
3247
- readonly "meta-llama/llama-4-scout": {
4488
+ readonly "google/gemini-2.0-flash-exp:free": {
3248
4489
  id: string;
3249
4490
  name: string;
3250
4491
  api: "openai-completions";
3251
4492
  provider: string;
3252
4493
  baseUrl: string;
3253
4494
  reasoning: false;
3254
- input: ("text" | "image")[];
4495
+ input: ("image" | "text")[];
3255
4496
  cost: {
3256
4497
  input: number;
3257
4498
  output: number;
@@ -3261,7 +4502,7 @@ export declare const MODELS: {
3261
4502
  contextWindow: number;
3262
4503
  maxTokens: number;
3263
4504
  };
3264
- readonly "deepseek/deepseek-chat-v3-0324:free": {
4505
+ readonly "meta-llama/llama-3.3-70b-instruct:free": {
3265
4506
  id: string;
3266
4507
  name: string;
3267
4508
  api: "openai-completions";
@@ -3278,13 +4519,13 @@ export declare const MODELS: {
3278
4519
  contextWindow: number;
3279
4520
  maxTokens: number;
3280
4521
  };
3281
- readonly "deepseek/deepseek-chat-v3-0324": {
4522
+ readonly "meta-llama/llama-3.3-70b-instruct": {
3282
4523
  id: string;
3283
4524
  name: string;
3284
4525
  api: "openai-completions";
3285
4526
  provider: string;
3286
4527
  baseUrl: string;
3287
- reasoning: true;
4528
+ reasoning: false;
3288
4529
  input: "text"[];
3289
4530
  cost: {
3290
4531
  input: number;
@@ -3295,14 +4536,14 @@ export declare const MODELS: {
3295
4536
  contextWindow: number;
3296
4537
  maxTokens: number;
3297
4538
  };
3298
- readonly "mistralai/mistral-small-3.1-24b-instruct:free": {
4539
+ readonly "amazon/nova-lite-v1": {
3299
4540
  id: string;
3300
4541
  name: string;
3301
4542
  api: "openai-completions";
3302
4543
  provider: string;
3303
4544
  baseUrl: string;
3304
4545
  reasoning: false;
3305
- input: ("text" | "image")[];
4546
+ input: ("image" | "text")[];
3306
4547
  cost: {
3307
4548
  input: number;
3308
4549
  output: number;
@@ -3312,14 +4553,14 @@ export declare const MODELS: {
3312
4553
  contextWindow: number;
3313
4554
  maxTokens: number;
3314
4555
  };
3315
- readonly "mistralai/mistral-small-3.1-24b-instruct": {
4556
+ readonly "amazon/nova-micro-v1": {
3316
4557
  id: string;
3317
4558
  name: string;
3318
4559
  api: "openai-completions";
3319
4560
  provider: string;
3320
4561
  baseUrl: string;
3321
4562
  reasoning: false;
3322
- input: ("text" | "image")[];
4563
+ input: "text"[];
3323
4564
  cost: {
3324
4565
  input: number;
3325
4566
  output: number;
@@ -3329,14 +4570,31 @@ export declare const MODELS: {
3329
4570
  contextWindow: number;
3330
4571
  maxTokens: number;
3331
4572
  };
3332
- readonly "qwen/qwq-32b": {
4573
+ readonly "amazon/nova-pro-v1": {
3333
4574
  id: string;
3334
4575
  name: string;
3335
4576
  api: "openai-completions";
3336
4577
  provider: string;
3337
4578
  baseUrl: string;
3338
- reasoning: true;
3339
- input: "text"[];
4579
+ reasoning: false;
4580
+ input: ("image" | "text")[];
4581
+ cost: {
4582
+ input: number;
4583
+ output: number;
4584
+ cacheRead: number;
4585
+ cacheWrite: number;
4586
+ };
4587
+ contextWindow: number;
4588
+ maxTokens: number;
4589
+ };
4590
+ readonly "openai/gpt-4o-2024-11-20": {
4591
+ id: string;
4592
+ name: string;
4593
+ api: "openai-completions";
4594
+ provider: string;
4595
+ baseUrl: string;
4596
+ reasoning: false;
4597
+ input: ("image" | "text")[];
3340
4598
  cost: {
3341
4599
  input: number;
3342
4600
  output: number;
@@ -3346,7 +4604,7 @@ export declare const MODELS: {
3346
4604
  contextWindow: number;
3347
4605
  maxTokens: number;
3348
4606
  };
3349
- readonly "nousresearch/deephermes-3-llama-3-8b-preview": {
4607
+ readonly "mistralai/mistral-large-2411": {
3350
4608
  id: string;
3351
4609
  name: string;
3352
4610
  api: "openai-completions";
@@ -3363,7 +4621,7 @@ export declare const MODELS: {
3363
4621
  contextWindow: number;
3364
4622
  maxTokens: number;
3365
4623
  };
3366
- readonly "mistralai/mistral-saba": {
4624
+ readonly "mistralai/mistral-large-2407": {
3367
4625
  id: string;
3368
4626
  name: string;
3369
4627
  api: "openai-completions";
@@ -3380,14 +4638,14 @@ export declare const MODELS: {
3380
4638
  contextWindow: number;
3381
4639
  maxTokens: number;
3382
4640
  };
3383
- readonly "qwen/qwen-vl-max": {
4641
+ readonly "mistralai/pixtral-large-2411": {
3384
4642
  id: string;
3385
4643
  name: string;
3386
4644
  api: "openai-completions";
3387
4645
  provider: string;
3388
4646
  baseUrl: string;
3389
4647
  reasoning: false;
3390
- input: ("text" | "image")[];
4648
+ input: ("image" | "text")[];
3391
4649
  cost: {
3392
4650
  input: number;
3393
4651
  output: number;
@@ -3397,7 +4655,7 @@ export declare const MODELS: {
3397
4655
  contextWindow: number;
3398
4656
  maxTokens: number;
3399
4657
  };
3400
- readonly "qwen/qwen-turbo": {
4658
+ readonly "thedrummer/unslopnemo-12b": {
3401
4659
  id: string;
3402
4660
  name: string;
3403
4661
  api: "openai-completions";
@@ -3414,14 +4672,14 @@ export declare const MODELS: {
3414
4672
  contextWindow: number;
3415
4673
  maxTokens: number;
3416
4674
  };
3417
- readonly "qwen/qwen-plus": {
4675
+ readonly "anthropic/claude-3.5-haiku-20241022": {
3418
4676
  id: string;
3419
4677
  name: string;
3420
4678
  api: "openai-completions";
3421
4679
  provider: string;
3422
4680
  baseUrl: string;
3423
4681
  reasoning: false;
3424
- input: "text"[];
4682
+ input: ("image" | "text")[];
3425
4683
  cost: {
3426
4684
  input: number;
3427
4685
  output: number;
@@ -3431,14 +4689,14 @@ export declare const MODELS: {
3431
4689
  contextWindow: number;
3432
4690
  maxTokens: number;
3433
4691
  };
3434
- readonly "qwen/qwen-max": {
4692
+ readonly "anthropic/claude-3.5-haiku": {
3435
4693
  id: string;
3436
4694
  name: string;
3437
4695
  api: "openai-completions";
3438
4696
  provider: string;
3439
4697
  baseUrl: string;
3440
4698
  reasoning: false;
3441
- input: "text"[];
4699
+ input: ("image" | "text")[];
3442
4700
  cost: {
3443
4701
  input: number;
3444
4702
  output: number;
@@ -3448,7 +4706,24 @@ export declare const MODELS: {
3448
4706
  contextWindow: number;
3449
4707
  maxTokens: number;
3450
4708
  };
3451
- readonly "mistralai/mistral-small-24b-instruct-2501": {
4709
+ readonly "anthropic/claude-3.5-sonnet": {
4710
+ id: string;
4711
+ name: string;
4712
+ api: "openai-completions";
4713
+ provider: string;
4714
+ baseUrl: string;
4715
+ reasoning: false;
4716
+ input: ("image" | "text")[];
4717
+ cost: {
4718
+ input: number;
4719
+ output: number;
4720
+ cacheRead: number;
4721
+ cacheWrite: number;
4722
+ };
4723
+ contextWindow: number;
4724
+ maxTokens: number;
4725
+ };
4726
+ readonly "mistralai/ministral-8b": {
3452
4727
  id: string;
3453
4728
  name: string;
3454
4729
  api: "openai-completions";
@@ -3465,13 +4740,13 @@ export declare const MODELS: {
3465
4740
  contextWindow: number;
3466
4741
  maxTokens: number;
3467
4742
  };
3468
- readonly "deepseek/deepseek-r1-distill-llama-70b": {
4743
+ readonly "mistralai/ministral-3b": {
3469
4744
  id: string;
3470
4745
  name: string;
3471
4746
  api: "openai-completions";
3472
4747
  provider: string;
3473
4748
  baseUrl: string;
3474
- reasoning: true;
4749
+ reasoning: false;
3475
4750
  input: "text"[];
3476
4751
  cost: {
3477
4752
  input: number;
@@ -3482,13 +4757,13 @@ export declare const MODELS: {
3482
4757
  contextWindow: number;
3483
4758
  maxTokens: number;
3484
4759
  };
3485
- readonly "deepseek/deepseek-r1": {
4760
+ readonly "qwen/qwen-2.5-7b-instruct": {
3486
4761
  id: string;
3487
4762
  name: string;
3488
4763
  api: "openai-completions";
3489
4764
  provider: string;
3490
4765
  baseUrl: string;
3491
- reasoning: true;
4766
+ reasoning: false;
3492
4767
  input: "text"[];
3493
4768
  cost: {
3494
4769
  input: number;
@@ -3499,7 +4774,7 @@ export declare const MODELS: {
3499
4774
  contextWindow: number;
3500
4775
  maxTokens: number;
3501
4776
  };
3502
- readonly "mistralai/codestral-2501": {
4777
+ readonly "nvidia/llama-3.1-nemotron-70b-instruct": {
3503
4778
  id: string;
3504
4779
  name: string;
3505
4780
  api: "openai-completions";
@@ -3516,7 +4791,7 @@ export declare const MODELS: {
3516
4791
  contextWindow: number;
3517
4792
  maxTokens: number;
3518
4793
  };
3519
- readonly "deepseek/deepseek-chat": {
4794
+ readonly "thedrummer/rocinante-12b": {
3520
4795
  id: string;
3521
4796
  name: string;
3522
4797
  api: "openai-completions";
@@ -3533,7 +4808,7 @@ export declare const MODELS: {
3533
4808
  contextWindow: number;
3534
4809
  maxTokens: number;
3535
4810
  };
3536
- readonly "meta-llama/llama-3.3-70b-instruct:free": {
4811
+ readonly "meta-llama/llama-3.2-3b-instruct": {
3537
4812
  id: string;
3538
4813
  name: string;
3539
4814
  api: "openai-completions";
@@ -3550,7 +4825,7 @@ export declare const MODELS: {
3550
4825
  contextWindow: number;
3551
4826
  maxTokens: number;
3552
4827
  };
3553
- readonly "meta-llama/llama-3.3-70b-instruct": {
4828
+ readonly "qwen/qwen-2.5-72b-instruct": {
3554
4829
  id: string;
3555
4830
  name: string;
3556
4831
  api: "openai-completions";
@@ -3567,14 +4842,14 @@ export declare const MODELS: {
3567
4842
  contextWindow: number;
3568
4843
  maxTokens: number;
3569
4844
  };
3570
- readonly "amazon/nova-lite-v1": {
4845
+ readonly "mistralai/pixtral-12b": {
3571
4846
  id: string;
3572
4847
  name: string;
3573
4848
  api: "openai-completions";
3574
4849
  provider: string;
3575
4850
  baseUrl: string;
3576
4851
  reasoning: false;
3577
- input: ("text" | "image")[];
4852
+ input: ("image" | "text")[];
3578
4853
  cost: {
3579
4854
  input: number;
3580
4855
  output: number;
@@ -3584,7 +4859,7 @@ export declare const MODELS: {
3584
4859
  contextWindow: number;
3585
4860
  maxTokens: number;
3586
4861
  };
3587
- readonly "amazon/nova-micro-v1": {
4862
+ readonly "cohere/command-r-08-2024": {
3588
4863
  id: string;
3589
4864
  name: string;
3590
4865
  api: "openai-completions";
@@ -3601,14 +4876,14 @@ export declare const MODELS: {
3601
4876
  contextWindow: number;
3602
4877
  maxTokens: number;
3603
4878
  };
3604
- readonly "amazon/nova-pro-v1": {
4879
+ readonly "cohere/command-r-plus-08-2024": {
3605
4880
  id: string;
3606
4881
  name: string;
3607
4882
  api: "openai-completions";
3608
4883
  provider: string;
3609
4884
  baseUrl: string;
3610
4885
  reasoning: false;
3611
- input: ("text" | "image")[];
4886
+ input: "text"[];
3612
4887
  cost: {
3613
4888
  input: number;
3614
4889
  output: number;
@@ -3618,7 +4893,7 @@ export declare const MODELS: {
3618
4893
  contextWindow: number;
3619
4894
  maxTokens: number;
3620
4895
  };
3621
- readonly "mistralai/mistral-large-2411": {
4896
+ readonly "sao10k/l3.1-euryale-70b": {
3622
4897
  id: string;
3623
4898
  name: string;
3624
4899
  api: "openai-completions";
@@ -3635,7 +4910,7 @@ export declare const MODELS: {
3635
4910
  contextWindow: number;
3636
4911
  maxTokens: number;
3637
4912
  };
3638
- readonly "mistralai/mistral-large-2407": {
4913
+ readonly "microsoft/phi-3.5-mini-128k-instruct": {
3639
4914
  id: string;
3640
4915
  name: string;
3641
4916
  api: "openai-completions";
@@ -3652,14 +4927,14 @@ export declare const MODELS: {
3652
4927
  contextWindow: number;
3653
4928
  maxTokens: number;
3654
4929
  };
3655
- readonly "mistralai/pixtral-large-2411": {
4930
+ readonly "nousresearch/hermes-3-llama-3.1-70b": {
3656
4931
  id: string;
3657
4932
  name: string;
3658
4933
  api: "openai-completions";
3659
4934
  provider: string;
3660
4935
  baseUrl: string;
3661
4936
  reasoning: false;
3662
- input: ("text" | "image")[];
4937
+ input: "text"[];
3663
4938
  cost: {
3664
4939
  input: number;
3665
4940
  output: number;
@@ -3669,14 +4944,14 @@ export declare const MODELS: {
3669
4944
  contextWindow: number;
3670
4945
  maxTokens: number;
3671
4946
  };
3672
- readonly "thedrummer/unslopnemo-12b": {
4947
+ readonly "openai/gpt-4o-2024-08-06": {
3673
4948
  id: string;
3674
4949
  name: string;
3675
4950
  api: "openai-completions";
3676
4951
  provider: string;
3677
4952
  baseUrl: string;
3678
4953
  reasoning: false;
3679
- input: "text"[];
4954
+ input: ("image" | "text")[];
3680
4955
  cost: {
3681
4956
  input: number;
3682
4957
  output: number;
@@ -3686,7 +4961,7 @@ export declare const MODELS: {
3686
4961
  contextWindow: number;
3687
4962
  maxTokens: number;
3688
4963
  };
3689
- readonly "mistralai/ministral-8b": {
4964
+ readonly "meta-llama/llama-3.1-8b-instruct": {
3690
4965
  id: string;
3691
4966
  name: string;
3692
4967
  api: "openai-completions";
@@ -3703,7 +4978,7 @@ export declare const MODELS: {
3703
4978
  contextWindow: number;
3704
4979
  maxTokens: number;
3705
4980
  };
3706
- readonly "qwen/qwen-2.5-7b-instruct": {
4981
+ readonly "meta-llama/llama-3.1-405b-instruct": {
3707
4982
  id: string;
3708
4983
  name: string;
3709
4984
  api: "openai-completions";
@@ -3720,7 +4995,7 @@ export declare const MODELS: {
3720
4995
  contextWindow: number;
3721
4996
  maxTokens: number;
3722
4997
  };
3723
- readonly "nvidia/llama-3.1-nemotron-70b-instruct": {
4998
+ readonly "meta-llama/llama-3.1-70b-instruct": {
3724
4999
  id: string;
3725
5000
  name: string;
3726
5001
  api: "openai-completions";
@@ -3737,7 +5012,7 @@ export declare const MODELS: {
3737
5012
  contextWindow: number;
3738
5013
  maxTokens: number;
3739
5014
  };
3740
- readonly "thedrummer/rocinante-12b": {
5015
+ readonly "mistralai/mistral-nemo": {
3741
5016
  id: string;
3742
5017
  name: string;
3743
5018
  api: "openai-completions";
@@ -3754,14 +5029,14 @@ export declare const MODELS: {
3754
5029
  contextWindow: number;
3755
5030
  maxTokens: number;
3756
5031
  };
3757
- readonly "meta-llama/llama-3.2-3b-instruct": {
5032
+ readonly "openai/gpt-4o-mini-2024-07-18": {
3758
5033
  id: string;
3759
5034
  name: string;
3760
5035
  api: "openai-completions";
3761
5036
  provider: string;
3762
5037
  baseUrl: string;
3763
5038
  reasoning: false;
3764
- input: "text"[];
5039
+ input: ("image" | "text")[];
3765
5040
  cost: {
3766
5041
  input: number;
3767
5042
  output: number;
@@ -3771,14 +5046,14 @@ export declare const MODELS: {
3771
5046
  contextWindow: number;
3772
5047
  maxTokens: number;
3773
5048
  };
3774
- readonly "qwen/qwen-2.5-72b-instruct": {
5049
+ readonly "openai/gpt-4o-mini": {
3775
5050
  id: string;
3776
5051
  name: string;
3777
5052
  api: "openai-completions";
3778
5053
  provider: string;
3779
5054
  baseUrl: string;
3780
5055
  reasoning: false;
3781
- input: "text"[];
5056
+ input: ("image" | "text")[];
3782
5057
  cost: {
3783
5058
  input: number;
3784
5059
  output: number;
@@ -3788,14 +5063,14 @@ export declare const MODELS: {
3788
5063
  contextWindow: number;
3789
5064
  maxTokens: number;
3790
5065
  };
3791
- readonly "mistralai/pixtral-12b": {
5066
+ readonly "anthropic/claude-3.5-sonnet-20240620": {
3792
5067
  id: string;
3793
5068
  name: string;
3794
5069
  api: "openai-completions";
3795
5070
  provider: string;
3796
5071
  baseUrl: string;
3797
5072
  reasoning: false;
3798
- input: ("text" | "image")[];
5073
+ input: ("image" | "text")[];
3799
5074
  cost: {
3800
5075
  input: number;
3801
5076
  output: number;
@@ -3805,7 +5080,7 @@ export declare const MODELS: {
3805
5080
  contextWindow: number;
3806
5081
  maxTokens: number;
3807
5082
  };
3808
- readonly "cohere/command-r-plus-08-2024": {
5083
+ readonly "sao10k/l3-euryale-70b": {
3809
5084
  id: string;
3810
5085
  name: string;
3811
5086
  api: "openai-completions";
@@ -3822,7 +5097,7 @@ export declare const MODELS: {
3822
5097
  contextWindow: number;
3823
5098
  maxTokens: number;
3824
5099
  };
3825
- readonly "cohere/command-r-08-2024": {
5100
+ readonly "mistralai/mistral-7b-instruct:free": {
3826
5101
  id: string;
3827
5102
  name: string;
3828
5103
  api: "openai-completions";
@@ -3839,7 +5114,7 @@ export declare const MODELS: {
3839
5114
  contextWindow: number;
3840
5115
  maxTokens: number;
3841
5116
  };
3842
- readonly "sao10k/l3.1-euryale-70b": {
5117
+ readonly "mistralai/mistral-7b-instruct": {
3843
5118
  id: string;
3844
5119
  name: string;
3845
5120
  api: "openai-completions";
@@ -3856,7 +5131,7 @@ export declare const MODELS: {
3856
5131
  contextWindow: number;
3857
5132
  maxTokens: number;
3858
5133
  };
3859
- readonly "microsoft/phi-3.5-mini-128k-instruct": {
5134
+ readonly "microsoft/phi-3-mini-128k-instruct": {
3860
5135
  id: string;
3861
5136
  name: string;
3862
5137
  api: "openai-completions";
@@ -3873,7 +5148,7 @@ export declare const MODELS: {
3873
5148
  contextWindow: number;
3874
5149
  maxTokens: number;
3875
5150
  };
3876
- readonly "nousresearch/hermes-3-llama-3.1-70b": {
5151
+ readonly "microsoft/phi-3-medium-128k-instruct": {
3877
5152
  id: string;
3878
5153
  name: string;
3879
5154
  api: "openai-completions";
@@ -3890,14 +5165,14 @@ export declare const MODELS: {
3890
5165
  contextWindow: number;
3891
5166
  maxTokens: number;
3892
5167
  };
3893
- readonly "meta-llama/llama-3.1-8b-instruct": {
5168
+ readonly "openai/gpt-4o-2024-05-13": {
3894
5169
  id: string;
3895
5170
  name: string;
3896
5171
  api: "openai-completions";
3897
5172
  provider: string;
3898
5173
  baseUrl: string;
3899
5174
  reasoning: false;
3900
- input: "text"[];
5175
+ input: ("image" | "text")[];
3901
5176
  cost: {
3902
5177
  input: number;
3903
5178
  output: number;
@@ -3907,14 +5182,14 @@ export declare const MODELS: {
3907
5182
  contextWindow: number;
3908
5183
  maxTokens: number;
3909
5184
  };
3910
- readonly "meta-llama/llama-3.1-405b-instruct": {
5185
+ readonly "openai/gpt-4o": {
3911
5186
  id: string;
3912
5187
  name: string;
3913
5188
  api: "openai-completions";
3914
5189
  provider: string;
3915
5190
  baseUrl: string;
3916
5191
  reasoning: false;
3917
- input: "text"[];
5192
+ input: ("image" | "text")[];
3918
5193
  cost: {
3919
5194
  input: number;
3920
5195
  output: number;
@@ -3924,14 +5199,14 @@ export declare const MODELS: {
3924
5199
  contextWindow: number;
3925
5200
  maxTokens: number;
3926
5201
  };
3927
- readonly "meta-llama/llama-3.1-70b-instruct": {
5202
+ readonly "openai/gpt-4o:extended": {
3928
5203
  id: string;
3929
5204
  name: string;
3930
5205
  api: "openai-completions";
3931
5206
  provider: string;
3932
5207
  baseUrl: string;
3933
5208
  reasoning: false;
3934
- input: "text"[];
5209
+ input: ("image" | "text")[];
3935
5210
  cost: {
3936
5211
  input: number;
3937
5212
  output: number;
@@ -3941,7 +5216,7 @@ export declare const MODELS: {
3941
5216
  contextWindow: number;
3942
5217
  maxTokens: number;
3943
5218
  };
3944
- readonly "mistralai/mistral-nemo": {
5219
+ readonly "meta-llama/llama-3-70b-instruct": {
3945
5220
  id: string;
3946
5221
  name: string;
3947
5222
  api: "openai-completions";
@@ -3958,7 +5233,7 @@ export declare const MODELS: {
3958
5233
  contextWindow: number;
3959
5234
  maxTokens: number;
3960
5235
  };
3961
- readonly "sao10k/l3-euryale-70b": {
5236
+ readonly "meta-llama/llama-3-8b-instruct": {
3962
5237
  id: string;
3963
5238
  name: string;
3964
5239
  api: "openai-completions";
@@ -3975,7 +5250,7 @@ export declare const MODELS: {
3975
5250
  contextWindow: number;
3976
5251
  maxTokens: number;
3977
5252
  };
3978
- readonly "mistralai/mistral-7b-instruct:free": {
5253
+ readonly "mistralai/mixtral-8x22b-instruct": {
3979
5254
  id: string;
3980
5255
  name: string;
3981
5256
  api: "openai-completions";
@@ -3992,7 +5267,58 @@ export declare const MODELS: {
3992
5267
  contextWindow: number;
3993
5268
  maxTokens: number;
3994
5269
  };
3995
- readonly "mistralai/mistral-7b-instruct": {
5270
+ readonly "openai/gpt-4-turbo": {
5271
+ id: string;
5272
+ name: string;
5273
+ api: "openai-completions";
5274
+ provider: string;
5275
+ baseUrl: string;
5276
+ reasoning: false;
5277
+ input: ("image" | "text")[];
5278
+ cost: {
5279
+ input: number;
5280
+ output: number;
5281
+ cacheRead: number;
5282
+ cacheWrite: number;
5283
+ };
5284
+ contextWindow: number;
5285
+ maxTokens: number;
5286
+ };
5287
+ readonly "anthropic/claude-3-haiku": {
5288
+ id: string;
5289
+ name: string;
5290
+ api: "openai-completions";
5291
+ provider: string;
5292
+ baseUrl: string;
5293
+ reasoning: false;
5294
+ input: ("image" | "text")[];
5295
+ cost: {
5296
+ input: number;
5297
+ output: number;
5298
+ cacheRead: number;
5299
+ cacheWrite: number;
5300
+ };
5301
+ contextWindow: number;
5302
+ maxTokens: number;
5303
+ };
5304
+ readonly "anthropic/claude-3-opus": {
5305
+ id: string;
5306
+ name: string;
5307
+ api: "openai-completions";
5308
+ provider: string;
5309
+ baseUrl: string;
5310
+ reasoning: false;
5311
+ input: ("image" | "text")[];
5312
+ cost: {
5313
+ input: number;
5314
+ output: number;
5315
+ cacheRead: number;
5316
+ cacheWrite: number;
5317
+ };
5318
+ contextWindow: number;
5319
+ maxTokens: number;
5320
+ };
5321
+ readonly "mistralai/mistral-large": {
3996
5322
  id: string;
3997
5323
  name: string;
3998
5324
  api: "openai-completions";
@@ -4009,7 +5335,7 @@ export declare const MODELS: {
4009
5335
  contextWindow: number;
4010
5336
  maxTokens: number;
4011
5337
  };
4012
- readonly "mistralai/mistral-7b-instruct-v0.3": {
5338
+ readonly "openai/gpt-3.5-turbo-0613": {
4013
5339
  id: string;
4014
5340
  name: string;
4015
5341
  api: "openai-completions";
@@ -4026,7 +5352,7 @@ export declare const MODELS: {
4026
5352
  contextWindow: number;
4027
5353
  maxTokens: number;
4028
5354
  };
4029
- readonly "microsoft/phi-3-mini-128k-instruct": {
5355
+ readonly "openai/gpt-4-turbo-preview": {
4030
5356
  id: string;
4031
5357
  name: string;
4032
5358
  api: "openai-completions";
@@ -4043,7 +5369,7 @@ export declare const MODELS: {
4043
5369
  contextWindow: number;
4044
5370
  maxTokens: number;
4045
5371
  };
4046
- readonly "microsoft/phi-3-medium-128k-instruct": {
5372
+ readonly "mistralai/mistral-small": {
4047
5373
  id: string;
4048
5374
  name: string;
4049
5375
  api: "openai-completions";
@@ -4060,7 +5386,7 @@ export declare const MODELS: {
4060
5386
  contextWindow: number;
4061
5387
  maxTokens: number;
4062
5388
  };
4063
- readonly "meta-llama/llama-3-8b-instruct": {
5389
+ readonly "mistralai/mistral-tiny": {
4064
5390
  id: string;
4065
5391
  name: string;
4066
5392
  api: "openai-completions";
@@ -4077,7 +5403,7 @@ export declare const MODELS: {
4077
5403
  contextWindow: number;
4078
5404
  maxTokens: number;
4079
5405
  };
4080
- readonly "meta-llama/llama-3-70b-instruct": {
5406
+ readonly "mistralai/mixtral-8x7b-instruct": {
4081
5407
  id: string;
4082
5408
  name: string;
4083
5409
  api: "openai-completions";
@@ -4094,7 +5420,7 @@ export declare const MODELS: {
4094
5420
  contextWindow: number;
4095
5421
  maxTokens: number;
4096
5422
  };
4097
- readonly "mistralai/mixtral-8x22b-instruct": {
5423
+ readonly "openai/gpt-4-1106-preview": {
4098
5424
  id: string;
4099
5425
  name: string;
4100
5426
  api: "openai-completions";
@@ -4111,7 +5437,7 @@ export declare const MODELS: {
4111
5437
  contextWindow: number;
4112
5438
  maxTokens: number;
4113
5439
  };
4114
- readonly "mistralai/mistral-large": {
5440
+ readonly "mistralai/mistral-7b-instruct-v0.1": {
4115
5441
  id: string;
4116
5442
  name: string;
4117
5443
  api: "openai-completions";
@@ -4128,7 +5454,7 @@ export declare const MODELS: {
4128
5454
  contextWindow: number;
4129
5455
  maxTokens: number;
4130
5456
  };
4131
- readonly "mistralai/mistral-small": {
5457
+ readonly "openai/gpt-3.5-turbo-16k": {
4132
5458
  id: string;
4133
5459
  name: string;
4134
5460
  api: "openai-completions";
@@ -4145,7 +5471,7 @@ export declare const MODELS: {
4145
5471
  contextWindow: number;
4146
5472
  maxTokens: number;
4147
5473
  };
4148
- readonly "mistralai/mistral-tiny": {
5474
+ readonly "openai/gpt-4-0314": {
4149
5475
  id: string;
4150
5476
  name: string;
4151
5477
  api: "openai-completions";
@@ -4162,7 +5488,7 @@ export declare const MODELS: {
4162
5488
  contextWindow: number;
4163
5489
  maxTokens: number;
4164
5490
  };
4165
- readonly "mistralai/mixtral-8x7b-instruct": {
5491
+ readonly "openai/gpt-4": {
4166
5492
  id: string;
4167
5493
  name: string;
4168
5494
  api: "openai-completions";
@@ -4179,7 +5505,7 @@ export declare const MODELS: {
4179
5505
  contextWindow: number;
4180
5506
  maxTokens: number;
4181
5507
  };
4182
- readonly "mistralai/mistral-7b-instruct-v0.1": {
5508
+ readonly "openai/gpt-3.5-turbo": {
4183
5509
  id: string;
4184
5510
  name: string;
4185
5511
  api: "openai-completions";