@yeshwanthyk/ai 0.1.6 → 0.2.6

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -68,6 +68,23 @@ export declare const MODELS: {
68
68
  contextWindow: number;
69
69
  maxTokens: number;
70
70
  };
71
+ readonly "claude-sonnet-4-6": {
72
+ id: string;
73
+ name: string;
74
+ api: "anthropic-messages";
75
+ provider: string;
76
+ baseUrl: string;
77
+ reasoning: true;
78
+ input: ("image" | "text")[];
79
+ cost: {
80
+ input: number;
81
+ output: number;
82
+ cacheRead: number;
83
+ cacheWrite: number;
84
+ };
85
+ contextWindow: number;
86
+ maxTokens: number;
87
+ };
71
88
  readonly "claude-3-5-sonnet-20240620": {
72
89
  id: string;
73
90
  name: string;
@@ -85,6 +102,23 @@ export declare const MODELS: {
85
102
  contextWindow: number;
86
103
  maxTokens: number;
87
104
  };
105
+ readonly "claude-opus-4-6": {
106
+ id: string;
107
+ name: string;
108
+ api: "anthropic-messages";
109
+ provider: string;
110
+ baseUrl: string;
111
+ reasoning: true;
112
+ input: ("image" | "text")[];
113
+ cost: {
114
+ input: number;
115
+ output: number;
116
+ cacheRead: number;
117
+ cacheWrite: number;
118
+ };
119
+ contextWindow: number;
120
+ maxTokens: number;
121
+ };
88
122
  readonly "claude-3-5-haiku-latest": {
89
123
  id: string;
90
124
  name: string;
@@ -512,6 +546,23 @@ export declare const MODELS: {
512
546
  contextWindow: number;
513
547
  maxTokens: number;
514
548
  };
549
+ readonly "gemini-3.1-pro-preview-customtools": {
550
+ id: string;
551
+ name: string;
552
+ api: "google-generative-ai";
553
+ provider: string;
554
+ baseUrl: string;
555
+ reasoning: true;
556
+ input: ("image" | "text")[];
557
+ cost: {
558
+ input: number;
559
+ output: number;
560
+ cacheRead: number;
561
+ cacheWrite: number;
562
+ };
563
+ contextWindow: number;
564
+ maxTokens: number;
565
+ };
515
566
  readonly "gemini-2.0-flash": {
516
567
  id: string;
517
568
  name: string;
@@ -648,6 +699,23 @@ export declare const MODELS: {
648
699
  contextWindow: number;
649
700
  maxTokens: number;
650
701
  };
702
+ readonly "gemini-3.1-pro-preview": {
703
+ id: string;
704
+ name: string;
705
+ api: "google-generative-ai";
706
+ provider: string;
707
+ baseUrl: string;
708
+ reasoning: true;
709
+ input: ("image" | "text")[];
710
+ cost: {
711
+ input: number;
712
+ output: number;
713
+ cacheRead: number;
714
+ cacheWrite: number;
715
+ };
716
+ contextWindow: number;
717
+ maxTokens: number;
718
+ };
651
719
  readonly "gemini-1.5-flash": {
652
720
  id: string;
653
721
  name: string;
@@ -735,6 +803,23 @@ export declare const MODELS: {
735
803
  contextWindow: number;
736
804
  maxTokens: number;
737
805
  };
806
+ readonly "gpt-5.3-codex-spark": {
807
+ id: string;
808
+ name: string;
809
+ api: "openai-responses";
810
+ provider: string;
811
+ baseUrl: string;
812
+ reasoning: true;
813
+ input: ("image" | "text")[];
814
+ cost: {
815
+ input: number;
816
+ output: number;
817
+ cacheRead: number;
818
+ cacheWrite: number;
819
+ };
820
+ contextWindow: number;
821
+ maxTokens: number;
822
+ };
738
823
  readonly "gpt-4": {
739
824
  id: string;
740
825
  name: string;
@@ -1092,6 +1177,23 @@ export declare const MODELS: {
1092
1177
  contextWindow: number;
1093
1178
  maxTokens: number;
1094
1179
  };
1180
+ readonly "gpt-5.3-codex": {
1181
+ id: string;
1182
+ name: string;
1183
+ api: "openai-responses";
1184
+ provider: string;
1185
+ baseUrl: string;
1186
+ reasoning: true;
1187
+ input: ("image" | "text")[];
1188
+ cost: {
1189
+ input: number;
1190
+ output: number;
1191
+ cacheRead: number;
1192
+ cacheWrite: number;
1193
+ };
1194
+ contextWindow: number;
1195
+ maxTokens: number;
1196
+ };
1095
1197
  readonly "gpt-5-mini": {
1096
1198
  id: string;
1097
1199
  name: string;
@@ -1589,6 +1691,23 @@ export declare const MODELS: {
1589
1691
  contextWindow: number;
1590
1692
  maxTokens: number;
1591
1693
  };
1694
+ readonly "llama3.1-8b": {
1695
+ id: string;
1696
+ name: string;
1697
+ api: "openai-completions";
1698
+ provider: string;
1699
+ baseUrl: string;
1700
+ reasoning: false;
1701
+ input: "text"[];
1702
+ cost: {
1703
+ input: number;
1704
+ output: number;
1705
+ cacheRead: number;
1706
+ cacheWrite: number;
1707
+ };
1708
+ contextWindow: number;
1709
+ maxTokens: number;
1710
+ };
1592
1711
  readonly "gpt-oss-120b": {
1593
1712
  id: string;
1594
1713
  name: string;
@@ -2001,6 +2120,23 @@ export declare const MODELS: {
2001
2120
  contextWindow: number;
2002
2121
  maxTokens: number;
2003
2122
  };
2123
+ readonly "glm-5": {
2124
+ id: string;
2125
+ name: string;
2126
+ api: "anthropic-messages";
2127
+ provider: string;
2128
+ baseUrl: string;
2129
+ reasoning: true;
2130
+ input: "text"[];
2131
+ cost: {
2132
+ input: number;
2133
+ output: number;
2134
+ cacheRead: number;
2135
+ cacheWrite: number;
2136
+ };
2137
+ contextWindow: number;
2138
+ maxTokens: number;
2139
+ };
2004
2140
  readonly "glm-4.5-flash": {
2005
2141
  id: string;
2006
2142
  name: string;
@@ -2566,6 +2702,23 @@ export declare const MODELS: {
2566
2702
  contextWindow: number;
2567
2703
  maxTokens: number;
2568
2704
  };
2705
+ readonly "glm-5": {
2706
+ id: string;
2707
+ name: string;
2708
+ api: "openai-completions";
2709
+ provider: string;
2710
+ baseUrl: string;
2711
+ reasoning: true;
2712
+ input: "text"[];
2713
+ cost: {
2714
+ input: number;
2715
+ output: number;
2716
+ cacheRead: number;
2717
+ cacheWrite: number;
2718
+ };
2719
+ contextWindow: number;
2720
+ maxTokens: number;
2721
+ };
2569
2722
  readonly "qwen3-coder": {
2570
2723
  id: string;
2571
2724
  name: string;
@@ -2668,6 +2821,40 @@ export declare const MODELS: {
2668
2821
  contextWindow: number;
2669
2822
  maxTokens: number;
2670
2823
  };
2824
+ readonly "claude-sonnet-4-6": {
2825
+ id: string;
2826
+ name: string;
2827
+ api: "openai-completions";
2828
+ provider: string;
2829
+ baseUrl: string;
2830
+ reasoning: true;
2831
+ input: ("image" | "text")[];
2832
+ cost: {
2833
+ input: number;
2834
+ output: number;
2835
+ cacheRead: number;
2836
+ cacheWrite: number;
2837
+ };
2838
+ contextWindow: number;
2839
+ maxTokens: number;
2840
+ };
2841
+ readonly "claude-opus-4-6": {
2842
+ id: string;
2843
+ name: string;
2844
+ api: "openai-completions";
2845
+ provider: string;
2846
+ baseUrl: string;
2847
+ reasoning: true;
2848
+ input: ("image" | "text")[];
2849
+ cost: {
2850
+ input: number;
2851
+ output: number;
2852
+ cacheRead: number;
2853
+ cacheWrite: number;
2854
+ };
2855
+ contextWindow: number;
2856
+ maxTokens: number;
2857
+ };
2671
2858
  readonly "trinity-large-preview-free": {
2672
2859
  id: string;
2673
2860
  name: string;
@@ -2787,6 +2974,40 @@ export declare const MODELS: {
2787
2974
  contextWindow: number;
2788
2975
  maxTokens: number;
2789
2976
  };
2977
+ readonly "glm-5-free": {
2978
+ id: string;
2979
+ name: string;
2980
+ api: "openai-completions";
2981
+ provider: string;
2982
+ baseUrl: string;
2983
+ reasoning: true;
2984
+ input: "text"[];
2985
+ cost: {
2986
+ input: number;
2987
+ output: number;
2988
+ cacheRead: number;
2989
+ cacheWrite: number;
2990
+ };
2991
+ contextWindow: number;
2992
+ maxTokens: number;
2993
+ };
2994
+ readonly "gemini-3.1-pro": {
2995
+ id: string;
2996
+ name: string;
2997
+ api: "openai-completions";
2998
+ provider: string;
2999
+ baseUrl: string;
3000
+ reasoning: true;
3001
+ input: ("image" | "text")[];
3002
+ cost: {
3003
+ input: number;
3004
+ output: number;
3005
+ cacheRead: number;
3006
+ cacheWrite: number;
3007
+ };
3008
+ contextWindow: number;
3009
+ maxTokens: number;
3010
+ };
2790
3011
  readonly "gpt-5.1": {
2791
3012
  id: string;
2792
3013
  name: string;
@@ -3049,6 +3270,40 @@ export declare const MODELS: {
3049
3270
  contextWindow: number;
3050
3271
  maxTokens: number;
3051
3272
  };
3273
+ readonly "minimax-m2.5-free": {
3274
+ id: string;
3275
+ name: string;
3276
+ api: "openai-completions";
3277
+ provider: string;
3278
+ baseUrl: string;
3279
+ reasoning: true;
3280
+ input: "text"[];
3281
+ cost: {
3282
+ input: number;
3283
+ output: number;
3284
+ cacheRead: number;
3285
+ cacheWrite: number;
3286
+ };
3287
+ contextWindow: number;
3288
+ maxTokens: number;
3289
+ };
3290
+ readonly "minimax-m2.5": {
3291
+ id: string;
3292
+ name: string;
3293
+ api: "openai-completions";
3294
+ provider: string;
3295
+ baseUrl: string;
3296
+ reasoning: true;
3297
+ input: "text"[];
3298
+ cost: {
3299
+ input: number;
3300
+ output: number;
3301
+ cacheRead: number;
3302
+ cacheWrite: number;
3303
+ };
3304
+ contextWindow: number;
3305
+ maxTokens: number;
3306
+ };
3052
3307
  readonly "gpt-5.2": {
3053
3308
  id: string;
3054
3309
  name: string;
@@ -3351,7 +3606,7 @@ export declare const MODELS: {
3351
3606
  contextWindow: number;
3352
3607
  maxTokens: number;
3353
3608
  };
3354
- readonly "gemini-2.5-pro": {
3609
+ readonly "claude-sonnet-4.6": {
3355
3610
  id: string;
3356
3611
  name: string;
3357
3612
  api: "openai-completions";
@@ -3368,7 +3623,7 @@ export declare const MODELS: {
3368
3623
  supportsDeveloperRole: false;
3369
3624
  supportsReasoningEffort: false;
3370
3625
  };
3371
- reasoning: false;
3626
+ reasoning: true;
3372
3627
  input: ("image" | "text")[];
3373
3628
  cost: {
3374
3629
  input: number;
@@ -3379,10 +3634,10 @@ export declare const MODELS: {
3379
3634
  contextWindow: number;
3380
3635
  maxTokens: number;
3381
3636
  };
3382
- readonly "gpt-5.1-codex-max": {
3637
+ readonly "gemini-2.5-pro": {
3383
3638
  id: string;
3384
3639
  name: string;
3385
- api: "openai-responses";
3640
+ api: "openai-completions";
3386
3641
  provider: string;
3387
3642
  baseUrl: string;
3388
3643
  headers: {
@@ -3391,7 +3646,12 @@ export declare const MODELS: {
3391
3646
  "Editor-Plugin-Version": string;
3392
3647
  "Copilot-Integration-Id": string;
3393
3648
  };
3394
- reasoning: true;
3649
+ compat: {
3650
+ supportsStore: false;
3651
+ supportsDeveloperRole: false;
3652
+ supportsReasoningEffort: false;
3653
+ };
3654
+ reasoning: false;
3395
3655
  input: ("image" | "text")[];
3396
3656
  cost: {
3397
3657
  input: number;
@@ -3402,7 +3662,58 @@ export declare const MODELS: {
3402
3662
  contextWindow: number;
3403
3663
  maxTokens: number;
3404
3664
  };
3405
- readonly "claude-sonnet-4": {
3665
+ readonly "gpt-5.1-codex-max": {
3666
+ id: string;
3667
+ name: string;
3668
+ api: "openai-responses";
3669
+ provider: string;
3670
+ baseUrl: string;
3671
+ headers: {
3672
+ "User-Agent": string;
3673
+ "Editor-Version": string;
3674
+ "Editor-Plugin-Version": string;
3675
+ "Copilot-Integration-Id": string;
3676
+ };
3677
+ reasoning: true;
3678
+ input: ("image" | "text")[];
3679
+ cost: {
3680
+ input: number;
3681
+ output: number;
3682
+ cacheRead: number;
3683
+ cacheWrite: number;
3684
+ };
3685
+ contextWindow: number;
3686
+ maxTokens: number;
3687
+ };
3688
+ readonly "gemini-3.1-pro-preview": {
3689
+ id: string;
3690
+ name: string;
3691
+ api: "openai-completions";
3692
+ provider: string;
3693
+ baseUrl: string;
3694
+ headers: {
3695
+ "User-Agent": string;
3696
+ "Editor-Version": string;
3697
+ "Editor-Plugin-Version": string;
3698
+ "Copilot-Integration-Id": string;
3699
+ };
3700
+ compat: {
3701
+ supportsStore: false;
3702
+ supportsDeveloperRole: false;
3703
+ supportsReasoningEffort: false;
3704
+ };
3705
+ reasoning: true;
3706
+ input: ("image" | "text")[];
3707
+ cost: {
3708
+ input: number;
3709
+ output: number;
3710
+ cacheRead: number;
3711
+ cacheWrite: number;
3712
+ };
3713
+ contextWindow: number;
3714
+ maxTokens: number;
3715
+ };
3716
+ readonly "claude-sonnet-4": {
3406
3717
  id: string;
3407
3718
  name: string;
3408
3719
  api: "openai-completions";
@@ -3504,25 +3815,259 @@ export declare const MODELS: {
3504
3815
  contextWindow: number;
3505
3816
  maxTokens: number;
3506
3817
  };
3507
- readonly "claude-sonnet-4.5": {
3818
+ readonly "claude-sonnet-4.5": {
3819
+ id: string;
3820
+ name: string;
3821
+ api: "openai-completions";
3822
+ provider: string;
3823
+ baseUrl: string;
3824
+ headers: {
3825
+ "User-Agent": string;
3826
+ "Editor-Version": string;
3827
+ "Editor-Plugin-Version": string;
3828
+ "Copilot-Integration-Id": string;
3829
+ };
3830
+ compat: {
3831
+ supportsStore: false;
3832
+ supportsDeveloperRole: false;
3833
+ supportsReasoningEffort: false;
3834
+ };
3835
+ reasoning: true;
3836
+ input: ("image" | "text")[];
3837
+ cost: {
3838
+ input: number;
3839
+ output: number;
3840
+ cacheRead: number;
3841
+ cacheWrite: number;
3842
+ };
3843
+ contextWindow: number;
3844
+ maxTokens: number;
3845
+ };
3846
+ readonly "claude-opus-4.6": {
3847
+ id: string;
3848
+ name: string;
3849
+ api: "openai-completions";
3850
+ provider: string;
3851
+ baseUrl: string;
3852
+ headers: {
3853
+ "User-Agent": string;
3854
+ "Editor-Version": string;
3855
+ "Editor-Plugin-Version": string;
3856
+ "Copilot-Integration-Id": string;
3857
+ };
3858
+ compat: {
3859
+ supportsStore: false;
3860
+ supportsDeveloperRole: false;
3861
+ supportsReasoningEffort: false;
3862
+ };
3863
+ reasoning: true;
3864
+ input: ("image" | "text")[];
3865
+ cost: {
3866
+ input: number;
3867
+ output: number;
3868
+ cacheRead: number;
3869
+ cacheWrite: number;
3870
+ };
3871
+ contextWindow: number;
3872
+ maxTokens: number;
3873
+ };
3874
+ };
3875
+ readonly openrouter: {
3876
+ readonly "google/gemini-3.1-pro-preview": {
3877
+ id: string;
3878
+ name: string;
3879
+ api: "openai-completions";
3880
+ provider: string;
3881
+ baseUrl: string;
3882
+ reasoning: true;
3883
+ input: ("image" | "text")[];
3884
+ cost: {
3885
+ input: number;
3886
+ output: number;
3887
+ cacheRead: number;
3888
+ cacheWrite: number;
3889
+ };
3890
+ contextWindow: number;
3891
+ maxTokens: number;
3892
+ };
3893
+ readonly "anthropic/claude-sonnet-4.6": {
3894
+ id: string;
3895
+ name: string;
3896
+ api: "openai-completions";
3897
+ provider: string;
3898
+ baseUrl: string;
3899
+ reasoning: true;
3900
+ input: ("image" | "text")[];
3901
+ cost: {
3902
+ input: number;
3903
+ output: number;
3904
+ cacheRead: number;
3905
+ cacheWrite: number;
3906
+ };
3907
+ contextWindow: number;
3908
+ maxTokens: number;
3909
+ };
3910
+ readonly "qwen/qwen3.5-plus-02-15": {
3911
+ id: string;
3912
+ name: string;
3913
+ api: "openai-completions";
3914
+ provider: string;
3915
+ baseUrl: string;
3916
+ reasoning: true;
3917
+ input: ("image" | "text")[];
3918
+ cost: {
3919
+ input: number;
3920
+ output: number;
3921
+ cacheRead: number;
3922
+ cacheWrite: number;
3923
+ };
3924
+ contextWindow: number;
3925
+ maxTokens: number;
3926
+ };
3927
+ readonly "qwen/qwen3.5-397b-a17b": {
3928
+ id: string;
3929
+ name: string;
3930
+ api: "openai-completions";
3931
+ provider: string;
3932
+ baseUrl: string;
3933
+ reasoning: true;
3934
+ input: ("image" | "text")[];
3935
+ cost: {
3936
+ input: number;
3937
+ output: number;
3938
+ cacheRead: number;
3939
+ cacheWrite: number;
3940
+ };
3941
+ contextWindow: number;
3942
+ maxTokens: number;
3943
+ };
3944
+ readonly "minimax/minimax-m2.5": {
3945
+ id: string;
3946
+ name: string;
3947
+ api: "openai-completions";
3948
+ provider: string;
3949
+ baseUrl: string;
3950
+ reasoning: true;
3951
+ input: "text"[];
3952
+ cost: {
3953
+ input: number;
3954
+ output: number;
3955
+ cacheRead: number;
3956
+ cacheWrite: number;
3957
+ };
3958
+ contextWindow: number;
3959
+ maxTokens: number;
3960
+ };
3961
+ readonly "z-ai/glm-5": {
3962
+ id: string;
3963
+ name: string;
3964
+ api: "openai-completions";
3965
+ provider: string;
3966
+ baseUrl: string;
3967
+ reasoning: true;
3968
+ input: "text"[];
3969
+ cost: {
3970
+ input: number;
3971
+ output: number;
3972
+ cacheRead: number;
3973
+ cacheWrite: number;
3974
+ };
3975
+ contextWindow: number;
3976
+ maxTokens: number;
3977
+ };
3978
+ readonly "qwen/qwen3-max-thinking": {
3979
+ id: string;
3980
+ name: string;
3981
+ api: "openai-completions";
3982
+ provider: string;
3983
+ baseUrl: string;
3984
+ reasoning: true;
3985
+ input: "text"[];
3986
+ cost: {
3987
+ input: number;
3988
+ output: number;
3989
+ cacheRead: number;
3990
+ cacheWrite: number;
3991
+ };
3992
+ contextWindow: number;
3993
+ maxTokens: number;
3994
+ };
3995
+ readonly "anthropic/claude-opus-4.6": {
3996
+ id: string;
3997
+ name: string;
3998
+ api: "openai-completions";
3999
+ provider: string;
4000
+ baseUrl: string;
4001
+ reasoning: true;
4002
+ input: ("image" | "text")[];
4003
+ cost: {
4004
+ input: number;
4005
+ output: number;
4006
+ cacheRead: number;
4007
+ cacheWrite: number;
4008
+ };
4009
+ contextWindow: number;
4010
+ maxTokens: number;
4011
+ };
4012
+ readonly "qwen/qwen3-coder-next": {
4013
+ id: string;
4014
+ name: string;
4015
+ api: "openai-completions";
4016
+ provider: string;
4017
+ baseUrl: string;
4018
+ reasoning: false;
4019
+ input: "text"[];
4020
+ cost: {
4021
+ input: number;
4022
+ output: number;
4023
+ cacheRead: number;
4024
+ cacheWrite: number;
4025
+ };
4026
+ contextWindow: number;
4027
+ maxTokens: number;
4028
+ };
4029
+ readonly "openrouter/free": {
4030
+ id: string;
4031
+ name: string;
4032
+ api: "openai-completions";
4033
+ provider: string;
4034
+ baseUrl: string;
4035
+ reasoning: true;
4036
+ input: ("image" | "text")[];
4037
+ cost: {
4038
+ input: number;
4039
+ output: number;
4040
+ cacheRead: number;
4041
+ cacheWrite: number;
4042
+ };
4043
+ contextWindow: number;
4044
+ maxTokens: number;
4045
+ };
4046
+ readonly "stepfun/step-3.5-flash:free": {
4047
+ id: string;
4048
+ name: string;
4049
+ api: "openai-completions";
4050
+ provider: string;
4051
+ baseUrl: string;
4052
+ reasoning: true;
4053
+ input: "text"[];
4054
+ cost: {
4055
+ input: number;
4056
+ output: number;
4057
+ cacheRead: number;
4058
+ cacheWrite: number;
4059
+ };
4060
+ contextWindow: number;
4061
+ maxTokens: number;
4062
+ };
4063
+ readonly "stepfun/step-3.5-flash": {
3508
4064
  id: string;
3509
4065
  name: string;
3510
4066
  api: "openai-completions";
3511
4067
  provider: string;
3512
4068
  baseUrl: string;
3513
- headers: {
3514
- "User-Agent": string;
3515
- "Editor-Version": string;
3516
- "Editor-Plugin-Version": string;
3517
- "Copilot-Integration-Id": string;
3518
- };
3519
- compat: {
3520
- supportsStore: false;
3521
- supportsDeveloperRole: false;
3522
- supportsReasoningEffort: false;
3523
- };
3524
4069
  reasoning: true;
3525
- input: ("image" | "text")[];
4070
+ input: "text"[];
3526
4071
  cost: {
3527
4072
  input: number;
3528
4073
  output: number;
@@ -3532,8 +4077,6 @@ export declare const MODELS: {
3532
4077
  contextWindow: number;
3533
4078
  maxTokens: number;
3534
4079
  };
3535
- };
3536
- readonly openrouter: {
3537
4080
  readonly "arcee-ai/trinity-large-preview:free": {
3538
4081
  id: string;
3539
4082
  name: string;
@@ -4078,40 +4621,6 @@ export declare const MODELS: {
4078
4621
  contextWindow: number;
4079
4622
  maxTokens: number;
4080
4623
  };
4081
- readonly "tngtech/tng-r1t-chimera:free": {
4082
- id: string;
4083
- name: string;
4084
- api: "openai-completions";
4085
- provider: string;
4086
- baseUrl: string;
4087
- reasoning: true;
4088
- input: "text"[];
4089
- cost: {
4090
- input: number;
4091
- output: number;
4092
- cacheRead: number;
4093
- cacheWrite: number;
4094
- };
4095
- contextWindow: number;
4096
- maxTokens: number;
4097
- };
4098
- readonly "tngtech/tng-r1t-chimera": {
4099
- id: string;
4100
- name: string;
4101
- api: "openai-completions";
4102
- provider: string;
4103
- baseUrl: string;
4104
- reasoning: true;
4105
- input: "text"[];
4106
- cost: {
4107
- input: number;
4108
- output: number;
4109
- cacheRead: number;
4110
- cacheWrite: number;
4111
- };
4112
- contextWindow: number;
4113
- maxTokens: number;
4114
- };
4115
4624
  readonly "anthropic/claude-opus-4.5": {
4116
4625
  id: string;
4117
4626
  name: string;
@@ -4350,14 +4859,14 @@ export declare const MODELS: {
4350
4859
  contextWindow: number;
4351
4860
  maxTokens: number;
4352
4861
  };
4353
- readonly "deepcogito/cogito-v2-preview-llama-405b": {
4862
+ readonly "qwen/qwen3-vl-32b-instruct": {
4354
4863
  id: string;
4355
4864
  name: string;
4356
4865
  api: "openai-completions";
4357
4866
  provider: string;
4358
4867
  baseUrl: string;
4359
- reasoning: true;
4360
- input: "text"[];
4868
+ reasoning: false;
4869
+ input: ("image" | "text")[];
4361
4870
  cost: {
4362
4871
  input: number;
4363
4872
  output: number;
@@ -4622,23 +5131,6 @@ export declare const MODELS: {
4622
5131
  contextWindow: number;
4623
5132
  maxTokens: number;
4624
5133
  };
4625
- readonly "google/gemini-2.5-flash-preview-09-2025": {
4626
- id: string;
4627
- name: string;
4628
- api: "openai-completions";
4629
- provider: string;
4630
- baseUrl: string;
4631
- reasoning: true;
4632
- input: ("image" | "text")[];
4633
- cost: {
4634
- input: number;
4635
- output: number;
4636
- cacheRead: number;
4637
- cacheWrite: number;
4638
- };
4639
- contextWindow: number;
4640
- maxTokens: number;
4641
- };
4642
5134
  readonly "google/gemini-2.5-flash-lite-preview-09-2025": {
4643
5135
  id: string;
4644
5136
  name: string;
@@ -4979,57 +5471,6 @@ export declare const MODELS: {
4979
5471
  contextWindow: number;
4980
5472
  maxTokens: number;
4981
5473
  };
4982
- readonly "deepcogito/cogito-v2-preview-llama-70b": {
4983
- id: string;
4984
- name: string;
4985
- api: "openai-completions";
4986
- provider: string;
4987
- baseUrl: string;
4988
- reasoning: true;
4989
- input: "text"[];
4990
- cost: {
4991
- input: number;
4992
- output: number;
4993
- cacheRead: number;
4994
- cacheWrite: number;
4995
- };
4996
- contextWindow: number;
4997
- maxTokens: number;
4998
- };
4999
- readonly "deepcogito/cogito-v2-preview-llama-109b-moe": {
5000
- id: string;
5001
- name: string;
5002
- api: "openai-completions";
5003
- provider: string;
5004
- baseUrl: string;
5005
- reasoning: true;
5006
- input: ("image" | "text")[];
5007
- cost: {
5008
- input: number;
5009
- output: number;
5010
- cacheRead: number;
5011
- cacheWrite: number;
5012
- };
5013
- contextWindow: number;
5014
- maxTokens: number;
5015
- };
5016
- readonly "stepfun-ai/step3": {
5017
- id: string;
5018
- name: string;
5019
- api: "openai-completions";
5020
- provider: string;
5021
- baseUrl: string;
5022
- reasoning: true;
5023
- input: ("image" | "text")[];
5024
- cost: {
5025
- input: number;
5026
- output: number;
5027
- cacheRead: number;
5028
- cacheWrite: number;
5029
- };
5030
- contextWindow: number;
5031
- maxTokens: number;
5032
- };
5033
5474
  readonly "qwen/qwen3-30b-a3b-thinking-2507": {
5034
5475
  id: string;
5035
5476
  name: string;
@@ -5064,23 +5505,6 @@ export declare const MODELS: {
5064
5505
  contextWindow: number;
5065
5506
  maxTokens: number;
5066
5507
  };
5067
- readonly "nousresearch/hermes-4-70b": {
5068
- id: string;
5069
- name: string;
5070
- api: "openai-completions";
5071
- provider: string;
5072
- baseUrl: string;
5073
- reasoning: true;
5074
- input: "text"[];
5075
- cost: {
5076
- input: number;
5077
- output: number;
5078
- cacheRead: number;
5079
- cacheWrite: number;
5080
- };
5081
- contextWindow: number;
5082
- maxTokens: number;
5083
- };
5084
5508
  readonly "deepseek/deepseek-chat-v3.1": {
5085
5509
  id: string;
5086
5510
  name: string;
@@ -5183,23 +5607,6 @@ export declare const MODELS: {
5183
5607
  contextWindow: number;
5184
5608
  maxTokens: number;
5185
5609
  };
5186
- readonly "ai21/jamba-mini-1.7": {
5187
- id: string;
5188
- name: string;
5189
- api: "openai-completions";
5190
- provider: string;
5191
- baseUrl: string;
5192
- reasoning: false;
5193
- input: "text"[];
5194
- cost: {
5195
- input: number;
5196
- output: number;
5197
- cacheRead: number;
5198
- cacheWrite: number;
5199
- };
5200
- contextWindow: number;
5201
- maxTokens: number;
5202
- };
5203
5610
  readonly "ai21/jamba-large-1.7": {
5204
5611
  id: string;
5205
5612
  name: string;
@@ -5529,7 +5936,7 @@ export declare const MODELS: {
5529
5936
  api: "openai-completions";
5530
5937
  provider: string;
5531
5938
  baseUrl: string;
5532
- reasoning: true;
5939
+ reasoning: false;
5533
5940
  input: "text"[];
5534
5941
  cost: {
5535
5942
  input: number;
@@ -5546,7 +5953,7 @@ export declare const MODELS: {
5546
5953
  api: "openai-completions";
5547
5954
  provider: string;
5548
5955
  baseUrl: string;
5549
- reasoning: true;
5956
+ reasoning: false;
5550
5957
  input: "text"[];
5551
5958
  cost: {
5552
5959
  input: number;
@@ -5880,23 +6287,6 @@ export declare const MODELS: {
5880
6287
  contextWindow: number;
5881
6288
  maxTokens: number;
5882
6289
  };
5883
- readonly "nousresearch/deephermes-3-mistral-24b-preview": {
5884
- id: string;
5885
- name: string;
5886
- api: "openai-completions";
5887
- provider: string;
5888
- baseUrl: string;
5889
- reasoning: true;
5890
- input: "text"[];
5891
- cost: {
5892
- input: number;
5893
- output: number;
5894
- cacheRead: number;
5895
- cacheWrite: number;
5896
- };
5897
- contextWindow: number;
5898
- maxTokens: number;
5899
- };
5900
6290
  readonly "mistralai/mistral-medium-3": {
5901
6291
  id: string;
5902
6292
  name: string;
@@ -6271,23 +6661,6 @@ export declare const MODELS: {
6271
6661
  contextWindow: number;
6272
6662
  maxTokens: number;
6273
6663
  };
6274
- readonly "mistralai/mistral-small-3.1-24b-instruct": {
6275
- id: string;
6276
- name: string;
6277
- api: "openai-completions";
6278
- provider: string;
6279
- baseUrl: string;
6280
- reasoning: false;
6281
- input: ("image" | "text")[];
6282
- cost: {
6283
- input: number;
6284
- output: number;
6285
- cacheRead: number;
6286
- cacheWrite: number;
6287
- };
6288
- contextWindow: number;
6289
- maxTokens: number;
6290
- };
6291
6664
  readonly "google/gemma-3-27b-it:free": {
6292
6665
  id: string;
6293
6666
  name: string;
@@ -6356,7 +6729,7 @@ export declare const MODELS: {
6356
6729
  contextWindow: number;
6357
6730
  maxTokens: number;
6358
6731
  };
6359
- readonly "anthropic/claude-3.7-sonnet:thinking": {
6732
+ readonly "anthropic/claude-3.7-sonnet": {
6360
6733
  id: string;
6361
6734
  name: string;
6362
6735
  api: "openai-completions";
@@ -6373,7 +6746,7 @@ export declare const MODELS: {
6373
6746
  contextWindow: number;
6374
6747
  maxTokens: number;
6375
6748
  };
6376
- readonly "anthropic/claude-3.7-sonnet": {
6749
+ readonly "anthropic/claude-3.7-sonnet:thinking": {
6377
6750
  id: string;
6378
6751
  name: string;
6379
6752
  api: "openai-completions";
@@ -6798,40 +7171,6 @@ export declare const MODELS: {
6798
7171
  contextWindow: number;
6799
7172
  maxTokens: number;
6800
7173
  };
6801
- readonly "mistralai/ministral-8b": {
6802
- id: string;
6803
- name: string;
6804
- api: "openai-completions";
6805
- provider: string;
6806
- baseUrl: string;
6807
- reasoning: false;
6808
- input: "text"[];
6809
- cost: {
6810
- input: number;
6811
- output: number;
6812
- cacheRead: number;
6813
- cacheWrite: number;
6814
- };
6815
- contextWindow: number;
6816
- maxTokens: number;
6817
- };
6818
- readonly "mistralai/ministral-3b": {
6819
- id: string;
6820
- name: string;
6821
- api: "openai-completions";
6822
- provider: string;
6823
- baseUrl: string;
6824
- reasoning: false;
6825
- input: "text"[];
6826
- cost: {
6827
- input: number;
6828
- output: number;
6829
- cacheRead: number;
6830
- cacheWrite: number;
6831
- };
6832
- contextWindow: number;
6833
- maxTokens: number;
6834
- };
6835
7174
  readonly "qwen/qwen-2.5-7b-instruct": {
6836
7175
  id: string;
6837
7176
  name: string;
@@ -6900,23 +7239,6 @@ export declare const MODELS: {
6900
7239
  contextWindow: number;
6901
7240
  maxTokens: number;
6902
7241
  };
6903
- readonly "mistralai/pixtral-12b": {
6904
- id: string;
6905
- name: string;
6906
- api: "openai-completions";
6907
- provider: string;
6908
- baseUrl: string;
6909
- reasoning: false;
6910
- input: ("image" | "text")[];
6911
- cost: {
6912
- input: number;
6913
- output: number;
6914
- cacheRead: number;
6915
- cacheWrite: number;
6916
- };
6917
- contextWindow: number;
6918
- maxTokens: number;
6919
- };
6920
7242
  readonly "cohere/command-r-08-2024": {
6921
7243
  id: string;
6922
7244
  name: string;
@@ -7274,7 +7596,7 @@ export declare const MODELS: {
7274
7596
  contextWindow: number;
7275
7597
  maxTokens: number;
7276
7598
  };
7277
- readonly "mistralai/mistral-tiny": {
7599
+ readonly "mistralai/mixtral-8x7b-instruct": {
7278
7600
  id: string;
7279
7601
  name: string;
7280
7602
  api: "openai-completions";
@@ -7291,14 +7613,14 @@ export declare const MODELS: {
7291
7613
  contextWindow: number;
7292
7614
  maxTokens: number;
7293
7615
  };
7294
- readonly "mistralai/mixtral-8x7b-instruct": {
7616
+ readonly "openrouter/auto": {
7295
7617
  id: string;
7296
7618
  name: string;
7297
7619
  api: "openai-completions";
7298
7620
  provider: string;
7299
7621
  baseUrl: string;
7300
- reasoning: false;
7301
- input: "text"[];
7622
+ reasoning: true;
7623
+ input: ("image" | "text")[];
7302
7624
  cost: {
7303
7625
  input: number;
7304
7626
  output: number;
@@ -7393,10 +7715,12 @@ export declare const MODELS: {
7393
7715
  contextWindow: number;
7394
7716
  maxTokens: number;
7395
7717
  };
7396
- readonly "openrouter/auto": {
7718
+ };
7719
+ readonly codex: {
7720
+ readonly "gpt-5.2": {
7397
7721
  id: string;
7398
7722
  name: string;
7399
- api: "openai-completions";
7723
+ api: "openai-responses";
7400
7724
  provider: string;
7401
7725
  baseUrl: string;
7402
7726
  reasoning: true;
@@ -7410,9 +7734,7 @@ export declare const MODELS: {
7410
7734
  contextWindow: number;
7411
7735
  maxTokens: number;
7412
7736
  };
7413
- };
7414
- readonly codex: {
7415
- readonly "gpt-5.2": {
7737
+ readonly "gpt-5.2-codex": {
7416
7738
  id: string;
7417
7739
  name: string;
7418
7740
  api: "openai-responses";
@@ -7429,7 +7751,7 @@ export declare const MODELS: {
7429
7751
  contextWindow: number;
7430
7752
  maxTokens: number;
7431
7753
  };
7432
- readonly "gpt-5.2-codex": {
7754
+ readonly "gpt-5.2-mini": {
7433
7755
  id: string;
7434
7756
  name: string;
7435
7757
  api: "openai-responses";
@@ -7446,7 +7768,7 @@ export declare const MODELS: {
7446
7768
  contextWindow: number;
7447
7769
  maxTokens: number;
7448
7770
  };
7449
- readonly "gpt-5.2-mini": {
7771
+ readonly "gpt-5.3-codex": {
7450
7772
  id: string;
7451
7773
  name: string;
7452
7774
  api: "openai-responses";