@yeshwanthyk/ai 0.2.5 → 0.2.6

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -68,6 +68,23 @@ export declare const MODELS: {
68
68
  contextWindow: number;
69
69
  maxTokens: number;
70
70
  };
71
+ readonly "claude-sonnet-4-6": {
72
+ id: string;
73
+ name: string;
74
+ api: "anthropic-messages";
75
+ provider: string;
76
+ baseUrl: string;
77
+ reasoning: true;
78
+ input: ("image" | "text")[];
79
+ cost: {
80
+ input: number;
81
+ output: number;
82
+ cacheRead: number;
83
+ cacheWrite: number;
84
+ };
85
+ contextWindow: number;
86
+ maxTokens: number;
87
+ };
71
88
  readonly "claude-3-5-sonnet-20240620": {
72
89
  id: string;
73
90
  name: string;
@@ -529,6 +546,23 @@ export declare const MODELS: {
529
546
  contextWindow: number;
530
547
  maxTokens: number;
531
548
  };
549
+ readonly "gemini-3.1-pro-preview-customtools": {
550
+ id: string;
551
+ name: string;
552
+ api: "google-generative-ai";
553
+ provider: string;
554
+ baseUrl: string;
555
+ reasoning: true;
556
+ input: ("image" | "text")[];
557
+ cost: {
558
+ input: number;
559
+ output: number;
560
+ cacheRead: number;
561
+ cacheWrite: number;
562
+ };
563
+ contextWindow: number;
564
+ maxTokens: number;
565
+ };
532
566
  readonly "gemini-2.0-flash": {
533
567
  id: string;
534
568
  name: string;
@@ -665,6 +699,23 @@ export declare const MODELS: {
665
699
  contextWindow: number;
666
700
  maxTokens: number;
667
701
  };
702
+ readonly "gemini-3.1-pro-preview": {
703
+ id: string;
704
+ name: string;
705
+ api: "google-generative-ai";
706
+ provider: string;
707
+ baseUrl: string;
708
+ reasoning: true;
709
+ input: ("image" | "text")[];
710
+ cost: {
711
+ input: number;
712
+ output: number;
713
+ cacheRead: number;
714
+ cacheWrite: number;
715
+ };
716
+ contextWindow: number;
717
+ maxTokens: number;
718
+ };
668
719
  readonly "gemini-1.5-flash": {
669
720
  id: string;
670
721
  name: string;
@@ -752,6 +803,23 @@ export declare const MODELS: {
752
803
  contextWindow: number;
753
804
  maxTokens: number;
754
805
  };
806
+ readonly "gpt-5.3-codex-spark": {
807
+ id: string;
808
+ name: string;
809
+ api: "openai-responses";
810
+ provider: string;
811
+ baseUrl: string;
812
+ reasoning: true;
813
+ input: ("image" | "text")[];
814
+ cost: {
815
+ input: number;
816
+ output: number;
817
+ cacheRead: number;
818
+ cacheWrite: number;
819
+ };
820
+ contextWindow: number;
821
+ maxTokens: number;
822
+ };
755
823
  readonly "gpt-4": {
756
824
  id: string;
757
825
  name: string;
@@ -1623,6 +1691,23 @@ export declare const MODELS: {
1623
1691
  contextWindow: number;
1624
1692
  maxTokens: number;
1625
1693
  };
1694
+ readonly "llama3.1-8b": {
1695
+ id: string;
1696
+ name: string;
1697
+ api: "openai-completions";
1698
+ provider: string;
1699
+ baseUrl: string;
1700
+ reasoning: false;
1701
+ input: "text"[];
1702
+ cost: {
1703
+ input: number;
1704
+ output: number;
1705
+ cacheRead: number;
1706
+ cacheWrite: number;
1707
+ };
1708
+ contextWindow: number;
1709
+ maxTokens: number;
1710
+ };
1626
1711
  readonly "gpt-oss-120b": {
1627
1712
  id: string;
1628
1713
  name: string;
@@ -2035,6 +2120,23 @@ export declare const MODELS: {
2035
2120
  contextWindow: number;
2036
2121
  maxTokens: number;
2037
2122
  };
2123
+ readonly "glm-5": {
2124
+ id: string;
2125
+ name: string;
2126
+ api: "anthropic-messages";
2127
+ provider: string;
2128
+ baseUrl: string;
2129
+ reasoning: true;
2130
+ input: "text"[];
2131
+ cost: {
2132
+ input: number;
2133
+ output: number;
2134
+ cacheRead: number;
2135
+ cacheWrite: number;
2136
+ };
2137
+ contextWindow: number;
2138
+ maxTokens: number;
2139
+ };
2038
2140
  readonly "glm-4.5-flash": {
2039
2141
  id: string;
2040
2142
  name: string;
@@ -2600,6 +2702,23 @@ export declare const MODELS: {
2600
2702
  contextWindow: number;
2601
2703
  maxTokens: number;
2602
2704
  };
2705
+ readonly "glm-5": {
2706
+ id: string;
2707
+ name: string;
2708
+ api: "openai-completions";
2709
+ provider: string;
2710
+ baseUrl: string;
2711
+ reasoning: true;
2712
+ input: "text"[];
2713
+ cost: {
2714
+ input: number;
2715
+ output: number;
2716
+ cacheRead: number;
2717
+ cacheWrite: number;
2718
+ };
2719
+ contextWindow: number;
2720
+ maxTokens: number;
2721
+ };
2603
2722
  readonly "qwen3-coder": {
2604
2723
  id: string;
2605
2724
  name: string;
@@ -2702,6 +2821,23 @@ export declare const MODELS: {
2702
2821
  contextWindow: number;
2703
2822
  maxTokens: number;
2704
2823
  };
2824
+ readonly "claude-sonnet-4-6": {
2825
+ id: string;
2826
+ name: string;
2827
+ api: "openai-completions";
2828
+ provider: string;
2829
+ baseUrl: string;
2830
+ reasoning: true;
2831
+ input: ("image" | "text")[];
2832
+ cost: {
2833
+ input: number;
2834
+ output: number;
2835
+ cacheRead: number;
2836
+ cacheWrite: number;
2837
+ };
2838
+ contextWindow: number;
2839
+ maxTokens: number;
2840
+ };
2705
2841
  readonly "claude-opus-4-6": {
2706
2842
  id: string;
2707
2843
  name: string;
@@ -2838,6 +2974,40 @@ export declare const MODELS: {
2838
2974
  contextWindow: number;
2839
2975
  maxTokens: number;
2840
2976
  };
2977
+ readonly "glm-5-free": {
2978
+ id: string;
2979
+ name: string;
2980
+ api: "openai-completions";
2981
+ provider: string;
2982
+ baseUrl: string;
2983
+ reasoning: true;
2984
+ input: "text"[];
2985
+ cost: {
2986
+ input: number;
2987
+ output: number;
2988
+ cacheRead: number;
2989
+ cacheWrite: number;
2990
+ };
2991
+ contextWindow: number;
2992
+ maxTokens: number;
2993
+ };
2994
+ readonly "gemini-3.1-pro": {
2995
+ id: string;
2996
+ name: string;
2997
+ api: "openai-completions";
2998
+ provider: string;
2999
+ baseUrl: string;
3000
+ reasoning: true;
3001
+ input: ("image" | "text")[];
3002
+ cost: {
3003
+ input: number;
3004
+ output: number;
3005
+ cacheRead: number;
3006
+ cacheWrite: number;
3007
+ };
3008
+ contextWindow: number;
3009
+ maxTokens: number;
3010
+ };
2841
3011
  readonly "gpt-5.1": {
2842
3012
  id: string;
2843
3013
  name: string;
@@ -3100,6 +3270,40 @@ export declare const MODELS: {
3100
3270
  contextWindow: number;
3101
3271
  maxTokens: number;
3102
3272
  };
3273
+ readonly "minimax-m2.5-free": {
3274
+ id: string;
3275
+ name: string;
3276
+ api: "openai-completions";
3277
+ provider: string;
3278
+ baseUrl: string;
3279
+ reasoning: true;
3280
+ input: "text"[];
3281
+ cost: {
3282
+ input: number;
3283
+ output: number;
3284
+ cacheRead: number;
3285
+ cacheWrite: number;
3286
+ };
3287
+ contextWindow: number;
3288
+ maxTokens: number;
3289
+ };
3290
+ readonly "minimax-m2.5": {
3291
+ id: string;
3292
+ name: string;
3293
+ api: "openai-completions";
3294
+ provider: string;
3295
+ baseUrl: string;
3296
+ reasoning: true;
3297
+ input: "text"[];
3298
+ cost: {
3299
+ input: number;
3300
+ output: number;
3301
+ cacheRead: number;
3302
+ cacheWrite: number;
3303
+ };
3304
+ contextWindow: number;
3305
+ maxTokens: number;
3306
+ };
3103
3307
  readonly "gpt-5.2": {
3104
3308
  id: string;
3105
3309
  name: string;
@@ -3402,6 +3606,34 @@ export declare const MODELS: {
3402
3606
  contextWindow: number;
3403
3607
  maxTokens: number;
3404
3608
  };
3609
+ readonly "claude-sonnet-4.6": {
3610
+ id: string;
3611
+ name: string;
3612
+ api: "openai-completions";
3613
+ provider: string;
3614
+ baseUrl: string;
3615
+ headers: {
3616
+ "User-Agent": string;
3617
+ "Editor-Version": string;
3618
+ "Editor-Plugin-Version": string;
3619
+ "Copilot-Integration-Id": string;
3620
+ };
3621
+ compat: {
3622
+ supportsStore: false;
3623
+ supportsDeveloperRole: false;
3624
+ supportsReasoningEffort: false;
3625
+ };
3626
+ reasoning: true;
3627
+ input: ("image" | "text")[];
3628
+ cost: {
3629
+ input: number;
3630
+ output: number;
3631
+ cacheRead: number;
3632
+ cacheWrite: number;
3633
+ };
3634
+ contextWindow: number;
3635
+ maxTokens: number;
3636
+ };
3405
3637
  readonly "gemini-2.5-pro": {
3406
3638
  id: string;
3407
3639
  name: string;
@@ -3453,6 +3685,34 @@ export declare const MODELS: {
3453
3685
  contextWindow: number;
3454
3686
  maxTokens: number;
3455
3687
  };
3688
+ readonly "gemini-3.1-pro-preview": {
3689
+ id: string;
3690
+ name: string;
3691
+ api: "openai-completions";
3692
+ provider: string;
3693
+ baseUrl: string;
3694
+ headers: {
3695
+ "User-Agent": string;
3696
+ "Editor-Version": string;
3697
+ "Editor-Plugin-Version": string;
3698
+ "Copilot-Integration-Id": string;
3699
+ };
3700
+ compat: {
3701
+ supportsStore: false;
3702
+ supportsDeveloperRole: false;
3703
+ supportsReasoningEffort: false;
3704
+ };
3705
+ reasoning: true;
3706
+ input: ("image" | "text")[];
3707
+ cost: {
3708
+ input: number;
3709
+ output: number;
3710
+ cacheRead: number;
3711
+ cacheWrite: number;
3712
+ };
3713
+ contextWindow: number;
3714
+ maxTokens: number;
3715
+ };
3456
3716
  readonly "claude-sonnet-4": {
3457
3717
  id: string;
3458
3718
  name: string;
@@ -3613,14 +3873,14 @@ export declare const MODELS: {
3613
3873
  };
3614
3874
  };
3615
3875
  readonly openrouter: {
3616
- readonly "openrouter/pony-alpha": {
3876
+ readonly "google/gemini-3.1-pro-preview": {
3617
3877
  id: string;
3618
3878
  name: string;
3619
3879
  api: "openai-completions";
3620
3880
  provider: string;
3621
3881
  baseUrl: string;
3622
3882
  reasoning: true;
3623
- input: "text"[];
3883
+ input: ("image" | "text")[];
3624
3884
  cost: {
3625
3885
  input: number;
3626
3886
  output: number;
@@ -3630,7 +3890,7 @@ export declare const MODELS: {
3630
3890
  contextWindow: number;
3631
3891
  maxTokens: number;
3632
3892
  };
3633
- readonly "anthropic/claude-opus-4.6": {
3893
+ readonly "anthropic/claude-sonnet-4.6": {
3634
3894
  id: string;
3635
3895
  name: string;
3636
3896
  api: "openai-completions";
@@ -3647,14 +3907,14 @@ export declare const MODELS: {
3647
3907
  contextWindow: number;
3648
3908
  maxTokens: number;
3649
3909
  };
3650
- readonly "qwen/qwen3-coder-next": {
3910
+ readonly "qwen/qwen3.5-plus-02-15": {
3651
3911
  id: string;
3652
3912
  name: string;
3653
3913
  api: "openai-completions";
3654
3914
  provider: string;
3655
3915
  baseUrl: string;
3656
- reasoning: false;
3657
- input: "text"[];
3916
+ reasoning: true;
3917
+ input: ("image" | "text")[];
3658
3918
  cost: {
3659
3919
  input: number;
3660
3920
  output: number;
@@ -3664,7 +3924,7 @@ export declare const MODELS: {
3664
3924
  contextWindow: number;
3665
3925
  maxTokens: number;
3666
3926
  };
3667
- readonly "openrouter/free": {
3927
+ readonly "qwen/qwen3.5-397b-a17b": {
3668
3928
  id: string;
3669
3929
  name: string;
3670
3930
  api: "openai-completions";
@@ -3681,7 +3941,126 @@ export declare const MODELS: {
3681
3941
  contextWindow: number;
3682
3942
  maxTokens: number;
3683
3943
  };
3684
- readonly "stepfun/step-3.5-flash:free": {
3944
+ readonly "minimax/minimax-m2.5": {
3945
+ id: string;
3946
+ name: string;
3947
+ api: "openai-completions";
3948
+ provider: string;
3949
+ baseUrl: string;
3950
+ reasoning: true;
3951
+ input: "text"[];
3952
+ cost: {
3953
+ input: number;
3954
+ output: number;
3955
+ cacheRead: number;
3956
+ cacheWrite: number;
3957
+ };
3958
+ contextWindow: number;
3959
+ maxTokens: number;
3960
+ };
3961
+ readonly "z-ai/glm-5": {
3962
+ id: string;
3963
+ name: string;
3964
+ api: "openai-completions";
3965
+ provider: string;
3966
+ baseUrl: string;
3967
+ reasoning: true;
3968
+ input: "text"[];
3969
+ cost: {
3970
+ input: number;
3971
+ output: number;
3972
+ cacheRead: number;
3973
+ cacheWrite: number;
3974
+ };
3975
+ contextWindow: number;
3976
+ maxTokens: number;
3977
+ };
3978
+ readonly "qwen/qwen3-max-thinking": {
3979
+ id: string;
3980
+ name: string;
3981
+ api: "openai-completions";
3982
+ provider: string;
3983
+ baseUrl: string;
3984
+ reasoning: true;
3985
+ input: "text"[];
3986
+ cost: {
3987
+ input: number;
3988
+ output: number;
3989
+ cacheRead: number;
3990
+ cacheWrite: number;
3991
+ };
3992
+ contextWindow: number;
3993
+ maxTokens: number;
3994
+ };
3995
+ readonly "anthropic/claude-opus-4.6": {
3996
+ id: string;
3997
+ name: string;
3998
+ api: "openai-completions";
3999
+ provider: string;
4000
+ baseUrl: string;
4001
+ reasoning: true;
4002
+ input: ("image" | "text")[];
4003
+ cost: {
4004
+ input: number;
4005
+ output: number;
4006
+ cacheRead: number;
4007
+ cacheWrite: number;
4008
+ };
4009
+ contextWindow: number;
4010
+ maxTokens: number;
4011
+ };
4012
+ readonly "qwen/qwen3-coder-next": {
4013
+ id: string;
4014
+ name: string;
4015
+ api: "openai-completions";
4016
+ provider: string;
4017
+ baseUrl: string;
4018
+ reasoning: false;
4019
+ input: "text"[];
4020
+ cost: {
4021
+ input: number;
4022
+ output: number;
4023
+ cacheRead: number;
4024
+ cacheWrite: number;
4025
+ };
4026
+ contextWindow: number;
4027
+ maxTokens: number;
4028
+ };
4029
+ readonly "openrouter/free": {
4030
+ id: string;
4031
+ name: string;
4032
+ api: "openai-completions";
4033
+ provider: string;
4034
+ baseUrl: string;
4035
+ reasoning: true;
4036
+ input: ("image" | "text")[];
4037
+ cost: {
4038
+ input: number;
4039
+ output: number;
4040
+ cacheRead: number;
4041
+ cacheWrite: number;
4042
+ };
4043
+ contextWindow: number;
4044
+ maxTokens: number;
4045
+ };
4046
+ readonly "stepfun/step-3.5-flash:free": {
4047
+ id: string;
4048
+ name: string;
4049
+ api: "openai-completions";
4050
+ provider: string;
4051
+ baseUrl: string;
4052
+ reasoning: true;
4053
+ input: "text"[];
4054
+ cost: {
4055
+ input: number;
4056
+ output: number;
4057
+ cacheRead: number;
4058
+ cacheWrite: number;
4059
+ };
4060
+ contextWindow: number;
4061
+ maxTokens: number;
4062
+ };
4063
+ readonly "stepfun/step-3.5-flash": {
3685
4064
  id: string;
3686
4065
  name: string;
3687
4066
  api: "openai-completions";
@@ -4242,40 +4621,6 @@ export declare const MODELS: {
4242
4621
  contextWindow: number;
4243
4622
  maxTokens: number;
4244
4623
  };
4245
- readonly "tngtech/tng-r1t-chimera:free": {
4246
- id: string;
4247
- name: string;
4248
- api: "openai-completions";
4249
- provider: string;
4250
- baseUrl: string;
4251
- reasoning: true;
4252
- input: "text"[];
4253
- cost: {
4254
- input: number;
4255
- output: number;
4256
- cacheRead: number;
4257
- cacheWrite: number;
4258
- };
4259
- contextWindow: number;
4260
- maxTokens: number;
4261
- };
4262
- readonly "tngtech/tng-r1t-chimera": {
4263
- id: string;
4264
- name: string;
4265
- api: "openai-completions";
4266
- provider: string;
4267
- baseUrl: string;
4268
- reasoning: true;
4269
- input: "text"[];
4270
- cost: {
4271
- input: number;
4272
- output: number;
4273
- cacheRead: number;
4274
- cacheWrite: number;
4275
- };
4276
- contextWindow: number;
4277
- maxTokens: number;
4278
- };
4279
4624
  readonly "anthropic/claude-opus-4.5": {
4280
4625
  id: string;
4281
4626
  name: string;
@@ -4514,6 +4859,23 @@ export declare const MODELS: {
4514
4859
  contextWindow: number;
4515
4860
  maxTokens: number;
4516
4861
  };
4862
+ readonly "qwen/qwen3-vl-32b-instruct": {
4863
+ id: string;
4864
+ name: string;
4865
+ api: "openai-completions";
4866
+ provider: string;
4867
+ baseUrl: string;
4868
+ reasoning: false;
4869
+ input: ("image" | "text")[];
4870
+ cost: {
4871
+ input: number;
4872
+ output: number;
4873
+ cacheRead: number;
4874
+ cacheWrite: number;
4875
+ };
4876
+ contextWindow: number;
4877
+ maxTokens: number;
4878
+ };
4517
4879
  readonly "openai/gpt-5-image-mini": {
4518
4880
  id: string;
4519
4881
  name: string;
@@ -4769,23 +5131,6 @@ export declare const MODELS: {
4769
5131
  contextWindow: number;
4770
5132
  maxTokens: number;
4771
5133
  };
4772
- readonly "google/gemini-2.5-flash-preview-09-2025": {
4773
- id: string;
4774
- name: string;
4775
- api: "openai-completions";
4776
- provider: string;
4777
- baseUrl: string;
4778
- reasoning: true;
4779
- input: ("image" | "text")[];
4780
- cost: {
4781
- input: number;
4782
- output: number;
4783
- cacheRead: number;
4784
- cacheWrite: number;
4785
- };
4786
- contextWindow: number;
4787
- maxTokens: number;
4788
- };
4789
5134
  readonly "google/gemini-2.5-flash-lite-preview-09-2025": {
4790
5135
  id: string;
4791
5136
  name: string;
@@ -5126,23 +5471,6 @@ export declare const MODELS: {
5126
5471
  contextWindow: number;
5127
5472
  maxTokens: number;
5128
5473
  };
5129
- readonly "stepfun-ai/step3": {
5130
- id: string;
5131
- name: string;
5132
- api: "openai-completions";
5133
- provider: string;
5134
- baseUrl: string;
5135
- reasoning: true;
5136
- input: ("image" | "text")[];
5137
- cost: {
5138
- input: number;
5139
- output: number;
5140
- cacheRead: number;
5141
- cacheWrite: number;
5142
- };
5143
- contextWindow: number;
5144
- maxTokens: number;
5145
- };
5146
5474
  readonly "qwen/qwen3-30b-a3b-thinking-2507": {
5147
5475
  id: string;
5148
5476
  name: string;
@@ -5177,23 +5505,6 @@ export declare const MODELS: {
5177
5505
  contextWindow: number;
5178
5506
  maxTokens: number;
5179
5507
  };
5180
- readonly "nousresearch/hermes-4-70b": {
5181
- id: string;
5182
- name: string;
5183
- api: "openai-completions";
5184
- provider: string;
5185
- baseUrl: string;
5186
- reasoning: true;
5187
- input: "text"[];
5188
- cost: {
5189
- input: number;
5190
- output: number;
5191
- cacheRead: number;
5192
- cacheWrite: number;
5193
- };
5194
- contextWindow: number;
5195
- maxTokens: number;
5196
- };
5197
5508
  readonly "deepseek/deepseek-chat-v3.1": {
5198
5509
  id: string;
5199
5510
  name: string;
@@ -5296,23 +5607,6 @@ export declare const MODELS: {
5296
5607
  contextWindow: number;
5297
5608
  maxTokens: number;
5298
5609
  };
5299
- readonly "ai21/jamba-mini-1.7": {
5300
- id: string;
5301
- name: string;
5302
- api: "openai-completions";
5303
- provider: string;
5304
- baseUrl: string;
5305
- reasoning: false;
5306
- input: "text"[];
5307
- cost: {
5308
- input: number;
5309
- output: number;
5310
- cacheRead: number;
5311
- cacheWrite: number;
5312
- };
5313
- contextWindow: number;
5314
- maxTokens: number;
5315
- };
5316
5610
  readonly "ai21/jamba-large-1.7": {
5317
5611
  id: string;
5318
5612
  name: string;
@@ -5642,7 +5936,7 @@ export declare const MODELS: {
5642
5936
  api: "openai-completions";
5643
5937
  provider: string;
5644
5938
  baseUrl: string;
5645
- reasoning: true;
5939
+ reasoning: false;
5646
5940
  input: "text"[];
5647
5941
  cost: {
5648
5942
  input: number;
@@ -5993,23 +6287,6 @@ export declare const MODELS: {
5993
6287
  contextWindow: number;
5994
6288
  maxTokens: number;
5995
6289
  };
5996
- readonly "nousresearch/deephermes-3-mistral-24b-preview": {
5997
- id: string;
5998
- name: string;
5999
- api: "openai-completions";
6000
- provider: string;
6001
- baseUrl: string;
6002
- reasoning: true;
6003
- input: "text"[];
6004
- cost: {
6005
- input: number;
6006
- output: number;
6007
- cacheRead: number;
6008
- cacheWrite: number;
6009
- };
6010
- contextWindow: number;
6011
- maxTokens: number;
6012
- };
6013
6290
  readonly "mistralai/mistral-medium-3": {
6014
6291
  id: string;
6015
6292
  name: string;
@@ -6384,23 +6661,6 @@ export declare const MODELS: {
6384
6661
  contextWindow: number;
6385
6662
  maxTokens: number;
6386
6663
  };
6387
- readonly "mistralai/mistral-small-3.1-24b-instruct": {
6388
- id: string;
6389
- name: string;
6390
- api: "openai-completions";
6391
- provider: string;
6392
- baseUrl: string;
6393
- reasoning: false;
6394
- input: ("image" | "text")[];
6395
- cost: {
6396
- input: number;
6397
- output: number;
6398
- cacheRead: number;
6399
- cacheWrite: number;
6400
- };
6401
- contextWindow: number;
6402
- maxTokens: number;
6403
- };
6404
6664
  readonly "google/gemma-3-27b-it:free": {
6405
6665
  id: string;
6406
6666
  name: string;
@@ -6469,7 +6729,7 @@ export declare const MODELS: {
6469
6729
  contextWindow: number;
6470
6730
  maxTokens: number;
6471
6731
  };
6472
- readonly "anthropic/claude-3.7-sonnet:thinking": {
6732
+ readonly "anthropic/claude-3.7-sonnet": {
6473
6733
  id: string;
6474
6734
  name: string;
6475
6735
  api: "openai-completions";
@@ -6486,7 +6746,7 @@ export declare const MODELS: {
6486
6746
  contextWindow: number;
6487
6747
  maxTokens: number;
6488
6748
  };
6489
- readonly "anthropic/claude-3.7-sonnet": {
6749
+ readonly "anthropic/claude-3.7-sonnet:thinking": {
6490
6750
  id: string;
6491
6751
  name: string;
6492
6752
  api: "openai-completions";
@@ -6911,40 +7171,6 @@ export declare const MODELS: {
6911
7171
  contextWindow: number;
6912
7172
  maxTokens: number;
6913
7173
  };
6914
- readonly "mistralai/ministral-8b": {
6915
- id: string;
6916
- name: string;
6917
- api: "openai-completions";
6918
- provider: string;
6919
- baseUrl: string;
6920
- reasoning: false;
6921
- input: "text"[];
6922
- cost: {
6923
- input: number;
6924
- output: number;
6925
- cacheRead: number;
6926
- cacheWrite: number;
6927
- };
6928
- contextWindow: number;
6929
- maxTokens: number;
6930
- };
6931
- readonly "mistralai/ministral-3b": {
6932
- id: string;
6933
- name: string;
6934
- api: "openai-completions";
6935
- provider: string;
6936
- baseUrl: string;
6937
- reasoning: false;
6938
- input: "text"[];
6939
- cost: {
6940
- input: number;
6941
- output: number;
6942
- cacheRead: number;
6943
- cacheWrite: number;
6944
- };
6945
- contextWindow: number;
6946
- maxTokens: number;
6947
- };
6948
7174
  readonly "qwen/qwen-2.5-7b-instruct": {
6949
7175
  id: string;
6950
7176
  name: string;
@@ -7013,23 +7239,6 @@ export declare const MODELS: {
7013
7239
  contextWindow: number;
7014
7240
  maxTokens: number;
7015
7241
  };
7016
- readonly "mistralai/pixtral-12b": {
7017
- id: string;
7018
- name: string;
7019
- api: "openai-completions";
7020
- provider: string;
7021
- baseUrl: string;
7022
- reasoning: false;
7023
- input: ("image" | "text")[];
7024
- cost: {
7025
- input: number;
7026
- output: number;
7027
- cacheRead: number;
7028
- cacheWrite: number;
7029
- };
7030
- contextWindow: number;
7031
- maxTokens: number;
7032
- };
7033
7242
  readonly "cohere/command-r-08-2024": {
7034
7243
  id: string;
7035
7244
  name: string;
@@ -7387,23 +7596,6 @@ export declare const MODELS: {
7387
7596
  contextWindow: number;
7388
7597
  maxTokens: number;
7389
7598
  };
7390
- readonly "mistralai/mistral-tiny": {
7391
- id: string;
7392
- name: string;
7393
- api: "openai-completions";
7394
- provider: string;
7395
- baseUrl: string;
7396
- reasoning: false;
7397
- input: "text"[];
7398
- cost: {
7399
- input: number;
7400
- output: number;
7401
- cacheRead: number;
7402
- cacheWrite: number;
7403
- };
7404
- contextWindow: number;
7405
- maxTokens: number;
7406
- };
7407
7599
  readonly "mistralai/mixtral-8x7b-instruct": {
7408
7600
  id: string;
7409
7601
  name: string;