@yeshwanthyk/ai 0.2.5 → 0.2.6
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/models.generated.d.ts +407 -215
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +528 -354
- package/dist/models.generated.js.map +1 -1
- package/package.json +1 -1
|
@@ -68,6 +68,23 @@ export declare const MODELS: {
|
|
|
68
68
|
contextWindow: number;
|
|
69
69
|
maxTokens: number;
|
|
70
70
|
};
|
|
71
|
+
readonly "claude-sonnet-4-6": {
|
|
72
|
+
id: string;
|
|
73
|
+
name: string;
|
|
74
|
+
api: "anthropic-messages";
|
|
75
|
+
provider: string;
|
|
76
|
+
baseUrl: string;
|
|
77
|
+
reasoning: true;
|
|
78
|
+
input: ("image" | "text")[];
|
|
79
|
+
cost: {
|
|
80
|
+
input: number;
|
|
81
|
+
output: number;
|
|
82
|
+
cacheRead: number;
|
|
83
|
+
cacheWrite: number;
|
|
84
|
+
};
|
|
85
|
+
contextWindow: number;
|
|
86
|
+
maxTokens: number;
|
|
87
|
+
};
|
|
71
88
|
readonly "claude-3-5-sonnet-20240620": {
|
|
72
89
|
id: string;
|
|
73
90
|
name: string;
|
|
@@ -529,6 +546,23 @@ export declare const MODELS: {
|
|
|
529
546
|
contextWindow: number;
|
|
530
547
|
maxTokens: number;
|
|
531
548
|
};
|
|
549
|
+
readonly "gemini-3.1-pro-preview-customtools": {
|
|
550
|
+
id: string;
|
|
551
|
+
name: string;
|
|
552
|
+
api: "google-generative-ai";
|
|
553
|
+
provider: string;
|
|
554
|
+
baseUrl: string;
|
|
555
|
+
reasoning: true;
|
|
556
|
+
input: ("image" | "text")[];
|
|
557
|
+
cost: {
|
|
558
|
+
input: number;
|
|
559
|
+
output: number;
|
|
560
|
+
cacheRead: number;
|
|
561
|
+
cacheWrite: number;
|
|
562
|
+
};
|
|
563
|
+
contextWindow: number;
|
|
564
|
+
maxTokens: number;
|
|
565
|
+
};
|
|
532
566
|
readonly "gemini-2.0-flash": {
|
|
533
567
|
id: string;
|
|
534
568
|
name: string;
|
|
@@ -665,6 +699,23 @@ export declare const MODELS: {
|
|
|
665
699
|
contextWindow: number;
|
|
666
700
|
maxTokens: number;
|
|
667
701
|
};
|
|
702
|
+
readonly "gemini-3.1-pro-preview": {
|
|
703
|
+
id: string;
|
|
704
|
+
name: string;
|
|
705
|
+
api: "google-generative-ai";
|
|
706
|
+
provider: string;
|
|
707
|
+
baseUrl: string;
|
|
708
|
+
reasoning: true;
|
|
709
|
+
input: ("image" | "text")[];
|
|
710
|
+
cost: {
|
|
711
|
+
input: number;
|
|
712
|
+
output: number;
|
|
713
|
+
cacheRead: number;
|
|
714
|
+
cacheWrite: number;
|
|
715
|
+
};
|
|
716
|
+
contextWindow: number;
|
|
717
|
+
maxTokens: number;
|
|
718
|
+
};
|
|
668
719
|
readonly "gemini-1.5-flash": {
|
|
669
720
|
id: string;
|
|
670
721
|
name: string;
|
|
@@ -752,6 +803,23 @@ export declare const MODELS: {
|
|
|
752
803
|
contextWindow: number;
|
|
753
804
|
maxTokens: number;
|
|
754
805
|
};
|
|
806
|
+
readonly "gpt-5.3-codex-spark": {
|
|
807
|
+
id: string;
|
|
808
|
+
name: string;
|
|
809
|
+
api: "openai-responses";
|
|
810
|
+
provider: string;
|
|
811
|
+
baseUrl: string;
|
|
812
|
+
reasoning: true;
|
|
813
|
+
input: ("image" | "text")[];
|
|
814
|
+
cost: {
|
|
815
|
+
input: number;
|
|
816
|
+
output: number;
|
|
817
|
+
cacheRead: number;
|
|
818
|
+
cacheWrite: number;
|
|
819
|
+
};
|
|
820
|
+
contextWindow: number;
|
|
821
|
+
maxTokens: number;
|
|
822
|
+
};
|
|
755
823
|
readonly "gpt-4": {
|
|
756
824
|
id: string;
|
|
757
825
|
name: string;
|
|
@@ -1623,6 +1691,23 @@ export declare const MODELS: {
|
|
|
1623
1691
|
contextWindow: number;
|
|
1624
1692
|
maxTokens: number;
|
|
1625
1693
|
};
|
|
1694
|
+
readonly "llama3.1-8b": {
|
|
1695
|
+
id: string;
|
|
1696
|
+
name: string;
|
|
1697
|
+
api: "openai-completions";
|
|
1698
|
+
provider: string;
|
|
1699
|
+
baseUrl: string;
|
|
1700
|
+
reasoning: false;
|
|
1701
|
+
input: "text"[];
|
|
1702
|
+
cost: {
|
|
1703
|
+
input: number;
|
|
1704
|
+
output: number;
|
|
1705
|
+
cacheRead: number;
|
|
1706
|
+
cacheWrite: number;
|
|
1707
|
+
};
|
|
1708
|
+
contextWindow: number;
|
|
1709
|
+
maxTokens: number;
|
|
1710
|
+
};
|
|
1626
1711
|
readonly "gpt-oss-120b": {
|
|
1627
1712
|
id: string;
|
|
1628
1713
|
name: string;
|
|
@@ -2035,6 +2120,23 @@ export declare const MODELS: {
|
|
|
2035
2120
|
contextWindow: number;
|
|
2036
2121
|
maxTokens: number;
|
|
2037
2122
|
};
|
|
2123
|
+
readonly "glm-5": {
|
|
2124
|
+
id: string;
|
|
2125
|
+
name: string;
|
|
2126
|
+
api: "anthropic-messages";
|
|
2127
|
+
provider: string;
|
|
2128
|
+
baseUrl: string;
|
|
2129
|
+
reasoning: true;
|
|
2130
|
+
input: "text"[];
|
|
2131
|
+
cost: {
|
|
2132
|
+
input: number;
|
|
2133
|
+
output: number;
|
|
2134
|
+
cacheRead: number;
|
|
2135
|
+
cacheWrite: number;
|
|
2136
|
+
};
|
|
2137
|
+
contextWindow: number;
|
|
2138
|
+
maxTokens: number;
|
|
2139
|
+
};
|
|
2038
2140
|
readonly "glm-4.5-flash": {
|
|
2039
2141
|
id: string;
|
|
2040
2142
|
name: string;
|
|
@@ -2600,6 +2702,23 @@ export declare const MODELS: {
|
|
|
2600
2702
|
contextWindow: number;
|
|
2601
2703
|
maxTokens: number;
|
|
2602
2704
|
};
|
|
2705
|
+
readonly "glm-5": {
|
|
2706
|
+
id: string;
|
|
2707
|
+
name: string;
|
|
2708
|
+
api: "openai-completions";
|
|
2709
|
+
provider: string;
|
|
2710
|
+
baseUrl: string;
|
|
2711
|
+
reasoning: true;
|
|
2712
|
+
input: "text"[];
|
|
2713
|
+
cost: {
|
|
2714
|
+
input: number;
|
|
2715
|
+
output: number;
|
|
2716
|
+
cacheRead: number;
|
|
2717
|
+
cacheWrite: number;
|
|
2718
|
+
};
|
|
2719
|
+
contextWindow: number;
|
|
2720
|
+
maxTokens: number;
|
|
2721
|
+
};
|
|
2603
2722
|
readonly "qwen3-coder": {
|
|
2604
2723
|
id: string;
|
|
2605
2724
|
name: string;
|
|
@@ -2702,6 +2821,23 @@ export declare const MODELS: {
|
|
|
2702
2821
|
contextWindow: number;
|
|
2703
2822
|
maxTokens: number;
|
|
2704
2823
|
};
|
|
2824
|
+
readonly "claude-sonnet-4-6": {
|
|
2825
|
+
id: string;
|
|
2826
|
+
name: string;
|
|
2827
|
+
api: "openai-completions";
|
|
2828
|
+
provider: string;
|
|
2829
|
+
baseUrl: string;
|
|
2830
|
+
reasoning: true;
|
|
2831
|
+
input: ("image" | "text")[];
|
|
2832
|
+
cost: {
|
|
2833
|
+
input: number;
|
|
2834
|
+
output: number;
|
|
2835
|
+
cacheRead: number;
|
|
2836
|
+
cacheWrite: number;
|
|
2837
|
+
};
|
|
2838
|
+
contextWindow: number;
|
|
2839
|
+
maxTokens: number;
|
|
2840
|
+
};
|
|
2705
2841
|
readonly "claude-opus-4-6": {
|
|
2706
2842
|
id: string;
|
|
2707
2843
|
name: string;
|
|
@@ -2838,6 +2974,40 @@ export declare const MODELS: {
|
|
|
2838
2974
|
contextWindow: number;
|
|
2839
2975
|
maxTokens: number;
|
|
2840
2976
|
};
|
|
2977
|
+
readonly "glm-5-free": {
|
|
2978
|
+
id: string;
|
|
2979
|
+
name: string;
|
|
2980
|
+
api: "openai-completions";
|
|
2981
|
+
provider: string;
|
|
2982
|
+
baseUrl: string;
|
|
2983
|
+
reasoning: true;
|
|
2984
|
+
input: "text"[];
|
|
2985
|
+
cost: {
|
|
2986
|
+
input: number;
|
|
2987
|
+
output: number;
|
|
2988
|
+
cacheRead: number;
|
|
2989
|
+
cacheWrite: number;
|
|
2990
|
+
};
|
|
2991
|
+
contextWindow: number;
|
|
2992
|
+
maxTokens: number;
|
|
2993
|
+
};
|
|
2994
|
+
readonly "gemini-3.1-pro": {
|
|
2995
|
+
id: string;
|
|
2996
|
+
name: string;
|
|
2997
|
+
api: "openai-completions";
|
|
2998
|
+
provider: string;
|
|
2999
|
+
baseUrl: string;
|
|
3000
|
+
reasoning: true;
|
|
3001
|
+
input: ("image" | "text")[];
|
|
3002
|
+
cost: {
|
|
3003
|
+
input: number;
|
|
3004
|
+
output: number;
|
|
3005
|
+
cacheRead: number;
|
|
3006
|
+
cacheWrite: number;
|
|
3007
|
+
};
|
|
3008
|
+
contextWindow: number;
|
|
3009
|
+
maxTokens: number;
|
|
3010
|
+
};
|
|
2841
3011
|
readonly "gpt-5.1": {
|
|
2842
3012
|
id: string;
|
|
2843
3013
|
name: string;
|
|
@@ -3100,6 +3270,40 @@ export declare const MODELS: {
|
|
|
3100
3270
|
contextWindow: number;
|
|
3101
3271
|
maxTokens: number;
|
|
3102
3272
|
};
|
|
3273
|
+
readonly "minimax-m2.5-free": {
|
|
3274
|
+
id: string;
|
|
3275
|
+
name: string;
|
|
3276
|
+
api: "openai-completions";
|
|
3277
|
+
provider: string;
|
|
3278
|
+
baseUrl: string;
|
|
3279
|
+
reasoning: true;
|
|
3280
|
+
input: "text"[];
|
|
3281
|
+
cost: {
|
|
3282
|
+
input: number;
|
|
3283
|
+
output: number;
|
|
3284
|
+
cacheRead: number;
|
|
3285
|
+
cacheWrite: number;
|
|
3286
|
+
};
|
|
3287
|
+
contextWindow: number;
|
|
3288
|
+
maxTokens: number;
|
|
3289
|
+
};
|
|
3290
|
+
readonly "minimax-m2.5": {
|
|
3291
|
+
id: string;
|
|
3292
|
+
name: string;
|
|
3293
|
+
api: "openai-completions";
|
|
3294
|
+
provider: string;
|
|
3295
|
+
baseUrl: string;
|
|
3296
|
+
reasoning: true;
|
|
3297
|
+
input: "text"[];
|
|
3298
|
+
cost: {
|
|
3299
|
+
input: number;
|
|
3300
|
+
output: number;
|
|
3301
|
+
cacheRead: number;
|
|
3302
|
+
cacheWrite: number;
|
|
3303
|
+
};
|
|
3304
|
+
contextWindow: number;
|
|
3305
|
+
maxTokens: number;
|
|
3306
|
+
};
|
|
3103
3307
|
readonly "gpt-5.2": {
|
|
3104
3308
|
id: string;
|
|
3105
3309
|
name: string;
|
|
@@ -3402,6 +3606,34 @@ export declare const MODELS: {
|
|
|
3402
3606
|
contextWindow: number;
|
|
3403
3607
|
maxTokens: number;
|
|
3404
3608
|
};
|
|
3609
|
+
readonly "claude-sonnet-4.6": {
|
|
3610
|
+
id: string;
|
|
3611
|
+
name: string;
|
|
3612
|
+
api: "openai-completions";
|
|
3613
|
+
provider: string;
|
|
3614
|
+
baseUrl: string;
|
|
3615
|
+
headers: {
|
|
3616
|
+
"User-Agent": string;
|
|
3617
|
+
"Editor-Version": string;
|
|
3618
|
+
"Editor-Plugin-Version": string;
|
|
3619
|
+
"Copilot-Integration-Id": string;
|
|
3620
|
+
};
|
|
3621
|
+
compat: {
|
|
3622
|
+
supportsStore: false;
|
|
3623
|
+
supportsDeveloperRole: false;
|
|
3624
|
+
supportsReasoningEffort: false;
|
|
3625
|
+
};
|
|
3626
|
+
reasoning: true;
|
|
3627
|
+
input: ("image" | "text")[];
|
|
3628
|
+
cost: {
|
|
3629
|
+
input: number;
|
|
3630
|
+
output: number;
|
|
3631
|
+
cacheRead: number;
|
|
3632
|
+
cacheWrite: number;
|
|
3633
|
+
};
|
|
3634
|
+
contextWindow: number;
|
|
3635
|
+
maxTokens: number;
|
|
3636
|
+
};
|
|
3405
3637
|
readonly "gemini-2.5-pro": {
|
|
3406
3638
|
id: string;
|
|
3407
3639
|
name: string;
|
|
@@ -3453,6 +3685,34 @@ export declare const MODELS: {
|
|
|
3453
3685
|
contextWindow: number;
|
|
3454
3686
|
maxTokens: number;
|
|
3455
3687
|
};
|
|
3688
|
+
readonly "gemini-3.1-pro-preview": {
|
|
3689
|
+
id: string;
|
|
3690
|
+
name: string;
|
|
3691
|
+
api: "openai-completions";
|
|
3692
|
+
provider: string;
|
|
3693
|
+
baseUrl: string;
|
|
3694
|
+
headers: {
|
|
3695
|
+
"User-Agent": string;
|
|
3696
|
+
"Editor-Version": string;
|
|
3697
|
+
"Editor-Plugin-Version": string;
|
|
3698
|
+
"Copilot-Integration-Id": string;
|
|
3699
|
+
};
|
|
3700
|
+
compat: {
|
|
3701
|
+
supportsStore: false;
|
|
3702
|
+
supportsDeveloperRole: false;
|
|
3703
|
+
supportsReasoningEffort: false;
|
|
3704
|
+
};
|
|
3705
|
+
reasoning: true;
|
|
3706
|
+
input: ("image" | "text")[];
|
|
3707
|
+
cost: {
|
|
3708
|
+
input: number;
|
|
3709
|
+
output: number;
|
|
3710
|
+
cacheRead: number;
|
|
3711
|
+
cacheWrite: number;
|
|
3712
|
+
};
|
|
3713
|
+
contextWindow: number;
|
|
3714
|
+
maxTokens: number;
|
|
3715
|
+
};
|
|
3456
3716
|
readonly "claude-sonnet-4": {
|
|
3457
3717
|
id: string;
|
|
3458
3718
|
name: string;
|
|
@@ -3613,14 +3873,14 @@ export declare const MODELS: {
|
|
|
3613
3873
|
};
|
|
3614
3874
|
};
|
|
3615
3875
|
readonly openrouter: {
|
|
3616
|
-
readonly "
|
|
3876
|
+
readonly "google/gemini-3.1-pro-preview": {
|
|
3617
3877
|
id: string;
|
|
3618
3878
|
name: string;
|
|
3619
3879
|
api: "openai-completions";
|
|
3620
3880
|
provider: string;
|
|
3621
3881
|
baseUrl: string;
|
|
3622
3882
|
reasoning: true;
|
|
3623
|
-
input: "text"[];
|
|
3883
|
+
input: ("image" | "text")[];
|
|
3624
3884
|
cost: {
|
|
3625
3885
|
input: number;
|
|
3626
3886
|
output: number;
|
|
@@ -3630,7 +3890,7 @@ export declare const MODELS: {
|
|
|
3630
3890
|
contextWindow: number;
|
|
3631
3891
|
maxTokens: number;
|
|
3632
3892
|
};
|
|
3633
|
-
readonly "anthropic/claude-
|
|
3893
|
+
readonly "anthropic/claude-sonnet-4.6": {
|
|
3634
3894
|
id: string;
|
|
3635
3895
|
name: string;
|
|
3636
3896
|
api: "openai-completions";
|
|
@@ -3647,14 +3907,14 @@ export declare const MODELS: {
|
|
|
3647
3907
|
contextWindow: number;
|
|
3648
3908
|
maxTokens: number;
|
|
3649
3909
|
};
|
|
3650
|
-
readonly "qwen/qwen3-
|
|
3910
|
+
readonly "qwen/qwen3.5-plus-02-15": {
|
|
3651
3911
|
id: string;
|
|
3652
3912
|
name: string;
|
|
3653
3913
|
api: "openai-completions";
|
|
3654
3914
|
provider: string;
|
|
3655
3915
|
baseUrl: string;
|
|
3656
|
-
reasoning:
|
|
3657
|
-
input: "text"[];
|
|
3916
|
+
reasoning: true;
|
|
3917
|
+
input: ("image" | "text")[];
|
|
3658
3918
|
cost: {
|
|
3659
3919
|
input: number;
|
|
3660
3920
|
output: number;
|
|
@@ -3664,7 +3924,7 @@ export declare const MODELS: {
|
|
|
3664
3924
|
contextWindow: number;
|
|
3665
3925
|
maxTokens: number;
|
|
3666
3926
|
};
|
|
3667
|
-
readonly "
|
|
3927
|
+
readonly "qwen/qwen3.5-397b-a17b": {
|
|
3668
3928
|
id: string;
|
|
3669
3929
|
name: string;
|
|
3670
3930
|
api: "openai-completions";
|
|
@@ -3681,7 +3941,126 @@ export declare const MODELS: {
|
|
|
3681
3941
|
contextWindow: number;
|
|
3682
3942
|
maxTokens: number;
|
|
3683
3943
|
};
|
|
3684
|
-
readonly "
|
|
3944
|
+
readonly "minimax/minimax-m2.5": {
|
|
3945
|
+
id: string;
|
|
3946
|
+
name: string;
|
|
3947
|
+
api: "openai-completions";
|
|
3948
|
+
provider: string;
|
|
3949
|
+
baseUrl: string;
|
|
3950
|
+
reasoning: true;
|
|
3951
|
+
input: "text"[];
|
|
3952
|
+
cost: {
|
|
3953
|
+
input: number;
|
|
3954
|
+
output: number;
|
|
3955
|
+
cacheRead: number;
|
|
3956
|
+
cacheWrite: number;
|
|
3957
|
+
};
|
|
3958
|
+
contextWindow: number;
|
|
3959
|
+
maxTokens: number;
|
|
3960
|
+
};
|
|
3961
|
+
readonly "z-ai/glm-5": {
|
|
3962
|
+
id: string;
|
|
3963
|
+
name: string;
|
|
3964
|
+
api: "openai-completions";
|
|
3965
|
+
provider: string;
|
|
3966
|
+
baseUrl: string;
|
|
3967
|
+
reasoning: true;
|
|
3968
|
+
input: "text"[];
|
|
3969
|
+
cost: {
|
|
3970
|
+
input: number;
|
|
3971
|
+
output: number;
|
|
3972
|
+
cacheRead: number;
|
|
3973
|
+
cacheWrite: number;
|
|
3974
|
+
};
|
|
3975
|
+
contextWindow: number;
|
|
3976
|
+
maxTokens: number;
|
|
3977
|
+
};
|
|
3978
|
+
readonly "qwen/qwen3-max-thinking": {
|
|
3979
|
+
id: string;
|
|
3980
|
+
name: string;
|
|
3981
|
+
api: "openai-completions";
|
|
3982
|
+
provider: string;
|
|
3983
|
+
baseUrl: string;
|
|
3984
|
+
reasoning: true;
|
|
3985
|
+
input: "text"[];
|
|
3986
|
+
cost: {
|
|
3987
|
+
input: number;
|
|
3988
|
+
output: number;
|
|
3989
|
+
cacheRead: number;
|
|
3990
|
+
cacheWrite: number;
|
|
3991
|
+
};
|
|
3992
|
+
contextWindow: number;
|
|
3993
|
+
maxTokens: number;
|
|
3994
|
+
};
|
|
3995
|
+
readonly "anthropic/claude-opus-4.6": {
|
|
3996
|
+
id: string;
|
|
3997
|
+
name: string;
|
|
3998
|
+
api: "openai-completions";
|
|
3999
|
+
provider: string;
|
|
4000
|
+
baseUrl: string;
|
|
4001
|
+
reasoning: true;
|
|
4002
|
+
input: ("image" | "text")[];
|
|
4003
|
+
cost: {
|
|
4004
|
+
input: number;
|
|
4005
|
+
output: number;
|
|
4006
|
+
cacheRead: number;
|
|
4007
|
+
cacheWrite: number;
|
|
4008
|
+
};
|
|
4009
|
+
contextWindow: number;
|
|
4010
|
+
maxTokens: number;
|
|
4011
|
+
};
|
|
4012
|
+
readonly "qwen/qwen3-coder-next": {
|
|
4013
|
+
id: string;
|
|
4014
|
+
name: string;
|
|
4015
|
+
api: "openai-completions";
|
|
4016
|
+
provider: string;
|
|
4017
|
+
baseUrl: string;
|
|
4018
|
+
reasoning: false;
|
|
4019
|
+
input: "text"[];
|
|
4020
|
+
cost: {
|
|
4021
|
+
input: number;
|
|
4022
|
+
output: number;
|
|
4023
|
+
cacheRead: number;
|
|
4024
|
+
cacheWrite: number;
|
|
4025
|
+
};
|
|
4026
|
+
contextWindow: number;
|
|
4027
|
+
maxTokens: number;
|
|
4028
|
+
};
|
|
4029
|
+
readonly "openrouter/free": {
|
|
4030
|
+
id: string;
|
|
4031
|
+
name: string;
|
|
4032
|
+
api: "openai-completions";
|
|
4033
|
+
provider: string;
|
|
4034
|
+
baseUrl: string;
|
|
4035
|
+
reasoning: true;
|
|
4036
|
+
input: ("image" | "text")[];
|
|
4037
|
+
cost: {
|
|
4038
|
+
input: number;
|
|
4039
|
+
output: number;
|
|
4040
|
+
cacheRead: number;
|
|
4041
|
+
cacheWrite: number;
|
|
4042
|
+
};
|
|
4043
|
+
contextWindow: number;
|
|
4044
|
+
maxTokens: number;
|
|
4045
|
+
};
|
|
4046
|
+
readonly "stepfun/step-3.5-flash:free": {
|
|
4047
|
+
id: string;
|
|
4048
|
+
name: string;
|
|
4049
|
+
api: "openai-completions";
|
|
4050
|
+
provider: string;
|
|
4051
|
+
baseUrl: string;
|
|
4052
|
+
reasoning: true;
|
|
4053
|
+
input: "text"[];
|
|
4054
|
+
cost: {
|
|
4055
|
+
input: number;
|
|
4056
|
+
output: number;
|
|
4057
|
+
cacheRead: number;
|
|
4058
|
+
cacheWrite: number;
|
|
4059
|
+
};
|
|
4060
|
+
contextWindow: number;
|
|
4061
|
+
maxTokens: number;
|
|
4062
|
+
};
|
|
4063
|
+
readonly "stepfun/step-3.5-flash": {
|
|
3685
4064
|
id: string;
|
|
3686
4065
|
name: string;
|
|
3687
4066
|
api: "openai-completions";
|
|
@@ -4242,40 +4621,6 @@ export declare const MODELS: {
|
|
|
4242
4621
|
contextWindow: number;
|
|
4243
4622
|
maxTokens: number;
|
|
4244
4623
|
};
|
|
4245
|
-
readonly "tngtech/tng-r1t-chimera:free": {
|
|
4246
|
-
id: string;
|
|
4247
|
-
name: string;
|
|
4248
|
-
api: "openai-completions";
|
|
4249
|
-
provider: string;
|
|
4250
|
-
baseUrl: string;
|
|
4251
|
-
reasoning: true;
|
|
4252
|
-
input: "text"[];
|
|
4253
|
-
cost: {
|
|
4254
|
-
input: number;
|
|
4255
|
-
output: number;
|
|
4256
|
-
cacheRead: number;
|
|
4257
|
-
cacheWrite: number;
|
|
4258
|
-
};
|
|
4259
|
-
contextWindow: number;
|
|
4260
|
-
maxTokens: number;
|
|
4261
|
-
};
|
|
4262
|
-
readonly "tngtech/tng-r1t-chimera": {
|
|
4263
|
-
id: string;
|
|
4264
|
-
name: string;
|
|
4265
|
-
api: "openai-completions";
|
|
4266
|
-
provider: string;
|
|
4267
|
-
baseUrl: string;
|
|
4268
|
-
reasoning: true;
|
|
4269
|
-
input: "text"[];
|
|
4270
|
-
cost: {
|
|
4271
|
-
input: number;
|
|
4272
|
-
output: number;
|
|
4273
|
-
cacheRead: number;
|
|
4274
|
-
cacheWrite: number;
|
|
4275
|
-
};
|
|
4276
|
-
contextWindow: number;
|
|
4277
|
-
maxTokens: number;
|
|
4278
|
-
};
|
|
4279
4624
|
readonly "anthropic/claude-opus-4.5": {
|
|
4280
4625
|
id: string;
|
|
4281
4626
|
name: string;
|
|
@@ -4514,6 +4859,23 @@ export declare const MODELS: {
|
|
|
4514
4859
|
contextWindow: number;
|
|
4515
4860
|
maxTokens: number;
|
|
4516
4861
|
};
|
|
4862
|
+
readonly "qwen/qwen3-vl-32b-instruct": {
|
|
4863
|
+
id: string;
|
|
4864
|
+
name: string;
|
|
4865
|
+
api: "openai-completions";
|
|
4866
|
+
provider: string;
|
|
4867
|
+
baseUrl: string;
|
|
4868
|
+
reasoning: false;
|
|
4869
|
+
input: ("image" | "text")[];
|
|
4870
|
+
cost: {
|
|
4871
|
+
input: number;
|
|
4872
|
+
output: number;
|
|
4873
|
+
cacheRead: number;
|
|
4874
|
+
cacheWrite: number;
|
|
4875
|
+
};
|
|
4876
|
+
contextWindow: number;
|
|
4877
|
+
maxTokens: number;
|
|
4878
|
+
};
|
|
4517
4879
|
readonly "openai/gpt-5-image-mini": {
|
|
4518
4880
|
id: string;
|
|
4519
4881
|
name: string;
|
|
@@ -4769,23 +5131,6 @@ export declare const MODELS: {
|
|
|
4769
5131
|
contextWindow: number;
|
|
4770
5132
|
maxTokens: number;
|
|
4771
5133
|
};
|
|
4772
|
-
readonly "google/gemini-2.5-flash-preview-09-2025": {
|
|
4773
|
-
id: string;
|
|
4774
|
-
name: string;
|
|
4775
|
-
api: "openai-completions";
|
|
4776
|
-
provider: string;
|
|
4777
|
-
baseUrl: string;
|
|
4778
|
-
reasoning: true;
|
|
4779
|
-
input: ("image" | "text")[];
|
|
4780
|
-
cost: {
|
|
4781
|
-
input: number;
|
|
4782
|
-
output: number;
|
|
4783
|
-
cacheRead: number;
|
|
4784
|
-
cacheWrite: number;
|
|
4785
|
-
};
|
|
4786
|
-
contextWindow: number;
|
|
4787
|
-
maxTokens: number;
|
|
4788
|
-
};
|
|
4789
5134
|
readonly "google/gemini-2.5-flash-lite-preview-09-2025": {
|
|
4790
5135
|
id: string;
|
|
4791
5136
|
name: string;
|
|
@@ -5126,23 +5471,6 @@ export declare const MODELS: {
|
|
|
5126
5471
|
contextWindow: number;
|
|
5127
5472
|
maxTokens: number;
|
|
5128
5473
|
};
|
|
5129
|
-
readonly "stepfun-ai/step3": {
|
|
5130
|
-
id: string;
|
|
5131
|
-
name: string;
|
|
5132
|
-
api: "openai-completions";
|
|
5133
|
-
provider: string;
|
|
5134
|
-
baseUrl: string;
|
|
5135
|
-
reasoning: true;
|
|
5136
|
-
input: ("image" | "text")[];
|
|
5137
|
-
cost: {
|
|
5138
|
-
input: number;
|
|
5139
|
-
output: number;
|
|
5140
|
-
cacheRead: number;
|
|
5141
|
-
cacheWrite: number;
|
|
5142
|
-
};
|
|
5143
|
-
contextWindow: number;
|
|
5144
|
-
maxTokens: number;
|
|
5145
|
-
};
|
|
5146
5474
|
readonly "qwen/qwen3-30b-a3b-thinking-2507": {
|
|
5147
5475
|
id: string;
|
|
5148
5476
|
name: string;
|
|
@@ -5177,23 +5505,6 @@ export declare const MODELS: {
|
|
|
5177
5505
|
contextWindow: number;
|
|
5178
5506
|
maxTokens: number;
|
|
5179
5507
|
};
|
|
5180
|
-
readonly "nousresearch/hermes-4-70b": {
|
|
5181
|
-
id: string;
|
|
5182
|
-
name: string;
|
|
5183
|
-
api: "openai-completions";
|
|
5184
|
-
provider: string;
|
|
5185
|
-
baseUrl: string;
|
|
5186
|
-
reasoning: true;
|
|
5187
|
-
input: "text"[];
|
|
5188
|
-
cost: {
|
|
5189
|
-
input: number;
|
|
5190
|
-
output: number;
|
|
5191
|
-
cacheRead: number;
|
|
5192
|
-
cacheWrite: number;
|
|
5193
|
-
};
|
|
5194
|
-
contextWindow: number;
|
|
5195
|
-
maxTokens: number;
|
|
5196
|
-
};
|
|
5197
5508
|
readonly "deepseek/deepseek-chat-v3.1": {
|
|
5198
5509
|
id: string;
|
|
5199
5510
|
name: string;
|
|
@@ -5296,23 +5607,6 @@ export declare const MODELS: {
|
|
|
5296
5607
|
contextWindow: number;
|
|
5297
5608
|
maxTokens: number;
|
|
5298
5609
|
};
|
|
5299
|
-
readonly "ai21/jamba-mini-1.7": {
|
|
5300
|
-
id: string;
|
|
5301
|
-
name: string;
|
|
5302
|
-
api: "openai-completions";
|
|
5303
|
-
provider: string;
|
|
5304
|
-
baseUrl: string;
|
|
5305
|
-
reasoning: false;
|
|
5306
|
-
input: "text"[];
|
|
5307
|
-
cost: {
|
|
5308
|
-
input: number;
|
|
5309
|
-
output: number;
|
|
5310
|
-
cacheRead: number;
|
|
5311
|
-
cacheWrite: number;
|
|
5312
|
-
};
|
|
5313
|
-
contextWindow: number;
|
|
5314
|
-
maxTokens: number;
|
|
5315
|
-
};
|
|
5316
5610
|
readonly "ai21/jamba-large-1.7": {
|
|
5317
5611
|
id: string;
|
|
5318
5612
|
name: string;
|
|
@@ -5642,7 +5936,7 @@ export declare const MODELS: {
|
|
|
5642
5936
|
api: "openai-completions";
|
|
5643
5937
|
provider: string;
|
|
5644
5938
|
baseUrl: string;
|
|
5645
|
-
reasoning:
|
|
5939
|
+
reasoning: false;
|
|
5646
5940
|
input: "text"[];
|
|
5647
5941
|
cost: {
|
|
5648
5942
|
input: number;
|
|
@@ -5993,23 +6287,6 @@ export declare const MODELS: {
|
|
|
5993
6287
|
contextWindow: number;
|
|
5994
6288
|
maxTokens: number;
|
|
5995
6289
|
};
|
|
5996
|
-
readonly "nousresearch/deephermes-3-mistral-24b-preview": {
|
|
5997
|
-
id: string;
|
|
5998
|
-
name: string;
|
|
5999
|
-
api: "openai-completions";
|
|
6000
|
-
provider: string;
|
|
6001
|
-
baseUrl: string;
|
|
6002
|
-
reasoning: true;
|
|
6003
|
-
input: "text"[];
|
|
6004
|
-
cost: {
|
|
6005
|
-
input: number;
|
|
6006
|
-
output: number;
|
|
6007
|
-
cacheRead: number;
|
|
6008
|
-
cacheWrite: number;
|
|
6009
|
-
};
|
|
6010
|
-
contextWindow: number;
|
|
6011
|
-
maxTokens: number;
|
|
6012
|
-
};
|
|
6013
6290
|
readonly "mistralai/mistral-medium-3": {
|
|
6014
6291
|
id: string;
|
|
6015
6292
|
name: string;
|
|
@@ -6384,23 +6661,6 @@ export declare const MODELS: {
|
|
|
6384
6661
|
contextWindow: number;
|
|
6385
6662
|
maxTokens: number;
|
|
6386
6663
|
};
|
|
6387
|
-
readonly "mistralai/mistral-small-3.1-24b-instruct": {
|
|
6388
|
-
id: string;
|
|
6389
|
-
name: string;
|
|
6390
|
-
api: "openai-completions";
|
|
6391
|
-
provider: string;
|
|
6392
|
-
baseUrl: string;
|
|
6393
|
-
reasoning: false;
|
|
6394
|
-
input: ("image" | "text")[];
|
|
6395
|
-
cost: {
|
|
6396
|
-
input: number;
|
|
6397
|
-
output: number;
|
|
6398
|
-
cacheRead: number;
|
|
6399
|
-
cacheWrite: number;
|
|
6400
|
-
};
|
|
6401
|
-
contextWindow: number;
|
|
6402
|
-
maxTokens: number;
|
|
6403
|
-
};
|
|
6404
6664
|
readonly "google/gemma-3-27b-it:free": {
|
|
6405
6665
|
id: string;
|
|
6406
6666
|
name: string;
|
|
@@ -6469,7 +6729,7 @@ export declare const MODELS: {
|
|
|
6469
6729
|
contextWindow: number;
|
|
6470
6730
|
maxTokens: number;
|
|
6471
6731
|
};
|
|
6472
|
-
readonly "anthropic/claude-3.7-sonnet
|
|
6732
|
+
readonly "anthropic/claude-3.7-sonnet": {
|
|
6473
6733
|
id: string;
|
|
6474
6734
|
name: string;
|
|
6475
6735
|
api: "openai-completions";
|
|
@@ -6486,7 +6746,7 @@ export declare const MODELS: {
|
|
|
6486
6746
|
contextWindow: number;
|
|
6487
6747
|
maxTokens: number;
|
|
6488
6748
|
};
|
|
6489
|
-
readonly "anthropic/claude-3.7-sonnet": {
|
|
6749
|
+
readonly "anthropic/claude-3.7-sonnet:thinking": {
|
|
6490
6750
|
id: string;
|
|
6491
6751
|
name: string;
|
|
6492
6752
|
api: "openai-completions";
|
|
@@ -6911,40 +7171,6 @@ export declare const MODELS: {
|
|
|
6911
7171
|
contextWindow: number;
|
|
6912
7172
|
maxTokens: number;
|
|
6913
7173
|
};
|
|
6914
|
-
readonly "mistralai/ministral-8b": {
|
|
6915
|
-
id: string;
|
|
6916
|
-
name: string;
|
|
6917
|
-
api: "openai-completions";
|
|
6918
|
-
provider: string;
|
|
6919
|
-
baseUrl: string;
|
|
6920
|
-
reasoning: false;
|
|
6921
|
-
input: "text"[];
|
|
6922
|
-
cost: {
|
|
6923
|
-
input: number;
|
|
6924
|
-
output: number;
|
|
6925
|
-
cacheRead: number;
|
|
6926
|
-
cacheWrite: number;
|
|
6927
|
-
};
|
|
6928
|
-
contextWindow: number;
|
|
6929
|
-
maxTokens: number;
|
|
6930
|
-
};
|
|
6931
|
-
readonly "mistralai/ministral-3b": {
|
|
6932
|
-
id: string;
|
|
6933
|
-
name: string;
|
|
6934
|
-
api: "openai-completions";
|
|
6935
|
-
provider: string;
|
|
6936
|
-
baseUrl: string;
|
|
6937
|
-
reasoning: false;
|
|
6938
|
-
input: "text"[];
|
|
6939
|
-
cost: {
|
|
6940
|
-
input: number;
|
|
6941
|
-
output: number;
|
|
6942
|
-
cacheRead: number;
|
|
6943
|
-
cacheWrite: number;
|
|
6944
|
-
};
|
|
6945
|
-
contextWindow: number;
|
|
6946
|
-
maxTokens: number;
|
|
6947
|
-
};
|
|
6948
7174
|
readonly "qwen/qwen-2.5-7b-instruct": {
|
|
6949
7175
|
id: string;
|
|
6950
7176
|
name: string;
|
|
@@ -7013,23 +7239,6 @@ export declare const MODELS: {
|
|
|
7013
7239
|
contextWindow: number;
|
|
7014
7240
|
maxTokens: number;
|
|
7015
7241
|
};
|
|
7016
|
-
readonly "mistralai/pixtral-12b": {
|
|
7017
|
-
id: string;
|
|
7018
|
-
name: string;
|
|
7019
|
-
api: "openai-completions";
|
|
7020
|
-
provider: string;
|
|
7021
|
-
baseUrl: string;
|
|
7022
|
-
reasoning: false;
|
|
7023
|
-
input: ("image" | "text")[];
|
|
7024
|
-
cost: {
|
|
7025
|
-
input: number;
|
|
7026
|
-
output: number;
|
|
7027
|
-
cacheRead: number;
|
|
7028
|
-
cacheWrite: number;
|
|
7029
|
-
};
|
|
7030
|
-
contextWindow: number;
|
|
7031
|
-
maxTokens: number;
|
|
7032
|
-
};
|
|
7033
7242
|
readonly "cohere/command-r-08-2024": {
|
|
7034
7243
|
id: string;
|
|
7035
7244
|
name: string;
|
|
@@ -7387,23 +7596,6 @@ export declare const MODELS: {
|
|
|
7387
7596
|
contextWindow: number;
|
|
7388
7597
|
maxTokens: number;
|
|
7389
7598
|
};
|
|
7390
|
-
readonly "mistralai/mistral-tiny": {
|
|
7391
|
-
id: string;
|
|
7392
|
-
name: string;
|
|
7393
|
-
api: "openai-completions";
|
|
7394
|
-
provider: string;
|
|
7395
|
-
baseUrl: string;
|
|
7396
|
-
reasoning: false;
|
|
7397
|
-
input: "text"[];
|
|
7398
|
-
cost: {
|
|
7399
|
-
input: number;
|
|
7400
|
-
output: number;
|
|
7401
|
-
cacheRead: number;
|
|
7402
|
-
cacheWrite: number;
|
|
7403
|
-
};
|
|
7404
|
-
contextWindow: number;
|
|
7405
|
-
maxTokens: number;
|
|
7406
|
-
};
|
|
7407
7599
|
readonly "mistralai/mixtral-8x7b-instruct": {
|
|
7408
7600
|
id: string;
|
|
7409
7601
|
name: string;
|