@draht/ai 2026.3.25-1 → 2026.4.5
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +86 -0
- package/dist/index.d.ts +1 -0
- package/dist/index.d.ts.map +1 -1
- package/dist/index.js +1 -0
- package/dist/index.js.map +1 -1
- package/dist/models.generated.d.ts +467 -128
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +591 -266
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/faux.d.ts +56 -0
- package/dist/providers/faux.d.ts.map +1 -0
- package/dist/providers/faux.js +367 -0
- package/dist/providers/faux.js.map +1 -0
- package/dist/providers/openai-codex-responses.d.ts.map +1 -1
- package/dist/providers/openai-codex-responses.js +1 -1
- package/dist/providers/openai-codex-responses.js.map +1 -1
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +5 -0
- package/dist/providers/openai-completions.js.map +1 -1
- package/dist/providers/openai-responses-shared.d.ts.map +1 -1
- package/dist/providers/openai-responses-shared.js +12 -0
- package/dist/providers/openai-responses-shared.js.map +1 -1
- package/dist/types.d.ts +2 -0
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/dist/utils/overflow.d.ts +1 -1
- package/dist/utils/overflow.d.ts.map +1 -1
- package/dist/utils/overflow.js +4 -2
- package/dist/utils/overflow.js.map +1 -1
- package/package.json +1 -1
|
@@ -1054,6 +1054,23 @@ export declare const MODELS: {
|
|
|
1054
1054
|
contextWindow: number;
|
|
1055
1055
|
maxTokens: number;
|
|
1056
1056
|
};
|
|
1057
|
+
readonly "nvidia.nemotron-super-3-120b": {
|
|
1058
|
+
id: string;
|
|
1059
|
+
name: string;
|
|
1060
|
+
api: "bedrock-converse-stream";
|
|
1061
|
+
provider: string;
|
|
1062
|
+
baseUrl: string;
|
|
1063
|
+
reasoning: true;
|
|
1064
|
+
input: "text"[];
|
|
1065
|
+
cost: {
|
|
1066
|
+
input: number;
|
|
1067
|
+
output: number;
|
|
1068
|
+
cacheRead: number;
|
|
1069
|
+
cacheWrite: number;
|
|
1070
|
+
};
|
|
1071
|
+
contextWindow: number;
|
|
1072
|
+
maxTokens: number;
|
|
1073
|
+
};
|
|
1057
1074
|
readonly "openai.gpt-oss-120b-1:0": {
|
|
1058
1075
|
id: string;
|
|
1059
1076
|
name: string;
|
|
@@ -1532,23 +1549,6 @@ export declare const MODELS: {
|
|
|
1532
1549
|
contextWindow: number;
|
|
1533
1550
|
maxTokens: number;
|
|
1534
1551
|
};
|
|
1535
|
-
readonly "claude-3-7-sonnet-latest": {
|
|
1536
|
-
id: string;
|
|
1537
|
-
name: string;
|
|
1538
|
-
api: "anthropic-messages";
|
|
1539
|
-
provider: string;
|
|
1540
|
-
baseUrl: string;
|
|
1541
|
-
reasoning: true;
|
|
1542
|
-
input: ("image" | "text")[];
|
|
1543
|
-
cost: {
|
|
1544
|
-
input: number;
|
|
1545
|
-
output: number;
|
|
1546
|
-
cacheRead: number;
|
|
1547
|
-
cacheWrite: number;
|
|
1548
|
-
};
|
|
1549
|
-
contextWindow: number;
|
|
1550
|
-
maxTokens: number;
|
|
1551
|
-
};
|
|
1552
1552
|
readonly "claude-3-haiku-20240307": {
|
|
1553
1553
|
id: string;
|
|
1554
1554
|
name: string;
|
|
@@ -2282,6 +2282,23 @@ export declare const MODELS: {
|
|
|
2282
2282
|
contextWindow: number;
|
|
2283
2283
|
maxTokens: number;
|
|
2284
2284
|
};
|
|
2285
|
+
readonly "gpt-5.3-chat-latest": {
|
|
2286
|
+
id: string;
|
|
2287
|
+
name: string;
|
|
2288
|
+
api: "azure-openai-responses";
|
|
2289
|
+
provider: string;
|
|
2290
|
+
baseUrl: string;
|
|
2291
|
+
reasoning: false;
|
|
2292
|
+
input: ("image" | "text")[];
|
|
2293
|
+
cost: {
|
|
2294
|
+
input: number;
|
|
2295
|
+
output: number;
|
|
2296
|
+
cacheRead: number;
|
|
2297
|
+
cacheWrite: number;
|
|
2298
|
+
};
|
|
2299
|
+
contextWindow: number;
|
|
2300
|
+
maxTokens: number;
|
|
2301
|
+
};
|
|
2285
2302
|
readonly "gpt-5.3-codex": {
|
|
2286
2303
|
id: string;
|
|
2287
2304
|
name: string;
|
|
@@ -3589,6 +3606,57 @@ export declare const MODELS: {
|
|
|
3589
3606
|
contextWindow: number;
|
|
3590
3607
|
maxTokens: number;
|
|
3591
3608
|
};
|
|
3609
|
+
readonly "gemma-3-27b-it": {
|
|
3610
|
+
id: string;
|
|
3611
|
+
name: string;
|
|
3612
|
+
api: "google-generative-ai";
|
|
3613
|
+
provider: string;
|
|
3614
|
+
baseUrl: string;
|
|
3615
|
+
reasoning: false;
|
|
3616
|
+
input: ("image" | "text")[];
|
|
3617
|
+
cost: {
|
|
3618
|
+
input: number;
|
|
3619
|
+
output: number;
|
|
3620
|
+
cacheRead: number;
|
|
3621
|
+
cacheWrite: number;
|
|
3622
|
+
};
|
|
3623
|
+
contextWindow: number;
|
|
3624
|
+
maxTokens: number;
|
|
3625
|
+
};
|
|
3626
|
+
readonly "gemma-4-26b": {
|
|
3627
|
+
id: string;
|
|
3628
|
+
name: string;
|
|
3629
|
+
api: "google-generative-ai";
|
|
3630
|
+
provider: string;
|
|
3631
|
+
baseUrl: string;
|
|
3632
|
+
reasoning: true;
|
|
3633
|
+
input: ("image" | "text")[];
|
|
3634
|
+
cost: {
|
|
3635
|
+
input: number;
|
|
3636
|
+
output: number;
|
|
3637
|
+
cacheRead: number;
|
|
3638
|
+
cacheWrite: number;
|
|
3639
|
+
};
|
|
3640
|
+
contextWindow: number;
|
|
3641
|
+
maxTokens: number;
|
|
3642
|
+
};
|
|
3643
|
+
readonly "gemma-4-31b": {
|
|
3644
|
+
id: string;
|
|
3645
|
+
name: string;
|
|
3646
|
+
api: "google-generative-ai";
|
|
3647
|
+
provider: string;
|
|
3648
|
+
baseUrl: string;
|
|
3649
|
+
reasoning: true;
|
|
3650
|
+
input: ("image" | "text")[];
|
|
3651
|
+
cost: {
|
|
3652
|
+
input: number;
|
|
3653
|
+
output: number;
|
|
3654
|
+
cacheRead: number;
|
|
3655
|
+
cacheWrite: number;
|
|
3656
|
+
};
|
|
3657
|
+
contextWindow: number;
|
|
3658
|
+
maxTokens: number;
|
|
3659
|
+
};
|
|
3592
3660
|
};
|
|
3593
3661
|
readonly "google-antigravity": {
|
|
3594
3662
|
readonly "claude-opus-4-5-thinking": {
|
|
@@ -4090,6 +4158,40 @@ export declare const MODELS: {
|
|
|
4090
4158
|
contextWindow: number;
|
|
4091
4159
|
maxTokens: number;
|
|
4092
4160
|
};
|
|
4161
|
+
readonly "groq/compound": {
|
|
4162
|
+
id: string;
|
|
4163
|
+
name: string;
|
|
4164
|
+
api: "openai-completions";
|
|
4165
|
+
provider: string;
|
|
4166
|
+
baseUrl: string;
|
|
4167
|
+
reasoning: true;
|
|
4168
|
+
input: "text"[];
|
|
4169
|
+
cost: {
|
|
4170
|
+
input: number;
|
|
4171
|
+
output: number;
|
|
4172
|
+
cacheRead: number;
|
|
4173
|
+
cacheWrite: number;
|
|
4174
|
+
};
|
|
4175
|
+
contextWindow: number;
|
|
4176
|
+
maxTokens: number;
|
|
4177
|
+
};
|
|
4178
|
+
readonly "groq/compound-mini": {
|
|
4179
|
+
id: string;
|
|
4180
|
+
name: string;
|
|
4181
|
+
api: "openai-completions";
|
|
4182
|
+
provider: string;
|
|
4183
|
+
baseUrl: string;
|
|
4184
|
+
reasoning: true;
|
|
4185
|
+
input: "text"[];
|
|
4186
|
+
cost: {
|
|
4187
|
+
input: number;
|
|
4188
|
+
output: number;
|
|
4189
|
+
cacheRead: number;
|
|
4190
|
+
cacheWrite: number;
|
|
4191
|
+
};
|
|
4192
|
+
contextWindow: number;
|
|
4193
|
+
maxTokens: number;
|
|
4194
|
+
};
|
|
4093
4195
|
readonly "llama-3.1-8b-instant": {
|
|
4094
4196
|
id: string;
|
|
4095
4197
|
name: string;
|
|
@@ -4277,6 +4379,23 @@ export declare const MODELS: {
|
|
|
4277
4379
|
contextWindow: number;
|
|
4278
4380
|
maxTokens: number;
|
|
4279
4381
|
};
|
|
4382
|
+
readonly "openai/gpt-oss-safeguard-20b": {
|
|
4383
|
+
id: string;
|
|
4384
|
+
name: string;
|
|
4385
|
+
api: "openai-completions";
|
|
4386
|
+
provider: string;
|
|
4387
|
+
baseUrl: string;
|
|
4388
|
+
reasoning: true;
|
|
4389
|
+
input: "text"[];
|
|
4390
|
+
cost: {
|
|
4391
|
+
input: number;
|
|
4392
|
+
output: number;
|
|
4393
|
+
cacheRead: number;
|
|
4394
|
+
cacheWrite: number;
|
|
4395
|
+
};
|
|
4396
|
+
contextWindow: number;
|
|
4397
|
+
maxTokens: number;
|
|
4398
|
+
};
|
|
4280
4399
|
readonly "qwen-qwq-32b": {
|
|
4281
4400
|
id: string;
|
|
4282
4401
|
name: string;
|
|
@@ -5276,13 +5395,30 @@ export declare const MODELS: {
|
|
|
5276
5395
|
contextWindow: number;
|
|
5277
5396
|
maxTokens: number;
|
|
5278
5397
|
};
|
|
5398
|
+
readonly "mistral-small-2603": {
|
|
5399
|
+
id: string;
|
|
5400
|
+
name: string;
|
|
5401
|
+
api: "mistral-conversations";
|
|
5402
|
+
provider: string;
|
|
5403
|
+
baseUrl: string;
|
|
5404
|
+
reasoning: true;
|
|
5405
|
+
input: ("image" | "text")[];
|
|
5406
|
+
cost: {
|
|
5407
|
+
input: number;
|
|
5408
|
+
output: number;
|
|
5409
|
+
cacheRead: number;
|
|
5410
|
+
cacheWrite: number;
|
|
5411
|
+
};
|
|
5412
|
+
contextWindow: number;
|
|
5413
|
+
maxTokens: number;
|
|
5414
|
+
};
|
|
5279
5415
|
readonly "mistral-small-latest": {
|
|
5280
5416
|
id: string;
|
|
5281
5417
|
name: string;
|
|
5282
5418
|
api: "mistral-conversations";
|
|
5283
5419
|
provider: string;
|
|
5284
5420
|
baseUrl: string;
|
|
5285
|
-
reasoning:
|
|
5421
|
+
reasoning: true;
|
|
5286
5422
|
input: ("image" | "text")[];
|
|
5287
5423
|
cost: {
|
|
5288
5424
|
input: number;
|
|
@@ -5822,6 +5958,23 @@ export declare const MODELS: {
|
|
|
5822
5958
|
contextWindow: number;
|
|
5823
5959
|
maxTokens: number;
|
|
5824
5960
|
};
|
|
5961
|
+
readonly "gpt-5.3-chat-latest": {
|
|
5962
|
+
id: string;
|
|
5963
|
+
name: string;
|
|
5964
|
+
api: "openai-responses";
|
|
5965
|
+
provider: string;
|
|
5966
|
+
baseUrl: string;
|
|
5967
|
+
reasoning: false;
|
|
5968
|
+
input: ("image" | "text")[];
|
|
5969
|
+
cost: {
|
|
5970
|
+
input: number;
|
|
5971
|
+
output: number;
|
|
5972
|
+
cacheRead: number;
|
|
5973
|
+
cacheWrite: number;
|
|
5974
|
+
};
|
|
5975
|
+
contextWindow: number;
|
|
5976
|
+
maxTokens: number;
|
|
5977
|
+
};
|
|
5825
5978
|
readonly "gpt-5.3-codex": {
|
|
5826
5979
|
id: string;
|
|
5827
5980
|
name: string;
|
|
@@ -6676,14 +6829,14 @@ export declare const MODELS: {
|
|
|
6676
6829
|
contextWindow: number;
|
|
6677
6830
|
maxTokens: number;
|
|
6678
6831
|
};
|
|
6679
|
-
readonly "
|
|
6832
|
+
readonly "minimax-m2.5": {
|
|
6680
6833
|
id: string;
|
|
6681
6834
|
name: string;
|
|
6682
6835
|
api: "openai-completions";
|
|
6683
6836
|
provider: string;
|
|
6684
6837
|
baseUrl: string;
|
|
6685
6838
|
reasoning: true;
|
|
6686
|
-
input:
|
|
6839
|
+
input: "text"[];
|
|
6687
6840
|
cost: {
|
|
6688
6841
|
input: number;
|
|
6689
6842
|
output: number;
|
|
@@ -6693,10 +6846,10 @@ export declare const MODELS: {
|
|
|
6693
6846
|
contextWindow: number;
|
|
6694
6847
|
maxTokens: number;
|
|
6695
6848
|
};
|
|
6696
|
-
readonly "
|
|
6849
|
+
readonly "minimax-m2.5-free": {
|
|
6697
6850
|
id: string;
|
|
6698
6851
|
name: string;
|
|
6699
|
-
api: "
|
|
6852
|
+
api: "anthropic-messages";
|
|
6700
6853
|
provider: string;
|
|
6701
6854
|
baseUrl: string;
|
|
6702
6855
|
reasoning: true;
|
|
@@ -6710,7 +6863,7 @@ export declare const MODELS: {
|
|
|
6710
6863
|
contextWindow: number;
|
|
6711
6864
|
maxTokens: number;
|
|
6712
6865
|
};
|
|
6713
|
-
readonly "
|
|
6866
|
+
readonly "nemotron-3-super-free": {
|
|
6714
6867
|
id: string;
|
|
6715
6868
|
name: string;
|
|
6716
6869
|
api: "openai-completions";
|
|
@@ -6727,10 +6880,10 @@ export declare const MODELS: {
|
|
|
6727
6880
|
contextWindow: number;
|
|
6728
6881
|
maxTokens: number;
|
|
6729
6882
|
};
|
|
6730
|
-
readonly "
|
|
6883
|
+
readonly "qwen3.6-plus-free": {
|
|
6731
6884
|
id: string;
|
|
6732
6885
|
name: string;
|
|
6733
|
-
api: "
|
|
6886
|
+
api: "openai-completions";
|
|
6734
6887
|
provider: string;
|
|
6735
6888
|
baseUrl: string;
|
|
6736
6889
|
reasoning: true;
|
|
@@ -6744,7 +6897,9 @@ export declare const MODELS: {
|
|
|
6744
6897
|
contextWindow: number;
|
|
6745
6898
|
maxTokens: number;
|
|
6746
6899
|
};
|
|
6747
|
-
|
|
6900
|
+
};
|
|
6901
|
+
readonly "opencode-go": {
|
|
6902
|
+
readonly "glm-5": {
|
|
6748
6903
|
id: string;
|
|
6749
6904
|
name: string;
|
|
6750
6905
|
api: "openai-completions";
|
|
@@ -6761,16 +6916,14 @@ export declare const MODELS: {
|
|
|
6761
6916
|
contextWindow: number;
|
|
6762
6917
|
maxTokens: number;
|
|
6763
6918
|
};
|
|
6764
|
-
|
|
6765
|
-
readonly "opencode-go": {
|
|
6766
|
-
readonly "glm-5": {
|
|
6919
|
+
readonly "kimi-k2.5": {
|
|
6767
6920
|
id: string;
|
|
6768
6921
|
name: string;
|
|
6769
6922
|
api: "openai-completions";
|
|
6770
6923
|
provider: string;
|
|
6771
6924
|
baseUrl: string;
|
|
6772
6925
|
reasoning: true;
|
|
6773
|
-
input: "text"[];
|
|
6926
|
+
input: ("image" | "text")[];
|
|
6774
6927
|
cost: {
|
|
6775
6928
|
input: number;
|
|
6776
6929
|
output: number;
|
|
@@ -6780,7 +6933,7 @@ export declare const MODELS: {
|
|
|
6780
6933
|
contextWindow: number;
|
|
6781
6934
|
maxTokens: number;
|
|
6782
6935
|
};
|
|
6783
|
-
readonly "
|
|
6936
|
+
readonly "mimo-v2-omni": {
|
|
6784
6937
|
id: string;
|
|
6785
6938
|
name: string;
|
|
6786
6939
|
api: "openai-completions";
|
|
@@ -6797,10 +6950,27 @@ export declare const MODELS: {
|
|
|
6797
6950
|
contextWindow: number;
|
|
6798
6951
|
maxTokens: number;
|
|
6799
6952
|
};
|
|
6953
|
+
readonly "mimo-v2-pro": {
|
|
6954
|
+
id: string;
|
|
6955
|
+
name: string;
|
|
6956
|
+
api: "openai-completions";
|
|
6957
|
+
provider: string;
|
|
6958
|
+
baseUrl: string;
|
|
6959
|
+
reasoning: true;
|
|
6960
|
+
input: "text"[];
|
|
6961
|
+
cost: {
|
|
6962
|
+
input: number;
|
|
6963
|
+
output: number;
|
|
6964
|
+
cacheRead: number;
|
|
6965
|
+
cacheWrite: number;
|
|
6966
|
+
};
|
|
6967
|
+
contextWindow: number;
|
|
6968
|
+
maxTokens: number;
|
|
6969
|
+
};
|
|
6800
6970
|
readonly "minimax-m2.5": {
|
|
6801
6971
|
id: string;
|
|
6802
6972
|
name: string;
|
|
6803
|
-
api: "
|
|
6973
|
+
api: "openai-completions";
|
|
6804
6974
|
provider: string;
|
|
6805
6975
|
baseUrl: string;
|
|
6806
6976
|
reasoning: true;
|
|
@@ -7003,13 +7173,13 @@ export declare const MODELS: {
|
|
|
7003
7173
|
contextWindow: number;
|
|
7004
7174
|
maxTokens: number;
|
|
7005
7175
|
};
|
|
7006
|
-
readonly "anthropic/claude-3.
|
|
7176
|
+
readonly "anthropic/claude-3.7-sonnet": {
|
|
7007
7177
|
id: string;
|
|
7008
7178
|
name: string;
|
|
7009
7179
|
api: "openai-completions";
|
|
7010
7180
|
provider: string;
|
|
7011
7181
|
baseUrl: string;
|
|
7012
|
-
reasoning:
|
|
7182
|
+
reasoning: true;
|
|
7013
7183
|
input: ("image" | "text")[];
|
|
7014
7184
|
cost: {
|
|
7015
7185
|
input: number;
|
|
@@ -7020,7 +7190,7 @@ export declare const MODELS: {
|
|
|
7020
7190
|
contextWindow: number;
|
|
7021
7191
|
maxTokens: number;
|
|
7022
7192
|
};
|
|
7023
|
-
readonly "anthropic/claude-3.7-sonnet": {
|
|
7193
|
+
readonly "anthropic/claude-3.7-sonnet:thinking": {
|
|
7024
7194
|
id: string;
|
|
7025
7195
|
name: string;
|
|
7026
7196
|
api: "openai-completions";
|
|
@@ -7037,7 +7207,7 @@ export declare const MODELS: {
|
|
|
7037
7207
|
contextWindow: number;
|
|
7038
7208
|
maxTokens: number;
|
|
7039
7209
|
};
|
|
7040
|
-
readonly "anthropic/claude-
|
|
7210
|
+
readonly "anthropic/claude-haiku-4.5": {
|
|
7041
7211
|
id: string;
|
|
7042
7212
|
name: string;
|
|
7043
7213
|
api: "openai-completions";
|
|
@@ -7054,7 +7224,7 @@ export declare const MODELS: {
|
|
|
7054
7224
|
contextWindow: number;
|
|
7055
7225
|
maxTokens: number;
|
|
7056
7226
|
};
|
|
7057
|
-
readonly "anthropic/claude-
|
|
7227
|
+
readonly "anthropic/claude-opus-4": {
|
|
7058
7228
|
id: string;
|
|
7059
7229
|
name: string;
|
|
7060
7230
|
api: "openai-completions";
|
|
@@ -7071,7 +7241,7 @@ export declare const MODELS: {
|
|
|
7071
7241
|
contextWindow: number;
|
|
7072
7242
|
maxTokens: number;
|
|
7073
7243
|
};
|
|
7074
|
-
readonly "anthropic/claude-opus-4": {
|
|
7244
|
+
readonly "anthropic/claude-opus-4.1": {
|
|
7075
7245
|
id: string;
|
|
7076
7246
|
name: string;
|
|
7077
7247
|
api: "openai-completions";
|
|
@@ -7088,7 +7258,7 @@ export declare const MODELS: {
|
|
|
7088
7258
|
contextWindow: number;
|
|
7089
7259
|
maxTokens: number;
|
|
7090
7260
|
};
|
|
7091
|
-
readonly "anthropic/claude-opus-4.
|
|
7261
|
+
readonly "anthropic/claude-opus-4.5": {
|
|
7092
7262
|
id: string;
|
|
7093
7263
|
name: string;
|
|
7094
7264
|
api: "openai-completions";
|
|
@@ -7105,7 +7275,7 @@ export declare const MODELS: {
|
|
|
7105
7275
|
contextWindow: number;
|
|
7106
7276
|
maxTokens: number;
|
|
7107
7277
|
};
|
|
7108
|
-
readonly "anthropic/claude-opus-4.
|
|
7278
|
+
readonly "anthropic/claude-opus-4.6": {
|
|
7109
7279
|
id: string;
|
|
7110
7280
|
name: string;
|
|
7111
7281
|
api: "openai-completions";
|
|
@@ -7122,7 +7292,7 @@ export declare const MODELS: {
|
|
|
7122
7292
|
contextWindow: number;
|
|
7123
7293
|
maxTokens: number;
|
|
7124
7294
|
};
|
|
7125
|
-
readonly "anthropic/claude-
|
|
7295
|
+
readonly "anthropic/claude-sonnet-4": {
|
|
7126
7296
|
id: string;
|
|
7127
7297
|
name: string;
|
|
7128
7298
|
api: "openai-completions";
|
|
@@ -7139,7 +7309,7 @@ export declare const MODELS: {
|
|
|
7139
7309
|
contextWindow: number;
|
|
7140
7310
|
maxTokens: number;
|
|
7141
7311
|
};
|
|
7142
|
-
readonly "anthropic/claude-sonnet-4": {
|
|
7312
|
+
readonly "anthropic/claude-sonnet-4.5": {
|
|
7143
7313
|
id: string;
|
|
7144
7314
|
name: string;
|
|
7145
7315
|
api: "openai-completions";
|
|
@@ -7156,7 +7326,7 @@ export declare const MODELS: {
|
|
|
7156
7326
|
contextWindow: number;
|
|
7157
7327
|
maxTokens: number;
|
|
7158
7328
|
};
|
|
7159
|
-
readonly "anthropic/claude-sonnet-4.
|
|
7329
|
+
readonly "anthropic/claude-sonnet-4.6": {
|
|
7160
7330
|
id: string;
|
|
7161
7331
|
name: string;
|
|
7162
7332
|
api: "openai-completions";
|
|
@@ -7173,14 +7343,14 @@ export declare const MODELS: {
|
|
|
7173
7343
|
contextWindow: number;
|
|
7174
7344
|
maxTokens: number;
|
|
7175
7345
|
};
|
|
7176
|
-
readonly "
|
|
7346
|
+
readonly "arcee-ai/trinity-large-preview:free": {
|
|
7177
7347
|
id: string;
|
|
7178
7348
|
name: string;
|
|
7179
7349
|
api: "openai-completions";
|
|
7180
7350
|
provider: string;
|
|
7181
7351
|
baseUrl: string;
|
|
7182
|
-
reasoning:
|
|
7183
|
-
input:
|
|
7352
|
+
reasoning: false;
|
|
7353
|
+
input: "text"[];
|
|
7184
7354
|
cost: {
|
|
7185
7355
|
input: number;
|
|
7186
7356
|
output: number;
|
|
@@ -7190,13 +7360,13 @@ export declare const MODELS: {
|
|
|
7190
7360
|
contextWindow: number;
|
|
7191
7361
|
maxTokens: number;
|
|
7192
7362
|
};
|
|
7193
|
-
readonly "arcee-ai/trinity-large-
|
|
7363
|
+
readonly "arcee-ai/trinity-large-thinking": {
|
|
7194
7364
|
id: string;
|
|
7195
7365
|
name: string;
|
|
7196
7366
|
api: "openai-completions";
|
|
7197
7367
|
provider: string;
|
|
7198
7368
|
baseUrl: string;
|
|
7199
|
-
reasoning:
|
|
7369
|
+
reasoning: true;
|
|
7200
7370
|
input: "text"[];
|
|
7201
7371
|
cost: {
|
|
7202
7372
|
input: number;
|
|
@@ -7717,7 +7887,7 @@ export declare const MODELS: {
|
|
|
7717
7887
|
contextWindow: number;
|
|
7718
7888
|
maxTokens: number;
|
|
7719
7889
|
};
|
|
7720
|
-
readonly "google/gemini-3-
|
|
7890
|
+
readonly "google/gemini-3.1-flash-lite-preview": {
|
|
7721
7891
|
id: string;
|
|
7722
7892
|
name: string;
|
|
7723
7893
|
api: "openai-completions";
|
|
@@ -7734,7 +7904,7 @@ export declare const MODELS: {
|
|
|
7734
7904
|
contextWindow: number;
|
|
7735
7905
|
maxTokens: number;
|
|
7736
7906
|
};
|
|
7737
|
-
readonly "google/gemini-3.1-
|
|
7907
|
+
readonly "google/gemini-3.1-pro-preview": {
|
|
7738
7908
|
id: string;
|
|
7739
7909
|
name: string;
|
|
7740
7910
|
api: "openai-completions";
|
|
@@ -7751,7 +7921,7 @@ export declare const MODELS: {
|
|
|
7751
7921
|
contextWindow: number;
|
|
7752
7922
|
maxTokens: number;
|
|
7753
7923
|
};
|
|
7754
|
-
readonly "google/gemini-3.1-pro-preview": {
|
|
7924
|
+
readonly "google/gemini-3.1-pro-preview-customtools": {
|
|
7755
7925
|
id: string;
|
|
7756
7926
|
name: string;
|
|
7757
7927
|
api: "openai-completions";
|
|
@@ -7768,7 +7938,24 @@ export declare const MODELS: {
|
|
|
7768
7938
|
contextWindow: number;
|
|
7769
7939
|
maxTokens: number;
|
|
7770
7940
|
};
|
|
7771
|
-
readonly "google/
|
|
7941
|
+
readonly "google/gemma-4-26b-a4b-it": {
|
|
7942
|
+
id: string;
|
|
7943
|
+
name: string;
|
|
7944
|
+
api: "openai-completions";
|
|
7945
|
+
provider: string;
|
|
7946
|
+
baseUrl: string;
|
|
7947
|
+
reasoning: true;
|
|
7948
|
+
input: ("image" | "text")[];
|
|
7949
|
+
cost: {
|
|
7950
|
+
input: number;
|
|
7951
|
+
output: number;
|
|
7952
|
+
cacheRead: number;
|
|
7953
|
+
cacheWrite: number;
|
|
7954
|
+
};
|
|
7955
|
+
contextWindow: number;
|
|
7956
|
+
maxTokens: number;
|
|
7957
|
+
};
|
|
7958
|
+
readonly "google/gemma-4-31b-it": {
|
|
7772
7959
|
id: string;
|
|
7773
7960
|
name: string;
|
|
7774
7961
|
api: "openai-completions";
|
|
@@ -7836,7 +8023,7 @@ export declare const MODELS: {
|
|
|
7836
8023
|
contextWindow: number;
|
|
7837
8024
|
maxTokens: number;
|
|
7838
8025
|
};
|
|
7839
|
-
readonly "kwaipilot/kat-coder-pro": {
|
|
8026
|
+
readonly "kwaipilot/kat-coder-pro-v2": {
|
|
7840
8027
|
id: string;
|
|
7841
8028
|
name: string;
|
|
7842
8029
|
api: "openai-completions";
|
|
@@ -8346,23 +8533,6 @@ export declare const MODELS: {
|
|
|
8346
8533
|
contextWindow: number;
|
|
8347
8534
|
maxTokens: number;
|
|
8348
8535
|
};
|
|
8349
|
-
readonly "mistralai/mistral-small-24b-instruct-2501": {
|
|
8350
|
-
id: string;
|
|
8351
|
-
name: string;
|
|
8352
|
-
api: "openai-completions";
|
|
8353
|
-
provider: string;
|
|
8354
|
-
baseUrl: string;
|
|
8355
|
-
reasoning: false;
|
|
8356
|
-
input: "text"[];
|
|
8357
|
-
cost: {
|
|
8358
|
-
input: number;
|
|
8359
|
-
output: number;
|
|
8360
|
-
cacheRead: number;
|
|
8361
|
-
cacheWrite: number;
|
|
8362
|
-
};
|
|
8363
|
-
contextWindow: number;
|
|
8364
|
-
maxTokens: number;
|
|
8365
|
-
};
|
|
8366
8536
|
readonly "mistralai/mistral-small-2603": {
|
|
8367
8537
|
id: string;
|
|
8368
8538
|
name: string;
|
|
@@ -8380,23 +8550,6 @@ export declare const MODELS: {
|
|
|
8380
8550
|
contextWindow: number;
|
|
8381
8551
|
maxTokens: number;
|
|
8382
8552
|
};
|
|
8383
|
-
readonly "mistralai/mistral-small-3.1-24b-instruct:free": {
|
|
8384
|
-
id: string;
|
|
8385
|
-
name: string;
|
|
8386
|
-
api: "openai-completions";
|
|
8387
|
-
provider: string;
|
|
8388
|
-
baseUrl: string;
|
|
8389
|
-
reasoning: false;
|
|
8390
|
-
input: ("image" | "text")[];
|
|
8391
|
-
cost: {
|
|
8392
|
-
input: number;
|
|
8393
|
-
output: number;
|
|
8394
|
-
cacheRead: number;
|
|
8395
|
-
cacheWrite: number;
|
|
8396
|
-
};
|
|
8397
|
-
contextWindow: number;
|
|
8398
|
-
maxTokens: number;
|
|
8399
|
-
};
|
|
8400
8553
|
readonly "mistralai/mistral-small-3.2-24b-instruct": {
|
|
8401
8554
|
id: string;
|
|
8402
8555
|
name: string;
|
|
@@ -9434,6 +9587,40 @@ export declare const MODELS: {
|
|
|
9434
9587
|
contextWindow: number;
|
|
9435
9588
|
maxTokens: number;
|
|
9436
9589
|
};
|
|
9590
|
+
readonly "openai/gpt-audio": {
|
|
9591
|
+
id: string;
|
|
9592
|
+
name: string;
|
|
9593
|
+
api: "openai-completions";
|
|
9594
|
+
provider: string;
|
|
9595
|
+
baseUrl: string;
|
|
9596
|
+
reasoning: false;
|
|
9597
|
+
input: "text"[];
|
|
9598
|
+
cost: {
|
|
9599
|
+
input: number;
|
|
9600
|
+
output: number;
|
|
9601
|
+
cacheRead: number;
|
|
9602
|
+
cacheWrite: number;
|
|
9603
|
+
};
|
|
9604
|
+
contextWindow: number;
|
|
9605
|
+
maxTokens: number;
|
|
9606
|
+
};
|
|
9607
|
+
readonly "openai/gpt-audio-mini": {
|
|
9608
|
+
id: string;
|
|
9609
|
+
name: string;
|
|
9610
|
+
api: "openai-completions";
|
|
9611
|
+
provider: string;
|
|
9612
|
+
baseUrl: string;
|
|
9613
|
+
reasoning: false;
|
|
9614
|
+
input: "text"[];
|
|
9615
|
+
cost: {
|
|
9616
|
+
input: number;
|
|
9617
|
+
output: number;
|
|
9618
|
+
cacheRead: number;
|
|
9619
|
+
cacheWrite: number;
|
|
9620
|
+
};
|
|
9621
|
+
contextWindow: number;
|
|
9622
|
+
maxTokens: number;
|
|
9623
|
+
};
|
|
9437
9624
|
readonly "openai/gpt-oss-120b": {
|
|
9438
9625
|
id: string;
|
|
9439
9626
|
name: string;
|
|
@@ -9995,23 +10182,6 @@ export declare const MODELS: {
|
|
|
9995
10182
|
contextWindow: number;
|
|
9996
10183
|
maxTokens: number;
|
|
9997
10184
|
};
|
|
9998
|
-
readonly "qwen/qwen3-4b:free": {
|
|
9999
|
-
id: string;
|
|
10000
|
-
name: string;
|
|
10001
|
-
api: "openai-completions";
|
|
10002
|
-
provider: string;
|
|
10003
|
-
baseUrl: string;
|
|
10004
|
-
reasoning: true;
|
|
10005
|
-
input: "text"[];
|
|
10006
|
-
cost: {
|
|
10007
|
-
input: number;
|
|
10008
|
-
output: number;
|
|
10009
|
-
cacheRead: number;
|
|
10010
|
-
cacheWrite: number;
|
|
10011
|
-
};
|
|
10012
|
-
contextWindow: number;
|
|
10013
|
-
maxTokens: number;
|
|
10014
|
-
};
|
|
10015
10185
|
readonly "qwen/qwen3-8b": {
|
|
10016
10186
|
id: string;
|
|
10017
10187
|
name: string;
|
|
@@ -10454,6 +10624,23 @@ export declare const MODELS: {
|
|
|
10454
10624
|
contextWindow: number;
|
|
10455
10625
|
maxTokens: number;
|
|
10456
10626
|
};
|
|
10627
|
+
readonly "qwen/qwen3.6-plus:free": {
|
|
10628
|
+
id: string;
|
|
10629
|
+
name: string;
|
|
10630
|
+
api: "openai-completions";
|
|
10631
|
+
provider: string;
|
|
10632
|
+
baseUrl: string;
|
|
10633
|
+
reasoning: true;
|
|
10634
|
+
input: ("image" | "text")[];
|
|
10635
|
+
cost: {
|
|
10636
|
+
input: number;
|
|
10637
|
+
output: number;
|
|
10638
|
+
cacheRead: number;
|
|
10639
|
+
cacheWrite: number;
|
|
10640
|
+
};
|
|
10641
|
+
contextWindow: number;
|
|
10642
|
+
maxTokens: number;
|
|
10643
|
+
};
|
|
10457
10644
|
readonly "qwen/qwq-32b": {
|
|
10458
10645
|
id: string;
|
|
10459
10646
|
name: string;
|
|
@@ -10471,6 +10658,23 @@ export declare const MODELS: {
|
|
|
10471
10658
|
contextWindow: number;
|
|
10472
10659
|
maxTokens: number;
|
|
10473
10660
|
};
|
|
10661
|
+
readonly "rekaai/reka-edge": {
|
|
10662
|
+
id: string;
|
|
10663
|
+
name: string;
|
|
10664
|
+
api: "openai-completions";
|
|
10665
|
+
provider: string;
|
|
10666
|
+
baseUrl: string;
|
|
10667
|
+
reasoning: false;
|
|
10668
|
+
input: ("image" | "text")[];
|
|
10669
|
+
cost: {
|
|
10670
|
+
input: number;
|
|
10671
|
+
output: number;
|
|
10672
|
+
cacheRead: number;
|
|
10673
|
+
cacheWrite: number;
|
|
10674
|
+
};
|
|
10675
|
+
contextWindow: number;
|
|
10676
|
+
maxTokens: number;
|
|
10677
|
+
};
|
|
10474
10678
|
readonly "relace/relace-search": {
|
|
10475
10679
|
id: string;
|
|
10476
10680
|
name: string;
|
|
@@ -10743,7 +10947,7 @@ export declare const MODELS: {
|
|
|
10743
10947
|
contextWindow: number;
|
|
10744
10948
|
maxTokens: number;
|
|
10745
10949
|
};
|
|
10746
|
-
readonly "x-ai/grok-4.20
|
|
10950
|
+
readonly "x-ai/grok-4.20": {
|
|
10747
10951
|
id: string;
|
|
10748
10952
|
name: string;
|
|
10749
10953
|
api: "openai-completions";
|
|
@@ -11015,6 +11219,23 @@ export declare const MODELS: {
|
|
|
11015
11219
|
contextWindow: number;
|
|
11016
11220
|
maxTokens: number;
|
|
11017
11221
|
};
|
|
11222
|
+
readonly "z-ai/glm-5v-turbo": {
|
|
11223
|
+
id: string;
|
|
11224
|
+
name: string;
|
|
11225
|
+
api: "openai-completions";
|
|
11226
|
+
provider: string;
|
|
11227
|
+
baseUrl: string;
|
|
11228
|
+
reasoning: true;
|
|
11229
|
+
input: ("image" | "text")[];
|
|
11230
|
+
cost: {
|
|
11231
|
+
input: number;
|
|
11232
|
+
output: number;
|
|
11233
|
+
cacheRead: number;
|
|
11234
|
+
cacheWrite: number;
|
|
11235
|
+
};
|
|
11236
|
+
contextWindow: number;
|
|
11237
|
+
maxTokens: number;
|
|
11238
|
+
};
|
|
11018
11239
|
};
|
|
11019
11240
|
readonly "vercel-ai-gateway": {
|
|
11020
11241
|
readonly "alibaba/qwen-3-14b": {
|
|
@@ -11272,13 +11493,13 @@ export declare const MODELS: {
|
|
|
11272
11493
|
contextWindow: number;
|
|
11273
11494
|
maxTokens: number;
|
|
11274
11495
|
};
|
|
11275
|
-
readonly "
|
|
11496
|
+
readonly "alibaba/qwen3.6-plus": {
|
|
11276
11497
|
id: string;
|
|
11277
11498
|
name: string;
|
|
11278
11499
|
api: "anthropic-messages";
|
|
11279
11500
|
provider: string;
|
|
11280
11501
|
baseUrl: string;
|
|
11281
|
-
reasoning:
|
|
11502
|
+
reasoning: true;
|
|
11282
11503
|
input: ("image" | "text")[];
|
|
11283
11504
|
cost: {
|
|
11284
11505
|
input: number;
|
|
@@ -11289,7 +11510,7 @@ export declare const MODELS: {
|
|
|
11289
11510
|
contextWindow: number;
|
|
11290
11511
|
maxTokens: number;
|
|
11291
11512
|
};
|
|
11292
|
-
readonly "anthropic/claude-3
|
|
11513
|
+
readonly "anthropic/claude-3-haiku": {
|
|
11293
11514
|
id: string;
|
|
11294
11515
|
name: string;
|
|
11295
11516
|
api: "anthropic-messages";
|
|
@@ -11306,7 +11527,7 @@ export declare const MODELS: {
|
|
|
11306
11527
|
contextWindow: number;
|
|
11307
11528
|
maxTokens: number;
|
|
11308
11529
|
};
|
|
11309
|
-
readonly "anthropic/claude-3.5-
|
|
11530
|
+
readonly "anthropic/claude-3.5-haiku": {
|
|
11310
11531
|
id: string;
|
|
11311
11532
|
name: string;
|
|
11312
11533
|
api: "anthropic-messages";
|
|
@@ -11323,13 +11544,13 @@ export declare const MODELS: {
|
|
|
11323
11544
|
contextWindow: number;
|
|
11324
11545
|
maxTokens: number;
|
|
11325
11546
|
};
|
|
11326
|
-
readonly "anthropic/claude-3.
|
|
11547
|
+
readonly "anthropic/claude-3.7-sonnet": {
|
|
11327
11548
|
id: string;
|
|
11328
11549
|
name: string;
|
|
11329
11550
|
api: "anthropic-messages";
|
|
11330
11551
|
provider: string;
|
|
11331
11552
|
baseUrl: string;
|
|
11332
|
-
reasoning:
|
|
11553
|
+
reasoning: true;
|
|
11333
11554
|
input: ("image" | "text")[];
|
|
11334
11555
|
cost: {
|
|
11335
11556
|
input: number;
|
|
@@ -11340,7 +11561,7 @@ export declare const MODELS: {
|
|
|
11340
11561
|
contextWindow: number;
|
|
11341
11562
|
maxTokens: number;
|
|
11342
11563
|
};
|
|
11343
|
-
readonly "anthropic/claude-
|
|
11564
|
+
readonly "anthropic/claude-haiku-4.5": {
|
|
11344
11565
|
id: string;
|
|
11345
11566
|
name: string;
|
|
11346
11567
|
api: "anthropic-messages";
|
|
@@ -11357,7 +11578,7 @@ export declare const MODELS: {
|
|
|
11357
11578
|
contextWindow: number;
|
|
11358
11579
|
maxTokens: number;
|
|
11359
11580
|
};
|
|
11360
|
-
readonly "anthropic/claude-
|
|
11581
|
+
readonly "anthropic/claude-opus-4": {
|
|
11361
11582
|
id: string;
|
|
11362
11583
|
name: string;
|
|
11363
11584
|
api: "anthropic-messages";
|
|
@@ -11374,7 +11595,7 @@ export declare const MODELS: {
|
|
|
11374
11595
|
contextWindow: number;
|
|
11375
11596
|
maxTokens: number;
|
|
11376
11597
|
};
|
|
11377
|
-
readonly "anthropic/claude-opus-4": {
|
|
11598
|
+
readonly "anthropic/claude-opus-4.1": {
|
|
11378
11599
|
id: string;
|
|
11379
11600
|
name: string;
|
|
11380
11601
|
api: "anthropic-messages";
|
|
@@ -11391,7 +11612,7 @@ export declare const MODELS: {
|
|
|
11391
11612
|
contextWindow: number;
|
|
11392
11613
|
maxTokens: number;
|
|
11393
11614
|
};
|
|
11394
|
-
readonly "anthropic/claude-opus-4.
|
|
11615
|
+
readonly "anthropic/claude-opus-4.5": {
|
|
11395
11616
|
id: string;
|
|
11396
11617
|
name: string;
|
|
11397
11618
|
api: "anthropic-messages";
|
|
@@ -11408,7 +11629,7 @@ export declare const MODELS: {
|
|
|
11408
11629
|
contextWindow: number;
|
|
11409
11630
|
maxTokens: number;
|
|
11410
11631
|
};
|
|
11411
|
-
readonly "anthropic/claude-opus-4.
|
|
11632
|
+
readonly "anthropic/claude-opus-4.6": {
|
|
11412
11633
|
id: string;
|
|
11413
11634
|
name: string;
|
|
11414
11635
|
api: "anthropic-messages";
|
|
@@ -11425,7 +11646,7 @@ export declare const MODELS: {
|
|
|
11425
11646
|
contextWindow: number;
|
|
11426
11647
|
maxTokens: number;
|
|
11427
11648
|
};
|
|
11428
|
-
readonly "anthropic/claude-
|
|
11649
|
+
readonly "anthropic/claude-sonnet-4": {
|
|
11429
11650
|
id: string;
|
|
11430
11651
|
name: string;
|
|
11431
11652
|
api: "anthropic-messages";
|
|
@@ -11442,7 +11663,7 @@ export declare const MODELS: {
|
|
|
11442
11663
|
contextWindow: number;
|
|
11443
11664
|
maxTokens: number;
|
|
11444
11665
|
};
|
|
11445
|
-
readonly "anthropic/claude-sonnet-4": {
|
|
11666
|
+
readonly "anthropic/claude-sonnet-4.5": {
|
|
11446
11667
|
id: string;
|
|
11447
11668
|
name: string;
|
|
11448
11669
|
api: "anthropic-messages";
|
|
@@ -11459,7 +11680,7 @@ export declare const MODELS: {
|
|
|
11459
11680
|
contextWindow: number;
|
|
11460
11681
|
maxTokens: number;
|
|
11461
11682
|
};
|
|
11462
|
-
readonly "anthropic/claude-sonnet-4.
|
|
11683
|
+
readonly "anthropic/claude-sonnet-4.6": {
|
|
11463
11684
|
id: string;
|
|
11464
11685
|
name: string;
|
|
11465
11686
|
api: "anthropic-messages";
|
|
@@ -11476,14 +11697,14 @@ export declare const MODELS: {
|
|
|
11476
11697
|
contextWindow: number;
|
|
11477
11698
|
maxTokens: number;
|
|
11478
11699
|
};
|
|
11479
|
-
readonly "
|
|
11700
|
+
readonly "arcee-ai/trinity-large-preview": {
|
|
11480
11701
|
id: string;
|
|
11481
11702
|
name: string;
|
|
11482
11703
|
api: "anthropic-messages";
|
|
11483
11704
|
provider: string;
|
|
11484
11705
|
baseUrl: string;
|
|
11485
|
-
reasoning:
|
|
11486
|
-
input:
|
|
11706
|
+
reasoning: false;
|
|
11707
|
+
input: "text"[];
|
|
11487
11708
|
cost: {
|
|
11488
11709
|
input: number;
|
|
11489
11710
|
output: number;
|
|
@@ -11493,13 +11714,13 @@ export declare const MODELS: {
|
|
|
11493
11714
|
contextWindow: number;
|
|
11494
11715
|
maxTokens: number;
|
|
11495
11716
|
};
|
|
11496
|
-
readonly "arcee-ai/trinity-large-
|
|
11717
|
+
readonly "arcee-ai/trinity-large-thinking": {
|
|
11497
11718
|
id: string;
|
|
11498
11719
|
name: string;
|
|
11499
11720
|
api: "anthropic-messages";
|
|
11500
11721
|
provider: string;
|
|
11501
11722
|
baseUrl: string;
|
|
11502
|
-
reasoning:
|
|
11723
|
+
reasoning: true;
|
|
11503
11724
|
input: "text"[];
|
|
11504
11725
|
cost: {
|
|
11505
11726
|
input: number;
|
|
@@ -11799,6 +12020,40 @@ export declare const MODELS: {
|
|
|
11799
12020
|
contextWindow: number;
|
|
11800
12021
|
maxTokens: number;
|
|
11801
12022
|
};
|
|
12023
|
+
readonly "google/gemma-4-26b-a4b-it": {
|
|
12024
|
+
id: string;
|
|
12025
|
+
name: string;
|
|
12026
|
+
api: "anthropic-messages";
|
|
12027
|
+
provider: string;
|
|
12028
|
+
baseUrl: string;
|
|
12029
|
+
reasoning: false;
|
|
12030
|
+
input: ("image" | "text")[];
|
|
12031
|
+
cost: {
|
|
12032
|
+
input: number;
|
|
12033
|
+
output: number;
|
|
12034
|
+
cacheRead: number;
|
|
12035
|
+
cacheWrite: number;
|
|
12036
|
+
};
|
|
12037
|
+
contextWindow: number;
|
|
12038
|
+
maxTokens: number;
|
|
12039
|
+
};
|
|
12040
|
+
readonly "google/gemma-4-31b-it": {
|
|
12041
|
+
id: string;
|
|
12042
|
+
name: string;
|
|
12043
|
+
api: "anthropic-messages";
|
|
12044
|
+
provider: string;
|
|
12045
|
+
baseUrl: string;
|
|
12046
|
+
reasoning: false;
|
|
12047
|
+
input: ("image" | "text")[];
|
|
12048
|
+
cost: {
|
|
12049
|
+
input: number;
|
|
12050
|
+
output: number;
|
|
12051
|
+
cacheRead: number;
|
|
12052
|
+
cacheWrite: number;
|
|
12053
|
+
};
|
|
12054
|
+
contextWindow: number;
|
|
12055
|
+
maxTokens: number;
|
|
12056
|
+
};
|
|
11802
12057
|
readonly "inception/mercury-2": {
|
|
11803
12058
|
id: string;
|
|
11804
12059
|
name: string;
|
|
@@ -11833,13 +12088,13 @@ export declare const MODELS: {
|
|
|
11833
12088
|
contextWindow: number;
|
|
11834
12089
|
maxTokens: number;
|
|
11835
12090
|
};
|
|
11836
|
-
readonly "
|
|
12091
|
+
readonly "kwaipilot/kat-coder-pro-v2": {
|
|
11837
12092
|
id: string;
|
|
11838
12093
|
name: string;
|
|
11839
12094
|
api: "anthropic-messages";
|
|
11840
12095
|
provider: string;
|
|
11841
12096
|
baseUrl: string;
|
|
11842
|
-
reasoning:
|
|
12097
|
+
reasoning: true;
|
|
11843
12098
|
input: "text"[];
|
|
11844
12099
|
cost: {
|
|
11845
12100
|
input: number;
|
|
@@ -11850,13 +12105,13 @@ export declare const MODELS: {
|
|
|
11850
12105
|
contextWindow: number;
|
|
11851
12106
|
maxTokens: number;
|
|
11852
12107
|
};
|
|
11853
|
-
readonly "meituan/longcat-flash-
|
|
12108
|
+
readonly "meituan/longcat-flash-chat": {
|
|
11854
12109
|
id: string;
|
|
11855
12110
|
name: string;
|
|
11856
12111
|
api: "anthropic-messages";
|
|
11857
12112
|
provider: string;
|
|
11858
12113
|
baseUrl: string;
|
|
11859
|
-
reasoning:
|
|
12114
|
+
reasoning: false;
|
|
11860
12115
|
input: "text"[];
|
|
11861
12116
|
cost: {
|
|
11862
12117
|
input: number;
|
|
@@ -12870,6 +13125,23 @@ export declare const MODELS: {
|
|
|
12870
13125
|
contextWindow: number;
|
|
12871
13126
|
maxTokens: number;
|
|
12872
13127
|
};
|
|
13128
|
+
readonly "openai/gpt-oss-120b": {
|
|
13129
|
+
id: string;
|
|
13130
|
+
name: string;
|
|
13131
|
+
api: "anthropic-messages";
|
|
13132
|
+
provider: string;
|
|
13133
|
+
baseUrl: string;
|
|
13134
|
+
reasoning: true;
|
|
13135
|
+
input: "text"[];
|
|
13136
|
+
cost: {
|
|
13137
|
+
input: number;
|
|
13138
|
+
output: number;
|
|
13139
|
+
cacheRead: number;
|
|
13140
|
+
cacheWrite: number;
|
|
13141
|
+
};
|
|
13142
|
+
contextWindow: number;
|
|
13143
|
+
maxTokens: number;
|
|
13144
|
+
};
|
|
12873
13145
|
readonly "openai/gpt-oss-20b": {
|
|
12874
13146
|
id: string;
|
|
12875
13147
|
name: string;
|
|
@@ -13567,6 +13839,23 @@ export declare const MODELS: {
|
|
|
13567
13839
|
contextWindow: number;
|
|
13568
13840
|
maxTokens: number;
|
|
13569
13841
|
};
|
|
13842
|
+
readonly "zai/glm-5v-turbo": {
|
|
13843
|
+
id: string;
|
|
13844
|
+
name: string;
|
|
13845
|
+
api: "anthropic-messages";
|
|
13846
|
+
provider: string;
|
|
13847
|
+
baseUrl: string;
|
|
13848
|
+
reasoning: true;
|
|
13849
|
+
input: ("image" | "text")[];
|
|
13850
|
+
cost: {
|
|
13851
|
+
input: number;
|
|
13852
|
+
output: number;
|
|
13853
|
+
cacheRead: number;
|
|
13854
|
+
cacheWrite: number;
|
|
13855
|
+
};
|
|
13856
|
+
contextWindow: number;
|
|
13857
|
+
maxTokens: number;
|
|
13858
|
+
};
|
|
13570
13859
|
};
|
|
13571
13860
|
readonly xai: {
|
|
13572
13861
|
readonly "grok-2": {
|
|
@@ -14072,6 +14361,7 @@ export declare const MODELS: {
|
|
|
14072
14361
|
compat: {
|
|
14073
14362
|
supportsDeveloperRole: false;
|
|
14074
14363
|
thinkingFormat: "zai";
|
|
14364
|
+
zaiToolStream: true;
|
|
14075
14365
|
};
|
|
14076
14366
|
reasoning: true;
|
|
14077
14367
|
input: "text"[];
|
|
@@ -14093,6 +14383,7 @@ export declare const MODELS: {
|
|
|
14093
14383
|
compat: {
|
|
14094
14384
|
supportsDeveloperRole: false;
|
|
14095
14385
|
thinkingFormat: "zai";
|
|
14386
|
+
zaiToolStream: true;
|
|
14096
14387
|
};
|
|
14097
14388
|
reasoning: true;
|
|
14098
14389
|
input: ("image" | "text")[];
|
|
@@ -14114,6 +14405,7 @@ export declare const MODELS: {
|
|
|
14114
14405
|
compat: {
|
|
14115
14406
|
supportsDeveloperRole: false;
|
|
14116
14407
|
thinkingFormat: "zai";
|
|
14408
|
+
zaiToolStream: true;
|
|
14117
14409
|
};
|
|
14118
14410
|
reasoning: true;
|
|
14119
14411
|
input: "text"[];
|
|
@@ -14135,6 +14427,29 @@ export declare const MODELS: {
|
|
|
14135
14427
|
compat: {
|
|
14136
14428
|
supportsDeveloperRole: false;
|
|
14137
14429
|
thinkingFormat: "zai";
|
|
14430
|
+
zaiToolStream: true;
|
|
14431
|
+
};
|
|
14432
|
+
reasoning: true;
|
|
14433
|
+
input: "text"[];
|
|
14434
|
+
cost: {
|
|
14435
|
+
input: number;
|
|
14436
|
+
output: number;
|
|
14437
|
+
cacheRead: number;
|
|
14438
|
+
cacheWrite: number;
|
|
14439
|
+
};
|
|
14440
|
+
contextWindow: number;
|
|
14441
|
+
maxTokens: number;
|
|
14442
|
+
};
|
|
14443
|
+
readonly "glm-4.7-flashx": {
|
|
14444
|
+
id: string;
|
|
14445
|
+
name: string;
|
|
14446
|
+
api: "openai-completions";
|
|
14447
|
+
provider: string;
|
|
14448
|
+
baseUrl: string;
|
|
14449
|
+
compat: {
|
|
14450
|
+
supportsDeveloperRole: false;
|
|
14451
|
+
thinkingFormat: "zai";
|
|
14452
|
+
zaiToolStream: true;
|
|
14138
14453
|
};
|
|
14139
14454
|
reasoning: true;
|
|
14140
14455
|
input: "text"[];
|
|
@@ -14156,6 +14471,7 @@ export declare const MODELS: {
|
|
|
14156
14471
|
compat: {
|
|
14157
14472
|
supportsDeveloperRole: false;
|
|
14158
14473
|
thinkingFormat: "zai";
|
|
14474
|
+
zaiToolStream: true;
|
|
14159
14475
|
};
|
|
14160
14476
|
reasoning: true;
|
|
14161
14477
|
input: "text"[];
|
|
@@ -14177,6 +14493,7 @@ export declare const MODELS: {
|
|
|
14177
14493
|
compat: {
|
|
14178
14494
|
supportsDeveloperRole: false;
|
|
14179
14495
|
thinkingFormat: "zai";
|
|
14496
|
+
zaiToolStream: true;
|
|
14180
14497
|
};
|
|
14181
14498
|
reasoning: true;
|
|
14182
14499
|
input: "text"[];
|
|
@@ -14189,6 +14506,28 @@ export declare const MODELS: {
|
|
|
14189
14506
|
contextWindow: number;
|
|
14190
14507
|
maxTokens: number;
|
|
14191
14508
|
};
|
|
14509
|
+
readonly "glm-5v-turbo": {
|
|
14510
|
+
id: string;
|
|
14511
|
+
name: string;
|
|
14512
|
+
api: "openai-completions";
|
|
14513
|
+
provider: string;
|
|
14514
|
+
baseUrl: string;
|
|
14515
|
+
compat: {
|
|
14516
|
+
supportsDeveloperRole: false;
|
|
14517
|
+
thinkingFormat: "zai";
|
|
14518
|
+
zaiToolStream: true;
|
|
14519
|
+
};
|
|
14520
|
+
reasoning: true;
|
|
14521
|
+
input: ("image" | "text")[];
|
|
14522
|
+
cost: {
|
|
14523
|
+
input: number;
|
|
14524
|
+
output: number;
|
|
14525
|
+
cacheRead: number;
|
|
14526
|
+
cacheWrite: number;
|
|
14527
|
+
};
|
|
14528
|
+
contextWindow: number;
|
|
14529
|
+
maxTokens: number;
|
|
14530
|
+
};
|
|
14192
14531
|
};
|
|
14193
14532
|
};
|
|
14194
14533
|
//# sourceMappingURL=models.generated.d.ts.map
|