@avadisabelle/ava-pi-ai 0.65.0 → 0.65.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/models.generated.d.ts +1106 -524
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +937 -331
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/amazon-bedrock.d.ts.map +1 -1
- package/dist/providers/amazon-bedrock.js +57 -26
- package/dist/providers/amazon-bedrock.js.map +1 -1
- package/dist/providers/anthropic.d.ts.map +1 -1
- package/dist/providers/anthropic.js +7 -2
- package/dist/providers/anthropic.js.map +1 -1
- package/dist/providers/azure-openai-responses.d.ts.map +1 -1
- package/dist/providers/azure-openai-responses.js +7 -4
- package/dist/providers/azure-openai-responses.js.map +1 -1
- package/dist/providers/openai-codex-responses.d.ts.map +1 -1
- package/dist/providers/openai-codex-responses.js +1 -1
- package/dist/providers/openai-codex-responses.js.map +1 -1
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +7 -2
- package/dist/providers/openai-completions.js.map +1 -1
- package/dist/providers/openai-responses.d.ts.map +1 -1
- package/dist/providers/openai-responses.js +7 -4
- package/dist/providers/openai-responses.js.map +1 -1
- package/dist/providers/simple-options.d.ts.map +1 -1
- package/dist/providers/simple-options.js +2 -0
- package/dist/providers/simple-options.js.map +1 -1
- package/dist/types.d.ts +10 -0
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/package.json +1 -1
|
@@ -323,6 +323,40 @@ export declare const MODELS: {
|
|
|
323
323
|
contextWindow: number;
|
|
324
324
|
maxTokens: number;
|
|
325
325
|
};
|
|
326
|
+
readonly "au.anthropic.claude-opus-4-6-v1": {
|
|
327
|
+
id: string;
|
|
328
|
+
name: string;
|
|
329
|
+
api: "bedrock-converse-stream";
|
|
330
|
+
provider: string;
|
|
331
|
+
baseUrl: string;
|
|
332
|
+
reasoning: true;
|
|
333
|
+
input: ("image" | "text")[];
|
|
334
|
+
cost: {
|
|
335
|
+
input: number;
|
|
336
|
+
output: number;
|
|
337
|
+
cacheRead: number;
|
|
338
|
+
cacheWrite: number;
|
|
339
|
+
};
|
|
340
|
+
contextWindow: number;
|
|
341
|
+
maxTokens: number;
|
|
342
|
+
};
|
|
343
|
+
readonly "au.anthropic.claude-sonnet-4-6": {
|
|
344
|
+
id: string;
|
|
345
|
+
name: string;
|
|
346
|
+
api: "bedrock-converse-stream";
|
|
347
|
+
provider: string;
|
|
348
|
+
baseUrl: string;
|
|
349
|
+
reasoning: true;
|
|
350
|
+
input: ("image" | "text")[];
|
|
351
|
+
cost: {
|
|
352
|
+
input: number;
|
|
353
|
+
output: number;
|
|
354
|
+
cacheRead: number;
|
|
355
|
+
cacheWrite: number;
|
|
356
|
+
};
|
|
357
|
+
contextWindow: number;
|
|
358
|
+
maxTokens: number;
|
|
359
|
+
};
|
|
326
360
|
readonly "deepseek.r1-v1:0": {
|
|
327
361
|
id: string;
|
|
328
362
|
name: string;
|
|
@@ -2486,6 +2520,23 @@ export declare const MODELS: {
|
|
|
2486
2520
|
contextWindow: number;
|
|
2487
2521
|
maxTokens: number;
|
|
2488
2522
|
};
|
|
2523
|
+
readonly "gpt-5.5": {
|
|
2524
|
+
id: string;
|
|
2525
|
+
name: string;
|
|
2526
|
+
api: "azure-openai-responses";
|
|
2527
|
+
provider: string;
|
|
2528
|
+
baseUrl: string;
|
|
2529
|
+
reasoning: true;
|
|
2530
|
+
input: ("image" | "text")[];
|
|
2531
|
+
cost: {
|
|
2532
|
+
input: number;
|
|
2533
|
+
output: number;
|
|
2534
|
+
cacheRead: number;
|
|
2535
|
+
cacheWrite: number;
|
|
2536
|
+
};
|
|
2537
|
+
contextWindow: number;
|
|
2538
|
+
maxTokens: number;
|
|
2539
|
+
};
|
|
2489
2540
|
readonly o1: {
|
|
2490
2541
|
id: string;
|
|
2491
2542
|
name: string;
|
|
@@ -3276,6 +3327,29 @@ export declare const MODELS: {
|
|
|
3276
3327
|
contextWindow: number;
|
|
3277
3328
|
maxTokens: number;
|
|
3278
3329
|
};
|
|
3330
|
+
readonly "gpt-5.5": {
|
|
3331
|
+
id: string;
|
|
3332
|
+
name: string;
|
|
3333
|
+
api: "openai-responses";
|
|
3334
|
+
provider: string;
|
|
3335
|
+
baseUrl: string;
|
|
3336
|
+
headers: {
|
|
3337
|
+
"User-Agent": string;
|
|
3338
|
+
"Editor-Version": string;
|
|
3339
|
+
"Editor-Plugin-Version": string;
|
|
3340
|
+
"Copilot-Integration-Id": string;
|
|
3341
|
+
};
|
|
3342
|
+
reasoning: true;
|
|
3343
|
+
input: ("image" | "text")[];
|
|
3344
|
+
cost: {
|
|
3345
|
+
input: number;
|
|
3346
|
+
output: number;
|
|
3347
|
+
cacheRead: number;
|
|
3348
|
+
cacheWrite: number;
|
|
3349
|
+
};
|
|
3350
|
+
contextWindow: number;
|
|
3351
|
+
maxTokens: number;
|
|
3352
|
+
};
|
|
3279
3353
|
readonly "grok-code-fast-1": {
|
|
3280
3354
|
id: string;
|
|
3281
3355
|
name: string;
|
|
@@ -3731,7 +3805,7 @@ export declare const MODELS: {
|
|
|
3731
3805
|
contextWindow: number;
|
|
3732
3806
|
maxTokens: number;
|
|
3733
3807
|
};
|
|
3734
|
-
readonly "gemma-4-26b-it": {
|
|
3808
|
+
readonly "gemma-4-26b-a4b-it": {
|
|
3735
3809
|
id: string;
|
|
3736
3810
|
name: string;
|
|
3737
3811
|
api: "google-generative-ai";
|
|
@@ -4797,6 +4871,26 @@ export declare const MODELS: {
|
|
|
4797
4871
|
contextWindow: number;
|
|
4798
4872
|
maxTokens: number;
|
|
4799
4873
|
};
|
|
4874
|
+
readonly "deepseek-ai/DeepSeek-V4-Pro": {
|
|
4875
|
+
id: string;
|
|
4876
|
+
name: string;
|
|
4877
|
+
api: "openai-completions";
|
|
4878
|
+
provider: string;
|
|
4879
|
+
baseUrl: string;
|
|
4880
|
+
compat: {
|
|
4881
|
+
supportsDeveloperRole: false;
|
|
4882
|
+
};
|
|
4883
|
+
reasoning: true;
|
|
4884
|
+
input: "text"[];
|
|
4885
|
+
cost: {
|
|
4886
|
+
input: number;
|
|
4887
|
+
output: number;
|
|
4888
|
+
cacheRead: number;
|
|
4889
|
+
cacheWrite: number;
|
|
4890
|
+
};
|
|
4891
|
+
contextWindow: number;
|
|
4892
|
+
maxTokens: number;
|
|
4893
|
+
};
|
|
4800
4894
|
readonly "moonshotai/Kimi-K2-Instruct": {
|
|
4801
4895
|
id: string;
|
|
4802
4896
|
name: string;
|
|
@@ -4877,6 +4971,26 @@ export declare const MODELS: {
|
|
|
4877
4971
|
contextWindow: number;
|
|
4878
4972
|
maxTokens: number;
|
|
4879
4973
|
};
|
|
4974
|
+
readonly "moonshotai/Kimi-K2.6": {
|
|
4975
|
+
id: string;
|
|
4976
|
+
name: string;
|
|
4977
|
+
api: "openai-completions";
|
|
4978
|
+
provider: string;
|
|
4979
|
+
baseUrl: string;
|
|
4980
|
+
compat: {
|
|
4981
|
+
supportsDeveloperRole: false;
|
|
4982
|
+
};
|
|
4983
|
+
reasoning: true;
|
|
4984
|
+
input: ("image" | "text")[];
|
|
4985
|
+
cost: {
|
|
4986
|
+
input: number;
|
|
4987
|
+
output: number;
|
|
4988
|
+
cacheRead: number;
|
|
4989
|
+
cacheWrite: number;
|
|
4990
|
+
};
|
|
4991
|
+
contextWindow: number;
|
|
4992
|
+
maxTokens: number;
|
|
4993
|
+
};
|
|
4880
4994
|
readonly "zai-org/GLM-4.7": {
|
|
4881
4995
|
id: string;
|
|
4882
4996
|
name: string;
|
|
@@ -4959,6 +5073,23 @@ export declare const MODELS: {
|
|
|
4959
5073
|
};
|
|
4960
5074
|
};
|
|
4961
5075
|
readonly "kimi-coding": {
|
|
5076
|
+
readonly k2p6: {
|
|
5077
|
+
id: string;
|
|
5078
|
+
name: string;
|
|
5079
|
+
api: "anthropic-messages";
|
|
5080
|
+
provider: string;
|
|
5081
|
+
baseUrl: string;
|
|
5082
|
+
reasoning: true;
|
|
5083
|
+
input: ("image" | "text")[];
|
|
5084
|
+
cost: {
|
|
5085
|
+
input: number;
|
|
5086
|
+
output: number;
|
|
5087
|
+
cacheRead: number;
|
|
5088
|
+
cacheWrite: number;
|
|
5089
|
+
};
|
|
5090
|
+
contextWindow: number;
|
|
5091
|
+
maxTokens: number;
|
|
5092
|
+
};
|
|
4962
5093
|
readonly "kimi-for-coding": {
|
|
4963
5094
|
id: string;
|
|
4964
5095
|
name: string;
|
|
@@ -6055,6 +6186,23 @@ export declare const MODELS: {
|
|
|
6055
6186
|
contextWindow: number;
|
|
6056
6187
|
maxTokens: number;
|
|
6057
6188
|
};
|
|
6189
|
+
readonly "gpt-5.5": {
|
|
6190
|
+
id: string;
|
|
6191
|
+
name: string;
|
|
6192
|
+
api: "openai-responses";
|
|
6193
|
+
provider: string;
|
|
6194
|
+
baseUrl: string;
|
|
6195
|
+
reasoning: true;
|
|
6196
|
+
input: ("image" | "text")[];
|
|
6197
|
+
cost: {
|
|
6198
|
+
input: number;
|
|
6199
|
+
output: number;
|
|
6200
|
+
cacheRead: number;
|
|
6201
|
+
cacheWrite: number;
|
|
6202
|
+
};
|
|
6203
|
+
contextWindow: number;
|
|
6204
|
+
maxTokens: number;
|
|
6205
|
+
};
|
|
6058
6206
|
readonly o1: {
|
|
6059
6207
|
id: string;
|
|
6060
6208
|
name: string;
|
|
@@ -6365,23 +6513,6 @@ export declare const MODELS: {
|
|
|
6365
6513
|
contextWindow: number;
|
|
6366
6514
|
maxTokens: number;
|
|
6367
6515
|
};
|
|
6368
|
-
readonly "claude-3-5-haiku": {
|
|
6369
|
-
id: string;
|
|
6370
|
-
name: string;
|
|
6371
|
-
api: "anthropic-messages";
|
|
6372
|
-
provider: string;
|
|
6373
|
-
baseUrl: string;
|
|
6374
|
-
reasoning: false;
|
|
6375
|
-
input: ("image" | "text")[];
|
|
6376
|
-
cost: {
|
|
6377
|
-
input: number;
|
|
6378
|
-
output: number;
|
|
6379
|
-
cacheRead: number;
|
|
6380
|
-
cacheWrite: number;
|
|
6381
|
-
};
|
|
6382
|
-
contextWindow: number;
|
|
6383
|
-
maxTokens: number;
|
|
6384
|
-
};
|
|
6385
6516
|
readonly "claude-haiku-4-5": {
|
|
6386
6517
|
id: string;
|
|
6387
6518
|
name: string;
|
|
@@ -6824,10 +6955,10 @@ export declare const MODELS: {
|
|
|
6824
6955
|
contextWindow: number;
|
|
6825
6956
|
maxTokens: number;
|
|
6826
6957
|
};
|
|
6827
|
-
readonly "
|
|
6958
|
+
readonly "gpt-5.5": {
|
|
6828
6959
|
id: string;
|
|
6829
6960
|
name: string;
|
|
6830
|
-
api: "openai-
|
|
6961
|
+
api: "openai-responses";
|
|
6831
6962
|
provider: string;
|
|
6832
6963
|
baseUrl: string;
|
|
6833
6964
|
reasoning: true;
|
|
@@ -6841,14 +6972,14 @@ export declare const MODELS: {
|
|
|
6841
6972
|
contextWindow: number;
|
|
6842
6973
|
maxTokens: number;
|
|
6843
6974
|
};
|
|
6844
|
-
readonly "
|
|
6975
|
+
readonly "gpt-5.5-pro": {
|
|
6845
6976
|
id: string;
|
|
6846
6977
|
name: string;
|
|
6847
|
-
api: "openai-
|
|
6978
|
+
api: "openai-responses";
|
|
6848
6979
|
provider: string;
|
|
6849
6980
|
baseUrl: string;
|
|
6850
6981
|
reasoning: true;
|
|
6851
|
-
input: "text"[];
|
|
6982
|
+
input: ("image" | "text")[];
|
|
6852
6983
|
cost: {
|
|
6853
6984
|
input: number;
|
|
6854
6985
|
output: number;
|
|
@@ -6858,10 +6989,10 @@ export declare const MODELS: {
|
|
|
6858
6989
|
contextWindow: number;
|
|
6859
6990
|
maxTokens: number;
|
|
6860
6991
|
};
|
|
6861
|
-
readonly "
|
|
6992
|
+
readonly "hy3-preview-free": {
|
|
6862
6993
|
id: string;
|
|
6863
6994
|
name: string;
|
|
6864
|
-
api: "
|
|
6995
|
+
api: "openai-completions";
|
|
6865
6996
|
provider: string;
|
|
6866
6997
|
baseUrl: string;
|
|
6867
6998
|
reasoning: true;
|
|
@@ -6875,14 +7006,14 @@ export declare const MODELS: {
|
|
|
6875
7006
|
contextWindow: number;
|
|
6876
7007
|
maxTokens: number;
|
|
6877
7008
|
};
|
|
6878
|
-
readonly "
|
|
7009
|
+
readonly "kimi-k2.5": {
|
|
6879
7010
|
id: string;
|
|
6880
7011
|
name: string;
|
|
6881
7012
|
api: "openai-completions";
|
|
6882
7013
|
provider: string;
|
|
6883
7014
|
baseUrl: string;
|
|
6884
7015
|
reasoning: true;
|
|
6885
|
-
input: "text"[];
|
|
7016
|
+
input: ("image" | "text")[];
|
|
6886
7017
|
cost: {
|
|
6887
7018
|
input: number;
|
|
6888
7019
|
output: number;
|
|
@@ -6892,7 +7023,7 @@ export declare const MODELS: {
|
|
|
6892
7023
|
contextWindow: number;
|
|
6893
7024
|
maxTokens: number;
|
|
6894
7025
|
};
|
|
6895
|
-
readonly "
|
|
7026
|
+
readonly "kimi-k2.6": {
|
|
6896
7027
|
id: string;
|
|
6897
7028
|
name: string;
|
|
6898
7029
|
api: "openai-completions";
|
|
@@ -6909,14 +7040,14 @@ export declare const MODELS: {
|
|
|
6909
7040
|
contextWindow: number;
|
|
6910
7041
|
maxTokens: number;
|
|
6911
7042
|
};
|
|
6912
|
-
readonly "
|
|
7043
|
+
readonly "ling-2.6-flash-free": {
|
|
6913
7044
|
id: string;
|
|
6914
7045
|
name: string;
|
|
6915
7046
|
api: "openai-completions";
|
|
6916
7047
|
provider: string;
|
|
6917
7048
|
baseUrl: string;
|
|
6918
|
-
reasoning:
|
|
6919
|
-
input:
|
|
7049
|
+
reasoning: false;
|
|
7050
|
+
input: "text"[];
|
|
6920
7051
|
cost: {
|
|
6921
7052
|
input: number;
|
|
6922
7053
|
output: number;
|
|
@@ -6926,9 +7057,7 @@ export declare const MODELS: {
|
|
|
6926
7057
|
contextWindow: number;
|
|
6927
7058
|
maxTokens: number;
|
|
6928
7059
|
};
|
|
6929
|
-
|
|
6930
|
-
readonly "opencode-go": {
|
|
6931
|
-
readonly "glm-5": {
|
|
7060
|
+
readonly "minimax-m2.5": {
|
|
6932
7061
|
id: string;
|
|
6933
7062
|
name: string;
|
|
6934
7063
|
api: "openai-completions";
|
|
@@ -6945,10 +7074,10 @@ export declare const MODELS: {
|
|
|
6945
7074
|
contextWindow: number;
|
|
6946
7075
|
maxTokens: number;
|
|
6947
7076
|
};
|
|
6948
|
-
readonly "
|
|
7077
|
+
readonly "minimax-m2.5-free": {
|
|
6949
7078
|
id: string;
|
|
6950
7079
|
name: string;
|
|
6951
|
-
api: "
|
|
7080
|
+
api: "anthropic-messages";
|
|
6952
7081
|
provider: string;
|
|
6953
7082
|
baseUrl: string;
|
|
6954
7083
|
reasoning: true;
|
|
@@ -6962,14 +7091,14 @@ export declare const MODELS: {
|
|
|
6962
7091
|
contextWindow: number;
|
|
6963
7092
|
maxTokens: number;
|
|
6964
7093
|
};
|
|
6965
|
-
readonly "
|
|
7094
|
+
readonly "minimax-m2.7": {
|
|
6966
7095
|
id: string;
|
|
6967
7096
|
name: string;
|
|
6968
7097
|
api: "openai-completions";
|
|
6969
7098
|
provider: string;
|
|
6970
7099
|
baseUrl: string;
|
|
6971
7100
|
reasoning: true;
|
|
6972
|
-
input:
|
|
7101
|
+
input: "text"[];
|
|
6973
7102
|
cost: {
|
|
6974
7103
|
input: number;
|
|
6975
7104
|
output: number;
|
|
@@ -6979,14 +7108,14 @@ export declare const MODELS: {
|
|
|
6979
7108
|
contextWindow: number;
|
|
6980
7109
|
maxTokens: number;
|
|
6981
7110
|
};
|
|
6982
|
-
readonly "
|
|
7111
|
+
readonly "nemotron-3-super-free": {
|
|
6983
7112
|
id: string;
|
|
6984
7113
|
name: string;
|
|
6985
7114
|
api: "openai-completions";
|
|
6986
7115
|
provider: string;
|
|
6987
7116
|
baseUrl: string;
|
|
6988
7117
|
reasoning: true;
|
|
6989
|
-
input:
|
|
7118
|
+
input: "text"[];
|
|
6990
7119
|
cost: {
|
|
6991
7120
|
input: number;
|
|
6992
7121
|
output: number;
|
|
@@ -6996,14 +7125,14 @@ export declare const MODELS: {
|
|
|
6996
7125
|
contextWindow: number;
|
|
6997
7126
|
maxTokens: number;
|
|
6998
7127
|
};
|
|
6999
|
-
readonly "
|
|
7128
|
+
readonly "qwen3.5-plus": {
|
|
7000
7129
|
id: string;
|
|
7001
7130
|
name: string;
|
|
7002
|
-
api: "
|
|
7131
|
+
api: "anthropic-messages";
|
|
7003
7132
|
provider: string;
|
|
7004
7133
|
baseUrl: string;
|
|
7005
7134
|
reasoning: true;
|
|
7006
|
-
input: "text"[];
|
|
7135
|
+
input: ("image" | "text")[];
|
|
7007
7136
|
cost: {
|
|
7008
7137
|
input: number;
|
|
7009
7138
|
output: number;
|
|
@@ -7013,14 +7142,14 @@ export declare const MODELS: {
|
|
|
7013
7142
|
contextWindow: number;
|
|
7014
7143
|
maxTokens: number;
|
|
7015
7144
|
};
|
|
7016
|
-
readonly "
|
|
7145
|
+
readonly "qwen3.6-plus": {
|
|
7017
7146
|
id: string;
|
|
7018
7147
|
name: string;
|
|
7019
|
-
api: "
|
|
7148
|
+
api: "anthropic-messages";
|
|
7020
7149
|
provider: string;
|
|
7021
7150
|
baseUrl: string;
|
|
7022
7151
|
reasoning: true;
|
|
7023
|
-
input: "text"[];
|
|
7152
|
+
input: ("image" | "text")[];
|
|
7024
7153
|
cost: {
|
|
7025
7154
|
input: number;
|
|
7026
7155
|
output: number;
|
|
@@ -7030,10 +7159,12 @@ export declare const MODELS: {
|
|
|
7030
7159
|
contextWindow: number;
|
|
7031
7160
|
maxTokens: number;
|
|
7032
7161
|
};
|
|
7033
|
-
|
|
7162
|
+
};
|
|
7163
|
+
readonly "opencode-go": {
|
|
7164
|
+
readonly "deepseek-v4-flash": {
|
|
7034
7165
|
id: string;
|
|
7035
7166
|
name: string;
|
|
7036
|
-
api: "
|
|
7167
|
+
api: "openai-completions";
|
|
7037
7168
|
provider: string;
|
|
7038
7169
|
baseUrl: string;
|
|
7039
7170
|
reasoning: true;
|
|
@@ -7047,14 +7178,14 @@ export declare const MODELS: {
|
|
|
7047
7178
|
contextWindow: number;
|
|
7048
7179
|
maxTokens: number;
|
|
7049
7180
|
};
|
|
7050
|
-
readonly "
|
|
7181
|
+
readonly "deepseek-v4-pro": {
|
|
7051
7182
|
id: string;
|
|
7052
7183
|
name: string;
|
|
7053
7184
|
api: "openai-completions";
|
|
7054
7185
|
provider: string;
|
|
7055
7186
|
baseUrl: string;
|
|
7056
7187
|
reasoning: true;
|
|
7057
|
-
input:
|
|
7188
|
+
input: "text"[];
|
|
7058
7189
|
cost: {
|
|
7059
7190
|
input: number;
|
|
7060
7191
|
output: number;
|
|
@@ -7064,14 +7195,14 @@ export declare const MODELS: {
|
|
|
7064
7195
|
contextWindow: number;
|
|
7065
7196
|
maxTokens: number;
|
|
7066
7197
|
};
|
|
7067
|
-
readonly "
|
|
7198
|
+
readonly "glm-5": {
|
|
7068
7199
|
id: string;
|
|
7069
7200
|
name: string;
|
|
7070
7201
|
api: "openai-completions";
|
|
7071
7202
|
provider: string;
|
|
7072
7203
|
baseUrl: string;
|
|
7073
7204
|
reasoning: true;
|
|
7074
|
-
input:
|
|
7205
|
+
input: "text"[];
|
|
7075
7206
|
cost: {
|
|
7076
7207
|
input: number;
|
|
7077
7208
|
output: number;
|
|
@@ -7081,15 +7212,13 @@ export declare const MODELS: {
|
|
|
7081
7212
|
contextWindow: number;
|
|
7082
7213
|
maxTokens: number;
|
|
7083
7214
|
};
|
|
7084
|
-
|
|
7085
|
-
readonly openrouter: {
|
|
7086
|
-
readonly "ai21/jamba-large-1.7": {
|
|
7215
|
+
readonly "glm-5.1": {
|
|
7087
7216
|
id: string;
|
|
7088
7217
|
name: string;
|
|
7089
7218
|
api: "openai-completions";
|
|
7090
7219
|
provider: string;
|
|
7091
7220
|
baseUrl: string;
|
|
7092
|
-
reasoning:
|
|
7221
|
+
reasoning: true;
|
|
7093
7222
|
input: "text"[];
|
|
7094
7223
|
cost: {
|
|
7095
7224
|
input: number;
|
|
@@ -7100,14 +7229,14 @@ export declare const MODELS: {
|
|
|
7100
7229
|
contextWindow: number;
|
|
7101
7230
|
maxTokens: number;
|
|
7102
7231
|
};
|
|
7103
|
-
readonly "
|
|
7232
|
+
readonly "kimi-k2.5": {
|
|
7104
7233
|
id: string;
|
|
7105
7234
|
name: string;
|
|
7106
7235
|
api: "openai-completions";
|
|
7107
7236
|
provider: string;
|
|
7108
7237
|
baseUrl: string;
|
|
7109
7238
|
reasoning: true;
|
|
7110
|
-
input: "text"[];
|
|
7239
|
+
input: ("image" | "text")[];
|
|
7111
7240
|
cost: {
|
|
7112
7241
|
input: number;
|
|
7113
7242
|
output: number;
|
|
@@ -7117,14 +7246,14 @@ export declare const MODELS: {
|
|
|
7117
7246
|
contextWindow: number;
|
|
7118
7247
|
maxTokens: number;
|
|
7119
7248
|
};
|
|
7120
|
-
readonly "
|
|
7249
|
+
readonly "kimi-k2.6": {
|
|
7121
7250
|
id: string;
|
|
7122
7251
|
name: string;
|
|
7123
7252
|
api: "openai-completions";
|
|
7124
7253
|
provider: string;
|
|
7125
7254
|
baseUrl: string;
|
|
7126
|
-
reasoning:
|
|
7127
|
-
input: "text"[];
|
|
7255
|
+
reasoning: true;
|
|
7256
|
+
input: ("image" | "text")[];
|
|
7128
7257
|
cost: {
|
|
7129
7258
|
input: number;
|
|
7130
7259
|
output: number;
|
|
@@ -7134,7 +7263,7 @@ export declare const MODELS: {
|
|
|
7134
7263
|
contextWindow: number;
|
|
7135
7264
|
maxTokens: number;
|
|
7136
7265
|
};
|
|
7137
|
-
readonly "
|
|
7266
|
+
readonly "mimo-v2-omni": {
|
|
7138
7267
|
id: string;
|
|
7139
7268
|
name: string;
|
|
7140
7269
|
api: "openai-completions";
|
|
@@ -7151,14 +7280,14 @@ export declare const MODELS: {
|
|
|
7151
7280
|
contextWindow: number;
|
|
7152
7281
|
maxTokens: number;
|
|
7153
7282
|
};
|
|
7154
|
-
readonly "
|
|
7283
|
+
readonly "mimo-v2-pro": {
|
|
7155
7284
|
id: string;
|
|
7156
7285
|
name: string;
|
|
7157
7286
|
api: "openai-completions";
|
|
7158
7287
|
provider: string;
|
|
7159
7288
|
baseUrl: string;
|
|
7160
|
-
reasoning:
|
|
7161
|
-
input:
|
|
7289
|
+
reasoning: true;
|
|
7290
|
+
input: "text"[];
|
|
7162
7291
|
cost: {
|
|
7163
7292
|
input: number;
|
|
7164
7293
|
output: number;
|
|
@@ -7168,14 +7297,14 @@ export declare const MODELS: {
|
|
|
7168
7297
|
contextWindow: number;
|
|
7169
7298
|
maxTokens: number;
|
|
7170
7299
|
};
|
|
7171
|
-
readonly "
|
|
7300
|
+
readonly "mimo-v2.5": {
|
|
7172
7301
|
id: string;
|
|
7173
7302
|
name: string;
|
|
7174
7303
|
api: "openai-completions";
|
|
7175
7304
|
provider: string;
|
|
7176
7305
|
baseUrl: string;
|
|
7177
|
-
reasoning:
|
|
7178
|
-
input: "text"[];
|
|
7306
|
+
reasoning: true;
|
|
7307
|
+
input: ("image" | "text")[];
|
|
7179
7308
|
cost: {
|
|
7180
7309
|
input: number;
|
|
7181
7310
|
output: number;
|
|
@@ -7185,14 +7314,14 @@ export declare const MODELS: {
|
|
|
7185
7314
|
contextWindow: number;
|
|
7186
7315
|
maxTokens: number;
|
|
7187
7316
|
};
|
|
7188
|
-
readonly "
|
|
7317
|
+
readonly "mimo-v2.5-pro": {
|
|
7189
7318
|
id: string;
|
|
7190
7319
|
name: string;
|
|
7191
7320
|
api: "openai-completions";
|
|
7192
7321
|
provider: string;
|
|
7193
7322
|
baseUrl: string;
|
|
7194
|
-
reasoning:
|
|
7195
|
-
input:
|
|
7323
|
+
reasoning: true;
|
|
7324
|
+
input: "text"[];
|
|
7196
7325
|
cost: {
|
|
7197
7326
|
input: number;
|
|
7198
7327
|
output: number;
|
|
@@ -7202,14 +7331,203 @@ export declare const MODELS: {
|
|
|
7202
7331
|
contextWindow: number;
|
|
7203
7332
|
maxTokens: number;
|
|
7204
7333
|
};
|
|
7205
|
-
readonly "
|
|
7334
|
+
readonly "minimax-m2.5": {
|
|
7206
7335
|
id: string;
|
|
7207
7336
|
name: string;
|
|
7208
7337
|
api: "openai-completions";
|
|
7209
7338
|
provider: string;
|
|
7210
7339
|
baseUrl: string;
|
|
7211
|
-
reasoning:
|
|
7212
|
-
input:
|
|
7340
|
+
reasoning: true;
|
|
7341
|
+
input: "text"[];
|
|
7342
|
+
cost: {
|
|
7343
|
+
input: number;
|
|
7344
|
+
output: number;
|
|
7345
|
+
cacheRead: number;
|
|
7346
|
+
cacheWrite: number;
|
|
7347
|
+
};
|
|
7348
|
+
contextWindow: number;
|
|
7349
|
+
maxTokens: number;
|
|
7350
|
+
};
|
|
7351
|
+
readonly "minimax-m2.7": {
|
|
7352
|
+
id: string;
|
|
7353
|
+
name: string;
|
|
7354
|
+
api: "anthropic-messages";
|
|
7355
|
+
provider: string;
|
|
7356
|
+
baseUrl: string;
|
|
7357
|
+
reasoning: true;
|
|
7358
|
+
input: "text"[];
|
|
7359
|
+
cost: {
|
|
7360
|
+
input: number;
|
|
7361
|
+
output: number;
|
|
7362
|
+
cacheRead: number;
|
|
7363
|
+
cacheWrite: number;
|
|
7364
|
+
};
|
|
7365
|
+
contextWindow: number;
|
|
7366
|
+
maxTokens: number;
|
|
7367
|
+
};
|
|
7368
|
+
readonly "qwen3.5-plus": {
|
|
7369
|
+
id: string;
|
|
7370
|
+
name: string;
|
|
7371
|
+
api: "anthropic-messages";
|
|
7372
|
+
provider: string;
|
|
7373
|
+
baseUrl: string;
|
|
7374
|
+
reasoning: true;
|
|
7375
|
+
input: ("image" | "text")[];
|
|
7376
|
+
cost: {
|
|
7377
|
+
input: number;
|
|
7378
|
+
output: number;
|
|
7379
|
+
cacheRead: number;
|
|
7380
|
+
cacheWrite: number;
|
|
7381
|
+
};
|
|
7382
|
+
contextWindow: number;
|
|
7383
|
+
maxTokens: number;
|
|
7384
|
+
};
|
|
7385
|
+
readonly "qwen3.6-plus": {
|
|
7386
|
+
id: string;
|
|
7387
|
+
name: string;
|
|
7388
|
+
api: "anthropic-messages";
|
|
7389
|
+
provider: string;
|
|
7390
|
+
baseUrl: string;
|
|
7391
|
+
reasoning: true;
|
|
7392
|
+
input: ("image" | "text")[];
|
|
7393
|
+
cost: {
|
|
7394
|
+
input: number;
|
|
7395
|
+
output: number;
|
|
7396
|
+
cacheRead: number;
|
|
7397
|
+
cacheWrite: number;
|
|
7398
|
+
};
|
|
7399
|
+
contextWindow: number;
|
|
7400
|
+
maxTokens: number;
|
|
7401
|
+
};
|
|
7402
|
+
};
|
|
7403
|
+
readonly openrouter: {
|
|
7404
|
+
readonly "ai21/jamba-large-1.7": {
|
|
7405
|
+
id: string;
|
|
7406
|
+
name: string;
|
|
7407
|
+
api: "openai-completions";
|
|
7408
|
+
provider: string;
|
|
7409
|
+
baseUrl: string;
|
|
7410
|
+
reasoning: false;
|
|
7411
|
+
input: "text"[];
|
|
7412
|
+
cost: {
|
|
7413
|
+
input: number;
|
|
7414
|
+
output: number;
|
|
7415
|
+
cacheRead: number;
|
|
7416
|
+
cacheWrite: number;
|
|
7417
|
+
};
|
|
7418
|
+
contextWindow: number;
|
|
7419
|
+
maxTokens: number;
|
|
7420
|
+
};
|
|
7421
|
+
readonly "alibaba/tongyi-deepresearch-30b-a3b": {
|
|
7422
|
+
id: string;
|
|
7423
|
+
name: string;
|
|
7424
|
+
api: "openai-completions";
|
|
7425
|
+
provider: string;
|
|
7426
|
+
baseUrl: string;
|
|
7427
|
+
reasoning: true;
|
|
7428
|
+
input: "text"[];
|
|
7429
|
+
cost: {
|
|
7430
|
+
input: number;
|
|
7431
|
+
output: number;
|
|
7432
|
+
cacheRead: number;
|
|
7433
|
+
cacheWrite: number;
|
|
7434
|
+
};
|
|
7435
|
+
contextWindow: number;
|
|
7436
|
+
maxTokens: number;
|
|
7437
|
+
};
|
|
7438
|
+
readonly "allenai/olmo-3.1-32b-instruct": {
|
|
7439
|
+
id: string;
|
|
7440
|
+
name: string;
|
|
7441
|
+
api: "openai-completions";
|
|
7442
|
+
provider: string;
|
|
7443
|
+
baseUrl: string;
|
|
7444
|
+
reasoning: false;
|
|
7445
|
+
input: "text"[];
|
|
7446
|
+
cost: {
|
|
7447
|
+
input: number;
|
|
7448
|
+
output: number;
|
|
7449
|
+
cacheRead: number;
|
|
7450
|
+
cacheWrite: number;
|
|
7451
|
+
};
|
|
7452
|
+
contextWindow: number;
|
|
7453
|
+
maxTokens: number;
|
|
7454
|
+
};
|
|
7455
|
+
readonly "amazon/nova-2-lite-v1": {
|
|
7456
|
+
id: string;
|
|
7457
|
+
name: string;
|
|
7458
|
+
api: "openai-completions";
|
|
7459
|
+
provider: string;
|
|
7460
|
+
baseUrl: string;
|
|
7461
|
+
reasoning: true;
|
|
7462
|
+
input: ("image" | "text")[];
|
|
7463
|
+
cost: {
|
|
7464
|
+
input: number;
|
|
7465
|
+
output: number;
|
|
7466
|
+
cacheRead: number;
|
|
7467
|
+
cacheWrite: number;
|
|
7468
|
+
};
|
|
7469
|
+
contextWindow: number;
|
|
7470
|
+
maxTokens: number;
|
|
7471
|
+
};
|
|
7472
|
+
readonly "amazon/nova-lite-v1": {
|
|
7473
|
+
id: string;
|
|
7474
|
+
name: string;
|
|
7475
|
+
api: "openai-completions";
|
|
7476
|
+
provider: string;
|
|
7477
|
+
baseUrl: string;
|
|
7478
|
+
reasoning: false;
|
|
7479
|
+
input: ("image" | "text")[];
|
|
7480
|
+
cost: {
|
|
7481
|
+
input: number;
|
|
7482
|
+
output: number;
|
|
7483
|
+
cacheRead: number;
|
|
7484
|
+
cacheWrite: number;
|
|
7485
|
+
};
|
|
7486
|
+
contextWindow: number;
|
|
7487
|
+
maxTokens: number;
|
|
7488
|
+
};
|
|
7489
|
+
readonly "amazon/nova-micro-v1": {
|
|
7490
|
+
id: string;
|
|
7491
|
+
name: string;
|
|
7492
|
+
api: "openai-completions";
|
|
7493
|
+
provider: string;
|
|
7494
|
+
baseUrl: string;
|
|
7495
|
+
reasoning: false;
|
|
7496
|
+
input: "text"[];
|
|
7497
|
+
cost: {
|
|
7498
|
+
input: number;
|
|
7499
|
+
output: number;
|
|
7500
|
+
cacheRead: number;
|
|
7501
|
+
cacheWrite: number;
|
|
7502
|
+
};
|
|
7503
|
+
contextWindow: number;
|
|
7504
|
+
maxTokens: number;
|
|
7505
|
+
};
|
|
7506
|
+
readonly "amazon/nova-premier-v1": {
|
|
7507
|
+
id: string;
|
|
7508
|
+
name: string;
|
|
7509
|
+
api: "openai-completions";
|
|
7510
|
+
provider: string;
|
|
7511
|
+
baseUrl: string;
|
|
7512
|
+
reasoning: false;
|
|
7513
|
+
input: ("image" | "text")[];
|
|
7514
|
+
cost: {
|
|
7515
|
+
input: number;
|
|
7516
|
+
output: number;
|
|
7517
|
+
cacheRead: number;
|
|
7518
|
+
cacheWrite: number;
|
|
7519
|
+
};
|
|
7520
|
+
contextWindow: number;
|
|
7521
|
+
maxTokens: number;
|
|
7522
|
+
};
|
|
7523
|
+
readonly "amazon/nova-pro-v1": {
|
|
7524
|
+
id: string;
|
|
7525
|
+
name: string;
|
|
7526
|
+
api: "openai-completions";
|
|
7527
|
+
provider: string;
|
|
7528
|
+
baseUrl: string;
|
|
7529
|
+
reasoning: false;
|
|
7530
|
+
input: ("image" | "text")[];
|
|
7213
7531
|
cost: {
|
|
7214
7532
|
input: number;
|
|
7215
7533
|
output: number;
|
|
@@ -7457,7 +7775,7 @@ export declare const MODELS: {
|
|
|
7457
7775
|
contextWindow: number;
|
|
7458
7776
|
maxTokens: number;
|
|
7459
7777
|
};
|
|
7460
|
-
readonly "arcee-ai/trinity-large-preview
|
|
7778
|
+
readonly "arcee-ai/trinity-large-preview": {
|
|
7461
7779
|
id: string;
|
|
7462
7780
|
name: string;
|
|
7463
7781
|
api: "openai-completions";
|
|
@@ -7814,6 +8132,40 @@ export declare const MODELS: {
|
|
|
7814
8132
|
contextWindow: number;
|
|
7815
8133
|
maxTokens: number;
|
|
7816
8134
|
};
|
|
8135
|
+
readonly "deepseek/deepseek-v4-flash": {
|
|
8136
|
+
id: string;
|
|
8137
|
+
name: string;
|
|
8138
|
+
api: "openai-completions";
|
|
8139
|
+
provider: string;
|
|
8140
|
+
baseUrl: string;
|
|
8141
|
+
reasoning: true;
|
|
8142
|
+
input: "text"[];
|
|
8143
|
+
cost: {
|
|
8144
|
+
input: number;
|
|
8145
|
+
output: number;
|
|
8146
|
+
cacheRead: number;
|
|
8147
|
+
cacheWrite: number;
|
|
8148
|
+
};
|
|
8149
|
+
contextWindow: number;
|
|
8150
|
+
maxTokens: number;
|
|
8151
|
+
};
|
|
8152
|
+
readonly "deepseek/deepseek-v4-pro": {
|
|
8153
|
+
id: string;
|
|
8154
|
+
name: string;
|
|
8155
|
+
api: "openai-completions";
|
|
8156
|
+
provider: string;
|
|
8157
|
+
baseUrl: string;
|
|
8158
|
+
reasoning: true;
|
|
8159
|
+
input: "text"[];
|
|
8160
|
+
cost: {
|
|
8161
|
+
input: number;
|
|
8162
|
+
output: number;
|
|
8163
|
+
cacheRead: number;
|
|
8164
|
+
cacheWrite: number;
|
|
8165
|
+
};
|
|
8166
|
+
contextWindow: number;
|
|
8167
|
+
maxTokens: number;
|
|
8168
|
+
};
|
|
7817
8169
|
readonly "essentialai/rnj-1-instruct": {
|
|
7818
8170
|
id: string;
|
|
7819
8171
|
name: string;
|
|
@@ -8035,6 +8387,40 @@ export declare const MODELS: {
|
|
|
8035
8387
|
contextWindow: number;
|
|
8036
8388
|
maxTokens: number;
|
|
8037
8389
|
};
|
|
8390
|
+
readonly "google/gemma-3-12b-it": {
|
|
8391
|
+
id: string;
|
|
8392
|
+
name: string;
|
|
8393
|
+
api: "openai-completions";
|
|
8394
|
+
provider: string;
|
|
8395
|
+
baseUrl: string;
|
|
8396
|
+
reasoning: false;
|
|
8397
|
+
input: ("image" | "text")[];
|
|
8398
|
+
cost: {
|
|
8399
|
+
input: number;
|
|
8400
|
+
output: number;
|
|
8401
|
+
cacheRead: number;
|
|
8402
|
+
cacheWrite: number;
|
|
8403
|
+
};
|
|
8404
|
+
contextWindow: number;
|
|
8405
|
+
maxTokens: number;
|
|
8406
|
+
};
|
|
8407
|
+
readonly "google/gemma-3-27b-it": {
|
|
8408
|
+
id: string;
|
|
8409
|
+
name: string;
|
|
8410
|
+
api: "openai-completions";
|
|
8411
|
+
provider: string;
|
|
8412
|
+
baseUrl: string;
|
|
8413
|
+
reasoning: false;
|
|
8414
|
+
input: ("image" | "text")[];
|
|
8415
|
+
cost: {
|
|
8416
|
+
input: number;
|
|
8417
|
+
output: number;
|
|
8418
|
+
cacheRead: number;
|
|
8419
|
+
cacheWrite: number;
|
|
8420
|
+
};
|
|
8421
|
+
contextWindow: number;
|
|
8422
|
+
maxTokens: number;
|
|
8423
|
+
};
|
|
8038
8424
|
readonly "google/gemma-4-26b-a4b-it": {
|
|
8039
8425
|
id: string;
|
|
8040
8426
|
name: string;
|
|
@@ -8103,13 +8489,47 @@ export declare const MODELS: {
|
|
|
8103
8489
|
contextWindow: number;
|
|
8104
8490
|
maxTokens: number;
|
|
8105
8491
|
};
|
|
8106
|
-
readonly "inception/mercury-2": {
|
|
8492
|
+
readonly "inception/mercury-2": {
|
|
8493
|
+
id: string;
|
|
8494
|
+
name: string;
|
|
8495
|
+
api: "openai-completions";
|
|
8496
|
+
provider: string;
|
|
8497
|
+
baseUrl: string;
|
|
8498
|
+
reasoning: true;
|
|
8499
|
+
input: "text"[];
|
|
8500
|
+
cost: {
|
|
8501
|
+
input: number;
|
|
8502
|
+
output: number;
|
|
8503
|
+
cacheRead: number;
|
|
8504
|
+
cacheWrite: number;
|
|
8505
|
+
};
|
|
8506
|
+
contextWindow: number;
|
|
8507
|
+
maxTokens: number;
|
|
8508
|
+
};
|
|
8509
|
+
readonly "inclusionai/ling-2.6-1t:free": {
|
|
8510
|
+
id: string;
|
|
8511
|
+
name: string;
|
|
8512
|
+
api: "openai-completions";
|
|
8513
|
+
provider: string;
|
|
8514
|
+
baseUrl: string;
|
|
8515
|
+
reasoning: false;
|
|
8516
|
+
input: "text"[];
|
|
8517
|
+
cost: {
|
|
8518
|
+
input: number;
|
|
8519
|
+
output: number;
|
|
8520
|
+
cacheRead: number;
|
|
8521
|
+
cacheWrite: number;
|
|
8522
|
+
};
|
|
8523
|
+
contextWindow: number;
|
|
8524
|
+
maxTokens: number;
|
|
8525
|
+
};
|
|
8526
|
+
readonly "inclusionai/ling-2.6-flash:free": {
|
|
8107
8527
|
id: string;
|
|
8108
8528
|
name: string;
|
|
8109
8529
|
api: "openai-completions";
|
|
8110
8530
|
provider: string;
|
|
8111
8531
|
baseUrl: string;
|
|
8112
|
-
reasoning:
|
|
8532
|
+
reasoning: false;
|
|
8113
8533
|
input: "text"[];
|
|
8114
8534
|
cost: {
|
|
8115
8535
|
input: number;
|
|
@@ -8783,6 +9203,23 @@ export declare const MODELS: {
|
|
|
8783
9203
|
contextWindow: number;
|
|
8784
9204
|
maxTokens: number;
|
|
8785
9205
|
};
|
|
9206
|
+
readonly "moonshotai/kimi-k2.6": {
|
|
9207
|
+
id: string;
|
|
9208
|
+
name: string;
|
|
9209
|
+
api: "openai-completions";
|
|
9210
|
+
provider: string;
|
|
9211
|
+
baseUrl: string;
|
|
9212
|
+
reasoning: true;
|
|
9213
|
+
input: ("image" | "text")[];
|
|
9214
|
+
cost: {
|
|
9215
|
+
input: number;
|
|
9216
|
+
output: number;
|
|
9217
|
+
cacheRead: number;
|
|
9218
|
+
cacheWrite: number;
|
|
9219
|
+
};
|
|
9220
|
+
contextWindow: number;
|
|
9221
|
+
maxTokens: number;
|
|
9222
|
+
};
|
|
8786
9223
|
readonly "nex-agi/deepseek-v3.1-nex-n1": {
|
|
8787
9224
|
id: string;
|
|
8788
9225
|
name: string;
|
|
@@ -9191,13 +9628,251 @@ export declare const MODELS: {
|
|
|
9191
9628
|
contextWindow: number;
|
|
9192
9629
|
maxTokens: number;
|
|
9193
9630
|
};
|
|
9194
|
-
readonly "openai/gpt-4o-2024-11-20": {
|
|
9631
|
+
readonly "openai/gpt-4o-2024-11-20": {
|
|
9632
|
+
id: string;
|
|
9633
|
+
name: string;
|
|
9634
|
+
api: "openai-completions";
|
|
9635
|
+
provider: string;
|
|
9636
|
+
baseUrl: string;
|
|
9637
|
+
reasoning: false;
|
|
9638
|
+
input: ("image" | "text")[];
|
|
9639
|
+
cost: {
|
|
9640
|
+
input: number;
|
|
9641
|
+
output: number;
|
|
9642
|
+
cacheRead: number;
|
|
9643
|
+
cacheWrite: number;
|
|
9644
|
+
};
|
|
9645
|
+
contextWindow: number;
|
|
9646
|
+
maxTokens: number;
|
|
9647
|
+
};
|
|
9648
|
+
readonly "openai/gpt-4o-audio-preview": {
|
|
9649
|
+
id: string;
|
|
9650
|
+
name: string;
|
|
9651
|
+
api: "openai-completions";
|
|
9652
|
+
provider: string;
|
|
9653
|
+
baseUrl: string;
|
|
9654
|
+
reasoning: false;
|
|
9655
|
+
input: "text"[];
|
|
9656
|
+
cost: {
|
|
9657
|
+
input: number;
|
|
9658
|
+
output: number;
|
|
9659
|
+
cacheRead: number;
|
|
9660
|
+
cacheWrite: number;
|
|
9661
|
+
};
|
|
9662
|
+
contextWindow: number;
|
|
9663
|
+
maxTokens: number;
|
|
9664
|
+
};
|
|
9665
|
+
readonly "openai/gpt-4o-mini": {
|
|
9666
|
+
id: string;
|
|
9667
|
+
name: string;
|
|
9668
|
+
api: "openai-completions";
|
|
9669
|
+
provider: string;
|
|
9670
|
+
baseUrl: string;
|
|
9671
|
+
reasoning: false;
|
|
9672
|
+
input: ("image" | "text")[];
|
|
9673
|
+
cost: {
|
|
9674
|
+
input: number;
|
|
9675
|
+
output: number;
|
|
9676
|
+
cacheRead: number;
|
|
9677
|
+
cacheWrite: number;
|
|
9678
|
+
};
|
|
9679
|
+
contextWindow: number;
|
|
9680
|
+
maxTokens: number;
|
|
9681
|
+
};
|
|
9682
|
+
readonly "openai/gpt-4o-mini-2024-07-18": {
|
|
9683
|
+
id: string;
|
|
9684
|
+
name: string;
|
|
9685
|
+
api: "openai-completions";
|
|
9686
|
+
provider: string;
|
|
9687
|
+
baseUrl: string;
|
|
9688
|
+
reasoning: false;
|
|
9689
|
+
input: ("image" | "text")[];
|
|
9690
|
+
cost: {
|
|
9691
|
+
input: number;
|
|
9692
|
+
output: number;
|
|
9693
|
+
cacheRead: number;
|
|
9694
|
+
cacheWrite: number;
|
|
9695
|
+
};
|
|
9696
|
+
contextWindow: number;
|
|
9697
|
+
maxTokens: number;
|
|
9698
|
+
};
|
|
9699
|
+
readonly "openai/gpt-5": {
|
|
9700
|
+
id: string;
|
|
9701
|
+
name: string;
|
|
9702
|
+
api: "openai-completions";
|
|
9703
|
+
provider: string;
|
|
9704
|
+
baseUrl: string;
|
|
9705
|
+
reasoning: true;
|
|
9706
|
+
input: ("image" | "text")[];
|
|
9707
|
+
cost: {
|
|
9708
|
+
input: number;
|
|
9709
|
+
output: number;
|
|
9710
|
+
cacheRead: number;
|
|
9711
|
+
cacheWrite: number;
|
|
9712
|
+
};
|
|
9713
|
+
contextWindow: number;
|
|
9714
|
+
maxTokens: number;
|
|
9715
|
+
};
|
|
9716
|
+
readonly "openai/gpt-5-codex": {
|
|
9717
|
+
id: string;
|
|
9718
|
+
name: string;
|
|
9719
|
+
api: "openai-completions";
|
|
9720
|
+
provider: string;
|
|
9721
|
+
baseUrl: string;
|
|
9722
|
+
reasoning: true;
|
|
9723
|
+
input: ("image" | "text")[];
|
|
9724
|
+
cost: {
|
|
9725
|
+
input: number;
|
|
9726
|
+
output: number;
|
|
9727
|
+
cacheRead: number;
|
|
9728
|
+
cacheWrite: number;
|
|
9729
|
+
};
|
|
9730
|
+
contextWindow: number;
|
|
9731
|
+
maxTokens: number;
|
|
9732
|
+
};
|
|
9733
|
+
readonly "openai/gpt-5-mini": {
|
|
9734
|
+
id: string;
|
|
9735
|
+
name: string;
|
|
9736
|
+
api: "openai-completions";
|
|
9737
|
+
provider: string;
|
|
9738
|
+
baseUrl: string;
|
|
9739
|
+
reasoning: true;
|
|
9740
|
+
input: ("image" | "text")[];
|
|
9741
|
+
cost: {
|
|
9742
|
+
input: number;
|
|
9743
|
+
output: number;
|
|
9744
|
+
cacheRead: number;
|
|
9745
|
+
cacheWrite: number;
|
|
9746
|
+
};
|
|
9747
|
+
contextWindow: number;
|
|
9748
|
+
maxTokens: number;
|
|
9749
|
+
};
|
|
9750
|
+
readonly "openai/gpt-5-nano": {
|
|
9751
|
+
id: string;
|
|
9752
|
+
name: string;
|
|
9753
|
+
api: "openai-completions";
|
|
9754
|
+
provider: string;
|
|
9755
|
+
baseUrl: string;
|
|
9756
|
+
reasoning: true;
|
|
9757
|
+
input: ("image" | "text")[];
|
|
9758
|
+
cost: {
|
|
9759
|
+
input: number;
|
|
9760
|
+
output: number;
|
|
9761
|
+
cacheRead: number;
|
|
9762
|
+
cacheWrite: number;
|
|
9763
|
+
};
|
|
9764
|
+
contextWindow: number;
|
|
9765
|
+
maxTokens: number;
|
|
9766
|
+
};
|
|
9767
|
+
readonly "openai/gpt-5-pro": {
|
|
9768
|
+
id: string;
|
|
9769
|
+
name: string;
|
|
9770
|
+
api: "openai-completions";
|
|
9771
|
+
provider: string;
|
|
9772
|
+
baseUrl: string;
|
|
9773
|
+
reasoning: true;
|
|
9774
|
+
input: ("image" | "text")[];
|
|
9775
|
+
cost: {
|
|
9776
|
+
input: number;
|
|
9777
|
+
output: number;
|
|
9778
|
+
cacheRead: number;
|
|
9779
|
+
cacheWrite: number;
|
|
9780
|
+
};
|
|
9781
|
+
contextWindow: number;
|
|
9782
|
+
maxTokens: number;
|
|
9783
|
+
};
|
|
9784
|
+
readonly "openai/gpt-5.1": {
|
|
9785
|
+
id: string;
|
|
9786
|
+
name: string;
|
|
9787
|
+
api: "openai-completions";
|
|
9788
|
+
provider: string;
|
|
9789
|
+
baseUrl: string;
|
|
9790
|
+
reasoning: true;
|
|
9791
|
+
input: ("image" | "text")[];
|
|
9792
|
+
cost: {
|
|
9793
|
+
input: number;
|
|
9794
|
+
output: number;
|
|
9795
|
+
cacheRead: number;
|
|
9796
|
+
cacheWrite: number;
|
|
9797
|
+
};
|
|
9798
|
+
contextWindow: number;
|
|
9799
|
+
maxTokens: number;
|
|
9800
|
+
};
|
|
9801
|
+
readonly "openai/gpt-5.1-chat": {
|
|
9802
|
+
id: string;
|
|
9803
|
+
name: string;
|
|
9804
|
+
api: "openai-completions";
|
|
9805
|
+
provider: string;
|
|
9806
|
+
baseUrl: string;
|
|
9807
|
+
reasoning: false;
|
|
9808
|
+
input: ("image" | "text")[];
|
|
9809
|
+
cost: {
|
|
9810
|
+
input: number;
|
|
9811
|
+
output: number;
|
|
9812
|
+
cacheRead: number;
|
|
9813
|
+
cacheWrite: number;
|
|
9814
|
+
};
|
|
9815
|
+
contextWindow: number;
|
|
9816
|
+
maxTokens: number;
|
|
9817
|
+
};
|
|
9818
|
+
readonly "openai/gpt-5.1-codex": {
|
|
9819
|
+
id: string;
|
|
9820
|
+
name: string;
|
|
9821
|
+
api: "openai-completions";
|
|
9822
|
+
provider: string;
|
|
9823
|
+
baseUrl: string;
|
|
9824
|
+
reasoning: true;
|
|
9825
|
+
input: ("image" | "text")[];
|
|
9826
|
+
cost: {
|
|
9827
|
+
input: number;
|
|
9828
|
+
output: number;
|
|
9829
|
+
cacheRead: number;
|
|
9830
|
+
cacheWrite: number;
|
|
9831
|
+
};
|
|
9832
|
+
contextWindow: number;
|
|
9833
|
+
maxTokens: number;
|
|
9834
|
+
};
|
|
9835
|
+
readonly "openai/gpt-5.1-codex-max": {
|
|
9836
|
+
id: string;
|
|
9837
|
+
name: string;
|
|
9838
|
+
api: "openai-completions";
|
|
9839
|
+
provider: string;
|
|
9840
|
+
baseUrl: string;
|
|
9841
|
+
reasoning: true;
|
|
9842
|
+
input: ("image" | "text")[];
|
|
9843
|
+
cost: {
|
|
9844
|
+
input: number;
|
|
9845
|
+
output: number;
|
|
9846
|
+
cacheRead: number;
|
|
9847
|
+
cacheWrite: number;
|
|
9848
|
+
};
|
|
9849
|
+
contextWindow: number;
|
|
9850
|
+
maxTokens: number;
|
|
9851
|
+
};
|
|
9852
|
+
readonly "openai/gpt-5.1-codex-mini": {
|
|
9853
|
+
id: string;
|
|
9854
|
+
name: string;
|
|
9855
|
+
api: "openai-completions";
|
|
9856
|
+
provider: string;
|
|
9857
|
+
baseUrl: string;
|
|
9858
|
+
reasoning: true;
|
|
9859
|
+
input: ("image" | "text")[];
|
|
9860
|
+
cost: {
|
|
9861
|
+
input: number;
|
|
9862
|
+
output: number;
|
|
9863
|
+
cacheRead: number;
|
|
9864
|
+
cacheWrite: number;
|
|
9865
|
+
};
|
|
9866
|
+
contextWindow: number;
|
|
9867
|
+
maxTokens: number;
|
|
9868
|
+
};
|
|
9869
|
+
readonly "openai/gpt-5.2": {
|
|
9195
9870
|
id: string;
|
|
9196
9871
|
name: string;
|
|
9197
9872
|
api: "openai-completions";
|
|
9198
9873
|
provider: string;
|
|
9199
9874
|
baseUrl: string;
|
|
9200
|
-
reasoning:
|
|
9875
|
+
reasoning: true;
|
|
9201
9876
|
input: ("image" | "text")[];
|
|
9202
9877
|
cost: {
|
|
9203
9878
|
input: number;
|
|
@@ -9208,14 +9883,14 @@ export declare const MODELS: {
|
|
|
9208
9883
|
contextWindow: number;
|
|
9209
9884
|
maxTokens: number;
|
|
9210
9885
|
};
|
|
9211
|
-
readonly "openai/gpt-
|
|
9886
|
+
readonly "openai/gpt-5.2-chat": {
|
|
9212
9887
|
id: string;
|
|
9213
9888
|
name: string;
|
|
9214
9889
|
api: "openai-completions";
|
|
9215
9890
|
provider: string;
|
|
9216
9891
|
baseUrl: string;
|
|
9217
9892
|
reasoning: false;
|
|
9218
|
-
input: "text"[];
|
|
9893
|
+
input: ("image" | "text")[];
|
|
9219
9894
|
cost: {
|
|
9220
9895
|
input: number;
|
|
9221
9896
|
output: number;
|
|
@@ -9225,13 +9900,13 @@ export declare const MODELS: {
|
|
|
9225
9900
|
contextWindow: number;
|
|
9226
9901
|
maxTokens: number;
|
|
9227
9902
|
};
|
|
9228
|
-
readonly "openai/gpt-
|
|
9903
|
+
readonly "openai/gpt-5.2-codex": {
|
|
9229
9904
|
id: string;
|
|
9230
9905
|
name: string;
|
|
9231
9906
|
api: "openai-completions";
|
|
9232
9907
|
provider: string;
|
|
9233
9908
|
baseUrl: string;
|
|
9234
|
-
reasoning:
|
|
9909
|
+
reasoning: true;
|
|
9235
9910
|
input: ("image" | "text")[];
|
|
9236
9911
|
cost: {
|
|
9237
9912
|
input: number;
|
|
@@ -9242,13 +9917,13 @@ export declare const MODELS: {
|
|
|
9242
9917
|
contextWindow: number;
|
|
9243
9918
|
maxTokens: number;
|
|
9244
9919
|
};
|
|
9245
|
-
readonly "openai/gpt-
|
|
9920
|
+
readonly "openai/gpt-5.2-pro": {
|
|
9246
9921
|
id: string;
|
|
9247
9922
|
name: string;
|
|
9248
9923
|
api: "openai-completions";
|
|
9249
9924
|
provider: string;
|
|
9250
9925
|
baseUrl: string;
|
|
9251
|
-
reasoning:
|
|
9926
|
+
reasoning: true;
|
|
9252
9927
|
input: ("image" | "text")[];
|
|
9253
9928
|
cost: {
|
|
9254
9929
|
input: number;
|
|
@@ -9259,13 +9934,13 @@ export declare const MODELS: {
|
|
|
9259
9934
|
contextWindow: number;
|
|
9260
9935
|
maxTokens: number;
|
|
9261
9936
|
};
|
|
9262
|
-
readonly "openai/gpt-5": {
|
|
9937
|
+
readonly "openai/gpt-5.3-chat": {
|
|
9263
9938
|
id: string;
|
|
9264
9939
|
name: string;
|
|
9265
9940
|
api: "openai-completions";
|
|
9266
9941
|
provider: string;
|
|
9267
9942
|
baseUrl: string;
|
|
9268
|
-
reasoning:
|
|
9943
|
+
reasoning: false;
|
|
9269
9944
|
input: ("image" | "text")[];
|
|
9270
9945
|
cost: {
|
|
9271
9946
|
input: number;
|
|
@@ -9276,7 +9951,7 @@ export declare const MODELS: {
|
|
|
9276
9951
|
contextWindow: number;
|
|
9277
9952
|
maxTokens: number;
|
|
9278
9953
|
};
|
|
9279
|
-
readonly "openai/gpt-5-codex": {
|
|
9954
|
+
readonly "openai/gpt-5.3-codex": {
|
|
9280
9955
|
id: string;
|
|
9281
9956
|
name: string;
|
|
9282
9957
|
api: "openai-completions";
|
|
@@ -9293,7 +9968,7 @@ export declare const MODELS: {
|
|
|
9293
9968
|
contextWindow: number;
|
|
9294
9969
|
maxTokens: number;
|
|
9295
9970
|
};
|
|
9296
|
-
readonly "openai/gpt-5
|
|
9971
|
+
readonly "openai/gpt-5.4": {
|
|
9297
9972
|
id: string;
|
|
9298
9973
|
name: string;
|
|
9299
9974
|
api: "openai-completions";
|
|
@@ -9310,7 +9985,7 @@ export declare const MODELS: {
|
|
|
9310
9985
|
contextWindow: number;
|
|
9311
9986
|
maxTokens: number;
|
|
9312
9987
|
};
|
|
9313
|
-
readonly "openai/gpt-5-
|
|
9988
|
+
readonly "openai/gpt-5.4-mini": {
|
|
9314
9989
|
id: string;
|
|
9315
9990
|
name: string;
|
|
9316
9991
|
api: "openai-completions";
|
|
@@ -9327,7 +10002,7 @@ export declare const MODELS: {
|
|
|
9327
10002
|
contextWindow: number;
|
|
9328
10003
|
maxTokens: number;
|
|
9329
10004
|
};
|
|
9330
|
-
readonly "openai/gpt-5-
|
|
10005
|
+
readonly "openai/gpt-5.4-nano": {
|
|
9331
10006
|
id: string;
|
|
9332
10007
|
name: string;
|
|
9333
10008
|
api: "openai-completions";
|
|
@@ -9344,7 +10019,7 @@ export declare const MODELS: {
|
|
|
9344
10019
|
contextWindow: number;
|
|
9345
10020
|
maxTokens: number;
|
|
9346
10021
|
};
|
|
9347
|
-
readonly "openai/gpt-5-
|
|
10022
|
+
readonly "openai/gpt-5.4-pro": {
|
|
9348
10023
|
id: string;
|
|
9349
10024
|
name: string;
|
|
9350
10025
|
api: "openai-completions";
|
|
@@ -9361,7 +10036,7 @@ export declare const MODELS: {
|
|
|
9361
10036
|
contextWindow: number;
|
|
9362
10037
|
maxTokens: number;
|
|
9363
10038
|
};
|
|
9364
|
-
readonly "openai/gpt-5
|
|
10039
|
+
readonly "openai/gpt-5.5": {
|
|
9365
10040
|
id: string;
|
|
9366
10041
|
name: string;
|
|
9367
10042
|
api: "openai-completions";
|
|
@@ -9378,7 +10053,7 @@ export declare const MODELS: {
|
|
|
9378
10053
|
contextWindow: number;
|
|
9379
10054
|
maxTokens: number;
|
|
9380
10055
|
};
|
|
9381
|
-
readonly "openai/gpt-5.
|
|
10056
|
+
readonly "openai/gpt-5.5-pro": {
|
|
9382
10057
|
id: string;
|
|
9383
10058
|
name: string;
|
|
9384
10059
|
api: "openai-completions";
|
|
@@ -9395,14 +10070,14 @@ export declare const MODELS: {
|
|
|
9395
10070
|
contextWindow: number;
|
|
9396
10071
|
maxTokens: number;
|
|
9397
10072
|
};
|
|
9398
|
-
readonly "openai/gpt-
|
|
10073
|
+
readonly "openai/gpt-audio": {
|
|
9399
10074
|
id: string;
|
|
9400
10075
|
name: string;
|
|
9401
10076
|
api: "openai-completions";
|
|
9402
10077
|
provider: string;
|
|
9403
10078
|
baseUrl: string;
|
|
9404
10079
|
reasoning: false;
|
|
9405
|
-
input:
|
|
10080
|
+
input: "text"[];
|
|
9406
10081
|
cost: {
|
|
9407
10082
|
input: number;
|
|
9408
10083
|
output: number;
|
|
@@ -9412,14 +10087,14 @@ export declare const MODELS: {
|
|
|
9412
10087
|
contextWindow: number;
|
|
9413
10088
|
maxTokens: number;
|
|
9414
10089
|
};
|
|
9415
|
-
readonly "openai/gpt-
|
|
10090
|
+
readonly "openai/gpt-audio-mini": {
|
|
9416
10091
|
id: string;
|
|
9417
10092
|
name: string;
|
|
9418
10093
|
api: "openai-completions";
|
|
9419
10094
|
provider: string;
|
|
9420
10095
|
baseUrl: string;
|
|
9421
|
-
reasoning:
|
|
9422
|
-
input:
|
|
10096
|
+
reasoning: false;
|
|
10097
|
+
input: "text"[];
|
|
9423
10098
|
cost: {
|
|
9424
10099
|
input: number;
|
|
9425
10100
|
output: number;
|
|
@@ -9429,14 +10104,14 @@ export declare const MODELS: {
|
|
|
9429
10104
|
contextWindow: number;
|
|
9430
10105
|
maxTokens: number;
|
|
9431
10106
|
};
|
|
9432
|
-
readonly "openai/gpt-
|
|
10107
|
+
readonly "openai/gpt-oss-120b": {
|
|
9433
10108
|
id: string;
|
|
9434
10109
|
name: string;
|
|
9435
10110
|
api: "openai-completions";
|
|
9436
10111
|
provider: string;
|
|
9437
10112
|
baseUrl: string;
|
|
9438
10113
|
reasoning: true;
|
|
9439
|
-
input:
|
|
10114
|
+
input: "text"[];
|
|
9440
10115
|
cost: {
|
|
9441
10116
|
input: number;
|
|
9442
10117
|
output: number;
|
|
@@ -9446,14 +10121,14 @@ export declare const MODELS: {
|
|
|
9446
10121
|
contextWindow: number;
|
|
9447
10122
|
maxTokens: number;
|
|
9448
10123
|
};
|
|
9449
|
-
readonly "openai/gpt-
|
|
10124
|
+
readonly "openai/gpt-oss-120b:free": {
|
|
9450
10125
|
id: string;
|
|
9451
10126
|
name: string;
|
|
9452
10127
|
api: "openai-completions";
|
|
9453
10128
|
provider: string;
|
|
9454
10129
|
baseUrl: string;
|
|
9455
10130
|
reasoning: true;
|
|
9456
|
-
input:
|
|
10131
|
+
input: "text"[];
|
|
9457
10132
|
cost: {
|
|
9458
10133
|
input: number;
|
|
9459
10134
|
output: number;
|
|
@@ -9463,14 +10138,14 @@ export declare const MODELS: {
|
|
|
9463
10138
|
contextWindow: number;
|
|
9464
10139
|
maxTokens: number;
|
|
9465
10140
|
};
|
|
9466
|
-
readonly "openai/gpt-
|
|
10141
|
+
readonly "openai/gpt-oss-20b": {
|
|
9467
10142
|
id: string;
|
|
9468
10143
|
name: string;
|
|
9469
10144
|
api: "openai-completions";
|
|
9470
10145
|
provider: string;
|
|
9471
10146
|
baseUrl: string;
|
|
9472
10147
|
reasoning: true;
|
|
9473
|
-
input:
|
|
10148
|
+
input: "text"[];
|
|
9474
10149
|
cost: {
|
|
9475
10150
|
input: number;
|
|
9476
10151
|
output: number;
|
|
@@ -9480,14 +10155,14 @@ export declare const MODELS: {
|
|
|
9480
10155
|
contextWindow: number;
|
|
9481
10156
|
maxTokens: number;
|
|
9482
10157
|
};
|
|
9483
|
-
readonly "openai/gpt-
|
|
10158
|
+
readonly "openai/gpt-oss-20b:free": {
|
|
9484
10159
|
id: string;
|
|
9485
10160
|
name: string;
|
|
9486
10161
|
api: "openai-completions";
|
|
9487
10162
|
provider: string;
|
|
9488
10163
|
baseUrl: string;
|
|
9489
|
-
reasoning:
|
|
9490
|
-
input:
|
|
10164
|
+
reasoning: true;
|
|
10165
|
+
input: "text"[];
|
|
9491
10166
|
cost: {
|
|
9492
10167
|
input: number;
|
|
9493
10168
|
output: number;
|
|
@@ -9497,14 +10172,14 @@ export declare const MODELS: {
|
|
|
9497
10172
|
contextWindow: number;
|
|
9498
10173
|
maxTokens: number;
|
|
9499
10174
|
};
|
|
9500
|
-
readonly "openai/gpt-
|
|
10175
|
+
readonly "openai/gpt-oss-safeguard-20b": {
|
|
9501
10176
|
id: string;
|
|
9502
10177
|
name: string;
|
|
9503
10178
|
api: "openai-completions";
|
|
9504
10179
|
provider: string;
|
|
9505
10180
|
baseUrl: string;
|
|
9506
10181
|
reasoning: true;
|
|
9507
|
-
input:
|
|
10182
|
+
input: "text"[];
|
|
9508
10183
|
cost: {
|
|
9509
10184
|
input: number;
|
|
9510
10185
|
output: number;
|
|
@@ -9514,7 +10189,7 @@ export declare const MODELS: {
|
|
|
9514
10189
|
contextWindow: number;
|
|
9515
10190
|
maxTokens: number;
|
|
9516
10191
|
};
|
|
9517
|
-
readonly "openai/
|
|
10192
|
+
readonly "openai/o1": {
|
|
9518
10193
|
id: string;
|
|
9519
10194
|
name: string;
|
|
9520
10195
|
api: "openai-completions";
|
|
@@ -9531,13 +10206,13 @@ export declare const MODELS: {
|
|
|
9531
10206
|
contextWindow: number;
|
|
9532
10207
|
maxTokens: number;
|
|
9533
10208
|
};
|
|
9534
|
-
readonly "openai/
|
|
10209
|
+
readonly "openai/o3": {
|
|
9535
10210
|
id: string;
|
|
9536
10211
|
name: string;
|
|
9537
10212
|
api: "openai-completions";
|
|
9538
10213
|
provider: string;
|
|
9539
10214
|
baseUrl: string;
|
|
9540
|
-
reasoning:
|
|
10215
|
+
reasoning: true;
|
|
9541
10216
|
input: ("image" | "text")[];
|
|
9542
10217
|
cost: {
|
|
9543
10218
|
input: number;
|
|
@@ -9548,7 +10223,7 @@ export declare const MODELS: {
|
|
|
9548
10223
|
contextWindow: number;
|
|
9549
10224
|
maxTokens: number;
|
|
9550
10225
|
};
|
|
9551
|
-
readonly "openai/
|
|
10226
|
+
readonly "openai/o3-deep-research": {
|
|
9552
10227
|
id: string;
|
|
9553
10228
|
name: string;
|
|
9554
10229
|
api: "openai-completions";
|
|
@@ -9565,14 +10240,14 @@ export declare const MODELS: {
|
|
|
9565
10240
|
contextWindow: number;
|
|
9566
10241
|
maxTokens: number;
|
|
9567
10242
|
};
|
|
9568
|
-
readonly "openai/
|
|
10243
|
+
readonly "openai/o3-mini": {
|
|
9569
10244
|
id: string;
|
|
9570
10245
|
name: string;
|
|
9571
10246
|
api: "openai-completions";
|
|
9572
10247
|
provider: string;
|
|
9573
10248
|
baseUrl: string;
|
|
9574
10249
|
reasoning: true;
|
|
9575
|
-
input:
|
|
10250
|
+
input: "text"[];
|
|
9576
10251
|
cost: {
|
|
9577
10252
|
input: number;
|
|
9578
10253
|
output: number;
|
|
@@ -9582,14 +10257,14 @@ export declare const MODELS: {
|
|
|
9582
10257
|
contextWindow: number;
|
|
9583
10258
|
maxTokens: number;
|
|
9584
10259
|
};
|
|
9585
|
-
readonly "openai/
|
|
10260
|
+
readonly "openai/o3-mini-high": {
|
|
9586
10261
|
id: string;
|
|
9587
10262
|
name: string;
|
|
9588
10263
|
api: "openai-completions";
|
|
9589
10264
|
provider: string;
|
|
9590
10265
|
baseUrl: string;
|
|
9591
10266
|
reasoning: true;
|
|
9592
|
-
input:
|
|
10267
|
+
input: "text"[];
|
|
9593
10268
|
cost: {
|
|
9594
10269
|
input: number;
|
|
9595
10270
|
output: number;
|
|
@@ -9599,7 +10274,7 @@ export declare const MODELS: {
|
|
|
9599
10274
|
contextWindow: number;
|
|
9600
10275
|
maxTokens: number;
|
|
9601
10276
|
};
|
|
9602
|
-
readonly "openai/
|
|
10277
|
+
readonly "openai/o3-pro": {
|
|
9603
10278
|
id: string;
|
|
9604
10279
|
name: string;
|
|
9605
10280
|
api: "openai-completions";
|
|
@@ -9616,7 +10291,7 @@ export declare const MODELS: {
|
|
|
9616
10291
|
contextWindow: number;
|
|
9617
10292
|
maxTokens: number;
|
|
9618
10293
|
};
|
|
9619
|
-
readonly "openai/
|
|
10294
|
+
readonly "openai/o4-mini": {
|
|
9620
10295
|
id: string;
|
|
9621
10296
|
name: string;
|
|
9622
10297
|
api: "openai-completions";
|
|
@@ -9633,14 +10308,14 @@ export declare const MODELS: {
|
|
|
9633
10308
|
contextWindow: number;
|
|
9634
10309
|
maxTokens: number;
|
|
9635
10310
|
};
|
|
9636
|
-
readonly "openai/
|
|
10311
|
+
readonly "openai/o4-mini-deep-research": {
|
|
9637
10312
|
id: string;
|
|
9638
10313
|
name: string;
|
|
9639
10314
|
api: "openai-completions";
|
|
9640
10315
|
provider: string;
|
|
9641
10316
|
baseUrl: string;
|
|
9642
|
-
reasoning:
|
|
9643
|
-
input: "text"[];
|
|
10317
|
+
reasoning: true;
|
|
10318
|
+
input: ("image" | "text")[];
|
|
9644
10319
|
cost: {
|
|
9645
10320
|
input: number;
|
|
9646
10321
|
output: number;
|
|
@@ -9650,14 +10325,14 @@ export declare const MODELS: {
|
|
|
9650
10325
|
contextWindow: number;
|
|
9651
10326
|
maxTokens: number;
|
|
9652
10327
|
};
|
|
9653
|
-
readonly "openai/
|
|
10328
|
+
readonly "openai/o4-mini-high": {
|
|
9654
10329
|
id: string;
|
|
9655
10330
|
name: string;
|
|
9656
10331
|
api: "openai-completions";
|
|
9657
10332
|
provider: string;
|
|
9658
10333
|
baseUrl: string;
|
|
9659
|
-
reasoning:
|
|
9660
|
-
input: "text"[];
|
|
10334
|
+
reasoning: true;
|
|
10335
|
+
input: ("image" | "text")[];
|
|
9661
10336
|
cost: {
|
|
9662
10337
|
input: number;
|
|
9663
10338
|
output: number;
|
|
@@ -9667,14 +10342,14 @@ export declare const MODELS: {
|
|
|
9667
10342
|
contextWindow: number;
|
|
9668
10343
|
maxTokens: number;
|
|
9669
10344
|
};
|
|
9670
|
-
readonly "
|
|
10345
|
+
readonly "openrouter/auto": {
|
|
9671
10346
|
id: string;
|
|
9672
10347
|
name: string;
|
|
9673
10348
|
api: "openai-completions";
|
|
9674
10349
|
provider: string;
|
|
9675
10350
|
baseUrl: string;
|
|
9676
10351
|
reasoning: true;
|
|
9677
|
-
input: "text"[];
|
|
10352
|
+
input: ("image" | "text")[];
|
|
9678
10353
|
cost: {
|
|
9679
10354
|
input: number;
|
|
9680
10355
|
output: number;
|
|
@@ -9684,14 +10359,14 @@ export declare const MODELS: {
|
|
|
9684
10359
|
contextWindow: number;
|
|
9685
10360
|
maxTokens: number;
|
|
9686
10361
|
};
|
|
9687
|
-
readonly "
|
|
10362
|
+
readonly "openrouter/free": {
|
|
9688
10363
|
id: string;
|
|
9689
10364
|
name: string;
|
|
9690
10365
|
api: "openai-completions";
|
|
9691
10366
|
provider: string;
|
|
9692
10367
|
baseUrl: string;
|
|
9693
10368
|
reasoning: true;
|
|
9694
|
-
input: "text"[];
|
|
10369
|
+
input: ("image" | "text")[];
|
|
9695
10370
|
cost: {
|
|
9696
10371
|
input: number;
|
|
9697
10372
|
output: number;
|
|
@@ -9701,7 +10376,7 @@ export declare const MODELS: {
|
|
|
9701
10376
|
contextWindow: number;
|
|
9702
10377
|
maxTokens: number;
|
|
9703
10378
|
};
|
|
9704
|
-
readonly "
|
|
10379
|
+
readonly "prime-intellect/intellect-3": {
|
|
9705
10380
|
id: string;
|
|
9706
10381
|
name: string;
|
|
9707
10382
|
api: "openai-completions";
|
|
@@ -9718,13 +10393,13 @@ export declare const MODELS: {
|
|
|
9718
10393
|
contextWindow: number;
|
|
9719
10394
|
maxTokens: number;
|
|
9720
10395
|
};
|
|
9721
|
-
readonly "
|
|
10396
|
+
readonly "qwen/qwen-2.5-72b-instruct": {
|
|
9722
10397
|
id: string;
|
|
9723
10398
|
name: string;
|
|
9724
10399
|
api: "openai-completions";
|
|
9725
10400
|
provider: string;
|
|
9726
10401
|
baseUrl: string;
|
|
9727
|
-
reasoning:
|
|
10402
|
+
reasoning: false;
|
|
9728
10403
|
input: "text"[];
|
|
9729
10404
|
cost: {
|
|
9730
10405
|
input: number;
|
|
@@ -9735,13 +10410,13 @@ export declare const MODELS: {
|
|
|
9735
10410
|
contextWindow: number;
|
|
9736
10411
|
maxTokens: number;
|
|
9737
10412
|
};
|
|
9738
|
-
readonly "
|
|
10413
|
+
readonly "qwen/qwen-2.5-7b-instruct": {
|
|
9739
10414
|
id: string;
|
|
9740
10415
|
name: string;
|
|
9741
10416
|
api: "openai-completions";
|
|
9742
10417
|
provider: string;
|
|
9743
10418
|
baseUrl: string;
|
|
9744
|
-
reasoning:
|
|
10419
|
+
reasoning: false;
|
|
9745
10420
|
input: "text"[];
|
|
9746
10421
|
cost: {
|
|
9747
10422
|
input: number;
|
|
@@ -9752,14 +10427,14 @@ export declare const MODELS: {
|
|
|
9752
10427
|
contextWindow: number;
|
|
9753
10428
|
maxTokens: number;
|
|
9754
10429
|
};
|
|
9755
|
-
readonly "
|
|
10430
|
+
readonly "qwen/qwen-max": {
|
|
9756
10431
|
id: string;
|
|
9757
10432
|
name: string;
|
|
9758
10433
|
api: "openai-completions";
|
|
9759
10434
|
provider: string;
|
|
9760
10435
|
baseUrl: string;
|
|
9761
|
-
reasoning:
|
|
9762
|
-
input:
|
|
10436
|
+
reasoning: false;
|
|
10437
|
+
input: "text"[];
|
|
9763
10438
|
cost: {
|
|
9764
10439
|
input: number;
|
|
9765
10440
|
output: number;
|
|
@@ -9769,14 +10444,14 @@ export declare const MODELS: {
|
|
|
9769
10444
|
contextWindow: number;
|
|
9770
10445
|
maxTokens: number;
|
|
9771
10446
|
};
|
|
9772
|
-
readonly "
|
|
10447
|
+
readonly "qwen/qwen-plus": {
|
|
9773
10448
|
id: string;
|
|
9774
10449
|
name: string;
|
|
9775
10450
|
api: "openai-completions";
|
|
9776
10451
|
provider: string;
|
|
9777
10452
|
baseUrl: string;
|
|
9778
|
-
reasoning:
|
|
9779
|
-
input:
|
|
10453
|
+
reasoning: false;
|
|
10454
|
+
input: "text"[];
|
|
9780
10455
|
cost: {
|
|
9781
10456
|
input: number;
|
|
9782
10457
|
output: number;
|
|
@@ -9786,14 +10461,14 @@ export declare const MODELS: {
|
|
|
9786
10461
|
contextWindow: number;
|
|
9787
10462
|
maxTokens: number;
|
|
9788
10463
|
};
|
|
9789
|
-
readonly "
|
|
10464
|
+
readonly "qwen/qwen-plus-2025-07-28": {
|
|
9790
10465
|
id: string;
|
|
9791
10466
|
name: string;
|
|
9792
10467
|
api: "openai-completions";
|
|
9793
10468
|
provider: string;
|
|
9794
10469
|
baseUrl: string;
|
|
9795
|
-
reasoning:
|
|
9796
|
-
input:
|
|
10470
|
+
reasoning: false;
|
|
10471
|
+
input: "text"[];
|
|
9797
10472
|
cost: {
|
|
9798
10473
|
input: number;
|
|
9799
10474
|
output: number;
|
|
@@ -9803,7 +10478,7 @@ export declare const MODELS: {
|
|
|
9803
10478
|
contextWindow: number;
|
|
9804
10479
|
maxTokens: number;
|
|
9805
10480
|
};
|
|
9806
|
-
readonly "
|
|
10481
|
+
readonly "qwen/qwen-plus-2025-07-28:thinking": {
|
|
9807
10482
|
id: string;
|
|
9808
10483
|
name: string;
|
|
9809
10484
|
api: "openai-completions";
|
|
@@ -9820,13 +10495,13 @@ export declare const MODELS: {
|
|
|
9820
10495
|
contextWindow: number;
|
|
9821
10496
|
maxTokens: number;
|
|
9822
10497
|
};
|
|
9823
|
-
readonly "
|
|
10498
|
+
readonly "qwen/qwen-turbo": {
|
|
9824
10499
|
id: string;
|
|
9825
10500
|
name: string;
|
|
9826
10501
|
api: "openai-completions";
|
|
9827
10502
|
provider: string;
|
|
9828
10503
|
baseUrl: string;
|
|
9829
|
-
reasoning:
|
|
10504
|
+
reasoning: false;
|
|
9830
10505
|
input: "text"[];
|
|
9831
10506
|
cost: {
|
|
9832
10507
|
input: number;
|
|
@@ -9837,13 +10512,13 @@ export declare const MODELS: {
|
|
|
9837
10512
|
contextWindow: number;
|
|
9838
10513
|
maxTokens: number;
|
|
9839
10514
|
};
|
|
9840
|
-
readonly "
|
|
10515
|
+
readonly "qwen/qwen-vl-max": {
|
|
9841
10516
|
id: string;
|
|
9842
10517
|
name: string;
|
|
9843
10518
|
api: "openai-completions";
|
|
9844
10519
|
provider: string;
|
|
9845
10520
|
baseUrl: string;
|
|
9846
|
-
reasoning:
|
|
10521
|
+
reasoning: false;
|
|
9847
10522
|
input: ("image" | "text")[];
|
|
9848
10523
|
cost: {
|
|
9849
10524
|
input: number;
|
|
@@ -9854,14 +10529,14 @@ export declare const MODELS: {
|
|
|
9854
10529
|
contextWindow: number;
|
|
9855
10530
|
maxTokens: number;
|
|
9856
10531
|
};
|
|
9857
|
-
readonly "
|
|
10532
|
+
readonly "qwen/qwen3-14b": {
|
|
9858
10533
|
id: string;
|
|
9859
10534
|
name: string;
|
|
9860
10535
|
api: "openai-completions";
|
|
9861
10536
|
provider: string;
|
|
9862
10537
|
baseUrl: string;
|
|
9863
10538
|
reasoning: true;
|
|
9864
|
-
input:
|
|
10539
|
+
input: "text"[];
|
|
9865
10540
|
cost: {
|
|
9866
10541
|
input: number;
|
|
9867
10542
|
output: number;
|
|
@@ -9871,14 +10546,14 @@ export declare const MODELS: {
|
|
|
9871
10546
|
contextWindow: number;
|
|
9872
10547
|
maxTokens: number;
|
|
9873
10548
|
};
|
|
9874
|
-
readonly "
|
|
10549
|
+
readonly "qwen/qwen3-235b-a22b": {
|
|
9875
10550
|
id: string;
|
|
9876
10551
|
name: string;
|
|
9877
10552
|
api: "openai-completions";
|
|
9878
10553
|
provider: string;
|
|
9879
10554
|
baseUrl: string;
|
|
9880
10555
|
reasoning: true;
|
|
9881
|
-
input:
|
|
10556
|
+
input: "text"[];
|
|
9882
10557
|
cost: {
|
|
9883
10558
|
input: number;
|
|
9884
10559
|
output: number;
|
|
@@ -9888,14 +10563,14 @@ export declare const MODELS: {
|
|
|
9888
10563
|
contextWindow: number;
|
|
9889
10564
|
maxTokens: number;
|
|
9890
10565
|
};
|
|
9891
|
-
readonly "
|
|
10566
|
+
readonly "qwen/qwen3-235b-a22b-2507": {
|
|
9892
10567
|
id: string;
|
|
9893
10568
|
name: string;
|
|
9894
10569
|
api: "openai-completions";
|
|
9895
10570
|
provider: string;
|
|
9896
10571
|
baseUrl: string;
|
|
9897
|
-
reasoning:
|
|
9898
|
-
input:
|
|
10572
|
+
reasoning: false;
|
|
10573
|
+
input: "text"[];
|
|
9899
10574
|
cost: {
|
|
9900
10575
|
input: number;
|
|
9901
10576
|
output: number;
|
|
@@ -9905,14 +10580,14 @@ export declare const MODELS: {
|
|
|
9905
10580
|
contextWindow: number;
|
|
9906
10581
|
maxTokens: number;
|
|
9907
10582
|
};
|
|
9908
|
-
readonly "
|
|
10583
|
+
readonly "qwen/qwen3-235b-a22b-thinking-2507": {
|
|
9909
10584
|
id: string;
|
|
9910
10585
|
name: string;
|
|
9911
10586
|
api: "openai-completions";
|
|
9912
10587
|
provider: string;
|
|
9913
10588
|
baseUrl: string;
|
|
9914
10589
|
reasoning: true;
|
|
9915
|
-
input:
|
|
10590
|
+
input: "text"[];
|
|
9916
10591
|
cost: {
|
|
9917
10592
|
input: number;
|
|
9918
10593
|
output: number;
|
|
@@ -9922,13 +10597,13 @@ export declare const MODELS: {
|
|
|
9922
10597
|
contextWindow: number;
|
|
9923
10598
|
maxTokens: number;
|
|
9924
10599
|
};
|
|
9925
|
-
readonly "
|
|
10600
|
+
readonly "qwen/qwen3-30b-a3b": {
|
|
9926
10601
|
id: string;
|
|
9927
10602
|
name: string;
|
|
9928
10603
|
api: "openai-completions";
|
|
9929
10604
|
provider: string;
|
|
9930
10605
|
baseUrl: string;
|
|
9931
|
-
reasoning:
|
|
10606
|
+
reasoning: true;
|
|
9932
10607
|
input: "text"[];
|
|
9933
10608
|
cost: {
|
|
9934
10609
|
input: number;
|
|
@@ -9939,14 +10614,14 @@ export declare const MODELS: {
|
|
|
9939
10614
|
contextWindow: number;
|
|
9940
10615
|
maxTokens: number;
|
|
9941
10616
|
};
|
|
9942
|
-
readonly "
|
|
10617
|
+
readonly "qwen/qwen3-30b-a3b-instruct-2507": {
|
|
9943
10618
|
id: string;
|
|
9944
10619
|
name: string;
|
|
9945
10620
|
api: "openai-completions";
|
|
9946
10621
|
provider: string;
|
|
9947
10622
|
baseUrl: string;
|
|
9948
|
-
reasoning:
|
|
9949
|
-
input:
|
|
10623
|
+
reasoning: false;
|
|
10624
|
+
input: "text"[];
|
|
9950
10625
|
cost: {
|
|
9951
10626
|
input: number;
|
|
9952
10627
|
output: number;
|
|
@@ -9956,7 +10631,7 @@ export declare const MODELS: {
|
|
|
9956
10631
|
contextWindow: number;
|
|
9957
10632
|
maxTokens: number;
|
|
9958
10633
|
};
|
|
9959
|
-
readonly "
|
|
10634
|
+
readonly "qwen/qwen3-30b-a3b-thinking-2507": {
|
|
9960
10635
|
id: string;
|
|
9961
10636
|
name: string;
|
|
9962
10637
|
api: "openai-completions";
|
|
@@ -9973,13 +10648,13 @@ export declare const MODELS: {
|
|
|
9973
10648
|
contextWindow: number;
|
|
9974
10649
|
maxTokens: number;
|
|
9975
10650
|
};
|
|
9976
|
-
readonly "qwen/
|
|
10651
|
+
readonly "qwen/qwen3-32b": {
|
|
9977
10652
|
id: string;
|
|
9978
10653
|
name: string;
|
|
9979
10654
|
api: "openai-completions";
|
|
9980
10655
|
provider: string;
|
|
9981
10656
|
baseUrl: string;
|
|
9982
|
-
reasoning:
|
|
10657
|
+
reasoning: true;
|
|
9983
10658
|
input: "text"[];
|
|
9984
10659
|
cost: {
|
|
9985
10660
|
input: number;
|
|
@@ -9990,13 +10665,13 @@ export declare const MODELS: {
|
|
|
9990
10665
|
contextWindow: number;
|
|
9991
10666
|
maxTokens: number;
|
|
9992
10667
|
};
|
|
9993
|
-
readonly "qwen/
|
|
10668
|
+
readonly "qwen/qwen3-8b": {
|
|
9994
10669
|
id: string;
|
|
9995
10670
|
name: string;
|
|
9996
10671
|
api: "openai-completions";
|
|
9997
10672
|
provider: string;
|
|
9998
10673
|
baseUrl: string;
|
|
9999
|
-
reasoning:
|
|
10674
|
+
reasoning: true;
|
|
10000
10675
|
input: "text"[];
|
|
10001
10676
|
cost: {
|
|
10002
10677
|
input: number;
|
|
@@ -10007,7 +10682,7 @@ export declare const MODELS: {
|
|
|
10007
10682
|
contextWindow: number;
|
|
10008
10683
|
maxTokens: number;
|
|
10009
10684
|
};
|
|
10010
|
-
readonly "qwen/
|
|
10685
|
+
readonly "qwen/qwen3-coder": {
|
|
10011
10686
|
id: string;
|
|
10012
10687
|
name: string;
|
|
10013
10688
|
api: "openai-completions";
|
|
@@ -10024,7 +10699,7 @@ export declare const MODELS: {
|
|
|
10024
10699
|
contextWindow: number;
|
|
10025
10700
|
maxTokens: number;
|
|
10026
10701
|
};
|
|
10027
|
-
readonly "qwen/
|
|
10702
|
+
readonly "qwen/qwen3-coder-30b-a3b-instruct": {
|
|
10028
10703
|
id: string;
|
|
10029
10704
|
name: string;
|
|
10030
10705
|
api: "openai-completions";
|
|
@@ -10041,7 +10716,7 @@ export declare const MODELS: {
|
|
|
10041
10716
|
contextWindow: number;
|
|
10042
10717
|
maxTokens: number;
|
|
10043
10718
|
};
|
|
10044
|
-
readonly "qwen/
|
|
10719
|
+
readonly "qwen/qwen3-coder-flash": {
|
|
10045
10720
|
id: string;
|
|
10046
10721
|
name: string;
|
|
10047
10722
|
api: "openai-completions";
|
|
@@ -10058,13 +10733,13 @@ export declare const MODELS: {
|
|
|
10058
10733
|
contextWindow: number;
|
|
10059
10734
|
maxTokens: number;
|
|
10060
10735
|
};
|
|
10061
|
-
readonly "qwen/
|
|
10736
|
+
readonly "qwen/qwen3-coder-next": {
|
|
10062
10737
|
id: string;
|
|
10063
10738
|
name: string;
|
|
10064
10739
|
api: "openai-completions";
|
|
10065
10740
|
provider: string;
|
|
10066
10741
|
baseUrl: string;
|
|
10067
|
-
reasoning:
|
|
10742
|
+
reasoning: false;
|
|
10068
10743
|
input: "text"[];
|
|
10069
10744
|
cost: {
|
|
10070
10745
|
input: number;
|
|
@@ -10075,7 +10750,7 @@ export declare const MODELS: {
|
|
|
10075
10750
|
contextWindow: number;
|
|
10076
10751
|
maxTokens: number;
|
|
10077
10752
|
};
|
|
10078
|
-
readonly "qwen/
|
|
10753
|
+
readonly "qwen/qwen3-coder-plus": {
|
|
10079
10754
|
id: string;
|
|
10080
10755
|
name: string;
|
|
10081
10756
|
api: "openai-completions";
|
|
@@ -10092,14 +10767,14 @@ export declare const MODELS: {
|
|
|
10092
10767
|
contextWindow: number;
|
|
10093
10768
|
maxTokens: number;
|
|
10094
10769
|
};
|
|
10095
|
-
readonly "qwen/
|
|
10770
|
+
readonly "qwen/qwen3-coder:free": {
|
|
10096
10771
|
id: string;
|
|
10097
10772
|
name: string;
|
|
10098
10773
|
api: "openai-completions";
|
|
10099
10774
|
provider: string;
|
|
10100
10775
|
baseUrl: string;
|
|
10101
10776
|
reasoning: false;
|
|
10102
|
-
input:
|
|
10777
|
+
input: "text"[];
|
|
10103
10778
|
cost: {
|
|
10104
10779
|
input: number;
|
|
10105
10780
|
output: number;
|
|
@@ -10109,13 +10784,13 @@ export declare const MODELS: {
|
|
|
10109
10784
|
contextWindow: number;
|
|
10110
10785
|
maxTokens: number;
|
|
10111
10786
|
};
|
|
10112
|
-
readonly "qwen/qwen3-
|
|
10787
|
+
readonly "qwen/qwen3-max": {
|
|
10113
10788
|
id: string;
|
|
10114
10789
|
name: string;
|
|
10115
10790
|
api: "openai-completions";
|
|
10116
10791
|
provider: string;
|
|
10117
10792
|
baseUrl: string;
|
|
10118
|
-
reasoning:
|
|
10793
|
+
reasoning: false;
|
|
10119
10794
|
input: "text"[];
|
|
10120
10795
|
cost: {
|
|
10121
10796
|
input: number;
|
|
@@ -10126,7 +10801,7 @@ export declare const MODELS: {
|
|
|
10126
10801
|
contextWindow: number;
|
|
10127
10802
|
maxTokens: number;
|
|
10128
10803
|
};
|
|
10129
|
-
readonly "qwen/qwen3-
|
|
10804
|
+
readonly "qwen/qwen3-max-thinking": {
|
|
10130
10805
|
id: string;
|
|
10131
10806
|
name: string;
|
|
10132
10807
|
api: "openai-completions";
|
|
@@ -10143,7 +10818,7 @@ export declare const MODELS: {
|
|
|
10143
10818
|
contextWindow: number;
|
|
10144
10819
|
maxTokens: number;
|
|
10145
10820
|
};
|
|
10146
|
-
readonly "qwen/qwen3-
|
|
10821
|
+
readonly "qwen/qwen3-next-80b-a3b-instruct": {
|
|
10147
10822
|
id: string;
|
|
10148
10823
|
name: string;
|
|
10149
10824
|
api: "openai-completions";
|
|
@@ -10160,13 +10835,13 @@ export declare const MODELS: {
|
|
|
10160
10835
|
contextWindow: number;
|
|
10161
10836
|
maxTokens: number;
|
|
10162
10837
|
};
|
|
10163
|
-
readonly "qwen/qwen3-
|
|
10838
|
+
readonly "qwen/qwen3-next-80b-a3b-instruct:free": {
|
|
10164
10839
|
id: string;
|
|
10165
10840
|
name: string;
|
|
10166
10841
|
api: "openai-completions";
|
|
10167
10842
|
provider: string;
|
|
10168
10843
|
baseUrl: string;
|
|
10169
|
-
reasoning:
|
|
10844
|
+
reasoning: false;
|
|
10170
10845
|
input: "text"[];
|
|
10171
10846
|
cost: {
|
|
10172
10847
|
input: number;
|
|
@@ -10177,7 +10852,7 @@ export declare const MODELS: {
|
|
|
10177
10852
|
contextWindow: number;
|
|
10178
10853
|
maxTokens: number;
|
|
10179
10854
|
};
|
|
10180
|
-
readonly "qwen/qwen3-
|
|
10855
|
+
readonly "qwen/qwen3-next-80b-a3b-thinking": {
|
|
10181
10856
|
id: string;
|
|
10182
10857
|
name: string;
|
|
10183
10858
|
api: "openai-completions";
|
|
@@ -10194,14 +10869,14 @@ export declare const MODELS: {
|
|
|
10194
10869
|
contextWindow: number;
|
|
10195
10870
|
maxTokens: number;
|
|
10196
10871
|
};
|
|
10197
|
-
readonly "qwen/qwen3-
|
|
10872
|
+
readonly "qwen/qwen3-vl-235b-a22b-instruct": {
|
|
10198
10873
|
id: string;
|
|
10199
10874
|
name: string;
|
|
10200
10875
|
api: "openai-completions";
|
|
10201
10876
|
provider: string;
|
|
10202
10877
|
baseUrl: string;
|
|
10203
10878
|
reasoning: false;
|
|
10204
|
-
input: "text"[];
|
|
10879
|
+
input: ("image" | "text")[];
|
|
10205
10880
|
cost: {
|
|
10206
10881
|
input: number;
|
|
10207
10882
|
output: number;
|
|
@@ -10211,14 +10886,14 @@ export declare const MODELS: {
|
|
|
10211
10886
|
contextWindow: number;
|
|
10212
10887
|
maxTokens: number;
|
|
10213
10888
|
};
|
|
10214
|
-
readonly "qwen/qwen3-
|
|
10889
|
+
readonly "qwen/qwen3-vl-235b-a22b-thinking": {
|
|
10215
10890
|
id: string;
|
|
10216
10891
|
name: string;
|
|
10217
10892
|
api: "openai-completions";
|
|
10218
10893
|
provider: string;
|
|
10219
10894
|
baseUrl: string;
|
|
10220
10895
|
reasoning: true;
|
|
10221
|
-
input: "text"[];
|
|
10896
|
+
input: ("image" | "text")[];
|
|
10222
10897
|
cost: {
|
|
10223
10898
|
input: number;
|
|
10224
10899
|
output: number;
|
|
@@ -10228,14 +10903,14 @@ export declare const MODELS: {
|
|
|
10228
10903
|
contextWindow: number;
|
|
10229
10904
|
maxTokens: number;
|
|
10230
10905
|
};
|
|
10231
|
-
readonly "qwen/qwen3-
|
|
10906
|
+
readonly "qwen/qwen3-vl-30b-a3b-instruct": {
|
|
10232
10907
|
id: string;
|
|
10233
10908
|
name: string;
|
|
10234
10909
|
api: "openai-completions";
|
|
10235
10910
|
provider: string;
|
|
10236
10911
|
baseUrl: string;
|
|
10237
|
-
reasoning:
|
|
10238
|
-
input: "text"[];
|
|
10912
|
+
reasoning: false;
|
|
10913
|
+
input: ("image" | "text")[];
|
|
10239
10914
|
cost: {
|
|
10240
10915
|
input: number;
|
|
10241
10916
|
output: number;
|
|
@@ -10245,14 +10920,14 @@ export declare const MODELS: {
|
|
|
10245
10920
|
contextWindow: number;
|
|
10246
10921
|
maxTokens: number;
|
|
10247
10922
|
};
|
|
10248
|
-
readonly "qwen/qwen3-
|
|
10923
|
+
readonly "qwen/qwen3-vl-30b-a3b-thinking": {
|
|
10249
10924
|
id: string;
|
|
10250
10925
|
name: string;
|
|
10251
10926
|
api: "openai-completions";
|
|
10252
10927
|
provider: string;
|
|
10253
10928
|
baseUrl: string;
|
|
10254
10929
|
reasoning: true;
|
|
10255
|
-
input: "text"[];
|
|
10930
|
+
input: ("image" | "text")[];
|
|
10256
10931
|
cost: {
|
|
10257
10932
|
input: number;
|
|
10258
10933
|
output: number;
|
|
@@ -10262,14 +10937,14 @@ export declare const MODELS: {
|
|
|
10262
10937
|
contextWindow: number;
|
|
10263
10938
|
maxTokens: number;
|
|
10264
10939
|
};
|
|
10265
|
-
readonly "qwen/qwen3-
|
|
10940
|
+
readonly "qwen/qwen3-vl-32b-instruct": {
|
|
10266
10941
|
id: string;
|
|
10267
10942
|
name: string;
|
|
10268
10943
|
api: "openai-completions";
|
|
10269
10944
|
provider: string;
|
|
10270
10945
|
baseUrl: string;
|
|
10271
10946
|
reasoning: false;
|
|
10272
|
-
input: "text"[];
|
|
10947
|
+
input: ("image" | "text")[];
|
|
10273
10948
|
cost: {
|
|
10274
10949
|
input: number;
|
|
10275
10950
|
output: number;
|
|
@@ -10279,14 +10954,14 @@ export declare const MODELS: {
|
|
|
10279
10954
|
contextWindow: number;
|
|
10280
10955
|
maxTokens: number;
|
|
10281
10956
|
};
|
|
10282
|
-
readonly "qwen/qwen3-
|
|
10957
|
+
readonly "qwen/qwen3-vl-8b-instruct": {
|
|
10283
10958
|
id: string;
|
|
10284
10959
|
name: string;
|
|
10285
10960
|
api: "openai-completions";
|
|
10286
10961
|
provider: string;
|
|
10287
10962
|
baseUrl: string;
|
|
10288
10963
|
reasoning: false;
|
|
10289
|
-
input: "text"[];
|
|
10964
|
+
input: ("image" | "text")[];
|
|
10290
10965
|
cost: {
|
|
10291
10966
|
input: number;
|
|
10292
10967
|
output: number;
|
|
@@ -10296,14 +10971,14 @@ export declare const MODELS: {
|
|
|
10296
10971
|
contextWindow: number;
|
|
10297
10972
|
maxTokens: number;
|
|
10298
10973
|
};
|
|
10299
|
-
readonly "qwen/qwen3-
|
|
10974
|
+
readonly "qwen/qwen3-vl-8b-thinking": {
|
|
10300
10975
|
id: string;
|
|
10301
10976
|
name: string;
|
|
10302
10977
|
api: "openai-completions";
|
|
10303
10978
|
provider: string;
|
|
10304
10979
|
baseUrl: string;
|
|
10305
|
-
reasoning:
|
|
10306
|
-
input: "text"[];
|
|
10980
|
+
reasoning: true;
|
|
10981
|
+
input: ("image" | "text")[];
|
|
10307
10982
|
cost: {
|
|
10308
10983
|
input: number;
|
|
10309
10984
|
output: number;
|
|
@@ -10313,14 +10988,14 @@ export declare const MODELS: {
|
|
|
10313
10988
|
contextWindow: number;
|
|
10314
10989
|
maxTokens: number;
|
|
10315
10990
|
};
|
|
10316
|
-
readonly "qwen/qwen3-
|
|
10991
|
+
readonly "qwen/qwen3.5-122b-a10b": {
|
|
10317
10992
|
id: string;
|
|
10318
10993
|
name: string;
|
|
10319
10994
|
api: "openai-completions";
|
|
10320
10995
|
provider: string;
|
|
10321
10996
|
baseUrl: string;
|
|
10322
|
-
reasoning:
|
|
10323
|
-
input: "text"[];
|
|
10997
|
+
reasoning: true;
|
|
10998
|
+
input: ("image" | "text")[];
|
|
10324
10999
|
cost: {
|
|
10325
11000
|
input: number;
|
|
10326
11001
|
output: number;
|
|
@@ -10330,14 +11005,14 @@ export declare const MODELS: {
|
|
|
10330
11005
|
contextWindow: number;
|
|
10331
11006
|
maxTokens: number;
|
|
10332
11007
|
};
|
|
10333
|
-
readonly "qwen/qwen3-
|
|
11008
|
+
readonly "qwen/qwen3.5-27b": {
|
|
10334
11009
|
id: string;
|
|
10335
11010
|
name: string;
|
|
10336
11011
|
api: "openai-completions";
|
|
10337
11012
|
provider: string;
|
|
10338
11013
|
baseUrl: string;
|
|
10339
|
-
reasoning:
|
|
10340
|
-
input: "text"[];
|
|
11014
|
+
reasoning: true;
|
|
11015
|
+
input: ("image" | "text")[];
|
|
10341
11016
|
cost: {
|
|
10342
11017
|
input: number;
|
|
10343
11018
|
output: number;
|
|
@@ -10347,14 +11022,14 @@ export declare const MODELS: {
|
|
|
10347
11022
|
contextWindow: number;
|
|
10348
11023
|
maxTokens: number;
|
|
10349
11024
|
};
|
|
10350
|
-
readonly "qwen/qwen3-
|
|
11025
|
+
readonly "qwen/qwen3.5-35b-a3b": {
|
|
10351
11026
|
id: string;
|
|
10352
11027
|
name: string;
|
|
10353
11028
|
api: "openai-completions";
|
|
10354
11029
|
provider: string;
|
|
10355
11030
|
baseUrl: string;
|
|
10356
|
-
reasoning:
|
|
10357
|
-
input: "text"[];
|
|
11031
|
+
reasoning: true;
|
|
11032
|
+
input: ("image" | "text")[];
|
|
10358
11033
|
cost: {
|
|
10359
11034
|
input: number;
|
|
10360
11035
|
output: number;
|
|
@@ -10364,14 +11039,14 @@ export declare const MODELS: {
|
|
|
10364
11039
|
contextWindow: number;
|
|
10365
11040
|
maxTokens: number;
|
|
10366
11041
|
};
|
|
10367
|
-
readonly "qwen/qwen3-
|
|
11042
|
+
readonly "qwen/qwen3.5-397b-a17b": {
|
|
10368
11043
|
id: string;
|
|
10369
11044
|
name: string;
|
|
10370
11045
|
api: "openai-completions";
|
|
10371
11046
|
provider: string;
|
|
10372
11047
|
baseUrl: string;
|
|
10373
|
-
reasoning:
|
|
10374
|
-
input: "text"[];
|
|
11048
|
+
reasoning: true;
|
|
11049
|
+
input: ("image" | "text")[];
|
|
10375
11050
|
cost: {
|
|
10376
11051
|
input: number;
|
|
10377
11052
|
output: number;
|
|
@@ -10381,14 +11056,14 @@ export declare const MODELS: {
|
|
|
10381
11056
|
contextWindow: number;
|
|
10382
11057
|
maxTokens: number;
|
|
10383
11058
|
};
|
|
10384
|
-
readonly "qwen/qwen3-
|
|
11059
|
+
readonly "qwen/qwen3.5-9b": {
|
|
10385
11060
|
id: string;
|
|
10386
11061
|
name: string;
|
|
10387
11062
|
api: "openai-completions";
|
|
10388
11063
|
provider: string;
|
|
10389
11064
|
baseUrl: string;
|
|
10390
11065
|
reasoning: true;
|
|
10391
|
-
input: "text"[];
|
|
11066
|
+
input: ("image" | "text")[];
|
|
10392
11067
|
cost: {
|
|
10393
11068
|
input: number;
|
|
10394
11069
|
output: number;
|
|
@@ -10398,14 +11073,14 @@ export declare const MODELS: {
|
|
|
10398
11073
|
contextWindow: number;
|
|
10399
11074
|
maxTokens: number;
|
|
10400
11075
|
};
|
|
10401
|
-
readonly "qwen/qwen3-
|
|
11076
|
+
readonly "qwen/qwen3.5-flash-02-23": {
|
|
10402
11077
|
id: string;
|
|
10403
11078
|
name: string;
|
|
10404
11079
|
api: "openai-completions";
|
|
10405
11080
|
provider: string;
|
|
10406
11081
|
baseUrl: string;
|
|
10407
|
-
reasoning:
|
|
10408
|
-
input: "text"[];
|
|
11082
|
+
reasoning: true;
|
|
11083
|
+
input: ("image" | "text")[];
|
|
10409
11084
|
cost: {
|
|
10410
11085
|
input: number;
|
|
10411
11086
|
output: number;
|
|
@@ -10415,14 +11090,14 @@ export declare const MODELS: {
|
|
|
10415
11090
|
contextWindow: number;
|
|
10416
11091
|
maxTokens: number;
|
|
10417
11092
|
};
|
|
10418
|
-
readonly "qwen/qwen3-
|
|
11093
|
+
readonly "qwen/qwen3.5-plus-02-15": {
|
|
10419
11094
|
id: string;
|
|
10420
11095
|
name: string;
|
|
10421
11096
|
api: "openai-completions";
|
|
10422
11097
|
provider: string;
|
|
10423
11098
|
baseUrl: string;
|
|
10424
|
-
reasoning:
|
|
10425
|
-
input: "text"[];
|
|
11099
|
+
reasoning: true;
|
|
11100
|
+
input: ("image" | "text")[];
|
|
10426
11101
|
cost: {
|
|
10427
11102
|
input: number;
|
|
10428
11103
|
output: number;
|
|
@@ -10432,14 +11107,14 @@ export declare const MODELS: {
|
|
|
10432
11107
|
contextWindow: number;
|
|
10433
11108
|
maxTokens: number;
|
|
10434
11109
|
};
|
|
10435
|
-
readonly "qwen/qwen3-
|
|
11110
|
+
readonly "qwen/qwen3.5-plus-20260420": {
|
|
10436
11111
|
id: string;
|
|
10437
11112
|
name: string;
|
|
10438
11113
|
api: "openai-completions";
|
|
10439
11114
|
provider: string;
|
|
10440
11115
|
baseUrl: string;
|
|
10441
11116
|
reasoning: true;
|
|
10442
|
-
input: "text"[];
|
|
11117
|
+
input: ("image" | "text")[];
|
|
10443
11118
|
cost: {
|
|
10444
11119
|
input: number;
|
|
10445
11120
|
output: number;
|
|
@@ -10449,13 +11124,13 @@ export declare const MODELS: {
|
|
|
10449
11124
|
contextWindow: number;
|
|
10450
11125
|
maxTokens: number;
|
|
10451
11126
|
};
|
|
10452
|
-
readonly "qwen/qwen3-
|
|
11127
|
+
readonly "qwen/qwen3.6-27b": {
|
|
10453
11128
|
id: string;
|
|
10454
11129
|
name: string;
|
|
10455
11130
|
api: "openai-completions";
|
|
10456
11131
|
provider: string;
|
|
10457
11132
|
baseUrl: string;
|
|
10458
|
-
reasoning:
|
|
11133
|
+
reasoning: true;
|
|
10459
11134
|
input: ("image" | "text")[];
|
|
10460
11135
|
cost: {
|
|
10461
11136
|
input: number;
|
|
@@ -10466,7 +11141,7 @@ export declare const MODELS: {
|
|
|
10466
11141
|
contextWindow: number;
|
|
10467
11142
|
maxTokens: number;
|
|
10468
11143
|
};
|
|
10469
|
-
readonly "qwen/qwen3-
|
|
11144
|
+
readonly "qwen/qwen3.6-flash": {
|
|
10470
11145
|
id: string;
|
|
10471
11146
|
name: string;
|
|
10472
11147
|
api: "openai-completions";
|
|
@@ -10483,14 +11158,14 @@ export declare const MODELS: {
|
|
|
10483
11158
|
contextWindow: number;
|
|
10484
11159
|
maxTokens: number;
|
|
10485
11160
|
};
|
|
10486
|
-
readonly "qwen/qwen3-
|
|
11161
|
+
readonly "qwen/qwen3.6-max-preview": {
|
|
10487
11162
|
id: string;
|
|
10488
11163
|
name: string;
|
|
10489
11164
|
api: "openai-completions";
|
|
10490
11165
|
provider: string;
|
|
10491
11166
|
baseUrl: string;
|
|
10492
|
-
reasoning:
|
|
10493
|
-
input:
|
|
11167
|
+
reasoning: true;
|
|
11168
|
+
input: "text"[];
|
|
10494
11169
|
cost: {
|
|
10495
11170
|
input: number;
|
|
10496
11171
|
output: number;
|
|
@@ -10500,7 +11175,7 @@ export declare const MODELS: {
|
|
|
10500
11175
|
contextWindow: number;
|
|
10501
11176
|
maxTokens: number;
|
|
10502
11177
|
};
|
|
10503
|
-
readonly "qwen/qwen3-
|
|
11178
|
+
readonly "qwen/qwen3.6-plus": {
|
|
10504
11179
|
id: string;
|
|
10505
11180
|
name: string;
|
|
10506
11181
|
api: "openai-completions";
|
|
@@ -10517,14 +11192,14 @@ export declare const MODELS: {
|
|
|
10517
11192
|
contextWindow: number;
|
|
10518
11193
|
maxTokens: number;
|
|
10519
11194
|
};
|
|
10520
|
-
readonly "qwen/
|
|
11195
|
+
readonly "qwen/qwq-32b": {
|
|
10521
11196
|
id: string;
|
|
10522
11197
|
name: string;
|
|
10523
11198
|
api: "openai-completions";
|
|
10524
11199
|
provider: string;
|
|
10525
11200
|
baseUrl: string;
|
|
10526
|
-
reasoning:
|
|
10527
|
-
input:
|
|
11201
|
+
reasoning: true;
|
|
11202
|
+
input: "text"[];
|
|
10528
11203
|
cost: {
|
|
10529
11204
|
input: number;
|
|
10530
11205
|
output: number;
|
|
@@ -10534,7 +11209,7 @@ export declare const MODELS: {
|
|
|
10534
11209
|
contextWindow: number;
|
|
10535
11210
|
maxTokens: number;
|
|
10536
11211
|
};
|
|
10537
|
-
readonly "
|
|
11212
|
+
readonly "rekaai/reka-edge": {
|
|
10538
11213
|
id: string;
|
|
10539
11214
|
name: string;
|
|
10540
11215
|
api: "openai-completions";
|
|
@@ -10551,14 +11226,14 @@ export declare const MODELS: {
|
|
|
10551
11226
|
contextWindow: number;
|
|
10552
11227
|
maxTokens: number;
|
|
10553
11228
|
};
|
|
10554
|
-
readonly "
|
|
11229
|
+
readonly "relace/relace-search": {
|
|
10555
11230
|
id: string;
|
|
10556
11231
|
name: string;
|
|
10557
11232
|
api: "openai-completions";
|
|
10558
11233
|
provider: string;
|
|
10559
11234
|
baseUrl: string;
|
|
10560
|
-
reasoning:
|
|
10561
|
-
input:
|
|
11235
|
+
reasoning: false;
|
|
11236
|
+
input: "text"[];
|
|
10562
11237
|
cost: {
|
|
10563
11238
|
input: number;
|
|
10564
11239
|
output: number;
|
|
@@ -10568,14 +11243,14 @@ export declare const MODELS: {
|
|
|
10568
11243
|
contextWindow: number;
|
|
10569
11244
|
maxTokens: number;
|
|
10570
11245
|
};
|
|
10571
|
-
readonly "
|
|
11246
|
+
readonly "sao10k/l3-euryale-70b": {
|
|
10572
11247
|
id: string;
|
|
10573
11248
|
name: string;
|
|
10574
11249
|
api: "openai-completions";
|
|
10575
11250
|
provider: string;
|
|
10576
11251
|
baseUrl: string;
|
|
10577
|
-
reasoning:
|
|
10578
|
-
input:
|
|
11252
|
+
reasoning: false;
|
|
11253
|
+
input: "text"[];
|
|
10579
11254
|
cost: {
|
|
10580
11255
|
input: number;
|
|
10581
11256
|
output: number;
|
|
@@ -10585,14 +11260,14 @@ export declare const MODELS: {
|
|
|
10585
11260
|
contextWindow: number;
|
|
10586
11261
|
maxTokens: number;
|
|
10587
11262
|
};
|
|
10588
|
-
readonly "
|
|
11263
|
+
readonly "sao10k/l3.1-euryale-70b": {
|
|
10589
11264
|
id: string;
|
|
10590
11265
|
name: string;
|
|
10591
11266
|
api: "openai-completions";
|
|
10592
11267
|
provider: string;
|
|
10593
11268
|
baseUrl: string;
|
|
10594
|
-
reasoning:
|
|
10595
|
-
input:
|
|
11269
|
+
reasoning: false;
|
|
11270
|
+
input: "text"[];
|
|
10596
11271
|
cost: {
|
|
10597
11272
|
input: number;
|
|
10598
11273
|
output: number;
|
|
@@ -10602,14 +11277,14 @@ export declare const MODELS: {
|
|
|
10602
11277
|
contextWindow: number;
|
|
10603
11278
|
maxTokens: number;
|
|
10604
11279
|
};
|
|
10605
|
-
readonly "
|
|
11280
|
+
readonly "stepfun/step-3.5-flash": {
|
|
10606
11281
|
id: string;
|
|
10607
11282
|
name: string;
|
|
10608
11283
|
api: "openai-completions";
|
|
10609
11284
|
provider: string;
|
|
10610
11285
|
baseUrl: string;
|
|
10611
11286
|
reasoning: true;
|
|
10612
|
-
input:
|
|
11287
|
+
input: "text"[];
|
|
10613
11288
|
cost: {
|
|
10614
11289
|
input: number;
|
|
10615
11290
|
output: number;
|
|
@@ -10619,14 +11294,14 @@ export declare const MODELS: {
|
|
|
10619
11294
|
contextWindow: number;
|
|
10620
11295
|
maxTokens: number;
|
|
10621
11296
|
};
|
|
10622
|
-
readonly "
|
|
11297
|
+
readonly "tencent/hy3-preview:free": {
|
|
10623
11298
|
id: string;
|
|
10624
11299
|
name: string;
|
|
10625
11300
|
api: "openai-completions";
|
|
10626
11301
|
provider: string;
|
|
10627
11302
|
baseUrl: string;
|
|
10628
11303
|
reasoning: true;
|
|
10629
|
-
input:
|
|
11304
|
+
input: "text"[];
|
|
10630
11305
|
cost: {
|
|
10631
11306
|
input: number;
|
|
10632
11307
|
output: number;
|
|
@@ -10636,14 +11311,14 @@ export declare const MODELS: {
|
|
|
10636
11311
|
contextWindow: number;
|
|
10637
11312
|
maxTokens: number;
|
|
10638
11313
|
};
|
|
10639
|
-
readonly "
|
|
11314
|
+
readonly "thedrummer/rocinante-12b": {
|
|
10640
11315
|
id: string;
|
|
10641
11316
|
name: string;
|
|
10642
11317
|
api: "openai-completions";
|
|
10643
11318
|
provider: string;
|
|
10644
11319
|
baseUrl: string;
|
|
10645
|
-
reasoning:
|
|
10646
|
-
input:
|
|
11320
|
+
reasoning: false;
|
|
11321
|
+
input: "text"[];
|
|
10647
11322
|
cost: {
|
|
10648
11323
|
input: number;
|
|
10649
11324
|
output: number;
|
|
@@ -10653,14 +11328,14 @@ export declare const MODELS: {
|
|
|
10653
11328
|
contextWindow: number;
|
|
10654
11329
|
maxTokens: number;
|
|
10655
11330
|
};
|
|
10656
|
-
readonly "
|
|
11331
|
+
readonly "thedrummer/unslopnemo-12b": {
|
|
10657
11332
|
id: string;
|
|
10658
11333
|
name: string;
|
|
10659
11334
|
api: "openai-completions";
|
|
10660
11335
|
provider: string;
|
|
10661
11336
|
baseUrl: string;
|
|
10662
|
-
reasoning:
|
|
10663
|
-
input:
|
|
11337
|
+
reasoning: false;
|
|
11338
|
+
input: "text"[];
|
|
10664
11339
|
cost: {
|
|
10665
11340
|
input: number;
|
|
10666
11341
|
output: number;
|
|
@@ -10670,14 +11345,14 @@ export declare const MODELS: {
|
|
|
10670
11345
|
contextWindow: number;
|
|
10671
11346
|
maxTokens: number;
|
|
10672
11347
|
};
|
|
10673
|
-
readonly "
|
|
11348
|
+
readonly "tngtech/deepseek-r1t2-chimera": {
|
|
10674
11349
|
id: string;
|
|
10675
11350
|
name: string;
|
|
10676
11351
|
api: "openai-completions";
|
|
10677
11352
|
provider: string;
|
|
10678
11353
|
baseUrl: string;
|
|
10679
11354
|
reasoning: true;
|
|
10680
|
-
input:
|
|
11355
|
+
input: "text"[];
|
|
10681
11356
|
cost: {
|
|
10682
11357
|
input: number;
|
|
10683
11358
|
output: number;
|
|
@@ -10687,14 +11362,14 @@ export declare const MODELS: {
|
|
|
10687
11362
|
contextWindow: number;
|
|
10688
11363
|
maxTokens: number;
|
|
10689
11364
|
};
|
|
10690
|
-
readonly "
|
|
11365
|
+
readonly "upstage/solar-pro-3": {
|
|
10691
11366
|
id: string;
|
|
10692
11367
|
name: string;
|
|
10693
11368
|
api: "openai-completions";
|
|
10694
11369
|
provider: string;
|
|
10695
11370
|
baseUrl: string;
|
|
10696
11371
|
reasoning: true;
|
|
10697
|
-
input:
|
|
11372
|
+
input: "text"[];
|
|
10698
11373
|
cost: {
|
|
10699
11374
|
input: number;
|
|
10700
11375
|
output: number;
|
|
@@ -10704,13 +11379,13 @@ export declare const MODELS: {
|
|
|
10704
11379
|
contextWindow: number;
|
|
10705
11380
|
maxTokens: number;
|
|
10706
11381
|
};
|
|
10707
|
-
readonly "
|
|
11382
|
+
readonly "x-ai/grok-3": {
|
|
10708
11383
|
id: string;
|
|
10709
11384
|
name: string;
|
|
10710
11385
|
api: "openai-completions";
|
|
10711
11386
|
provider: string;
|
|
10712
11387
|
baseUrl: string;
|
|
10713
|
-
reasoning:
|
|
11388
|
+
reasoning: false;
|
|
10714
11389
|
input: "text"[];
|
|
10715
11390
|
cost: {
|
|
10716
11391
|
input: number;
|
|
@@ -10721,14 +11396,14 @@ export declare const MODELS: {
|
|
|
10721
11396
|
contextWindow: number;
|
|
10722
11397
|
maxTokens: number;
|
|
10723
11398
|
};
|
|
10724
|
-
readonly "
|
|
11399
|
+
readonly "x-ai/grok-3-beta": {
|
|
10725
11400
|
id: string;
|
|
10726
11401
|
name: string;
|
|
10727
11402
|
api: "openai-completions";
|
|
10728
11403
|
provider: string;
|
|
10729
11404
|
baseUrl: string;
|
|
10730
11405
|
reasoning: false;
|
|
10731
|
-
input:
|
|
11406
|
+
input: "text"[];
|
|
10732
11407
|
cost: {
|
|
10733
11408
|
input: number;
|
|
10734
11409
|
output: number;
|
|
@@ -10738,13 +11413,13 @@ export declare const MODELS: {
|
|
|
10738
11413
|
contextWindow: number;
|
|
10739
11414
|
maxTokens: number;
|
|
10740
11415
|
};
|
|
10741
|
-
readonly "
|
|
11416
|
+
readonly "x-ai/grok-3-mini": {
|
|
10742
11417
|
id: string;
|
|
10743
11418
|
name: string;
|
|
10744
11419
|
api: "openai-completions";
|
|
10745
11420
|
provider: string;
|
|
10746
11421
|
baseUrl: string;
|
|
10747
|
-
reasoning:
|
|
11422
|
+
reasoning: true;
|
|
10748
11423
|
input: "text"[];
|
|
10749
11424
|
cost: {
|
|
10750
11425
|
input: number;
|
|
@@ -10755,13 +11430,13 @@ export declare const MODELS: {
|
|
|
10755
11430
|
contextWindow: number;
|
|
10756
11431
|
maxTokens: number;
|
|
10757
11432
|
};
|
|
10758
|
-
readonly "
|
|
11433
|
+
readonly "x-ai/grok-3-mini-beta": {
|
|
10759
11434
|
id: string;
|
|
10760
11435
|
name: string;
|
|
10761
11436
|
api: "openai-completions";
|
|
10762
11437
|
provider: string;
|
|
10763
11438
|
baseUrl: string;
|
|
10764
|
-
reasoning:
|
|
11439
|
+
reasoning: true;
|
|
10765
11440
|
input: "text"[];
|
|
10766
11441
|
cost: {
|
|
10767
11442
|
input: number;
|
|
@@ -10772,14 +11447,14 @@ export declare const MODELS: {
|
|
|
10772
11447
|
contextWindow: number;
|
|
10773
11448
|
maxTokens: number;
|
|
10774
11449
|
};
|
|
10775
|
-
readonly "
|
|
11450
|
+
readonly "x-ai/grok-4": {
|
|
10776
11451
|
id: string;
|
|
10777
11452
|
name: string;
|
|
10778
11453
|
api: "openai-completions";
|
|
10779
11454
|
provider: string;
|
|
10780
11455
|
baseUrl: string;
|
|
10781
|
-
reasoning:
|
|
10782
|
-
input: "text"[];
|
|
11456
|
+
reasoning: true;
|
|
11457
|
+
input: ("image" | "text")[];
|
|
10783
11458
|
cost: {
|
|
10784
11459
|
input: number;
|
|
10785
11460
|
output: number;
|
|
@@ -10789,14 +11464,14 @@ export declare const MODELS: {
|
|
|
10789
11464
|
contextWindow: number;
|
|
10790
11465
|
maxTokens: number;
|
|
10791
11466
|
};
|
|
10792
|
-
readonly "
|
|
11467
|
+
readonly "x-ai/grok-4-fast": {
|
|
10793
11468
|
id: string;
|
|
10794
11469
|
name: string;
|
|
10795
11470
|
api: "openai-completions";
|
|
10796
11471
|
provider: string;
|
|
10797
11472
|
baseUrl: string;
|
|
10798
11473
|
reasoning: true;
|
|
10799
|
-
input: "text"[];
|
|
11474
|
+
input: ("image" | "text")[];
|
|
10800
11475
|
cost: {
|
|
10801
11476
|
input: number;
|
|
10802
11477
|
output: number;
|
|
@@ -10806,14 +11481,14 @@ export declare const MODELS: {
|
|
|
10806
11481
|
contextWindow: number;
|
|
10807
11482
|
maxTokens: number;
|
|
10808
11483
|
};
|
|
10809
|
-
readonly "
|
|
11484
|
+
readonly "x-ai/grok-4.1-fast": {
|
|
10810
11485
|
id: string;
|
|
10811
11486
|
name: string;
|
|
10812
11487
|
api: "openai-completions";
|
|
10813
11488
|
provider: string;
|
|
10814
11489
|
baseUrl: string;
|
|
10815
|
-
reasoning:
|
|
10816
|
-
input: "text"[];
|
|
11490
|
+
reasoning: true;
|
|
11491
|
+
input: ("image" | "text")[];
|
|
10817
11492
|
cost: {
|
|
10818
11493
|
input: number;
|
|
10819
11494
|
output: number;
|
|
@@ -10823,14 +11498,14 @@ export declare const MODELS: {
|
|
|
10823
11498
|
contextWindow: number;
|
|
10824
11499
|
maxTokens: number;
|
|
10825
11500
|
};
|
|
10826
|
-
readonly "
|
|
11501
|
+
readonly "x-ai/grok-4.20": {
|
|
10827
11502
|
id: string;
|
|
10828
11503
|
name: string;
|
|
10829
11504
|
api: "openai-completions";
|
|
10830
11505
|
provider: string;
|
|
10831
11506
|
baseUrl: string;
|
|
10832
|
-
reasoning:
|
|
10833
|
-
input: "text"[];
|
|
11507
|
+
reasoning: true;
|
|
11508
|
+
input: ("image" | "text")[];
|
|
10834
11509
|
cost: {
|
|
10835
11510
|
input: number;
|
|
10836
11511
|
output: number;
|
|
@@ -10840,7 +11515,7 @@ export declare const MODELS: {
|
|
|
10840
11515
|
contextWindow: number;
|
|
10841
11516
|
maxTokens: number;
|
|
10842
11517
|
};
|
|
10843
|
-
readonly "
|
|
11518
|
+
readonly "x-ai/grok-code-fast-1": {
|
|
10844
11519
|
id: string;
|
|
10845
11520
|
name: string;
|
|
10846
11521
|
api: "openai-completions";
|
|
@@ -10857,7 +11532,7 @@ export declare const MODELS: {
|
|
|
10857
11532
|
contextWindow: number;
|
|
10858
11533
|
maxTokens: number;
|
|
10859
11534
|
};
|
|
10860
|
-
readonly "
|
|
11535
|
+
readonly "xiaomi/mimo-v2-flash": {
|
|
10861
11536
|
id: string;
|
|
10862
11537
|
name: string;
|
|
10863
11538
|
api: "openai-completions";
|
|
@@ -10874,14 +11549,14 @@ export declare const MODELS: {
|
|
|
10874
11549
|
contextWindow: number;
|
|
10875
11550
|
maxTokens: number;
|
|
10876
11551
|
};
|
|
10877
|
-
readonly "
|
|
11552
|
+
readonly "xiaomi/mimo-v2-omni": {
|
|
10878
11553
|
id: string;
|
|
10879
11554
|
name: string;
|
|
10880
11555
|
api: "openai-completions";
|
|
10881
11556
|
provider: string;
|
|
10882
11557
|
baseUrl: string;
|
|
10883
|
-
reasoning:
|
|
10884
|
-
input: "text"[];
|
|
11558
|
+
reasoning: true;
|
|
11559
|
+
input: ("image" | "text")[];
|
|
10885
11560
|
cost: {
|
|
10886
11561
|
input: number;
|
|
10887
11562
|
output: number;
|
|
@@ -10891,13 +11566,13 @@ export declare const MODELS: {
|
|
|
10891
11566
|
contextWindow: number;
|
|
10892
11567
|
maxTokens: number;
|
|
10893
11568
|
};
|
|
10894
|
-
readonly "
|
|
11569
|
+
readonly "xiaomi/mimo-v2-pro": {
|
|
10895
11570
|
id: string;
|
|
10896
11571
|
name: string;
|
|
10897
11572
|
api: "openai-completions";
|
|
10898
11573
|
provider: string;
|
|
10899
11574
|
baseUrl: string;
|
|
10900
|
-
reasoning:
|
|
11575
|
+
reasoning: true;
|
|
10901
11576
|
input: "text"[];
|
|
10902
11577
|
cost: {
|
|
10903
11578
|
input: number;
|
|
@@ -10908,14 +11583,14 @@ export declare const MODELS: {
|
|
|
10908
11583
|
contextWindow: number;
|
|
10909
11584
|
maxTokens: number;
|
|
10910
11585
|
};
|
|
10911
|
-
readonly "
|
|
11586
|
+
readonly "xiaomi/mimo-v2.5": {
|
|
10912
11587
|
id: string;
|
|
10913
11588
|
name: string;
|
|
10914
11589
|
api: "openai-completions";
|
|
10915
11590
|
provider: string;
|
|
10916
11591
|
baseUrl: string;
|
|
10917
11592
|
reasoning: true;
|
|
10918
|
-
input: "text"[];
|
|
11593
|
+
input: ("image" | "text")[];
|
|
10919
11594
|
cost: {
|
|
10920
11595
|
input: number;
|
|
10921
11596
|
output: number;
|
|
@@ -10925,7 +11600,7 @@ export declare const MODELS: {
|
|
|
10925
11600
|
contextWindow: number;
|
|
10926
11601
|
maxTokens: number;
|
|
10927
11602
|
};
|
|
10928
|
-
readonly "
|
|
11603
|
+
readonly "xiaomi/mimo-v2.5-pro": {
|
|
10929
11604
|
id: string;
|
|
10930
11605
|
name: string;
|
|
10931
11606
|
api: "openai-completions";
|
|
@@ -10942,14 +11617,14 @@ export declare const MODELS: {
|
|
|
10942
11617
|
contextWindow: number;
|
|
10943
11618
|
maxTokens: number;
|
|
10944
11619
|
};
|
|
10945
|
-
readonly "
|
|
11620
|
+
readonly "z-ai/glm-4-32b": {
|
|
10946
11621
|
id: string;
|
|
10947
11622
|
name: string;
|
|
10948
11623
|
api: "openai-completions";
|
|
10949
11624
|
provider: string;
|
|
10950
11625
|
baseUrl: string;
|
|
10951
|
-
reasoning:
|
|
10952
|
-
input:
|
|
11626
|
+
reasoning: false;
|
|
11627
|
+
input: "text"[];
|
|
10953
11628
|
cost: {
|
|
10954
11629
|
input: number;
|
|
10955
11630
|
output: number;
|
|
@@ -10959,14 +11634,14 @@ export declare const MODELS: {
|
|
|
10959
11634
|
contextWindow: number;
|
|
10960
11635
|
maxTokens: number;
|
|
10961
11636
|
};
|
|
10962
|
-
readonly "
|
|
11637
|
+
readonly "z-ai/glm-4.5": {
|
|
10963
11638
|
id: string;
|
|
10964
11639
|
name: string;
|
|
10965
11640
|
api: "openai-completions";
|
|
10966
11641
|
provider: string;
|
|
10967
11642
|
baseUrl: string;
|
|
10968
11643
|
reasoning: true;
|
|
10969
|
-
input:
|
|
11644
|
+
input: "text"[];
|
|
10970
11645
|
cost: {
|
|
10971
11646
|
input: number;
|
|
10972
11647
|
output: number;
|
|
@@ -10976,14 +11651,14 @@ export declare const MODELS: {
|
|
|
10976
11651
|
contextWindow: number;
|
|
10977
11652
|
maxTokens: number;
|
|
10978
11653
|
};
|
|
10979
|
-
readonly "
|
|
11654
|
+
readonly "z-ai/glm-4.5-air": {
|
|
10980
11655
|
id: string;
|
|
10981
11656
|
name: string;
|
|
10982
11657
|
api: "openai-completions";
|
|
10983
11658
|
provider: string;
|
|
10984
11659
|
baseUrl: string;
|
|
10985
11660
|
reasoning: true;
|
|
10986
|
-
input:
|
|
11661
|
+
input: "text"[];
|
|
10987
11662
|
cost: {
|
|
10988
11663
|
input: number;
|
|
10989
11664
|
output: number;
|
|
@@ -10993,14 +11668,14 @@ export declare const MODELS: {
|
|
|
10993
11668
|
contextWindow: number;
|
|
10994
11669
|
maxTokens: number;
|
|
10995
11670
|
};
|
|
10996
|
-
readonly "
|
|
11671
|
+
readonly "z-ai/glm-4.5-air:free": {
|
|
10997
11672
|
id: string;
|
|
10998
11673
|
name: string;
|
|
10999
11674
|
api: "openai-completions";
|
|
11000
11675
|
provider: string;
|
|
11001
11676
|
baseUrl: string;
|
|
11002
11677
|
reasoning: true;
|
|
11003
|
-
input:
|
|
11678
|
+
input: "text"[];
|
|
11004
11679
|
cost: {
|
|
11005
11680
|
input: number;
|
|
11006
11681
|
output: number;
|
|
@@ -11010,14 +11685,14 @@ export declare const MODELS: {
|
|
|
11010
11685
|
contextWindow: number;
|
|
11011
11686
|
maxTokens: number;
|
|
11012
11687
|
};
|
|
11013
|
-
readonly "
|
|
11688
|
+
readonly "z-ai/glm-4.5v": {
|
|
11014
11689
|
id: string;
|
|
11015
11690
|
name: string;
|
|
11016
11691
|
api: "openai-completions";
|
|
11017
11692
|
provider: string;
|
|
11018
11693
|
baseUrl: string;
|
|
11019
11694
|
reasoning: true;
|
|
11020
|
-
input: "text"[];
|
|
11695
|
+
input: ("image" | "text")[];
|
|
11021
11696
|
cost: {
|
|
11022
11697
|
input: number;
|
|
11023
11698
|
output: number;
|
|
@@ -11027,7 +11702,7 @@ export declare const MODELS: {
|
|
|
11027
11702
|
contextWindow: number;
|
|
11028
11703
|
maxTokens: number;
|
|
11029
11704
|
};
|
|
11030
|
-
readonly "
|
|
11705
|
+
readonly "z-ai/glm-4.6": {
|
|
11031
11706
|
id: string;
|
|
11032
11707
|
name: string;
|
|
11033
11708
|
api: "openai-completions";
|
|
@@ -11044,7 +11719,7 @@ export declare const MODELS: {
|
|
|
11044
11719
|
contextWindow: number;
|
|
11045
11720
|
maxTokens: number;
|
|
11046
11721
|
};
|
|
11047
|
-
readonly "
|
|
11722
|
+
readonly "z-ai/glm-4.6v": {
|
|
11048
11723
|
id: string;
|
|
11049
11724
|
name: string;
|
|
11050
11725
|
api: "openai-completions";
|
|
@@ -11061,7 +11736,7 @@ export declare const MODELS: {
|
|
|
11061
11736
|
contextWindow: number;
|
|
11062
11737
|
maxTokens: number;
|
|
11063
11738
|
};
|
|
11064
|
-
readonly "
|
|
11739
|
+
readonly "z-ai/glm-4.7": {
|
|
11065
11740
|
id: string;
|
|
11066
11741
|
name: string;
|
|
11067
11742
|
api: "openai-completions";
|
|
@@ -11078,13 +11753,13 @@ export declare const MODELS: {
|
|
|
11078
11753
|
contextWindow: number;
|
|
11079
11754
|
maxTokens: number;
|
|
11080
11755
|
};
|
|
11081
|
-
readonly "z-ai/glm-4-
|
|
11756
|
+
readonly "z-ai/glm-4.7-flash": {
|
|
11082
11757
|
id: string;
|
|
11083
11758
|
name: string;
|
|
11084
11759
|
api: "openai-completions";
|
|
11085
11760
|
provider: string;
|
|
11086
11761
|
baseUrl: string;
|
|
11087
|
-
reasoning:
|
|
11762
|
+
reasoning: true;
|
|
11088
11763
|
input: "text"[];
|
|
11089
11764
|
cost: {
|
|
11090
11765
|
input: number;
|
|
@@ -11095,7 +11770,7 @@ export declare const MODELS: {
|
|
|
11095
11770
|
contextWindow: number;
|
|
11096
11771
|
maxTokens: number;
|
|
11097
11772
|
};
|
|
11098
|
-
readonly "z-ai/glm-
|
|
11773
|
+
readonly "z-ai/glm-5": {
|
|
11099
11774
|
id: string;
|
|
11100
11775
|
name: string;
|
|
11101
11776
|
api: "openai-completions";
|
|
@@ -11112,7 +11787,7 @@ export declare const MODELS: {
|
|
|
11112
11787
|
contextWindow: number;
|
|
11113
11788
|
maxTokens: number;
|
|
11114
11789
|
};
|
|
11115
|
-
readonly "z-ai/glm-
|
|
11790
|
+
readonly "z-ai/glm-5-turbo": {
|
|
11116
11791
|
id: string;
|
|
11117
11792
|
name: string;
|
|
11118
11793
|
api: "openai-completions";
|
|
@@ -11129,7 +11804,7 @@ export declare const MODELS: {
|
|
|
11129
11804
|
contextWindow: number;
|
|
11130
11805
|
maxTokens: number;
|
|
11131
11806
|
};
|
|
11132
|
-
readonly "z-ai/glm-
|
|
11807
|
+
readonly "z-ai/glm-5.1": {
|
|
11133
11808
|
id: string;
|
|
11134
11809
|
name: string;
|
|
11135
11810
|
api: "openai-completions";
|
|
@@ -11146,7 +11821,7 @@ export declare const MODELS: {
|
|
|
11146
11821
|
contextWindow: number;
|
|
11147
11822
|
maxTokens: number;
|
|
11148
11823
|
};
|
|
11149
|
-
readonly "z-ai/glm-
|
|
11824
|
+
readonly "z-ai/glm-5v-turbo": {
|
|
11150
11825
|
id: string;
|
|
11151
11826
|
name: string;
|
|
11152
11827
|
api: "openai-completions";
|
|
@@ -11163,14 +11838,14 @@ export declare const MODELS: {
|
|
|
11163
11838
|
contextWindow: number;
|
|
11164
11839
|
maxTokens: number;
|
|
11165
11840
|
};
|
|
11166
|
-
readonly "
|
|
11841
|
+
readonly "~anthropic/claude-haiku-latest": {
|
|
11167
11842
|
id: string;
|
|
11168
11843
|
name: string;
|
|
11169
11844
|
api: "openai-completions";
|
|
11170
11845
|
provider: string;
|
|
11171
11846
|
baseUrl: string;
|
|
11172
11847
|
reasoning: true;
|
|
11173
|
-
input: "text"[];
|
|
11848
|
+
input: ("image" | "text")[];
|
|
11174
11849
|
cost: {
|
|
11175
11850
|
input: number;
|
|
11176
11851
|
output: number;
|
|
@@ -11180,7 +11855,7 @@ export declare const MODELS: {
|
|
|
11180
11855
|
contextWindow: number;
|
|
11181
11856
|
maxTokens: number;
|
|
11182
11857
|
};
|
|
11183
|
-
readonly "
|
|
11858
|
+
readonly "~anthropic/claude-opus-latest": {
|
|
11184
11859
|
id: string;
|
|
11185
11860
|
name: string;
|
|
11186
11861
|
api: "openai-completions";
|
|
@@ -11197,14 +11872,14 @@ export declare const MODELS: {
|
|
|
11197
11872
|
contextWindow: number;
|
|
11198
11873
|
maxTokens: number;
|
|
11199
11874
|
};
|
|
11200
|
-
readonly "
|
|
11875
|
+
readonly "~anthropic/claude-sonnet-latest": {
|
|
11201
11876
|
id: string;
|
|
11202
11877
|
name: string;
|
|
11203
11878
|
api: "openai-completions";
|
|
11204
11879
|
provider: string;
|
|
11205
11880
|
baseUrl: string;
|
|
11206
11881
|
reasoning: true;
|
|
11207
|
-
input: "text"[];
|
|
11882
|
+
input: ("image" | "text")[];
|
|
11208
11883
|
cost: {
|
|
11209
11884
|
input: number;
|
|
11210
11885
|
output: number;
|
|
@@ -11214,14 +11889,14 @@ export declare const MODELS: {
|
|
|
11214
11889
|
contextWindow: number;
|
|
11215
11890
|
maxTokens: number;
|
|
11216
11891
|
};
|
|
11217
|
-
readonly "
|
|
11892
|
+
readonly "~google/gemini-flash-latest": {
|
|
11218
11893
|
id: string;
|
|
11219
11894
|
name: string;
|
|
11220
11895
|
api: "openai-completions";
|
|
11221
11896
|
provider: string;
|
|
11222
11897
|
baseUrl: string;
|
|
11223
11898
|
reasoning: true;
|
|
11224
|
-
input: "text"[];
|
|
11899
|
+
input: ("image" | "text")[];
|
|
11225
11900
|
cost: {
|
|
11226
11901
|
input: number;
|
|
11227
11902
|
output: number;
|
|
@@ -11231,14 +11906,14 @@ export declare const MODELS: {
|
|
|
11231
11906
|
contextWindow: number;
|
|
11232
11907
|
maxTokens: number;
|
|
11233
11908
|
};
|
|
11234
|
-
readonly "
|
|
11909
|
+
readonly "~google/gemini-pro-latest": {
|
|
11235
11910
|
id: string;
|
|
11236
11911
|
name: string;
|
|
11237
11912
|
api: "openai-completions";
|
|
11238
11913
|
provider: string;
|
|
11239
11914
|
baseUrl: string;
|
|
11240
11915
|
reasoning: true;
|
|
11241
|
-
input: "text"[];
|
|
11916
|
+
input: ("image" | "text")[];
|
|
11242
11917
|
cost: {
|
|
11243
11918
|
input: number;
|
|
11244
11919
|
output: number;
|
|
@@ -11248,14 +11923,14 @@ export declare const MODELS: {
|
|
|
11248
11923
|
contextWindow: number;
|
|
11249
11924
|
maxTokens: number;
|
|
11250
11925
|
};
|
|
11251
|
-
readonly "
|
|
11926
|
+
readonly "~moonshotai/kimi-latest": {
|
|
11252
11927
|
id: string;
|
|
11253
11928
|
name: string;
|
|
11254
11929
|
api: "openai-completions";
|
|
11255
11930
|
provider: string;
|
|
11256
11931
|
baseUrl: string;
|
|
11257
11932
|
reasoning: true;
|
|
11258
|
-
input: "text"[];
|
|
11933
|
+
input: ("image" | "text")[];
|
|
11259
11934
|
cost: {
|
|
11260
11935
|
input: number;
|
|
11261
11936
|
output: number;
|
|
@@ -11265,14 +11940,14 @@ export declare const MODELS: {
|
|
|
11265
11940
|
contextWindow: number;
|
|
11266
11941
|
maxTokens: number;
|
|
11267
11942
|
};
|
|
11268
|
-
readonly "
|
|
11943
|
+
readonly "~openai/gpt-latest": {
|
|
11269
11944
|
id: string;
|
|
11270
11945
|
name: string;
|
|
11271
11946
|
api: "openai-completions";
|
|
11272
11947
|
provider: string;
|
|
11273
11948
|
baseUrl: string;
|
|
11274
11949
|
reasoning: true;
|
|
11275
|
-
input: "text"[];
|
|
11950
|
+
input: ("image" | "text")[];
|
|
11276
11951
|
cost: {
|
|
11277
11952
|
input: number;
|
|
11278
11953
|
output: number;
|
|
@@ -11282,7 +11957,7 @@ export declare const MODELS: {
|
|
|
11282
11957
|
contextWindow: number;
|
|
11283
11958
|
maxTokens: number;
|
|
11284
11959
|
};
|
|
11285
|
-
readonly "
|
|
11960
|
+
readonly "~openai/gpt-mini-latest": {
|
|
11286
11961
|
id: string;
|
|
11287
11962
|
name: string;
|
|
11288
11963
|
api: "openai-completions";
|
|
@@ -11369,6 +12044,23 @@ export declare const MODELS: {
|
|
|
11369
12044
|
contextWindow: number;
|
|
11370
12045
|
maxTokens: number;
|
|
11371
12046
|
};
|
|
12047
|
+
readonly "alibaba/qwen-3.6-max-preview": {
|
|
12048
|
+
id: string;
|
|
12049
|
+
name: string;
|
|
12050
|
+
api: "anthropic-messages";
|
|
12051
|
+
provider: string;
|
|
12052
|
+
baseUrl: string;
|
|
12053
|
+
reasoning: true;
|
|
12054
|
+
input: ("image" | "text")[];
|
|
12055
|
+
cost: {
|
|
12056
|
+
input: number;
|
|
12057
|
+
output: number;
|
|
12058
|
+
cacheRead: number;
|
|
12059
|
+
cacheWrite: number;
|
|
12060
|
+
};
|
|
12061
|
+
contextWindow: number;
|
|
12062
|
+
maxTokens: number;
|
|
12063
|
+
};
|
|
11372
12064
|
readonly "alibaba/qwen3-235b-a22b-thinking": {
|
|
11373
12065
|
id: string;
|
|
11374
12066
|
name: string;
|
|
@@ -11828,7 +12520,41 @@ export declare const MODELS: {
|
|
|
11828
12520
|
contextWindow: number;
|
|
11829
12521
|
maxTokens: number;
|
|
11830
12522
|
};
|
|
11831
|
-
readonly "cohere/command-a": {
|
|
12523
|
+
readonly "cohere/command-a": {
|
|
12524
|
+
id: string;
|
|
12525
|
+
name: string;
|
|
12526
|
+
api: "anthropic-messages";
|
|
12527
|
+
provider: string;
|
|
12528
|
+
baseUrl: string;
|
|
12529
|
+
reasoning: false;
|
|
12530
|
+
input: "text"[];
|
|
12531
|
+
cost: {
|
|
12532
|
+
input: number;
|
|
12533
|
+
output: number;
|
|
12534
|
+
cacheRead: number;
|
|
12535
|
+
cacheWrite: number;
|
|
12536
|
+
};
|
|
12537
|
+
contextWindow: number;
|
|
12538
|
+
maxTokens: number;
|
|
12539
|
+
};
|
|
12540
|
+
readonly "deepseek/deepseek-r1": {
|
|
12541
|
+
id: string;
|
|
12542
|
+
name: string;
|
|
12543
|
+
api: "anthropic-messages";
|
|
12544
|
+
provider: string;
|
|
12545
|
+
baseUrl: string;
|
|
12546
|
+
reasoning: true;
|
|
12547
|
+
input: "text"[];
|
|
12548
|
+
cost: {
|
|
12549
|
+
input: number;
|
|
12550
|
+
output: number;
|
|
12551
|
+
cacheRead: number;
|
|
12552
|
+
cacheWrite: number;
|
|
12553
|
+
};
|
|
12554
|
+
contextWindow: number;
|
|
12555
|
+
maxTokens: number;
|
|
12556
|
+
};
|
|
12557
|
+
readonly "deepseek/deepseek-v3": {
|
|
11832
12558
|
id: string;
|
|
11833
12559
|
name: string;
|
|
11834
12560
|
api: "anthropic-messages";
|
|
@@ -11845,7 +12571,7 @@ export declare const MODELS: {
|
|
|
11845
12571
|
contextWindow: number;
|
|
11846
12572
|
maxTokens: number;
|
|
11847
12573
|
};
|
|
11848
|
-
readonly "deepseek/deepseek-
|
|
12574
|
+
readonly "deepseek/deepseek-v3.1": {
|
|
11849
12575
|
id: string;
|
|
11850
12576
|
name: string;
|
|
11851
12577
|
api: "anthropic-messages";
|
|
@@ -11862,13 +12588,13 @@ export declare const MODELS: {
|
|
|
11862
12588
|
contextWindow: number;
|
|
11863
12589
|
maxTokens: number;
|
|
11864
12590
|
};
|
|
11865
|
-
readonly "deepseek/deepseek-v3": {
|
|
12591
|
+
readonly "deepseek/deepseek-v3.1-terminus": {
|
|
11866
12592
|
id: string;
|
|
11867
12593
|
name: string;
|
|
11868
12594
|
api: "anthropic-messages";
|
|
11869
12595
|
provider: string;
|
|
11870
12596
|
baseUrl: string;
|
|
11871
|
-
reasoning:
|
|
12597
|
+
reasoning: true;
|
|
11872
12598
|
input: "text"[];
|
|
11873
12599
|
cost: {
|
|
11874
12600
|
input: number;
|
|
@@ -11879,13 +12605,13 @@ export declare const MODELS: {
|
|
|
11879
12605
|
contextWindow: number;
|
|
11880
12606
|
maxTokens: number;
|
|
11881
12607
|
};
|
|
11882
|
-
readonly "deepseek/deepseek-v3.
|
|
12608
|
+
readonly "deepseek/deepseek-v3.2": {
|
|
11883
12609
|
id: string;
|
|
11884
12610
|
name: string;
|
|
11885
12611
|
api: "anthropic-messages";
|
|
11886
12612
|
provider: string;
|
|
11887
12613
|
baseUrl: string;
|
|
11888
|
-
reasoning:
|
|
12614
|
+
reasoning: false;
|
|
11889
12615
|
input: "text"[];
|
|
11890
12616
|
cost: {
|
|
11891
12617
|
input: number;
|
|
@@ -11896,7 +12622,7 @@ export declare const MODELS: {
|
|
|
11896
12622
|
contextWindow: number;
|
|
11897
12623
|
maxTokens: number;
|
|
11898
12624
|
};
|
|
11899
|
-
readonly "deepseek/deepseek-v3.
|
|
12625
|
+
readonly "deepseek/deepseek-v3.2-thinking": {
|
|
11900
12626
|
id: string;
|
|
11901
12627
|
name: string;
|
|
11902
12628
|
api: "anthropic-messages";
|
|
@@ -11913,13 +12639,13 @@ export declare const MODELS: {
|
|
|
11913
12639
|
contextWindow: number;
|
|
11914
12640
|
maxTokens: number;
|
|
11915
12641
|
};
|
|
11916
|
-
readonly "deepseek/deepseek-
|
|
12642
|
+
readonly "deepseek/deepseek-v4-flash": {
|
|
11917
12643
|
id: string;
|
|
11918
12644
|
name: string;
|
|
11919
12645
|
api: "anthropic-messages";
|
|
11920
12646
|
provider: string;
|
|
11921
12647
|
baseUrl: string;
|
|
11922
|
-
reasoning:
|
|
12648
|
+
reasoning: true;
|
|
11923
12649
|
input: "text"[];
|
|
11924
12650
|
cost: {
|
|
11925
12651
|
input: number;
|
|
@@ -11930,7 +12656,7 @@ export declare const MODELS: {
|
|
|
11930
12656
|
contextWindow: number;
|
|
11931
12657
|
maxTokens: number;
|
|
11932
12658
|
};
|
|
11933
|
-
readonly "deepseek/deepseek-
|
|
12659
|
+
readonly "deepseek/deepseek-v4-pro": {
|
|
11934
12660
|
id: string;
|
|
11935
12661
|
name: string;
|
|
11936
12662
|
api: "anthropic-messages";
|
|
@@ -12712,6 +13438,23 @@ export declare const MODELS: {
|
|
|
12712
13438
|
contextWindow: number;
|
|
12713
13439
|
maxTokens: number;
|
|
12714
13440
|
};
|
|
13441
|
+
readonly "moonshotai/kimi-k2.6": {
|
|
13442
|
+
id: string;
|
|
13443
|
+
name: string;
|
|
13444
|
+
api: "anthropic-messages";
|
|
13445
|
+
provider: string;
|
|
13446
|
+
baseUrl: string;
|
|
13447
|
+
reasoning: true;
|
|
13448
|
+
input: ("image" | "text")[];
|
|
13449
|
+
cost: {
|
|
13450
|
+
input: number;
|
|
13451
|
+
output: number;
|
|
13452
|
+
cacheRead: number;
|
|
13453
|
+
cacheWrite: number;
|
|
13454
|
+
};
|
|
13455
|
+
contextWindow: number;
|
|
13456
|
+
maxTokens: number;
|
|
13457
|
+
};
|
|
12715
13458
|
readonly "nvidia/nemotron-nano-12b-v2-vl": {
|
|
12716
13459
|
id: string;
|
|
12717
13460
|
name: string;
|
|
@@ -13205,6 +13948,40 @@ export declare const MODELS: {
|
|
|
13205
13948
|
contextWindow: number;
|
|
13206
13949
|
maxTokens: number;
|
|
13207
13950
|
};
|
|
13951
|
+
readonly "openai/gpt-5.5": {
|
|
13952
|
+
id: string;
|
|
13953
|
+
name: string;
|
|
13954
|
+
api: "anthropic-messages";
|
|
13955
|
+
provider: string;
|
|
13956
|
+
baseUrl: string;
|
|
13957
|
+
reasoning: true;
|
|
13958
|
+
input: ("image" | "text")[];
|
|
13959
|
+
cost: {
|
|
13960
|
+
input: number;
|
|
13961
|
+
output: number;
|
|
13962
|
+
cacheRead: number;
|
|
13963
|
+
cacheWrite: number;
|
|
13964
|
+
};
|
|
13965
|
+
contextWindow: number;
|
|
13966
|
+
maxTokens: number;
|
|
13967
|
+
};
|
|
13968
|
+
readonly "openai/gpt-5.5-pro": {
|
|
13969
|
+
id: string;
|
|
13970
|
+
name: string;
|
|
13971
|
+
api: "anthropic-messages";
|
|
13972
|
+
provider: string;
|
|
13973
|
+
baseUrl: string;
|
|
13974
|
+
reasoning: true;
|
|
13975
|
+
input: ("image" | "text")[];
|
|
13976
|
+
cost: {
|
|
13977
|
+
input: number;
|
|
13978
|
+
output: number;
|
|
13979
|
+
cacheRead: number;
|
|
13980
|
+
cacheWrite: number;
|
|
13981
|
+
};
|
|
13982
|
+
contextWindow: number;
|
|
13983
|
+
maxTokens: number;
|
|
13984
|
+
};
|
|
13208
13985
|
readonly "openai/gpt-oss-20b": {
|
|
13209
13986
|
id: string;
|
|
13210
13987
|
name: string;
|
|
@@ -13892,7 +14669,7 @@ export declare const MODELS: {
|
|
|
13892
14669
|
provider: string;
|
|
13893
14670
|
baseUrl: string;
|
|
13894
14671
|
reasoning: true;
|
|
13895
|
-
input:
|
|
14672
|
+
input: "text"[];
|
|
13896
14673
|
cost: {
|
|
13897
14674
|
input: number;
|
|
13898
14675
|
output: number;
|
|
@@ -14331,27 +15108,6 @@ export declare const MODELS: {
|
|
|
14331
15108
|
};
|
|
14332
15109
|
};
|
|
14333
15110
|
readonly zai: {
|
|
14334
|
-
readonly "glm-4.5": {
|
|
14335
|
-
id: string;
|
|
14336
|
-
name: string;
|
|
14337
|
-
api: "openai-completions";
|
|
14338
|
-
provider: string;
|
|
14339
|
-
baseUrl: string;
|
|
14340
|
-
compat: {
|
|
14341
|
-
supportsDeveloperRole: false;
|
|
14342
|
-
thinkingFormat: "zai";
|
|
14343
|
-
};
|
|
14344
|
-
reasoning: true;
|
|
14345
|
-
input: "text"[];
|
|
14346
|
-
cost: {
|
|
14347
|
-
input: number;
|
|
14348
|
-
output: number;
|
|
14349
|
-
cacheRead: number;
|
|
14350
|
-
cacheWrite: number;
|
|
14351
|
-
};
|
|
14352
|
-
contextWindow: number;
|
|
14353
|
-
maxTokens: number;
|
|
14354
|
-
};
|
|
14355
15111
|
readonly "glm-4.5-air": {
|
|
14356
15112
|
id: string;
|
|
14357
15113
|
name: string;
|
|
@@ -14373,92 +15129,6 @@ export declare const MODELS: {
|
|
|
14373
15129
|
contextWindow: number;
|
|
14374
15130
|
maxTokens: number;
|
|
14375
15131
|
};
|
|
14376
|
-
readonly "glm-4.5-flash": {
|
|
14377
|
-
id: string;
|
|
14378
|
-
name: string;
|
|
14379
|
-
api: "openai-completions";
|
|
14380
|
-
provider: string;
|
|
14381
|
-
baseUrl: string;
|
|
14382
|
-
compat: {
|
|
14383
|
-
supportsDeveloperRole: false;
|
|
14384
|
-
thinkingFormat: "zai";
|
|
14385
|
-
};
|
|
14386
|
-
reasoning: true;
|
|
14387
|
-
input: "text"[];
|
|
14388
|
-
cost: {
|
|
14389
|
-
input: number;
|
|
14390
|
-
output: number;
|
|
14391
|
-
cacheRead: number;
|
|
14392
|
-
cacheWrite: number;
|
|
14393
|
-
};
|
|
14394
|
-
contextWindow: number;
|
|
14395
|
-
maxTokens: number;
|
|
14396
|
-
};
|
|
14397
|
-
readonly "glm-4.5v": {
|
|
14398
|
-
id: string;
|
|
14399
|
-
name: string;
|
|
14400
|
-
api: "openai-completions";
|
|
14401
|
-
provider: string;
|
|
14402
|
-
baseUrl: string;
|
|
14403
|
-
compat: {
|
|
14404
|
-
supportsDeveloperRole: false;
|
|
14405
|
-
thinkingFormat: "zai";
|
|
14406
|
-
};
|
|
14407
|
-
reasoning: true;
|
|
14408
|
-
input: ("image" | "text")[];
|
|
14409
|
-
cost: {
|
|
14410
|
-
input: number;
|
|
14411
|
-
output: number;
|
|
14412
|
-
cacheRead: number;
|
|
14413
|
-
cacheWrite: number;
|
|
14414
|
-
};
|
|
14415
|
-
contextWindow: number;
|
|
14416
|
-
maxTokens: number;
|
|
14417
|
-
};
|
|
14418
|
-
readonly "glm-4.6": {
|
|
14419
|
-
id: string;
|
|
14420
|
-
name: string;
|
|
14421
|
-
api: "openai-completions";
|
|
14422
|
-
provider: string;
|
|
14423
|
-
baseUrl: string;
|
|
14424
|
-
compat: {
|
|
14425
|
-
supportsDeveloperRole: false;
|
|
14426
|
-
thinkingFormat: "zai";
|
|
14427
|
-
zaiToolStream: true;
|
|
14428
|
-
};
|
|
14429
|
-
reasoning: true;
|
|
14430
|
-
input: "text"[];
|
|
14431
|
-
cost: {
|
|
14432
|
-
input: number;
|
|
14433
|
-
output: number;
|
|
14434
|
-
cacheRead: number;
|
|
14435
|
-
cacheWrite: number;
|
|
14436
|
-
};
|
|
14437
|
-
contextWindow: number;
|
|
14438
|
-
maxTokens: number;
|
|
14439
|
-
};
|
|
14440
|
-
readonly "glm-4.6v": {
|
|
14441
|
-
id: string;
|
|
14442
|
-
name: string;
|
|
14443
|
-
api: "openai-completions";
|
|
14444
|
-
provider: string;
|
|
14445
|
-
baseUrl: string;
|
|
14446
|
-
compat: {
|
|
14447
|
-
supportsDeveloperRole: false;
|
|
14448
|
-
thinkingFormat: "zai";
|
|
14449
|
-
zaiToolStream: true;
|
|
14450
|
-
};
|
|
14451
|
-
reasoning: true;
|
|
14452
|
-
input: ("image" | "text")[];
|
|
14453
|
-
cost: {
|
|
14454
|
-
input: number;
|
|
14455
|
-
output: number;
|
|
14456
|
-
cacheRead: number;
|
|
14457
|
-
cacheWrite: number;
|
|
14458
|
-
};
|
|
14459
|
-
contextWindow: number;
|
|
14460
|
-
maxTokens: number;
|
|
14461
|
-
};
|
|
14462
15132
|
readonly "glm-4.7": {
|
|
14463
15133
|
id: string;
|
|
14464
15134
|
name: string;
|
|
@@ -14481,72 +15151,6 @@ export declare const MODELS: {
|
|
|
14481
15151
|
contextWindow: number;
|
|
14482
15152
|
maxTokens: number;
|
|
14483
15153
|
};
|
|
14484
|
-
readonly "glm-4.7-flash": {
|
|
14485
|
-
id: string;
|
|
14486
|
-
name: string;
|
|
14487
|
-
api: "openai-completions";
|
|
14488
|
-
provider: string;
|
|
14489
|
-
baseUrl: string;
|
|
14490
|
-
compat: {
|
|
14491
|
-
supportsDeveloperRole: false;
|
|
14492
|
-
thinkingFormat: "zai";
|
|
14493
|
-
zaiToolStream: true;
|
|
14494
|
-
};
|
|
14495
|
-
reasoning: true;
|
|
14496
|
-
input: "text"[];
|
|
14497
|
-
cost: {
|
|
14498
|
-
input: number;
|
|
14499
|
-
output: number;
|
|
14500
|
-
cacheRead: number;
|
|
14501
|
-
cacheWrite: number;
|
|
14502
|
-
};
|
|
14503
|
-
contextWindow: number;
|
|
14504
|
-
maxTokens: number;
|
|
14505
|
-
};
|
|
14506
|
-
readonly "glm-4.7-flashx": {
|
|
14507
|
-
id: string;
|
|
14508
|
-
name: string;
|
|
14509
|
-
api: "openai-completions";
|
|
14510
|
-
provider: string;
|
|
14511
|
-
baseUrl: string;
|
|
14512
|
-
compat: {
|
|
14513
|
-
supportsDeveloperRole: false;
|
|
14514
|
-
thinkingFormat: "zai";
|
|
14515
|
-
zaiToolStream: true;
|
|
14516
|
-
};
|
|
14517
|
-
reasoning: true;
|
|
14518
|
-
input: "text"[];
|
|
14519
|
-
cost: {
|
|
14520
|
-
input: number;
|
|
14521
|
-
output: number;
|
|
14522
|
-
cacheRead: number;
|
|
14523
|
-
cacheWrite: number;
|
|
14524
|
-
};
|
|
14525
|
-
contextWindow: number;
|
|
14526
|
-
maxTokens: number;
|
|
14527
|
-
};
|
|
14528
|
-
readonly "glm-5": {
|
|
14529
|
-
id: string;
|
|
14530
|
-
name: string;
|
|
14531
|
-
api: "openai-completions";
|
|
14532
|
-
provider: string;
|
|
14533
|
-
baseUrl: string;
|
|
14534
|
-
compat: {
|
|
14535
|
-
supportsDeveloperRole: false;
|
|
14536
|
-
thinkingFormat: "zai";
|
|
14537
|
-
zaiToolStream: true;
|
|
14538
|
-
};
|
|
14539
|
-
reasoning: true;
|
|
14540
|
-
input: "text"[];
|
|
14541
|
-
cost: {
|
|
14542
|
-
input: number;
|
|
14543
|
-
output: number;
|
|
14544
|
-
cacheRead: number;
|
|
14545
|
-
cacheWrite: number;
|
|
14546
|
-
};
|
|
14547
|
-
contextWindow: number;
|
|
14548
|
-
maxTokens: number;
|
|
14549
|
-
};
|
|
14550
15154
|
readonly "glm-5-turbo": {
|
|
14551
15155
|
id: string;
|
|
14552
15156
|
name: string;
|
|
@@ -14591,28 +15195,6 @@ export declare const MODELS: {
|
|
|
14591
15195
|
contextWindow: number;
|
|
14592
15196
|
maxTokens: number;
|
|
14593
15197
|
};
|
|
14594
|
-
readonly "glm-5v-turbo": {
|
|
14595
|
-
id: string;
|
|
14596
|
-
name: string;
|
|
14597
|
-
api: "openai-completions";
|
|
14598
|
-
provider: string;
|
|
14599
|
-
baseUrl: string;
|
|
14600
|
-
compat: {
|
|
14601
|
-
supportsDeveloperRole: false;
|
|
14602
|
-
thinkingFormat: "zai";
|
|
14603
|
-
zaiToolStream: true;
|
|
14604
|
-
};
|
|
14605
|
-
reasoning: true;
|
|
14606
|
-
input: ("image" | "text")[];
|
|
14607
|
-
cost: {
|
|
14608
|
-
input: number;
|
|
14609
|
-
output: number;
|
|
14610
|
-
cacheRead: number;
|
|
14611
|
-
cacheWrite: number;
|
|
14612
|
-
};
|
|
14613
|
-
contextWindow: number;
|
|
14614
|
-
maxTokens: number;
|
|
14615
|
-
};
|
|
14616
15198
|
};
|
|
14617
15199
|
};
|
|
14618
15200
|
//# sourceMappingURL=models.generated.d.ts.map
|