@avadisabelle/ava-pi-ai 0.65.0 → 0.65.3
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/models.generated.d.ts +1168 -394
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +1167 -373
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/amazon-bedrock.d.ts.map +1 -1
- package/dist/providers/amazon-bedrock.js +57 -26
- package/dist/providers/amazon-bedrock.js.map +1 -1
- package/dist/providers/anthropic.d.ts.map +1 -1
- package/dist/providers/anthropic.js +7 -2
- package/dist/providers/anthropic.js.map +1 -1
- package/dist/providers/azure-openai-responses.d.ts.map +1 -1
- package/dist/providers/azure-openai-responses.js +7 -4
- package/dist/providers/azure-openai-responses.js.map +1 -1
- package/dist/providers/openai-codex-responses.d.ts.map +1 -1
- package/dist/providers/openai-codex-responses.js +1 -1
- package/dist/providers/openai-codex-responses.js.map +1 -1
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +7 -2
- package/dist/providers/openai-completions.js.map +1 -1
- package/dist/providers/openai-responses.d.ts.map +1 -1
- package/dist/providers/openai-responses.js +7 -4
- package/dist/providers/openai-responses.js.map +1 -1
- package/dist/providers/simple-options.d.ts.map +1 -1
- package/dist/providers/simple-options.js +2 -0
- package/dist/providers/simple-options.js.map +1 -1
- package/dist/types.d.ts +10 -0
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/package.json +1 -1
|
@@ -323,6 +323,40 @@ export declare const MODELS: {
|
|
|
323
323
|
contextWindow: number;
|
|
324
324
|
maxTokens: number;
|
|
325
325
|
};
|
|
326
|
+
readonly "au.anthropic.claude-opus-4-6-v1": {
|
|
327
|
+
id: string;
|
|
328
|
+
name: string;
|
|
329
|
+
api: "bedrock-converse-stream";
|
|
330
|
+
provider: string;
|
|
331
|
+
baseUrl: string;
|
|
332
|
+
reasoning: true;
|
|
333
|
+
input: ("image" | "text")[];
|
|
334
|
+
cost: {
|
|
335
|
+
input: number;
|
|
336
|
+
output: number;
|
|
337
|
+
cacheRead: number;
|
|
338
|
+
cacheWrite: number;
|
|
339
|
+
};
|
|
340
|
+
contextWindow: number;
|
|
341
|
+
maxTokens: number;
|
|
342
|
+
};
|
|
343
|
+
readonly "au.anthropic.claude-sonnet-4-6": {
|
|
344
|
+
id: string;
|
|
345
|
+
name: string;
|
|
346
|
+
api: "bedrock-converse-stream";
|
|
347
|
+
provider: string;
|
|
348
|
+
baseUrl: string;
|
|
349
|
+
reasoning: true;
|
|
350
|
+
input: ("image" | "text")[];
|
|
351
|
+
cost: {
|
|
352
|
+
input: number;
|
|
353
|
+
output: number;
|
|
354
|
+
cacheRead: number;
|
|
355
|
+
cacheWrite: number;
|
|
356
|
+
};
|
|
357
|
+
contextWindow: number;
|
|
358
|
+
maxTokens: number;
|
|
359
|
+
};
|
|
326
360
|
readonly "deepseek.r1-v1:0": {
|
|
327
361
|
id: string;
|
|
328
362
|
name: string;
|
|
@@ -2486,6 +2520,40 @@ export declare const MODELS: {
|
|
|
2486
2520
|
contextWindow: number;
|
|
2487
2521
|
maxTokens: number;
|
|
2488
2522
|
};
|
|
2523
|
+
readonly "gpt-5.5": {
|
|
2524
|
+
id: string;
|
|
2525
|
+
name: string;
|
|
2526
|
+
api: "azure-openai-responses";
|
|
2527
|
+
provider: string;
|
|
2528
|
+
baseUrl: string;
|
|
2529
|
+
reasoning: true;
|
|
2530
|
+
input: ("image" | "text")[];
|
|
2531
|
+
cost: {
|
|
2532
|
+
input: number;
|
|
2533
|
+
output: number;
|
|
2534
|
+
cacheRead: number;
|
|
2535
|
+
cacheWrite: number;
|
|
2536
|
+
};
|
|
2537
|
+
contextWindow: number;
|
|
2538
|
+
maxTokens: number;
|
|
2539
|
+
};
|
|
2540
|
+
readonly "gpt-5.5-pro": {
|
|
2541
|
+
id: string;
|
|
2542
|
+
name: string;
|
|
2543
|
+
api: "azure-openai-responses";
|
|
2544
|
+
provider: string;
|
|
2545
|
+
baseUrl: string;
|
|
2546
|
+
reasoning: true;
|
|
2547
|
+
input: ("image" | "text")[];
|
|
2548
|
+
cost: {
|
|
2549
|
+
input: number;
|
|
2550
|
+
output: number;
|
|
2551
|
+
cacheRead: number;
|
|
2552
|
+
cacheWrite: number;
|
|
2553
|
+
};
|
|
2554
|
+
contextWindow: number;
|
|
2555
|
+
maxTokens: number;
|
|
2556
|
+
};
|
|
2489
2557
|
readonly o1: {
|
|
2490
2558
|
id: string;
|
|
2491
2559
|
name: string;
|
|
@@ -3276,6 +3344,29 @@ export declare const MODELS: {
|
|
|
3276
3344
|
contextWindow: number;
|
|
3277
3345
|
maxTokens: number;
|
|
3278
3346
|
};
|
|
3347
|
+
readonly "gpt-5.5": {
|
|
3348
|
+
id: string;
|
|
3349
|
+
name: string;
|
|
3350
|
+
api: "openai-responses";
|
|
3351
|
+
provider: string;
|
|
3352
|
+
baseUrl: string;
|
|
3353
|
+
headers: {
|
|
3354
|
+
"User-Agent": string;
|
|
3355
|
+
"Editor-Version": string;
|
|
3356
|
+
"Editor-Plugin-Version": string;
|
|
3357
|
+
"Copilot-Integration-Id": string;
|
|
3358
|
+
};
|
|
3359
|
+
reasoning: true;
|
|
3360
|
+
input: ("image" | "text")[];
|
|
3361
|
+
cost: {
|
|
3362
|
+
input: number;
|
|
3363
|
+
output: number;
|
|
3364
|
+
cacheRead: number;
|
|
3365
|
+
cacheWrite: number;
|
|
3366
|
+
};
|
|
3367
|
+
contextWindow: number;
|
|
3368
|
+
maxTokens: number;
|
|
3369
|
+
};
|
|
3279
3370
|
readonly "grok-code-fast-1": {
|
|
3280
3371
|
id: string;
|
|
3281
3372
|
name: string;
|
|
@@ -3731,7 +3822,7 @@ export declare const MODELS: {
|
|
|
3731
3822
|
contextWindow: number;
|
|
3732
3823
|
maxTokens: number;
|
|
3733
3824
|
};
|
|
3734
|
-
readonly "gemma-4-26b-it": {
|
|
3825
|
+
readonly "gemma-4-26b-a4b-it": {
|
|
3735
3826
|
id: string;
|
|
3736
3827
|
name: string;
|
|
3737
3828
|
api: "google-generative-ai";
|
|
@@ -4797,6 +4888,26 @@ export declare const MODELS: {
|
|
|
4797
4888
|
contextWindow: number;
|
|
4798
4889
|
maxTokens: number;
|
|
4799
4890
|
};
|
|
4891
|
+
readonly "deepseek-ai/DeepSeek-V4-Pro": {
|
|
4892
|
+
id: string;
|
|
4893
|
+
name: string;
|
|
4894
|
+
api: "openai-completions";
|
|
4895
|
+
provider: string;
|
|
4896
|
+
baseUrl: string;
|
|
4897
|
+
compat: {
|
|
4898
|
+
supportsDeveloperRole: false;
|
|
4899
|
+
};
|
|
4900
|
+
reasoning: true;
|
|
4901
|
+
input: "text"[];
|
|
4902
|
+
cost: {
|
|
4903
|
+
input: number;
|
|
4904
|
+
output: number;
|
|
4905
|
+
cacheRead: number;
|
|
4906
|
+
cacheWrite: number;
|
|
4907
|
+
};
|
|
4908
|
+
contextWindow: number;
|
|
4909
|
+
maxTokens: number;
|
|
4910
|
+
};
|
|
4800
4911
|
readonly "moonshotai/Kimi-K2-Instruct": {
|
|
4801
4912
|
id: string;
|
|
4802
4913
|
name: string;
|
|
@@ -4877,6 +4988,26 @@ export declare const MODELS: {
|
|
|
4877
4988
|
contextWindow: number;
|
|
4878
4989
|
maxTokens: number;
|
|
4879
4990
|
};
|
|
4991
|
+
readonly "moonshotai/Kimi-K2.6": {
|
|
4992
|
+
id: string;
|
|
4993
|
+
name: string;
|
|
4994
|
+
api: "openai-completions";
|
|
4995
|
+
provider: string;
|
|
4996
|
+
baseUrl: string;
|
|
4997
|
+
compat: {
|
|
4998
|
+
supportsDeveloperRole: false;
|
|
4999
|
+
};
|
|
5000
|
+
reasoning: true;
|
|
5001
|
+
input: ("image" | "text")[];
|
|
5002
|
+
cost: {
|
|
5003
|
+
input: number;
|
|
5004
|
+
output: number;
|
|
5005
|
+
cacheRead: number;
|
|
5006
|
+
cacheWrite: number;
|
|
5007
|
+
};
|
|
5008
|
+
contextWindow: number;
|
|
5009
|
+
maxTokens: number;
|
|
5010
|
+
};
|
|
4880
5011
|
readonly "zai-org/GLM-4.7": {
|
|
4881
5012
|
id: string;
|
|
4882
5013
|
name: string;
|
|
@@ -4959,6 +5090,23 @@ export declare const MODELS: {
|
|
|
4959
5090
|
};
|
|
4960
5091
|
};
|
|
4961
5092
|
readonly "kimi-coding": {
|
|
5093
|
+
readonly k2p6: {
|
|
5094
|
+
id: string;
|
|
5095
|
+
name: string;
|
|
5096
|
+
api: "anthropic-messages";
|
|
5097
|
+
provider: string;
|
|
5098
|
+
baseUrl: string;
|
|
5099
|
+
reasoning: true;
|
|
5100
|
+
input: ("image" | "text")[];
|
|
5101
|
+
cost: {
|
|
5102
|
+
input: number;
|
|
5103
|
+
output: number;
|
|
5104
|
+
cacheRead: number;
|
|
5105
|
+
cacheWrite: number;
|
|
5106
|
+
};
|
|
5107
|
+
contextWindow: number;
|
|
5108
|
+
maxTokens: number;
|
|
5109
|
+
};
|
|
4962
5110
|
readonly "kimi-for-coding": {
|
|
4963
5111
|
id: string;
|
|
4964
5112
|
name: string;
|
|
@@ -5339,13 +5487,30 @@ export declare const MODELS: {
|
|
|
5339
5487
|
contextWindow: number;
|
|
5340
5488
|
maxTokens: number;
|
|
5341
5489
|
};
|
|
5490
|
+
readonly "mistral-medium-2604": {
|
|
5491
|
+
id: string;
|
|
5492
|
+
name: string;
|
|
5493
|
+
api: "mistral-conversations";
|
|
5494
|
+
provider: string;
|
|
5495
|
+
baseUrl: string;
|
|
5496
|
+
reasoning: true;
|
|
5497
|
+
input: ("image" | "text")[];
|
|
5498
|
+
cost: {
|
|
5499
|
+
input: number;
|
|
5500
|
+
output: number;
|
|
5501
|
+
cacheRead: number;
|
|
5502
|
+
cacheWrite: number;
|
|
5503
|
+
};
|
|
5504
|
+
contextWindow: number;
|
|
5505
|
+
maxTokens: number;
|
|
5506
|
+
};
|
|
5342
5507
|
readonly "mistral-medium-latest": {
|
|
5343
5508
|
id: string;
|
|
5344
5509
|
name: string;
|
|
5345
5510
|
api: "mistral-conversations";
|
|
5346
5511
|
provider: string;
|
|
5347
5512
|
baseUrl: string;
|
|
5348
|
-
reasoning:
|
|
5513
|
+
reasoning: true;
|
|
5349
5514
|
input: ("image" | "text")[];
|
|
5350
5515
|
cost: {
|
|
5351
5516
|
input: number;
|
|
@@ -6055,6 +6220,40 @@ export declare const MODELS: {
|
|
|
6055
6220
|
contextWindow: number;
|
|
6056
6221
|
maxTokens: number;
|
|
6057
6222
|
};
|
|
6223
|
+
readonly "gpt-5.5": {
|
|
6224
|
+
id: string;
|
|
6225
|
+
name: string;
|
|
6226
|
+
api: "openai-responses";
|
|
6227
|
+
provider: string;
|
|
6228
|
+
baseUrl: string;
|
|
6229
|
+
reasoning: true;
|
|
6230
|
+
input: ("image" | "text")[];
|
|
6231
|
+
cost: {
|
|
6232
|
+
input: number;
|
|
6233
|
+
output: number;
|
|
6234
|
+
cacheRead: number;
|
|
6235
|
+
cacheWrite: number;
|
|
6236
|
+
};
|
|
6237
|
+
contextWindow: number;
|
|
6238
|
+
maxTokens: number;
|
|
6239
|
+
};
|
|
6240
|
+
readonly "gpt-5.5-pro": {
|
|
6241
|
+
id: string;
|
|
6242
|
+
name: string;
|
|
6243
|
+
api: "openai-responses";
|
|
6244
|
+
provider: string;
|
|
6245
|
+
baseUrl: string;
|
|
6246
|
+
reasoning: true;
|
|
6247
|
+
input: ("image" | "text")[];
|
|
6248
|
+
cost: {
|
|
6249
|
+
input: number;
|
|
6250
|
+
output: number;
|
|
6251
|
+
cacheRead: number;
|
|
6252
|
+
cacheWrite: number;
|
|
6253
|
+
};
|
|
6254
|
+
contextWindow: number;
|
|
6255
|
+
maxTokens: number;
|
|
6256
|
+
};
|
|
6058
6257
|
readonly o1: {
|
|
6059
6258
|
id: string;
|
|
6060
6259
|
name: string;
|
|
@@ -6365,23 +6564,6 @@ export declare const MODELS: {
|
|
|
6365
6564
|
contextWindow: number;
|
|
6366
6565
|
maxTokens: number;
|
|
6367
6566
|
};
|
|
6368
|
-
readonly "claude-3-5-haiku": {
|
|
6369
|
-
id: string;
|
|
6370
|
-
name: string;
|
|
6371
|
-
api: "anthropic-messages";
|
|
6372
|
-
provider: string;
|
|
6373
|
-
baseUrl: string;
|
|
6374
|
-
reasoning: false;
|
|
6375
|
-
input: ("image" | "text")[];
|
|
6376
|
-
cost: {
|
|
6377
|
-
input: number;
|
|
6378
|
-
output: number;
|
|
6379
|
-
cacheRead: number;
|
|
6380
|
-
cacheWrite: number;
|
|
6381
|
-
};
|
|
6382
|
-
contextWindow: number;
|
|
6383
|
-
maxTokens: number;
|
|
6384
|
-
};
|
|
6385
6567
|
readonly "claude-haiku-4-5": {
|
|
6386
6568
|
id: string;
|
|
6387
6569
|
name: string;
|
|
@@ -6824,10 +7006,10 @@ export declare const MODELS: {
|
|
|
6824
7006
|
contextWindow: number;
|
|
6825
7007
|
maxTokens: number;
|
|
6826
7008
|
};
|
|
6827
|
-
readonly "
|
|
7009
|
+
readonly "gpt-5.5": {
|
|
6828
7010
|
id: string;
|
|
6829
7011
|
name: string;
|
|
6830
|
-
api: "openai-
|
|
7012
|
+
api: "openai-responses";
|
|
6831
7013
|
provider: string;
|
|
6832
7014
|
baseUrl: string;
|
|
6833
7015
|
reasoning: true;
|
|
@@ -6841,14 +7023,14 @@ export declare const MODELS: {
|
|
|
6841
7023
|
contextWindow: number;
|
|
6842
7024
|
maxTokens: number;
|
|
6843
7025
|
};
|
|
6844
|
-
readonly "
|
|
7026
|
+
readonly "gpt-5.5-pro": {
|
|
6845
7027
|
id: string;
|
|
6846
7028
|
name: string;
|
|
6847
|
-
api: "openai-
|
|
7029
|
+
api: "openai-responses";
|
|
6848
7030
|
provider: string;
|
|
6849
7031
|
baseUrl: string;
|
|
6850
7032
|
reasoning: true;
|
|
6851
|
-
input: "text"[];
|
|
7033
|
+
input: ("image" | "text")[];
|
|
6852
7034
|
cost: {
|
|
6853
7035
|
input: number;
|
|
6854
7036
|
output: number;
|
|
@@ -6858,10 +7040,10 @@ export declare const MODELS: {
|
|
|
6858
7040
|
contextWindow: number;
|
|
6859
7041
|
maxTokens: number;
|
|
6860
7042
|
};
|
|
6861
|
-
readonly "
|
|
7043
|
+
readonly "hy3-preview-free": {
|
|
6862
7044
|
id: string;
|
|
6863
7045
|
name: string;
|
|
6864
|
-
api: "
|
|
7046
|
+
api: "openai-completions";
|
|
6865
7047
|
provider: string;
|
|
6866
7048
|
baseUrl: string;
|
|
6867
7049
|
reasoning: true;
|
|
@@ -6875,14 +7057,14 @@ export declare const MODELS: {
|
|
|
6875
7057
|
contextWindow: number;
|
|
6876
7058
|
maxTokens: number;
|
|
6877
7059
|
};
|
|
6878
|
-
readonly "
|
|
7060
|
+
readonly "kimi-k2.5": {
|
|
6879
7061
|
id: string;
|
|
6880
7062
|
name: string;
|
|
6881
7063
|
api: "openai-completions";
|
|
6882
7064
|
provider: string;
|
|
6883
7065
|
baseUrl: string;
|
|
6884
7066
|
reasoning: true;
|
|
6885
|
-
input: "text"[];
|
|
7067
|
+
input: ("image" | "text")[];
|
|
6886
7068
|
cost: {
|
|
6887
7069
|
input: number;
|
|
6888
7070
|
output: number;
|
|
@@ -6892,7 +7074,7 @@ export declare const MODELS: {
|
|
|
6892
7074
|
contextWindow: number;
|
|
6893
7075
|
maxTokens: number;
|
|
6894
7076
|
};
|
|
6895
|
-
readonly "
|
|
7077
|
+
readonly "kimi-k2.6": {
|
|
6896
7078
|
id: string;
|
|
6897
7079
|
name: string;
|
|
6898
7080
|
api: "openai-completions";
|
|
@@ -6909,14 +7091,14 @@ export declare const MODELS: {
|
|
|
6909
7091
|
contextWindow: number;
|
|
6910
7092
|
maxTokens: number;
|
|
6911
7093
|
};
|
|
6912
|
-
readonly "
|
|
7094
|
+
readonly "minimax-m2.5": {
|
|
6913
7095
|
id: string;
|
|
6914
7096
|
name: string;
|
|
6915
7097
|
api: "openai-completions";
|
|
6916
7098
|
provider: string;
|
|
6917
7099
|
baseUrl: string;
|
|
6918
7100
|
reasoning: true;
|
|
6919
|
-
input:
|
|
7101
|
+
input: "text"[];
|
|
6920
7102
|
cost: {
|
|
6921
7103
|
input: number;
|
|
6922
7104
|
output: number;
|
|
@@ -6926,12 +7108,10 @@ export declare const MODELS: {
|
|
|
6926
7108
|
contextWindow: number;
|
|
6927
7109
|
maxTokens: number;
|
|
6928
7110
|
};
|
|
6929
|
-
|
|
6930
|
-
readonly "opencode-go": {
|
|
6931
|
-
readonly "glm-5": {
|
|
7111
|
+
readonly "minimax-m2.5-free": {
|
|
6932
7112
|
id: string;
|
|
6933
7113
|
name: string;
|
|
6934
|
-
api: "
|
|
7114
|
+
api: "anthropic-messages";
|
|
6935
7115
|
provider: string;
|
|
6936
7116
|
baseUrl: string;
|
|
6937
7117
|
reasoning: true;
|
|
@@ -6945,7 +7125,7 @@ export declare const MODELS: {
|
|
|
6945
7125
|
contextWindow: number;
|
|
6946
7126
|
maxTokens: number;
|
|
6947
7127
|
};
|
|
6948
|
-
readonly "
|
|
7128
|
+
readonly "minimax-m2.7": {
|
|
6949
7129
|
id: string;
|
|
6950
7130
|
name: string;
|
|
6951
7131
|
api: "openai-completions";
|
|
@@ -6962,14 +7142,14 @@ export declare const MODELS: {
|
|
|
6962
7142
|
contextWindow: number;
|
|
6963
7143
|
maxTokens: number;
|
|
6964
7144
|
};
|
|
6965
|
-
readonly "
|
|
7145
|
+
readonly "nemotron-3-super-free": {
|
|
6966
7146
|
id: string;
|
|
6967
7147
|
name: string;
|
|
6968
7148
|
api: "openai-completions";
|
|
6969
7149
|
provider: string;
|
|
6970
7150
|
baseUrl: string;
|
|
6971
7151
|
reasoning: true;
|
|
6972
|
-
input:
|
|
7152
|
+
input: "text"[];
|
|
6973
7153
|
cost: {
|
|
6974
7154
|
input: number;
|
|
6975
7155
|
output: number;
|
|
@@ -6979,10 +7159,10 @@ export declare const MODELS: {
|
|
|
6979
7159
|
contextWindow: number;
|
|
6980
7160
|
maxTokens: number;
|
|
6981
7161
|
};
|
|
6982
|
-
readonly "
|
|
7162
|
+
readonly "qwen3.5-plus": {
|
|
6983
7163
|
id: string;
|
|
6984
7164
|
name: string;
|
|
6985
|
-
api: "
|
|
7165
|
+
api: "anthropic-messages";
|
|
6986
7166
|
provider: string;
|
|
6987
7167
|
baseUrl: string;
|
|
6988
7168
|
reasoning: true;
|
|
@@ -6996,14 +7176,14 @@ export declare const MODELS: {
|
|
|
6996
7176
|
contextWindow: number;
|
|
6997
7177
|
maxTokens: number;
|
|
6998
7178
|
};
|
|
6999
|
-
readonly "
|
|
7179
|
+
readonly "qwen3.6-plus": {
|
|
7000
7180
|
id: string;
|
|
7001
7181
|
name: string;
|
|
7002
|
-
api: "
|
|
7182
|
+
api: "anthropic-messages";
|
|
7003
7183
|
provider: string;
|
|
7004
7184
|
baseUrl: string;
|
|
7005
7185
|
reasoning: true;
|
|
7006
|
-
input: "text"[];
|
|
7186
|
+
input: ("image" | "text")[];
|
|
7007
7187
|
cost: {
|
|
7008
7188
|
input: number;
|
|
7009
7189
|
output: number;
|
|
@@ -7013,7 +7193,9 @@ export declare const MODELS: {
|
|
|
7013
7193
|
contextWindow: number;
|
|
7014
7194
|
maxTokens: number;
|
|
7015
7195
|
};
|
|
7016
|
-
|
|
7196
|
+
};
|
|
7197
|
+
readonly "opencode-go": {
|
|
7198
|
+
readonly "deepseek-v4-flash": {
|
|
7017
7199
|
id: string;
|
|
7018
7200
|
name: string;
|
|
7019
7201
|
api: "openai-completions";
|
|
@@ -7030,10 +7212,10 @@ export declare const MODELS: {
|
|
|
7030
7212
|
contextWindow: number;
|
|
7031
7213
|
maxTokens: number;
|
|
7032
7214
|
};
|
|
7033
|
-
readonly "
|
|
7215
|
+
readonly "deepseek-v4-pro": {
|
|
7034
7216
|
id: string;
|
|
7035
7217
|
name: string;
|
|
7036
|
-
api: "
|
|
7218
|
+
api: "openai-completions";
|
|
7037
7219
|
provider: string;
|
|
7038
7220
|
baseUrl: string;
|
|
7039
7221
|
reasoning: true;
|
|
@@ -7047,7 +7229,41 @@ export declare const MODELS: {
|
|
|
7047
7229
|
contextWindow: number;
|
|
7048
7230
|
maxTokens: number;
|
|
7049
7231
|
};
|
|
7050
|
-
readonly "
|
|
7232
|
+
readonly "glm-5": {
|
|
7233
|
+
id: string;
|
|
7234
|
+
name: string;
|
|
7235
|
+
api: "openai-completions";
|
|
7236
|
+
provider: string;
|
|
7237
|
+
baseUrl: string;
|
|
7238
|
+
reasoning: true;
|
|
7239
|
+
input: "text"[];
|
|
7240
|
+
cost: {
|
|
7241
|
+
input: number;
|
|
7242
|
+
output: number;
|
|
7243
|
+
cacheRead: number;
|
|
7244
|
+
cacheWrite: number;
|
|
7245
|
+
};
|
|
7246
|
+
contextWindow: number;
|
|
7247
|
+
maxTokens: number;
|
|
7248
|
+
};
|
|
7249
|
+
readonly "glm-5.1": {
|
|
7250
|
+
id: string;
|
|
7251
|
+
name: string;
|
|
7252
|
+
api: "openai-completions";
|
|
7253
|
+
provider: string;
|
|
7254
|
+
baseUrl: string;
|
|
7255
|
+
reasoning: true;
|
|
7256
|
+
input: "text"[];
|
|
7257
|
+
cost: {
|
|
7258
|
+
input: number;
|
|
7259
|
+
output: number;
|
|
7260
|
+
cacheRead: number;
|
|
7261
|
+
cacheWrite: number;
|
|
7262
|
+
};
|
|
7263
|
+
contextWindow: number;
|
|
7264
|
+
maxTokens: number;
|
|
7265
|
+
};
|
|
7266
|
+
readonly "kimi-k2.5": {
|
|
7051
7267
|
id: string;
|
|
7052
7268
|
name: string;
|
|
7053
7269
|
api: "openai-completions";
|
|
@@ -7064,13 +7280,149 @@ export declare const MODELS: {
|
|
|
7064
7280
|
contextWindow: number;
|
|
7065
7281
|
maxTokens: number;
|
|
7066
7282
|
};
|
|
7067
|
-
readonly "
|
|
7283
|
+
readonly "kimi-k2.6": {
|
|
7284
|
+
id: string;
|
|
7285
|
+
name: string;
|
|
7286
|
+
api: "openai-completions";
|
|
7287
|
+
provider: string;
|
|
7288
|
+
baseUrl: string;
|
|
7289
|
+
reasoning: true;
|
|
7290
|
+
input: ("image" | "text")[];
|
|
7291
|
+
cost: {
|
|
7292
|
+
input: number;
|
|
7293
|
+
output: number;
|
|
7294
|
+
cacheRead: number;
|
|
7295
|
+
cacheWrite: number;
|
|
7296
|
+
};
|
|
7297
|
+
contextWindow: number;
|
|
7298
|
+
maxTokens: number;
|
|
7299
|
+
};
|
|
7300
|
+
readonly "mimo-v2-omni": {
|
|
7301
|
+
id: string;
|
|
7302
|
+
name: string;
|
|
7303
|
+
api: "openai-completions";
|
|
7304
|
+
provider: string;
|
|
7305
|
+
baseUrl: string;
|
|
7306
|
+
reasoning: true;
|
|
7307
|
+
input: ("image" | "text")[];
|
|
7308
|
+
cost: {
|
|
7309
|
+
input: number;
|
|
7310
|
+
output: number;
|
|
7311
|
+
cacheRead: number;
|
|
7312
|
+
cacheWrite: number;
|
|
7313
|
+
};
|
|
7314
|
+
contextWindow: number;
|
|
7315
|
+
maxTokens: number;
|
|
7316
|
+
};
|
|
7317
|
+
readonly "mimo-v2-pro": {
|
|
7318
|
+
id: string;
|
|
7319
|
+
name: string;
|
|
7320
|
+
api: "openai-completions";
|
|
7321
|
+
provider: string;
|
|
7322
|
+
baseUrl: string;
|
|
7323
|
+
reasoning: true;
|
|
7324
|
+
input: "text"[];
|
|
7325
|
+
cost: {
|
|
7326
|
+
input: number;
|
|
7327
|
+
output: number;
|
|
7328
|
+
cacheRead: number;
|
|
7329
|
+
cacheWrite: number;
|
|
7330
|
+
};
|
|
7331
|
+
contextWindow: number;
|
|
7332
|
+
maxTokens: number;
|
|
7333
|
+
};
|
|
7334
|
+
readonly "mimo-v2.5": {
|
|
7335
|
+
id: string;
|
|
7336
|
+
name: string;
|
|
7337
|
+
api: "openai-completions";
|
|
7338
|
+
provider: string;
|
|
7339
|
+
baseUrl: string;
|
|
7340
|
+
reasoning: true;
|
|
7341
|
+
input: ("image" | "text")[];
|
|
7342
|
+
cost: {
|
|
7343
|
+
input: number;
|
|
7344
|
+
output: number;
|
|
7345
|
+
cacheRead: number;
|
|
7346
|
+
cacheWrite: number;
|
|
7347
|
+
};
|
|
7348
|
+
contextWindow: number;
|
|
7349
|
+
maxTokens: number;
|
|
7350
|
+
};
|
|
7351
|
+
readonly "mimo-v2.5-pro": {
|
|
7352
|
+
id: string;
|
|
7353
|
+
name: string;
|
|
7354
|
+
api: "openai-completions";
|
|
7355
|
+
provider: string;
|
|
7356
|
+
baseUrl: string;
|
|
7357
|
+
reasoning: true;
|
|
7358
|
+
input: "text"[];
|
|
7359
|
+
cost: {
|
|
7360
|
+
input: number;
|
|
7361
|
+
output: number;
|
|
7362
|
+
cacheRead: number;
|
|
7363
|
+
cacheWrite: number;
|
|
7364
|
+
};
|
|
7365
|
+
contextWindow: number;
|
|
7366
|
+
maxTokens: number;
|
|
7367
|
+
};
|
|
7368
|
+
readonly "minimax-m2.5": {
|
|
7068
7369
|
id: string;
|
|
7069
7370
|
name: string;
|
|
7070
7371
|
api: "openai-completions";
|
|
7071
7372
|
provider: string;
|
|
7072
7373
|
baseUrl: string;
|
|
7073
7374
|
reasoning: true;
|
|
7375
|
+
input: "text"[];
|
|
7376
|
+
cost: {
|
|
7377
|
+
input: number;
|
|
7378
|
+
output: number;
|
|
7379
|
+
cacheRead: number;
|
|
7380
|
+
cacheWrite: number;
|
|
7381
|
+
};
|
|
7382
|
+
contextWindow: number;
|
|
7383
|
+
maxTokens: number;
|
|
7384
|
+
};
|
|
7385
|
+
readonly "minimax-m2.7": {
|
|
7386
|
+
id: string;
|
|
7387
|
+
name: string;
|
|
7388
|
+
api: "anthropic-messages";
|
|
7389
|
+
provider: string;
|
|
7390
|
+
baseUrl: string;
|
|
7391
|
+
reasoning: true;
|
|
7392
|
+
input: "text"[];
|
|
7393
|
+
cost: {
|
|
7394
|
+
input: number;
|
|
7395
|
+
output: number;
|
|
7396
|
+
cacheRead: number;
|
|
7397
|
+
cacheWrite: number;
|
|
7398
|
+
};
|
|
7399
|
+
contextWindow: number;
|
|
7400
|
+
maxTokens: number;
|
|
7401
|
+
};
|
|
7402
|
+
readonly "qwen3.5-plus": {
|
|
7403
|
+
id: string;
|
|
7404
|
+
name: string;
|
|
7405
|
+
api: "anthropic-messages";
|
|
7406
|
+
provider: string;
|
|
7407
|
+
baseUrl: string;
|
|
7408
|
+
reasoning: true;
|
|
7409
|
+
input: ("image" | "text")[];
|
|
7410
|
+
cost: {
|
|
7411
|
+
input: number;
|
|
7412
|
+
output: number;
|
|
7413
|
+
cacheRead: number;
|
|
7414
|
+
cacheWrite: number;
|
|
7415
|
+
};
|
|
7416
|
+
contextWindow: number;
|
|
7417
|
+
maxTokens: number;
|
|
7418
|
+
};
|
|
7419
|
+
readonly "qwen3.6-plus": {
|
|
7420
|
+
id: string;
|
|
7421
|
+
name: string;
|
|
7422
|
+
api: "anthropic-messages";
|
|
7423
|
+
provider: string;
|
|
7424
|
+
baseUrl: string;
|
|
7425
|
+
reasoning: true;
|
|
7074
7426
|
input: ("image" | "text")[];
|
|
7075
7427
|
cost: {
|
|
7076
7428
|
input: number;
|
|
@@ -7457,7 +7809,7 @@ export declare const MODELS: {
|
|
|
7457
7809
|
contextWindow: number;
|
|
7458
7810
|
maxTokens: number;
|
|
7459
7811
|
};
|
|
7460
|
-
readonly "arcee-ai/trinity-large-preview
|
|
7812
|
+
readonly "arcee-ai/trinity-large-preview": {
|
|
7461
7813
|
id: string;
|
|
7462
7814
|
name: string;
|
|
7463
7815
|
api: "openai-completions";
|
|
@@ -7814,6 +8166,40 @@ export declare const MODELS: {
|
|
|
7814
8166
|
contextWindow: number;
|
|
7815
8167
|
maxTokens: number;
|
|
7816
8168
|
};
|
|
8169
|
+
readonly "deepseek/deepseek-v4-flash": {
|
|
8170
|
+
id: string;
|
|
8171
|
+
name: string;
|
|
8172
|
+
api: "openai-completions";
|
|
8173
|
+
provider: string;
|
|
8174
|
+
baseUrl: string;
|
|
8175
|
+
reasoning: true;
|
|
8176
|
+
input: "text"[];
|
|
8177
|
+
cost: {
|
|
8178
|
+
input: number;
|
|
8179
|
+
output: number;
|
|
8180
|
+
cacheRead: number;
|
|
8181
|
+
cacheWrite: number;
|
|
8182
|
+
};
|
|
8183
|
+
contextWindow: number;
|
|
8184
|
+
maxTokens: number;
|
|
8185
|
+
};
|
|
8186
|
+
readonly "deepseek/deepseek-v4-pro": {
|
|
8187
|
+
id: string;
|
|
8188
|
+
name: string;
|
|
8189
|
+
api: "openai-completions";
|
|
8190
|
+
provider: string;
|
|
8191
|
+
baseUrl: string;
|
|
8192
|
+
reasoning: true;
|
|
8193
|
+
input: "text"[];
|
|
8194
|
+
cost: {
|
|
8195
|
+
input: number;
|
|
8196
|
+
output: number;
|
|
8197
|
+
cacheRead: number;
|
|
8198
|
+
cacheWrite: number;
|
|
8199
|
+
};
|
|
8200
|
+
contextWindow: number;
|
|
8201
|
+
maxTokens: number;
|
|
8202
|
+
};
|
|
7817
8203
|
readonly "essentialai/rnj-1-instruct": {
|
|
7818
8204
|
id: string;
|
|
7819
8205
|
name: string;
|
|
@@ -8035,13 +8421,13 @@ export declare const MODELS: {
|
|
|
8035
8421
|
contextWindow: number;
|
|
8036
8422
|
maxTokens: number;
|
|
8037
8423
|
};
|
|
8038
|
-
readonly "google/gemma-
|
|
8424
|
+
readonly "google/gemma-3-12b-it": {
|
|
8039
8425
|
id: string;
|
|
8040
8426
|
name: string;
|
|
8041
8427
|
api: "openai-completions";
|
|
8042
8428
|
provider: string;
|
|
8043
8429
|
baseUrl: string;
|
|
8044
|
-
reasoning:
|
|
8430
|
+
reasoning: false;
|
|
8045
8431
|
input: ("image" | "text")[];
|
|
8046
8432
|
cost: {
|
|
8047
8433
|
input: number;
|
|
@@ -8052,13 +8438,13 @@ export declare const MODELS: {
|
|
|
8052
8438
|
contextWindow: number;
|
|
8053
8439
|
maxTokens: number;
|
|
8054
8440
|
};
|
|
8055
|
-
readonly "google/gemma-
|
|
8441
|
+
readonly "google/gemma-3-27b-it": {
|
|
8056
8442
|
id: string;
|
|
8057
8443
|
name: string;
|
|
8058
8444
|
api: "openai-completions";
|
|
8059
8445
|
provider: string;
|
|
8060
8446
|
baseUrl: string;
|
|
8061
|
-
reasoning:
|
|
8447
|
+
reasoning: false;
|
|
8062
8448
|
input: ("image" | "text")[];
|
|
8063
8449
|
cost: {
|
|
8064
8450
|
input: number;
|
|
@@ -8069,7 +8455,7 @@ export declare const MODELS: {
|
|
|
8069
8455
|
contextWindow: number;
|
|
8070
8456
|
maxTokens: number;
|
|
8071
8457
|
};
|
|
8072
|
-
readonly "google/gemma-4-
|
|
8458
|
+
readonly "google/gemma-4-26b-a4b-it": {
|
|
8073
8459
|
id: string;
|
|
8074
8460
|
name: string;
|
|
8075
8461
|
api: "openai-completions";
|
|
@@ -8086,7 +8472,7 @@ export declare const MODELS: {
|
|
|
8086
8472
|
contextWindow: number;
|
|
8087
8473
|
maxTokens: number;
|
|
8088
8474
|
};
|
|
8089
|
-
readonly "google/gemma-4-
|
|
8475
|
+
readonly "google/gemma-4-26b-a4b-it:free": {
|
|
8090
8476
|
id: string;
|
|
8091
8477
|
name: string;
|
|
8092
8478
|
api: "openai-completions";
|
|
@@ -8103,14 +8489,14 @@ export declare const MODELS: {
|
|
|
8103
8489
|
contextWindow: number;
|
|
8104
8490
|
maxTokens: number;
|
|
8105
8491
|
};
|
|
8106
|
-
readonly "
|
|
8492
|
+
readonly "google/gemma-4-31b-it": {
|
|
8107
8493
|
id: string;
|
|
8108
8494
|
name: string;
|
|
8109
8495
|
api: "openai-completions";
|
|
8110
8496
|
provider: string;
|
|
8111
8497
|
baseUrl: string;
|
|
8112
8498
|
reasoning: true;
|
|
8113
|
-
input: "text"[];
|
|
8499
|
+
input: ("image" | "text")[];
|
|
8114
8500
|
cost: {
|
|
8115
8501
|
input: number;
|
|
8116
8502
|
output: number;
|
|
@@ -8120,14 +8506,14 @@ export declare const MODELS: {
|
|
|
8120
8506
|
contextWindow: number;
|
|
8121
8507
|
maxTokens: number;
|
|
8122
8508
|
};
|
|
8123
|
-
readonly "
|
|
8509
|
+
readonly "google/gemma-4-31b-it:free": {
|
|
8124
8510
|
id: string;
|
|
8125
8511
|
name: string;
|
|
8126
8512
|
api: "openai-completions";
|
|
8127
8513
|
provider: string;
|
|
8128
8514
|
baseUrl: string;
|
|
8129
|
-
reasoning:
|
|
8130
|
-
input: "text"[];
|
|
8515
|
+
reasoning: true;
|
|
8516
|
+
input: ("image" | "text")[];
|
|
8131
8517
|
cost: {
|
|
8132
8518
|
input: number;
|
|
8133
8519
|
output: number;
|
|
@@ -8137,7 +8523,7 @@ export declare const MODELS: {
|
|
|
8137
8523
|
contextWindow: number;
|
|
8138
8524
|
maxTokens: number;
|
|
8139
8525
|
};
|
|
8140
|
-
readonly "
|
|
8526
|
+
readonly "ibm-granite/granite-4.1-8b": {
|
|
8141
8527
|
id: string;
|
|
8142
8528
|
name: string;
|
|
8143
8529
|
api: "openai-completions";
|
|
@@ -8154,13 +8540,13 @@ export declare const MODELS: {
|
|
|
8154
8540
|
contextWindow: number;
|
|
8155
8541
|
maxTokens: number;
|
|
8156
8542
|
};
|
|
8157
|
-
readonly "
|
|
8543
|
+
readonly "inception/mercury-2": {
|
|
8158
8544
|
id: string;
|
|
8159
8545
|
name: string;
|
|
8160
8546
|
api: "openai-completions";
|
|
8161
8547
|
provider: string;
|
|
8162
8548
|
baseUrl: string;
|
|
8163
|
-
reasoning:
|
|
8549
|
+
reasoning: true;
|
|
8164
8550
|
input: "text"[];
|
|
8165
8551
|
cost: {
|
|
8166
8552
|
input: number;
|
|
@@ -8171,7 +8557,7 @@ export declare const MODELS: {
|
|
|
8171
8557
|
contextWindow: number;
|
|
8172
8558
|
maxTokens: number;
|
|
8173
8559
|
};
|
|
8174
|
-
readonly "
|
|
8560
|
+
readonly "inclusionai/ling-2.6-1t:free": {
|
|
8175
8561
|
id: string;
|
|
8176
8562
|
name: string;
|
|
8177
8563
|
api: "openai-completions";
|
|
@@ -8188,7 +8574,7 @@ export declare const MODELS: {
|
|
|
8188
8574
|
contextWindow: number;
|
|
8189
8575
|
maxTokens: number;
|
|
8190
8576
|
};
|
|
8191
|
-
readonly "
|
|
8577
|
+
readonly "inclusionai/ling-2.6-flash": {
|
|
8192
8578
|
id: string;
|
|
8193
8579
|
name: string;
|
|
8194
8580
|
api: "openai-completions";
|
|
@@ -8205,7 +8591,7 @@ export declare const MODELS: {
|
|
|
8205
8591
|
contextWindow: number;
|
|
8206
8592
|
maxTokens: number;
|
|
8207
8593
|
};
|
|
8208
|
-
readonly "
|
|
8594
|
+
readonly "kwaipilot/kat-coder-pro-v2": {
|
|
8209
8595
|
id: string;
|
|
8210
8596
|
name: string;
|
|
8211
8597
|
api: "openai-completions";
|
|
@@ -8222,14 +8608,14 @@ export declare const MODELS: {
|
|
|
8222
8608
|
contextWindow: number;
|
|
8223
8609
|
maxTokens: number;
|
|
8224
8610
|
};
|
|
8225
|
-
readonly "meta-llama/llama-
|
|
8611
|
+
readonly "meta-llama/llama-3-8b-instruct": {
|
|
8226
8612
|
id: string;
|
|
8227
8613
|
name: string;
|
|
8228
8614
|
api: "openai-completions";
|
|
8229
8615
|
provider: string;
|
|
8230
8616
|
baseUrl: string;
|
|
8231
8617
|
reasoning: false;
|
|
8232
|
-
input:
|
|
8618
|
+
input: "text"[];
|
|
8233
8619
|
cost: {
|
|
8234
8620
|
input: number;
|
|
8235
8621
|
output: number;
|
|
@@ -8239,13 +8625,13 @@ export declare const MODELS: {
|
|
|
8239
8625
|
contextWindow: number;
|
|
8240
8626
|
maxTokens: number;
|
|
8241
8627
|
};
|
|
8242
|
-
readonly "
|
|
8628
|
+
readonly "meta-llama/llama-3.1-70b-instruct": {
|
|
8243
8629
|
id: string;
|
|
8244
8630
|
name: string;
|
|
8245
8631
|
api: "openai-completions";
|
|
8246
8632
|
provider: string;
|
|
8247
8633
|
baseUrl: string;
|
|
8248
|
-
reasoning:
|
|
8634
|
+
reasoning: false;
|
|
8249
8635
|
input: "text"[];
|
|
8250
8636
|
cost: {
|
|
8251
8637
|
input: number;
|
|
@@ -8256,13 +8642,13 @@ export declare const MODELS: {
|
|
|
8256
8642
|
contextWindow: number;
|
|
8257
8643
|
maxTokens: number;
|
|
8258
8644
|
};
|
|
8259
|
-
readonly "
|
|
8645
|
+
readonly "meta-llama/llama-3.1-8b-instruct": {
|
|
8260
8646
|
id: string;
|
|
8261
8647
|
name: string;
|
|
8262
8648
|
api: "openai-completions";
|
|
8263
8649
|
provider: string;
|
|
8264
8650
|
baseUrl: string;
|
|
8265
|
-
reasoning:
|
|
8651
|
+
reasoning: false;
|
|
8266
8652
|
input: "text"[];
|
|
8267
8653
|
cost: {
|
|
8268
8654
|
input: number;
|
|
@@ -8273,13 +8659,13 @@ export declare const MODELS: {
|
|
|
8273
8659
|
contextWindow: number;
|
|
8274
8660
|
maxTokens: number;
|
|
8275
8661
|
};
|
|
8276
|
-
readonly "
|
|
8662
|
+
readonly "meta-llama/llama-3.3-70b-instruct": {
|
|
8277
8663
|
id: string;
|
|
8278
8664
|
name: string;
|
|
8279
8665
|
api: "openai-completions";
|
|
8280
8666
|
provider: string;
|
|
8281
8667
|
baseUrl: string;
|
|
8282
|
-
reasoning:
|
|
8668
|
+
reasoning: false;
|
|
8283
8669
|
input: "text"[];
|
|
8284
8670
|
cost: {
|
|
8285
8671
|
input: number;
|
|
@@ -8290,13 +8676,13 @@ export declare const MODELS: {
|
|
|
8290
8676
|
contextWindow: number;
|
|
8291
8677
|
maxTokens: number;
|
|
8292
8678
|
};
|
|
8293
|
-
readonly "
|
|
8679
|
+
readonly "meta-llama/llama-3.3-70b-instruct:free": {
|
|
8294
8680
|
id: string;
|
|
8295
8681
|
name: string;
|
|
8296
8682
|
api: "openai-completions";
|
|
8297
8683
|
provider: string;
|
|
8298
8684
|
baseUrl: string;
|
|
8299
|
-
reasoning:
|
|
8685
|
+
reasoning: false;
|
|
8300
8686
|
input: "text"[];
|
|
8301
8687
|
cost: {
|
|
8302
8688
|
input: number;
|
|
@@ -8307,15 +8693,100 @@ export declare const MODELS: {
|
|
|
8307
8693
|
contextWindow: number;
|
|
8308
8694
|
maxTokens: number;
|
|
8309
8695
|
};
|
|
8310
|
-
readonly "
|
|
8696
|
+
readonly "meta-llama/llama-4-scout": {
|
|
8311
8697
|
id: string;
|
|
8312
8698
|
name: string;
|
|
8313
8699
|
api: "openai-completions";
|
|
8314
8700
|
provider: string;
|
|
8315
8701
|
baseUrl: string;
|
|
8316
|
-
reasoning:
|
|
8317
|
-
input: "text"[];
|
|
8318
|
-
cost: {
|
|
8702
|
+
reasoning: false;
|
|
8703
|
+
input: ("image" | "text")[];
|
|
8704
|
+
cost: {
|
|
8705
|
+
input: number;
|
|
8706
|
+
output: number;
|
|
8707
|
+
cacheRead: number;
|
|
8708
|
+
cacheWrite: number;
|
|
8709
|
+
};
|
|
8710
|
+
contextWindow: number;
|
|
8711
|
+
maxTokens: number;
|
|
8712
|
+
};
|
|
8713
|
+
readonly "minimax/minimax-m1": {
|
|
8714
|
+
id: string;
|
|
8715
|
+
name: string;
|
|
8716
|
+
api: "openai-completions";
|
|
8717
|
+
provider: string;
|
|
8718
|
+
baseUrl: string;
|
|
8719
|
+
reasoning: true;
|
|
8720
|
+
input: "text"[];
|
|
8721
|
+
cost: {
|
|
8722
|
+
input: number;
|
|
8723
|
+
output: number;
|
|
8724
|
+
cacheRead: number;
|
|
8725
|
+
cacheWrite: number;
|
|
8726
|
+
};
|
|
8727
|
+
contextWindow: number;
|
|
8728
|
+
maxTokens: number;
|
|
8729
|
+
};
|
|
8730
|
+
readonly "minimax/minimax-m2": {
|
|
8731
|
+
id: string;
|
|
8732
|
+
name: string;
|
|
8733
|
+
api: "openai-completions";
|
|
8734
|
+
provider: string;
|
|
8735
|
+
baseUrl: string;
|
|
8736
|
+
reasoning: true;
|
|
8737
|
+
input: "text"[];
|
|
8738
|
+
cost: {
|
|
8739
|
+
input: number;
|
|
8740
|
+
output: number;
|
|
8741
|
+
cacheRead: number;
|
|
8742
|
+
cacheWrite: number;
|
|
8743
|
+
};
|
|
8744
|
+
contextWindow: number;
|
|
8745
|
+
maxTokens: number;
|
|
8746
|
+
};
|
|
8747
|
+
readonly "minimax/minimax-m2.1": {
|
|
8748
|
+
id: string;
|
|
8749
|
+
name: string;
|
|
8750
|
+
api: "openai-completions";
|
|
8751
|
+
provider: string;
|
|
8752
|
+
baseUrl: string;
|
|
8753
|
+
reasoning: true;
|
|
8754
|
+
input: "text"[];
|
|
8755
|
+
cost: {
|
|
8756
|
+
input: number;
|
|
8757
|
+
output: number;
|
|
8758
|
+
cacheRead: number;
|
|
8759
|
+
cacheWrite: number;
|
|
8760
|
+
};
|
|
8761
|
+
contextWindow: number;
|
|
8762
|
+
maxTokens: number;
|
|
8763
|
+
};
|
|
8764
|
+
readonly "minimax/minimax-m2.5": {
|
|
8765
|
+
id: string;
|
|
8766
|
+
name: string;
|
|
8767
|
+
api: "openai-completions";
|
|
8768
|
+
provider: string;
|
|
8769
|
+
baseUrl: string;
|
|
8770
|
+
reasoning: true;
|
|
8771
|
+
input: "text"[];
|
|
8772
|
+
cost: {
|
|
8773
|
+
input: number;
|
|
8774
|
+
output: number;
|
|
8775
|
+
cacheRead: number;
|
|
8776
|
+
cacheWrite: number;
|
|
8777
|
+
};
|
|
8778
|
+
contextWindow: number;
|
|
8779
|
+
maxTokens: number;
|
|
8780
|
+
};
|
|
8781
|
+
readonly "minimax/minimax-m2.5:free": {
|
|
8782
|
+
id: string;
|
|
8783
|
+
name: string;
|
|
8784
|
+
api: "openai-completions";
|
|
8785
|
+
provider: string;
|
|
8786
|
+
baseUrl: string;
|
|
8787
|
+
reasoning: true;
|
|
8788
|
+
input: "text"[];
|
|
8789
|
+
cost: {
|
|
8319
8790
|
input: number;
|
|
8320
8791
|
output: number;
|
|
8321
8792
|
cacheRead: number;
|
|
@@ -8630,7 +9101,7 @@ export declare const MODELS: {
|
|
|
8630
9101
|
contextWindow: number;
|
|
8631
9102
|
maxTokens: number;
|
|
8632
9103
|
};
|
|
8633
|
-
readonly "mistralai/
|
|
9104
|
+
readonly "mistralai/mixtral-8x22b-instruct": {
|
|
8634
9105
|
id: string;
|
|
8635
9106
|
name: string;
|
|
8636
9107
|
api: "openai-completions";
|
|
@@ -8647,7 +9118,7 @@ export declare const MODELS: {
|
|
|
8647
9118
|
contextWindow: number;
|
|
8648
9119
|
maxTokens: number;
|
|
8649
9120
|
};
|
|
8650
|
-
readonly "mistralai/mixtral-
|
|
9121
|
+
readonly "mistralai/mixtral-8x7b-instruct": {
|
|
8651
9122
|
id: string;
|
|
8652
9123
|
name: string;
|
|
8653
9124
|
api: "openai-completions";
|
|
@@ -8664,14 +9135,14 @@ export declare const MODELS: {
|
|
|
8664
9135
|
contextWindow: number;
|
|
8665
9136
|
maxTokens: number;
|
|
8666
9137
|
};
|
|
8667
|
-
readonly "mistralai/
|
|
9138
|
+
readonly "mistralai/pixtral-large-2411": {
|
|
8668
9139
|
id: string;
|
|
8669
9140
|
name: string;
|
|
8670
9141
|
api: "openai-completions";
|
|
8671
9142
|
provider: string;
|
|
8672
9143
|
baseUrl: string;
|
|
8673
9144
|
reasoning: false;
|
|
8674
|
-
input: "text"[];
|
|
9145
|
+
input: ("image" | "text")[];
|
|
8675
9146
|
cost: {
|
|
8676
9147
|
input: number;
|
|
8677
9148
|
output: number;
|
|
@@ -8681,14 +9152,14 @@ export declare const MODELS: {
|
|
|
8681
9152
|
contextWindow: number;
|
|
8682
9153
|
maxTokens: number;
|
|
8683
9154
|
};
|
|
8684
|
-
readonly "mistralai/
|
|
9155
|
+
readonly "mistralai/voxtral-small-24b-2507": {
|
|
8685
9156
|
id: string;
|
|
8686
9157
|
name: string;
|
|
8687
9158
|
api: "openai-completions";
|
|
8688
9159
|
provider: string;
|
|
8689
9160
|
baseUrl: string;
|
|
8690
9161
|
reasoning: false;
|
|
8691
|
-
input:
|
|
9162
|
+
input: "text"[];
|
|
8692
9163
|
cost: {
|
|
8693
9164
|
input: number;
|
|
8694
9165
|
output: number;
|
|
@@ -8698,7 +9169,7 @@ export declare const MODELS: {
|
|
|
8698
9169
|
contextWindow: number;
|
|
8699
9170
|
maxTokens: number;
|
|
8700
9171
|
};
|
|
8701
|
-
readonly "
|
|
9172
|
+
readonly "moonshotai/kimi-k2": {
|
|
8702
9173
|
id: string;
|
|
8703
9174
|
name: string;
|
|
8704
9175
|
api: "openai-completions";
|
|
@@ -8715,7 +9186,7 @@ export declare const MODELS: {
|
|
|
8715
9186
|
contextWindow: number;
|
|
8716
9187
|
maxTokens: number;
|
|
8717
9188
|
};
|
|
8718
|
-
readonly "moonshotai/kimi-k2": {
|
|
9189
|
+
readonly "moonshotai/kimi-k2-0905": {
|
|
8719
9190
|
id: string;
|
|
8720
9191
|
name: string;
|
|
8721
9192
|
api: "openai-completions";
|
|
@@ -8732,13 +9203,13 @@ export declare const MODELS: {
|
|
|
8732
9203
|
contextWindow: number;
|
|
8733
9204
|
maxTokens: number;
|
|
8734
9205
|
};
|
|
8735
|
-
readonly "moonshotai/kimi-k2-
|
|
9206
|
+
readonly "moonshotai/kimi-k2-thinking": {
|
|
8736
9207
|
id: string;
|
|
8737
9208
|
name: string;
|
|
8738
9209
|
api: "openai-completions";
|
|
8739
9210
|
provider: string;
|
|
8740
9211
|
baseUrl: string;
|
|
8741
|
-
reasoning:
|
|
9212
|
+
reasoning: true;
|
|
8742
9213
|
input: "text"[];
|
|
8743
9214
|
cost: {
|
|
8744
9215
|
input: number;
|
|
@@ -8749,14 +9220,14 @@ export declare const MODELS: {
|
|
|
8749
9220
|
contextWindow: number;
|
|
8750
9221
|
maxTokens: number;
|
|
8751
9222
|
};
|
|
8752
|
-
readonly "moonshotai/kimi-k2
|
|
9223
|
+
readonly "moonshotai/kimi-k2.5": {
|
|
8753
9224
|
id: string;
|
|
8754
9225
|
name: string;
|
|
8755
9226
|
api: "openai-completions";
|
|
8756
9227
|
provider: string;
|
|
8757
9228
|
baseUrl: string;
|
|
8758
9229
|
reasoning: true;
|
|
8759
|
-
input: "text"[];
|
|
9230
|
+
input: ("image" | "text")[];
|
|
8760
9231
|
cost: {
|
|
8761
9232
|
input: number;
|
|
8762
9233
|
output: number;
|
|
@@ -8766,7 +9237,7 @@ export declare const MODELS: {
|
|
|
8766
9237
|
contextWindow: number;
|
|
8767
9238
|
maxTokens: number;
|
|
8768
9239
|
};
|
|
8769
|
-
readonly "moonshotai/kimi-k2.
|
|
9240
|
+
readonly "moonshotai/kimi-k2.6": {
|
|
8770
9241
|
id: string;
|
|
8771
9242
|
name: string;
|
|
8772
9243
|
api: "openai-completions";
|
|
@@ -8868,6 +9339,23 @@ export declare const MODELS: {
|
|
|
8868
9339
|
contextWindow: number;
|
|
8869
9340
|
maxTokens: number;
|
|
8870
9341
|
};
|
|
9342
|
+
readonly "nvidia/nemotron-3-nano-omni-30b-a3b-reasoning:free": {
|
|
9343
|
+
id: string;
|
|
9344
|
+
name: string;
|
|
9345
|
+
api: "openai-completions";
|
|
9346
|
+
provider: string;
|
|
9347
|
+
baseUrl: string;
|
|
9348
|
+
reasoning: true;
|
|
9349
|
+
input: ("image" | "text")[];
|
|
9350
|
+
cost: {
|
|
9351
|
+
input: number;
|
|
9352
|
+
output: number;
|
|
9353
|
+
cacheRead: number;
|
|
9354
|
+
cacheWrite: number;
|
|
9355
|
+
};
|
|
9356
|
+
contextWindow: number;
|
|
9357
|
+
maxTokens: number;
|
|
9358
|
+
};
|
|
8871
9359
|
readonly "nvidia/nemotron-3-super-120b-a12b": {
|
|
8872
9360
|
id: string;
|
|
8873
9361
|
name: string;
|
|
@@ -9293,7 +9781,7 @@ export declare const MODELS: {
|
|
|
9293
9781
|
contextWindow: number;
|
|
9294
9782
|
maxTokens: number;
|
|
9295
9783
|
};
|
|
9296
|
-
readonly "openai/gpt-5-
|
|
9784
|
+
readonly "openai/gpt-5-mini": {
|
|
9297
9785
|
id: string;
|
|
9298
9786
|
name: string;
|
|
9299
9787
|
api: "openai-completions";
|
|
@@ -9310,7 +9798,7 @@ export declare const MODELS: {
|
|
|
9310
9798
|
contextWindow: number;
|
|
9311
9799
|
maxTokens: number;
|
|
9312
9800
|
};
|
|
9313
|
-
readonly "openai/gpt-5-
|
|
9801
|
+
readonly "openai/gpt-5-nano": {
|
|
9314
9802
|
id: string;
|
|
9315
9803
|
name: string;
|
|
9316
9804
|
api: "openai-completions";
|
|
@@ -9327,7 +9815,7 @@ export declare const MODELS: {
|
|
|
9327
9815
|
contextWindow: number;
|
|
9328
9816
|
maxTokens: number;
|
|
9329
9817
|
};
|
|
9330
|
-
readonly "openai/gpt-5-
|
|
9818
|
+
readonly "openai/gpt-5-pro": {
|
|
9331
9819
|
id: string;
|
|
9332
9820
|
name: string;
|
|
9333
9821
|
api: "openai-completions";
|
|
@@ -9344,7 +9832,7 @@ export declare const MODELS: {
|
|
|
9344
9832
|
contextWindow: number;
|
|
9345
9833
|
maxTokens: number;
|
|
9346
9834
|
};
|
|
9347
|
-
readonly "openai/gpt-5
|
|
9835
|
+
readonly "openai/gpt-5.1": {
|
|
9348
9836
|
id: string;
|
|
9349
9837
|
name: string;
|
|
9350
9838
|
api: "openai-completions";
|
|
@@ -9361,13 +9849,13 @@ export declare const MODELS: {
|
|
|
9361
9849
|
contextWindow: number;
|
|
9362
9850
|
maxTokens: number;
|
|
9363
9851
|
};
|
|
9364
|
-
readonly "openai/gpt-5-
|
|
9852
|
+
readonly "openai/gpt-5.1-chat": {
|
|
9365
9853
|
id: string;
|
|
9366
9854
|
name: string;
|
|
9367
9855
|
api: "openai-completions";
|
|
9368
9856
|
provider: string;
|
|
9369
9857
|
baseUrl: string;
|
|
9370
|
-
reasoning:
|
|
9858
|
+
reasoning: false;
|
|
9371
9859
|
input: ("image" | "text")[];
|
|
9372
9860
|
cost: {
|
|
9373
9861
|
input: number;
|
|
@@ -9378,7 +9866,7 @@ export declare const MODELS: {
|
|
|
9378
9866
|
contextWindow: number;
|
|
9379
9867
|
maxTokens: number;
|
|
9380
9868
|
};
|
|
9381
|
-
readonly "openai/gpt-5.1": {
|
|
9869
|
+
readonly "openai/gpt-5.1-codex": {
|
|
9382
9870
|
id: string;
|
|
9383
9871
|
name: string;
|
|
9384
9872
|
api: "openai-completions";
|
|
@@ -9395,13 +9883,13 @@ export declare const MODELS: {
|
|
|
9395
9883
|
contextWindow: number;
|
|
9396
9884
|
maxTokens: number;
|
|
9397
9885
|
};
|
|
9398
|
-
readonly "openai/gpt-5.1-
|
|
9886
|
+
readonly "openai/gpt-5.1-codex-max": {
|
|
9399
9887
|
id: string;
|
|
9400
9888
|
name: string;
|
|
9401
9889
|
api: "openai-completions";
|
|
9402
9890
|
provider: string;
|
|
9403
9891
|
baseUrl: string;
|
|
9404
|
-
reasoning:
|
|
9892
|
+
reasoning: true;
|
|
9405
9893
|
input: ("image" | "text")[];
|
|
9406
9894
|
cost: {
|
|
9407
9895
|
input: number;
|
|
@@ -9412,7 +9900,7 @@ export declare const MODELS: {
|
|
|
9412
9900
|
contextWindow: number;
|
|
9413
9901
|
maxTokens: number;
|
|
9414
9902
|
};
|
|
9415
|
-
readonly "openai/gpt-5.1-codex": {
|
|
9903
|
+
readonly "openai/gpt-5.1-codex-mini": {
|
|
9416
9904
|
id: string;
|
|
9417
9905
|
name: string;
|
|
9418
9906
|
api: "openai-completions";
|
|
@@ -9429,7 +9917,7 @@ export declare const MODELS: {
|
|
|
9429
9917
|
contextWindow: number;
|
|
9430
9918
|
maxTokens: number;
|
|
9431
9919
|
};
|
|
9432
|
-
readonly "openai/gpt-5.
|
|
9920
|
+
readonly "openai/gpt-5.2": {
|
|
9433
9921
|
id: string;
|
|
9434
9922
|
name: string;
|
|
9435
9923
|
api: "openai-completions";
|
|
@@ -9446,13 +9934,13 @@ export declare const MODELS: {
|
|
|
9446
9934
|
contextWindow: number;
|
|
9447
9935
|
maxTokens: number;
|
|
9448
9936
|
};
|
|
9449
|
-
readonly "openai/gpt-5.
|
|
9937
|
+
readonly "openai/gpt-5.2-chat": {
|
|
9450
9938
|
id: string;
|
|
9451
9939
|
name: string;
|
|
9452
9940
|
api: "openai-completions";
|
|
9453
9941
|
provider: string;
|
|
9454
9942
|
baseUrl: string;
|
|
9455
|
-
reasoning:
|
|
9943
|
+
reasoning: false;
|
|
9456
9944
|
input: ("image" | "text")[];
|
|
9457
9945
|
cost: {
|
|
9458
9946
|
input: number;
|
|
@@ -9463,7 +9951,7 @@ export declare const MODELS: {
|
|
|
9463
9951
|
contextWindow: number;
|
|
9464
9952
|
maxTokens: number;
|
|
9465
9953
|
};
|
|
9466
|
-
readonly "openai/gpt-5.2": {
|
|
9954
|
+
readonly "openai/gpt-5.2-codex": {
|
|
9467
9955
|
id: string;
|
|
9468
9956
|
name: string;
|
|
9469
9957
|
api: "openai-completions";
|
|
@@ -9480,13 +9968,13 @@ export declare const MODELS: {
|
|
|
9480
9968
|
contextWindow: number;
|
|
9481
9969
|
maxTokens: number;
|
|
9482
9970
|
};
|
|
9483
|
-
readonly "openai/gpt-5.2-
|
|
9971
|
+
readonly "openai/gpt-5.2-pro": {
|
|
9484
9972
|
id: string;
|
|
9485
9973
|
name: string;
|
|
9486
9974
|
api: "openai-completions";
|
|
9487
9975
|
provider: string;
|
|
9488
9976
|
baseUrl: string;
|
|
9489
|
-
reasoning:
|
|
9977
|
+
reasoning: true;
|
|
9490
9978
|
input: ("image" | "text")[];
|
|
9491
9979
|
cost: {
|
|
9492
9980
|
input: number;
|
|
@@ -9497,13 +9985,13 @@ export declare const MODELS: {
|
|
|
9497
9985
|
contextWindow: number;
|
|
9498
9986
|
maxTokens: number;
|
|
9499
9987
|
};
|
|
9500
|
-
readonly "openai/gpt-5.
|
|
9988
|
+
readonly "openai/gpt-5.3-chat": {
|
|
9501
9989
|
id: string;
|
|
9502
9990
|
name: string;
|
|
9503
9991
|
api: "openai-completions";
|
|
9504
9992
|
provider: string;
|
|
9505
9993
|
baseUrl: string;
|
|
9506
|
-
reasoning:
|
|
9994
|
+
reasoning: false;
|
|
9507
9995
|
input: ("image" | "text")[];
|
|
9508
9996
|
cost: {
|
|
9509
9997
|
input: number;
|
|
@@ -9514,7 +10002,7 @@ export declare const MODELS: {
|
|
|
9514
10002
|
contextWindow: number;
|
|
9515
10003
|
maxTokens: number;
|
|
9516
10004
|
};
|
|
9517
|
-
readonly "openai/gpt-5.
|
|
10005
|
+
readonly "openai/gpt-5.3-codex": {
|
|
9518
10006
|
id: string;
|
|
9519
10007
|
name: string;
|
|
9520
10008
|
api: "openai-completions";
|
|
@@ -9531,13 +10019,13 @@ export declare const MODELS: {
|
|
|
9531
10019
|
contextWindow: number;
|
|
9532
10020
|
maxTokens: number;
|
|
9533
10021
|
};
|
|
9534
|
-
readonly "openai/gpt-5.
|
|
10022
|
+
readonly "openai/gpt-5.4": {
|
|
9535
10023
|
id: string;
|
|
9536
10024
|
name: string;
|
|
9537
10025
|
api: "openai-completions";
|
|
9538
10026
|
provider: string;
|
|
9539
10027
|
baseUrl: string;
|
|
9540
|
-
reasoning:
|
|
10028
|
+
reasoning: true;
|
|
9541
10029
|
input: ("image" | "text")[];
|
|
9542
10030
|
cost: {
|
|
9543
10031
|
input: number;
|
|
@@ -9548,7 +10036,7 @@ export declare const MODELS: {
|
|
|
9548
10036
|
contextWindow: number;
|
|
9549
10037
|
maxTokens: number;
|
|
9550
10038
|
};
|
|
9551
|
-
readonly "openai/gpt-5.
|
|
10039
|
+
readonly "openai/gpt-5.4-mini": {
|
|
9552
10040
|
id: string;
|
|
9553
10041
|
name: string;
|
|
9554
10042
|
api: "openai-completions";
|
|
@@ -9565,7 +10053,7 @@ export declare const MODELS: {
|
|
|
9565
10053
|
contextWindow: number;
|
|
9566
10054
|
maxTokens: number;
|
|
9567
10055
|
};
|
|
9568
|
-
readonly "openai/gpt-5.4": {
|
|
10056
|
+
readonly "openai/gpt-5.4-nano": {
|
|
9569
10057
|
id: string;
|
|
9570
10058
|
name: string;
|
|
9571
10059
|
api: "openai-completions";
|
|
@@ -9582,7 +10070,7 @@ export declare const MODELS: {
|
|
|
9582
10070
|
contextWindow: number;
|
|
9583
10071
|
maxTokens: number;
|
|
9584
10072
|
};
|
|
9585
|
-
readonly "openai/gpt-5.4-
|
|
10073
|
+
readonly "openai/gpt-5.4-pro": {
|
|
9586
10074
|
id: string;
|
|
9587
10075
|
name: string;
|
|
9588
10076
|
api: "openai-completions";
|
|
@@ -9599,7 +10087,7 @@ export declare const MODELS: {
|
|
|
9599
10087
|
contextWindow: number;
|
|
9600
10088
|
maxTokens: number;
|
|
9601
10089
|
};
|
|
9602
|
-
readonly "openai/gpt-5.
|
|
10090
|
+
readonly "openai/gpt-5.5": {
|
|
9603
10091
|
id: string;
|
|
9604
10092
|
name: string;
|
|
9605
10093
|
api: "openai-completions";
|
|
@@ -9616,7 +10104,7 @@ export declare const MODELS: {
|
|
|
9616
10104
|
contextWindow: number;
|
|
9617
10105
|
maxTokens: number;
|
|
9618
10106
|
};
|
|
9619
|
-
readonly "openai/gpt-5.
|
|
10107
|
+
readonly "openai/gpt-5.5-pro": {
|
|
9620
10108
|
id: string;
|
|
9621
10109
|
name: string;
|
|
9622
10110
|
api: "openai-completions";
|
|
@@ -9922,7 +10410,24 @@ export declare const MODELS: {
|
|
|
9922
10410
|
contextWindow: number;
|
|
9923
10411
|
maxTokens: number;
|
|
9924
10412
|
};
|
|
9925
|
-
readonly "openrouter/
|
|
10413
|
+
readonly "openrouter/free": {
|
|
10414
|
+
id: string;
|
|
10415
|
+
name: string;
|
|
10416
|
+
api: "openai-completions";
|
|
10417
|
+
provider: string;
|
|
10418
|
+
baseUrl: string;
|
|
10419
|
+
reasoning: true;
|
|
10420
|
+
input: ("image" | "text")[];
|
|
10421
|
+
cost: {
|
|
10422
|
+
input: number;
|
|
10423
|
+
output: number;
|
|
10424
|
+
cacheRead: number;
|
|
10425
|
+
cacheWrite: number;
|
|
10426
|
+
};
|
|
10427
|
+
contextWindow: number;
|
|
10428
|
+
maxTokens: number;
|
|
10429
|
+
};
|
|
10430
|
+
readonly "openrouter/owl-alpha": {
|
|
9926
10431
|
id: string;
|
|
9927
10432
|
name: string;
|
|
9928
10433
|
api: "openai-completions";
|
|
@@ -9939,14 +10444,31 @@ export declare const MODELS: {
|
|
|
9939
10444
|
contextWindow: number;
|
|
9940
10445
|
maxTokens: number;
|
|
9941
10446
|
};
|
|
9942
|
-
readonly "
|
|
10447
|
+
readonly "poolside/laguna-m.1:free": {
|
|
9943
10448
|
id: string;
|
|
9944
10449
|
name: string;
|
|
9945
10450
|
api: "openai-completions";
|
|
9946
10451
|
provider: string;
|
|
9947
10452
|
baseUrl: string;
|
|
9948
10453
|
reasoning: true;
|
|
9949
|
-
input:
|
|
10454
|
+
input: "text"[];
|
|
10455
|
+
cost: {
|
|
10456
|
+
input: number;
|
|
10457
|
+
output: number;
|
|
10458
|
+
cacheRead: number;
|
|
10459
|
+
cacheWrite: number;
|
|
10460
|
+
};
|
|
10461
|
+
contextWindow: number;
|
|
10462
|
+
maxTokens: number;
|
|
10463
|
+
};
|
|
10464
|
+
readonly "poolside/laguna-xs.2:free": {
|
|
10465
|
+
id: string;
|
|
10466
|
+
name: string;
|
|
10467
|
+
api: "openai-completions";
|
|
10468
|
+
provider: string;
|
|
10469
|
+
baseUrl: string;
|
|
10470
|
+
reasoning: true;
|
|
10471
|
+
input: "text"[];
|
|
9950
10472
|
cost: {
|
|
9951
10473
|
input: number;
|
|
9952
10474
|
output: number;
|
|
@@ -10625,8 +11147,263 @@ export declare const MODELS: {
|
|
|
10625
11147
|
api: "openai-completions";
|
|
10626
11148
|
provider: string;
|
|
10627
11149
|
baseUrl: string;
|
|
10628
|
-
reasoning: true;
|
|
10629
|
-
input: ("image" | "text")[];
|
|
11150
|
+
reasoning: true;
|
|
11151
|
+
input: ("image" | "text")[];
|
|
11152
|
+
cost: {
|
|
11153
|
+
input: number;
|
|
11154
|
+
output: number;
|
|
11155
|
+
cacheRead: number;
|
|
11156
|
+
cacheWrite: number;
|
|
11157
|
+
};
|
|
11158
|
+
contextWindow: number;
|
|
11159
|
+
maxTokens: number;
|
|
11160
|
+
};
|
|
11161
|
+
readonly "qwen/qwen3.5-9b": {
|
|
11162
|
+
id: string;
|
|
11163
|
+
name: string;
|
|
11164
|
+
api: "openai-completions";
|
|
11165
|
+
provider: string;
|
|
11166
|
+
baseUrl: string;
|
|
11167
|
+
reasoning: true;
|
|
11168
|
+
input: ("image" | "text")[];
|
|
11169
|
+
cost: {
|
|
11170
|
+
input: number;
|
|
11171
|
+
output: number;
|
|
11172
|
+
cacheRead: number;
|
|
11173
|
+
cacheWrite: number;
|
|
11174
|
+
};
|
|
11175
|
+
contextWindow: number;
|
|
11176
|
+
maxTokens: number;
|
|
11177
|
+
};
|
|
11178
|
+
readonly "qwen/qwen3.5-flash-02-23": {
|
|
11179
|
+
id: string;
|
|
11180
|
+
name: string;
|
|
11181
|
+
api: "openai-completions";
|
|
11182
|
+
provider: string;
|
|
11183
|
+
baseUrl: string;
|
|
11184
|
+
reasoning: true;
|
|
11185
|
+
input: ("image" | "text")[];
|
|
11186
|
+
cost: {
|
|
11187
|
+
input: number;
|
|
11188
|
+
output: number;
|
|
11189
|
+
cacheRead: number;
|
|
11190
|
+
cacheWrite: number;
|
|
11191
|
+
};
|
|
11192
|
+
contextWindow: number;
|
|
11193
|
+
maxTokens: number;
|
|
11194
|
+
};
|
|
11195
|
+
readonly "qwen/qwen3.5-plus-02-15": {
|
|
11196
|
+
id: string;
|
|
11197
|
+
name: string;
|
|
11198
|
+
api: "openai-completions";
|
|
11199
|
+
provider: string;
|
|
11200
|
+
baseUrl: string;
|
|
11201
|
+
reasoning: true;
|
|
11202
|
+
input: ("image" | "text")[];
|
|
11203
|
+
cost: {
|
|
11204
|
+
input: number;
|
|
11205
|
+
output: number;
|
|
11206
|
+
cacheRead: number;
|
|
11207
|
+
cacheWrite: number;
|
|
11208
|
+
};
|
|
11209
|
+
contextWindow: number;
|
|
11210
|
+
maxTokens: number;
|
|
11211
|
+
};
|
|
11212
|
+
readonly "qwen/qwen3.5-plus-20260420": {
|
|
11213
|
+
id: string;
|
|
11214
|
+
name: string;
|
|
11215
|
+
api: "openai-completions";
|
|
11216
|
+
provider: string;
|
|
11217
|
+
baseUrl: string;
|
|
11218
|
+
reasoning: true;
|
|
11219
|
+
input: ("image" | "text")[];
|
|
11220
|
+
cost: {
|
|
11221
|
+
input: number;
|
|
11222
|
+
output: number;
|
|
11223
|
+
cacheRead: number;
|
|
11224
|
+
cacheWrite: number;
|
|
11225
|
+
};
|
|
11226
|
+
contextWindow: number;
|
|
11227
|
+
maxTokens: number;
|
|
11228
|
+
};
|
|
11229
|
+
readonly "qwen/qwen3.6-27b": {
|
|
11230
|
+
id: string;
|
|
11231
|
+
name: string;
|
|
11232
|
+
api: "openai-completions";
|
|
11233
|
+
provider: string;
|
|
11234
|
+
baseUrl: string;
|
|
11235
|
+
reasoning: true;
|
|
11236
|
+
input: ("image" | "text")[];
|
|
11237
|
+
cost: {
|
|
11238
|
+
input: number;
|
|
11239
|
+
output: number;
|
|
11240
|
+
cacheRead: number;
|
|
11241
|
+
cacheWrite: number;
|
|
11242
|
+
};
|
|
11243
|
+
contextWindow: number;
|
|
11244
|
+
maxTokens: number;
|
|
11245
|
+
};
|
|
11246
|
+
readonly "qwen/qwen3.6-flash": {
|
|
11247
|
+
id: string;
|
|
11248
|
+
name: string;
|
|
11249
|
+
api: "openai-completions";
|
|
11250
|
+
provider: string;
|
|
11251
|
+
baseUrl: string;
|
|
11252
|
+
reasoning: true;
|
|
11253
|
+
input: ("image" | "text")[];
|
|
11254
|
+
cost: {
|
|
11255
|
+
input: number;
|
|
11256
|
+
output: number;
|
|
11257
|
+
cacheRead: number;
|
|
11258
|
+
cacheWrite: number;
|
|
11259
|
+
};
|
|
11260
|
+
contextWindow: number;
|
|
11261
|
+
maxTokens: number;
|
|
11262
|
+
};
|
|
11263
|
+
readonly "qwen/qwen3.6-max-preview": {
|
|
11264
|
+
id: string;
|
|
11265
|
+
name: string;
|
|
11266
|
+
api: "openai-completions";
|
|
11267
|
+
provider: string;
|
|
11268
|
+
baseUrl: string;
|
|
11269
|
+
reasoning: true;
|
|
11270
|
+
input: "text"[];
|
|
11271
|
+
cost: {
|
|
11272
|
+
input: number;
|
|
11273
|
+
output: number;
|
|
11274
|
+
cacheRead: number;
|
|
11275
|
+
cacheWrite: number;
|
|
11276
|
+
};
|
|
11277
|
+
contextWindow: number;
|
|
11278
|
+
maxTokens: number;
|
|
11279
|
+
};
|
|
11280
|
+
readonly "qwen/qwen3.6-plus": {
|
|
11281
|
+
id: string;
|
|
11282
|
+
name: string;
|
|
11283
|
+
api: "openai-completions";
|
|
11284
|
+
provider: string;
|
|
11285
|
+
baseUrl: string;
|
|
11286
|
+
reasoning: true;
|
|
11287
|
+
input: ("image" | "text")[];
|
|
11288
|
+
cost: {
|
|
11289
|
+
input: number;
|
|
11290
|
+
output: number;
|
|
11291
|
+
cacheRead: number;
|
|
11292
|
+
cacheWrite: number;
|
|
11293
|
+
};
|
|
11294
|
+
contextWindow: number;
|
|
11295
|
+
maxTokens: number;
|
|
11296
|
+
};
|
|
11297
|
+
readonly "rekaai/reka-edge": {
|
|
11298
|
+
id: string;
|
|
11299
|
+
name: string;
|
|
11300
|
+
api: "openai-completions";
|
|
11301
|
+
provider: string;
|
|
11302
|
+
baseUrl: string;
|
|
11303
|
+
reasoning: false;
|
|
11304
|
+
input: ("image" | "text")[];
|
|
11305
|
+
cost: {
|
|
11306
|
+
input: number;
|
|
11307
|
+
output: number;
|
|
11308
|
+
cacheRead: number;
|
|
11309
|
+
cacheWrite: number;
|
|
11310
|
+
};
|
|
11311
|
+
contextWindow: number;
|
|
11312
|
+
maxTokens: number;
|
|
11313
|
+
};
|
|
11314
|
+
readonly "relace/relace-search": {
|
|
11315
|
+
id: string;
|
|
11316
|
+
name: string;
|
|
11317
|
+
api: "openai-completions";
|
|
11318
|
+
provider: string;
|
|
11319
|
+
baseUrl: string;
|
|
11320
|
+
reasoning: false;
|
|
11321
|
+
input: "text"[];
|
|
11322
|
+
cost: {
|
|
11323
|
+
input: number;
|
|
11324
|
+
output: number;
|
|
11325
|
+
cacheRead: number;
|
|
11326
|
+
cacheWrite: number;
|
|
11327
|
+
};
|
|
11328
|
+
contextWindow: number;
|
|
11329
|
+
maxTokens: number;
|
|
11330
|
+
};
|
|
11331
|
+
readonly "sao10k/l3-euryale-70b": {
|
|
11332
|
+
id: string;
|
|
11333
|
+
name: string;
|
|
11334
|
+
api: "openai-completions";
|
|
11335
|
+
provider: string;
|
|
11336
|
+
baseUrl: string;
|
|
11337
|
+
reasoning: false;
|
|
11338
|
+
input: "text"[];
|
|
11339
|
+
cost: {
|
|
11340
|
+
input: number;
|
|
11341
|
+
output: number;
|
|
11342
|
+
cacheRead: number;
|
|
11343
|
+
cacheWrite: number;
|
|
11344
|
+
};
|
|
11345
|
+
contextWindow: number;
|
|
11346
|
+
maxTokens: number;
|
|
11347
|
+
};
|
|
11348
|
+
readonly "sao10k/l3.1-euryale-70b": {
|
|
11349
|
+
id: string;
|
|
11350
|
+
name: string;
|
|
11351
|
+
api: "openai-completions";
|
|
11352
|
+
provider: string;
|
|
11353
|
+
baseUrl: string;
|
|
11354
|
+
reasoning: false;
|
|
11355
|
+
input: "text"[];
|
|
11356
|
+
cost: {
|
|
11357
|
+
input: number;
|
|
11358
|
+
output: number;
|
|
11359
|
+
cacheRead: number;
|
|
11360
|
+
cacheWrite: number;
|
|
11361
|
+
};
|
|
11362
|
+
contextWindow: number;
|
|
11363
|
+
maxTokens: number;
|
|
11364
|
+
};
|
|
11365
|
+
readonly "stepfun/step-3.5-flash": {
|
|
11366
|
+
id: string;
|
|
11367
|
+
name: string;
|
|
11368
|
+
api: "openai-completions";
|
|
11369
|
+
provider: string;
|
|
11370
|
+
baseUrl: string;
|
|
11371
|
+
reasoning: true;
|
|
11372
|
+
input: "text"[];
|
|
11373
|
+
cost: {
|
|
11374
|
+
input: number;
|
|
11375
|
+
output: number;
|
|
11376
|
+
cacheRead: number;
|
|
11377
|
+
cacheWrite: number;
|
|
11378
|
+
};
|
|
11379
|
+
contextWindow: number;
|
|
11380
|
+
maxTokens: number;
|
|
11381
|
+
};
|
|
11382
|
+
readonly "tencent/hy3-preview:free": {
|
|
11383
|
+
id: string;
|
|
11384
|
+
name: string;
|
|
11385
|
+
api: "openai-completions";
|
|
11386
|
+
provider: string;
|
|
11387
|
+
baseUrl: string;
|
|
11388
|
+
reasoning: true;
|
|
11389
|
+
input: "text"[];
|
|
11390
|
+
cost: {
|
|
11391
|
+
input: number;
|
|
11392
|
+
output: number;
|
|
11393
|
+
cacheRead: number;
|
|
11394
|
+
cacheWrite: number;
|
|
11395
|
+
};
|
|
11396
|
+
contextWindow: number;
|
|
11397
|
+
maxTokens: number;
|
|
11398
|
+
};
|
|
11399
|
+
readonly "thedrummer/rocinante-12b": {
|
|
11400
|
+
id: string;
|
|
11401
|
+
name: string;
|
|
11402
|
+
api: "openai-completions";
|
|
11403
|
+
provider: string;
|
|
11404
|
+
baseUrl: string;
|
|
11405
|
+
reasoning: false;
|
|
11406
|
+
input: "text"[];
|
|
10630
11407
|
cost: {
|
|
10631
11408
|
input: number;
|
|
10632
11409
|
output: number;
|
|
@@ -10636,14 +11413,14 @@ export declare const MODELS: {
|
|
|
10636
11413
|
contextWindow: number;
|
|
10637
11414
|
maxTokens: number;
|
|
10638
11415
|
};
|
|
10639
|
-
readonly "
|
|
11416
|
+
readonly "thedrummer/unslopnemo-12b": {
|
|
10640
11417
|
id: string;
|
|
10641
11418
|
name: string;
|
|
10642
11419
|
api: "openai-completions";
|
|
10643
11420
|
provider: string;
|
|
10644
11421
|
baseUrl: string;
|
|
10645
|
-
reasoning:
|
|
10646
|
-
input:
|
|
11422
|
+
reasoning: false;
|
|
11423
|
+
input: "text"[];
|
|
10647
11424
|
cost: {
|
|
10648
11425
|
input: number;
|
|
10649
11426
|
output: number;
|
|
@@ -10653,14 +11430,14 @@ export declare const MODELS: {
|
|
|
10653
11430
|
contextWindow: number;
|
|
10654
11431
|
maxTokens: number;
|
|
10655
11432
|
};
|
|
10656
|
-
readonly "
|
|
11433
|
+
readonly "tngtech/deepseek-r1t2-chimera": {
|
|
10657
11434
|
id: string;
|
|
10658
11435
|
name: string;
|
|
10659
11436
|
api: "openai-completions";
|
|
10660
11437
|
provider: string;
|
|
10661
11438
|
baseUrl: string;
|
|
10662
11439
|
reasoning: true;
|
|
10663
|
-
input:
|
|
11440
|
+
input: "text"[];
|
|
10664
11441
|
cost: {
|
|
10665
11442
|
input: number;
|
|
10666
11443
|
output: number;
|
|
@@ -10670,14 +11447,14 @@ export declare const MODELS: {
|
|
|
10670
11447
|
contextWindow: number;
|
|
10671
11448
|
maxTokens: number;
|
|
10672
11449
|
};
|
|
10673
|
-
readonly "
|
|
11450
|
+
readonly "upstage/solar-pro-3": {
|
|
10674
11451
|
id: string;
|
|
10675
11452
|
name: string;
|
|
10676
11453
|
api: "openai-completions";
|
|
10677
11454
|
provider: string;
|
|
10678
11455
|
baseUrl: string;
|
|
10679
11456
|
reasoning: true;
|
|
10680
|
-
input:
|
|
11457
|
+
input: "text"[];
|
|
10681
11458
|
cost: {
|
|
10682
11459
|
input: number;
|
|
10683
11460
|
output: number;
|
|
@@ -10687,14 +11464,14 @@ export declare const MODELS: {
|
|
|
10687
11464
|
contextWindow: number;
|
|
10688
11465
|
maxTokens: number;
|
|
10689
11466
|
};
|
|
10690
|
-
readonly "
|
|
11467
|
+
readonly "x-ai/grok-3": {
|
|
10691
11468
|
id: string;
|
|
10692
11469
|
name: string;
|
|
10693
11470
|
api: "openai-completions";
|
|
10694
11471
|
provider: string;
|
|
10695
11472
|
baseUrl: string;
|
|
10696
|
-
reasoning:
|
|
10697
|
-
input:
|
|
11473
|
+
reasoning: false;
|
|
11474
|
+
input: "text"[];
|
|
10698
11475
|
cost: {
|
|
10699
11476
|
input: number;
|
|
10700
11477
|
output: number;
|
|
@@ -10704,13 +11481,13 @@ export declare const MODELS: {
|
|
|
10704
11481
|
contextWindow: number;
|
|
10705
11482
|
maxTokens: number;
|
|
10706
11483
|
};
|
|
10707
|
-
readonly "
|
|
11484
|
+
readonly "x-ai/grok-3-beta": {
|
|
10708
11485
|
id: string;
|
|
10709
11486
|
name: string;
|
|
10710
11487
|
api: "openai-completions";
|
|
10711
11488
|
provider: string;
|
|
10712
11489
|
baseUrl: string;
|
|
10713
|
-
reasoning:
|
|
11490
|
+
reasoning: false;
|
|
10714
11491
|
input: "text"[];
|
|
10715
11492
|
cost: {
|
|
10716
11493
|
input: number;
|
|
@@ -10721,14 +11498,14 @@ export declare const MODELS: {
|
|
|
10721
11498
|
contextWindow: number;
|
|
10722
11499
|
maxTokens: number;
|
|
10723
11500
|
};
|
|
10724
|
-
readonly "
|
|
11501
|
+
readonly "x-ai/grok-3-mini": {
|
|
10725
11502
|
id: string;
|
|
10726
11503
|
name: string;
|
|
10727
11504
|
api: "openai-completions";
|
|
10728
11505
|
provider: string;
|
|
10729
11506
|
baseUrl: string;
|
|
10730
|
-
reasoning:
|
|
10731
|
-
input:
|
|
11507
|
+
reasoning: true;
|
|
11508
|
+
input: "text"[];
|
|
10732
11509
|
cost: {
|
|
10733
11510
|
input: number;
|
|
10734
11511
|
output: number;
|
|
@@ -10738,13 +11515,13 @@ export declare const MODELS: {
|
|
|
10738
11515
|
contextWindow: number;
|
|
10739
11516
|
maxTokens: number;
|
|
10740
11517
|
};
|
|
10741
|
-
readonly "
|
|
11518
|
+
readonly "x-ai/grok-3-mini-beta": {
|
|
10742
11519
|
id: string;
|
|
10743
11520
|
name: string;
|
|
10744
11521
|
api: "openai-completions";
|
|
10745
11522
|
provider: string;
|
|
10746
11523
|
baseUrl: string;
|
|
10747
|
-
reasoning:
|
|
11524
|
+
reasoning: true;
|
|
10748
11525
|
input: "text"[];
|
|
10749
11526
|
cost: {
|
|
10750
11527
|
input: number;
|
|
@@ -10755,14 +11532,14 @@ export declare const MODELS: {
|
|
|
10755
11532
|
contextWindow: number;
|
|
10756
11533
|
maxTokens: number;
|
|
10757
11534
|
};
|
|
10758
|
-
readonly "
|
|
11535
|
+
readonly "x-ai/grok-4": {
|
|
10759
11536
|
id: string;
|
|
10760
11537
|
name: string;
|
|
10761
11538
|
api: "openai-completions";
|
|
10762
11539
|
provider: string;
|
|
10763
11540
|
baseUrl: string;
|
|
10764
|
-
reasoning:
|
|
10765
|
-
input: "text"[];
|
|
11541
|
+
reasoning: true;
|
|
11542
|
+
input: ("image" | "text")[];
|
|
10766
11543
|
cost: {
|
|
10767
11544
|
input: number;
|
|
10768
11545
|
output: number;
|
|
@@ -10772,14 +11549,14 @@ export declare const MODELS: {
|
|
|
10772
11549
|
contextWindow: number;
|
|
10773
11550
|
maxTokens: number;
|
|
10774
11551
|
};
|
|
10775
|
-
readonly "
|
|
11552
|
+
readonly "x-ai/grok-4-fast": {
|
|
10776
11553
|
id: string;
|
|
10777
11554
|
name: string;
|
|
10778
11555
|
api: "openai-completions";
|
|
10779
11556
|
provider: string;
|
|
10780
11557
|
baseUrl: string;
|
|
10781
|
-
reasoning:
|
|
10782
|
-
input: "text"[];
|
|
11558
|
+
reasoning: true;
|
|
11559
|
+
input: ("image" | "text")[];
|
|
10783
11560
|
cost: {
|
|
10784
11561
|
input: number;
|
|
10785
11562
|
output: number;
|
|
@@ -10789,14 +11566,14 @@ export declare const MODELS: {
|
|
|
10789
11566
|
contextWindow: number;
|
|
10790
11567
|
maxTokens: number;
|
|
10791
11568
|
};
|
|
10792
|
-
readonly "
|
|
11569
|
+
readonly "x-ai/grok-4.1-fast": {
|
|
10793
11570
|
id: string;
|
|
10794
11571
|
name: string;
|
|
10795
11572
|
api: "openai-completions";
|
|
10796
11573
|
provider: string;
|
|
10797
11574
|
baseUrl: string;
|
|
10798
11575
|
reasoning: true;
|
|
10799
|
-
input: "text"[];
|
|
11576
|
+
input: ("image" | "text")[];
|
|
10800
11577
|
cost: {
|
|
10801
11578
|
input: number;
|
|
10802
11579
|
output: number;
|
|
@@ -10806,14 +11583,14 @@ export declare const MODELS: {
|
|
|
10806
11583
|
contextWindow: number;
|
|
10807
11584
|
maxTokens: number;
|
|
10808
11585
|
};
|
|
10809
|
-
readonly "
|
|
11586
|
+
readonly "x-ai/grok-4.20": {
|
|
10810
11587
|
id: string;
|
|
10811
11588
|
name: string;
|
|
10812
11589
|
api: "openai-completions";
|
|
10813
11590
|
provider: string;
|
|
10814
11591
|
baseUrl: string;
|
|
10815
|
-
reasoning:
|
|
10816
|
-
input: "text"[];
|
|
11592
|
+
reasoning: true;
|
|
11593
|
+
input: ("image" | "text")[];
|
|
10817
11594
|
cost: {
|
|
10818
11595
|
input: number;
|
|
10819
11596
|
output: number;
|
|
@@ -10823,14 +11600,14 @@ export declare const MODELS: {
|
|
|
10823
11600
|
contextWindow: number;
|
|
10824
11601
|
maxTokens: number;
|
|
10825
11602
|
};
|
|
10826
|
-
readonly "
|
|
11603
|
+
readonly "x-ai/grok-4.3": {
|
|
10827
11604
|
id: string;
|
|
10828
11605
|
name: string;
|
|
10829
11606
|
api: "openai-completions";
|
|
10830
11607
|
provider: string;
|
|
10831
11608
|
baseUrl: string;
|
|
10832
|
-
reasoning:
|
|
10833
|
-
input: "text"[];
|
|
11609
|
+
reasoning: true;
|
|
11610
|
+
input: ("image" | "text")[];
|
|
10834
11611
|
cost: {
|
|
10835
11612
|
input: number;
|
|
10836
11613
|
output: number;
|
|
@@ -10840,7 +11617,7 @@ export declare const MODELS: {
|
|
|
10840
11617
|
contextWindow: number;
|
|
10841
11618
|
maxTokens: number;
|
|
10842
11619
|
};
|
|
10843
|
-
readonly "
|
|
11620
|
+
readonly "x-ai/grok-code-fast-1": {
|
|
10844
11621
|
id: string;
|
|
10845
11622
|
name: string;
|
|
10846
11623
|
api: "openai-completions";
|
|
@@ -10857,7 +11634,7 @@ export declare const MODELS: {
|
|
|
10857
11634
|
contextWindow: number;
|
|
10858
11635
|
maxTokens: number;
|
|
10859
11636
|
};
|
|
10860
|
-
readonly "
|
|
11637
|
+
readonly "xiaomi/mimo-v2-flash": {
|
|
10861
11638
|
id: string;
|
|
10862
11639
|
name: string;
|
|
10863
11640
|
api: "openai-completions";
|
|
@@ -10874,14 +11651,14 @@ export declare const MODELS: {
|
|
|
10874
11651
|
contextWindow: number;
|
|
10875
11652
|
maxTokens: number;
|
|
10876
11653
|
};
|
|
10877
|
-
readonly "
|
|
11654
|
+
readonly "xiaomi/mimo-v2-omni": {
|
|
10878
11655
|
id: string;
|
|
10879
11656
|
name: string;
|
|
10880
11657
|
api: "openai-completions";
|
|
10881
11658
|
provider: string;
|
|
10882
11659
|
baseUrl: string;
|
|
10883
|
-
reasoning:
|
|
10884
|
-
input: "text"[];
|
|
11660
|
+
reasoning: true;
|
|
11661
|
+
input: ("image" | "text")[];
|
|
10885
11662
|
cost: {
|
|
10886
11663
|
input: number;
|
|
10887
11664
|
output: number;
|
|
@@ -10891,13 +11668,13 @@ export declare const MODELS: {
|
|
|
10891
11668
|
contextWindow: number;
|
|
10892
11669
|
maxTokens: number;
|
|
10893
11670
|
};
|
|
10894
|
-
readonly "
|
|
11671
|
+
readonly "xiaomi/mimo-v2-pro": {
|
|
10895
11672
|
id: string;
|
|
10896
11673
|
name: string;
|
|
10897
11674
|
api: "openai-completions";
|
|
10898
11675
|
provider: string;
|
|
10899
11676
|
baseUrl: string;
|
|
10900
|
-
reasoning:
|
|
11677
|
+
reasoning: true;
|
|
10901
11678
|
input: "text"[];
|
|
10902
11679
|
cost: {
|
|
10903
11680
|
input: number;
|
|
@@ -10908,14 +11685,14 @@ export declare const MODELS: {
|
|
|
10908
11685
|
contextWindow: number;
|
|
10909
11686
|
maxTokens: number;
|
|
10910
11687
|
};
|
|
10911
|
-
readonly "
|
|
11688
|
+
readonly "xiaomi/mimo-v2.5": {
|
|
10912
11689
|
id: string;
|
|
10913
11690
|
name: string;
|
|
10914
11691
|
api: "openai-completions";
|
|
10915
11692
|
provider: string;
|
|
10916
11693
|
baseUrl: string;
|
|
10917
11694
|
reasoning: true;
|
|
10918
|
-
input: "text"[];
|
|
11695
|
+
input: ("image" | "text")[];
|
|
10919
11696
|
cost: {
|
|
10920
11697
|
input: number;
|
|
10921
11698
|
output: number;
|
|
@@ -10925,7 +11702,7 @@ export declare const MODELS: {
|
|
|
10925
11702
|
contextWindow: number;
|
|
10926
11703
|
maxTokens: number;
|
|
10927
11704
|
};
|
|
10928
|
-
readonly "
|
|
11705
|
+
readonly "xiaomi/mimo-v2.5-pro": {
|
|
10929
11706
|
id: string;
|
|
10930
11707
|
name: string;
|
|
10931
11708
|
api: "openai-completions";
|
|
@@ -10942,14 +11719,14 @@ export declare const MODELS: {
|
|
|
10942
11719
|
contextWindow: number;
|
|
10943
11720
|
maxTokens: number;
|
|
10944
11721
|
};
|
|
10945
|
-
readonly "
|
|
11722
|
+
readonly "z-ai/glm-4-32b": {
|
|
10946
11723
|
id: string;
|
|
10947
11724
|
name: string;
|
|
10948
11725
|
api: "openai-completions";
|
|
10949
11726
|
provider: string;
|
|
10950
11727
|
baseUrl: string;
|
|
10951
|
-
reasoning:
|
|
10952
|
-
input:
|
|
11728
|
+
reasoning: false;
|
|
11729
|
+
input: "text"[];
|
|
10953
11730
|
cost: {
|
|
10954
11731
|
input: number;
|
|
10955
11732
|
output: number;
|
|
@@ -10959,14 +11736,14 @@ export declare const MODELS: {
|
|
|
10959
11736
|
contextWindow: number;
|
|
10960
11737
|
maxTokens: number;
|
|
10961
11738
|
};
|
|
10962
|
-
readonly "
|
|
11739
|
+
readonly "z-ai/glm-4.5": {
|
|
10963
11740
|
id: string;
|
|
10964
11741
|
name: string;
|
|
10965
11742
|
api: "openai-completions";
|
|
10966
11743
|
provider: string;
|
|
10967
11744
|
baseUrl: string;
|
|
10968
11745
|
reasoning: true;
|
|
10969
|
-
input:
|
|
11746
|
+
input: "text"[];
|
|
10970
11747
|
cost: {
|
|
10971
11748
|
input: number;
|
|
10972
11749
|
output: number;
|
|
@@ -10976,14 +11753,14 @@ export declare const MODELS: {
|
|
|
10976
11753
|
contextWindow: number;
|
|
10977
11754
|
maxTokens: number;
|
|
10978
11755
|
};
|
|
10979
|
-
readonly "
|
|
11756
|
+
readonly "z-ai/glm-4.5-air": {
|
|
10980
11757
|
id: string;
|
|
10981
11758
|
name: string;
|
|
10982
11759
|
api: "openai-completions";
|
|
10983
11760
|
provider: string;
|
|
10984
11761
|
baseUrl: string;
|
|
10985
11762
|
reasoning: true;
|
|
10986
|
-
input:
|
|
11763
|
+
input: "text"[];
|
|
10987
11764
|
cost: {
|
|
10988
11765
|
input: number;
|
|
10989
11766
|
output: number;
|
|
@@ -10993,14 +11770,14 @@ export declare const MODELS: {
|
|
|
10993
11770
|
contextWindow: number;
|
|
10994
11771
|
maxTokens: number;
|
|
10995
11772
|
};
|
|
10996
|
-
readonly "
|
|
11773
|
+
readonly "z-ai/glm-4.5-air:free": {
|
|
10997
11774
|
id: string;
|
|
10998
11775
|
name: string;
|
|
10999
11776
|
api: "openai-completions";
|
|
11000
11777
|
provider: string;
|
|
11001
11778
|
baseUrl: string;
|
|
11002
11779
|
reasoning: true;
|
|
11003
|
-
input:
|
|
11780
|
+
input: "text"[];
|
|
11004
11781
|
cost: {
|
|
11005
11782
|
input: number;
|
|
11006
11783
|
output: number;
|
|
@@ -11010,14 +11787,14 @@ export declare const MODELS: {
|
|
|
11010
11787
|
contextWindow: number;
|
|
11011
11788
|
maxTokens: number;
|
|
11012
11789
|
};
|
|
11013
|
-
readonly "
|
|
11790
|
+
readonly "z-ai/glm-4.5v": {
|
|
11014
11791
|
id: string;
|
|
11015
11792
|
name: string;
|
|
11016
11793
|
api: "openai-completions";
|
|
11017
11794
|
provider: string;
|
|
11018
11795
|
baseUrl: string;
|
|
11019
11796
|
reasoning: true;
|
|
11020
|
-
input: "text"[];
|
|
11797
|
+
input: ("image" | "text")[];
|
|
11021
11798
|
cost: {
|
|
11022
11799
|
input: number;
|
|
11023
11800
|
output: number;
|
|
@@ -11027,7 +11804,7 @@ export declare const MODELS: {
|
|
|
11027
11804
|
contextWindow: number;
|
|
11028
11805
|
maxTokens: number;
|
|
11029
11806
|
};
|
|
11030
|
-
readonly "
|
|
11807
|
+
readonly "z-ai/glm-4.6": {
|
|
11031
11808
|
id: string;
|
|
11032
11809
|
name: string;
|
|
11033
11810
|
api: "openai-completions";
|
|
@@ -11044,7 +11821,7 @@ export declare const MODELS: {
|
|
|
11044
11821
|
contextWindow: number;
|
|
11045
11822
|
maxTokens: number;
|
|
11046
11823
|
};
|
|
11047
|
-
readonly "
|
|
11824
|
+
readonly "z-ai/glm-4.6v": {
|
|
11048
11825
|
id: string;
|
|
11049
11826
|
name: string;
|
|
11050
11827
|
api: "openai-completions";
|
|
@@ -11061,7 +11838,7 @@ export declare const MODELS: {
|
|
|
11061
11838
|
contextWindow: number;
|
|
11062
11839
|
maxTokens: number;
|
|
11063
11840
|
};
|
|
11064
|
-
readonly "
|
|
11841
|
+
readonly "z-ai/glm-4.7": {
|
|
11065
11842
|
id: string;
|
|
11066
11843
|
name: string;
|
|
11067
11844
|
api: "openai-completions";
|
|
@@ -11078,13 +11855,13 @@ export declare const MODELS: {
|
|
|
11078
11855
|
contextWindow: number;
|
|
11079
11856
|
maxTokens: number;
|
|
11080
11857
|
};
|
|
11081
|
-
readonly "z-ai/glm-4-
|
|
11858
|
+
readonly "z-ai/glm-4.7-flash": {
|
|
11082
11859
|
id: string;
|
|
11083
11860
|
name: string;
|
|
11084
11861
|
api: "openai-completions";
|
|
11085
11862
|
provider: string;
|
|
11086
11863
|
baseUrl: string;
|
|
11087
|
-
reasoning:
|
|
11864
|
+
reasoning: true;
|
|
11088
11865
|
input: "text"[];
|
|
11089
11866
|
cost: {
|
|
11090
11867
|
input: number;
|
|
@@ -11095,7 +11872,7 @@ export declare const MODELS: {
|
|
|
11095
11872
|
contextWindow: number;
|
|
11096
11873
|
maxTokens: number;
|
|
11097
11874
|
};
|
|
11098
|
-
readonly "z-ai/glm-
|
|
11875
|
+
readonly "z-ai/glm-5": {
|
|
11099
11876
|
id: string;
|
|
11100
11877
|
name: string;
|
|
11101
11878
|
api: "openai-completions";
|
|
@@ -11112,7 +11889,7 @@ export declare const MODELS: {
|
|
|
11112
11889
|
contextWindow: number;
|
|
11113
11890
|
maxTokens: number;
|
|
11114
11891
|
};
|
|
11115
|
-
readonly "z-ai/glm-
|
|
11892
|
+
readonly "z-ai/glm-5-turbo": {
|
|
11116
11893
|
id: string;
|
|
11117
11894
|
name: string;
|
|
11118
11895
|
api: "openai-completions";
|
|
@@ -11129,7 +11906,7 @@ export declare const MODELS: {
|
|
|
11129
11906
|
contextWindow: number;
|
|
11130
11907
|
maxTokens: number;
|
|
11131
11908
|
};
|
|
11132
|
-
readonly "z-ai/glm-
|
|
11909
|
+
readonly "z-ai/glm-5.1": {
|
|
11133
11910
|
id: string;
|
|
11134
11911
|
name: string;
|
|
11135
11912
|
api: "openai-completions";
|
|
@@ -11146,7 +11923,7 @@ export declare const MODELS: {
|
|
|
11146
11923
|
contextWindow: number;
|
|
11147
11924
|
maxTokens: number;
|
|
11148
11925
|
};
|
|
11149
|
-
readonly "z-ai/glm-
|
|
11926
|
+
readonly "z-ai/glm-5v-turbo": {
|
|
11150
11927
|
id: string;
|
|
11151
11928
|
name: string;
|
|
11152
11929
|
api: "openai-completions";
|
|
@@ -11163,14 +11940,14 @@ export declare const MODELS: {
|
|
|
11163
11940
|
contextWindow: number;
|
|
11164
11941
|
maxTokens: number;
|
|
11165
11942
|
};
|
|
11166
|
-
readonly "
|
|
11943
|
+
readonly "~anthropic/claude-haiku-latest": {
|
|
11167
11944
|
id: string;
|
|
11168
11945
|
name: string;
|
|
11169
11946
|
api: "openai-completions";
|
|
11170
11947
|
provider: string;
|
|
11171
11948
|
baseUrl: string;
|
|
11172
11949
|
reasoning: true;
|
|
11173
|
-
input: "text"[];
|
|
11950
|
+
input: ("image" | "text")[];
|
|
11174
11951
|
cost: {
|
|
11175
11952
|
input: number;
|
|
11176
11953
|
output: number;
|
|
@@ -11180,7 +11957,7 @@ export declare const MODELS: {
|
|
|
11180
11957
|
contextWindow: number;
|
|
11181
11958
|
maxTokens: number;
|
|
11182
11959
|
};
|
|
11183
|
-
readonly "
|
|
11960
|
+
readonly "~anthropic/claude-opus-latest": {
|
|
11184
11961
|
id: string;
|
|
11185
11962
|
name: string;
|
|
11186
11963
|
api: "openai-completions";
|
|
@@ -11197,14 +11974,14 @@ export declare const MODELS: {
|
|
|
11197
11974
|
contextWindow: number;
|
|
11198
11975
|
maxTokens: number;
|
|
11199
11976
|
};
|
|
11200
|
-
readonly "
|
|
11977
|
+
readonly "~anthropic/claude-sonnet-latest": {
|
|
11201
11978
|
id: string;
|
|
11202
11979
|
name: string;
|
|
11203
11980
|
api: "openai-completions";
|
|
11204
11981
|
provider: string;
|
|
11205
11982
|
baseUrl: string;
|
|
11206
11983
|
reasoning: true;
|
|
11207
|
-
input: "text"[];
|
|
11984
|
+
input: ("image" | "text")[];
|
|
11208
11985
|
cost: {
|
|
11209
11986
|
input: number;
|
|
11210
11987
|
output: number;
|
|
@@ -11214,14 +11991,14 @@ export declare const MODELS: {
|
|
|
11214
11991
|
contextWindow: number;
|
|
11215
11992
|
maxTokens: number;
|
|
11216
11993
|
};
|
|
11217
|
-
readonly "
|
|
11994
|
+
readonly "~google/gemini-flash-latest": {
|
|
11218
11995
|
id: string;
|
|
11219
11996
|
name: string;
|
|
11220
11997
|
api: "openai-completions";
|
|
11221
11998
|
provider: string;
|
|
11222
11999
|
baseUrl: string;
|
|
11223
12000
|
reasoning: true;
|
|
11224
|
-
input: "text"[];
|
|
12001
|
+
input: ("image" | "text")[];
|
|
11225
12002
|
cost: {
|
|
11226
12003
|
input: number;
|
|
11227
12004
|
output: number;
|
|
@@ -11231,14 +12008,14 @@ export declare const MODELS: {
|
|
|
11231
12008
|
contextWindow: number;
|
|
11232
12009
|
maxTokens: number;
|
|
11233
12010
|
};
|
|
11234
|
-
readonly "
|
|
12011
|
+
readonly "~google/gemini-pro-latest": {
|
|
11235
12012
|
id: string;
|
|
11236
12013
|
name: string;
|
|
11237
12014
|
api: "openai-completions";
|
|
11238
12015
|
provider: string;
|
|
11239
12016
|
baseUrl: string;
|
|
11240
12017
|
reasoning: true;
|
|
11241
|
-
input: "text"[];
|
|
12018
|
+
input: ("image" | "text")[];
|
|
11242
12019
|
cost: {
|
|
11243
12020
|
input: number;
|
|
11244
12021
|
output: number;
|
|
@@ -11248,14 +12025,14 @@ export declare const MODELS: {
|
|
|
11248
12025
|
contextWindow: number;
|
|
11249
12026
|
maxTokens: number;
|
|
11250
12027
|
};
|
|
11251
|
-
readonly "
|
|
12028
|
+
readonly "~moonshotai/kimi-latest": {
|
|
11252
12029
|
id: string;
|
|
11253
12030
|
name: string;
|
|
11254
12031
|
api: "openai-completions";
|
|
11255
12032
|
provider: string;
|
|
11256
12033
|
baseUrl: string;
|
|
11257
12034
|
reasoning: true;
|
|
11258
|
-
input: "text"[];
|
|
12035
|
+
input: ("image" | "text")[];
|
|
11259
12036
|
cost: {
|
|
11260
12037
|
input: number;
|
|
11261
12038
|
output: number;
|
|
@@ -11265,14 +12042,14 @@ export declare const MODELS: {
|
|
|
11265
12042
|
contextWindow: number;
|
|
11266
12043
|
maxTokens: number;
|
|
11267
12044
|
};
|
|
11268
|
-
readonly "
|
|
12045
|
+
readonly "~openai/gpt-latest": {
|
|
11269
12046
|
id: string;
|
|
11270
12047
|
name: string;
|
|
11271
12048
|
api: "openai-completions";
|
|
11272
12049
|
provider: string;
|
|
11273
12050
|
baseUrl: string;
|
|
11274
12051
|
reasoning: true;
|
|
11275
|
-
input: "text"[];
|
|
12052
|
+
input: ("image" | "text")[];
|
|
11276
12053
|
cost: {
|
|
11277
12054
|
input: number;
|
|
11278
12055
|
output: number;
|
|
@@ -11282,7 +12059,7 @@ export declare const MODELS: {
|
|
|
11282
12059
|
contextWindow: number;
|
|
11283
12060
|
maxTokens: number;
|
|
11284
12061
|
};
|
|
11285
|
-
readonly "
|
|
12062
|
+
readonly "~openai/gpt-mini-latest": {
|
|
11286
12063
|
id: string;
|
|
11287
12064
|
name: string;
|
|
11288
12065
|
api: "openai-completions";
|
|
@@ -11369,6 +12146,23 @@ export declare const MODELS: {
|
|
|
11369
12146
|
contextWindow: number;
|
|
11370
12147
|
maxTokens: number;
|
|
11371
12148
|
};
|
|
12149
|
+
readonly "alibaba/qwen-3.6-max-preview": {
|
|
12150
|
+
id: string;
|
|
12151
|
+
name: string;
|
|
12152
|
+
api: "anthropic-messages";
|
|
12153
|
+
provider: string;
|
|
12154
|
+
baseUrl: string;
|
|
12155
|
+
reasoning: true;
|
|
12156
|
+
input: ("image" | "text")[];
|
|
12157
|
+
cost: {
|
|
12158
|
+
input: number;
|
|
12159
|
+
output: number;
|
|
12160
|
+
cacheRead: number;
|
|
12161
|
+
cacheWrite: number;
|
|
12162
|
+
};
|
|
12163
|
+
contextWindow: number;
|
|
12164
|
+
maxTokens: number;
|
|
12165
|
+
};
|
|
11372
12166
|
readonly "alibaba/qwen3-235b-a22b-thinking": {
|
|
11373
12167
|
id: string;
|
|
11374
12168
|
name: string;
|
|
@@ -11556,6 +12350,23 @@ export declare const MODELS: {
|
|
|
11556
12350
|
contextWindow: number;
|
|
11557
12351
|
maxTokens: number;
|
|
11558
12352
|
};
|
|
12353
|
+
readonly "alibaba/qwen3.6-27b": {
|
|
12354
|
+
id: string;
|
|
12355
|
+
name: string;
|
|
12356
|
+
api: "anthropic-messages";
|
|
12357
|
+
provider: string;
|
|
12358
|
+
baseUrl: string;
|
|
12359
|
+
reasoning: true;
|
|
12360
|
+
input: ("image" | "text")[];
|
|
12361
|
+
cost: {
|
|
12362
|
+
input: number;
|
|
12363
|
+
output: number;
|
|
12364
|
+
cacheRead: number;
|
|
12365
|
+
cacheWrite: number;
|
|
12366
|
+
};
|
|
12367
|
+
contextWindow: number;
|
|
12368
|
+
maxTokens: number;
|
|
12369
|
+
};
|
|
11559
12370
|
readonly "alibaba/qwen3.6-plus": {
|
|
11560
12371
|
id: string;
|
|
11561
12372
|
name: string;
|
|
@@ -11931,6 +12742,40 @@ export declare const MODELS: {
|
|
|
11931
12742
|
maxTokens: number;
|
|
11932
12743
|
};
|
|
11933
12744
|
readonly "deepseek/deepseek-v3.2-thinking": {
|
|
12745
|
+
id: string;
|
|
12746
|
+
name: string;
|
|
12747
|
+
api: "anthropic-messages";
|
|
12748
|
+
provider: string;
|
|
12749
|
+
baseUrl: string;
|
|
12750
|
+
reasoning: false;
|
|
12751
|
+
input: "text"[];
|
|
12752
|
+
cost: {
|
|
12753
|
+
input: number;
|
|
12754
|
+
output: number;
|
|
12755
|
+
cacheRead: number;
|
|
12756
|
+
cacheWrite: number;
|
|
12757
|
+
};
|
|
12758
|
+
contextWindow: number;
|
|
12759
|
+
maxTokens: number;
|
|
12760
|
+
};
|
|
12761
|
+
readonly "deepseek/deepseek-v4-flash": {
|
|
12762
|
+
id: string;
|
|
12763
|
+
name: string;
|
|
12764
|
+
api: "anthropic-messages";
|
|
12765
|
+
provider: string;
|
|
12766
|
+
baseUrl: string;
|
|
12767
|
+
reasoning: true;
|
|
12768
|
+
input: "text"[];
|
|
12769
|
+
cost: {
|
|
12770
|
+
input: number;
|
|
12771
|
+
output: number;
|
|
12772
|
+
cacheRead: number;
|
|
12773
|
+
cacheWrite: number;
|
|
12774
|
+
};
|
|
12775
|
+
contextWindow: number;
|
|
12776
|
+
maxTokens: number;
|
|
12777
|
+
};
|
|
12778
|
+
readonly "deepseek/deepseek-v4-pro": {
|
|
11934
12779
|
id: string;
|
|
11935
12780
|
name: string;
|
|
11936
12781
|
api: "anthropic-messages";
|
|
@@ -12712,6 +13557,23 @@ export declare const MODELS: {
|
|
|
12712
13557
|
contextWindow: number;
|
|
12713
13558
|
maxTokens: number;
|
|
12714
13559
|
};
|
|
13560
|
+
readonly "moonshotai/kimi-k2.6": {
|
|
13561
|
+
id: string;
|
|
13562
|
+
name: string;
|
|
13563
|
+
api: "anthropic-messages";
|
|
13564
|
+
provider: string;
|
|
13565
|
+
baseUrl: string;
|
|
13566
|
+
reasoning: true;
|
|
13567
|
+
input: ("image" | "text")[];
|
|
13568
|
+
cost: {
|
|
13569
|
+
input: number;
|
|
13570
|
+
output: number;
|
|
13571
|
+
cacheRead: number;
|
|
13572
|
+
cacheWrite: number;
|
|
13573
|
+
};
|
|
13574
|
+
contextWindow: number;
|
|
13575
|
+
maxTokens: number;
|
|
13576
|
+
};
|
|
12715
13577
|
readonly "nvidia/nemotron-nano-12b-v2-vl": {
|
|
12716
13578
|
id: string;
|
|
12717
13579
|
name: string;
|
|
@@ -13001,7 +13863,41 @@ export declare const MODELS: {
|
|
|
13001
13863
|
contextWindow: number;
|
|
13002
13864
|
maxTokens: number;
|
|
13003
13865
|
};
|
|
13004
|
-
readonly "openai/gpt-5.1-instant": {
|
|
13866
|
+
readonly "openai/gpt-5.1-instant": {
|
|
13867
|
+
id: string;
|
|
13868
|
+
name: string;
|
|
13869
|
+
api: "anthropic-messages";
|
|
13870
|
+
provider: string;
|
|
13871
|
+
baseUrl: string;
|
|
13872
|
+
reasoning: true;
|
|
13873
|
+
input: ("image" | "text")[];
|
|
13874
|
+
cost: {
|
|
13875
|
+
input: number;
|
|
13876
|
+
output: number;
|
|
13877
|
+
cacheRead: number;
|
|
13878
|
+
cacheWrite: number;
|
|
13879
|
+
};
|
|
13880
|
+
contextWindow: number;
|
|
13881
|
+
maxTokens: number;
|
|
13882
|
+
};
|
|
13883
|
+
readonly "openai/gpt-5.1-thinking": {
|
|
13884
|
+
id: string;
|
|
13885
|
+
name: string;
|
|
13886
|
+
api: "anthropic-messages";
|
|
13887
|
+
provider: string;
|
|
13888
|
+
baseUrl: string;
|
|
13889
|
+
reasoning: true;
|
|
13890
|
+
input: ("image" | "text")[];
|
|
13891
|
+
cost: {
|
|
13892
|
+
input: number;
|
|
13893
|
+
output: number;
|
|
13894
|
+
cacheRead: number;
|
|
13895
|
+
cacheWrite: number;
|
|
13896
|
+
};
|
|
13897
|
+
contextWindow: number;
|
|
13898
|
+
maxTokens: number;
|
|
13899
|
+
};
|
|
13900
|
+
readonly "openai/gpt-5.2": {
|
|
13005
13901
|
id: string;
|
|
13006
13902
|
name: string;
|
|
13007
13903
|
api: "anthropic-messages";
|
|
@@ -13018,7 +13914,7 @@ export declare const MODELS: {
|
|
|
13018
13914
|
contextWindow: number;
|
|
13019
13915
|
maxTokens: number;
|
|
13020
13916
|
};
|
|
13021
|
-
readonly "openai/gpt-5.
|
|
13917
|
+
readonly "openai/gpt-5.2-chat": {
|
|
13022
13918
|
id: string;
|
|
13023
13919
|
name: string;
|
|
13024
13920
|
api: "anthropic-messages";
|
|
@@ -13035,7 +13931,7 @@ export declare const MODELS: {
|
|
|
13035
13931
|
contextWindow: number;
|
|
13036
13932
|
maxTokens: number;
|
|
13037
13933
|
};
|
|
13038
|
-
readonly "openai/gpt-5.2": {
|
|
13934
|
+
readonly "openai/gpt-5.2-codex": {
|
|
13039
13935
|
id: string;
|
|
13040
13936
|
name: string;
|
|
13041
13937
|
api: "anthropic-messages";
|
|
@@ -13052,7 +13948,7 @@ export declare const MODELS: {
|
|
|
13052
13948
|
contextWindow: number;
|
|
13053
13949
|
maxTokens: number;
|
|
13054
13950
|
};
|
|
13055
|
-
readonly "openai/gpt-5.2-
|
|
13951
|
+
readonly "openai/gpt-5.2-pro": {
|
|
13056
13952
|
id: string;
|
|
13057
13953
|
name: string;
|
|
13058
13954
|
api: "anthropic-messages";
|
|
@@ -13069,7 +13965,7 @@ export declare const MODELS: {
|
|
|
13069
13965
|
contextWindow: number;
|
|
13070
13966
|
maxTokens: number;
|
|
13071
13967
|
};
|
|
13072
|
-
readonly "openai/gpt-5.
|
|
13968
|
+
readonly "openai/gpt-5.3-chat": {
|
|
13073
13969
|
id: string;
|
|
13074
13970
|
name: string;
|
|
13075
13971
|
api: "anthropic-messages";
|
|
@@ -13086,7 +13982,7 @@ export declare const MODELS: {
|
|
|
13086
13982
|
contextWindow: number;
|
|
13087
13983
|
maxTokens: number;
|
|
13088
13984
|
};
|
|
13089
|
-
readonly "openai/gpt-5.
|
|
13985
|
+
readonly "openai/gpt-5.3-codex": {
|
|
13090
13986
|
id: string;
|
|
13091
13987
|
name: string;
|
|
13092
13988
|
api: "anthropic-messages";
|
|
@@ -13103,7 +13999,7 @@ export declare const MODELS: {
|
|
|
13103
13999
|
contextWindow: number;
|
|
13104
14000
|
maxTokens: number;
|
|
13105
14001
|
};
|
|
13106
|
-
readonly "openai/gpt-5.
|
|
14002
|
+
readonly "openai/gpt-5.4": {
|
|
13107
14003
|
id: string;
|
|
13108
14004
|
name: string;
|
|
13109
14005
|
api: "anthropic-messages";
|
|
@@ -13120,7 +14016,7 @@ export declare const MODELS: {
|
|
|
13120
14016
|
contextWindow: number;
|
|
13121
14017
|
maxTokens: number;
|
|
13122
14018
|
};
|
|
13123
|
-
readonly "openai/gpt-5.
|
|
14019
|
+
readonly "openai/gpt-5.4-mini": {
|
|
13124
14020
|
id: string;
|
|
13125
14021
|
name: string;
|
|
13126
14022
|
api: "anthropic-messages";
|
|
@@ -13137,7 +14033,7 @@ export declare const MODELS: {
|
|
|
13137
14033
|
contextWindow: number;
|
|
13138
14034
|
maxTokens: number;
|
|
13139
14035
|
};
|
|
13140
|
-
readonly "openai/gpt-5.4": {
|
|
14036
|
+
readonly "openai/gpt-5.4-nano": {
|
|
13141
14037
|
id: string;
|
|
13142
14038
|
name: string;
|
|
13143
14039
|
api: "anthropic-messages";
|
|
@@ -13154,7 +14050,7 @@ export declare const MODELS: {
|
|
|
13154
14050
|
contextWindow: number;
|
|
13155
14051
|
maxTokens: number;
|
|
13156
14052
|
};
|
|
13157
|
-
readonly "openai/gpt-5.4-
|
|
14053
|
+
readonly "openai/gpt-5.4-pro": {
|
|
13158
14054
|
id: string;
|
|
13159
14055
|
name: string;
|
|
13160
14056
|
api: "anthropic-messages";
|
|
@@ -13171,7 +14067,7 @@ export declare const MODELS: {
|
|
|
13171
14067
|
contextWindow: number;
|
|
13172
14068
|
maxTokens: number;
|
|
13173
14069
|
};
|
|
13174
|
-
readonly "openai/gpt-5.
|
|
14070
|
+
readonly "openai/gpt-5.5": {
|
|
13175
14071
|
id: string;
|
|
13176
14072
|
name: string;
|
|
13177
14073
|
api: "anthropic-messages";
|
|
@@ -13188,7 +14084,7 @@ export declare const MODELS: {
|
|
|
13188
14084
|
contextWindow: number;
|
|
13189
14085
|
maxTokens: number;
|
|
13190
14086
|
};
|
|
13191
|
-
readonly "openai/gpt-5.
|
|
14087
|
+
readonly "openai/gpt-5.5-pro": {
|
|
13192
14088
|
id: string;
|
|
13193
14089
|
name: string;
|
|
13194
14090
|
api: "anthropic-messages";
|
|
@@ -13375,23 +14271,6 @@ export declare const MODELS: {
|
|
|
13375
14271
|
contextWindow: number;
|
|
13376
14272
|
maxTokens: number;
|
|
13377
14273
|
};
|
|
13378
|
-
readonly "prime-intellect/intellect-3": {
|
|
13379
|
-
id: string;
|
|
13380
|
-
name: string;
|
|
13381
|
-
api: "anthropic-messages";
|
|
13382
|
-
provider: string;
|
|
13383
|
-
baseUrl: string;
|
|
13384
|
-
reasoning: true;
|
|
13385
|
-
input: "text"[];
|
|
13386
|
-
cost: {
|
|
13387
|
-
input: number;
|
|
13388
|
-
output: number;
|
|
13389
|
-
cacheRead: number;
|
|
13390
|
-
cacheWrite: number;
|
|
13391
|
-
};
|
|
13392
|
-
contextWindow: number;
|
|
13393
|
-
maxTokens: number;
|
|
13394
|
-
};
|
|
13395
14274
|
readonly "xai/grok-3": {
|
|
13396
14275
|
id: string;
|
|
13397
14276
|
name: string;
|
|
@@ -13647,6 +14526,23 @@ export declare const MODELS: {
|
|
|
13647
14526
|
contextWindow: number;
|
|
13648
14527
|
maxTokens: number;
|
|
13649
14528
|
};
|
|
14529
|
+
readonly "xai/grok-4.3": {
|
|
14530
|
+
id: string;
|
|
14531
|
+
name: string;
|
|
14532
|
+
api: "anthropic-messages";
|
|
14533
|
+
provider: string;
|
|
14534
|
+
baseUrl: string;
|
|
14535
|
+
reasoning: true;
|
|
14536
|
+
input: ("image" | "text")[];
|
|
14537
|
+
cost: {
|
|
14538
|
+
input: number;
|
|
14539
|
+
output: number;
|
|
14540
|
+
cacheRead: number;
|
|
14541
|
+
cacheWrite: number;
|
|
14542
|
+
};
|
|
14543
|
+
contextWindow: number;
|
|
14544
|
+
maxTokens: number;
|
|
14545
|
+
};
|
|
13650
14546
|
readonly "xai/grok-code-fast-1": {
|
|
13651
14547
|
id: string;
|
|
13652
14548
|
name: string;
|
|
@@ -13698,6 +14594,40 @@ export declare const MODELS: {
|
|
|
13698
14594
|
contextWindow: number;
|
|
13699
14595
|
maxTokens: number;
|
|
13700
14596
|
};
|
|
14597
|
+
readonly "xiaomi/mimo-v2.5": {
|
|
14598
|
+
id: string;
|
|
14599
|
+
name: string;
|
|
14600
|
+
api: "anthropic-messages";
|
|
14601
|
+
provider: string;
|
|
14602
|
+
baseUrl: string;
|
|
14603
|
+
reasoning: true;
|
|
14604
|
+
input: ("image" | "text")[];
|
|
14605
|
+
cost: {
|
|
14606
|
+
input: number;
|
|
14607
|
+
output: number;
|
|
14608
|
+
cacheRead: number;
|
|
14609
|
+
cacheWrite: number;
|
|
14610
|
+
};
|
|
14611
|
+
contextWindow: number;
|
|
14612
|
+
maxTokens: number;
|
|
14613
|
+
};
|
|
14614
|
+
readonly "xiaomi/mimo-v2.5-pro": {
|
|
14615
|
+
id: string;
|
|
14616
|
+
name: string;
|
|
14617
|
+
api: "anthropic-messages";
|
|
14618
|
+
provider: string;
|
|
14619
|
+
baseUrl: string;
|
|
14620
|
+
reasoning: true;
|
|
14621
|
+
input: ("image" | "text")[];
|
|
14622
|
+
cost: {
|
|
14623
|
+
input: number;
|
|
14624
|
+
output: number;
|
|
14625
|
+
cacheRead: number;
|
|
14626
|
+
cacheWrite: number;
|
|
14627
|
+
};
|
|
14628
|
+
contextWindow: number;
|
|
14629
|
+
maxTokens: number;
|
|
14630
|
+
};
|
|
13701
14631
|
readonly "zai/glm-4.5": {
|
|
13702
14632
|
id: string;
|
|
13703
14633
|
name: string;
|
|
@@ -13892,7 +14822,7 @@ export declare const MODELS: {
|
|
|
13892
14822
|
provider: string;
|
|
13893
14823
|
baseUrl: string;
|
|
13894
14824
|
reasoning: true;
|
|
13895
|
-
input:
|
|
14825
|
+
input: "text"[];
|
|
13896
14826
|
cost: {
|
|
13897
14827
|
input: number;
|
|
13898
14828
|
output: number;
|
|
@@ -14278,47 +15208,13 @@ export declare const MODELS: {
|
|
|
14278
15208
|
contextWindow: number;
|
|
14279
15209
|
maxTokens: number;
|
|
14280
15210
|
};
|
|
14281
|
-
readonly "grok-
|
|
14282
|
-
id: string;
|
|
14283
|
-
name: string;
|
|
14284
|
-
api: "openai-completions";
|
|
14285
|
-
provider: string;
|
|
14286
|
-
baseUrl: string;
|
|
14287
|
-
reasoning: false;
|
|
14288
|
-
input: "text"[];
|
|
14289
|
-
cost: {
|
|
14290
|
-
input: number;
|
|
14291
|
-
output: number;
|
|
14292
|
-
cacheRead: number;
|
|
14293
|
-
cacheWrite: number;
|
|
14294
|
-
};
|
|
14295
|
-
contextWindow: number;
|
|
14296
|
-
maxTokens: number;
|
|
14297
|
-
};
|
|
14298
|
-
readonly "grok-code-fast-1": {
|
|
15211
|
+
readonly "grok-4.3": {
|
|
14299
15212
|
id: string;
|
|
14300
15213
|
name: string;
|
|
14301
15214
|
api: "openai-completions";
|
|
14302
15215
|
provider: string;
|
|
14303
15216
|
baseUrl: string;
|
|
14304
15217
|
reasoning: true;
|
|
14305
|
-
input: "text"[];
|
|
14306
|
-
cost: {
|
|
14307
|
-
input: number;
|
|
14308
|
-
output: number;
|
|
14309
|
-
cacheRead: number;
|
|
14310
|
-
cacheWrite: number;
|
|
14311
|
-
};
|
|
14312
|
-
contextWindow: number;
|
|
14313
|
-
maxTokens: number;
|
|
14314
|
-
};
|
|
14315
|
-
readonly "grok-vision-beta": {
|
|
14316
|
-
id: string;
|
|
14317
|
-
name: string;
|
|
14318
|
-
api: "openai-completions";
|
|
14319
|
-
provider: string;
|
|
14320
|
-
baseUrl: string;
|
|
14321
|
-
reasoning: false;
|
|
14322
15218
|
input: ("image" | "text")[];
|
|
14323
15219
|
cost: {
|
|
14324
15220
|
input: number;
|
|
@@ -14329,40 +15225,13 @@ export declare const MODELS: {
|
|
|
14329
15225
|
contextWindow: number;
|
|
14330
15226
|
maxTokens: number;
|
|
14331
15227
|
};
|
|
14332
|
-
|
|
14333
|
-
readonly zai: {
|
|
14334
|
-
readonly "glm-4.5": {
|
|
14335
|
-
id: string;
|
|
14336
|
-
name: string;
|
|
14337
|
-
api: "openai-completions";
|
|
14338
|
-
provider: string;
|
|
14339
|
-
baseUrl: string;
|
|
14340
|
-
compat: {
|
|
14341
|
-
supportsDeveloperRole: false;
|
|
14342
|
-
thinkingFormat: "zai";
|
|
14343
|
-
};
|
|
14344
|
-
reasoning: true;
|
|
14345
|
-
input: "text"[];
|
|
14346
|
-
cost: {
|
|
14347
|
-
input: number;
|
|
14348
|
-
output: number;
|
|
14349
|
-
cacheRead: number;
|
|
14350
|
-
cacheWrite: number;
|
|
14351
|
-
};
|
|
14352
|
-
contextWindow: number;
|
|
14353
|
-
maxTokens: number;
|
|
14354
|
-
};
|
|
14355
|
-
readonly "glm-4.5-air": {
|
|
15228
|
+
readonly "grok-beta": {
|
|
14356
15229
|
id: string;
|
|
14357
15230
|
name: string;
|
|
14358
15231
|
api: "openai-completions";
|
|
14359
15232
|
provider: string;
|
|
14360
15233
|
baseUrl: string;
|
|
14361
|
-
|
|
14362
|
-
supportsDeveloperRole: false;
|
|
14363
|
-
thinkingFormat: "zai";
|
|
14364
|
-
};
|
|
14365
|
-
reasoning: true;
|
|
15234
|
+
reasoning: false;
|
|
14366
15235
|
input: "text"[];
|
|
14367
15236
|
cost: {
|
|
14368
15237
|
input: number;
|
|
@@ -14373,16 +15242,12 @@ export declare const MODELS: {
|
|
|
14373
15242
|
contextWindow: number;
|
|
14374
15243
|
maxTokens: number;
|
|
14375
15244
|
};
|
|
14376
|
-
readonly "
|
|
15245
|
+
readonly "grok-code-fast-1": {
|
|
14377
15246
|
id: string;
|
|
14378
15247
|
name: string;
|
|
14379
15248
|
api: "openai-completions";
|
|
14380
15249
|
provider: string;
|
|
14381
15250
|
baseUrl: string;
|
|
14382
|
-
compat: {
|
|
14383
|
-
supportsDeveloperRole: false;
|
|
14384
|
-
thinkingFormat: "zai";
|
|
14385
|
-
};
|
|
14386
15251
|
reasoning: true;
|
|
14387
15252
|
input: "text"[];
|
|
14388
15253
|
cost: {
|
|
@@ -14394,17 +15259,13 @@ export declare const MODELS: {
|
|
|
14394
15259
|
contextWindow: number;
|
|
14395
15260
|
maxTokens: number;
|
|
14396
15261
|
};
|
|
14397
|
-
readonly "
|
|
15262
|
+
readonly "grok-vision-beta": {
|
|
14398
15263
|
id: string;
|
|
14399
15264
|
name: string;
|
|
14400
15265
|
api: "openai-completions";
|
|
14401
15266
|
provider: string;
|
|
14402
15267
|
baseUrl: string;
|
|
14403
|
-
|
|
14404
|
-
supportsDeveloperRole: false;
|
|
14405
|
-
thinkingFormat: "zai";
|
|
14406
|
-
};
|
|
14407
|
-
reasoning: true;
|
|
15268
|
+
reasoning: false;
|
|
14408
15269
|
input: ("image" | "text")[];
|
|
14409
15270
|
cost: {
|
|
14410
15271
|
input: number;
|
|
@@ -14415,7 +15276,9 @@ export declare const MODELS: {
|
|
|
14415
15276
|
contextWindow: number;
|
|
14416
15277
|
maxTokens: number;
|
|
14417
15278
|
};
|
|
14418
|
-
|
|
15279
|
+
};
|
|
15280
|
+
readonly zai: {
|
|
15281
|
+
readonly "glm-4.5-air": {
|
|
14419
15282
|
id: string;
|
|
14420
15283
|
name: string;
|
|
14421
15284
|
api: "openai-completions";
|
|
@@ -14424,7 +15287,6 @@ export declare const MODELS: {
|
|
|
14424
15287
|
compat: {
|
|
14425
15288
|
supportsDeveloperRole: false;
|
|
14426
15289
|
thinkingFormat: "zai";
|
|
14427
|
-
zaiToolStream: true;
|
|
14428
15290
|
};
|
|
14429
15291
|
reasoning: true;
|
|
14430
15292
|
input: "text"[];
|
|
@@ -14437,28 +15299,6 @@ export declare const MODELS: {
|
|
|
14437
15299
|
contextWindow: number;
|
|
14438
15300
|
maxTokens: number;
|
|
14439
15301
|
};
|
|
14440
|
-
readonly "glm-4.6v": {
|
|
14441
|
-
id: string;
|
|
14442
|
-
name: string;
|
|
14443
|
-
api: "openai-completions";
|
|
14444
|
-
provider: string;
|
|
14445
|
-
baseUrl: string;
|
|
14446
|
-
compat: {
|
|
14447
|
-
supportsDeveloperRole: false;
|
|
14448
|
-
thinkingFormat: "zai";
|
|
14449
|
-
zaiToolStream: true;
|
|
14450
|
-
};
|
|
14451
|
-
reasoning: true;
|
|
14452
|
-
input: ("image" | "text")[];
|
|
14453
|
-
cost: {
|
|
14454
|
-
input: number;
|
|
14455
|
-
output: number;
|
|
14456
|
-
cacheRead: number;
|
|
14457
|
-
cacheWrite: number;
|
|
14458
|
-
};
|
|
14459
|
-
contextWindow: number;
|
|
14460
|
-
maxTokens: number;
|
|
14461
|
-
};
|
|
14462
15302
|
readonly "glm-4.7": {
|
|
14463
15303
|
id: string;
|
|
14464
15304
|
name: string;
|
|
@@ -14481,72 +15321,6 @@ export declare const MODELS: {
|
|
|
14481
15321
|
contextWindow: number;
|
|
14482
15322
|
maxTokens: number;
|
|
14483
15323
|
};
|
|
14484
|
-
readonly "glm-4.7-flash": {
|
|
14485
|
-
id: string;
|
|
14486
|
-
name: string;
|
|
14487
|
-
api: "openai-completions";
|
|
14488
|
-
provider: string;
|
|
14489
|
-
baseUrl: string;
|
|
14490
|
-
compat: {
|
|
14491
|
-
supportsDeveloperRole: false;
|
|
14492
|
-
thinkingFormat: "zai";
|
|
14493
|
-
zaiToolStream: true;
|
|
14494
|
-
};
|
|
14495
|
-
reasoning: true;
|
|
14496
|
-
input: "text"[];
|
|
14497
|
-
cost: {
|
|
14498
|
-
input: number;
|
|
14499
|
-
output: number;
|
|
14500
|
-
cacheRead: number;
|
|
14501
|
-
cacheWrite: number;
|
|
14502
|
-
};
|
|
14503
|
-
contextWindow: number;
|
|
14504
|
-
maxTokens: number;
|
|
14505
|
-
};
|
|
14506
|
-
readonly "glm-4.7-flashx": {
|
|
14507
|
-
id: string;
|
|
14508
|
-
name: string;
|
|
14509
|
-
api: "openai-completions";
|
|
14510
|
-
provider: string;
|
|
14511
|
-
baseUrl: string;
|
|
14512
|
-
compat: {
|
|
14513
|
-
supportsDeveloperRole: false;
|
|
14514
|
-
thinkingFormat: "zai";
|
|
14515
|
-
zaiToolStream: true;
|
|
14516
|
-
};
|
|
14517
|
-
reasoning: true;
|
|
14518
|
-
input: "text"[];
|
|
14519
|
-
cost: {
|
|
14520
|
-
input: number;
|
|
14521
|
-
output: number;
|
|
14522
|
-
cacheRead: number;
|
|
14523
|
-
cacheWrite: number;
|
|
14524
|
-
};
|
|
14525
|
-
contextWindow: number;
|
|
14526
|
-
maxTokens: number;
|
|
14527
|
-
};
|
|
14528
|
-
readonly "glm-5": {
|
|
14529
|
-
id: string;
|
|
14530
|
-
name: string;
|
|
14531
|
-
api: "openai-completions";
|
|
14532
|
-
provider: string;
|
|
14533
|
-
baseUrl: string;
|
|
14534
|
-
compat: {
|
|
14535
|
-
supportsDeveloperRole: false;
|
|
14536
|
-
thinkingFormat: "zai";
|
|
14537
|
-
zaiToolStream: true;
|
|
14538
|
-
};
|
|
14539
|
-
reasoning: true;
|
|
14540
|
-
input: "text"[];
|
|
14541
|
-
cost: {
|
|
14542
|
-
input: number;
|
|
14543
|
-
output: number;
|
|
14544
|
-
cacheRead: number;
|
|
14545
|
-
cacheWrite: number;
|
|
14546
|
-
};
|
|
14547
|
-
contextWindow: number;
|
|
14548
|
-
maxTokens: number;
|
|
14549
|
-
};
|
|
14550
15324
|
readonly "glm-5-turbo": {
|
|
14551
15325
|
id: string;
|
|
14552
15326
|
name: string;
|