@draht/ai 2026.3.14 → 2026.3.25
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +28 -0
- package/dist/cli.js +0 -0
- package/dist/index.d.ts +10 -8
- package/dist/index.d.ts.map +1 -1
- package/dist/index.js +0 -8
- package/dist/index.js.map +1 -1
- package/dist/models.d.ts +1 -1
- package/dist/models.d.ts.map +1 -1
- package/dist/models.generated.d.ts +593 -175
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +781 -371
- package/dist/models.generated.js.map +1 -1
- package/dist/models.js +3 -3
- package/dist/models.js.map +1 -1
- package/dist/providers/amazon-bedrock.d.ts +5 -0
- package/dist/providers/amazon-bedrock.d.ts.map +1 -1
- package/dist/providers/amazon-bedrock.js +32 -9
- package/dist/providers/amazon-bedrock.js.map +1 -1
- package/dist/providers/anthropic.d.ts +7 -0
- package/dist/providers/anthropic.d.ts.map +1 -1
- package/dist/providers/anthropic.js +42 -25
- package/dist/providers/anthropic.js.map +1 -1
- package/dist/providers/azure-openai-responses.d.ts.map +1 -1
- package/dist/providers/azure-openai-responses.js +1 -12
- package/dist/providers/azure-openai-responses.js.map +1 -1
- package/dist/providers/google-gemini-cli.d.ts.map +1 -1
- package/dist/providers/google-gemini-cli.js +19 -0
- package/dist/providers/google-gemini-cli.js.map +1 -1
- package/dist/providers/google-shared.d.ts.map +1 -1
- package/dist/providers/google-shared.js +20 -8
- package/dist/providers/google-shared.js.map +1 -1
- package/dist/providers/google-vertex.d.ts.map +1 -1
- package/dist/providers/google-vertex.js +28 -1
- package/dist/providers/google-vertex.js.map +1 -1
- package/dist/providers/google.d.ts.map +1 -1
- package/dist/providers/google.js +19 -0
- package/dist/providers/google.js.map +1 -1
- package/dist/providers/mistral.d.ts.map +1 -1
- package/dist/providers/mistral.js +3 -0
- package/dist/providers/mistral.js.map +1 -1
- package/dist/providers/openai-codex-responses.d.ts.map +1 -1
- package/dist/providers/openai-codex-responses.js +72 -35
- package/dist/providers/openai-codex-responses.js.map +1 -1
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +53 -16
- package/dist/providers/openai-completions.js.map +1 -1
- package/dist/providers/openai-responses-shared.d.ts.map +1 -1
- package/dist/providers/openai-responses-shared.js +43 -39
- package/dist/providers/openai-responses-shared.js.map +1 -1
- package/dist/providers/openai-responses.d.ts.map +1 -1
- package/dist/providers/openai-responses.js +2 -13
- package/dist/providers/openai-responses.js.map +1 -1
- package/dist/providers/register-builtins.d.ts +28 -1
- package/dist/providers/register-builtins.d.ts.map +1 -1
- package/dist/providers/register-builtins.js +170 -47
- package/dist/providers/register-builtins.js.map +1 -1
- package/dist/types.d.ts +11 -2
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/dist/utils/oauth/anthropic.d.ts +14 -6
- package/dist/utils/oauth/anthropic.d.ts.map +1 -1
- package/dist/utils/oauth/anthropic.js +288 -57
- package/dist/utils/oauth/anthropic.js.map +1 -1
- package/dist/utils/oauth/google-antigravity.d.ts.map +1 -1
- package/dist/utils/oauth/google-antigravity.js +22 -19
- package/dist/utils/oauth/google-antigravity.js.map +1 -1
- package/dist/utils/oauth/google-gemini-cli.d.ts.map +1 -1
- package/dist/utils/oauth/google-gemini-cli.js +22 -19
- package/dist/utils/oauth/google-gemini-cli.js.map +1 -1
- package/dist/utils/oauth/oauth-page.d.ts +3 -0
- package/dist/utils/oauth/oauth-page.d.ts.map +1 -0
- package/dist/utils/oauth/oauth-page.js +105 -0
- package/dist/utils/oauth/oauth-page.js.map +1 -0
- package/dist/utils/oauth/openai-codex.d.ts.map +1 -1
- package/dist/utils/oauth/openai-codex.js +24 -31
- package/dist/utils/oauth/openai-codex.js.map +1 -1
- package/dist/utils/validation.d.ts.map +1 -1
- package/dist/utils/validation.js +17 -9
- package/dist/utils/validation.js.map +1 -1
- package/package.json +39 -5
- package/bedrock-provider.d.ts +0 -1
- package/bedrock-provider.js +0 -1
|
@@ -799,6 +799,23 @@ export declare const MODELS: {
|
|
|
799
799
|
contextWindow: number;
|
|
800
800
|
maxTokens: number;
|
|
801
801
|
};
|
|
802
|
+
readonly "minimax.minimax-m2.5": {
|
|
803
|
+
id: string;
|
|
804
|
+
name: string;
|
|
805
|
+
api: "bedrock-converse-stream";
|
|
806
|
+
provider: string;
|
|
807
|
+
baseUrl: string;
|
|
808
|
+
reasoning: true;
|
|
809
|
+
input: "text"[];
|
|
810
|
+
cost: {
|
|
811
|
+
input: number;
|
|
812
|
+
output: number;
|
|
813
|
+
cacheRead: number;
|
|
814
|
+
cacheWrite: number;
|
|
815
|
+
};
|
|
816
|
+
contextWindow: number;
|
|
817
|
+
maxTokens: number;
|
|
818
|
+
};
|
|
802
819
|
readonly "mistral.devstral-2-123b": {
|
|
803
820
|
id: string;
|
|
804
821
|
name: string;
|
|
@@ -1411,6 +1428,23 @@ export declare const MODELS: {
|
|
|
1411
1428
|
contextWindow: number;
|
|
1412
1429
|
maxTokens: number;
|
|
1413
1430
|
};
|
|
1431
|
+
readonly "zai.glm-5": {
|
|
1432
|
+
id: string;
|
|
1433
|
+
name: string;
|
|
1434
|
+
api: "bedrock-converse-stream";
|
|
1435
|
+
provider: string;
|
|
1436
|
+
baseUrl: string;
|
|
1437
|
+
reasoning: true;
|
|
1438
|
+
input: "text"[];
|
|
1439
|
+
cost: {
|
|
1440
|
+
input: number;
|
|
1441
|
+
output: number;
|
|
1442
|
+
cacheRead: number;
|
|
1443
|
+
cacheWrite: number;
|
|
1444
|
+
};
|
|
1445
|
+
contextWindow: number;
|
|
1446
|
+
maxTokens: number;
|
|
1447
|
+
};
|
|
1414
1448
|
};
|
|
1415
1449
|
readonly anthropic: {
|
|
1416
1450
|
readonly "claude-3-5-haiku-20241022": {
|
|
@@ -2299,6 +2333,40 @@ export declare const MODELS: {
|
|
|
2299
2333
|
contextWindow: number;
|
|
2300
2334
|
maxTokens: number;
|
|
2301
2335
|
};
|
|
2336
|
+
readonly "gpt-5.4-mini": {
|
|
2337
|
+
id: string;
|
|
2338
|
+
name: string;
|
|
2339
|
+
api: "azure-openai-responses";
|
|
2340
|
+
provider: string;
|
|
2341
|
+
baseUrl: string;
|
|
2342
|
+
reasoning: true;
|
|
2343
|
+
input: ("image" | "text")[];
|
|
2344
|
+
cost: {
|
|
2345
|
+
input: number;
|
|
2346
|
+
output: number;
|
|
2347
|
+
cacheRead: number;
|
|
2348
|
+
cacheWrite: number;
|
|
2349
|
+
};
|
|
2350
|
+
contextWindow: number;
|
|
2351
|
+
maxTokens: number;
|
|
2352
|
+
};
|
|
2353
|
+
readonly "gpt-5.4-nano": {
|
|
2354
|
+
id: string;
|
|
2355
|
+
name: string;
|
|
2356
|
+
api: "azure-openai-responses";
|
|
2357
|
+
provider: string;
|
|
2358
|
+
baseUrl: string;
|
|
2359
|
+
reasoning: true;
|
|
2360
|
+
input: ("image" | "text")[];
|
|
2361
|
+
cost: {
|
|
2362
|
+
input: number;
|
|
2363
|
+
output: number;
|
|
2364
|
+
cacheRead: number;
|
|
2365
|
+
cacheWrite: number;
|
|
2366
|
+
};
|
|
2367
|
+
contextWindow: number;
|
|
2368
|
+
maxTokens: number;
|
|
2369
|
+
};
|
|
2302
2370
|
readonly "gpt-5.4-pro": {
|
|
2303
2371
|
id: string;
|
|
2304
2372
|
name: string;
|
|
@@ -3060,6 +3128,29 @@ export declare const MODELS: {
|
|
|
3060
3128
|
contextWindow: number;
|
|
3061
3129
|
maxTokens: number;
|
|
3062
3130
|
};
|
|
3131
|
+
readonly "gpt-5.4-mini": {
|
|
3132
|
+
id: string;
|
|
3133
|
+
name: string;
|
|
3134
|
+
api: "openai-responses";
|
|
3135
|
+
provider: string;
|
|
3136
|
+
baseUrl: string;
|
|
3137
|
+
headers: {
|
|
3138
|
+
"User-Agent": string;
|
|
3139
|
+
"Editor-Version": string;
|
|
3140
|
+
"Editor-Plugin-Version": string;
|
|
3141
|
+
"Copilot-Integration-Id": string;
|
|
3142
|
+
};
|
|
3143
|
+
reasoning: true;
|
|
3144
|
+
input: ("image" | "text")[];
|
|
3145
|
+
cost: {
|
|
3146
|
+
input: number;
|
|
3147
|
+
output: number;
|
|
3148
|
+
cacheRead: number;
|
|
3149
|
+
cacheWrite: number;
|
|
3150
|
+
};
|
|
3151
|
+
contextWindow: number;
|
|
3152
|
+
maxTokens: number;
|
|
3153
|
+
};
|
|
3063
3154
|
readonly "grok-code-fast-1": {
|
|
3064
3155
|
id: string;
|
|
3065
3156
|
name: string;
|
|
@@ -4654,6 +4745,23 @@ export declare const MODELS: {
|
|
|
4654
4745
|
contextWindow: number;
|
|
4655
4746
|
maxTokens: number;
|
|
4656
4747
|
};
|
|
4748
|
+
readonly "MiniMax-M2.1-highspeed": {
|
|
4749
|
+
id: string;
|
|
4750
|
+
name: string;
|
|
4751
|
+
api: "anthropic-messages";
|
|
4752
|
+
provider: string;
|
|
4753
|
+
baseUrl: string;
|
|
4754
|
+
reasoning: true;
|
|
4755
|
+
input: "text"[];
|
|
4756
|
+
cost: {
|
|
4757
|
+
input: number;
|
|
4758
|
+
output: number;
|
|
4759
|
+
cacheRead: number;
|
|
4760
|
+
cacheWrite: number;
|
|
4761
|
+
};
|
|
4762
|
+
contextWindow: number;
|
|
4763
|
+
maxTokens: number;
|
|
4764
|
+
};
|
|
4657
4765
|
readonly "MiniMax-M2.5": {
|
|
4658
4766
|
id: string;
|
|
4659
4767
|
name: string;
|
|
@@ -4688,6 +4796,40 @@ export declare const MODELS: {
|
|
|
4688
4796
|
contextWindow: number;
|
|
4689
4797
|
maxTokens: number;
|
|
4690
4798
|
};
|
|
4799
|
+
readonly "MiniMax-M2.7": {
|
|
4800
|
+
id: string;
|
|
4801
|
+
name: string;
|
|
4802
|
+
api: "anthropic-messages";
|
|
4803
|
+
provider: string;
|
|
4804
|
+
baseUrl: string;
|
|
4805
|
+
reasoning: true;
|
|
4806
|
+
input: "text"[];
|
|
4807
|
+
cost: {
|
|
4808
|
+
input: number;
|
|
4809
|
+
output: number;
|
|
4810
|
+
cacheRead: number;
|
|
4811
|
+
cacheWrite: number;
|
|
4812
|
+
};
|
|
4813
|
+
contextWindow: number;
|
|
4814
|
+
maxTokens: number;
|
|
4815
|
+
};
|
|
4816
|
+
readonly "MiniMax-M2.7-highspeed": {
|
|
4817
|
+
id: string;
|
|
4818
|
+
name: string;
|
|
4819
|
+
api: "anthropic-messages";
|
|
4820
|
+
provider: string;
|
|
4821
|
+
baseUrl: string;
|
|
4822
|
+
reasoning: true;
|
|
4823
|
+
input: "text"[];
|
|
4824
|
+
cost: {
|
|
4825
|
+
input: number;
|
|
4826
|
+
output: number;
|
|
4827
|
+
cacheRead: number;
|
|
4828
|
+
cacheWrite: number;
|
|
4829
|
+
};
|
|
4830
|
+
contextWindow: number;
|
|
4831
|
+
maxTokens: number;
|
|
4832
|
+
};
|
|
4691
4833
|
};
|
|
4692
4834
|
readonly "minimax-cn": {
|
|
4693
4835
|
readonly "MiniMax-M2": {
|
|
@@ -4724,6 +4866,23 @@ export declare const MODELS: {
|
|
|
4724
4866
|
contextWindow: number;
|
|
4725
4867
|
maxTokens: number;
|
|
4726
4868
|
};
|
|
4869
|
+
readonly "MiniMax-M2.1-highspeed": {
|
|
4870
|
+
id: string;
|
|
4871
|
+
name: string;
|
|
4872
|
+
api: "anthropic-messages";
|
|
4873
|
+
provider: string;
|
|
4874
|
+
baseUrl: string;
|
|
4875
|
+
reasoning: true;
|
|
4876
|
+
input: "text"[];
|
|
4877
|
+
cost: {
|
|
4878
|
+
input: number;
|
|
4879
|
+
output: number;
|
|
4880
|
+
cacheRead: number;
|
|
4881
|
+
cacheWrite: number;
|
|
4882
|
+
};
|
|
4883
|
+
contextWindow: number;
|
|
4884
|
+
maxTokens: number;
|
|
4885
|
+
};
|
|
4727
4886
|
readonly "MiniMax-M2.5": {
|
|
4728
4887
|
id: string;
|
|
4729
4888
|
name: string;
|
|
@@ -4758,6 +4917,40 @@ export declare const MODELS: {
|
|
|
4758
4917
|
contextWindow: number;
|
|
4759
4918
|
maxTokens: number;
|
|
4760
4919
|
};
|
|
4920
|
+
readonly "MiniMax-M2.7": {
|
|
4921
|
+
id: string;
|
|
4922
|
+
name: string;
|
|
4923
|
+
api: "anthropic-messages";
|
|
4924
|
+
provider: string;
|
|
4925
|
+
baseUrl: string;
|
|
4926
|
+
reasoning: true;
|
|
4927
|
+
input: "text"[];
|
|
4928
|
+
cost: {
|
|
4929
|
+
input: number;
|
|
4930
|
+
output: number;
|
|
4931
|
+
cacheRead: number;
|
|
4932
|
+
cacheWrite: number;
|
|
4933
|
+
};
|
|
4934
|
+
contextWindow: number;
|
|
4935
|
+
maxTokens: number;
|
|
4936
|
+
};
|
|
4937
|
+
readonly "MiniMax-M2.7-highspeed": {
|
|
4938
|
+
id: string;
|
|
4939
|
+
name: string;
|
|
4940
|
+
api: "anthropic-messages";
|
|
4941
|
+
provider: string;
|
|
4942
|
+
baseUrl: string;
|
|
4943
|
+
reasoning: true;
|
|
4944
|
+
input: "text"[];
|
|
4945
|
+
cost: {
|
|
4946
|
+
input: number;
|
|
4947
|
+
output: number;
|
|
4948
|
+
cacheRead: number;
|
|
4949
|
+
cacheWrite: number;
|
|
4950
|
+
};
|
|
4951
|
+
contextWindow: number;
|
|
4952
|
+
maxTokens: number;
|
|
4953
|
+
};
|
|
4761
4954
|
};
|
|
4762
4955
|
readonly mistral: {
|
|
4763
4956
|
readonly "codestral-latest": {
|
|
@@ -5680,6 +5873,40 @@ export declare const MODELS: {
|
|
|
5680
5873
|
contextWindow: number;
|
|
5681
5874
|
maxTokens: number;
|
|
5682
5875
|
};
|
|
5876
|
+
readonly "gpt-5.4-mini": {
|
|
5877
|
+
id: string;
|
|
5878
|
+
name: string;
|
|
5879
|
+
api: "openai-responses";
|
|
5880
|
+
provider: string;
|
|
5881
|
+
baseUrl: string;
|
|
5882
|
+
reasoning: true;
|
|
5883
|
+
input: ("image" | "text")[];
|
|
5884
|
+
cost: {
|
|
5885
|
+
input: number;
|
|
5886
|
+
output: number;
|
|
5887
|
+
cacheRead: number;
|
|
5888
|
+
cacheWrite: number;
|
|
5889
|
+
};
|
|
5890
|
+
contextWindow: number;
|
|
5891
|
+
maxTokens: number;
|
|
5892
|
+
};
|
|
5893
|
+
readonly "gpt-5.4-nano": {
|
|
5894
|
+
id: string;
|
|
5895
|
+
name: string;
|
|
5896
|
+
api: "openai-responses";
|
|
5897
|
+
provider: string;
|
|
5898
|
+
baseUrl: string;
|
|
5899
|
+
reasoning: true;
|
|
5900
|
+
input: ("image" | "text")[];
|
|
5901
|
+
cost: {
|
|
5902
|
+
input: number;
|
|
5903
|
+
output: number;
|
|
5904
|
+
cacheRead: number;
|
|
5905
|
+
cacheWrite: number;
|
|
5906
|
+
};
|
|
5907
|
+
contextWindow: number;
|
|
5908
|
+
maxTokens: number;
|
|
5909
|
+
};
|
|
5683
5910
|
readonly "gpt-5.4-pro": {
|
|
5684
5911
|
id: string;
|
|
5685
5912
|
name: string;
|
|
@@ -5971,6 +6198,23 @@ export declare const MODELS: {
|
|
|
5971
6198
|
contextWindow: number;
|
|
5972
6199
|
maxTokens: number;
|
|
5973
6200
|
};
|
|
6201
|
+
readonly "gpt-5.4-mini": {
|
|
6202
|
+
id: string;
|
|
6203
|
+
name: string;
|
|
6204
|
+
api: "openai-codex-responses";
|
|
6205
|
+
provider: string;
|
|
6206
|
+
baseUrl: string;
|
|
6207
|
+
reasoning: true;
|
|
6208
|
+
input: ("image" | "text")[];
|
|
6209
|
+
cost: {
|
|
6210
|
+
input: number;
|
|
6211
|
+
output: number;
|
|
6212
|
+
cacheRead: number;
|
|
6213
|
+
cacheWrite: number;
|
|
6214
|
+
};
|
|
6215
|
+
contextWindow: number;
|
|
6216
|
+
maxTokens: number;
|
|
6217
|
+
};
|
|
5974
6218
|
};
|
|
5975
6219
|
readonly opencode: {
|
|
5976
6220
|
readonly "big-pickle": {
|
|
@@ -6143,7 +6387,7 @@ export declare const MODELS: {
|
|
|
6143
6387
|
contextWindow: number;
|
|
6144
6388
|
maxTokens: number;
|
|
6145
6389
|
};
|
|
6146
|
-
readonly "gemini-3-pro": {
|
|
6390
|
+
readonly "gemini-3.1-pro": {
|
|
6147
6391
|
id: string;
|
|
6148
6392
|
name: string;
|
|
6149
6393
|
api: "google-generative-ai";
|
|
@@ -6160,14 +6404,14 @@ export declare const MODELS: {
|
|
|
6160
6404
|
contextWindow: number;
|
|
6161
6405
|
maxTokens: number;
|
|
6162
6406
|
};
|
|
6163
|
-
readonly "
|
|
6407
|
+
readonly "glm-5": {
|
|
6164
6408
|
id: string;
|
|
6165
6409
|
name: string;
|
|
6166
|
-
api: "
|
|
6410
|
+
api: "openai-completions";
|
|
6167
6411
|
provider: string;
|
|
6168
6412
|
baseUrl: string;
|
|
6169
6413
|
reasoning: true;
|
|
6170
|
-
input:
|
|
6414
|
+
input: "text"[];
|
|
6171
6415
|
cost: {
|
|
6172
6416
|
input: number;
|
|
6173
6417
|
output: number;
|
|
@@ -6177,14 +6421,14 @@ export declare const MODELS: {
|
|
|
6177
6421
|
contextWindow: number;
|
|
6178
6422
|
maxTokens: number;
|
|
6179
6423
|
};
|
|
6180
|
-
readonly "
|
|
6424
|
+
readonly "gpt-5": {
|
|
6181
6425
|
id: string;
|
|
6182
6426
|
name: string;
|
|
6183
|
-
api: "openai-
|
|
6427
|
+
api: "openai-responses";
|
|
6184
6428
|
provider: string;
|
|
6185
6429
|
baseUrl: string;
|
|
6186
6430
|
reasoning: true;
|
|
6187
|
-
input: "text"[];
|
|
6431
|
+
input: ("image" | "text")[];
|
|
6188
6432
|
cost: {
|
|
6189
6433
|
input: number;
|
|
6190
6434
|
output: number;
|
|
@@ -6194,14 +6438,14 @@ export declare const MODELS: {
|
|
|
6194
6438
|
contextWindow: number;
|
|
6195
6439
|
maxTokens: number;
|
|
6196
6440
|
};
|
|
6197
|
-
readonly "
|
|
6441
|
+
readonly "gpt-5-codex": {
|
|
6198
6442
|
id: string;
|
|
6199
6443
|
name: string;
|
|
6200
|
-
api: "openai-
|
|
6444
|
+
api: "openai-responses";
|
|
6201
6445
|
provider: string;
|
|
6202
6446
|
baseUrl: string;
|
|
6203
6447
|
reasoning: true;
|
|
6204
|
-
input: "text"[];
|
|
6448
|
+
input: ("image" | "text")[];
|
|
6205
6449
|
cost: {
|
|
6206
6450
|
input: number;
|
|
6207
6451
|
output: number;
|
|
@@ -6211,14 +6455,14 @@ export declare const MODELS: {
|
|
|
6211
6455
|
contextWindow: number;
|
|
6212
6456
|
maxTokens: number;
|
|
6213
6457
|
};
|
|
6214
|
-
readonly "
|
|
6458
|
+
readonly "gpt-5-nano": {
|
|
6215
6459
|
id: string;
|
|
6216
6460
|
name: string;
|
|
6217
|
-
api: "openai-
|
|
6461
|
+
api: "openai-responses";
|
|
6218
6462
|
provider: string;
|
|
6219
6463
|
baseUrl: string;
|
|
6220
6464
|
reasoning: true;
|
|
6221
|
-
input: "text"[];
|
|
6465
|
+
input: ("image" | "text")[];
|
|
6222
6466
|
cost: {
|
|
6223
6467
|
input: number;
|
|
6224
6468
|
output: number;
|
|
@@ -6228,7 +6472,7 @@ export declare const MODELS: {
|
|
|
6228
6472
|
contextWindow: number;
|
|
6229
6473
|
maxTokens: number;
|
|
6230
6474
|
};
|
|
6231
|
-
readonly "gpt-5": {
|
|
6475
|
+
readonly "gpt-5.1": {
|
|
6232
6476
|
id: string;
|
|
6233
6477
|
name: string;
|
|
6234
6478
|
api: "openai-responses";
|
|
@@ -6245,7 +6489,7 @@ export declare const MODELS: {
|
|
|
6245
6489
|
contextWindow: number;
|
|
6246
6490
|
maxTokens: number;
|
|
6247
6491
|
};
|
|
6248
|
-
readonly "gpt-5-codex": {
|
|
6492
|
+
readonly "gpt-5.1-codex": {
|
|
6249
6493
|
id: string;
|
|
6250
6494
|
name: string;
|
|
6251
6495
|
api: "openai-responses";
|
|
@@ -6262,7 +6506,7 @@ export declare const MODELS: {
|
|
|
6262
6506
|
contextWindow: number;
|
|
6263
6507
|
maxTokens: number;
|
|
6264
6508
|
};
|
|
6265
|
-
readonly "gpt-5-
|
|
6509
|
+
readonly "gpt-5.1-codex-max": {
|
|
6266
6510
|
id: string;
|
|
6267
6511
|
name: string;
|
|
6268
6512
|
api: "openai-responses";
|
|
@@ -6279,7 +6523,7 @@ export declare const MODELS: {
|
|
|
6279
6523
|
contextWindow: number;
|
|
6280
6524
|
maxTokens: number;
|
|
6281
6525
|
};
|
|
6282
|
-
readonly "gpt-5.1": {
|
|
6526
|
+
readonly "gpt-5.1-codex-mini": {
|
|
6283
6527
|
id: string;
|
|
6284
6528
|
name: string;
|
|
6285
6529
|
api: "openai-responses";
|
|
@@ -6296,7 +6540,7 @@ export declare const MODELS: {
|
|
|
6296
6540
|
contextWindow: number;
|
|
6297
6541
|
maxTokens: number;
|
|
6298
6542
|
};
|
|
6299
|
-
readonly "gpt-5.
|
|
6543
|
+
readonly "gpt-5.2": {
|
|
6300
6544
|
id: string;
|
|
6301
6545
|
name: string;
|
|
6302
6546
|
api: "openai-responses";
|
|
@@ -6313,7 +6557,7 @@ export declare const MODELS: {
|
|
|
6313
6557
|
contextWindow: number;
|
|
6314
6558
|
maxTokens: number;
|
|
6315
6559
|
};
|
|
6316
|
-
readonly "gpt-5.
|
|
6560
|
+
readonly "gpt-5.2-codex": {
|
|
6317
6561
|
id: string;
|
|
6318
6562
|
name: string;
|
|
6319
6563
|
api: "openai-responses";
|
|
@@ -6330,7 +6574,7 @@ export declare const MODELS: {
|
|
|
6330
6574
|
contextWindow: number;
|
|
6331
6575
|
maxTokens: number;
|
|
6332
6576
|
};
|
|
6333
|
-
readonly "gpt-5.
|
|
6577
|
+
readonly "gpt-5.3-codex": {
|
|
6334
6578
|
id: string;
|
|
6335
6579
|
name: string;
|
|
6336
6580
|
api: "openai-responses";
|
|
@@ -6347,7 +6591,7 @@ export declare const MODELS: {
|
|
|
6347
6591
|
contextWindow: number;
|
|
6348
6592
|
maxTokens: number;
|
|
6349
6593
|
};
|
|
6350
|
-
readonly "gpt-5.
|
|
6594
|
+
readonly "gpt-5.4": {
|
|
6351
6595
|
id: string;
|
|
6352
6596
|
name: string;
|
|
6353
6597
|
api: "openai-responses";
|
|
@@ -6364,7 +6608,7 @@ export declare const MODELS: {
|
|
|
6364
6608
|
contextWindow: number;
|
|
6365
6609
|
maxTokens: number;
|
|
6366
6610
|
};
|
|
6367
|
-
readonly "gpt-5.
|
|
6611
|
+
readonly "gpt-5.4-mini": {
|
|
6368
6612
|
id: string;
|
|
6369
6613
|
name: string;
|
|
6370
6614
|
api: "openai-responses";
|
|
@@ -6381,7 +6625,7 @@ export declare const MODELS: {
|
|
|
6381
6625
|
contextWindow: number;
|
|
6382
6626
|
maxTokens: number;
|
|
6383
6627
|
};
|
|
6384
|
-
readonly "gpt-5.
|
|
6628
|
+
readonly "gpt-5.4-nano": {
|
|
6385
6629
|
id: string;
|
|
6386
6630
|
name: string;
|
|
6387
6631
|
api: "openai-responses";
|
|
@@ -6398,7 +6642,7 @@ export declare const MODELS: {
|
|
|
6398
6642
|
contextWindow: number;
|
|
6399
6643
|
maxTokens: number;
|
|
6400
6644
|
};
|
|
6401
|
-
readonly "gpt-5.4": {
|
|
6645
|
+
readonly "gpt-5.4-pro": {
|
|
6402
6646
|
id: string;
|
|
6403
6647
|
name: string;
|
|
6404
6648
|
api: "openai-responses";
|
|
@@ -6415,10 +6659,10 @@ export declare const MODELS: {
|
|
|
6415
6659
|
contextWindow: number;
|
|
6416
6660
|
maxTokens: number;
|
|
6417
6661
|
};
|
|
6418
|
-
readonly "
|
|
6662
|
+
readonly "kimi-k2.5": {
|
|
6419
6663
|
id: string;
|
|
6420
6664
|
name: string;
|
|
6421
|
-
api: "openai-
|
|
6665
|
+
api: "openai-completions";
|
|
6422
6666
|
provider: string;
|
|
6423
6667
|
baseUrl: string;
|
|
6424
6668
|
reasoning: true;
|
|
@@ -6432,7 +6676,7 @@ export declare const MODELS: {
|
|
|
6432
6676
|
contextWindow: number;
|
|
6433
6677
|
maxTokens: number;
|
|
6434
6678
|
};
|
|
6435
|
-
readonly "
|
|
6679
|
+
readonly "mimo-v2-omni-free": {
|
|
6436
6680
|
id: string;
|
|
6437
6681
|
name: string;
|
|
6438
6682
|
api: "openai-completions";
|
|
@@ -6449,7 +6693,7 @@ export declare const MODELS: {
|
|
|
6449
6693
|
contextWindow: number;
|
|
6450
6694
|
maxTokens: number;
|
|
6451
6695
|
};
|
|
6452
|
-
readonly "mimo-v2-
|
|
6696
|
+
readonly "mimo-v2-pro-free": {
|
|
6453
6697
|
id: string;
|
|
6454
6698
|
name: string;
|
|
6455
6699
|
api: "openai-completions";
|
|
@@ -6466,7 +6710,7 @@ export declare const MODELS: {
|
|
|
6466
6710
|
contextWindow: number;
|
|
6467
6711
|
maxTokens: number;
|
|
6468
6712
|
};
|
|
6469
|
-
readonly "minimax-m2.
|
|
6713
|
+
readonly "minimax-m2.5": {
|
|
6470
6714
|
id: string;
|
|
6471
6715
|
name: string;
|
|
6472
6716
|
api: "openai-completions";
|
|
@@ -6483,10 +6727,10 @@ export declare const MODELS: {
|
|
|
6483
6727
|
contextWindow: number;
|
|
6484
6728
|
maxTokens: number;
|
|
6485
6729
|
};
|
|
6486
|
-
readonly "minimax-m2.5": {
|
|
6730
|
+
readonly "minimax-m2.5-free": {
|
|
6487
6731
|
id: string;
|
|
6488
6732
|
name: string;
|
|
6489
|
-
api: "
|
|
6733
|
+
api: "anthropic-messages";
|
|
6490
6734
|
provider: string;
|
|
6491
6735
|
baseUrl: string;
|
|
6492
6736
|
reasoning: true;
|
|
@@ -6500,10 +6744,10 @@ export declare const MODELS: {
|
|
|
6500
6744
|
contextWindow: number;
|
|
6501
6745
|
maxTokens: number;
|
|
6502
6746
|
};
|
|
6503
|
-
readonly "
|
|
6747
|
+
readonly "nemotron-3-super-free": {
|
|
6504
6748
|
id: string;
|
|
6505
6749
|
name: string;
|
|
6506
|
-
api: "
|
|
6750
|
+
api: "openai-completions";
|
|
6507
6751
|
provider: string;
|
|
6508
6752
|
baseUrl: string;
|
|
6509
6753
|
reasoning: true;
|
|
@@ -6517,7 +6761,9 @@ export declare const MODELS: {
|
|
|
6517
6761
|
contextWindow: number;
|
|
6518
6762
|
maxTokens: number;
|
|
6519
6763
|
};
|
|
6520
|
-
|
|
6764
|
+
};
|
|
6765
|
+
readonly "opencode-go": {
|
|
6766
|
+
readonly "glm-5": {
|
|
6521
6767
|
id: string;
|
|
6522
6768
|
name: string;
|
|
6523
6769
|
api: "openai-completions";
|
|
@@ -6534,16 +6780,14 @@ export declare const MODELS: {
|
|
|
6534
6780
|
contextWindow: number;
|
|
6535
6781
|
maxTokens: number;
|
|
6536
6782
|
};
|
|
6537
|
-
|
|
6538
|
-
readonly "opencode-go": {
|
|
6539
|
-
readonly "glm-5": {
|
|
6783
|
+
readonly "kimi-k2.5": {
|
|
6540
6784
|
id: string;
|
|
6541
6785
|
name: string;
|
|
6542
6786
|
api: "openai-completions";
|
|
6543
6787
|
provider: string;
|
|
6544
6788
|
baseUrl: string;
|
|
6545
6789
|
reasoning: true;
|
|
6546
|
-
input: "text"[];
|
|
6790
|
+
input: ("image" | "text")[];
|
|
6547
6791
|
cost: {
|
|
6548
6792
|
input: number;
|
|
6549
6793
|
output: number;
|
|
@@ -6553,14 +6797,14 @@ export declare const MODELS: {
|
|
|
6553
6797
|
contextWindow: number;
|
|
6554
6798
|
maxTokens: number;
|
|
6555
6799
|
};
|
|
6556
|
-
readonly "
|
|
6800
|
+
readonly "minimax-m2.5": {
|
|
6557
6801
|
id: string;
|
|
6558
6802
|
name: string;
|
|
6559
|
-
api: "
|
|
6803
|
+
api: "anthropic-messages";
|
|
6560
6804
|
provider: string;
|
|
6561
6805
|
baseUrl: string;
|
|
6562
6806
|
reasoning: true;
|
|
6563
|
-
input:
|
|
6807
|
+
input: "text"[];
|
|
6564
6808
|
cost: {
|
|
6565
6809
|
input: number;
|
|
6566
6810
|
output: number;
|
|
@@ -6570,7 +6814,7 @@ export declare const MODELS: {
|
|
|
6570
6814
|
contextWindow: number;
|
|
6571
6815
|
maxTokens: number;
|
|
6572
6816
|
};
|
|
6573
|
-
readonly "minimax-m2.
|
|
6817
|
+
readonly "minimax-m2.7": {
|
|
6574
6818
|
id: string;
|
|
6575
6819
|
name: string;
|
|
6576
6820
|
api: "anthropic-messages";
|
|
@@ -7541,14 +7785,14 @@ export declare const MODELS: {
|
|
|
7541
7785
|
contextWindow: number;
|
|
7542
7786
|
maxTokens: number;
|
|
7543
7787
|
};
|
|
7544
|
-
readonly "
|
|
7788
|
+
readonly "inception/mercury": {
|
|
7545
7789
|
id: string;
|
|
7546
7790
|
name: string;
|
|
7547
7791
|
api: "openai-completions";
|
|
7548
7792
|
provider: string;
|
|
7549
7793
|
baseUrl: string;
|
|
7550
7794
|
reasoning: false;
|
|
7551
|
-
input:
|
|
7795
|
+
input: "text"[];
|
|
7552
7796
|
cost: {
|
|
7553
7797
|
input: number;
|
|
7554
7798
|
output: number;
|
|
@@ -7558,14 +7802,14 @@ export declare const MODELS: {
|
|
|
7558
7802
|
contextWindow: number;
|
|
7559
7803
|
maxTokens: number;
|
|
7560
7804
|
};
|
|
7561
|
-
readonly "
|
|
7805
|
+
readonly "inception/mercury-2": {
|
|
7562
7806
|
id: string;
|
|
7563
7807
|
name: string;
|
|
7564
7808
|
api: "openai-completions";
|
|
7565
7809
|
provider: string;
|
|
7566
7810
|
baseUrl: string;
|
|
7567
|
-
reasoning:
|
|
7568
|
-
input:
|
|
7811
|
+
reasoning: true;
|
|
7812
|
+
input: "text"[];
|
|
7569
7813
|
cost: {
|
|
7570
7814
|
input: number;
|
|
7571
7815
|
output: number;
|
|
@@ -7575,7 +7819,7 @@ export declare const MODELS: {
|
|
|
7575
7819
|
contextWindow: number;
|
|
7576
7820
|
maxTokens: number;
|
|
7577
7821
|
};
|
|
7578
|
-
readonly "inception/mercury": {
|
|
7822
|
+
readonly "inception/mercury-coder": {
|
|
7579
7823
|
id: string;
|
|
7580
7824
|
name: string;
|
|
7581
7825
|
api: "openai-completions";
|
|
@@ -7592,13 +7836,13 @@ export declare const MODELS: {
|
|
|
7592
7836
|
contextWindow: number;
|
|
7593
7837
|
maxTokens: number;
|
|
7594
7838
|
};
|
|
7595
|
-
readonly "
|
|
7839
|
+
readonly "kwaipilot/kat-coder-pro": {
|
|
7596
7840
|
id: string;
|
|
7597
7841
|
name: string;
|
|
7598
7842
|
api: "openai-completions";
|
|
7599
7843
|
provider: string;
|
|
7600
7844
|
baseUrl: string;
|
|
7601
|
-
reasoning:
|
|
7845
|
+
reasoning: false;
|
|
7602
7846
|
input: "text"[];
|
|
7603
7847
|
cost: {
|
|
7604
7848
|
input: number;
|
|
@@ -7609,7 +7853,7 @@ export declare const MODELS: {
|
|
|
7609
7853
|
contextWindow: number;
|
|
7610
7854
|
maxTokens: number;
|
|
7611
7855
|
};
|
|
7612
|
-
readonly "
|
|
7856
|
+
readonly "meituan/longcat-flash-chat": {
|
|
7613
7857
|
id: string;
|
|
7614
7858
|
name: string;
|
|
7615
7859
|
api: "openai-completions";
|
|
@@ -7626,7 +7870,7 @@ export declare const MODELS: {
|
|
|
7626
7870
|
contextWindow: number;
|
|
7627
7871
|
maxTokens: number;
|
|
7628
7872
|
};
|
|
7629
|
-
readonly "
|
|
7873
|
+
readonly "meta-llama/llama-3-8b-instruct": {
|
|
7630
7874
|
id: string;
|
|
7631
7875
|
name: string;
|
|
7632
7876
|
api: "openai-completions";
|
|
@@ -7643,7 +7887,7 @@ export declare const MODELS: {
|
|
|
7643
7887
|
contextWindow: number;
|
|
7644
7888
|
maxTokens: number;
|
|
7645
7889
|
};
|
|
7646
|
-
readonly "
|
|
7890
|
+
readonly "meta-llama/llama-3.1-70b-instruct": {
|
|
7647
7891
|
id: string;
|
|
7648
7892
|
name: string;
|
|
7649
7893
|
api: "openai-completions";
|
|
@@ -7660,7 +7904,7 @@ export declare const MODELS: {
|
|
|
7660
7904
|
contextWindow: number;
|
|
7661
7905
|
maxTokens: number;
|
|
7662
7906
|
};
|
|
7663
|
-
readonly "meta-llama/llama-3-8b-instruct": {
|
|
7907
|
+
readonly "meta-llama/llama-3.1-8b-instruct": {
|
|
7664
7908
|
id: string;
|
|
7665
7909
|
name: string;
|
|
7666
7910
|
api: "openai-completions";
|
|
@@ -7677,7 +7921,7 @@ export declare const MODELS: {
|
|
|
7677
7921
|
contextWindow: number;
|
|
7678
7922
|
maxTokens: number;
|
|
7679
7923
|
};
|
|
7680
|
-
readonly "meta-llama/llama-3.
|
|
7924
|
+
readonly "meta-llama/llama-3.3-70b-instruct": {
|
|
7681
7925
|
id: string;
|
|
7682
7926
|
name: string;
|
|
7683
7927
|
api: "openai-completions";
|
|
@@ -7694,7 +7938,7 @@ export declare const MODELS: {
|
|
|
7694
7938
|
contextWindow: number;
|
|
7695
7939
|
maxTokens: number;
|
|
7696
7940
|
};
|
|
7697
|
-
readonly "meta-llama/llama-3.
|
|
7941
|
+
readonly "meta-llama/llama-3.3-70b-instruct:free": {
|
|
7698
7942
|
id: string;
|
|
7699
7943
|
name: string;
|
|
7700
7944
|
api: "openai-completions";
|
|
@@ -7711,14 +7955,14 @@ export declare const MODELS: {
|
|
|
7711
7955
|
contextWindow: number;
|
|
7712
7956
|
maxTokens: number;
|
|
7713
7957
|
};
|
|
7714
|
-
readonly "meta-llama/llama-
|
|
7958
|
+
readonly "meta-llama/llama-4-maverick": {
|
|
7715
7959
|
id: string;
|
|
7716
7960
|
name: string;
|
|
7717
7961
|
api: "openai-completions";
|
|
7718
7962
|
provider: string;
|
|
7719
7963
|
baseUrl: string;
|
|
7720
7964
|
reasoning: false;
|
|
7721
|
-
input: "text"[];
|
|
7965
|
+
input: ("image" | "text")[];
|
|
7722
7966
|
cost: {
|
|
7723
7967
|
input: number;
|
|
7724
7968
|
output: number;
|
|
@@ -7728,14 +7972,14 @@ export declare const MODELS: {
|
|
|
7728
7972
|
contextWindow: number;
|
|
7729
7973
|
maxTokens: number;
|
|
7730
7974
|
};
|
|
7731
|
-
readonly "meta-llama/llama-
|
|
7975
|
+
readonly "meta-llama/llama-4-scout": {
|
|
7732
7976
|
id: string;
|
|
7733
7977
|
name: string;
|
|
7734
7978
|
api: "openai-completions";
|
|
7735
7979
|
provider: string;
|
|
7736
7980
|
baseUrl: string;
|
|
7737
7981
|
reasoning: false;
|
|
7738
|
-
input: "text"[];
|
|
7982
|
+
input: ("image" | "text")[];
|
|
7739
7983
|
cost: {
|
|
7740
7984
|
input: number;
|
|
7741
7985
|
output: number;
|
|
@@ -7745,14 +7989,14 @@ export declare const MODELS: {
|
|
|
7745
7989
|
contextWindow: number;
|
|
7746
7990
|
maxTokens: number;
|
|
7747
7991
|
};
|
|
7748
|
-
readonly "
|
|
7992
|
+
readonly "minimax/minimax-m1": {
|
|
7749
7993
|
id: string;
|
|
7750
7994
|
name: string;
|
|
7751
7995
|
api: "openai-completions";
|
|
7752
7996
|
provider: string;
|
|
7753
7997
|
baseUrl: string;
|
|
7754
|
-
reasoning:
|
|
7755
|
-
input:
|
|
7998
|
+
reasoning: true;
|
|
7999
|
+
input: "text"[];
|
|
7756
8000
|
cost: {
|
|
7757
8001
|
input: number;
|
|
7758
8002
|
output: number;
|
|
@@ -7762,14 +8006,14 @@ export declare const MODELS: {
|
|
|
7762
8006
|
contextWindow: number;
|
|
7763
8007
|
maxTokens: number;
|
|
7764
8008
|
};
|
|
7765
|
-
readonly "
|
|
8009
|
+
readonly "minimax/minimax-m2": {
|
|
7766
8010
|
id: string;
|
|
7767
8011
|
name: string;
|
|
7768
8012
|
api: "openai-completions";
|
|
7769
8013
|
provider: string;
|
|
7770
8014
|
baseUrl: string;
|
|
7771
|
-
reasoning:
|
|
7772
|
-
input:
|
|
8015
|
+
reasoning: true;
|
|
8016
|
+
input: "text"[];
|
|
7773
8017
|
cost: {
|
|
7774
8018
|
input: number;
|
|
7775
8019
|
output: number;
|
|
@@ -7779,7 +8023,7 @@ export declare const MODELS: {
|
|
|
7779
8023
|
contextWindow: number;
|
|
7780
8024
|
maxTokens: number;
|
|
7781
8025
|
};
|
|
7782
|
-
readonly "minimax/minimax-
|
|
8026
|
+
readonly "minimax/minimax-m2.1": {
|
|
7783
8027
|
id: string;
|
|
7784
8028
|
name: string;
|
|
7785
8029
|
api: "openai-completions";
|
|
@@ -7796,7 +8040,7 @@ export declare const MODELS: {
|
|
|
7796
8040
|
contextWindow: number;
|
|
7797
8041
|
maxTokens: number;
|
|
7798
8042
|
};
|
|
7799
|
-
readonly "minimax/minimax-m2": {
|
|
8043
|
+
readonly "minimax/minimax-m2.5": {
|
|
7800
8044
|
id: string;
|
|
7801
8045
|
name: string;
|
|
7802
8046
|
api: "openai-completions";
|
|
@@ -7813,7 +8057,7 @@ export declare const MODELS: {
|
|
|
7813
8057
|
contextWindow: number;
|
|
7814
8058
|
maxTokens: number;
|
|
7815
8059
|
};
|
|
7816
|
-
readonly "minimax/minimax-m2.
|
|
8060
|
+
readonly "minimax/minimax-m2.5:free": {
|
|
7817
8061
|
id: string;
|
|
7818
8062
|
name: string;
|
|
7819
8063
|
api: "openai-completions";
|
|
@@ -7830,7 +8074,7 @@ export declare const MODELS: {
|
|
|
7830
8074
|
contextWindow: number;
|
|
7831
8075
|
maxTokens: number;
|
|
7832
8076
|
};
|
|
7833
|
-
readonly "minimax/minimax-m2.
|
|
8077
|
+
readonly "minimax/minimax-m2.7": {
|
|
7834
8078
|
id: string;
|
|
7835
8079
|
name: string;
|
|
7836
8080
|
api: "openai-completions";
|
|
@@ -8119,6 +8363,23 @@ export declare const MODELS: {
|
|
|
8119
8363
|
contextWindow: number;
|
|
8120
8364
|
maxTokens: number;
|
|
8121
8365
|
};
|
|
8366
|
+
readonly "mistralai/mistral-small-2603": {
|
|
8367
|
+
id: string;
|
|
8368
|
+
name: string;
|
|
8369
|
+
api: "openai-completions";
|
|
8370
|
+
provider: string;
|
|
8371
|
+
baseUrl: string;
|
|
8372
|
+
reasoning: true;
|
|
8373
|
+
input: ("image" | "text")[];
|
|
8374
|
+
cost: {
|
|
8375
|
+
input: number;
|
|
8376
|
+
output: number;
|
|
8377
|
+
cacheRead: number;
|
|
8378
|
+
cacheWrite: number;
|
|
8379
|
+
};
|
|
8380
|
+
contextWindow: number;
|
|
8381
|
+
maxTokens: number;
|
|
8382
|
+
};
|
|
8122
8383
|
readonly "mistralai/mistral-small-3.1-24b-instruct:free": {
|
|
8123
8384
|
id: string;
|
|
8124
8385
|
name: string;
|
|
@@ -8391,6 +8652,23 @@ export declare const MODELS: {
|
|
|
8391
8652
|
contextWindow: number;
|
|
8392
8653
|
maxTokens: number;
|
|
8393
8654
|
};
|
|
8655
|
+
readonly "nvidia/nemotron-3-super-120b-a12b": {
|
|
8656
|
+
id: string;
|
|
8657
|
+
name: string;
|
|
8658
|
+
api: "openai-completions";
|
|
8659
|
+
provider: string;
|
|
8660
|
+
baseUrl: string;
|
|
8661
|
+
reasoning: true;
|
|
8662
|
+
input: "text"[];
|
|
8663
|
+
cost: {
|
|
8664
|
+
input: number;
|
|
8665
|
+
output: number;
|
|
8666
|
+
cacheRead: number;
|
|
8667
|
+
cacheWrite: number;
|
|
8668
|
+
};
|
|
8669
|
+
contextWindow: number;
|
|
8670
|
+
maxTokens: number;
|
|
8671
|
+
};
|
|
8394
8672
|
readonly "nvidia/nemotron-3-super-120b-a12b:free": {
|
|
8395
8673
|
id: string;
|
|
8396
8674
|
name: string;
|
|
@@ -9105,7 +9383,7 @@ export declare const MODELS: {
|
|
|
9105
9383
|
contextWindow: number;
|
|
9106
9384
|
maxTokens: number;
|
|
9107
9385
|
};
|
|
9108
|
-
readonly "openai/gpt-5.4-
|
|
9386
|
+
readonly "openai/gpt-5.4-mini": {
|
|
9109
9387
|
id: string;
|
|
9110
9388
|
name: string;
|
|
9111
9389
|
api: "openai-completions";
|
|
@@ -9122,14 +9400,14 @@ export declare const MODELS: {
|
|
|
9122
9400
|
contextWindow: number;
|
|
9123
9401
|
maxTokens: number;
|
|
9124
9402
|
};
|
|
9125
|
-
readonly "openai/gpt-
|
|
9403
|
+
readonly "openai/gpt-5.4-nano": {
|
|
9126
9404
|
id: string;
|
|
9127
9405
|
name: string;
|
|
9128
9406
|
api: "openai-completions";
|
|
9129
9407
|
provider: string;
|
|
9130
9408
|
baseUrl: string;
|
|
9131
9409
|
reasoning: true;
|
|
9132
|
-
input: "text"[];
|
|
9410
|
+
input: ("image" | "text")[];
|
|
9133
9411
|
cost: {
|
|
9134
9412
|
input: number;
|
|
9135
9413
|
output: number;
|
|
@@ -9139,14 +9417,14 @@ export declare const MODELS: {
|
|
|
9139
9417
|
contextWindow: number;
|
|
9140
9418
|
maxTokens: number;
|
|
9141
9419
|
};
|
|
9142
|
-
readonly "openai/gpt-
|
|
9420
|
+
readonly "openai/gpt-5.4-pro": {
|
|
9143
9421
|
id: string;
|
|
9144
9422
|
name: string;
|
|
9145
9423
|
api: "openai-completions";
|
|
9146
9424
|
provider: string;
|
|
9147
9425
|
baseUrl: string;
|
|
9148
9426
|
reasoning: true;
|
|
9149
|
-
input: "text"[];
|
|
9427
|
+
input: ("image" | "text")[];
|
|
9150
9428
|
cost: {
|
|
9151
9429
|
input: number;
|
|
9152
9430
|
output: number;
|
|
@@ -9156,7 +9434,7 @@ export declare const MODELS: {
|
|
|
9156
9434
|
contextWindow: number;
|
|
9157
9435
|
maxTokens: number;
|
|
9158
9436
|
};
|
|
9159
|
-
readonly "openai/gpt-oss-
|
|
9437
|
+
readonly "openai/gpt-oss-120b": {
|
|
9160
9438
|
id: string;
|
|
9161
9439
|
name: string;
|
|
9162
9440
|
api: "openai-completions";
|
|
@@ -9173,7 +9451,7 @@ export declare const MODELS: {
|
|
|
9173
9451
|
contextWindow: number;
|
|
9174
9452
|
maxTokens: number;
|
|
9175
9453
|
};
|
|
9176
|
-
readonly "openai/gpt-oss-
|
|
9454
|
+
readonly "openai/gpt-oss-120b:free": {
|
|
9177
9455
|
id: string;
|
|
9178
9456
|
name: string;
|
|
9179
9457
|
api: "openai-completions";
|
|
@@ -9190,7 +9468,7 @@ export declare const MODELS: {
|
|
|
9190
9468
|
contextWindow: number;
|
|
9191
9469
|
maxTokens: number;
|
|
9192
9470
|
};
|
|
9193
|
-
readonly "openai/gpt-oss-
|
|
9471
|
+
readonly "openai/gpt-oss-20b": {
|
|
9194
9472
|
id: string;
|
|
9195
9473
|
name: string;
|
|
9196
9474
|
api: "openai-completions";
|
|
@@ -9207,14 +9485,14 @@ export declare const MODELS: {
|
|
|
9207
9485
|
contextWindow: number;
|
|
9208
9486
|
maxTokens: number;
|
|
9209
9487
|
};
|
|
9210
|
-
readonly "openai/
|
|
9488
|
+
readonly "openai/gpt-oss-20b:free": {
|
|
9211
9489
|
id: string;
|
|
9212
9490
|
name: string;
|
|
9213
9491
|
api: "openai-completions";
|
|
9214
9492
|
provider: string;
|
|
9215
9493
|
baseUrl: string;
|
|
9216
|
-
reasoning:
|
|
9217
|
-
input:
|
|
9494
|
+
reasoning: true;
|
|
9495
|
+
input: "text"[];
|
|
9218
9496
|
cost: {
|
|
9219
9497
|
input: number;
|
|
9220
9498
|
output: number;
|
|
@@ -9224,14 +9502,14 @@ export declare const MODELS: {
|
|
|
9224
9502
|
contextWindow: number;
|
|
9225
9503
|
maxTokens: number;
|
|
9226
9504
|
};
|
|
9227
|
-
readonly "openai/
|
|
9505
|
+
readonly "openai/gpt-oss-safeguard-20b": {
|
|
9228
9506
|
id: string;
|
|
9229
9507
|
name: string;
|
|
9230
9508
|
api: "openai-completions";
|
|
9231
9509
|
provider: string;
|
|
9232
9510
|
baseUrl: string;
|
|
9233
9511
|
reasoning: true;
|
|
9234
|
-
input:
|
|
9512
|
+
input: "text"[];
|
|
9235
9513
|
cost: {
|
|
9236
9514
|
input: number;
|
|
9237
9515
|
output: number;
|
|
@@ -9241,7 +9519,7 @@ export declare const MODELS: {
|
|
|
9241
9519
|
contextWindow: number;
|
|
9242
9520
|
maxTokens: number;
|
|
9243
9521
|
};
|
|
9244
|
-
readonly "openai/
|
|
9522
|
+
readonly "openai/o1": {
|
|
9245
9523
|
id: string;
|
|
9246
9524
|
name: string;
|
|
9247
9525
|
api: "openai-completions";
|
|
@@ -9258,14 +9536,14 @@ export declare const MODELS: {
|
|
|
9258
9536
|
contextWindow: number;
|
|
9259
9537
|
maxTokens: number;
|
|
9260
9538
|
};
|
|
9261
|
-
readonly "openai/o3
|
|
9539
|
+
readonly "openai/o3": {
|
|
9262
9540
|
id: string;
|
|
9263
9541
|
name: string;
|
|
9264
9542
|
api: "openai-completions";
|
|
9265
9543
|
provider: string;
|
|
9266
9544
|
baseUrl: string;
|
|
9267
|
-
reasoning:
|
|
9268
|
-
input: "text"[];
|
|
9545
|
+
reasoning: true;
|
|
9546
|
+
input: ("image" | "text")[];
|
|
9269
9547
|
cost: {
|
|
9270
9548
|
input: number;
|
|
9271
9549
|
output: number;
|
|
@@ -9275,14 +9553,14 @@ export declare const MODELS: {
|
|
|
9275
9553
|
contextWindow: number;
|
|
9276
9554
|
maxTokens: number;
|
|
9277
9555
|
};
|
|
9278
|
-
readonly "openai/o3-
|
|
9556
|
+
readonly "openai/o3-deep-research": {
|
|
9279
9557
|
id: string;
|
|
9280
9558
|
name: string;
|
|
9281
9559
|
api: "openai-completions";
|
|
9282
9560
|
provider: string;
|
|
9283
9561
|
baseUrl: string;
|
|
9284
|
-
reasoning:
|
|
9285
|
-
input: "text"[];
|
|
9562
|
+
reasoning: true;
|
|
9563
|
+
input: ("image" | "text")[];
|
|
9286
9564
|
cost: {
|
|
9287
9565
|
input: number;
|
|
9288
9566
|
output: number;
|
|
@@ -9292,14 +9570,14 @@ export declare const MODELS: {
|
|
|
9292
9570
|
contextWindow: number;
|
|
9293
9571
|
maxTokens: number;
|
|
9294
9572
|
};
|
|
9295
|
-
readonly "openai/o3-
|
|
9573
|
+
readonly "openai/o3-mini": {
|
|
9296
9574
|
id: string;
|
|
9297
9575
|
name: string;
|
|
9298
9576
|
api: "openai-completions";
|
|
9299
9577
|
provider: string;
|
|
9300
9578
|
baseUrl: string;
|
|
9301
9579
|
reasoning: true;
|
|
9302
|
-
input:
|
|
9580
|
+
input: "text"[];
|
|
9303
9581
|
cost: {
|
|
9304
9582
|
input: number;
|
|
9305
9583
|
output: number;
|
|
@@ -9309,14 +9587,14 @@ export declare const MODELS: {
|
|
|
9309
9587
|
contextWindow: number;
|
|
9310
9588
|
maxTokens: number;
|
|
9311
9589
|
};
|
|
9312
|
-
readonly "openai/
|
|
9590
|
+
readonly "openai/o3-mini-high": {
|
|
9313
9591
|
id: string;
|
|
9314
9592
|
name: string;
|
|
9315
9593
|
api: "openai-completions";
|
|
9316
9594
|
provider: string;
|
|
9317
9595
|
baseUrl: string;
|
|
9318
9596
|
reasoning: true;
|
|
9319
|
-
input:
|
|
9597
|
+
input: "text"[];
|
|
9320
9598
|
cost: {
|
|
9321
9599
|
input: number;
|
|
9322
9600
|
output: number;
|
|
@@ -9326,7 +9604,7 @@ export declare const MODELS: {
|
|
|
9326
9604
|
contextWindow: number;
|
|
9327
9605
|
maxTokens: number;
|
|
9328
9606
|
};
|
|
9329
|
-
readonly "openai/
|
|
9607
|
+
readonly "openai/o3-pro": {
|
|
9330
9608
|
id: string;
|
|
9331
9609
|
name: string;
|
|
9332
9610
|
api: "openai-completions";
|
|
@@ -9343,7 +9621,7 @@ export declare const MODELS: {
|
|
|
9343
9621
|
contextWindow: number;
|
|
9344
9622
|
maxTokens: number;
|
|
9345
9623
|
};
|
|
9346
|
-
readonly "openai/o4-mini
|
|
9624
|
+
readonly "openai/o4-mini": {
|
|
9347
9625
|
id: string;
|
|
9348
9626
|
name: string;
|
|
9349
9627
|
api: "openai-completions";
|
|
@@ -9360,7 +9638,7 @@ export declare const MODELS: {
|
|
|
9360
9638
|
contextWindow: number;
|
|
9361
9639
|
maxTokens: number;
|
|
9362
9640
|
};
|
|
9363
|
-
readonly "
|
|
9641
|
+
readonly "openai/o4-mini-deep-research": {
|
|
9364
9642
|
id: string;
|
|
9365
9643
|
name: string;
|
|
9366
9644
|
api: "openai-completions";
|
|
@@ -9377,7 +9655,7 @@ export declare const MODELS: {
|
|
|
9377
9655
|
contextWindow: number;
|
|
9378
9656
|
maxTokens: number;
|
|
9379
9657
|
};
|
|
9380
|
-
readonly "
|
|
9658
|
+
readonly "openai/o4-mini-high": {
|
|
9381
9659
|
id: string;
|
|
9382
9660
|
name: string;
|
|
9383
9661
|
api: "openai-completions";
|
|
@@ -9394,7 +9672,7 @@ export declare const MODELS: {
|
|
|
9394
9672
|
contextWindow: number;
|
|
9395
9673
|
maxTokens: number;
|
|
9396
9674
|
};
|
|
9397
|
-
readonly "openrouter/
|
|
9675
|
+
readonly "openrouter/auto": {
|
|
9398
9676
|
id: string;
|
|
9399
9677
|
name: string;
|
|
9400
9678
|
api: "openai-completions";
|
|
@@ -9411,14 +9689,14 @@ export declare const MODELS: {
|
|
|
9411
9689
|
contextWindow: number;
|
|
9412
9690
|
maxTokens: number;
|
|
9413
9691
|
};
|
|
9414
|
-
readonly "openrouter/
|
|
9692
|
+
readonly "openrouter/free": {
|
|
9415
9693
|
id: string;
|
|
9416
9694
|
name: string;
|
|
9417
9695
|
api: "openai-completions";
|
|
9418
9696
|
provider: string;
|
|
9419
9697
|
baseUrl: string;
|
|
9420
9698
|
reasoning: true;
|
|
9421
|
-
input: "text"[];
|
|
9699
|
+
input: ("image" | "text")[];
|
|
9422
9700
|
cost: {
|
|
9423
9701
|
input: number;
|
|
9424
9702
|
output: number;
|
|
@@ -10516,6 +10794,40 @@ export declare const MODELS: {
|
|
|
10516
10794
|
contextWindow: number;
|
|
10517
10795
|
maxTokens: number;
|
|
10518
10796
|
};
|
|
10797
|
+
readonly "xiaomi/mimo-v2-omni": {
|
|
10798
|
+
id: string;
|
|
10799
|
+
name: string;
|
|
10800
|
+
api: "openai-completions";
|
|
10801
|
+
provider: string;
|
|
10802
|
+
baseUrl: string;
|
|
10803
|
+
reasoning: true;
|
|
10804
|
+
input: ("image" | "text")[];
|
|
10805
|
+
cost: {
|
|
10806
|
+
input: number;
|
|
10807
|
+
output: number;
|
|
10808
|
+
cacheRead: number;
|
|
10809
|
+
cacheWrite: number;
|
|
10810
|
+
};
|
|
10811
|
+
contextWindow: number;
|
|
10812
|
+
maxTokens: number;
|
|
10813
|
+
};
|
|
10814
|
+
readonly "xiaomi/mimo-v2-pro": {
|
|
10815
|
+
id: string;
|
|
10816
|
+
name: string;
|
|
10817
|
+
api: "openai-completions";
|
|
10818
|
+
provider: string;
|
|
10819
|
+
baseUrl: string;
|
|
10820
|
+
reasoning: true;
|
|
10821
|
+
input: "text"[];
|
|
10822
|
+
cost: {
|
|
10823
|
+
input: number;
|
|
10824
|
+
output: number;
|
|
10825
|
+
cacheRead: number;
|
|
10826
|
+
cacheWrite: number;
|
|
10827
|
+
};
|
|
10828
|
+
contextWindow: number;
|
|
10829
|
+
maxTokens: number;
|
|
10830
|
+
};
|
|
10519
10831
|
readonly "z-ai/glm-4-32b": {
|
|
10520
10832
|
id: string;
|
|
10521
10833
|
name: string;
|
|
@@ -10686,6 +10998,23 @@ export declare const MODELS: {
|
|
|
10686
10998
|
contextWindow: number;
|
|
10687
10999
|
maxTokens: number;
|
|
10688
11000
|
};
|
|
11001
|
+
readonly "z-ai/glm-5-turbo": {
|
|
11002
|
+
id: string;
|
|
11003
|
+
name: string;
|
|
11004
|
+
api: "openai-completions";
|
|
11005
|
+
provider: string;
|
|
11006
|
+
baseUrl: string;
|
|
11007
|
+
reasoning: true;
|
|
11008
|
+
input: "text"[];
|
|
11009
|
+
cost: {
|
|
11010
|
+
input: number;
|
|
11011
|
+
output: number;
|
|
11012
|
+
cacheRead: number;
|
|
11013
|
+
cacheWrite: number;
|
|
11014
|
+
};
|
|
11015
|
+
contextWindow: number;
|
|
11016
|
+
maxTokens: number;
|
|
11017
|
+
};
|
|
10689
11018
|
};
|
|
10690
11019
|
readonly "vercel-ai-gateway": {
|
|
10691
11020
|
readonly "alibaba/qwen-3-14b": {
|
|
@@ -11385,40 +11714,6 @@ export declare const MODELS: {
|
|
|
11385
11714
|
contextWindow: number;
|
|
11386
11715
|
maxTokens: number;
|
|
11387
11716
|
};
|
|
11388
|
-
readonly "google/gemini-2.5-flash-lite-preview-09-2025": {
|
|
11389
|
-
id: string;
|
|
11390
|
-
name: string;
|
|
11391
|
-
api: "anthropic-messages";
|
|
11392
|
-
provider: string;
|
|
11393
|
-
baseUrl: string;
|
|
11394
|
-
reasoning: true;
|
|
11395
|
-
input: ("image" | "text")[];
|
|
11396
|
-
cost: {
|
|
11397
|
-
input: number;
|
|
11398
|
-
output: number;
|
|
11399
|
-
cacheRead: number;
|
|
11400
|
-
cacheWrite: number;
|
|
11401
|
-
};
|
|
11402
|
-
contextWindow: number;
|
|
11403
|
-
maxTokens: number;
|
|
11404
|
-
};
|
|
11405
|
-
readonly "google/gemini-2.5-flash-preview-09-2025": {
|
|
11406
|
-
id: string;
|
|
11407
|
-
name: string;
|
|
11408
|
-
api: "anthropic-messages";
|
|
11409
|
-
provider: string;
|
|
11410
|
-
baseUrl: string;
|
|
11411
|
-
reasoning: true;
|
|
11412
|
-
input: ("image" | "text")[];
|
|
11413
|
-
cost: {
|
|
11414
|
-
input: number;
|
|
11415
|
-
output: number;
|
|
11416
|
-
cacheRead: number;
|
|
11417
|
-
cacheWrite: number;
|
|
11418
|
-
};
|
|
11419
|
-
contextWindow: number;
|
|
11420
|
-
maxTokens: number;
|
|
11421
|
-
};
|
|
11422
11717
|
readonly "google/gemini-2.5-pro": {
|
|
11423
11718
|
id: string;
|
|
11424
11719
|
name: string;
|
|
@@ -11776,6 +12071,40 @@ export declare const MODELS: {
|
|
|
11776
12071
|
contextWindow: number;
|
|
11777
12072
|
maxTokens: number;
|
|
11778
12073
|
};
|
|
12074
|
+
readonly "minimax/minimax-m2.7": {
|
|
12075
|
+
id: string;
|
|
12076
|
+
name: string;
|
|
12077
|
+
api: "anthropic-messages";
|
|
12078
|
+
provider: string;
|
|
12079
|
+
baseUrl: string;
|
|
12080
|
+
reasoning: true;
|
|
12081
|
+
input: ("image" | "text")[];
|
|
12082
|
+
cost: {
|
|
12083
|
+
input: number;
|
|
12084
|
+
output: number;
|
|
12085
|
+
cacheRead: number;
|
|
12086
|
+
cacheWrite: number;
|
|
12087
|
+
};
|
|
12088
|
+
contextWindow: number;
|
|
12089
|
+
maxTokens: number;
|
|
12090
|
+
};
|
|
12091
|
+
readonly "minimax/minimax-m2.7-highspeed": {
|
|
12092
|
+
id: string;
|
|
12093
|
+
name: string;
|
|
12094
|
+
api: "anthropic-messages";
|
|
12095
|
+
provider: string;
|
|
12096
|
+
baseUrl: string;
|
|
12097
|
+
reasoning: true;
|
|
12098
|
+
input: ("image" | "text")[];
|
|
12099
|
+
cost: {
|
|
12100
|
+
input: number;
|
|
12101
|
+
output: number;
|
|
12102
|
+
cacheRead: number;
|
|
12103
|
+
cacheWrite: number;
|
|
12104
|
+
};
|
|
12105
|
+
contextWindow: number;
|
|
12106
|
+
maxTokens: number;
|
|
12107
|
+
};
|
|
11779
12108
|
readonly "mistral/codestral": {
|
|
11780
12109
|
id: string;
|
|
11781
12110
|
name: string;
|
|
@@ -12490,6 +12819,40 @@ export declare const MODELS: {
|
|
|
12490
12819
|
contextWindow: number;
|
|
12491
12820
|
maxTokens: number;
|
|
12492
12821
|
};
|
|
12822
|
+
readonly "openai/gpt-5.4-mini": {
|
|
12823
|
+
id: string;
|
|
12824
|
+
name: string;
|
|
12825
|
+
api: "anthropic-messages";
|
|
12826
|
+
provider: string;
|
|
12827
|
+
baseUrl: string;
|
|
12828
|
+
reasoning: true;
|
|
12829
|
+
input: ("image" | "text")[];
|
|
12830
|
+
cost: {
|
|
12831
|
+
input: number;
|
|
12832
|
+
output: number;
|
|
12833
|
+
cacheRead: number;
|
|
12834
|
+
cacheWrite: number;
|
|
12835
|
+
};
|
|
12836
|
+
contextWindow: number;
|
|
12837
|
+
maxTokens: number;
|
|
12838
|
+
};
|
|
12839
|
+
readonly "openai/gpt-5.4-nano": {
|
|
12840
|
+
id: string;
|
|
12841
|
+
name: string;
|
|
12842
|
+
api: "anthropic-messages";
|
|
12843
|
+
provider: string;
|
|
12844
|
+
baseUrl: string;
|
|
12845
|
+
reasoning: true;
|
|
12846
|
+
input: ("image" | "text")[];
|
|
12847
|
+
cost: {
|
|
12848
|
+
input: number;
|
|
12849
|
+
output: number;
|
|
12850
|
+
cacheRead: number;
|
|
12851
|
+
cacheWrite: number;
|
|
12852
|
+
};
|
|
12853
|
+
contextWindow: number;
|
|
12854
|
+
maxTokens: number;
|
|
12855
|
+
};
|
|
12493
12856
|
readonly "openai/gpt-5.4-pro": {
|
|
12494
12857
|
id: string;
|
|
12495
12858
|
name: string;
|
|
@@ -12694,7 +13057,7 @@ export declare const MODELS: {
|
|
|
12694
13057
|
contextWindow: number;
|
|
12695
13058
|
maxTokens: number;
|
|
12696
13059
|
};
|
|
12697
|
-
readonly "
|
|
13060
|
+
readonly "xai/grok-2-vision": {
|
|
12698
13061
|
id: string;
|
|
12699
13062
|
name: string;
|
|
12700
13063
|
api: "anthropic-messages";
|
|
@@ -12711,14 +13074,14 @@ export declare const MODELS: {
|
|
|
12711
13074
|
contextWindow: number;
|
|
12712
13075
|
maxTokens: number;
|
|
12713
13076
|
};
|
|
12714
|
-
readonly "
|
|
13077
|
+
readonly "xai/grok-3": {
|
|
12715
13078
|
id: string;
|
|
12716
13079
|
name: string;
|
|
12717
13080
|
api: "anthropic-messages";
|
|
12718
13081
|
provider: string;
|
|
12719
13082
|
baseUrl: string;
|
|
12720
13083
|
reasoning: false;
|
|
12721
|
-
input:
|
|
13084
|
+
input: "text"[];
|
|
12722
13085
|
cost: {
|
|
12723
13086
|
input: number;
|
|
12724
13087
|
output: number;
|
|
@@ -12728,14 +13091,14 @@ export declare const MODELS: {
|
|
|
12728
13091
|
contextWindow: number;
|
|
12729
13092
|
maxTokens: number;
|
|
12730
13093
|
};
|
|
12731
|
-
readonly "xai/grok-
|
|
13094
|
+
readonly "xai/grok-3-fast": {
|
|
12732
13095
|
id: string;
|
|
12733
13096
|
name: string;
|
|
12734
13097
|
api: "anthropic-messages";
|
|
12735
13098
|
provider: string;
|
|
12736
13099
|
baseUrl: string;
|
|
12737
13100
|
reasoning: false;
|
|
12738
|
-
input:
|
|
13101
|
+
input: "text"[];
|
|
12739
13102
|
cost: {
|
|
12740
13103
|
input: number;
|
|
12741
13104
|
output: number;
|
|
@@ -12745,7 +13108,7 @@ export declare const MODELS: {
|
|
|
12745
13108
|
contextWindow: number;
|
|
12746
13109
|
maxTokens: number;
|
|
12747
13110
|
};
|
|
12748
|
-
readonly "xai/grok-3": {
|
|
13111
|
+
readonly "xai/grok-3-mini": {
|
|
12749
13112
|
id: string;
|
|
12750
13113
|
name: string;
|
|
12751
13114
|
api: "anthropic-messages";
|
|
@@ -12762,7 +13125,7 @@ export declare const MODELS: {
|
|
|
12762
13125
|
contextWindow: number;
|
|
12763
13126
|
maxTokens: number;
|
|
12764
13127
|
};
|
|
12765
|
-
readonly "xai/grok-3-fast": {
|
|
13128
|
+
readonly "xai/grok-3-mini-fast": {
|
|
12766
13129
|
id: string;
|
|
12767
13130
|
name: string;
|
|
12768
13131
|
api: "anthropic-messages";
|
|
@@ -12779,14 +13142,14 @@ export declare const MODELS: {
|
|
|
12779
13142
|
contextWindow: number;
|
|
12780
13143
|
maxTokens: number;
|
|
12781
13144
|
};
|
|
12782
|
-
readonly "xai/grok-
|
|
13145
|
+
readonly "xai/grok-4": {
|
|
12783
13146
|
id: string;
|
|
12784
13147
|
name: string;
|
|
12785
13148
|
api: "anthropic-messages";
|
|
12786
13149
|
provider: string;
|
|
12787
13150
|
baseUrl: string;
|
|
12788
|
-
reasoning:
|
|
12789
|
-
input: "text"[];
|
|
13151
|
+
reasoning: true;
|
|
13152
|
+
input: ("image" | "text")[];
|
|
12790
13153
|
cost: {
|
|
12791
13154
|
input: number;
|
|
12792
13155
|
output: number;
|
|
@@ -12796,7 +13159,7 @@ export declare const MODELS: {
|
|
|
12796
13159
|
contextWindow: number;
|
|
12797
13160
|
maxTokens: number;
|
|
12798
13161
|
};
|
|
12799
|
-
readonly "xai/grok-
|
|
13162
|
+
readonly "xai/grok-4-fast-non-reasoning": {
|
|
12800
13163
|
id: string;
|
|
12801
13164
|
name: string;
|
|
12802
13165
|
api: "anthropic-messages";
|
|
@@ -12813,14 +13176,14 @@ export declare const MODELS: {
|
|
|
12813
13176
|
contextWindow: number;
|
|
12814
13177
|
maxTokens: number;
|
|
12815
13178
|
};
|
|
12816
|
-
readonly "xai/grok-4": {
|
|
13179
|
+
readonly "xai/grok-4-fast-reasoning": {
|
|
12817
13180
|
id: string;
|
|
12818
13181
|
name: string;
|
|
12819
13182
|
api: "anthropic-messages";
|
|
12820
13183
|
provider: string;
|
|
12821
13184
|
baseUrl: string;
|
|
12822
13185
|
reasoning: true;
|
|
12823
|
-
input:
|
|
13186
|
+
input: "text"[];
|
|
12824
13187
|
cost: {
|
|
12825
13188
|
input: number;
|
|
12826
13189
|
output: number;
|
|
@@ -12830,7 +13193,7 @@ export declare const MODELS: {
|
|
|
12830
13193
|
contextWindow: number;
|
|
12831
13194
|
maxTokens: number;
|
|
12832
13195
|
};
|
|
12833
|
-
readonly "xai/grok-4-fast-non-reasoning": {
|
|
13196
|
+
readonly "xai/grok-4.1-fast-non-reasoning": {
|
|
12834
13197
|
id: string;
|
|
12835
13198
|
name: string;
|
|
12836
13199
|
api: "anthropic-messages";
|
|
@@ -12847,7 +13210,7 @@ export declare const MODELS: {
|
|
|
12847
13210
|
contextWindow: number;
|
|
12848
13211
|
maxTokens: number;
|
|
12849
13212
|
};
|
|
12850
|
-
readonly "xai/grok-4-fast-reasoning": {
|
|
13213
|
+
readonly "xai/grok-4.1-fast-reasoning": {
|
|
12851
13214
|
id: string;
|
|
12852
13215
|
name: string;
|
|
12853
13216
|
api: "anthropic-messages";
|
|
@@ -12864,13 +13227,13 @@ export declare const MODELS: {
|
|
|
12864
13227
|
contextWindow: number;
|
|
12865
13228
|
maxTokens: number;
|
|
12866
13229
|
};
|
|
12867
|
-
readonly "xai/grok-4.
|
|
13230
|
+
readonly "xai/grok-4.20-multi-agent": {
|
|
12868
13231
|
id: string;
|
|
12869
13232
|
name: string;
|
|
12870
13233
|
api: "anthropic-messages";
|
|
12871
13234
|
provider: string;
|
|
12872
13235
|
baseUrl: string;
|
|
12873
|
-
reasoning:
|
|
13236
|
+
reasoning: true;
|
|
12874
13237
|
input: "text"[];
|
|
12875
13238
|
cost: {
|
|
12876
13239
|
input: number;
|
|
@@ -12881,7 +13244,7 @@ export declare const MODELS: {
|
|
|
12881
13244
|
contextWindow: number;
|
|
12882
13245
|
maxTokens: number;
|
|
12883
13246
|
};
|
|
12884
|
-
readonly "xai/grok-4.
|
|
13247
|
+
readonly "xai/grok-4.20-multi-agent-beta": {
|
|
12885
13248
|
id: string;
|
|
12886
13249
|
name: string;
|
|
12887
13250
|
api: "anthropic-messages";
|
|
@@ -12898,14 +13261,14 @@ export declare const MODELS: {
|
|
|
12898
13261
|
contextWindow: number;
|
|
12899
13262
|
maxTokens: number;
|
|
12900
13263
|
};
|
|
12901
|
-
readonly "xai/grok-4.20-
|
|
13264
|
+
readonly "xai/grok-4.20-non-reasoning": {
|
|
12902
13265
|
id: string;
|
|
12903
13266
|
name: string;
|
|
12904
13267
|
api: "anthropic-messages";
|
|
12905
13268
|
provider: string;
|
|
12906
13269
|
baseUrl: string;
|
|
12907
|
-
reasoning:
|
|
12908
|
-
input: "text"[];
|
|
13270
|
+
reasoning: false;
|
|
13271
|
+
input: ("image" | "text")[];
|
|
12909
13272
|
cost: {
|
|
12910
13273
|
input: number;
|
|
12911
13274
|
output: number;
|
|
@@ -12932,6 +13295,23 @@ export declare const MODELS: {
|
|
|
12932
13295
|
contextWindow: number;
|
|
12933
13296
|
maxTokens: number;
|
|
12934
13297
|
};
|
|
13298
|
+
readonly "xai/grok-4.20-reasoning": {
|
|
13299
|
+
id: string;
|
|
13300
|
+
name: string;
|
|
13301
|
+
api: "anthropic-messages";
|
|
13302
|
+
provider: string;
|
|
13303
|
+
baseUrl: string;
|
|
13304
|
+
reasoning: true;
|
|
13305
|
+
input: ("image" | "text")[];
|
|
13306
|
+
cost: {
|
|
13307
|
+
input: number;
|
|
13308
|
+
output: number;
|
|
13309
|
+
cacheRead: number;
|
|
13310
|
+
cacheWrite: number;
|
|
13311
|
+
};
|
|
13312
|
+
contextWindow: number;
|
|
13313
|
+
maxTokens: number;
|
|
13314
|
+
};
|
|
12935
13315
|
readonly "xai/grok-4.20-reasoning-beta": {
|
|
12936
13316
|
id: string;
|
|
12937
13317
|
name: string;
|
|
@@ -12983,6 +13363,23 @@ export declare const MODELS: {
|
|
|
12983
13363
|
contextWindow: number;
|
|
12984
13364
|
maxTokens: number;
|
|
12985
13365
|
};
|
|
13366
|
+
readonly "xiaomi/mimo-v2-pro": {
|
|
13367
|
+
id: string;
|
|
13368
|
+
name: string;
|
|
13369
|
+
api: "anthropic-messages";
|
|
13370
|
+
provider: string;
|
|
13371
|
+
baseUrl: string;
|
|
13372
|
+
reasoning: true;
|
|
13373
|
+
input: "text"[];
|
|
13374
|
+
cost: {
|
|
13375
|
+
input: number;
|
|
13376
|
+
output: number;
|
|
13377
|
+
cacheRead: number;
|
|
13378
|
+
cacheWrite: number;
|
|
13379
|
+
};
|
|
13380
|
+
contextWindow: number;
|
|
13381
|
+
maxTokens: number;
|
|
13382
|
+
};
|
|
12986
13383
|
readonly "zai/glm-4.5": {
|
|
12987
13384
|
id: string;
|
|
12988
13385
|
name: string;
|
|
@@ -13153,6 +13550,23 @@ export declare const MODELS: {
|
|
|
13153
13550
|
contextWindow: number;
|
|
13154
13551
|
maxTokens: number;
|
|
13155
13552
|
};
|
|
13553
|
+
readonly "zai/glm-5-turbo": {
|
|
13554
|
+
id: string;
|
|
13555
|
+
name: string;
|
|
13556
|
+
api: "anthropic-messages";
|
|
13557
|
+
provider: string;
|
|
13558
|
+
baseUrl: string;
|
|
13559
|
+
reasoning: true;
|
|
13560
|
+
input: "text"[];
|
|
13561
|
+
cost: {
|
|
13562
|
+
input: number;
|
|
13563
|
+
output: number;
|
|
13564
|
+
cacheRead: number;
|
|
13565
|
+
cacheWrite: number;
|
|
13566
|
+
};
|
|
13567
|
+
contextWindow: number;
|
|
13568
|
+
maxTokens: number;
|
|
13569
|
+
};
|
|
13156
13570
|
};
|
|
13157
13571
|
readonly xai: {
|
|
13158
13572
|
readonly "grok-2": {
|
|
@@ -13478,7 +13892,7 @@ export declare const MODELS: {
|
|
|
13478
13892
|
contextWindow: number;
|
|
13479
13893
|
maxTokens: number;
|
|
13480
13894
|
};
|
|
13481
|
-
readonly "grok-4.20-
|
|
13895
|
+
readonly "grok-4.20-0309-non-reasoning": {
|
|
13482
13896
|
id: string;
|
|
13483
13897
|
name: string;
|
|
13484
13898
|
api: "openai-completions";
|
|
@@ -13495,24 +13909,7 @@ export declare const MODELS: {
|
|
|
13495
13909
|
contextWindow: number;
|
|
13496
13910
|
maxTokens: number;
|
|
13497
13911
|
};
|
|
13498
|
-
readonly "grok-4.20-
|
|
13499
|
-
id: string;
|
|
13500
|
-
name: string;
|
|
13501
|
-
api: "openai-completions";
|
|
13502
|
-
provider: string;
|
|
13503
|
-
baseUrl: string;
|
|
13504
|
-
reasoning: true;
|
|
13505
|
-
input: ("image" | "text")[];
|
|
13506
|
-
cost: {
|
|
13507
|
-
input: number;
|
|
13508
|
-
output: number;
|
|
13509
|
-
cacheRead: number;
|
|
13510
|
-
cacheWrite: number;
|
|
13511
|
-
};
|
|
13512
|
-
contextWindow: number;
|
|
13513
|
-
maxTokens: number;
|
|
13514
|
-
};
|
|
13515
|
-
readonly "grok-4.20-multi-agent-beta-latest": {
|
|
13912
|
+
readonly "grok-4.20-0309-reasoning": {
|
|
13516
13913
|
id: string;
|
|
13517
13914
|
name: string;
|
|
13518
13915
|
api: "openai-completions";
|
|
@@ -13771,6 +14168,27 @@ export declare const MODELS: {
|
|
|
13771
14168
|
contextWindow: number;
|
|
13772
14169
|
maxTokens: number;
|
|
13773
14170
|
};
|
|
14171
|
+
readonly "glm-5-turbo": {
|
|
14172
|
+
id: string;
|
|
14173
|
+
name: string;
|
|
14174
|
+
api: "openai-completions";
|
|
14175
|
+
provider: string;
|
|
14176
|
+
baseUrl: string;
|
|
14177
|
+
compat: {
|
|
14178
|
+
supportsDeveloperRole: false;
|
|
14179
|
+
thinkingFormat: "zai";
|
|
14180
|
+
};
|
|
14181
|
+
reasoning: true;
|
|
14182
|
+
input: "text"[];
|
|
14183
|
+
cost: {
|
|
14184
|
+
input: number;
|
|
14185
|
+
output: number;
|
|
14186
|
+
cacheRead: number;
|
|
14187
|
+
cacheWrite: number;
|
|
14188
|
+
};
|
|
14189
|
+
contextWindow: number;
|
|
14190
|
+
maxTokens: number;
|
|
14191
|
+
};
|
|
13774
14192
|
};
|
|
13775
14193
|
};
|
|
13776
14194
|
//# sourceMappingURL=models.generated.d.ts.map
|