@draht/ai 2026.3.25-1 → 2026.4.5
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +86 -0
- package/dist/index.d.ts +1 -0
- package/dist/index.d.ts.map +1 -1
- package/dist/index.js +1 -0
- package/dist/index.js.map +1 -1
- package/dist/models.generated.d.ts +467 -128
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +591 -266
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/faux.d.ts +56 -0
- package/dist/providers/faux.d.ts.map +1 -0
- package/dist/providers/faux.js +367 -0
- package/dist/providers/faux.js.map +1 -0
- package/dist/providers/openai-codex-responses.d.ts.map +1 -1
- package/dist/providers/openai-codex-responses.js +1 -1
- package/dist/providers/openai-codex-responses.js.map +1 -1
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +5 -0
- package/dist/providers/openai-completions.js.map +1 -1
- package/dist/providers/openai-responses-shared.d.ts.map +1 -1
- package/dist/providers/openai-responses-shared.js +12 -0
- package/dist/providers/openai-responses-shared.js.map +1 -1
- package/dist/types.d.ts +2 -0
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/dist/utils/overflow.d.ts +1 -1
- package/dist/utils/overflow.d.ts.map +1 -1
- package/dist/utils/overflow.js +4 -2
- package/dist/utils/overflow.js.map +1 -1
- package/package.json +1 -1
package/dist/models.generated.js
CHANGED
|
@@ -815,8 +815,8 @@ export const MODELS = {
|
|
|
815
815
|
cacheRead: 0,
|
|
816
816
|
cacheWrite: 0,
|
|
817
817
|
},
|
|
818
|
-
contextWindow:
|
|
819
|
-
maxTokens:
|
|
818
|
+
contextWindow: 196608,
|
|
819
|
+
maxTokens: 98304,
|
|
820
820
|
},
|
|
821
821
|
"mistral.devstral-2-123b": {
|
|
822
822
|
id: "mistral.devstral-2-123b",
|
|
@@ -1056,6 +1056,23 @@ export const MODELS = {
|
|
|
1056
1056
|
contextWindow: 128000,
|
|
1057
1057
|
maxTokens: 4096,
|
|
1058
1058
|
},
|
|
1059
|
+
"nvidia.nemotron-super-3-120b": {
|
|
1060
|
+
id: "nvidia.nemotron-super-3-120b",
|
|
1061
|
+
name: "NVIDIA Nemotron 3 Super 120B A12B",
|
|
1062
|
+
api: "bedrock-converse-stream",
|
|
1063
|
+
provider: "amazon-bedrock",
|
|
1064
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
1065
|
+
reasoning: true,
|
|
1066
|
+
input: ["text"],
|
|
1067
|
+
cost: {
|
|
1068
|
+
input: 0.15,
|
|
1069
|
+
output: 0.65,
|
|
1070
|
+
cacheRead: 0,
|
|
1071
|
+
cacheWrite: 0,
|
|
1072
|
+
},
|
|
1073
|
+
contextWindow: 262144,
|
|
1074
|
+
maxTokens: 131072,
|
|
1075
|
+
},
|
|
1059
1076
|
"openai.gpt-oss-120b-1:0": {
|
|
1060
1077
|
id: "openai.gpt-oss-120b-1:0",
|
|
1061
1078
|
name: "gpt-oss-120b",
|
|
@@ -1444,8 +1461,8 @@ export const MODELS = {
|
|
|
1444
1461
|
cacheRead: 0,
|
|
1445
1462
|
cacheWrite: 0,
|
|
1446
1463
|
},
|
|
1447
|
-
contextWindow:
|
|
1448
|
-
maxTokens:
|
|
1464
|
+
contextWindow: 202752,
|
|
1465
|
+
maxTokens: 101376,
|
|
1449
1466
|
},
|
|
1450
1467
|
},
|
|
1451
1468
|
"anthropic": {
|
|
@@ -1534,23 +1551,6 @@ export const MODELS = {
|
|
|
1534
1551
|
contextWindow: 200000,
|
|
1535
1552
|
maxTokens: 64000,
|
|
1536
1553
|
},
|
|
1537
|
-
"claude-3-7-sonnet-latest": {
|
|
1538
|
-
id: "claude-3-7-sonnet-latest",
|
|
1539
|
-
name: "Claude Sonnet 3.7 (latest)",
|
|
1540
|
-
api: "anthropic-messages",
|
|
1541
|
-
provider: "anthropic",
|
|
1542
|
-
baseUrl: "https://api.anthropic.com",
|
|
1543
|
-
reasoning: true,
|
|
1544
|
-
input: ["text", "image"],
|
|
1545
|
-
cost: {
|
|
1546
|
-
input: 3,
|
|
1547
|
-
output: 15,
|
|
1548
|
-
cacheRead: 0.3,
|
|
1549
|
-
cacheWrite: 3.75,
|
|
1550
|
-
},
|
|
1551
|
-
contextWindow: 200000,
|
|
1552
|
-
maxTokens: 64000,
|
|
1553
|
-
},
|
|
1554
1554
|
"claude-3-haiku-20240307": {
|
|
1555
1555
|
id: "claude-3-haiku-20240307",
|
|
1556
1556
|
name: "Claude Haiku 3",
|
|
@@ -2284,6 +2284,23 @@ export const MODELS = {
|
|
|
2284
2284
|
contextWindow: 400000,
|
|
2285
2285
|
maxTokens: 128000,
|
|
2286
2286
|
},
|
|
2287
|
+
"gpt-5.3-chat-latest": {
|
|
2288
|
+
id: "gpt-5.3-chat-latest",
|
|
2289
|
+
name: "GPT-5.3 Chat (latest)",
|
|
2290
|
+
api: "azure-openai-responses",
|
|
2291
|
+
provider: "azure-openai-responses",
|
|
2292
|
+
baseUrl: "",
|
|
2293
|
+
reasoning: false,
|
|
2294
|
+
input: ["text", "image"],
|
|
2295
|
+
cost: {
|
|
2296
|
+
input: 1.75,
|
|
2297
|
+
output: 14,
|
|
2298
|
+
cacheRead: 0.175,
|
|
2299
|
+
cacheWrite: 0,
|
|
2300
|
+
},
|
|
2301
|
+
contextWindow: 128000,
|
|
2302
|
+
maxTokens: 16384,
|
|
2303
|
+
},
|
|
2287
2304
|
"gpt-5.3-codex": {
|
|
2288
2305
|
id: "gpt-5.3-codex",
|
|
2289
2306
|
name: "GPT-5.3 Codex",
|
|
@@ -3443,6 +3460,57 @@ export const MODELS = {
|
|
|
3443
3460
|
contextWindow: 131072,
|
|
3444
3461
|
maxTokens: 65536,
|
|
3445
3462
|
},
|
|
3463
|
+
"gemma-3-27b-it": {
|
|
3464
|
+
id: "gemma-3-27b-it",
|
|
3465
|
+
name: "Gemma 3 27B",
|
|
3466
|
+
api: "google-generative-ai",
|
|
3467
|
+
provider: "google",
|
|
3468
|
+
baseUrl: "https://generativelanguage.googleapis.com/v1beta",
|
|
3469
|
+
reasoning: false,
|
|
3470
|
+
input: ["text", "image"],
|
|
3471
|
+
cost: {
|
|
3472
|
+
input: 0,
|
|
3473
|
+
output: 0,
|
|
3474
|
+
cacheRead: 0,
|
|
3475
|
+
cacheWrite: 0,
|
|
3476
|
+
},
|
|
3477
|
+
contextWindow: 131072,
|
|
3478
|
+
maxTokens: 8192,
|
|
3479
|
+
},
|
|
3480
|
+
"gemma-4-26b": {
|
|
3481
|
+
id: "gemma-4-26b",
|
|
3482
|
+
name: "Gemma 4 26B",
|
|
3483
|
+
api: "google-generative-ai",
|
|
3484
|
+
provider: "google",
|
|
3485
|
+
baseUrl: "https://generativelanguage.googleapis.com/v1beta",
|
|
3486
|
+
reasoning: true,
|
|
3487
|
+
input: ["text", "image"],
|
|
3488
|
+
cost: {
|
|
3489
|
+
input: 0,
|
|
3490
|
+
output: 0,
|
|
3491
|
+
cacheRead: 0,
|
|
3492
|
+
cacheWrite: 0,
|
|
3493
|
+
},
|
|
3494
|
+
contextWindow: 256000,
|
|
3495
|
+
maxTokens: 8192,
|
|
3496
|
+
},
|
|
3497
|
+
"gemma-4-31b": {
|
|
3498
|
+
id: "gemma-4-31b",
|
|
3499
|
+
name: "Gemma 4 31B",
|
|
3500
|
+
api: "google-generative-ai",
|
|
3501
|
+
provider: "google",
|
|
3502
|
+
baseUrl: "https://generativelanguage.googleapis.com/v1beta",
|
|
3503
|
+
reasoning: true,
|
|
3504
|
+
input: ["text", "image"],
|
|
3505
|
+
cost: {
|
|
3506
|
+
input: 0,
|
|
3507
|
+
output: 0,
|
|
3508
|
+
cacheRead: 0,
|
|
3509
|
+
cacheWrite: 0,
|
|
3510
|
+
},
|
|
3511
|
+
contextWindow: 256000,
|
|
3512
|
+
maxTokens: 8192,
|
|
3513
|
+
},
|
|
3446
3514
|
},
|
|
3447
3515
|
"google-antigravity": {
|
|
3448
3516
|
"claude-opus-4-5-thinking": {
|
|
@@ -3944,6 +4012,40 @@ export const MODELS = {
|
|
|
3944
4012
|
contextWindow: 8192,
|
|
3945
4013
|
maxTokens: 8192,
|
|
3946
4014
|
},
|
|
4015
|
+
"groq/compound": {
|
|
4016
|
+
id: "groq/compound",
|
|
4017
|
+
name: "Compound",
|
|
4018
|
+
api: "openai-completions",
|
|
4019
|
+
provider: "groq",
|
|
4020
|
+
baseUrl: "https://api.groq.com/openai/v1",
|
|
4021
|
+
reasoning: true,
|
|
4022
|
+
input: ["text"],
|
|
4023
|
+
cost: {
|
|
4024
|
+
input: 0,
|
|
4025
|
+
output: 0,
|
|
4026
|
+
cacheRead: 0,
|
|
4027
|
+
cacheWrite: 0,
|
|
4028
|
+
},
|
|
4029
|
+
contextWindow: 131072,
|
|
4030
|
+
maxTokens: 8192,
|
|
4031
|
+
},
|
|
4032
|
+
"groq/compound-mini": {
|
|
4033
|
+
id: "groq/compound-mini",
|
|
4034
|
+
name: "Compound Mini",
|
|
4035
|
+
api: "openai-completions",
|
|
4036
|
+
provider: "groq",
|
|
4037
|
+
baseUrl: "https://api.groq.com/openai/v1",
|
|
4038
|
+
reasoning: true,
|
|
4039
|
+
input: ["text"],
|
|
4040
|
+
cost: {
|
|
4041
|
+
input: 0,
|
|
4042
|
+
output: 0,
|
|
4043
|
+
cacheRead: 0,
|
|
4044
|
+
cacheWrite: 0,
|
|
4045
|
+
},
|
|
4046
|
+
contextWindow: 131072,
|
|
4047
|
+
maxTokens: 8192,
|
|
4048
|
+
},
|
|
3947
4049
|
"llama-3.1-8b-instant": {
|
|
3948
4050
|
id: "llama-3.1-8b-instant",
|
|
3949
4051
|
name: "Llama 3.1 8B Instant",
|
|
@@ -4131,6 +4233,23 @@ export const MODELS = {
|
|
|
4131
4233
|
contextWindow: 131072,
|
|
4132
4234
|
maxTokens: 65536,
|
|
4133
4235
|
},
|
|
4236
|
+
"openai/gpt-oss-safeguard-20b": {
|
|
4237
|
+
id: "openai/gpt-oss-safeguard-20b",
|
|
4238
|
+
name: "Safety GPT OSS 20B",
|
|
4239
|
+
api: "openai-completions",
|
|
4240
|
+
provider: "groq",
|
|
4241
|
+
baseUrl: "https://api.groq.com/openai/v1",
|
|
4242
|
+
reasoning: true,
|
|
4243
|
+
input: ["text"],
|
|
4244
|
+
cost: {
|
|
4245
|
+
input: 0.075,
|
|
4246
|
+
output: 0.3,
|
|
4247
|
+
cacheRead: 0.037,
|
|
4248
|
+
cacheWrite: 0,
|
|
4249
|
+
},
|
|
4250
|
+
contextWindow: 131072,
|
|
4251
|
+
maxTokens: 65536,
|
|
4252
|
+
},
|
|
4134
4253
|
"qwen-qwq-32b": {
|
|
4135
4254
|
id: "qwen-qwq-32b",
|
|
4136
4255
|
name: "Qwen QwQ 32B",
|
|
@@ -4163,7 +4282,7 @@ export const MODELS = {
|
|
|
4163
4282
|
cacheWrite: 0,
|
|
4164
4283
|
},
|
|
4165
4284
|
contextWindow: 131072,
|
|
4166
|
-
maxTokens:
|
|
4285
|
+
maxTokens: 40960,
|
|
4167
4286
|
},
|
|
4168
4287
|
},
|
|
4169
4288
|
"huggingface": {
|
|
@@ -5094,22 +5213,39 @@ export const MODELS = {
|
|
|
5094
5213
|
contextWindow: 128000,
|
|
5095
5214
|
maxTokens: 16384,
|
|
5096
5215
|
},
|
|
5216
|
+
"mistral-small-2603": {
|
|
5217
|
+
id: "mistral-small-2603",
|
|
5218
|
+
name: "Mistral Small 4",
|
|
5219
|
+
api: "mistral-conversations",
|
|
5220
|
+
provider: "mistral",
|
|
5221
|
+
baseUrl: "https://api.mistral.ai",
|
|
5222
|
+
reasoning: true,
|
|
5223
|
+
input: ["text", "image"],
|
|
5224
|
+
cost: {
|
|
5225
|
+
input: 0.15,
|
|
5226
|
+
output: 0.6,
|
|
5227
|
+
cacheRead: 0,
|
|
5228
|
+
cacheWrite: 0,
|
|
5229
|
+
},
|
|
5230
|
+
contextWindow: 256000,
|
|
5231
|
+
maxTokens: 256000,
|
|
5232
|
+
},
|
|
5097
5233
|
"mistral-small-latest": {
|
|
5098
5234
|
id: "mistral-small-latest",
|
|
5099
5235
|
name: "Mistral Small (latest)",
|
|
5100
5236
|
api: "mistral-conversations",
|
|
5101
5237
|
provider: "mistral",
|
|
5102
5238
|
baseUrl: "https://api.mistral.ai",
|
|
5103
|
-
reasoning:
|
|
5239
|
+
reasoning: true,
|
|
5104
5240
|
input: ["text", "image"],
|
|
5105
5241
|
cost: {
|
|
5106
|
-
input: 0.
|
|
5107
|
-
output: 0.
|
|
5242
|
+
input: 0.15,
|
|
5243
|
+
output: 0.6,
|
|
5108
5244
|
cacheRead: 0,
|
|
5109
5245
|
cacheWrite: 0,
|
|
5110
5246
|
},
|
|
5111
|
-
contextWindow:
|
|
5112
|
-
maxTokens:
|
|
5247
|
+
contextWindow: 256000,
|
|
5248
|
+
maxTokens: 256000,
|
|
5113
5249
|
},
|
|
5114
5250
|
"open-mistral-7b": {
|
|
5115
5251
|
id: "open-mistral-7b",
|
|
@@ -5640,6 +5776,23 @@ export const MODELS = {
|
|
|
5640
5776
|
contextWindow: 400000,
|
|
5641
5777
|
maxTokens: 128000,
|
|
5642
5778
|
},
|
|
5779
|
+
"gpt-5.3-chat-latest": {
|
|
5780
|
+
id: "gpt-5.3-chat-latest",
|
|
5781
|
+
name: "GPT-5.3 Chat (latest)",
|
|
5782
|
+
api: "openai-responses",
|
|
5783
|
+
provider: "openai",
|
|
5784
|
+
baseUrl: "https://api.openai.com/v1",
|
|
5785
|
+
reasoning: false,
|
|
5786
|
+
input: ["text", "image"],
|
|
5787
|
+
cost: {
|
|
5788
|
+
input: 1.75,
|
|
5789
|
+
output: 14,
|
|
5790
|
+
cacheRead: 0.175,
|
|
5791
|
+
cacheWrite: 0,
|
|
5792
|
+
},
|
|
5793
|
+
contextWindow: 128000,
|
|
5794
|
+
maxTokens: 16384,
|
|
5795
|
+
},
|
|
5643
5796
|
"gpt-5.3-codex": {
|
|
5644
5797
|
id: "gpt-5.3-codex",
|
|
5645
5798
|
name: "GPT-5.3 Codex",
|
|
@@ -6494,40 +6647,6 @@ export const MODELS = {
|
|
|
6494
6647
|
contextWindow: 262144,
|
|
6495
6648
|
maxTokens: 65536,
|
|
6496
6649
|
},
|
|
6497
|
-
"mimo-v2-omni-free": {
|
|
6498
|
-
id: "mimo-v2-omni-free",
|
|
6499
|
-
name: "MiMo V2 Omni Free",
|
|
6500
|
-
api: "openai-completions",
|
|
6501
|
-
provider: "opencode",
|
|
6502
|
-
baseUrl: "https://opencode.ai/zen/v1",
|
|
6503
|
-
reasoning: true,
|
|
6504
|
-
input: ["text", "image"],
|
|
6505
|
-
cost: {
|
|
6506
|
-
input: 0,
|
|
6507
|
-
output: 0,
|
|
6508
|
-
cacheRead: 0,
|
|
6509
|
-
cacheWrite: 0,
|
|
6510
|
-
},
|
|
6511
|
-
contextWindow: 262144,
|
|
6512
|
-
maxTokens: 64000,
|
|
6513
|
-
},
|
|
6514
|
-
"mimo-v2-pro-free": {
|
|
6515
|
-
id: "mimo-v2-pro-free",
|
|
6516
|
-
name: "MiMo V2 Pro Free",
|
|
6517
|
-
api: "openai-completions",
|
|
6518
|
-
provider: "opencode",
|
|
6519
|
-
baseUrl: "https://opencode.ai/zen/v1",
|
|
6520
|
-
reasoning: true,
|
|
6521
|
-
input: ["text"],
|
|
6522
|
-
cost: {
|
|
6523
|
-
input: 0,
|
|
6524
|
-
output: 0,
|
|
6525
|
-
cacheRead: 0,
|
|
6526
|
-
cacheWrite: 0,
|
|
6527
|
-
},
|
|
6528
|
-
contextWindow: 1048576,
|
|
6529
|
-
maxTokens: 64000,
|
|
6530
|
-
},
|
|
6531
6650
|
"minimax-m2.5": {
|
|
6532
6651
|
id: "minimax-m2.5",
|
|
6533
6652
|
name: "MiniMax M2.5",
|
|
@@ -6576,9 +6695,26 @@ export const MODELS = {
|
|
|
6576
6695
|
cacheRead: 0,
|
|
6577
6696
|
cacheWrite: 0,
|
|
6578
6697
|
},
|
|
6579
|
-
contextWindow:
|
|
6698
|
+
contextWindow: 204800,
|
|
6580
6699
|
maxTokens: 128000,
|
|
6581
6700
|
},
|
|
6701
|
+
"qwen3.6-plus-free": {
|
|
6702
|
+
id: "qwen3.6-plus-free",
|
|
6703
|
+
name: "Qwen3.6 Plus Free",
|
|
6704
|
+
api: "openai-completions",
|
|
6705
|
+
provider: "opencode",
|
|
6706
|
+
baseUrl: "https://opencode.ai/zen/v1",
|
|
6707
|
+
reasoning: true,
|
|
6708
|
+
input: ["text"],
|
|
6709
|
+
cost: {
|
|
6710
|
+
input: 0,
|
|
6711
|
+
output: 0,
|
|
6712
|
+
cacheRead: 0,
|
|
6713
|
+
cacheWrite: 0,
|
|
6714
|
+
},
|
|
6715
|
+
contextWindow: 1048576,
|
|
6716
|
+
maxTokens: 64000,
|
|
6717
|
+
},
|
|
6582
6718
|
},
|
|
6583
6719
|
"opencode-go": {
|
|
6584
6720
|
"glm-5": {
|
|
@@ -6615,12 +6751,46 @@ export const MODELS = {
|
|
|
6615
6751
|
contextWindow: 262144,
|
|
6616
6752
|
maxTokens: 65536,
|
|
6617
6753
|
},
|
|
6754
|
+
"mimo-v2-omni": {
|
|
6755
|
+
id: "mimo-v2-omni",
|
|
6756
|
+
name: "MiMo V2 Omni",
|
|
6757
|
+
api: "openai-completions",
|
|
6758
|
+
provider: "opencode-go",
|
|
6759
|
+
baseUrl: "https://opencode.ai/zen/go/v1",
|
|
6760
|
+
reasoning: true,
|
|
6761
|
+
input: ["text", "image"],
|
|
6762
|
+
cost: {
|
|
6763
|
+
input: 0.4,
|
|
6764
|
+
output: 2,
|
|
6765
|
+
cacheRead: 0.08,
|
|
6766
|
+
cacheWrite: 0,
|
|
6767
|
+
},
|
|
6768
|
+
contextWindow: 262144,
|
|
6769
|
+
maxTokens: 64000,
|
|
6770
|
+
},
|
|
6771
|
+
"mimo-v2-pro": {
|
|
6772
|
+
id: "mimo-v2-pro",
|
|
6773
|
+
name: "MiMo V2 Pro",
|
|
6774
|
+
api: "openai-completions",
|
|
6775
|
+
provider: "opencode-go",
|
|
6776
|
+
baseUrl: "https://opencode.ai/zen/go/v1",
|
|
6777
|
+
reasoning: true,
|
|
6778
|
+
input: ["text"],
|
|
6779
|
+
cost: {
|
|
6780
|
+
input: 1,
|
|
6781
|
+
output: 3,
|
|
6782
|
+
cacheRead: 0.2,
|
|
6783
|
+
cacheWrite: 0,
|
|
6784
|
+
},
|
|
6785
|
+
contextWindow: 1048576,
|
|
6786
|
+
maxTokens: 64000,
|
|
6787
|
+
},
|
|
6618
6788
|
"minimax-m2.5": {
|
|
6619
6789
|
id: "minimax-m2.5",
|
|
6620
6790
|
name: "MiniMax M2.5",
|
|
6621
|
-
api: "
|
|
6791
|
+
api: "openai-completions",
|
|
6622
6792
|
provider: "opencode-go",
|
|
6623
|
-
baseUrl: "https://opencode.ai/zen/go",
|
|
6793
|
+
baseUrl: "https://opencode.ai/zen/go/v1",
|
|
6624
6794
|
reasoning: true,
|
|
6625
6795
|
input: ["text"],
|
|
6626
6796
|
cost: {
|
|
@@ -6821,30 +6991,13 @@ export const MODELS = {
|
|
|
6821
6991
|
contextWindow: 200000,
|
|
6822
6992
|
maxTokens: 8192,
|
|
6823
6993
|
},
|
|
6824
|
-
"anthropic/claude-3.
|
|
6825
|
-
id: "anthropic/claude-3.
|
|
6826
|
-
name: "Anthropic: Claude 3.
|
|
6994
|
+
"anthropic/claude-3.7-sonnet": {
|
|
6995
|
+
id: "anthropic/claude-3.7-sonnet",
|
|
6996
|
+
name: "Anthropic: Claude 3.7 Sonnet",
|
|
6827
6997
|
api: "openai-completions",
|
|
6828
6998
|
provider: "openrouter",
|
|
6829
6999
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
6830
|
-
reasoning:
|
|
6831
|
-
input: ["text", "image"],
|
|
6832
|
-
cost: {
|
|
6833
|
-
input: 6,
|
|
6834
|
-
output: 30,
|
|
6835
|
-
cacheRead: 0.6,
|
|
6836
|
-
cacheWrite: 7.5,
|
|
6837
|
-
},
|
|
6838
|
-
contextWindow: 200000,
|
|
6839
|
-
maxTokens: 8192,
|
|
6840
|
-
},
|
|
6841
|
-
"anthropic/claude-3.7-sonnet": {
|
|
6842
|
-
id: "anthropic/claude-3.7-sonnet",
|
|
6843
|
-
name: "Anthropic: Claude 3.7 Sonnet",
|
|
6844
|
-
api: "openai-completions",
|
|
6845
|
-
provider: "openrouter",
|
|
6846
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
6847
|
-
reasoning: true,
|
|
7000
|
+
reasoning: true,
|
|
6848
7001
|
input: ["text", "image"],
|
|
6849
7002
|
cost: {
|
|
6850
7003
|
input: 3,
|
|
@@ -7025,6 +7178,23 @@ export const MODELS = {
|
|
|
7025
7178
|
contextWindow: 131000,
|
|
7026
7179
|
maxTokens: 4096,
|
|
7027
7180
|
},
|
|
7181
|
+
"arcee-ai/trinity-large-thinking": {
|
|
7182
|
+
id: "arcee-ai/trinity-large-thinking",
|
|
7183
|
+
name: "Arcee AI: Trinity Large Thinking",
|
|
7184
|
+
api: "openai-completions",
|
|
7185
|
+
provider: "openrouter",
|
|
7186
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
7187
|
+
reasoning: true,
|
|
7188
|
+
input: ["text"],
|
|
7189
|
+
cost: {
|
|
7190
|
+
input: 0.22,
|
|
7191
|
+
output: 0.85,
|
|
7192
|
+
cacheRead: 0,
|
|
7193
|
+
cacheWrite: 0,
|
|
7194
|
+
},
|
|
7195
|
+
contextWindow: 262144,
|
|
7196
|
+
maxTokens: 262144,
|
|
7197
|
+
},
|
|
7028
7198
|
"arcee-ai/trinity-mini": {
|
|
7029
7199
|
id: "arcee-ai/trinity-mini",
|
|
7030
7200
|
name: "Arcee AI: Trinity Mini",
|
|
@@ -7465,7 +7635,7 @@ export const MODELS = {
|
|
|
7465
7635
|
cacheWrite: 0.08333333333333334,
|
|
7466
7636
|
},
|
|
7467
7637
|
contextWindow: 1048576,
|
|
7468
|
-
maxTokens:
|
|
7638
|
+
maxTokens: 65535,
|
|
7469
7639
|
},
|
|
7470
7640
|
"google/gemini-2.5-pro": {
|
|
7471
7641
|
id: "google/gemini-2.5-pro",
|
|
@@ -7535,23 +7705,6 @@ export const MODELS = {
|
|
|
7535
7705
|
contextWindow: 1048576,
|
|
7536
7706
|
maxTokens: 65536,
|
|
7537
7707
|
},
|
|
7538
|
-
"google/gemini-3-pro-preview": {
|
|
7539
|
-
id: "google/gemini-3-pro-preview",
|
|
7540
|
-
name: "Google: Gemini 3 Pro Preview",
|
|
7541
|
-
api: "openai-completions",
|
|
7542
|
-
provider: "openrouter",
|
|
7543
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
7544
|
-
reasoning: true,
|
|
7545
|
-
input: ["text", "image"],
|
|
7546
|
-
cost: {
|
|
7547
|
-
input: 2,
|
|
7548
|
-
output: 12,
|
|
7549
|
-
cacheRead: 0.19999999999999998,
|
|
7550
|
-
cacheWrite: 0.375,
|
|
7551
|
-
},
|
|
7552
|
-
contextWindow: 1048576,
|
|
7553
|
-
maxTokens: 65536,
|
|
7554
|
-
},
|
|
7555
7708
|
"google/gemini-3.1-flash-lite-preview": {
|
|
7556
7709
|
id: "google/gemini-3.1-flash-lite-preview",
|
|
7557
7710
|
name: "Google: Gemini 3.1 Flash Lite Preview",
|
|
@@ -7603,6 +7756,40 @@ export const MODELS = {
|
|
|
7603
7756
|
contextWindow: 1048576,
|
|
7604
7757
|
maxTokens: 65536,
|
|
7605
7758
|
},
|
|
7759
|
+
"google/gemma-4-26b-a4b-it": {
|
|
7760
|
+
id: "google/gemma-4-26b-a4b-it",
|
|
7761
|
+
name: "Google: Gemma 4 26B A4B ",
|
|
7762
|
+
api: "openai-completions",
|
|
7763
|
+
provider: "openrouter",
|
|
7764
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
7765
|
+
reasoning: true,
|
|
7766
|
+
input: ["text", "image"],
|
|
7767
|
+
cost: {
|
|
7768
|
+
input: 0.13,
|
|
7769
|
+
output: 0.39999999999999997,
|
|
7770
|
+
cacheRead: 0,
|
|
7771
|
+
cacheWrite: 0,
|
|
7772
|
+
},
|
|
7773
|
+
contextWindow: 262144,
|
|
7774
|
+
maxTokens: 262144,
|
|
7775
|
+
},
|
|
7776
|
+
"google/gemma-4-31b-it": {
|
|
7777
|
+
id: "google/gemma-4-31b-it",
|
|
7778
|
+
name: "Google: Gemma 4 31B",
|
|
7779
|
+
api: "openai-completions",
|
|
7780
|
+
provider: "openrouter",
|
|
7781
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
7782
|
+
reasoning: true,
|
|
7783
|
+
input: ["text", "image"],
|
|
7784
|
+
cost: {
|
|
7785
|
+
input: 0.14,
|
|
7786
|
+
output: 0.39999999999999997,
|
|
7787
|
+
cacheRead: 0,
|
|
7788
|
+
cacheWrite: 0,
|
|
7789
|
+
},
|
|
7790
|
+
contextWindow: 262144,
|
|
7791
|
+
maxTokens: 131072,
|
|
7792
|
+
},
|
|
7606
7793
|
"inception/mercury": {
|
|
7607
7794
|
id: "inception/mercury",
|
|
7608
7795
|
name: "Inception: Mercury",
|
|
@@ -7654,22 +7841,22 @@ export const MODELS = {
|
|
|
7654
7841
|
contextWindow: 128000,
|
|
7655
7842
|
maxTokens: 32000,
|
|
7656
7843
|
},
|
|
7657
|
-
"kwaipilot/kat-coder-pro": {
|
|
7658
|
-
id: "kwaipilot/kat-coder-pro",
|
|
7659
|
-
name: "Kwaipilot: KAT-Coder-Pro
|
|
7844
|
+
"kwaipilot/kat-coder-pro-v2": {
|
|
7845
|
+
id: "kwaipilot/kat-coder-pro-v2",
|
|
7846
|
+
name: "Kwaipilot: KAT-Coder-Pro V2",
|
|
7660
7847
|
api: "openai-completions",
|
|
7661
7848
|
provider: "openrouter",
|
|
7662
7849
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
7663
7850
|
reasoning: false,
|
|
7664
7851
|
input: ["text"],
|
|
7665
7852
|
cost: {
|
|
7666
|
-
input: 0.
|
|
7667
|
-
output:
|
|
7668
|
-
cacheRead: 0.
|
|
7853
|
+
input: 0.3,
|
|
7854
|
+
output: 1.2,
|
|
7855
|
+
cacheRead: 0.06,
|
|
7669
7856
|
cacheWrite: 0,
|
|
7670
7857
|
},
|
|
7671
7858
|
contextWindow: 256000,
|
|
7672
|
-
maxTokens:
|
|
7859
|
+
maxTokens: 80000,
|
|
7673
7860
|
},
|
|
7674
7861
|
"meituan/longcat-flash-chat": {
|
|
7675
7862
|
id: "meituan/longcat-flash-chat",
|
|
@@ -7867,9 +8054,9 @@ export const MODELS = {
|
|
|
7867
8054
|
reasoning: true,
|
|
7868
8055
|
input: ["text"],
|
|
7869
8056
|
cost: {
|
|
7870
|
-
input: 0.
|
|
7871
|
-
output:
|
|
7872
|
-
cacheRead: 0.
|
|
8057
|
+
input: 0.118,
|
|
8058
|
+
output: 0.9900000000000001,
|
|
8059
|
+
cacheRead: 0.059,
|
|
7873
8060
|
cacheWrite: 0,
|
|
7874
8061
|
},
|
|
7875
8062
|
contextWindow: 196608,
|
|
@@ -8164,23 +8351,6 @@ export const MODELS = {
|
|
|
8164
8351
|
contextWindow: 32768,
|
|
8165
8352
|
maxTokens: 4096,
|
|
8166
8353
|
},
|
|
8167
|
-
"mistralai/mistral-small-24b-instruct-2501": {
|
|
8168
|
-
id: "mistralai/mistral-small-24b-instruct-2501",
|
|
8169
|
-
name: "Mistral: Mistral Small 3",
|
|
8170
|
-
api: "openai-completions",
|
|
8171
|
-
provider: "openrouter",
|
|
8172
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
8173
|
-
reasoning: false,
|
|
8174
|
-
input: ["text"],
|
|
8175
|
-
cost: {
|
|
8176
|
-
input: 0.049999999999999996,
|
|
8177
|
-
output: 0.08,
|
|
8178
|
-
cacheRead: 0,
|
|
8179
|
-
cacheWrite: 0,
|
|
8180
|
-
},
|
|
8181
|
-
contextWindow: 32768,
|
|
8182
|
-
maxTokens: 16384,
|
|
8183
|
-
},
|
|
8184
8354
|
"mistralai/mistral-small-2603": {
|
|
8185
8355
|
id: "mistralai/mistral-small-2603",
|
|
8186
8356
|
name: "Mistral: Mistral Small 4",
|
|
@@ -8198,23 +8368,6 @@ export const MODELS = {
|
|
|
8198
8368
|
contextWindow: 262144,
|
|
8199
8369
|
maxTokens: 4096,
|
|
8200
8370
|
},
|
|
8201
|
-
"mistralai/mistral-small-3.1-24b-instruct:free": {
|
|
8202
|
-
id: "mistralai/mistral-small-3.1-24b-instruct:free",
|
|
8203
|
-
name: "Mistral: Mistral Small 3.1 24B (free)",
|
|
8204
|
-
api: "openai-completions",
|
|
8205
|
-
provider: "openrouter",
|
|
8206
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
8207
|
-
reasoning: false,
|
|
8208
|
-
input: ["text", "image"],
|
|
8209
|
-
cost: {
|
|
8210
|
-
input: 0,
|
|
8211
|
-
output: 0,
|
|
8212
|
-
cacheRead: 0,
|
|
8213
|
-
cacheWrite: 0,
|
|
8214
|
-
},
|
|
8215
|
-
contextWindow: 128000,
|
|
8216
|
-
maxTokens: 4096,
|
|
8217
|
-
},
|
|
8218
8371
|
"mistralai/mistral-small-3.2-24b-instruct": {
|
|
8219
8372
|
id: "mistralai/mistral-small-3.2-24b-instruct",
|
|
8220
8373
|
name: "Mistral: Mistral Small 3.2 24B",
|
|
@@ -9252,6 +9405,40 @@ export const MODELS = {
|
|
|
9252
9405
|
contextWindow: 1050000,
|
|
9253
9406
|
maxTokens: 128000,
|
|
9254
9407
|
},
|
|
9408
|
+
"openai/gpt-audio": {
|
|
9409
|
+
id: "openai/gpt-audio",
|
|
9410
|
+
name: "OpenAI: GPT Audio",
|
|
9411
|
+
api: "openai-completions",
|
|
9412
|
+
provider: "openrouter",
|
|
9413
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
9414
|
+
reasoning: false,
|
|
9415
|
+
input: ["text"],
|
|
9416
|
+
cost: {
|
|
9417
|
+
input: 2.5,
|
|
9418
|
+
output: 10,
|
|
9419
|
+
cacheRead: 0,
|
|
9420
|
+
cacheWrite: 0,
|
|
9421
|
+
},
|
|
9422
|
+
contextWindow: 128000,
|
|
9423
|
+
maxTokens: 16384,
|
|
9424
|
+
},
|
|
9425
|
+
"openai/gpt-audio-mini": {
|
|
9426
|
+
id: "openai/gpt-audio-mini",
|
|
9427
|
+
name: "OpenAI: GPT Audio Mini",
|
|
9428
|
+
api: "openai-completions",
|
|
9429
|
+
provider: "openrouter",
|
|
9430
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
9431
|
+
reasoning: false,
|
|
9432
|
+
input: ["text"],
|
|
9433
|
+
cost: {
|
|
9434
|
+
input: 0.6,
|
|
9435
|
+
output: 2.4,
|
|
9436
|
+
cacheRead: 0,
|
|
9437
|
+
cacheWrite: 0,
|
|
9438
|
+
},
|
|
9439
|
+
contextWindow: 128000,
|
|
9440
|
+
maxTokens: 16384,
|
|
9441
|
+
},
|
|
9255
9442
|
"openai/gpt-oss-120b": {
|
|
9256
9443
|
id: "openai/gpt-oss-120b",
|
|
9257
9444
|
name: "OpenAI: gpt-oss-120b",
|
|
@@ -9813,23 +10000,6 @@ export const MODELS = {
|
|
|
9813
10000
|
contextWindow: 40960,
|
|
9814
10001
|
maxTokens: 40960,
|
|
9815
10002
|
},
|
|
9816
|
-
"qwen/qwen3-4b:free": {
|
|
9817
|
-
id: "qwen/qwen3-4b:free",
|
|
9818
|
-
name: "Qwen: Qwen3 4B (free)",
|
|
9819
|
-
api: "openai-completions",
|
|
9820
|
-
provider: "openrouter",
|
|
9821
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
9822
|
-
reasoning: true,
|
|
9823
|
-
input: ["text"],
|
|
9824
|
-
cost: {
|
|
9825
|
-
input: 0,
|
|
9826
|
-
output: 0,
|
|
9827
|
-
cacheRead: 0,
|
|
9828
|
-
cacheWrite: 0,
|
|
9829
|
-
},
|
|
9830
|
-
contextWindow: 40960,
|
|
9831
|
-
maxTokens: 4096,
|
|
9832
|
-
},
|
|
9833
10003
|
"qwen/qwen3-8b": {
|
|
9834
10004
|
id: "qwen/qwen3-8b",
|
|
9835
10005
|
name: "Qwen: Qwen3 8B",
|
|
@@ -10236,7 +10406,7 @@ export const MODELS = {
|
|
|
10236
10406
|
cacheWrite: 0,
|
|
10237
10407
|
},
|
|
10238
10408
|
contextWindow: 256000,
|
|
10239
|
-
maxTokens:
|
|
10409
|
+
maxTokens: 32768,
|
|
10240
10410
|
},
|
|
10241
10411
|
"qwen/qwen3.5-flash-02-23": {
|
|
10242
10412
|
id: "qwen/qwen3.5-flash-02-23",
|
|
@@ -10272,6 +10442,23 @@ export const MODELS = {
|
|
|
10272
10442
|
contextWindow: 1000000,
|
|
10273
10443
|
maxTokens: 65536,
|
|
10274
10444
|
},
|
|
10445
|
+
"qwen/qwen3.6-plus:free": {
|
|
10446
|
+
id: "qwen/qwen3.6-plus:free",
|
|
10447
|
+
name: "Qwen: Qwen3.6 Plus (free)",
|
|
10448
|
+
api: "openai-completions",
|
|
10449
|
+
provider: "openrouter",
|
|
10450
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
10451
|
+
reasoning: true,
|
|
10452
|
+
input: ["text", "image"],
|
|
10453
|
+
cost: {
|
|
10454
|
+
input: 0,
|
|
10455
|
+
output: 0,
|
|
10456
|
+
cacheRead: 0,
|
|
10457
|
+
cacheWrite: 0,
|
|
10458
|
+
},
|
|
10459
|
+
contextWindow: 1000000,
|
|
10460
|
+
maxTokens: 65536,
|
|
10461
|
+
},
|
|
10275
10462
|
"qwen/qwq-32b": {
|
|
10276
10463
|
id: "qwen/qwq-32b",
|
|
10277
10464
|
name: "Qwen: QwQ 32B",
|
|
@@ -10289,6 +10476,23 @@ export const MODELS = {
|
|
|
10289
10476
|
contextWindow: 131072,
|
|
10290
10477
|
maxTokens: 131072,
|
|
10291
10478
|
},
|
|
10479
|
+
"rekaai/reka-edge": {
|
|
10480
|
+
id: "rekaai/reka-edge",
|
|
10481
|
+
name: "Reka Edge",
|
|
10482
|
+
api: "openai-completions",
|
|
10483
|
+
provider: "openrouter",
|
|
10484
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
10485
|
+
reasoning: false,
|
|
10486
|
+
input: ["text", "image"],
|
|
10487
|
+
cost: {
|
|
10488
|
+
input: 0.09999999999999999,
|
|
10489
|
+
output: 0.09999999999999999,
|
|
10490
|
+
cacheRead: 0,
|
|
10491
|
+
cacheWrite: 0,
|
|
10492
|
+
},
|
|
10493
|
+
contextWindow: 16384,
|
|
10494
|
+
maxTokens: 16384,
|
|
10495
|
+
},
|
|
10292
10496
|
"relace/relace-search": {
|
|
10293
10497
|
id: "relace/relace-search",
|
|
10294
10498
|
name: "Relace: Relace Search",
|
|
@@ -10351,11 +10555,11 @@ export const MODELS = {
|
|
|
10351
10555
|
cost: {
|
|
10352
10556
|
input: 0.09999999999999999,
|
|
10353
10557
|
output: 0.3,
|
|
10354
|
-
cacheRead: 0
|
|
10558
|
+
cacheRead: 0,
|
|
10355
10559
|
cacheWrite: 0,
|
|
10356
10560
|
},
|
|
10357
|
-
contextWindow:
|
|
10358
|
-
maxTokens:
|
|
10561
|
+
contextWindow: 262144,
|
|
10562
|
+
maxTokens: 65536,
|
|
10359
10563
|
},
|
|
10360
10564
|
"stepfun/step-3.5-flash:free": {
|
|
10361
10565
|
id: "stepfun/step-3.5-flash:free",
|
|
@@ -10561,9 +10765,9 @@ export const MODELS = {
|
|
|
10561
10765
|
contextWindow: 2000000,
|
|
10562
10766
|
maxTokens: 30000,
|
|
10563
10767
|
},
|
|
10564
|
-
"x-ai/grok-4.20
|
|
10565
|
-
id: "x-ai/grok-4.20
|
|
10566
|
-
name: "xAI: Grok 4.20
|
|
10768
|
+
"x-ai/grok-4.20": {
|
|
10769
|
+
id: "x-ai/grok-4.20",
|
|
10770
|
+
name: "xAI: Grok 4.20",
|
|
10567
10771
|
api: "openai-completions",
|
|
10568
10772
|
provider: "openrouter",
|
|
10569
10773
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -10833,6 +11037,23 @@ export const MODELS = {
|
|
|
10833
11037
|
contextWindow: 202752,
|
|
10834
11038
|
maxTokens: 131072,
|
|
10835
11039
|
},
|
|
11040
|
+
"z-ai/glm-5v-turbo": {
|
|
11041
|
+
id: "z-ai/glm-5v-turbo",
|
|
11042
|
+
name: "Z.ai: GLM 5V Turbo",
|
|
11043
|
+
api: "openai-completions",
|
|
11044
|
+
provider: "openrouter",
|
|
11045
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
11046
|
+
reasoning: true,
|
|
11047
|
+
input: ["text", "image"],
|
|
11048
|
+
cost: {
|
|
11049
|
+
input: 1.2,
|
|
11050
|
+
output: 4,
|
|
11051
|
+
cacheRead: 0.24,
|
|
11052
|
+
cacheWrite: 0,
|
|
11053
|
+
},
|
|
11054
|
+
contextWindow: 202752,
|
|
11055
|
+
maxTokens: 131072,
|
|
11056
|
+
},
|
|
10836
11057
|
},
|
|
10837
11058
|
"vercel-ai-gateway": {
|
|
10838
11059
|
"alibaba/qwen-3-14b": {
|
|
@@ -10854,19 +11075,19 @@ export const MODELS = {
|
|
|
10854
11075
|
},
|
|
10855
11076
|
"alibaba/qwen-3-235b": {
|
|
10856
11077
|
id: "alibaba/qwen-3-235b",
|
|
10857
|
-
name: "Qwen3
|
|
11078
|
+
name: "Qwen3 235B A22b Instruct 2507",
|
|
10858
11079
|
api: "anthropic-messages",
|
|
10859
11080
|
provider: "vercel-ai-gateway",
|
|
10860
11081
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
10861
11082
|
reasoning: false,
|
|
10862
11083
|
input: ["text"],
|
|
10863
11084
|
cost: {
|
|
10864
|
-
input: 0.
|
|
10865
|
-
output: 0.
|
|
10866
|
-
cacheRead: 0,
|
|
11085
|
+
input: 0.22,
|
|
11086
|
+
output: 0.88,
|
|
11087
|
+
cacheRead: 0.11,
|
|
10867
11088
|
cacheWrite: 0,
|
|
10868
11089
|
},
|
|
10869
|
-
contextWindow:
|
|
11090
|
+
contextWindow: 32768,
|
|
10870
11091
|
maxTokens: 16384,
|
|
10871
11092
|
},
|
|
10872
11093
|
"alibaba/qwen-3-30b": {
|
|
@@ -10895,13 +11116,13 @@ export const MODELS = {
|
|
|
10895
11116
|
reasoning: true,
|
|
10896
11117
|
input: ["text"],
|
|
10897
11118
|
cost: {
|
|
10898
|
-
input: 0.
|
|
10899
|
-
output: 0.
|
|
10900
|
-
cacheRead: 0
|
|
11119
|
+
input: 0.16,
|
|
11120
|
+
output: 0.64,
|
|
11121
|
+
cacheRead: 0,
|
|
10901
11122
|
cacheWrite: 0,
|
|
10902
11123
|
},
|
|
10903
|
-
contextWindow:
|
|
10904
|
-
maxTokens:
|
|
11124
|
+
contextWindow: 128000,
|
|
11125
|
+
maxTokens: 8192,
|
|
10905
11126
|
},
|
|
10906
11127
|
"alibaba/qwen3-235b-a22b-thinking": {
|
|
10907
11128
|
id: "alibaba/qwen3-235b-a22b-thinking",
|
|
@@ -10929,13 +11150,13 @@ export const MODELS = {
|
|
|
10929
11150
|
reasoning: false,
|
|
10930
11151
|
input: ["text"],
|
|
10931
11152
|
cost: {
|
|
10932
|
-
input:
|
|
10933
|
-
output:
|
|
10934
|
-
cacheRead: 0.
|
|
11153
|
+
input: 1.5,
|
|
11154
|
+
output: 7.5,
|
|
11155
|
+
cacheRead: 0.3,
|
|
10935
11156
|
cacheWrite: 0,
|
|
10936
11157
|
},
|
|
10937
11158
|
contextWindow: 262144,
|
|
10938
|
-
maxTokens:
|
|
11159
|
+
maxTokens: 65536,
|
|
10939
11160
|
},
|
|
10940
11161
|
"alibaba/qwen3-coder-30b-a3b": {
|
|
10941
11162
|
id: "alibaba/qwen3-coder-30b-a3b",
|
|
@@ -11048,13 +11269,13 @@ export const MODELS = {
|
|
|
11048
11269
|
reasoning: true,
|
|
11049
11270
|
input: ["text", "image"],
|
|
11050
11271
|
cost: {
|
|
11051
|
-
input: 0.
|
|
11052
|
-
output:
|
|
11272
|
+
input: 0.39999999999999997,
|
|
11273
|
+
output: 4,
|
|
11053
11274
|
cacheRead: 0,
|
|
11054
11275
|
cacheWrite: 0,
|
|
11055
11276
|
},
|
|
11056
|
-
contextWindow:
|
|
11057
|
-
maxTokens:
|
|
11277
|
+
contextWindow: 131072,
|
|
11278
|
+
maxTokens: 32768,
|
|
11058
11279
|
},
|
|
11059
11280
|
"alibaba/qwen3.5-flash": {
|
|
11060
11281
|
id: "alibaba/qwen3.5-flash",
|
|
@@ -11090,6 +11311,23 @@ export const MODELS = {
|
|
|
11090
11311
|
contextWindow: 1000000,
|
|
11091
11312
|
maxTokens: 64000,
|
|
11092
11313
|
},
|
|
11314
|
+
"alibaba/qwen3.6-plus": {
|
|
11315
|
+
id: "alibaba/qwen3.6-plus",
|
|
11316
|
+
name: "Qwen 3.6 Plus",
|
|
11317
|
+
api: "anthropic-messages",
|
|
11318
|
+
provider: "vercel-ai-gateway",
|
|
11319
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11320
|
+
reasoning: true,
|
|
11321
|
+
input: ["text", "image"],
|
|
11322
|
+
cost: {
|
|
11323
|
+
input: 0.5,
|
|
11324
|
+
output: 3,
|
|
11325
|
+
cacheRead: 0.09999999999999999,
|
|
11326
|
+
cacheWrite: 0.625,
|
|
11327
|
+
},
|
|
11328
|
+
contextWindow: 1000000,
|
|
11329
|
+
maxTokens: 64000,
|
|
11330
|
+
},
|
|
11093
11331
|
"anthropic/claude-3-haiku": {
|
|
11094
11332
|
id: "anthropic/claude-3-haiku",
|
|
11095
11333
|
name: "Claude 3 Haiku",
|
|
@@ -11124,40 +11362,6 @@ export const MODELS = {
|
|
|
11124
11362
|
contextWindow: 200000,
|
|
11125
11363
|
maxTokens: 8192,
|
|
11126
11364
|
},
|
|
11127
|
-
"anthropic/claude-3.5-sonnet": {
|
|
11128
|
-
id: "anthropic/claude-3.5-sonnet",
|
|
11129
|
-
name: "Claude 3.5 Sonnet",
|
|
11130
|
-
api: "anthropic-messages",
|
|
11131
|
-
provider: "vercel-ai-gateway",
|
|
11132
|
-
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11133
|
-
reasoning: false,
|
|
11134
|
-
input: ["text", "image"],
|
|
11135
|
-
cost: {
|
|
11136
|
-
input: 3,
|
|
11137
|
-
output: 15,
|
|
11138
|
-
cacheRead: 0.3,
|
|
11139
|
-
cacheWrite: 3.75,
|
|
11140
|
-
},
|
|
11141
|
-
contextWindow: 200000,
|
|
11142
|
-
maxTokens: 8192,
|
|
11143
|
-
},
|
|
11144
|
-
"anthropic/claude-3.5-sonnet-20240620": {
|
|
11145
|
-
id: "anthropic/claude-3.5-sonnet-20240620",
|
|
11146
|
-
name: "Claude 3.5 Sonnet (2024-06-20)",
|
|
11147
|
-
api: "anthropic-messages",
|
|
11148
|
-
provider: "vercel-ai-gateway",
|
|
11149
|
-
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11150
|
-
reasoning: false,
|
|
11151
|
-
input: ["text", "image"],
|
|
11152
|
-
cost: {
|
|
11153
|
-
input: 3,
|
|
11154
|
-
output: 15,
|
|
11155
|
-
cacheRead: 0.3,
|
|
11156
|
-
cacheWrite: 3.75,
|
|
11157
|
-
},
|
|
11158
|
-
contextWindow: 200000,
|
|
11159
|
-
maxTokens: 8192,
|
|
11160
|
-
},
|
|
11161
11365
|
"anthropic/claude-3.7-sonnet": {
|
|
11162
11366
|
id: "anthropic/claude-3.7-sonnet",
|
|
11163
11367
|
name: "Claude 3.7 Sonnet",
|
|
@@ -11328,6 +11532,23 @@ export const MODELS = {
|
|
|
11328
11532
|
contextWindow: 131000,
|
|
11329
11533
|
maxTokens: 131000,
|
|
11330
11534
|
},
|
|
11535
|
+
"arcee-ai/trinity-large-thinking": {
|
|
11536
|
+
id: "arcee-ai/trinity-large-thinking",
|
|
11537
|
+
name: "Trinity Large Thinking",
|
|
11538
|
+
api: "anthropic-messages",
|
|
11539
|
+
provider: "vercel-ai-gateway",
|
|
11540
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11541
|
+
reasoning: true,
|
|
11542
|
+
input: ["text"],
|
|
11543
|
+
cost: {
|
|
11544
|
+
input: 0.25,
|
|
11545
|
+
output: 0.8999999999999999,
|
|
11546
|
+
cacheRead: 0,
|
|
11547
|
+
cacheWrite: 0,
|
|
11548
|
+
},
|
|
11549
|
+
contextWindow: 262100,
|
|
11550
|
+
maxTokens: 80000,
|
|
11551
|
+
},
|
|
11331
11552
|
"bytedance/seed-1.6": {
|
|
11332
11553
|
id: "bytedance/seed-1.6",
|
|
11333
11554
|
name: "Seed 1.6",
|
|
@@ -11405,13 +11626,13 @@ export const MODELS = {
|
|
|
11405
11626
|
reasoning: true,
|
|
11406
11627
|
input: ["text"],
|
|
11407
11628
|
cost: {
|
|
11408
|
-
input: 0.
|
|
11409
|
-
output: 1.
|
|
11410
|
-
cacheRead: 0,
|
|
11629
|
+
input: 0.56,
|
|
11630
|
+
output: 1.68,
|
|
11631
|
+
cacheRead: 0.28,
|
|
11411
11632
|
cacheWrite: 0,
|
|
11412
11633
|
},
|
|
11413
11634
|
contextWindow: 163840,
|
|
11414
|
-
maxTokens:
|
|
11635
|
+
maxTokens: 8192,
|
|
11415
11636
|
},
|
|
11416
11637
|
"deepseek/deepseek-v3.1-terminus": {
|
|
11417
11638
|
id: "deepseek/deepseek-v3.1-terminus",
|
|
@@ -11617,6 +11838,40 @@ export const MODELS = {
|
|
|
11617
11838
|
contextWindow: 1000000,
|
|
11618
11839
|
maxTokens: 64000,
|
|
11619
11840
|
},
|
|
11841
|
+
"google/gemma-4-26b-a4b-it": {
|
|
11842
|
+
id: "google/gemma-4-26b-a4b-it",
|
|
11843
|
+
name: "Gemma 4 26B A4B IT",
|
|
11844
|
+
api: "anthropic-messages",
|
|
11845
|
+
provider: "vercel-ai-gateway",
|
|
11846
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11847
|
+
reasoning: false,
|
|
11848
|
+
input: ["text", "image"],
|
|
11849
|
+
cost: {
|
|
11850
|
+
input: 0.13,
|
|
11851
|
+
output: 0.39999999999999997,
|
|
11852
|
+
cacheRead: 0,
|
|
11853
|
+
cacheWrite: 0,
|
|
11854
|
+
},
|
|
11855
|
+
contextWindow: 262144,
|
|
11856
|
+
maxTokens: 131072,
|
|
11857
|
+
},
|
|
11858
|
+
"google/gemma-4-31b-it": {
|
|
11859
|
+
id: "google/gemma-4-31b-it",
|
|
11860
|
+
name: "Gemma 4 31B IT",
|
|
11861
|
+
api: "anthropic-messages",
|
|
11862
|
+
provider: "vercel-ai-gateway",
|
|
11863
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11864
|
+
reasoning: false,
|
|
11865
|
+
input: ["text", "image"],
|
|
11866
|
+
cost: {
|
|
11867
|
+
input: 0.14,
|
|
11868
|
+
output: 0.39999999999999997,
|
|
11869
|
+
cacheRead: 0,
|
|
11870
|
+
cacheWrite: 0,
|
|
11871
|
+
},
|
|
11872
|
+
contextWindow: 262144,
|
|
11873
|
+
maxTokens: 131072,
|
|
11874
|
+
},
|
|
11620
11875
|
"inception/mercury-2": {
|
|
11621
11876
|
id: "inception/mercury-2",
|
|
11622
11877
|
name: "Mercury 2",
|
|
@@ -11651,34 +11906,34 @@ export const MODELS = {
|
|
|
11651
11906
|
contextWindow: 32000,
|
|
11652
11907
|
maxTokens: 16384,
|
|
11653
11908
|
},
|
|
11654
|
-
"
|
|
11655
|
-
id: "
|
|
11656
|
-
name: "
|
|
11909
|
+
"kwaipilot/kat-coder-pro-v2": {
|
|
11910
|
+
id: "kwaipilot/kat-coder-pro-v2",
|
|
11911
|
+
name: "Kat Coder Pro V2",
|
|
11657
11912
|
api: "anthropic-messages",
|
|
11658
11913
|
provider: "vercel-ai-gateway",
|
|
11659
11914
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11660
|
-
reasoning:
|
|
11915
|
+
reasoning: true,
|
|
11661
11916
|
input: ["text"],
|
|
11662
11917
|
cost: {
|
|
11663
|
-
input: 0,
|
|
11664
|
-
output:
|
|
11665
|
-
cacheRead: 0,
|
|
11918
|
+
input: 0.3,
|
|
11919
|
+
output: 1.2,
|
|
11920
|
+
cacheRead: 0.06,
|
|
11666
11921
|
cacheWrite: 0,
|
|
11667
11922
|
},
|
|
11668
|
-
contextWindow:
|
|
11669
|
-
maxTokens:
|
|
11923
|
+
contextWindow: 256000,
|
|
11924
|
+
maxTokens: 256000,
|
|
11670
11925
|
},
|
|
11671
|
-
"meituan/longcat-flash-
|
|
11672
|
-
id: "meituan/longcat-flash-
|
|
11673
|
-
name: "LongCat Flash
|
|
11926
|
+
"meituan/longcat-flash-chat": {
|
|
11927
|
+
id: "meituan/longcat-flash-chat",
|
|
11928
|
+
name: "LongCat Flash Chat",
|
|
11674
11929
|
api: "anthropic-messages",
|
|
11675
11930
|
provider: "vercel-ai-gateway",
|
|
11676
11931
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11677
|
-
reasoning:
|
|
11932
|
+
reasoning: false,
|
|
11678
11933
|
input: ["text"],
|
|
11679
11934
|
cost: {
|
|
11680
|
-
input: 0
|
|
11681
|
-
output:
|
|
11935
|
+
input: 0,
|
|
11936
|
+
output: 0,
|
|
11682
11937
|
cacheRead: 0,
|
|
11683
11938
|
cacheWrite: 0,
|
|
11684
11939
|
},
|
|
@@ -11711,13 +11966,13 @@ export const MODELS = {
|
|
|
11711
11966
|
reasoning: false,
|
|
11712
11967
|
input: ["text"],
|
|
11713
11968
|
cost: {
|
|
11714
|
-
input: 0.
|
|
11715
|
-
output: 0.
|
|
11716
|
-
cacheRead: 0
|
|
11969
|
+
input: 0.22,
|
|
11970
|
+
output: 0.22,
|
|
11971
|
+
cacheRead: 0,
|
|
11717
11972
|
cacheWrite: 0,
|
|
11718
11973
|
},
|
|
11719
11974
|
contextWindow: 128000,
|
|
11720
|
-
maxTokens:
|
|
11975
|
+
maxTokens: 8192,
|
|
11721
11976
|
},
|
|
11722
11977
|
"meta/llama-3.2-11b": {
|
|
11723
11978
|
id: "meta/llama-3.2-11b",
|
|
@@ -12095,20 +12350,20 @@ export const MODELS = {
|
|
|
12095
12350
|
},
|
|
12096
12351
|
"moonshotai/kimi-k2": {
|
|
12097
12352
|
id: "moonshotai/kimi-k2",
|
|
12098
|
-
name: "Kimi K2",
|
|
12353
|
+
name: "Kimi K2 Instruct",
|
|
12099
12354
|
api: "anthropic-messages",
|
|
12100
12355
|
provider: "vercel-ai-gateway",
|
|
12101
12356
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12102
12357
|
reasoning: false,
|
|
12103
12358
|
input: ["text"],
|
|
12104
12359
|
cost: {
|
|
12105
|
-
input: 0.
|
|
12106
|
-
output: 2.
|
|
12107
|
-
cacheRead: 0
|
|
12360
|
+
input: 0.5700000000000001,
|
|
12361
|
+
output: 2.3,
|
|
12362
|
+
cacheRead: 0,
|
|
12108
12363
|
cacheWrite: 0,
|
|
12109
12364
|
},
|
|
12110
12365
|
contextWindow: 131072,
|
|
12111
|
-
maxTokens:
|
|
12366
|
+
maxTokens: 131072,
|
|
12112
12367
|
},
|
|
12113
12368
|
"moonshotai/kimi-k2-0905": {
|
|
12114
12369
|
id: "moonshotai/kimi-k2-0905",
|
|
@@ -12121,11 +12376,11 @@ export const MODELS = {
|
|
|
12121
12376
|
cost: {
|
|
12122
12377
|
input: 0.6,
|
|
12123
12378
|
output: 2.5,
|
|
12124
|
-
cacheRead: 0.
|
|
12379
|
+
cacheRead: 0.3,
|
|
12125
12380
|
cacheWrite: 0,
|
|
12126
12381
|
},
|
|
12127
12382
|
contextWindow: 256000,
|
|
12128
|
-
maxTokens:
|
|
12383
|
+
maxTokens: 128000,
|
|
12129
12384
|
},
|
|
12130
12385
|
"moonshotai/kimi-k2-thinking": {
|
|
12131
12386
|
id: "moonshotai/kimi-k2-thinking",
|
|
@@ -12688,6 +12943,23 @@ export const MODELS = {
|
|
|
12688
12943
|
contextWindow: 1050000,
|
|
12689
12944
|
maxTokens: 128000,
|
|
12690
12945
|
},
|
|
12946
|
+
"openai/gpt-oss-120b": {
|
|
12947
|
+
id: "openai/gpt-oss-120b",
|
|
12948
|
+
name: "gpt-oss-120b",
|
|
12949
|
+
api: "anthropic-messages",
|
|
12950
|
+
provider: "vercel-ai-gateway",
|
|
12951
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12952
|
+
reasoning: true,
|
|
12953
|
+
input: ["text"],
|
|
12954
|
+
cost: {
|
|
12955
|
+
input: 0.15,
|
|
12956
|
+
output: 0.6,
|
|
12957
|
+
cacheRead: 0,
|
|
12958
|
+
cacheWrite: 0,
|
|
12959
|
+
},
|
|
12960
|
+
contextWindow: 128000,
|
|
12961
|
+
maxTokens: 8192,
|
|
12962
|
+
},
|
|
12691
12963
|
"openai/gpt-oss-20b": {
|
|
12692
12964
|
id: "openai/gpt-oss-20b",
|
|
12693
12965
|
name: "gpt-oss-20b",
|
|
@@ -12697,12 +12969,12 @@ export const MODELS = {
|
|
|
12697
12969
|
reasoning: true,
|
|
12698
12970
|
input: ["text"],
|
|
12699
12971
|
cost: {
|
|
12700
|
-
input: 0.
|
|
12701
|
-
output: 0.
|
|
12972
|
+
input: 0.049999999999999996,
|
|
12973
|
+
output: 0.19999999999999998,
|
|
12702
12974
|
cacheRead: 0,
|
|
12703
12975
|
cacheWrite: 0,
|
|
12704
12976
|
},
|
|
12705
|
-
contextWindow:
|
|
12977
|
+
contextWindow: 131072,
|
|
12706
12978
|
maxTokens: 8192,
|
|
12707
12979
|
},
|
|
12708
12980
|
"openai/gpt-oss-safeguard-20b": {
|
|
@@ -13173,9 +13445,9 @@ export const MODELS = {
|
|
|
13173
13445
|
reasoning: true,
|
|
13174
13446
|
input: ["text"],
|
|
13175
13447
|
cost: {
|
|
13176
|
-
input: 0.
|
|
13177
|
-
output: 0.
|
|
13178
|
-
cacheRead: 0.
|
|
13448
|
+
input: 0.09,
|
|
13449
|
+
output: 0.29,
|
|
13450
|
+
cacheRead: 0.045,
|
|
13179
13451
|
cacheWrite: 0,
|
|
13180
13452
|
},
|
|
13181
13453
|
contextWindow: 262144,
|
|
@@ -13311,7 +13583,7 @@ export const MODELS = {
|
|
|
13311
13583
|
cost: {
|
|
13312
13584
|
input: 0.6,
|
|
13313
13585
|
output: 2.2,
|
|
13314
|
-
cacheRead: 0,
|
|
13586
|
+
cacheRead: 0.11,
|
|
13315
13587
|
cacheWrite: 0,
|
|
13316
13588
|
},
|
|
13317
13589
|
contextWindow: 200000,
|
|
@@ -13385,6 +13657,23 @@ export const MODELS = {
|
|
|
13385
13657
|
contextWindow: 202800,
|
|
13386
13658
|
maxTokens: 131100,
|
|
13387
13659
|
},
|
|
13660
|
+
"zai/glm-5v-turbo": {
|
|
13661
|
+
id: "zai/glm-5v-turbo",
|
|
13662
|
+
name: "GLM 5V Turbo",
|
|
13663
|
+
api: "anthropic-messages",
|
|
13664
|
+
provider: "vercel-ai-gateway",
|
|
13665
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
13666
|
+
reasoning: true,
|
|
13667
|
+
input: ["text", "image"],
|
|
13668
|
+
cost: {
|
|
13669
|
+
input: 1.2,
|
|
13670
|
+
output: 4,
|
|
13671
|
+
cacheRead: 0.24,
|
|
13672
|
+
cacheWrite: 0,
|
|
13673
|
+
},
|
|
13674
|
+
contextWindow: 200000,
|
|
13675
|
+
maxTokens: 128000,
|
|
13676
|
+
},
|
|
13388
13677
|
},
|
|
13389
13678
|
"xai": {
|
|
13390
13679
|
"grok-2": {
|
|
@@ -13875,7 +14164,7 @@ export const MODELS = {
|
|
|
13875
14164
|
api: "openai-completions",
|
|
13876
14165
|
provider: "zai",
|
|
13877
14166
|
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
13878
|
-
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
|
|
14167
|
+
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
13879
14168
|
reasoning: true,
|
|
13880
14169
|
input: ["text"],
|
|
13881
14170
|
cost: {
|
|
@@ -13893,7 +14182,7 @@ export const MODELS = {
|
|
|
13893
14182
|
api: "openai-completions",
|
|
13894
14183
|
provider: "zai",
|
|
13895
14184
|
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
13896
|
-
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
|
|
14185
|
+
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
13897
14186
|
reasoning: true,
|
|
13898
14187
|
input: ["text", "image"],
|
|
13899
14188
|
cost: {
|
|
@@ -13911,7 +14200,7 @@ export const MODELS = {
|
|
|
13911
14200
|
api: "openai-completions",
|
|
13912
14201
|
provider: "zai",
|
|
13913
14202
|
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
13914
|
-
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
|
|
14203
|
+
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
13915
14204
|
reasoning: true,
|
|
13916
14205
|
input: ["text"],
|
|
13917
14206
|
cost: {
|
|
@@ -13929,7 +14218,7 @@ export const MODELS = {
|
|
|
13929
14218
|
api: "openai-completions",
|
|
13930
14219
|
provider: "zai",
|
|
13931
14220
|
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
13932
|
-
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
|
|
14221
|
+
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
13933
14222
|
reasoning: true,
|
|
13934
14223
|
input: ["text"],
|
|
13935
14224
|
cost: {
|
|
@@ -13941,13 +14230,31 @@ export const MODELS = {
|
|
|
13941
14230
|
contextWindow: 200000,
|
|
13942
14231
|
maxTokens: 131072,
|
|
13943
14232
|
},
|
|
14233
|
+
"glm-4.7-flashx": {
|
|
14234
|
+
id: "glm-4.7-flashx",
|
|
14235
|
+
name: "GLM-4.7-FlashX",
|
|
14236
|
+
api: "openai-completions",
|
|
14237
|
+
provider: "zai",
|
|
14238
|
+
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
14239
|
+
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
14240
|
+
reasoning: true,
|
|
14241
|
+
input: ["text"],
|
|
14242
|
+
cost: {
|
|
14243
|
+
input: 0.07,
|
|
14244
|
+
output: 0.4,
|
|
14245
|
+
cacheRead: 0.01,
|
|
14246
|
+
cacheWrite: 0,
|
|
14247
|
+
},
|
|
14248
|
+
contextWindow: 200000,
|
|
14249
|
+
maxTokens: 131072,
|
|
14250
|
+
},
|
|
13944
14251
|
"glm-5": {
|
|
13945
14252
|
id: "glm-5",
|
|
13946
14253
|
name: "GLM-5",
|
|
13947
14254
|
api: "openai-completions",
|
|
13948
14255
|
provider: "zai",
|
|
13949
14256
|
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
13950
|
-
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
|
|
14257
|
+
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
13951
14258
|
reasoning: true,
|
|
13952
14259
|
input: ["text"],
|
|
13953
14260
|
cost: {
|
|
@@ -13965,7 +14272,7 @@ export const MODELS = {
|
|
|
13965
14272
|
api: "openai-completions",
|
|
13966
14273
|
provider: "zai",
|
|
13967
14274
|
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
13968
|
-
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
|
|
14275
|
+
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
13969
14276
|
reasoning: true,
|
|
13970
14277
|
input: ["text"],
|
|
13971
14278
|
cost: {
|
|
@@ -13977,6 +14284,24 @@ export const MODELS = {
|
|
|
13977
14284
|
contextWindow: 200000,
|
|
13978
14285
|
maxTokens: 131072,
|
|
13979
14286
|
},
|
|
14287
|
+
"glm-5v-turbo": {
|
|
14288
|
+
id: "glm-5v-turbo",
|
|
14289
|
+
name: "glm-5v-turbo",
|
|
14290
|
+
api: "openai-completions",
|
|
14291
|
+
provider: "zai",
|
|
14292
|
+
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
14293
|
+
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
14294
|
+
reasoning: true,
|
|
14295
|
+
input: ["text", "image"],
|
|
14296
|
+
cost: {
|
|
14297
|
+
input: 1.2,
|
|
14298
|
+
output: 4,
|
|
14299
|
+
cacheRead: 0.24,
|
|
14300
|
+
cacheWrite: 0,
|
|
14301
|
+
},
|
|
14302
|
+
contextWindow: 200000,
|
|
14303
|
+
maxTokens: 131072,
|
|
14304
|
+
},
|
|
13980
14305
|
},
|
|
13981
14306
|
};
|
|
13982
14307
|
//# sourceMappingURL=models.generated.js.map
|