@draht/ai 2026.3.25-1 → 2026.4.5

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -815,8 +815,8 @@ export const MODELS = {
815
815
  cacheRead: 0,
816
816
  cacheWrite: 0,
817
817
  },
818
- contextWindow: 1000000,
819
- maxTokens: 131072,
818
+ contextWindow: 196608,
819
+ maxTokens: 98304,
820
820
  },
821
821
  "mistral.devstral-2-123b": {
822
822
  id: "mistral.devstral-2-123b",
@@ -1056,6 +1056,23 @@ export const MODELS = {
1056
1056
  contextWindow: 128000,
1057
1057
  maxTokens: 4096,
1058
1058
  },
1059
+ "nvidia.nemotron-super-3-120b": {
1060
+ id: "nvidia.nemotron-super-3-120b",
1061
+ name: "NVIDIA Nemotron 3 Super 120B A12B",
1062
+ api: "bedrock-converse-stream",
1063
+ provider: "amazon-bedrock",
1064
+ baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
1065
+ reasoning: true,
1066
+ input: ["text"],
1067
+ cost: {
1068
+ input: 0.15,
1069
+ output: 0.65,
1070
+ cacheRead: 0,
1071
+ cacheWrite: 0,
1072
+ },
1073
+ contextWindow: 262144,
1074
+ maxTokens: 131072,
1075
+ },
1059
1076
  "openai.gpt-oss-120b-1:0": {
1060
1077
  id: "openai.gpt-oss-120b-1:0",
1061
1078
  name: "gpt-oss-120b",
@@ -1444,8 +1461,8 @@ export const MODELS = {
1444
1461
  cacheRead: 0,
1445
1462
  cacheWrite: 0,
1446
1463
  },
1447
- contextWindow: 200000,
1448
- maxTokens: 131072,
1464
+ contextWindow: 202752,
1465
+ maxTokens: 101376,
1449
1466
  },
1450
1467
  },
1451
1468
  "anthropic": {
@@ -1534,23 +1551,6 @@ export const MODELS = {
1534
1551
  contextWindow: 200000,
1535
1552
  maxTokens: 64000,
1536
1553
  },
1537
- "claude-3-7-sonnet-latest": {
1538
- id: "claude-3-7-sonnet-latest",
1539
- name: "Claude Sonnet 3.7 (latest)",
1540
- api: "anthropic-messages",
1541
- provider: "anthropic",
1542
- baseUrl: "https://api.anthropic.com",
1543
- reasoning: true,
1544
- input: ["text", "image"],
1545
- cost: {
1546
- input: 3,
1547
- output: 15,
1548
- cacheRead: 0.3,
1549
- cacheWrite: 3.75,
1550
- },
1551
- contextWindow: 200000,
1552
- maxTokens: 64000,
1553
- },
1554
1554
  "claude-3-haiku-20240307": {
1555
1555
  id: "claude-3-haiku-20240307",
1556
1556
  name: "Claude Haiku 3",
@@ -2284,6 +2284,23 @@ export const MODELS = {
2284
2284
  contextWindow: 400000,
2285
2285
  maxTokens: 128000,
2286
2286
  },
2287
+ "gpt-5.3-chat-latest": {
2288
+ id: "gpt-5.3-chat-latest",
2289
+ name: "GPT-5.3 Chat (latest)",
2290
+ api: "azure-openai-responses",
2291
+ provider: "azure-openai-responses",
2292
+ baseUrl: "",
2293
+ reasoning: false,
2294
+ input: ["text", "image"],
2295
+ cost: {
2296
+ input: 1.75,
2297
+ output: 14,
2298
+ cacheRead: 0.175,
2299
+ cacheWrite: 0,
2300
+ },
2301
+ contextWindow: 128000,
2302
+ maxTokens: 16384,
2303
+ },
2287
2304
  "gpt-5.3-codex": {
2288
2305
  id: "gpt-5.3-codex",
2289
2306
  name: "GPT-5.3 Codex",
@@ -3443,6 +3460,57 @@ export const MODELS = {
3443
3460
  contextWindow: 131072,
3444
3461
  maxTokens: 65536,
3445
3462
  },
3463
+ "gemma-3-27b-it": {
3464
+ id: "gemma-3-27b-it",
3465
+ name: "Gemma 3 27B",
3466
+ api: "google-generative-ai",
3467
+ provider: "google",
3468
+ baseUrl: "https://generativelanguage.googleapis.com/v1beta",
3469
+ reasoning: false,
3470
+ input: ["text", "image"],
3471
+ cost: {
3472
+ input: 0,
3473
+ output: 0,
3474
+ cacheRead: 0,
3475
+ cacheWrite: 0,
3476
+ },
3477
+ contextWindow: 131072,
3478
+ maxTokens: 8192,
3479
+ },
3480
+ "gemma-4-26b": {
3481
+ id: "gemma-4-26b",
3482
+ name: "Gemma 4 26B",
3483
+ api: "google-generative-ai",
3484
+ provider: "google",
3485
+ baseUrl: "https://generativelanguage.googleapis.com/v1beta",
3486
+ reasoning: true,
3487
+ input: ["text", "image"],
3488
+ cost: {
3489
+ input: 0,
3490
+ output: 0,
3491
+ cacheRead: 0,
3492
+ cacheWrite: 0,
3493
+ },
3494
+ contextWindow: 256000,
3495
+ maxTokens: 8192,
3496
+ },
3497
+ "gemma-4-31b": {
3498
+ id: "gemma-4-31b",
3499
+ name: "Gemma 4 31B",
3500
+ api: "google-generative-ai",
3501
+ provider: "google",
3502
+ baseUrl: "https://generativelanguage.googleapis.com/v1beta",
3503
+ reasoning: true,
3504
+ input: ["text", "image"],
3505
+ cost: {
3506
+ input: 0,
3507
+ output: 0,
3508
+ cacheRead: 0,
3509
+ cacheWrite: 0,
3510
+ },
3511
+ contextWindow: 256000,
3512
+ maxTokens: 8192,
3513
+ },
3446
3514
  },
3447
3515
  "google-antigravity": {
3448
3516
  "claude-opus-4-5-thinking": {
@@ -3944,6 +4012,40 @@ export const MODELS = {
3944
4012
  contextWindow: 8192,
3945
4013
  maxTokens: 8192,
3946
4014
  },
4015
+ "groq/compound": {
4016
+ id: "groq/compound",
4017
+ name: "Compound",
4018
+ api: "openai-completions",
4019
+ provider: "groq",
4020
+ baseUrl: "https://api.groq.com/openai/v1",
4021
+ reasoning: true,
4022
+ input: ["text"],
4023
+ cost: {
4024
+ input: 0,
4025
+ output: 0,
4026
+ cacheRead: 0,
4027
+ cacheWrite: 0,
4028
+ },
4029
+ contextWindow: 131072,
4030
+ maxTokens: 8192,
4031
+ },
4032
+ "groq/compound-mini": {
4033
+ id: "groq/compound-mini",
4034
+ name: "Compound Mini",
4035
+ api: "openai-completions",
4036
+ provider: "groq",
4037
+ baseUrl: "https://api.groq.com/openai/v1",
4038
+ reasoning: true,
4039
+ input: ["text"],
4040
+ cost: {
4041
+ input: 0,
4042
+ output: 0,
4043
+ cacheRead: 0,
4044
+ cacheWrite: 0,
4045
+ },
4046
+ contextWindow: 131072,
4047
+ maxTokens: 8192,
4048
+ },
3947
4049
  "llama-3.1-8b-instant": {
3948
4050
  id: "llama-3.1-8b-instant",
3949
4051
  name: "Llama 3.1 8B Instant",
@@ -4131,6 +4233,23 @@ export const MODELS = {
4131
4233
  contextWindow: 131072,
4132
4234
  maxTokens: 65536,
4133
4235
  },
4236
+ "openai/gpt-oss-safeguard-20b": {
4237
+ id: "openai/gpt-oss-safeguard-20b",
4238
+ name: "Safety GPT OSS 20B",
4239
+ api: "openai-completions",
4240
+ provider: "groq",
4241
+ baseUrl: "https://api.groq.com/openai/v1",
4242
+ reasoning: true,
4243
+ input: ["text"],
4244
+ cost: {
4245
+ input: 0.075,
4246
+ output: 0.3,
4247
+ cacheRead: 0.037,
4248
+ cacheWrite: 0,
4249
+ },
4250
+ contextWindow: 131072,
4251
+ maxTokens: 65536,
4252
+ },
4134
4253
  "qwen-qwq-32b": {
4135
4254
  id: "qwen-qwq-32b",
4136
4255
  name: "Qwen QwQ 32B",
@@ -4163,7 +4282,7 @@ export const MODELS = {
4163
4282
  cacheWrite: 0,
4164
4283
  },
4165
4284
  contextWindow: 131072,
4166
- maxTokens: 16384,
4285
+ maxTokens: 40960,
4167
4286
  },
4168
4287
  },
4169
4288
  "huggingface": {
@@ -5094,22 +5213,39 @@ export const MODELS = {
5094
5213
  contextWindow: 128000,
5095
5214
  maxTokens: 16384,
5096
5215
  },
5216
+ "mistral-small-2603": {
5217
+ id: "mistral-small-2603",
5218
+ name: "Mistral Small 4",
5219
+ api: "mistral-conversations",
5220
+ provider: "mistral",
5221
+ baseUrl: "https://api.mistral.ai",
5222
+ reasoning: true,
5223
+ input: ["text", "image"],
5224
+ cost: {
5225
+ input: 0.15,
5226
+ output: 0.6,
5227
+ cacheRead: 0,
5228
+ cacheWrite: 0,
5229
+ },
5230
+ contextWindow: 256000,
5231
+ maxTokens: 256000,
5232
+ },
5097
5233
  "mistral-small-latest": {
5098
5234
  id: "mistral-small-latest",
5099
5235
  name: "Mistral Small (latest)",
5100
5236
  api: "mistral-conversations",
5101
5237
  provider: "mistral",
5102
5238
  baseUrl: "https://api.mistral.ai",
5103
- reasoning: false,
5239
+ reasoning: true,
5104
5240
  input: ["text", "image"],
5105
5241
  cost: {
5106
- input: 0.1,
5107
- output: 0.3,
5242
+ input: 0.15,
5243
+ output: 0.6,
5108
5244
  cacheRead: 0,
5109
5245
  cacheWrite: 0,
5110
5246
  },
5111
- contextWindow: 128000,
5112
- maxTokens: 16384,
5247
+ contextWindow: 256000,
5248
+ maxTokens: 256000,
5113
5249
  },
5114
5250
  "open-mistral-7b": {
5115
5251
  id: "open-mistral-7b",
@@ -5640,6 +5776,23 @@ export const MODELS = {
5640
5776
  contextWindow: 400000,
5641
5777
  maxTokens: 128000,
5642
5778
  },
5779
+ "gpt-5.3-chat-latest": {
5780
+ id: "gpt-5.3-chat-latest",
5781
+ name: "GPT-5.3 Chat (latest)",
5782
+ api: "openai-responses",
5783
+ provider: "openai",
5784
+ baseUrl: "https://api.openai.com/v1",
5785
+ reasoning: false,
5786
+ input: ["text", "image"],
5787
+ cost: {
5788
+ input: 1.75,
5789
+ output: 14,
5790
+ cacheRead: 0.175,
5791
+ cacheWrite: 0,
5792
+ },
5793
+ contextWindow: 128000,
5794
+ maxTokens: 16384,
5795
+ },
5643
5796
  "gpt-5.3-codex": {
5644
5797
  id: "gpt-5.3-codex",
5645
5798
  name: "GPT-5.3 Codex",
@@ -6494,40 +6647,6 @@ export const MODELS = {
6494
6647
  contextWindow: 262144,
6495
6648
  maxTokens: 65536,
6496
6649
  },
6497
- "mimo-v2-omni-free": {
6498
- id: "mimo-v2-omni-free",
6499
- name: "MiMo V2 Omni Free",
6500
- api: "openai-completions",
6501
- provider: "opencode",
6502
- baseUrl: "https://opencode.ai/zen/v1",
6503
- reasoning: true,
6504
- input: ["text", "image"],
6505
- cost: {
6506
- input: 0,
6507
- output: 0,
6508
- cacheRead: 0,
6509
- cacheWrite: 0,
6510
- },
6511
- contextWindow: 262144,
6512
- maxTokens: 64000,
6513
- },
6514
- "mimo-v2-pro-free": {
6515
- id: "mimo-v2-pro-free",
6516
- name: "MiMo V2 Pro Free",
6517
- api: "openai-completions",
6518
- provider: "opencode",
6519
- baseUrl: "https://opencode.ai/zen/v1",
6520
- reasoning: true,
6521
- input: ["text"],
6522
- cost: {
6523
- input: 0,
6524
- output: 0,
6525
- cacheRead: 0,
6526
- cacheWrite: 0,
6527
- },
6528
- contextWindow: 1048576,
6529
- maxTokens: 64000,
6530
- },
6531
6650
  "minimax-m2.5": {
6532
6651
  id: "minimax-m2.5",
6533
6652
  name: "MiniMax M2.5",
@@ -6576,9 +6695,26 @@ export const MODELS = {
6576
6695
  cacheRead: 0,
6577
6696
  cacheWrite: 0,
6578
6697
  },
6579
- contextWindow: 1000000,
6698
+ contextWindow: 204800,
6580
6699
  maxTokens: 128000,
6581
6700
  },
6701
+ "qwen3.6-plus-free": {
6702
+ id: "qwen3.6-plus-free",
6703
+ name: "Qwen3.6 Plus Free",
6704
+ api: "openai-completions",
6705
+ provider: "opencode",
6706
+ baseUrl: "https://opencode.ai/zen/v1",
6707
+ reasoning: true,
6708
+ input: ["text"],
6709
+ cost: {
6710
+ input: 0,
6711
+ output: 0,
6712
+ cacheRead: 0,
6713
+ cacheWrite: 0,
6714
+ },
6715
+ contextWindow: 1048576,
6716
+ maxTokens: 64000,
6717
+ },
6582
6718
  },
6583
6719
  "opencode-go": {
6584
6720
  "glm-5": {
@@ -6615,12 +6751,46 @@ export const MODELS = {
6615
6751
  contextWindow: 262144,
6616
6752
  maxTokens: 65536,
6617
6753
  },
6754
+ "mimo-v2-omni": {
6755
+ id: "mimo-v2-omni",
6756
+ name: "MiMo V2 Omni",
6757
+ api: "openai-completions",
6758
+ provider: "opencode-go",
6759
+ baseUrl: "https://opencode.ai/zen/go/v1",
6760
+ reasoning: true,
6761
+ input: ["text", "image"],
6762
+ cost: {
6763
+ input: 0.4,
6764
+ output: 2,
6765
+ cacheRead: 0.08,
6766
+ cacheWrite: 0,
6767
+ },
6768
+ contextWindow: 262144,
6769
+ maxTokens: 64000,
6770
+ },
6771
+ "mimo-v2-pro": {
6772
+ id: "mimo-v2-pro",
6773
+ name: "MiMo V2 Pro",
6774
+ api: "openai-completions",
6775
+ provider: "opencode-go",
6776
+ baseUrl: "https://opencode.ai/zen/go/v1",
6777
+ reasoning: true,
6778
+ input: ["text"],
6779
+ cost: {
6780
+ input: 1,
6781
+ output: 3,
6782
+ cacheRead: 0.2,
6783
+ cacheWrite: 0,
6784
+ },
6785
+ contextWindow: 1048576,
6786
+ maxTokens: 64000,
6787
+ },
6618
6788
  "minimax-m2.5": {
6619
6789
  id: "minimax-m2.5",
6620
6790
  name: "MiniMax M2.5",
6621
- api: "anthropic-messages",
6791
+ api: "openai-completions",
6622
6792
  provider: "opencode-go",
6623
- baseUrl: "https://opencode.ai/zen/go",
6793
+ baseUrl: "https://opencode.ai/zen/go/v1",
6624
6794
  reasoning: true,
6625
6795
  input: ["text"],
6626
6796
  cost: {
@@ -6821,30 +6991,13 @@ export const MODELS = {
6821
6991
  contextWindow: 200000,
6822
6992
  maxTokens: 8192,
6823
6993
  },
6824
- "anthropic/claude-3.5-sonnet": {
6825
- id: "anthropic/claude-3.5-sonnet",
6826
- name: "Anthropic: Claude 3.5 Sonnet",
6994
+ "anthropic/claude-3.7-sonnet": {
6995
+ id: "anthropic/claude-3.7-sonnet",
6996
+ name: "Anthropic: Claude 3.7 Sonnet",
6827
6997
  api: "openai-completions",
6828
6998
  provider: "openrouter",
6829
6999
  baseUrl: "https://openrouter.ai/api/v1",
6830
- reasoning: false,
6831
- input: ["text", "image"],
6832
- cost: {
6833
- input: 6,
6834
- output: 30,
6835
- cacheRead: 0.6,
6836
- cacheWrite: 7.5,
6837
- },
6838
- contextWindow: 200000,
6839
- maxTokens: 8192,
6840
- },
6841
- "anthropic/claude-3.7-sonnet": {
6842
- id: "anthropic/claude-3.7-sonnet",
6843
- name: "Anthropic: Claude 3.7 Sonnet",
6844
- api: "openai-completions",
6845
- provider: "openrouter",
6846
- baseUrl: "https://openrouter.ai/api/v1",
6847
- reasoning: true,
7000
+ reasoning: true,
6848
7001
  input: ["text", "image"],
6849
7002
  cost: {
6850
7003
  input: 3,
@@ -7025,6 +7178,23 @@ export const MODELS = {
7025
7178
  contextWindow: 131000,
7026
7179
  maxTokens: 4096,
7027
7180
  },
7181
+ "arcee-ai/trinity-large-thinking": {
7182
+ id: "arcee-ai/trinity-large-thinking",
7183
+ name: "Arcee AI: Trinity Large Thinking",
7184
+ api: "openai-completions",
7185
+ provider: "openrouter",
7186
+ baseUrl: "https://openrouter.ai/api/v1",
7187
+ reasoning: true,
7188
+ input: ["text"],
7189
+ cost: {
7190
+ input: 0.22,
7191
+ output: 0.85,
7192
+ cacheRead: 0,
7193
+ cacheWrite: 0,
7194
+ },
7195
+ contextWindow: 262144,
7196
+ maxTokens: 262144,
7197
+ },
7028
7198
  "arcee-ai/trinity-mini": {
7029
7199
  id: "arcee-ai/trinity-mini",
7030
7200
  name: "Arcee AI: Trinity Mini",
@@ -7465,7 +7635,7 @@ export const MODELS = {
7465
7635
  cacheWrite: 0.08333333333333334,
7466
7636
  },
7467
7637
  contextWindow: 1048576,
7468
- maxTokens: 65536,
7638
+ maxTokens: 65535,
7469
7639
  },
7470
7640
  "google/gemini-2.5-pro": {
7471
7641
  id: "google/gemini-2.5-pro",
@@ -7535,23 +7705,6 @@ export const MODELS = {
7535
7705
  contextWindow: 1048576,
7536
7706
  maxTokens: 65536,
7537
7707
  },
7538
- "google/gemini-3-pro-preview": {
7539
- id: "google/gemini-3-pro-preview",
7540
- name: "Google: Gemini 3 Pro Preview",
7541
- api: "openai-completions",
7542
- provider: "openrouter",
7543
- baseUrl: "https://openrouter.ai/api/v1",
7544
- reasoning: true,
7545
- input: ["text", "image"],
7546
- cost: {
7547
- input: 2,
7548
- output: 12,
7549
- cacheRead: 0.19999999999999998,
7550
- cacheWrite: 0.375,
7551
- },
7552
- contextWindow: 1048576,
7553
- maxTokens: 65536,
7554
- },
7555
7708
  "google/gemini-3.1-flash-lite-preview": {
7556
7709
  id: "google/gemini-3.1-flash-lite-preview",
7557
7710
  name: "Google: Gemini 3.1 Flash Lite Preview",
@@ -7603,6 +7756,40 @@ export const MODELS = {
7603
7756
  contextWindow: 1048576,
7604
7757
  maxTokens: 65536,
7605
7758
  },
7759
+ "google/gemma-4-26b-a4b-it": {
7760
+ id: "google/gemma-4-26b-a4b-it",
7761
+ name: "Google: Gemma 4 26B A4B ",
7762
+ api: "openai-completions",
7763
+ provider: "openrouter",
7764
+ baseUrl: "https://openrouter.ai/api/v1",
7765
+ reasoning: true,
7766
+ input: ["text", "image"],
7767
+ cost: {
7768
+ input: 0.13,
7769
+ output: 0.39999999999999997,
7770
+ cacheRead: 0,
7771
+ cacheWrite: 0,
7772
+ },
7773
+ contextWindow: 262144,
7774
+ maxTokens: 262144,
7775
+ },
7776
+ "google/gemma-4-31b-it": {
7777
+ id: "google/gemma-4-31b-it",
7778
+ name: "Google: Gemma 4 31B",
7779
+ api: "openai-completions",
7780
+ provider: "openrouter",
7781
+ baseUrl: "https://openrouter.ai/api/v1",
7782
+ reasoning: true,
7783
+ input: ["text", "image"],
7784
+ cost: {
7785
+ input: 0.14,
7786
+ output: 0.39999999999999997,
7787
+ cacheRead: 0,
7788
+ cacheWrite: 0,
7789
+ },
7790
+ contextWindow: 262144,
7791
+ maxTokens: 131072,
7792
+ },
7606
7793
  "inception/mercury": {
7607
7794
  id: "inception/mercury",
7608
7795
  name: "Inception: Mercury",
@@ -7654,22 +7841,22 @@ export const MODELS = {
7654
7841
  contextWindow: 128000,
7655
7842
  maxTokens: 32000,
7656
7843
  },
7657
- "kwaipilot/kat-coder-pro": {
7658
- id: "kwaipilot/kat-coder-pro",
7659
- name: "Kwaipilot: KAT-Coder-Pro V1",
7844
+ "kwaipilot/kat-coder-pro-v2": {
7845
+ id: "kwaipilot/kat-coder-pro-v2",
7846
+ name: "Kwaipilot: KAT-Coder-Pro V2",
7660
7847
  api: "openai-completions",
7661
7848
  provider: "openrouter",
7662
7849
  baseUrl: "https://openrouter.ai/api/v1",
7663
7850
  reasoning: false,
7664
7851
  input: ["text"],
7665
7852
  cost: {
7666
- input: 0.207,
7667
- output: 0.828,
7668
- cacheRead: 0.0414,
7853
+ input: 0.3,
7854
+ output: 1.2,
7855
+ cacheRead: 0.06,
7669
7856
  cacheWrite: 0,
7670
7857
  },
7671
7858
  contextWindow: 256000,
7672
- maxTokens: 128000,
7859
+ maxTokens: 80000,
7673
7860
  },
7674
7861
  "meituan/longcat-flash-chat": {
7675
7862
  id: "meituan/longcat-flash-chat",
@@ -7867,9 +8054,9 @@ export const MODELS = {
7867
8054
  reasoning: true,
7868
8055
  input: ["text"],
7869
8056
  cost: {
7870
- input: 0.19999999999999998,
7871
- output: 1.17,
7872
- cacheRead: 0.09999999999999999,
8057
+ input: 0.118,
8058
+ output: 0.9900000000000001,
8059
+ cacheRead: 0.059,
7873
8060
  cacheWrite: 0,
7874
8061
  },
7875
8062
  contextWindow: 196608,
@@ -8164,23 +8351,6 @@ export const MODELS = {
8164
8351
  contextWindow: 32768,
8165
8352
  maxTokens: 4096,
8166
8353
  },
8167
- "mistralai/mistral-small-24b-instruct-2501": {
8168
- id: "mistralai/mistral-small-24b-instruct-2501",
8169
- name: "Mistral: Mistral Small 3",
8170
- api: "openai-completions",
8171
- provider: "openrouter",
8172
- baseUrl: "https://openrouter.ai/api/v1",
8173
- reasoning: false,
8174
- input: ["text"],
8175
- cost: {
8176
- input: 0.049999999999999996,
8177
- output: 0.08,
8178
- cacheRead: 0,
8179
- cacheWrite: 0,
8180
- },
8181
- contextWindow: 32768,
8182
- maxTokens: 16384,
8183
- },
8184
8354
  "mistralai/mistral-small-2603": {
8185
8355
  id: "mistralai/mistral-small-2603",
8186
8356
  name: "Mistral: Mistral Small 4",
@@ -8198,23 +8368,6 @@ export const MODELS = {
8198
8368
  contextWindow: 262144,
8199
8369
  maxTokens: 4096,
8200
8370
  },
8201
- "mistralai/mistral-small-3.1-24b-instruct:free": {
8202
- id: "mistralai/mistral-small-3.1-24b-instruct:free",
8203
- name: "Mistral: Mistral Small 3.1 24B (free)",
8204
- api: "openai-completions",
8205
- provider: "openrouter",
8206
- baseUrl: "https://openrouter.ai/api/v1",
8207
- reasoning: false,
8208
- input: ["text", "image"],
8209
- cost: {
8210
- input: 0,
8211
- output: 0,
8212
- cacheRead: 0,
8213
- cacheWrite: 0,
8214
- },
8215
- contextWindow: 128000,
8216
- maxTokens: 4096,
8217
- },
8218
8371
  "mistralai/mistral-small-3.2-24b-instruct": {
8219
8372
  id: "mistralai/mistral-small-3.2-24b-instruct",
8220
8373
  name: "Mistral: Mistral Small 3.2 24B",
@@ -9252,6 +9405,40 @@ export const MODELS = {
9252
9405
  contextWindow: 1050000,
9253
9406
  maxTokens: 128000,
9254
9407
  },
9408
+ "openai/gpt-audio": {
9409
+ id: "openai/gpt-audio",
9410
+ name: "OpenAI: GPT Audio",
9411
+ api: "openai-completions",
9412
+ provider: "openrouter",
9413
+ baseUrl: "https://openrouter.ai/api/v1",
9414
+ reasoning: false,
9415
+ input: ["text"],
9416
+ cost: {
9417
+ input: 2.5,
9418
+ output: 10,
9419
+ cacheRead: 0,
9420
+ cacheWrite: 0,
9421
+ },
9422
+ contextWindow: 128000,
9423
+ maxTokens: 16384,
9424
+ },
9425
+ "openai/gpt-audio-mini": {
9426
+ id: "openai/gpt-audio-mini",
9427
+ name: "OpenAI: GPT Audio Mini",
9428
+ api: "openai-completions",
9429
+ provider: "openrouter",
9430
+ baseUrl: "https://openrouter.ai/api/v1",
9431
+ reasoning: false,
9432
+ input: ["text"],
9433
+ cost: {
9434
+ input: 0.6,
9435
+ output: 2.4,
9436
+ cacheRead: 0,
9437
+ cacheWrite: 0,
9438
+ },
9439
+ contextWindow: 128000,
9440
+ maxTokens: 16384,
9441
+ },
9255
9442
  "openai/gpt-oss-120b": {
9256
9443
  id: "openai/gpt-oss-120b",
9257
9444
  name: "OpenAI: gpt-oss-120b",
@@ -9813,23 +10000,6 @@ export const MODELS = {
9813
10000
  contextWindow: 40960,
9814
10001
  maxTokens: 40960,
9815
10002
  },
9816
- "qwen/qwen3-4b:free": {
9817
- id: "qwen/qwen3-4b:free",
9818
- name: "Qwen: Qwen3 4B (free)",
9819
- api: "openai-completions",
9820
- provider: "openrouter",
9821
- baseUrl: "https://openrouter.ai/api/v1",
9822
- reasoning: true,
9823
- input: ["text"],
9824
- cost: {
9825
- input: 0,
9826
- output: 0,
9827
- cacheRead: 0,
9828
- cacheWrite: 0,
9829
- },
9830
- contextWindow: 40960,
9831
- maxTokens: 4096,
9832
- },
9833
10003
  "qwen/qwen3-8b": {
9834
10004
  id: "qwen/qwen3-8b",
9835
10005
  name: "Qwen: Qwen3 8B",
@@ -10236,7 +10406,7 @@ export const MODELS = {
10236
10406
  cacheWrite: 0,
10237
10407
  },
10238
10408
  contextWindow: 256000,
10239
- maxTokens: 65536,
10409
+ maxTokens: 32768,
10240
10410
  },
10241
10411
  "qwen/qwen3.5-flash-02-23": {
10242
10412
  id: "qwen/qwen3.5-flash-02-23",
@@ -10272,6 +10442,23 @@ export const MODELS = {
10272
10442
  contextWindow: 1000000,
10273
10443
  maxTokens: 65536,
10274
10444
  },
10445
+ "qwen/qwen3.6-plus:free": {
10446
+ id: "qwen/qwen3.6-plus:free",
10447
+ name: "Qwen: Qwen3.6 Plus (free)",
10448
+ api: "openai-completions",
10449
+ provider: "openrouter",
10450
+ baseUrl: "https://openrouter.ai/api/v1",
10451
+ reasoning: true,
10452
+ input: ["text", "image"],
10453
+ cost: {
10454
+ input: 0,
10455
+ output: 0,
10456
+ cacheRead: 0,
10457
+ cacheWrite: 0,
10458
+ },
10459
+ contextWindow: 1000000,
10460
+ maxTokens: 65536,
10461
+ },
10275
10462
  "qwen/qwq-32b": {
10276
10463
  id: "qwen/qwq-32b",
10277
10464
  name: "Qwen: QwQ 32B",
@@ -10289,6 +10476,23 @@ export const MODELS = {
10289
10476
  contextWindow: 131072,
10290
10477
  maxTokens: 131072,
10291
10478
  },
10479
+ "rekaai/reka-edge": {
10480
+ id: "rekaai/reka-edge",
10481
+ name: "Reka Edge",
10482
+ api: "openai-completions",
10483
+ provider: "openrouter",
10484
+ baseUrl: "https://openrouter.ai/api/v1",
10485
+ reasoning: false,
10486
+ input: ["text", "image"],
10487
+ cost: {
10488
+ input: 0.09999999999999999,
10489
+ output: 0.09999999999999999,
10490
+ cacheRead: 0,
10491
+ cacheWrite: 0,
10492
+ },
10493
+ contextWindow: 16384,
10494
+ maxTokens: 16384,
10495
+ },
10292
10496
  "relace/relace-search": {
10293
10497
  id: "relace/relace-search",
10294
10498
  name: "Relace: Relace Search",
@@ -10351,11 +10555,11 @@ export const MODELS = {
10351
10555
  cost: {
10352
10556
  input: 0.09999999999999999,
10353
10557
  output: 0.3,
10354
- cacheRead: 0.02,
10558
+ cacheRead: 0,
10355
10559
  cacheWrite: 0,
10356
10560
  },
10357
- contextWindow: 256000,
10358
- maxTokens: 256000,
10561
+ contextWindow: 262144,
10562
+ maxTokens: 65536,
10359
10563
  },
10360
10564
  "stepfun/step-3.5-flash:free": {
10361
10565
  id: "stepfun/step-3.5-flash:free",
@@ -10561,9 +10765,9 @@ export const MODELS = {
10561
10765
  contextWindow: 2000000,
10562
10766
  maxTokens: 30000,
10563
10767
  },
10564
- "x-ai/grok-4.20-beta": {
10565
- id: "x-ai/grok-4.20-beta",
10566
- name: "xAI: Grok 4.20 Beta",
10768
+ "x-ai/grok-4.20": {
10769
+ id: "x-ai/grok-4.20",
10770
+ name: "xAI: Grok 4.20",
10567
10771
  api: "openai-completions",
10568
10772
  provider: "openrouter",
10569
10773
  baseUrl: "https://openrouter.ai/api/v1",
@@ -10833,6 +11037,23 @@ export const MODELS = {
10833
11037
  contextWindow: 202752,
10834
11038
  maxTokens: 131072,
10835
11039
  },
11040
+ "z-ai/glm-5v-turbo": {
11041
+ id: "z-ai/glm-5v-turbo",
11042
+ name: "Z.ai: GLM 5V Turbo",
11043
+ api: "openai-completions",
11044
+ provider: "openrouter",
11045
+ baseUrl: "https://openrouter.ai/api/v1",
11046
+ reasoning: true,
11047
+ input: ["text", "image"],
11048
+ cost: {
11049
+ input: 1.2,
11050
+ output: 4,
11051
+ cacheRead: 0.24,
11052
+ cacheWrite: 0,
11053
+ },
11054
+ contextWindow: 202752,
11055
+ maxTokens: 131072,
11056
+ },
10836
11057
  },
10837
11058
  "vercel-ai-gateway": {
10838
11059
  "alibaba/qwen-3-14b": {
@@ -10854,19 +11075,19 @@ export const MODELS = {
10854
11075
  },
10855
11076
  "alibaba/qwen-3-235b": {
10856
11077
  id: "alibaba/qwen-3-235b",
10857
- name: "Qwen3-235B-A22B",
11078
+ name: "Qwen3 235B A22b Instruct 2507",
10858
11079
  api: "anthropic-messages",
10859
11080
  provider: "vercel-ai-gateway",
10860
11081
  baseUrl: "https://ai-gateway.vercel.sh",
10861
11082
  reasoning: false,
10862
11083
  input: ["text"],
10863
11084
  cost: {
10864
- input: 0.071,
10865
- output: 0.463,
10866
- cacheRead: 0,
11085
+ input: 0.22,
11086
+ output: 0.88,
11087
+ cacheRead: 0.11,
10867
11088
  cacheWrite: 0,
10868
11089
  },
10869
- contextWindow: 40960,
11090
+ contextWindow: 32768,
10870
11091
  maxTokens: 16384,
10871
11092
  },
10872
11093
  "alibaba/qwen-3-30b": {
@@ -10895,13 +11116,13 @@ export const MODELS = {
10895
11116
  reasoning: true,
10896
11117
  input: ["text"],
10897
11118
  cost: {
10898
- input: 0.29,
10899
- output: 0.59,
10900
- cacheRead: 0.145,
11119
+ input: 0.16,
11120
+ output: 0.64,
11121
+ cacheRead: 0,
10901
11122
  cacheWrite: 0,
10902
11123
  },
10903
- contextWindow: 131072,
10904
- maxTokens: 40960,
11124
+ contextWindow: 128000,
11125
+ maxTokens: 8192,
10905
11126
  },
10906
11127
  "alibaba/qwen3-235b-a22b-thinking": {
10907
11128
  id: "alibaba/qwen3-235b-a22b-thinking",
@@ -10929,13 +11150,13 @@ export const MODELS = {
10929
11150
  reasoning: false,
10930
11151
  input: ["text"],
10931
11152
  cost: {
10932
- input: 0.39999999999999997,
10933
- output: 1.5999999999999999,
10934
- cacheRead: 0.022,
11153
+ input: 1.5,
11154
+ output: 7.5,
11155
+ cacheRead: 0.3,
10935
11156
  cacheWrite: 0,
10936
11157
  },
10937
11158
  contextWindow: 262144,
10938
- maxTokens: 66536,
11159
+ maxTokens: 65536,
10939
11160
  },
10940
11161
  "alibaba/qwen3-coder-30b-a3b": {
10941
11162
  id: "alibaba/qwen3-coder-30b-a3b",
@@ -11048,13 +11269,13 @@ export const MODELS = {
11048
11269
  reasoning: true,
11049
11270
  input: ["text", "image"],
11050
11271
  cost: {
11051
- input: 0.22,
11052
- output: 0.88,
11272
+ input: 0.39999999999999997,
11273
+ output: 4,
11053
11274
  cacheRead: 0,
11054
11275
  cacheWrite: 0,
11055
11276
  },
11056
- contextWindow: 256000,
11057
- maxTokens: 256000,
11277
+ contextWindow: 131072,
11278
+ maxTokens: 32768,
11058
11279
  },
11059
11280
  "alibaba/qwen3.5-flash": {
11060
11281
  id: "alibaba/qwen3.5-flash",
@@ -11090,6 +11311,23 @@ export const MODELS = {
11090
11311
  contextWindow: 1000000,
11091
11312
  maxTokens: 64000,
11092
11313
  },
11314
+ "alibaba/qwen3.6-plus": {
11315
+ id: "alibaba/qwen3.6-plus",
11316
+ name: "Qwen 3.6 Plus",
11317
+ api: "anthropic-messages",
11318
+ provider: "vercel-ai-gateway",
11319
+ baseUrl: "https://ai-gateway.vercel.sh",
11320
+ reasoning: true,
11321
+ input: ["text", "image"],
11322
+ cost: {
11323
+ input: 0.5,
11324
+ output: 3,
11325
+ cacheRead: 0.09999999999999999,
11326
+ cacheWrite: 0.625,
11327
+ },
11328
+ contextWindow: 1000000,
11329
+ maxTokens: 64000,
11330
+ },
11093
11331
  "anthropic/claude-3-haiku": {
11094
11332
  id: "anthropic/claude-3-haiku",
11095
11333
  name: "Claude 3 Haiku",
@@ -11124,40 +11362,6 @@ export const MODELS = {
11124
11362
  contextWindow: 200000,
11125
11363
  maxTokens: 8192,
11126
11364
  },
11127
- "anthropic/claude-3.5-sonnet": {
11128
- id: "anthropic/claude-3.5-sonnet",
11129
- name: "Claude 3.5 Sonnet",
11130
- api: "anthropic-messages",
11131
- provider: "vercel-ai-gateway",
11132
- baseUrl: "https://ai-gateway.vercel.sh",
11133
- reasoning: false,
11134
- input: ["text", "image"],
11135
- cost: {
11136
- input: 3,
11137
- output: 15,
11138
- cacheRead: 0.3,
11139
- cacheWrite: 3.75,
11140
- },
11141
- contextWindow: 200000,
11142
- maxTokens: 8192,
11143
- },
11144
- "anthropic/claude-3.5-sonnet-20240620": {
11145
- id: "anthropic/claude-3.5-sonnet-20240620",
11146
- name: "Claude 3.5 Sonnet (2024-06-20)",
11147
- api: "anthropic-messages",
11148
- provider: "vercel-ai-gateway",
11149
- baseUrl: "https://ai-gateway.vercel.sh",
11150
- reasoning: false,
11151
- input: ["text", "image"],
11152
- cost: {
11153
- input: 3,
11154
- output: 15,
11155
- cacheRead: 0.3,
11156
- cacheWrite: 3.75,
11157
- },
11158
- contextWindow: 200000,
11159
- maxTokens: 8192,
11160
- },
11161
11365
  "anthropic/claude-3.7-sonnet": {
11162
11366
  id: "anthropic/claude-3.7-sonnet",
11163
11367
  name: "Claude 3.7 Sonnet",
@@ -11328,6 +11532,23 @@ export const MODELS = {
11328
11532
  contextWindow: 131000,
11329
11533
  maxTokens: 131000,
11330
11534
  },
11535
+ "arcee-ai/trinity-large-thinking": {
11536
+ id: "arcee-ai/trinity-large-thinking",
11537
+ name: "Trinity Large Thinking",
11538
+ api: "anthropic-messages",
11539
+ provider: "vercel-ai-gateway",
11540
+ baseUrl: "https://ai-gateway.vercel.sh",
11541
+ reasoning: true,
11542
+ input: ["text"],
11543
+ cost: {
11544
+ input: 0.25,
11545
+ output: 0.8999999999999999,
11546
+ cacheRead: 0,
11547
+ cacheWrite: 0,
11548
+ },
11549
+ contextWindow: 262100,
11550
+ maxTokens: 80000,
11551
+ },
11331
11552
  "bytedance/seed-1.6": {
11332
11553
  id: "bytedance/seed-1.6",
11333
11554
  name: "Seed 1.6",
@@ -11405,13 +11626,13 @@ export const MODELS = {
11405
11626
  reasoning: true,
11406
11627
  input: ["text"],
11407
11628
  cost: {
11408
- input: 0.5,
11409
- output: 1.5,
11410
- cacheRead: 0,
11629
+ input: 0.56,
11630
+ output: 1.68,
11631
+ cacheRead: 0.28,
11411
11632
  cacheWrite: 0,
11412
11633
  },
11413
11634
  contextWindow: 163840,
11414
- maxTokens: 16384,
11635
+ maxTokens: 8192,
11415
11636
  },
11416
11637
  "deepseek/deepseek-v3.1-terminus": {
11417
11638
  id: "deepseek/deepseek-v3.1-terminus",
@@ -11617,6 +11838,40 @@ export const MODELS = {
11617
11838
  contextWindow: 1000000,
11618
11839
  maxTokens: 64000,
11619
11840
  },
11841
+ "google/gemma-4-26b-a4b-it": {
11842
+ id: "google/gemma-4-26b-a4b-it",
11843
+ name: "Gemma 4 26B A4B IT",
11844
+ api: "anthropic-messages",
11845
+ provider: "vercel-ai-gateway",
11846
+ baseUrl: "https://ai-gateway.vercel.sh",
11847
+ reasoning: false,
11848
+ input: ["text", "image"],
11849
+ cost: {
11850
+ input: 0.13,
11851
+ output: 0.39999999999999997,
11852
+ cacheRead: 0,
11853
+ cacheWrite: 0,
11854
+ },
11855
+ contextWindow: 262144,
11856
+ maxTokens: 131072,
11857
+ },
11858
+ "google/gemma-4-31b-it": {
11859
+ id: "google/gemma-4-31b-it",
11860
+ name: "Gemma 4 31B IT",
11861
+ api: "anthropic-messages",
11862
+ provider: "vercel-ai-gateway",
11863
+ baseUrl: "https://ai-gateway.vercel.sh",
11864
+ reasoning: false,
11865
+ input: ["text", "image"],
11866
+ cost: {
11867
+ input: 0.14,
11868
+ output: 0.39999999999999997,
11869
+ cacheRead: 0,
11870
+ cacheWrite: 0,
11871
+ },
11872
+ contextWindow: 262144,
11873
+ maxTokens: 131072,
11874
+ },
11620
11875
  "inception/mercury-2": {
11621
11876
  id: "inception/mercury-2",
11622
11877
  name: "Mercury 2",
@@ -11651,34 +11906,34 @@ export const MODELS = {
11651
11906
  contextWindow: 32000,
11652
11907
  maxTokens: 16384,
11653
11908
  },
11654
- "meituan/longcat-flash-chat": {
11655
- id: "meituan/longcat-flash-chat",
11656
- name: "LongCat Flash Chat",
11909
+ "kwaipilot/kat-coder-pro-v2": {
11910
+ id: "kwaipilot/kat-coder-pro-v2",
11911
+ name: "Kat Coder Pro V2",
11657
11912
  api: "anthropic-messages",
11658
11913
  provider: "vercel-ai-gateway",
11659
11914
  baseUrl: "https://ai-gateway.vercel.sh",
11660
- reasoning: false,
11915
+ reasoning: true,
11661
11916
  input: ["text"],
11662
11917
  cost: {
11663
- input: 0,
11664
- output: 0,
11665
- cacheRead: 0,
11918
+ input: 0.3,
11919
+ output: 1.2,
11920
+ cacheRead: 0.06,
11666
11921
  cacheWrite: 0,
11667
11922
  },
11668
- contextWindow: 128000,
11669
- maxTokens: 100000,
11923
+ contextWindow: 256000,
11924
+ maxTokens: 256000,
11670
11925
  },
11671
- "meituan/longcat-flash-thinking": {
11672
- id: "meituan/longcat-flash-thinking",
11673
- name: "LongCat Flash Thinking",
11926
+ "meituan/longcat-flash-chat": {
11927
+ id: "meituan/longcat-flash-chat",
11928
+ name: "LongCat Flash Chat",
11674
11929
  api: "anthropic-messages",
11675
11930
  provider: "vercel-ai-gateway",
11676
11931
  baseUrl: "https://ai-gateway.vercel.sh",
11677
- reasoning: true,
11932
+ reasoning: false,
11678
11933
  input: ["text"],
11679
11934
  cost: {
11680
- input: 0.15,
11681
- output: 1.5,
11935
+ input: 0,
11936
+ output: 0,
11682
11937
  cacheRead: 0,
11683
11938
  cacheWrite: 0,
11684
11939
  },
@@ -11711,13 +11966,13 @@ export const MODELS = {
11711
11966
  reasoning: false,
11712
11967
  input: ["text"],
11713
11968
  cost: {
11714
- input: 0.09999999999999999,
11715
- output: 0.09999999999999999,
11716
- cacheRead: 0.09999999999999999,
11969
+ input: 0.22,
11970
+ output: 0.22,
11971
+ cacheRead: 0,
11717
11972
  cacheWrite: 0,
11718
11973
  },
11719
11974
  contextWindow: 128000,
11720
- maxTokens: 16384,
11975
+ maxTokens: 8192,
11721
11976
  },
11722
11977
  "meta/llama-3.2-11b": {
11723
11978
  id: "meta/llama-3.2-11b",
@@ -12095,20 +12350,20 @@ export const MODELS = {
12095
12350
  },
12096
12351
  "moonshotai/kimi-k2": {
12097
12352
  id: "moonshotai/kimi-k2",
12098
- name: "Kimi K2",
12353
+ name: "Kimi K2 Instruct",
12099
12354
  api: "anthropic-messages",
12100
12355
  provider: "vercel-ai-gateway",
12101
12356
  baseUrl: "https://ai-gateway.vercel.sh",
12102
12357
  reasoning: false,
12103
12358
  input: ["text"],
12104
12359
  cost: {
12105
- input: 0.6,
12106
- output: 2.5,
12107
- cacheRead: 0.15,
12360
+ input: 0.5700000000000001,
12361
+ output: 2.3,
12362
+ cacheRead: 0,
12108
12363
  cacheWrite: 0,
12109
12364
  },
12110
12365
  contextWindow: 131072,
12111
- maxTokens: 16384,
12366
+ maxTokens: 131072,
12112
12367
  },
12113
12368
  "moonshotai/kimi-k2-0905": {
12114
12369
  id: "moonshotai/kimi-k2-0905",
@@ -12121,11 +12376,11 @@ export const MODELS = {
12121
12376
  cost: {
12122
12377
  input: 0.6,
12123
12378
  output: 2.5,
12124
- cacheRead: 0.15,
12379
+ cacheRead: 0.3,
12125
12380
  cacheWrite: 0,
12126
12381
  },
12127
12382
  contextWindow: 256000,
12128
- maxTokens: 16384,
12383
+ maxTokens: 128000,
12129
12384
  },
12130
12385
  "moonshotai/kimi-k2-thinking": {
12131
12386
  id: "moonshotai/kimi-k2-thinking",
@@ -12688,6 +12943,23 @@ export const MODELS = {
12688
12943
  contextWindow: 1050000,
12689
12944
  maxTokens: 128000,
12690
12945
  },
12946
+ "openai/gpt-oss-120b": {
12947
+ id: "openai/gpt-oss-120b",
12948
+ name: "gpt-oss-120b",
12949
+ api: "anthropic-messages",
12950
+ provider: "vercel-ai-gateway",
12951
+ baseUrl: "https://ai-gateway.vercel.sh",
12952
+ reasoning: true,
12953
+ input: ["text"],
12954
+ cost: {
12955
+ input: 0.15,
12956
+ output: 0.6,
12957
+ cacheRead: 0,
12958
+ cacheWrite: 0,
12959
+ },
12960
+ contextWindow: 128000,
12961
+ maxTokens: 8192,
12962
+ },
12691
12963
  "openai/gpt-oss-20b": {
12692
12964
  id: "openai/gpt-oss-20b",
12693
12965
  name: "gpt-oss-20b",
@@ -12697,12 +12969,12 @@ export const MODELS = {
12697
12969
  reasoning: true,
12698
12970
  input: ["text"],
12699
12971
  cost: {
12700
- input: 0.07,
12701
- output: 0.3,
12972
+ input: 0.049999999999999996,
12973
+ output: 0.19999999999999998,
12702
12974
  cacheRead: 0,
12703
12975
  cacheWrite: 0,
12704
12976
  },
12705
- contextWindow: 128000,
12977
+ contextWindow: 131072,
12706
12978
  maxTokens: 8192,
12707
12979
  },
12708
12980
  "openai/gpt-oss-safeguard-20b": {
@@ -13173,9 +13445,9 @@ export const MODELS = {
13173
13445
  reasoning: true,
13174
13446
  input: ["text"],
13175
13447
  cost: {
13176
- input: 0.09999999999999999,
13177
- output: 0.3,
13178
- cacheRead: 0.02,
13448
+ input: 0.09,
13449
+ output: 0.29,
13450
+ cacheRead: 0.045,
13179
13451
  cacheWrite: 0,
13180
13452
  },
13181
13453
  contextWindow: 262144,
@@ -13311,7 +13583,7 @@ export const MODELS = {
13311
13583
  cost: {
13312
13584
  input: 0.6,
13313
13585
  output: 2.2,
13314
- cacheRead: 0,
13586
+ cacheRead: 0.11,
13315
13587
  cacheWrite: 0,
13316
13588
  },
13317
13589
  contextWindow: 200000,
@@ -13385,6 +13657,23 @@ export const MODELS = {
13385
13657
  contextWindow: 202800,
13386
13658
  maxTokens: 131100,
13387
13659
  },
13660
+ "zai/glm-5v-turbo": {
13661
+ id: "zai/glm-5v-turbo",
13662
+ name: "GLM 5V Turbo",
13663
+ api: "anthropic-messages",
13664
+ provider: "vercel-ai-gateway",
13665
+ baseUrl: "https://ai-gateway.vercel.sh",
13666
+ reasoning: true,
13667
+ input: ["text", "image"],
13668
+ cost: {
13669
+ input: 1.2,
13670
+ output: 4,
13671
+ cacheRead: 0.24,
13672
+ cacheWrite: 0,
13673
+ },
13674
+ contextWindow: 200000,
13675
+ maxTokens: 128000,
13676
+ },
13388
13677
  },
13389
13678
  "xai": {
13390
13679
  "grok-2": {
@@ -13875,7 +14164,7 @@ export const MODELS = {
13875
14164
  api: "openai-completions",
13876
14165
  provider: "zai",
13877
14166
  baseUrl: "https://api.z.ai/api/coding/paas/v4",
13878
- compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
14167
+ compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
13879
14168
  reasoning: true,
13880
14169
  input: ["text"],
13881
14170
  cost: {
@@ -13893,7 +14182,7 @@ export const MODELS = {
13893
14182
  api: "openai-completions",
13894
14183
  provider: "zai",
13895
14184
  baseUrl: "https://api.z.ai/api/coding/paas/v4",
13896
- compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
14185
+ compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
13897
14186
  reasoning: true,
13898
14187
  input: ["text", "image"],
13899
14188
  cost: {
@@ -13911,7 +14200,7 @@ export const MODELS = {
13911
14200
  api: "openai-completions",
13912
14201
  provider: "zai",
13913
14202
  baseUrl: "https://api.z.ai/api/coding/paas/v4",
13914
- compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
14203
+ compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
13915
14204
  reasoning: true,
13916
14205
  input: ["text"],
13917
14206
  cost: {
@@ -13929,7 +14218,7 @@ export const MODELS = {
13929
14218
  api: "openai-completions",
13930
14219
  provider: "zai",
13931
14220
  baseUrl: "https://api.z.ai/api/coding/paas/v4",
13932
- compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
14221
+ compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
13933
14222
  reasoning: true,
13934
14223
  input: ["text"],
13935
14224
  cost: {
@@ -13941,13 +14230,31 @@ export const MODELS = {
13941
14230
  contextWindow: 200000,
13942
14231
  maxTokens: 131072,
13943
14232
  },
14233
+ "glm-4.7-flashx": {
14234
+ id: "glm-4.7-flashx",
14235
+ name: "GLM-4.7-FlashX",
14236
+ api: "openai-completions",
14237
+ provider: "zai",
14238
+ baseUrl: "https://api.z.ai/api/coding/paas/v4",
14239
+ compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
14240
+ reasoning: true,
14241
+ input: ["text"],
14242
+ cost: {
14243
+ input: 0.07,
14244
+ output: 0.4,
14245
+ cacheRead: 0.01,
14246
+ cacheWrite: 0,
14247
+ },
14248
+ contextWindow: 200000,
14249
+ maxTokens: 131072,
14250
+ },
13944
14251
  "glm-5": {
13945
14252
  id: "glm-5",
13946
14253
  name: "GLM-5",
13947
14254
  api: "openai-completions",
13948
14255
  provider: "zai",
13949
14256
  baseUrl: "https://api.z.ai/api/coding/paas/v4",
13950
- compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
14257
+ compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
13951
14258
  reasoning: true,
13952
14259
  input: ["text"],
13953
14260
  cost: {
@@ -13965,7 +14272,7 @@ export const MODELS = {
13965
14272
  api: "openai-completions",
13966
14273
  provider: "zai",
13967
14274
  baseUrl: "https://api.z.ai/api/coding/paas/v4",
13968
- compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
14275
+ compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
13969
14276
  reasoning: true,
13970
14277
  input: ["text"],
13971
14278
  cost: {
@@ -13977,6 +14284,24 @@ export const MODELS = {
13977
14284
  contextWindow: 200000,
13978
14285
  maxTokens: 131072,
13979
14286
  },
14287
+ "glm-5v-turbo": {
14288
+ id: "glm-5v-turbo",
14289
+ name: "glm-5v-turbo",
14290
+ api: "openai-completions",
14291
+ provider: "zai",
14292
+ baseUrl: "https://api.z.ai/api/coding/paas/v4",
14293
+ compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
14294
+ reasoning: true,
14295
+ input: ["text", "image"],
14296
+ cost: {
14297
+ input: 1.2,
14298
+ output: 4,
14299
+ cacheRead: 0.24,
14300
+ cacheWrite: 0,
14301
+ },
14302
+ contextWindow: 200000,
14303
+ maxTokens: 131072,
14304
+ },
13980
14305
  },
13981
14306
  };
13982
14307
  //# sourceMappingURL=models.generated.js.map