@mariozechner/pi-ai 0.64.0 → 0.65.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1551,23 +1551,6 @@ export const MODELS = {
1551
1551
  contextWindow: 200000,
1552
1552
  maxTokens: 64000,
1553
1553
  },
1554
- "claude-3-7-sonnet-latest": {
1555
- id: "claude-3-7-sonnet-latest",
1556
- name: "Claude Sonnet 3.7 (latest)",
1557
- api: "anthropic-messages",
1558
- provider: "anthropic",
1559
- baseUrl: "https://api.anthropic.com",
1560
- reasoning: true,
1561
- input: ["text", "image"],
1562
- cost: {
1563
- input: 3,
1564
- output: 15,
1565
- cacheRead: 0.3,
1566
- cacheWrite: 3.75,
1567
- },
1568
- contextWindow: 200000,
1569
- maxTokens: 64000,
1570
- },
1571
1554
  "claude-3-haiku-20240307": {
1572
1555
  id: "claude-3-haiku-20240307",
1573
1556
  name: "Claude Haiku 3",
@@ -2301,6 +2284,23 @@ export const MODELS = {
2301
2284
  contextWindow: 400000,
2302
2285
  maxTokens: 128000,
2303
2286
  },
2287
+ "gpt-5.3-chat-latest": {
2288
+ id: "gpt-5.3-chat-latest",
2289
+ name: "GPT-5.3 Chat (latest)",
2290
+ api: "azure-openai-responses",
2291
+ provider: "azure-openai-responses",
2292
+ baseUrl: "",
2293
+ reasoning: false,
2294
+ input: ["text", "image"],
2295
+ cost: {
2296
+ input: 1.75,
2297
+ output: 14,
2298
+ cacheRead: 0.175,
2299
+ cacheWrite: 0,
2300
+ },
2301
+ contextWindow: 128000,
2302
+ maxTokens: 16384,
2303
+ },
2304
2304
  "gpt-5.3-codex": {
2305
2305
  id: "gpt-5.3-codex",
2306
2306
  name: "GPT-5.3 Codex",
@@ -5026,22 +5026,39 @@ export const MODELS = {
5026
5026
  contextWindow: 128000,
5027
5027
  maxTokens: 16384,
5028
5028
  },
5029
+ "mistral-small-2603": {
5030
+ id: "mistral-small-2603",
5031
+ name: "Mistral Small 4",
5032
+ api: "mistral-conversations",
5033
+ provider: "mistral",
5034
+ baseUrl: "https://api.mistral.ai",
5035
+ reasoning: true,
5036
+ input: ["text", "image"],
5037
+ cost: {
5038
+ input: 0.15,
5039
+ output: 0.6,
5040
+ cacheRead: 0,
5041
+ cacheWrite: 0,
5042
+ },
5043
+ contextWindow: 256000,
5044
+ maxTokens: 256000,
5045
+ },
5029
5046
  "mistral-small-latest": {
5030
5047
  id: "mistral-small-latest",
5031
5048
  name: "Mistral Small (latest)",
5032
5049
  api: "mistral-conversations",
5033
5050
  provider: "mistral",
5034
5051
  baseUrl: "https://api.mistral.ai",
5035
- reasoning: false,
5052
+ reasoning: true,
5036
5053
  input: ["text", "image"],
5037
5054
  cost: {
5038
- input: 0.1,
5039
- output: 0.3,
5055
+ input: 0.15,
5056
+ output: 0.6,
5040
5057
  cacheRead: 0,
5041
5058
  cacheWrite: 0,
5042
5059
  },
5043
- contextWindow: 128000,
5044
- maxTokens: 16384,
5060
+ contextWindow: 256000,
5061
+ maxTokens: 256000,
5045
5062
  },
5046
5063
  "open-mistral-7b": {
5047
5064
  id: "open-mistral-7b",
@@ -5572,6 +5589,23 @@ export const MODELS = {
5572
5589
  contextWindow: 400000,
5573
5590
  maxTokens: 128000,
5574
5591
  },
5592
+ "gpt-5.3-chat-latest": {
5593
+ id: "gpt-5.3-chat-latest",
5594
+ name: "GPT-5.3 Chat (latest)",
5595
+ api: "openai-responses",
5596
+ provider: "openai",
5597
+ baseUrl: "https://api.openai.com/v1",
5598
+ reasoning: false,
5599
+ input: ["text", "image"],
5600
+ cost: {
5601
+ input: 1.75,
5602
+ output: 14,
5603
+ cacheRead: 0.175,
5604
+ cacheWrite: 0,
5605
+ },
5606
+ contextWindow: 128000,
5607
+ maxTokens: 16384,
5608
+ },
5575
5609
  "gpt-5.3-codex": {
5576
5610
  id: "gpt-5.3-codex",
5577
5611
  name: "GPT-5.3 Codex",
@@ -6426,40 +6460,6 @@ export const MODELS = {
6426
6460
  contextWindow: 262144,
6427
6461
  maxTokens: 65536,
6428
6462
  },
6429
- "mimo-v2-omni-free": {
6430
- id: "mimo-v2-omni-free",
6431
- name: "MiMo V2 Omni Free",
6432
- api: "openai-completions",
6433
- provider: "opencode",
6434
- baseUrl: "https://opencode.ai/zen/v1",
6435
- reasoning: true,
6436
- input: ["text", "image"],
6437
- cost: {
6438
- input: 0,
6439
- output: 0,
6440
- cacheRead: 0,
6441
- cacheWrite: 0,
6442
- },
6443
- contextWindow: 262144,
6444
- maxTokens: 64000,
6445
- },
6446
- "mimo-v2-pro-free": {
6447
- id: "mimo-v2-pro-free",
6448
- name: "MiMo V2 Pro Free",
6449
- api: "openai-completions",
6450
- provider: "opencode",
6451
- baseUrl: "https://opencode.ai/zen/v1",
6452
- reasoning: true,
6453
- input: ["text"],
6454
- cost: {
6455
- input: 0,
6456
- output: 0,
6457
- cacheRead: 0,
6458
- cacheWrite: 0,
6459
- },
6460
- contextWindow: 1048576,
6461
- maxTokens: 64000,
6462
- },
6463
6463
  "minimax-m2.5": {
6464
6464
  id: "minimax-m2.5",
6465
6465
  name: "MiniMax M2.5",
@@ -6508,9 +6508,26 @@ export const MODELS = {
6508
6508
  cacheRead: 0,
6509
6509
  cacheWrite: 0,
6510
6510
  },
6511
- contextWindow: 1000000,
6511
+ contextWindow: 204800,
6512
6512
  maxTokens: 128000,
6513
6513
  },
6514
+ "qwen3.6-plus-free": {
6515
+ id: "qwen3.6-plus-free",
6516
+ name: "Qwen3.6 Plus Free",
6517
+ api: "openai-completions",
6518
+ provider: "opencode",
6519
+ baseUrl: "https://opencode.ai/zen/v1",
6520
+ reasoning: true,
6521
+ input: ["text"],
6522
+ cost: {
6523
+ input: 0,
6524
+ output: 0,
6525
+ cacheRead: 0,
6526
+ cacheWrite: 0,
6527
+ },
6528
+ contextWindow: 1048576,
6529
+ maxTokens: 64000,
6530
+ },
6514
6531
  },
6515
6532
  "opencode-go": {
6516
6533
  "glm-5": {
@@ -6547,12 +6564,46 @@ export const MODELS = {
6547
6564
  contextWindow: 262144,
6548
6565
  maxTokens: 65536,
6549
6566
  },
6567
+ "mimo-v2-omni": {
6568
+ id: "mimo-v2-omni",
6569
+ name: "MiMo V2 Omni",
6570
+ api: "openai-completions",
6571
+ provider: "opencode-go",
6572
+ baseUrl: "https://opencode.ai/zen/go/v1",
6573
+ reasoning: true,
6574
+ input: ["text", "image"],
6575
+ cost: {
6576
+ input: 0.4,
6577
+ output: 2,
6578
+ cacheRead: 0.08,
6579
+ cacheWrite: 0,
6580
+ },
6581
+ contextWindow: 262144,
6582
+ maxTokens: 64000,
6583
+ },
6584
+ "mimo-v2-pro": {
6585
+ id: "mimo-v2-pro",
6586
+ name: "MiMo V2 Pro",
6587
+ api: "openai-completions",
6588
+ provider: "opencode-go",
6589
+ baseUrl: "https://opencode.ai/zen/go/v1",
6590
+ reasoning: true,
6591
+ input: ["text"],
6592
+ cost: {
6593
+ input: 1,
6594
+ output: 3,
6595
+ cacheRead: 0.2,
6596
+ cacheWrite: 0,
6597
+ },
6598
+ contextWindow: 1048576,
6599
+ maxTokens: 64000,
6600
+ },
6550
6601
  "minimax-m2.5": {
6551
6602
  id: "minimax-m2.5",
6552
6603
  name: "MiniMax M2.5",
6553
- api: "anthropic-messages",
6604
+ api: "openai-completions",
6554
6605
  provider: "opencode-go",
6555
- baseUrl: "https://opencode.ai/zen/go",
6606
+ baseUrl: "https://opencode.ai/zen/go/v1",
6556
6607
  reasoning: true,
6557
6608
  input: ["text"],
6558
6609
  cost: {
@@ -6957,6 +7008,23 @@ export const MODELS = {
6957
7008
  contextWindow: 131000,
6958
7009
  maxTokens: 4096,
6959
7010
  },
7011
+ "arcee-ai/trinity-large-thinking": {
7012
+ id: "arcee-ai/trinity-large-thinking",
7013
+ name: "Arcee AI: Trinity Large Thinking",
7014
+ api: "openai-completions",
7015
+ provider: "openrouter",
7016
+ baseUrl: "https://openrouter.ai/api/v1",
7017
+ reasoning: true,
7018
+ input: ["text"],
7019
+ cost: {
7020
+ input: 0.22,
7021
+ output: 0.85,
7022
+ cacheRead: 0,
7023
+ cacheWrite: 0,
7024
+ },
7025
+ contextWindow: 262144,
7026
+ maxTokens: 262144,
7027
+ },
6960
7028
  "arcee-ai/trinity-mini": {
6961
7029
  id: "arcee-ai/trinity-mini",
6962
7030
  name: "Arcee AI: Trinity Mini",
@@ -7397,7 +7465,7 @@ export const MODELS = {
7397
7465
  cacheWrite: 0.08333333333333334,
7398
7466
  },
7399
7467
  contextWindow: 1048576,
7400
- maxTokens: 65536,
7468
+ maxTokens: 65535,
7401
7469
  },
7402
7470
  "google/gemini-2.5-pro": {
7403
7471
  id: "google/gemini-2.5-pro",
@@ -7518,6 +7586,40 @@ export const MODELS = {
7518
7586
  contextWindow: 1048576,
7519
7587
  maxTokens: 65536,
7520
7588
  },
7589
+ "google/gemma-4-26b-a4b-it": {
7590
+ id: "google/gemma-4-26b-a4b-it",
7591
+ name: "Google: Gemma 4 26B A4B ",
7592
+ api: "openai-completions",
7593
+ provider: "openrouter",
7594
+ baseUrl: "https://openrouter.ai/api/v1",
7595
+ reasoning: true,
7596
+ input: ["text", "image"],
7597
+ cost: {
7598
+ input: 0.13,
7599
+ output: 0.39999999999999997,
7600
+ cacheRead: 0,
7601
+ cacheWrite: 0,
7602
+ },
7603
+ contextWindow: 262144,
7604
+ maxTokens: 262144,
7605
+ },
7606
+ "google/gemma-4-31b-it": {
7607
+ id: "google/gemma-4-31b-it",
7608
+ name: "Google: Gemma 4 31B",
7609
+ api: "openai-completions",
7610
+ provider: "openrouter",
7611
+ baseUrl: "https://openrouter.ai/api/v1",
7612
+ reasoning: true,
7613
+ input: ["text", "image"],
7614
+ cost: {
7615
+ input: 0.14,
7616
+ output: 0.39999999999999997,
7617
+ cacheRead: 0,
7618
+ cacheWrite: 0,
7619
+ },
7620
+ contextWindow: 262144,
7621
+ maxTokens: 131072,
7622
+ },
7521
7623
  "inception/mercury": {
7522
7624
  id: "inception/mercury",
7523
7625
  name: "Inception: Mercury",
@@ -7569,23 +7671,6 @@ export const MODELS = {
7569
7671
  contextWindow: 128000,
7570
7672
  maxTokens: 32000,
7571
7673
  },
7572
- "kwaipilot/kat-coder-pro": {
7573
- id: "kwaipilot/kat-coder-pro",
7574
- name: "Kwaipilot: KAT-Coder-Pro V1",
7575
- api: "openai-completions",
7576
- provider: "openrouter",
7577
- baseUrl: "https://openrouter.ai/api/v1",
7578
- reasoning: false,
7579
- input: ["text"],
7580
- cost: {
7581
- input: 0.207,
7582
- output: 0.828,
7583
- cacheRead: 0.0414,
7584
- cacheWrite: 0,
7585
- },
7586
- contextWindow: 256000,
7587
- maxTokens: 128000,
7588
- },
7589
7674
  "kwaipilot/kat-coder-pro-v2": {
7590
7675
  id: "kwaipilot/kat-coder-pro-v2",
7591
7676
  name: "Kwaipilot: KAT-Coder-Pro V2",
@@ -7799,9 +7884,9 @@ export const MODELS = {
7799
7884
  reasoning: true,
7800
7885
  input: ["text"],
7801
7886
  cost: {
7802
- input: 0.19,
7803
- output: 1.15,
7804
- cacheRead: 0.095,
7887
+ input: 0.118,
7888
+ output: 0.9900000000000001,
7889
+ cacheRead: 0.059,
7805
7890
  cacheWrite: 0,
7806
7891
  },
7807
7892
  contextWindow: 196608,
@@ -8096,23 +8181,6 @@ export const MODELS = {
8096
8181
  contextWindow: 32768,
8097
8182
  maxTokens: 4096,
8098
8183
  },
8099
- "mistralai/mistral-small-24b-instruct-2501": {
8100
- id: "mistralai/mistral-small-24b-instruct-2501",
8101
- name: "Mistral: Mistral Small 3",
8102
- api: "openai-completions",
8103
- provider: "openrouter",
8104
- baseUrl: "https://openrouter.ai/api/v1",
8105
- reasoning: false,
8106
- input: ["text"],
8107
- cost: {
8108
- input: 0.049999999999999996,
8109
- output: 0.08,
8110
- cacheRead: 0,
8111
- cacheWrite: 0,
8112
- },
8113
- contextWindow: 32768,
8114
- maxTokens: 16384,
8115
- },
8116
8184
  "mistralai/mistral-small-2603": {
8117
8185
  id: "mistralai/mistral-small-2603",
8118
8186
  name: "Mistral: Mistral Small 4",
@@ -9167,6 +9235,40 @@ export const MODELS = {
9167
9235
  contextWindow: 1050000,
9168
9236
  maxTokens: 128000,
9169
9237
  },
9238
+ "openai/gpt-audio": {
9239
+ id: "openai/gpt-audio",
9240
+ name: "OpenAI: GPT Audio",
9241
+ api: "openai-completions",
9242
+ provider: "openrouter",
9243
+ baseUrl: "https://openrouter.ai/api/v1",
9244
+ reasoning: false,
9245
+ input: ["text"],
9246
+ cost: {
9247
+ input: 2.5,
9248
+ output: 10,
9249
+ cacheRead: 0,
9250
+ cacheWrite: 0,
9251
+ },
9252
+ contextWindow: 128000,
9253
+ maxTokens: 16384,
9254
+ },
9255
+ "openai/gpt-audio-mini": {
9256
+ id: "openai/gpt-audio-mini",
9257
+ name: "OpenAI: GPT Audio Mini",
9258
+ api: "openai-completions",
9259
+ provider: "openrouter",
9260
+ baseUrl: "https://openrouter.ai/api/v1",
9261
+ reasoning: false,
9262
+ input: ["text"],
9263
+ cost: {
9264
+ input: 0.6,
9265
+ output: 2.4,
9266
+ cacheRead: 0,
9267
+ cacheWrite: 0,
9268
+ },
9269
+ contextWindow: 128000,
9270
+ maxTokens: 16384,
9271
+ },
9170
9272
  "openai/gpt-oss-120b": {
9171
9273
  id: "openai/gpt-oss-120b",
9172
9274
  name: "OpenAI: gpt-oss-120b",
@@ -10134,7 +10236,7 @@ export const MODELS = {
10134
10236
  cacheWrite: 0,
10135
10237
  },
10136
10238
  contextWindow: 256000,
10137
- maxTokens: 65536,
10239
+ maxTokens: 32768,
10138
10240
  },
10139
10241
  "qwen/qwen3.5-flash-02-23": {
10140
10242
  id: "qwen/qwen3.5-flash-02-23",
@@ -10170,6 +10272,23 @@ export const MODELS = {
10170
10272
  contextWindow: 1000000,
10171
10273
  maxTokens: 65536,
10172
10274
  },
10275
+ "qwen/qwen3.6-plus:free": {
10276
+ id: "qwen/qwen3.6-plus:free",
10277
+ name: "Qwen: Qwen3.6 Plus (free)",
10278
+ api: "openai-completions",
10279
+ provider: "openrouter",
10280
+ baseUrl: "https://openrouter.ai/api/v1",
10281
+ reasoning: true,
10282
+ input: ["text", "image"],
10283
+ cost: {
10284
+ input: 0,
10285
+ output: 0,
10286
+ cacheRead: 0,
10287
+ cacheWrite: 0,
10288
+ },
10289
+ contextWindow: 1000000,
10290
+ maxTokens: 65536,
10291
+ },
10173
10292
  "qwen/qwq-32b": {
10174
10293
  id: "qwen/qwq-32b",
10175
10294
  name: "Qwen: QwQ 32B",
@@ -10187,8 +10306,8 @@ export const MODELS = {
10187
10306
  contextWindow: 131072,
10188
10307
  maxTokens: 131072,
10189
10308
  },
10190
- "reka/reka-edge": {
10191
- id: "reka/reka-edge",
10309
+ "rekaai/reka-edge": {
10310
+ id: "rekaai/reka-edge",
10192
10311
  name: "Reka Edge",
10193
10312
  api: "openai-completions",
10194
10313
  provider: "openrouter",
@@ -10266,11 +10385,11 @@ export const MODELS = {
10266
10385
  cost: {
10267
10386
  input: 0.09999999999999999,
10268
10387
  output: 0.3,
10269
- cacheRead: 0.02,
10388
+ cacheRead: 0,
10270
10389
  cacheWrite: 0,
10271
10390
  },
10272
10391
  contextWindow: 262144,
10273
- maxTokens: 4096,
10392
+ maxTokens: 65536,
10274
10393
  },
10275
10394
  "stepfun/step-3.5-flash:free": {
10276
10395
  id: "stepfun/step-3.5-flash:free",
@@ -10476,9 +10595,9 @@ export const MODELS = {
10476
10595
  contextWindow: 2000000,
10477
10596
  maxTokens: 30000,
10478
10597
  },
10479
- "x-ai/grok-4.20-beta": {
10480
- id: "x-ai/grok-4.20-beta",
10481
- name: "xAI: Grok 4.20 Beta",
10598
+ "x-ai/grok-4.20": {
10599
+ id: "x-ai/grok-4.20",
10600
+ name: "xAI: Grok 4.20",
10482
10601
  api: "openai-completions",
10483
10602
  provider: "openrouter",
10484
10603
  baseUrl: "https://openrouter.ai/api/v1",
@@ -10748,6 +10867,23 @@ export const MODELS = {
10748
10867
  contextWindow: 202752,
10749
10868
  maxTokens: 131072,
10750
10869
  },
10870
+ "z-ai/glm-5v-turbo": {
10871
+ id: "z-ai/glm-5v-turbo",
10872
+ name: "Z.ai: GLM 5V Turbo",
10873
+ api: "openai-completions",
10874
+ provider: "openrouter",
10875
+ baseUrl: "https://openrouter.ai/api/v1",
10876
+ reasoning: true,
10877
+ input: ["text", "image"],
10878
+ cost: {
10879
+ input: 1.2,
10880
+ output: 4,
10881
+ cacheRead: 0.24,
10882
+ cacheWrite: 0,
10883
+ },
10884
+ contextWindow: 202752,
10885
+ maxTokens: 131072,
10886
+ },
10751
10887
  },
10752
10888
  "vercel-ai-gateway": {
10753
10889
  "alibaba/qwen-3-14b": {
@@ -10776,12 +10912,12 @@ export const MODELS = {
10776
10912
  reasoning: false,
10777
10913
  input: ["text"],
10778
10914
  cost: {
10779
- input: 0.071,
10780
- output: 0.463,
10781
- cacheRead: 0,
10915
+ input: 0.22,
10916
+ output: 0.88,
10917
+ cacheRead: 0.11,
10782
10918
  cacheWrite: 0,
10783
10919
  },
10784
- contextWindow: 40960,
10920
+ contextWindow: 32768,
10785
10921
  maxTokens: 16384,
10786
10922
  },
10787
10923
  "alibaba/qwen-3-30b": {
@@ -10810,13 +10946,13 @@ export const MODELS = {
10810
10946
  reasoning: true,
10811
10947
  input: ["text"],
10812
10948
  cost: {
10813
- input: 0.29,
10814
- output: 0.59,
10815
- cacheRead: 0.145,
10949
+ input: 0.16,
10950
+ output: 0.64,
10951
+ cacheRead: 0,
10816
10952
  cacheWrite: 0,
10817
10953
  },
10818
- contextWindow: 131072,
10819
- maxTokens: 40960,
10954
+ contextWindow: 128000,
10955
+ maxTokens: 8192,
10820
10956
  },
10821
10957
  "alibaba/qwen3-235b-a22b-thinking": {
10822
10958
  id: "alibaba/qwen3-235b-a22b-thinking",
@@ -10844,13 +10980,13 @@ export const MODELS = {
10844
10980
  reasoning: false,
10845
10981
  input: ["text"],
10846
10982
  cost: {
10847
- input: 0.39999999999999997,
10848
- output: 1.5999999999999999,
10849
- cacheRead: 0.022,
10983
+ input: 1.5,
10984
+ output: 7.5,
10985
+ cacheRead: 0.3,
10850
10986
  cacheWrite: 0,
10851
10987
  },
10852
10988
  contextWindow: 262144,
10853
- maxTokens: 66536,
10989
+ maxTokens: 65536,
10854
10990
  },
10855
10991
  "alibaba/qwen3-coder-30b-a3b": {
10856
10992
  id: "alibaba/qwen3-coder-30b-a3b",
@@ -11005,6 +11141,23 @@ export const MODELS = {
11005
11141
  contextWindow: 1000000,
11006
11142
  maxTokens: 64000,
11007
11143
  },
11144
+ "alibaba/qwen3.6-plus": {
11145
+ id: "alibaba/qwen3.6-plus",
11146
+ name: "Qwen 3.6 Plus",
11147
+ api: "anthropic-messages",
11148
+ provider: "vercel-ai-gateway",
11149
+ baseUrl: "https://ai-gateway.vercel.sh",
11150
+ reasoning: true,
11151
+ input: ["text", "image"],
11152
+ cost: {
11153
+ input: 0.5,
11154
+ output: 3,
11155
+ cacheRead: 0.09999999999999999,
11156
+ cacheWrite: 0.625,
11157
+ },
11158
+ contextWindow: 1000000,
11159
+ maxTokens: 64000,
11160
+ },
11008
11161
  "anthropic/claude-3-haiku": {
11009
11162
  id: "anthropic/claude-3-haiku",
11010
11163
  name: "Claude 3 Haiku",
@@ -11243,6 +11396,23 @@ export const MODELS = {
11243
11396
  contextWindow: 131000,
11244
11397
  maxTokens: 131000,
11245
11398
  },
11399
+ "arcee-ai/trinity-large-thinking": {
11400
+ id: "arcee-ai/trinity-large-thinking",
11401
+ name: "Trinity Large Thinking",
11402
+ api: "anthropic-messages",
11403
+ provider: "vercel-ai-gateway",
11404
+ baseUrl: "https://ai-gateway.vercel.sh",
11405
+ reasoning: true,
11406
+ input: ["text"],
11407
+ cost: {
11408
+ input: 0.25,
11409
+ output: 0.8999999999999999,
11410
+ cacheRead: 0,
11411
+ cacheWrite: 0,
11412
+ },
11413
+ contextWindow: 262100,
11414
+ maxTokens: 80000,
11415
+ },
11246
11416
  "bytedance/seed-1.6": {
11247
11417
  id: "bytedance/seed-1.6",
11248
11418
  name: "Seed 1.6",
@@ -11320,13 +11490,13 @@ export const MODELS = {
11320
11490
  reasoning: true,
11321
11491
  input: ["text"],
11322
11492
  cost: {
11323
- input: 0.5,
11324
- output: 1.5,
11325
- cacheRead: 0,
11493
+ input: 0.56,
11494
+ output: 1.68,
11495
+ cacheRead: 0.28,
11326
11496
  cacheWrite: 0,
11327
11497
  },
11328
11498
  contextWindow: 163840,
11329
- maxTokens: 16384,
11499
+ maxTokens: 8192,
11330
11500
  },
11331
11501
  "deepseek/deepseek-v3.1-terminus": {
11332
11502
  id: "deepseek/deepseek-v3.1-terminus",
@@ -11532,6 +11702,40 @@ export const MODELS = {
11532
11702
  contextWindow: 1000000,
11533
11703
  maxTokens: 64000,
11534
11704
  },
11705
+ "google/gemma-4-26b-a4b-it": {
11706
+ id: "google/gemma-4-26b-a4b-it",
11707
+ name: "Gemma 4 26B A4B IT",
11708
+ api: "anthropic-messages",
11709
+ provider: "vercel-ai-gateway",
11710
+ baseUrl: "https://ai-gateway.vercel.sh",
11711
+ reasoning: true,
11712
+ input: ["text", "image"],
11713
+ cost: {
11714
+ input: 0.13,
11715
+ output: 0.39999999999999997,
11716
+ cacheRead: 0,
11717
+ cacheWrite: 0,
11718
+ },
11719
+ contextWindow: 262144,
11720
+ maxTokens: 131072,
11721
+ },
11722
+ "google/gemma-4-31b-it": {
11723
+ id: "google/gemma-4-31b-it",
11724
+ name: "Gemma 4 31B IT",
11725
+ api: "anthropic-messages",
11726
+ provider: "vercel-ai-gateway",
11727
+ baseUrl: "https://ai-gateway.vercel.sh",
11728
+ reasoning: true,
11729
+ input: ["text", "image"],
11730
+ cost: {
11731
+ input: 0.14,
11732
+ output: 0.39999999999999997,
11733
+ cacheRead: 0,
11734
+ cacheWrite: 0,
11735
+ },
11736
+ contextWindow: 262144,
11737
+ maxTokens: 131072,
11738
+ },
11535
11739
  "inception/mercury-2": {
11536
11740
  id: "inception/mercury-2",
11537
11741
  name: "Mercury 2",
@@ -11598,7 +11802,7 @@ export const MODELS = {
11598
11802
  cacheWrite: 0,
11599
11803
  },
11600
11804
  contextWindow: 128000,
11601
- maxTokens: 100000,
11805
+ maxTokens: 8192,
11602
11806
  },
11603
11807
  "meituan/longcat-flash-thinking": {
11604
11808
  id: "meituan/longcat-flash-thinking",
@@ -11643,13 +11847,13 @@ export const MODELS = {
11643
11847
  reasoning: false,
11644
11848
  input: ["text"],
11645
11849
  cost: {
11646
- input: 0.09999999999999999,
11647
- output: 0.09999999999999999,
11648
- cacheRead: 0.09999999999999999,
11850
+ input: 0.22,
11851
+ output: 0.22,
11852
+ cacheRead: 0,
11649
11853
  cacheWrite: 0,
11650
11854
  },
11651
11855
  contextWindow: 128000,
11652
- maxTokens: 16384,
11856
+ maxTokens: 8192,
11653
11857
  },
11654
11858
  "meta/llama-3.2-11b": {
11655
11859
  id: "meta/llama-3.2-11b",
@@ -11711,12 +11915,12 @@ export const MODELS = {
11711
11915
  reasoning: false,
11712
11916
  input: ["text", "image"],
11713
11917
  cost: {
11714
- input: 0.24,
11715
- output: 0.9700000000000001,
11918
+ input: 0.35,
11919
+ output: 1.15,
11716
11920
  cacheRead: 0,
11717
11921
  cacheWrite: 0,
11718
11922
  },
11719
- contextWindow: 128000,
11923
+ contextWindow: 524288,
11720
11924
  maxTokens: 8192,
11721
11925
  },
11722
11926
  "meta/llama-4-scout": {
@@ -12036,7 +12240,7 @@ export const MODELS = {
12036
12240
  cost: {
12037
12241
  input: 0.6,
12038
12242
  output: 2.5,
12039
- cacheRead: 0.15,
12243
+ cacheRead: 0,
12040
12244
  cacheWrite: 0,
12041
12245
  },
12042
12246
  contextWindow: 131072,
@@ -12053,11 +12257,11 @@ export const MODELS = {
12053
12257
  cost: {
12054
12258
  input: 0.6,
12055
12259
  output: 2.5,
12056
- cacheRead: 0.15,
12260
+ cacheRead: 0.3,
12057
12261
  cacheWrite: 0,
12058
12262
  },
12059
12263
  contextWindow: 256000,
12060
- maxTokens: 16384,
12264
+ maxTokens: 128000,
12061
12265
  },
12062
12266
  "moonshotai/kimi-k2-thinking": {
12063
12267
  id: "moonshotai/kimi-k2-thinking",
@@ -12620,6 +12824,23 @@ export const MODELS = {
12620
12824
  contextWindow: 1050000,
12621
12825
  maxTokens: 128000,
12622
12826
  },
12827
+ "openai/gpt-oss-120b": {
12828
+ id: "openai/gpt-oss-120b",
12829
+ name: "gpt-oss-120b",
12830
+ api: "anthropic-messages",
12831
+ provider: "vercel-ai-gateway",
12832
+ baseUrl: "https://ai-gateway.vercel.sh",
12833
+ reasoning: true,
12834
+ input: ["text"],
12835
+ cost: {
12836
+ input: 0.15,
12837
+ output: 0.6,
12838
+ cacheRead: 0,
12839
+ cacheWrite: 0,
12840
+ },
12841
+ contextWindow: 128000,
12842
+ maxTokens: 8192,
12843
+ },
12623
12844
  "openai/gpt-oss-20b": {
12624
12845
  id: "openai/gpt-oss-20b",
12625
12846
  name: "gpt-oss-20b",
@@ -12629,12 +12850,12 @@ export const MODELS = {
12629
12850
  reasoning: true,
12630
12851
  input: ["text"],
12631
12852
  cost: {
12632
- input: 0.07,
12633
- output: 0.3,
12853
+ input: 0.049999999999999996,
12854
+ output: 0.19999999999999998,
12634
12855
  cacheRead: 0,
12635
12856
  cacheWrite: 0,
12636
12857
  },
12637
- contextWindow: 128000,
12858
+ contextWindow: 131072,
12638
12859
  maxTokens: 8192,
12639
12860
  },
12640
12861
  "openai/gpt-oss-safeguard-20b": {
@@ -13105,9 +13326,9 @@ export const MODELS = {
13105
13326
  reasoning: true,
13106
13327
  input: ["text"],
13107
13328
  cost: {
13108
- input: 0.09999999999999999,
13109
- output: 0.3,
13110
- cacheRead: 0.02,
13329
+ input: 0.09,
13330
+ output: 0.29,
13331
+ cacheRead: 0.045,
13111
13332
  cacheWrite: 0,
13112
13333
  },
13113
13334
  contextWindow: 262144,
@@ -13243,7 +13464,7 @@ export const MODELS = {
13243
13464
  cost: {
13244
13465
  input: 0.6,
13245
13466
  output: 2.2,
13246
- cacheRead: 0,
13467
+ cacheRead: 0.11,
13247
13468
  cacheWrite: 0,
13248
13469
  },
13249
13470
  contextWindow: 200000,
@@ -13283,6 +13504,23 @@ export const MODELS = {
13283
13504
  contextWindow: 200000,
13284
13505
  maxTokens: 128000,
13285
13506
  },
13507
+ "zai/glm-5": {
13508
+ id: "zai/glm-5",
13509
+ name: "GLM 5",
13510
+ api: "anthropic-messages",
13511
+ provider: "vercel-ai-gateway",
13512
+ baseUrl: "https://ai-gateway.vercel.sh",
13513
+ reasoning: true,
13514
+ input: ["text"],
13515
+ cost: {
13516
+ input: 1,
13517
+ output: 3.1999999999999997,
13518
+ cacheRead: 0.19999999999999998,
13519
+ cacheWrite: 0,
13520
+ },
13521
+ contextWindow: 202800,
13522
+ maxTokens: 131100,
13523
+ },
13286
13524
  "zai/glm-5-turbo": {
13287
13525
  id: "zai/glm-5-turbo",
13288
13526
  name: "GLM 5 Turbo",
@@ -13300,6 +13538,23 @@ export const MODELS = {
13300
13538
  contextWindow: 202800,
13301
13539
  maxTokens: 131100,
13302
13540
  },
13541
+ "zai/glm-5v-turbo": {
13542
+ id: "zai/glm-5v-turbo",
13543
+ name: "GLM 5V Turbo",
13544
+ api: "anthropic-messages",
13545
+ provider: "vercel-ai-gateway",
13546
+ baseUrl: "https://ai-gateway.vercel.sh",
13547
+ reasoning: true,
13548
+ input: ["text", "image"],
13549
+ cost: {
13550
+ input: 1.2,
13551
+ output: 4,
13552
+ cacheRead: 0.24,
13553
+ cacheWrite: 0,
13554
+ },
13555
+ contextWindow: 200000,
13556
+ maxTokens: 128000,
13557
+ },
13303
13558
  },
13304
13559
  "xai": {
13305
13560
  "grok-2": {
@@ -13790,7 +14045,7 @@ export const MODELS = {
13790
14045
  api: "openai-completions",
13791
14046
  provider: "zai",
13792
14047
  baseUrl: "https://api.z.ai/api/coding/paas/v4",
13793
- compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
14048
+ compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
13794
14049
  reasoning: true,
13795
14050
  input: ["text"],
13796
14051
  cost: {
@@ -13808,7 +14063,7 @@ export const MODELS = {
13808
14063
  api: "openai-completions",
13809
14064
  provider: "zai",
13810
14065
  baseUrl: "https://api.z.ai/api/coding/paas/v4",
13811
- compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
14066
+ compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
13812
14067
  reasoning: true,
13813
14068
  input: ["text", "image"],
13814
14069
  cost: {
@@ -13826,7 +14081,7 @@ export const MODELS = {
13826
14081
  api: "openai-completions",
13827
14082
  provider: "zai",
13828
14083
  baseUrl: "https://api.z.ai/api/coding/paas/v4",
13829
- compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
14084
+ compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
13830
14085
  reasoning: true,
13831
14086
  input: ["text"],
13832
14087
  cost: {
@@ -13844,7 +14099,7 @@ export const MODELS = {
13844
14099
  api: "openai-completions",
13845
14100
  provider: "zai",
13846
14101
  baseUrl: "https://api.z.ai/api/coding/paas/v4",
13847
- compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
14102
+ compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
13848
14103
  reasoning: true,
13849
14104
  input: ["text"],
13850
14105
  cost: {
@@ -13862,7 +14117,7 @@ export const MODELS = {
13862
14117
  api: "openai-completions",
13863
14118
  provider: "zai",
13864
14119
  baseUrl: "https://api.z.ai/api/coding/paas/v4",
13865
- compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
14120
+ compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
13866
14121
  reasoning: true,
13867
14122
  input: ["text"],
13868
14123
  cost: {
@@ -13880,7 +14135,7 @@ export const MODELS = {
13880
14135
  api: "openai-completions",
13881
14136
  provider: "zai",
13882
14137
  baseUrl: "https://api.z.ai/api/coding/paas/v4",
13883
- compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
14138
+ compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
13884
14139
  reasoning: true,
13885
14140
  input: ["text"],
13886
14141
  cost: {
@@ -13898,7 +14153,7 @@ export const MODELS = {
13898
14153
  api: "openai-completions",
13899
14154
  provider: "zai",
13900
14155
  baseUrl: "https://api.z.ai/api/coding/paas/v4",
13901
- compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
14156
+ compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
13902
14157
  reasoning: true,
13903
14158
  input: ["text"],
13904
14159
  cost: {
@@ -13910,6 +14165,24 @@ export const MODELS = {
13910
14165
  contextWindow: 200000,
13911
14166
  maxTokens: 131072,
13912
14167
  },
14168
+ "glm-5v-turbo": {
14169
+ id: "glm-5v-turbo",
14170
+ name: "glm-5v-turbo",
14171
+ api: "openai-completions",
14172
+ provider: "zai",
14173
+ baseUrl: "https://api.z.ai/api/coding/paas/v4",
14174
+ compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
14175
+ reasoning: true,
14176
+ input: ["text", "image"],
14177
+ cost: {
14178
+ input: 1.2,
14179
+ output: 4,
14180
+ cacheRead: 0.24,
14181
+ cacheWrite: 0,
14182
+ },
14183
+ contextWindow: 200000,
14184
+ maxTokens: 131072,
14185
+ },
13913
14186
  },
13914
14187
  };
13915
14188
  //# sourceMappingURL=models.generated.js.map