@mariozechner/pi-ai 0.62.0 → 0.63.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -801,6 +801,23 @@ export const MODELS = {
801
801
  contextWindow: 204800,
802
802
  maxTokens: 131072,
803
803
  },
804
+ "minimax.minimax-m2.5": {
805
+ id: "minimax.minimax-m2.5",
806
+ name: "MiniMax M2.5",
807
+ api: "bedrock-converse-stream",
808
+ provider: "amazon-bedrock",
809
+ baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
810
+ reasoning: true,
811
+ input: ["text"],
812
+ cost: {
813
+ input: 0.3,
814
+ output: 1.2,
815
+ cacheRead: 0,
816
+ cacheWrite: 0,
817
+ },
818
+ contextWindow: 196608,
819
+ maxTokens: 98304,
820
+ },
804
821
  "mistral.devstral-2-123b": {
805
822
  id: "mistral.devstral-2-123b",
806
823
  name: "Devstral 2 123B",
@@ -1039,6 +1056,23 @@ export const MODELS = {
1039
1056
  contextWindow: 128000,
1040
1057
  maxTokens: 4096,
1041
1058
  },
1059
+ "nvidia.nemotron-super-3-120b": {
1060
+ id: "nvidia.nemotron-super-3-120b",
1061
+ name: "NVIDIA Nemotron 3 Super 120B A12B",
1062
+ api: "bedrock-converse-stream",
1063
+ provider: "amazon-bedrock",
1064
+ baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
1065
+ reasoning: true,
1066
+ input: ["text"],
1067
+ cost: {
1068
+ input: 0.15,
1069
+ output: 0.65,
1070
+ cacheRead: 0,
1071
+ cacheWrite: 0,
1072
+ },
1073
+ contextWindow: 262144,
1074
+ maxTokens: 131072,
1075
+ },
1042
1076
  "openai.gpt-oss-120b-1:0": {
1043
1077
  id: "openai.gpt-oss-120b-1:0",
1044
1078
  name: "gpt-oss-120b",
@@ -1413,6 +1447,23 @@ export const MODELS = {
1413
1447
  contextWindow: 200000,
1414
1448
  maxTokens: 131072,
1415
1449
  },
1450
+ "zai.glm-5": {
1451
+ id: "zai.glm-5",
1452
+ name: "GLM-5",
1453
+ api: "bedrock-converse-stream",
1454
+ provider: "amazon-bedrock",
1455
+ baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
1456
+ reasoning: true,
1457
+ input: ["text"],
1458
+ cost: {
1459
+ input: 1,
1460
+ output: 3.2,
1461
+ cacheRead: 0,
1462
+ cacheWrite: 0,
1463
+ },
1464
+ contextWindow: 202752,
1465
+ maxTokens: 101376,
1466
+ },
1416
1467
  },
1417
1468
  "anthropic": {
1418
1469
  "claude-3-5-haiku-20241022": {
@@ -3874,6 +3925,23 @@ export const MODELS = {
3874
3925
  contextWindow: 1048576,
3875
3926
  maxTokens: 65536,
3876
3927
  },
3928
+ "gemini-3.1-pro-preview-customtools": {
3929
+ id: "gemini-3.1-pro-preview-customtools",
3930
+ name: "Gemini 3.1 Pro Preview Custom Tools (Vertex)",
3931
+ api: "google-vertex",
3932
+ provider: "google-vertex",
3933
+ baseUrl: "https://{location}-aiplatform.googleapis.com",
3934
+ reasoning: true,
3935
+ input: ["text", "image"],
3936
+ cost: {
3937
+ input: 2,
3938
+ output: 12,
3939
+ cacheRead: 0.2,
3940
+ cacheWrite: 0,
3941
+ },
3942
+ contextWindow: 1048576,
3943
+ maxTokens: 65536,
3944
+ },
3877
3945
  },
3878
3946
  "groq": {
3879
3947
  "deepseek-r1-distill-llama-70b": {
@@ -4495,91 +4563,6 @@ export const MODELS = {
4495
4563
  },
4496
4564
  },
4497
4565
  "minimax": {
4498
- "MiniMax-M2": {
4499
- id: "MiniMax-M2",
4500
- name: "MiniMax-M2",
4501
- api: "anthropic-messages",
4502
- provider: "minimax",
4503
- baseUrl: "https://api.minimax.io/anthropic",
4504
- reasoning: true,
4505
- input: ["text"],
4506
- cost: {
4507
- input: 0.3,
4508
- output: 1.2,
4509
- cacheRead: 0,
4510
- cacheWrite: 0,
4511
- },
4512
- contextWindow: 204800,
4513
- maxTokens: 131072,
4514
- },
4515
- "MiniMax-M2.1": {
4516
- id: "MiniMax-M2.1",
4517
- name: "MiniMax-M2.1",
4518
- api: "anthropic-messages",
4519
- provider: "minimax",
4520
- baseUrl: "https://api.minimax.io/anthropic",
4521
- reasoning: true,
4522
- input: ["text"],
4523
- cost: {
4524
- input: 0.3,
4525
- output: 1.2,
4526
- cacheRead: 0,
4527
- cacheWrite: 0,
4528
- },
4529
- contextWindow: 204800,
4530
- maxTokens: 131072,
4531
- },
4532
- "MiniMax-M2.1-highspeed": {
4533
- id: "MiniMax-M2.1-highspeed",
4534
- name: "MiniMax-M2.1-highspeed",
4535
- api: "anthropic-messages",
4536
- provider: "minimax",
4537
- baseUrl: "https://api.minimax.io/anthropic",
4538
- reasoning: true,
4539
- input: ["text"],
4540
- cost: {
4541
- input: 0.6,
4542
- output: 2.4,
4543
- cacheRead: 0,
4544
- cacheWrite: 0,
4545
- },
4546
- contextWindow: 204800,
4547
- maxTokens: 131072,
4548
- },
4549
- "MiniMax-M2.5": {
4550
- id: "MiniMax-M2.5",
4551
- name: "MiniMax-M2.5",
4552
- api: "anthropic-messages",
4553
- provider: "minimax",
4554
- baseUrl: "https://api.minimax.io/anthropic",
4555
- reasoning: true,
4556
- input: ["text"],
4557
- cost: {
4558
- input: 0.3,
4559
- output: 1.2,
4560
- cacheRead: 0.03,
4561
- cacheWrite: 0.375,
4562
- },
4563
- contextWindow: 204800,
4564
- maxTokens: 131072,
4565
- },
4566
- "MiniMax-M2.5-highspeed": {
4567
- id: "MiniMax-M2.5-highspeed",
4568
- name: "MiniMax-M2.5-highspeed",
4569
- api: "anthropic-messages",
4570
- provider: "minimax",
4571
- baseUrl: "https://api.minimax.io/anthropic",
4572
- reasoning: true,
4573
- input: ["text"],
4574
- cost: {
4575
- input: 0.6,
4576
- output: 2.4,
4577
- cacheRead: 0.06,
4578
- cacheWrite: 0.375,
4579
- },
4580
- contextWindow: 204800,
4581
- maxTokens: 131072,
4582
- },
4583
4566
  "MiniMax-M2.7": {
4584
4567
  id: "MiniMax-M2.7",
4585
4568
  name: "MiniMax-M2.7",
@@ -4616,91 +4599,6 @@ export const MODELS = {
4616
4599
  },
4617
4600
  },
4618
4601
  "minimax-cn": {
4619
- "MiniMax-M2": {
4620
- id: "MiniMax-M2",
4621
- name: "MiniMax-M2",
4622
- api: "anthropic-messages",
4623
- provider: "minimax-cn",
4624
- baseUrl: "https://api.minimaxi.com/anthropic",
4625
- reasoning: true,
4626
- input: ["text"],
4627
- cost: {
4628
- input: 0.3,
4629
- output: 1.2,
4630
- cacheRead: 0,
4631
- cacheWrite: 0,
4632
- },
4633
- contextWindow: 204800,
4634
- maxTokens: 131072,
4635
- },
4636
- "MiniMax-M2.1": {
4637
- id: "MiniMax-M2.1",
4638
- name: "MiniMax-M2.1",
4639
- api: "anthropic-messages",
4640
- provider: "minimax-cn",
4641
- baseUrl: "https://api.minimaxi.com/anthropic",
4642
- reasoning: true,
4643
- input: ["text"],
4644
- cost: {
4645
- input: 0.3,
4646
- output: 1.2,
4647
- cacheRead: 0,
4648
- cacheWrite: 0,
4649
- },
4650
- contextWindow: 204800,
4651
- maxTokens: 131072,
4652
- },
4653
- "MiniMax-M2.1-highspeed": {
4654
- id: "MiniMax-M2.1-highspeed",
4655
- name: "MiniMax-M2.1-highspeed",
4656
- api: "anthropic-messages",
4657
- provider: "minimax-cn",
4658
- baseUrl: "https://api.minimaxi.com/anthropic",
4659
- reasoning: true,
4660
- input: ["text"],
4661
- cost: {
4662
- input: 0.6,
4663
- output: 2.4,
4664
- cacheRead: 0,
4665
- cacheWrite: 0,
4666
- },
4667
- contextWindow: 204800,
4668
- maxTokens: 131072,
4669
- },
4670
- "MiniMax-M2.5": {
4671
- id: "MiniMax-M2.5",
4672
- name: "MiniMax-M2.5",
4673
- api: "anthropic-messages",
4674
- provider: "minimax-cn",
4675
- baseUrl: "https://api.minimaxi.com/anthropic",
4676
- reasoning: true,
4677
- input: ["text"],
4678
- cost: {
4679
- input: 0.3,
4680
- output: 1.2,
4681
- cacheRead: 0.03,
4682
- cacheWrite: 0.375,
4683
- },
4684
- contextWindow: 204800,
4685
- maxTokens: 131072,
4686
- },
4687
- "MiniMax-M2.5-highspeed": {
4688
- id: "MiniMax-M2.5-highspeed",
4689
- name: "MiniMax-M2.5-highspeed",
4690
- api: "anthropic-messages",
4691
- provider: "minimax-cn",
4692
- baseUrl: "https://api.minimaxi.com/anthropic",
4693
- reasoning: true,
4694
- input: ["text"],
4695
- cost: {
4696
- input: 0.6,
4697
- output: 2.4,
4698
- cacheRead: 0.06,
4699
- cacheWrite: 0.375,
4700
- },
4701
- contextWindow: 204800,
4702
- maxTokens: 131072,
4703
- },
4704
4602
  "MiniMax-M2.7": {
4705
4603
  id: "MiniMax-M2.7",
4706
4604
  name: "MiniMax-M2.7",
@@ -7501,23 +7399,6 @@ export const MODELS = {
7501
7399
  contextWindow: 1048576,
7502
7400
  maxTokens: 65536,
7503
7401
  },
7504
- "google/gemini-3-pro-preview": {
7505
- id: "google/gemini-3-pro-preview",
7506
- name: "Google: Gemini 3 Pro Preview",
7507
- api: "openai-completions",
7508
- provider: "openrouter",
7509
- baseUrl: "https://openrouter.ai/api/v1",
7510
- reasoning: true,
7511
- input: ["text", "image"],
7512
- cost: {
7513
- input: 2,
7514
- output: 12,
7515
- cacheRead: 0.19999999999999998,
7516
- cacheWrite: 0.375,
7517
- },
7518
- contextWindow: 1048576,
7519
- maxTokens: 65536,
7520
- },
7521
7402
  "google/gemini-3.1-flash-lite-preview": {
7522
7403
  id: "google/gemini-3.1-flash-lite-preview",
7523
7404
  name: "Google: Gemini 3.1 Flash Lite Preview",
@@ -8292,13 +8173,13 @@ export const MODELS = {
8292
8173
  reasoning: false,
8293
8174
  input: ["text"],
8294
8175
  cost: {
8295
- input: 0.55,
8296
- output: 2.2,
8176
+ input: 0.5700000000000001,
8177
+ output: 2.3,
8297
8178
  cacheRead: 0,
8298
8179
  cacheWrite: 0,
8299
8180
  },
8300
- contextWindow: 131000,
8301
- maxTokens: 4096,
8181
+ contextWindow: 131072,
8182
+ maxTokens: 131072,
8302
8183
  },
8303
8184
  "moonshotai/kimi-k2-0905": {
8304
8185
  id: "moonshotai/kimi-k2-0905",
@@ -8360,8 +8241,8 @@ export const MODELS = {
8360
8241
  reasoning: false,
8361
8242
  input: ["text"],
8362
8243
  cost: {
8363
- input: 0.27,
8364
- output: 1,
8244
+ input: 0.135,
8245
+ output: 0.5,
8365
8246
  cacheRead: 0,
8366
8247
  cacheWrite: 0,
8367
8248
  },
@@ -8447,7 +8328,7 @@ export const MODELS = {
8447
8328
  cost: {
8448
8329
  input: 0.09999999999999999,
8449
8330
  output: 0.5,
8450
- cacheRead: 0.04,
8331
+ cacheRead: 0.09999999999999999,
8451
8332
  cacheWrite: 0,
8452
8333
  },
8453
8334
  contextWindow: 262144,
@@ -9539,7 +9420,7 @@ export const MODELS = {
9539
9420
  cacheWrite: 0,
9540
9421
  },
9541
9422
  contextWindow: 32768,
9542
- maxTokens: 4096,
9423
+ maxTokens: 32768,
9543
9424
  },
9544
9425
  "qwen/qwen-max": {
9545
9426
  id: "qwen/qwen-max",
@@ -10202,7 +10083,7 @@ export const MODELS = {
10202
10083
  cacheWrite: 0,
10203
10084
  },
10204
10085
  contextWindow: 256000,
10205
- maxTokens: 4096,
10086
+ maxTokens: 65536,
10206
10087
  },
10207
10088
  "qwen/qwen3.5-flash-02-23": {
10208
10089
  id: "qwen/qwen3.5-flash-02-23",
@@ -10255,6 +10136,23 @@ export const MODELS = {
10255
10136
  contextWindow: 131072,
10256
10137
  maxTokens: 131072,
10257
10138
  },
10139
+ "reka/reka-edge": {
10140
+ id: "reka/reka-edge",
10141
+ name: "Reka Edge",
10142
+ api: "openai-completions",
10143
+ provider: "openrouter",
10144
+ baseUrl: "https://openrouter.ai/api/v1",
10145
+ reasoning: false,
10146
+ input: ["text", "image"],
10147
+ cost: {
10148
+ input: 0.09999999999999999,
10149
+ output: 0.09999999999999999,
10150
+ cacheRead: 0,
10151
+ cacheWrite: 0,
10152
+ },
10153
+ contextWindow: 16384,
10154
+ maxTokens: 16384,
10155
+ },
10258
10156
  "relace/relace-search": {
10259
10157
  id: "relace/relace-search",
10260
10158
  name: "Relace: Relace Search",
@@ -10320,8 +10218,8 @@ export const MODELS = {
10320
10218
  cacheRead: 0.02,
10321
10219
  cacheWrite: 0,
10322
10220
  },
10323
- contextWindow: 256000,
10324
- maxTokens: 256000,
10221
+ contextWindow: 262144,
10222
+ maxTokens: 4096,
10325
10223
  },
10326
10224
  "stepfun/step-3.5-flash:free": {
10327
10225
  id: "stepfun/step-3.5-flash:free",
@@ -10791,9 +10689,9 @@ export const MODELS = {
10791
10689
  reasoning: true,
10792
10690
  input: ["text"],
10793
10691
  cost: {
10794
- input: 0.96,
10795
- output: 3.1999999999999997,
10796
- cacheRead: 0.192,
10692
+ input: 1.2,
10693
+ output: 4,
10694
+ cacheRead: 0.24,
10797
10695
  cacheWrite: 0,
10798
10696
  },
10799
10697
  contextWindow: 202752,
@@ -12799,8 +12697,8 @@ export const MODELS = {
12799
12697
  reasoning: false,
12800
12698
  input: ["text", "image"],
12801
12699
  cost: {
12802
- input: 1,
12803
- output: 1,
12700
+ input: 0,
12701
+ output: 0,
12804
12702
  cacheRead: 0,
12805
12703
  cacheWrite: 0,
12806
12704
  },
@@ -12816,8 +12714,8 @@ export const MODELS = {
12816
12714
  reasoning: false,
12817
12715
  input: ["text", "image"],
12818
12716
  cost: {
12819
- input: 3,
12820
- output: 15,
12717
+ input: 0,
12718
+ output: 0,
12821
12719
  cacheRead: 0,
12822
12720
  cacheWrite: 0,
12823
12721
  },
@@ -13011,6 +12909,23 @@ export const MODELS = {
13011
12909
  contextWindow: 2000000,
13012
12910
  maxTokens: 30000,
13013
12911
  },
12912
+ "xai/grok-4.20-multi-agent": {
12913
+ id: "xai/grok-4.20-multi-agent",
12914
+ name: "Grok 4.20 Multi-Agent",
12915
+ api: "anthropic-messages",
12916
+ provider: "vercel-ai-gateway",
12917
+ baseUrl: "https://ai-gateway.vercel.sh",
12918
+ reasoning: true,
12919
+ input: ["text"],
12920
+ cost: {
12921
+ input: 2,
12922
+ output: 6,
12923
+ cacheRead: 0.19999999999999998,
12924
+ cacheWrite: 0,
12925
+ },
12926
+ contextWindow: 2000000,
12927
+ maxTokens: 2000000,
12928
+ },
13014
12929
  "xai/grok-4.20-multi-agent-beta": {
13015
12930
  id: "xai/grok-4.20-multi-agent-beta",
13016
12931
  name: "Grok 4.20 Multi Agent Beta",
@@ -13028,6 +12943,23 @@ export const MODELS = {
13028
12943
  contextWindow: 2000000,
13029
12944
  maxTokens: 2000000,
13030
12945
  },
12946
+ "xai/grok-4.20-non-reasoning": {
12947
+ id: "xai/grok-4.20-non-reasoning",
12948
+ name: "Grok 4.20 Non-Reasoning",
12949
+ api: "anthropic-messages",
12950
+ provider: "vercel-ai-gateway",
12951
+ baseUrl: "https://ai-gateway.vercel.sh",
12952
+ reasoning: false,
12953
+ input: ["text", "image"],
12954
+ cost: {
12955
+ input: 2,
12956
+ output: 6,
12957
+ cacheRead: 0.19999999999999998,
12958
+ cacheWrite: 0,
12959
+ },
12960
+ contextWindow: 2000000,
12961
+ maxTokens: 2000000,
12962
+ },
13031
12963
  "xai/grok-4.20-non-reasoning-beta": {
13032
12964
  id: "xai/grok-4.20-non-reasoning-beta",
13033
12965
  name: "Grok 4.20 Beta Non-Reasoning",
@@ -13045,6 +12977,23 @@ export const MODELS = {
13045
12977
  contextWindow: 2000000,
13046
12978
  maxTokens: 2000000,
13047
12979
  },
12980
+ "xai/grok-4.20-reasoning": {
12981
+ id: "xai/grok-4.20-reasoning",
12982
+ name: "Grok 4.20 Reasoning",
12983
+ api: "anthropic-messages",
12984
+ provider: "vercel-ai-gateway",
12985
+ baseUrl: "https://ai-gateway.vercel.sh",
12986
+ reasoning: true,
12987
+ input: ["text", "image"],
12988
+ cost: {
12989
+ input: 2,
12990
+ output: 6,
12991
+ cacheRead: 0.19999999999999998,
12992
+ cacheWrite: 0,
12993
+ },
12994
+ contextWindow: 2000000,
12995
+ maxTokens: 2000000,
12996
+ },
13048
12997
  "xai/grok-4.20-reasoning-beta": {
13049
12998
  id: "xai/grok-4.20-reasoning-beta",
13050
12999
  name: "Grok 4.20 Beta Reasoning",
@@ -13856,6 +13805,24 @@ export const MODELS = {
13856
13805
  contextWindow: 200000,
13857
13806
  maxTokens: 131072,
13858
13807
  },
13808
+ "glm-4.7-flashx": {
13809
+ id: "glm-4.7-flashx",
13810
+ name: "GLM-4.7-FlashX",
13811
+ api: "openai-completions",
13812
+ provider: "zai",
13813
+ baseUrl: "https://api.z.ai/api/coding/paas/v4",
13814
+ compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
13815
+ reasoning: true,
13816
+ input: ["text"],
13817
+ cost: {
13818
+ input: 0.07,
13819
+ output: 0.4,
13820
+ cacheRead: 0.01,
13821
+ cacheWrite: 0,
13822
+ },
13823
+ contextWindow: 200000,
13824
+ maxTokens: 131072,
13825
+ },
13859
13826
  "glm-5": {
13860
13827
  id: "glm-5",
13861
13828
  name: "GLM-5",