@mariozechner/pi-ai 0.52.11 → 0.53.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -427,6 +427,23 @@ export const MODELS = {
427
427
  contextWindow: 163840,
428
428
  maxTokens: 81920,
429
429
  },
430
+ "deepseek.v3.2-v1:0": {
431
+ id: "deepseek.v3.2-v1:0",
432
+ name: "DeepSeek-V3.2",
433
+ api: "bedrock-converse-stream",
434
+ provider: "amazon-bedrock",
435
+ baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
436
+ reasoning: true,
437
+ input: ["text"],
438
+ cost: {
439
+ input: 0.62,
440
+ output: 1.85,
441
+ cacheRead: 0,
442
+ cacheWrite: 0,
443
+ },
444
+ contextWindow: 163840,
445
+ maxTokens: 81920,
446
+ },
430
447
  "eu.anthropic.claude-haiku-4-5-20251001-v1:0": {
431
448
  id: "eu.anthropic.claude-haiku-4-5-20251001-v1:0",
432
449
  name: "Claude Haiku 4.5 (EU)",
@@ -1704,6 +1721,23 @@ export const MODELS = {
1704
1721
  contextWindow: 200000,
1705
1722
  maxTokens: 64000,
1706
1723
  },
1724
+ "claude-sonnet-4-6": {
1725
+ id: "claude-sonnet-4-6",
1726
+ name: "Claude Sonnet 4.6",
1727
+ api: "anthropic-messages",
1728
+ provider: "anthropic",
1729
+ baseUrl: "https://api.anthropic.com",
1730
+ reasoning: true,
1731
+ input: ["text", "image"],
1732
+ cost: {
1733
+ input: 3,
1734
+ output: 15,
1735
+ cacheRead: 0.3,
1736
+ cacheWrite: 3.75,
1737
+ },
1738
+ contextWindow: 200000,
1739
+ maxTokens: 64000,
1740
+ },
1707
1741
  },
1708
1742
  "azure-openai-responses": {
1709
1743
  "codex-mini-latest": {
@@ -2380,8 +2414,8 @@ export const MODELS = {
2380
2414
  reasoning: false,
2381
2415
  input: ["text"],
2382
2416
  cost: {
2383
- input: 0,
2384
- output: 0,
2417
+ input: 2.25,
2418
+ output: 2.75,
2385
2419
  cacheRead: 0,
2386
2420
  cacheWrite: 0,
2387
2421
  },
@@ -3771,6 +3805,24 @@ export const MODELS = {
3771
3805
  contextWindow: 204800,
3772
3806
  maxTokens: 131072,
3773
3807
  },
3808
+ "MiniMaxAI/MiniMax-M2.5": {
3809
+ id: "MiniMaxAI/MiniMax-M2.5",
3810
+ name: "MiniMax-M2.5",
3811
+ api: "openai-completions",
3812
+ provider: "huggingface",
3813
+ baseUrl: "https://router.huggingface.co/v1",
3814
+ compat: { "supportsDeveloperRole": false },
3815
+ reasoning: true,
3816
+ input: ["text"],
3817
+ cost: {
3818
+ input: 0.3,
3819
+ output: 1.2,
3820
+ cacheRead: 0.03,
3821
+ cacheWrite: 0,
3822
+ },
3823
+ contextWindow: 204800,
3824
+ maxTokens: 131072,
3825
+ },
3774
3826
  "Qwen/Qwen3-235B-A22B-Thinking-2507": {
3775
3827
  id: "Qwen/Qwen3-235B-A22B-Thinking-2507",
3776
3828
  name: "Qwen3-235B-A22B-Thinking-2507",
@@ -3807,6 +3859,24 @@ export const MODELS = {
3807
3859
  contextWindow: 262144,
3808
3860
  maxTokens: 66536,
3809
3861
  },
3862
+ "Qwen/Qwen3-Coder-Next": {
3863
+ id: "Qwen/Qwen3-Coder-Next",
3864
+ name: "Qwen3-Coder-Next",
3865
+ api: "openai-completions",
3866
+ provider: "huggingface",
3867
+ baseUrl: "https://router.huggingface.co/v1",
3868
+ compat: { "supportsDeveloperRole": false },
3869
+ reasoning: false,
3870
+ input: ["text"],
3871
+ cost: {
3872
+ input: 0.2,
3873
+ output: 1.5,
3874
+ cacheRead: 0,
3875
+ cacheWrite: 0,
3876
+ },
3877
+ contextWindow: 262144,
3878
+ maxTokens: 65536,
3879
+ },
3810
3880
  "Qwen/Qwen3-Next-80B-A3B-Instruct": {
3811
3881
  id: "Qwen/Qwen3-Next-80B-A3B-Instruct",
3812
3882
  name: "Qwen3-Next-80B-A3B-Instruct",
@@ -3843,6 +3913,24 @@ export const MODELS = {
3843
3913
  contextWindow: 262144,
3844
3914
  maxTokens: 131072,
3845
3915
  },
3916
+ "Qwen/Qwen3.5-397B-A17B": {
3917
+ id: "Qwen/Qwen3.5-397B-A17B",
3918
+ name: "Qwen3.5-397B-A17B",
3919
+ api: "openai-completions",
3920
+ provider: "huggingface",
3921
+ baseUrl: "https://router.huggingface.co/v1",
3922
+ compat: { "supportsDeveloperRole": false },
3923
+ reasoning: true,
3924
+ input: ["text", "image"],
3925
+ cost: {
3926
+ input: 0.6,
3927
+ output: 3.6,
3928
+ cacheRead: 0,
3929
+ cacheWrite: 0,
3930
+ },
3931
+ contextWindow: 262144,
3932
+ maxTokens: 32768,
3933
+ },
3846
3934
  "XiaomiMiMo/MiMo-V2-Flash": {
3847
3935
  id: "XiaomiMiMo/MiMo-V2-Flash",
3848
3936
  name: "MiMo-V2-Flash",
@@ -4112,6 +4200,23 @@ export const MODELS = {
4112
4200
  contextWindow: 204800,
4113
4201
  maxTokens: 131072,
4114
4202
  },
4203
+ "MiniMax-M2.5-highspeed": {
4204
+ id: "MiniMax-M2.5-highspeed",
4205
+ name: "MiniMax-M2.5-highspeed",
4206
+ api: "anthropic-messages",
4207
+ provider: "minimax",
4208
+ baseUrl: "https://api.minimax.io/anthropic",
4209
+ reasoning: true,
4210
+ input: ["text"],
4211
+ cost: {
4212
+ input: 0.6,
4213
+ output: 2.4,
4214
+ cacheRead: 0.06,
4215
+ cacheWrite: 0.375,
4216
+ },
4217
+ contextWindow: 204800,
4218
+ maxTokens: 131072,
4219
+ },
4115
4220
  },
4116
4221
  "minimax-cn": {
4117
4222
  "MiniMax-M2": {
@@ -4165,6 +4270,23 @@ export const MODELS = {
4165
4270
  contextWindow: 204800,
4166
4271
  maxTokens: 131072,
4167
4272
  },
4273
+ "MiniMax-M2.5-highspeed": {
4274
+ id: "MiniMax-M2.5-highspeed",
4275
+ name: "MiniMax-M2.5-highspeed",
4276
+ api: "anthropic-messages",
4277
+ provider: "minimax-cn",
4278
+ baseUrl: "https://api.minimaxi.com/anthropic",
4279
+ reasoning: true,
4280
+ input: ["text"],
4281
+ cost: {
4282
+ input: 0.6,
4283
+ output: 2.4,
4284
+ cacheRead: 0.06,
4285
+ cacheWrite: 0.375,
4286
+ },
4287
+ contextWindow: 204800,
4288
+ maxTokens: 131072,
4289
+ },
4168
4290
  },
4169
4291
  "mistral": {
4170
4292
  "codestral-latest": {
@@ -5533,6 +5655,40 @@ export const MODELS = {
5533
5655
  contextWindow: 204800,
5534
5656
  maxTokens: 131072,
5535
5657
  },
5658
+ "glm-5": {
5659
+ id: "glm-5",
5660
+ name: "GLM-5",
5661
+ api: "openai-completions",
5662
+ provider: "opencode",
5663
+ baseUrl: "https://opencode.ai/zen/v1",
5664
+ reasoning: true,
5665
+ input: ["text"],
5666
+ cost: {
5667
+ input: 1,
5668
+ output: 3.2,
5669
+ cacheRead: 0.2,
5670
+ cacheWrite: 0,
5671
+ },
5672
+ contextWindow: 204800,
5673
+ maxTokens: 131072,
5674
+ },
5675
+ "glm-5-free": {
5676
+ id: "glm-5-free",
5677
+ name: "GLM-5 Free",
5678
+ api: "openai-completions",
5679
+ provider: "opencode",
5680
+ baseUrl: "https://opencode.ai/zen/v1",
5681
+ reasoning: true,
5682
+ input: ["text"],
5683
+ cost: {
5684
+ input: 0,
5685
+ output: 0,
5686
+ cacheRead: 0,
5687
+ cacheWrite: 0,
5688
+ },
5689
+ contextWindow: 204800,
5690
+ maxTokens: 131072,
5691
+ },
5536
5692
  "gpt-5": {
5537
5693
  id: "gpt-5",
5538
5694
  name: "GPT-5",
@@ -5771,6 +5927,23 @@ export const MODELS = {
5771
5927
  contextWindow: 204800,
5772
5928
  maxTokens: 131072,
5773
5929
  },
5930
+ "minimax-m2.5": {
5931
+ id: "minimax-m2.5",
5932
+ name: "MiniMax M2.5",
5933
+ api: "openai-completions",
5934
+ provider: "opencode",
5935
+ baseUrl: "https://opencode.ai/zen/v1",
5936
+ reasoning: true,
5937
+ input: ["text"],
5938
+ cost: {
5939
+ input: 0.3,
5940
+ output: 1.2,
5941
+ cacheRead: 0.06,
5942
+ cacheWrite: 0,
5943
+ },
5944
+ contextWindow: 204800,
5945
+ maxTokens: 131072,
5946
+ },
5774
5947
  "minimax-m2.5-free": {
5775
5948
  id: "minimax-m2.5-free",
5776
5949
  name: "MiniMax M2.5 Free",
@@ -6130,6 +6303,23 @@ export const MODELS = {
6130
6303
  contextWindow: 1000000,
6131
6304
  maxTokens: 64000,
6132
6305
  },
6306
+ "anthropic/claude-sonnet-4.6": {
6307
+ id: "anthropic/claude-sonnet-4.6",
6308
+ name: "Anthropic: Claude Sonnet 4.6",
6309
+ api: "openai-completions",
6310
+ provider: "openrouter",
6311
+ baseUrl: "https://openrouter.ai/api/v1",
6312
+ reasoning: true,
6313
+ input: ["text", "image"],
6314
+ cost: {
6315
+ input: 3,
6316
+ output: 15,
6317
+ cacheRead: 0.3,
6318
+ cacheWrite: 3.75,
6319
+ },
6320
+ contextWindow: 1000000,
6321
+ maxTokens: 128000,
6322
+ },
6133
6323
  "arcee-ai/trinity-large-preview:free": {
6134
6324
  id: "arcee-ai/trinity-large-preview:free",
6135
6325
  name: "Arcee AI: Trinity Large Preview (free)",
@@ -6445,13 +6635,13 @@ export const MODELS = {
6445
6635
  reasoning: true,
6446
6636
  input: ["text"],
6447
6637
  cost: {
6448
- input: 0.25,
6638
+ input: 0.26,
6449
6639
  output: 0.38,
6450
- cacheRead: 0.125,
6640
+ cacheRead: 0.13,
6451
6641
  cacheWrite: 0,
6452
6642
  },
6453
6643
  contextWindow: 163840,
6454
- maxTokens: 65536,
6644
+ maxTokens: 4096,
6455
6645
  },
6456
6646
  "deepseek/deepseek-v3.2-exp": {
6457
6647
  id: "deepseek/deepseek-v3.2-exp",
@@ -6939,12 +7129,12 @@ export const MODELS = {
6939
7129
  input: ["text"],
6940
7130
  cost: {
6941
7131
  input: 0.3,
6942
- output: 1.2,
6943
- cacheRead: 0.03,
7132
+ output: 1.1,
7133
+ cacheRead: 0.15,
6944
7134
  cacheWrite: 0,
6945
7135
  },
6946
- contextWindow: 204800,
6947
- maxTokens: 131072,
7136
+ contextWindow: 196608,
7137
+ maxTokens: 65536,
6948
7138
  },
6949
7139
  "mistralai/codestral-2508": {
6950
7140
  id: "mistralai/codestral-2508",
@@ -7380,13 +7570,13 @@ export const MODELS = {
7380
7570
  reasoning: false,
7381
7571
  input: ["text"],
7382
7572
  cost: {
7383
- input: 0.39,
7384
- output: 1.9,
7385
- cacheRead: 0.195,
7573
+ input: 0.39999999999999997,
7574
+ output: 2,
7575
+ cacheRead: 0.15,
7386
7576
  cacheWrite: 0,
7387
7577
  },
7388
- contextWindow: 262144,
7389
- maxTokens: 262144,
7578
+ contextWindow: 131072,
7579
+ maxTokens: 4096,
7390
7580
  },
7391
7581
  "moonshotai/kimi-k2-0905:exacto": {
7392
7582
  id: "moonshotai/kimi-k2-0905:exacto",
@@ -7431,13 +7621,13 @@ export const MODELS = {
7431
7621
  reasoning: true,
7432
7622
  input: ["text", "image"],
7433
7623
  cost: {
7434
- input: 0.44999999999999996,
7435
- output: 2.25,
7436
- cacheRead: 0.070000002,
7624
+ input: 0.22999999999999998,
7625
+ output: 3,
7626
+ cacheRead: 0,
7437
7627
  cacheWrite: 0,
7438
7628
  },
7439
7629
  contextWindow: 262144,
7440
- maxTokens: 4096,
7630
+ maxTokens: 262144,
7441
7631
  },
7442
7632
  "nex-agi/deepseek-v3.1-nex-n1": {
7443
7633
  id: "nex-agi/deepseek-v3.1-nex-n1",
@@ -8910,9 +9100,9 @@ export const MODELS = {
8910
9100
  reasoning: false,
8911
9101
  input: ["text"],
8912
9102
  cost: {
8913
- input: 0.07,
8914
- output: 0.3,
8915
- cacheRead: 0.035,
9103
+ input: 0.12,
9104
+ output: 0.75,
9105
+ cacheRead: 0.06,
8916
9106
  cacheWrite: 0,
8917
9107
  },
8918
9108
  contextWindow: 262144,
@@ -9173,6 +9363,40 @@ export const MODELS = {
9173
9363
  contextWindow: 131072,
9174
9364
  maxTokens: 32768,
9175
9365
  },
9366
+ "qwen/qwen3.5-397b-a17b": {
9367
+ id: "qwen/qwen3.5-397b-a17b",
9368
+ name: "Qwen: Qwen3.5 397B A17B",
9369
+ api: "openai-completions",
9370
+ provider: "openrouter",
9371
+ baseUrl: "https://openrouter.ai/api/v1",
9372
+ reasoning: true,
9373
+ input: ["text", "image"],
9374
+ cost: {
9375
+ input: 0.15,
9376
+ output: 1,
9377
+ cacheRead: 0.15,
9378
+ cacheWrite: 0,
9379
+ },
9380
+ contextWindow: 262144,
9381
+ maxTokens: 65536,
9382
+ },
9383
+ "qwen/qwen3.5-plus-02-15": {
9384
+ id: "qwen/qwen3.5-plus-02-15",
9385
+ name: "Qwen: Qwen3.5 Plus 2026-02-15",
9386
+ api: "openai-completions",
9387
+ provider: "openrouter",
9388
+ baseUrl: "https://openrouter.ai/api/v1",
9389
+ reasoning: true,
9390
+ input: ["text", "image"],
9391
+ cost: {
9392
+ input: 0.39999999999999997,
9393
+ output: 2.4,
9394
+ cacheRead: 0,
9395
+ cacheWrite: 0,
9396
+ },
9397
+ contextWindow: 1000000,
9398
+ maxTokens: 65536,
9399
+ },
9176
9400
  "qwen/qwq-32b": {
9177
9401
  id: "qwen/qwq-32b",
9178
9402
  name: "Qwen: QwQ 32B",
@@ -9343,23 +9567,6 @@ export const MODELS = {
9343
9567
  contextWindow: 163840,
9344
9568
  maxTokens: 65536,
9345
9569
  },
9346
- "tngtech/tng-r1t-chimera:free": {
9347
- id: "tngtech/tng-r1t-chimera:free",
9348
- name: "TNG: R1T Chimera (free)",
9349
- api: "openai-completions",
9350
- provider: "openrouter",
9351
- baseUrl: "https://openrouter.ai/api/v1",
9352
- reasoning: true,
9353
- input: ["text"],
9354
- cost: {
9355
- input: 0,
9356
- output: 0,
9357
- cacheRead: 0,
9358
- cacheWrite: 0,
9359
- },
9360
- contextWindow: 163840,
9361
- maxTokens: 65536,
9362
- },
9363
9570
  "upstage/solar-pro-3:free": {
9364
9571
  id: "upstage/solar-pro-3:free",
9365
9572
  name: "Upstage: Solar Pro 3 (free)",
@@ -9624,9 +9831,9 @@ export const MODELS = {
9624
9831
  reasoning: true,
9625
9832
  input: ["text"],
9626
9833
  cost: {
9627
- input: 0.35,
9628
- output: 1.5,
9629
- cacheRead: 0.175,
9834
+ input: 0.33999999999999997,
9835
+ output: 1.7,
9836
+ cacheRead: 0.16999999999999998,
9630
9837
  cacheWrite: 0,
9631
9838
  },
9632
9839
  contextWindow: 202752,
@@ -9675,9 +9882,9 @@ export const MODELS = {
9675
9882
  reasoning: true,
9676
9883
  input: ["text"],
9677
9884
  cost: {
9678
- input: 0.39999999999999997,
9679
- output: 1.5,
9680
- cacheRead: 0.19999999999999998,
9885
+ input: 0.38,
9886
+ output: 1.7,
9887
+ cacheRead: 0.19,
9681
9888
  cacheWrite: 0,
9682
9889
  },
9683
9890
  contextWindow: 202752,
@@ -9709,12 +9916,12 @@ export const MODELS = {
9709
9916
  reasoning: true,
9710
9917
  input: ["text"],
9711
9918
  cost: {
9712
- input: 0.7999999999999999,
9713
- output: 2.56,
9714
- cacheRead: 0.16,
9919
+ input: 0.3,
9920
+ output: 2.5500000000000003,
9921
+ cacheRead: 0,
9715
9922
  cacheWrite: 0,
9716
9923
  },
9717
- contextWindow: 202752,
9924
+ contextWindow: 204800,
9718
9925
  maxTokens: 131072,
9719
9926
  },
9720
9927
  },
@@ -9923,6 +10130,23 @@ export const MODELS = {
9923
10130
  contextWindow: 256000,
9924
10131
  maxTokens: 256000,
9925
10132
  },
10133
+ "alibaba/qwen3.5-plus": {
10134
+ id: "alibaba/qwen3.5-plus",
10135
+ name: "Qwen 3.5 Plus",
10136
+ api: "anthropic-messages",
10137
+ provider: "vercel-ai-gateway",
10138
+ baseUrl: "https://ai-gateway.vercel.sh",
10139
+ reasoning: true,
10140
+ input: ["text", "image"],
10141
+ cost: {
10142
+ input: 0.39999999999999997,
10143
+ output: 2.4,
10144
+ cacheRead: 0.04,
10145
+ cacheWrite: 0.5,
10146
+ },
10147
+ contextWindow: 1000000,
10148
+ maxTokens: 64000,
10149
+ },
9926
10150
  "anthropic/claude-3-haiku": {
9927
10151
  id: "anthropic/claude-3-haiku",
9928
10152
  name: "Claude 3 Haiku",
@@ -10127,6 +10351,23 @@ export const MODELS = {
10127
10351
  contextWindow: 1000000,
10128
10352
  maxTokens: 64000,
10129
10353
  },
10354
+ "anthropic/claude-sonnet-4.6": {
10355
+ id: "anthropic/claude-sonnet-4.6",
10356
+ name: "Claude Sonnet 4.6",
10357
+ api: "anthropic-messages",
10358
+ provider: "vercel-ai-gateway",
10359
+ baseUrl: "https://ai-gateway.vercel.sh",
10360
+ reasoning: true,
10361
+ input: ["text", "image"],
10362
+ cost: {
10363
+ input: 3,
10364
+ output: 15,
10365
+ cacheRead: 0.3,
10366
+ cacheWrite: 3.75,
10367
+ },
10368
+ contextWindow: 1000000,
10369
+ maxTokens: 64000,
10370
+ },
10130
10371
  "arcee-ai/trinity-large-preview": {
10131
10372
  id: "arcee-ai/trinity-large-preview",
10132
10373
  name: "Trinity Large Preview",