@mariozechner/pi-ai 0.52.12 → 0.53.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -425,6 +425,23 @@ export declare const MODELS: {
425
425
  contextWindow: number;
426
426
  maxTokens: number;
427
427
  };
428
+ readonly "deepseek.v3.2-v1:0": {
429
+ id: string;
430
+ name: string;
431
+ api: "bedrock-converse-stream";
432
+ provider: string;
433
+ baseUrl: string;
434
+ reasoning: true;
435
+ input: "text"[];
436
+ cost: {
437
+ input: number;
438
+ output: number;
439
+ cacheRead: number;
440
+ cacheWrite: number;
441
+ };
442
+ contextWindow: number;
443
+ maxTokens: number;
444
+ };
428
445
  readonly "eu.anthropic.claude-haiku-4-5-20251001-v1:0": {
429
446
  id: string;
430
447
  name: string;
@@ -1702,6 +1719,23 @@ export declare const MODELS: {
1702
1719
  contextWindow: number;
1703
1720
  maxTokens: number;
1704
1721
  };
1722
+ readonly "claude-sonnet-4-6": {
1723
+ id: string;
1724
+ name: string;
1725
+ api: "anthropic-messages";
1726
+ provider: string;
1727
+ baseUrl: string;
1728
+ reasoning: true;
1729
+ input: ("image" | "text")[];
1730
+ cost: {
1731
+ input: number;
1732
+ output: number;
1733
+ cacheRead: number;
1734
+ cacheWrite: number;
1735
+ };
1736
+ contextWindow: number;
1737
+ maxTokens: number;
1738
+ };
1705
1739
  };
1706
1740
  readonly "azure-openai-responses": {
1707
1741
  readonly "codex-mini-latest": {
@@ -3890,6 +3924,26 @@ export declare const MODELS: {
3890
3924
  contextWindow: number;
3891
3925
  maxTokens: number;
3892
3926
  };
3927
+ readonly "MiniMaxAI/MiniMax-M2.5": {
3928
+ id: string;
3929
+ name: string;
3930
+ api: "openai-completions";
3931
+ provider: string;
3932
+ baseUrl: string;
3933
+ compat: {
3934
+ supportsDeveloperRole: false;
3935
+ };
3936
+ reasoning: true;
3937
+ input: "text"[];
3938
+ cost: {
3939
+ input: number;
3940
+ output: number;
3941
+ cacheRead: number;
3942
+ cacheWrite: number;
3943
+ };
3944
+ contextWindow: number;
3945
+ maxTokens: number;
3946
+ };
3893
3947
  readonly "Qwen/Qwen3-235B-A22B-Thinking-2507": {
3894
3948
  id: string;
3895
3949
  name: string;
@@ -3930,6 +3984,26 @@ export declare const MODELS: {
3930
3984
  contextWindow: number;
3931
3985
  maxTokens: number;
3932
3986
  };
3987
+ readonly "Qwen/Qwen3-Coder-Next": {
3988
+ id: string;
3989
+ name: string;
3990
+ api: "openai-completions";
3991
+ provider: string;
3992
+ baseUrl: string;
3993
+ compat: {
3994
+ supportsDeveloperRole: false;
3995
+ };
3996
+ reasoning: false;
3997
+ input: "text"[];
3998
+ cost: {
3999
+ input: number;
4000
+ output: number;
4001
+ cacheRead: number;
4002
+ cacheWrite: number;
4003
+ };
4004
+ contextWindow: number;
4005
+ maxTokens: number;
4006
+ };
3933
4007
  readonly "Qwen/Qwen3-Next-80B-A3B-Instruct": {
3934
4008
  id: string;
3935
4009
  name: string;
@@ -3970,6 +4044,26 @@ export declare const MODELS: {
3970
4044
  contextWindow: number;
3971
4045
  maxTokens: number;
3972
4046
  };
4047
+ readonly "Qwen/Qwen3.5-397B-A17B": {
4048
+ id: string;
4049
+ name: string;
4050
+ api: "openai-completions";
4051
+ provider: string;
4052
+ baseUrl: string;
4053
+ compat: {
4054
+ supportsDeveloperRole: false;
4055
+ };
4056
+ reasoning: true;
4057
+ input: ("image" | "text")[];
4058
+ cost: {
4059
+ input: number;
4060
+ output: number;
4061
+ cacheRead: number;
4062
+ cacheWrite: number;
4063
+ };
4064
+ contextWindow: number;
4065
+ maxTokens: number;
4066
+ };
3973
4067
  readonly "XiaomiMiMo/MiMo-V2-Flash": {
3974
4068
  id: string;
3975
4069
  name: string;
@@ -4259,6 +4353,23 @@ export declare const MODELS: {
4259
4353
  contextWindow: number;
4260
4354
  maxTokens: number;
4261
4355
  };
4356
+ readonly "MiniMax-M2.5-highspeed": {
4357
+ id: string;
4358
+ name: string;
4359
+ api: "anthropic-messages";
4360
+ provider: string;
4361
+ baseUrl: string;
4362
+ reasoning: true;
4363
+ input: "text"[];
4364
+ cost: {
4365
+ input: number;
4366
+ output: number;
4367
+ cacheRead: number;
4368
+ cacheWrite: number;
4369
+ };
4370
+ contextWindow: number;
4371
+ maxTokens: number;
4372
+ };
4262
4373
  };
4263
4374
  readonly "minimax-cn": {
4264
4375
  readonly "MiniMax-M2": {
@@ -4312,6 +4423,23 @@ export declare const MODELS: {
4312
4423
  contextWindow: number;
4313
4424
  maxTokens: number;
4314
4425
  };
4426
+ readonly "MiniMax-M2.5-highspeed": {
4427
+ id: string;
4428
+ name: string;
4429
+ api: "anthropic-messages";
4430
+ provider: string;
4431
+ baseUrl: string;
4432
+ reasoning: true;
4433
+ input: "text"[];
4434
+ cost: {
4435
+ input: number;
4436
+ output: number;
4437
+ cacheRead: number;
4438
+ cacheWrite: number;
4439
+ };
4440
+ contextWindow: number;
4441
+ maxTokens: number;
4442
+ };
4315
4443
  };
4316
4444
  readonly mistral: {
4317
4445
  readonly "codestral-latest": {
@@ -5680,6 +5808,40 @@ export declare const MODELS: {
5680
5808
  contextWindow: number;
5681
5809
  maxTokens: number;
5682
5810
  };
5811
+ readonly "glm-5": {
5812
+ id: string;
5813
+ name: string;
5814
+ api: "openai-completions";
5815
+ provider: string;
5816
+ baseUrl: string;
5817
+ reasoning: true;
5818
+ input: "text"[];
5819
+ cost: {
5820
+ input: number;
5821
+ output: number;
5822
+ cacheRead: number;
5823
+ cacheWrite: number;
5824
+ };
5825
+ contextWindow: number;
5826
+ maxTokens: number;
5827
+ };
5828
+ readonly "glm-5-free": {
5829
+ id: string;
5830
+ name: string;
5831
+ api: "openai-completions";
5832
+ provider: string;
5833
+ baseUrl: string;
5834
+ reasoning: true;
5835
+ input: "text"[];
5836
+ cost: {
5837
+ input: number;
5838
+ output: number;
5839
+ cacheRead: number;
5840
+ cacheWrite: number;
5841
+ };
5842
+ contextWindow: number;
5843
+ maxTokens: number;
5844
+ };
5683
5845
  readonly "gpt-5": {
5684
5846
  id: string;
5685
5847
  name: string;
@@ -5918,6 +6080,23 @@ export declare const MODELS: {
5918
6080
  contextWindow: number;
5919
6081
  maxTokens: number;
5920
6082
  };
6083
+ readonly "minimax-m2.5": {
6084
+ id: string;
6085
+ name: string;
6086
+ api: "openai-completions";
6087
+ provider: string;
6088
+ baseUrl: string;
6089
+ reasoning: true;
6090
+ input: "text"[];
6091
+ cost: {
6092
+ input: number;
6093
+ output: number;
6094
+ cacheRead: number;
6095
+ cacheWrite: number;
6096
+ };
6097
+ contextWindow: number;
6098
+ maxTokens: number;
6099
+ };
5921
6100
  readonly "minimax-m2.5-free": {
5922
6101
  id: string;
5923
6102
  name: string;
@@ -6277,6 +6456,23 @@ export declare const MODELS: {
6277
6456
  contextWindow: number;
6278
6457
  maxTokens: number;
6279
6458
  };
6459
+ readonly "anthropic/claude-sonnet-4.6": {
6460
+ id: string;
6461
+ name: string;
6462
+ api: "openai-completions";
6463
+ provider: string;
6464
+ baseUrl: string;
6465
+ reasoning: true;
6466
+ input: ("image" | "text")[];
6467
+ cost: {
6468
+ input: number;
6469
+ output: number;
6470
+ cacheRead: number;
6471
+ cacheWrite: number;
6472
+ };
6473
+ contextWindow: number;
6474
+ maxTokens: number;
6475
+ };
6280
6476
  readonly "arcee-ai/trinity-large-preview:free": {
6281
6477
  id: string;
6282
6478
  name: string;
@@ -9320,6 +9516,40 @@ export declare const MODELS: {
9320
9516
  contextWindow: number;
9321
9517
  maxTokens: number;
9322
9518
  };
9519
+ readonly "qwen/qwen3.5-397b-a17b": {
9520
+ id: string;
9521
+ name: string;
9522
+ api: "openai-completions";
9523
+ provider: string;
9524
+ baseUrl: string;
9525
+ reasoning: true;
9526
+ input: ("image" | "text")[];
9527
+ cost: {
9528
+ input: number;
9529
+ output: number;
9530
+ cacheRead: number;
9531
+ cacheWrite: number;
9532
+ };
9533
+ contextWindow: number;
9534
+ maxTokens: number;
9535
+ };
9536
+ readonly "qwen/qwen3.5-plus-02-15": {
9537
+ id: string;
9538
+ name: string;
9539
+ api: "openai-completions";
9540
+ provider: string;
9541
+ baseUrl: string;
9542
+ reasoning: true;
9543
+ input: ("image" | "text")[];
9544
+ cost: {
9545
+ input: number;
9546
+ output: number;
9547
+ cacheRead: number;
9548
+ cacheWrite: number;
9549
+ };
9550
+ contextWindow: number;
9551
+ maxTokens: number;
9552
+ };
9323
9553
  readonly "qwen/qwq-32b": {
9324
9554
  id: string;
9325
9555
  name: string;
@@ -10053,6 +10283,23 @@ export declare const MODELS: {
10053
10283
  contextWindow: number;
10054
10284
  maxTokens: number;
10055
10285
  };
10286
+ readonly "alibaba/qwen3.5-plus": {
10287
+ id: string;
10288
+ name: string;
10289
+ api: "anthropic-messages";
10290
+ provider: string;
10291
+ baseUrl: string;
10292
+ reasoning: true;
10293
+ input: ("image" | "text")[];
10294
+ cost: {
10295
+ input: number;
10296
+ output: number;
10297
+ cacheRead: number;
10298
+ cacheWrite: number;
10299
+ };
10300
+ contextWindow: number;
10301
+ maxTokens: number;
10302
+ };
10056
10303
  readonly "anthropic/claude-3-haiku": {
10057
10304
  id: string;
10058
10305
  name: string;
@@ -10257,6 +10504,23 @@ export declare const MODELS: {
10257
10504
  contextWindow: number;
10258
10505
  maxTokens: number;
10259
10506
  };
10507
+ readonly "anthropic/claude-sonnet-4.6": {
10508
+ id: string;
10509
+ name: string;
10510
+ api: "anthropic-messages";
10511
+ provider: string;
10512
+ baseUrl: string;
10513
+ reasoning: true;
10514
+ input: ("image" | "text")[];
10515
+ cost: {
10516
+ input: number;
10517
+ output: number;
10518
+ cacheRead: number;
10519
+ cacheWrite: number;
10520
+ };
10521
+ contextWindow: number;
10522
+ maxTokens: number;
10523
+ };
10260
10524
  readonly "arcee-ai/trinity-large-preview": {
10261
10525
  id: string;
10262
10526
  name: string;