@mariozechner/pi-ai 0.52.11 → 0.53.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/README.md CHANGED
@@ -1086,7 +1086,7 @@ const response = await complete(model, {
1086
1086
 
1087
1087
  ### Provider Notes
1088
1088
 
1089
- **OpenAI Codex**: Requires a ChatGPT Plus or Pro subscription. Provides access to GPT-5.x Codex models with extended context windows and reasoning capabilities. The library automatically handles session-based prompt caching when `sessionId` is provided in stream options.
1089
+ **OpenAI Codex**: Requires a ChatGPT Plus or Pro subscription. Provides access to GPT-5.x Codex models with extended context windows and reasoning capabilities. The library automatically handles session-based prompt caching when `sessionId` is provided in stream options. You can set `transport` in stream options to `"sse"`, `"websocket"`, or `"auto"` for Codex Responses transport selection. When using WebSocket with a `sessionId`, connections are reused per session and expire after 5 minutes of inactivity.
1090
1090
 
1091
1091
  **Azure OpenAI (Responses)**: Uses the Responses API only. Set `AZURE_OPENAI_API_KEY` and either `AZURE_OPENAI_BASE_URL` or `AZURE_OPENAI_RESOURCE_NAME`. Use `AZURE_OPENAI_API_VERSION` (defaults to `v1`) to override the API version if needed. Deployment names are treated as model IDs by default, override with `azureDeploymentName` or `AZURE_OPENAI_DEPLOYMENT_NAME_MAP` using comma-separated `model-id=deployment` pairs (for example `gpt-4o-mini=my-deployment,gpt-4o=prod`). Legacy deployment-based URLs are intentionally unsupported.
1092
1092
 
@@ -425,6 +425,23 @@ export declare const MODELS: {
425
425
  contextWindow: number;
426
426
  maxTokens: number;
427
427
  };
428
+ readonly "deepseek.v3.2-v1:0": {
429
+ id: string;
430
+ name: string;
431
+ api: "bedrock-converse-stream";
432
+ provider: string;
433
+ baseUrl: string;
434
+ reasoning: true;
435
+ input: "text"[];
436
+ cost: {
437
+ input: number;
438
+ output: number;
439
+ cacheRead: number;
440
+ cacheWrite: number;
441
+ };
442
+ contextWindow: number;
443
+ maxTokens: number;
444
+ };
428
445
  readonly "eu.anthropic.claude-haiku-4-5-20251001-v1:0": {
429
446
  id: string;
430
447
  name: string;
@@ -1702,6 +1719,23 @@ export declare const MODELS: {
1702
1719
  contextWindow: number;
1703
1720
  maxTokens: number;
1704
1721
  };
1722
+ readonly "claude-sonnet-4-6": {
1723
+ id: string;
1724
+ name: string;
1725
+ api: "anthropic-messages";
1726
+ provider: string;
1727
+ baseUrl: string;
1728
+ reasoning: true;
1729
+ input: ("image" | "text")[];
1730
+ cost: {
1731
+ input: number;
1732
+ output: number;
1733
+ cacheRead: number;
1734
+ cacheWrite: number;
1735
+ };
1736
+ contextWindow: number;
1737
+ maxTokens: number;
1738
+ };
1705
1739
  };
1706
1740
  readonly "azure-openai-responses": {
1707
1741
  readonly "codex-mini-latest": {
@@ -3890,6 +3924,26 @@ export declare const MODELS: {
3890
3924
  contextWindow: number;
3891
3925
  maxTokens: number;
3892
3926
  };
3927
+ readonly "MiniMaxAI/MiniMax-M2.5": {
3928
+ id: string;
3929
+ name: string;
3930
+ api: "openai-completions";
3931
+ provider: string;
3932
+ baseUrl: string;
3933
+ compat: {
3934
+ supportsDeveloperRole: false;
3935
+ };
3936
+ reasoning: true;
3937
+ input: "text"[];
3938
+ cost: {
3939
+ input: number;
3940
+ output: number;
3941
+ cacheRead: number;
3942
+ cacheWrite: number;
3943
+ };
3944
+ contextWindow: number;
3945
+ maxTokens: number;
3946
+ };
3893
3947
  readonly "Qwen/Qwen3-235B-A22B-Thinking-2507": {
3894
3948
  id: string;
3895
3949
  name: string;
@@ -3930,6 +3984,26 @@ export declare const MODELS: {
3930
3984
  contextWindow: number;
3931
3985
  maxTokens: number;
3932
3986
  };
3987
+ readonly "Qwen/Qwen3-Coder-Next": {
3988
+ id: string;
3989
+ name: string;
3990
+ api: "openai-completions";
3991
+ provider: string;
3992
+ baseUrl: string;
3993
+ compat: {
3994
+ supportsDeveloperRole: false;
3995
+ };
3996
+ reasoning: false;
3997
+ input: "text"[];
3998
+ cost: {
3999
+ input: number;
4000
+ output: number;
4001
+ cacheRead: number;
4002
+ cacheWrite: number;
4003
+ };
4004
+ contextWindow: number;
4005
+ maxTokens: number;
4006
+ };
3933
4007
  readonly "Qwen/Qwen3-Next-80B-A3B-Instruct": {
3934
4008
  id: string;
3935
4009
  name: string;
@@ -3970,6 +4044,26 @@ export declare const MODELS: {
3970
4044
  contextWindow: number;
3971
4045
  maxTokens: number;
3972
4046
  };
4047
+ readonly "Qwen/Qwen3.5-397B-A17B": {
4048
+ id: string;
4049
+ name: string;
4050
+ api: "openai-completions";
4051
+ provider: string;
4052
+ baseUrl: string;
4053
+ compat: {
4054
+ supportsDeveloperRole: false;
4055
+ };
4056
+ reasoning: true;
4057
+ input: ("image" | "text")[];
4058
+ cost: {
4059
+ input: number;
4060
+ output: number;
4061
+ cacheRead: number;
4062
+ cacheWrite: number;
4063
+ };
4064
+ contextWindow: number;
4065
+ maxTokens: number;
4066
+ };
3973
4067
  readonly "XiaomiMiMo/MiMo-V2-Flash": {
3974
4068
  id: string;
3975
4069
  name: string;
@@ -4259,6 +4353,23 @@ export declare const MODELS: {
4259
4353
  contextWindow: number;
4260
4354
  maxTokens: number;
4261
4355
  };
4356
+ readonly "MiniMax-M2.5-highspeed": {
4357
+ id: string;
4358
+ name: string;
4359
+ api: "anthropic-messages";
4360
+ provider: string;
4361
+ baseUrl: string;
4362
+ reasoning: true;
4363
+ input: "text"[];
4364
+ cost: {
4365
+ input: number;
4366
+ output: number;
4367
+ cacheRead: number;
4368
+ cacheWrite: number;
4369
+ };
4370
+ contextWindow: number;
4371
+ maxTokens: number;
4372
+ };
4262
4373
  };
4263
4374
  readonly "minimax-cn": {
4264
4375
  readonly "MiniMax-M2": {
@@ -4312,6 +4423,23 @@ export declare const MODELS: {
4312
4423
  contextWindow: number;
4313
4424
  maxTokens: number;
4314
4425
  };
4426
+ readonly "MiniMax-M2.5-highspeed": {
4427
+ id: string;
4428
+ name: string;
4429
+ api: "anthropic-messages";
4430
+ provider: string;
4431
+ baseUrl: string;
4432
+ reasoning: true;
4433
+ input: "text"[];
4434
+ cost: {
4435
+ input: number;
4436
+ output: number;
4437
+ cacheRead: number;
4438
+ cacheWrite: number;
4439
+ };
4440
+ contextWindow: number;
4441
+ maxTokens: number;
4442
+ };
4315
4443
  };
4316
4444
  readonly mistral: {
4317
4445
  readonly "codestral-latest": {
@@ -5680,6 +5808,40 @@ export declare const MODELS: {
5680
5808
  contextWindow: number;
5681
5809
  maxTokens: number;
5682
5810
  };
5811
+ readonly "glm-5": {
5812
+ id: string;
5813
+ name: string;
5814
+ api: "openai-completions";
5815
+ provider: string;
5816
+ baseUrl: string;
5817
+ reasoning: true;
5818
+ input: "text"[];
5819
+ cost: {
5820
+ input: number;
5821
+ output: number;
5822
+ cacheRead: number;
5823
+ cacheWrite: number;
5824
+ };
5825
+ contextWindow: number;
5826
+ maxTokens: number;
5827
+ };
5828
+ readonly "glm-5-free": {
5829
+ id: string;
5830
+ name: string;
5831
+ api: "openai-completions";
5832
+ provider: string;
5833
+ baseUrl: string;
5834
+ reasoning: true;
5835
+ input: "text"[];
5836
+ cost: {
5837
+ input: number;
5838
+ output: number;
5839
+ cacheRead: number;
5840
+ cacheWrite: number;
5841
+ };
5842
+ contextWindow: number;
5843
+ maxTokens: number;
5844
+ };
5683
5845
  readonly "gpt-5": {
5684
5846
  id: string;
5685
5847
  name: string;
@@ -5918,6 +6080,23 @@ export declare const MODELS: {
5918
6080
  contextWindow: number;
5919
6081
  maxTokens: number;
5920
6082
  };
6083
+ readonly "minimax-m2.5": {
6084
+ id: string;
6085
+ name: string;
6086
+ api: "openai-completions";
6087
+ provider: string;
6088
+ baseUrl: string;
6089
+ reasoning: true;
6090
+ input: "text"[];
6091
+ cost: {
6092
+ input: number;
6093
+ output: number;
6094
+ cacheRead: number;
6095
+ cacheWrite: number;
6096
+ };
6097
+ contextWindow: number;
6098
+ maxTokens: number;
6099
+ };
5921
6100
  readonly "minimax-m2.5-free": {
5922
6101
  id: string;
5923
6102
  name: string;
@@ -6277,6 +6456,23 @@ export declare const MODELS: {
6277
6456
  contextWindow: number;
6278
6457
  maxTokens: number;
6279
6458
  };
6459
+ readonly "anthropic/claude-sonnet-4.6": {
6460
+ id: string;
6461
+ name: string;
6462
+ api: "openai-completions";
6463
+ provider: string;
6464
+ baseUrl: string;
6465
+ reasoning: true;
6466
+ input: ("image" | "text")[];
6467
+ cost: {
6468
+ input: number;
6469
+ output: number;
6470
+ cacheRead: number;
6471
+ cacheWrite: number;
6472
+ };
6473
+ contextWindow: number;
6474
+ maxTokens: number;
6475
+ };
6280
6476
  readonly "arcee-ai/trinity-large-preview:free": {
6281
6477
  id: string;
6282
6478
  name: string;
@@ -9320,14 +9516,14 @@ export declare const MODELS: {
9320
9516
  contextWindow: number;
9321
9517
  maxTokens: number;
9322
9518
  };
9323
- readonly "qwen/qwq-32b": {
9519
+ readonly "qwen/qwen3.5-397b-a17b": {
9324
9520
  id: string;
9325
9521
  name: string;
9326
9522
  api: "openai-completions";
9327
9523
  provider: string;
9328
9524
  baseUrl: string;
9329
9525
  reasoning: true;
9330
- input: "text"[];
9526
+ input: ("image" | "text")[];
9331
9527
  cost: {
9332
9528
  input: number;
9333
9529
  output: number;
@@ -9337,13 +9533,30 @@ export declare const MODELS: {
9337
9533
  contextWindow: number;
9338
9534
  maxTokens: number;
9339
9535
  };
9340
- readonly "relace/relace-search": {
9536
+ readonly "qwen/qwen3.5-plus-02-15": {
9341
9537
  id: string;
9342
9538
  name: string;
9343
9539
  api: "openai-completions";
9344
9540
  provider: string;
9345
9541
  baseUrl: string;
9346
- reasoning: false;
9542
+ reasoning: true;
9543
+ input: ("image" | "text")[];
9544
+ cost: {
9545
+ input: number;
9546
+ output: number;
9547
+ cacheRead: number;
9548
+ cacheWrite: number;
9549
+ };
9550
+ contextWindow: number;
9551
+ maxTokens: number;
9552
+ };
9553
+ readonly "qwen/qwq-32b": {
9554
+ id: string;
9555
+ name: string;
9556
+ api: "openai-completions";
9557
+ provider: string;
9558
+ baseUrl: string;
9559
+ reasoning: true;
9347
9560
  input: "text"[];
9348
9561
  cost: {
9349
9562
  input: number;
@@ -9354,7 +9567,7 @@ export declare const MODELS: {
9354
9567
  contextWindow: number;
9355
9568
  maxTokens: number;
9356
9569
  };
9357
- readonly "sao10k/l3-euryale-70b": {
9570
+ readonly "relace/relace-search": {
9358
9571
  id: string;
9359
9572
  name: string;
9360
9573
  api: "openai-completions";
@@ -9371,7 +9584,7 @@ export declare const MODELS: {
9371
9584
  contextWindow: number;
9372
9585
  maxTokens: number;
9373
9586
  };
9374
- readonly "sao10k/l3.1-euryale-70b": {
9587
+ readonly "sao10k/l3-euryale-70b": {
9375
9588
  id: string;
9376
9589
  name: string;
9377
9590
  api: "openai-completions";
@@ -9388,13 +9601,13 @@ export declare const MODELS: {
9388
9601
  contextWindow: number;
9389
9602
  maxTokens: number;
9390
9603
  };
9391
- readonly "stepfun/step-3.5-flash": {
9604
+ readonly "sao10k/l3.1-euryale-70b": {
9392
9605
  id: string;
9393
9606
  name: string;
9394
9607
  api: "openai-completions";
9395
9608
  provider: string;
9396
9609
  baseUrl: string;
9397
- reasoning: true;
9610
+ reasoning: false;
9398
9611
  input: "text"[];
9399
9612
  cost: {
9400
9613
  input: number;
@@ -9405,7 +9618,7 @@ export declare const MODELS: {
9405
9618
  contextWindow: number;
9406
9619
  maxTokens: number;
9407
9620
  };
9408
- readonly "stepfun/step-3.5-flash:free": {
9621
+ readonly "stepfun/step-3.5-flash": {
9409
9622
  id: string;
9410
9623
  name: string;
9411
9624
  api: "openai-completions";
@@ -9422,13 +9635,13 @@ export declare const MODELS: {
9422
9635
  contextWindow: number;
9423
9636
  maxTokens: number;
9424
9637
  };
9425
- readonly "thedrummer/rocinante-12b": {
9638
+ readonly "stepfun/step-3.5-flash:free": {
9426
9639
  id: string;
9427
9640
  name: string;
9428
9641
  api: "openai-completions";
9429
9642
  provider: string;
9430
9643
  baseUrl: string;
9431
- reasoning: false;
9644
+ reasoning: true;
9432
9645
  input: "text"[];
9433
9646
  cost: {
9434
9647
  input: number;
@@ -9439,7 +9652,7 @@ export declare const MODELS: {
9439
9652
  contextWindow: number;
9440
9653
  maxTokens: number;
9441
9654
  };
9442
- readonly "thedrummer/unslopnemo-12b": {
9655
+ readonly "thedrummer/rocinante-12b": {
9443
9656
  id: string;
9444
9657
  name: string;
9445
9658
  api: "openai-completions";
@@ -9456,13 +9669,13 @@ export declare const MODELS: {
9456
9669
  contextWindow: number;
9457
9670
  maxTokens: number;
9458
9671
  };
9459
- readonly "tngtech/deepseek-r1t2-chimera": {
9672
+ readonly "thedrummer/unslopnemo-12b": {
9460
9673
  id: string;
9461
9674
  name: string;
9462
9675
  api: "openai-completions";
9463
9676
  provider: string;
9464
9677
  baseUrl: string;
9465
- reasoning: true;
9678
+ reasoning: false;
9466
9679
  input: "text"[];
9467
9680
  cost: {
9468
9681
  input: number;
@@ -9473,7 +9686,7 @@ export declare const MODELS: {
9473
9686
  contextWindow: number;
9474
9687
  maxTokens: number;
9475
9688
  };
9476
- readonly "tngtech/tng-r1t-chimera": {
9689
+ readonly "tngtech/deepseek-r1t2-chimera": {
9477
9690
  id: string;
9478
9691
  name: string;
9479
9692
  api: "openai-completions";
@@ -9490,7 +9703,7 @@ export declare const MODELS: {
9490
9703
  contextWindow: number;
9491
9704
  maxTokens: number;
9492
9705
  };
9493
- readonly "tngtech/tng-r1t-chimera:free": {
9706
+ readonly "tngtech/tng-r1t-chimera": {
9494
9707
  id: string;
9495
9708
  name: string;
9496
9709
  api: "openai-completions";
@@ -10070,6 +10283,23 @@ export declare const MODELS: {
10070
10283
  contextWindow: number;
10071
10284
  maxTokens: number;
10072
10285
  };
10286
+ readonly "alibaba/qwen3.5-plus": {
10287
+ id: string;
10288
+ name: string;
10289
+ api: "anthropic-messages";
10290
+ provider: string;
10291
+ baseUrl: string;
10292
+ reasoning: true;
10293
+ input: ("image" | "text")[];
10294
+ cost: {
10295
+ input: number;
10296
+ output: number;
10297
+ cacheRead: number;
10298
+ cacheWrite: number;
10299
+ };
10300
+ contextWindow: number;
10301
+ maxTokens: number;
10302
+ };
10073
10303
  readonly "anthropic/claude-3-haiku": {
10074
10304
  id: string;
10075
10305
  name: string;
@@ -10274,6 +10504,23 @@ export declare const MODELS: {
10274
10504
  contextWindow: number;
10275
10505
  maxTokens: number;
10276
10506
  };
10507
+ readonly "anthropic/claude-sonnet-4.6": {
10508
+ id: string;
10509
+ name: string;
10510
+ api: "anthropic-messages";
10511
+ provider: string;
10512
+ baseUrl: string;
10513
+ reasoning: true;
10514
+ input: ("image" | "text")[];
10515
+ cost: {
10516
+ input: number;
10517
+ output: number;
10518
+ cacheRead: number;
10519
+ cacheWrite: number;
10520
+ };
10521
+ contextWindow: number;
10522
+ maxTokens: number;
10523
+ };
10277
10524
  readonly "arcee-ai/trinity-large-preview": {
10278
10525
  id: string;
10279
10526
  name: string;