@mariozechner/pi-ai 0.68.1 → 0.70.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (48) hide show
  1. package/README.md +1 -1
  2. package/dist/env-api-keys.d.ts +9 -0
  3. package/dist/env-api-keys.d.ts.map +1 -1
  4. package/dist/env-api-keys.js +41 -31
  5. package/dist/env-api-keys.js.map +1 -1
  6. package/dist/index.d.ts +2 -2
  7. package/dist/index.d.ts.map +1 -1
  8. package/dist/index.js +1 -1
  9. package/dist/index.js.map +1 -1
  10. package/dist/models.d.ts +1 -1
  11. package/dist/models.d.ts.map +1 -1
  12. package/dist/models.generated.d.ts +212 -19
  13. package/dist/models.generated.d.ts.map +1 -1
  14. package/dist/models.generated.js +238 -49
  15. package/dist/models.generated.js.map +1 -1
  16. package/dist/models.js +5 -2
  17. package/dist/models.js.map +1 -1
  18. package/dist/providers/amazon-bedrock.d.ts.map +1 -1
  19. package/dist/providers/amazon-bedrock.js.map +1 -1
  20. package/dist/providers/anthropic.d.ts.map +1 -1
  21. package/dist/providers/anthropic.js +28 -11
  22. package/dist/providers/anthropic.js.map +1 -1
  23. package/dist/providers/google-vertex.d.ts.map +1 -1
  24. package/dist/providers/google-vertex.js +34 -13
  25. package/dist/providers/google-vertex.js.map +1 -1
  26. package/dist/providers/openai-codex-responses.d.ts.map +1 -1
  27. package/dist/providers/openai-codex-responses.js +8 -7
  28. package/dist/providers/openai-codex-responses.js.map +1 -1
  29. package/dist/providers/openai-completions.d.ts.map +1 -1
  30. package/dist/providers/openai-completions.js +53 -31
  31. package/dist/providers/openai-completions.js.map +1 -1
  32. package/dist/providers/openai-responses.d.ts.map +1 -1
  33. package/dist/providers/openai-responses.js +19 -19
  34. package/dist/providers/openai-responses.js.map +1 -1
  35. package/dist/providers/transform-messages.d.ts.map +1 -1
  36. package/dist/providers/transform-messages.js +2 -0
  37. package/dist/providers/transform-messages.js.map +1 -1
  38. package/dist/types.d.ts +21 -2
  39. package/dist/types.d.ts.map +1 -1
  40. package/dist/types.js.map +1 -1
  41. package/dist/utils/typebox-helpers.d.ts +1 -1
  42. package/dist/utils/typebox-helpers.d.ts.map +1 -1
  43. package/dist/utils/typebox-helpers.js +1 -1
  44. package/dist/utils/typebox-helpers.js.map +1 -1
  45. package/dist/utils/validation.d.ts.map +1 -1
  46. package/dist/utils/validation.js +242 -41
  47. package/dist/utils/validation.js.map +1 -1
  48. package/package.json +2 -4
@@ -2486,6 +2486,23 @@ export declare const MODELS: {
2486
2486
  contextWindow: number;
2487
2487
  maxTokens: number;
2488
2488
  };
2489
+ readonly "gpt-5.5": {
2490
+ id: string;
2491
+ name: string;
2492
+ api: "azure-openai-responses";
2493
+ provider: string;
2494
+ baseUrl: string;
2495
+ reasoning: true;
2496
+ input: ("image" | "text")[];
2497
+ cost: {
2498
+ input: number;
2499
+ output: number;
2500
+ cacheRead: number;
2501
+ cacheWrite: number;
2502
+ };
2503
+ contextWindow: number;
2504
+ maxTokens: number;
2505
+ };
2489
2506
  readonly o1: {
2490
2507
  id: string;
2491
2508
  name: string;
@@ -3014,6 +3031,9 @@ export declare const MODELS: {
3014
3031
  "Editor-Plugin-Version": string;
3015
3032
  "Copilot-Integration-Id": string;
3016
3033
  };
3034
+ compat: {
3035
+ supportsEagerToolInputStreaming: false;
3036
+ };
3017
3037
  reasoning: true;
3018
3038
  input: ("image" | "text")[];
3019
3039
  cost: {
@@ -3106,6 +3126,9 @@ export declare const MODELS: {
3106
3126
  "Editor-Plugin-Version": string;
3107
3127
  "Copilot-Integration-Id": string;
3108
3128
  };
3129
+ compat: {
3130
+ supportsEagerToolInputStreaming: false;
3131
+ };
3109
3132
  reasoning: true;
3110
3133
  input: ("image" | "text")[];
3111
3134
  cost: {
@@ -3129,6 +3152,9 @@ export declare const MODELS: {
3129
3152
  "Editor-Plugin-Version": string;
3130
3153
  "Copilot-Integration-Id": string;
3131
3154
  };
3155
+ compat: {
3156
+ supportsEagerToolInputStreaming: false;
3157
+ };
3132
3158
  reasoning: true;
3133
3159
  input: ("image" | "text")[];
3134
3160
  cost: {
@@ -4039,7 +4065,7 @@ export declare const MODELS: {
4039
4065
  contextWindow: number;
4040
4066
  maxTokens: number;
4041
4067
  };
4042
- readonly "gemma-4-26b-it": {
4068
+ readonly "gemma-4-26b-a4b-it": {
4043
4069
  id: string;
4044
4070
  name: string;
4045
4071
  api: "google-generative-ai";
@@ -4315,6 +4341,23 @@ export declare const MODELS: {
4315
4341
  contextWindow: number;
4316
4342
  maxTokens: number;
4317
4343
  };
4344
+ readonly "gemini-3.1-flash-lite-preview": {
4345
+ id: string;
4346
+ name: string;
4347
+ api: "google-gemini-cli";
4348
+ provider: string;
4349
+ baseUrl: string;
4350
+ reasoning: true;
4351
+ input: ("image" | "text")[];
4352
+ cost: {
4353
+ input: number;
4354
+ output: number;
4355
+ cacheRead: number;
4356
+ cacheWrite: number;
4357
+ };
4358
+ contextWindow: number;
4359
+ maxTokens: number;
4360
+ };
4318
4361
  readonly "gemini-3.1-pro-preview": {
4319
4362
  id: string;
4320
4363
  name: string;
@@ -5293,6 +5336,9 @@ export declare const MODELS: {
5293
5336
  api: "anthropic-messages";
5294
5337
  provider: string;
5295
5338
  baseUrl: string;
5339
+ headers: {
5340
+ "User-Agent": string;
5341
+ };
5296
5342
  reasoning: true;
5297
5343
  input: ("image" | "text")[];
5298
5344
  cost: {
@@ -5310,6 +5356,9 @@ export declare const MODELS: {
5310
5356
  api: "anthropic-messages";
5311
5357
  provider: string;
5312
5358
  baseUrl: string;
5359
+ headers: {
5360
+ "User-Agent": string;
5361
+ };
5313
5362
  reasoning: true;
5314
5363
  input: ("image" | "text")[];
5315
5364
  cost: {
@@ -5327,6 +5376,9 @@ export declare const MODELS: {
5327
5376
  api: "anthropic-messages";
5328
5377
  provider: string;
5329
5378
  baseUrl: string;
5379
+ headers: {
5380
+ "User-Agent": string;
5381
+ };
5330
5382
  reasoning: true;
5331
5383
  input: "text"[];
5332
5384
  cost: {
@@ -6400,6 +6452,23 @@ export declare const MODELS: {
6400
6452
  contextWindow: number;
6401
6453
  maxTokens: number;
6402
6454
  };
6455
+ readonly "gpt-5.5": {
6456
+ id: string;
6457
+ name: string;
6458
+ api: "openai-responses";
6459
+ provider: string;
6460
+ baseUrl: string;
6461
+ reasoning: true;
6462
+ input: ("image" | "text")[];
6463
+ cost: {
6464
+ input: number;
6465
+ output: number;
6466
+ cacheRead: number;
6467
+ cacheWrite: number;
6468
+ };
6469
+ contextWindow: number;
6470
+ maxTokens: number;
6471
+ };
6403
6472
  readonly o1: {
6404
6473
  id: string;
6405
6474
  name: string;
@@ -6691,6 +6760,23 @@ export declare const MODELS: {
6691
6760
  contextWindow: number;
6692
6761
  maxTokens: number;
6693
6762
  };
6763
+ readonly "gpt-5.5": {
6764
+ id: string;
6765
+ name: string;
6766
+ api: "openai-codex-responses";
6767
+ provider: string;
6768
+ baseUrl: string;
6769
+ reasoning: true;
6770
+ input: ("image" | "text")[];
6771
+ cost: {
6772
+ input: number;
6773
+ output: number;
6774
+ cacheRead: number;
6775
+ cacheWrite: number;
6776
+ };
6777
+ contextWindow: number;
6778
+ maxTokens: number;
6779
+ };
6694
6780
  };
6695
6781
  readonly opencode: {
6696
6782
  readonly "big-pickle": {
@@ -7169,6 +7255,23 @@ export declare const MODELS: {
7169
7255
  contextWindow: number;
7170
7256
  maxTokens: number;
7171
7257
  };
7258
+ readonly "hy3-preview-free": {
7259
+ id: string;
7260
+ name: string;
7261
+ api: "openai-completions";
7262
+ provider: string;
7263
+ baseUrl: string;
7264
+ reasoning: true;
7265
+ input: "text"[];
7266
+ cost: {
7267
+ input: number;
7268
+ output: number;
7269
+ cacheRead: number;
7270
+ cacheWrite: number;
7271
+ };
7272
+ contextWindow: number;
7273
+ maxTokens: number;
7274
+ };
7172
7275
  readonly "kimi-k2.5": {
7173
7276
  id: string;
7174
7277
  name: string;
@@ -7291,12 +7394,9 @@ export declare const MODELS: {
7291
7394
  readonly "qwen3.5-plus": {
7292
7395
  id: string;
7293
7396
  name: string;
7294
- api: "openai-completions";
7397
+ api: "anthropic-messages";
7295
7398
  provider: string;
7296
7399
  baseUrl: string;
7297
- compat: {
7298
- cacheControlFormat: "anthropic";
7299
- };
7300
7400
  reasoning: true;
7301
7401
  input: ("image" | "text")[];
7302
7402
  cost: {
@@ -7311,12 +7411,9 @@ export declare const MODELS: {
7311
7411
  readonly "qwen3.6-plus": {
7312
7412
  id: string;
7313
7413
  name: string;
7314
- api: "openai-completions";
7414
+ api: "anthropic-messages";
7315
7415
  provider: string;
7316
7416
  baseUrl: string;
7317
- compat: {
7318
- cacheControlFormat: "anthropic";
7319
- };
7320
7417
  reasoning: true;
7321
7418
  input: ("image" | "text")[];
7322
7419
  cost: {
@@ -7432,6 +7529,40 @@ export declare const MODELS: {
7432
7529
  contextWindow: number;
7433
7530
  maxTokens: number;
7434
7531
  };
7532
+ readonly "mimo-v2.5": {
7533
+ id: string;
7534
+ name: string;
7535
+ api: "openai-completions";
7536
+ provider: string;
7537
+ baseUrl: string;
7538
+ reasoning: true;
7539
+ input: ("image" | "text")[];
7540
+ cost: {
7541
+ input: number;
7542
+ output: number;
7543
+ cacheRead: number;
7544
+ cacheWrite: number;
7545
+ };
7546
+ contextWindow: number;
7547
+ maxTokens: number;
7548
+ };
7549
+ readonly "mimo-v2.5-pro": {
7550
+ id: string;
7551
+ name: string;
7552
+ api: "openai-completions";
7553
+ provider: string;
7554
+ baseUrl: string;
7555
+ reasoning: true;
7556
+ input: "text"[];
7557
+ cost: {
7558
+ input: number;
7559
+ output: number;
7560
+ cacheRead: number;
7561
+ cacheWrite: number;
7562
+ };
7563
+ contextWindow: number;
7564
+ maxTokens: number;
7565
+ };
7435
7566
  readonly "minimax-m2.5": {
7436
7567
  id: string;
7437
7568
  name: string;
@@ -7469,12 +7600,9 @@ export declare const MODELS: {
7469
7600
  readonly "qwen3.5-plus": {
7470
7601
  id: string;
7471
7602
  name: string;
7472
- api: "openai-completions";
7603
+ api: "anthropic-messages";
7473
7604
  provider: string;
7474
7605
  baseUrl: string;
7475
- compat: {
7476
- cacheControlFormat: "anthropic";
7477
- };
7478
7606
  reasoning: true;
7479
7607
  input: ("image" | "text")[];
7480
7608
  cost: {
@@ -7489,12 +7617,9 @@ export declare const MODELS: {
7489
7617
  readonly "qwen3.6-plus": {
7490
7618
  id: string;
7491
7619
  name: string;
7492
- api: "openai-completions";
7620
+ api: "anthropic-messages";
7493
7621
  provider: string;
7494
7622
  baseUrl: string;
7495
- compat: {
7496
- cacheControlFormat: "anthropic";
7497
- };
7498
7623
  reasoning: true;
7499
7624
  input: ("image" | "text")[];
7500
7625
  cost: {
@@ -7882,7 +8007,7 @@ export declare const MODELS: {
7882
8007
  contextWindow: number;
7883
8008
  maxTokens: number;
7884
8009
  };
7885
- readonly "arcee-ai/trinity-large-preview:free": {
8010
+ readonly "arcee-ai/trinity-large-preview": {
7886
8011
  id: string;
7887
8012
  name: string;
7888
8013
  api: "openai-completions";
@@ -8545,6 +8670,23 @@ export declare const MODELS: {
8545
8670
  contextWindow: number;
8546
8671
  maxTokens: number;
8547
8672
  };
8673
+ readonly "inclusionai/ling-2.6-1t:free": {
8674
+ id: string;
8675
+ name: string;
8676
+ api: "openai-completions";
8677
+ provider: string;
8678
+ baseUrl: string;
8679
+ reasoning: false;
8680
+ input: "text"[];
8681
+ cost: {
8682
+ input: number;
8683
+ output: number;
8684
+ cacheRead: number;
8685
+ cacheWrite: number;
8686
+ };
8687
+ contextWindow: number;
8688
+ maxTokens: number;
8689
+ };
8548
8690
  readonly "inclusionai/ling-2.6-flash:free": {
8549
8691
  id: string;
8550
8692
  name: string;
@@ -11214,6 +11356,23 @@ export declare const MODELS: {
11214
11356
  contextWindow: number;
11215
11357
  maxTokens: number;
11216
11358
  };
11359
+ readonly "tencent/hy3-preview:free": {
11360
+ id: string;
11361
+ name: string;
11362
+ api: "openai-completions";
11363
+ provider: string;
11364
+ baseUrl: string;
11365
+ reasoning: true;
11366
+ input: "text"[];
11367
+ cost: {
11368
+ input: number;
11369
+ output: number;
11370
+ cacheRead: number;
11371
+ cacheWrite: number;
11372
+ };
11373
+ contextWindow: number;
11374
+ maxTokens: number;
11375
+ };
11217
11376
  readonly "thedrummer/rocinante-12b": {
11218
11377
  id: string;
11219
11378
  name: string;
@@ -11486,6 +11645,40 @@ export declare const MODELS: {
11486
11645
  contextWindow: number;
11487
11646
  maxTokens: number;
11488
11647
  };
11648
+ readonly "xiaomi/mimo-v2.5": {
11649
+ id: string;
11650
+ name: string;
11651
+ api: "openai-completions";
11652
+ provider: string;
11653
+ baseUrl: string;
11654
+ reasoning: true;
11655
+ input: ("image" | "text")[];
11656
+ cost: {
11657
+ input: number;
11658
+ output: number;
11659
+ cacheRead: number;
11660
+ cacheWrite: number;
11661
+ };
11662
+ contextWindow: number;
11663
+ maxTokens: number;
11664
+ };
11665
+ readonly "xiaomi/mimo-v2.5-pro": {
11666
+ id: string;
11667
+ name: string;
11668
+ api: "openai-completions";
11669
+ provider: string;
11670
+ baseUrl: string;
11671
+ reasoning: true;
11672
+ input: "text"[];
11673
+ cost: {
11674
+ input: number;
11675
+ output: number;
11676
+ cacheRead: number;
11677
+ cacheWrite: number;
11678
+ };
11679
+ contextWindow: number;
11680
+ maxTokens: number;
11681
+ };
11489
11682
  readonly "z-ai/glm-4-32b": {
11490
11683
  id: string;
11491
11684
  name: string;
@@ -14351,7 +14544,7 @@ export declare const MODELS: {
14351
14544
  provider: string;
14352
14545
  baseUrl: string;
14353
14546
  reasoning: true;
14354
- input: ("image" | "text")[];
14547
+ input: "text"[];
14355
14548
  cost: {
14356
14549
  input: number;
14357
14550
  output: number;