@mariozechner/pi-ai 0.56.1 → 0.56.3

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (49) hide show
  1. package/README.md +4 -3
  2. package/dist/index.d.ts +1 -0
  3. package/dist/index.d.ts.map +1 -1
  4. package/dist/index.js +1 -0
  5. package/dist/index.js.map +1 -1
  6. package/dist/models.d.ts +1 -1
  7. package/dist/models.d.ts.map +1 -1
  8. package/dist/models.generated.d.ts +285 -35
  9. package/dist/models.generated.d.ts.map +1 -1
  10. package/dist/models.generated.js +327 -87
  11. package/dist/models.generated.js.map +1 -1
  12. package/dist/models.js +2 -2
  13. package/dist/models.js.map +1 -1
  14. package/dist/providers/google-gemini-cli.d.ts.map +1 -1
  15. package/dist/providers/google-gemini-cli.js +25 -15
  16. package/dist/providers/google-gemini-cli.js.map +1 -1
  17. package/dist/providers/google-shared.d.ts.map +1 -1
  18. package/dist/providers/google-shared.js +16 -22
  19. package/dist/providers/google-shared.js.map +1 -1
  20. package/dist/providers/mistral.d.ts +22 -0
  21. package/dist/providers/mistral.d.ts.map +1 -0
  22. package/dist/providers/mistral.js +495 -0
  23. package/dist/providers/mistral.js.map +1 -0
  24. package/dist/providers/openai-codex-responses.d.ts.map +1 -1
  25. package/dist/providers/openai-codex-responses.js +1 -1
  26. package/dist/providers/openai-codex-responses.js.map +1 -1
  27. package/dist/providers/openai-completions.d.ts.map +1 -1
  28. package/dist/providers/openai-completions.js +8 -33
  29. package/dist/providers/openai-completions.js.map +1 -1
  30. package/dist/providers/openai-responses-shared.d.ts.map +1 -1
  31. package/dist/providers/openai-responses-shared.js +28 -13
  32. package/dist/providers/openai-responses-shared.js.map +1 -1
  33. package/dist/providers/register-builtins.d.ts.map +1 -1
  34. package/dist/providers/register-builtins.js +6 -0
  35. package/dist/providers/register-builtins.js.map +1 -1
  36. package/dist/types.d.ts +6 -3
  37. package/dist/types.d.ts.map +1 -1
  38. package/dist/types.js.map +1 -1
  39. package/dist/utils/hash.d.ts +3 -0
  40. package/dist/utils/hash.d.ts.map +1 -0
  41. package/dist/utils/hash.js +14 -0
  42. package/dist/utils/hash.js.map +1 -0
  43. package/dist/utils/overflow.d.ts +1 -1
  44. package/dist/utils/overflow.d.ts.map +1 -1
  45. package/dist/utils/overflow.js +4 -3
  46. package/dist/utils/overflow.js.map +1 -1
  47. package/package.json +5 -7
  48. package/oauth.d.ts +0 -1
  49. package/oauth.js +0 -1
@@ -2282,6 +2282,40 @@ export declare const MODELS: {
2282
2282
  contextWindow: number;
2283
2283
  maxTokens: number;
2284
2284
  };
2285
+ readonly "gpt-5.4": {
2286
+ id: string;
2287
+ name: string;
2288
+ api: "azure-openai-responses";
2289
+ provider: string;
2290
+ baseUrl: string;
2291
+ reasoning: true;
2292
+ input: ("image" | "text")[];
2293
+ cost: {
2294
+ input: number;
2295
+ output: number;
2296
+ cacheRead: number;
2297
+ cacheWrite: number;
2298
+ };
2299
+ contextWindow: number;
2300
+ maxTokens: number;
2301
+ };
2302
+ readonly "gpt-5.4-pro": {
2303
+ id: string;
2304
+ name: string;
2305
+ api: "azure-openai-responses";
2306
+ provider: string;
2307
+ baseUrl: string;
2308
+ reasoning: true;
2309
+ input: ("image" | "text")[];
2310
+ cost: {
2311
+ input: number;
2312
+ output: number;
2313
+ cacheRead: number;
2314
+ cacheWrite: number;
2315
+ };
2316
+ contextWindow: number;
2317
+ maxTokens: number;
2318
+ };
2285
2319
  readonly o1: {
2286
2320
  id: string;
2287
2321
  name: string;
@@ -2980,6 +3014,52 @@ export declare const MODELS: {
2980
3014
  contextWindow: number;
2981
3015
  maxTokens: number;
2982
3016
  };
3017
+ readonly "gpt-5.3-codex": {
3018
+ id: string;
3019
+ name: string;
3020
+ api: "openai-responses";
3021
+ provider: string;
3022
+ baseUrl: string;
3023
+ headers: {
3024
+ "User-Agent": string;
3025
+ "Editor-Version": string;
3026
+ "Editor-Plugin-Version": string;
3027
+ "Copilot-Integration-Id": string;
3028
+ };
3029
+ reasoning: true;
3030
+ input: ("image" | "text")[];
3031
+ cost: {
3032
+ input: number;
3033
+ output: number;
3034
+ cacheRead: number;
3035
+ cacheWrite: number;
3036
+ };
3037
+ contextWindow: number;
3038
+ maxTokens: number;
3039
+ };
3040
+ readonly "gpt-5.4": {
3041
+ id: string;
3042
+ name: string;
3043
+ api: "openai-responses";
3044
+ provider: string;
3045
+ baseUrl: string;
3046
+ headers: {
3047
+ "User-Agent": string;
3048
+ "Editor-Version": string;
3049
+ "Editor-Plugin-Version": string;
3050
+ "Copilot-Integration-Id": string;
3051
+ };
3052
+ reasoning: true;
3053
+ input: ("image" | "text")[];
3054
+ cost: {
3055
+ input: number;
3056
+ output: number;
3057
+ cacheRead: number;
3058
+ cacheWrite: number;
3059
+ };
3060
+ contextWindow: number;
3061
+ maxTokens: number;
3062
+ };
2983
3063
  readonly "grok-code-fast-1": {
2984
3064
  id: string;
2985
3065
  name: string;
@@ -3488,6 +3568,23 @@ export declare const MODELS: {
3488
3568
  contextWindow: number;
3489
3569
  maxTokens: number;
3490
3570
  };
3571
+ readonly "claude-sonnet-4-6": {
3572
+ id: string;
3573
+ name: string;
3574
+ api: "google-gemini-cli";
3575
+ provider: string;
3576
+ baseUrl: string;
3577
+ reasoning: true;
3578
+ input: ("image" | "text")[];
3579
+ cost: {
3580
+ input: number;
3581
+ output: number;
3582
+ cacheRead: number;
3583
+ cacheWrite: number;
3584
+ };
3585
+ contextWindow: number;
3586
+ maxTokens: number;
3587
+ };
3491
3588
  readonly "gemini-3-flash": {
3492
3589
  id: string;
3493
3590
  name: string;
@@ -4666,7 +4763,7 @@ export declare const MODELS: {
4666
4763
  readonly "codestral-latest": {
4667
4764
  id: string;
4668
4765
  name: string;
4669
- api: "openai-completions";
4766
+ api: "mistral-conversations";
4670
4767
  provider: string;
4671
4768
  baseUrl: string;
4672
4769
  reasoning: false;
@@ -4683,7 +4780,7 @@ export declare const MODELS: {
4683
4780
  readonly "devstral-2512": {
4684
4781
  id: string;
4685
4782
  name: string;
4686
- api: "openai-completions";
4783
+ api: "mistral-conversations";
4687
4784
  provider: string;
4688
4785
  baseUrl: string;
4689
4786
  reasoning: false;
@@ -4700,7 +4797,7 @@ export declare const MODELS: {
4700
4797
  readonly "devstral-medium-2507": {
4701
4798
  id: string;
4702
4799
  name: string;
4703
- api: "openai-completions";
4800
+ api: "mistral-conversations";
4704
4801
  provider: string;
4705
4802
  baseUrl: string;
4706
4803
  reasoning: false;
@@ -4717,7 +4814,7 @@ export declare const MODELS: {
4717
4814
  readonly "devstral-medium-latest": {
4718
4815
  id: string;
4719
4816
  name: string;
4720
- api: "openai-completions";
4817
+ api: "mistral-conversations";
4721
4818
  provider: string;
4722
4819
  baseUrl: string;
4723
4820
  reasoning: false;
@@ -4734,7 +4831,7 @@ export declare const MODELS: {
4734
4831
  readonly "devstral-small-2505": {
4735
4832
  id: string;
4736
4833
  name: string;
4737
- api: "openai-completions";
4834
+ api: "mistral-conversations";
4738
4835
  provider: string;
4739
4836
  baseUrl: string;
4740
4837
  reasoning: false;
@@ -4751,7 +4848,7 @@ export declare const MODELS: {
4751
4848
  readonly "devstral-small-2507": {
4752
4849
  id: string;
4753
4850
  name: string;
4754
- api: "openai-completions";
4851
+ api: "mistral-conversations";
4755
4852
  provider: string;
4756
4853
  baseUrl: string;
4757
4854
  reasoning: false;
@@ -4768,7 +4865,7 @@ export declare const MODELS: {
4768
4865
  readonly "labs-devstral-small-2512": {
4769
4866
  id: string;
4770
4867
  name: string;
4771
- api: "openai-completions";
4868
+ api: "mistral-conversations";
4772
4869
  provider: string;
4773
4870
  baseUrl: string;
4774
4871
  reasoning: false;
@@ -4785,7 +4882,7 @@ export declare const MODELS: {
4785
4882
  readonly "magistral-medium-latest": {
4786
4883
  id: string;
4787
4884
  name: string;
4788
- api: "openai-completions";
4885
+ api: "mistral-conversations";
4789
4886
  provider: string;
4790
4887
  baseUrl: string;
4791
4888
  reasoning: true;
@@ -4802,7 +4899,7 @@ export declare const MODELS: {
4802
4899
  readonly "magistral-small": {
4803
4900
  id: string;
4804
4901
  name: string;
4805
- api: "openai-completions";
4902
+ api: "mistral-conversations";
4806
4903
  provider: string;
4807
4904
  baseUrl: string;
4808
4905
  reasoning: true;
@@ -4819,7 +4916,7 @@ export declare const MODELS: {
4819
4916
  readonly "ministral-3b-latest": {
4820
4917
  id: string;
4821
4918
  name: string;
4822
- api: "openai-completions";
4919
+ api: "mistral-conversations";
4823
4920
  provider: string;
4824
4921
  baseUrl: string;
4825
4922
  reasoning: false;
@@ -4836,7 +4933,7 @@ export declare const MODELS: {
4836
4933
  readonly "ministral-8b-latest": {
4837
4934
  id: string;
4838
4935
  name: string;
4839
- api: "openai-completions";
4936
+ api: "mistral-conversations";
4840
4937
  provider: string;
4841
4938
  baseUrl: string;
4842
4939
  reasoning: false;
@@ -4853,7 +4950,7 @@ export declare const MODELS: {
4853
4950
  readonly "mistral-large-2411": {
4854
4951
  id: string;
4855
4952
  name: string;
4856
- api: "openai-completions";
4953
+ api: "mistral-conversations";
4857
4954
  provider: string;
4858
4955
  baseUrl: string;
4859
4956
  reasoning: false;
@@ -4870,7 +4967,7 @@ export declare const MODELS: {
4870
4967
  readonly "mistral-large-2512": {
4871
4968
  id: string;
4872
4969
  name: string;
4873
- api: "openai-completions";
4970
+ api: "mistral-conversations";
4874
4971
  provider: string;
4875
4972
  baseUrl: string;
4876
4973
  reasoning: false;
@@ -4887,7 +4984,7 @@ export declare const MODELS: {
4887
4984
  readonly "mistral-large-latest": {
4888
4985
  id: string;
4889
4986
  name: string;
4890
- api: "openai-completions";
4987
+ api: "mistral-conversations";
4891
4988
  provider: string;
4892
4989
  baseUrl: string;
4893
4990
  reasoning: false;
@@ -4904,7 +5001,7 @@ export declare const MODELS: {
4904
5001
  readonly "mistral-medium-2505": {
4905
5002
  id: string;
4906
5003
  name: string;
4907
- api: "openai-completions";
5004
+ api: "mistral-conversations";
4908
5005
  provider: string;
4909
5006
  baseUrl: string;
4910
5007
  reasoning: false;
@@ -4921,7 +5018,7 @@ export declare const MODELS: {
4921
5018
  readonly "mistral-medium-2508": {
4922
5019
  id: string;
4923
5020
  name: string;
4924
- api: "openai-completions";
5021
+ api: "mistral-conversations";
4925
5022
  provider: string;
4926
5023
  baseUrl: string;
4927
5024
  reasoning: false;
@@ -4938,7 +5035,7 @@ export declare const MODELS: {
4938
5035
  readonly "mistral-medium-latest": {
4939
5036
  id: string;
4940
5037
  name: string;
4941
- api: "openai-completions";
5038
+ api: "mistral-conversations";
4942
5039
  provider: string;
4943
5040
  baseUrl: string;
4944
5041
  reasoning: false;
@@ -4955,7 +5052,7 @@ export declare const MODELS: {
4955
5052
  readonly "mistral-nemo": {
4956
5053
  id: string;
4957
5054
  name: string;
4958
- api: "openai-completions";
5055
+ api: "mistral-conversations";
4959
5056
  provider: string;
4960
5057
  baseUrl: string;
4961
5058
  reasoning: false;
@@ -4972,7 +5069,7 @@ export declare const MODELS: {
4972
5069
  readonly "mistral-small-2506": {
4973
5070
  id: string;
4974
5071
  name: string;
4975
- api: "openai-completions";
5072
+ api: "mistral-conversations";
4976
5073
  provider: string;
4977
5074
  baseUrl: string;
4978
5075
  reasoning: false;
@@ -4989,7 +5086,7 @@ export declare const MODELS: {
4989
5086
  readonly "mistral-small-latest": {
4990
5087
  id: string;
4991
5088
  name: string;
4992
- api: "openai-completions";
5089
+ api: "mistral-conversations";
4993
5090
  provider: string;
4994
5091
  baseUrl: string;
4995
5092
  reasoning: false;
@@ -5006,7 +5103,7 @@ export declare const MODELS: {
5006
5103
  readonly "open-mistral-7b": {
5007
5104
  id: string;
5008
5105
  name: string;
5009
- api: "openai-completions";
5106
+ api: "mistral-conversations";
5010
5107
  provider: string;
5011
5108
  baseUrl: string;
5012
5109
  reasoning: false;
@@ -5023,7 +5120,7 @@ export declare const MODELS: {
5023
5120
  readonly "open-mixtral-8x22b": {
5024
5121
  id: string;
5025
5122
  name: string;
5026
- api: "openai-completions";
5123
+ api: "mistral-conversations";
5027
5124
  provider: string;
5028
5125
  baseUrl: string;
5029
5126
  reasoning: false;
@@ -5040,7 +5137,7 @@ export declare const MODELS: {
5040
5137
  readonly "open-mixtral-8x7b": {
5041
5138
  id: string;
5042
5139
  name: string;
5043
- api: "openai-completions";
5140
+ api: "mistral-conversations";
5044
5141
  provider: string;
5045
5142
  baseUrl: string;
5046
5143
  reasoning: false;
@@ -5057,7 +5154,7 @@ export declare const MODELS: {
5057
5154
  readonly "pixtral-12b": {
5058
5155
  id: string;
5059
5156
  name: string;
5060
- api: "openai-completions";
5157
+ api: "mistral-conversations";
5061
5158
  provider: string;
5062
5159
  baseUrl: string;
5063
5160
  reasoning: false;
@@ -5074,7 +5171,7 @@ export declare const MODELS: {
5074
5171
  readonly "pixtral-large-latest": {
5075
5172
  id: string;
5076
5173
  name: string;
5077
- api: "openai-completions";
5174
+ api: "mistral-conversations";
5078
5175
  provider: string;
5079
5176
  baseUrl: string;
5080
5177
  reasoning: false;
@@ -5566,6 +5663,40 @@ export declare const MODELS: {
5566
5663
  contextWindow: number;
5567
5664
  maxTokens: number;
5568
5665
  };
5666
+ readonly "gpt-5.4": {
5667
+ id: string;
5668
+ name: string;
5669
+ api: "openai-responses";
5670
+ provider: string;
5671
+ baseUrl: string;
5672
+ reasoning: true;
5673
+ input: ("image" | "text")[];
5674
+ cost: {
5675
+ input: number;
5676
+ output: number;
5677
+ cacheRead: number;
5678
+ cacheWrite: number;
5679
+ };
5680
+ contextWindow: number;
5681
+ maxTokens: number;
5682
+ };
5683
+ readonly "gpt-5.4-pro": {
5684
+ id: string;
5685
+ name: string;
5686
+ api: "openai-responses";
5687
+ provider: string;
5688
+ baseUrl: string;
5689
+ reasoning: true;
5690
+ input: ("image" | "text")[];
5691
+ cost: {
5692
+ input: number;
5693
+ output: number;
5694
+ cacheRead: number;
5695
+ cacheWrite: number;
5696
+ };
5697
+ contextWindow: number;
5698
+ maxTokens: number;
5699
+ };
5569
5700
  readonly o1: {
5570
5701
  id: string;
5571
5702
  name: string;
@@ -5823,6 +5954,23 @@ export declare const MODELS: {
5823
5954
  contextWindow: number;
5824
5955
  maxTokens: number;
5825
5956
  };
5957
+ readonly "gpt-5.4": {
5958
+ id: string;
5959
+ name: string;
5960
+ api: "openai-codex-responses";
5961
+ provider: string;
5962
+ baseUrl: string;
5963
+ reasoning: true;
5964
+ input: ("image" | "text")[];
5965
+ cost: {
5966
+ input: number;
5967
+ output: number;
5968
+ cacheRead: number;
5969
+ cacheWrite: number;
5970
+ };
5971
+ contextWindow: number;
5972
+ maxTokens: number;
5973
+ };
5826
5974
  };
5827
5975
  readonly opencode: {
5828
5976
  readonly "big-pickle": {
@@ -6250,10 +6398,10 @@ export declare const MODELS: {
6250
6398
  contextWindow: number;
6251
6399
  maxTokens: number;
6252
6400
  };
6253
- readonly "kimi-k2.5": {
6401
+ readonly "gpt-5.4": {
6254
6402
  id: string;
6255
6403
  name: string;
6256
- api: "openai-completions";
6404
+ api: "openai-responses";
6257
6405
  provider: string;
6258
6406
  baseUrl: string;
6259
6407
  reasoning: true;
@@ -6267,14 +6415,31 @@ export declare const MODELS: {
6267
6415
  contextWindow: number;
6268
6416
  maxTokens: number;
6269
6417
  };
6270
- readonly "minimax-m2.1": {
6418
+ readonly "gpt-5.4-pro": {
6419
+ id: string;
6420
+ name: string;
6421
+ api: "openai-responses";
6422
+ provider: string;
6423
+ baseUrl: string;
6424
+ reasoning: true;
6425
+ input: ("image" | "text")[];
6426
+ cost: {
6427
+ input: number;
6428
+ output: number;
6429
+ cacheRead: number;
6430
+ cacheWrite: number;
6431
+ };
6432
+ contextWindow: number;
6433
+ maxTokens: number;
6434
+ };
6435
+ readonly "kimi-k2.5": {
6271
6436
  id: string;
6272
6437
  name: string;
6273
6438
  api: "openai-completions";
6274
6439
  provider: string;
6275
6440
  baseUrl: string;
6276
6441
  reasoning: true;
6277
- input: "text"[];
6442
+ input: ("image" | "text")[];
6278
6443
  cost: {
6279
6444
  input: number;
6280
6445
  output: number;
@@ -6284,7 +6449,7 @@ export declare const MODELS: {
6284
6449
  contextWindow: number;
6285
6450
  maxTokens: number;
6286
6451
  };
6287
- readonly "minimax-m2.5": {
6452
+ readonly "minimax-m2.1": {
6288
6453
  id: string;
6289
6454
  name: string;
6290
6455
  api: "openai-completions";
@@ -6301,10 +6466,10 @@ export declare const MODELS: {
6301
6466
  contextWindow: number;
6302
6467
  maxTokens: number;
6303
6468
  };
6304
- readonly "minimax-m2.5-free": {
6469
+ readonly "minimax-m2.5": {
6305
6470
  id: string;
6306
6471
  name: string;
6307
- api: "anthropic-messages";
6472
+ api: "openai-completions";
6308
6473
  provider: string;
6309
6474
  baseUrl: string;
6310
6475
  reasoning: true;
@@ -6318,13 +6483,13 @@ export declare const MODELS: {
6318
6483
  contextWindow: number;
6319
6484
  maxTokens: number;
6320
6485
  };
6321
- readonly "trinity-large-preview-free": {
6486
+ readonly "minimax-m2.5-free": {
6322
6487
  id: string;
6323
6488
  name: string;
6324
- api: "openai-completions";
6489
+ api: "anthropic-messages";
6325
6490
  provider: string;
6326
6491
  baseUrl: string;
6327
- reasoning: false;
6492
+ reasoning: true;
6328
6493
  input: "text"[];
6329
6494
  cost: {
6330
6495
  input: number;
@@ -7393,6 +7558,23 @@ export declare const MODELS: {
7393
7558
  contextWindow: number;
7394
7559
  maxTokens: number;
7395
7560
  };
7561
+ readonly "inception/mercury-2": {
7562
+ id: string;
7563
+ name: string;
7564
+ api: "openai-completions";
7565
+ provider: string;
7566
+ baseUrl: string;
7567
+ reasoning: true;
7568
+ input: "text"[];
7569
+ cost: {
7570
+ input: number;
7571
+ output: number;
7572
+ cacheRead: number;
7573
+ cacheWrite: number;
7574
+ };
7575
+ contextWindow: number;
7576
+ maxTokens: number;
7577
+ };
7396
7578
  readonly "inception/mercury-coder": {
7397
7579
  id: string;
7398
7580
  name: string;
@@ -8889,6 +9071,40 @@ export declare const MODELS: {
8889
9071
  contextWindow: number;
8890
9072
  maxTokens: number;
8891
9073
  };
9074
+ readonly "openai/gpt-5.4": {
9075
+ id: string;
9076
+ name: string;
9077
+ api: "openai-completions";
9078
+ provider: string;
9079
+ baseUrl: string;
9080
+ reasoning: true;
9081
+ input: ("image" | "text")[];
9082
+ cost: {
9083
+ input: number;
9084
+ output: number;
9085
+ cacheRead: number;
9086
+ cacheWrite: number;
9087
+ };
9088
+ contextWindow: number;
9089
+ maxTokens: number;
9090
+ };
9091
+ readonly "openai/gpt-5.4-pro": {
9092
+ id: string;
9093
+ name: string;
9094
+ api: "openai-completions";
9095
+ provider: string;
9096
+ baseUrl: string;
9097
+ reasoning: true;
9098
+ input: ("image" | "text")[];
9099
+ cost: {
9100
+ input: number;
9101
+ output: number;
9102
+ cacheRead: number;
9103
+ cacheWrite: number;
9104
+ };
9105
+ contextWindow: number;
9106
+ maxTokens: number;
9107
+ };
8892
9108
  readonly "openai/gpt-oss-120b": {
8893
9109
  id: string;
8894
9110
  name: string;
@@ -12138,6 +12354,40 @@ export declare const MODELS: {
12138
12354
  contextWindow: number;
12139
12355
  maxTokens: number;
12140
12356
  };
12357
+ readonly "openai/gpt-5.4": {
12358
+ id: string;
12359
+ name: string;
12360
+ api: "anthropic-messages";
12361
+ provider: string;
12362
+ baseUrl: string;
12363
+ reasoning: true;
12364
+ input: ("image" | "text")[];
12365
+ cost: {
12366
+ input: number;
12367
+ output: number;
12368
+ cacheRead: number;
12369
+ cacheWrite: number;
12370
+ };
12371
+ contextWindow: number;
12372
+ maxTokens: number;
12373
+ };
12374
+ readonly "openai/gpt-5.4-pro": {
12375
+ id: string;
12376
+ name: string;
12377
+ api: "anthropic-messages";
12378
+ provider: string;
12379
+ baseUrl: string;
12380
+ reasoning: true;
12381
+ input: ("image" | "text")[];
12382
+ cost: {
12383
+ input: number;
12384
+ output: number;
12385
+ cacheRead: number;
12386
+ cacheWrite: number;
12387
+ };
12388
+ contextWindow: number;
12389
+ maxTokens: number;
12390
+ };
12141
12391
  readonly "openai/gpt-oss-120b": {
12142
12392
  id: string;
12143
12393
  name: string;