@mariozechner/pi-ai 0.59.0 → 0.61.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (45) hide show
  1. package/dist/models.generated.d.ts +398 -86
  2. package/dist/models.generated.d.ts.map +1 -1
  3. package/dist/models.generated.js +445 -138
  4. package/dist/models.generated.js.map +1 -1
  5. package/dist/providers/amazon-bedrock.d.ts.map +1 -1
  6. package/dist/providers/amazon-bedrock.js +13 -1
  7. package/dist/providers/amazon-bedrock.js.map +1 -1
  8. package/dist/providers/google-shared.d.ts.map +1 -1
  9. package/dist/providers/google-shared.js +20 -8
  10. package/dist/providers/google-shared.js.map +1 -1
  11. package/dist/providers/google-vertex.d.ts.map +1 -1
  12. package/dist/providers/google-vertex.js +8 -1
  13. package/dist/providers/google-vertex.js.map +1 -1
  14. package/dist/providers/openai-completions.d.ts.map +1 -1
  15. package/dist/providers/openai-completions.js +34 -13
  16. package/dist/providers/openai-completions.js.map +1 -1
  17. package/dist/providers/openai-responses-shared.d.ts.map +1 -1
  18. package/dist/providers/openai-responses-shared.js +11 -11
  19. package/dist/providers/openai-responses-shared.js.map +1 -1
  20. package/dist/providers/register-builtins.d.ts.map +1 -1
  21. package/dist/providers/register-builtins.js +11 -21
  22. package/dist/providers/register-builtins.js.map +1 -1
  23. package/dist/types.d.ts +2 -2
  24. package/dist/types.d.ts.map +1 -1
  25. package/dist/types.js.map +1 -1
  26. package/dist/utils/oauth/anthropic.d.ts.map +1 -1
  27. package/dist/utils/oauth/anthropic.js +20 -28
  28. package/dist/utils/oauth/anthropic.js.map +1 -1
  29. package/dist/utils/oauth/google-antigravity.d.ts.map +1 -1
  30. package/dist/utils/oauth/google-antigravity.js +22 -19
  31. package/dist/utils/oauth/google-antigravity.js.map +1 -1
  32. package/dist/utils/oauth/google-gemini-cli.d.ts.map +1 -1
  33. package/dist/utils/oauth/google-gemini-cli.js +22 -19
  34. package/dist/utils/oauth/google-gemini-cli.js.map +1 -1
  35. package/dist/utils/oauth/oauth-page.d.ts +3 -0
  36. package/dist/utils/oauth/oauth-page.d.ts.map +1 -0
  37. package/dist/utils/oauth/oauth-page.js +105 -0
  38. package/dist/utils/oauth/oauth-page.js.map +1 -0
  39. package/dist/utils/oauth/openai-codex.d.ts.map +1 -1
  40. package/dist/utils/oauth/openai-codex.js +24 -31
  41. package/dist/utils/oauth/openai-codex.js.map +1 -1
  42. package/dist/utils/validation.d.ts.map +1 -1
  43. package/dist/utils/validation.js +17 -9
  44. package/dist/utils/validation.js.map +1 -1
  45. package/package.json +1 -1
@@ -2299,6 +2299,40 @@ export declare const MODELS: {
2299
2299
  contextWindow: number;
2300
2300
  maxTokens: number;
2301
2301
  };
2302
+ readonly "gpt-5.4-mini": {
2303
+ id: string;
2304
+ name: string;
2305
+ api: "azure-openai-responses";
2306
+ provider: string;
2307
+ baseUrl: string;
2308
+ reasoning: true;
2309
+ input: ("image" | "text")[];
2310
+ cost: {
2311
+ input: number;
2312
+ output: number;
2313
+ cacheRead: number;
2314
+ cacheWrite: number;
2315
+ };
2316
+ contextWindow: number;
2317
+ maxTokens: number;
2318
+ };
2319
+ readonly "gpt-5.4-nano": {
2320
+ id: string;
2321
+ name: string;
2322
+ api: "azure-openai-responses";
2323
+ provider: string;
2324
+ baseUrl: string;
2325
+ reasoning: true;
2326
+ input: ("image" | "text")[];
2327
+ cost: {
2328
+ input: number;
2329
+ output: number;
2330
+ cacheRead: number;
2331
+ cacheWrite: number;
2332
+ };
2333
+ contextWindow: number;
2334
+ maxTokens: number;
2335
+ };
2302
2336
  readonly "gpt-5.4-pro": {
2303
2337
  id: string;
2304
2338
  name: string;
@@ -3060,6 +3094,29 @@ export declare const MODELS: {
3060
3094
  contextWindow: number;
3061
3095
  maxTokens: number;
3062
3096
  };
3097
+ readonly "gpt-5.4-mini": {
3098
+ id: string;
3099
+ name: string;
3100
+ api: "openai-responses";
3101
+ provider: string;
3102
+ baseUrl: string;
3103
+ headers: {
3104
+ "User-Agent": string;
3105
+ "Editor-Version": string;
3106
+ "Editor-Plugin-Version": string;
3107
+ "Copilot-Integration-Id": string;
3108
+ };
3109
+ reasoning: true;
3110
+ input: ("image" | "text")[];
3111
+ cost: {
3112
+ input: number;
3113
+ output: number;
3114
+ cacheRead: number;
3115
+ cacheWrite: number;
3116
+ };
3117
+ contextWindow: number;
3118
+ maxTokens: number;
3119
+ };
3063
3120
  readonly "grok-code-fast-1": {
3064
3121
  id: string;
3065
3122
  name: string;
@@ -4688,6 +4745,40 @@ export declare const MODELS: {
4688
4745
  contextWindow: number;
4689
4746
  maxTokens: number;
4690
4747
  };
4748
+ readonly "MiniMax-M2.7": {
4749
+ id: string;
4750
+ name: string;
4751
+ api: "anthropic-messages";
4752
+ provider: string;
4753
+ baseUrl: string;
4754
+ reasoning: true;
4755
+ input: "text"[];
4756
+ cost: {
4757
+ input: number;
4758
+ output: number;
4759
+ cacheRead: number;
4760
+ cacheWrite: number;
4761
+ };
4762
+ contextWindow: number;
4763
+ maxTokens: number;
4764
+ };
4765
+ readonly "MiniMax-M2.7-highspeed": {
4766
+ id: string;
4767
+ name: string;
4768
+ api: "anthropic-messages";
4769
+ provider: string;
4770
+ baseUrl: string;
4771
+ reasoning: true;
4772
+ input: "text"[];
4773
+ cost: {
4774
+ input: number;
4775
+ output: number;
4776
+ cacheRead: number;
4777
+ cacheWrite: number;
4778
+ };
4779
+ contextWindow: number;
4780
+ maxTokens: number;
4781
+ };
4691
4782
  };
4692
4783
  readonly "minimax-cn": {
4693
4784
  readonly "MiniMax-M2": {
@@ -4758,6 +4849,40 @@ export declare const MODELS: {
4758
4849
  contextWindow: number;
4759
4850
  maxTokens: number;
4760
4851
  };
4852
+ readonly "MiniMax-M2.7": {
4853
+ id: string;
4854
+ name: string;
4855
+ api: "anthropic-messages";
4856
+ provider: string;
4857
+ baseUrl: string;
4858
+ reasoning: true;
4859
+ input: "text"[];
4860
+ cost: {
4861
+ input: number;
4862
+ output: number;
4863
+ cacheRead: number;
4864
+ cacheWrite: number;
4865
+ };
4866
+ contextWindow: number;
4867
+ maxTokens: number;
4868
+ };
4869
+ readonly "MiniMax-M2.7-highspeed": {
4870
+ id: string;
4871
+ name: string;
4872
+ api: "anthropic-messages";
4873
+ provider: string;
4874
+ baseUrl: string;
4875
+ reasoning: true;
4876
+ input: "text"[];
4877
+ cost: {
4878
+ input: number;
4879
+ output: number;
4880
+ cacheRead: number;
4881
+ cacheWrite: number;
4882
+ };
4883
+ contextWindow: number;
4884
+ maxTokens: number;
4885
+ };
4761
4886
  };
4762
4887
  readonly mistral: {
4763
4888
  readonly "codestral-latest": {
@@ -5680,6 +5805,40 @@ export declare const MODELS: {
5680
5805
  contextWindow: number;
5681
5806
  maxTokens: number;
5682
5807
  };
5808
+ readonly "gpt-5.4-mini": {
5809
+ id: string;
5810
+ name: string;
5811
+ api: "openai-responses";
5812
+ provider: string;
5813
+ baseUrl: string;
5814
+ reasoning: true;
5815
+ input: ("image" | "text")[];
5816
+ cost: {
5817
+ input: number;
5818
+ output: number;
5819
+ cacheRead: number;
5820
+ cacheWrite: number;
5821
+ };
5822
+ contextWindow: number;
5823
+ maxTokens: number;
5824
+ };
5825
+ readonly "gpt-5.4-nano": {
5826
+ id: string;
5827
+ name: string;
5828
+ api: "openai-responses";
5829
+ provider: string;
5830
+ baseUrl: string;
5831
+ reasoning: true;
5832
+ input: ("image" | "text")[];
5833
+ cost: {
5834
+ input: number;
5835
+ output: number;
5836
+ cacheRead: number;
5837
+ cacheWrite: number;
5838
+ };
5839
+ contextWindow: number;
5840
+ maxTokens: number;
5841
+ };
5683
5842
  readonly "gpt-5.4-pro": {
5684
5843
  id: string;
5685
5844
  name: string;
@@ -5971,6 +6130,23 @@ export declare const MODELS: {
5971
6130
  contextWindow: number;
5972
6131
  maxTokens: number;
5973
6132
  };
6133
+ readonly "gpt-5.4-mini": {
6134
+ id: string;
6135
+ name: string;
6136
+ api: "openai-codex-responses";
6137
+ provider: string;
6138
+ baseUrl: string;
6139
+ reasoning: true;
6140
+ input: ("image" | "text")[];
6141
+ cost: {
6142
+ input: number;
6143
+ output: number;
6144
+ cacheRead: number;
6145
+ cacheWrite: number;
6146
+ };
6147
+ contextWindow: number;
6148
+ maxTokens: number;
6149
+ };
5974
6150
  };
5975
6151
  readonly opencode: {
5976
6152
  readonly "big-pickle": {
@@ -6364,6 +6540,40 @@ export declare const MODELS: {
6364
6540
  contextWindow: number;
6365
6541
  maxTokens: number;
6366
6542
  };
6543
+ readonly "gpt-5.4-mini": {
6544
+ id: string;
6545
+ name: string;
6546
+ api: "openai-responses";
6547
+ provider: string;
6548
+ baseUrl: string;
6549
+ reasoning: true;
6550
+ input: ("image" | "text")[];
6551
+ cost: {
6552
+ input: number;
6553
+ output: number;
6554
+ cacheRead: number;
6555
+ cacheWrite: number;
6556
+ };
6557
+ contextWindow: number;
6558
+ maxTokens: number;
6559
+ };
6560
+ readonly "gpt-5.4-nano": {
6561
+ id: string;
6562
+ name: string;
6563
+ api: "openai-responses";
6564
+ provider: string;
6565
+ baseUrl: string;
6566
+ reasoning: true;
6567
+ input: ("image" | "text")[];
6568
+ cost: {
6569
+ input: number;
6570
+ output: number;
6571
+ cacheRead: number;
6572
+ cacheWrite: number;
6573
+ };
6574
+ contextWindow: number;
6575
+ maxTokens: number;
6576
+ };
6367
6577
  readonly "gpt-5.4-pro": {
6368
6578
  id: string;
6369
6579
  name: string;
@@ -6398,7 +6608,24 @@ export declare const MODELS: {
6398
6608
  contextWindow: number;
6399
6609
  maxTokens: number;
6400
6610
  };
6401
- readonly "mimo-v2-flash-free": {
6611
+ readonly "mimo-v2-omni-free": {
6612
+ id: string;
6613
+ name: string;
6614
+ api: "openai-completions";
6615
+ provider: string;
6616
+ baseUrl: string;
6617
+ reasoning: true;
6618
+ input: ("image" | "text")[];
6619
+ cost: {
6620
+ input: number;
6621
+ output: number;
6622
+ cacheRead: number;
6623
+ cacheWrite: number;
6624
+ };
6625
+ contextWindow: number;
6626
+ maxTokens: number;
6627
+ };
6628
+ readonly "mimo-v2-pro-free": {
6402
6629
  id: string;
6403
6630
  name: string;
6404
6631
  api: "openai-completions";
@@ -6519,6 +6746,23 @@ export declare const MODELS: {
6519
6746
  contextWindow: number;
6520
6747
  maxTokens: number;
6521
6748
  };
6749
+ readonly "minimax-m2.7": {
6750
+ id: string;
6751
+ name: string;
6752
+ api: "anthropic-messages";
6753
+ provider: string;
6754
+ baseUrl: string;
6755
+ reasoning: true;
6756
+ input: "text"[];
6757
+ cost: {
6758
+ input: number;
6759
+ output: number;
6760
+ cacheRead: number;
6761
+ cacheWrite: number;
6762
+ };
6763
+ contextWindow: number;
6764
+ maxTokens: number;
6765
+ };
6522
6766
  };
6523
6767
  readonly openrouter: {
6524
6768
  readonly "ai21/jamba-large-1.7": {
@@ -7473,40 +7717,6 @@ export declare const MODELS: {
7473
7717
  contextWindow: number;
7474
7718
  maxTokens: number;
7475
7719
  };
7476
- readonly "google/gemma-3-27b-it": {
7477
- id: string;
7478
- name: string;
7479
- api: "openai-completions";
7480
- provider: string;
7481
- baseUrl: string;
7482
- reasoning: false;
7483
- input: ("image" | "text")[];
7484
- cost: {
7485
- input: number;
7486
- output: number;
7487
- cacheRead: number;
7488
- cacheWrite: number;
7489
- };
7490
- contextWindow: number;
7491
- maxTokens: number;
7492
- };
7493
- readonly "google/gemma-3-27b-it:free": {
7494
- id: string;
7495
- name: string;
7496
- api: "openai-completions";
7497
- provider: string;
7498
- baseUrl: string;
7499
- reasoning: false;
7500
- input: ("image" | "text")[];
7501
- cost: {
7502
- input: number;
7503
- output: number;
7504
- cacheRead: number;
7505
- cacheWrite: number;
7506
- };
7507
- contextWindow: number;
7508
- maxTokens: number;
7509
- };
7510
7720
  readonly "inception/mercury": {
7511
7721
  id: string;
7512
7722
  name: string;
@@ -7796,6 +8006,23 @@ export declare const MODELS: {
7796
8006
  contextWindow: number;
7797
8007
  maxTokens: number;
7798
8008
  };
8009
+ readonly "minimax/minimax-m2.7": {
8010
+ id: string;
8011
+ name: string;
8012
+ api: "openai-completions";
8013
+ provider: string;
8014
+ baseUrl: string;
8015
+ reasoning: true;
8016
+ input: "text"[];
8017
+ cost: {
8018
+ input: number;
8019
+ output: number;
8020
+ cacheRead: number;
8021
+ cacheWrite: number;
8022
+ };
8023
+ contextWindow: number;
8024
+ maxTokens: number;
8025
+ };
7799
8026
  readonly "mistralai/codestral-2508": {
7800
8027
  id: string;
7801
8028
  name: string;
@@ -8357,6 +8584,23 @@ export declare const MODELS: {
8357
8584
  contextWindow: number;
8358
8585
  maxTokens: number;
8359
8586
  };
8587
+ readonly "nvidia/nemotron-3-super-120b-a12b": {
8588
+ id: string;
8589
+ name: string;
8590
+ api: "openai-completions";
8591
+ provider: string;
8592
+ baseUrl: string;
8593
+ reasoning: true;
8594
+ input: "text"[];
8595
+ cost: {
8596
+ input: number;
8597
+ output: number;
8598
+ cacheRead: number;
8599
+ cacheWrite: number;
8600
+ };
8601
+ contextWindow: number;
8602
+ maxTokens: number;
8603
+ };
8360
8604
  readonly "nvidia/nemotron-3-super-120b-a12b:free": {
8361
8605
  id: string;
8362
8606
  name: string;
@@ -9394,40 +9638,6 @@ export declare const MODELS: {
9394
9638
  contextWindow: number;
9395
9639
  maxTokens: number;
9396
9640
  };
9397
- readonly "openrouter/healer-alpha": {
9398
- id: string;
9399
- name: string;
9400
- api: "openai-completions";
9401
- provider: string;
9402
- baseUrl: string;
9403
- reasoning: true;
9404
- input: ("image" | "text")[];
9405
- cost: {
9406
- input: number;
9407
- output: number;
9408
- cacheRead: number;
9409
- cacheWrite: number;
9410
- };
9411
- contextWindow: number;
9412
- maxTokens: number;
9413
- };
9414
- readonly "openrouter/hunter-alpha": {
9415
- id: string;
9416
- name: string;
9417
- api: "openai-completions";
9418
- provider: string;
9419
- baseUrl: string;
9420
- reasoning: true;
9421
- input: ("image" | "text")[];
9422
- cost: {
9423
- input: number;
9424
- output: number;
9425
- cacheRead: number;
9426
- cacheWrite: number;
9427
- };
9428
- contextWindow: number;
9429
- maxTokens: number;
9430
- };
9431
9641
  readonly "prime-intellect/intellect-3": {
9432
9642
  id: string;
9433
9643
  name: string;
@@ -10516,6 +10726,40 @@ export declare const MODELS: {
10516
10726
  contextWindow: number;
10517
10727
  maxTokens: number;
10518
10728
  };
10729
+ readonly "xiaomi/mimo-v2-omni": {
10730
+ id: string;
10731
+ name: string;
10732
+ api: "openai-completions";
10733
+ provider: string;
10734
+ baseUrl: string;
10735
+ reasoning: true;
10736
+ input: ("image" | "text")[];
10737
+ cost: {
10738
+ input: number;
10739
+ output: number;
10740
+ cacheRead: number;
10741
+ cacheWrite: number;
10742
+ };
10743
+ contextWindow: number;
10744
+ maxTokens: number;
10745
+ };
10746
+ readonly "xiaomi/mimo-v2-pro": {
10747
+ id: string;
10748
+ name: string;
10749
+ api: "openai-completions";
10750
+ provider: string;
10751
+ baseUrl: string;
10752
+ reasoning: true;
10753
+ input: "text"[];
10754
+ cost: {
10755
+ input: number;
10756
+ output: number;
10757
+ cacheRead: number;
10758
+ cacheWrite: number;
10759
+ };
10760
+ contextWindow: number;
10761
+ maxTokens: number;
10762
+ };
10519
10763
  readonly "z-ai/glm-4-32b": {
10520
10764
  id: string;
10521
10765
  name: string;
@@ -11759,6 +12003,40 @@ export declare const MODELS: {
11759
12003
  contextWindow: number;
11760
12004
  maxTokens: number;
11761
12005
  };
12006
+ readonly "minimax/minimax-m2.7": {
12007
+ id: string;
12008
+ name: string;
12009
+ api: "anthropic-messages";
12010
+ provider: string;
12011
+ baseUrl: string;
12012
+ reasoning: true;
12013
+ input: ("image" | "text")[];
12014
+ cost: {
12015
+ input: number;
12016
+ output: number;
12017
+ cacheRead: number;
12018
+ cacheWrite: number;
12019
+ };
12020
+ contextWindow: number;
12021
+ maxTokens: number;
12022
+ };
12023
+ readonly "minimax/minimax-m2.7-highspeed": {
12024
+ id: string;
12025
+ name: string;
12026
+ api: "anthropic-messages";
12027
+ provider: string;
12028
+ baseUrl: string;
12029
+ reasoning: true;
12030
+ input: ("image" | "text")[];
12031
+ cost: {
12032
+ input: number;
12033
+ output: number;
12034
+ cacheRead: number;
12035
+ cacheWrite: number;
12036
+ };
12037
+ contextWindow: number;
12038
+ maxTokens: number;
12039
+ };
11762
12040
  readonly "mistral/codestral": {
11763
12041
  id: string;
11764
12042
  name: string;
@@ -12473,6 +12751,40 @@ export declare const MODELS: {
12473
12751
  contextWindow: number;
12474
12752
  maxTokens: number;
12475
12753
  };
12754
+ readonly "openai/gpt-5.4-mini": {
12755
+ id: string;
12756
+ name: string;
12757
+ api: "anthropic-messages";
12758
+ provider: string;
12759
+ baseUrl: string;
12760
+ reasoning: true;
12761
+ input: ("image" | "text")[];
12762
+ cost: {
12763
+ input: number;
12764
+ output: number;
12765
+ cacheRead: number;
12766
+ cacheWrite: number;
12767
+ };
12768
+ contextWindow: number;
12769
+ maxTokens: number;
12770
+ };
12771
+ readonly "openai/gpt-5.4-nano": {
12772
+ id: string;
12773
+ name: string;
12774
+ api: "anthropic-messages";
12775
+ provider: string;
12776
+ baseUrl: string;
12777
+ reasoning: true;
12778
+ input: ("image" | "text")[];
12779
+ cost: {
12780
+ input: number;
12781
+ output: number;
12782
+ cacheRead: number;
12783
+ cacheWrite: number;
12784
+ };
12785
+ contextWindow: number;
12786
+ maxTokens: number;
12787
+ };
12476
12788
  readonly "openai/gpt-5.4-pro": {
12477
12789
  id: string;
12478
12790
  name: string;
@@ -12932,6 +13244,23 @@ export declare const MODELS: {
12932
13244
  contextWindow: number;
12933
13245
  maxTokens: number;
12934
13246
  };
13247
+ readonly "xiaomi/mimo-v2-pro": {
13248
+ id: string;
13249
+ name: string;
13250
+ api: "anthropic-messages";
13251
+ provider: string;
13252
+ baseUrl: string;
13253
+ reasoning: true;
13254
+ input: "text"[];
13255
+ cost: {
13256
+ input: number;
13257
+ output: number;
13258
+ cacheRead: number;
13259
+ cacheWrite: number;
13260
+ };
13261
+ contextWindow: number;
13262
+ maxTokens: number;
13263
+ };
12935
13264
  readonly "zai/glm-4.5": {
12936
13265
  id: string;
12937
13266
  name: string;
@@ -13478,23 +13807,6 @@ export declare const MODELS: {
13478
13807
  contextWindow: number;
13479
13808
  maxTokens: number;
13480
13809
  };
13481
- readonly "grok-4.20-multi-agent-beta-latest": {
13482
- id: string;
13483
- name: string;
13484
- api: "openai-completions";
13485
- provider: string;
13486
- baseUrl: string;
13487
- reasoning: true;
13488
- input: ("image" | "text")[];
13489
- cost: {
13490
- input: number;
13491
- output: number;
13492
- cacheRead: number;
13493
- cacheWrite: number;
13494
- };
13495
- contextWindow: number;
13496
- maxTokens: number;
13497
- };
13498
13810
  readonly "grok-beta": {
13499
13811
  id: string;
13500
13812
  name: string;