@mariozechner/pi-ai 0.60.0 → 0.61.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -3117,29 +3117,6 @@ export declare const MODELS: {
3117
3117
  contextWindow: number;
3118
3118
  maxTokens: number;
3119
3119
  };
3120
- readonly "gpt-5.4-nano": {
3121
- id: string;
3122
- name: string;
3123
- api: "openai-responses";
3124
- provider: string;
3125
- baseUrl: string;
3126
- headers: {
3127
- "User-Agent": string;
3128
- "Editor-Version": string;
3129
- "Editor-Plugin-Version": string;
3130
- "Copilot-Integration-Id": string;
3131
- };
3132
- reasoning: true;
3133
- input: ("image" | "text")[];
3134
- cost: {
3135
- input: number;
3136
- output: number;
3137
- cacheRead: number;
3138
- cacheWrite: number;
3139
- };
3140
- contextWindow: number;
3141
- maxTokens: number;
3142
- };
3143
3120
  readonly "grok-code-fast-1": {
3144
3121
  id: string;
3145
3122
  name: string;
@@ -4768,6 +4745,40 @@ export declare const MODELS: {
4768
4745
  contextWindow: number;
4769
4746
  maxTokens: number;
4770
4747
  };
4748
+ readonly "MiniMax-M2.7": {
4749
+ id: string;
4750
+ name: string;
4751
+ api: "anthropic-messages";
4752
+ provider: string;
4753
+ baseUrl: string;
4754
+ reasoning: true;
4755
+ input: "text"[];
4756
+ cost: {
4757
+ input: number;
4758
+ output: number;
4759
+ cacheRead: number;
4760
+ cacheWrite: number;
4761
+ };
4762
+ contextWindow: number;
4763
+ maxTokens: number;
4764
+ };
4765
+ readonly "MiniMax-M2.7-highspeed": {
4766
+ id: string;
4767
+ name: string;
4768
+ api: "anthropic-messages";
4769
+ provider: string;
4770
+ baseUrl: string;
4771
+ reasoning: true;
4772
+ input: "text"[];
4773
+ cost: {
4774
+ input: number;
4775
+ output: number;
4776
+ cacheRead: number;
4777
+ cacheWrite: number;
4778
+ };
4779
+ contextWindow: number;
4780
+ maxTokens: number;
4781
+ };
4771
4782
  };
4772
4783
  readonly "minimax-cn": {
4773
4784
  readonly "MiniMax-M2": {
@@ -4838,6 +4849,40 @@ export declare const MODELS: {
4838
4849
  contextWindow: number;
4839
4850
  maxTokens: number;
4840
4851
  };
4852
+ readonly "MiniMax-M2.7": {
4853
+ id: string;
4854
+ name: string;
4855
+ api: "anthropic-messages";
4856
+ provider: string;
4857
+ baseUrl: string;
4858
+ reasoning: true;
4859
+ input: "text"[];
4860
+ cost: {
4861
+ input: number;
4862
+ output: number;
4863
+ cacheRead: number;
4864
+ cacheWrite: number;
4865
+ };
4866
+ contextWindow: number;
4867
+ maxTokens: number;
4868
+ };
4869
+ readonly "MiniMax-M2.7-highspeed": {
4870
+ id: string;
4871
+ name: string;
4872
+ api: "anthropic-messages";
4873
+ provider: string;
4874
+ baseUrl: string;
4875
+ reasoning: true;
4876
+ input: "text"[];
4877
+ cost: {
4878
+ input: number;
4879
+ output: number;
4880
+ cacheRead: number;
4881
+ cacheWrite: number;
4882
+ };
4883
+ contextWindow: number;
4884
+ maxTokens: number;
4885
+ };
4841
4886
  };
4842
4887
  readonly mistral: {
4843
4888
  readonly "codestral-latest": {
@@ -6085,6 +6130,23 @@ export declare const MODELS: {
6085
6130
  contextWindow: number;
6086
6131
  maxTokens: number;
6087
6132
  };
6133
+ readonly "gpt-5.4-mini": {
6134
+ id: string;
6135
+ name: string;
6136
+ api: "openai-codex-responses";
6137
+ provider: string;
6138
+ baseUrl: string;
6139
+ reasoning: true;
6140
+ input: ("image" | "text")[];
6141
+ cost: {
6142
+ input: number;
6143
+ output: number;
6144
+ cacheRead: number;
6145
+ cacheWrite: number;
6146
+ };
6147
+ contextWindow: number;
6148
+ maxTokens: number;
6149
+ };
6088
6150
  };
6089
6151
  readonly opencode: {
6090
6152
  readonly "big-pickle": {
@@ -6478,6 +6540,40 @@ export declare const MODELS: {
6478
6540
  contextWindow: number;
6479
6541
  maxTokens: number;
6480
6542
  };
6543
+ readonly "gpt-5.4-mini": {
6544
+ id: string;
6545
+ name: string;
6546
+ api: "openai-responses";
6547
+ provider: string;
6548
+ baseUrl: string;
6549
+ reasoning: true;
6550
+ input: ("image" | "text")[];
6551
+ cost: {
6552
+ input: number;
6553
+ output: number;
6554
+ cacheRead: number;
6555
+ cacheWrite: number;
6556
+ };
6557
+ contextWindow: number;
6558
+ maxTokens: number;
6559
+ };
6560
+ readonly "gpt-5.4-nano": {
6561
+ id: string;
6562
+ name: string;
6563
+ api: "openai-responses";
6564
+ provider: string;
6565
+ baseUrl: string;
6566
+ reasoning: true;
6567
+ input: ("image" | "text")[];
6568
+ cost: {
6569
+ input: number;
6570
+ output: number;
6571
+ cacheRead: number;
6572
+ cacheWrite: number;
6573
+ };
6574
+ contextWindow: number;
6575
+ maxTokens: number;
6576
+ };
6481
6577
  readonly "gpt-5.4-pro": {
6482
6578
  id: string;
6483
6579
  name: string;
@@ -6512,7 +6608,24 @@ export declare const MODELS: {
6512
6608
  contextWindow: number;
6513
6609
  maxTokens: number;
6514
6610
  };
6515
- readonly "mimo-v2-flash-free": {
6611
+ readonly "mimo-v2-omni-free": {
6612
+ id: string;
6613
+ name: string;
6614
+ api: "openai-completions";
6615
+ provider: string;
6616
+ baseUrl: string;
6617
+ reasoning: true;
6618
+ input: ("image" | "text")[];
6619
+ cost: {
6620
+ input: number;
6621
+ output: number;
6622
+ cacheRead: number;
6623
+ cacheWrite: number;
6624
+ };
6625
+ contextWindow: number;
6626
+ maxTokens: number;
6627
+ };
6628
+ readonly "mimo-v2-pro-free": {
6516
6629
  id: string;
6517
6630
  name: string;
6518
6631
  api: "openai-completions";
@@ -6633,6 +6746,23 @@ export declare const MODELS: {
6633
6746
  contextWindow: number;
6634
6747
  maxTokens: number;
6635
6748
  };
6749
+ readonly "minimax-m2.7": {
6750
+ id: string;
6751
+ name: string;
6752
+ api: "anthropic-messages";
6753
+ provider: string;
6754
+ baseUrl: string;
6755
+ reasoning: true;
6756
+ input: "text"[];
6757
+ cost: {
6758
+ input: number;
6759
+ output: number;
6760
+ cacheRead: number;
6761
+ cacheWrite: number;
6762
+ };
6763
+ contextWindow: number;
6764
+ maxTokens: number;
6765
+ };
6636
6766
  };
6637
6767
  readonly openrouter: {
6638
6768
  readonly "ai21/jamba-large-1.7": {
@@ -7587,40 +7717,6 @@ export declare const MODELS: {
7587
7717
  contextWindow: number;
7588
7718
  maxTokens: number;
7589
7719
  };
7590
- readonly "google/gemma-3-27b-it": {
7591
- id: string;
7592
- name: string;
7593
- api: "openai-completions";
7594
- provider: string;
7595
- baseUrl: string;
7596
- reasoning: false;
7597
- input: ("image" | "text")[];
7598
- cost: {
7599
- input: number;
7600
- output: number;
7601
- cacheRead: number;
7602
- cacheWrite: number;
7603
- };
7604
- contextWindow: number;
7605
- maxTokens: number;
7606
- };
7607
- readonly "google/gemma-3-27b-it:free": {
7608
- id: string;
7609
- name: string;
7610
- api: "openai-completions";
7611
- provider: string;
7612
- baseUrl: string;
7613
- reasoning: false;
7614
- input: ("image" | "text")[];
7615
- cost: {
7616
- input: number;
7617
- output: number;
7618
- cacheRead: number;
7619
- cacheWrite: number;
7620
- };
7621
- contextWindow: number;
7622
- maxTokens: number;
7623
- };
7624
7720
  readonly "inception/mercury": {
7625
7721
  id: string;
7626
7722
  name: string;
@@ -7910,6 +8006,23 @@ export declare const MODELS: {
7910
8006
  contextWindow: number;
7911
8007
  maxTokens: number;
7912
8008
  };
8009
+ readonly "minimax/minimax-m2.7": {
8010
+ id: string;
8011
+ name: string;
8012
+ api: "openai-completions";
8013
+ provider: string;
8014
+ baseUrl: string;
8015
+ reasoning: true;
8016
+ input: "text"[];
8017
+ cost: {
8018
+ input: number;
8019
+ output: number;
8020
+ cacheRead: number;
8021
+ cacheWrite: number;
8022
+ };
8023
+ contextWindow: number;
8024
+ maxTokens: number;
8025
+ };
7913
8026
  readonly "mistralai/codestral-2508": {
7914
8027
  id: string;
7915
8028
  name: string;
@@ -8471,6 +8584,23 @@ export declare const MODELS: {
8471
8584
  contextWindow: number;
8472
8585
  maxTokens: number;
8473
8586
  };
8587
+ readonly "nvidia/nemotron-3-super-120b-a12b": {
8588
+ id: string;
8589
+ name: string;
8590
+ api: "openai-completions";
8591
+ provider: string;
8592
+ baseUrl: string;
8593
+ reasoning: true;
8594
+ input: "text"[];
8595
+ cost: {
8596
+ input: number;
8597
+ output: number;
8598
+ cacheRead: number;
8599
+ cacheWrite: number;
8600
+ };
8601
+ contextWindow: number;
8602
+ maxTokens: number;
8603
+ };
8474
8604
  readonly "nvidia/nemotron-3-super-120b-a12b:free": {
8475
8605
  id: string;
8476
8606
  name: string;
@@ -9508,40 +9638,6 @@ export declare const MODELS: {
9508
9638
  contextWindow: number;
9509
9639
  maxTokens: number;
9510
9640
  };
9511
- readonly "openrouter/healer-alpha": {
9512
- id: string;
9513
- name: string;
9514
- api: "openai-completions";
9515
- provider: string;
9516
- baseUrl: string;
9517
- reasoning: true;
9518
- input: ("image" | "text")[];
9519
- cost: {
9520
- input: number;
9521
- output: number;
9522
- cacheRead: number;
9523
- cacheWrite: number;
9524
- };
9525
- contextWindow: number;
9526
- maxTokens: number;
9527
- };
9528
- readonly "openrouter/hunter-alpha": {
9529
- id: string;
9530
- name: string;
9531
- api: "openai-completions";
9532
- provider: string;
9533
- baseUrl: string;
9534
- reasoning: true;
9535
- input: ("image" | "text")[];
9536
- cost: {
9537
- input: number;
9538
- output: number;
9539
- cacheRead: number;
9540
- cacheWrite: number;
9541
- };
9542
- contextWindow: number;
9543
- maxTokens: number;
9544
- };
9545
9641
  readonly "prime-intellect/intellect-3": {
9546
9642
  id: string;
9547
9643
  name: string;
@@ -10630,6 +10726,40 @@ export declare const MODELS: {
10630
10726
  contextWindow: number;
10631
10727
  maxTokens: number;
10632
10728
  };
10729
+ readonly "xiaomi/mimo-v2-omni": {
10730
+ id: string;
10731
+ name: string;
10732
+ api: "openai-completions";
10733
+ provider: string;
10734
+ baseUrl: string;
10735
+ reasoning: true;
10736
+ input: ("image" | "text")[];
10737
+ cost: {
10738
+ input: number;
10739
+ output: number;
10740
+ cacheRead: number;
10741
+ cacheWrite: number;
10742
+ };
10743
+ contextWindow: number;
10744
+ maxTokens: number;
10745
+ };
10746
+ readonly "xiaomi/mimo-v2-pro": {
10747
+ id: string;
10748
+ name: string;
10749
+ api: "openai-completions";
10750
+ provider: string;
10751
+ baseUrl: string;
10752
+ reasoning: true;
10753
+ input: "text"[];
10754
+ cost: {
10755
+ input: number;
10756
+ output: number;
10757
+ cacheRead: number;
10758
+ cacheWrite: number;
10759
+ };
10760
+ contextWindow: number;
10761
+ maxTokens: number;
10762
+ };
10633
10763
  readonly "z-ai/glm-4-32b": {
10634
10764
  id: string;
10635
10765
  name: string;
@@ -11873,6 +12003,40 @@ export declare const MODELS: {
11873
12003
  contextWindow: number;
11874
12004
  maxTokens: number;
11875
12005
  };
12006
+ readonly "minimax/minimax-m2.7": {
12007
+ id: string;
12008
+ name: string;
12009
+ api: "anthropic-messages";
12010
+ provider: string;
12011
+ baseUrl: string;
12012
+ reasoning: true;
12013
+ input: ("image" | "text")[];
12014
+ cost: {
12015
+ input: number;
12016
+ output: number;
12017
+ cacheRead: number;
12018
+ cacheWrite: number;
12019
+ };
12020
+ contextWindow: number;
12021
+ maxTokens: number;
12022
+ };
12023
+ readonly "minimax/minimax-m2.7-highspeed": {
12024
+ id: string;
12025
+ name: string;
12026
+ api: "anthropic-messages";
12027
+ provider: string;
12028
+ baseUrl: string;
12029
+ reasoning: true;
12030
+ input: ("image" | "text")[];
12031
+ cost: {
12032
+ input: number;
12033
+ output: number;
12034
+ cacheRead: number;
12035
+ cacheWrite: number;
12036
+ };
12037
+ contextWindow: number;
12038
+ maxTokens: number;
12039
+ };
11876
12040
  readonly "mistral/codestral": {
11877
12041
  id: string;
11878
12042
  name: string;
@@ -13080,6 +13244,23 @@ export declare const MODELS: {
13080
13244
  contextWindow: number;
13081
13245
  maxTokens: number;
13082
13246
  };
13247
+ readonly "xiaomi/mimo-v2-pro": {
13248
+ id: string;
13249
+ name: string;
13250
+ api: "anthropic-messages";
13251
+ provider: string;
13252
+ baseUrl: string;
13253
+ reasoning: true;
13254
+ input: "text"[];
13255
+ cost: {
13256
+ input: number;
13257
+ output: number;
13258
+ cacheRead: number;
13259
+ cacheWrite: number;
13260
+ };
13261
+ contextWindow: number;
13262
+ maxTokens: number;
13263
+ };
13083
13264
  readonly "zai/glm-4.5": {
13084
13265
  id: string;
13085
13266
  name: string;
@@ -13626,23 +13807,6 @@ export declare const MODELS: {
13626
13807
  contextWindow: number;
13627
13808
  maxTokens: number;
13628
13809
  };
13629
- readonly "grok-4.20-multi-agent-beta-latest": {
13630
- id: string;
13631
- name: string;
13632
- api: "openai-completions";
13633
- provider: string;
13634
- baseUrl: string;
13635
- reasoning: true;
13636
- input: ("image" | "text")[];
13637
- cost: {
13638
- input: number;
13639
- output: number;
13640
- cacheRead: number;
13641
- cacheWrite: number;
13642
- };
13643
- contextWindow: number;
13644
- maxTokens: number;
13645
- };
13646
13810
  readonly "grok-beta": {
13647
13811
  id: string;
13648
13812
  name: string;