@mariozechner/pi-ai 0.62.0 → 0.63.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -799,6 +799,23 @@ export declare const MODELS: {
799
799
  contextWindow: number;
800
800
  maxTokens: number;
801
801
  };
802
+ readonly "minimax.minimax-m2.5": {
803
+ id: string;
804
+ name: string;
805
+ api: "bedrock-converse-stream";
806
+ provider: string;
807
+ baseUrl: string;
808
+ reasoning: true;
809
+ input: "text"[];
810
+ cost: {
811
+ input: number;
812
+ output: number;
813
+ cacheRead: number;
814
+ cacheWrite: number;
815
+ };
816
+ contextWindow: number;
817
+ maxTokens: number;
818
+ };
802
819
  readonly "mistral.devstral-2-123b": {
803
820
  id: string;
804
821
  name: string;
@@ -1037,6 +1054,23 @@ export declare const MODELS: {
1037
1054
  contextWindow: number;
1038
1055
  maxTokens: number;
1039
1056
  };
1057
+ readonly "nvidia.nemotron-super-3-120b": {
1058
+ id: string;
1059
+ name: string;
1060
+ api: "bedrock-converse-stream";
1061
+ provider: string;
1062
+ baseUrl: string;
1063
+ reasoning: true;
1064
+ input: "text"[];
1065
+ cost: {
1066
+ input: number;
1067
+ output: number;
1068
+ cacheRead: number;
1069
+ cacheWrite: number;
1070
+ };
1071
+ contextWindow: number;
1072
+ maxTokens: number;
1073
+ };
1040
1074
  readonly "openai.gpt-oss-120b-1:0": {
1041
1075
  id: string;
1042
1076
  name: string;
@@ -1411,6 +1445,23 @@ export declare const MODELS: {
1411
1445
  contextWindow: number;
1412
1446
  maxTokens: number;
1413
1447
  };
1448
+ readonly "zai.glm-5": {
1449
+ id: string;
1450
+ name: string;
1451
+ api: "bedrock-converse-stream";
1452
+ provider: string;
1453
+ baseUrl: string;
1454
+ reasoning: true;
1455
+ input: "text"[];
1456
+ cost: {
1457
+ input: number;
1458
+ output: number;
1459
+ cacheRead: number;
1460
+ cacheWrite: number;
1461
+ };
1462
+ contextWindow: number;
1463
+ maxTokens: number;
1464
+ };
1414
1465
  };
1415
1466
  readonly anthropic: {
1416
1467
  readonly "claude-3-5-haiku-20241022": {
@@ -4020,6 +4071,23 @@ export declare const MODELS: {
4020
4071
  contextWindow: number;
4021
4072
  maxTokens: number;
4022
4073
  };
4074
+ readonly "gemini-3.1-pro-preview-customtools": {
4075
+ id: string;
4076
+ name: string;
4077
+ api: "google-vertex";
4078
+ provider: string;
4079
+ baseUrl: string;
4080
+ reasoning: true;
4081
+ input: ("image" | "text")[];
4082
+ cost: {
4083
+ input: number;
4084
+ output: number;
4085
+ cacheRead: number;
4086
+ cacheWrite: number;
4087
+ };
4088
+ contextWindow: number;
4089
+ maxTokens: number;
4090
+ };
4023
4091
  };
4024
4092
  readonly groq: {
4025
4093
  readonly "deepseek-r1-distill-llama-70b": {
@@ -4677,91 +4745,6 @@ export declare const MODELS: {
4677
4745
  };
4678
4746
  };
4679
4747
  readonly minimax: {
4680
- readonly "MiniMax-M2": {
4681
- id: string;
4682
- name: string;
4683
- api: "anthropic-messages";
4684
- provider: string;
4685
- baseUrl: string;
4686
- reasoning: true;
4687
- input: "text"[];
4688
- cost: {
4689
- input: number;
4690
- output: number;
4691
- cacheRead: number;
4692
- cacheWrite: number;
4693
- };
4694
- contextWindow: number;
4695
- maxTokens: number;
4696
- };
4697
- readonly "MiniMax-M2.1": {
4698
- id: string;
4699
- name: string;
4700
- api: "anthropic-messages";
4701
- provider: string;
4702
- baseUrl: string;
4703
- reasoning: true;
4704
- input: "text"[];
4705
- cost: {
4706
- input: number;
4707
- output: number;
4708
- cacheRead: number;
4709
- cacheWrite: number;
4710
- };
4711
- contextWindow: number;
4712
- maxTokens: number;
4713
- };
4714
- readonly "MiniMax-M2.1-highspeed": {
4715
- id: string;
4716
- name: string;
4717
- api: "anthropic-messages";
4718
- provider: string;
4719
- baseUrl: string;
4720
- reasoning: true;
4721
- input: "text"[];
4722
- cost: {
4723
- input: number;
4724
- output: number;
4725
- cacheRead: number;
4726
- cacheWrite: number;
4727
- };
4728
- contextWindow: number;
4729
- maxTokens: number;
4730
- };
4731
- readonly "MiniMax-M2.5": {
4732
- id: string;
4733
- name: string;
4734
- api: "anthropic-messages";
4735
- provider: string;
4736
- baseUrl: string;
4737
- reasoning: true;
4738
- input: "text"[];
4739
- cost: {
4740
- input: number;
4741
- output: number;
4742
- cacheRead: number;
4743
- cacheWrite: number;
4744
- };
4745
- contextWindow: number;
4746
- maxTokens: number;
4747
- };
4748
- readonly "MiniMax-M2.5-highspeed": {
4749
- id: string;
4750
- name: string;
4751
- api: "anthropic-messages";
4752
- provider: string;
4753
- baseUrl: string;
4754
- reasoning: true;
4755
- input: "text"[];
4756
- cost: {
4757
- input: number;
4758
- output: number;
4759
- cacheRead: number;
4760
- cacheWrite: number;
4761
- };
4762
- contextWindow: number;
4763
- maxTokens: number;
4764
- };
4765
4748
  readonly "MiniMax-M2.7": {
4766
4749
  id: string;
4767
4750
  name: string;
@@ -4798,91 +4781,6 @@ export declare const MODELS: {
4798
4781
  };
4799
4782
  };
4800
4783
  readonly "minimax-cn": {
4801
- readonly "MiniMax-M2": {
4802
- id: string;
4803
- name: string;
4804
- api: "anthropic-messages";
4805
- provider: string;
4806
- baseUrl: string;
4807
- reasoning: true;
4808
- input: "text"[];
4809
- cost: {
4810
- input: number;
4811
- output: number;
4812
- cacheRead: number;
4813
- cacheWrite: number;
4814
- };
4815
- contextWindow: number;
4816
- maxTokens: number;
4817
- };
4818
- readonly "MiniMax-M2.1": {
4819
- id: string;
4820
- name: string;
4821
- api: "anthropic-messages";
4822
- provider: string;
4823
- baseUrl: string;
4824
- reasoning: true;
4825
- input: "text"[];
4826
- cost: {
4827
- input: number;
4828
- output: number;
4829
- cacheRead: number;
4830
- cacheWrite: number;
4831
- };
4832
- contextWindow: number;
4833
- maxTokens: number;
4834
- };
4835
- readonly "MiniMax-M2.1-highspeed": {
4836
- id: string;
4837
- name: string;
4838
- api: "anthropic-messages";
4839
- provider: string;
4840
- baseUrl: string;
4841
- reasoning: true;
4842
- input: "text"[];
4843
- cost: {
4844
- input: number;
4845
- output: number;
4846
- cacheRead: number;
4847
- cacheWrite: number;
4848
- };
4849
- contextWindow: number;
4850
- maxTokens: number;
4851
- };
4852
- readonly "MiniMax-M2.5": {
4853
- id: string;
4854
- name: string;
4855
- api: "anthropic-messages";
4856
- provider: string;
4857
- baseUrl: string;
4858
- reasoning: true;
4859
- input: "text"[];
4860
- cost: {
4861
- input: number;
4862
- output: number;
4863
- cacheRead: number;
4864
- cacheWrite: number;
4865
- };
4866
- contextWindow: number;
4867
- maxTokens: number;
4868
- };
4869
- readonly "MiniMax-M2.5-highspeed": {
4870
- id: string;
4871
- name: string;
4872
- api: "anthropic-messages";
4873
- provider: string;
4874
- baseUrl: string;
4875
- reasoning: true;
4876
- input: "text"[];
4877
- cost: {
4878
- input: number;
4879
- output: number;
4880
- cacheRead: number;
4881
- cacheWrite: number;
4882
- };
4883
- contextWindow: number;
4884
- maxTokens: number;
4885
- };
4886
4784
  readonly "MiniMax-M2.7": {
4887
4785
  id: string;
4888
4786
  name: string;
@@ -7683,23 +7581,6 @@ export declare const MODELS: {
7683
7581
  contextWindow: number;
7684
7582
  maxTokens: number;
7685
7583
  };
7686
- readonly "google/gemini-3-pro-preview": {
7687
- id: string;
7688
- name: string;
7689
- api: "openai-completions";
7690
- provider: string;
7691
- baseUrl: string;
7692
- reasoning: true;
7693
- input: ("image" | "text")[];
7694
- cost: {
7695
- input: number;
7696
- output: number;
7697
- cacheRead: number;
7698
- cacheWrite: number;
7699
- };
7700
- contextWindow: number;
7701
- maxTokens: number;
7702
- };
7703
7584
  readonly "google/gemini-3.1-flash-lite-preview": {
7704
7585
  id: string;
7705
7586
  name: string;
@@ -10437,6 +10318,23 @@ export declare const MODELS: {
10437
10318
  contextWindow: number;
10438
10319
  maxTokens: number;
10439
10320
  };
10321
+ readonly "reka/reka-edge": {
10322
+ id: string;
10323
+ name: string;
10324
+ api: "openai-completions";
10325
+ provider: string;
10326
+ baseUrl: string;
10327
+ reasoning: false;
10328
+ input: ("image" | "text")[];
10329
+ cost: {
10330
+ input: number;
10331
+ output: number;
10332
+ cacheRead: number;
10333
+ cacheWrite: number;
10334
+ };
10335
+ contextWindow: number;
10336
+ maxTokens: number;
10337
+ };
10440
10338
  readonly "relace/relace-search": {
10441
10339
  id: string;
10442
10340
  name: string;
@@ -13193,6 +13091,23 @@ export declare const MODELS: {
13193
13091
  contextWindow: number;
13194
13092
  maxTokens: number;
13195
13093
  };
13094
+ readonly "xai/grok-4.20-multi-agent": {
13095
+ id: string;
13096
+ name: string;
13097
+ api: "anthropic-messages";
13098
+ provider: string;
13099
+ baseUrl: string;
13100
+ reasoning: true;
13101
+ input: "text"[];
13102
+ cost: {
13103
+ input: number;
13104
+ output: number;
13105
+ cacheRead: number;
13106
+ cacheWrite: number;
13107
+ };
13108
+ contextWindow: number;
13109
+ maxTokens: number;
13110
+ };
13196
13111
  readonly "xai/grok-4.20-multi-agent-beta": {
13197
13112
  id: string;
13198
13113
  name: string;
@@ -13210,6 +13125,23 @@ export declare const MODELS: {
13210
13125
  contextWindow: number;
13211
13126
  maxTokens: number;
13212
13127
  };
13128
+ readonly "xai/grok-4.20-non-reasoning": {
13129
+ id: string;
13130
+ name: string;
13131
+ api: "anthropic-messages";
13132
+ provider: string;
13133
+ baseUrl: string;
13134
+ reasoning: false;
13135
+ input: ("image" | "text")[];
13136
+ cost: {
13137
+ input: number;
13138
+ output: number;
13139
+ cacheRead: number;
13140
+ cacheWrite: number;
13141
+ };
13142
+ contextWindow: number;
13143
+ maxTokens: number;
13144
+ };
13213
13145
  readonly "xai/grok-4.20-non-reasoning-beta": {
13214
13146
  id: string;
13215
13147
  name: string;
@@ -13227,6 +13159,23 @@ export declare const MODELS: {
13227
13159
  contextWindow: number;
13228
13160
  maxTokens: number;
13229
13161
  };
13162
+ readonly "xai/grok-4.20-reasoning": {
13163
+ id: string;
13164
+ name: string;
13165
+ api: "anthropic-messages";
13166
+ provider: string;
13167
+ baseUrl: string;
13168
+ reasoning: true;
13169
+ input: ("image" | "text")[];
13170
+ cost: {
13171
+ input: number;
13172
+ output: number;
13173
+ cacheRead: number;
13174
+ cacheWrite: number;
13175
+ };
13176
+ contextWindow: number;
13177
+ maxTokens: number;
13178
+ };
13230
13179
  readonly "xai/grok-4.20-reasoning-beta": {
13231
13180
  id: string;
13232
13181
  name: string;
@@ -14062,6 +14011,27 @@ export declare const MODELS: {
14062
14011
  contextWindow: number;
14063
14012
  maxTokens: number;
14064
14013
  };
14014
+ readonly "glm-4.7-flashx": {
14015
+ id: string;
14016
+ name: string;
14017
+ api: "openai-completions";
14018
+ provider: string;
14019
+ baseUrl: string;
14020
+ compat: {
14021
+ supportsDeveloperRole: false;
14022
+ thinkingFormat: "zai";
14023
+ };
14024
+ reasoning: true;
14025
+ input: "text"[];
14026
+ cost: {
14027
+ input: number;
14028
+ output: number;
14029
+ cacheRead: number;
14030
+ cacheWrite: number;
14031
+ };
14032
+ contextWindow: number;
14033
+ maxTokens: number;
14034
+ };
14065
14035
  readonly "glm-5": {
14066
14036
  id: string;
14067
14037
  name: string;