@mariozechner/pi-ai 0.45.6 → 0.46.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1802,6 +1802,29 @@ export declare const MODELS: {
1802
1802
  contextWindow: number;
1803
1803
  maxTokens: number;
1804
1804
  };
1805
+ readonly "gpt-5.2-codex": {
1806
+ id: string;
1807
+ name: string;
1808
+ api: "openai-responses";
1809
+ provider: string;
1810
+ baseUrl: string;
1811
+ headers: {
1812
+ "User-Agent": string;
1813
+ "Editor-Version": string;
1814
+ "Editor-Plugin-Version": string;
1815
+ "Copilot-Integration-Id": string;
1816
+ };
1817
+ reasoning: true;
1818
+ input: ("image" | "text")[];
1819
+ cost: {
1820
+ input: number;
1821
+ output: number;
1822
+ cacheRead: number;
1823
+ cacheWrite: number;
1824
+ };
1825
+ contextWindow: number;
1826
+ maxTokens: number;
1827
+ };
1805
1828
  readonly "grok-code-fast-1": {
1806
1829
  id: string;
1807
1830
  name: string;
@@ -2903,6 +2926,42 @@ export declare const MODELS: {
2903
2926
  maxTokens: number;
2904
2927
  };
2905
2928
  };
2929
+ readonly "minimax-cn": {
2930
+ readonly "MiniMax-M2": {
2931
+ id: string;
2932
+ name: string;
2933
+ api: "anthropic-messages";
2934
+ provider: string;
2935
+ baseUrl: string;
2936
+ reasoning: true;
2937
+ input: "text"[];
2938
+ cost: {
2939
+ input: number;
2940
+ output: number;
2941
+ cacheRead: number;
2942
+ cacheWrite: number;
2943
+ };
2944
+ contextWindow: number;
2945
+ maxTokens: number;
2946
+ };
2947
+ readonly "MiniMax-M2.1": {
2948
+ id: string;
2949
+ name: string;
2950
+ api: "anthropic-messages";
2951
+ provider: string;
2952
+ baseUrl: string;
2953
+ reasoning: true;
2954
+ input: "text"[];
2955
+ cost: {
2956
+ input: number;
2957
+ output: number;
2958
+ cacheRead: number;
2959
+ cacheWrite: number;
2960
+ };
2961
+ contextWindow: number;
2962
+ maxTokens: number;
2963
+ };
2964
+ };
2906
2965
  readonly mistral: {
2907
2966
  readonly "codestral-latest": {
2908
2967
  id: string;
@@ -3739,6 +3798,23 @@ export declare const MODELS: {
3739
3798
  contextWindow: number;
3740
3799
  maxTokens: number;
3741
3800
  };
3801
+ readonly "gpt-5.2-codex": {
3802
+ id: string;
3803
+ name: string;
3804
+ api: "openai-responses";
3805
+ provider: string;
3806
+ baseUrl: string;
3807
+ reasoning: true;
3808
+ input: ("image" | "text")[];
3809
+ cost: {
3810
+ input: number;
3811
+ output: number;
3812
+ cacheRead: number;
3813
+ cacheWrite: number;
3814
+ };
3815
+ contextWindow: number;
3816
+ maxTokens: number;
3817
+ };
3742
3818
  readonly "gpt-5.2-pro": {
3743
3819
  id: string;
3744
3820
  name: string;
@@ -4338,6 +4414,23 @@ export declare const MODELS: {
4338
4414
  contextWindow: number;
4339
4415
  maxTokens: number;
4340
4416
  };
4417
+ readonly "gpt-5.2-codex": {
4418
+ id: string;
4419
+ name: string;
4420
+ api: "openai-responses";
4421
+ provider: string;
4422
+ baseUrl: string;
4423
+ reasoning: true;
4424
+ input: ("image" | "text")[];
4425
+ cost: {
4426
+ input: number;
4427
+ output: number;
4428
+ cacheRead: number;
4429
+ cacheWrite: number;
4430
+ };
4431
+ contextWindow: number;
4432
+ maxTokens: number;
4433
+ };
4341
4434
  readonly "grok-code": {
4342
4435
  id: string;
4343
4436
  name: string;
@@ -4629,23 +4722,6 @@ export declare const MODELS: {
4629
4722
  contextWindow: number;
4630
4723
  maxTokens: number;
4631
4724
  };
4632
- readonly "anthropic/claude-3.5-haiku-20241022": {
4633
- id: string;
4634
- name: string;
4635
- api: "openai-completions";
4636
- provider: string;
4637
- baseUrl: string;
4638
- reasoning: false;
4639
- input: ("image" | "text")[];
4640
- cost: {
4641
- input: number;
4642
- output: number;
4643
- cacheRead: number;
4644
- cacheWrite: number;
4645
- };
4646
- contextWindow: number;
4647
- maxTokens: number;
4648
- };
4649
4725
  readonly "anthropic/claude-3.5-sonnet": {
4650
4726
  id: string;
4651
4727
  name: string;
@@ -5275,23 +5351,6 @@ export declare const MODELS: {
5275
5351
  contextWindow: number;
5276
5352
  maxTokens: number;
5277
5353
  };
5278
- readonly "google/gemini-2.5-flash-preview-09-2025": {
5279
- id: string;
5280
- name: string;
5281
- api: "openai-completions";
5282
- provider: string;
5283
- baseUrl: string;
5284
- reasoning: true;
5285
- input: ("image" | "text")[];
5286
- cost: {
5287
- input: number;
5288
- output: number;
5289
- cacheRead: number;
5290
- cacheWrite: number;
5291
- };
5292
- contextWindow: number;
5293
- maxTokens: number;
5294
- };
5295
5354
  readonly "google/gemini-2.5-pro": {
5296
5355
  id: string;
5297
5356
  name: string;
@@ -5547,23 +5606,6 @@ export declare const MODELS: {
5547
5606
  contextWindow: number;
5548
5607
  maxTokens: number;
5549
5608
  };
5550
- readonly "meta-llama/llama-3.2-3b-instruct": {
5551
- id: string;
5552
- name: string;
5553
- api: "openai-completions";
5554
- provider: string;
5555
- baseUrl: string;
5556
- reasoning: false;
5557
- input: "text"[];
5558
- cost: {
5559
- input: number;
5560
- output: number;
5561
- cacheRead: number;
5562
- cacheWrite: number;
5563
- };
5564
- contextWindow: number;
5565
- maxTokens: number;
5566
- };
5567
5609
  readonly "meta-llama/llama-3.3-70b-instruct": {
5568
5610
  id: string;
5569
5611
  name: string;
@@ -7026,6 +7068,23 @@ export declare const MODELS: {
7026
7068
  contextWindow: number;
7027
7069
  maxTokens: number;
7028
7070
  };
7071
+ readonly "openai/gpt-5.2-codex": {
7072
+ id: string;
7073
+ name: string;
7074
+ api: "openai-completions";
7075
+ provider: string;
7076
+ baseUrl: string;
7077
+ reasoning: true;
7078
+ input: ("image" | "text")[];
7079
+ cost: {
7080
+ input: number;
7081
+ output: number;
7082
+ cacheRead: number;
7083
+ cacheWrite: number;
7084
+ };
7085
+ contextWindow: number;
7086
+ maxTokens: number;
7087
+ };
7029
7088
  readonly "openai/gpt-5.2-pro": {
7030
7089
  id: string;
7031
7090
  name: string;
@@ -8199,6 +8258,23 @@ export declare const MODELS: {
8199
8258
  contextWindow: number;
8200
8259
  maxTokens: number;
8201
8260
  };
8261
+ readonly "xiaomi/mimo-v2-flash": {
8262
+ id: string;
8263
+ name: string;
8264
+ api: "openai-completions";
8265
+ provider: string;
8266
+ baseUrl: string;
8267
+ reasoning: true;
8268
+ input: "text"[];
8269
+ cost: {
8270
+ input: number;
8271
+ output: number;
8272
+ cacheRead: number;
8273
+ cacheWrite: number;
8274
+ };
8275
+ contextWindow: number;
8276
+ maxTokens: number;
8277
+ };
8202
8278
  readonly "xiaomi/mimo-v2-flash:free": {
8203
8279
  id: string;
8204
8280
  name: string;
@@ -9357,23 +9433,6 @@ export declare const MODELS: {
9357
9433
  contextWindow: number;
9358
9434
  maxTokens: number;
9359
9435
  };
9360
- readonly "mistral/mistral-nemo": {
9361
- id: string;
9362
- name: string;
9363
- api: "anthropic-messages";
9364
- provider: string;
9365
- baseUrl: string;
9366
- reasoning: false;
9367
- input: "text"[];
9368
- cost: {
9369
- input: number;
9370
- output: number;
9371
- cacheRead: number;
9372
- cacheWrite: number;
9373
- };
9374
- contextWindow: number;
9375
- maxTokens: number;
9376
- };
9377
9436
  readonly "mistral/mistral-small": {
9378
9437
  id: string;
9379
9438
  name: string;
@@ -9867,6 +9926,23 @@ export declare const MODELS: {
9867
9926
  contextWindow: number;
9868
9927
  maxTokens: number;
9869
9928
  };
9929
+ readonly "openai/gpt-5.2-codex": {
9930
+ id: string;
9931
+ name: string;
9932
+ api: "anthropic-messages";
9933
+ provider: string;
9934
+ baseUrl: string;
9935
+ reasoning: true;
9936
+ input: ("image" | "text")[];
9937
+ cost: {
9938
+ input: number;
9939
+ output: number;
9940
+ cacheRead: number;
9941
+ cacheWrite: number;
9942
+ };
9943
+ contextWindow: number;
9944
+ maxTokens: number;
9945
+ };
9870
9946
  readonly "openai/gpt-5.2-pro": {
9871
9947
  id: string;
9872
9948
  name: string;