@mariozechner/pi-ai 0.53.0 → 0.53.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -357,6 +357,23 @@ export declare const MODELS: {
357
357
  contextWindow: number;
358
358
  maxTokens: number;
359
359
  };
360
+ readonly "anthropic.claude-sonnet-4-6": {
361
+ id: string;
362
+ name: string;
363
+ api: "bedrock-converse-stream";
364
+ provider: string;
365
+ baseUrl: string;
366
+ reasoning: true;
367
+ input: ("image" | "text")[];
368
+ cost: {
369
+ input: number;
370
+ output: number;
371
+ cacheRead: number;
372
+ cacheWrite: number;
373
+ };
374
+ contextWindow: number;
375
+ maxTokens: number;
376
+ };
360
377
  readonly "cohere.command-r-plus-v1:0": {
361
378
  id: string;
362
379
  name: string;
@@ -527,6 +544,23 @@ export declare const MODELS: {
527
544
  contextWindow: number;
528
545
  maxTokens: number;
529
546
  };
547
+ readonly "eu.anthropic.claude-sonnet-4-6": {
548
+ id: string;
549
+ name: string;
550
+ api: "bedrock-converse-stream";
551
+ provider: string;
552
+ baseUrl: string;
553
+ reasoning: true;
554
+ input: ("image" | "text")[];
555
+ cost: {
556
+ input: number;
557
+ output: number;
558
+ cacheRead: number;
559
+ cacheWrite: number;
560
+ };
561
+ contextWindow: number;
562
+ maxTokens: number;
563
+ };
530
564
  readonly "global.anthropic.claude-haiku-4-5-20251001-v1:0": {
531
565
  id: string;
532
566
  name: string;
@@ -612,6 +646,23 @@ export declare const MODELS: {
612
646
  contextWindow: number;
613
647
  maxTokens: number;
614
648
  };
649
+ readonly "global.anthropic.claude-sonnet-4-6": {
650
+ id: string;
651
+ name: string;
652
+ api: "bedrock-converse-stream";
653
+ provider: string;
654
+ baseUrl: string;
655
+ reasoning: true;
656
+ input: ("image" | "text")[];
657
+ cost: {
658
+ input: number;
659
+ output: number;
660
+ cacheRead: number;
661
+ cacheWrite: number;
662
+ };
663
+ contextWindow: number;
664
+ maxTokens: number;
665
+ };
615
666
  readonly "google.gemma-3-27b-it": {
616
667
  id: string;
617
668
  name: string;
@@ -1275,6 +1326,23 @@ export declare const MODELS: {
1275
1326
  contextWindow: number;
1276
1327
  maxTokens: number;
1277
1328
  };
1329
+ readonly "us.anthropic.claude-sonnet-4-6": {
1330
+ id: string;
1331
+ name: string;
1332
+ api: "bedrock-converse-stream";
1333
+ provider: string;
1334
+ baseUrl: string;
1335
+ reasoning: true;
1336
+ input: ("image" | "text")[];
1337
+ cost: {
1338
+ input: number;
1339
+ output: number;
1340
+ cacheRead: number;
1341
+ cacheWrite: number;
1342
+ };
1343
+ contextWindow: number;
1344
+ maxTokens: number;
1345
+ };
1278
1346
  readonly "writer.palmyra-x4-v1:0": {
1279
1347
  id: string;
1280
1348
  name: string;
@@ -2537,6 +2605,29 @@ export declare const MODELS: {
2537
2605
  contextWindow: number;
2538
2606
  maxTokens: number;
2539
2607
  };
2608
+ readonly "claude-sonnet-4.6": {
2609
+ id: string;
2610
+ name: string;
2611
+ api: "anthropic-messages";
2612
+ provider: string;
2613
+ baseUrl: string;
2614
+ headers: {
2615
+ "User-Agent": string;
2616
+ "Editor-Version": string;
2617
+ "Editor-Plugin-Version": string;
2618
+ "Copilot-Integration-Id": string;
2619
+ };
2620
+ reasoning: true;
2621
+ input: ("image" | "text")[];
2622
+ cost: {
2623
+ input: number;
2624
+ output: number;
2625
+ cacheRead: number;
2626
+ cacheWrite: number;
2627
+ };
2628
+ contextWindow: number;
2629
+ maxTokens: number;
2630
+ };
2540
2631
  readonly "gemini-2.5-pro": {
2541
2632
  id: string;
2542
2633
  name: string;
@@ -3180,6 +3271,40 @@ export declare const MODELS: {
3180
3271
  contextWindow: number;
3181
3272
  maxTokens: number;
3182
3273
  };
3274
+ readonly "gemini-3.1-pro-preview": {
3275
+ id: string;
3276
+ name: string;
3277
+ api: "google-generative-ai";
3278
+ provider: string;
3279
+ baseUrl: string;
3280
+ reasoning: true;
3281
+ input: ("image" | "text")[];
3282
+ cost: {
3283
+ input: number;
3284
+ output: number;
3285
+ cacheRead: number;
3286
+ cacheWrite: number;
3287
+ };
3288
+ contextWindow: number;
3289
+ maxTokens: number;
3290
+ };
3291
+ readonly "gemini-3.1-pro-preview-customtools": {
3292
+ id: string;
3293
+ name: string;
3294
+ api: "google-generative-ai";
3295
+ provider: string;
3296
+ baseUrl: string;
3297
+ reasoning: true;
3298
+ input: ("image" | "text")[];
3299
+ cost: {
3300
+ input: number;
3301
+ output: number;
3302
+ cacheRead: number;
3303
+ cacheWrite: number;
3304
+ };
3305
+ contextWindow: number;
3306
+ maxTokens: number;
3307
+ };
3183
3308
  readonly "gemini-flash-latest": {
3184
3309
  id: string;
3185
3310
  name: string;
@@ -3267,6 +3392,23 @@ export declare const MODELS: {
3267
3392
  contextWindow: number;
3268
3393
  maxTokens: number;
3269
3394
  };
3395
+ readonly "claude-opus-4-6-thinking": {
3396
+ id: string;
3397
+ name: string;
3398
+ api: "google-gemini-cli";
3399
+ provider: string;
3400
+ baseUrl: string;
3401
+ reasoning: true;
3402
+ input: ("image" | "text")[];
3403
+ cost: {
3404
+ input: number;
3405
+ output: number;
3406
+ cacheRead: number;
3407
+ cacheWrite: number;
3408
+ };
3409
+ contextWindow: number;
3410
+ maxTokens: number;
3411
+ };
3270
3412
  readonly "claude-sonnet-4-5": {
3271
3413
  id: string;
3272
3414
  name: string;
@@ -3645,6 +3787,23 @@ export declare const MODELS: {
3645
3787
  contextWindow: number;
3646
3788
  maxTokens: number;
3647
3789
  };
3790
+ readonly "gemini-3.1-pro-preview": {
3791
+ id: string;
3792
+ name: string;
3793
+ api: "google-vertex";
3794
+ provider: string;
3795
+ baseUrl: string;
3796
+ reasoning: true;
3797
+ input: ("image" | "text")[];
3798
+ cost: {
3799
+ input: number;
3800
+ output: number;
3801
+ cacheRead: number;
3802
+ cacheWrite: number;
3803
+ };
3804
+ contextWindow: number;
3805
+ maxTokens: number;
3806
+ };
3648
3807
  };
3649
3808
  readonly groq: {
3650
3809
  readonly "deepseek-r1-distill-llama-70b": {
@@ -5607,7 +5766,7 @@ export declare const MODELS: {
5607
5766
  readonly "big-pickle": {
5608
5767
  id: string;
5609
5768
  name: string;
5610
- api: "openai-completions";
5769
+ api: "anthropic-messages";
5611
5770
  provider: string;
5612
5771
  baseUrl: string;
5613
5772
  reasoning: true;
@@ -5740,6 +5899,23 @@ export declare const MODELS: {
5740
5899
  contextWindow: number;
5741
5900
  maxTokens: number;
5742
5901
  };
5902
+ readonly "claude-sonnet-4-6": {
5903
+ id: string;
5904
+ name: string;
5905
+ api: "anthropic-messages";
5906
+ provider: string;
5907
+ baseUrl: string;
5908
+ reasoning: true;
5909
+ input: ("image" | "text")[];
5910
+ cost: {
5911
+ input: number;
5912
+ output: number;
5913
+ cacheRead: number;
5914
+ cacheWrite: number;
5915
+ };
5916
+ contextWindow: number;
5917
+ maxTokens: number;
5918
+ };
5743
5919
  readonly "gemini-3-flash": {
5744
5920
  id: string;
5745
5921
  name: string;
@@ -5774,6 +5950,23 @@ export declare const MODELS: {
5774
5950
  contextWindow: number;
5775
5951
  maxTokens: number;
5776
5952
  };
5953
+ readonly "gemini-3.1-pro": {
5954
+ id: string;
5955
+ name: string;
5956
+ api: "google-generative-ai";
5957
+ provider: string;
5958
+ baseUrl: string;
5959
+ reasoning: true;
5960
+ input: ("image" | "text")[];
5961
+ cost: {
5962
+ input: number;
5963
+ output: number;
5964
+ cacheRead: number;
5965
+ cacheWrite: number;
5966
+ };
5967
+ contextWindow: number;
5968
+ maxTokens: number;
5969
+ };
5777
5970
  readonly "glm-4.6": {
5778
5971
  id: string;
5779
5972
  name: string;
@@ -6046,14 +6239,14 @@ export declare const MODELS: {
6046
6239
  contextWindow: number;
6047
6240
  maxTokens: number;
6048
6241
  };
6049
- readonly "kimi-k2.5-free": {
6242
+ readonly "minimax-m2.1": {
6050
6243
  id: string;
6051
6244
  name: string;
6052
6245
  api: "openai-completions";
6053
6246
  provider: string;
6054
6247
  baseUrl: string;
6055
6248
  reasoning: true;
6056
- input: ("image" | "text")[];
6249
+ input: "text"[];
6057
6250
  cost: {
6058
6251
  input: number;
6059
6252
  output: number;
@@ -6063,7 +6256,7 @@ export declare const MODELS: {
6063
6256
  contextWindow: number;
6064
6257
  maxTokens: number;
6065
6258
  };
6066
- readonly "minimax-m2.1": {
6259
+ readonly "minimax-m2.5": {
6067
6260
  id: string;
6068
6261
  name: string;
6069
6262
  api: "openai-completions";
@@ -6080,10 +6273,10 @@ export declare const MODELS: {
6080
6273
  contextWindow: number;
6081
6274
  maxTokens: number;
6082
6275
  };
6083
- readonly "minimax-m2.5": {
6276
+ readonly "minimax-m2.5-free": {
6084
6277
  id: string;
6085
6278
  name: string;
6086
- api: "openai-completions";
6279
+ api: "anthropic-messages";
6087
6280
  provider: string;
6088
6281
  baseUrl: string;
6089
6282
  reasoning: true;
@@ -6097,13 +6290,13 @@ export declare const MODELS: {
6097
6290
  contextWindow: number;
6098
6291
  maxTokens: number;
6099
6292
  };
6100
- readonly "minimax-m2.5-free": {
6293
+ readonly "trinity-large-preview-free": {
6101
6294
  id: string;
6102
6295
  name: string;
6103
6296
  api: "openai-completions";
6104
6297
  provider: string;
6105
6298
  baseUrl: string;
6106
- reasoning: true;
6299
+ reasoning: false;
6107
6300
  input: "text"[];
6108
6301
  cost: {
6109
6302
  input: number;
@@ -6898,7 +7091,7 @@ export declare const MODELS: {
6898
7091
  contextWindow: number;
6899
7092
  maxTokens: number;
6900
7093
  };
6901
- readonly "google/gemini-2.5-flash-preview-09-2025": {
7094
+ readonly "google/gemini-2.5-pro": {
6902
7095
  id: string;
6903
7096
  name: string;
6904
7097
  api: "openai-completions";
@@ -6915,7 +7108,7 @@ export declare const MODELS: {
6915
7108
  contextWindow: number;
6916
7109
  maxTokens: number;
6917
7110
  };
6918
- readonly "google/gemini-2.5-pro": {
7111
+ readonly "google/gemini-2.5-pro-preview": {
6919
7112
  id: string;
6920
7113
  name: string;
6921
7114
  api: "openai-completions";
@@ -6932,7 +7125,7 @@ export declare const MODELS: {
6932
7125
  contextWindow: number;
6933
7126
  maxTokens: number;
6934
7127
  };
6935
- readonly "google/gemini-2.5-pro-preview": {
7128
+ readonly "google/gemini-2.5-pro-preview-05-06": {
6936
7129
  id: string;
6937
7130
  name: string;
6938
7131
  api: "openai-completions";
@@ -6949,7 +7142,7 @@ export declare const MODELS: {
6949
7142
  contextWindow: number;
6950
7143
  maxTokens: number;
6951
7144
  };
6952
- readonly "google/gemini-2.5-pro-preview-05-06": {
7145
+ readonly "google/gemini-3-flash-preview": {
6953
7146
  id: string;
6954
7147
  name: string;
6955
7148
  api: "openai-completions";
@@ -6966,7 +7159,7 @@ export declare const MODELS: {
6966
7159
  contextWindow: number;
6967
7160
  maxTokens: number;
6968
7161
  };
6969
- readonly "google/gemini-3-flash-preview": {
7162
+ readonly "google/gemini-3-pro-preview": {
6970
7163
  id: string;
6971
7164
  name: string;
6972
7165
  api: "openai-completions";
@@ -6983,7 +7176,7 @@ export declare const MODELS: {
6983
7176
  contextWindow: number;
6984
7177
  maxTokens: number;
6985
7178
  };
6986
- readonly "google/gemini-3-pro-preview": {
7179
+ readonly "google/gemini-3.1-pro-preview": {
6987
7180
  id: string;
6988
7181
  name: string;
6989
7182
  api: "openai-completions";
@@ -7561,23 +7754,6 @@ export declare const MODELS: {
7561
7754
  contextWindow: number;
7562
7755
  maxTokens: number;
7563
7756
  };
7564
- readonly "mistralai/mistral-small-3.1-24b-instruct": {
7565
- id: string;
7566
- name: string;
7567
- api: "openai-completions";
7568
- provider: string;
7569
- baseUrl: string;
7570
- reasoning: false;
7571
- input: ("image" | "text")[];
7572
- cost: {
7573
- input: number;
7574
- output: number;
7575
- cacheRead: number;
7576
- cacheWrite: number;
7577
- };
7578
- contextWindow: number;
7579
- maxTokens: number;
7580
- };
7581
7757
  readonly "mistralai/mistral-small-3.1-24b-instruct:free": {
7582
7758
  id: string;
7583
7759
  name: string;
@@ -7799,40 +7975,6 @@ export declare const MODELS: {
7799
7975
  contextWindow: number;
7800
7976
  maxTokens: number;
7801
7977
  };
7802
- readonly "nousresearch/deephermes-3-mistral-24b-preview": {
7803
- id: string;
7804
- name: string;
7805
- api: "openai-completions";
7806
- provider: string;
7807
- baseUrl: string;
7808
- reasoning: true;
7809
- input: "text"[];
7810
- cost: {
7811
- input: number;
7812
- output: number;
7813
- cacheRead: number;
7814
- cacheWrite: number;
7815
- };
7816
- contextWindow: number;
7817
- maxTokens: number;
7818
- };
7819
- readonly "nousresearch/hermes-4-70b": {
7820
- id: string;
7821
- name: string;
7822
- api: "openai-completions";
7823
- provider: string;
7824
- baseUrl: string;
7825
- reasoning: true;
7826
- input: "text"[];
7827
- cost: {
7828
- input: number;
7829
- output: number;
7830
- cacheRead: number;
7831
- cacheWrite: number;
7832
- };
7833
- contextWindow: number;
7834
- maxTokens: number;
7835
- };
7836
7978
  readonly "nvidia/llama-3.1-nemotron-70b-instruct": {
7837
7979
  id: string;
7838
7980
  name: string;
@@ -9142,23 +9284,6 @@ export declare const MODELS: {
9142
9284
  contextWindow: number;
9143
9285
  maxTokens: number;
9144
9286
  };
9145
- readonly "qwen/qwen3-4b": {
9146
- id: string;
9147
- name: string;
9148
- api: "openai-completions";
9149
- provider: string;
9150
- baseUrl: string;
9151
- reasoning: true;
9152
- input: "text"[];
9153
- cost: {
9154
- input: number;
9155
- output: number;
9156
- cacheRead: number;
9157
- cacheWrite: number;
9158
- };
9159
- contextWindow: number;
9160
- maxTokens: number;
9161
- };
9162
9287
  readonly "qwen/qwen3-4b:free": {
9163
9288
  id: string;
9164
9289
  name: string;
@@ -9703,23 +9828,6 @@ export declare const MODELS: {
9703
9828
  contextWindow: number;
9704
9829
  maxTokens: number;
9705
9830
  };
9706
- readonly "tngtech/tng-r1t-chimera": {
9707
- id: string;
9708
- name: string;
9709
- api: "openai-completions";
9710
- provider: string;
9711
- baseUrl: string;
9712
- reasoning: true;
9713
- input: "text"[];
9714
- cost: {
9715
- input: number;
9716
- output: number;
9717
- cacheRead: number;
9718
- cacheWrite: number;
9719
- };
9720
- contextWindow: number;
9721
- maxTokens: number;
9722
- };
9723
9831
  readonly "upstage/solar-pro-3:free": {
9724
9832
  id: string;
9725
9833
  name: string;
@@ -10776,6 +10884,23 @@ export declare const MODELS: {
10776
10884
  contextWindow: number;
10777
10885
  maxTokens: number;
10778
10886
  };
10887
+ readonly "google/gemini-3.1-pro-preview": {
10888
+ id: string;
10889
+ name: string;
10890
+ api: "anthropic-messages";
10891
+ provider: string;
10892
+ baseUrl: string;
10893
+ reasoning: true;
10894
+ input: ("image" | "text")[];
10895
+ cost: {
10896
+ input: number;
10897
+ output: number;
10898
+ cacheRead: number;
10899
+ cacheWrite: number;
10900
+ };
10901
+ contextWindow: number;
10902
+ maxTokens: number;
10903
+ };
10779
10904
  readonly "inception/mercury-coder-small": {
10780
10905
  id: string;
10781
10906
  name: string;