@mariozechner/pi-ai 0.53.0 → 0.54.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -357,6 +357,23 @@ export declare const MODELS: {
357
357
  contextWindow: number;
358
358
  maxTokens: number;
359
359
  };
360
+ readonly "anthropic.claude-sonnet-4-6": {
361
+ id: string;
362
+ name: string;
363
+ api: "bedrock-converse-stream";
364
+ provider: string;
365
+ baseUrl: string;
366
+ reasoning: true;
367
+ input: ("image" | "text")[];
368
+ cost: {
369
+ input: number;
370
+ output: number;
371
+ cacheRead: number;
372
+ cacheWrite: number;
373
+ };
374
+ contextWindow: number;
375
+ maxTokens: number;
376
+ };
360
377
  readonly "cohere.command-r-plus-v1:0": {
361
378
  id: string;
362
379
  name: string;
@@ -527,6 +544,23 @@ export declare const MODELS: {
527
544
  contextWindow: number;
528
545
  maxTokens: number;
529
546
  };
547
+ readonly "eu.anthropic.claude-sonnet-4-6": {
548
+ id: string;
549
+ name: string;
550
+ api: "bedrock-converse-stream";
551
+ provider: string;
552
+ baseUrl: string;
553
+ reasoning: true;
554
+ input: ("image" | "text")[];
555
+ cost: {
556
+ input: number;
557
+ output: number;
558
+ cacheRead: number;
559
+ cacheWrite: number;
560
+ };
561
+ contextWindow: number;
562
+ maxTokens: number;
563
+ };
530
564
  readonly "global.anthropic.claude-haiku-4-5-20251001-v1:0": {
531
565
  id: string;
532
566
  name: string;
@@ -612,6 +646,23 @@ export declare const MODELS: {
612
646
  contextWindow: number;
613
647
  maxTokens: number;
614
648
  };
649
+ readonly "global.anthropic.claude-sonnet-4-6": {
650
+ id: string;
651
+ name: string;
652
+ api: "bedrock-converse-stream";
653
+ provider: string;
654
+ baseUrl: string;
655
+ reasoning: true;
656
+ input: ("image" | "text")[];
657
+ cost: {
658
+ input: number;
659
+ output: number;
660
+ cacheRead: number;
661
+ cacheWrite: number;
662
+ };
663
+ contextWindow: number;
664
+ maxTokens: number;
665
+ };
615
666
  readonly "google.gemma-3-27b-it": {
616
667
  id: string;
617
668
  name: string;
@@ -1275,6 +1326,23 @@ export declare const MODELS: {
1275
1326
  contextWindow: number;
1276
1327
  maxTokens: number;
1277
1328
  };
1329
+ readonly "us.anthropic.claude-sonnet-4-6": {
1330
+ id: string;
1331
+ name: string;
1332
+ api: "bedrock-converse-stream";
1333
+ provider: string;
1334
+ baseUrl: string;
1335
+ reasoning: true;
1336
+ input: ("image" | "text")[];
1337
+ cost: {
1338
+ input: number;
1339
+ output: number;
1340
+ cacheRead: number;
1341
+ cacheWrite: number;
1342
+ };
1343
+ contextWindow: number;
1344
+ maxTokens: number;
1345
+ };
1278
1346
  readonly "writer.palmyra-x4-v1:0": {
1279
1347
  id: string;
1280
1348
  name: string;
@@ -2537,6 +2605,29 @@ export declare const MODELS: {
2537
2605
  contextWindow: number;
2538
2606
  maxTokens: number;
2539
2607
  };
2608
+ readonly "claude-sonnet-4.6": {
2609
+ id: string;
2610
+ name: string;
2611
+ api: "anthropic-messages";
2612
+ provider: string;
2613
+ baseUrl: string;
2614
+ headers: {
2615
+ "User-Agent": string;
2616
+ "Editor-Version": string;
2617
+ "Editor-Plugin-Version": string;
2618
+ "Copilot-Integration-Id": string;
2619
+ };
2620
+ reasoning: true;
2621
+ input: ("image" | "text")[];
2622
+ cost: {
2623
+ input: number;
2624
+ output: number;
2625
+ cacheRead: number;
2626
+ cacheWrite: number;
2627
+ };
2628
+ contextWindow: number;
2629
+ maxTokens: number;
2630
+ };
2540
2631
  readonly "gemini-2.5-pro": {
2541
2632
  id: string;
2542
2633
  name: string;
@@ -2621,6 +2712,34 @@ export declare const MODELS: {
2621
2712
  contextWindow: number;
2622
2713
  maxTokens: number;
2623
2714
  };
2715
+ readonly "gemini-3.1-pro-preview": {
2716
+ id: string;
2717
+ name: string;
2718
+ api: "openai-completions";
2719
+ provider: string;
2720
+ baseUrl: string;
2721
+ headers: {
2722
+ "User-Agent": string;
2723
+ "Editor-Version": string;
2724
+ "Editor-Plugin-Version": string;
2725
+ "Copilot-Integration-Id": string;
2726
+ };
2727
+ compat: {
2728
+ supportsStore: false;
2729
+ supportsDeveloperRole: false;
2730
+ supportsReasoningEffort: false;
2731
+ };
2732
+ reasoning: true;
2733
+ input: ("image" | "text")[];
2734
+ cost: {
2735
+ input: number;
2736
+ output: number;
2737
+ cacheRead: number;
2738
+ cacheWrite: number;
2739
+ };
2740
+ contextWindow: number;
2741
+ maxTokens: number;
2742
+ };
2624
2743
  readonly "gpt-4.1": {
2625
2744
  id: string;
2626
2745
  name: string;
@@ -3180,6 +3299,40 @@ export declare const MODELS: {
3180
3299
  contextWindow: number;
3181
3300
  maxTokens: number;
3182
3301
  };
3302
+ readonly "gemini-3.1-pro-preview": {
3303
+ id: string;
3304
+ name: string;
3305
+ api: "google-generative-ai";
3306
+ provider: string;
3307
+ baseUrl: string;
3308
+ reasoning: true;
3309
+ input: ("image" | "text")[];
3310
+ cost: {
3311
+ input: number;
3312
+ output: number;
3313
+ cacheRead: number;
3314
+ cacheWrite: number;
3315
+ };
3316
+ contextWindow: number;
3317
+ maxTokens: number;
3318
+ };
3319
+ readonly "gemini-3.1-pro-preview-customtools": {
3320
+ id: string;
3321
+ name: string;
3322
+ api: "google-generative-ai";
3323
+ provider: string;
3324
+ baseUrl: string;
3325
+ reasoning: true;
3326
+ input: ("image" | "text")[];
3327
+ cost: {
3328
+ input: number;
3329
+ output: number;
3330
+ cacheRead: number;
3331
+ cacheWrite: number;
3332
+ };
3333
+ contextWindow: number;
3334
+ maxTokens: number;
3335
+ };
3183
3336
  readonly "gemini-flash-latest": {
3184
3337
  id: string;
3185
3338
  name: string;
@@ -3267,6 +3420,23 @@ export declare const MODELS: {
3267
3420
  contextWindow: number;
3268
3421
  maxTokens: number;
3269
3422
  };
3423
+ readonly "claude-opus-4-6-thinking": {
3424
+ id: string;
3425
+ name: string;
3426
+ api: "google-gemini-cli";
3427
+ provider: string;
3428
+ baseUrl: string;
3429
+ reasoning: true;
3430
+ input: ("image" | "text")[];
3431
+ cost: {
3432
+ input: number;
3433
+ output: number;
3434
+ cacheRead: number;
3435
+ cacheWrite: number;
3436
+ };
3437
+ contextWindow: number;
3438
+ maxTokens: number;
3439
+ };
3270
3440
  readonly "claude-sonnet-4-5": {
3271
3441
  id: string;
3272
3442
  name: string;
@@ -3645,6 +3815,23 @@ export declare const MODELS: {
3645
3815
  contextWindow: number;
3646
3816
  maxTokens: number;
3647
3817
  };
3818
+ readonly "gemini-3.1-pro-preview": {
3819
+ id: string;
3820
+ name: string;
3821
+ api: "google-vertex";
3822
+ provider: string;
3823
+ baseUrl: string;
3824
+ reasoning: true;
3825
+ input: ("image" | "text")[];
3826
+ cost: {
3827
+ input: number;
3828
+ output: number;
3829
+ cacheRead: number;
3830
+ cacheWrite: number;
3831
+ };
3832
+ contextWindow: number;
3833
+ maxTokens: number;
3834
+ };
3648
3835
  };
3649
3836
  readonly groq: {
3650
3837
  readonly "deepseek-r1-distill-llama-70b": {
@@ -5607,7 +5794,7 @@ export declare const MODELS: {
5607
5794
  readonly "big-pickle": {
5608
5795
  id: string;
5609
5796
  name: string;
5610
- api: "openai-completions";
5797
+ api: "anthropic-messages";
5611
5798
  provider: string;
5612
5799
  baseUrl: string;
5613
5800
  reasoning: true;
@@ -5740,6 +5927,23 @@ export declare const MODELS: {
5740
5927
  contextWindow: number;
5741
5928
  maxTokens: number;
5742
5929
  };
5930
+ readonly "claude-sonnet-4-6": {
5931
+ id: string;
5932
+ name: string;
5933
+ api: "anthropic-messages";
5934
+ provider: string;
5935
+ baseUrl: string;
5936
+ reasoning: true;
5937
+ input: ("image" | "text")[];
5938
+ cost: {
5939
+ input: number;
5940
+ output: number;
5941
+ cacheRead: number;
5942
+ cacheWrite: number;
5943
+ };
5944
+ contextWindow: number;
5945
+ maxTokens: number;
5946
+ };
5743
5947
  readonly "gemini-3-flash": {
5744
5948
  id: string;
5745
5949
  name: string;
@@ -5774,6 +5978,23 @@ export declare const MODELS: {
5774
5978
  contextWindow: number;
5775
5979
  maxTokens: number;
5776
5980
  };
5981
+ readonly "gemini-3.1-pro": {
5982
+ id: string;
5983
+ name: string;
5984
+ api: "google-generative-ai";
5985
+ provider: string;
5986
+ baseUrl: string;
5987
+ reasoning: true;
5988
+ input: ("image" | "text")[];
5989
+ cost: {
5990
+ input: number;
5991
+ output: number;
5992
+ cacheRead: number;
5993
+ cacheWrite: number;
5994
+ };
5995
+ contextWindow: number;
5996
+ maxTokens: number;
5997
+ };
5777
5998
  readonly "glm-4.6": {
5778
5999
  id: string;
5779
6000
  name: string;
@@ -6046,14 +6267,14 @@ export declare const MODELS: {
6046
6267
  contextWindow: number;
6047
6268
  maxTokens: number;
6048
6269
  };
6049
- readonly "kimi-k2.5-free": {
6270
+ readonly "minimax-m2.1": {
6050
6271
  id: string;
6051
6272
  name: string;
6052
6273
  api: "openai-completions";
6053
6274
  provider: string;
6054
6275
  baseUrl: string;
6055
6276
  reasoning: true;
6056
- input: ("image" | "text")[];
6277
+ input: "text"[];
6057
6278
  cost: {
6058
6279
  input: number;
6059
6280
  output: number;
@@ -6063,7 +6284,7 @@ export declare const MODELS: {
6063
6284
  contextWindow: number;
6064
6285
  maxTokens: number;
6065
6286
  };
6066
- readonly "minimax-m2.1": {
6287
+ readonly "minimax-m2.5": {
6067
6288
  id: string;
6068
6289
  name: string;
6069
6290
  api: "openai-completions";
@@ -6080,10 +6301,10 @@ export declare const MODELS: {
6080
6301
  contextWindow: number;
6081
6302
  maxTokens: number;
6082
6303
  };
6083
- readonly "minimax-m2.5": {
6304
+ readonly "minimax-m2.5-free": {
6084
6305
  id: string;
6085
6306
  name: string;
6086
- api: "openai-completions";
6307
+ api: "anthropic-messages";
6087
6308
  provider: string;
6088
6309
  baseUrl: string;
6089
6310
  reasoning: true;
@@ -6097,13 +6318,13 @@ export declare const MODELS: {
6097
6318
  contextWindow: number;
6098
6319
  maxTokens: number;
6099
6320
  };
6100
- readonly "minimax-m2.5-free": {
6321
+ readonly "trinity-large-preview-free": {
6101
6322
  id: string;
6102
6323
  name: string;
6103
6324
  api: "openai-completions";
6104
6325
  provider: string;
6105
6326
  baseUrl: string;
6106
- reasoning: true;
6327
+ reasoning: false;
6107
6328
  input: "text"[];
6108
6329
  cost: {
6109
6330
  input: number;
@@ -6898,7 +7119,7 @@ export declare const MODELS: {
6898
7119
  contextWindow: number;
6899
7120
  maxTokens: number;
6900
7121
  };
6901
- readonly "google/gemini-2.5-flash-preview-09-2025": {
7122
+ readonly "google/gemini-2.5-pro": {
6902
7123
  id: string;
6903
7124
  name: string;
6904
7125
  api: "openai-completions";
@@ -6915,7 +7136,7 @@ export declare const MODELS: {
6915
7136
  contextWindow: number;
6916
7137
  maxTokens: number;
6917
7138
  };
6918
- readonly "google/gemini-2.5-pro": {
7139
+ readonly "google/gemini-2.5-pro-preview": {
6919
7140
  id: string;
6920
7141
  name: string;
6921
7142
  api: "openai-completions";
@@ -6932,7 +7153,7 @@ export declare const MODELS: {
6932
7153
  contextWindow: number;
6933
7154
  maxTokens: number;
6934
7155
  };
6935
- readonly "google/gemini-2.5-pro-preview": {
7156
+ readonly "google/gemini-2.5-pro-preview-05-06": {
6936
7157
  id: string;
6937
7158
  name: string;
6938
7159
  api: "openai-completions";
@@ -6949,7 +7170,7 @@ export declare const MODELS: {
6949
7170
  contextWindow: number;
6950
7171
  maxTokens: number;
6951
7172
  };
6952
- readonly "google/gemini-2.5-pro-preview-05-06": {
7173
+ readonly "google/gemini-3-flash-preview": {
6953
7174
  id: string;
6954
7175
  name: string;
6955
7176
  api: "openai-completions";
@@ -6966,7 +7187,7 @@ export declare const MODELS: {
6966
7187
  contextWindow: number;
6967
7188
  maxTokens: number;
6968
7189
  };
6969
- readonly "google/gemini-3-flash-preview": {
7190
+ readonly "google/gemini-3-pro-preview": {
6970
7191
  id: string;
6971
7192
  name: string;
6972
7193
  api: "openai-completions";
@@ -6983,7 +7204,7 @@ export declare const MODELS: {
6983
7204
  contextWindow: number;
6984
7205
  maxTokens: number;
6985
7206
  };
6986
- readonly "google/gemini-3-pro-preview": {
7207
+ readonly "google/gemini-3.1-pro-preview": {
6987
7208
  id: string;
6988
7209
  name: string;
6989
7210
  api: "openai-completions";
@@ -7561,23 +7782,6 @@ export declare const MODELS: {
7561
7782
  contextWindow: number;
7562
7783
  maxTokens: number;
7563
7784
  };
7564
- readonly "mistralai/mistral-small-3.1-24b-instruct": {
7565
- id: string;
7566
- name: string;
7567
- api: "openai-completions";
7568
- provider: string;
7569
- baseUrl: string;
7570
- reasoning: false;
7571
- input: ("image" | "text")[];
7572
- cost: {
7573
- input: number;
7574
- output: number;
7575
- cacheRead: number;
7576
- cacheWrite: number;
7577
- };
7578
- contextWindow: number;
7579
- maxTokens: number;
7580
- };
7581
7785
  readonly "mistralai/mistral-small-3.1-24b-instruct:free": {
7582
7786
  id: string;
7583
7787
  name: string;
@@ -7799,40 +8003,6 @@ export declare const MODELS: {
7799
8003
  contextWindow: number;
7800
8004
  maxTokens: number;
7801
8005
  };
7802
- readonly "nousresearch/deephermes-3-mistral-24b-preview": {
7803
- id: string;
7804
- name: string;
7805
- api: "openai-completions";
7806
- provider: string;
7807
- baseUrl: string;
7808
- reasoning: true;
7809
- input: "text"[];
7810
- cost: {
7811
- input: number;
7812
- output: number;
7813
- cacheRead: number;
7814
- cacheWrite: number;
7815
- };
7816
- contextWindow: number;
7817
- maxTokens: number;
7818
- };
7819
- readonly "nousresearch/hermes-4-70b": {
7820
- id: string;
7821
- name: string;
7822
- api: "openai-completions";
7823
- provider: string;
7824
- baseUrl: string;
7825
- reasoning: true;
7826
- input: "text"[];
7827
- cost: {
7828
- input: number;
7829
- output: number;
7830
- cacheRead: number;
7831
- cacheWrite: number;
7832
- };
7833
- contextWindow: number;
7834
- maxTokens: number;
7835
- };
7836
8006
  readonly "nvidia/llama-3.1-nemotron-70b-instruct": {
7837
8007
  id: string;
7838
8008
  name: string;
@@ -8802,23 +8972,6 @@ export declare const MODELS: {
8802
8972
  contextWindow: number;
8803
8973
  maxTokens: number;
8804
8974
  };
8805
- readonly "openrouter/aurora-alpha": {
8806
- id: string;
8807
- name: string;
8808
- api: "openai-completions";
8809
- provider: string;
8810
- baseUrl: string;
8811
- reasoning: true;
8812
- input: "text"[];
8813
- cost: {
8814
- input: number;
8815
- output: number;
8816
- cacheRead: number;
8817
- cacheWrite: number;
8818
- };
8819
- contextWindow: number;
8820
- maxTokens: number;
8821
- };
8822
8975
  readonly "openrouter/auto": {
8823
8976
  id: string;
8824
8977
  name: string;
@@ -9142,23 +9295,6 @@ export declare const MODELS: {
9142
9295
  contextWindow: number;
9143
9296
  maxTokens: number;
9144
9297
  };
9145
- readonly "qwen/qwen3-4b": {
9146
- id: string;
9147
- name: string;
9148
- api: "openai-completions";
9149
- provider: string;
9150
- baseUrl: string;
9151
- reasoning: true;
9152
- input: "text"[];
9153
- cost: {
9154
- input: number;
9155
- output: number;
9156
- cacheRead: number;
9157
- cacheWrite: number;
9158
- };
9159
- contextWindow: number;
9160
- maxTokens: number;
9161
- };
9162
9298
  readonly "qwen/qwen3-4b:free": {
9163
9299
  id: string;
9164
9300
  name: string;
@@ -9703,23 +9839,6 @@ export declare const MODELS: {
9703
9839
  contextWindow: number;
9704
9840
  maxTokens: number;
9705
9841
  };
9706
- readonly "tngtech/tng-r1t-chimera": {
9707
- id: string;
9708
- name: string;
9709
- api: "openai-completions";
9710
- provider: string;
9711
- baseUrl: string;
9712
- reasoning: true;
9713
- input: "text"[];
9714
- cost: {
9715
- input: number;
9716
- output: number;
9717
- cacheRead: number;
9718
- cacheWrite: number;
9719
- };
9720
- contextWindow: number;
9721
- maxTokens: number;
9722
- };
9723
9842
  readonly "upstage/solar-pro-3:free": {
9724
9843
  id: string;
9725
9844
  name: string;
@@ -10776,6 +10895,23 @@ export declare const MODELS: {
10776
10895
  contextWindow: number;
10777
10896
  maxTokens: number;
10778
10897
  };
10898
+ readonly "google/gemini-3.1-pro-preview": {
10899
+ id: string;
10900
+ name: string;
10901
+ api: "anthropic-messages";
10902
+ provider: string;
10903
+ baseUrl: string;
10904
+ reasoning: true;
10905
+ input: ("image" | "text")[];
10906
+ cost: {
10907
+ input: number;
10908
+ output: number;
10909
+ cacheRead: number;
10910
+ cacheWrite: number;
10911
+ };
10912
+ contextWindow: number;
10913
+ maxTokens: number;
10914
+ };
10779
10915
  readonly "inception/mercury-coder-small": {
10780
10916
  id: string;
10781
10917
  name: string;