@mariozechner/pi-ai 0.69.0 → 0.70.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (38) hide show
  1. package/README.md +4 -1
  2. package/dist/env-api-keys.d.ts +9 -0
  3. package/dist/env-api-keys.d.ts.map +1 -1
  4. package/dist/env-api-keys.js +42 -31
  5. package/dist/env-api-keys.js.map +1 -1
  6. package/dist/models.d.ts +1 -1
  7. package/dist/models.d.ts.map +1 -1
  8. package/dist/models.generated.d.ts +282 -19
  9. package/dist/models.generated.d.ts.map +1 -1
  10. package/dist/models.generated.js +278 -47
  11. package/dist/models.generated.js.map +1 -1
  12. package/dist/models.js +5 -2
  13. package/dist/models.js.map +1 -1
  14. package/dist/providers/anthropic.d.ts.map +1 -1
  15. package/dist/providers/anthropic.js +33 -12
  16. package/dist/providers/anthropic.js.map +1 -1
  17. package/dist/providers/azure-openai-responses.d.ts.map +1 -1
  18. package/dist/providers/azure-openai-responses.js +5 -1
  19. package/dist/providers/azure-openai-responses.js.map +1 -1
  20. package/dist/providers/google-vertex.d.ts.map +1 -1
  21. package/dist/providers/google-vertex.js +34 -13
  22. package/dist/providers/google-vertex.js.map +1 -1
  23. package/dist/providers/openai-codex-responses.d.ts.map +1 -1
  24. package/dist/providers/openai-codex-responses.js +8 -7
  25. package/dist/providers/openai-codex-responses.js.map +1 -1
  26. package/dist/providers/openai-completions.d.ts.map +1 -1
  27. package/dist/providers/openai-completions.js +95 -44
  28. package/dist/providers/openai-completions.js.map +1 -1
  29. package/dist/providers/openai-responses.d.ts.map +1 -1
  30. package/dist/providers/openai-responses.js +24 -20
  31. package/dist/providers/openai-responses.js.map +1 -1
  32. package/dist/providers/simple-options.d.ts.map +1 -1
  33. package/dist/providers/simple-options.js +2 -0
  34. package/dist/providers/simple-options.js.map +1 -1
  35. package/dist/types.d.ts +35 -4
  36. package/dist/types.d.ts.map +1 -1
  37. package/dist/types.js.map +1 -1
  38. package/package.json +1 -1
@@ -323,6 +323,40 @@ export declare const MODELS: {
323
323
  contextWindow: number;
324
324
  maxTokens: number;
325
325
  };
326
+ readonly "au.anthropic.claude-opus-4-6-v1": {
327
+ id: string;
328
+ name: string;
329
+ api: "bedrock-converse-stream";
330
+ provider: string;
331
+ baseUrl: string;
332
+ reasoning: true;
333
+ input: ("image" | "text")[];
334
+ cost: {
335
+ input: number;
336
+ output: number;
337
+ cacheRead: number;
338
+ cacheWrite: number;
339
+ };
340
+ contextWindow: number;
341
+ maxTokens: number;
342
+ };
343
+ readonly "au.anthropic.claude-sonnet-4-6": {
344
+ id: string;
345
+ name: string;
346
+ api: "bedrock-converse-stream";
347
+ provider: string;
348
+ baseUrl: string;
349
+ reasoning: true;
350
+ input: ("image" | "text")[];
351
+ cost: {
352
+ input: number;
353
+ output: number;
354
+ cacheRead: number;
355
+ cacheWrite: number;
356
+ };
357
+ contextWindow: number;
358
+ maxTokens: number;
359
+ };
326
360
  readonly "deepseek.r1-v1:0": {
327
361
  id: string;
328
362
  name: string;
@@ -2486,6 +2520,23 @@ export declare const MODELS: {
2486
2520
  contextWindow: number;
2487
2521
  maxTokens: number;
2488
2522
  };
2523
+ readonly "gpt-5.5": {
2524
+ id: string;
2525
+ name: string;
2526
+ api: "azure-openai-responses";
2527
+ provider: string;
2528
+ baseUrl: string;
2529
+ reasoning: true;
2530
+ input: ("image" | "text")[];
2531
+ cost: {
2532
+ input: number;
2533
+ output: number;
2534
+ cacheRead: number;
2535
+ cacheWrite: number;
2536
+ };
2537
+ contextWindow: number;
2538
+ maxTokens: number;
2539
+ };
2489
2540
  readonly o1: {
2490
2541
  id: string;
2491
2542
  name: string;
@@ -2693,6 +2744,64 @@ export declare const MODELS: {
2693
2744
  maxTokens: number;
2694
2745
  };
2695
2746
  };
2747
+ readonly deepseek: {
2748
+ readonly "deepseek-v4-flash": {
2749
+ id: string;
2750
+ name: string;
2751
+ api: "openai-completions";
2752
+ provider: string;
2753
+ baseUrl: string;
2754
+ compat: {
2755
+ requiresReasoningContentOnAssistantMessages: true;
2756
+ thinkingFormat: "deepseek";
2757
+ reasoningEffortMap: {
2758
+ minimal: string;
2759
+ low: string;
2760
+ medium: string;
2761
+ high: string;
2762
+ xhigh: string;
2763
+ };
2764
+ };
2765
+ reasoning: true;
2766
+ input: "text"[];
2767
+ cost: {
2768
+ input: number;
2769
+ output: number;
2770
+ cacheRead: number;
2771
+ cacheWrite: number;
2772
+ };
2773
+ contextWindow: number;
2774
+ maxTokens: number;
2775
+ };
2776
+ readonly "deepseek-v4-pro": {
2777
+ id: string;
2778
+ name: string;
2779
+ api: "openai-completions";
2780
+ provider: string;
2781
+ baseUrl: string;
2782
+ compat: {
2783
+ requiresReasoningContentOnAssistantMessages: true;
2784
+ thinkingFormat: "deepseek";
2785
+ reasoningEffortMap: {
2786
+ minimal: string;
2787
+ low: string;
2788
+ medium: string;
2789
+ high: string;
2790
+ xhigh: string;
2791
+ };
2792
+ };
2793
+ reasoning: true;
2794
+ input: "text"[];
2795
+ cost: {
2796
+ input: number;
2797
+ output: number;
2798
+ cacheRead: number;
2799
+ cacheWrite: number;
2800
+ };
2801
+ contextWindow: number;
2802
+ maxTokens: number;
2803
+ };
2804
+ };
2696
2805
  readonly fireworks: {
2697
2806
  readonly "accounts/fireworks/models/deepseek-v3p1": {
2698
2807
  id: string;
@@ -3014,6 +3123,9 @@ export declare const MODELS: {
3014
3123
  "Editor-Plugin-Version": string;
3015
3124
  "Copilot-Integration-Id": string;
3016
3125
  };
3126
+ compat: {
3127
+ supportsEagerToolInputStreaming: false;
3128
+ };
3017
3129
  reasoning: true;
3018
3130
  input: ("image" | "text")[];
3019
3131
  cost: {
@@ -3106,6 +3218,9 @@ export declare const MODELS: {
3106
3218
  "Editor-Plugin-Version": string;
3107
3219
  "Copilot-Integration-Id": string;
3108
3220
  };
3221
+ compat: {
3222
+ supportsEagerToolInputStreaming: false;
3223
+ };
3109
3224
  reasoning: true;
3110
3225
  input: ("image" | "text")[];
3111
3226
  cost: {
@@ -3129,6 +3244,9 @@ export declare const MODELS: {
3129
3244
  "Editor-Plugin-Version": string;
3130
3245
  "Copilot-Integration-Id": string;
3131
3246
  };
3247
+ compat: {
3248
+ supportsEagerToolInputStreaming: false;
3249
+ };
3132
3250
  reasoning: true;
3133
3251
  input: ("image" | "text")[];
3134
3252
  cost: {
@@ -4039,7 +4157,7 @@ export declare const MODELS: {
4039
4157
  contextWindow: number;
4040
4158
  maxTokens: number;
4041
4159
  };
4042
- readonly "gemma-4-26b-it": {
4160
+ readonly "gemma-4-26b-a4b-it": {
4043
4161
  id: string;
4044
4162
  name: string;
4045
4163
  api: "google-generative-ai";
@@ -5310,6 +5428,9 @@ export declare const MODELS: {
5310
5428
  api: "anthropic-messages";
5311
5429
  provider: string;
5312
5430
  baseUrl: string;
5431
+ headers: {
5432
+ "User-Agent": string;
5433
+ };
5313
5434
  reasoning: true;
5314
5435
  input: ("image" | "text")[];
5315
5436
  cost: {
@@ -5327,6 +5448,9 @@ export declare const MODELS: {
5327
5448
  api: "anthropic-messages";
5328
5449
  provider: string;
5329
5450
  baseUrl: string;
5451
+ headers: {
5452
+ "User-Agent": string;
5453
+ };
5330
5454
  reasoning: true;
5331
5455
  input: ("image" | "text")[];
5332
5456
  cost: {
@@ -5344,6 +5468,9 @@ export declare const MODELS: {
5344
5468
  api: "anthropic-messages";
5345
5469
  provider: string;
5346
5470
  baseUrl: string;
5471
+ headers: {
5472
+ "User-Agent": string;
5473
+ };
5347
5474
  reasoning: true;
5348
5475
  input: "text"[];
5349
5476
  cost: {
@@ -6417,6 +6544,23 @@ export declare const MODELS: {
6417
6544
  contextWindow: number;
6418
6545
  maxTokens: number;
6419
6546
  };
6547
+ readonly "gpt-5.5": {
6548
+ id: string;
6549
+ name: string;
6550
+ api: "openai-responses";
6551
+ provider: string;
6552
+ baseUrl: string;
6553
+ reasoning: true;
6554
+ input: ("image" | "text")[];
6555
+ cost: {
6556
+ input: number;
6557
+ output: number;
6558
+ cacheRead: number;
6559
+ cacheWrite: number;
6560
+ };
6561
+ contextWindow: number;
6562
+ maxTokens: number;
6563
+ };
6420
6564
  readonly o1: {
6421
6565
  id: string;
6422
6566
  name: string;
@@ -6708,6 +6852,23 @@ export declare const MODELS: {
6708
6852
  contextWindow: number;
6709
6853
  maxTokens: number;
6710
6854
  };
6855
+ readonly "gpt-5.5": {
6856
+ id: string;
6857
+ name: string;
6858
+ api: "openai-codex-responses";
6859
+ provider: string;
6860
+ baseUrl: string;
6861
+ reasoning: true;
6862
+ input: ("image" | "text")[];
6863
+ cost: {
6864
+ input: number;
6865
+ output: number;
6866
+ cacheRead: number;
6867
+ cacheWrite: number;
6868
+ };
6869
+ contextWindow: number;
6870
+ maxTokens: number;
6871
+ };
6711
6872
  };
6712
6873
  readonly opencode: {
6713
6874
  readonly "big-pickle": {
@@ -7186,6 +7347,23 @@ export declare const MODELS: {
7186
7347
  contextWindow: number;
7187
7348
  maxTokens: number;
7188
7349
  };
7350
+ readonly "hy3-preview-free": {
7351
+ id: string;
7352
+ name: string;
7353
+ api: "openai-completions";
7354
+ provider: string;
7355
+ baseUrl: string;
7356
+ reasoning: true;
7357
+ input: "text"[];
7358
+ cost: {
7359
+ input: number;
7360
+ output: number;
7361
+ cacheRead: number;
7362
+ cacheWrite: number;
7363
+ };
7364
+ contextWindow: number;
7365
+ maxTokens: number;
7366
+ };
7189
7367
  readonly "kimi-k2.5": {
7190
7368
  id: string;
7191
7369
  name: string;
@@ -7938,23 +8116,6 @@ export declare const MODELS: {
7938
8116
  contextWindow: number;
7939
8117
  maxTokens: number;
7940
8118
  };
7941
- readonly "arcee-ai/trinity-large-preview:free": {
7942
- id: string;
7943
- name: string;
7944
- api: "openai-completions";
7945
- provider: string;
7946
- baseUrl: string;
7947
- reasoning: false;
7948
- input: "text"[];
7949
- cost: {
7950
- input: number;
7951
- output: number;
7952
- cacheRead: number;
7953
- cacheWrite: number;
7954
- };
7955
- contextWindow: number;
7956
- maxTokens: number;
7957
- };
7958
8119
  readonly "arcee-ai/trinity-large-thinking": {
7959
8120
  id: string;
7960
8121
  name: string;
@@ -8295,6 +8456,40 @@ export declare const MODELS: {
8295
8456
  contextWindow: number;
8296
8457
  maxTokens: number;
8297
8458
  };
8459
+ readonly "deepseek/deepseek-v4-flash": {
8460
+ id: string;
8461
+ name: string;
8462
+ api: "openai-completions";
8463
+ provider: string;
8464
+ baseUrl: string;
8465
+ reasoning: true;
8466
+ input: "text"[];
8467
+ cost: {
8468
+ input: number;
8469
+ output: number;
8470
+ cacheRead: number;
8471
+ cacheWrite: number;
8472
+ };
8473
+ contextWindow: number;
8474
+ maxTokens: number;
8475
+ };
8476
+ readonly "deepseek/deepseek-v4-pro": {
8477
+ id: string;
8478
+ name: string;
8479
+ api: "openai-completions";
8480
+ provider: string;
8481
+ baseUrl: string;
8482
+ reasoning: true;
8483
+ input: "text"[];
8484
+ cost: {
8485
+ input: number;
8486
+ output: number;
8487
+ cacheRead: number;
8488
+ cacheWrite: number;
8489
+ };
8490
+ contextWindow: number;
8491
+ maxTokens: number;
8492
+ };
8298
8493
  readonly "essentialai/rnj-1-instruct": {
8299
8494
  id: string;
8300
8495
  name: string;
@@ -8601,6 +8796,23 @@ export declare const MODELS: {
8601
8796
  contextWindow: number;
8602
8797
  maxTokens: number;
8603
8798
  };
8799
+ readonly "inclusionai/ling-2.6-1t:free": {
8800
+ id: string;
8801
+ name: string;
8802
+ api: "openai-completions";
8803
+ provider: string;
8804
+ baseUrl: string;
8805
+ reasoning: false;
8806
+ input: "text"[];
8807
+ cost: {
8808
+ input: number;
8809
+ output: number;
8810
+ cacheRead: number;
8811
+ cacheWrite: number;
8812
+ };
8813
+ contextWindow: number;
8814
+ maxTokens: number;
8815
+ };
8604
8816
  readonly "inclusionai/ling-2.6-flash:free": {
8605
8817
  id: string;
8606
8818
  name: string;
@@ -11270,6 +11482,23 @@ export declare const MODELS: {
11270
11482
  contextWindow: number;
11271
11483
  maxTokens: number;
11272
11484
  };
11485
+ readonly "tencent/hy3-preview:free": {
11486
+ id: string;
11487
+ name: string;
11488
+ api: "openai-completions";
11489
+ provider: string;
11490
+ baseUrl: string;
11491
+ reasoning: true;
11492
+ input: "text"[];
11493
+ cost: {
11494
+ input: number;
11495
+ output: number;
11496
+ cacheRead: number;
11497
+ cacheWrite: number;
11498
+ };
11499
+ contextWindow: number;
11500
+ maxTokens: number;
11501
+ };
11273
11502
  readonly "thedrummer/rocinante-12b": {
11274
11503
  id: string;
11275
11504
  name: string;
@@ -12479,6 +12708,40 @@ export declare const MODELS: {
12479
12708
  contextWindow: number;
12480
12709
  maxTokens: number;
12481
12710
  };
12711
+ readonly "deepseek/deepseek-v4-flash": {
12712
+ id: string;
12713
+ name: string;
12714
+ api: "anthropic-messages";
12715
+ provider: string;
12716
+ baseUrl: string;
12717
+ reasoning: true;
12718
+ input: "text"[];
12719
+ cost: {
12720
+ input: number;
12721
+ output: number;
12722
+ cacheRead: number;
12723
+ cacheWrite: number;
12724
+ };
12725
+ contextWindow: number;
12726
+ maxTokens: number;
12727
+ };
12728
+ readonly "deepseek/deepseek-v4-pro": {
12729
+ id: string;
12730
+ name: string;
12731
+ api: "anthropic-messages";
12732
+ provider: string;
12733
+ baseUrl: string;
12734
+ reasoning: true;
12735
+ input: "text"[];
12736
+ cost: {
12737
+ input: number;
12738
+ output: number;
12739
+ cacheRead: number;
12740
+ cacheWrite: number;
12741
+ };
12742
+ contextWindow: number;
12743
+ maxTokens: number;
12744
+ };
12482
12745
  readonly "google/gemini-2.0-flash": {
12483
12746
  id: string;
12484
12747
  name: string;
@@ -14441,7 +14704,7 @@ export declare const MODELS: {
14441
14704
  provider: string;
14442
14705
  baseUrl: string;
14443
14706
  reasoning: true;
14444
- input: ("image" | "text")[];
14707
+ input: "text"[];
14445
14708
  cost: {
14446
14709
  input: number;
14447
14710
  output: number;