@mariozechner/pi-ai 0.37.8 → 0.39.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (37) hide show
  1. package/dist/models.generated.d.ts +50 -135
  2. package/dist/models.generated.d.ts.map +1 -1
  3. package/dist/models.generated.js +78 -163
  4. package/dist/models.generated.js.map +1 -1
  5. package/dist/providers/google-gemini-cli.d.ts.map +1 -1
  6. package/dist/providers/google-gemini-cli.js +250 -130
  7. package/dist/providers/google-gemini-cli.js.map +1 -1
  8. package/dist/providers/google-shared.d.ts.map +1 -1
  9. package/dist/providers/google-shared.js +10 -5
  10. package/dist/providers/google-shared.js.map +1 -1
  11. package/dist/providers/openai-codex/prompts/codex.d.ts +2 -2
  12. package/dist/providers/openai-codex/prompts/codex.d.ts.map +1 -1
  13. package/dist/providers/openai-codex/prompts/codex.js +7 -7
  14. package/dist/providers/openai-codex/prompts/codex.js.map +1 -1
  15. package/dist/providers/openai-codex/request-transformer.d.ts +0 -1
  16. package/dist/providers/openai-codex/request-transformer.d.ts.map +1 -1
  17. package/dist/providers/openai-codex/request-transformer.js +12 -130
  18. package/dist/providers/openai-codex/request-transformer.js.map +1 -1
  19. package/dist/providers/openai-codex-responses.d.ts.map +1 -1
  20. package/dist/providers/openai-codex-responses.js +65 -8
  21. package/dist/providers/openai-codex-responses.js.map +1 -1
  22. package/dist/providers/openai-completions.d.ts.map +1 -1
  23. package/dist/providers/openai-completions.js +2 -4
  24. package/dist/providers/openai-completions.js.map +1 -1
  25. package/dist/providers/transorm-messages.d.ts.map +1 -1
  26. package/dist/providers/transorm-messages.js +5 -3
  27. package/dist/providers/transorm-messages.js.map +1 -1
  28. package/dist/stream.d.ts.map +1 -1
  29. package/dist/stream.js +13 -6
  30. package/dist/stream.js.map +1 -1
  31. package/dist/types.d.ts +9 -0
  32. package/dist/types.d.ts.map +1 -1
  33. package/dist/types.js.map +1 -1
  34. package/dist/utils/overflow.d.ts.map +1 -1
  35. package/dist/utils/overflow.js +1 -1
  36. package/dist/utils/overflow.js.map +1 -1
  37. package/package.json +1 -1
@@ -412,6 +412,23 @@ export const MODELS = {
412
412
  contextWindow: 131072,
413
413
  maxTokens: 40960,
414
414
  },
415
+ "zai-glm-4.7": {
416
+ id: "zai-glm-4.7",
417
+ name: "Z.AI GLM-4.7",
418
+ api: "openai-completions",
419
+ provider: "cerebras",
420
+ baseUrl: "https://api.cerebras.ai/v1",
421
+ reasoning: false,
422
+ input: ["text"],
423
+ cost: {
424
+ input: 0,
425
+ output: 0,
426
+ cacheRead: 0,
427
+ cacheWrite: 0,
428
+ },
429
+ contextWindow: 131072,
430
+ maxTokens: 40000,
431
+ },
415
432
  },
416
433
  "github-copilot": {
417
434
  "claude-haiku-4.5": {
@@ -2771,108 +2788,6 @@ export const MODELS = {
2771
2788
  },
2772
2789
  },
2773
2790
  "openai-codex": {
2774
- "codex-mini-latest": {
2775
- id: "codex-mini-latest",
2776
- name: "Codex Mini Latest",
2777
- api: "openai-codex-responses",
2778
- provider: "openai-codex",
2779
- baseUrl: "https://chatgpt.com/backend-api",
2780
- reasoning: true,
2781
- input: ["text", "image"],
2782
- cost: {
2783
- input: 1.5,
2784
- output: 6,
2785
- cacheRead: 0.375,
2786
- cacheWrite: 0,
2787
- },
2788
- contextWindow: 400000,
2789
- maxTokens: 128000,
2790
- },
2791
- "gpt-5": {
2792
- id: "gpt-5",
2793
- name: "gpt-5",
2794
- api: "openai-codex-responses",
2795
- provider: "openai-codex",
2796
- baseUrl: "https://chatgpt.com/backend-api",
2797
- reasoning: true,
2798
- input: ["text", "image"],
2799
- cost: {
2800
- input: 1.25,
2801
- output: 10,
2802
- cacheRead: 0.125,
2803
- cacheWrite: 0,
2804
- },
2805
- contextWindow: 400000,
2806
- maxTokens: 128000,
2807
- },
2808
- "gpt-5-codex": {
2809
- id: "gpt-5-codex",
2810
- name: "gpt-5-codex",
2811
- api: "openai-codex-responses",
2812
- provider: "openai-codex",
2813
- baseUrl: "https://chatgpt.com/backend-api",
2814
- reasoning: true,
2815
- input: ["text", "image"],
2816
- cost: {
2817
- input: 1.25,
2818
- output: 10,
2819
- cacheRead: 0.125,
2820
- cacheWrite: 0,
2821
- },
2822
- contextWindow: 400000,
2823
- maxTokens: 128000,
2824
- },
2825
- "gpt-5-codex-mini": {
2826
- id: "gpt-5-codex-mini",
2827
- name: "gpt-5-codex-mini",
2828
- api: "openai-codex-responses",
2829
- provider: "openai-codex",
2830
- baseUrl: "https://chatgpt.com/backend-api",
2831
- reasoning: true,
2832
- input: ["text", "image"],
2833
- cost: {
2834
- input: 0.25,
2835
- output: 2,
2836
- cacheRead: 0.025,
2837
- cacheWrite: 0,
2838
- },
2839
- contextWindow: 400000,
2840
- maxTokens: 128000,
2841
- },
2842
- "gpt-5-mini": {
2843
- id: "gpt-5-mini",
2844
- name: "gpt-5-mini",
2845
- api: "openai-codex-responses",
2846
- provider: "openai-codex",
2847
- baseUrl: "https://chatgpt.com/backend-api",
2848
- reasoning: true,
2849
- input: ["text", "image"],
2850
- cost: {
2851
- input: 0.25,
2852
- output: 2,
2853
- cacheRead: 0.025,
2854
- cacheWrite: 0,
2855
- },
2856
- contextWindow: 400000,
2857
- maxTokens: 128000,
2858
- },
2859
- "gpt-5-nano": {
2860
- id: "gpt-5-nano",
2861
- name: "gpt-5-nano",
2862
- api: "openai-codex-responses",
2863
- provider: "openai-codex",
2864
- baseUrl: "https://chatgpt.com/backend-api",
2865
- reasoning: true,
2866
- input: ["text", "image"],
2867
- cost: {
2868
- input: 0.05,
2869
- output: 0.4,
2870
- cacheRead: 0.005,
2871
- cacheWrite: 0,
2872
- },
2873
- contextWindow: 400000,
2874
- maxTokens: 128000,
2875
- },
2876
2791
  "gpt-5.1": {
2877
2792
  id: "gpt-5.1",
2878
2793
  name: "GPT-5.1",
@@ -2887,41 +2802,7 @@ export const MODELS = {
2887
2802
  cacheRead: 0.125,
2888
2803
  cacheWrite: 0,
2889
2804
  },
2890
- contextWindow: 400000,
2891
- maxTokens: 128000,
2892
- },
2893
- "gpt-5.1-chat-latest": {
2894
- id: "gpt-5.1-chat-latest",
2895
- name: "gpt-5.1-chat-latest",
2896
- api: "openai-codex-responses",
2897
- provider: "openai-codex",
2898
- baseUrl: "https://chatgpt.com/backend-api",
2899
- reasoning: true,
2900
- input: ["text", "image"],
2901
- cost: {
2902
- input: 1.25,
2903
- output: 10,
2904
- cacheRead: 0.125,
2905
- cacheWrite: 0,
2906
- },
2907
- contextWindow: 400000,
2908
- maxTokens: 128000,
2909
- },
2910
- "gpt-5.1-codex": {
2911
- id: "gpt-5.1-codex",
2912
- name: "GPT-5.1 Codex",
2913
- api: "openai-codex-responses",
2914
- provider: "openai-codex",
2915
- baseUrl: "https://chatgpt.com/backend-api",
2916
- reasoning: true,
2917
- input: ["text", "image"],
2918
- cost: {
2919
- input: 1.25,
2920
- output: 10,
2921
- cacheRead: 0.125,
2922
- cacheWrite: 0,
2923
- },
2924
- contextWindow: 400000,
2805
+ contextWindow: 272000,
2925
2806
  maxTokens: 128000,
2926
2807
  },
2927
2808
  "gpt-5.1-codex-max": {
@@ -2938,7 +2819,7 @@ export const MODELS = {
2938
2819
  cacheRead: 0.125,
2939
2820
  cacheWrite: 0,
2940
2821
  },
2941
- contextWindow: 400000,
2822
+ contextWindow: 272000,
2942
2823
  maxTokens: 128000,
2943
2824
  },
2944
2825
  "gpt-5.1-codex-mini": {
@@ -2955,7 +2836,7 @@ export const MODELS = {
2955
2836
  cacheRead: 0.025,
2956
2837
  cacheWrite: 0,
2957
2838
  },
2958
- contextWindow: 400000,
2839
+ contextWindow: 272000,
2959
2840
  maxTokens: 128000,
2960
2841
  },
2961
2842
  "gpt-5.2": {
@@ -2972,7 +2853,7 @@ export const MODELS = {
2972
2853
  cacheRead: 0.175,
2973
2854
  cacheWrite: 0,
2974
2855
  },
2975
- contextWindow: 400000,
2856
+ contextWindow: 272000,
2976
2857
  maxTokens: 128000,
2977
2858
  },
2978
2859
  "gpt-5.2-codex": {
@@ -2989,7 +2870,7 @@ export const MODELS = {
2989
2870
  cacheRead: 0.175,
2990
2871
  cacheWrite: 0,
2991
2872
  },
2992
- contextWindow: 400000,
2873
+ contextWindow: 272000,
2993
2874
  maxTokens: 128000,
2994
2875
  },
2995
2876
  },
@@ -3062,6 +2943,23 @@ export const MODELS = {
3062
2943
  contextWindow: 65536,
3063
2944
  maxTokens: 65536,
3064
2945
  },
2946
+ "allenai/olmo-3.1-32b-instruct": {
2947
+ id: "allenai/olmo-3.1-32b-instruct",
2948
+ name: "AllenAI: Olmo 3.1 32B Instruct",
2949
+ api: "openai-completions",
2950
+ provider: "openrouter",
2951
+ baseUrl: "https://openrouter.ai/api/v1",
2952
+ reasoning: false,
2953
+ input: ["text"],
2954
+ cost: {
2955
+ input: 0.19999999999999998,
2956
+ output: 0.6,
2957
+ cacheRead: 0,
2958
+ cacheWrite: 0,
2959
+ },
2960
+ contextWindow: 65536,
2961
+ maxTokens: 4096,
2962
+ },
3065
2963
  "amazon/nova-2-lite-v1": {
3066
2964
  id: "amazon/nova-2-lite-v1",
3067
2965
  name: "Amazon: Nova 2 Lite",
@@ -3938,13 +3836,13 @@ export const MODELS = {
3938
3836
  reasoning: false,
3939
3837
  input: ["text", "image"],
3940
3838
  cost: {
3941
- input: 0.036,
3942
- output: 0.064,
3839
+ input: 0.04,
3840
+ output: 0.15,
3943
3841
  cacheRead: 0,
3944
3842
  cacheWrite: 0,
3945
3843
  },
3946
- contextWindow: 131072,
3947
- maxTokens: 4096,
3844
+ contextWindow: 96000,
3845
+ maxTokens: 96000,
3948
3846
  },
3949
3847
  "google/gemma-3-27b-it:free": {
3950
3848
  id: "google/gemma-3-27b-it:free",
@@ -4244,9 +4142,9 @@ export const MODELS = {
4244
4142
  reasoning: true,
4245
4143
  input: ["text"],
4246
4144
  cost: {
4247
- input: 0.12,
4248
- output: 0.48,
4249
- cacheRead: 0,
4145
+ input: 0.28,
4146
+ output: 1.2,
4147
+ cacheRead: 0.14,
4250
4148
  cacheWrite: 0,
4251
4149
  },
4252
4150
  contextWindow: 196608,
@@ -4839,13 +4737,13 @@ export const MODELS = {
4839
4737
  reasoning: true,
4840
4738
  input: ["text"],
4841
4739
  cost: {
4842
- input: 0.32,
4843
- output: 0.48,
4740
+ input: 0.39999999999999997,
4741
+ output: 1.75,
4844
4742
  cacheRead: 0,
4845
4743
  cacheWrite: 0,
4846
4744
  },
4847
4745
  contextWindow: 262144,
4848
- maxTokens: 4096,
4746
+ maxTokens: 65535,
4849
4747
  },
4850
4748
  "nex-agi/deepseek-v3.1-nex-n1:free": {
4851
4749
  id: "nex-agi/deepseek-v3.1-nex-n1:free",
@@ -5621,8 +5519,8 @@ export const MODELS = {
5621
5519
  reasoning: true,
5622
5520
  input: ["text"],
5623
5521
  cost: {
5624
- input: 0.02,
5625
- output: 0.09999999999999999,
5522
+ input: 0.039,
5523
+ output: 0.19,
5626
5524
  cacheRead: 0,
5627
5525
  cacheWrite: 0,
5628
5526
  },
@@ -5672,13 +5570,13 @@ export const MODELS = {
5672
5570
  reasoning: true,
5673
5571
  input: ["text"],
5674
5572
  cost: {
5675
- input: 0.016,
5676
- output: 0.06,
5573
+ input: 0.02,
5574
+ output: 0.09999999999999999,
5677
5575
  cacheRead: 0,
5678
5576
  cacheWrite: 0,
5679
5577
  },
5680
5578
  contextWindow: 131072,
5681
- maxTokens: 4096,
5579
+ maxTokens: 131072,
5682
5580
  },
5683
5581
  "openai/gpt-oss-20b:free": {
5684
5582
  id: "openai/gpt-oss-20b:free",
@@ -5918,6 +5816,23 @@ export const MODELS = {
5918
5816
  contextWindow: 32768,
5919
5817
  maxTokens: 16384,
5920
5818
  },
5819
+ "qwen/qwen-2.5-7b-instruct": {
5820
+ id: "qwen/qwen-2.5-7b-instruct",
5821
+ name: "Qwen: Qwen2.5 7B Instruct",
5822
+ api: "openai-completions",
5823
+ provider: "openrouter",
5824
+ baseUrl: "https://openrouter.ai/api/v1",
5825
+ reasoning: false,
5826
+ input: ["text"],
5827
+ cost: {
5828
+ input: 0.04,
5829
+ output: 0.09999999999999999,
5830
+ cacheRead: 0,
5831
+ cacheWrite: 0,
5832
+ },
5833
+ contextWindow: 32768,
5834
+ maxTokens: 4096,
5835
+ },
5921
5836
  "qwen/qwen-max": {
5922
5837
  id: "qwen/qwen-max",
5923
5838
  name: "Qwen: Qwen-Max ",
@@ -6318,8 +6233,8 @@ export const MODELS = {
6318
6233
  reasoning: false,
6319
6234
  input: ["text"],
6320
6235
  cost: {
6321
- input: 0.06,
6322
- output: 0.6,
6236
+ input: 0.09,
6237
+ output: 1.1,
6323
6238
  cacheRead: 0,
6324
6239
  cacheWrite: 0,
6325
6240
  },
@@ -6352,8 +6267,8 @@ export const MODELS = {
6352
6267
  reasoning: false,
6353
6268
  input: ["text", "image"],
6354
6269
  cost: {
6355
- input: 0.12,
6356
- output: 0.56,
6270
+ input: 0.19999999999999998,
6271
+ output: 1.2,
6357
6272
  cacheRead: 0,
6358
6273
  cacheWrite: 0,
6359
6274
  },
@@ -6913,13 +6828,13 @@ export const MODELS = {
6913
6828
  reasoning: true,
6914
6829
  input: ["text"],
6915
6830
  cost: {
6916
- input: 0.16,
6917
- output: 0.7999999999999999,
6831
+ input: 0.39999999999999997,
6832
+ output: 1.5,
6918
6833
  cacheRead: 0,
6919
6834
  cacheWrite: 0,
6920
6835
  },
6921
6836
  contextWindow: 202752,
6922
- maxTokens: 4096,
6837
+ maxTokens: 65535,
6923
6838
  },
6924
6839
  },
6925
6840
  "xai": {