@mariozechner/pi-ai 0.37.8 → 0.39.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/models.generated.d.ts +50 -135
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +78 -163
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/google-gemini-cli.d.ts.map +1 -1
- package/dist/providers/google-gemini-cli.js +250 -130
- package/dist/providers/google-gemini-cli.js.map +1 -1
- package/dist/providers/google-shared.d.ts.map +1 -1
- package/dist/providers/google-shared.js +10 -5
- package/dist/providers/google-shared.js.map +1 -1
- package/dist/providers/openai-codex/prompts/codex.d.ts +2 -2
- package/dist/providers/openai-codex/prompts/codex.d.ts.map +1 -1
- package/dist/providers/openai-codex/prompts/codex.js +7 -7
- package/dist/providers/openai-codex/prompts/codex.js.map +1 -1
- package/dist/providers/openai-codex/request-transformer.d.ts +0 -1
- package/dist/providers/openai-codex/request-transformer.d.ts.map +1 -1
- package/dist/providers/openai-codex/request-transformer.js +12 -130
- package/dist/providers/openai-codex/request-transformer.js.map +1 -1
- package/dist/providers/openai-codex-responses.d.ts.map +1 -1
- package/dist/providers/openai-codex-responses.js +65 -8
- package/dist/providers/openai-codex-responses.js.map +1 -1
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +2 -4
- package/dist/providers/openai-completions.js.map +1 -1
- package/dist/providers/transorm-messages.d.ts.map +1 -1
- package/dist/providers/transorm-messages.js +5 -3
- package/dist/providers/transorm-messages.js.map +1 -1
- package/dist/stream.d.ts.map +1 -1
- package/dist/stream.js +13 -6
- package/dist/stream.js.map +1 -1
- package/dist/types.d.ts +9 -0
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/dist/utils/overflow.d.ts.map +1 -1
- package/dist/utils/overflow.js +1 -1
- package/dist/utils/overflow.js.map +1 -1
- package/package.json +1 -1
package/dist/models.generated.js
CHANGED
|
@@ -412,6 +412,23 @@ export const MODELS = {
|
|
|
412
412
|
contextWindow: 131072,
|
|
413
413
|
maxTokens: 40960,
|
|
414
414
|
},
|
|
415
|
+
"zai-glm-4.7": {
|
|
416
|
+
id: "zai-glm-4.7",
|
|
417
|
+
name: "Z.AI GLM-4.7",
|
|
418
|
+
api: "openai-completions",
|
|
419
|
+
provider: "cerebras",
|
|
420
|
+
baseUrl: "https://api.cerebras.ai/v1",
|
|
421
|
+
reasoning: false,
|
|
422
|
+
input: ["text"],
|
|
423
|
+
cost: {
|
|
424
|
+
input: 0,
|
|
425
|
+
output: 0,
|
|
426
|
+
cacheRead: 0,
|
|
427
|
+
cacheWrite: 0,
|
|
428
|
+
},
|
|
429
|
+
contextWindow: 131072,
|
|
430
|
+
maxTokens: 40000,
|
|
431
|
+
},
|
|
415
432
|
},
|
|
416
433
|
"github-copilot": {
|
|
417
434
|
"claude-haiku-4.5": {
|
|
@@ -2771,108 +2788,6 @@ export const MODELS = {
|
|
|
2771
2788
|
},
|
|
2772
2789
|
},
|
|
2773
2790
|
"openai-codex": {
|
|
2774
|
-
"codex-mini-latest": {
|
|
2775
|
-
id: "codex-mini-latest",
|
|
2776
|
-
name: "Codex Mini Latest",
|
|
2777
|
-
api: "openai-codex-responses",
|
|
2778
|
-
provider: "openai-codex",
|
|
2779
|
-
baseUrl: "https://chatgpt.com/backend-api",
|
|
2780
|
-
reasoning: true,
|
|
2781
|
-
input: ["text", "image"],
|
|
2782
|
-
cost: {
|
|
2783
|
-
input: 1.5,
|
|
2784
|
-
output: 6,
|
|
2785
|
-
cacheRead: 0.375,
|
|
2786
|
-
cacheWrite: 0,
|
|
2787
|
-
},
|
|
2788
|
-
contextWindow: 400000,
|
|
2789
|
-
maxTokens: 128000,
|
|
2790
|
-
},
|
|
2791
|
-
"gpt-5": {
|
|
2792
|
-
id: "gpt-5",
|
|
2793
|
-
name: "gpt-5",
|
|
2794
|
-
api: "openai-codex-responses",
|
|
2795
|
-
provider: "openai-codex",
|
|
2796
|
-
baseUrl: "https://chatgpt.com/backend-api",
|
|
2797
|
-
reasoning: true,
|
|
2798
|
-
input: ["text", "image"],
|
|
2799
|
-
cost: {
|
|
2800
|
-
input: 1.25,
|
|
2801
|
-
output: 10,
|
|
2802
|
-
cacheRead: 0.125,
|
|
2803
|
-
cacheWrite: 0,
|
|
2804
|
-
},
|
|
2805
|
-
contextWindow: 400000,
|
|
2806
|
-
maxTokens: 128000,
|
|
2807
|
-
},
|
|
2808
|
-
"gpt-5-codex": {
|
|
2809
|
-
id: "gpt-5-codex",
|
|
2810
|
-
name: "gpt-5-codex",
|
|
2811
|
-
api: "openai-codex-responses",
|
|
2812
|
-
provider: "openai-codex",
|
|
2813
|
-
baseUrl: "https://chatgpt.com/backend-api",
|
|
2814
|
-
reasoning: true,
|
|
2815
|
-
input: ["text", "image"],
|
|
2816
|
-
cost: {
|
|
2817
|
-
input: 1.25,
|
|
2818
|
-
output: 10,
|
|
2819
|
-
cacheRead: 0.125,
|
|
2820
|
-
cacheWrite: 0,
|
|
2821
|
-
},
|
|
2822
|
-
contextWindow: 400000,
|
|
2823
|
-
maxTokens: 128000,
|
|
2824
|
-
},
|
|
2825
|
-
"gpt-5-codex-mini": {
|
|
2826
|
-
id: "gpt-5-codex-mini",
|
|
2827
|
-
name: "gpt-5-codex-mini",
|
|
2828
|
-
api: "openai-codex-responses",
|
|
2829
|
-
provider: "openai-codex",
|
|
2830
|
-
baseUrl: "https://chatgpt.com/backend-api",
|
|
2831
|
-
reasoning: true,
|
|
2832
|
-
input: ["text", "image"],
|
|
2833
|
-
cost: {
|
|
2834
|
-
input: 0.25,
|
|
2835
|
-
output: 2,
|
|
2836
|
-
cacheRead: 0.025,
|
|
2837
|
-
cacheWrite: 0,
|
|
2838
|
-
},
|
|
2839
|
-
contextWindow: 400000,
|
|
2840
|
-
maxTokens: 128000,
|
|
2841
|
-
},
|
|
2842
|
-
"gpt-5-mini": {
|
|
2843
|
-
id: "gpt-5-mini",
|
|
2844
|
-
name: "gpt-5-mini",
|
|
2845
|
-
api: "openai-codex-responses",
|
|
2846
|
-
provider: "openai-codex",
|
|
2847
|
-
baseUrl: "https://chatgpt.com/backend-api",
|
|
2848
|
-
reasoning: true,
|
|
2849
|
-
input: ["text", "image"],
|
|
2850
|
-
cost: {
|
|
2851
|
-
input: 0.25,
|
|
2852
|
-
output: 2,
|
|
2853
|
-
cacheRead: 0.025,
|
|
2854
|
-
cacheWrite: 0,
|
|
2855
|
-
},
|
|
2856
|
-
contextWindow: 400000,
|
|
2857
|
-
maxTokens: 128000,
|
|
2858
|
-
},
|
|
2859
|
-
"gpt-5-nano": {
|
|
2860
|
-
id: "gpt-5-nano",
|
|
2861
|
-
name: "gpt-5-nano",
|
|
2862
|
-
api: "openai-codex-responses",
|
|
2863
|
-
provider: "openai-codex",
|
|
2864
|
-
baseUrl: "https://chatgpt.com/backend-api",
|
|
2865
|
-
reasoning: true,
|
|
2866
|
-
input: ["text", "image"],
|
|
2867
|
-
cost: {
|
|
2868
|
-
input: 0.05,
|
|
2869
|
-
output: 0.4,
|
|
2870
|
-
cacheRead: 0.005,
|
|
2871
|
-
cacheWrite: 0,
|
|
2872
|
-
},
|
|
2873
|
-
contextWindow: 400000,
|
|
2874
|
-
maxTokens: 128000,
|
|
2875
|
-
},
|
|
2876
2791
|
"gpt-5.1": {
|
|
2877
2792
|
id: "gpt-5.1",
|
|
2878
2793
|
name: "GPT-5.1",
|
|
@@ -2887,41 +2802,7 @@ export const MODELS = {
|
|
|
2887
2802
|
cacheRead: 0.125,
|
|
2888
2803
|
cacheWrite: 0,
|
|
2889
2804
|
},
|
|
2890
|
-
contextWindow:
|
|
2891
|
-
maxTokens: 128000,
|
|
2892
|
-
},
|
|
2893
|
-
"gpt-5.1-chat-latest": {
|
|
2894
|
-
id: "gpt-5.1-chat-latest",
|
|
2895
|
-
name: "gpt-5.1-chat-latest",
|
|
2896
|
-
api: "openai-codex-responses",
|
|
2897
|
-
provider: "openai-codex",
|
|
2898
|
-
baseUrl: "https://chatgpt.com/backend-api",
|
|
2899
|
-
reasoning: true,
|
|
2900
|
-
input: ["text", "image"],
|
|
2901
|
-
cost: {
|
|
2902
|
-
input: 1.25,
|
|
2903
|
-
output: 10,
|
|
2904
|
-
cacheRead: 0.125,
|
|
2905
|
-
cacheWrite: 0,
|
|
2906
|
-
},
|
|
2907
|
-
contextWindow: 400000,
|
|
2908
|
-
maxTokens: 128000,
|
|
2909
|
-
},
|
|
2910
|
-
"gpt-5.1-codex": {
|
|
2911
|
-
id: "gpt-5.1-codex",
|
|
2912
|
-
name: "GPT-5.1 Codex",
|
|
2913
|
-
api: "openai-codex-responses",
|
|
2914
|
-
provider: "openai-codex",
|
|
2915
|
-
baseUrl: "https://chatgpt.com/backend-api",
|
|
2916
|
-
reasoning: true,
|
|
2917
|
-
input: ["text", "image"],
|
|
2918
|
-
cost: {
|
|
2919
|
-
input: 1.25,
|
|
2920
|
-
output: 10,
|
|
2921
|
-
cacheRead: 0.125,
|
|
2922
|
-
cacheWrite: 0,
|
|
2923
|
-
},
|
|
2924
|
-
contextWindow: 400000,
|
|
2805
|
+
contextWindow: 272000,
|
|
2925
2806
|
maxTokens: 128000,
|
|
2926
2807
|
},
|
|
2927
2808
|
"gpt-5.1-codex-max": {
|
|
@@ -2938,7 +2819,7 @@ export const MODELS = {
|
|
|
2938
2819
|
cacheRead: 0.125,
|
|
2939
2820
|
cacheWrite: 0,
|
|
2940
2821
|
},
|
|
2941
|
-
contextWindow:
|
|
2822
|
+
contextWindow: 272000,
|
|
2942
2823
|
maxTokens: 128000,
|
|
2943
2824
|
},
|
|
2944
2825
|
"gpt-5.1-codex-mini": {
|
|
@@ -2955,7 +2836,7 @@ export const MODELS = {
|
|
|
2955
2836
|
cacheRead: 0.025,
|
|
2956
2837
|
cacheWrite: 0,
|
|
2957
2838
|
},
|
|
2958
|
-
contextWindow:
|
|
2839
|
+
contextWindow: 272000,
|
|
2959
2840
|
maxTokens: 128000,
|
|
2960
2841
|
},
|
|
2961
2842
|
"gpt-5.2": {
|
|
@@ -2972,7 +2853,7 @@ export const MODELS = {
|
|
|
2972
2853
|
cacheRead: 0.175,
|
|
2973
2854
|
cacheWrite: 0,
|
|
2974
2855
|
},
|
|
2975
|
-
contextWindow:
|
|
2856
|
+
contextWindow: 272000,
|
|
2976
2857
|
maxTokens: 128000,
|
|
2977
2858
|
},
|
|
2978
2859
|
"gpt-5.2-codex": {
|
|
@@ -2989,7 +2870,7 @@ export const MODELS = {
|
|
|
2989
2870
|
cacheRead: 0.175,
|
|
2990
2871
|
cacheWrite: 0,
|
|
2991
2872
|
},
|
|
2992
|
-
contextWindow:
|
|
2873
|
+
contextWindow: 272000,
|
|
2993
2874
|
maxTokens: 128000,
|
|
2994
2875
|
},
|
|
2995
2876
|
},
|
|
@@ -3062,6 +2943,23 @@ export const MODELS = {
|
|
|
3062
2943
|
contextWindow: 65536,
|
|
3063
2944
|
maxTokens: 65536,
|
|
3064
2945
|
},
|
|
2946
|
+
"allenai/olmo-3.1-32b-instruct": {
|
|
2947
|
+
id: "allenai/olmo-3.1-32b-instruct",
|
|
2948
|
+
name: "AllenAI: Olmo 3.1 32B Instruct",
|
|
2949
|
+
api: "openai-completions",
|
|
2950
|
+
provider: "openrouter",
|
|
2951
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
2952
|
+
reasoning: false,
|
|
2953
|
+
input: ["text"],
|
|
2954
|
+
cost: {
|
|
2955
|
+
input: 0.19999999999999998,
|
|
2956
|
+
output: 0.6,
|
|
2957
|
+
cacheRead: 0,
|
|
2958
|
+
cacheWrite: 0,
|
|
2959
|
+
},
|
|
2960
|
+
contextWindow: 65536,
|
|
2961
|
+
maxTokens: 4096,
|
|
2962
|
+
},
|
|
3065
2963
|
"amazon/nova-2-lite-v1": {
|
|
3066
2964
|
id: "amazon/nova-2-lite-v1",
|
|
3067
2965
|
name: "Amazon: Nova 2 Lite",
|
|
@@ -3938,13 +3836,13 @@ export const MODELS = {
|
|
|
3938
3836
|
reasoning: false,
|
|
3939
3837
|
input: ["text", "image"],
|
|
3940
3838
|
cost: {
|
|
3941
|
-
input: 0.
|
|
3942
|
-
output: 0.
|
|
3839
|
+
input: 0.04,
|
|
3840
|
+
output: 0.15,
|
|
3943
3841
|
cacheRead: 0,
|
|
3944
3842
|
cacheWrite: 0,
|
|
3945
3843
|
},
|
|
3946
|
-
contextWindow:
|
|
3947
|
-
maxTokens:
|
|
3844
|
+
contextWindow: 96000,
|
|
3845
|
+
maxTokens: 96000,
|
|
3948
3846
|
},
|
|
3949
3847
|
"google/gemma-3-27b-it:free": {
|
|
3950
3848
|
id: "google/gemma-3-27b-it:free",
|
|
@@ -4244,9 +4142,9 @@ export const MODELS = {
|
|
|
4244
4142
|
reasoning: true,
|
|
4245
4143
|
input: ["text"],
|
|
4246
4144
|
cost: {
|
|
4247
|
-
input: 0.
|
|
4248
|
-
output:
|
|
4249
|
-
cacheRead: 0,
|
|
4145
|
+
input: 0.28,
|
|
4146
|
+
output: 1.2,
|
|
4147
|
+
cacheRead: 0.14,
|
|
4250
4148
|
cacheWrite: 0,
|
|
4251
4149
|
},
|
|
4252
4150
|
contextWindow: 196608,
|
|
@@ -4839,13 +4737,13 @@ export const MODELS = {
|
|
|
4839
4737
|
reasoning: true,
|
|
4840
4738
|
input: ["text"],
|
|
4841
4739
|
cost: {
|
|
4842
|
-
input: 0.
|
|
4843
|
-
output:
|
|
4740
|
+
input: 0.39999999999999997,
|
|
4741
|
+
output: 1.75,
|
|
4844
4742
|
cacheRead: 0,
|
|
4845
4743
|
cacheWrite: 0,
|
|
4846
4744
|
},
|
|
4847
4745
|
contextWindow: 262144,
|
|
4848
|
-
maxTokens:
|
|
4746
|
+
maxTokens: 65535,
|
|
4849
4747
|
},
|
|
4850
4748
|
"nex-agi/deepseek-v3.1-nex-n1:free": {
|
|
4851
4749
|
id: "nex-agi/deepseek-v3.1-nex-n1:free",
|
|
@@ -5621,8 +5519,8 @@ export const MODELS = {
|
|
|
5621
5519
|
reasoning: true,
|
|
5622
5520
|
input: ["text"],
|
|
5623
5521
|
cost: {
|
|
5624
|
-
input: 0.
|
|
5625
|
-
output: 0.
|
|
5522
|
+
input: 0.039,
|
|
5523
|
+
output: 0.19,
|
|
5626
5524
|
cacheRead: 0,
|
|
5627
5525
|
cacheWrite: 0,
|
|
5628
5526
|
},
|
|
@@ -5672,13 +5570,13 @@ export const MODELS = {
|
|
|
5672
5570
|
reasoning: true,
|
|
5673
5571
|
input: ["text"],
|
|
5674
5572
|
cost: {
|
|
5675
|
-
input: 0.
|
|
5676
|
-
output: 0.
|
|
5573
|
+
input: 0.02,
|
|
5574
|
+
output: 0.09999999999999999,
|
|
5677
5575
|
cacheRead: 0,
|
|
5678
5576
|
cacheWrite: 0,
|
|
5679
5577
|
},
|
|
5680
5578
|
contextWindow: 131072,
|
|
5681
|
-
maxTokens:
|
|
5579
|
+
maxTokens: 131072,
|
|
5682
5580
|
},
|
|
5683
5581
|
"openai/gpt-oss-20b:free": {
|
|
5684
5582
|
id: "openai/gpt-oss-20b:free",
|
|
@@ -5918,6 +5816,23 @@ export const MODELS = {
|
|
|
5918
5816
|
contextWindow: 32768,
|
|
5919
5817
|
maxTokens: 16384,
|
|
5920
5818
|
},
|
|
5819
|
+
"qwen/qwen-2.5-7b-instruct": {
|
|
5820
|
+
id: "qwen/qwen-2.5-7b-instruct",
|
|
5821
|
+
name: "Qwen: Qwen2.5 7B Instruct",
|
|
5822
|
+
api: "openai-completions",
|
|
5823
|
+
provider: "openrouter",
|
|
5824
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
5825
|
+
reasoning: false,
|
|
5826
|
+
input: ["text"],
|
|
5827
|
+
cost: {
|
|
5828
|
+
input: 0.04,
|
|
5829
|
+
output: 0.09999999999999999,
|
|
5830
|
+
cacheRead: 0,
|
|
5831
|
+
cacheWrite: 0,
|
|
5832
|
+
},
|
|
5833
|
+
contextWindow: 32768,
|
|
5834
|
+
maxTokens: 4096,
|
|
5835
|
+
},
|
|
5921
5836
|
"qwen/qwen-max": {
|
|
5922
5837
|
id: "qwen/qwen-max",
|
|
5923
5838
|
name: "Qwen: Qwen-Max ",
|
|
@@ -6318,8 +6233,8 @@ export const MODELS = {
|
|
|
6318
6233
|
reasoning: false,
|
|
6319
6234
|
input: ["text"],
|
|
6320
6235
|
cost: {
|
|
6321
|
-
input: 0.
|
|
6322
|
-
output:
|
|
6236
|
+
input: 0.09,
|
|
6237
|
+
output: 1.1,
|
|
6323
6238
|
cacheRead: 0,
|
|
6324
6239
|
cacheWrite: 0,
|
|
6325
6240
|
},
|
|
@@ -6352,8 +6267,8 @@ export const MODELS = {
|
|
|
6352
6267
|
reasoning: false,
|
|
6353
6268
|
input: ["text", "image"],
|
|
6354
6269
|
cost: {
|
|
6355
|
-
input: 0.
|
|
6356
|
-
output:
|
|
6270
|
+
input: 0.19999999999999998,
|
|
6271
|
+
output: 1.2,
|
|
6357
6272
|
cacheRead: 0,
|
|
6358
6273
|
cacheWrite: 0,
|
|
6359
6274
|
},
|
|
@@ -6913,13 +6828,13 @@ export const MODELS = {
|
|
|
6913
6828
|
reasoning: true,
|
|
6914
6829
|
input: ["text"],
|
|
6915
6830
|
cost: {
|
|
6916
|
-
input: 0.
|
|
6917
|
-
output:
|
|
6831
|
+
input: 0.39999999999999997,
|
|
6832
|
+
output: 1.5,
|
|
6918
6833
|
cacheRead: 0,
|
|
6919
6834
|
cacheWrite: 0,
|
|
6920
6835
|
},
|
|
6921
6836
|
contextWindow: 202752,
|
|
6922
|
-
maxTokens:
|
|
6837
|
+
maxTokens: 65535,
|
|
6923
6838
|
},
|
|
6924
6839
|
},
|
|
6925
6840
|
"xai": {
|