@mariozechner/pi-ai 0.37.8 → 0.39.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/models.generated.d.ts +50 -135
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +78 -163
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/google-gemini-cli.d.ts.map +1 -1
- package/dist/providers/google-gemini-cli.js +250 -130
- package/dist/providers/google-gemini-cli.js.map +1 -1
- package/dist/providers/google-shared.d.ts.map +1 -1
- package/dist/providers/google-shared.js +10 -5
- package/dist/providers/google-shared.js.map +1 -1
- package/dist/providers/openai-codex/prompts/codex.d.ts +2 -2
- package/dist/providers/openai-codex/prompts/codex.d.ts.map +1 -1
- package/dist/providers/openai-codex/prompts/codex.js +7 -7
- package/dist/providers/openai-codex/prompts/codex.js.map +1 -1
- package/dist/providers/openai-codex/request-transformer.d.ts +0 -1
- package/dist/providers/openai-codex/request-transformer.d.ts.map +1 -1
- package/dist/providers/openai-codex/request-transformer.js +12 -130
- package/dist/providers/openai-codex/request-transformer.js.map +1 -1
- package/dist/providers/openai-codex-responses.d.ts.map +1 -1
- package/dist/providers/openai-codex-responses.js +65 -8
- package/dist/providers/openai-codex-responses.js.map +1 -1
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +2 -4
- package/dist/providers/openai-completions.js.map +1 -1
- package/dist/providers/transorm-messages.d.ts.map +1 -1
- package/dist/providers/transorm-messages.js +5 -3
- package/dist/providers/transorm-messages.js.map +1 -1
- package/dist/stream.d.ts.map +1 -1
- package/dist/stream.js +13 -6
- package/dist/stream.js.map +1 -1
- package/dist/types.d.ts +9 -0
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/dist/utils/overflow.d.ts.map +1 -1
- package/dist/utils/overflow.js +1 -1
- package/dist/utils/overflow.js.map +1 -1
- package/package.json +1 -1
|
@@ -410,6 +410,23 @@ export declare const MODELS: {
|
|
|
410
410
|
contextWindow: number;
|
|
411
411
|
maxTokens: number;
|
|
412
412
|
};
|
|
413
|
+
readonly "zai-glm-4.7": {
|
|
414
|
+
id: string;
|
|
415
|
+
name: string;
|
|
416
|
+
api: "openai-completions";
|
|
417
|
+
provider: string;
|
|
418
|
+
baseUrl: string;
|
|
419
|
+
reasoning: false;
|
|
420
|
+
input: "text"[];
|
|
421
|
+
cost: {
|
|
422
|
+
input: number;
|
|
423
|
+
output: number;
|
|
424
|
+
cacheRead: number;
|
|
425
|
+
cacheWrite: number;
|
|
426
|
+
};
|
|
427
|
+
contextWindow: number;
|
|
428
|
+
maxTokens: number;
|
|
429
|
+
};
|
|
413
430
|
};
|
|
414
431
|
readonly "github-copilot": {
|
|
415
432
|
readonly "claude-haiku-4.5": {
|
|
@@ -2904,108 +2921,6 @@ export declare const MODELS: {
|
|
|
2904
2921
|
};
|
|
2905
2922
|
};
|
|
2906
2923
|
readonly "openai-codex": {
|
|
2907
|
-
readonly "codex-mini-latest": {
|
|
2908
|
-
id: string;
|
|
2909
|
-
name: string;
|
|
2910
|
-
api: "openai-codex-responses";
|
|
2911
|
-
provider: string;
|
|
2912
|
-
baseUrl: string;
|
|
2913
|
-
reasoning: true;
|
|
2914
|
-
input: ("image" | "text")[];
|
|
2915
|
-
cost: {
|
|
2916
|
-
input: number;
|
|
2917
|
-
output: number;
|
|
2918
|
-
cacheRead: number;
|
|
2919
|
-
cacheWrite: number;
|
|
2920
|
-
};
|
|
2921
|
-
contextWindow: number;
|
|
2922
|
-
maxTokens: number;
|
|
2923
|
-
};
|
|
2924
|
-
readonly "gpt-5": {
|
|
2925
|
-
id: string;
|
|
2926
|
-
name: string;
|
|
2927
|
-
api: "openai-codex-responses";
|
|
2928
|
-
provider: string;
|
|
2929
|
-
baseUrl: string;
|
|
2930
|
-
reasoning: true;
|
|
2931
|
-
input: ("image" | "text")[];
|
|
2932
|
-
cost: {
|
|
2933
|
-
input: number;
|
|
2934
|
-
output: number;
|
|
2935
|
-
cacheRead: number;
|
|
2936
|
-
cacheWrite: number;
|
|
2937
|
-
};
|
|
2938
|
-
contextWindow: number;
|
|
2939
|
-
maxTokens: number;
|
|
2940
|
-
};
|
|
2941
|
-
readonly "gpt-5-codex": {
|
|
2942
|
-
id: string;
|
|
2943
|
-
name: string;
|
|
2944
|
-
api: "openai-codex-responses";
|
|
2945
|
-
provider: string;
|
|
2946
|
-
baseUrl: string;
|
|
2947
|
-
reasoning: true;
|
|
2948
|
-
input: ("image" | "text")[];
|
|
2949
|
-
cost: {
|
|
2950
|
-
input: number;
|
|
2951
|
-
output: number;
|
|
2952
|
-
cacheRead: number;
|
|
2953
|
-
cacheWrite: number;
|
|
2954
|
-
};
|
|
2955
|
-
contextWindow: number;
|
|
2956
|
-
maxTokens: number;
|
|
2957
|
-
};
|
|
2958
|
-
readonly "gpt-5-codex-mini": {
|
|
2959
|
-
id: string;
|
|
2960
|
-
name: string;
|
|
2961
|
-
api: "openai-codex-responses";
|
|
2962
|
-
provider: string;
|
|
2963
|
-
baseUrl: string;
|
|
2964
|
-
reasoning: true;
|
|
2965
|
-
input: ("image" | "text")[];
|
|
2966
|
-
cost: {
|
|
2967
|
-
input: number;
|
|
2968
|
-
output: number;
|
|
2969
|
-
cacheRead: number;
|
|
2970
|
-
cacheWrite: number;
|
|
2971
|
-
};
|
|
2972
|
-
contextWindow: number;
|
|
2973
|
-
maxTokens: number;
|
|
2974
|
-
};
|
|
2975
|
-
readonly "gpt-5-mini": {
|
|
2976
|
-
id: string;
|
|
2977
|
-
name: string;
|
|
2978
|
-
api: "openai-codex-responses";
|
|
2979
|
-
provider: string;
|
|
2980
|
-
baseUrl: string;
|
|
2981
|
-
reasoning: true;
|
|
2982
|
-
input: ("image" | "text")[];
|
|
2983
|
-
cost: {
|
|
2984
|
-
input: number;
|
|
2985
|
-
output: number;
|
|
2986
|
-
cacheRead: number;
|
|
2987
|
-
cacheWrite: number;
|
|
2988
|
-
};
|
|
2989
|
-
contextWindow: number;
|
|
2990
|
-
maxTokens: number;
|
|
2991
|
-
};
|
|
2992
|
-
readonly "gpt-5-nano": {
|
|
2993
|
-
id: string;
|
|
2994
|
-
name: string;
|
|
2995
|
-
api: "openai-codex-responses";
|
|
2996
|
-
provider: string;
|
|
2997
|
-
baseUrl: string;
|
|
2998
|
-
reasoning: true;
|
|
2999
|
-
input: ("image" | "text")[];
|
|
3000
|
-
cost: {
|
|
3001
|
-
input: number;
|
|
3002
|
-
output: number;
|
|
3003
|
-
cacheRead: number;
|
|
3004
|
-
cacheWrite: number;
|
|
3005
|
-
};
|
|
3006
|
-
contextWindow: number;
|
|
3007
|
-
maxTokens: number;
|
|
3008
|
-
};
|
|
3009
2924
|
readonly "gpt-5.1": {
|
|
3010
2925
|
id: string;
|
|
3011
2926
|
name: string;
|
|
@@ -3023,24 +2938,7 @@ export declare const MODELS: {
|
|
|
3023
2938
|
contextWindow: number;
|
|
3024
2939
|
maxTokens: number;
|
|
3025
2940
|
};
|
|
3026
|
-
readonly "gpt-5.1-
|
|
3027
|
-
id: string;
|
|
3028
|
-
name: string;
|
|
3029
|
-
api: "openai-codex-responses";
|
|
3030
|
-
provider: string;
|
|
3031
|
-
baseUrl: string;
|
|
3032
|
-
reasoning: true;
|
|
3033
|
-
input: ("image" | "text")[];
|
|
3034
|
-
cost: {
|
|
3035
|
-
input: number;
|
|
3036
|
-
output: number;
|
|
3037
|
-
cacheRead: number;
|
|
3038
|
-
cacheWrite: number;
|
|
3039
|
-
};
|
|
3040
|
-
contextWindow: number;
|
|
3041
|
-
maxTokens: number;
|
|
3042
|
-
};
|
|
3043
|
-
readonly "gpt-5.1-codex": {
|
|
2941
|
+
readonly "gpt-5.1-codex-max": {
|
|
3044
2942
|
id: string;
|
|
3045
2943
|
name: string;
|
|
3046
2944
|
api: "openai-codex-responses";
|
|
@@ -3057,7 +2955,7 @@ export declare const MODELS: {
|
|
|
3057
2955
|
contextWindow: number;
|
|
3058
2956
|
maxTokens: number;
|
|
3059
2957
|
};
|
|
3060
|
-
readonly "gpt-5.1-codex-
|
|
2958
|
+
readonly "gpt-5.1-codex-mini": {
|
|
3061
2959
|
id: string;
|
|
3062
2960
|
name: string;
|
|
3063
2961
|
api: "openai-codex-responses";
|
|
@@ -3074,7 +2972,7 @@ export declare const MODELS: {
|
|
|
3074
2972
|
contextWindow: number;
|
|
3075
2973
|
maxTokens: number;
|
|
3076
2974
|
};
|
|
3077
|
-
readonly "gpt-5.
|
|
2975
|
+
readonly "gpt-5.2": {
|
|
3078
2976
|
id: string;
|
|
3079
2977
|
name: string;
|
|
3080
2978
|
api: "openai-codex-responses";
|
|
@@ -3091,7 +2989,7 @@ export declare const MODELS: {
|
|
|
3091
2989
|
contextWindow: number;
|
|
3092
2990
|
maxTokens: number;
|
|
3093
2991
|
};
|
|
3094
|
-
readonly "gpt-5.2": {
|
|
2992
|
+
readonly "gpt-5.2-codex": {
|
|
3095
2993
|
id: string;
|
|
3096
2994
|
name: string;
|
|
3097
2995
|
api: "openai-codex-responses";
|
|
@@ -3108,14 +3006,16 @@ export declare const MODELS: {
|
|
|
3108
3006
|
contextWindow: number;
|
|
3109
3007
|
maxTokens: number;
|
|
3110
3008
|
};
|
|
3111
|
-
|
|
3009
|
+
};
|
|
3010
|
+
readonly openrouter: {
|
|
3011
|
+
readonly "ai21/jamba-large-1.7": {
|
|
3112
3012
|
id: string;
|
|
3113
3013
|
name: string;
|
|
3114
|
-
api: "openai-
|
|
3014
|
+
api: "openai-completions";
|
|
3115
3015
|
provider: string;
|
|
3116
3016
|
baseUrl: string;
|
|
3117
|
-
reasoning:
|
|
3118
|
-
input:
|
|
3017
|
+
reasoning: false;
|
|
3018
|
+
input: "text"[];
|
|
3119
3019
|
cost: {
|
|
3120
3020
|
input: number;
|
|
3121
3021
|
output: number;
|
|
@@ -3125,9 +3025,7 @@ export declare const MODELS: {
|
|
|
3125
3025
|
contextWindow: number;
|
|
3126
3026
|
maxTokens: number;
|
|
3127
3027
|
};
|
|
3128
|
-
|
|
3129
|
-
readonly openrouter: {
|
|
3130
|
-
readonly "ai21/jamba-large-1.7": {
|
|
3028
|
+
readonly "ai21/jamba-mini-1.7": {
|
|
3131
3029
|
id: string;
|
|
3132
3030
|
name: string;
|
|
3133
3031
|
api: "openai-completions";
|
|
@@ -3144,13 +3042,13 @@ export declare const MODELS: {
|
|
|
3144
3042
|
contextWindow: number;
|
|
3145
3043
|
maxTokens: number;
|
|
3146
3044
|
};
|
|
3147
|
-
readonly "
|
|
3045
|
+
readonly "alibaba/tongyi-deepresearch-30b-a3b": {
|
|
3148
3046
|
id: string;
|
|
3149
3047
|
name: string;
|
|
3150
3048
|
api: "openai-completions";
|
|
3151
3049
|
provider: string;
|
|
3152
3050
|
baseUrl: string;
|
|
3153
|
-
reasoning:
|
|
3051
|
+
reasoning: true;
|
|
3154
3052
|
input: "text"[];
|
|
3155
3053
|
cost: {
|
|
3156
3054
|
input: number;
|
|
@@ -3161,13 +3059,13 @@ export declare const MODELS: {
|
|
|
3161
3059
|
contextWindow: number;
|
|
3162
3060
|
maxTokens: number;
|
|
3163
3061
|
};
|
|
3164
|
-
readonly "
|
|
3062
|
+
readonly "allenai/olmo-3-7b-instruct": {
|
|
3165
3063
|
id: string;
|
|
3166
3064
|
name: string;
|
|
3167
3065
|
api: "openai-completions";
|
|
3168
3066
|
provider: string;
|
|
3169
3067
|
baseUrl: string;
|
|
3170
|
-
reasoning:
|
|
3068
|
+
reasoning: false;
|
|
3171
3069
|
input: "text"[];
|
|
3172
3070
|
cost: {
|
|
3173
3071
|
input: number;
|
|
@@ -3178,7 +3076,7 @@ export declare const MODELS: {
|
|
|
3178
3076
|
contextWindow: number;
|
|
3179
3077
|
maxTokens: number;
|
|
3180
3078
|
};
|
|
3181
|
-
readonly "allenai/olmo-3-
|
|
3079
|
+
readonly "allenai/olmo-3.1-32b-instruct": {
|
|
3182
3080
|
id: string;
|
|
3183
3081
|
name: string;
|
|
3184
3082
|
api: "openai-completions";
|
|
@@ -6051,6 +5949,23 @@ export declare const MODELS: {
|
|
|
6051
5949
|
contextWindow: number;
|
|
6052
5950
|
maxTokens: number;
|
|
6053
5951
|
};
|
|
5952
|
+
readonly "qwen/qwen-2.5-7b-instruct": {
|
|
5953
|
+
id: string;
|
|
5954
|
+
name: string;
|
|
5955
|
+
api: "openai-completions";
|
|
5956
|
+
provider: string;
|
|
5957
|
+
baseUrl: string;
|
|
5958
|
+
reasoning: false;
|
|
5959
|
+
input: "text"[];
|
|
5960
|
+
cost: {
|
|
5961
|
+
input: number;
|
|
5962
|
+
output: number;
|
|
5963
|
+
cacheRead: number;
|
|
5964
|
+
cacheWrite: number;
|
|
5965
|
+
};
|
|
5966
|
+
contextWindow: number;
|
|
5967
|
+
maxTokens: number;
|
|
5968
|
+
};
|
|
6054
5969
|
readonly "qwen/qwen-max": {
|
|
6055
5970
|
id: string;
|
|
6056
5971
|
name: string;
|