@mariozechner/pi-ai 0.12.9 → 0.12.11
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
|
@@ -1075,6 +1075,23 @@ export declare const MODELS: {
|
|
|
1075
1075
|
contextWindow: number;
|
|
1076
1076
|
maxTokens: number;
|
|
1077
1077
|
};
|
|
1078
|
+
readonly "gpt-5.1-codex-max": {
|
|
1079
|
+
id: string;
|
|
1080
|
+
name: string;
|
|
1081
|
+
api: "openai-responses";
|
|
1082
|
+
provider: string;
|
|
1083
|
+
baseUrl: string;
|
|
1084
|
+
reasoning: true;
|
|
1085
|
+
input: ("image" | "text")[];
|
|
1086
|
+
cost: {
|
|
1087
|
+
input: number;
|
|
1088
|
+
output: number;
|
|
1089
|
+
cacheRead: number;
|
|
1090
|
+
cacheWrite: number;
|
|
1091
|
+
};
|
|
1092
|
+
contextWindow: number;
|
|
1093
|
+
maxTokens: number;
|
|
1094
|
+
};
|
|
1078
1095
|
readonly o3: {
|
|
1079
1096
|
id: string;
|
|
1080
1097
|
name: string;
|
|
@@ -2003,6 +2020,57 @@ export declare const MODELS: {
|
|
|
2003
2020
|
contextWindow: number;
|
|
2004
2021
|
maxTokens: number;
|
|
2005
2022
|
};
|
|
2023
|
+
readonly "mistralai/ministral-14b-2512": {
|
|
2024
|
+
id: string;
|
|
2025
|
+
name: string;
|
|
2026
|
+
api: "openai-completions";
|
|
2027
|
+
provider: string;
|
|
2028
|
+
baseUrl: string;
|
|
2029
|
+
reasoning: false;
|
|
2030
|
+
input: ("image" | "text")[];
|
|
2031
|
+
cost: {
|
|
2032
|
+
input: number;
|
|
2033
|
+
output: number;
|
|
2034
|
+
cacheRead: number;
|
|
2035
|
+
cacheWrite: number;
|
|
2036
|
+
};
|
|
2037
|
+
contextWindow: number;
|
|
2038
|
+
maxTokens: number;
|
|
2039
|
+
};
|
|
2040
|
+
readonly "mistralai/ministral-8b-2512": {
|
|
2041
|
+
id: string;
|
|
2042
|
+
name: string;
|
|
2043
|
+
api: "openai-completions";
|
|
2044
|
+
provider: string;
|
|
2045
|
+
baseUrl: string;
|
|
2046
|
+
reasoning: false;
|
|
2047
|
+
input: ("image" | "text")[];
|
|
2048
|
+
cost: {
|
|
2049
|
+
input: number;
|
|
2050
|
+
output: number;
|
|
2051
|
+
cacheRead: number;
|
|
2052
|
+
cacheWrite: number;
|
|
2053
|
+
};
|
|
2054
|
+
contextWindow: number;
|
|
2055
|
+
maxTokens: number;
|
|
2056
|
+
};
|
|
2057
|
+
readonly "mistralai/ministral-3b-2512": {
|
|
2058
|
+
id: string;
|
|
2059
|
+
name: string;
|
|
2060
|
+
api: "openai-completions";
|
|
2061
|
+
provider: string;
|
|
2062
|
+
baseUrl: string;
|
|
2063
|
+
reasoning: false;
|
|
2064
|
+
input: ("image" | "text")[];
|
|
2065
|
+
cost: {
|
|
2066
|
+
input: number;
|
|
2067
|
+
output: number;
|
|
2068
|
+
cacheRead: number;
|
|
2069
|
+
cacheWrite: number;
|
|
2070
|
+
};
|
|
2071
|
+
contextWindow: number;
|
|
2072
|
+
maxTokens: number;
|
|
2073
|
+
};
|
|
2006
2074
|
readonly "mistralai/mistral-large-2512": {
|
|
2007
2075
|
id: string;
|
|
2008
2076
|
name: string;
|
|
@@ -3329,7 +3397,7 @@ export declare const MODELS: {
|
|
|
3329
3397
|
contextWindow: number;
|
|
3330
3398
|
maxTokens: number;
|
|
3331
3399
|
};
|
|
3332
|
-
readonly "openai/gpt-oss-120b
|
|
3400
|
+
readonly "openai/gpt-oss-120b": {
|
|
3333
3401
|
id: string;
|
|
3334
3402
|
name: string;
|
|
3335
3403
|
api: "openai-completions";
|
|
@@ -3346,7 +3414,7 @@ export declare const MODELS: {
|
|
|
3346
3414
|
contextWindow: number;
|
|
3347
3415
|
maxTokens: number;
|
|
3348
3416
|
};
|
|
3349
|
-
readonly "openai/gpt-oss-120b": {
|
|
3417
|
+
readonly "openai/gpt-oss-120b:exacto": {
|
|
3350
3418
|
id: string;
|
|
3351
3419
|
name: string;
|
|
3352
3420
|
api: "openai-completions";
|
|
@@ -3856,23 +3924,6 @@ export declare const MODELS: {
|
|
|
3856
3924
|
contextWindow: number;
|
|
3857
3925
|
maxTokens: number;
|
|
3858
3926
|
};
|
|
3859
|
-
readonly "mistralai/magistral-small-2506": {
|
|
3860
|
-
id: string;
|
|
3861
|
-
name: string;
|
|
3862
|
-
api: "openai-completions";
|
|
3863
|
-
provider: string;
|
|
3864
|
-
baseUrl: string;
|
|
3865
|
-
reasoning: true;
|
|
3866
|
-
input: "text"[];
|
|
3867
|
-
cost: {
|
|
3868
|
-
input: number;
|
|
3869
|
-
output: number;
|
|
3870
|
-
cacheRead: number;
|
|
3871
|
-
cacheWrite: number;
|
|
3872
|
-
};
|
|
3873
|
-
contextWindow: number;
|
|
3874
|
-
maxTokens: number;
|
|
3875
|
-
};
|
|
3876
3927
|
readonly "mistralai/magistral-medium-2506:thinking": {
|
|
3877
3928
|
id: string;
|
|
3878
3929
|
name: string;
|
|
@@ -3890,23 +3941,6 @@ export declare const MODELS: {
|
|
|
3890
3941
|
contextWindow: number;
|
|
3891
3942
|
maxTokens: number;
|
|
3892
3943
|
};
|
|
3893
|
-
readonly "mistralai/magistral-medium-2506": {
|
|
3894
|
-
id: string;
|
|
3895
|
-
name: string;
|
|
3896
|
-
api: "openai-completions";
|
|
3897
|
-
provider: string;
|
|
3898
|
-
baseUrl: string;
|
|
3899
|
-
reasoning: true;
|
|
3900
|
-
input: "text"[];
|
|
3901
|
-
cost: {
|
|
3902
|
-
input: number;
|
|
3903
|
-
output: number;
|
|
3904
|
-
cacheRead: number;
|
|
3905
|
-
cacheWrite: number;
|
|
3906
|
-
};
|
|
3907
|
-
contextWindow: number;
|
|
3908
|
-
maxTokens: number;
|
|
3909
|
-
};
|
|
3910
3944
|
readonly "google/gemini-2.5-pro-preview": {
|
|
3911
3945
|
id: string;
|
|
3912
3946
|
name: string;
|
|
@@ -3975,23 +4009,6 @@ export declare const MODELS: {
|
|
|
3975
4009
|
contextWindow: number;
|
|
3976
4010
|
maxTokens: number;
|
|
3977
4011
|
};
|
|
3978
|
-
readonly "mistralai/devstral-small-2505": {
|
|
3979
|
-
id: string;
|
|
3980
|
-
name: string;
|
|
3981
|
-
api: "openai-completions";
|
|
3982
|
-
provider: string;
|
|
3983
|
-
baseUrl: string;
|
|
3984
|
-
reasoning: false;
|
|
3985
|
-
input: "text"[];
|
|
3986
|
-
cost: {
|
|
3987
|
-
input: number;
|
|
3988
|
-
output: number;
|
|
3989
|
-
cacheRead: number;
|
|
3990
|
-
cacheWrite: number;
|
|
3991
|
-
};
|
|
3992
|
-
contextWindow: number;
|
|
3993
|
-
maxTokens: number;
|
|
3994
|
-
};
|
|
3995
4012
|
readonly "openai/codex-mini": {
|
|
3996
4013
|
id: string;
|
|
3997
4014
|
name: string;
|
|
@@ -4706,23 +4723,6 @@ export declare const MODELS: {
|
|
|
4706
4723
|
contextWindow: number;
|
|
4707
4724
|
maxTokens: number;
|
|
4708
4725
|
};
|
|
4709
|
-
readonly "mistralai/codestral-2501": {
|
|
4710
|
-
id: string;
|
|
4711
|
-
name: string;
|
|
4712
|
-
api: "openai-completions";
|
|
4713
|
-
provider: string;
|
|
4714
|
-
baseUrl: string;
|
|
4715
|
-
reasoning: false;
|
|
4716
|
-
input: "text"[];
|
|
4717
|
-
cost: {
|
|
4718
|
-
input: number;
|
|
4719
|
-
output: number;
|
|
4720
|
-
cacheRead: number;
|
|
4721
|
-
cacheWrite: number;
|
|
4722
|
-
};
|
|
4723
|
-
contextWindow: number;
|
|
4724
|
-
maxTokens: number;
|
|
4725
|
-
};
|
|
4726
4726
|
readonly "deepseek/deepseek-chat": {
|
|
4727
4727
|
id: string;
|
|
4728
4728
|
name: string;
|
|
@@ -5590,23 +5590,6 @@ export declare const MODELS: {
|
|
|
5590
5590
|
contextWindow: number;
|
|
5591
5591
|
maxTokens: number;
|
|
5592
5592
|
};
|
|
5593
|
-
readonly "mistralai/mistral-small": {
|
|
5594
|
-
id: string;
|
|
5595
|
-
name: string;
|
|
5596
|
-
api: "openai-completions";
|
|
5597
|
-
provider: string;
|
|
5598
|
-
baseUrl: string;
|
|
5599
|
-
reasoning: false;
|
|
5600
|
-
input: "text"[];
|
|
5601
|
-
cost: {
|
|
5602
|
-
input: number;
|
|
5603
|
-
output: number;
|
|
5604
|
-
cacheRead: number;
|
|
5605
|
-
cacheWrite: number;
|
|
5606
|
-
};
|
|
5607
|
-
contextWindow: number;
|
|
5608
|
-
maxTokens: number;
|
|
5609
|
-
};
|
|
5610
5593
|
readonly "mistralai/mistral-tiny": {
|
|
5611
5594
|
id: string;
|
|
5612
5595
|
name: string;
|