@mariozechner/pi-ai 0.12.10 → 0.12.12
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
|
@@ -1075,7 +1075,7 @@ export declare const MODELS: {
|
|
|
1075
1075
|
contextWindow: number;
|
|
1076
1076
|
maxTokens: number;
|
|
1077
1077
|
};
|
|
1078
|
-
readonly
|
|
1078
|
+
readonly "gpt-5.1-codex-max": {
|
|
1079
1079
|
id: string;
|
|
1080
1080
|
name: string;
|
|
1081
1081
|
api: "openai-responses";
|
|
@@ -1092,7 +1092,7 @@ export declare const MODELS: {
|
|
|
1092
1092
|
contextWindow: number;
|
|
1093
1093
|
maxTokens: number;
|
|
1094
1094
|
};
|
|
1095
|
-
readonly
|
|
1095
|
+
readonly o3: {
|
|
1096
1096
|
id: string;
|
|
1097
1097
|
name: string;
|
|
1098
1098
|
api: "openai-responses";
|
|
@@ -1109,13 +1109,13 @@ export declare const MODELS: {
|
|
|
1109
1109
|
contextWindow: number;
|
|
1110
1110
|
maxTokens: number;
|
|
1111
1111
|
};
|
|
1112
|
-
readonly "
|
|
1112
|
+
readonly "o4-mini-deep-research": {
|
|
1113
1113
|
id: string;
|
|
1114
1114
|
name: string;
|
|
1115
1115
|
api: "openai-responses";
|
|
1116
1116
|
provider: string;
|
|
1117
1117
|
baseUrl: string;
|
|
1118
|
-
reasoning:
|
|
1118
|
+
reasoning: true;
|
|
1119
1119
|
input: ("image" | "text")[];
|
|
1120
1120
|
cost: {
|
|
1121
1121
|
input: number;
|
|
@@ -1126,13 +1126,13 @@ export declare const MODELS: {
|
|
|
1126
1126
|
contextWindow: number;
|
|
1127
1127
|
maxTokens: number;
|
|
1128
1128
|
};
|
|
1129
|
-
readonly "gpt-
|
|
1129
|
+
readonly "gpt-4o-mini": {
|
|
1130
1130
|
id: string;
|
|
1131
1131
|
name: string;
|
|
1132
1132
|
api: "openai-responses";
|
|
1133
1133
|
provider: string;
|
|
1134
1134
|
baseUrl: string;
|
|
1135
|
-
reasoning:
|
|
1135
|
+
reasoning: false;
|
|
1136
1136
|
input: ("image" | "text")[];
|
|
1137
1137
|
cost: {
|
|
1138
1138
|
input: number;
|
|
@@ -1143,7 +1143,7 @@ export declare const MODELS: {
|
|
|
1143
1143
|
contextWindow: number;
|
|
1144
1144
|
maxTokens: number;
|
|
1145
1145
|
};
|
|
1146
|
-
readonly "gpt-5
|
|
1146
|
+
readonly "gpt-5": {
|
|
1147
1147
|
id: string;
|
|
1148
1148
|
name: string;
|
|
1149
1149
|
api: "openai-responses";
|
|
@@ -1160,7 +1160,7 @@ export declare const MODELS: {
|
|
|
1160
1160
|
contextWindow: number;
|
|
1161
1161
|
maxTokens: number;
|
|
1162
1162
|
};
|
|
1163
|
-
readonly "gpt-5
|
|
1163
|
+
readonly "gpt-5-pro": {
|
|
1164
1164
|
id: string;
|
|
1165
1165
|
name: string;
|
|
1166
1166
|
api: "openai-responses";
|
|
@@ -1177,13 +1177,13 @@ export declare const MODELS: {
|
|
|
1177
1177
|
contextWindow: number;
|
|
1178
1178
|
maxTokens: number;
|
|
1179
1179
|
};
|
|
1180
|
-
readonly "gpt-5-chat-latest": {
|
|
1180
|
+
readonly "gpt-5.1-chat-latest": {
|
|
1181
1181
|
id: string;
|
|
1182
1182
|
name: string;
|
|
1183
1183
|
api: "openai-responses";
|
|
1184
1184
|
provider: string;
|
|
1185
1185
|
baseUrl: string;
|
|
1186
|
-
reasoning:
|
|
1186
|
+
reasoning: true;
|
|
1187
1187
|
input: ("image" | "text")[];
|
|
1188
1188
|
cost: {
|
|
1189
1189
|
input: number;
|
|
@@ -1194,13 +1194,13 @@ export declare const MODELS: {
|
|
|
1194
1194
|
contextWindow: number;
|
|
1195
1195
|
maxTokens: number;
|
|
1196
1196
|
};
|
|
1197
|
-
readonly "gpt-5
|
|
1197
|
+
readonly "gpt-5-chat-latest": {
|
|
1198
1198
|
id: string;
|
|
1199
1199
|
name: string;
|
|
1200
1200
|
api: "openai-responses";
|
|
1201
1201
|
provider: string;
|
|
1202
1202
|
baseUrl: string;
|
|
1203
|
-
reasoning:
|
|
1203
|
+
reasoning: false;
|
|
1204
1204
|
input: ("image" | "text")[];
|
|
1205
1205
|
cost: {
|
|
1206
1206
|
input: number;
|
|
@@ -1986,6 +1986,23 @@ export declare const MODELS: {
|
|
|
1986
1986
|
};
|
|
1987
1987
|
};
|
|
1988
1988
|
readonly openrouter: {
|
|
1989
|
+
readonly "openai/gpt-5.1-codex-max": {
|
|
1990
|
+
id: string;
|
|
1991
|
+
name: string;
|
|
1992
|
+
api: "openai-completions";
|
|
1993
|
+
provider: string;
|
|
1994
|
+
baseUrl: string;
|
|
1995
|
+
reasoning: true;
|
|
1996
|
+
input: ("image" | "text")[];
|
|
1997
|
+
cost: {
|
|
1998
|
+
input: number;
|
|
1999
|
+
output: number;
|
|
2000
|
+
cacheRead: number;
|
|
2001
|
+
cacheWrite: number;
|
|
2002
|
+
};
|
|
2003
|
+
contextWindow: number;
|
|
2004
|
+
maxTokens: number;
|
|
2005
|
+
};
|
|
1989
2006
|
readonly "amazon/nova-2-lite-v1:free": {
|
|
1990
2007
|
id: string;
|
|
1991
2008
|
name: string;
|
|
@@ -2020,6 +2037,57 @@ export declare const MODELS: {
|
|
|
2020
2037
|
contextWindow: number;
|
|
2021
2038
|
maxTokens: number;
|
|
2022
2039
|
};
|
|
2040
|
+
readonly "mistralai/ministral-14b-2512": {
|
|
2041
|
+
id: string;
|
|
2042
|
+
name: string;
|
|
2043
|
+
api: "openai-completions";
|
|
2044
|
+
provider: string;
|
|
2045
|
+
baseUrl: string;
|
|
2046
|
+
reasoning: false;
|
|
2047
|
+
input: ("image" | "text")[];
|
|
2048
|
+
cost: {
|
|
2049
|
+
input: number;
|
|
2050
|
+
output: number;
|
|
2051
|
+
cacheRead: number;
|
|
2052
|
+
cacheWrite: number;
|
|
2053
|
+
};
|
|
2054
|
+
contextWindow: number;
|
|
2055
|
+
maxTokens: number;
|
|
2056
|
+
};
|
|
2057
|
+
readonly "mistralai/ministral-8b-2512": {
|
|
2058
|
+
id: string;
|
|
2059
|
+
name: string;
|
|
2060
|
+
api: "openai-completions";
|
|
2061
|
+
provider: string;
|
|
2062
|
+
baseUrl: string;
|
|
2063
|
+
reasoning: false;
|
|
2064
|
+
input: ("image" | "text")[];
|
|
2065
|
+
cost: {
|
|
2066
|
+
input: number;
|
|
2067
|
+
output: number;
|
|
2068
|
+
cacheRead: number;
|
|
2069
|
+
cacheWrite: number;
|
|
2070
|
+
};
|
|
2071
|
+
contextWindow: number;
|
|
2072
|
+
maxTokens: number;
|
|
2073
|
+
};
|
|
2074
|
+
readonly "mistralai/ministral-3b-2512": {
|
|
2075
|
+
id: string;
|
|
2076
|
+
name: string;
|
|
2077
|
+
api: "openai-completions";
|
|
2078
|
+
provider: string;
|
|
2079
|
+
baseUrl: string;
|
|
2080
|
+
reasoning: false;
|
|
2081
|
+
input: ("image" | "text")[];
|
|
2082
|
+
cost: {
|
|
2083
|
+
input: number;
|
|
2084
|
+
output: number;
|
|
2085
|
+
cacheRead: number;
|
|
2086
|
+
cacheWrite: number;
|
|
2087
|
+
};
|
|
2088
|
+
contextWindow: number;
|
|
2089
|
+
maxTokens: number;
|
|
2090
|
+
};
|
|
2023
2091
|
readonly "mistralai/mistral-large-2512": {
|
|
2024
2092
|
id: string;
|
|
2025
2093
|
name: string;
|
|
@@ -3346,7 +3414,7 @@ export declare const MODELS: {
|
|
|
3346
3414
|
contextWindow: number;
|
|
3347
3415
|
maxTokens: number;
|
|
3348
3416
|
};
|
|
3349
|
-
readonly "openai/gpt-oss-120b:
|
|
3417
|
+
readonly "openai/gpt-oss-120b:free": {
|
|
3350
3418
|
id: string;
|
|
3351
3419
|
name: string;
|
|
3352
3420
|
api: "openai-completions";
|
|
@@ -3380,6 +3448,23 @@ export declare const MODELS: {
|
|
|
3380
3448
|
contextWindow: number;
|
|
3381
3449
|
maxTokens: number;
|
|
3382
3450
|
};
|
|
3451
|
+
readonly "openai/gpt-oss-120b:exacto": {
|
|
3452
|
+
id: string;
|
|
3453
|
+
name: string;
|
|
3454
|
+
api: "openai-completions";
|
|
3455
|
+
provider: string;
|
|
3456
|
+
baseUrl: string;
|
|
3457
|
+
reasoning: true;
|
|
3458
|
+
input: "text"[];
|
|
3459
|
+
cost: {
|
|
3460
|
+
input: number;
|
|
3461
|
+
output: number;
|
|
3462
|
+
cacheRead: number;
|
|
3463
|
+
cacheWrite: number;
|
|
3464
|
+
};
|
|
3465
|
+
contextWindow: number;
|
|
3466
|
+
maxTokens: number;
|
|
3467
|
+
};
|
|
3383
3468
|
readonly "openai/gpt-oss-20b:free": {
|
|
3384
3469
|
id: string;
|
|
3385
3470
|
name: string;
|