@mariozechner/pi-ai 0.12.10 → 0.12.11

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1075,7 +1075,7 @@ export declare const MODELS: {
1075
1075
  contextWindow: number;
1076
1076
  maxTokens: number;
1077
1077
  };
1078
- readonly o3: {
1078
+ readonly "gpt-5.1-codex-max": {
1079
1079
  id: string;
1080
1080
  name: string;
1081
1081
  api: "openai-responses";
@@ -1092,7 +1092,7 @@ export declare const MODELS: {
1092
1092
  contextWindow: number;
1093
1093
  maxTokens: number;
1094
1094
  };
1095
- readonly "o4-mini-deep-research": {
1095
+ readonly o3: {
1096
1096
  id: string;
1097
1097
  name: string;
1098
1098
  api: "openai-responses";
@@ -1109,13 +1109,13 @@ export declare const MODELS: {
1109
1109
  contextWindow: number;
1110
1110
  maxTokens: number;
1111
1111
  };
1112
- readonly "gpt-4o-mini": {
1112
+ readonly "o4-mini-deep-research": {
1113
1113
  id: string;
1114
1114
  name: string;
1115
1115
  api: "openai-responses";
1116
1116
  provider: string;
1117
1117
  baseUrl: string;
1118
- reasoning: false;
1118
+ reasoning: true;
1119
1119
  input: ("image" | "text")[];
1120
1120
  cost: {
1121
1121
  input: number;
@@ -1126,13 +1126,13 @@ export declare const MODELS: {
1126
1126
  contextWindow: number;
1127
1127
  maxTokens: number;
1128
1128
  };
1129
- readonly "gpt-5": {
1129
+ readonly "gpt-4o-mini": {
1130
1130
  id: string;
1131
1131
  name: string;
1132
1132
  api: "openai-responses";
1133
1133
  provider: string;
1134
1134
  baseUrl: string;
1135
- reasoning: true;
1135
+ reasoning: false;
1136
1136
  input: ("image" | "text")[];
1137
1137
  cost: {
1138
1138
  input: number;
@@ -1143,7 +1143,7 @@ export declare const MODELS: {
1143
1143
  contextWindow: number;
1144
1144
  maxTokens: number;
1145
1145
  };
1146
- readonly "gpt-5-pro": {
1146
+ readonly "gpt-5": {
1147
1147
  id: string;
1148
1148
  name: string;
1149
1149
  api: "openai-responses";
@@ -1160,7 +1160,7 @@ export declare const MODELS: {
1160
1160
  contextWindow: number;
1161
1161
  maxTokens: number;
1162
1162
  };
1163
- readonly "gpt-5.1-chat-latest": {
1163
+ readonly "gpt-5-pro": {
1164
1164
  id: string;
1165
1165
  name: string;
1166
1166
  api: "openai-responses";
@@ -1177,13 +1177,13 @@ export declare const MODELS: {
1177
1177
  contextWindow: number;
1178
1178
  maxTokens: number;
1179
1179
  };
1180
- readonly "gpt-5-chat-latest": {
1180
+ readonly "gpt-5.1-chat-latest": {
1181
1181
  id: string;
1182
1182
  name: string;
1183
1183
  api: "openai-responses";
1184
1184
  provider: string;
1185
1185
  baseUrl: string;
1186
- reasoning: false;
1186
+ reasoning: true;
1187
1187
  input: ("image" | "text")[];
1188
1188
  cost: {
1189
1189
  input: number;
@@ -1194,13 +1194,13 @@ export declare const MODELS: {
1194
1194
  contextWindow: number;
1195
1195
  maxTokens: number;
1196
1196
  };
1197
- readonly "gpt-5.1-codex-max": {
1197
+ readonly "gpt-5-chat-latest": {
1198
1198
  id: string;
1199
1199
  name: string;
1200
1200
  api: "openai-responses";
1201
1201
  provider: string;
1202
1202
  baseUrl: string;
1203
- reasoning: true;
1203
+ reasoning: false;
1204
1204
  input: ("image" | "text")[];
1205
1205
  cost: {
1206
1206
  input: number;
@@ -2020,6 +2020,57 @@ export declare const MODELS: {
2020
2020
  contextWindow: number;
2021
2021
  maxTokens: number;
2022
2022
  };
2023
+ readonly "mistralai/ministral-14b-2512": {
2024
+ id: string;
2025
+ name: string;
2026
+ api: "openai-completions";
2027
+ provider: string;
2028
+ baseUrl: string;
2029
+ reasoning: false;
2030
+ input: ("image" | "text")[];
2031
+ cost: {
2032
+ input: number;
2033
+ output: number;
2034
+ cacheRead: number;
2035
+ cacheWrite: number;
2036
+ };
2037
+ contextWindow: number;
2038
+ maxTokens: number;
2039
+ };
2040
+ readonly "mistralai/ministral-8b-2512": {
2041
+ id: string;
2042
+ name: string;
2043
+ api: "openai-completions";
2044
+ provider: string;
2045
+ baseUrl: string;
2046
+ reasoning: false;
2047
+ input: ("image" | "text")[];
2048
+ cost: {
2049
+ input: number;
2050
+ output: number;
2051
+ cacheRead: number;
2052
+ cacheWrite: number;
2053
+ };
2054
+ contextWindow: number;
2055
+ maxTokens: number;
2056
+ };
2057
+ readonly "mistralai/ministral-3b-2512": {
2058
+ id: string;
2059
+ name: string;
2060
+ api: "openai-completions";
2061
+ provider: string;
2062
+ baseUrl: string;
2063
+ reasoning: false;
2064
+ input: ("image" | "text")[];
2065
+ cost: {
2066
+ input: number;
2067
+ output: number;
2068
+ cacheRead: number;
2069
+ cacheWrite: number;
2070
+ };
2071
+ contextWindow: number;
2072
+ maxTokens: number;
2073
+ };
2023
2074
  readonly "mistralai/mistral-large-2512": {
2024
2075
  id: string;
2025
2076
  name: string;
@@ -3346,7 +3397,7 @@ export declare const MODELS: {
3346
3397
  contextWindow: number;
3347
3398
  maxTokens: number;
3348
3399
  };
3349
- readonly "openai/gpt-oss-120b:exacto": {
3400
+ readonly "openai/gpt-oss-120b": {
3350
3401
  id: string;
3351
3402
  name: string;
3352
3403
  api: "openai-completions";
@@ -3363,7 +3414,7 @@ export declare const MODELS: {
3363
3414
  contextWindow: number;
3364
3415
  maxTokens: number;
3365
3416
  };
3366
- readonly "openai/gpt-oss-120b": {
3417
+ readonly "openai/gpt-oss-120b:exacto": {
3367
3418
  id: string;
3368
3419
  name: string;
3369
3420
  api: "openai-completions";