@mariozechner/pi-ai 0.52.9 → 0.52.11
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/models.generated.d.ts +373 -170
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +469 -251
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/amazon-bedrock.d.ts.map +1 -1
- package/dist/providers/amazon-bedrock.js +2 -2
- package/dist/providers/amazon-bedrock.js.map +1 -1
- package/dist/providers/anthropic.d.ts.map +1 -1
- package/dist/providers/anthropic.js +51 -19
- package/dist/providers/anthropic.js.map +1 -1
- package/dist/providers/github-copilot-headers.d.ts +8 -0
- package/dist/providers/github-copilot-headers.d.ts.map +1 -0
- package/dist/providers/github-copilot-headers.js +29 -0
- package/dist/providers/github-copilot-headers.js.map +1 -0
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +7 -25
- package/dist/providers/openai-completions.js.map +1 -1
- package/dist/providers/openai-responses-shared.d.ts.map +1 -1
- package/dist/providers/openai-responses-shared.js +2 -2
- package/dist/providers/openai-responses-shared.js.map +1 -1
- package/dist/providers/openai-responses.d.ts.map +1 -1
- package/dist/providers/openai-responses.js +6 -20
- package/dist/providers/openai-responses.js.map +1 -1
- package/dist/providers/simple-options.d.ts.map +1 -1
- package/dist/providers/simple-options.js +1 -0
- package/dist/providers/simple-options.js.map +1 -1
- package/dist/types.d.ts +6 -0
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/package.json +1 -1
|
@@ -799,6 +799,23 @@ export declare const MODELS: {
|
|
|
799
799
|
contextWindow: number;
|
|
800
800
|
maxTokens: number;
|
|
801
801
|
};
|
|
802
|
+
readonly "minimax.minimax-m2.1": {
|
|
803
|
+
id: string;
|
|
804
|
+
name: string;
|
|
805
|
+
api: "bedrock-converse-stream";
|
|
806
|
+
provider: string;
|
|
807
|
+
baseUrl: string;
|
|
808
|
+
reasoning: true;
|
|
809
|
+
input: "text"[];
|
|
810
|
+
cost: {
|
|
811
|
+
input: number;
|
|
812
|
+
output: number;
|
|
813
|
+
cacheRead: number;
|
|
814
|
+
cacheWrite: number;
|
|
815
|
+
};
|
|
816
|
+
contextWindow: number;
|
|
817
|
+
maxTokens: number;
|
|
818
|
+
};
|
|
802
819
|
readonly "mistral.ministral-3-14b-instruct": {
|
|
803
820
|
id: string;
|
|
804
821
|
name: string;
|
|
@@ -901,6 +918,23 @@ export declare const MODELS: {
|
|
|
901
918
|
contextWindow: number;
|
|
902
919
|
maxTokens: number;
|
|
903
920
|
};
|
|
921
|
+
readonly "moonshotai.kimi-k2.5": {
|
|
922
|
+
id: string;
|
|
923
|
+
name: string;
|
|
924
|
+
api: "bedrock-converse-stream";
|
|
925
|
+
provider: string;
|
|
926
|
+
baseUrl: string;
|
|
927
|
+
reasoning: true;
|
|
928
|
+
input: ("image" | "text")[];
|
|
929
|
+
cost: {
|
|
930
|
+
input: number;
|
|
931
|
+
output: number;
|
|
932
|
+
cacheRead: number;
|
|
933
|
+
cacheWrite: number;
|
|
934
|
+
};
|
|
935
|
+
contextWindow: number;
|
|
936
|
+
maxTokens: number;
|
|
937
|
+
};
|
|
904
938
|
readonly "nvidia.nemotron-nano-12b-v2": {
|
|
905
939
|
id: string;
|
|
906
940
|
name: string;
|
|
@@ -1224,6 +1258,74 @@ export declare const MODELS: {
|
|
|
1224
1258
|
contextWindow: number;
|
|
1225
1259
|
maxTokens: number;
|
|
1226
1260
|
};
|
|
1261
|
+
readonly "writer.palmyra-x4-v1:0": {
|
|
1262
|
+
id: string;
|
|
1263
|
+
name: string;
|
|
1264
|
+
api: "bedrock-converse-stream";
|
|
1265
|
+
provider: string;
|
|
1266
|
+
baseUrl: string;
|
|
1267
|
+
reasoning: true;
|
|
1268
|
+
input: "text"[];
|
|
1269
|
+
cost: {
|
|
1270
|
+
input: number;
|
|
1271
|
+
output: number;
|
|
1272
|
+
cacheRead: number;
|
|
1273
|
+
cacheWrite: number;
|
|
1274
|
+
};
|
|
1275
|
+
contextWindow: number;
|
|
1276
|
+
maxTokens: number;
|
|
1277
|
+
};
|
|
1278
|
+
readonly "writer.palmyra-x5-v1:0": {
|
|
1279
|
+
id: string;
|
|
1280
|
+
name: string;
|
|
1281
|
+
api: "bedrock-converse-stream";
|
|
1282
|
+
provider: string;
|
|
1283
|
+
baseUrl: string;
|
|
1284
|
+
reasoning: true;
|
|
1285
|
+
input: "text"[];
|
|
1286
|
+
cost: {
|
|
1287
|
+
input: number;
|
|
1288
|
+
output: number;
|
|
1289
|
+
cacheRead: number;
|
|
1290
|
+
cacheWrite: number;
|
|
1291
|
+
};
|
|
1292
|
+
contextWindow: number;
|
|
1293
|
+
maxTokens: number;
|
|
1294
|
+
};
|
|
1295
|
+
readonly "zai.glm-4.7": {
|
|
1296
|
+
id: string;
|
|
1297
|
+
name: string;
|
|
1298
|
+
api: "bedrock-converse-stream";
|
|
1299
|
+
provider: string;
|
|
1300
|
+
baseUrl: string;
|
|
1301
|
+
reasoning: true;
|
|
1302
|
+
input: "text"[];
|
|
1303
|
+
cost: {
|
|
1304
|
+
input: number;
|
|
1305
|
+
output: number;
|
|
1306
|
+
cacheRead: number;
|
|
1307
|
+
cacheWrite: number;
|
|
1308
|
+
};
|
|
1309
|
+
contextWindow: number;
|
|
1310
|
+
maxTokens: number;
|
|
1311
|
+
};
|
|
1312
|
+
readonly "zai.glm-4.7-flash": {
|
|
1313
|
+
id: string;
|
|
1314
|
+
name: string;
|
|
1315
|
+
api: "bedrock-converse-stream";
|
|
1316
|
+
provider: string;
|
|
1317
|
+
baseUrl: string;
|
|
1318
|
+
reasoning: true;
|
|
1319
|
+
input: "text"[];
|
|
1320
|
+
cost: {
|
|
1321
|
+
input: number;
|
|
1322
|
+
output: number;
|
|
1323
|
+
cacheRead: number;
|
|
1324
|
+
cacheWrite: number;
|
|
1325
|
+
};
|
|
1326
|
+
contextWindow: number;
|
|
1327
|
+
maxTokens: number;
|
|
1328
|
+
};
|
|
1227
1329
|
};
|
|
1228
1330
|
readonly anthropic: {
|
|
1229
1331
|
readonly "claude-3-5-haiku-20241022": {
|
|
@@ -2061,6 +2163,23 @@ export declare const MODELS: {
|
|
|
2061
2163
|
contextWindow: number;
|
|
2062
2164
|
maxTokens: number;
|
|
2063
2165
|
};
|
|
2166
|
+
readonly "gpt-5.3-codex-spark": {
|
|
2167
|
+
id: string;
|
|
2168
|
+
name: string;
|
|
2169
|
+
api: "azure-openai-responses";
|
|
2170
|
+
provider: string;
|
|
2171
|
+
baseUrl: string;
|
|
2172
|
+
reasoning: true;
|
|
2173
|
+
input: ("image" | "text")[];
|
|
2174
|
+
cost: {
|
|
2175
|
+
input: number;
|
|
2176
|
+
output: number;
|
|
2177
|
+
cacheRead: number;
|
|
2178
|
+
cacheWrite: number;
|
|
2179
|
+
};
|
|
2180
|
+
contextWindow: number;
|
|
2181
|
+
maxTokens: number;
|
|
2182
|
+
};
|
|
2064
2183
|
readonly o1: {
|
|
2065
2184
|
id: string;
|
|
2066
2185
|
name: string;
|
|
@@ -2216,6 +2335,23 @@ export declare const MODELS: {
|
|
|
2216
2335
|
contextWindow: number;
|
|
2217
2336
|
maxTokens: number;
|
|
2218
2337
|
};
|
|
2338
|
+
readonly "llama3.1-8b": {
|
|
2339
|
+
id: string;
|
|
2340
|
+
name: string;
|
|
2341
|
+
api: "openai-completions";
|
|
2342
|
+
provider: string;
|
|
2343
|
+
baseUrl: string;
|
|
2344
|
+
reasoning: false;
|
|
2345
|
+
input: "text"[];
|
|
2346
|
+
cost: {
|
|
2347
|
+
input: number;
|
|
2348
|
+
output: number;
|
|
2349
|
+
cacheRead: number;
|
|
2350
|
+
cacheWrite: number;
|
|
2351
|
+
};
|
|
2352
|
+
contextWindow: number;
|
|
2353
|
+
maxTokens: number;
|
|
2354
|
+
};
|
|
2219
2355
|
readonly "qwen-3-235b-a22b-instruct-2507": {
|
|
2220
2356
|
id: string;
|
|
2221
2357
|
name: string;
|
|
@@ -2255,7 +2391,7 @@ export declare const MODELS: {
|
|
|
2255
2391
|
readonly "claude-haiku-4.5": {
|
|
2256
2392
|
id: string;
|
|
2257
2393
|
name: string;
|
|
2258
|
-
api: "
|
|
2394
|
+
api: "anthropic-messages";
|
|
2259
2395
|
provider: string;
|
|
2260
2396
|
baseUrl: string;
|
|
2261
2397
|
headers: {
|
|
@@ -2264,11 +2400,6 @@ export declare const MODELS: {
|
|
|
2264
2400
|
"Editor-Plugin-Version": string;
|
|
2265
2401
|
"Copilot-Integration-Id": string;
|
|
2266
2402
|
};
|
|
2267
|
-
compat: {
|
|
2268
|
-
supportsStore: false;
|
|
2269
|
-
supportsDeveloperRole: false;
|
|
2270
|
-
supportsReasoningEffort: false;
|
|
2271
|
-
};
|
|
2272
2403
|
reasoning: true;
|
|
2273
2404
|
input: ("image" | "text")[];
|
|
2274
2405
|
cost: {
|
|
@@ -2283,7 +2414,7 @@ export declare const MODELS: {
|
|
|
2283
2414
|
readonly "claude-opus-4.5": {
|
|
2284
2415
|
id: string;
|
|
2285
2416
|
name: string;
|
|
2286
|
-
api: "
|
|
2417
|
+
api: "anthropic-messages";
|
|
2287
2418
|
provider: string;
|
|
2288
2419
|
baseUrl: string;
|
|
2289
2420
|
headers: {
|
|
@@ -2292,11 +2423,6 @@ export declare const MODELS: {
|
|
|
2292
2423
|
"Editor-Plugin-Version": string;
|
|
2293
2424
|
"Copilot-Integration-Id": string;
|
|
2294
2425
|
};
|
|
2295
|
-
compat: {
|
|
2296
|
-
supportsStore: false;
|
|
2297
|
-
supportsDeveloperRole: false;
|
|
2298
|
-
supportsReasoningEffort: false;
|
|
2299
|
-
};
|
|
2300
2426
|
reasoning: true;
|
|
2301
2427
|
input: ("image" | "text")[];
|
|
2302
2428
|
cost: {
|
|
@@ -2311,7 +2437,7 @@ export declare const MODELS: {
|
|
|
2311
2437
|
readonly "claude-opus-4.6": {
|
|
2312
2438
|
id: string;
|
|
2313
2439
|
name: string;
|
|
2314
|
-
api: "
|
|
2440
|
+
api: "anthropic-messages";
|
|
2315
2441
|
provider: string;
|
|
2316
2442
|
baseUrl: string;
|
|
2317
2443
|
headers: {
|
|
@@ -2320,11 +2446,6 @@ export declare const MODELS: {
|
|
|
2320
2446
|
"Editor-Plugin-Version": string;
|
|
2321
2447
|
"Copilot-Integration-Id": string;
|
|
2322
2448
|
};
|
|
2323
|
-
compat: {
|
|
2324
|
-
supportsStore: false;
|
|
2325
|
-
supportsDeveloperRole: false;
|
|
2326
|
-
supportsReasoningEffort: false;
|
|
2327
|
-
};
|
|
2328
2449
|
reasoning: true;
|
|
2329
2450
|
input: ("image" | "text")[];
|
|
2330
2451
|
cost: {
|
|
@@ -2339,7 +2460,7 @@ export declare const MODELS: {
|
|
|
2339
2460
|
readonly "claude-sonnet-4": {
|
|
2340
2461
|
id: string;
|
|
2341
2462
|
name: string;
|
|
2342
|
-
api: "
|
|
2463
|
+
api: "anthropic-messages";
|
|
2343
2464
|
provider: string;
|
|
2344
2465
|
baseUrl: string;
|
|
2345
2466
|
headers: {
|
|
@@ -2348,11 +2469,6 @@ export declare const MODELS: {
|
|
|
2348
2469
|
"Editor-Plugin-Version": string;
|
|
2349
2470
|
"Copilot-Integration-Id": string;
|
|
2350
2471
|
};
|
|
2351
|
-
compat: {
|
|
2352
|
-
supportsStore: false;
|
|
2353
|
-
supportsDeveloperRole: false;
|
|
2354
|
-
supportsReasoningEffort: false;
|
|
2355
|
-
};
|
|
2356
2472
|
reasoning: true;
|
|
2357
2473
|
input: ("image" | "text")[];
|
|
2358
2474
|
cost: {
|
|
@@ -2367,7 +2483,7 @@ export declare const MODELS: {
|
|
|
2367
2483
|
readonly "claude-sonnet-4.5": {
|
|
2368
2484
|
id: string;
|
|
2369
2485
|
name: string;
|
|
2370
|
-
api: "
|
|
2486
|
+
api: "anthropic-messages";
|
|
2371
2487
|
provider: string;
|
|
2372
2488
|
baseUrl: string;
|
|
2373
2489
|
headers: {
|
|
@@ -2376,11 +2492,6 @@ export declare const MODELS: {
|
|
|
2376
2492
|
"Editor-Plugin-Version": string;
|
|
2377
2493
|
"Copilot-Integration-Id": string;
|
|
2378
2494
|
};
|
|
2379
|
-
compat: {
|
|
2380
|
-
supportsStore: false;
|
|
2381
|
-
supportsDeveloperRole: false;
|
|
2382
|
-
supportsReasoningEffort: false;
|
|
2383
|
-
};
|
|
2384
2495
|
reasoning: true;
|
|
2385
2496
|
input: ("image" | "text")[];
|
|
2386
2497
|
cost: {
|
|
@@ -4039,6 +4150,26 @@ export declare const MODELS: {
|
|
|
4039
4150
|
contextWindow: number;
|
|
4040
4151
|
maxTokens: number;
|
|
4041
4152
|
};
|
|
4153
|
+
readonly "zai-org/GLM-5": {
|
|
4154
|
+
id: string;
|
|
4155
|
+
name: string;
|
|
4156
|
+
api: "openai-completions";
|
|
4157
|
+
provider: string;
|
|
4158
|
+
baseUrl: string;
|
|
4159
|
+
compat: {
|
|
4160
|
+
supportsDeveloperRole: false;
|
|
4161
|
+
};
|
|
4162
|
+
reasoning: true;
|
|
4163
|
+
input: "text"[];
|
|
4164
|
+
cost: {
|
|
4165
|
+
input: number;
|
|
4166
|
+
output: number;
|
|
4167
|
+
cacheRead: number;
|
|
4168
|
+
cacheWrite: number;
|
|
4169
|
+
};
|
|
4170
|
+
contextWindow: number;
|
|
4171
|
+
maxTokens: number;
|
|
4172
|
+
};
|
|
4042
4173
|
};
|
|
4043
4174
|
readonly "kimi-coding": {
|
|
4044
4175
|
readonly k2p5: {
|
|
@@ -4111,6 +4242,23 @@ export declare const MODELS: {
|
|
|
4111
4242
|
contextWindow: number;
|
|
4112
4243
|
maxTokens: number;
|
|
4113
4244
|
};
|
|
4245
|
+
readonly "MiniMax-M2.5": {
|
|
4246
|
+
id: string;
|
|
4247
|
+
name: string;
|
|
4248
|
+
api: "anthropic-messages";
|
|
4249
|
+
provider: string;
|
|
4250
|
+
baseUrl: string;
|
|
4251
|
+
reasoning: true;
|
|
4252
|
+
input: "text"[];
|
|
4253
|
+
cost: {
|
|
4254
|
+
input: number;
|
|
4255
|
+
output: number;
|
|
4256
|
+
cacheRead: number;
|
|
4257
|
+
cacheWrite: number;
|
|
4258
|
+
};
|
|
4259
|
+
contextWindow: number;
|
|
4260
|
+
maxTokens: number;
|
|
4261
|
+
};
|
|
4114
4262
|
};
|
|
4115
4263
|
readonly "minimax-cn": {
|
|
4116
4264
|
readonly "MiniMax-M2": {
|
|
@@ -4147,6 +4295,23 @@ export declare const MODELS: {
|
|
|
4147
4295
|
contextWindow: number;
|
|
4148
4296
|
maxTokens: number;
|
|
4149
4297
|
};
|
|
4298
|
+
readonly "MiniMax-M2.5": {
|
|
4299
|
+
id: string;
|
|
4300
|
+
name: string;
|
|
4301
|
+
api: "anthropic-messages";
|
|
4302
|
+
provider: string;
|
|
4303
|
+
baseUrl: string;
|
|
4304
|
+
reasoning: true;
|
|
4305
|
+
input: "text"[];
|
|
4306
|
+
cost: {
|
|
4307
|
+
input: number;
|
|
4308
|
+
output: number;
|
|
4309
|
+
cacheRead: number;
|
|
4310
|
+
cacheWrite: number;
|
|
4311
|
+
};
|
|
4312
|
+
contextWindow: number;
|
|
4313
|
+
maxTokens: number;
|
|
4314
|
+
};
|
|
4150
4315
|
};
|
|
4151
4316
|
readonly mistral: {
|
|
4152
4317
|
readonly "codestral-latest": {
|
|
@@ -5035,6 +5200,23 @@ export declare const MODELS: {
|
|
|
5035
5200
|
contextWindow: number;
|
|
5036
5201
|
maxTokens: number;
|
|
5037
5202
|
};
|
|
5203
|
+
readonly "gpt-5.3-codex-spark": {
|
|
5204
|
+
id: string;
|
|
5205
|
+
name: string;
|
|
5206
|
+
api: "openai-responses";
|
|
5207
|
+
provider: string;
|
|
5208
|
+
baseUrl: string;
|
|
5209
|
+
reasoning: true;
|
|
5210
|
+
input: ("image" | "text")[];
|
|
5211
|
+
cost: {
|
|
5212
|
+
input: number;
|
|
5213
|
+
output: number;
|
|
5214
|
+
cacheRead: number;
|
|
5215
|
+
cacheWrite: number;
|
|
5216
|
+
};
|
|
5217
|
+
contextWindow: number;
|
|
5218
|
+
maxTokens: number;
|
|
5219
|
+
};
|
|
5038
5220
|
readonly o1: {
|
|
5039
5221
|
id: string;
|
|
5040
5222
|
name: string;
|
|
@@ -5275,6 +5457,23 @@ export declare const MODELS: {
|
|
|
5275
5457
|
contextWindow: number;
|
|
5276
5458
|
maxTokens: number;
|
|
5277
5459
|
};
|
|
5460
|
+
readonly "gpt-5.3-codex-spark": {
|
|
5461
|
+
id: string;
|
|
5462
|
+
name: string;
|
|
5463
|
+
api: "openai-codex-responses";
|
|
5464
|
+
provider: string;
|
|
5465
|
+
baseUrl: string;
|
|
5466
|
+
reasoning: true;
|
|
5467
|
+
input: "text"[];
|
|
5468
|
+
cost: {
|
|
5469
|
+
input: number;
|
|
5470
|
+
output: number;
|
|
5471
|
+
cacheRead: number;
|
|
5472
|
+
cacheWrite: number;
|
|
5473
|
+
};
|
|
5474
|
+
contextWindow: number;
|
|
5475
|
+
maxTokens: number;
|
|
5476
|
+
};
|
|
5278
5477
|
};
|
|
5279
5478
|
readonly opencode: {
|
|
5280
5479
|
readonly "big-pickle": {
|
|
@@ -5481,23 +5680,6 @@ export declare const MODELS: {
|
|
|
5481
5680
|
contextWindow: number;
|
|
5482
5681
|
maxTokens: number;
|
|
5483
5682
|
};
|
|
5484
|
-
readonly "glm-4.7-free": {
|
|
5485
|
-
id: string;
|
|
5486
|
-
name: string;
|
|
5487
|
-
api: "openai-completions";
|
|
5488
|
-
provider: string;
|
|
5489
|
-
baseUrl: string;
|
|
5490
|
-
reasoning: true;
|
|
5491
|
-
input: "text"[];
|
|
5492
|
-
cost: {
|
|
5493
|
-
input: number;
|
|
5494
|
-
output: number;
|
|
5495
|
-
cacheRead: number;
|
|
5496
|
-
cacheWrite: number;
|
|
5497
|
-
};
|
|
5498
|
-
contextWindow: number;
|
|
5499
|
-
maxTokens: number;
|
|
5500
|
-
};
|
|
5501
5683
|
readonly "gpt-5": {
|
|
5502
5684
|
id: string;
|
|
5503
5685
|
name: string;
|
|
@@ -5736,47 +5918,13 @@ export declare const MODELS: {
|
|
|
5736
5918
|
contextWindow: number;
|
|
5737
5919
|
maxTokens: number;
|
|
5738
5920
|
};
|
|
5739
|
-
readonly "minimax-m2.
|
|
5740
|
-
id: string;
|
|
5741
|
-
name: string;
|
|
5742
|
-
api: "anthropic-messages";
|
|
5743
|
-
provider: string;
|
|
5744
|
-
baseUrl: string;
|
|
5745
|
-
reasoning: true;
|
|
5746
|
-
input: "text"[];
|
|
5747
|
-
cost: {
|
|
5748
|
-
input: number;
|
|
5749
|
-
output: number;
|
|
5750
|
-
cacheRead: number;
|
|
5751
|
-
cacheWrite: number;
|
|
5752
|
-
};
|
|
5753
|
-
contextWindow: number;
|
|
5754
|
-
maxTokens: number;
|
|
5755
|
-
};
|
|
5756
|
-
readonly "qwen3-coder": {
|
|
5757
|
-
id: string;
|
|
5758
|
-
name: string;
|
|
5759
|
-
api: "openai-completions";
|
|
5760
|
-
provider: string;
|
|
5761
|
-
baseUrl: string;
|
|
5762
|
-
reasoning: false;
|
|
5763
|
-
input: "text"[];
|
|
5764
|
-
cost: {
|
|
5765
|
-
input: number;
|
|
5766
|
-
output: number;
|
|
5767
|
-
cacheRead: number;
|
|
5768
|
-
cacheWrite: number;
|
|
5769
|
-
};
|
|
5770
|
-
contextWindow: number;
|
|
5771
|
-
maxTokens: number;
|
|
5772
|
-
};
|
|
5773
|
-
readonly "trinity-large-preview-free": {
|
|
5921
|
+
readonly "minimax-m2.5-free": {
|
|
5774
5922
|
id: string;
|
|
5775
5923
|
name: string;
|
|
5776
5924
|
api: "openai-completions";
|
|
5777
5925
|
provider: string;
|
|
5778
5926
|
baseUrl: string;
|
|
5779
|
-
reasoning:
|
|
5927
|
+
reasoning: true;
|
|
5780
5928
|
input: "text"[];
|
|
5781
5929
|
cost: {
|
|
5782
5930
|
input: number;
|
|
@@ -5806,23 +5954,6 @@ export declare const MODELS: {
|
|
|
5806
5954
|
contextWindow: number;
|
|
5807
5955
|
maxTokens: number;
|
|
5808
5956
|
};
|
|
5809
|
-
readonly "ai21/jamba-mini-1.7": {
|
|
5810
|
-
id: string;
|
|
5811
|
-
name: string;
|
|
5812
|
-
api: "openai-completions";
|
|
5813
|
-
provider: string;
|
|
5814
|
-
baseUrl: string;
|
|
5815
|
-
reasoning: false;
|
|
5816
|
-
input: "text"[];
|
|
5817
|
-
cost: {
|
|
5818
|
-
input: number;
|
|
5819
|
-
output: number;
|
|
5820
|
-
cacheRead: number;
|
|
5821
|
-
cacheWrite: number;
|
|
5822
|
-
};
|
|
5823
|
-
contextWindow: number;
|
|
5824
|
-
maxTokens: number;
|
|
5825
|
-
};
|
|
5826
5957
|
readonly "alibaba/tongyi-deepresearch-30b-a3b": {
|
|
5827
5958
|
id: string;
|
|
5828
5959
|
name: string;
|
|
@@ -6945,13 +7076,13 @@ export declare const MODELS: {
|
|
|
6945
7076
|
contextWindow: number;
|
|
6946
7077
|
maxTokens: number;
|
|
6947
7078
|
};
|
|
6948
|
-
readonly "
|
|
7079
|
+
readonly "minimax/minimax-m2.5": {
|
|
6949
7080
|
id: string;
|
|
6950
7081
|
name: string;
|
|
6951
7082
|
api: "openai-completions";
|
|
6952
7083
|
provider: string;
|
|
6953
7084
|
baseUrl: string;
|
|
6954
|
-
reasoning:
|
|
7085
|
+
reasoning: true;
|
|
6955
7086
|
input: "text"[];
|
|
6956
7087
|
cost: {
|
|
6957
7088
|
input: number;
|
|
@@ -6962,7 +7093,7 @@ export declare const MODELS: {
|
|
|
6962
7093
|
contextWindow: number;
|
|
6963
7094
|
maxTokens: number;
|
|
6964
7095
|
};
|
|
6965
|
-
readonly "mistralai/
|
|
7096
|
+
readonly "mistralai/codestral-2508": {
|
|
6966
7097
|
id: string;
|
|
6967
7098
|
name: string;
|
|
6968
7099
|
api: "openai-completions";
|
|
@@ -6979,7 +7110,7 @@ export declare const MODELS: {
|
|
|
6979
7110
|
contextWindow: number;
|
|
6980
7111
|
maxTokens: number;
|
|
6981
7112
|
};
|
|
6982
|
-
readonly "mistralai/devstral-
|
|
7113
|
+
readonly "mistralai/devstral-2512": {
|
|
6983
7114
|
id: string;
|
|
6984
7115
|
name: string;
|
|
6985
7116
|
api: "openai-completions";
|
|
@@ -6996,7 +7127,7 @@ export declare const MODELS: {
|
|
|
6996
7127
|
contextWindow: number;
|
|
6997
7128
|
maxTokens: number;
|
|
6998
7129
|
};
|
|
6999
|
-
readonly "mistralai/devstral-
|
|
7130
|
+
readonly "mistralai/devstral-medium": {
|
|
7000
7131
|
id: string;
|
|
7001
7132
|
name: string;
|
|
7002
7133
|
api: "openai-completions";
|
|
@@ -7013,24 +7144,7 @@ export declare const MODELS: {
|
|
|
7013
7144
|
contextWindow: number;
|
|
7014
7145
|
maxTokens: number;
|
|
7015
7146
|
};
|
|
7016
|
-
readonly "mistralai/
|
|
7017
|
-
id: string;
|
|
7018
|
-
name: string;
|
|
7019
|
-
api: "openai-completions";
|
|
7020
|
-
provider: string;
|
|
7021
|
-
baseUrl: string;
|
|
7022
|
-
reasoning: false;
|
|
7023
|
-
input: ("image" | "text")[];
|
|
7024
|
-
cost: {
|
|
7025
|
-
input: number;
|
|
7026
|
-
output: number;
|
|
7027
|
-
cacheRead: number;
|
|
7028
|
-
cacheWrite: number;
|
|
7029
|
-
};
|
|
7030
|
-
contextWindow: number;
|
|
7031
|
-
maxTokens: number;
|
|
7032
|
-
};
|
|
7033
|
-
readonly "mistralai/ministral-3b": {
|
|
7147
|
+
readonly "mistralai/devstral-small": {
|
|
7034
7148
|
id: string;
|
|
7035
7149
|
name: string;
|
|
7036
7150
|
api: "openai-completions";
|
|
@@ -7047,7 +7161,7 @@ export declare const MODELS: {
|
|
|
7047
7161
|
contextWindow: number;
|
|
7048
7162
|
maxTokens: number;
|
|
7049
7163
|
};
|
|
7050
|
-
readonly "mistralai/ministral-
|
|
7164
|
+
readonly "mistralai/ministral-14b-2512": {
|
|
7051
7165
|
id: string;
|
|
7052
7166
|
name: string;
|
|
7053
7167
|
api: "openai-completions";
|
|
@@ -7064,14 +7178,14 @@ export declare const MODELS: {
|
|
|
7064
7178
|
contextWindow: number;
|
|
7065
7179
|
maxTokens: number;
|
|
7066
7180
|
};
|
|
7067
|
-
readonly "mistralai/ministral-
|
|
7181
|
+
readonly "mistralai/ministral-3b-2512": {
|
|
7068
7182
|
id: string;
|
|
7069
7183
|
name: string;
|
|
7070
7184
|
api: "openai-completions";
|
|
7071
7185
|
provider: string;
|
|
7072
7186
|
baseUrl: string;
|
|
7073
7187
|
reasoning: false;
|
|
7074
|
-
input: "text"[];
|
|
7188
|
+
input: ("image" | "text")[];
|
|
7075
7189
|
cost: {
|
|
7076
7190
|
input: number;
|
|
7077
7191
|
output: number;
|
|
@@ -7319,23 +7433,6 @@ export declare const MODELS: {
|
|
|
7319
7433
|
contextWindow: number;
|
|
7320
7434
|
maxTokens: number;
|
|
7321
7435
|
};
|
|
7322
|
-
readonly "mistralai/mistral-tiny": {
|
|
7323
|
-
id: string;
|
|
7324
|
-
name: string;
|
|
7325
|
-
api: "openai-completions";
|
|
7326
|
-
provider: string;
|
|
7327
|
-
baseUrl: string;
|
|
7328
|
-
reasoning: false;
|
|
7329
|
-
input: "text"[];
|
|
7330
|
-
cost: {
|
|
7331
|
-
input: number;
|
|
7332
|
-
output: number;
|
|
7333
|
-
cacheRead: number;
|
|
7334
|
-
cacheWrite: number;
|
|
7335
|
-
};
|
|
7336
|
-
contextWindow: number;
|
|
7337
|
-
maxTokens: number;
|
|
7338
|
-
};
|
|
7339
7436
|
readonly "mistralai/mixtral-8x22b-instruct": {
|
|
7340
7437
|
id: string;
|
|
7341
7438
|
name: string;
|
|
@@ -7370,23 +7467,6 @@ export declare const MODELS: {
|
|
|
7370
7467
|
contextWindow: number;
|
|
7371
7468
|
maxTokens: number;
|
|
7372
7469
|
};
|
|
7373
|
-
readonly "mistralai/pixtral-12b": {
|
|
7374
|
-
id: string;
|
|
7375
|
-
name: string;
|
|
7376
|
-
api: "openai-completions";
|
|
7377
|
-
provider: string;
|
|
7378
|
-
baseUrl: string;
|
|
7379
|
-
reasoning: false;
|
|
7380
|
-
input: ("image" | "text")[];
|
|
7381
|
-
cost: {
|
|
7382
|
-
input: number;
|
|
7383
|
-
output: number;
|
|
7384
|
-
cacheRead: number;
|
|
7385
|
-
cacheWrite: number;
|
|
7386
|
-
};
|
|
7387
|
-
contextWindow: number;
|
|
7388
|
-
maxTokens: number;
|
|
7389
|
-
};
|
|
7390
7470
|
readonly "mistralai/pixtral-large-2411": {
|
|
7391
7471
|
id: string;
|
|
7392
7472
|
name: string;
|
|
@@ -8526,14 +8606,14 @@ export declare const MODELS: {
|
|
|
8526
8606
|
contextWindow: number;
|
|
8527
8607
|
maxTokens: number;
|
|
8528
8608
|
};
|
|
8529
|
-
readonly "openrouter/
|
|
8609
|
+
readonly "openrouter/aurora-alpha": {
|
|
8530
8610
|
id: string;
|
|
8531
8611
|
name: string;
|
|
8532
8612
|
api: "openai-completions";
|
|
8533
8613
|
provider: string;
|
|
8534
8614
|
baseUrl: string;
|
|
8535
8615
|
reasoning: true;
|
|
8536
|
-
input:
|
|
8616
|
+
input: "text"[];
|
|
8537
8617
|
cost: {
|
|
8538
8618
|
input: number;
|
|
8539
8619
|
output: number;
|
|
@@ -8543,7 +8623,7 @@ export declare const MODELS: {
|
|
|
8543
8623
|
contextWindow: number;
|
|
8544
8624
|
maxTokens: number;
|
|
8545
8625
|
};
|
|
8546
|
-
readonly "openrouter/
|
|
8626
|
+
readonly "openrouter/auto": {
|
|
8547
8627
|
id: string;
|
|
8548
8628
|
name: string;
|
|
8549
8629
|
api: "openai-completions";
|
|
@@ -8560,14 +8640,14 @@ export declare const MODELS: {
|
|
|
8560
8640
|
contextWindow: number;
|
|
8561
8641
|
maxTokens: number;
|
|
8562
8642
|
};
|
|
8563
|
-
readonly "openrouter/
|
|
8643
|
+
readonly "openrouter/free": {
|
|
8564
8644
|
id: string;
|
|
8565
8645
|
name: string;
|
|
8566
8646
|
api: "openai-completions";
|
|
8567
8647
|
provider: string;
|
|
8568
8648
|
baseUrl: string;
|
|
8569
8649
|
reasoning: true;
|
|
8570
|
-
input: "text"[];
|
|
8650
|
+
input: ("image" | "text")[];
|
|
8571
8651
|
cost: {
|
|
8572
8652
|
input: number;
|
|
8573
8653
|
output: number;
|
|
@@ -8866,6 +8946,23 @@ export declare const MODELS: {
|
|
|
8866
8946
|
contextWindow: number;
|
|
8867
8947
|
maxTokens: number;
|
|
8868
8948
|
};
|
|
8949
|
+
readonly "qwen/qwen3-4b": {
|
|
8950
|
+
id: string;
|
|
8951
|
+
name: string;
|
|
8952
|
+
api: "openai-completions";
|
|
8953
|
+
provider: string;
|
|
8954
|
+
baseUrl: string;
|
|
8955
|
+
reasoning: true;
|
|
8956
|
+
input: "text"[];
|
|
8957
|
+
cost: {
|
|
8958
|
+
input: number;
|
|
8959
|
+
output: number;
|
|
8960
|
+
cacheRead: number;
|
|
8961
|
+
cacheWrite: number;
|
|
8962
|
+
};
|
|
8963
|
+
contextWindow: number;
|
|
8964
|
+
maxTokens: number;
|
|
8965
|
+
};
|
|
8869
8966
|
readonly "qwen/qwen3-4b:free": {
|
|
8870
8967
|
id: string;
|
|
8871
8968
|
name: string;
|
|
@@ -8906,7 +9003,7 @@ export declare const MODELS: {
|
|
|
8906
9003
|
api: "openai-completions";
|
|
8907
9004
|
provider: string;
|
|
8908
9005
|
baseUrl: string;
|
|
8909
|
-
reasoning:
|
|
9006
|
+
reasoning: false;
|
|
8910
9007
|
input: "text"[];
|
|
8911
9008
|
cost: {
|
|
8912
9009
|
input: number;
|
|
@@ -9036,6 +9133,23 @@ export declare const MODELS: {
|
|
|
9036
9133
|
contextWindow: number;
|
|
9037
9134
|
maxTokens: number;
|
|
9038
9135
|
};
|
|
9136
|
+
readonly "qwen/qwen3-max-thinking": {
|
|
9137
|
+
id: string;
|
|
9138
|
+
name: string;
|
|
9139
|
+
api: "openai-completions";
|
|
9140
|
+
provider: string;
|
|
9141
|
+
baseUrl: string;
|
|
9142
|
+
reasoning: true;
|
|
9143
|
+
input: "text"[];
|
|
9144
|
+
cost: {
|
|
9145
|
+
input: number;
|
|
9146
|
+
output: number;
|
|
9147
|
+
cacheRead: number;
|
|
9148
|
+
cacheWrite: number;
|
|
9149
|
+
};
|
|
9150
|
+
contextWindow: number;
|
|
9151
|
+
maxTokens: number;
|
|
9152
|
+
};
|
|
9039
9153
|
readonly "qwen/qwen3-next-80b-a3b-instruct": {
|
|
9040
9154
|
id: string;
|
|
9041
9155
|
name: string;
|
|
@@ -9155,6 +9269,23 @@ export declare const MODELS: {
|
|
|
9155
9269
|
contextWindow: number;
|
|
9156
9270
|
maxTokens: number;
|
|
9157
9271
|
};
|
|
9272
|
+
readonly "qwen/qwen3-vl-32b-instruct": {
|
|
9273
|
+
id: string;
|
|
9274
|
+
name: string;
|
|
9275
|
+
api: "openai-completions";
|
|
9276
|
+
provider: string;
|
|
9277
|
+
baseUrl: string;
|
|
9278
|
+
reasoning: false;
|
|
9279
|
+
input: ("image" | "text")[];
|
|
9280
|
+
cost: {
|
|
9281
|
+
input: number;
|
|
9282
|
+
output: number;
|
|
9283
|
+
cacheRead: number;
|
|
9284
|
+
cacheWrite: number;
|
|
9285
|
+
};
|
|
9286
|
+
contextWindow: number;
|
|
9287
|
+
maxTokens: number;
|
|
9288
|
+
};
|
|
9158
9289
|
readonly "qwen/qwen3-vl-8b-instruct": {
|
|
9159
9290
|
id: string;
|
|
9160
9291
|
name: string;
|
|
@@ -9257,14 +9388,14 @@ export declare const MODELS: {
|
|
|
9257
9388
|
contextWindow: number;
|
|
9258
9389
|
maxTokens: number;
|
|
9259
9390
|
};
|
|
9260
|
-
readonly "stepfun-
|
|
9391
|
+
readonly "stepfun/step-3.5-flash": {
|
|
9261
9392
|
id: string;
|
|
9262
9393
|
name: string;
|
|
9263
9394
|
api: "openai-completions";
|
|
9264
9395
|
provider: string;
|
|
9265
9396
|
baseUrl: string;
|
|
9266
9397
|
reasoning: true;
|
|
9267
|
-
input:
|
|
9398
|
+
input: "text"[];
|
|
9268
9399
|
cost: {
|
|
9269
9400
|
input: number;
|
|
9270
9401
|
output: number;
|
|
@@ -9716,6 +9847,23 @@ export declare const MODELS: {
|
|
|
9716
9847
|
contextWindow: number;
|
|
9717
9848
|
maxTokens: number;
|
|
9718
9849
|
};
|
|
9850
|
+
readonly "z-ai/glm-5": {
|
|
9851
|
+
id: string;
|
|
9852
|
+
name: string;
|
|
9853
|
+
api: "openai-completions";
|
|
9854
|
+
provider: string;
|
|
9855
|
+
baseUrl: string;
|
|
9856
|
+
reasoning: true;
|
|
9857
|
+
input: "text"[];
|
|
9858
|
+
cost: {
|
|
9859
|
+
input: number;
|
|
9860
|
+
output: number;
|
|
9861
|
+
cacheRead: number;
|
|
9862
|
+
cacheWrite: number;
|
|
9863
|
+
};
|
|
9864
|
+
contextWindow: number;
|
|
9865
|
+
maxTokens: number;
|
|
9866
|
+
};
|
|
9719
9867
|
};
|
|
9720
9868
|
readonly "vercel-ai-gateway": {
|
|
9721
9869
|
readonly "alibaba/qwen-3-14b": {
|
|
@@ -10228,13 +10376,13 @@ export declare const MODELS: {
|
|
|
10228
10376
|
contextWindow: number;
|
|
10229
10377
|
maxTokens: number;
|
|
10230
10378
|
};
|
|
10231
|
-
readonly "deepseek/deepseek-v3.2
|
|
10379
|
+
readonly "deepseek/deepseek-v3.2": {
|
|
10232
10380
|
id: string;
|
|
10233
10381
|
name: string;
|
|
10234
10382
|
api: "anthropic-messages";
|
|
10235
10383
|
provider: string;
|
|
10236
10384
|
baseUrl: string;
|
|
10237
|
-
reasoning:
|
|
10385
|
+
reasoning: false;
|
|
10238
10386
|
input: "text"[];
|
|
10239
10387
|
cost: {
|
|
10240
10388
|
input: number;
|
|
@@ -10602,6 +10750,23 @@ export declare const MODELS: {
|
|
|
10602
10750
|
contextWindow: number;
|
|
10603
10751
|
maxTokens: number;
|
|
10604
10752
|
};
|
|
10753
|
+
readonly "minimax/minimax-m2.5": {
|
|
10754
|
+
id: string;
|
|
10755
|
+
name: string;
|
|
10756
|
+
api: "anthropic-messages";
|
|
10757
|
+
provider: string;
|
|
10758
|
+
baseUrl: string;
|
|
10759
|
+
reasoning: true;
|
|
10760
|
+
input: "text"[];
|
|
10761
|
+
cost: {
|
|
10762
|
+
input: number;
|
|
10763
|
+
output: number;
|
|
10764
|
+
cacheRead: number;
|
|
10765
|
+
cacheWrite: number;
|
|
10766
|
+
};
|
|
10767
|
+
contextWindow: number;
|
|
10768
|
+
maxTokens: number;
|
|
10769
|
+
};
|
|
10605
10770
|
readonly "mistral/codestral": {
|
|
10606
10771
|
id: string;
|
|
10607
10772
|
name: string;
|
|
@@ -11843,6 +12008,23 @@ export declare const MODELS: {
|
|
|
11843
12008
|
contextWindow: number;
|
|
11844
12009
|
maxTokens: number;
|
|
11845
12010
|
};
|
|
12011
|
+
readonly "zai/glm-5": {
|
|
12012
|
+
id: string;
|
|
12013
|
+
name: string;
|
|
12014
|
+
api: "anthropic-messages";
|
|
12015
|
+
provider: string;
|
|
12016
|
+
baseUrl: string;
|
|
12017
|
+
reasoning: true;
|
|
12018
|
+
input: "text"[];
|
|
12019
|
+
cost: {
|
|
12020
|
+
input: number;
|
|
12021
|
+
output: number;
|
|
12022
|
+
cacheRead: number;
|
|
12023
|
+
cacheWrite: number;
|
|
12024
|
+
};
|
|
12025
|
+
contextWindow: number;
|
|
12026
|
+
maxTokens: number;
|
|
12027
|
+
};
|
|
11846
12028
|
};
|
|
11847
12029
|
readonly xai: {
|
|
11848
12030
|
readonly "grok-2": {
|
|
@@ -12389,6 +12571,27 @@ export declare const MODELS: {
|
|
|
12389
12571
|
contextWindow: number;
|
|
12390
12572
|
maxTokens: number;
|
|
12391
12573
|
};
|
|
12574
|
+
readonly "glm-5": {
|
|
12575
|
+
id: string;
|
|
12576
|
+
name: string;
|
|
12577
|
+
api: "openai-completions";
|
|
12578
|
+
provider: string;
|
|
12579
|
+
baseUrl: string;
|
|
12580
|
+
compat: {
|
|
12581
|
+
supportsDeveloperRole: false;
|
|
12582
|
+
thinkingFormat: "zai";
|
|
12583
|
+
};
|
|
12584
|
+
reasoning: true;
|
|
12585
|
+
input: "text"[];
|
|
12586
|
+
cost: {
|
|
12587
|
+
input: number;
|
|
12588
|
+
output: number;
|
|
12589
|
+
cacheRead: number;
|
|
12590
|
+
cacheWrite: number;
|
|
12591
|
+
};
|
|
12592
|
+
contextWindow: number;
|
|
12593
|
+
maxTokens: number;
|
|
12594
|
+
};
|
|
12392
12595
|
};
|
|
12393
12596
|
};
|
|
12394
12597
|
//# sourceMappingURL=models.generated.d.ts.map
|