@mariozechner/pi-ai 0.52.9 → 0.52.11

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (30) hide show
  1. package/dist/models.generated.d.ts +373 -170
  2. package/dist/models.generated.d.ts.map +1 -1
  3. package/dist/models.generated.js +469 -251
  4. package/dist/models.generated.js.map +1 -1
  5. package/dist/providers/amazon-bedrock.d.ts.map +1 -1
  6. package/dist/providers/amazon-bedrock.js +2 -2
  7. package/dist/providers/amazon-bedrock.js.map +1 -1
  8. package/dist/providers/anthropic.d.ts.map +1 -1
  9. package/dist/providers/anthropic.js +51 -19
  10. package/dist/providers/anthropic.js.map +1 -1
  11. package/dist/providers/github-copilot-headers.d.ts +8 -0
  12. package/dist/providers/github-copilot-headers.d.ts.map +1 -0
  13. package/dist/providers/github-copilot-headers.js +29 -0
  14. package/dist/providers/github-copilot-headers.js.map +1 -0
  15. package/dist/providers/openai-completions.d.ts.map +1 -1
  16. package/dist/providers/openai-completions.js +7 -25
  17. package/dist/providers/openai-completions.js.map +1 -1
  18. package/dist/providers/openai-responses-shared.d.ts.map +1 -1
  19. package/dist/providers/openai-responses-shared.js +2 -2
  20. package/dist/providers/openai-responses-shared.js.map +1 -1
  21. package/dist/providers/openai-responses.d.ts.map +1 -1
  22. package/dist/providers/openai-responses.js +6 -20
  23. package/dist/providers/openai-responses.js.map +1 -1
  24. package/dist/providers/simple-options.d.ts.map +1 -1
  25. package/dist/providers/simple-options.js +1 -0
  26. package/dist/providers/simple-options.js.map +1 -1
  27. package/dist/types.d.ts +6 -0
  28. package/dist/types.d.ts.map +1 -1
  29. package/dist/types.js.map +1 -1
  30. package/package.json +1 -1
@@ -799,6 +799,23 @@ export declare const MODELS: {
799
799
  contextWindow: number;
800
800
  maxTokens: number;
801
801
  };
802
+ readonly "minimax.minimax-m2.1": {
803
+ id: string;
804
+ name: string;
805
+ api: "bedrock-converse-stream";
806
+ provider: string;
807
+ baseUrl: string;
808
+ reasoning: true;
809
+ input: "text"[];
810
+ cost: {
811
+ input: number;
812
+ output: number;
813
+ cacheRead: number;
814
+ cacheWrite: number;
815
+ };
816
+ contextWindow: number;
817
+ maxTokens: number;
818
+ };
802
819
  readonly "mistral.ministral-3-14b-instruct": {
803
820
  id: string;
804
821
  name: string;
@@ -901,6 +918,23 @@ export declare const MODELS: {
901
918
  contextWindow: number;
902
919
  maxTokens: number;
903
920
  };
921
+ readonly "moonshotai.kimi-k2.5": {
922
+ id: string;
923
+ name: string;
924
+ api: "bedrock-converse-stream";
925
+ provider: string;
926
+ baseUrl: string;
927
+ reasoning: true;
928
+ input: ("image" | "text")[];
929
+ cost: {
930
+ input: number;
931
+ output: number;
932
+ cacheRead: number;
933
+ cacheWrite: number;
934
+ };
935
+ contextWindow: number;
936
+ maxTokens: number;
937
+ };
904
938
  readonly "nvidia.nemotron-nano-12b-v2": {
905
939
  id: string;
906
940
  name: string;
@@ -1224,6 +1258,74 @@ export declare const MODELS: {
1224
1258
  contextWindow: number;
1225
1259
  maxTokens: number;
1226
1260
  };
1261
+ readonly "writer.palmyra-x4-v1:0": {
1262
+ id: string;
1263
+ name: string;
1264
+ api: "bedrock-converse-stream";
1265
+ provider: string;
1266
+ baseUrl: string;
1267
+ reasoning: true;
1268
+ input: "text"[];
1269
+ cost: {
1270
+ input: number;
1271
+ output: number;
1272
+ cacheRead: number;
1273
+ cacheWrite: number;
1274
+ };
1275
+ contextWindow: number;
1276
+ maxTokens: number;
1277
+ };
1278
+ readonly "writer.palmyra-x5-v1:0": {
1279
+ id: string;
1280
+ name: string;
1281
+ api: "bedrock-converse-stream";
1282
+ provider: string;
1283
+ baseUrl: string;
1284
+ reasoning: true;
1285
+ input: "text"[];
1286
+ cost: {
1287
+ input: number;
1288
+ output: number;
1289
+ cacheRead: number;
1290
+ cacheWrite: number;
1291
+ };
1292
+ contextWindow: number;
1293
+ maxTokens: number;
1294
+ };
1295
+ readonly "zai.glm-4.7": {
1296
+ id: string;
1297
+ name: string;
1298
+ api: "bedrock-converse-stream";
1299
+ provider: string;
1300
+ baseUrl: string;
1301
+ reasoning: true;
1302
+ input: "text"[];
1303
+ cost: {
1304
+ input: number;
1305
+ output: number;
1306
+ cacheRead: number;
1307
+ cacheWrite: number;
1308
+ };
1309
+ contextWindow: number;
1310
+ maxTokens: number;
1311
+ };
1312
+ readonly "zai.glm-4.7-flash": {
1313
+ id: string;
1314
+ name: string;
1315
+ api: "bedrock-converse-stream";
1316
+ provider: string;
1317
+ baseUrl: string;
1318
+ reasoning: true;
1319
+ input: "text"[];
1320
+ cost: {
1321
+ input: number;
1322
+ output: number;
1323
+ cacheRead: number;
1324
+ cacheWrite: number;
1325
+ };
1326
+ contextWindow: number;
1327
+ maxTokens: number;
1328
+ };
1227
1329
  };
1228
1330
  readonly anthropic: {
1229
1331
  readonly "claude-3-5-haiku-20241022": {
@@ -2061,6 +2163,23 @@ export declare const MODELS: {
2061
2163
  contextWindow: number;
2062
2164
  maxTokens: number;
2063
2165
  };
2166
+ readonly "gpt-5.3-codex-spark": {
2167
+ id: string;
2168
+ name: string;
2169
+ api: "azure-openai-responses";
2170
+ provider: string;
2171
+ baseUrl: string;
2172
+ reasoning: true;
2173
+ input: ("image" | "text")[];
2174
+ cost: {
2175
+ input: number;
2176
+ output: number;
2177
+ cacheRead: number;
2178
+ cacheWrite: number;
2179
+ };
2180
+ contextWindow: number;
2181
+ maxTokens: number;
2182
+ };
2064
2183
  readonly o1: {
2065
2184
  id: string;
2066
2185
  name: string;
@@ -2216,6 +2335,23 @@ export declare const MODELS: {
2216
2335
  contextWindow: number;
2217
2336
  maxTokens: number;
2218
2337
  };
2338
+ readonly "llama3.1-8b": {
2339
+ id: string;
2340
+ name: string;
2341
+ api: "openai-completions";
2342
+ provider: string;
2343
+ baseUrl: string;
2344
+ reasoning: false;
2345
+ input: "text"[];
2346
+ cost: {
2347
+ input: number;
2348
+ output: number;
2349
+ cacheRead: number;
2350
+ cacheWrite: number;
2351
+ };
2352
+ contextWindow: number;
2353
+ maxTokens: number;
2354
+ };
2219
2355
  readonly "qwen-3-235b-a22b-instruct-2507": {
2220
2356
  id: string;
2221
2357
  name: string;
@@ -2255,7 +2391,7 @@ export declare const MODELS: {
2255
2391
  readonly "claude-haiku-4.5": {
2256
2392
  id: string;
2257
2393
  name: string;
2258
- api: "openai-completions";
2394
+ api: "anthropic-messages";
2259
2395
  provider: string;
2260
2396
  baseUrl: string;
2261
2397
  headers: {
@@ -2264,11 +2400,6 @@ export declare const MODELS: {
2264
2400
  "Editor-Plugin-Version": string;
2265
2401
  "Copilot-Integration-Id": string;
2266
2402
  };
2267
- compat: {
2268
- supportsStore: false;
2269
- supportsDeveloperRole: false;
2270
- supportsReasoningEffort: false;
2271
- };
2272
2403
  reasoning: true;
2273
2404
  input: ("image" | "text")[];
2274
2405
  cost: {
@@ -2283,7 +2414,7 @@ export declare const MODELS: {
2283
2414
  readonly "claude-opus-4.5": {
2284
2415
  id: string;
2285
2416
  name: string;
2286
- api: "openai-completions";
2417
+ api: "anthropic-messages";
2287
2418
  provider: string;
2288
2419
  baseUrl: string;
2289
2420
  headers: {
@@ -2292,11 +2423,6 @@ export declare const MODELS: {
2292
2423
  "Editor-Plugin-Version": string;
2293
2424
  "Copilot-Integration-Id": string;
2294
2425
  };
2295
- compat: {
2296
- supportsStore: false;
2297
- supportsDeveloperRole: false;
2298
- supportsReasoningEffort: false;
2299
- };
2300
2426
  reasoning: true;
2301
2427
  input: ("image" | "text")[];
2302
2428
  cost: {
@@ -2311,7 +2437,7 @@ export declare const MODELS: {
2311
2437
  readonly "claude-opus-4.6": {
2312
2438
  id: string;
2313
2439
  name: string;
2314
- api: "openai-completions";
2440
+ api: "anthropic-messages";
2315
2441
  provider: string;
2316
2442
  baseUrl: string;
2317
2443
  headers: {
@@ -2320,11 +2446,6 @@ export declare const MODELS: {
2320
2446
  "Editor-Plugin-Version": string;
2321
2447
  "Copilot-Integration-Id": string;
2322
2448
  };
2323
- compat: {
2324
- supportsStore: false;
2325
- supportsDeveloperRole: false;
2326
- supportsReasoningEffort: false;
2327
- };
2328
2449
  reasoning: true;
2329
2450
  input: ("image" | "text")[];
2330
2451
  cost: {
@@ -2339,7 +2460,7 @@ export declare const MODELS: {
2339
2460
  readonly "claude-sonnet-4": {
2340
2461
  id: string;
2341
2462
  name: string;
2342
- api: "openai-completions";
2463
+ api: "anthropic-messages";
2343
2464
  provider: string;
2344
2465
  baseUrl: string;
2345
2466
  headers: {
@@ -2348,11 +2469,6 @@ export declare const MODELS: {
2348
2469
  "Editor-Plugin-Version": string;
2349
2470
  "Copilot-Integration-Id": string;
2350
2471
  };
2351
- compat: {
2352
- supportsStore: false;
2353
- supportsDeveloperRole: false;
2354
- supportsReasoningEffort: false;
2355
- };
2356
2472
  reasoning: true;
2357
2473
  input: ("image" | "text")[];
2358
2474
  cost: {
@@ -2367,7 +2483,7 @@ export declare const MODELS: {
2367
2483
  readonly "claude-sonnet-4.5": {
2368
2484
  id: string;
2369
2485
  name: string;
2370
- api: "openai-completions";
2486
+ api: "anthropic-messages";
2371
2487
  provider: string;
2372
2488
  baseUrl: string;
2373
2489
  headers: {
@@ -2376,11 +2492,6 @@ export declare const MODELS: {
2376
2492
  "Editor-Plugin-Version": string;
2377
2493
  "Copilot-Integration-Id": string;
2378
2494
  };
2379
- compat: {
2380
- supportsStore: false;
2381
- supportsDeveloperRole: false;
2382
- supportsReasoningEffort: false;
2383
- };
2384
2495
  reasoning: true;
2385
2496
  input: ("image" | "text")[];
2386
2497
  cost: {
@@ -4039,6 +4150,26 @@ export declare const MODELS: {
4039
4150
  contextWindow: number;
4040
4151
  maxTokens: number;
4041
4152
  };
4153
+ readonly "zai-org/GLM-5": {
4154
+ id: string;
4155
+ name: string;
4156
+ api: "openai-completions";
4157
+ provider: string;
4158
+ baseUrl: string;
4159
+ compat: {
4160
+ supportsDeveloperRole: false;
4161
+ };
4162
+ reasoning: true;
4163
+ input: "text"[];
4164
+ cost: {
4165
+ input: number;
4166
+ output: number;
4167
+ cacheRead: number;
4168
+ cacheWrite: number;
4169
+ };
4170
+ contextWindow: number;
4171
+ maxTokens: number;
4172
+ };
4042
4173
  };
4043
4174
  readonly "kimi-coding": {
4044
4175
  readonly k2p5: {
@@ -4111,6 +4242,23 @@ export declare const MODELS: {
4111
4242
  contextWindow: number;
4112
4243
  maxTokens: number;
4113
4244
  };
4245
+ readonly "MiniMax-M2.5": {
4246
+ id: string;
4247
+ name: string;
4248
+ api: "anthropic-messages";
4249
+ provider: string;
4250
+ baseUrl: string;
4251
+ reasoning: true;
4252
+ input: "text"[];
4253
+ cost: {
4254
+ input: number;
4255
+ output: number;
4256
+ cacheRead: number;
4257
+ cacheWrite: number;
4258
+ };
4259
+ contextWindow: number;
4260
+ maxTokens: number;
4261
+ };
4114
4262
  };
4115
4263
  readonly "minimax-cn": {
4116
4264
  readonly "MiniMax-M2": {
@@ -4147,6 +4295,23 @@ export declare const MODELS: {
4147
4295
  contextWindow: number;
4148
4296
  maxTokens: number;
4149
4297
  };
4298
+ readonly "MiniMax-M2.5": {
4299
+ id: string;
4300
+ name: string;
4301
+ api: "anthropic-messages";
4302
+ provider: string;
4303
+ baseUrl: string;
4304
+ reasoning: true;
4305
+ input: "text"[];
4306
+ cost: {
4307
+ input: number;
4308
+ output: number;
4309
+ cacheRead: number;
4310
+ cacheWrite: number;
4311
+ };
4312
+ contextWindow: number;
4313
+ maxTokens: number;
4314
+ };
4150
4315
  };
4151
4316
  readonly mistral: {
4152
4317
  readonly "codestral-latest": {
@@ -5035,6 +5200,23 @@ export declare const MODELS: {
5035
5200
  contextWindow: number;
5036
5201
  maxTokens: number;
5037
5202
  };
5203
+ readonly "gpt-5.3-codex-spark": {
5204
+ id: string;
5205
+ name: string;
5206
+ api: "openai-responses";
5207
+ provider: string;
5208
+ baseUrl: string;
5209
+ reasoning: true;
5210
+ input: ("image" | "text")[];
5211
+ cost: {
5212
+ input: number;
5213
+ output: number;
5214
+ cacheRead: number;
5215
+ cacheWrite: number;
5216
+ };
5217
+ contextWindow: number;
5218
+ maxTokens: number;
5219
+ };
5038
5220
  readonly o1: {
5039
5221
  id: string;
5040
5222
  name: string;
@@ -5275,6 +5457,23 @@ export declare const MODELS: {
5275
5457
  contextWindow: number;
5276
5458
  maxTokens: number;
5277
5459
  };
5460
+ readonly "gpt-5.3-codex-spark": {
5461
+ id: string;
5462
+ name: string;
5463
+ api: "openai-codex-responses";
5464
+ provider: string;
5465
+ baseUrl: string;
5466
+ reasoning: true;
5467
+ input: "text"[];
5468
+ cost: {
5469
+ input: number;
5470
+ output: number;
5471
+ cacheRead: number;
5472
+ cacheWrite: number;
5473
+ };
5474
+ contextWindow: number;
5475
+ maxTokens: number;
5476
+ };
5278
5477
  };
5279
5478
  readonly opencode: {
5280
5479
  readonly "big-pickle": {
@@ -5481,23 +5680,6 @@ export declare const MODELS: {
5481
5680
  contextWindow: number;
5482
5681
  maxTokens: number;
5483
5682
  };
5484
- readonly "glm-4.7-free": {
5485
- id: string;
5486
- name: string;
5487
- api: "openai-completions";
5488
- provider: string;
5489
- baseUrl: string;
5490
- reasoning: true;
5491
- input: "text"[];
5492
- cost: {
5493
- input: number;
5494
- output: number;
5495
- cacheRead: number;
5496
- cacheWrite: number;
5497
- };
5498
- contextWindow: number;
5499
- maxTokens: number;
5500
- };
5501
5683
  readonly "gpt-5": {
5502
5684
  id: string;
5503
5685
  name: string;
@@ -5736,47 +5918,13 @@ export declare const MODELS: {
5736
5918
  contextWindow: number;
5737
5919
  maxTokens: number;
5738
5920
  };
5739
- readonly "minimax-m2.1-free": {
5740
- id: string;
5741
- name: string;
5742
- api: "anthropic-messages";
5743
- provider: string;
5744
- baseUrl: string;
5745
- reasoning: true;
5746
- input: "text"[];
5747
- cost: {
5748
- input: number;
5749
- output: number;
5750
- cacheRead: number;
5751
- cacheWrite: number;
5752
- };
5753
- contextWindow: number;
5754
- maxTokens: number;
5755
- };
5756
- readonly "qwen3-coder": {
5757
- id: string;
5758
- name: string;
5759
- api: "openai-completions";
5760
- provider: string;
5761
- baseUrl: string;
5762
- reasoning: false;
5763
- input: "text"[];
5764
- cost: {
5765
- input: number;
5766
- output: number;
5767
- cacheRead: number;
5768
- cacheWrite: number;
5769
- };
5770
- contextWindow: number;
5771
- maxTokens: number;
5772
- };
5773
- readonly "trinity-large-preview-free": {
5921
+ readonly "minimax-m2.5-free": {
5774
5922
  id: string;
5775
5923
  name: string;
5776
5924
  api: "openai-completions";
5777
5925
  provider: string;
5778
5926
  baseUrl: string;
5779
- reasoning: false;
5927
+ reasoning: true;
5780
5928
  input: "text"[];
5781
5929
  cost: {
5782
5930
  input: number;
@@ -5806,23 +5954,6 @@ export declare const MODELS: {
5806
5954
  contextWindow: number;
5807
5955
  maxTokens: number;
5808
5956
  };
5809
- readonly "ai21/jamba-mini-1.7": {
5810
- id: string;
5811
- name: string;
5812
- api: "openai-completions";
5813
- provider: string;
5814
- baseUrl: string;
5815
- reasoning: false;
5816
- input: "text"[];
5817
- cost: {
5818
- input: number;
5819
- output: number;
5820
- cacheRead: number;
5821
- cacheWrite: number;
5822
- };
5823
- contextWindow: number;
5824
- maxTokens: number;
5825
- };
5826
5957
  readonly "alibaba/tongyi-deepresearch-30b-a3b": {
5827
5958
  id: string;
5828
5959
  name: string;
@@ -6945,13 +7076,13 @@ export declare const MODELS: {
6945
7076
  contextWindow: number;
6946
7077
  maxTokens: number;
6947
7078
  };
6948
- readonly "mistralai/codestral-2508": {
7079
+ readonly "minimax/minimax-m2.5": {
6949
7080
  id: string;
6950
7081
  name: string;
6951
7082
  api: "openai-completions";
6952
7083
  provider: string;
6953
7084
  baseUrl: string;
6954
- reasoning: false;
7085
+ reasoning: true;
6955
7086
  input: "text"[];
6956
7087
  cost: {
6957
7088
  input: number;
@@ -6962,7 +7093,7 @@ export declare const MODELS: {
6962
7093
  contextWindow: number;
6963
7094
  maxTokens: number;
6964
7095
  };
6965
- readonly "mistralai/devstral-2512": {
7096
+ readonly "mistralai/codestral-2508": {
6966
7097
  id: string;
6967
7098
  name: string;
6968
7099
  api: "openai-completions";
@@ -6979,7 +7110,7 @@ export declare const MODELS: {
6979
7110
  contextWindow: number;
6980
7111
  maxTokens: number;
6981
7112
  };
6982
- readonly "mistralai/devstral-medium": {
7113
+ readonly "mistralai/devstral-2512": {
6983
7114
  id: string;
6984
7115
  name: string;
6985
7116
  api: "openai-completions";
@@ -6996,7 +7127,7 @@ export declare const MODELS: {
6996
7127
  contextWindow: number;
6997
7128
  maxTokens: number;
6998
7129
  };
6999
- readonly "mistralai/devstral-small": {
7130
+ readonly "mistralai/devstral-medium": {
7000
7131
  id: string;
7001
7132
  name: string;
7002
7133
  api: "openai-completions";
@@ -7013,24 +7144,7 @@ export declare const MODELS: {
7013
7144
  contextWindow: number;
7014
7145
  maxTokens: number;
7015
7146
  };
7016
- readonly "mistralai/ministral-14b-2512": {
7017
- id: string;
7018
- name: string;
7019
- api: "openai-completions";
7020
- provider: string;
7021
- baseUrl: string;
7022
- reasoning: false;
7023
- input: ("image" | "text")[];
7024
- cost: {
7025
- input: number;
7026
- output: number;
7027
- cacheRead: number;
7028
- cacheWrite: number;
7029
- };
7030
- contextWindow: number;
7031
- maxTokens: number;
7032
- };
7033
- readonly "mistralai/ministral-3b": {
7147
+ readonly "mistralai/devstral-small": {
7034
7148
  id: string;
7035
7149
  name: string;
7036
7150
  api: "openai-completions";
@@ -7047,7 +7161,7 @@ export declare const MODELS: {
7047
7161
  contextWindow: number;
7048
7162
  maxTokens: number;
7049
7163
  };
7050
- readonly "mistralai/ministral-3b-2512": {
7164
+ readonly "mistralai/ministral-14b-2512": {
7051
7165
  id: string;
7052
7166
  name: string;
7053
7167
  api: "openai-completions";
@@ -7064,14 +7178,14 @@ export declare const MODELS: {
7064
7178
  contextWindow: number;
7065
7179
  maxTokens: number;
7066
7180
  };
7067
- readonly "mistralai/ministral-8b": {
7181
+ readonly "mistralai/ministral-3b-2512": {
7068
7182
  id: string;
7069
7183
  name: string;
7070
7184
  api: "openai-completions";
7071
7185
  provider: string;
7072
7186
  baseUrl: string;
7073
7187
  reasoning: false;
7074
- input: "text"[];
7188
+ input: ("image" | "text")[];
7075
7189
  cost: {
7076
7190
  input: number;
7077
7191
  output: number;
@@ -7319,23 +7433,6 @@ export declare const MODELS: {
7319
7433
  contextWindow: number;
7320
7434
  maxTokens: number;
7321
7435
  };
7322
- readonly "mistralai/mistral-tiny": {
7323
- id: string;
7324
- name: string;
7325
- api: "openai-completions";
7326
- provider: string;
7327
- baseUrl: string;
7328
- reasoning: false;
7329
- input: "text"[];
7330
- cost: {
7331
- input: number;
7332
- output: number;
7333
- cacheRead: number;
7334
- cacheWrite: number;
7335
- };
7336
- contextWindow: number;
7337
- maxTokens: number;
7338
- };
7339
7436
  readonly "mistralai/mixtral-8x22b-instruct": {
7340
7437
  id: string;
7341
7438
  name: string;
@@ -7370,23 +7467,6 @@ export declare const MODELS: {
7370
7467
  contextWindow: number;
7371
7468
  maxTokens: number;
7372
7469
  };
7373
- readonly "mistralai/pixtral-12b": {
7374
- id: string;
7375
- name: string;
7376
- api: "openai-completions";
7377
- provider: string;
7378
- baseUrl: string;
7379
- reasoning: false;
7380
- input: ("image" | "text")[];
7381
- cost: {
7382
- input: number;
7383
- output: number;
7384
- cacheRead: number;
7385
- cacheWrite: number;
7386
- };
7387
- contextWindow: number;
7388
- maxTokens: number;
7389
- };
7390
7470
  readonly "mistralai/pixtral-large-2411": {
7391
7471
  id: string;
7392
7472
  name: string;
@@ -8526,14 +8606,14 @@ export declare const MODELS: {
8526
8606
  contextWindow: number;
8527
8607
  maxTokens: number;
8528
8608
  };
8529
- readonly "openrouter/auto": {
8609
+ readonly "openrouter/aurora-alpha": {
8530
8610
  id: string;
8531
8611
  name: string;
8532
8612
  api: "openai-completions";
8533
8613
  provider: string;
8534
8614
  baseUrl: string;
8535
8615
  reasoning: true;
8536
- input: ("image" | "text")[];
8616
+ input: "text"[];
8537
8617
  cost: {
8538
8618
  input: number;
8539
8619
  output: number;
@@ -8543,7 +8623,7 @@ export declare const MODELS: {
8543
8623
  contextWindow: number;
8544
8624
  maxTokens: number;
8545
8625
  };
8546
- readonly "openrouter/free": {
8626
+ readonly "openrouter/auto": {
8547
8627
  id: string;
8548
8628
  name: string;
8549
8629
  api: "openai-completions";
@@ -8560,14 +8640,14 @@ export declare const MODELS: {
8560
8640
  contextWindow: number;
8561
8641
  maxTokens: number;
8562
8642
  };
8563
- readonly "openrouter/pony-alpha": {
8643
+ readonly "openrouter/free": {
8564
8644
  id: string;
8565
8645
  name: string;
8566
8646
  api: "openai-completions";
8567
8647
  provider: string;
8568
8648
  baseUrl: string;
8569
8649
  reasoning: true;
8570
- input: "text"[];
8650
+ input: ("image" | "text")[];
8571
8651
  cost: {
8572
8652
  input: number;
8573
8653
  output: number;
@@ -8866,6 +8946,23 @@ export declare const MODELS: {
8866
8946
  contextWindow: number;
8867
8947
  maxTokens: number;
8868
8948
  };
8949
+ readonly "qwen/qwen3-4b": {
8950
+ id: string;
8951
+ name: string;
8952
+ api: "openai-completions";
8953
+ provider: string;
8954
+ baseUrl: string;
8955
+ reasoning: true;
8956
+ input: "text"[];
8957
+ cost: {
8958
+ input: number;
8959
+ output: number;
8960
+ cacheRead: number;
8961
+ cacheWrite: number;
8962
+ };
8963
+ contextWindow: number;
8964
+ maxTokens: number;
8965
+ };
8869
8966
  readonly "qwen/qwen3-4b:free": {
8870
8967
  id: string;
8871
8968
  name: string;
@@ -8906,7 +9003,7 @@ export declare const MODELS: {
8906
9003
  api: "openai-completions";
8907
9004
  provider: string;
8908
9005
  baseUrl: string;
8909
- reasoning: true;
9006
+ reasoning: false;
8910
9007
  input: "text"[];
8911
9008
  cost: {
8912
9009
  input: number;
@@ -9036,6 +9133,23 @@ export declare const MODELS: {
9036
9133
  contextWindow: number;
9037
9134
  maxTokens: number;
9038
9135
  };
9136
+ readonly "qwen/qwen3-max-thinking": {
9137
+ id: string;
9138
+ name: string;
9139
+ api: "openai-completions";
9140
+ provider: string;
9141
+ baseUrl: string;
9142
+ reasoning: true;
9143
+ input: "text"[];
9144
+ cost: {
9145
+ input: number;
9146
+ output: number;
9147
+ cacheRead: number;
9148
+ cacheWrite: number;
9149
+ };
9150
+ contextWindow: number;
9151
+ maxTokens: number;
9152
+ };
9039
9153
  readonly "qwen/qwen3-next-80b-a3b-instruct": {
9040
9154
  id: string;
9041
9155
  name: string;
@@ -9155,6 +9269,23 @@ export declare const MODELS: {
9155
9269
  contextWindow: number;
9156
9270
  maxTokens: number;
9157
9271
  };
9272
+ readonly "qwen/qwen3-vl-32b-instruct": {
9273
+ id: string;
9274
+ name: string;
9275
+ api: "openai-completions";
9276
+ provider: string;
9277
+ baseUrl: string;
9278
+ reasoning: false;
9279
+ input: ("image" | "text")[];
9280
+ cost: {
9281
+ input: number;
9282
+ output: number;
9283
+ cacheRead: number;
9284
+ cacheWrite: number;
9285
+ };
9286
+ contextWindow: number;
9287
+ maxTokens: number;
9288
+ };
9158
9289
  readonly "qwen/qwen3-vl-8b-instruct": {
9159
9290
  id: string;
9160
9291
  name: string;
@@ -9257,14 +9388,14 @@ export declare const MODELS: {
9257
9388
  contextWindow: number;
9258
9389
  maxTokens: number;
9259
9390
  };
9260
- readonly "stepfun-ai/step3": {
9391
+ readonly "stepfun/step-3.5-flash": {
9261
9392
  id: string;
9262
9393
  name: string;
9263
9394
  api: "openai-completions";
9264
9395
  provider: string;
9265
9396
  baseUrl: string;
9266
9397
  reasoning: true;
9267
- input: ("image" | "text")[];
9398
+ input: "text"[];
9268
9399
  cost: {
9269
9400
  input: number;
9270
9401
  output: number;
@@ -9716,6 +9847,23 @@ export declare const MODELS: {
9716
9847
  contextWindow: number;
9717
9848
  maxTokens: number;
9718
9849
  };
9850
+ readonly "z-ai/glm-5": {
9851
+ id: string;
9852
+ name: string;
9853
+ api: "openai-completions";
9854
+ provider: string;
9855
+ baseUrl: string;
9856
+ reasoning: true;
9857
+ input: "text"[];
9858
+ cost: {
9859
+ input: number;
9860
+ output: number;
9861
+ cacheRead: number;
9862
+ cacheWrite: number;
9863
+ };
9864
+ contextWindow: number;
9865
+ maxTokens: number;
9866
+ };
9719
9867
  };
9720
9868
  readonly "vercel-ai-gateway": {
9721
9869
  readonly "alibaba/qwen-3-14b": {
@@ -10228,13 +10376,13 @@ export declare const MODELS: {
10228
10376
  contextWindow: number;
10229
10377
  maxTokens: number;
10230
10378
  };
10231
- readonly "deepseek/deepseek-v3.2-exp": {
10379
+ readonly "deepseek/deepseek-v3.2": {
10232
10380
  id: string;
10233
10381
  name: string;
10234
10382
  api: "anthropic-messages";
10235
10383
  provider: string;
10236
10384
  baseUrl: string;
10237
- reasoning: true;
10385
+ reasoning: false;
10238
10386
  input: "text"[];
10239
10387
  cost: {
10240
10388
  input: number;
@@ -10602,6 +10750,23 @@ export declare const MODELS: {
10602
10750
  contextWindow: number;
10603
10751
  maxTokens: number;
10604
10752
  };
10753
+ readonly "minimax/minimax-m2.5": {
10754
+ id: string;
10755
+ name: string;
10756
+ api: "anthropic-messages";
10757
+ provider: string;
10758
+ baseUrl: string;
10759
+ reasoning: true;
10760
+ input: "text"[];
10761
+ cost: {
10762
+ input: number;
10763
+ output: number;
10764
+ cacheRead: number;
10765
+ cacheWrite: number;
10766
+ };
10767
+ contextWindow: number;
10768
+ maxTokens: number;
10769
+ };
10605
10770
  readonly "mistral/codestral": {
10606
10771
  id: string;
10607
10772
  name: string;
@@ -11843,6 +12008,23 @@ export declare const MODELS: {
11843
12008
  contextWindow: number;
11844
12009
  maxTokens: number;
11845
12010
  };
12011
+ readonly "zai/glm-5": {
12012
+ id: string;
12013
+ name: string;
12014
+ api: "anthropic-messages";
12015
+ provider: string;
12016
+ baseUrl: string;
12017
+ reasoning: true;
12018
+ input: "text"[];
12019
+ cost: {
12020
+ input: number;
12021
+ output: number;
12022
+ cacheRead: number;
12023
+ cacheWrite: number;
12024
+ };
12025
+ contextWindow: number;
12026
+ maxTokens: number;
12027
+ };
11846
12028
  };
11847
12029
  readonly xai: {
11848
12030
  readonly "grok-2": {
@@ -12389,6 +12571,27 @@ export declare const MODELS: {
12389
12571
  contextWindow: number;
12390
12572
  maxTokens: number;
12391
12573
  };
12574
+ readonly "glm-5": {
12575
+ id: string;
12576
+ name: string;
12577
+ api: "openai-completions";
12578
+ provider: string;
12579
+ baseUrl: string;
12580
+ compat: {
12581
+ supportsDeveloperRole: false;
12582
+ thinkingFormat: "zai";
12583
+ };
12584
+ reasoning: true;
12585
+ input: "text"[];
12586
+ cost: {
12587
+ input: number;
12588
+ output: number;
12589
+ cacheRead: number;
12590
+ cacheWrite: number;
12591
+ };
12592
+ contextWindow: number;
12593
+ maxTokens: number;
12594
+ };
12392
12595
  };
12393
12596
  };
12394
12597
  //# sourceMappingURL=models.generated.d.ts.map