@avadisabelle/ava-pi-ai 0.65.0 → 0.65.3

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -323,6 +323,40 @@ export declare const MODELS: {
323
323
  contextWindow: number;
324
324
  maxTokens: number;
325
325
  };
326
+ readonly "au.anthropic.claude-opus-4-6-v1": {
327
+ id: string;
328
+ name: string;
329
+ api: "bedrock-converse-stream";
330
+ provider: string;
331
+ baseUrl: string;
332
+ reasoning: true;
333
+ input: ("image" | "text")[];
334
+ cost: {
335
+ input: number;
336
+ output: number;
337
+ cacheRead: number;
338
+ cacheWrite: number;
339
+ };
340
+ contextWindow: number;
341
+ maxTokens: number;
342
+ };
343
+ readonly "au.anthropic.claude-sonnet-4-6": {
344
+ id: string;
345
+ name: string;
346
+ api: "bedrock-converse-stream";
347
+ provider: string;
348
+ baseUrl: string;
349
+ reasoning: true;
350
+ input: ("image" | "text")[];
351
+ cost: {
352
+ input: number;
353
+ output: number;
354
+ cacheRead: number;
355
+ cacheWrite: number;
356
+ };
357
+ contextWindow: number;
358
+ maxTokens: number;
359
+ };
326
360
  readonly "deepseek.r1-v1:0": {
327
361
  id: string;
328
362
  name: string;
@@ -2486,6 +2520,40 @@ export declare const MODELS: {
2486
2520
  contextWindow: number;
2487
2521
  maxTokens: number;
2488
2522
  };
2523
+ readonly "gpt-5.5": {
2524
+ id: string;
2525
+ name: string;
2526
+ api: "azure-openai-responses";
2527
+ provider: string;
2528
+ baseUrl: string;
2529
+ reasoning: true;
2530
+ input: ("image" | "text")[];
2531
+ cost: {
2532
+ input: number;
2533
+ output: number;
2534
+ cacheRead: number;
2535
+ cacheWrite: number;
2536
+ };
2537
+ contextWindow: number;
2538
+ maxTokens: number;
2539
+ };
2540
+ readonly "gpt-5.5-pro": {
2541
+ id: string;
2542
+ name: string;
2543
+ api: "azure-openai-responses";
2544
+ provider: string;
2545
+ baseUrl: string;
2546
+ reasoning: true;
2547
+ input: ("image" | "text")[];
2548
+ cost: {
2549
+ input: number;
2550
+ output: number;
2551
+ cacheRead: number;
2552
+ cacheWrite: number;
2553
+ };
2554
+ contextWindow: number;
2555
+ maxTokens: number;
2556
+ };
2489
2557
  readonly o1: {
2490
2558
  id: string;
2491
2559
  name: string;
@@ -3276,6 +3344,29 @@ export declare const MODELS: {
3276
3344
  contextWindow: number;
3277
3345
  maxTokens: number;
3278
3346
  };
3347
+ readonly "gpt-5.5": {
3348
+ id: string;
3349
+ name: string;
3350
+ api: "openai-responses";
3351
+ provider: string;
3352
+ baseUrl: string;
3353
+ headers: {
3354
+ "User-Agent": string;
3355
+ "Editor-Version": string;
3356
+ "Editor-Plugin-Version": string;
3357
+ "Copilot-Integration-Id": string;
3358
+ };
3359
+ reasoning: true;
3360
+ input: ("image" | "text")[];
3361
+ cost: {
3362
+ input: number;
3363
+ output: number;
3364
+ cacheRead: number;
3365
+ cacheWrite: number;
3366
+ };
3367
+ contextWindow: number;
3368
+ maxTokens: number;
3369
+ };
3279
3370
  readonly "grok-code-fast-1": {
3280
3371
  id: string;
3281
3372
  name: string;
@@ -3731,7 +3822,7 @@ export declare const MODELS: {
3731
3822
  contextWindow: number;
3732
3823
  maxTokens: number;
3733
3824
  };
3734
- readonly "gemma-4-26b-it": {
3825
+ readonly "gemma-4-26b-a4b-it": {
3735
3826
  id: string;
3736
3827
  name: string;
3737
3828
  api: "google-generative-ai";
@@ -4797,6 +4888,26 @@ export declare const MODELS: {
4797
4888
  contextWindow: number;
4798
4889
  maxTokens: number;
4799
4890
  };
4891
+ readonly "deepseek-ai/DeepSeek-V4-Pro": {
4892
+ id: string;
4893
+ name: string;
4894
+ api: "openai-completions";
4895
+ provider: string;
4896
+ baseUrl: string;
4897
+ compat: {
4898
+ supportsDeveloperRole: false;
4899
+ };
4900
+ reasoning: true;
4901
+ input: "text"[];
4902
+ cost: {
4903
+ input: number;
4904
+ output: number;
4905
+ cacheRead: number;
4906
+ cacheWrite: number;
4907
+ };
4908
+ contextWindow: number;
4909
+ maxTokens: number;
4910
+ };
4800
4911
  readonly "moonshotai/Kimi-K2-Instruct": {
4801
4912
  id: string;
4802
4913
  name: string;
@@ -4877,6 +4988,26 @@ export declare const MODELS: {
4877
4988
  contextWindow: number;
4878
4989
  maxTokens: number;
4879
4990
  };
4991
+ readonly "moonshotai/Kimi-K2.6": {
4992
+ id: string;
4993
+ name: string;
4994
+ api: "openai-completions";
4995
+ provider: string;
4996
+ baseUrl: string;
4997
+ compat: {
4998
+ supportsDeveloperRole: false;
4999
+ };
5000
+ reasoning: true;
5001
+ input: ("image" | "text")[];
5002
+ cost: {
5003
+ input: number;
5004
+ output: number;
5005
+ cacheRead: number;
5006
+ cacheWrite: number;
5007
+ };
5008
+ contextWindow: number;
5009
+ maxTokens: number;
5010
+ };
4880
5011
  readonly "zai-org/GLM-4.7": {
4881
5012
  id: string;
4882
5013
  name: string;
@@ -4959,6 +5090,23 @@ export declare const MODELS: {
4959
5090
  };
4960
5091
  };
4961
5092
  readonly "kimi-coding": {
5093
+ readonly k2p6: {
5094
+ id: string;
5095
+ name: string;
5096
+ api: "anthropic-messages";
5097
+ provider: string;
5098
+ baseUrl: string;
5099
+ reasoning: true;
5100
+ input: ("image" | "text")[];
5101
+ cost: {
5102
+ input: number;
5103
+ output: number;
5104
+ cacheRead: number;
5105
+ cacheWrite: number;
5106
+ };
5107
+ contextWindow: number;
5108
+ maxTokens: number;
5109
+ };
4962
5110
  readonly "kimi-for-coding": {
4963
5111
  id: string;
4964
5112
  name: string;
@@ -5339,13 +5487,30 @@ export declare const MODELS: {
5339
5487
  contextWindow: number;
5340
5488
  maxTokens: number;
5341
5489
  };
5490
+ readonly "mistral-medium-2604": {
5491
+ id: string;
5492
+ name: string;
5493
+ api: "mistral-conversations";
5494
+ provider: string;
5495
+ baseUrl: string;
5496
+ reasoning: true;
5497
+ input: ("image" | "text")[];
5498
+ cost: {
5499
+ input: number;
5500
+ output: number;
5501
+ cacheRead: number;
5502
+ cacheWrite: number;
5503
+ };
5504
+ contextWindow: number;
5505
+ maxTokens: number;
5506
+ };
5342
5507
  readonly "mistral-medium-latest": {
5343
5508
  id: string;
5344
5509
  name: string;
5345
5510
  api: "mistral-conversations";
5346
5511
  provider: string;
5347
5512
  baseUrl: string;
5348
- reasoning: false;
5513
+ reasoning: true;
5349
5514
  input: ("image" | "text")[];
5350
5515
  cost: {
5351
5516
  input: number;
@@ -6055,6 +6220,40 @@ export declare const MODELS: {
6055
6220
  contextWindow: number;
6056
6221
  maxTokens: number;
6057
6222
  };
6223
+ readonly "gpt-5.5": {
6224
+ id: string;
6225
+ name: string;
6226
+ api: "openai-responses";
6227
+ provider: string;
6228
+ baseUrl: string;
6229
+ reasoning: true;
6230
+ input: ("image" | "text")[];
6231
+ cost: {
6232
+ input: number;
6233
+ output: number;
6234
+ cacheRead: number;
6235
+ cacheWrite: number;
6236
+ };
6237
+ contextWindow: number;
6238
+ maxTokens: number;
6239
+ };
6240
+ readonly "gpt-5.5-pro": {
6241
+ id: string;
6242
+ name: string;
6243
+ api: "openai-responses";
6244
+ provider: string;
6245
+ baseUrl: string;
6246
+ reasoning: true;
6247
+ input: ("image" | "text")[];
6248
+ cost: {
6249
+ input: number;
6250
+ output: number;
6251
+ cacheRead: number;
6252
+ cacheWrite: number;
6253
+ };
6254
+ contextWindow: number;
6255
+ maxTokens: number;
6256
+ };
6058
6257
  readonly o1: {
6059
6258
  id: string;
6060
6259
  name: string;
@@ -6365,23 +6564,6 @@ export declare const MODELS: {
6365
6564
  contextWindow: number;
6366
6565
  maxTokens: number;
6367
6566
  };
6368
- readonly "claude-3-5-haiku": {
6369
- id: string;
6370
- name: string;
6371
- api: "anthropic-messages";
6372
- provider: string;
6373
- baseUrl: string;
6374
- reasoning: false;
6375
- input: ("image" | "text")[];
6376
- cost: {
6377
- input: number;
6378
- output: number;
6379
- cacheRead: number;
6380
- cacheWrite: number;
6381
- };
6382
- contextWindow: number;
6383
- maxTokens: number;
6384
- };
6385
6567
  readonly "claude-haiku-4-5": {
6386
6568
  id: string;
6387
6569
  name: string;
@@ -6824,10 +7006,10 @@ export declare const MODELS: {
6824
7006
  contextWindow: number;
6825
7007
  maxTokens: number;
6826
7008
  };
6827
- readonly "kimi-k2.5": {
7009
+ readonly "gpt-5.5": {
6828
7010
  id: string;
6829
7011
  name: string;
6830
- api: "openai-completions";
7012
+ api: "openai-responses";
6831
7013
  provider: string;
6832
7014
  baseUrl: string;
6833
7015
  reasoning: true;
@@ -6841,14 +7023,14 @@ export declare const MODELS: {
6841
7023
  contextWindow: number;
6842
7024
  maxTokens: number;
6843
7025
  };
6844
- readonly "minimax-m2.5": {
7026
+ readonly "gpt-5.5-pro": {
6845
7027
  id: string;
6846
7028
  name: string;
6847
- api: "openai-completions";
7029
+ api: "openai-responses";
6848
7030
  provider: string;
6849
7031
  baseUrl: string;
6850
7032
  reasoning: true;
6851
- input: "text"[];
7033
+ input: ("image" | "text")[];
6852
7034
  cost: {
6853
7035
  input: number;
6854
7036
  output: number;
@@ -6858,10 +7040,10 @@ export declare const MODELS: {
6858
7040
  contextWindow: number;
6859
7041
  maxTokens: number;
6860
7042
  };
6861
- readonly "minimax-m2.5-free": {
7043
+ readonly "hy3-preview-free": {
6862
7044
  id: string;
6863
7045
  name: string;
6864
- api: "anthropic-messages";
7046
+ api: "openai-completions";
6865
7047
  provider: string;
6866
7048
  baseUrl: string;
6867
7049
  reasoning: true;
@@ -6875,14 +7057,14 @@ export declare const MODELS: {
6875
7057
  contextWindow: number;
6876
7058
  maxTokens: number;
6877
7059
  };
6878
- readonly "nemotron-3-super-free": {
7060
+ readonly "kimi-k2.5": {
6879
7061
  id: string;
6880
7062
  name: string;
6881
7063
  api: "openai-completions";
6882
7064
  provider: string;
6883
7065
  baseUrl: string;
6884
7066
  reasoning: true;
6885
- input: "text"[];
7067
+ input: ("image" | "text")[];
6886
7068
  cost: {
6887
7069
  input: number;
6888
7070
  output: number;
@@ -6892,7 +7074,7 @@ export declare const MODELS: {
6892
7074
  contextWindow: number;
6893
7075
  maxTokens: number;
6894
7076
  };
6895
- readonly "qwen3.5-plus": {
7077
+ readonly "kimi-k2.6": {
6896
7078
  id: string;
6897
7079
  name: string;
6898
7080
  api: "openai-completions";
@@ -6909,14 +7091,14 @@ export declare const MODELS: {
6909
7091
  contextWindow: number;
6910
7092
  maxTokens: number;
6911
7093
  };
6912
- readonly "qwen3.6-plus": {
7094
+ readonly "minimax-m2.5": {
6913
7095
  id: string;
6914
7096
  name: string;
6915
7097
  api: "openai-completions";
6916
7098
  provider: string;
6917
7099
  baseUrl: string;
6918
7100
  reasoning: true;
6919
- input: ("image" | "text")[];
7101
+ input: "text"[];
6920
7102
  cost: {
6921
7103
  input: number;
6922
7104
  output: number;
@@ -6926,12 +7108,10 @@ export declare const MODELS: {
6926
7108
  contextWindow: number;
6927
7109
  maxTokens: number;
6928
7110
  };
6929
- };
6930
- readonly "opencode-go": {
6931
- readonly "glm-5": {
7111
+ readonly "minimax-m2.5-free": {
6932
7112
  id: string;
6933
7113
  name: string;
6934
- api: "openai-completions";
7114
+ api: "anthropic-messages";
6935
7115
  provider: string;
6936
7116
  baseUrl: string;
6937
7117
  reasoning: true;
@@ -6945,7 +7125,7 @@ export declare const MODELS: {
6945
7125
  contextWindow: number;
6946
7126
  maxTokens: number;
6947
7127
  };
6948
- readonly "glm-5.1": {
7128
+ readonly "minimax-m2.7": {
6949
7129
  id: string;
6950
7130
  name: string;
6951
7131
  api: "openai-completions";
@@ -6962,14 +7142,14 @@ export declare const MODELS: {
6962
7142
  contextWindow: number;
6963
7143
  maxTokens: number;
6964
7144
  };
6965
- readonly "kimi-k2.5": {
7145
+ readonly "nemotron-3-super-free": {
6966
7146
  id: string;
6967
7147
  name: string;
6968
7148
  api: "openai-completions";
6969
7149
  provider: string;
6970
7150
  baseUrl: string;
6971
7151
  reasoning: true;
6972
- input: ("image" | "text")[];
7152
+ input: "text"[];
6973
7153
  cost: {
6974
7154
  input: number;
6975
7155
  output: number;
@@ -6979,10 +7159,10 @@ export declare const MODELS: {
6979
7159
  contextWindow: number;
6980
7160
  maxTokens: number;
6981
7161
  };
6982
- readonly "mimo-v2-omni": {
7162
+ readonly "qwen3.5-plus": {
6983
7163
  id: string;
6984
7164
  name: string;
6985
- api: "openai-completions";
7165
+ api: "anthropic-messages";
6986
7166
  provider: string;
6987
7167
  baseUrl: string;
6988
7168
  reasoning: true;
@@ -6996,14 +7176,14 @@ export declare const MODELS: {
6996
7176
  contextWindow: number;
6997
7177
  maxTokens: number;
6998
7178
  };
6999
- readonly "mimo-v2-pro": {
7179
+ readonly "qwen3.6-plus": {
7000
7180
  id: string;
7001
7181
  name: string;
7002
- api: "openai-completions";
7182
+ api: "anthropic-messages";
7003
7183
  provider: string;
7004
7184
  baseUrl: string;
7005
7185
  reasoning: true;
7006
- input: "text"[];
7186
+ input: ("image" | "text")[];
7007
7187
  cost: {
7008
7188
  input: number;
7009
7189
  output: number;
@@ -7013,7 +7193,9 @@ export declare const MODELS: {
7013
7193
  contextWindow: number;
7014
7194
  maxTokens: number;
7015
7195
  };
7016
- readonly "minimax-m2.5": {
7196
+ };
7197
+ readonly "opencode-go": {
7198
+ readonly "deepseek-v4-flash": {
7017
7199
  id: string;
7018
7200
  name: string;
7019
7201
  api: "openai-completions";
@@ -7030,10 +7212,10 @@ export declare const MODELS: {
7030
7212
  contextWindow: number;
7031
7213
  maxTokens: number;
7032
7214
  };
7033
- readonly "minimax-m2.7": {
7215
+ readonly "deepseek-v4-pro": {
7034
7216
  id: string;
7035
7217
  name: string;
7036
- api: "anthropic-messages";
7218
+ api: "openai-completions";
7037
7219
  provider: string;
7038
7220
  baseUrl: string;
7039
7221
  reasoning: true;
@@ -7047,7 +7229,41 @@ export declare const MODELS: {
7047
7229
  contextWindow: number;
7048
7230
  maxTokens: number;
7049
7231
  };
7050
- readonly "qwen3.5-plus": {
7232
+ readonly "glm-5": {
7233
+ id: string;
7234
+ name: string;
7235
+ api: "openai-completions";
7236
+ provider: string;
7237
+ baseUrl: string;
7238
+ reasoning: true;
7239
+ input: "text"[];
7240
+ cost: {
7241
+ input: number;
7242
+ output: number;
7243
+ cacheRead: number;
7244
+ cacheWrite: number;
7245
+ };
7246
+ contextWindow: number;
7247
+ maxTokens: number;
7248
+ };
7249
+ readonly "glm-5.1": {
7250
+ id: string;
7251
+ name: string;
7252
+ api: "openai-completions";
7253
+ provider: string;
7254
+ baseUrl: string;
7255
+ reasoning: true;
7256
+ input: "text"[];
7257
+ cost: {
7258
+ input: number;
7259
+ output: number;
7260
+ cacheRead: number;
7261
+ cacheWrite: number;
7262
+ };
7263
+ contextWindow: number;
7264
+ maxTokens: number;
7265
+ };
7266
+ readonly "kimi-k2.5": {
7051
7267
  id: string;
7052
7268
  name: string;
7053
7269
  api: "openai-completions";
@@ -7064,13 +7280,149 @@ export declare const MODELS: {
7064
7280
  contextWindow: number;
7065
7281
  maxTokens: number;
7066
7282
  };
7067
- readonly "qwen3.6-plus": {
7283
+ readonly "kimi-k2.6": {
7284
+ id: string;
7285
+ name: string;
7286
+ api: "openai-completions";
7287
+ provider: string;
7288
+ baseUrl: string;
7289
+ reasoning: true;
7290
+ input: ("image" | "text")[];
7291
+ cost: {
7292
+ input: number;
7293
+ output: number;
7294
+ cacheRead: number;
7295
+ cacheWrite: number;
7296
+ };
7297
+ contextWindow: number;
7298
+ maxTokens: number;
7299
+ };
7300
+ readonly "mimo-v2-omni": {
7301
+ id: string;
7302
+ name: string;
7303
+ api: "openai-completions";
7304
+ provider: string;
7305
+ baseUrl: string;
7306
+ reasoning: true;
7307
+ input: ("image" | "text")[];
7308
+ cost: {
7309
+ input: number;
7310
+ output: number;
7311
+ cacheRead: number;
7312
+ cacheWrite: number;
7313
+ };
7314
+ contextWindow: number;
7315
+ maxTokens: number;
7316
+ };
7317
+ readonly "mimo-v2-pro": {
7318
+ id: string;
7319
+ name: string;
7320
+ api: "openai-completions";
7321
+ provider: string;
7322
+ baseUrl: string;
7323
+ reasoning: true;
7324
+ input: "text"[];
7325
+ cost: {
7326
+ input: number;
7327
+ output: number;
7328
+ cacheRead: number;
7329
+ cacheWrite: number;
7330
+ };
7331
+ contextWindow: number;
7332
+ maxTokens: number;
7333
+ };
7334
+ readonly "mimo-v2.5": {
7335
+ id: string;
7336
+ name: string;
7337
+ api: "openai-completions";
7338
+ provider: string;
7339
+ baseUrl: string;
7340
+ reasoning: true;
7341
+ input: ("image" | "text")[];
7342
+ cost: {
7343
+ input: number;
7344
+ output: number;
7345
+ cacheRead: number;
7346
+ cacheWrite: number;
7347
+ };
7348
+ contextWindow: number;
7349
+ maxTokens: number;
7350
+ };
7351
+ readonly "mimo-v2.5-pro": {
7352
+ id: string;
7353
+ name: string;
7354
+ api: "openai-completions";
7355
+ provider: string;
7356
+ baseUrl: string;
7357
+ reasoning: true;
7358
+ input: "text"[];
7359
+ cost: {
7360
+ input: number;
7361
+ output: number;
7362
+ cacheRead: number;
7363
+ cacheWrite: number;
7364
+ };
7365
+ contextWindow: number;
7366
+ maxTokens: number;
7367
+ };
7368
+ readonly "minimax-m2.5": {
7068
7369
  id: string;
7069
7370
  name: string;
7070
7371
  api: "openai-completions";
7071
7372
  provider: string;
7072
7373
  baseUrl: string;
7073
7374
  reasoning: true;
7375
+ input: "text"[];
7376
+ cost: {
7377
+ input: number;
7378
+ output: number;
7379
+ cacheRead: number;
7380
+ cacheWrite: number;
7381
+ };
7382
+ contextWindow: number;
7383
+ maxTokens: number;
7384
+ };
7385
+ readonly "minimax-m2.7": {
7386
+ id: string;
7387
+ name: string;
7388
+ api: "anthropic-messages";
7389
+ provider: string;
7390
+ baseUrl: string;
7391
+ reasoning: true;
7392
+ input: "text"[];
7393
+ cost: {
7394
+ input: number;
7395
+ output: number;
7396
+ cacheRead: number;
7397
+ cacheWrite: number;
7398
+ };
7399
+ contextWindow: number;
7400
+ maxTokens: number;
7401
+ };
7402
+ readonly "qwen3.5-plus": {
7403
+ id: string;
7404
+ name: string;
7405
+ api: "anthropic-messages";
7406
+ provider: string;
7407
+ baseUrl: string;
7408
+ reasoning: true;
7409
+ input: ("image" | "text")[];
7410
+ cost: {
7411
+ input: number;
7412
+ output: number;
7413
+ cacheRead: number;
7414
+ cacheWrite: number;
7415
+ };
7416
+ contextWindow: number;
7417
+ maxTokens: number;
7418
+ };
7419
+ readonly "qwen3.6-plus": {
7420
+ id: string;
7421
+ name: string;
7422
+ api: "anthropic-messages";
7423
+ provider: string;
7424
+ baseUrl: string;
7425
+ reasoning: true;
7074
7426
  input: ("image" | "text")[];
7075
7427
  cost: {
7076
7428
  input: number;
@@ -7457,7 +7809,7 @@ export declare const MODELS: {
7457
7809
  contextWindow: number;
7458
7810
  maxTokens: number;
7459
7811
  };
7460
- readonly "arcee-ai/trinity-large-preview:free": {
7812
+ readonly "arcee-ai/trinity-large-preview": {
7461
7813
  id: string;
7462
7814
  name: string;
7463
7815
  api: "openai-completions";
@@ -7814,6 +8166,40 @@ export declare const MODELS: {
7814
8166
  contextWindow: number;
7815
8167
  maxTokens: number;
7816
8168
  };
8169
+ readonly "deepseek/deepseek-v4-flash": {
8170
+ id: string;
8171
+ name: string;
8172
+ api: "openai-completions";
8173
+ provider: string;
8174
+ baseUrl: string;
8175
+ reasoning: true;
8176
+ input: "text"[];
8177
+ cost: {
8178
+ input: number;
8179
+ output: number;
8180
+ cacheRead: number;
8181
+ cacheWrite: number;
8182
+ };
8183
+ contextWindow: number;
8184
+ maxTokens: number;
8185
+ };
8186
+ readonly "deepseek/deepseek-v4-pro": {
8187
+ id: string;
8188
+ name: string;
8189
+ api: "openai-completions";
8190
+ provider: string;
8191
+ baseUrl: string;
8192
+ reasoning: true;
8193
+ input: "text"[];
8194
+ cost: {
8195
+ input: number;
8196
+ output: number;
8197
+ cacheRead: number;
8198
+ cacheWrite: number;
8199
+ };
8200
+ contextWindow: number;
8201
+ maxTokens: number;
8202
+ };
7817
8203
  readonly "essentialai/rnj-1-instruct": {
7818
8204
  id: string;
7819
8205
  name: string;
@@ -8035,13 +8421,13 @@ export declare const MODELS: {
8035
8421
  contextWindow: number;
8036
8422
  maxTokens: number;
8037
8423
  };
8038
- readonly "google/gemma-4-26b-a4b-it": {
8424
+ readonly "google/gemma-3-12b-it": {
8039
8425
  id: string;
8040
8426
  name: string;
8041
8427
  api: "openai-completions";
8042
8428
  provider: string;
8043
8429
  baseUrl: string;
8044
- reasoning: true;
8430
+ reasoning: false;
8045
8431
  input: ("image" | "text")[];
8046
8432
  cost: {
8047
8433
  input: number;
@@ -8052,13 +8438,13 @@ export declare const MODELS: {
8052
8438
  contextWindow: number;
8053
8439
  maxTokens: number;
8054
8440
  };
8055
- readonly "google/gemma-4-26b-a4b-it:free": {
8441
+ readonly "google/gemma-3-27b-it": {
8056
8442
  id: string;
8057
8443
  name: string;
8058
8444
  api: "openai-completions";
8059
8445
  provider: string;
8060
8446
  baseUrl: string;
8061
- reasoning: true;
8447
+ reasoning: false;
8062
8448
  input: ("image" | "text")[];
8063
8449
  cost: {
8064
8450
  input: number;
@@ -8069,7 +8455,7 @@ export declare const MODELS: {
8069
8455
  contextWindow: number;
8070
8456
  maxTokens: number;
8071
8457
  };
8072
- readonly "google/gemma-4-31b-it": {
8458
+ readonly "google/gemma-4-26b-a4b-it": {
8073
8459
  id: string;
8074
8460
  name: string;
8075
8461
  api: "openai-completions";
@@ -8086,7 +8472,7 @@ export declare const MODELS: {
8086
8472
  contextWindow: number;
8087
8473
  maxTokens: number;
8088
8474
  };
8089
- readonly "google/gemma-4-31b-it:free": {
8475
+ readonly "google/gemma-4-26b-a4b-it:free": {
8090
8476
  id: string;
8091
8477
  name: string;
8092
8478
  api: "openai-completions";
@@ -8103,14 +8489,14 @@ export declare const MODELS: {
8103
8489
  contextWindow: number;
8104
8490
  maxTokens: number;
8105
8491
  };
8106
- readonly "inception/mercury-2": {
8492
+ readonly "google/gemma-4-31b-it": {
8107
8493
  id: string;
8108
8494
  name: string;
8109
8495
  api: "openai-completions";
8110
8496
  provider: string;
8111
8497
  baseUrl: string;
8112
8498
  reasoning: true;
8113
- input: "text"[];
8499
+ input: ("image" | "text")[];
8114
8500
  cost: {
8115
8501
  input: number;
8116
8502
  output: number;
@@ -8120,14 +8506,14 @@ export declare const MODELS: {
8120
8506
  contextWindow: number;
8121
8507
  maxTokens: number;
8122
8508
  };
8123
- readonly "kwaipilot/kat-coder-pro-v2": {
8509
+ readonly "google/gemma-4-31b-it:free": {
8124
8510
  id: string;
8125
8511
  name: string;
8126
8512
  api: "openai-completions";
8127
8513
  provider: string;
8128
8514
  baseUrl: string;
8129
- reasoning: false;
8130
- input: "text"[];
8515
+ reasoning: true;
8516
+ input: ("image" | "text")[];
8131
8517
  cost: {
8132
8518
  input: number;
8133
8519
  output: number;
@@ -8137,7 +8523,7 @@ export declare const MODELS: {
8137
8523
  contextWindow: number;
8138
8524
  maxTokens: number;
8139
8525
  };
8140
- readonly "meta-llama/llama-3-8b-instruct": {
8526
+ readonly "ibm-granite/granite-4.1-8b": {
8141
8527
  id: string;
8142
8528
  name: string;
8143
8529
  api: "openai-completions";
@@ -8154,13 +8540,13 @@ export declare const MODELS: {
8154
8540
  contextWindow: number;
8155
8541
  maxTokens: number;
8156
8542
  };
8157
- readonly "meta-llama/llama-3.1-70b-instruct": {
8543
+ readonly "inception/mercury-2": {
8158
8544
  id: string;
8159
8545
  name: string;
8160
8546
  api: "openai-completions";
8161
8547
  provider: string;
8162
8548
  baseUrl: string;
8163
- reasoning: false;
8549
+ reasoning: true;
8164
8550
  input: "text"[];
8165
8551
  cost: {
8166
8552
  input: number;
@@ -8171,7 +8557,7 @@ export declare const MODELS: {
8171
8557
  contextWindow: number;
8172
8558
  maxTokens: number;
8173
8559
  };
8174
- readonly "meta-llama/llama-3.1-8b-instruct": {
8560
+ readonly "inclusionai/ling-2.6-1t:free": {
8175
8561
  id: string;
8176
8562
  name: string;
8177
8563
  api: "openai-completions";
@@ -8188,7 +8574,7 @@ export declare const MODELS: {
8188
8574
  contextWindow: number;
8189
8575
  maxTokens: number;
8190
8576
  };
8191
- readonly "meta-llama/llama-3.3-70b-instruct": {
8577
+ readonly "inclusionai/ling-2.6-flash": {
8192
8578
  id: string;
8193
8579
  name: string;
8194
8580
  api: "openai-completions";
@@ -8205,7 +8591,7 @@ export declare const MODELS: {
8205
8591
  contextWindow: number;
8206
8592
  maxTokens: number;
8207
8593
  };
8208
- readonly "meta-llama/llama-3.3-70b-instruct:free": {
8594
+ readonly "kwaipilot/kat-coder-pro-v2": {
8209
8595
  id: string;
8210
8596
  name: string;
8211
8597
  api: "openai-completions";
@@ -8222,14 +8608,14 @@ export declare const MODELS: {
8222
8608
  contextWindow: number;
8223
8609
  maxTokens: number;
8224
8610
  };
8225
- readonly "meta-llama/llama-4-scout": {
8611
+ readonly "meta-llama/llama-3-8b-instruct": {
8226
8612
  id: string;
8227
8613
  name: string;
8228
8614
  api: "openai-completions";
8229
8615
  provider: string;
8230
8616
  baseUrl: string;
8231
8617
  reasoning: false;
8232
- input: ("image" | "text")[];
8618
+ input: "text"[];
8233
8619
  cost: {
8234
8620
  input: number;
8235
8621
  output: number;
@@ -8239,13 +8625,13 @@ export declare const MODELS: {
8239
8625
  contextWindow: number;
8240
8626
  maxTokens: number;
8241
8627
  };
8242
- readonly "minimax/minimax-m1": {
8628
+ readonly "meta-llama/llama-3.1-70b-instruct": {
8243
8629
  id: string;
8244
8630
  name: string;
8245
8631
  api: "openai-completions";
8246
8632
  provider: string;
8247
8633
  baseUrl: string;
8248
- reasoning: true;
8634
+ reasoning: false;
8249
8635
  input: "text"[];
8250
8636
  cost: {
8251
8637
  input: number;
@@ -8256,13 +8642,13 @@ export declare const MODELS: {
8256
8642
  contextWindow: number;
8257
8643
  maxTokens: number;
8258
8644
  };
8259
- readonly "minimax/minimax-m2": {
8645
+ readonly "meta-llama/llama-3.1-8b-instruct": {
8260
8646
  id: string;
8261
8647
  name: string;
8262
8648
  api: "openai-completions";
8263
8649
  provider: string;
8264
8650
  baseUrl: string;
8265
- reasoning: true;
8651
+ reasoning: false;
8266
8652
  input: "text"[];
8267
8653
  cost: {
8268
8654
  input: number;
@@ -8273,13 +8659,13 @@ export declare const MODELS: {
8273
8659
  contextWindow: number;
8274
8660
  maxTokens: number;
8275
8661
  };
8276
- readonly "minimax/minimax-m2.1": {
8662
+ readonly "meta-llama/llama-3.3-70b-instruct": {
8277
8663
  id: string;
8278
8664
  name: string;
8279
8665
  api: "openai-completions";
8280
8666
  provider: string;
8281
8667
  baseUrl: string;
8282
- reasoning: true;
8668
+ reasoning: false;
8283
8669
  input: "text"[];
8284
8670
  cost: {
8285
8671
  input: number;
@@ -8290,13 +8676,13 @@ export declare const MODELS: {
8290
8676
  contextWindow: number;
8291
8677
  maxTokens: number;
8292
8678
  };
8293
- readonly "minimax/minimax-m2.5": {
8679
+ readonly "meta-llama/llama-3.3-70b-instruct:free": {
8294
8680
  id: string;
8295
8681
  name: string;
8296
8682
  api: "openai-completions";
8297
8683
  provider: string;
8298
8684
  baseUrl: string;
8299
- reasoning: true;
8685
+ reasoning: false;
8300
8686
  input: "text"[];
8301
8687
  cost: {
8302
8688
  input: number;
@@ -8307,15 +8693,100 @@ export declare const MODELS: {
8307
8693
  contextWindow: number;
8308
8694
  maxTokens: number;
8309
8695
  };
8310
- readonly "minimax/minimax-m2.5:free": {
8696
+ readonly "meta-llama/llama-4-scout": {
8311
8697
  id: string;
8312
8698
  name: string;
8313
8699
  api: "openai-completions";
8314
8700
  provider: string;
8315
8701
  baseUrl: string;
8316
- reasoning: true;
8317
- input: "text"[];
8318
- cost: {
8702
+ reasoning: false;
8703
+ input: ("image" | "text")[];
8704
+ cost: {
8705
+ input: number;
8706
+ output: number;
8707
+ cacheRead: number;
8708
+ cacheWrite: number;
8709
+ };
8710
+ contextWindow: number;
8711
+ maxTokens: number;
8712
+ };
8713
+ readonly "minimax/minimax-m1": {
8714
+ id: string;
8715
+ name: string;
8716
+ api: "openai-completions";
8717
+ provider: string;
8718
+ baseUrl: string;
8719
+ reasoning: true;
8720
+ input: "text"[];
8721
+ cost: {
8722
+ input: number;
8723
+ output: number;
8724
+ cacheRead: number;
8725
+ cacheWrite: number;
8726
+ };
8727
+ contextWindow: number;
8728
+ maxTokens: number;
8729
+ };
8730
+ readonly "minimax/minimax-m2": {
8731
+ id: string;
8732
+ name: string;
8733
+ api: "openai-completions";
8734
+ provider: string;
8735
+ baseUrl: string;
8736
+ reasoning: true;
8737
+ input: "text"[];
8738
+ cost: {
8739
+ input: number;
8740
+ output: number;
8741
+ cacheRead: number;
8742
+ cacheWrite: number;
8743
+ };
8744
+ contextWindow: number;
8745
+ maxTokens: number;
8746
+ };
8747
+ readonly "minimax/minimax-m2.1": {
8748
+ id: string;
8749
+ name: string;
8750
+ api: "openai-completions";
8751
+ provider: string;
8752
+ baseUrl: string;
8753
+ reasoning: true;
8754
+ input: "text"[];
8755
+ cost: {
8756
+ input: number;
8757
+ output: number;
8758
+ cacheRead: number;
8759
+ cacheWrite: number;
8760
+ };
8761
+ contextWindow: number;
8762
+ maxTokens: number;
8763
+ };
8764
+ readonly "minimax/minimax-m2.5": {
8765
+ id: string;
8766
+ name: string;
8767
+ api: "openai-completions";
8768
+ provider: string;
8769
+ baseUrl: string;
8770
+ reasoning: true;
8771
+ input: "text"[];
8772
+ cost: {
8773
+ input: number;
8774
+ output: number;
8775
+ cacheRead: number;
8776
+ cacheWrite: number;
8777
+ };
8778
+ contextWindow: number;
8779
+ maxTokens: number;
8780
+ };
8781
+ readonly "minimax/minimax-m2.5:free": {
8782
+ id: string;
8783
+ name: string;
8784
+ api: "openai-completions";
8785
+ provider: string;
8786
+ baseUrl: string;
8787
+ reasoning: true;
8788
+ input: "text"[];
8789
+ cost: {
8319
8790
  input: number;
8320
8791
  output: number;
8321
8792
  cacheRead: number;
@@ -8630,7 +9101,7 @@ export declare const MODELS: {
8630
9101
  contextWindow: number;
8631
9102
  maxTokens: number;
8632
9103
  };
8633
- readonly "mistralai/mistral-small-creative": {
9104
+ readonly "mistralai/mixtral-8x22b-instruct": {
8634
9105
  id: string;
8635
9106
  name: string;
8636
9107
  api: "openai-completions";
@@ -8647,7 +9118,7 @@ export declare const MODELS: {
8647
9118
  contextWindow: number;
8648
9119
  maxTokens: number;
8649
9120
  };
8650
- readonly "mistralai/mixtral-8x22b-instruct": {
9121
+ readonly "mistralai/mixtral-8x7b-instruct": {
8651
9122
  id: string;
8652
9123
  name: string;
8653
9124
  api: "openai-completions";
@@ -8664,14 +9135,14 @@ export declare const MODELS: {
8664
9135
  contextWindow: number;
8665
9136
  maxTokens: number;
8666
9137
  };
8667
- readonly "mistralai/mixtral-8x7b-instruct": {
9138
+ readonly "mistralai/pixtral-large-2411": {
8668
9139
  id: string;
8669
9140
  name: string;
8670
9141
  api: "openai-completions";
8671
9142
  provider: string;
8672
9143
  baseUrl: string;
8673
9144
  reasoning: false;
8674
- input: "text"[];
9145
+ input: ("image" | "text")[];
8675
9146
  cost: {
8676
9147
  input: number;
8677
9148
  output: number;
@@ -8681,14 +9152,14 @@ export declare const MODELS: {
8681
9152
  contextWindow: number;
8682
9153
  maxTokens: number;
8683
9154
  };
8684
- readonly "mistralai/pixtral-large-2411": {
9155
+ readonly "mistralai/voxtral-small-24b-2507": {
8685
9156
  id: string;
8686
9157
  name: string;
8687
9158
  api: "openai-completions";
8688
9159
  provider: string;
8689
9160
  baseUrl: string;
8690
9161
  reasoning: false;
8691
- input: ("image" | "text")[];
9162
+ input: "text"[];
8692
9163
  cost: {
8693
9164
  input: number;
8694
9165
  output: number;
@@ -8698,7 +9169,7 @@ export declare const MODELS: {
8698
9169
  contextWindow: number;
8699
9170
  maxTokens: number;
8700
9171
  };
8701
- readonly "mistralai/voxtral-small-24b-2507": {
9172
+ readonly "moonshotai/kimi-k2": {
8702
9173
  id: string;
8703
9174
  name: string;
8704
9175
  api: "openai-completions";
@@ -8715,7 +9186,7 @@ export declare const MODELS: {
8715
9186
  contextWindow: number;
8716
9187
  maxTokens: number;
8717
9188
  };
8718
- readonly "moonshotai/kimi-k2": {
9189
+ readonly "moonshotai/kimi-k2-0905": {
8719
9190
  id: string;
8720
9191
  name: string;
8721
9192
  api: "openai-completions";
@@ -8732,13 +9203,13 @@ export declare const MODELS: {
8732
9203
  contextWindow: number;
8733
9204
  maxTokens: number;
8734
9205
  };
8735
- readonly "moonshotai/kimi-k2-0905": {
9206
+ readonly "moonshotai/kimi-k2-thinking": {
8736
9207
  id: string;
8737
9208
  name: string;
8738
9209
  api: "openai-completions";
8739
9210
  provider: string;
8740
9211
  baseUrl: string;
8741
- reasoning: false;
9212
+ reasoning: true;
8742
9213
  input: "text"[];
8743
9214
  cost: {
8744
9215
  input: number;
@@ -8749,14 +9220,14 @@ export declare const MODELS: {
8749
9220
  contextWindow: number;
8750
9221
  maxTokens: number;
8751
9222
  };
8752
- readonly "moonshotai/kimi-k2-thinking": {
9223
+ readonly "moonshotai/kimi-k2.5": {
8753
9224
  id: string;
8754
9225
  name: string;
8755
9226
  api: "openai-completions";
8756
9227
  provider: string;
8757
9228
  baseUrl: string;
8758
9229
  reasoning: true;
8759
- input: "text"[];
9230
+ input: ("image" | "text")[];
8760
9231
  cost: {
8761
9232
  input: number;
8762
9233
  output: number;
@@ -8766,7 +9237,7 @@ export declare const MODELS: {
8766
9237
  contextWindow: number;
8767
9238
  maxTokens: number;
8768
9239
  };
8769
- readonly "moonshotai/kimi-k2.5": {
9240
+ readonly "moonshotai/kimi-k2.6": {
8770
9241
  id: string;
8771
9242
  name: string;
8772
9243
  api: "openai-completions";
@@ -8868,6 +9339,23 @@ export declare const MODELS: {
8868
9339
  contextWindow: number;
8869
9340
  maxTokens: number;
8870
9341
  };
9342
+ readonly "nvidia/nemotron-3-nano-omni-30b-a3b-reasoning:free": {
9343
+ id: string;
9344
+ name: string;
9345
+ api: "openai-completions";
9346
+ provider: string;
9347
+ baseUrl: string;
9348
+ reasoning: true;
9349
+ input: ("image" | "text")[];
9350
+ cost: {
9351
+ input: number;
9352
+ output: number;
9353
+ cacheRead: number;
9354
+ cacheWrite: number;
9355
+ };
9356
+ contextWindow: number;
9357
+ maxTokens: number;
9358
+ };
8871
9359
  readonly "nvidia/nemotron-3-super-120b-a12b": {
8872
9360
  id: string;
8873
9361
  name: string;
@@ -9293,7 +9781,7 @@ export declare const MODELS: {
9293
9781
  contextWindow: number;
9294
9782
  maxTokens: number;
9295
9783
  };
9296
- readonly "openai/gpt-5-image": {
9784
+ readonly "openai/gpt-5-mini": {
9297
9785
  id: string;
9298
9786
  name: string;
9299
9787
  api: "openai-completions";
@@ -9310,7 +9798,7 @@ export declare const MODELS: {
9310
9798
  contextWindow: number;
9311
9799
  maxTokens: number;
9312
9800
  };
9313
- readonly "openai/gpt-5-image-mini": {
9801
+ readonly "openai/gpt-5-nano": {
9314
9802
  id: string;
9315
9803
  name: string;
9316
9804
  api: "openai-completions";
@@ -9327,7 +9815,7 @@ export declare const MODELS: {
9327
9815
  contextWindow: number;
9328
9816
  maxTokens: number;
9329
9817
  };
9330
- readonly "openai/gpt-5-mini": {
9818
+ readonly "openai/gpt-5-pro": {
9331
9819
  id: string;
9332
9820
  name: string;
9333
9821
  api: "openai-completions";
@@ -9344,7 +9832,7 @@ export declare const MODELS: {
9344
9832
  contextWindow: number;
9345
9833
  maxTokens: number;
9346
9834
  };
9347
- readonly "openai/gpt-5-nano": {
9835
+ readonly "openai/gpt-5.1": {
9348
9836
  id: string;
9349
9837
  name: string;
9350
9838
  api: "openai-completions";
@@ -9361,13 +9849,13 @@ export declare const MODELS: {
9361
9849
  contextWindow: number;
9362
9850
  maxTokens: number;
9363
9851
  };
9364
- readonly "openai/gpt-5-pro": {
9852
+ readonly "openai/gpt-5.1-chat": {
9365
9853
  id: string;
9366
9854
  name: string;
9367
9855
  api: "openai-completions";
9368
9856
  provider: string;
9369
9857
  baseUrl: string;
9370
- reasoning: true;
9858
+ reasoning: false;
9371
9859
  input: ("image" | "text")[];
9372
9860
  cost: {
9373
9861
  input: number;
@@ -9378,7 +9866,7 @@ export declare const MODELS: {
9378
9866
  contextWindow: number;
9379
9867
  maxTokens: number;
9380
9868
  };
9381
- readonly "openai/gpt-5.1": {
9869
+ readonly "openai/gpt-5.1-codex": {
9382
9870
  id: string;
9383
9871
  name: string;
9384
9872
  api: "openai-completions";
@@ -9395,13 +9883,13 @@ export declare const MODELS: {
9395
9883
  contextWindow: number;
9396
9884
  maxTokens: number;
9397
9885
  };
9398
- readonly "openai/gpt-5.1-chat": {
9886
+ readonly "openai/gpt-5.1-codex-max": {
9399
9887
  id: string;
9400
9888
  name: string;
9401
9889
  api: "openai-completions";
9402
9890
  provider: string;
9403
9891
  baseUrl: string;
9404
- reasoning: false;
9892
+ reasoning: true;
9405
9893
  input: ("image" | "text")[];
9406
9894
  cost: {
9407
9895
  input: number;
@@ -9412,7 +9900,7 @@ export declare const MODELS: {
9412
9900
  contextWindow: number;
9413
9901
  maxTokens: number;
9414
9902
  };
9415
- readonly "openai/gpt-5.1-codex": {
9903
+ readonly "openai/gpt-5.1-codex-mini": {
9416
9904
  id: string;
9417
9905
  name: string;
9418
9906
  api: "openai-completions";
@@ -9429,7 +9917,7 @@ export declare const MODELS: {
9429
9917
  contextWindow: number;
9430
9918
  maxTokens: number;
9431
9919
  };
9432
- readonly "openai/gpt-5.1-codex-max": {
9920
+ readonly "openai/gpt-5.2": {
9433
9921
  id: string;
9434
9922
  name: string;
9435
9923
  api: "openai-completions";
@@ -9446,13 +9934,13 @@ export declare const MODELS: {
9446
9934
  contextWindow: number;
9447
9935
  maxTokens: number;
9448
9936
  };
9449
- readonly "openai/gpt-5.1-codex-mini": {
9937
+ readonly "openai/gpt-5.2-chat": {
9450
9938
  id: string;
9451
9939
  name: string;
9452
9940
  api: "openai-completions";
9453
9941
  provider: string;
9454
9942
  baseUrl: string;
9455
- reasoning: true;
9943
+ reasoning: false;
9456
9944
  input: ("image" | "text")[];
9457
9945
  cost: {
9458
9946
  input: number;
@@ -9463,7 +9951,7 @@ export declare const MODELS: {
9463
9951
  contextWindow: number;
9464
9952
  maxTokens: number;
9465
9953
  };
9466
- readonly "openai/gpt-5.2": {
9954
+ readonly "openai/gpt-5.2-codex": {
9467
9955
  id: string;
9468
9956
  name: string;
9469
9957
  api: "openai-completions";
@@ -9480,13 +9968,13 @@ export declare const MODELS: {
9480
9968
  contextWindow: number;
9481
9969
  maxTokens: number;
9482
9970
  };
9483
- readonly "openai/gpt-5.2-chat": {
9971
+ readonly "openai/gpt-5.2-pro": {
9484
9972
  id: string;
9485
9973
  name: string;
9486
9974
  api: "openai-completions";
9487
9975
  provider: string;
9488
9976
  baseUrl: string;
9489
- reasoning: false;
9977
+ reasoning: true;
9490
9978
  input: ("image" | "text")[];
9491
9979
  cost: {
9492
9980
  input: number;
@@ -9497,13 +9985,13 @@ export declare const MODELS: {
9497
9985
  contextWindow: number;
9498
9986
  maxTokens: number;
9499
9987
  };
9500
- readonly "openai/gpt-5.2-codex": {
9988
+ readonly "openai/gpt-5.3-chat": {
9501
9989
  id: string;
9502
9990
  name: string;
9503
9991
  api: "openai-completions";
9504
9992
  provider: string;
9505
9993
  baseUrl: string;
9506
- reasoning: true;
9994
+ reasoning: false;
9507
9995
  input: ("image" | "text")[];
9508
9996
  cost: {
9509
9997
  input: number;
@@ -9514,7 +10002,7 @@ export declare const MODELS: {
9514
10002
  contextWindow: number;
9515
10003
  maxTokens: number;
9516
10004
  };
9517
- readonly "openai/gpt-5.2-pro": {
10005
+ readonly "openai/gpt-5.3-codex": {
9518
10006
  id: string;
9519
10007
  name: string;
9520
10008
  api: "openai-completions";
@@ -9531,13 +10019,13 @@ export declare const MODELS: {
9531
10019
  contextWindow: number;
9532
10020
  maxTokens: number;
9533
10021
  };
9534
- readonly "openai/gpt-5.3-chat": {
10022
+ readonly "openai/gpt-5.4": {
9535
10023
  id: string;
9536
10024
  name: string;
9537
10025
  api: "openai-completions";
9538
10026
  provider: string;
9539
10027
  baseUrl: string;
9540
- reasoning: false;
10028
+ reasoning: true;
9541
10029
  input: ("image" | "text")[];
9542
10030
  cost: {
9543
10031
  input: number;
@@ -9548,7 +10036,7 @@ export declare const MODELS: {
9548
10036
  contextWindow: number;
9549
10037
  maxTokens: number;
9550
10038
  };
9551
- readonly "openai/gpt-5.3-codex": {
10039
+ readonly "openai/gpt-5.4-mini": {
9552
10040
  id: string;
9553
10041
  name: string;
9554
10042
  api: "openai-completions";
@@ -9565,7 +10053,7 @@ export declare const MODELS: {
9565
10053
  contextWindow: number;
9566
10054
  maxTokens: number;
9567
10055
  };
9568
- readonly "openai/gpt-5.4": {
10056
+ readonly "openai/gpt-5.4-nano": {
9569
10057
  id: string;
9570
10058
  name: string;
9571
10059
  api: "openai-completions";
@@ -9582,7 +10070,7 @@ export declare const MODELS: {
9582
10070
  contextWindow: number;
9583
10071
  maxTokens: number;
9584
10072
  };
9585
- readonly "openai/gpt-5.4-mini": {
10073
+ readonly "openai/gpt-5.4-pro": {
9586
10074
  id: string;
9587
10075
  name: string;
9588
10076
  api: "openai-completions";
@@ -9599,7 +10087,7 @@ export declare const MODELS: {
9599
10087
  contextWindow: number;
9600
10088
  maxTokens: number;
9601
10089
  };
9602
- readonly "openai/gpt-5.4-nano": {
10090
+ readonly "openai/gpt-5.5": {
9603
10091
  id: string;
9604
10092
  name: string;
9605
10093
  api: "openai-completions";
@@ -9616,7 +10104,7 @@ export declare const MODELS: {
9616
10104
  contextWindow: number;
9617
10105
  maxTokens: number;
9618
10106
  };
9619
- readonly "openai/gpt-5.4-pro": {
10107
+ readonly "openai/gpt-5.5-pro": {
9620
10108
  id: string;
9621
10109
  name: string;
9622
10110
  api: "openai-completions";
@@ -9922,7 +10410,24 @@ export declare const MODELS: {
9922
10410
  contextWindow: number;
9923
10411
  maxTokens: number;
9924
10412
  };
9925
- readonly "openrouter/elephant-alpha": {
10413
+ readonly "openrouter/free": {
10414
+ id: string;
10415
+ name: string;
10416
+ api: "openai-completions";
10417
+ provider: string;
10418
+ baseUrl: string;
10419
+ reasoning: true;
10420
+ input: ("image" | "text")[];
10421
+ cost: {
10422
+ input: number;
10423
+ output: number;
10424
+ cacheRead: number;
10425
+ cacheWrite: number;
10426
+ };
10427
+ contextWindow: number;
10428
+ maxTokens: number;
10429
+ };
10430
+ readonly "openrouter/owl-alpha": {
9926
10431
  id: string;
9927
10432
  name: string;
9928
10433
  api: "openai-completions";
@@ -9939,14 +10444,31 @@ export declare const MODELS: {
9939
10444
  contextWindow: number;
9940
10445
  maxTokens: number;
9941
10446
  };
9942
- readonly "openrouter/free": {
10447
+ readonly "poolside/laguna-m.1:free": {
9943
10448
  id: string;
9944
10449
  name: string;
9945
10450
  api: "openai-completions";
9946
10451
  provider: string;
9947
10452
  baseUrl: string;
9948
10453
  reasoning: true;
9949
- input: ("image" | "text")[];
10454
+ input: "text"[];
10455
+ cost: {
10456
+ input: number;
10457
+ output: number;
10458
+ cacheRead: number;
10459
+ cacheWrite: number;
10460
+ };
10461
+ contextWindow: number;
10462
+ maxTokens: number;
10463
+ };
10464
+ readonly "poolside/laguna-xs.2:free": {
10465
+ id: string;
10466
+ name: string;
10467
+ api: "openai-completions";
10468
+ provider: string;
10469
+ baseUrl: string;
10470
+ reasoning: true;
10471
+ input: "text"[];
9950
10472
  cost: {
9951
10473
  input: number;
9952
10474
  output: number;
@@ -10625,8 +11147,263 @@ export declare const MODELS: {
10625
11147
  api: "openai-completions";
10626
11148
  provider: string;
10627
11149
  baseUrl: string;
10628
- reasoning: true;
10629
- input: ("image" | "text")[];
11150
+ reasoning: true;
11151
+ input: ("image" | "text")[];
11152
+ cost: {
11153
+ input: number;
11154
+ output: number;
11155
+ cacheRead: number;
11156
+ cacheWrite: number;
11157
+ };
11158
+ contextWindow: number;
11159
+ maxTokens: number;
11160
+ };
11161
+ readonly "qwen/qwen3.5-9b": {
11162
+ id: string;
11163
+ name: string;
11164
+ api: "openai-completions";
11165
+ provider: string;
11166
+ baseUrl: string;
11167
+ reasoning: true;
11168
+ input: ("image" | "text")[];
11169
+ cost: {
11170
+ input: number;
11171
+ output: number;
11172
+ cacheRead: number;
11173
+ cacheWrite: number;
11174
+ };
11175
+ contextWindow: number;
11176
+ maxTokens: number;
11177
+ };
11178
+ readonly "qwen/qwen3.5-flash-02-23": {
11179
+ id: string;
11180
+ name: string;
11181
+ api: "openai-completions";
11182
+ provider: string;
11183
+ baseUrl: string;
11184
+ reasoning: true;
11185
+ input: ("image" | "text")[];
11186
+ cost: {
11187
+ input: number;
11188
+ output: number;
11189
+ cacheRead: number;
11190
+ cacheWrite: number;
11191
+ };
11192
+ contextWindow: number;
11193
+ maxTokens: number;
11194
+ };
11195
+ readonly "qwen/qwen3.5-plus-02-15": {
11196
+ id: string;
11197
+ name: string;
11198
+ api: "openai-completions";
11199
+ provider: string;
11200
+ baseUrl: string;
11201
+ reasoning: true;
11202
+ input: ("image" | "text")[];
11203
+ cost: {
11204
+ input: number;
11205
+ output: number;
11206
+ cacheRead: number;
11207
+ cacheWrite: number;
11208
+ };
11209
+ contextWindow: number;
11210
+ maxTokens: number;
11211
+ };
11212
+ readonly "qwen/qwen3.5-plus-20260420": {
11213
+ id: string;
11214
+ name: string;
11215
+ api: "openai-completions";
11216
+ provider: string;
11217
+ baseUrl: string;
11218
+ reasoning: true;
11219
+ input: ("image" | "text")[];
11220
+ cost: {
11221
+ input: number;
11222
+ output: number;
11223
+ cacheRead: number;
11224
+ cacheWrite: number;
11225
+ };
11226
+ contextWindow: number;
11227
+ maxTokens: number;
11228
+ };
11229
+ readonly "qwen/qwen3.6-27b": {
11230
+ id: string;
11231
+ name: string;
11232
+ api: "openai-completions";
11233
+ provider: string;
11234
+ baseUrl: string;
11235
+ reasoning: true;
11236
+ input: ("image" | "text")[];
11237
+ cost: {
11238
+ input: number;
11239
+ output: number;
11240
+ cacheRead: number;
11241
+ cacheWrite: number;
11242
+ };
11243
+ contextWindow: number;
11244
+ maxTokens: number;
11245
+ };
11246
+ readonly "qwen/qwen3.6-flash": {
11247
+ id: string;
11248
+ name: string;
11249
+ api: "openai-completions";
11250
+ provider: string;
11251
+ baseUrl: string;
11252
+ reasoning: true;
11253
+ input: ("image" | "text")[];
11254
+ cost: {
11255
+ input: number;
11256
+ output: number;
11257
+ cacheRead: number;
11258
+ cacheWrite: number;
11259
+ };
11260
+ contextWindow: number;
11261
+ maxTokens: number;
11262
+ };
11263
+ readonly "qwen/qwen3.6-max-preview": {
11264
+ id: string;
11265
+ name: string;
11266
+ api: "openai-completions";
11267
+ provider: string;
11268
+ baseUrl: string;
11269
+ reasoning: true;
11270
+ input: "text"[];
11271
+ cost: {
11272
+ input: number;
11273
+ output: number;
11274
+ cacheRead: number;
11275
+ cacheWrite: number;
11276
+ };
11277
+ contextWindow: number;
11278
+ maxTokens: number;
11279
+ };
11280
+ readonly "qwen/qwen3.6-plus": {
11281
+ id: string;
11282
+ name: string;
11283
+ api: "openai-completions";
11284
+ provider: string;
11285
+ baseUrl: string;
11286
+ reasoning: true;
11287
+ input: ("image" | "text")[];
11288
+ cost: {
11289
+ input: number;
11290
+ output: number;
11291
+ cacheRead: number;
11292
+ cacheWrite: number;
11293
+ };
11294
+ contextWindow: number;
11295
+ maxTokens: number;
11296
+ };
11297
+ readonly "rekaai/reka-edge": {
11298
+ id: string;
11299
+ name: string;
11300
+ api: "openai-completions";
11301
+ provider: string;
11302
+ baseUrl: string;
11303
+ reasoning: false;
11304
+ input: ("image" | "text")[];
11305
+ cost: {
11306
+ input: number;
11307
+ output: number;
11308
+ cacheRead: number;
11309
+ cacheWrite: number;
11310
+ };
11311
+ contextWindow: number;
11312
+ maxTokens: number;
11313
+ };
11314
+ readonly "relace/relace-search": {
11315
+ id: string;
11316
+ name: string;
11317
+ api: "openai-completions";
11318
+ provider: string;
11319
+ baseUrl: string;
11320
+ reasoning: false;
11321
+ input: "text"[];
11322
+ cost: {
11323
+ input: number;
11324
+ output: number;
11325
+ cacheRead: number;
11326
+ cacheWrite: number;
11327
+ };
11328
+ contextWindow: number;
11329
+ maxTokens: number;
11330
+ };
11331
+ readonly "sao10k/l3-euryale-70b": {
11332
+ id: string;
11333
+ name: string;
11334
+ api: "openai-completions";
11335
+ provider: string;
11336
+ baseUrl: string;
11337
+ reasoning: false;
11338
+ input: "text"[];
11339
+ cost: {
11340
+ input: number;
11341
+ output: number;
11342
+ cacheRead: number;
11343
+ cacheWrite: number;
11344
+ };
11345
+ contextWindow: number;
11346
+ maxTokens: number;
11347
+ };
11348
+ readonly "sao10k/l3.1-euryale-70b": {
11349
+ id: string;
11350
+ name: string;
11351
+ api: "openai-completions";
11352
+ provider: string;
11353
+ baseUrl: string;
11354
+ reasoning: false;
11355
+ input: "text"[];
11356
+ cost: {
11357
+ input: number;
11358
+ output: number;
11359
+ cacheRead: number;
11360
+ cacheWrite: number;
11361
+ };
11362
+ contextWindow: number;
11363
+ maxTokens: number;
11364
+ };
11365
+ readonly "stepfun/step-3.5-flash": {
11366
+ id: string;
11367
+ name: string;
11368
+ api: "openai-completions";
11369
+ provider: string;
11370
+ baseUrl: string;
11371
+ reasoning: true;
11372
+ input: "text"[];
11373
+ cost: {
11374
+ input: number;
11375
+ output: number;
11376
+ cacheRead: number;
11377
+ cacheWrite: number;
11378
+ };
11379
+ contextWindow: number;
11380
+ maxTokens: number;
11381
+ };
11382
+ readonly "tencent/hy3-preview:free": {
11383
+ id: string;
11384
+ name: string;
11385
+ api: "openai-completions";
11386
+ provider: string;
11387
+ baseUrl: string;
11388
+ reasoning: true;
11389
+ input: "text"[];
11390
+ cost: {
11391
+ input: number;
11392
+ output: number;
11393
+ cacheRead: number;
11394
+ cacheWrite: number;
11395
+ };
11396
+ contextWindow: number;
11397
+ maxTokens: number;
11398
+ };
11399
+ readonly "thedrummer/rocinante-12b": {
11400
+ id: string;
11401
+ name: string;
11402
+ api: "openai-completions";
11403
+ provider: string;
11404
+ baseUrl: string;
11405
+ reasoning: false;
11406
+ input: "text"[];
10630
11407
  cost: {
10631
11408
  input: number;
10632
11409
  output: number;
@@ -10636,14 +11413,14 @@ export declare const MODELS: {
10636
11413
  contextWindow: number;
10637
11414
  maxTokens: number;
10638
11415
  };
10639
- readonly "qwen/qwen3.5-9b": {
11416
+ readonly "thedrummer/unslopnemo-12b": {
10640
11417
  id: string;
10641
11418
  name: string;
10642
11419
  api: "openai-completions";
10643
11420
  provider: string;
10644
11421
  baseUrl: string;
10645
- reasoning: true;
10646
- input: ("image" | "text")[];
11422
+ reasoning: false;
11423
+ input: "text"[];
10647
11424
  cost: {
10648
11425
  input: number;
10649
11426
  output: number;
@@ -10653,14 +11430,14 @@ export declare const MODELS: {
10653
11430
  contextWindow: number;
10654
11431
  maxTokens: number;
10655
11432
  };
10656
- readonly "qwen/qwen3.5-flash-02-23": {
11433
+ readonly "tngtech/deepseek-r1t2-chimera": {
10657
11434
  id: string;
10658
11435
  name: string;
10659
11436
  api: "openai-completions";
10660
11437
  provider: string;
10661
11438
  baseUrl: string;
10662
11439
  reasoning: true;
10663
- input: ("image" | "text")[];
11440
+ input: "text"[];
10664
11441
  cost: {
10665
11442
  input: number;
10666
11443
  output: number;
@@ -10670,14 +11447,14 @@ export declare const MODELS: {
10670
11447
  contextWindow: number;
10671
11448
  maxTokens: number;
10672
11449
  };
10673
- readonly "qwen/qwen3.5-plus-02-15": {
11450
+ readonly "upstage/solar-pro-3": {
10674
11451
  id: string;
10675
11452
  name: string;
10676
11453
  api: "openai-completions";
10677
11454
  provider: string;
10678
11455
  baseUrl: string;
10679
11456
  reasoning: true;
10680
- input: ("image" | "text")[];
11457
+ input: "text"[];
10681
11458
  cost: {
10682
11459
  input: number;
10683
11460
  output: number;
@@ -10687,14 +11464,14 @@ export declare const MODELS: {
10687
11464
  contextWindow: number;
10688
11465
  maxTokens: number;
10689
11466
  };
10690
- readonly "qwen/qwen3.6-plus": {
11467
+ readonly "x-ai/grok-3": {
10691
11468
  id: string;
10692
11469
  name: string;
10693
11470
  api: "openai-completions";
10694
11471
  provider: string;
10695
11472
  baseUrl: string;
10696
- reasoning: true;
10697
- input: ("image" | "text")[];
11473
+ reasoning: false;
11474
+ input: "text"[];
10698
11475
  cost: {
10699
11476
  input: number;
10700
11477
  output: number;
@@ -10704,13 +11481,13 @@ export declare const MODELS: {
10704
11481
  contextWindow: number;
10705
11482
  maxTokens: number;
10706
11483
  };
10707
- readonly "qwen/qwq-32b": {
11484
+ readonly "x-ai/grok-3-beta": {
10708
11485
  id: string;
10709
11486
  name: string;
10710
11487
  api: "openai-completions";
10711
11488
  provider: string;
10712
11489
  baseUrl: string;
10713
- reasoning: true;
11490
+ reasoning: false;
10714
11491
  input: "text"[];
10715
11492
  cost: {
10716
11493
  input: number;
@@ -10721,14 +11498,14 @@ export declare const MODELS: {
10721
11498
  contextWindow: number;
10722
11499
  maxTokens: number;
10723
11500
  };
10724
- readonly "rekaai/reka-edge": {
11501
+ readonly "x-ai/grok-3-mini": {
10725
11502
  id: string;
10726
11503
  name: string;
10727
11504
  api: "openai-completions";
10728
11505
  provider: string;
10729
11506
  baseUrl: string;
10730
- reasoning: false;
10731
- input: ("image" | "text")[];
11507
+ reasoning: true;
11508
+ input: "text"[];
10732
11509
  cost: {
10733
11510
  input: number;
10734
11511
  output: number;
@@ -10738,13 +11515,13 @@ export declare const MODELS: {
10738
11515
  contextWindow: number;
10739
11516
  maxTokens: number;
10740
11517
  };
10741
- readonly "relace/relace-search": {
11518
+ readonly "x-ai/grok-3-mini-beta": {
10742
11519
  id: string;
10743
11520
  name: string;
10744
11521
  api: "openai-completions";
10745
11522
  provider: string;
10746
11523
  baseUrl: string;
10747
- reasoning: false;
11524
+ reasoning: true;
10748
11525
  input: "text"[];
10749
11526
  cost: {
10750
11527
  input: number;
@@ -10755,14 +11532,14 @@ export declare const MODELS: {
10755
11532
  contextWindow: number;
10756
11533
  maxTokens: number;
10757
11534
  };
10758
- readonly "sao10k/l3-euryale-70b": {
11535
+ readonly "x-ai/grok-4": {
10759
11536
  id: string;
10760
11537
  name: string;
10761
11538
  api: "openai-completions";
10762
11539
  provider: string;
10763
11540
  baseUrl: string;
10764
- reasoning: false;
10765
- input: "text"[];
11541
+ reasoning: true;
11542
+ input: ("image" | "text")[];
10766
11543
  cost: {
10767
11544
  input: number;
10768
11545
  output: number;
@@ -10772,14 +11549,14 @@ export declare const MODELS: {
10772
11549
  contextWindow: number;
10773
11550
  maxTokens: number;
10774
11551
  };
10775
- readonly "sao10k/l3.1-euryale-70b": {
11552
+ readonly "x-ai/grok-4-fast": {
10776
11553
  id: string;
10777
11554
  name: string;
10778
11555
  api: "openai-completions";
10779
11556
  provider: string;
10780
11557
  baseUrl: string;
10781
- reasoning: false;
10782
- input: "text"[];
11558
+ reasoning: true;
11559
+ input: ("image" | "text")[];
10783
11560
  cost: {
10784
11561
  input: number;
10785
11562
  output: number;
@@ -10789,14 +11566,14 @@ export declare const MODELS: {
10789
11566
  contextWindow: number;
10790
11567
  maxTokens: number;
10791
11568
  };
10792
- readonly "stepfun/step-3.5-flash": {
11569
+ readonly "x-ai/grok-4.1-fast": {
10793
11570
  id: string;
10794
11571
  name: string;
10795
11572
  api: "openai-completions";
10796
11573
  provider: string;
10797
11574
  baseUrl: string;
10798
11575
  reasoning: true;
10799
- input: "text"[];
11576
+ input: ("image" | "text")[];
10800
11577
  cost: {
10801
11578
  input: number;
10802
11579
  output: number;
@@ -10806,14 +11583,14 @@ export declare const MODELS: {
10806
11583
  contextWindow: number;
10807
11584
  maxTokens: number;
10808
11585
  };
10809
- readonly "thedrummer/rocinante-12b": {
11586
+ readonly "x-ai/grok-4.20": {
10810
11587
  id: string;
10811
11588
  name: string;
10812
11589
  api: "openai-completions";
10813
11590
  provider: string;
10814
11591
  baseUrl: string;
10815
- reasoning: false;
10816
- input: "text"[];
11592
+ reasoning: true;
11593
+ input: ("image" | "text")[];
10817
11594
  cost: {
10818
11595
  input: number;
10819
11596
  output: number;
@@ -10823,14 +11600,14 @@ export declare const MODELS: {
10823
11600
  contextWindow: number;
10824
11601
  maxTokens: number;
10825
11602
  };
10826
- readonly "thedrummer/unslopnemo-12b": {
11603
+ readonly "x-ai/grok-4.3": {
10827
11604
  id: string;
10828
11605
  name: string;
10829
11606
  api: "openai-completions";
10830
11607
  provider: string;
10831
11608
  baseUrl: string;
10832
- reasoning: false;
10833
- input: "text"[];
11609
+ reasoning: true;
11610
+ input: ("image" | "text")[];
10834
11611
  cost: {
10835
11612
  input: number;
10836
11613
  output: number;
@@ -10840,7 +11617,7 @@ export declare const MODELS: {
10840
11617
  contextWindow: number;
10841
11618
  maxTokens: number;
10842
11619
  };
10843
- readonly "tngtech/deepseek-r1t2-chimera": {
11620
+ readonly "x-ai/grok-code-fast-1": {
10844
11621
  id: string;
10845
11622
  name: string;
10846
11623
  api: "openai-completions";
@@ -10857,7 +11634,7 @@ export declare const MODELS: {
10857
11634
  contextWindow: number;
10858
11635
  maxTokens: number;
10859
11636
  };
10860
- readonly "upstage/solar-pro-3": {
11637
+ readonly "xiaomi/mimo-v2-flash": {
10861
11638
  id: string;
10862
11639
  name: string;
10863
11640
  api: "openai-completions";
@@ -10874,14 +11651,14 @@ export declare const MODELS: {
10874
11651
  contextWindow: number;
10875
11652
  maxTokens: number;
10876
11653
  };
10877
- readonly "x-ai/grok-3": {
11654
+ readonly "xiaomi/mimo-v2-omni": {
10878
11655
  id: string;
10879
11656
  name: string;
10880
11657
  api: "openai-completions";
10881
11658
  provider: string;
10882
11659
  baseUrl: string;
10883
- reasoning: false;
10884
- input: "text"[];
11660
+ reasoning: true;
11661
+ input: ("image" | "text")[];
10885
11662
  cost: {
10886
11663
  input: number;
10887
11664
  output: number;
@@ -10891,13 +11668,13 @@ export declare const MODELS: {
10891
11668
  contextWindow: number;
10892
11669
  maxTokens: number;
10893
11670
  };
10894
- readonly "x-ai/grok-3-beta": {
11671
+ readonly "xiaomi/mimo-v2-pro": {
10895
11672
  id: string;
10896
11673
  name: string;
10897
11674
  api: "openai-completions";
10898
11675
  provider: string;
10899
11676
  baseUrl: string;
10900
- reasoning: false;
11677
+ reasoning: true;
10901
11678
  input: "text"[];
10902
11679
  cost: {
10903
11680
  input: number;
@@ -10908,14 +11685,14 @@ export declare const MODELS: {
10908
11685
  contextWindow: number;
10909
11686
  maxTokens: number;
10910
11687
  };
10911
- readonly "x-ai/grok-3-mini": {
11688
+ readonly "xiaomi/mimo-v2.5": {
10912
11689
  id: string;
10913
11690
  name: string;
10914
11691
  api: "openai-completions";
10915
11692
  provider: string;
10916
11693
  baseUrl: string;
10917
11694
  reasoning: true;
10918
- input: "text"[];
11695
+ input: ("image" | "text")[];
10919
11696
  cost: {
10920
11697
  input: number;
10921
11698
  output: number;
@@ -10925,7 +11702,7 @@ export declare const MODELS: {
10925
11702
  contextWindow: number;
10926
11703
  maxTokens: number;
10927
11704
  };
10928
- readonly "x-ai/grok-3-mini-beta": {
11705
+ readonly "xiaomi/mimo-v2.5-pro": {
10929
11706
  id: string;
10930
11707
  name: string;
10931
11708
  api: "openai-completions";
@@ -10942,14 +11719,14 @@ export declare const MODELS: {
10942
11719
  contextWindow: number;
10943
11720
  maxTokens: number;
10944
11721
  };
10945
- readonly "x-ai/grok-4": {
11722
+ readonly "z-ai/glm-4-32b": {
10946
11723
  id: string;
10947
11724
  name: string;
10948
11725
  api: "openai-completions";
10949
11726
  provider: string;
10950
11727
  baseUrl: string;
10951
- reasoning: true;
10952
- input: ("image" | "text")[];
11728
+ reasoning: false;
11729
+ input: "text"[];
10953
11730
  cost: {
10954
11731
  input: number;
10955
11732
  output: number;
@@ -10959,14 +11736,14 @@ export declare const MODELS: {
10959
11736
  contextWindow: number;
10960
11737
  maxTokens: number;
10961
11738
  };
10962
- readonly "x-ai/grok-4-fast": {
11739
+ readonly "z-ai/glm-4.5": {
10963
11740
  id: string;
10964
11741
  name: string;
10965
11742
  api: "openai-completions";
10966
11743
  provider: string;
10967
11744
  baseUrl: string;
10968
11745
  reasoning: true;
10969
- input: ("image" | "text")[];
11746
+ input: "text"[];
10970
11747
  cost: {
10971
11748
  input: number;
10972
11749
  output: number;
@@ -10976,14 +11753,14 @@ export declare const MODELS: {
10976
11753
  contextWindow: number;
10977
11754
  maxTokens: number;
10978
11755
  };
10979
- readonly "x-ai/grok-4.1-fast": {
11756
+ readonly "z-ai/glm-4.5-air": {
10980
11757
  id: string;
10981
11758
  name: string;
10982
11759
  api: "openai-completions";
10983
11760
  provider: string;
10984
11761
  baseUrl: string;
10985
11762
  reasoning: true;
10986
- input: ("image" | "text")[];
11763
+ input: "text"[];
10987
11764
  cost: {
10988
11765
  input: number;
10989
11766
  output: number;
@@ -10993,14 +11770,14 @@ export declare const MODELS: {
10993
11770
  contextWindow: number;
10994
11771
  maxTokens: number;
10995
11772
  };
10996
- readonly "x-ai/grok-4.20": {
11773
+ readonly "z-ai/glm-4.5-air:free": {
10997
11774
  id: string;
10998
11775
  name: string;
10999
11776
  api: "openai-completions";
11000
11777
  provider: string;
11001
11778
  baseUrl: string;
11002
11779
  reasoning: true;
11003
- input: ("image" | "text")[];
11780
+ input: "text"[];
11004
11781
  cost: {
11005
11782
  input: number;
11006
11783
  output: number;
@@ -11010,14 +11787,14 @@ export declare const MODELS: {
11010
11787
  contextWindow: number;
11011
11788
  maxTokens: number;
11012
11789
  };
11013
- readonly "x-ai/grok-code-fast-1": {
11790
+ readonly "z-ai/glm-4.5v": {
11014
11791
  id: string;
11015
11792
  name: string;
11016
11793
  api: "openai-completions";
11017
11794
  provider: string;
11018
11795
  baseUrl: string;
11019
11796
  reasoning: true;
11020
- input: "text"[];
11797
+ input: ("image" | "text")[];
11021
11798
  cost: {
11022
11799
  input: number;
11023
11800
  output: number;
@@ -11027,7 +11804,7 @@ export declare const MODELS: {
11027
11804
  contextWindow: number;
11028
11805
  maxTokens: number;
11029
11806
  };
11030
- readonly "xiaomi/mimo-v2-flash": {
11807
+ readonly "z-ai/glm-4.6": {
11031
11808
  id: string;
11032
11809
  name: string;
11033
11810
  api: "openai-completions";
@@ -11044,7 +11821,7 @@ export declare const MODELS: {
11044
11821
  contextWindow: number;
11045
11822
  maxTokens: number;
11046
11823
  };
11047
- readonly "xiaomi/mimo-v2-omni": {
11824
+ readonly "z-ai/glm-4.6v": {
11048
11825
  id: string;
11049
11826
  name: string;
11050
11827
  api: "openai-completions";
@@ -11061,7 +11838,7 @@ export declare const MODELS: {
11061
11838
  contextWindow: number;
11062
11839
  maxTokens: number;
11063
11840
  };
11064
- readonly "xiaomi/mimo-v2-pro": {
11841
+ readonly "z-ai/glm-4.7": {
11065
11842
  id: string;
11066
11843
  name: string;
11067
11844
  api: "openai-completions";
@@ -11078,13 +11855,13 @@ export declare const MODELS: {
11078
11855
  contextWindow: number;
11079
11856
  maxTokens: number;
11080
11857
  };
11081
- readonly "z-ai/glm-4-32b": {
11858
+ readonly "z-ai/glm-4.7-flash": {
11082
11859
  id: string;
11083
11860
  name: string;
11084
11861
  api: "openai-completions";
11085
11862
  provider: string;
11086
11863
  baseUrl: string;
11087
- reasoning: false;
11864
+ reasoning: true;
11088
11865
  input: "text"[];
11089
11866
  cost: {
11090
11867
  input: number;
@@ -11095,7 +11872,7 @@ export declare const MODELS: {
11095
11872
  contextWindow: number;
11096
11873
  maxTokens: number;
11097
11874
  };
11098
- readonly "z-ai/glm-4.5": {
11875
+ readonly "z-ai/glm-5": {
11099
11876
  id: string;
11100
11877
  name: string;
11101
11878
  api: "openai-completions";
@@ -11112,7 +11889,7 @@ export declare const MODELS: {
11112
11889
  contextWindow: number;
11113
11890
  maxTokens: number;
11114
11891
  };
11115
- readonly "z-ai/glm-4.5-air": {
11892
+ readonly "z-ai/glm-5-turbo": {
11116
11893
  id: string;
11117
11894
  name: string;
11118
11895
  api: "openai-completions";
@@ -11129,7 +11906,7 @@ export declare const MODELS: {
11129
11906
  contextWindow: number;
11130
11907
  maxTokens: number;
11131
11908
  };
11132
- readonly "z-ai/glm-4.5-air:free": {
11909
+ readonly "z-ai/glm-5.1": {
11133
11910
  id: string;
11134
11911
  name: string;
11135
11912
  api: "openai-completions";
@@ -11146,7 +11923,7 @@ export declare const MODELS: {
11146
11923
  contextWindow: number;
11147
11924
  maxTokens: number;
11148
11925
  };
11149
- readonly "z-ai/glm-4.5v": {
11926
+ readonly "z-ai/glm-5v-turbo": {
11150
11927
  id: string;
11151
11928
  name: string;
11152
11929
  api: "openai-completions";
@@ -11163,14 +11940,14 @@ export declare const MODELS: {
11163
11940
  contextWindow: number;
11164
11941
  maxTokens: number;
11165
11942
  };
11166
- readonly "z-ai/glm-4.6": {
11943
+ readonly "~anthropic/claude-haiku-latest": {
11167
11944
  id: string;
11168
11945
  name: string;
11169
11946
  api: "openai-completions";
11170
11947
  provider: string;
11171
11948
  baseUrl: string;
11172
11949
  reasoning: true;
11173
- input: "text"[];
11950
+ input: ("image" | "text")[];
11174
11951
  cost: {
11175
11952
  input: number;
11176
11953
  output: number;
@@ -11180,7 +11957,7 @@ export declare const MODELS: {
11180
11957
  contextWindow: number;
11181
11958
  maxTokens: number;
11182
11959
  };
11183
- readonly "z-ai/glm-4.6v": {
11960
+ readonly "~anthropic/claude-opus-latest": {
11184
11961
  id: string;
11185
11962
  name: string;
11186
11963
  api: "openai-completions";
@@ -11197,14 +11974,14 @@ export declare const MODELS: {
11197
11974
  contextWindow: number;
11198
11975
  maxTokens: number;
11199
11976
  };
11200
- readonly "z-ai/glm-4.7": {
11977
+ readonly "~anthropic/claude-sonnet-latest": {
11201
11978
  id: string;
11202
11979
  name: string;
11203
11980
  api: "openai-completions";
11204
11981
  provider: string;
11205
11982
  baseUrl: string;
11206
11983
  reasoning: true;
11207
- input: "text"[];
11984
+ input: ("image" | "text")[];
11208
11985
  cost: {
11209
11986
  input: number;
11210
11987
  output: number;
@@ -11214,14 +11991,14 @@ export declare const MODELS: {
11214
11991
  contextWindow: number;
11215
11992
  maxTokens: number;
11216
11993
  };
11217
- readonly "z-ai/glm-4.7-flash": {
11994
+ readonly "~google/gemini-flash-latest": {
11218
11995
  id: string;
11219
11996
  name: string;
11220
11997
  api: "openai-completions";
11221
11998
  provider: string;
11222
11999
  baseUrl: string;
11223
12000
  reasoning: true;
11224
- input: "text"[];
12001
+ input: ("image" | "text")[];
11225
12002
  cost: {
11226
12003
  input: number;
11227
12004
  output: number;
@@ -11231,14 +12008,14 @@ export declare const MODELS: {
11231
12008
  contextWindow: number;
11232
12009
  maxTokens: number;
11233
12010
  };
11234
- readonly "z-ai/glm-5": {
12011
+ readonly "~google/gemini-pro-latest": {
11235
12012
  id: string;
11236
12013
  name: string;
11237
12014
  api: "openai-completions";
11238
12015
  provider: string;
11239
12016
  baseUrl: string;
11240
12017
  reasoning: true;
11241
- input: "text"[];
12018
+ input: ("image" | "text")[];
11242
12019
  cost: {
11243
12020
  input: number;
11244
12021
  output: number;
@@ -11248,14 +12025,14 @@ export declare const MODELS: {
11248
12025
  contextWindow: number;
11249
12026
  maxTokens: number;
11250
12027
  };
11251
- readonly "z-ai/glm-5-turbo": {
12028
+ readonly "~moonshotai/kimi-latest": {
11252
12029
  id: string;
11253
12030
  name: string;
11254
12031
  api: "openai-completions";
11255
12032
  provider: string;
11256
12033
  baseUrl: string;
11257
12034
  reasoning: true;
11258
- input: "text"[];
12035
+ input: ("image" | "text")[];
11259
12036
  cost: {
11260
12037
  input: number;
11261
12038
  output: number;
@@ -11265,14 +12042,14 @@ export declare const MODELS: {
11265
12042
  contextWindow: number;
11266
12043
  maxTokens: number;
11267
12044
  };
11268
- readonly "z-ai/glm-5.1": {
12045
+ readonly "~openai/gpt-latest": {
11269
12046
  id: string;
11270
12047
  name: string;
11271
12048
  api: "openai-completions";
11272
12049
  provider: string;
11273
12050
  baseUrl: string;
11274
12051
  reasoning: true;
11275
- input: "text"[];
12052
+ input: ("image" | "text")[];
11276
12053
  cost: {
11277
12054
  input: number;
11278
12055
  output: number;
@@ -11282,7 +12059,7 @@ export declare const MODELS: {
11282
12059
  contextWindow: number;
11283
12060
  maxTokens: number;
11284
12061
  };
11285
- readonly "z-ai/glm-5v-turbo": {
12062
+ readonly "~openai/gpt-mini-latest": {
11286
12063
  id: string;
11287
12064
  name: string;
11288
12065
  api: "openai-completions";
@@ -11369,6 +12146,23 @@ export declare const MODELS: {
11369
12146
  contextWindow: number;
11370
12147
  maxTokens: number;
11371
12148
  };
12149
+ readonly "alibaba/qwen-3.6-max-preview": {
12150
+ id: string;
12151
+ name: string;
12152
+ api: "anthropic-messages";
12153
+ provider: string;
12154
+ baseUrl: string;
12155
+ reasoning: true;
12156
+ input: ("image" | "text")[];
12157
+ cost: {
12158
+ input: number;
12159
+ output: number;
12160
+ cacheRead: number;
12161
+ cacheWrite: number;
12162
+ };
12163
+ contextWindow: number;
12164
+ maxTokens: number;
12165
+ };
11372
12166
  readonly "alibaba/qwen3-235b-a22b-thinking": {
11373
12167
  id: string;
11374
12168
  name: string;
@@ -11556,6 +12350,23 @@ export declare const MODELS: {
11556
12350
  contextWindow: number;
11557
12351
  maxTokens: number;
11558
12352
  };
12353
+ readonly "alibaba/qwen3.6-27b": {
12354
+ id: string;
12355
+ name: string;
12356
+ api: "anthropic-messages";
12357
+ provider: string;
12358
+ baseUrl: string;
12359
+ reasoning: true;
12360
+ input: ("image" | "text")[];
12361
+ cost: {
12362
+ input: number;
12363
+ output: number;
12364
+ cacheRead: number;
12365
+ cacheWrite: number;
12366
+ };
12367
+ contextWindow: number;
12368
+ maxTokens: number;
12369
+ };
11559
12370
  readonly "alibaba/qwen3.6-plus": {
11560
12371
  id: string;
11561
12372
  name: string;
@@ -11931,6 +12742,40 @@ export declare const MODELS: {
11931
12742
  maxTokens: number;
11932
12743
  };
11933
12744
  readonly "deepseek/deepseek-v3.2-thinking": {
12745
+ id: string;
12746
+ name: string;
12747
+ api: "anthropic-messages";
12748
+ provider: string;
12749
+ baseUrl: string;
12750
+ reasoning: false;
12751
+ input: "text"[];
12752
+ cost: {
12753
+ input: number;
12754
+ output: number;
12755
+ cacheRead: number;
12756
+ cacheWrite: number;
12757
+ };
12758
+ contextWindow: number;
12759
+ maxTokens: number;
12760
+ };
12761
+ readonly "deepseek/deepseek-v4-flash": {
12762
+ id: string;
12763
+ name: string;
12764
+ api: "anthropic-messages";
12765
+ provider: string;
12766
+ baseUrl: string;
12767
+ reasoning: true;
12768
+ input: "text"[];
12769
+ cost: {
12770
+ input: number;
12771
+ output: number;
12772
+ cacheRead: number;
12773
+ cacheWrite: number;
12774
+ };
12775
+ contextWindow: number;
12776
+ maxTokens: number;
12777
+ };
12778
+ readonly "deepseek/deepseek-v4-pro": {
11934
12779
  id: string;
11935
12780
  name: string;
11936
12781
  api: "anthropic-messages";
@@ -12712,6 +13557,23 @@ export declare const MODELS: {
12712
13557
  contextWindow: number;
12713
13558
  maxTokens: number;
12714
13559
  };
13560
+ readonly "moonshotai/kimi-k2.6": {
13561
+ id: string;
13562
+ name: string;
13563
+ api: "anthropic-messages";
13564
+ provider: string;
13565
+ baseUrl: string;
13566
+ reasoning: true;
13567
+ input: ("image" | "text")[];
13568
+ cost: {
13569
+ input: number;
13570
+ output: number;
13571
+ cacheRead: number;
13572
+ cacheWrite: number;
13573
+ };
13574
+ contextWindow: number;
13575
+ maxTokens: number;
13576
+ };
12715
13577
  readonly "nvidia/nemotron-nano-12b-v2-vl": {
12716
13578
  id: string;
12717
13579
  name: string;
@@ -13001,7 +13863,41 @@ export declare const MODELS: {
13001
13863
  contextWindow: number;
13002
13864
  maxTokens: number;
13003
13865
  };
13004
- readonly "openai/gpt-5.1-instant": {
13866
+ readonly "openai/gpt-5.1-instant": {
13867
+ id: string;
13868
+ name: string;
13869
+ api: "anthropic-messages";
13870
+ provider: string;
13871
+ baseUrl: string;
13872
+ reasoning: true;
13873
+ input: ("image" | "text")[];
13874
+ cost: {
13875
+ input: number;
13876
+ output: number;
13877
+ cacheRead: number;
13878
+ cacheWrite: number;
13879
+ };
13880
+ contextWindow: number;
13881
+ maxTokens: number;
13882
+ };
13883
+ readonly "openai/gpt-5.1-thinking": {
13884
+ id: string;
13885
+ name: string;
13886
+ api: "anthropic-messages";
13887
+ provider: string;
13888
+ baseUrl: string;
13889
+ reasoning: true;
13890
+ input: ("image" | "text")[];
13891
+ cost: {
13892
+ input: number;
13893
+ output: number;
13894
+ cacheRead: number;
13895
+ cacheWrite: number;
13896
+ };
13897
+ contextWindow: number;
13898
+ maxTokens: number;
13899
+ };
13900
+ readonly "openai/gpt-5.2": {
13005
13901
  id: string;
13006
13902
  name: string;
13007
13903
  api: "anthropic-messages";
@@ -13018,7 +13914,7 @@ export declare const MODELS: {
13018
13914
  contextWindow: number;
13019
13915
  maxTokens: number;
13020
13916
  };
13021
- readonly "openai/gpt-5.1-thinking": {
13917
+ readonly "openai/gpt-5.2-chat": {
13022
13918
  id: string;
13023
13919
  name: string;
13024
13920
  api: "anthropic-messages";
@@ -13035,7 +13931,7 @@ export declare const MODELS: {
13035
13931
  contextWindow: number;
13036
13932
  maxTokens: number;
13037
13933
  };
13038
- readonly "openai/gpt-5.2": {
13934
+ readonly "openai/gpt-5.2-codex": {
13039
13935
  id: string;
13040
13936
  name: string;
13041
13937
  api: "anthropic-messages";
@@ -13052,7 +13948,7 @@ export declare const MODELS: {
13052
13948
  contextWindow: number;
13053
13949
  maxTokens: number;
13054
13950
  };
13055
- readonly "openai/gpt-5.2-chat": {
13951
+ readonly "openai/gpt-5.2-pro": {
13056
13952
  id: string;
13057
13953
  name: string;
13058
13954
  api: "anthropic-messages";
@@ -13069,7 +13965,7 @@ export declare const MODELS: {
13069
13965
  contextWindow: number;
13070
13966
  maxTokens: number;
13071
13967
  };
13072
- readonly "openai/gpt-5.2-codex": {
13968
+ readonly "openai/gpt-5.3-chat": {
13073
13969
  id: string;
13074
13970
  name: string;
13075
13971
  api: "anthropic-messages";
@@ -13086,7 +13982,7 @@ export declare const MODELS: {
13086
13982
  contextWindow: number;
13087
13983
  maxTokens: number;
13088
13984
  };
13089
- readonly "openai/gpt-5.2-pro": {
13985
+ readonly "openai/gpt-5.3-codex": {
13090
13986
  id: string;
13091
13987
  name: string;
13092
13988
  api: "anthropic-messages";
@@ -13103,7 +13999,7 @@ export declare const MODELS: {
13103
13999
  contextWindow: number;
13104
14000
  maxTokens: number;
13105
14001
  };
13106
- readonly "openai/gpt-5.3-chat": {
14002
+ readonly "openai/gpt-5.4": {
13107
14003
  id: string;
13108
14004
  name: string;
13109
14005
  api: "anthropic-messages";
@@ -13120,7 +14016,7 @@ export declare const MODELS: {
13120
14016
  contextWindow: number;
13121
14017
  maxTokens: number;
13122
14018
  };
13123
- readonly "openai/gpt-5.3-codex": {
14019
+ readonly "openai/gpt-5.4-mini": {
13124
14020
  id: string;
13125
14021
  name: string;
13126
14022
  api: "anthropic-messages";
@@ -13137,7 +14033,7 @@ export declare const MODELS: {
13137
14033
  contextWindow: number;
13138
14034
  maxTokens: number;
13139
14035
  };
13140
- readonly "openai/gpt-5.4": {
14036
+ readonly "openai/gpt-5.4-nano": {
13141
14037
  id: string;
13142
14038
  name: string;
13143
14039
  api: "anthropic-messages";
@@ -13154,7 +14050,7 @@ export declare const MODELS: {
13154
14050
  contextWindow: number;
13155
14051
  maxTokens: number;
13156
14052
  };
13157
- readonly "openai/gpt-5.4-mini": {
14053
+ readonly "openai/gpt-5.4-pro": {
13158
14054
  id: string;
13159
14055
  name: string;
13160
14056
  api: "anthropic-messages";
@@ -13171,7 +14067,7 @@ export declare const MODELS: {
13171
14067
  contextWindow: number;
13172
14068
  maxTokens: number;
13173
14069
  };
13174
- readonly "openai/gpt-5.4-nano": {
14070
+ readonly "openai/gpt-5.5": {
13175
14071
  id: string;
13176
14072
  name: string;
13177
14073
  api: "anthropic-messages";
@@ -13188,7 +14084,7 @@ export declare const MODELS: {
13188
14084
  contextWindow: number;
13189
14085
  maxTokens: number;
13190
14086
  };
13191
- readonly "openai/gpt-5.4-pro": {
14087
+ readonly "openai/gpt-5.5-pro": {
13192
14088
  id: string;
13193
14089
  name: string;
13194
14090
  api: "anthropic-messages";
@@ -13375,23 +14271,6 @@ export declare const MODELS: {
13375
14271
  contextWindow: number;
13376
14272
  maxTokens: number;
13377
14273
  };
13378
- readonly "prime-intellect/intellect-3": {
13379
- id: string;
13380
- name: string;
13381
- api: "anthropic-messages";
13382
- provider: string;
13383
- baseUrl: string;
13384
- reasoning: true;
13385
- input: "text"[];
13386
- cost: {
13387
- input: number;
13388
- output: number;
13389
- cacheRead: number;
13390
- cacheWrite: number;
13391
- };
13392
- contextWindow: number;
13393
- maxTokens: number;
13394
- };
13395
14274
  readonly "xai/grok-3": {
13396
14275
  id: string;
13397
14276
  name: string;
@@ -13647,6 +14526,23 @@ export declare const MODELS: {
13647
14526
  contextWindow: number;
13648
14527
  maxTokens: number;
13649
14528
  };
14529
+ readonly "xai/grok-4.3": {
14530
+ id: string;
14531
+ name: string;
14532
+ api: "anthropic-messages";
14533
+ provider: string;
14534
+ baseUrl: string;
14535
+ reasoning: true;
14536
+ input: ("image" | "text")[];
14537
+ cost: {
14538
+ input: number;
14539
+ output: number;
14540
+ cacheRead: number;
14541
+ cacheWrite: number;
14542
+ };
14543
+ contextWindow: number;
14544
+ maxTokens: number;
14545
+ };
13650
14546
  readonly "xai/grok-code-fast-1": {
13651
14547
  id: string;
13652
14548
  name: string;
@@ -13698,6 +14594,40 @@ export declare const MODELS: {
13698
14594
  contextWindow: number;
13699
14595
  maxTokens: number;
13700
14596
  };
14597
+ readonly "xiaomi/mimo-v2.5": {
14598
+ id: string;
14599
+ name: string;
14600
+ api: "anthropic-messages";
14601
+ provider: string;
14602
+ baseUrl: string;
14603
+ reasoning: true;
14604
+ input: ("image" | "text")[];
14605
+ cost: {
14606
+ input: number;
14607
+ output: number;
14608
+ cacheRead: number;
14609
+ cacheWrite: number;
14610
+ };
14611
+ contextWindow: number;
14612
+ maxTokens: number;
14613
+ };
14614
+ readonly "xiaomi/mimo-v2.5-pro": {
14615
+ id: string;
14616
+ name: string;
14617
+ api: "anthropic-messages";
14618
+ provider: string;
14619
+ baseUrl: string;
14620
+ reasoning: true;
14621
+ input: ("image" | "text")[];
14622
+ cost: {
14623
+ input: number;
14624
+ output: number;
14625
+ cacheRead: number;
14626
+ cacheWrite: number;
14627
+ };
14628
+ contextWindow: number;
14629
+ maxTokens: number;
14630
+ };
13701
14631
  readonly "zai/glm-4.5": {
13702
14632
  id: string;
13703
14633
  name: string;
@@ -13892,7 +14822,7 @@ export declare const MODELS: {
13892
14822
  provider: string;
13893
14823
  baseUrl: string;
13894
14824
  reasoning: true;
13895
- input: ("image" | "text")[];
14825
+ input: "text"[];
13896
14826
  cost: {
13897
14827
  input: number;
13898
14828
  output: number;
@@ -14278,47 +15208,13 @@ export declare const MODELS: {
14278
15208
  contextWindow: number;
14279
15209
  maxTokens: number;
14280
15210
  };
14281
- readonly "grok-beta": {
14282
- id: string;
14283
- name: string;
14284
- api: "openai-completions";
14285
- provider: string;
14286
- baseUrl: string;
14287
- reasoning: false;
14288
- input: "text"[];
14289
- cost: {
14290
- input: number;
14291
- output: number;
14292
- cacheRead: number;
14293
- cacheWrite: number;
14294
- };
14295
- contextWindow: number;
14296
- maxTokens: number;
14297
- };
14298
- readonly "grok-code-fast-1": {
15211
+ readonly "grok-4.3": {
14299
15212
  id: string;
14300
15213
  name: string;
14301
15214
  api: "openai-completions";
14302
15215
  provider: string;
14303
15216
  baseUrl: string;
14304
15217
  reasoning: true;
14305
- input: "text"[];
14306
- cost: {
14307
- input: number;
14308
- output: number;
14309
- cacheRead: number;
14310
- cacheWrite: number;
14311
- };
14312
- contextWindow: number;
14313
- maxTokens: number;
14314
- };
14315
- readonly "grok-vision-beta": {
14316
- id: string;
14317
- name: string;
14318
- api: "openai-completions";
14319
- provider: string;
14320
- baseUrl: string;
14321
- reasoning: false;
14322
15218
  input: ("image" | "text")[];
14323
15219
  cost: {
14324
15220
  input: number;
@@ -14329,40 +15225,13 @@ export declare const MODELS: {
14329
15225
  contextWindow: number;
14330
15226
  maxTokens: number;
14331
15227
  };
14332
- };
14333
- readonly zai: {
14334
- readonly "glm-4.5": {
14335
- id: string;
14336
- name: string;
14337
- api: "openai-completions";
14338
- provider: string;
14339
- baseUrl: string;
14340
- compat: {
14341
- supportsDeveloperRole: false;
14342
- thinkingFormat: "zai";
14343
- };
14344
- reasoning: true;
14345
- input: "text"[];
14346
- cost: {
14347
- input: number;
14348
- output: number;
14349
- cacheRead: number;
14350
- cacheWrite: number;
14351
- };
14352
- contextWindow: number;
14353
- maxTokens: number;
14354
- };
14355
- readonly "glm-4.5-air": {
15228
+ readonly "grok-beta": {
14356
15229
  id: string;
14357
15230
  name: string;
14358
15231
  api: "openai-completions";
14359
15232
  provider: string;
14360
15233
  baseUrl: string;
14361
- compat: {
14362
- supportsDeveloperRole: false;
14363
- thinkingFormat: "zai";
14364
- };
14365
- reasoning: true;
15234
+ reasoning: false;
14366
15235
  input: "text"[];
14367
15236
  cost: {
14368
15237
  input: number;
@@ -14373,16 +15242,12 @@ export declare const MODELS: {
14373
15242
  contextWindow: number;
14374
15243
  maxTokens: number;
14375
15244
  };
14376
- readonly "glm-4.5-flash": {
15245
+ readonly "grok-code-fast-1": {
14377
15246
  id: string;
14378
15247
  name: string;
14379
15248
  api: "openai-completions";
14380
15249
  provider: string;
14381
15250
  baseUrl: string;
14382
- compat: {
14383
- supportsDeveloperRole: false;
14384
- thinkingFormat: "zai";
14385
- };
14386
15251
  reasoning: true;
14387
15252
  input: "text"[];
14388
15253
  cost: {
@@ -14394,17 +15259,13 @@ export declare const MODELS: {
14394
15259
  contextWindow: number;
14395
15260
  maxTokens: number;
14396
15261
  };
14397
- readonly "glm-4.5v": {
15262
+ readonly "grok-vision-beta": {
14398
15263
  id: string;
14399
15264
  name: string;
14400
15265
  api: "openai-completions";
14401
15266
  provider: string;
14402
15267
  baseUrl: string;
14403
- compat: {
14404
- supportsDeveloperRole: false;
14405
- thinkingFormat: "zai";
14406
- };
14407
- reasoning: true;
15268
+ reasoning: false;
14408
15269
  input: ("image" | "text")[];
14409
15270
  cost: {
14410
15271
  input: number;
@@ -14415,7 +15276,9 @@ export declare const MODELS: {
14415
15276
  contextWindow: number;
14416
15277
  maxTokens: number;
14417
15278
  };
14418
- readonly "glm-4.6": {
15279
+ };
15280
+ readonly zai: {
15281
+ readonly "glm-4.5-air": {
14419
15282
  id: string;
14420
15283
  name: string;
14421
15284
  api: "openai-completions";
@@ -14424,7 +15287,6 @@ export declare const MODELS: {
14424
15287
  compat: {
14425
15288
  supportsDeveloperRole: false;
14426
15289
  thinkingFormat: "zai";
14427
- zaiToolStream: true;
14428
15290
  };
14429
15291
  reasoning: true;
14430
15292
  input: "text"[];
@@ -14437,28 +15299,6 @@ export declare const MODELS: {
14437
15299
  contextWindow: number;
14438
15300
  maxTokens: number;
14439
15301
  };
14440
- readonly "glm-4.6v": {
14441
- id: string;
14442
- name: string;
14443
- api: "openai-completions";
14444
- provider: string;
14445
- baseUrl: string;
14446
- compat: {
14447
- supportsDeveloperRole: false;
14448
- thinkingFormat: "zai";
14449
- zaiToolStream: true;
14450
- };
14451
- reasoning: true;
14452
- input: ("image" | "text")[];
14453
- cost: {
14454
- input: number;
14455
- output: number;
14456
- cacheRead: number;
14457
- cacheWrite: number;
14458
- };
14459
- contextWindow: number;
14460
- maxTokens: number;
14461
- };
14462
15302
  readonly "glm-4.7": {
14463
15303
  id: string;
14464
15304
  name: string;
@@ -14481,72 +15321,6 @@ export declare const MODELS: {
14481
15321
  contextWindow: number;
14482
15322
  maxTokens: number;
14483
15323
  };
14484
- readonly "glm-4.7-flash": {
14485
- id: string;
14486
- name: string;
14487
- api: "openai-completions";
14488
- provider: string;
14489
- baseUrl: string;
14490
- compat: {
14491
- supportsDeveloperRole: false;
14492
- thinkingFormat: "zai";
14493
- zaiToolStream: true;
14494
- };
14495
- reasoning: true;
14496
- input: "text"[];
14497
- cost: {
14498
- input: number;
14499
- output: number;
14500
- cacheRead: number;
14501
- cacheWrite: number;
14502
- };
14503
- contextWindow: number;
14504
- maxTokens: number;
14505
- };
14506
- readonly "glm-4.7-flashx": {
14507
- id: string;
14508
- name: string;
14509
- api: "openai-completions";
14510
- provider: string;
14511
- baseUrl: string;
14512
- compat: {
14513
- supportsDeveloperRole: false;
14514
- thinkingFormat: "zai";
14515
- zaiToolStream: true;
14516
- };
14517
- reasoning: true;
14518
- input: "text"[];
14519
- cost: {
14520
- input: number;
14521
- output: number;
14522
- cacheRead: number;
14523
- cacheWrite: number;
14524
- };
14525
- contextWindow: number;
14526
- maxTokens: number;
14527
- };
14528
- readonly "glm-5": {
14529
- id: string;
14530
- name: string;
14531
- api: "openai-completions";
14532
- provider: string;
14533
- baseUrl: string;
14534
- compat: {
14535
- supportsDeveloperRole: false;
14536
- thinkingFormat: "zai";
14537
- zaiToolStream: true;
14538
- };
14539
- reasoning: true;
14540
- input: "text"[];
14541
- cost: {
14542
- input: number;
14543
- output: number;
14544
- cacheRead: number;
14545
- cacheWrite: number;
14546
- };
14547
- contextWindow: number;
14548
- maxTokens: number;
14549
- };
14550
15324
  readonly "glm-5-turbo": {
14551
15325
  id: string;
14552
15326
  name: string;