@avadisabelle/ava-pi-ai 0.65.0 → 0.65.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -323,6 +323,40 @@ export declare const MODELS: {
323
323
  contextWindow: number;
324
324
  maxTokens: number;
325
325
  };
326
+ readonly "au.anthropic.claude-opus-4-6-v1": {
327
+ id: string;
328
+ name: string;
329
+ api: "bedrock-converse-stream";
330
+ provider: string;
331
+ baseUrl: string;
332
+ reasoning: true;
333
+ input: ("image" | "text")[];
334
+ cost: {
335
+ input: number;
336
+ output: number;
337
+ cacheRead: number;
338
+ cacheWrite: number;
339
+ };
340
+ contextWindow: number;
341
+ maxTokens: number;
342
+ };
343
+ readonly "au.anthropic.claude-sonnet-4-6": {
344
+ id: string;
345
+ name: string;
346
+ api: "bedrock-converse-stream";
347
+ provider: string;
348
+ baseUrl: string;
349
+ reasoning: true;
350
+ input: ("image" | "text")[];
351
+ cost: {
352
+ input: number;
353
+ output: number;
354
+ cacheRead: number;
355
+ cacheWrite: number;
356
+ };
357
+ contextWindow: number;
358
+ maxTokens: number;
359
+ };
326
360
  readonly "deepseek.r1-v1:0": {
327
361
  id: string;
328
362
  name: string;
@@ -2486,6 +2520,23 @@ export declare const MODELS: {
2486
2520
  contextWindow: number;
2487
2521
  maxTokens: number;
2488
2522
  };
2523
+ readonly "gpt-5.5": {
2524
+ id: string;
2525
+ name: string;
2526
+ api: "azure-openai-responses";
2527
+ provider: string;
2528
+ baseUrl: string;
2529
+ reasoning: true;
2530
+ input: ("image" | "text")[];
2531
+ cost: {
2532
+ input: number;
2533
+ output: number;
2534
+ cacheRead: number;
2535
+ cacheWrite: number;
2536
+ };
2537
+ contextWindow: number;
2538
+ maxTokens: number;
2539
+ };
2489
2540
  readonly o1: {
2490
2541
  id: string;
2491
2542
  name: string;
@@ -3276,6 +3327,29 @@ export declare const MODELS: {
3276
3327
  contextWindow: number;
3277
3328
  maxTokens: number;
3278
3329
  };
3330
+ readonly "gpt-5.5": {
3331
+ id: string;
3332
+ name: string;
3333
+ api: "openai-responses";
3334
+ provider: string;
3335
+ baseUrl: string;
3336
+ headers: {
3337
+ "User-Agent": string;
3338
+ "Editor-Version": string;
3339
+ "Editor-Plugin-Version": string;
3340
+ "Copilot-Integration-Id": string;
3341
+ };
3342
+ reasoning: true;
3343
+ input: ("image" | "text")[];
3344
+ cost: {
3345
+ input: number;
3346
+ output: number;
3347
+ cacheRead: number;
3348
+ cacheWrite: number;
3349
+ };
3350
+ contextWindow: number;
3351
+ maxTokens: number;
3352
+ };
3279
3353
  readonly "grok-code-fast-1": {
3280
3354
  id: string;
3281
3355
  name: string;
@@ -3731,7 +3805,7 @@ export declare const MODELS: {
3731
3805
  contextWindow: number;
3732
3806
  maxTokens: number;
3733
3807
  };
3734
- readonly "gemma-4-26b-it": {
3808
+ readonly "gemma-4-26b-a4b-it": {
3735
3809
  id: string;
3736
3810
  name: string;
3737
3811
  api: "google-generative-ai";
@@ -4797,6 +4871,26 @@ export declare const MODELS: {
4797
4871
  contextWindow: number;
4798
4872
  maxTokens: number;
4799
4873
  };
4874
+ readonly "deepseek-ai/DeepSeek-V4-Pro": {
4875
+ id: string;
4876
+ name: string;
4877
+ api: "openai-completions";
4878
+ provider: string;
4879
+ baseUrl: string;
4880
+ compat: {
4881
+ supportsDeveloperRole: false;
4882
+ };
4883
+ reasoning: true;
4884
+ input: "text"[];
4885
+ cost: {
4886
+ input: number;
4887
+ output: number;
4888
+ cacheRead: number;
4889
+ cacheWrite: number;
4890
+ };
4891
+ contextWindow: number;
4892
+ maxTokens: number;
4893
+ };
4800
4894
  readonly "moonshotai/Kimi-K2-Instruct": {
4801
4895
  id: string;
4802
4896
  name: string;
@@ -4877,6 +4971,26 @@ export declare const MODELS: {
4877
4971
  contextWindow: number;
4878
4972
  maxTokens: number;
4879
4973
  };
4974
+ readonly "moonshotai/Kimi-K2.6": {
4975
+ id: string;
4976
+ name: string;
4977
+ api: "openai-completions";
4978
+ provider: string;
4979
+ baseUrl: string;
4980
+ compat: {
4981
+ supportsDeveloperRole: false;
4982
+ };
4983
+ reasoning: true;
4984
+ input: ("image" | "text")[];
4985
+ cost: {
4986
+ input: number;
4987
+ output: number;
4988
+ cacheRead: number;
4989
+ cacheWrite: number;
4990
+ };
4991
+ contextWindow: number;
4992
+ maxTokens: number;
4993
+ };
4880
4994
  readonly "zai-org/GLM-4.7": {
4881
4995
  id: string;
4882
4996
  name: string;
@@ -4959,6 +5073,23 @@ export declare const MODELS: {
4959
5073
  };
4960
5074
  };
4961
5075
  readonly "kimi-coding": {
5076
+ readonly k2p6: {
5077
+ id: string;
5078
+ name: string;
5079
+ api: "anthropic-messages";
5080
+ provider: string;
5081
+ baseUrl: string;
5082
+ reasoning: true;
5083
+ input: ("image" | "text")[];
5084
+ cost: {
5085
+ input: number;
5086
+ output: number;
5087
+ cacheRead: number;
5088
+ cacheWrite: number;
5089
+ };
5090
+ contextWindow: number;
5091
+ maxTokens: number;
5092
+ };
4962
5093
  readonly "kimi-for-coding": {
4963
5094
  id: string;
4964
5095
  name: string;
@@ -6055,6 +6186,23 @@ export declare const MODELS: {
6055
6186
  contextWindow: number;
6056
6187
  maxTokens: number;
6057
6188
  };
6189
+ readonly "gpt-5.5": {
6190
+ id: string;
6191
+ name: string;
6192
+ api: "openai-responses";
6193
+ provider: string;
6194
+ baseUrl: string;
6195
+ reasoning: true;
6196
+ input: ("image" | "text")[];
6197
+ cost: {
6198
+ input: number;
6199
+ output: number;
6200
+ cacheRead: number;
6201
+ cacheWrite: number;
6202
+ };
6203
+ contextWindow: number;
6204
+ maxTokens: number;
6205
+ };
6058
6206
  readonly o1: {
6059
6207
  id: string;
6060
6208
  name: string;
@@ -6365,23 +6513,6 @@ export declare const MODELS: {
6365
6513
  contextWindow: number;
6366
6514
  maxTokens: number;
6367
6515
  };
6368
- readonly "claude-3-5-haiku": {
6369
- id: string;
6370
- name: string;
6371
- api: "anthropic-messages";
6372
- provider: string;
6373
- baseUrl: string;
6374
- reasoning: false;
6375
- input: ("image" | "text")[];
6376
- cost: {
6377
- input: number;
6378
- output: number;
6379
- cacheRead: number;
6380
- cacheWrite: number;
6381
- };
6382
- contextWindow: number;
6383
- maxTokens: number;
6384
- };
6385
6516
  readonly "claude-haiku-4-5": {
6386
6517
  id: string;
6387
6518
  name: string;
@@ -6824,10 +6955,10 @@ export declare const MODELS: {
6824
6955
  contextWindow: number;
6825
6956
  maxTokens: number;
6826
6957
  };
6827
- readonly "kimi-k2.5": {
6958
+ readonly "gpt-5.5": {
6828
6959
  id: string;
6829
6960
  name: string;
6830
- api: "openai-completions";
6961
+ api: "openai-responses";
6831
6962
  provider: string;
6832
6963
  baseUrl: string;
6833
6964
  reasoning: true;
@@ -6841,14 +6972,14 @@ export declare const MODELS: {
6841
6972
  contextWindow: number;
6842
6973
  maxTokens: number;
6843
6974
  };
6844
- readonly "minimax-m2.5": {
6975
+ readonly "gpt-5.5-pro": {
6845
6976
  id: string;
6846
6977
  name: string;
6847
- api: "openai-completions";
6978
+ api: "openai-responses";
6848
6979
  provider: string;
6849
6980
  baseUrl: string;
6850
6981
  reasoning: true;
6851
- input: "text"[];
6982
+ input: ("image" | "text")[];
6852
6983
  cost: {
6853
6984
  input: number;
6854
6985
  output: number;
@@ -6858,10 +6989,10 @@ export declare const MODELS: {
6858
6989
  contextWindow: number;
6859
6990
  maxTokens: number;
6860
6991
  };
6861
- readonly "minimax-m2.5-free": {
6992
+ readonly "hy3-preview-free": {
6862
6993
  id: string;
6863
6994
  name: string;
6864
- api: "anthropic-messages";
6995
+ api: "openai-completions";
6865
6996
  provider: string;
6866
6997
  baseUrl: string;
6867
6998
  reasoning: true;
@@ -6875,14 +7006,14 @@ export declare const MODELS: {
6875
7006
  contextWindow: number;
6876
7007
  maxTokens: number;
6877
7008
  };
6878
- readonly "nemotron-3-super-free": {
7009
+ readonly "kimi-k2.5": {
6879
7010
  id: string;
6880
7011
  name: string;
6881
7012
  api: "openai-completions";
6882
7013
  provider: string;
6883
7014
  baseUrl: string;
6884
7015
  reasoning: true;
6885
- input: "text"[];
7016
+ input: ("image" | "text")[];
6886
7017
  cost: {
6887
7018
  input: number;
6888
7019
  output: number;
@@ -6892,7 +7023,7 @@ export declare const MODELS: {
6892
7023
  contextWindow: number;
6893
7024
  maxTokens: number;
6894
7025
  };
6895
- readonly "qwen3.5-plus": {
7026
+ readonly "kimi-k2.6": {
6896
7027
  id: string;
6897
7028
  name: string;
6898
7029
  api: "openai-completions";
@@ -6909,14 +7040,14 @@ export declare const MODELS: {
6909
7040
  contextWindow: number;
6910
7041
  maxTokens: number;
6911
7042
  };
6912
- readonly "qwen3.6-plus": {
7043
+ readonly "ling-2.6-flash-free": {
6913
7044
  id: string;
6914
7045
  name: string;
6915
7046
  api: "openai-completions";
6916
7047
  provider: string;
6917
7048
  baseUrl: string;
6918
- reasoning: true;
6919
- input: ("image" | "text")[];
7049
+ reasoning: false;
7050
+ input: "text"[];
6920
7051
  cost: {
6921
7052
  input: number;
6922
7053
  output: number;
@@ -6926,9 +7057,7 @@ export declare const MODELS: {
6926
7057
  contextWindow: number;
6927
7058
  maxTokens: number;
6928
7059
  };
6929
- };
6930
- readonly "opencode-go": {
6931
- readonly "glm-5": {
7060
+ readonly "minimax-m2.5": {
6932
7061
  id: string;
6933
7062
  name: string;
6934
7063
  api: "openai-completions";
@@ -6945,10 +7074,10 @@ export declare const MODELS: {
6945
7074
  contextWindow: number;
6946
7075
  maxTokens: number;
6947
7076
  };
6948
- readonly "glm-5.1": {
7077
+ readonly "minimax-m2.5-free": {
6949
7078
  id: string;
6950
7079
  name: string;
6951
- api: "openai-completions";
7080
+ api: "anthropic-messages";
6952
7081
  provider: string;
6953
7082
  baseUrl: string;
6954
7083
  reasoning: true;
@@ -6962,14 +7091,14 @@ export declare const MODELS: {
6962
7091
  contextWindow: number;
6963
7092
  maxTokens: number;
6964
7093
  };
6965
- readonly "kimi-k2.5": {
7094
+ readonly "minimax-m2.7": {
6966
7095
  id: string;
6967
7096
  name: string;
6968
7097
  api: "openai-completions";
6969
7098
  provider: string;
6970
7099
  baseUrl: string;
6971
7100
  reasoning: true;
6972
- input: ("image" | "text")[];
7101
+ input: "text"[];
6973
7102
  cost: {
6974
7103
  input: number;
6975
7104
  output: number;
@@ -6979,14 +7108,14 @@ export declare const MODELS: {
6979
7108
  contextWindow: number;
6980
7109
  maxTokens: number;
6981
7110
  };
6982
- readonly "mimo-v2-omni": {
7111
+ readonly "nemotron-3-super-free": {
6983
7112
  id: string;
6984
7113
  name: string;
6985
7114
  api: "openai-completions";
6986
7115
  provider: string;
6987
7116
  baseUrl: string;
6988
7117
  reasoning: true;
6989
- input: ("image" | "text")[];
7118
+ input: "text"[];
6990
7119
  cost: {
6991
7120
  input: number;
6992
7121
  output: number;
@@ -6996,14 +7125,14 @@ export declare const MODELS: {
6996
7125
  contextWindow: number;
6997
7126
  maxTokens: number;
6998
7127
  };
6999
- readonly "mimo-v2-pro": {
7128
+ readonly "qwen3.5-plus": {
7000
7129
  id: string;
7001
7130
  name: string;
7002
- api: "openai-completions";
7131
+ api: "anthropic-messages";
7003
7132
  provider: string;
7004
7133
  baseUrl: string;
7005
7134
  reasoning: true;
7006
- input: "text"[];
7135
+ input: ("image" | "text")[];
7007
7136
  cost: {
7008
7137
  input: number;
7009
7138
  output: number;
@@ -7013,14 +7142,14 @@ export declare const MODELS: {
7013
7142
  contextWindow: number;
7014
7143
  maxTokens: number;
7015
7144
  };
7016
- readonly "minimax-m2.5": {
7145
+ readonly "qwen3.6-plus": {
7017
7146
  id: string;
7018
7147
  name: string;
7019
- api: "openai-completions";
7148
+ api: "anthropic-messages";
7020
7149
  provider: string;
7021
7150
  baseUrl: string;
7022
7151
  reasoning: true;
7023
- input: "text"[];
7152
+ input: ("image" | "text")[];
7024
7153
  cost: {
7025
7154
  input: number;
7026
7155
  output: number;
@@ -7030,10 +7159,12 @@ export declare const MODELS: {
7030
7159
  contextWindow: number;
7031
7160
  maxTokens: number;
7032
7161
  };
7033
- readonly "minimax-m2.7": {
7162
+ };
7163
+ readonly "opencode-go": {
7164
+ readonly "deepseek-v4-flash": {
7034
7165
  id: string;
7035
7166
  name: string;
7036
- api: "anthropic-messages";
7167
+ api: "openai-completions";
7037
7168
  provider: string;
7038
7169
  baseUrl: string;
7039
7170
  reasoning: true;
@@ -7047,14 +7178,14 @@ export declare const MODELS: {
7047
7178
  contextWindow: number;
7048
7179
  maxTokens: number;
7049
7180
  };
7050
- readonly "qwen3.5-plus": {
7181
+ readonly "deepseek-v4-pro": {
7051
7182
  id: string;
7052
7183
  name: string;
7053
7184
  api: "openai-completions";
7054
7185
  provider: string;
7055
7186
  baseUrl: string;
7056
7187
  reasoning: true;
7057
- input: ("image" | "text")[];
7188
+ input: "text"[];
7058
7189
  cost: {
7059
7190
  input: number;
7060
7191
  output: number;
@@ -7064,14 +7195,14 @@ export declare const MODELS: {
7064
7195
  contextWindow: number;
7065
7196
  maxTokens: number;
7066
7197
  };
7067
- readonly "qwen3.6-plus": {
7198
+ readonly "glm-5": {
7068
7199
  id: string;
7069
7200
  name: string;
7070
7201
  api: "openai-completions";
7071
7202
  provider: string;
7072
7203
  baseUrl: string;
7073
7204
  reasoning: true;
7074
- input: ("image" | "text")[];
7205
+ input: "text"[];
7075
7206
  cost: {
7076
7207
  input: number;
7077
7208
  output: number;
@@ -7081,15 +7212,13 @@ export declare const MODELS: {
7081
7212
  contextWindow: number;
7082
7213
  maxTokens: number;
7083
7214
  };
7084
- };
7085
- readonly openrouter: {
7086
- readonly "ai21/jamba-large-1.7": {
7215
+ readonly "glm-5.1": {
7087
7216
  id: string;
7088
7217
  name: string;
7089
7218
  api: "openai-completions";
7090
7219
  provider: string;
7091
7220
  baseUrl: string;
7092
- reasoning: false;
7221
+ reasoning: true;
7093
7222
  input: "text"[];
7094
7223
  cost: {
7095
7224
  input: number;
@@ -7100,14 +7229,14 @@ export declare const MODELS: {
7100
7229
  contextWindow: number;
7101
7230
  maxTokens: number;
7102
7231
  };
7103
- readonly "alibaba/tongyi-deepresearch-30b-a3b": {
7232
+ readonly "kimi-k2.5": {
7104
7233
  id: string;
7105
7234
  name: string;
7106
7235
  api: "openai-completions";
7107
7236
  provider: string;
7108
7237
  baseUrl: string;
7109
7238
  reasoning: true;
7110
- input: "text"[];
7239
+ input: ("image" | "text")[];
7111
7240
  cost: {
7112
7241
  input: number;
7113
7242
  output: number;
@@ -7117,14 +7246,14 @@ export declare const MODELS: {
7117
7246
  contextWindow: number;
7118
7247
  maxTokens: number;
7119
7248
  };
7120
- readonly "allenai/olmo-3.1-32b-instruct": {
7249
+ readonly "kimi-k2.6": {
7121
7250
  id: string;
7122
7251
  name: string;
7123
7252
  api: "openai-completions";
7124
7253
  provider: string;
7125
7254
  baseUrl: string;
7126
- reasoning: false;
7127
- input: "text"[];
7255
+ reasoning: true;
7256
+ input: ("image" | "text")[];
7128
7257
  cost: {
7129
7258
  input: number;
7130
7259
  output: number;
@@ -7134,7 +7263,7 @@ export declare const MODELS: {
7134
7263
  contextWindow: number;
7135
7264
  maxTokens: number;
7136
7265
  };
7137
- readonly "amazon/nova-2-lite-v1": {
7266
+ readonly "mimo-v2-omni": {
7138
7267
  id: string;
7139
7268
  name: string;
7140
7269
  api: "openai-completions";
@@ -7151,14 +7280,14 @@ export declare const MODELS: {
7151
7280
  contextWindow: number;
7152
7281
  maxTokens: number;
7153
7282
  };
7154
- readonly "amazon/nova-lite-v1": {
7283
+ readonly "mimo-v2-pro": {
7155
7284
  id: string;
7156
7285
  name: string;
7157
7286
  api: "openai-completions";
7158
7287
  provider: string;
7159
7288
  baseUrl: string;
7160
- reasoning: false;
7161
- input: ("image" | "text")[];
7289
+ reasoning: true;
7290
+ input: "text"[];
7162
7291
  cost: {
7163
7292
  input: number;
7164
7293
  output: number;
@@ -7168,14 +7297,14 @@ export declare const MODELS: {
7168
7297
  contextWindow: number;
7169
7298
  maxTokens: number;
7170
7299
  };
7171
- readonly "amazon/nova-micro-v1": {
7300
+ readonly "mimo-v2.5": {
7172
7301
  id: string;
7173
7302
  name: string;
7174
7303
  api: "openai-completions";
7175
7304
  provider: string;
7176
7305
  baseUrl: string;
7177
- reasoning: false;
7178
- input: "text"[];
7306
+ reasoning: true;
7307
+ input: ("image" | "text")[];
7179
7308
  cost: {
7180
7309
  input: number;
7181
7310
  output: number;
@@ -7185,14 +7314,14 @@ export declare const MODELS: {
7185
7314
  contextWindow: number;
7186
7315
  maxTokens: number;
7187
7316
  };
7188
- readonly "amazon/nova-premier-v1": {
7317
+ readonly "mimo-v2.5-pro": {
7189
7318
  id: string;
7190
7319
  name: string;
7191
7320
  api: "openai-completions";
7192
7321
  provider: string;
7193
7322
  baseUrl: string;
7194
- reasoning: false;
7195
- input: ("image" | "text")[];
7323
+ reasoning: true;
7324
+ input: "text"[];
7196
7325
  cost: {
7197
7326
  input: number;
7198
7327
  output: number;
@@ -7202,14 +7331,203 @@ export declare const MODELS: {
7202
7331
  contextWindow: number;
7203
7332
  maxTokens: number;
7204
7333
  };
7205
- readonly "amazon/nova-pro-v1": {
7334
+ readonly "minimax-m2.5": {
7206
7335
  id: string;
7207
7336
  name: string;
7208
7337
  api: "openai-completions";
7209
7338
  provider: string;
7210
7339
  baseUrl: string;
7211
- reasoning: false;
7212
- input: ("image" | "text")[];
7340
+ reasoning: true;
7341
+ input: "text"[];
7342
+ cost: {
7343
+ input: number;
7344
+ output: number;
7345
+ cacheRead: number;
7346
+ cacheWrite: number;
7347
+ };
7348
+ contextWindow: number;
7349
+ maxTokens: number;
7350
+ };
7351
+ readonly "minimax-m2.7": {
7352
+ id: string;
7353
+ name: string;
7354
+ api: "anthropic-messages";
7355
+ provider: string;
7356
+ baseUrl: string;
7357
+ reasoning: true;
7358
+ input: "text"[];
7359
+ cost: {
7360
+ input: number;
7361
+ output: number;
7362
+ cacheRead: number;
7363
+ cacheWrite: number;
7364
+ };
7365
+ contextWindow: number;
7366
+ maxTokens: number;
7367
+ };
7368
+ readonly "qwen3.5-plus": {
7369
+ id: string;
7370
+ name: string;
7371
+ api: "anthropic-messages";
7372
+ provider: string;
7373
+ baseUrl: string;
7374
+ reasoning: true;
7375
+ input: ("image" | "text")[];
7376
+ cost: {
7377
+ input: number;
7378
+ output: number;
7379
+ cacheRead: number;
7380
+ cacheWrite: number;
7381
+ };
7382
+ contextWindow: number;
7383
+ maxTokens: number;
7384
+ };
7385
+ readonly "qwen3.6-plus": {
7386
+ id: string;
7387
+ name: string;
7388
+ api: "anthropic-messages";
7389
+ provider: string;
7390
+ baseUrl: string;
7391
+ reasoning: true;
7392
+ input: ("image" | "text")[];
7393
+ cost: {
7394
+ input: number;
7395
+ output: number;
7396
+ cacheRead: number;
7397
+ cacheWrite: number;
7398
+ };
7399
+ contextWindow: number;
7400
+ maxTokens: number;
7401
+ };
7402
+ };
7403
+ readonly openrouter: {
7404
+ readonly "ai21/jamba-large-1.7": {
7405
+ id: string;
7406
+ name: string;
7407
+ api: "openai-completions";
7408
+ provider: string;
7409
+ baseUrl: string;
7410
+ reasoning: false;
7411
+ input: "text"[];
7412
+ cost: {
7413
+ input: number;
7414
+ output: number;
7415
+ cacheRead: number;
7416
+ cacheWrite: number;
7417
+ };
7418
+ contextWindow: number;
7419
+ maxTokens: number;
7420
+ };
7421
+ readonly "alibaba/tongyi-deepresearch-30b-a3b": {
7422
+ id: string;
7423
+ name: string;
7424
+ api: "openai-completions";
7425
+ provider: string;
7426
+ baseUrl: string;
7427
+ reasoning: true;
7428
+ input: "text"[];
7429
+ cost: {
7430
+ input: number;
7431
+ output: number;
7432
+ cacheRead: number;
7433
+ cacheWrite: number;
7434
+ };
7435
+ contextWindow: number;
7436
+ maxTokens: number;
7437
+ };
7438
+ readonly "allenai/olmo-3.1-32b-instruct": {
7439
+ id: string;
7440
+ name: string;
7441
+ api: "openai-completions";
7442
+ provider: string;
7443
+ baseUrl: string;
7444
+ reasoning: false;
7445
+ input: "text"[];
7446
+ cost: {
7447
+ input: number;
7448
+ output: number;
7449
+ cacheRead: number;
7450
+ cacheWrite: number;
7451
+ };
7452
+ contextWindow: number;
7453
+ maxTokens: number;
7454
+ };
7455
+ readonly "amazon/nova-2-lite-v1": {
7456
+ id: string;
7457
+ name: string;
7458
+ api: "openai-completions";
7459
+ provider: string;
7460
+ baseUrl: string;
7461
+ reasoning: true;
7462
+ input: ("image" | "text")[];
7463
+ cost: {
7464
+ input: number;
7465
+ output: number;
7466
+ cacheRead: number;
7467
+ cacheWrite: number;
7468
+ };
7469
+ contextWindow: number;
7470
+ maxTokens: number;
7471
+ };
7472
+ readonly "amazon/nova-lite-v1": {
7473
+ id: string;
7474
+ name: string;
7475
+ api: "openai-completions";
7476
+ provider: string;
7477
+ baseUrl: string;
7478
+ reasoning: false;
7479
+ input: ("image" | "text")[];
7480
+ cost: {
7481
+ input: number;
7482
+ output: number;
7483
+ cacheRead: number;
7484
+ cacheWrite: number;
7485
+ };
7486
+ contextWindow: number;
7487
+ maxTokens: number;
7488
+ };
7489
+ readonly "amazon/nova-micro-v1": {
7490
+ id: string;
7491
+ name: string;
7492
+ api: "openai-completions";
7493
+ provider: string;
7494
+ baseUrl: string;
7495
+ reasoning: false;
7496
+ input: "text"[];
7497
+ cost: {
7498
+ input: number;
7499
+ output: number;
7500
+ cacheRead: number;
7501
+ cacheWrite: number;
7502
+ };
7503
+ contextWindow: number;
7504
+ maxTokens: number;
7505
+ };
7506
+ readonly "amazon/nova-premier-v1": {
7507
+ id: string;
7508
+ name: string;
7509
+ api: "openai-completions";
7510
+ provider: string;
7511
+ baseUrl: string;
7512
+ reasoning: false;
7513
+ input: ("image" | "text")[];
7514
+ cost: {
7515
+ input: number;
7516
+ output: number;
7517
+ cacheRead: number;
7518
+ cacheWrite: number;
7519
+ };
7520
+ contextWindow: number;
7521
+ maxTokens: number;
7522
+ };
7523
+ readonly "amazon/nova-pro-v1": {
7524
+ id: string;
7525
+ name: string;
7526
+ api: "openai-completions";
7527
+ provider: string;
7528
+ baseUrl: string;
7529
+ reasoning: false;
7530
+ input: ("image" | "text")[];
7213
7531
  cost: {
7214
7532
  input: number;
7215
7533
  output: number;
@@ -7457,7 +7775,7 @@ export declare const MODELS: {
7457
7775
  contextWindow: number;
7458
7776
  maxTokens: number;
7459
7777
  };
7460
- readonly "arcee-ai/trinity-large-preview:free": {
7778
+ readonly "arcee-ai/trinity-large-preview": {
7461
7779
  id: string;
7462
7780
  name: string;
7463
7781
  api: "openai-completions";
@@ -7814,6 +8132,40 @@ export declare const MODELS: {
7814
8132
  contextWindow: number;
7815
8133
  maxTokens: number;
7816
8134
  };
8135
+ readonly "deepseek/deepseek-v4-flash": {
8136
+ id: string;
8137
+ name: string;
8138
+ api: "openai-completions";
8139
+ provider: string;
8140
+ baseUrl: string;
8141
+ reasoning: true;
8142
+ input: "text"[];
8143
+ cost: {
8144
+ input: number;
8145
+ output: number;
8146
+ cacheRead: number;
8147
+ cacheWrite: number;
8148
+ };
8149
+ contextWindow: number;
8150
+ maxTokens: number;
8151
+ };
8152
+ readonly "deepseek/deepseek-v4-pro": {
8153
+ id: string;
8154
+ name: string;
8155
+ api: "openai-completions";
8156
+ provider: string;
8157
+ baseUrl: string;
8158
+ reasoning: true;
8159
+ input: "text"[];
8160
+ cost: {
8161
+ input: number;
8162
+ output: number;
8163
+ cacheRead: number;
8164
+ cacheWrite: number;
8165
+ };
8166
+ contextWindow: number;
8167
+ maxTokens: number;
8168
+ };
7817
8169
  readonly "essentialai/rnj-1-instruct": {
7818
8170
  id: string;
7819
8171
  name: string;
@@ -8035,6 +8387,40 @@ export declare const MODELS: {
8035
8387
  contextWindow: number;
8036
8388
  maxTokens: number;
8037
8389
  };
8390
+ readonly "google/gemma-3-12b-it": {
8391
+ id: string;
8392
+ name: string;
8393
+ api: "openai-completions";
8394
+ provider: string;
8395
+ baseUrl: string;
8396
+ reasoning: false;
8397
+ input: ("image" | "text")[];
8398
+ cost: {
8399
+ input: number;
8400
+ output: number;
8401
+ cacheRead: number;
8402
+ cacheWrite: number;
8403
+ };
8404
+ contextWindow: number;
8405
+ maxTokens: number;
8406
+ };
8407
+ readonly "google/gemma-3-27b-it": {
8408
+ id: string;
8409
+ name: string;
8410
+ api: "openai-completions";
8411
+ provider: string;
8412
+ baseUrl: string;
8413
+ reasoning: false;
8414
+ input: ("image" | "text")[];
8415
+ cost: {
8416
+ input: number;
8417
+ output: number;
8418
+ cacheRead: number;
8419
+ cacheWrite: number;
8420
+ };
8421
+ contextWindow: number;
8422
+ maxTokens: number;
8423
+ };
8038
8424
  readonly "google/gemma-4-26b-a4b-it": {
8039
8425
  id: string;
8040
8426
  name: string;
@@ -8103,13 +8489,47 @@ export declare const MODELS: {
8103
8489
  contextWindow: number;
8104
8490
  maxTokens: number;
8105
8491
  };
8106
- readonly "inception/mercury-2": {
8492
+ readonly "inception/mercury-2": {
8493
+ id: string;
8494
+ name: string;
8495
+ api: "openai-completions";
8496
+ provider: string;
8497
+ baseUrl: string;
8498
+ reasoning: true;
8499
+ input: "text"[];
8500
+ cost: {
8501
+ input: number;
8502
+ output: number;
8503
+ cacheRead: number;
8504
+ cacheWrite: number;
8505
+ };
8506
+ contextWindow: number;
8507
+ maxTokens: number;
8508
+ };
8509
+ readonly "inclusionai/ling-2.6-1t:free": {
8510
+ id: string;
8511
+ name: string;
8512
+ api: "openai-completions";
8513
+ provider: string;
8514
+ baseUrl: string;
8515
+ reasoning: false;
8516
+ input: "text"[];
8517
+ cost: {
8518
+ input: number;
8519
+ output: number;
8520
+ cacheRead: number;
8521
+ cacheWrite: number;
8522
+ };
8523
+ contextWindow: number;
8524
+ maxTokens: number;
8525
+ };
8526
+ readonly "inclusionai/ling-2.6-flash:free": {
8107
8527
  id: string;
8108
8528
  name: string;
8109
8529
  api: "openai-completions";
8110
8530
  provider: string;
8111
8531
  baseUrl: string;
8112
- reasoning: true;
8532
+ reasoning: false;
8113
8533
  input: "text"[];
8114
8534
  cost: {
8115
8535
  input: number;
@@ -8783,6 +9203,23 @@ export declare const MODELS: {
8783
9203
  contextWindow: number;
8784
9204
  maxTokens: number;
8785
9205
  };
9206
+ readonly "moonshotai/kimi-k2.6": {
9207
+ id: string;
9208
+ name: string;
9209
+ api: "openai-completions";
9210
+ provider: string;
9211
+ baseUrl: string;
9212
+ reasoning: true;
9213
+ input: ("image" | "text")[];
9214
+ cost: {
9215
+ input: number;
9216
+ output: number;
9217
+ cacheRead: number;
9218
+ cacheWrite: number;
9219
+ };
9220
+ contextWindow: number;
9221
+ maxTokens: number;
9222
+ };
8786
9223
  readonly "nex-agi/deepseek-v3.1-nex-n1": {
8787
9224
  id: string;
8788
9225
  name: string;
@@ -9191,13 +9628,251 @@ export declare const MODELS: {
9191
9628
  contextWindow: number;
9192
9629
  maxTokens: number;
9193
9630
  };
9194
- readonly "openai/gpt-4o-2024-11-20": {
9631
+ readonly "openai/gpt-4o-2024-11-20": {
9632
+ id: string;
9633
+ name: string;
9634
+ api: "openai-completions";
9635
+ provider: string;
9636
+ baseUrl: string;
9637
+ reasoning: false;
9638
+ input: ("image" | "text")[];
9639
+ cost: {
9640
+ input: number;
9641
+ output: number;
9642
+ cacheRead: number;
9643
+ cacheWrite: number;
9644
+ };
9645
+ contextWindow: number;
9646
+ maxTokens: number;
9647
+ };
9648
+ readonly "openai/gpt-4o-audio-preview": {
9649
+ id: string;
9650
+ name: string;
9651
+ api: "openai-completions";
9652
+ provider: string;
9653
+ baseUrl: string;
9654
+ reasoning: false;
9655
+ input: "text"[];
9656
+ cost: {
9657
+ input: number;
9658
+ output: number;
9659
+ cacheRead: number;
9660
+ cacheWrite: number;
9661
+ };
9662
+ contextWindow: number;
9663
+ maxTokens: number;
9664
+ };
9665
+ readonly "openai/gpt-4o-mini": {
9666
+ id: string;
9667
+ name: string;
9668
+ api: "openai-completions";
9669
+ provider: string;
9670
+ baseUrl: string;
9671
+ reasoning: false;
9672
+ input: ("image" | "text")[];
9673
+ cost: {
9674
+ input: number;
9675
+ output: number;
9676
+ cacheRead: number;
9677
+ cacheWrite: number;
9678
+ };
9679
+ contextWindow: number;
9680
+ maxTokens: number;
9681
+ };
9682
+ readonly "openai/gpt-4o-mini-2024-07-18": {
9683
+ id: string;
9684
+ name: string;
9685
+ api: "openai-completions";
9686
+ provider: string;
9687
+ baseUrl: string;
9688
+ reasoning: false;
9689
+ input: ("image" | "text")[];
9690
+ cost: {
9691
+ input: number;
9692
+ output: number;
9693
+ cacheRead: number;
9694
+ cacheWrite: number;
9695
+ };
9696
+ contextWindow: number;
9697
+ maxTokens: number;
9698
+ };
9699
+ readonly "openai/gpt-5": {
9700
+ id: string;
9701
+ name: string;
9702
+ api: "openai-completions";
9703
+ provider: string;
9704
+ baseUrl: string;
9705
+ reasoning: true;
9706
+ input: ("image" | "text")[];
9707
+ cost: {
9708
+ input: number;
9709
+ output: number;
9710
+ cacheRead: number;
9711
+ cacheWrite: number;
9712
+ };
9713
+ contextWindow: number;
9714
+ maxTokens: number;
9715
+ };
9716
+ readonly "openai/gpt-5-codex": {
9717
+ id: string;
9718
+ name: string;
9719
+ api: "openai-completions";
9720
+ provider: string;
9721
+ baseUrl: string;
9722
+ reasoning: true;
9723
+ input: ("image" | "text")[];
9724
+ cost: {
9725
+ input: number;
9726
+ output: number;
9727
+ cacheRead: number;
9728
+ cacheWrite: number;
9729
+ };
9730
+ contextWindow: number;
9731
+ maxTokens: number;
9732
+ };
9733
+ readonly "openai/gpt-5-mini": {
9734
+ id: string;
9735
+ name: string;
9736
+ api: "openai-completions";
9737
+ provider: string;
9738
+ baseUrl: string;
9739
+ reasoning: true;
9740
+ input: ("image" | "text")[];
9741
+ cost: {
9742
+ input: number;
9743
+ output: number;
9744
+ cacheRead: number;
9745
+ cacheWrite: number;
9746
+ };
9747
+ contextWindow: number;
9748
+ maxTokens: number;
9749
+ };
9750
+ readonly "openai/gpt-5-nano": {
9751
+ id: string;
9752
+ name: string;
9753
+ api: "openai-completions";
9754
+ provider: string;
9755
+ baseUrl: string;
9756
+ reasoning: true;
9757
+ input: ("image" | "text")[];
9758
+ cost: {
9759
+ input: number;
9760
+ output: number;
9761
+ cacheRead: number;
9762
+ cacheWrite: number;
9763
+ };
9764
+ contextWindow: number;
9765
+ maxTokens: number;
9766
+ };
9767
+ readonly "openai/gpt-5-pro": {
9768
+ id: string;
9769
+ name: string;
9770
+ api: "openai-completions";
9771
+ provider: string;
9772
+ baseUrl: string;
9773
+ reasoning: true;
9774
+ input: ("image" | "text")[];
9775
+ cost: {
9776
+ input: number;
9777
+ output: number;
9778
+ cacheRead: number;
9779
+ cacheWrite: number;
9780
+ };
9781
+ contextWindow: number;
9782
+ maxTokens: number;
9783
+ };
9784
+ readonly "openai/gpt-5.1": {
9785
+ id: string;
9786
+ name: string;
9787
+ api: "openai-completions";
9788
+ provider: string;
9789
+ baseUrl: string;
9790
+ reasoning: true;
9791
+ input: ("image" | "text")[];
9792
+ cost: {
9793
+ input: number;
9794
+ output: number;
9795
+ cacheRead: number;
9796
+ cacheWrite: number;
9797
+ };
9798
+ contextWindow: number;
9799
+ maxTokens: number;
9800
+ };
9801
+ readonly "openai/gpt-5.1-chat": {
9802
+ id: string;
9803
+ name: string;
9804
+ api: "openai-completions";
9805
+ provider: string;
9806
+ baseUrl: string;
9807
+ reasoning: false;
9808
+ input: ("image" | "text")[];
9809
+ cost: {
9810
+ input: number;
9811
+ output: number;
9812
+ cacheRead: number;
9813
+ cacheWrite: number;
9814
+ };
9815
+ contextWindow: number;
9816
+ maxTokens: number;
9817
+ };
9818
+ readonly "openai/gpt-5.1-codex": {
9819
+ id: string;
9820
+ name: string;
9821
+ api: "openai-completions";
9822
+ provider: string;
9823
+ baseUrl: string;
9824
+ reasoning: true;
9825
+ input: ("image" | "text")[];
9826
+ cost: {
9827
+ input: number;
9828
+ output: number;
9829
+ cacheRead: number;
9830
+ cacheWrite: number;
9831
+ };
9832
+ contextWindow: number;
9833
+ maxTokens: number;
9834
+ };
9835
+ readonly "openai/gpt-5.1-codex-max": {
9836
+ id: string;
9837
+ name: string;
9838
+ api: "openai-completions";
9839
+ provider: string;
9840
+ baseUrl: string;
9841
+ reasoning: true;
9842
+ input: ("image" | "text")[];
9843
+ cost: {
9844
+ input: number;
9845
+ output: number;
9846
+ cacheRead: number;
9847
+ cacheWrite: number;
9848
+ };
9849
+ contextWindow: number;
9850
+ maxTokens: number;
9851
+ };
9852
+ readonly "openai/gpt-5.1-codex-mini": {
9853
+ id: string;
9854
+ name: string;
9855
+ api: "openai-completions";
9856
+ provider: string;
9857
+ baseUrl: string;
9858
+ reasoning: true;
9859
+ input: ("image" | "text")[];
9860
+ cost: {
9861
+ input: number;
9862
+ output: number;
9863
+ cacheRead: number;
9864
+ cacheWrite: number;
9865
+ };
9866
+ contextWindow: number;
9867
+ maxTokens: number;
9868
+ };
9869
+ readonly "openai/gpt-5.2": {
9195
9870
  id: string;
9196
9871
  name: string;
9197
9872
  api: "openai-completions";
9198
9873
  provider: string;
9199
9874
  baseUrl: string;
9200
- reasoning: false;
9875
+ reasoning: true;
9201
9876
  input: ("image" | "text")[];
9202
9877
  cost: {
9203
9878
  input: number;
@@ -9208,14 +9883,14 @@ export declare const MODELS: {
9208
9883
  contextWindow: number;
9209
9884
  maxTokens: number;
9210
9885
  };
9211
- readonly "openai/gpt-4o-audio-preview": {
9886
+ readonly "openai/gpt-5.2-chat": {
9212
9887
  id: string;
9213
9888
  name: string;
9214
9889
  api: "openai-completions";
9215
9890
  provider: string;
9216
9891
  baseUrl: string;
9217
9892
  reasoning: false;
9218
- input: "text"[];
9893
+ input: ("image" | "text")[];
9219
9894
  cost: {
9220
9895
  input: number;
9221
9896
  output: number;
@@ -9225,13 +9900,13 @@ export declare const MODELS: {
9225
9900
  contextWindow: number;
9226
9901
  maxTokens: number;
9227
9902
  };
9228
- readonly "openai/gpt-4o-mini": {
9903
+ readonly "openai/gpt-5.2-codex": {
9229
9904
  id: string;
9230
9905
  name: string;
9231
9906
  api: "openai-completions";
9232
9907
  provider: string;
9233
9908
  baseUrl: string;
9234
- reasoning: false;
9909
+ reasoning: true;
9235
9910
  input: ("image" | "text")[];
9236
9911
  cost: {
9237
9912
  input: number;
@@ -9242,13 +9917,13 @@ export declare const MODELS: {
9242
9917
  contextWindow: number;
9243
9918
  maxTokens: number;
9244
9919
  };
9245
- readonly "openai/gpt-4o-mini-2024-07-18": {
9920
+ readonly "openai/gpt-5.2-pro": {
9246
9921
  id: string;
9247
9922
  name: string;
9248
9923
  api: "openai-completions";
9249
9924
  provider: string;
9250
9925
  baseUrl: string;
9251
- reasoning: false;
9926
+ reasoning: true;
9252
9927
  input: ("image" | "text")[];
9253
9928
  cost: {
9254
9929
  input: number;
@@ -9259,13 +9934,13 @@ export declare const MODELS: {
9259
9934
  contextWindow: number;
9260
9935
  maxTokens: number;
9261
9936
  };
9262
- readonly "openai/gpt-5": {
9937
+ readonly "openai/gpt-5.3-chat": {
9263
9938
  id: string;
9264
9939
  name: string;
9265
9940
  api: "openai-completions";
9266
9941
  provider: string;
9267
9942
  baseUrl: string;
9268
- reasoning: true;
9943
+ reasoning: false;
9269
9944
  input: ("image" | "text")[];
9270
9945
  cost: {
9271
9946
  input: number;
@@ -9276,7 +9951,7 @@ export declare const MODELS: {
9276
9951
  contextWindow: number;
9277
9952
  maxTokens: number;
9278
9953
  };
9279
- readonly "openai/gpt-5-codex": {
9954
+ readonly "openai/gpt-5.3-codex": {
9280
9955
  id: string;
9281
9956
  name: string;
9282
9957
  api: "openai-completions";
@@ -9293,7 +9968,7 @@ export declare const MODELS: {
9293
9968
  contextWindow: number;
9294
9969
  maxTokens: number;
9295
9970
  };
9296
- readonly "openai/gpt-5-image": {
9971
+ readonly "openai/gpt-5.4": {
9297
9972
  id: string;
9298
9973
  name: string;
9299
9974
  api: "openai-completions";
@@ -9310,7 +9985,7 @@ export declare const MODELS: {
9310
9985
  contextWindow: number;
9311
9986
  maxTokens: number;
9312
9987
  };
9313
- readonly "openai/gpt-5-image-mini": {
9988
+ readonly "openai/gpt-5.4-mini": {
9314
9989
  id: string;
9315
9990
  name: string;
9316
9991
  api: "openai-completions";
@@ -9327,7 +10002,7 @@ export declare const MODELS: {
9327
10002
  contextWindow: number;
9328
10003
  maxTokens: number;
9329
10004
  };
9330
- readonly "openai/gpt-5-mini": {
10005
+ readonly "openai/gpt-5.4-nano": {
9331
10006
  id: string;
9332
10007
  name: string;
9333
10008
  api: "openai-completions";
@@ -9344,7 +10019,7 @@ export declare const MODELS: {
9344
10019
  contextWindow: number;
9345
10020
  maxTokens: number;
9346
10021
  };
9347
- readonly "openai/gpt-5-nano": {
10022
+ readonly "openai/gpt-5.4-pro": {
9348
10023
  id: string;
9349
10024
  name: string;
9350
10025
  api: "openai-completions";
@@ -9361,7 +10036,7 @@ export declare const MODELS: {
9361
10036
  contextWindow: number;
9362
10037
  maxTokens: number;
9363
10038
  };
9364
- readonly "openai/gpt-5-pro": {
10039
+ readonly "openai/gpt-5.5": {
9365
10040
  id: string;
9366
10041
  name: string;
9367
10042
  api: "openai-completions";
@@ -9378,7 +10053,7 @@ export declare const MODELS: {
9378
10053
  contextWindow: number;
9379
10054
  maxTokens: number;
9380
10055
  };
9381
- readonly "openai/gpt-5.1": {
10056
+ readonly "openai/gpt-5.5-pro": {
9382
10057
  id: string;
9383
10058
  name: string;
9384
10059
  api: "openai-completions";
@@ -9395,14 +10070,14 @@ export declare const MODELS: {
9395
10070
  contextWindow: number;
9396
10071
  maxTokens: number;
9397
10072
  };
9398
- readonly "openai/gpt-5.1-chat": {
10073
+ readonly "openai/gpt-audio": {
9399
10074
  id: string;
9400
10075
  name: string;
9401
10076
  api: "openai-completions";
9402
10077
  provider: string;
9403
10078
  baseUrl: string;
9404
10079
  reasoning: false;
9405
- input: ("image" | "text")[];
10080
+ input: "text"[];
9406
10081
  cost: {
9407
10082
  input: number;
9408
10083
  output: number;
@@ -9412,14 +10087,14 @@ export declare const MODELS: {
9412
10087
  contextWindow: number;
9413
10088
  maxTokens: number;
9414
10089
  };
9415
- readonly "openai/gpt-5.1-codex": {
10090
+ readonly "openai/gpt-audio-mini": {
9416
10091
  id: string;
9417
10092
  name: string;
9418
10093
  api: "openai-completions";
9419
10094
  provider: string;
9420
10095
  baseUrl: string;
9421
- reasoning: true;
9422
- input: ("image" | "text")[];
10096
+ reasoning: false;
10097
+ input: "text"[];
9423
10098
  cost: {
9424
10099
  input: number;
9425
10100
  output: number;
@@ -9429,14 +10104,14 @@ export declare const MODELS: {
9429
10104
  contextWindow: number;
9430
10105
  maxTokens: number;
9431
10106
  };
9432
- readonly "openai/gpt-5.1-codex-max": {
10107
+ readonly "openai/gpt-oss-120b": {
9433
10108
  id: string;
9434
10109
  name: string;
9435
10110
  api: "openai-completions";
9436
10111
  provider: string;
9437
10112
  baseUrl: string;
9438
10113
  reasoning: true;
9439
- input: ("image" | "text")[];
10114
+ input: "text"[];
9440
10115
  cost: {
9441
10116
  input: number;
9442
10117
  output: number;
@@ -9446,14 +10121,14 @@ export declare const MODELS: {
9446
10121
  contextWindow: number;
9447
10122
  maxTokens: number;
9448
10123
  };
9449
- readonly "openai/gpt-5.1-codex-mini": {
10124
+ readonly "openai/gpt-oss-120b:free": {
9450
10125
  id: string;
9451
10126
  name: string;
9452
10127
  api: "openai-completions";
9453
10128
  provider: string;
9454
10129
  baseUrl: string;
9455
10130
  reasoning: true;
9456
- input: ("image" | "text")[];
10131
+ input: "text"[];
9457
10132
  cost: {
9458
10133
  input: number;
9459
10134
  output: number;
@@ -9463,14 +10138,14 @@ export declare const MODELS: {
9463
10138
  contextWindow: number;
9464
10139
  maxTokens: number;
9465
10140
  };
9466
- readonly "openai/gpt-5.2": {
10141
+ readonly "openai/gpt-oss-20b": {
9467
10142
  id: string;
9468
10143
  name: string;
9469
10144
  api: "openai-completions";
9470
10145
  provider: string;
9471
10146
  baseUrl: string;
9472
10147
  reasoning: true;
9473
- input: ("image" | "text")[];
10148
+ input: "text"[];
9474
10149
  cost: {
9475
10150
  input: number;
9476
10151
  output: number;
@@ -9480,14 +10155,14 @@ export declare const MODELS: {
9480
10155
  contextWindow: number;
9481
10156
  maxTokens: number;
9482
10157
  };
9483
- readonly "openai/gpt-5.2-chat": {
10158
+ readonly "openai/gpt-oss-20b:free": {
9484
10159
  id: string;
9485
10160
  name: string;
9486
10161
  api: "openai-completions";
9487
10162
  provider: string;
9488
10163
  baseUrl: string;
9489
- reasoning: false;
9490
- input: ("image" | "text")[];
10164
+ reasoning: true;
10165
+ input: "text"[];
9491
10166
  cost: {
9492
10167
  input: number;
9493
10168
  output: number;
@@ -9497,14 +10172,14 @@ export declare const MODELS: {
9497
10172
  contextWindow: number;
9498
10173
  maxTokens: number;
9499
10174
  };
9500
- readonly "openai/gpt-5.2-codex": {
10175
+ readonly "openai/gpt-oss-safeguard-20b": {
9501
10176
  id: string;
9502
10177
  name: string;
9503
10178
  api: "openai-completions";
9504
10179
  provider: string;
9505
10180
  baseUrl: string;
9506
10181
  reasoning: true;
9507
- input: ("image" | "text")[];
10182
+ input: "text"[];
9508
10183
  cost: {
9509
10184
  input: number;
9510
10185
  output: number;
@@ -9514,7 +10189,7 @@ export declare const MODELS: {
9514
10189
  contextWindow: number;
9515
10190
  maxTokens: number;
9516
10191
  };
9517
- readonly "openai/gpt-5.2-pro": {
10192
+ readonly "openai/o1": {
9518
10193
  id: string;
9519
10194
  name: string;
9520
10195
  api: "openai-completions";
@@ -9531,13 +10206,13 @@ export declare const MODELS: {
9531
10206
  contextWindow: number;
9532
10207
  maxTokens: number;
9533
10208
  };
9534
- readonly "openai/gpt-5.3-chat": {
10209
+ readonly "openai/o3": {
9535
10210
  id: string;
9536
10211
  name: string;
9537
10212
  api: "openai-completions";
9538
10213
  provider: string;
9539
10214
  baseUrl: string;
9540
- reasoning: false;
10215
+ reasoning: true;
9541
10216
  input: ("image" | "text")[];
9542
10217
  cost: {
9543
10218
  input: number;
@@ -9548,7 +10223,7 @@ export declare const MODELS: {
9548
10223
  contextWindow: number;
9549
10224
  maxTokens: number;
9550
10225
  };
9551
- readonly "openai/gpt-5.3-codex": {
10226
+ readonly "openai/o3-deep-research": {
9552
10227
  id: string;
9553
10228
  name: string;
9554
10229
  api: "openai-completions";
@@ -9565,14 +10240,14 @@ export declare const MODELS: {
9565
10240
  contextWindow: number;
9566
10241
  maxTokens: number;
9567
10242
  };
9568
- readonly "openai/gpt-5.4": {
10243
+ readonly "openai/o3-mini": {
9569
10244
  id: string;
9570
10245
  name: string;
9571
10246
  api: "openai-completions";
9572
10247
  provider: string;
9573
10248
  baseUrl: string;
9574
10249
  reasoning: true;
9575
- input: ("image" | "text")[];
10250
+ input: "text"[];
9576
10251
  cost: {
9577
10252
  input: number;
9578
10253
  output: number;
@@ -9582,14 +10257,14 @@ export declare const MODELS: {
9582
10257
  contextWindow: number;
9583
10258
  maxTokens: number;
9584
10259
  };
9585
- readonly "openai/gpt-5.4-mini": {
10260
+ readonly "openai/o3-mini-high": {
9586
10261
  id: string;
9587
10262
  name: string;
9588
10263
  api: "openai-completions";
9589
10264
  provider: string;
9590
10265
  baseUrl: string;
9591
10266
  reasoning: true;
9592
- input: ("image" | "text")[];
10267
+ input: "text"[];
9593
10268
  cost: {
9594
10269
  input: number;
9595
10270
  output: number;
@@ -9599,7 +10274,7 @@ export declare const MODELS: {
9599
10274
  contextWindow: number;
9600
10275
  maxTokens: number;
9601
10276
  };
9602
- readonly "openai/gpt-5.4-nano": {
10277
+ readonly "openai/o3-pro": {
9603
10278
  id: string;
9604
10279
  name: string;
9605
10280
  api: "openai-completions";
@@ -9616,7 +10291,7 @@ export declare const MODELS: {
9616
10291
  contextWindow: number;
9617
10292
  maxTokens: number;
9618
10293
  };
9619
- readonly "openai/gpt-5.4-pro": {
10294
+ readonly "openai/o4-mini": {
9620
10295
  id: string;
9621
10296
  name: string;
9622
10297
  api: "openai-completions";
@@ -9633,14 +10308,14 @@ export declare const MODELS: {
9633
10308
  contextWindow: number;
9634
10309
  maxTokens: number;
9635
10310
  };
9636
- readonly "openai/gpt-audio": {
10311
+ readonly "openai/o4-mini-deep-research": {
9637
10312
  id: string;
9638
10313
  name: string;
9639
10314
  api: "openai-completions";
9640
10315
  provider: string;
9641
10316
  baseUrl: string;
9642
- reasoning: false;
9643
- input: "text"[];
10317
+ reasoning: true;
10318
+ input: ("image" | "text")[];
9644
10319
  cost: {
9645
10320
  input: number;
9646
10321
  output: number;
@@ -9650,14 +10325,14 @@ export declare const MODELS: {
9650
10325
  contextWindow: number;
9651
10326
  maxTokens: number;
9652
10327
  };
9653
- readonly "openai/gpt-audio-mini": {
10328
+ readonly "openai/o4-mini-high": {
9654
10329
  id: string;
9655
10330
  name: string;
9656
10331
  api: "openai-completions";
9657
10332
  provider: string;
9658
10333
  baseUrl: string;
9659
- reasoning: false;
9660
- input: "text"[];
10334
+ reasoning: true;
10335
+ input: ("image" | "text")[];
9661
10336
  cost: {
9662
10337
  input: number;
9663
10338
  output: number;
@@ -9667,14 +10342,14 @@ export declare const MODELS: {
9667
10342
  contextWindow: number;
9668
10343
  maxTokens: number;
9669
10344
  };
9670
- readonly "openai/gpt-oss-120b": {
10345
+ readonly "openrouter/auto": {
9671
10346
  id: string;
9672
10347
  name: string;
9673
10348
  api: "openai-completions";
9674
10349
  provider: string;
9675
10350
  baseUrl: string;
9676
10351
  reasoning: true;
9677
- input: "text"[];
10352
+ input: ("image" | "text")[];
9678
10353
  cost: {
9679
10354
  input: number;
9680
10355
  output: number;
@@ -9684,14 +10359,14 @@ export declare const MODELS: {
9684
10359
  contextWindow: number;
9685
10360
  maxTokens: number;
9686
10361
  };
9687
- readonly "openai/gpt-oss-120b:free": {
10362
+ readonly "openrouter/free": {
9688
10363
  id: string;
9689
10364
  name: string;
9690
10365
  api: "openai-completions";
9691
10366
  provider: string;
9692
10367
  baseUrl: string;
9693
10368
  reasoning: true;
9694
- input: "text"[];
10369
+ input: ("image" | "text")[];
9695
10370
  cost: {
9696
10371
  input: number;
9697
10372
  output: number;
@@ -9701,7 +10376,7 @@ export declare const MODELS: {
9701
10376
  contextWindow: number;
9702
10377
  maxTokens: number;
9703
10378
  };
9704
- readonly "openai/gpt-oss-20b": {
10379
+ readonly "prime-intellect/intellect-3": {
9705
10380
  id: string;
9706
10381
  name: string;
9707
10382
  api: "openai-completions";
@@ -9718,13 +10393,13 @@ export declare const MODELS: {
9718
10393
  contextWindow: number;
9719
10394
  maxTokens: number;
9720
10395
  };
9721
- readonly "openai/gpt-oss-20b:free": {
10396
+ readonly "qwen/qwen-2.5-72b-instruct": {
9722
10397
  id: string;
9723
10398
  name: string;
9724
10399
  api: "openai-completions";
9725
10400
  provider: string;
9726
10401
  baseUrl: string;
9727
- reasoning: true;
10402
+ reasoning: false;
9728
10403
  input: "text"[];
9729
10404
  cost: {
9730
10405
  input: number;
@@ -9735,13 +10410,13 @@ export declare const MODELS: {
9735
10410
  contextWindow: number;
9736
10411
  maxTokens: number;
9737
10412
  };
9738
- readonly "openai/gpt-oss-safeguard-20b": {
10413
+ readonly "qwen/qwen-2.5-7b-instruct": {
9739
10414
  id: string;
9740
10415
  name: string;
9741
10416
  api: "openai-completions";
9742
10417
  provider: string;
9743
10418
  baseUrl: string;
9744
- reasoning: true;
10419
+ reasoning: false;
9745
10420
  input: "text"[];
9746
10421
  cost: {
9747
10422
  input: number;
@@ -9752,14 +10427,14 @@ export declare const MODELS: {
9752
10427
  contextWindow: number;
9753
10428
  maxTokens: number;
9754
10429
  };
9755
- readonly "openai/o1": {
10430
+ readonly "qwen/qwen-max": {
9756
10431
  id: string;
9757
10432
  name: string;
9758
10433
  api: "openai-completions";
9759
10434
  provider: string;
9760
10435
  baseUrl: string;
9761
- reasoning: true;
9762
- input: ("image" | "text")[];
10436
+ reasoning: false;
10437
+ input: "text"[];
9763
10438
  cost: {
9764
10439
  input: number;
9765
10440
  output: number;
@@ -9769,14 +10444,14 @@ export declare const MODELS: {
9769
10444
  contextWindow: number;
9770
10445
  maxTokens: number;
9771
10446
  };
9772
- readonly "openai/o3": {
10447
+ readonly "qwen/qwen-plus": {
9773
10448
  id: string;
9774
10449
  name: string;
9775
10450
  api: "openai-completions";
9776
10451
  provider: string;
9777
10452
  baseUrl: string;
9778
- reasoning: true;
9779
- input: ("image" | "text")[];
10453
+ reasoning: false;
10454
+ input: "text"[];
9780
10455
  cost: {
9781
10456
  input: number;
9782
10457
  output: number;
@@ -9786,14 +10461,14 @@ export declare const MODELS: {
9786
10461
  contextWindow: number;
9787
10462
  maxTokens: number;
9788
10463
  };
9789
- readonly "openai/o3-deep-research": {
10464
+ readonly "qwen/qwen-plus-2025-07-28": {
9790
10465
  id: string;
9791
10466
  name: string;
9792
10467
  api: "openai-completions";
9793
10468
  provider: string;
9794
10469
  baseUrl: string;
9795
- reasoning: true;
9796
- input: ("image" | "text")[];
10470
+ reasoning: false;
10471
+ input: "text"[];
9797
10472
  cost: {
9798
10473
  input: number;
9799
10474
  output: number;
@@ -9803,7 +10478,7 @@ export declare const MODELS: {
9803
10478
  contextWindow: number;
9804
10479
  maxTokens: number;
9805
10480
  };
9806
- readonly "openai/o3-mini": {
10481
+ readonly "qwen/qwen-plus-2025-07-28:thinking": {
9807
10482
  id: string;
9808
10483
  name: string;
9809
10484
  api: "openai-completions";
@@ -9820,13 +10495,13 @@ export declare const MODELS: {
9820
10495
  contextWindow: number;
9821
10496
  maxTokens: number;
9822
10497
  };
9823
- readonly "openai/o3-mini-high": {
10498
+ readonly "qwen/qwen-turbo": {
9824
10499
  id: string;
9825
10500
  name: string;
9826
10501
  api: "openai-completions";
9827
10502
  provider: string;
9828
10503
  baseUrl: string;
9829
- reasoning: true;
10504
+ reasoning: false;
9830
10505
  input: "text"[];
9831
10506
  cost: {
9832
10507
  input: number;
@@ -9837,13 +10512,13 @@ export declare const MODELS: {
9837
10512
  contextWindow: number;
9838
10513
  maxTokens: number;
9839
10514
  };
9840
- readonly "openai/o3-pro": {
10515
+ readonly "qwen/qwen-vl-max": {
9841
10516
  id: string;
9842
10517
  name: string;
9843
10518
  api: "openai-completions";
9844
10519
  provider: string;
9845
10520
  baseUrl: string;
9846
- reasoning: true;
10521
+ reasoning: false;
9847
10522
  input: ("image" | "text")[];
9848
10523
  cost: {
9849
10524
  input: number;
@@ -9854,14 +10529,14 @@ export declare const MODELS: {
9854
10529
  contextWindow: number;
9855
10530
  maxTokens: number;
9856
10531
  };
9857
- readonly "openai/o4-mini": {
10532
+ readonly "qwen/qwen3-14b": {
9858
10533
  id: string;
9859
10534
  name: string;
9860
10535
  api: "openai-completions";
9861
10536
  provider: string;
9862
10537
  baseUrl: string;
9863
10538
  reasoning: true;
9864
- input: ("image" | "text")[];
10539
+ input: "text"[];
9865
10540
  cost: {
9866
10541
  input: number;
9867
10542
  output: number;
@@ -9871,14 +10546,14 @@ export declare const MODELS: {
9871
10546
  contextWindow: number;
9872
10547
  maxTokens: number;
9873
10548
  };
9874
- readonly "openai/o4-mini-deep-research": {
10549
+ readonly "qwen/qwen3-235b-a22b": {
9875
10550
  id: string;
9876
10551
  name: string;
9877
10552
  api: "openai-completions";
9878
10553
  provider: string;
9879
10554
  baseUrl: string;
9880
10555
  reasoning: true;
9881
- input: ("image" | "text")[];
10556
+ input: "text"[];
9882
10557
  cost: {
9883
10558
  input: number;
9884
10559
  output: number;
@@ -9888,14 +10563,14 @@ export declare const MODELS: {
9888
10563
  contextWindow: number;
9889
10564
  maxTokens: number;
9890
10565
  };
9891
- readonly "openai/o4-mini-high": {
10566
+ readonly "qwen/qwen3-235b-a22b-2507": {
9892
10567
  id: string;
9893
10568
  name: string;
9894
10569
  api: "openai-completions";
9895
10570
  provider: string;
9896
10571
  baseUrl: string;
9897
- reasoning: true;
9898
- input: ("image" | "text")[];
10572
+ reasoning: false;
10573
+ input: "text"[];
9899
10574
  cost: {
9900
10575
  input: number;
9901
10576
  output: number;
@@ -9905,14 +10580,14 @@ export declare const MODELS: {
9905
10580
  contextWindow: number;
9906
10581
  maxTokens: number;
9907
10582
  };
9908
- readonly "openrouter/auto": {
10583
+ readonly "qwen/qwen3-235b-a22b-thinking-2507": {
9909
10584
  id: string;
9910
10585
  name: string;
9911
10586
  api: "openai-completions";
9912
10587
  provider: string;
9913
10588
  baseUrl: string;
9914
10589
  reasoning: true;
9915
- input: ("image" | "text")[];
10590
+ input: "text"[];
9916
10591
  cost: {
9917
10592
  input: number;
9918
10593
  output: number;
@@ -9922,13 +10597,13 @@ export declare const MODELS: {
9922
10597
  contextWindow: number;
9923
10598
  maxTokens: number;
9924
10599
  };
9925
- readonly "openrouter/elephant-alpha": {
10600
+ readonly "qwen/qwen3-30b-a3b": {
9926
10601
  id: string;
9927
10602
  name: string;
9928
10603
  api: "openai-completions";
9929
10604
  provider: string;
9930
10605
  baseUrl: string;
9931
- reasoning: false;
10606
+ reasoning: true;
9932
10607
  input: "text"[];
9933
10608
  cost: {
9934
10609
  input: number;
@@ -9939,14 +10614,14 @@ export declare const MODELS: {
9939
10614
  contextWindow: number;
9940
10615
  maxTokens: number;
9941
10616
  };
9942
- readonly "openrouter/free": {
10617
+ readonly "qwen/qwen3-30b-a3b-instruct-2507": {
9943
10618
  id: string;
9944
10619
  name: string;
9945
10620
  api: "openai-completions";
9946
10621
  provider: string;
9947
10622
  baseUrl: string;
9948
- reasoning: true;
9949
- input: ("image" | "text")[];
10623
+ reasoning: false;
10624
+ input: "text"[];
9950
10625
  cost: {
9951
10626
  input: number;
9952
10627
  output: number;
@@ -9956,7 +10631,7 @@ export declare const MODELS: {
9956
10631
  contextWindow: number;
9957
10632
  maxTokens: number;
9958
10633
  };
9959
- readonly "prime-intellect/intellect-3": {
10634
+ readonly "qwen/qwen3-30b-a3b-thinking-2507": {
9960
10635
  id: string;
9961
10636
  name: string;
9962
10637
  api: "openai-completions";
@@ -9973,13 +10648,13 @@ export declare const MODELS: {
9973
10648
  contextWindow: number;
9974
10649
  maxTokens: number;
9975
10650
  };
9976
- readonly "qwen/qwen-2.5-72b-instruct": {
10651
+ readonly "qwen/qwen3-32b": {
9977
10652
  id: string;
9978
10653
  name: string;
9979
10654
  api: "openai-completions";
9980
10655
  provider: string;
9981
10656
  baseUrl: string;
9982
- reasoning: false;
10657
+ reasoning: true;
9983
10658
  input: "text"[];
9984
10659
  cost: {
9985
10660
  input: number;
@@ -9990,13 +10665,13 @@ export declare const MODELS: {
9990
10665
  contextWindow: number;
9991
10666
  maxTokens: number;
9992
10667
  };
9993
- readonly "qwen/qwen-2.5-7b-instruct": {
10668
+ readonly "qwen/qwen3-8b": {
9994
10669
  id: string;
9995
10670
  name: string;
9996
10671
  api: "openai-completions";
9997
10672
  provider: string;
9998
10673
  baseUrl: string;
9999
- reasoning: false;
10674
+ reasoning: true;
10000
10675
  input: "text"[];
10001
10676
  cost: {
10002
10677
  input: number;
@@ -10007,7 +10682,7 @@ export declare const MODELS: {
10007
10682
  contextWindow: number;
10008
10683
  maxTokens: number;
10009
10684
  };
10010
- readonly "qwen/qwen-max": {
10685
+ readonly "qwen/qwen3-coder": {
10011
10686
  id: string;
10012
10687
  name: string;
10013
10688
  api: "openai-completions";
@@ -10024,7 +10699,7 @@ export declare const MODELS: {
10024
10699
  contextWindow: number;
10025
10700
  maxTokens: number;
10026
10701
  };
10027
- readonly "qwen/qwen-plus": {
10702
+ readonly "qwen/qwen3-coder-30b-a3b-instruct": {
10028
10703
  id: string;
10029
10704
  name: string;
10030
10705
  api: "openai-completions";
@@ -10041,7 +10716,7 @@ export declare const MODELS: {
10041
10716
  contextWindow: number;
10042
10717
  maxTokens: number;
10043
10718
  };
10044
- readonly "qwen/qwen-plus-2025-07-28": {
10719
+ readonly "qwen/qwen3-coder-flash": {
10045
10720
  id: string;
10046
10721
  name: string;
10047
10722
  api: "openai-completions";
@@ -10058,13 +10733,13 @@ export declare const MODELS: {
10058
10733
  contextWindow: number;
10059
10734
  maxTokens: number;
10060
10735
  };
10061
- readonly "qwen/qwen-plus-2025-07-28:thinking": {
10736
+ readonly "qwen/qwen3-coder-next": {
10062
10737
  id: string;
10063
10738
  name: string;
10064
10739
  api: "openai-completions";
10065
10740
  provider: string;
10066
10741
  baseUrl: string;
10067
- reasoning: true;
10742
+ reasoning: false;
10068
10743
  input: "text"[];
10069
10744
  cost: {
10070
10745
  input: number;
@@ -10075,7 +10750,7 @@ export declare const MODELS: {
10075
10750
  contextWindow: number;
10076
10751
  maxTokens: number;
10077
10752
  };
10078
- readonly "qwen/qwen-turbo": {
10753
+ readonly "qwen/qwen3-coder-plus": {
10079
10754
  id: string;
10080
10755
  name: string;
10081
10756
  api: "openai-completions";
@@ -10092,14 +10767,14 @@ export declare const MODELS: {
10092
10767
  contextWindow: number;
10093
10768
  maxTokens: number;
10094
10769
  };
10095
- readonly "qwen/qwen-vl-max": {
10770
+ readonly "qwen/qwen3-coder:free": {
10096
10771
  id: string;
10097
10772
  name: string;
10098
10773
  api: "openai-completions";
10099
10774
  provider: string;
10100
10775
  baseUrl: string;
10101
10776
  reasoning: false;
10102
- input: ("image" | "text")[];
10777
+ input: "text"[];
10103
10778
  cost: {
10104
10779
  input: number;
10105
10780
  output: number;
@@ -10109,13 +10784,13 @@ export declare const MODELS: {
10109
10784
  contextWindow: number;
10110
10785
  maxTokens: number;
10111
10786
  };
10112
- readonly "qwen/qwen3-14b": {
10787
+ readonly "qwen/qwen3-max": {
10113
10788
  id: string;
10114
10789
  name: string;
10115
10790
  api: "openai-completions";
10116
10791
  provider: string;
10117
10792
  baseUrl: string;
10118
- reasoning: true;
10793
+ reasoning: false;
10119
10794
  input: "text"[];
10120
10795
  cost: {
10121
10796
  input: number;
@@ -10126,7 +10801,7 @@ export declare const MODELS: {
10126
10801
  contextWindow: number;
10127
10802
  maxTokens: number;
10128
10803
  };
10129
- readonly "qwen/qwen3-235b-a22b": {
10804
+ readonly "qwen/qwen3-max-thinking": {
10130
10805
  id: string;
10131
10806
  name: string;
10132
10807
  api: "openai-completions";
@@ -10143,7 +10818,7 @@ export declare const MODELS: {
10143
10818
  contextWindow: number;
10144
10819
  maxTokens: number;
10145
10820
  };
10146
- readonly "qwen/qwen3-235b-a22b-2507": {
10821
+ readonly "qwen/qwen3-next-80b-a3b-instruct": {
10147
10822
  id: string;
10148
10823
  name: string;
10149
10824
  api: "openai-completions";
@@ -10160,13 +10835,13 @@ export declare const MODELS: {
10160
10835
  contextWindow: number;
10161
10836
  maxTokens: number;
10162
10837
  };
10163
- readonly "qwen/qwen3-235b-a22b-thinking-2507": {
10838
+ readonly "qwen/qwen3-next-80b-a3b-instruct:free": {
10164
10839
  id: string;
10165
10840
  name: string;
10166
10841
  api: "openai-completions";
10167
10842
  provider: string;
10168
10843
  baseUrl: string;
10169
- reasoning: true;
10844
+ reasoning: false;
10170
10845
  input: "text"[];
10171
10846
  cost: {
10172
10847
  input: number;
@@ -10177,7 +10852,7 @@ export declare const MODELS: {
10177
10852
  contextWindow: number;
10178
10853
  maxTokens: number;
10179
10854
  };
10180
- readonly "qwen/qwen3-30b-a3b": {
10855
+ readonly "qwen/qwen3-next-80b-a3b-thinking": {
10181
10856
  id: string;
10182
10857
  name: string;
10183
10858
  api: "openai-completions";
@@ -10194,14 +10869,14 @@ export declare const MODELS: {
10194
10869
  contextWindow: number;
10195
10870
  maxTokens: number;
10196
10871
  };
10197
- readonly "qwen/qwen3-30b-a3b-instruct-2507": {
10872
+ readonly "qwen/qwen3-vl-235b-a22b-instruct": {
10198
10873
  id: string;
10199
10874
  name: string;
10200
10875
  api: "openai-completions";
10201
10876
  provider: string;
10202
10877
  baseUrl: string;
10203
10878
  reasoning: false;
10204
- input: "text"[];
10879
+ input: ("image" | "text")[];
10205
10880
  cost: {
10206
10881
  input: number;
10207
10882
  output: number;
@@ -10211,14 +10886,14 @@ export declare const MODELS: {
10211
10886
  contextWindow: number;
10212
10887
  maxTokens: number;
10213
10888
  };
10214
- readonly "qwen/qwen3-30b-a3b-thinking-2507": {
10889
+ readonly "qwen/qwen3-vl-235b-a22b-thinking": {
10215
10890
  id: string;
10216
10891
  name: string;
10217
10892
  api: "openai-completions";
10218
10893
  provider: string;
10219
10894
  baseUrl: string;
10220
10895
  reasoning: true;
10221
- input: "text"[];
10896
+ input: ("image" | "text")[];
10222
10897
  cost: {
10223
10898
  input: number;
10224
10899
  output: number;
@@ -10228,14 +10903,14 @@ export declare const MODELS: {
10228
10903
  contextWindow: number;
10229
10904
  maxTokens: number;
10230
10905
  };
10231
- readonly "qwen/qwen3-32b": {
10906
+ readonly "qwen/qwen3-vl-30b-a3b-instruct": {
10232
10907
  id: string;
10233
10908
  name: string;
10234
10909
  api: "openai-completions";
10235
10910
  provider: string;
10236
10911
  baseUrl: string;
10237
- reasoning: true;
10238
- input: "text"[];
10912
+ reasoning: false;
10913
+ input: ("image" | "text")[];
10239
10914
  cost: {
10240
10915
  input: number;
10241
10916
  output: number;
@@ -10245,14 +10920,14 @@ export declare const MODELS: {
10245
10920
  contextWindow: number;
10246
10921
  maxTokens: number;
10247
10922
  };
10248
- readonly "qwen/qwen3-8b": {
10923
+ readonly "qwen/qwen3-vl-30b-a3b-thinking": {
10249
10924
  id: string;
10250
10925
  name: string;
10251
10926
  api: "openai-completions";
10252
10927
  provider: string;
10253
10928
  baseUrl: string;
10254
10929
  reasoning: true;
10255
- input: "text"[];
10930
+ input: ("image" | "text")[];
10256
10931
  cost: {
10257
10932
  input: number;
10258
10933
  output: number;
@@ -10262,14 +10937,14 @@ export declare const MODELS: {
10262
10937
  contextWindow: number;
10263
10938
  maxTokens: number;
10264
10939
  };
10265
- readonly "qwen/qwen3-coder": {
10940
+ readonly "qwen/qwen3-vl-32b-instruct": {
10266
10941
  id: string;
10267
10942
  name: string;
10268
10943
  api: "openai-completions";
10269
10944
  provider: string;
10270
10945
  baseUrl: string;
10271
10946
  reasoning: false;
10272
- input: "text"[];
10947
+ input: ("image" | "text")[];
10273
10948
  cost: {
10274
10949
  input: number;
10275
10950
  output: number;
@@ -10279,14 +10954,14 @@ export declare const MODELS: {
10279
10954
  contextWindow: number;
10280
10955
  maxTokens: number;
10281
10956
  };
10282
- readonly "qwen/qwen3-coder-30b-a3b-instruct": {
10957
+ readonly "qwen/qwen3-vl-8b-instruct": {
10283
10958
  id: string;
10284
10959
  name: string;
10285
10960
  api: "openai-completions";
10286
10961
  provider: string;
10287
10962
  baseUrl: string;
10288
10963
  reasoning: false;
10289
- input: "text"[];
10964
+ input: ("image" | "text")[];
10290
10965
  cost: {
10291
10966
  input: number;
10292
10967
  output: number;
@@ -10296,14 +10971,14 @@ export declare const MODELS: {
10296
10971
  contextWindow: number;
10297
10972
  maxTokens: number;
10298
10973
  };
10299
- readonly "qwen/qwen3-coder-flash": {
10974
+ readonly "qwen/qwen3-vl-8b-thinking": {
10300
10975
  id: string;
10301
10976
  name: string;
10302
10977
  api: "openai-completions";
10303
10978
  provider: string;
10304
10979
  baseUrl: string;
10305
- reasoning: false;
10306
- input: "text"[];
10980
+ reasoning: true;
10981
+ input: ("image" | "text")[];
10307
10982
  cost: {
10308
10983
  input: number;
10309
10984
  output: number;
@@ -10313,14 +10988,14 @@ export declare const MODELS: {
10313
10988
  contextWindow: number;
10314
10989
  maxTokens: number;
10315
10990
  };
10316
- readonly "qwen/qwen3-coder-next": {
10991
+ readonly "qwen/qwen3.5-122b-a10b": {
10317
10992
  id: string;
10318
10993
  name: string;
10319
10994
  api: "openai-completions";
10320
10995
  provider: string;
10321
10996
  baseUrl: string;
10322
- reasoning: false;
10323
- input: "text"[];
10997
+ reasoning: true;
10998
+ input: ("image" | "text")[];
10324
10999
  cost: {
10325
11000
  input: number;
10326
11001
  output: number;
@@ -10330,14 +11005,14 @@ export declare const MODELS: {
10330
11005
  contextWindow: number;
10331
11006
  maxTokens: number;
10332
11007
  };
10333
- readonly "qwen/qwen3-coder-plus": {
11008
+ readonly "qwen/qwen3.5-27b": {
10334
11009
  id: string;
10335
11010
  name: string;
10336
11011
  api: "openai-completions";
10337
11012
  provider: string;
10338
11013
  baseUrl: string;
10339
- reasoning: false;
10340
- input: "text"[];
11014
+ reasoning: true;
11015
+ input: ("image" | "text")[];
10341
11016
  cost: {
10342
11017
  input: number;
10343
11018
  output: number;
@@ -10347,14 +11022,14 @@ export declare const MODELS: {
10347
11022
  contextWindow: number;
10348
11023
  maxTokens: number;
10349
11024
  };
10350
- readonly "qwen/qwen3-coder:free": {
11025
+ readonly "qwen/qwen3.5-35b-a3b": {
10351
11026
  id: string;
10352
11027
  name: string;
10353
11028
  api: "openai-completions";
10354
11029
  provider: string;
10355
11030
  baseUrl: string;
10356
- reasoning: false;
10357
- input: "text"[];
11031
+ reasoning: true;
11032
+ input: ("image" | "text")[];
10358
11033
  cost: {
10359
11034
  input: number;
10360
11035
  output: number;
@@ -10364,14 +11039,14 @@ export declare const MODELS: {
10364
11039
  contextWindow: number;
10365
11040
  maxTokens: number;
10366
11041
  };
10367
- readonly "qwen/qwen3-max": {
11042
+ readonly "qwen/qwen3.5-397b-a17b": {
10368
11043
  id: string;
10369
11044
  name: string;
10370
11045
  api: "openai-completions";
10371
11046
  provider: string;
10372
11047
  baseUrl: string;
10373
- reasoning: false;
10374
- input: "text"[];
11048
+ reasoning: true;
11049
+ input: ("image" | "text")[];
10375
11050
  cost: {
10376
11051
  input: number;
10377
11052
  output: number;
@@ -10381,14 +11056,14 @@ export declare const MODELS: {
10381
11056
  contextWindow: number;
10382
11057
  maxTokens: number;
10383
11058
  };
10384
- readonly "qwen/qwen3-max-thinking": {
11059
+ readonly "qwen/qwen3.5-9b": {
10385
11060
  id: string;
10386
11061
  name: string;
10387
11062
  api: "openai-completions";
10388
11063
  provider: string;
10389
11064
  baseUrl: string;
10390
11065
  reasoning: true;
10391
- input: "text"[];
11066
+ input: ("image" | "text")[];
10392
11067
  cost: {
10393
11068
  input: number;
10394
11069
  output: number;
@@ -10398,14 +11073,14 @@ export declare const MODELS: {
10398
11073
  contextWindow: number;
10399
11074
  maxTokens: number;
10400
11075
  };
10401
- readonly "qwen/qwen3-next-80b-a3b-instruct": {
11076
+ readonly "qwen/qwen3.5-flash-02-23": {
10402
11077
  id: string;
10403
11078
  name: string;
10404
11079
  api: "openai-completions";
10405
11080
  provider: string;
10406
11081
  baseUrl: string;
10407
- reasoning: false;
10408
- input: "text"[];
11082
+ reasoning: true;
11083
+ input: ("image" | "text")[];
10409
11084
  cost: {
10410
11085
  input: number;
10411
11086
  output: number;
@@ -10415,14 +11090,14 @@ export declare const MODELS: {
10415
11090
  contextWindow: number;
10416
11091
  maxTokens: number;
10417
11092
  };
10418
- readonly "qwen/qwen3-next-80b-a3b-instruct:free": {
11093
+ readonly "qwen/qwen3.5-plus-02-15": {
10419
11094
  id: string;
10420
11095
  name: string;
10421
11096
  api: "openai-completions";
10422
11097
  provider: string;
10423
11098
  baseUrl: string;
10424
- reasoning: false;
10425
- input: "text"[];
11099
+ reasoning: true;
11100
+ input: ("image" | "text")[];
10426
11101
  cost: {
10427
11102
  input: number;
10428
11103
  output: number;
@@ -10432,14 +11107,14 @@ export declare const MODELS: {
10432
11107
  contextWindow: number;
10433
11108
  maxTokens: number;
10434
11109
  };
10435
- readonly "qwen/qwen3-next-80b-a3b-thinking": {
11110
+ readonly "qwen/qwen3.5-plus-20260420": {
10436
11111
  id: string;
10437
11112
  name: string;
10438
11113
  api: "openai-completions";
10439
11114
  provider: string;
10440
11115
  baseUrl: string;
10441
11116
  reasoning: true;
10442
- input: "text"[];
11117
+ input: ("image" | "text")[];
10443
11118
  cost: {
10444
11119
  input: number;
10445
11120
  output: number;
@@ -10449,13 +11124,13 @@ export declare const MODELS: {
10449
11124
  contextWindow: number;
10450
11125
  maxTokens: number;
10451
11126
  };
10452
- readonly "qwen/qwen3-vl-235b-a22b-instruct": {
11127
+ readonly "qwen/qwen3.6-27b": {
10453
11128
  id: string;
10454
11129
  name: string;
10455
11130
  api: "openai-completions";
10456
11131
  provider: string;
10457
11132
  baseUrl: string;
10458
- reasoning: false;
11133
+ reasoning: true;
10459
11134
  input: ("image" | "text")[];
10460
11135
  cost: {
10461
11136
  input: number;
@@ -10466,7 +11141,7 @@ export declare const MODELS: {
10466
11141
  contextWindow: number;
10467
11142
  maxTokens: number;
10468
11143
  };
10469
- readonly "qwen/qwen3-vl-235b-a22b-thinking": {
11144
+ readonly "qwen/qwen3.6-flash": {
10470
11145
  id: string;
10471
11146
  name: string;
10472
11147
  api: "openai-completions";
@@ -10483,14 +11158,14 @@ export declare const MODELS: {
10483
11158
  contextWindow: number;
10484
11159
  maxTokens: number;
10485
11160
  };
10486
- readonly "qwen/qwen3-vl-30b-a3b-instruct": {
11161
+ readonly "qwen/qwen3.6-max-preview": {
10487
11162
  id: string;
10488
11163
  name: string;
10489
11164
  api: "openai-completions";
10490
11165
  provider: string;
10491
11166
  baseUrl: string;
10492
- reasoning: false;
10493
- input: ("image" | "text")[];
11167
+ reasoning: true;
11168
+ input: "text"[];
10494
11169
  cost: {
10495
11170
  input: number;
10496
11171
  output: number;
@@ -10500,7 +11175,7 @@ export declare const MODELS: {
10500
11175
  contextWindow: number;
10501
11176
  maxTokens: number;
10502
11177
  };
10503
- readonly "qwen/qwen3-vl-30b-a3b-thinking": {
11178
+ readonly "qwen/qwen3.6-plus": {
10504
11179
  id: string;
10505
11180
  name: string;
10506
11181
  api: "openai-completions";
@@ -10517,14 +11192,14 @@ export declare const MODELS: {
10517
11192
  contextWindow: number;
10518
11193
  maxTokens: number;
10519
11194
  };
10520
- readonly "qwen/qwen3-vl-32b-instruct": {
11195
+ readonly "qwen/qwq-32b": {
10521
11196
  id: string;
10522
11197
  name: string;
10523
11198
  api: "openai-completions";
10524
11199
  provider: string;
10525
11200
  baseUrl: string;
10526
- reasoning: false;
10527
- input: ("image" | "text")[];
11201
+ reasoning: true;
11202
+ input: "text"[];
10528
11203
  cost: {
10529
11204
  input: number;
10530
11205
  output: number;
@@ -10534,7 +11209,7 @@ export declare const MODELS: {
10534
11209
  contextWindow: number;
10535
11210
  maxTokens: number;
10536
11211
  };
10537
- readonly "qwen/qwen3-vl-8b-instruct": {
11212
+ readonly "rekaai/reka-edge": {
10538
11213
  id: string;
10539
11214
  name: string;
10540
11215
  api: "openai-completions";
@@ -10551,14 +11226,14 @@ export declare const MODELS: {
10551
11226
  contextWindow: number;
10552
11227
  maxTokens: number;
10553
11228
  };
10554
- readonly "qwen/qwen3-vl-8b-thinking": {
11229
+ readonly "relace/relace-search": {
10555
11230
  id: string;
10556
11231
  name: string;
10557
11232
  api: "openai-completions";
10558
11233
  provider: string;
10559
11234
  baseUrl: string;
10560
- reasoning: true;
10561
- input: ("image" | "text")[];
11235
+ reasoning: false;
11236
+ input: "text"[];
10562
11237
  cost: {
10563
11238
  input: number;
10564
11239
  output: number;
@@ -10568,14 +11243,14 @@ export declare const MODELS: {
10568
11243
  contextWindow: number;
10569
11244
  maxTokens: number;
10570
11245
  };
10571
- readonly "qwen/qwen3.5-122b-a10b": {
11246
+ readonly "sao10k/l3-euryale-70b": {
10572
11247
  id: string;
10573
11248
  name: string;
10574
11249
  api: "openai-completions";
10575
11250
  provider: string;
10576
11251
  baseUrl: string;
10577
- reasoning: true;
10578
- input: ("image" | "text")[];
11252
+ reasoning: false;
11253
+ input: "text"[];
10579
11254
  cost: {
10580
11255
  input: number;
10581
11256
  output: number;
@@ -10585,14 +11260,14 @@ export declare const MODELS: {
10585
11260
  contextWindow: number;
10586
11261
  maxTokens: number;
10587
11262
  };
10588
- readonly "qwen/qwen3.5-27b": {
11263
+ readonly "sao10k/l3.1-euryale-70b": {
10589
11264
  id: string;
10590
11265
  name: string;
10591
11266
  api: "openai-completions";
10592
11267
  provider: string;
10593
11268
  baseUrl: string;
10594
- reasoning: true;
10595
- input: ("image" | "text")[];
11269
+ reasoning: false;
11270
+ input: "text"[];
10596
11271
  cost: {
10597
11272
  input: number;
10598
11273
  output: number;
@@ -10602,14 +11277,14 @@ export declare const MODELS: {
10602
11277
  contextWindow: number;
10603
11278
  maxTokens: number;
10604
11279
  };
10605
- readonly "qwen/qwen3.5-35b-a3b": {
11280
+ readonly "stepfun/step-3.5-flash": {
10606
11281
  id: string;
10607
11282
  name: string;
10608
11283
  api: "openai-completions";
10609
11284
  provider: string;
10610
11285
  baseUrl: string;
10611
11286
  reasoning: true;
10612
- input: ("image" | "text")[];
11287
+ input: "text"[];
10613
11288
  cost: {
10614
11289
  input: number;
10615
11290
  output: number;
@@ -10619,14 +11294,14 @@ export declare const MODELS: {
10619
11294
  contextWindow: number;
10620
11295
  maxTokens: number;
10621
11296
  };
10622
- readonly "qwen/qwen3.5-397b-a17b": {
11297
+ readonly "tencent/hy3-preview:free": {
10623
11298
  id: string;
10624
11299
  name: string;
10625
11300
  api: "openai-completions";
10626
11301
  provider: string;
10627
11302
  baseUrl: string;
10628
11303
  reasoning: true;
10629
- input: ("image" | "text")[];
11304
+ input: "text"[];
10630
11305
  cost: {
10631
11306
  input: number;
10632
11307
  output: number;
@@ -10636,14 +11311,14 @@ export declare const MODELS: {
10636
11311
  contextWindow: number;
10637
11312
  maxTokens: number;
10638
11313
  };
10639
- readonly "qwen/qwen3.5-9b": {
11314
+ readonly "thedrummer/rocinante-12b": {
10640
11315
  id: string;
10641
11316
  name: string;
10642
11317
  api: "openai-completions";
10643
11318
  provider: string;
10644
11319
  baseUrl: string;
10645
- reasoning: true;
10646
- input: ("image" | "text")[];
11320
+ reasoning: false;
11321
+ input: "text"[];
10647
11322
  cost: {
10648
11323
  input: number;
10649
11324
  output: number;
@@ -10653,14 +11328,14 @@ export declare const MODELS: {
10653
11328
  contextWindow: number;
10654
11329
  maxTokens: number;
10655
11330
  };
10656
- readonly "qwen/qwen3.5-flash-02-23": {
11331
+ readonly "thedrummer/unslopnemo-12b": {
10657
11332
  id: string;
10658
11333
  name: string;
10659
11334
  api: "openai-completions";
10660
11335
  provider: string;
10661
11336
  baseUrl: string;
10662
- reasoning: true;
10663
- input: ("image" | "text")[];
11337
+ reasoning: false;
11338
+ input: "text"[];
10664
11339
  cost: {
10665
11340
  input: number;
10666
11341
  output: number;
@@ -10670,14 +11345,14 @@ export declare const MODELS: {
10670
11345
  contextWindow: number;
10671
11346
  maxTokens: number;
10672
11347
  };
10673
- readonly "qwen/qwen3.5-plus-02-15": {
11348
+ readonly "tngtech/deepseek-r1t2-chimera": {
10674
11349
  id: string;
10675
11350
  name: string;
10676
11351
  api: "openai-completions";
10677
11352
  provider: string;
10678
11353
  baseUrl: string;
10679
11354
  reasoning: true;
10680
- input: ("image" | "text")[];
11355
+ input: "text"[];
10681
11356
  cost: {
10682
11357
  input: number;
10683
11358
  output: number;
@@ -10687,14 +11362,14 @@ export declare const MODELS: {
10687
11362
  contextWindow: number;
10688
11363
  maxTokens: number;
10689
11364
  };
10690
- readonly "qwen/qwen3.6-plus": {
11365
+ readonly "upstage/solar-pro-3": {
10691
11366
  id: string;
10692
11367
  name: string;
10693
11368
  api: "openai-completions";
10694
11369
  provider: string;
10695
11370
  baseUrl: string;
10696
11371
  reasoning: true;
10697
- input: ("image" | "text")[];
11372
+ input: "text"[];
10698
11373
  cost: {
10699
11374
  input: number;
10700
11375
  output: number;
@@ -10704,13 +11379,13 @@ export declare const MODELS: {
10704
11379
  contextWindow: number;
10705
11380
  maxTokens: number;
10706
11381
  };
10707
- readonly "qwen/qwq-32b": {
11382
+ readonly "x-ai/grok-3": {
10708
11383
  id: string;
10709
11384
  name: string;
10710
11385
  api: "openai-completions";
10711
11386
  provider: string;
10712
11387
  baseUrl: string;
10713
- reasoning: true;
11388
+ reasoning: false;
10714
11389
  input: "text"[];
10715
11390
  cost: {
10716
11391
  input: number;
@@ -10721,14 +11396,14 @@ export declare const MODELS: {
10721
11396
  contextWindow: number;
10722
11397
  maxTokens: number;
10723
11398
  };
10724
- readonly "rekaai/reka-edge": {
11399
+ readonly "x-ai/grok-3-beta": {
10725
11400
  id: string;
10726
11401
  name: string;
10727
11402
  api: "openai-completions";
10728
11403
  provider: string;
10729
11404
  baseUrl: string;
10730
11405
  reasoning: false;
10731
- input: ("image" | "text")[];
11406
+ input: "text"[];
10732
11407
  cost: {
10733
11408
  input: number;
10734
11409
  output: number;
@@ -10738,13 +11413,13 @@ export declare const MODELS: {
10738
11413
  contextWindow: number;
10739
11414
  maxTokens: number;
10740
11415
  };
10741
- readonly "relace/relace-search": {
11416
+ readonly "x-ai/grok-3-mini": {
10742
11417
  id: string;
10743
11418
  name: string;
10744
11419
  api: "openai-completions";
10745
11420
  provider: string;
10746
11421
  baseUrl: string;
10747
- reasoning: false;
11422
+ reasoning: true;
10748
11423
  input: "text"[];
10749
11424
  cost: {
10750
11425
  input: number;
@@ -10755,13 +11430,13 @@ export declare const MODELS: {
10755
11430
  contextWindow: number;
10756
11431
  maxTokens: number;
10757
11432
  };
10758
- readonly "sao10k/l3-euryale-70b": {
11433
+ readonly "x-ai/grok-3-mini-beta": {
10759
11434
  id: string;
10760
11435
  name: string;
10761
11436
  api: "openai-completions";
10762
11437
  provider: string;
10763
11438
  baseUrl: string;
10764
- reasoning: false;
11439
+ reasoning: true;
10765
11440
  input: "text"[];
10766
11441
  cost: {
10767
11442
  input: number;
@@ -10772,14 +11447,14 @@ export declare const MODELS: {
10772
11447
  contextWindow: number;
10773
11448
  maxTokens: number;
10774
11449
  };
10775
- readonly "sao10k/l3.1-euryale-70b": {
11450
+ readonly "x-ai/grok-4": {
10776
11451
  id: string;
10777
11452
  name: string;
10778
11453
  api: "openai-completions";
10779
11454
  provider: string;
10780
11455
  baseUrl: string;
10781
- reasoning: false;
10782
- input: "text"[];
11456
+ reasoning: true;
11457
+ input: ("image" | "text")[];
10783
11458
  cost: {
10784
11459
  input: number;
10785
11460
  output: number;
@@ -10789,14 +11464,14 @@ export declare const MODELS: {
10789
11464
  contextWindow: number;
10790
11465
  maxTokens: number;
10791
11466
  };
10792
- readonly "stepfun/step-3.5-flash": {
11467
+ readonly "x-ai/grok-4-fast": {
10793
11468
  id: string;
10794
11469
  name: string;
10795
11470
  api: "openai-completions";
10796
11471
  provider: string;
10797
11472
  baseUrl: string;
10798
11473
  reasoning: true;
10799
- input: "text"[];
11474
+ input: ("image" | "text")[];
10800
11475
  cost: {
10801
11476
  input: number;
10802
11477
  output: number;
@@ -10806,14 +11481,14 @@ export declare const MODELS: {
10806
11481
  contextWindow: number;
10807
11482
  maxTokens: number;
10808
11483
  };
10809
- readonly "thedrummer/rocinante-12b": {
11484
+ readonly "x-ai/grok-4.1-fast": {
10810
11485
  id: string;
10811
11486
  name: string;
10812
11487
  api: "openai-completions";
10813
11488
  provider: string;
10814
11489
  baseUrl: string;
10815
- reasoning: false;
10816
- input: "text"[];
11490
+ reasoning: true;
11491
+ input: ("image" | "text")[];
10817
11492
  cost: {
10818
11493
  input: number;
10819
11494
  output: number;
@@ -10823,14 +11498,14 @@ export declare const MODELS: {
10823
11498
  contextWindow: number;
10824
11499
  maxTokens: number;
10825
11500
  };
10826
- readonly "thedrummer/unslopnemo-12b": {
11501
+ readonly "x-ai/grok-4.20": {
10827
11502
  id: string;
10828
11503
  name: string;
10829
11504
  api: "openai-completions";
10830
11505
  provider: string;
10831
11506
  baseUrl: string;
10832
- reasoning: false;
10833
- input: "text"[];
11507
+ reasoning: true;
11508
+ input: ("image" | "text")[];
10834
11509
  cost: {
10835
11510
  input: number;
10836
11511
  output: number;
@@ -10840,7 +11515,7 @@ export declare const MODELS: {
10840
11515
  contextWindow: number;
10841
11516
  maxTokens: number;
10842
11517
  };
10843
- readonly "tngtech/deepseek-r1t2-chimera": {
11518
+ readonly "x-ai/grok-code-fast-1": {
10844
11519
  id: string;
10845
11520
  name: string;
10846
11521
  api: "openai-completions";
@@ -10857,7 +11532,7 @@ export declare const MODELS: {
10857
11532
  contextWindow: number;
10858
11533
  maxTokens: number;
10859
11534
  };
10860
- readonly "upstage/solar-pro-3": {
11535
+ readonly "xiaomi/mimo-v2-flash": {
10861
11536
  id: string;
10862
11537
  name: string;
10863
11538
  api: "openai-completions";
@@ -10874,14 +11549,14 @@ export declare const MODELS: {
10874
11549
  contextWindow: number;
10875
11550
  maxTokens: number;
10876
11551
  };
10877
- readonly "x-ai/grok-3": {
11552
+ readonly "xiaomi/mimo-v2-omni": {
10878
11553
  id: string;
10879
11554
  name: string;
10880
11555
  api: "openai-completions";
10881
11556
  provider: string;
10882
11557
  baseUrl: string;
10883
- reasoning: false;
10884
- input: "text"[];
11558
+ reasoning: true;
11559
+ input: ("image" | "text")[];
10885
11560
  cost: {
10886
11561
  input: number;
10887
11562
  output: number;
@@ -10891,13 +11566,13 @@ export declare const MODELS: {
10891
11566
  contextWindow: number;
10892
11567
  maxTokens: number;
10893
11568
  };
10894
- readonly "x-ai/grok-3-beta": {
11569
+ readonly "xiaomi/mimo-v2-pro": {
10895
11570
  id: string;
10896
11571
  name: string;
10897
11572
  api: "openai-completions";
10898
11573
  provider: string;
10899
11574
  baseUrl: string;
10900
- reasoning: false;
11575
+ reasoning: true;
10901
11576
  input: "text"[];
10902
11577
  cost: {
10903
11578
  input: number;
@@ -10908,14 +11583,14 @@ export declare const MODELS: {
10908
11583
  contextWindow: number;
10909
11584
  maxTokens: number;
10910
11585
  };
10911
- readonly "x-ai/grok-3-mini": {
11586
+ readonly "xiaomi/mimo-v2.5": {
10912
11587
  id: string;
10913
11588
  name: string;
10914
11589
  api: "openai-completions";
10915
11590
  provider: string;
10916
11591
  baseUrl: string;
10917
11592
  reasoning: true;
10918
- input: "text"[];
11593
+ input: ("image" | "text")[];
10919
11594
  cost: {
10920
11595
  input: number;
10921
11596
  output: number;
@@ -10925,7 +11600,7 @@ export declare const MODELS: {
10925
11600
  contextWindow: number;
10926
11601
  maxTokens: number;
10927
11602
  };
10928
- readonly "x-ai/grok-3-mini-beta": {
11603
+ readonly "xiaomi/mimo-v2.5-pro": {
10929
11604
  id: string;
10930
11605
  name: string;
10931
11606
  api: "openai-completions";
@@ -10942,14 +11617,14 @@ export declare const MODELS: {
10942
11617
  contextWindow: number;
10943
11618
  maxTokens: number;
10944
11619
  };
10945
- readonly "x-ai/grok-4": {
11620
+ readonly "z-ai/glm-4-32b": {
10946
11621
  id: string;
10947
11622
  name: string;
10948
11623
  api: "openai-completions";
10949
11624
  provider: string;
10950
11625
  baseUrl: string;
10951
- reasoning: true;
10952
- input: ("image" | "text")[];
11626
+ reasoning: false;
11627
+ input: "text"[];
10953
11628
  cost: {
10954
11629
  input: number;
10955
11630
  output: number;
@@ -10959,14 +11634,14 @@ export declare const MODELS: {
10959
11634
  contextWindow: number;
10960
11635
  maxTokens: number;
10961
11636
  };
10962
- readonly "x-ai/grok-4-fast": {
11637
+ readonly "z-ai/glm-4.5": {
10963
11638
  id: string;
10964
11639
  name: string;
10965
11640
  api: "openai-completions";
10966
11641
  provider: string;
10967
11642
  baseUrl: string;
10968
11643
  reasoning: true;
10969
- input: ("image" | "text")[];
11644
+ input: "text"[];
10970
11645
  cost: {
10971
11646
  input: number;
10972
11647
  output: number;
@@ -10976,14 +11651,14 @@ export declare const MODELS: {
10976
11651
  contextWindow: number;
10977
11652
  maxTokens: number;
10978
11653
  };
10979
- readonly "x-ai/grok-4.1-fast": {
11654
+ readonly "z-ai/glm-4.5-air": {
10980
11655
  id: string;
10981
11656
  name: string;
10982
11657
  api: "openai-completions";
10983
11658
  provider: string;
10984
11659
  baseUrl: string;
10985
11660
  reasoning: true;
10986
- input: ("image" | "text")[];
11661
+ input: "text"[];
10987
11662
  cost: {
10988
11663
  input: number;
10989
11664
  output: number;
@@ -10993,14 +11668,14 @@ export declare const MODELS: {
10993
11668
  contextWindow: number;
10994
11669
  maxTokens: number;
10995
11670
  };
10996
- readonly "x-ai/grok-4.20": {
11671
+ readonly "z-ai/glm-4.5-air:free": {
10997
11672
  id: string;
10998
11673
  name: string;
10999
11674
  api: "openai-completions";
11000
11675
  provider: string;
11001
11676
  baseUrl: string;
11002
11677
  reasoning: true;
11003
- input: ("image" | "text")[];
11678
+ input: "text"[];
11004
11679
  cost: {
11005
11680
  input: number;
11006
11681
  output: number;
@@ -11010,14 +11685,14 @@ export declare const MODELS: {
11010
11685
  contextWindow: number;
11011
11686
  maxTokens: number;
11012
11687
  };
11013
- readonly "x-ai/grok-code-fast-1": {
11688
+ readonly "z-ai/glm-4.5v": {
11014
11689
  id: string;
11015
11690
  name: string;
11016
11691
  api: "openai-completions";
11017
11692
  provider: string;
11018
11693
  baseUrl: string;
11019
11694
  reasoning: true;
11020
- input: "text"[];
11695
+ input: ("image" | "text")[];
11021
11696
  cost: {
11022
11697
  input: number;
11023
11698
  output: number;
@@ -11027,7 +11702,7 @@ export declare const MODELS: {
11027
11702
  contextWindow: number;
11028
11703
  maxTokens: number;
11029
11704
  };
11030
- readonly "xiaomi/mimo-v2-flash": {
11705
+ readonly "z-ai/glm-4.6": {
11031
11706
  id: string;
11032
11707
  name: string;
11033
11708
  api: "openai-completions";
@@ -11044,7 +11719,7 @@ export declare const MODELS: {
11044
11719
  contextWindow: number;
11045
11720
  maxTokens: number;
11046
11721
  };
11047
- readonly "xiaomi/mimo-v2-omni": {
11722
+ readonly "z-ai/glm-4.6v": {
11048
11723
  id: string;
11049
11724
  name: string;
11050
11725
  api: "openai-completions";
@@ -11061,7 +11736,7 @@ export declare const MODELS: {
11061
11736
  contextWindow: number;
11062
11737
  maxTokens: number;
11063
11738
  };
11064
- readonly "xiaomi/mimo-v2-pro": {
11739
+ readonly "z-ai/glm-4.7": {
11065
11740
  id: string;
11066
11741
  name: string;
11067
11742
  api: "openai-completions";
@@ -11078,13 +11753,13 @@ export declare const MODELS: {
11078
11753
  contextWindow: number;
11079
11754
  maxTokens: number;
11080
11755
  };
11081
- readonly "z-ai/glm-4-32b": {
11756
+ readonly "z-ai/glm-4.7-flash": {
11082
11757
  id: string;
11083
11758
  name: string;
11084
11759
  api: "openai-completions";
11085
11760
  provider: string;
11086
11761
  baseUrl: string;
11087
- reasoning: false;
11762
+ reasoning: true;
11088
11763
  input: "text"[];
11089
11764
  cost: {
11090
11765
  input: number;
@@ -11095,7 +11770,7 @@ export declare const MODELS: {
11095
11770
  contextWindow: number;
11096
11771
  maxTokens: number;
11097
11772
  };
11098
- readonly "z-ai/glm-4.5": {
11773
+ readonly "z-ai/glm-5": {
11099
11774
  id: string;
11100
11775
  name: string;
11101
11776
  api: "openai-completions";
@@ -11112,7 +11787,7 @@ export declare const MODELS: {
11112
11787
  contextWindow: number;
11113
11788
  maxTokens: number;
11114
11789
  };
11115
- readonly "z-ai/glm-4.5-air": {
11790
+ readonly "z-ai/glm-5-turbo": {
11116
11791
  id: string;
11117
11792
  name: string;
11118
11793
  api: "openai-completions";
@@ -11129,7 +11804,7 @@ export declare const MODELS: {
11129
11804
  contextWindow: number;
11130
11805
  maxTokens: number;
11131
11806
  };
11132
- readonly "z-ai/glm-4.5-air:free": {
11807
+ readonly "z-ai/glm-5.1": {
11133
11808
  id: string;
11134
11809
  name: string;
11135
11810
  api: "openai-completions";
@@ -11146,7 +11821,7 @@ export declare const MODELS: {
11146
11821
  contextWindow: number;
11147
11822
  maxTokens: number;
11148
11823
  };
11149
- readonly "z-ai/glm-4.5v": {
11824
+ readonly "z-ai/glm-5v-turbo": {
11150
11825
  id: string;
11151
11826
  name: string;
11152
11827
  api: "openai-completions";
@@ -11163,14 +11838,14 @@ export declare const MODELS: {
11163
11838
  contextWindow: number;
11164
11839
  maxTokens: number;
11165
11840
  };
11166
- readonly "z-ai/glm-4.6": {
11841
+ readonly "~anthropic/claude-haiku-latest": {
11167
11842
  id: string;
11168
11843
  name: string;
11169
11844
  api: "openai-completions";
11170
11845
  provider: string;
11171
11846
  baseUrl: string;
11172
11847
  reasoning: true;
11173
- input: "text"[];
11848
+ input: ("image" | "text")[];
11174
11849
  cost: {
11175
11850
  input: number;
11176
11851
  output: number;
@@ -11180,7 +11855,7 @@ export declare const MODELS: {
11180
11855
  contextWindow: number;
11181
11856
  maxTokens: number;
11182
11857
  };
11183
- readonly "z-ai/glm-4.6v": {
11858
+ readonly "~anthropic/claude-opus-latest": {
11184
11859
  id: string;
11185
11860
  name: string;
11186
11861
  api: "openai-completions";
@@ -11197,14 +11872,14 @@ export declare const MODELS: {
11197
11872
  contextWindow: number;
11198
11873
  maxTokens: number;
11199
11874
  };
11200
- readonly "z-ai/glm-4.7": {
11875
+ readonly "~anthropic/claude-sonnet-latest": {
11201
11876
  id: string;
11202
11877
  name: string;
11203
11878
  api: "openai-completions";
11204
11879
  provider: string;
11205
11880
  baseUrl: string;
11206
11881
  reasoning: true;
11207
- input: "text"[];
11882
+ input: ("image" | "text")[];
11208
11883
  cost: {
11209
11884
  input: number;
11210
11885
  output: number;
@@ -11214,14 +11889,14 @@ export declare const MODELS: {
11214
11889
  contextWindow: number;
11215
11890
  maxTokens: number;
11216
11891
  };
11217
- readonly "z-ai/glm-4.7-flash": {
11892
+ readonly "~google/gemini-flash-latest": {
11218
11893
  id: string;
11219
11894
  name: string;
11220
11895
  api: "openai-completions";
11221
11896
  provider: string;
11222
11897
  baseUrl: string;
11223
11898
  reasoning: true;
11224
- input: "text"[];
11899
+ input: ("image" | "text")[];
11225
11900
  cost: {
11226
11901
  input: number;
11227
11902
  output: number;
@@ -11231,14 +11906,14 @@ export declare const MODELS: {
11231
11906
  contextWindow: number;
11232
11907
  maxTokens: number;
11233
11908
  };
11234
- readonly "z-ai/glm-5": {
11909
+ readonly "~google/gemini-pro-latest": {
11235
11910
  id: string;
11236
11911
  name: string;
11237
11912
  api: "openai-completions";
11238
11913
  provider: string;
11239
11914
  baseUrl: string;
11240
11915
  reasoning: true;
11241
- input: "text"[];
11916
+ input: ("image" | "text")[];
11242
11917
  cost: {
11243
11918
  input: number;
11244
11919
  output: number;
@@ -11248,14 +11923,14 @@ export declare const MODELS: {
11248
11923
  contextWindow: number;
11249
11924
  maxTokens: number;
11250
11925
  };
11251
- readonly "z-ai/glm-5-turbo": {
11926
+ readonly "~moonshotai/kimi-latest": {
11252
11927
  id: string;
11253
11928
  name: string;
11254
11929
  api: "openai-completions";
11255
11930
  provider: string;
11256
11931
  baseUrl: string;
11257
11932
  reasoning: true;
11258
- input: "text"[];
11933
+ input: ("image" | "text")[];
11259
11934
  cost: {
11260
11935
  input: number;
11261
11936
  output: number;
@@ -11265,14 +11940,14 @@ export declare const MODELS: {
11265
11940
  contextWindow: number;
11266
11941
  maxTokens: number;
11267
11942
  };
11268
- readonly "z-ai/glm-5.1": {
11943
+ readonly "~openai/gpt-latest": {
11269
11944
  id: string;
11270
11945
  name: string;
11271
11946
  api: "openai-completions";
11272
11947
  provider: string;
11273
11948
  baseUrl: string;
11274
11949
  reasoning: true;
11275
- input: "text"[];
11950
+ input: ("image" | "text")[];
11276
11951
  cost: {
11277
11952
  input: number;
11278
11953
  output: number;
@@ -11282,7 +11957,7 @@ export declare const MODELS: {
11282
11957
  contextWindow: number;
11283
11958
  maxTokens: number;
11284
11959
  };
11285
- readonly "z-ai/glm-5v-turbo": {
11960
+ readonly "~openai/gpt-mini-latest": {
11286
11961
  id: string;
11287
11962
  name: string;
11288
11963
  api: "openai-completions";
@@ -11369,6 +12044,23 @@ export declare const MODELS: {
11369
12044
  contextWindow: number;
11370
12045
  maxTokens: number;
11371
12046
  };
12047
+ readonly "alibaba/qwen-3.6-max-preview": {
12048
+ id: string;
12049
+ name: string;
12050
+ api: "anthropic-messages";
12051
+ provider: string;
12052
+ baseUrl: string;
12053
+ reasoning: true;
12054
+ input: ("image" | "text")[];
12055
+ cost: {
12056
+ input: number;
12057
+ output: number;
12058
+ cacheRead: number;
12059
+ cacheWrite: number;
12060
+ };
12061
+ contextWindow: number;
12062
+ maxTokens: number;
12063
+ };
11372
12064
  readonly "alibaba/qwen3-235b-a22b-thinking": {
11373
12065
  id: string;
11374
12066
  name: string;
@@ -11828,7 +12520,41 @@ export declare const MODELS: {
11828
12520
  contextWindow: number;
11829
12521
  maxTokens: number;
11830
12522
  };
11831
- readonly "cohere/command-a": {
12523
+ readonly "cohere/command-a": {
12524
+ id: string;
12525
+ name: string;
12526
+ api: "anthropic-messages";
12527
+ provider: string;
12528
+ baseUrl: string;
12529
+ reasoning: false;
12530
+ input: "text"[];
12531
+ cost: {
12532
+ input: number;
12533
+ output: number;
12534
+ cacheRead: number;
12535
+ cacheWrite: number;
12536
+ };
12537
+ contextWindow: number;
12538
+ maxTokens: number;
12539
+ };
12540
+ readonly "deepseek/deepseek-r1": {
12541
+ id: string;
12542
+ name: string;
12543
+ api: "anthropic-messages";
12544
+ provider: string;
12545
+ baseUrl: string;
12546
+ reasoning: true;
12547
+ input: "text"[];
12548
+ cost: {
12549
+ input: number;
12550
+ output: number;
12551
+ cacheRead: number;
12552
+ cacheWrite: number;
12553
+ };
12554
+ contextWindow: number;
12555
+ maxTokens: number;
12556
+ };
12557
+ readonly "deepseek/deepseek-v3": {
11832
12558
  id: string;
11833
12559
  name: string;
11834
12560
  api: "anthropic-messages";
@@ -11845,7 +12571,7 @@ export declare const MODELS: {
11845
12571
  contextWindow: number;
11846
12572
  maxTokens: number;
11847
12573
  };
11848
- readonly "deepseek/deepseek-r1": {
12574
+ readonly "deepseek/deepseek-v3.1": {
11849
12575
  id: string;
11850
12576
  name: string;
11851
12577
  api: "anthropic-messages";
@@ -11862,13 +12588,13 @@ export declare const MODELS: {
11862
12588
  contextWindow: number;
11863
12589
  maxTokens: number;
11864
12590
  };
11865
- readonly "deepseek/deepseek-v3": {
12591
+ readonly "deepseek/deepseek-v3.1-terminus": {
11866
12592
  id: string;
11867
12593
  name: string;
11868
12594
  api: "anthropic-messages";
11869
12595
  provider: string;
11870
12596
  baseUrl: string;
11871
- reasoning: false;
12597
+ reasoning: true;
11872
12598
  input: "text"[];
11873
12599
  cost: {
11874
12600
  input: number;
@@ -11879,13 +12605,13 @@ export declare const MODELS: {
11879
12605
  contextWindow: number;
11880
12606
  maxTokens: number;
11881
12607
  };
11882
- readonly "deepseek/deepseek-v3.1": {
12608
+ readonly "deepseek/deepseek-v3.2": {
11883
12609
  id: string;
11884
12610
  name: string;
11885
12611
  api: "anthropic-messages";
11886
12612
  provider: string;
11887
12613
  baseUrl: string;
11888
- reasoning: true;
12614
+ reasoning: false;
11889
12615
  input: "text"[];
11890
12616
  cost: {
11891
12617
  input: number;
@@ -11896,7 +12622,7 @@ export declare const MODELS: {
11896
12622
  contextWindow: number;
11897
12623
  maxTokens: number;
11898
12624
  };
11899
- readonly "deepseek/deepseek-v3.1-terminus": {
12625
+ readonly "deepseek/deepseek-v3.2-thinking": {
11900
12626
  id: string;
11901
12627
  name: string;
11902
12628
  api: "anthropic-messages";
@@ -11913,13 +12639,13 @@ export declare const MODELS: {
11913
12639
  contextWindow: number;
11914
12640
  maxTokens: number;
11915
12641
  };
11916
- readonly "deepseek/deepseek-v3.2": {
12642
+ readonly "deepseek/deepseek-v4-flash": {
11917
12643
  id: string;
11918
12644
  name: string;
11919
12645
  api: "anthropic-messages";
11920
12646
  provider: string;
11921
12647
  baseUrl: string;
11922
- reasoning: false;
12648
+ reasoning: true;
11923
12649
  input: "text"[];
11924
12650
  cost: {
11925
12651
  input: number;
@@ -11930,7 +12656,7 @@ export declare const MODELS: {
11930
12656
  contextWindow: number;
11931
12657
  maxTokens: number;
11932
12658
  };
11933
- readonly "deepseek/deepseek-v3.2-thinking": {
12659
+ readonly "deepseek/deepseek-v4-pro": {
11934
12660
  id: string;
11935
12661
  name: string;
11936
12662
  api: "anthropic-messages";
@@ -12712,6 +13438,23 @@ export declare const MODELS: {
12712
13438
  contextWindow: number;
12713
13439
  maxTokens: number;
12714
13440
  };
13441
+ readonly "moonshotai/kimi-k2.6": {
13442
+ id: string;
13443
+ name: string;
13444
+ api: "anthropic-messages";
13445
+ provider: string;
13446
+ baseUrl: string;
13447
+ reasoning: true;
13448
+ input: ("image" | "text")[];
13449
+ cost: {
13450
+ input: number;
13451
+ output: number;
13452
+ cacheRead: number;
13453
+ cacheWrite: number;
13454
+ };
13455
+ contextWindow: number;
13456
+ maxTokens: number;
13457
+ };
12715
13458
  readonly "nvidia/nemotron-nano-12b-v2-vl": {
12716
13459
  id: string;
12717
13460
  name: string;
@@ -13205,6 +13948,40 @@ export declare const MODELS: {
13205
13948
  contextWindow: number;
13206
13949
  maxTokens: number;
13207
13950
  };
13951
+ readonly "openai/gpt-5.5": {
13952
+ id: string;
13953
+ name: string;
13954
+ api: "anthropic-messages";
13955
+ provider: string;
13956
+ baseUrl: string;
13957
+ reasoning: true;
13958
+ input: ("image" | "text")[];
13959
+ cost: {
13960
+ input: number;
13961
+ output: number;
13962
+ cacheRead: number;
13963
+ cacheWrite: number;
13964
+ };
13965
+ contextWindow: number;
13966
+ maxTokens: number;
13967
+ };
13968
+ readonly "openai/gpt-5.5-pro": {
13969
+ id: string;
13970
+ name: string;
13971
+ api: "anthropic-messages";
13972
+ provider: string;
13973
+ baseUrl: string;
13974
+ reasoning: true;
13975
+ input: ("image" | "text")[];
13976
+ cost: {
13977
+ input: number;
13978
+ output: number;
13979
+ cacheRead: number;
13980
+ cacheWrite: number;
13981
+ };
13982
+ contextWindow: number;
13983
+ maxTokens: number;
13984
+ };
13208
13985
  readonly "openai/gpt-oss-20b": {
13209
13986
  id: string;
13210
13987
  name: string;
@@ -13892,7 +14669,7 @@ export declare const MODELS: {
13892
14669
  provider: string;
13893
14670
  baseUrl: string;
13894
14671
  reasoning: true;
13895
- input: ("image" | "text")[];
14672
+ input: "text"[];
13896
14673
  cost: {
13897
14674
  input: number;
13898
14675
  output: number;
@@ -14331,27 +15108,6 @@ export declare const MODELS: {
14331
15108
  };
14332
15109
  };
14333
15110
  readonly zai: {
14334
- readonly "glm-4.5": {
14335
- id: string;
14336
- name: string;
14337
- api: "openai-completions";
14338
- provider: string;
14339
- baseUrl: string;
14340
- compat: {
14341
- supportsDeveloperRole: false;
14342
- thinkingFormat: "zai";
14343
- };
14344
- reasoning: true;
14345
- input: "text"[];
14346
- cost: {
14347
- input: number;
14348
- output: number;
14349
- cacheRead: number;
14350
- cacheWrite: number;
14351
- };
14352
- contextWindow: number;
14353
- maxTokens: number;
14354
- };
14355
15111
  readonly "glm-4.5-air": {
14356
15112
  id: string;
14357
15113
  name: string;
@@ -14373,92 +15129,6 @@ export declare const MODELS: {
14373
15129
  contextWindow: number;
14374
15130
  maxTokens: number;
14375
15131
  };
14376
- readonly "glm-4.5-flash": {
14377
- id: string;
14378
- name: string;
14379
- api: "openai-completions";
14380
- provider: string;
14381
- baseUrl: string;
14382
- compat: {
14383
- supportsDeveloperRole: false;
14384
- thinkingFormat: "zai";
14385
- };
14386
- reasoning: true;
14387
- input: "text"[];
14388
- cost: {
14389
- input: number;
14390
- output: number;
14391
- cacheRead: number;
14392
- cacheWrite: number;
14393
- };
14394
- contextWindow: number;
14395
- maxTokens: number;
14396
- };
14397
- readonly "glm-4.5v": {
14398
- id: string;
14399
- name: string;
14400
- api: "openai-completions";
14401
- provider: string;
14402
- baseUrl: string;
14403
- compat: {
14404
- supportsDeveloperRole: false;
14405
- thinkingFormat: "zai";
14406
- };
14407
- reasoning: true;
14408
- input: ("image" | "text")[];
14409
- cost: {
14410
- input: number;
14411
- output: number;
14412
- cacheRead: number;
14413
- cacheWrite: number;
14414
- };
14415
- contextWindow: number;
14416
- maxTokens: number;
14417
- };
14418
- readonly "glm-4.6": {
14419
- id: string;
14420
- name: string;
14421
- api: "openai-completions";
14422
- provider: string;
14423
- baseUrl: string;
14424
- compat: {
14425
- supportsDeveloperRole: false;
14426
- thinkingFormat: "zai";
14427
- zaiToolStream: true;
14428
- };
14429
- reasoning: true;
14430
- input: "text"[];
14431
- cost: {
14432
- input: number;
14433
- output: number;
14434
- cacheRead: number;
14435
- cacheWrite: number;
14436
- };
14437
- contextWindow: number;
14438
- maxTokens: number;
14439
- };
14440
- readonly "glm-4.6v": {
14441
- id: string;
14442
- name: string;
14443
- api: "openai-completions";
14444
- provider: string;
14445
- baseUrl: string;
14446
- compat: {
14447
- supportsDeveloperRole: false;
14448
- thinkingFormat: "zai";
14449
- zaiToolStream: true;
14450
- };
14451
- reasoning: true;
14452
- input: ("image" | "text")[];
14453
- cost: {
14454
- input: number;
14455
- output: number;
14456
- cacheRead: number;
14457
- cacheWrite: number;
14458
- };
14459
- contextWindow: number;
14460
- maxTokens: number;
14461
- };
14462
15132
  readonly "glm-4.7": {
14463
15133
  id: string;
14464
15134
  name: string;
@@ -14481,72 +15151,6 @@ export declare const MODELS: {
14481
15151
  contextWindow: number;
14482
15152
  maxTokens: number;
14483
15153
  };
14484
- readonly "glm-4.7-flash": {
14485
- id: string;
14486
- name: string;
14487
- api: "openai-completions";
14488
- provider: string;
14489
- baseUrl: string;
14490
- compat: {
14491
- supportsDeveloperRole: false;
14492
- thinkingFormat: "zai";
14493
- zaiToolStream: true;
14494
- };
14495
- reasoning: true;
14496
- input: "text"[];
14497
- cost: {
14498
- input: number;
14499
- output: number;
14500
- cacheRead: number;
14501
- cacheWrite: number;
14502
- };
14503
- contextWindow: number;
14504
- maxTokens: number;
14505
- };
14506
- readonly "glm-4.7-flashx": {
14507
- id: string;
14508
- name: string;
14509
- api: "openai-completions";
14510
- provider: string;
14511
- baseUrl: string;
14512
- compat: {
14513
- supportsDeveloperRole: false;
14514
- thinkingFormat: "zai";
14515
- zaiToolStream: true;
14516
- };
14517
- reasoning: true;
14518
- input: "text"[];
14519
- cost: {
14520
- input: number;
14521
- output: number;
14522
- cacheRead: number;
14523
- cacheWrite: number;
14524
- };
14525
- contextWindow: number;
14526
- maxTokens: number;
14527
- };
14528
- readonly "glm-5": {
14529
- id: string;
14530
- name: string;
14531
- api: "openai-completions";
14532
- provider: string;
14533
- baseUrl: string;
14534
- compat: {
14535
- supportsDeveloperRole: false;
14536
- thinkingFormat: "zai";
14537
- zaiToolStream: true;
14538
- };
14539
- reasoning: true;
14540
- input: "text"[];
14541
- cost: {
14542
- input: number;
14543
- output: number;
14544
- cacheRead: number;
14545
- cacheWrite: number;
14546
- };
14547
- contextWindow: number;
14548
- maxTokens: number;
14549
- };
14550
15154
  readonly "glm-5-turbo": {
14551
15155
  id: string;
14552
15156
  name: string;
@@ -14591,28 +15195,6 @@ export declare const MODELS: {
14591
15195
  contextWindow: number;
14592
15196
  maxTokens: number;
14593
15197
  };
14594
- readonly "glm-5v-turbo": {
14595
- id: string;
14596
- name: string;
14597
- api: "openai-completions";
14598
- provider: string;
14599
- baseUrl: string;
14600
- compat: {
14601
- supportsDeveloperRole: false;
14602
- thinkingFormat: "zai";
14603
- zaiToolStream: true;
14604
- };
14605
- reasoning: true;
14606
- input: ("image" | "text")[];
14607
- cost: {
14608
- input: number;
14609
- output: number;
14610
- cacheRead: number;
14611
- cacheWrite: number;
14612
- };
14613
- contextWindow: number;
14614
- maxTokens: number;
14615
- };
14616
15198
  };
14617
15199
  };
14618
15200
  //# sourceMappingURL=models.generated.d.ts.map