@draht/ai 2026.3.25 → 2026.4.5

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1054,6 +1054,23 @@ export declare const MODELS: {
1054
1054
  contextWindow: number;
1055
1055
  maxTokens: number;
1056
1056
  };
1057
+ readonly "nvidia.nemotron-super-3-120b": {
1058
+ id: string;
1059
+ name: string;
1060
+ api: "bedrock-converse-stream";
1061
+ provider: string;
1062
+ baseUrl: string;
1063
+ reasoning: true;
1064
+ input: "text"[];
1065
+ cost: {
1066
+ input: number;
1067
+ output: number;
1068
+ cacheRead: number;
1069
+ cacheWrite: number;
1070
+ };
1071
+ contextWindow: number;
1072
+ maxTokens: number;
1073
+ };
1057
1074
  readonly "openai.gpt-oss-120b-1:0": {
1058
1075
  id: string;
1059
1076
  name: string;
@@ -1532,23 +1549,6 @@ export declare const MODELS: {
1532
1549
  contextWindow: number;
1533
1550
  maxTokens: number;
1534
1551
  };
1535
- readonly "claude-3-7-sonnet-latest": {
1536
- id: string;
1537
- name: string;
1538
- api: "anthropic-messages";
1539
- provider: string;
1540
- baseUrl: string;
1541
- reasoning: true;
1542
- input: ("image" | "text")[];
1543
- cost: {
1544
- input: number;
1545
- output: number;
1546
- cacheRead: number;
1547
- cacheWrite: number;
1548
- };
1549
- contextWindow: number;
1550
- maxTokens: number;
1551
- };
1552
1552
  readonly "claude-3-haiku-20240307": {
1553
1553
  id: string;
1554
1554
  name: string;
@@ -2282,6 +2282,23 @@ export declare const MODELS: {
2282
2282
  contextWindow: number;
2283
2283
  maxTokens: number;
2284
2284
  };
2285
+ readonly "gpt-5.3-chat-latest": {
2286
+ id: string;
2287
+ name: string;
2288
+ api: "azure-openai-responses";
2289
+ provider: string;
2290
+ baseUrl: string;
2291
+ reasoning: false;
2292
+ input: ("image" | "text")[];
2293
+ cost: {
2294
+ input: number;
2295
+ output: number;
2296
+ cacheRead: number;
2297
+ cacheWrite: number;
2298
+ };
2299
+ contextWindow: number;
2300
+ maxTokens: number;
2301
+ };
2285
2302
  readonly "gpt-5.3-codex": {
2286
2303
  id: string;
2287
2304
  name: string;
@@ -3589,6 +3606,57 @@ export declare const MODELS: {
3589
3606
  contextWindow: number;
3590
3607
  maxTokens: number;
3591
3608
  };
3609
+ readonly "gemma-3-27b-it": {
3610
+ id: string;
3611
+ name: string;
3612
+ api: "google-generative-ai";
3613
+ provider: string;
3614
+ baseUrl: string;
3615
+ reasoning: false;
3616
+ input: ("image" | "text")[];
3617
+ cost: {
3618
+ input: number;
3619
+ output: number;
3620
+ cacheRead: number;
3621
+ cacheWrite: number;
3622
+ };
3623
+ contextWindow: number;
3624
+ maxTokens: number;
3625
+ };
3626
+ readonly "gemma-4-26b": {
3627
+ id: string;
3628
+ name: string;
3629
+ api: "google-generative-ai";
3630
+ provider: string;
3631
+ baseUrl: string;
3632
+ reasoning: true;
3633
+ input: ("image" | "text")[];
3634
+ cost: {
3635
+ input: number;
3636
+ output: number;
3637
+ cacheRead: number;
3638
+ cacheWrite: number;
3639
+ };
3640
+ contextWindow: number;
3641
+ maxTokens: number;
3642
+ };
3643
+ readonly "gemma-4-31b": {
3644
+ id: string;
3645
+ name: string;
3646
+ api: "google-generative-ai";
3647
+ provider: string;
3648
+ baseUrl: string;
3649
+ reasoning: true;
3650
+ input: ("image" | "text")[];
3651
+ cost: {
3652
+ input: number;
3653
+ output: number;
3654
+ cacheRead: number;
3655
+ cacheWrite: number;
3656
+ };
3657
+ contextWindow: number;
3658
+ maxTokens: number;
3659
+ };
3592
3660
  };
3593
3661
  readonly "google-antigravity": {
3594
3662
  readonly "claude-opus-4-5-thinking": {
@@ -4090,6 +4158,40 @@ export declare const MODELS: {
4090
4158
  contextWindow: number;
4091
4159
  maxTokens: number;
4092
4160
  };
4161
+ readonly "groq/compound": {
4162
+ id: string;
4163
+ name: string;
4164
+ api: "openai-completions";
4165
+ provider: string;
4166
+ baseUrl: string;
4167
+ reasoning: true;
4168
+ input: "text"[];
4169
+ cost: {
4170
+ input: number;
4171
+ output: number;
4172
+ cacheRead: number;
4173
+ cacheWrite: number;
4174
+ };
4175
+ contextWindow: number;
4176
+ maxTokens: number;
4177
+ };
4178
+ readonly "groq/compound-mini": {
4179
+ id: string;
4180
+ name: string;
4181
+ api: "openai-completions";
4182
+ provider: string;
4183
+ baseUrl: string;
4184
+ reasoning: true;
4185
+ input: "text"[];
4186
+ cost: {
4187
+ input: number;
4188
+ output: number;
4189
+ cacheRead: number;
4190
+ cacheWrite: number;
4191
+ };
4192
+ contextWindow: number;
4193
+ maxTokens: number;
4194
+ };
4093
4195
  readonly "llama-3.1-8b-instant": {
4094
4196
  id: string;
4095
4197
  name: string;
@@ -4277,6 +4379,23 @@ export declare const MODELS: {
4277
4379
  contextWindow: number;
4278
4380
  maxTokens: number;
4279
4381
  };
4382
+ readonly "openai/gpt-oss-safeguard-20b": {
4383
+ id: string;
4384
+ name: string;
4385
+ api: "openai-completions";
4386
+ provider: string;
4387
+ baseUrl: string;
4388
+ reasoning: true;
4389
+ input: "text"[];
4390
+ cost: {
4391
+ input: number;
4392
+ output: number;
4393
+ cacheRead: number;
4394
+ cacheWrite: number;
4395
+ };
4396
+ contextWindow: number;
4397
+ maxTokens: number;
4398
+ };
4280
4399
  readonly "qwen-qwq-32b": {
4281
4400
  id: string;
4282
4401
  name: string;
@@ -5276,13 +5395,30 @@ export declare const MODELS: {
5276
5395
  contextWindow: number;
5277
5396
  maxTokens: number;
5278
5397
  };
5398
+ readonly "mistral-small-2603": {
5399
+ id: string;
5400
+ name: string;
5401
+ api: "mistral-conversations";
5402
+ provider: string;
5403
+ baseUrl: string;
5404
+ reasoning: true;
5405
+ input: ("image" | "text")[];
5406
+ cost: {
5407
+ input: number;
5408
+ output: number;
5409
+ cacheRead: number;
5410
+ cacheWrite: number;
5411
+ };
5412
+ contextWindow: number;
5413
+ maxTokens: number;
5414
+ };
5279
5415
  readonly "mistral-small-latest": {
5280
5416
  id: string;
5281
5417
  name: string;
5282
5418
  api: "mistral-conversations";
5283
5419
  provider: string;
5284
5420
  baseUrl: string;
5285
- reasoning: false;
5421
+ reasoning: true;
5286
5422
  input: ("image" | "text")[];
5287
5423
  cost: {
5288
5424
  input: number;
@@ -5822,6 +5958,23 @@ export declare const MODELS: {
5822
5958
  contextWindow: number;
5823
5959
  maxTokens: number;
5824
5960
  };
5961
+ readonly "gpt-5.3-chat-latest": {
5962
+ id: string;
5963
+ name: string;
5964
+ api: "openai-responses";
5965
+ provider: string;
5966
+ baseUrl: string;
5967
+ reasoning: false;
5968
+ input: ("image" | "text")[];
5969
+ cost: {
5970
+ input: number;
5971
+ output: number;
5972
+ cacheRead: number;
5973
+ cacheWrite: number;
5974
+ };
5975
+ contextWindow: number;
5976
+ maxTokens: number;
5977
+ };
5825
5978
  readonly "gpt-5.3-codex": {
5826
5979
  id: string;
5827
5980
  name: string;
@@ -6676,14 +6829,14 @@ export declare const MODELS: {
6676
6829
  contextWindow: number;
6677
6830
  maxTokens: number;
6678
6831
  };
6679
- readonly "mimo-v2-omni-free": {
6832
+ readonly "minimax-m2.5": {
6680
6833
  id: string;
6681
6834
  name: string;
6682
6835
  api: "openai-completions";
6683
6836
  provider: string;
6684
6837
  baseUrl: string;
6685
6838
  reasoning: true;
6686
- input: ("image" | "text")[];
6839
+ input: "text"[];
6687
6840
  cost: {
6688
6841
  input: number;
6689
6842
  output: number;
@@ -6693,10 +6846,10 @@ export declare const MODELS: {
6693
6846
  contextWindow: number;
6694
6847
  maxTokens: number;
6695
6848
  };
6696
- readonly "mimo-v2-pro-free": {
6849
+ readonly "minimax-m2.5-free": {
6697
6850
  id: string;
6698
6851
  name: string;
6699
- api: "openai-completions";
6852
+ api: "anthropic-messages";
6700
6853
  provider: string;
6701
6854
  baseUrl: string;
6702
6855
  reasoning: true;
@@ -6710,7 +6863,7 @@ export declare const MODELS: {
6710
6863
  contextWindow: number;
6711
6864
  maxTokens: number;
6712
6865
  };
6713
- readonly "minimax-m2.5": {
6866
+ readonly "nemotron-3-super-free": {
6714
6867
  id: string;
6715
6868
  name: string;
6716
6869
  api: "openai-completions";
@@ -6727,10 +6880,10 @@ export declare const MODELS: {
6727
6880
  contextWindow: number;
6728
6881
  maxTokens: number;
6729
6882
  };
6730
- readonly "minimax-m2.5-free": {
6883
+ readonly "qwen3.6-plus-free": {
6731
6884
  id: string;
6732
6885
  name: string;
6733
- api: "anthropic-messages";
6886
+ api: "openai-completions";
6734
6887
  provider: string;
6735
6888
  baseUrl: string;
6736
6889
  reasoning: true;
@@ -6744,7 +6897,9 @@ export declare const MODELS: {
6744
6897
  contextWindow: number;
6745
6898
  maxTokens: number;
6746
6899
  };
6747
- readonly "nemotron-3-super-free": {
6900
+ };
6901
+ readonly "opencode-go": {
6902
+ readonly "glm-5": {
6748
6903
  id: string;
6749
6904
  name: string;
6750
6905
  api: "openai-completions";
@@ -6761,16 +6916,14 @@ export declare const MODELS: {
6761
6916
  contextWindow: number;
6762
6917
  maxTokens: number;
6763
6918
  };
6764
- };
6765
- readonly "opencode-go": {
6766
- readonly "glm-5": {
6919
+ readonly "kimi-k2.5": {
6767
6920
  id: string;
6768
6921
  name: string;
6769
6922
  api: "openai-completions";
6770
6923
  provider: string;
6771
6924
  baseUrl: string;
6772
6925
  reasoning: true;
6773
- input: "text"[];
6926
+ input: ("image" | "text")[];
6774
6927
  cost: {
6775
6928
  input: number;
6776
6929
  output: number;
@@ -6780,7 +6933,7 @@ export declare const MODELS: {
6780
6933
  contextWindow: number;
6781
6934
  maxTokens: number;
6782
6935
  };
6783
- readonly "kimi-k2.5": {
6936
+ readonly "mimo-v2-omni": {
6784
6937
  id: string;
6785
6938
  name: string;
6786
6939
  api: "openai-completions";
@@ -6797,10 +6950,27 @@ export declare const MODELS: {
6797
6950
  contextWindow: number;
6798
6951
  maxTokens: number;
6799
6952
  };
6953
+ readonly "mimo-v2-pro": {
6954
+ id: string;
6955
+ name: string;
6956
+ api: "openai-completions";
6957
+ provider: string;
6958
+ baseUrl: string;
6959
+ reasoning: true;
6960
+ input: "text"[];
6961
+ cost: {
6962
+ input: number;
6963
+ output: number;
6964
+ cacheRead: number;
6965
+ cacheWrite: number;
6966
+ };
6967
+ contextWindow: number;
6968
+ maxTokens: number;
6969
+ };
6800
6970
  readonly "minimax-m2.5": {
6801
6971
  id: string;
6802
6972
  name: string;
6803
- api: "anthropic-messages";
6973
+ api: "openai-completions";
6804
6974
  provider: string;
6805
6975
  baseUrl: string;
6806
6976
  reasoning: true;
@@ -7003,13 +7173,13 @@ export declare const MODELS: {
7003
7173
  contextWindow: number;
7004
7174
  maxTokens: number;
7005
7175
  };
7006
- readonly "anthropic/claude-3.5-sonnet": {
7176
+ readonly "anthropic/claude-3.7-sonnet": {
7007
7177
  id: string;
7008
7178
  name: string;
7009
7179
  api: "openai-completions";
7010
7180
  provider: string;
7011
7181
  baseUrl: string;
7012
- reasoning: false;
7182
+ reasoning: true;
7013
7183
  input: ("image" | "text")[];
7014
7184
  cost: {
7015
7185
  input: number;
@@ -7020,7 +7190,7 @@ export declare const MODELS: {
7020
7190
  contextWindow: number;
7021
7191
  maxTokens: number;
7022
7192
  };
7023
- readonly "anthropic/claude-3.7-sonnet": {
7193
+ readonly "anthropic/claude-3.7-sonnet:thinking": {
7024
7194
  id: string;
7025
7195
  name: string;
7026
7196
  api: "openai-completions";
@@ -7037,7 +7207,7 @@ export declare const MODELS: {
7037
7207
  contextWindow: number;
7038
7208
  maxTokens: number;
7039
7209
  };
7040
- readonly "anthropic/claude-3.7-sonnet:thinking": {
7210
+ readonly "anthropic/claude-haiku-4.5": {
7041
7211
  id: string;
7042
7212
  name: string;
7043
7213
  api: "openai-completions";
@@ -7054,7 +7224,7 @@ export declare const MODELS: {
7054
7224
  contextWindow: number;
7055
7225
  maxTokens: number;
7056
7226
  };
7057
- readonly "anthropic/claude-haiku-4.5": {
7227
+ readonly "anthropic/claude-opus-4": {
7058
7228
  id: string;
7059
7229
  name: string;
7060
7230
  api: "openai-completions";
@@ -7071,7 +7241,7 @@ export declare const MODELS: {
7071
7241
  contextWindow: number;
7072
7242
  maxTokens: number;
7073
7243
  };
7074
- readonly "anthropic/claude-opus-4": {
7244
+ readonly "anthropic/claude-opus-4.1": {
7075
7245
  id: string;
7076
7246
  name: string;
7077
7247
  api: "openai-completions";
@@ -7088,7 +7258,7 @@ export declare const MODELS: {
7088
7258
  contextWindow: number;
7089
7259
  maxTokens: number;
7090
7260
  };
7091
- readonly "anthropic/claude-opus-4.1": {
7261
+ readonly "anthropic/claude-opus-4.5": {
7092
7262
  id: string;
7093
7263
  name: string;
7094
7264
  api: "openai-completions";
@@ -7105,7 +7275,7 @@ export declare const MODELS: {
7105
7275
  contextWindow: number;
7106
7276
  maxTokens: number;
7107
7277
  };
7108
- readonly "anthropic/claude-opus-4.5": {
7278
+ readonly "anthropic/claude-opus-4.6": {
7109
7279
  id: string;
7110
7280
  name: string;
7111
7281
  api: "openai-completions";
@@ -7122,7 +7292,7 @@ export declare const MODELS: {
7122
7292
  contextWindow: number;
7123
7293
  maxTokens: number;
7124
7294
  };
7125
- readonly "anthropic/claude-opus-4.6": {
7295
+ readonly "anthropic/claude-sonnet-4": {
7126
7296
  id: string;
7127
7297
  name: string;
7128
7298
  api: "openai-completions";
@@ -7139,7 +7309,7 @@ export declare const MODELS: {
7139
7309
  contextWindow: number;
7140
7310
  maxTokens: number;
7141
7311
  };
7142
- readonly "anthropic/claude-sonnet-4": {
7312
+ readonly "anthropic/claude-sonnet-4.5": {
7143
7313
  id: string;
7144
7314
  name: string;
7145
7315
  api: "openai-completions";
@@ -7156,7 +7326,7 @@ export declare const MODELS: {
7156
7326
  contextWindow: number;
7157
7327
  maxTokens: number;
7158
7328
  };
7159
- readonly "anthropic/claude-sonnet-4.5": {
7329
+ readonly "anthropic/claude-sonnet-4.6": {
7160
7330
  id: string;
7161
7331
  name: string;
7162
7332
  api: "openai-completions";
@@ -7173,14 +7343,14 @@ export declare const MODELS: {
7173
7343
  contextWindow: number;
7174
7344
  maxTokens: number;
7175
7345
  };
7176
- readonly "anthropic/claude-sonnet-4.6": {
7346
+ readonly "arcee-ai/trinity-large-preview:free": {
7177
7347
  id: string;
7178
7348
  name: string;
7179
7349
  api: "openai-completions";
7180
7350
  provider: string;
7181
7351
  baseUrl: string;
7182
- reasoning: true;
7183
- input: ("image" | "text")[];
7352
+ reasoning: false;
7353
+ input: "text"[];
7184
7354
  cost: {
7185
7355
  input: number;
7186
7356
  output: number;
@@ -7190,13 +7360,13 @@ export declare const MODELS: {
7190
7360
  contextWindow: number;
7191
7361
  maxTokens: number;
7192
7362
  };
7193
- readonly "arcee-ai/trinity-large-preview:free": {
7363
+ readonly "arcee-ai/trinity-large-thinking": {
7194
7364
  id: string;
7195
7365
  name: string;
7196
7366
  api: "openai-completions";
7197
7367
  provider: string;
7198
7368
  baseUrl: string;
7199
- reasoning: false;
7369
+ reasoning: true;
7200
7370
  input: "text"[];
7201
7371
  cost: {
7202
7372
  input: number;
@@ -7717,7 +7887,7 @@ export declare const MODELS: {
7717
7887
  contextWindow: number;
7718
7888
  maxTokens: number;
7719
7889
  };
7720
- readonly "google/gemini-3-pro-preview": {
7890
+ readonly "google/gemini-3.1-flash-lite-preview": {
7721
7891
  id: string;
7722
7892
  name: string;
7723
7893
  api: "openai-completions";
@@ -7734,7 +7904,7 @@ export declare const MODELS: {
7734
7904
  contextWindow: number;
7735
7905
  maxTokens: number;
7736
7906
  };
7737
- readonly "google/gemini-3.1-flash-lite-preview": {
7907
+ readonly "google/gemini-3.1-pro-preview": {
7738
7908
  id: string;
7739
7909
  name: string;
7740
7910
  api: "openai-completions";
@@ -7751,7 +7921,7 @@ export declare const MODELS: {
7751
7921
  contextWindow: number;
7752
7922
  maxTokens: number;
7753
7923
  };
7754
- readonly "google/gemini-3.1-pro-preview": {
7924
+ readonly "google/gemini-3.1-pro-preview-customtools": {
7755
7925
  id: string;
7756
7926
  name: string;
7757
7927
  api: "openai-completions";
@@ -7768,7 +7938,24 @@ export declare const MODELS: {
7768
7938
  contextWindow: number;
7769
7939
  maxTokens: number;
7770
7940
  };
7771
- readonly "google/gemini-3.1-pro-preview-customtools": {
7941
+ readonly "google/gemma-4-26b-a4b-it": {
7942
+ id: string;
7943
+ name: string;
7944
+ api: "openai-completions";
7945
+ provider: string;
7946
+ baseUrl: string;
7947
+ reasoning: true;
7948
+ input: ("image" | "text")[];
7949
+ cost: {
7950
+ input: number;
7951
+ output: number;
7952
+ cacheRead: number;
7953
+ cacheWrite: number;
7954
+ };
7955
+ contextWindow: number;
7956
+ maxTokens: number;
7957
+ };
7958
+ readonly "google/gemma-4-31b-it": {
7772
7959
  id: string;
7773
7960
  name: string;
7774
7961
  api: "openai-completions";
@@ -7836,7 +8023,7 @@ export declare const MODELS: {
7836
8023
  contextWindow: number;
7837
8024
  maxTokens: number;
7838
8025
  };
7839
- readonly "kwaipilot/kat-coder-pro": {
8026
+ readonly "kwaipilot/kat-coder-pro-v2": {
7840
8027
  id: string;
7841
8028
  name: string;
7842
8029
  api: "openai-completions";
@@ -8346,23 +8533,6 @@ export declare const MODELS: {
8346
8533
  contextWindow: number;
8347
8534
  maxTokens: number;
8348
8535
  };
8349
- readonly "mistralai/mistral-small-24b-instruct-2501": {
8350
- id: string;
8351
- name: string;
8352
- api: "openai-completions";
8353
- provider: string;
8354
- baseUrl: string;
8355
- reasoning: false;
8356
- input: "text"[];
8357
- cost: {
8358
- input: number;
8359
- output: number;
8360
- cacheRead: number;
8361
- cacheWrite: number;
8362
- };
8363
- contextWindow: number;
8364
- maxTokens: number;
8365
- };
8366
8536
  readonly "mistralai/mistral-small-2603": {
8367
8537
  id: string;
8368
8538
  name: string;
@@ -8380,23 +8550,6 @@ export declare const MODELS: {
8380
8550
  contextWindow: number;
8381
8551
  maxTokens: number;
8382
8552
  };
8383
- readonly "mistralai/mistral-small-3.1-24b-instruct:free": {
8384
- id: string;
8385
- name: string;
8386
- api: "openai-completions";
8387
- provider: string;
8388
- baseUrl: string;
8389
- reasoning: false;
8390
- input: ("image" | "text")[];
8391
- cost: {
8392
- input: number;
8393
- output: number;
8394
- cacheRead: number;
8395
- cacheWrite: number;
8396
- };
8397
- contextWindow: number;
8398
- maxTokens: number;
8399
- };
8400
8553
  readonly "mistralai/mistral-small-3.2-24b-instruct": {
8401
8554
  id: string;
8402
8555
  name: string;
@@ -9434,6 +9587,40 @@ export declare const MODELS: {
9434
9587
  contextWindow: number;
9435
9588
  maxTokens: number;
9436
9589
  };
9590
+ readonly "openai/gpt-audio": {
9591
+ id: string;
9592
+ name: string;
9593
+ api: "openai-completions";
9594
+ provider: string;
9595
+ baseUrl: string;
9596
+ reasoning: false;
9597
+ input: "text"[];
9598
+ cost: {
9599
+ input: number;
9600
+ output: number;
9601
+ cacheRead: number;
9602
+ cacheWrite: number;
9603
+ };
9604
+ contextWindow: number;
9605
+ maxTokens: number;
9606
+ };
9607
+ readonly "openai/gpt-audio-mini": {
9608
+ id: string;
9609
+ name: string;
9610
+ api: "openai-completions";
9611
+ provider: string;
9612
+ baseUrl: string;
9613
+ reasoning: false;
9614
+ input: "text"[];
9615
+ cost: {
9616
+ input: number;
9617
+ output: number;
9618
+ cacheRead: number;
9619
+ cacheWrite: number;
9620
+ };
9621
+ contextWindow: number;
9622
+ maxTokens: number;
9623
+ };
9437
9624
  readonly "openai/gpt-oss-120b": {
9438
9625
  id: string;
9439
9626
  name: string;
@@ -9995,23 +10182,6 @@ export declare const MODELS: {
9995
10182
  contextWindow: number;
9996
10183
  maxTokens: number;
9997
10184
  };
9998
- readonly "qwen/qwen3-4b:free": {
9999
- id: string;
10000
- name: string;
10001
- api: "openai-completions";
10002
- provider: string;
10003
- baseUrl: string;
10004
- reasoning: true;
10005
- input: "text"[];
10006
- cost: {
10007
- input: number;
10008
- output: number;
10009
- cacheRead: number;
10010
- cacheWrite: number;
10011
- };
10012
- contextWindow: number;
10013
- maxTokens: number;
10014
- };
10015
10185
  readonly "qwen/qwen3-8b": {
10016
10186
  id: string;
10017
10187
  name: string;
@@ -10454,6 +10624,23 @@ export declare const MODELS: {
10454
10624
  contextWindow: number;
10455
10625
  maxTokens: number;
10456
10626
  };
10627
+ readonly "qwen/qwen3.6-plus:free": {
10628
+ id: string;
10629
+ name: string;
10630
+ api: "openai-completions";
10631
+ provider: string;
10632
+ baseUrl: string;
10633
+ reasoning: true;
10634
+ input: ("image" | "text")[];
10635
+ cost: {
10636
+ input: number;
10637
+ output: number;
10638
+ cacheRead: number;
10639
+ cacheWrite: number;
10640
+ };
10641
+ contextWindow: number;
10642
+ maxTokens: number;
10643
+ };
10457
10644
  readonly "qwen/qwq-32b": {
10458
10645
  id: string;
10459
10646
  name: string;
@@ -10471,6 +10658,23 @@ export declare const MODELS: {
10471
10658
  contextWindow: number;
10472
10659
  maxTokens: number;
10473
10660
  };
10661
+ readonly "rekaai/reka-edge": {
10662
+ id: string;
10663
+ name: string;
10664
+ api: "openai-completions";
10665
+ provider: string;
10666
+ baseUrl: string;
10667
+ reasoning: false;
10668
+ input: ("image" | "text")[];
10669
+ cost: {
10670
+ input: number;
10671
+ output: number;
10672
+ cacheRead: number;
10673
+ cacheWrite: number;
10674
+ };
10675
+ contextWindow: number;
10676
+ maxTokens: number;
10677
+ };
10474
10678
  readonly "relace/relace-search": {
10475
10679
  id: string;
10476
10680
  name: string;
@@ -10743,7 +10947,7 @@ export declare const MODELS: {
10743
10947
  contextWindow: number;
10744
10948
  maxTokens: number;
10745
10949
  };
10746
- readonly "x-ai/grok-4.20-beta": {
10950
+ readonly "x-ai/grok-4.20": {
10747
10951
  id: string;
10748
10952
  name: string;
10749
10953
  api: "openai-completions";
@@ -11015,6 +11219,23 @@ export declare const MODELS: {
11015
11219
  contextWindow: number;
11016
11220
  maxTokens: number;
11017
11221
  };
11222
+ readonly "z-ai/glm-5v-turbo": {
11223
+ id: string;
11224
+ name: string;
11225
+ api: "openai-completions";
11226
+ provider: string;
11227
+ baseUrl: string;
11228
+ reasoning: true;
11229
+ input: ("image" | "text")[];
11230
+ cost: {
11231
+ input: number;
11232
+ output: number;
11233
+ cacheRead: number;
11234
+ cacheWrite: number;
11235
+ };
11236
+ contextWindow: number;
11237
+ maxTokens: number;
11238
+ };
11018
11239
  };
11019
11240
  readonly "vercel-ai-gateway": {
11020
11241
  readonly "alibaba/qwen-3-14b": {
@@ -11272,13 +11493,13 @@ export declare const MODELS: {
11272
11493
  contextWindow: number;
11273
11494
  maxTokens: number;
11274
11495
  };
11275
- readonly "anthropic/claude-3-haiku": {
11496
+ readonly "alibaba/qwen3.6-plus": {
11276
11497
  id: string;
11277
11498
  name: string;
11278
11499
  api: "anthropic-messages";
11279
11500
  provider: string;
11280
11501
  baseUrl: string;
11281
- reasoning: false;
11502
+ reasoning: true;
11282
11503
  input: ("image" | "text")[];
11283
11504
  cost: {
11284
11505
  input: number;
@@ -11289,7 +11510,7 @@ export declare const MODELS: {
11289
11510
  contextWindow: number;
11290
11511
  maxTokens: number;
11291
11512
  };
11292
- readonly "anthropic/claude-3.5-haiku": {
11513
+ readonly "anthropic/claude-3-haiku": {
11293
11514
  id: string;
11294
11515
  name: string;
11295
11516
  api: "anthropic-messages";
@@ -11306,7 +11527,7 @@ export declare const MODELS: {
11306
11527
  contextWindow: number;
11307
11528
  maxTokens: number;
11308
11529
  };
11309
- readonly "anthropic/claude-3.5-sonnet": {
11530
+ readonly "anthropic/claude-3.5-haiku": {
11310
11531
  id: string;
11311
11532
  name: string;
11312
11533
  api: "anthropic-messages";
@@ -11323,13 +11544,13 @@ export declare const MODELS: {
11323
11544
  contextWindow: number;
11324
11545
  maxTokens: number;
11325
11546
  };
11326
- readonly "anthropic/claude-3.5-sonnet-20240620": {
11547
+ readonly "anthropic/claude-3.7-sonnet": {
11327
11548
  id: string;
11328
11549
  name: string;
11329
11550
  api: "anthropic-messages";
11330
11551
  provider: string;
11331
11552
  baseUrl: string;
11332
- reasoning: false;
11553
+ reasoning: true;
11333
11554
  input: ("image" | "text")[];
11334
11555
  cost: {
11335
11556
  input: number;
@@ -11340,7 +11561,7 @@ export declare const MODELS: {
11340
11561
  contextWindow: number;
11341
11562
  maxTokens: number;
11342
11563
  };
11343
- readonly "anthropic/claude-3.7-sonnet": {
11564
+ readonly "anthropic/claude-haiku-4.5": {
11344
11565
  id: string;
11345
11566
  name: string;
11346
11567
  api: "anthropic-messages";
@@ -11357,7 +11578,7 @@ export declare const MODELS: {
11357
11578
  contextWindow: number;
11358
11579
  maxTokens: number;
11359
11580
  };
11360
- readonly "anthropic/claude-haiku-4.5": {
11581
+ readonly "anthropic/claude-opus-4": {
11361
11582
  id: string;
11362
11583
  name: string;
11363
11584
  api: "anthropic-messages";
@@ -11374,7 +11595,7 @@ export declare const MODELS: {
11374
11595
  contextWindow: number;
11375
11596
  maxTokens: number;
11376
11597
  };
11377
- readonly "anthropic/claude-opus-4": {
11598
+ readonly "anthropic/claude-opus-4.1": {
11378
11599
  id: string;
11379
11600
  name: string;
11380
11601
  api: "anthropic-messages";
@@ -11391,7 +11612,7 @@ export declare const MODELS: {
11391
11612
  contextWindow: number;
11392
11613
  maxTokens: number;
11393
11614
  };
11394
- readonly "anthropic/claude-opus-4.1": {
11615
+ readonly "anthropic/claude-opus-4.5": {
11395
11616
  id: string;
11396
11617
  name: string;
11397
11618
  api: "anthropic-messages";
@@ -11408,7 +11629,7 @@ export declare const MODELS: {
11408
11629
  contextWindow: number;
11409
11630
  maxTokens: number;
11410
11631
  };
11411
- readonly "anthropic/claude-opus-4.5": {
11632
+ readonly "anthropic/claude-opus-4.6": {
11412
11633
  id: string;
11413
11634
  name: string;
11414
11635
  api: "anthropic-messages";
@@ -11425,7 +11646,7 @@ export declare const MODELS: {
11425
11646
  contextWindow: number;
11426
11647
  maxTokens: number;
11427
11648
  };
11428
- readonly "anthropic/claude-opus-4.6": {
11649
+ readonly "anthropic/claude-sonnet-4": {
11429
11650
  id: string;
11430
11651
  name: string;
11431
11652
  api: "anthropic-messages";
@@ -11442,7 +11663,7 @@ export declare const MODELS: {
11442
11663
  contextWindow: number;
11443
11664
  maxTokens: number;
11444
11665
  };
11445
- readonly "anthropic/claude-sonnet-4": {
11666
+ readonly "anthropic/claude-sonnet-4.5": {
11446
11667
  id: string;
11447
11668
  name: string;
11448
11669
  api: "anthropic-messages";
@@ -11459,7 +11680,7 @@ export declare const MODELS: {
11459
11680
  contextWindow: number;
11460
11681
  maxTokens: number;
11461
11682
  };
11462
- readonly "anthropic/claude-sonnet-4.5": {
11683
+ readonly "anthropic/claude-sonnet-4.6": {
11463
11684
  id: string;
11464
11685
  name: string;
11465
11686
  api: "anthropic-messages";
@@ -11476,14 +11697,14 @@ export declare const MODELS: {
11476
11697
  contextWindow: number;
11477
11698
  maxTokens: number;
11478
11699
  };
11479
- readonly "anthropic/claude-sonnet-4.6": {
11700
+ readonly "arcee-ai/trinity-large-preview": {
11480
11701
  id: string;
11481
11702
  name: string;
11482
11703
  api: "anthropic-messages";
11483
11704
  provider: string;
11484
11705
  baseUrl: string;
11485
- reasoning: true;
11486
- input: ("image" | "text")[];
11706
+ reasoning: false;
11707
+ input: "text"[];
11487
11708
  cost: {
11488
11709
  input: number;
11489
11710
  output: number;
@@ -11493,13 +11714,13 @@ export declare const MODELS: {
11493
11714
  contextWindow: number;
11494
11715
  maxTokens: number;
11495
11716
  };
11496
- readonly "arcee-ai/trinity-large-preview": {
11717
+ readonly "arcee-ai/trinity-large-thinking": {
11497
11718
  id: string;
11498
11719
  name: string;
11499
11720
  api: "anthropic-messages";
11500
11721
  provider: string;
11501
11722
  baseUrl: string;
11502
- reasoning: false;
11723
+ reasoning: true;
11503
11724
  input: "text"[];
11504
11725
  cost: {
11505
11726
  input: number;
@@ -11799,6 +12020,40 @@ export declare const MODELS: {
11799
12020
  contextWindow: number;
11800
12021
  maxTokens: number;
11801
12022
  };
12023
+ readonly "google/gemma-4-26b-a4b-it": {
12024
+ id: string;
12025
+ name: string;
12026
+ api: "anthropic-messages";
12027
+ provider: string;
12028
+ baseUrl: string;
12029
+ reasoning: false;
12030
+ input: ("image" | "text")[];
12031
+ cost: {
12032
+ input: number;
12033
+ output: number;
12034
+ cacheRead: number;
12035
+ cacheWrite: number;
12036
+ };
12037
+ contextWindow: number;
12038
+ maxTokens: number;
12039
+ };
12040
+ readonly "google/gemma-4-31b-it": {
12041
+ id: string;
12042
+ name: string;
12043
+ api: "anthropic-messages";
12044
+ provider: string;
12045
+ baseUrl: string;
12046
+ reasoning: false;
12047
+ input: ("image" | "text")[];
12048
+ cost: {
12049
+ input: number;
12050
+ output: number;
12051
+ cacheRead: number;
12052
+ cacheWrite: number;
12053
+ };
12054
+ contextWindow: number;
12055
+ maxTokens: number;
12056
+ };
11802
12057
  readonly "inception/mercury-2": {
11803
12058
  id: string;
11804
12059
  name: string;
@@ -11833,13 +12088,13 @@ export declare const MODELS: {
11833
12088
  contextWindow: number;
11834
12089
  maxTokens: number;
11835
12090
  };
11836
- readonly "meituan/longcat-flash-chat": {
12091
+ readonly "kwaipilot/kat-coder-pro-v2": {
11837
12092
  id: string;
11838
12093
  name: string;
11839
12094
  api: "anthropic-messages";
11840
12095
  provider: string;
11841
12096
  baseUrl: string;
11842
- reasoning: false;
12097
+ reasoning: true;
11843
12098
  input: "text"[];
11844
12099
  cost: {
11845
12100
  input: number;
@@ -11850,13 +12105,13 @@ export declare const MODELS: {
11850
12105
  contextWindow: number;
11851
12106
  maxTokens: number;
11852
12107
  };
11853
- readonly "meituan/longcat-flash-thinking": {
12108
+ readonly "meituan/longcat-flash-chat": {
11854
12109
  id: string;
11855
12110
  name: string;
11856
12111
  api: "anthropic-messages";
11857
12112
  provider: string;
11858
12113
  baseUrl: string;
11859
- reasoning: true;
12114
+ reasoning: false;
11860
12115
  input: "text"[];
11861
12116
  cost: {
11862
12117
  input: number;
@@ -12870,6 +13125,23 @@ export declare const MODELS: {
12870
13125
  contextWindow: number;
12871
13126
  maxTokens: number;
12872
13127
  };
13128
+ readonly "openai/gpt-oss-120b": {
13129
+ id: string;
13130
+ name: string;
13131
+ api: "anthropic-messages";
13132
+ provider: string;
13133
+ baseUrl: string;
13134
+ reasoning: true;
13135
+ input: "text"[];
13136
+ cost: {
13137
+ input: number;
13138
+ output: number;
13139
+ cacheRead: number;
13140
+ cacheWrite: number;
13141
+ };
13142
+ contextWindow: number;
13143
+ maxTokens: number;
13144
+ };
12873
13145
  readonly "openai/gpt-oss-20b": {
12874
13146
  id: string;
12875
13147
  name: string;
@@ -13567,6 +13839,23 @@ export declare const MODELS: {
13567
13839
  contextWindow: number;
13568
13840
  maxTokens: number;
13569
13841
  };
13842
+ readonly "zai/glm-5v-turbo": {
13843
+ id: string;
13844
+ name: string;
13845
+ api: "anthropic-messages";
13846
+ provider: string;
13847
+ baseUrl: string;
13848
+ reasoning: true;
13849
+ input: ("image" | "text")[];
13850
+ cost: {
13851
+ input: number;
13852
+ output: number;
13853
+ cacheRead: number;
13854
+ cacheWrite: number;
13855
+ };
13856
+ contextWindow: number;
13857
+ maxTokens: number;
13858
+ };
13570
13859
  };
13571
13860
  readonly xai: {
13572
13861
  readonly "grok-2": {
@@ -14072,6 +14361,7 @@ export declare const MODELS: {
14072
14361
  compat: {
14073
14362
  supportsDeveloperRole: false;
14074
14363
  thinkingFormat: "zai";
14364
+ zaiToolStream: true;
14075
14365
  };
14076
14366
  reasoning: true;
14077
14367
  input: "text"[];
@@ -14093,6 +14383,7 @@ export declare const MODELS: {
14093
14383
  compat: {
14094
14384
  supportsDeveloperRole: false;
14095
14385
  thinkingFormat: "zai";
14386
+ zaiToolStream: true;
14096
14387
  };
14097
14388
  reasoning: true;
14098
14389
  input: ("image" | "text")[];
@@ -14114,6 +14405,7 @@ export declare const MODELS: {
14114
14405
  compat: {
14115
14406
  supportsDeveloperRole: false;
14116
14407
  thinkingFormat: "zai";
14408
+ zaiToolStream: true;
14117
14409
  };
14118
14410
  reasoning: true;
14119
14411
  input: "text"[];
@@ -14135,6 +14427,29 @@ export declare const MODELS: {
14135
14427
  compat: {
14136
14428
  supportsDeveloperRole: false;
14137
14429
  thinkingFormat: "zai";
14430
+ zaiToolStream: true;
14431
+ };
14432
+ reasoning: true;
14433
+ input: "text"[];
14434
+ cost: {
14435
+ input: number;
14436
+ output: number;
14437
+ cacheRead: number;
14438
+ cacheWrite: number;
14439
+ };
14440
+ contextWindow: number;
14441
+ maxTokens: number;
14442
+ };
14443
+ readonly "glm-4.7-flashx": {
14444
+ id: string;
14445
+ name: string;
14446
+ api: "openai-completions";
14447
+ provider: string;
14448
+ baseUrl: string;
14449
+ compat: {
14450
+ supportsDeveloperRole: false;
14451
+ thinkingFormat: "zai";
14452
+ zaiToolStream: true;
14138
14453
  };
14139
14454
  reasoning: true;
14140
14455
  input: "text"[];
@@ -14156,6 +14471,7 @@ export declare const MODELS: {
14156
14471
  compat: {
14157
14472
  supportsDeveloperRole: false;
14158
14473
  thinkingFormat: "zai";
14474
+ zaiToolStream: true;
14159
14475
  };
14160
14476
  reasoning: true;
14161
14477
  input: "text"[];
@@ -14177,6 +14493,7 @@ export declare const MODELS: {
14177
14493
  compat: {
14178
14494
  supportsDeveloperRole: false;
14179
14495
  thinkingFormat: "zai";
14496
+ zaiToolStream: true;
14180
14497
  };
14181
14498
  reasoning: true;
14182
14499
  input: "text"[];
@@ -14189,6 +14506,28 @@ export declare const MODELS: {
14189
14506
  contextWindow: number;
14190
14507
  maxTokens: number;
14191
14508
  };
14509
+ readonly "glm-5v-turbo": {
14510
+ id: string;
14511
+ name: string;
14512
+ api: "openai-completions";
14513
+ provider: string;
14514
+ baseUrl: string;
14515
+ compat: {
14516
+ supportsDeveloperRole: false;
14517
+ thinkingFormat: "zai";
14518
+ zaiToolStream: true;
14519
+ };
14520
+ reasoning: true;
14521
+ input: ("image" | "text")[];
14522
+ cost: {
14523
+ input: number;
14524
+ output: number;
14525
+ cacheRead: number;
14526
+ cacheWrite: number;
14527
+ };
14528
+ contextWindow: number;
14529
+ maxTokens: number;
14530
+ };
14192
14531
  };
14193
14532
  };
14194
14533
  //# sourceMappingURL=models.generated.d.ts.map