@avadisabelle/ava-pi-ai 0.63.2 → 0.64.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1549,23 +1549,6 @@ export declare const MODELS: {
1549
1549
  contextWindow: number;
1550
1550
  maxTokens: number;
1551
1551
  };
1552
- readonly "claude-3-7-sonnet-latest": {
1553
- id: string;
1554
- name: string;
1555
- api: "anthropic-messages";
1556
- provider: string;
1557
- baseUrl: string;
1558
- reasoning: true;
1559
- input: ("image" | "text")[];
1560
- cost: {
1561
- input: number;
1562
- output: number;
1563
- cacheRead: number;
1564
- cacheWrite: number;
1565
- };
1566
- contextWindow: number;
1567
- maxTokens: number;
1568
- };
1569
1552
  readonly "claude-3-haiku-20240307": {
1570
1553
  id: string;
1571
1554
  name: string;
@@ -3640,6 +3623,40 @@ export declare const MODELS: {
3640
3623
  contextWindow: number;
3641
3624
  maxTokens: number;
3642
3625
  };
3626
+ readonly "gemma-4-26b": {
3627
+ id: string;
3628
+ name: string;
3629
+ api: "google-generative-ai";
3630
+ provider: string;
3631
+ baseUrl: string;
3632
+ reasoning: true;
3633
+ input: ("image" | "text")[];
3634
+ cost: {
3635
+ input: number;
3636
+ output: number;
3637
+ cacheRead: number;
3638
+ cacheWrite: number;
3639
+ };
3640
+ contextWindow: number;
3641
+ maxTokens: number;
3642
+ };
3643
+ readonly "gemma-4-31b": {
3644
+ id: string;
3645
+ name: string;
3646
+ api: "google-generative-ai";
3647
+ provider: string;
3648
+ baseUrl: string;
3649
+ reasoning: true;
3650
+ input: ("image" | "text")[];
3651
+ cost: {
3652
+ input: number;
3653
+ output: number;
3654
+ cacheRead: number;
3655
+ cacheWrite: number;
3656
+ };
3657
+ contextWindow: number;
3658
+ maxTokens: number;
3659
+ };
3643
3660
  };
3644
3661
  readonly "google-antigravity": {
3645
3662
  readonly "claude-opus-4-5-thinking": {
@@ -6812,14 +6829,14 @@ export declare const MODELS: {
6812
6829
  contextWindow: number;
6813
6830
  maxTokens: number;
6814
6831
  };
6815
- readonly "mimo-v2-omni-free": {
6832
+ readonly "minimax-m2.5": {
6816
6833
  id: string;
6817
6834
  name: string;
6818
6835
  api: "openai-completions";
6819
6836
  provider: string;
6820
6837
  baseUrl: string;
6821
6838
  reasoning: true;
6822
- input: ("image" | "text")[];
6839
+ input: "text"[];
6823
6840
  cost: {
6824
6841
  input: number;
6825
6842
  output: number;
@@ -6829,10 +6846,10 @@ export declare const MODELS: {
6829
6846
  contextWindow: number;
6830
6847
  maxTokens: number;
6831
6848
  };
6832
- readonly "mimo-v2-pro-free": {
6849
+ readonly "minimax-m2.5-free": {
6833
6850
  id: string;
6834
6851
  name: string;
6835
- api: "openai-completions";
6852
+ api: "anthropic-messages";
6836
6853
  provider: string;
6837
6854
  baseUrl: string;
6838
6855
  reasoning: true;
@@ -6846,7 +6863,7 @@ export declare const MODELS: {
6846
6863
  contextWindow: number;
6847
6864
  maxTokens: number;
6848
6865
  };
6849
- readonly "minimax-m2.5": {
6866
+ readonly "nemotron-3-super-free": {
6850
6867
  id: string;
6851
6868
  name: string;
6852
6869
  api: "openai-completions";
@@ -6863,10 +6880,10 @@ export declare const MODELS: {
6863
6880
  contextWindow: number;
6864
6881
  maxTokens: number;
6865
6882
  };
6866
- readonly "minimax-m2.5-free": {
6883
+ readonly "qwen3.6-plus-free": {
6867
6884
  id: string;
6868
6885
  name: string;
6869
- api: "anthropic-messages";
6886
+ api: "openai-completions";
6870
6887
  provider: string;
6871
6888
  baseUrl: string;
6872
6889
  reasoning: true;
@@ -6880,7 +6897,9 @@ export declare const MODELS: {
6880
6897
  contextWindow: number;
6881
6898
  maxTokens: number;
6882
6899
  };
6883
- readonly "nemotron-3-super-free": {
6900
+ };
6901
+ readonly "opencode-go": {
6902
+ readonly "glm-5": {
6884
6903
  id: string;
6885
6904
  name: string;
6886
6905
  api: "openai-completions";
@@ -6897,14 +6916,14 @@ export declare const MODELS: {
6897
6916
  contextWindow: number;
6898
6917
  maxTokens: number;
6899
6918
  };
6900
- readonly "qwen3.6-plus-free": {
6919
+ readonly "kimi-k2.5": {
6901
6920
  id: string;
6902
6921
  name: string;
6903
6922
  api: "openai-completions";
6904
6923
  provider: string;
6905
6924
  baseUrl: string;
6906
6925
  reasoning: true;
6907
- input: "text"[];
6926
+ input: ("image" | "text")[];
6908
6927
  cost: {
6909
6928
  input: number;
6910
6929
  output: number;
@@ -6914,16 +6933,14 @@ export declare const MODELS: {
6914
6933
  contextWindow: number;
6915
6934
  maxTokens: number;
6916
6935
  };
6917
- };
6918
- readonly "opencode-go": {
6919
- readonly "glm-5": {
6936
+ readonly "mimo-v2-omni": {
6920
6937
  id: string;
6921
6938
  name: string;
6922
6939
  api: "openai-completions";
6923
6940
  provider: string;
6924
6941
  baseUrl: string;
6925
6942
  reasoning: true;
6926
- input: "text"[];
6943
+ input: ("image" | "text")[];
6927
6944
  cost: {
6928
6945
  input: number;
6929
6946
  output: number;
@@ -6933,14 +6950,14 @@ export declare const MODELS: {
6933
6950
  contextWindow: number;
6934
6951
  maxTokens: number;
6935
6952
  };
6936
- readonly "kimi-k2.5": {
6953
+ readonly "mimo-v2-pro": {
6937
6954
  id: string;
6938
6955
  name: string;
6939
6956
  api: "openai-completions";
6940
6957
  provider: string;
6941
6958
  baseUrl: string;
6942
6959
  reasoning: true;
6943
- input: ("image" | "text")[];
6960
+ input: "text"[];
6944
6961
  cost: {
6945
6962
  input: number;
6946
6963
  output: number;
@@ -6953,7 +6970,7 @@ export declare const MODELS: {
6953
6970
  readonly "minimax-m2.5": {
6954
6971
  id: string;
6955
6972
  name: string;
6956
- api: "anthropic-messages";
6973
+ api: "openai-completions";
6957
6974
  provider: string;
6958
6975
  baseUrl: string;
6959
6976
  reasoning: true;
@@ -7156,13 +7173,13 @@ export declare const MODELS: {
7156
7173
  contextWindow: number;
7157
7174
  maxTokens: number;
7158
7175
  };
7159
- readonly "anthropic/claude-3.5-sonnet": {
7176
+ readonly "anthropic/claude-3.7-sonnet": {
7160
7177
  id: string;
7161
7178
  name: string;
7162
7179
  api: "openai-completions";
7163
7180
  provider: string;
7164
7181
  baseUrl: string;
7165
- reasoning: false;
7182
+ reasoning: true;
7166
7183
  input: ("image" | "text")[];
7167
7184
  cost: {
7168
7185
  input: number;
@@ -7173,7 +7190,7 @@ export declare const MODELS: {
7173
7190
  contextWindow: number;
7174
7191
  maxTokens: number;
7175
7192
  };
7176
- readonly "anthropic/claude-3.7-sonnet": {
7193
+ readonly "anthropic/claude-3.7-sonnet:thinking": {
7177
7194
  id: string;
7178
7195
  name: string;
7179
7196
  api: "openai-completions";
@@ -7190,7 +7207,7 @@ export declare const MODELS: {
7190
7207
  contextWindow: number;
7191
7208
  maxTokens: number;
7192
7209
  };
7193
- readonly "anthropic/claude-3.7-sonnet:thinking": {
7210
+ readonly "anthropic/claude-haiku-4.5": {
7194
7211
  id: string;
7195
7212
  name: string;
7196
7213
  api: "openai-completions";
@@ -7207,7 +7224,7 @@ export declare const MODELS: {
7207
7224
  contextWindow: number;
7208
7225
  maxTokens: number;
7209
7226
  };
7210
- readonly "anthropic/claude-haiku-4.5": {
7227
+ readonly "anthropic/claude-opus-4": {
7211
7228
  id: string;
7212
7229
  name: string;
7213
7230
  api: "openai-completions";
@@ -7224,7 +7241,7 @@ export declare const MODELS: {
7224
7241
  contextWindow: number;
7225
7242
  maxTokens: number;
7226
7243
  };
7227
- readonly "anthropic/claude-opus-4": {
7244
+ readonly "anthropic/claude-opus-4.1": {
7228
7245
  id: string;
7229
7246
  name: string;
7230
7247
  api: "openai-completions";
@@ -7241,7 +7258,7 @@ export declare const MODELS: {
7241
7258
  contextWindow: number;
7242
7259
  maxTokens: number;
7243
7260
  };
7244
- readonly "anthropic/claude-opus-4.1": {
7261
+ readonly "anthropic/claude-opus-4.5": {
7245
7262
  id: string;
7246
7263
  name: string;
7247
7264
  api: "openai-completions";
@@ -7258,7 +7275,7 @@ export declare const MODELS: {
7258
7275
  contextWindow: number;
7259
7276
  maxTokens: number;
7260
7277
  };
7261
- readonly "anthropic/claude-opus-4.5": {
7278
+ readonly "anthropic/claude-opus-4.6": {
7262
7279
  id: string;
7263
7280
  name: string;
7264
7281
  api: "openai-completions";
@@ -7275,7 +7292,7 @@ export declare const MODELS: {
7275
7292
  contextWindow: number;
7276
7293
  maxTokens: number;
7277
7294
  };
7278
- readonly "anthropic/claude-opus-4.6": {
7295
+ readonly "anthropic/claude-sonnet-4": {
7279
7296
  id: string;
7280
7297
  name: string;
7281
7298
  api: "openai-completions";
@@ -7292,7 +7309,7 @@ export declare const MODELS: {
7292
7309
  contextWindow: number;
7293
7310
  maxTokens: number;
7294
7311
  };
7295
- readonly "anthropic/claude-sonnet-4": {
7312
+ readonly "anthropic/claude-sonnet-4.5": {
7296
7313
  id: string;
7297
7314
  name: string;
7298
7315
  api: "openai-completions";
@@ -7309,7 +7326,7 @@ export declare const MODELS: {
7309
7326
  contextWindow: number;
7310
7327
  maxTokens: number;
7311
7328
  };
7312
- readonly "anthropic/claude-sonnet-4.5": {
7329
+ readonly "anthropic/claude-sonnet-4.6": {
7313
7330
  id: string;
7314
7331
  name: string;
7315
7332
  api: "openai-completions";
@@ -7326,14 +7343,14 @@ export declare const MODELS: {
7326
7343
  contextWindow: number;
7327
7344
  maxTokens: number;
7328
7345
  };
7329
- readonly "anthropic/claude-sonnet-4.6": {
7346
+ readonly "arcee-ai/trinity-large-preview:free": {
7330
7347
  id: string;
7331
7348
  name: string;
7332
7349
  api: "openai-completions";
7333
7350
  provider: string;
7334
7351
  baseUrl: string;
7335
- reasoning: true;
7336
- input: ("image" | "text")[];
7352
+ reasoning: false;
7353
+ input: "text"[];
7337
7354
  cost: {
7338
7355
  input: number;
7339
7356
  output: number;
@@ -7343,13 +7360,13 @@ export declare const MODELS: {
7343
7360
  contextWindow: number;
7344
7361
  maxTokens: number;
7345
7362
  };
7346
- readonly "arcee-ai/trinity-large-preview:free": {
7363
+ readonly "arcee-ai/trinity-large-thinking": {
7347
7364
  id: string;
7348
7365
  name: string;
7349
7366
  api: "openai-completions";
7350
7367
  provider: string;
7351
7368
  baseUrl: string;
7352
- reasoning: false;
7369
+ reasoning: true;
7353
7370
  input: "text"[];
7354
7371
  cost: {
7355
7372
  input: number;
@@ -7921,6 +7938,40 @@ export declare const MODELS: {
7921
7938
  contextWindow: number;
7922
7939
  maxTokens: number;
7923
7940
  };
7941
+ readonly "google/gemma-4-26b-a4b-it": {
7942
+ id: string;
7943
+ name: string;
7944
+ api: "openai-completions";
7945
+ provider: string;
7946
+ baseUrl: string;
7947
+ reasoning: true;
7948
+ input: ("image" | "text")[];
7949
+ cost: {
7950
+ input: number;
7951
+ output: number;
7952
+ cacheRead: number;
7953
+ cacheWrite: number;
7954
+ };
7955
+ contextWindow: number;
7956
+ maxTokens: number;
7957
+ };
7958
+ readonly "google/gemma-4-31b-it": {
7959
+ id: string;
7960
+ name: string;
7961
+ api: "openai-completions";
7962
+ provider: string;
7963
+ baseUrl: string;
7964
+ reasoning: true;
7965
+ input: ("image" | "text")[];
7966
+ cost: {
7967
+ input: number;
7968
+ output: number;
7969
+ cacheRead: number;
7970
+ cacheWrite: number;
7971
+ };
7972
+ contextWindow: number;
7973
+ maxTokens: number;
7974
+ };
7924
7975
  readonly "inception/mercury": {
7925
7976
  id: string;
7926
7977
  name: string;
@@ -8482,23 +8533,6 @@ export declare const MODELS: {
8482
8533
  contextWindow: number;
8483
8534
  maxTokens: number;
8484
8535
  };
8485
- readonly "mistralai/mistral-small-24b-instruct-2501": {
8486
- id: string;
8487
- name: string;
8488
- api: "openai-completions";
8489
- provider: string;
8490
- baseUrl: string;
8491
- reasoning: false;
8492
- input: "text"[];
8493
- cost: {
8494
- input: number;
8495
- output: number;
8496
- cacheRead: number;
8497
- cacheWrite: number;
8498
- };
8499
- contextWindow: number;
8500
- maxTokens: number;
8501
- };
8502
8536
  readonly "mistralai/mistral-small-2603": {
8503
8537
  id: string;
8504
8538
  name: string;
@@ -10590,14 +10624,14 @@ export declare const MODELS: {
10590
10624
  contextWindow: number;
10591
10625
  maxTokens: number;
10592
10626
  };
10593
- readonly "qwen/qwen3.6-plus-preview:free": {
10627
+ readonly "qwen/qwen3.6-plus:free": {
10594
10628
  id: string;
10595
10629
  name: string;
10596
10630
  api: "openai-completions";
10597
10631
  provider: string;
10598
10632
  baseUrl: string;
10599
10633
  reasoning: true;
10600
- input: "text"[];
10634
+ input: ("image" | "text")[];
10601
10635
  cost: {
10602
10636
  input: number;
10603
10637
  output: number;
@@ -10624,7 +10658,7 @@ export declare const MODELS: {
10624
10658
  contextWindow: number;
10625
10659
  maxTokens: number;
10626
10660
  };
10627
- readonly "reka/reka-edge": {
10661
+ readonly "rekaai/reka-edge": {
10628
10662
  id: string;
10629
10663
  name: string;
10630
10664
  api: "openai-completions";
@@ -11185,6 +11219,23 @@ export declare const MODELS: {
11185
11219
  contextWindow: number;
11186
11220
  maxTokens: number;
11187
11221
  };
11222
+ readonly "z-ai/glm-5v-turbo": {
11223
+ id: string;
11224
+ name: string;
11225
+ api: "openai-completions";
11226
+ provider: string;
11227
+ baseUrl: string;
11228
+ reasoning: true;
11229
+ input: ("image" | "text")[];
11230
+ cost: {
11231
+ input: number;
11232
+ output: number;
11233
+ cacheRead: number;
11234
+ cacheWrite: number;
11235
+ };
11236
+ contextWindow: number;
11237
+ maxTokens: number;
11238
+ };
11188
11239
  };
11189
11240
  readonly "vercel-ai-gateway": {
11190
11241
  readonly "alibaba/qwen-3-14b": {
@@ -11442,13 +11493,13 @@ export declare const MODELS: {
11442
11493
  contextWindow: number;
11443
11494
  maxTokens: number;
11444
11495
  };
11445
- readonly "anthropic/claude-3-haiku": {
11496
+ readonly "alibaba/qwen3.6-plus": {
11446
11497
  id: string;
11447
11498
  name: string;
11448
11499
  api: "anthropic-messages";
11449
11500
  provider: string;
11450
11501
  baseUrl: string;
11451
- reasoning: false;
11502
+ reasoning: true;
11452
11503
  input: ("image" | "text")[];
11453
11504
  cost: {
11454
11505
  input: number;
@@ -11459,7 +11510,7 @@ export declare const MODELS: {
11459
11510
  contextWindow: number;
11460
11511
  maxTokens: number;
11461
11512
  };
11462
- readonly "anthropic/claude-3.5-haiku": {
11513
+ readonly "anthropic/claude-3-haiku": {
11463
11514
  id: string;
11464
11515
  name: string;
11465
11516
  api: "anthropic-messages";
@@ -11476,7 +11527,7 @@ export declare const MODELS: {
11476
11527
  contextWindow: number;
11477
11528
  maxTokens: number;
11478
11529
  };
11479
- readonly "anthropic/claude-3.5-sonnet": {
11530
+ readonly "anthropic/claude-3.5-haiku": {
11480
11531
  id: string;
11481
11532
  name: string;
11482
11533
  api: "anthropic-messages";
@@ -11493,13 +11544,13 @@ export declare const MODELS: {
11493
11544
  contextWindow: number;
11494
11545
  maxTokens: number;
11495
11546
  };
11496
- readonly "anthropic/claude-3.5-sonnet-20240620": {
11547
+ readonly "anthropic/claude-3.7-sonnet": {
11497
11548
  id: string;
11498
11549
  name: string;
11499
11550
  api: "anthropic-messages";
11500
11551
  provider: string;
11501
11552
  baseUrl: string;
11502
- reasoning: false;
11553
+ reasoning: true;
11503
11554
  input: ("image" | "text")[];
11504
11555
  cost: {
11505
11556
  input: number;
@@ -11510,7 +11561,7 @@ export declare const MODELS: {
11510
11561
  contextWindow: number;
11511
11562
  maxTokens: number;
11512
11563
  };
11513
- readonly "anthropic/claude-3.7-sonnet": {
11564
+ readonly "anthropic/claude-haiku-4.5": {
11514
11565
  id: string;
11515
11566
  name: string;
11516
11567
  api: "anthropic-messages";
@@ -11527,7 +11578,7 @@ export declare const MODELS: {
11527
11578
  contextWindow: number;
11528
11579
  maxTokens: number;
11529
11580
  };
11530
- readonly "anthropic/claude-haiku-4.5": {
11581
+ readonly "anthropic/claude-opus-4": {
11531
11582
  id: string;
11532
11583
  name: string;
11533
11584
  api: "anthropic-messages";
@@ -11544,7 +11595,7 @@ export declare const MODELS: {
11544
11595
  contextWindow: number;
11545
11596
  maxTokens: number;
11546
11597
  };
11547
- readonly "anthropic/claude-opus-4": {
11598
+ readonly "anthropic/claude-opus-4.1": {
11548
11599
  id: string;
11549
11600
  name: string;
11550
11601
  api: "anthropic-messages";
@@ -11561,7 +11612,7 @@ export declare const MODELS: {
11561
11612
  contextWindow: number;
11562
11613
  maxTokens: number;
11563
11614
  };
11564
- readonly "anthropic/claude-opus-4.1": {
11615
+ readonly "anthropic/claude-opus-4.5": {
11565
11616
  id: string;
11566
11617
  name: string;
11567
11618
  api: "anthropic-messages";
@@ -11578,7 +11629,7 @@ export declare const MODELS: {
11578
11629
  contextWindow: number;
11579
11630
  maxTokens: number;
11580
11631
  };
11581
- readonly "anthropic/claude-opus-4.5": {
11632
+ readonly "anthropic/claude-opus-4.6": {
11582
11633
  id: string;
11583
11634
  name: string;
11584
11635
  api: "anthropic-messages";
@@ -11595,7 +11646,7 @@ export declare const MODELS: {
11595
11646
  contextWindow: number;
11596
11647
  maxTokens: number;
11597
11648
  };
11598
- readonly "anthropic/claude-opus-4.6": {
11649
+ readonly "anthropic/claude-sonnet-4": {
11599
11650
  id: string;
11600
11651
  name: string;
11601
11652
  api: "anthropic-messages";
@@ -11612,7 +11663,7 @@ export declare const MODELS: {
11612
11663
  contextWindow: number;
11613
11664
  maxTokens: number;
11614
11665
  };
11615
- readonly "anthropic/claude-sonnet-4": {
11666
+ readonly "anthropic/claude-sonnet-4.5": {
11616
11667
  id: string;
11617
11668
  name: string;
11618
11669
  api: "anthropic-messages";
@@ -11629,7 +11680,7 @@ export declare const MODELS: {
11629
11680
  contextWindow: number;
11630
11681
  maxTokens: number;
11631
11682
  };
11632
- readonly "anthropic/claude-sonnet-4.5": {
11683
+ readonly "anthropic/claude-sonnet-4.6": {
11633
11684
  id: string;
11634
11685
  name: string;
11635
11686
  api: "anthropic-messages";
@@ -11646,14 +11697,14 @@ export declare const MODELS: {
11646
11697
  contextWindow: number;
11647
11698
  maxTokens: number;
11648
11699
  };
11649
- readonly "anthropic/claude-sonnet-4.6": {
11700
+ readonly "arcee-ai/trinity-large-preview": {
11650
11701
  id: string;
11651
11702
  name: string;
11652
11703
  api: "anthropic-messages";
11653
11704
  provider: string;
11654
11705
  baseUrl: string;
11655
- reasoning: true;
11656
- input: ("image" | "text")[];
11706
+ reasoning: false;
11707
+ input: "text"[];
11657
11708
  cost: {
11658
11709
  input: number;
11659
11710
  output: number;
@@ -11663,13 +11714,13 @@ export declare const MODELS: {
11663
11714
  contextWindow: number;
11664
11715
  maxTokens: number;
11665
11716
  };
11666
- readonly "arcee-ai/trinity-large-preview": {
11717
+ readonly "arcee-ai/trinity-large-thinking": {
11667
11718
  id: string;
11668
11719
  name: string;
11669
11720
  api: "anthropic-messages";
11670
11721
  provider: string;
11671
11722
  baseUrl: string;
11672
- reasoning: false;
11723
+ reasoning: true;
11673
11724
  input: "text"[];
11674
11725
  cost: {
11675
11726
  input: number;
@@ -11969,14 +12020,14 @@ export declare const MODELS: {
11969
12020
  contextWindow: number;
11970
12021
  maxTokens: number;
11971
12022
  };
11972
- readonly "inception/mercury-2": {
12023
+ readonly "google/gemma-4-26b-a4b-it": {
11973
12024
  id: string;
11974
12025
  name: string;
11975
12026
  api: "anthropic-messages";
11976
12027
  provider: string;
11977
12028
  baseUrl: string;
11978
- reasoning: true;
11979
- input: "text"[];
12029
+ reasoning: false;
12030
+ input: ("image" | "text")[];
11980
12031
  cost: {
11981
12032
  input: number;
11982
12033
  output: number;
@@ -11986,14 +12037,14 @@ export declare const MODELS: {
11986
12037
  contextWindow: number;
11987
12038
  maxTokens: number;
11988
12039
  };
11989
- readonly "inception/mercury-coder-small": {
12040
+ readonly "google/gemma-4-31b-it": {
11990
12041
  id: string;
11991
12042
  name: string;
11992
12043
  api: "anthropic-messages";
11993
12044
  provider: string;
11994
12045
  baseUrl: string;
11995
12046
  reasoning: false;
11996
- input: "text"[];
12047
+ input: ("image" | "text")[];
11997
12048
  cost: {
11998
12049
  input: number;
11999
12050
  output: number;
@@ -12003,7 +12054,7 @@ export declare const MODELS: {
12003
12054
  contextWindow: number;
12004
12055
  maxTokens: number;
12005
12056
  };
12006
- readonly "kwaipilot/kat-coder-pro-v2": {
12057
+ readonly "inception/mercury-2": {
12007
12058
  id: string;
12008
12059
  name: string;
12009
12060
  api: "anthropic-messages";
@@ -12020,7 +12071,7 @@ export declare const MODELS: {
12020
12071
  contextWindow: number;
12021
12072
  maxTokens: number;
12022
12073
  };
12023
- readonly "meituan/longcat-flash-chat": {
12074
+ readonly "inception/mercury-coder-small": {
12024
12075
  id: string;
12025
12076
  name: string;
12026
12077
  api: "anthropic-messages";
@@ -12037,7 +12088,7 @@ export declare const MODELS: {
12037
12088
  contextWindow: number;
12038
12089
  maxTokens: number;
12039
12090
  };
12040
- readonly "meituan/longcat-flash-thinking": {
12091
+ readonly "kwaipilot/kat-coder-pro-v2": {
12041
12092
  id: string;
12042
12093
  name: string;
12043
12094
  api: "anthropic-messages";
@@ -12054,6 +12105,23 @@ export declare const MODELS: {
12054
12105
  contextWindow: number;
12055
12106
  maxTokens: number;
12056
12107
  };
12108
+ readonly "meituan/longcat-flash-chat": {
12109
+ id: string;
12110
+ name: string;
12111
+ api: "anthropic-messages";
12112
+ provider: string;
12113
+ baseUrl: string;
12114
+ reasoning: false;
12115
+ input: "text"[];
12116
+ cost: {
12117
+ input: number;
12118
+ output: number;
12119
+ cacheRead: number;
12120
+ cacheWrite: number;
12121
+ };
12122
+ contextWindow: number;
12123
+ maxTokens: number;
12124
+ };
12057
12125
  readonly "meta/llama-3.1-70b": {
12058
12126
  id: string;
12059
12127
  name: string;
@@ -13771,6 +13839,23 @@ export declare const MODELS: {
13771
13839
  contextWindow: number;
13772
13840
  maxTokens: number;
13773
13841
  };
13842
+ readonly "zai/glm-5v-turbo": {
13843
+ id: string;
13844
+ name: string;
13845
+ api: "anthropic-messages";
13846
+ provider: string;
13847
+ baseUrl: string;
13848
+ reasoning: true;
13849
+ input: ("image" | "text")[];
13850
+ cost: {
13851
+ input: number;
13852
+ output: number;
13853
+ cacheRead: number;
13854
+ cacheWrite: number;
13855
+ };
13856
+ contextWindow: number;
13857
+ maxTokens: number;
13858
+ };
13774
13859
  };
13775
13860
  readonly xai: {
13776
13861
  readonly "grok-2": {
@@ -14414,6 +14499,27 @@ export declare const MODELS: {
14414
14499
  contextWindow: number;
14415
14500
  maxTokens: number;
14416
14501
  };
14502
+ readonly "glm-5v-turbo": {
14503
+ id: string;
14504
+ name: string;
14505
+ api: "openai-completions";
14506
+ provider: string;
14507
+ baseUrl: string;
14508
+ compat: {
14509
+ supportsDeveloperRole: false;
14510
+ thinkingFormat: "zai";
14511
+ };
14512
+ reasoning: true;
14513
+ input: ("image" | "text")[];
14514
+ cost: {
14515
+ input: number;
14516
+ output: number;
14517
+ cacheRead: number;
14518
+ cacheWrite: number;
14519
+ };
14520
+ contextWindow: number;
14521
+ maxTokens: number;
14522
+ };
14417
14523
  };
14418
14524
  };
14419
14525
  //# sourceMappingURL=models.generated.d.ts.map