@avadisabelle/ava-pi-ai 0.64.1 → 0.64.3

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -255,6 +255,23 @@ export declare const MODELS: {
255
255
  contextWindow: number;
256
256
  maxTokens: number;
257
257
  };
258
+ readonly "anthropic.claude-opus-4-7": {
259
+ id: string;
260
+ name: string;
261
+ api: "bedrock-converse-stream";
262
+ provider: string;
263
+ baseUrl: string;
264
+ reasoning: true;
265
+ input: ("image" | "text")[];
266
+ cost: {
267
+ input: number;
268
+ output: number;
269
+ cacheRead: number;
270
+ cacheWrite: number;
271
+ };
272
+ contextWindow: number;
273
+ maxTokens: number;
274
+ };
258
275
  readonly "anthropic.claude-sonnet-4-20250514-v1:0": {
259
276
  id: string;
260
277
  name: string;
@@ -408,6 +425,23 @@ export declare const MODELS: {
408
425
  contextWindow: number;
409
426
  maxTokens: number;
410
427
  };
428
+ readonly "eu.anthropic.claude-opus-4-7": {
429
+ id: string;
430
+ name: string;
431
+ api: "bedrock-converse-stream";
432
+ provider: string;
433
+ baseUrl: string;
434
+ reasoning: true;
435
+ input: ("image" | "text")[];
436
+ cost: {
437
+ input: number;
438
+ output: number;
439
+ cacheRead: number;
440
+ cacheWrite: number;
441
+ };
442
+ contextWindow: number;
443
+ maxTokens: number;
444
+ };
411
445
  readonly "eu.anthropic.claude-sonnet-4-20250514-v1:0": {
412
446
  id: string;
413
447
  name: string;
@@ -510,6 +544,23 @@ export declare const MODELS: {
510
544
  contextWindow: number;
511
545
  maxTokens: number;
512
546
  };
547
+ readonly "global.anthropic.claude-opus-4-7": {
548
+ id: string;
549
+ name: string;
550
+ api: "bedrock-converse-stream";
551
+ provider: string;
552
+ baseUrl: string;
553
+ reasoning: true;
554
+ input: ("image" | "text")[];
555
+ cost: {
556
+ input: number;
557
+ output: number;
558
+ cacheRead: number;
559
+ cacheWrite: number;
560
+ };
561
+ contextWindow: number;
562
+ maxTokens: number;
563
+ };
513
564
  readonly "global.anthropic.claude-sonnet-4-20250514-v1:0": {
514
565
  id: string;
515
566
  name: string;
@@ -1207,6 +1258,23 @@ export declare const MODELS: {
1207
1258
  contextWindow: number;
1208
1259
  maxTokens: number;
1209
1260
  };
1261
+ readonly "qwen.qwen3-coder-next": {
1262
+ id: string;
1263
+ name: string;
1264
+ api: "bedrock-converse-stream";
1265
+ provider: string;
1266
+ baseUrl: string;
1267
+ reasoning: true;
1268
+ input: "text"[];
1269
+ cost: {
1270
+ input: number;
1271
+ output: number;
1272
+ cacheRead: number;
1273
+ cacheWrite: number;
1274
+ };
1275
+ contextWindow: number;
1276
+ maxTokens: number;
1277
+ };
1210
1278
  readonly "qwen.qwen3-next-80b-a3b": {
1211
1279
  id: string;
1212
1280
  name: string;
@@ -1326,6 +1394,23 @@ export declare const MODELS: {
1326
1394
  contextWindow: number;
1327
1395
  maxTokens: number;
1328
1396
  };
1397
+ readonly "us.anthropic.claude-opus-4-7": {
1398
+ id: string;
1399
+ name: string;
1400
+ api: "bedrock-converse-stream";
1401
+ provider: string;
1402
+ baseUrl: string;
1403
+ reasoning: true;
1404
+ input: ("image" | "text")[];
1405
+ cost: {
1406
+ input: number;
1407
+ output: number;
1408
+ cacheRead: number;
1409
+ cacheWrite: number;
1410
+ };
1411
+ contextWindow: number;
1412
+ maxTokens: number;
1413
+ };
1329
1414
  readonly "us.anthropic.claude-sonnet-4-20250514-v1:0": {
1330
1415
  id: string;
1331
1416
  name: string;
@@ -1753,6 +1838,23 @@ export declare const MODELS: {
1753
1838
  contextWindow: number;
1754
1839
  maxTokens: number;
1755
1840
  };
1841
+ readonly "claude-opus-4-7": {
1842
+ id: string;
1843
+ name: string;
1844
+ api: "anthropic-messages";
1845
+ provider: string;
1846
+ baseUrl: string;
1847
+ reasoning: true;
1848
+ input: ("image" | "text")[];
1849
+ cost: {
1850
+ input: number;
1851
+ output: number;
1852
+ cacheRead: number;
1853
+ cacheWrite: number;
1854
+ };
1855
+ contextWindow: number;
1856
+ maxTokens: number;
1857
+ };
1756
1858
  readonly "claude-sonnet-4-0": {
1757
1859
  id: string;
1758
1860
  name: string;
@@ -3623,7 +3725,7 @@ export declare const MODELS: {
3623
3725
  contextWindow: number;
3624
3726
  maxTokens: number;
3625
3727
  };
3626
- readonly "gemma-4-26b": {
3728
+ readonly "gemma-4-26b-it": {
3627
3729
  id: string;
3628
3730
  name: string;
3629
3731
  api: "google-generative-ai";
@@ -3640,7 +3742,7 @@ export declare const MODELS: {
3640
3742
  contextWindow: number;
3641
3743
  maxTokens: number;
3642
3744
  };
3643
- readonly "gemma-4-31b": {
3745
+ readonly "gemma-4-31b-it": {
3644
3746
  id: string;
3645
3747
  name: string;
3646
3748
  api: "google-generative-ai";
@@ -4472,6 +4574,26 @@ export declare const MODELS: {
4472
4574
  contextWindow: number;
4473
4575
  maxTokens: number;
4474
4576
  };
4577
+ readonly "MiniMaxAI/MiniMax-M2.7": {
4578
+ id: string;
4579
+ name: string;
4580
+ api: "openai-completions";
4581
+ provider: string;
4582
+ baseUrl: string;
4583
+ compat: {
4584
+ supportsDeveloperRole: false;
4585
+ };
4586
+ reasoning: true;
4587
+ input: "text"[];
4588
+ cost: {
4589
+ input: number;
4590
+ output: number;
4591
+ cacheRead: number;
4592
+ cacheWrite: number;
4593
+ };
4594
+ contextWindow: number;
4595
+ maxTokens: number;
4596
+ };
4475
4597
  readonly "Qwen/Qwen3-235B-A22B-Thinking-2507": {
4476
4598
  id: string;
4477
4599
  name: string;
@@ -4792,6 +4914,26 @@ export declare const MODELS: {
4792
4914
  contextWindow: number;
4793
4915
  maxTokens: number;
4794
4916
  };
4917
+ readonly "zai-org/GLM-5.1": {
4918
+ id: string;
4919
+ name: string;
4920
+ api: "openai-completions";
4921
+ provider: string;
4922
+ baseUrl: string;
4923
+ compat: {
4924
+ supportsDeveloperRole: false;
4925
+ };
4926
+ reasoning: true;
4927
+ input: "text"[];
4928
+ cost: {
4929
+ input: number;
4930
+ output: number;
4931
+ cacheRead: number;
4932
+ cacheWrite: number;
4933
+ };
4934
+ contextWindow: number;
4935
+ maxTokens: number;
4936
+ };
4795
4937
  };
4796
4938
  readonly "kimi-coding": {
4797
4939
  readonly k2p5: {
@@ -6472,6 +6614,23 @@ export declare const MODELS: {
6472
6614
  contextWindow: number;
6473
6615
  maxTokens: number;
6474
6616
  };
6617
+ readonly "claude-opus-4-7": {
6618
+ id: string;
6619
+ name: string;
6620
+ api: "anthropic-messages";
6621
+ provider: string;
6622
+ baseUrl: string;
6623
+ reasoning: true;
6624
+ input: ("image" | "text")[];
6625
+ cost: {
6626
+ input: number;
6627
+ output: number;
6628
+ cacheRead: number;
6629
+ cacheWrite: number;
6630
+ };
6631
+ contextWindow: number;
6632
+ maxTokens: number;
6633
+ };
6475
6634
  readonly "claude-sonnet-4": {
6476
6635
  id: string;
6477
6636
  name: string;
@@ -6574,6 +6733,23 @@ export declare const MODELS: {
6574
6733
  contextWindow: number;
6575
6734
  maxTokens: number;
6576
6735
  };
6736
+ readonly "glm-5.1": {
6737
+ id: string;
6738
+ name: string;
6739
+ api: "openai-completions";
6740
+ provider: string;
6741
+ baseUrl: string;
6742
+ reasoning: true;
6743
+ input: "text"[];
6744
+ cost: {
6745
+ input: number;
6746
+ output: number;
6747
+ cacheRead: number;
6748
+ cacheWrite: number;
6749
+ };
6750
+ contextWindow: number;
6751
+ maxTokens: number;
6752
+ };
6577
6753
  readonly "gpt-5": {
6578
6754
  id: string;
6579
6755
  name: string;
@@ -6880,14 +7056,31 @@ export declare const MODELS: {
6880
7056
  contextWindow: number;
6881
7057
  maxTokens: number;
6882
7058
  };
6883
- readonly "qwen3.6-plus-free": {
7059
+ readonly "qwen3.5-plus": {
6884
7060
  id: string;
6885
7061
  name: string;
6886
7062
  api: "openai-completions";
6887
7063
  provider: string;
6888
7064
  baseUrl: string;
6889
7065
  reasoning: true;
6890
- input: "text"[];
7066
+ input: ("image" | "text")[];
7067
+ cost: {
7068
+ input: number;
7069
+ output: number;
7070
+ cacheRead: number;
7071
+ cacheWrite: number;
7072
+ };
7073
+ contextWindow: number;
7074
+ maxTokens: number;
7075
+ };
7076
+ readonly "qwen3.6-plus": {
7077
+ id: string;
7078
+ name: string;
7079
+ api: "openai-completions";
7080
+ provider: string;
7081
+ baseUrl: string;
7082
+ reasoning: true;
7083
+ input: ("image" | "text")[];
6891
7084
  cost: {
6892
7085
  input: number;
6893
7086
  output: number;
@@ -6916,6 +7109,23 @@ export declare const MODELS: {
6916
7109
  contextWindow: number;
6917
7110
  maxTokens: number;
6918
7111
  };
7112
+ readonly "glm-5.1": {
7113
+ id: string;
7114
+ name: string;
7115
+ api: "openai-completions";
7116
+ provider: string;
7117
+ baseUrl: string;
7118
+ reasoning: true;
7119
+ input: "text"[];
7120
+ cost: {
7121
+ input: number;
7122
+ output: number;
7123
+ cacheRead: number;
7124
+ cacheWrite: number;
7125
+ };
7126
+ contextWindow: number;
7127
+ maxTokens: number;
7128
+ };
6919
7129
  readonly "kimi-k2.5": {
6920
7130
  id: string;
6921
7131
  name: string;
@@ -6970,7 +7180,7 @@ export declare const MODELS: {
6970
7180
  readonly "minimax-m2.5": {
6971
7181
  id: string;
6972
7182
  name: string;
6973
- api: "openai-completions";
7183
+ api: "anthropic-messages";
6974
7184
  provider: string;
6975
7185
  baseUrl: string;
6976
7186
  reasoning: true;
@@ -7001,6 +7211,40 @@ export declare const MODELS: {
7001
7211
  contextWindow: number;
7002
7212
  maxTokens: number;
7003
7213
  };
7214
+ readonly "qwen3.5-plus": {
7215
+ id: string;
7216
+ name: string;
7217
+ api: "openai-completions";
7218
+ provider: string;
7219
+ baseUrl: string;
7220
+ reasoning: true;
7221
+ input: ("image" | "text")[];
7222
+ cost: {
7223
+ input: number;
7224
+ output: number;
7225
+ cacheRead: number;
7226
+ cacheWrite: number;
7227
+ };
7228
+ contextWindow: number;
7229
+ maxTokens: number;
7230
+ };
7231
+ readonly "qwen3.6-plus": {
7232
+ id: string;
7233
+ name: string;
7234
+ api: "openai-completions";
7235
+ provider: string;
7236
+ baseUrl: string;
7237
+ reasoning: true;
7238
+ input: ("image" | "text")[];
7239
+ cost: {
7240
+ input: number;
7241
+ output: number;
7242
+ cacheRead: number;
7243
+ cacheWrite: number;
7244
+ };
7245
+ contextWindow: number;
7246
+ maxTokens: number;
7247
+ };
7004
7248
  };
7005
7249
  readonly openrouter: {
7006
7250
  readonly "ai21/jamba-large-1.7": {
@@ -7292,7 +7536,7 @@ export declare const MODELS: {
7292
7536
  contextWindow: number;
7293
7537
  maxTokens: number;
7294
7538
  };
7295
- readonly "anthropic/claude-sonnet-4": {
7539
+ readonly "anthropic/claude-opus-4.6-fast": {
7296
7540
  id: string;
7297
7541
  name: string;
7298
7542
  api: "openai-completions";
@@ -7309,7 +7553,7 @@ export declare const MODELS: {
7309
7553
  contextWindow: number;
7310
7554
  maxTokens: number;
7311
7555
  };
7312
- readonly "anthropic/claude-sonnet-4.5": {
7556
+ readonly "anthropic/claude-opus-4.7": {
7313
7557
  id: string;
7314
7558
  name: string;
7315
7559
  api: "openai-completions";
@@ -7326,7 +7570,7 @@ export declare const MODELS: {
7326
7570
  contextWindow: number;
7327
7571
  maxTokens: number;
7328
7572
  };
7329
- readonly "anthropic/claude-sonnet-4.6": {
7573
+ readonly "anthropic/claude-sonnet-4": {
7330
7574
  id: string;
7331
7575
  name: string;
7332
7576
  api: "openai-completions";
@@ -7343,14 +7587,14 @@ export declare const MODELS: {
7343
7587
  contextWindow: number;
7344
7588
  maxTokens: number;
7345
7589
  };
7346
- readonly "arcee-ai/trinity-large-preview:free": {
7590
+ readonly "anthropic/claude-sonnet-4.5": {
7347
7591
  id: string;
7348
7592
  name: string;
7349
7593
  api: "openai-completions";
7350
7594
  provider: string;
7351
7595
  baseUrl: string;
7352
- reasoning: false;
7353
- input: "text"[];
7596
+ reasoning: true;
7597
+ input: ("image" | "text")[];
7354
7598
  cost: {
7355
7599
  input: number;
7356
7600
  output: number;
@@ -7360,13 +7604,30 @@ export declare const MODELS: {
7360
7604
  contextWindow: number;
7361
7605
  maxTokens: number;
7362
7606
  };
7363
- readonly "arcee-ai/trinity-large-thinking": {
7607
+ readonly "anthropic/claude-sonnet-4.6": {
7364
7608
  id: string;
7365
7609
  name: string;
7366
7610
  api: "openai-completions";
7367
7611
  provider: string;
7368
7612
  baseUrl: string;
7369
7613
  reasoning: true;
7614
+ input: ("image" | "text")[];
7615
+ cost: {
7616
+ input: number;
7617
+ output: number;
7618
+ cacheRead: number;
7619
+ cacheWrite: number;
7620
+ };
7621
+ contextWindow: number;
7622
+ maxTokens: number;
7623
+ };
7624
+ readonly "arcee-ai/trinity-large-preview:free": {
7625
+ id: string;
7626
+ name: string;
7627
+ api: "openai-completions";
7628
+ provider: string;
7629
+ baseUrl: string;
7630
+ reasoning: false;
7370
7631
  input: "text"[];
7371
7632
  cost: {
7372
7633
  input: number;
@@ -7377,7 +7638,7 @@ export declare const MODELS: {
7377
7638
  contextWindow: number;
7378
7639
  maxTokens: number;
7379
7640
  };
7380
- readonly "arcee-ai/trinity-mini": {
7641
+ readonly "arcee-ai/trinity-large-thinking": {
7381
7642
  id: string;
7382
7643
  name: string;
7383
7644
  api: "openai-completions";
@@ -7394,7 +7655,7 @@ export declare const MODELS: {
7394
7655
  contextWindow: number;
7395
7656
  maxTokens: number;
7396
7657
  };
7397
- readonly "arcee-ai/trinity-mini:free": {
7658
+ readonly "arcee-ai/trinity-mini": {
7398
7659
  id: string;
7399
7660
  name: string;
7400
7661
  api: "openai-completions";
@@ -7955,7 +8216,7 @@ export declare const MODELS: {
7955
8216
  contextWindow: number;
7956
8217
  maxTokens: number;
7957
8218
  };
7958
- readonly "google/gemma-4-31b-it": {
8219
+ readonly "google/gemma-4-26b-a4b-it:free": {
7959
8220
  id: string;
7960
8221
  name: string;
7961
8222
  api: "openai-completions";
@@ -7972,14 +8233,14 @@ export declare const MODELS: {
7972
8233
  contextWindow: number;
7973
8234
  maxTokens: number;
7974
8235
  };
7975
- readonly "inception/mercury": {
8236
+ readonly "google/gemma-4-31b-it": {
7976
8237
  id: string;
7977
8238
  name: string;
7978
8239
  api: "openai-completions";
7979
8240
  provider: string;
7980
8241
  baseUrl: string;
7981
- reasoning: false;
7982
- input: "text"[];
8242
+ reasoning: true;
8243
+ input: ("image" | "text")[];
7983
8244
  cost: {
7984
8245
  input: number;
7985
8246
  output: number;
@@ -7989,14 +8250,14 @@ export declare const MODELS: {
7989
8250
  contextWindow: number;
7990
8251
  maxTokens: number;
7991
8252
  };
7992
- readonly "inception/mercury-2": {
8253
+ readonly "google/gemma-4-31b-it:free": {
7993
8254
  id: string;
7994
8255
  name: string;
7995
8256
  api: "openai-completions";
7996
8257
  provider: string;
7997
8258
  baseUrl: string;
7998
8259
  reasoning: true;
7999
- input: "text"[];
8260
+ input: ("image" | "text")[];
8000
8261
  cost: {
8001
8262
  input: number;
8002
8263
  output: number;
@@ -8006,13 +8267,13 @@ export declare const MODELS: {
8006
8267
  contextWindow: number;
8007
8268
  maxTokens: number;
8008
8269
  };
8009
- readonly "inception/mercury-coder": {
8270
+ readonly "inception/mercury-2": {
8010
8271
  id: string;
8011
8272
  name: string;
8012
8273
  api: "openai-completions";
8013
8274
  provider: string;
8014
8275
  baseUrl: string;
8015
- reasoning: false;
8276
+ reasoning: true;
8016
8277
  input: "text"[];
8017
8278
  cost: {
8018
8279
  input: number;
@@ -8040,23 +8301,6 @@ export declare const MODELS: {
8040
8301
  contextWindow: number;
8041
8302
  maxTokens: number;
8042
8303
  };
8043
- readonly "meituan/longcat-flash-chat": {
8044
- id: string;
8045
- name: string;
8046
- api: "openai-completions";
8047
- provider: string;
8048
- baseUrl: string;
8049
- reasoning: false;
8050
- input: "text"[];
8051
- cost: {
8052
- input: number;
8053
- output: number;
8054
- cacheRead: number;
8055
- cacheWrite: number;
8056
- };
8057
- contextWindow: number;
8058
- maxTokens: number;
8059
- };
8060
8304
  readonly "meta-llama/llama-3-8b-instruct": {
8061
8305
  id: string;
8062
8306
  name: string;
@@ -9876,6 +10120,23 @@ export declare const MODELS: {
9876
10120
  contextWindow: number;
9877
10121
  maxTokens: number;
9878
10122
  };
10123
+ readonly "openrouter/elephant-alpha": {
10124
+ id: string;
10125
+ name: string;
10126
+ api: "openai-completions";
10127
+ provider: string;
10128
+ baseUrl: string;
10129
+ reasoning: false;
10130
+ input: "text"[];
10131
+ cost: {
10132
+ input: number;
10133
+ output: number;
10134
+ cacheRead: number;
10135
+ cacheWrite: number;
10136
+ };
10137
+ contextWindow: number;
10138
+ maxTokens: number;
10139
+ };
9879
10140
  readonly "openrouter/free": {
9880
10141
  id: string;
9881
10142
  name: string;
@@ -10624,7 +10885,7 @@ export declare const MODELS: {
10624
10885
  contextWindow: number;
10625
10886
  maxTokens: number;
10626
10887
  };
10627
- readonly "qwen/qwen3.6-plus:free": {
10888
+ readonly "qwen/qwen3.6-plus": {
10628
10889
  id: string;
10629
10890
  name: string;
10630
10891
  api: "openai-completions";
@@ -10743,23 +11004,6 @@ export declare const MODELS: {
10743
11004
  contextWindow: number;
10744
11005
  maxTokens: number;
10745
11006
  };
10746
- readonly "stepfun/step-3.5-flash:free": {
10747
- id: string;
10748
- name: string;
10749
- api: "openai-completions";
10750
- provider: string;
10751
- baseUrl: string;
10752
- reasoning: true;
10753
- input: "text"[];
10754
- cost: {
10755
- input: number;
10756
- output: number;
10757
- cacheRead: number;
10758
- cacheWrite: number;
10759
- };
10760
- contextWindow: number;
10761
- maxTokens: number;
10762
- };
10763
11007
  readonly "thedrummer/rocinante-12b": {
10764
11008
  id: string;
10765
11009
  name: string;
@@ -11219,6 +11463,23 @@ export declare const MODELS: {
11219
11463
  contextWindow: number;
11220
11464
  maxTokens: number;
11221
11465
  };
11466
+ readonly "z-ai/glm-5.1": {
11467
+ id: string;
11468
+ name: string;
11469
+ api: "openai-completions";
11470
+ provider: string;
11471
+ baseUrl: string;
11472
+ reasoning: true;
11473
+ input: "text"[];
11474
+ cost: {
11475
+ input: number;
11476
+ output: number;
11477
+ cacheRead: number;
11478
+ cacheWrite: number;
11479
+ };
11480
+ contextWindow: number;
11481
+ maxTokens: number;
11482
+ };
11222
11483
  readonly "z-ai/glm-5v-turbo": {
11223
11484
  id: string;
11224
11485
  name: string;
@@ -11646,6 +11907,23 @@ export declare const MODELS: {
11646
11907
  contextWindow: number;
11647
11908
  maxTokens: number;
11648
11909
  };
11910
+ readonly "anthropic/claude-opus-4.7": {
11911
+ id: string;
11912
+ name: string;
11913
+ api: "anthropic-messages";
11914
+ provider: string;
11915
+ baseUrl: string;
11916
+ reasoning: true;
11917
+ input: ("image" | "text")[];
11918
+ cost: {
11919
+ input: number;
11920
+ output: number;
11921
+ cacheRead: number;
11922
+ cacheWrite: number;
11923
+ };
11924
+ contextWindow: number;
11925
+ maxTokens: number;
11926
+ };
11649
11927
  readonly "anthropic/claude-sonnet-4": {
11650
11928
  id: string;
11651
11929
  name: string;
@@ -13125,23 +13403,6 @@ export declare const MODELS: {
13125
13403
  contextWindow: number;
13126
13404
  maxTokens: number;
13127
13405
  };
13128
- readonly "openai/gpt-oss-120b": {
13129
- id: string;
13130
- name: string;
13131
- api: "anthropic-messages";
13132
- provider: string;
13133
- baseUrl: string;
13134
- reasoning: true;
13135
- input: "text"[];
13136
- cost: {
13137
- input: number;
13138
- output: number;
13139
- cacheRead: number;
13140
- cacheWrite: number;
13141
- };
13142
- contextWindow: number;
13143
- maxTokens: number;
13144
- };
13145
13406
  readonly "openai/gpt-oss-20b": {
13146
13407
  id: string;
13147
13408
  name: string;
@@ -13329,23 +13590,6 @@ export declare const MODELS: {
13329
13590
  contextWindow: number;
13330
13591
  maxTokens: number;
13331
13592
  };
13332
- readonly "xai/grok-2-vision": {
13333
- id: string;
13334
- name: string;
13335
- api: "anthropic-messages";
13336
- provider: string;
13337
- baseUrl: string;
13338
- reasoning: false;
13339
- input: ("image" | "text")[];
13340
- cost: {
13341
- input: number;
13342
- output: number;
13343
- cacheRead: number;
13344
- cacheWrite: number;
13345
- };
13346
- contextWindow: number;
13347
- maxTokens: number;
13348
- };
13349
13593
  readonly "xai/grok-3": {
13350
13594
  id: string;
13351
13595
  name: string;
@@ -13438,7 +13682,7 @@ export declare const MODELS: {
13438
13682
  provider: string;
13439
13683
  baseUrl: string;
13440
13684
  reasoning: false;
13441
- input: "text"[];
13685
+ input: ("image" | "text")[];
13442
13686
  cost: {
13443
13687
  input: number;
13444
13688
  output: number;
@@ -13455,7 +13699,7 @@ export declare const MODELS: {
13455
13699
  provider: string;
13456
13700
  baseUrl: string;
13457
13701
  reasoning: true;
13458
- input: "text"[];
13702
+ input: ("image" | "text")[];
13459
13703
  cost: {
13460
13704
  input: number;
13461
13705
  output: number;
@@ -13472,7 +13716,7 @@ export declare const MODELS: {
13472
13716
  provider: string;
13473
13717
  baseUrl: string;
13474
13718
  reasoning: false;
13475
- input: "text"[];
13719
+ input: ("image" | "text")[];
13476
13720
  cost: {
13477
13721
  input: number;
13478
13722
  output: number;
@@ -13489,7 +13733,7 @@ export declare const MODELS: {
13489
13733
  provider: string;
13490
13734
  baseUrl: string;
13491
13735
  reasoning: true;
13492
- input: "text"[];
13736
+ input: ("image" | "text")[];
13493
13737
  cost: {
13494
13738
  input: number;
13495
13739
  output: number;
@@ -13506,7 +13750,7 @@ export declare const MODELS: {
13506
13750
  provider: string;
13507
13751
  baseUrl: string;
13508
13752
  reasoning: true;
13509
- input: "text"[];
13753
+ input: ("image" | "text")[];
13510
13754
  cost: {
13511
13755
  input: number;
13512
13756
  output: number;
@@ -13523,7 +13767,7 @@ export declare const MODELS: {
13523
13767
  provider: string;
13524
13768
  baseUrl: string;
13525
13769
  reasoning: true;
13526
- input: "text"[];
13770
+ input: ("image" | "text")[];
13527
13771
  cost: {
13528
13772
  input: number;
13529
13773
  output: number;
@@ -13839,6 +14083,23 @@ export declare const MODELS: {
13839
14083
  contextWindow: number;
13840
14084
  maxTokens: number;
13841
14085
  };
14086
+ readonly "zai/glm-5.1": {
14087
+ id: string;
14088
+ name: string;
14089
+ api: "anthropic-messages";
14090
+ provider: string;
14091
+ baseUrl: string;
14092
+ reasoning: true;
14093
+ input: ("image" | "text")[];
14094
+ cost: {
14095
+ input: number;
14096
+ output: number;
14097
+ cacheRead: number;
14098
+ cacheWrite: number;
14099
+ };
14100
+ contextWindow: number;
14101
+ maxTokens: number;
14102
+ };
13842
14103
  readonly "zai/glm-5v-turbo": {
13843
14104
  id: string;
13844
14105
  name: string;
@@ -14499,6 +14760,27 @@ export declare const MODELS: {
14499
14760
  contextWindow: number;
14500
14761
  maxTokens: number;
14501
14762
  };
14763
+ readonly "glm-5.1": {
14764
+ id: string;
14765
+ name: string;
14766
+ api: "openai-completions";
14767
+ provider: string;
14768
+ baseUrl: string;
14769
+ compat: {
14770
+ supportsDeveloperRole: false;
14771
+ thinkingFormat: "zai";
14772
+ };
14773
+ reasoning: true;
14774
+ input: "text"[];
14775
+ cost: {
14776
+ input: number;
14777
+ output: number;
14778
+ cacheRead: number;
14779
+ cacheWrite: number;
14780
+ };
14781
+ contextWindow: number;
14782
+ maxTokens: number;
14783
+ };
14502
14784
  readonly "glm-5v-turbo": {
14503
14785
  id: string;
14504
14786
  name: string;