@mariozechner/pi-ai 0.64.0 → 0.65.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1549,23 +1549,6 @@ export declare const MODELS: {
1549
1549
  contextWindow: number;
1550
1550
  maxTokens: number;
1551
1551
  };
1552
- readonly "claude-3-7-sonnet-latest": {
1553
- id: string;
1554
- name: string;
1555
- api: "anthropic-messages";
1556
- provider: string;
1557
- baseUrl: string;
1558
- reasoning: true;
1559
- input: ("image" | "text")[];
1560
- cost: {
1561
- input: number;
1562
- output: number;
1563
- cacheRead: number;
1564
- cacheWrite: number;
1565
- };
1566
- contextWindow: number;
1567
- maxTokens: number;
1568
- };
1569
1552
  readonly "claude-3-haiku-20240307": {
1570
1553
  id: string;
1571
1554
  name: string;
@@ -2299,6 +2282,23 @@ export declare const MODELS: {
2299
2282
  contextWindow: number;
2300
2283
  maxTokens: number;
2301
2284
  };
2285
+ readonly "gpt-5.3-chat-latest": {
2286
+ id: string;
2287
+ name: string;
2288
+ api: "azure-openai-responses";
2289
+ provider: string;
2290
+ baseUrl: string;
2291
+ reasoning: false;
2292
+ input: ("image" | "text")[];
2293
+ cost: {
2294
+ input: number;
2295
+ output: number;
2296
+ cacheRead: number;
2297
+ cacheWrite: number;
2298
+ };
2299
+ contextWindow: number;
2300
+ maxTokens: number;
2301
+ };
2302
2302
  readonly "gpt-5.3-codex": {
2303
2303
  id: string;
2304
2304
  name: string;
@@ -3623,6 +3623,40 @@ export declare const MODELS: {
3623
3623
  contextWindow: number;
3624
3624
  maxTokens: number;
3625
3625
  };
3626
+ readonly "gemma-4-26b": {
3627
+ id: string;
3628
+ name: string;
3629
+ api: "google-generative-ai";
3630
+ provider: string;
3631
+ baseUrl: string;
3632
+ reasoning: true;
3633
+ input: ("image" | "text")[];
3634
+ cost: {
3635
+ input: number;
3636
+ output: number;
3637
+ cacheRead: number;
3638
+ cacheWrite: number;
3639
+ };
3640
+ contextWindow: number;
3641
+ maxTokens: number;
3642
+ };
3643
+ readonly "gemma-4-31b": {
3644
+ id: string;
3645
+ name: string;
3646
+ api: "google-generative-ai";
3647
+ provider: string;
3648
+ baseUrl: string;
3649
+ reasoning: true;
3650
+ input: ("image" | "text")[];
3651
+ cost: {
3652
+ input: number;
3653
+ output: number;
3654
+ cacheRead: number;
3655
+ cacheWrite: number;
3656
+ };
3657
+ contextWindow: number;
3658
+ maxTokens: number;
3659
+ };
3626
3660
  };
3627
3661
  readonly "google-antigravity": {
3628
3662
  readonly "claude-opus-4-5-thinking": {
@@ -5208,13 +5242,30 @@ export declare const MODELS: {
5208
5242
  contextWindow: number;
5209
5243
  maxTokens: number;
5210
5244
  };
5245
+ readonly "mistral-small-2603": {
5246
+ id: string;
5247
+ name: string;
5248
+ api: "mistral-conversations";
5249
+ provider: string;
5250
+ baseUrl: string;
5251
+ reasoning: true;
5252
+ input: ("image" | "text")[];
5253
+ cost: {
5254
+ input: number;
5255
+ output: number;
5256
+ cacheRead: number;
5257
+ cacheWrite: number;
5258
+ };
5259
+ contextWindow: number;
5260
+ maxTokens: number;
5261
+ };
5211
5262
  readonly "mistral-small-latest": {
5212
5263
  id: string;
5213
5264
  name: string;
5214
5265
  api: "mistral-conversations";
5215
5266
  provider: string;
5216
5267
  baseUrl: string;
5217
- reasoning: false;
5268
+ reasoning: true;
5218
5269
  input: ("image" | "text")[];
5219
5270
  cost: {
5220
5271
  input: number;
@@ -5754,6 +5805,23 @@ export declare const MODELS: {
5754
5805
  contextWindow: number;
5755
5806
  maxTokens: number;
5756
5807
  };
5808
+ readonly "gpt-5.3-chat-latest": {
5809
+ id: string;
5810
+ name: string;
5811
+ api: "openai-responses";
5812
+ provider: string;
5813
+ baseUrl: string;
5814
+ reasoning: false;
5815
+ input: ("image" | "text")[];
5816
+ cost: {
5817
+ input: number;
5818
+ output: number;
5819
+ cacheRead: number;
5820
+ cacheWrite: number;
5821
+ };
5822
+ contextWindow: number;
5823
+ maxTokens: number;
5824
+ };
5757
5825
  readonly "gpt-5.3-codex": {
5758
5826
  id: string;
5759
5827
  name: string;
@@ -6608,14 +6676,14 @@ export declare const MODELS: {
6608
6676
  contextWindow: number;
6609
6677
  maxTokens: number;
6610
6678
  };
6611
- readonly "mimo-v2-omni-free": {
6679
+ readonly "minimax-m2.5": {
6612
6680
  id: string;
6613
6681
  name: string;
6614
6682
  api: "openai-completions";
6615
6683
  provider: string;
6616
6684
  baseUrl: string;
6617
6685
  reasoning: true;
6618
- input: ("image" | "text")[];
6686
+ input: "text"[];
6619
6687
  cost: {
6620
6688
  input: number;
6621
6689
  output: number;
@@ -6625,10 +6693,10 @@ export declare const MODELS: {
6625
6693
  contextWindow: number;
6626
6694
  maxTokens: number;
6627
6695
  };
6628
- readonly "mimo-v2-pro-free": {
6696
+ readonly "minimax-m2.5-free": {
6629
6697
  id: string;
6630
6698
  name: string;
6631
- api: "openai-completions";
6699
+ api: "anthropic-messages";
6632
6700
  provider: string;
6633
6701
  baseUrl: string;
6634
6702
  reasoning: true;
@@ -6642,7 +6710,7 @@ export declare const MODELS: {
6642
6710
  contextWindow: number;
6643
6711
  maxTokens: number;
6644
6712
  };
6645
- readonly "minimax-m2.5": {
6713
+ readonly "nemotron-3-super-free": {
6646
6714
  id: string;
6647
6715
  name: string;
6648
6716
  api: "openai-completions";
@@ -6659,10 +6727,10 @@ export declare const MODELS: {
6659
6727
  contextWindow: number;
6660
6728
  maxTokens: number;
6661
6729
  };
6662
- readonly "minimax-m2.5-free": {
6730
+ readonly "qwen3.6-plus-free": {
6663
6731
  id: string;
6664
6732
  name: string;
6665
- api: "anthropic-messages";
6733
+ api: "openai-completions";
6666
6734
  provider: string;
6667
6735
  baseUrl: string;
6668
6736
  reasoning: true;
@@ -6676,7 +6744,9 @@ export declare const MODELS: {
6676
6744
  contextWindow: number;
6677
6745
  maxTokens: number;
6678
6746
  };
6679
- readonly "nemotron-3-super-free": {
6747
+ };
6748
+ readonly "opencode-go": {
6749
+ readonly "glm-5": {
6680
6750
  id: string;
6681
6751
  name: string;
6682
6752
  api: "openai-completions";
@@ -6693,16 +6763,14 @@ export declare const MODELS: {
6693
6763
  contextWindow: number;
6694
6764
  maxTokens: number;
6695
6765
  };
6696
- };
6697
- readonly "opencode-go": {
6698
- readonly "glm-5": {
6766
+ readonly "kimi-k2.5": {
6699
6767
  id: string;
6700
6768
  name: string;
6701
6769
  api: "openai-completions";
6702
6770
  provider: string;
6703
6771
  baseUrl: string;
6704
6772
  reasoning: true;
6705
- input: "text"[];
6773
+ input: ("image" | "text")[];
6706
6774
  cost: {
6707
6775
  input: number;
6708
6776
  output: number;
@@ -6712,7 +6780,7 @@ export declare const MODELS: {
6712
6780
  contextWindow: number;
6713
6781
  maxTokens: number;
6714
6782
  };
6715
- readonly "kimi-k2.5": {
6783
+ readonly "mimo-v2-omni": {
6716
6784
  id: string;
6717
6785
  name: string;
6718
6786
  api: "openai-completions";
@@ -6729,10 +6797,27 @@ export declare const MODELS: {
6729
6797
  contextWindow: number;
6730
6798
  maxTokens: number;
6731
6799
  };
6800
+ readonly "mimo-v2-pro": {
6801
+ id: string;
6802
+ name: string;
6803
+ api: "openai-completions";
6804
+ provider: string;
6805
+ baseUrl: string;
6806
+ reasoning: true;
6807
+ input: "text"[];
6808
+ cost: {
6809
+ input: number;
6810
+ output: number;
6811
+ cacheRead: number;
6812
+ cacheWrite: number;
6813
+ };
6814
+ contextWindow: number;
6815
+ maxTokens: number;
6816
+ };
6732
6817
  readonly "minimax-m2.5": {
6733
6818
  id: string;
6734
6819
  name: string;
6735
- api: "anthropic-messages";
6820
+ api: "openai-completions";
6736
6821
  provider: string;
6737
6822
  baseUrl: string;
6738
6823
  reasoning: true;
@@ -6935,13 +7020,13 @@ export declare const MODELS: {
6935
7020
  contextWindow: number;
6936
7021
  maxTokens: number;
6937
7022
  };
6938
- readonly "anthropic/claude-3.5-sonnet": {
7023
+ readonly "anthropic/claude-3.7-sonnet": {
6939
7024
  id: string;
6940
7025
  name: string;
6941
7026
  api: "openai-completions";
6942
7027
  provider: string;
6943
7028
  baseUrl: string;
6944
- reasoning: false;
7029
+ reasoning: true;
6945
7030
  input: ("image" | "text")[];
6946
7031
  cost: {
6947
7032
  input: number;
@@ -6952,7 +7037,7 @@ export declare const MODELS: {
6952
7037
  contextWindow: number;
6953
7038
  maxTokens: number;
6954
7039
  };
6955
- readonly "anthropic/claude-3.7-sonnet": {
7040
+ readonly "anthropic/claude-3.7-sonnet:thinking": {
6956
7041
  id: string;
6957
7042
  name: string;
6958
7043
  api: "openai-completions";
@@ -6969,7 +7054,7 @@ export declare const MODELS: {
6969
7054
  contextWindow: number;
6970
7055
  maxTokens: number;
6971
7056
  };
6972
- readonly "anthropic/claude-3.7-sonnet:thinking": {
7057
+ readonly "anthropic/claude-haiku-4.5": {
6973
7058
  id: string;
6974
7059
  name: string;
6975
7060
  api: "openai-completions";
@@ -6986,7 +7071,7 @@ export declare const MODELS: {
6986
7071
  contextWindow: number;
6987
7072
  maxTokens: number;
6988
7073
  };
6989
- readonly "anthropic/claude-haiku-4.5": {
7074
+ readonly "anthropic/claude-opus-4": {
6990
7075
  id: string;
6991
7076
  name: string;
6992
7077
  api: "openai-completions";
@@ -7003,7 +7088,7 @@ export declare const MODELS: {
7003
7088
  contextWindow: number;
7004
7089
  maxTokens: number;
7005
7090
  };
7006
- readonly "anthropic/claude-opus-4": {
7091
+ readonly "anthropic/claude-opus-4.1": {
7007
7092
  id: string;
7008
7093
  name: string;
7009
7094
  api: "openai-completions";
@@ -7020,7 +7105,7 @@ export declare const MODELS: {
7020
7105
  contextWindow: number;
7021
7106
  maxTokens: number;
7022
7107
  };
7023
- readonly "anthropic/claude-opus-4.1": {
7108
+ readonly "anthropic/claude-opus-4.5": {
7024
7109
  id: string;
7025
7110
  name: string;
7026
7111
  api: "openai-completions";
@@ -7037,7 +7122,7 @@ export declare const MODELS: {
7037
7122
  contextWindow: number;
7038
7123
  maxTokens: number;
7039
7124
  };
7040
- readonly "anthropic/claude-opus-4.5": {
7125
+ readonly "anthropic/claude-opus-4.6": {
7041
7126
  id: string;
7042
7127
  name: string;
7043
7128
  api: "openai-completions";
@@ -7054,7 +7139,7 @@ export declare const MODELS: {
7054
7139
  contextWindow: number;
7055
7140
  maxTokens: number;
7056
7141
  };
7057
- readonly "anthropic/claude-opus-4.6": {
7142
+ readonly "anthropic/claude-sonnet-4": {
7058
7143
  id: string;
7059
7144
  name: string;
7060
7145
  api: "openai-completions";
@@ -7071,7 +7156,7 @@ export declare const MODELS: {
7071
7156
  contextWindow: number;
7072
7157
  maxTokens: number;
7073
7158
  };
7074
- readonly "anthropic/claude-sonnet-4": {
7159
+ readonly "anthropic/claude-sonnet-4.5": {
7075
7160
  id: string;
7076
7161
  name: string;
7077
7162
  api: "openai-completions";
@@ -7088,7 +7173,7 @@ export declare const MODELS: {
7088
7173
  contextWindow: number;
7089
7174
  maxTokens: number;
7090
7175
  };
7091
- readonly "anthropic/claude-sonnet-4.5": {
7176
+ readonly "anthropic/claude-sonnet-4.6": {
7092
7177
  id: string;
7093
7178
  name: string;
7094
7179
  api: "openai-completions";
@@ -7105,14 +7190,14 @@ export declare const MODELS: {
7105
7190
  contextWindow: number;
7106
7191
  maxTokens: number;
7107
7192
  };
7108
- readonly "anthropic/claude-sonnet-4.6": {
7193
+ readonly "arcee-ai/trinity-large-preview:free": {
7109
7194
  id: string;
7110
7195
  name: string;
7111
7196
  api: "openai-completions";
7112
7197
  provider: string;
7113
7198
  baseUrl: string;
7114
- reasoning: true;
7115
- input: ("image" | "text")[];
7199
+ reasoning: false;
7200
+ input: "text"[];
7116
7201
  cost: {
7117
7202
  input: number;
7118
7203
  output: number;
@@ -7122,13 +7207,13 @@ export declare const MODELS: {
7122
7207
  contextWindow: number;
7123
7208
  maxTokens: number;
7124
7209
  };
7125
- readonly "arcee-ai/trinity-large-preview:free": {
7210
+ readonly "arcee-ai/trinity-large-thinking": {
7126
7211
  id: string;
7127
7212
  name: string;
7128
7213
  api: "openai-completions";
7129
7214
  provider: string;
7130
7215
  baseUrl: string;
7131
- reasoning: false;
7216
+ reasoning: true;
7132
7217
  input: "text"[];
7133
7218
  cost: {
7134
7219
  input: number;
@@ -7700,14 +7785,14 @@ export declare const MODELS: {
7700
7785
  contextWindow: number;
7701
7786
  maxTokens: number;
7702
7787
  };
7703
- readonly "inception/mercury": {
7788
+ readonly "google/gemma-4-26b-a4b-it": {
7704
7789
  id: string;
7705
7790
  name: string;
7706
7791
  api: "openai-completions";
7707
7792
  provider: string;
7708
7793
  baseUrl: string;
7709
- reasoning: false;
7710
- input: "text"[];
7794
+ reasoning: true;
7795
+ input: ("image" | "text")[];
7711
7796
  cost: {
7712
7797
  input: number;
7713
7798
  output: number;
@@ -7717,14 +7802,14 @@ export declare const MODELS: {
7717
7802
  contextWindow: number;
7718
7803
  maxTokens: number;
7719
7804
  };
7720
- readonly "inception/mercury-2": {
7805
+ readonly "google/gemma-4-31b-it": {
7721
7806
  id: string;
7722
7807
  name: string;
7723
7808
  api: "openai-completions";
7724
7809
  provider: string;
7725
7810
  baseUrl: string;
7726
7811
  reasoning: true;
7727
- input: "text"[];
7812
+ input: ("image" | "text")[];
7728
7813
  cost: {
7729
7814
  input: number;
7730
7815
  output: number;
@@ -7734,7 +7819,7 @@ export declare const MODELS: {
7734
7819
  contextWindow: number;
7735
7820
  maxTokens: number;
7736
7821
  };
7737
- readonly "inception/mercury-coder": {
7822
+ readonly "inception/mercury": {
7738
7823
  id: string;
7739
7824
  name: string;
7740
7825
  api: "openai-completions";
@@ -7751,7 +7836,24 @@ export declare const MODELS: {
7751
7836
  contextWindow: number;
7752
7837
  maxTokens: number;
7753
7838
  };
7754
- readonly "kwaipilot/kat-coder-pro": {
7839
+ readonly "inception/mercury-2": {
7840
+ id: string;
7841
+ name: string;
7842
+ api: "openai-completions";
7843
+ provider: string;
7844
+ baseUrl: string;
7845
+ reasoning: true;
7846
+ input: "text"[];
7847
+ cost: {
7848
+ input: number;
7849
+ output: number;
7850
+ cacheRead: number;
7851
+ cacheWrite: number;
7852
+ };
7853
+ contextWindow: number;
7854
+ maxTokens: number;
7855
+ };
7856
+ readonly "inception/mercury-coder": {
7755
7857
  id: string;
7756
7858
  name: string;
7757
7859
  api: "openai-completions";
@@ -8278,23 +8380,6 @@ export declare const MODELS: {
8278
8380
  contextWindow: number;
8279
8381
  maxTokens: number;
8280
8382
  };
8281
- readonly "mistralai/mistral-small-24b-instruct-2501": {
8282
- id: string;
8283
- name: string;
8284
- api: "openai-completions";
8285
- provider: string;
8286
- baseUrl: string;
8287
- reasoning: false;
8288
- input: "text"[];
8289
- cost: {
8290
- input: number;
8291
- output: number;
8292
- cacheRead: number;
8293
- cacheWrite: number;
8294
- };
8295
- contextWindow: number;
8296
- maxTokens: number;
8297
- };
8298
8383
  readonly "mistralai/mistral-small-2603": {
8299
8384
  id: string;
8300
8385
  name: string;
@@ -9349,6 +9434,40 @@ export declare const MODELS: {
9349
9434
  contextWindow: number;
9350
9435
  maxTokens: number;
9351
9436
  };
9437
+ readonly "openai/gpt-audio": {
9438
+ id: string;
9439
+ name: string;
9440
+ api: "openai-completions";
9441
+ provider: string;
9442
+ baseUrl: string;
9443
+ reasoning: false;
9444
+ input: "text"[];
9445
+ cost: {
9446
+ input: number;
9447
+ output: number;
9448
+ cacheRead: number;
9449
+ cacheWrite: number;
9450
+ };
9451
+ contextWindow: number;
9452
+ maxTokens: number;
9453
+ };
9454
+ readonly "openai/gpt-audio-mini": {
9455
+ id: string;
9456
+ name: string;
9457
+ api: "openai-completions";
9458
+ provider: string;
9459
+ baseUrl: string;
9460
+ reasoning: false;
9461
+ input: "text"[];
9462
+ cost: {
9463
+ input: number;
9464
+ output: number;
9465
+ cacheRead: number;
9466
+ cacheWrite: number;
9467
+ };
9468
+ contextWindow: number;
9469
+ maxTokens: number;
9470
+ };
9352
9471
  readonly "openai/gpt-oss-120b": {
9353
9472
  id: string;
9354
9473
  name: string;
@@ -10352,6 +10471,23 @@ export declare const MODELS: {
10352
10471
  contextWindow: number;
10353
10472
  maxTokens: number;
10354
10473
  };
10474
+ readonly "qwen/qwen3.6-plus:free": {
10475
+ id: string;
10476
+ name: string;
10477
+ api: "openai-completions";
10478
+ provider: string;
10479
+ baseUrl: string;
10480
+ reasoning: true;
10481
+ input: ("image" | "text")[];
10482
+ cost: {
10483
+ input: number;
10484
+ output: number;
10485
+ cacheRead: number;
10486
+ cacheWrite: number;
10487
+ };
10488
+ contextWindow: number;
10489
+ maxTokens: number;
10490
+ };
10355
10491
  readonly "qwen/qwq-32b": {
10356
10492
  id: string;
10357
10493
  name: string;
@@ -10369,7 +10505,7 @@ export declare const MODELS: {
10369
10505
  contextWindow: number;
10370
10506
  maxTokens: number;
10371
10507
  };
10372
- readonly "reka/reka-edge": {
10508
+ readonly "rekaai/reka-edge": {
10373
10509
  id: string;
10374
10510
  name: string;
10375
10511
  api: "openai-completions";
@@ -10658,7 +10794,7 @@ export declare const MODELS: {
10658
10794
  contextWindow: number;
10659
10795
  maxTokens: number;
10660
10796
  };
10661
- readonly "x-ai/grok-4.20-beta": {
10797
+ readonly "x-ai/grok-4.20": {
10662
10798
  id: string;
10663
10799
  name: string;
10664
10800
  api: "openai-completions";
@@ -10930,6 +11066,23 @@ export declare const MODELS: {
10930
11066
  contextWindow: number;
10931
11067
  maxTokens: number;
10932
11068
  };
11069
+ readonly "z-ai/glm-5v-turbo": {
11070
+ id: string;
11071
+ name: string;
11072
+ api: "openai-completions";
11073
+ provider: string;
11074
+ baseUrl: string;
11075
+ reasoning: true;
11076
+ input: ("image" | "text")[];
11077
+ cost: {
11078
+ input: number;
11079
+ output: number;
11080
+ cacheRead: number;
11081
+ cacheWrite: number;
11082
+ };
11083
+ contextWindow: number;
11084
+ maxTokens: number;
11085
+ };
10933
11086
  };
10934
11087
  readonly "vercel-ai-gateway": {
10935
11088
  readonly "alibaba/qwen-3-14b": {
@@ -11187,13 +11340,13 @@ export declare const MODELS: {
11187
11340
  contextWindow: number;
11188
11341
  maxTokens: number;
11189
11342
  };
11190
- readonly "anthropic/claude-3-haiku": {
11343
+ readonly "alibaba/qwen3.6-plus": {
11191
11344
  id: string;
11192
11345
  name: string;
11193
11346
  api: "anthropic-messages";
11194
11347
  provider: string;
11195
11348
  baseUrl: string;
11196
- reasoning: false;
11349
+ reasoning: true;
11197
11350
  input: ("image" | "text")[];
11198
11351
  cost: {
11199
11352
  input: number;
@@ -11204,7 +11357,7 @@ export declare const MODELS: {
11204
11357
  contextWindow: number;
11205
11358
  maxTokens: number;
11206
11359
  };
11207
- readonly "anthropic/claude-3.5-haiku": {
11360
+ readonly "anthropic/claude-3-haiku": {
11208
11361
  id: string;
11209
11362
  name: string;
11210
11363
  api: "anthropic-messages";
@@ -11221,7 +11374,7 @@ export declare const MODELS: {
11221
11374
  contextWindow: number;
11222
11375
  maxTokens: number;
11223
11376
  };
11224
- readonly "anthropic/claude-3.5-sonnet": {
11377
+ readonly "anthropic/claude-3.5-haiku": {
11225
11378
  id: string;
11226
11379
  name: string;
11227
11380
  api: "anthropic-messages";
@@ -11238,13 +11391,13 @@ export declare const MODELS: {
11238
11391
  contextWindow: number;
11239
11392
  maxTokens: number;
11240
11393
  };
11241
- readonly "anthropic/claude-3.5-sonnet-20240620": {
11394
+ readonly "anthropic/claude-3.7-sonnet": {
11242
11395
  id: string;
11243
11396
  name: string;
11244
11397
  api: "anthropic-messages";
11245
11398
  provider: string;
11246
11399
  baseUrl: string;
11247
- reasoning: false;
11400
+ reasoning: true;
11248
11401
  input: ("image" | "text")[];
11249
11402
  cost: {
11250
11403
  input: number;
@@ -11255,7 +11408,7 @@ export declare const MODELS: {
11255
11408
  contextWindow: number;
11256
11409
  maxTokens: number;
11257
11410
  };
11258
- readonly "anthropic/claude-3.7-sonnet": {
11411
+ readonly "anthropic/claude-haiku-4.5": {
11259
11412
  id: string;
11260
11413
  name: string;
11261
11414
  api: "anthropic-messages";
@@ -11272,7 +11425,7 @@ export declare const MODELS: {
11272
11425
  contextWindow: number;
11273
11426
  maxTokens: number;
11274
11427
  };
11275
- readonly "anthropic/claude-haiku-4.5": {
11428
+ readonly "anthropic/claude-opus-4": {
11276
11429
  id: string;
11277
11430
  name: string;
11278
11431
  api: "anthropic-messages";
@@ -11289,7 +11442,7 @@ export declare const MODELS: {
11289
11442
  contextWindow: number;
11290
11443
  maxTokens: number;
11291
11444
  };
11292
- readonly "anthropic/claude-opus-4": {
11445
+ readonly "anthropic/claude-opus-4.1": {
11293
11446
  id: string;
11294
11447
  name: string;
11295
11448
  api: "anthropic-messages";
@@ -11306,7 +11459,7 @@ export declare const MODELS: {
11306
11459
  contextWindow: number;
11307
11460
  maxTokens: number;
11308
11461
  };
11309
- readonly "anthropic/claude-opus-4.1": {
11462
+ readonly "anthropic/claude-opus-4.5": {
11310
11463
  id: string;
11311
11464
  name: string;
11312
11465
  api: "anthropic-messages";
@@ -11323,7 +11476,7 @@ export declare const MODELS: {
11323
11476
  contextWindow: number;
11324
11477
  maxTokens: number;
11325
11478
  };
11326
- readonly "anthropic/claude-opus-4.5": {
11479
+ readonly "anthropic/claude-opus-4.6": {
11327
11480
  id: string;
11328
11481
  name: string;
11329
11482
  api: "anthropic-messages";
@@ -11340,7 +11493,7 @@ export declare const MODELS: {
11340
11493
  contextWindow: number;
11341
11494
  maxTokens: number;
11342
11495
  };
11343
- readonly "anthropic/claude-opus-4.6": {
11496
+ readonly "anthropic/claude-sonnet-4": {
11344
11497
  id: string;
11345
11498
  name: string;
11346
11499
  api: "anthropic-messages";
@@ -11357,7 +11510,7 @@ export declare const MODELS: {
11357
11510
  contextWindow: number;
11358
11511
  maxTokens: number;
11359
11512
  };
11360
- readonly "anthropic/claude-sonnet-4": {
11513
+ readonly "anthropic/claude-sonnet-4.5": {
11361
11514
  id: string;
11362
11515
  name: string;
11363
11516
  api: "anthropic-messages";
@@ -11374,7 +11527,7 @@ export declare const MODELS: {
11374
11527
  contextWindow: number;
11375
11528
  maxTokens: number;
11376
11529
  };
11377
- readonly "anthropic/claude-sonnet-4.5": {
11530
+ readonly "anthropic/claude-sonnet-4.6": {
11378
11531
  id: string;
11379
11532
  name: string;
11380
11533
  api: "anthropic-messages";
@@ -11391,14 +11544,14 @@ export declare const MODELS: {
11391
11544
  contextWindow: number;
11392
11545
  maxTokens: number;
11393
11546
  };
11394
- readonly "anthropic/claude-sonnet-4.6": {
11547
+ readonly "arcee-ai/trinity-large-preview": {
11395
11548
  id: string;
11396
11549
  name: string;
11397
11550
  api: "anthropic-messages";
11398
11551
  provider: string;
11399
11552
  baseUrl: string;
11400
- reasoning: true;
11401
- input: ("image" | "text")[];
11553
+ reasoning: false;
11554
+ input: "text"[];
11402
11555
  cost: {
11403
11556
  input: number;
11404
11557
  output: number;
@@ -11408,13 +11561,13 @@ export declare const MODELS: {
11408
11561
  contextWindow: number;
11409
11562
  maxTokens: number;
11410
11563
  };
11411
- readonly "arcee-ai/trinity-large-preview": {
11564
+ readonly "arcee-ai/trinity-large-thinking": {
11412
11565
  id: string;
11413
11566
  name: string;
11414
11567
  api: "anthropic-messages";
11415
11568
  provider: string;
11416
11569
  baseUrl: string;
11417
- reasoning: false;
11570
+ reasoning: true;
11418
11571
  input: "text"[];
11419
11572
  cost: {
11420
11573
  input: number;
@@ -11714,14 +11867,14 @@ export declare const MODELS: {
11714
11867
  contextWindow: number;
11715
11868
  maxTokens: number;
11716
11869
  };
11717
- readonly "inception/mercury-2": {
11870
+ readonly "google/gemma-4-26b-a4b-it": {
11718
11871
  id: string;
11719
11872
  name: string;
11720
11873
  api: "anthropic-messages";
11721
11874
  provider: string;
11722
11875
  baseUrl: string;
11723
- reasoning: true;
11724
- input: "text"[];
11876
+ reasoning: false;
11877
+ input: ("image" | "text")[];
11725
11878
  cost: {
11726
11879
  input: number;
11727
11880
  output: number;
@@ -11731,14 +11884,14 @@ export declare const MODELS: {
11731
11884
  contextWindow: number;
11732
11885
  maxTokens: number;
11733
11886
  };
11734
- readonly "inception/mercury-coder-small": {
11887
+ readonly "google/gemma-4-31b-it": {
11735
11888
  id: string;
11736
11889
  name: string;
11737
11890
  api: "anthropic-messages";
11738
11891
  provider: string;
11739
11892
  baseUrl: string;
11740
11893
  reasoning: false;
11741
- input: "text"[];
11894
+ input: ("image" | "text")[];
11742
11895
  cost: {
11743
11896
  input: number;
11744
11897
  output: number;
@@ -11748,7 +11901,7 @@ export declare const MODELS: {
11748
11901
  contextWindow: number;
11749
11902
  maxTokens: number;
11750
11903
  };
11751
- readonly "kwaipilot/kat-coder-pro-v2": {
11904
+ readonly "inception/mercury-2": {
11752
11905
  id: string;
11753
11906
  name: string;
11754
11907
  api: "anthropic-messages";
@@ -11765,7 +11918,7 @@ export declare const MODELS: {
11765
11918
  contextWindow: number;
11766
11919
  maxTokens: number;
11767
11920
  };
11768
- readonly "meituan/longcat-flash-chat": {
11921
+ readonly "inception/mercury-coder-small": {
11769
11922
  id: string;
11770
11923
  name: string;
11771
11924
  api: "anthropic-messages";
@@ -11782,7 +11935,7 @@ export declare const MODELS: {
11782
11935
  contextWindow: number;
11783
11936
  maxTokens: number;
11784
11937
  };
11785
- readonly "meituan/longcat-flash-thinking": {
11938
+ readonly "kwaipilot/kat-coder-pro-v2": {
11786
11939
  id: string;
11787
11940
  name: string;
11788
11941
  api: "anthropic-messages";
@@ -11799,6 +11952,23 @@ export declare const MODELS: {
11799
11952
  contextWindow: number;
11800
11953
  maxTokens: number;
11801
11954
  };
11955
+ readonly "meituan/longcat-flash-chat": {
11956
+ id: string;
11957
+ name: string;
11958
+ api: "anthropic-messages";
11959
+ provider: string;
11960
+ baseUrl: string;
11961
+ reasoning: false;
11962
+ input: "text"[];
11963
+ cost: {
11964
+ input: number;
11965
+ output: number;
11966
+ cacheRead: number;
11967
+ cacheWrite: number;
11968
+ };
11969
+ contextWindow: number;
11970
+ maxTokens: number;
11971
+ };
11802
11972
  readonly "meta/llama-3.1-70b": {
11803
11973
  id: string;
11804
11974
  name: string;
@@ -12802,6 +12972,23 @@ export declare const MODELS: {
12802
12972
  contextWindow: number;
12803
12973
  maxTokens: number;
12804
12974
  };
12975
+ readonly "openai/gpt-oss-120b": {
12976
+ id: string;
12977
+ name: string;
12978
+ api: "anthropic-messages";
12979
+ provider: string;
12980
+ baseUrl: string;
12981
+ reasoning: true;
12982
+ input: "text"[];
12983
+ cost: {
12984
+ input: number;
12985
+ output: number;
12986
+ cacheRead: number;
12987
+ cacheWrite: number;
12988
+ };
12989
+ contextWindow: number;
12990
+ maxTokens: number;
12991
+ };
12805
12992
  readonly "openai/gpt-oss-20b": {
12806
12993
  id: string;
12807
12994
  name: string;
@@ -13465,6 +13652,23 @@ export declare const MODELS: {
13465
13652
  contextWindow: number;
13466
13653
  maxTokens: number;
13467
13654
  };
13655
+ readonly "zai/glm-5": {
13656
+ id: string;
13657
+ name: string;
13658
+ api: "anthropic-messages";
13659
+ provider: string;
13660
+ baseUrl: string;
13661
+ reasoning: true;
13662
+ input: "text"[];
13663
+ cost: {
13664
+ input: number;
13665
+ output: number;
13666
+ cacheRead: number;
13667
+ cacheWrite: number;
13668
+ };
13669
+ contextWindow: number;
13670
+ maxTokens: number;
13671
+ };
13468
13672
  readonly "zai/glm-5-turbo": {
13469
13673
  id: string;
13470
13674
  name: string;
@@ -13482,6 +13686,23 @@ export declare const MODELS: {
13482
13686
  contextWindow: number;
13483
13687
  maxTokens: number;
13484
13688
  };
13689
+ readonly "zai/glm-5v-turbo": {
13690
+ id: string;
13691
+ name: string;
13692
+ api: "anthropic-messages";
13693
+ provider: string;
13694
+ baseUrl: string;
13695
+ reasoning: true;
13696
+ input: ("image" | "text")[];
13697
+ cost: {
13698
+ input: number;
13699
+ output: number;
13700
+ cacheRead: number;
13701
+ cacheWrite: number;
13702
+ };
13703
+ contextWindow: number;
13704
+ maxTokens: number;
13705
+ };
13485
13706
  };
13486
13707
  readonly xai: {
13487
13708
  readonly "grok-2": {
@@ -13987,6 +14208,7 @@ export declare const MODELS: {
13987
14208
  compat: {
13988
14209
  supportsDeveloperRole: false;
13989
14210
  thinkingFormat: "zai";
14211
+ zaiToolStream: true;
13990
14212
  };
13991
14213
  reasoning: true;
13992
14214
  input: "text"[];
@@ -14008,6 +14230,7 @@ export declare const MODELS: {
14008
14230
  compat: {
14009
14231
  supportsDeveloperRole: false;
14010
14232
  thinkingFormat: "zai";
14233
+ zaiToolStream: true;
14011
14234
  };
14012
14235
  reasoning: true;
14013
14236
  input: ("image" | "text")[];
@@ -14029,6 +14252,7 @@ export declare const MODELS: {
14029
14252
  compat: {
14030
14253
  supportsDeveloperRole: false;
14031
14254
  thinkingFormat: "zai";
14255
+ zaiToolStream: true;
14032
14256
  };
14033
14257
  reasoning: true;
14034
14258
  input: "text"[];
@@ -14050,6 +14274,7 @@ export declare const MODELS: {
14050
14274
  compat: {
14051
14275
  supportsDeveloperRole: false;
14052
14276
  thinkingFormat: "zai";
14277
+ zaiToolStream: true;
14053
14278
  };
14054
14279
  reasoning: true;
14055
14280
  input: "text"[];
@@ -14071,6 +14296,7 @@ export declare const MODELS: {
14071
14296
  compat: {
14072
14297
  supportsDeveloperRole: false;
14073
14298
  thinkingFormat: "zai";
14299
+ zaiToolStream: true;
14074
14300
  };
14075
14301
  reasoning: true;
14076
14302
  input: "text"[];
@@ -14092,6 +14318,7 @@ export declare const MODELS: {
14092
14318
  compat: {
14093
14319
  supportsDeveloperRole: false;
14094
14320
  thinkingFormat: "zai";
14321
+ zaiToolStream: true;
14095
14322
  };
14096
14323
  reasoning: true;
14097
14324
  input: "text"[];
@@ -14113,6 +14340,7 @@ export declare const MODELS: {
14113
14340
  compat: {
14114
14341
  supportsDeveloperRole: false;
14115
14342
  thinkingFormat: "zai";
14343
+ zaiToolStream: true;
14116
14344
  };
14117
14345
  reasoning: true;
14118
14346
  input: "text"[];
@@ -14125,6 +14353,50 @@ export declare const MODELS: {
14125
14353
  contextWindow: number;
14126
14354
  maxTokens: number;
14127
14355
  };
14356
+ readonly "glm-5.1": {
14357
+ id: string;
14358
+ name: string;
14359
+ api: "openai-completions";
14360
+ provider: string;
14361
+ baseUrl: string;
14362
+ compat: {
14363
+ supportsDeveloperRole: false;
14364
+ thinkingFormat: "zai";
14365
+ zaiToolStream: true;
14366
+ };
14367
+ reasoning: true;
14368
+ input: "text"[];
14369
+ cost: {
14370
+ input: number;
14371
+ output: number;
14372
+ cacheRead: number;
14373
+ cacheWrite: number;
14374
+ };
14375
+ contextWindow: number;
14376
+ maxTokens: number;
14377
+ };
14378
+ readonly "glm-5v-turbo": {
14379
+ id: string;
14380
+ name: string;
14381
+ api: "openai-completions";
14382
+ provider: string;
14383
+ baseUrl: string;
14384
+ compat: {
14385
+ supportsDeveloperRole: false;
14386
+ thinkingFormat: "zai";
14387
+ zaiToolStream: true;
14388
+ };
14389
+ reasoning: true;
14390
+ input: ("image" | "text")[];
14391
+ cost: {
14392
+ input: number;
14393
+ output: number;
14394
+ cacheRead: number;
14395
+ cacheWrite: number;
14396
+ };
14397
+ contextWindow: number;
14398
+ maxTokens: number;
14399
+ };
14128
14400
  };
14129
14401
  };
14130
14402
  //# sourceMappingURL=models.generated.d.ts.map