@hyperspaceng/neural-ai 0.64.1 → 0.65.3

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1549,23 +1549,6 @@ export declare const MODELS: {
1549
1549
  contextWindow: number;
1550
1550
  maxTokens: number;
1551
1551
  };
1552
- readonly "claude-3-7-sonnet-latest": {
1553
- id: string;
1554
- name: string;
1555
- api: "anthropic-messages";
1556
- provider: string;
1557
- baseUrl: string;
1558
- reasoning: true;
1559
- input: ("image" | "text")[];
1560
- cost: {
1561
- input: number;
1562
- output: number;
1563
- cacheRead: number;
1564
- cacheWrite: number;
1565
- };
1566
- contextWindow: number;
1567
- maxTokens: number;
1568
- };
1569
1552
  readonly "claude-3-haiku-20240307": {
1570
1553
  id: string;
1571
1554
  name: string;
@@ -3640,6 +3623,40 @@ export declare const MODELS: {
3640
3623
  contextWindow: number;
3641
3624
  maxTokens: number;
3642
3625
  };
3626
+ readonly "gemma-4-26b-it": {
3627
+ id: string;
3628
+ name: string;
3629
+ api: "google-generative-ai";
3630
+ provider: string;
3631
+ baseUrl: string;
3632
+ reasoning: true;
3633
+ input: ("image" | "text")[];
3634
+ cost: {
3635
+ input: number;
3636
+ output: number;
3637
+ cacheRead: number;
3638
+ cacheWrite: number;
3639
+ };
3640
+ contextWindow: number;
3641
+ maxTokens: number;
3642
+ };
3643
+ readonly "gemma-4-31b-it": {
3644
+ id: string;
3645
+ name: string;
3646
+ api: "google-generative-ai";
3647
+ provider: string;
3648
+ baseUrl: string;
3649
+ reasoning: true;
3650
+ input: ("image" | "text")[];
3651
+ cost: {
3652
+ input: number;
3653
+ output: number;
3654
+ cacheRead: number;
3655
+ cacheWrite: number;
3656
+ };
3657
+ contextWindow: number;
3658
+ maxTokens: number;
3659
+ };
3643
3660
  };
3644
3661
  readonly "google-antigravity": {
3645
3662
  readonly "claude-opus-4-5-thinking": {
@@ -6659,14 +6676,14 @@ export declare const MODELS: {
6659
6676
  contextWindow: number;
6660
6677
  maxTokens: number;
6661
6678
  };
6662
- readonly "mimo-v2-omni-free": {
6679
+ readonly "minimax-m2.5": {
6663
6680
  id: string;
6664
6681
  name: string;
6665
6682
  api: "openai-completions";
6666
6683
  provider: string;
6667
6684
  baseUrl: string;
6668
6685
  reasoning: true;
6669
- input: ("image" | "text")[];
6686
+ input: "text"[];
6670
6687
  cost: {
6671
6688
  input: number;
6672
6689
  output: number;
@@ -6676,10 +6693,10 @@ export declare const MODELS: {
6676
6693
  contextWindow: number;
6677
6694
  maxTokens: number;
6678
6695
  };
6679
- readonly "mimo-v2-pro-free": {
6696
+ readonly "minimax-m2.5-free": {
6680
6697
  id: string;
6681
6698
  name: string;
6682
- api: "openai-completions";
6699
+ api: "anthropic-messages";
6683
6700
  provider: string;
6684
6701
  baseUrl: string;
6685
6702
  reasoning: true;
@@ -6693,7 +6710,7 @@ export declare const MODELS: {
6693
6710
  contextWindow: number;
6694
6711
  maxTokens: number;
6695
6712
  };
6696
- readonly "minimax-m2.5": {
6713
+ readonly "nemotron-3-super-free": {
6697
6714
  id: string;
6698
6715
  name: string;
6699
6716
  api: "openai-completions";
@@ -6710,10 +6727,10 @@ export declare const MODELS: {
6710
6727
  contextWindow: number;
6711
6728
  maxTokens: number;
6712
6729
  };
6713
- readonly "minimax-m2.5-free": {
6730
+ readonly "qwen3.6-plus-free": {
6714
6731
  id: string;
6715
6732
  name: string;
6716
- api: "anthropic-messages";
6733
+ api: "openai-completions";
6717
6734
  provider: string;
6718
6735
  baseUrl: string;
6719
6736
  reasoning: true;
@@ -6727,7 +6744,9 @@ export declare const MODELS: {
6727
6744
  contextWindow: number;
6728
6745
  maxTokens: number;
6729
6746
  };
6730
- readonly "nemotron-3-super-free": {
6747
+ };
6748
+ readonly "opencode-go": {
6749
+ readonly "glm-5": {
6731
6750
  id: string;
6732
6751
  name: string;
6733
6752
  api: "openai-completions";
@@ -6744,14 +6763,14 @@ export declare const MODELS: {
6744
6763
  contextWindow: number;
6745
6764
  maxTokens: number;
6746
6765
  };
6747
- readonly "qwen3.6-plus-free": {
6766
+ readonly "kimi-k2.5": {
6748
6767
  id: string;
6749
6768
  name: string;
6750
6769
  api: "openai-completions";
6751
6770
  provider: string;
6752
6771
  baseUrl: string;
6753
6772
  reasoning: true;
6754
- input: "text"[];
6773
+ input: ("image" | "text")[];
6755
6774
  cost: {
6756
6775
  input: number;
6757
6776
  output: number;
@@ -6761,16 +6780,14 @@ export declare const MODELS: {
6761
6780
  contextWindow: number;
6762
6781
  maxTokens: number;
6763
6782
  };
6764
- };
6765
- readonly "opencode-go": {
6766
- readonly "glm-5": {
6783
+ readonly "mimo-v2-omni": {
6767
6784
  id: string;
6768
6785
  name: string;
6769
6786
  api: "openai-completions";
6770
6787
  provider: string;
6771
6788
  baseUrl: string;
6772
6789
  reasoning: true;
6773
- input: "text"[];
6790
+ input: ("image" | "text")[];
6774
6791
  cost: {
6775
6792
  input: number;
6776
6793
  output: number;
@@ -6780,14 +6797,14 @@ export declare const MODELS: {
6780
6797
  contextWindow: number;
6781
6798
  maxTokens: number;
6782
6799
  };
6783
- readonly "kimi-k2.5": {
6800
+ readonly "mimo-v2-pro": {
6784
6801
  id: string;
6785
6802
  name: string;
6786
6803
  api: "openai-completions";
6787
6804
  provider: string;
6788
6805
  baseUrl: string;
6789
6806
  reasoning: true;
6790
- input: ("image" | "text")[];
6807
+ input: "text"[];
6791
6808
  cost: {
6792
6809
  input: number;
6793
6810
  output: number;
@@ -6800,7 +6817,7 @@ export declare const MODELS: {
6800
6817
  readonly "minimax-m2.5": {
6801
6818
  id: string;
6802
6819
  name: string;
6803
- api: "anthropic-messages";
6820
+ api: "openai-completions";
6804
6821
  provider: string;
6805
6822
  baseUrl: string;
6806
6823
  reasoning: true;
@@ -7003,13 +7020,13 @@ export declare const MODELS: {
7003
7020
  contextWindow: number;
7004
7021
  maxTokens: number;
7005
7022
  };
7006
- readonly "anthropic/claude-3.5-sonnet": {
7023
+ readonly "anthropic/claude-3.7-sonnet": {
7007
7024
  id: string;
7008
7025
  name: string;
7009
7026
  api: "openai-completions";
7010
7027
  provider: string;
7011
7028
  baseUrl: string;
7012
- reasoning: false;
7029
+ reasoning: true;
7013
7030
  input: ("image" | "text")[];
7014
7031
  cost: {
7015
7032
  input: number;
@@ -7020,7 +7037,7 @@ export declare const MODELS: {
7020
7037
  contextWindow: number;
7021
7038
  maxTokens: number;
7022
7039
  };
7023
- readonly "anthropic/claude-3.7-sonnet": {
7040
+ readonly "anthropic/claude-3.7-sonnet:thinking": {
7024
7041
  id: string;
7025
7042
  name: string;
7026
7043
  api: "openai-completions";
@@ -7037,7 +7054,7 @@ export declare const MODELS: {
7037
7054
  contextWindow: number;
7038
7055
  maxTokens: number;
7039
7056
  };
7040
- readonly "anthropic/claude-3.7-sonnet:thinking": {
7057
+ readonly "anthropic/claude-haiku-4.5": {
7041
7058
  id: string;
7042
7059
  name: string;
7043
7060
  api: "openai-completions";
@@ -7054,7 +7071,7 @@ export declare const MODELS: {
7054
7071
  contextWindow: number;
7055
7072
  maxTokens: number;
7056
7073
  };
7057
- readonly "anthropic/claude-haiku-4.5": {
7074
+ readonly "anthropic/claude-opus-4": {
7058
7075
  id: string;
7059
7076
  name: string;
7060
7077
  api: "openai-completions";
@@ -7071,7 +7088,7 @@ export declare const MODELS: {
7071
7088
  contextWindow: number;
7072
7089
  maxTokens: number;
7073
7090
  };
7074
- readonly "anthropic/claude-opus-4": {
7091
+ readonly "anthropic/claude-opus-4.1": {
7075
7092
  id: string;
7076
7093
  name: string;
7077
7094
  api: "openai-completions";
@@ -7088,7 +7105,7 @@ export declare const MODELS: {
7088
7105
  contextWindow: number;
7089
7106
  maxTokens: number;
7090
7107
  };
7091
- readonly "anthropic/claude-opus-4.1": {
7108
+ readonly "anthropic/claude-opus-4.5": {
7092
7109
  id: string;
7093
7110
  name: string;
7094
7111
  api: "openai-completions";
@@ -7105,7 +7122,7 @@ export declare const MODELS: {
7105
7122
  contextWindow: number;
7106
7123
  maxTokens: number;
7107
7124
  };
7108
- readonly "anthropic/claude-opus-4.5": {
7125
+ readonly "anthropic/claude-opus-4.6": {
7109
7126
  id: string;
7110
7127
  name: string;
7111
7128
  api: "openai-completions";
@@ -7122,7 +7139,7 @@ export declare const MODELS: {
7122
7139
  contextWindow: number;
7123
7140
  maxTokens: number;
7124
7141
  };
7125
- readonly "anthropic/claude-opus-4.6": {
7142
+ readonly "anthropic/claude-sonnet-4": {
7126
7143
  id: string;
7127
7144
  name: string;
7128
7145
  api: "openai-completions";
@@ -7139,7 +7156,7 @@ export declare const MODELS: {
7139
7156
  contextWindow: number;
7140
7157
  maxTokens: number;
7141
7158
  };
7142
- readonly "anthropic/claude-sonnet-4": {
7159
+ readonly "anthropic/claude-sonnet-4.5": {
7143
7160
  id: string;
7144
7161
  name: string;
7145
7162
  api: "openai-completions";
@@ -7156,7 +7173,7 @@ export declare const MODELS: {
7156
7173
  contextWindow: number;
7157
7174
  maxTokens: number;
7158
7175
  };
7159
- readonly "anthropic/claude-sonnet-4.5": {
7176
+ readonly "anthropic/claude-sonnet-4.6": {
7160
7177
  id: string;
7161
7178
  name: string;
7162
7179
  api: "openai-completions";
@@ -7173,14 +7190,14 @@ export declare const MODELS: {
7173
7190
  contextWindow: number;
7174
7191
  maxTokens: number;
7175
7192
  };
7176
- readonly "anthropic/claude-sonnet-4.6": {
7193
+ readonly "arcee-ai/trinity-large-preview:free": {
7177
7194
  id: string;
7178
7195
  name: string;
7179
7196
  api: "openai-completions";
7180
7197
  provider: string;
7181
7198
  baseUrl: string;
7182
- reasoning: true;
7183
- input: ("image" | "text")[];
7199
+ reasoning: false;
7200
+ input: "text"[];
7184
7201
  cost: {
7185
7202
  input: number;
7186
7203
  output: number;
@@ -7190,13 +7207,13 @@ export declare const MODELS: {
7190
7207
  contextWindow: number;
7191
7208
  maxTokens: number;
7192
7209
  };
7193
- readonly "arcee-ai/trinity-large-preview:free": {
7210
+ readonly "arcee-ai/trinity-large-thinking": {
7194
7211
  id: string;
7195
7212
  name: string;
7196
7213
  api: "openai-completions";
7197
7214
  provider: string;
7198
7215
  baseUrl: string;
7199
- reasoning: false;
7216
+ reasoning: true;
7200
7217
  input: "text"[];
7201
7218
  cost: {
7202
7219
  input: number;
@@ -7768,6 +7785,40 @@ export declare const MODELS: {
7768
7785
  contextWindow: number;
7769
7786
  maxTokens: number;
7770
7787
  };
7788
+ readonly "google/gemma-4-26b-a4b-it": {
7789
+ id: string;
7790
+ name: string;
7791
+ api: "openai-completions";
7792
+ provider: string;
7793
+ baseUrl: string;
7794
+ reasoning: true;
7795
+ input: ("image" | "text")[];
7796
+ cost: {
7797
+ input: number;
7798
+ output: number;
7799
+ cacheRead: number;
7800
+ cacheWrite: number;
7801
+ };
7802
+ contextWindow: number;
7803
+ maxTokens: number;
7804
+ };
7805
+ readonly "google/gemma-4-31b-it": {
7806
+ id: string;
7807
+ name: string;
7808
+ api: "openai-completions";
7809
+ provider: string;
7810
+ baseUrl: string;
7811
+ reasoning: true;
7812
+ input: ("image" | "text")[];
7813
+ cost: {
7814
+ input: number;
7815
+ output: number;
7816
+ cacheRead: number;
7817
+ cacheWrite: number;
7818
+ };
7819
+ contextWindow: number;
7820
+ maxTokens: number;
7821
+ };
7771
7822
  readonly "inception/mercury": {
7772
7823
  id: string;
7773
7824
  name: string;
@@ -8329,23 +8380,6 @@ export declare const MODELS: {
8329
8380
  contextWindow: number;
8330
8381
  maxTokens: number;
8331
8382
  };
8332
- readonly "mistralai/mistral-small-24b-instruct-2501": {
8333
- id: string;
8334
- name: string;
8335
- api: "openai-completions";
8336
- provider: string;
8337
- baseUrl: string;
8338
- reasoning: false;
8339
- input: "text"[];
8340
- cost: {
8341
- input: number;
8342
- output: number;
8343
- cacheRead: number;
8344
- cacheWrite: number;
8345
- };
8346
- contextWindow: number;
8347
- maxTokens: number;
8348
- };
8349
8383
  readonly "mistralai/mistral-small-2603": {
8350
8384
  id: string;
8351
8385
  name: string;
@@ -9400,6 +9434,40 @@ export declare const MODELS: {
9400
9434
  contextWindow: number;
9401
9435
  maxTokens: number;
9402
9436
  };
9437
+ readonly "openai/gpt-audio": {
9438
+ id: string;
9439
+ name: string;
9440
+ api: "openai-completions";
9441
+ provider: string;
9442
+ baseUrl: string;
9443
+ reasoning: false;
9444
+ input: "text"[];
9445
+ cost: {
9446
+ input: number;
9447
+ output: number;
9448
+ cacheRead: number;
9449
+ cacheWrite: number;
9450
+ };
9451
+ contextWindow: number;
9452
+ maxTokens: number;
9453
+ };
9454
+ readonly "openai/gpt-audio-mini": {
9455
+ id: string;
9456
+ name: string;
9457
+ api: "openai-completions";
9458
+ provider: string;
9459
+ baseUrl: string;
9460
+ reasoning: false;
9461
+ input: "text"[];
9462
+ cost: {
9463
+ input: number;
9464
+ output: number;
9465
+ cacheRead: number;
9466
+ cacheWrite: number;
9467
+ };
9468
+ contextWindow: number;
9469
+ maxTokens: number;
9470
+ };
9403
9471
  readonly "openai/gpt-oss-120b": {
9404
9472
  id: string;
9405
9473
  name: string;
@@ -10403,14 +10471,14 @@ export declare const MODELS: {
10403
10471
  contextWindow: number;
10404
10472
  maxTokens: number;
10405
10473
  };
10406
- readonly "qwen/qwen3.6-plus-preview:free": {
10474
+ readonly "qwen/qwen3.6-plus:free": {
10407
10475
  id: string;
10408
10476
  name: string;
10409
10477
  api: "openai-completions";
10410
10478
  provider: string;
10411
10479
  baseUrl: string;
10412
10480
  reasoning: true;
10413
- input: "text"[];
10481
+ input: ("image" | "text")[];
10414
10482
  cost: {
10415
10483
  input: number;
10416
10484
  output: number;
@@ -10437,7 +10505,7 @@ export declare const MODELS: {
10437
10505
  contextWindow: number;
10438
10506
  maxTokens: number;
10439
10507
  };
10440
- readonly "reka/reka-edge": {
10508
+ readonly "rekaai/reka-edge": {
10441
10509
  id: string;
10442
10510
  name: string;
10443
10511
  api: "openai-completions";
@@ -10726,7 +10794,7 @@ export declare const MODELS: {
10726
10794
  contextWindow: number;
10727
10795
  maxTokens: number;
10728
10796
  };
10729
- readonly "x-ai/grok-4.20-beta": {
10797
+ readonly "x-ai/grok-4.20": {
10730
10798
  id: string;
10731
10799
  name: string;
10732
10800
  api: "openai-completions";
@@ -10998,6 +11066,23 @@ export declare const MODELS: {
10998
11066
  contextWindow: number;
10999
11067
  maxTokens: number;
11000
11068
  };
11069
+ readonly "z-ai/glm-5v-turbo": {
11070
+ id: string;
11071
+ name: string;
11072
+ api: "openai-completions";
11073
+ provider: string;
11074
+ baseUrl: string;
11075
+ reasoning: true;
11076
+ input: ("image" | "text")[];
11077
+ cost: {
11078
+ input: number;
11079
+ output: number;
11080
+ cacheRead: number;
11081
+ cacheWrite: number;
11082
+ };
11083
+ contextWindow: number;
11084
+ maxTokens: number;
11085
+ };
11001
11086
  };
11002
11087
  readonly "vercel-ai-gateway": {
11003
11088
  readonly "alibaba/qwen-3-14b": {
@@ -11255,13 +11340,13 @@ export declare const MODELS: {
11255
11340
  contextWindow: number;
11256
11341
  maxTokens: number;
11257
11342
  };
11258
- readonly "anthropic/claude-3-haiku": {
11343
+ readonly "alibaba/qwen3.6-plus": {
11259
11344
  id: string;
11260
11345
  name: string;
11261
11346
  api: "anthropic-messages";
11262
11347
  provider: string;
11263
11348
  baseUrl: string;
11264
- reasoning: false;
11349
+ reasoning: true;
11265
11350
  input: ("image" | "text")[];
11266
11351
  cost: {
11267
11352
  input: number;
@@ -11272,7 +11357,7 @@ export declare const MODELS: {
11272
11357
  contextWindow: number;
11273
11358
  maxTokens: number;
11274
11359
  };
11275
- readonly "anthropic/claude-3.5-haiku": {
11360
+ readonly "anthropic/claude-3-haiku": {
11276
11361
  id: string;
11277
11362
  name: string;
11278
11363
  api: "anthropic-messages";
@@ -11289,7 +11374,7 @@ export declare const MODELS: {
11289
11374
  contextWindow: number;
11290
11375
  maxTokens: number;
11291
11376
  };
11292
- readonly "anthropic/claude-3.5-sonnet": {
11377
+ readonly "anthropic/claude-3.5-haiku": {
11293
11378
  id: string;
11294
11379
  name: string;
11295
11380
  api: "anthropic-messages";
@@ -11306,13 +11391,13 @@ export declare const MODELS: {
11306
11391
  contextWindow: number;
11307
11392
  maxTokens: number;
11308
11393
  };
11309
- readonly "anthropic/claude-3.5-sonnet-20240620": {
11394
+ readonly "anthropic/claude-3.7-sonnet": {
11310
11395
  id: string;
11311
11396
  name: string;
11312
11397
  api: "anthropic-messages";
11313
11398
  provider: string;
11314
11399
  baseUrl: string;
11315
- reasoning: false;
11400
+ reasoning: true;
11316
11401
  input: ("image" | "text")[];
11317
11402
  cost: {
11318
11403
  input: number;
@@ -11323,7 +11408,7 @@ export declare const MODELS: {
11323
11408
  contextWindow: number;
11324
11409
  maxTokens: number;
11325
11410
  };
11326
- readonly "anthropic/claude-3.7-sonnet": {
11411
+ readonly "anthropic/claude-haiku-4.5": {
11327
11412
  id: string;
11328
11413
  name: string;
11329
11414
  api: "anthropic-messages";
@@ -11340,7 +11425,7 @@ export declare const MODELS: {
11340
11425
  contextWindow: number;
11341
11426
  maxTokens: number;
11342
11427
  };
11343
- readonly "anthropic/claude-haiku-4.5": {
11428
+ readonly "anthropic/claude-opus-4": {
11344
11429
  id: string;
11345
11430
  name: string;
11346
11431
  api: "anthropic-messages";
@@ -11357,7 +11442,7 @@ export declare const MODELS: {
11357
11442
  contextWindow: number;
11358
11443
  maxTokens: number;
11359
11444
  };
11360
- readonly "anthropic/claude-opus-4": {
11445
+ readonly "anthropic/claude-opus-4.1": {
11361
11446
  id: string;
11362
11447
  name: string;
11363
11448
  api: "anthropic-messages";
@@ -11374,7 +11459,7 @@ export declare const MODELS: {
11374
11459
  contextWindow: number;
11375
11460
  maxTokens: number;
11376
11461
  };
11377
- readonly "anthropic/claude-opus-4.1": {
11462
+ readonly "anthropic/claude-opus-4.5": {
11378
11463
  id: string;
11379
11464
  name: string;
11380
11465
  api: "anthropic-messages";
@@ -11391,7 +11476,7 @@ export declare const MODELS: {
11391
11476
  contextWindow: number;
11392
11477
  maxTokens: number;
11393
11478
  };
11394
- readonly "anthropic/claude-opus-4.5": {
11479
+ readonly "anthropic/claude-opus-4.6": {
11395
11480
  id: string;
11396
11481
  name: string;
11397
11482
  api: "anthropic-messages";
@@ -11408,7 +11493,7 @@ export declare const MODELS: {
11408
11493
  contextWindow: number;
11409
11494
  maxTokens: number;
11410
11495
  };
11411
- readonly "anthropic/claude-opus-4.6": {
11496
+ readonly "anthropic/claude-sonnet-4": {
11412
11497
  id: string;
11413
11498
  name: string;
11414
11499
  api: "anthropic-messages";
@@ -11425,7 +11510,7 @@ export declare const MODELS: {
11425
11510
  contextWindow: number;
11426
11511
  maxTokens: number;
11427
11512
  };
11428
- readonly "anthropic/claude-sonnet-4": {
11513
+ readonly "anthropic/claude-sonnet-4.5": {
11429
11514
  id: string;
11430
11515
  name: string;
11431
11516
  api: "anthropic-messages";
@@ -11442,7 +11527,7 @@ export declare const MODELS: {
11442
11527
  contextWindow: number;
11443
11528
  maxTokens: number;
11444
11529
  };
11445
- readonly "anthropic/claude-sonnet-4.5": {
11530
+ readonly "anthropic/claude-sonnet-4.6": {
11446
11531
  id: string;
11447
11532
  name: string;
11448
11533
  api: "anthropic-messages";
@@ -11459,14 +11544,14 @@ export declare const MODELS: {
11459
11544
  contextWindow: number;
11460
11545
  maxTokens: number;
11461
11546
  };
11462
- readonly "anthropic/claude-sonnet-4.6": {
11547
+ readonly "arcee-ai/trinity-large-preview": {
11463
11548
  id: string;
11464
11549
  name: string;
11465
11550
  api: "anthropic-messages";
11466
11551
  provider: string;
11467
11552
  baseUrl: string;
11468
- reasoning: true;
11469
- input: ("image" | "text")[];
11553
+ reasoning: false;
11554
+ input: "text"[];
11470
11555
  cost: {
11471
11556
  input: number;
11472
11557
  output: number;
@@ -11476,13 +11561,13 @@ export declare const MODELS: {
11476
11561
  contextWindow: number;
11477
11562
  maxTokens: number;
11478
11563
  };
11479
- readonly "arcee-ai/trinity-large-preview": {
11564
+ readonly "arcee-ai/trinity-large-thinking": {
11480
11565
  id: string;
11481
11566
  name: string;
11482
11567
  api: "anthropic-messages";
11483
11568
  provider: string;
11484
11569
  baseUrl: string;
11485
- reasoning: false;
11570
+ reasoning: true;
11486
11571
  input: "text"[];
11487
11572
  cost: {
11488
11573
  input: number;
@@ -11782,14 +11867,14 @@ export declare const MODELS: {
11782
11867
  contextWindow: number;
11783
11868
  maxTokens: number;
11784
11869
  };
11785
- readonly "inception/mercury-2": {
11870
+ readonly "google/gemma-4-26b-a4b-it": {
11786
11871
  id: string;
11787
11872
  name: string;
11788
11873
  api: "anthropic-messages";
11789
11874
  provider: string;
11790
11875
  baseUrl: string;
11791
- reasoning: true;
11792
- input: "text"[];
11876
+ reasoning: false;
11877
+ input: ("image" | "text")[];
11793
11878
  cost: {
11794
11879
  input: number;
11795
11880
  output: number;
@@ -11799,14 +11884,14 @@ export declare const MODELS: {
11799
11884
  contextWindow: number;
11800
11885
  maxTokens: number;
11801
11886
  };
11802
- readonly "inception/mercury-coder-small": {
11887
+ readonly "google/gemma-4-31b-it": {
11803
11888
  id: string;
11804
11889
  name: string;
11805
11890
  api: "anthropic-messages";
11806
11891
  provider: string;
11807
11892
  baseUrl: string;
11808
11893
  reasoning: false;
11809
- input: "text"[];
11894
+ input: ("image" | "text")[];
11810
11895
  cost: {
11811
11896
  input: number;
11812
11897
  output: number;
@@ -11816,7 +11901,7 @@ export declare const MODELS: {
11816
11901
  contextWindow: number;
11817
11902
  maxTokens: number;
11818
11903
  };
11819
- readonly "kwaipilot/kat-coder-pro-v2": {
11904
+ readonly "inception/mercury-2": {
11820
11905
  id: string;
11821
11906
  name: string;
11822
11907
  api: "anthropic-messages";
@@ -11833,7 +11918,7 @@ export declare const MODELS: {
11833
11918
  contextWindow: number;
11834
11919
  maxTokens: number;
11835
11920
  };
11836
- readonly "meituan/longcat-flash-chat": {
11921
+ readonly "inception/mercury-coder-small": {
11837
11922
  id: string;
11838
11923
  name: string;
11839
11924
  api: "anthropic-messages";
@@ -11850,7 +11935,7 @@ export declare const MODELS: {
11850
11935
  contextWindow: number;
11851
11936
  maxTokens: number;
11852
11937
  };
11853
- readonly "meituan/longcat-flash-thinking": {
11938
+ readonly "kwaipilot/kat-coder-pro-v2": {
11854
11939
  id: string;
11855
11940
  name: string;
11856
11941
  api: "anthropic-messages";
@@ -11867,6 +11952,23 @@ export declare const MODELS: {
11867
11952
  contextWindow: number;
11868
11953
  maxTokens: number;
11869
11954
  };
11955
+ readonly "meituan/longcat-flash-chat": {
11956
+ id: string;
11957
+ name: string;
11958
+ api: "anthropic-messages";
11959
+ provider: string;
11960
+ baseUrl: string;
11961
+ reasoning: false;
11962
+ input: "text"[];
11963
+ cost: {
11964
+ input: number;
11965
+ output: number;
11966
+ cacheRead: number;
11967
+ cacheWrite: number;
11968
+ };
11969
+ contextWindow: number;
11970
+ maxTokens: number;
11971
+ };
11870
11972
  readonly "meta/llama-3.1-70b": {
11871
11973
  id: string;
11872
11974
  name: string;
@@ -13584,6 +13686,23 @@ export declare const MODELS: {
13584
13686
  contextWindow: number;
13585
13687
  maxTokens: number;
13586
13688
  };
13689
+ readonly "zai/glm-5v-turbo": {
13690
+ id: string;
13691
+ name: string;
13692
+ api: "anthropic-messages";
13693
+ provider: string;
13694
+ baseUrl: string;
13695
+ reasoning: true;
13696
+ input: ("image" | "text")[];
13697
+ cost: {
13698
+ input: number;
13699
+ output: number;
13700
+ cacheRead: number;
13701
+ cacheWrite: number;
13702
+ };
13703
+ contextWindow: number;
13704
+ maxTokens: number;
13705
+ };
13587
13706
  };
13588
13707
  readonly xai: {
13589
13708
  readonly "grok-2": {
@@ -14089,6 +14208,7 @@ export declare const MODELS: {
14089
14208
  compat: {
14090
14209
  supportsDeveloperRole: false;
14091
14210
  thinkingFormat: "zai";
14211
+ zaiToolStream: true;
14092
14212
  };
14093
14213
  reasoning: true;
14094
14214
  input: "text"[];
@@ -14110,6 +14230,7 @@ export declare const MODELS: {
14110
14230
  compat: {
14111
14231
  supportsDeveloperRole: false;
14112
14232
  thinkingFormat: "zai";
14233
+ zaiToolStream: true;
14113
14234
  };
14114
14235
  reasoning: true;
14115
14236
  input: ("image" | "text")[];
@@ -14131,6 +14252,7 @@ export declare const MODELS: {
14131
14252
  compat: {
14132
14253
  supportsDeveloperRole: false;
14133
14254
  thinkingFormat: "zai";
14255
+ zaiToolStream: true;
14134
14256
  };
14135
14257
  reasoning: true;
14136
14258
  input: "text"[];
@@ -14152,6 +14274,7 @@ export declare const MODELS: {
14152
14274
  compat: {
14153
14275
  supportsDeveloperRole: false;
14154
14276
  thinkingFormat: "zai";
14277
+ zaiToolStream: true;
14155
14278
  };
14156
14279
  reasoning: true;
14157
14280
  input: "text"[];
@@ -14173,6 +14296,7 @@ export declare const MODELS: {
14173
14296
  compat: {
14174
14297
  supportsDeveloperRole: false;
14175
14298
  thinkingFormat: "zai";
14299
+ zaiToolStream: true;
14176
14300
  };
14177
14301
  reasoning: true;
14178
14302
  input: "text"[];
@@ -14194,6 +14318,7 @@ export declare const MODELS: {
14194
14318
  compat: {
14195
14319
  supportsDeveloperRole: false;
14196
14320
  thinkingFormat: "zai";
14321
+ zaiToolStream: true;
14197
14322
  };
14198
14323
  reasoning: true;
14199
14324
  input: "text"[];
@@ -14215,6 +14340,7 @@ export declare const MODELS: {
14215
14340
  compat: {
14216
14341
  supportsDeveloperRole: false;
14217
14342
  thinkingFormat: "zai";
14343
+ zaiToolStream: true;
14218
14344
  };
14219
14345
  reasoning: true;
14220
14346
  input: "text"[];
@@ -14227,6 +14353,50 @@ export declare const MODELS: {
14227
14353
  contextWindow: number;
14228
14354
  maxTokens: number;
14229
14355
  };
14356
+ readonly "glm-5.1": {
14357
+ id: string;
14358
+ name: string;
14359
+ api: "openai-completions";
14360
+ provider: string;
14361
+ baseUrl: string;
14362
+ compat: {
14363
+ supportsDeveloperRole: false;
14364
+ thinkingFormat: "zai";
14365
+ zaiToolStream: true;
14366
+ };
14367
+ reasoning: true;
14368
+ input: "text"[];
14369
+ cost: {
14370
+ input: number;
14371
+ output: number;
14372
+ cacheRead: number;
14373
+ cacheWrite: number;
14374
+ };
14375
+ contextWindow: number;
14376
+ maxTokens: number;
14377
+ };
14378
+ readonly "glm-5v-turbo": {
14379
+ id: string;
14380
+ name: string;
14381
+ api: "openai-completions";
14382
+ provider: string;
14383
+ baseUrl: string;
14384
+ compat: {
14385
+ supportsDeveloperRole: false;
14386
+ thinkingFormat: "zai";
14387
+ zaiToolStream: true;
14388
+ };
14389
+ reasoning: true;
14390
+ input: ("image" | "text")[];
14391
+ cost: {
14392
+ input: number;
14393
+ output: number;
14394
+ cacheRead: number;
14395
+ cacheWrite: number;
14396
+ };
14397
+ contextWindow: number;
14398
+ maxTokens: number;
14399
+ };
14230
14400
  };
14231
14401
  };
14232
14402
  //# sourceMappingURL=models.generated.d.ts.map