@mariozechner/pi-ai 0.70.2 → 0.70.4

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -3702,6 +3702,29 @@ export declare const MODELS: {
3702
3702
  contextWindow: number;
3703
3703
  maxTokens: number;
3704
3704
  };
3705
+ readonly "gpt-5.5": {
3706
+ id: string;
3707
+ name: string;
3708
+ api: "openai-responses";
3709
+ provider: string;
3710
+ baseUrl: string;
3711
+ headers: {
3712
+ "User-Agent": string;
3713
+ "Editor-Version": string;
3714
+ "Editor-Plugin-Version": string;
3715
+ "Copilot-Integration-Id": string;
3716
+ };
3717
+ reasoning: true;
3718
+ input: ("image" | "text")[];
3719
+ cost: {
3720
+ input: number;
3721
+ output: number;
3722
+ cacheRead: number;
3723
+ cacheWrite: number;
3724
+ };
3725
+ contextWindow: number;
3726
+ maxTokens: number;
3727
+ };
3705
3728
  readonly "grok-code-fast-1": {
3706
3729
  id: string;
3707
3730
  name: string;
@@ -5240,6 +5263,26 @@ export declare const MODELS: {
5240
5263
  contextWindow: number;
5241
5264
  maxTokens: number;
5242
5265
  };
5266
+ readonly "deepseek-ai/DeepSeek-V4-Pro": {
5267
+ id: string;
5268
+ name: string;
5269
+ api: "openai-completions";
5270
+ provider: string;
5271
+ baseUrl: string;
5272
+ compat: {
5273
+ supportsDeveloperRole: false;
5274
+ };
5275
+ reasoning: true;
5276
+ input: "text"[];
5277
+ cost: {
5278
+ input: number;
5279
+ output: number;
5280
+ cacheRead: number;
5281
+ cacheWrite: number;
5282
+ };
5283
+ contextWindow: number;
5284
+ maxTokens: number;
5285
+ };
5243
5286
  readonly "moonshotai/Kimi-K2-Instruct": {
5244
5287
  id: string;
5245
5288
  name: string;
@@ -6888,23 +6931,6 @@ export declare const MODELS: {
6888
6931
  contextWindow: number;
6889
6932
  maxTokens: number;
6890
6933
  };
6891
- readonly "claude-3-5-haiku": {
6892
- id: string;
6893
- name: string;
6894
- api: "anthropic-messages";
6895
- provider: string;
6896
- baseUrl: string;
6897
- reasoning: false;
6898
- input: ("image" | "text")[];
6899
- cost: {
6900
- input: number;
6901
- output: number;
6902
- cacheRead: number;
6903
- cacheWrite: number;
6904
- };
6905
- contextWindow: number;
6906
- maxTokens: number;
6907
- };
6908
6934
  readonly "claude-haiku-4-5": {
6909
6935
  id: string;
6910
6936
  name: string;
@@ -7347,6 +7373,40 @@ export declare const MODELS: {
7347
7373
  contextWindow: number;
7348
7374
  maxTokens: number;
7349
7375
  };
7376
+ readonly "gpt-5.5": {
7377
+ id: string;
7378
+ name: string;
7379
+ api: "openai-responses";
7380
+ provider: string;
7381
+ baseUrl: string;
7382
+ reasoning: true;
7383
+ input: ("image" | "text")[];
7384
+ cost: {
7385
+ input: number;
7386
+ output: number;
7387
+ cacheRead: number;
7388
+ cacheWrite: number;
7389
+ };
7390
+ contextWindow: number;
7391
+ maxTokens: number;
7392
+ };
7393
+ readonly "gpt-5.5-pro": {
7394
+ id: string;
7395
+ name: string;
7396
+ api: "openai-responses";
7397
+ provider: string;
7398
+ baseUrl: string;
7399
+ reasoning: true;
7400
+ input: ("image" | "text")[];
7401
+ cost: {
7402
+ input: number;
7403
+ output: number;
7404
+ cacheRead: number;
7405
+ cacheWrite: number;
7406
+ };
7407
+ contextWindow: number;
7408
+ maxTokens: number;
7409
+ };
7350
7410
  readonly "hy3-preview-free": {
7351
7411
  id: string;
7352
7412
  name: string;
@@ -7519,6 +7579,46 @@ export declare const MODELS: {
7519
7579
  };
7520
7580
  };
7521
7581
  readonly "opencode-go": {
7582
+ readonly "deepseek-v4-flash": {
7583
+ id: string;
7584
+ name: string;
7585
+ api: "openai-completions";
7586
+ provider: string;
7587
+ baseUrl: string;
7588
+ compat: {
7589
+ requiresReasoningContentOnAssistantMessages: true;
7590
+ };
7591
+ reasoning: true;
7592
+ input: "text"[];
7593
+ cost: {
7594
+ input: number;
7595
+ output: number;
7596
+ cacheRead: number;
7597
+ cacheWrite: number;
7598
+ };
7599
+ contextWindow: number;
7600
+ maxTokens: number;
7601
+ };
7602
+ readonly "deepseek-v4-pro": {
7603
+ id: string;
7604
+ name: string;
7605
+ api: "openai-completions";
7606
+ provider: string;
7607
+ baseUrl: string;
7608
+ compat: {
7609
+ requiresReasoningContentOnAssistantMessages: true;
7610
+ };
7611
+ reasoning: true;
7612
+ input: "text"[];
7613
+ cost: {
7614
+ input: number;
7615
+ output: number;
7616
+ cacheRead: number;
7617
+ cacheWrite: number;
7618
+ };
7619
+ contextWindow: number;
7620
+ maxTokens: number;
7621
+ };
7522
7622
  readonly "glm-5": {
7523
7623
  id: string;
7524
7624
  name: string;
@@ -8462,6 +8562,9 @@ export declare const MODELS: {
8462
8562
  api: "openai-completions";
8463
8563
  provider: string;
8464
8564
  baseUrl: string;
8565
+ compat: {
8566
+ requiresReasoningContentOnAssistantMessages: true;
8567
+ };
8465
8568
  reasoning: true;
8466
8569
  input: "text"[];
8467
8570
  cost: {
@@ -8479,6 +8582,9 @@ export declare const MODELS: {
8479
8582
  api: "openai-completions";
8480
8583
  provider: string;
8481
8584
  baseUrl: string;
8585
+ compat: {
8586
+ requiresReasoningContentOnAssistantMessages: true;
8587
+ };
8482
8588
  reasoning: true;
8483
8589
  input: "text"[];
8484
8590
  cost: {
@@ -8711,6 +8817,40 @@ export declare const MODELS: {
8711
8817
  contextWindow: number;
8712
8818
  maxTokens: number;
8713
8819
  };
8820
+ readonly "google/gemma-3-12b-it": {
8821
+ id: string;
8822
+ name: string;
8823
+ api: "openai-completions";
8824
+ provider: string;
8825
+ baseUrl: string;
8826
+ reasoning: false;
8827
+ input: ("image" | "text")[];
8828
+ cost: {
8829
+ input: number;
8830
+ output: number;
8831
+ cacheRead: number;
8832
+ cacheWrite: number;
8833
+ };
8834
+ contextWindow: number;
8835
+ maxTokens: number;
8836
+ };
8837
+ readonly "google/gemma-3-27b-it": {
8838
+ id: string;
8839
+ name: string;
8840
+ api: "openai-completions";
8841
+ provider: string;
8842
+ baseUrl: string;
8843
+ reasoning: false;
8844
+ input: ("image" | "text")[];
8845
+ cost: {
8846
+ input: number;
8847
+ output: number;
8848
+ cacheRead: number;
8849
+ cacheWrite: number;
8850
+ };
8851
+ contextWindow: number;
8852
+ maxTokens: number;
8853
+ };
8714
8854
  readonly "google/gemma-4-26b-a4b-it": {
8715
8855
  id: string;
8716
8856
  name: string;
@@ -10326,6 +10466,40 @@ export declare const MODELS: {
10326
10466
  contextWindow: number;
10327
10467
  maxTokens: number;
10328
10468
  };
10469
+ readonly "openai/gpt-5.5": {
10470
+ id: string;
10471
+ name: string;
10472
+ api: "openai-completions";
10473
+ provider: string;
10474
+ baseUrl: string;
10475
+ reasoning: true;
10476
+ input: ("image" | "text")[];
10477
+ cost: {
10478
+ input: number;
10479
+ output: number;
10480
+ cacheRead: number;
10481
+ cacheWrite: number;
10482
+ };
10483
+ contextWindow: number;
10484
+ maxTokens: number;
10485
+ };
10486
+ readonly "openai/gpt-5.5-pro": {
10487
+ id: string;
10488
+ name: string;
10489
+ api: "openai-completions";
10490
+ provider: string;
10491
+ baseUrl: string;
10492
+ reasoning: true;
10493
+ input: ("image" | "text")[];
10494
+ cost: {
10495
+ input: number;
10496
+ output: number;
10497
+ cacheRead: number;
10498
+ cacheWrite: number;
10499
+ };
10500
+ contextWindow: number;
10501
+ maxTokens: number;
10502
+ };
10329
10503
  readonly "openai/gpt-audio": {
10330
10504
  id: string;
10331
10505
  name: string;
@@ -11363,7 +11537,7 @@ export declare const MODELS: {
11363
11537
  contextWindow: number;
11364
11538
  maxTokens: number;
11365
11539
  };
11366
- readonly "qwen/qwen3.6-plus": {
11540
+ readonly "qwen/qwen3.5-plus-20260420": {
11367
11541
  id: string;
11368
11542
  name: string;
11369
11543
  api: "openai-completions";
@@ -11380,14 +11554,14 @@ export declare const MODELS: {
11380
11554
  contextWindow: number;
11381
11555
  maxTokens: number;
11382
11556
  };
11383
- readonly "qwen/qwq-32b": {
11557
+ readonly "qwen/qwen3.6-27b": {
11384
11558
  id: string;
11385
11559
  name: string;
11386
11560
  api: "openai-completions";
11387
11561
  provider: string;
11388
11562
  baseUrl: string;
11389
11563
  reasoning: true;
11390
- input: "text"[];
11564
+ input: ("image" | "text")[];
11391
11565
  cost: {
11392
11566
  input: number;
11393
11567
  output: number;
@@ -11397,13 +11571,13 @@ export declare const MODELS: {
11397
11571
  contextWindow: number;
11398
11572
  maxTokens: number;
11399
11573
  };
11400
- readonly "rekaai/reka-edge": {
11574
+ readonly "qwen/qwen3.6-flash": {
11401
11575
  id: string;
11402
11576
  name: string;
11403
11577
  api: "openai-completions";
11404
11578
  provider: string;
11405
11579
  baseUrl: string;
11406
- reasoning: false;
11580
+ reasoning: true;
11407
11581
  input: ("image" | "text")[];
11408
11582
  cost: {
11409
11583
  input: number;
@@ -11414,13 +11588,13 @@ export declare const MODELS: {
11414
11588
  contextWindow: number;
11415
11589
  maxTokens: number;
11416
11590
  };
11417
- readonly "relace/relace-search": {
11591
+ readonly "qwen/qwen3.6-max-preview": {
11418
11592
  id: string;
11419
11593
  name: string;
11420
11594
  api: "openai-completions";
11421
11595
  provider: string;
11422
11596
  baseUrl: string;
11423
- reasoning: false;
11597
+ reasoning: true;
11424
11598
  input: "text"[];
11425
11599
  cost: {
11426
11600
  input: number;
@@ -11431,14 +11605,14 @@ export declare const MODELS: {
11431
11605
  contextWindow: number;
11432
11606
  maxTokens: number;
11433
11607
  };
11434
- readonly "sao10k/l3-euryale-70b": {
11608
+ readonly "qwen/qwen3.6-plus": {
11435
11609
  id: string;
11436
11610
  name: string;
11437
11611
  api: "openai-completions";
11438
11612
  provider: string;
11439
11613
  baseUrl: string;
11440
- reasoning: false;
11441
- input: "text"[];
11614
+ reasoning: true;
11615
+ input: ("image" | "text")[];
11442
11616
  cost: {
11443
11617
  input: number;
11444
11618
  output: number;
@@ -11448,13 +11622,13 @@ export declare const MODELS: {
11448
11622
  contextWindow: number;
11449
11623
  maxTokens: number;
11450
11624
  };
11451
- readonly "sao10k/l3.1-euryale-70b": {
11625
+ readonly "qwen/qwq-32b": {
11452
11626
  id: string;
11453
11627
  name: string;
11454
11628
  api: "openai-completions";
11455
11629
  provider: string;
11456
11630
  baseUrl: string;
11457
- reasoning: false;
11631
+ reasoning: true;
11458
11632
  input: "text"[];
11459
11633
  cost: {
11460
11634
  input: number;
@@ -11465,14 +11639,14 @@ export declare const MODELS: {
11465
11639
  contextWindow: number;
11466
11640
  maxTokens: number;
11467
11641
  };
11468
- readonly "stepfun/step-3.5-flash": {
11642
+ readonly "rekaai/reka-edge": {
11469
11643
  id: string;
11470
11644
  name: string;
11471
11645
  api: "openai-completions";
11472
11646
  provider: string;
11473
11647
  baseUrl: string;
11474
- reasoning: true;
11475
- input: "text"[];
11648
+ reasoning: false;
11649
+ input: ("image" | "text")[];
11476
11650
  cost: {
11477
11651
  input: number;
11478
11652
  output: number;
@@ -11482,13 +11656,13 @@ export declare const MODELS: {
11482
11656
  contextWindow: number;
11483
11657
  maxTokens: number;
11484
11658
  };
11485
- readonly "tencent/hy3-preview:free": {
11659
+ readonly "relace/relace-search": {
11486
11660
  id: string;
11487
11661
  name: string;
11488
11662
  api: "openai-completions";
11489
11663
  provider: string;
11490
11664
  baseUrl: string;
11491
- reasoning: true;
11665
+ reasoning: false;
11492
11666
  input: "text"[];
11493
11667
  cost: {
11494
11668
  input: number;
@@ -11499,7 +11673,7 @@ export declare const MODELS: {
11499
11673
  contextWindow: number;
11500
11674
  maxTokens: number;
11501
11675
  };
11502
- readonly "thedrummer/rocinante-12b": {
11676
+ readonly "sao10k/l3-euryale-70b": {
11503
11677
  id: string;
11504
11678
  name: string;
11505
11679
  api: "openai-completions";
@@ -11516,7 +11690,7 @@ export declare const MODELS: {
11516
11690
  contextWindow: number;
11517
11691
  maxTokens: number;
11518
11692
  };
11519
- readonly "thedrummer/unslopnemo-12b": {
11693
+ readonly "sao10k/l3.1-euryale-70b": {
11520
11694
  id: string;
11521
11695
  name: string;
11522
11696
  api: "openai-completions";
@@ -11533,7 +11707,7 @@ export declare const MODELS: {
11533
11707
  contextWindow: number;
11534
11708
  maxTokens: number;
11535
11709
  };
11536
- readonly "tngtech/deepseek-r1t2-chimera": {
11710
+ readonly "stepfun/step-3.5-flash": {
11537
11711
  id: string;
11538
11712
  name: string;
11539
11713
  api: "openai-completions";
@@ -11550,7 +11724,7 @@ export declare const MODELS: {
11550
11724
  contextWindow: number;
11551
11725
  maxTokens: number;
11552
11726
  };
11553
- readonly "upstage/solar-pro-3": {
11727
+ readonly "tencent/hy3-preview:free": {
11554
11728
  id: string;
11555
11729
  name: string;
11556
11730
  api: "openai-completions";
@@ -11567,7 +11741,7 @@ export declare const MODELS: {
11567
11741
  contextWindow: number;
11568
11742
  maxTokens: number;
11569
11743
  };
11570
- readonly "x-ai/grok-3": {
11744
+ readonly "thedrummer/rocinante-12b": {
11571
11745
  id: string;
11572
11746
  name: string;
11573
11747
  api: "openai-completions";
@@ -11584,7 +11758,7 @@ export declare const MODELS: {
11584
11758
  contextWindow: number;
11585
11759
  maxTokens: number;
11586
11760
  };
11587
- readonly "x-ai/grok-3-beta": {
11761
+ readonly "thedrummer/unslopnemo-12b": {
11588
11762
  id: string;
11589
11763
  name: string;
11590
11764
  api: "openai-completions";
@@ -11601,7 +11775,7 @@ export declare const MODELS: {
11601
11775
  contextWindow: number;
11602
11776
  maxTokens: number;
11603
11777
  };
11604
- readonly "x-ai/grok-3-mini": {
11778
+ readonly "tngtech/deepseek-r1t2-chimera": {
11605
11779
  id: string;
11606
11780
  name: string;
11607
11781
  api: "openai-completions";
@@ -11618,7 +11792,7 @@ export declare const MODELS: {
11618
11792
  contextWindow: number;
11619
11793
  maxTokens: number;
11620
11794
  };
11621
- readonly "x-ai/grok-3-mini-beta": {
11795
+ readonly "upstage/solar-pro-3": {
11622
11796
  id: string;
11623
11797
  name: string;
11624
11798
  api: "openai-completions";
@@ -11635,14 +11809,14 @@ export declare const MODELS: {
11635
11809
  contextWindow: number;
11636
11810
  maxTokens: number;
11637
11811
  };
11638
- readonly "x-ai/grok-4": {
11812
+ readonly "x-ai/grok-3": {
11639
11813
  id: string;
11640
11814
  name: string;
11641
11815
  api: "openai-completions";
11642
11816
  provider: string;
11643
11817
  baseUrl: string;
11644
- reasoning: true;
11645
- input: ("image" | "text")[];
11818
+ reasoning: false;
11819
+ input: "text"[];
11646
11820
  cost: {
11647
11821
  input: number;
11648
11822
  output: number;
@@ -11652,13 +11826,81 @@ export declare const MODELS: {
11652
11826
  contextWindow: number;
11653
11827
  maxTokens: number;
11654
11828
  };
11655
- readonly "x-ai/grok-4-fast": {
11829
+ readonly "x-ai/grok-3-beta": {
11656
11830
  id: string;
11657
11831
  name: string;
11658
11832
  api: "openai-completions";
11659
11833
  provider: string;
11660
11834
  baseUrl: string;
11661
- reasoning: true;
11835
+ reasoning: false;
11836
+ input: "text"[];
11837
+ cost: {
11838
+ input: number;
11839
+ output: number;
11840
+ cacheRead: number;
11841
+ cacheWrite: number;
11842
+ };
11843
+ contextWindow: number;
11844
+ maxTokens: number;
11845
+ };
11846
+ readonly "x-ai/grok-3-mini": {
11847
+ id: string;
11848
+ name: string;
11849
+ api: "openai-completions";
11850
+ provider: string;
11851
+ baseUrl: string;
11852
+ reasoning: true;
11853
+ input: "text"[];
11854
+ cost: {
11855
+ input: number;
11856
+ output: number;
11857
+ cacheRead: number;
11858
+ cacheWrite: number;
11859
+ };
11860
+ contextWindow: number;
11861
+ maxTokens: number;
11862
+ };
11863
+ readonly "x-ai/grok-3-mini-beta": {
11864
+ id: string;
11865
+ name: string;
11866
+ api: "openai-completions";
11867
+ provider: string;
11868
+ baseUrl: string;
11869
+ reasoning: true;
11870
+ input: "text"[];
11871
+ cost: {
11872
+ input: number;
11873
+ output: number;
11874
+ cacheRead: number;
11875
+ cacheWrite: number;
11876
+ };
11877
+ contextWindow: number;
11878
+ maxTokens: number;
11879
+ };
11880
+ readonly "x-ai/grok-4": {
11881
+ id: string;
11882
+ name: string;
11883
+ api: "openai-completions";
11884
+ provider: string;
11885
+ baseUrl: string;
11886
+ reasoning: true;
11887
+ input: ("image" | "text")[];
11888
+ cost: {
11889
+ input: number;
11890
+ output: number;
11891
+ cacheRead: number;
11892
+ cacheWrite: number;
11893
+ };
11894
+ contextWindow: number;
11895
+ maxTokens: number;
11896
+ };
11897
+ readonly "x-ai/grok-4-fast": {
11898
+ id: string;
11899
+ name: string;
11900
+ api: "openai-completions";
11901
+ provider: string;
11902
+ baseUrl: string;
11903
+ reasoning: true;
11662
11904
  input: ("image" | "text")[];
11663
11905
  cost: {
11664
11906
  input: number;
@@ -12026,6 +12268,23 @@ export declare const MODELS: {
12026
12268
  contextWindow: number;
12027
12269
  maxTokens: number;
12028
12270
  };
12271
+ readonly "~anthropic/claude-haiku-latest": {
12272
+ id: string;
12273
+ name: string;
12274
+ api: "openai-completions";
12275
+ provider: string;
12276
+ baseUrl: string;
12277
+ reasoning: true;
12278
+ input: ("image" | "text")[];
12279
+ cost: {
12280
+ input: number;
12281
+ output: number;
12282
+ cacheRead: number;
12283
+ cacheWrite: number;
12284
+ };
12285
+ contextWindow: number;
12286
+ maxTokens: number;
12287
+ };
12029
12288
  readonly "~anthropic/claude-opus-latest": {
12030
12289
  id: string;
12031
12290
  name: string;
@@ -12043,6 +12302,108 @@ export declare const MODELS: {
12043
12302
  contextWindow: number;
12044
12303
  maxTokens: number;
12045
12304
  };
12305
+ readonly "~anthropic/claude-sonnet-latest": {
12306
+ id: string;
12307
+ name: string;
12308
+ api: "openai-completions";
12309
+ provider: string;
12310
+ baseUrl: string;
12311
+ reasoning: true;
12312
+ input: ("image" | "text")[];
12313
+ cost: {
12314
+ input: number;
12315
+ output: number;
12316
+ cacheRead: number;
12317
+ cacheWrite: number;
12318
+ };
12319
+ contextWindow: number;
12320
+ maxTokens: number;
12321
+ };
12322
+ readonly "~google/gemini-flash-latest": {
12323
+ id: string;
12324
+ name: string;
12325
+ api: "openai-completions";
12326
+ provider: string;
12327
+ baseUrl: string;
12328
+ reasoning: true;
12329
+ input: ("image" | "text")[];
12330
+ cost: {
12331
+ input: number;
12332
+ output: number;
12333
+ cacheRead: number;
12334
+ cacheWrite: number;
12335
+ };
12336
+ contextWindow: number;
12337
+ maxTokens: number;
12338
+ };
12339
+ readonly "~google/gemini-pro-latest": {
12340
+ id: string;
12341
+ name: string;
12342
+ api: "openai-completions";
12343
+ provider: string;
12344
+ baseUrl: string;
12345
+ reasoning: true;
12346
+ input: ("image" | "text")[];
12347
+ cost: {
12348
+ input: number;
12349
+ output: number;
12350
+ cacheRead: number;
12351
+ cacheWrite: number;
12352
+ };
12353
+ contextWindow: number;
12354
+ maxTokens: number;
12355
+ };
12356
+ readonly "~moonshotai/kimi-latest": {
12357
+ id: string;
12358
+ name: string;
12359
+ api: "openai-completions";
12360
+ provider: string;
12361
+ baseUrl: string;
12362
+ reasoning: true;
12363
+ input: ("image" | "text")[];
12364
+ cost: {
12365
+ input: number;
12366
+ output: number;
12367
+ cacheRead: number;
12368
+ cacheWrite: number;
12369
+ };
12370
+ contextWindow: number;
12371
+ maxTokens: number;
12372
+ };
12373
+ readonly "~openai/gpt-latest": {
12374
+ id: string;
12375
+ name: string;
12376
+ api: "openai-completions";
12377
+ provider: string;
12378
+ baseUrl: string;
12379
+ reasoning: true;
12380
+ input: ("image" | "text")[];
12381
+ cost: {
12382
+ input: number;
12383
+ output: number;
12384
+ cacheRead: number;
12385
+ cacheWrite: number;
12386
+ };
12387
+ contextWindow: number;
12388
+ maxTokens: number;
12389
+ };
12390
+ readonly "~openai/gpt-mini-latest": {
12391
+ id: string;
12392
+ name: string;
12393
+ api: "openai-completions";
12394
+ provider: string;
12395
+ baseUrl: string;
12396
+ reasoning: true;
12397
+ input: ("image" | "text")[];
12398
+ cost: {
12399
+ input: number;
12400
+ output: number;
12401
+ cacheRead: number;
12402
+ cacheWrite: number;
12403
+ };
12404
+ contextWindow: number;
12405
+ maxTokens: number;
12406
+ };
12046
12407
  };
12047
12408
  readonly "vercel-ai-gateway": {
12048
12409
  readonly "alibaba/qwen-3-14b": {
@@ -14017,6 +14378,40 @@ export declare const MODELS: {
14017
14378
  contextWindow: number;
14018
14379
  maxTokens: number;
14019
14380
  };
14381
+ readonly "openai/gpt-5.5": {
14382
+ id: string;
14383
+ name: string;
14384
+ api: "anthropic-messages";
14385
+ provider: string;
14386
+ baseUrl: string;
14387
+ reasoning: true;
14388
+ input: ("image" | "text")[];
14389
+ cost: {
14390
+ input: number;
14391
+ output: number;
14392
+ cacheRead: number;
14393
+ cacheWrite: number;
14394
+ };
14395
+ contextWindow: number;
14396
+ maxTokens: number;
14397
+ };
14398
+ readonly "openai/gpt-5.5-pro": {
14399
+ id: string;
14400
+ name: string;
14401
+ api: "anthropic-messages";
14402
+ provider: string;
14403
+ baseUrl: string;
14404
+ reasoning: true;
14405
+ input: ("image" | "text")[];
14406
+ cost: {
14407
+ input: number;
14408
+ output: number;
14409
+ cacheRead: number;
14410
+ cacheWrite: number;
14411
+ };
14412
+ contextWindow: number;
14413
+ maxTokens: number;
14414
+ };
14020
14415
  readonly "openai/gpt-oss-20b": {
14021
14416
  id: string;
14022
14417
  name: string;
@@ -15143,27 +15538,6 @@ export declare const MODELS: {
15143
15538
  };
15144
15539
  };
15145
15540
  readonly zai: {
15146
- readonly "glm-4.5": {
15147
- id: string;
15148
- name: string;
15149
- api: "openai-completions";
15150
- provider: string;
15151
- baseUrl: string;
15152
- compat: {
15153
- supportsDeveloperRole: false;
15154
- thinkingFormat: "zai";
15155
- };
15156
- reasoning: true;
15157
- input: "text"[];
15158
- cost: {
15159
- input: number;
15160
- output: number;
15161
- cacheRead: number;
15162
- cacheWrite: number;
15163
- };
15164
- contextWindow: number;
15165
- maxTokens: number;
15166
- };
15167
15541
  readonly "glm-4.5-air": {
15168
15542
  id: string;
15169
15543
  name: string;
@@ -15185,92 +15559,6 @@ export declare const MODELS: {
15185
15559
  contextWindow: number;
15186
15560
  maxTokens: number;
15187
15561
  };
15188
- readonly "glm-4.5-flash": {
15189
- id: string;
15190
- name: string;
15191
- api: "openai-completions";
15192
- provider: string;
15193
- baseUrl: string;
15194
- compat: {
15195
- supportsDeveloperRole: false;
15196
- thinkingFormat: "zai";
15197
- };
15198
- reasoning: true;
15199
- input: "text"[];
15200
- cost: {
15201
- input: number;
15202
- output: number;
15203
- cacheRead: number;
15204
- cacheWrite: number;
15205
- };
15206
- contextWindow: number;
15207
- maxTokens: number;
15208
- };
15209
- readonly "glm-4.5v": {
15210
- id: string;
15211
- name: string;
15212
- api: "openai-completions";
15213
- provider: string;
15214
- baseUrl: string;
15215
- compat: {
15216
- supportsDeveloperRole: false;
15217
- thinkingFormat: "zai";
15218
- };
15219
- reasoning: true;
15220
- input: ("image" | "text")[];
15221
- cost: {
15222
- input: number;
15223
- output: number;
15224
- cacheRead: number;
15225
- cacheWrite: number;
15226
- };
15227
- contextWindow: number;
15228
- maxTokens: number;
15229
- };
15230
- readonly "glm-4.6": {
15231
- id: string;
15232
- name: string;
15233
- api: "openai-completions";
15234
- provider: string;
15235
- baseUrl: string;
15236
- compat: {
15237
- supportsDeveloperRole: false;
15238
- thinkingFormat: "zai";
15239
- zaiToolStream: true;
15240
- };
15241
- reasoning: true;
15242
- input: "text"[];
15243
- cost: {
15244
- input: number;
15245
- output: number;
15246
- cacheRead: number;
15247
- cacheWrite: number;
15248
- };
15249
- contextWindow: number;
15250
- maxTokens: number;
15251
- };
15252
- readonly "glm-4.6v": {
15253
- id: string;
15254
- name: string;
15255
- api: "openai-completions";
15256
- provider: string;
15257
- baseUrl: string;
15258
- compat: {
15259
- supportsDeveloperRole: false;
15260
- thinkingFormat: "zai";
15261
- zaiToolStream: true;
15262
- };
15263
- reasoning: true;
15264
- input: ("image" | "text")[];
15265
- cost: {
15266
- input: number;
15267
- output: number;
15268
- cacheRead: number;
15269
- cacheWrite: number;
15270
- };
15271
- contextWindow: number;
15272
- maxTokens: number;
15273
- };
15274
15562
  readonly "glm-4.7": {
15275
15563
  id: string;
15276
15564
  name: string;
@@ -15293,72 +15581,6 @@ export declare const MODELS: {
15293
15581
  contextWindow: number;
15294
15582
  maxTokens: number;
15295
15583
  };
15296
- readonly "glm-4.7-flash": {
15297
- id: string;
15298
- name: string;
15299
- api: "openai-completions";
15300
- provider: string;
15301
- baseUrl: string;
15302
- compat: {
15303
- supportsDeveloperRole: false;
15304
- thinkingFormat: "zai";
15305
- zaiToolStream: true;
15306
- };
15307
- reasoning: true;
15308
- input: "text"[];
15309
- cost: {
15310
- input: number;
15311
- output: number;
15312
- cacheRead: number;
15313
- cacheWrite: number;
15314
- };
15315
- contextWindow: number;
15316
- maxTokens: number;
15317
- };
15318
- readonly "glm-4.7-flashx": {
15319
- id: string;
15320
- name: string;
15321
- api: "openai-completions";
15322
- provider: string;
15323
- baseUrl: string;
15324
- compat: {
15325
- supportsDeveloperRole: false;
15326
- thinkingFormat: "zai";
15327
- zaiToolStream: true;
15328
- };
15329
- reasoning: true;
15330
- input: "text"[];
15331
- cost: {
15332
- input: number;
15333
- output: number;
15334
- cacheRead: number;
15335
- cacheWrite: number;
15336
- };
15337
- contextWindow: number;
15338
- maxTokens: number;
15339
- };
15340
- readonly "glm-5": {
15341
- id: string;
15342
- name: string;
15343
- api: "openai-completions";
15344
- provider: string;
15345
- baseUrl: string;
15346
- compat: {
15347
- supportsDeveloperRole: false;
15348
- thinkingFormat: "zai";
15349
- zaiToolStream: true;
15350
- };
15351
- reasoning: true;
15352
- input: "text"[];
15353
- cost: {
15354
- input: number;
15355
- output: number;
15356
- cacheRead: number;
15357
- cacheWrite: number;
15358
- };
15359
- contextWindow: number;
15360
- maxTokens: number;
15361
- };
15362
15584
  readonly "glm-5-turbo": {
15363
15585
  id: string;
15364
15586
  name: string;
@@ -15403,28 +15625,6 @@ export declare const MODELS: {
15403
15625
  contextWindow: number;
15404
15626
  maxTokens: number;
15405
15627
  };
15406
- readonly "glm-5v-turbo": {
15407
- id: string;
15408
- name: string;
15409
- api: "openai-completions";
15410
- provider: string;
15411
- baseUrl: string;
15412
- compat: {
15413
- supportsDeveloperRole: false;
15414
- thinkingFormat: "zai";
15415
- zaiToolStream: true;
15416
- };
15417
- reasoning: true;
15418
- input: ("image" | "text")[];
15419
- cost: {
15420
- input: number;
15421
- output: number;
15422
- cacheRead: number;
15423
- cacheWrite: number;
15424
- };
15425
- contextWindow: number;
15426
- maxTokens: number;
15427
- };
15428
15628
  };
15429
15629
  };
15430
15630
  //# sourceMappingURL=models.generated.d.ts.map