@hyperspaceng/neural-ai 0.70.3 → 0.70.7

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -2537,6 +2537,23 @@ export declare const MODELS: {
2537
2537
  contextWindow: number;
2538
2538
  maxTokens: number;
2539
2539
  };
2540
+ readonly "gpt-5.5-pro": {
2541
+ id: string;
2542
+ name: string;
2543
+ api: "azure-openai-responses";
2544
+ provider: string;
2545
+ baseUrl: string;
2546
+ reasoning: true;
2547
+ input: ("image" | "text")[];
2548
+ cost: {
2549
+ input: number;
2550
+ output: number;
2551
+ cacheRead: number;
2552
+ cacheWrite: number;
2553
+ };
2554
+ contextWindow: number;
2555
+ maxTokens: number;
2556
+ };
2540
2557
  readonly o1: {
2541
2558
  id: string;
2542
2559
  name: string;
@@ -2744,6 +2761,168 @@ export declare const MODELS: {
2744
2761
  maxTokens: number;
2745
2762
  };
2746
2763
  };
2764
+ readonly "cloudflare-workers-ai": {
2765
+ readonly "@cf/google/gemma-4-26b-a4b-it": {
2766
+ id: string;
2767
+ name: string;
2768
+ api: "openai-completions";
2769
+ provider: string;
2770
+ baseUrl: string;
2771
+ compat: {
2772
+ sendSessionAffinityHeaders: true;
2773
+ };
2774
+ reasoning: true;
2775
+ input: ("image" | "text")[];
2776
+ cost: {
2777
+ input: number;
2778
+ output: number;
2779
+ cacheRead: number;
2780
+ cacheWrite: number;
2781
+ };
2782
+ contextWindow: number;
2783
+ maxTokens: number;
2784
+ };
2785
+ readonly "@cf/meta/llama-4-scout-17b-16e-instruct": {
2786
+ id: string;
2787
+ name: string;
2788
+ api: "openai-completions";
2789
+ provider: string;
2790
+ baseUrl: string;
2791
+ compat: {
2792
+ sendSessionAffinityHeaders: true;
2793
+ };
2794
+ reasoning: false;
2795
+ input: ("image" | "text")[];
2796
+ cost: {
2797
+ input: number;
2798
+ output: number;
2799
+ cacheRead: number;
2800
+ cacheWrite: number;
2801
+ };
2802
+ contextWindow: number;
2803
+ maxTokens: number;
2804
+ };
2805
+ readonly "@cf/moonshotai/kimi-k2.5": {
2806
+ id: string;
2807
+ name: string;
2808
+ api: "openai-completions";
2809
+ provider: string;
2810
+ baseUrl: string;
2811
+ compat: {
2812
+ sendSessionAffinityHeaders: true;
2813
+ };
2814
+ reasoning: true;
2815
+ input: ("image" | "text")[];
2816
+ cost: {
2817
+ input: number;
2818
+ output: number;
2819
+ cacheRead: number;
2820
+ cacheWrite: number;
2821
+ };
2822
+ contextWindow: number;
2823
+ maxTokens: number;
2824
+ };
2825
+ readonly "@cf/moonshotai/kimi-k2.6": {
2826
+ id: string;
2827
+ name: string;
2828
+ api: "openai-completions";
2829
+ provider: string;
2830
+ baseUrl: string;
2831
+ compat: {
2832
+ sendSessionAffinityHeaders: true;
2833
+ };
2834
+ reasoning: true;
2835
+ input: ("image" | "text")[];
2836
+ cost: {
2837
+ input: number;
2838
+ output: number;
2839
+ cacheRead: number;
2840
+ cacheWrite: number;
2841
+ };
2842
+ contextWindow: number;
2843
+ maxTokens: number;
2844
+ };
2845
+ readonly "@cf/nvidia/nemotron-3-120b-a12b": {
2846
+ id: string;
2847
+ name: string;
2848
+ api: "openai-completions";
2849
+ provider: string;
2850
+ baseUrl: string;
2851
+ compat: {
2852
+ sendSessionAffinityHeaders: true;
2853
+ };
2854
+ reasoning: true;
2855
+ input: "text"[];
2856
+ cost: {
2857
+ input: number;
2858
+ output: number;
2859
+ cacheRead: number;
2860
+ cacheWrite: number;
2861
+ };
2862
+ contextWindow: number;
2863
+ maxTokens: number;
2864
+ };
2865
+ readonly "@cf/openai/gpt-oss-120b": {
2866
+ id: string;
2867
+ name: string;
2868
+ api: "openai-completions";
2869
+ provider: string;
2870
+ baseUrl: string;
2871
+ compat: {
2872
+ sendSessionAffinityHeaders: true;
2873
+ };
2874
+ reasoning: true;
2875
+ input: "text"[];
2876
+ cost: {
2877
+ input: number;
2878
+ output: number;
2879
+ cacheRead: number;
2880
+ cacheWrite: number;
2881
+ };
2882
+ contextWindow: number;
2883
+ maxTokens: number;
2884
+ };
2885
+ readonly "@cf/openai/gpt-oss-20b": {
2886
+ id: string;
2887
+ name: string;
2888
+ api: "openai-completions";
2889
+ provider: string;
2890
+ baseUrl: string;
2891
+ compat: {
2892
+ sendSessionAffinityHeaders: true;
2893
+ };
2894
+ reasoning: true;
2895
+ input: "text"[];
2896
+ cost: {
2897
+ input: number;
2898
+ output: number;
2899
+ cacheRead: number;
2900
+ cacheWrite: number;
2901
+ };
2902
+ contextWindow: number;
2903
+ maxTokens: number;
2904
+ };
2905
+ readonly "@cf/zai-org/glm-4.7-flash": {
2906
+ id: string;
2907
+ name: string;
2908
+ api: "openai-completions";
2909
+ provider: string;
2910
+ baseUrl: string;
2911
+ compat: {
2912
+ sendSessionAffinityHeaders: true;
2913
+ };
2914
+ reasoning: true;
2915
+ input: "text"[];
2916
+ cost: {
2917
+ input: number;
2918
+ output: number;
2919
+ cacheRead: number;
2920
+ cacheWrite: number;
2921
+ };
2922
+ contextWindow: number;
2923
+ maxTokens: number;
2924
+ };
2925
+ };
2747
2926
  readonly deepseek: {
2748
2927
  readonly "deepseek-v4-flash": {
2749
2928
  id: string;
@@ -2837,6 +3016,23 @@ export declare const MODELS: {
2837
3016
  contextWindow: number;
2838
3017
  maxTokens: number;
2839
3018
  };
3019
+ readonly "accounts/fireworks/models/deepseek-v4-pro": {
3020
+ id: string;
3021
+ name: string;
3022
+ api: "anthropic-messages";
3023
+ provider: string;
3024
+ baseUrl: string;
3025
+ reasoning: true;
3026
+ input: "text"[];
3027
+ cost: {
3028
+ input: number;
3029
+ output: number;
3030
+ cacheRead: number;
3031
+ cacheWrite: number;
3032
+ };
3033
+ contextWindow: number;
3034
+ maxTokens: number;
3035
+ };
2840
3036
  readonly "accounts/fireworks/models/glm-4p5": {
2841
3037
  id: string;
2842
3038
  name: string;
@@ -5263,6 +5459,26 @@ export declare const MODELS: {
5263
5459
  contextWindow: number;
5264
5460
  maxTokens: number;
5265
5461
  };
5462
+ readonly "deepseek-ai/DeepSeek-V4-Pro": {
5463
+ id: string;
5464
+ name: string;
5465
+ api: "openai-completions";
5466
+ provider: string;
5467
+ baseUrl: string;
5468
+ compat: {
5469
+ supportsDeveloperRole: false;
5470
+ };
5471
+ reasoning: true;
5472
+ input: "text"[];
5473
+ cost: {
5474
+ input: number;
5475
+ output: number;
5476
+ cacheRead: number;
5477
+ cacheWrite: number;
5478
+ };
5479
+ contextWindow: number;
5480
+ maxTokens: number;
5481
+ };
5266
5482
  readonly "moonshotai/Kimi-K2-Instruct": {
5267
5483
  id: string;
5268
5484
  name: string;
@@ -6584,6 +6800,23 @@ export declare const MODELS: {
6584
6800
  contextWindow: number;
6585
6801
  maxTokens: number;
6586
6802
  };
6803
+ readonly "gpt-5.5-pro": {
6804
+ id: string;
6805
+ name: string;
6806
+ api: "openai-responses";
6807
+ provider: string;
6808
+ baseUrl: string;
6809
+ reasoning: true;
6810
+ input: ("image" | "text")[];
6811
+ cost: {
6812
+ input: number;
6813
+ output: number;
6814
+ cacheRead: number;
6815
+ cacheWrite: number;
6816
+ };
6817
+ contextWindow: number;
6818
+ maxTokens: number;
6819
+ };
6587
6820
  readonly o1: {
6588
6821
  id: string;
6589
6822
  name: string;
@@ -6911,23 +7144,6 @@ export declare const MODELS: {
6911
7144
  contextWindow: number;
6912
7145
  maxTokens: number;
6913
7146
  };
6914
- readonly "claude-3-5-haiku": {
6915
- id: string;
6916
- name: string;
6917
- api: "anthropic-messages";
6918
- provider: string;
6919
- baseUrl: string;
6920
- reasoning: false;
6921
- input: ("image" | "text")[];
6922
- cost: {
6923
- input: number;
6924
- output: number;
6925
- cacheRead: number;
6926
- cacheWrite: number;
6927
- };
6928
- contextWindow: number;
6929
- maxTokens: number;
6930
- };
6931
7147
  readonly "claude-haiku-4-5": {
6932
7148
  id: string;
6933
7149
  name: string;
@@ -7455,23 +7671,6 @@ export declare const MODELS: {
7455
7671
  contextWindow: number;
7456
7672
  maxTokens: number;
7457
7673
  };
7458
- readonly "ling-2.6-flash-free": {
7459
- id: string;
7460
- name: string;
7461
- api: "openai-completions";
7462
- provider: string;
7463
- baseUrl: string;
7464
- reasoning: false;
7465
- input: "text"[];
7466
- cost: {
7467
- input: number;
7468
- output: number;
7469
- cacheRead: number;
7470
- cacheWrite: number;
7471
- };
7472
- contextWindow: number;
7473
- maxTokens: number;
7474
- };
7475
7674
  readonly "minimax-m2.5": {
7476
7675
  id: string;
7477
7676
  name: string;
@@ -8950,7 +9149,7 @@ export declare const MODELS: {
8950
9149
  contextWindow: number;
8951
9150
  maxTokens: number;
8952
9151
  };
8953
- readonly "inclusionai/ling-2.6-flash:free": {
9152
+ readonly "inclusionai/ling-2.6-flash": {
8954
9153
  id: string;
8955
9154
  name: string;
8956
9155
  api: "openai-completions";
@@ -9477,7 +9676,7 @@ export declare const MODELS: {
9477
9676
  contextWindow: number;
9478
9677
  maxTokens: number;
9479
9678
  };
9480
- readonly "mistralai/mistral-small-creative": {
9679
+ readonly "mistralai/mixtral-8x22b-instruct": {
9481
9680
  id: string;
9482
9681
  name: string;
9483
9682
  api: "openai-completions";
@@ -9494,7 +9693,7 @@ export declare const MODELS: {
9494
9693
  contextWindow: number;
9495
9694
  maxTokens: number;
9496
9695
  };
9497
- readonly "mistralai/mixtral-8x22b-instruct": {
9696
+ readonly "mistralai/mixtral-8x7b-instruct": {
9498
9697
  id: string;
9499
9698
  name: string;
9500
9699
  api: "openai-completions";
@@ -9511,14 +9710,14 @@ export declare const MODELS: {
9511
9710
  contextWindow: number;
9512
9711
  maxTokens: number;
9513
9712
  };
9514
- readonly "mistralai/mixtral-8x7b-instruct": {
9713
+ readonly "mistralai/pixtral-large-2411": {
9515
9714
  id: string;
9516
9715
  name: string;
9517
9716
  api: "openai-completions";
9518
9717
  provider: string;
9519
9718
  baseUrl: string;
9520
9719
  reasoning: false;
9521
- input: "text"[];
9720
+ input: ("image" | "text")[];
9522
9721
  cost: {
9523
9722
  input: number;
9524
9723
  output: number;
@@ -9528,24 +9727,7 @@ export declare const MODELS: {
9528
9727
  contextWindow: number;
9529
9728
  maxTokens: number;
9530
9729
  };
9531
- readonly "mistralai/pixtral-large-2411": {
9532
- id: string;
9533
- name: string;
9534
- api: "openai-completions";
9535
- provider: string;
9536
- baseUrl: string;
9537
- reasoning: false;
9538
- input: ("image" | "text")[];
9539
- cost: {
9540
- input: number;
9541
- output: number;
9542
- cacheRead: number;
9543
- cacheWrite: number;
9544
- };
9545
- contextWindow: number;
9546
- maxTokens: number;
9547
- };
9548
- readonly "mistralai/voxtral-small-24b-2507": {
9730
+ readonly "mistralai/voxtral-small-24b-2507": {
9549
9731
  id: string;
9550
9732
  name: string;
9551
9733
  api: "openai-completions";
@@ -9732,6 +9914,23 @@ export declare const MODELS: {
9732
9914
  contextWindow: number;
9733
9915
  maxTokens: number;
9734
9916
  };
9917
+ readonly "nvidia/nemotron-3-nano-omni-30b-a3b-reasoning:free": {
9918
+ id: string;
9919
+ name: string;
9920
+ api: "openai-completions";
9921
+ provider: string;
9922
+ baseUrl: string;
9923
+ reasoning: true;
9924
+ input: ("image" | "text")[];
9925
+ cost: {
9926
+ input: number;
9927
+ output: number;
9928
+ cacheRead: number;
9929
+ cacheWrite: number;
9930
+ };
9931
+ contextWindow: number;
9932
+ maxTokens: number;
9933
+ };
9735
9934
  readonly "nvidia/nemotron-3-super-120b-a12b": {
9736
9935
  id: string;
9737
9936
  name: string;
@@ -10803,6 +11002,40 @@ export declare const MODELS: {
10803
11002
  contextWindow: number;
10804
11003
  maxTokens: number;
10805
11004
  };
11005
+ readonly "poolside/laguna-m.1:free": {
11006
+ id: string;
11007
+ name: string;
11008
+ api: "openai-completions";
11009
+ provider: string;
11010
+ baseUrl: string;
11011
+ reasoning: true;
11012
+ input: "text"[];
11013
+ cost: {
11014
+ input: number;
11015
+ output: number;
11016
+ cacheRead: number;
11017
+ cacheWrite: number;
11018
+ };
11019
+ contextWindow: number;
11020
+ maxTokens: number;
11021
+ };
11022
+ readonly "poolside/laguna-xs.2:free": {
11023
+ id: string;
11024
+ name: string;
11025
+ api: "openai-completions";
11026
+ provider: string;
11027
+ baseUrl: string;
11028
+ reasoning: true;
11029
+ input: "text"[];
11030
+ cost: {
11031
+ input: number;
11032
+ output: number;
11033
+ cacheRead: number;
11034
+ cacheWrite: number;
11035
+ };
11036
+ contextWindow: number;
11037
+ maxTokens: number;
11038
+ };
10806
11039
  readonly "prime-intellect/intellect-3": {
10807
11040
  id: string;
10808
11041
  name: string;
@@ -11534,7 +11767,41 @@ export declare const MODELS: {
11534
11767
  contextWindow: number;
11535
11768
  maxTokens: number;
11536
11769
  };
11537
- readonly "qwen/qwen3.6-plus": {
11770
+ readonly "qwen/qwen3.5-plus-20260420": {
11771
+ id: string;
11772
+ name: string;
11773
+ api: "openai-completions";
11774
+ provider: string;
11775
+ baseUrl: string;
11776
+ reasoning: true;
11777
+ input: ("image" | "text")[];
11778
+ cost: {
11779
+ input: number;
11780
+ output: number;
11781
+ cacheRead: number;
11782
+ cacheWrite: number;
11783
+ };
11784
+ contextWindow: number;
11785
+ maxTokens: number;
11786
+ };
11787
+ readonly "qwen/qwen3.6-27b": {
11788
+ id: string;
11789
+ name: string;
11790
+ api: "openai-completions";
11791
+ provider: string;
11792
+ baseUrl: string;
11793
+ reasoning: true;
11794
+ input: ("image" | "text")[];
11795
+ cost: {
11796
+ input: number;
11797
+ output: number;
11798
+ cacheRead: number;
11799
+ cacheWrite: number;
11800
+ };
11801
+ contextWindow: number;
11802
+ maxTokens: number;
11803
+ };
11804
+ readonly "qwen/qwen3.6-flash": {
11538
11805
  id: string;
11539
11806
  name: string;
11540
11807
  api: "openai-completions";
@@ -11551,7 +11818,7 @@ export declare const MODELS: {
11551
11818
  contextWindow: number;
11552
11819
  maxTokens: number;
11553
11820
  };
11554
- readonly "qwen/qwq-32b": {
11821
+ readonly "qwen/qwen3.6-max-preview": {
11555
11822
  id: string;
11556
11823
  name: string;
11557
11824
  api: "openai-completions";
@@ -11568,6 +11835,23 @@ export declare const MODELS: {
11568
11835
  contextWindow: number;
11569
11836
  maxTokens: number;
11570
11837
  };
11838
+ readonly "qwen/qwen3.6-plus": {
11839
+ id: string;
11840
+ name: string;
11841
+ api: "openai-completions";
11842
+ provider: string;
11843
+ baseUrl: string;
11844
+ reasoning: true;
11845
+ input: ("image" | "text")[];
11846
+ cost: {
11847
+ input: number;
11848
+ output: number;
11849
+ cacheRead: number;
11850
+ cacheWrite: number;
11851
+ };
11852
+ contextWindow: number;
11853
+ maxTokens: number;
11854
+ };
11571
11855
  readonly "rekaai/reka-edge": {
11572
11856
  id: string;
11573
11857
  name: string;
@@ -12197,6 +12481,23 @@ export declare const MODELS: {
12197
12481
  contextWindow: number;
12198
12482
  maxTokens: number;
12199
12483
  };
12484
+ readonly "~anthropic/claude-haiku-latest": {
12485
+ id: string;
12486
+ name: string;
12487
+ api: "openai-completions";
12488
+ provider: string;
12489
+ baseUrl: string;
12490
+ reasoning: true;
12491
+ input: ("image" | "text")[];
12492
+ cost: {
12493
+ input: number;
12494
+ output: number;
12495
+ cacheRead: number;
12496
+ cacheWrite: number;
12497
+ };
12498
+ contextWindow: number;
12499
+ maxTokens: number;
12500
+ };
12200
12501
  readonly "~anthropic/claude-opus-latest": {
12201
12502
  id: string;
12202
12503
  name: string;
@@ -12214,6 +12515,108 @@ export declare const MODELS: {
12214
12515
  contextWindow: number;
12215
12516
  maxTokens: number;
12216
12517
  };
12518
+ readonly "~anthropic/claude-sonnet-latest": {
12519
+ id: string;
12520
+ name: string;
12521
+ api: "openai-completions";
12522
+ provider: string;
12523
+ baseUrl: string;
12524
+ reasoning: true;
12525
+ input: ("image" | "text")[];
12526
+ cost: {
12527
+ input: number;
12528
+ output: number;
12529
+ cacheRead: number;
12530
+ cacheWrite: number;
12531
+ };
12532
+ contextWindow: number;
12533
+ maxTokens: number;
12534
+ };
12535
+ readonly "~google/gemini-flash-latest": {
12536
+ id: string;
12537
+ name: string;
12538
+ api: "openai-completions";
12539
+ provider: string;
12540
+ baseUrl: string;
12541
+ reasoning: true;
12542
+ input: ("image" | "text")[];
12543
+ cost: {
12544
+ input: number;
12545
+ output: number;
12546
+ cacheRead: number;
12547
+ cacheWrite: number;
12548
+ };
12549
+ contextWindow: number;
12550
+ maxTokens: number;
12551
+ };
12552
+ readonly "~google/gemini-pro-latest": {
12553
+ id: string;
12554
+ name: string;
12555
+ api: "openai-completions";
12556
+ provider: string;
12557
+ baseUrl: string;
12558
+ reasoning: true;
12559
+ input: ("image" | "text")[];
12560
+ cost: {
12561
+ input: number;
12562
+ output: number;
12563
+ cacheRead: number;
12564
+ cacheWrite: number;
12565
+ };
12566
+ contextWindow: number;
12567
+ maxTokens: number;
12568
+ };
12569
+ readonly "~moonshotai/kimi-latest": {
12570
+ id: string;
12571
+ name: string;
12572
+ api: "openai-completions";
12573
+ provider: string;
12574
+ baseUrl: string;
12575
+ reasoning: true;
12576
+ input: ("image" | "text")[];
12577
+ cost: {
12578
+ input: number;
12579
+ output: number;
12580
+ cacheRead: number;
12581
+ cacheWrite: number;
12582
+ };
12583
+ contextWindow: number;
12584
+ maxTokens: number;
12585
+ };
12586
+ readonly "~openai/gpt-latest": {
12587
+ id: string;
12588
+ name: string;
12589
+ api: "openai-completions";
12590
+ provider: string;
12591
+ baseUrl: string;
12592
+ reasoning: true;
12593
+ input: ("image" | "text")[];
12594
+ cost: {
12595
+ input: number;
12596
+ output: number;
12597
+ cacheRead: number;
12598
+ cacheWrite: number;
12599
+ };
12600
+ contextWindow: number;
12601
+ maxTokens: number;
12602
+ };
12603
+ readonly "~openai/gpt-mini-latest": {
12604
+ id: string;
12605
+ name: string;
12606
+ api: "openai-completions";
12607
+ provider: string;
12608
+ baseUrl: string;
12609
+ reasoning: true;
12610
+ input: ("image" | "text")[];
12611
+ cost: {
12612
+ input: number;
12613
+ output: number;
12614
+ cacheRead: number;
12615
+ cacheWrite: number;
12616
+ };
12617
+ contextWindow: number;
12618
+ maxTokens: number;
12619
+ };
12217
12620
  };
12218
12621
  readonly "vercel-ai-gateway": {
12219
12622
  readonly "alibaba/qwen-3-14b": {
@@ -12488,6 +12891,23 @@ export declare const MODELS: {
12488
12891
  contextWindow: number;
12489
12892
  maxTokens: number;
12490
12893
  };
12894
+ readonly "alibaba/qwen3.6-27b": {
12895
+ id: string;
12896
+ name: string;
12897
+ api: "anthropic-messages";
12898
+ provider: string;
12899
+ baseUrl: string;
12900
+ reasoning: true;
12901
+ input: ("image" | "text")[];
12902
+ cost: {
12903
+ input: number;
12904
+ output: number;
12905
+ cacheRead: number;
12906
+ cacheWrite: number;
12907
+ };
12908
+ contextWindow: number;
12909
+ maxTokens: number;
12910
+ };
12491
12911
  readonly "alibaba/qwen3.6-plus": {
12492
12912
  id: string;
12493
12913
  name: string;
@@ -12868,7 +13288,7 @@ export declare const MODELS: {
12868
13288
  api: "anthropic-messages";
12869
13289
  provider: string;
12870
13290
  baseUrl: string;
12871
- reasoning: true;
13291
+ reasoning: false;
12872
13292
  input: "text"[];
12873
13293
  cost: {
12874
13294
  input: number;
@@ -14715,6 +15135,40 @@ export declare const MODELS: {
14715
15135
  contextWindow: number;
14716
15136
  maxTokens: number;
14717
15137
  };
15138
+ readonly "xiaomi/mimo-v2.5": {
15139
+ id: string;
15140
+ name: string;
15141
+ api: "anthropic-messages";
15142
+ provider: string;
15143
+ baseUrl: string;
15144
+ reasoning: true;
15145
+ input: ("image" | "text")[];
15146
+ cost: {
15147
+ input: number;
15148
+ output: number;
15149
+ cacheRead: number;
15150
+ cacheWrite: number;
15151
+ };
15152
+ contextWindow: number;
15153
+ maxTokens: number;
15154
+ };
15155
+ readonly "xiaomi/mimo-v2.5-pro": {
15156
+ id: string;
15157
+ name: string;
15158
+ api: "anthropic-messages";
15159
+ provider: string;
15160
+ baseUrl: string;
15161
+ reasoning: true;
15162
+ input: ("image" | "text")[];
15163
+ cost: {
15164
+ input: number;
15165
+ output: number;
15166
+ cacheRead: number;
15167
+ cacheWrite: number;
15168
+ };
15169
+ contextWindow: number;
15170
+ maxTokens: number;
15171
+ };
14718
15172
  readonly "zai/glm-4.5": {
14719
15173
  id: string;
14720
15174
  name: string;
@@ -15435,6 +15889,28 @@ export declare const MODELS: {
15435
15889
  contextWindow: number;
15436
15890
  maxTokens: number;
15437
15891
  };
15892
+ readonly "glm-5v-turbo": {
15893
+ id: string;
15894
+ name: string;
15895
+ api: "openai-completions";
15896
+ provider: string;
15897
+ baseUrl: string;
15898
+ compat: {
15899
+ supportsDeveloperRole: false;
15900
+ thinkingFormat: "zai";
15901
+ zaiToolStream: true;
15902
+ };
15903
+ reasoning: true;
15904
+ input: ("image" | "text")[];
15905
+ cost: {
15906
+ input: number;
15907
+ output: number;
15908
+ cacheRead: number;
15909
+ cacheWrite: number;
15910
+ };
15911
+ contextWindow: number;
15912
+ maxTokens: number;
15913
+ };
15438
15914
  };
15439
15915
  };
15440
15916
  //# sourceMappingURL=models.generated.d.ts.map