@hyperspaceng/neural-ai 0.70.3 → 0.70.6

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -2537,6 +2537,23 @@ export declare const MODELS: {
2537
2537
  contextWindow: number;
2538
2538
  maxTokens: number;
2539
2539
  };
2540
+ readonly "gpt-5.5-pro": {
2541
+ id: string;
2542
+ name: string;
2543
+ api: "azure-openai-responses";
2544
+ provider: string;
2545
+ baseUrl: string;
2546
+ reasoning: true;
2547
+ input: ("image" | "text")[];
2548
+ cost: {
2549
+ input: number;
2550
+ output: number;
2551
+ cacheRead: number;
2552
+ cacheWrite: number;
2553
+ };
2554
+ contextWindow: number;
2555
+ maxTokens: number;
2556
+ };
2540
2557
  readonly o1: {
2541
2558
  id: string;
2542
2559
  name: string;
@@ -2744,6 +2761,168 @@ export declare const MODELS: {
2744
2761
  maxTokens: number;
2745
2762
  };
2746
2763
  };
2764
+ readonly "cloudflare-workers-ai": {
2765
+ readonly "@cf/google/gemma-4-26b-a4b-it": {
2766
+ id: string;
2767
+ name: string;
2768
+ api: "openai-completions";
2769
+ provider: string;
2770
+ baseUrl: string;
2771
+ compat: {
2772
+ sendSessionAffinityHeaders: true;
2773
+ };
2774
+ reasoning: true;
2775
+ input: ("image" | "text")[];
2776
+ cost: {
2777
+ input: number;
2778
+ output: number;
2779
+ cacheRead: number;
2780
+ cacheWrite: number;
2781
+ };
2782
+ contextWindow: number;
2783
+ maxTokens: number;
2784
+ };
2785
+ readonly "@cf/meta/llama-4-scout-17b-16e-instruct": {
2786
+ id: string;
2787
+ name: string;
2788
+ api: "openai-completions";
2789
+ provider: string;
2790
+ baseUrl: string;
2791
+ compat: {
2792
+ sendSessionAffinityHeaders: true;
2793
+ };
2794
+ reasoning: false;
2795
+ input: ("image" | "text")[];
2796
+ cost: {
2797
+ input: number;
2798
+ output: number;
2799
+ cacheRead: number;
2800
+ cacheWrite: number;
2801
+ };
2802
+ contextWindow: number;
2803
+ maxTokens: number;
2804
+ };
2805
+ readonly "@cf/moonshotai/kimi-k2.5": {
2806
+ id: string;
2807
+ name: string;
2808
+ api: "openai-completions";
2809
+ provider: string;
2810
+ baseUrl: string;
2811
+ compat: {
2812
+ sendSessionAffinityHeaders: true;
2813
+ };
2814
+ reasoning: true;
2815
+ input: ("image" | "text")[];
2816
+ cost: {
2817
+ input: number;
2818
+ output: number;
2819
+ cacheRead: number;
2820
+ cacheWrite: number;
2821
+ };
2822
+ contextWindow: number;
2823
+ maxTokens: number;
2824
+ };
2825
+ readonly "@cf/moonshotai/kimi-k2.6": {
2826
+ id: string;
2827
+ name: string;
2828
+ api: "openai-completions";
2829
+ provider: string;
2830
+ baseUrl: string;
2831
+ compat: {
2832
+ sendSessionAffinityHeaders: true;
2833
+ };
2834
+ reasoning: true;
2835
+ input: ("image" | "text")[];
2836
+ cost: {
2837
+ input: number;
2838
+ output: number;
2839
+ cacheRead: number;
2840
+ cacheWrite: number;
2841
+ };
2842
+ contextWindow: number;
2843
+ maxTokens: number;
2844
+ };
2845
+ readonly "@cf/nvidia/nemotron-3-120b-a12b": {
2846
+ id: string;
2847
+ name: string;
2848
+ api: "openai-completions";
2849
+ provider: string;
2850
+ baseUrl: string;
2851
+ compat: {
2852
+ sendSessionAffinityHeaders: true;
2853
+ };
2854
+ reasoning: true;
2855
+ input: "text"[];
2856
+ cost: {
2857
+ input: number;
2858
+ output: number;
2859
+ cacheRead: number;
2860
+ cacheWrite: number;
2861
+ };
2862
+ contextWindow: number;
2863
+ maxTokens: number;
2864
+ };
2865
+ readonly "@cf/openai/gpt-oss-120b": {
2866
+ id: string;
2867
+ name: string;
2868
+ api: "openai-completions";
2869
+ provider: string;
2870
+ baseUrl: string;
2871
+ compat: {
2872
+ sendSessionAffinityHeaders: true;
2873
+ };
2874
+ reasoning: true;
2875
+ input: "text"[];
2876
+ cost: {
2877
+ input: number;
2878
+ output: number;
2879
+ cacheRead: number;
2880
+ cacheWrite: number;
2881
+ };
2882
+ contextWindow: number;
2883
+ maxTokens: number;
2884
+ };
2885
+ readonly "@cf/openai/gpt-oss-20b": {
2886
+ id: string;
2887
+ name: string;
2888
+ api: "openai-completions";
2889
+ provider: string;
2890
+ baseUrl: string;
2891
+ compat: {
2892
+ sendSessionAffinityHeaders: true;
2893
+ };
2894
+ reasoning: true;
2895
+ input: "text"[];
2896
+ cost: {
2897
+ input: number;
2898
+ output: number;
2899
+ cacheRead: number;
2900
+ cacheWrite: number;
2901
+ };
2902
+ contextWindow: number;
2903
+ maxTokens: number;
2904
+ };
2905
+ readonly "@cf/zai-org/glm-4.7-flash": {
2906
+ id: string;
2907
+ name: string;
2908
+ api: "openai-completions";
2909
+ provider: string;
2910
+ baseUrl: string;
2911
+ compat: {
2912
+ sendSessionAffinityHeaders: true;
2913
+ };
2914
+ reasoning: true;
2915
+ input: "text"[];
2916
+ cost: {
2917
+ input: number;
2918
+ output: number;
2919
+ cacheRead: number;
2920
+ cacheWrite: number;
2921
+ };
2922
+ contextWindow: number;
2923
+ maxTokens: number;
2924
+ };
2925
+ };
2747
2926
  readonly deepseek: {
2748
2927
  readonly "deepseek-v4-flash": {
2749
2928
  id: string;
@@ -2837,6 +3016,23 @@ export declare const MODELS: {
2837
3016
  contextWindow: number;
2838
3017
  maxTokens: number;
2839
3018
  };
3019
+ readonly "accounts/fireworks/models/deepseek-v4-pro": {
3020
+ id: string;
3021
+ name: string;
3022
+ api: "anthropic-messages";
3023
+ provider: string;
3024
+ baseUrl: string;
3025
+ reasoning: true;
3026
+ input: "text"[];
3027
+ cost: {
3028
+ input: number;
3029
+ output: number;
3030
+ cacheRead: number;
3031
+ cacheWrite: number;
3032
+ };
3033
+ contextWindow: number;
3034
+ maxTokens: number;
3035
+ };
2840
3036
  readonly "accounts/fireworks/models/glm-4p5": {
2841
3037
  id: string;
2842
3038
  name: string;
@@ -5263,6 +5459,26 @@ export declare const MODELS: {
5263
5459
  contextWindow: number;
5264
5460
  maxTokens: number;
5265
5461
  };
5462
+ readonly "deepseek-ai/DeepSeek-V4-Pro": {
5463
+ id: string;
5464
+ name: string;
5465
+ api: "openai-completions";
5466
+ provider: string;
5467
+ baseUrl: string;
5468
+ compat: {
5469
+ supportsDeveloperRole: false;
5470
+ };
5471
+ reasoning: true;
5472
+ input: "text"[];
5473
+ cost: {
5474
+ input: number;
5475
+ output: number;
5476
+ cacheRead: number;
5477
+ cacheWrite: number;
5478
+ };
5479
+ contextWindow: number;
5480
+ maxTokens: number;
5481
+ };
5266
5482
  readonly "moonshotai/Kimi-K2-Instruct": {
5267
5483
  id: string;
5268
5484
  name: string;
@@ -6584,6 +6800,23 @@ export declare const MODELS: {
6584
6800
  contextWindow: number;
6585
6801
  maxTokens: number;
6586
6802
  };
6803
+ readonly "gpt-5.5-pro": {
6804
+ id: string;
6805
+ name: string;
6806
+ api: "openai-responses";
6807
+ provider: string;
6808
+ baseUrl: string;
6809
+ reasoning: true;
6810
+ input: ("image" | "text")[];
6811
+ cost: {
6812
+ input: number;
6813
+ output: number;
6814
+ cacheRead: number;
6815
+ cacheWrite: number;
6816
+ };
6817
+ contextWindow: number;
6818
+ maxTokens: number;
6819
+ };
6587
6820
  readonly o1: {
6588
6821
  id: string;
6589
6822
  name: string;
@@ -6911,23 +7144,6 @@ export declare const MODELS: {
6911
7144
  contextWindow: number;
6912
7145
  maxTokens: number;
6913
7146
  };
6914
- readonly "claude-3-5-haiku": {
6915
- id: string;
6916
- name: string;
6917
- api: "anthropic-messages";
6918
- provider: string;
6919
- baseUrl: string;
6920
- reasoning: false;
6921
- input: ("image" | "text")[];
6922
- cost: {
6923
- input: number;
6924
- output: number;
6925
- cacheRead: number;
6926
- cacheWrite: number;
6927
- };
6928
- contextWindow: number;
6929
- maxTokens: number;
6930
- };
6931
7147
  readonly "claude-haiku-4-5": {
6932
7148
  id: string;
6933
7149
  name: string;
@@ -8950,7 +9166,7 @@ export declare const MODELS: {
8950
9166
  contextWindow: number;
8951
9167
  maxTokens: number;
8952
9168
  };
8953
- readonly "inclusionai/ling-2.6-flash:free": {
9169
+ readonly "inclusionai/ling-2.6-flash": {
8954
9170
  id: string;
8955
9171
  name: string;
8956
9172
  api: "openai-completions";
@@ -9732,6 +9948,23 @@ export declare const MODELS: {
9732
9948
  contextWindow: number;
9733
9949
  maxTokens: number;
9734
9950
  };
9951
+ readonly "nvidia/nemotron-3-nano-omni-30b-a3b-reasoning:free": {
9952
+ id: string;
9953
+ name: string;
9954
+ api: "openai-completions";
9955
+ provider: string;
9956
+ baseUrl: string;
9957
+ reasoning: true;
9958
+ input: ("image" | "text")[];
9959
+ cost: {
9960
+ input: number;
9961
+ output: number;
9962
+ cacheRead: number;
9963
+ cacheWrite: number;
9964
+ };
9965
+ contextWindow: number;
9966
+ maxTokens: number;
9967
+ };
9735
9968
  readonly "nvidia/nemotron-3-super-120b-a12b": {
9736
9969
  id: string;
9737
9970
  name: string;
@@ -10803,6 +11036,40 @@ export declare const MODELS: {
10803
11036
  contextWindow: number;
10804
11037
  maxTokens: number;
10805
11038
  };
11039
+ readonly "poolside/laguna-m.1:free": {
11040
+ id: string;
11041
+ name: string;
11042
+ api: "openai-completions";
11043
+ provider: string;
11044
+ baseUrl: string;
11045
+ reasoning: true;
11046
+ input: "text"[];
11047
+ cost: {
11048
+ input: number;
11049
+ output: number;
11050
+ cacheRead: number;
11051
+ cacheWrite: number;
11052
+ };
11053
+ contextWindow: number;
11054
+ maxTokens: number;
11055
+ };
11056
+ readonly "poolside/laguna-xs.2:free": {
11057
+ id: string;
11058
+ name: string;
11059
+ api: "openai-completions";
11060
+ provider: string;
11061
+ baseUrl: string;
11062
+ reasoning: true;
11063
+ input: "text"[];
11064
+ cost: {
11065
+ input: number;
11066
+ output: number;
11067
+ cacheRead: number;
11068
+ cacheWrite: number;
11069
+ };
11070
+ contextWindow: number;
11071
+ maxTokens: number;
11072
+ };
10806
11073
  readonly "prime-intellect/intellect-3": {
10807
11074
  id: string;
10808
11075
  name: string;
@@ -11534,7 +11801,7 @@ export declare const MODELS: {
11534
11801
  contextWindow: number;
11535
11802
  maxTokens: number;
11536
11803
  };
11537
- readonly "qwen/qwen3.6-plus": {
11804
+ readonly "qwen/qwen3.5-plus-20260420": {
11538
11805
  id: string;
11539
11806
  name: string;
11540
11807
  api: "openai-completions";
@@ -11551,7 +11818,41 @@ export declare const MODELS: {
11551
11818
  contextWindow: number;
11552
11819
  maxTokens: number;
11553
11820
  };
11554
- readonly "qwen/qwq-32b": {
11821
+ readonly "qwen/qwen3.6-27b": {
11822
+ id: string;
11823
+ name: string;
11824
+ api: "openai-completions";
11825
+ provider: string;
11826
+ baseUrl: string;
11827
+ reasoning: true;
11828
+ input: ("image" | "text")[];
11829
+ cost: {
11830
+ input: number;
11831
+ output: number;
11832
+ cacheRead: number;
11833
+ cacheWrite: number;
11834
+ };
11835
+ contextWindow: number;
11836
+ maxTokens: number;
11837
+ };
11838
+ readonly "qwen/qwen3.6-flash": {
11839
+ id: string;
11840
+ name: string;
11841
+ api: "openai-completions";
11842
+ provider: string;
11843
+ baseUrl: string;
11844
+ reasoning: true;
11845
+ input: ("image" | "text")[];
11846
+ cost: {
11847
+ input: number;
11848
+ output: number;
11849
+ cacheRead: number;
11850
+ cacheWrite: number;
11851
+ };
11852
+ contextWindow: number;
11853
+ maxTokens: number;
11854
+ };
11855
+ readonly "qwen/qwen3.6-max-preview": {
11555
11856
  id: string;
11556
11857
  name: string;
11557
11858
  api: "openai-completions";
@@ -11568,6 +11869,23 @@ export declare const MODELS: {
11568
11869
  contextWindow: number;
11569
11870
  maxTokens: number;
11570
11871
  };
11872
+ readonly "qwen/qwen3.6-plus": {
11873
+ id: string;
11874
+ name: string;
11875
+ api: "openai-completions";
11876
+ provider: string;
11877
+ baseUrl: string;
11878
+ reasoning: true;
11879
+ input: ("image" | "text")[];
11880
+ cost: {
11881
+ input: number;
11882
+ output: number;
11883
+ cacheRead: number;
11884
+ cacheWrite: number;
11885
+ };
11886
+ contextWindow: number;
11887
+ maxTokens: number;
11888
+ };
11571
11889
  readonly "rekaai/reka-edge": {
11572
11890
  id: string;
11573
11891
  name: string;
@@ -12197,6 +12515,23 @@ export declare const MODELS: {
12197
12515
  contextWindow: number;
12198
12516
  maxTokens: number;
12199
12517
  };
12518
+ readonly "~anthropic/claude-haiku-latest": {
12519
+ id: string;
12520
+ name: string;
12521
+ api: "openai-completions";
12522
+ provider: string;
12523
+ baseUrl: string;
12524
+ reasoning: true;
12525
+ input: ("image" | "text")[];
12526
+ cost: {
12527
+ input: number;
12528
+ output: number;
12529
+ cacheRead: number;
12530
+ cacheWrite: number;
12531
+ };
12532
+ contextWindow: number;
12533
+ maxTokens: number;
12534
+ };
12200
12535
  readonly "~anthropic/claude-opus-latest": {
12201
12536
  id: string;
12202
12537
  name: string;
@@ -12214,6 +12549,108 @@ export declare const MODELS: {
12214
12549
  contextWindow: number;
12215
12550
  maxTokens: number;
12216
12551
  };
12552
+ readonly "~anthropic/claude-sonnet-latest": {
12553
+ id: string;
12554
+ name: string;
12555
+ api: "openai-completions";
12556
+ provider: string;
12557
+ baseUrl: string;
12558
+ reasoning: true;
12559
+ input: ("image" | "text")[];
12560
+ cost: {
12561
+ input: number;
12562
+ output: number;
12563
+ cacheRead: number;
12564
+ cacheWrite: number;
12565
+ };
12566
+ contextWindow: number;
12567
+ maxTokens: number;
12568
+ };
12569
+ readonly "~google/gemini-flash-latest": {
12570
+ id: string;
12571
+ name: string;
12572
+ api: "openai-completions";
12573
+ provider: string;
12574
+ baseUrl: string;
12575
+ reasoning: true;
12576
+ input: ("image" | "text")[];
12577
+ cost: {
12578
+ input: number;
12579
+ output: number;
12580
+ cacheRead: number;
12581
+ cacheWrite: number;
12582
+ };
12583
+ contextWindow: number;
12584
+ maxTokens: number;
12585
+ };
12586
+ readonly "~google/gemini-pro-latest": {
12587
+ id: string;
12588
+ name: string;
12589
+ api: "openai-completions";
12590
+ provider: string;
12591
+ baseUrl: string;
12592
+ reasoning: true;
12593
+ input: ("image" | "text")[];
12594
+ cost: {
12595
+ input: number;
12596
+ output: number;
12597
+ cacheRead: number;
12598
+ cacheWrite: number;
12599
+ };
12600
+ contextWindow: number;
12601
+ maxTokens: number;
12602
+ };
12603
+ readonly "~moonshotai/kimi-latest": {
12604
+ id: string;
12605
+ name: string;
12606
+ api: "openai-completions";
12607
+ provider: string;
12608
+ baseUrl: string;
12609
+ reasoning: true;
12610
+ input: ("image" | "text")[];
12611
+ cost: {
12612
+ input: number;
12613
+ output: number;
12614
+ cacheRead: number;
12615
+ cacheWrite: number;
12616
+ };
12617
+ contextWindow: number;
12618
+ maxTokens: number;
12619
+ };
12620
+ readonly "~openai/gpt-latest": {
12621
+ id: string;
12622
+ name: string;
12623
+ api: "openai-completions";
12624
+ provider: string;
12625
+ baseUrl: string;
12626
+ reasoning: true;
12627
+ input: ("image" | "text")[];
12628
+ cost: {
12629
+ input: number;
12630
+ output: number;
12631
+ cacheRead: number;
12632
+ cacheWrite: number;
12633
+ };
12634
+ contextWindow: number;
12635
+ maxTokens: number;
12636
+ };
12637
+ readonly "~openai/gpt-mini-latest": {
12638
+ id: string;
12639
+ name: string;
12640
+ api: "openai-completions";
12641
+ provider: string;
12642
+ baseUrl: string;
12643
+ reasoning: true;
12644
+ input: ("image" | "text")[];
12645
+ cost: {
12646
+ input: number;
12647
+ output: number;
12648
+ cacheRead: number;
12649
+ cacheWrite: number;
12650
+ };
12651
+ contextWindow: number;
12652
+ maxTokens: number;
12653
+ };
12217
12654
  };
12218
12655
  readonly "vercel-ai-gateway": {
12219
12656
  readonly "alibaba/qwen-3-14b": {
@@ -15435,6 +15872,28 @@ export declare const MODELS: {
15435
15872
  contextWindow: number;
15436
15873
  maxTokens: number;
15437
15874
  };
15875
+ readonly "glm-5v-turbo": {
15876
+ id: string;
15877
+ name: string;
15878
+ api: "openai-completions";
15879
+ provider: string;
15880
+ baseUrl: string;
15881
+ compat: {
15882
+ supportsDeveloperRole: false;
15883
+ thinkingFormat: "zai";
15884
+ zaiToolStream: true;
15885
+ };
15886
+ reasoning: true;
15887
+ input: ("image" | "text")[];
15888
+ cost: {
15889
+ input: number;
15890
+ output: number;
15891
+ cacheRead: number;
15892
+ cacheWrite: number;
15893
+ };
15894
+ contextWindow: number;
15895
+ maxTokens: number;
15896
+ };
15438
15897
  };
15439
15898
  };
15440
15899
  //# sourceMappingURL=models.generated.d.ts.map