@hyperspaceng/neural-ai 0.70.3 → 0.70.6
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +5 -3
- package/dist/env-api-keys.d.ts.map +1 -1
- package/dist/env-api-keys.js +50 -7
- package/dist/env-api-keys.js.map +1 -1
- package/dist/models.generated.d.ts +479 -20
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +490 -53
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/amazon-bedrock.d.ts.map +1 -1
- package/dist/providers/amazon-bedrock.js +49 -25
- package/dist/providers/amazon-bedrock.js.map +1 -1
- package/dist/providers/anthropic.d.ts.map +1 -1
- package/dist/providers/anthropic.js +11 -3
- package/dist/providers/anthropic.js.map +1 -1
- package/dist/providers/azure-openai-responses.d.ts.map +1 -1
- package/dist/providers/azure-openai-responses.js +17 -1
- package/dist/providers/azure-openai-responses.js.map +1 -1
- package/dist/providers/cloudflare.d.ts +7 -0
- package/dist/providers/cloudflare.d.ts.map +1 -0
- package/dist/providers/cloudflare.js +19 -0
- package/dist/providers/cloudflare.js.map +1 -0
- package/dist/providers/openai-codex-responses.d.ts.map +1 -1
- package/dist/providers/openai-codex-responses.js +1 -1
- package/dist/providers/openai-codex-responses.js.map +1 -1
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +5 -2
- package/dist/providers/openai-completions.js.map +1 -1
- package/dist/types.d.ts +1 -1
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/package.json +1 -1
|
@@ -2537,6 +2537,23 @@ export declare const MODELS: {
|
|
|
2537
2537
|
contextWindow: number;
|
|
2538
2538
|
maxTokens: number;
|
|
2539
2539
|
};
|
|
2540
|
+
readonly "gpt-5.5-pro": {
|
|
2541
|
+
id: string;
|
|
2542
|
+
name: string;
|
|
2543
|
+
api: "azure-openai-responses";
|
|
2544
|
+
provider: string;
|
|
2545
|
+
baseUrl: string;
|
|
2546
|
+
reasoning: true;
|
|
2547
|
+
input: ("image" | "text")[];
|
|
2548
|
+
cost: {
|
|
2549
|
+
input: number;
|
|
2550
|
+
output: number;
|
|
2551
|
+
cacheRead: number;
|
|
2552
|
+
cacheWrite: number;
|
|
2553
|
+
};
|
|
2554
|
+
contextWindow: number;
|
|
2555
|
+
maxTokens: number;
|
|
2556
|
+
};
|
|
2540
2557
|
readonly o1: {
|
|
2541
2558
|
id: string;
|
|
2542
2559
|
name: string;
|
|
@@ -2744,6 +2761,168 @@ export declare const MODELS: {
|
|
|
2744
2761
|
maxTokens: number;
|
|
2745
2762
|
};
|
|
2746
2763
|
};
|
|
2764
|
+
readonly "cloudflare-workers-ai": {
|
|
2765
|
+
readonly "@cf/google/gemma-4-26b-a4b-it": {
|
|
2766
|
+
id: string;
|
|
2767
|
+
name: string;
|
|
2768
|
+
api: "openai-completions";
|
|
2769
|
+
provider: string;
|
|
2770
|
+
baseUrl: string;
|
|
2771
|
+
compat: {
|
|
2772
|
+
sendSessionAffinityHeaders: true;
|
|
2773
|
+
};
|
|
2774
|
+
reasoning: true;
|
|
2775
|
+
input: ("image" | "text")[];
|
|
2776
|
+
cost: {
|
|
2777
|
+
input: number;
|
|
2778
|
+
output: number;
|
|
2779
|
+
cacheRead: number;
|
|
2780
|
+
cacheWrite: number;
|
|
2781
|
+
};
|
|
2782
|
+
contextWindow: number;
|
|
2783
|
+
maxTokens: number;
|
|
2784
|
+
};
|
|
2785
|
+
readonly "@cf/meta/llama-4-scout-17b-16e-instruct": {
|
|
2786
|
+
id: string;
|
|
2787
|
+
name: string;
|
|
2788
|
+
api: "openai-completions";
|
|
2789
|
+
provider: string;
|
|
2790
|
+
baseUrl: string;
|
|
2791
|
+
compat: {
|
|
2792
|
+
sendSessionAffinityHeaders: true;
|
|
2793
|
+
};
|
|
2794
|
+
reasoning: false;
|
|
2795
|
+
input: ("image" | "text")[];
|
|
2796
|
+
cost: {
|
|
2797
|
+
input: number;
|
|
2798
|
+
output: number;
|
|
2799
|
+
cacheRead: number;
|
|
2800
|
+
cacheWrite: number;
|
|
2801
|
+
};
|
|
2802
|
+
contextWindow: number;
|
|
2803
|
+
maxTokens: number;
|
|
2804
|
+
};
|
|
2805
|
+
readonly "@cf/moonshotai/kimi-k2.5": {
|
|
2806
|
+
id: string;
|
|
2807
|
+
name: string;
|
|
2808
|
+
api: "openai-completions";
|
|
2809
|
+
provider: string;
|
|
2810
|
+
baseUrl: string;
|
|
2811
|
+
compat: {
|
|
2812
|
+
sendSessionAffinityHeaders: true;
|
|
2813
|
+
};
|
|
2814
|
+
reasoning: true;
|
|
2815
|
+
input: ("image" | "text")[];
|
|
2816
|
+
cost: {
|
|
2817
|
+
input: number;
|
|
2818
|
+
output: number;
|
|
2819
|
+
cacheRead: number;
|
|
2820
|
+
cacheWrite: number;
|
|
2821
|
+
};
|
|
2822
|
+
contextWindow: number;
|
|
2823
|
+
maxTokens: number;
|
|
2824
|
+
};
|
|
2825
|
+
readonly "@cf/moonshotai/kimi-k2.6": {
|
|
2826
|
+
id: string;
|
|
2827
|
+
name: string;
|
|
2828
|
+
api: "openai-completions";
|
|
2829
|
+
provider: string;
|
|
2830
|
+
baseUrl: string;
|
|
2831
|
+
compat: {
|
|
2832
|
+
sendSessionAffinityHeaders: true;
|
|
2833
|
+
};
|
|
2834
|
+
reasoning: true;
|
|
2835
|
+
input: ("image" | "text")[];
|
|
2836
|
+
cost: {
|
|
2837
|
+
input: number;
|
|
2838
|
+
output: number;
|
|
2839
|
+
cacheRead: number;
|
|
2840
|
+
cacheWrite: number;
|
|
2841
|
+
};
|
|
2842
|
+
contextWindow: number;
|
|
2843
|
+
maxTokens: number;
|
|
2844
|
+
};
|
|
2845
|
+
readonly "@cf/nvidia/nemotron-3-120b-a12b": {
|
|
2846
|
+
id: string;
|
|
2847
|
+
name: string;
|
|
2848
|
+
api: "openai-completions";
|
|
2849
|
+
provider: string;
|
|
2850
|
+
baseUrl: string;
|
|
2851
|
+
compat: {
|
|
2852
|
+
sendSessionAffinityHeaders: true;
|
|
2853
|
+
};
|
|
2854
|
+
reasoning: true;
|
|
2855
|
+
input: "text"[];
|
|
2856
|
+
cost: {
|
|
2857
|
+
input: number;
|
|
2858
|
+
output: number;
|
|
2859
|
+
cacheRead: number;
|
|
2860
|
+
cacheWrite: number;
|
|
2861
|
+
};
|
|
2862
|
+
contextWindow: number;
|
|
2863
|
+
maxTokens: number;
|
|
2864
|
+
};
|
|
2865
|
+
readonly "@cf/openai/gpt-oss-120b": {
|
|
2866
|
+
id: string;
|
|
2867
|
+
name: string;
|
|
2868
|
+
api: "openai-completions";
|
|
2869
|
+
provider: string;
|
|
2870
|
+
baseUrl: string;
|
|
2871
|
+
compat: {
|
|
2872
|
+
sendSessionAffinityHeaders: true;
|
|
2873
|
+
};
|
|
2874
|
+
reasoning: true;
|
|
2875
|
+
input: "text"[];
|
|
2876
|
+
cost: {
|
|
2877
|
+
input: number;
|
|
2878
|
+
output: number;
|
|
2879
|
+
cacheRead: number;
|
|
2880
|
+
cacheWrite: number;
|
|
2881
|
+
};
|
|
2882
|
+
contextWindow: number;
|
|
2883
|
+
maxTokens: number;
|
|
2884
|
+
};
|
|
2885
|
+
readonly "@cf/openai/gpt-oss-20b": {
|
|
2886
|
+
id: string;
|
|
2887
|
+
name: string;
|
|
2888
|
+
api: "openai-completions";
|
|
2889
|
+
provider: string;
|
|
2890
|
+
baseUrl: string;
|
|
2891
|
+
compat: {
|
|
2892
|
+
sendSessionAffinityHeaders: true;
|
|
2893
|
+
};
|
|
2894
|
+
reasoning: true;
|
|
2895
|
+
input: "text"[];
|
|
2896
|
+
cost: {
|
|
2897
|
+
input: number;
|
|
2898
|
+
output: number;
|
|
2899
|
+
cacheRead: number;
|
|
2900
|
+
cacheWrite: number;
|
|
2901
|
+
};
|
|
2902
|
+
contextWindow: number;
|
|
2903
|
+
maxTokens: number;
|
|
2904
|
+
};
|
|
2905
|
+
readonly "@cf/zai-org/glm-4.7-flash": {
|
|
2906
|
+
id: string;
|
|
2907
|
+
name: string;
|
|
2908
|
+
api: "openai-completions";
|
|
2909
|
+
provider: string;
|
|
2910
|
+
baseUrl: string;
|
|
2911
|
+
compat: {
|
|
2912
|
+
sendSessionAffinityHeaders: true;
|
|
2913
|
+
};
|
|
2914
|
+
reasoning: true;
|
|
2915
|
+
input: "text"[];
|
|
2916
|
+
cost: {
|
|
2917
|
+
input: number;
|
|
2918
|
+
output: number;
|
|
2919
|
+
cacheRead: number;
|
|
2920
|
+
cacheWrite: number;
|
|
2921
|
+
};
|
|
2922
|
+
contextWindow: number;
|
|
2923
|
+
maxTokens: number;
|
|
2924
|
+
};
|
|
2925
|
+
};
|
|
2747
2926
|
readonly deepseek: {
|
|
2748
2927
|
readonly "deepseek-v4-flash": {
|
|
2749
2928
|
id: string;
|
|
@@ -2837,6 +3016,23 @@ export declare const MODELS: {
|
|
|
2837
3016
|
contextWindow: number;
|
|
2838
3017
|
maxTokens: number;
|
|
2839
3018
|
};
|
|
3019
|
+
readonly "accounts/fireworks/models/deepseek-v4-pro": {
|
|
3020
|
+
id: string;
|
|
3021
|
+
name: string;
|
|
3022
|
+
api: "anthropic-messages";
|
|
3023
|
+
provider: string;
|
|
3024
|
+
baseUrl: string;
|
|
3025
|
+
reasoning: true;
|
|
3026
|
+
input: "text"[];
|
|
3027
|
+
cost: {
|
|
3028
|
+
input: number;
|
|
3029
|
+
output: number;
|
|
3030
|
+
cacheRead: number;
|
|
3031
|
+
cacheWrite: number;
|
|
3032
|
+
};
|
|
3033
|
+
contextWindow: number;
|
|
3034
|
+
maxTokens: number;
|
|
3035
|
+
};
|
|
2840
3036
|
readonly "accounts/fireworks/models/glm-4p5": {
|
|
2841
3037
|
id: string;
|
|
2842
3038
|
name: string;
|
|
@@ -5263,6 +5459,26 @@ export declare const MODELS: {
|
|
|
5263
5459
|
contextWindow: number;
|
|
5264
5460
|
maxTokens: number;
|
|
5265
5461
|
};
|
|
5462
|
+
readonly "deepseek-ai/DeepSeek-V4-Pro": {
|
|
5463
|
+
id: string;
|
|
5464
|
+
name: string;
|
|
5465
|
+
api: "openai-completions";
|
|
5466
|
+
provider: string;
|
|
5467
|
+
baseUrl: string;
|
|
5468
|
+
compat: {
|
|
5469
|
+
supportsDeveloperRole: false;
|
|
5470
|
+
};
|
|
5471
|
+
reasoning: true;
|
|
5472
|
+
input: "text"[];
|
|
5473
|
+
cost: {
|
|
5474
|
+
input: number;
|
|
5475
|
+
output: number;
|
|
5476
|
+
cacheRead: number;
|
|
5477
|
+
cacheWrite: number;
|
|
5478
|
+
};
|
|
5479
|
+
contextWindow: number;
|
|
5480
|
+
maxTokens: number;
|
|
5481
|
+
};
|
|
5266
5482
|
readonly "moonshotai/Kimi-K2-Instruct": {
|
|
5267
5483
|
id: string;
|
|
5268
5484
|
name: string;
|
|
@@ -6584,6 +6800,23 @@ export declare const MODELS: {
|
|
|
6584
6800
|
contextWindow: number;
|
|
6585
6801
|
maxTokens: number;
|
|
6586
6802
|
};
|
|
6803
|
+
readonly "gpt-5.5-pro": {
|
|
6804
|
+
id: string;
|
|
6805
|
+
name: string;
|
|
6806
|
+
api: "openai-responses";
|
|
6807
|
+
provider: string;
|
|
6808
|
+
baseUrl: string;
|
|
6809
|
+
reasoning: true;
|
|
6810
|
+
input: ("image" | "text")[];
|
|
6811
|
+
cost: {
|
|
6812
|
+
input: number;
|
|
6813
|
+
output: number;
|
|
6814
|
+
cacheRead: number;
|
|
6815
|
+
cacheWrite: number;
|
|
6816
|
+
};
|
|
6817
|
+
contextWindow: number;
|
|
6818
|
+
maxTokens: number;
|
|
6819
|
+
};
|
|
6587
6820
|
readonly o1: {
|
|
6588
6821
|
id: string;
|
|
6589
6822
|
name: string;
|
|
@@ -6911,23 +7144,6 @@ export declare const MODELS: {
|
|
|
6911
7144
|
contextWindow: number;
|
|
6912
7145
|
maxTokens: number;
|
|
6913
7146
|
};
|
|
6914
|
-
readonly "claude-3-5-haiku": {
|
|
6915
|
-
id: string;
|
|
6916
|
-
name: string;
|
|
6917
|
-
api: "anthropic-messages";
|
|
6918
|
-
provider: string;
|
|
6919
|
-
baseUrl: string;
|
|
6920
|
-
reasoning: false;
|
|
6921
|
-
input: ("image" | "text")[];
|
|
6922
|
-
cost: {
|
|
6923
|
-
input: number;
|
|
6924
|
-
output: number;
|
|
6925
|
-
cacheRead: number;
|
|
6926
|
-
cacheWrite: number;
|
|
6927
|
-
};
|
|
6928
|
-
contextWindow: number;
|
|
6929
|
-
maxTokens: number;
|
|
6930
|
-
};
|
|
6931
7147
|
readonly "claude-haiku-4-5": {
|
|
6932
7148
|
id: string;
|
|
6933
7149
|
name: string;
|
|
@@ -8950,7 +9166,7 @@ export declare const MODELS: {
|
|
|
8950
9166
|
contextWindow: number;
|
|
8951
9167
|
maxTokens: number;
|
|
8952
9168
|
};
|
|
8953
|
-
readonly "inclusionai/ling-2.6-flash
|
|
9169
|
+
readonly "inclusionai/ling-2.6-flash": {
|
|
8954
9170
|
id: string;
|
|
8955
9171
|
name: string;
|
|
8956
9172
|
api: "openai-completions";
|
|
@@ -9732,6 +9948,23 @@ export declare const MODELS: {
|
|
|
9732
9948
|
contextWindow: number;
|
|
9733
9949
|
maxTokens: number;
|
|
9734
9950
|
};
|
|
9951
|
+
readonly "nvidia/nemotron-3-nano-omni-30b-a3b-reasoning:free": {
|
|
9952
|
+
id: string;
|
|
9953
|
+
name: string;
|
|
9954
|
+
api: "openai-completions";
|
|
9955
|
+
provider: string;
|
|
9956
|
+
baseUrl: string;
|
|
9957
|
+
reasoning: true;
|
|
9958
|
+
input: ("image" | "text")[];
|
|
9959
|
+
cost: {
|
|
9960
|
+
input: number;
|
|
9961
|
+
output: number;
|
|
9962
|
+
cacheRead: number;
|
|
9963
|
+
cacheWrite: number;
|
|
9964
|
+
};
|
|
9965
|
+
contextWindow: number;
|
|
9966
|
+
maxTokens: number;
|
|
9967
|
+
};
|
|
9735
9968
|
readonly "nvidia/nemotron-3-super-120b-a12b": {
|
|
9736
9969
|
id: string;
|
|
9737
9970
|
name: string;
|
|
@@ -10803,6 +11036,40 @@ export declare const MODELS: {
|
|
|
10803
11036
|
contextWindow: number;
|
|
10804
11037
|
maxTokens: number;
|
|
10805
11038
|
};
|
|
11039
|
+
readonly "poolside/laguna-m.1:free": {
|
|
11040
|
+
id: string;
|
|
11041
|
+
name: string;
|
|
11042
|
+
api: "openai-completions";
|
|
11043
|
+
provider: string;
|
|
11044
|
+
baseUrl: string;
|
|
11045
|
+
reasoning: true;
|
|
11046
|
+
input: "text"[];
|
|
11047
|
+
cost: {
|
|
11048
|
+
input: number;
|
|
11049
|
+
output: number;
|
|
11050
|
+
cacheRead: number;
|
|
11051
|
+
cacheWrite: number;
|
|
11052
|
+
};
|
|
11053
|
+
contextWindow: number;
|
|
11054
|
+
maxTokens: number;
|
|
11055
|
+
};
|
|
11056
|
+
readonly "poolside/laguna-xs.2:free": {
|
|
11057
|
+
id: string;
|
|
11058
|
+
name: string;
|
|
11059
|
+
api: "openai-completions";
|
|
11060
|
+
provider: string;
|
|
11061
|
+
baseUrl: string;
|
|
11062
|
+
reasoning: true;
|
|
11063
|
+
input: "text"[];
|
|
11064
|
+
cost: {
|
|
11065
|
+
input: number;
|
|
11066
|
+
output: number;
|
|
11067
|
+
cacheRead: number;
|
|
11068
|
+
cacheWrite: number;
|
|
11069
|
+
};
|
|
11070
|
+
contextWindow: number;
|
|
11071
|
+
maxTokens: number;
|
|
11072
|
+
};
|
|
10806
11073
|
readonly "prime-intellect/intellect-3": {
|
|
10807
11074
|
id: string;
|
|
10808
11075
|
name: string;
|
|
@@ -11534,7 +11801,7 @@ export declare const MODELS: {
|
|
|
11534
11801
|
contextWindow: number;
|
|
11535
11802
|
maxTokens: number;
|
|
11536
11803
|
};
|
|
11537
|
-
readonly "qwen/qwen3.
|
|
11804
|
+
readonly "qwen/qwen3.5-plus-20260420": {
|
|
11538
11805
|
id: string;
|
|
11539
11806
|
name: string;
|
|
11540
11807
|
api: "openai-completions";
|
|
@@ -11551,7 +11818,41 @@ export declare const MODELS: {
|
|
|
11551
11818
|
contextWindow: number;
|
|
11552
11819
|
maxTokens: number;
|
|
11553
11820
|
};
|
|
11554
|
-
readonly "qwen/
|
|
11821
|
+
readonly "qwen/qwen3.6-27b": {
|
|
11822
|
+
id: string;
|
|
11823
|
+
name: string;
|
|
11824
|
+
api: "openai-completions";
|
|
11825
|
+
provider: string;
|
|
11826
|
+
baseUrl: string;
|
|
11827
|
+
reasoning: true;
|
|
11828
|
+
input: ("image" | "text")[];
|
|
11829
|
+
cost: {
|
|
11830
|
+
input: number;
|
|
11831
|
+
output: number;
|
|
11832
|
+
cacheRead: number;
|
|
11833
|
+
cacheWrite: number;
|
|
11834
|
+
};
|
|
11835
|
+
contextWindow: number;
|
|
11836
|
+
maxTokens: number;
|
|
11837
|
+
};
|
|
11838
|
+
readonly "qwen/qwen3.6-flash": {
|
|
11839
|
+
id: string;
|
|
11840
|
+
name: string;
|
|
11841
|
+
api: "openai-completions";
|
|
11842
|
+
provider: string;
|
|
11843
|
+
baseUrl: string;
|
|
11844
|
+
reasoning: true;
|
|
11845
|
+
input: ("image" | "text")[];
|
|
11846
|
+
cost: {
|
|
11847
|
+
input: number;
|
|
11848
|
+
output: number;
|
|
11849
|
+
cacheRead: number;
|
|
11850
|
+
cacheWrite: number;
|
|
11851
|
+
};
|
|
11852
|
+
contextWindow: number;
|
|
11853
|
+
maxTokens: number;
|
|
11854
|
+
};
|
|
11855
|
+
readonly "qwen/qwen3.6-max-preview": {
|
|
11555
11856
|
id: string;
|
|
11556
11857
|
name: string;
|
|
11557
11858
|
api: "openai-completions";
|
|
@@ -11568,6 +11869,23 @@ export declare const MODELS: {
|
|
|
11568
11869
|
contextWindow: number;
|
|
11569
11870
|
maxTokens: number;
|
|
11570
11871
|
};
|
|
11872
|
+
readonly "qwen/qwen3.6-plus": {
|
|
11873
|
+
id: string;
|
|
11874
|
+
name: string;
|
|
11875
|
+
api: "openai-completions";
|
|
11876
|
+
provider: string;
|
|
11877
|
+
baseUrl: string;
|
|
11878
|
+
reasoning: true;
|
|
11879
|
+
input: ("image" | "text")[];
|
|
11880
|
+
cost: {
|
|
11881
|
+
input: number;
|
|
11882
|
+
output: number;
|
|
11883
|
+
cacheRead: number;
|
|
11884
|
+
cacheWrite: number;
|
|
11885
|
+
};
|
|
11886
|
+
contextWindow: number;
|
|
11887
|
+
maxTokens: number;
|
|
11888
|
+
};
|
|
11571
11889
|
readonly "rekaai/reka-edge": {
|
|
11572
11890
|
id: string;
|
|
11573
11891
|
name: string;
|
|
@@ -12197,6 +12515,23 @@ export declare const MODELS: {
|
|
|
12197
12515
|
contextWindow: number;
|
|
12198
12516
|
maxTokens: number;
|
|
12199
12517
|
};
|
|
12518
|
+
readonly "~anthropic/claude-haiku-latest": {
|
|
12519
|
+
id: string;
|
|
12520
|
+
name: string;
|
|
12521
|
+
api: "openai-completions";
|
|
12522
|
+
provider: string;
|
|
12523
|
+
baseUrl: string;
|
|
12524
|
+
reasoning: true;
|
|
12525
|
+
input: ("image" | "text")[];
|
|
12526
|
+
cost: {
|
|
12527
|
+
input: number;
|
|
12528
|
+
output: number;
|
|
12529
|
+
cacheRead: number;
|
|
12530
|
+
cacheWrite: number;
|
|
12531
|
+
};
|
|
12532
|
+
contextWindow: number;
|
|
12533
|
+
maxTokens: number;
|
|
12534
|
+
};
|
|
12200
12535
|
readonly "~anthropic/claude-opus-latest": {
|
|
12201
12536
|
id: string;
|
|
12202
12537
|
name: string;
|
|
@@ -12214,6 +12549,108 @@ export declare const MODELS: {
|
|
|
12214
12549
|
contextWindow: number;
|
|
12215
12550
|
maxTokens: number;
|
|
12216
12551
|
};
|
|
12552
|
+
readonly "~anthropic/claude-sonnet-latest": {
|
|
12553
|
+
id: string;
|
|
12554
|
+
name: string;
|
|
12555
|
+
api: "openai-completions";
|
|
12556
|
+
provider: string;
|
|
12557
|
+
baseUrl: string;
|
|
12558
|
+
reasoning: true;
|
|
12559
|
+
input: ("image" | "text")[];
|
|
12560
|
+
cost: {
|
|
12561
|
+
input: number;
|
|
12562
|
+
output: number;
|
|
12563
|
+
cacheRead: number;
|
|
12564
|
+
cacheWrite: number;
|
|
12565
|
+
};
|
|
12566
|
+
contextWindow: number;
|
|
12567
|
+
maxTokens: number;
|
|
12568
|
+
};
|
|
12569
|
+
readonly "~google/gemini-flash-latest": {
|
|
12570
|
+
id: string;
|
|
12571
|
+
name: string;
|
|
12572
|
+
api: "openai-completions";
|
|
12573
|
+
provider: string;
|
|
12574
|
+
baseUrl: string;
|
|
12575
|
+
reasoning: true;
|
|
12576
|
+
input: ("image" | "text")[];
|
|
12577
|
+
cost: {
|
|
12578
|
+
input: number;
|
|
12579
|
+
output: number;
|
|
12580
|
+
cacheRead: number;
|
|
12581
|
+
cacheWrite: number;
|
|
12582
|
+
};
|
|
12583
|
+
contextWindow: number;
|
|
12584
|
+
maxTokens: number;
|
|
12585
|
+
};
|
|
12586
|
+
readonly "~google/gemini-pro-latest": {
|
|
12587
|
+
id: string;
|
|
12588
|
+
name: string;
|
|
12589
|
+
api: "openai-completions";
|
|
12590
|
+
provider: string;
|
|
12591
|
+
baseUrl: string;
|
|
12592
|
+
reasoning: true;
|
|
12593
|
+
input: ("image" | "text")[];
|
|
12594
|
+
cost: {
|
|
12595
|
+
input: number;
|
|
12596
|
+
output: number;
|
|
12597
|
+
cacheRead: number;
|
|
12598
|
+
cacheWrite: number;
|
|
12599
|
+
};
|
|
12600
|
+
contextWindow: number;
|
|
12601
|
+
maxTokens: number;
|
|
12602
|
+
};
|
|
12603
|
+
readonly "~moonshotai/kimi-latest": {
|
|
12604
|
+
id: string;
|
|
12605
|
+
name: string;
|
|
12606
|
+
api: "openai-completions";
|
|
12607
|
+
provider: string;
|
|
12608
|
+
baseUrl: string;
|
|
12609
|
+
reasoning: true;
|
|
12610
|
+
input: ("image" | "text")[];
|
|
12611
|
+
cost: {
|
|
12612
|
+
input: number;
|
|
12613
|
+
output: number;
|
|
12614
|
+
cacheRead: number;
|
|
12615
|
+
cacheWrite: number;
|
|
12616
|
+
};
|
|
12617
|
+
contextWindow: number;
|
|
12618
|
+
maxTokens: number;
|
|
12619
|
+
};
|
|
12620
|
+
readonly "~openai/gpt-latest": {
|
|
12621
|
+
id: string;
|
|
12622
|
+
name: string;
|
|
12623
|
+
api: "openai-completions";
|
|
12624
|
+
provider: string;
|
|
12625
|
+
baseUrl: string;
|
|
12626
|
+
reasoning: true;
|
|
12627
|
+
input: ("image" | "text")[];
|
|
12628
|
+
cost: {
|
|
12629
|
+
input: number;
|
|
12630
|
+
output: number;
|
|
12631
|
+
cacheRead: number;
|
|
12632
|
+
cacheWrite: number;
|
|
12633
|
+
};
|
|
12634
|
+
contextWindow: number;
|
|
12635
|
+
maxTokens: number;
|
|
12636
|
+
};
|
|
12637
|
+
readonly "~openai/gpt-mini-latest": {
|
|
12638
|
+
id: string;
|
|
12639
|
+
name: string;
|
|
12640
|
+
api: "openai-completions";
|
|
12641
|
+
provider: string;
|
|
12642
|
+
baseUrl: string;
|
|
12643
|
+
reasoning: true;
|
|
12644
|
+
input: ("image" | "text")[];
|
|
12645
|
+
cost: {
|
|
12646
|
+
input: number;
|
|
12647
|
+
output: number;
|
|
12648
|
+
cacheRead: number;
|
|
12649
|
+
cacheWrite: number;
|
|
12650
|
+
};
|
|
12651
|
+
contextWindow: number;
|
|
12652
|
+
maxTokens: number;
|
|
12653
|
+
};
|
|
12217
12654
|
};
|
|
12218
12655
|
readonly "vercel-ai-gateway": {
|
|
12219
12656
|
readonly "alibaba/qwen-3-14b": {
|
|
@@ -15435,6 +15872,28 @@ export declare const MODELS: {
|
|
|
15435
15872
|
contextWindow: number;
|
|
15436
15873
|
maxTokens: number;
|
|
15437
15874
|
};
|
|
15875
|
+
readonly "glm-5v-turbo": {
|
|
15876
|
+
id: string;
|
|
15877
|
+
name: string;
|
|
15878
|
+
api: "openai-completions";
|
|
15879
|
+
provider: string;
|
|
15880
|
+
baseUrl: string;
|
|
15881
|
+
compat: {
|
|
15882
|
+
supportsDeveloperRole: false;
|
|
15883
|
+
thinkingFormat: "zai";
|
|
15884
|
+
zaiToolStream: true;
|
|
15885
|
+
};
|
|
15886
|
+
reasoning: true;
|
|
15887
|
+
input: ("image" | "text")[];
|
|
15888
|
+
cost: {
|
|
15889
|
+
input: number;
|
|
15890
|
+
output: number;
|
|
15891
|
+
cacheRead: number;
|
|
15892
|
+
cacheWrite: number;
|
|
15893
|
+
};
|
|
15894
|
+
contextWindow: number;
|
|
15895
|
+
maxTokens: number;
|
|
15896
|
+
};
|
|
15438
15897
|
};
|
|
15439
15898
|
};
|
|
15440
15899
|
//# sourceMappingURL=models.generated.d.ts.map
|