@hyperspaceng/neural-ai 0.70.3 → 0.70.7
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +5 -3
- package/dist/env-api-keys.d.ts.map +1 -1
- package/dist/env-api-keys.js +50 -7
- package/dist/env-api-keys.js.map +1 -1
- package/dist/models.generated.d.ts +536 -60
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +562 -108
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/amazon-bedrock.d.ts.map +1 -1
- package/dist/providers/amazon-bedrock.js +48 -25
- package/dist/providers/amazon-bedrock.js.map +1 -1
- package/dist/providers/anthropic.d.ts.map +1 -1
- package/dist/providers/anthropic.js +11 -3
- package/dist/providers/anthropic.js.map +1 -1
- package/dist/providers/azure-openai-responses.d.ts.map +1 -1
- package/dist/providers/azure-openai-responses.js +17 -1
- package/dist/providers/azure-openai-responses.js.map +1 -1
- package/dist/providers/cloudflare.d.ts +7 -0
- package/dist/providers/cloudflare.d.ts.map +1 -0
- package/dist/providers/cloudflare.js +19 -0
- package/dist/providers/cloudflare.js.map +1 -0
- package/dist/providers/openai-codex-responses.d.ts.map +1 -1
- package/dist/providers/openai-codex-responses.js +1 -1
- package/dist/providers/openai-codex-responses.js.map +1 -1
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +5 -2
- package/dist/providers/openai-completions.js.map +1 -1
- package/dist/types.d.ts +1 -1
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/package.json +1 -1
|
@@ -2537,6 +2537,23 @@ export declare const MODELS: {
|
|
|
2537
2537
|
contextWindow: number;
|
|
2538
2538
|
maxTokens: number;
|
|
2539
2539
|
};
|
|
2540
|
+
readonly "gpt-5.5-pro": {
|
|
2541
|
+
id: string;
|
|
2542
|
+
name: string;
|
|
2543
|
+
api: "azure-openai-responses";
|
|
2544
|
+
provider: string;
|
|
2545
|
+
baseUrl: string;
|
|
2546
|
+
reasoning: true;
|
|
2547
|
+
input: ("image" | "text")[];
|
|
2548
|
+
cost: {
|
|
2549
|
+
input: number;
|
|
2550
|
+
output: number;
|
|
2551
|
+
cacheRead: number;
|
|
2552
|
+
cacheWrite: number;
|
|
2553
|
+
};
|
|
2554
|
+
contextWindow: number;
|
|
2555
|
+
maxTokens: number;
|
|
2556
|
+
};
|
|
2540
2557
|
readonly o1: {
|
|
2541
2558
|
id: string;
|
|
2542
2559
|
name: string;
|
|
@@ -2744,6 +2761,168 @@ export declare const MODELS: {
|
|
|
2744
2761
|
maxTokens: number;
|
|
2745
2762
|
};
|
|
2746
2763
|
};
|
|
2764
|
+
readonly "cloudflare-workers-ai": {
|
|
2765
|
+
readonly "@cf/google/gemma-4-26b-a4b-it": {
|
|
2766
|
+
id: string;
|
|
2767
|
+
name: string;
|
|
2768
|
+
api: "openai-completions";
|
|
2769
|
+
provider: string;
|
|
2770
|
+
baseUrl: string;
|
|
2771
|
+
compat: {
|
|
2772
|
+
sendSessionAffinityHeaders: true;
|
|
2773
|
+
};
|
|
2774
|
+
reasoning: true;
|
|
2775
|
+
input: ("image" | "text")[];
|
|
2776
|
+
cost: {
|
|
2777
|
+
input: number;
|
|
2778
|
+
output: number;
|
|
2779
|
+
cacheRead: number;
|
|
2780
|
+
cacheWrite: number;
|
|
2781
|
+
};
|
|
2782
|
+
contextWindow: number;
|
|
2783
|
+
maxTokens: number;
|
|
2784
|
+
};
|
|
2785
|
+
readonly "@cf/meta/llama-4-scout-17b-16e-instruct": {
|
|
2786
|
+
id: string;
|
|
2787
|
+
name: string;
|
|
2788
|
+
api: "openai-completions";
|
|
2789
|
+
provider: string;
|
|
2790
|
+
baseUrl: string;
|
|
2791
|
+
compat: {
|
|
2792
|
+
sendSessionAffinityHeaders: true;
|
|
2793
|
+
};
|
|
2794
|
+
reasoning: false;
|
|
2795
|
+
input: ("image" | "text")[];
|
|
2796
|
+
cost: {
|
|
2797
|
+
input: number;
|
|
2798
|
+
output: number;
|
|
2799
|
+
cacheRead: number;
|
|
2800
|
+
cacheWrite: number;
|
|
2801
|
+
};
|
|
2802
|
+
contextWindow: number;
|
|
2803
|
+
maxTokens: number;
|
|
2804
|
+
};
|
|
2805
|
+
readonly "@cf/moonshotai/kimi-k2.5": {
|
|
2806
|
+
id: string;
|
|
2807
|
+
name: string;
|
|
2808
|
+
api: "openai-completions";
|
|
2809
|
+
provider: string;
|
|
2810
|
+
baseUrl: string;
|
|
2811
|
+
compat: {
|
|
2812
|
+
sendSessionAffinityHeaders: true;
|
|
2813
|
+
};
|
|
2814
|
+
reasoning: true;
|
|
2815
|
+
input: ("image" | "text")[];
|
|
2816
|
+
cost: {
|
|
2817
|
+
input: number;
|
|
2818
|
+
output: number;
|
|
2819
|
+
cacheRead: number;
|
|
2820
|
+
cacheWrite: number;
|
|
2821
|
+
};
|
|
2822
|
+
contextWindow: number;
|
|
2823
|
+
maxTokens: number;
|
|
2824
|
+
};
|
|
2825
|
+
readonly "@cf/moonshotai/kimi-k2.6": {
|
|
2826
|
+
id: string;
|
|
2827
|
+
name: string;
|
|
2828
|
+
api: "openai-completions";
|
|
2829
|
+
provider: string;
|
|
2830
|
+
baseUrl: string;
|
|
2831
|
+
compat: {
|
|
2832
|
+
sendSessionAffinityHeaders: true;
|
|
2833
|
+
};
|
|
2834
|
+
reasoning: true;
|
|
2835
|
+
input: ("image" | "text")[];
|
|
2836
|
+
cost: {
|
|
2837
|
+
input: number;
|
|
2838
|
+
output: number;
|
|
2839
|
+
cacheRead: number;
|
|
2840
|
+
cacheWrite: number;
|
|
2841
|
+
};
|
|
2842
|
+
contextWindow: number;
|
|
2843
|
+
maxTokens: number;
|
|
2844
|
+
};
|
|
2845
|
+
readonly "@cf/nvidia/nemotron-3-120b-a12b": {
|
|
2846
|
+
id: string;
|
|
2847
|
+
name: string;
|
|
2848
|
+
api: "openai-completions";
|
|
2849
|
+
provider: string;
|
|
2850
|
+
baseUrl: string;
|
|
2851
|
+
compat: {
|
|
2852
|
+
sendSessionAffinityHeaders: true;
|
|
2853
|
+
};
|
|
2854
|
+
reasoning: true;
|
|
2855
|
+
input: "text"[];
|
|
2856
|
+
cost: {
|
|
2857
|
+
input: number;
|
|
2858
|
+
output: number;
|
|
2859
|
+
cacheRead: number;
|
|
2860
|
+
cacheWrite: number;
|
|
2861
|
+
};
|
|
2862
|
+
contextWindow: number;
|
|
2863
|
+
maxTokens: number;
|
|
2864
|
+
};
|
|
2865
|
+
readonly "@cf/openai/gpt-oss-120b": {
|
|
2866
|
+
id: string;
|
|
2867
|
+
name: string;
|
|
2868
|
+
api: "openai-completions";
|
|
2869
|
+
provider: string;
|
|
2870
|
+
baseUrl: string;
|
|
2871
|
+
compat: {
|
|
2872
|
+
sendSessionAffinityHeaders: true;
|
|
2873
|
+
};
|
|
2874
|
+
reasoning: true;
|
|
2875
|
+
input: "text"[];
|
|
2876
|
+
cost: {
|
|
2877
|
+
input: number;
|
|
2878
|
+
output: number;
|
|
2879
|
+
cacheRead: number;
|
|
2880
|
+
cacheWrite: number;
|
|
2881
|
+
};
|
|
2882
|
+
contextWindow: number;
|
|
2883
|
+
maxTokens: number;
|
|
2884
|
+
};
|
|
2885
|
+
readonly "@cf/openai/gpt-oss-20b": {
|
|
2886
|
+
id: string;
|
|
2887
|
+
name: string;
|
|
2888
|
+
api: "openai-completions";
|
|
2889
|
+
provider: string;
|
|
2890
|
+
baseUrl: string;
|
|
2891
|
+
compat: {
|
|
2892
|
+
sendSessionAffinityHeaders: true;
|
|
2893
|
+
};
|
|
2894
|
+
reasoning: true;
|
|
2895
|
+
input: "text"[];
|
|
2896
|
+
cost: {
|
|
2897
|
+
input: number;
|
|
2898
|
+
output: number;
|
|
2899
|
+
cacheRead: number;
|
|
2900
|
+
cacheWrite: number;
|
|
2901
|
+
};
|
|
2902
|
+
contextWindow: number;
|
|
2903
|
+
maxTokens: number;
|
|
2904
|
+
};
|
|
2905
|
+
readonly "@cf/zai-org/glm-4.7-flash": {
|
|
2906
|
+
id: string;
|
|
2907
|
+
name: string;
|
|
2908
|
+
api: "openai-completions";
|
|
2909
|
+
provider: string;
|
|
2910
|
+
baseUrl: string;
|
|
2911
|
+
compat: {
|
|
2912
|
+
sendSessionAffinityHeaders: true;
|
|
2913
|
+
};
|
|
2914
|
+
reasoning: true;
|
|
2915
|
+
input: "text"[];
|
|
2916
|
+
cost: {
|
|
2917
|
+
input: number;
|
|
2918
|
+
output: number;
|
|
2919
|
+
cacheRead: number;
|
|
2920
|
+
cacheWrite: number;
|
|
2921
|
+
};
|
|
2922
|
+
contextWindow: number;
|
|
2923
|
+
maxTokens: number;
|
|
2924
|
+
};
|
|
2925
|
+
};
|
|
2747
2926
|
readonly deepseek: {
|
|
2748
2927
|
readonly "deepseek-v4-flash": {
|
|
2749
2928
|
id: string;
|
|
@@ -2837,6 +3016,23 @@ export declare const MODELS: {
|
|
|
2837
3016
|
contextWindow: number;
|
|
2838
3017
|
maxTokens: number;
|
|
2839
3018
|
};
|
|
3019
|
+
readonly "accounts/fireworks/models/deepseek-v4-pro": {
|
|
3020
|
+
id: string;
|
|
3021
|
+
name: string;
|
|
3022
|
+
api: "anthropic-messages";
|
|
3023
|
+
provider: string;
|
|
3024
|
+
baseUrl: string;
|
|
3025
|
+
reasoning: true;
|
|
3026
|
+
input: "text"[];
|
|
3027
|
+
cost: {
|
|
3028
|
+
input: number;
|
|
3029
|
+
output: number;
|
|
3030
|
+
cacheRead: number;
|
|
3031
|
+
cacheWrite: number;
|
|
3032
|
+
};
|
|
3033
|
+
contextWindow: number;
|
|
3034
|
+
maxTokens: number;
|
|
3035
|
+
};
|
|
2840
3036
|
readonly "accounts/fireworks/models/glm-4p5": {
|
|
2841
3037
|
id: string;
|
|
2842
3038
|
name: string;
|
|
@@ -5263,6 +5459,26 @@ export declare const MODELS: {
|
|
|
5263
5459
|
contextWindow: number;
|
|
5264
5460
|
maxTokens: number;
|
|
5265
5461
|
};
|
|
5462
|
+
readonly "deepseek-ai/DeepSeek-V4-Pro": {
|
|
5463
|
+
id: string;
|
|
5464
|
+
name: string;
|
|
5465
|
+
api: "openai-completions";
|
|
5466
|
+
provider: string;
|
|
5467
|
+
baseUrl: string;
|
|
5468
|
+
compat: {
|
|
5469
|
+
supportsDeveloperRole: false;
|
|
5470
|
+
};
|
|
5471
|
+
reasoning: true;
|
|
5472
|
+
input: "text"[];
|
|
5473
|
+
cost: {
|
|
5474
|
+
input: number;
|
|
5475
|
+
output: number;
|
|
5476
|
+
cacheRead: number;
|
|
5477
|
+
cacheWrite: number;
|
|
5478
|
+
};
|
|
5479
|
+
contextWindow: number;
|
|
5480
|
+
maxTokens: number;
|
|
5481
|
+
};
|
|
5266
5482
|
readonly "moonshotai/Kimi-K2-Instruct": {
|
|
5267
5483
|
id: string;
|
|
5268
5484
|
name: string;
|
|
@@ -6584,6 +6800,23 @@ export declare const MODELS: {
|
|
|
6584
6800
|
contextWindow: number;
|
|
6585
6801
|
maxTokens: number;
|
|
6586
6802
|
};
|
|
6803
|
+
readonly "gpt-5.5-pro": {
|
|
6804
|
+
id: string;
|
|
6805
|
+
name: string;
|
|
6806
|
+
api: "openai-responses";
|
|
6807
|
+
provider: string;
|
|
6808
|
+
baseUrl: string;
|
|
6809
|
+
reasoning: true;
|
|
6810
|
+
input: ("image" | "text")[];
|
|
6811
|
+
cost: {
|
|
6812
|
+
input: number;
|
|
6813
|
+
output: number;
|
|
6814
|
+
cacheRead: number;
|
|
6815
|
+
cacheWrite: number;
|
|
6816
|
+
};
|
|
6817
|
+
contextWindow: number;
|
|
6818
|
+
maxTokens: number;
|
|
6819
|
+
};
|
|
6587
6820
|
readonly o1: {
|
|
6588
6821
|
id: string;
|
|
6589
6822
|
name: string;
|
|
@@ -6911,23 +7144,6 @@ export declare const MODELS: {
|
|
|
6911
7144
|
contextWindow: number;
|
|
6912
7145
|
maxTokens: number;
|
|
6913
7146
|
};
|
|
6914
|
-
readonly "claude-3-5-haiku": {
|
|
6915
|
-
id: string;
|
|
6916
|
-
name: string;
|
|
6917
|
-
api: "anthropic-messages";
|
|
6918
|
-
provider: string;
|
|
6919
|
-
baseUrl: string;
|
|
6920
|
-
reasoning: false;
|
|
6921
|
-
input: ("image" | "text")[];
|
|
6922
|
-
cost: {
|
|
6923
|
-
input: number;
|
|
6924
|
-
output: number;
|
|
6925
|
-
cacheRead: number;
|
|
6926
|
-
cacheWrite: number;
|
|
6927
|
-
};
|
|
6928
|
-
contextWindow: number;
|
|
6929
|
-
maxTokens: number;
|
|
6930
|
-
};
|
|
6931
7147
|
readonly "claude-haiku-4-5": {
|
|
6932
7148
|
id: string;
|
|
6933
7149
|
name: string;
|
|
@@ -7455,23 +7671,6 @@ export declare const MODELS: {
|
|
|
7455
7671
|
contextWindow: number;
|
|
7456
7672
|
maxTokens: number;
|
|
7457
7673
|
};
|
|
7458
|
-
readonly "ling-2.6-flash-free": {
|
|
7459
|
-
id: string;
|
|
7460
|
-
name: string;
|
|
7461
|
-
api: "openai-completions";
|
|
7462
|
-
provider: string;
|
|
7463
|
-
baseUrl: string;
|
|
7464
|
-
reasoning: false;
|
|
7465
|
-
input: "text"[];
|
|
7466
|
-
cost: {
|
|
7467
|
-
input: number;
|
|
7468
|
-
output: number;
|
|
7469
|
-
cacheRead: number;
|
|
7470
|
-
cacheWrite: number;
|
|
7471
|
-
};
|
|
7472
|
-
contextWindow: number;
|
|
7473
|
-
maxTokens: number;
|
|
7474
|
-
};
|
|
7475
7674
|
readonly "minimax-m2.5": {
|
|
7476
7675
|
id: string;
|
|
7477
7676
|
name: string;
|
|
@@ -8950,7 +9149,7 @@ export declare const MODELS: {
|
|
|
8950
9149
|
contextWindow: number;
|
|
8951
9150
|
maxTokens: number;
|
|
8952
9151
|
};
|
|
8953
|
-
readonly "inclusionai/ling-2.6-flash
|
|
9152
|
+
readonly "inclusionai/ling-2.6-flash": {
|
|
8954
9153
|
id: string;
|
|
8955
9154
|
name: string;
|
|
8956
9155
|
api: "openai-completions";
|
|
@@ -9477,7 +9676,7 @@ export declare const MODELS: {
|
|
|
9477
9676
|
contextWindow: number;
|
|
9478
9677
|
maxTokens: number;
|
|
9479
9678
|
};
|
|
9480
|
-
readonly "mistralai/
|
|
9679
|
+
readonly "mistralai/mixtral-8x22b-instruct": {
|
|
9481
9680
|
id: string;
|
|
9482
9681
|
name: string;
|
|
9483
9682
|
api: "openai-completions";
|
|
@@ -9494,7 +9693,7 @@ export declare const MODELS: {
|
|
|
9494
9693
|
contextWindow: number;
|
|
9495
9694
|
maxTokens: number;
|
|
9496
9695
|
};
|
|
9497
|
-
readonly "mistralai/mixtral-
|
|
9696
|
+
readonly "mistralai/mixtral-8x7b-instruct": {
|
|
9498
9697
|
id: string;
|
|
9499
9698
|
name: string;
|
|
9500
9699
|
api: "openai-completions";
|
|
@@ -9511,14 +9710,14 @@ export declare const MODELS: {
|
|
|
9511
9710
|
contextWindow: number;
|
|
9512
9711
|
maxTokens: number;
|
|
9513
9712
|
};
|
|
9514
|
-
readonly "mistralai/
|
|
9713
|
+
readonly "mistralai/pixtral-large-2411": {
|
|
9515
9714
|
id: string;
|
|
9516
9715
|
name: string;
|
|
9517
9716
|
api: "openai-completions";
|
|
9518
9717
|
provider: string;
|
|
9519
9718
|
baseUrl: string;
|
|
9520
9719
|
reasoning: false;
|
|
9521
|
-
input: "text"[];
|
|
9720
|
+
input: ("image" | "text")[];
|
|
9522
9721
|
cost: {
|
|
9523
9722
|
input: number;
|
|
9524
9723
|
output: number;
|
|
@@ -9528,24 +9727,7 @@ export declare const MODELS: {
|
|
|
9528
9727
|
contextWindow: number;
|
|
9529
9728
|
maxTokens: number;
|
|
9530
9729
|
};
|
|
9531
|
-
readonly "mistralai/
|
|
9532
|
-
id: string;
|
|
9533
|
-
name: string;
|
|
9534
|
-
api: "openai-completions";
|
|
9535
|
-
provider: string;
|
|
9536
|
-
baseUrl: string;
|
|
9537
|
-
reasoning: false;
|
|
9538
|
-
input: ("image" | "text")[];
|
|
9539
|
-
cost: {
|
|
9540
|
-
input: number;
|
|
9541
|
-
output: number;
|
|
9542
|
-
cacheRead: number;
|
|
9543
|
-
cacheWrite: number;
|
|
9544
|
-
};
|
|
9545
|
-
contextWindow: number;
|
|
9546
|
-
maxTokens: number;
|
|
9547
|
-
};
|
|
9548
|
-
readonly "mistralai/voxtral-small-24b-2507": {
|
|
9730
|
+
readonly "mistralai/voxtral-small-24b-2507": {
|
|
9549
9731
|
id: string;
|
|
9550
9732
|
name: string;
|
|
9551
9733
|
api: "openai-completions";
|
|
@@ -9732,6 +9914,23 @@ export declare const MODELS: {
|
|
|
9732
9914
|
contextWindow: number;
|
|
9733
9915
|
maxTokens: number;
|
|
9734
9916
|
};
|
|
9917
|
+
readonly "nvidia/nemotron-3-nano-omni-30b-a3b-reasoning:free": {
|
|
9918
|
+
id: string;
|
|
9919
|
+
name: string;
|
|
9920
|
+
api: "openai-completions";
|
|
9921
|
+
provider: string;
|
|
9922
|
+
baseUrl: string;
|
|
9923
|
+
reasoning: true;
|
|
9924
|
+
input: ("image" | "text")[];
|
|
9925
|
+
cost: {
|
|
9926
|
+
input: number;
|
|
9927
|
+
output: number;
|
|
9928
|
+
cacheRead: number;
|
|
9929
|
+
cacheWrite: number;
|
|
9930
|
+
};
|
|
9931
|
+
contextWindow: number;
|
|
9932
|
+
maxTokens: number;
|
|
9933
|
+
};
|
|
9735
9934
|
readonly "nvidia/nemotron-3-super-120b-a12b": {
|
|
9736
9935
|
id: string;
|
|
9737
9936
|
name: string;
|
|
@@ -10803,6 +11002,40 @@ export declare const MODELS: {
|
|
|
10803
11002
|
contextWindow: number;
|
|
10804
11003
|
maxTokens: number;
|
|
10805
11004
|
};
|
|
11005
|
+
readonly "poolside/laguna-m.1:free": {
|
|
11006
|
+
id: string;
|
|
11007
|
+
name: string;
|
|
11008
|
+
api: "openai-completions";
|
|
11009
|
+
provider: string;
|
|
11010
|
+
baseUrl: string;
|
|
11011
|
+
reasoning: true;
|
|
11012
|
+
input: "text"[];
|
|
11013
|
+
cost: {
|
|
11014
|
+
input: number;
|
|
11015
|
+
output: number;
|
|
11016
|
+
cacheRead: number;
|
|
11017
|
+
cacheWrite: number;
|
|
11018
|
+
};
|
|
11019
|
+
contextWindow: number;
|
|
11020
|
+
maxTokens: number;
|
|
11021
|
+
};
|
|
11022
|
+
readonly "poolside/laguna-xs.2:free": {
|
|
11023
|
+
id: string;
|
|
11024
|
+
name: string;
|
|
11025
|
+
api: "openai-completions";
|
|
11026
|
+
provider: string;
|
|
11027
|
+
baseUrl: string;
|
|
11028
|
+
reasoning: true;
|
|
11029
|
+
input: "text"[];
|
|
11030
|
+
cost: {
|
|
11031
|
+
input: number;
|
|
11032
|
+
output: number;
|
|
11033
|
+
cacheRead: number;
|
|
11034
|
+
cacheWrite: number;
|
|
11035
|
+
};
|
|
11036
|
+
contextWindow: number;
|
|
11037
|
+
maxTokens: number;
|
|
11038
|
+
};
|
|
10806
11039
|
readonly "prime-intellect/intellect-3": {
|
|
10807
11040
|
id: string;
|
|
10808
11041
|
name: string;
|
|
@@ -11534,7 +11767,41 @@ export declare const MODELS: {
|
|
|
11534
11767
|
contextWindow: number;
|
|
11535
11768
|
maxTokens: number;
|
|
11536
11769
|
};
|
|
11537
|
-
readonly "qwen/qwen3.
|
|
11770
|
+
readonly "qwen/qwen3.5-plus-20260420": {
|
|
11771
|
+
id: string;
|
|
11772
|
+
name: string;
|
|
11773
|
+
api: "openai-completions";
|
|
11774
|
+
provider: string;
|
|
11775
|
+
baseUrl: string;
|
|
11776
|
+
reasoning: true;
|
|
11777
|
+
input: ("image" | "text")[];
|
|
11778
|
+
cost: {
|
|
11779
|
+
input: number;
|
|
11780
|
+
output: number;
|
|
11781
|
+
cacheRead: number;
|
|
11782
|
+
cacheWrite: number;
|
|
11783
|
+
};
|
|
11784
|
+
contextWindow: number;
|
|
11785
|
+
maxTokens: number;
|
|
11786
|
+
};
|
|
11787
|
+
readonly "qwen/qwen3.6-27b": {
|
|
11788
|
+
id: string;
|
|
11789
|
+
name: string;
|
|
11790
|
+
api: "openai-completions";
|
|
11791
|
+
provider: string;
|
|
11792
|
+
baseUrl: string;
|
|
11793
|
+
reasoning: true;
|
|
11794
|
+
input: ("image" | "text")[];
|
|
11795
|
+
cost: {
|
|
11796
|
+
input: number;
|
|
11797
|
+
output: number;
|
|
11798
|
+
cacheRead: number;
|
|
11799
|
+
cacheWrite: number;
|
|
11800
|
+
};
|
|
11801
|
+
contextWindow: number;
|
|
11802
|
+
maxTokens: number;
|
|
11803
|
+
};
|
|
11804
|
+
readonly "qwen/qwen3.6-flash": {
|
|
11538
11805
|
id: string;
|
|
11539
11806
|
name: string;
|
|
11540
11807
|
api: "openai-completions";
|
|
@@ -11551,7 +11818,7 @@ export declare const MODELS: {
|
|
|
11551
11818
|
contextWindow: number;
|
|
11552
11819
|
maxTokens: number;
|
|
11553
11820
|
};
|
|
11554
|
-
readonly "qwen/
|
|
11821
|
+
readonly "qwen/qwen3.6-max-preview": {
|
|
11555
11822
|
id: string;
|
|
11556
11823
|
name: string;
|
|
11557
11824
|
api: "openai-completions";
|
|
@@ -11568,6 +11835,23 @@ export declare const MODELS: {
|
|
|
11568
11835
|
contextWindow: number;
|
|
11569
11836
|
maxTokens: number;
|
|
11570
11837
|
};
|
|
11838
|
+
readonly "qwen/qwen3.6-plus": {
|
|
11839
|
+
id: string;
|
|
11840
|
+
name: string;
|
|
11841
|
+
api: "openai-completions";
|
|
11842
|
+
provider: string;
|
|
11843
|
+
baseUrl: string;
|
|
11844
|
+
reasoning: true;
|
|
11845
|
+
input: ("image" | "text")[];
|
|
11846
|
+
cost: {
|
|
11847
|
+
input: number;
|
|
11848
|
+
output: number;
|
|
11849
|
+
cacheRead: number;
|
|
11850
|
+
cacheWrite: number;
|
|
11851
|
+
};
|
|
11852
|
+
contextWindow: number;
|
|
11853
|
+
maxTokens: number;
|
|
11854
|
+
};
|
|
11571
11855
|
readonly "rekaai/reka-edge": {
|
|
11572
11856
|
id: string;
|
|
11573
11857
|
name: string;
|
|
@@ -12197,6 +12481,23 @@ export declare const MODELS: {
|
|
|
12197
12481
|
contextWindow: number;
|
|
12198
12482
|
maxTokens: number;
|
|
12199
12483
|
};
|
|
12484
|
+
readonly "~anthropic/claude-haiku-latest": {
|
|
12485
|
+
id: string;
|
|
12486
|
+
name: string;
|
|
12487
|
+
api: "openai-completions";
|
|
12488
|
+
provider: string;
|
|
12489
|
+
baseUrl: string;
|
|
12490
|
+
reasoning: true;
|
|
12491
|
+
input: ("image" | "text")[];
|
|
12492
|
+
cost: {
|
|
12493
|
+
input: number;
|
|
12494
|
+
output: number;
|
|
12495
|
+
cacheRead: number;
|
|
12496
|
+
cacheWrite: number;
|
|
12497
|
+
};
|
|
12498
|
+
contextWindow: number;
|
|
12499
|
+
maxTokens: number;
|
|
12500
|
+
};
|
|
12200
12501
|
readonly "~anthropic/claude-opus-latest": {
|
|
12201
12502
|
id: string;
|
|
12202
12503
|
name: string;
|
|
@@ -12214,6 +12515,108 @@ export declare const MODELS: {
|
|
|
12214
12515
|
contextWindow: number;
|
|
12215
12516
|
maxTokens: number;
|
|
12216
12517
|
};
|
|
12518
|
+
readonly "~anthropic/claude-sonnet-latest": {
|
|
12519
|
+
id: string;
|
|
12520
|
+
name: string;
|
|
12521
|
+
api: "openai-completions";
|
|
12522
|
+
provider: string;
|
|
12523
|
+
baseUrl: string;
|
|
12524
|
+
reasoning: true;
|
|
12525
|
+
input: ("image" | "text")[];
|
|
12526
|
+
cost: {
|
|
12527
|
+
input: number;
|
|
12528
|
+
output: number;
|
|
12529
|
+
cacheRead: number;
|
|
12530
|
+
cacheWrite: number;
|
|
12531
|
+
};
|
|
12532
|
+
contextWindow: number;
|
|
12533
|
+
maxTokens: number;
|
|
12534
|
+
};
|
|
12535
|
+
readonly "~google/gemini-flash-latest": {
|
|
12536
|
+
id: string;
|
|
12537
|
+
name: string;
|
|
12538
|
+
api: "openai-completions";
|
|
12539
|
+
provider: string;
|
|
12540
|
+
baseUrl: string;
|
|
12541
|
+
reasoning: true;
|
|
12542
|
+
input: ("image" | "text")[];
|
|
12543
|
+
cost: {
|
|
12544
|
+
input: number;
|
|
12545
|
+
output: number;
|
|
12546
|
+
cacheRead: number;
|
|
12547
|
+
cacheWrite: number;
|
|
12548
|
+
};
|
|
12549
|
+
contextWindow: number;
|
|
12550
|
+
maxTokens: number;
|
|
12551
|
+
};
|
|
12552
|
+
readonly "~google/gemini-pro-latest": {
|
|
12553
|
+
id: string;
|
|
12554
|
+
name: string;
|
|
12555
|
+
api: "openai-completions";
|
|
12556
|
+
provider: string;
|
|
12557
|
+
baseUrl: string;
|
|
12558
|
+
reasoning: true;
|
|
12559
|
+
input: ("image" | "text")[];
|
|
12560
|
+
cost: {
|
|
12561
|
+
input: number;
|
|
12562
|
+
output: number;
|
|
12563
|
+
cacheRead: number;
|
|
12564
|
+
cacheWrite: number;
|
|
12565
|
+
};
|
|
12566
|
+
contextWindow: number;
|
|
12567
|
+
maxTokens: number;
|
|
12568
|
+
};
|
|
12569
|
+
readonly "~moonshotai/kimi-latest": {
|
|
12570
|
+
id: string;
|
|
12571
|
+
name: string;
|
|
12572
|
+
api: "openai-completions";
|
|
12573
|
+
provider: string;
|
|
12574
|
+
baseUrl: string;
|
|
12575
|
+
reasoning: true;
|
|
12576
|
+
input: ("image" | "text")[];
|
|
12577
|
+
cost: {
|
|
12578
|
+
input: number;
|
|
12579
|
+
output: number;
|
|
12580
|
+
cacheRead: number;
|
|
12581
|
+
cacheWrite: number;
|
|
12582
|
+
};
|
|
12583
|
+
contextWindow: number;
|
|
12584
|
+
maxTokens: number;
|
|
12585
|
+
};
|
|
12586
|
+
readonly "~openai/gpt-latest": {
|
|
12587
|
+
id: string;
|
|
12588
|
+
name: string;
|
|
12589
|
+
api: "openai-completions";
|
|
12590
|
+
provider: string;
|
|
12591
|
+
baseUrl: string;
|
|
12592
|
+
reasoning: true;
|
|
12593
|
+
input: ("image" | "text")[];
|
|
12594
|
+
cost: {
|
|
12595
|
+
input: number;
|
|
12596
|
+
output: number;
|
|
12597
|
+
cacheRead: number;
|
|
12598
|
+
cacheWrite: number;
|
|
12599
|
+
};
|
|
12600
|
+
contextWindow: number;
|
|
12601
|
+
maxTokens: number;
|
|
12602
|
+
};
|
|
12603
|
+
readonly "~openai/gpt-mini-latest": {
|
|
12604
|
+
id: string;
|
|
12605
|
+
name: string;
|
|
12606
|
+
api: "openai-completions";
|
|
12607
|
+
provider: string;
|
|
12608
|
+
baseUrl: string;
|
|
12609
|
+
reasoning: true;
|
|
12610
|
+
input: ("image" | "text")[];
|
|
12611
|
+
cost: {
|
|
12612
|
+
input: number;
|
|
12613
|
+
output: number;
|
|
12614
|
+
cacheRead: number;
|
|
12615
|
+
cacheWrite: number;
|
|
12616
|
+
};
|
|
12617
|
+
contextWindow: number;
|
|
12618
|
+
maxTokens: number;
|
|
12619
|
+
};
|
|
12217
12620
|
};
|
|
12218
12621
|
readonly "vercel-ai-gateway": {
|
|
12219
12622
|
readonly "alibaba/qwen-3-14b": {
|
|
@@ -12488,6 +12891,23 @@ export declare const MODELS: {
|
|
|
12488
12891
|
contextWindow: number;
|
|
12489
12892
|
maxTokens: number;
|
|
12490
12893
|
};
|
|
12894
|
+
readonly "alibaba/qwen3.6-27b": {
|
|
12895
|
+
id: string;
|
|
12896
|
+
name: string;
|
|
12897
|
+
api: "anthropic-messages";
|
|
12898
|
+
provider: string;
|
|
12899
|
+
baseUrl: string;
|
|
12900
|
+
reasoning: true;
|
|
12901
|
+
input: ("image" | "text")[];
|
|
12902
|
+
cost: {
|
|
12903
|
+
input: number;
|
|
12904
|
+
output: number;
|
|
12905
|
+
cacheRead: number;
|
|
12906
|
+
cacheWrite: number;
|
|
12907
|
+
};
|
|
12908
|
+
contextWindow: number;
|
|
12909
|
+
maxTokens: number;
|
|
12910
|
+
};
|
|
12491
12911
|
readonly "alibaba/qwen3.6-plus": {
|
|
12492
12912
|
id: string;
|
|
12493
12913
|
name: string;
|
|
@@ -12868,7 +13288,7 @@ export declare const MODELS: {
|
|
|
12868
13288
|
api: "anthropic-messages";
|
|
12869
13289
|
provider: string;
|
|
12870
13290
|
baseUrl: string;
|
|
12871
|
-
reasoning:
|
|
13291
|
+
reasoning: false;
|
|
12872
13292
|
input: "text"[];
|
|
12873
13293
|
cost: {
|
|
12874
13294
|
input: number;
|
|
@@ -14715,6 +15135,40 @@ export declare const MODELS: {
|
|
|
14715
15135
|
contextWindow: number;
|
|
14716
15136
|
maxTokens: number;
|
|
14717
15137
|
};
|
|
15138
|
+
readonly "xiaomi/mimo-v2.5": {
|
|
15139
|
+
id: string;
|
|
15140
|
+
name: string;
|
|
15141
|
+
api: "anthropic-messages";
|
|
15142
|
+
provider: string;
|
|
15143
|
+
baseUrl: string;
|
|
15144
|
+
reasoning: true;
|
|
15145
|
+
input: ("image" | "text")[];
|
|
15146
|
+
cost: {
|
|
15147
|
+
input: number;
|
|
15148
|
+
output: number;
|
|
15149
|
+
cacheRead: number;
|
|
15150
|
+
cacheWrite: number;
|
|
15151
|
+
};
|
|
15152
|
+
contextWindow: number;
|
|
15153
|
+
maxTokens: number;
|
|
15154
|
+
};
|
|
15155
|
+
readonly "xiaomi/mimo-v2.5-pro": {
|
|
15156
|
+
id: string;
|
|
15157
|
+
name: string;
|
|
15158
|
+
api: "anthropic-messages";
|
|
15159
|
+
provider: string;
|
|
15160
|
+
baseUrl: string;
|
|
15161
|
+
reasoning: true;
|
|
15162
|
+
input: ("image" | "text")[];
|
|
15163
|
+
cost: {
|
|
15164
|
+
input: number;
|
|
15165
|
+
output: number;
|
|
15166
|
+
cacheRead: number;
|
|
15167
|
+
cacheWrite: number;
|
|
15168
|
+
};
|
|
15169
|
+
contextWindow: number;
|
|
15170
|
+
maxTokens: number;
|
|
15171
|
+
};
|
|
14718
15172
|
readonly "zai/glm-4.5": {
|
|
14719
15173
|
id: string;
|
|
14720
15174
|
name: string;
|
|
@@ -15435,6 +15889,28 @@ export declare const MODELS: {
|
|
|
15435
15889
|
contextWindow: number;
|
|
15436
15890
|
maxTokens: number;
|
|
15437
15891
|
};
|
|
15892
|
+
readonly "glm-5v-turbo": {
|
|
15893
|
+
id: string;
|
|
15894
|
+
name: string;
|
|
15895
|
+
api: "openai-completions";
|
|
15896
|
+
provider: string;
|
|
15897
|
+
baseUrl: string;
|
|
15898
|
+
compat: {
|
|
15899
|
+
supportsDeveloperRole: false;
|
|
15900
|
+
thinkingFormat: "zai";
|
|
15901
|
+
zaiToolStream: true;
|
|
15902
|
+
};
|
|
15903
|
+
reasoning: true;
|
|
15904
|
+
input: ("image" | "text")[];
|
|
15905
|
+
cost: {
|
|
15906
|
+
input: number;
|
|
15907
|
+
output: number;
|
|
15908
|
+
cacheRead: number;
|
|
15909
|
+
cacheWrite: number;
|
|
15910
|
+
};
|
|
15911
|
+
contextWindow: number;
|
|
15912
|
+
maxTokens: number;
|
|
15913
|
+
};
|
|
15438
15914
|
};
|
|
15439
15915
|
};
|
|
15440
15916
|
//# sourceMappingURL=models.generated.d.ts.map
|