@hyperspaceng/neural-ai 0.70.1 → 0.70.6
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +9 -4
- package/dist/env-api-keys.d.ts.map +1 -1
- package/dist/env-api-keys.js +51 -7
- package/dist/env-api-keys.js.map +1 -1
- package/dist/models.d.ts +1 -0
- package/dist/models.d.ts.map +1 -1
- package/dist/models.generated.d.ts +606 -249
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +539 -204
- package/dist/models.generated.js.map +1 -1
- package/dist/models.js +3 -1
- package/dist/models.js.map +1 -1
- package/dist/providers/amazon-bedrock.d.ts.map +1 -1
- package/dist/providers/amazon-bedrock.js +49 -25
- package/dist/providers/amazon-bedrock.js.map +1 -1
- package/dist/providers/anthropic.d.ts.map +1 -1
- package/dist/providers/anthropic.js +18 -7
- package/dist/providers/anthropic.js.map +1 -1
- package/dist/providers/azure-openai-responses.d.ts.map +1 -1
- package/dist/providers/azure-openai-responses.js +24 -5
- package/dist/providers/azure-openai-responses.js.map +1 -1
- package/dist/providers/cloudflare.d.ts +7 -0
- package/dist/providers/cloudflare.d.ts.map +1 -0
- package/dist/providers/cloudflare.js +19 -0
- package/dist/providers/cloudflare.js.map +1 -0
- package/dist/providers/openai-codex-responses.d.ts.map +1 -1
- package/dist/providers/openai-codex-responses.js +2 -2
- package/dist/providers/openai-codex-responses.js.map +1 -1
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +49 -16
- package/dist/providers/openai-completions.js.map +1 -1
- package/dist/providers/openai-responses.d.ts.map +1 -1
- package/dist/providers/openai-responses.js +7 -4
- package/dist/providers/openai-responses.js.map +1 -1
- package/dist/providers/simple-options.d.ts.map +1 -1
- package/dist/providers/simple-options.js +2 -0
- package/dist/providers/simple-options.js.map +1 -1
- package/dist/types.d.ts +15 -3
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/package.json +1 -1
|
@@ -2537,6 +2537,23 @@ export declare const MODELS: {
|
|
|
2537
2537
|
contextWindow: number;
|
|
2538
2538
|
maxTokens: number;
|
|
2539
2539
|
};
|
|
2540
|
+
readonly "gpt-5.5-pro": {
|
|
2541
|
+
id: string;
|
|
2542
|
+
name: string;
|
|
2543
|
+
api: "azure-openai-responses";
|
|
2544
|
+
provider: string;
|
|
2545
|
+
baseUrl: string;
|
|
2546
|
+
reasoning: true;
|
|
2547
|
+
input: ("image" | "text")[];
|
|
2548
|
+
cost: {
|
|
2549
|
+
input: number;
|
|
2550
|
+
output: number;
|
|
2551
|
+
cacheRead: number;
|
|
2552
|
+
cacheWrite: number;
|
|
2553
|
+
};
|
|
2554
|
+
contextWindow: number;
|
|
2555
|
+
maxTokens: number;
|
|
2556
|
+
};
|
|
2540
2557
|
readonly o1: {
|
|
2541
2558
|
id: string;
|
|
2542
2559
|
name: string;
|
|
@@ -2744,6 +2761,226 @@ export declare const MODELS: {
|
|
|
2744
2761
|
maxTokens: number;
|
|
2745
2762
|
};
|
|
2746
2763
|
};
|
|
2764
|
+
readonly "cloudflare-workers-ai": {
|
|
2765
|
+
readonly "@cf/google/gemma-4-26b-a4b-it": {
|
|
2766
|
+
id: string;
|
|
2767
|
+
name: string;
|
|
2768
|
+
api: "openai-completions";
|
|
2769
|
+
provider: string;
|
|
2770
|
+
baseUrl: string;
|
|
2771
|
+
compat: {
|
|
2772
|
+
sendSessionAffinityHeaders: true;
|
|
2773
|
+
};
|
|
2774
|
+
reasoning: true;
|
|
2775
|
+
input: ("image" | "text")[];
|
|
2776
|
+
cost: {
|
|
2777
|
+
input: number;
|
|
2778
|
+
output: number;
|
|
2779
|
+
cacheRead: number;
|
|
2780
|
+
cacheWrite: number;
|
|
2781
|
+
};
|
|
2782
|
+
contextWindow: number;
|
|
2783
|
+
maxTokens: number;
|
|
2784
|
+
};
|
|
2785
|
+
readonly "@cf/meta/llama-4-scout-17b-16e-instruct": {
|
|
2786
|
+
id: string;
|
|
2787
|
+
name: string;
|
|
2788
|
+
api: "openai-completions";
|
|
2789
|
+
provider: string;
|
|
2790
|
+
baseUrl: string;
|
|
2791
|
+
compat: {
|
|
2792
|
+
sendSessionAffinityHeaders: true;
|
|
2793
|
+
};
|
|
2794
|
+
reasoning: false;
|
|
2795
|
+
input: ("image" | "text")[];
|
|
2796
|
+
cost: {
|
|
2797
|
+
input: number;
|
|
2798
|
+
output: number;
|
|
2799
|
+
cacheRead: number;
|
|
2800
|
+
cacheWrite: number;
|
|
2801
|
+
};
|
|
2802
|
+
contextWindow: number;
|
|
2803
|
+
maxTokens: number;
|
|
2804
|
+
};
|
|
2805
|
+
readonly "@cf/moonshotai/kimi-k2.5": {
|
|
2806
|
+
id: string;
|
|
2807
|
+
name: string;
|
|
2808
|
+
api: "openai-completions";
|
|
2809
|
+
provider: string;
|
|
2810
|
+
baseUrl: string;
|
|
2811
|
+
compat: {
|
|
2812
|
+
sendSessionAffinityHeaders: true;
|
|
2813
|
+
};
|
|
2814
|
+
reasoning: true;
|
|
2815
|
+
input: ("image" | "text")[];
|
|
2816
|
+
cost: {
|
|
2817
|
+
input: number;
|
|
2818
|
+
output: number;
|
|
2819
|
+
cacheRead: number;
|
|
2820
|
+
cacheWrite: number;
|
|
2821
|
+
};
|
|
2822
|
+
contextWindow: number;
|
|
2823
|
+
maxTokens: number;
|
|
2824
|
+
};
|
|
2825
|
+
readonly "@cf/moonshotai/kimi-k2.6": {
|
|
2826
|
+
id: string;
|
|
2827
|
+
name: string;
|
|
2828
|
+
api: "openai-completions";
|
|
2829
|
+
provider: string;
|
|
2830
|
+
baseUrl: string;
|
|
2831
|
+
compat: {
|
|
2832
|
+
sendSessionAffinityHeaders: true;
|
|
2833
|
+
};
|
|
2834
|
+
reasoning: true;
|
|
2835
|
+
input: ("image" | "text")[];
|
|
2836
|
+
cost: {
|
|
2837
|
+
input: number;
|
|
2838
|
+
output: number;
|
|
2839
|
+
cacheRead: number;
|
|
2840
|
+
cacheWrite: number;
|
|
2841
|
+
};
|
|
2842
|
+
contextWindow: number;
|
|
2843
|
+
maxTokens: number;
|
|
2844
|
+
};
|
|
2845
|
+
readonly "@cf/nvidia/nemotron-3-120b-a12b": {
|
|
2846
|
+
id: string;
|
|
2847
|
+
name: string;
|
|
2848
|
+
api: "openai-completions";
|
|
2849
|
+
provider: string;
|
|
2850
|
+
baseUrl: string;
|
|
2851
|
+
compat: {
|
|
2852
|
+
sendSessionAffinityHeaders: true;
|
|
2853
|
+
};
|
|
2854
|
+
reasoning: true;
|
|
2855
|
+
input: "text"[];
|
|
2856
|
+
cost: {
|
|
2857
|
+
input: number;
|
|
2858
|
+
output: number;
|
|
2859
|
+
cacheRead: number;
|
|
2860
|
+
cacheWrite: number;
|
|
2861
|
+
};
|
|
2862
|
+
contextWindow: number;
|
|
2863
|
+
maxTokens: number;
|
|
2864
|
+
};
|
|
2865
|
+
readonly "@cf/openai/gpt-oss-120b": {
|
|
2866
|
+
id: string;
|
|
2867
|
+
name: string;
|
|
2868
|
+
api: "openai-completions";
|
|
2869
|
+
provider: string;
|
|
2870
|
+
baseUrl: string;
|
|
2871
|
+
compat: {
|
|
2872
|
+
sendSessionAffinityHeaders: true;
|
|
2873
|
+
};
|
|
2874
|
+
reasoning: true;
|
|
2875
|
+
input: "text"[];
|
|
2876
|
+
cost: {
|
|
2877
|
+
input: number;
|
|
2878
|
+
output: number;
|
|
2879
|
+
cacheRead: number;
|
|
2880
|
+
cacheWrite: number;
|
|
2881
|
+
};
|
|
2882
|
+
contextWindow: number;
|
|
2883
|
+
maxTokens: number;
|
|
2884
|
+
};
|
|
2885
|
+
readonly "@cf/openai/gpt-oss-20b": {
|
|
2886
|
+
id: string;
|
|
2887
|
+
name: string;
|
|
2888
|
+
api: "openai-completions";
|
|
2889
|
+
provider: string;
|
|
2890
|
+
baseUrl: string;
|
|
2891
|
+
compat: {
|
|
2892
|
+
sendSessionAffinityHeaders: true;
|
|
2893
|
+
};
|
|
2894
|
+
reasoning: true;
|
|
2895
|
+
input: "text"[];
|
|
2896
|
+
cost: {
|
|
2897
|
+
input: number;
|
|
2898
|
+
output: number;
|
|
2899
|
+
cacheRead: number;
|
|
2900
|
+
cacheWrite: number;
|
|
2901
|
+
};
|
|
2902
|
+
contextWindow: number;
|
|
2903
|
+
maxTokens: number;
|
|
2904
|
+
};
|
|
2905
|
+
readonly "@cf/zai-org/glm-4.7-flash": {
|
|
2906
|
+
id: string;
|
|
2907
|
+
name: string;
|
|
2908
|
+
api: "openai-completions";
|
|
2909
|
+
provider: string;
|
|
2910
|
+
baseUrl: string;
|
|
2911
|
+
compat: {
|
|
2912
|
+
sendSessionAffinityHeaders: true;
|
|
2913
|
+
};
|
|
2914
|
+
reasoning: true;
|
|
2915
|
+
input: "text"[];
|
|
2916
|
+
cost: {
|
|
2917
|
+
input: number;
|
|
2918
|
+
output: number;
|
|
2919
|
+
cacheRead: number;
|
|
2920
|
+
cacheWrite: number;
|
|
2921
|
+
};
|
|
2922
|
+
contextWindow: number;
|
|
2923
|
+
maxTokens: number;
|
|
2924
|
+
};
|
|
2925
|
+
};
|
|
2926
|
+
readonly deepseek: {
|
|
2927
|
+
readonly "deepseek-v4-flash": {
|
|
2928
|
+
id: string;
|
|
2929
|
+
name: string;
|
|
2930
|
+
api: "openai-completions";
|
|
2931
|
+
provider: string;
|
|
2932
|
+
baseUrl: string;
|
|
2933
|
+
compat: {
|
|
2934
|
+
requiresReasoningContentOnAssistantMessages: true;
|
|
2935
|
+
thinkingFormat: "deepseek";
|
|
2936
|
+
reasoningEffortMap: {
|
|
2937
|
+
minimal: string;
|
|
2938
|
+
low: string;
|
|
2939
|
+
medium: string;
|
|
2940
|
+
high: string;
|
|
2941
|
+
xhigh: string;
|
|
2942
|
+
};
|
|
2943
|
+
};
|
|
2944
|
+
reasoning: true;
|
|
2945
|
+
input: "text"[];
|
|
2946
|
+
cost: {
|
|
2947
|
+
input: number;
|
|
2948
|
+
output: number;
|
|
2949
|
+
cacheRead: number;
|
|
2950
|
+
cacheWrite: number;
|
|
2951
|
+
};
|
|
2952
|
+
contextWindow: number;
|
|
2953
|
+
maxTokens: number;
|
|
2954
|
+
};
|
|
2955
|
+
readonly "deepseek-v4-pro": {
|
|
2956
|
+
id: string;
|
|
2957
|
+
name: string;
|
|
2958
|
+
api: "openai-completions";
|
|
2959
|
+
provider: string;
|
|
2960
|
+
baseUrl: string;
|
|
2961
|
+
compat: {
|
|
2962
|
+
requiresReasoningContentOnAssistantMessages: true;
|
|
2963
|
+
thinkingFormat: "deepseek";
|
|
2964
|
+
reasoningEffortMap: {
|
|
2965
|
+
minimal: string;
|
|
2966
|
+
low: string;
|
|
2967
|
+
medium: string;
|
|
2968
|
+
high: string;
|
|
2969
|
+
xhigh: string;
|
|
2970
|
+
};
|
|
2971
|
+
};
|
|
2972
|
+
reasoning: true;
|
|
2973
|
+
input: "text"[];
|
|
2974
|
+
cost: {
|
|
2975
|
+
input: number;
|
|
2976
|
+
output: number;
|
|
2977
|
+
cacheRead: number;
|
|
2978
|
+
cacheWrite: number;
|
|
2979
|
+
};
|
|
2980
|
+
contextWindow: number;
|
|
2981
|
+
maxTokens: number;
|
|
2982
|
+
};
|
|
2983
|
+
};
|
|
2747
2984
|
readonly fireworks: {
|
|
2748
2985
|
readonly "accounts/fireworks/models/deepseek-v3p1": {
|
|
2749
2986
|
id: string;
|
|
@@ -2779,6 +3016,23 @@ export declare const MODELS: {
|
|
|
2779
3016
|
contextWindow: number;
|
|
2780
3017
|
maxTokens: number;
|
|
2781
3018
|
};
|
|
3019
|
+
readonly "accounts/fireworks/models/deepseek-v4-pro": {
|
|
3020
|
+
id: string;
|
|
3021
|
+
name: string;
|
|
3022
|
+
api: "anthropic-messages";
|
|
3023
|
+
provider: string;
|
|
3024
|
+
baseUrl: string;
|
|
3025
|
+
reasoning: true;
|
|
3026
|
+
input: "text"[];
|
|
3027
|
+
cost: {
|
|
3028
|
+
input: number;
|
|
3029
|
+
output: number;
|
|
3030
|
+
cacheRead: number;
|
|
3031
|
+
cacheWrite: number;
|
|
3032
|
+
};
|
|
3033
|
+
contextWindow: number;
|
|
3034
|
+
maxTokens: number;
|
|
3035
|
+
};
|
|
2782
3036
|
readonly "accounts/fireworks/models/glm-4p5": {
|
|
2783
3037
|
id: string;
|
|
2784
3038
|
name: string;
|
|
@@ -3644,6 +3898,29 @@ export declare const MODELS: {
|
|
|
3644
3898
|
contextWindow: number;
|
|
3645
3899
|
maxTokens: number;
|
|
3646
3900
|
};
|
|
3901
|
+
readonly "gpt-5.5": {
|
|
3902
|
+
id: string;
|
|
3903
|
+
name: string;
|
|
3904
|
+
api: "openai-responses";
|
|
3905
|
+
provider: string;
|
|
3906
|
+
baseUrl: string;
|
|
3907
|
+
headers: {
|
|
3908
|
+
"User-Agent": string;
|
|
3909
|
+
"Editor-Version": string;
|
|
3910
|
+
"Editor-Plugin-Version": string;
|
|
3911
|
+
"Copilot-Integration-Id": string;
|
|
3912
|
+
};
|
|
3913
|
+
reasoning: true;
|
|
3914
|
+
input: ("image" | "text")[];
|
|
3915
|
+
cost: {
|
|
3916
|
+
input: number;
|
|
3917
|
+
output: number;
|
|
3918
|
+
cacheRead: number;
|
|
3919
|
+
cacheWrite: number;
|
|
3920
|
+
};
|
|
3921
|
+
contextWindow: number;
|
|
3922
|
+
maxTokens: number;
|
|
3923
|
+
};
|
|
3647
3924
|
readonly "grok-code-fast-1": {
|
|
3648
3925
|
id: string;
|
|
3649
3926
|
name: string;
|
|
@@ -5182,6 +5459,26 @@ export declare const MODELS: {
|
|
|
5182
5459
|
contextWindow: number;
|
|
5183
5460
|
maxTokens: number;
|
|
5184
5461
|
};
|
|
5462
|
+
readonly "deepseek-ai/DeepSeek-V4-Pro": {
|
|
5463
|
+
id: string;
|
|
5464
|
+
name: string;
|
|
5465
|
+
api: "openai-completions";
|
|
5466
|
+
provider: string;
|
|
5467
|
+
baseUrl: string;
|
|
5468
|
+
compat: {
|
|
5469
|
+
supportsDeveloperRole: false;
|
|
5470
|
+
};
|
|
5471
|
+
reasoning: true;
|
|
5472
|
+
input: "text"[];
|
|
5473
|
+
cost: {
|
|
5474
|
+
input: number;
|
|
5475
|
+
output: number;
|
|
5476
|
+
cacheRead: number;
|
|
5477
|
+
cacheWrite: number;
|
|
5478
|
+
};
|
|
5479
|
+
contextWindow: number;
|
|
5480
|
+
maxTokens: number;
|
|
5481
|
+
};
|
|
5185
5482
|
readonly "moonshotai/Kimi-K2-Instruct": {
|
|
5186
5483
|
id: string;
|
|
5187
5484
|
name: string;
|
|
@@ -6373,7 +6670,24 @@ export declare const MODELS: {
|
|
|
6373
6670
|
api: "openai-responses";
|
|
6374
6671
|
provider: string;
|
|
6375
6672
|
baseUrl: string;
|
|
6376
|
-
reasoning: false;
|
|
6673
|
+
reasoning: false;
|
|
6674
|
+
input: ("image" | "text")[];
|
|
6675
|
+
cost: {
|
|
6676
|
+
input: number;
|
|
6677
|
+
output: number;
|
|
6678
|
+
cacheRead: number;
|
|
6679
|
+
cacheWrite: number;
|
|
6680
|
+
};
|
|
6681
|
+
contextWindow: number;
|
|
6682
|
+
maxTokens: number;
|
|
6683
|
+
};
|
|
6684
|
+
readonly "gpt-5.3-codex": {
|
|
6685
|
+
id: string;
|
|
6686
|
+
name: string;
|
|
6687
|
+
api: "openai-responses";
|
|
6688
|
+
provider: string;
|
|
6689
|
+
baseUrl: string;
|
|
6690
|
+
reasoning: true;
|
|
6377
6691
|
input: ("image" | "text")[];
|
|
6378
6692
|
cost: {
|
|
6379
6693
|
input: number;
|
|
@@ -6384,7 +6698,7 @@ export declare const MODELS: {
|
|
|
6384
6698
|
contextWindow: number;
|
|
6385
6699
|
maxTokens: number;
|
|
6386
6700
|
};
|
|
6387
|
-
readonly "gpt-5.3-codex": {
|
|
6701
|
+
readonly "gpt-5.3-codex-spark": {
|
|
6388
6702
|
id: string;
|
|
6389
6703
|
name: string;
|
|
6390
6704
|
api: "openai-responses";
|
|
@@ -6401,7 +6715,7 @@ export declare const MODELS: {
|
|
|
6401
6715
|
contextWindow: number;
|
|
6402
6716
|
maxTokens: number;
|
|
6403
6717
|
};
|
|
6404
|
-
readonly "gpt-5.
|
|
6718
|
+
readonly "gpt-5.4": {
|
|
6405
6719
|
id: string;
|
|
6406
6720
|
name: string;
|
|
6407
6721
|
api: "openai-responses";
|
|
@@ -6418,7 +6732,7 @@ export declare const MODELS: {
|
|
|
6418
6732
|
contextWindow: number;
|
|
6419
6733
|
maxTokens: number;
|
|
6420
6734
|
};
|
|
6421
|
-
readonly "gpt-5.4": {
|
|
6735
|
+
readonly "gpt-5.4-mini": {
|
|
6422
6736
|
id: string;
|
|
6423
6737
|
name: string;
|
|
6424
6738
|
api: "openai-responses";
|
|
@@ -6435,7 +6749,7 @@ export declare const MODELS: {
|
|
|
6435
6749
|
contextWindow: number;
|
|
6436
6750
|
maxTokens: number;
|
|
6437
6751
|
};
|
|
6438
|
-
readonly "gpt-5.4-
|
|
6752
|
+
readonly "gpt-5.4-nano": {
|
|
6439
6753
|
id: string;
|
|
6440
6754
|
name: string;
|
|
6441
6755
|
api: "openai-responses";
|
|
@@ -6452,7 +6766,7 @@ export declare const MODELS: {
|
|
|
6452
6766
|
contextWindow: number;
|
|
6453
6767
|
maxTokens: number;
|
|
6454
6768
|
};
|
|
6455
|
-
readonly "gpt-5.4-
|
|
6769
|
+
readonly "gpt-5.4-pro": {
|
|
6456
6770
|
id: string;
|
|
6457
6771
|
name: string;
|
|
6458
6772
|
api: "openai-responses";
|
|
@@ -6469,7 +6783,7 @@ export declare const MODELS: {
|
|
|
6469
6783
|
contextWindow: number;
|
|
6470
6784
|
maxTokens: number;
|
|
6471
6785
|
};
|
|
6472
|
-
readonly "gpt-5.
|
|
6786
|
+
readonly "gpt-5.5": {
|
|
6473
6787
|
id: string;
|
|
6474
6788
|
name: string;
|
|
6475
6789
|
api: "openai-responses";
|
|
@@ -6486,7 +6800,7 @@ export declare const MODELS: {
|
|
|
6486
6800
|
contextWindow: number;
|
|
6487
6801
|
maxTokens: number;
|
|
6488
6802
|
};
|
|
6489
|
-
readonly "gpt-5.5": {
|
|
6803
|
+
readonly "gpt-5.5-pro": {
|
|
6490
6804
|
id: string;
|
|
6491
6805
|
name: string;
|
|
6492
6806
|
api: "openai-responses";
|
|
@@ -6830,23 +7144,6 @@ export declare const MODELS: {
|
|
|
6830
7144
|
contextWindow: number;
|
|
6831
7145
|
maxTokens: number;
|
|
6832
7146
|
};
|
|
6833
|
-
readonly "claude-3-5-haiku": {
|
|
6834
|
-
id: string;
|
|
6835
|
-
name: string;
|
|
6836
|
-
api: "anthropic-messages";
|
|
6837
|
-
provider: string;
|
|
6838
|
-
baseUrl: string;
|
|
6839
|
-
reasoning: false;
|
|
6840
|
-
input: ("image" | "text")[];
|
|
6841
|
-
cost: {
|
|
6842
|
-
input: number;
|
|
6843
|
-
output: number;
|
|
6844
|
-
cacheRead: number;
|
|
6845
|
-
cacheWrite: number;
|
|
6846
|
-
};
|
|
6847
|
-
contextWindow: number;
|
|
6848
|
-
maxTokens: number;
|
|
6849
|
-
};
|
|
6850
7147
|
readonly "claude-haiku-4-5": {
|
|
6851
7148
|
id: string;
|
|
6852
7149
|
name: string;
|
|
@@ -7501,6 +7798,9 @@ export declare const MODELS: {
|
|
|
7501
7798
|
api: "openai-completions";
|
|
7502
7799
|
provider: string;
|
|
7503
7800
|
baseUrl: string;
|
|
7801
|
+
compat: {
|
|
7802
|
+
requiresReasoningContentOnAssistantMessages: true;
|
|
7803
|
+
};
|
|
7504
7804
|
reasoning: true;
|
|
7505
7805
|
input: "text"[];
|
|
7506
7806
|
cost: {
|
|
@@ -7518,6 +7818,9 @@ export declare const MODELS: {
|
|
|
7518
7818
|
api: "openai-completions";
|
|
7519
7819
|
provider: string;
|
|
7520
7820
|
baseUrl: string;
|
|
7821
|
+
compat: {
|
|
7822
|
+
requiresReasoningContentOnAssistantMessages: true;
|
|
7823
|
+
};
|
|
7521
7824
|
reasoning: true;
|
|
7522
7825
|
input: "text"[];
|
|
7523
7826
|
cost: {
|
|
@@ -8472,6 +8775,9 @@ export declare const MODELS: {
|
|
|
8472
8775
|
api: "openai-completions";
|
|
8473
8776
|
provider: string;
|
|
8474
8777
|
baseUrl: string;
|
|
8778
|
+
compat: {
|
|
8779
|
+
requiresReasoningContentOnAssistantMessages: true;
|
|
8780
|
+
};
|
|
8475
8781
|
reasoning: true;
|
|
8476
8782
|
input: "text"[];
|
|
8477
8783
|
cost: {
|
|
@@ -8489,6 +8795,9 @@ export declare const MODELS: {
|
|
|
8489
8795
|
api: "openai-completions";
|
|
8490
8796
|
provider: string;
|
|
8491
8797
|
baseUrl: string;
|
|
8798
|
+
compat: {
|
|
8799
|
+
requiresReasoningContentOnAssistantMessages: true;
|
|
8800
|
+
};
|
|
8492
8801
|
reasoning: true;
|
|
8493
8802
|
input: "text"[];
|
|
8494
8803
|
cost: {
|
|
@@ -8857,7 +9166,7 @@ export declare const MODELS: {
|
|
|
8857
9166
|
contextWindow: number;
|
|
8858
9167
|
maxTokens: number;
|
|
8859
9168
|
};
|
|
8860
|
-
readonly "inclusionai/ling-2.6-flash
|
|
9169
|
+
readonly "inclusionai/ling-2.6-flash": {
|
|
8861
9170
|
id: string;
|
|
8862
9171
|
name: string;
|
|
8863
9172
|
api: "openai-completions";
|
|
@@ -9639,6 +9948,23 @@ export declare const MODELS: {
|
|
|
9639
9948
|
contextWindow: number;
|
|
9640
9949
|
maxTokens: number;
|
|
9641
9950
|
};
|
|
9951
|
+
readonly "nvidia/nemotron-3-nano-omni-30b-a3b-reasoning:free": {
|
|
9952
|
+
id: string;
|
|
9953
|
+
name: string;
|
|
9954
|
+
api: "openai-completions";
|
|
9955
|
+
provider: string;
|
|
9956
|
+
baseUrl: string;
|
|
9957
|
+
reasoning: true;
|
|
9958
|
+
input: ("image" | "text")[];
|
|
9959
|
+
cost: {
|
|
9960
|
+
input: number;
|
|
9961
|
+
output: number;
|
|
9962
|
+
cacheRead: number;
|
|
9963
|
+
cacheWrite: number;
|
|
9964
|
+
};
|
|
9965
|
+
contextWindow: number;
|
|
9966
|
+
maxTokens: number;
|
|
9967
|
+
};
|
|
9642
9968
|
readonly "nvidia/nemotron-3-super-120b-a12b": {
|
|
9643
9969
|
id: string;
|
|
9644
9970
|
name: string;
|
|
@@ -10710,6 +11036,40 @@ export declare const MODELS: {
|
|
|
10710
11036
|
contextWindow: number;
|
|
10711
11037
|
maxTokens: number;
|
|
10712
11038
|
};
|
|
11039
|
+
readonly "poolside/laguna-m.1:free": {
|
|
11040
|
+
id: string;
|
|
11041
|
+
name: string;
|
|
11042
|
+
api: "openai-completions";
|
|
11043
|
+
provider: string;
|
|
11044
|
+
baseUrl: string;
|
|
11045
|
+
reasoning: true;
|
|
11046
|
+
input: "text"[];
|
|
11047
|
+
cost: {
|
|
11048
|
+
input: number;
|
|
11049
|
+
output: number;
|
|
11050
|
+
cacheRead: number;
|
|
11051
|
+
cacheWrite: number;
|
|
11052
|
+
};
|
|
11053
|
+
contextWindow: number;
|
|
11054
|
+
maxTokens: number;
|
|
11055
|
+
};
|
|
11056
|
+
readonly "poolside/laguna-xs.2:free": {
|
|
11057
|
+
id: string;
|
|
11058
|
+
name: string;
|
|
11059
|
+
api: "openai-completions";
|
|
11060
|
+
provider: string;
|
|
11061
|
+
baseUrl: string;
|
|
11062
|
+
reasoning: true;
|
|
11063
|
+
input: "text"[];
|
|
11064
|
+
cost: {
|
|
11065
|
+
input: number;
|
|
11066
|
+
output: number;
|
|
11067
|
+
cacheRead: number;
|
|
11068
|
+
cacheWrite: number;
|
|
11069
|
+
};
|
|
11070
|
+
contextWindow: number;
|
|
11071
|
+
maxTokens: number;
|
|
11072
|
+
};
|
|
10713
11073
|
readonly "prime-intellect/intellect-3": {
|
|
10714
11074
|
id: string;
|
|
10715
11075
|
name: string;
|
|
@@ -11441,7 +11801,41 @@ export declare const MODELS: {
|
|
|
11441
11801
|
contextWindow: number;
|
|
11442
11802
|
maxTokens: number;
|
|
11443
11803
|
};
|
|
11444
|
-
readonly "qwen/qwen3.
|
|
11804
|
+
readonly "qwen/qwen3.5-plus-20260420": {
|
|
11805
|
+
id: string;
|
|
11806
|
+
name: string;
|
|
11807
|
+
api: "openai-completions";
|
|
11808
|
+
provider: string;
|
|
11809
|
+
baseUrl: string;
|
|
11810
|
+
reasoning: true;
|
|
11811
|
+
input: ("image" | "text")[];
|
|
11812
|
+
cost: {
|
|
11813
|
+
input: number;
|
|
11814
|
+
output: number;
|
|
11815
|
+
cacheRead: number;
|
|
11816
|
+
cacheWrite: number;
|
|
11817
|
+
};
|
|
11818
|
+
contextWindow: number;
|
|
11819
|
+
maxTokens: number;
|
|
11820
|
+
};
|
|
11821
|
+
readonly "qwen/qwen3.6-27b": {
|
|
11822
|
+
id: string;
|
|
11823
|
+
name: string;
|
|
11824
|
+
api: "openai-completions";
|
|
11825
|
+
provider: string;
|
|
11826
|
+
baseUrl: string;
|
|
11827
|
+
reasoning: true;
|
|
11828
|
+
input: ("image" | "text")[];
|
|
11829
|
+
cost: {
|
|
11830
|
+
input: number;
|
|
11831
|
+
output: number;
|
|
11832
|
+
cacheRead: number;
|
|
11833
|
+
cacheWrite: number;
|
|
11834
|
+
};
|
|
11835
|
+
contextWindow: number;
|
|
11836
|
+
maxTokens: number;
|
|
11837
|
+
};
|
|
11838
|
+
readonly "qwen/qwen3.6-flash": {
|
|
11445
11839
|
id: string;
|
|
11446
11840
|
name: string;
|
|
11447
11841
|
api: "openai-completions";
|
|
@@ -11458,7 +11852,7 @@ export declare const MODELS: {
|
|
|
11458
11852
|
contextWindow: number;
|
|
11459
11853
|
maxTokens: number;
|
|
11460
11854
|
};
|
|
11461
|
-
readonly "qwen/
|
|
11855
|
+
readonly "qwen/qwen3.6-max-preview": {
|
|
11462
11856
|
id: string;
|
|
11463
11857
|
name: string;
|
|
11464
11858
|
api: "openai-completions";
|
|
@@ -11475,6 +11869,23 @@ export declare const MODELS: {
|
|
|
11475
11869
|
contextWindow: number;
|
|
11476
11870
|
maxTokens: number;
|
|
11477
11871
|
};
|
|
11872
|
+
readonly "qwen/qwen3.6-plus": {
|
|
11873
|
+
id: string;
|
|
11874
|
+
name: string;
|
|
11875
|
+
api: "openai-completions";
|
|
11876
|
+
provider: string;
|
|
11877
|
+
baseUrl: string;
|
|
11878
|
+
reasoning: true;
|
|
11879
|
+
input: ("image" | "text")[];
|
|
11880
|
+
cost: {
|
|
11881
|
+
input: number;
|
|
11882
|
+
output: number;
|
|
11883
|
+
cacheRead: number;
|
|
11884
|
+
cacheWrite: number;
|
|
11885
|
+
};
|
|
11886
|
+
contextWindow: number;
|
|
11887
|
+
maxTokens: number;
|
|
11888
|
+
};
|
|
11478
11889
|
readonly "rekaai/reka-edge": {
|
|
11479
11890
|
id: string;
|
|
11480
11891
|
name: string;
|
|
@@ -11645,14 +12056,133 @@ export declare const MODELS: {
|
|
|
11645
12056
|
contextWindow: number;
|
|
11646
12057
|
maxTokens: number;
|
|
11647
12058
|
};
|
|
11648
|
-
readonly "x-ai/grok-3": {
|
|
12059
|
+
readonly "x-ai/grok-3": {
|
|
12060
|
+
id: string;
|
|
12061
|
+
name: string;
|
|
12062
|
+
api: "openai-completions";
|
|
12063
|
+
provider: string;
|
|
12064
|
+
baseUrl: string;
|
|
12065
|
+
reasoning: false;
|
|
12066
|
+
input: "text"[];
|
|
12067
|
+
cost: {
|
|
12068
|
+
input: number;
|
|
12069
|
+
output: number;
|
|
12070
|
+
cacheRead: number;
|
|
12071
|
+
cacheWrite: number;
|
|
12072
|
+
};
|
|
12073
|
+
contextWindow: number;
|
|
12074
|
+
maxTokens: number;
|
|
12075
|
+
};
|
|
12076
|
+
readonly "x-ai/grok-3-beta": {
|
|
12077
|
+
id: string;
|
|
12078
|
+
name: string;
|
|
12079
|
+
api: "openai-completions";
|
|
12080
|
+
provider: string;
|
|
12081
|
+
baseUrl: string;
|
|
12082
|
+
reasoning: false;
|
|
12083
|
+
input: "text"[];
|
|
12084
|
+
cost: {
|
|
12085
|
+
input: number;
|
|
12086
|
+
output: number;
|
|
12087
|
+
cacheRead: number;
|
|
12088
|
+
cacheWrite: number;
|
|
12089
|
+
};
|
|
12090
|
+
contextWindow: number;
|
|
12091
|
+
maxTokens: number;
|
|
12092
|
+
};
|
|
12093
|
+
readonly "x-ai/grok-3-mini": {
|
|
12094
|
+
id: string;
|
|
12095
|
+
name: string;
|
|
12096
|
+
api: "openai-completions";
|
|
12097
|
+
provider: string;
|
|
12098
|
+
baseUrl: string;
|
|
12099
|
+
reasoning: true;
|
|
12100
|
+
input: "text"[];
|
|
12101
|
+
cost: {
|
|
12102
|
+
input: number;
|
|
12103
|
+
output: number;
|
|
12104
|
+
cacheRead: number;
|
|
12105
|
+
cacheWrite: number;
|
|
12106
|
+
};
|
|
12107
|
+
contextWindow: number;
|
|
12108
|
+
maxTokens: number;
|
|
12109
|
+
};
|
|
12110
|
+
readonly "x-ai/grok-3-mini-beta": {
|
|
12111
|
+
id: string;
|
|
12112
|
+
name: string;
|
|
12113
|
+
api: "openai-completions";
|
|
12114
|
+
provider: string;
|
|
12115
|
+
baseUrl: string;
|
|
12116
|
+
reasoning: true;
|
|
12117
|
+
input: "text"[];
|
|
12118
|
+
cost: {
|
|
12119
|
+
input: number;
|
|
12120
|
+
output: number;
|
|
12121
|
+
cacheRead: number;
|
|
12122
|
+
cacheWrite: number;
|
|
12123
|
+
};
|
|
12124
|
+
contextWindow: number;
|
|
12125
|
+
maxTokens: number;
|
|
12126
|
+
};
|
|
12127
|
+
readonly "x-ai/grok-4": {
|
|
12128
|
+
id: string;
|
|
12129
|
+
name: string;
|
|
12130
|
+
api: "openai-completions";
|
|
12131
|
+
provider: string;
|
|
12132
|
+
baseUrl: string;
|
|
12133
|
+
reasoning: true;
|
|
12134
|
+
input: ("image" | "text")[];
|
|
12135
|
+
cost: {
|
|
12136
|
+
input: number;
|
|
12137
|
+
output: number;
|
|
12138
|
+
cacheRead: number;
|
|
12139
|
+
cacheWrite: number;
|
|
12140
|
+
};
|
|
12141
|
+
contextWindow: number;
|
|
12142
|
+
maxTokens: number;
|
|
12143
|
+
};
|
|
12144
|
+
readonly "x-ai/grok-4-fast": {
|
|
12145
|
+
id: string;
|
|
12146
|
+
name: string;
|
|
12147
|
+
api: "openai-completions";
|
|
12148
|
+
provider: string;
|
|
12149
|
+
baseUrl: string;
|
|
12150
|
+
reasoning: true;
|
|
12151
|
+
input: ("image" | "text")[];
|
|
12152
|
+
cost: {
|
|
12153
|
+
input: number;
|
|
12154
|
+
output: number;
|
|
12155
|
+
cacheRead: number;
|
|
12156
|
+
cacheWrite: number;
|
|
12157
|
+
};
|
|
12158
|
+
contextWindow: number;
|
|
12159
|
+
maxTokens: number;
|
|
12160
|
+
};
|
|
12161
|
+
readonly "x-ai/grok-4.1-fast": {
|
|
12162
|
+
id: string;
|
|
12163
|
+
name: string;
|
|
12164
|
+
api: "openai-completions";
|
|
12165
|
+
provider: string;
|
|
12166
|
+
baseUrl: string;
|
|
12167
|
+
reasoning: true;
|
|
12168
|
+
input: ("image" | "text")[];
|
|
12169
|
+
cost: {
|
|
12170
|
+
input: number;
|
|
12171
|
+
output: number;
|
|
12172
|
+
cacheRead: number;
|
|
12173
|
+
cacheWrite: number;
|
|
12174
|
+
};
|
|
12175
|
+
contextWindow: number;
|
|
12176
|
+
maxTokens: number;
|
|
12177
|
+
};
|
|
12178
|
+
readonly "x-ai/grok-4.20": {
|
|
11649
12179
|
id: string;
|
|
11650
12180
|
name: string;
|
|
11651
12181
|
api: "openai-completions";
|
|
11652
12182
|
provider: string;
|
|
11653
12183
|
baseUrl: string;
|
|
11654
|
-
reasoning:
|
|
11655
|
-
input: "text"[];
|
|
12184
|
+
reasoning: true;
|
|
12185
|
+
input: ("image" | "text")[];
|
|
11656
12186
|
cost: {
|
|
11657
12187
|
input: number;
|
|
11658
12188
|
output: number;
|
|
@@ -11662,13 +12192,13 @@ export declare const MODELS: {
|
|
|
11662
12192
|
contextWindow: number;
|
|
11663
12193
|
maxTokens: number;
|
|
11664
12194
|
};
|
|
11665
|
-
readonly "x-ai/grok-
|
|
12195
|
+
readonly "x-ai/grok-code-fast-1": {
|
|
11666
12196
|
id: string;
|
|
11667
12197
|
name: string;
|
|
11668
12198
|
api: "openai-completions";
|
|
11669
12199
|
provider: string;
|
|
11670
12200
|
baseUrl: string;
|
|
11671
|
-
reasoning:
|
|
12201
|
+
reasoning: true;
|
|
11672
12202
|
input: "text"[];
|
|
11673
12203
|
cost: {
|
|
11674
12204
|
input: number;
|
|
@@ -11679,7 +12209,7 @@ export declare const MODELS: {
|
|
|
11679
12209
|
contextWindow: number;
|
|
11680
12210
|
maxTokens: number;
|
|
11681
12211
|
};
|
|
11682
|
-
readonly "
|
|
12212
|
+
readonly "xiaomi/mimo-v2-flash": {
|
|
11683
12213
|
id: string;
|
|
11684
12214
|
name: string;
|
|
11685
12215
|
api: "openai-completions";
|
|
@@ -11696,14 +12226,14 @@ export declare const MODELS: {
|
|
|
11696
12226
|
contextWindow: number;
|
|
11697
12227
|
maxTokens: number;
|
|
11698
12228
|
};
|
|
11699
|
-
readonly "
|
|
12229
|
+
readonly "xiaomi/mimo-v2-omni": {
|
|
11700
12230
|
id: string;
|
|
11701
12231
|
name: string;
|
|
11702
12232
|
api: "openai-completions";
|
|
11703
12233
|
provider: string;
|
|
11704
12234
|
baseUrl: string;
|
|
11705
12235
|
reasoning: true;
|
|
11706
|
-
input: "text"[];
|
|
12236
|
+
input: ("image" | "text")[];
|
|
11707
12237
|
cost: {
|
|
11708
12238
|
input: number;
|
|
11709
12239
|
output: number;
|
|
@@ -11713,14 +12243,14 @@ export declare const MODELS: {
|
|
|
11713
12243
|
contextWindow: number;
|
|
11714
12244
|
maxTokens: number;
|
|
11715
12245
|
};
|
|
11716
|
-
readonly "
|
|
12246
|
+
readonly "xiaomi/mimo-v2-pro": {
|
|
11717
12247
|
id: string;
|
|
11718
12248
|
name: string;
|
|
11719
12249
|
api: "openai-completions";
|
|
11720
12250
|
provider: string;
|
|
11721
12251
|
baseUrl: string;
|
|
11722
12252
|
reasoning: true;
|
|
11723
|
-
input:
|
|
12253
|
+
input: "text"[];
|
|
11724
12254
|
cost: {
|
|
11725
12255
|
input: number;
|
|
11726
12256
|
output: number;
|
|
@@ -11730,7 +12260,7 @@ export declare const MODELS: {
|
|
|
11730
12260
|
contextWindow: number;
|
|
11731
12261
|
maxTokens: number;
|
|
11732
12262
|
};
|
|
11733
|
-
readonly "
|
|
12263
|
+
readonly "xiaomi/mimo-v2.5": {
|
|
11734
12264
|
id: string;
|
|
11735
12265
|
name: string;
|
|
11736
12266
|
api: "openai-completions";
|
|
@@ -11747,14 +12277,14 @@ export declare const MODELS: {
|
|
|
11747
12277
|
contextWindow: number;
|
|
11748
12278
|
maxTokens: number;
|
|
11749
12279
|
};
|
|
11750
|
-
readonly "
|
|
12280
|
+
readonly "xiaomi/mimo-v2.5-pro": {
|
|
11751
12281
|
id: string;
|
|
11752
12282
|
name: string;
|
|
11753
12283
|
api: "openai-completions";
|
|
11754
12284
|
provider: string;
|
|
11755
12285
|
baseUrl: string;
|
|
11756
12286
|
reasoning: true;
|
|
11757
|
-
input:
|
|
12287
|
+
input: "text"[];
|
|
11758
12288
|
cost: {
|
|
11759
12289
|
input: number;
|
|
11760
12290
|
output: number;
|
|
@@ -11764,14 +12294,14 @@ export declare const MODELS: {
|
|
|
11764
12294
|
contextWindow: number;
|
|
11765
12295
|
maxTokens: number;
|
|
11766
12296
|
};
|
|
11767
|
-
readonly "
|
|
12297
|
+
readonly "z-ai/glm-4-32b": {
|
|
11768
12298
|
id: string;
|
|
11769
12299
|
name: string;
|
|
11770
12300
|
api: "openai-completions";
|
|
11771
12301
|
provider: string;
|
|
11772
12302
|
baseUrl: string;
|
|
11773
|
-
reasoning:
|
|
11774
|
-
input:
|
|
12303
|
+
reasoning: false;
|
|
12304
|
+
input: "text"[];
|
|
11775
12305
|
cost: {
|
|
11776
12306
|
input: number;
|
|
11777
12307
|
output: number;
|
|
@@ -11781,7 +12311,7 @@ export declare const MODELS: {
|
|
|
11781
12311
|
contextWindow: number;
|
|
11782
12312
|
maxTokens: number;
|
|
11783
12313
|
};
|
|
11784
|
-
readonly "
|
|
12314
|
+
readonly "z-ai/glm-4.5": {
|
|
11785
12315
|
id: string;
|
|
11786
12316
|
name: string;
|
|
11787
12317
|
api: "openai-completions";
|
|
@@ -11798,7 +12328,7 @@ export declare const MODELS: {
|
|
|
11798
12328
|
contextWindow: number;
|
|
11799
12329
|
maxTokens: number;
|
|
11800
12330
|
};
|
|
11801
|
-
readonly "
|
|
12331
|
+
readonly "z-ai/glm-4.5-air": {
|
|
11802
12332
|
id: string;
|
|
11803
12333
|
name: string;
|
|
11804
12334
|
api: "openai-completions";
|
|
@@ -11815,14 +12345,14 @@ export declare const MODELS: {
|
|
|
11815
12345
|
contextWindow: number;
|
|
11816
12346
|
maxTokens: number;
|
|
11817
12347
|
};
|
|
11818
|
-
readonly "
|
|
12348
|
+
readonly "z-ai/glm-4.5-air:free": {
|
|
11819
12349
|
id: string;
|
|
11820
12350
|
name: string;
|
|
11821
12351
|
api: "openai-completions";
|
|
11822
12352
|
provider: string;
|
|
11823
12353
|
baseUrl: string;
|
|
11824
12354
|
reasoning: true;
|
|
11825
|
-
input:
|
|
12355
|
+
input: "text"[];
|
|
11826
12356
|
cost: {
|
|
11827
12357
|
input: number;
|
|
11828
12358
|
output: number;
|
|
@@ -11832,14 +12362,14 @@ export declare const MODELS: {
|
|
|
11832
12362
|
contextWindow: number;
|
|
11833
12363
|
maxTokens: number;
|
|
11834
12364
|
};
|
|
11835
|
-
readonly "
|
|
12365
|
+
readonly "z-ai/glm-4.5v": {
|
|
11836
12366
|
id: string;
|
|
11837
12367
|
name: string;
|
|
11838
12368
|
api: "openai-completions";
|
|
11839
12369
|
provider: string;
|
|
11840
12370
|
baseUrl: string;
|
|
11841
12371
|
reasoning: true;
|
|
11842
|
-
input: "text"[];
|
|
12372
|
+
input: ("image" | "text")[];
|
|
11843
12373
|
cost: {
|
|
11844
12374
|
input: number;
|
|
11845
12375
|
output: number;
|
|
@@ -11849,14 +12379,14 @@ export declare const MODELS: {
|
|
|
11849
12379
|
contextWindow: number;
|
|
11850
12380
|
maxTokens: number;
|
|
11851
12381
|
};
|
|
11852
|
-
readonly "
|
|
12382
|
+
readonly "z-ai/glm-4.6": {
|
|
11853
12383
|
id: string;
|
|
11854
12384
|
name: string;
|
|
11855
12385
|
api: "openai-completions";
|
|
11856
12386
|
provider: string;
|
|
11857
12387
|
baseUrl: string;
|
|
11858
12388
|
reasoning: true;
|
|
11859
|
-
input:
|
|
12389
|
+
input: "text"[];
|
|
11860
12390
|
cost: {
|
|
11861
12391
|
input: number;
|
|
11862
12392
|
output: number;
|
|
@@ -11866,14 +12396,14 @@ export declare const MODELS: {
|
|
|
11866
12396
|
contextWindow: number;
|
|
11867
12397
|
maxTokens: number;
|
|
11868
12398
|
};
|
|
11869
|
-
readonly "
|
|
12399
|
+
readonly "z-ai/glm-4.6v": {
|
|
11870
12400
|
id: string;
|
|
11871
12401
|
name: string;
|
|
11872
12402
|
api: "openai-completions";
|
|
11873
12403
|
provider: string;
|
|
11874
12404
|
baseUrl: string;
|
|
11875
12405
|
reasoning: true;
|
|
11876
|
-
input: "text"[];
|
|
12406
|
+
input: ("image" | "text")[];
|
|
11877
12407
|
cost: {
|
|
11878
12408
|
input: number;
|
|
11879
12409
|
output: number;
|
|
@@ -11883,13 +12413,13 @@ export declare const MODELS: {
|
|
|
11883
12413
|
contextWindow: number;
|
|
11884
12414
|
maxTokens: number;
|
|
11885
12415
|
};
|
|
11886
|
-
readonly "z-ai/glm-4
|
|
12416
|
+
readonly "z-ai/glm-4.7": {
|
|
11887
12417
|
id: string;
|
|
11888
12418
|
name: string;
|
|
11889
12419
|
api: "openai-completions";
|
|
11890
12420
|
provider: string;
|
|
11891
12421
|
baseUrl: string;
|
|
11892
|
-
reasoning:
|
|
12422
|
+
reasoning: true;
|
|
11893
12423
|
input: "text"[];
|
|
11894
12424
|
cost: {
|
|
11895
12425
|
input: number;
|
|
@@ -11900,7 +12430,7 @@ export declare const MODELS: {
|
|
|
11900
12430
|
contextWindow: number;
|
|
11901
12431
|
maxTokens: number;
|
|
11902
12432
|
};
|
|
11903
|
-
readonly "z-ai/glm-4.
|
|
12433
|
+
readonly "z-ai/glm-4.7-flash": {
|
|
11904
12434
|
id: string;
|
|
11905
12435
|
name: string;
|
|
11906
12436
|
api: "openai-completions";
|
|
@@ -11917,7 +12447,7 @@ export declare const MODELS: {
|
|
|
11917
12447
|
contextWindow: number;
|
|
11918
12448
|
maxTokens: number;
|
|
11919
12449
|
};
|
|
11920
|
-
readonly "z-ai/glm-
|
|
12450
|
+
readonly "z-ai/glm-5": {
|
|
11921
12451
|
id: string;
|
|
11922
12452
|
name: string;
|
|
11923
12453
|
api: "openai-completions";
|
|
@@ -11934,7 +12464,7 @@ export declare const MODELS: {
|
|
|
11934
12464
|
contextWindow: number;
|
|
11935
12465
|
maxTokens: number;
|
|
11936
12466
|
};
|
|
11937
|
-
readonly "z-ai/glm-
|
|
12467
|
+
readonly "z-ai/glm-5-turbo": {
|
|
11938
12468
|
id: string;
|
|
11939
12469
|
name: string;
|
|
11940
12470
|
api: "openai-completions";
|
|
@@ -11951,14 +12481,14 @@ export declare const MODELS: {
|
|
|
11951
12481
|
contextWindow: number;
|
|
11952
12482
|
maxTokens: number;
|
|
11953
12483
|
};
|
|
11954
|
-
readonly "z-ai/glm-
|
|
12484
|
+
readonly "z-ai/glm-5.1": {
|
|
11955
12485
|
id: string;
|
|
11956
12486
|
name: string;
|
|
11957
12487
|
api: "openai-completions";
|
|
11958
12488
|
provider: string;
|
|
11959
12489
|
baseUrl: string;
|
|
11960
12490
|
reasoning: true;
|
|
11961
|
-
input:
|
|
12491
|
+
input: "text"[];
|
|
11962
12492
|
cost: {
|
|
11963
12493
|
input: number;
|
|
11964
12494
|
output: number;
|
|
@@ -11968,14 +12498,14 @@ export declare const MODELS: {
|
|
|
11968
12498
|
contextWindow: number;
|
|
11969
12499
|
maxTokens: number;
|
|
11970
12500
|
};
|
|
11971
|
-
readonly "z-ai/glm-
|
|
12501
|
+
readonly "z-ai/glm-5v-turbo": {
|
|
11972
12502
|
id: string;
|
|
11973
12503
|
name: string;
|
|
11974
12504
|
api: "openai-completions";
|
|
11975
12505
|
provider: string;
|
|
11976
12506
|
baseUrl: string;
|
|
11977
12507
|
reasoning: true;
|
|
11978
|
-
input: "text"[];
|
|
12508
|
+
input: ("image" | "text")[];
|
|
11979
12509
|
cost: {
|
|
11980
12510
|
input: number;
|
|
11981
12511
|
output: number;
|
|
@@ -11985,7 +12515,7 @@ export declare const MODELS: {
|
|
|
11985
12515
|
contextWindow: number;
|
|
11986
12516
|
maxTokens: number;
|
|
11987
12517
|
};
|
|
11988
|
-
readonly "
|
|
12518
|
+
readonly "~anthropic/claude-haiku-latest": {
|
|
11989
12519
|
id: string;
|
|
11990
12520
|
name: string;
|
|
11991
12521
|
api: "openai-completions";
|
|
@@ -12002,14 +12532,14 @@ export declare const MODELS: {
|
|
|
12002
12532
|
contextWindow: number;
|
|
12003
12533
|
maxTokens: number;
|
|
12004
12534
|
};
|
|
12005
|
-
readonly "
|
|
12535
|
+
readonly "~anthropic/claude-opus-latest": {
|
|
12006
12536
|
id: string;
|
|
12007
12537
|
name: string;
|
|
12008
12538
|
api: "openai-completions";
|
|
12009
12539
|
provider: string;
|
|
12010
12540
|
baseUrl: string;
|
|
12011
12541
|
reasoning: true;
|
|
12012
|
-
input: "text"[];
|
|
12542
|
+
input: ("image" | "text")[];
|
|
12013
12543
|
cost: {
|
|
12014
12544
|
input: number;
|
|
12015
12545
|
output: number;
|
|
@@ -12019,14 +12549,14 @@ export declare const MODELS: {
|
|
|
12019
12549
|
contextWindow: number;
|
|
12020
12550
|
maxTokens: number;
|
|
12021
12551
|
};
|
|
12022
|
-
readonly "
|
|
12552
|
+
readonly "~anthropic/claude-sonnet-latest": {
|
|
12023
12553
|
id: string;
|
|
12024
12554
|
name: string;
|
|
12025
12555
|
api: "openai-completions";
|
|
12026
12556
|
provider: string;
|
|
12027
12557
|
baseUrl: string;
|
|
12028
12558
|
reasoning: true;
|
|
12029
|
-
input: "text"[];
|
|
12559
|
+
input: ("image" | "text")[];
|
|
12030
12560
|
cost: {
|
|
12031
12561
|
input: number;
|
|
12032
12562
|
output: number;
|
|
@@ -12036,14 +12566,14 @@ export declare const MODELS: {
|
|
|
12036
12566
|
contextWindow: number;
|
|
12037
12567
|
maxTokens: number;
|
|
12038
12568
|
};
|
|
12039
|
-
readonly "
|
|
12569
|
+
readonly "~google/gemini-flash-latest": {
|
|
12040
12570
|
id: string;
|
|
12041
12571
|
name: string;
|
|
12042
12572
|
api: "openai-completions";
|
|
12043
12573
|
provider: string;
|
|
12044
12574
|
baseUrl: string;
|
|
12045
12575
|
reasoning: true;
|
|
12046
|
-
input: "text"[];
|
|
12576
|
+
input: ("image" | "text")[];
|
|
12047
12577
|
cost: {
|
|
12048
12578
|
input: number;
|
|
12049
12579
|
output: number;
|
|
@@ -12053,14 +12583,14 @@ export declare const MODELS: {
|
|
|
12053
12583
|
contextWindow: number;
|
|
12054
12584
|
maxTokens: number;
|
|
12055
12585
|
};
|
|
12056
|
-
readonly "
|
|
12586
|
+
readonly "~google/gemini-pro-latest": {
|
|
12057
12587
|
id: string;
|
|
12058
12588
|
name: string;
|
|
12059
12589
|
api: "openai-completions";
|
|
12060
12590
|
provider: string;
|
|
12061
12591
|
baseUrl: string;
|
|
12062
12592
|
reasoning: true;
|
|
12063
|
-
input: "text"[];
|
|
12593
|
+
input: ("image" | "text")[];
|
|
12064
12594
|
cost: {
|
|
12065
12595
|
input: number;
|
|
12066
12596
|
output: number;
|
|
@@ -12070,14 +12600,14 @@ export declare const MODELS: {
|
|
|
12070
12600
|
contextWindow: number;
|
|
12071
12601
|
maxTokens: number;
|
|
12072
12602
|
};
|
|
12073
|
-
readonly "
|
|
12603
|
+
readonly "~moonshotai/kimi-latest": {
|
|
12074
12604
|
id: string;
|
|
12075
12605
|
name: string;
|
|
12076
12606
|
api: "openai-completions";
|
|
12077
12607
|
provider: string;
|
|
12078
12608
|
baseUrl: string;
|
|
12079
12609
|
reasoning: true;
|
|
12080
|
-
input: "text"[];
|
|
12610
|
+
input: ("image" | "text")[];
|
|
12081
12611
|
cost: {
|
|
12082
12612
|
input: number;
|
|
12083
12613
|
output: number;
|
|
@@ -12087,7 +12617,7 @@ export declare const MODELS: {
|
|
|
12087
12617
|
contextWindow: number;
|
|
12088
12618
|
maxTokens: number;
|
|
12089
12619
|
};
|
|
12090
|
-
readonly "
|
|
12620
|
+
readonly "~openai/gpt-latest": {
|
|
12091
12621
|
id: string;
|
|
12092
12622
|
name: string;
|
|
12093
12623
|
api: "openai-completions";
|
|
@@ -12104,7 +12634,7 @@ export declare const MODELS: {
|
|
|
12104
12634
|
contextWindow: number;
|
|
12105
12635
|
maxTokens: number;
|
|
12106
12636
|
};
|
|
12107
|
-
readonly "~
|
|
12637
|
+
readonly "~openai/gpt-mini-latest": {
|
|
12108
12638
|
id: string;
|
|
12109
12639
|
name: string;
|
|
12110
12640
|
api: "openai-completions";
|
|
@@ -15255,27 +15785,6 @@ export declare const MODELS: {
|
|
|
15255
15785
|
};
|
|
15256
15786
|
};
|
|
15257
15787
|
readonly zai: {
|
|
15258
|
-
readonly "glm-4.5": {
|
|
15259
|
-
id: string;
|
|
15260
|
-
name: string;
|
|
15261
|
-
api: "openai-completions";
|
|
15262
|
-
provider: string;
|
|
15263
|
-
baseUrl: string;
|
|
15264
|
-
compat: {
|
|
15265
|
-
supportsDeveloperRole: false;
|
|
15266
|
-
thinkingFormat: "zai";
|
|
15267
|
-
};
|
|
15268
|
-
reasoning: true;
|
|
15269
|
-
input: "text"[];
|
|
15270
|
-
cost: {
|
|
15271
|
-
input: number;
|
|
15272
|
-
output: number;
|
|
15273
|
-
cacheRead: number;
|
|
15274
|
-
cacheWrite: number;
|
|
15275
|
-
};
|
|
15276
|
-
contextWindow: number;
|
|
15277
|
-
maxTokens: number;
|
|
15278
|
-
};
|
|
15279
15788
|
readonly "glm-4.5-air": {
|
|
15280
15789
|
id: string;
|
|
15281
15790
|
name: string;
|
|
@@ -15297,92 +15806,6 @@ export declare const MODELS: {
|
|
|
15297
15806
|
contextWindow: number;
|
|
15298
15807
|
maxTokens: number;
|
|
15299
15808
|
};
|
|
15300
|
-
readonly "glm-4.5-flash": {
|
|
15301
|
-
id: string;
|
|
15302
|
-
name: string;
|
|
15303
|
-
api: "openai-completions";
|
|
15304
|
-
provider: string;
|
|
15305
|
-
baseUrl: string;
|
|
15306
|
-
compat: {
|
|
15307
|
-
supportsDeveloperRole: false;
|
|
15308
|
-
thinkingFormat: "zai";
|
|
15309
|
-
};
|
|
15310
|
-
reasoning: true;
|
|
15311
|
-
input: "text"[];
|
|
15312
|
-
cost: {
|
|
15313
|
-
input: number;
|
|
15314
|
-
output: number;
|
|
15315
|
-
cacheRead: number;
|
|
15316
|
-
cacheWrite: number;
|
|
15317
|
-
};
|
|
15318
|
-
contextWindow: number;
|
|
15319
|
-
maxTokens: number;
|
|
15320
|
-
};
|
|
15321
|
-
readonly "glm-4.5v": {
|
|
15322
|
-
id: string;
|
|
15323
|
-
name: string;
|
|
15324
|
-
api: "openai-completions";
|
|
15325
|
-
provider: string;
|
|
15326
|
-
baseUrl: string;
|
|
15327
|
-
compat: {
|
|
15328
|
-
supportsDeveloperRole: false;
|
|
15329
|
-
thinkingFormat: "zai";
|
|
15330
|
-
};
|
|
15331
|
-
reasoning: true;
|
|
15332
|
-
input: ("image" | "text")[];
|
|
15333
|
-
cost: {
|
|
15334
|
-
input: number;
|
|
15335
|
-
output: number;
|
|
15336
|
-
cacheRead: number;
|
|
15337
|
-
cacheWrite: number;
|
|
15338
|
-
};
|
|
15339
|
-
contextWindow: number;
|
|
15340
|
-
maxTokens: number;
|
|
15341
|
-
};
|
|
15342
|
-
readonly "glm-4.6": {
|
|
15343
|
-
id: string;
|
|
15344
|
-
name: string;
|
|
15345
|
-
api: "openai-completions";
|
|
15346
|
-
provider: string;
|
|
15347
|
-
baseUrl: string;
|
|
15348
|
-
compat: {
|
|
15349
|
-
supportsDeveloperRole: false;
|
|
15350
|
-
thinkingFormat: "zai";
|
|
15351
|
-
zaiToolStream: true;
|
|
15352
|
-
};
|
|
15353
|
-
reasoning: true;
|
|
15354
|
-
input: "text"[];
|
|
15355
|
-
cost: {
|
|
15356
|
-
input: number;
|
|
15357
|
-
output: number;
|
|
15358
|
-
cacheRead: number;
|
|
15359
|
-
cacheWrite: number;
|
|
15360
|
-
};
|
|
15361
|
-
contextWindow: number;
|
|
15362
|
-
maxTokens: number;
|
|
15363
|
-
};
|
|
15364
|
-
readonly "glm-4.6v": {
|
|
15365
|
-
id: string;
|
|
15366
|
-
name: string;
|
|
15367
|
-
api: "openai-completions";
|
|
15368
|
-
provider: string;
|
|
15369
|
-
baseUrl: string;
|
|
15370
|
-
compat: {
|
|
15371
|
-
supportsDeveloperRole: false;
|
|
15372
|
-
thinkingFormat: "zai";
|
|
15373
|
-
zaiToolStream: true;
|
|
15374
|
-
};
|
|
15375
|
-
reasoning: true;
|
|
15376
|
-
input: ("image" | "text")[];
|
|
15377
|
-
cost: {
|
|
15378
|
-
input: number;
|
|
15379
|
-
output: number;
|
|
15380
|
-
cacheRead: number;
|
|
15381
|
-
cacheWrite: number;
|
|
15382
|
-
};
|
|
15383
|
-
contextWindow: number;
|
|
15384
|
-
maxTokens: number;
|
|
15385
|
-
};
|
|
15386
15809
|
readonly "glm-4.7": {
|
|
15387
15810
|
id: string;
|
|
15388
15811
|
name: string;
|
|
@@ -15405,72 +15828,6 @@ export declare const MODELS: {
|
|
|
15405
15828
|
contextWindow: number;
|
|
15406
15829
|
maxTokens: number;
|
|
15407
15830
|
};
|
|
15408
|
-
readonly "glm-4.7-flash": {
|
|
15409
|
-
id: string;
|
|
15410
|
-
name: string;
|
|
15411
|
-
api: "openai-completions";
|
|
15412
|
-
provider: string;
|
|
15413
|
-
baseUrl: string;
|
|
15414
|
-
compat: {
|
|
15415
|
-
supportsDeveloperRole: false;
|
|
15416
|
-
thinkingFormat: "zai";
|
|
15417
|
-
zaiToolStream: true;
|
|
15418
|
-
};
|
|
15419
|
-
reasoning: true;
|
|
15420
|
-
input: "text"[];
|
|
15421
|
-
cost: {
|
|
15422
|
-
input: number;
|
|
15423
|
-
output: number;
|
|
15424
|
-
cacheRead: number;
|
|
15425
|
-
cacheWrite: number;
|
|
15426
|
-
};
|
|
15427
|
-
contextWindow: number;
|
|
15428
|
-
maxTokens: number;
|
|
15429
|
-
};
|
|
15430
|
-
readonly "glm-4.7-flashx": {
|
|
15431
|
-
id: string;
|
|
15432
|
-
name: string;
|
|
15433
|
-
api: "openai-completions";
|
|
15434
|
-
provider: string;
|
|
15435
|
-
baseUrl: string;
|
|
15436
|
-
compat: {
|
|
15437
|
-
supportsDeveloperRole: false;
|
|
15438
|
-
thinkingFormat: "zai";
|
|
15439
|
-
zaiToolStream: true;
|
|
15440
|
-
};
|
|
15441
|
-
reasoning: true;
|
|
15442
|
-
input: "text"[];
|
|
15443
|
-
cost: {
|
|
15444
|
-
input: number;
|
|
15445
|
-
output: number;
|
|
15446
|
-
cacheRead: number;
|
|
15447
|
-
cacheWrite: number;
|
|
15448
|
-
};
|
|
15449
|
-
contextWindow: number;
|
|
15450
|
-
maxTokens: number;
|
|
15451
|
-
};
|
|
15452
|
-
readonly "glm-5": {
|
|
15453
|
-
id: string;
|
|
15454
|
-
name: string;
|
|
15455
|
-
api: "openai-completions";
|
|
15456
|
-
provider: string;
|
|
15457
|
-
baseUrl: string;
|
|
15458
|
-
compat: {
|
|
15459
|
-
supportsDeveloperRole: false;
|
|
15460
|
-
thinkingFormat: "zai";
|
|
15461
|
-
zaiToolStream: true;
|
|
15462
|
-
};
|
|
15463
|
-
reasoning: true;
|
|
15464
|
-
input: "text"[];
|
|
15465
|
-
cost: {
|
|
15466
|
-
input: number;
|
|
15467
|
-
output: number;
|
|
15468
|
-
cacheRead: number;
|
|
15469
|
-
cacheWrite: number;
|
|
15470
|
-
};
|
|
15471
|
-
contextWindow: number;
|
|
15472
|
-
maxTokens: number;
|
|
15473
|
-
};
|
|
15474
15831
|
readonly "glm-5-turbo": {
|
|
15475
15832
|
id: string;
|
|
15476
15833
|
name: string;
|