@hyperspaceng/neural-ai 0.70.1 → 0.70.6
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +9 -4
- package/dist/env-api-keys.d.ts.map +1 -1
- package/dist/env-api-keys.js +51 -7
- package/dist/env-api-keys.js.map +1 -1
- package/dist/models.d.ts +1 -0
- package/dist/models.d.ts.map +1 -1
- package/dist/models.generated.d.ts +606 -249
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +539 -204
- package/dist/models.generated.js.map +1 -1
- package/dist/models.js +3 -1
- package/dist/models.js.map +1 -1
- package/dist/providers/amazon-bedrock.d.ts.map +1 -1
- package/dist/providers/amazon-bedrock.js +49 -25
- package/dist/providers/amazon-bedrock.js.map +1 -1
- package/dist/providers/anthropic.d.ts.map +1 -1
- package/dist/providers/anthropic.js +18 -7
- package/dist/providers/anthropic.js.map +1 -1
- package/dist/providers/azure-openai-responses.d.ts.map +1 -1
- package/dist/providers/azure-openai-responses.js +24 -5
- package/dist/providers/azure-openai-responses.js.map +1 -1
- package/dist/providers/cloudflare.d.ts +7 -0
- package/dist/providers/cloudflare.d.ts.map +1 -0
- package/dist/providers/cloudflare.js +19 -0
- package/dist/providers/cloudflare.js.map +1 -0
- package/dist/providers/openai-codex-responses.d.ts.map +1 -1
- package/dist/providers/openai-codex-responses.js +2 -2
- package/dist/providers/openai-codex-responses.js.map +1 -1
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +49 -16
- package/dist/providers/openai-completions.js.map +1 -1
- package/dist/providers/openai-responses.d.ts.map +1 -1
- package/dist/providers/openai-responses.js +7 -4
- package/dist/providers/openai-responses.js.map +1 -1
- package/dist/providers/simple-options.d.ts.map +1 -1
- package/dist/providers/simple-options.js +2 -0
- package/dist/providers/simple-options.js.map +1 -1
- package/dist/types.d.ts +15 -3
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/package.json +1 -1
package/dist/models.generated.js
CHANGED
|
@@ -2536,8 +2536,25 @@ export const MODELS = {
|
|
|
2536
2536
|
cacheRead: 0.5,
|
|
2537
2537
|
cacheWrite: 0,
|
|
2538
2538
|
},
|
|
2539
|
+
contextWindow: 272000,
|
|
2540
|
+
maxTokens: 128000,
|
|
2541
|
+
},
|
|
2542
|
+
"gpt-5.5-pro": {
|
|
2543
|
+
id: "gpt-5.5-pro",
|
|
2544
|
+
name: "GPT-5.5 Pro",
|
|
2545
|
+
api: "azure-openai-responses",
|
|
2546
|
+
provider: "azure-openai-responses",
|
|
2547
|
+
baseUrl: "",
|
|
2548
|
+
reasoning: true,
|
|
2549
|
+
input: ["text", "image"],
|
|
2550
|
+
cost: {
|
|
2551
|
+
input: 30,
|
|
2552
|
+
output: 180,
|
|
2553
|
+
cacheRead: 0,
|
|
2554
|
+
cacheWrite: 0,
|
|
2555
|
+
},
|
|
2539
2556
|
contextWindow: 1050000,
|
|
2540
|
-
maxTokens:
|
|
2557
|
+
maxTokens: 128000,
|
|
2541
2558
|
},
|
|
2542
2559
|
"o1": {
|
|
2543
2560
|
id: "o1",
|
|
@@ -2746,6 +2763,190 @@ export const MODELS = {
|
|
|
2746
2763
|
maxTokens: 40000,
|
|
2747
2764
|
},
|
|
2748
2765
|
},
|
|
2766
|
+
"cloudflare-workers-ai": {
|
|
2767
|
+
"@cf/google/gemma-4-26b-a4b-it": {
|
|
2768
|
+
id: "@cf/google/gemma-4-26b-a4b-it",
|
|
2769
|
+
name: "Gemma 4 26B A4B IT",
|
|
2770
|
+
api: "openai-completions",
|
|
2771
|
+
provider: "cloudflare-workers-ai",
|
|
2772
|
+
baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
|
|
2773
|
+
compat: { "sendSessionAffinityHeaders": true },
|
|
2774
|
+
reasoning: true,
|
|
2775
|
+
input: ["text", "image"],
|
|
2776
|
+
cost: {
|
|
2777
|
+
input: 0.1,
|
|
2778
|
+
output: 0.3,
|
|
2779
|
+
cacheRead: 0,
|
|
2780
|
+
cacheWrite: 0,
|
|
2781
|
+
},
|
|
2782
|
+
contextWindow: 256000,
|
|
2783
|
+
maxTokens: 16384,
|
|
2784
|
+
},
|
|
2785
|
+
"@cf/meta/llama-4-scout-17b-16e-instruct": {
|
|
2786
|
+
id: "@cf/meta/llama-4-scout-17b-16e-instruct",
|
|
2787
|
+
name: "Llama 4 Scout 17B 16E Instruct",
|
|
2788
|
+
api: "openai-completions",
|
|
2789
|
+
provider: "cloudflare-workers-ai",
|
|
2790
|
+
baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
|
|
2791
|
+
compat: { "sendSessionAffinityHeaders": true },
|
|
2792
|
+
reasoning: false,
|
|
2793
|
+
input: ["text", "image"],
|
|
2794
|
+
cost: {
|
|
2795
|
+
input: 0.27,
|
|
2796
|
+
output: 0.85,
|
|
2797
|
+
cacheRead: 0,
|
|
2798
|
+
cacheWrite: 0,
|
|
2799
|
+
},
|
|
2800
|
+
contextWindow: 128000,
|
|
2801
|
+
maxTokens: 16384,
|
|
2802
|
+
},
|
|
2803
|
+
"@cf/moonshotai/kimi-k2.5": {
|
|
2804
|
+
id: "@cf/moonshotai/kimi-k2.5",
|
|
2805
|
+
name: "Kimi K2.5",
|
|
2806
|
+
api: "openai-completions",
|
|
2807
|
+
provider: "cloudflare-workers-ai",
|
|
2808
|
+
baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
|
|
2809
|
+
compat: { "sendSessionAffinityHeaders": true },
|
|
2810
|
+
reasoning: true,
|
|
2811
|
+
input: ["text", "image"],
|
|
2812
|
+
cost: {
|
|
2813
|
+
input: 0.6,
|
|
2814
|
+
output: 3,
|
|
2815
|
+
cacheRead: 0.1,
|
|
2816
|
+
cacheWrite: 0,
|
|
2817
|
+
},
|
|
2818
|
+
contextWindow: 256000,
|
|
2819
|
+
maxTokens: 256000,
|
|
2820
|
+
},
|
|
2821
|
+
"@cf/moonshotai/kimi-k2.6": {
|
|
2822
|
+
id: "@cf/moonshotai/kimi-k2.6",
|
|
2823
|
+
name: "Kimi K2.6",
|
|
2824
|
+
api: "openai-completions",
|
|
2825
|
+
provider: "cloudflare-workers-ai",
|
|
2826
|
+
baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
|
|
2827
|
+
compat: { "sendSessionAffinityHeaders": true },
|
|
2828
|
+
reasoning: true,
|
|
2829
|
+
input: ["text", "image"],
|
|
2830
|
+
cost: {
|
|
2831
|
+
input: 0.95,
|
|
2832
|
+
output: 4,
|
|
2833
|
+
cacheRead: 0.16,
|
|
2834
|
+
cacheWrite: 0,
|
|
2835
|
+
},
|
|
2836
|
+
contextWindow: 256000,
|
|
2837
|
+
maxTokens: 256000,
|
|
2838
|
+
},
|
|
2839
|
+
"@cf/nvidia/nemotron-3-120b-a12b": {
|
|
2840
|
+
id: "@cf/nvidia/nemotron-3-120b-a12b",
|
|
2841
|
+
name: "Nemotron 3 Super 120B",
|
|
2842
|
+
api: "openai-completions",
|
|
2843
|
+
provider: "cloudflare-workers-ai",
|
|
2844
|
+
baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
|
|
2845
|
+
compat: { "sendSessionAffinityHeaders": true },
|
|
2846
|
+
reasoning: true,
|
|
2847
|
+
input: ["text"],
|
|
2848
|
+
cost: {
|
|
2849
|
+
input: 0.5,
|
|
2850
|
+
output: 1.5,
|
|
2851
|
+
cacheRead: 0,
|
|
2852
|
+
cacheWrite: 0,
|
|
2853
|
+
},
|
|
2854
|
+
contextWindow: 256000,
|
|
2855
|
+
maxTokens: 256000,
|
|
2856
|
+
},
|
|
2857
|
+
"@cf/openai/gpt-oss-120b": {
|
|
2858
|
+
id: "@cf/openai/gpt-oss-120b",
|
|
2859
|
+
name: "GPT OSS 120B",
|
|
2860
|
+
api: "openai-completions",
|
|
2861
|
+
provider: "cloudflare-workers-ai",
|
|
2862
|
+
baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
|
|
2863
|
+
compat: { "sendSessionAffinityHeaders": true },
|
|
2864
|
+
reasoning: true,
|
|
2865
|
+
input: ["text"],
|
|
2866
|
+
cost: {
|
|
2867
|
+
input: 0.35,
|
|
2868
|
+
output: 0.75,
|
|
2869
|
+
cacheRead: 0,
|
|
2870
|
+
cacheWrite: 0,
|
|
2871
|
+
},
|
|
2872
|
+
contextWindow: 128000,
|
|
2873
|
+
maxTokens: 16384,
|
|
2874
|
+
},
|
|
2875
|
+
"@cf/openai/gpt-oss-20b": {
|
|
2876
|
+
id: "@cf/openai/gpt-oss-20b",
|
|
2877
|
+
name: "GPT OSS 20B",
|
|
2878
|
+
api: "openai-completions",
|
|
2879
|
+
provider: "cloudflare-workers-ai",
|
|
2880
|
+
baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
|
|
2881
|
+
compat: { "sendSessionAffinityHeaders": true },
|
|
2882
|
+
reasoning: true,
|
|
2883
|
+
input: ["text"],
|
|
2884
|
+
cost: {
|
|
2885
|
+
input: 0.2,
|
|
2886
|
+
output: 0.3,
|
|
2887
|
+
cacheRead: 0,
|
|
2888
|
+
cacheWrite: 0,
|
|
2889
|
+
},
|
|
2890
|
+
contextWindow: 128000,
|
|
2891
|
+
maxTokens: 16384,
|
|
2892
|
+
},
|
|
2893
|
+
"@cf/zai-org/glm-4.7-flash": {
|
|
2894
|
+
id: "@cf/zai-org/glm-4.7-flash",
|
|
2895
|
+
name: "GLM-4.7-Flash",
|
|
2896
|
+
api: "openai-completions",
|
|
2897
|
+
provider: "cloudflare-workers-ai",
|
|
2898
|
+
baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
|
|
2899
|
+
compat: { "sendSessionAffinityHeaders": true },
|
|
2900
|
+
reasoning: true,
|
|
2901
|
+
input: ["text"],
|
|
2902
|
+
cost: {
|
|
2903
|
+
input: 0.06,
|
|
2904
|
+
output: 0.4,
|
|
2905
|
+
cacheRead: 0,
|
|
2906
|
+
cacheWrite: 0,
|
|
2907
|
+
},
|
|
2908
|
+
contextWindow: 131072,
|
|
2909
|
+
maxTokens: 131072,
|
|
2910
|
+
},
|
|
2911
|
+
},
|
|
2912
|
+
"deepseek": {
|
|
2913
|
+
"deepseek-v4-flash": {
|
|
2914
|
+
id: "deepseek-v4-flash",
|
|
2915
|
+
name: "DeepSeek V4 Flash",
|
|
2916
|
+
api: "openai-completions",
|
|
2917
|
+
provider: "deepseek",
|
|
2918
|
+
baseUrl: "https://api.deepseek.com",
|
|
2919
|
+
compat: { "requiresReasoningContentOnAssistantMessages": true, "thinkingFormat": "deepseek", "reasoningEffortMap": { "minimal": "high", "low": "high", "medium": "high", "high": "high", "xhigh": "max" } },
|
|
2920
|
+
reasoning: true,
|
|
2921
|
+
input: ["text"],
|
|
2922
|
+
cost: {
|
|
2923
|
+
input: 0.14,
|
|
2924
|
+
output: 0.28,
|
|
2925
|
+
cacheRead: 0.028,
|
|
2926
|
+
cacheWrite: 0,
|
|
2927
|
+
},
|
|
2928
|
+
contextWindow: 1000000,
|
|
2929
|
+
maxTokens: 384000,
|
|
2930
|
+
},
|
|
2931
|
+
"deepseek-v4-pro": {
|
|
2932
|
+
id: "deepseek-v4-pro",
|
|
2933
|
+
name: "DeepSeek V4 Pro",
|
|
2934
|
+
api: "openai-completions",
|
|
2935
|
+
provider: "deepseek",
|
|
2936
|
+
baseUrl: "https://api.deepseek.com",
|
|
2937
|
+
compat: { "requiresReasoningContentOnAssistantMessages": true, "thinkingFormat": "deepseek", "reasoningEffortMap": { "minimal": "high", "low": "high", "medium": "high", "high": "high", "xhigh": "max" } },
|
|
2938
|
+
reasoning: true,
|
|
2939
|
+
input: ["text"],
|
|
2940
|
+
cost: {
|
|
2941
|
+
input: 1.74,
|
|
2942
|
+
output: 3.48,
|
|
2943
|
+
cacheRead: 0.145,
|
|
2944
|
+
cacheWrite: 0,
|
|
2945
|
+
},
|
|
2946
|
+
contextWindow: 1000000,
|
|
2947
|
+
maxTokens: 384000,
|
|
2948
|
+
},
|
|
2949
|
+
},
|
|
2749
2950
|
"fireworks": {
|
|
2750
2951
|
"accounts/fireworks/models/deepseek-v3p1": {
|
|
2751
2952
|
id: "accounts/fireworks/models/deepseek-v3p1",
|
|
@@ -2781,6 +2982,23 @@ export const MODELS = {
|
|
|
2781
2982
|
contextWindow: 160000,
|
|
2782
2983
|
maxTokens: 160000,
|
|
2783
2984
|
},
|
|
2985
|
+
"accounts/fireworks/models/deepseek-v4-pro": {
|
|
2986
|
+
id: "accounts/fireworks/models/deepseek-v4-pro",
|
|
2987
|
+
name: "DeepSeek V4 Pro",
|
|
2988
|
+
api: "anthropic-messages",
|
|
2989
|
+
provider: "fireworks",
|
|
2990
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2991
|
+
reasoning: true,
|
|
2992
|
+
input: ["text"],
|
|
2993
|
+
cost: {
|
|
2994
|
+
input: 1.74,
|
|
2995
|
+
output: 3.48,
|
|
2996
|
+
cacheRead: 0.15,
|
|
2997
|
+
cacheWrite: 0,
|
|
2998
|
+
},
|
|
2999
|
+
contextWindow: 1000000,
|
|
3000
|
+
maxTokens: 384000,
|
|
3001
|
+
},
|
|
2784
3002
|
"accounts/fireworks/models/glm-4p5": {
|
|
2785
3003
|
id: "accounts/fireworks/models/glm-4p5",
|
|
2786
3004
|
name: "GLM 4.5",
|
|
@@ -3496,6 +3714,24 @@ export const MODELS = {
|
|
|
3496
3714
|
contextWindow: 400000,
|
|
3497
3715
|
maxTokens: 128000,
|
|
3498
3716
|
},
|
|
3717
|
+
"gpt-5.5": {
|
|
3718
|
+
id: "gpt-5.5",
|
|
3719
|
+
name: "GPT-5.5",
|
|
3720
|
+
api: "openai-responses",
|
|
3721
|
+
provider: "github-copilot",
|
|
3722
|
+
baseUrl: "https://api.individual.githubcopilot.com",
|
|
3723
|
+
headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
|
|
3724
|
+
reasoning: true,
|
|
3725
|
+
input: ["text", "image"],
|
|
3726
|
+
cost: {
|
|
3727
|
+
input: 0,
|
|
3728
|
+
output: 0,
|
|
3729
|
+
cacheRead: 0,
|
|
3730
|
+
cacheWrite: 0,
|
|
3731
|
+
},
|
|
3732
|
+
contextWindow: 400000,
|
|
3733
|
+
maxTokens: 128000,
|
|
3734
|
+
},
|
|
3499
3735
|
"grok-code-fast-1": {
|
|
3500
3736
|
id: "grok-code-fast-1",
|
|
3501
3737
|
name: "Grok Code Fast 1",
|
|
@@ -5001,6 +5237,24 @@ export const MODELS = {
|
|
|
5001
5237
|
contextWindow: 163840,
|
|
5002
5238
|
maxTokens: 65536,
|
|
5003
5239
|
},
|
|
5240
|
+
"deepseek-ai/DeepSeek-V4-Pro": {
|
|
5241
|
+
id: "deepseek-ai/DeepSeek-V4-Pro",
|
|
5242
|
+
name: "DeepSeek V4 Pro",
|
|
5243
|
+
api: "openai-completions",
|
|
5244
|
+
provider: "huggingface",
|
|
5245
|
+
baseUrl: "https://router.huggingface.co/v1",
|
|
5246
|
+
compat: { "supportsDeveloperRole": false },
|
|
5247
|
+
reasoning: true,
|
|
5248
|
+
input: ["text"],
|
|
5249
|
+
cost: {
|
|
5250
|
+
input: 1.74,
|
|
5251
|
+
output: 3.48,
|
|
5252
|
+
cacheRead: 0.145,
|
|
5253
|
+
cacheWrite: 0,
|
|
5254
|
+
},
|
|
5255
|
+
contextWindow: 1048576,
|
|
5256
|
+
maxTokens: 393216,
|
|
5257
|
+
},
|
|
5004
5258
|
"moonshotai/Kimi-K2-Instruct": {
|
|
5005
5259
|
id: "moonshotai/Kimi-K2-Instruct",
|
|
5006
5260
|
name: "Kimi-K2-Instruct",
|
|
@@ -6295,8 +6549,25 @@ export const MODELS = {
|
|
|
6295
6549
|
cacheRead: 0.5,
|
|
6296
6550
|
cacheWrite: 0,
|
|
6297
6551
|
},
|
|
6552
|
+
contextWindow: 272000,
|
|
6553
|
+
maxTokens: 128000,
|
|
6554
|
+
},
|
|
6555
|
+
"gpt-5.5-pro": {
|
|
6556
|
+
id: "gpt-5.5-pro",
|
|
6557
|
+
name: "GPT-5.5 Pro",
|
|
6558
|
+
api: "openai-responses",
|
|
6559
|
+
provider: "openai",
|
|
6560
|
+
baseUrl: "https://api.openai.com/v1",
|
|
6561
|
+
reasoning: true,
|
|
6562
|
+
input: ["text", "image"],
|
|
6563
|
+
cost: {
|
|
6564
|
+
input: 30,
|
|
6565
|
+
output: 180,
|
|
6566
|
+
cacheRead: 0,
|
|
6567
|
+
cacheWrite: 0,
|
|
6568
|
+
},
|
|
6298
6569
|
contextWindow: 1050000,
|
|
6299
|
-
maxTokens:
|
|
6570
|
+
maxTokens: 128000,
|
|
6300
6571
|
},
|
|
6301
6572
|
"o1": {
|
|
6302
6573
|
id: "o1",
|
|
@@ -6603,7 +6874,7 @@ export const MODELS = {
|
|
|
6603
6874
|
cacheRead: 0.5,
|
|
6604
6875
|
cacheWrite: 0,
|
|
6605
6876
|
},
|
|
6606
|
-
contextWindow:
|
|
6877
|
+
contextWindow: 272000,
|
|
6607
6878
|
maxTokens: 128000,
|
|
6608
6879
|
},
|
|
6609
6880
|
},
|
|
@@ -6625,23 +6896,6 @@ export const MODELS = {
|
|
|
6625
6896
|
contextWindow: 200000,
|
|
6626
6897
|
maxTokens: 128000,
|
|
6627
6898
|
},
|
|
6628
|
-
"claude-3-5-haiku": {
|
|
6629
|
-
id: "claude-3-5-haiku",
|
|
6630
|
-
name: "Claude Haiku 3.5",
|
|
6631
|
-
api: "anthropic-messages",
|
|
6632
|
-
provider: "opencode",
|
|
6633
|
-
baseUrl: "https://opencode.ai/zen",
|
|
6634
|
-
reasoning: false,
|
|
6635
|
-
input: ["text", "image"],
|
|
6636
|
-
cost: {
|
|
6637
|
-
input: 0.8,
|
|
6638
|
-
output: 4,
|
|
6639
|
-
cacheRead: 0.08,
|
|
6640
|
-
cacheWrite: 1,
|
|
6641
|
-
},
|
|
6642
|
-
contextWindow: 200000,
|
|
6643
|
-
maxTokens: 8192,
|
|
6644
|
-
},
|
|
6645
6899
|
"claude-haiku-4-5": {
|
|
6646
6900
|
id: "claude-haiku-4-5",
|
|
6647
6901
|
name: "Claude Haiku 4.5",
|
|
@@ -7099,7 +7353,7 @@ export const MODELS = {
|
|
|
7099
7353
|
cacheWrite: 0,
|
|
7100
7354
|
},
|
|
7101
7355
|
contextWindow: 1050000,
|
|
7102
|
-
maxTokens:
|
|
7356
|
+
maxTokens: 128000,
|
|
7103
7357
|
},
|
|
7104
7358
|
"gpt-5.5-pro": {
|
|
7105
7359
|
id: "gpt-5.5-pro",
|
|
@@ -7296,12 +7550,13 @@ export const MODELS = {
|
|
|
7296
7550
|
api: "openai-completions",
|
|
7297
7551
|
provider: "opencode-go",
|
|
7298
7552
|
baseUrl: "https://opencode.ai/zen/go/v1",
|
|
7553
|
+
compat: { "requiresReasoningContentOnAssistantMessages": true },
|
|
7299
7554
|
reasoning: true,
|
|
7300
7555
|
input: ["text"],
|
|
7301
7556
|
cost: {
|
|
7302
7557
|
input: 0.14,
|
|
7303
7558
|
output: 0.28,
|
|
7304
|
-
cacheRead: 0.
|
|
7559
|
+
cacheRead: 0.0028,
|
|
7305
7560
|
cacheWrite: 0,
|
|
7306
7561
|
},
|
|
7307
7562
|
contextWindow: 1000000,
|
|
@@ -7313,12 +7568,13 @@ export const MODELS = {
|
|
|
7313
7568
|
api: "openai-completions",
|
|
7314
7569
|
provider: "opencode-go",
|
|
7315
7570
|
baseUrl: "https://opencode.ai/zen/go/v1",
|
|
7571
|
+
compat: { "requiresReasoningContentOnAssistantMessages": true },
|
|
7316
7572
|
reasoning: true,
|
|
7317
7573
|
input: ["text"],
|
|
7318
7574
|
cost: {
|
|
7319
7575
|
input: 1.74,
|
|
7320
7576
|
output: 3.48,
|
|
7321
|
-
cacheRead: 0.
|
|
7577
|
+
cacheRead: 0.0145,
|
|
7322
7578
|
cacheWrite: 0,
|
|
7323
7579
|
},
|
|
7324
7580
|
contextWindow: 1000000,
|
|
@@ -7338,8 +7594,8 @@ export const MODELS = {
|
|
|
7338
7594
|
cacheRead: 0.2,
|
|
7339
7595
|
cacheWrite: 0,
|
|
7340
7596
|
},
|
|
7341
|
-
contextWindow:
|
|
7342
|
-
maxTokens:
|
|
7597
|
+
contextWindow: 202752,
|
|
7598
|
+
maxTokens: 32768,
|
|
7343
7599
|
},
|
|
7344
7600
|
"glm-5.1": {
|
|
7345
7601
|
id: "glm-5.1",
|
|
@@ -7355,8 +7611,8 @@ export const MODELS = {
|
|
|
7355
7611
|
cacheRead: 0.26,
|
|
7356
7612
|
cacheWrite: 0,
|
|
7357
7613
|
},
|
|
7358
|
-
contextWindow:
|
|
7359
|
-
maxTokens:
|
|
7614
|
+
contextWindow: 202752,
|
|
7615
|
+
maxTokens: 32768,
|
|
7360
7616
|
},
|
|
7361
7617
|
"kimi-k2.5": {
|
|
7362
7618
|
id: "kimi-k2.5",
|
|
@@ -7440,7 +7696,7 @@ export const MODELS = {
|
|
|
7440
7696
|
cacheRead: 0.08,
|
|
7441
7697
|
cacheWrite: 0,
|
|
7442
7698
|
},
|
|
7443
|
-
contextWindow:
|
|
7699
|
+
contextWindow: 1000000,
|
|
7444
7700
|
maxTokens: 128000,
|
|
7445
7701
|
},
|
|
7446
7702
|
"mimo-v2.5-pro": {
|
|
@@ -8267,12 +8523,13 @@ export const MODELS = {
|
|
|
8267
8523
|
api: "openai-completions",
|
|
8268
8524
|
provider: "openrouter",
|
|
8269
8525
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
8526
|
+
compat: { "requiresReasoningContentOnAssistantMessages": true },
|
|
8270
8527
|
reasoning: true,
|
|
8271
8528
|
input: ["text"],
|
|
8272
8529
|
cost: {
|
|
8273
8530
|
input: 0.14,
|
|
8274
8531
|
output: 0.28,
|
|
8275
|
-
cacheRead: 0.
|
|
8532
|
+
cacheRead: 0.0028,
|
|
8276
8533
|
cacheWrite: 0,
|
|
8277
8534
|
},
|
|
8278
8535
|
contextWindow: 1048576,
|
|
@@ -8284,12 +8541,13 @@ export const MODELS = {
|
|
|
8284
8541
|
api: "openai-completions",
|
|
8285
8542
|
provider: "openrouter",
|
|
8286
8543
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
8544
|
+
compat: { "requiresReasoningContentOnAssistantMessages": true },
|
|
8287
8545
|
reasoning: true,
|
|
8288
8546
|
input: ["text"],
|
|
8289
8547
|
cost: {
|
|
8290
|
-
input:
|
|
8291
|
-
output:
|
|
8292
|
-
cacheRead: 0.
|
|
8548
|
+
input: 0.435,
|
|
8549
|
+
output: 0.87,
|
|
8550
|
+
cacheRead: 0.003625,
|
|
8293
8551
|
cacheWrite: 0,
|
|
8294
8552
|
},
|
|
8295
8553
|
contextWindow: 1048576,
|
|
@@ -8652,18 +8910,18 @@ export const MODELS = {
|
|
|
8652
8910
|
contextWindow: 262144,
|
|
8653
8911
|
maxTokens: 32768,
|
|
8654
8912
|
},
|
|
8655
|
-
"inclusionai/ling-2.6-flash
|
|
8656
|
-
id: "inclusionai/ling-2.6-flash
|
|
8657
|
-
name: "inclusionAI: Ling-2.6-flash
|
|
8913
|
+
"inclusionai/ling-2.6-flash": {
|
|
8914
|
+
id: "inclusionai/ling-2.6-flash",
|
|
8915
|
+
name: "inclusionAI: Ling-2.6-flash",
|
|
8658
8916
|
api: "openai-completions",
|
|
8659
8917
|
provider: "openrouter",
|
|
8660
8918
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
8661
8919
|
reasoning: false,
|
|
8662
8920
|
input: ["text"],
|
|
8663
8921
|
cost: {
|
|
8664
|
-
input: 0,
|
|
8665
|
-
output: 0,
|
|
8666
|
-
cacheRead: 0,
|
|
8922
|
+
input: 0.08,
|
|
8923
|
+
output: 0.24,
|
|
8924
|
+
cacheRead: 0.016,
|
|
8667
8925
|
cacheWrite: 0,
|
|
8668
8926
|
},
|
|
8669
8927
|
contextWindow: 262144,
|
|
@@ -9120,7 +9378,7 @@ export const MODELS = {
|
|
|
9120
9378
|
reasoning: false,
|
|
9121
9379
|
input: ["text"],
|
|
9122
9380
|
cost: {
|
|
9123
|
-
input: 0.
|
|
9381
|
+
input: 0.02,
|
|
9124
9382
|
output: 0.03,
|
|
9125
9383
|
cacheRead: 0,
|
|
9126
9384
|
cacheWrite: 0,
|
|
@@ -9434,6 +9692,23 @@ export const MODELS = {
|
|
|
9434
9692
|
contextWindow: 256000,
|
|
9435
9693
|
maxTokens: 4096,
|
|
9436
9694
|
},
|
|
9695
|
+
"nvidia/nemotron-3-nano-omni-30b-a3b-reasoning:free": {
|
|
9696
|
+
id: "nvidia/nemotron-3-nano-omni-30b-a3b-reasoning:free",
|
|
9697
|
+
name: "NVIDIA: Nemotron 3 Nano Omni (free)",
|
|
9698
|
+
api: "openai-completions",
|
|
9699
|
+
provider: "openrouter",
|
|
9700
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
9701
|
+
reasoning: true,
|
|
9702
|
+
input: ["text", "image"],
|
|
9703
|
+
cost: {
|
|
9704
|
+
input: 0,
|
|
9705
|
+
output: 0,
|
|
9706
|
+
cacheRead: 0,
|
|
9707
|
+
cacheWrite: 0,
|
|
9708
|
+
},
|
|
9709
|
+
contextWindow: 256000,
|
|
9710
|
+
maxTokens: 65536,
|
|
9711
|
+
},
|
|
9437
9712
|
"nvidia/nemotron-3-super-120b-a12b": {
|
|
9438
9713
|
id: "nvidia/nemotron-3-super-120b-a12b",
|
|
9439
9714
|
name: "NVIDIA: Nemotron 3 Super",
|
|
@@ -10505,6 +10780,40 @@ export const MODELS = {
|
|
|
10505
10780
|
contextWindow: 200000,
|
|
10506
10781
|
maxTokens: 4096,
|
|
10507
10782
|
},
|
|
10783
|
+
"poolside/laguna-m.1:free": {
|
|
10784
|
+
id: "poolside/laguna-m.1:free",
|
|
10785
|
+
name: "Poolside: Laguna M.1 (free)",
|
|
10786
|
+
api: "openai-completions",
|
|
10787
|
+
provider: "openrouter",
|
|
10788
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
10789
|
+
reasoning: true,
|
|
10790
|
+
input: ["text"],
|
|
10791
|
+
cost: {
|
|
10792
|
+
input: 0,
|
|
10793
|
+
output: 0,
|
|
10794
|
+
cacheRead: 0,
|
|
10795
|
+
cacheWrite: 0,
|
|
10796
|
+
},
|
|
10797
|
+
contextWindow: 131072,
|
|
10798
|
+
maxTokens: 8192,
|
|
10799
|
+
},
|
|
10800
|
+
"poolside/laguna-xs.2:free": {
|
|
10801
|
+
id: "poolside/laguna-xs.2:free",
|
|
10802
|
+
name: "Poolside: Laguna XS.2 (free)",
|
|
10803
|
+
api: "openai-completions",
|
|
10804
|
+
provider: "openrouter",
|
|
10805
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
10806
|
+
reasoning: true,
|
|
10807
|
+
input: ["text"],
|
|
10808
|
+
cost: {
|
|
10809
|
+
input: 0,
|
|
10810
|
+
output: 0,
|
|
10811
|
+
cacheRead: 0,
|
|
10812
|
+
cacheWrite: 0,
|
|
10813
|
+
},
|
|
10814
|
+
contextWindow: 131072,
|
|
10815
|
+
maxTokens: 8192,
|
|
10816
|
+
},
|
|
10508
10817
|
"prime-intellect/intellect-3": {
|
|
10509
10818
|
id: "prime-intellect/intellect-3",
|
|
10510
10819
|
name: "Prime Intellect: INTELLECT-3",
|
|
@@ -10531,8 +10840,8 @@ export const MODELS = {
|
|
|
10531
10840
|
reasoning: false,
|
|
10532
10841
|
input: ["text"],
|
|
10533
10842
|
cost: {
|
|
10534
|
-
input: 0.
|
|
10535
|
-
output: 0.
|
|
10843
|
+
input: 0.36,
|
|
10844
|
+
output: 0.39999999999999997,
|
|
10536
10845
|
cacheRead: 0,
|
|
10537
10846
|
cacheWrite: 0,
|
|
10538
10847
|
},
|
|
@@ -10821,8 +11130,8 @@ export const MODELS = {
|
|
|
10821
11130
|
input: ["text"],
|
|
10822
11131
|
cost: {
|
|
10823
11132
|
input: 0.22,
|
|
10824
|
-
output: 1,
|
|
10825
|
-
cacheRead: 0
|
|
11133
|
+
output: 1.7999999999999998,
|
|
11134
|
+
cacheRead: 0,
|
|
10826
11135
|
cacheWrite: 0,
|
|
10827
11136
|
},
|
|
10828
11137
|
contextWindow: 262144,
|
|
@@ -11236,9 +11545,26 @@ export const MODELS = {
|
|
|
11236
11545
|
contextWindow: 1000000,
|
|
11237
11546
|
maxTokens: 65536,
|
|
11238
11547
|
},
|
|
11239
|
-
"qwen/qwen3.
|
|
11240
|
-
id: "qwen/qwen3.
|
|
11241
|
-
name: "Qwen: Qwen3.
|
|
11548
|
+
"qwen/qwen3.5-plus-20260420": {
|
|
11549
|
+
id: "qwen/qwen3.5-plus-20260420",
|
|
11550
|
+
name: "Qwen: Qwen3.5 Plus 2026-04-20",
|
|
11551
|
+
api: "openai-completions",
|
|
11552
|
+
provider: "openrouter",
|
|
11553
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
11554
|
+
reasoning: true,
|
|
11555
|
+
input: ["text", "image"],
|
|
11556
|
+
cost: {
|
|
11557
|
+
input: 0.39999999999999997,
|
|
11558
|
+
output: 2.4,
|
|
11559
|
+
cacheRead: 0,
|
|
11560
|
+
cacheWrite: 0,
|
|
11561
|
+
},
|
|
11562
|
+
contextWindow: 1000000,
|
|
11563
|
+
maxTokens: 65536,
|
|
11564
|
+
},
|
|
11565
|
+
"qwen/qwen3.6-27b": {
|
|
11566
|
+
id: "qwen/qwen3.6-27b",
|
|
11567
|
+
name: "Qwen: Qwen3.6 27B",
|
|
11242
11568
|
api: "openai-completions",
|
|
11243
11569
|
provider: "openrouter",
|
|
11244
11570
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -11246,29 +11572,63 @@ export const MODELS = {
|
|
|
11246
11572
|
input: ["text", "image"],
|
|
11247
11573
|
cost: {
|
|
11248
11574
|
input: 0.325,
|
|
11249
|
-
output:
|
|
11575
|
+
output: 3.25,
|
|
11250
11576
|
cacheRead: 0,
|
|
11251
|
-
cacheWrite: 0
|
|
11577
|
+
cacheWrite: 0,
|
|
11578
|
+
},
|
|
11579
|
+
contextWindow: 256000,
|
|
11580
|
+
maxTokens: 65536,
|
|
11581
|
+
},
|
|
11582
|
+
"qwen/qwen3.6-flash": {
|
|
11583
|
+
id: "qwen/qwen3.6-flash",
|
|
11584
|
+
name: "Qwen: Qwen3.6 Flash",
|
|
11585
|
+
api: "openai-completions",
|
|
11586
|
+
provider: "openrouter",
|
|
11587
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
11588
|
+
reasoning: true,
|
|
11589
|
+
input: ["text", "image"],
|
|
11590
|
+
cost: {
|
|
11591
|
+
input: 0.25,
|
|
11592
|
+
output: 1.5,
|
|
11593
|
+
cacheRead: 0,
|
|
11594
|
+
cacheWrite: 0.3125,
|
|
11252
11595
|
},
|
|
11253
11596
|
contextWindow: 1000000,
|
|
11254
11597
|
maxTokens: 65536,
|
|
11255
11598
|
},
|
|
11256
|
-
"qwen/
|
|
11257
|
-
id: "qwen/
|
|
11258
|
-
name: "Qwen:
|
|
11599
|
+
"qwen/qwen3.6-max-preview": {
|
|
11600
|
+
id: "qwen/qwen3.6-max-preview",
|
|
11601
|
+
name: "Qwen: Qwen3.6 Max Preview",
|
|
11259
11602
|
api: "openai-completions",
|
|
11260
11603
|
provider: "openrouter",
|
|
11261
11604
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
11262
11605
|
reasoning: true,
|
|
11263
11606
|
input: ["text"],
|
|
11264
11607
|
cost: {
|
|
11265
|
-
input:
|
|
11266
|
-
output:
|
|
11608
|
+
input: 1.04,
|
|
11609
|
+
output: 6.24,
|
|
11267
11610
|
cacheRead: 0,
|
|
11268
|
-
cacheWrite:
|
|
11611
|
+
cacheWrite: 1.3,
|
|
11269
11612
|
},
|
|
11270
|
-
contextWindow:
|
|
11271
|
-
maxTokens:
|
|
11613
|
+
contextWindow: 262144,
|
|
11614
|
+
maxTokens: 65536,
|
|
11615
|
+
},
|
|
11616
|
+
"qwen/qwen3.6-plus": {
|
|
11617
|
+
id: "qwen/qwen3.6-plus",
|
|
11618
|
+
name: "Qwen: Qwen3.6 Plus",
|
|
11619
|
+
api: "openai-completions",
|
|
11620
|
+
provider: "openrouter",
|
|
11621
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
11622
|
+
reasoning: true,
|
|
11623
|
+
input: ["text", "image"],
|
|
11624
|
+
cost: {
|
|
11625
|
+
input: 0.325,
|
|
11626
|
+
output: 1.95,
|
|
11627
|
+
cacheRead: 0,
|
|
11628
|
+
cacheWrite: 0.40625,
|
|
11629
|
+
},
|
|
11630
|
+
contextWindow: 1000000,
|
|
11631
|
+
maxTokens: 65536,
|
|
11272
11632
|
},
|
|
11273
11633
|
"rekaai/reka-edge": {
|
|
11274
11634
|
id: "rekaai/reka-edge",
|
|
@@ -11899,6 +12259,23 @@ export const MODELS = {
|
|
|
11899
12259
|
contextWindow: 202752,
|
|
11900
12260
|
maxTokens: 131072,
|
|
11901
12261
|
},
|
|
12262
|
+
"~anthropic/claude-haiku-latest": {
|
|
12263
|
+
id: "~anthropic/claude-haiku-latest",
|
|
12264
|
+
name: "Anthropic Claude Haiku Latest",
|
|
12265
|
+
api: "openai-completions",
|
|
12266
|
+
provider: "openrouter",
|
|
12267
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
12268
|
+
reasoning: true,
|
|
12269
|
+
input: ["text", "image"],
|
|
12270
|
+
cost: {
|
|
12271
|
+
input: 1,
|
|
12272
|
+
output: 5,
|
|
12273
|
+
cacheRead: 0.09999999999999999,
|
|
12274
|
+
cacheWrite: 1.25,
|
|
12275
|
+
},
|
|
12276
|
+
contextWindow: 200000,
|
|
12277
|
+
maxTokens: 64000,
|
|
12278
|
+
},
|
|
11902
12279
|
"~anthropic/claude-opus-latest": {
|
|
11903
12280
|
id: "~anthropic/claude-opus-latest",
|
|
11904
12281
|
name: "Anthropic: Claude Opus Latest",
|
|
@@ -11916,6 +12293,108 @@ export const MODELS = {
|
|
|
11916
12293
|
contextWindow: 1000000,
|
|
11917
12294
|
maxTokens: 128000,
|
|
11918
12295
|
},
|
|
12296
|
+
"~anthropic/claude-sonnet-latest": {
|
|
12297
|
+
id: "~anthropic/claude-sonnet-latest",
|
|
12298
|
+
name: "Anthropic Claude Sonnet Latest",
|
|
12299
|
+
api: "openai-completions",
|
|
12300
|
+
provider: "openrouter",
|
|
12301
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
12302
|
+
reasoning: true,
|
|
12303
|
+
input: ["text", "image"],
|
|
12304
|
+
cost: {
|
|
12305
|
+
input: 3,
|
|
12306
|
+
output: 15,
|
|
12307
|
+
cacheRead: 0.3,
|
|
12308
|
+
cacheWrite: 3.75,
|
|
12309
|
+
},
|
|
12310
|
+
contextWindow: 1000000,
|
|
12311
|
+
maxTokens: 128000,
|
|
12312
|
+
},
|
|
12313
|
+
"~google/gemini-flash-latest": {
|
|
12314
|
+
id: "~google/gemini-flash-latest",
|
|
12315
|
+
name: "Google Gemini Flash Latest",
|
|
12316
|
+
api: "openai-completions",
|
|
12317
|
+
provider: "openrouter",
|
|
12318
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
12319
|
+
reasoning: true,
|
|
12320
|
+
input: ["text", "image"],
|
|
12321
|
+
cost: {
|
|
12322
|
+
input: 0.5,
|
|
12323
|
+
output: 3,
|
|
12324
|
+
cacheRead: 0.049999999999999996,
|
|
12325
|
+
cacheWrite: 0.08333333333333334,
|
|
12326
|
+
},
|
|
12327
|
+
contextWindow: 1048576,
|
|
12328
|
+
maxTokens: 65536,
|
|
12329
|
+
},
|
|
12330
|
+
"~google/gemini-pro-latest": {
|
|
12331
|
+
id: "~google/gemini-pro-latest",
|
|
12332
|
+
name: "Google Gemini Pro Latest",
|
|
12333
|
+
api: "openai-completions",
|
|
12334
|
+
provider: "openrouter",
|
|
12335
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
12336
|
+
reasoning: true,
|
|
12337
|
+
input: ["text", "image"],
|
|
12338
|
+
cost: {
|
|
12339
|
+
input: 2,
|
|
12340
|
+
output: 12,
|
|
12341
|
+
cacheRead: 0.19999999999999998,
|
|
12342
|
+
cacheWrite: 0.375,
|
|
12343
|
+
},
|
|
12344
|
+
contextWindow: 1048576,
|
|
12345
|
+
maxTokens: 65536,
|
|
12346
|
+
},
|
|
12347
|
+
"~moonshotai/kimi-latest": {
|
|
12348
|
+
id: "~moonshotai/kimi-latest",
|
|
12349
|
+
name: "MoonshotAI Kimi Latest",
|
|
12350
|
+
api: "openai-completions",
|
|
12351
|
+
provider: "openrouter",
|
|
12352
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
12353
|
+
reasoning: true,
|
|
12354
|
+
input: ["text", "image"],
|
|
12355
|
+
cost: {
|
|
12356
|
+
input: 0.7448,
|
|
12357
|
+
output: 4.655,
|
|
12358
|
+
cacheRead: 0.1463,
|
|
12359
|
+
cacheWrite: 0,
|
|
12360
|
+
},
|
|
12361
|
+
contextWindow: 256000,
|
|
12362
|
+
maxTokens: 65536,
|
|
12363
|
+
},
|
|
12364
|
+
"~openai/gpt-latest": {
|
|
12365
|
+
id: "~openai/gpt-latest",
|
|
12366
|
+
name: "OpenAI GPT Latest",
|
|
12367
|
+
api: "openai-completions",
|
|
12368
|
+
provider: "openrouter",
|
|
12369
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
12370
|
+
reasoning: true,
|
|
12371
|
+
input: ["text", "image"],
|
|
12372
|
+
cost: {
|
|
12373
|
+
input: 5,
|
|
12374
|
+
output: 30,
|
|
12375
|
+
cacheRead: 0.5,
|
|
12376
|
+
cacheWrite: 0,
|
|
12377
|
+
},
|
|
12378
|
+
contextWindow: 1050000,
|
|
12379
|
+
maxTokens: 128000,
|
|
12380
|
+
},
|
|
12381
|
+
"~openai/gpt-mini-latest": {
|
|
12382
|
+
id: "~openai/gpt-mini-latest",
|
|
12383
|
+
name: "OpenAI GPT Mini Latest",
|
|
12384
|
+
api: "openai-completions",
|
|
12385
|
+
provider: "openrouter",
|
|
12386
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
12387
|
+
reasoning: true,
|
|
12388
|
+
input: ["text", "image"],
|
|
12389
|
+
cost: {
|
|
12390
|
+
input: 0.75,
|
|
12391
|
+
output: 4.5,
|
|
12392
|
+
cacheRead: 0.075,
|
|
12393
|
+
cacheWrite: 0,
|
|
12394
|
+
},
|
|
12395
|
+
contextWindow: 400000,
|
|
12396
|
+
maxTokens: 128000,
|
|
12397
|
+
},
|
|
11919
12398
|
},
|
|
11920
12399
|
"vercel-ai-gateway": {
|
|
11921
12400
|
"alibaba/qwen-3-14b": {
|
|
@@ -12592,7 +13071,7 @@ export const MODELS = {
|
|
|
12592
13071
|
cost: {
|
|
12593
13072
|
input: 0.14,
|
|
12594
13073
|
output: 0.28,
|
|
12595
|
-
cacheRead: 0.
|
|
13074
|
+
cacheRead: 0.0028,
|
|
12596
13075
|
cacheWrite: 0,
|
|
12597
13076
|
},
|
|
12598
13077
|
contextWindow: 1000000,
|
|
@@ -12607,9 +13086,9 @@ export const MODELS = {
|
|
|
12607
13086
|
reasoning: true,
|
|
12608
13087
|
input: ["text"],
|
|
12609
13088
|
cost: {
|
|
12610
|
-
input:
|
|
12611
|
-
output:
|
|
12612
|
-
cacheRead: 0.
|
|
13089
|
+
input: 0.435,
|
|
13090
|
+
output: 0.87,
|
|
13091
|
+
cacheRead: 0.0036,
|
|
12613
13092
|
cacheWrite: 0,
|
|
12614
13093
|
},
|
|
12615
13094
|
contextWindow: 1000000,
|
|
@@ -15050,24 +15529,6 @@ export const MODELS = {
|
|
|
15050
15529
|
},
|
|
15051
15530
|
},
|
|
15052
15531
|
"zai": {
|
|
15053
|
-
"glm-4.5": {
|
|
15054
|
-
id: "glm-4.5",
|
|
15055
|
-
name: "GLM-4.5",
|
|
15056
|
-
api: "openai-completions",
|
|
15057
|
-
provider: "zai",
|
|
15058
|
-
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
15059
|
-
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
|
|
15060
|
-
reasoning: true,
|
|
15061
|
-
input: ["text"],
|
|
15062
|
-
cost: {
|
|
15063
|
-
input: 0,
|
|
15064
|
-
output: 0,
|
|
15065
|
-
cacheRead: 0,
|
|
15066
|
-
cacheWrite: 0,
|
|
15067
|
-
},
|
|
15068
|
-
contextWindow: 131072,
|
|
15069
|
-
maxTokens: 98304,
|
|
15070
|
-
},
|
|
15071
15532
|
"glm-4.5-air": {
|
|
15072
15533
|
id: "glm-4.5-air",
|
|
15073
15534
|
name: "GLM-4.5-Air",
|
|
@@ -15086,78 +15547,6 @@ export const MODELS = {
|
|
|
15086
15547
|
contextWindow: 131072,
|
|
15087
15548
|
maxTokens: 98304,
|
|
15088
15549
|
},
|
|
15089
|
-
"glm-4.5-flash": {
|
|
15090
|
-
id: "glm-4.5-flash",
|
|
15091
|
-
name: "GLM-4.5-Flash",
|
|
15092
|
-
api: "openai-completions",
|
|
15093
|
-
provider: "zai",
|
|
15094
|
-
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
15095
|
-
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
|
|
15096
|
-
reasoning: true,
|
|
15097
|
-
input: ["text"],
|
|
15098
|
-
cost: {
|
|
15099
|
-
input: 0,
|
|
15100
|
-
output: 0,
|
|
15101
|
-
cacheRead: 0,
|
|
15102
|
-
cacheWrite: 0,
|
|
15103
|
-
},
|
|
15104
|
-
contextWindow: 131072,
|
|
15105
|
-
maxTokens: 98304,
|
|
15106
|
-
},
|
|
15107
|
-
"glm-4.5v": {
|
|
15108
|
-
id: "glm-4.5v",
|
|
15109
|
-
name: "GLM-4.5V",
|
|
15110
|
-
api: "openai-completions",
|
|
15111
|
-
provider: "zai",
|
|
15112
|
-
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
15113
|
-
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
|
|
15114
|
-
reasoning: true,
|
|
15115
|
-
input: ["text", "image"],
|
|
15116
|
-
cost: {
|
|
15117
|
-
input: 0,
|
|
15118
|
-
output: 0,
|
|
15119
|
-
cacheRead: 0,
|
|
15120
|
-
cacheWrite: 0,
|
|
15121
|
-
},
|
|
15122
|
-
contextWindow: 64000,
|
|
15123
|
-
maxTokens: 16384,
|
|
15124
|
-
},
|
|
15125
|
-
"glm-4.6": {
|
|
15126
|
-
id: "glm-4.6",
|
|
15127
|
-
name: "GLM-4.6",
|
|
15128
|
-
api: "openai-completions",
|
|
15129
|
-
provider: "zai",
|
|
15130
|
-
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
15131
|
-
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
15132
|
-
reasoning: true,
|
|
15133
|
-
input: ["text"],
|
|
15134
|
-
cost: {
|
|
15135
|
-
input: 0,
|
|
15136
|
-
output: 0,
|
|
15137
|
-
cacheRead: 0,
|
|
15138
|
-
cacheWrite: 0,
|
|
15139
|
-
},
|
|
15140
|
-
contextWindow: 204800,
|
|
15141
|
-
maxTokens: 131072,
|
|
15142
|
-
},
|
|
15143
|
-
"glm-4.6v": {
|
|
15144
|
-
id: "glm-4.6v",
|
|
15145
|
-
name: "GLM-4.6V",
|
|
15146
|
-
api: "openai-completions",
|
|
15147
|
-
provider: "zai",
|
|
15148
|
-
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
15149
|
-
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
15150
|
-
reasoning: true,
|
|
15151
|
-
input: ["text", "image"],
|
|
15152
|
-
cost: {
|
|
15153
|
-
input: 0,
|
|
15154
|
-
output: 0,
|
|
15155
|
-
cacheRead: 0,
|
|
15156
|
-
cacheWrite: 0,
|
|
15157
|
-
},
|
|
15158
|
-
contextWindow: 128000,
|
|
15159
|
-
maxTokens: 32768,
|
|
15160
|
-
},
|
|
15161
15550
|
"glm-4.7": {
|
|
15162
15551
|
id: "glm-4.7",
|
|
15163
15552
|
name: "GLM-4.7",
|
|
@@ -15176,60 +15565,6 @@ export const MODELS = {
|
|
|
15176
15565
|
contextWindow: 204800,
|
|
15177
15566
|
maxTokens: 131072,
|
|
15178
15567
|
},
|
|
15179
|
-
"glm-4.7-flash": {
|
|
15180
|
-
id: "glm-4.7-flash",
|
|
15181
|
-
name: "GLM-4.7-Flash",
|
|
15182
|
-
api: "openai-completions",
|
|
15183
|
-
provider: "zai",
|
|
15184
|
-
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
15185
|
-
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
15186
|
-
reasoning: true,
|
|
15187
|
-
input: ["text"],
|
|
15188
|
-
cost: {
|
|
15189
|
-
input: 0,
|
|
15190
|
-
output: 0,
|
|
15191
|
-
cacheRead: 0,
|
|
15192
|
-
cacheWrite: 0,
|
|
15193
|
-
},
|
|
15194
|
-
contextWindow: 200000,
|
|
15195
|
-
maxTokens: 131072,
|
|
15196
|
-
},
|
|
15197
|
-
"glm-4.7-flashx": {
|
|
15198
|
-
id: "glm-4.7-flashx",
|
|
15199
|
-
name: "GLM-4.7-FlashX",
|
|
15200
|
-
api: "openai-completions",
|
|
15201
|
-
provider: "zai",
|
|
15202
|
-
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
15203
|
-
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
15204
|
-
reasoning: true,
|
|
15205
|
-
input: ["text"],
|
|
15206
|
-
cost: {
|
|
15207
|
-
input: 0.07,
|
|
15208
|
-
output: 0.4,
|
|
15209
|
-
cacheRead: 0.01,
|
|
15210
|
-
cacheWrite: 0,
|
|
15211
|
-
},
|
|
15212
|
-
contextWindow: 200000,
|
|
15213
|
-
maxTokens: 131072,
|
|
15214
|
-
},
|
|
15215
|
-
"glm-5": {
|
|
15216
|
-
id: "glm-5",
|
|
15217
|
-
name: "GLM-5",
|
|
15218
|
-
api: "openai-completions",
|
|
15219
|
-
provider: "zai",
|
|
15220
|
-
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
15221
|
-
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
15222
|
-
reasoning: true,
|
|
15223
|
-
input: ["text"],
|
|
15224
|
-
cost: {
|
|
15225
|
-
input: 0,
|
|
15226
|
-
output: 0,
|
|
15227
|
-
cacheRead: 0,
|
|
15228
|
-
cacheWrite: 0,
|
|
15229
|
-
},
|
|
15230
|
-
contextWindow: 204800,
|
|
15231
|
-
maxTokens: 131072,
|
|
15232
|
-
},
|
|
15233
15568
|
"glm-5-turbo": {
|
|
15234
15569
|
id: "glm-5-turbo",
|
|
15235
15570
|
name: "GLM-5-Turbo",
|