@hyperspaceng/neural-ai 0.70.3 → 0.70.6
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +5 -3
- package/dist/env-api-keys.d.ts.map +1 -1
- package/dist/env-api-keys.js +50 -7
- package/dist/env-api-keys.js.map +1 -1
- package/dist/models.generated.d.ts +479 -20
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +490 -53
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/amazon-bedrock.d.ts.map +1 -1
- package/dist/providers/amazon-bedrock.js +49 -25
- package/dist/providers/amazon-bedrock.js.map +1 -1
- package/dist/providers/anthropic.d.ts.map +1 -1
- package/dist/providers/anthropic.js +11 -3
- package/dist/providers/anthropic.js.map +1 -1
- package/dist/providers/azure-openai-responses.d.ts.map +1 -1
- package/dist/providers/azure-openai-responses.js +17 -1
- package/dist/providers/azure-openai-responses.js.map +1 -1
- package/dist/providers/cloudflare.d.ts +7 -0
- package/dist/providers/cloudflare.d.ts.map +1 -0
- package/dist/providers/cloudflare.js +19 -0
- package/dist/providers/cloudflare.js.map +1 -0
- package/dist/providers/openai-codex-responses.d.ts.map +1 -1
- package/dist/providers/openai-codex-responses.js +1 -1
- package/dist/providers/openai-codex-responses.js.map +1 -1
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +5 -2
- package/dist/providers/openai-completions.js.map +1 -1
- package/dist/types.d.ts +1 -1
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/package.json +1 -1
package/dist/models.generated.js
CHANGED
|
@@ -2539,6 +2539,23 @@ export const MODELS = {
|
|
|
2539
2539
|
contextWindow: 272000,
|
|
2540
2540
|
maxTokens: 128000,
|
|
2541
2541
|
},
|
|
2542
|
+
"gpt-5.5-pro": {
|
|
2543
|
+
id: "gpt-5.5-pro",
|
|
2544
|
+
name: "GPT-5.5 Pro",
|
|
2545
|
+
api: "azure-openai-responses",
|
|
2546
|
+
provider: "azure-openai-responses",
|
|
2547
|
+
baseUrl: "",
|
|
2548
|
+
reasoning: true,
|
|
2549
|
+
input: ["text", "image"],
|
|
2550
|
+
cost: {
|
|
2551
|
+
input: 30,
|
|
2552
|
+
output: 180,
|
|
2553
|
+
cacheRead: 0,
|
|
2554
|
+
cacheWrite: 0,
|
|
2555
|
+
},
|
|
2556
|
+
contextWindow: 1050000,
|
|
2557
|
+
maxTokens: 128000,
|
|
2558
|
+
},
|
|
2542
2559
|
"o1": {
|
|
2543
2560
|
id: "o1",
|
|
2544
2561
|
name: "o1",
|
|
@@ -2746,6 +2763,152 @@ export const MODELS = {
|
|
|
2746
2763
|
maxTokens: 40000,
|
|
2747
2764
|
},
|
|
2748
2765
|
},
|
|
2766
|
+
"cloudflare-workers-ai": {
|
|
2767
|
+
"@cf/google/gemma-4-26b-a4b-it": {
|
|
2768
|
+
id: "@cf/google/gemma-4-26b-a4b-it",
|
|
2769
|
+
name: "Gemma 4 26B A4B IT",
|
|
2770
|
+
api: "openai-completions",
|
|
2771
|
+
provider: "cloudflare-workers-ai",
|
|
2772
|
+
baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
|
|
2773
|
+
compat: { "sendSessionAffinityHeaders": true },
|
|
2774
|
+
reasoning: true,
|
|
2775
|
+
input: ["text", "image"],
|
|
2776
|
+
cost: {
|
|
2777
|
+
input: 0.1,
|
|
2778
|
+
output: 0.3,
|
|
2779
|
+
cacheRead: 0,
|
|
2780
|
+
cacheWrite: 0,
|
|
2781
|
+
},
|
|
2782
|
+
contextWindow: 256000,
|
|
2783
|
+
maxTokens: 16384,
|
|
2784
|
+
},
|
|
2785
|
+
"@cf/meta/llama-4-scout-17b-16e-instruct": {
|
|
2786
|
+
id: "@cf/meta/llama-4-scout-17b-16e-instruct",
|
|
2787
|
+
name: "Llama 4 Scout 17B 16E Instruct",
|
|
2788
|
+
api: "openai-completions",
|
|
2789
|
+
provider: "cloudflare-workers-ai",
|
|
2790
|
+
baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
|
|
2791
|
+
compat: { "sendSessionAffinityHeaders": true },
|
|
2792
|
+
reasoning: false,
|
|
2793
|
+
input: ["text", "image"],
|
|
2794
|
+
cost: {
|
|
2795
|
+
input: 0.27,
|
|
2796
|
+
output: 0.85,
|
|
2797
|
+
cacheRead: 0,
|
|
2798
|
+
cacheWrite: 0,
|
|
2799
|
+
},
|
|
2800
|
+
contextWindow: 128000,
|
|
2801
|
+
maxTokens: 16384,
|
|
2802
|
+
},
|
|
2803
|
+
"@cf/moonshotai/kimi-k2.5": {
|
|
2804
|
+
id: "@cf/moonshotai/kimi-k2.5",
|
|
2805
|
+
name: "Kimi K2.5",
|
|
2806
|
+
api: "openai-completions",
|
|
2807
|
+
provider: "cloudflare-workers-ai",
|
|
2808
|
+
baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
|
|
2809
|
+
compat: { "sendSessionAffinityHeaders": true },
|
|
2810
|
+
reasoning: true,
|
|
2811
|
+
input: ["text", "image"],
|
|
2812
|
+
cost: {
|
|
2813
|
+
input: 0.6,
|
|
2814
|
+
output: 3,
|
|
2815
|
+
cacheRead: 0.1,
|
|
2816
|
+
cacheWrite: 0,
|
|
2817
|
+
},
|
|
2818
|
+
contextWindow: 256000,
|
|
2819
|
+
maxTokens: 256000,
|
|
2820
|
+
},
|
|
2821
|
+
"@cf/moonshotai/kimi-k2.6": {
|
|
2822
|
+
id: "@cf/moonshotai/kimi-k2.6",
|
|
2823
|
+
name: "Kimi K2.6",
|
|
2824
|
+
api: "openai-completions",
|
|
2825
|
+
provider: "cloudflare-workers-ai",
|
|
2826
|
+
baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
|
|
2827
|
+
compat: { "sendSessionAffinityHeaders": true },
|
|
2828
|
+
reasoning: true,
|
|
2829
|
+
input: ["text", "image"],
|
|
2830
|
+
cost: {
|
|
2831
|
+
input: 0.95,
|
|
2832
|
+
output: 4,
|
|
2833
|
+
cacheRead: 0.16,
|
|
2834
|
+
cacheWrite: 0,
|
|
2835
|
+
},
|
|
2836
|
+
contextWindow: 256000,
|
|
2837
|
+
maxTokens: 256000,
|
|
2838
|
+
},
|
|
2839
|
+
"@cf/nvidia/nemotron-3-120b-a12b": {
|
|
2840
|
+
id: "@cf/nvidia/nemotron-3-120b-a12b",
|
|
2841
|
+
name: "Nemotron 3 Super 120B",
|
|
2842
|
+
api: "openai-completions",
|
|
2843
|
+
provider: "cloudflare-workers-ai",
|
|
2844
|
+
baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
|
|
2845
|
+
compat: { "sendSessionAffinityHeaders": true },
|
|
2846
|
+
reasoning: true,
|
|
2847
|
+
input: ["text"],
|
|
2848
|
+
cost: {
|
|
2849
|
+
input: 0.5,
|
|
2850
|
+
output: 1.5,
|
|
2851
|
+
cacheRead: 0,
|
|
2852
|
+
cacheWrite: 0,
|
|
2853
|
+
},
|
|
2854
|
+
contextWindow: 256000,
|
|
2855
|
+
maxTokens: 256000,
|
|
2856
|
+
},
|
|
2857
|
+
"@cf/openai/gpt-oss-120b": {
|
|
2858
|
+
id: "@cf/openai/gpt-oss-120b",
|
|
2859
|
+
name: "GPT OSS 120B",
|
|
2860
|
+
api: "openai-completions",
|
|
2861
|
+
provider: "cloudflare-workers-ai",
|
|
2862
|
+
baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
|
|
2863
|
+
compat: { "sendSessionAffinityHeaders": true },
|
|
2864
|
+
reasoning: true,
|
|
2865
|
+
input: ["text"],
|
|
2866
|
+
cost: {
|
|
2867
|
+
input: 0.35,
|
|
2868
|
+
output: 0.75,
|
|
2869
|
+
cacheRead: 0,
|
|
2870
|
+
cacheWrite: 0,
|
|
2871
|
+
},
|
|
2872
|
+
contextWindow: 128000,
|
|
2873
|
+
maxTokens: 16384,
|
|
2874
|
+
},
|
|
2875
|
+
"@cf/openai/gpt-oss-20b": {
|
|
2876
|
+
id: "@cf/openai/gpt-oss-20b",
|
|
2877
|
+
name: "GPT OSS 20B",
|
|
2878
|
+
api: "openai-completions",
|
|
2879
|
+
provider: "cloudflare-workers-ai",
|
|
2880
|
+
baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
|
|
2881
|
+
compat: { "sendSessionAffinityHeaders": true },
|
|
2882
|
+
reasoning: true,
|
|
2883
|
+
input: ["text"],
|
|
2884
|
+
cost: {
|
|
2885
|
+
input: 0.2,
|
|
2886
|
+
output: 0.3,
|
|
2887
|
+
cacheRead: 0,
|
|
2888
|
+
cacheWrite: 0,
|
|
2889
|
+
},
|
|
2890
|
+
contextWindow: 128000,
|
|
2891
|
+
maxTokens: 16384,
|
|
2892
|
+
},
|
|
2893
|
+
"@cf/zai-org/glm-4.7-flash": {
|
|
2894
|
+
id: "@cf/zai-org/glm-4.7-flash",
|
|
2895
|
+
name: "GLM-4.7-Flash",
|
|
2896
|
+
api: "openai-completions",
|
|
2897
|
+
provider: "cloudflare-workers-ai",
|
|
2898
|
+
baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
|
|
2899
|
+
compat: { "sendSessionAffinityHeaders": true },
|
|
2900
|
+
reasoning: true,
|
|
2901
|
+
input: ["text"],
|
|
2902
|
+
cost: {
|
|
2903
|
+
input: 0.06,
|
|
2904
|
+
output: 0.4,
|
|
2905
|
+
cacheRead: 0,
|
|
2906
|
+
cacheWrite: 0,
|
|
2907
|
+
},
|
|
2908
|
+
contextWindow: 131072,
|
|
2909
|
+
maxTokens: 131072,
|
|
2910
|
+
},
|
|
2911
|
+
},
|
|
2749
2912
|
"deepseek": {
|
|
2750
2913
|
"deepseek-v4-flash": {
|
|
2751
2914
|
id: "deepseek-v4-flash",
|
|
@@ -2819,6 +2982,23 @@ export const MODELS = {
|
|
|
2819
2982
|
contextWindow: 160000,
|
|
2820
2983
|
maxTokens: 160000,
|
|
2821
2984
|
},
|
|
2985
|
+
"accounts/fireworks/models/deepseek-v4-pro": {
|
|
2986
|
+
id: "accounts/fireworks/models/deepseek-v4-pro",
|
|
2987
|
+
name: "DeepSeek V4 Pro",
|
|
2988
|
+
api: "anthropic-messages",
|
|
2989
|
+
provider: "fireworks",
|
|
2990
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2991
|
+
reasoning: true,
|
|
2992
|
+
input: ["text"],
|
|
2993
|
+
cost: {
|
|
2994
|
+
input: 1.74,
|
|
2995
|
+
output: 3.48,
|
|
2996
|
+
cacheRead: 0.15,
|
|
2997
|
+
cacheWrite: 0,
|
|
2998
|
+
},
|
|
2999
|
+
contextWindow: 1000000,
|
|
3000
|
+
maxTokens: 384000,
|
|
3001
|
+
},
|
|
2822
3002
|
"accounts/fireworks/models/glm-4p5": {
|
|
2823
3003
|
id: "accounts/fireworks/models/glm-4p5",
|
|
2824
3004
|
name: "GLM 4.5",
|
|
@@ -5057,6 +5237,24 @@ export const MODELS = {
|
|
|
5057
5237
|
contextWindow: 163840,
|
|
5058
5238
|
maxTokens: 65536,
|
|
5059
5239
|
},
|
|
5240
|
+
"deepseek-ai/DeepSeek-V4-Pro": {
|
|
5241
|
+
id: "deepseek-ai/DeepSeek-V4-Pro",
|
|
5242
|
+
name: "DeepSeek V4 Pro",
|
|
5243
|
+
api: "openai-completions",
|
|
5244
|
+
provider: "huggingface",
|
|
5245
|
+
baseUrl: "https://router.huggingface.co/v1",
|
|
5246
|
+
compat: { "supportsDeveloperRole": false },
|
|
5247
|
+
reasoning: true,
|
|
5248
|
+
input: ["text"],
|
|
5249
|
+
cost: {
|
|
5250
|
+
input: 1.74,
|
|
5251
|
+
output: 3.48,
|
|
5252
|
+
cacheRead: 0.145,
|
|
5253
|
+
cacheWrite: 0,
|
|
5254
|
+
},
|
|
5255
|
+
contextWindow: 1048576,
|
|
5256
|
+
maxTokens: 393216,
|
|
5257
|
+
},
|
|
5060
5258
|
"moonshotai/Kimi-K2-Instruct": {
|
|
5061
5259
|
id: "moonshotai/Kimi-K2-Instruct",
|
|
5062
5260
|
name: "Kimi-K2-Instruct",
|
|
@@ -6354,6 +6552,23 @@ export const MODELS = {
|
|
|
6354
6552
|
contextWindow: 272000,
|
|
6355
6553
|
maxTokens: 128000,
|
|
6356
6554
|
},
|
|
6555
|
+
"gpt-5.5-pro": {
|
|
6556
|
+
id: "gpt-5.5-pro",
|
|
6557
|
+
name: "GPT-5.5 Pro",
|
|
6558
|
+
api: "openai-responses",
|
|
6559
|
+
provider: "openai",
|
|
6560
|
+
baseUrl: "https://api.openai.com/v1",
|
|
6561
|
+
reasoning: true,
|
|
6562
|
+
input: ["text", "image"],
|
|
6563
|
+
cost: {
|
|
6564
|
+
input: 30,
|
|
6565
|
+
output: 180,
|
|
6566
|
+
cacheRead: 0,
|
|
6567
|
+
cacheWrite: 0,
|
|
6568
|
+
},
|
|
6569
|
+
contextWindow: 1050000,
|
|
6570
|
+
maxTokens: 128000,
|
|
6571
|
+
},
|
|
6357
6572
|
"o1": {
|
|
6358
6573
|
id: "o1",
|
|
6359
6574
|
name: "o1",
|
|
@@ -6681,23 +6896,6 @@ export const MODELS = {
|
|
|
6681
6896
|
contextWindow: 200000,
|
|
6682
6897
|
maxTokens: 128000,
|
|
6683
6898
|
},
|
|
6684
|
-
"claude-3-5-haiku": {
|
|
6685
|
-
id: "claude-3-5-haiku",
|
|
6686
|
-
name: "Claude Haiku 3.5",
|
|
6687
|
-
api: "anthropic-messages",
|
|
6688
|
-
provider: "opencode",
|
|
6689
|
-
baseUrl: "https://opencode.ai/zen",
|
|
6690
|
-
reasoning: false,
|
|
6691
|
-
input: ["text", "image"],
|
|
6692
|
-
cost: {
|
|
6693
|
-
input: 0.8,
|
|
6694
|
-
output: 4,
|
|
6695
|
-
cacheRead: 0.08,
|
|
6696
|
-
cacheWrite: 1,
|
|
6697
|
-
},
|
|
6698
|
-
contextWindow: 200000,
|
|
6699
|
-
maxTokens: 8192,
|
|
6700
|
-
},
|
|
6701
6899
|
"claude-haiku-4-5": {
|
|
6702
6900
|
id: "claude-haiku-4-5",
|
|
6703
6901
|
name: "Claude Haiku 4.5",
|
|
@@ -7155,7 +7353,7 @@ export const MODELS = {
|
|
|
7155
7353
|
cacheWrite: 0,
|
|
7156
7354
|
},
|
|
7157
7355
|
contextWindow: 1050000,
|
|
7158
|
-
maxTokens:
|
|
7356
|
+
maxTokens: 128000,
|
|
7159
7357
|
},
|
|
7160
7358
|
"gpt-5.5-pro": {
|
|
7161
7359
|
id: "gpt-5.5-pro",
|
|
@@ -7358,7 +7556,7 @@ export const MODELS = {
|
|
|
7358
7556
|
cost: {
|
|
7359
7557
|
input: 0.14,
|
|
7360
7558
|
output: 0.28,
|
|
7361
|
-
cacheRead: 0.
|
|
7559
|
+
cacheRead: 0.0028,
|
|
7362
7560
|
cacheWrite: 0,
|
|
7363
7561
|
},
|
|
7364
7562
|
contextWindow: 1000000,
|
|
@@ -7376,7 +7574,7 @@ export const MODELS = {
|
|
|
7376
7574
|
cost: {
|
|
7377
7575
|
input: 1.74,
|
|
7378
7576
|
output: 3.48,
|
|
7379
|
-
cacheRead: 0.
|
|
7577
|
+
cacheRead: 0.0145,
|
|
7380
7578
|
cacheWrite: 0,
|
|
7381
7579
|
},
|
|
7382
7580
|
contextWindow: 1000000,
|
|
@@ -7396,8 +7594,8 @@ export const MODELS = {
|
|
|
7396
7594
|
cacheRead: 0.2,
|
|
7397
7595
|
cacheWrite: 0,
|
|
7398
7596
|
},
|
|
7399
|
-
contextWindow:
|
|
7400
|
-
maxTokens:
|
|
7597
|
+
contextWindow: 202752,
|
|
7598
|
+
maxTokens: 32768,
|
|
7401
7599
|
},
|
|
7402
7600
|
"glm-5.1": {
|
|
7403
7601
|
id: "glm-5.1",
|
|
@@ -7413,8 +7611,8 @@ export const MODELS = {
|
|
|
7413
7611
|
cacheRead: 0.26,
|
|
7414
7612
|
cacheWrite: 0,
|
|
7415
7613
|
},
|
|
7416
|
-
contextWindow:
|
|
7417
|
-
maxTokens:
|
|
7614
|
+
contextWindow: 202752,
|
|
7615
|
+
maxTokens: 32768,
|
|
7418
7616
|
},
|
|
7419
7617
|
"kimi-k2.5": {
|
|
7420
7618
|
id: "kimi-k2.5",
|
|
@@ -7498,7 +7696,7 @@ export const MODELS = {
|
|
|
7498
7696
|
cacheRead: 0.08,
|
|
7499
7697
|
cacheWrite: 0,
|
|
7500
7698
|
},
|
|
7501
|
-
contextWindow:
|
|
7699
|
+
contextWindow: 1000000,
|
|
7502
7700
|
maxTokens: 128000,
|
|
7503
7701
|
},
|
|
7504
7702
|
"mimo-v2.5-pro": {
|
|
@@ -8331,7 +8529,7 @@ export const MODELS = {
|
|
|
8331
8529
|
cost: {
|
|
8332
8530
|
input: 0.14,
|
|
8333
8531
|
output: 0.28,
|
|
8334
|
-
cacheRead: 0.
|
|
8532
|
+
cacheRead: 0.0028,
|
|
8335
8533
|
cacheWrite: 0,
|
|
8336
8534
|
},
|
|
8337
8535
|
contextWindow: 1048576,
|
|
@@ -8349,7 +8547,7 @@ export const MODELS = {
|
|
|
8349
8547
|
cost: {
|
|
8350
8548
|
input: 0.435,
|
|
8351
8549
|
output: 0.87,
|
|
8352
|
-
cacheRead: 0.
|
|
8550
|
+
cacheRead: 0.003625,
|
|
8353
8551
|
cacheWrite: 0,
|
|
8354
8552
|
},
|
|
8355
8553
|
contextWindow: 1048576,
|
|
@@ -8712,18 +8910,18 @@ export const MODELS = {
|
|
|
8712
8910
|
contextWindow: 262144,
|
|
8713
8911
|
maxTokens: 32768,
|
|
8714
8912
|
},
|
|
8715
|
-
"inclusionai/ling-2.6-flash
|
|
8716
|
-
id: "inclusionai/ling-2.6-flash
|
|
8717
|
-
name: "inclusionAI: Ling-2.6-flash
|
|
8913
|
+
"inclusionai/ling-2.6-flash": {
|
|
8914
|
+
id: "inclusionai/ling-2.6-flash",
|
|
8915
|
+
name: "inclusionAI: Ling-2.6-flash",
|
|
8718
8916
|
api: "openai-completions",
|
|
8719
8917
|
provider: "openrouter",
|
|
8720
8918
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
8721
8919
|
reasoning: false,
|
|
8722
8920
|
input: ["text"],
|
|
8723
8921
|
cost: {
|
|
8724
|
-
input: 0,
|
|
8725
|
-
output: 0,
|
|
8726
|
-
cacheRead: 0,
|
|
8922
|
+
input: 0.08,
|
|
8923
|
+
output: 0.24,
|
|
8924
|
+
cacheRead: 0.016,
|
|
8727
8925
|
cacheWrite: 0,
|
|
8728
8926
|
},
|
|
8729
8927
|
contextWindow: 262144,
|
|
@@ -9180,7 +9378,7 @@ export const MODELS = {
|
|
|
9180
9378
|
reasoning: false,
|
|
9181
9379
|
input: ["text"],
|
|
9182
9380
|
cost: {
|
|
9183
|
-
input: 0.
|
|
9381
|
+
input: 0.02,
|
|
9184
9382
|
output: 0.03,
|
|
9185
9383
|
cacheRead: 0,
|
|
9186
9384
|
cacheWrite: 0,
|
|
@@ -9494,6 +9692,23 @@ export const MODELS = {
|
|
|
9494
9692
|
contextWindow: 256000,
|
|
9495
9693
|
maxTokens: 4096,
|
|
9496
9694
|
},
|
|
9695
|
+
"nvidia/nemotron-3-nano-omni-30b-a3b-reasoning:free": {
|
|
9696
|
+
id: "nvidia/nemotron-3-nano-omni-30b-a3b-reasoning:free",
|
|
9697
|
+
name: "NVIDIA: Nemotron 3 Nano Omni (free)",
|
|
9698
|
+
api: "openai-completions",
|
|
9699
|
+
provider: "openrouter",
|
|
9700
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
9701
|
+
reasoning: true,
|
|
9702
|
+
input: ["text", "image"],
|
|
9703
|
+
cost: {
|
|
9704
|
+
input: 0,
|
|
9705
|
+
output: 0,
|
|
9706
|
+
cacheRead: 0,
|
|
9707
|
+
cacheWrite: 0,
|
|
9708
|
+
},
|
|
9709
|
+
contextWindow: 256000,
|
|
9710
|
+
maxTokens: 65536,
|
|
9711
|
+
},
|
|
9497
9712
|
"nvidia/nemotron-3-super-120b-a12b": {
|
|
9498
9713
|
id: "nvidia/nemotron-3-super-120b-a12b",
|
|
9499
9714
|
name: "NVIDIA: Nemotron 3 Super",
|
|
@@ -10565,6 +10780,40 @@ export const MODELS = {
|
|
|
10565
10780
|
contextWindow: 200000,
|
|
10566
10781
|
maxTokens: 4096,
|
|
10567
10782
|
},
|
|
10783
|
+
"poolside/laguna-m.1:free": {
|
|
10784
|
+
id: "poolside/laguna-m.1:free",
|
|
10785
|
+
name: "Poolside: Laguna M.1 (free)",
|
|
10786
|
+
api: "openai-completions",
|
|
10787
|
+
provider: "openrouter",
|
|
10788
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
10789
|
+
reasoning: true,
|
|
10790
|
+
input: ["text"],
|
|
10791
|
+
cost: {
|
|
10792
|
+
input: 0,
|
|
10793
|
+
output: 0,
|
|
10794
|
+
cacheRead: 0,
|
|
10795
|
+
cacheWrite: 0,
|
|
10796
|
+
},
|
|
10797
|
+
contextWindow: 131072,
|
|
10798
|
+
maxTokens: 8192,
|
|
10799
|
+
},
|
|
10800
|
+
"poolside/laguna-xs.2:free": {
|
|
10801
|
+
id: "poolside/laguna-xs.2:free",
|
|
10802
|
+
name: "Poolside: Laguna XS.2 (free)",
|
|
10803
|
+
api: "openai-completions",
|
|
10804
|
+
provider: "openrouter",
|
|
10805
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
10806
|
+
reasoning: true,
|
|
10807
|
+
input: ["text"],
|
|
10808
|
+
cost: {
|
|
10809
|
+
input: 0,
|
|
10810
|
+
output: 0,
|
|
10811
|
+
cacheRead: 0,
|
|
10812
|
+
cacheWrite: 0,
|
|
10813
|
+
},
|
|
10814
|
+
contextWindow: 131072,
|
|
10815
|
+
maxTokens: 8192,
|
|
10816
|
+
},
|
|
10568
10817
|
"prime-intellect/intellect-3": {
|
|
10569
10818
|
id: "prime-intellect/intellect-3",
|
|
10570
10819
|
name: "Prime Intellect: INTELLECT-3",
|
|
@@ -10591,8 +10840,8 @@ export const MODELS = {
|
|
|
10591
10840
|
reasoning: false,
|
|
10592
10841
|
input: ["text"],
|
|
10593
10842
|
cost: {
|
|
10594
|
-
input: 0.
|
|
10595
|
-
output: 0.
|
|
10843
|
+
input: 0.36,
|
|
10844
|
+
output: 0.39999999999999997,
|
|
10596
10845
|
cacheRead: 0,
|
|
10597
10846
|
cacheWrite: 0,
|
|
10598
10847
|
},
|
|
@@ -11296,9 +11545,26 @@ export const MODELS = {
|
|
|
11296
11545
|
contextWindow: 1000000,
|
|
11297
11546
|
maxTokens: 65536,
|
|
11298
11547
|
},
|
|
11299
|
-
"qwen/qwen3.
|
|
11300
|
-
id: "qwen/qwen3.
|
|
11301
|
-
name: "Qwen: Qwen3.
|
|
11548
|
+
"qwen/qwen3.5-plus-20260420": {
|
|
11549
|
+
id: "qwen/qwen3.5-plus-20260420",
|
|
11550
|
+
name: "Qwen: Qwen3.5 Plus 2026-04-20",
|
|
11551
|
+
api: "openai-completions",
|
|
11552
|
+
provider: "openrouter",
|
|
11553
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
11554
|
+
reasoning: true,
|
|
11555
|
+
input: ["text", "image"],
|
|
11556
|
+
cost: {
|
|
11557
|
+
input: 0.39999999999999997,
|
|
11558
|
+
output: 2.4,
|
|
11559
|
+
cacheRead: 0,
|
|
11560
|
+
cacheWrite: 0,
|
|
11561
|
+
},
|
|
11562
|
+
contextWindow: 1000000,
|
|
11563
|
+
maxTokens: 65536,
|
|
11564
|
+
},
|
|
11565
|
+
"qwen/qwen3.6-27b": {
|
|
11566
|
+
id: "qwen/qwen3.6-27b",
|
|
11567
|
+
name: "Qwen: Qwen3.6 27B",
|
|
11302
11568
|
api: "openai-completions",
|
|
11303
11569
|
provider: "openrouter",
|
|
11304
11570
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -11306,29 +11572,63 @@ export const MODELS = {
|
|
|
11306
11572
|
input: ["text", "image"],
|
|
11307
11573
|
cost: {
|
|
11308
11574
|
input: 0.325,
|
|
11309
|
-
output:
|
|
11575
|
+
output: 3.25,
|
|
11310
11576
|
cacheRead: 0,
|
|
11311
|
-
cacheWrite: 0
|
|
11577
|
+
cacheWrite: 0,
|
|
11578
|
+
},
|
|
11579
|
+
contextWindow: 256000,
|
|
11580
|
+
maxTokens: 65536,
|
|
11581
|
+
},
|
|
11582
|
+
"qwen/qwen3.6-flash": {
|
|
11583
|
+
id: "qwen/qwen3.6-flash",
|
|
11584
|
+
name: "Qwen: Qwen3.6 Flash",
|
|
11585
|
+
api: "openai-completions",
|
|
11586
|
+
provider: "openrouter",
|
|
11587
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
11588
|
+
reasoning: true,
|
|
11589
|
+
input: ["text", "image"],
|
|
11590
|
+
cost: {
|
|
11591
|
+
input: 0.25,
|
|
11592
|
+
output: 1.5,
|
|
11593
|
+
cacheRead: 0,
|
|
11594
|
+
cacheWrite: 0.3125,
|
|
11312
11595
|
},
|
|
11313
11596
|
contextWindow: 1000000,
|
|
11314
11597
|
maxTokens: 65536,
|
|
11315
11598
|
},
|
|
11316
|
-
"qwen/
|
|
11317
|
-
id: "qwen/
|
|
11318
|
-
name: "Qwen:
|
|
11599
|
+
"qwen/qwen3.6-max-preview": {
|
|
11600
|
+
id: "qwen/qwen3.6-max-preview",
|
|
11601
|
+
name: "Qwen: Qwen3.6 Max Preview",
|
|
11319
11602
|
api: "openai-completions",
|
|
11320
11603
|
provider: "openrouter",
|
|
11321
11604
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
11322
11605
|
reasoning: true,
|
|
11323
11606
|
input: ["text"],
|
|
11324
11607
|
cost: {
|
|
11325
|
-
input:
|
|
11326
|
-
output:
|
|
11608
|
+
input: 1.04,
|
|
11609
|
+
output: 6.24,
|
|
11327
11610
|
cacheRead: 0,
|
|
11328
|
-
cacheWrite:
|
|
11611
|
+
cacheWrite: 1.3,
|
|
11329
11612
|
},
|
|
11330
|
-
contextWindow:
|
|
11331
|
-
maxTokens:
|
|
11613
|
+
contextWindow: 262144,
|
|
11614
|
+
maxTokens: 65536,
|
|
11615
|
+
},
|
|
11616
|
+
"qwen/qwen3.6-plus": {
|
|
11617
|
+
id: "qwen/qwen3.6-plus",
|
|
11618
|
+
name: "Qwen: Qwen3.6 Plus",
|
|
11619
|
+
api: "openai-completions",
|
|
11620
|
+
provider: "openrouter",
|
|
11621
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
11622
|
+
reasoning: true,
|
|
11623
|
+
input: ["text", "image"],
|
|
11624
|
+
cost: {
|
|
11625
|
+
input: 0.325,
|
|
11626
|
+
output: 1.95,
|
|
11627
|
+
cacheRead: 0,
|
|
11628
|
+
cacheWrite: 0.40625,
|
|
11629
|
+
},
|
|
11630
|
+
contextWindow: 1000000,
|
|
11631
|
+
maxTokens: 65536,
|
|
11332
11632
|
},
|
|
11333
11633
|
"rekaai/reka-edge": {
|
|
11334
11634
|
id: "rekaai/reka-edge",
|
|
@@ -11959,6 +12259,23 @@ export const MODELS = {
|
|
|
11959
12259
|
contextWindow: 202752,
|
|
11960
12260
|
maxTokens: 131072,
|
|
11961
12261
|
},
|
|
12262
|
+
"~anthropic/claude-haiku-latest": {
|
|
12263
|
+
id: "~anthropic/claude-haiku-latest",
|
|
12264
|
+
name: "Anthropic Claude Haiku Latest",
|
|
12265
|
+
api: "openai-completions",
|
|
12266
|
+
provider: "openrouter",
|
|
12267
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
12268
|
+
reasoning: true,
|
|
12269
|
+
input: ["text", "image"],
|
|
12270
|
+
cost: {
|
|
12271
|
+
input: 1,
|
|
12272
|
+
output: 5,
|
|
12273
|
+
cacheRead: 0.09999999999999999,
|
|
12274
|
+
cacheWrite: 1.25,
|
|
12275
|
+
},
|
|
12276
|
+
contextWindow: 200000,
|
|
12277
|
+
maxTokens: 64000,
|
|
12278
|
+
},
|
|
11962
12279
|
"~anthropic/claude-opus-latest": {
|
|
11963
12280
|
id: "~anthropic/claude-opus-latest",
|
|
11964
12281
|
name: "Anthropic: Claude Opus Latest",
|
|
@@ -11976,6 +12293,108 @@ export const MODELS = {
|
|
|
11976
12293
|
contextWindow: 1000000,
|
|
11977
12294
|
maxTokens: 128000,
|
|
11978
12295
|
},
|
|
12296
|
+
"~anthropic/claude-sonnet-latest": {
|
|
12297
|
+
id: "~anthropic/claude-sonnet-latest",
|
|
12298
|
+
name: "Anthropic Claude Sonnet Latest",
|
|
12299
|
+
api: "openai-completions",
|
|
12300
|
+
provider: "openrouter",
|
|
12301
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
12302
|
+
reasoning: true,
|
|
12303
|
+
input: ["text", "image"],
|
|
12304
|
+
cost: {
|
|
12305
|
+
input: 3,
|
|
12306
|
+
output: 15,
|
|
12307
|
+
cacheRead: 0.3,
|
|
12308
|
+
cacheWrite: 3.75,
|
|
12309
|
+
},
|
|
12310
|
+
contextWindow: 1000000,
|
|
12311
|
+
maxTokens: 128000,
|
|
12312
|
+
},
|
|
12313
|
+
"~google/gemini-flash-latest": {
|
|
12314
|
+
id: "~google/gemini-flash-latest",
|
|
12315
|
+
name: "Google Gemini Flash Latest",
|
|
12316
|
+
api: "openai-completions",
|
|
12317
|
+
provider: "openrouter",
|
|
12318
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
12319
|
+
reasoning: true,
|
|
12320
|
+
input: ["text", "image"],
|
|
12321
|
+
cost: {
|
|
12322
|
+
input: 0.5,
|
|
12323
|
+
output: 3,
|
|
12324
|
+
cacheRead: 0.049999999999999996,
|
|
12325
|
+
cacheWrite: 0.08333333333333334,
|
|
12326
|
+
},
|
|
12327
|
+
contextWindow: 1048576,
|
|
12328
|
+
maxTokens: 65536,
|
|
12329
|
+
},
|
|
12330
|
+
"~google/gemini-pro-latest": {
|
|
12331
|
+
id: "~google/gemini-pro-latest",
|
|
12332
|
+
name: "Google Gemini Pro Latest",
|
|
12333
|
+
api: "openai-completions",
|
|
12334
|
+
provider: "openrouter",
|
|
12335
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
12336
|
+
reasoning: true,
|
|
12337
|
+
input: ["text", "image"],
|
|
12338
|
+
cost: {
|
|
12339
|
+
input: 2,
|
|
12340
|
+
output: 12,
|
|
12341
|
+
cacheRead: 0.19999999999999998,
|
|
12342
|
+
cacheWrite: 0.375,
|
|
12343
|
+
},
|
|
12344
|
+
contextWindow: 1048576,
|
|
12345
|
+
maxTokens: 65536,
|
|
12346
|
+
},
|
|
12347
|
+
"~moonshotai/kimi-latest": {
|
|
12348
|
+
id: "~moonshotai/kimi-latest",
|
|
12349
|
+
name: "MoonshotAI Kimi Latest",
|
|
12350
|
+
api: "openai-completions",
|
|
12351
|
+
provider: "openrouter",
|
|
12352
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
12353
|
+
reasoning: true,
|
|
12354
|
+
input: ["text", "image"],
|
|
12355
|
+
cost: {
|
|
12356
|
+
input: 0.7448,
|
|
12357
|
+
output: 4.655,
|
|
12358
|
+
cacheRead: 0.1463,
|
|
12359
|
+
cacheWrite: 0,
|
|
12360
|
+
},
|
|
12361
|
+
contextWindow: 256000,
|
|
12362
|
+
maxTokens: 65536,
|
|
12363
|
+
},
|
|
12364
|
+
"~openai/gpt-latest": {
|
|
12365
|
+
id: "~openai/gpt-latest",
|
|
12366
|
+
name: "OpenAI GPT Latest",
|
|
12367
|
+
api: "openai-completions",
|
|
12368
|
+
provider: "openrouter",
|
|
12369
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
12370
|
+
reasoning: true,
|
|
12371
|
+
input: ["text", "image"],
|
|
12372
|
+
cost: {
|
|
12373
|
+
input: 5,
|
|
12374
|
+
output: 30,
|
|
12375
|
+
cacheRead: 0.5,
|
|
12376
|
+
cacheWrite: 0,
|
|
12377
|
+
},
|
|
12378
|
+
contextWindow: 1050000,
|
|
12379
|
+
maxTokens: 128000,
|
|
12380
|
+
},
|
|
12381
|
+
"~openai/gpt-mini-latest": {
|
|
12382
|
+
id: "~openai/gpt-mini-latest",
|
|
12383
|
+
name: "OpenAI GPT Mini Latest",
|
|
12384
|
+
api: "openai-completions",
|
|
12385
|
+
provider: "openrouter",
|
|
12386
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
12387
|
+
reasoning: true,
|
|
12388
|
+
input: ["text", "image"],
|
|
12389
|
+
cost: {
|
|
12390
|
+
input: 0.75,
|
|
12391
|
+
output: 4.5,
|
|
12392
|
+
cacheRead: 0.075,
|
|
12393
|
+
cacheWrite: 0,
|
|
12394
|
+
},
|
|
12395
|
+
contextWindow: 400000,
|
|
12396
|
+
maxTokens: 128000,
|
|
12397
|
+
},
|
|
11979
12398
|
},
|
|
11980
12399
|
"vercel-ai-gateway": {
|
|
11981
12400
|
"alibaba/qwen-3-14b": {
|
|
@@ -12652,7 +13071,7 @@ export const MODELS = {
|
|
|
12652
13071
|
cost: {
|
|
12653
13072
|
input: 0.14,
|
|
12654
13073
|
output: 0.28,
|
|
12655
|
-
cacheRead: 0.
|
|
13074
|
+
cacheRead: 0.0028,
|
|
12656
13075
|
cacheWrite: 0,
|
|
12657
13076
|
},
|
|
12658
13077
|
contextWindow: 1000000,
|
|
@@ -12667,9 +13086,9 @@ export const MODELS = {
|
|
|
12667
13086
|
reasoning: true,
|
|
12668
13087
|
input: ["text"],
|
|
12669
13088
|
cost: {
|
|
12670
|
-
input:
|
|
12671
|
-
output:
|
|
12672
|
-
cacheRead: 0.
|
|
13089
|
+
input: 0.435,
|
|
13090
|
+
output: 0.87,
|
|
13091
|
+
cacheRead: 0.0036,
|
|
12673
13092
|
cacheWrite: 0,
|
|
12674
13093
|
},
|
|
12675
13094
|
contextWindow: 1000000,
|
|
@@ -15182,6 +15601,24 @@ export const MODELS = {
|
|
|
15182
15601
|
contextWindow: 200000,
|
|
15183
15602
|
maxTokens: 131072,
|
|
15184
15603
|
},
|
|
15604
|
+
"glm-5v-turbo": {
|
|
15605
|
+
id: "glm-5v-turbo",
|
|
15606
|
+
name: "glm-5v-turbo",
|
|
15607
|
+
api: "openai-completions",
|
|
15608
|
+
provider: "zai",
|
|
15609
|
+
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
15610
|
+
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
15611
|
+
reasoning: true,
|
|
15612
|
+
input: ["text", "image"],
|
|
15613
|
+
cost: {
|
|
15614
|
+
input: 0,
|
|
15615
|
+
output: 0,
|
|
15616
|
+
cacheRead: 0,
|
|
15617
|
+
cacheWrite: 0,
|
|
15618
|
+
},
|
|
15619
|
+
contextWindow: 200000,
|
|
15620
|
+
maxTokens: 131072,
|
|
15621
|
+
},
|
|
15185
15622
|
},
|
|
15186
15623
|
};
|
|
15187
15624
|
//# sourceMappingURL=models.generated.js.map
|