@hyperspaceng/neural-ai 0.68.1 → 0.69.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +3 -1
- package/dist/env-api-keys.d.ts.map +1 -1
- package/dist/env-api-keys.js +1 -0
- package/dist/env-api-keys.js.map +1 -1
- package/dist/index.d.ts +2 -2
- package/dist/index.d.ts.map +1 -1
- package/dist/index.js +1 -1
- package/dist/index.js.map +1 -1
- package/dist/models.generated.d.ts +610 -30
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +624 -44
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/amazon-bedrock.d.ts.map +1 -1
- package/dist/providers/amazon-bedrock.js +49 -9
- package/dist/providers/amazon-bedrock.js.map +1 -1
- package/dist/providers/anthropic.d.ts.map +1 -1
- package/dist/providers/anthropic.js +140 -8
- package/dist/providers/anthropic.js.map +1 -1
- package/dist/providers/transform-messages.d.ts.map +1 -1
- package/dist/providers/transform-messages.js +2 -0
- package/dist/providers/transform-messages.js.map +1 -1
- package/dist/types.d.ts +2 -2
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/dist/utils/json-parse.d.ts +8 -1
- package/dist/utils/json-parse.d.ts.map +1 -1
- package/dist/utils/json-parse.js +89 -5
- package/dist/utils/json-parse.js.map +1 -1
- package/dist/utils/typebox-helpers.d.ts +1 -1
- package/dist/utils/typebox-helpers.d.ts.map +1 -1
- package/dist/utils/typebox-helpers.js +1 -1
- package/dist/utils/typebox-helpers.js.map +1 -1
- package/dist/utils/validation.d.ts.map +1 -1
- package/dist/utils/validation.js +242 -41
- package/dist/utils/validation.js.map +1 -1
- package/package.json +2 -4
package/dist/models.generated.js
CHANGED
|
@@ -325,6 +325,40 @@ export const MODELS = {
|
|
|
325
325
|
contextWindow: 1000000,
|
|
326
326
|
maxTokens: 64000,
|
|
327
327
|
},
|
|
328
|
+
"au.anthropic.claude-opus-4-6-v1": {
|
|
329
|
+
id: "au.anthropic.claude-opus-4-6-v1",
|
|
330
|
+
name: "AU Anthropic Claude Opus 4.6",
|
|
331
|
+
api: "bedrock-converse-stream",
|
|
332
|
+
provider: "amazon-bedrock",
|
|
333
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
334
|
+
reasoning: true,
|
|
335
|
+
input: ["text", "image"],
|
|
336
|
+
cost: {
|
|
337
|
+
input: 16.5,
|
|
338
|
+
output: 82.5,
|
|
339
|
+
cacheRead: 0.5,
|
|
340
|
+
cacheWrite: 6.25,
|
|
341
|
+
},
|
|
342
|
+
contextWindow: 1000000,
|
|
343
|
+
maxTokens: 128000,
|
|
344
|
+
},
|
|
345
|
+
"au.anthropic.claude-sonnet-4-6": {
|
|
346
|
+
id: "au.anthropic.claude-sonnet-4-6",
|
|
347
|
+
name: "AU Anthropic Claude Sonnet 4.6",
|
|
348
|
+
api: "bedrock-converse-stream",
|
|
349
|
+
provider: "amazon-bedrock",
|
|
350
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
351
|
+
reasoning: true,
|
|
352
|
+
input: ["text", "image"],
|
|
353
|
+
cost: {
|
|
354
|
+
input: 3.3,
|
|
355
|
+
output: 16.5,
|
|
356
|
+
cacheRead: 0.33,
|
|
357
|
+
cacheWrite: 4.125,
|
|
358
|
+
},
|
|
359
|
+
contextWindow: 1000000,
|
|
360
|
+
maxTokens: 128000,
|
|
361
|
+
},
|
|
328
362
|
"deepseek.r1-v1:0": {
|
|
329
363
|
id: "deepseek.r1-v1:0",
|
|
330
364
|
name: "DeepSeek-R1",
|
|
@@ -381,7 +415,7 @@ export const MODELS = {
|
|
|
381
415
|
name: "Claude Haiku 4.5 (EU)",
|
|
382
416
|
api: "bedrock-converse-stream",
|
|
383
417
|
provider: "amazon-bedrock",
|
|
384
|
-
baseUrl: "https://bedrock-runtime.
|
|
418
|
+
baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
|
|
385
419
|
reasoning: true,
|
|
386
420
|
input: ["text", "image"],
|
|
387
421
|
cost: {
|
|
@@ -398,7 +432,7 @@ export const MODELS = {
|
|
|
398
432
|
name: "Claude Opus 4.5 (EU)",
|
|
399
433
|
api: "bedrock-converse-stream",
|
|
400
434
|
provider: "amazon-bedrock",
|
|
401
|
-
baseUrl: "https://bedrock-runtime.
|
|
435
|
+
baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
|
|
402
436
|
reasoning: true,
|
|
403
437
|
input: ["text", "image"],
|
|
404
438
|
cost: {
|
|
@@ -415,7 +449,7 @@ export const MODELS = {
|
|
|
415
449
|
name: "Claude Opus 4.6 (EU)",
|
|
416
450
|
api: "bedrock-converse-stream",
|
|
417
451
|
provider: "amazon-bedrock",
|
|
418
|
-
baseUrl: "https://bedrock-runtime.
|
|
452
|
+
baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
|
|
419
453
|
reasoning: true,
|
|
420
454
|
input: ["text", "image"],
|
|
421
455
|
cost: {
|
|
@@ -432,7 +466,7 @@ export const MODELS = {
|
|
|
432
466
|
name: "Claude Opus 4.7 (EU)",
|
|
433
467
|
api: "bedrock-converse-stream",
|
|
434
468
|
provider: "amazon-bedrock",
|
|
435
|
-
baseUrl: "https://bedrock-runtime.
|
|
469
|
+
baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
|
|
436
470
|
reasoning: true,
|
|
437
471
|
input: ["text", "image"],
|
|
438
472
|
cost: {
|
|
@@ -449,7 +483,7 @@ export const MODELS = {
|
|
|
449
483
|
name: "Claude Sonnet 4 (EU)",
|
|
450
484
|
api: "bedrock-converse-stream",
|
|
451
485
|
provider: "amazon-bedrock",
|
|
452
|
-
baseUrl: "https://bedrock-runtime.
|
|
486
|
+
baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
|
|
453
487
|
reasoning: true,
|
|
454
488
|
input: ["text", "image"],
|
|
455
489
|
cost: {
|
|
@@ -466,7 +500,7 @@ export const MODELS = {
|
|
|
466
500
|
name: "Claude Sonnet 4.5 (EU)",
|
|
467
501
|
api: "bedrock-converse-stream",
|
|
468
502
|
provider: "amazon-bedrock",
|
|
469
|
-
baseUrl: "https://bedrock-runtime.
|
|
503
|
+
baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
|
|
470
504
|
reasoning: true,
|
|
471
505
|
input: ["text", "image"],
|
|
472
506
|
cost: {
|
|
@@ -483,7 +517,7 @@ export const MODELS = {
|
|
|
483
517
|
name: "Claude Sonnet 4.6 (EU)",
|
|
484
518
|
api: "bedrock-converse-stream",
|
|
485
519
|
provider: "amazon-bedrock",
|
|
486
|
-
baseUrl: "https://bedrock-runtime.
|
|
520
|
+
baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
|
|
487
521
|
reasoning: true,
|
|
488
522
|
input: ["text", "image"],
|
|
489
523
|
cost: {
|
|
@@ -2488,6 +2522,23 @@ export const MODELS = {
|
|
|
2488
2522
|
contextWindow: 1050000,
|
|
2489
2523
|
maxTokens: 128000,
|
|
2490
2524
|
},
|
|
2525
|
+
"gpt-5.5": {
|
|
2526
|
+
id: "gpt-5.5",
|
|
2527
|
+
name: "GPT-5.5",
|
|
2528
|
+
api: "azure-openai-responses",
|
|
2529
|
+
provider: "azure-openai-responses",
|
|
2530
|
+
baseUrl: "",
|
|
2531
|
+
reasoning: true,
|
|
2532
|
+
input: ["text", "image"],
|
|
2533
|
+
cost: {
|
|
2534
|
+
input: 5,
|
|
2535
|
+
output: 30,
|
|
2536
|
+
cacheRead: 0.5,
|
|
2537
|
+
cacheWrite: 0,
|
|
2538
|
+
},
|
|
2539
|
+
contextWindow: 1050000,
|
|
2540
|
+
maxTokens: 130000,
|
|
2541
|
+
},
|
|
2491
2542
|
"o1": {
|
|
2492
2543
|
id: "o1",
|
|
2493
2544
|
name: "o1",
|
|
@@ -2695,6 +2746,314 @@ export const MODELS = {
|
|
|
2695
2746
|
maxTokens: 40000,
|
|
2696
2747
|
},
|
|
2697
2748
|
},
|
|
2749
|
+
"fireworks": {
|
|
2750
|
+
"accounts/fireworks/models/deepseek-v3p1": {
|
|
2751
|
+
id: "accounts/fireworks/models/deepseek-v3p1",
|
|
2752
|
+
name: "DeepSeek V3.1",
|
|
2753
|
+
api: "anthropic-messages",
|
|
2754
|
+
provider: "fireworks",
|
|
2755
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2756
|
+
reasoning: true,
|
|
2757
|
+
input: ["text"],
|
|
2758
|
+
cost: {
|
|
2759
|
+
input: 0.56,
|
|
2760
|
+
output: 1.68,
|
|
2761
|
+
cacheRead: 0,
|
|
2762
|
+
cacheWrite: 0,
|
|
2763
|
+
},
|
|
2764
|
+
contextWindow: 163840,
|
|
2765
|
+
maxTokens: 163840,
|
|
2766
|
+
},
|
|
2767
|
+
"accounts/fireworks/models/deepseek-v3p2": {
|
|
2768
|
+
id: "accounts/fireworks/models/deepseek-v3p2",
|
|
2769
|
+
name: "DeepSeek V3.2",
|
|
2770
|
+
api: "anthropic-messages",
|
|
2771
|
+
provider: "fireworks",
|
|
2772
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2773
|
+
reasoning: true,
|
|
2774
|
+
input: ["text"],
|
|
2775
|
+
cost: {
|
|
2776
|
+
input: 0.56,
|
|
2777
|
+
output: 1.68,
|
|
2778
|
+
cacheRead: 0.28,
|
|
2779
|
+
cacheWrite: 0,
|
|
2780
|
+
},
|
|
2781
|
+
contextWindow: 160000,
|
|
2782
|
+
maxTokens: 160000,
|
|
2783
|
+
},
|
|
2784
|
+
"accounts/fireworks/models/glm-4p5": {
|
|
2785
|
+
id: "accounts/fireworks/models/glm-4p5",
|
|
2786
|
+
name: "GLM 4.5",
|
|
2787
|
+
api: "anthropic-messages",
|
|
2788
|
+
provider: "fireworks",
|
|
2789
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2790
|
+
reasoning: true,
|
|
2791
|
+
input: ["text"],
|
|
2792
|
+
cost: {
|
|
2793
|
+
input: 0.55,
|
|
2794
|
+
output: 2.19,
|
|
2795
|
+
cacheRead: 0,
|
|
2796
|
+
cacheWrite: 0,
|
|
2797
|
+
},
|
|
2798
|
+
contextWindow: 131072,
|
|
2799
|
+
maxTokens: 131072,
|
|
2800
|
+
},
|
|
2801
|
+
"accounts/fireworks/models/glm-4p5-air": {
|
|
2802
|
+
id: "accounts/fireworks/models/glm-4p5-air",
|
|
2803
|
+
name: "GLM 4.5 Air",
|
|
2804
|
+
api: "anthropic-messages",
|
|
2805
|
+
provider: "fireworks",
|
|
2806
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2807
|
+
reasoning: true,
|
|
2808
|
+
input: ["text"],
|
|
2809
|
+
cost: {
|
|
2810
|
+
input: 0.22,
|
|
2811
|
+
output: 0.88,
|
|
2812
|
+
cacheRead: 0,
|
|
2813
|
+
cacheWrite: 0,
|
|
2814
|
+
},
|
|
2815
|
+
contextWindow: 131072,
|
|
2816
|
+
maxTokens: 131072,
|
|
2817
|
+
},
|
|
2818
|
+
"accounts/fireworks/models/glm-4p7": {
|
|
2819
|
+
id: "accounts/fireworks/models/glm-4p7",
|
|
2820
|
+
name: "GLM 4.7",
|
|
2821
|
+
api: "anthropic-messages",
|
|
2822
|
+
provider: "fireworks",
|
|
2823
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2824
|
+
reasoning: true,
|
|
2825
|
+
input: ["text"],
|
|
2826
|
+
cost: {
|
|
2827
|
+
input: 0.6,
|
|
2828
|
+
output: 2.2,
|
|
2829
|
+
cacheRead: 0.3,
|
|
2830
|
+
cacheWrite: 0,
|
|
2831
|
+
},
|
|
2832
|
+
contextWindow: 198000,
|
|
2833
|
+
maxTokens: 198000,
|
|
2834
|
+
},
|
|
2835
|
+
"accounts/fireworks/models/glm-5": {
|
|
2836
|
+
id: "accounts/fireworks/models/glm-5",
|
|
2837
|
+
name: "GLM 5",
|
|
2838
|
+
api: "anthropic-messages",
|
|
2839
|
+
provider: "fireworks",
|
|
2840
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2841
|
+
reasoning: true,
|
|
2842
|
+
input: ["text"],
|
|
2843
|
+
cost: {
|
|
2844
|
+
input: 1,
|
|
2845
|
+
output: 3.2,
|
|
2846
|
+
cacheRead: 0.5,
|
|
2847
|
+
cacheWrite: 0,
|
|
2848
|
+
},
|
|
2849
|
+
contextWindow: 202752,
|
|
2850
|
+
maxTokens: 131072,
|
|
2851
|
+
},
|
|
2852
|
+
"accounts/fireworks/models/glm-5p1": {
|
|
2853
|
+
id: "accounts/fireworks/models/glm-5p1",
|
|
2854
|
+
name: "GLM 5.1",
|
|
2855
|
+
api: "anthropic-messages",
|
|
2856
|
+
provider: "fireworks",
|
|
2857
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2858
|
+
reasoning: true,
|
|
2859
|
+
input: ["text"],
|
|
2860
|
+
cost: {
|
|
2861
|
+
input: 1.4,
|
|
2862
|
+
output: 4.4,
|
|
2863
|
+
cacheRead: 0.26,
|
|
2864
|
+
cacheWrite: 0,
|
|
2865
|
+
},
|
|
2866
|
+
contextWindow: 202800,
|
|
2867
|
+
maxTokens: 131072,
|
|
2868
|
+
},
|
|
2869
|
+
"accounts/fireworks/models/gpt-oss-120b": {
|
|
2870
|
+
id: "accounts/fireworks/models/gpt-oss-120b",
|
|
2871
|
+
name: "GPT OSS 120B",
|
|
2872
|
+
api: "anthropic-messages",
|
|
2873
|
+
provider: "fireworks",
|
|
2874
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2875
|
+
reasoning: true,
|
|
2876
|
+
input: ["text"],
|
|
2877
|
+
cost: {
|
|
2878
|
+
input: 0.15,
|
|
2879
|
+
output: 0.6,
|
|
2880
|
+
cacheRead: 0,
|
|
2881
|
+
cacheWrite: 0,
|
|
2882
|
+
},
|
|
2883
|
+
contextWindow: 131072,
|
|
2884
|
+
maxTokens: 32768,
|
|
2885
|
+
},
|
|
2886
|
+
"accounts/fireworks/models/gpt-oss-20b": {
|
|
2887
|
+
id: "accounts/fireworks/models/gpt-oss-20b",
|
|
2888
|
+
name: "GPT OSS 20B",
|
|
2889
|
+
api: "anthropic-messages",
|
|
2890
|
+
provider: "fireworks",
|
|
2891
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2892
|
+
reasoning: true,
|
|
2893
|
+
input: ["text"],
|
|
2894
|
+
cost: {
|
|
2895
|
+
input: 0.05,
|
|
2896
|
+
output: 0.2,
|
|
2897
|
+
cacheRead: 0,
|
|
2898
|
+
cacheWrite: 0,
|
|
2899
|
+
},
|
|
2900
|
+
contextWindow: 131072,
|
|
2901
|
+
maxTokens: 32768,
|
|
2902
|
+
},
|
|
2903
|
+
"accounts/fireworks/models/kimi-k2-instruct": {
|
|
2904
|
+
id: "accounts/fireworks/models/kimi-k2-instruct",
|
|
2905
|
+
name: "Kimi K2 Instruct",
|
|
2906
|
+
api: "anthropic-messages",
|
|
2907
|
+
provider: "fireworks",
|
|
2908
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2909
|
+
reasoning: false,
|
|
2910
|
+
input: ["text"],
|
|
2911
|
+
cost: {
|
|
2912
|
+
input: 1,
|
|
2913
|
+
output: 3,
|
|
2914
|
+
cacheRead: 0,
|
|
2915
|
+
cacheWrite: 0,
|
|
2916
|
+
},
|
|
2917
|
+
contextWindow: 128000,
|
|
2918
|
+
maxTokens: 16384,
|
|
2919
|
+
},
|
|
2920
|
+
"accounts/fireworks/models/kimi-k2-thinking": {
|
|
2921
|
+
id: "accounts/fireworks/models/kimi-k2-thinking",
|
|
2922
|
+
name: "Kimi K2 Thinking",
|
|
2923
|
+
api: "anthropic-messages",
|
|
2924
|
+
provider: "fireworks",
|
|
2925
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2926
|
+
reasoning: true,
|
|
2927
|
+
input: ["text"],
|
|
2928
|
+
cost: {
|
|
2929
|
+
input: 0.6,
|
|
2930
|
+
output: 2.5,
|
|
2931
|
+
cacheRead: 0.3,
|
|
2932
|
+
cacheWrite: 0,
|
|
2933
|
+
},
|
|
2934
|
+
contextWindow: 256000,
|
|
2935
|
+
maxTokens: 256000,
|
|
2936
|
+
},
|
|
2937
|
+
"accounts/fireworks/models/kimi-k2p5": {
|
|
2938
|
+
id: "accounts/fireworks/models/kimi-k2p5",
|
|
2939
|
+
name: "Kimi K2.5",
|
|
2940
|
+
api: "anthropic-messages",
|
|
2941
|
+
provider: "fireworks",
|
|
2942
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2943
|
+
reasoning: true,
|
|
2944
|
+
input: ["text", "image"],
|
|
2945
|
+
cost: {
|
|
2946
|
+
input: 0.6,
|
|
2947
|
+
output: 3,
|
|
2948
|
+
cacheRead: 0.1,
|
|
2949
|
+
cacheWrite: 0,
|
|
2950
|
+
},
|
|
2951
|
+
contextWindow: 256000,
|
|
2952
|
+
maxTokens: 256000,
|
|
2953
|
+
},
|
|
2954
|
+
"accounts/fireworks/models/kimi-k2p6": {
|
|
2955
|
+
id: "accounts/fireworks/models/kimi-k2p6",
|
|
2956
|
+
name: "Kimi K2.6",
|
|
2957
|
+
api: "anthropic-messages",
|
|
2958
|
+
provider: "fireworks",
|
|
2959
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2960
|
+
reasoning: true,
|
|
2961
|
+
input: ["text", "image"],
|
|
2962
|
+
cost: {
|
|
2963
|
+
input: 0.95,
|
|
2964
|
+
output: 4,
|
|
2965
|
+
cacheRead: 0.16,
|
|
2966
|
+
cacheWrite: 0,
|
|
2967
|
+
},
|
|
2968
|
+
contextWindow: 262000,
|
|
2969
|
+
maxTokens: 262000,
|
|
2970
|
+
},
|
|
2971
|
+
"accounts/fireworks/models/minimax-m2p1": {
|
|
2972
|
+
id: "accounts/fireworks/models/minimax-m2p1",
|
|
2973
|
+
name: "MiniMax-M2.1",
|
|
2974
|
+
api: "anthropic-messages",
|
|
2975
|
+
provider: "fireworks",
|
|
2976
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2977
|
+
reasoning: true,
|
|
2978
|
+
input: ["text"],
|
|
2979
|
+
cost: {
|
|
2980
|
+
input: 0.3,
|
|
2981
|
+
output: 1.2,
|
|
2982
|
+
cacheRead: 0.03,
|
|
2983
|
+
cacheWrite: 0,
|
|
2984
|
+
},
|
|
2985
|
+
contextWindow: 200000,
|
|
2986
|
+
maxTokens: 200000,
|
|
2987
|
+
},
|
|
2988
|
+
"accounts/fireworks/models/minimax-m2p5": {
|
|
2989
|
+
id: "accounts/fireworks/models/minimax-m2p5",
|
|
2990
|
+
name: "MiniMax-M2.5",
|
|
2991
|
+
api: "anthropic-messages",
|
|
2992
|
+
provider: "fireworks",
|
|
2993
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2994
|
+
reasoning: true,
|
|
2995
|
+
input: ["text"],
|
|
2996
|
+
cost: {
|
|
2997
|
+
input: 0.3,
|
|
2998
|
+
output: 1.2,
|
|
2999
|
+
cacheRead: 0.03,
|
|
3000
|
+
cacheWrite: 0,
|
|
3001
|
+
},
|
|
3002
|
+
contextWindow: 196608,
|
|
3003
|
+
maxTokens: 196608,
|
|
3004
|
+
},
|
|
3005
|
+
"accounts/fireworks/models/minimax-m2p7": {
|
|
3006
|
+
id: "accounts/fireworks/models/minimax-m2p7",
|
|
3007
|
+
name: "MiniMax-M2.7",
|
|
3008
|
+
api: "anthropic-messages",
|
|
3009
|
+
provider: "fireworks",
|
|
3010
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
3011
|
+
reasoning: true,
|
|
3012
|
+
input: ["text"],
|
|
3013
|
+
cost: {
|
|
3014
|
+
input: 0.3,
|
|
3015
|
+
output: 1.2,
|
|
3016
|
+
cacheRead: 0.03,
|
|
3017
|
+
cacheWrite: 0,
|
|
3018
|
+
},
|
|
3019
|
+
contextWindow: 196608,
|
|
3020
|
+
maxTokens: 196608,
|
|
3021
|
+
},
|
|
3022
|
+
"accounts/fireworks/models/qwen3p6-plus": {
|
|
3023
|
+
id: "accounts/fireworks/models/qwen3p6-plus",
|
|
3024
|
+
name: "Qwen 3.6 Plus",
|
|
3025
|
+
api: "anthropic-messages",
|
|
3026
|
+
provider: "fireworks",
|
|
3027
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
3028
|
+
reasoning: true,
|
|
3029
|
+
input: ["text", "image"],
|
|
3030
|
+
cost: {
|
|
3031
|
+
input: 0.5,
|
|
3032
|
+
output: 3,
|
|
3033
|
+
cacheRead: 0.1,
|
|
3034
|
+
cacheWrite: 0,
|
|
3035
|
+
},
|
|
3036
|
+
contextWindow: 128000,
|
|
3037
|
+
maxTokens: 8192,
|
|
3038
|
+
},
|
|
3039
|
+
"accounts/fireworks/routers/kimi-k2p5-turbo": {
|
|
3040
|
+
id: "accounts/fireworks/routers/kimi-k2p5-turbo",
|
|
3041
|
+
name: "Kimi K2.5 Turbo (firepass)",
|
|
3042
|
+
api: "anthropic-messages",
|
|
3043
|
+
provider: "fireworks",
|
|
3044
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
3045
|
+
reasoning: true,
|
|
3046
|
+
input: ["text", "image"],
|
|
3047
|
+
cost: {
|
|
3048
|
+
input: 0,
|
|
3049
|
+
output: 0,
|
|
3050
|
+
cacheRead: 0,
|
|
3051
|
+
cacheWrite: 0,
|
|
3052
|
+
},
|
|
3053
|
+
contextWindow: 256000,
|
|
3054
|
+
maxTokens: 256000,
|
|
3055
|
+
},
|
|
3056
|
+
},
|
|
2698
3057
|
"github-copilot": {
|
|
2699
3058
|
"claude-haiku-4.5": {
|
|
2700
3059
|
id: "claude-haiku-4.5",
|
|
@@ -3580,8 +3939,8 @@ export const MODELS = {
|
|
|
3580
3939
|
contextWindow: 131072,
|
|
3581
3940
|
maxTokens: 8192,
|
|
3582
3941
|
},
|
|
3583
|
-
"gemma-4-26b-it": {
|
|
3584
|
-
id: "gemma-4-26b-it",
|
|
3942
|
+
"gemma-4-26b-a4b-it": {
|
|
3943
|
+
id: "gemma-4-26b-a4b-it",
|
|
3585
3944
|
name: "Gemma 4 26B",
|
|
3586
3945
|
api: "google-generative-ai",
|
|
3587
3946
|
provider: "google",
|
|
@@ -3856,6 +4215,23 @@ export const MODELS = {
|
|
|
3856
4215
|
contextWindow: 1048576,
|
|
3857
4216
|
maxTokens: 65535,
|
|
3858
4217
|
},
|
|
4218
|
+
"gemini-3.1-flash-lite-preview": {
|
|
4219
|
+
id: "gemini-3.1-flash-lite-preview",
|
|
4220
|
+
name: "Gemini 3.1 Flash Lite Preview (Cloud Code Assist)",
|
|
4221
|
+
api: "google-gemini-cli",
|
|
4222
|
+
provider: "google-gemini-cli",
|
|
4223
|
+
baseUrl: "https://cloudcode-pa.googleapis.com",
|
|
4224
|
+
reasoning: true,
|
|
4225
|
+
input: ["text", "image"],
|
|
4226
|
+
cost: {
|
|
4227
|
+
input: 0,
|
|
4228
|
+
output: 0,
|
|
4229
|
+
cacheRead: 0,
|
|
4230
|
+
cacheWrite: 0,
|
|
4231
|
+
},
|
|
4232
|
+
contextWindow: 1048576,
|
|
4233
|
+
maxTokens: 65535,
|
|
4234
|
+
},
|
|
3859
4235
|
"gemini-3.1-pro-preview": {
|
|
3860
4236
|
id: "gemini-3.1-pro-preview",
|
|
3861
4237
|
name: "Gemini 3.1 Pro Preview (Cloud Code Assist)",
|
|
@@ -5899,6 +6275,23 @@ export const MODELS = {
|
|
|
5899
6275
|
contextWindow: 1050000,
|
|
5900
6276
|
maxTokens: 128000,
|
|
5901
6277
|
},
|
|
6278
|
+
"gpt-5.5": {
|
|
6279
|
+
id: "gpt-5.5",
|
|
6280
|
+
name: "GPT-5.5",
|
|
6281
|
+
api: "openai-responses",
|
|
6282
|
+
provider: "openai",
|
|
6283
|
+
baseUrl: "https://api.openai.com/v1",
|
|
6284
|
+
reasoning: true,
|
|
6285
|
+
input: ["text", "image"],
|
|
6286
|
+
cost: {
|
|
6287
|
+
input: 5,
|
|
6288
|
+
output: 30,
|
|
6289
|
+
cacheRead: 0.5,
|
|
6290
|
+
cacheWrite: 0,
|
|
6291
|
+
},
|
|
6292
|
+
contextWindow: 1050000,
|
|
6293
|
+
maxTokens: 130000,
|
|
6294
|
+
},
|
|
5902
6295
|
"o1": {
|
|
5903
6296
|
id: "o1",
|
|
5904
6297
|
name: "o1",
|
|
@@ -6668,6 +7061,23 @@ export const MODELS = {
|
|
|
6668
7061
|
contextWindow: 1050000,
|
|
6669
7062
|
maxTokens: 128000,
|
|
6670
7063
|
},
|
|
7064
|
+
"hy3-preview-free": {
|
|
7065
|
+
id: "hy3-preview-free",
|
|
7066
|
+
name: "Hy3 preview Free",
|
|
7067
|
+
api: "openai-completions",
|
|
7068
|
+
provider: "opencode",
|
|
7069
|
+
baseUrl: "https://opencode.ai/zen/v1",
|
|
7070
|
+
reasoning: true,
|
|
7071
|
+
input: ["text"],
|
|
7072
|
+
cost: {
|
|
7073
|
+
input: 0,
|
|
7074
|
+
output: 0,
|
|
7075
|
+
cacheRead: 0,
|
|
7076
|
+
cacheWrite: 0,
|
|
7077
|
+
},
|
|
7078
|
+
contextWindow: 256000,
|
|
7079
|
+
maxTokens: 64000,
|
|
7080
|
+
},
|
|
6671
7081
|
"kimi-k2.5": {
|
|
6672
7082
|
id: "kimi-k2.5",
|
|
6673
7083
|
name: "Kimi K2.5",
|
|
@@ -6906,7 +7316,7 @@ export const MODELS = {
|
|
|
6906
7316
|
cacheWrite: 0,
|
|
6907
7317
|
},
|
|
6908
7318
|
contextWindow: 262144,
|
|
6909
|
-
maxTokens:
|
|
7319
|
+
maxTokens: 128000,
|
|
6910
7320
|
},
|
|
6911
7321
|
"mimo-v2-pro": {
|
|
6912
7322
|
id: "mimo-v2-pro",
|
|
@@ -6923,7 +7333,41 @@ export const MODELS = {
|
|
|
6923
7333
|
cacheWrite: 0,
|
|
6924
7334
|
},
|
|
6925
7335
|
contextWindow: 1048576,
|
|
6926
|
-
maxTokens:
|
|
7336
|
+
maxTokens: 128000,
|
|
7337
|
+
},
|
|
7338
|
+
"mimo-v2.5": {
|
|
7339
|
+
id: "mimo-v2.5",
|
|
7340
|
+
name: "MiMo V2.5",
|
|
7341
|
+
api: "openai-completions",
|
|
7342
|
+
provider: "opencode-go",
|
|
7343
|
+
baseUrl: "https://opencode.ai/zen/go/v1",
|
|
7344
|
+
reasoning: true,
|
|
7345
|
+
input: ["text", "image"],
|
|
7346
|
+
cost: {
|
|
7347
|
+
input: 0.4,
|
|
7348
|
+
output: 2,
|
|
7349
|
+
cacheRead: 0.08,
|
|
7350
|
+
cacheWrite: 0,
|
|
7351
|
+
},
|
|
7352
|
+
contextWindow: 262144,
|
|
7353
|
+
maxTokens: 128000,
|
|
7354
|
+
},
|
|
7355
|
+
"mimo-v2.5-pro": {
|
|
7356
|
+
id: "mimo-v2.5-pro",
|
|
7357
|
+
name: "MiMo V2.5 Pro",
|
|
7358
|
+
api: "openai-completions",
|
|
7359
|
+
provider: "opencode-go",
|
|
7360
|
+
baseUrl: "https://opencode.ai/zen/go/v1",
|
|
7361
|
+
reasoning: true,
|
|
7362
|
+
input: ["text"],
|
|
7363
|
+
cost: {
|
|
7364
|
+
input: 1,
|
|
7365
|
+
output: 3,
|
|
7366
|
+
cacheRead: 0.2,
|
|
7367
|
+
cacheWrite: 0,
|
|
7368
|
+
},
|
|
7369
|
+
contextWindow: 1048576,
|
|
7370
|
+
maxTokens: 128000,
|
|
6927
7371
|
},
|
|
6928
7372
|
"minimax-m2.5": {
|
|
6929
7373
|
id: "minimax-m2.5",
|
|
@@ -7369,17 +7813,17 @@ export const MODELS = {
|
|
|
7369
7813
|
contextWindow: 1000000,
|
|
7370
7814
|
maxTokens: 128000,
|
|
7371
7815
|
},
|
|
7372
|
-
"arcee-ai/trinity-large-preview
|
|
7373
|
-
id: "arcee-ai/trinity-large-preview
|
|
7374
|
-
name: "Arcee AI: Trinity Large Preview
|
|
7816
|
+
"arcee-ai/trinity-large-preview": {
|
|
7817
|
+
id: "arcee-ai/trinity-large-preview",
|
|
7818
|
+
name: "Arcee AI: Trinity Large Preview",
|
|
7375
7819
|
api: "openai-completions",
|
|
7376
7820
|
provider: "openrouter",
|
|
7377
7821
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
7378
7822
|
reasoning: false,
|
|
7379
7823
|
input: ["text"],
|
|
7380
7824
|
cost: {
|
|
7381
|
-
input: 0,
|
|
7382
|
-
output: 0,
|
|
7825
|
+
input: 0.15,
|
|
7826
|
+
output: 0.44999999999999996,
|
|
7383
7827
|
cacheRead: 0,
|
|
7384
7828
|
cacheWrite: 0,
|
|
7385
7829
|
},
|
|
@@ -7707,7 +8151,7 @@ export const MODELS = {
|
|
|
7707
8151
|
cacheWrite: 0,
|
|
7708
8152
|
},
|
|
7709
8153
|
contextWindow: 131072,
|
|
7710
|
-
maxTokens:
|
|
8154
|
+
maxTokens: 65536,
|
|
7711
8155
|
},
|
|
7712
8156
|
"deepseek/deepseek-v3.2-exp": {
|
|
7713
8157
|
id: "deepseek/deepseek-v3.2-exp",
|
|
@@ -7726,6 +8170,40 @@ export const MODELS = {
|
|
|
7726
8170
|
contextWindow: 163840,
|
|
7727
8171
|
maxTokens: 65536,
|
|
7728
8172
|
},
|
|
8173
|
+
"deepseek/deepseek-v4-flash": {
|
|
8174
|
+
id: "deepseek/deepseek-v4-flash",
|
|
8175
|
+
name: "DeepSeek: DeepSeek V4 Flash",
|
|
8176
|
+
api: "openai-completions",
|
|
8177
|
+
provider: "openrouter",
|
|
8178
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
8179
|
+
reasoning: true,
|
|
8180
|
+
input: ["text"],
|
|
8181
|
+
cost: {
|
|
8182
|
+
input: 0.14,
|
|
8183
|
+
output: 0.28,
|
|
8184
|
+
cacheRead: 0.028,
|
|
8185
|
+
cacheWrite: 0,
|
|
8186
|
+
},
|
|
8187
|
+
contextWindow: 1048576,
|
|
8188
|
+
maxTokens: 384000,
|
|
8189
|
+
},
|
|
8190
|
+
"deepseek/deepseek-v4-pro": {
|
|
8191
|
+
id: "deepseek/deepseek-v4-pro",
|
|
8192
|
+
name: "DeepSeek: DeepSeek V4 Pro",
|
|
8193
|
+
api: "openai-completions",
|
|
8194
|
+
provider: "openrouter",
|
|
8195
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
8196
|
+
reasoning: true,
|
|
8197
|
+
input: ["text"],
|
|
8198
|
+
cost: {
|
|
8199
|
+
input: 1.74,
|
|
8200
|
+
output: 3.48,
|
|
8201
|
+
cacheRead: 0.145,
|
|
8202
|
+
cacheWrite: 0,
|
|
8203
|
+
},
|
|
8204
|
+
contextWindow: 1048576,
|
|
8205
|
+
maxTokens: 384000,
|
|
8206
|
+
},
|
|
7729
8207
|
"essentialai/rnj-1-instruct": {
|
|
7730
8208
|
id: "essentialai/rnj-1-instruct",
|
|
7731
8209
|
name: "EssentialAI: Rnj 1 Instruct",
|
|
@@ -7956,13 +8434,13 @@ export const MODELS = {
|
|
|
7956
8434
|
reasoning: true,
|
|
7957
8435
|
input: ["text", "image"],
|
|
7958
8436
|
cost: {
|
|
7959
|
-
input: 0.
|
|
7960
|
-
output: 0.
|
|
7961
|
-
cacheRead: 0
|
|
8437
|
+
input: 0.06,
|
|
8438
|
+
output: 0.33,
|
|
8439
|
+
cacheRead: 0,
|
|
7962
8440
|
cacheWrite: 0,
|
|
7963
8441
|
},
|
|
7964
8442
|
contextWindow: 262144,
|
|
7965
|
-
maxTokens:
|
|
8443
|
+
maxTokens: 4096,
|
|
7966
8444
|
},
|
|
7967
8445
|
"google/gemma-4-26b-a4b-it:free": {
|
|
7968
8446
|
id: "google/gemma-4-26b-a4b-it:free",
|
|
@@ -7992,7 +8470,7 @@ export const MODELS = {
|
|
|
7992
8470
|
cost: {
|
|
7993
8471
|
input: 0.13,
|
|
7994
8472
|
output: 0.38,
|
|
7995
|
-
cacheRead: 0
|
|
8473
|
+
cacheRead: 0,
|
|
7996
8474
|
cacheWrite: 0,
|
|
7997
8475
|
},
|
|
7998
8476
|
contextWindow: 262144,
|
|
@@ -8032,6 +8510,23 @@ export const MODELS = {
|
|
|
8032
8510
|
contextWindow: 128000,
|
|
8033
8511
|
maxTokens: 50000,
|
|
8034
8512
|
},
|
|
8513
|
+
"inclusionai/ling-2.6-1t:free": {
|
|
8514
|
+
id: "inclusionai/ling-2.6-1t:free",
|
|
8515
|
+
name: "inclusionAI: Ling-2.6-1T (free)",
|
|
8516
|
+
api: "openai-completions",
|
|
8517
|
+
provider: "openrouter",
|
|
8518
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
8519
|
+
reasoning: false,
|
|
8520
|
+
input: ["text"],
|
|
8521
|
+
cost: {
|
|
8522
|
+
input: 0,
|
|
8523
|
+
output: 0,
|
|
8524
|
+
cacheRead: 0,
|
|
8525
|
+
cacheWrite: 0,
|
|
8526
|
+
},
|
|
8527
|
+
contextWindow: 262144,
|
|
8528
|
+
maxTokens: 32768,
|
|
8529
|
+
},
|
|
8035
8530
|
"inclusionai/ling-2.6-flash:free": {
|
|
8036
8531
|
id: "inclusionai/ling-2.6-flash:free",
|
|
8037
8532
|
name: "inclusionAI: Ling-2.6-flash (free)",
|
|
@@ -8229,12 +8724,12 @@ export const MODELS = {
|
|
|
8229
8724
|
input: ["text"],
|
|
8230
8725
|
cost: {
|
|
8231
8726
|
input: 0.15,
|
|
8232
|
-
output: 1.
|
|
8233
|
-
cacheRead: 0.
|
|
8727
|
+
output: 1.15,
|
|
8728
|
+
cacheRead: 0.03,
|
|
8234
8729
|
cacheWrite: 0,
|
|
8235
8730
|
},
|
|
8236
8731
|
contextWindow: 196608,
|
|
8237
|
-
maxTokens:
|
|
8732
|
+
maxTokens: 4096,
|
|
8238
8733
|
},
|
|
8239
8734
|
"minimax/minimax-m2.5:free": {
|
|
8240
8735
|
id: "minimax/minimax-m2.5:free",
|
|
@@ -8500,13 +8995,13 @@ export const MODELS = {
|
|
|
8500
8995
|
reasoning: false,
|
|
8501
8996
|
input: ["text"],
|
|
8502
8997
|
cost: {
|
|
8503
|
-
input: 0.
|
|
8504
|
-
output: 0.
|
|
8998
|
+
input: 0.01,
|
|
8999
|
+
output: 0.03,
|
|
8505
9000
|
cacheRead: 0,
|
|
8506
9001
|
cacheWrite: 0,
|
|
8507
9002
|
},
|
|
8508
9003
|
contextWindow: 131072,
|
|
8509
|
-
maxTokens:
|
|
9004
|
+
maxTokens: 4096,
|
|
8510
9005
|
},
|
|
8511
9006
|
"mistralai/mistral-saba": {
|
|
8512
9007
|
id: "mistralai/mistral-saba",
|
|
@@ -8721,13 +9216,13 @@ export const MODELS = {
|
|
|
8721
9216
|
reasoning: true,
|
|
8722
9217
|
input: ["text", "image"],
|
|
8723
9218
|
cost: {
|
|
8724
|
-
input: 0.
|
|
8725
|
-
output:
|
|
8726
|
-
cacheRead: 0.
|
|
9219
|
+
input: 0.7448,
|
|
9220
|
+
output: 4.655,
|
|
9221
|
+
cacheRead: 0.1463,
|
|
8727
9222
|
cacheWrite: 0,
|
|
8728
9223
|
},
|
|
8729
|
-
contextWindow:
|
|
8730
|
-
maxTokens:
|
|
9224
|
+
contextWindow: 256000,
|
|
9225
|
+
maxTokens: 65536,
|
|
8731
9226
|
},
|
|
8732
9227
|
"nex-agi/deepseek-v3.1-nex-n1": {
|
|
8733
9228
|
id: "nex-agi/deepseek-v3.1-nex-n1",
|
|
@@ -10064,13 +10559,13 @@ export const MODELS = {
|
|
|
10064
10559
|
reasoning: true,
|
|
10065
10560
|
input: ["text"],
|
|
10066
10561
|
cost: {
|
|
10067
|
-
input: 0.
|
|
10068
|
-
output:
|
|
10562
|
+
input: 0.14950000000000002,
|
|
10563
|
+
output: 1.495,
|
|
10069
10564
|
cacheRead: 0,
|
|
10070
10565
|
cacheWrite: 0,
|
|
10071
10566
|
},
|
|
10072
|
-
contextWindow:
|
|
10073
|
-
maxTokens:
|
|
10567
|
+
contextWindow: 131072,
|
|
10568
|
+
maxTokens: 4096,
|
|
10074
10569
|
},
|
|
10075
10570
|
"qwen/qwen3-30b-a3b": {
|
|
10076
10571
|
id: "qwen/qwen3-30b-a3b",
|
|
@@ -10701,6 +11196,23 @@ export const MODELS = {
|
|
|
10701
11196
|
contextWindow: 262144,
|
|
10702
11197
|
maxTokens: 65536,
|
|
10703
11198
|
},
|
|
11199
|
+
"tencent/hy3-preview:free": {
|
|
11200
|
+
id: "tencent/hy3-preview:free",
|
|
11201
|
+
name: "Tencent: Hy3 preview (free)",
|
|
11202
|
+
api: "openai-completions",
|
|
11203
|
+
provider: "openrouter",
|
|
11204
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
11205
|
+
reasoning: true,
|
|
11206
|
+
input: ["text"],
|
|
11207
|
+
cost: {
|
|
11208
|
+
input: 0,
|
|
11209
|
+
output: 0,
|
|
11210
|
+
cacheRead: 0,
|
|
11211
|
+
cacheWrite: 0,
|
|
11212
|
+
},
|
|
11213
|
+
contextWindow: 262144,
|
|
11214
|
+
maxTokens: 262144,
|
|
11215
|
+
},
|
|
10704
11216
|
"thedrummer/rocinante-12b": {
|
|
10705
11217
|
id: "thedrummer/rocinante-12b",
|
|
10706
11218
|
name: "TheDrummer: Rocinante 12B",
|
|
@@ -10973,6 +11485,40 @@ export const MODELS = {
|
|
|
10973
11485
|
contextWindow: 1048576,
|
|
10974
11486
|
maxTokens: 131072,
|
|
10975
11487
|
},
|
|
11488
|
+
"xiaomi/mimo-v2.5": {
|
|
11489
|
+
id: "xiaomi/mimo-v2.5",
|
|
11490
|
+
name: "Xiaomi: MiMo-V2.5",
|
|
11491
|
+
api: "openai-completions",
|
|
11492
|
+
provider: "openrouter",
|
|
11493
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
11494
|
+
reasoning: true,
|
|
11495
|
+
input: ["text", "image"],
|
|
11496
|
+
cost: {
|
|
11497
|
+
input: 0.39999999999999997,
|
|
11498
|
+
output: 2,
|
|
11499
|
+
cacheRead: 0.08,
|
|
11500
|
+
cacheWrite: 0,
|
|
11501
|
+
},
|
|
11502
|
+
contextWindow: 1048576,
|
|
11503
|
+
maxTokens: 131072,
|
|
11504
|
+
},
|
|
11505
|
+
"xiaomi/mimo-v2.5-pro": {
|
|
11506
|
+
id: "xiaomi/mimo-v2.5-pro",
|
|
11507
|
+
name: "Xiaomi: MiMo-V2.5-Pro",
|
|
11508
|
+
api: "openai-completions",
|
|
11509
|
+
provider: "openrouter",
|
|
11510
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
11511
|
+
reasoning: true,
|
|
11512
|
+
input: ["text"],
|
|
11513
|
+
cost: {
|
|
11514
|
+
input: 1,
|
|
11515
|
+
output: 3,
|
|
11516
|
+
cacheRead: 0.19999999999999998,
|
|
11517
|
+
cacheWrite: 0,
|
|
11518
|
+
},
|
|
11519
|
+
contextWindow: 1048576,
|
|
11520
|
+
maxTokens: 131072,
|
|
11521
|
+
},
|
|
10976
11522
|
"z-ai/glm-4-32b": {
|
|
10977
11523
|
id: "z-ai/glm-4-32b",
|
|
10978
11524
|
name: "Z.ai: GLM 4 32B ",
|
|
@@ -11497,7 +12043,7 @@ export const MODELS = {
|
|
|
11497
12043
|
input: 0.5,
|
|
11498
12044
|
output: 3,
|
|
11499
12045
|
cacheRead: 0.09999999999999999,
|
|
11500
|
-
cacheWrite: 0,
|
|
12046
|
+
cacheWrite: 0.625,
|
|
11501
12047
|
},
|
|
11502
12048
|
contextWindow: 1000000,
|
|
11503
12049
|
maxTokens: 64000,
|
|
@@ -11876,6 +12422,40 @@ export const MODELS = {
|
|
|
11876
12422
|
contextWindow: 128000,
|
|
11877
12423
|
maxTokens: 64000,
|
|
11878
12424
|
},
|
|
12425
|
+
"deepseek/deepseek-v4-flash": {
|
|
12426
|
+
id: "deepseek/deepseek-v4-flash",
|
|
12427
|
+
name: "DeepSeek V4 Flash",
|
|
12428
|
+
api: "anthropic-messages",
|
|
12429
|
+
provider: "vercel-ai-gateway",
|
|
12430
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12431
|
+
reasoning: true,
|
|
12432
|
+
input: ["text"],
|
|
12433
|
+
cost: {
|
|
12434
|
+
input: 0.14,
|
|
12435
|
+
output: 0.28,
|
|
12436
|
+
cacheRead: 0.014,
|
|
12437
|
+
cacheWrite: 0,
|
|
12438
|
+
},
|
|
12439
|
+
contextWindow: 1000000,
|
|
12440
|
+
maxTokens: 384000,
|
|
12441
|
+
},
|
|
12442
|
+
"deepseek/deepseek-v4-pro": {
|
|
12443
|
+
id: "deepseek/deepseek-v4-pro",
|
|
12444
|
+
name: "DeepSeek V4 Pro",
|
|
12445
|
+
api: "anthropic-messages",
|
|
12446
|
+
provider: "vercel-ai-gateway",
|
|
12447
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12448
|
+
reasoning: true,
|
|
12449
|
+
input: ["text"],
|
|
12450
|
+
cost: {
|
|
12451
|
+
input: 1.74,
|
|
12452
|
+
output: 3.48,
|
|
12453
|
+
cacheRead: 0.145,
|
|
12454
|
+
cacheWrite: 0,
|
|
12455
|
+
},
|
|
12456
|
+
contextWindow: 1000000,
|
|
12457
|
+
maxTokens: 384000,
|
|
12458
|
+
},
|
|
11879
12459
|
"google/gemini-2.0-flash": {
|
|
11880
12460
|
id: "google/gemini-2.0-flash",
|
|
11881
12461
|
name: "Gemini 2.0 Flash",
|
|
@@ -13619,9 +14199,9 @@ export const MODELS = {
|
|
|
13619
14199
|
reasoning: true,
|
|
13620
14200
|
input: ["text"],
|
|
13621
14201
|
cost: {
|
|
13622
|
-
input: 0.
|
|
13623
|
-
output: 0.
|
|
13624
|
-
cacheRead: 0.
|
|
14202
|
+
input: 0.09999999999999999,
|
|
14203
|
+
output: 0.3,
|
|
14204
|
+
cacheRead: 0.01,
|
|
13625
14205
|
cacheWrite: 0,
|
|
13626
14206
|
},
|
|
13627
14207
|
contextWindow: 262144,
|
|
@@ -13838,15 +14418,15 @@ export const MODELS = {
|
|
|
13838
14418
|
provider: "vercel-ai-gateway",
|
|
13839
14419
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
13840
14420
|
reasoning: true,
|
|
13841
|
-
input: ["text"
|
|
14421
|
+
input: ["text"],
|
|
13842
14422
|
cost: {
|
|
13843
14423
|
input: 1.4,
|
|
13844
14424
|
output: 4.4,
|
|
13845
14425
|
cacheRead: 0.26,
|
|
13846
14426
|
cacheWrite: 0,
|
|
13847
14427
|
},
|
|
13848
|
-
contextWindow:
|
|
13849
|
-
maxTokens:
|
|
14428
|
+
contextWindow: 202800,
|
|
14429
|
+
maxTokens: 64000,
|
|
13850
14430
|
},
|
|
13851
14431
|
"zai/glm-5v-turbo": {
|
|
13852
14432
|
id: "zai/glm-5v-turbo",
|