@hyperspaceng/neural-ai 0.67.69 → 0.68.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +5 -1
- package/dist/env-api-keys.d.ts.map +1 -1
- package/dist/env-api-keys.js +1 -0
- package/dist/env-api-keys.js.map +1 -1
- package/dist/models.generated.d.ts +618 -103
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +627 -114
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/amazon-bedrock.d.ts.map +1 -1
- package/dist/providers/amazon-bedrock.js +55 -7
- package/dist/providers/amazon-bedrock.js.map +1 -1
- package/dist/providers/anthropic.d.ts.map +1 -1
- package/dist/providers/anthropic.js +141 -10
- package/dist/providers/anthropic.js.map +1 -1
- package/dist/providers/google-shared.d.ts.map +1 -1
- package/dist/providers/google-shared.js +30 -4
- package/dist/providers/google-shared.js.map +1 -1
- package/dist/providers/openai-completions.d.ts +5 -1
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +149 -60
- package/dist/providers/openai-completions.js.map +1 -1
- package/dist/providers/openai-responses-shared.d.ts.map +1 -1
- package/dist/providers/openai-responses-shared.js +2 -5
- package/dist/providers/openai-responses-shared.js.map +1 -1
- package/dist/providers/simple-options.d.ts.map +1 -1
- package/dist/providers/simple-options.js +1 -1
- package/dist/providers/simple-options.js.map +1 -1
- package/dist/providers/transform-messages.d.ts.map +1 -1
- package/dist/providers/transform-messages.js +41 -2
- package/dist/providers/transform-messages.js.map +1 -1
- package/dist/types.d.ts +5 -1
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/dist/utils/json-parse.d.ts +8 -1
- package/dist/utils/json-parse.d.ts.map +1 -1
- package/dist/utils/json-parse.js +89 -5
- package/dist/utils/json-parse.js.map +1 -1
- package/dist/utils/oauth/anthropic.d.ts.map +1 -1
- package/dist/utils/oauth/anthropic.js +1 -1
- package/dist/utils/oauth/anthropic.js.map +1 -1
- package/dist/utils/oauth/google-antigravity.d.ts.map +1 -1
- package/dist/utils/oauth/google-antigravity.js +2 -1
- package/dist/utils/oauth/google-antigravity.js.map +1 -1
- package/dist/utils/oauth/google-gemini-cli.d.ts.map +1 -1
- package/dist/utils/oauth/google-gemini-cli.js +2 -1
- package/dist/utils/oauth/google-gemini-cli.js.map +1 -1
- package/dist/utils/oauth/openai-codex.d.ts.map +1 -1
- package/dist/utils/oauth/openai-codex.js +3 -2
- package/dist/utils/oauth/openai-codex.js.map +1 -1
- package/package.json +1 -1
package/dist/models.generated.js
CHANGED
|
@@ -381,7 +381,7 @@ export const MODELS = {
|
|
|
381
381
|
name: "Claude Haiku 4.5 (EU)",
|
|
382
382
|
api: "bedrock-converse-stream",
|
|
383
383
|
provider: "amazon-bedrock",
|
|
384
|
-
baseUrl: "https://bedrock-runtime.
|
|
384
|
+
baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
|
|
385
385
|
reasoning: true,
|
|
386
386
|
input: ["text", "image"],
|
|
387
387
|
cost: {
|
|
@@ -398,7 +398,7 @@ export const MODELS = {
|
|
|
398
398
|
name: "Claude Opus 4.5 (EU)",
|
|
399
399
|
api: "bedrock-converse-stream",
|
|
400
400
|
provider: "amazon-bedrock",
|
|
401
|
-
baseUrl: "https://bedrock-runtime.
|
|
401
|
+
baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
|
|
402
402
|
reasoning: true,
|
|
403
403
|
input: ["text", "image"],
|
|
404
404
|
cost: {
|
|
@@ -415,7 +415,7 @@ export const MODELS = {
|
|
|
415
415
|
name: "Claude Opus 4.6 (EU)",
|
|
416
416
|
api: "bedrock-converse-stream",
|
|
417
417
|
provider: "amazon-bedrock",
|
|
418
|
-
baseUrl: "https://bedrock-runtime.
|
|
418
|
+
baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
|
|
419
419
|
reasoning: true,
|
|
420
420
|
input: ["text", "image"],
|
|
421
421
|
cost: {
|
|
@@ -432,7 +432,7 @@ export const MODELS = {
|
|
|
432
432
|
name: "Claude Opus 4.7 (EU)",
|
|
433
433
|
api: "bedrock-converse-stream",
|
|
434
434
|
provider: "amazon-bedrock",
|
|
435
|
-
baseUrl: "https://bedrock-runtime.
|
|
435
|
+
baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
|
|
436
436
|
reasoning: true,
|
|
437
437
|
input: ["text", "image"],
|
|
438
438
|
cost: {
|
|
@@ -449,7 +449,7 @@ export const MODELS = {
|
|
|
449
449
|
name: "Claude Sonnet 4 (EU)",
|
|
450
450
|
api: "bedrock-converse-stream",
|
|
451
451
|
provider: "amazon-bedrock",
|
|
452
|
-
baseUrl: "https://bedrock-runtime.
|
|
452
|
+
baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
|
|
453
453
|
reasoning: true,
|
|
454
454
|
input: ["text", "image"],
|
|
455
455
|
cost: {
|
|
@@ -466,7 +466,7 @@ export const MODELS = {
|
|
|
466
466
|
name: "Claude Sonnet 4.5 (EU)",
|
|
467
467
|
api: "bedrock-converse-stream",
|
|
468
468
|
provider: "amazon-bedrock",
|
|
469
|
-
baseUrl: "https://bedrock-runtime.
|
|
469
|
+
baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
|
|
470
470
|
reasoning: true,
|
|
471
471
|
input: ["text", "image"],
|
|
472
472
|
cost: {
|
|
@@ -483,7 +483,7 @@ export const MODELS = {
|
|
|
483
483
|
name: "Claude Sonnet 4.6 (EU)",
|
|
484
484
|
api: "bedrock-converse-stream",
|
|
485
485
|
provider: "amazon-bedrock",
|
|
486
|
-
baseUrl: "https://bedrock-runtime.
|
|
486
|
+
baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
|
|
487
487
|
reasoning: true,
|
|
488
488
|
input: ["text", "image"],
|
|
489
489
|
cost: {
|
|
@@ -2695,6 +2695,314 @@ export const MODELS = {
|
|
|
2695
2695
|
maxTokens: 40000,
|
|
2696
2696
|
},
|
|
2697
2697
|
},
|
|
2698
|
+
"fireworks": {
|
|
2699
|
+
"accounts/fireworks/models/deepseek-v3p1": {
|
|
2700
|
+
id: "accounts/fireworks/models/deepseek-v3p1",
|
|
2701
|
+
name: "DeepSeek V3.1",
|
|
2702
|
+
api: "anthropic-messages",
|
|
2703
|
+
provider: "fireworks",
|
|
2704
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2705
|
+
reasoning: true,
|
|
2706
|
+
input: ["text"],
|
|
2707
|
+
cost: {
|
|
2708
|
+
input: 0.56,
|
|
2709
|
+
output: 1.68,
|
|
2710
|
+
cacheRead: 0,
|
|
2711
|
+
cacheWrite: 0,
|
|
2712
|
+
},
|
|
2713
|
+
contextWindow: 163840,
|
|
2714
|
+
maxTokens: 163840,
|
|
2715
|
+
},
|
|
2716
|
+
"accounts/fireworks/models/deepseek-v3p2": {
|
|
2717
|
+
id: "accounts/fireworks/models/deepseek-v3p2",
|
|
2718
|
+
name: "DeepSeek V3.2",
|
|
2719
|
+
api: "anthropic-messages",
|
|
2720
|
+
provider: "fireworks",
|
|
2721
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2722
|
+
reasoning: true,
|
|
2723
|
+
input: ["text"],
|
|
2724
|
+
cost: {
|
|
2725
|
+
input: 0.56,
|
|
2726
|
+
output: 1.68,
|
|
2727
|
+
cacheRead: 0.28,
|
|
2728
|
+
cacheWrite: 0,
|
|
2729
|
+
},
|
|
2730
|
+
contextWindow: 160000,
|
|
2731
|
+
maxTokens: 160000,
|
|
2732
|
+
},
|
|
2733
|
+
"accounts/fireworks/models/glm-4p5": {
|
|
2734
|
+
id: "accounts/fireworks/models/glm-4p5",
|
|
2735
|
+
name: "GLM 4.5",
|
|
2736
|
+
api: "anthropic-messages",
|
|
2737
|
+
provider: "fireworks",
|
|
2738
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2739
|
+
reasoning: true,
|
|
2740
|
+
input: ["text"],
|
|
2741
|
+
cost: {
|
|
2742
|
+
input: 0.55,
|
|
2743
|
+
output: 2.19,
|
|
2744
|
+
cacheRead: 0,
|
|
2745
|
+
cacheWrite: 0,
|
|
2746
|
+
},
|
|
2747
|
+
contextWindow: 131072,
|
|
2748
|
+
maxTokens: 131072,
|
|
2749
|
+
},
|
|
2750
|
+
"accounts/fireworks/models/glm-4p5-air": {
|
|
2751
|
+
id: "accounts/fireworks/models/glm-4p5-air",
|
|
2752
|
+
name: "GLM 4.5 Air",
|
|
2753
|
+
api: "anthropic-messages",
|
|
2754
|
+
provider: "fireworks",
|
|
2755
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2756
|
+
reasoning: true,
|
|
2757
|
+
input: ["text"],
|
|
2758
|
+
cost: {
|
|
2759
|
+
input: 0.22,
|
|
2760
|
+
output: 0.88,
|
|
2761
|
+
cacheRead: 0,
|
|
2762
|
+
cacheWrite: 0,
|
|
2763
|
+
},
|
|
2764
|
+
contextWindow: 131072,
|
|
2765
|
+
maxTokens: 131072,
|
|
2766
|
+
},
|
|
2767
|
+
"accounts/fireworks/models/glm-4p7": {
|
|
2768
|
+
id: "accounts/fireworks/models/glm-4p7",
|
|
2769
|
+
name: "GLM 4.7",
|
|
2770
|
+
api: "anthropic-messages",
|
|
2771
|
+
provider: "fireworks",
|
|
2772
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2773
|
+
reasoning: true,
|
|
2774
|
+
input: ["text"],
|
|
2775
|
+
cost: {
|
|
2776
|
+
input: 0.6,
|
|
2777
|
+
output: 2.2,
|
|
2778
|
+
cacheRead: 0.3,
|
|
2779
|
+
cacheWrite: 0,
|
|
2780
|
+
},
|
|
2781
|
+
contextWindow: 198000,
|
|
2782
|
+
maxTokens: 198000,
|
|
2783
|
+
},
|
|
2784
|
+
"accounts/fireworks/models/glm-5": {
|
|
2785
|
+
id: "accounts/fireworks/models/glm-5",
|
|
2786
|
+
name: "GLM 5",
|
|
2787
|
+
api: "anthropic-messages",
|
|
2788
|
+
provider: "fireworks",
|
|
2789
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2790
|
+
reasoning: true,
|
|
2791
|
+
input: ["text"],
|
|
2792
|
+
cost: {
|
|
2793
|
+
input: 1,
|
|
2794
|
+
output: 3.2,
|
|
2795
|
+
cacheRead: 0.5,
|
|
2796
|
+
cacheWrite: 0,
|
|
2797
|
+
},
|
|
2798
|
+
contextWindow: 202752,
|
|
2799
|
+
maxTokens: 131072,
|
|
2800
|
+
},
|
|
2801
|
+
"accounts/fireworks/models/glm-5p1": {
|
|
2802
|
+
id: "accounts/fireworks/models/glm-5p1",
|
|
2803
|
+
name: "GLM 5.1",
|
|
2804
|
+
api: "anthropic-messages",
|
|
2805
|
+
provider: "fireworks",
|
|
2806
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2807
|
+
reasoning: true,
|
|
2808
|
+
input: ["text"],
|
|
2809
|
+
cost: {
|
|
2810
|
+
input: 1.4,
|
|
2811
|
+
output: 4.4,
|
|
2812
|
+
cacheRead: 0.26,
|
|
2813
|
+
cacheWrite: 0,
|
|
2814
|
+
},
|
|
2815
|
+
contextWindow: 202800,
|
|
2816
|
+
maxTokens: 131072,
|
|
2817
|
+
},
|
|
2818
|
+
"accounts/fireworks/models/gpt-oss-120b": {
|
|
2819
|
+
id: "accounts/fireworks/models/gpt-oss-120b",
|
|
2820
|
+
name: "GPT OSS 120B",
|
|
2821
|
+
api: "anthropic-messages",
|
|
2822
|
+
provider: "fireworks",
|
|
2823
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2824
|
+
reasoning: true,
|
|
2825
|
+
input: ["text"],
|
|
2826
|
+
cost: {
|
|
2827
|
+
input: 0.15,
|
|
2828
|
+
output: 0.6,
|
|
2829
|
+
cacheRead: 0,
|
|
2830
|
+
cacheWrite: 0,
|
|
2831
|
+
},
|
|
2832
|
+
contextWindow: 131072,
|
|
2833
|
+
maxTokens: 32768,
|
|
2834
|
+
},
|
|
2835
|
+
"accounts/fireworks/models/gpt-oss-20b": {
|
|
2836
|
+
id: "accounts/fireworks/models/gpt-oss-20b",
|
|
2837
|
+
name: "GPT OSS 20B",
|
|
2838
|
+
api: "anthropic-messages",
|
|
2839
|
+
provider: "fireworks",
|
|
2840
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2841
|
+
reasoning: true,
|
|
2842
|
+
input: ["text"],
|
|
2843
|
+
cost: {
|
|
2844
|
+
input: 0.05,
|
|
2845
|
+
output: 0.2,
|
|
2846
|
+
cacheRead: 0,
|
|
2847
|
+
cacheWrite: 0,
|
|
2848
|
+
},
|
|
2849
|
+
contextWindow: 131072,
|
|
2850
|
+
maxTokens: 32768,
|
|
2851
|
+
},
|
|
2852
|
+
"accounts/fireworks/models/kimi-k2-instruct": {
|
|
2853
|
+
id: "accounts/fireworks/models/kimi-k2-instruct",
|
|
2854
|
+
name: "Kimi K2 Instruct",
|
|
2855
|
+
api: "anthropic-messages",
|
|
2856
|
+
provider: "fireworks",
|
|
2857
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2858
|
+
reasoning: false,
|
|
2859
|
+
input: ["text"],
|
|
2860
|
+
cost: {
|
|
2861
|
+
input: 1,
|
|
2862
|
+
output: 3,
|
|
2863
|
+
cacheRead: 0,
|
|
2864
|
+
cacheWrite: 0,
|
|
2865
|
+
},
|
|
2866
|
+
contextWindow: 128000,
|
|
2867
|
+
maxTokens: 16384,
|
|
2868
|
+
},
|
|
2869
|
+
"accounts/fireworks/models/kimi-k2-thinking": {
|
|
2870
|
+
id: "accounts/fireworks/models/kimi-k2-thinking",
|
|
2871
|
+
name: "Kimi K2 Thinking",
|
|
2872
|
+
api: "anthropic-messages",
|
|
2873
|
+
provider: "fireworks",
|
|
2874
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2875
|
+
reasoning: true,
|
|
2876
|
+
input: ["text"],
|
|
2877
|
+
cost: {
|
|
2878
|
+
input: 0.6,
|
|
2879
|
+
output: 2.5,
|
|
2880
|
+
cacheRead: 0.3,
|
|
2881
|
+
cacheWrite: 0,
|
|
2882
|
+
},
|
|
2883
|
+
contextWindow: 256000,
|
|
2884
|
+
maxTokens: 256000,
|
|
2885
|
+
},
|
|
2886
|
+
"accounts/fireworks/models/kimi-k2p5": {
|
|
2887
|
+
id: "accounts/fireworks/models/kimi-k2p5",
|
|
2888
|
+
name: "Kimi K2.5",
|
|
2889
|
+
api: "anthropic-messages",
|
|
2890
|
+
provider: "fireworks",
|
|
2891
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2892
|
+
reasoning: true,
|
|
2893
|
+
input: ["text", "image"],
|
|
2894
|
+
cost: {
|
|
2895
|
+
input: 0.6,
|
|
2896
|
+
output: 3,
|
|
2897
|
+
cacheRead: 0.1,
|
|
2898
|
+
cacheWrite: 0,
|
|
2899
|
+
},
|
|
2900
|
+
contextWindow: 256000,
|
|
2901
|
+
maxTokens: 256000,
|
|
2902
|
+
},
|
|
2903
|
+
"accounts/fireworks/models/kimi-k2p6": {
|
|
2904
|
+
id: "accounts/fireworks/models/kimi-k2p6",
|
|
2905
|
+
name: "Kimi K2.6",
|
|
2906
|
+
api: "anthropic-messages",
|
|
2907
|
+
provider: "fireworks",
|
|
2908
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2909
|
+
reasoning: true,
|
|
2910
|
+
input: ["text", "image"],
|
|
2911
|
+
cost: {
|
|
2912
|
+
input: 0.95,
|
|
2913
|
+
output: 4,
|
|
2914
|
+
cacheRead: 0.16,
|
|
2915
|
+
cacheWrite: 0,
|
|
2916
|
+
},
|
|
2917
|
+
contextWindow: 262000,
|
|
2918
|
+
maxTokens: 262000,
|
|
2919
|
+
},
|
|
2920
|
+
"accounts/fireworks/models/minimax-m2p1": {
|
|
2921
|
+
id: "accounts/fireworks/models/minimax-m2p1",
|
|
2922
|
+
name: "MiniMax-M2.1",
|
|
2923
|
+
api: "anthropic-messages",
|
|
2924
|
+
provider: "fireworks",
|
|
2925
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2926
|
+
reasoning: true,
|
|
2927
|
+
input: ["text"],
|
|
2928
|
+
cost: {
|
|
2929
|
+
input: 0.3,
|
|
2930
|
+
output: 1.2,
|
|
2931
|
+
cacheRead: 0.03,
|
|
2932
|
+
cacheWrite: 0,
|
|
2933
|
+
},
|
|
2934
|
+
contextWindow: 200000,
|
|
2935
|
+
maxTokens: 200000,
|
|
2936
|
+
},
|
|
2937
|
+
"accounts/fireworks/models/minimax-m2p5": {
|
|
2938
|
+
id: "accounts/fireworks/models/minimax-m2p5",
|
|
2939
|
+
name: "MiniMax-M2.5",
|
|
2940
|
+
api: "anthropic-messages",
|
|
2941
|
+
provider: "fireworks",
|
|
2942
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2943
|
+
reasoning: true,
|
|
2944
|
+
input: ["text"],
|
|
2945
|
+
cost: {
|
|
2946
|
+
input: 0.3,
|
|
2947
|
+
output: 1.2,
|
|
2948
|
+
cacheRead: 0.03,
|
|
2949
|
+
cacheWrite: 0,
|
|
2950
|
+
},
|
|
2951
|
+
contextWindow: 196608,
|
|
2952
|
+
maxTokens: 196608,
|
|
2953
|
+
},
|
|
2954
|
+
"accounts/fireworks/models/minimax-m2p7": {
|
|
2955
|
+
id: "accounts/fireworks/models/minimax-m2p7",
|
|
2956
|
+
name: "MiniMax-M2.7",
|
|
2957
|
+
api: "anthropic-messages",
|
|
2958
|
+
provider: "fireworks",
|
|
2959
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2960
|
+
reasoning: true,
|
|
2961
|
+
input: ["text"],
|
|
2962
|
+
cost: {
|
|
2963
|
+
input: 0.3,
|
|
2964
|
+
output: 1.2,
|
|
2965
|
+
cacheRead: 0.03,
|
|
2966
|
+
cacheWrite: 0,
|
|
2967
|
+
},
|
|
2968
|
+
contextWindow: 196608,
|
|
2969
|
+
maxTokens: 196608,
|
|
2970
|
+
},
|
|
2971
|
+
"accounts/fireworks/models/qwen3p6-plus": {
|
|
2972
|
+
id: "accounts/fireworks/models/qwen3p6-plus",
|
|
2973
|
+
name: "Qwen 3.6 Plus",
|
|
2974
|
+
api: "anthropic-messages",
|
|
2975
|
+
provider: "fireworks",
|
|
2976
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2977
|
+
reasoning: true,
|
|
2978
|
+
input: ["text", "image"],
|
|
2979
|
+
cost: {
|
|
2980
|
+
input: 0.5,
|
|
2981
|
+
output: 3,
|
|
2982
|
+
cacheRead: 0.1,
|
|
2983
|
+
cacheWrite: 0,
|
|
2984
|
+
},
|
|
2985
|
+
contextWindow: 128000,
|
|
2986
|
+
maxTokens: 8192,
|
|
2987
|
+
},
|
|
2988
|
+
"accounts/fireworks/routers/kimi-k2p5-turbo": {
|
|
2989
|
+
id: "accounts/fireworks/routers/kimi-k2p5-turbo",
|
|
2990
|
+
name: "Kimi K2.5 Turbo (firepass)",
|
|
2991
|
+
api: "anthropic-messages",
|
|
2992
|
+
provider: "fireworks",
|
|
2993
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2994
|
+
reasoning: true,
|
|
2995
|
+
input: ["text", "image"],
|
|
2996
|
+
cost: {
|
|
2997
|
+
input: 0,
|
|
2998
|
+
output: 0,
|
|
2999
|
+
cacheRead: 0,
|
|
3000
|
+
cacheWrite: 0,
|
|
3001
|
+
},
|
|
3002
|
+
contextWindow: 256000,
|
|
3003
|
+
maxTokens: 256000,
|
|
3004
|
+
},
|
|
3005
|
+
},
|
|
2698
3006
|
"github-copilot": {
|
|
2699
3007
|
"claude-haiku-4.5": {
|
|
2700
3008
|
id: "claude-haiku-4.5",
|
|
@@ -3251,7 +3559,7 @@ export const MODELS = {
|
|
|
3251
3559
|
cost: {
|
|
3252
3560
|
input: 0.3,
|
|
3253
3561
|
output: 2.5,
|
|
3254
|
-
cacheRead: 0.
|
|
3562
|
+
cacheRead: 0.03,
|
|
3255
3563
|
cacheWrite: 0,
|
|
3256
3564
|
},
|
|
3257
3565
|
contextWindow: 1048576,
|
|
@@ -3370,7 +3678,7 @@ export const MODELS = {
|
|
|
3370
3678
|
cost: {
|
|
3371
3679
|
input: 1.25,
|
|
3372
3680
|
output: 10,
|
|
3373
|
-
cacheRead: 0.
|
|
3681
|
+
cacheRead: 0.125,
|
|
3374
3682
|
cacheWrite: 0,
|
|
3375
3683
|
},
|
|
3376
3684
|
contextWindow: 1048576,
|
|
@@ -3580,8 +3888,8 @@ export const MODELS = {
|
|
|
3580
3888
|
contextWindow: 131072,
|
|
3581
3889
|
maxTokens: 8192,
|
|
3582
3890
|
},
|
|
3583
|
-
"gemma-4-26b-it": {
|
|
3584
|
-
id: "gemma-4-26b-it",
|
|
3891
|
+
"gemma-4-26b-a4b-it": {
|
|
3892
|
+
id: "gemma-4-26b-a4b-it",
|
|
3585
3893
|
name: "Gemma 4 26B",
|
|
3586
3894
|
api: "google-generative-ai",
|
|
3587
3895
|
provider: "google",
|
|
@@ -4694,6 +5002,24 @@ export const MODELS = {
|
|
|
4694
5002
|
contextWindow: 262144,
|
|
4695
5003
|
maxTokens: 262144,
|
|
4696
5004
|
},
|
|
5005
|
+
"moonshotai/Kimi-K2.6": {
|
|
5006
|
+
id: "moonshotai/Kimi-K2.6",
|
|
5007
|
+
name: "Kimi-K2.6",
|
|
5008
|
+
api: "openai-completions",
|
|
5009
|
+
provider: "huggingface",
|
|
5010
|
+
baseUrl: "https://router.huggingface.co/v1",
|
|
5011
|
+
compat: { "supportsDeveloperRole": false },
|
|
5012
|
+
reasoning: true,
|
|
5013
|
+
input: ["text", "image"],
|
|
5014
|
+
cost: {
|
|
5015
|
+
input: 0.95,
|
|
5016
|
+
output: 4,
|
|
5017
|
+
cacheRead: 0.16,
|
|
5018
|
+
cacheWrite: 0,
|
|
5019
|
+
},
|
|
5020
|
+
contextWindow: 262144,
|
|
5021
|
+
maxTokens: 262144,
|
|
5022
|
+
},
|
|
4697
5023
|
"zai-org/GLM-4.7": {
|
|
4698
5024
|
id: "zai-org/GLM-4.7",
|
|
4699
5025
|
name: "GLM-4.7",
|
|
@@ -4768,6 +5094,23 @@ export const MODELS = {
|
|
|
4768
5094
|
},
|
|
4769
5095
|
},
|
|
4770
5096
|
"kimi-coding": {
|
|
5097
|
+
"k2p6": {
|
|
5098
|
+
id: "k2p6",
|
|
5099
|
+
name: "Kimi K2.6",
|
|
5100
|
+
api: "anthropic-messages",
|
|
5101
|
+
provider: "kimi-coding",
|
|
5102
|
+
baseUrl: "https://api.kimi.com/coding",
|
|
5103
|
+
reasoning: true,
|
|
5104
|
+
input: ["text", "image"],
|
|
5105
|
+
cost: {
|
|
5106
|
+
input: 0,
|
|
5107
|
+
output: 0,
|
|
5108
|
+
cacheRead: 0,
|
|
5109
|
+
cacheWrite: 0,
|
|
5110
|
+
},
|
|
5111
|
+
contextWindow: 262144,
|
|
5112
|
+
maxTokens: 32768,
|
|
5113
|
+
},
|
|
4771
5114
|
"kimi-for-coding": {
|
|
4772
5115
|
id: "kimi-for-coding",
|
|
4773
5116
|
name: "Kimi For Coding",
|
|
@@ -6650,6 +6993,40 @@ export const MODELS = {
|
|
|
6650
6993
|
contextWindow: 262144,
|
|
6651
6994
|
maxTokens: 65536,
|
|
6652
6995
|
},
|
|
6996
|
+
"kimi-k2.6": {
|
|
6997
|
+
id: "kimi-k2.6",
|
|
6998
|
+
name: "Kimi K2.6",
|
|
6999
|
+
api: "openai-completions",
|
|
7000
|
+
provider: "opencode",
|
|
7001
|
+
baseUrl: "https://opencode.ai/zen/v1",
|
|
7002
|
+
reasoning: true,
|
|
7003
|
+
input: ["text", "image"],
|
|
7004
|
+
cost: {
|
|
7005
|
+
input: 0.95,
|
|
7006
|
+
output: 4,
|
|
7007
|
+
cacheRead: 0.16,
|
|
7008
|
+
cacheWrite: 0,
|
|
7009
|
+
},
|
|
7010
|
+
contextWindow: 262144,
|
|
7011
|
+
maxTokens: 65536,
|
|
7012
|
+
},
|
|
7013
|
+
"ling-2.6-flash-free": {
|
|
7014
|
+
id: "ling-2.6-flash-free",
|
|
7015
|
+
name: "Ling 2.6 Flash Free",
|
|
7016
|
+
api: "openai-completions",
|
|
7017
|
+
provider: "opencode",
|
|
7018
|
+
baseUrl: "https://opencode.ai/zen/v1",
|
|
7019
|
+
reasoning: false,
|
|
7020
|
+
input: ["text"],
|
|
7021
|
+
cost: {
|
|
7022
|
+
input: 0,
|
|
7023
|
+
output: 0,
|
|
7024
|
+
cacheRead: 0,
|
|
7025
|
+
cacheWrite: 0,
|
|
7026
|
+
},
|
|
7027
|
+
contextWindow: 262100,
|
|
7028
|
+
maxTokens: 32800,
|
|
7029
|
+
},
|
|
6653
7030
|
"minimax-m2.5": {
|
|
6654
7031
|
id: "minimax-m2.5",
|
|
6655
7032
|
name: "MiniMax M2.5",
|
|
@@ -6684,6 +7061,23 @@ export const MODELS = {
|
|
|
6684
7061
|
contextWindow: 204800,
|
|
6685
7062
|
maxTokens: 131072,
|
|
6686
7063
|
},
|
|
7064
|
+
"minimax-m2.7": {
|
|
7065
|
+
id: "minimax-m2.7",
|
|
7066
|
+
name: "MiniMax M2.7",
|
|
7067
|
+
api: "openai-completions",
|
|
7068
|
+
provider: "opencode",
|
|
7069
|
+
baseUrl: "https://opencode.ai/zen/v1",
|
|
7070
|
+
reasoning: true,
|
|
7071
|
+
input: ["text"],
|
|
7072
|
+
cost: {
|
|
7073
|
+
input: 0.3,
|
|
7074
|
+
output: 1.2,
|
|
7075
|
+
cacheRead: 0.06,
|
|
7076
|
+
cacheWrite: 0,
|
|
7077
|
+
},
|
|
7078
|
+
contextWindow: 204800,
|
|
7079
|
+
maxTokens: 131072,
|
|
7080
|
+
},
|
|
6687
7081
|
"nemotron-3-super-free": {
|
|
6688
7082
|
id: "nemotron-3-super-free",
|
|
6689
7083
|
name: "Nemotron 3 Super Free",
|
|
@@ -6704,9 +7098,9 @@ export const MODELS = {
|
|
|
6704
7098
|
"qwen3.5-plus": {
|
|
6705
7099
|
id: "qwen3.5-plus",
|
|
6706
7100
|
name: "Qwen3.5 Plus",
|
|
6707
|
-
api: "
|
|
7101
|
+
api: "anthropic-messages",
|
|
6708
7102
|
provider: "opencode",
|
|
6709
|
-
baseUrl: "https://opencode.ai/zen
|
|
7103
|
+
baseUrl: "https://opencode.ai/zen",
|
|
6710
7104
|
reasoning: true,
|
|
6711
7105
|
input: ["text", "image"],
|
|
6712
7106
|
cost: {
|
|
@@ -6721,9 +7115,9 @@ export const MODELS = {
|
|
|
6721
7115
|
"qwen3.6-plus": {
|
|
6722
7116
|
id: "qwen3.6-plus",
|
|
6723
7117
|
name: "Qwen3.6 Plus",
|
|
6724
|
-
api: "
|
|
7118
|
+
api: "anthropic-messages",
|
|
6725
7119
|
provider: "opencode",
|
|
6726
|
-
baseUrl: "https://opencode.ai/zen
|
|
7120
|
+
baseUrl: "https://opencode.ai/zen",
|
|
6727
7121
|
reasoning: true,
|
|
6728
7122
|
input: ["text", "image"],
|
|
6729
7123
|
cost: {
|
|
@@ -6788,6 +7182,23 @@ export const MODELS = {
|
|
|
6788
7182
|
contextWindow: 262144,
|
|
6789
7183
|
maxTokens: 65536,
|
|
6790
7184
|
},
|
|
7185
|
+
"kimi-k2.6": {
|
|
7186
|
+
id: "kimi-k2.6",
|
|
7187
|
+
name: "Kimi K2.6 (3x limits)",
|
|
7188
|
+
api: "openai-completions",
|
|
7189
|
+
provider: "opencode-go",
|
|
7190
|
+
baseUrl: "https://opencode.ai/zen/go/v1",
|
|
7191
|
+
reasoning: true,
|
|
7192
|
+
input: ["text", "image"],
|
|
7193
|
+
cost: {
|
|
7194
|
+
input: 0.32,
|
|
7195
|
+
output: 1.34,
|
|
7196
|
+
cacheRead: 0.054,
|
|
7197
|
+
cacheWrite: 0,
|
|
7198
|
+
},
|
|
7199
|
+
contextWindow: 262144,
|
|
7200
|
+
maxTokens: 65536,
|
|
7201
|
+
},
|
|
6791
7202
|
"mimo-v2-omni": {
|
|
6792
7203
|
id: "mimo-v2-omni",
|
|
6793
7204
|
name: "MiMo V2 Omni",
|
|
@@ -6803,7 +7214,7 @@ export const MODELS = {
|
|
|
6803
7214
|
cacheWrite: 0,
|
|
6804
7215
|
},
|
|
6805
7216
|
contextWindow: 262144,
|
|
6806
|
-
maxTokens:
|
|
7217
|
+
maxTokens: 128000,
|
|
6807
7218
|
},
|
|
6808
7219
|
"mimo-v2-pro": {
|
|
6809
7220
|
id: "mimo-v2-pro",
|
|
@@ -6820,7 +7231,41 @@ export const MODELS = {
|
|
|
6820
7231
|
cacheWrite: 0,
|
|
6821
7232
|
},
|
|
6822
7233
|
contextWindow: 1048576,
|
|
6823
|
-
maxTokens:
|
|
7234
|
+
maxTokens: 128000,
|
|
7235
|
+
},
|
|
7236
|
+
"mimo-v2.5": {
|
|
7237
|
+
id: "mimo-v2.5",
|
|
7238
|
+
name: "MiMo V2.5",
|
|
7239
|
+
api: "openai-completions",
|
|
7240
|
+
provider: "opencode-go",
|
|
7241
|
+
baseUrl: "https://opencode.ai/zen/go/v1",
|
|
7242
|
+
reasoning: true,
|
|
7243
|
+
input: ["text", "image"],
|
|
7244
|
+
cost: {
|
|
7245
|
+
input: 0.4,
|
|
7246
|
+
output: 2,
|
|
7247
|
+
cacheRead: 0.08,
|
|
7248
|
+
cacheWrite: 0,
|
|
7249
|
+
},
|
|
7250
|
+
contextWindow: 262144,
|
|
7251
|
+
maxTokens: 128000,
|
|
7252
|
+
},
|
|
7253
|
+
"mimo-v2.5-pro": {
|
|
7254
|
+
id: "mimo-v2.5-pro",
|
|
7255
|
+
name: "MiMo V2.5 Pro",
|
|
7256
|
+
api: "openai-completions",
|
|
7257
|
+
provider: "opencode-go",
|
|
7258
|
+
baseUrl: "https://opencode.ai/zen/go/v1",
|
|
7259
|
+
reasoning: true,
|
|
7260
|
+
input: ["text"],
|
|
7261
|
+
cost: {
|
|
7262
|
+
input: 1,
|
|
7263
|
+
output: 3,
|
|
7264
|
+
cacheRead: 0.2,
|
|
7265
|
+
cacheWrite: 0,
|
|
7266
|
+
},
|
|
7267
|
+
contextWindow: 1048576,
|
|
7268
|
+
maxTokens: 128000,
|
|
6824
7269
|
},
|
|
6825
7270
|
"minimax-m2.5": {
|
|
6826
7271
|
id: "minimax-m2.5",
|
|
@@ -6859,9 +7304,9 @@ export const MODELS = {
|
|
|
6859
7304
|
"qwen3.5-plus": {
|
|
6860
7305
|
id: "qwen3.5-plus",
|
|
6861
7306
|
name: "Qwen3.5 Plus",
|
|
6862
|
-
api: "
|
|
7307
|
+
api: "anthropic-messages",
|
|
6863
7308
|
provider: "opencode-go",
|
|
6864
|
-
baseUrl: "https://opencode.ai/zen/go
|
|
7309
|
+
baseUrl: "https://opencode.ai/zen/go",
|
|
6865
7310
|
reasoning: true,
|
|
6866
7311
|
input: ["text", "image"],
|
|
6867
7312
|
cost: {
|
|
@@ -6876,9 +7321,9 @@ export const MODELS = {
|
|
|
6876
7321
|
"qwen3.6-plus": {
|
|
6877
7322
|
id: "qwen3.6-plus",
|
|
6878
7323
|
name: "Qwen3.6 Plus",
|
|
6879
|
-
api: "
|
|
7324
|
+
api: "anthropic-messages",
|
|
6880
7325
|
provider: "opencode-go",
|
|
6881
|
-
baseUrl: "https://opencode.ai/zen/go
|
|
7326
|
+
baseUrl: "https://opencode.ai/zen/go",
|
|
6882
7327
|
reasoning: true,
|
|
6883
7328
|
input: ["text", "image"],
|
|
6884
7329
|
cost: {
|
|
@@ -7266,17 +7711,17 @@ export const MODELS = {
|
|
|
7266
7711
|
contextWindow: 1000000,
|
|
7267
7712
|
maxTokens: 128000,
|
|
7268
7713
|
},
|
|
7269
|
-
"arcee-ai/trinity-large-preview
|
|
7270
|
-
id: "arcee-ai/trinity-large-preview
|
|
7271
|
-
name: "Arcee AI: Trinity Large Preview
|
|
7714
|
+
"arcee-ai/trinity-large-preview": {
|
|
7715
|
+
id: "arcee-ai/trinity-large-preview",
|
|
7716
|
+
name: "Arcee AI: Trinity Large Preview",
|
|
7272
7717
|
api: "openai-completions",
|
|
7273
7718
|
provider: "openrouter",
|
|
7274
7719
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
7275
7720
|
reasoning: false,
|
|
7276
7721
|
input: ["text"],
|
|
7277
7722
|
cost: {
|
|
7278
|
-
input: 0,
|
|
7279
|
-
output: 0,
|
|
7723
|
+
input: 0.15,
|
|
7724
|
+
output: 0.44999999999999996,
|
|
7280
7725
|
cacheRead: 0,
|
|
7281
7726
|
cacheWrite: 0,
|
|
7282
7727
|
},
|
|
@@ -7598,13 +8043,13 @@ export const MODELS = {
|
|
|
7598
8043
|
reasoning: true,
|
|
7599
8044
|
input: ["text"],
|
|
7600
8045
|
cost: {
|
|
7601
|
-
input: 0.
|
|
7602
|
-
output: 0.
|
|
7603
|
-
cacheRead: 0.
|
|
8046
|
+
input: 0.252,
|
|
8047
|
+
output: 0.378,
|
|
8048
|
+
cacheRead: 0.0252,
|
|
7604
8049
|
cacheWrite: 0,
|
|
7605
8050
|
},
|
|
7606
|
-
contextWindow:
|
|
7607
|
-
maxTokens:
|
|
8051
|
+
contextWindow: 131072,
|
|
8052
|
+
maxTokens: 65536,
|
|
7608
8053
|
},
|
|
7609
8054
|
"deepseek/deepseek-v3.2-exp": {
|
|
7610
8055
|
id: "deepseek/deepseek-v3.2-exp",
|
|
@@ -7654,7 +8099,7 @@ export const MODELS = {
|
|
|
7654
8099
|
cacheRead: 0.024999999999999998,
|
|
7655
8100
|
cacheWrite: 0.08333333333333334,
|
|
7656
8101
|
},
|
|
7657
|
-
contextWindow:
|
|
8102
|
+
contextWindow: 1000000,
|
|
7658
8103
|
maxTokens: 8192,
|
|
7659
8104
|
},
|
|
7660
8105
|
"google/gemini-2.0-flash-lite-001": {
|
|
@@ -7853,9 +8298,9 @@ export const MODELS = {
|
|
|
7853
8298
|
reasoning: true,
|
|
7854
8299
|
input: ["text", "image"],
|
|
7855
8300
|
cost: {
|
|
7856
|
-
input: 0.
|
|
7857
|
-
output: 0.
|
|
7858
|
-
cacheRead: 0
|
|
8301
|
+
input: 0.07,
|
|
8302
|
+
output: 0.33999999999999997,
|
|
8303
|
+
cacheRead: 0,
|
|
7859
8304
|
cacheWrite: 0,
|
|
7860
8305
|
},
|
|
7861
8306
|
contextWindow: 262144,
|
|
@@ -7889,7 +8334,7 @@ export const MODELS = {
|
|
|
7889
8334
|
cost: {
|
|
7890
8335
|
input: 0.13,
|
|
7891
8336
|
output: 0.38,
|
|
7892
|
-
cacheRead: 0
|
|
8337
|
+
cacheRead: 0,
|
|
7893
8338
|
cacheWrite: 0,
|
|
7894
8339
|
},
|
|
7895
8340
|
contextWindow: 262144,
|
|
@@ -7929,6 +8374,23 @@ export const MODELS = {
|
|
|
7929
8374
|
contextWindow: 128000,
|
|
7930
8375
|
maxTokens: 50000,
|
|
7931
8376
|
},
|
|
8377
|
+
"inclusionai/ling-2.6-flash:free": {
|
|
8378
|
+
id: "inclusionai/ling-2.6-flash:free",
|
|
8379
|
+
name: "inclusionAI: Ling-2.6-flash (free)",
|
|
8380
|
+
api: "openai-completions",
|
|
8381
|
+
provider: "openrouter",
|
|
8382
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
8383
|
+
reasoning: false,
|
|
8384
|
+
input: ["text"],
|
|
8385
|
+
cost: {
|
|
8386
|
+
input: 0,
|
|
8387
|
+
output: 0,
|
|
8388
|
+
cacheRead: 0,
|
|
8389
|
+
cacheWrite: 0,
|
|
8390
|
+
},
|
|
8391
|
+
contextWindow: 262144,
|
|
8392
|
+
maxTokens: 32768,
|
|
8393
|
+
},
|
|
7932
8394
|
"kwaipilot/kat-coder-pro-v2": {
|
|
7933
8395
|
id: "kwaipilot/kat-coder-pro-v2",
|
|
7934
8396
|
name: "Kwaipilot: KAT-Coder-Pro V2",
|
|
@@ -8006,13 +8468,13 @@ export const MODELS = {
|
|
|
8006
8468
|
reasoning: false,
|
|
8007
8469
|
input: ["text"],
|
|
8008
8470
|
cost: {
|
|
8009
|
-
input: 0.
|
|
8010
|
-
output: 0.
|
|
8471
|
+
input: 0.09999999999999999,
|
|
8472
|
+
output: 0.32,
|
|
8011
8473
|
cacheRead: 0,
|
|
8012
8474
|
cacheWrite: 0,
|
|
8013
8475
|
},
|
|
8014
8476
|
contextWindow: 131072,
|
|
8015
|
-
maxTokens:
|
|
8477
|
+
maxTokens: 16384,
|
|
8016
8478
|
},
|
|
8017
8479
|
"meta-llama/llama-3.3-70b-instruct:free": {
|
|
8018
8480
|
id: "meta-llama/llama-3.3-70b-instruct:free",
|
|
@@ -8108,9 +8570,9 @@ export const MODELS = {
|
|
|
8108
8570
|
reasoning: true,
|
|
8109
8571
|
input: ["text"],
|
|
8110
8572
|
cost: {
|
|
8111
|
-
input: 0.
|
|
8112
|
-
output:
|
|
8113
|
-
cacheRead: 0.
|
|
8573
|
+
input: 0.15,
|
|
8574
|
+
output: 1.2,
|
|
8575
|
+
cacheRead: 0.075,
|
|
8114
8576
|
cacheWrite: 0,
|
|
8115
8577
|
},
|
|
8116
8578
|
contextWindow: 196608,
|
|
@@ -8592,6 +9054,23 @@ export const MODELS = {
|
|
|
8592
9054
|
contextWindow: 262144,
|
|
8593
9055
|
maxTokens: 4096,
|
|
8594
9056
|
},
|
|
9057
|
+
"moonshotai/kimi-k2.6": {
|
|
9058
|
+
id: "moonshotai/kimi-k2.6",
|
|
9059
|
+
name: "MoonshotAI: Kimi K2.6",
|
|
9060
|
+
api: "openai-completions",
|
|
9061
|
+
provider: "openrouter",
|
|
9062
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
9063
|
+
reasoning: true,
|
|
9064
|
+
input: ["text", "image"],
|
|
9065
|
+
cost: {
|
|
9066
|
+
input: 0.75,
|
|
9067
|
+
output: 3.5,
|
|
9068
|
+
cacheRead: 0.15,
|
|
9069
|
+
cacheWrite: 0,
|
|
9070
|
+
},
|
|
9071
|
+
contextWindow: 262144,
|
|
9072
|
+
maxTokens: 4096,
|
|
9073
|
+
},
|
|
8595
9074
|
"nex-agi/deepseek-v3.1-nex-n1": {
|
|
8596
9075
|
id: "nex-agi/deepseek-v3.1-nex-n1",
|
|
8597
9076
|
name: "Nex AGI: DeepSeek V3.1 Nex N1",
|
|
@@ -9102,40 +9581,6 @@ export const MODELS = {
|
|
|
9102
9581
|
contextWindow: 400000,
|
|
9103
9582
|
maxTokens: 128000,
|
|
9104
9583
|
},
|
|
9105
|
-
"openai/gpt-5-image": {
|
|
9106
|
-
id: "openai/gpt-5-image",
|
|
9107
|
-
name: "OpenAI: GPT-5 Image",
|
|
9108
|
-
api: "openai-completions",
|
|
9109
|
-
provider: "openrouter",
|
|
9110
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
9111
|
-
reasoning: true,
|
|
9112
|
-
input: ["text", "image"],
|
|
9113
|
-
cost: {
|
|
9114
|
-
input: 10,
|
|
9115
|
-
output: 10,
|
|
9116
|
-
cacheRead: 1.25,
|
|
9117
|
-
cacheWrite: 0,
|
|
9118
|
-
},
|
|
9119
|
-
contextWindow: 400000,
|
|
9120
|
-
maxTokens: 128000,
|
|
9121
|
-
},
|
|
9122
|
-
"openai/gpt-5-image-mini": {
|
|
9123
|
-
id: "openai/gpt-5-image-mini",
|
|
9124
|
-
name: "OpenAI: GPT-5 Image Mini",
|
|
9125
|
-
api: "openai-completions",
|
|
9126
|
-
provider: "openrouter",
|
|
9127
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
9128
|
-
reasoning: true,
|
|
9129
|
-
input: ["text", "image"],
|
|
9130
|
-
cost: {
|
|
9131
|
-
input: 2.5,
|
|
9132
|
-
output: 2,
|
|
9133
|
-
cacheRead: 0.25,
|
|
9134
|
-
cacheWrite: 0,
|
|
9135
|
-
},
|
|
9136
|
-
contextWindow: 400000,
|
|
9137
|
-
maxTokens: 128000,
|
|
9138
|
-
},
|
|
9139
9584
|
"openai/gpt-5-mini": {
|
|
9140
9585
|
id: "openai/gpt-5-mini",
|
|
9141
9586
|
name: "OpenAI: GPT-5 Mini",
|
|
@@ -9731,23 +10176,6 @@ export const MODELS = {
|
|
|
9731
10176
|
contextWindow: 2000000,
|
|
9732
10177
|
maxTokens: 4096,
|
|
9733
10178
|
},
|
|
9734
|
-
"openrouter/elephant-alpha": {
|
|
9735
|
-
id: "openrouter/elephant-alpha",
|
|
9736
|
-
name: "Elephant",
|
|
9737
|
-
api: "openai-completions",
|
|
9738
|
-
provider: "openrouter",
|
|
9739
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
9740
|
-
reasoning: false,
|
|
9741
|
-
input: ["text"],
|
|
9742
|
-
cost: {
|
|
9743
|
-
input: 0,
|
|
9744
|
-
output: 0,
|
|
9745
|
-
cacheRead: 0,
|
|
9746
|
-
cacheWrite: 0,
|
|
9747
|
-
},
|
|
9748
|
-
contextWindow: 262144,
|
|
9749
|
-
maxTokens: 32768,
|
|
9750
|
-
},
|
|
9751
10179
|
"openrouter/free": {
|
|
9752
10180
|
id: "openrouter/free",
|
|
9753
10181
|
name: "Free Models Router",
|
|
@@ -9978,13 +10406,13 @@ export const MODELS = {
|
|
|
9978
10406
|
reasoning: true,
|
|
9979
10407
|
input: ["text"],
|
|
9980
10408
|
cost: {
|
|
9981
|
-
input: 0.
|
|
9982
|
-
output:
|
|
10409
|
+
input: 0.14950000000000002,
|
|
10410
|
+
output: 1.495,
|
|
9983
10411
|
cacheRead: 0,
|
|
9984
10412
|
cacheWrite: 0,
|
|
9985
10413
|
},
|
|
9986
|
-
contextWindow:
|
|
9987
|
-
maxTokens:
|
|
10414
|
+
contextWindow: 131072,
|
|
10415
|
+
maxTokens: 4096,
|
|
9988
10416
|
},
|
|
9989
10417
|
"qwen/qwen3-30b-a3b": {
|
|
9990
10418
|
id: "qwen/qwen3-30b-a3b",
|
|
@@ -10133,7 +10561,7 @@ export const MODELS = {
|
|
|
10133
10561
|
cost: {
|
|
10134
10562
|
input: 0.15,
|
|
10135
10563
|
output: 0.7999999999999999,
|
|
10136
|
-
cacheRead: 0.
|
|
10564
|
+
cacheRead: 0.11,
|
|
10137
10565
|
cacheWrite: 0,
|
|
10138
10566
|
},
|
|
10139
10567
|
contextWindow: 262144,
|
|
@@ -10887,6 +11315,40 @@ export const MODELS = {
|
|
|
10887
11315
|
contextWindow: 1048576,
|
|
10888
11316
|
maxTokens: 131072,
|
|
10889
11317
|
},
|
|
11318
|
+
"xiaomi/mimo-v2.5": {
|
|
11319
|
+
id: "xiaomi/mimo-v2.5",
|
|
11320
|
+
name: "Xiaomi: MiMo-V2.5",
|
|
11321
|
+
api: "openai-completions",
|
|
11322
|
+
provider: "openrouter",
|
|
11323
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
11324
|
+
reasoning: true,
|
|
11325
|
+
input: ["text", "image"],
|
|
11326
|
+
cost: {
|
|
11327
|
+
input: 0.39999999999999997,
|
|
11328
|
+
output: 2,
|
|
11329
|
+
cacheRead: 0.08,
|
|
11330
|
+
cacheWrite: 0,
|
|
11331
|
+
},
|
|
11332
|
+
contextWindow: 1048576,
|
|
11333
|
+
maxTokens: 131072,
|
|
11334
|
+
},
|
|
11335
|
+
"xiaomi/mimo-v2.5-pro": {
|
|
11336
|
+
id: "xiaomi/mimo-v2.5-pro",
|
|
11337
|
+
name: "Xiaomi: MiMo-V2.5-Pro",
|
|
11338
|
+
api: "openai-completions",
|
|
11339
|
+
provider: "openrouter",
|
|
11340
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
11341
|
+
reasoning: true,
|
|
11342
|
+
input: ["text"],
|
|
11343
|
+
cost: {
|
|
11344
|
+
input: 1,
|
|
11345
|
+
output: 3,
|
|
11346
|
+
cacheRead: 0.19999999999999998,
|
|
11347
|
+
cacheWrite: 0,
|
|
11348
|
+
},
|
|
11349
|
+
contextWindow: 1048576,
|
|
11350
|
+
maxTokens: 131072,
|
|
11351
|
+
},
|
|
10890
11352
|
"z-ai/glm-4-32b": {
|
|
10891
11353
|
id: "z-ai/glm-4-32b",
|
|
10892
11354
|
name: "Z.ai: GLM 4 32B ",
|
|
@@ -11015,13 +11477,13 @@ export const MODELS = {
|
|
|
11015
11477
|
reasoning: true,
|
|
11016
11478
|
input: ["text"],
|
|
11017
11479
|
cost: {
|
|
11018
|
-
input: 0.
|
|
11019
|
-
output: 1.
|
|
11020
|
-
cacheRead: 0
|
|
11480
|
+
input: 0.38,
|
|
11481
|
+
output: 1.74,
|
|
11482
|
+
cacheRead: 0,
|
|
11021
11483
|
cacheWrite: 0,
|
|
11022
11484
|
},
|
|
11023
11485
|
contextWindow: 202752,
|
|
11024
|
-
maxTokens:
|
|
11486
|
+
maxTokens: 4096,
|
|
11025
11487
|
},
|
|
11026
11488
|
"z-ai/glm-4.7-flash": {
|
|
11027
11489
|
id: "z-ai/glm-4.7-flash",
|
|
@@ -11054,8 +11516,8 @@ export const MODELS = {
|
|
|
11054
11516
|
cacheRead: 0.119,
|
|
11055
11517
|
cacheWrite: 0,
|
|
11056
11518
|
},
|
|
11057
|
-
contextWindow:
|
|
11058
|
-
maxTokens:
|
|
11519
|
+
contextWindow: 202752,
|
|
11520
|
+
maxTokens: 4096,
|
|
11059
11521
|
},
|
|
11060
11522
|
"z-ai/glm-5-turbo": {
|
|
11061
11523
|
id: "z-ai/glm-5-turbo",
|
|
@@ -11083,9 +11545,9 @@ export const MODELS = {
|
|
|
11083
11545
|
reasoning: true,
|
|
11084
11546
|
input: ["text"],
|
|
11085
11547
|
cost: {
|
|
11086
|
-
input:
|
|
11087
|
-
output: 3.
|
|
11088
|
-
cacheRead: 0.
|
|
11548
|
+
input: 1.0499999999999998,
|
|
11549
|
+
output: 3.5,
|
|
11550
|
+
cacheRead: 0.5249999999999999,
|
|
11089
11551
|
cacheWrite: 0,
|
|
11090
11552
|
},
|
|
11091
11553
|
contextWindow: 202752,
|
|
@@ -11108,6 +11570,23 @@ export const MODELS = {
|
|
|
11108
11570
|
contextWindow: 202752,
|
|
11109
11571
|
maxTokens: 131072,
|
|
11110
11572
|
},
|
|
11573
|
+
"~anthropic/claude-opus-latest": {
|
|
11574
|
+
id: "~anthropic/claude-opus-latest",
|
|
11575
|
+
name: "Anthropic: Claude Opus Latest",
|
|
11576
|
+
api: "openai-completions",
|
|
11577
|
+
provider: "openrouter",
|
|
11578
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
11579
|
+
reasoning: true,
|
|
11580
|
+
input: ["text", "image"],
|
|
11581
|
+
cost: {
|
|
11582
|
+
input: 5,
|
|
11583
|
+
output: 25,
|
|
11584
|
+
cacheRead: 0.5,
|
|
11585
|
+
cacheWrite: 6.25,
|
|
11586
|
+
},
|
|
11587
|
+
contextWindow: 1000000,
|
|
11588
|
+
maxTokens: 128000,
|
|
11589
|
+
},
|
|
11111
11590
|
},
|
|
11112
11591
|
"vercel-ai-gateway": {
|
|
11113
11592
|
"alibaba/qwen-3-14b": {
|
|
@@ -11178,6 +11657,23 @@ export const MODELS = {
|
|
|
11178
11657
|
contextWindow: 128000,
|
|
11179
11658
|
maxTokens: 8192,
|
|
11180
11659
|
},
|
|
11660
|
+
"alibaba/qwen-3.6-max-preview": {
|
|
11661
|
+
id: "alibaba/qwen-3.6-max-preview",
|
|
11662
|
+
name: "Qwen 3.6 Max Preview",
|
|
11663
|
+
api: "anthropic-messages",
|
|
11664
|
+
provider: "vercel-ai-gateway",
|
|
11665
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11666
|
+
reasoning: true,
|
|
11667
|
+
input: ["text", "image"],
|
|
11668
|
+
cost: {
|
|
11669
|
+
input: 1.3,
|
|
11670
|
+
output: 7.8,
|
|
11671
|
+
cacheRead: 0.26,
|
|
11672
|
+
cacheWrite: 1.625,
|
|
11673
|
+
},
|
|
11674
|
+
contextWindow: 240000,
|
|
11675
|
+
maxTokens: 64000,
|
|
11676
|
+
},
|
|
11181
11677
|
"alibaba/qwen3-235b-a22b-thinking": {
|
|
11182
11678
|
id: "alibaba/qwen3-235b-a22b-thinking",
|
|
11183
11679
|
name: "Qwen3 235B A22B Thinking 2507",
|
|
@@ -11377,7 +11873,7 @@ export const MODELS = {
|
|
|
11377
11873
|
input: 0.5,
|
|
11378
11874
|
output: 3,
|
|
11379
11875
|
cacheRead: 0.09999999999999999,
|
|
11380
|
-
cacheWrite: 0,
|
|
11876
|
+
cacheWrite: 0.625,
|
|
11381
11877
|
},
|
|
11382
11878
|
contextWindow: 1000000,
|
|
11383
11879
|
maxTokens: 64000,
|
|
@@ -12521,6 +13017,23 @@ export const MODELS = {
|
|
|
12521
13017
|
contextWindow: 262114,
|
|
12522
13018
|
maxTokens: 262114,
|
|
12523
13019
|
},
|
|
13020
|
+
"moonshotai/kimi-k2.6": {
|
|
13021
|
+
id: "moonshotai/kimi-k2.6",
|
|
13022
|
+
name: "Kimi K2.6",
|
|
13023
|
+
api: "anthropic-messages",
|
|
13024
|
+
provider: "vercel-ai-gateway",
|
|
13025
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
13026
|
+
reasoning: true,
|
|
13027
|
+
input: ["text", "image"],
|
|
13028
|
+
cost: {
|
|
13029
|
+
input: 0.95,
|
|
13030
|
+
output: 4,
|
|
13031
|
+
cacheRead: 0.16,
|
|
13032
|
+
cacheWrite: 0,
|
|
13033
|
+
},
|
|
13034
|
+
contextWindow: 262000,
|
|
13035
|
+
maxTokens: 262000,
|
|
13036
|
+
},
|
|
12524
13037
|
"nvidia/nemotron-nano-12b-v2-vl": {
|
|
12525
13038
|
id: "nvidia/nemotron-nano-12b-v2-vl",
|
|
12526
13039
|
name: "Nvidia Nemotron Nano 12B V2 VL",
|
|
@@ -13482,9 +13995,9 @@ export const MODELS = {
|
|
|
13482
13995
|
reasoning: true,
|
|
13483
13996
|
input: ["text"],
|
|
13484
13997
|
cost: {
|
|
13485
|
-
input: 0.
|
|
13486
|
-
output: 0.
|
|
13487
|
-
cacheRead: 0.
|
|
13998
|
+
input: 0.09999999999999999,
|
|
13999
|
+
output: 0.3,
|
|
14000
|
+
cacheRead: 0.01,
|
|
13488
14001
|
cacheWrite: 0,
|
|
13489
14002
|
},
|
|
13490
14003
|
contextWindow: 262144,
|
|
@@ -13701,15 +14214,15 @@ export const MODELS = {
|
|
|
13701
14214
|
provider: "vercel-ai-gateway",
|
|
13702
14215
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
13703
14216
|
reasoning: true,
|
|
13704
|
-
input: ["text"
|
|
14217
|
+
input: ["text"],
|
|
13705
14218
|
cost: {
|
|
13706
14219
|
input: 1.4,
|
|
13707
14220
|
output: 4.4,
|
|
13708
14221
|
cacheRead: 0.26,
|
|
13709
14222
|
cacheWrite: 0,
|
|
13710
14223
|
},
|
|
13711
|
-
contextWindow:
|
|
13712
|
-
maxTokens:
|
|
14224
|
+
contextWindow: 202800,
|
|
14225
|
+
maxTokens: 64000,
|
|
13713
14226
|
},
|
|
13714
14227
|
"zai/glm-5v-turbo": {
|
|
13715
14228
|
id: "zai/glm-5v-turbo",
|