@mariozechner/pi-ai 0.67.68 → 0.68.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +5 -1
- package/dist/env-api-keys.d.ts.map +1 -1
- package/dist/env-api-keys.js +1 -0
- package/dist/env-api-keys.js.map +1 -1
- package/dist/models.generated.d.ts +502 -77
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +532 -117
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/amazon-bedrock.d.ts.map +1 -1
- package/dist/providers/amazon-bedrock.js +55 -7
- package/dist/providers/amazon-bedrock.js.map +1 -1
- package/dist/providers/anthropic.d.ts.map +1 -1
- package/dist/providers/anthropic.js +141 -10
- package/dist/providers/anthropic.js.map +1 -1
- package/dist/providers/google-shared.d.ts.map +1 -1
- package/dist/providers/google-shared.js +30 -4
- package/dist/providers/google-shared.js.map +1 -1
- package/dist/providers/openai-completions.d.ts +5 -1
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +149 -60
- package/dist/providers/openai-completions.js.map +1 -1
- package/dist/providers/openai-responses-shared.d.ts.map +1 -1
- package/dist/providers/openai-responses-shared.js +2 -5
- package/dist/providers/openai-responses-shared.js.map +1 -1
- package/dist/providers/simple-options.d.ts.map +1 -1
- package/dist/providers/simple-options.js +1 -1
- package/dist/providers/simple-options.js.map +1 -1
- package/dist/providers/transform-messages.d.ts.map +1 -1
- package/dist/providers/transform-messages.js +41 -2
- package/dist/providers/transform-messages.js.map +1 -1
- package/dist/types.d.ts +5 -1
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/dist/utils/json-parse.d.ts +8 -1
- package/dist/utils/json-parse.d.ts.map +1 -1
- package/dist/utils/json-parse.js +89 -5
- package/dist/utils/json-parse.js.map +1 -1
- package/dist/utils/oauth/anthropic.d.ts.map +1 -1
- package/dist/utils/oauth/anthropic.js +1 -1
- package/dist/utils/oauth/anthropic.js.map +1 -1
- package/dist/utils/oauth/google-antigravity.d.ts.map +1 -1
- package/dist/utils/oauth/google-antigravity.js +2 -1
- package/dist/utils/oauth/google-antigravity.js.map +1 -1
- package/dist/utils/oauth/google-gemini-cli.d.ts.map +1 -1
- package/dist/utils/oauth/google-gemini-cli.js +2 -1
- package/dist/utils/oauth/google-gemini-cli.js.map +1 -1
- package/dist/utils/oauth/openai-codex.d.ts.map +1 -1
- package/dist/utils/oauth/openai-codex.js +3 -2
- package/dist/utils/oauth/openai-codex.js.map +1 -1
- package/package.json +1 -1
package/dist/models.generated.js
CHANGED
|
@@ -381,7 +381,7 @@ export const MODELS = {
|
|
|
381
381
|
name: "Claude Haiku 4.5 (EU)",
|
|
382
382
|
api: "bedrock-converse-stream",
|
|
383
383
|
provider: "amazon-bedrock",
|
|
384
|
-
baseUrl: "https://bedrock-runtime.
|
|
384
|
+
baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
|
|
385
385
|
reasoning: true,
|
|
386
386
|
input: ["text", "image"],
|
|
387
387
|
cost: {
|
|
@@ -398,7 +398,7 @@ export const MODELS = {
|
|
|
398
398
|
name: "Claude Opus 4.5 (EU)",
|
|
399
399
|
api: "bedrock-converse-stream",
|
|
400
400
|
provider: "amazon-bedrock",
|
|
401
|
-
baseUrl: "https://bedrock-runtime.
|
|
401
|
+
baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
|
|
402
402
|
reasoning: true,
|
|
403
403
|
input: ["text", "image"],
|
|
404
404
|
cost: {
|
|
@@ -415,7 +415,7 @@ export const MODELS = {
|
|
|
415
415
|
name: "Claude Opus 4.6 (EU)",
|
|
416
416
|
api: "bedrock-converse-stream",
|
|
417
417
|
provider: "amazon-bedrock",
|
|
418
|
-
baseUrl: "https://bedrock-runtime.
|
|
418
|
+
baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
|
|
419
419
|
reasoning: true,
|
|
420
420
|
input: ["text", "image"],
|
|
421
421
|
cost: {
|
|
@@ -432,7 +432,7 @@ export const MODELS = {
|
|
|
432
432
|
name: "Claude Opus 4.7 (EU)",
|
|
433
433
|
api: "bedrock-converse-stream",
|
|
434
434
|
provider: "amazon-bedrock",
|
|
435
|
-
baseUrl: "https://bedrock-runtime.
|
|
435
|
+
baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
|
|
436
436
|
reasoning: true,
|
|
437
437
|
input: ["text", "image"],
|
|
438
438
|
cost: {
|
|
@@ -449,7 +449,7 @@ export const MODELS = {
|
|
|
449
449
|
name: "Claude Sonnet 4 (EU)",
|
|
450
450
|
api: "bedrock-converse-stream",
|
|
451
451
|
provider: "amazon-bedrock",
|
|
452
|
-
baseUrl: "https://bedrock-runtime.
|
|
452
|
+
baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
|
|
453
453
|
reasoning: true,
|
|
454
454
|
input: ["text", "image"],
|
|
455
455
|
cost: {
|
|
@@ -466,7 +466,7 @@ export const MODELS = {
|
|
|
466
466
|
name: "Claude Sonnet 4.5 (EU)",
|
|
467
467
|
api: "bedrock-converse-stream",
|
|
468
468
|
provider: "amazon-bedrock",
|
|
469
|
-
baseUrl: "https://bedrock-runtime.
|
|
469
|
+
baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
|
|
470
470
|
reasoning: true,
|
|
471
471
|
input: ["text", "image"],
|
|
472
472
|
cost: {
|
|
@@ -483,7 +483,7 @@ export const MODELS = {
|
|
|
483
483
|
name: "Claude Sonnet 4.6 (EU)",
|
|
484
484
|
api: "bedrock-converse-stream",
|
|
485
485
|
provider: "amazon-bedrock",
|
|
486
|
-
baseUrl: "https://bedrock-runtime.
|
|
486
|
+
baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
|
|
487
487
|
reasoning: true,
|
|
488
488
|
input: ["text", "image"],
|
|
489
489
|
cost: {
|
|
@@ -2695,6 +2695,314 @@ export const MODELS = {
|
|
|
2695
2695
|
maxTokens: 40000,
|
|
2696
2696
|
},
|
|
2697
2697
|
},
|
|
2698
|
+
"fireworks": {
|
|
2699
|
+
"accounts/fireworks/models/deepseek-v3p1": {
|
|
2700
|
+
id: "accounts/fireworks/models/deepseek-v3p1",
|
|
2701
|
+
name: "DeepSeek V3.1",
|
|
2702
|
+
api: "anthropic-messages",
|
|
2703
|
+
provider: "fireworks",
|
|
2704
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2705
|
+
reasoning: true,
|
|
2706
|
+
input: ["text"],
|
|
2707
|
+
cost: {
|
|
2708
|
+
input: 0.56,
|
|
2709
|
+
output: 1.68,
|
|
2710
|
+
cacheRead: 0,
|
|
2711
|
+
cacheWrite: 0,
|
|
2712
|
+
},
|
|
2713
|
+
contextWindow: 163840,
|
|
2714
|
+
maxTokens: 163840,
|
|
2715
|
+
},
|
|
2716
|
+
"accounts/fireworks/models/deepseek-v3p2": {
|
|
2717
|
+
id: "accounts/fireworks/models/deepseek-v3p2",
|
|
2718
|
+
name: "DeepSeek V3.2",
|
|
2719
|
+
api: "anthropic-messages",
|
|
2720
|
+
provider: "fireworks",
|
|
2721
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2722
|
+
reasoning: true,
|
|
2723
|
+
input: ["text"],
|
|
2724
|
+
cost: {
|
|
2725
|
+
input: 0.56,
|
|
2726
|
+
output: 1.68,
|
|
2727
|
+
cacheRead: 0.28,
|
|
2728
|
+
cacheWrite: 0,
|
|
2729
|
+
},
|
|
2730
|
+
contextWindow: 160000,
|
|
2731
|
+
maxTokens: 160000,
|
|
2732
|
+
},
|
|
2733
|
+
"accounts/fireworks/models/glm-4p5": {
|
|
2734
|
+
id: "accounts/fireworks/models/glm-4p5",
|
|
2735
|
+
name: "GLM 4.5",
|
|
2736
|
+
api: "anthropic-messages",
|
|
2737
|
+
provider: "fireworks",
|
|
2738
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2739
|
+
reasoning: true,
|
|
2740
|
+
input: ["text"],
|
|
2741
|
+
cost: {
|
|
2742
|
+
input: 0.55,
|
|
2743
|
+
output: 2.19,
|
|
2744
|
+
cacheRead: 0,
|
|
2745
|
+
cacheWrite: 0,
|
|
2746
|
+
},
|
|
2747
|
+
contextWindow: 131072,
|
|
2748
|
+
maxTokens: 131072,
|
|
2749
|
+
},
|
|
2750
|
+
"accounts/fireworks/models/glm-4p5-air": {
|
|
2751
|
+
id: "accounts/fireworks/models/glm-4p5-air",
|
|
2752
|
+
name: "GLM 4.5 Air",
|
|
2753
|
+
api: "anthropic-messages",
|
|
2754
|
+
provider: "fireworks",
|
|
2755
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2756
|
+
reasoning: true,
|
|
2757
|
+
input: ["text"],
|
|
2758
|
+
cost: {
|
|
2759
|
+
input: 0.22,
|
|
2760
|
+
output: 0.88,
|
|
2761
|
+
cacheRead: 0,
|
|
2762
|
+
cacheWrite: 0,
|
|
2763
|
+
},
|
|
2764
|
+
contextWindow: 131072,
|
|
2765
|
+
maxTokens: 131072,
|
|
2766
|
+
},
|
|
2767
|
+
"accounts/fireworks/models/glm-4p7": {
|
|
2768
|
+
id: "accounts/fireworks/models/glm-4p7",
|
|
2769
|
+
name: "GLM 4.7",
|
|
2770
|
+
api: "anthropic-messages",
|
|
2771
|
+
provider: "fireworks",
|
|
2772
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2773
|
+
reasoning: true,
|
|
2774
|
+
input: ["text"],
|
|
2775
|
+
cost: {
|
|
2776
|
+
input: 0.6,
|
|
2777
|
+
output: 2.2,
|
|
2778
|
+
cacheRead: 0.3,
|
|
2779
|
+
cacheWrite: 0,
|
|
2780
|
+
},
|
|
2781
|
+
contextWindow: 198000,
|
|
2782
|
+
maxTokens: 198000,
|
|
2783
|
+
},
|
|
2784
|
+
"accounts/fireworks/models/glm-5": {
|
|
2785
|
+
id: "accounts/fireworks/models/glm-5",
|
|
2786
|
+
name: "GLM 5",
|
|
2787
|
+
api: "anthropic-messages",
|
|
2788
|
+
provider: "fireworks",
|
|
2789
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2790
|
+
reasoning: true,
|
|
2791
|
+
input: ["text"],
|
|
2792
|
+
cost: {
|
|
2793
|
+
input: 1,
|
|
2794
|
+
output: 3.2,
|
|
2795
|
+
cacheRead: 0.5,
|
|
2796
|
+
cacheWrite: 0,
|
|
2797
|
+
},
|
|
2798
|
+
contextWindow: 202752,
|
|
2799
|
+
maxTokens: 131072,
|
|
2800
|
+
},
|
|
2801
|
+
"accounts/fireworks/models/glm-5p1": {
|
|
2802
|
+
id: "accounts/fireworks/models/glm-5p1",
|
|
2803
|
+
name: "GLM 5.1",
|
|
2804
|
+
api: "anthropic-messages",
|
|
2805
|
+
provider: "fireworks",
|
|
2806
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2807
|
+
reasoning: true,
|
|
2808
|
+
input: ["text"],
|
|
2809
|
+
cost: {
|
|
2810
|
+
input: 1.4,
|
|
2811
|
+
output: 4.4,
|
|
2812
|
+
cacheRead: 0.26,
|
|
2813
|
+
cacheWrite: 0,
|
|
2814
|
+
},
|
|
2815
|
+
contextWindow: 202800,
|
|
2816
|
+
maxTokens: 131072,
|
|
2817
|
+
},
|
|
2818
|
+
"accounts/fireworks/models/gpt-oss-120b": {
|
|
2819
|
+
id: "accounts/fireworks/models/gpt-oss-120b",
|
|
2820
|
+
name: "GPT OSS 120B",
|
|
2821
|
+
api: "anthropic-messages",
|
|
2822
|
+
provider: "fireworks",
|
|
2823
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2824
|
+
reasoning: true,
|
|
2825
|
+
input: ["text"],
|
|
2826
|
+
cost: {
|
|
2827
|
+
input: 0.15,
|
|
2828
|
+
output: 0.6,
|
|
2829
|
+
cacheRead: 0,
|
|
2830
|
+
cacheWrite: 0,
|
|
2831
|
+
},
|
|
2832
|
+
contextWindow: 131072,
|
|
2833
|
+
maxTokens: 32768,
|
|
2834
|
+
},
|
|
2835
|
+
"accounts/fireworks/models/gpt-oss-20b": {
|
|
2836
|
+
id: "accounts/fireworks/models/gpt-oss-20b",
|
|
2837
|
+
name: "GPT OSS 20B",
|
|
2838
|
+
api: "anthropic-messages",
|
|
2839
|
+
provider: "fireworks",
|
|
2840
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2841
|
+
reasoning: true,
|
|
2842
|
+
input: ["text"],
|
|
2843
|
+
cost: {
|
|
2844
|
+
input: 0.05,
|
|
2845
|
+
output: 0.2,
|
|
2846
|
+
cacheRead: 0,
|
|
2847
|
+
cacheWrite: 0,
|
|
2848
|
+
},
|
|
2849
|
+
contextWindow: 131072,
|
|
2850
|
+
maxTokens: 32768,
|
|
2851
|
+
},
|
|
2852
|
+
"accounts/fireworks/models/kimi-k2-instruct": {
|
|
2853
|
+
id: "accounts/fireworks/models/kimi-k2-instruct",
|
|
2854
|
+
name: "Kimi K2 Instruct",
|
|
2855
|
+
api: "anthropic-messages",
|
|
2856
|
+
provider: "fireworks",
|
|
2857
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2858
|
+
reasoning: false,
|
|
2859
|
+
input: ["text"],
|
|
2860
|
+
cost: {
|
|
2861
|
+
input: 1,
|
|
2862
|
+
output: 3,
|
|
2863
|
+
cacheRead: 0,
|
|
2864
|
+
cacheWrite: 0,
|
|
2865
|
+
},
|
|
2866
|
+
contextWindow: 128000,
|
|
2867
|
+
maxTokens: 16384,
|
|
2868
|
+
},
|
|
2869
|
+
"accounts/fireworks/models/kimi-k2-thinking": {
|
|
2870
|
+
id: "accounts/fireworks/models/kimi-k2-thinking",
|
|
2871
|
+
name: "Kimi K2 Thinking",
|
|
2872
|
+
api: "anthropic-messages",
|
|
2873
|
+
provider: "fireworks",
|
|
2874
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2875
|
+
reasoning: true,
|
|
2876
|
+
input: ["text"],
|
|
2877
|
+
cost: {
|
|
2878
|
+
input: 0.6,
|
|
2879
|
+
output: 2.5,
|
|
2880
|
+
cacheRead: 0.3,
|
|
2881
|
+
cacheWrite: 0,
|
|
2882
|
+
},
|
|
2883
|
+
contextWindow: 256000,
|
|
2884
|
+
maxTokens: 256000,
|
|
2885
|
+
},
|
|
2886
|
+
"accounts/fireworks/models/kimi-k2p5": {
|
|
2887
|
+
id: "accounts/fireworks/models/kimi-k2p5",
|
|
2888
|
+
name: "Kimi K2.5",
|
|
2889
|
+
api: "anthropic-messages",
|
|
2890
|
+
provider: "fireworks",
|
|
2891
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2892
|
+
reasoning: true,
|
|
2893
|
+
input: ["text", "image"],
|
|
2894
|
+
cost: {
|
|
2895
|
+
input: 0.6,
|
|
2896
|
+
output: 3,
|
|
2897
|
+
cacheRead: 0.1,
|
|
2898
|
+
cacheWrite: 0,
|
|
2899
|
+
},
|
|
2900
|
+
contextWindow: 256000,
|
|
2901
|
+
maxTokens: 256000,
|
|
2902
|
+
},
|
|
2903
|
+
"accounts/fireworks/models/kimi-k2p6": {
|
|
2904
|
+
id: "accounts/fireworks/models/kimi-k2p6",
|
|
2905
|
+
name: "Kimi K2.6",
|
|
2906
|
+
api: "anthropic-messages",
|
|
2907
|
+
provider: "fireworks",
|
|
2908
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2909
|
+
reasoning: true,
|
|
2910
|
+
input: ["text", "image"],
|
|
2911
|
+
cost: {
|
|
2912
|
+
input: 0.95,
|
|
2913
|
+
output: 4,
|
|
2914
|
+
cacheRead: 0.16,
|
|
2915
|
+
cacheWrite: 0,
|
|
2916
|
+
},
|
|
2917
|
+
contextWindow: 262000,
|
|
2918
|
+
maxTokens: 262000,
|
|
2919
|
+
},
|
|
2920
|
+
"accounts/fireworks/models/minimax-m2p1": {
|
|
2921
|
+
id: "accounts/fireworks/models/minimax-m2p1",
|
|
2922
|
+
name: "MiniMax-M2.1",
|
|
2923
|
+
api: "anthropic-messages",
|
|
2924
|
+
provider: "fireworks",
|
|
2925
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2926
|
+
reasoning: true,
|
|
2927
|
+
input: ["text"],
|
|
2928
|
+
cost: {
|
|
2929
|
+
input: 0.3,
|
|
2930
|
+
output: 1.2,
|
|
2931
|
+
cacheRead: 0.03,
|
|
2932
|
+
cacheWrite: 0,
|
|
2933
|
+
},
|
|
2934
|
+
contextWindow: 200000,
|
|
2935
|
+
maxTokens: 200000,
|
|
2936
|
+
},
|
|
2937
|
+
"accounts/fireworks/models/minimax-m2p5": {
|
|
2938
|
+
id: "accounts/fireworks/models/minimax-m2p5",
|
|
2939
|
+
name: "MiniMax-M2.5",
|
|
2940
|
+
api: "anthropic-messages",
|
|
2941
|
+
provider: "fireworks",
|
|
2942
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2943
|
+
reasoning: true,
|
|
2944
|
+
input: ["text"],
|
|
2945
|
+
cost: {
|
|
2946
|
+
input: 0.3,
|
|
2947
|
+
output: 1.2,
|
|
2948
|
+
cacheRead: 0.03,
|
|
2949
|
+
cacheWrite: 0,
|
|
2950
|
+
},
|
|
2951
|
+
contextWindow: 196608,
|
|
2952
|
+
maxTokens: 196608,
|
|
2953
|
+
},
|
|
2954
|
+
"accounts/fireworks/models/minimax-m2p7": {
|
|
2955
|
+
id: "accounts/fireworks/models/minimax-m2p7",
|
|
2956
|
+
name: "MiniMax-M2.7",
|
|
2957
|
+
api: "anthropic-messages",
|
|
2958
|
+
provider: "fireworks",
|
|
2959
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2960
|
+
reasoning: true,
|
|
2961
|
+
input: ["text"],
|
|
2962
|
+
cost: {
|
|
2963
|
+
input: 0.3,
|
|
2964
|
+
output: 1.2,
|
|
2965
|
+
cacheRead: 0.03,
|
|
2966
|
+
cacheWrite: 0,
|
|
2967
|
+
},
|
|
2968
|
+
contextWindow: 196608,
|
|
2969
|
+
maxTokens: 196608,
|
|
2970
|
+
},
|
|
2971
|
+
"accounts/fireworks/models/qwen3p6-plus": {
|
|
2972
|
+
id: "accounts/fireworks/models/qwen3p6-plus",
|
|
2973
|
+
name: "Qwen 3.6 Plus",
|
|
2974
|
+
api: "anthropic-messages",
|
|
2975
|
+
provider: "fireworks",
|
|
2976
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2977
|
+
reasoning: true,
|
|
2978
|
+
input: ["text", "image"],
|
|
2979
|
+
cost: {
|
|
2980
|
+
input: 0.5,
|
|
2981
|
+
output: 3,
|
|
2982
|
+
cacheRead: 0.1,
|
|
2983
|
+
cacheWrite: 0,
|
|
2984
|
+
},
|
|
2985
|
+
contextWindow: 128000,
|
|
2986
|
+
maxTokens: 8192,
|
|
2987
|
+
},
|
|
2988
|
+
"accounts/fireworks/routers/kimi-k2p5-turbo": {
|
|
2989
|
+
id: "accounts/fireworks/routers/kimi-k2p5-turbo",
|
|
2990
|
+
name: "Kimi K2.5 Turbo (firepass)",
|
|
2991
|
+
api: "anthropic-messages",
|
|
2992
|
+
provider: "fireworks",
|
|
2993
|
+
baseUrl: "https://api.fireworks.ai/inference",
|
|
2994
|
+
reasoning: true,
|
|
2995
|
+
input: ["text", "image"],
|
|
2996
|
+
cost: {
|
|
2997
|
+
input: 0,
|
|
2998
|
+
output: 0,
|
|
2999
|
+
cacheRead: 0,
|
|
3000
|
+
cacheWrite: 0,
|
|
3001
|
+
},
|
|
3002
|
+
contextWindow: 256000,
|
|
3003
|
+
maxTokens: 256000,
|
|
3004
|
+
},
|
|
3005
|
+
},
|
|
2698
3006
|
"github-copilot": {
|
|
2699
3007
|
"claude-haiku-4.5": {
|
|
2700
3008
|
id: "claude-haiku-4.5",
|
|
@@ -3251,7 +3559,7 @@ export const MODELS = {
|
|
|
3251
3559
|
cost: {
|
|
3252
3560
|
input: 0.3,
|
|
3253
3561
|
output: 2.5,
|
|
3254
|
-
cacheRead: 0.
|
|
3562
|
+
cacheRead: 0.03,
|
|
3255
3563
|
cacheWrite: 0,
|
|
3256
3564
|
},
|
|
3257
3565
|
contextWindow: 1048576,
|
|
@@ -3370,7 +3678,7 @@ export const MODELS = {
|
|
|
3370
3678
|
cost: {
|
|
3371
3679
|
input: 1.25,
|
|
3372
3680
|
output: 10,
|
|
3373
|
-
cacheRead: 0.
|
|
3681
|
+
cacheRead: 0.125,
|
|
3374
3682
|
cacheWrite: 0,
|
|
3375
3683
|
},
|
|
3376
3684
|
contextWindow: 1048576,
|
|
@@ -4694,6 +5002,24 @@ export const MODELS = {
|
|
|
4694
5002
|
contextWindow: 262144,
|
|
4695
5003
|
maxTokens: 262144,
|
|
4696
5004
|
},
|
|
5005
|
+
"moonshotai/Kimi-K2.6": {
|
|
5006
|
+
id: "moonshotai/Kimi-K2.6",
|
|
5007
|
+
name: "Kimi-K2.6",
|
|
5008
|
+
api: "openai-completions",
|
|
5009
|
+
provider: "huggingface",
|
|
5010
|
+
baseUrl: "https://router.huggingface.co/v1",
|
|
5011
|
+
compat: { "supportsDeveloperRole": false },
|
|
5012
|
+
reasoning: true,
|
|
5013
|
+
input: ["text", "image"],
|
|
5014
|
+
cost: {
|
|
5015
|
+
input: 0.95,
|
|
5016
|
+
output: 4,
|
|
5017
|
+
cacheRead: 0.16,
|
|
5018
|
+
cacheWrite: 0,
|
|
5019
|
+
},
|
|
5020
|
+
contextWindow: 262144,
|
|
5021
|
+
maxTokens: 262144,
|
|
5022
|
+
},
|
|
4697
5023
|
"zai-org/GLM-4.7": {
|
|
4698
5024
|
id: "zai-org/GLM-4.7",
|
|
4699
5025
|
name: "GLM-4.7",
|
|
@@ -4768,6 +5094,23 @@ export const MODELS = {
|
|
|
4768
5094
|
},
|
|
4769
5095
|
},
|
|
4770
5096
|
"kimi-coding": {
|
|
5097
|
+
"k2p6": {
|
|
5098
|
+
id: "k2p6",
|
|
5099
|
+
name: "Kimi K2.6",
|
|
5100
|
+
api: "anthropic-messages",
|
|
5101
|
+
provider: "kimi-coding",
|
|
5102
|
+
baseUrl: "https://api.kimi.com/coding",
|
|
5103
|
+
reasoning: true,
|
|
5104
|
+
input: ["text", "image"],
|
|
5105
|
+
cost: {
|
|
5106
|
+
input: 0,
|
|
5107
|
+
output: 0,
|
|
5108
|
+
cacheRead: 0,
|
|
5109
|
+
cacheWrite: 0,
|
|
5110
|
+
},
|
|
5111
|
+
contextWindow: 262144,
|
|
5112
|
+
maxTokens: 32768,
|
|
5113
|
+
},
|
|
4771
5114
|
"kimi-for-coding": {
|
|
4772
5115
|
id: "kimi-for-coding",
|
|
4773
5116
|
name: "Kimi For Coding",
|
|
@@ -6650,6 +6993,40 @@ export const MODELS = {
|
|
|
6650
6993
|
contextWindow: 262144,
|
|
6651
6994
|
maxTokens: 65536,
|
|
6652
6995
|
},
|
|
6996
|
+
"kimi-k2.6": {
|
|
6997
|
+
id: "kimi-k2.6",
|
|
6998
|
+
name: "Kimi K2.6",
|
|
6999
|
+
api: "openai-completions",
|
|
7000
|
+
provider: "opencode",
|
|
7001
|
+
baseUrl: "https://opencode.ai/zen/v1",
|
|
7002
|
+
reasoning: true,
|
|
7003
|
+
input: ["text", "image"],
|
|
7004
|
+
cost: {
|
|
7005
|
+
input: 0.95,
|
|
7006
|
+
output: 4,
|
|
7007
|
+
cacheRead: 0.16,
|
|
7008
|
+
cacheWrite: 0,
|
|
7009
|
+
},
|
|
7010
|
+
contextWindow: 262144,
|
|
7011
|
+
maxTokens: 65536,
|
|
7012
|
+
},
|
|
7013
|
+
"ling-2.6-flash-free": {
|
|
7014
|
+
id: "ling-2.6-flash-free",
|
|
7015
|
+
name: "Ling 2.6 Flash Free",
|
|
7016
|
+
api: "openai-completions",
|
|
7017
|
+
provider: "opencode",
|
|
7018
|
+
baseUrl: "https://opencode.ai/zen/v1",
|
|
7019
|
+
reasoning: false,
|
|
7020
|
+
input: ["text"],
|
|
7021
|
+
cost: {
|
|
7022
|
+
input: 0,
|
|
7023
|
+
output: 0,
|
|
7024
|
+
cacheRead: 0,
|
|
7025
|
+
cacheWrite: 0,
|
|
7026
|
+
},
|
|
7027
|
+
contextWindow: 262100,
|
|
7028
|
+
maxTokens: 32800,
|
|
7029
|
+
},
|
|
6653
7030
|
"minimax-m2.5": {
|
|
6654
7031
|
id: "minimax-m2.5",
|
|
6655
7032
|
name: "MiniMax M2.5",
|
|
@@ -6684,6 +7061,23 @@ export const MODELS = {
|
|
|
6684
7061
|
contextWindow: 204800,
|
|
6685
7062
|
maxTokens: 131072,
|
|
6686
7063
|
},
|
|
7064
|
+
"minimax-m2.7": {
|
|
7065
|
+
id: "minimax-m2.7",
|
|
7066
|
+
name: "MiniMax M2.7",
|
|
7067
|
+
api: "openai-completions",
|
|
7068
|
+
provider: "opencode",
|
|
7069
|
+
baseUrl: "https://opencode.ai/zen/v1",
|
|
7070
|
+
reasoning: true,
|
|
7071
|
+
input: ["text"],
|
|
7072
|
+
cost: {
|
|
7073
|
+
input: 0.3,
|
|
7074
|
+
output: 1.2,
|
|
7075
|
+
cacheRead: 0.06,
|
|
7076
|
+
cacheWrite: 0,
|
|
7077
|
+
},
|
|
7078
|
+
contextWindow: 204800,
|
|
7079
|
+
maxTokens: 131072,
|
|
7080
|
+
},
|
|
6687
7081
|
"nemotron-3-super-free": {
|
|
6688
7082
|
id: "nemotron-3-super-free",
|
|
6689
7083
|
name: "Nemotron 3 Super Free",
|
|
@@ -6707,6 +7101,7 @@ export const MODELS = {
|
|
|
6707
7101
|
api: "openai-completions",
|
|
6708
7102
|
provider: "opencode",
|
|
6709
7103
|
baseUrl: "https://opencode.ai/zen/v1",
|
|
7104
|
+
compat: { "cacheControlFormat": "anthropic" },
|
|
6710
7105
|
reasoning: true,
|
|
6711
7106
|
input: ["text", "image"],
|
|
6712
7107
|
cost: {
|
|
@@ -6724,6 +7119,7 @@ export const MODELS = {
|
|
|
6724
7119
|
api: "openai-completions",
|
|
6725
7120
|
provider: "opencode",
|
|
6726
7121
|
baseUrl: "https://opencode.ai/zen/v1",
|
|
7122
|
+
compat: { "cacheControlFormat": "anthropic" },
|
|
6727
7123
|
reasoning: true,
|
|
6728
7124
|
input: ["text", "image"],
|
|
6729
7125
|
cost: {
|
|
@@ -6788,6 +7184,23 @@ export const MODELS = {
|
|
|
6788
7184
|
contextWindow: 262144,
|
|
6789
7185
|
maxTokens: 65536,
|
|
6790
7186
|
},
|
|
7187
|
+
"kimi-k2.6": {
|
|
7188
|
+
id: "kimi-k2.6",
|
|
7189
|
+
name: "Kimi K2.6 (3x limits)",
|
|
7190
|
+
api: "openai-completions",
|
|
7191
|
+
provider: "opencode-go",
|
|
7192
|
+
baseUrl: "https://opencode.ai/zen/go/v1",
|
|
7193
|
+
reasoning: true,
|
|
7194
|
+
input: ["text", "image"],
|
|
7195
|
+
cost: {
|
|
7196
|
+
input: 0.32,
|
|
7197
|
+
output: 1.34,
|
|
7198
|
+
cacheRead: 0.054,
|
|
7199
|
+
cacheWrite: 0,
|
|
7200
|
+
},
|
|
7201
|
+
contextWindow: 262144,
|
|
7202
|
+
maxTokens: 65536,
|
|
7203
|
+
},
|
|
6791
7204
|
"mimo-v2-omni": {
|
|
6792
7205
|
id: "mimo-v2-omni",
|
|
6793
7206
|
name: "MiMo V2 Omni",
|
|
@@ -6862,6 +7275,7 @@ export const MODELS = {
|
|
|
6862
7275
|
api: "openai-completions",
|
|
6863
7276
|
provider: "opencode-go",
|
|
6864
7277
|
baseUrl: "https://opencode.ai/zen/go/v1",
|
|
7278
|
+
compat: { "cacheControlFormat": "anthropic" },
|
|
6865
7279
|
reasoning: true,
|
|
6866
7280
|
input: ["text", "image"],
|
|
6867
7281
|
cost: {
|
|
@@ -6879,6 +7293,7 @@ export const MODELS = {
|
|
|
6879
7293
|
api: "openai-completions",
|
|
6880
7294
|
provider: "opencode-go",
|
|
6881
7295
|
baseUrl: "https://opencode.ai/zen/go/v1",
|
|
7296
|
+
compat: { "cacheControlFormat": "anthropic" },
|
|
6882
7297
|
reasoning: true,
|
|
6883
7298
|
input: ["text", "image"],
|
|
6884
7299
|
cost: {
|
|
@@ -7598,13 +8013,13 @@ export const MODELS = {
|
|
|
7598
8013
|
reasoning: true,
|
|
7599
8014
|
input: ["text"],
|
|
7600
8015
|
cost: {
|
|
7601
|
-
input: 0.
|
|
7602
|
-
output: 0.
|
|
7603
|
-
cacheRead: 0.
|
|
8016
|
+
input: 0.252,
|
|
8017
|
+
output: 0.378,
|
|
8018
|
+
cacheRead: 0.0252,
|
|
7604
8019
|
cacheWrite: 0,
|
|
7605
8020
|
},
|
|
7606
|
-
contextWindow:
|
|
7607
|
-
maxTokens:
|
|
8021
|
+
contextWindow: 131072,
|
|
8022
|
+
maxTokens: 32768,
|
|
7608
8023
|
},
|
|
7609
8024
|
"deepseek/deepseek-v3.2-exp": {
|
|
7610
8025
|
id: "deepseek/deepseek-v3.2-exp",
|
|
@@ -7853,13 +8268,13 @@ export const MODELS = {
|
|
|
7853
8268
|
reasoning: true,
|
|
7854
8269
|
input: ["text", "image"],
|
|
7855
8270
|
cost: {
|
|
7856
|
-
input: 0.
|
|
8271
|
+
input: 0.07,
|
|
7857
8272
|
output: 0.35,
|
|
7858
|
-
cacheRead: 0.
|
|
8273
|
+
cacheRead: 0.04,
|
|
7859
8274
|
cacheWrite: 0,
|
|
7860
8275
|
},
|
|
7861
8276
|
contextWindow: 262144,
|
|
7862
|
-
maxTokens:
|
|
8277
|
+
maxTokens: 262144,
|
|
7863
8278
|
},
|
|
7864
8279
|
"google/gemma-4-26b-a4b-it:free": {
|
|
7865
8280
|
id: "google/gemma-4-26b-a4b-it:free",
|
|
@@ -7929,6 +8344,23 @@ export const MODELS = {
|
|
|
7929
8344
|
contextWindow: 128000,
|
|
7930
8345
|
maxTokens: 50000,
|
|
7931
8346
|
},
|
|
8347
|
+
"inclusionai/ling-2.6-flash:free": {
|
|
8348
|
+
id: "inclusionai/ling-2.6-flash:free",
|
|
8349
|
+
name: "inclusionAI: Ling-2.6-flash (free)",
|
|
8350
|
+
api: "openai-completions",
|
|
8351
|
+
provider: "openrouter",
|
|
8352
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
8353
|
+
reasoning: false,
|
|
8354
|
+
input: ["text"],
|
|
8355
|
+
cost: {
|
|
8356
|
+
input: 0,
|
|
8357
|
+
output: 0,
|
|
8358
|
+
cacheRead: 0,
|
|
8359
|
+
cacheWrite: 0,
|
|
8360
|
+
},
|
|
8361
|
+
contextWindow: 262144,
|
|
8362
|
+
maxTokens: 32768,
|
|
8363
|
+
},
|
|
7932
8364
|
"kwaipilot/kat-coder-pro-v2": {
|
|
7933
8365
|
id: "kwaipilot/kat-coder-pro-v2",
|
|
7934
8366
|
name: "Kwaipilot: KAT-Coder-Pro V2",
|
|
@@ -8031,23 +8463,6 @@ export const MODELS = {
|
|
|
8031
8463
|
contextWindow: 65536,
|
|
8032
8464
|
maxTokens: 4096,
|
|
8033
8465
|
},
|
|
8034
|
-
"meta-llama/llama-4-maverick": {
|
|
8035
|
-
id: "meta-llama/llama-4-maverick",
|
|
8036
|
-
name: "Meta: Llama 4 Maverick",
|
|
8037
|
-
api: "openai-completions",
|
|
8038
|
-
provider: "openrouter",
|
|
8039
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
8040
|
-
reasoning: false,
|
|
8041
|
-
input: ["text", "image"],
|
|
8042
|
-
cost: {
|
|
8043
|
-
input: 0.15,
|
|
8044
|
-
output: 0.6,
|
|
8045
|
-
cacheRead: 0,
|
|
8046
|
-
cacheWrite: 0,
|
|
8047
|
-
},
|
|
8048
|
-
contextWindow: 1048576,
|
|
8049
|
-
maxTokens: 16384,
|
|
8050
|
-
},
|
|
8051
8466
|
"meta-llama/llama-4-scout": {
|
|
8052
8467
|
id: "meta-llama/llama-4-scout",
|
|
8053
8468
|
name: "Meta: Llama 4 Scout",
|
|
@@ -8125,9 +8540,9 @@ export const MODELS = {
|
|
|
8125
8540
|
reasoning: true,
|
|
8126
8541
|
input: ["text"],
|
|
8127
8542
|
cost: {
|
|
8128
|
-
input: 0.
|
|
8129
|
-
output:
|
|
8130
|
-
cacheRead: 0.
|
|
8543
|
+
input: 0.15,
|
|
8544
|
+
output: 1.2,
|
|
8545
|
+
cacheRead: 0.075,
|
|
8131
8546
|
cacheWrite: 0,
|
|
8132
8547
|
},
|
|
8133
8548
|
contextWindow: 196608,
|
|
@@ -8606,9 +9021,26 @@ export const MODELS = {
|
|
|
8606
9021
|
cacheRead: 0.07,
|
|
8607
9022
|
cacheWrite: 0,
|
|
8608
9023
|
},
|
|
8609
|
-
contextWindow:
|
|
9024
|
+
contextWindow: 262144,
|
|
8610
9025
|
maxTokens: 4096,
|
|
8611
9026
|
},
|
|
9027
|
+
"moonshotai/kimi-k2.6": {
|
|
9028
|
+
id: "moonshotai/kimi-k2.6",
|
|
9029
|
+
name: "MoonshotAI: Kimi K2.6",
|
|
9030
|
+
api: "openai-completions",
|
|
9031
|
+
provider: "openrouter",
|
|
9032
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
9033
|
+
reasoning: true,
|
|
9034
|
+
input: ["text", "image"],
|
|
9035
|
+
cost: {
|
|
9036
|
+
input: 0.7999999999999999,
|
|
9037
|
+
output: 3.5,
|
|
9038
|
+
cacheRead: 0.19999999999999998,
|
|
9039
|
+
cacheWrite: 0,
|
|
9040
|
+
},
|
|
9041
|
+
contextWindow: 262144,
|
|
9042
|
+
maxTokens: 262144,
|
|
9043
|
+
},
|
|
8612
9044
|
"nex-agi/deepseek-v3.1-nex-n1": {
|
|
8613
9045
|
id: "nex-agi/deepseek-v3.1-nex-n1",
|
|
8614
9046
|
name: "Nex AGI: DeepSeek V3.1 Nex N1",
|
|
@@ -9085,23 +9517,6 @@ export const MODELS = {
|
|
|
9085
9517
|
contextWindow: 128000,
|
|
9086
9518
|
maxTokens: 16384,
|
|
9087
9519
|
},
|
|
9088
|
-
"openai/gpt-4o:extended": {
|
|
9089
|
-
id: "openai/gpt-4o:extended",
|
|
9090
|
-
name: "OpenAI: GPT-4o (extended)",
|
|
9091
|
-
api: "openai-completions",
|
|
9092
|
-
provider: "openrouter",
|
|
9093
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
9094
|
-
reasoning: false,
|
|
9095
|
-
input: ["text", "image"],
|
|
9096
|
-
cost: {
|
|
9097
|
-
input: 6,
|
|
9098
|
-
output: 18,
|
|
9099
|
-
cacheRead: 0,
|
|
9100
|
-
cacheWrite: 0,
|
|
9101
|
-
},
|
|
9102
|
-
contextWindow: 128000,
|
|
9103
|
-
maxTokens: 64000,
|
|
9104
|
-
},
|
|
9105
9520
|
"openai/gpt-5": {
|
|
9106
9521
|
id: "openai/gpt-5",
|
|
9107
9522
|
name: "OpenAI: GPT-5",
|
|
@@ -9136,40 +9551,6 @@ export const MODELS = {
|
|
|
9136
9551
|
contextWindow: 400000,
|
|
9137
9552
|
maxTokens: 128000,
|
|
9138
9553
|
},
|
|
9139
|
-
"openai/gpt-5-image": {
|
|
9140
|
-
id: "openai/gpt-5-image",
|
|
9141
|
-
name: "OpenAI: GPT-5 Image",
|
|
9142
|
-
api: "openai-completions",
|
|
9143
|
-
provider: "openrouter",
|
|
9144
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
9145
|
-
reasoning: true,
|
|
9146
|
-
input: ["text", "image"],
|
|
9147
|
-
cost: {
|
|
9148
|
-
input: 10,
|
|
9149
|
-
output: 10,
|
|
9150
|
-
cacheRead: 1.25,
|
|
9151
|
-
cacheWrite: 0,
|
|
9152
|
-
},
|
|
9153
|
-
contextWindow: 400000,
|
|
9154
|
-
maxTokens: 128000,
|
|
9155
|
-
},
|
|
9156
|
-
"openai/gpt-5-image-mini": {
|
|
9157
|
-
id: "openai/gpt-5-image-mini",
|
|
9158
|
-
name: "OpenAI: GPT-5 Image Mini",
|
|
9159
|
-
api: "openai-completions",
|
|
9160
|
-
provider: "openrouter",
|
|
9161
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
9162
|
-
reasoning: true,
|
|
9163
|
-
input: ["text", "image"],
|
|
9164
|
-
cost: {
|
|
9165
|
-
input: 2.5,
|
|
9166
|
-
output: 2,
|
|
9167
|
-
cacheRead: 0.25,
|
|
9168
|
-
cacheWrite: 0,
|
|
9169
|
-
},
|
|
9170
|
-
contextWindow: 400000,
|
|
9171
|
-
maxTokens: 128000,
|
|
9172
|
-
},
|
|
9173
9554
|
"openai/gpt-5-mini": {
|
|
9174
9555
|
id: "openai/gpt-5-mini",
|
|
9175
9556
|
name: "OpenAI: GPT-5 Mini",
|
|
@@ -9765,23 +10146,6 @@ export const MODELS = {
|
|
|
9765
10146
|
contextWindow: 2000000,
|
|
9766
10147
|
maxTokens: 4096,
|
|
9767
10148
|
},
|
|
9768
|
-
"openrouter/elephant-alpha": {
|
|
9769
|
-
id: "openrouter/elephant-alpha",
|
|
9770
|
-
name: "Elephant",
|
|
9771
|
-
api: "openai-completions",
|
|
9772
|
-
provider: "openrouter",
|
|
9773
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
9774
|
-
reasoning: false,
|
|
9775
|
-
input: ["text"],
|
|
9776
|
-
cost: {
|
|
9777
|
-
input: 0,
|
|
9778
|
-
output: 0,
|
|
9779
|
-
cacheRead: 0,
|
|
9780
|
-
cacheWrite: 0,
|
|
9781
|
-
},
|
|
9782
|
-
contextWindow: 262144,
|
|
9783
|
-
maxTokens: 32768,
|
|
9784
|
-
},
|
|
9785
10149
|
"openrouter/free": {
|
|
9786
10150
|
id: "openrouter/free",
|
|
9787
10151
|
name: "Free Models Router",
|
|
@@ -9992,7 +10356,7 @@ export const MODELS = {
|
|
|
9992
10356
|
api: "openai-completions",
|
|
9993
10357
|
provider: "openrouter",
|
|
9994
10358
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
9995
|
-
reasoning:
|
|
10359
|
+
reasoning: false,
|
|
9996
10360
|
input: ["text"],
|
|
9997
10361
|
cost: {
|
|
9998
10362
|
input: 0.071,
|
|
@@ -10167,7 +10531,7 @@ export const MODELS = {
|
|
|
10167
10531
|
cost: {
|
|
10168
10532
|
input: 0.15,
|
|
10169
10533
|
output: 0.7999999999999999,
|
|
10170
|
-
cacheRead: 0.
|
|
10534
|
+
cacheRead: 0.11,
|
|
10171
10535
|
cacheWrite: 0,
|
|
10172
10536
|
},
|
|
10173
10537
|
contextWindow: 262144,
|
|
@@ -11049,13 +11413,13 @@ export const MODELS = {
|
|
|
11049
11413
|
reasoning: true,
|
|
11050
11414
|
input: ["text"],
|
|
11051
11415
|
cost: {
|
|
11052
|
-
input: 0.
|
|
11053
|
-
output: 1.
|
|
11054
|
-
cacheRead: 0
|
|
11416
|
+
input: 0.38,
|
|
11417
|
+
output: 1.74,
|
|
11418
|
+
cacheRead: 0,
|
|
11055
11419
|
cacheWrite: 0,
|
|
11056
11420
|
},
|
|
11057
11421
|
contextWindow: 202752,
|
|
11058
|
-
maxTokens:
|
|
11422
|
+
maxTokens: 4096,
|
|
11059
11423
|
},
|
|
11060
11424
|
"z-ai/glm-4.7-flash": {
|
|
11061
11425
|
id: "z-ai/glm-4.7-flash",
|
|
@@ -11088,8 +11452,8 @@ export const MODELS = {
|
|
|
11088
11452
|
cacheRead: 0.119,
|
|
11089
11453
|
cacheWrite: 0,
|
|
11090
11454
|
},
|
|
11091
|
-
contextWindow:
|
|
11092
|
-
maxTokens:
|
|
11455
|
+
contextWindow: 202752,
|
|
11456
|
+
maxTokens: 4096,
|
|
11093
11457
|
},
|
|
11094
11458
|
"z-ai/glm-5-turbo": {
|
|
11095
11459
|
id: "z-ai/glm-5-turbo",
|
|
@@ -11117,9 +11481,9 @@ export const MODELS = {
|
|
|
11117
11481
|
reasoning: true,
|
|
11118
11482
|
input: ["text"],
|
|
11119
11483
|
cost: {
|
|
11120
|
-
input:
|
|
11121
|
-
output: 3.
|
|
11122
|
-
cacheRead: 0.
|
|
11484
|
+
input: 1.0499999999999998,
|
|
11485
|
+
output: 3.5,
|
|
11486
|
+
cacheRead: 0.5249999999999999,
|
|
11123
11487
|
cacheWrite: 0,
|
|
11124
11488
|
},
|
|
11125
11489
|
contextWindow: 202752,
|
|
@@ -11142,6 +11506,23 @@ export const MODELS = {
|
|
|
11142
11506
|
contextWindow: 202752,
|
|
11143
11507
|
maxTokens: 131072,
|
|
11144
11508
|
},
|
|
11509
|
+
"~anthropic/claude-opus-latest": {
|
|
11510
|
+
id: "~anthropic/claude-opus-latest",
|
|
11511
|
+
name: "Anthropic: Claude Opus Latest",
|
|
11512
|
+
api: "openai-completions",
|
|
11513
|
+
provider: "openrouter",
|
|
11514
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
11515
|
+
reasoning: true,
|
|
11516
|
+
input: ["text", "image"],
|
|
11517
|
+
cost: {
|
|
11518
|
+
input: 5,
|
|
11519
|
+
output: 25,
|
|
11520
|
+
cacheRead: 0.5,
|
|
11521
|
+
cacheWrite: 6.25,
|
|
11522
|
+
},
|
|
11523
|
+
contextWindow: 1000000,
|
|
11524
|
+
maxTokens: 128000,
|
|
11525
|
+
},
|
|
11145
11526
|
},
|
|
11146
11527
|
"vercel-ai-gateway": {
|
|
11147
11528
|
"alibaba/qwen-3-14b": {
|
|
@@ -11212,6 +11593,23 @@ export const MODELS = {
|
|
|
11212
11593
|
contextWindow: 128000,
|
|
11213
11594
|
maxTokens: 8192,
|
|
11214
11595
|
},
|
|
11596
|
+
"alibaba/qwen-3.6-max-preview": {
|
|
11597
|
+
id: "alibaba/qwen-3.6-max-preview",
|
|
11598
|
+
name: "Qwen 3.6 Max Preview",
|
|
11599
|
+
api: "anthropic-messages",
|
|
11600
|
+
provider: "vercel-ai-gateway",
|
|
11601
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11602
|
+
reasoning: true,
|
|
11603
|
+
input: ["text", "image"],
|
|
11604
|
+
cost: {
|
|
11605
|
+
input: 1.3,
|
|
11606
|
+
output: 7.8,
|
|
11607
|
+
cacheRead: 0.26,
|
|
11608
|
+
cacheWrite: 1.625,
|
|
11609
|
+
},
|
|
11610
|
+
contextWindow: 240000,
|
|
11611
|
+
maxTokens: 64000,
|
|
11612
|
+
},
|
|
11215
11613
|
"alibaba/qwen3-235b-a22b-thinking": {
|
|
11216
11614
|
id: "alibaba/qwen3-235b-a22b-thinking",
|
|
11217
11615
|
name: "Qwen3 235B A22B Thinking 2507",
|
|
@@ -12555,6 +12953,23 @@ export const MODELS = {
|
|
|
12555
12953
|
contextWindow: 262114,
|
|
12556
12954
|
maxTokens: 262114,
|
|
12557
12955
|
},
|
|
12956
|
+
"moonshotai/kimi-k2.6": {
|
|
12957
|
+
id: "moonshotai/kimi-k2.6",
|
|
12958
|
+
name: "Kimi K2.6",
|
|
12959
|
+
api: "anthropic-messages",
|
|
12960
|
+
provider: "vercel-ai-gateway",
|
|
12961
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12962
|
+
reasoning: true,
|
|
12963
|
+
input: ["text", "image"],
|
|
12964
|
+
cost: {
|
|
12965
|
+
input: 0.95,
|
|
12966
|
+
output: 4,
|
|
12967
|
+
cacheRead: 0.16,
|
|
12968
|
+
cacheWrite: 0,
|
|
12969
|
+
},
|
|
12970
|
+
contextWindow: 262000,
|
|
12971
|
+
maxTokens: 262000,
|
|
12972
|
+
},
|
|
12558
12973
|
"nvidia/nemotron-nano-12b-v2-vl": {
|
|
12559
12974
|
id: "nvidia/nemotron-nano-12b-v2-vl",
|
|
12560
12975
|
name: "Nvidia Nemotron Nano 12B V2 VL",
|