@hyperspaceng/neural-ai 0.67.69 → 0.68.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (50) hide show
  1. package/README.md +5 -1
  2. package/dist/env-api-keys.d.ts.map +1 -1
  3. package/dist/env-api-keys.js +1 -0
  4. package/dist/env-api-keys.js.map +1 -1
  5. package/dist/models.generated.d.ts +618 -103
  6. package/dist/models.generated.d.ts.map +1 -1
  7. package/dist/models.generated.js +627 -114
  8. package/dist/models.generated.js.map +1 -1
  9. package/dist/providers/amazon-bedrock.d.ts.map +1 -1
  10. package/dist/providers/amazon-bedrock.js +55 -7
  11. package/dist/providers/amazon-bedrock.js.map +1 -1
  12. package/dist/providers/anthropic.d.ts.map +1 -1
  13. package/dist/providers/anthropic.js +141 -10
  14. package/dist/providers/anthropic.js.map +1 -1
  15. package/dist/providers/google-shared.d.ts.map +1 -1
  16. package/dist/providers/google-shared.js +30 -4
  17. package/dist/providers/google-shared.js.map +1 -1
  18. package/dist/providers/openai-completions.d.ts +5 -1
  19. package/dist/providers/openai-completions.d.ts.map +1 -1
  20. package/dist/providers/openai-completions.js +149 -60
  21. package/dist/providers/openai-completions.js.map +1 -1
  22. package/dist/providers/openai-responses-shared.d.ts.map +1 -1
  23. package/dist/providers/openai-responses-shared.js +2 -5
  24. package/dist/providers/openai-responses-shared.js.map +1 -1
  25. package/dist/providers/simple-options.d.ts.map +1 -1
  26. package/dist/providers/simple-options.js +1 -1
  27. package/dist/providers/simple-options.js.map +1 -1
  28. package/dist/providers/transform-messages.d.ts.map +1 -1
  29. package/dist/providers/transform-messages.js +41 -2
  30. package/dist/providers/transform-messages.js.map +1 -1
  31. package/dist/types.d.ts +5 -1
  32. package/dist/types.d.ts.map +1 -1
  33. package/dist/types.js.map +1 -1
  34. package/dist/utils/json-parse.d.ts +8 -1
  35. package/dist/utils/json-parse.d.ts.map +1 -1
  36. package/dist/utils/json-parse.js +89 -5
  37. package/dist/utils/json-parse.js.map +1 -1
  38. package/dist/utils/oauth/anthropic.d.ts.map +1 -1
  39. package/dist/utils/oauth/anthropic.js +1 -1
  40. package/dist/utils/oauth/anthropic.js.map +1 -1
  41. package/dist/utils/oauth/google-antigravity.d.ts.map +1 -1
  42. package/dist/utils/oauth/google-antigravity.js +2 -1
  43. package/dist/utils/oauth/google-antigravity.js.map +1 -1
  44. package/dist/utils/oauth/google-gemini-cli.d.ts.map +1 -1
  45. package/dist/utils/oauth/google-gemini-cli.js +2 -1
  46. package/dist/utils/oauth/google-gemini-cli.js.map +1 -1
  47. package/dist/utils/oauth/openai-codex.d.ts.map +1 -1
  48. package/dist/utils/oauth/openai-codex.js +3 -2
  49. package/dist/utils/oauth/openai-codex.js.map +1 -1
  50. package/package.json +1 -1
@@ -381,7 +381,7 @@ export const MODELS = {
381
381
  name: "Claude Haiku 4.5 (EU)",
382
382
  api: "bedrock-converse-stream",
383
383
  provider: "amazon-bedrock",
384
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
384
+ baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
385
385
  reasoning: true,
386
386
  input: ["text", "image"],
387
387
  cost: {
@@ -398,7 +398,7 @@ export const MODELS = {
398
398
  name: "Claude Opus 4.5 (EU)",
399
399
  api: "bedrock-converse-stream",
400
400
  provider: "amazon-bedrock",
401
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
401
+ baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
402
402
  reasoning: true,
403
403
  input: ["text", "image"],
404
404
  cost: {
@@ -415,7 +415,7 @@ export const MODELS = {
415
415
  name: "Claude Opus 4.6 (EU)",
416
416
  api: "bedrock-converse-stream",
417
417
  provider: "amazon-bedrock",
418
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
418
+ baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
419
419
  reasoning: true,
420
420
  input: ["text", "image"],
421
421
  cost: {
@@ -432,7 +432,7 @@ export const MODELS = {
432
432
  name: "Claude Opus 4.7 (EU)",
433
433
  api: "bedrock-converse-stream",
434
434
  provider: "amazon-bedrock",
435
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
435
+ baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
436
436
  reasoning: true,
437
437
  input: ["text", "image"],
438
438
  cost: {
@@ -449,7 +449,7 @@ export const MODELS = {
449
449
  name: "Claude Sonnet 4 (EU)",
450
450
  api: "bedrock-converse-stream",
451
451
  provider: "amazon-bedrock",
452
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
452
+ baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
453
453
  reasoning: true,
454
454
  input: ["text", "image"],
455
455
  cost: {
@@ -466,7 +466,7 @@ export const MODELS = {
466
466
  name: "Claude Sonnet 4.5 (EU)",
467
467
  api: "bedrock-converse-stream",
468
468
  provider: "amazon-bedrock",
469
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
469
+ baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
470
470
  reasoning: true,
471
471
  input: ["text", "image"],
472
472
  cost: {
@@ -483,7 +483,7 @@ export const MODELS = {
483
483
  name: "Claude Sonnet 4.6 (EU)",
484
484
  api: "bedrock-converse-stream",
485
485
  provider: "amazon-bedrock",
486
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
486
+ baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
487
487
  reasoning: true,
488
488
  input: ["text", "image"],
489
489
  cost: {
@@ -2695,6 +2695,314 @@ export const MODELS = {
2695
2695
  maxTokens: 40000,
2696
2696
  },
2697
2697
  },
2698
+ "fireworks": {
2699
+ "accounts/fireworks/models/deepseek-v3p1": {
2700
+ id: "accounts/fireworks/models/deepseek-v3p1",
2701
+ name: "DeepSeek V3.1",
2702
+ api: "anthropic-messages",
2703
+ provider: "fireworks",
2704
+ baseUrl: "https://api.fireworks.ai/inference",
2705
+ reasoning: true,
2706
+ input: ["text"],
2707
+ cost: {
2708
+ input: 0.56,
2709
+ output: 1.68,
2710
+ cacheRead: 0,
2711
+ cacheWrite: 0,
2712
+ },
2713
+ contextWindow: 163840,
2714
+ maxTokens: 163840,
2715
+ },
2716
+ "accounts/fireworks/models/deepseek-v3p2": {
2717
+ id: "accounts/fireworks/models/deepseek-v3p2",
2718
+ name: "DeepSeek V3.2",
2719
+ api: "anthropic-messages",
2720
+ provider: "fireworks",
2721
+ baseUrl: "https://api.fireworks.ai/inference",
2722
+ reasoning: true,
2723
+ input: ["text"],
2724
+ cost: {
2725
+ input: 0.56,
2726
+ output: 1.68,
2727
+ cacheRead: 0.28,
2728
+ cacheWrite: 0,
2729
+ },
2730
+ contextWindow: 160000,
2731
+ maxTokens: 160000,
2732
+ },
2733
+ "accounts/fireworks/models/glm-4p5": {
2734
+ id: "accounts/fireworks/models/glm-4p5",
2735
+ name: "GLM 4.5",
2736
+ api: "anthropic-messages",
2737
+ provider: "fireworks",
2738
+ baseUrl: "https://api.fireworks.ai/inference",
2739
+ reasoning: true,
2740
+ input: ["text"],
2741
+ cost: {
2742
+ input: 0.55,
2743
+ output: 2.19,
2744
+ cacheRead: 0,
2745
+ cacheWrite: 0,
2746
+ },
2747
+ contextWindow: 131072,
2748
+ maxTokens: 131072,
2749
+ },
2750
+ "accounts/fireworks/models/glm-4p5-air": {
2751
+ id: "accounts/fireworks/models/glm-4p5-air",
2752
+ name: "GLM 4.5 Air",
2753
+ api: "anthropic-messages",
2754
+ provider: "fireworks",
2755
+ baseUrl: "https://api.fireworks.ai/inference",
2756
+ reasoning: true,
2757
+ input: ["text"],
2758
+ cost: {
2759
+ input: 0.22,
2760
+ output: 0.88,
2761
+ cacheRead: 0,
2762
+ cacheWrite: 0,
2763
+ },
2764
+ contextWindow: 131072,
2765
+ maxTokens: 131072,
2766
+ },
2767
+ "accounts/fireworks/models/glm-4p7": {
2768
+ id: "accounts/fireworks/models/glm-4p7",
2769
+ name: "GLM 4.7",
2770
+ api: "anthropic-messages",
2771
+ provider: "fireworks",
2772
+ baseUrl: "https://api.fireworks.ai/inference",
2773
+ reasoning: true,
2774
+ input: ["text"],
2775
+ cost: {
2776
+ input: 0.6,
2777
+ output: 2.2,
2778
+ cacheRead: 0.3,
2779
+ cacheWrite: 0,
2780
+ },
2781
+ contextWindow: 198000,
2782
+ maxTokens: 198000,
2783
+ },
2784
+ "accounts/fireworks/models/glm-5": {
2785
+ id: "accounts/fireworks/models/glm-5",
2786
+ name: "GLM 5",
2787
+ api: "anthropic-messages",
2788
+ provider: "fireworks",
2789
+ baseUrl: "https://api.fireworks.ai/inference",
2790
+ reasoning: true,
2791
+ input: ["text"],
2792
+ cost: {
2793
+ input: 1,
2794
+ output: 3.2,
2795
+ cacheRead: 0.5,
2796
+ cacheWrite: 0,
2797
+ },
2798
+ contextWindow: 202752,
2799
+ maxTokens: 131072,
2800
+ },
2801
+ "accounts/fireworks/models/glm-5p1": {
2802
+ id: "accounts/fireworks/models/glm-5p1",
2803
+ name: "GLM 5.1",
2804
+ api: "anthropic-messages",
2805
+ provider: "fireworks",
2806
+ baseUrl: "https://api.fireworks.ai/inference",
2807
+ reasoning: true,
2808
+ input: ["text"],
2809
+ cost: {
2810
+ input: 1.4,
2811
+ output: 4.4,
2812
+ cacheRead: 0.26,
2813
+ cacheWrite: 0,
2814
+ },
2815
+ contextWindow: 202800,
2816
+ maxTokens: 131072,
2817
+ },
2818
+ "accounts/fireworks/models/gpt-oss-120b": {
2819
+ id: "accounts/fireworks/models/gpt-oss-120b",
2820
+ name: "GPT OSS 120B",
2821
+ api: "anthropic-messages",
2822
+ provider: "fireworks",
2823
+ baseUrl: "https://api.fireworks.ai/inference",
2824
+ reasoning: true,
2825
+ input: ["text"],
2826
+ cost: {
2827
+ input: 0.15,
2828
+ output: 0.6,
2829
+ cacheRead: 0,
2830
+ cacheWrite: 0,
2831
+ },
2832
+ contextWindow: 131072,
2833
+ maxTokens: 32768,
2834
+ },
2835
+ "accounts/fireworks/models/gpt-oss-20b": {
2836
+ id: "accounts/fireworks/models/gpt-oss-20b",
2837
+ name: "GPT OSS 20B",
2838
+ api: "anthropic-messages",
2839
+ provider: "fireworks",
2840
+ baseUrl: "https://api.fireworks.ai/inference",
2841
+ reasoning: true,
2842
+ input: ["text"],
2843
+ cost: {
2844
+ input: 0.05,
2845
+ output: 0.2,
2846
+ cacheRead: 0,
2847
+ cacheWrite: 0,
2848
+ },
2849
+ contextWindow: 131072,
2850
+ maxTokens: 32768,
2851
+ },
2852
+ "accounts/fireworks/models/kimi-k2-instruct": {
2853
+ id: "accounts/fireworks/models/kimi-k2-instruct",
2854
+ name: "Kimi K2 Instruct",
2855
+ api: "anthropic-messages",
2856
+ provider: "fireworks",
2857
+ baseUrl: "https://api.fireworks.ai/inference",
2858
+ reasoning: false,
2859
+ input: ["text"],
2860
+ cost: {
2861
+ input: 1,
2862
+ output: 3,
2863
+ cacheRead: 0,
2864
+ cacheWrite: 0,
2865
+ },
2866
+ contextWindow: 128000,
2867
+ maxTokens: 16384,
2868
+ },
2869
+ "accounts/fireworks/models/kimi-k2-thinking": {
2870
+ id: "accounts/fireworks/models/kimi-k2-thinking",
2871
+ name: "Kimi K2 Thinking",
2872
+ api: "anthropic-messages",
2873
+ provider: "fireworks",
2874
+ baseUrl: "https://api.fireworks.ai/inference",
2875
+ reasoning: true,
2876
+ input: ["text"],
2877
+ cost: {
2878
+ input: 0.6,
2879
+ output: 2.5,
2880
+ cacheRead: 0.3,
2881
+ cacheWrite: 0,
2882
+ },
2883
+ contextWindow: 256000,
2884
+ maxTokens: 256000,
2885
+ },
2886
+ "accounts/fireworks/models/kimi-k2p5": {
2887
+ id: "accounts/fireworks/models/kimi-k2p5",
2888
+ name: "Kimi K2.5",
2889
+ api: "anthropic-messages",
2890
+ provider: "fireworks",
2891
+ baseUrl: "https://api.fireworks.ai/inference",
2892
+ reasoning: true,
2893
+ input: ["text", "image"],
2894
+ cost: {
2895
+ input: 0.6,
2896
+ output: 3,
2897
+ cacheRead: 0.1,
2898
+ cacheWrite: 0,
2899
+ },
2900
+ contextWindow: 256000,
2901
+ maxTokens: 256000,
2902
+ },
2903
+ "accounts/fireworks/models/kimi-k2p6": {
2904
+ id: "accounts/fireworks/models/kimi-k2p6",
2905
+ name: "Kimi K2.6",
2906
+ api: "anthropic-messages",
2907
+ provider: "fireworks",
2908
+ baseUrl: "https://api.fireworks.ai/inference",
2909
+ reasoning: true,
2910
+ input: ["text", "image"],
2911
+ cost: {
2912
+ input: 0.95,
2913
+ output: 4,
2914
+ cacheRead: 0.16,
2915
+ cacheWrite: 0,
2916
+ },
2917
+ contextWindow: 262000,
2918
+ maxTokens: 262000,
2919
+ },
2920
+ "accounts/fireworks/models/minimax-m2p1": {
2921
+ id: "accounts/fireworks/models/minimax-m2p1",
2922
+ name: "MiniMax-M2.1",
2923
+ api: "anthropic-messages",
2924
+ provider: "fireworks",
2925
+ baseUrl: "https://api.fireworks.ai/inference",
2926
+ reasoning: true,
2927
+ input: ["text"],
2928
+ cost: {
2929
+ input: 0.3,
2930
+ output: 1.2,
2931
+ cacheRead: 0.03,
2932
+ cacheWrite: 0,
2933
+ },
2934
+ contextWindow: 200000,
2935
+ maxTokens: 200000,
2936
+ },
2937
+ "accounts/fireworks/models/minimax-m2p5": {
2938
+ id: "accounts/fireworks/models/minimax-m2p5",
2939
+ name: "MiniMax-M2.5",
2940
+ api: "anthropic-messages",
2941
+ provider: "fireworks",
2942
+ baseUrl: "https://api.fireworks.ai/inference",
2943
+ reasoning: true,
2944
+ input: ["text"],
2945
+ cost: {
2946
+ input: 0.3,
2947
+ output: 1.2,
2948
+ cacheRead: 0.03,
2949
+ cacheWrite: 0,
2950
+ },
2951
+ contextWindow: 196608,
2952
+ maxTokens: 196608,
2953
+ },
2954
+ "accounts/fireworks/models/minimax-m2p7": {
2955
+ id: "accounts/fireworks/models/minimax-m2p7",
2956
+ name: "MiniMax-M2.7",
2957
+ api: "anthropic-messages",
2958
+ provider: "fireworks",
2959
+ baseUrl: "https://api.fireworks.ai/inference",
2960
+ reasoning: true,
2961
+ input: ["text"],
2962
+ cost: {
2963
+ input: 0.3,
2964
+ output: 1.2,
2965
+ cacheRead: 0.03,
2966
+ cacheWrite: 0,
2967
+ },
2968
+ contextWindow: 196608,
2969
+ maxTokens: 196608,
2970
+ },
2971
+ "accounts/fireworks/models/qwen3p6-plus": {
2972
+ id: "accounts/fireworks/models/qwen3p6-plus",
2973
+ name: "Qwen 3.6 Plus",
2974
+ api: "anthropic-messages",
2975
+ provider: "fireworks",
2976
+ baseUrl: "https://api.fireworks.ai/inference",
2977
+ reasoning: true,
2978
+ input: ["text", "image"],
2979
+ cost: {
2980
+ input: 0.5,
2981
+ output: 3,
2982
+ cacheRead: 0.1,
2983
+ cacheWrite: 0,
2984
+ },
2985
+ contextWindow: 128000,
2986
+ maxTokens: 8192,
2987
+ },
2988
+ "accounts/fireworks/routers/kimi-k2p5-turbo": {
2989
+ id: "accounts/fireworks/routers/kimi-k2p5-turbo",
2990
+ name: "Kimi K2.5 Turbo (firepass)",
2991
+ api: "anthropic-messages",
2992
+ provider: "fireworks",
2993
+ baseUrl: "https://api.fireworks.ai/inference",
2994
+ reasoning: true,
2995
+ input: ["text", "image"],
2996
+ cost: {
2997
+ input: 0,
2998
+ output: 0,
2999
+ cacheRead: 0,
3000
+ cacheWrite: 0,
3001
+ },
3002
+ contextWindow: 256000,
3003
+ maxTokens: 256000,
3004
+ },
3005
+ },
2698
3006
  "github-copilot": {
2699
3007
  "claude-haiku-4.5": {
2700
3008
  id: "claude-haiku-4.5",
@@ -3251,7 +3559,7 @@ export const MODELS = {
3251
3559
  cost: {
3252
3560
  input: 0.3,
3253
3561
  output: 2.5,
3254
- cacheRead: 0.075,
3562
+ cacheRead: 0.03,
3255
3563
  cacheWrite: 0,
3256
3564
  },
3257
3565
  contextWindow: 1048576,
@@ -3370,7 +3678,7 @@ export const MODELS = {
3370
3678
  cost: {
3371
3679
  input: 1.25,
3372
3680
  output: 10,
3373
- cacheRead: 0.31,
3681
+ cacheRead: 0.125,
3374
3682
  cacheWrite: 0,
3375
3683
  },
3376
3684
  contextWindow: 1048576,
@@ -3580,8 +3888,8 @@ export const MODELS = {
3580
3888
  contextWindow: 131072,
3581
3889
  maxTokens: 8192,
3582
3890
  },
3583
- "gemma-4-26b-it": {
3584
- id: "gemma-4-26b-it",
3891
+ "gemma-4-26b-a4b-it": {
3892
+ id: "gemma-4-26b-a4b-it",
3585
3893
  name: "Gemma 4 26B",
3586
3894
  api: "google-generative-ai",
3587
3895
  provider: "google",
@@ -4694,6 +5002,24 @@ export const MODELS = {
4694
5002
  contextWindow: 262144,
4695
5003
  maxTokens: 262144,
4696
5004
  },
5005
+ "moonshotai/Kimi-K2.6": {
5006
+ id: "moonshotai/Kimi-K2.6",
5007
+ name: "Kimi-K2.6",
5008
+ api: "openai-completions",
5009
+ provider: "huggingface",
5010
+ baseUrl: "https://router.huggingface.co/v1",
5011
+ compat: { "supportsDeveloperRole": false },
5012
+ reasoning: true,
5013
+ input: ["text", "image"],
5014
+ cost: {
5015
+ input: 0.95,
5016
+ output: 4,
5017
+ cacheRead: 0.16,
5018
+ cacheWrite: 0,
5019
+ },
5020
+ contextWindow: 262144,
5021
+ maxTokens: 262144,
5022
+ },
4697
5023
  "zai-org/GLM-4.7": {
4698
5024
  id: "zai-org/GLM-4.7",
4699
5025
  name: "GLM-4.7",
@@ -4768,6 +5094,23 @@ export const MODELS = {
4768
5094
  },
4769
5095
  },
4770
5096
  "kimi-coding": {
5097
+ "k2p6": {
5098
+ id: "k2p6",
5099
+ name: "Kimi K2.6",
5100
+ api: "anthropic-messages",
5101
+ provider: "kimi-coding",
5102
+ baseUrl: "https://api.kimi.com/coding",
5103
+ reasoning: true,
5104
+ input: ["text", "image"],
5105
+ cost: {
5106
+ input: 0,
5107
+ output: 0,
5108
+ cacheRead: 0,
5109
+ cacheWrite: 0,
5110
+ },
5111
+ contextWindow: 262144,
5112
+ maxTokens: 32768,
5113
+ },
4771
5114
  "kimi-for-coding": {
4772
5115
  id: "kimi-for-coding",
4773
5116
  name: "Kimi For Coding",
@@ -6650,6 +6993,40 @@ export const MODELS = {
6650
6993
  contextWindow: 262144,
6651
6994
  maxTokens: 65536,
6652
6995
  },
6996
+ "kimi-k2.6": {
6997
+ id: "kimi-k2.6",
6998
+ name: "Kimi K2.6",
6999
+ api: "openai-completions",
7000
+ provider: "opencode",
7001
+ baseUrl: "https://opencode.ai/zen/v1",
7002
+ reasoning: true,
7003
+ input: ["text", "image"],
7004
+ cost: {
7005
+ input: 0.95,
7006
+ output: 4,
7007
+ cacheRead: 0.16,
7008
+ cacheWrite: 0,
7009
+ },
7010
+ contextWindow: 262144,
7011
+ maxTokens: 65536,
7012
+ },
7013
+ "ling-2.6-flash-free": {
7014
+ id: "ling-2.6-flash-free",
7015
+ name: "Ling 2.6 Flash Free",
7016
+ api: "openai-completions",
7017
+ provider: "opencode",
7018
+ baseUrl: "https://opencode.ai/zen/v1",
7019
+ reasoning: false,
7020
+ input: ["text"],
7021
+ cost: {
7022
+ input: 0,
7023
+ output: 0,
7024
+ cacheRead: 0,
7025
+ cacheWrite: 0,
7026
+ },
7027
+ contextWindow: 262100,
7028
+ maxTokens: 32800,
7029
+ },
6653
7030
  "minimax-m2.5": {
6654
7031
  id: "minimax-m2.5",
6655
7032
  name: "MiniMax M2.5",
@@ -6684,6 +7061,23 @@ export const MODELS = {
6684
7061
  contextWindow: 204800,
6685
7062
  maxTokens: 131072,
6686
7063
  },
7064
+ "minimax-m2.7": {
7065
+ id: "minimax-m2.7",
7066
+ name: "MiniMax M2.7",
7067
+ api: "openai-completions",
7068
+ provider: "opencode",
7069
+ baseUrl: "https://opencode.ai/zen/v1",
7070
+ reasoning: true,
7071
+ input: ["text"],
7072
+ cost: {
7073
+ input: 0.3,
7074
+ output: 1.2,
7075
+ cacheRead: 0.06,
7076
+ cacheWrite: 0,
7077
+ },
7078
+ contextWindow: 204800,
7079
+ maxTokens: 131072,
7080
+ },
6687
7081
  "nemotron-3-super-free": {
6688
7082
  id: "nemotron-3-super-free",
6689
7083
  name: "Nemotron 3 Super Free",
@@ -6704,9 +7098,9 @@ export const MODELS = {
6704
7098
  "qwen3.5-plus": {
6705
7099
  id: "qwen3.5-plus",
6706
7100
  name: "Qwen3.5 Plus",
6707
- api: "openai-completions",
7101
+ api: "anthropic-messages",
6708
7102
  provider: "opencode",
6709
- baseUrl: "https://opencode.ai/zen/v1",
7103
+ baseUrl: "https://opencode.ai/zen",
6710
7104
  reasoning: true,
6711
7105
  input: ["text", "image"],
6712
7106
  cost: {
@@ -6721,9 +7115,9 @@ export const MODELS = {
6721
7115
  "qwen3.6-plus": {
6722
7116
  id: "qwen3.6-plus",
6723
7117
  name: "Qwen3.6 Plus",
6724
- api: "openai-completions",
7118
+ api: "anthropic-messages",
6725
7119
  provider: "opencode",
6726
- baseUrl: "https://opencode.ai/zen/v1",
7120
+ baseUrl: "https://opencode.ai/zen",
6727
7121
  reasoning: true,
6728
7122
  input: ["text", "image"],
6729
7123
  cost: {
@@ -6788,6 +7182,23 @@ export const MODELS = {
6788
7182
  contextWindow: 262144,
6789
7183
  maxTokens: 65536,
6790
7184
  },
7185
+ "kimi-k2.6": {
7186
+ id: "kimi-k2.6",
7187
+ name: "Kimi K2.6 (3x limits)",
7188
+ api: "openai-completions",
7189
+ provider: "opencode-go",
7190
+ baseUrl: "https://opencode.ai/zen/go/v1",
7191
+ reasoning: true,
7192
+ input: ["text", "image"],
7193
+ cost: {
7194
+ input: 0.32,
7195
+ output: 1.34,
7196
+ cacheRead: 0.054,
7197
+ cacheWrite: 0,
7198
+ },
7199
+ contextWindow: 262144,
7200
+ maxTokens: 65536,
7201
+ },
6791
7202
  "mimo-v2-omni": {
6792
7203
  id: "mimo-v2-omni",
6793
7204
  name: "MiMo V2 Omni",
@@ -6803,7 +7214,7 @@ export const MODELS = {
6803
7214
  cacheWrite: 0,
6804
7215
  },
6805
7216
  contextWindow: 262144,
6806
- maxTokens: 64000,
7217
+ maxTokens: 128000,
6807
7218
  },
6808
7219
  "mimo-v2-pro": {
6809
7220
  id: "mimo-v2-pro",
@@ -6820,7 +7231,41 @@ export const MODELS = {
6820
7231
  cacheWrite: 0,
6821
7232
  },
6822
7233
  contextWindow: 1048576,
6823
- maxTokens: 64000,
7234
+ maxTokens: 128000,
7235
+ },
7236
+ "mimo-v2.5": {
7237
+ id: "mimo-v2.5",
7238
+ name: "MiMo V2.5",
7239
+ api: "openai-completions",
7240
+ provider: "opencode-go",
7241
+ baseUrl: "https://opencode.ai/zen/go/v1",
7242
+ reasoning: true,
7243
+ input: ["text", "image"],
7244
+ cost: {
7245
+ input: 0.4,
7246
+ output: 2,
7247
+ cacheRead: 0.08,
7248
+ cacheWrite: 0,
7249
+ },
7250
+ contextWindow: 262144,
7251
+ maxTokens: 128000,
7252
+ },
7253
+ "mimo-v2.5-pro": {
7254
+ id: "mimo-v2.5-pro",
7255
+ name: "MiMo V2.5 Pro",
7256
+ api: "openai-completions",
7257
+ provider: "opencode-go",
7258
+ baseUrl: "https://opencode.ai/zen/go/v1",
7259
+ reasoning: true,
7260
+ input: ["text"],
7261
+ cost: {
7262
+ input: 1,
7263
+ output: 3,
7264
+ cacheRead: 0.2,
7265
+ cacheWrite: 0,
7266
+ },
7267
+ contextWindow: 1048576,
7268
+ maxTokens: 128000,
6824
7269
  },
6825
7270
  "minimax-m2.5": {
6826
7271
  id: "minimax-m2.5",
@@ -6859,9 +7304,9 @@ export const MODELS = {
6859
7304
  "qwen3.5-plus": {
6860
7305
  id: "qwen3.5-plus",
6861
7306
  name: "Qwen3.5 Plus",
6862
- api: "openai-completions",
7307
+ api: "anthropic-messages",
6863
7308
  provider: "opencode-go",
6864
- baseUrl: "https://opencode.ai/zen/go/v1",
7309
+ baseUrl: "https://opencode.ai/zen/go",
6865
7310
  reasoning: true,
6866
7311
  input: ["text", "image"],
6867
7312
  cost: {
@@ -6876,9 +7321,9 @@ export const MODELS = {
6876
7321
  "qwen3.6-plus": {
6877
7322
  id: "qwen3.6-plus",
6878
7323
  name: "Qwen3.6 Plus",
6879
- api: "openai-completions",
7324
+ api: "anthropic-messages",
6880
7325
  provider: "opencode-go",
6881
- baseUrl: "https://opencode.ai/zen/go/v1",
7326
+ baseUrl: "https://opencode.ai/zen/go",
6882
7327
  reasoning: true,
6883
7328
  input: ["text", "image"],
6884
7329
  cost: {
@@ -7266,17 +7711,17 @@ export const MODELS = {
7266
7711
  contextWindow: 1000000,
7267
7712
  maxTokens: 128000,
7268
7713
  },
7269
- "arcee-ai/trinity-large-preview:free": {
7270
- id: "arcee-ai/trinity-large-preview:free",
7271
- name: "Arcee AI: Trinity Large Preview (free)",
7714
+ "arcee-ai/trinity-large-preview": {
7715
+ id: "arcee-ai/trinity-large-preview",
7716
+ name: "Arcee AI: Trinity Large Preview",
7272
7717
  api: "openai-completions",
7273
7718
  provider: "openrouter",
7274
7719
  baseUrl: "https://openrouter.ai/api/v1",
7275
7720
  reasoning: false,
7276
7721
  input: ["text"],
7277
7722
  cost: {
7278
- input: 0,
7279
- output: 0,
7723
+ input: 0.15,
7724
+ output: 0.44999999999999996,
7280
7725
  cacheRead: 0,
7281
7726
  cacheWrite: 0,
7282
7727
  },
@@ -7598,13 +8043,13 @@ export const MODELS = {
7598
8043
  reasoning: true,
7599
8044
  input: ["text"],
7600
8045
  cost: {
7601
- input: 0.25899999999999995,
7602
- output: 0.42,
7603
- cacheRead: 0.135,
8046
+ input: 0.252,
8047
+ output: 0.378,
8048
+ cacheRead: 0.0252,
7604
8049
  cacheWrite: 0,
7605
8050
  },
7606
- contextWindow: 163840,
7607
- maxTokens: 163840,
8051
+ contextWindow: 131072,
8052
+ maxTokens: 65536,
7608
8053
  },
7609
8054
  "deepseek/deepseek-v3.2-exp": {
7610
8055
  id: "deepseek/deepseek-v3.2-exp",
@@ -7654,7 +8099,7 @@ export const MODELS = {
7654
8099
  cacheRead: 0.024999999999999998,
7655
8100
  cacheWrite: 0.08333333333333334,
7656
8101
  },
7657
- contextWindow: 1048576,
8102
+ contextWindow: 1000000,
7658
8103
  maxTokens: 8192,
7659
8104
  },
7660
8105
  "google/gemini-2.0-flash-lite-001": {
@@ -7853,9 +8298,9 @@ export const MODELS = {
7853
8298
  reasoning: true,
7854
8299
  input: ["text", "image"],
7855
8300
  cost: {
7856
- input: 0.08,
7857
- output: 0.35,
7858
- cacheRead: 0.01,
8301
+ input: 0.07,
8302
+ output: 0.33999999999999997,
8303
+ cacheRead: 0,
7859
8304
  cacheWrite: 0,
7860
8305
  },
7861
8306
  contextWindow: 262144,
@@ -7889,7 +8334,7 @@ export const MODELS = {
7889
8334
  cost: {
7890
8335
  input: 0.13,
7891
8336
  output: 0.38,
7892
- cacheRead: 0.019999999499999997,
8337
+ cacheRead: 0,
7893
8338
  cacheWrite: 0,
7894
8339
  },
7895
8340
  contextWindow: 262144,
@@ -7929,6 +8374,23 @@ export const MODELS = {
7929
8374
  contextWindow: 128000,
7930
8375
  maxTokens: 50000,
7931
8376
  },
8377
+ "inclusionai/ling-2.6-flash:free": {
8378
+ id: "inclusionai/ling-2.6-flash:free",
8379
+ name: "inclusionAI: Ling-2.6-flash (free)",
8380
+ api: "openai-completions",
8381
+ provider: "openrouter",
8382
+ baseUrl: "https://openrouter.ai/api/v1",
8383
+ reasoning: false,
8384
+ input: ["text"],
8385
+ cost: {
8386
+ input: 0,
8387
+ output: 0,
8388
+ cacheRead: 0,
8389
+ cacheWrite: 0,
8390
+ },
8391
+ contextWindow: 262144,
8392
+ maxTokens: 32768,
8393
+ },
7932
8394
  "kwaipilot/kat-coder-pro-v2": {
7933
8395
  id: "kwaipilot/kat-coder-pro-v2",
7934
8396
  name: "Kwaipilot: KAT-Coder-Pro V2",
@@ -8006,13 +8468,13 @@ export const MODELS = {
8006
8468
  reasoning: false,
8007
8469
  input: ["text"],
8008
8470
  cost: {
8009
- input: 0.12,
8010
- output: 0.38,
8471
+ input: 0.09999999999999999,
8472
+ output: 0.32,
8011
8473
  cacheRead: 0,
8012
8474
  cacheWrite: 0,
8013
8475
  },
8014
8476
  contextWindow: 131072,
8015
- maxTokens: 131072,
8477
+ maxTokens: 16384,
8016
8478
  },
8017
8479
  "meta-llama/llama-3.3-70b-instruct:free": {
8018
8480
  id: "meta-llama/llama-3.3-70b-instruct:free",
@@ -8108,9 +8570,9 @@ export const MODELS = {
8108
8570
  reasoning: true,
8109
8571
  input: ["text"],
8110
8572
  cost: {
8111
- input: 0.118,
8112
- output: 0.9900000000000001,
8113
- cacheRead: 0.059,
8573
+ input: 0.15,
8574
+ output: 1.2,
8575
+ cacheRead: 0.075,
8114
8576
  cacheWrite: 0,
8115
8577
  },
8116
8578
  contextWindow: 196608,
@@ -8592,6 +9054,23 @@ export const MODELS = {
8592
9054
  contextWindow: 262144,
8593
9055
  maxTokens: 4096,
8594
9056
  },
9057
+ "moonshotai/kimi-k2.6": {
9058
+ id: "moonshotai/kimi-k2.6",
9059
+ name: "MoonshotAI: Kimi K2.6",
9060
+ api: "openai-completions",
9061
+ provider: "openrouter",
9062
+ baseUrl: "https://openrouter.ai/api/v1",
9063
+ reasoning: true,
9064
+ input: ["text", "image"],
9065
+ cost: {
9066
+ input: 0.75,
9067
+ output: 3.5,
9068
+ cacheRead: 0.15,
9069
+ cacheWrite: 0,
9070
+ },
9071
+ contextWindow: 262144,
9072
+ maxTokens: 4096,
9073
+ },
8595
9074
  "nex-agi/deepseek-v3.1-nex-n1": {
8596
9075
  id: "nex-agi/deepseek-v3.1-nex-n1",
8597
9076
  name: "Nex AGI: DeepSeek V3.1 Nex N1",
@@ -9102,40 +9581,6 @@ export const MODELS = {
9102
9581
  contextWindow: 400000,
9103
9582
  maxTokens: 128000,
9104
9583
  },
9105
- "openai/gpt-5-image": {
9106
- id: "openai/gpt-5-image",
9107
- name: "OpenAI: GPT-5 Image",
9108
- api: "openai-completions",
9109
- provider: "openrouter",
9110
- baseUrl: "https://openrouter.ai/api/v1",
9111
- reasoning: true,
9112
- input: ["text", "image"],
9113
- cost: {
9114
- input: 10,
9115
- output: 10,
9116
- cacheRead: 1.25,
9117
- cacheWrite: 0,
9118
- },
9119
- contextWindow: 400000,
9120
- maxTokens: 128000,
9121
- },
9122
- "openai/gpt-5-image-mini": {
9123
- id: "openai/gpt-5-image-mini",
9124
- name: "OpenAI: GPT-5 Image Mini",
9125
- api: "openai-completions",
9126
- provider: "openrouter",
9127
- baseUrl: "https://openrouter.ai/api/v1",
9128
- reasoning: true,
9129
- input: ["text", "image"],
9130
- cost: {
9131
- input: 2.5,
9132
- output: 2,
9133
- cacheRead: 0.25,
9134
- cacheWrite: 0,
9135
- },
9136
- contextWindow: 400000,
9137
- maxTokens: 128000,
9138
- },
9139
9584
  "openai/gpt-5-mini": {
9140
9585
  id: "openai/gpt-5-mini",
9141
9586
  name: "OpenAI: GPT-5 Mini",
@@ -9731,23 +10176,6 @@ export const MODELS = {
9731
10176
  contextWindow: 2000000,
9732
10177
  maxTokens: 4096,
9733
10178
  },
9734
- "openrouter/elephant-alpha": {
9735
- id: "openrouter/elephant-alpha",
9736
- name: "Elephant",
9737
- api: "openai-completions",
9738
- provider: "openrouter",
9739
- baseUrl: "https://openrouter.ai/api/v1",
9740
- reasoning: false,
9741
- input: ["text"],
9742
- cost: {
9743
- input: 0,
9744
- output: 0,
9745
- cacheRead: 0,
9746
- cacheWrite: 0,
9747
- },
9748
- contextWindow: 262144,
9749
- maxTokens: 32768,
9750
- },
9751
10179
  "openrouter/free": {
9752
10180
  id: "openrouter/free",
9753
10181
  name: "Free Models Router",
@@ -9978,13 +10406,13 @@ export const MODELS = {
9978
10406
  reasoning: true,
9979
10407
  input: ["text"],
9980
10408
  cost: {
9981
- input: 0.13,
9982
- output: 0.6,
10409
+ input: 0.14950000000000002,
10410
+ output: 1.495,
9983
10411
  cacheRead: 0,
9984
10412
  cacheWrite: 0,
9985
10413
  },
9986
- contextWindow: 262144,
9987
- maxTokens: 262144,
10414
+ contextWindow: 131072,
10415
+ maxTokens: 4096,
9988
10416
  },
9989
10417
  "qwen/qwen3-30b-a3b": {
9990
10418
  id: "qwen/qwen3-30b-a3b",
@@ -10133,7 +10561,7 @@ export const MODELS = {
10133
10561
  cost: {
10134
10562
  input: 0.15,
10135
10563
  output: 0.7999999999999999,
10136
- cacheRead: 0.12,
10564
+ cacheRead: 0.11,
10137
10565
  cacheWrite: 0,
10138
10566
  },
10139
10567
  contextWindow: 262144,
@@ -10887,6 +11315,40 @@ export const MODELS = {
10887
11315
  contextWindow: 1048576,
10888
11316
  maxTokens: 131072,
10889
11317
  },
11318
+ "xiaomi/mimo-v2.5": {
11319
+ id: "xiaomi/mimo-v2.5",
11320
+ name: "Xiaomi: MiMo-V2.5",
11321
+ api: "openai-completions",
11322
+ provider: "openrouter",
11323
+ baseUrl: "https://openrouter.ai/api/v1",
11324
+ reasoning: true,
11325
+ input: ["text", "image"],
11326
+ cost: {
11327
+ input: 0.39999999999999997,
11328
+ output: 2,
11329
+ cacheRead: 0.08,
11330
+ cacheWrite: 0,
11331
+ },
11332
+ contextWindow: 1048576,
11333
+ maxTokens: 131072,
11334
+ },
11335
+ "xiaomi/mimo-v2.5-pro": {
11336
+ id: "xiaomi/mimo-v2.5-pro",
11337
+ name: "Xiaomi: MiMo-V2.5-Pro",
11338
+ api: "openai-completions",
11339
+ provider: "openrouter",
11340
+ baseUrl: "https://openrouter.ai/api/v1",
11341
+ reasoning: true,
11342
+ input: ["text"],
11343
+ cost: {
11344
+ input: 1,
11345
+ output: 3,
11346
+ cacheRead: 0.19999999999999998,
11347
+ cacheWrite: 0,
11348
+ },
11349
+ contextWindow: 1048576,
11350
+ maxTokens: 131072,
11351
+ },
10890
11352
  "z-ai/glm-4-32b": {
10891
11353
  id: "z-ai/glm-4-32b",
10892
11354
  name: "Z.ai: GLM 4 32B ",
@@ -11015,13 +11477,13 @@ export const MODELS = {
11015
11477
  reasoning: true,
11016
11478
  input: ["text"],
11017
11479
  cost: {
11018
- input: 0.39,
11019
- output: 1.75,
11020
- cacheRead: 0.195,
11480
+ input: 0.38,
11481
+ output: 1.74,
11482
+ cacheRead: 0,
11021
11483
  cacheWrite: 0,
11022
11484
  },
11023
11485
  contextWindow: 202752,
11024
- maxTokens: 65535,
11486
+ maxTokens: 4096,
11025
11487
  },
11026
11488
  "z-ai/glm-4.7-flash": {
11027
11489
  id: "z-ai/glm-4.7-flash",
@@ -11054,8 +11516,8 @@ export const MODELS = {
11054
11516
  cacheRead: 0.119,
11055
11517
  cacheWrite: 0,
11056
11518
  },
11057
- contextWindow: 80000,
11058
- maxTokens: 131072,
11519
+ contextWindow: 202752,
11520
+ maxTokens: 4096,
11059
11521
  },
11060
11522
  "z-ai/glm-5-turbo": {
11061
11523
  id: "z-ai/glm-5-turbo",
@@ -11083,9 +11545,9 @@ export const MODELS = {
11083
11545
  reasoning: true,
11084
11546
  input: ["text"],
11085
11547
  cost: {
11086
- input: 0.95,
11087
- output: 3.15,
11088
- cacheRead: 0.475,
11548
+ input: 1.0499999999999998,
11549
+ output: 3.5,
11550
+ cacheRead: 0.5249999999999999,
11089
11551
  cacheWrite: 0,
11090
11552
  },
11091
11553
  contextWindow: 202752,
@@ -11108,6 +11570,23 @@ export const MODELS = {
11108
11570
  contextWindow: 202752,
11109
11571
  maxTokens: 131072,
11110
11572
  },
11573
+ "~anthropic/claude-opus-latest": {
11574
+ id: "~anthropic/claude-opus-latest",
11575
+ name: "Anthropic: Claude Opus Latest",
11576
+ api: "openai-completions",
11577
+ provider: "openrouter",
11578
+ baseUrl: "https://openrouter.ai/api/v1",
11579
+ reasoning: true,
11580
+ input: ["text", "image"],
11581
+ cost: {
11582
+ input: 5,
11583
+ output: 25,
11584
+ cacheRead: 0.5,
11585
+ cacheWrite: 6.25,
11586
+ },
11587
+ contextWindow: 1000000,
11588
+ maxTokens: 128000,
11589
+ },
11111
11590
  },
11112
11591
  "vercel-ai-gateway": {
11113
11592
  "alibaba/qwen-3-14b": {
@@ -11178,6 +11657,23 @@ export const MODELS = {
11178
11657
  contextWindow: 128000,
11179
11658
  maxTokens: 8192,
11180
11659
  },
11660
+ "alibaba/qwen-3.6-max-preview": {
11661
+ id: "alibaba/qwen-3.6-max-preview",
11662
+ name: "Qwen 3.6 Max Preview",
11663
+ api: "anthropic-messages",
11664
+ provider: "vercel-ai-gateway",
11665
+ baseUrl: "https://ai-gateway.vercel.sh",
11666
+ reasoning: true,
11667
+ input: ["text", "image"],
11668
+ cost: {
11669
+ input: 1.3,
11670
+ output: 7.8,
11671
+ cacheRead: 0.26,
11672
+ cacheWrite: 1.625,
11673
+ },
11674
+ contextWindow: 240000,
11675
+ maxTokens: 64000,
11676
+ },
11181
11677
  "alibaba/qwen3-235b-a22b-thinking": {
11182
11678
  id: "alibaba/qwen3-235b-a22b-thinking",
11183
11679
  name: "Qwen3 235B A22B Thinking 2507",
@@ -11377,7 +11873,7 @@ export const MODELS = {
11377
11873
  input: 0.5,
11378
11874
  output: 3,
11379
11875
  cacheRead: 0.09999999999999999,
11380
- cacheWrite: 0,
11876
+ cacheWrite: 0.625,
11381
11877
  },
11382
11878
  contextWindow: 1000000,
11383
11879
  maxTokens: 64000,
@@ -12521,6 +13017,23 @@ export const MODELS = {
12521
13017
  contextWindow: 262114,
12522
13018
  maxTokens: 262114,
12523
13019
  },
13020
+ "moonshotai/kimi-k2.6": {
13021
+ id: "moonshotai/kimi-k2.6",
13022
+ name: "Kimi K2.6",
13023
+ api: "anthropic-messages",
13024
+ provider: "vercel-ai-gateway",
13025
+ baseUrl: "https://ai-gateway.vercel.sh",
13026
+ reasoning: true,
13027
+ input: ["text", "image"],
13028
+ cost: {
13029
+ input: 0.95,
13030
+ output: 4,
13031
+ cacheRead: 0.16,
13032
+ cacheWrite: 0,
13033
+ },
13034
+ contextWindow: 262000,
13035
+ maxTokens: 262000,
13036
+ },
12524
13037
  "nvidia/nemotron-nano-12b-v2-vl": {
12525
13038
  id: "nvidia/nemotron-nano-12b-v2-vl",
12526
13039
  name: "Nvidia Nemotron Nano 12B V2 VL",
@@ -13482,9 +13995,9 @@ export const MODELS = {
13482
13995
  reasoning: true,
13483
13996
  input: ["text"],
13484
13997
  cost: {
13485
- input: 0.09,
13486
- output: 0.29,
13487
- cacheRead: 0.045,
13998
+ input: 0.09999999999999999,
13999
+ output: 0.3,
14000
+ cacheRead: 0.01,
13488
14001
  cacheWrite: 0,
13489
14002
  },
13490
14003
  contextWindow: 262144,
@@ -13701,15 +14214,15 @@ export const MODELS = {
13701
14214
  provider: "vercel-ai-gateway",
13702
14215
  baseUrl: "https://ai-gateway.vercel.sh",
13703
14216
  reasoning: true,
13704
- input: ["text", "image"],
14217
+ input: ["text"],
13705
14218
  cost: {
13706
14219
  input: 1.4,
13707
14220
  output: 4.4,
13708
14221
  cacheRead: 0.26,
13709
14222
  cacheWrite: 0,
13710
14223
  },
13711
- contextWindow: 202752,
13712
- maxTokens: 202752,
14224
+ contextWindow: 202800,
14225
+ maxTokens: 64000,
13713
14226
  },
13714
14227
  "zai/glm-5v-turbo": {
13715
14228
  id: "zai/glm-5v-turbo",