@hyperspaceng/neural-ai 0.68.1 → 0.69.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (36) hide show
  1. package/README.md +3 -1
  2. package/dist/env-api-keys.d.ts.map +1 -1
  3. package/dist/env-api-keys.js +1 -0
  4. package/dist/env-api-keys.js.map +1 -1
  5. package/dist/index.d.ts +2 -2
  6. package/dist/index.d.ts.map +1 -1
  7. package/dist/index.js +1 -1
  8. package/dist/index.js.map +1 -1
  9. package/dist/models.generated.d.ts +610 -30
  10. package/dist/models.generated.d.ts.map +1 -1
  11. package/dist/models.generated.js +624 -44
  12. package/dist/models.generated.js.map +1 -1
  13. package/dist/providers/amazon-bedrock.d.ts.map +1 -1
  14. package/dist/providers/amazon-bedrock.js +49 -9
  15. package/dist/providers/amazon-bedrock.js.map +1 -1
  16. package/dist/providers/anthropic.d.ts.map +1 -1
  17. package/dist/providers/anthropic.js +140 -8
  18. package/dist/providers/anthropic.js.map +1 -1
  19. package/dist/providers/transform-messages.d.ts.map +1 -1
  20. package/dist/providers/transform-messages.js +2 -0
  21. package/dist/providers/transform-messages.js.map +1 -1
  22. package/dist/types.d.ts +2 -2
  23. package/dist/types.d.ts.map +1 -1
  24. package/dist/types.js.map +1 -1
  25. package/dist/utils/json-parse.d.ts +8 -1
  26. package/dist/utils/json-parse.d.ts.map +1 -1
  27. package/dist/utils/json-parse.js +89 -5
  28. package/dist/utils/json-parse.js.map +1 -1
  29. package/dist/utils/typebox-helpers.d.ts +1 -1
  30. package/dist/utils/typebox-helpers.d.ts.map +1 -1
  31. package/dist/utils/typebox-helpers.js +1 -1
  32. package/dist/utils/typebox-helpers.js.map +1 -1
  33. package/dist/utils/validation.d.ts.map +1 -1
  34. package/dist/utils/validation.js +242 -41
  35. package/dist/utils/validation.js.map +1 -1
  36. package/package.json +2 -4
@@ -325,6 +325,40 @@ export const MODELS = {
325
325
  contextWindow: 1000000,
326
326
  maxTokens: 64000,
327
327
  },
328
+ "au.anthropic.claude-opus-4-6-v1": {
329
+ id: "au.anthropic.claude-opus-4-6-v1",
330
+ name: "AU Anthropic Claude Opus 4.6",
331
+ api: "bedrock-converse-stream",
332
+ provider: "amazon-bedrock",
333
+ baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
334
+ reasoning: true,
335
+ input: ["text", "image"],
336
+ cost: {
337
+ input: 16.5,
338
+ output: 82.5,
339
+ cacheRead: 0.5,
340
+ cacheWrite: 6.25,
341
+ },
342
+ contextWindow: 1000000,
343
+ maxTokens: 128000,
344
+ },
345
+ "au.anthropic.claude-sonnet-4-6": {
346
+ id: "au.anthropic.claude-sonnet-4-6",
347
+ name: "AU Anthropic Claude Sonnet 4.6",
348
+ api: "bedrock-converse-stream",
349
+ provider: "amazon-bedrock",
350
+ baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
351
+ reasoning: true,
352
+ input: ["text", "image"],
353
+ cost: {
354
+ input: 3.3,
355
+ output: 16.5,
356
+ cacheRead: 0.33,
357
+ cacheWrite: 4.125,
358
+ },
359
+ contextWindow: 1000000,
360
+ maxTokens: 128000,
361
+ },
328
362
  "deepseek.r1-v1:0": {
329
363
  id: "deepseek.r1-v1:0",
330
364
  name: "DeepSeek-R1",
@@ -381,7 +415,7 @@ export const MODELS = {
381
415
  name: "Claude Haiku 4.5 (EU)",
382
416
  api: "bedrock-converse-stream",
383
417
  provider: "amazon-bedrock",
384
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
418
+ baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
385
419
  reasoning: true,
386
420
  input: ["text", "image"],
387
421
  cost: {
@@ -398,7 +432,7 @@ export const MODELS = {
398
432
  name: "Claude Opus 4.5 (EU)",
399
433
  api: "bedrock-converse-stream",
400
434
  provider: "amazon-bedrock",
401
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
435
+ baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
402
436
  reasoning: true,
403
437
  input: ["text", "image"],
404
438
  cost: {
@@ -415,7 +449,7 @@ export const MODELS = {
415
449
  name: "Claude Opus 4.6 (EU)",
416
450
  api: "bedrock-converse-stream",
417
451
  provider: "amazon-bedrock",
418
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
452
+ baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
419
453
  reasoning: true,
420
454
  input: ["text", "image"],
421
455
  cost: {
@@ -432,7 +466,7 @@ export const MODELS = {
432
466
  name: "Claude Opus 4.7 (EU)",
433
467
  api: "bedrock-converse-stream",
434
468
  provider: "amazon-bedrock",
435
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
469
+ baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
436
470
  reasoning: true,
437
471
  input: ["text", "image"],
438
472
  cost: {
@@ -449,7 +483,7 @@ export const MODELS = {
449
483
  name: "Claude Sonnet 4 (EU)",
450
484
  api: "bedrock-converse-stream",
451
485
  provider: "amazon-bedrock",
452
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
486
+ baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
453
487
  reasoning: true,
454
488
  input: ["text", "image"],
455
489
  cost: {
@@ -466,7 +500,7 @@ export const MODELS = {
466
500
  name: "Claude Sonnet 4.5 (EU)",
467
501
  api: "bedrock-converse-stream",
468
502
  provider: "amazon-bedrock",
469
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
503
+ baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
470
504
  reasoning: true,
471
505
  input: ["text", "image"],
472
506
  cost: {
@@ -483,7 +517,7 @@ export const MODELS = {
483
517
  name: "Claude Sonnet 4.6 (EU)",
484
518
  api: "bedrock-converse-stream",
485
519
  provider: "amazon-bedrock",
486
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
520
+ baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
487
521
  reasoning: true,
488
522
  input: ["text", "image"],
489
523
  cost: {
@@ -2488,6 +2522,23 @@ export const MODELS = {
2488
2522
  contextWindow: 1050000,
2489
2523
  maxTokens: 128000,
2490
2524
  },
2525
+ "gpt-5.5": {
2526
+ id: "gpt-5.5",
2527
+ name: "GPT-5.5",
2528
+ api: "azure-openai-responses",
2529
+ provider: "azure-openai-responses",
2530
+ baseUrl: "",
2531
+ reasoning: true,
2532
+ input: ["text", "image"],
2533
+ cost: {
2534
+ input: 5,
2535
+ output: 30,
2536
+ cacheRead: 0.5,
2537
+ cacheWrite: 0,
2538
+ },
2539
+ contextWindow: 1050000,
2540
+ maxTokens: 130000,
2541
+ },
2491
2542
  "o1": {
2492
2543
  id: "o1",
2493
2544
  name: "o1",
@@ -2695,6 +2746,314 @@ export const MODELS = {
2695
2746
  maxTokens: 40000,
2696
2747
  },
2697
2748
  },
2749
+ "fireworks": {
2750
+ "accounts/fireworks/models/deepseek-v3p1": {
2751
+ id: "accounts/fireworks/models/deepseek-v3p1",
2752
+ name: "DeepSeek V3.1",
2753
+ api: "anthropic-messages",
2754
+ provider: "fireworks",
2755
+ baseUrl: "https://api.fireworks.ai/inference",
2756
+ reasoning: true,
2757
+ input: ["text"],
2758
+ cost: {
2759
+ input: 0.56,
2760
+ output: 1.68,
2761
+ cacheRead: 0,
2762
+ cacheWrite: 0,
2763
+ },
2764
+ contextWindow: 163840,
2765
+ maxTokens: 163840,
2766
+ },
2767
+ "accounts/fireworks/models/deepseek-v3p2": {
2768
+ id: "accounts/fireworks/models/deepseek-v3p2",
2769
+ name: "DeepSeek V3.2",
2770
+ api: "anthropic-messages",
2771
+ provider: "fireworks",
2772
+ baseUrl: "https://api.fireworks.ai/inference",
2773
+ reasoning: true,
2774
+ input: ["text"],
2775
+ cost: {
2776
+ input: 0.56,
2777
+ output: 1.68,
2778
+ cacheRead: 0.28,
2779
+ cacheWrite: 0,
2780
+ },
2781
+ contextWindow: 160000,
2782
+ maxTokens: 160000,
2783
+ },
2784
+ "accounts/fireworks/models/glm-4p5": {
2785
+ id: "accounts/fireworks/models/glm-4p5",
2786
+ name: "GLM 4.5",
2787
+ api: "anthropic-messages",
2788
+ provider: "fireworks",
2789
+ baseUrl: "https://api.fireworks.ai/inference",
2790
+ reasoning: true,
2791
+ input: ["text"],
2792
+ cost: {
2793
+ input: 0.55,
2794
+ output: 2.19,
2795
+ cacheRead: 0,
2796
+ cacheWrite: 0,
2797
+ },
2798
+ contextWindow: 131072,
2799
+ maxTokens: 131072,
2800
+ },
2801
+ "accounts/fireworks/models/glm-4p5-air": {
2802
+ id: "accounts/fireworks/models/glm-4p5-air",
2803
+ name: "GLM 4.5 Air",
2804
+ api: "anthropic-messages",
2805
+ provider: "fireworks",
2806
+ baseUrl: "https://api.fireworks.ai/inference",
2807
+ reasoning: true,
2808
+ input: ["text"],
2809
+ cost: {
2810
+ input: 0.22,
2811
+ output: 0.88,
2812
+ cacheRead: 0,
2813
+ cacheWrite: 0,
2814
+ },
2815
+ contextWindow: 131072,
2816
+ maxTokens: 131072,
2817
+ },
2818
+ "accounts/fireworks/models/glm-4p7": {
2819
+ id: "accounts/fireworks/models/glm-4p7",
2820
+ name: "GLM 4.7",
2821
+ api: "anthropic-messages",
2822
+ provider: "fireworks",
2823
+ baseUrl: "https://api.fireworks.ai/inference",
2824
+ reasoning: true,
2825
+ input: ["text"],
2826
+ cost: {
2827
+ input: 0.6,
2828
+ output: 2.2,
2829
+ cacheRead: 0.3,
2830
+ cacheWrite: 0,
2831
+ },
2832
+ contextWindow: 198000,
2833
+ maxTokens: 198000,
2834
+ },
2835
+ "accounts/fireworks/models/glm-5": {
2836
+ id: "accounts/fireworks/models/glm-5",
2837
+ name: "GLM 5",
2838
+ api: "anthropic-messages",
2839
+ provider: "fireworks",
2840
+ baseUrl: "https://api.fireworks.ai/inference",
2841
+ reasoning: true,
2842
+ input: ["text"],
2843
+ cost: {
2844
+ input: 1,
2845
+ output: 3.2,
2846
+ cacheRead: 0.5,
2847
+ cacheWrite: 0,
2848
+ },
2849
+ contextWindow: 202752,
2850
+ maxTokens: 131072,
2851
+ },
2852
+ "accounts/fireworks/models/glm-5p1": {
2853
+ id: "accounts/fireworks/models/glm-5p1",
2854
+ name: "GLM 5.1",
2855
+ api: "anthropic-messages",
2856
+ provider: "fireworks",
2857
+ baseUrl: "https://api.fireworks.ai/inference",
2858
+ reasoning: true,
2859
+ input: ["text"],
2860
+ cost: {
2861
+ input: 1.4,
2862
+ output: 4.4,
2863
+ cacheRead: 0.26,
2864
+ cacheWrite: 0,
2865
+ },
2866
+ contextWindow: 202800,
2867
+ maxTokens: 131072,
2868
+ },
2869
+ "accounts/fireworks/models/gpt-oss-120b": {
2870
+ id: "accounts/fireworks/models/gpt-oss-120b",
2871
+ name: "GPT OSS 120B",
2872
+ api: "anthropic-messages",
2873
+ provider: "fireworks",
2874
+ baseUrl: "https://api.fireworks.ai/inference",
2875
+ reasoning: true,
2876
+ input: ["text"],
2877
+ cost: {
2878
+ input: 0.15,
2879
+ output: 0.6,
2880
+ cacheRead: 0,
2881
+ cacheWrite: 0,
2882
+ },
2883
+ contextWindow: 131072,
2884
+ maxTokens: 32768,
2885
+ },
2886
+ "accounts/fireworks/models/gpt-oss-20b": {
2887
+ id: "accounts/fireworks/models/gpt-oss-20b",
2888
+ name: "GPT OSS 20B",
2889
+ api: "anthropic-messages",
2890
+ provider: "fireworks",
2891
+ baseUrl: "https://api.fireworks.ai/inference",
2892
+ reasoning: true,
2893
+ input: ["text"],
2894
+ cost: {
2895
+ input: 0.05,
2896
+ output: 0.2,
2897
+ cacheRead: 0,
2898
+ cacheWrite: 0,
2899
+ },
2900
+ contextWindow: 131072,
2901
+ maxTokens: 32768,
2902
+ },
2903
+ "accounts/fireworks/models/kimi-k2-instruct": {
2904
+ id: "accounts/fireworks/models/kimi-k2-instruct",
2905
+ name: "Kimi K2 Instruct",
2906
+ api: "anthropic-messages",
2907
+ provider: "fireworks",
2908
+ baseUrl: "https://api.fireworks.ai/inference",
2909
+ reasoning: false,
2910
+ input: ["text"],
2911
+ cost: {
2912
+ input: 1,
2913
+ output: 3,
2914
+ cacheRead: 0,
2915
+ cacheWrite: 0,
2916
+ },
2917
+ contextWindow: 128000,
2918
+ maxTokens: 16384,
2919
+ },
2920
+ "accounts/fireworks/models/kimi-k2-thinking": {
2921
+ id: "accounts/fireworks/models/kimi-k2-thinking",
2922
+ name: "Kimi K2 Thinking",
2923
+ api: "anthropic-messages",
2924
+ provider: "fireworks",
2925
+ baseUrl: "https://api.fireworks.ai/inference",
2926
+ reasoning: true,
2927
+ input: ["text"],
2928
+ cost: {
2929
+ input: 0.6,
2930
+ output: 2.5,
2931
+ cacheRead: 0.3,
2932
+ cacheWrite: 0,
2933
+ },
2934
+ contextWindow: 256000,
2935
+ maxTokens: 256000,
2936
+ },
2937
+ "accounts/fireworks/models/kimi-k2p5": {
2938
+ id: "accounts/fireworks/models/kimi-k2p5",
2939
+ name: "Kimi K2.5",
2940
+ api: "anthropic-messages",
2941
+ provider: "fireworks",
2942
+ baseUrl: "https://api.fireworks.ai/inference",
2943
+ reasoning: true,
2944
+ input: ["text", "image"],
2945
+ cost: {
2946
+ input: 0.6,
2947
+ output: 3,
2948
+ cacheRead: 0.1,
2949
+ cacheWrite: 0,
2950
+ },
2951
+ contextWindow: 256000,
2952
+ maxTokens: 256000,
2953
+ },
2954
+ "accounts/fireworks/models/kimi-k2p6": {
2955
+ id: "accounts/fireworks/models/kimi-k2p6",
2956
+ name: "Kimi K2.6",
2957
+ api: "anthropic-messages",
2958
+ provider: "fireworks",
2959
+ baseUrl: "https://api.fireworks.ai/inference",
2960
+ reasoning: true,
2961
+ input: ["text", "image"],
2962
+ cost: {
2963
+ input: 0.95,
2964
+ output: 4,
2965
+ cacheRead: 0.16,
2966
+ cacheWrite: 0,
2967
+ },
2968
+ contextWindow: 262000,
2969
+ maxTokens: 262000,
2970
+ },
2971
+ "accounts/fireworks/models/minimax-m2p1": {
2972
+ id: "accounts/fireworks/models/minimax-m2p1",
2973
+ name: "MiniMax-M2.1",
2974
+ api: "anthropic-messages",
2975
+ provider: "fireworks",
2976
+ baseUrl: "https://api.fireworks.ai/inference",
2977
+ reasoning: true,
2978
+ input: ["text"],
2979
+ cost: {
2980
+ input: 0.3,
2981
+ output: 1.2,
2982
+ cacheRead: 0.03,
2983
+ cacheWrite: 0,
2984
+ },
2985
+ contextWindow: 200000,
2986
+ maxTokens: 200000,
2987
+ },
2988
+ "accounts/fireworks/models/minimax-m2p5": {
2989
+ id: "accounts/fireworks/models/minimax-m2p5",
2990
+ name: "MiniMax-M2.5",
2991
+ api: "anthropic-messages",
2992
+ provider: "fireworks",
2993
+ baseUrl: "https://api.fireworks.ai/inference",
2994
+ reasoning: true,
2995
+ input: ["text"],
2996
+ cost: {
2997
+ input: 0.3,
2998
+ output: 1.2,
2999
+ cacheRead: 0.03,
3000
+ cacheWrite: 0,
3001
+ },
3002
+ contextWindow: 196608,
3003
+ maxTokens: 196608,
3004
+ },
3005
+ "accounts/fireworks/models/minimax-m2p7": {
3006
+ id: "accounts/fireworks/models/minimax-m2p7",
3007
+ name: "MiniMax-M2.7",
3008
+ api: "anthropic-messages",
3009
+ provider: "fireworks",
3010
+ baseUrl: "https://api.fireworks.ai/inference",
3011
+ reasoning: true,
3012
+ input: ["text"],
3013
+ cost: {
3014
+ input: 0.3,
3015
+ output: 1.2,
3016
+ cacheRead: 0.03,
3017
+ cacheWrite: 0,
3018
+ },
3019
+ contextWindow: 196608,
3020
+ maxTokens: 196608,
3021
+ },
3022
+ "accounts/fireworks/models/qwen3p6-plus": {
3023
+ id: "accounts/fireworks/models/qwen3p6-plus",
3024
+ name: "Qwen 3.6 Plus",
3025
+ api: "anthropic-messages",
3026
+ provider: "fireworks",
3027
+ baseUrl: "https://api.fireworks.ai/inference",
3028
+ reasoning: true,
3029
+ input: ["text", "image"],
3030
+ cost: {
3031
+ input: 0.5,
3032
+ output: 3,
3033
+ cacheRead: 0.1,
3034
+ cacheWrite: 0,
3035
+ },
3036
+ contextWindow: 128000,
3037
+ maxTokens: 8192,
3038
+ },
3039
+ "accounts/fireworks/routers/kimi-k2p5-turbo": {
3040
+ id: "accounts/fireworks/routers/kimi-k2p5-turbo",
3041
+ name: "Kimi K2.5 Turbo (firepass)",
3042
+ api: "anthropic-messages",
3043
+ provider: "fireworks",
3044
+ baseUrl: "https://api.fireworks.ai/inference",
3045
+ reasoning: true,
3046
+ input: ["text", "image"],
3047
+ cost: {
3048
+ input: 0,
3049
+ output: 0,
3050
+ cacheRead: 0,
3051
+ cacheWrite: 0,
3052
+ },
3053
+ contextWindow: 256000,
3054
+ maxTokens: 256000,
3055
+ },
3056
+ },
2698
3057
  "github-copilot": {
2699
3058
  "claude-haiku-4.5": {
2700
3059
  id: "claude-haiku-4.5",
@@ -3580,8 +3939,8 @@ export const MODELS = {
3580
3939
  contextWindow: 131072,
3581
3940
  maxTokens: 8192,
3582
3941
  },
3583
- "gemma-4-26b-it": {
3584
- id: "gemma-4-26b-it",
3942
+ "gemma-4-26b-a4b-it": {
3943
+ id: "gemma-4-26b-a4b-it",
3585
3944
  name: "Gemma 4 26B",
3586
3945
  api: "google-generative-ai",
3587
3946
  provider: "google",
@@ -3856,6 +4215,23 @@ export const MODELS = {
3856
4215
  contextWindow: 1048576,
3857
4216
  maxTokens: 65535,
3858
4217
  },
4218
+ "gemini-3.1-flash-lite-preview": {
4219
+ id: "gemini-3.1-flash-lite-preview",
4220
+ name: "Gemini 3.1 Flash Lite Preview (Cloud Code Assist)",
4221
+ api: "google-gemini-cli",
4222
+ provider: "google-gemini-cli",
4223
+ baseUrl: "https://cloudcode-pa.googleapis.com",
4224
+ reasoning: true,
4225
+ input: ["text", "image"],
4226
+ cost: {
4227
+ input: 0,
4228
+ output: 0,
4229
+ cacheRead: 0,
4230
+ cacheWrite: 0,
4231
+ },
4232
+ contextWindow: 1048576,
4233
+ maxTokens: 65535,
4234
+ },
3859
4235
  "gemini-3.1-pro-preview": {
3860
4236
  id: "gemini-3.1-pro-preview",
3861
4237
  name: "Gemini 3.1 Pro Preview (Cloud Code Assist)",
@@ -5899,6 +6275,23 @@ export const MODELS = {
5899
6275
  contextWindow: 1050000,
5900
6276
  maxTokens: 128000,
5901
6277
  },
6278
+ "gpt-5.5": {
6279
+ id: "gpt-5.5",
6280
+ name: "GPT-5.5",
6281
+ api: "openai-responses",
6282
+ provider: "openai",
6283
+ baseUrl: "https://api.openai.com/v1",
6284
+ reasoning: true,
6285
+ input: ["text", "image"],
6286
+ cost: {
6287
+ input: 5,
6288
+ output: 30,
6289
+ cacheRead: 0.5,
6290
+ cacheWrite: 0,
6291
+ },
6292
+ contextWindow: 1050000,
6293
+ maxTokens: 130000,
6294
+ },
5902
6295
  "o1": {
5903
6296
  id: "o1",
5904
6297
  name: "o1",
@@ -6668,6 +7061,23 @@ export const MODELS = {
6668
7061
  contextWindow: 1050000,
6669
7062
  maxTokens: 128000,
6670
7063
  },
7064
+ "hy3-preview-free": {
7065
+ id: "hy3-preview-free",
7066
+ name: "Hy3 preview Free",
7067
+ api: "openai-completions",
7068
+ provider: "opencode",
7069
+ baseUrl: "https://opencode.ai/zen/v1",
7070
+ reasoning: true,
7071
+ input: ["text"],
7072
+ cost: {
7073
+ input: 0,
7074
+ output: 0,
7075
+ cacheRead: 0,
7076
+ cacheWrite: 0,
7077
+ },
7078
+ contextWindow: 256000,
7079
+ maxTokens: 64000,
7080
+ },
6671
7081
  "kimi-k2.5": {
6672
7082
  id: "kimi-k2.5",
6673
7083
  name: "Kimi K2.5",
@@ -6906,7 +7316,7 @@ export const MODELS = {
6906
7316
  cacheWrite: 0,
6907
7317
  },
6908
7318
  contextWindow: 262144,
6909
- maxTokens: 64000,
7319
+ maxTokens: 128000,
6910
7320
  },
6911
7321
  "mimo-v2-pro": {
6912
7322
  id: "mimo-v2-pro",
@@ -6923,7 +7333,41 @@ export const MODELS = {
6923
7333
  cacheWrite: 0,
6924
7334
  },
6925
7335
  contextWindow: 1048576,
6926
- maxTokens: 64000,
7336
+ maxTokens: 128000,
7337
+ },
7338
+ "mimo-v2.5": {
7339
+ id: "mimo-v2.5",
7340
+ name: "MiMo V2.5",
7341
+ api: "openai-completions",
7342
+ provider: "opencode-go",
7343
+ baseUrl: "https://opencode.ai/zen/go/v1",
7344
+ reasoning: true,
7345
+ input: ["text", "image"],
7346
+ cost: {
7347
+ input: 0.4,
7348
+ output: 2,
7349
+ cacheRead: 0.08,
7350
+ cacheWrite: 0,
7351
+ },
7352
+ contextWindow: 262144,
7353
+ maxTokens: 128000,
7354
+ },
7355
+ "mimo-v2.5-pro": {
7356
+ id: "mimo-v2.5-pro",
7357
+ name: "MiMo V2.5 Pro",
7358
+ api: "openai-completions",
7359
+ provider: "opencode-go",
7360
+ baseUrl: "https://opencode.ai/zen/go/v1",
7361
+ reasoning: true,
7362
+ input: ["text"],
7363
+ cost: {
7364
+ input: 1,
7365
+ output: 3,
7366
+ cacheRead: 0.2,
7367
+ cacheWrite: 0,
7368
+ },
7369
+ contextWindow: 1048576,
7370
+ maxTokens: 128000,
6927
7371
  },
6928
7372
  "minimax-m2.5": {
6929
7373
  id: "minimax-m2.5",
@@ -7369,17 +7813,17 @@ export const MODELS = {
7369
7813
  contextWindow: 1000000,
7370
7814
  maxTokens: 128000,
7371
7815
  },
7372
- "arcee-ai/trinity-large-preview:free": {
7373
- id: "arcee-ai/trinity-large-preview:free",
7374
- name: "Arcee AI: Trinity Large Preview (free)",
7816
+ "arcee-ai/trinity-large-preview": {
7817
+ id: "arcee-ai/trinity-large-preview",
7818
+ name: "Arcee AI: Trinity Large Preview",
7375
7819
  api: "openai-completions",
7376
7820
  provider: "openrouter",
7377
7821
  baseUrl: "https://openrouter.ai/api/v1",
7378
7822
  reasoning: false,
7379
7823
  input: ["text"],
7380
7824
  cost: {
7381
- input: 0,
7382
- output: 0,
7825
+ input: 0.15,
7826
+ output: 0.44999999999999996,
7383
7827
  cacheRead: 0,
7384
7828
  cacheWrite: 0,
7385
7829
  },
@@ -7707,7 +8151,7 @@ export const MODELS = {
7707
8151
  cacheWrite: 0,
7708
8152
  },
7709
8153
  contextWindow: 131072,
7710
- maxTokens: 32768,
8154
+ maxTokens: 65536,
7711
8155
  },
7712
8156
  "deepseek/deepseek-v3.2-exp": {
7713
8157
  id: "deepseek/deepseek-v3.2-exp",
@@ -7726,6 +8170,40 @@ export const MODELS = {
7726
8170
  contextWindow: 163840,
7727
8171
  maxTokens: 65536,
7728
8172
  },
8173
+ "deepseek/deepseek-v4-flash": {
8174
+ id: "deepseek/deepseek-v4-flash",
8175
+ name: "DeepSeek: DeepSeek V4 Flash",
8176
+ api: "openai-completions",
8177
+ provider: "openrouter",
8178
+ baseUrl: "https://openrouter.ai/api/v1",
8179
+ reasoning: true,
8180
+ input: ["text"],
8181
+ cost: {
8182
+ input: 0.14,
8183
+ output: 0.28,
8184
+ cacheRead: 0.028,
8185
+ cacheWrite: 0,
8186
+ },
8187
+ contextWindow: 1048576,
8188
+ maxTokens: 384000,
8189
+ },
8190
+ "deepseek/deepseek-v4-pro": {
8191
+ id: "deepseek/deepseek-v4-pro",
8192
+ name: "DeepSeek: DeepSeek V4 Pro",
8193
+ api: "openai-completions",
8194
+ provider: "openrouter",
8195
+ baseUrl: "https://openrouter.ai/api/v1",
8196
+ reasoning: true,
8197
+ input: ["text"],
8198
+ cost: {
8199
+ input: 1.74,
8200
+ output: 3.48,
8201
+ cacheRead: 0.145,
8202
+ cacheWrite: 0,
8203
+ },
8204
+ contextWindow: 1048576,
8205
+ maxTokens: 384000,
8206
+ },
7729
8207
  "essentialai/rnj-1-instruct": {
7730
8208
  id: "essentialai/rnj-1-instruct",
7731
8209
  name: "EssentialAI: Rnj 1 Instruct",
@@ -7956,13 +8434,13 @@ export const MODELS = {
7956
8434
  reasoning: true,
7957
8435
  input: ["text", "image"],
7958
8436
  cost: {
7959
- input: 0.07,
7960
- output: 0.35,
7961
- cacheRead: 0.04,
8437
+ input: 0.06,
8438
+ output: 0.33,
8439
+ cacheRead: 0,
7962
8440
  cacheWrite: 0,
7963
8441
  },
7964
8442
  contextWindow: 262144,
7965
- maxTokens: 262144,
8443
+ maxTokens: 4096,
7966
8444
  },
7967
8445
  "google/gemma-4-26b-a4b-it:free": {
7968
8446
  id: "google/gemma-4-26b-a4b-it:free",
@@ -7992,7 +8470,7 @@ export const MODELS = {
7992
8470
  cost: {
7993
8471
  input: 0.13,
7994
8472
  output: 0.38,
7995
- cacheRead: 0.019999999499999997,
8473
+ cacheRead: 0,
7996
8474
  cacheWrite: 0,
7997
8475
  },
7998
8476
  contextWindow: 262144,
@@ -8032,6 +8510,23 @@ export const MODELS = {
8032
8510
  contextWindow: 128000,
8033
8511
  maxTokens: 50000,
8034
8512
  },
8513
+ "inclusionai/ling-2.6-1t:free": {
8514
+ id: "inclusionai/ling-2.6-1t:free",
8515
+ name: "inclusionAI: Ling-2.6-1T (free)",
8516
+ api: "openai-completions",
8517
+ provider: "openrouter",
8518
+ baseUrl: "https://openrouter.ai/api/v1",
8519
+ reasoning: false,
8520
+ input: ["text"],
8521
+ cost: {
8522
+ input: 0,
8523
+ output: 0,
8524
+ cacheRead: 0,
8525
+ cacheWrite: 0,
8526
+ },
8527
+ contextWindow: 262144,
8528
+ maxTokens: 32768,
8529
+ },
8035
8530
  "inclusionai/ling-2.6-flash:free": {
8036
8531
  id: "inclusionai/ling-2.6-flash:free",
8037
8532
  name: "inclusionAI: Ling-2.6-flash (free)",
@@ -8229,12 +8724,12 @@ export const MODELS = {
8229
8724
  input: ["text"],
8230
8725
  cost: {
8231
8726
  input: 0.15,
8232
- output: 1.2,
8233
- cacheRead: 0.075,
8727
+ output: 1.15,
8728
+ cacheRead: 0.03,
8234
8729
  cacheWrite: 0,
8235
8730
  },
8236
8731
  contextWindow: 196608,
8237
- maxTokens: 65536,
8732
+ maxTokens: 4096,
8238
8733
  },
8239
8734
  "minimax/minimax-m2.5:free": {
8240
8735
  id: "minimax/minimax-m2.5:free",
@@ -8500,13 +8995,13 @@ export const MODELS = {
8500
8995
  reasoning: false,
8501
8996
  input: ["text"],
8502
8997
  cost: {
8503
- input: 0.02,
8504
- output: 0.04,
8998
+ input: 0.01,
8999
+ output: 0.03,
8505
9000
  cacheRead: 0,
8506
9001
  cacheWrite: 0,
8507
9002
  },
8508
9003
  contextWindow: 131072,
8509
- maxTokens: 16384,
9004
+ maxTokens: 4096,
8510
9005
  },
8511
9006
  "mistralai/mistral-saba": {
8512
9007
  id: "mistralai/mistral-saba",
@@ -8721,13 +9216,13 @@ export const MODELS = {
8721
9216
  reasoning: true,
8722
9217
  input: ["text", "image"],
8723
9218
  cost: {
8724
- input: 0.7999999999999999,
8725
- output: 3.5,
8726
- cacheRead: 0.19999999999999998,
9219
+ input: 0.7448,
9220
+ output: 4.655,
9221
+ cacheRead: 0.1463,
8727
9222
  cacheWrite: 0,
8728
9223
  },
8729
- contextWindow: 262144,
8730
- maxTokens: 262144,
9224
+ contextWindow: 256000,
9225
+ maxTokens: 65536,
8731
9226
  },
8732
9227
  "nex-agi/deepseek-v3.1-nex-n1": {
8733
9228
  id: "nex-agi/deepseek-v3.1-nex-n1",
@@ -10064,13 +10559,13 @@ export const MODELS = {
10064
10559
  reasoning: true,
10065
10560
  input: ["text"],
10066
10561
  cost: {
10067
- input: 0.13,
10068
- output: 0.6,
10562
+ input: 0.14950000000000002,
10563
+ output: 1.495,
10069
10564
  cacheRead: 0,
10070
10565
  cacheWrite: 0,
10071
10566
  },
10072
- contextWindow: 262144,
10073
- maxTokens: 262144,
10567
+ contextWindow: 131072,
10568
+ maxTokens: 4096,
10074
10569
  },
10075
10570
  "qwen/qwen3-30b-a3b": {
10076
10571
  id: "qwen/qwen3-30b-a3b",
@@ -10701,6 +11196,23 @@ export const MODELS = {
10701
11196
  contextWindow: 262144,
10702
11197
  maxTokens: 65536,
10703
11198
  },
11199
+ "tencent/hy3-preview:free": {
11200
+ id: "tencent/hy3-preview:free",
11201
+ name: "Tencent: Hy3 preview (free)",
11202
+ api: "openai-completions",
11203
+ provider: "openrouter",
11204
+ baseUrl: "https://openrouter.ai/api/v1",
11205
+ reasoning: true,
11206
+ input: ["text"],
11207
+ cost: {
11208
+ input: 0,
11209
+ output: 0,
11210
+ cacheRead: 0,
11211
+ cacheWrite: 0,
11212
+ },
11213
+ contextWindow: 262144,
11214
+ maxTokens: 262144,
11215
+ },
10704
11216
  "thedrummer/rocinante-12b": {
10705
11217
  id: "thedrummer/rocinante-12b",
10706
11218
  name: "TheDrummer: Rocinante 12B",
@@ -10973,6 +11485,40 @@ export const MODELS = {
10973
11485
  contextWindow: 1048576,
10974
11486
  maxTokens: 131072,
10975
11487
  },
11488
+ "xiaomi/mimo-v2.5": {
11489
+ id: "xiaomi/mimo-v2.5",
11490
+ name: "Xiaomi: MiMo-V2.5",
11491
+ api: "openai-completions",
11492
+ provider: "openrouter",
11493
+ baseUrl: "https://openrouter.ai/api/v1",
11494
+ reasoning: true,
11495
+ input: ["text", "image"],
11496
+ cost: {
11497
+ input: 0.39999999999999997,
11498
+ output: 2,
11499
+ cacheRead: 0.08,
11500
+ cacheWrite: 0,
11501
+ },
11502
+ contextWindow: 1048576,
11503
+ maxTokens: 131072,
11504
+ },
11505
+ "xiaomi/mimo-v2.5-pro": {
11506
+ id: "xiaomi/mimo-v2.5-pro",
11507
+ name: "Xiaomi: MiMo-V2.5-Pro",
11508
+ api: "openai-completions",
11509
+ provider: "openrouter",
11510
+ baseUrl: "https://openrouter.ai/api/v1",
11511
+ reasoning: true,
11512
+ input: ["text"],
11513
+ cost: {
11514
+ input: 1,
11515
+ output: 3,
11516
+ cacheRead: 0.19999999999999998,
11517
+ cacheWrite: 0,
11518
+ },
11519
+ contextWindow: 1048576,
11520
+ maxTokens: 131072,
11521
+ },
10976
11522
  "z-ai/glm-4-32b": {
10977
11523
  id: "z-ai/glm-4-32b",
10978
11524
  name: "Z.ai: GLM 4 32B ",
@@ -11497,7 +12043,7 @@ export const MODELS = {
11497
12043
  input: 0.5,
11498
12044
  output: 3,
11499
12045
  cacheRead: 0.09999999999999999,
11500
- cacheWrite: 0,
12046
+ cacheWrite: 0.625,
11501
12047
  },
11502
12048
  contextWindow: 1000000,
11503
12049
  maxTokens: 64000,
@@ -11876,6 +12422,40 @@ export const MODELS = {
11876
12422
  contextWindow: 128000,
11877
12423
  maxTokens: 64000,
11878
12424
  },
12425
+ "deepseek/deepseek-v4-flash": {
12426
+ id: "deepseek/deepseek-v4-flash",
12427
+ name: "DeepSeek V4 Flash",
12428
+ api: "anthropic-messages",
12429
+ provider: "vercel-ai-gateway",
12430
+ baseUrl: "https://ai-gateway.vercel.sh",
12431
+ reasoning: true,
12432
+ input: ["text"],
12433
+ cost: {
12434
+ input: 0.14,
12435
+ output: 0.28,
12436
+ cacheRead: 0.014,
12437
+ cacheWrite: 0,
12438
+ },
12439
+ contextWindow: 1000000,
12440
+ maxTokens: 384000,
12441
+ },
12442
+ "deepseek/deepseek-v4-pro": {
12443
+ id: "deepseek/deepseek-v4-pro",
12444
+ name: "DeepSeek V4 Pro",
12445
+ api: "anthropic-messages",
12446
+ provider: "vercel-ai-gateway",
12447
+ baseUrl: "https://ai-gateway.vercel.sh",
12448
+ reasoning: true,
12449
+ input: ["text"],
12450
+ cost: {
12451
+ input: 1.74,
12452
+ output: 3.48,
12453
+ cacheRead: 0.145,
12454
+ cacheWrite: 0,
12455
+ },
12456
+ contextWindow: 1000000,
12457
+ maxTokens: 384000,
12458
+ },
11879
12459
  "google/gemini-2.0-flash": {
11880
12460
  id: "google/gemini-2.0-flash",
11881
12461
  name: "Gemini 2.0 Flash",
@@ -13619,9 +14199,9 @@ export const MODELS = {
13619
14199
  reasoning: true,
13620
14200
  input: ["text"],
13621
14201
  cost: {
13622
- input: 0.09,
13623
- output: 0.29,
13624
- cacheRead: 0.045,
14202
+ input: 0.09999999999999999,
14203
+ output: 0.3,
14204
+ cacheRead: 0.01,
13625
14205
  cacheWrite: 0,
13626
14206
  },
13627
14207
  contextWindow: 262144,
@@ -13838,15 +14418,15 @@ export const MODELS = {
13838
14418
  provider: "vercel-ai-gateway",
13839
14419
  baseUrl: "https://ai-gateway.vercel.sh",
13840
14420
  reasoning: true,
13841
- input: ["text", "image"],
14421
+ input: ["text"],
13842
14422
  cost: {
13843
14423
  input: 1.4,
13844
14424
  output: 4.4,
13845
14425
  cacheRead: 0.26,
13846
14426
  cacheWrite: 0,
13847
14427
  },
13848
- contextWindow: 202752,
13849
- maxTokens: 202752,
14428
+ contextWindow: 202800,
14429
+ maxTokens: 64000,
13850
14430
  },
13851
14431
  "zai/glm-5v-turbo": {
13852
14432
  id: "zai/glm-5v-turbo",