@hyperspaceng/neural-ai 0.67.69 → 0.68.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (50) hide show
  1. package/README.md +5 -1
  2. package/dist/env-api-keys.d.ts.map +1 -1
  3. package/dist/env-api-keys.js +1 -0
  4. package/dist/env-api-keys.js.map +1 -1
  5. package/dist/models.generated.d.ts +618 -103
  6. package/dist/models.generated.d.ts.map +1 -1
  7. package/dist/models.generated.js +627 -114
  8. package/dist/models.generated.js.map +1 -1
  9. package/dist/providers/amazon-bedrock.d.ts.map +1 -1
  10. package/dist/providers/amazon-bedrock.js +55 -7
  11. package/dist/providers/amazon-bedrock.js.map +1 -1
  12. package/dist/providers/anthropic.d.ts.map +1 -1
  13. package/dist/providers/anthropic.js +141 -10
  14. package/dist/providers/anthropic.js.map +1 -1
  15. package/dist/providers/google-shared.d.ts.map +1 -1
  16. package/dist/providers/google-shared.js +30 -4
  17. package/dist/providers/google-shared.js.map +1 -1
  18. package/dist/providers/openai-completions.d.ts +5 -1
  19. package/dist/providers/openai-completions.d.ts.map +1 -1
  20. package/dist/providers/openai-completions.js +149 -60
  21. package/dist/providers/openai-completions.js.map +1 -1
  22. package/dist/providers/openai-responses-shared.d.ts.map +1 -1
  23. package/dist/providers/openai-responses-shared.js +2 -5
  24. package/dist/providers/openai-responses-shared.js.map +1 -1
  25. package/dist/providers/simple-options.d.ts.map +1 -1
  26. package/dist/providers/simple-options.js +1 -1
  27. package/dist/providers/simple-options.js.map +1 -1
  28. package/dist/providers/transform-messages.d.ts.map +1 -1
  29. package/dist/providers/transform-messages.js +41 -2
  30. package/dist/providers/transform-messages.js.map +1 -1
  31. package/dist/types.d.ts +5 -1
  32. package/dist/types.d.ts.map +1 -1
  33. package/dist/types.js.map +1 -1
  34. package/dist/utils/json-parse.d.ts +8 -1
  35. package/dist/utils/json-parse.d.ts.map +1 -1
  36. package/dist/utils/json-parse.js +89 -5
  37. package/dist/utils/json-parse.js.map +1 -1
  38. package/dist/utils/oauth/anthropic.d.ts.map +1 -1
  39. package/dist/utils/oauth/anthropic.js +1 -1
  40. package/dist/utils/oauth/anthropic.js.map +1 -1
  41. package/dist/utils/oauth/google-antigravity.d.ts.map +1 -1
  42. package/dist/utils/oauth/google-antigravity.js +2 -1
  43. package/dist/utils/oauth/google-antigravity.js.map +1 -1
  44. package/dist/utils/oauth/google-gemini-cli.d.ts.map +1 -1
  45. package/dist/utils/oauth/google-gemini-cli.js +2 -1
  46. package/dist/utils/oauth/google-gemini-cli.js.map +1 -1
  47. package/dist/utils/oauth/openai-codex.d.ts.map +1 -1
  48. package/dist/utils/oauth/openai-codex.js +3 -2
  49. package/dist/utils/oauth/openai-codex.js.map +1 -1
  50. package/package.json +1 -1
@@ -2693,6 +2693,314 @@ export declare const MODELS: {
2693
2693
  maxTokens: number;
2694
2694
  };
2695
2695
  };
2696
+ readonly fireworks: {
2697
+ readonly "accounts/fireworks/models/deepseek-v3p1": {
2698
+ id: string;
2699
+ name: string;
2700
+ api: "anthropic-messages";
2701
+ provider: string;
2702
+ baseUrl: string;
2703
+ reasoning: true;
2704
+ input: "text"[];
2705
+ cost: {
2706
+ input: number;
2707
+ output: number;
2708
+ cacheRead: number;
2709
+ cacheWrite: number;
2710
+ };
2711
+ contextWindow: number;
2712
+ maxTokens: number;
2713
+ };
2714
+ readonly "accounts/fireworks/models/deepseek-v3p2": {
2715
+ id: string;
2716
+ name: string;
2717
+ api: "anthropic-messages";
2718
+ provider: string;
2719
+ baseUrl: string;
2720
+ reasoning: true;
2721
+ input: "text"[];
2722
+ cost: {
2723
+ input: number;
2724
+ output: number;
2725
+ cacheRead: number;
2726
+ cacheWrite: number;
2727
+ };
2728
+ contextWindow: number;
2729
+ maxTokens: number;
2730
+ };
2731
+ readonly "accounts/fireworks/models/glm-4p5": {
2732
+ id: string;
2733
+ name: string;
2734
+ api: "anthropic-messages";
2735
+ provider: string;
2736
+ baseUrl: string;
2737
+ reasoning: true;
2738
+ input: "text"[];
2739
+ cost: {
2740
+ input: number;
2741
+ output: number;
2742
+ cacheRead: number;
2743
+ cacheWrite: number;
2744
+ };
2745
+ contextWindow: number;
2746
+ maxTokens: number;
2747
+ };
2748
+ readonly "accounts/fireworks/models/glm-4p5-air": {
2749
+ id: string;
2750
+ name: string;
2751
+ api: "anthropic-messages";
2752
+ provider: string;
2753
+ baseUrl: string;
2754
+ reasoning: true;
2755
+ input: "text"[];
2756
+ cost: {
2757
+ input: number;
2758
+ output: number;
2759
+ cacheRead: number;
2760
+ cacheWrite: number;
2761
+ };
2762
+ contextWindow: number;
2763
+ maxTokens: number;
2764
+ };
2765
+ readonly "accounts/fireworks/models/glm-4p7": {
2766
+ id: string;
2767
+ name: string;
2768
+ api: "anthropic-messages";
2769
+ provider: string;
2770
+ baseUrl: string;
2771
+ reasoning: true;
2772
+ input: "text"[];
2773
+ cost: {
2774
+ input: number;
2775
+ output: number;
2776
+ cacheRead: number;
2777
+ cacheWrite: number;
2778
+ };
2779
+ contextWindow: number;
2780
+ maxTokens: number;
2781
+ };
2782
+ readonly "accounts/fireworks/models/glm-5": {
2783
+ id: string;
2784
+ name: string;
2785
+ api: "anthropic-messages";
2786
+ provider: string;
2787
+ baseUrl: string;
2788
+ reasoning: true;
2789
+ input: "text"[];
2790
+ cost: {
2791
+ input: number;
2792
+ output: number;
2793
+ cacheRead: number;
2794
+ cacheWrite: number;
2795
+ };
2796
+ contextWindow: number;
2797
+ maxTokens: number;
2798
+ };
2799
+ readonly "accounts/fireworks/models/glm-5p1": {
2800
+ id: string;
2801
+ name: string;
2802
+ api: "anthropic-messages";
2803
+ provider: string;
2804
+ baseUrl: string;
2805
+ reasoning: true;
2806
+ input: "text"[];
2807
+ cost: {
2808
+ input: number;
2809
+ output: number;
2810
+ cacheRead: number;
2811
+ cacheWrite: number;
2812
+ };
2813
+ contextWindow: number;
2814
+ maxTokens: number;
2815
+ };
2816
+ readonly "accounts/fireworks/models/gpt-oss-120b": {
2817
+ id: string;
2818
+ name: string;
2819
+ api: "anthropic-messages";
2820
+ provider: string;
2821
+ baseUrl: string;
2822
+ reasoning: true;
2823
+ input: "text"[];
2824
+ cost: {
2825
+ input: number;
2826
+ output: number;
2827
+ cacheRead: number;
2828
+ cacheWrite: number;
2829
+ };
2830
+ contextWindow: number;
2831
+ maxTokens: number;
2832
+ };
2833
+ readonly "accounts/fireworks/models/gpt-oss-20b": {
2834
+ id: string;
2835
+ name: string;
2836
+ api: "anthropic-messages";
2837
+ provider: string;
2838
+ baseUrl: string;
2839
+ reasoning: true;
2840
+ input: "text"[];
2841
+ cost: {
2842
+ input: number;
2843
+ output: number;
2844
+ cacheRead: number;
2845
+ cacheWrite: number;
2846
+ };
2847
+ contextWindow: number;
2848
+ maxTokens: number;
2849
+ };
2850
+ readonly "accounts/fireworks/models/kimi-k2-instruct": {
2851
+ id: string;
2852
+ name: string;
2853
+ api: "anthropic-messages";
2854
+ provider: string;
2855
+ baseUrl: string;
2856
+ reasoning: false;
2857
+ input: "text"[];
2858
+ cost: {
2859
+ input: number;
2860
+ output: number;
2861
+ cacheRead: number;
2862
+ cacheWrite: number;
2863
+ };
2864
+ contextWindow: number;
2865
+ maxTokens: number;
2866
+ };
2867
+ readonly "accounts/fireworks/models/kimi-k2-thinking": {
2868
+ id: string;
2869
+ name: string;
2870
+ api: "anthropic-messages";
2871
+ provider: string;
2872
+ baseUrl: string;
2873
+ reasoning: true;
2874
+ input: "text"[];
2875
+ cost: {
2876
+ input: number;
2877
+ output: number;
2878
+ cacheRead: number;
2879
+ cacheWrite: number;
2880
+ };
2881
+ contextWindow: number;
2882
+ maxTokens: number;
2883
+ };
2884
+ readonly "accounts/fireworks/models/kimi-k2p5": {
2885
+ id: string;
2886
+ name: string;
2887
+ api: "anthropic-messages";
2888
+ provider: string;
2889
+ baseUrl: string;
2890
+ reasoning: true;
2891
+ input: ("image" | "text")[];
2892
+ cost: {
2893
+ input: number;
2894
+ output: number;
2895
+ cacheRead: number;
2896
+ cacheWrite: number;
2897
+ };
2898
+ contextWindow: number;
2899
+ maxTokens: number;
2900
+ };
2901
+ readonly "accounts/fireworks/models/kimi-k2p6": {
2902
+ id: string;
2903
+ name: string;
2904
+ api: "anthropic-messages";
2905
+ provider: string;
2906
+ baseUrl: string;
2907
+ reasoning: true;
2908
+ input: ("image" | "text")[];
2909
+ cost: {
2910
+ input: number;
2911
+ output: number;
2912
+ cacheRead: number;
2913
+ cacheWrite: number;
2914
+ };
2915
+ contextWindow: number;
2916
+ maxTokens: number;
2917
+ };
2918
+ readonly "accounts/fireworks/models/minimax-m2p1": {
2919
+ id: string;
2920
+ name: string;
2921
+ api: "anthropic-messages";
2922
+ provider: string;
2923
+ baseUrl: string;
2924
+ reasoning: true;
2925
+ input: "text"[];
2926
+ cost: {
2927
+ input: number;
2928
+ output: number;
2929
+ cacheRead: number;
2930
+ cacheWrite: number;
2931
+ };
2932
+ contextWindow: number;
2933
+ maxTokens: number;
2934
+ };
2935
+ readonly "accounts/fireworks/models/minimax-m2p5": {
2936
+ id: string;
2937
+ name: string;
2938
+ api: "anthropic-messages";
2939
+ provider: string;
2940
+ baseUrl: string;
2941
+ reasoning: true;
2942
+ input: "text"[];
2943
+ cost: {
2944
+ input: number;
2945
+ output: number;
2946
+ cacheRead: number;
2947
+ cacheWrite: number;
2948
+ };
2949
+ contextWindow: number;
2950
+ maxTokens: number;
2951
+ };
2952
+ readonly "accounts/fireworks/models/minimax-m2p7": {
2953
+ id: string;
2954
+ name: string;
2955
+ api: "anthropic-messages";
2956
+ provider: string;
2957
+ baseUrl: string;
2958
+ reasoning: true;
2959
+ input: "text"[];
2960
+ cost: {
2961
+ input: number;
2962
+ output: number;
2963
+ cacheRead: number;
2964
+ cacheWrite: number;
2965
+ };
2966
+ contextWindow: number;
2967
+ maxTokens: number;
2968
+ };
2969
+ readonly "accounts/fireworks/models/qwen3p6-plus": {
2970
+ id: string;
2971
+ name: string;
2972
+ api: "anthropic-messages";
2973
+ provider: string;
2974
+ baseUrl: string;
2975
+ reasoning: true;
2976
+ input: ("image" | "text")[];
2977
+ cost: {
2978
+ input: number;
2979
+ output: number;
2980
+ cacheRead: number;
2981
+ cacheWrite: number;
2982
+ };
2983
+ contextWindow: number;
2984
+ maxTokens: number;
2985
+ };
2986
+ readonly "accounts/fireworks/routers/kimi-k2p5-turbo": {
2987
+ id: string;
2988
+ name: string;
2989
+ api: "anthropic-messages";
2990
+ provider: string;
2991
+ baseUrl: string;
2992
+ reasoning: true;
2993
+ input: ("image" | "text")[];
2994
+ cost: {
2995
+ input: number;
2996
+ output: number;
2997
+ cacheRead: number;
2998
+ cacheWrite: number;
2999
+ };
3000
+ contextWindow: number;
3001
+ maxTokens: number;
3002
+ };
3003
+ };
2696
3004
  readonly "github-copilot": {
2697
3005
  readonly "claude-haiku-4.5": {
2698
3006
  id: string;
@@ -3731,7 +4039,7 @@ export declare const MODELS: {
3731
4039
  contextWindow: number;
3732
4040
  maxTokens: number;
3733
4041
  };
3734
- readonly "gemma-4-26b-it": {
4042
+ readonly "gemma-4-26b-a4b-it": {
3735
4043
  id: string;
3736
4044
  name: string;
3737
4045
  api: "google-generative-ai";
@@ -4877,6 +5185,26 @@ export declare const MODELS: {
4877
5185
  contextWindow: number;
4878
5186
  maxTokens: number;
4879
5187
  };
5188
+ readonly "moonshotai/Kimi-K2.6": {
5189
+ id: string;
5190
+ name: string;
5191
+ api: "openai-completions";
5192
+ provider: string;
5193
+ baseUrl: string;
5194
+ compat: {
5195
+ supportsDeveloperRole: false;
5196
+ };
5197
+ reasoning: true;
5198
+ input: ("image" | "text")[];
5199
+ cost: {
5200
+ input: number;
5201
+ output: number;
5202
+ cacheRead: number;
5203
+ cacheWrite: number;
5204
+ };
5205
+ contextWindow: number;
5206
+ maxTokens: number;
5207
+ };
4880
5208
  readonly "zai-org/GLM-4.7": {
4881
5209
  id: string;
4882
5210
  name: string;
@@ -4959,6 +5287,23 @@ export declare const MODELS: {
4959
5287
  };
4960
5288
  };
4961
5289
  readonly "kimi-coding": {
5290
+ readonly k2p6: {
5291
+ id: string;
5292
+ name: string;
5293
+ api: "anthropic-messages";
5294
+ provider: string;
5295
+ baseUrl: string;
5296
+ reasoning: true;
5297
+ input: ("image" | "text")[];
5298
+ cost: {
5299
+ input: number;
5300
+ output: number;
5301
+ cacheRead: number;
5302
+ cacheWrite: number;
5303
+ };
5304
+ contextWindow: number;
5305
+ maxTokens: number;
5306
+ };
4962
5307
  readonly "kimi-for-coding": {
4963
5308
  id: string;
4964
5309
  name: string;
@@ -6576,7 +6921,109 @@ export declare const MODELS: {
6576
6921
  provider: string;
6577
6922
  baseUrl: string;
6578
6923
  reasoning: true;
6579
- input: "text"[];
6924
+ input: "text"[];
6925
+ cost: {
6926
+ input: number;
6927
+ output: number;
6928
+ cacheRead: number;
6929
+ cacheWrite: number;
6930
+ };
6931
+ contextWindow: number;
6932
+ maxTokens: number;
6933
+ };
6934
+ readonly "gpt-5": {
6935
+ id: string;
6936
+ name: string;
6937
+ api: "openai-responses";
6938
+ provider: string;
6939
+ baseUrl: string;
6940
+ reasoning: true;
6941
+ input: ("image" | "text")[];
6942
+ cost: {
6943
+ input: number;
6944
+ output: number;
6945
+ cacheRead: number;
6946
+ cacheWrite: number;
6947
+ };
6948
+ contextWindow: number;
6949
+ maxTokens: number;
6950
+ };
6951
+ readonly "gpt-5-codex": {
6952
+ id: string;
6953
+ name: string;
6954
+ api: "openai-responses";
6955
+ provider: string;
6956
+ baseUrl: string;
6957
+ reasoning: true;
6958
+ input: ("image" | "text")[];
6959
+ cost: {
6960
+ input: number;
6961
+ output: number;
6962
+ cacheRead: number;
6963
+ cacheWrite: number;
6964
+ };
6965
+ contextWindow: number;
6966
+ maxTokens: number;
6967
+ };
6968
+ readonly "gpt-5-nano": {
6969
+ id: string;
6970
+ name: string;
6971
+ api: "openai-responses";
6972
+ provider: string;
6973
+ baseUrl: string;
6974
+ reasoning: true;
6975
+ input: ("image" | "text")[];
6976
+ cost: {
6977
+ input: number;
6978
+ output: number;
6979
+ cacheRead: number;
6980
+ cacheWrite: number;
6981
+ };
6982
+ contextWindow: number;
6983
+ maxTokens: number;
6984
+ };
6985
+ readonly "gpt-5.1": {
6986
+ id: string;
6987
+ name: string;
6988
+ api: "openai-responses";
6989
+ provider: string;
6990
+ baseUrl: string;
6991
+ reasoning: true;
6992
+ input: ("image" | "text")[];
6993
+ cost: {
6994
+ input: number;
6995
+ output: number;
6996
+ cacheRead: number;
6997
+ cacheWrite: number;
6998
+ };
6999
+ contextWindow: number;
7000
+ maxTokens: number;
7001
+ };
7002
+ readonly "gpt-5.1-codex": {
7003
+ id: string;
7004
+ name: string;
7005
+ api: "openai-responses";
7006
+ provider: string;
7007
+ baseUrl: string;
7008
+ reasoning: true;
7009
+ input: ("image" | "text")[];
7010
+ cost: {
7011
+ input: number;
7012
+ output: number;
7013
+ cacheRead: number;
7014
+ cacheWrite: number;
7015
+ };
7016
+ contextWindow: number;
7017
+ maxTokens: number;
7018
+ };
7019
+ readonly "gpt-5.1-codex-max": {
7020
+ id: string;
7021
+ name: string;
7022
+ api: "openai-responses";
7023
+ provider: string;
7024
+ baseUrl: string;
7025
+ reasoning: true;
7026
+ input: ("image" | "text")[];
6580
7027
  cost: {
6581
7028
  input: number;
6582
7029
  output: number;
@@ -6586,7 +7033,7 @@ export declare const MODELS: {
6586
7033
  contextWindow: number;
6587
7034
  maxTokens: number;
6588
7035
  };
6589
- readonly "gpt-5": {
7036
+ readonly "gpt-5.1-codex-mini": {
6590
7037
  id: string;
6591
7038
  name: string;
6592
7039
  api: "openai-responses";
@@ -6603,7 +7050,7 @@ export declare const MODELS: {
6603
7050
  contextWindow: number;
6604
7051
  maxTokens: number;
6605
7052
  };
6606
- readonly "gpt-5-codex": {
7053
+ readonly "gpt-5.2": {
6607
7054
  id: string;
6608
7055
  name: string;
6609
7056
  api: "openai-responses";
@@ -6620,7 +7067,7 @@ export declare const MODELS: {
6620
7067
  contextWindow: number;
6621
7068
  maxTokens: number;
6622
7069
  };
6623
- readonly "gpt-5-nano": {
7070
+ readonly "gpt-5.2-codex": {
6624
7071
  id: string;
6625
7072
  name: string;
6626
7073
  api: "openai-responses";
@@ -6637,7 +7084,7 @@ export declare const MODELS: {
6637
7084
  contextWindow: number;
6638
7085
  maxTokens: number;
6639
7086
  };
6640
- readonly "gpt-5.1": {
7087
+ readonly "gpt-5.3-codex": {
6641
7088
  id: string;
6642
7089
  name: string;
6643
7090
  api: "openai-responses";
@@ -6654,7 +7101,7 @@ export declare const MODELS: {
6654
7101
  contextWindow: number;
6655
7102
  maxTokens: number;
6656
7103
  };
6657
- readonly "gpt-5.1-codex": {
7104
+ readonly "gpt-5.4": {
6658
7105
  id: string;
6659
7106
  name: string;
6660
7107
  api: "openai-responses";
@@ -6671,7 +7118,7 @@ export declare const MODELS: {
6671
7118
  contextWindow: number;
6672
7119
  maxTokens: number;
6673
7120
  };
6674
- readonly "gpt-5.1-codex-max": {
7121
+ readonly "gpt-5.4-mini": {
6675
7122
  id: string;
6676
7123
  name: string;
6677
7124
  api: "openai-responses";
@@ -6688,7 +7135,7 @@ export declare const MODELS: {
6688
7135
  contextWindow: number;
6689
7136
  maxTokens: number;
6690
7137
  };
6691
- readonly "gpt-5.1-codex-mini": {
7138
+ readonly "gpt-5.4-nano": {
6692
7139
  id: string;
6693
7140
  name: string;
6694
7141
  api: "openai-responses";
@@ -6705,7 +7152,7 @@ export declare const MODELS: {
6705
7152
  contextWindow: number;
6706
7153
  maxTokens: number;
6707
7154
  };
6708
- readonly "gpt-5.2": {
7155
+ readonly "gpt-5.4-pro": {
6709
7156
  id: string;
6710
7157
  name: string;
6711
7158
  api: "openai-responses";
@@ -6722,10 +7169,10 @@ export declare const MODELS: {
6722
7169
  contextWindow: number;
6723
7170
  maxTokens: number;
6724
7171
  };
6725
- readonly "gpt-5.2-codex": {
7172
+ readonly "kimi-k2.5": {
6726
7173
  id: string;
6727
7174
  name: string;
6728
- api: "openai-responses";
7175
+ api: "openai-completions";
6729
7176
  provider: string;
6730
7177
  baseUrl: string;
6731
7178
  reasoning: true;
@@ -6739,10 +7186,10 @@ export declare const MODELS: {
6739
7186
  contextWindow: number;
6740
7187
  maxTokens: number;
6741
7188
  };
6742
- readonly "gpt-5.3-codex": {
7189
+ readonly "kimi-k2.6": {
6743
7190
  id: string;
6744
7191
  name: string;
6745
- api: "openai-responses";
7192
+ api: "openai-completions";
6746
7193
  provider: string;
6747
7194
  baseUrl: string;
6748
7195
  reasoning: true;
@@ -6756,14 +7203,14 @@ export declare const MODELS: {
6756
7203
  contextWindow: number;
6757
7204
  maxTokens: number;
6758
7205
  };
6759
- readonly "gpt-5.4": {
7206
+ readonly "ling-2.6-flash-free": {
6760
7207
  id: string;
6761
7208
  name: string;
6762
- api: "openai-responses";
7209
+ api: "openai-completions";
6763
7210
  provider: string;
6764
7211
  baseUrl: string;
6765
- reasoning: true;
6766
- input: ("image" | "text")[];
7212
+ reasoning: false;
7213
+ input: "text"[];
6767
7214
  cost: {
6768
7215
  input: number;
6769
7216
  output: number;
@@ -6773,14 +7220,14 @@ export declare const MODELS: {
6773
7220
  contextWindow: number;
6774
7221
  maxTokens: number;
6775
7222
  };
6776
- readonly "gpt-5.4-mini": {
7223
+ readonly "minimax-m2.5": {
6777
7224
  id: string;
6778
7225
  name: string;
6779
- api: "openai-responses";
7226
+ api: "openai-completions";
6780
7227
  provider: string;
6781
7228
  baseUrl: string;
6782
7229
  reasoning: true;
6783
- input: ("image" | "text")[];
7230
+ input: "text"[];
6784
7231
  cost: {
6785
7232
  input: number;
6786
7233
  output: number;
@@ -6790,14 +7237,14 @@ export declare const MODELS: {
6790
7237
  contextWindow: number;
6791
7238
  maxTokens: number;
6792
7239
  };
6793
- readonly "gpt-5.4-nano": {
7240
+ readonly "minimax-m2.5-free": {
6794
7241
  id: string;
6795
7242
  name: string;
6796
- api: "openai-responses";
7243
+ api: "anthropic-messages";
6797
7244
  provider: string;
6798
7245
  baseUrl: string;
6799
7246
  reasoning: true;
6800
- input: ("image" | "text")[];
7247
+ input: "text"[];
6801
7248
  cost: {
6802
7249
  input: number;
6803
7250
  output: number;
@@ -6807,14 +7254,14 @@ export declare const MODELS: {
6807
7254
  contextWindow: number;
6808
7255
  maxTokens: number;
6809
7256
  };
6810
- readonly "gpt-5.4-pro": {
7257
+ readonly "minimax-m2.7": {
6811
7258
  id: string;
6812
7259
  name: string;
6813
- api: "openai-responses";
7260
+ api: "openai-completions";
6814
7261
  provider: string;
6815
7262
  baseUrl: string;
6816
7263
  reasoning: true;
6817
- input: ("image" | "text")[];
7264
+ input: "text"[];
6818
7265
  cost: {
6819
7266
  input: number;
6820
7267
  output: number;
@@ -6824,14 +7271,14 @@ export declare const MODELS: {
6824
7271
  contextWindow: number;
6825
7272
  maxTokens: number;
6826
7273
  };
6827
- readonly "kimi-k2.5": {
7274
+ readonly "nemotron-3-super-free": {
6828
7275
  id: string;
6829
7276
  name: string;
6830
7277
  api: "openai-completions";
6831
7278
  provider: string;
6832
7279
  baseUrl: string;
6833
7280
  reasoning: true;
6834
- input: ("image" | "text")[];
7281
+ input: "text"[];
6835
7282
  cost: {
6836
7283
  input: number;
6837
7284
  output: number;
@@ -6841,14 +7288,14 @@ export declare const MODELS: {
6841
7288
  contextWindow: number;
6842
7289
  maxTokens: number;
6843
7290
  };
6844
- readonly "minimax-m2.5": {
7291
+ readonly "qwen3.5-plus": {
6845
7292
  id: string;
6846
7293
  name: string;
6847
- api: "openai-completions";
7294
+ api: "anthropic-messages";
6848
7295
  provider: string;
6849
7296
  baseUrl: string;
6850
7297
  reasoning: true;
6851
- input: "text"[];
7298
+ input: ("image" | "text")[];
6852
7299
  cost: {
6853
7300
  input: number;
6854
7301
  output: number;
@@ -6858,14 +7305,14 @@ export declare const MODELS: {
6858
7305
  contextWindow: number;
6859
7306
  maxTokens: number;
6860
7307
  };
6861
- readonly "minimax-m2.5-free": {
7308
+ readonly "qwen3.6-plus": {
6862
7309
  id: string;
6863
7310
  name: string;
6864
7311
  api: "anthropic-messages";
6865
7312
  provider: string;
6866
7313
  baseUrl: string;
6867
7314
  reasoning: true;
6868
- input: "text"[];
7315
+ input: ("image" | "text")[];
6869
7316
  cost: {
6870
7317
  input: number;
6871
7318
  output: number;
@@ -6875,7 +7322,9 @@ export declare const MODELS: {
6875
7322
  contextWindow: number;
6876
7323
  maxTokens: number;
6877
7324
  };
6878
- readonly "nemotron-3-super-free": {
7325
+ };
7326
+ readonly "opencode-go": {
7327
+ readonly "glm-5": {
6879
7328
  id: string;
6880
7329
  name: string;
6881
7330
  api: "openai-completions";
@@ -6892,14 +7341,14 @@ export declare const MODELS: {
6892
7341
  contextWindow: number;
6893
7342
  maxTokens: number;
6894
7343
  };
6895
- readonly "qwen3.5-plus": {
7344
+ readonly "glm-5.1": {
6896
7345
  id: string;
6897
7346
  name: string;
6898
7347
  api: "openai-completions";
6899
7348
  provider: string;
6900
7349
  baseUrl: string;
6901
7350
  reasoning: true;
6902
- input: ("image" | "text")[];
7351
+ input: "text"[];
6903
7352
  cost: {
6904
7353
  input: number;
6905
7354
  output: number;
@@ -6909,7 +7358,7 @@ export declare const MODELS: {
6909
7358
  contextWindow: number;
6910
7359
  maxTokens: number;
6911
7360
  };
6912
- readonly "qwen3.6-plus": {
7361
+ readonly "kimi-k2.5": {
6913
7362
  id: string;
6914
7363
  name: string;
6915
7364
  api: "openai-completions";
@@ -6926,16 +7375,14 @@ export declare const MODELS: {
6926
7375
  contextWindow: number;
6927
7376
  maxTokens: number;
6928
7377
  };
6929
- };
6930
- readonly "opencode-go": {
6931
- readonly "glm-5": {
7378
+ readonly "kimi-k2.6": {
6932
7379
  id: string;
6933
7380
  name: string;
6934
7381
  api: "openai-completions";
6935
7382
  provider: string;
6936
7383
  baseUrl: string;
6937
7384
  reasoning: true;
6938
- input: "text"[];
7385
+ input: ("image" | "text")[];
6939
7386
  cost: {
6940
7387
  input: number;
6941
7388
  output: number;
@@ -6945,14 +7392,14 @@ export declare const MODELS: {
6945
7392
  contextWindow: number;
6946
7393
  maxTokens: number;
6947
7394
  };
6948
- readonly "glm-5.1": {
7395
+ readonly "mimo-v2-omni": {
6949
7396
  id: string;
6950
7397
  name: string;
6951
7398
  api: "openai-completions";
6952
7399
  provider: string;
6953
7400
  baseUrl: string;
6954
7401
  reasoning: true;
6955
- input: "text"[];
7402
+ input: ("image" | "text")[];
6956
7403
  cost: {
6957
7404
  input: number;
6958
7405
  output: number;
@@ -6962,14 +7409,14 @@ export declare const MODELS: {
6962
7409
  contextWindow: number;
6963
7410
  maxTokens: number;
6964
7411
  };
6965
- readonly "kimi-k2.5": {
7412
+ readonly "mimo-v2-pro": {
6966
7413
  id: string;
6967
7414
  name: string;
6968
7415
  api: "openai-completions";
6969
7416
  provider: string;
6970
7417
  baseUrl: string;
6971
7418
  reasoning: true;
6972
- input: ("image" | "text")[];
7419
+ input: "text"[];
6973
7420
  cost: {
6974
7421
  input: number;
6975
7422
  output: number;
@@ -6979,7 +7426,7 @@ export declare const MODELS: {
6979
7426
  contextWindow: number;
6980
7427
  maxTokens: number;
6981
7428
  };
6982
- readonly "mimo-v2-omni": {
7429
+ readonly "mimo-v2.5": {
6983
7430
  id: string;
6984
7431
  name: string;
6985
7432
  api: "openai-completions";
@@ -6996,7 +7443,7 @@ export declare const MODELS: {
6996
7443
  contextWindow: number;
6997
7444
  maxTokens: number;
6998
7445
  };
6999
- readonly "mimo-v2-pro": {
7446
+ readonly "mimo-v2.5-pro": {
7000
7447
  id: string;
7001
7448
  name: string;
7002
7449
  api: "openai-completions";
@@ -7050,7 +7497,7 @@ export declare const MODELS: {
7050
7497
  readonly "qwen3.5-plus": {
7051
7498
  id: string;
7052
7499
  name: string;
7053
- api: "openai-completions";
7500
+ api: "anthropic-messages";
7054
7501
  provider: string;
7055
7502
  baseUrl: string;
7056
7503
  reasoning: true;
@@ -7067,7 +7514,7 @@ export declare const MODELS: {
7067
7514
  readonly "qwen3.6-plus": {
7068
7515
  id: string;
7069
7516
  name: string;
7070
- api: "openai-completions";
7517
+ api: "anthropic-messages";
7071
7518
  provider: string;
7072
7519
  baseUrl: string;
7073
7520
  reasoning: true;
@@ -7457,7 +7904,7 @@ export declare const MODELS: {
7457
7904
  contextWindow: number;
7458
7905
  maxTokens: number;
7459
7906
  };
7460
- readonly "arcee-ai/trinity-large-preview:free": {
7907
+ readonly "arcee-ai/trinity-large-preview": {
7461
7908
  id: string;
7462
7909
  name: string;
7463
7910
  api: "openai-completions";
@@ -8120,6 +8567,23 @@ export declare const MODELS: {
8120
8567
  contextWindow: number;
8121
8568
  maxTokens: number;
8122
8569
  };
8570
+ readonly "inclusionai/ling-2.6-flash:free": {
8571
+ id: string;
8572
+ name: string;
8573
+ api: "openai-completions";
8574
+ provider: string;
8575
+ baseUrl: string;
8576
+ reasoning: false;
8577
+ input: "text"[];
8578
+ cost: {
8579
+ input: number;
8580
+ output: number;
8581
+ cacheRead: number;
8582
+ cacheWrite: number;
8583
+ };
8584
+ contextWindow: number;
8585
+ maxTokens: number;
8586
+ };
8123
8587
  readonly "kwaipilot/kat-coder-pro-v2": {
8124
8588
  id: string;
8125
8589
  name: string;
@@ -8783,6 +9247,23 @@ export declare const MODELS: {
8783
9247
  contextWindow: number;
8784
9248
  maxTokens: number;
8785
9249
  };
9250
+ readonly "moonshotai/kimi-k2.6": {
9251
+ id: string;
9252
+ name: string;
9253
+ api: "openai-completions";
9254
+ provider: string;
9255
+ baseUrl: string;
9256
+ reasoning: true;
9257
+ input: ("image" | "text")[];
9258
+ cost: {
9259
+ input: number;
9260
+ output: number;
9261
+ cacheRead: number;
9262
+ cacheWrite: number;
9263
+ };
9264
+ contextWindow: number;
9265
+ maxTokens: number;
9266
+ };
8786
9267
  readonly "nex-agi/deepseek-v3.1-nex-n1": {
8787
9268
  id: string;
8788
9269
  name: string;
@@ -9293,40 +9774,6 @@ export declare const MODELS: {
9293
9774
  contextWindow: number;
9294
9775
  maxTokens: number;
9295
9776
  };
9296
- readonly "openai/gpt-5-image": {
9297
- id: string;
9298
- name: string;
9299
- api: "openai-completions";
9300
- provider: string;
9301
- baseUrl: string;
9302
- reasoning: true;
9303
- input: ("image" | "text")[];
9304
- cost: {
9305
- input: number;
9306
- output: number;
9307
- cacheRead: number;
9308
- cacheWrite: number;
9309
- };
9310
- contextWindow: number;
9311
- maxTokens: number;
9312
- };
9313
- readonly "openai/gpt-5-image-mini": {
9314
- id: string;
9315
- name: string;
9316
- api: "openai-completions";
9317
- provider: string;
9318
- baseUrl: string;
9319
- reasoning: true;
9320
- input: ("image" | "text")[];
9321
- cost: {
9322
- input: number;
9323
- output: number;
9324
- cacheRead: number;
9325
- cacheWrite: number;
9326
- };
9327
- contextWindow: number;
9328
- maxTokens: number;
9329
- };
9330
9777
  readonly "openai/gpt-5-mini": {
9331
9778
  id: string;
9332
9779
  name: string;
@@ -9922,23 +10369,6 @@ export declare const MODELS: {
9922
10369
  contextWindow: number;
9923
10370
  maxTokens: number;
9924
10371
  };
9925
- readonly "openrouter/elephant-alpha": {
9926
- id: string;
9927
- name: string;
9928
- api: "openai-completions";
9929
- provider: string;
9930
- baseUrl: string;
9931
- reasoning: false;
9932
- input: "text"[];
9933
- cost: {
9934
- input: number;
9935
- output: number;
9936
- cacheRead: number;
9937
- cacheWrite: number;
9938
- };
9939
- contextWindow: number;
9940
- maxTokens: number;
9941
- };
9942
10372
  readonly "openrouter/free": {
9943
10373
  id: string;
9944
10374
  name: string;
@@ -11078,6 +11508,40 @@ export declare const MODELS: {
11078
11508
  contextWindow: number;
11079
11509
  maxTokens: number;
11080
11510
  };
11511
+ readonly "xiaomi/mimo-v2.5": {
11512
+ id: string;
11513
+ name: string;
11514
+ api: "openai-completions";
11515
+ provider: string;
11516
+ baseUrl: string;
11517
+ reasoning: true;
11518
+ input: ("image" | "text")[];
11519
+ cost: {
11520
+ input: number;
11521
+ output: number;
11522
+ cacheRead: number;
11523
+ cacheWrite: number;
11524
+ };
11525
+ contextWindow: number;
11526
+ maxTokens: number;
11527
+ };
11528
+ readonly "xiaomi/mimo-v2.5-pro": {
11529
+ id: string;
11530
+ name: string;
11531
+ api: "openai-completions";
11532
+ provider: string;
11533
+ baseUrl: string;
11534
+ reasoning: true;
11535
+ input: "text"[];
11536
+ cost: {
11537
+ input: number;
11538
+ output: number;
11539
+ cacheRead: number;
11540
+ cacheWrite: number;
11541
+ };
11542
+ contextWindow: number;
11543
+ maxTokens: number;
11544
+ };
11081
11545
  readonly "z-ai/glm-4-32b": {
11082
11546
  id: string;
11083
11547
  name: string;
@@ -11299,6 +11763,23 @@ export declare const MODELS: {
11299
11763
  contextWindow: number;
11300
11764
  maxTokens: number;
11301
11765
  };
11766
+ readonly "~anthropic/claude-opus-latest": {
11767
+ id: string;
11768
+ name: string;
11769
+ api: "openai-completions";
11770
+ provider: string;
11771
+ baseUrl: string;
11772
+ reasoning: true;
11773
+ input: ("image" | "text")[];
11774
+ cost: {
11775
+ input: number;
11776
+ output: number;
11777
+ cacheRead: number;
11778
+ cacheWrite: number;
11779
+ };
11780
+ contextWindow: number;
11781
+ maxTokens: number;
11782
+ };
11302
11783
  };
11303
11784
  readonly "vercel-ai-gateway": {
11304
11785
  readonly "alibaba/qwen-3-14b": {
@@ -11369,6 +11850,23 @@ export declare const MODELS: {
11369
11850
  contextWindow: number;
11370
11851
  maxTokens: number;
11371
11852
  };
11853
+ readonly "alibaba/qwen-3.6-max-preview": {
11854
+ id: string;
11855
+ name: string;
11856
+ api: "anthropic-messages";
11857
+ provider: string;
11858
+ baseUrl: string;
11859
+ reasoning: true;
11860
+ input: ("image" | "text")[];
11861
+ cost: {
11862
+ input: number;
11863
+ output: number;
11864
+ cacheRead: number;
11865
+ cacheWrite: number;
11866
+ };
11867
+ contextWindow: number;
11868
+ maxTokens: number;
11869
+ };
11372
11870
  readonly "alibaba/qwen3-235b-a22b-thinking": {
11373
11871
  id: string;
11374
11872
  name: string;
@@ -12712,6 +13210,23 @@ export declare const MODELS: {
12712
13210
  contextWindow: number;
12713
13211
  maxTokens: number;
12714
13212
  };
13213
+ readonly "moonshotai/kimi-k2.6": {
13214
+ id: string;
13215
+ name: string;
13216
+ api: "anthropic-messages";
13217
+ provider: string;
13218
+ baseUrl: string;
13219
+ reasoning: true;
13220
+ input: ("image" | "text")[];
13221
+ cost: {
13222
+ input: number;
13223
+ output: number;
13224
+ cacheRead: number;
13225
+ cacheWrite: number;
13226
+ };
13227
+ contextWindow: number;
13228
+ maxTokens: number;
13229
+ };
12715
13230
  readonly "nvidia/nemotron-nano-12b-v2-vl": {
12716
13231
  id: string;
12717
13232
  name: string;
@@ -13892,7 +14407,7 @@ export declare const MODELS: {
13892
14407
  provider: string;
13893
14408
  baseUrl: string;
13894
14409
  reasoning: true;
13895
- input: ("image" | "text")[];
14410
+ input: "text"[];
13896
14411
  cost: {
13897
14412
  input: number;
13898
14413
  output: number;