@hyperspaceng/neural-ai 0.70.1 → 0.70.6

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (41) hide show
  1. package/README.md +9 -4
  2. package/dist/env-api-keys.d.ts.map +1 -1
  3. package/dist/env-api-keys.js +51 -7
  4. package/dist/env-api-keys.js.map +1 -1
  5. package/dist/models.d.ts +1 -0
  6. package/dist/models.d.ts.map +1 -1
  7. package/dist/models.generated.d.ts +606 -249
  8. package/dist/models.generated.d.ts.map +1 -1
  9. package/dist/models.generated.js +539 -204
  10. package/dist/models.generated.js.map +1 -1
  11. package/dist/models.js +3 -1
  12. package/dist/models.js.map +1 -1
  13. package/dist/providers/amazon-bedrock.d.ts.map +1 -1
  14. package/dist/providers/amazon-bedrock.js +49 -25
  15. package/dist/providers/amazon-bedrock.js.map +1 -1
  16. package/dist/providers/anthropic.d.ts.map +1 -1
  17. package/dist/providers/anthropic.js +18 -7
  18. package/dist/providers/anthropic.js.map +1 -1
  19. package/dist/providers/azure-openai-responses.d.ts.map +1 -1
  20. package/dist/providers/azure-openai-responses.js +24 -5
  21. package/dist/providers/azure-openai-responses.js.map +1 -1
  22. package/dist/providers/cloudflare.d.ts +7 -0
  23. package/dist/providers/cloudflare.d.ts.map +1 -0
  24. package/dist/providers/cloudflare.js +19 -0
  25. package/dist/providers/cloudflare.js.map +1 -0
  26. package/dist/providers/openai-codex-responses.d.ts.map +1 -1
  27. package/dist/providers/openai-codex-responses.js +2 -2
  28. package/dist/providers/openai-codex-responses.js.map +1 -1
  29. package/dist/providers/openai-completions.d.ts.map +1 -1
  30. package/dist/providers/openai-completions.js +49 -16
  31. package/dist/providers/openai-completions.js.map +1 -1
  32. package/dist/providers/openai-responses.d.ts.map +1 -1
  33. package/dist/providers/openai-responses.js +7 -4
  34. package/dist/providers/openai-responses.js.map +1 -1
  35. package/dist/providers/simple-options.d.ts.map +1 -1
  36. package/dist/providers/simple-options.js +2 -0
  37. package/dist/providers/simple-options.js.map +1 -1
  38. package/dist/types.d.ts +15 -3
  39. package/dist/types.d.ts.map +1 -1
  40. package/dist/types.js.map +1 -1
  41. package/package.json +1 -1
@@ -2537,6 +2537,23 @@ export declare const MODELS: {
2537
2537
  contextWindow: number;
2538
2538
  maxTokens: number;
2539
2539
  };
2540
+ readonly "gpt-5.5-pro": {
2541
+ id: string;
2542
+ name: string;
2543
+ api: "azure-openai-responses";
2544
+ provider: string;
2545
+ baseUrl: string;
2546
+ reasoning: true;
2547
+ input: ("image" | "text")[];
2548
+ cost: {
2549
+ input: number;
2550
+ output: number;
2551
+ cacheRead: number;
2552
+ cacheWrite: number;
2553
+ };
2554
+ contextWindow: number;
2555
+ maxTokens: number;
2556
+ };
2540
2557
  readonly o1: {
2541
2558
  id: string;
2542
2559
  name: string;
@@ -2744,6 +2761,226 @@ export declare const MODELS: {
2744
2761
  maxTokens: number;
2745
2762
  };
2746
2763
  };
2764
+ readonly "cloudflare-workers-ai": {
2765
+ readonly "@cf/google/gemma-4-26b-a4b-it": {
2766
+ id: string;
2767
+ name: string;
2768
+ api: "openai-completions";
2769
+ provider: string;
2770
+ baseUrl: string;
2771
+ compat: {
2772
+ sendSessionAffinityHeaders: true;
2773
+ };
2774
+ reasoning: true;
2775
+ input: ("image" | "text")[];
2776
+ cost: {
2777
+ input: number;
2778
+ output: number;
2779
+ cacheRead: number;
2780
+ cacheWrite: number;
2781
+ };
2782
+ contextWindow: number;
2783
+ maxTokens: number;
2784
+ };
2785
+ readonly "@cf/meta/llama-4-scout-17b-16e-instruct": {
2786
+ id: string;
2787
+ name: string;
2788
+ api: "openai-completions";
2789
+ provider: string;
2790
+ baseUrl: string;
2791
+ compat: {
2792
+ sendSessionAffinityHeaders: true;
2793
+ };
2794
+ reasoning: false;
2795
+ input: ("image" | "text")[];
2796
+ cost: {
2797
+ input: number;
2798
+ output: number;
2799
+ cacheRead: number;
2800
+ cacheWrite: number;
2801
+ };
2802
+ contextWindow: number;
2803
+ maxTokens: number;
2804
+ };
2805
+ readonly "@cf/moonshotai/kimi-k2.5": {
2806
+ id: string;
2807
+ name: string;
2808
+ api: "openai-completions";
2809
+ provider: string;
2810
+ baseUrl: string;
2811
+ compat: {
2812
+ sendSessionAffinityHeaders: true;
2813
+ };
2814
+ reasoning: true;
2815
+ input: ("image" | "text")[];
2816
+ cost: {
2817
+ input: number;
2818
+ output: number;
2819
+ cacheRead: number;
2820
+ cacheWrite: number;
2821
+ };
2822
+ contextWindow: number;
2823
+ maxTokens: number;
2824
+ };
2825
+ readonly "@cf/moonshotai/kimi-k2.6": {
2826
+ id: string;
2827
+ name: string;
2828
+ api: "openai-completions";
2829
+ provider: string;
2830
+ baseUrl: string;
2831
+ compat: {
2832
+ sendSessionAffinityHeaders: true;
2833
+ };
2834
+ reasoning: true;
2835
+ input: ("image" | "text")[];
2836
+ cost: {
2837
+ input: number;
2838
+ output: number;
2839
+ cacheRead: number;
2840
+ cacheWrite: number;
2841
+ };
2842
+ contextWindow: number;
2843
+ maxTokens: number;
2844
+ };
2845
+ readonly "@cf/nvidia/nemotron-3-120b-a12b": {
2846
+ id: string;
2847
+ name: string;
2848
+ api: "openai-completions";
2849
+ provider: string;
2850
+ baseUrl: string;
2851
+ compat: {
2852
+ sendSessionAffinityHeaders: true;
2853
+ };
2854
+ reasoning: true;
2855
+ input: "text"[];
2856
+ cost: {
2857
+ input: number;
2858
+ output: number;
2859
+ cacheRead: number;
2860
+ cacheWrite: number;
2861
+ };
2862
+ contextWindow: number;
2863
+ maxTokens: number;
2864
+ };
2865
+ readonly "@cf/openai/gpt-oss-120b": {
2866
+ id: string;
2867
+ name: string;
2868
+ api: "openai-completions";
2869
+ provider: string;
2870
+ baseUrl: string;
2871
+ compat: {
2872
+ sendSessionAffinityHeaders: true;
2873
+ };
2874
+ reasoning: true;
2875
+ input: "text"[];
2876
+ cost: {
2877
+ input: number;
2878
+ output: number;
2879
+ cacheRead: number;
2880
+ cacheWrite: number;
2881
+ };
2882
+ contextWindow: number;
2883
+ maxTokens: number;
2884
+ };
2885
+ readonly "@cf/openai/gpt-oss-20b": {
2886
+ id: string;
2887
+ name: string;
2888
+ api: "openai-completions";
2889
+ provider: string;
2890
+ baseUrl: string;
2891
+ compat: {
2892
+ sendSessionAffinityHeaders: true;
2893
+ };
2894
+ reasoning: true;
2895
+ input: "text"[];
2896
+ cost: {
2897
+ input: number;
2898
+ output: number;
2899
+ cacheRead: number;
2900
+ cacheWrite: number;
2901
+ };
2902
+ contextWindow: number;
2903
+ maxTokens: number;
2904
+ };
2905
+ readonly "@cf/zai-org/glm-4.7-flash": {
2906
+ id: string;
2907
+ name: string;
2908
+ api: "openai-completions";
2909
+ provider: string;
2910
+ baseUrl: string;
2911
+ compat: {
2912
+ sendSessionAffinityHeaders: true;
2913
+ };
2914
+ reasoning: true;
2915
+ input: "text"[];
2916
+ cost: {
2917
+ input: number;
2918
+ output: number;
2919
+ cacheRead: number;
2920
+ cacheWrite: number;
2921
+ };
2922
+ contextWindow: number;
2923
+ maxTokens: number;
2924
+ };
2925
+ };
2926
+ readonly deepseek: {
2927
+ readonly "deepseek-v4-flash": {
2928
+ id: string;
2929
+ name: string;
2930
+ api: "openai-completions";
2931
+ provider: string;
2932
+ baseUrl: string;
2933
+ compat: {
2934
+ requiresReasoningContentOnAssistantMessages: true;
2935
+ thinkingFormat: "deepseek";
2936
+ reasoningEffortMap: {
2937
+ minimal: string;
2938
+ low: string;
2939
+ medium: string;
2940
+ high: string;
2941
+ xhigh: string;
2942
+ };
2943
+ };
2944
+ reasoning: true;
2945
+ input: "text"[];
2946
+ cost: {
2947
+ input: number;
2948
+ output: number;
2949
+ cacheRead: number;
2950
+ cacheWrite: number;
2951
+ };
2952
+ contextWindow: number;
2953
+ maxTokens: number;
2954
+ };
2955
+ readonly "deepseek-v4-pro": {
2956
+ id: string;
2957
+ name: string;
2958
+ api: "openai-completions";
2959
+ provider: string;
2960
+ baseUrl: string;
2961
+ compat: {
2962
+ requiresReasoningContentOnAssistantMessages: true;
2963
+ thinkingFormat: "deepseek";
2964
+ reasoningEffortMap: {
2965
+ minimal: string;
2966
+ low: string;
2967
+ medium: string;
2968
+ high: string;
2969
+ xhigh: string;
2970
+ };
2971
+ };
2972
+ reasoning: true;
2973
+ input: "text"[];
2974
+ cost: {
2975
+ input: number;
2976
+ output: number;
2977
+ cacheRead: number;
2978
+ cacheWrite: number;
2979
+ };
2980
+ contextWindow: number;
2981
+ maxTokens: number;
2982
+ };
2983
+ };
2747
2984
  readonly fireworks: {
2748
2985
  readonly "accounts/fireworks/models/deepseek-v3p1": {
2749
2986
  id: string;
@@ -2779,6 +3016,23 @@ export declare const MODELS: {
2779
3016
  contextWindow: number;
2780
3017
  maxTokens: number;
2781
3018
  };
3019
+ readonly "accounts/fireworks/models/deepseek-v4-pro": {
3020
+ id: string;
3021
+ name: string;
3022
+ api: "anthropic-messages";
3023
+ provider: string;
3024
+ baseUrl: string;
3025
+ reasoning: true;
3026
+ input: "text"[];
3027
+ cost: {
3028
+ input: number;
3029
+ output: number;
3030
+ cacheRead: number;
3031
+ cacheWrite: number;
3032
+ };
3033
+ contextWindow: number;
3034
+ maxTokens: number;
3035
+ };
2782
3036
  readonly "accounts/fireworks/models/glm-4p5": {
2783
3037
  id: string;
2784
3038
  name: string;
@@ -3644,6 +3898,29 @@ export declare const MODELS: {
3644
3898
  contextWindow: number;
3645
3899
  maxTokens: number;
3646
3900
  };
3901
+ readonly "gpt-5.5": {
3902
+ id: string;
3903
+ name: string;
3904
+ api: "openai-responses";
3905
+ provider: string;
3906
+ baseUrl: string;
3907
+ headers: {
3908
+ "User-Agent": string;
3909
+ "Editor-Version": string;
3910
+ "Editor-Plugin-Version": string;
3911
+ "Copilot-Integration-Id": string;
3912
+ };
3913
+ reasoning: true;
3914
+ input: ("image" | "text")[];
3915
+ cost: {
3916
+ input: number;
3917
+ output: number;
3918
+ cacheRead: number;
3919
+ cacheWrite: number;
3920
+ };
3921
+ contextWindow: number;
3922
+ maxTokens: number;
3923
+ };
3647
3924
  readonly "grok-code-fast-1": {
3648
3925
  id: string;
3649
3926
  name: string;
@@ -5182,6 +5459,26 @@ export declare const MODELS: {
5182
5459
  contextWindow: number;
5183
5460
  maxTokens: number;
5184
5461
  };
5462
+ readonly "deepseek-ai/DeepSeek-V4-Pro": {
5463
+ id: string;
5464
+ name: string;
5465
+ api: "openai-completions";
5466
+ provider: string;
5467
+ baseUrl: string;
5468
+ compat: {
5469
+ supportsDeveloperRole: false;
5470
+ };
5471
+ reasoning: true;
5472
+ input: "text"[];
5473
+ cost: {
5474
+ input: number;
5475
+ output: number;
5476
+ cacheRead: number;
5477
+ cacheWrite: number;
5478
+ };
5479
+ contextWindow: number;
5480
+ maxTokens: number;
5481
+ };
5185
5482
  readonly "moonshotai/Kimi-K2-Instruct": {
5186
5483
  id: string;
5187
5484
  name: string;
@@ -6373,7 +6670,24 @@ export declare const MODELS: {
6373
6670
  api: "openai-responses";
6374
6671
  provider: string;
6375
6672
  baseUrl: string;
6376
- reasoning: false;
6673
+ reasoning: false;
6674
+ input: ("image" | "text")[];
6675
+ cost: {
6676
+ input: number;
6677
+ output: number;
6678
+ cacheRead: number;
6679
+ cacheWrite: number;
6680
+ };
6681
+ contextWindow: number;
6682
+ maxTokens: number;
6683
+ };
6684
+ readonly "gpt-5.3-codex": {
6685
+ id: string;
6686
+ name: string;
6687
+ api: "openai-responses";
6688
+ provider: string;
6689
+ baseUrl: string;
6690
+ reasoning: true;
6377
6691
  input: ("image" | "text")[];
6378
6692
  cost: {
6379
6693
  input: number;
@@ -6384,7 +6698,7 @@ export declare const MODELS: {
6384
6698
  contextWindow: number;
6385
6699
  maxTokens: number;
6386
6700
  };
6387
- readonly "gpt-5.3-codex": {
6701
+ readonly "gpt-5.3-codex-spark": {
6388
6702
  id: string;
6389
6703
  name: string;
6390
6704
  api: "openai-responses";
@@ -6401,7 +6715,7 @@ export declare const MODELS: {
6401
6715
  contextWindow: number;
6402
6716
  maxTokens: number;
6403
6717
  };
6404
- readonly "gpt-5.3-codex-spark": {
6718
+ readonly "gpt-5.4": {
6405
6719
  id: string;
6406
6720
  name: string;
6407
6721
  api: "openai-responses";
@@ -6418,7 +6732,7 @@ export declare const MODELS: {
6418
6732
  contextWindow: number;
6419
6733
  maxTokens: number;
6420
6734
  };
6421
- readonly "gpt-5.4": {
6735
+ readonly "gpt-5.4-mini": {
6422
6736
  id: string;
6423
6737
  name: string;
6424
6738
  api: "openai-responses";
@@ -6435,7 +6749,7 @@ export declare const MODELS: {
6435
6749
  contextWindow: number;
6436
6750
  maxTokens: number;
6437
6751
  };
6438
- readonly "gpt-5.4-mini": {
6752
+ readonly "gpt-5.4-nano": {
6439
6753
  id: string;
6440
6754
  name: string;
6441
6755
  api: "openai-responses";
@@ -6452,7 +6766,7 @@ export declare const MODELS: {
6452
6766
  contextWindow: number;
6453
6767
  maxTokens: number;
6454
6768
  };
6455
- readonly "gpt-5.4-nano": {
6769
+ readonly "gpt-5.4-pro": {
6456
6770
  id: string;
6457
6771
  name: string;
6458
6772
  api: "openai-responses";
@@ -6469,7 +6783,7 @@ export declare const MODELS: {
6469
6783
  contextWindow: number;
6470
6784
  maxTokens: number;
6471
6785
  };
6472
- readonly "gpt-5.4-pro": {
6786
+ readonly "gpt-5.5": {
6473
6787
  id: string;
6474
6788
  name: string;
6475
6789
  api: "openai-responses";
@@ -6486,7 +6800,7 @@ export declare const MODELS: {
6486
6800
  contextWindow: number;
6487
6801
  maxTokens: number;
6488
6802
  };
6489
- readonly "gpt-5.5": {
6803
+ readonly "gpt-5.5-pro": {
6490
6804
  id: string;
6491
6805
  name: string;
6492
6806
  api: "openai-responses";
@@ -6830,23 +7144,6 @@ export declare const MODELS: {
6830
7144
  contextWindow: number;
6831
7145
  maxTokens: number;
6832
7146
  };
6833
- readonly "claude-3-5-haiku": {
6834
- id: string;
6835
- name: string;
6836
- api: "anthropic-messages";
6837
- provider: string;
6838
- baseUrl: string;
6839
- reasoning: false;
6840
- input: ("image" | "text")[];
6841
- cost: {
6842
- input: number;
6843
- output: number;
6844
- cacheRead: number;
6845
- cacheWrite: number;
6846
- };
6847
- contextWindow: number;
6848
- maxTokens: number;
6849
- };
6850
7147
  readonly "claude-haiku-4-5": {
6851
7148
  id: string;
6852
7149
  name: string;
@@ -7501,6 +7798,9 @@ export declare const MODELS: {
7501
7798
  api: "openai-completions";
7502
7799
  provider: string;
7503
7800
  baseUrl: string;
7801
+ compat: {
7802
+ requiresReasoningContentOnAssistantMessages: true;
7803
+ };
7504
7804
  reasoning: true;
7505
7805
  input: "text"[];
7506
7806
  cost: {
@@ -7518,6 +7818,9 @@ export declare const MODELS: {
7518
7818
  api: "openai-completions";
7519
7819
  provider: string;
7520
7820
  baseUrl: string;
7821
+ compat: {
7822
+ requiresReasoningContentOnAssistantMessages: true;
7823
+ };
7521
7824
  reasoning: true;
7522
7825
  input: "text"[];
7523
7826
  cost: {
@@ -8472,6 +8775,9 @@ export declare const MODELS: {
8472
8775
  api: "openai-completions";
8473
8776
  provider: string;
8474
8777
  baseUrl: string;
8778
+ compat: {
8779
+ requiresReasoningContentOnAssistantMessages: true;
8780
+ };
8475
8781
  reasoning: true;
8476
8782
  input: "text"[];
8477
8783
  cost: {
@@ -8489,6 +8795,9 @@ export declare const MODELS: {
8489
8795
  api: "openai-completions";
8490
8796
  provider: string;
8491
8797
  baseUrl: string;
8798
+ compat: {
8799
+ requiresReasoningContentOnAssistantMessages: true;
8800
+ };
8492
8801
  reasoning: true;
8493
8802
  input: "text"[];
8494
8803
  cost: {
@@ -8857,7 +9166,7 @@ export declare const MODELS: {
8857
9166
  contextWindow: number;
8858
9167
  maxTokens: number;
8859
9168
  };
8860
- readonly "inclusionai/ling-2.6-flash:free": {
9169
+ readonly "inclusionai/ling-2.6-flash": {
8861
9170
  id: string;
8862
9171
  name: string;
8863
9172
  api: "openai-completions";
@@ -9639,6 +9948,23 @@ export declare const MODELS: {
9639
9948
  contextWindow: number;
9640
9949
  maxTokens: number;
9641
9950
  };
9951
+ readonly "nvidia/nemotron-3-nano-omni-30b-a3b-reasoning:free": {
9952
+ id: string;
9953
+ name: string;
9954
+ api: "openai-completions";
9955
+ provider: string;
9956
+ baseUrl: string;
9957
+ reasoning: true;
9958
+ input: ("image" | "text")[];
9959
+ cost: {
9960
+ input: number;
9961
+ output: number;
9962
+ cacheRead: number;
9963
+ cacheWrite: number;
9964
+ };
9965
+ contextWindow: number;
9966
+ maxTokens: number;
9967
+ };
9642
9968
  readonly "nvidia/nemotron-3-super-120b-a12b": {
9643
9969
  id: string;
9644
9970
  name: string;
@@ -10710,6 +11036,40 @@ export declare const MODELS: {
10710
11036
  contextWindow: number;
10711
11037
  maxTokens: number;
10712
11038
  };
11039
+ readonly "poolside/laguna-m.1:free": {
11040
+ id: string;
11041
+ name: string;
11042
+ api: "openai-completions";
11043
+ provider: string;
11044
+ baseUrl: string;
11045
+ reasoning: true;
11046
+ input: "text"[];
11047
+ cost: {
11048
+ input: number;
11049
+ output: number;
11050
+ cacheRead: number;
11051
+ cacheWrite: number;
11052
+ };
11053
+ contextWindow: number;
11054
+ maxTokens: number;
11055
+ };
11056
+ readonly "poolside/laguna-xs.2:free": {
11057
+ id: string;
11058
+ name: string;
11059
+ api: "openai-completions";
11060
+ provider: string;
11061
+ baseUrl: string;
11062
+ reasoning: true;
11063
+ input: "text"[];
11064
+ cost: {
11065
+ input: number;
11066
+ output: number;
11067
+ cacheRead: number;
11068
+ cacheWrite: number;
11069
+ };
11070
+ contextWindow: number;
11071
+ maxTokens: number;
11072
+ };
10713
11073
  readonly "prime-intellect/intellect-3": {
10714
11074
  id: string;
10715
11075
  name: string;
@@ -11441,7 +11801,41 @@ export declare const MODELS: {
11441
11801
  contextWindow: number;
11442
11802
  maxTokens: number;
11443
11803
  };
11444
- readonly "qwen/qwen3.6-plus": {
11804
+ readonly "qwen/qwen3.5-plus-20260420": {
11805
+ id: string;
11806
+ name: string;
11807
+ api: "openai-completions";
11808
+ provider: string;
11809
+ baseUrl: string;
11810
+ reasoning: true;
11811
+ input: ("image" | "text")[];
11812
+ cost: {
11813
+ input: number;
11814
+ output: number;
11815
+ cacheRead: number;
11816
+ cacheWrite: number;
11817
+ };
11818
+ contextWindow: number;
11819
+ maxTokens: number;
11820
+ };
11821
+ readonly "qwen/qwen3.6-27b": {
11822
+ id: string;
11823
+ name: string;
11824
+ api: "openai-completions";
11825
+ provider: string;
11826
+ baseUrl: string;
11827
+ reasoning: true;
11828
+ input: ("image" | "text")[];
11829
+ cost: {
11830
+ input: number;
11831
+ output: number;
11832
+ cacheRead: number;
11833
+ cacheWrite: number;
11834
+ };
11835
+ contextWindow: number;
11836
+ maxTokens: number;
11837
+ };
11838
+ readonly "qwen/qwen3.6-flash": {
11445
11839
  id: string;
11446
11840
  name: string;
11447
11841
  api: "openai-completions";
@@ -11458,7 +11852,7 @@ export declare const MODELS: {
11458
11852
  contextWindow: number;
11459
11853
  maxTokens: number;
11460
11854
  };
11461
- readonly "qwen/qwq-32b": {
11855
+ readonly "qwen/qwen3.6-max-preview": {
11462
11856
  id: string;
11463
11857
  name: string;
11464
11858
  api: "openai-completions";
@@ -11475,6 +11869,23 @@ export declare const MODELS: {
11475
11869
  contextWindow: number;
11476
11870
  maxTokens: number;
11477
11871
  };
11872
+ readonly "qwen/qwen3.6-plus": {
11873
+ id: string;
11874
+ name: string;
11875
+ api: "openai-completions";
11876
+ provider: string;
11877
+ baseUrl: string;
11878
+ reasoning: true;
11879
+ input: ("image" | "text")[];
11880
+ cost: {
11881
+ input: number;
11882
+ output: number;
11883
+ cacheRead: number;
11884
+ cacheWrite: number;
11885
+ };
11886
+ contextWindow: number;
11887
+ maxTokens: number;
11888
+ };
11478
11889
  readonly "rekaai/reka-edge": {
11479
11890
  id: string;
11480
11891
  name: string;
@@ -11645,14 +12056,133 @@ export declare const MODELS: {
11645
12056
  contextWindow: number;
11646
12057
  maxTokens: number;
11647
12058
  };
11648
- readonly "x-ai/grok-3": {
12059
+ readonly "x-ai/grok-3": {
12060
+ id: string;
12061
+ name: string;
12062
+ api: "openai-completions";
12063
+ provider: string;
12064
+ baseUrl: string;
12065
+ reasoning: false;
12066
+ input: "text"[];
12067
+ cost: {
12068
+ input: number;
12069
+ output: number;
12070
+ cacheRead: number;
12071
+ cacheWrite: number;
12072
+ };
12073
+ contextWindow: number;
12074
+ maxTokens: number;
12075
+ };
12076
+ readonly "x-ai/grok-3-beta": {
12077
+ id: string;
12078
+ name: string;
12079
+ api: "openai-completions";
12080
+ provider: string;
12081
+ baseUrl: string;
12082
+ reasoning: false;
12083
+ input: "text"[];
12084
+ cost: {
12085
+ input: number;
12086
+ output: number;
12087
+ cacheRead: number;
12088
+ cacheWrite: number;
12089
+ };
12090
+ contextWindow: number;
12091
+ maxTokens: number;
12092
+ };
12093
+ readonly "x-ai/grok-3-mini": {
12094
+ id: string;
12095
+ name: string;
12096
+ api: "openai-completions";
12097
+ provider: string;
12098
+ baseUrl: string;
12099
+ reasoning: true;
12100
+ input: "text"[];
12101
+ cost: {
12102
+ input: number;
12103
+ output: number;
12104
+ cacheRead: number;
12105
+ cacheWrite: number;
12106
+ };
12107
+ contextWindow: number;
12108
+ maxTokens: number;
12109
+ };
12110
+ readonly "x-ai/grok-3-mini-beta": {
12111
+ id: string;
12112
+ name: string;
12113
+ api: "openai-completions";
12114
+ provider: string;
12115
+ baseUrl: string;
12116
+ reasoning: true;
12117
+ input: "text"[];
12118
+ cost: {
12119
+ input: number;
12120
+ output: number;
12121
+ cacheRead: number;
12122
+ cacheWrite: number;
12123
+ };
12124
+ contextWindow: number;
12125
+ maxTokens: number;
12126
+ };
12127
+ readonly "x-ai/grok-4": {
12128
+ id: string;
12129
+ name: string;
12130
+ api: "openai-completions";
12131
+ provider: string;
12132
+ baseUrl: string;
12133
+ reasoning: true;
12134
+ input: ("image" | "text")[];
12135
+ cost: {
12136
+ input: number;
12137
+ output: number;
12138
+ cacheRead: number;
12139
+ cacheWrite: number;
12140
+ };
12141
+ contextWindow: number;
12142
+ maxTokens: number;
12143
+ };
12144
+ readonly "x-ai/grok-4-fast": {
12145
+ id: string;
12146
+ name: string;
12147
+ api: "openai-completions";
12148
+ provider: string;
12149
+ baseUrl: string;
12150
+ reasoning: true;
12151
+ input: ("image" | "text")[];
12152
+ cost: {
12153
+ input: number;
12154
+ output: number;
12155
+ cacheRead: number;
12156
+ cacheWrite: number;
12157
+ };
12158
+ contextWindow: number;
12159
+ maxTokens: number;
12160
+ };
12161
+ readonly "x-ai/grok-4.1-fast": {
12162
+ id: string;
12163
+ name: string;
12164
+ api: "openai-completions";
12165
+ provider: string;
12166
+ baseUrl: string;
12167
+ reasoning: true;
12168
+ input: ("image" | "text")[];
12169
+ cost: {
12170
+ input: number;
12171
+ output: number;
12172
+ cacheRead: number;
12173
+ cacheWrite: number;
12174
+ };
12175
+ contextWindow: number;
12176
+ maxTokens: number;
12177
+ };
12178
+ readonly "x-ai/grok-4.20": {
11649
12179
  id: string;
11650
12180
  name: string;
11651
12181
  api: "openai-completions";
11652
12182
  provider: string;
11653
12183
  baseUrl: string;
11654
- reasoning: false;
11655
- input: "text"[];
12184
+ reasoning: true;
12185
+ input: ("image" | "text")[];
11656
12186
  cost: {
11657
12187
  input: number;
11658
12188
  output: number;
@@ -11662,13 +12192,13 @@ export declare const MODELS: {
11662
12192
  contextWindow: number;
11663
12193
  maxTokens: number;
11664
12194
  };
11665
- readonly "x-ai/grok-3-beta": {
12195
+ readonly "x-ai/grok-code-fast-1": {
11666
12196
  id: string;
11667
12197
  name: string;
11668
12198
  api: "openai-completions";
11669
12199
  provider: string;
11670
12200
  baseUrl: string;
11671
- reasoning: false;
12201
+ reasoning: true;
11672
12202
  input: "text"[];
11673
12203
  cost: {
11674
12204
  input: number;
@@ -11679,7 +12209,7 @@ export declare const MODELS: {
11679
12209
  contextWindow: number;
11680
12210
  maxTokens: number;
11681
12211
  };
11682
- readonly "x-ai/grok-3-mini": {
12212
+ readonly "xiaomi/mimo-v2-flash": {
11683
12213
  id: string;
11684
12214
  name: string;
11685
12215
  api: "openai-completions";
@@ -11696,14 +12226,14 @@ export declare const MODELS: {
11696
12226
  contextWindow: number;
11697
12227
  maxTokens: number;
11698
12228
  };
11699
- readonly "x-ai/grok-3-mini-beta": {
12229
+ readonly "xiaomi/mimo-v2-omni": {
11700
12230
  id: string;
11701
12231
  name: string;
11702
12232
  api: "openai-completions";
11703
12233
  provider: string;
11704
12234
  baseUrl: string;
11705
12235
  reasoning: true;
11706
- input: "text"[];
12236
+ input: ("image" | "text")[];
11707
12237
  cost: {
11708
12238
  input: number;
11709
12239
  output: number;
@@ -11713,14 +12243,14 @@ export declare const MODELS: {
11713
12243
  contextWindow: number;
11714
12244
  maxTokens: number;
11715
12245
  };
11716
- readonly "x-ai/grok-4": {
12246
+ readonly "xiaomi/mimo-v2-pro": {
11717
12247
  id: string;
11718
12248
  name: string;
11719
12249
  api: "openai-completions";
11720
12250
  provider: string;
11721
12251
  baseUrl: string;
11722
12252
  reasoning: true;
11723
- input: ("image" | "text")[];
12253
+ input: "text"[];
11724
12254
  cost: {
11725
12255
  input: number;
11726
12256
  output: number;
@@ -11730,7 +12260,7 @@ export declare const MODELS: {
11730
12260
  contextWindow: number;
11731
12261
  maxTokens: number;
11732
12262
  };
11733
- readonly "x-ai/grok-4-fast": {
12263
+ readonly "xiaomi/mimo-v2.5": {
11734
12264
  id: string;
11735
12265
  name: string;
11736
12266
  api: "openai-completions";
@@ -11747,14 +12277,14 @@ export declare const MODELS: {
11747
12277
  contextWindow: number;
11748
12278
  maxTokens: number;
11749
12279
  };
11750
- readonly "x-ai/grok-4.1-fast": {
12280
+ readonly "xiaomi/mimo-v2.5-pro": {
11751
12281
  id: string;
11752
12282
  name: string;
11753
12283
  api: "openai-completions";
11754
12284
  provider: string;
11755
12285
  baseUrl: string;
11756
12286
  reasoning: true;
11757
- input: ("image" | "text")[];
12287
+ input: "text"[];
11758
12288
  cost: {
11759
12289
  input: number;
11760
12290
  output: number;
@@ -11764,14 +12294,14 @@ export declare const MODELS: {
11764
12294
  contextWindow: number;
11765
12295
  maxTokens: number;
11766
12296
  };
11767
- readonly "x-ai/grok-4.20": {
12297
+ readonly "z-ai/glm-4-32b": {
11768
12298
  id: string;
11769
12299
  name: string;
11770
12300
  api: "openai-completions";
11771
12301
  provider: string;
11772
12302
  baseUrl: string;
11773
- reasoning: true;
11774
- input: ("image" | "text")[];
12303
+ reasoning: false;
12304
+ input: "text"[];
11775
12305
  cost: {
11776
12306
  input: number;
11777
12307
  output: number;
@@ -11781,7 +12311,7 @@ export declare const MODELS: {
11781
12311
  contextWindow: number;
11782
12312
  maxTokens: number;
11783
12313
  };
11784
- readonly "x-ai/grok-code-fast-1": {
12314
+ readonly "z-ai/glm-4.5": {
11785
12315
  id: string;
11786
12316
  name: string;
11787
12317
  api: "openai-completions";
@@ -11798,7 +12328,7 @@ export declare const MODELS: {
11798
12328
  contextWindow: number;
11799
12329
  maxTokens: number;
11800
12330
  };
11801
- readonly "xiaomi/mimo-v2-flash": {
12331
+ readonly "z-ai/glm-4.5-air": {
11802
12332
  id: string;
11803
12333
  name: string;
11804
12334
  api: "openai-completions";
@@ -11815,14 +12345,14 @@ export declare const MODELS: {
11815
12345
  contextWindow: number;
11816
12346
  maxTokens: number;
11817
12347
  };
11818
- readonly "xiaomi/mimo-v2-omni": {
12348
+ readonly "z-ai/glm-4.5-air:free": {
11819
12349
  id: string;
11820
12350
  name: string;
11821
12351
  api: "openai-completions";
11822
12352
  provider: string;
11823
12353
  baseUrl: string;
11824
12354
  reasoning: true;
11825
- input: ("image" | "text")[];
12355
+ input: "text"[];
11826
12356
  cost: {
11827
12357
  input: number;
11828
12358
  output: number;
@@ -11832,14 +12362,14 @@ export declare const MODELS: {
11832
12362
  contextWindow: number;
11833
12363
  maxTokens: number;
11834
12364
  };
11835
- readonly "xiaomi/mimo-v2-pro": {
12365
+ readonly "z-ai/glm-4.5v": {
11836
12366
  id: string;
11837
12367
  name: string;
11838
12368
  api: "openai-completions";
11839
12369
  provider: string;
11840
12370
  baseUrl: string;
11841
12371
  reasoning: true;
11842
- input: "text"[];
12372
+ input: ("image" | "text")[];
11843
12373
  cost: {
11844
12374
  input: number;
11845
12375
  output: number;
@@ -11849,14 +12379,14 @@ export declare const MODELS: {
11849
12379
  contextWindow: number;
11850
12380
  maxTokens: number;
11851
12381
  };
11852
- readonly "xiaomi/mimo-v2.5": {
12382
+ readonly "z-ai/glm-4.6": {
11853
12383
  id: string;
11854
12384
  name: string;
11855
12385
  api: "openai-completions";
11856
12386
  provider: string;
11857
12387
  baseUrl: string;
11858
12388
  reasoning: true;
11859
- input: ("image" | "text")[];
12389
+ input: "text"[];
11860
12390
  cost: {
11861
12391
  input: number;
11862
12392
  output: number;
@@ -11866,14 +12396,14 @@ export declare const MODELS: {
11866
12396
  contextWindow: number;
11867
12397
  maxTokens: number;
11868
12398
  };
11869
- readonly "xiaomi/mimo-v2.5-pro": {
12399
+ readonly "z-ai/glm-4.6v": {
11870
12400
  id: string;
11871
12401
  name: string;
11872
12402
  api: "openai-completions";
11873
12403
  provider: string;
11874
12404
  baseUrl: string;
11875
12405
  reasoning: true;
11876
- input: "text"[];
12406
+ input: ("image" | "text")[];
11877
12407
  cost: {
11878
12408
  input: number;
11879
12409
  output: number;
@@ -11883,13 +12413,13 @@ export declare const MODELS: {
11883
12413
  contextWindow: number;
11884
12414
  maxTokens: number;
11885
12415
  };
11886
- readonly "z-ai/glm-4-32b": {
12416
+ readonly "z-ai/glm-4.7": {
11887
12417
  id: string;
11888
12418
  name: string;
11889
12419
  api: "openai-completions";
11890
12420
  provider: string;
11891
12421
  baseUrl: string;
11892
- reasoning: false;
12422
+ reasoning: true;
11893
12423
  input: "text"[];
11894
12424
  cost: {
11895
12425
  input: number;
@@ -11900,7 +12430,7 @@ export declare const MODELS: {
11900
12430
  contextWindow: number;
11901
12431
  maxTokens: number;
11902
12432
  };
11903
- readonly "z-ai/glm-4.5": {
12433
+ readonly "z-ai/glm-4.7-flash": {
11904
12434
  id: string;
11905
12435
  name: string;
11906
12436
  api: "openai-completions";
@@ -11917,7 +12447,7 @@ export declare const MODELS: {
11917
12447
  contextWindow: number;
11918
12448
  maxTokens: number;
11919
12449
  };
11920
- readonly "z-ai/glm-4.5-air": {
12450
+ readonly "z-ai/glm-5": {
11921
12451
  id: string;
11922
12452
  name: string;
11923
12453
  api: "openai-completions";
@@ -11934,7 +12464,7 @@ export declare const MODELS: {
11934
12464
  contextWindow: number;
11935
12465
  maxTokens: number;
11936
12466
  };
11937
- readonly "z-ai/glm-4.5-air:free": {
12467
+ readonly "z-ai/glm-5-turbo": {
11938
12468
  id: string;
11939
12469
  name: string;
11940
12470
  api: "openai-completions";
@@ -11951,14 +12481,14 @@ export declare const MODELS: {
11951
12481
  contextWindow: number;
11952
12482
  maxTokens: number;
11953
12483
  };
11954
- readonly "z-ai/glm-4.5v": {
12484
+ readonly "z-ai/glm-5.1": {
11955
12485
  id: string;
11956
12486
  name: string;
11957
12487
  api: "openai-completions";
11958
12488
  provider: string;
11959
12489
  baseUrl: string;
11960
12490
  reasoning: true;
11961
- input: ("image" | "text")[];
12491
+ input: "text"[];
11962
12492
  cost: {
11963
12493
  input: number;
11964
12494
  output: number;
@@ -11968,14 +12498,14 @@ export declare const MODELS: {
11968
12498
  contextWindow: number;
11969
12499
  maxTokens: number;
11970
12500
  };
11971
- readonly "z-ai/glm-4.6": {
12501
+ readonly "z-ai/glm-5v-turbo": {
11972
12502
  id: string;
11973
12503
  name: string;
11974
12504
  api: "openai-completions";
11975
12505
  provider: string;
11976
12506
  baseUrl: string;
11977
12507
  reasoning: true;
11978
- input: "text"[];
12508
+ input: ("image" | "text")[];
11979
12509
  cost: {
11980
12510
  input: number;
11981
12511
  output: number;
@@ -11985,7 +12515,7 @@ export declare const MODELS: {
11985
12515
  contextWindow: number;
11986
12516
  maxTokens: number;
11987
12517
  };
11988
- readonly "z-ai/glm-4.6v": {
12518
+ readonly "~anthropic/claude-haiku-latest": {
11989
12519
  id: string;
11990
12520
  name: string;
11991
12521
  api: "openai-completions";
@@ -12002,14 +12532,14 @@ export declare const MODELS: {
12002
12532
  contextWindow: number;
12003
12533
  maxTokens: number;
12004
12534
  };
12005
- readonly "z-ai/glm-4.7": {
12535
+ readonly "~anthropic/claude-opus-latest": {
12006
12536
  id: string;
12007
12537
  name: string;
12008
12538
  api: "openai-completions";
12009
12539
  provider: string;
12010
12540
  baseUrl: string;
12011
12541
  reasoning: true;
12012
- input: "text"[];
12542
+ input: ("image" | "text")[];
12013
12543
  cost: {
12014
12544
  input: number;
12015
12545
  output: number;
@@ -12019,14 +12549,14 @@ export declare const MODELS: {
12019
12549
  contextWindow: number;
12020
12550
  maxTokens: number;
12021
12551
  };
12022
- readonly "z-ai/glm-4.7-flash": {
12552
+ readonly "~anthropic/claude-sonnet-latest": {
12023
12553
  id: string;
12024
12554
  name: string;
12025
12555
  api: "openai-completions";
12026
12556
  provider: string;
12027
12557
  baseUrl: string;
12028
12558
  reasoning: true;
12029
- input: "text"[];
12559
+ input: ("image" | "text")[];
12030
12560
  cost: {
12031
12561
  input: number;
12032
12562
  output: number;
@@ -12036,14 +12566,14 @@ export declare const MODELS: {
12036
12566
  contextWindow: number;
12037
12567
  maxTokens: number;
12038
12568
  };
12039
- readonly "z-ai/glm-5": {
12569
+ readonly "~google/gemini-flash-latest": {
12040
12570
  id: string;
12041
12571
  name: string;
12042
12572
  api: "openai-completions";
12043
12573
  provider: string;
12044
12574
  baseUrl: string;
12045
12575
  reasoning: true;
12046
- input: "text"[];
12576
+ input: ("image" | "text")[];
12047
12577
  cost: {
12048
12578
  input: number;
12049
12579
  output: number;
@@ -12053,14 +12583,14 @@ export declare const MODELS: {
12053
12583
  contextWindow: number;
12054
12584
  maxTokens: number;
12055
12585
  };
12056
- readonly "z-ai/glm-5-turbo": {
12586
+ readonly "~google/gemini-pro-latest": {
12057
12587
  id: string;
12058
12588
  name: string;
12059
12589
  api: "openai-completions";
12060
12590
  provider: string;
12061
12591
  baseUrl: string;
12062
12592
  reasoning: true;
12063
- input: "text"[];
12593
+ input: ("image" | "text")[];
12064
12594
  cost: {
12065
12595
  input: number;
12066
12596
  output: number;
@@ -12070,14 +12600,14 @@ export declare const MODELS: {
12070
12600
  contextWindow: number;
12071
12601
  maxTokens: number;
12072
12602
  };
12073
- readonly "z-ai/glm-5.1": {
12603
+ readonly "~moonshotai/kimi-latest": {
12074
12604
  id: string;
12075
12605
  name: string;
12076
12606
  api: "openai-completions";
12077
12607
  provider: string;
12078
12608
  baseUrl: string;
12079
12609
  reasoning: true;
12080
- input: "text"[];
12610
+ input: ("image" | "text")[];
12081
12611
  cost: {
12082
12612
  input: number;
12083
12613
  output: number;
@@ -12087,7 +12617,7 @@ export declare const MODELS: {
12087
12617
  contextWindow: number;
12088
12618
  maxTokens: number;
12089
12619
  };
12090
- readonly "z-ai/glm-5v-turbo": {
12620
+ readonly "~openai/gpt-latest": {
12091
12621
  id: string;
12092
12622
  name: string;
12093
12623
  api: "openai-completions";
@@ -12104,7 +12634,7 @@ export declare const MODELS: {
12104
12634
  contextWindow: number;
12105
12635
  maxTokens: number;
12106
12636
  };
12107
- readonly "~anthropic/claude-opus-latest": {
12637
+ readonly "~openai/gpt-mini-latest": {
12108
12638
  id: string;
12109
12639
  name: string;
12110
12640
  api: "openai-completions";
@@ -15255,27 +15785,6 @@ export declare const MODELS: {
15255
15785
  };
15256
15786
  };
15257
15787
  readonly zai: {
15258
- readonly "glm-4.5": {
15259
- id: string;
15260
- name: string;
15261
- api: "openai-completions";
15262
- provider: string;
15263
- baseUrl: string;
15264
- compat: {
15265
- supportsDeveloperRole: false;
15266
- thinkingFormat: "zai";
15267
- };
15268
- reasoning: true;
15269
- input: "text"[];
15270
- cost: {
15271
- input: number;
15272
- output: number;
15273
- cacheRead: number;
15274
- cacheWrite: number;
15275
- };
15276
- contextWindow: number;
15277
- maxTokens: number;
15278
- };
15279
15788
  readonly "glm-4.5-air": {
15280
15789
  id: string;
15281
15790
  name: string;
@@ -15297,92 +15806,6 @@ export declare const MODELS: {
15297
15806
  contextWindow: number;
15298
15807
  maxTokens: number;
15299
15808
  };
15300
- readonly "glm-4.5-flash": {
15301
- id: string;
15302
- name: string;
15303
- api: "openai-completions";
15304
- provider: string;
15305
- baseUrl: string;
15306
- compat: {
15307
- supportsDeveloperRole: false;
15308
- thinkingFormat: "zai";
15309
- };
15310
- reasoning: true;
15311
- input: "text"[];
15312
- cost: {
15313
- input: number;
15314
- output: number;
15315
- cacheRead: number;
15316
- cacheWrite: number;
15317
- };
15318
- contextWindow: number;
15319
- maxTokens: number;
15320
- };
15321
- readonly "glm-4.5v": {
15322
- id: string;
15323
- name: string;
15324
- api: "openai-completions";
15325
- provider: string;
15326
- baseUrl: string;
15327
- compat: {
15328
- supportsDeveloperRole: false;
15329
- thinkingFormat: "zai";
15330
- };
15331
- reasoning: true;
15332
- input: ("image" | "text")[];
15333
- cost: {
15334
- input: number;
15335
- output: number;
15336
- cacheRead: number;
15337
- cacheWrite: number;
15338
- };
15339
- contextWindow: number;
15340
- maxTokens: number;
15341
- };
15342
- readonly "glm-4.6": {
15343
- id: string;
15344
- name: string;
15345
- api: "openai-completions";
15346
- provider: string;
15347
- baseUrl: string;
15348
- compat: {
15349
- supportsDeveloperRole: false;
15350
- thinkingFormat: "zai";
15351
- zaiToolStream: true;
15352
- };
15353
- reasoning: true;
15354
- input: "text"[];
15355
- cost: {
15356
- input: number;
15357
- output: number;
15358
- cacheRead: number;
15359
- cacheWrite: number;
15360
- };
15361
- contextWindow: number;
15362
- maxTokens: number;
15363
- };
15364
- readonly "glm-4.6v": {
15365
- id: string;
15366
- name: string;
15367
- api: "openai-completions";
15368
- provider: string;
15369
- baseUrl: string;
15370
- compat: {
15371
- supportsDeveloperRole: false;
15372
- thinkingFormat: "zai";
15373
- zaiToolStream: true;
15374
- };
15375
- reasoning: true;
15376
- input: ("image" | "text")[];
15377
- cost: {
15378
- input: number;
15379
- output: number;
15380
- cacheRead: number;
15381
- cacheWrite: number;
15382
- };
15383
- contextWindow: number;
15384
- maxTokens: number;
15385
- };
15386
15809
  readonly "glm-4.7": {
15387
15810
  id: string;
15388
15811
  name: string;
@@ -15405,72 +15828,6 @@ export declare const MODELS: {
15405
15828
  contextWindow: number;
15406
15829
  maxTokens: number;
15407
15830
  };
15408
- readonly "glm-4.7-flash": {
15409
- id: string;
15410
- name: string;
15411
- api: "openai-completions";
15412
- provider: string;
15413
- baseUrl: string;
15414
- compat: {
15415
- supportsDeveloperRole: false;
15416
- thinkingFormat: "zai";
15417
- zaiToolStream: true;
15418
- };
15419
- reasoning: true;
15420
- input: "text"[];
15421
- cost: {
15422
- input: number;
15423
- output: number;
15424
- cacheRead: number;
15425
- cacheWrite: number;
15426
- };
15427
- contextWindow: number;
15428
- maxTokens: number;
15429
- };
15430
- readonly "glm-4.7-flashx": {
15431
- id: string;
15432
- name: string;
15433
- api: "openai-completions";
15434
- provider: string;
15435
- baseUrl: string;
15436
- compat: {
15437
- supportsDeveloperRole: false;
15438
- thinkingFormat: "zai";
15439
- zaiToolStream: true;
15440
- };
15441
- reasoning: true;
15442
- input: "text"[];
15443
- cost: {
15444
- input: number;
15445
- output: number;
15446
- cacheRead: number;
15447
- cacheWrite: number;
15448
- };
15449
- contextWindow: number;
15450
- maxTokens: number;
15451
- };
15452
- readonly "glm-5": {
15453
- id: string;
15454
- name: string;
15455
- api: "openai-completions";
15456
- provider: string;
15457
- baseUrl: string;
15458
- compat: {
15459
- supportsDeveloperRole: false;
15460
- thinkingFormat: "zai";
15461
- zaiToolStream: true;
15462
- };
15463
- reasoning: true;
15464
- input: "text"[];
15465
- cost: {
15466
- input: number;
15467
- output: number;
15468
- cacheRead: number;
15469
- cacheWrite: number;
15470
- };
15471
- contextWindow: number;
15472
- maxTokens: number;
15473
- };
15474
15831
  readonly "glm-5-turbo": {
15475
15832
  id: string;
15476
15833
  name: string;