@hyperspaceng/neural-ai 0.68.1 → 0.69.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (36) hide show
  1. package/README.md +3 -1
  2. package/dist/env-api-keys.d.ts.map +1 -1
  3. package/dist/env-api-keys.js +1 -0
  4. package/dist/env-api-keys.js.map +1 -1
  5. package/dist/index.d.ts +2 -2
  6. package/dist/index.d.ts.map +1 -1
  7. package/dist/index.js +1 -1
  8. package/dist/index.js.map +1 -1
  9. package/dist/models.generated.d.ts +610 -30
  10. package/dist/models.generated.d.ts.map +1 -1
  11. package/dist/models.generated.js +624 -44
  12. package/dist/models.generated.js.map +1 -1
  13. package/dist/providers/amazon-bedrock.d.ts.map +1 -1
  14. package/dist/providers/amazon-bedrock.js +49 -9
  15. package/dist/providers/amazon-bedrock.js.map +1 -1
  16. package/dist/providers/anthropic.d.ts.map +1 -1
  17. package/dist/providers/anthropic.js +140 -8
  18. package/dist/providers/anthropic.js.map +1 -1
  19. package/dist/providers/transform-messages.d.ts.map +1 -1
  20. package/dist/providers/transform-messages.js +2 -0
  21. package/dist/providers/transform-messages.js.map +1 -1
  22. package/dist/types.d.ts +2 -2
  23. package/dist/types.d.ts.map +1 -1
  24. package/dist/types.js.map +1 -1
  25. package/dist/utils/json-parse.d.ts +8 -1
  26. package/dist/utils/json-parse.d.ts.map +1 -1
  27. package/dist/utils/json-parse.js +89 -5
  28. package/dist/utils/json-parse.js.map +1 -1
  29. package/dist/utils/typebox-helpers.d.ts +1 -1
  30. package/dist/utils/typebox-helpers.d.ts.map +1 -1
  31. package/dist/utils/typebox-helpers.js +1 -1
  32. package/dist/utils/typebox-helpers.js.map +1 -1
  33. package/dist/utils/validation.d.ts.map +1 -1
  34. package/dist/utils/validation.js +242 -41
  35. package/dist/utils/validation.js.map +1 -1
  36. package/package.json +2 -4
@@ -323,6 +323,40 @@ export declare const MODELS: {
323
323
  contextWindow: number;
324
324
  maxTokens: number;
325
325
  };
326
+ readonly "au.anthropic.claude-opus-4-6-v1": {
327
+ id: string;
328
+ name: string;
329
+ api: "bedrock-converse-stream";
330
+ provider: string;
331
+ baseUrl: string;
332
+ reasoning: true;
333
+ input: ("image" | "text")[];
334
+ cost: {
335
+ input: number;
336
+ output: number;
337
+ cacheRead: number;
338
+ cacheWrite: number;
339
+ };
340
+ contextWindow: number;
341
+ maxTokens: number;
342
+ };
343
+ readonly "au.anthropic.claude-sonnet-4-6": {
344
+ id: string;
345
+ name: string;
346
+ api: "bedrock-converse-stream";
347
+ provider: string;
348
+ baseUrl: string;
349
+ reasoning: true;
350
+ input: ("image" | "text")[];
351
+ cost: {
352
+ input: number;
353
+ output: number;
354
+ cacheRead: number;
355
+ cacheWrite: number;
356
+ };
357
+ contextWindow: number;
358
+ maxTokens: number;
359
+ };
326
360
  readonly "deepseek.r1-v1:0": {
327
361
  id: string;
328
362
  name: string;
@@ -2486,6 +2520,23 @@ export declare const MODELS: {
2486
2520
  contextWindow: number;
2487
2521
  maxTokens: number;
2488
2522
  };
2523
+ readonly "gpt-5.5": {
2524
+ id: string;
2525
+ name: string;
2526
+ api: "azure-openai-responses";
2527
+ provider: string;
2528
+ baseUrl: string;
2529
+ reasoning: true;
2530
+ input: ("image" | "text")[];
2531
+ cost: {
2532
+ input: number;
2533
+ output: number;
2534
+ cacheRead: number;
2535
+ cacheWrite: number;
2536
+ };
2537
+ contextWindow: number;
2538
+ maxTokens: number;
2539
+ };
2489
2540
  readonly o1: {
2490
2541
  id: string;
2491
2542
  name: string;
@@ -2540,11 +2591,321 @@ export declare const MODELS: {
2540
2591
  readonly "o3-deep-research": {
2541
2592
  id: string;
2542
2593
  name: string;
2543
- api: "azure-openai-responses";
2594
+ api: "azure-openai-responses";
2595
+ provider: string;
2596
+ baseUrl: string;
2597
+ reasoning: true;
2598
+ input: ("image" | "text")[];
2599
+ cost: {
2600
+ input: number;
2601
+ output: number;
2602
+ cacheRead: number;
2603
+ cacheWrite: number;
2604
+ };
2605
+ contextWindow: number;
2606
+ maxTokens: number;
2607
+ };
2608
+ readonly "o3-mini": {
2609
+ id: string;
2610
+ name: string;
2611
+ api: "azure-openai-responses";
2612
+ provider: string;
2613
+ baseUrl: string;
2614
+ reasoning: true;
2615
+ input: "text"[];
2616
+ cost: {
2617
+ input: number;
2618
+ output: number;
2619
+ cacheRead: number;
2620
+ cacheWrite: number;
2621
+ };
2622
+ contextWindow: number;
2623
+ maxTokens: number;
2624
+ };
2625
+ readonly "o3-pro": {
2626
+ id: string;
2627
+ name: string;
2628
+ api: "azure-openai-responses";
2629
+ provider: string;
2630
+ baseUrl: string;
2631
+ reasoning: true;
2632
+ input: ("image" | "text")[];
2633
+ cost: {
2634
+ input: number;
2635
+ output: number;
2636
+ cacheRead: number;
2637
+ cacheWrite: number;
2638
+ };
2639
+ contextWindow: number;
2640
+ maxTokens: number;
2641
+ };
2642
+ readonly "o4-mini": {
2643
+ id: string;
2644
+ name: string;
2645
+ api: "azure-openai-responses";
2646
+ provider: string;
2647
+ baseUrl: string;
2648
+ reasoning: true;
2649
+ input: ("image" | "text")[];
2650
+ cost: {
2651
+ input: number;
2652
+ output: number;
2653
+ cacheRead: number;
2654
+ cacheWrite: number;
2655
+ };
2656
+ contextWindow: number;
2657
+ maxTokens: number;
2658
+ };
2659
+ readonly "o4-mini-deep-research": {
2660
+ id: string;
2661
+ name: string;
2662
+ api: "azure-openai-responses";
2663
+ provider: string;
2664
+ baseUrl: string;
2665
+ reasoning: true;
2666
+ input: ("image" | "text")[];
2667
+ cost: {
2668
+ input: number;
2669
+ output: number;
2670
+ cacheRead: number;
2671
+ cacheWrite: number;
2672
+ };
2673
+ contextWindow: number;
2674
+ maxTokens: number;
2675
+ };
2676
+ };
2677
+ readonly cerebras: {
2678
+ readonly "gpt-oss-120b": {
2679
+ id: string;
2680
+ name: string;
2681
+ api: "openai-completions";
2682
+ provider: string;
2683
+ baseUrl: string;
2684
+ reasoning: true;
2685
+ input: "text"[];
2686
+ cost: {
2687
+ input: number;
2688
+ output: number;
2689
+ cacheRead: number;
2690
+ cacheWrite: number;
2691
+ };
2692
+ contextWindow: number;
2693
+ maxTokens: number;
2694
+ };
2695
+ readonly "llama3.1-8b": {
2696
+ id: string;
2697
+ name: string;
2698
+ api: "openai-completions";
2699
+ provider: string;
2700
+ baseUrl: string;
2701
+ reasoning: false;
2702
+ input: "text"[];
2703
+ cost: {
2704
+ input: number;
2705
+ output: number;
2706
+ cacheRead: number;
2707
+ cacheWrite: number;
2708
+ };
2709
+ contextWindow: number;
2710
+ maxTokens: number;
2711
+ };
2712
+ readonly "qwen-3-235b-a22b-instruct-2507": {
2713
+ id: string;
2714
+ name: string;
2715
+ api: "openai-completions";
2716
+ provider: string;
2717
+ baseUrl: string;
2718
+ reasoning: false;
2719
+ input: "text"[];
2720
+ cost: {
2721
+ input: number;
2722
+ output: number;
2723
+ cacheRead: number;
2724
+ cacheWrite: number;
2725
+ };
2726
+ contextWindow: number;
2727
+ maxTokens: number;
2728
+ };
2729
+ readonly "zai-glm-4.7": {
2730
+ id: string;
2731
+ name: string;
2732
+ api: "openai-completions";
2733
+ provider: string;
2734
+ baseUrl: string;
2735
+ reasoning: false;
2736
+ input: "text"[];
2737
+ cost: {
2738
+ input: number;
2739
+ output: number;
2740
+ cacheRead: number;
2741
+ cacheWrite: number;
2742
+ };
2743
+ contextWindow: number;
2744
+ maxTokens: number;
2745
+ };
2746
+ };
2747
+ readonly fireworks: {
2748
+ readonly "accounts/fireworks/models/deepseek-v3p1": {
2749
+ id: string;
2750
+ name: string;
2751
+ api: "anthropic-messages";
2752
+ provider: string;
2753
+ baseUrl: string;
2754
+ reasoning: true;
2755
+ input: "text"[];
2756
+ cost: {
2757
+ input: number;
2758
+ output: number;
2759
+ cacheRead: number;
2760
+ cacheWrite: number;
2761
+ };
2762
+ contextWindow: number;
2763
+ maxTokens: number;
2764
+ };
2765
+ readonly "accounts/fireworks/models/deepseek-v3p2": {
2766
+ id: string;
2767
+ name: string;
2768
+ api: "anthropic-messages";
2769
+ provider: string;
2770
+ baseUrl: string;
2771
+ reasoning: true;
2772
+ input: "text"[];
2773
+ cost: {
2774
+ input: number;
2775
+ output: number;
2776
+ cacheRead: number;
2777
+ cacheWrite: number;
2778
+ };
2779
+ contextWindow: number;
2780
+ maxTokens: number;
2781
+ };
2782
+ readonly "accounts/fireworks/models/glm-4p5": {
2783
+ id: string;
2784
+ name: string;
2785
+ api: "anthropic-messages";
2786
+ provider: string;
2787
+ baseUrl: string;
2788
+ reasoning: true;
2789
+ input: "text"[];
2790
+ cost: {
2791
+ input: number;
2792
+ output: number;
2793
+ cacheRead: number;
2794
+ cacheWrite: number;
2795
+ };
2796
+ contextWindow: number;
2797
+ maxTokens: number;
2798
+ };
2799
+ readonly "accounts/fireworks/models/glm-4p5-air": {
2800
+ id: string;
2801
+ name: string;
2802
+ api: "anthropic-messages";
2803
+ provider: string;
2804
+ baseUrl: string;
2805
+ reasoning: true;
2806
+ input: "text"[];
2807
+ cost: {
2808
+ input: number;
2809
+ output: number;
2810
+ cacheRead: number;
2811
+ cacheWrite: number;
2812
+ };
2813
+ contextWindow: number;
2814
+ maxTokens: number;
2815
+ };
2816
+ readonly "accounts/fireworks/models/glm-4p7": {
2817
+ id: string;
2818
+ name: string;
2819
+ api: "anthropic-messages";
2820
+ provider: string;
2821
+ baseUrl: string;
2822
+ reasoning: true;
2823
+ input: "text"[];
2824
+ cost: {
2825
+ input: number;
2826
+ output: number;
2827
+ cacheRead: number;
2828
+ cacheWrite: number;
2829
+ };
2830
+ contextWindow: number;
2831
+ maxTokens: number;
2832
+ };
2833
+ readonly "accounts/fireworks/models/glm-5": {
2834
+ id: string;
2835
+ name: string;
2836
+ api: "anthropic-messages";
2837
+ provider: string;
2838
+ baseUrl: string;
2839
+ reasoning: true;
2840
+ input: "text"[];
2841
+ cost: {
2842
+ input: number;
2843
+ output: number;
2844
+ cacheRead: number;
2845
+ cacheWrite: number;
2846
+ };
2847
+ contextWindow: number;
2848
+ maxTokens: number;
2849
+ };
2850
+ readonly "accounts/fireworks/models/glm-5p1": {
2851
+ id: string;
2852
+ name: string;
2853
+ api: "anthropic-messages";
2854
+ provider: string;
2855
+ baseUrl: string;
2856
+ reasoning: true;
2857
+ input: "text"[];
2858
+ cost: {
2859
+ input: number;
2860
+ output: number;
2861
+ cacheRead: number;
2862
+ cacheWrite: number;
2863
+ };
2864
+ contextWindow: number;
2865
+ maxTokens: number;
2866
+ };
2867
+ readonly "accounts/fireworks/models/gpt-oss-120b": {
2868
+ id: string;
2869
+ name: string;
2870
+ api: "anthropic-messages";
2871
+ provider: string;
2872
+ baseUrl: string;
2873
+ reasoning: true;
2874
+ input: "text"[];
2875
+ cost: {
2876
+ input: number;
2877
+ output: number;
2878
+ cacheRead: number;
2879
+ cacheWrite: number;
2880
+ };
2881
+ contextWindow: number;
2882
+ maxTokens: number;
2883
+ };
2884
+ readonly "accounts/fireworks/models/gpt-oss-20b": {
2885
+ id: string;
2886
+ name: string;
2887
+ api: "anthropic-messages";
2888
+ provider: string;
2889
+ baseUrl: string;
2890
+ reasoning: true;
2891
+ input: "text"[];
2892
+ cost: {
2893
+ input: number;
2894
+ output: number;
2895
+ cacheRead: number;
2896
+ cacheWrite: number;
2897
+ };
2898
+ contextWindow: number;
2899
+ maxTokens: number;
2900
+ };
2901
+ readonly "accounts/fireworks/models/kimi-k2-instruct": {
2902
+ id: string;
2903
+ name: string;
2904
+ api: "anthropic-messages";
2544
2905
  provider: string;
2545
2906
  baseUrl: string;
2546
- reasoning: true;
2547
- input: ("image" | "text")[];
2907
+ reasoning: false;
2908
+ input: "text"[];
2548
2909
  cost: {
2549
2910
  input: number;
2550
2911
  output: number;
@@ -2554,10 +2915,10 @@ export declare const MODELS: {
2554
2915
  contextWindow: number;
2555
2916
  maxTokens: number;
2556
2917
  };
2557
- readonly "o3-mini": {
2918
+ readonly "accounts/fireworks/models/kimi-k2-thinking": {
2558
2919
  id: string;
2559
2920
  name: string;
2560
- api: "azure-openai-responses";
2921
+ api: "anthropic-messages";
2561
2922
  provider: string;
2562
2923
  baseUrl: string;
2563
2924
  reasoning: true;
@@ -2571,10 +2932,10 @@ export declare const MODELS: {
2571
2932
  contextWindow: number;
2572
2933
  maxTokens: number;
2573
2934
  };
2574
- readonly "o3-pro": {
2935
+ readonly "accounts/fireworks/models/kimi-k2p5": {
2575
2936
  id: string;
2576
2937
  name: string;
2577
- api: "azure-openai-responses";
2938
+ api: "anthropic-messages";
2578
2939
  provider: string;
2579
2940
  baseUrl: string;
2580
2941
  reasoning: true;
@@ -2588,10 +2949,10 @@ export declare const MODELS: {
2588
2949
  contextWindow: number;
2589
2950
  maxTokens: number;
2590
2951
  };
2591
- readonly "o4-mini": {
2952
+ readonly "accounts/fireworks/models/kimi-k2p6": {
2592
2953
  id: string;
2593
2954
  name: string;
2594
- api: "azure-openai-responses";
2955
+ api: "anthropic-messages";
2595
2956
  provider: string;
2596
2957
  baseUrl: string;
2597
2958
  reasoning: true;
@@ -2605,14 +2966,14 @@ export declare const MODELS: {
2605
2966
  contextWindow: number;
2606
2967
  maxTokens: number;
2607
2968
  };
2608
- readonly "o4-mini-deep-research": {
2969
+ readonly "accounts/fireworks/models/minimax-m2p1": {
2609
2970
  id: string;
2610
2971
  name: string;
2611
- api: "azure-openai-responses";
2972
+ api: "anthropic-messages";
2612
2973
  provider: string;
2613
2974
  baseUrl: string;
2614
2975
  reasoning: true;
2615
- input: ("image" | "text")[];
2976
+ input: "text"[];
2616
2977
  cost: {
2617
2978
  input: number;
2618
2979
  output: number;
@@ -2622,12 +2983,10 @@ export declare const MODELS: {
2622
2983
  contextWindow: number;
2623
2984
  maxTokens: number;
2624
2985
  };
2625
- };
2626
- readonly cerebras: {
2627
- readonly "gpt-oss-120b": {
2986
+ readonly "accounts/fireworks/models/minimax-m2p5": {
2628
2987
  id: string;
2629
2988
  name: string;
2630
- api: "openai-completions";
2989
+ api: "anthropic-messages";
2631
2990
  provider: string;
2632
2991
  baseUrl: string;
2633
2992
  reasoning: true;
@@ -2641,13 +3000,13 @@ export declare const MODELS: {
2641
3000
  contextWindow: number;
2642
3001
  maxTokens: number;
2643
3002
  };
2644
- readonly "llama3.1-8b": {
3003
+ readonly "accounts/fireworks/models/minimax-m2p7": {
2645
3004
  id: string;
2646
3005
  name: string;
2647
- api: "openai-completions";
3006
+ api: "anthropic-messages";
2648
3007
  provider: string;
2649
3008
  baseUrl: string;
2650
- reasoning: false;
3009
+ reasoning: true;
2651
3010
  input: "text"[];
2652
3011
  cost: {
2653
3012
  input: number;
@@ -2658,14 +3017,14 @@ export declare const MODELS: {
2658
3017
  contextWindow: number;
2659
3018
  maxTokens: number;
2660
3019
  };
2661
- readonly "qwen-3-235b-a22b-instruct-2507": {
3020
+ readonly "accounts/fireworks/models/qwen3p6-plus": {
2662
3021
  id: string;
2663
3022
  name: string;
2664
- api: "openai-completions";
3023
+ api: "anthropic-messages";
2665
3024
  provider: string;
2666
3025
  baseUrl: string;
2667
- reasoning: false;
2668
- input: "text"[];
3026
+ reasoning: true;
3027
+ input: ("image" | "text")[];
2669
3028
  cost: {
2670
3029
  input: number;
2671
3030
  output: number;
@@ -2675,14 +3034,14 @@ export declare const MODELS: {
2675
3034
  contextWindow: number;
2676
3035
  maxTokens: number;
2677
3036
  };
2678
- readonly "zai-glm-4.7": {
3037
+ readonly "accounts/fireworks/routers/kimi-k2p5-turbo": {
2679
3038
  id: string;
2680
3039
  name: string;
2681
- api: "openai-completions";
3040
+ api: "anthropic-messages";
2682
3041
  provider: string;
2683
3042
  baseUrl: string;
2684
- reasoning: false;
2685
- input: "text"[];
3043
+ reasoning: true;
3044
+ input: ("image" | "text")[];
2686
3045
  cost: {
2687
3046
  input: number;
2688
3047
  output: number;
@@ -3731,7 +4090,7 @@ export declare const MODELS: {
3731
4090
  contextWindow: number;
3732
4091
  maxTokens: number;
3733
4092
  };
3734
- readonly "gemma-4-26b-it": {
4093
+ readonly "gemma-4-26b-a4b-it": {
3735
4094
  id: string;
3736
4095
  name: string;
3737
4096
  api: "google-generative-ai";
@@ -4007,6 +4366,23 @@ export declare const MODELS: {
4007
4366
  contextWindow: number;
4008
4367
  maxTokens: number;
4009
4368
  };
4369
+ readonly "gemini-3.1-flash-lite-preview": {
4370
+ id: string;
4371
+ name: string;
4372
+ api: "google-gemini-cli";
4373
+ provider: string;
4374
+ baseUrl: string;
4375
+ reasoning: true;
4376
+ input: ("image" | "text")[];
4377
+ cost: {
4378
+ input: number;
4379
+ output: number;
4380
+ cacheRead: number;
4381
+ cacheWrite: number;
4382
+ };
4383
+ contextWindow: number;
4384
+ maxTokens: number;
4385
+ };
4010
4386
  readonly "gemini-3.1-pro-preview": {
4011
4387
  id: string;
4012
4388
  name: string;
@@ -6092,6 +6468,23 @@ export declare const MODELS: {
6092
6468
  contextWindow: number;
6093
6469
  maxTokens: number;
6094
6470
  };
6471
+ readonly "gpt-5.5": {
6472
+ id: string;
6473
+ name: string;
6474
+ api: "openai-responses";
6475
+ provider: string;
6476
+ baseUrl: string;
6477
+ reasoning: true;
6478
+ input: ("image" | "text")[];
6479
+ cost: {
6480
+ input: number;
6481
+ output: number;
6482
+ cacheRead: number;
6483
+ cacheWrite: number;
6484
+ };
6485
+ contextWindow: number;
6486
+ maxTokens: number;
6487
+ };
6095
6488
  readonly o1: {
6096
6489
  id: string;
6097
6490
  name: string;
@@ -6861,6 +7254,23 @@ export declare const MODELS: {
6861
7254
  contextWindow: number;
6862
7255
  maxTokens: number;
6863
7256
  };
7257
+ readonly "hy3-preview-free": {
7258
+ id: string;
7259
+ name: string;
7260
+ api: "openai-completions";
7261
+ provider: string;
7262
+ baseUrl: string;
7263
+ reasoning: true;
7264
+ input: "text"[];
7265
+ cost: {
7266
+ input: number;
7267
+ output: number;
7268
+ cacheRead: number;
7269
+ cacheWrite: number;
7270
+ };
7271
+ contextWindow: number;
7272
+ maxTokens: number;
7273
+ };
6864
7274
  readonly "kimi-k2.5": {
6865
7275
  id: string;
6866
7276
  name: string;
@@ -7118,6 +7528,40 @@ export declare const MODELS: {
7118
7528
  contextWindow: number;
7119
7529
  maxTokens: number;
7120
7530
  };
7531
+ readonly "mimo-v2.5": {
7532
+ id: string;
7533
+ name: string;
7534
+ api: "openai-completions";
7535
+ provider: string;
7536
+ baseUrl: string;
7537
+ reasoning: true;
7538
+ input: ("image" | "text")[];
7539
+ cost: {
7540
+ input: number;
7541
+ output: number;
7542
+ cacheRead: number;
7543
+ cacheWrite: number;
7544
+ };
7545
+ contextWindow: number;
7546
+ maxTokens: number;
7547
+ };
7548
+ readonly "mimo-v2.5-pro": {
7549
+ id: string;
7550
+ name: string;
7551
+ api: "openai-completions";
7552
+ provider: string;
7553
+ baseUrl: string;
7554
+ reasoning: true;
7555
+ input: "text"[];
7556
+ cost: {
7557
+ input: number;
7558
+ output: number;
7559
+ cacheRead: number;
7560
+ cacheWrite: number;
7561
+ };
7562
+ contextWindow: number;
7563
+ maxTokens: number;
7564
+ };
7121
7565
  readonly "minimax-m2.5": {
7122
7566
  id: string;
7123
7567
  name: string;
@@ -7562,7 +8006,7 @@ export declare const MODELS: {
7562
8006
  contextWindow: number;
7563
8007
  maxTokens: number;
7564
8008
  };
7565
- readonly "arcee-ai/trinity-large-preview:free": {
8009
+ readonly "arcee-ai/trinity-large-preview": {
7566
8010
  id: string;
7567
8011
  name: string;
7568
8012
  api: "openai-completions";
@@ -7919,6 +8363,40 @@ export declare const MODELS: {
7919
8363
  contextWindow: number;
7920
8364
  maxTokens: number;
7921
8365
  };
8366
+ readonly "deepseek/deepseek-v4-flash": {
8367
+ id: string;
8368
+ name: string;
8369
+ api: "openai-completions";
8370
+ provider: string;
8371
+ baseUrl: string;
8372
+ reasoning: true;
8373
+ input: "text"[];
8374
+ cost: {
8375
+ input: number;
8376
+ output: number;
8377
+ cacheRead: number;
8378
+ cacheWrite: number;
8379
+ };
8380
+ contextWindow: number;
8381
+ maxTokens: number;
8382
+ };
8383
+ readonly "deepseek/deepseek-v4-pro": {
8384
+ id: string;
8385
+ name: string;
8386
+ api: "openai-completions";
8387
+ provider: string;
8388
+ baseUrl: string;
8389
+ reasoning: true;
8390
+ input: "text"[];
8391
+ cost: {
8392
+ input: number;
8393
+ output: number;
8394
+ cacheRead: number;
8395
+ cacheWrite: number;
8396
+ };
8397
+ contextWindow: number;
8398
+ maxTokens: number;
8399
+ };
7922
8400
  readonly "essentialai/rnj-1-instruct": {
7923
8401
  id: string;
7924
8402
  name: string;
@@ -8225,6 +8703,23 @@ export declare const MODELS: {
8225
8703
  contextWindow: number;
8226
8704
  maxTokens: number;
8227
8705
  };
8706
+ readonly "inclusionai/ling-2.6-1t:free": {
8707
+ id: string;
8708
+ name: string;
8709
+ api: "openai-completions";
8710
+ provider: string;
8711
+ baseUrl: string;
8712
+ reasoning: false;
8713
+ input: "text"[];
8714
+ cost: {
8715
+ input: number;
8716
+ output: number;
8717
+ cacheRead: number;
8718
+ cacheWrite: number;
8719
+ };
8720
+ contextWindow: number;
8721
+ maxTokens: number;
8722
+ };
8228
8723
  readonly "inclusionai/ling-2.6-flash:free": {
8229
8724
  id: string;
8230
8725
  name: string;
@@ -10894,6 +11389,23 @@ export declare const MODELS: {
10894
11389
  contextWindow: number;
10895
11390
  maxTokens: number;
10896
11391
  };
11392
+ readonly "tencent/hy3-preview:free": {
11393
+ id: string;
11394
+ name: string;
11395
+ api: "openai-completions";
11396
+ provider: string;
11397
+ baseUrl: string;
11398
+ reasoning: true;
11399
+ input: "text"[];
11400
+ cost: {
11401
+ input: number;
11402
+ output: number;
11403
+ cacheRead: number;
11404
+ cacheWrite: number;
11405
+ };
11406
+ contextWindow: number;
11407
+ maxTokens: number;
11408
+ };
10897
11409
  readonly "thedrummer/rocinante-12b": {
10898
11410
  id: string;
10899
11411
  name: string;
@@ -11166,6 +11678,40 @@ export declare const MODELS: {
11166
11678
  contextWindow: number;
11167
11679
  maxTokens: number;
11168
11680
  };
11681
+ readonly "xiaomi/mimo-v2.5": {
11682
+ id: string;
11683
+ name: string;
11684
+ api: "openai-completions";
11685
+ provider: string;
11686
+ baseUrl: string;
11687
+ reasoning: true;
11688
+ input: ("image" | "text")[];
11689
+ cost: {
11690
+ input: number;
11691
+ output: number;
11692
+ cacheRead: number;
11693
+ cacheWrite: number;
11694
+ };
11695
+ contextWindow: number;
11696
+ maxTokens: number;
11697
+ };
11698
+ readonly "xiaomi/mimo-v2.5-pro": {
11699
+ id: string;
11700
+ name: string;
11701
+ api: "openai-completions";
11702
+ provider: string;
11703
+ baseUrl: string;
11704
+ reasoning: true;
11705
+ input: "text"[];
11706
+ cost: {
11707
+ input: number;
11708
+ output: number;
11709
+ cacheRead: number;
11710
+ cacheWrite: number;
11711
+ };
11712
+ contextWindow: number;
11713
+ maxTokens: number;
11714
+ };
11169
11715
  readonly "z-ai/glm-4-32b": {
11170
11716
  id: string;
11171
11717
  name: string;
@@ -12069,6 +12615,40 @@ export declare const MODELS: {
12069
12615
  contextWindow: number;
12070
12616
  maxTokens: number;
12071
12617
  };
12618
+ readonly "deepseek/deepseek-v4-flash": {
12619
+ id: string;
12620
+ name: string;
12621
+ api: "anthropic-messages";
12622
+ provider: string;
12623
+ baseUrl: string;
12624
+ reasoning: true;
12625
+ input: "text"[];
12626
+ cost: {
12627
+ input: number;
12628
+ output: number;
12629
+ cacheRead: number;
12630
+ cacheWrite: number;
12631
+ };
12632
+ contextWindow: number;
12633
+ maxTokens: number;
12634
+ };
12635
+ readonly "deepseek/deepseek-v4-pro": {
12636
+ id: string;
12637
+ name: string;
12638
+ api: "anthropic-messages";
12639
+ provider: string;
12640
+ baseUrl: string;
12641
+ reasoning: true;
12642
+ input: "text"[];
12643
+ cost: {
12644
+ input: number;
12645
+ output: number;
12646
+ cacheRead: number;
12647
+ cacheWrite: number;
12648
+ };
12649
+ contextWindow: number;
12650
+ maxTokens: number;
12651
+ };
12072
12652
  readonly "google/gemini-2.0-flash": {
12073
12653
  id: string;
12074
12654
  name: string;
@@ -14031,7 +14611,7 @@ export declare const MODELS: {
14031
14611
  provider: string;
14032
14612
  baseUrl: string;
14033
14613
  reasoning: true;
14034
- input: ("image" | "text")[];
14614
+ input: "text"[];
14035
14615
  cost: {
14036
14616
  input: number;
14037
14617
  output: number;