@llmgateway/ai-sdk-provider 2.5.0 → 3.0.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -8,7 +8,7 @@ type LLMGatewayProviderOptions = {
8
8
  * One of `max_tokens` or `effort` is required.
9
9
  * If `exclude` is true, reasoning will be removed from the response. Default is false.
10
10
  */
11
- reasoning?: {
11
+ reasoningText?: {
12
12
  enabled?: boolean;
13
13
  exclude?: boolean;
14
14
  } & ({
@@ -16,22 +16,6 @@ type LLMGatewayProviderOptions = {
16
16
  } | {
17
17
  effort: 'high' | 'medium' | 'low';
18
18
  });
19
- /**
20
- * Image generation configuration for supported models (e.g., Google's image generation models).
21
- * Specifies aspect ratio and image resolution for generated images.
22
- */
23
- image_config?: {
24
- /**
25
- * The aspect ratio of the generated image.
26
- * Examples: "1:1", "16:9", "4:3", "5:4"
27
- */
28
- aspect_ratio?: string;
29
- /**
30
- * The resolution of the generated image.
31
- * Options: "1K" (1024x1024), "2K" (2048x2048), "4K"
32
- */
33
- image_size?: string;
34
- };
35
19
  /**
36
20
  * A unique identifier representing your end-user, which can
37
21
  * help LLMGateway to monitor and detect abuse.
@@ -43,11 +27,6 @@ type LLMGatewaySharedSettings = LLMGatewayProviderOptions & {
43
27
  * @deprecated use `reasoning` instead
44
28
  */
45
29
  includeReasoning?: boolean;
46
- /**
47
- * Reasoning effort level for models that support it.
48
- * Controls the computational effort applied to reasoning tasks.
49
- */
50
- reasoning_effort?: 'minimal' | 'low' | 'medium' | 'high';
51
30
  extraBody?: Record<string, unknown>;
52
31
  /**
53
32
  * Enable usage accounting to get detailed token usage information.
@@ -368,7 +347,7 @@ interface ProviderModelMapping {
368
347
  /**
369
348
  * Whether this model supports reasoning mode
370
349
  */
371
- reasoning?: boolean;
350
+ reasoningText?: boolean;
372
351
  /**
373
352
  * Whether this model supports the OpenAI responses API (defaults to true if reasoning is true)
374
353
  */
@@ -632,7 +611,7 @@ declare const models: [{
632
611
  readonly maxOutput: undefined;
633
612
  readonly streaming: true;
634
613
  readonly vision: true;
635
- readonly reasoning: true;
614
+ readonly reasoningText: true;
636
615
  readonly tools: false;
637
616
  }];
638
617
  readonly jsonOutput: true;
@@ -653,7 +632,7 @@ declare const models: [{
653
632
  readonly streaming: false;
654
633
  readonly vision: false;
655
634
  readonly tools: false;
656
- readonly reasoning: false;
635
+ readonly reasoningText: false;
657
636
  readonly supportsResponsesApi: false;
658
637
  }];
659
638
  readonly jsonOutput: true;
@@ -753,7 +732,7 @@ declare const models: [{
753
732
  readonly streaming: true;
754
733
  readonly vision: false;
755
734
  readonly tools: true;
756
- readonly reasoning: true;
735
+ readonly reasoningText: true;
757
736
  }];
758
737
  readonly jsonOutput: true;
759
738
  }, {
@@ -773,7 +752,7 @@ declare const models: [{
773
752
  readonly streaming: true;
774
753
  readonly vision: false;
775
754
  readonly tools: true;
776
- readonly reasoning: true;
755
+ readonly reasoningText: true;
777
756
  }];
778
757
  readonly jsonOutput: true;
779
758
  }, {
@@ -791,7 +770,7 @@ declare const models: [{
791
770
  readonly requestPrice: 0;
792
771
  readonly contextSize: 400000;
793
772
  readonly maxOutput: 128000;
794
- readonly reasoning: true;
773
+ readonly reasoningText: true;
795
774
  readonly streaming: true;
796
775
  readonly vision: true;
797
776
  readonly tools: true;
@@ -813,7 +792,7 @@ declare const models: [{
813
792
  readonly requestPrice: 0;
814
793
  readonly contextSize: 400000;
815
794
  readonly maxOutput: 128000;
816
- readonly reasoning: true;
795
+ readonly reasoningText: true;
817
796
  readonly streaming: true;
818
797
  readonly vision: true;
819
798
  readonly tools: true;
@@ -835,7 +814,7 @@ declare const models: [{
835
814
  readonly requestPrice: 0;
836
815
  readonly contextSize: 400000;
837
816
  readonly maxOutput: 128000;
838
- readonly reasoning: true;
817
+ readonly reasoningText: true;
839
818
  readonly streaming: true;
840
819
  readonly vision: false;
841
820
  readonly tools: true;
@@ -899,7 +878,7 @@ declare const models: [{
899
878
  readonly maxOutput: 8192;
900
879
  readonly streaming: true;
901
880
  readonly vision: false;
902
- readonly reasoning: true;
881
+ readonly reasoningText: true;
903
882
  readonly tools: true;
904
883
  }];
905
884
  }, {
@@ -994,7 +973,7 @@ declare const models: [{
994
973
  readonly maxOutput: 32000;
995
974
  readonly streaming: true;
996
975
  readonly vision: true;
997
- readonly reasoning: true;
976
+ readonly reasoningText: true;
998
977
  readonly tools: true;
999
978
  }];
1000
979
  }, {
@@ -1378,7 +1357,7 @@ declare const models: [{
1378
1357
  readonly contextSize: 1000000;
1379
1358
  readonly maxOutput: undefined;
1380
1359
  readonly streaming: true;
1381
- readonly reasoning: false;
1360
+ readonly reasoningText: false;
1382
1361
  readonly vision: false;
1383
1362
  readonly tools: false;
1384
1363
  }];
@@ -1397,7 +1376,7 @@ declare const models: [{
1397
1376
  readonly contextSize: 1000000;
1398
1377
  readonly maxOutput: undefined;
1399
1378
  readonly streaming: true;
1400
- readonly reasoning: false;
1379
+ readonly reasoningText: false;
1401
1380
  readonly vision: false;
1402
1381
  readonly tools: false;
1403
1382
  }];
@@ -2308,7 +2287,7 @@ declare const models: [{
2308
2287
  readonly requestPrice: 0;
2309
2288
  readonly contextSize: 262000;
2310
2289
  readonly maxOutput: 8192;
2311
- readonly reasoning: false;
2290
+ readonly reasoningText: false;
2312
2291
  readonly streaming: true;
2313
2292
  readonly vision: false;
2314
2293
  readonly tools: true;
@@ -2328,7 +2307,7 @@ declare const models: [{
2328
2307
  readonly requestPrice: 0;
2329
2308
  readonly contextSize: 262000;
2330
2309
  readonly maxOutput: 8192;
2331
- readonly reasoning: true;
2310
+ readonly reasoningText: true;
2332
2311
  readonly streaming: true;
2333
2312
  readonly vision: false;
2334
2313
  readonly tools: true;
@@ -2560,7 +2539,7 @@ declare const models: [{
2560
2539
  readonly streaming: true;
2561
2540
  readonly vision: false;
2562
2541
  readonly tools: true;
2563
- readonly reasoning: true;
2542
+ readonly reasoningText: true;
2564
2543
  }];
2565
2544
  readonly jsonOutput: true;
2566
2545
  }, {
@@ -2615,7 +2594,7 @@ declare const models: [{
2615
2594
  readonly requestPrice: 0;
2616
2595
  readonly contextSize: 131072;
2617
2596
  readonly maxOutput: 32768;
2618
- readonly reasoning: true;
2597
+ readonly reasoningText: true;
2619
2598
  readonly reasoningOutput: "omit";
2620
2599
  readonly streaming: true;
2621
2600
  readonly vision: false;
@@ -2656,7 +2635,7 @@ declare const models: [{
2656
2635
  readonly requestPrice: 0;
2657
2636
  readonly contextSize: 256000;
2658
2637
  readonly maxOutput: 32800;
2659
- readonly reasoning: true;
2638
+ readonly reasoningText: true;
2660
2639
  readonly reasoningOutput: "omit";
2661
2640
  readonly streaming: true;
2662
2641
  readonly vision: true;
@@ -2877,7 +2856,7 @@ declare const models: [{
2877
2856
  readonly contextSize: 128000;
2878
2857
  readonly maxOutput: undefined;
2879
2858
  readonly streaming: true;
2880
- readonly reasoning: true;
2859
+ readonly reasoningText: true;
2881
2860
  readonly vision: false;
2882
2861
  readonly tools: true;
2883
2862
  }];
@@ -2897,7 +2876,7 @@ declare const models: [{
2897
2876
  readonly contextSize: 128000;
2898
2877
  readonly maxOutput: undefined;
2899
2878
  readonly streaming: true;
2900
- readonly reasoning: true;
2879
+ readonly reasoningText: true;
2901
2880
  readonly reasoningOutput: "omit";
2902
2881
  readonly vision: true;
2903
2882
  readonly tools: true;
@@ -2918,7 +2897,7 @@ declare const models: [{
2918
2897
  readonly contextSize: 128000;
2919
2898
  readonly maxOutput: undefined;
2920
2899
  readonly streaming: true;
2921
- readonly reasoning: false;
2900
+ readonly reasoningText: false;
2922
2901
  readonly vision: false;
2923
2902
  readonly tools: true;
2924
2903
  }];
@@ -2938,7 +2917,7 @@ declare const models: [{
2938
2917
  readonly contextSize: 128000;
2939
2918
  readonly maxOutput: undefined;
2940
2919
  readonly streaming: true;
2941
- readonly reasoning: true;
2920
+ readonly reasoningText: true;
2942
2921
  readonly vision: false;
2943
2922
  readonly tools: true;
2944
2923
  }];
@@ -2958,7 +2937,7 @@ declare const models: [{
2958
2937
  readonly contextSize: 128000;
2959
2938
  readonly maxOutput: undefined;
2960
2939
  readonly streaming: true;
2961
- readonly reasoning: false;
2940
+ readonly reasoningText: false;
2962
2941
  readonly vision: false;
2963
2942
  readonly tools: true;
2964
2943
  }];
@@ -2978,7 +2957,7 @@ declare const models: [{
2978
2957
  readonly contextSize: 128000;
2979
2958
  readonly maxOutput: undefined;
2980
2959
  readonly streaming: true;
2981
- readonly reasoning: false;
2960
+ readonly reasoningText: false;
2982
2961
  readonly vision: false;
2983
2962
  readonly tools: true;
2984
2963
  }];
@@ -2998,7 +2977,7 @@ declare const models: [{
2998
2977
  readonly contextSize: 128000;
2999
2978
  readonly maxOutput: undefined;
3000
2979
  readonly streaming: true;
3001
- readonly reasoning: false;
2980
+ readonly reasoningText: false;
3002
2981
  readonly vision: false;
3003
2982
  readonly tools: true;
3004
2983
  }];
@@ -8,7 +8,7 @@ type LLMGatewayProviderOptions = {
8
8
  * One of `max_tokens` or `effort` is required.
9
9
  * If `exclude` is true, reasoning will be removed from the response. Default is false.
10
10
  */
11
- reasoning?: {
11
+ reasoningText?: {
12
12
  enabled?: boolean;
13
13
  exclude?: boolean;
14
14
  } & ({
@@ -16,22 +16,6 @@ type LLMGatewayProviderOptions = {
16
16
  } | {
17
17
  effort: 'high' | 'medium' | 'low';
18
18
  });
19
- /**
20
- * Image generation configuration for supported models (e.g., Google's image generation models).
21
- * Specifies aspect ratio and image resolution for generated images.
22
- */
23
- image_config?: {
24
- /**
25
- * The aspect ratio of the generated image.
26
- * Examples: "1:1", "16:9", "4:3", "5:4"
27
- */
28
- aspect_ratio?: string;
29
- /**
30
- * The resolution of the generated image.
31
- * Options: "1K" (1024x1024), "2K" (2048x2048), "4K"
32
- */
33
- image_size?: string;
34
- };
35
19
  /**
36
20
  * A unique identifier representing your end-user, which can
37
21
  * help LLMGateway to monitor and detect abuse.
@@ -43,11 +27,6 @@ type LLMGatewaySharedSettings = LLMGatewayProviderOptions & {
43
27
  * @deprecated use `reasoning` instead
44
28
  */
45
29
  includeReasoning?: boolean;
46
- /**
47
- * Reasoning effort level for models that support it.
48
- * Controls the computational effort applied to reasoning tasks.
49
- */
50
- reasoning_effort?: 'minimal' | 'low' | 'medium' | 'high';
51
30
  extraBody?: Record<string, unknown>;
52
31
  /**
53
32
  * Enable usage accounting to get detailed token usage information.
@@ -368,7 +347,7 @@ interface ProviderModelMapping {
368
347
  /**
369
348
  * Whether this model supports reasoning mode
370
349
  */
371
- reasoning?: boolean;
350
+ reasoningText?: boolean;
372
351
  /**
373
352
  * Whether this model supports the OpenAI responses API (defaults to true if reasoning is true)
374
353
  */
@@ -632,7 +611,7 @@ declare const models: [{
632
611
  readonly maxOutput: undefined;
633
612
  readonly streaming: true;
634
613
  readonly vision: true;
635
- readonly reasoning: true;
614
+ readonly reasoningText: true;
636
615
  readonly tools: false;
637
616
  }];
638
617
  readonly jsonOutput: true;
@@ -653,7 +632,7 @@ declare const models: [{
653
632
  readonly streaming: false;
654
633
  readonly vision: false;
655
634
  readonly tools: false;
656
- readonly reasoning: false;
635
+ readonly reasoningText: false;
657
636
  readonly supportsResponsesApi: false;
658
637
  }];
659
638
  readonly jsonOutput: true;
@@ -753,7 +732,7 @@ declare const models: [{
753
732
  readonly streaming: true;
754
733
  readonly vision: false;
755
734
  readonly tools: true;
756
- readonly reasoning: true;
735
+ readonly reasoningText: true;
757
736
  }];
758
737
  readonly jsonOutput: true;
759
738
  }, {
@@ -773,7 +752,7 @@ declare const models: [{
773
752
  readonly streaming: true;
774
753
  readonly vision: false;
775
754
  readonly tools: true;
776
- readonly reasoning: true;
755
+ readonly reasoningText: true;
777
756
  }];
778
757
  readonly jsonOutput: true;
779
758
  }, {
@@ -791,7 +770,7 @@ declare const models: [{
791
770
  readonly requestPrice: 0;
792
771
  readonly contextSize: 400000;
793
772
  readonly maxOutput: 128000;
794
- readonly reasoning: true;
773
+ readonly reasoningText: true;
795
774
  readonly streaming: true;
796
775
  readonly vision: true;
797
776
  readonly tools: true;
@@ -813,7 +792,7 @@ declare const models: [{
813
792
  readonly requestPrice: 0;
814
793
  readonly contextSize: 400000;
815
794
  readonly maxOutput: 128000;
816
- readonly reasoning: true;
795
+ readonly reasoningText: true;
817
796
  readonly streaming: true;
818
797
  readonly vision: true;
819
798
  readonly tools: true;
@@ -835,7 +814,7 @@ declare const models: [{
835
814
  readonly requestPrice: 0;
836
815
  readonly contextSize: 400000;
837
816
  readonly maxOutput: 128000;
838
- readonly reasoning: true;
817
+ readonly reasoningText: true;
839
818
  readonly streaming: true;
840
819
  readonly vision: false;
841
820
  readonly tools: true;
@@ -899,7 +878,7 @@ declare const models: [{
899
878
  readonly maxOutput: 8192;
900
879
  readonly streaming: true;
901
880
  readonly vision: false;
902
- readonly reasoning: true;
881
+ readonly reasoningText: true;
903
882
  readonly tools: true;
904
883
  }];
905
884
  }, {
@@ -994,7 +973,7 @@ declare const models: [{
994
973
  readonly maxOutput: 32000;
995
974
  readonly streaming: true;
996
975
  readonly vision: true;
997
- readonly reasoning: true;
976
+ readonly reasoningText: true;
998
977
  readonly tools: true;
999
978
  }];
1000
979
  }, {
@@ -1378,7 +1357,7 @@ declare const models: [{
1378
1357
  readonly contextSize: 1000000;
1379
1358
  readonly maxOutput: undefined;
1380
1359
  readonly streaming: true;
1381
- readonly reasoning: false;
1360
+ readonly reasoningText: false;
1382
1361
  readonly vision: false;
1383
1362
  readonly tools: false;
1384
1363
  }];
@@ -1397,7 +1376,7 @@ declare const models: [{
1397
1376
  readonly contextSize: 1000000;
1398
1377
  readonly maxOutput: undefined;
1399
1378
  readonly streaming: true;
1400
- readonly reasoning: false;
1379
+ readonly reasoningText: false;
1401
1380
  readonly vision: false;
1402
1381
  readonly tools: false;
1403
1382
  }];
@@ -2308,7 +2287,7 @@ declare const models: [{
2308
2287
  readonly requestPrice: 0;
2309
2288
  readonly contextSize: 262000;
2310
2289
  readonly maxOutput: 8192;
2311
- readonly reasoning: false;
2290
+ readonly reasoningText: false;
2312
2291
  readonly streaming: true;
2313
2292
  readonly vision: false;
2314
2293
  readonly tools: true;
@@ -2328,7 +2307,7 @@ declare const models: [{
2328
2307
  readonly requestPrice: 0;
2329
2308
  readonly contextSize: 262000;
2330
2309
  readonly maxOutput: 8192;
2331
- readonly reasoning: true;
2310
+ readonly reasoningText: true;
2332
2311
  readonly streaming: true;
2333
2312
  readonly vision: false;
2334
2313
  readonly tools: true;
@@ -2560,7 +2539,7 @@ declare const models: [{
2560
2539
  readonly streaming: true;
2561
2540
  readonly vision: false;
2562
2541
  readonly tools: true;
2563
- readonly reasoning: true;
2542
+ readonly reasoningText: true;
2564
2543
  }];
2565
2544
  readonly jsonOutput: true;
2566
2545
  }, {
@@ -2615,7 +2594,7 @@ declare const models: [{
2615
2594
  readonly requestPrice: 0;
2616
2595
  readonly contextSize: 131072;
2617
2596
  readonly maxOutput: 32768;
2618
- readonly reasoning: true;
2597
+ readonly reasoningText: true;
2619
2598
  readonly reasoningOutput: "omit";
2620
2599
  readonly streaming: true;
2621
2600
  readonly vision: false;
@@ -2656,7 +2635,7 @@ declare const models: [{
2656
2635
  readonly requestPrice: 0;
2657
2636
  readonly contextSize: 256000;
2658
2637
  readonly maxOutput: 32800;
2659
- readonly reasoning: true;
2638
+ readonly reasoningText: true;
2660
2639
  readonly reasoningOutput: "omit";
2661
2640
  readonly streaming: true;
2662
2641
  readonly vision: true;
@@ -2877,7 +2856,7 @@ declare const models: [{
2877
2856
  readonly contextSize: 128000;
2878
2857
  readonly maxOutput: undefined;
2879
2858
  readonly streaming: true;
2880
- readonly reasoning: true;
2859
+ readonly reasoningText: true;
2881
2860
  readonly vision: false;
2882
2861
  readonly tools: true;
2883
2862
  }];
@@ -2897,7 +2876,7 @@ declare const models: [{
2897
2876
  readonly contextSize: 128000;
2898
2877
  readonly maxOutput: undefined;
2899
2878
  readonly streaming: true;
2900
- readonly reasoning: true;
2879
+ readonly reasoningText: true;
2901
2880
  readonly reasoningOutput: "omit";
2902
2881
  readonly vision: true;
2903
2882
  readonly tools: true;
@@ -2918,7 +2897,7 @@ declare const models: [{
2918
2897
  readonly contextSize: 128000;
2919
2898
  readonly maxOutput: undefined;
2920
2899
  readonly streaming: true;
2921
- readonly reasoning: false;
2900
+ readonly reasoningText: false;
2922
2901
  readonly vision: false;
2923
2902
  readonly tools: true;
2924
2903
  }];
@@ -2938,7 +2917,7 @@ declare const models: [{
2938
2917
  readonly contextSize: 128000;
2939
2918
  readonly maxOutput: undefined;
2940
2919
  readonly streaming: true;
2941
- readonly reasoning: true;
2920
+ readonly reasoningText: true;
2942
2921
  readonly vision: false;
2943
2922
  readonly tools: true;
2944
2923
  }];
@@ -2958,7 +2937,7 @@ declare const models: [{
2958
2937
  readonly contextSize: 128000;
2959
2938
  readonly maxOutput: undefined;
2960
2939
  readonly streaming: true;
2961
- readonly reasoning: false;
2940
+ readonly reasoningText: false;
2962
2941
  readonly vision: false;
2963
2942
  readonly tools: true;
2964
2943
  }];
@@ -2978,7 +2957,7 @@ declare const models: [{
2978
2957
  readonly contextSize: 128000;
2979
2958
  readonly maxOutput: undefined;
2980
2959
  readonly streaming: true;
2981
- readonly reasoning: false;
2960
+ readonly reasoningText: false;
2982
2961
  readonly vision: false;
2983
2962
  readonly tools: true;
2984
2963
  }];
@@ -2998,7 +2977,7 @@ declare const models: [{
2998
2977
  readonly contextSize: 128000;
2999
2978
  readonly maxOutput: undefined;
3000
2979
  readonly streaming: true;
3001
- readonly reasoning: false;
2980
+ readonly reasoningText: false;
3002
2981
  readonly vision: false;
3003
2982
  readonly tools: true;
3004
2983
  }];