backend-manager 5.0.126 → 5.0.127

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "backend-manager",
3
- "version": "5.0.126",
3
+ "version": "5.0.127",
4
4
  "description": "Quick tools for developing Firebase functions",
5
5
  "main": "src/manager/index.js",
6
6
  "bin": {
@@ -21,6 +21,7 @@ const MODEL_TABLE = {
21
21
  provider: 'openai',
22
22
  features: {
23
23
  json: true,
24
+ temperature: false,
24
25
  },
25
26
  },
26
27
  'gpt-5.2': {
@@ -29,6 +30,7 @@ const MODEL_TABLE = {
29
30
  provider: 'openai',
30
31
  features: {
31
32
  json: true,
33
+ temperature: false,
32
34
  },
33
35
  },
34
36
  'gpt-5.1': {
@@ -37,6 +39,7 @@ const MODEL_TABLE = {
37
39
  provider: 'openai',
38
40
  features: {
39
41
  json: true,
42
+ temperature: false,
40
43
  },
41
44
  },
42
45
  'gpt-5': {
@@ -45,6 +48,7 @@ const MODEL_TABLE = {
45
48
  provider: 'openai',
46
49
  features: {
47
50
  json: true,
51
+ temperature: false,
48
52
  },
49
53
  },
50
54
  'gpt-5-mini': {
@@ -53,6 +57,7 @@ const MODEL_TABLE = {
53
57
  provider: 'openai',
54
58
  features: {
55
59
  json: true,
60
+ temperature: false,
56
61
  },
57
62
  },
58
63
  'gpt-5-nano': {
@@ -61,6 +66,7 @@ const MODEL_TABLE = {
61
66
  provider: 'openai',
62
67
  features: {
63
68
  json: true,
69
+ temperature: false,
64
70
  },
65
71
  },
66
72
  // GPT-4.5
@@ -793,16 +799,22 @@ function makeRequest(mode, options, self, prompt, message, user, _log) {
793
799
  const history = formatHistory(options, prompt, message, _log);
794
800
 
795
801
  // Set request
802
+ const modelConfig = getModelConfig(options.model);
803
+
796
804
  request.url = 'https://api.openai.com/v1/responses';
797
805
  request.body = {
798
806
  model: options.model,
799
807
  input: history,
800
808
  user: user,
801
- temperature: options.temperature,
802
809
  max_output_tokens: options.maxTokens,
803
810
  text: resolveFormatting(options),
804
811
  reasoning: resolveReasoning(options),
805
812
  }
813
+
814
+ // Only include temperature if the model supports it
815
+ if (modelConfig.features?.temperature !== false) {
816
+ request.body.temperature = options.temperature;
817
+ }
806
818
  }
807
819
 
808
820
  // Request