ide-assi 0.271.0 → 0.273.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -193447,6 +193447,7 @@ class IdeAi
193447
193447
  configuration: {
193448
193448
  baseURL: "https://openrouter.ai/api/v1",
193449
193449
  },
193450
+ temperature: 0,
193450
193451
  });
193451
193452
  break;
193452
193453
  case "gemini":
@@ -193459,6 +193460,7 @@ class IdeAi
193459
193460
  this.#model = new Ollama({
193460
193461
  model: this.#parent.settings.model,
193461
193462
  host: this.#parent.settings.ollamaUrl,
193463
+ temperature: 0,
193462
193464
  });
193463
193465
  break;
193464
193466
  }
@@ -193606,13 +193608,14 @@ class IdeAi
193606
193608
  #invoke = async (path, params) => {
193607
193609
  const prompt = await IdeUtils.generatePrompt(path, params);
193608
193610
 
193609
-
193610
193611
  try {
193611
193612
  const response = await this.#model.invoke([
193612
193613
  //new SystemMessage(systemMessage),
193613
193614
  new HumanMessage(prompt),
193614
193615
  ]);
193615
193616
 
193617
+ //console.log(response);
193618
+
193616
193619
  return IdeUtils.extractResponse(response, this.#parent.settings.server);
193617
193620
  }
193618
193621
  catch (error) {
@@ -193443,6 +193443,7 @@ class IdeAi
193443
193443
  configuration: {
193444
193444
  baseURL: "https://openrouter.ai/api/v1",
193445
193445
  },
193446
+ temperature: 0,
193446
193447
  });
193447
193448
  break;
193448
193449
  case "gemini":
@@ -193455,6 +193456,7 @@ class IdeAi
193455
193456
  this.#model = new Ollama({
193456
193457
  model: this.#parent.settings.model,
193457
193458
  host: this.#parent.settings.ollamaUrl,
193459
+ temperature: 0,
193458
193460
  });
193459
193461
  break;
193460
193462
  }
@@ -193602,13 +193604,14 @@ class IdeAi
193602
193604
  #invoke = async (path, params) => {
193603
193605
  const prompt = await IdeUtils.generatePrompt(path, params);
193604
193606
 
193605
-
193606
193607
  try {
193607
193608
  const response = await this.#model.invoke([
193608
193609
  //new SystemMessage(systemMessage),
193609
193610
  new HumanMessage(prompt),
193610
193611
  ]);
193611
193612
 
193613
+ //console.log(response);
193614
+
193612
193615
  return IdeUtils.extractResponse(response, this.#parent.settings.server);
193613
193616
  }
193614
193617
  catch (error) {
@@ -28,6 +28,7 @@ export class IdeAi
28
28
  configuration: {
29
29
  baseURL: "https://openrouter.ai/api/v1",
30
30
  },
31
+ temperature: 0,
31
32
  });
32
33
  break;
33
34
  case "gemini":
@@ -40,6 +41,7 @@ export class IdeAi
40
41
  this.#model = new Ollama({
41
42
  model: this.#parent.settings.model,
42
43
  host: this.#parent.settings.ollamaUrl,
44
+ temperature: 0,
43
45
  });
44
46
  break;
45
47
  }
@@ -187,13 +189,14 @@ export class IdeAi
187
189
  #invoke = async (path, params) => {
188
190
  const prompt = await IdeUtils.generatePrompt(path, params);
189
191
 
190
-
191
192
  try {
192
193
  const response = await this.#model.invoke([
193
194
  //new SystemMessage(systemMessage),
194
195
  new HumanMessage(prompt),
195
196
  ]);
196
197
 
198
+ //console.log(response);
199
+
197
200
  return IdeUtils.extractResponse(response, this.#parent.settings.server);
198
201
  }
199
202
  catch (error) {
package/package.json CHANGED
@@ -1,7 +1,7 @@
1
1
  {
2
2
  "name": "ide-assi",
3
3
  "type": "module",
4
- "version": "0.271.0",
4
+ "version": "0.273.0",
5
5
  "main": "dist/index.js",
6
6
  "types": "dist/index.d.ts",
7
7
  "exports": {
@@ -28,6 +28,7 @@ export class IdeAi
28
28
  configuration: {
29
29
  baseURL: "https://openrouter.ai/api/v1",
30
30
  },
31
+ temperature: 0,
31
32
  });
32
33
  break;
33
34
  case "gemini":
@@ -40,6 +41,7 @@ export class IdeAi
40
41
  this.#model = new Ollama({
41
42
  model: this.#parent.settings.model,
42
43
  host: this.#parent.settings.ollamaUrl,
44
+ temperature: 0,
43
45
  });
44
46
  break;
45
47
  }
@@ -187,13 +189,14 @@ export class IdeAi
187
189
  #invoke = async (path, params) => {
188
190
  const prompt = await IdeUtils.generatePrompt(path, params);
189
191
 
190
-
191
192
  try {
192
193
  const response = await this.#model.invoke([
193
194
  //new SystemMessage(systemMessage),
194
195
  new HumanMessage(prompt),
195
196
  ]);
196
197
 
198
+ //console.log(response);
199
+
197
200
  return IdeUtils.extractResponse(response, this.#parent.settings.server);
198
201
  }
199
202
  catch (error) {