@tiledesk/tiledesk-tybot-connector 2.0.27 → 2.0.28-rc1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@tiledesk/tiledesk-tybot-connector",
3
- "version": "2.0.27",
3
+ "version": "2.0.28-rc1",
4
4
  "description": "Tiledesk Tybot connector",
5
5
  "main": "index.js",
6
6
  "scripts": {
@@ -12,6 +12,7 @@ const utils = require("../../utils/HttpUtils");
12
12
  const httpUtils = require("../../utils/HttpUtils");
13
13
  const integrationService = require("../../services/IntegrationService");
14
14
  const { Logger } = require("../../Logger");
15
+ const assert = require("assert");
15
16
 
16
17
 
17
18
  class DirAiPrompt {
@@ -89,7 +90,8 @@ class DirAiPrompt {
89
90
  const filler = new Filler();
90
91
  const filled_question = filler.fill(action.question, requestVariables);
91
92
  const filled_context = filler.fill(action.context, requestVariables);
92
-
93
+ const filled_model = filler.fill(action.model, requestVariables);
94
+
93
95
  if (action.history) {
94
96
  this.logger.native("[AI Prompt] using chat transcript");
95
97
  let transcript_string = await TiledeskChatbot.getParameterStatic(
@@ -151,7 +153,7 @@ class DirAiPrompt {
151
153
  let json = {
152
154
  question: filled_question,
153
155
  llm: action.llm,
154
- model: action.model,
156
+ model: filled_model,
155
157
  llm_key: key,
156
158
  temperature: action.temperature,
157
159
  max_tokens: action.max_tokens
@@ -519,11 +519,14 @@ class DirAskGPTV2 {
519
519
  let isHybrid = hybrid === true;
520
520
  return new Promise((resolve) => {
521
521
  let engine = {
522
- name: "pinecone",
523
- type: isHybrid ? "serverless" : process.env.PINECONE_TYPE,
522
+ name: process.env.VECTOR_STORE_NAME || "pinecone",
523
+ type: isHybrid ? process.env.INDEX_TYPE_HYBRID || "serverless" : process.env.INDEX_TYPE || process.env.PINECONE_TYPE || 'serverless',
524
524
  apikey: "",
525
- vector_size: 1536,
526
- index_name: isHybrid ? process.env.PINECONE_INDEX_HYBRID : process.env.PINECONE_INDEX
525
+ vector_size: process.env.VECTOR_SIZE || 1536,
526
+ index_name: isHybrid ? process.env.INDEX_NAME_HYBRID || process.env.PINECONE_INDEX_HYBRID || "llm-sample-index-hybrid" : process.env.INDEX_NAME || process.env.PINECONE_INDEX || "llm-sample-index",
527
+ ...(process.env.VECTOR_STORE_HOST && { host: process.env.VECTOR_STORE_HOST }),
528
+ ...(process.env.VECTOR_STORE_PORT && { port: process.env.VECTOR_STORE_PORT }),
529
+ ...(process.env.VECTOR_STORE_DEPLOYMENT && { deployment: process.env.VECTOR_STORE_DEPLOYMENT })
527
530
  }
528
531
  resolve(engine);
529
532
  })