@cognigy/rest-api-client 2025.12.0 → 2025.13.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (55) hide show
  1. package/CHANGELOG.md +5 -0
  2. package/build/apigroups/ResourcesAPIGroup_2_0.js +4 -0
  3. package/build/shared/charts/descriptors/analytics/trackGoal.js +3 -1
  4. package/build/shared/charts/descriptors/index.js +5 -0
  5. package/build/shared/charts/descriptors/knowledgeSearch/searchExtractOutput.js +4 -2
  6. package/build/shared/charts/descriptors/message/question/question.js +12 -1
  7. package/build/shared/charts/descriptors/service/GPTPrompt.js +15 -1
  8. package/build/shared/charts/descriptors/service/aiAgent/aiAgentJob.js +32 -173
  9. package/build/shared/charts/descriptors/service/aiAgent/aiAgentJobTool.js +2 -2
  10. package/build/shared/charts/descriptors/service/aiAgent/helpers/createToolDefinitions.js +175 -0
  11. package/build/shared/charts/descriptors/service/aiAgent/loadAiAgent.js +194 -0
  12. package/build/shared/charts/descriptors/service/handoverV2.js +1 -1
  13. package/build/shared/charts/descriptors/service/index.js +11 -1
  14. package/build/shared/charts/descriptors/service/llmPrompt/LLMPromptV2.js +922 -0
  15. package/build/shared/charts/descriptors/service/llmPrompt/llmPromptDefault.js +31 -0
  16. package/build/shared/charts/descriptors/service/llmPrompt/llmPromptMCPTool.js +196 -0
  17. package/build/shared/charts/descriptors/service/llmPrompt/llmPromptTool.js +139 -0
  18. package/build/shared/constants.js +1 -5
  19. package/build/shared/interfaces/debugEvents/IGoalCompletedEventPayload.js +3 -0
  20. package/build/shared/interfaces/debugEvents/TDebugEventType.js +1 -0
  21. package/build/shared/interfaces/generativeAI/IGenerativeAIModels.js +12 -1
  22. package/build/shared/interfaces/resources/knowledgeStore/IKnowledgeChunk.js +2 -1
  23. package/build/shared/interfaces/resources/settings/IGenerativeAISettings.js +5 -18
  24. package/build/shared/interfaces/restAPI/operations/generateOutput/v2.0/index.js +3 -0
  25. package/build/shared/interfaces/security/IPermission.js +2 -0
  26. package/build/shared/interfaces/security/IRole.js +3 -1
  27. package/build/shared/interfaces/security/index.js +1 -1
  28. package/dist/esm/apigroups/ResourcesAPIGroup_2_0.js +4 -0
  29. package/dist/esm/shared/charts/descriptors/analytics/trackGoal.js +3 -1
  30. package/dist/esm/shared/charts/descriptors/index.js +6 -1
  31. package/dist/esm/shared/charts/descriptors/knowledgeSearch/searchExtractOutput.js +4 -2
  32. package/dist/esm/shared/charts/descriptors/message/question/question.js +12 -1
  33. package/dist/esm/shared/charts/descriptors/service/GPTPrompt.js +15 -1
  34. package/dist/esm/shared/charts/descriptors/service/aiAgent/aiAgentJob.js +33 -174
  35. package/dist/esm/shared/charts/descriptors/service/aiAgent/aiAgentJobTool.js +2 -2
  36. package/dist/esm/shared/charts/descriptors/service/aiAgent/helpers/createToolDefinitions.js +172 -0
  37. package/dist/esm/shared/charts/descriptors/service/aiAgent/loadAiAgent.js +192 -0
  38. package/dist/esm/shared/charts/descriptors/service/handoverV2.js +1 -1
  39. package/dist/esm/shared/charts/descriptors/service/index.js +5 -0
  40. package/dist/esm/shared/charts/descriptors/service/llmPrompt/LLMPromptV2.js +909 -0
  41. package/dist/esm/shared/charts/descriptors/service/llmPrompt/llmPromptDefault.js +28 -0
  42. package/dist/esm/shared/charts/descriptors/service/llmPrompt/llmPromptMCPTool.js +193 -0
  43. package/dist/esm/shared/charts/descriptors/service/llmPrompt/llmPromptTool.js +136 -0
  44. package/dist/esm/shared/constants.js +1 -5
  45. package/dist/esm/shared/interfaces/debugEvents/IGoalCompletedEventPayload.js +2 -0
  46. package/dist/esm/shared/interfaces/debugEvents/TDebugEventType.js +1 -0
  47. package/dist/esm/shared/interfaces/generativeAI/IGenerativeAIModels.js +12 -1
  48. package/dist/esm/shared/interfaces/resources/knowledgeStore/IKnowledgeChunk.js +2 -1
  49. package/dist/esm/shared/interfaces/resources/settings/IGenerativeAISettings.js +4 -17
  50. package/dist/esm/shared/interfaces/restAPI/operations/generateOutput/v2.0/index.js +2 -0
  51. package/dist/esm/shared/interfaces/security/IPermission.js +2 -0
  52. package/dist/esm/shared/interfaces/security/IRole.js +3 -1
  53. package/dist/esm/shared/interfaces/security/index.js +1 -1
  54. package/package.json +1 -1
  55. package/types/index.d.ts +83 -42
@@ -0,0 +1,31 @@
1
+ "use strict";
2
+ Object.defineProperty(exports, "__esModule", { value: true });
3
+ exports.LLM_PROMPT_DEFAULT = void 0;
4
+ /* Custom modules */
5
+ const createNodeDescriptor_1 = require("../../../createNodeDescriptor");
6
+ exports.LLM_PROMPT_DEFAULT = (0, createNodeDescriptor_1.createNodeDescriptor)({
7
+ type: "llmPromptDefault",
8
+ defaultLabel: "Default",
9
+ parentType: "llmPromptV2",
10
+ constraints: {
11
+ editable: false,
12
+ deletable: false,
13
+ collapsable: true,
14
+ creatable: false,
15
+ movable: false,
16
+ placement: {
17
+ predecessor: {
18
+ whitelist: []
19
+ }
20
+ }
21
+ },
22
+ fields: [],
23
+ sections: [],
24
+ form: [],
25
+ appearance: {
26
+ color: "white",
27
+ textColor: "#252525",
28
+ variant: "mini",
29
+ },
30
+ });
31
+ //# sourceMappingURL=llmPromptDefault.js.map
@@ -0,0 +1,196 @@
1
+ "use strict";
2
+ Object.defineProperty(exports, "__esModule", { value: true });
3
+ exports.LLM_PROMPT_MCP_TOOL = void 0;
4
+ /* Custom modules */
5
+ const createNodeDescriptor_1 = require("../../../createNodeDescriptor");
6
+ exports.LLM_PROMPT_MCP_TOOL = (0, createNodeDescriptor_1.createNodeDescriptor)({
7
+ type: "llmPromptMCPTool",
8
+ defaultLabel: "MCP Tool",
9
+ summary: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__SUMMARY",
10
+ parentType: "llmPromptV2",
11
+ constraints: {
12
+ editable: true,
13
+ deletable: true,
14
+ collapsable: true,
15
+ creatable: true,
16
+ movable: true,
17
+ placement: {
18
+ predecessor: {
19
+ whitelist: [],
20
+ },
21
+ },
22
+ childFlowCreatable: false,
23
+ },
24
+ preview: {
25
+ type: "text",
26
+ key: "name",
27
+ },
28
+ fields: [
29
+ {
30
+ type: "description",
31
+ key: "cachingWarning",
32
+ label: " ",
33
+ params: {
34
+ text: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__CACHING_WARNING__TEXT",
35
+ },
36
+ condition: {
37
+ key: "cacheTools",
38
+ value: false,
39
+ }
40
+ },
41
+ {
42
+ key: "name",
43
+ label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__NAME__LABEL",
44
+ description: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__NAME__DESCRIPTION",
45
+ type: "cognigyText",
46
+ params: {
47
+ required: true,
48
+ rows: 1,
49
+ multiline: false,
50
+ maxLength: 64,
51
+ },
52
+ },
53
+ {
54
+ type: "description",
55
+ key: "mcpWarning",
56
+ label: " ",
57
+ params: {
58
+ text: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__MCP_WARNING__TEXT",
59
+ },
60
+ },
61
+ {
62
+ key: "mcpServerUrl",
63
+ label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__MCP_SERVER_URL__LABEL",
64
+ description: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__MCP_SERVER_URL__DESCRIPTION",
65
+ type: "cognigyText",
66
+ params: {
67
+ required: true,
68
+ rows: 5,
69
+ multiline: true,
70
+ },
71
+ },
72
+ {
73
+ key: "timeout",
74
+ label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__TIMEOUT__LABEL",
75
+ description: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__TIMEOUT__DESCRIPTION",
76
+ type: "slider",
77
+ defaultValue: 12,
78
+ params: {
79
+ min: 1,
80
+ max: 30,
81
+ step: 1,
82
+ },
83
+ },
84
+ {
85
+ key: "debugMessageFetchedTools",
86
+ type: "toggle",
87
+ label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__DEBUG_TOOLS__LABEL",
88
+ description: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__DEBUG_TOOLS__DESCRIPTION",
89
+ defaultValue: true,
90
+ },
91
+ {
92
+ key: "debugMessageParameters",
93
+ type: "toggle",
94
+ label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__DEBUG_PARAMETERS__LABEL",
95
+ description: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__DEBUG_PARAMETERS__DESCRIPTION",
96
+ defaultValue: false,
97
+ condition: {
98
+ key: "debugMessageFetchedTools",
99
+ value: true,
100
+ },
101
+ },
102
+ {
103
+ key: "debugMessage",
104
+ type: "toggle",
105
+ label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__DEBUG_MESSAGE__LABEL",
106
+ description: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__DEBUG_MESSAGE__DESCRIPTION",
107
+ defaultValue: true,
108
+ },
109
+ {
110
+ key: "cacheTools",
111
+ type: "toggle",
112
+ label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__CACHE_TOOLS__LABEL",
113
+ description: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__CACHE_TOOLS__DESCRIPTION",
114
+ defaultValue: true,
115
+ },
116
+ {
117
+ key: "condition",
118
+ label: "UI__NODE_EDITOR__SERVICE__AI_AGENT_TOOL__FIELDS__CONDITION__LABEL",
119
+ description: "UI__NODE_EDITOR__SERVICE__AI_AGENT_TOOL__FIELDS__CONDITION__DESCRIPTION",
120
+ type: "cognigyText",
121
+ defaultValue: "",
122
+ },
123
+ {
124
+ key: "toolFilter",
125
+ type: "select",
126
+ label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__TOOL_FILTER__LABEL",
127
+ description: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__TOOL_FILTER__DESCRIPTION",
128
+ defaultValue: "none",
129
+ params: {
130
+ options: [
131
+ {
132
+ label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__TOOL_FILTER__OPTIONS__NONE__LABEL",
133
+ value: "none",
134
+ },
135
+ {
136
+ label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__TOOL_FILTER__OPTIONS__WHITELIST__LABEL",
137
+ value: "whitelist",
138
+ },
139
+ {
140
+ label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__TOOL_FILTER__OPTIONS__BLACKLIST__LABEL",
141
+ value: "blacklist",
142
+ },
143
+ ],
144
+ },
145
+ },
146
+ {
147
+ key: "whitelist",
148
+ type: "cognigyTextArray",
149
+ label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__WHITELIST__LABEL",
150
+ description: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__WHITELIST__DESCRIPTION",
151
+ condition: {
152
+ key: "toolFilter",
153
+ value: "whitelist",
154
+ },
155
+ },
156
+ {
157
+ key: "blacklist",
158
+ type: "cognigyTextArray",
159
+ label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__BLACKLIST__LABEL",
160
+ description: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__BLACKLIST__DESCRIPTION",
161
+ condition: {
162
+ key: "toolFilter",
163
+ value: "blacklist",
164
+ },
165
+ },
166
+ ],
167
+ sections: [
168
+ {
169
+ key: "debugging",
170
+ label: "UI__NODE_EDITOR__SERVICE__AI_AGENT_JOB__SECTIONS__DEBUG_SETTINGS__LABEL",
171
+ defaultCollapsed: true,
172
+ fields: ["debugMessageFetchedTools", "debugMessageParameters", "debugMessage"],
173
+ },
174
+ {
175
+ key: "advanced",
176
+ label: "UI__NODE_EDITOR__SERVICE__AI_AGENT_JOB__SECTIONS__ADVANCED__LABEL",
177
+ defaultCollapsed: true,
178
+ fields: ["cacheTools", "condition", "toolFilter", "whitelist", "blacklist"],
179
+ },
180
+ ],
181
+ form: [
182
+ { type: "field", key: "cachingWarning" },
183
+ { type: "field", key: "name" },
184
+ { type: "field", key: "mcpWarning" },
185
+ { type: "field", key: "mcpServerUrl" },
186
+ { type: "field", key: "timeout" },
187
+ { type: "section", key: "debugging" },
188
+ { type: "section", key: "advanced" },
189
+ ],
190
+ appearance: {
191
+ color: "white",
192
+ textColor: "#252525",
193
+ variant: "mini",
194
+ },
195
+ });
196
+ //# sourceMappingURL=llmPromptMCPTool.js.map
@@ -0,0 +1,139 @@
1
+ "use strict";
2
+ Object.defineProperty(exports, "__esModule", { value: true });
3
+ exports.LLM_PROMPT_TOOL = void 0;
4
+ /* Custom modules */
5
+ const createNodeDescriptor_1 = require("../../../createNodeDescriptor");
6
+ exports.LLM_PROMPT_TOOL = (0, createNodeDescriptor_1.createNodeDescriptor)({
7
+ type: "llmPromptTool",
8
+ defaultLabel: "Tool",
9
+ parentType: "llmPromptV2",
10
+ constraints: {
11
+ editable: true,
12
+ deletable: true,
13
+ collapsable: true,
14
+ creatable: true,
15
+ movable: true,
16
+ placement: {
17
+ predecessor: {
18
+ whitelist: []
19
+ }
20
+ },
21
+ childFlowCreatable: false
22
+ },
23
+ preview: {
24
+ type: "text",
25
+ key: "toolId"
26
+ },
27
+ fields: [
28
+ {
29
+ key: "toolId",
30
+ label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_TOOL__FIELDS__TOOL_ID__LABEL",
31
+ description: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_TOOL__FIELDS__TOOL_ID__DESCRIPTION",
32
+ type: "cognigyLLMText",
33
+ defaultValue: "unlock_account",
34
+ params: {
35
+ required: true,
36
+ rows: 1,
37
+ multiline: false,
38
+ maxLength: 64,
39
+ }
40
+ },
41
+ {
42
+ key: "description",
43
+ label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_TOOL__FIELDS__DESCRIPTION__LABEL",
44
+ description: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_TOOL__FIELDS__DESCRIPTION__DESCRIPTION",
45
+ type: "cognigyLLMText",
46
+ defaultValue: "This tool unlocks a locked user account.",
47
+ params: {
48
+ required: true,
49
+ rows: 5,
50
+ multiline: true
51
+ }
52
+ },
53
+ {
54
+ key: "useParameters",
55
+ label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_TOOL__FIELDS__USE_PARAMETERS__LABEL",
56
+ description: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_TOOL__FIELDS__USE_PARAMETERS__DESCRIPTION",
57
+ type: "toggle",
58
+ defaultValue: false
59
+ },
60
+ {
61
+ key: "parameters",
62
+ label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_TOOL__FIELDS__PARAMETERS__LABEL",
63
+ description: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_TOOL__FIELDS__PARAMETERS__DESCRIPTION",
64
+ type: "toolParameters",
65
+ defaultValue: `{
66
+ "type": "object",
67
+ "properties": {
68
+ "email": {
69
+ "type": "string",
70
+ "description": "User's login email for their account."
71
+ }
72
+ },
73
+ "required": ["email"],
74
+ "additionalProperties": false
75
+ }`,
76
+ params: {
77
+ required: false,
78
+ },
79
+ },
80
+ {
81
+ key: "debugMessage",
82
+ type: "toggle",
83
+ label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_TOOL__FIELDS__DEBUG_MESSAGE__LABEL",
84
+ description: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_TOOL__FIELDS__DEBUG_MESSAGE__DESCRIPTION",
85
+ defaultValue: true,
86
+ },
87
+ {
88
+ key: "condition",
89
+ label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_TOOL__FIELDS__CONDITION__LABEL",
90
+ description: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_TOOL__FIELDS__CONDITION__DESCRIPTION",
91
+ type: "cognigyText",
92
+ defaultValue: "",
93
+ },
94
+ ],
95
+ sections: [
96
+ {
97
+ key: "parameters",
98
+ label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_TOOL__SECTIONS__PARAMETERS__LABEL",
99
+ defaultCollapsed: true,
100
+ fields: [
101
+ "parameters",
102
+ ],
103
+ condition: {
104
+ key: "useParameters",
105
+ value: true
106
+ },
107
+ },
108
+ {
109
+ key: "debugging",
110
+ label: "UI__NODE_EDITOR__SERVICE__AI_AGENT_JOB__SECTIONS__DEBUG_SETTINGS__LABEL",
111
+ defaultCollapsed: true,
112
+ fields: [
113
+ "debugMessage",
114
+ ],
115
+ },
116
+ {
117
+ key: "advanced",
118
+ label: "UI__NODE_EDITOR__SERVICE__AI_AGENT_JOB__SECTIONS__ADVANCED__LABEL",
119
+ defaultCollapsed: true,
120
+ fields: [
121
+ "condition",
122
+ ],
123
+ },
124
+ ],
125
+ form: [
126
+ { type: "field", key: "toolId" },
127
+ { type: "field", key: "description" },
128
+ { type: "field", key: "useParameters" },
129
+ { type: "section", key: "parameters" },
130
+ { type: "section", key: "debugging" },
131
+ { type: "section", key: "advanced" },
132
+ ],
133
+ appearance: {
134
+ color: "white",
135
+ textColor: "#252525",
136
+ variant: "mini",
137
+ },
138
+ });
139
+ //# sourceMappingURL=llmPromptTool.js.map
@@ -54,17 +54,13 @@ exports.IP_NOT_AVAILABLE = "not available";
54
54
  exports.SECONDS_PER_MINUTE = 60;
55
55
  // Platform provided LLM
56
56
  exports.PLATFORM_PROVIDED_LLM_USE_CASES = [
57
- "intentSentenceGeneration",
58
- "flowGeneration",
59
- "lexiconGeneration",
60
- "generateNodeOutput",
61
57
  "designTimeGeneration",
62
58
  ];
63
59
  exports.DEPRECATED_GENAI_USE_CASES = [
64
60
  "intentSentenceGeneration",
65
61
  "flowGeneration",
66
- "lexiconGeneration",
67
62
  "generateNodeOutput",
63
+ "lexiconGeneration",
68
64
  ];
69
65
  exports.PLATFORM_PROVIDED_LLM_ID = "aaaaaaaa-aaaa-aaaa-aaaa-000000000000";
70
66
  exports.PLATFORM_PROVIDED_LLM_CONNECTION_ID = "aaaaaaaa-aaaa-aaaa-aaaa-000000000001";
@@ -0,0 +1,3 @@
1
+ "use strict";
2
+ Object.defineProperty(exports, "__esModule", { value: true });
3
+ //# sourceMappingURL=IGoalCompletedEventPayload.js.map
@@ -15,5 +15,6 @@ exports.arrayTDebugEventTypes = [
15
15
  "nluWarning",
16
16
  "debugMessage",
17
17
  "debugError",
18
+ "goalCompleted",
18
19
  ];
19
20
  //# sourceMappingURL=TDebugEventType.js.map
@@ -19,6 +19,9 @@ exports.generativeAIModels = [
19
19
  "gpt-4",
20
20
  "gpt-4o",
21
21
  "gpt-4o-mini",
22
+ "gpt-4.1",
23
+ "gpt-4.1-mini",
24
+ "gpt-4.1-nano",
22
25
  "luminous-extended-control",
23
26
  "claude-v1-100k",
24
27
  "claude-instant-v1",
@@ -26,6 +29,11 @@ exports.generativeAIModels = [
26
29
  "claude-3-haiku-20240307",
27
30
  "claude-3-sonnet-20240229",
28
31
  "claude-3-5-sonnet-20241022",
32
+ "claude-3-7-sonnet-20250219",
33
+ "claude-3-5-sonnet-latest",
34
+ "claude-3-7-sonnet-latest",
35
+ "claude-opus-4-0",
36
+ "claude-sonnet-4-0",
29
37
  "text-bison@001",
30
38
  "custom-model",
31
39
  "gemini-1.0-pro",
@@ -37,11 +45,14 @@ exports.generativeAIModels = [
37
45
  "amazon.nova-pro-v1:0",
38
46
  "amazon.nova-micro-v1:0",
39
47
  "anthropic.claude-3-5-sonnet-20240620-v1:0",
40
- "claude-3-7-sonnet-20250219",
41
48
  "mistral-large-2411",
42
49
  "mistral-small-2503",
43
50
  "pixtral-large-2411",
44
51
  "pixtral-12b-2409",
52
+ "mistral-large-latest",
53
+ "pixtral-large-latest",
54
+ "mistral-medium-latest",
55
+ "mistral-small-latest",
45
56
  ...exports.embeddingModels,
46
57
  ];
47
58
  exports.generativeAIProviders = [
@@ -12,7 +12,8 @@ exports.knowledgeChunkDataSchema = {
12
12
  order: { type: "number" },
13
13
  text: {
14
14
  type: "string",
15
- minLength: 0,
15
+ minLength: 1,
16
+ pattern: "\\S",
16
17
  maxLength: parseInt(process.env.MAX_CHUNK_SIZE, 10) || 2000,
17
18
  },
18
19
  embedding: {
@@ -1,6 +1,9 @@
1
1
  "use strict";
2
2
  Object.defineProperty(exports, "__esModule", { value: true });
3
- exports.generativeAISettingsDataSchema = exports.generativeAIUseCaseDataSchema = void 0;
3
+ exports.generativeAISettingsDataSchema = exports.generativeAIUseCaseDataSchema = exports.configurableGenerativeAIUseCases = void 0;
4
+ const IGenerativeAIModels_1 = require("../../generativeAI/IGenerativeAIModels");
5
+ const constants_1 = require("../../../constants");
6
+ exports.configurableGenerativeAIUseCases = IGenerativeAIModels_1.generativeAIUseCases.filter((useCase) => !constants_1.DEPRECATED_GENAI_USE_CASES.includes(useCase));
4
7
  exports.generativeAIUseCaseDataSchema = {
5
8
  title: "generativeAIMetaAzureOpenAIDataSchema",
6
9
  type: "object",
@@ -34,22 +37,10 @@ exports.generativeAISettingsDataSchema = {
34
37
  type: "object",
35
38
  properties: Object.assign({}, exports.generativeAIUseCaseDataSchema.properties)
36
39
  },
37
- intentSentenceGeneration: {
38
- type: "object",
39
- properties: Object.assign({}, exports.generativeAIUseCaseDataSchema.properties)
40
- },
41
40
  aiEnhancedOutputs: {
42
41
  type: "object",
43
42
  properties: Object.assign({}, exports.generativeAIUseCaseDataSchema.properties)
44
43
  },
45
- lexiconGeneration: {
46
- type: "object",
47
- properties: Object.assign({}, exports.generativeAIUseCaseDataSchema.properties)
48
- },
49
- flowGeneration: {
50
- type: "object",
51
- properties: Object.assign({}, exports.generativeAIUseCaseDataSchema.properties)
52
- },
53
44
  gptConversation: {
54
45
  type: "object",
55
46
  properties: Object.assign({}, exports.generativeAIUseCaseDataSchema.properties)
@@ -58,10 +49,6 @@ exports.generativeAISettingsDataSchema = {
58
49
  type: "object",
59
50
  properties: Object.assign({}, exports.generativeAIUseCaseDataSchema.properties)
60
51
  },
61
- generateNodeOutput: {
62
- type: "object",
63
- properties: Object.assign({}, exports.generativeAIUseCaseDataSchema.properties)
64
- },
65
52
  knowledgeSearch: {
66
53
  type: "object",
67
54
  properties: Object.assign({}, exports.generativeAIUseCaseDataSchema.properties)
@@ -81,7 +68,7 @@ exports.generativeAISettingsDataSchema = {
81
68
  aiAgent: {
82
69
  type: "object",
83
70
  properties: Object.assign({}, exports.generativeAIUseCaseDataSchema.properties)
84
- },
71
+ }
85
72
  }
86
73
  }
87
74
  }
@@ -0,0 +1,3 @@
1
+ "use strict";
2
+ Object.defineProperty(exports, "__esModule", { value: true });
3
+ //# sourceMappingURL=index.js.map
@@ -74,6 +74,7 @@ exports.organisationWidePermissions = [
74
74
  * - functions
75
75
  * - yesNoIntents
76
76
  * - dataPrivacySettings
77
+ * - simulator
77
78
  */
78
79
  exports.projectWidePermissions = [
79
80
  "agentAssistConfigs",
@@ -113,6 +114,7 @@ exports.projectWidePermissions = [
113
114
  "tokens",
114
115
  "yesNoIntents",
115
116
  "dataPrivacySettings",
117
+ "simulator"
116
118
  ];
117
119
  exports.availablePermissions = [
118
120
  ...exports.organisationWidePermissions,
@@ -88,6 +88,7 @@ exports.organisationWideRoles = [
88
88
  * - tokenEditor
89
89
  * - function_admin
90
90
  * - function_editor
91
+ * - simulator_admin
91
92
  */
92
93
  exports.projectWideRoles = [
93
94
  "agentAssistConfigAdmin",
@@ -129,7 +130,8 @@ exports.projectWideRoles = [
129
130
  "tokenEditor",
130
131
  "data_privacy_admin",
131
132
  "data_privacy_editor",
132
- "data_privacy_viewer"
133
+ "data_privacy_viewer",
134
+ "simulator_admin"
133
135
  ];
134
136
  exports.availableRoles = [
135
137
  ...exports.organisationWideRoles,
@@ -1,7 +1,7 @@
1
1
  "use strict";
2
2
  Object.defineProperty(exports, "__esModule", { value: true });
3
3
  exports.identityProviderSchema = exports.oidcIdentityProviderDataSchema = exports.samlIdentityProviderDataSchema = exports.operations = exports.availableRoles = exports.projectWideRoles = exports.organisationWideRoles = exports.availablePermissions = exports.COMPUTED_ACL_HASH_VERSION = void 0;
4
- exports.COMPUTED_ACL_HASH_VERSION = "v19";
4
+ exports.COMPUTED_ACL_HASH_VERSION = "v20";
5
5
  var IPermission_1 = require("./IPermission");
6
6
  Object.defineProperty(exports, "availablePermissions", { enumerable: true, get: function () { return IPermission_1.availablePermissions; } });
7
7
  var IRole_1 = require("./IRole");
@@ -615,6 +615,10 @@ export const ResourcesAPIGroup_2_0 = (instance) => {
615
615
  generateNluScores: (_a) => {
616
616
  var { projectId } = _a, args = __rest(_a, ["projectId"]);
617
617
  return GenericAPIFn(`/new/v2.0/projects/${projectId}/nlu/scores`, "POST", self)(args);
618
+ },
619
+ generateDesignTimeLLMOutput: (_a) => {
620
+ var { projectId } = _a, args = __rest(_a, ["projectId"]);
621
+ return GenericAPIFn(`/new/v2.0/projects/${projectId}/generate-output/design-time-llm`, "POST", self)(args);
618
622
  }
619
623
  };
620
624
  };
@@ -45,7 +45,7 @@ export const TRACK_GOAL = createNodeDescriptor({
45
45
  return;
46
46
  }
47
47
  const sessionState = api.getSessionStateCopy();
48
- const { selectedSteps, version, name, goalId, } = config.goal;
48
+ const { selectedSteps, referenceId, version, name, goalId, } = config.goal;
49
49
  const activeCycleIds = ((_a = sessionState.analytics) === null || _a === void 0 ? void 0 : _a.goalCycleIds) || {};
50
50
  let cycleId = activeCycleIds[goalId];
51
51
  const hasStartStep = selectedSteps.some(step => step.type === "start");
@@ -73,10 +73,12 @@ export const TRACK_GOAL = createNodeDescriptor({
73
73
  organisationId,
74
74
  sessionId,
75
75
  version,
76
+ referenceId,
76
77
  timestamp: new Date(),
77
78
  goalCycleId: cycleId,
78
79
  goalId,
79
80
  stepId: step.stepId,
81
+ stepType: step.type,
80
82
  endpointUrlToken,
81
83
  endpointName,
82
84
  endpointType,
@@ -11,7 +11,7 @@ import { REGEX_SLOT_FILLER, EXECUTE_COGNIGY_NLU, ADD_LEXICON_KEYPHRASE, FUZZY_SE
11
11
  import { KNOWLEDGE_SEARCH, KNOWLEDGE_SEARCH_V2, SEARCH_EXTRACT_OUTPUT } from "./knowledgeSearch";
12
12
  import { CONTINUOUS_ASR, DTMF, HANG_UP, PLAY, TRANSFER_VOICE, SESSION_SPEECH_PARAMETERS, USER_INPUT_TIMEOUT, SEND_METADATA, BARGE_IN, MUTE_SPEECH_INPUT, } from "./voice";
13
13
  import { ACTIVATE_PROFILE, COMPLETE_GOAL, DEACTIVATE_PROFILE, DELETE_PROFILE, MERGE_PROFILE, UPDATE_PROFILE, ADD_MEMORY, BLIND_MODE, OVERWRITE_ANALYTICS, SET_RATING, REQUEST_RATING, TRACK_GOAL, } from "./analytics";
14
- import { HANDOVER, HANDOVER_V2, CHECK_AGENT_AVAILABILITY, HTTP_REQUEST, HTTP_CONNECTION_BASIC, HTTP_CONNECTION_APIKEYAUTHKEY, HTTP_CONNECTION_APIKEYXKEY, HTTP_CONNECTION_OAUTH2, JWT_SECRET_CONNECTION, TRIGGER_FUNCTION, ON_SCHEDULING_ERROR, ON_SCHEDULED, GPT_PROMPT, CLOSE_HANDOVER, HANDOVER_INACTIVITY_TIMER, GPT_CONVERSATION, GPT_CONVERSATION_SUMMARY, LLM_ENTITY_EXTRACT, AI_AGENT_JOB, AI_AGENT_JOB_DEFAULT, AI_AGENT_JOB_TOOL, AI_AGENT_JOB_MCP_TOOL, AI_AGENT_JOB_CALL_MCP_TOOL, AI_AGENT_TOOL_ANSWER, AI_AGENT_HANDOVER, LIVE_AGENT_CONNECTION, RINGCENTRAL_ENGAGE_CONNECTION, CHATWOOT_CONNECTION, EIGHT_BY_EIGHT_CONNECTION, GENESYS_CLOUD_CONNECTION, GENESYS_CLOUD_CONNECTION_OM, LLM_MODERATE, NICECXONEAAH_AUTHENTICATION_CONNECTION, AIOPS_CENTER_WEBHOOKS_CONNECTION } from "./service";
14
+ import { HANDOVER, HANDOVER_V2, CHECK_AGENT_AVAILABILITY, HTTP_REQUEST, HTTP_CONNECTION_BASIC, HTTP_CONNECTION_APIKEYAUTHKEY, HTTP_CONNECTION_APIKEYXKEY, HTTP_CONNECTION_OAUTH2, JWT_SECRET_CONNECTION, TRIGGER_FUNCTION, ON_SCHEDULING_ERROR, ON_SCHEDULED, GPT_PROMPT, LLM_PROMPT_V2, LLM_PROMPT_DEFAULT, LLM_PROMPT_MCP_TOOL, LLM_PROMPT_TOOL, CLOSE_HANDOVER, HANDOVER_INACTIVITY_TIMER, GPT_CONVERSATION, GPT_CONVERSATION_SUMMARY, LLM_ENTITY_EXTRACT, AI_AGENT_JOB, AI_AGENT_JOB_DEFAULT, AI_AGENT_JOB_TOOL, AI_AGENT_JOB_MCP_TOOL, AI_AGENT_JOB_CALL_MCP_TOOL, AI_AGENT_TOOL_ANSWER, AI_AGENT_HANDOVER, LIVE_AGENT_CONNECTION, RINGCENTRAL_ENGAGE_CONNECTION, CHATWOOT_CONNECTION, EIGHT_BY_EIGHT_CONNECTION, GENESYS_CLOUD_CONNECTION, GENESYS_CLOUD_CONNECTION_OM, LLM_MODERATE, NICECXONEAAH_AUTHENTICATION_CONNECTION, LOAD_AI_AGENT, AIOPS_CENTER_WEBHOOKS_CONNECTION } from "./service";
15
15
  import { INIT_APP_SESSION, GET_APP_SESSION_PIN, SET_HTML_APP_STATE, SET_ADAPTIVE_CARD_APP_STATE, } from "./apps";
16
16
  import { SET_IFRAME_TILE, SET_HTML_TILE, SEND_TILE_DATA, SET_SECURE_FORMS_TILE, SET_ADAPTIVE_CARD_TILE, SET_AGENT_ASSIST_GRID, NEXT_ACTION_ASSIST, SENTIMENT_ASSIST, TRANSCRIPT_ASSIST, IDENTITY_ASSIST, KNOWLEDGE_ASSIST, } from "./agentAssist";
17
17
  import { ASSIST_INFO } from "./liveAgent";
@@ -88,6 +88,10 @@ const nodes = [
88
88
  CHECK_AGENT_AVAILABILITY,
89
89
  HTTP_REQUEST,
90
90
  GPT_PROMPT,
91
+ LLM_PROMPT_V2,
92
+ LLM_PROMPT_DEFAULT,
93
+ LLM_PROMPT_MCP_TOOL,
94
+ LLM_PROMPT_TOOL,
91
95
  GPT_CONVERSATION,
92
96
  GPT_CONVERSATION_SUMMARY,
93
97
  LLM_ENTITY_EXTRACT,
@@ -143,6 +147,7 @@ if (process.env.DISABLE_FEATURE_TRANSCRIPT_MANAGER !== "true") {
143
147
  nodes.push(AI_AGENT_JOB_CALL_MCP_TOOL);
144
148
  nodes.push(AI_AGENT_TOOL_ANSWER);
145
149
  nodes.push(AI_AGENT_HANDOVER);
150
+ nodes.push(LOAD_AI_AGENT);
146
151
  }
147
152
  if (process.env.FEATURE_USE_COGNIGY_LIVE_AGENT === "true") {
148
153
  nodes.push(ASSIST_INFO);
@@ -582,10 +582,12 @@ export const SEARCH_EXTRACT_OUTPUT = createNodeDescriptor({
582
582
  // as it doesn't make sense to check for follow ups in the first execution
583
583
  if (input.execution > 1) {
584
584
  // always remember the last thing the user said (needed later)
585
- lastRoundTrip = (_b = cognigy.lastConversationEntries) === null || _b === void 0 ? void 0 : _b.slice(1, 3).reverse().map(entry => "- " + (entry.source === "user" ? "USER: " : "BOT: ") + entry.text).join("\n");
585
+ lastRoundTrip = (_b = cognigy
586
+ .lastConversationEntries) === null || _b === void 0 ? void 0 : _b.slice(1, followUpDetectionSteps + 1).reverse().map(entry => "- " + (entry.source === "user" ? "USER: " : "BOT: ") + entry.text).join("\n");
586
587
  // if follow up detection is set to 2 or more, we use the conversation transcript
587
588
  // as reference. Start at the second entry, because the first one is the current
588
- const recentConversation = (_c = cognigy.lastConversationEntries) === null || _c === void 0 ? void 0 : _c.slice(1, followUpDetectionSteps + 1).reverse().map(entry => "- " + (entry.source === "user" ? "USER: " : "BOT: ") + entry.text).join("\n");
589
+ const recentConversation = (_c = cognigy
590
+ .lastConversationEntries) === null || _c === void 0 ? void 0 : _c.slice(1, followUpDetectionSteps + 1).reverse().map(entry => "- " + (entry.source === "user" ? "USER: " : "BOT: ") + entry.text).join("\n");
589
591
  prompt = `Below is the transcript of a conversation:
590
592
  ${recentConversation}
591
593
  USER: ${searchString}
@@ -307,7 +307,7 @@ export const QUESTION = createNodeDescriptor({
307
307
  },
308
308
  {
309
309
  key: "repromptLLMPrompt",
310
- type: "cognigyText",
310
+ type: "cognigyLLMText",
311
311
  label: "UI__NODE_EDITOR__MESSAGE__QUESTION__QUESTION__FIELDS__REPROMPT_LLM_PROMPT__LABEL",
312
312
  description: "UI__NODE_EDITOR__MESSAGE__QUESTION__QUESTION__FIELDS__REPROMPT_LLM_PROMPT__DESCRIPTION",
313
313
  defaultValue: `You are a chatbot that helps a user.
@@ -740,6 +740,17 @@ DO NOT talk about other topics. Do not offer general assistance.`,
740
740
  },
741
741
  defaultValue: true,
742
742
  },
743
+ {
744
+ key: "escalateAnswersRepeatHandoverMessage",
745
+ type: "toggle",
746
+ label: "UI__NODE_EDITOR__MESSAGE__QUESTION__QUESTION__FIELDS__ESCALATE_ANSWERS_REPEAT_HANDOVER_MESSAGE__LABEL",
747
+ description: "UI__NODE_EDITOR__MESSAGE__QUESTION__QUESTION__FIELDS__ESCALATE_ANSWERS_REPEAT_HANDOVER_MESSAGE__DESCRIPTION",
748
+ defaultValue: false,
749
+ condition: {
750
+ key: "escalateAnswersAction",
751
+ value: "handover"
752
+ }
753
+ },
743
754
  {
744
755
  key: "escalateAnswersHandoverCancelIntent",
745
756
  type: "cognigyText",