@cognigy/rest-api-client 2025.12.0 → 2025.13.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +5 -0
- package/build/apigroups/ResourcesAPIGroup_2_0.js +4 -0
- package/build/shared/charts/descriptors/analytics/trackGoal.js +3 -1
- package/build/shared/charts/descriptors/index.js +5 -0
- package/build/shared/charts/descriptors/knowledgeSearch/searchExtractOutput.js +4 -2
- package/build/shared/charts/descriptors/message/question/question.js +12 -1
- package/build/shared/charts/descriptors/service/GPTPrompt.js +15 -1
- package/build/shared/charts/descriptors/service/aiAgent/aiAgentJob.js +32 -173
- package/build/shared/charts/descriptors/service/aiAgent/aiAgentJobTool.js +2 -2
- package/build/shared/charts/descriptors/service/aiAgent/helpers/createToolDefinitions.js +175 -0
- package/build/shared/charts/descriptors/service/aiAgent/loadAiAgent.js +194 -0
- package/build/shared/charts/descriptors/service/handoverV2.js +1 -1
- package/build/shared/charts/descriptors/service/index.js +11 -1
- package/build/shared/charts/descriptors/service/llmPrompt/LLMPromptV2.js +922 -0
- package/build/shared/charts/descriptors/service/llmPrompt/llmPromptDefault.js +31 -0
- package/build/shared/charts/descriptors/service/llmPrompt/llmPromptMCPTool.js +196 -0
- package/build/shared/charts/descriptors/service/llmPrompt/llmPromptTool.js +139 -0
- package/build/shared/constants.js +1 -5
- package/build/shared/interfaces/debugEvents/IGoalCompletedEventPayload.js +3 -0
- package/build/shared/interfaces/debugEvents/TDebugEventType.js +1 -0
- package/build/shared/interfaces/generativeAI/IGenerativeAIModels.js +12 -1
- package/build/shared/interfaces/resources/knowledgeStore/IKnowledgeChunk.js +2 -1
- package/build/shared/interfaces/resources/settings/IGenerativeAISettings.js +5 -18
- package/build/shared/interfaces/restAPI/operations/generateOutput/v2.0/index.js +3 -0
- package/build/shared/interfaces/security/IPermission.js +2 -0
- package/build/shared/interfaces/security/IRole.js +3 -1
- package/build/shared/interfaces/security/index.js +1 -1
- package/dist/esm/apigroups/ResourcesAPIGroup_2_0.js +4 -0
- package/dist/esm/shared/charts/descriptors/analytics/trackGoal.js +3 -1
- package/dist/esm/shared/charts/descriptors/index.js +6 -1
- package/dist/esm/shared/charts/descriptors/knowledgeSearch/searchExtractOutput.js +4 -2
- package/dist/esm/shared/charts/descriptors/message/question/question.js +12 -1
- package/dist/esm/shared/charts/descriptors/service/GPTPrompt.js +15 -1
- package/dist/esm/shared/charts/descriptors/service/aiAgent/aiAgentJob.js +33 -174
- package/dist/esm/shared/charts/descriptors/service/aiAgent/aiAgentJobTool.js +2 -2
- package/dist/esm/shared/charts/descriptors/service/aiAgent/helpers/createToolDefinitions.js +172 -0
- package/dist/esm/shared/charts/descriptors/service/aiAgent/loadAiAgent.js +192 -0
- package/dist/esm/shared/charts/descriptors/service/handoverV2.js +1 -1
- package/dist/esm/shared/charts/descriptors/service/index.js +5 -0
- package/dist/esm/shared/charts/descriptors/service/llmPrompt/LLMPromptV2.js +909 -0
- package/dist/esm/shared/charts/descriptors/service/llmPrompt/llmPromptDefault.js +28 -0
- package/dist/esm/shared/charts/descriptors/service/llmPrompt/llmPromptMCPTool.js +193 -0
- package/dist/esm/shared/charts/descriptors/service/llmPrompt/llmPromptTool.js +136 -0
- package/dist/esm/shared/constants.js +1 -5
- package/dist/esm/shared/interfaces/debugEvents/IGoalCompletedEventPayload.js +2 -0
- package/dist/esm/shared/interfaces/debugEvents/TDebugEventType.js +1 -0
- package/dist/esm/shared/interfaces/generativeAI/IGenerativeAIModels.js +12 -1
- package/dist/esm/shared/interfaces/resources/knowledgeStore/IKnowledgeChunk.js +2 -1
- package/dist/esm/shared/interfaces/resources/settings/IGenerativeAISettings.js +4 -17
- package/dist/esm/shared/interfaces/restAPI/operations/generateOutput/v2.0/index.js +2 -0
- package/dist/esm/shared/interfaces/security/IPermission.js +2 -0
- package/dist/esm/shared/interfaces/security/IRole.js +3 -1
- package/dist/esm/shared/interfaces/security/index.js +1 -1
- package/package.json +1 -1
- package/types/index.d.ts +83 -42
|
@@ -0,0 +1,31 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
3
|
+
exports.LLM_PROMPT_DEFAULT = void 0;
|
|
4
|
+
/* Custom modules */
|
|
5
|
+
const createNodeDescriptor_1 = require("../../../createNodeDescriptor");
|
|
6
|
+
exports.LLM_PROMPT_DEFAULT = (0, createNodeDescriptor_1.createNodeDescriptor)({
|
|
7
|
+
type: "llmPromptDefault",
|
|
8
|
+
defaultLabel: "Default",
|
|
9
|
+
parentType: "llmPromptV2",
|
|
10
|
+
constraints: {
|
|
11
|
+
editable: false,
|
|
12
|
+
deletable: false,
|
|
13
|
+
collapsable: true,
|
|
14
|
+
creatable: false,
|
|
15
|
+
movable: false,
|
|
16
|
+
placement: {
|
|
17
|
+
predecessor: {
|
|
18
|
+
whitelist: []
|
|
19
|
+
}
|
|
20
|
+
}
|
|
21
|
+
},
|
|
22
|
+
fields: [],
|
|
23
|
+
sections: [],
|
|
24
|
+
form: [],
|
|
25
|
+
appearance: {
|
|
26
|
+
color: "white",
|
|
27
|
+
textColor: "#252525",
|
|
28
|
+
variant: "mini",
|
|
29
|
+
},
|
|
30
|
+
});
|
|
31
|
+
//# sourceMappingURL=llmPromptDefault.js.map
|
|
@@ -0,0 +1,196 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
3
|
+
exports.LLM_PROMPT_MCP_TOOL = void 0;
|
|
4
|
+
/* Custom modules */
|
|
5
|
+
const createNodeDescriptor_1 = require("../../../createNodeDescriptor");
|
|
6
|
+
exports.LLM_PROMPT_MCP_TOOL = (0, createNodeDescriptor_1.createNodeDescriptor)({
|
|
7
|
+
type: "llmPromptMCPTool",
|
|
8
|
+
defaultLabel: "MCP Tool",
|
|
9
|
+
summary: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__SUMMARY",
|
|
10
|
+
parentType: "llmPromptV2",
|
|
11
|
+
constraints: {
|
|
12
|
+
editable: true,
|
|
13
|
+
deletable: true,
|
|
14
|
+
collapsable: true,
|
|
15
|
+
creatable: true,
|
|
16
|
+
movable: true,
|
|
17
|
+
placement: {
|
|
18
|
+
predecessor: {
|
|
19
|
+
whitelist: [],
|
|
20
|
+
},
|
|
21
|
+
},
|
|
22
|
+
childFlowCreatable: false,
|
|
23
|
+
},
|
|
24
|
+
preview: {
|
|
25
|
+
type: "text",
|
|
26
|
+
key: "name",
|
|
27
|
+
},
|
|
28
|
+
fields: [
|
|
29
|
+
{
|
|
30
|
+
type: "description",
|
|
31
|
+
key: "cachingWarning",
|
|
32
|
+
label: " ",
|
|
33
|
+
params: {
|
|
34
|
+
text: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__CACHING_WARNING__TEXT",
|
|
35
|
+
},
|
|
36
|
+
condition: {
|
|
37
|
+
key: "cacheTools",
|
|
38
|
+
value: false,
|
|
39
|
+
}
|
|
40
|
+
},
|
|
41
|
+
{
|
|
42
|
+
key: "name",
|
|
43
|
+
label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__NAME__LABEL",
|
|
44
|
+
description: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__NAME__DESCRIPTION",
|
|
45
|
+
type: "cognigyText",
|
|
46
|
+
params: {
|
|
47
|
+
required: true,
|
|
48
|
+
rows: 1,
|
|
49
|
+
multiline: false,
|
|
50
|
+
maxLength: 64,
|
|
51
|
+
},
|
|
52
|
+
},
|
|
53
|
+
{
|
|
54
|
+
type: "description",
|
|
55
|
+
key: "mcpWarning",
|
|
56
|
+
label: " ",
|
|
57
|
+
params: {
|
|
58
|
+
text: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__MCP_WARNING__TEXT",
|
|
59
|
+
},
|
|
60
|
+
},
|
|
61
|
+
{
|
|
62
|
+
key: "mcpServerUrl",
|
|
63
|
+
label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__MCP_SERVER_URL__LABEL",
|
|
64
|
+
description: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__MCP_SERVER_URL__DESCRIPTION",
|
|
65
|
+
type: "cognigyText",
|
|
66
|
+
params: {
|
|
67
|
+
required: true,
|
|
68
|
+
rows: 5,
|
|
69
|
+
multiline: true,
|
|
70
|
+
},
|
|
71
|
+
},
|
|
72
|
+
{
|
|
73
|
+
key: "timeout",
|
|
74
|
+
label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__TIMEOUT__LABEL",
|
|
75
|
+
description: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__TIMEOUT__DESCRIPTION",
|
|
76
|
+
type: "slider",
|
|
77
|
+
defaultValue: 12,
|
|
78
|
+
params: {
|
|
79
|
+
min: 1,
|
|
80
|
+
max: 30,
|
|
81
|
+
step: 1,
|
|
82
|
+
},
|
|
83
|
+
},
|
|
84
|
+
{
|
|
85
|
+
key: "debugMessageFetchedTools",
|
|
86
|
+
type: "toggle",
|
|
87
|
+
label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__DEBUG_TOOLS__LABEL",
|
|
88
|
+
description: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__DEBUG_TOOLS__DESCRIPTION",
|
|
89
|
+
defaultValue: true,
|
|
90
|
+
},
|
|
91
|
+
{
|
|
92
|
+
key: "debugMessageParameters",
|
|
93
|
+
type: "toggle",
|
|
94
|
+
label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__DEBUG_PARAMETERS__LABEL",
|
|
95
|
+
description: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__DEBUG_PARAMETERS__DESCRIPTION",
|
|
96
|
+
defaultValue: false,
|
|
97
|
+
condition: {
|
|
98
|
+
key: "debugMessageFetchedTools",
|
|
99
|
+
value: true,
|
|
100
|
+
},
|
|
101
|
+
},
|
|
102
|
+
{
|
|
103
|
+
key: "debugMessage",
|
|
104
|
+
type: "toggle",
|
|
105
|
+
label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__DEBUG_MESSAGE__LABEL",
|
|
106
|
+
description: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__DEBUG_MESSAGE__DESCRIPTION",
|
|
107
|
+
defaultValue: true,
|
|
108
|
+
},
|
|
109
|
+
{
|
|
110
|
+
key: "cacheTools",
|
|
111
|
+
type: "toggle",
|
|
112
|
+
label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__CACHE_TOOLS__LABEL",
|
|
113
|
+
description: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__CACHE_TOOLS__DESCRIPTION",
|
|
114
|
+
defaultValue: true,
|
|
115
|
+
},
|
|
116
|
+
{
|
|
117
|
+
key: "condition",
|
|
118
|
+
label: "UI__NODE_EDITOR__SERVICE__AI_AGENT_TOOL__FIELDS__CONDITION__LABEL",
|
|
119
|
+
description: "UI__NODE_EDITOR__SERVICE__AI_AGENT_TOOL__FIELDS__CONDITION__DESCRIPTION",
|
|
120
|
+
type: "cognigyText",
|
|
121
|
+
defaultValue: "",
|
|
122
|
+
},
|
|
123
|
+
{
|
|
124
|
+
key: "toolFilter",
|
|
125
|
+
type: "select",
|
|
126
|
+
label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__TOOL_FILTER__LABEL",
|
|
127
|
+
description: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__TOOL_FILTER__DESCRIPTION",
|
|
128
|
+
defaultValue: "none",
|
|
129
|
+
params: {
|
|
130
|
+
options: [
|
|
131
|
+
{
|
|
132
|
+
label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__TOOL_FILTER__OPTIONS__NONE__LABEL",
|
|
133
|
+
value: "none",
|
|
134
|
+
},
|
|
135
|
+
{
|
|
136
|
+
label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__TOOL_FILTER__OPTIONS__WHITELIST__LABEL",
|
|
137
|
+
value: "whitelist",
|
|
138
|
+
},
|
|
139
|
+
{
|
|
140
|
+
label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__TOOL_FILTER__OPTIONS__BLACKLIST__LABEL",
|
|
141
|
+
value: "blacklist",
|
|
142
|
+
},
|
|
143
|
+
],
|
|
144
|
+
},
|
|
145
|
+
},
|
|
146
|
+
{
|
|
147
|
+
key: "whitelist",
|
|
148
|
+
type: "cognigyTextArray",
|
|
149
|
+
label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__WHITELIST__LABEL",
|
|
150
|
+
description: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__WHITELIST__DESCRIPTION",
|
|
151
|
+
condition: {
|
|
152
|
+
key: "toolFilter",
|
|
153
|
+
value: "whitelist",
|
|
154
|
+
},
|
|
155
|
+
},
|
|
156
|
+
{
|
|
157
|
+
key: "blacklist",
|
|
158
|
+
type: "cognigyTextArray",
|
|
159
|
+
label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__BLACKLIST__LABEL",
|
|
160
|
+
description: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_MCP_TOOL__FIELDS__BLACKLIST__DESCRIPTION",
|
|
161
|
+
condition: {
|
|
162
|
+
key: "toolFilter",
|
|
163
|
+
value: "blacklist",
|
|
164
|
+
},
|
|
165
|
+
},
|
|
166
|
+
],
|
|
167
|
+
sections: [
|
|
168
|
+
{
|
|
169
|
+
key: "debugging",
|
|
170
|
+
label: "UI__NODE_EDITOR__SERVICE__AI_AGENT_JOB__SECTIONS__DEBUG_SETTINGS__LABEL",
|
|
171
|
+
defaultCollapsed: true,
|
|
172
|
+
fields: ["debugMessageFetchedTools", "debugMessageParameters", "debugMessage"],
|
|
173
|
+
},
|
|
174
|
+
{
|
|
175
|
+
key: "advanced",
|
|
176
|
+
label: "UI__NODE_EDITOR__SERVICE__AI_AGENT_JOB__SECTIONS__ADVANCED__LABEL",
|
|
177
|
+
defaultCollapsed: true,
|
|
178
|
+
fields: ["cacheTools", "condition", "toolFilter", "whitelist", "blacklist"],
|
|
179
|
+
},
|
|
180
|
+
],
|
|
181
|
+
form: [
|
|
182
|
+
{ type: "field", key: "cachingWarning" },
|
|
183
|
+
{ type: "field", key: "name" },
|
|
184
|
+
{ type: "field", key: "mcpWarning" },
|
|
185
|
+
{ type: "field", key: "mcpServerUrl" },
|
|
186
|
+
{ type: "field", key: "timeout" },
|
|
187
|
+
{ type: "section", key: "debugging" },
|
|
188
|
+
{ type: "section", key: "advanced" },
|
|
189
|
+
],
|
|
190
|
+
appearance: {
|
|
191
|
+
color: "white",
|
|
192
|
+
textColor: "#252525",
|
|
193
|
+
variant: "mini",
|
|
194
|
+
},
|
|
195
|
+
});
|
|
196
|
+
//# sourceMappingURL=llmPromptMCPTool.js.map
|
|
@@ -0,0 +1,139 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
3
|
+
exports.LLM_PROMPT_TOOL = void 0;
|
|
4
|
+
/* Custom modules */
|
|
5
|
+
const createNodeDescriptor_1 = require("../../../createNodeDescriptor");
|
|
6
|
+
exports.LLM_PROMPT_TOOL = (0, createNodeDescriptor_1.createNodeDescriptor)({
|
|
7
|
+
type: "llmPromptTool",
|
|
8
|
+
defaultLabel: "Tool",
|
|
9
|
+
parentType: "llmPromptV2",
|
|
10
|
+
constraints: {
|
|
11
|
+
editable: true,
|
|
12
|
+
deletable: true,
|
|
13
|
+
collapsable: true,
|
|
14
|
+
creatable: true,
|
|
15
|
+
movable: true,
|
|
16
|
+
placement: {
|
|
17
|
+
predecessor: {
|
|
18
|
+
whitelist: []
|
|
19
|
+
}
|
|
20
|
+
},
|
|
21
|
+
childFlowCreatable: false
|
|
22
|
+
},
|
|
23
|
+
preview: {
|
|
24
|
+
type: "text",
|
|
25
|
+
key: "toolId"
|
|
26
|
+
},
|
|
27
|
+
fields: [
|
|
28
|
+
{
|
|
29
|
+
key: "toolId",
|
|
30
|
+
label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_TOOL__FIELDS__TOOL_ID__LABEL",
|
|
31
|
+
description: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_TOOL__FIELDS__TOOL_ID__DESCRIPTION",
|
|
32
|
+
type: "cognigyLLMText",
|
|
33
|
+
defaultValue: "unlock_account",
|
|
34
|
+
params: {
|
|
35
|
+
required: true,
|
|
36
|
+
rows: 1,
|
|
37
|
+
multiline: false,
|
|
38
|
+
maxLength: 64,
|
|
39
|
+
}
|
|
40
|
+
},
|
|
41
|
+
{
|
|
42
|
+
key: "description",
|
|
43
|
+
label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_TOOL__FIELDS__DESCRIPTION__LABEL",
|
|
44
|
+
description: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_TOOL__FIELDS__DESCRIPTION__DESCRIPTION",
|
|
45
|
+
type: "cognigyLLMText",
|
|
46
|
+
defaultValue: "This tool unlocks a locked user account.",
|
|
47
|
+
params: {
|
|
48
|
+
required: true,
|
|
49
|
+
rows: 5,
|
|
50
|
+
multiline: true
|
|
51
|
+
}
|
|
52
|
+
},
|
|
53
|
+
{
|
|
54
|
+
key: "useParameters",
|
|
55
|
+
label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_TOOL__FIELDS__USE_PARAMETERS__LABEL",
|
|
56
|
+
description: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_TOOL__FIELDS__USE_PARAMETERS__DESCRIPTION",
|
|
57
|
+
type: "toggle",
|
|
58
|
+
defaultValue: false
|
|
59
|
+
},
|
|
60
|
+
{
|
|
61
|
+
key: "parameters",
|
|
62
|
+
label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_TOOL__FIELDS__PARAMETERS__LABEL",
|
|
63
|
+
description: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_TOOL__FIELDS__PARAMETERS__DESCRIPTION",
|
|
64
|
+
type: "toolParameters",
|
|
65
|
+
defaultValue: `{
|
|
66
|
+
"type": "object",
|
|
67
|
+
"properties": {
|
|
68
|
+
"email": {
|
|
69
|
+
"type": "string",
|
|
70
|
+
"description": "User's login email for their account."
|
|
71
|
+
}
|
|
72
|
+
},
|
|
73
|
+
"required": ["email"],
|
|
74
|
+
"additionalProperties": false
|
|
75
|
+
}`,
|
|
76
|
+
params: {
|
|
77
|
+
required: false,
|
|
78
|
+
},
|
|
79
|
+
},
|
|
80
|
+
{
|
|
81
|
+
key: "debugMessage",
|
|
82
|
+
type: "toggle",
|
|
83
|
+
label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_TOOL__FIELDS__DEBUG_MESSAGE__LABEL",
|
|
84
|
+
description: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_TOOL__FIELDS__DEBUG_MESSAGE__DESCRIPTION",
|
|
85
|
+
defaultValue: true,
|
|
86
|
+
},
|
|
87
|
+
{
|
|
88
|
+
key: "condition",
|
|
89
|
+
label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_TOOL__FIELDS__CONDITION__LABEL",
|
|
90
|
+
description: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_TOOL__FIELDS__CONDITION__DESCRIPTION",
|
|
91
|
+
type: "cognigyText",
|
|
92
|
+
defaultValue: "",
|
|
93
|
+
},
|
|
94
|
+
],
|
|
95
|
+
sections: [
|
|
96
|
+
{
|
|
97
|
+
key: "parameters",
|
|
98
|
+
label: "UI__NODE_EDITOR__SERVICE__LLM_PROMPT_TOOL__SECTIONS__PARAMETERS__LABEL",
|
|
99
|
+
defaultCollapsed: true,
|
|
100
|
+
fields: [
|
|
101
|
+
"parameters",
|
|
102
|
+
],
|
|
103
|
+
condition: {
|
|
104
|
+
key: "useParameters",
|
|
105
|
+
value: true
|
|
106
|
+
},
|
|
107
|
+
},
|
|
108
|
+
{
|
|
109
|
+
key: "debugging",
|
|
110
|
+
label: "UI__NODE_EDITOR__SERVICE__AI_AGENT_JOB__SECTIONS__DEBUG_SETTINGS__LABEL",
|
|
111
|
+
defaultCollapsed: true,
|
|
112
|
+
fields: [
|
|
113
|
+
"debugMessage",
|
|
114
|
+
],
|
|
115
|
+
},
|
|
116
|
+
{
|
|
117
|
+
key: "advanced",
|
|
118
|
+
label: "UI__NODE_EDITOR__SERVICE__AI_AGENT_JOB__SECTIONS__ADVANCED__LABEL",
|
|
119
|
+
defaultCollapsed: true,
|
|
120
|
+
fields: [
|
|
121
|
+
"condition",
|
|
122
|
+
],
|
|
123
|
+
},
|
|
124
|
+
],
|
|
125
|
+
form: [
|
|
126
|
+
{ type: "field", key: "toolId" },
|
|
127
|
+
{ type: "field", key: "description" },
|
|
128
|
+
{ type: "field", key: "useParameters" },
|
|
129
|
+
{ type: "section", key: "parameters" },
|
|
130
|
+
{ type: "section", key: "debugging" },
|
|
131
|
+
{ type: "section", key: "advanced" },
|
|
132
|
+
],
|
|
133
|
+
appearance: {
|
|
134
|
+
color: "white",
|
|
135
|
+
textColor: "#252525",
|
|
136
|
+
variant: "mini",
|
|
137
|
+
},
|
|
138
|
+
});
|
|
139
|
+
//# sourceMappingURL=llmPromptTool.js.map
|
|
@@ -54,17 +54,13 @@ exports.IP_NOT_AVAILABLE = "not available";
|
|
|
54
54
|
exports.SECONDS_PER_MINUTE = 60;
|
|
55
55
|
// Platform provided LLM
|
|
56
56
|
exports.PLATFORM_PROVIDED_LLM_USE_CASES = [
|
|
57
|
-
"intentSentenceGeneration",
|
|
58
|
-
"flowGeneration",
|
|
59
|
-
"lexiconGeneration",
|
|
60
|
-
"generateNodeOutput",
|
|
61
57
|
"designTimeGeneration",
|
|
62
58
|
];
|
|
63
59
|
exports.DEPRECATED_GENAI_USE_CASES = [
|
|
64
60
|
"intentSentenceGeneration",
|
|
65
61
|
"flowGeneration",
|
|
66
|
-
"lexiconGeneration",
|
|
67
62
|
"generateNodeOutput",
|
|
63
|
+
"lexiconGeneration",
|
|
68
64
|
];
|
|
69
65
|
exports.PLATFORM_PROVIDED_LLM_ID = "aaaaaaaa-aaaa-aaaa-aaaa-000000000000";
|
|
70
66
|
exports.PLATFORM_PROVIDED_LLM_CONNECTION_ID = "aaaaaaaa-aaaa-aaaa-aaaa-000000000001";
|
|
@@ -19,6 +19,9 @@ exports.generativeAIModels = [
|
|
|
19
19
|
"gpt-4",
|
|
20
20
|
"gpt-4o",
|
|
21
21
|
"gpt-4o-mini",
|
|
22
|
+
"gpt-4.1",
|
|
23
|
+
"gpt-4.1-mini",
|
|
24
|
+
"gpt-4.1-nano",
|
|
22
25
|
"luminous-extended-control",
|
|
23
26
|
"claude-v1-100k",
|
|
24
27
|
"claude-instant-v1",
|
|
@@ -26,6 +29,11 @@ exports.generativeAIModels = [
|
|
|
26
29
|
"claude-3-haiku-20240307",
|
|
27
30
|
"claude-3-sonnet-20240229",
|
|
28
31
|
"claude-3-5-sonnet-20241022",
|
|
32
|
+
"claude-3-7-sonnet-20250219",
|
|
33
|
+
"claude-3-5-sonnet-latest",
|
|
34
|
+
"claude-3-7-sonnet-latest",
|
|
35
|
+
"claude-opus-4-0",
|
|
36
|
+
"claude-sonnet-4-0",
|
|
29
37
|
"text-bison@001",
|
|
30
38
|
"custom-model",
|
|
31
39
|
"gemini-1.0-pro",
|
|
@@ -37,11 +45,14 @@ exports.generativeAIModels = [
|
|
|
37
45
|
"amazon.nova-pro-v1:0",
|
|
38
46
|
"amazon.nova-micro-v1:0",
|
|
39
47
|
"anthropic.claude-3-5-sonnet-20240620-v1:0",
|
|
40
|
-
"claude-3-7-sonnet-20250219",
|
|
41
48
|
"mistral-large-2411",
|
|
42
49
|
"mistral-small-2503",
|
|
43
50
|
"pixtral-large-2411",
|
|
44
51
|
"pixtral-12b-2409",
|
|
52
|
+
"mistral-large-latest",
|
|
53
|
+
"pixtral-large-latest",
|
|
54
|
+
"mistral-medium-latest",
|
|
55
|
+
"mistral-small-latest",
|
|
45
56
|
...exports.embeddingModels,
|
|
46
57
|
];
|
|
47
58
|
exports.generativeAIProviders = [
|
|
@@ -1,6 +1,9 @@
|
|
|
1
1
|
"use strict";
|
|
2
2
|
Object.defineProperty(exports, "__esModule", { value: true });
|
|
3
|
-
exports.generativeAISettingsDataSchema = exports.generativeAIUseCaseDataSchema = void 0;
|
|
3
|
+
exports.generativeAISettingsDataSchema = exports.generativeAIUseCaseDataSchema = exports.configurableGenerativeAIUseCases = void 0;
|
|
4
|
+
const IGenerativeAIModels_1 = require("../../generativeAI/IGenerativeAIModels");
|
|
5
|
+
const constants_1 = require("../../../constants");
|
|
6
|
+
exports.configurableGenerativeAIUseCases = IGenerativeAIModels_1.generativeAIUseCases.filter((useCase) => !constants_1.DEPRECATED_GENAI_USE_CASES.includes(useCase));
|
|
4
7
|
exports.generativeAIUseCaseDataSchema = {
|
|
5
8
|
title: "generativeAIMetaAzureOpenAIDataSchema",
|
|
6
9
|
type: "object",
|
|
@@ -34,22 +37,10 @@ exports.generativeAISettingsDataSchema = {
|
|
|
34
37
|
type: "object",
|
|
35
38
|
properties: Object.assign({}, exports.generativeAIUseCaseDataSchema.properties)
|
|
36
39
|
},
|
|
37
|
-
intentSentenceGeneration: {
|
|
38
|
-
type: "object",
|
|
39
|
-
properties: Object.assign({}, exports.generativeAIUseCaseDataSchema.properties)
|
|
40
|
-
},
|
|
41
40
|
aiEnhancedOutputs: {
|
|
42
41
|
type: "object",
|
|
43
42
|
properties: Object.assign({}, exports.generativeAIUseCaseDataSchema.properties)
|
|
44
43
|
},
|
|
45
|
-
lexiconGeneration: {
|
|
46
|
-
type: "object",
|
|
47
|
-
properties: Object.assign({}, exports.generativeAIUseCaseDataSchema.properties)
|
|
48
|
-
},
|
|
49
|
-
flowGeneration: {
|
|
50
|
-
type: "object",
|
|
51
|
-
properties: Object.assign({}, exports.generativeAIUseCaseDataSchema.properties)
|
|
52
|
-
},
|
|
53
44
|
gptConversation: {
|
|
54
45
|
type: "object",
|
|
55
46
|
properties: Object.assign({}, exports.generativeAIUseCaseDataSchema.properties)
|
|
@@ -58,10 +49,6 @@ exports.generativeAISettingsDataSchema = {
|
|
|
58
49
|
type: "object",
|
|
59
50
|
properties: Object.assign({}, exports.generativeAIUseCaseDataSchema.properties)
|
|
60
51
|
},
|
|
61
|
-
generateNodeOutput: {
|
|
62
|
-
type: "object",
|
|
63
|
-
properties: Object.assign({}, exports.generativeAIUseCaseDataSchema.properties)
|
|
64
|
-
},
|
|
65
52
|
knowledgeSearch: {
|
|
66
53
|
type: "object",
|
|
67
54
|
properties: Object.assign({}, exports.generativeAIUseCaseDataSchema.properties)
|
|
@@ -81,7 +68,7 @@ exports.generativeAISettingsDataSchema = {
|
|
|
81
68
|
aiAgent: {
|
|
82
69
|
type: "object",
|
|
83
70
|
properties: Object.assign({}, exports.generativeAIUseCaseDataSchema.properties)
|
|
84
|
-
}
|
|
71
|
+
}
|
|
85
72
|
}
|
|
86
73
|
}
|
|
87
74
|
}
|
|
@@ -74,6 +74,7 @@ exports.organisationWidePermissions = [
|
|
|
74
74
|
* - functions
|
|
75
75
|
* - yesNoIntents
|
|
76
76
|
* - dataPrivacySettings
|
|
77
|
+
* - simulator
|
|
77
78
|
*/
|
|
78
79
|
exports.projectWidePermissions = [
|
|
79
80
|
"agentAssistConfigs",
|
|
@@ -113,6 +114,7 @@ exports.projectWidePermissions = [
|
|
|
113
114
|
"tokens",
|
|
114
115
|
"yesNoIntents",
|
|
115
116
|
"dataPrivacySettings",
|
|
117
|
+
"simulator"
|
|
116
118
|
];
|
|
117
119
|
exports.availablePermissions = [
|
|
118
120
|
...exports.organisationWidePermissions,
|
|
@@ -88,6 +88,7 @@ exports.organisationWideRoles = [
|
|
|
88
88
|
* - tokenEditor
|
|
89
89
|
* - function_admin
|
|
90
90
|
* - function_editor
|
|
91
|
+
* - simulator_admin
|
|
91
92
|
*/
|
|
92
93
|
exports.projectWideRoles = [
|
|
93
94
|
"agentAssistConfigAdmin",
|
|
@@ -129,7 +130,8 @@ exports.projectWideRoles = [
|
|
|
129
130
|
"tokenEditor",
|
|
130
131
|
"data_privacy_admin",
|
|
131
132
|
"data_privacy_editor",
|
|
132
|
-
"data_privacy_viewer"
|
|
133
|
+
"data_privacy_viewer",
|
|
134
|
+
"simulator_admin"
|
|
133
135
|
];
|
|
134
136
|
exports.availableRoles = [
|
|
135
137
|
...exports.organisationWideRoles,
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
"use strict";
|
|
2
2
|
Object.defineProperty(exports, "__esModule", { value: true });
|
|
3
3
|
exports.identityProviderSchema = exports.oidcIdentityProviderDataSchema = exports.samlIdentityProviderDataSchema = exports.operations = exports.availableRoles = exports.projectWideRoles = exports.organisationWideRoles = exports.availablePermissions = exports.COMPUTED_ACL_HASH_VERSION = void 0;
|
|
4
|
-
exports.COMPUTED_ACL_HASH_VERSION = "
|
|
4
|
+
exports.COMPUTED_ACL_HASH_VERSION = "v20";
|
|
5
5
|
var IPermission_1 = require("./IPermission");
|
|
6
6
|
Object.defineProperty(exports, "availablePermissions", { enumerable: true, get: function () { return IPermission_1.availablePermissions; } });
|
|
7
7
|
var IRole_1 = require("./IRole");
|
|
@@ -615,6 +615,10 @@ export const ResourcesAPIGroup_2_0 = (instance) => {
|
|
|
615
615
|
generateNluScores: (_a) => {
|
|
616
616
|
var { projectId } = _a, args = __rest(_a, ["projectId"]);
|
|
617
617
|
return GenericAPIFn(`/new/v2.0/projects/${projectId}/nlu/scores`, "POST", self)(args);
|
|
618
|
+
},
|
|
619
|
+
generateDesignTimeLLMOutput: (_a) => {
|
|
620
|
+
var { projectId } = _a, args = __rest(_a, ["projectId"]);
|
|
621
|
+
return GenericAPIFn(`/new/v2.0/projects/${projectId}/generate-output/design-time-llm`, "POST", self)(args);
|
|
618
622
|
}
|
|
619
623
|
};
|
|
620
624
|
};
|
|
@@ -45,7 +45,7 @@ export const TRACK_GOAL = createNodeDescriptor({
|
|
|
45
45
|
return;
|
|
46
46
|
}
|
|
47
47
|
const sessionState = api.getSessionStateCopy();
|
|
48
|
-
const { selectedSteps, version, name, goalId, } = config.goal;
|
|
48
|
+
const { selectedSteps, referenceId, version, name, goalId, } = config.goal;
|
|
49
49
|
const activeCycleIds = ((_a = sessionState.analytics) === null || _a === void 0 ? void 0 : _a.goalCycleIds) || {};
|
|
50
50
|
let cycleId = activeCycleIds[goalId];
|
|
51
51
|
const hasStartStep = selectedSteps.some(step => step.type === "start");
|
|
@@ -73,10 +73,12 @@ export const TRACK_GOAL = createNodeDescriptor({
|
|
|
73
73
|
organisationId,
|
|
74
74
|
sessionId,
|
|
75
75
|
version,
|
|
76
|
+
referenceId,
|
|
76
77
|
timestamp: new Date(),
|
|
77
78
|
goalCycleId: cycleId,
|
|
78
79
|
goalId,
|
|
79
80
|
stepId: step.stepId,
|
|
81
|
+
stepType: step.type,
|
|
80
82
|
endpointUrlToken,
|
|
81
83
|
endpointName,
|
|
82
84
|
endpointType,
|
|
@@ -11,7 +11,7 @@ import { REGEX_SLOT_FILLER, EXECUTE_COGNIGY_NLU, ADD_LEXICON_KEYPHRASE, FUZZY_SE
|
|
|
11
11
|
import { KNOWLEDGE_SEARCH, KNOWLEDGE_SEARCH_V2, SEARCH_EXTRACT_OUTPUT } from "./knowledgeSearch";
|
|
12
12
|
import { CONTINUOUS_ASR, DTMF, HANG_UP, PLAY, TRANSFER_VOICE, SESSION_SPEECH_PARAMETERS, USER_INPUT_TIMEOUT, SEND_METADATA, BARGE_IN, MUTE_SPEECH_INPUT, } from "./voice";
|
|
13
13
|
import { ACTIVATE_PROFILE, COMPLETE_GOAL, DEACTIVATE_PROFILE, DELETE_PROFILE, MERGE_PROFILE, UPDATE_PROFILE, ADD_MEMORY, BLIND_MODE, OVERWRITE_ANALYTICS, SET_RATING, REQUEST_RATING, TRACK_GOAL, } from "./analytics";
|
|
14
|
-
import { HANDOVER, HANDOVER_V2, CHECK_AGENT_AVAILABILITY, HTTP_REQUEST, HTTP_CONNECTION_BASIC, HTTP_CONNECTION_APIKEYAUTHKEY, HTTP_CONNECTION_APIKEYXKEY, HTTP_CONNECTION_OAUTH2, JWT_SECRET_CONNECTION, TRIGGER_FUNCTION, ON_SCHEDULING_ERROR, ON_SCHEDULED, GPT_PROMPT, CLOSE_HANDOVER, HANDOVER_INACTIVITY_TIMER, GPT_CONVERSATION, GPT_CONVERSATION_SUMMARY, LLM_ENTITY_EXTRACT, AI_AGENT_JOB, AI_AGENT_JOB_DEFAULT, AI_AGENT_JOB_TOOL, AI_AGENT_JOB_MCP_TOOL, AI_AGENT_JOB_CALL_MCP_TOOL, AI_AGENT_TOOL_ANSWER, AI_AGENT_HANDOVER, LIVE_AGENT_CONNECTION, RINGCENTRAL_ENGAGE_CONNECTION, CHATWOOT_CONNECTION, EIGHT_BY_EIGHT_CONNECTION, GENESYS_CLOUD_CONNECTION, GENESYS_CLOUD_CONNECTION_OM, LLM_MODERATE, NICECXONEAAH_AUTHENTICATION_CONNECTION, AIOPS_CENTER_WEBHOOKS_CONNECTION } from "./service";
|
|
14
|
+
import { HANDOVER, HANDOVER_V2, CHECK_AGENT_AVAILABILITY, HTTP_REQUEST, HTTP_CONNECTION_BASIC, HTTP_CONNECTION_APIKEYAUTHKEY, HTTP_CONNECTION_APIKEYXKEY, HTTP_CONNECTION_OAUTH2, JWT_SECRET_CONNECTION, TRIGGER_FUNCTION, ON_SCHEDULING_ERROR, ON_SCHEDULED, GPT_PROMPT, LLM_PROMPT_V2, LLM_PROMPT_DEFAULT, LLM_PROMPT_MCP_TOOL, LLM_PROMPT_TOOL, CLOSE_HANDOVER, HANDOVER_INACTIVITY_TIMER, GPT_CONVERSATION, GPT_CONVERSATION_SUMMARY, LLM_ENTITY_EXTRACT, AI_AGENT_JOB, AI_AGENT_JOB_DEFAULT, AI_AGENT_JOB_TOOL, AI_AGENT_JOB_MCP_TOOL, AI_AGENT_JOB_CALL_MCP_TOOL, AI_AGENT_TOOL_ANSWER, AI_AGENT_HANDOVER, LIVE_AGENT_CONNECTION, RINGCENTRAL_ENGAGE_CONNECTION, CHATWOOT_CONNECTION, EIGHT_BY_EIGHT_CONNECTION, GENESYS_CLOUD_CONNECTION, GENESYS_CLOUD_CONNECTION_OM, LLM_MODERATE, NICECXONEAAH_AUTHENTICATION_CONNECTION, LOAD_AI_AGENT, AIOPS_CENTER_WEBHOOKS_CONNECTION } from "./service";
|
|
15
15
|
import { INIT_APP_SESSION, GET_APP_SESSION_PIN, SET_HTML_APP_STATE, SET_ADAPTIVE_CARD_APP_STATE, } from "./apps";
|
|
16
16
|
import { SET_IFRAME_TILE, SET_HTML_TILE, SEND_TILE_DATA, SET_SECURE_FORMS_TILE, SET_ADAPTIVE_CARD_TILE, SET_AGENT_ASSIST_GRID, NEXT_ACTION_ASSIST, SENTIMENT_ASSIST, TRANSCRIPT_ASSIST, IDENTITY_ASSIST, KNOWLEDGE_ASSIST, } from "./agentAssist";
|
|
17
17
|
import { ASSIST_INFO } from "./liveAgent";
|
|
@@ -88,6 +88,10 @@ const nodes = [
|
|
|
88
88
|
CHECK_AGENT_AVAILABILITY,
|
|
89
89
|
HTTP_REQUEST,
|
|
90
90
|
GPT_PROMPT,
|
|
91
|
+
LLM_PROMPT_V2,
|
|
92
|
+
LLM_PROMPT_DEFAULT,
|
|
93
|
+
LLM_PROMPT_MCP_TOOL,
|
|
94
|
+
LLM_PROMPT_TOOL,
|
|
91
95
|
GPT_CONVERSATION,
|
|
92
96
|
GPT_CONVERSATION_SUMMARY,
|
|
93
97
|
LLM_ENTITY_EXTRACT,
|
|
@@ -143,6 +147,7 @@ if (process.env.DISABLE_FEATURE_TRANSCRIPT_MANAGER !== "true") {
|
|
|
143
147
|
nodes.push(AI_AGENT_JOB_CALL_MCP_TOOL);
|
|
144
148
|
nodes.push(AI_AGENT_TOOL_ANSWER);
|
|
145
149
|
nodes.push(AI_AGENT_HANDOVER);
|
|
150
|
+
nodes.push(LOAD_AI_AGENT);
|
|
146
151
|
}
|
|
147
152
|
if (process.env.FEATURE_USE_COGNIGY_LIVE_AGENT === "true") {
|
|
148
153
|
nodes.push(ASSIST_INFO);
|
|
@@ -582,10 +582,12 @@ export const SEARCH_EXTRACT_OUTPUT = createNodeDescriptor({
|
|
|
582
582
|
// as it doesn't make sense to check for follow ups in the first execution
|
|
583
583
|
if (input.execution > 1) {
|
|
584
584
|
// always remember the last thing the user said (needed later)
|
|
585
|
-
lastRoundTrip = (_b = cognigy
|
|
585
|
+
lastRoundTrip = (_b = cognigy
|
|
586
|
+
.lastConversationEntries) === null || _b === void 0 ? void 0 : _b.slice(1, followUpDetectionSteps + 1).reverse().map(entry => "- " + (entry.source === "user" ? "USER: " : "BOT: ") + entry.text).join("\n");
|
|
586
587
|
// if follow up detection is set to 2 or more, we use the conversation transcript
|
|
587
588
|
// as reference. Start at the second entry, because the first one is the current
|
|
588
|
-
const recentConversation = (_c = cognigy
|
|
589
|
+
const recentConversation = (_c = cognigy
|
|
590
|
+
.lastConversationEntries) === null || _c === void 0 ? void 0 : _c.slice(1, followUpDetectionSteps + 1).reverse().map(entry => "- " + (entry.source === "user" ? "USER: " : "BOT: ") + entry.text).join("\n");
|
|
589
591
|
prompt = `Below is the transcript of a conversation:
|
|
590
592
|
${recentConversation}
|
|
591
593
|
USER: ${searchString}
|
|
@@ -307,7 +307,7 @@ export const QUESTION = createNodeDescriptor({
|
|
|
307
307
|
},
|
|
308
308
|
{
|
|
309
309
|
key: "repromptLLMPrompt",
|
|
310
|
-
type: "
|
|
310
|
+
type: "cognigyLLMText",
|
|
311
311
|
label: "UI__NODE_EDITOR__MESSAGE__QUESTION__QUESTION__FIELDS__REPROMPT_LLM_PROMPT__LABEL",
|
|
312
312
|
description: "UI__NODE_EDITOR__MESSAGE__QUESTION__QUESTION__FIELDS__REPROMPT_LLM_PROMPT__DESCRIPTION",
|
|
313
313
|
defaultValue: `You are a chatbot that helps a user.
|
|
@@ -740,6 +740,17 @@ DO NOT talk about other topics. Do not offer general assistance.`,
|
|
|
740
740
|
},
|
|
741
741
|
defaultValue: true,
|
|
742
742
|
},
|
|
743
|
+
{
|
|
744
|
+
key: "escalateAnswersRepeatHandoverMessage",
|
|
745
|
+
type: "toggle",
|
|
746
|
+
label: "UI__NODE_EDITOR__MESSAGE__QUESTION__QUESTION__FIELDS__ESCALATE_ANSWERS_REPEAT_HANDOVER_MESSAGE__LABEL",
|
|
747
|
+
description: "UI__NODE_EDITOR__MESSAGE__QUESTION__QUESTION__FIELDS__ESCALATE_ANSWERS_REPEAT_HANDOVER_MESSAGE__DESCRIPTION",
|
|
748
|
+
defaultValue: false,
|
|
749
|
+
condition: {
|
|
750
|
+
key: "escalateAnswersAction",
|
|
751
|
+
value: "handover"
|
|
752
|
+
}
|
|
753
|
+
},
|
|
743
754
|
{
|
|
744
755
|
key: "escalateAnswersHandoverCancelIntent",
|
|
745
756
|
type: "cognigyText",
|