@smythos/sre 1.5.53 → 1.5.54

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (223) hide show
  1. package/CHANGELOG +98 -98
  2. package/LICENSE +18 -18
  3. package/README.md +135 -135
  4. package/dist/bundle-analysis-lazy.html +4949 -0
  5. package/dist/bundle-analysis.html +4949 -0
  6. package/dist/index.js +3 -3
  7. package/dist/index.js.map +1 -1
  8. package/dist/types/subsystems/LLMManager/LLM.service/connectors/openai/OpenAIConnector.class.d.ts +1 -6
  9. package/dist/types/utils/package-manager.utils.d.ts +26 -0
  10. package/package.json +1 -1
  11. package/src/Components/APICall/APICall.class.ts +157 -157
  12. package/src/Components/APICall/AccessTokenManager.ts +166 -166
  13. package/src/Components/APICall/ArrayBufferResponse.helper.ts +58 -58
  14. package/src/Components/APICall/OAuth.helper.ts +447 -447
  15. package/src/Components/APICall/mimeTypeCategories.ts +46 -46
  16. package/src/Components/APICall/parseData.ts +167 -167
  17. package/src/Components/APICall/parseHeaders.ts +41 -41
  18. package/src/Components/APICall/parseProxy.ts +68 -68
  19. package/src/Components/APICall/parseUrl.ts +91 -91
  20. package/src/Components/APIEndpoint.class.ts +234 -234
  21. package/src/Components/APIOutput.class.ts +58 -58
  22. package/src/Components/AgentPlugin.class.ts +102 -102
  23. package/src/Components/Async.class.ts +155 -155
  24. package/src/Components/Await.class.ts +90 -90
  25. package/src/Components/Classifier.class.ts +158 -158
  26. package/src/Components/Component.class.ts +132 -132
  27. package/src/Components/ComponentHost.class.ts +38 -38
  28. package/src/Components/DataSourceCleaner.class.ts +92 -92
  29. package/src/Components/DataSourceIndexer.class.ts +181 -181
  30. package/src/Components/DataSourceLookup.class.ts +161 -161
  31. package/src/Components/ECMASandbox.class.ts +71 -71
  32. package/src/Components/FEncDec.class.ts +29 -29
  33. package/src/Components/FHash.class.ts +33 -33
  34. package/src/Components/FSign.class.ts +80 -80
  35. package/src/Components/FSleep.class.ts +25 -25
  36. package/src/Components/FTimestamp.class.ts +25 -25
  37. package/src/Components/FileStore.class.ts +78 -78
  38. package/src/Components/ForEach.class.ts +97 -97
  39. package/src/Components/GPTPlugin.class.ts +70 -70
  40. package/src/Components/GenAILLM.class.ts +586 -586
  41. package/src/Components/HuggingFace.class.ts +314 -314
  42. package/src/Components/Image/imageSettings.config.ts +70 -70
  43. package/src/Components/ImageGenerator.class.ts +502 -502
  44. package/src/Components/JSONFilter.class.ts +54 -54
  45. package/src/Components/LLMAssistant.class.ts +213 -213
  46. package/src/Components/LogicAND.class.ts +28 -28
  47. package/src/Components/LogicAtLeast.class.ts +85 -85
  48. package/src/Components/LogicAtMost.class.ts +86 -86
  49. package/src/Components/LogicOR.class.ts +29 -29
  50. package/src/Components/LogicXOR.class.ts +34 -34
  51. package/src/Components/MCPClient.class.ts +138 -138
  52. package/src/Components/MemoryDeleteKeyVal.class.ts +70 -70
  53. package/src/Components/MemoryReadKeyVal.class.ts +66 -66
  54. package/src/Components/MemoryWriteKeyVal.class.ts +62 -62
  55. package/src/Components/MemoryWriteObject.class.ts +97 -97
  56. package/src/Components/MultimodalLLM.class.ts +128 -128
  57. package/src/Components/OpenAPI.class.ts +72 -72
  58. package/src/Components/PromptGenerator.class.ts +122 -122
  59. package/src/Components/ScrapflyWebScrape.class.ts +159 -159
  60. package/src/Components/ServerlessCode.class.ts +123 -123
  61. package/src/Components/TavilyWebSearch.class.ts +98 -98
  62. package/src/Components/VisionLLM.class.ts +104 -104
  63. package/src/Components/ZapierAction.class.ts +127 -127
  64. package/src/Components/index.ts +97 -97
  65. package/src/Core/AgentProcess.helper.ts +240 -240
  66. package/src/Core/Connector.class.ts +123 -123
  67. package/src/Core/ConnectorsService.ts +197 -197
  68. package/src/Core/DummyConnector.ts +49 -49
  69. package/src/Core/HookService.ts +105 -105
  70. package/src/Core/SmythRuntime.class.ts +235 -235
  71. package/src/Core/SystemEvents.ts +16 -16
  72. package/src/Core/boot.ts +56 -56
  73. package/src/config.ts +15 -15
  74. package/src/constants.ts +126 -126
  75. package/src/data/hugging-face.params.json +579 -579
  76. package/src/helpers/AWSLambdaCode.helper.ts +590 -590
  77. package/src/helpers/BinaryInput.helper.ts +331 -331
  78. package/src/helpers/Conversation.helper.ts +1119 -1119
  79. package/src/helpers/ECMASandbox.helper.ts +54 -54
  80. package/src/helpers/JsonContent.helper.ts +97 -97
  81. package/src/helpers/LocalCache.helper.ts +97 -97
  82. package/src/helpers/Log.helper.ts +274 -274
  83. package/src/helpers/OpenApiParser.helper.ts +150 -150
  84. package/src/helpers/S3Cache.helper.ts +147 -147
  85. package/src/helpers/SmythURI.helper.ts +5 -5
  86. package/src/helpers/Sysconfig.helper.ts +77 -77
  87. package/src/helpers/TemplateString.helper.ts +243 -243
  88. package/src/helpers/TypeChecker.helper.ts +329 -329
  89. package/src/index.ts +3 -3
  90. package/src/index.ts.bak +3 -3
  91. package/src/subsystems/AgentManager/Agent.class.ts +1114 -1114
  92. package/src/subsystems/AgentManager/Agent.helper.ts +3 -3
  93. package/src/subsystems/AgentManager/AgentData.service/AgentDataConnector.ts +230 -230
  94. package/src/subsystems/AgentManager/AgentData.service/connectors/CLIAgentDataConnector.class.ts +66 -66
  95. package/src/subsystems/AgentManager/AgentData.service/connectors/LocalAgentDataConnector.class.ts +142 -142
  96. package/src/subsystems/AgentManager/AgentData.service/connectors/NullAgentData.class.ts +39 -39
  97. package/src/subsystems/AgentManager/AgentData.service/index.ts +18 -18
  98. package/src/subsystems/AgentManager/AgentLogger.class.ts +301 -297
  99. package/src/subsystems/AgentManager/AgentRequest.class.ts +51 -51
  100. package/src/subsystems/AgentManager/AgentRuntime.class.ts +559 -559
  101. package/src/subsystems/AgentManager/AgentSSE.class.ts +101 -101
  102. package/src/subsystems/AgentManager/AgentSettings.class.ts +52 -52
  103. package/src/subsystems/AgentManager/Component.service/ComponentConnector.ts +32 -32
  104. package/src/subsystems/AgentManager/Component.service/connectors/LocalComponentConnector.class.ts +60 -60
  105. package/src/subsystems/AgentManager/Component.service/index.ts +11 -11
  106. package/src/subsystems/AgentManager/EmbodimentSettings.class.ts +47 -47
  107. package/src/subsystems/AgentManager/ForkedAgent.class.ts +154 -154
  108. package/src/subsystems/AgentManager/OSResourceMonitor.ts +77 -77
  109. package/src/subsystems/ComputeManager/Code.service/CodeConnector.ts +98 -98
  110. package/src/subsystems/ComputeManager/Code.service/connectors/AWSLambdaCode.class.ts +172 -172
  111. package/src/subsystems/ComputeManager/Code.service/connectors/ECMASandbox.class.ts +131 -131
  112. package/src/subsystems/ComputeManager/Code.service/index.ts +13 -13
  113. package/src/subsystems/IO/CLI.service/CLIConnector.ts +47 -47
  114. package/src/subsystems/IO/CLI.service/index.ts +9 -9
  115. package/src/subsystems/IO/Log.service/LogConnector.ts +32 -32
  116. package/src/subsystems/IO/Log.service/connectors/ConsoleLog.class.ts +28 -28
  117. package/src/subsystems/IO/Log.service/index.ts +13 -13
  118. package/src/subsystems/IO/NKV.service/NKVConnector.ts +43 -43
  119. package/src/subsystems/IO/NKV.service/connectors/NKVLocalStorage.class.ts +234 -234
  120. package/src/subsystems/IO/NKV.service/connectors/NKVRAM.class.ts +204 -204
  121. package/src/subsystems/IO/NKV.service/connectors/NKVRedis.class.ts +182 -182
  122. package/src/subsystems/IO/NKV.service/index.ts +14 -14
  123. package/src/subsystems/IO/Router.service/RouterConnector.ts +21 -21
  124. package/src/subsystems/IO/Router.service/connectors/ExpressRouter.class.ts +48 -48
  125. package/src/subsystems/IO/Router.service/connectors/NullRouter.class.ts +40 -40
  126. package/src/subsystems/IO/Router.service/index.ts +11 -11
  127. package/src/subsystems/IO/Storage.service/SmythFS.class.ts +489 -489
  128. package/src/subsystems/IO/Storage.service/StorageConnector.ts +66 -66
  129. package/src/subsystems/IO/Storage.service/connectors/LocalStorage.class.ts +327 -327
  130. package/src/subsystems/IO/Storage.service/connectors/S3Storage.class.ts +482 -482
  131. package/src/subsystems/IO/Storage.service/index.ts +13 -13
  132. package/src/subsystems/IO/VectorDB.service/VectorDBConnector.ts +108 -108
  133. package/src/subsystems/IO/VectorDB.service/connectors/MilvusVectorDB.class.ts +454 -454
  134. package/src/subsystems/IO/VectorDB.service/connectors/PineconeVectorDB.class.ts +384 -384
  135. package/src/subsystems/IO/VectorDB.service/connectors/RAMVecrtorDB.class.ts +421 -421
  136. package/src/subsystems/IO/VectorDB.service/embed/BaseEmbedding.ts +107 -107
  137. package/src/subsystems/IO/VectorDB.service/embed/OpenAIEmbedding.ts +109 -109
  138. package/src/subsystems/IO/VectorDB.service/embed/index.ts +21 -21
  139. package/src/subsystems/IO/VectorDB.service/index.ts +14 -14
  140. package/src/subsystems/LLMManager/LLM.helper.ts +251 -251
  141. package/src/subsystems/LLMManager/LLM.inference.ts +339 -339
  142. package/src/subsystems/LLMManager/LLM.service/LLMConnector.ts +489 -489
  143. package/src/subsystems/LLMManager/LLM.service/LLMCredentials.helper.ts +171 -171
  144. package/src/subsystems/LLMManager/LLM.service/connectors/Anthropic.class.ts +659 -659
  145. package/src/subsystems/LLMManager/LLM.service/connectors/Bedrock.class.ts +400 -400
  146. package/src/subsystems/LLMManager/LLM.service/connectors/Echo.class.ts +77 -77
  147. package/src/subsystems/LLMManager/LLM.service/connectors/GoogleAI.class.ts +757 -757
  148. package/src/subsystems/LLMManager/LLM.service/connectors/Groq.class.ts +304 -304
  149. package/src/subsystems/LLMManager/LLM.service/connectors/Perplexity.class.ts +250 -250
  150. package/src/subsystems/LLMManager/LLM.service/connectors/VertexAI.class.ts +423 -423
  151. package/src/subsystems/LLMManager/LLM.service/connectors/openai/OpenAIConnector.class.ts +488 -488
  152. package/src/subsystems/LLMManager/LLM.service/connectors/openai/apiInterfaces/ChatCompletionsApiInterface.ts +524 -524
  153. package/src/subsystems/LLMManager/LLM.service/connectors/openai/apiInterfaces/OpenAIApiInterface.ts +100 -100
  154. package/src/subsystems/LLMManager/LLM.service/connectors/openai/apiInterfaces/OpenAIApiInterfaceFactory.ts +81 -81
  155. package/src/subsystems/LLMManager/LLM.service/connectors/openai/apiInterfaces/ResponsesApiInterface.ts +1145 -1145
  156. package/src/subsystems/LLMManager/LLM.service/connectors/openai/apiInterfaces/constants.ts +13 -13
  157. package/src/subsystems/LLMManager/LLM.service/connectors/openai/apiInterfaces/index.ts +4 -4
  158. package/src/subsystems/LLMManager/LLM.service/connectors/openai/apiInterfaces/utils.ts +11 -11
  159. package/src/subsystems/LLMManager/LLM.service/connectors/openai/types.ts +32 -32
  160. package/src/subsystems/LLMManager/LLM.service/connectors/xAI.class.ts +471 -471
  161. package/src/subsystems/LLMManager/LLM.service/index.ts +44 -44
  162. package/src/subsystems/LLMManager/ModelsProvider.service/ModelsProviderConnector.ts +300 -300
  163. package/src/subsystems/LLMManager/ModelsProvider.service/connectors/JSONModelsProvider.class.ts +252 -252
  164. package/src/subsystems/LLMManager/ModelsProvider.service/index.ts +11 -11
  165. package/src/subsystems/LLMManager/custom-models.ts +854 -854
  166. package/src/subsystems/LLMManager/models.ts +2540 -2540
  167. package/src/subsystems/LLMManager/paramMappings.ts +69 -69
  168. package/src/subsystems/MemoryManager/Cache.service/CacheConnector.ts +86 -86
  169. package/src/subsystems/MemoryManager/Cache.service/connectors/LocalStorageCache.class.ts +297 -297
  170. package/src/subsystems/MemoryManager/Cache.service/connectors/RAMCache.class.ts +201 -201
  171. package/src/subsystems/MemoryManager/Cache.service/connectors/RedisCache.class.ts +252 -252
  172. package/src/subsystems/MemoryManager/Cache.service/connectors/S3Cache.class.ts +373 -373
  173. package/src/subsystems/MemoryManager/Cache.service/index.ts +15 -15
  174. package/src/subsystems/MemoryManager/LLMCache.ts +72 -72
  175. package/src/subsystems/MemoryManager/LLMContext.ts +124 -124
  176. package/src/subsystems/MemoryManager/LLMMemory.service/LLMMemoryConnector.ts +26 -26
  177. package/src/subsystems/MemoryManager/RuntimeContext.ts +266 -266
  178. package/src/subsystems/Security/AccessControl/ACL.class.ts +208 -208
  179. package/src/subsystems/Security/AccessControl/AccessCandidate.class.ts +82 -82
  180. package/src/subsystems/Security/AccessControl/AccessRequest.class.ts +52 -52
  181. package/src/subsystems/Security/Account.service/AccountConnector.ts +44 -44
  182. package/src/subsystems/Security/Account.service/connectors/AWSAccount.class.ts +76 -76
  183. package/src/subsystems/Security/Account.service/connectors/DummyAccount.class.ts +130 -130
  184. package/src/subsystems/Security/Account.service/connectors/JSONFileAccount.class.ts +159 -159
  185. package/src/subsystems/Security/Account.service/index.ts +14 -14
  186. package/src/subsystems/Security/Credentials.helper.ts +62 -62
  187. package/src/subsystems/Security/ManagedVault.service/ManagedVaultConnector.ts +38 -38
  188. package/src/subsystems/Security/ManagedVault.service/connectors/NullManagedVault.class.ts +53 -53
  189. package/src/subsystems/Security/ManagedVault.service/connectors/SecretManagerManagedVault.ts +154 -154
  190. package/src/subsystems/Security/ManagedVault.service/index.ts +12 -12
  191. package/src/subsystems/Security/SecureConnector.class.ts +110 -110
  192. package/src/subsystems/Security/Vault.service/Vault.helper.ts +30 -30
  193. package/src/subsystems/Security/Vault.service/VaultConnector.ts +29 -29
  194. package/src/subsystems/Security/Vault.service/connectors/HashicorpVault.class.ts +46 -46
  195. package/src/subsystems/Security/Vault.service/connectors/JSONFileVault.class.ts +221 -221
  196. package/src/subsystems/Security/Vault.service/connectors/NullVault.class.ts +54 -54
  197. package/src/subsystems/Security/Vault.service/connectors/SecretsManager.class.ts +140 -140
  198. package/src/subsystems/Security/Vault.service/index.ts +12 -12
  199. package/src/types/ACL.types.ts +104 -104
  200. package/src/types/AWS.types.ts +10 -10
  201. package/src/types/Agent.types.ts +61 -61
  202. package/src/types/AgentLogger.types.ts +17 -17
  203. package/src/types/Cache.types.ts +1 -1
  204. package/src/types/Common.types.ts +2 -2
  205. package/src/types/LLM.types.ts +496 -496
  206. package/src/types/Redis.types.ts +8 -8
  207. package/src/types/SRE.types.ts +64 -64
  208. package/src/types/Security.types.ts +14 -14
  209. package/src/types/Storage.types.ts +5 -5
  210. package/src/types/VectorDB.types.ts +86 -86
  211. package/src/utils/base64.utils.ts +275 -275
  212. package/src/utils/cli.utils.ts +68 -68
  213. package/src/utils/data.utils.ts +322 -322
  214. package/src/utils/date-time.utils.ts +22 -22
  215. package/src/utils/general.utils.ts +238 -238
  216. package/src/utils/index.ts +12 -12
  217. package/src/utils/lazy-client.ts +261 -261
  218. package/src/utils/numbers.utils.ts +13 -13
  219. package/src/utils/oauth.utils.ts +35 -35
  220. package/src/utils/string.utils.ts +414 -414
  221. package/src/utils/url.utils.ts +19 -19
  222. package/src/utils/validation.utils.ts +74 -74
  223. package/dist/types/subsystems/LLMManager/ModelsProvider.service/connectors/SmythModelsProvider.class.d.ts +0 -39
@@ -1,54 +1,54 @@
1
- import Joi from 'joi';
2
-
3
- import { IAgent as Agent } from '@sre/types/Agent.types';
4
- import { Component } from './Component.class';
5
-
6
- export class JSONFilter extends Component {
7
- protected configSchema = Joi.object({
8
- fields: Joi.string().max(30000).allow('').label('Prompt'),
9
- });
10
- constructor() {
11
- super();
12
- }
13
- init() {}
14
- async process(input, config, agent: Agent) {
15
- await super.process(input, config, agent);
16
-
17
- const logger = this.createComponentLogger(agent, config);
18
- logger.debug(`=== JSONFilter Log ===`);
19
- let Output = {};
20
- let _error = null;
21
- try {
22
- const componentId = config.id;
23
- const fields = config.data.fields;
24
- const obj = input.Input;
25
-
26
- Output = filterFields(obj, fields);
27
- logger.debug(`Output filtered`);
28
- } catch (error: any) {
29
- _error = error;
30
- logger.error(` JSONFilter Error`, error);
31
- }
32
- return { Output, _error, _debug: logger.output };
33
- }
34
- }
35
-
36
- function filterFields(obj, fields) {
37
- const fieldList = fields?.split(',').map((field) => field.trim());
38
-
39
- function filterObject(obj) {
40
- if (Array.isArray(obj)) {
41
- return obj.map(filterObject);
42
- } else if (obj !== null && typeof obj === 'object') {
43
- return Object.keys(obj)
44
- .filter((key) => fieldList.includes(key))
45
- .reduce((acc, key) => {
46
- acc[key] = filterObject(obj[key]);
47
- return acc;
48
- }, {});
49
- }
50
- return obj;
51
- }
52
-
53
- return filterObject(obj);
54
- }
1
+ import Joi from 'joi';
2
+
3
+ import { IAgent as Agent } from '@sre/types/Agent.types';
4
+ import { Component } from './Component.class';
5
+
6
+ export class JSONFilter extends Component {
7
+ protected configSchema = Joi.object({
8
+ fields: Joi.string().max(30000).allow('').label('Prompt'),
9
+ });
10
+ constructor() {
11
+ super();
12
+ }
13
+ init() {}
14
+ async process(input, config, agent: Agent) {
15
+ await super.process(input, config, agent);
16
+
17
+ const logger = this.createComponentLogger(agent, config);
18
+ logger.debug(`=== JSONFilter Log ===`);
19
+ let Output = {};
20
+ let _error = null;
21
+ try {
22
+ const componentId = config.id;
23
+ const fields = config.data.fields;
24
+ const obj = input.Input;
25
+
26
+ Output = filterFields(obj, fields);
27
+ logger.debug(`Output filtered`);
28
+ } catch (error: any) {
29
+ _error = error;
30
+ logger.error(` JSONFilter Error`, error);
31
+ }
32
+ return { Output, _error, _debug: logger.output };
33
+ }
34
+ }
35
+
36
+ function filterFields(obj, fields) {
37
+ const fieldList = fields?.split(',').map((field) => field.trim());
38
+
39
+ function filterObject(obj) {
40
+ if (Array.isArray(obj)) {
41
+ return obj.map(filterObject);
42
+ } else if (obj !== null && typeof obj === 'object') {
43
+ return Object.keys(obj)
44
+ .filter((key) => fieldList.includes(key))
45
+ .reduce((acc, key) => {
46
+ acc[key] = filterObject(obj[key]);
47
+ return acc;
48
+ }, {});
49
+ }
50
+ return obj;
51
+ }
52
+
53
+ return filterObject(obj);
54
+ }
@@ -1,213 +1,213 @@
1
- import Joi from 'joi';
2
-
3
- import { IAgent as Agent } from '@sre/types/Agent.types';
4
- import { ConnectorService } from '@sre/Core/ConnectorsService';
5
- import { CacheConnector } from '@sre/MemoryManager/Cache.service/CacheConnector';
6
- import { AccessCandidate } from '@sre/Security/AccessControl/AccessCandidate.class';
7
- import { DEFAULT_MAX_TOKENS_FOR_LLM } from '@sre/constants';
8
- import { TemplateString } from '@sre/helpers/TemplateString.helper';
9
- import { encode } from 'gpt-tokenizer';
10
- import { Component } from './Component.class';
11
- import { JSONContent } from '@sre/helpers/JsonContent.helper';
12
- import { LLMInference } from '@sre/LLMManager/LLM.inference';
13
- import { TLLMMessageRole } from '@sre/types/LLM.types';
14
- import { VaultHelper } from '@sre/Security/Vault.service/Vault.helper';
15
- import path from 'path';
16
- import config from '@sre/config';
17
- import fs from 'fs/promises';
18
-
19
- //const sessions = {};
20
- let cacheConnector: CacheConnector;
21
- function getCacheConnector() {
22
- if (!cacheConnector) {
23
- cacheConnector = ConnectorService.getCacheConnector();
24
- }
25
- return cacheConnector;
26
- }
27
-
28
- async function saveMessagesToSession(agentId, userId, conversationId, messages, ttl?) {
29
- if (!userId && !conversationId) return;
30
- const cacheConnector = getCacheConnector();
31
- const conv_uid = `${agentId}:conv-u${userId}-c${conversationId}`;
32
-
33
- cacheConnector.requester(AccessCandidate.agent(agentId)).set(conv_uid, JSON.stringify(messages), null, null, ttl);
34
- }
35
-
36
- async function readMessagesFromSession(agentId, userId, conversationId, maxTokens = DEFAULT_MAX_TOKENS_FOR_LLM) {
37
- if (!userId && !conversationId) return [];
38
- const cacheConnector = getCacheConnector();
39
-
40
- const conv_uid = `${agentId}:conv-u${userId}-c${conversationId}`;
41
- //read the last messages from a given session and ensure that the total chat tokens are within the limit
42
- //start from the last message and keep adding messages until the total tokens exceed the limit
43
- //if (!sessions[agentId]) return [];
44
- //if (!sessions[agentId][conv_uid]) return [];
45
-
46
- const sessionData = await cacheConnector.requester(AccessCandidate.agent(agentId))?.get(conv_uid);
47
-
48
- let messages = sessionData ? JSONContent(sessionData).tryParse() : [];
49
-
50
- //const messages = sessions[agentId][conv_uid].messages;
51
-
52
- const filteredMessages: any[] = [];
53
-
54
- let tokens = 0;
55
- if (messages[0]?.role == 'system') {
56
- const encoded = encode(messages[0]?.content);
57
- const messageTokens = encoded.length + 3;
58
- tokens += messageTokens;
59
- }
60
-
61
- for (let i = messages.length - 1; i >= 0; i--) {
62
- if (messages[i].role == 'system') continue;
63
- const message = messages[i];
64
- const encoded = encode(message?.content);
65
- const messageTokens = encoded.length + 3;
66
- if (tokens + messageTokens > maxTokens) break;
67
- filteredMessages.unshift(message);
68
- tokens += messageTokens;
69
- }
70
-
71
- if (messages[0]?.role == 'system') filteredMessages.unshift(messages[0]);
72
-
73
- return filteredMessages;
74
- }
75
-
76
- //TODO : update this implementation to use ConversationManager
77
- // This will allow better context management and support for tool calls
78
- export class LLMAssistant extends Component {
79
- protected configSchema = Joi.object({
80
- model: Joi.string().max(200).required(),
81
- behavior: Joi.string().max(30000).allow('').label('Behavior'),
82
- passthrough: Joi.boolean().optional().label('Passthrough'),
83
- });
84
- constructor() {
85
- super();
86
- }
87
- init() {}
88
- async process(input, config, agent: Agent) {
89
- await super.process(input, config, agent);
90
- const logger = this.createComponentLogger(agent, config);
91
- try {
92
- logger.debug('== LLM Assistant Log ==\n');
93
-
94
- const passThrough: boolean = config.data.passthrough || false;
95
- const model: string = config.data.model || 'echo';
96
- const ttl = config.data.ttl || undefined;
97
- let teamId = agent?.teamId;
98
-
99
- const llmInference: LLMInference = await LLMInference.getInstance(model, AccessCandidate.agent(agent.id));
100
- // if the llm is undefined, then it means we removed the model from our system
101
- if (!llmInference.connector) {
102
- return {
103
- _error: `The model '${model}' is not available. Please try a different one.`,
104
- _debug: logger.output,
105
- };
106
- }
107
-
108
- const modelId = await agent.modelsProvider.getModelId(model);
109
- logger.debug(` Model : ${modelId || model}`);
110
-
111
- const userInput = input.UserInput;
112
- const userId = input.UserId;
113
- const conversationId = input.ConversationId;
114
-
115
- let behavior = TemplateString(config.data.behavior).parse(input).result;
116
- logger.debug(`[Parsed Behavior] \n${behavior}\n\n`);
117
-
118
- //#region get max tokens
119
- let maxTokens = 2048;
120
-
121
- const isStandardLLM = await agent.modelsProvider.isStandardLLM(model);
122
- const hasKey = true; //TODO : check if the user has a key
123
- //const modelInfo = await agent.modelsProvider.getModelInfo(model, hasKey);
124
- maxTokens = await agent.modelsProvider.getMaxCompletionTokens(model, hasKey);
125
-
126
- // if (isStandardLLM) {
127
- // const provider = LLMRegistry.getProvider(model);
128
- // const apiKey = await VaultHelper.getAgentKey(provider, agent?.id);
129
- // maxTokens = LLMRegistry.getMaxCompletionTokens(model, !!apiKey);
130
- // } else {
131
- // const team = AccessCandidate.team(teamId);
132
- // const customLLMRegistry = await CustomLLMRegistry.getInstance(team);
133
- // maxTokens = await customLLMRegistry.getMaxCompletionTokens(model);
134
- // }
135
- //#endregion get max tokens
136
-
137
- const messages: any[] = await readMessagesFromSession(agent.id, userId, conversationId, Math.round(maxTokens / 2));
138
-
139
- messages.push({ role: TLLMMessageRole.User, content: userInput });
140
-
141
- if (messages[0]?.role != TLLMMessageRole.System) {
142
- messages.unshift({ role: TLLMMessageRole.System, content: behavior });
143
- }
144
-
145
- const customParams = {
146
- messages,
147
- };
148
-
149
- let response: any;
150
- if (passThrough) {
151
- const contentPromise = new Promise(async (resolve, reject) => {
152
- let _content = '';
153
- const eventEmitter: any = await llmInference
154
- .promptStream({
155
- contextWindow: messages,
156
- params: { ...config, model, agentId: agent.id },
157
- })
158
- .catch((error) => {
159
- console.error('Error on promptStream: ', error);
160
- reject(error);
161
- });
162
- eventEmitter.on('content', (content) => {
163
- if (typeof agent.callback === 'function') {
164
- agent.callback({ content });
165
- }
166
- agent.sse.send('llm/passthrough/content', content);
167
- _content += content;
168
- });
169
- eventEmitter.on('thinking', (thinking) => {
170
- if (typeof agent.callback === 'function') {
171
- agent.callback({ thinking });
172
- }
173
- agent.sse.send('llm/passthrough/thinking', thinking);
174
- });
175
- eventEmitter.on('end', () => {
176
- console.log('end');
177
- resolve(_content);
178
- });
179
- });
180
- response = await contentPromise;
181
- } else {
182
- response = await llmInference
183
- .prompt({ contextWindow: messages, params: { ...config, agentId: agent.id } })
184
- .catch((error) => ({ error: error }));
185
- }
186
-
187
- // in case we have the response but it's empty string, undefined or null
188
- if (!response) {
189
- return { _error: ' LLM Error = Empty Response!', _debug: logger.output };
190
- }
191
-
192
- if (response?.error) {
193
- const error = response?.error + ' ' + (response?.details || '');
194
- logger.error(` LLM Error=`, error);
195
-
196
- return { Response: response?.data, _error: error, _debug: logger.output };
197
- }
198
-
199
- messages.push({ role: 'assistant', content: response });
200
- saveMessagesToSession(agent.id, userId, conversationId, messages, ttl);
201
-
202
- logger.debug(' Response \n', response);
203
-
204
- const result = { Response: response };
205
-
206
- result['_debug'] = logger.output;
207
-
208
- return result;
209
- } catch (error) {
210
- return { _error: error.message, _debug: logger.output };
211
- }
212
- }
213
- }
1
+ import Joi from 'joi';
2
+
3
+ import { IAgent as Agent } from '@sre/types/Agent.types';
4
+ import { ConnectorService } from '@sre/Core/ConnectorsService';
5
+ import { CacheConnector } from '@sre/MemoryManager/Cache.service/CacheConnector';
6
+ import { AccessCandidate } from '@sre/Security/AccessControl/AccessCandidate.class';
7
+ import { DEFAULT_MAX_TOKENS_FOR_LLM } from '@sre/constants';
8
+ import { TemplateString } from '@sre/helpers/TemplateString.helper';
9
+ import { encode } from 'gpt-tokenizer';
10
+ import { Component } from './Component.class';
11
+ import { JSONContent } from '@sre/helpers/JsonContent.helper';
12
+ import { LLMInference } from '@sre/LLMManager/LLM.inference';
13
+ import { TLLMMessageRole } from '@sre/types/LLM.types';
14
+ import { VaultHelper } from '@sre/Security/Vault.service/Vault.helper';
15
+ import path from 'path';
16
+ import config from '@sre/config';
17
+ import fs from 'fs/promises';
18
+
19
+ //const sessions = {};
20
+ let cacheConnector: CacheConnector;
21
+ function getCacheConnector() {
22
+ if (!cacheConnector) {
23
+ cacheConnector = ConnectorService.getCacheConnector();
24
+ }
25
+ return cacheConnector;
26
+ }
27
+
28
+ async function saveMessagesToSession(agentId, userId, conversationId, messages, ttl?) {
29
+ if (!userId && !conversationId) return;
30
+ const cacheConnector = getCacheConnector();
31
+ const conv_uid = `${agentId}:conv-u${userId}-c${conversationId}`;
32
+
33
+ cacheConnector.requester(AccessCandidate.agent(agentId)).set(conv_uid, JSON.stringify(messages), null, null, ttl);
34
+ }
35
+
36
+ async function readMessagesFromSession(agentId, userId, conversationId, maxTokens = DEFAULT_MAX_TOKENS_FOR_LLM) {
37
+ if (!userId && !conversationId) return [];
38
+ const cacheConnector = getCacheConnector();
39
+
40
+ const conv_uid = `${agentId}:conv-u${userId}-c${conversationId}`;
41
+ //read the last messages from a given session and ensure that the total chat tokens are within the limit
42
+ //start from the last message and keep adding messages until the total tokens exceed the limit
43
+ //if (!sessions[agentId]) return [];
44
+ //if (!sessions[agentId][conv_uid]) return [];
45
+
46
+ const sessionData = await cacheConnector.requester(AccessCandidate.agent(agentId))?.get(conv_uid);
47
+
48
+ let messages = sessionData ? JSONContent(sessionData).tryParse() : [];
49
+
50
+ //const messages = sessions[agentId][conv_uid].messages;
51
+
52
+ const filteredMessages: any[] = [];
53
+
54
+ let tokens = 0;
55
+ if (messages[0]?.role == 'system') {
56
+ const encoded = encode(messages[0]?.content);
57
+ const messageTokens = encoded.length + 3;
58
+ tokens += messageTokens;
59
+ }
60
+
61
+ for (let i = messages.length - 1; i >= 0; i--) {
62
+ if (messages[i].role == 'system') continue;
63
+ const message = messages[i];
64
+ const encoded = encode(message?.content);
65
+ const messageTokens = encoded.length + 3;
66
+ if (tokens + messageTokens > maxTokens) break;
67
+ filteredMessages.unshift(message);
68
+ tokens += messageTokens;
69
+ }
70
+
71
+ if (messages[0]?.role == 'system') filteredMessages.unshift(messages[0]);
72
+
73
+ return filteredMessages;
74
+ }
75
+
76
+ //TODO : update this implementation to use ConversationManager
77
+ // This will allow better context management and support for tool calls
78
+ export class LLMAssistant extends Component {
79
+ protected configSchema = Joi.object({
80
+ model: Joi.string().max(200).required(),
81
+ behavior: Joi.string().max(30000).allow('').label('Behavior'),
82
+ passthrough: Joi.boolean().optional().label('Passthrough'),
83
+ });
84
+ constructor() {
85
+ super();
86
+ }
87
+ init() {}
88
+ async process(input, config, agent: Agent) {
89
+ await super.process(input, config, agent);
90
+ const logger = this.createComponentLogger(agent, config);
91
+ try {
92
+ logger.debug('== LLM Assistant Log ==\n');
93
+
94
+ const passThrough: boolean = config.data.passthrough || false;
95
+ const model: string = config.data.model || 'echo';
96
+ const ttl = config.data.ttl || undefined;
97
+ let teamId = agent?.teamId;
98
+
99
+ const llmInference: LLMInference = await LLMInference.getInstance(model, AccessCandidate.agent(agent.id));
100
+ // if the llm is undefined, then it means we removed the model from our system
101
+ if (!llmInference.connector) {
102
+ return {
103
+ _error: `The model '${model}' is not available. Please try a different one.`,
104
+ _debug: logger.output,
105
+ };
106
+ }
107
+
108
+ const modelId = await agent.modelsProvider.getModelId(model);
109
+ logger.debug(` Model : ${modelId || model}`);
110
+
111
+ const userInput = input.UserInput;
112
+ const userId = input.UserId;
113
+ const conversationId = input.ConversationId;
114
+
115
+ let behavior = TemplateString(config.data.behavior).parse(input).result;
116
+ logger.debug(`[Parsed Behavior] \n${behavior}\n\n`);
117
+
118
+ //#region get max tokens
119
+ let maxTokens = 2048;
120
+
121
+ const isStandardLLM = await agent.modelsProvider.isStandardLLM(model);
122
+ const hasKey = true; //TODO : check if the user has a key
123
+ //const modelInfo = await agent.modelsProvider.getModelInfo(model, hasKey);
124
+ maxTokens = await agent.modelsProvider.getMaxCompletionTokens(model, hasKey);
125
+
126
+ // if (isStandardLLM) {
127
+ // const provider = LLMRegistry.getProvider(model);
128
+ // const apiKey = await VaultHelper.getAgentKey(provider, agent?.id);
129
+ // maxTokens = LLMRegistry.getMaxCompletionTokens(model, !!apiKey);
130
+ // } else {
131
+ // const team = AccessCandidate.team(teamId);
132
+ // const customLLMRegistry = await CustomLLMRegistry.getInstance(team);
133
+ // maxTokens = await customLLMRegistry.getMaxCompletionTokens(model);
134
+ // }
135
+ //#endregion get max tokens
136
+
137
+ const messages: any[] = await readMessagesFromSession(agent.id, userId, conversationId, Math.round(maxTokens / 2));
138
+
139
+ messages.push({ role: TLLMMessageRole.User, content: userInput });
140
+
141
+ if (messages[0]?.role != TLLMMessageRole.System) {
142
+ messages.unshift({ role: TLLMMessageRole.System, content: behavior });
143
+ }
144
+
145
+ const customParams = {
146
+ messages,
147
+ };
148
+
149
+ let response: any;
150
+ if (passThrough) {
151
+ const contentPromise = new Promise(async (resolve, reject) => {
152
+ let _content = '';
153
+ const eventEmitter: any = await llmInference
154
+ .promptStream({
155
+ contextWindow: messages,
156
+ params: { ...config, model, agentId: agent.id },
157
+ })
158
+ .catch((error) => {
159
+ console.error('Error on promptStream: ', error);
160
+ reject(error);
161
+ });
162
+ eventEmitter.on('content', (content) => {
163
+ if (typeof agent.callback === 'function') {
164
+ agent.callback({ content });
165
+ }
166
+ agent.sse.send('llm/passthrough/content', content);
167
+ _content += content;
168
+ });
169
+ eventEmitter.on('thinking', (thinking) => {
170
+ if (typeof agent.callback === 'function') {
171
+ agent.callback({ thinking });
172
+ }
173
+ agent.sse.send('llm/passthrough/thinking', thinking);
174
+ });
175
+ eventEmitter.on('end', () => {
176
+ console.log('end');
177
+ resolve(_content);
178
+ });
179
+ });
180
+ response = await contentPromise;
181
+ } else {
182
+ response = await llmInference
183
+ .prompt({ contextWindow: messages, params: { ...config, agentId: agent.id } })
184
+ .catch((error) => ({ error: error }));
185
+ }
186
+
187
+ // in case we have the response but it's empty string, undefined or null
188
+ if (!response) {
189
+ return { _error: ' LLM Error = Empty Response!', _debug: logger.output };
190
+ }
191
+
192
+ if (response?.error) {
193
+ const error = response?.error + ' ' + (response?.details || '');
194
+ logger.error(` LLM Error=`, error);
195
+
196
+ return { Response: response?.data, _error: error, _debug: logger.output };
197
+ }
198
+
199
+ messages.push({ role: 'assistant', content: response });
200
+ saveMessagesToSession(agent.id, userId, conversationId, messages, ttl);
201
+
202
+ logger.debug(' Response \n', response);
203
+
204
+ const result = { Response: response };
205
+
206
+ result['_debug'] = logger.output;
207
+
208
+ return result;
209
+ } catch (error) {
210
+ return { _error: error.message, _debug: logger.output };
211
+ }
212
+ }
213
+ }
@@ -1,28 +1,28 @@
1
- import { IAgent as Agent } from '@sre/types/Agent.types';
2
- import { Component } from './Component.class';
3
-
4
- export class LogicAND extends Component {
5
- constructor() {
6
- super();
7
- }
8
- init() {}
9
- async process(input, config, agent: Agent) {
10
- await super.process(input, config, agent);
11
- const logger = this.createComponentLogger(agent, config);
12
- const result: any = { Output: true };
13
-
14
- for (let cfgInput of config.inputs) {
15
- // check if all inputs are set (expected inputs are in "config.inputs" actual inputs are in "input")
16
- if (!input[cfgInput.name]) {
17
- result.Output = undefined;
18
- break;
19
- }
20
- }
21
-
22
- result.Verified = result.Output !== undefined;
23
- result.Unverified = !result.Verified;
24
- if (!result.Verified) delete result.Verified;
25
- if (!result.Unverified) delete result.Unverified;
26
- return result;
27
- }
28
- }
1
+ import { IAgent as Agent } from '@sre/types/Agent.types';
2
+ import { Component } from './Component.class';
3
+
4
+ export class LogicAND extends Component {
5
+ constructor() {
6
+ super();
7
+ }
8
+ init() {}
9
+ async process(input, config, agent: Agent) {
10
+ await super.process(input, config, agent);
11
+ const logger = this.createComponentLogger(agent, config);
12
+ const result: any = { Output: true };
13
+
14
+ for (let cfgInput of config.inputs) {
15
+ // check if all inputs are set (expected inputs are in "config.inputs" actual inputs are in "input")
16
+ if (!input[cfgInput.name]) {
17
+ result.Output = undefined;
18
+ break;
19
+ }
20
+ }
21
+
22
+ result.Verified = result.Output !== undefined;
23
+ result.Unverified = !result.Verified;
24
+ if (!result.Verified) delete result.Verified;
25
+ if (!result.Unverified) delete result.Unverified;
26
+ return result;
27
+ }
28
+ }