@smythos/sre 1.5.46 → 1.5.50
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG +98 -90
- package/LICENSE +18 -18
- package/README.md +135 -135
- package/dist/bundle-analysis-lazy.html +4949 -0
- package/dist/bundle-analysis.html +4949 -0
- package/dist/index.js +3 -3
- package/dist/index.js.map +1 -1
- package/dist/types/Components/MCPClient.class.d.ts +1 -0
- package/dist/types/subsystems/LLMManager/LLM.service/connectors/openai/OpenAIConnector.class.d.ts +1 -6
- package/dist/types/utils/package-manager.utils.d.ts +26 -0
- package/package.json +1 -1
- package/src/Components/APICall/APICall.class.ts +156 -156
- package/src/Components/APICall/AccessTokenManager.ts +130 -130
- package/src/Components/APICall/ArrayBufferResponse.helper.ts +58 -58
- package/src/Components/APICall/OAuth.helper.ts +294 -294
- package/src/Components/APICall/mimeTypeCategories.ts +46 -46
- package/src/Components/APICall/parseData.ts +167 -167
- package/src/Components/APICall/parseHeaders.ts +41 -41
- package/src/Components/APICall/parseProxy.ts +68 -68
- package/src/Components/APICall/parseUrl.ts +91 -91
- package/src/Components/APIEndpoint.class.ts +234 -234
- package/src/Components/APIOutput.class.ts +58 -58
- package/src/Components/AgentPlugin.class.ts +102 -102
- package/src/Components/Async.class.ts +155 -155
- package/src/Components/Await.class.ts +90 -90
- package/src/Components/Classifier.class.ts +158 -158
- package/src/Components/Component.class.ts +132 -132
- package/src/Components/ComponentHost.class.ts +38 -38
- package/src/Components/DataSourceCleaner.class.ts +92 -92
- package/src/Components/DataSourceIndexer.class.ts +181 -181
- package/src/Components/DataSourceLookup.class.ts +161 -161
- package/src/Components/ECMASandbox.class.ts +71 -71
- package/src/Components/FEncDec.class.ts +29 -29
- package/src/Components/FHash.class.ts +33 -33
- package/src/Components/FSign.class.ts +80 -80
- package/src/Components/FSleep.class.ts +25 -25
- package/src/Components/FTimestamp.class.ts +25 -25
- package/src/Components/FileStore.class.ts +78 -78
- package/src/Components/ForEach.class.ts +97 -97
- package/src/Components/GPTPlugin.class.ts +70 -70
- package/src/Components/GenAILLM.class.ts +586 -586
- package/src/Components/HuggingFace.class.ts +314 -314
- package/src/Components/Image/imageSettings.config.ts +70 -70
- package/src/Components/ImageGenerator.class.ts +502 -502
- package/src/Components/JSONFilter.class.ts +54 -54
- package/src/Components/LLMAssistant.class.ts +213 -213
- package/src/Components/LogicAND.class.ts +28 -28
- package/src/Components/LogicAtLeast.class.ts +85 -85
- package/src/Components/LogicAtMost.class.ts +86 -86
- package/src/Components/LogicOR.class.ts +29 -29
- package/src/Components/LogicXOR.class.ts +34 -34
- package/src/Components/MCPClient.class.ts +138 -112
- package/src/Components/MemoryDeleteKeyVal.class.ts +70 -70
- package/src/Components/MemoryReadKeyVal.class.ts +66 -66
- package/src/Components/MemoryWriteKeyVal.class.ts +62 -62
- package/src/Components/MemoryWriteObject.class.ts +97 -97
- package/src/Components/MultimodalLLM.class.ts +128 -128
- package/src/Components/OpenAPI.class.ts +72 -72
- package/src/Components/PromptGenerator.class.ts +122 -122
- package/src/Components/ScrapflyWebScrape.class.ts +159 -159
- package/src/Components/ServerlessCode.class.ts +123 -123
- package/src/Components/TavilyWebSearch.class.ts +98 -98
- package/src/Components/VisionLLM.class.ts +104 -104
- package/src/Components/ZapierAction.class.ts +127 -127
- package/src/Components/index.ts +97 -97
- package/src/Core/AgentProcess.helper.ts +240 -240
- package/src/Core/Connector.class.ts +123 -123
- package/src/Core/ConnectorsService.ts +197 -197
- package/src/Core/DummyConnector.ts +49 -49
- package/src/Core/HookService.ts +105 -105
- package/src/Core/SmythRuntime.class.ts +235 -235
- package/src/Core/SystemEvents.ts +16 -16
- package/src/Core/boot.ts +56 -56
- package/src/config.ts +15 -15
- package/src/constants.ts +126 -126
- package/src/data/hugging-face.params.json +579 -579
- package/src/helpers/AWSLambdaCode.helper.ts +587 -587
- package/src/helpers/BinaryInput.helper.ts +331 -331
- package/src/helpers/Conversation.helper.ts +1119 -1119
- package/src/helpers/ECMASandbox.helper.ts +54 -54
- package/src/helpers/JsonContent.helper.ts +97 -97
- package/src/helpers/LocalCache.helper.ts +97 -97
- package/src/helpers/Log.helper.ts +274 -274
- package/src/helpers/OpenApiParser.helper.ts +150 -150
- package/src/helpers/S3Cache.helper.ts +147 -147
- package/src/helpers/SmythURI.helper.ts +5 -5
- package/src/helpers/Sysconfig.helper.ts +77 -77
- package/src/helpers/TemplateString.helper.ts +243 -243
- package/src/helpers/TypeChecker.helper.ts +329 -329
- package/src/index.ts +3 -3
- package/src/index.ts.bak +3 -3
- package/src/subsystems/AgentManager/Agent.class.ts +1114 -1114
- package/src/subsystems/AgentManager/Agent.helper.ts +3 -3
- package/src/subsystems/AgentManager/AgentData.service/AgentDataConnector.ts +230 -230
- package/src/subsystems/AgentManager/AgentData.service/connectors/CLIAgentDataConnector.class.ts +66 -66
- package/src/subsystems/AgentManager/AgentData.service/connectors/LocalAgentDataConnector.class.ts +142 -142
- package/src/subsystems/AgentManager/AgentData.service/connectors/NullAgentData.class.ts +39 -39
- package/src/subsystems/AgentManager/AgentData.service/index.ts +18 -18
- package/src/subsystems/AgentManager/AgentLogger.class.ts +297 -297
- package/src/subsystems/AgentManager/AgentRequest.class.ts +51 -51
- package/src/subsystems/AgentManager/AgentRuntime.class.ts +559 -559
- package/src/subsystems/AgentManager/AgentSSE.class.ts +101 -101
- package/src/subsystems/AgentManager/AgentSettings.class.ts +52 -52
- package/src/subsystems/AgentManager/Component.service/ComponentConnector.ts +32 -32
- package/src/subsystems/AgentManager/Component.service/connectors/LocalComponentConnector.class.ts +60 -60
- package/src/subsystems/AgentManager/Component.service/index.ts +11 -11
- package/src/subsystems/AgentManager/EmbodimentSettings.class.ts +47 -47
- package/src/subsystems/AgentManager/ForkedAgent.class.ts +154 -154
- package/src/subsystems/AgentManager/OSResourceMonitor.ts +77 -77
- package/src/subsystems/ComputeManager/Code.service/CodeConnector.ts +98 -98
- package/src/subsystems/ComputeManager/Code.service/connectors/AWSLambdaCode.class.ts +172 -172
- package/src/subsystems/ComputeManager/Code.service/connectors/ECMASandbox.class.ts +131 -131
- package/src/subsystems/ComputeManager/Code.service/index.ts +13 -13
- package/src/subsystems/IO/CLI.service/CLIConnector.ts +47 -47
- package/src/subsystems/IO/CLI.service/index.ts +9 -9
- package/src/subsystems/IO/Log.service/LogConnector.ts +32 -32
- package/src/subsystems/IO/Log.service/connectors/ConsoleLog.class.ts +28 -28
- package/src/subsystems/IO/Log.service/index.ts +13 -13
- package/src/subsystems/IO/NKV.service/NKVConnector.ts +43 -43
- package/src/subsystems/IO/NKV.service/connectors/NKVLocalStorage.class.ts +234 -234
- package/src/subsystems/IO/NKV.service/connectors/NKVRAM.class.ts +204 -204
- package/src/subsystems/IO/NKV.service/connectors/NKVRedis.class.ts +182 -182
- package/src/subsystems/IO/NKV.service/index.ts +14 -14
- package/src/subsystems/IO/Router.service/RouterConnector.ts +21 -21
- package/src/subsystems/IO/Router.service/connectors/ExpressRouter.class.ts +48 -48
- package/src/subsystems/IO/Router.service/connectors/NullRouter.class.ts +40 -40
- package/src/subsystems/IO/Router.service/index.ts +11 -11
- package/src/subsystems/IO/Storage.service/SmythFS.class.ts +489 -489
- package/src/subsystems/IO/Storage.service/StorageConnector.ts +66 -66
- package/src/subsystems/IO/Storage.service/connectors/LocalStorage.class.ts +327 -327
- package/src/subsystems/IO/Storage.service/connectors/S3Storage.class.ts +482 -482
- package/src/subsystems/IO/Storage.service/index.ts +13 -13
- package/src/subsystems/IO/VectorDB.service/VectorDBConnector.ts +108 -108
- package/src/subsystems/IO/VectorDB.service/connectors/MilvusVectorDB.class.ts +454 -454
- package/src/subsystems/IO/VectorDB.service/connectors/PineconeVectorDB.class.ts +384 -384
- package/src/subsystems/IO/VectorDB.service/connectors/RAMVecrtorDB.class.ts +421 -421
- package/src/subsystems/IO/VectorDB.service/embed/BaseEmbedding.ts +107 -107
- package/src/subsystems/IO/VectorDB.service/embed/OpenAIEmbedding.ts +109 -109
- package/src/subsystems/IO/VectorDB.service/embed/index.ts +21 -21
- package/src/subsystems/IO/VectorDB.service/index.ts +14 -14
- package/src/subsystems/LLMManager/LLM.helper.ts +251 -251
- package/src/subsystems/LLMManager/LLM.inference.ts +339 -339
- package/src/subsystems/LLMManager/LLM.service/LLMConnector.ts +489 -489
- package/src/subsystems/LLMManager/LLM.service/LLMCredentials.helper.ts +171 -171
- package/src/subsystems/LLMManager/LLM.service/connectors/Anthropic.class.ts +659 -659
- package/src/subsystems/LLMManager/LLM.service/connectors/Bedrock.class.ts +400 -400
- package/src/subsystems/LLMManager/LLM.service/connectors/Echo.class.ts +77 -77
- package/src/subsystems/LLMManager/LLM.service/connectors/GoogleAI.class.ts +757 -757
- package/src/subsystems/LLMManager/LLM.service/connectors/Groq.class.ts +304 -304
- package/src/subsystems/LLMManager/LLM.service/connectors/Perplexity.class.ts +250 -250
- package/src/subsystems/LLMManager/LLM.service/connectors/VertexAI.class.ts +423 -423
- package/src/subsystems/LLMManager/LLM.service/connectors/openai/OpenAIConnector.class.ts +488 -488
- package/src/subsystems/LLMManager/LLM.service/connectors/openai/apiInterfaces/ChatCompletionsApiInterface.ts +524 -524
- package/src/subsystems/LLMManager/LLM.service/connectors/openai/apiInterfaces/OpenAIApiInterface.ts +100 -100
- package/src/subsystems/LLMManager/LLM.service/connectors/openai/apiInterfaces/OpenAIApiInterfaceFactory.ts +81 -81
- package/src/subsystems/LLMManager/LLM.service/connectors/openai/apiInterfaces/ResponsesApiInterface.ts +1145 -1145
- package/src/subsystems/LLMManager/LLM.service/connectors/openai/apiInterfaces/constants.ts +13 -13
- package/src/subsystems/LLMManager/LLM.service/connectors/openai/apiInterfaces/index.ts +4 -4
- package/src/subsystems/LLMManager/LLM.service/connectors/openai/apiInterfaces/utils.ts +11 -11
- package/src/subsystems/LLMManager/LLM.service/connectors/openai/types.ts +32 -32
- package/src/subsystems/LLMManager/LLM.service/connectors/xAI.class.ts +471 -471
- package/src/subsystems/LLMManager/LLM.service/index.ts +44 -44
- package/src/subsystems/LLMManager/ModelsProvider.service/ModelsProviderConnector.ts +300 -300
- package/src/subsystems/LLMManager/ModelsProvider.service/connectors/JSONModelsProvider.class.ts +252 -252
- package/src/subsystems/LLMManager/ModelsProvider.service/index.ts +11 -11
- package/src/subsystems/LLMManager/custom-models.ts +854 -854
- package/src/subsystems/LLMManager/models.ts +2540 -2540
- package/src/subsystems/LLMManager/paramMappings.ts +69 -69
- package/src/subsystems/MemoryManager/Cache.service/CacheConnector.ts +86 -86
- package/src/subsystems/MemoryManager/Cache.service/connectors/LocalStorageCache.class.ts +297 -297
- package/src/subsystems/MemoryManager/Cache.service/connectors/RAMCache.class.ts +201 -201
- package/src/subsystems/MemoryManager/Cache.service/connectors/RedisCache.class.ts +252 -252
- package/src/subsystems/MemoryManager/Cache.service/connectors/S3Cache.class.ts +373 -373
- package/src/subsystems/MemoryManager/Cache.service/index.ts +15 -15
- package/src/subsystems/MemoryManager/LLMCache.ts +72 -72
- package/src/subsystems/MemoryManager/LLMContext.ts +124 -124
- package/src/subsystems/MemoryManager/LLMMemory.service/LLMMemoryConnector.ts +26 -26
- package/src/subsystems/MemoryManager/RuntimeContext.ts +266 -266
- package/src/subsystems/Security/AccessControl/ACL.class.ts +208 -208
- package/src/subsystems/Security/AccessControl/AccessCandidate.class.ts +82 -82
- package/src/subsystems/Security/AccessControl/AccessRequest.class.ts +52 -52
- package/src/subsystems/Security/Account.service/AccountConnector.ts +44 -44
- package/src/subsystems/Security/Account.service/connectors/AWSAccount.class.ts +76 -76
- package/src/subsystems/Security/Account.service/connectors/DummyAccount.class.ts +130 -130
- package/src/subsystems/Security/Account.service/connectors/JSONFileAccount.class.ts +159 -159
- package/src/subsystems/Security/Account.service/index.ts +14 -14
- package/src/subsystems/Security/Credentials.helper.ts +62 -62
- package/src/subsystems/Security/ManagedVault.service/ManagedVaultConnector.ts +38 -38
- package/src/subsystems/Security/ManagedVault.service/connectors/NullManagedVault.class.ts +53 -53
- package/src/subsystems/Security/ManagedVault.service/connectors/SecretManagerManagedVault.ts +154 -154
- package/src/subsystems/Security/ManagedVault.service/index.ts +12 -12
- package/src/subsystems/Security/SecureConnector.class.ts +110 -110
- package/src/subsystems/Security/Vault.service/Vault.helper.ts +30 -30
- package/src/subsystems/Security/Vault.service/VaultConnector.ts +29 -29
- package/src/subsystems/Security/Vault.service/connectors/HashicorpVault.class.ts +46 -46
- package/src/subsystems/Security/Vault.service/connectors/JSONFileVault.class.ts +221 -221
- package/src/subsystems/Security/Vault.service/connectors/NullVault.class.ts +54 -54
- package/src/subsystems/Security/Vault.service/connectors/SecretsManager.class.ts +140 -140
- package/src/subsystems/Security/Vault.service/index.ts +12 -12
- package/src/types/ACL.types.ts +104 -104
- package/src/types/AWS.types.ts +10 -10
- package/src/types/Agent.types.ts +61 -61
- package/src/types/AgentLogger.types.ts +17 -17
- package/src/types/Cache.types.ts +1 -1
- package/src/types/Common.types.ts +2 -2
- package/src/types/LLM.types.ts +496 -496
- package/src/types/Redis.types.ts +8 -8
- package/src/types/SRE.types.ts +64 -64
- package/src/types/Security.types.ts +14 -14
- package/src/types/Storage.types.ts +5 -5
- package/src/types/VectorDB.types.ts +86 -86
- package/src/utils/base64.utils.ts +275 -275
- package/src/utils/cli.utils.ts +68 -68
- package/src/utils/data.utils.ts +322 -322
- package/src/utils/date-time.utils.ts +22 -22
- package/src/utils/general.utils.ts +238 -238
- package/src/utils/index.ts +12 -12
- package/src/utils/lazy-client.ts +261 -261
- package/src/utils/numbers.utils.ts +13 -13
- package/src/utils/oauth.utils.ts +35 -35
- package/src/utils/string.utils.ts +414 -414
- package/src/utils/url.utils.ts +19 -19
- package/src/utils/validation.utils.ts +74 -74
- package/dist/types/subsystems/LLMManager/ModelsProvider.service/connectors/SmythModelsProvider.class.d.ts +0 -39
|
@@ -1,54 +1,54 @@
|
|
|
1
|
-
import Joi from 'joi';
|
|
2
|
-
|
|
3
|
-
import { IAgent as Agent } from '@sre/types/Agent.types';
|
|
4
|
-
import { Component } from './Component.class';
|
|
5
|
-
|
|
6
|
-
export class JSONFilter extends Component {
|
|
7
|
-
protected configSchema = Joi.object({
|
|
8
|
-
fields: Joi.string().max(30000).allow('').label('Prompt'),
|
|
9
|
-
});
|
|
10
|
-
constructor() {
|
|
11
|
-
super();
|
|
12
|
-
}
|
|
13
|
-
init() {}
|
|
14
|
-
async process(input, config, agent: Agent) {
|
|
15
|
-
await super.process(input, config, agent);
|
|
16
|
-
|
|
17
|
-
const logger = this.createComponentLogger(agent, config);
|
|
18
|
-
logger.debug(`=== JSONFilter Log ===`);
|
|
19
|
-
let Output = {};
|
|
20
|
-
let _error = null;
|
|
21
|
-
try {
|
|
22
|
-
const componentId = config.id;
|
|
23
|
-
const fields = config.data.fields;
|
|
24
|
-
const obj = input.Input;
|
|
25
|
-
|
|
26
|
-
Output = filterFields(obj, fields);
|
|
27
|
-
logger.debug(`Output filtered`);
|
|
28
|
-
} catch (error: any) {
|
|
29
|
-
_error = error;
|
|
30
|
-
logger.error(` JSONFilter Error`, error);
|
|
31
|
-
}
|
|
32
|
-
return { Output, _error, _debug: logger.output };
|
|
33
|
-
}
|
|
34
|
-
}
|
|
35
|
-
|
|
36
|
-
function filterFields(obj, fields) {
|
|
37
|
-
const fieldList = fields?.split(',').map((field) => field.trim());
|
|
38
|
-
|
|
39
|
-
function filterObject(obj) {
|
|
40
|
-
if (Array.isArray(obj)) {
|
|
41
|
-
return obj.map(filterObject);
|
|
42
|
-
} else if (obj !== null && typeof obj === 'object') {
|
|
43
|
-
return Object.keys(obj)
|
|
44
|
-
.filter((key) => fieldList.includes(key))
|
|
45
|
-
.reduce((acc, key) => {
|
|
46
|
-
acc[key] = filterObject(obj[key]);
|
|
47
|
-
return acc;
|
|
48
|
-
}, {});
|
|
49
|
-
}
|
|
50
|
-
return obj;
|
|
51
|
-
}
|
|
52
|
-
|
|
53
|
-
return filterObject(obj);
|
|
54
|
-
}
|
|
1
|
+
import Joi from 'joi';
|
|
2
|
+
|
|
3
|
+
import { IAgent as Agent } from '@sre/types/Agent.types';
|
|
4
|
+
import { Component } from './Component.class';
|
|
5
|
+
|
|
6
|
+
export class JSONFilter extends Component {
|
|
7
|
+
protected configSchema = Joi.object({
|
|
8
|
+
fields: Joi.string().max(30000).allow('').label('Prompt'),
|
|
9
|
+
});
|
|
10
|
+
constructor() {
|
|
11
|
+
super();
|
|
12
|
+
}
|
|
13
|
+
init() {}
|
|
14
|
+
async process(input, config, agent: Agent) {
|
|
15
|
+
await super.process(input, config, agent);
|
|
16
|
+
|
|
17
|
+
const logger = this.createComponentLogger(agent, config);
|
|
18
|
+
logger.debug(`=== JSONFilter Log ===`);
|
|
19
|
+
let Output = {};
|
|
20
|
+
let _error = null;
|
|
21
|
+
try {
|
|
22
|
+
const componentId = config.id;
|
|
23
|
+
const fields = config.data.fields;
|
|
24
|
+
const obj = input.Input;
|
|
25
|
+
|
|
26
|
+
Output = filterFields(obj, fields);
|
|
27
|
+
logger.debug(`Output filtered`);
|
|
28
|
+
} catch (error: any) {
|
|
29
|
+
_error = error;
|
|
30
|
+
logger.error(` JSONFilter Error`, error);
|
|
31
|
+
}
|
|
32
|
+
return { Output, _error, _debug: logger.output };
|
|
33
|
+
}
|
|
34
|
+
}
|
|
35
|
+
|
|
36
|
+
function filterFields(obj, fields) {
|
|
37
|
+
const fieldList = fields?.split(',').map((field) => field.trim());
|
|
38
|
+
|
|
39
|
+
function filterObject(obj) {
|
|
40
|
+
if (Array.isArray(obj)) {
|
|
41
|
+
return obj.map(filterObject);
|
|
42
|
+
} else if (obj !== null && typeof obj === 'object') {
|
|
43
|
+
return Object.keys(obj)
|
|
44
|
+
.filter((key) => fieldList.includes(key))
|
|
45
|
+
.reduce((acc, key) => {
|
|
46
|
+
acc[key] = filterObject(obj[key]);
|
|
47
|
+
return acc;
|
|
48
|
+
}, {});
|
|
49
|
+
}
|
|
50
|
+
return obj;
|
|
51
|
+
}
|
|
52
|
+
|
|
53
|
+
return filterObject(obj);
|
|
54
|
+
}
|
|
@@ -1,213 +1,213 @@
|
|
|
1
|
-
import Joi from 'joi';
|
|
2
|
-
|
|
3
|
-
import { IAgent as Agent } from '@sre/types/Agent.types';
|
|
4
|
-
import { ConnectorService } from '@sre/Core/ConnectorsService';
|
|
5
|
-
import { CacheConnector } from '@sre/MemoryManager/Cache.service/CacheConnector';
|
|
6
|
-
import { AccessCandidate } from '@sre/Security/AccessControl/AccessCandidate.class';
|
|
7
|
-
import { DEFAULT_MAX_TOKENS_FOR_LLM } from '@sre/constants';
|
|
8
|
-
import { TemplateString } from '@sre/helpers/TemplateString.helper';
|
|
9
|
-
import { encode } from 'gpt-tokenizer';
|
|
10
|
-
import { Component } from './Component.class';
|
|
11
|
-
import { JSONContent } from '@sre/helpers/JsonContent.helper';
|
|
12
|
-
import { LLMInference } from '@sre/LLMManager/LLM.inference';
|
|
13
|
-
import { TLLMMessageRole } from '@sre/types/LLM.types';
|
|
14
|
-
import { VaultHelper } from '@sre/Security/Vault.service/Vault.helper';
|
|
15
|
-
import path from 'path';
|
|
16
|
-
import config from '@sre/config';
|
|
17
|
-
import fs from 'fs/promises';
|
|
18
|
-
|
|
19
|
-
//const sessions = {};
|
|
20
|
-
let cacheConnector: CacheConnector;
|
|
21
|
-
function getCacheConnector() {
|
|
22
|
-
if (!cacheConnector) {
|
|
23
|
-
cacheConnector = ConnectorService.getCacheConnector();
|
|
24
|
-
}
|
|
25
|
-
return cacheConnector;
|
|
26
|
-
}
|
|
27
|
-
|
|
28
|
-
async function saveMessagesToSession(agentId, userId, conversationId, messages, ttl?) {
|
|
29
|
-
if (!userId && !conversationId) return;
|
|
30
|
-
const cacheConnector = getCacheConnector();
|
|
31
|
-
const conv_uid = `${agentId}:conv-u${userId}-c${conversationId}`;
|
|
32
|
-
|
|
33
|
-
cacheConnector.requester(AccessCandidate.agent(agentId)).set(conv_uid, JSON.stringify(messages), null, null, ttl);
|
|
34
|
-
}
|
|
35
|
-
|
|
36
|
-
async function readMessagesFromSession(agentId, userId, conversationId, maxTokens = DEFAULT_MAX_TOKENS_FOR_LLM) {
|
|
37
|
-
if (!userId && !conversationId) return [];
|
|
38
|
-
const cacheConnector = getCacheConnector();
|
|
39
|
-
|
|
40
|
-
const conv_uid = `${agentId}:conv-u${userId}-c${conversationId}`;
|
|
41
|
-
//read the last messages from a given session and ensure that the total chat tokens are within the limit
|
|
42
|
-
//start from the last message and keep adding messages until the total tokens exceed the limit
|
|
43
|
-
//if (!sessions[agentId]) return [];
|
|
44
|
-
//if (!sessions[agentId][conv_uid]) return [];
|
|
45
|
-
|
|
46
|
-
const sessionData = await cacheConnector.requester(AccessCandidate.agent(agentId))?.get(conv_uid);
|
|
47
|
-
|
|
48
|
-
let messages = sessionData ? JSONContent(sessionData).tryParse() : [];
|
|
49
|
-
|
|
50
|
-
//const messages = sessions[agentId][conv_uid].messages;
|
|
51
|
-
|
|
52
|
-
const filteredMessages: any[] = [];
|
|
53
|
-
|
|
54
|
-
let tokens = 0;
|
|
55
|
-
if (messages[0]?.role == 'system') {
|
|
56
|
-
const encoded = encode(messages[0]?.content);
|
|
57
|
-
const messageTokens = encoded.length + 3;
|
|
58
|
-
tokens += messageTokens;
|
|
59
|
-
}
|
|
60
|
-
|
|
61
|
-
for (let i = messages.length - 1; i >= 0; i--) {
|
|
62
|
-
if (messages[i].role == 'system') continue;
|
|
63
|
-
const message = messages[i];
|
|
64
|
-
const encoded = encode(message?.content);
|
|
65
|
-
const messageTokens = encoded.length + 3;
|
|
66
|
-
if (tokens + messageTokens > maxTokens) break;
|
|
67
|
-
filteredMessages.unshift(message);
|
|
68
|
-
tokens += messageTokens;
|
|
69
|
-
}
|
|
70
|
-
|
|
71
|
-
if (messages[0]?.role == 'system') filteredMessages.unshift(messages[0]);
|
|
72
|
-
|
|
73
|
-
return filteredMessages;
|
|
74
|
-
}
|
|
75
|
-
|
|
76
|
-
//TODO : update this implementation to use ConversationManager
|
|
77
|
-
// This will allow better context management and support for tool calls
|
|
78
|
-
export class LLMAssistant extends Component {
|
|
79
|
-
protected configSchema = Joi.object({
|
|
80
|
-
model: Joi.string().max(200).required(),
|
|
81
|
-
behavior: Joi.string().max(30000).allow('').label('Behavior'),
|
|
82
|
-
passthrough: Joi.boolean().optional().label('Passthrough'),
|
|
83
|
-
});
|
|
84
|
-
constructor() {
|
|
85
|
-
super();
|
|
86
|
-
}
|
|
87
|
-
init() {}
|
|
88
|
-
async process(input, config, agent: Agent) {
|
|
89
|
-
await super.process(input, config, agent);
|
|
90
|
-
const logger = this.createComponentLogger(agent, config);
|
|
91
|
-
try {
|
|
92
|
-
logger.debug('== LLM Assistant Log ==\n');
|
|
93
|
-
|
|
94
|
-
const passThrough: boolean = config.data.passthrough || false;
|
|
95
|
-
const model: string = config.data.model || 'echo';
|
|
96
|
-
const ttl = config.data.ttl || undefined;
|
|
97
|
-
let teamId = agent?.teamId;
|
|
98
|
-
|
|
99
|
-
const llmInference: LLMInference = await LLMInference.getInstance(model, AccessCandidate.agent(agent.id));
|
|
100
|
-
// if the llm is undefined, then it means we removed the model from our system
|
|
101
|
-
if (!llmInference.connector) {
|
|
102
|
-
return {
|
|
103
|
-
_error: `The model '${model}' is not available. Please try a different one.`,
|
|
104
|
-
_debug: logger.output,
|
|
105
|
-
};
|
|
106
|
-
}
|
|
107
|
-
|
|
108
|
-
const modelId = await agent.modelsProvider.getModelId(model);
|
|
109
|
-
logger.debug(` Model : ${modelId || model}`);
|
|
110
|
-
|
|
111
|
-
const userInput = input.UserInput;
|
|
112
|
-
const userId = input.UserId;
|
|
113
|
-
const conversationId = input.ConversationId;
|
|
114
|
-
|
|
115
|
-
let behavior = TemplateString(config.data.behavior).parse(input).result;
|
|
116
|
-
logger.debug(`[Parsed Behavior] \n${behavior}\n\n`);
|
|
117
|
-
|
|
118
|
-
//#region get max tokens
|
|
119
|
-
let maxTokens = 2048;
|
|
120
|
-
|
|
121
|
-
const isStandardLLM = await agent.modelsProvider.isStandardLLM(model);
|
|
122
|
-
const hasKey = true; //TODO : check if the user has a key
|
|
123
|
-
//const modelInfo = await agent.modelsProvider.getModelInfo(model, hasKey);
|
|
124
|
-
maxTokens = await agent.modelsProvider.getMaxCompletionTokens(model, hasKey);
|
|
125
|
-
|
|
126
|
-
// if (isStandardLLM) {
|
|
127
|
-
// const provider = LLMRegistry.getProvider(model);
|
|
128
|
-
// const apiKey = await VaultHelper.getAgentKey(provider, agent?.id);
|
|
129
|
-
// maxTokens = LLMRegistry.getMaxCompletionTokens(model, !!apiKey);
|
|
130
|
-
// } else {
|
|
131
|
-
// const team = AccessCandidate.team(teamId);
|
|
132
|
-
// const customLLMRegistry = await CustomLLMRegistry.getInstance(team);
|
|
133
|
-
// maxTokens = await customLLMRegistry.getMaxCompletionTokens(model);
|
|
134
|
-
// }
|
|
135
|
-
//#endregion get max tokens
|
|
136
|
-
|
|
137
|
-
const messages: any[] = await readMessagesFromSession(agent.id, userId, conversationId, Math.round(maxTokens / 2));
|
|
138
|
-
|
|
139
|
-
messages.push({ role: TLLMMessageRole.User, content: userInput });
|
|
140
|
-
|
|
141
|
-
if (messages[0]?.role != TLLMMessageRole.System) {
|
|
142
|
-
messages.unshift({ role: TLLMMessageRole.System, content: behavior });
|
|
143
|
-
}
|
|
144
|
-
|
|
145
|
-
const customParams = {
|
|
146
|
-
messages,
|
|
147
|
-
};
|
|
148
|
-
|
|
149
|
-
let response: any;
|
|
150
|
-
if (passThrough) {
|
|
151
|
-
const contentPromise = new Promise(async (resolve, reject) => {
|
|
152
|
-
let _content = '';
|
|
153
|
-
const eventEmitter: any = await llmInference
|
|
154
|
-
.promptStream({
|
|
155
|
-
contextWindow: messages,
|
|
156
|
-
params: { ...config, model, agentId: agent.id },
|
|
157
|
-
})
|
|
158
|
-
.catch((error) => {
|
|
159
|
-
console.error('Error on promptStream: ', error);
|
|
160
|
-
reject(error);
|
|
161
|
-
});
|
|
162
|
-
eventEmitter.on('content', (content) => {
|
|
163
|
-
if (typeof agent.callback === 'function') {
|
|
164
|
-
agent.callback({ content });
|
|
165
|
-
}
|
|
166
|
-
agent.sse.send('llm/passthrough/content', content);
|
|
167
|
-
_content += content;
|
|
168
|
-
});
|
|
169
|
-
eventEmitter.on('thinking', (thinking) => {
|
|
170
|
-
if (typeof agent.callback === 'function') {
|
|
171
|
-
agent.callback({ thinking });
|
|
172
|
-
}
|
|
173
|
-
agent.sse.send('llm/passthrough/thinking', thinking);
|
|
174
|
-
});
|
|
175
|
-
eventEmitter.on('end', () => {
|
|
176
|
-
console.log('end');
|
|
177
|
-
resolve(_content);
|
|
178
|
-
});
|
|
179
|
-
});
|
|
180
|
-
response = await contentPromise;
|
|
181
|
-
} else {
|
|
182
|
-
response = await llmInference
|
|
183
|
-
.prompt({ contextWindow: messages, params: { ...config, agentId: agent.id } })
|
|
184
|
-
.catch((error) => ({ error: error }));
|
|
185
|
-
}
|
|
186
|
-
|
|
187
|
-
// in case we have the response but it's empty string, undefined or null
|
|
188
|
-
if (!response) {
|
|
189
|
-
return { _error: ' LLM Error = Empty Response!', _debug: logger.output };
|
|
190
|
-
}
|
|
191
|
-
|
|
192
|
-
if (response?.error) {
|
|
193
|
-
const error = response?.error + ' ' + (response?.details || '');
|
|
194
|
-
logger.error(` LLM Error=`, error);
|
|
195
|
-
|
|
196
|
-
return { Response: response?.data, _error: error, _debug: logger.output };
|
|
197
|
-
}
|
|
198
|
-
|
|
199
|
-
messages.push({ role: 'assistant', content: response });
|
|
200
|
-
saveMessagesToSession(agent.id, userId, conversationId, messages, ttl);
|
|
201
|
-
|
|
202
|
-
logger.debug(' Response \n', response);
|
|
203
|
-
|
|
204
|
-
const result = { Response: response };
|
|
205
|
-
|
|
206
|
-
result['_debug'] = logger.output;
|
|
207
|
-
|
|
208
|
-
return result;
|
|
209
|
-
} catch (error) {
|
|
210
|
-
return { _error: error.message, _debug: logger.output };
|
|
211
|
-
}
|
|
212
|
-
}
|
|
213
|
-
}
|
|
1
|
+
import Joi from 'joi';
|
|
2
|
+
|
|
3
|
+
import { IAgent as Agent } from '@sre/types/Agent.types';
|
|
4
|
+
import { ConnectorService } from '@sre/Core/ConnectorsService';
|
|
5
|
+
import { CacheConnector } from '@sre/MemoryManager/Cache.service/CacheConnector';
|
|
6
|
+
import { AccessCandidate } from '@sre/Security/AccessControl/AccessCandidate.class';
|
|
7
|
+
import { DEFAULT_MAX_TOKENS_FOR_LLM } from '@sre/constants';
|
|
8
|
+
import { TemplateString } from '@sre/helpers/TemplateString.helper';
|
|
9
|
+
import { encode } from 'gpt-tokenizer';
|
|
10
|
+
import { Component } from './Component.class';
|
|
11
|
+
import { JSONContent } from '@sre/helpers/JsonContent.helper';
|
|
12
|
+
import { LLMInference } from '@sre/LLMManager/LLM.inference';
|
|
13
|
+
import { TLLMMessageRole } from '@sre/types/LLM.types';
|
|
14
|
+
import { VaultHelper } from '@sre/Security/Vault.service/Vault.helper';
|
|
15
|
+
import path from 'path';
|
|
16
|
+
import config from '@sre/config';
|
|
17
|
+
import fs from 'fs/promises';
|
|
18
|
+
|
|
19
|
+
//const sessions = {};
|
|
20
|
+
let cacheConnector: CacheConnector;
|
|
21
|
+
function getCacheConnector() {
|
|
22
|
+
if (!cacheConnector) {
|
|
23
|
+
cacheConnector = ConnectorService.getCacheConnector();
|
|
24
|
+
}
|
|
25
|
+
return cacheConnector;
|
|
26
|
+
}
|
|
27
|
+
|
|
28
|
+
async function saveMessagesToSession(agentId, userId, conversationId, messages, ttl?) {
|
|
29
|
+
if (!userId && !conversationId) return;
|
|
30
|
+
const cacheConnector = getCacheConnector();
|
|
31
|
+
const conv_uid = `${agentId}:conv-u${userId}-c${conversationId}`;
|
|
32
|
+
|
|
33
|
+
cacheConnector.requester(AccessCandidate.agent(agentId)).set(conv_uid, JSON.stringify(messages), null, null, ttl);
|
|
34
|
+
}
|
|
35
|
+
|
|
36
|
+
async function readMessagesFromSession(agentId, userId, conversationId, maxTokens = DEFAULT_MAX_TOKENS_FOR_LLM) {
|
|
37
|
+
if (!userId && !conversationId) return [];
|
|
38
|
+
const cacheConnector = getCacheConnector();
|
|
39
|
+
|
|
40
|
+
const conv_uid = `${agentId}:conv-u${userId}-c${conversationId}`;
|
|
41
|
+
//read the last messages from a given session and ensure that the total chat tokens are within the limit
|
|
42
|
+
//start from the last message and keep adding messages until the total tokens exceed the limit
|
|
43
|
+
//if (!sessions[agentId]) return [];
|
|
44
|
+
//if (!sessions[agentId][conv_uid]) return [];
|
|
45
|
+
|
|
46
|
+
const sessionData = await cacheConnector.requester(AccessCandidate.agent(agentId))?.get(conv_uid);
|
|
47
|
+
|
|
48
|
+
let messages = sessionData ? JSONContent(sessionData).tryParse() : [];
|
|
49
|
+
|
|
50
|
+
//const messages = sessions[agentId][conv_uid].messages;
|
|
51
|
+
|
|
52
|
+
const filteredMessages: any[] = [];
|
|
53
|
+
|
|
54
|
+
let tokens = 0;
|
|
55
|
+
if (messages[0]?.role == 'system') {
|
|
56
|
+
const encoded = encode(messages[0]?.content);
|
|
57
|
+
const messageTokens = encoded.length + 3;
|
|
58
|
+
tokens += messageTokens;
|
|
59
|
+
}
|
|
60
|
+
|
|
61
|
+
for (let i = messages.length - 1; i >= 0; i--) {
|
|
62
|
+
if (messages[i].role == 'system') continue;
|
|
63
|
+
const message = messages[i];
|
|
64
|
+
const encoded = encode(message?.content);
|
|
65
|
+
const messageTokens = encoded.length + 3;
|
|
66
|
+
if (tokens + messageTokens > maxTokens) break;
|
|
67
|
+
filteredMessages.unshift(message);
|
|
68
|
+
tokens += messageTokens;
|
|
69
|
+
}
|
|
70
|
+
|
|
71
|
+
if (messages[0]?.role == 'system') filteredMessages.unshift(messages[0]);
|
|
72
|
+
|
|
73
|
+
return filteredMessages;
|
|
74
|
+
}
|
|
75
|
+
|
|
76
|
+
//TODO : update this implementation to use ConversationManager
|
|
77
|
+
// This will allow better context management and support for tool calls
|
|
78
|
+
export class LLMAssistant extends Component {
|
|
79
|
+
protected configSchema = Joi.object({
|
|
80
|
+
model: Joi.string().max(200).required(),
|
|
81
|
+
behavior: Joi.string().max(30000).allow('').label('Behavior'),
|
|
82
|
+
passthrough: Joi.boolean().optional().label('Passthrough'),
|
|
83
|
+
});
|
|
84
|
+
constructor() {
|
|
85
|
+
super();
|
|
86
|
+
}
|
|
87
|
+
init() {}
|
|
88
|
+
async process(input, config, agent: Agent) {
|
|
89
|
+
await super.process(input, config, agent);
|
|
90
|
+
const logger = this.createComponentLogger(agent, config);
|
|
91
|
+
try {
|
|
92
|
+
logger.debug('== LLM Assistant Log ==\n');
|
|
93
|
+
|
|
94
|
+
const passThrough: boolean = config.data.passthrough || false;
|
|
95
|
+
const model: string = config.data.model || 'echo';
|
|
96
|
+
const ttl = config.data.ttl || undefined;
|
|
97
|
+
let teamId = agent?.teamId;
|
|
98
|
+
|
|
99
|
+
const llmInference: LLMInference = await LLMInference.getInstance(model, AccessCandidate.agent(agent.id));
|
|
100
|
+
// if the llm is undefined, then it means we removed the model from our system
|
|
101
|
+
if (!llmInference.connector) {
|
|
102
|
+
return {
|
|
103
|
+
_error: `The model '${model}' is not available. Please try a different one.`,
|
|
104
|
+
_debug: logger.output,
|
|
105
|
+
};
|
|
106
|
+
}
|
|
107
|
+
|
|
108
|
+
const modelId = await agent.modelsProvider.getModelId(model);
|
|
109
|
+
logger.debug(` Model : ${modelId || model}`);
|
|
110
|
+
|
|
111
|
+
const userInput = input.UserInput;
|
|
112
|
+
const userId = input.UserId;
|
|
113
|
+
const conversationId = input.ConversationId;
|
|
114
|
+
|
|
115
|
+
let behavior = TemplateString(config.data.behavior).parse(input).result;
|
|
116
|
+
logger.debug(`[Parsed Behavior] \n${behavior}\n\n`);
|
|
117
|
+
|
|
118
|
+
//#region get max tokens
|
|
119
|
+
let maxTokens = 2048;
|
|
120
|
+
|
|
121
|
+
const isStandardLLM = await agent.modelsProvider.isStandardLLM(model);
|
|
122
|
+
const hasKey = true; //TODO : check if the user has a key
|
|
123
|
+
//const modelInfo = await agent.modelsProvider.getModelInfo(model, hasKey);
|
|
124
|
+
maxTokens = await agent.modelsProvider.getMaxCompletionTokens(model, hasKey);
|
|
125
|
+
|
|
126
|
+
// if (isStandardLLM) {
|
|
127
|
+
// const provider = LLMRegistry.getProvider(model);
|
|
128
|
+
// const apiKey = await VaultHelper.getAgentKey(provider, agent?.id);
|
|
129
|
+
// maxTokens = LLMRegistry.getMaxCompletionTokens(model, !!apiKey);
|
|
130
|
+
// } else {
|
|
131
|
+
// const team = AccessCandidate.team(teamId);
|
|
132
|
+
// const customLLMRegistry = await CustomLLMRegistry.getInstance(team);
|
|
133
|
+
// maxTokens = await customLLMRegistry.getMaxCompletionTokens(model);
|
|
134
|
+
// }
|
|
135
|
+
//#endregion get max tokens
|
|
136
|
+
|
|
137
|
+
const messages: any[] = await readMessagesFromSession(agent.id, userId, conversationId, Math.round(maxTokens / 2));
|
|
138
|
+
|
|
139
|
+
messages.push({ role: TLLMMessageRole.User, content: userInput });
|
|
140
|
+
|
|
141
|
+
if (messages[0]?.role != TLLMMessageRole.System) {
|
|
142
|
+
messages.unshift({ role: TLLMMessageRole.System, content: behavior });
|
|
143
|
+
}
|
|
144
|
+
|
|
145
|
+
const customParams = {
|
|
146
|
+
messages,
|
|
147
|
+
};
|
|
148
|
+
|
|
149
|
+
let response: any;
|
|
150
|
+
if (passThrough) {
|
|
151
|
+
const contentPromise = new Promise(async (resolve, reject) => {
|
|
152
|
+
let _content = '';
|
|
153
|
+
const eventEmitter: any = await llmInference
|
|
154
|
+
.promptStream({
|
|
155
|
+
contextWindow: messages,
|
|
156
|
+
params: { ...config, model, agentId: agent.id },
|
|
157
|
+
})
|
|
158
|
+
.catch((error) => {
|
|
159
|
+
console.error('Error on promptStream: ', error);
|
|
160
|
+
reject(error);
|
|
161
|
+
});
|
|
162
|
+
eventEmitter.on('content', (content) => {
|
|
163
|
+
if (typeof agent.callback === 'function') {
|
|
164
|
+
agent.callback({ content });
|
|
165
|
+
}
|
|
166
|
+
agent.sse.send('llm/passthrough/content', content);
|
|
167
|
+
_content += content;
|
|
168
|
+
});
|
|
169
|
+
eventEmitter.on('thinking', (thinking) => {
|
|
170
|
+
if (typeof agent.callback === 'function') {
|
|
171
|
+
agent.callback({ thinking });
|
|
172
|
+
}
|
|
173
|
+
agent.sse.send('llm/passthrough/thinking', thinking);
|
|
174
|
+
});
|
|
175
|
+
eventEmitter.on('end', () => {
|
|
176
|
+
console.log('end');
|
|
177
|
+
resolve(_content);
|
|
178
|
+
});
|
|
179
|
+
});
|
|
180
|
+
response = await contentPromise;
|
|
181
|
+
} else {
|
|
182
|
+
response = await llmInference
|
|
183
|
+
.prompt({ contextWindow: messages, params: { ...config, agentId: agent.id } })
|
|
184
|
+
.catch((error) => ({ error: error }));
|
|
185
|
+
}
|
|
186
|
+
|
|
187
|
+
// in case we have the response but it's empty string, undefined or null
|
|
188
|
+
if (!response) {
|
|
189
|
+
return { _error: ' LLM Error = Empty Response!', _debug: logger.output };
|
|
190
|
+
}
|
|
191
|
+
|
|
192
|
+
if (response?.error) {
|
|
193
|
+
const error = response?.error + ' ' + (response?.details || '');
|
|
194
|
+
logger.error(` LLM Error=`, error);
|
|
195
|
+
|
|
196
|
+
return { Response: response?.data, _error: error, _debug: logger.output };
|
|
197
|
+
}
|
|
198
|
+
|
|
199
|
+
messages.push({ role: 'assistant', content: response });
|
|
200
|
+
saveMessagesToSession(agent.id, userId, conversationId, messages, ttl);
|
|
201
|
+
|
|
202
|
+
logger.debug(' Response \n', response);
|
|
203
|
+
|
|
204
|
+
const result = { Response: response };
|
|
205
|
+
|
|
206
|
+
result['_debug'] = logger.output;
|
|
207
|
+
|
|
208
|
+
return result;
|
|
209
|
+
} catch (error) {
|
|
210
|
+
return { _error: error.message, _debug: logger.output };
|
|
211
|
+
}
|
|
212
|
+
}
|
|
213
|
+
}
|
|
@@ -1,28 +1,28 @@
|
|
|
1
|
-
import { IAgent as Agent } from '@sre/types/Agent.types';
|
|
2
|
-
import { Component } from './Component.class';
|
|
3
|
-
|
|
4
|
-
export class LogicAND extends Component {
|
|
5
|
-
constructor() {
|
|
6
|
-
super();
|
|
7
|
-
}
|
|
8
|
-
init() {}
|
|
9
|
-
async process(input, config, agent: Agent) {
|
|
10
|
-
await super.process(input, config, agent);
|
|
11
|
-
const logger = this.createComponentLogger(agent, config);
|
|
12
|
-
const result: any = { Output: true };
|
|
13
|
-
|
|
14
|
-
for (let cfgInput of config.inputs) {
|
|
15
|
-
// check if all inputs are set (expected inputs are in "config.inputs" actual inputs are in "input")
|
|
16
|
-
if (!input[cfgInput.name]) {
|
|
17
|
-
result.Output = undefined;
|
|
18
|
-
break;
|
|
19
|
-
}
|
|
20
|
-
}
|
|
21
|
-
|
|
22
|
-
result.Verified = result.Output !== undefined;
|
|
23
|
-
result.Unverified = !result.Verified;
|
|
24
|
-
if (!result.Verified) delete result.Verified;
|
|
25
|
-
if (!result.Unverified) delete result.Unverified;
|
|
26
|
-
return result;
|
|
27
|
-
}
|
|
28
|
-
}
|
|
1
|
+
import { IAgent as Agent } from '@sre/types/Agent.types';
|
|
2
|
+
import { Component } from './Component.class';
|
|
3
|
+
|
|
4
|
+
export class LogicAND extends Component {
|
|
5
|
+
constructor() {
|
|
6
|
+
super();
|
|
7
|
+
}
|
|
8
|
+
init() {}
|
|
9
|
+
async process(input, config, agent: Agent) {
|
|
10
|
+
await super.process(input, config, agent);
|
|
11
|
+
const logger = this.createComponentLogger(agent, config);
|
|
12
|
+
const result: any = { Output: true };
|
|
13
|
+
|
|
14
|
+
for (let cfgInput of config.inputs) {
|
|
15
|
+
// check if all inputs are set (expected inputs are in "config.inputs" actual inputs are in "input")
|
|
16
|
+
if (!input[cfgInput.name]) {
|
|
17
|
+
result.Output = undefined;
|
|
18
|
+
break;
|
|
19
|
+
}
|
|
20
|
+
}
|
|
21
|
+
|
|
22
|
+
result.Verified = result.Output !== undefined;
|
|
23
|
+
result.Unverified = !result.Verified;
|
|
24
|
+
if (!result.Verified) delete result.Verified;
|
|
25
|
+
if (!result.Unverified) delete result.Unverified;
|
|
26
|
+
return result;
|
|
27
|
+
}
|
|
28
|
+
}
|