@smythos/sre 1.5.46 → 1.5.50

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (224) hide show
  1. package/CHANGELOG +98 -90
  2. package/LICENSE +18 -18
  3. package/README.md +135 -135
  4. package/dist/bundle-analysis-lazy.html +4949 -0
  5. package/dist/bundle-analysis.html +4949 -0
  6. package/dist/index.js +3 -3
  7. package/dist/index.js.map +1 -1
  8. package/dist/types/Components/MCPClient.class.d.ts +1 -0
  9. package/dist/types/subsystems/LLMManager/LLM.service/connectors/openai/OpenAIConnector.class.d.ts +1 -6
  10. package/dist/types/utils/package-manager.utils.d.ts +26 -0
  11. package/package.json +1 -1
  12. package/src/Components/APICall/APICall.class.ts +156 -156
  13. package/src/Components/APICall/AccessTokenManager.ts +130 -130
  14. package/src/Components/APICall/ArrayBufferResponse.helper.ts +58 -58
  15. package/src/Components/APICall/OAuth.helper.ts +294 -294
  16. package/src/Components/APICall/mimeTypeCategories.ts +46 -46
  17. package/src/Components/APICall/parseData.ts +167 -167
  18. package/src/Components/APICall/parseHeaders.ts +41 -41
  19. package/src/Components/APICall/parseProxy.ts +68 -68
  20. package/src/Components/APICall/parseUrl.ts +91 -91
  21. package/src/Components/APIEndpoint.class.ts +234 -234
  22. package/src/Components/APIOutput.class.ts +58 -58
  23. package/src/Components/AgentPlugin.class.ts +102 -102
  24. package/src/Components/Async.class.ts +155 -155
  25. package/src/Components/Await.class.ts +90 -90
  26. package/src/Components/Classifier.class.ts +158 -158
  27. package/src/Components/Component.class.ts +132 -132
  28. package/src/Components/ComponentHost.class.ts +38 -38
  29. package/src/Components/DataSourceCleaner.class.ts +92 -92
  30. package/src/Components/DataSourceIndexer.class.ts +181 -181
  31. package/src/Components/DataSourceLookup.class.ts +161 -161
  32. package/src/Components/ECMASandbox.class.ts +71 -71
  33. package/src/Components/FEncDec.class.ts +29 -29
  34. package/src/Components/FHash.class.ts +33 -33
  35. package/src/Components/FSign.class.ts +80 -80
  36. package/src/Components/FSleep.class.ts +25 -25
  37. package/src/Components/FTimestamp.class.ts +25 -25
  38. package/src/Components/FileStore.class.ts +78 -78
  39. package/src/Components/ForEach.class.ts +97 -97
  40. package/src/Components/GPTPlugin.class.ts +70 -70
  41. package/src/Components/GenAILLM.class.ts +586 -586
  42. package/src/Components/HuggingFace.class.ts +314 -314
  43. package/src/Components/Image/imageSettings.config.ts +70 -70
  44. package/src/Components/ImageGenerator.class.ts +502 -502
  45. package/src/Components/JSONFilter.class.ts +54 -54
  46. package/src/Components/LLMAssistant.class.ts +213 -213
  47. package/src/Components/LogicAND.class.ts +28 -28
  48. package/src/Components/LogicAtLeast.class.ts +85 -85
  49. package/src/Components/LogicAtMost.class.ts +86 -86
  50. package/src/Components/LogicOR.class.ts +29 -29
  51. package/src/Components/LogicXOR.class.ts +34 -34
  52. package/src/Components/MCPClient.class.ts +138 -112
  53. package/src/Components/MemoryDeleteKeyVal.class.ts +70 -70
  54. package/src/Components/MemoryReadKeyVal.class.ts +66 -66
  55. package/src/Components/MemoryWriteKeyVal.class.ts +62 -62
  56. package/src/Components/MemoryWriteObject.class.ts +97 -97
  57. package/src/Components/MultimodalLLM.class.ts +128 -128
  58. package/src/Components/OpenAPI.class.ts +72 -72
  59. package/src/Components/PromptGenerator.class.ts +122 -122
  60. package/src/Components/ScrapflyWebScrape.class.ts +159 -159
  61. package/src/Components/ServerlessCode.class.ts +123 -123
  62. package/src/Components/TavilyWebSearch.class.ts +98 -98
  63. package/src/Components/VisionLLM.class.ts +104 -104
  64. package/src/Components/ZapierAction.class.ts +127 -127
  65. package/src/Components/index.ts +97 -97
  66. package/src/Core/AgentProcess.helper.ts +240 -240
  67. package/src/Core/Connector.class.ts +123 -123
  68. package/src/Core/ConnectorsService.ts +197 -197
  69. package/src/Core/DummyConnector.ts +49 -49
  70. package/src/Core/HookService.ts +105 -105
  71. package/src/Core/SmythRuntime.class.ts +235 -235
  72. package/src/Core/SystemEvents.ts +16 -16
  73. package/src/Core/boot.ts +56 -56
  74. package/src/config.ts +15 -15
  75. package/src/constants.ts +126 -126
  76. package/src/data/hugging-face.params.json +579 -579
  77. package/src/helpers/AWSLambdaCode.helper.ts +587 -587
  78. package/src/helpers/BinaryInput.helper.ts +331 -331
  79. package/src/helpers/Conversation.helper.ts +1119 -1119
  80. package/src/helpers/ECMASandbox.helper.ts +54 -54
  81. package/src/helpers/JsonContent.helper.ts +97 -97
  82. package/src/helpers/LocalCache.helper.ts +97 -97
  83. package/src/helpers/Log.helper.ts +274 -274
  84. package/src/helpers/OpenApiParser.helper.ts +150 -150
  85. package/src/helpers/S3Cache.helper.ts +147 -147
  86. package/src/helpers/SmythURI.helper.ts +5 -5
  87. package/src/helpers/Sysconfig.helper.ts +77 -77
  88. package/src/helpers/TemplateString.helper.ts +243 -243
  89. package/src/helpers/TypeChecker.helper.ts +329 -329
  90. package/src/index.ts +3 -3
  91. package/src/index.ts.bak +3 -3
  92. package/src/subsystems/AgentManager/Agent.class.ts +1114 -1114
  93. package/src/subsystems/AgentManager/Agent.helper.ts +3 -3
  94. package/src/subsystems/AgentManager/AgentData.service/AgentDataConnector.ts +230 -230
  95. package/src/subsystems/AgentManager/AgentData.service/connectors/CLIAgentDataConnector.class.ts +66 -66
  96. package/src/subsystems/AgentManager/AgentData.service/connectors/LocalAgentDataConnector.class.ts +142 -142
  97. package/src/subsystems/AgentManager/AgentData.service/connectors/NullAgentData.class.ts +39 -39
  98. package/src/subsystems/AgentManager/AgentData.service/index.ts +18 -18
  99. package/src/subsystems/AgentManager/AgentLogger.class.ts +297 -297
  100. package/src/subsystems/AgentManager/AgentRequest.class.ts +51 -51
  101. package/src/subsystems/AgentManager/AgentRuntime.class.ts +559 -559
  102. package/src/subsystems/AgentManager/AgentSSE.class.ts +101 -101
  103. package/src/subsystems/AgentManager/AgentSettings.class.ts +52 -52
  104. package/src/subsystems/AgentManager/Component.service/ComponentConnector.ts +32 -32
  105. package/src/subsystems/AgentManager/Component.service/connectors/LocalComponentConnector.class.ts +60 -60
  106. package/src/subsystems/AgentManager/Component.service/index.ts +11 -11
  107. package/src/subsystems/AgentManager/EmbodimentSettings.class.ts +47 -47
  108. package/src/subsystems/AgentManager/ForkedAgent.class.ts +154 -154
  109. package/src/subsystems/AgentManager/OSResourceMonitor.ts +77 -77
  110. package/src/subsystems/ComputeManager/Code.service/CodeConnector.ts +98 -98
  111. package/src/subsystems/ComputeManager/Code.service/connectors/AWSLambdaCode.class.ts +172 -172
  112. package/src/subsystems/ComputeManager/Code.service/connectors/ECMASandbox.class.ts +131 -131
  113. package/src/subsystems/ComputeManager/Code.service/index.ts +13 -13
  114. package/src/subsystems/IO/CLI.service/CLIConnector.ts +47 -47
  115. package/src/subsystems/IO/CLI.service/index.ts +9 -9
  116. package/src/subsystems/IO/Log.service/LogConnector.ts +32 -32
  117. package/src/subsystems/IO/Log.service/connectors/ConsoleLog.class.ts +28 -28
  118. package/src/subsystems/IO/Log.service/index.ts +13 -13
  119. package/src/subsystems/IO/NKV.service/NKVConnector.ts +43 -43
  120. package/src/subsystems/IO/NKV.service/connectors/NKVLocalStorage.class.ts +234 -234
  121. package/src/subsystems/IO/NKV.service/connectors/NKVRAM.class.ts +204 -204
  122. package/src/subsystems/IO/NKV.service/connectors/NKVRedis.class.ts +182 -182
  123. package/src/subsystems/IO/NKV.service/index.ts +14 -14
  124. package/src/subsystems/IO/Router.service/RouterConnector.ts +21 -21
  125. package/src/subsystems/IO/Router.service/connectors/ExpressRouter.class.ts +48 -48
  126. package/src/subsystems/IO/Router.service/connectors/NullRouter.class.ts +40 -40
  127. package/src/subsystems/IO/Router.service/index.ts +11 -11
  128. package/src/subsystems/IO/Storage.service/SmythFS.class.ts +489 -489
  129. package/src/subsystems/IO/Storage.service/StorageConnector.ts +66 -66
  130. package/src/subsystems/IO/Storage.service/connectors/LocalStorage.class.ts +327 -327
  131. package/src/subsystems/IO/Storage.service/connectors/S3Storage.class.ts +482 -482
  132. package/src/subsystems/IO/Storage.service/index.ts +13 -13
  133. package/src/subsystems/IO/VectorDB.service/VectorDBConnector.ts +108 -108
  134. package/src/subsystems/IO/VectorDB.service/connectors/MilvusVectorDB.class.ts +454 -454
  135. package/src/subsystems/IO/VectorDB.service/connectors/PineconeVectorDB.class.ts +384 -384
  136. package/src/subsystems/IO/VectorDB.service/connectors/RAMVecrtorDB.class.ts +421 -421
  137. package/src/subsystems/IO/VectorDB.service/embed/BaseEmbedding.ts +107 -107
  138. package/src/subsystems/IO/VectorDB.service/embed/OpenAIEmbedding.ts +109 -109
  139. package/src/subsystems/IO/VectorDB.service/embed/index.ts +21 -21
  140. package/src/subsystems/IO/VectorDB.service/index.ts +14 -14
  141. package/src/subsystems/LLMManager/LLM.helper.ts +251 -251
  142. package/src/subsystems/LLMManager/LLM.inference.ts +339 -339
  143. package/src/subsystems/LLMManager/LLM.service/LLMConnector.ts +489 -489
  144. package/src/subsystems/LLMManager/LLM.service/LLMCredentials.helper.ts +171 -171
  145. package/src/subsystems/LLMManager/LLM.service/connectors/Anthropic.class.ts +659 -659
  146. package/src/subsystems/LLMManager/LLM.service/connectors/Bedrock.class.ts +400 -400
  147. package/src/subsystems/LLMManager/LLM.service/connectors/Echo.class.ts +77 -77
  148. package/src/subsystems/LLMManager/LLM.service/connectors/GoogleAI.class.ts +757 -757
  149. package/src/subsystems/LLMManager/LLM.service/connectors/Groq.class.ts +304 -304
  150. package/src/subsystems/LLMManager/LLM.service/connectors/Perplexity.class.ts +250 -250
  151. package/src/subsystems/LLMManager/LLM.service/connectors/VertexAI.class.ts +423 -423
  152. package/src/subsystems/LLMManager/LLM.service/connectors/openai/OpenAIConnector.class.ts +488 -488
  153. package/src/subsystems/LLMManager/LLM.service/connectors/openai/apiInterfaces/ChatCompletionsApiInterface.ts +524 -524
  154. package/src/subsystems/LLMManager/LLM.service/connectors/openai/apiInterfaces/OpenAIApiInterface.ts +100 -100
  155. package/src/subsystems/LLMManager/LLM.service/connectors/openai/apiInterfaces/OpenAIApiInterfaceFactory.ts +81 -81
  156. package/src/subsystems/LLMManager/LLM.service/connectors/openai/apiInterfaces/ResponsesApiInterface.ts +1145 -1145
  157. package/src/subsystems/LLMManager/LLM.service/connectors/openai/apiInterfaces/constants.ts +13 -13
  158. package/src/subsystems/LLMManager/LLM.service/connectors/openai/apiInterfaces/index.ts +4 -4
  159. package/src/subsystems/LLMManager/LLM.service/connectors/openai/apiInterfaces/utils.ts +11 -11
  160. package/src/subsystems/LLMManager/LLM.service/connectors/openai/types.ts +32 -32
  161. package/src/subsystems/LLMManager/LLM.service/connectors/xAI.class.ts +471 -471
  162. package/src/subsystems/LLMManager/LLM.service/index.ts +44 -44
  163. package/src/subsystems/LLMManager/ModelsProvider.service/ModelsProviderConnector.ts +300 -300
  164. package/src/subsystems/LLMManager/ModelsProvider.service/connectors/JSONModelsProvider.class.ts +252 -252
  165. package/src/subsystems/LLMManager/ModelsProvider.service/index.ts +11 -11
  166. package/src/subsystems/LLMManager/custom-models.ts +854 -854
  167. package/src/subsystems/LLMManager/models.ts +2540 -2540
  168. package/src/subsystems/LLMManager/paramMappings.ts +69 -69
  169. package/src/subsystems/MemoryManager/Cache.service/CacheConnector.ts +86 -86
  170. package/src/subsystems/MemoryManager/Cache.service/connectors/LocalStorageCache.class.ts +297 -297
  171. package/src/subsystems/MemoryManager/Cache.service/connectors/RAMCache.class.ts +201 -201
  172. package/src/subsystems/MemoryManager/Cache.service/connectors/RedisCache.class.ts +252 -252
  173. package/src/subsystems/MemoryManager/Cache.service/connectors/S3Cache.class.ts +373 -373
  174. package/src/subsystems/MemoryManager/Cache.service/index.ts +15 -15
  175. package/src/subsystems/MemoryManager/LLMCache.ts +72 -72
  176. package/src/subsystems/MemoryManager/LLMContext.ts +124 -124
  177. package/src/subsystems/MemoryManager/LLMMemory.service/LLMMemoryConnector.ts +26 -26
  178. package/src/subsystems/MemoryManager/RuntimeContext.ts +266 -266
  179. package/src/subsystems/Security/AccessControl/ACL.class.ts +208 -208
  180. package/src/subsystems/Security/AccessControl/AccessCandidate.class.ts +82 -82
  181. package/src/subsystems/Security/AccessControl/AccessRequest.class.ts +52 -52
  182. package/src/subsystems/Security/Account.service/AccountConnector.ts +44 -44
  183. package/src/subsystems/Security/Account.service/connectors/AWSAccount.class.ts +76 -76
  184. package/src/subsystems/Security/Account.service/connectors/DummyAccount.class.ts +130 -130
  185. package/src/subsystems/Security/Account.service/connectors/JSONFileAccount.class.ts +159 -159
  186. package/src/subsystems/Security/Account.service/index.ts +14 -14
  187. package/src/subsystems/Security/Credentials.helper.ts +62 -62
  188. package/src/subsystems/Security/ManagedVault.service/ManagedVaultConnector.ts +38 -38
  189. package/src/subsystems/Security/ManagedVault.service/connectors/NullManagedVault.class.ts +53 -53
  190. package/src/subsystems/Security/ManagedVault.service/connectors/SecretManagerManagedVault.ts +154 -154
  191. package/src/subsystems/Security/ManagedVault.service/index.ts +12 -12
  192. package/src/subsystems/Security/SecureConnector.class.ts +110 -110
  193. package/src/subsystems/Security/Vault.service/Vault.helper.ts +30 -30
  194. package/src/subsystems/Security/Vault.service/VaultConnector.ts +29 -29
  195. package/src/subsystems/Security/Vault.service/connectors/HashicorpVault.class.ts +46 -46
  196. package/src/subsystems/Security/Vault.service/connectors/JSONFileVault.class.ts +221 -221
  197. package/src/subsystems/Security/Vault.service/connectors/NullVault.class.ts +54 -54
  198. package/src/subsystems/Security/Vault.service/connectors/SecretsManager.class.ts +140 -140
  199. package/src/subsystems/Security/Vault.service/index.ts +12 -12
  200. package/src/types/ACL.types.ts +104 -104
  201. package/src/types/AWS.types.ts +10 -10
  202. package/src/types/Agent.types.ts +61 -61
  203. package/src/types/AgentLogger.types.ts +17 -17
  204. package/src/types/Cache.types.ts +1 -1
  205. package/src/types/Common.types.ts +2 -2
  206. package/src/types/LLM.types.ts +496 -496
  207. package/src/types/Redis.types.ts +8 -8
  208. package/src/types/SRE.types.ts +64 -64
  209. package/src/types/Security.types.ts +14 -14
  210. package/src/types/Storage.types.ts +5 -5
  211. package/src/types/VectorDB.types.ts +86 -86
  212. package/src/utils/base64.utils.ts +275 -275
  213. package/src/utils/cli.utils.ts +68 -68
  214. package/src/utils/data.utils.ts +322 -322
  215. package/src/utils/date-time.utils.ts +22 -22
  216. package/src/utils/general.utils.ts +238 -238
  217. package/src/utils/index.ts +12 -12
  218. package/src/utils/lazy-client.ts +261 -261
  219. package/src/utils/numbers.utils.ts +13 -13
  220. package/src/utils/oauth.utils.ts +35 -35
  221. package/src/utils/string.utils.ts +414 -414
  222. package/src/utils/url.utils.ts +19 -19
  223. package/src/utils/validation.utils.ts +74 -74
  224. package/dist/types/subsystems/LLMManager/ModelsProvider.service/connectors/SmythModelsProvider.class.d.ts +0 -39
@@ -1,300 +1,300 @@
1
- import { ConnectorService } from '@sre/Core/ConnectorsService';
2
- import { AccessCandidate } from '@sre/Security/AccessControl/AccessCandidate.class';
3
- import { AccessRequest } from '@sre/Security/AccessControl/AccessRequest.class';
4
- import { ACL } from '@sre/Security/AccessControl/ACL.class';
5
- import { SecureConnector } from '@sre/Security/SecureConnector.class';
6
- import { IAccessCandidate } from '@sre/types/ACL.types';
7
- import { TCustomLLMModel, TLLMCredentials, TLLMModel, TLLMModelsList, TLLMProvider } from '@sre/types/LLM.types';
8
- import { customModels } from '../custom-models';
9
- import { LocalCache } from '@sre/helpers/LocalCache.helper';
10
-
11
- export interface IModelsProviderRequest {
12
- getModels(): Promise<any>;
13
- getCustomModels(): Promise<any>;
14
- getMaxContextTokens(model: string, hasAPIKey?: boolean): Promise<number>;
15
- addModels(models: TLLMModelsList): Promise<void>;
16
- getModelInfo(model: string | TLLMModel | TCustomLLMModel, hasAPIKey?: boolean): Promise<TLLMModel>;
17
- getModelId(model: string | TLLMModel | TCustomLLMModel): Promise<string>;
18
- getProvider(model: string | TLLMModel | TCustomLLMModel): Promise<string>;
19
- isStandardLLM(model: string | TLLMModel | TCustomLLMModel): Promise<boolean>;
20
- adjustMaxCompletionTokens(model: string | TLLMModel | TCustomLLMModel, maxCompletionTokens: number, hasAPIKey?: boolean): Promise<number>;
21
- getMaxContextTokens(model: string | TLLMModel | TCustomLLMModel, hasAPIKey?: boolean): Promise<number>;
22
- getMaxCompletionTokens(model: string | TLLMModel | TCustomLLMModel, hasAPIKey?: boolean): Promise<number>;
23
- validateTokensLimit({
24
- model,
25
- promptTokens,
26
- completionTokens,
27
- hasAPIKey,
28
- }: {
29
- model: TLLMModel | TCustomLLMModel;
30
- promptTokens: number;
31
- completionTokens: number;
32
- hasAPIKey?: boolean;
33
- }): Promise<void>;
34
- }
35
-
36
- export abstract class ModelsProviderConnector extends SecureConnector {
37
- protected static localCache = new LocalCache();
38
- public abstract getResourceACL(resourceId: string, candidate: IAccessCandidate): Promise<ACL>;
39
- public abstract getModels(acRequest: AccessRequest): Promise<TLLMModelsList>;
40
- public abstract addModels(acRequest: AccessRequest, models: TLLMModelsList): Promise<void>;
41
-
42
- public requester(candidate: AccessCandidate): IModelsProviderRequest {
43
- const cacheKey = `ModelsProviderConnector:${candidate.toString()}`;
44
- if (ModelsProviderConnector.localCache.has(cacheKey)) {
45
- //update the TTL every time the requester is called
46
- return ModelsProviderConnector.localCache.get(cacheKey, 10 * 60 * 1000) as IModelsProviderRequest;
47
- }
48
-
49
- let teamModels = null;
50
- let customModels = null;
51
-
52
- const loadTeamModels = async () => {
53
- if (!teamModels) {
54
- try {
55
- const builtinmodels = await this.getModels(candidate.readRequest);
56
- customModels = await this.getCustomModels(candidate);
57
- teamModels = { ...builtinmodels, ...customModels };
58
- } catch (error) {
59
- return null;
60
- }
61
- }
62
- //Workaround : non-blocking auto-refresh of team models
63
- //this will force team models to refresh for the next request
64
- //TODO: we need a more elegant cache invalidation mechanism, and only refresh the team models if the custom models have changed
65
- setImmediate(async () => {
66
- const _customModels = await this.getCustomModels(candidate);
67
- teamModels = { ...teamModels, ..._customModels };
68
- });
69
-
70
- //immediatelly return the team models
71
- return teamModels;
72
- };
73
- loadTeamModels();
74
-
75
- const instance: IModelsProviderRequest = {
76
- getModels: async () => {
77
- return await loadTeamModels();
78
- },
79
- getCustomModels: async () => {
80
- return await this.getCustomModels(candidate);
81
- },
82
- addModels: async (models: TLLMModelsList) => {
83
- return await this.addModels(candidate.readRequest, models);
84
- },
85
- getModelInfo: async (model: string | TLLMModel | TCustomLLMModel, hasAPIKey: boolean = false) => {
86
- const teamModels = typeof model === 'string' ? await loadTeamModels() : {};
87
- const modelInfo = await this.getModelInfo(candidate.readRequest, teamModels, model, hasAPIKey);
88
- return modelInfo;
89
- },
90
-
91
- getModelId: async (model: string | TLLMModel | TCustomLLMModel) => {
92
- const teamModels = typeof model === 'string' ? await loadTeamModels() : {};
93
- return this.getModelId(candidate.readRequest, teamModels, model);
94
- },
95
- getProvider: async (model: string | TLLMModel | TCustomLLMModel) => {
96
- const teamModels = typeof model === 'string' ? await loadTeamModels() : {};
97
- return this.getProvider(candidate.readRequest, teamModels, model);
98
- },
99
- isStandardLLM: async (model: string | TLLMModel | TCustomLLMModel) => {
100
- const teamModels = typeof model === 'string' ? await loadTeamModels() : {};
101
- const modelInfo = await this.getModelInfo(candidate.readRequest, teamModels, model);
102
- return !modelInfo.isCustomLLM;
103
- },
104
- adjustMaxCompletionTokens: async (
105
- model: string | TLLMModel | TCustomLLMModel,
106
- maxCompletionTokens: number,
107
- hasAPIKey: boolean = false
108
- ) => {
109
- const teamModels = typeof model === 'string' ? await loadTeamModels() : {};
110
- const modelInfo = await this.getModelInfo(candidate.readRequest, teamModels, model, hasAPIKey);
111
- return Math.min(maxCompletionTokens || 512, modelInfo?.completionTokens || modelInfo?.tokens || maxCompletionTokens || 512);
112
- },
113
- getMaxContextTokens: async (model: string | TLLMModel | TCustomLLMModel, hasAPIKey: boolean = false) => {
114
- const teamModels = typeof model === 'string' ? await loadTeamModels() : {};
115
- const modelInfo = await this.getModelInfo(candidate.readRequest, teamModels, model, hasAPIKey);
116
- return modelInfo?.tokens || 1024;
117
- },
118
- getMaxCompletionTokens: async (model: string | TLLMModel | TCustomLLMModel, hasAPIKey: boolean = false) => {
119
- const teamModels = typeof model === 'string' ? await loadTeamModels() : {};
120
- const modelInfo = await this.getModelInfo(candidate.readRequest, teamModels, model, hasAPIKey);
121
- return modelInfo?.completionTokens || modelInfo?.tokens || 512;
122
- },
123
- validateTokensLimit: async ({
124
- model,
125
- promptTokens,
126
- completionTokens,
127
- hasAPIKey,
128
- }: {
129
- model: TLLMModel | TCustomLLMModel;
130
- promptTokens: number;
131
- completionTokens: number;
132
- hasAPIKey: boolean;
133
- }) => {
134
- //const teamModels = typeof model === 'string' ? await loadTeamModels() : {};
135
- const modelInfo = await this.getModelInfo(candidate.readRequest, {}, model, hasAPIKey);
136
- const allowedContextTokens = modelInfo?.tokens;
137
- const totalTokens = promptTokens + completionTokens;
138
-
139
- const teamAPIKeyExceededMessage = `This models' maximum content length is ${allowedContextTokens} tokens. (This is the sum of your prompt with all variables and the maximum output tokens you've set in Advanced Settings) However, you requested approx ${totalTokens} tokens (${promptTokens} in the prompt, ${completionTokens} in the output). Please reduce the length of either the input prompt or the Maximum output tokens.`;
140
- const noAPIKeyExceededMessage = `Input exceeds max tokens limit of ${allowedContextTokens}. Please add your API key and select Personal tagged models to unlock full length.`;
141
-
142
- if (totalTokens > allowedContextTokens) {
143
- throw new Error(hasAPIKey ? teamAPIKeyExceededMessage : noAPIKeyExceededMessage);
144
- }
145
- },
146
- };
147
- ModelsProviderConnector.localCache.set(cacheKey, instance, 10 * 60 * 1000); // cache for 10 minutes
148
- return instance;
149
- }
150
-
151
- protected async getModelInfo(
152
- acRequest: AccessRequest,
153
- models: TLLMModelsList,
154
- model: string | TLLMModel | TCustomLLMModel,
155
- hasAPIKey: boolean = false
156
- ): Promise<TLLMModel> {
157
- //model can be passed directly, in which case we do not need to look it up in the models list
158
-
159
- let modelId, alias, aliasModelInfo, modelInfo;
160
-
161
- if (typeof model === 'object' && model.modelId) {
162
- //return model;
163
- modelId = model.modelId;
164
- alias = model.alias;
165
- aliasModelInfo = models?.[alias];
166
- modelInfo = model;
167
- } else {
168
- //model can be passed as a string, in which case we need to look it up in the models list
169
-
170
- modelId = await this.getModelId(acRequest, models, model);
171
- alias = models?.[model as string]?.alias;
172
- aliasModelInfo = models?.[alias];
173
- modelInfo = models?.[model as string];
174
- }
175
-
176
- const aliasKeyOptions = aliasModelInfo && hasAPIKey ? aliasModelInfo?.keyOptions : null;
177
-
178
- const modelKeyOptions = modelInfo?.keyOptions || aliasKeyOptions;
179
-
180
- return { ...aliasModelInfo, ...modelInfo, ...aliasKeyOptions, ...modelKeyOptions, modelId };
181
- }
182
-
183
- protected async getModelId(acRequest: AccessRequest, models: TLLMModelsList, model: string | TLLMModel | TCustomLLMModel): Promise<string> {
184
- //model can be passed directly, in which case we do not need to look it up in the models list
185
- if (typeof model === 'object' && model.modelId) {
186
- return model.modelId;
187
- }
188
-
189
- //model can be passed as a string, in which case we need to look it up in the models list
190
- const modelId = models?.[model as string]?.modelId || (model as string);
191
- const alias = models?.[model as string]?.alias;
192
- if (alias) {
193
- const aliasModelId = models?.[alias]?.modelId || alias || (model as string);
194
- return aliasModelId;
195
- }
196
-
197
- return modelId;
198
- }
199
-
200
- // public static async validateTokensLimit({
201
- // model,
202
- // promptTokens,
203
- // completionTokens,
204
- // hasAPIKey = false,
205
- // }: {
206
- // model: string;
207
- // promptTokens: number;
208
- // completionTokens: number;
209
- // hasAPIKey?: boolean;
210
- // }): Promise<void> {
211
-
212
- // const allowedContextTokens = this.getMaxContextTokens(model, hasAPIKey);
213
- // const totalTokens = promptTokens + completionTokens;
214
-
215
- // const teamAPIKeyExceededMessage = `This models' maximum content length is ${allowedContextTokens} tokens. (This is the sum of your prompt with all variables and the maximum output tokens you've set in Advanced Settings) However, you requested approx ${totalTokens} tokens (${promptTokens} in the prompt, ${completionTokens} in the output). Please reduce the length of either the input prompt or the Maximum output tokens.`;
216
- // const noAPIKeyExceededMessage = `Input exceeds max tokens limit of ${allowedContextTokens}. Please add your API key to unlock full length.`;
217
-
218
- // if (totalTokens > allowedContextTokens) {
219
- // throw new Error(hasAPIKey ? teamAPIKeyExceededMessage : noAPIKeyExceededMessage);
220
- // }
221
- // }
222
-
223
- protected async getProvider(acRequest: AccessRequest, models: TLLMModelsList, model: string | TLLMModel | TCustomLLMModel): Promise<string> {
224
- //model can be passed directly, in which case we do not need to look it up in the models list
225
- if (typeof model === 'object' && model.provider) {
226
- return model.provider;
227
- }
228
-
229
- //model can be passed as a string, in which case we need to look it up in the models list
230
-
231
- const modelId = await this.getModelId(acRequest, models, model);
232
-
233
- return models?.[modelId]?.provider || models?.[model as string]?.provider || models?.[modelId]?.llm || models?.[model as string]?.llm;
234
- }
235
- protected async getCustomModels(candidate: IAccessCandidate): Promise<Record<string, any>> {
236
- const models = {};
237
- const settingsKey = 'custom-llm';
238
-
239
- try {
240
- const accountConnector = ConnectorService.getAccountConnector();
241
- const team = await accountConnector.requester(candidate as AccessCandidate).getTeam();
242
-
243
- const teamSettings = await accountConnector.team(team).getTeamSetting(settingsKey);
244
- const savedCustomModelsData = JSON.parse(teamSettings || '{}') as Record<string, any>;
245
-
246
- for (const [entryId, entry] of Object.entries(savedCustomModelsData)) {
247
- const foundationModel = entry.settings.foundationModel;
248
- const customModel = entry.settings.customModel;
249
- const supportsSystemPrompt = customModels[foundationModel]?.supportsSystemPrompt || entry.settings.supportsSystemPrompt;
250
- const customModelData = customModels[foundationModel] || {};
251
-
252
- let credentials = null;
253
- switch (entry.provider) {
254
- case TLLMProvider.Bedrock:
255
- credentials = TLLMCredentials.BedrockVault;
256
- break;
257
- case TLLMProvider.VertexAI:
258
- credentials = TLLMCredentials.VertexAIVault;
259
- break;
260
- default:
261
- credentials = TLLMCredentials.Internal;
262
- break;
263
- }
264
- models[entry.name] = {
265
- label: entry.name,
266
- modelId: customModel || foundationModel,
267
- provider: entry.provider,
268
- features: entry.features?.map((feature) => {
269
- switch (feature) {
270
- case 'text-completion':
271
- return 'text';
272
- case 'tool-use':
273
- return 'tools';
274
- default:
275
- return feature;
276
- }
277
- }),
278
- tags: Array.isArray(entry?.tags) ? ['Enterprise', ...entry?.tags] : ['Enterprise'],
279
- tokens: customModelData?.tokens ?? 100000,
280
- completionTokens: customModelData?.completionTokens ?? 4096,
281
- enabled: true,
282
-
283
- id: entryId,
284
- name: entry.name,
285
- alias: foundationModel,
286
- llm: entry.provider,
287
- components: customModelData?.components ?? [],
288
- isCustomLLM: true,
289
- supportsSystemPrompt,
290
- settings: entry.settings,
291
- credentials,
292
- };
293
- }
294
-
295
- return models;
296
- } catch (error) {
297
- return {};
298
- }
299
- }
300
- }
1
+ import { ConnectorService } from '@sre/Core/ConnectorsService';
2
+ import { AccessCandidate } from '@sre/Security/AccessControl/AccessCandidate.class';
3
+ import { AccessRequest } from '@sre/Security/AccessControl/AccessRequest.class';
4
+ import { ACL } from '@sre/Security/AccessControl/ACL.class';
5
+ import { SecureConnector } from '@sre/Security/SecureConnector.class';
6
+ import { IAccessCandidate } from '@sre/types/ACL.types';
7
+ import { TCustomLLMModel, TLLMCredentials, TLLMModel, TLLMModelsList, TLLMProvider } from '@sre/types/LLM.types';
8
+ import { customModels } from '../custom-models';
9
+ import { LocalCache } from '@sre/helpers/LocalCache.helper';
10
+
11
+ export interface IModelsProviderRequest {
12
+ getModels(): Promise<any>;
13
+ getCustomModels(): Promise<any>;
14
+ getMaxContextTokens(model: string, hasAPIKey?: boolean): Promise<number>;
15
+ addModels(models: TLLMModelsList): Promise<void>;
16
+ getModelInfo(model: string | TLLMModel | TCustomLLMModel, hasAPIKey?: boolean): Promise<TLLMModel>;
17
+ getModelId(model: string | TLLMModel | TCustomLLMModel): Promise<string>;
18
+ getProvider(model: string | TLLMModel | TCustomLLMModel): Promise<string>;
19
+ isStandardLLM(model: string | TLLMModel | TCustomLLMModel): Promise<boolean>;
20
+ adjustMaxCompletionTokens(model: string | TLLMModel | TCustomLLMModel, maxCompletionTokens: number, hasAPIKey?: boolean): Promise<number>;
21
+ getMaxContextTokens(model: string | TLLMModel | TCustomLLMModel, hasAPIKey?: boolean): Promise<number>;
22
+ getMaxCompletionTokens(model: string | TLLMModel | TCustomLLMModel, hasAPIKey?: boolean): Promise<number>;
23
+ validateTokensLimit({
24
+ model,
25
+ promptTokens,
26
+ completionTokens,
27
+ hasAPIKey,
28
+ }: {
29
+ model: TLLMModel | TCustomLLMModel;
30
+ promptTokens: number;
31
+ completionTokens: number;
32
+ hasAPIKey?: boolean;
33
+ }): Promise<void>;
34
+ }
35
+
36
+ export abstract class ModelsProviderConnector extends SecureConnector {
37
+ protected static localCache = new LocalCache();
38
+ public abstract getResourceACL(resourceId: string, candidate: IAccessCandidate): Promise<ACL>;
39
+ public abstract getModels(acRequest: AccessRequest): Promise<TLLMModelsList>;
40
+ public abstract addModels(acRequest: AccessRequest, models: TLLMModelsList): Promise<void>;
41
+
42
+ public requester(candidate: AccessCandidate): IModelsProviderRequest {
43
+ const cacheKey = `ModelsProviderConnector:${candidate.toString()}`;
44
+ if (ModelsProviderConnector.localCache.has(cacheKey)) {
45
+ //update the TTL every time the requester is called
46
+ return ModelsProviderConnector.localCache.get(cacheKey, 10 * 60 * 1000) as IModelsProviderRequest;
47
+ }
48
+
49
+ let teamModels = null;
50
+ let customModels = null;
51
+
52
+ const loadTeamModels = async () => {
53
+ if (!teamModels) {
54
+ try {
55
+ const builtinmodels = await this.getModels(candidate.readRequest);
56
+ customModels = await this.getCustomModels(candidate);
57
+ teamModels = { ...builtinmodels, ...customModels };
58
+ } catch (error) {
59
+ return null;
60
+ }
61
+ }
62
+ //Workaround : non-blocking auto-refresh of team models
63
+ //this will force team models to refresh for the next request
64
+ //TODO: we need a more elegant cache invalidation mechanism, and only refresh the team models if the custom models have changed
65
+ setImmediate(async () => {
66
+ const _customModels = await this.getCustomModels(candidate);
67
+ teamModels = { ...teamModels, ..._customModels };
68
+ });
69
+
70
+ //immediatelly return the team models
71
+ return teamModels;
72
+ };
73
+ loadTeamModels();
74
+
75
+ const instance: IModelsProviderRequest = {
76
+ getModels: async () => {
77
+ return await loadTeamModels();
78
+ },
79
+ getCustomModels: async () => {
80
+ return await this.getCustomModels(candidate);
81
+ },
82
+ addModels: async (models: TLLMModelsList) => {
83
+ return await this.addModels(candidate.readRequest, models);
84
+ },
85
+ getModelInfo: async (model: string | TLLMModel | TCustomLLMModel, hasAPIKey: boolean = false) => {
86
+ const teamModels = typeof model === 'string' ? await loadTeamModels() : {};
87
+ const modelInfo = await this.getModelInfo(candidate.readRequest, teamModels, model, hasAPIKey);
88
+ return modelInfo;
89
+ },
90
+
91
+ getModelId: async (model: string | TLLMModel | TCustomLLMModel) => {
92
+ const teamModels = typeof model === 'string' ? await loadTeamModels() : {};
93
+ return this.getModelId(candidate.readRequest, teamModels, model);
94
+ },
95
+ getProvider: async (model: string | TLLMModel | TCustomLLMModel) => {
96
+ const teamModels = typeof model === 'string' ? await loadTeamModels() : {};
97
+ return this.getProvider(candidate.readRequest, teamModels, model);
98
+ },
99
+ isStandardLLM: async (model: string | TLLMModel | TCustomLLMModel) => {
100
+ const teamModels = typeof model === 'string' ? await loadTeamModels() : {};
101
+ const modelInfo = await this.getModelInfo(candidate.readRequest, teamModels, model);
102
+ return !modelInfo.isCustomLLM;
103
+ },
104
+ adjustMaxCompletionTokens: async (
105
+ model: string | TLLMModel | TCustomLLMModel,
106
+ maxCompletionTokens: number,
107
+ hasAPIKey: boolean = false
108
+ ) => {
109
+ const teamModels = typeof model === 'string' ? await loadTeamModels() : {};
110
+ const modelInfo = await this.getModelInfo(candidate.readRequest, teamModels, model, hasAPIKey);
111
+ return Math.min(maxCompletionTokens || 512, modelInfo?.completionTokens || modelInfo?.tokens || maxCompletionTokens || 512);
112
+ },
113
+ getMaxContextTokens: async (model: string | TLLMModel | TCustomLLMModel, hasAPIKey: boolean = false) => {
114
+ const teamModels = typeof model === 'string' ? await loadTeamModels() : {};
115
+ const modelInfo = await this.getModelInfo(candidate.readRequest, teamModels, model, hasAPIKey);
116
+ return modelInfo?.tokens || 1024;
117
+ },
118
+ getMaxCompletionTokens: async (model: string | TLLMModel | TCustomLLMModel, hasAPIKey: boolean = false) => {
119
+ const teamModels = typeof model === 'string' ? await loadTeamModels() : {};
120
+ const modelInfo = await this.getModelInfo(candidate.readRequest, teamModels, model, hasAPIKey);
121
+ return modelInfo?.completionTokens || modelInfo?.tokens || 512;
122
+ },
123
+ validateTokensLimit: async ({
124
+ model,
125
+ promptTokens,
126
+ completionTokens,
127
+ hasAPIKey,
128
+ }: {
129
+ model: TLLMModel | TCustomLLMModel;
130
+ promptTokens: number;
131
+ completionTokens: number;
132
+ hasAPIKey: boolean;
133
+ }) => {
134
+ //const teamModels = typeof model === 'string' ? await loadTeamModels() : {};
135
+ const modelInfo = await this.getModelInfo(candidate.readRequest, {}, model, hasAPIKey);
136
+ const allowedContextTokens = modelInfo?.tokens;
137
+ const totalTokens = promptTokens + completionTokens;
138
+
139
+ const teamAPIKeyExceededMessage = `This models' maximum content length is ${allowedContextTokens} tokens. (This is the sum of your prompt with all variables and the maximum output tokens you've set in Advanced Settings) However, you requested approx ${totalTokens} tokens (${promptTokens} in the prompt, ${completionTokens} in the output). Please reduce the length of either the input prompt or the Maximum output tokens.`;
140
+ const noAPIKeyExceededMessage = `Input exceeds max tokens limit of ${allowedContextTokens}. Please add your API key and select Personal tagged models to unlock full length.`;
141
+
142
+ if (totalTokens > allowedContextTokens) {
143
+ throw new Error(hasAPIKey ? teamAPIKeyExceededMessage : noAPIKeyExceededMessage);
144
+ }
145
+ },
146
+ };
147
+ ModelsProviderConnector.localCache.set(cacheKey, instance, 10 * 60 * 1000); // cache for 10 minutes
148
+ return instance;
149
+ }
150
+
151
+ protected async getModelInfo(
152
+ acRequest: AccessRequest,
153
+ models: TLLMModelsList,
154
+ model: string | TLLMModel | TCustomLLMModel,
155
+ hasAPIKey: boolean = false
156
+ ): Promise<TLLMModel> {
157
+ //model can be passed directly, in which case we do not need to look it up in the models list
158
+
159
+ let modelId, alias, aliasModelInfo, modelInfo;
160
+
161
+ if (typeof model === 'object' && model.modelId) {
162
+ //return model;
163
+ modelId = model.modelId;
164
+ alias = model.alias;
165
+ aliasModelInfo = models?.[alias];
166
+ modelInfo = model;
167
+ } else {
168
+ //model can be passed as a string, in which case we need to look it up in the models list
169
+
170
+ modelId = await this.getModelId(acRequest, models, model);
171
+ alias = models?.[model as string]?.alias;
172
+ aliasModelInfo = models?.[alias];
173
+ modelInfo = models?.[model as string];
174
+ }
175
+
176
+ const aliasKeyOptions = aliasModelInfo && hasAPIKey ? aliasModelInfo?.keyOptions : null;
177
+
178
+ const modelKeyOptions = modelInfo?.keyOptions || aliasKeyOptions;
179
+
180
+ return { ...aliasModelInfo, ...modelInfo, ...aliasKeyOptions, ...modelKeyOptions, modelId };
181
+ }
182
+
183
+ protected async getModelId(acRequest: AccessRequest, models: TLLMModelsList, model: string | TLLMModel | TCustomLLMModel): Promise<string> {
184
+ //model can be passed directly, in which case we do not need to look it up in the models list
185
+ if (typeof model === 'object' && model.modelId) {
186
+ return model.modelId;
187
+ }
188
+
189
+ //model can be passed as a string, in which case we need to look it up in the models list
190
+ const modelId = models?.[model as string]?.modelId || (model as string);
191
+ const alias = models?.[model as string]?.alias;
192
+ if (alias) {
193
+ const aliasModelId = models?.[alias]?.modelId || alias || (model as string);
194
+ return aliasModelId;
195
+ }
196
+
197
+ return modelId;
198
+ }
199
+
200
+ // public static async validateTokensLimit({
201
+ // model,
202
+ // promptTokens,
203
+ // completionTokens,
204
+ // hasAPIKey = false,
205
+ // }: {
206
+ // model: string;
207
+ // promptTokens: number;
208
+ // completionTokens: number;
209
+ // hasAPIKey?: boolean;
210
+ // }): Promise<void> {
211
+
212
+ // const allowedContextTokens = this.getMaxContextTokens(model, hasAPIKey);
213
+ // const totalTokens = promptTokens + completionTokens;
214
+
215
+ // const teamAPIKeyExceededMessage = `This models' maximum content length is ${allowedContextTokens} tokens. (This is the sum of your prompt with all variables and the maximum output tokens you've set in Advanced Settings) However, you requested approx ${totalTokens} tokens (${promptTokens} in the prompt, ${completionTokens} in the output). Please reduce the length of either the input prompt or the Maximum output tokens.`;
216
+ // const noAPIKeyExceededMessage = `Input exceeds max tokens limit of ${allowedContextTokens}. Please add your API key to unlock full length.`;
217
+
218
+ // if (totalTokens > allowedContextTokens) {
219
+ // throw new Error(hasAPIKey ? teamAPIKeyExceededMessage : noAPIKeyExceededMessage);
220
+ // }
221
+ // }
222
+
223
+ protected async getProvider(acRequest: AccessRequest, models: TLLMModelsList, model: string | TLLMModel | TCustomLLMModel): Promise<string> {
224
+ //model can be passed directly, in which case we do not need to look it up in the models list
225
+ if (typeof model === 'object' && model.provider) {
226
+ return model.provider;
227
+ }
228
+
229
+ //model can be passed as a string, in which case we need to look it up in the models list
230
+
231
+ const modelId = await this.getModelId(acRequest, models, model);
232
+
233
+ return models?.[modelId]?.provider || models?.[model as string]?.provider || models?.[modelId]?.llm || models?.[model as string]?.llm;
234
+ }
235
+ protected async getCustomModels(candidate: IAccessCandidate): Promise<Record<string, any>> {
236
+ const models = {};
237
+ const settingsKey = 'custom-llm';
238
+
239
+ try {
240
+ const accountConnector = ConnectorService.getAccountConnector();
241
+ const team = await accountConnector.requester(candidate as AccessCandidate).getTeam();
242
+
243
+ const teamSettings = await accountConnector.team(team).getTeamSetting(settingsKey);
244
+ const savedCustomModelsData = JSON.parse(teamSettings || '{}') as Record<string, any>;
245
+
246
+ for (const [entryId, entry] of Object.entries(savedCustomModelsData)) {
247
+ const foundationModel = entry.settings.foundationModel;
248
+ const customModel = entry.settings.customModel;
249
+ const supportsSystemPrompt = customModels[foundationModel]?.supportsSystemPrompt || entry.settings.supportsSystemPrompt;
250
+ const customModelData = customModels[foundationModel] || {};
251
+
252
+ let credentials = null;
253
+ switch (entry.provider) {
254
+ case TLLMProvider.Bedrock:
255
+ credentials = TLLMCredentials.BedrockVault;
256
+ break;
257
+ case TLLMProvider.VertexAI:
258
+ credentials = TLLMCredentials.VertexAIVault;
259
+ break;
260
+ default:
261
+ credentials = TLLMCredentials.Internal;
262
+ break;
263
+ }
264
+ models[entry.name] = {
265
+ label: entry.name,
266
+ modelId: customModel || foundationModel,
267
+ provider: entry.provider,
268
+ features: entry.features?.map((feature) => {
269
+ switch (feature) {
270
+ case 'text-completion':
271
+ return 'text';
272
+ case 'tool-use':
273
+ return 'tools';
274
+ default:
275
+ return feature;
276
+ }
277
+ }),
278
+ tags: Array.isArray(entry?.tags) ? ['Enterprise', ...entry?.tags] : ['Enterprise'],
279
+ tokens: customModelData?.tokens ?? 100000,
280
+ completionTokens: customModelData?.completionTokens ?? 4096,
281
+ enabled: true,
282
+
283
+ id: entryId,
284
+ name: entry.name,
285
+ alias: foundationModel,
286
+ llm: entry.provider,
287
+ components: customModelData?.components ?? [],
288
+ isCustomLLM: true,
289
+ supportsSystemPrompt,
290
+ settings: entry.settings,
291
+ credentials,
292
+ };
293
+ }
294
+
295
+ return models;
296
+ } catch (error) {
297
+ return {};
298
+ }
299
+ }
300
+ }