@smythos/sre 1.5.44 → 1.5.45

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (227) hide show
  1. package/CHANGELOG +90 -90
  2. package/LICENSE +18 -18
  3. package/README.md +135 -135
  4. package/dist/index.js +3 -3
  5. package/dist/index.js.map +1 -1
  6. package/dist/types/subsystems/LLMManager/LLM.service/connectors/openai/OpenAIConnector.class.d.ts +13 -1
  7. package/dist/types/subsystems/LLMManager/LLM.service/connectors/openai/apiInterfaces/ResponsesApiInterface.d.ts +46 -27
  8. package/dist/types/subsystems/LLMManager/LLM.service/connectors/openai/apiInterfaces/constants.d.ts +4 -2
  9. package/dist/types/subsystems/LLMManager/LLM.service/connectors/openai/types.d.ts +0 -4
  10. package/dist/types/subsystems/LLMManager/ModelsProvider.service/connectors/SmythModelsProvider.class.d.ts +39 -0
  11. package/dist/types/types/LLM.types.d.ts +1 -0
  12. package/package.json +1 -1
  13. package/src/Components/APICall/APICall.class.ts +156 -156
  14. package/src/Components/APICall/AccessTokenManager.ts +130 -130
  15. package/src/Components/APICall/ArrayBufferResponse.helper.ts +58 -58
  16. package/src/Components/APICall/OAuth.helper.ts +294 -294
  17. package/src/Components/APICall/mimeTypeCategories.ts +46 -46
  18. package/src/Components/APICall/parseData.ts +167 -167
  19. package/src/Components/APICall/parseHeaders.ts +41 -41
  20. package/src/Components/APICall/parseProxy.ts +68 -68
  21. package/src/Components/APICall/parseUrl.ts +91 -91
  22. package/src/Components/APIEndpoint.class.ts +234 -234
  23. package/src/Components/APIOutput.class.ts +58 -58
  24. package/src/Components/AgentPlugin.class.ts +102 -102
  25. package/src/Components/Async.class.ts +155 -155
  26. package/src/Components/Await.class.ts +90 -90
  27. package/src/Components/Classifier.class.ts +158 -158
  28. package/src/Components/Component.class.ts +132 -132
  29. package/src/Components/ComponentHost.class.ts +38 -38
  30. package/src/Components/DataSourceCleaner.class.ts +92 -92
  31. package/src/Components/DataSourceIndexer.class.ts +181 -181
  32. package/src/Components/DataSourceLookup.class.ts +161 -161
  33. package/src/Components/ECMASandbox.class.ts +71 -71
  34. package/src/Components/FEncDec.class.ts +29 -29
  35. package/src/Components/FHash.class.ts +33 -33
  36. package/src/Components/FSign.class.ts +80 -80
  37. package/src/Components/FSleep.class.ts +25 -25
  38. package/src/Components/FTimestamp.class.ts +25 -25
  39. package/src/Components/FileStore.class.ts +78 -78
  40. package/src/Components/ForEach.class.ts +97 -97
  41. package/src/Components/GPTPlugin.class.ts +70 -70
  42. package/src/Components/GenAILLM.class.ts +586 -586
  43. package/src/Components/HuggingFace.class.ts +314 -314
  44. package/src/Components/Image/imageSettings.config.ts +70 -70
  45. package/src/Components/ImageGenerator.class.ts +502 -502
  46. package/src/Components/JSONFilter.class.ts +54 -54
  47. package/src/Components/LLMAssistant.class.ts +213 -213
  48. package/src/Components/LogicAND.class.ts +28 -28
  49. package/src/Components/LogicAtLeast.class.ts +85 -85
  50. package/src/Components/LogicAtMost.class.ts +86 -86
  51. package/src/Components/LogicOR.class.ts +29 -29
  52. package/src/Components/LogicXOR.class.ts +34 -34
  53. package/src/Components/MCPClient.class.ts +112 -112
  54. package/src/Components/MemoryDeleteKeyVal.class.ts +70 -70
  55. package/src/Components/MemoryReadKeyVal.class.ts +66 -66
  56. package/src/Components/MemoryWriteKeyVal.class.ts +62 -62
  57. package/src/Components/MemoryWriteObject.class.ts +97 -97
  58. package/src/Components/MultimodalLLM.class.ts +128 -128
  59. package/src/Components/OpenAPI.class.ts +72 -72
  60. package/src/Components/PromptGenerator.class.ts +122 -122
  61. package/src/Components/ScrapflyWebScrape.class.ts +159 -159
  62. package/src/Components/ServerlessCode.class.ts +123 -123
  63. package/src/Components/TavilyWebSearch.class.ts +98 -98
  64. package/src/Components/VisionLLM.class.ts +104 -104
  65. package/src/Components/ZapierAction.class.ts +127 -127
  66. package/src/Components/index.ts +97 -97
  67. package/src/Core/AgentProcess.helper.ts +240 -240
  68. package/src/Core/Connector.class.ts +123 -123
  69. package/src/Core/ConnectorsService.ts +197 -197
  70. package/src/Core/DummyConnector.ts +49 -49
  71. package/src/Core/HookService.ts +105 -105
  72. package/src/Core/SmythRuntime.class.ts +235 -235
  73. package/src/Core/SystemEvents.ts +16 -16
  74. package/src/Core/boot.ts +56 -56
  75. package/src/config.ts +15 -15
  76. package/src/constants.ts +126 -126
  77. package/src/data/hugging-face.params.json +579 -579
  78. package/src/helpers/AWSLambdaCode.helper.ts +587 -587
  79. package/src/helpers/BinaryInput.helper.ts +331 -331
  80. package/src/helpers/Conversation.helper.ts +1119 -1119
  81. package/src/helpers/ECMASandbox.helper.ts +54 -54
  82. package/src/helpers/JsonContent.helper.ts +97 -97
  83. package/src/helpers/LocalCache.helper.ts +97 -97
  84. package/src/helpers/Log.helper.ts +274 -274
  85. package/src/helpers/OpenApiParser.helper.ts +150 -150
  86. package/src/helpers/S3Cache.helper.ts +147 -147
  87. package/src/helpers/SmythURI.helper.ts +5 -5
  88. package/src/helpers/Sysconfig.helper.ts +77 -77
  89. package/src/helpers/TemplateString.helper.ts +243 -243
  90. package/src/helpers/TypeChecker.helper.ts +329 -329
  91. package/src/index.ts +196 -196
  92. package/src/index.ts.bak +196 -196
  93. package/src/subsystems/AgentManager/Agent.class.ts +1114 -1114
  94. package/src/subsystems/AgentManager/Agent.helper.ts +3 -3
  95. package/src/subsystems/AgentManager/AgentData.service/AgentDataConnector.ts +230 -230
  96. package/src/subsystems/AgentManager/AgentData.service/connectors/CLIAgentDataConnector.class.ts +66 -66
  97. package/src/subsystems/AgentManager/AgentData.service/connectors/LocalAgentDataConnector.class.ts +142 -142
  98. package/src/subsystems/AgentManager/AgentData.service/connectors/NullAgentData.class.ts +39 -39
  99. package/src/subsystems/AgentManager/AgentData.service/index.ts +18 -18
  100. package/src/subsystems/AgentManager/AgentLogger.class.ts +297 -297
  101. package/src/subsystems/AgentManager/AgentRequest.class.ts +51 -51
  102. package/src/subsystems/AgentManager/AgentRuntime.class.ts +559 -559
  103. package/src/subsystems/AgentManager/AgentSSE.class.ts +101 -101
  104. package/src/subsystems/AgentManager/AgentSettings.class.ts +52 -52
  105. package/src/subsystems/AgentManager/Component.service/ComponentConnector.ts +32 -32
  106. package/src/subsystems/AgentManager/Component.service/connectors/LocalComponentConnector.class.ts +60 -60
  107. package/src/subsystems/AgentManager/Component.service/index.ts +11 -11
  108. package/src/subsystems/AgentManager/EmbodimentSettings.class.ts +47 -47
  109. package/src/subsystems/AgentManager/ForkedAgent.class.ts +154 -154
  110. package/src/subsystems/AgentManager/OSResourceMonitor.ts +77 -77
  111. package/src/subsystems/ComputeManager/Code.service/CodeConnector.ts +98 -98
  112. package/src/subsystems/ComputeManager/Code.service/connectors/AWSLambdaCode.class.ts +172 -172
  113. package/src/subsystems/ComputeManager/Code.service/connectors/ECMASandbox.class.ts +131 -131
  114. package/src/subsystems/ComputeManager/Code.service/index.ts +13 -13
  115. package/src/subsystems/IO/CLI.service/CLIConnector.ts +47 -47
  116. package/src/subsystems/IO/CLI.service/index.ts +9 -9
  117. package/src/subsystems/IO/Log.service/LogConnector.ts +32 -32
  118. package/src/subsystems/IO/Log.service/connectors/ConsoleLog.class.ts +28 -28
  119. package/src/subsystems/IO/Log.service/index.ts +13 -13
  120. package/src/subsystems/IO/NKV.service/NKVConnector.ts +43 -43
  121. package/src/subsystems/IO/NKV.service/connectors/NKVLocalStorage.class.ts +234 -234
  122. package/src/subsystems/IO/NKV.service/connectors/NKVRAM.class.ts +204 -204
  123. package/src/subsystems/IO/NKV.service/connectors/NKVRedis.class.ts +182 -182
  124. package/src/subsystems/IO/NKV.service/index.ts +14 -14
  125. package/src/subsystems/IO/Router.service/RouterConnector.ts +21 -21
  126. package/src/subsystems/IO/Router.service/connectors/ExpressRouter.class.ts +48 -48
  127. package/src/subsystems/IO/Router.service/connectors/NullRouter.class.ts +40 -40
  128. package/src/subsystems/IO/Router.service/index.ts +11 -11
  129. package/src/subsystems/IO/Storage.service/SmythFS.class.ts +489 -489
  130. package/src/subsystems/IO/Storage.service/StorageConnector.ts +66 -66
  131. package/src/subsystems/IO/Storage.service/connectors/LocalStorage.class.ts +327 -327
  132. package/src/subsystems/IO/Storage.service/connectors/S3Storage.class.ts +482 -482
  133. package/src/subsystems/IO/Storage.service/index.ts +13 -13
  134. package/src/subsystems/IO/VectorDB.service/VectorDBConnector.ts +108 -108
  135. package/src/subsystems/IO/VectorDB.service/connectors/MilvusVectorDB.class.ts +454 -454
  136. package/src/subsystems/IO/VectorDB.service/connectors/PineconeVectorDB.class.ts +384 -384
  137. package/src/subsystems/IO/VectorDB.service/connectors/RAMVecrtorDB.class.ts +421 -421
  138. package/src/subsystems/IO/VectorDB.service/embed/BaseEmbedding.ts +107 -107
  139. package/src/subsystems/IO/VectorDB.service/embed/OpenAIEmbedding.ts +109 -109
  140. package/src/subsystems/IO/VectorDB.service/embed/index.ts +21 -21
  141. package/src/subsystems/IO/VectorDB.service/index.ts +14 -14
  142. package/src/subsystems/LLMManager/LLM.helper.ts +251 -251
  143. package/src/subsystems/LLMManager/LLM.inference.ts +339 -339
  144. package/src/subsystems/LLMManager/LLM.service/LLMConnector.ts +489 -489
  145. package/src/subsystems/LLMManager/LLM.service/LLMCredentials.helper.ts +171 -171
  146. package/src/subsystems/LLMManager/LLM.service/connectors/Anthropic.class.ts +659 -659
  147. package/src/subsystems/LLMManager/LLM.service/connectors/Bedrock.class.ts +400 -400
  148. package/src/subsystems/LLMManager/LLM.service/connectors/Echo.class.ts +77 -77
  149. package/src/subsystems/LLMManager/LLM.service/connectors/GoogleAI.class.ts +757 -757
  150. package/src/subsystems/LLMManager/LLM.service/connectors/Groq.class.ts +304 -304
  151. package/src/subsystems/LLMManager/LLM.service/connectors/Perplexity.class.ts +250 -250
  152. package/src/subsystems/LLMManager/LLM.service/connectors/VertexAI.class.ts +423 -423
  153. package/src/subsystems/LLMManager/LLM.service/connectors/openai/OpenAIConnector.class.ts +488 -455
  154. package/src/subsystems/LLMManager/LLM.service/connectors/openai/apiInterfaces/ChatCompletionsApiInterface.ts +528 -528
  155. package/src/subsystems/LLMManager/LLM.service/connectors/openai/apiInterfaces/OpenAIApiInterface.ts +100 -100
  156. package/src/subsystems/LLMManager/LLM.service/connectors/openai/apiInterfaces/OpenAIApiInterfaceFactory.ts +81 -81
  157. package/src/subsystems/LLMManager/LLM.service/connectors/openai/apiInterfaces/ResponsesApiInterface.ts +1168 -862
  158. package/src/subsystems/LLMManager/LLM.service/connectors/openai/apiInterfaces/constants.ts +13 -37
  159. package/src/subsystems/LLMManager/LLM.service/connectors/openai/apiInterfaces/index.ts +4 -4
  160. package/src/subsystems/LLMManager/LLM.service/connectors/openai/apiInterfaces/utils.ts +11 -11
  161. package/src/subsystems/LLMManager/LLM.service/connectors/openai/types.ts +32 -37
  162. package/src/subsystems/LLMManager/LLM.service/connectors/xAI.class.ts +471 -471
  163. package/src/subsystems/LLMManager/LLM.service/index.ts +44 -44
  164. package/src/subsystems/LLMManager/ModelsProvider.service/ModelsProviderConnector.ts +300 -300
  165. package/src/subsystems/LLMManager/ModelsProvider.service/connectors/JSONModelsProvider.class.ts +252 -252
  166. package/src/subsystems/LLMManager/ModelsProvider.service/index.ts +11 -11
  167. package/src/subsystems/LLMManager/custom-models.ts +854 -854
  168. package/src/subsystems/LLMManager/models.ts +2540 -2540
  169. package/src/subsystems/LLMManager/paramMappings.ts +69 -69
  170. package/src/subsystems/MemoryManager/Cache.service/CacheConnector.ts +86 -86
  171. package/src/subsystems/MemoryManager/Cache.service/connectors/LocalStorageCache.class.ts +297 -297
  172. package/src/subsystems/MemoryManager/Cache.service/connectors/RAMCache.class.ts +201 -201
  173. package/src/subsystems/MemoryManager/Cache.service/connectors/RedisCache.class.ts +252 -252
  174. package/src/subsystems/MemoryManager/Cache.service/connectors/S3Cache.class.ts +373 -373
  175. package/src/subsystems/MemoryManager/Cache.service/index.ts +15 -15
  176. package/src/subsystems/MemoryManager/LLMCache.ts +72 -72
  177. package/src/subsystems/MemoryManager/LLMContext.ts +124 -124
  178. package/src/subsystems/MemoryManager/LLMMemory.service/LLMMemoryConnector.ts +26 -26
  179. package/src/subsystems/MemoryManager/RuntimeContext.ts +266 -266
  180. package/src/subsystems/Security/AccessControl/ACL.class.ts +208 -208
  181. package/src/subsystems/Security/AccessControl/AccessCandidate.class.ts +82 -82
  182. package/src/subsystems/Security/AccessControl/AccessRequest.class.ts +52 -52
  183. package/src/subsystems/Security/Account.service/AccountConnector.ts +44 -44
  184. package/src/subsystems/Security/Account.service/connectors/AWSAccount.class.ts +76 -76
  185. package/src/subsystems/Security/Account.service/connectors/DummyAccount.class.ts +130 -130
  186. package/src/subsystems/Security/Account.service/connectors/JSONFileAccount.class.ts +159 -159
  187. package/src/subsystems/Security/Account.service/index.ts +14 -14
  188. package/src/subsystems/Security/Credentials.helper.ts +62 -62
  189. package/src/subsystems/Security/ManagedVault.service/ManagedVaultConnector.ts +38 -38
  190. package/src/subsystems/Security/ManagedVault.service/connectors/NullManagedVault.class.ts +53 -53
  191. package/src/subsystems/Security/ManagedVault.service/connectors/SecretManagerManagedVault.ts +154 -154
  192. package/src/subsystems/Security/ManagedVault.service/index.ts +12 -12
  193. package/src/subsystems/Security/SecureConnector.class.ts +110 -110
  194. package/src/subsystems/Security/Vault.service/Vault.helper.ts +30 -30
  195. package/src/subsystems/Security/Vault.service/VaultConnector.ts +29 -29
  196. package/src/subsystems/Security/Vault.service/connectors/HashicorpVault.class.ts +46 -46
  197. package/src/subsystems/Security/Vault.service/connectors/JSONFileVault.class.ts +221 -221
  198. package/src/subsystems/Security/Vault.service/connectors/NullVault.class.ts +54 -54
  199. package/src/subsystems/Security/Vault.service/connectors/SecretsManager.class.ts +140 -140
  200. package/src/subsystems/Security/Vault.service/index.ts +12 -12
  201. package/src/types/ACL.types.ts +104 -104
  202. package/src/types/AWS.types.ts +10 -10
  203. package/src/types/Agent.types.ts +61 -61
  204. package/src/types/AgentLogger.types.ts +17 -17
  205. package/src/types/Cache.types.ts +1 -1
  206. package/src/types/Common.types.ts +2 -2
  207. package/src/types/LLM.types.ts +496 -495
  208. package/src/types/Redis.types.ts +8 -8
  209. package/src/types/SRE.types.ts +64 -64
  210. package/src/types/Security.types.ts +14 -14
  211. package/src/types/Storage.types.ts +5 -5
  212. package/src/types/VectorDB.types.ts +86 -86
  213. package/src/utils/base64.utils.ts +275 -275
  214. package/src/utils/cli.utils.ts +68 -68
  215. package/src/utils/data.utils.ts +322 -322
  216. package/src/utils/date-time.utils.ts +22 -22
  217. package/src/utils/general.utils.ts +238 -238
  218. package/src/utils/index.ts +12 -12
  219. package/src/utils/lazy-client.ts +261 -261
  220. package/src/utils/numbers.utils.ts +13 -13
  221. package/src/utils/oauth.utils.ts +35 -35
  222. package/src/utils/string.utils.ts +414 -414
  223. package/src/utils/url.utils.ts +19 -19
  224. package/src/utils/validation.utils.ts +74 -74
  225. package/dist/bundle-analysis-lazy.html +0 -4949
  226. package/dist/bundle-analysis.html +0 -4949
  227. package/dist/types/utils/package-manager.utils.d.ts +0 -26
@@ -1,757 +1,757 @@
1
- import os from 'os';
2
- import path from 'path';
3
- import EventEmitter from 'events';
4
- import fs from 'fs';
5
-
6
- import { GoogleGenerativeAI, ModelParams, GenerationConfig, GenerateContentRequest, UsageMetadata, FunctionCallingMode } from '@google/generative-ai';
7
- import { GoogleAIFileManager, FileState } from '@google/generative-ai/server';
8
- import { GoogleGenAI } from '@google/genai';
9
-
10
- import { JSON_RESPONSE_INSTRUCTION, BUILT_IN_MODEL_PREFIX } from '@sre/constants';
11
- import { BinaryInput } from '@sre/helpers/BinaryInput.helper';
12
- import { AccessCandidate } from '@sre/Security/AccessControl/AccessCandidate.class';
13
- import { uid } from '@sre/utils';
14
-
15
- import { processWithConcurrencyLimit } from '@sre/utils';
16
-
17
- import {
18
- TLLMMessageBlock,
19
- ToolData,
20
- TLLMMessageRole,
21
- TLLMToolResultMessageBlock,
22
- APIKeySource,
23
- TLLMEvent,
24
- BasicCredentials,
25
- ILLMRequestFuncParams,
26
- TLLMChatResponse,
27
- TGoogleAIRequestBody,
28
- ILLMRequestContext,
29
- TLLMPreparedParams,
30
- } from '@sre/types/LLM.types';
31
- import { LLMHelper } from '@sre/LLMManager/LLM.helper';
32
-
33
- import { SystemEvents } from '@sre/Core/SystemEvents';
34
- import { SUPPORTED_MIME_TYPES_MAP } from '@sre/constants';
35
-
36
- import { LLMConnector } from '../LLMConnector';
37
-
38
- const MODELS_SUPPORT_SYSTEM_INSTRUCTION = [
39
- 'gemini-1.5-pro-exp-0801',
40
- 'gemini-1.5-pro-latest',
41
- 'gemini-1.5-pro-latest',
42
- 'gemini-1.5-pro',
43
- 'gemini-1.5-pro-001',
44
- 'gemini-1.5-flash-latest',
45
- 'gemini-1.5-flash-001',
46
- 'gemini-1.5-flash',
47
- ];
48
- const MODELS_SUPPORT_JSON_RESPONSE = MODELS_SUPPORT_SYSTEM_INSTRUCTION;
49
-
50
- // Supported file MIME types for Google AI's Gemini models
51
- const VALID_MIME_TYPES = [
52
- ...SUPPORTED_MIME_TYPES_MAP.GoogleAI.image,
53
- ...SUPPORTED_MIME_TYPES_MAP.GoogleAI.audio,
54
- ...SUPPORTED_MIME_TYPES_MAP.GoogleAI.video,
55
- ...SUPPORTED_MIME_TYPES_MAP.GoogleAI.document,
56
- ];
57
-
58
- // will be removed after updating the SDK
59
- type UsageMetadataWithThoughtsToken = UsageMetadata & { thoughtsTokenCount: number };
60
-
61
- export class GoogleAIConnector extends LLMConnector {
62
- public name = 'LLM:GoogleAI';
63
-
64
- private validMimeTypes = {
65
- all: VALID_MIME_TYPES,
66
- image: SUPPORTED_MIME_TYPES_MAP.GoogleAI.image,
67
- };
68
-
69
- private async getClient(params: ILLMRequestContext): Promise<GoogleGenerativeAI> {
70
- const apiKey = (params.credentials as BasicCredentials)?.apiKey;
71
-
72
- if (!apiKey) throw new Error('Please provide an API key for Google AI');
73
-
74
- return new GoogleGenerativeAI(apiKey);
75
- }
76
-
77
- protected async request({ acRequest, body, context }: ILLMRequestFuncParams): Promise<TLLMChatResponse> {
78
- try {
79
- const prompt = body.messages;
80
- delete body.messages;
81
-
82
- const genAI = await this.getClient(context);
83
- const $model = genAI.getGenerativeModel(body);
84
-
85
- const result = await $model.generateContent(prompt);
86
-
87
- const response = await result.response;
88
- const content = response.text();
89
- const finishReason = response.candidates[0].finishReason || 'stop';
90
- const usage = response?.usageMetadata as UsageMetadataWithThoughtsToken;
91
- this.reportUsage(usage, {
92
- modelEntryName: context.modelEntryName,
93
- keySource: context.isUserKey ? APIKeySource.User : APIKeySource.Smyth,
94
- agentId: context.agentId,
95
- teamId: context.teamId,
96
- });
97
-
98
- const toolCalls = response.candidates[0]?.content?.parts?.filter((part) => part.functionCall);
99
-
100
- let toolsData: ToolData[] = [];
101
- let useTool = false;
102
-
103
- if (toolCalls && toolCalls.length > 0) {
104
- toolsData = toolCalls.map((toolCall, index) => ({
105
- index,
106
- id: `tool-${index}`,
107
- type: 'function',
108
- name: toolCall.functionCall.name,
109
- arguments: JSON.stringify(toolCall.functionCall.args),
110
- role: TLLMMessageRole.Assistant,
111
- }));
112
- useTool = true;
113
- }
114
-
115
- return {
116
- content,
117
- finishReason: finishReason.toLowerCase(),
118
- useTool,
119
- toolsData,
120
- message: { content, role: 'assistant' },
121
- usage,
122
- };
123
- } catch (error: any) {
124
- throw error;
125
- }
126
- }
127
-
128
- protected async streamRequest({ acRequest, body, context }: ILLMRequestFuncParams): Promise<EventEmitter> {
129
- const emitter = new EventEmitter();
130
-
131
- const prompt = body.messages;
132
- delete body.messages;
133
-
134
- const genAI = await this.getClient(context);
135
- const $model = genAI.getGenerativeModel(body);
136
-
137
- try {
138
- const result = await $model.generateContentStream(prompt);
139
-
140
- let toolsData: ToolData[] = [];
141
- let usage: UsageMetadataWithThoughtsToken;
142
-
143
- // Process stream asynchronously while as we need to return emitter immediately
144
- (async () => {
145
- for await (const chunk of result.stream) {
146
- const chunkText = chunk.text();
147
- emitter.emit('content', chunkText);
148
-
149
- if (chunk.candidates[0]?.content?.parts) {
150
- const toolCalls = chunk.candidates[0].content.parts.filter((part) => part.functionCall);
151
- if (toolCalls.length > 0) {
152
- toolsData = toolCalls.map((toolCall, index) => ({
153
- index,
154
- id: `tool-${index}`,
155
- type: 'function',
156
- name: toolCall.functionCall.name,
157
- arguments: JSON.stringify(toolCall.functionCall.args),
158
- role: TLLMMessageRole.Assistant,
159
- }));
160
- emitter.emit(TLLMEvent.ToolInfo, toolsData);
161
- }
162
- }
163
-
164
- // the same usage is sent on each emit. IMPORTANT: google does not send usage for each chunk but
165
- // rather just sends the same usage for the entire request.
166
- // notice that the output tokens are only sent in the last chunk usage metadata.
167
- // so we will just update a var to hold the latest usage and report it when the stream ends.
168
- // e.g emit1: { input_tokens: 500, output_tokens: undefined } -> same input_tokens
169
- // e.g emit2: { input_tokens: 500, output_tokens: undefined } -> same input_tokens
170
- // e.g emit3: { input_tokens: 500, output_tokens: 10 } -> same input_tokens, new output_tokens in the last chunk
171
- if (chunk?.usageMetadata) {
172
- usage = chunk.usageMetadata as UsageMetadataWithThoughtsToken;
173
- }
174
- }
175
-
176
- if (usage) {
177
- this.reportUsage(usage, {
178
- modelEntryName: context.modelEntryName,
179
- keySource: context.isUserKey ? APIKeySource.User : APIKeySource.Smyth,
180
- agentId: context.agentId,
181
- teamId: context.teamId,
182
- });
183
- }
184
-
185
- setTimeout(() => {
186
- emitter.emit('end', toolsData);
187
- }, 100);
188
- })();
189
-
190
- return emitter;
191
- } catch (error: any) {
192
- throw error;
193
- }
194
- }
195
- // #region Image Generation, will be moved to a different subsystem/service
196
- protected async imageGenRequest({ body, context }: ILLMRequestFuncParams): Promise<any> {
197
- try {
198
- const apiKey = (context.credentials as BasicCredentials)?.apiKey;
199
- if (!apiKey) throw new Error('Please provide an API key for Google AI');
200
-
201
- const model = body.model || 'imagen-3.0-generate-001';
202
-
203
- // Use Imagen models via GoogleGenAI
204
- const ai = new GoogleGenAI({ apiKey });
205
-
206
- // Prepare the configuration for image generation
207
- const config = {
208
- numberOfImages: body.n || 1,
209
- aspectRatio: body.aspect_ratio || body.size || '1:1',
210
- personGeneration: body.person_generation || 'allow_adult',
211
- };
212
-
213
- // Generate images using the SDK
214
- const response = await ai.models.generateImages({
215
- model,
216
- prompt: body.prompt,
217
- config,
218
- });
219
-
220
- // Transform the response to match OpenAI format for compatibility
221
- return {
222
- created: Math.floor(Date.now() / 1000),
223
- data:
224
- response.generatedImages?.map((generatedImage: any) => ({
225
- url: generatedImage.image.imageBytes ? `data:image/png;base64,${generatedImage.image.imageBytes}` : undefined,
226
- b64_json: generatedImage.image.imageBytes,
227
- revised_prompt: body.prompt,
228
- })) || [],
229
- };
230
- } catch (error: any) {
231
- throw error;
232
- }
233
- }
234
-
235
- protected async imageEditRequest({ body, context }: ILLMRequestFuncParams): Promise<any> {
236
- throw new Error('Image editing is not supported for Google AI. Imagen models only support image generation.');
237
- }
238
-
239
- protected async reqBodyAdapter(params: TLLMPreparedParams): Promise<TGoogleAIRequestBody> {
240
- const model = params?.model;
241
-
242
- // Check if this is an image generation request based on capabilities
243
- if (params?.capabilities?.imageGeneration) {
244
- return this.prepareBodyForImageGenRequest(params) as any;
245
- }
246
-
247
- const messages = await this.prepareMessages(params);
248
-
249
- let body: ModelParams & { messages: string | TLLMMessageBlock[] | GenerateContentRequest } = {
250
- model: model as string,
251
- messages,
252
- };
253
-
254
- const responseFormat = params?.responseFormat || '';
255
- let responseMimeType = '';
256
- let systemInstruction = '';
257
-
258
- if (responseFormat === 'json') {
259
- systemInstruction += JSON_RESPONSE_INSTRUCTION;
260
-
261
- if (MODELS_SUPPORT_JSON_RESPONSE.includes(model as string)) {
262
- responseMimeType = 'application/json';
263
- }
264
- }
265
-
266
- const config: GenerationConfig = {};
267
-
268
- if (params.maxTokens !== undefined) config.maxOutputTokens = params.maxTokens;
269
- if (params.temperature !== undefined) config.temperature = params.temperature;
270
- if (params.topP !== undefined) config.topP = params.topP;
271
- if (params.topK !== undefined) config.topK = params.topK;
272
- if (params.stopSequences?.length) config.stopSequences = params.stopSequences;
273
- if (responseMimeType) config.responseMimeType = responseMimeType;
274
-
275
- if (systemInstruction) body.systemInstruction = systemInstruction;
276
- if (Object.keys(config).length > 0) {
277
- body.generationConfig = config;
278
- }
279
-
280
- return body;
281
- }
282
-
283
- protected reportUsage(
284
- usage: UsageMetadataWithThoughtsToken,
285
- metadata: { modelEntryName: string; keySource: APIKeySource; agentId: string; teamId: string }
286
- ) {
287
- const modelEntryName = metadata.modelEntryName;
288
- let tier = '';
289
-
290
- const tierThresholds = {
291
- 'gemini-1.5-pro': 128_000,
292
- 'gemini-2.5-pro': 200_000,
293
- };
294
-
295
- const textInputTokens =
296
- usage?.['promptTokensDetails']?.find((detail) => detail.modality === 'TEXT')?.tokenCount || usage?.promptTokenCount || 0;
297
- const audioInputTokens = usage?.['promptTokensDetails']?.find((detail) => detail.modality === 'AUDIO')?.tokenCount || 0;
298
-
299
- // Find matching model and set tier based on threshold
300
- const modelWithTier = Object.keys(tierThresholds).find((model) => modelEntryName.includes(model));
301
- if (modelWithTier) {
302
- tier = textInputTokens < tierThresholds[modelWithTier] ? 'tier1' : 'tier2';
303
- }
304
-
305
- // #endregion
306
-
307
- // SmythOS (built-in) models have a prefix, so we need to remove it to get the model name
308
- const modelName = metadata.modelEntryName.replace(BUILT_IN_MODEL_PREFIX, '');
309
-
310
- const usageData = {
311
- sourceId: `llm:${modelName}`,
312
- input_tokens: textInputTokens,
313
- output_tokens: usage.candidatesTokenCount,
314
- input_tokens_audio: audioInputTokens,
315
- input_tokens_cache_read: usage.cachedContentTokenCount || 0,
316
- input_tokens_cache_write: 0,
317
- reasoning_tokens: usage.thoughtsTokenCount,
318
- keySource: metadata.keySource,
319
- agentId: metadata.agentId,
320
- teamId: metadata.teamId,
321
- tier,
322
- };
323
- SystemEvents.emit('USAGE:LLM', usageData);
324
-
325
- return usageData;
326
- }
327
-
328
- public formatToolsConfig({ toolDefinitions, toolChoice = 'auto' }) {
329
- const tools = toolDefinitions.map((tool) => {
330
- const { name, description, properties, requiredFields } = tool;
331
-
332
- // Ensure the function name is valid
333
- const validName = this.sanitizeFunctionName(name);
334
-
335
- // Ensure properties are non-empty for OBJECT type
336
- const validProperties = properties && Object.keys(properties).length > 0 ? properties : { dummy: { type: 'string' } };
337
-
338
- return {
339
- functionDeclarations: [
340
- {
341
- name: validName,
342
- description: description || '',
343
- parameters: {
344
- type: 'OBJECT',
345
- properties: validProperties,
346
- required: requiredFields || [],
347
- },
348
- },
349
- ],
350
- };
351
- });
352
-
353
- return {
354
- tools,
355
- toolChoice: {
356
- type: toolChoice,
357
- },
358
- };
359
- }
360
-
361
- public transformToolMessageBlocks({
362
- messageBlock,
363
- toolsData,
364
- }: {
365
- messageBlock: TLLMMessageBlock;
366
- toolsData: ToolData[];
367
- }): TLLMToolResultMessageBlock[] {
368
- const messageBlocks: TLLMToolResultMessageBlock[] = [];
369
-
370
- if (messageBlock) {
371
- const content = [];
372
- if (typeof messageBlock.content === 'string') {
373
- content.push({ text: messageBlock.content });
374
- } else if (Array.isArray(messageBlock.content)) {
375
- content.push(...messageBlock.content);
376
- }
377
-
378
- if (messageBlock.parts) {
379
- const functionCalls = messageBlock.parts.filter((part) => part.functionCall);
380
- if (functionCalls.length > 0) {
381
- content.push(
382
- ...functionCalls.map((call) => ({
383
- functionCall: {
384
- name: call.functionCall.name,
385
- args: JSON.parse(call.functionCall.args),
386
- },
387
- }))
388
- );
389
- }
390
- }
391
-
392
- messageBlocks.push({
393
- role: messageBlock.role,
394
- parts: content,
395
- });
396
- }
397
-
398
- const transformedToolsData = toolsData.map(
399
- (toolData): TLLMToolResultMessageBlock => ({
400
- role: TLLMMessageRole.User,
401
- parts: [
402
- {
403
- functionResponse: {
404
- name: toolData.name,
405
- response: {
406
- name: toolData.name,
407
- content: typeof toolData.result === 'string' ? toolData.result : JSON.stringify(toolData.result),
408
- },
409
- },
410
- },
411
- ],
412
- })
413
- );
414
-
415
- return [...messageBlocks, ...transformedToolsData];
416
- }
417
-
418
- public getConsistentMessages(messages: TLLMMessageBlock[]): TLLMMessageBlock[] {
419
- const _messages = LLMHelper.removeDuplicateUserMessages(messages);
420
-
421
- return _messages.map((message) => {
422
- const _message = { ...message };
423
- let textContent = '';
424
-
425
- // Map roles to valid Google AI roles
426
- switch (_message.role) {
427
- case TLLMMessageRole.Assistant:
428
- case TLLMMessageRole.System:
429
- _message.role = TLLMMessageRole.Model;
430
- break;
431
- case TLLMMessageRole.User:
432
- // User role is already valid
433
- break;
434
- default:
435
- _message.role = TLLMMessageRole.User; // Default to user for unknown roles
436
- }
437
-
438
- // * empty text causes error that's why we added '...'
439
-
440
- if (_message?.parts) {
441
- textContent = _message.parts.map((textBlock) => textBlock?.text || '...').join(' ');
442
- } else if (Array.isArray(_message?.content)) {
443
- textContent = _message.content.map((textBlock) => textBlock?.text || '...').join(' ');
444
- } else if (_message?.content) {
445
- textContent = (_message.content as string) || '...';
446
- }
447
-
448
- _message.parts = [{ text: textContent || '...' }];
449
-
450
- delete _message.content; // Remove content to avoid error
451
-
452
- return _message;
453
- });
454
- }
455
-
456
- private async prepareMessages(params: TLLMPreparedParams): Promise<string | TLLMMessageBlock[] | GenerateContentRequest> {
457
- let messages: string | TLLMMessageBlock[] | GenerateContentRequest = params?.messages || '';
458
-
459
- const files: BinaryInput[] = params?.files || [];
460
-
461
- if (files.length > 0) {
462
- messages = await this.prepareMessagesWithFiles(params);
463
- } else if (params?.toolsConfig?.tools?.length > 0) {
464
- messages = await this.prepareMessagesWithTools(params);
465
- } else {
466
- messages = await this.prepareMessagesWithTextQuery(params);
467
- }
468
-
469
- return messages;
470
- }
471
-
472
- private async prepareMessagesWithFiles(params: TLLMPreparedParams): Promise<string> {
473
- const model = params.model;
474
-
475
- let messages: string | TLLMMessageBlock[] = params?.messages || '';
476
- let systemInstruction = '';
477
- const files: BinaryInput[] = params?.files || [];
478
-
479
- // #region Upload files
480
- const promises = [];
481
- const _files = [];
482
-
483
- for (let image of files) {
484
- const binaryInput = BinaryInput.from(image);
485
- promises.push(binaryInput.upload(AccessCandidate.agent(params.agentId)));
486
-
487
- _files.push(binaryInput);
488
- }
489
-
490
- await Promise.all(promises);
491
- // #endregion Upload files
492
-
493
- // If user provide mix of valid and invalid files, we will only process the valid files
494
- const validFiles = this.getValidFiles(_files, 'all');
495
-
496
- const hasVideo = validFiles.some((file) => file?.mimetype?.includes('video'));
497
-
498
- // GoogleAI only supports one video file at a time
499
- if (hasVideo && validFiles.length > 1) {
500
- throw new Error('Only one video file is supported at a time.');
501
- }
502
-
503
- const fileUploadingTasks = validFiles.map((file) => async () => {
504
- try {
505
- const uploadedFile = await this.uploadFile({
506
- file,
507
- apiKey: (params.credentials as BasicCredentials).apiKey,
508
- agentId: params.agentId,
509
- });
510
-
511
- return { url: uploadedFile.url, mimetype: file.mimetype };
512
- } catch {
513
- return null;
514
- }
515
- });
516
-
517
- const uploadedFiles = await processWithConcurrencyLimit(fileUploadingTasks);
518
-
519
- // We throw error when there are no valid uploaded files,
520
- if (uploadedFiles && uploadedFiles?.length === 0) {
521
- throw new Error(`There is an issue during upload file in Google AI Server!`);
522
- }
523
-
524
- const fileData = this.getFileData(uploadedFiles);
525
-
526
- const userMessage: TLLMMessageBlock = Array.isArray(messages) ? messages.pop() : { role: TLLMMessageRole.User, content: '' };
527
- let prompt = userMessage?.content || '';
528
-
529
- // if the the model does not support system instruction, we will add it to the prompt
530
- if (!MODELS_SUPPORT_SYSTEM_INSTRUCTION.includes(model as string)) {
531
- prompt = `${prompt}\n${systemInstruction}`;
532
- }
533
- //#endregion Separate system message and add JSON response instruction if needed
534
-
535
- // Adjust input structure handling for multiple image files to accommodate variations.
536
- messages = fileData.length === 1 ? ([...fileData, { text: prompt }] as any) : ([prompt, ...fileData] as any);
537
-
538
- return messages as string;
539
- }
540
-
541
- private async prepareMessagesWithTools(params: TLLMPreparedParams): Promise<GenerateContentRequest> {
542
- let formattedMessages: TLLMMessageBlock[];
543
- let systemInstruction = '';
544
-
545
- let messages = params?.messages || [];
546
-
547
- const hasSystemMessage = LLMHelper.hasSystemMessage(messages);
548
-
549
- if (hasSystemMessage) {
550
- const separateMessages = LLMHelper.separateSystemMessages(messages);
551
- const systemMessageContent = (separateMessages.systemMessage as TLLMMessageBlock)?.content;
552
- systemInstruction = typeof systemMessageContent === 'string' ? systemMessageContent : '';
553
- formattedMessages = separateMessages.otherMessages;
554
- } else {
555
- formattedMessages = messages;
556
- }
557
-
558
- const toolsPrompt: GenerateContentRequest = {
559
- contents: formattedMessages as any,
560
- };
561
-
562
- if (systemInstruction) {
563
- toolsPrompt.systemInstruction = systemInstruction;
564
- }
565
-
566
- if (params?.toolsConfig?.tools) toolsPrompt.tools = params?.toolsConfig?.tools as any;
567
- if (params?.toolsConfig?.tool_choice) {
568
- // Map tool choice to valid Google AI function calling modes
569
- let mode: FunctionCallingMode = FunctionCallingMode.AUTO; // default
570
- const toolChoice = params?.toolsConfig?.tool_choice;
571
-
572
- if (toolChoice === 'auto') {
573
- mode = FunctionCallingMode.AUTO;
574
- } else if (toolChoice === 'required') {
575
- mode = FunctionCallingMode.ANY;
576
- } else if (toolChoice === 'none') {
577
- mode = FunctionCallingMode.NONE;
578
- } else if (typeof toolChoice === 'object' && toolChoice.type === 'function') {
579
- // Handle OpenAI-style named tool choice - force any function call
580
- mode = FunctionCallingMode.ANY;
581
- }
582
-
583
- toolsPrompt.toolConfig = {
584
- functionCallingConfig: { mode },
585
- };
586
- }
587
-
588
- return toolsPrompt;
589
- }
590
-
591
- private async prepareMessagesWithTextQuery(params: TLLMPreparedParams): Promise<string> {
592
- const model = params.model;
593
- let systemInstruction = '';
594
- let prompt = '';
595
-
596
- const { systemMessage, otherMessages } = LLMHelper.separateSystemMessages(params?.messages as TLLMMessageBlock[]);
597
-
598
- if ('content' in systemMessage) {
599
- systemInstruction = systemMessage.content as string;
600
- }
601
-
602
- const responseFormat = params?.responseFormat || '';
603
- let responseMimeType = '';
604
-
605
- if (responseFormat === 'json') {
606
- systemInstruction += JSON_RESPONSE_INSTRUCTION;
607
-
608
- if (MODELS_SUPPORT_JSON_RESPONSE.includes(model as string)) {
609
- responseMimeType = 'application/json';
610
- }
611
- }
612
-
613
- if (otherMessages?.length > 0) {
614
- // Concatenate messages with prompt and remove messages from params as it's not supported
615
- prompt += otherMessages.map((message) => message?.parts?.[0]?.text || '').join('\n');
616
- }
617
-
618
- // if the the model does not support system instruction, we will add it to the prompt
619
- if (!MODELS_SUPPORT_SYSTEM_INSTRUCTION.includes(model as string)) {
620
- prompt = `${prompt}\n${systemInstruction}`;
621
- }
622
- //#endregion Separate system message and add JSON response instruction if needed
623
-
624
- return prompt;
625
- }
626
-
627
- private async prepareBodyForImageGenRequest(params: TLLMPreparedParams): Promise<any> {
628
- return {
629
- prompt: params.prompt,
630
- model: params.model,
631
- aspectRatio: (params as any).aspectRatio,
632
- personGeneration: (params as any).personGeneration,
633
- };
634
- }
635
-
636
- // Add this helper method to sanitize function names
637
- private sanitizeFunctionName(name: string): string {
638
- // Check if name is undefined or null
639
- if (name == null) {
640
- return '_unnamed_function';
641
- }
642
-
643
- // Remove any characters that are not alphanumeric, underscore, dot, or dash
644
- let sanitized = name.replace(/[^a-zA-Z0-9_.-]/g, '');
645
-
646
- // Ensure the name starts with a letter or underscore
647
- if (!/^[a-zA-Z_]/.test(sanitized)) {
648
- sanitized = '_' + sanitized;
649
- }
650
-
651
- // If sanitized is empty after removing invalid characters, use a default name
652
- if (sanitized === '') {
653
- sanitized = '_unnamed_function';
654
- }
655
-
656
- // Truncate to 64 characters if longer
657
- sanitized = sanitized.slice(0, 64);
658
-
659
- return sanitized;
660
- }
661
-
662
- private async uploadFile({ file, apiKey, agentId }: { file: BinaryInput; apiKey: string; agentId: string }): Promise<{ url: string }> {
663
- try {
664
- if (!apiKey || !file?.mimetype) {
665
- throw new Error('Missing required parameters to save file for Google AI!');
666
- }
667
-
668
- // Create a temporary directory
669
- const tempDir = os.tmpdir();
670
- const fileName = uid();
671
- const tempFilePath = path.join(tempDir, fileName);
672
-
673
- const bufferData = await file.readData(AccessCandidate.agent(agentId));
674
-
675
- // Write buffer data to temp file
676
- await fs.promises.writeFile(tempFilePath, new Uint8Array(bufferData));
677
-
678
- // Upload the file to the Google File Manager
679
- const fileManager = new GoogleAIFileManager(apiKey);
680
-
681
- const uploadResponse = await fileManager.uploadFile(tempFilePath, {
682
- mimeType: file.mimetype,
683
- displayName: fileName,
684
- });
685
-
686
- const name = uploadResponse.file.name;
687
-
688
- // Poll getFile() on a set interval (10 seconds here) to check file state.
689
- let uploadedFile = await fileManager.getFile(name);
690
- while (uploadedFile.state === FileState.PROCESSING) {
691
- process.stdout.write('.');
692
- // Sleep for 10 seconds
693
- await new Promise((resolve) => setTimeout(resolve, 10_000));
694
- // Fetch the file from the API again
695
- uploadedFile = await fileManager.getFile(name);
696
- }
697
-
698
- if (uploadedFile.state === FileState.FAILED) {
699
- throw new Error('File processing failed.');
700
- }
701
-
702
- // Clean up temp file
703
- await fs.promises.unlink(tempFilePath);
704
-
705
- return {
706
- url: uploadResponse.file.uri || '',
707
- };
708
- } catch (error) {
709
- throw new Error(`Error uploading file for Google AI: ${error.message}`);
710
- }
711
- }
712
-
713
- private getValidFiles(files: BinaryInput[], type: 'image' | 'all') {
714
- const validSources = [];
715
-
716
- for (let file of files) {
717
- if (this.validMimeTypes[type].includes(file?.mimetype)) {
718
- validSources.push(file);
719
- }
720
- }
721
-
722
- if (validSources?.length === 0) {
723
- throw new Error(`Unsupported file(s). Please make sure your file is one of the following types: ${this.validMimeTypes[type].join(', ')}`);
724
- }
725
-
726
- return validSources;
727
- }
728
-
729
- private getFileData(
730
- files: {
731
- url: string;
732
- mimetype: string;
733
- }[]
734
- ): {
735
- fileData: {
736
- mimeType: string;
737
- fileUri: string;
738
- };
739
- }[] {
740
- try {
741
- const imageData = [];
742
-
743
- for (let file of files) {
744
- imageData.push({
745
- fileData: {
746
- mimeType: file.mimetype,
747
- fileUri: file.url,
748
- },
749
- });
750
- }
751
-
752
- return imageData;
753
- } catch (error) {
754
- throw error;
755
- }
756
- }
757
- }
1
+ import os from 'os';
2
+ import path from 'path';
3
+ import EventEmitter from 'events';
4
+ import fs from 'fs';
5
+
6
+ import { GoogleGenerativeAI, ModelParams, GenerationConfig, GenerateContentRequest, UsageMetadata, FunctionCallingMode } from '@google/generative-ai';
7
+ import { GoogleAIFileManager, FileState } from '@google/generative-ai/server';
8
+ import { GoogleGenAI } from '@google/genai';
9
+
10
+ import { JSON_RESPONSE_INSTRUCTION, BUILT_IN_MODEL_PREFIX } from '@sre/constants';
11
+ import { BinaryInput } from '@sre/helpers/BinaryInput.helper';
12
+ import { AccessCandidate } from '@sre/Security/AccessControl/AccessCandidate.class';
13
+ import { uid } from '@sre/utils';
14
+
15
+ import { processWithConcurrencyLimit } from '@sre/utils';
16
+
17
+ import {
18
+ TLLMMessageBlock,
19
+ ToolData,
20
+ TLLMMessageRole,
21
+ TLLMToolResultMessageBlock,
22
+ APIKeySource,
23
+ TLLMEvent,
24
+ BasicCredentials,
25
+ ILLMRequestFuncParams,
26
+ TLLMChatResponse,
27
+ TGoogleAIRequestBody,
28
+ ILLMRequestContext,
29
+ TLLMPreparedParams,
30
+ } from '@sre/types/LLM.types';
31
+ import { LLMHelper } from '@sre/LLMManager/LLM.helper';
32
+
33
+ import { SystemEvents } from '@sre/Core/SystemEvents';
34
+ import { SUPPORTED_MIME_TYPES_MAP } from '@sre/constants';
35
+
36
+ import { LLMConnector } from '../LLMConnector';
37
+
38
+ const MODELS_SUPPORT_SYSTEM_INSTRUCTION = [
39
+ 'gemini-1.5-pro-exp-0801',
40
+ 'gemini-1.5-pro-latest',
41
+ 'gemini-1.5-pro-latest',
42
+ 'gemini-1.5-pro',
43
+ 'gemini-1.5-pro-001',
44
+ 'gemini-1.5-flash-latest',
45
+ 'gemini-1.5-flash-001',
46
+ 'gemini-1.5-flash',
47
+ ];
48
+ const MODELS_SUPPORT_JSON_RESPONSE = MODELS_SUPPORT_SYSTEM_INSTRUCTION;
49
+
50
+ // Supported file MIME types for Google AI's Gemini models
51
+ const VALID_MIME_TYPES = [
52
+ ...SUPPORTED_MIME_TYPES_MAP.GoogleAI.image,
53
+ ...SUPPORTED_MIME_TYPES_MAP.GoogleAI.audio,
54
+ ...SUPPORTED_MIME_TYPES_MAP.GoogleAI.video,
55
+ ...SUPPORTED_MIME_TYPES_MAP.GoogleAI.document,
56
+ ];
57
+
58
+ // will be removed after updating the SDK
59
+ type UsageMetadataWithThoughtsToken = UsageMetadata & { thoughtsTokenCount: number };
60
+
61
+ export class GoogleAIConnector extends LLMConnector {
62
+ public name = 'LLM:GoogleAI';
63
+
64
+ private validMimeTypes = {
65
+ all: VALID_MIME_TYPES,
66
+ image: SUPPORTED_MIME_TYPES_MAP.GoogleAI.image,
67
+ };
68
+
69
+ private async getClient(params: ILLMRequestContext): Promise<GoogleGenerativeAI> {
70
+ const apiKey = (params.credentials as BasicCredentials)?.apiKey;
71
+
72
+ if (!apiKey) throw new Error('Please provide an API key for Google AI');
73
+
74
+ return new GoogleGenerativeAI(apiKey);
75
+ }
76
+
77
+ protected async request({ acRequest, body, context }: ILLMRequestFuncParams): Promise<TLLMChatResponse> {
78
+ try {
79
+ const prompt = body.messages;
80
+ delete body.messages;
81
+
82
+ const genAI = await this.getClient(context);
83
+ const $model = genAI.getGenerativeModel(body);
84
+
85
+ const result = await $model.generateContent(prompt);
86
+
87
+ const response = await result.response;
88
+ const content = response.text();
89
+ const finishReason = response.candidates[0].finishReason || 'stop';
90
+ const usage = response?.usageMetadata as UsageMetadataWithThoughtsToken;
91
+ this.reportUsage(usage, {
92
+ modelEntryName: context.modelEntryName,
93
+ keySource: context.isUserKey ? APIKeySource.User : APIKeySource.Smyth,
94
+ agentId: context.agentId,
95
+ teamId: context.teamId,
96
+ });
97
+
98
+ const toolCalls = response.candidates[0]?.content?.parts?.filter((part) => part.functionCall);
99
+
100
+ let toolsData: ToolData[] = [];
101
+ let useTool = false;
102
+
103
+ if (toolCalls && toolCalls.length > 0) {
104
+ toolsData = toolCalls.map((toolCall, index) => ({
105
+ index,
106
+ id: `tool-${index}`,
107
+ type: 'function',
108
+ name: toolCall.functionCall.name,
109
+ arguments: JSON.stringify(toolCall.functionCall.args),
110
+ role: TLLMMessageRole.Assistant,
111
+ }));
112
+ useTool = true;
113
+ }
114
+
115
+ return {
116
+ content,
117
+ finishReason: finishReason.toLowerCase(),
118
+ useTool,
119
+ toolsData,
120
+ message: { content, role: 'assistant' },
121
+ usage,
122
+ };
123
+ } catch (error: any) {
124
+ throw error;
125
+ }
126
+ }
127
+
128
+ protected async streamRequest({ acRequest, body, context }: ILLMRequestFuncParams): Promise<EventEmitter> {
129
+ const emitter = new EventEmitter();
130
+
131
+ const prompt = body.messages;
132
+ delete body.messages;
133
+
134
+ const genAI = await this.getClient(context);
135
+ const $model = genAI.getGenerativeModel(body);
136
+
137
+ try {
138
+ const result = await $model.generateContentStream(prompt);
139
+
140
+ let toolsData: ToolData[] = [];
141
+ let usage: UsageMetadataWithThoughtsToken;
142
+
143
+ // Process stream asynchronously while as we need to return emitter immediately
144
+ (async () => {
145
+ for await (const chunk of result.stream) {
146
+ const chunkText = chunk.text();
147
+ emitter.emit('content', chunkText);
148
+
149
+ if (chunk.candidates[0]?.content?.parts) {
150
+ const toolCalls = chunk.candidates[0].content.parts.filter((part) => part.functionCall);
151
+ if (toolCalls.length > 0) {
152
+ toolsData = toolCalls.map((toolCall, index) => ({
153
+ index,
154
+ id: `tool-${index}`,
155
+ type: 'function',
156
+ name: toolCall.functionCall.name,
157
+ arguments: JSON.stringify(toolCall.functionCall.args),
158
+ role: TLLMMessageRole.Assistant,
159
+ }));
160
+ emitter.emit(TLLMEvent.ToolInfo, toolsData);
161
+ }
162
+ }
163
+
164
+ // the same usage is sent on each emit. IMPORTANT: google does not send usage for each chunk but
165
+ // rather just sends the same usage for the entire request.
166
+ // notice that the output tokens are only sent in the last chunk usage metadata.
167
+ // so we will just update a var to hold the latest usage and report it when the stream ends.
168
+ // e.g emit1: { input_tokens: 500, output_tokens: undefined } -> same input_tokens
169
+ // e.g emit2: { input_tokens: 500, output_tokens: undefined } -> same input_tokens
170
+ // e.g emit3: { input_tokens: 500, output_tokens: 10 } -> same input_tokens, new output_tokens in the last chunk
171
+ if (chunk?.usageMetadata) {
172
+ usage = chunk.usageMetadata as UsageMetadataWithThoughtsToken;
173
+ }
174
+ }
175
+
176
+ if (usage) {
177
+ this.reportUsage(usage, {
178
+ modelEntryName: context.modelEntryName,
179
+ keySource: context.isUserKey ? APIKeySource.User : APIKeySource.Smyth,
180
+ agentId: context.agentId,
181
+ teamId: context.teamId,
182
+ });
183
+ }
184
+
185
+ setTimeout(() => {
186
+ emitter.emit('end', toolsData);
187
+ }, 100);
188
+ })();
189
+
190
+ return emitter;
191
+ } catch (error: any) {
192
+ throw error;
193
+ }
194
+ }
195
+ // #region Image Generation, will be moved to a different subsystem/service
196
+ protected async imageGenRequest({ body, context }: ILLMRequestFuncParams): Promise<any> {
197
+ try {
198
+ const apiKey = (context.credentials as BasicCredentials)?.apiKey;
199
+ if (!apiKey) throw new Error('Please provide an API key for Google AI');
200
+
201
+ const model = body.model || 'imagen-3.0-generate-001';
202
+
203
+ // Use Imagen models via GoogleGenAI
204
+ const ai = new GoogleGenAI({ apiKey });
205
+
206
+ // Prepare the configuration for image generation
207
+ const config = {
208
+ numberOfImages: body.n || 1,
209
+ aspectRatio: body.aspect_ratio || body.size || '1:1',
210
+ personGeneration: body.person_generation || 'allow_adult',
211
+ };
212
+
213
+ // Generate images using the SDK
214
+ const response = await ai.models.generateImages({
215
+ model,
216
+ prompt: body.prompt,
217
+ config,
218
+ });
219
+
220
+ // Transform the response to match OpenAI format for compatibility
221
+ return {
222
+ created: Math.floor(Date.now() / 1000),
223
+ data:
224
+ response.generatedImages?.map((generatedImage: any) => ({
225
+ url: generatedImage.image.imageBytes ? `data:image/png;base64,${generatedImage.image.imageBytes}` : undefined,
226
+ b64_json: generatedImage.image.imageBytes,
227
+ revised_prompt: body.prompt,
228
+ })) || [],
229
+ };
230
+ } catch (error: any) {
231
+ throw error;
232
+ }
233
+ }
234
+
235
+ protected async imageEditRequest({ body, context }: ILLMRequestFuncParams): Promise<any> {
236
+ throw new Error('Image editing is not supported for Google AI. Imagen models only support image generation.');
237
+ }
238
+
239
+ protected async reqBodyAdapter(params: TLLMPreparedParams): Promise<TGoogleAIRequestBody> {
240
+ const model = params?.model;
241
+
242
+ // Check if this is an image generation request based on capabilities
243
+ if (params?.capabilities?.imageGeneration) {
244
+ return this.prepareBodyForImageGenRequest(params) as any;
245
+ }
246
+
247
+ const messages = await this.prepareMessages(params);
248
+
249
+ let body: ModelParams & { messages: string | TLLMMessageBlock[] | GenerateContentRequest } = {
250
+ model: model as string,
251
+ messages,
252
+ };
253
+
254
+ const responseFormat = params?.responseFormat || '';
255
+ let responseMimeType = '';
256
+ let systemInstruction = '';
257
+
258
+ if (responseFormat === 'json') {
259
+ systemInstruction += JSON_RESPONSE_INSTRUCTION;
260
+
261
+ if (MODELS_SUPPORT_JSON_RESPONSE.includes(model as string)) {
262
+ responseMimeType = 'application/json';
263
+ }
264
+ }
265
+
266
+ const config: GenerationConfig = {};
267
+
268
+ if (params.maxTokens !== undefined) config.maxOutputTokens = params.maxTokens;
269
+ if (params.temperature !== undefined) config.temperature = params.temperature;
270
+ if (params.topP !== undefined) config.topP = params.topP;
271
+ if (params.topK !== undefined) config.topK = params.topK;
272
+ if (params.stopSequences?.length) config.stopSequences = params.stopSequences;
273
+ if (responseMimeType) config.responseMimeType = responseMimeType;
274
+
275
+ if (systemInstruction) body.systemInstruction = systemInstruction;
276
+ if (Object.keys(config).length > 0) {
277
+ body.generationConfig = config;
278
+ }
279
+
280
+ return body;
281
+ }
282
+
283
+ protected reportUsage(
284
+ usage: UsageMetadataWithThoughtsToken,
285
+ metadata: { modelEntryName: string; keySource: APIKeySource; agentId: string; teamId: string }
286
+ ) {
287
+ const modelEntryName = metadata.modelEntryName;
288
+ let tier = '';
289
+
290
+ const tierThresholds = {
291
+ 'gemini-1.5-pro': 128_000,
292
+ 'gemini-2.5-pro': 200_000,
293
+ };
294
+
295
+ const textInputTokens =
296
+ usage?.['promptTokensDetails']?.find((detail) => detail.modality === 'TEXT')?.tokenCount || usage?.promptTokenCount || 0;
297
+ const audioInputTokens = usage?.['promptTokensDetails']?.find((detail) => detail.modality === 'AUDIO')?.tokenCount || 0;
298
+
299
+ // Find matching model and set tier based on threshold
300
+ const modelWithTier = Object.keys(tierThresholds).find((model) => modelEntryName.includes(model));
301
+ if (modelWithTier) {
302
+ tier = textInputTokens < tierThresholds[modelWithTier] ? 'tier1' : 'tier2';
303
+ }
304
+
305
+ // #endregion
306
+
307
+ // SmythOS (built-in) models have a prefix, so we need to remove it to get the model name
308
+ const modelName = metadata.modelEntryName.replace(BUILT_IN_MODEL_PREFIX, '');
309
+
310
+ const usageData = {
311
+ sourceId: `llm:${modelName}`,
312
+ input_tokens: textInputTokens,
313
+ output_tokens: usage.candidatesTokenCount,
314
+ input_tokens_audio: audioInputTokens,
315
+ input_tokens_cache_read: usage.cachedContentTokenCount || 0,
316
+ input_tokens_cache_write: 0,
317
+ reasoning_tokens: usage.thoughtsTokenCount,
318
+ keySource: metadata.keySource,
319
+ agentId: metadata.agentId,
320
+ teamId: metadata.teamId,
321
+ tier,
322
+ };
323
+ SystemEvents.emit('USAGE:LLM', usageData);
324
+
325
+ return usageData;
326
+ }
327
+
328
+ public formatToolsConfig({ toolDefinitions, toolChoice = 'auto' }) {
329
+ const tools = toolDefinitions.map((tool) => {
330
+ const { name, description, properties, requiredFields } = tool;
331
+
332
+ // Ensure the function name is valid
333
+ const validName = this.sanitizeFunctionName(name);
334
+
335
+ // Ensure properties are non-empty for OBJECT type
336
+ const validProperties = properties && Object.keys(properties).length > 0 ? properties : { dummy: { type: 'string' } };
337
+
338
+ return {
339
+ functionDeclarations: [
340
+ {
341
+ name: validName,
342
+ description: description || '',
343
+ parameters: {
344
+ type: 'OBJECT',
345
+ properties: validProperties,
346
+ required: requiredFields || [],
347
+ },
348
+ },
349
+ ],
350
+ };
351
+ });
352
+
353
+ return {
354
+ tools,
355
+ toolChoice: {
356
+ type: toolChoice,
357
+ },
358
+ };
359
+ }
360
+
361
+ public transformToolMessageBlocks({
362
+ messageBlock,
363
+ toolsData,
364
+ }: {
365
+ messageBlock: TLLMMessageBlock;
366
+ toolsData: ToolData[];
367
+ }): TLLMToolResultMessageBlock[] {
368
+ const messageBlocks: TLLMToolResultMessageBlock[] = [];
369
+
370
+ if (messageBlock) {
371
+ const content = [];
372
+ if (typeof messageBlock.content === 'string') {
373
+ content.push({ text: messageBlock.content });
374
+ } else if (Array.isArray(messageBlock.content)) {
375
+ content.push(...messageBlock.content);
376
+ }
377
+
378
+ if (messageBlock.parts) {
379
+ const functionCalls = messageBlock.parts.filter((part) => part.functionCall);
380
+ if (functionCalls.length > 0) {
381
+ content.push(
382
+ ...functionCalls.map((call) => ({
383
+ functionCall: {
384
+ name: call.functionCall.name,
385
+ args: JSON.parse(call.functionCall.args),
386
+ },
387
+ }))
388
+ );
389
+ }
390
+ }
391
+
392
+ messageBlocks.push({
393
+ role: messageBlock.role,
394
+ parts: content,
395
+ });
396
+ }
397
+
398
+ const transformedToolsData = toolsData.map(
399
+ (toolData): TLLMToolResultMessageBlock => ({
400
+ role: TLLMMessageRole.User,
401
+ parts: [
402
+ {
403
+ functionResponse: {
404
+ name: toolData.name,
405
+ response: {
406
+ name: toolData.name,
407
+ content: typeof toolData.result === 'string' ? toolData.result : JSON.stringify(toolData.result),
408
+ },
409
+ },
410
+ },
411
+ ],
412
+ })
413
+ );
414
+
415
+ return [...messageBlocks, ...transformedToolsData];
416
+ }
417
+
418
+ public getConsistentMessages(messages: TLLMMessageBlock[]): TLLMMessageBlock[] {
419
+ const _messages = LLMHelper.removeDuplicateUserMessages(messages);
420
+
421
+ return _messages.map((message) => {
422
+ const _message = { ...message };
423
+ let textContent = '';
424
+
425
+ // Map roles to valid Google AI roles
426
+ switch (_message.role) {
427
+ case TLLMMessageRole.Assistant:
428
+ case TLLMMessageRole.System:
429
+ _message.role = TLLMMessageRole.Model;
430
+ break;
431
+ case TLLMMessageRole.User:
432
+ // User role is already valid
433
+ break;
434
+ default:
435
+ _message.role = TLLMMessageRole.User; // Default to user for unknown roles
436
+ }
437
+
438
+ // * empty text causes error that's why we added '...'
439
+
440
+ if (_message?.parts) {
441
+ textContent = _message.parts.map((textBlock) => textBlock?.text || '...').join(' ');
442
+ } else if (Array.isArray(_message?.content)) {
443
+ textContent = _message.content.map((textBlock) => textBlock?.text || '...').join(' ');
444
+ } else if (_message?.content) {
445
+ textContent = (_message.content as string) || '...';
446
+ }
447
+
448
+ _message.parts = [{ text: textContent || '...' }];
449
+
450
+ delete _message.content; // Remove content to avoid error
451
+
452
+ return _message;
453
+ });
454
+ }
455
+
456
+ private async prepareMessages(params: TLLMPreparedParams): Promise<string | TLLMMessageBlock[] | GenerateContentRequest> {
457
+ let messages: string | TLLMMessageBlock[] | GenerateContentRequest = params?.messages || '';
458
+
459
+ const files: BinaryInput[] = params?.files || [];
460
+
461
+ if (files.length > 0) {
462
+ messages = await this.prepareMessagesWithFiles(params);
463
+ } else if (params?.toolsConfig?.tools?.length > 0) {
464
+ messages = await this.prepareMessagesWithTools(params);
465
+ } else {
466
+ messages = await this.prepareMessagesWithTextQuery(params);
467
+ }
468
+
469
+ return messages;
470
+ }
471
+
472
+ private async prepareMessagesWithFiles(params: TLLMPreparedParams): Promise<string> {
473
+ const model = params.model;
474
+
475
+ let messages: string | TLLMMessageBlock[] = params?.messages || '';
476
+ let systemInstruction = '';
477
+ const files: BinaryInput[] = params?.files || [];
478
+
479
+ // #region Upload files
480
+ const promises = [];
481
+ const _files = [];
482
+
483
+ for (let image of files) {
484
+ const binaryInput = BinaryInput.from(image);
485
+ promises.push(binaryInput.upload(AccessCandidate.agent(params.agentId)));
486
+
487
+ _files.push(binaryInput);
488
+ }
489
+
490
+ await Promise.all(promises);
491
+ // #endregion Upload files
492
+
493
+ // If user provide mix of valid and invalid files, we will only process the valid files
494
+ const validFiles = this.getValidFiles(_files, 'all');
495
+
496
+ const hasVideo = validFiles.some((file) => file?.mimetype?.includes('video'));
497
+
498
+ // GoogleAI only supports one video file at a time
499
+ if (hasVideo && validFiles.length > 1) {
500
+ throw new Error('Only one video file is supported at a time.');
501
+ }
502
+
503
+ const fileUploadingTasks = validFiles.map((file) => async () => {
504
+ try {
505
+ const uploadedFile = await this.uploadFile({
506
+ file,
507
+ apiKey: (params.credentials as BasicCredentials).apiKey,
508
+ agentId: params.agentId,
509
+ });
510
+
511
+ return { url: uploadedFile.url, mimetype: file.mimetype };
512
+ } catch {
513
+ return null;
514
+ }
515
+ });
516
+
517
+ const uploadedFiles = await processWithConcurrencyLimit(fileUploadingTasks);
518
+
519
+ // We throw error when there are no valid uploaded files,
520
+ if (uploadedFiles && uploadedFiles?.length === 0) {
521
+ throw new Error(`There is an issue during upload file in Google AI Server!`);
522
+ }
523
+
524
+ const fileData = this.getFileData(uploadedFiles);
525
+
526
+ const userMessage: TLLMMessageBlock = Array.isArray(messages) ? messages.pop() : { role: TLLMMessageRole.User, content: '' };
527
+ let prompt = userMessage?.content || '';
528
+
529
+ // if the the model does not support system instruction, we will add it to the prompt
530
+ if (!MODELS_SUPPORT_SYSTEM_INSTRUCTION.includes(model as string)) {
531
+ prompt = `${prompt}\n${systemInstruction}`;
532
+ }
533
+ //#endregion Separate system message and add JSON response instruction if needed
534
+
535
+ // Adjust input structure handling for multiple image files to accommodate variations.
536
+ messages = fileData.length === 1 ? ([...fileData, { text: prompt }] as any) : ([prompt, ...fileData] as any);
537
+
538
+ return messages as string;
539
+ }
540
+
541
+ private async prepareMessagesWithTools(params: TLLMPreparedParams): Promise<GenerateContentRequest> {
542
+ let formattedMessages: TLLMMessageBlock[];
543
+ let systemInstruction = '';
544
+
545
+ let messages = params?.messages || [];
546
+
547
+ const hasSystemMessage = LLMHelper.hasSystemMessage(messages);
548
+
549
+ if (hasSystemMessage) {
550
+ const separateMessages = LLMHelper.separateSystemMessages(messages);
551
+ const systemMessageContent = (separateMessages.systemMessage as TLLMMessageBlock)?.content;
552
+ systemInstruction = typeof systemMessageContent === 'string' ? systemMessageContent : '';
553
+ formattedMessages = separateMessages.otherMessages;
554
+ } else {
555
+ formattedMessages = messages;
556
+ }
557
+
558
+ const toolsPrompt: GenerateContentRequest = {
559
+ contents: formattedMessages as any,
560
+ };
561
+
562
+ if (systemInstruction) {
563
+ toolsPrompt.systemInstruction = systemInstruction;
564
+ }
565
+
566
+ if (params?.toolsConfig?.tools) toolsPrompt.tools = params?.toolsConfig?.tools as any;
567
+ if (params?.toolsConfig?.tool_choice) {
568
+ // Map tool choice to valid Google AI function calling modes
569
+ let mode: FunctionCallingMode = FunctionCallingMode.AUTO; // default
570
+ const toolChoice = params?.toolsConfig?.tool_choice;
571
+
572
+ if (toolChoice === 'auto') {
573
+ mode = FunctionCallingMode.AUTO;
574
+ } else if (toolChoice === 'required') {
575
+ mode = FunctionCallingMode.ANY;
576
+ } else if (toolChoice === 'none') {
577
+ mode = FunctionCallingMode.NONE;
578
+ } else if (typeof toolChoice === 'object' && toolChoice.type === 'function') {
579
+ // Handle OpenAI-style named tool choice - force any function call
580
+ mode = FunctionCallingMode.ANY;
581
+ }
582
+
583
+ toolsPrompt.toolConfig = {
584
+ functionCallingConfig: { mode },
585
+ };
586
+ }
587
+
588
+ return toolsPrompt;
589
+ }
590
+
591
+ private async prepareMessagesWithTextQuery(params: TLLMPreparedParams): Promise<string> {
592
+ const model = params.model;
593
+ let systemInstruction = '';
594
+ let prompt = '';
595
+
596
+ const { systemMessage, otherMessages } = LLMHelper.separateSystemMessages(params?.messages as TLLMMessageBlock[]);
597
+
598
+ if ('content' in systemMessage) {
599
+ systemInstruction = systemMessage.content as string;
600
+ }
601
+
602
+ const responseFormat = params?.responseFormat || '';
603
+ let responseMimeType = '';
604
+
605
+ if (responseFormat === 'json') {
606
+ systemInstruction += JSON_RESPONSE_INSTRUCTION;
607
+
608
+ if (MODELS_SUPPORT_JSON_RESPONSE.includes(model as string)) {
609
+ responseMimeType = 'application/json';
610
+ }
611
+ }
612
+
613
+ if (otherMessages?.length > 0) {
614
+ // Concatenate messages with prompt and remove messages from params as it's not supported
615
+ prompt += otherMessages.map((message) => message?.parts?.[0]?.text || '').join('\n');
616
+ }
617
+
618
+ // if the the model does not support system instruction, we will add it to the prompt
619
+ if (!MODELS_SUPPORT_SYSTEM_INSTRUCTION.includes(model as string)) {
620
+ prompt = `${prompt}\n${systemInstruction}`;
621
+ }
622
+ //#endregion Separate system message and add JSON response instruction if needed
623
+
624
+ return prompt;
625
+ }
626
+
627
+ private async prepareBodyForImageGenRequest(params: TLLMPreparedParams): Promise<any> {
628
+ return {
629
+ prompt: params.prompt,
630
+ model: params.model,
631
+ aspectRatio: (params as any).aspectRatio,
632
+ personGeneration: (params as any).personGeneration,
633
+ };
634
+ }
635
+
636
+ // Add this helper method to sanitize function names
637
+ private sanitizeFunctionName(name: string): string {
638
+ // Check if name is undefined or null
639
+ if (name == null) {
640
+ return '_unnamed_function';
641
+ }
642
+
643
+ // Remove any characters that are not alphanumeric, underscore, dot, or dash
644
+ let sanitized = name.replace(/[^a-zA-Z0-9_.-]/g, '');
645
+
646
+ // Ensure the name starts with a letter or underscore
647
+ if (!/^[a-zA-Z_]/.test(sanitized)) {
648
+ sanitized = '_' + sanitized;
649
+ }
650
+
651
+ // If sanitized is empty after removing invalid characters, use a default name
652
+ if (sanitized === '') {
653
+ sanitized = '_unnamed_function';
654
+ }
655
+
656
+ // Truncate to 64 characters if longer
657
+ sanitized = sanitized.slice(0, 64);
658
+
659
+ return sanitized;
660
+ }
661
+
662
+ private async uploadFile({ file, apiKey, agentId }: { file: BinaryInput; apiKey: string; agentId: string }): Promise<{ url: string }> {
663
+ try {
664
+ if (!apiKey || !file?.mimetype) {
665
+ throw new Error('Missing required parameters to save file for Google AI!');
666
+ }
667
+
668
+ // Create a temporary directory
669
+ const tempDir = os.tmpdir();
670
+ const fileName = uid();
671
+ const tempFilePath = path.join(tempDir, fileName);
672
+
673
+ const bufferData = await file.readData(AccessCandidate.agent(agentId));
674
+
675
+ // Write buffer data to temp file
676
+ await fs.promises.writeFile(tempFilePath, new Uint8Array(bufferData));
677
+
678
+ // Upload the file to the Google File Manager
679
+ const fileManager = new GoogleAIFileManager(apiKey);
680
+
681
+ const uploadResponse = await fileManager.uploadFile(tempFilePath, {
682
+ mimeType: file.mimetype,
683
+ displayName: fileName,
684
+ });
685
+
686
+ const name = uploadResponse.file.name;
687
+
688
+ // Poll getFile() on a set interval (10 seconds here) to check file state.
689
+ let uploadedFile = await fileManager.getFile(name);
690
+ while (uploadedFile.state === FileState.PROCESSING) {
691
+ process.stdout.write('.');
692
+ // Sleep for 10 seconds
693
+ await new Promise((resolve) => setTimeout(resolve, 10_000));
694
+ // Fetch the file from the API again
695
+ uploadedFile = await fileManager.getFile(name);
696
+ }
697
+
698
+ if (uploadedFile.state === FileState.FAILED) {
699
+ throw new Error('File processing failed.');
700
+ }
701
+
702
+ // Clean up temp file
703
+ await fs.promises.unlink(tempFilePath);
704
+
705
+ return {
706
+ url: uploadResponse.file.uri || '',
707
+ };
708
+ } catch (error) {
709
+ throw new Error(`Error uploading file for Google AI: ${error.message}`);
710
+ }
711
+ }
712
+
713
+ private getValidFiles(files: BinaryInput[], type: 'image' | 'all') {
714
+ const validSources = [];
715
+
716
+ for (let file of files) {
717
+ if (this.validMimeTypes[type].includes(file?.mimetype)) {
718
+ validSources.push(file);
719
+ }
720
+ }
721
+
722
+ if (validSources?.length === 0) {
723
+ throw new Error(`Unsupported file(s). Please make sure your file is one of the following types: ${this.validMimeTypes[type].join(', ')}`);
724
+ }
725
+
726
+ return validSources;
727
+ }
728
+
729
+ private getFileData(
730
+ files: {
731
+ url: string;
732
+ mimetype: string;
733
+ }[]
734
+ ): {
735
+ fileData: {
736
+ mimeType: string;
737
+ fileUri: string;
738
+ };
739
+ }[] {
740
+ try {
741
+ const imageData = [];
742
+
743
+ for (let file of files) {
744
+ imageData.push({
745
+ fileData: {
746
+ mimeType: file.mimetype,
747
+ fileUri: file.url,
748
+ },
749
+ });
750
+ }
751
+
752
+ return imageData;
753
+ } catch (error) {
754
+ throw error;
755
+ }
756
+ }
757
+ }