@llumiverse/drivers 0.22.2 → 0.22.3

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (225) hide show
  1. package/lib/cjs/adobe/firefly.js +2 -2
  2. package/lib/cjs/adobe/firefly.js.map +1 -1
  3. package/lib/cjs/azure/azure_foundry.js +11 -11
  4. package/lib/cjs/azure/azure_foundry.js.map +1 -1
  5. package/lib/cjs/bedrock/index.js +7 -7
  6. package/lib/cjs/bedrock/index.js.map +1 -1
  7. package/lib/cjs/groq/index.js +2 -2
  8. package/lib/cjs/groq/index.js.map +1 -1
  9. package/lib/cjs/huggingface_ie.js +3 -4
  10. package/lib/cjs/huggingface_ie.js.map +1 -1
  11. package/lib/cjs/index.js +2 -2
  12. package/lib/cjs/index.js.map +1 -1
  13. package/lib/cjs/mistral/index.js +2 -2
  14. package/lib/cjs/mistral/index.js.map +1 -1
  15. package/lib/cjs/openai/azure_openai.js +1 -1
  16. package/lib/cjs/openai/azure_openai.js.map +1 -1
  17. package/lib/cjs/openai/index.js +3 -3
  18. package/lib/cjs/openai/index.js.map +1 -1
  19. package/lib/cjs/replicate.js +4 -4
  20. package/lib/cjs/replicate.js.map +1 -1
  21. package/lib/cjs/togetherai/index.js +2 -2
  22. package/lib/cjs/togetherai/index.js.map +1 -1
  23. package/lib/cjs/vertexai/index.js +82 -40
  24. package/lib/cjs/vertexai/index.js.map +1 -1
  25. package/lib/cjs/vertexai/models/claude.js +4 -4
  26. package/lib/cjs/vertexai/models/claude.js.map +1 -1
  27. package/lib/cjs/vertexai/models/gemini.js +15 -5
  28. package/lib/cjs/vertexai/models/gemini.js.map +1 -1
  29. package/lib/cjs/vertexai/models/imagen.js +2 -5
  30. package/lib/cjs/vertexai/models/imagen.js.map +1 -1
  31. package/lib/cjs/watsonx/index.js +3 -3
  32. package/lib/cjs/watsonx/index.js.map +1 -1
  33. package/lib/esm/adobe/firefly.js +2 -2
  34. package/lib/esm/adobe/firefly.js.map +1 -1
  35. package/lib/esm/azure/azure_foundry.js +11 -11
  36. package/lib/esm/azure/azure_foundry.js.map +1 -1
  37. package/lib/esm/bedrock/index.js +8 -8
  38. package/lib/esm/bedrock/index.js.map +1 -1
  39. package/lib/esm/groq/index.js +2 -2
  40. package/lib/esm/groq/index.js.map +1 -1
  41. package/lib/esm/huggingface_ie.js +4 -5
  42. package/lib/esm/huggingface_ie.js.map +1 -1
  43. package/lib/esm/index.js +2 -2
  44. package/lib/esm/index.js.map +1 -1
  45. package/lib/esm/mistral/index.js +2 -2
  46. package/lib/esm/mistral/index.js.map +1 -1
  47. package/lib/esm/openai/azure_openai.js +1 -1
  48. package/lib/esm/openai/azure_openai.js.map +1 -1
  49. package/lib/esm/openai/index.js +3 -3
  50. package/lib/esm/openai/index.js.map +1 -1
  51. package/lib/esm/replicate.js +4 -4
  52. package/lib/esm/replicate.js.map +1 -1
  53. package/lib/esm/togetherai/index.js +2 -2
  54. package/lib/esm/togetherai/index.js.map +1 -1
  55. package/lib/esm/vertexai/index.js +83 -41
  56. package/lib/esm/vertexai/index.js.map +1 -1
  57. package/lib/esm/vertexai/models/claude.js +4 -4
  58. package/lib/esm/vertexai/models/claude.js.map +1 -1
  59. package/lib/esm/vertexai/models/gemini.js +15 -5
  60. package/lib/esm/vertexai/models/gemini.js.map +1 -1
  61. package/lib/esm/vertexai/models/imagen.js +3 -6
  62. package/lib/esm/vertexai/models/imagen.js.map +1 -1
  63. package/lib/esm/watsonx/index.js +3 -3
  64. package/lib/esm/watsonx/index.js.map +1 -1
  65. package/lib/types/bedrock/index.d.ts +1 -0
  66. package/lib/types/bedrock/index.d.ts.map +1 -1
  67. package/lib/types/huggingface_ie.d.ts +3 -3
  68. package/lib/types/huggingface_ie.d.ts.map +1 -1
  69. package/lib/types/index.d.ts +2 -2
  70. package/lib/types/index.d.ts.map +1 -1
  71. package/lib/types/replicate.d.ts.map +1 -1
  72. package/lib/types/vertexai/index.d.ts +12 -11
  73. package/lib/types/vertexai/index.d.ts.map +1 -1
  74. package/lib/types/vertexai/models/gemini.d.ts.map +1 -1
  75. package/lib/types/vertexai/models/imagen.d.ts.map +1 -1
  76. package/package.json +3 -3
  77. package/src/bedrock/index.ts +5 -4
  78. package/src/vertexai/index.ts +53 -6
  79. package/src/vertexai/models/gemini.ts +18 -6
  80. package/src/vertexai/models/imagen.ts +3 -7
  81. package/lib/cjs/test-driver/TestErrorCompletionStream.js +0 -20
  82. package/lib/cjs/test-driver/TestErrorCompletionStream.js.map +0 -1
  83. package/lib/cjs/test-driver/TestValidationErrorCompletionStream.js +0 -24
  84. package/lib/cjs/test-driver/TestValidationErrorCompletionStream.js.map +0 -1
  85. package/lib/cjs/test-driver/index.js +0 -109
  86. package/lib/cjs/test-driver/index.js.map +0 -1
  87. package/lib/cjs/test-driver/utils.js +0 -30
  88. package/lib/cjs/test-driver/utils.js.map +0 -1
  89. package/lib/esm/src/adobe/firefly.js +0 -116
  90. package/lib/esm/src/adobe/firefly.js.map +0 -1
  91. package/lib/esm/src/azure/azure_foundry.js +0 -382
  92. package/lib/esm/src/azure/azure_foundry.js.map +0 -1
  93. package/lib/esm/src/bedrock/converse.js +0 -278
  94. package/lib/esm/src/bedrock/converse.js.map +0 -1
  95. package/lib/esm/src/bedrock/index.js +0 -962
  96. package/lib/esm/src/bedrock/index.js.map +0 -1
  97. package/lib/esm/src/bedrock/nova-image-payload.js +0 -203
  98. package/lib/esm/src/bedrock/nova-image-payload.js.map +0 -1
  99. package/lib/esm/src/bedrock/payloads.js +0 -2
  100. package/lib/esm/src/bedrock/payloads.js.map +0 -1
  101. package/lib/esm/src/bedrock/s3.js +0 -99
  102. package/lib/esm/src/bedrock/s3.js.map +0 -1
  103. package/lib/esm/src/bedrock/twelvelabs.js +0 -84
  104. package/lib/esm/src/bedrock/twelvelabs.js.map +0 -1
  105. package/lib/esm/src/groq/index.js +0 -286
  106. package/lib/esm/src/groq/index.js.map +0 -1
  107. package/lib/esm/src/huggingface_ie.js +0 -197
  108. package/lib/esm/src/huggingface_ie.js.map +0 -1
  109. package/lib/esm/src/index.js +0 -14
  110. package/lib/esm/src/index.js.map +0 -1
  111. package/lib/esm/src/mistral/index.js +0 -169
  112. package/lib/esm/src/mistral/index.js.map +0 -1
  113. package/lib/esm/src/mistral/types.js +0 -80
  114. package/lib/esm/src/mistral/types.js.map +0 -1
  115. package/lib/esm/src/openai/azure_openai.js +0 -68
  116. package/lib/esm/src/openai/azure_openai.js.map +0 -1
  117. package/lib/esm/src/openai/index.js +0 -464
  118. package/lib/esm/src/openai/index.js.map +0 -1
  119. package/lib/esm/src/openai/openai.js +0 -14
  120. package/lib/esm/src/openai/openai.js.map +0 -1
  121. package/lib/esm/src/openai/openai_format.js +0 -134
  122. package/lib/esm/src/openai/openai_format.js.map +0 -1
  123. package/lib/esm/src/replicate.js +0 -268
  124. package/lib/esm/src/replicate.js.map +0 -1
  125. package/lib/esm/src/test/TestErrorCompletionStream.js +0 -16
  126. package/lib/esm/src/test/TestErrorCompletionStream.js.map +0 -1
  127. package/lib/esm/src/test/TestValidationErrorCompletionStream.js +0 -20
  128. package/lib/esm/src/test/TestValidationErrorCompletionStream.js.map +0 -1
  129. package/lib/esm/src/test/index.js +0 -91
  130. package/lib/esm/src/test/index.js.map +0 -1
  131. package/lib/esm/src/test/utils.js +0 -25
  132. package/lib/esm/src/test/utils.js.map +0 -1
  133. package/lib/esm/src/test-driver/TestErrorCompletionStream.js +0 -16
  134. package/lib/esm/src/test-driver/TestErrorCompletionStream.js.map +0 -1
  135. package/lib/esm/src/test-driver/TestValidationErrorCompletionStream.js +0 -20
  136. package/lib/esm/src/test-driver/TestValidationErrorCompletionStream.js.map +0 -1
  137. package/lib/esm/src/test-driver/index.js +0 -91
  138. package/lib/esm/src/test-driver/index.js.map +0 -1
  139. package/lib/esm/src/test-driver/utils.js +0 -25
  140. package/lib/esm/src/test-driver/utils.js.map +0 -1
  141. package/lib/esm/src/togetherai/index.js +0 -122
  142. package/lib/esm/src/togetherai/index.js.map +0 -1
  143. package/lib/esm/src/togetherai/interfaces.js +0 -2
  144. package/lib/esm/src/togetherai/interfaces.js.map +0 -1
  145. package/lib/esm/src/vertexai/debug.js +0 -6
  146. package/lib/esm/src/vertexai/debug.js.map +0 -1
  147. package/lib/esm/src/vertexai/embeddings/embeddings-image.js +0 -24
  148. package/lib/esm/src/vertexai/embeddings/embeddings-image.js.map +0 -1
  149. package/lib/esm/src/vertexai/embeddings/embeddings-text.js +0 -20
  150. package/lib/esm/src/vertexai/embeddings/embeddings-text.js.map +0 -1
  151. package/lib/esm/src/vertexai/index.js +0 -383
  152. package/lib/esm/src/vertexai/index.js.map +0 -1
  153. package/lib/esm/src/vertexai/models/claude.js +0 -394
  154. package/lib/esm/src/vertexai/models/claude.js.map +0 -1
  155. package/lib/esm/src/vertexai/models/gemini.js +0 -817
  156. package/lib/esm/src/vertexai/models/gemini.js.map +0 -1
  157. package/lib/esm/src/vertexai/models/imagen.js +0 -302
  158. package/lib/esm/src/vertexai/models/imagen.js.map +0 -1
  159. package/lib/esm/src/vertexai/models/llama.js +0 -179
  160. package/lib/esm/src/vertexai/models/llama.js.map +0 -1
  161. package/lib/esm/src/vertexai/models.js +0 -32
  162. package/lib/esm/src/vertexai/models.js.map +0 -1
  163. package/lib/esm/src/watsonx/index.js +0 -157
  164. package/lib/esm/src/watsonx/index.js.map +0 -1
  165. package/lib/esm/src/watsonx/interfaces.js +0 -2
  166. package/lib/esm/src/watsonx/interfaces.js.map +0 -1
  167. package/lib/esm/src/xai/index.js +0 -64
  168. package/lib/esm/src/xai/index.js.map +0 -1
  169. package/lib/esm/test-driver/TestErrorCompletionStream.js +0 -16
  170. package/lib/esm/test-driver/TestErrorCompletionStream.js.map +0 -1
  171. package/lib/esm/test-driver/TestValidationErrorCompletionStream.js +0 -20
  172. package/lib/esm/test-driver/TestValidationErrorCompletionStream.js.map +0 -1
  173. package/lib/esm/test-driver/index.js +0 -91
  174. package/lib/esm/test-driver/index.js.map +0 -1
  175. package/lib/esm/test-driver/utils.js +0 -25
  176. package/lib/esm/test-driver/utils.js.map +0 -1
  177. package/lib/esm/tsconfig.tsbuildinfo +0 -1
  178. package/lib/types/src/adobe/firefly.d.ts +0 -29
  179. package/lib/types/src/azure/azure_foundry.d.ts +0 -49
  180. package/lib/types/src/bedrock/converse.d.ts +0 -8
  181. package/lib/types/src/bedrock/index.d.ts +0 -61
  182. package/lib/types/src/bedrock/nova-image-payload.d.ts +0 -73
  183. package/lib/types/src/bedrock/payloads.d.ts +0 -11
  184. package/lib/types/src/bedrock/s3.d.ts +0 -22
  185. package/lib/types/src/bedrock/twelvelabs.d.ts +0 -49
  186. package/lib/types/src/groq/index.d.ts +0 -26
  187. package/lib/types/src/huggingface_ie.d.ts +0 -34
  188. package/lib/types/src/index.d.ts +0 -13
  189. package/lib/types/src/mistral/index.d.ts +0 -24
  190. package/lib/types/src/mistral/types.d.ts +0 -131
  191. package/lib/types/src/openai/azure_openai.d.ts +0 -24
  192. package/lib/types/src/openai/index.d.ts +0 -24
  193. package/lib/types/src/openai/openai.d.ts +0 -14
  194. package/lib/types/src/openai/openai_format.d.ts +0 -18
  195. package/lib/types/src/replicate.d.ts +0 -47
  196. package/lib/types/src/test/TestErrorCompletionStream.d.ts +0 -8
  197. package/lib/types/src/test/TestValidationErrorCompletionStream.d.ts +0 -8
  198. package/lib/types/src/test/index.d.ts +0 -23
  199. package/lib/types/src/test/utils.d.ts +0 -4
  200. package/lib/types/src/test-driver/TestErrorCompletionStream.d.ts +0 -8
  201. package/lib/types/src/test-driver/TestValidationErrorCompletionStream.d.ts +0 -8
  202. package/lib/types/src/test-driver/index.d.ts +0 -23
  203. package/lib/types/src/test-driver/utils.d.ts +0 -4
  204. package/lib/types/src/togetherai/index.d.ts +0 -22
  205. package/lib/types/src/togetherai/interfaces.d.ts +0 -95
  206. package/lib/types/src/vertexai/debug.d.ts +0 -1
  207. package/lib/types/src/vertexai/embeddings/embeddings-image.d.ts +0 -10
  208. package/lib/types/src/vertexai/embeddings/embeddings-text.d.ts +0 -9
  209. package/lib/types/src/vertexai/index.d.ts +0 -52
  210. package/lib/types/src/vertexai/models/claude.d.ts +0 -19
  211. package/lib/types/src/vertexai/models/gemini.d.ts +0 -17
  212. package/lib/types/src/vertexai/models/imagen.d.ts +0 -74
  213. package/lib/types/src/vertexai/models/llama.d.ts +0 -19
  214. package/lib/types/src/vertexai/models.d.ts +0 -14
  215. package/lib/types/src/watsonx/index.d.ts +0 -26
  216. package/lib/types/src/watsonx/interfaces.d.ts +0 -64
  217. package/lib/types/src/xai/index.d.ts +0 -18
  218. package/lib/types/test-driver/TestErrorCompletionStream.d.ts +0 -9
  219. package/lib/types/test-driver/TestErrorCompletionStream.d.ts.map +0 -1
  220. package/lib/types/test-driver/TestValidationErrorCompletionStream.d.ts +0 -9
  221. package/lib/types/test-driver/TestValidationErrorCompletionStream.d.ts.map +0 -1
  222. package/lib/types/test-driver/index.d.ts +0 -24
  223. package/lib/types/test-driver/index.d.ts.map +0 -1
  224. package/lib/types/test-driver/utils.d.ts +0 -5
  225. package/lib/types/test-driver/utils.d.ts.map +0 -1
@@ -1,29 +0,0 @@
1
- import { AbstractDriver, AIModel, Completion, CompletionChunkObject, DriverOptions, EmbeddingsOptions, EmbeddingsResult, ExecutionOptions, ModelSearchPayload, PromptSegment } from "@llumiverse/core";
2
- interface FireflySize {
3
- width: number;
4
- height: number;
5
- }
6
- export interface FireflyDriverOptions extends DriverOptions {
7
- /**
8
- * Adobe Firefly API key
9
- */
10
- apiKey: string;
11
- /**
12
- * Optional API endpoint override
13
- */
14
- endpoint?: string;
15
- }
16
- export declare class FireflyDriver extends AbstractDriver<FireflyDriverOptions> {
17
- static PROVIDER: string;
18
- provider: string;
19
- private readonly endpoint;
20
- constructor(options: FireflyDriverOptions);
21
- requestTextCompletion(_prompt: string, _options: ExecutionOptions): Promise<Completion>;
22
- requestTextCompletionStream(_prompt: string, _options: ExecutionOptions): Promise<AsyncIterable<CompletionChunkObject>>;
23
- requestImageGeneration(segments: PromptSegment[], options: ExecutionOptions): Promise<Completion>;
24
- mapSize(size?: string): FireflySize;
25
- listModels(_params?: ModelSearchPayload): Promise<AIModel[]>;
26
- validateConnection(): Promise<boolean>;
27
- generateEmbeddings(_options: EmbeddingsOptions): Promise<EmbeddingsResult>;
28
- }
29
- export {};
@@ -1,49 +0,0 @@
1
- import { TokenCredential } from "@azure/identity";
2
- import { AbstractDriver, AIModel, Completion, CompletionChunkObject, DriverOptions, EmbeddingsOptions, EmbeddingsResult, ExecutionOptions, Providers } from "@llumiverse/core";
3
- import { AIProjectClient, ModelDeployment } from '@azure/ai-projects';
4
- import { ChatCompletionMessageParam } from "openai/resources";
5
- import type { ChatRequestMessage } from "@azure-rest/ai-inference";
6
- export interface AzureFoundryDriverOptions extends DriverOptions {
7
- /**
8
- * The credentials to use to access Azure AI Foundry
9
- */
10
- azureADTokenProvider?: TokenCredential;
11
- endpoint?: string;
12
- apiVersion?: string;
13
- }
14
- export interface AzureFoundryInferencePrompt {
15
- messages: ChatRequestMessage[];
16
- }
17
- export interface AzureFoundryOpenAIPrompt {
18
- messages: ChatCompletionMessageParam[];
19
- }
20
- export type AzureFoundryPrompt = AzureFoundryInferencePrompt | AzureFoundryOpenAIPrompt;
21
- export declare class AzureFoundryDriver extends AbstractDriver<AzureFoundryDriverOptions, ChatCompletionMessageParam[]> {
22
- service: AIProjectClient;
23
- readonly provider = Providers.azure_foundry;
24
- OPENAI_API_VERSION: string;
25
- INFERENCE_API_VERSION: string;
26
- constructor(opts: AzureFoundryDriverOptions);
27
- /**
28
- * Get default authentication for Azure AI Foundry API
29
- */
30
- getDefaultAIFoundryAuth(): () => Promise<string>;
31
- isOpenAIDeployment(model: string): Promise<boolean>;
32
- protected canStream(_options: ExecutionOptions): Promise<boolean>;
33
- requestTextCompletion(prompt: ChatCompletionMessageParam[], options: ExecutionOptions): Promise<Completion>;
34
- requestTextCompletionStream(prompt: ChatCompletionMessageParam[], options: ExecutionOptions): Promise<AsyncIterable<CompletionChunkObject>>;
35
- private processStreamResponse;
36
- private extractDataFromResponse;
37
- private convertFinishReason;
38
- validateConnection(): Promise<boolean>;
39
- generateEmbeddings(options: EmbeddingsOptions): Promise<EmbeddingsResult>;
40
- generateTextEmbeddings(options: EmbeddingsOptions): Promise<EmbeddingsResult>;
41
- generateImageEmbeddings(options: EmbeddingsOptions): Promise<EmbeddingsResult>;
42
- listModels(): Promise<AIModel[]>;
43
- _listModels(filter?: (m: ModelDeployment) => boolean): Promise<AIModel[]>;
44
- }
45
- export declare function parseAzureFoundryModelId(compositeId: string): {
46
- deploymentName: string;
47
- baseModel: string;
48
- };
49
- export declare function isCompositeModelId(modelId: string): boolean;
@@ -1,8 +0,0 @@
1
- import { ExecutionOptions } from "@llumiverse/core";
2
- import { PromptSegment } from "@llumiverse/core";
3
- import { ConverseRequest, Message, SystemContentBlock } from "@aws-sdk/client-bedrock-runtime";
4
- export declare function converseConcatMessages(messages: Message[] | undefined): Message[];
5
- export declare function converseSystemToMessages(system: SystemContentBlock[]): Message;
6
- export declare function converseRemoveJSONprefill(messages: Message[] | undefined): Message[];
7
- export declare function converseJSONprefill(messages: Message[] | undefined): Message[];
8
- export declare function formatConversePrompt(segments: PromptSegment[], options: ExecutionOptions): Promise<ConverseRequest>;
@@ -1,61 +0,0 @@
1
- import { Bedrock, FoundationModelSummary } from "@aws-sdk/client-bedrock";
2
- import { BedrockRuntime, ConverseRequest, ConverseResponse, ConverseStreamOutput } from "@aws-sdk/client-bedrock-runtime";
3
- import { AwsCredentialIdentity, Provider } from "@aws-sdk/types";
4
- import { AbstractDriver, AIModel, Completion, CompletionChunkObject, DataSource, DriverOptions, EmbeddingsOptions, EmbeddingsResult, ExecutionOptions, PromptSegment, TrainingJob, TrainingOptions } from "@llumiverse/core";
5
- import { NovaMessagesPrompt } from "@llumiverse/core/formatters";
6
- import { TwelvelabsPegasusRequest } from "./twelvelabs.js";
7
- export interface BedrockModelCapabilities {
8
- name: string;
9
- canStream: boolean;
10
- }
11
- export interface BedrockDriverOptions extends DriverOptions {
12
- /**
13
- * The AWS region
14
- */
15
- region: string;
16
- /**
17
- * The bucket name to be used for training.
18
- * It will be created if does not already exist.
19
- */
20
- training_bucket?: string;
21
- /**
22
- * The role ARN to be used for training
23
- */
24
- training_role_arn?: string;
25
- /**
26
- * The credentials to use to access AWS
27
- */
28
- credentials?: AwsCredentialIdentity | Provider<AwsCredentialIdentity>;
29
- }
30
- export type BedrockPrompt = NovaMessagesPrompt | ConverseRequest | TwelvelabsPegasusRequest;
31
- export declare class BedrockDriver extends AbstractDriver<BedrockDriverOptions, BedrockPrompt> {
32
- static PROVIDER: string;
33
- provider: string;
34
- private _executor?;
35
- private _service?;
36
- private _service_region?;
37
- constructor(options: BedrockDriverOptions);
38
- getExecutor(): BedrockRuntime;
39
- getService(region?: string): Bedrock;
40
- protected formatPrompt(segments: PromptSegment[], opts: ExecutionOptions): Promise<BedrockPrompt>;
41
- getExtractedExecution(result: ConverseResponse, _prompt?: BedrockPrompt, options?: ExecutionOptions): CompletionChunkObject;
42
- getExtractedStream(result: ConverseStreamOutput, _prompt?: BedrockPrompt, options?: ExecutionOptions): CompletionChunkObject;
43
- extractRegion(modelString: string, defaultRegion: string): string;
44
- private getCanStream;
45
- protected canStream(options: ExecutionOptions): Promise<boolean>;
46
- requestTextCompletion(prompt: BedrockPrompt, options: ExecutionOptions): Promise<Completion>;
47
- private requestTwelvelabsPegasusCompletion;
48
- private requestTwelvelabsPegasusCompletionStream;
49
- requestTextCompletionStream(prompt: BedrockPrompt, options: ExecutionOptions): Promise<AsyncIterable<CompletionChunkObject>>;
50
- preparePayload(prompt: ConverseRequest, options: ExecutionOptions): ConverseRequest;
51
- requestImageGeneration(prompt: NovaMessagesPrompt, options: ExecutionOptions): Promise<Completion>;
52
- startTraining(dataset: DataSource, options: TrainingOptions): Promise<TrainingJob>;
53
- cancelTraining(jobId: string): Promise<TrainingJob>;
54
- getTrainingJob(jobId: string): Promise<TrainingJob>;
55
- validateConnection(): Promise<boolean>;
56
- listTrainableModels(): Promise<AIModel<string>[]>;
57
- listModels(): Promise<AIModel[]>;
58
- _listModels(foundationFilter?: (m: FoundationModelSummary) => boolean): Promise<AIModel[]>;
59
- generateEmbeddings({ text, image, model }: EmbeddingsOptions): Promise<EmbeddingsResult>;
60
- private generateTwelvelabsMarengoEmbeddings;
61
- }
@@ -1,73 +0,0 @@
1
- import { ExecutionOptions } from "@llumiverse/core";
2
- import { NovaMessagesPrompt } from "@llumiverse/core/formatters";
3
- export declare function formatNovaImageGenerationPayload(taskType: string, prompt: NovaMessagesPrompt, options: ExecutionOptions): Promise<NovaTextToImagePayload> | Promise<NovaColorGuidedGenerationPayload> | Promise<NovaImageVariationPayload> | Promise<NovaInpaintingPayload> | Promise<NovaOutpaintingPayload> | Promise<NovaBackgroundRemovalPayload>;
4
- export interface InvokeModelPayloadBase {
5
- taskType: NovaImageGenerationTaskType;
6
- imageGenerationConfig: {
7
- width?: number;
8
- height?: number;
9
- quality?: "standard" | "premium";
10
- cfgScale?: number;
11
- seed?: number;
12
- numberOfImages?: number;
13
- };
14
- }
15
- export interface NovaTextToImagePayload extends InvokeModelPayloadBase {
16
- textToImageParams: {
17
- conditionImage?: string;
18
- controlMode?: "CANNY_EDGE" | "SEGMENTATION";
19
- controlStrength?: number;
20
- text: string;
21
- negativeText?: string;
22
- };
23
- }
24
- export interface NovaImageVariationPayload extends InvokeModelPayloadBase {
25
- imageVariationParams: {
26
- images: string[];
27
- similarityStrength?: number;
28
- text?: string;
29
- negativeText?: string;
30
- };
31
- }
32
- export interface NovaColorGuidedGenerationPayload extends InvokeModelPayloadBase {
33
- colorGuidedGenerationParams: {
34
- colors: string[];
35
- text: string;
36
- referenceImage?: string;
37
- negativeText?: string;
38
- };
39
- }
40
- export interface NovaInpaintingPayload extends InvokeModelPayloadBase {
41
- inPaintingParams: {
42
- image: string;
43
- maskImage?: string;
44
- maskPrompt?: string;
45
- negativeText?: string;
46
- text?: string;
47
- };
48
- }
49
- export interface NovaOutpaintingPayload extends InvokeModelPayloadBase {
50
- outPaintingParams: {
51
- image: string;
52
- maskImage?: string;
53
- maskPrompt?: string;
54
- negativeText?: string;
55
- text?: string;
56
- outPaintingMode: "DEFAULT" | "PRECISE";
57
- };
58
- }
59
- export interface NovaBackgroundRemovalPayload {
60
- taskType: NovaImageGenerationTaskType.BACKGROUND_REMOVAL;
61
- backgroundRemovalParams: {
62
- image: string;
63
- };
64
- }
65
- export declare enum NovaImageGenerationTaskType {
66
- TEXT_IMAGE = "TEXT_IMAGE",
67
- TEXT_IMAGE_WITH_IMAGE_CONDITIONING = "TEXT_IMAGE_WITH_IMAGE_CONDITIONING",
68
- COLOR_GUIDED_GENERATION = "COLOR_GUIDED_GENERATION",
69
- IMAGE_VARIATION = "IMAGE_VARIATION",
70
- INPAINTING = "INPAINTING",
71
- OUTPAINTING = "OUTPAINTING",
72
- BACKGROUND_REMOVAL = "BACKGROUND_REMOVAL"
73
- }
@@ -1,11 +0,0 @@
1
- import { NovaMessagesPrompt } from "@llumiverse/core/formatters";
2
- export interface NovaPayload extends NovaMessagesPrompt {
3
- schemaVersion: string;
4
- inferenceConfig?: {
5
- max_new_tokens?: number;
6
- temperature?: number;
7
- top_p?: number;
8
- top_k?: number;
9
- stopSequences?: [string];
10
- };
11
- }
@@ -1,22 +0,0 @@
1
- import { S3Client } from "@aws-sdk/client-s3";
2
- import { Progress } from "@aws-sdk/lib-storage";
3
- export declare function doesBucketExist(s3: S3Client, bucketName: string): Promise<boolean>;
4
- export declare function createBucket(s3: S3Client, bucketName: string): Promise<import("@aws-sdk/client-s3").CreateBucketCommandOutput>;
5
- export declare function tryCreateBucket(s3: S3Client, bucketName: string): Promise<import("@aws-sdk/client-s3").CreateBucketCommandOutput | undefined>;
6
- export declare function uploadFile(s3: S3Client, source: ReadableStream, bucketName: string, file: string, onProgress?: (progress: Progress) => void): Promise<import("@aws-sdk/client-s3").CompleteMultipartUploadCommandOutput>;
7
- /**
8
- * Create the bucket if not already exists and then upload the file.
9
- * @param s3
10
- * @param source
11
- * @param bucketName
12
- * @param file
13
- * @param onProgress
14
- * @returns
15
- */
16
- export declare function forceUploadFile(s3: S3Client, source: ReadableStream, bucketName: string, file: string, onProgress?: (progress: Progress) => void): Promise<import("@aws-sdk/client-s3").CompleteMultipartUploadCommandOutput>;
17
- /**
18
- * Parse an S3 HTTPS URL into an S3 URI format
19
- * s3Url - The S3 HTTPS URL (e.g., https://bucket.s3.region.amazonaws.com/key)
20
- * returns The S3 URI (e.g., s3://bucket/key)
21
- */
22
- export declare function parseS3UrlToUri(s3Url: URL): string;
@@ -1,49 +0,0 @@
1
- import { ExecutionOptions } from "@llumiverse/core";
2
- import { PromptSegment } from "@llumiverse/core";
3
- export interface TwelvelabsPegasusRequest {
4
- inputPrompt: string;
5
- temperature?: number;
6
- responseFormat?: {
7
- type: "json_schema";
8
- json_schema: {
9
- name: string;
10
- schema: any;
11
- };
12
- };
13
- mediaSource: {
14
- base64String?: string;
15
- s3Location?: {
16
- uri: string;
17
- bucketOwner?: string;
18
- };
19
- };
20
- maxOutputTokens?: number;
21
- }
22
- export interface TwelvelabsPegasusResponse {
23
- message: string;
24
- finishReason: "stop" | "length";
25
- }
26
- export interface TwelvelabsMarengoRequest {
27
- inputType: "text" | "image" | "video" | "audio";
28
- inputText?: string;
29
- textTruncate?: "start" | "end";
30
- mediaSource?: {
31
- base64String?: string;
32
- s3Location?: {
33
- uri: string;
34
- bucketOwner?: string;
35
- };
36
- };
37
- embeddingOption?: "visual-text" | "visual-image" | "audio";
38
- startSec?: number;
39
- lengthSec?: number;
40
- useFixedLengthSec?: boolean;
41
- minClipSec?: number;
42
- }
43
- export interface TwelvelabsMarengoResponse {
44
- embedding: number[];
45
- embeddingOption: "visual-text" | "visual-image" | "audio";
46
- startSec: number;
47
- endSec: number;
48
- }
49
- export declare function formatTwelvelabsPegasusPrompt(segments: PromptSegment[], options: ExecutionOptions): Promise<TwelvelabsPegasusRequest>;
@@ -1,26 +0,0 @@
1
- import { AIModel, AbstractDriver, Completion, CompletionChunkObject, DriverOptions, EmbeddingsOptions, EmbeddingsResult, ExecutionOptions, PromptSegment } from "@llumiverse/core";
2
- import Groq from "groq-sdk";
3
- import type { ChatCompletionMessageParam } from "groq-sdk/resources/chat/completions";
4
- interface GroqDriverOptions extends DriverOptions {
5
- apiKey: string;
6
- endpoint_url?: string;
7
- }
8
- export declare class GroqDriver extends AbstractDriver<GroqDriverOptions, ChatCompletionMessageParam[]> {
9
- static PROVIDER: string;
10
- provider: string;
11
- apiKey: string;
12
- client: Groq;
13
- endpointUrl?: string;
14
- constructor(options: GroqDriverOptions);
15
- getResponseFormat(_options: ExecutionOptions): undefined;
16
- protected formatPrompt(segments: PromptSegment[], opts: ExecutionOptions): Promise<ChatCompletionMessageParam[]>;
17
- private getToolDefinitions;
18
- private extractToolUse;
19
- private sanitizeMessagesForGroq;
20
- requestTextCompletion(messages: ChatCompletionMessageParam[], options: ExecutionOptions): Promise<Completion>;
21
- requestTextCompletionStream(messages: ChatCompletionMessageParam[], options: ExecutionOptions): Promise<AsyncIterable<CompletionChunkObject>>;
22
- listModels(): Promise<AIModel<string>[]>;
23
- validateConnection(): Promise<boolean>;
24
- generateEmbeddings({}: EmbeddingsOptions): Promise<EmbeddingsResult>;
25
- }
26
- export {};
@@ -1,34 +0,0 @@
1
- import { InferenceClient } from "@huggingface/inference";
2
- import { AIModel, AbstractDriver, CompletionChunkObject, DriverOptions, EmbeddingsResult, ExecutionOptions } from "@llumiverse/core";
3
- import { FetchClient } from "@vertesia/api-fetch-client";
4
- export interface HuggingFaceIEDriverOptions extends DriverOptions {
5
- apiKey: string;
6
- endpoint_url: string;
7
- }
8
- export declare class HuggingFaceIEDriver extends AbstractDriver<HuggingFaceIEDriverOptions, string> {
9
- static PROVIDER: string;
10
- provider: string;
11
- service: FetchClient;
12
- _executor?: InferenceClient;
13
- constructor(options: HuggingFaceIEDriverOptions);
14
- getModelURLEndpoint(modelId: string): Promise<{
15
- url: string;
16
- status: string;
17
- }>;
18
- getExecutor(model: string): Promise<InferenceClient>;
19
- requestTextCompletionStream(prompt: string, options: ExecutionOptions): Promise<AsyncIterable<CompletionChunkObject>>;
20
- requestTextCompletion(prompt: string, options: ExecutionOptions): Promise<{
21
- result: {
22
- type: "text";
23
- value: string;
24
- }[];
25
- finish_reason: string;
26
- token_usage: {
27
- result: number | undefined;
28
- };
29
- original_response: import("@huggingface/inference").TextGenerationOutput | undefined;
30
- }>;
31
- listModels(): Promise<AIModel[]>;
32
- validateConnection(): Promise<boolean>;
33
- generateEmbeddings(): Promise<EmbeddingsResult>;
34
- }
@@ -1,13 +0,0 @@
1
- export * from "./azure/azure_foundry.js";
2
- export * from "./bedrock/index.js";
3
- export * from "./groq/index.js";
4
- export * from "./huggingface_ie.js";
5
- export * from "./mistral/index.js";
6
- export * from "./openai/azure_openai.js";
7
- export * from "./openai/openai.js";
8
- export * from "./replicate.js";
9
- export * from "./test-driver/index.js";
10
- export * from "./togetherai/index.js";
11
- export * from "./vertexai/index.js";
12
- export * from "./watsonx/index.js";
13
- export * from "./xai/index.js";
@@ -1,24 +0,0 @@
1
- import { AIModel, AbstractDriver, Completion, CompletionChunkObject, DriverOptions, EmbeddingsOptions, EmbeddingsResult, ExecutionOptions, PromptSegment } from "@llumiverse/core";
2
- import { OpenAITextMessage } from "../openai/openai_format.js";
3
- import { FetchClient } from "@vertesia/api-fetch-client";
4
- import { ResponseFormat } from "./types.js";
5
- interface MistralAIDriverOptions extends DriverOptions {
6
- apiKey: string;
7
- endpoint_url?: string;
8
- }
9
- export declare class MistralAIDriver extends AbstractDriver<MistralAIDriverOptions, OpenAITextMessage[]> {
10
- static PROVIDER: string;
11
- provider: string;
12
- apiKey: string;
13
- client: FetchClient;
14
- endpointUrl?: string;
15
- constructor(options: MistralAIDriverOptions);
16
- getResponseFormat: (_options: ExecutionOptions) => ResponseFormat | undefined;
17
- protected formatPrompt(segments: PromptSegment[], opts: ExecutionOptions): Promise<OpenAITextMessage[]>;
18
- requestTextCompletion(messages: OpenAITextMessage[], options: ExecutionOptions): Promise<Completion>;
19
- requestTextCompletionStream(messages: OpenAITextMessage[], options: ExecutionOptions): Promise<AsyncIterable<CompletionChunkObject>>;
20
- listModels(): Promise<AIModel<string>[]>;
21
- validateConnection(): Promise<boolean>;
22
- generateEmbeddings({ text, model }: EmbeddingsOptions): Promise<EmbeddingsResult>;
23
- }
24
- export {};
@@ -1,131 +0,0 @@
1
- export interface ModelPermission {
2
- id: string;
3
- object: 'model_permission';
4
- created: number;
5
- allow_create_engine: boolean;
6
- allow_sampling: boolean;
7
- allow_logprobs: boolean;
8
- allow_search_indices: boolean;
9
- allow_view: boolean;
10
- allow_fine_tuning: boolean;
11
- organization: string;
12
- group: string | null;
13
- is_blocking: boolean;
14
- }
15
- export interface Model {
16
- id: string;
17
- object: 'model';
18
- created: number;
19
- owned_by: string;
20
- root: string | null;
21
- parent: string | null;
22
- permission: ModelPermission[];
23
- }
24
- export interface ListModelsResponse {
25
- object: 'list';
26
- data: Model[];
27
- }
28
- export interface Function {
29
- name: string;
30
- description: string;
31
- parameters: object;
32
- }
33
- export declare enum ToolType {
34
- function = "function"
35
- }
36
- export interface FunctionCall {
37
- name: string;
38
- arguments: string;
39
- }
40
- export interface ToolCalls {
41
- id: 'null';
42
- type: ToolType;
43
- function: FunctionCall;
44
- }
45
- export declare enum ResponseFormats {
46
- text = "text",
47
- json_object = "json_object"
48
- }
49
- export declare enum ToolChoice {
50
- auto = "auto",
51
- any = "any",
52
- none = "none"
53
- }
54
- export interface ResponseFormat {
55
- type: ResponseFormats;
56
- }
57
- export interface TokenUsage {
58
- prompt_tokens: number;
59
- completion_tokens: number;
60
- total_tokens: number;
61
- }
62
- export interface ChatCompletionResponseChoice {
63
- index: number;
64
- message: {
65
- role: string;
66
- content: string;
67
- };
68
- finish_reason: string;
69
- }
70
- export interface ChatCompletionResponseChunkChoice {
71
- index: number;
72
- delta: {
73
- role?: string;
74
- content?: string;
75
- tool_calls?: ToolCalls[];
76
- };
77
- finish_reason: string;
78
- }
79
- export interface ChatCompletionResponse {
80
- id: string;
81
- object: 'chat.completion';
82
- created: number;
83
- model: string;
84
- choices: ChatCompletionResponseChoice[];
85
- usage: TokenUsage;
86
- }
87
- export interface ChatCompletionResponseChunk {
88
- id: string;
89
- object: 'chat.completion.chunk';
90
- created: number;
91
- model: string;
92
- choices: ChatCompletionResponseChunkChoice[];
93
- }
94
- export interface Embedding {
95
- id: string;
96
- object: 'embedding';
97
- embedding: number[];
98
- }
99
- export interface EmbeddingResponse {
100
- id: string;
101
- object: 'list';
102
- data: Embedding[];
103
- model: string;
104
- usage: TokenUsage;
105
- }
106
- export interface CompletionRequestParams {
107
- model: string;
108
- messages: Array<{
109
- role: string;
110
- name?: string;
111
- content: string | string[];
112
- tool_calls?: ToolCalls[];
113
- }>;
114
- tools?: Array<{
115
- type: string;
116
- function: Function;
117
- }>;
118
- temperature?: number;
119
- maxTokens?: number;
120
- topP?: number;
121
- randomSeed?: number;
122
- stream?: boolean;
123
- /**
124
- * @deprecated use safePrompt instead
125
- */
126
- safeMode?: boolean;
127
- safePrompt?: boolean;
128
- toolChoice?: ToolChoice;
129
- responseFormat?: ResponseFormat;
130
- stopSequences?: string[];
131
- }
@@ -1,24 +0,0 @@
1
- import { AIModel, DriverOptions, Providers } from "@llumiverse/core";
2
- import OpenAI, { AzureOpenAI } from "openai";
3
- import { BaseOpenAIDriver } from "./index.js";
4
- export interface AzureOpenAIDriverOptions extends DriverOptions {
5
- /**
6
- * The credentials to use to access Azure OpenAI
7
- */
8
- azureADTokenProvider?: any;
9
- apiKey?: string;
10
- endpoint?: string;
11
- apiVersion?: string;
12
- deployment?: string;
13
- }
14
- export declare class AzureOpenAIDriver extends BaseOpenAIDriver {
15
- service: AzureOpenAI;
16
- readonly provider = Providers.azure_openai;
17
- constructor(serviceOrOpts: AzureOpenAI | AzureOpenAIDriverOptions);
18
- /**
19
- * Get default authentication for Azure Cognitive Services API
20
- */
21
- getDefaultCognitiveServicesAuth(): () => Promise<string>;
22
- listModels(): Promise<AIModel[]>;
23
- _listModels(_filter?: (m: OpenAI.Models.Model) => boolean): Promise<AIModel[]>;
24
- }
@@ -1,24 +0,0 @@
1
- import { AIModel, AbstractDriver, Completion, CompletionChunkObject, DataSource, DriverOptions, EmbeddingsOptions, EmbeddingsResult, ExecutionOptions, Providers, ToolUse, TrainingJob, TrainingOptions, TrainingPromptOptions } from "@llumiverse/core";
2
- import OpenAI, { AzureOpenAI } from "openai";
3
- import { ChatCompletionMessageParam } from "openai/resources/chat/completions";
4
- export interface BaseOpenAIDriverOptions extends DriverOptions {
5
- }
6
- export declare abstract class BaseOpenAIDriver extends AbstractDriver<BaseOpenAIDriverOptions, ChatCompletionMessageParam[]> {
7
- abstract provider: Providers.openai | Providers.azure_openai | "xai" | Providers.azure_foundry;
8
- abstract service: OpenAI | AzureOpenAI;
9
- constructor(opts: BaseOpenAIDriverOptions);
10
- extractDataFromResponse(_options: ExecutionOptions, result: OpenAI.Chat.Completions.ChatCompletion): Completion;
11
- requestTextCompletionStream(prompt: ChatCompletionMessageParam[], options: ExecutionOptions): Promise<AsyncIterable<CompletionChunkObject>>;
12
- requestTextCompletion(prompt: ChatCompletionMessageParam[], options: ExecutionOptions): Promise<Completion>;
13
- protected canStream(_options: ExecutionOptions): Promise<boolean>;
14
- createTrainingPrompt(options: TrainingPromptOptions): Promise<string>;
15
- startTraining(dataset: DataSource, options: TrainingOptions): Promise<TrainingJob>;
16
- cancelTraining(jobId: string): Promise<TrainingJob>;
17
- getTrainingJob(jobId: string): Promise<TrainingJob>;
18
- validateConnection(): Promise<boolean>;
19
- listTrainableModels(): Promise<AIModel<string>[]>;
20
- listModels(): Promise<AIModel[]>;
21
- _listModels(filter?: (m: OpenAI.Models.Model) => boolean): Promise<AIModel[]>;
22
- generateEmbeddings({ text, image, model }: EmbeddingsOptions): Promise<EmbeddingsResult>;
23
- }
24
- export declare function collectTools(toolCalls?: OpenAI.Chat.Completions.ChatCompletionMessageToolCall[]): ToolUse[] | undefined;
@@ -1,14 +0,0 @@
1
- import { DriverOptions, Providers } from "@llumiverse/core";
2
- import OpenAI from "openai";
3
- import { BaseOpenAIDriver } from "./index.js";
4
- export interface OpenAIDriverOptions extends DriverOptions {
5
- /**
6
- * The OpenAI api key
7
- */
8
- apiKey?: string;
9
- }
10
- export declare class OpenAIDriver extends BaseOpenAIDriver {
11
- service: OpenAI;
12
- readonly provider = Providers.openai;
13
- constructor(opts: OpenAIDriverOptions);
14
- }
@@ -1,18 +0,0 @@
1
- import { PromptOptions, PromptSegment } from "@llumiverse/common";
2
- import type { ChatCompletionMessageParam } from 'openai/resources/chat/completions';
3
- export interface OpenAITextMessage {
4
- content: string;
5
- role: 'system' | 'user' | 'assistant' | 'developer';
6
- }
7
- /**
8
- * OpenAI text only prompts
9
- * @param segments
10
- * @returns
11
- */
12
- export declare function formatOpenAILikeTextPrompt(segments: PromptSegment[]): OpenAITextMessage[];
13
- export declare function formatOpenAILikeMultimodalPrompt(segments: PromptSegment[], opts: PromptOptions & OpenAIPromptFormatterOptions): Promise<ChatCompletionMessageParam[]>;
14
- export interface OpenAIPromptFormatterOptions {
15
- multimodal?: boolean;
16
- useToolForFormatting?: boolean;
17
- schema?: Object;
18
- }