@llumiverse/drivers 0.22.2 → 0.22.3
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/lib/cjs/adobe/firefly.js +2 -2
- package/lib/cjs/adobe/firefly.js.map +1 -1
- package/lib/cjs/azure/azure_foundry.js +11 -11
- package/lib/cjs/azure/azure_foundry.js.map +1 -1
- package/lib/cjs/bedrock/index.js +7 -7
- package/lib/cjs/bedrock/index.js.map +1 -1
- package/lib/cjs/groq/index.js +2 -2
- package/lib/cjs/groq/index.js.map +1 -1
- package/lib/cjs/huggingface_ie.js +3 -4
- package/lib/cjs/huggingface_ie.js.map +1 -1
- package/lib/cjs/index.js +2 -2
- package/lib/cjs/index.js.map +1 -1
- package/lib/cjs/mistral/index.js +2 -2
- package/lib/cjs/mistral/index.js.map +1 -1
- package/lib/cjs/openai/azure_openai.js +1 -1
- package/lib/cjs/openai/azure_openai.js.map +1 -1
- package/lib/cjs/openai/index.js +3 -3
- package/lib/cjs/openai/index.js.map +1 -1
- package/lib/cjs/replicate.js +4 -4
- package/lib/cjs/replicate.js.map +1 -1
- package/lib/cjs/togetherai/index.js +2 -2
- package/lib/cjs/togetherai/index.js.map +1 -1
- package/lib/cjs/vertexai/index.js +82 -40
- package/lib/cjs/vertexai/index.js.map +1 -1
- package/lib/cjs/vertexai/models/claude.js +4 -4
- package/lib/cjs/vertexai/models/claude.js.map +1 -1
- package/lib/cjs/vertexai/models/gemini.js +15 -5
- package/lib/cjs/vertexai/models/gemini.js.map +1 -1
- package/lib/cjs/vertexai/models/imagen.js +2 -5
- package/lib/cjs/vertexai/models/imagen.js.map +1 -1
- package/lib/cjs/watsonx/index.js +3 -3
- package/lib/cjs/watsonx/index.js.map +1 -1
- package/lib/esm/adobe/firefly.js +2 -2
- package/lib/esm/adobe/firefly.js.map +1 -1
- package/lib/esm/azure/azure_foundry.js +11 -11
- package/lib/esm/azure/azure_foundry.js.map +1 -1
- package/lib/esm/bedrock/index.js +8 -8
- package/lib/esm/bedrock/index.js.map +1 -1
- package/lib/esm/groq/index.js +2 -2
- package/lib/esm/groq/index.js.map +1 -1
- package/lib/esm/huggingface_ie.js +4 -5
- package/lib/esm/huggingface_ie.js.map +1 -1
- package/lib/esm/index.js +2 -2
- package/lib/esm/index.js.map +1 -1
- package/lib/esm/mistral/index.js +2 -2
- package/lib/esm/mistral/index.js.map +1 -1
- package/lib/esm/openai/azure_openai.js +1 -1
- package/lib/esm/openai/azure_openai.js.map +1 -1
- package/lib/esm/openai/index.js +3 -3
- package/lib/esm/openai/index.js.map +1 -1
- package/lib/esm/replicate.js +4 -4
- package/lib/esm/replicate.js.map +1 -1
- package/lib/esm/togetherai/index.js +2 -2
- package/lib/esm/togetherai/index.js.map +1 -1
- package/lib/esm/vertexai/index.js +83 -41
- package/lib/esm/vertexai/index.js.map +1 -1
- package/lib/esm/vertexai/models/claude.js +4 -4
- package/lib/esm/vertexai/models/claude.js.map +1 -1
- package/lib/esm/vertexai/models/gemini.js +15 -5
- package/lib/esm/vertexai/models/gemini.js.map +1 -1
- package/lib/esm/vertexai/models/imagen.js +3 -6
- package/lib/esm/vertexai/models/imagen.js.map +1 -1
- package/lib/esm/watsonx/index.js +3 -3
- package/lib/esm/watsonx/index.js.map +1 -1
- package/lib/types/bedrock/index.d.ts +1 -0
- package/lib/types/bedrock/index.d.ts.map +1 -1
- package/lib/types/huggingface_ie.d.ts +3 -3
- package/lib/types/huggingface_ie.d.ts.map +1 -1
- package/lib/types/index.d.ts +2 -2
- package/lib/types/index.d.ts.map +1 -1
- package/lib/types/replicate.d.ts.map +1 -1
- package/lib/types/vertexai/index.d.ts +12 -11
- package/lib/types/vertexai/index.d.ts.map +1 -1
- package/lib/types/vertexai/models/gemini.d.ts.map +1 -1
- package/lib/types/vertexai/models/imagen.d.ts.map +1 -1
- package/package.json +3 -3
- package/src/bedrock/index.ts +5 -4
- package/src/vertexai/index.ts +53 -6
- package/src/vertexai/models/gemini.ts +18 -6
- package/src/vertexai/models/imagen.ts +3 -7
- package/lib/cjs/test-driver/TestErrorCompletionStream.js +0 -20
- package/lib/cjs/test-driver/TestErrorCompletionStream.js.map +0 -1
- package/lib/cjs/test-driver/TestValidationErrorCompletionStream.js +0 -24
- package/lib/cjs/test-driver/TestValidationErrorCompletionStream.js.map +0 -1
- package/lib/cjs/test-driver/index.js +0 -109
- package/lib/cjs/test-driver/index.js.map +0 -1
- package/lib/cjs/test-driver/utils.js +0 -30
- package/lib/cjs/test-driver/utils.js.map +0 -1
- package/lib/esm/src/adobe/firefly.js +0 -116
- package/lib/esm/src/adobe/firefly.js.map +0 -1
- package/lib/esm/src/azure/azure_foundry.js +0 -382
- package/lib/esm/src/azure/azure_foundry.js.map +0 -1
- package/lib/esm/src/bedrock/converse.js +0 -278
- package/lib/esm/src/bedrock/converse.js.map +0 -1
- package/lib/esm/src/bedrock/index.js +0 -962
- package/lib/esm/src/bedrock/index.js.map +0 -1
- package/lib/esm/src/bedrock/nova-image-payload.js +0 -203
- package/lib/esm/src/bedrock/nova-image-payload.js.map +0 -1
- package/lib/esm/src/bedrock/payloads.js +0 -2
- package/lib/esm/src/bedrock/payloads.js.map +0 -1
- package/lib/esm/src/bedrock/s3.js +0 -99
- package/lib/esm/src/bedrock/s3.js.map +0 -1
- package/lib/esm/src/bedrock/twelvelabs.js +0 -84
- package/lib/esm/src/bedrock/twelvelabs.js.map +0 -1
- package/lib/esm/src/groq/index.js +0 -286
- package/lib/esm/src/groq/index.js.map +0 -1
- package/lib/esm/src/huggingface_ie.js +0 -197
- package/lib/esm/src/huggingface_ie.js.map +0 -1
- package/lib/esm/src/index.js +0 -14
- package/lib/esm/src/index.js.map +0 -1
- package/lib/esm/src/mistral/index.js +0 -169
- package/lib/esm/src/mistral/index.js.map +0 -1
- package/lib/esm/src/mistral/types.js +0 -80
- package/lib/esm/src/mistral/types.js.map +0 -1
- package/lib/esm/src/openai/azure_openai.js +0 -68
- package/lib/esm/src/openai/azure_openai.js.map +0 -1
- package/lib/esm/src/openai/index.js +0 -464
- package/lib/esm/src/openai/index.js.map +0 -1
- package/lib/esm/src/openai/openai.js +0 -14
- package/lib/esm/src/openai/openai.js.map +0 -1
- package/lib/esm/src/openai/openai_format.js +0 -134
- package/lib/esm/src/openai/openai_format.js.map +0 -1
- package/lib/esm/src/replicate.js +0 -268
- package/lib/esm/src/replicate.js.map +0 -1
- package/lib/esm/src/test/TestErrorCompletionStream.js +0 -16
- package/lib/esm/src/test/TestErrorCompletionStream.js.map +0 -1
- package/lib/esm/src/test/TestValidationErrorCompletionStream.js +0 -20
- package/lib/esm/src/test/TestValidationErrorCompletionStream.js.map +0 -1
- package/lib/esm/src/test/index.js +0 -91
- package/lib/esm/src/test/index.js.map +0 -1
- package/lib/esm/src/test/utils.js +0 -25
- package/lib/esm/src/test/utils.js.map +0 -1
- package/lib/esm/src/test-driver/TestErrorCompletionStream.js +0 -16
- package/lib/esm/src/test-driver/TestErrorCompletionStream.js.map +0 -1
- package/lib/esm/src/test-driver/TestValidationErrorCompletionStream.js +0 -20
- package/lib/esm/src/test-driver/TestValidationErrorCompletionStream.js.map +0 -1
- package/lib/esm/src/test-driver/index.js +0 -91
- package/lib/esm/src/test-driver/index.js.map +0 -1
- package/lib/esm/src/test-driver/utils.js +0 -25
- package/lib/esm/src/test-driver/utils.js.map +0 -1
- package/lib/esm/src/togetherai/index.js +0 -122
- package/lib/esm/src/togetherai/index.js.map +0 -1
- package/lib/esm/src/togetherai/interfaces.js +0 -2
- package/lib/esm/src/togetherai/interfaces.js.map +0 -1
- package/lib/esm/src/vertexai/debug.js +0 -6
- package/lib/esm/src/vertexai/debug.js.map +0 -1
- package/lib/esm/src/vertexai/embeddings/embeddings-image.js +0 -24
- package/lib/esm/src/vertexai/embeddings/embeddings-image.js.map +0 -1
- package/lib/esm/src/vertexai/embeddings/embeddings-text.js +0 -20
- package/lib/esm/src/vertexai/embeddings/embeddings-text.js.map +0 -1
- package/lib/esm/src/vertexai/index.js +0 -383
- package/lib/esm/src/vertexai/index.js.map +0 -1
- package/lib/esm/src/vertexai/models/claude.js +0 -394
- package/lib/esm/src/vertexai/models/claude.js.map +0 -1
- package/lib/esm/src/vertexai/models/gemini.js +0 -817
- package/lib/esm/src/vertexai/models/gemini.js.map +0 -1
- package/lib/esm/src/vertexai/models/imagen.js +0 -302
- package/lib/esm/src/vertexai/models/imagen.js.map +0 -1
- package/lib/esm/src/vertexai/models/llama.js +0 -179
- package/lib/esm/src/vertexai/models/llama.js.map +0 -1
- package/lib/esm/src/vertexai/models.js +0 -32
- package/lib/esm/src/vertexai/models.js.map +0 -1
- package/lib/esm/src/watsonx/index.js +0 -157
- package/lib/esm/src/watsonx/index.js.map +0 -1
- package/lib/esm/src/watsonx/interfaces.js +0 -2
- package/lib/esm/src/watsonx/interfaces.js.map +0 -1
- package/lib/esm/src/xai/index.js +0 -64
- package/lib/esm/src/xai/index.js.map +0 -1
- package/lib/esm/test-driver/TestErrorCompletionStream.js +0 -16
- package/lib/esm/test-driver/TestErrorCompletionStream.js.map +0 -1
- package/lib/esm/test-driver/TestValidationErrorCompletionStream.js +0 -20
- package/lib/esm/test-driver/TestValidationErrorCompletionStream.js.map +0 -1
- package/lib/esm/test-driver/index.js +0 -91
- package/lib/esm/test-driver/index.js.map +0 -1
- package/lib/esm/test-driver/utils.js +0 -25
- package/lib/esm/test-driver/utils.js.map +0 -1
- package/lib/esm/tsconfig.tsbuildinfo +0 -1
- package/lib/types/src/adobe/firefly.d.ts +0 -29
- package/lib/types/src/azure/azure_foundry.d.ts +0 -49
- package/lib/types/src/bedrock/converse.d.ts +0 -8
- package/lib/types/src/bedrock/index.d.ts +0 -61
- package/lib/types/src/bedrock/nova-image-payload.d.ts +0 -73
- package/lib/types/src/bedrock/payloads.d.ts +0 -11
- package/lib/types/src/bedrock/s3.d.ts +0 -22
- package/lib/types/src/bedrock/twelvelabs.d.ts +0 -49
- package/lib/types/src/groq/index.d.ts +0 -26
- package/lib/types/src/huggingface_ie.d.ts +0 -34
- package/lib/types/src/index.d.ts +0 -13
- package/lib/types/src/mistral/index.d.ts +0 -24
- package/lib/types/src/mistral/types.d.ts +0 -131
- package/lib/types/src/openai/azure_openai.d.ts +0 -24
- package/lib/types/src/openai/index.d.ts +0 -24
- package/lib/types/src/openai/openai.d.ts +0 -14
- package/lib/types/src/openai/openai_format.d.ts +0 -18
- package/lib/types/src/replicate.d.ts +0 -47
- package/lib/types/src/test/TestErrorCompletionStream.d.ts +0 -8
- package/lib/types/src/test/TestValidationErrorCompletionStream.d.ts +0 -8
- package/lib/types/src/test/index.d.ts +0 -23
- package/lib/types/src/test/utils.d.ts +0 -4
- package/lib/types/src/test-driver/TestErrorCompletionStream.d.ts +0 -8
- package/lib/types/src/test-driver/TestValidationErrorCompletionStream.d.ts +0 -8
- package/lib/types/src/test-driver/index.d.ts +0 -23
- package/lib/types/src/test-driver/utils.d.ts +0 -4
- package/lib/types/src/togetherai/index.d.ts +0 -22
- package/lib/types/src/togetherai/interfaces.d.ts +0 -95
- package/lib/types/src/vertexai/debug.d.ts +0 -1
- package/lib/types/src/vertexai/embeddings/embeddings-image.d.ts +0 -10
- package/lib/types/src/vertexai/embeddings/embeddings-text.d.ts +0 -9
- package/lib/types/src/vertexai/index.d.ts +0 -52
- package/lib/types/src/vertexai/models/claude.d.ts +0 -19
- package/lib/types/src/vertexai/models/gemini.d.ts +0 -17
- package/lib/types/src/vertexai/models/imagen.d.ts +0 -74
- package/lib/types/src/vertexai/models/llama.d.ts +0 -19
- package/lib/types/src/vertexai/models.d.ts +0 -14
- package/lib/types/src/watsonx/index.d.ts +0 -26
- package/lib/types/src/watsonx/interfaces.d.ts +0 -64
- package/lib/types/src/xai/index.d.ts +0 -18
- package/lib/types/test-driver/TestErrorCompletionStream.d.ts +0 -9
- package/lib/types/test-driver/TestErrorCompletionStream.d.ts.map +0 -1
- package/lib/types/test-driver/TestValidationErrorCompletionStream.d.ts +0 -9
- package/lib/types/test-driver/TestValidationErrorCompletionStream.d.ts.map +0 -1
- package/lib/types/test-driver/index.d.ts +0 -24
- package/lib/types/test-driver/index.d.ts.map +0 -1
- package/lib/types/test-driver/utils.d.ts +0 -5
- package/lib/types/test-driver/utils.d.ts.map +0 -1
|
@@ -1,29 +0,0 @@
|
|
|
1
|
-
import { AbstractDriver, AIModel, Completion, CompletionChunkObject, DriverOptions, EmbeddingsOptions, EmbeddingsResult, ExecutionOptions, ModelSearchPayload, PromptSegment } from "@llumiverse/core";
|
|
2
|
-
interface FireflySize {
|
|
3
|
-
width: number;
|
|
4
|
-
height: number;
|
|
5
|
-
}
|
|
6
|
-
export interface FireflyDriverOptions extends DriverOptions {
|
|
7
|
-
/**
|
|
8
|
-
* Adobe Firefly API key
|
|
9
|
-
*/
|
|
10
|
-
apiKey: string;
|
|
11
|
-
/**
|
|
12
|
-
* Optional API endpoint override
|
|
13
|
-
*/
|
|
14
|
-
endpoint?: string;
|
|
15
|
-
}
|
|
16
|
-
export declare class FireflyDriver extends AbstractDriver<FireflyDriverOptions> {
|
|
17
|
-
static PROVIDER: string;
|
|
18
|
-
provider: string;
|
|
19
|
-
private readonly endpoint;
|
|
20
|
-
constructor(options: FireflyDriverOptions);
|
|
21
|
-
requestTextCompletion(_prompt: string, _options: ExecutionOptions): Promise<Completion>;
|
|
22
|
-
requestTextCompletionStream(_prompt: string, _options: ExecutionOptions): Promise<AsyncIterable<CompletionChunkObject>>;
|
|
23
|
-
requestImageGeneration(segments: PromptSegment[], options: ExecutionOptions): Promise<Completion>;
|
|
24
|
-
mapSize(size?: string): FireflySize;
|
|
25
|
-
listModels(_params?: ModelSearchPayload): Promise<AIModel[]>;
|
|
26
|
-
validateConnection(): Promise<boolean>;
|
|
27
|
-
generateEmbeddings(_options: EmbeddingsOptions): Promise<EmbeddingsResult>;
|
|
28
|
-
}
|
|
29
|
-
export {};
|
|
@@ -1,49 +0,0 @@
|
|
|
1
|
-
import { TokenCredential } from "@azure/identity";
|
|
2
|
-
import { AbstractDriver, AIModel, Completion, CompletionChunkObject, DriverOptions, EmbeddingsOptions, EmbeddingsResult, ExecutionOptions, Providers } from "@llumiverse/core";
|
|
3
|
-
import { AIProjectClient, ModelDeployment } from '@azure/ai-projects';
|
|
4
|
-
import { ChatCompletionMessageParam } from "openai/resources";
|
|
5
|
-
import type { ChatRequestMessage } from "@azure-rest/ai-inference";
|
|
6
|
-
export interface AzureFoundryDriverOptions extends DriverOptions {
|
|
7
|
-
/**
|
|
8
|
-
* The credentials to use to access Azure AI Foundry
|
|
9
|
-
*/
|
|
10
|
-
azureADTokenProvider?: TokenCredential;
|
|
11
|
-
endpoint?: string;
|
|
12
|
-
apiVersion?: string;
|
|
13
|
-
}
|
|
14
|
-
export interface AzureFoundryInferencePrompt {
|
|
15
|
-
messages: ChatRequestMessage[];
|
|
16
|
-
}
|
|
17
|
-
export interface AzureFoundryOpenAIPrompt {
|
|
18
|
-
messages: ChatCompletionMessageParam[];
|
|
19
|
-
}
|
|
20
|
-
export type AzureFoundryPrompt = AzureFoundryInferencePrompt | AzureFoundryOpenAIPrompt;
|
|
21
|
-
export declare class AzureFoundryDriver extends AbstractDriver<AzureFoundryDriverOptions, ChatCompletionMessageParam[]> {
|
|
22
|
-
service: AIProjectClient;
|
|
23
|
-
readonly provider = Providers.azure_foundry;
|
|
24
|
-
OPENAI_API_VERSION: string;
|
|
25
|
-
INFERENCE_API_VERSION: string;
|
|
26
|
-
constructor(opts: AzureFoundryDriverOptions);
|
|
27
|
-
/**
|
|
28
|
-
* Get default authentication for Azure AI Foundry API
|
|
29
|
-
*/
|
|
30
|
-
getDefaultAIFoundryAuth(): () => Promise<string>;
|
|
31
|
-
isOpenAIDeployment(model: string): Promise<boolean>;
|
|
32
|
-
protected canStream(_options: ExecutionOptions): Promise<boolean>;
|
|
33
|
-
requestTextCompletion(prompt: ChatCompletionMessageParam[], options: ExecutionOptions): Promise<Completion>;
|
|
34
|
-
requestTextCompletionStream(prompt: ChatCompletionMessageParam[], options: ExecutionOptions): Promise<AsyncIterable<CompletionChunkObject>>;
|
|
35
|
-
private processStreamResponse;
|
|
36
|
-
private extractDataFromResponse;
|
|
37
|
-
private convertFinishReason;
|
|
38
|
-
validateConnection(): Promise<boolean>;
|
|
39
|
-
generateEmbeddings(options: EmbeddingsOptions): Promise<EmbeddingsResult>;
|
|
40
|
-
generateTextEmbeddings(options: EmbeddingsOptions): Promise<EmbeddingsResult>;
|
|
41
|
-
generateImageEmbeddings(options: EmbeddingsOptions): Promise<EmbeddingsResult>;
|
|
42
|
-
listModels(): Promise<AIModel[]>;
|
|
43
|
-
_listModels(filter?: (m: ModelDeployment) => boolean): Promise<AIModel[]>;
|
|
44
|
-
}
|
|
45
|
-
export declare function parseAzureFoundryModelId(compositeId: string): {
|
|
46
|
-
deploymentName: string;
|
|
47
|
-
baseModel: string;
|
|
48
|
-
};
|
|
49
|
-
export declare function isCompositeModelId(modelId: string): boolean;
|
|
@@ -1,8 +0,0 @@
|
|
|
1
|
-
import { ExecutionOptions } from "@llumiverse/core";
|
|
2
|
-
import { PromptSegment } from "@llumiverse/core";
|
|
3
|
-
import { ConverseRequest, Message, SystemContentBlock } from "@aws-sdk/client-bedrock-runtime";
|
|
4
|
-
export declare function converseConcatMessages(messages: Message[] | undefined): Message[];
|
|
5
|
-
export declare function converseSystemToMessages(system: SystemContentBlock[]): Message;
|
|
6
|
-
export declare function converseRemoveJSONprefill(messages: Message[] | undefined): Message[];
|
|
7
|
-
export declare function converseJSONprefill(messages: Message[] | undefined): Message[];
|
|
8
|
-
export declare function formatConversePrompt(segments: PromptSegment[], options: ExecutionOptions): Promise<ConverseRequest>;
|
|
@@ -1,61 +0,0 @@
|
|
|
1
|
-
import { Bedrock, FoundationModelSummary } from "@aws-sdk/client-bedrock";
|
|
2
|
-
import { BedrockRuntime, ConverseRequest, ConverseResponse, ConverseStreamOutput } from "@aws-sdk/client-bedrock-runtime";
|
|
3
|
-
import { AwsCredentialIdentity, Provider } from "@aws-sdk/types";
|
|
4
|
-
import { AbstractDriver, AIModel, Completion, CompletionChunkObject, DataSource, DriverOptions, EmbeddingsOptions, EmbeddingsResult, ExecutionOptions, PromptSegment, TrainingJob, TrainingOptions } from "@llumiverse/core";
|
|
5
|
-
import { NovaMessagesPrompt } from "@llumiverse/core/formatters";
|
|
6
|
-
import { TwelvelabsPegasusRequest } from "./twelvelabs.js";
|
|
7
|
-
export interface BedrockModelCapabilities {
|
|
8
|
-
name: string;
|
|
9
|
-
canStream: boolean;
|
|
10
|
-
}
|
|
11
|
-
export interface BedrockDriverOptions extends DriverOptions {
|
|
12
|
-
/**
|
|
13
|
-
* The AWS region
|
|
14
|
-
*/
|
|
15
|
-
region: string;
|
|
16
|
-
/**
|
|
17
|
-
* The bucket name to be used for training.
|
|
18
|
-
* It will be created if does not already exist.
|
|
19
|
-
*/
|
|
20
|
-
training_bucket?: string;
|
|
21
|
-
/**
|
|
22
|
-
* The role ARN to be used for training
|
|
23
|
-
*/
|
|
24
|
-
training_role_arn?: string;
|
|
25
|
-
/**
|
|
26
|
-
* The credentials to use to access AWS
|
|
27
|
-
*/
|
|
28
|
-
credentials?: AwsCredentialIdentity | Provider<AwsCredentialIdentity>;
|
|
29
|
-
}
|
|
30
|
-
export type BedrockPrompt = NovaMessagesPrompt | ConverseRequest | TwelvelabsPegasusRequest;
|
|
31
|
-
export declare class BedrockDriver extends AbstractDriver<BedrockDriverOptions, BedrockPrompt> {
|
|
32
|
-
static PROVIDER: string;
|
|
33
|
-
provider: string;
|
|
34
|
-
private _executor?;
|
|
35
|
-
private _service?;
|
|
36
|
-
private _service_region?;
|
|
37
|
-
constructor(options: BedrockDriverOptions);
|
|
38
|
-
getExecutor(): BedrockRuntime;
|
|
39
|
-
getService(region?: string): Bedrock;
|
|
40
|
-
protected formatPrompt(segments: PromptSegment[], opts: ExecutionOptions): Promise<BedrockPrompt>;
|
|
41
|
-
getExtractedExecution(result: ConverseResponse, _prompt?: BedrockPrompt, options?: ExecutionOptions): CompletionChunkObject;
|
|
42
|
-
getExtractedStream(result: ConverseStreamOutput, _prompt?: BedrockPrompt, options?: ExecutionOptions): CompletionChunkObject;
|
|
43
|
-
extractRegion(modelString: string, defaultRegion: string): string;
|
|
44
|
-
private getCanStream;
|
|
45
|
-
protected canStream(options: ExecutionOptions): Promise<boolean>;
|
|
46
|
-
requestTextCompletion(prompt: BedrockPrompt, options: ExecutionOptions): Promise<Completion>;
|
|
47
|
-
private requestTwelvelabsPegasusCompletion;
|
|
48
|
-
private requestTwelvelabsPegasusCompletionStream;
|
|
49
|
-
requestTextCompletionStream(prompt: BedrockPrompt, options: ExecutionOptions): Promise<AsyncIterable<CompletionChunkObject>>;
|
|
50
|
-
preparePayload(prompt: ConverseRequest, options: ExecutionOptions): ConverseRequest;
|
|
51
|
-
requestImageGeneration(prompt: NovaMessagesPrompt, options: ExecutionOptions): Promise<Completion>;
|
|
52
|
-
startTraining(dataset: DataSource, options: TrainingOptions): Promise<TrainingJob>;
|
|
53
|
-
cancelTraining(jobId: string): Promise<TrainingJob>;
|
|
54
|
-
getTrainingJob(jobId: string): Promise<TrainingJob>;
|
|
55
|
-
validateConnection(): Promise<boolean>;
|
|
56
|
-
listTrainableModels(): Promise<AIModel<string>[]>;
|
|
57
|
-
listModels(): Promise<AIModel[]>;
|
|
58
|
-
_listModels(foundationFilter?: (m: FoundationModelSummary) => boolean): Promise<AIModel[]>;
|
|
59
|
-
generateEmbeddings({ text, image, model }: EmbeddingsOptions): Promise<EmbeddingsResult>;
|
|
60
|
-
private generateTwelvelabsMarengoEmbeddings;
|
|
61
|
-
}
|
|
@@ -1,73 +0,0 @@
|
|
|
1
|
-
import { ExecutionOptions } from "@llumiverse/core";
|
|
2
|
-
import { NovaMessagesPrompt } from "@llumiverse/core/formatters";
|
|
3
|
-
export declare function formatNovaImageGenerationPayload(taskType: string, prompt: NovaMessagesPrompt, options: ExecutionOptions): Promise<NovaTextToImagePayload> | Promise<NovaColorGuidedGenerationPayload> | Promise<NovaImageVariationPayload> | Promise<NovaInpaintingPayload> | Promise<NovaOutpaintingPayload> | Promise<NovaBackgroundRemovalPayload>;
|
|
4
|
-
export interface InvokeModelPayloadBase {
|
|
5
|
-
taskType: NovaImageGenerationTaskType;
|
|
6
|
-
imageGenerationConfig: {
|
|
7
|
-
width?: number;
|
|
8
|
-
height?: number;
|
|
9
|
-
quality?: "standard" | "premium";
|
|
10
|
-
cfgScale?: number;
|
|
11
|
-
seed?: number;
|
|
12
|
-
numberOfImages?: number;
|
|
13
|
-
};
|
|
14
|
-
}
|
|
15
|
-
export interface NovaTextToImagePayload extends InvokeModelPayloadBase {
|
|
16
|
-
textToImageParams: {
|
|
17
|
-
conditionImage?: string;
|
|
18
|
-
controlMode?: "CANNY_EDGE" | "SEGMENTATION";
|
|
19
|
-
controlStrength?: number;
|
|
20
|
-
text: string;
|
|
21
|
-
negativeText?: string;
|
|
22
|
-
};
|
|
23
|
-
}
|
|
24
|
-
export interface NovaImageVariationPayload extends InvokeModelPayloadBase {
|
|
25
|
-
imageVariationParams: {
|
|
26
|
-
images: string[];
|
|
27
|
-
similarityStrength?: number;
|
|
28
|
-
text?: string;
|
|
29
|
-
negativeText?: string;
|
|
30
|
-
};
|
|
31
|
-
}
|
|
32
|
-
export interface NovaColorGuidedGenerationPayload extends InvokeModelPayloadBase {
|
|
33
|
-
colorGuidedGenerationParams: {
|
|
34
|
-
colors: string[];
|
|
35
|
-
text: string;
|
|
36
|
-
referenceImage?: string;
|
|
37
|
-
negativeText?: string;
|
|
38
|
-
};
|
|
39
|
-
}
|
|
40
|
-
export interface NovaInpaintingPayload extends InvokeModelPayloadBase {
|
|
41
|
-
inPaintingParams: {
|
|
42
|
-
image: string;
|
|
43
|
-
maskImage?: string;
|
|
44
|
-
maskPrompt?: string;
|
|
45
|
-
negativeText?: string;
|
|
46
|
-
text?: string;
|
|
47
|
-
};
|
|
48
|
-
}
|
|
49
|
-
export interface NovaOutpaintingPayload extends InvokeModelPayloadBase {
|
|
50
|
-
outPaintingParams: {
|
|
51
|
-
image: string;
|
|
52
|
-
maskImage?: string;
|
|
53
|
-
maskPrompt?: string;
|
|
54
|
-
negativeText?: string;
|
|
55
|
-
text?: string;
|
|
56
|
-
outPaintingMode: "DEFAULT" | "PRECISE";
|
|
57
|
-
};
|
|
58
|
-
}
|
|
59
|
-
export interface NovaBackgroundRemovalPayload {
|
|
60
|
-
taskType: NovaImageGenerationTaskType.BACKGROUND_REMOVAL;
|
|
61
|
-
backgroundRemovalParams: {
|
|
62
|
-
image: string;
|
|
63
|
-
};
|
|
64
|
-
}
|
|
65
|
-
export declare enum NovaImageGenerationTaskType {
|
|
66
|
-
TEXT_IMAGE = "TEXT_IMAGE",
|
|
67
|
-
TEXT_IMAGE_WITH_IMAGE_CONDITIONING = "TEXT_IMAGE_WITH_IMAGE_CONDITIONING",
|
|
68
|
-
COLOR_GUIDED_GENERATION = "COLOR_GUIDED_GENERATION",
|
|
69
|
-
IMAGE_VARIATION = "IMAGE_VARIATION",
|
|
70
|
-
INPAINTING = "INPAINTING",
|
|
71
|
-
OUTPAINTING = "OUTPAINTING",
|
|
72
|
-
BACKGROUND_REMOVAL = "BACKGROUND_REMOVAL"
|
|
73
|
-
}
|
|
@@ -1,11 +0,0 @@
|
|
|
1
|
-
import { NovaMessagesPrompt } from "@llumiverse/core/formatters";
|
|
2
|
-
export interface NovaPayload extends NovaMessagesPrompt {
|
|
3
|
-
schemaVersion: string;
|
|
4
|
-
inferenceConfig?: {
|
|
5
|
-
max_new_tokens?: number;
|
|
6
|
-
temperature?: number;
|
|
7
|
-
top_p?: number;
|
|
8
|
-
top_k?: number;
|
|
9
|
-
stopSequences?: [string];
|
|
10
|
-
};
|
|
11
|
-
}
|
|
@@ -1,22 +0,0 @@
|
|
|
1
|
-
import { S3Client } from "@aws-sdk/client-s3";
|
|
2
|
-
import { Progress } from "@aws-sdk/lib-storage";
|
|
3
|
-
export declare function doesBucketExist(s3: S3Client, bucketName: string): Promise<boolean>;
|
|
4
|
-
export declare function createBucket(s3: S3Client, bucketName: string): Promise<import("@aws-sdk/client-s3").CreateBucketCommandOutput>;
|
|
5
|
-
export declare function tryCreateBucket(s3: S3Client, bucketName: string): Promise<import("@aws-sdk/client-s3").CreateBucketCommandOutput | undefined>;
|
|
6
|
-
export declare function uploadFile(s3: S3Client, source: ReadableStream, bucketName: string, file: string, onProgress?: (progress: Progress) => void): Promise<import("@aws-sdk/client-s3").CompleteMultipartUploadCommandOutput>;
|
|
7
|
-
/**
|
|
8
|
-
* Create the bucket if not already exists and then upload the file.
|
|
9
|
-
* @param s3
|
|
10
|
-
* @param source
|
|
11
|
-
* @param bucketName
|
|
12
|
-
* @param file
|
|
13
|
-
* @param onProgress
|
|
14
|
-
* @returns
|
|
15
|
-
*/
|
|
16
|
-
export declare function forceUploadFile(s3: S3Client, source: ReadableStream, bucketName: string, file: string, onProgress?: (progress: Progress) => void): Promise<import("@aws-sdk/client-s3").CompleteMultipartUploadCommandOutput>;
|
|
17
|
-
/**
|
|
18
|
-
* Parse an S3 HTTPS URL into an S3 URI format
|
|
19
|
-
* s3Url - The S3 HTTPS URL (e.g., https://bucket.s3.region.amazonaws.com/key)
|
|
20
|
-
* returns The S3 URI (e.g., s3://bucket/key)
|
|
21
|
-
*/
|
|
22
|
-
export declare function parseS3UrlToUri(s3Url: URL): string;
|
|
@@ -1,49 +0,0 @@
|
|
|
1
|
-
import { ExecutionOptions } from "@llumiverse/core";
|
|
2
|
-
import { PromptSegment } from "@llumiverse/core";
|
|
3
|
-
export interface TwelvelabsPegasusRequest {
|
|
4
|
-
inputPrompt: string;
|
|
5
|
-
temperature?: number;
|
|
6
|
-
responseFormat?: {
|
|
7
|
-
type: "json_schema";
|
|
8
|
-
json_schema: {
|
|
9
|
-
name: string;
|
|
10
|
-
schema: any;
|
|
11
|
-
};
|
|
12
|
-
};
|
|
13
|
-
mediaSource: {
|
|
14
|
-
base64String?: string;
|
|
15
|
-
s3Location?: {
|
|
16
|
-
uri: string;
|
|
17
|
-
bucketOwner?: string;
|
|
18
|
-
};
|
|
19
|
-
};
|
|
20
|
-
maxOutputTokens?: number;
|
|
21
|
-
}
|
|
22
|
-
export interface TwelvelabsPegasusResponse {
|
|
23
|
-
message: string;
|
|
24
|
-
finishReason: "stop" | "length";
|
|
25
|
-
}
|
|
26
|
-
export interface TwelvelabsMarengoRequest {
|
|
27
|
-
inputType: "text" | "image" | "video" | "audio";
|
|
28
|
-
inputText?: string;
|
|
29
|
-
textTruncate?: "start" | "end";
|
|
30
|
-
mediaSource?: {
|
|
31
|
-
base64String?: string;
|
|
32
|
-
s3Location?: {
|
|
33
|
-
uri: string;
|
|
34
|
-
bucketOwner?: string;
|
|
35
|
-
};
|
|
36
|
-
};
|
|
37
|
-
embeddingOption?: "visual-text" | "visual-image" | "audio";
|
|
38
|
-
startSec?: number;
|
|
39
|
-
lengthSec?: number;
|
|
40
|
-
useFixedLengthSec?: boolean;
|
|
41
|
-
minClipSec?: number;
|
|
42
|
-
}
|
|
43
|
-
export interface TwelvelabsMarengoResponse {
|
|
44
|
-
embedding: number[];
|
|
45
|
-
embeddingOption: "visual-text" | "visual-image" | "audio";
|
|
46
|
-
startSec: number;
|
|
47
|
-
endSec: number;
|
|
48
|
-
}
|
|
49
|
-
export declare function formatTwelvelabsPegasusPrompt(segments: PromptSegment[], options: ExecutionOptions): Promise<TwelvelabsPegasusRequest>;
|
|
@@ -1,26 +0,0 @@
|
|
|
1
|
-
import { AIModel, AbstractDriver, Completion, CompletionChunkObject, DriverOptions, EmbeddingsOptions, EmbeddingsResult, ExecutionOptions, PromptSegment } from "@llumiverse/core";
|
|
2
|
-
import Groq from "groq-sdk";
|
|
3
|
-
import type { ChatCompletionMessageParam } from "groq-sdk/resources/chat/completions";
|
|
4
|
-
interface GroqDriverOptions extends DriverOptions {
|
|
5
|
-
apiKey: string;
|
|
6
|
-
endpoint_url?: string;
|
|
7
|
-
}
|
|
8
|
-
export declare class GroqDriver extends AbstractDriver<GroqDriverOptions, ChatCompletionMessageParam[]> {
|
|
9
|
-
static PROVIDER: string;
|
|
10
|
-
provider: string;
|
|
11
|
-
apiKey: string;
|
|
12
|
-
client: Groq;
|
|
13
|
-
endpointUrl?: string;
|
|
14
|
-
constructor(options: GroqDriverOptions);
|
|
15
|
-
getResponseFormat(_options: ExecutionOptions): undefined;
|
|
16
|
-
protected formatPrompt(segments: PromptSegment[], opts: ExecutionOptions): Promise<ChatCompletionMessageParam[]>;
|
|
17
|
-
private getToolDefinitions;
|
|
18
|
-
private extractToolUse;
|
|
19
|
-
private sanitizeMessagesForGroq;
|
|
20
|
-
requestTextCompletion(messages: ChatCompletionMessageParam[], options: ExecutionOptions): Promise<Completion>;
|
|
21
|
-
requestTextCompletionStream(messages: ChatCompletionMessageParam[], options: ExecutionOptions): Promise<AsyncIterable<CompletionChunkObject>>;
|
|
22
|
-
listModels(): Promise<AIModel<string>[]>;
|
|
23
|
-
validateConnection(): Promise<boolean>;
|
|
24
|
-
generateEmbeddings({}: EmbeddingsOptions): Promise<EmbeddingsResult>;
|
|
25
|
-
}
|
|
26
|
-
export {};
|
|
@@ -1,34 +0,0 @@
|
|
|
1
|
-
import { InferenceClient } from "@huggingface/inference";
|
|
2
|
-
import { AIModel, AbstractDriver, CompletionChunkObject, DriverOptions, EmbeddingsResult, ExecutionOptions } from "@llumiverse/core";
|
|
3
|
-
import { FetchClient } from "@vertesia/api-fetch-client";
|
|
4
|
-
export interface HuggingFaceIEDriverOptions extends DriverOptions {
|
|
5
|
-
apiKey: string;
|
|
6
|
-
endpoint_url: string;
|
|
7
|
-
}
|
|
8
|
-
export declare class HuggingFaceIEDriver extends AbstractDriver<HuggingFaceIEDriverOptions, string> {
|
|
9
|
-
static PROVIDER: string;
|
|
10
|
-
provider: string;
|
|
11
|
-
service: FetchClient;
|
|
12
|
-
_executor?: InferenceClient;
|
|
13
|
-
constructor(options: HuggingFaceIEDriverOptions);
|
|
14
|
-
getModelURLEndpoint(modelId: string): Promise<{
|
|
15
|
-
url: string;
|
|
16
|
-
status: string;
|
|
17
|
-
}>;
|
|
18
|
-
getExecutor(model: string): Promise<InferenceClient>;
|
|
19
|
-
requestTextCompletionStream(prompt: string, options: ExecutionOptions): Promise<AsyncIterable<CompletionChunkObject>>;
|
|
20
|
-
requestTextCompletion(prompt: string, options: ExecutionOptions): Promise<{
|
|
21
|
-
result: {
|
|
22
|
-
type: "text";
|
|
23
|
-
value: string;
|
|
24
|
-
}[];
|
|
25
|
-
finish_reason: string;
|
|
26
|
-
token_usage: {
|
|
27
|
-
result: number | undefined;
|
|
28
|
-
};
|
|
29
|
-
original_response: import("@huggingface/inference").TextGenerationOutput | undefined;
|
|
30
|
-
}>;
|
|
31
|
-
listModels(): Promise<AIModel[]>;
|
|
32
|
-
validateConnection(): Promise<boolean>;
|
|
33
|
-
generateEmbeddings(): Promise<EmbeddingsResult>;
|
|
34
|
-
}
|
package/lib/types/src/index.d.ts
DELETED
|
@@ -1,13 +0,0 @@
|
|
|
1
|
-
export * from "./azure/azure_foundry.js";
|
|
2
|
-
export * from "./bedrock/index.js";
|
|
3
|
-
export * from "./groq/index.js";
|
|
4
|
-
export * from "./huggingface_ie.js";
|
|
5
|
-
export * from "./mistral/index.js";
|
|
6
|
-
export * from "./openai/azure_openai.js";
|
|
7
|
-
export * from "./openai/openai.js";
|
|
8
|
-
export * from "./replicate.js";
|
|
9
|
-
export * from "./test-driver/index.js";
|
|
10
|
-
export * from "./togetherai/index.js";
|
|
11
|
-
export * from "./vertexai/index.js";
|
|
12
|
-
export * from "./watsonx/index.js";
|
|
13
|
-
export * from "./xai/index.js";
|
|
@@ -1,24 +0,0 @@
|
|
|
1
|
-
import { AIModel, AbstractDriver, Completion, CompletionChunkObject, DriverOptions, EmbeddingsOptions, EmbeddingsResult, ExecutionOptions, PromptSegment } from "@llumiverse/core";
|
|
2
|
-
import { OpenAITextMessage } from "../openai/openai_format.js";
|
|
3
|
-
import { FetchClient } from "@vertesia/api-fetch-client";
|
|
4
|
-
import { ResponseFormat } from "./types.js";
|
|
5
|
-
interface MistralAIDriverOptions extends DriverOptions {
|
|
6
|
-
apiKey: string;
|
|
7
|
-
endpoint_url?: string;
|
|
8
|
-
}
|
|
9
|
-
export declare class MistralAIDriver extends AbstractDriver<MistralAIDriverOptions, OpenAITextMessage[]> {
|
|
10
|
-
static PROVIDER: string;
|
|
11
|
-
provider: string;
|
|
12
|
-
apiKey: string;
|
|
13
|
-
client: FetchClient;
|
|
14
|
-
endpointUrl?: string;
|
|
15
|
-
constructor(options: MistralAIDriverOptions);
|
|
16
|
-
getResponseFormat: (_options: ExecutionOptions) => ResponseFormat | undefined;
|
|
17
|
-
protected formatPrompt(segments: PromptSegment[], opts: ExecutionOptions): Promise<OpenAITextMessage[]>;
|
|
18
|
-
requestTextCompletion(messages: OpenAITextMessage[], options: ExecutionOptions): Promise<Completion>;
|
|
19
|
-
requestTextCompletionStream(messages: OpenAITextMessage[], options: ExecutionOptions): Promise<AsyncIterable<CompletionChunkObject>>;
|
|
20
|
-
listModels(): Promise<AIModel<string>[]>;
|
|
21
|
-
validateConnection(): Promise<boolean>;
|
|
22
|
-
generateEmbeddings({ text, model }: EmbeddingsOptions): Promise<EmbeddingsResult>;
|
|
23
|
-
}
|
|
24
|
-
export {};
|
|
@@ -1,131 +0,0 @@
|
|
|
1
|
-
export interface ModelPermission {
|
|
2
|
-
id: string;
|
|
3
|
-
object: 'model_permission';
|
|
4
|
-
created: number;
|
|
5
|
-
allow_create_engine: boolean;
|
|
6
|
-
allow_sampling: boolean;
|
|
7
|
-
allow_logprobs: boolean;
|
|
8
|
-
allow_search_indices: boolean;
|
|
9
|
-
allow_view: boolean;
|
|
10
|
-
allow_fine_tuning: boolean;
|
|
11
|
-
organization: string;
|
|
12
|
-
group: string | null;
|
|
13
|
-
is_blocking: boolean;
|
|
14
|
-
}
|
|
15
|
-
export interface Model {
|
|
16
|
-
id: string;
|
|
17
|
-
object: 'model';
|
|
18
|
-
created: number;
|
|
19
|
-
owned_by: string;
|
|
20
|
-
root: string | null;
|
|
21
|
-
parent: string | null;
|
|
22
|
-
permission: ModelPermission[];
|
|
23
|
-
}
|
|
24
|
-
export interface ListModelsResponse {
|
|
25
|
-
object: 'list';
|
|
26
|
-
data: Model[];
|
|
27
|
-
}
|
|
28
|
-
export interface Function {
|
|
29
|
-
name: string;
|
|
30
|
-
description: string;
|
|
31
|
-
parameters: object;
|
|
32
|
-
}
|
|
33
|
-
export declare enum ToolType {
|
|
34
|
-
function = "function"
|
|
35
|
-
}
|
|
36
|
-
export interface FunctionCall {
|
|
37
|
-
name: string;
|
|
38
|
-
arguments: string;
|
|
39
|
-
}
|
|
40
|
-
export interface ToolCalls {
|
|
41
|
-
id: 'null';
|
|
42
|
-
type: ToolType;
|
|
43
|
-
function: FunctionCall;
|
|
44
|
-
}
|
|
45
|
-
export declare enum ResponseFormats {
|
|
46
|
-
text = "text",
|
|
47
|
-
json_object = "json_object"
|
|
48
|
-
}
|
|
49
|
-
export declare enum ToolChoice {
|
|
50
|
-
auto = "auto",
|
|
51
|
-
any = "any",
|
|
52
|
-
none = "none"
|
|
53
|
-
}
|
|
54
|
-
export interface ResponseFormat {
|
|
55
|
-
type: ResponseFormats;
|
|
56
|
-
}
|
|
57
|
-
export interface TokenUsage {
|
|
58
|
-
prompt_tokens: number;
|
|
59
|
-
completion_tokens: number;
|
|
60
|
-
total_tokens: number;
|
|
61
|
-
}
|
|
62
|
-
export interface ChatCompletionResponseChoice {
|
|
63
|
-
index: number;
|
|
64
|
-
message: {
|
|
65
|
-
role: string;
|
|
66
|
-
content: string;
|
|
67
|
-
};
|
|
68
|
-
finish_reason: string;
|
|
69
|
-
}
|
|
70
|
-
export interface ChatCompletionResponseChunkChoice {
|
|
71
|
-
index: number;
|
|
72
|
-
delta: {
|
|
73
|
-
role?: string;
|
|
74
|
-
content?: string;
|
|
75
|
-
tool_calls?: ToolCalls[];
|
|
76
|
-
};
|
|
77
|
-
finish_reason: string;
|
|
78
|
-
}
|
|
79
|
-
export interface ChatCompletionResponse {
|
|
80
|
-
id: string;
|
|
81
|
-
object: 'chat.completion';
|
|
82
|
-
created: number;
|
|
83
|
-
model: string;
|
|
84
|
-
choices: ChatCompletionResponseChoice[];
|
|
85
|
-
usage: TokenUsage;
|
|
86
|
-
}
|
|
87
|
-
export interface ChatCompletionResponseChunk {
|
|
88
|
-
id: string;
|
|
89
|
-
object: 'chat.completion.chunk';
|
|
90
|
-
created: number;
|
|
91
|
-
model: string;
|
|
92
|
-
choices: ChatCompletionResponseChunkChoice[];
|
|
93
|
-
}
|
|
94
|
-
export interface Embedding {
|
|
95
|
-
id: string;
|
|
96
|
-
object: 'embedding';
|
|
97
|
-
embedding: number[];
|
|
98
|
-
}
|
|
99
|
-
export interface EmbeddingResponse {
|
|
100
|
-
id: string;
|
|
101
|
-
object: 'list';
|
|
102
|
-
data: Embedding[];
|
|
103
|
-
model: string;
|
|
104
|
-
usage: TokenUsage;
|
|
105
|
-
}
|
|
106
|
-
export interface CompletionRequestParams {
|
|
107
|
-
model: string;
|
|
108
|
-
messages: Array<{
|
|
109
|
-
role: string;
|
|
110
|
-
name?: string;
|
|
111
|
-
content: string | string[];
|
|
112
|
-
tool_calls?: ToolCalls[];
|
|
113
|
-
}>;
|
|
114
|
-
tools?: Array<{
|
|
115
|
-
type: string;
|
|
116
|
-
function: Function;
|
|
117
|
-
}>;
|
|
118
|
-
temperature?: number;
|
|
119
|
-
maxTokens?: number;
|
|
120
|
-
topP?: number;
|
|
121
|
-
randomSeed?: number;
|
|
122
|
-
stream?: boolean;
|
|
123
|
-
/**
|
|
124
|
-
* @deprecated use safePrompt instead
|
|
125
|
-
*/
|
|
126
|
-
safeMode?: boolean;
|
|
127
|
-
safePrompt?: boolean;
|
|
128
|
-
toolChoice?: ToolChoice;
|
|
129
|
-
responseFormat?: ResponseFormat;
|
|
130
|
-
stopSequences?: string[];
|
|
131
|
-
}
|
|
@@ -1,24 +0,0 @@
|
|
|
1
|
-
import { AIModel, DriverOptions, Providers } from "@llumiverse/core";
|
|
2
|
-
import OpenAI, { AzureOpenAI } from "openai";
|
|
3
|
-
import { BaseOpenAIDriver } from "./index.js";
|
|
4
|
-
export interface AzureOpenAIDriverOptions extends DriverOptions {
|
|
5
|
-
/**
|
|
6
|
-
* The credentials to use to access Azure OpenAI
|
|
7
|
-
*/
|
|
8
|
-
azureADTokenProvider?: any;
|
|
9
|
-
apiKey?: string;
|
|
10
|
-
endpoint?: string;
|
|
11
|
-
apiVersion?: string;
|
|
12
|
-
deployment?: string;
|
|
13
|
-
}
|
|
14
|
-
export declare class AzureOpenAIDriver extends BaseOpenAIDriver {
|
|
15
|
-
service: AzureOpenAI;
|
|
16
|
-
readonly provider = Providers.azure_openai;
|
|
17
|
-
constructor(serviceOrOpts: AzureOpenAI | AzureOpenAIDriverOptions);
|
|
18
|
-
/**
|
|
19
|
-
* Get default authentication for Azure Cognitive Services API
|
|
20
|
-
*/
|
|
21
|
-
getDefaultCognitiveServicesAuth(): () => Promise<string>;
|
|
22
|
-
listModels(): Promise<AIModel[]>;
|
|
23
|
-
_listModels(_filter?: (m: OpenAI.Models.Model) => boolean): Promise<AIModel[]>;
|
|
24
|
-
}
|
|
@@ -1,24 +0,0 @@
|
|
|
1
|
-
import { AIModel, AbstractDriver, Completion, CompletionChunkObject, DataSource, DriverOptions, EmbeddingsOptions, EmbeddingsResult, ExecutionOptions, Providers, ToolUse, TrainingJob, TrainingOptions, TrainingPromptOptions } from "@llumiverse/core";
|
|
2
|
-
import OpenAI, { AzureOpenAI } from "openai";
|
|
3
|
-
import { ChatCompletionMessageParam } from "openai/resources/chat/completions";
|
|
4
|
-
export interface BaseOpenAIDriverOptions extends DriverOptions {
|
|
5
|
-
}
|
|
6
|
-
export declare abstract class BaseOpenAIDriver extends AbstractDriver<BaseOpenAIDriverOptions, ChatCompletionMessageParam[]> {
|
|
7
|
-
abstract provider: Providers.openai | Providers.azure_openai | "xai" | Providers.azure_foundry;
|
|
8
|
-
abstract service: OpenAI | AzureOpenAI;
|
|
9
|
-
constructor(opts: BaseOpenAIDriverOptions);
|
|
10
|
-
extractDataFromResponse(_options: ExecutionOptions, result: OpenAI.Chat.Completions.ChatCompletion): Completion;
|
|
11
|
-
requestTextCompletionStream(prompt: ChatCompletionMessageParam[], options: ExecutionOptions): Promise<AsyncIterable<CompletionChunkObject>>;
|
|
12
|
-
requestTextCompletion(prompt: ChatCompletionMessageParam[], options: ExecutionOptions): Promise<Completion>;
|
|
13
|
-
protected canStream(_options: ExecutionOptions): Promise<boolean>;
|
|
14
|
-
createTrainingPrompt(options: TrainingPromptOptions): Promise<string>;
|
|
15
|
-
startTraining(dataset: DataSource, options: TrainingOptions): Promise<TrainingJob>;
|
|
16
|
-
cancelTraining(jobId: string): Promise<TrainingJob>;
|
|
17
|
-
getTrainingJob(jobId: string): Promise<TrainingJob>;
|
|
18
|
-
validateConnection(): Promise<boolean>;
|
|
19
|
-
listTrainableModels(): Promise<AIModel<string>[]>;
|
|
20
|
-
listModels(): Promise<AIModel[]>;
|
|
21
|
-
_listModels(filter?: (m: OpenAI.Models.Model) => boolean): Promise<AIModel[]>;
|
|
22
|
-
generateEmbeddings({ text, image, model }: EmbeddingsOptions): Promise<EmbeddingsResult>;
|
|
23
|
-
}
|
|
24
|
-
export declare function collectTools(toolCalls?: OpenAI.Chat.Completions.ChatCompletionMessageToolCall[]): ToolUse[] | undefined;
|
|
@@ -1,14 +0,0 @@
|
|
|
1
|
-
import { DriverOptions, Providers } from "@llumiverse/core";
|
|
2
|
-
import OpenAI from "openai";
|
|
3
|
-
import { BaseOpenAIDriver } from "./index.js";
|
|
4
|
-
export interface OpenAIDriverOptions extends DriverOptions {
|
|
5
|
-
/**
|
|
6
|
-
* The OpenAI api key
|
|
7
|
-
*/
|
|
8
|
-
apiKey?: string;
|
|
9
|
-
}
|
|
10
|
-
export declare class OpenAIDriver extends BaseOpenAIDriver {
|
|
11
|
-
service: OpenAI;
|
|
12
|
-
readonly provider = Providers.openai;
|
|
13
|
-
constructor(opts: OpenAIDriverOptions);
|
|
14
|
-
}
|
|
@@ -1,18 +0,0 @@
|
|
|
1
|
-
import { PromptOptions, PromptSegment } from "@llumiverse/common";
|
|
2
|
-
import type { ChatCompletionMessageParam } from 'openai/resources/chat/completions';
|
|
3
|
-
export interface OpenAITextMessage {
|
|
4
|
-
content: string;
|
|
5
|
-
role: 'system' | 'user' | 'assistant' | 'developer';
|
|
6
|
-
}
|
|
7
|
-
/**
|
|
8
|
-
* OpenAI text only prompts
|
|
9
|
-
* @param segments
|
|
10
|
-
* @returns
|
|
11
|
-
*/
|
|
12
|
-
export declare function formatOpenAILikeTextPrompt(segments: PromptSegment[]): OpenAITextMessage[];
|
|
13
|
-
export declare function formatOpenAILikeMultimodalPrompt(segments: PromptSegment[], opts: PromptOptions & OpenAIPromptFormatterOptions): Promise<ChatCompletionMessageParam[]>;
|
|
14
|
-
export interface OpenAIPromptFormatterOptions {
|
|
15
|
-
multimodal?: boolean;
|
|
16
|
-
useToolForFormatting?: boolean;
|
|
17
|
-
schema?: Object;
|
|
18
|
-
}
|