@llumiverse/drivers 0.10.0 → 0.11.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +1 -7
- package/lib/cjs/bedrock/index.js +28 -5
- package/lib/cjs/bedrock/index.js.map +1 -1
- package/lib/cjs/bedrock/s3.js.map +1 -1
- package/lib/cjs/huggingface_ie.js +0 -1
- package/lib/cjs/huggingface_ie.js.map +1 -1
- package/lib/cjs/mistral/index.js +27 -21
- package/lib/cjs/mistral/index.js.map +1 -1
- package/lib/cjs/openai.js +2 -5
- package/lib/cjs/openai.js.map +1 -1
- package/lib/cjs/replicate.js +0 -1
- package/lib/cjs/replicate.js.map +1 -1
- package/lib/cjs/togetherai/index.js +0 -3
- package/lib/cjs/togetherai/index.js.map +1 -1
- package/lib/cjs/vertexai/index.js +2 -3
- package/lib/cjs/vertexai/index.js.map +1 -1
- package/lib/cjs/vertexai/models/codey-chat.js +2 -2
- package/lib/cjs/vertexai/models/codey-chat.js.map +1 -1
- package/lib/cjs/vertexai/models/codey-text.js +2 -2
- package/lib/cjs/vertexai/models/codey-text.js.map +1 -1
- package/lib/cjs/vertexai/models/gemini.js +2 -2
- package/lib/cjs/vertexai/models/gemini.js.map +1 -1
- package/lib/cjs/vertexai/models/palm-model-base.js.map +1 -1
- package/lib/cjs/vertexai/models/palm2-chat.js +2 -2
- package/lib/cjs/vertexai/models/palm2-chat.js.map +1 -1
- package/lib/cjs/vertexai/models/palm2-text.js +2 -2
- package/lib/cjs/vertexai/models/palm2-text.js.map +1 -1
- package/lib/esm/bedrock/index.js +29 -6
- package/lib/esm/bedrock/index.js.map +1 -1
- package/lib/esm/bedrock/s3.js.map +1 -1
- package/lib/esm/huggingface_ie.js +1 -2
- package/lib/esm/huggingface_ie.js.map +1 -1
- package/lib/esm/mistral/index.js +28 -22
- package/lib/esm/mistral/index.js.map +1 -1
- package/lib/esm/openai.js +3 -6
- package/lib/esm/openai.js.map +1 -1
- package/lib/esm/replicate.js +1 -2
- package/lib/esm/replicate.js.map +1 -1
- package/lib/esm/src/bedrock/index.js +375 -0
- package/lib/esm/src/bedrock/index.js.map +1 -0
- package/lib/esm/src/bedrock/s3.js +53 -0
- package/lib/esm/src/bedrock/s3.js.map +1 -0
- package/lib/esm/src/huggingface_ie.js +173 -0
- package/lib/esm/src/huggingface_ie.js.map +1 -0
- package/lib/esm/src/index.js +9 -0
- package/lib/esm/src/index.js.map +1 -0
- package/lib/esm/src/mistral/index.js +145 -0
- package/lib/esm/src/mistral/index.js.map +1 -0
- package/lib/esm/src/mistral/types.js +80 -0
- package/lib/esm/src/mistral/types.js.map +1 -0
- package/lib/esm/src/openai.js +195 -0
- package/lib/esm/src/openai.js.map +1 -0
- package/lib/esm/src/replicate.js +281 -0
- package/lib/esm/src/replicate.js.map +1 -0
- package/lib/esm/src/test/TestErrorCompletionStream.js +16 -0
- package/lib/esm/src/test/TestErrorCompletionStream.js.map +1 -0
- package/lib/esm/src/test/TestValidationErrorCompletionStream.js +20 -0
- package/lib/esm/src/test/TestValidationErrorCompletionStream.js.map +1 -0
- package/lib/esm/src/test/index.js +91 -0
- package/lib/esm/src/test/index.js.map +1 -0
- package/lib/esm/src/test/utils.js +25 -0
- package/lib/esm/src/test/utils.js.map +1 -0
- package/lib/esm/src/togetherai/index.js +89 -0
- package/lib/esm/src/togetherai/index.js.map +1 -0
- package/lib/esm/src/togetherai/interfaces.js +2 -0
- package/lib/esm/src/togetherai/interfaces.js.map +1 -0
- package/lib/esm/src/vertexai/debug.js +6 -0
- package/lib/esm/src/vertexai/debug.js.map +1 -0
- package/lib/esm/src/vertexai/embeddings/embeddings-text.js +19 -0
- package/lib/esm/src/vertexai/embeddings/embeddings-text.js.map +1 -0
- package/lib/esm/src/vertexai/index.js +73 -0
- package/lib/esm/src/vertexai/index.js.map +1 -0
- package/lib/esm/src/vertexai/models/codey-chat.js +61 -0
- package/lib/esm/src/vertexai/models/codey-chat.js.map +1 -0
- package/lib/esm/src/vertexai/models/codey-text.js +31 -0
- package/lib/esm/src/vertexai/models/codey-text.js.map +1 -0
- package/lib/esm/src/vertexai/models/gemini.js +136 -0
- package/lib/esm/src/vertexai/models/gemini.js.map +1 -0
- package/lib/esm/src/vertexai/models/palm-model-base.js +53 -0
- package/lib/esm/src/vertexai/models/palm-model-base.js.map +1 -0
- package/lib/esm/src/vertexai/models/palm2-chat.js +61 -0
- package/lib/esm/src/vertexai/models/palm2-chat.js.map +1 -0
- package/lib/esm/src/vertexai/models/palm2-text.js +31 -0
- package/lib/esm/src/vertexai/models/palm2-text.js.map +1 -0
- package/lib/esm/src/vertexai/models.js +87 -0
- package/lib/esm/src/vertexai/models.js.map +1 -0
- package/{src/vertexai/utils/prompts.ts → lib/esm/src/vertexai/utils/prompts.js} +10 -29
- package/lib/esm/src/vertexai/utils/prompts.js.map +1 -0
- package/lib/esm/src/vertexai/utils/tensor.js +82 -0
- package/lib/esm/src/vertexai/utils/tensor.js.map +1 -0
- package/lib/esm/togetherai/index.js +1 -4
- package/lib/esm/togetherai/index.js.map +1 -1
- package/lib/esm/tsconfig.tsbuildinfo +1 -0
- package/lib/esm/vertexai/index.js +3 -4
- package/lib/esm/vertexai/index.js.map +1 -1
- package/lib/esm/vertexai/models/codey-chat.js +1 -1
- package/lib/esm/vertexai/models/codey-chat.js.map +1 -1
- package/lib/esm/vertexai/models/codey-text.js +2 -2
- package/lib/esm/vertexai/models/codey-text.js.map +1 -1
- package/lib/esm/vertexai/models/gemini.js +2 -2
- package/lib/esm/vertexai/models/gemini.js.map +1 -1
- package/lib/esm/vertexai/models/palm-model-base.js.map +1 -1
- package/lib/esm/vertexai/models/palm2-chat.js +1 -1
- package/lib/esm/vertexai/models/palm2-chat.js.map +1 -1
- package/lib/esm/vertexai/models/palm2-text.js +2 -2
- package/lib/esm/vertexai/models/palm2-text.js.map +1 -1
- package/lib/types/bedrock/index.d.ts +11 -5
- package/lib/types/bedrock/index.d.ts.map +1 -1
- package/lib/types/bedrock/s3.d.ts +2 -5
- package/lib/types/bedrock/s3.d.ts.map +1 -1
- package/lib/types/huggingface_ie.d.ts +4 -5
- package/lib/types/huggingface_ie.d.ts.map +1 -1
- package/lib/types/mistral/index.d.ts +7 -11
- package/lib/types/mistral/index.d.ts.map +1 -1
- package/lib/types/openai.d.ts +1 -3
- package/lib/types/openai.d.ts.map +1 -1
- package/lib/types/replicate.d.ts +1 -2
- package/lib/types/replicate.d.ts.map +1 -1
- package/lib/types/src/bedrock/index.d.ts +94 -0
- package/lib/types/src/bedrock/s3.d.ts +16 -0
- package/lib/types/src/huggingface_ie.d.ts +30 -0
- package/lib/types/src/index.d.ts +8 -0
- package/lib/types/src/mistral/index.d.ts +23 -0
- package/lib/types/src/mistral/types.d.ts +130 -0
- package/lib/types/src/openai.d.ts +30 -0
- package/lib/types/src/replicate.d.ts +47 -0
- package/lib/types/src/test/TestErrorCompletionStream.d.ts +8 -0
- package/lib/types/src/test/TestValidationErrorCompletionStream.d.ts +8 -0
- package/lib/types/src/test/index.d.ts +23 -0
- package/lib/types/src/test/utils.d.ts +4 -0
- package/lib/types/src/togetherai/index.d.ts +21 -0
- package/lib/types/src/togetherai/interfaces.d.ts +80 -0
- package/lib/types/src/vertexai/debug.d.ts +1 -0
- package/lib/types/src/vertexai/embeddings/embeddings-text.d.ts +9 -0
- package/lib/types/src/vertexai/index.d.ts +21 -0
- package/lib/types/src/vertexai/models/codey-chat.d.ts +50 -0
- package/lib/types/src/vertexai/models/codey-text.d.ts +38 -0
- package/lib/types/src/vertexai/models/gemini.d.ts +10 -0
- package/lib/types/src/vertexai/models/palm-model-base.d.ts +60 -0
- package/lib/types/src/vertexai/models/palm2-chat.d.ts +60 -0
- package/lib/types/src/vertexai/models/palm2-text.d.ts +38 -0
- package/lib/types/src/vertexai/models.d.ts +13 -0
- package/lib/types/src/vertexai/utils/prompts.d.ts +19 -0
- package/lib/types/src/vertexai/utils/tensor.d.ts +5 -0
- package/lib/types/togetherai/index.d.ts +1 -2
- package/lib/types/togetherai/index.d.ts.map +1 -1
- package/lib/types/vertexai/index.d.ts +1 -2
- package/lib/types/vertexai/index.d.ts.map +1 -1
- package/lib/types/vertexai/models/codey-chat.d.ts.map +1 -1
- package/lib/types/vertexai/models/codey-text.d.ts.map +1 -1
- package/lib/types/vertexai/models/gemini.d.ts.map +1 -1
- package/lib/types/vertexai/models/palm-model-base.d.ts +15 -1
- package/lib/types/vertexai/models/palm-model-base.d.ts.map +1 -1
- package/lib/types/vertexai/models/palm2-chat.d.ts.map +1 -1
- package/lib/types/vertexai/models/palm2-text.d.ts.map +1 -1
- package/package.json +15 -16
- package/src/bedrock/index.ts +33 -9
- package/src/bedrock/s3.ts +2 -3
- package/src/huggingface_ie.ts +1 -3
- package/src/mistral/index.ts +32 -35
- package/src/openai.ts +3 -9
- package/src/replicate.ts +0 -2
- package/src/togetherai/index.ts +1 -4
- package/src/vertexai/index.ts +4 -5
- package/src/vertexai/models/codey-chat.ts +1 -1
- package/src/vertexai/models/codey-text.ts +2 -2
- package/src/vertexai/models/gemini.ts +4 -4
- package/src/vertexai/models/palm-model-base.ts +17 -1
- package/src/vertexai/models/palm2-chat.ts +1 -1
- package/src/vertexai/models/palm2-text.ts +2 -2
|
@@ -1,10 +1,10 @@
|
|
|
1
|
-
import { Content, GenerateContentRequest,
|
|
1
|
+
import { Content, GenerateContentRequest, HarmBlockThreshold, HarmCategory, TextPart } from "@google-cloud/vertexai";
|
|
2
2
|
import { AIModel, Completion, ExecutionOptions, ExecutionTokenUsage, ModelType, PromptOptions, PromptRole, PromptSegment } from "@llumiverse/core";
|
|
3
3
|
import { asyncMap } from "@llumiverse/core/async";
|
|
4
4
|
import { VertexAIDriver } from "../index.js";
|
|
5
5
|
import { ModelDefinition } from "../models.js";
|
|
6
6
|
|
|
7
|
-
function getGenerativeModel(driver: VertexAIDriver, options: ExecutionOptions)
|
|
7
|
+
function getGenerativeModel(driver: VertexAIDriver, options: ExecutionOptions) {
|
|
8
8
|
return driver.vertexai.preview.getGenerativeModel({
|
|
9
9
|
model: options.model,
|
|
10
10
|
//TODO pass in the options
|
|
@@ -106,8 +106,8 @@ export class GeminiModelDefinition implements ModelDefinition<GenerateContentReq
|
|
|
106
106
|
const response = await r.response;
|
|
107
107
|
const usage = response.usageMetadata;
|
|
108
108
|
const token_usage: ExecutionTokenUsage = {
|
|
109
|
-
prompt: usage?.
|
|
110
|
-
result: usage?.
|
|
109
|
+
prompt: usage?.promptTokenCount,
|
|
110
|
+
result: usage?.candidatesTokenCount,
|
|
111
111
|
total: usage?.totalTokenCount,
|
|
112
112
|
}
|
|
113
113
|
|
|
@@ -2,9 +2,25 @@ import { AIModel, Completion, ExecutionOptions, PromptOptions, PromptSegment } f
|
|
|
2
2
|
import { transformSSEStream } from "@llumiverse/core/async";
|
|
3
3
|
import { VertexAIDriver } from "../index.js";
|
|
4
4
|
import { ModelDefinition } from "../models.js";
|
|
5
|
-
import { PromptParamatersBase } from "../utils/prompts.js";
|
|
6
5
|
import { generateStreamingPrompt } from "../utils/tensor.js";
|
|
7
6
|
|
|
7
|
+
|
|
8
|
+
export interface PromptParamatersBase {
|
|
9
|
+
temperature?: number,
|
|
10
|
+
maxOutputTokens?: number,
|
|
11
|
+
topK?: number,
|
|
12
|
+
topP?: number,
|
|
13
|
+
groundingConfig?: string,
|
|
14
|
+
stopSequences?: string[],
|
|
15
|
+
candidateCount?: number,
|
|
16
|
+
logprobs?: number,
|
|
17
|
+
presencePenalty?: number,
|
|
18
|
+
frequencyPenalty?: number,
|
|
19
|
+
logitBias?: Record<string, number>,
|
|
20
|
+
seed?: number,
|
|
21
|
+
echo?: boolean
|
|
22
|
+
}
|
|
23
|
+
|
|
8
24
|
export interface NonStreamingPromptBase<InstanceType = any> {
|
|
9
25
|
instances: InstanceType[];
|
|
10
26
|
parameters: PromptParamatersBase;
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
import { AIModel, ModelType, PromptOptions, PromptRole, PromptSegment } from "@llumiverse/core";
|
|
2
|
+
import { getJSONSafetyNotice } from "@llumiverse/core/formatters";
|
|
2
3
|
import { VertexAIDriver } from "../index.js";
|
|
3
|
-
import { getJSONSafetyNotice } from "../utils/prompts.js";
|
|
4
4
|
import { AbstractPalmModelDefinition, NonStreamingPromptBase, PalmResponseMetadata, StreamingPromptBase } from "./palm-model-base.js";
|
|
5
5
|
|
|
6
6
|
export interface Palm2ChatMessage {
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
import { AIModel, ModelType, PromptOptions, PromptSegment } from "@llumiverse/core";
|
|
2
|
+
import { formatTextPrompt } from "@llumiverse/core/formatters";
|
|
2
3
|
import { VertexAIDriver } from "../index.js";
|
|
3
|
-
import { getPromptAsText } from "../utils/prompts.js";
|
|
4
4
|
import { AbstractPalmModelDefinition, NonStreamingPromptBase, PalmResponseMetadata, StreamingPromptBase } from "./palm-model-base.js";
|
|
5
5
|
|
|
6
6
|
export type Palm2TextPrompt = NonStreamingPromptBase<{
|
|
@@ -50,7 +50,7 @@ export class Palm2TextDefinition extends AbstractPalmModelDefinition<Palm2TextPr
|
|
|
50
50
|
createNonStreamingPrompt(_driver: VertexAIDriver, segments: PromptSegment[], opts: PromptOptions): Palm2TextPrompt {
|
|
51
51
|
return {
|
|
52
52
|
instances: [{
|
|
53
|
-
prompt:
|
|
53
|
+
prompt: formatTextPrompt(segments, opts.resultSchema)
|
|
54
54
|
}],
|
|
55
55
|
parameters: {
|
|
56
56
|
// put defauilts here
|