modelfusion 0.101.0 → 0.103.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +41 -0
- package/model-function/generate-image/ImageGenerationModel.d.ts +12 -2
- package/model-function/generate-image/PromptTemplateImageGenerationModel.cjs +3 -3
- package/model-function/generate-image/PromptTemplateImageGenerationModel.d.ts +2 -2
- package/model-function/generate-image/PromptTemplateImageGenerationModel.js +3 -3
- package/model-function/generate-image/generateImage.cjs +9 -7
- package/model-function/generate-image/generateImage.d.ts +2 -0
- package/model-function/generate-image/generateImage.js +9 -7
- package/model-function/generate-text/PromptTemplateTextGenerationModel.d.ts +1 -1
- package/model-function/generate-text/TextGenerationModel.cjs +7 -0
- package/model-function/generate-text/TextGenerationModel.d.ts +3 -1
- package/model-function/generate-text/TextGenerationModel.js +6 -1
- package/model-function/generate-text/TextGenerationResult.cjs +2 -0
- package/model-function/generate-text/TextGenerationResult.d.ts +11 -0
- package/model-function/generate-text/TextGenerationResult.js +1 -0
- package/model-function/generate-text/generateText.cjs +14 -9
- package/model-function/generate-text/generateText.d.ts +3 -0
- package/model-function/generate-text/generateText.js +14 -9
- package/model-function/generate-text/index.cjs +1 -0
- package/model-function/generate-text/index.d.ts +1 -0
- package/model-function/generate-text/index.js +1 -0
- package/model-provider/anthropic/AnthropicTextGenerationModel.cjs +23 -8
- package/model-provider/anthropic/AnthropicTextGenerationModel.d.ts +6 -1
- package/model-provider/anthropic/AnthropicTextGenerationModel.js +24 -9
- package/model-provider/automatic1111/Automatic1111ImageGenerationModel.cjs +7 -5
- package/model-provider/automatic1111/Automatic1111ImageGenerationModel.d.ts +16 -2
- package/model-provider/automatic1111/Automatic1111ImageGenerationModel.js +7 -5
- package/model-provider/automatic1111/Automatic1111ImageGenerationPrompt.d.ts +0 -1
- package/model-provider/cohere/CohereTextGenerationModel.cjs +22 -6
- package/model-provider/cohere/CohereTextGenerationModel.d.ts +6 -1
- package/model-provider/cohere/CohereTextGenerationModel.js +22 -6
- package/model-provider/elevenlabs/ElevenLabsSpeechModel.cjs +2 -2
- package/model-provider/elevenlabs/ElevenLabsSpeechModel.js +2 -2
- package/model-provider/huggingface/HuggingFaceTextGenerationModel.cjs +9 -8
- package/model-provider/huggingface/HuggingFaceTextGenerationModel.d.ts +4 -5
- package/model-provider/huggingface/HuggingFaceTextGenerationModel.js +9 -8
- package/model-provider/llamacpp/LlamaCppTextGenerationModel.cjs +51 -51
- package/model-provider/llamacpp/LlamaCppTextGenerationModel.d.ts +20 -17
- package/model-provider/llamacpp/LlamaCppTextGenerationModel.js +51 -51
- package/model-provider/mistral/MistralChatModel.cjs +19 -2
- package/model-provider/mistral/MistralChatModel.d.ts +6 -1
- package/model-provider/mistral/MistralChatModel.js +19 -2
- package/model-provider/ollama/OllamaChatModel.cjs +8 -3
- package/model-provider/ollama/OllamaChatModel.d.ts +4 -1
- package/model-provider/ollama/OllamaChatModel.js +8 -3
- package/model-provider/ollama/OllamaCompletionModel.cjs +8 -3
- package/model-provider/ollama/OllamaCompletionModel.d.ts +4 -1
- package/model-provider/ollama/OllamaCompletionModel.js +8 -3
- package/model-provider/openai/OpenAICompletionModel.cjs +20 -4
- package/model-provider/openai/OpenAICompletionModel.d.ts +6 -1
- package/model-provider/openai/OpenAICompletionModel.js +20 -4
- package/model-provider/openai/OpenAIImageGenerationModel.cjs +25 -31
- package/model-provider/openai/OpenAIImageGenerationModel.d.ts +2 -3
- package/model-provider/openai/OpenAIImageGenerationModel.js +25 -31
- package/model-provider/openai/chat/AbstractOpenAIChatModel.cjs +19 -1
- package/model-provider/openai/chat/AbstractOpenAIChatModel.d.ts +6 -1
- package/model-provider/openai/chat/AbstractOpenAIChatModel.js +19 -1
- package/model-provider/openai/chat/OpenAIChatModel.cjs +2 -3
- package/model-provider/openai/chat/OpenAIChatModel.js +2 -3
- package/model-provider/openai-compatible/OpenAICompatibleChatModel.cjs +2 -3
- package/model-provider/openai-compatible/OpenAICompatibleChatModel.js +2 -3
- package/model-provider/stability/StabilityApiConfiguration.cjs +12 -5
- package/model-provider/stability/StabilityApiConfiguration.d.ts +7 -8
- package/model-provider/stability/StabilityApiConfiguration.js +12 -5
- package/model-provider/stability/StabilityError.cjs +7 -31
- package/model-provider/stability/StabilityError.d.ts +2 -11
- package/model-provider/stability/StabilityError.js +6 -28
- package/model-provider/stability/StabilityFacade.cjs +11 -3
- package/model-provider/stability/StabilityFacade.d.ts +10 -2
- package/model-provider/stability/StabilityFacade.js +9 -2
- package/model-provider/stability/StabilityImageGenerationModel.cjs +39 -50
- package/model-provider/stability/StabilityImageGenerationModel.d.ts +42 -27
- package/model-provider/stability/StabilityImageGenerationModel.js +39 -50
- package/model-provider/stability/index.cjs +1 -3
- package/model-provider/stability/index.d.ts +1 -1
- package/model-provider/stability/index.js +0 -1
- package/package.json +2 -2
@@ -34,13 +34,15 @@ export class Automatic1111ImageGenerationModel extends AbstractModel {
|
|
34
34
|
url: api.assembleUrl(`/txt2img`),
|
35
35
|
headers: api.headers,
|
36
36
|
body: {
|
37
|
-
height: this.settings.height,
|
38
|
-
width: this.settings.width,
|
39
37
|
prompt: input.prompt,
|
40
38
|
negative_prompt: input.negativePrompt,
|
39
|
+
seed: this.settings.seed,
|
40
|
+
batch_size: this.settings.numberOfGenerations,
|
41
|
+
height: this.settings.height,
|
42
|
+
width: this.settings.width,
|
43
|
+
cfg_scale: this.settings.cfgScale,
|
41
44
|
sampler_index: this.settings.sampler,
|
42
45
|
steps: this.settings.steps,
|
43
|
-
seed: input.seed,
|
44
46
|
override_settings: {
|
45
47
|
sd_model_checkpoint: this.settings.model,
|
46
48
|
},
|
@@ -59,11 +61,11 @@ export class Automatic1111ImageGenerationModel extends AbstractModel {
|
|
59
61
|
steps: this.settings.steps,
|
60
62
|
};
|
61
63
|
}
|
62
|
-
async
|
64
|
+
async doGenerateImages(prompt, options) {
|
63
65
|
const response = await this.callAPI(prompt, options);
|
64
66
|
return {
|
65
67
|
response,
|
66
|
-
|
68
|
+
base64Images: response.images,
|
67
69
|
};
|
68
70
|
}
|
69
71
|
withTextPrompt() {
|
@@ -7,6 +7,7 @@ const postToApi_js_1 = require("../../core/api/postToApi.cjs");
|
|
7
7
|
const ZodSchema_js_1 = require("../../core/schema/ZodSchema.cjs");
|
8
8
|
const AbstractModel_js_1 = require("../../model-function/AbstractModel.cjs");
|
9
9
|
const PromptTemplateTextStreamingModel_js_1 = require("../../model-function/generate-text/PromptTemplateTextStreamingModel.cjs");
|
10
|
+
const TextGenerationModel_js_1 = require("../../model-function/generate-text/TextGenerationModel.cjs");
|
10
11
|
const TextPromptTemplate_js_1 = require("../../model-function/generate-text/prompt-template/TextPromptTemplate.cjs");
|
11
12
|
const countTokens_js_1 = require("../../model-function/tokenize-text/countTokens.cjs");
|
12
13
|
const AsyncQueue_js_1 = require("../../util/AsyncQueue.cjs");
|
@@ -84,8 +85,8 @@ class CohereTextGenerationModel extends AbstractModel_js_1.AbstractModel {
|
|
84
85
|
const responseFormat = options.responseFormat;
|
85
86
|
const abortSignal = options.run?.abortSignal;
|
86
87
|
return (0, callWithRetryAndThrottle_js_1.callWithRetryAndThrottle)({
|
87
|
-
retry:
|
88
|
-
throttle:
|
88
|
+
retry: api.retry,
|
89
|
+
throttle: api.throttle,
|
89
90
|
call: async () => {
|
90
91
|
return (0, postToApi_js_1.postJsonToApi)({
|
91
92
|
url: api.assembleUrl(`/generate`),
|
@@ -116,9 +117,7 @@ class CohereTextGenerationModel extends AbstractModel_js_1.AbstractModel {
|
|
116
117
|
}
|
117
118
|
get settingsForEvent() {
|
118
119
|
const eventSettingProperties = [
|
119
|
-
|
120
|
-
"stopSequences",
|
121
|
-
"numberOfGenerations",
|
120
|
+
...TextGenerationModel_js_1.textGenerationModelProperties,
|
122
121
|
"temperature",
|
123
122
|
"k",
|
124
123
|
"p",
|
@@ -138,9 +137,26 @@ class CohereTextGenerationModel extends AbstractModel_js_1.AbstractModel {
|
|
138
137
|
});
|
139
138
|
return {
|
140
139
|
response,
|
141
|
-
|
140
|
+
textGenerationResults: response.generations.map((generation) => ({
|
141
|
+
text: generation.text,
|
142
|
+
finishReason: this.translateFinishReason(generation.finish_reason),
|
143
|
+
})),
|
142
144
|
};
|
143
145
|
}
|
146
|
+
translateFinishReason(finishReason) {
|
147
|
+
switch (finishReason) {
|
148
|
+
case "COMPLETE":
|
149
|
+
return "stop";
|
150
|
+
case "MAX_TOKENS":
|
151
|
+
return "length";
|
152
|
+
case "ERROR_TOXIC":
|
153
|
+
return "content-filter";
|
154
|
+
case "ERROR":
|
155
|
+
return "error";
|
156
|
+
default:
|
157
|
+
return "unknown";
|
158
|
+
}
|
159
|
+
}
|
144
160
|
doStreamText(prompt, options) {
|
145
161
|
return this.callAPI(prompt, {
|
146
162
|
...options,
|
@@ -7,6 +7,7 @@ import { Delta } from "../../model-function/Delta.js";
|
|
7
7
|
import { PromptTemplateTextStreamingModel } from "../../model-function/generate-text/PromptTemplateTextStreamingModel.js";
|
8
8
|
import { TextGenerationModelSettings, TextStreamingModel } from "../../model-function/generate-text/TextGenerationModel.js";
|
9
9
|
import { TextGenerationPromptTemplate } from "../../model-function/generate-text/TextGenerationPromptTemplate.js";
|
10
|
+
import { TextGenerationFinishReason } from "../../model-function/generate-text/TextGenerationResult.js";
|
10
11
|
import { CohereTokenizer } from "./CohereTokenizer.js";
|
11
12
|
export declare const COHERE_TEXT_GENERATION_MODELS: {
|
12
13
|
command: {
|
@@ -79,8 +80,12 @@ export declare class CohereTextGenerationModel extends AbstractModel<CohereTextG
|
|
79
80
|
};
|
80
81
|
} | undefined;
|
81
82
|
};
|
82
|
-
|
83
|
+
textGenerationResults: {
|
84
|
+
text: string;
|
85
|
+
finishReason: TextGenerationFinishReason;
|
86
|
+
}[];
|
83
87
|
}>;
|
88
|
+
private translateFinishReason;
|
84
89
|
doStreamText(prompt: string, options?: FunctionOptions): Promise<AsyncIterable<Delta<string>>>;
|
85
90
|
extractTextDelta(fullDelta: CohereTextGenerationDelta): string | undefined;
|
86
91
|
/**
|
@@ -4,6 +4,7 @@ import { createJsonResponseHandler, postJsonToApi, } from "../../core/api/postTo
|
|
4
4
|
import { ZodSchema } from "../../core/schema/ZodSchema.js";
|
5
5
|
import { AbstractModel } from "../../model-function/AbstractModel.js";
|
6
6
|
import { PromptTemplateTextStreamingModel } from "../../model-function/generate-text/PromptTemplateTextStreamingModel.js";
|
7
|
+
import { textGenerationModelProperties, } from "../../model-function/generate-text/TextGenerationModel.js";
|
7
8
|
import { chat, instruction, } from "../../model-function/generate-text/prompt-template/TextPromptTemplate.js";
|
8
9
|
import { countTokens } from "../../model-function/tokenize-text/countTokens.js";
|
9
10
|
import { AsyncQueue } from "../../util/AsyncQueue.js";
|
@@ -81,8 +82,8 @@ export class CohereTextGenerationModel extends AbstractModel {
|
|
81
82
|
const responseFormat = options.responseFormat;
|
82
83
|
const abortSignal = options.run?.abortSignal;
|
83
84
|
return callWithRetryAndThrottle({
|
84
|
-
retry:
|
85
|
-
throttle:
|
85
|
+
retry: api.retry,
|
86
|
+
throttle: api.throttle,
|
86
87
|
call: async () => {
|
87
88
|
return postJsonToApi({
|
88
89
|
url: api.assembleUrl(`/generate`),
|
@@ -113,9 +114,7 @@ export class CohereTextGenerationModel extends AbstractModel {
|
|
113
114
|
}
|
114
115
|
get settingsForEvent() {
|
115
116
|
const eventSettingProperties = [
|
116
|
-
|
117
|
-
"stopSequences",
|
118
|
-
"numberOfGenerations",
|
117
|
+
...textGenerationModelProperties,
|
119
118
|
"temperature",
|
120
119
|
"k",
|
121
120
|
"p",
|
@@ -135,9 +134,26 @@ export class CohereTextGenerationModel extends AbstractModel {
|
|
135
134
|
});
|
136
135
|
return {
|
137
136
|
response,
|
138
|
-
|
137
|
+
textGenerationResults: response.generations.map((generation) => ({
|
138
|
+
text: generation.text,
|
139
|
+
finishReason: this.translateFinishReason(generation.finish_reason),
|
140
|
+
})),
|
139
141
|
};
|
140
142
|
}
|
143
|
+
translateFinishReason(finishReason) {
|
144
|
+
switch (finishReason) {
|
145
|
+
case "COMPLETE":
|
146
|
+
return "stop";
|
147
|
+
case "MAX_TOKENS":
|
148
|
+
return "length";
|
149
|
+
case "ERROR_TOXIC":
|
150
|
+
return "content-filter";
|
151
|
+
case "ERROR":
|
152
|
+
return "error";
|
153
|
+
default:
|
154
|
+
return "unknown";
|
155
|
+
}
|
156
|
+
}
|
141
157
|
doStreamText(prompt, options) {
|
142
158
|
return this.callAPI(prompt, {
|
143
159
|
...options,
|
@@ -101,7 +101,7 @@ class ElevenLabsSpeechModel extends AbstractModel_js_1.AbstractModel {
|
|
101
101
|
// The JS WebSocket API does not support authorization headers, so we send the API key in the BOS message.
|
102
102
|
// See https://stackoverflow.com/questions/4361173/http-headers-in-websockets-client-api
|
103
103
|
xi_api_key: api.apiKey,
|
104
|
-
text: " ",
|
104
|
+
text: " ", // first message
|
105
105
|
voice_settings: toApiVoiceSettings(this.settings.voiceSettings),
|
106
106
|
generation_config: toGenerationConfig(this.settings.generationConfig),
|
107
107
|
}));
|
@@ -125,7 +125,7 @@ class ElevenLabsSpeechModel extends AbstractModel_js_1.AbstractModel {
|
|
125
125
|
// send remaining text:
|
126
126
|
if (textBuffer.length > 0) {
|
127
127
|
socket.send(JSON.stringify({
|
128
|
-
text: `${textBuffer} `,
|
128
|
+
text: `${textBuffer} `, // append space
|
129
129
|
try_trigger_generation: true,
|
130
130
|
}));
|
131
131
|
}
|
@@ -98,7 +98,7 @@ export class ElevenLabsSpeechModel extends AbstractModel {
|
|
98
98
|
// The JS WebSocket API does not support authorization headers, so we send the API key in the BOS message.
|
99
99
|
// See https://stackoverflow.com/questions/4361173/http-headers-in-websockets-client-api
|
100
100
|
xi_api_key: api.apiKey,
|
101
|
-
text: " ",
|
101
|
+
text: " ", // first message
|
102
102
|
voice_settings: toApiVoiceSettings(this.settings.voiceSettings),
|
103
103
|
generation_config: toGenerationConfig(this.settings.generationConfig),
|
104
104
|
}));
|
@@ -122,7 +122,7 @@ export class ElevenLabsSpeechModel extends AbstractModel {
|
|
122
122
|
// send remaining text:
|
123
123
|
if (textBuffer.length > 0) {
|
124
124
|
socket.send(JSON.stringify({
|
125
|
-
text: `${textBuffer} `,
|
125
|
+
text: `${textBuffer} `, // append space
|
126
126
|
try_trigger_generation: true,
|
127
127
|
}));
|
128
128
|
}
|
@@ -6,6 +6,7 @@ const callWithRetryAndThrottle_js_1 = require("../../core/api/callWithRetryAndTh
|
|
6
6
|
const postToApi_js_1 = require("../../core/api/postToApi.cjs");
|
7
7
|
const AbstractModel_js_1 = require("../../model-function/AbstractModel.cjs");
|
8
8
|
const PromptTemplateTextGenerationModel_js_1 = require("../../model-function/generate-text/PromptTemplateTextGenerationModel.cjs");
|
9
|
+
const TextGenerationModel_js_1 = require("../../model-function/generate-text/TextGenerationModel.cjs");
|
9
10
|
const HuggingFaceApiConfiguration_js_1 = require("./HuggingFaceApiConfiguration.cjs");
|
10
11
|
const HuggingFaceError_js_1 = require("./HuggingFaceError.cjs");
|
11
12
|
/**
|
@@ -61,8 +62,8 @@ class HuggingFaceTextGenerationModel extends AbstractModel_js_1.AbstractModel {
|
|
61
62
|
const api = this.settings.api ?? new HuggingFaceApiConfiguration_js_1.HuggingFaceApiConfiguration();
|
62
63
|
const abortSignal = options?.run?.abortSignal;
|
63
64
|
return (0, callWithRetryAndThrottle_js_1.callWithRetryAndThrottle)({
|
64
|
-
retry:
|
65
|
-
throttle:
|
65
|
+
retry: api.retry,
|
66
|
+
throttle: api.throttle,
|
66
67
|
call: async () => {
|
67
68
|
return (0, postToApi_js_1.postJsonToApi)({
|
68
69
|
url: api.assembleUrl(`/${this.settings.model}`),
|
@@ -91,16 +92,13 @@ class HuggingFaceTextGenerationModel extends AbstractModel_js_1.AbstractModel {
|
|
91
92
|
}
|
92
93
|
get settingsForEvent() {
|
93
94
|
const eventSettingProperties = [
|
94
|
-
|
95
|
-
"maxGenerationTokens",
|
96
|
-
"numberOfGenerations",
|
95
|
+
...TextGenerationModel_js_1.textGenerationModelProperties,
|
97
96
|
"topK",
|
98
97
|
"topP",
|
99
98
|
"temperature",
|
100
99
|
"repetitionPenalty",
|
101
100
|
"maxTime",
|
102
101
|
"doSample",
|
103
|
-
"options",
|
104
102
|
];
|
105
103
|
return Object.fromEntries(Object.entries(this.settings).filter(([key]) => eventSettingProperties.includes(key)));
|
106
104
|
}
|
@@ -108,12 +106,15 @@ class HuggingFaceTextGenerationModel extends AbstractModel_js_1.AbstractModel {
|
|
108
106
|
const response = await this.callAPI(prompt, options);
|
109
107
|
return {
|
110
108
|
response,
|
111
|
-
|
109
|
+
textGenerationResults: response.map((response) => ({
|
110
|
+
text: response.generated_text,
|
111
|
+
finishReason: "unknown",
|
112
|
+
})),
|
112
113
|
};
|
113
114
|
}
|
114
115
|
withPromptTemplate(promptTemplate) {
|
115
116
|
return new PromptTemplateTextGenerationModel_js_1.PromptTemplateTextGenerationModel({
|
116
|
-
model: this,
|
117
|
+
model: this, // stop tokens are not supported by this model
|
117
118
|
promptTemplate,
|
118
119
|
});
|
119
120
|
}
|
@@ -14,10 +14,6 @@ export interface HuggingFaceTextGenerationModelSettings extends TextGenerationMo
|
|
14
14
|
repetitionPenalty?: number;
|
15
15
|
maxTime?: number;
|
16
16
|
doSample?: boolean;
|
17
|
-
options?: {
|
18
|
-
useCache?: boolean;
|
19
|
-
waitForModel?: boolean;
|
20
|
-
};
|
21
17
|
}
|
22
18
|
/**
|
23
19
|
* Create a text generation model that calls a Hugging Face Inference API Text Generation Task.
|
@@ -50,7 +46,10 @@ export declare class HuggingFaceTextGenerationModel extends AbstractModel<Huggin
|
|
50
46
|
response: {
|
51
47
|
generated_text: string;
|
52
48
|
}[];
|
53
|
-
|
49
|
+
textGenerationResults: {
|
50
|
+
text: string;
|
51
|
+
finishReason: "unknown";
|
52
|
+
}[];
|
54
53
|
}>;
|
55
54
|
withPromptTemplate<INPUT_PROMPT>(promptTemplate: TextGenerationPromptTemplate<INPUT_PROMPT, string>): PromptTemplateTextGenerationModel<INPUT_PROMPT, string, HuggingFaceTextGenerationModelSettings, this>;
|
56
55
|
withSettings(additionalSettings: Partial<HuggingFaceTextGenerationModelSettings>): this;
|
@@ -3,6 +3,7 @@ import { callWithRetryAndThrottle } from "../../core/api/callWithRetryAndThrottl
|
|
3
3
|
import { createJsonResponseHandler, postJsonToApi, } from "../../core/api/postToApi.js";
|
4
4
|
import { AbstractModel } from "../../model-function/AbstractModel.js";
|
5
5
|
import { PromptTemplateTextGenerationModel } from "../../model-function/generate-text/PromptTemplateTextGenerationModel.js";
|
6
|
+
import { textGenerationModelProperties, } from "../../model-function/generate-text/TextGenerationModel.js";
|
6
7
|
import { HuggingFaceApiConfiguration } from "./HuggingFaceApiConfiguration.js";
|
7
8
|
import { failedHuggingFaceCallResponseHandler } from "./HuggingFaceError.js";
|
8
9
|
/**
|
@@ -58,8 +59,8 @@ export class HuggingFaceTextGenerationModel extends AbstractModel {
|
|
58
59
|
const api = this.settings.api ?? new HuggingFaceApiConfiguration();
|
59
60
|
const abortSignal = options?.run?.abortSignal;
|
60
61
|
return callWithRetryAndThrottle({
|
61
|
-
retry:
|
62
|
-
throttle:
|
62
|
+
retry: api.retry,
|
63
|
+
throttle: api.throttle,
|
63
64
|
call: async () => {
|
64
65
|
return postJsonToApi({
|
65
66
|
url: api.assembleUrl(`/${this.settings.model}`),
|
@@ -88,16 +89,13 @@ export class HuggingFaceTextGenerationModel extends AbstractModel {
|
|
88
89
|
}
|
89
90
|
get settingsForEvent() {
|
90
91
|
const eventSettingProperties = [
|
91
|
-
|
92
|
-
"maxGenerationTokens",
|
93
|
-
"numberOfGenerations",
|
92
|
+
...textGenerationModelProperties,
|
94
93
|
"topK",
|
95
94
|
"topP",
|
96
95
|
"temperature",
|
97
96
|
"repetitionPenalty",
|
98
97
|
"maxTime",
|
99
98
|
"doSample",
|
100
|
-
"options",
|
101
99
|
];
|
102
100
|
return Object.fromEntries(Object.entries(this.settings).filter(([key]) => eventSettingProperties.includes(key)));
|
103
101
|
}
|
@@ -105,12 +103,15 @@ export class HuggingFaceTextGenerationModel extends AbstractModel {
|
|
105
103
|
const response = await this.callAPI(prompt, options);
|
106
104
|
return {
|
107
105
|
response,
|
108
|
-
|
106
|
+
textGenerationResults: response.map((response) => ({
|
107
|
+
text: response.generated_text,
|
108
|
+
finishReason: "unknown",
|
109
|
+
})),
|
109
110
|
};
|
110
111
|
}
|
111
112
|
withPromptTemplate(promptTemplate) {
|
112
113
|
return new PromptTemplateTextGenerationModel({
|
113
|
-
model: this,
|
114
|
+
model: this, // stop tokens are not supported by this model
|
114
115
|
promptTemplate,
|
115
116
|
});
|
116
117
|
}
|
@@ -7,6 +7,7 @@ const postToApi_js_1 = require("../../core/api/postToApi.cjs");
|
|
7
7
|
const ZodSchema_js_1 = require("../../core/schema/ZodSchema.cjs");
|
8
8
|
const AbstractModel_js_1 = require("../../model-function/AbstractModel.cjs");
|
9
9
|
const PromptTemplateTextStreamingModel_js_1 = require("../../model-function/generate-text/PromptTemplateTextStreamingModel.cjs");
|
10
|
+
const TextGenerationModel_js_1 = require("../../model-function/generate-text/TextGenerationModel.cjs");
|
10
11
|
const AsyncQueue_js_1 = require("../../util/AsyncQueue.cjs");
|
11
12
|
const parseJSON_js_1 = require("../../core/schema/parseJSON.cjs");
|
12
13
|
const parseEventSourceStream_js_1 = require("../../util/streaming/parseEventSourceStream.cjs");
|
@@ -37,25 +38,52 @@ class LlamaCppTextGenerationModel extends AbstractModel_js_1.AbstractModel {
|
|
37
38
|
return this.settings.contextWindowSize;
|
38
39
|
}
|
39
40
|
async callAPI(prompt, options) {
|
41
|
+
const api = this.settings.api ?? new LlamaCppApiConfiguration_js_1.LlamaCppApiConfiguration();
|
42
|
+
const responseFormat = options.responseFormat;
|
43
|
+
const abortSignal = options.run?.abortSignal;
|
40
44
|
return (0, callWithRetryAndThrottle_js_1.callWithRetryAndThrottle)({
|
41
|
-
retry:
|
42
|
-
throttle:
|
43
|
-
call: async () =>
|
44
|
-
|
45
|
-
|
46
|
-
|
47
|
-
|
48
|
-
|
49
|
-
|
50
|
-
|
51
|
-
|
45
|
+
retry: api.retry,
|
46
|
+
throttle: api.throttle,
|
47
|
+
call: async () => (0, postToApi_js_1.postJsonToApi)({
|
48
|
+
url: api.assembleUrl(`/completion`),
|
49
|
+
headers: api.headers,
|
50
|
+
body: {
|
51
|
+
stream: responseFormat.stream,
|
52
|
+
prompt: prompt.text,
|
53
|
+
image_data: prompt.images != null
|
54
|
+
? Object.entries(prompt.images).map(([id, data]) => ({
|
55
|
+
id: +id,
|
56
|
+
data,
|
57
|
+
}))
|
58
|
+
: undefined,
|
59
|
+
cache_prompt: this.settings.cachePrompt,
|
60
|
+
temperature: this.settings.temperature,
|
61
|
+
top_k: this.settings.topK,
|
62
|
+
top_p: this.settings.topP,
|
63
|
+
n_predict: this.settings.maxGenerationTokens,
|
64
|
+
n_keep: this.settings.nKeep,
|
65
|
+
stop: this.settings.stopSequences,
|
66
|
+
tfs_z: this.settings.tfsZ,
|
67
|
+
typical_p: this.settings.typicalP,
|
68
|
+
repeat_penalty: this.settings.repeatPenalty,
|
69
|
+
repeat_last_n: this.settings.repeatLastN,
|
70
|
+
penalize_nl: this.settings.penalizeNl,
|
71
|
+
mirostat: this.settings.mirostat,
|
72
|
+
mirostat_tau: this.settings.mirostatTau,
|
73
|
+
mirostat_eta: this.settings.mirostatEta,
|
74
|
+
seed: this.settings.seed,
|
75
|
+
ignore_eos: this.settings.ignoreEos,
|
76
|
+
logit_bias: this.settings.logitBias,
|
77
|
+
},
|
78
|
+
failedResponseHandler: LlamaCppError_js_1.failedLlamaCppCallResponseHandler,
|
79
|
+
successfulResponseHandler: responseFormat.handler,
|
80
|
+
abortSignal,
|
52
81
|
}),
|
53
82
|
});
|
54
83
|
}
|
55
84
|
get settingsForEvent() {
|
56
85
|
const eventSettingProperties = [
|
57
|
-
|
58
|
-
"stopSequences",
|
86
|
+
...TextGenerationModel_js_1.textGenerationModelProperties,
|
59
87
|
"contextWindowSize",
|
60
88
|
"cachePrompt",
|
61
89
|
"temperature",
|
@@ -87,7 +115,16 @@ class LlamaCppTextGenerationModel extends AbstractModel_js_1.AbstractModel {
|
|
87
115
|
});
|
88
116
|
return {
|
89
117
|
response,
|
90
|
-
|
118
|
+
textGenerationResults: [
|
119
|
+
{
|
120
|
+
text: response.content,
|
121
|
+
finishReason: response.stopped_eos || response.stopped_word
|
122
|
+
? "stop"
|
123
|
+
: response.stopped_limit
|
124
|
+
? "length"
|
125
|
+
: "unknown",
|
126
|
+
},
|
127
|
+
],
|
91
128
|
usage: {
|
92
129
|
promptTokens: response.tokens_evaluated,
|
93
130
|
completionTokens: response.tokens_predicted,
|
@@ -198,43 +235,6 @@ const llamaCppTextStreamingResponseSchema = new ZodSchema_js_1.ZodSchema(zod_1.z
|
|
198
235
|
}),
|
199
236
|
llamaCppTextGenerationResponseSchema,
|
200
237
|
]));
|
201
|
-
async function callLlamaCppTextGenerationAPI({ api = new LlamaCppApiConfiguration_js_1.LlamaCppApiConfiguration(), abortSignal, responseFormat, prompt, cachePrompt, temperature, topK, topP, nPredict, nKeep, stop, tfsZ, typicalP, repeatPenalty, repeatLastN, penalizeNl, mirostat, mirostatTau, mirostatEta, seed, ignoreEos, logitBias, }) {
|
202
|
-
return (0, postToApi_js_1.postJsonToApi)({
|
203
|
-
url: api.assembleUrl(`/completion`),
|
204
|
-
headers: api.headers,
|
205
|
-
body: {
|
206
|
-
stream: responseFormat.stream,
|
207
|
-
prompt: prompt.text,
|
208
|
-
cache_prompt: cachePrompt,
|
209
|
-
temperature,
|
210
|
-
top_k: topK,
|
211
|
-
top_p: topP,
|
212
|
-
n_predict: nPredict,
|
213
|
-
n_keep: nKeep,
|
214
|
-
stop,
|
215
|
-
tfs_z: tfsZ,
|
216
|
-
typical_p: typicalP,
|
217
|
-
repeat_penalty: repeatPenalty,
|
218
|
-
repeat_last_n: repeatLastN,
|
219
|
-
penalize_nl: penalizeNl,
|
220
|
-
mirostat,
|
221
|
-
mirostat_tau: mirostatTau,
|
222
|
-
mirostat_eta: mirostatEta,
|
223
|
-
seed,
|
224
|
-
ignore_eos: ignoreEos,
|
225
|
-
logit_bias: logitBias,
|
226
|
-
image_data: prompt.images != null
|
227
|
-
? Object.entries(prompt.images).map(([id, data]) => ({
|
228
|
-
id: +id,
|
229
|
-
data,
|
230
|
-
}))
|
231
|
-
: undefined,
|
232
|
-
},
|
233
|
-
failedResponseHandler: LlamaCppError_js_1.failedLlamaCppCallResponseHandler,
|
234
|
-
successfulResponseHandler: responseFormat.handler,
|
235
|
-
abortSignal,
|
236
|
-
});
|
237
|
-
}
|
238
238
|
async function createLlamaCppFullDeltaIterableQueue(stream) {
|
239
239
|
const queue = new AsyncQueue_js_1.AsyncQueue();
|
240
240
|
let content = "";
|
@@ -58,16 +58,16 @@ export declare class LlamaCppTextGenerationModel<CONTEXT_WINDOW_SIZE extends num
|
|
58
58
|
countPromptTokens(prompt: LlamaCppTextGenerationPrompt): Promise<number>;
|
59
59
|
doGenerateTexts(prompt: LlamaCppTextGenerationPrompt, options?: FunctionOptions): Promise<{
|
60
60
|
response: {
|
61
|
+
stop: true;
|
61
62
|
model: string;
|
62
63
|
prompt: string;
|
63
64
|
content: string;
|
64
|
-
stop: true;
|
65
65
|
generation_settings: {
|
66
|
+
stop: string[];
|
66
67
|
model: string;
|
67
68
|
stream: boolean;
|
68
|
-
mirostat: number;
|
69
69
|
seed: number;
|
70
|
-
|
70
|
+
mirostat: number;
|
71
71
|
frequency_penalty: number;
|
72
72
|
ignore_eos: boolean;
|
73
73
|
logit_bias: number[];
|
@@ -106,7 +106,10 @@ export declare class LlamaCppTextGenerationModel<CONTEXT_WINDOW_SIZE extends num
|
|
106
106
|
tokens_predicted: number;
|
107
107
|
truncated: boolean;
|
108
108
|
};
|
109
|
-
|
109
|
+
textGenerationResults: {
|
110
|
+
text: string;
|
111
|
+
finishReason: "length" | "stop" | "unknown";
|
112
|
+
}[];
|
110
113
|
usage: {
|
111
114
|
promptTokens: number;
|
112
115
|
completionTokens: number;
|
@@ -153,11 +156,11 @@ declare const llamaCppTextGenerationResponseSchema: z.ZodObject<{
|
|
153
156
|
top_p: z.ZodNumber;
|
154
157
|
typical_p: z.ZodNumber;
|
155
158
|
}, "strip", z.ZodTypeAny, {
|
159
|
+
stop: string[];
|
156
160
|
model: string;
|
157
161
|
stream: boolean;
|
158
|
-
mirostat: number;
|
159
162
|
seed: number;
|
160
|
-
|
163
|
+
mirostat: number;
|
161
164
|
frequency_penalty: number;
|
162
165
|
ignore_eos: boolean;
|
163
166
|
logit_bias: number[];
|
@@ -177,11 +180,11 @@ declare const llamaCppTextGenerationResponseSchema: z.ZodObject<{
|
|
177
180
|
top_p: number;
|
178
181
|
typical_p: number;
|
179
182
|
}, {
|
183
|
+
stop: string[];
|
180
184
|
model: string;
|
181
185
|
stream: boolean;
|
182
|
-
mirostat: number;
|
183
186
|
seed: number;
|
184
|
-
|
187
|
+
mirostat: number;
|
185
188
|
frequency_penalty: number;
|
186
189
|
ignore_eos: boolean;
|
187
190
|
logit_bias: number[];
|
@@ -240,16 +243,16 @@ declare const llamaCppTextGenerationResponseSchema: z.ZodObject<{
|
|
240
243
|
tokens_predicted: z.ZodNumber;
|
241
244
|
truncated: z.ZodBoolean;
|
242
245
|
}, "strip", z.ZodTypeAny, {
|
246
|
+
stop: true;
|
243
247
|
model: string;
|
244
248
|
prompt: string;
|
245
249
|
content: string;
|
246
|
-
stop: true;
|
247
250
|
generation_settings: {
|
251
|
+
stop: string[];
|
248
252
|
model: string;
|
249
253
|
stream: boolean;
|
250
|
-
mirostat: number;
|
251
254
|
seed: number;
|
252
|
-
|
255
|
+
mirostat: number;
|
253
256
|
frequency_penalty: number;
|
254
257
|
ignore_eos: boolean;
|
255
258
|
logit_bias: number[];
|
@@ -288,16 +291,16 @@ declare const llamaCppTextGenerationResponseSchema: z.ZodObject<{
|
|
288
291
|
tokens_predicted: number;
|
289
292
|
truncated: boolean;
|
290
293
|
}, {
|
294
|
+
stop: true;
|
291
295
|
model: string;
|
292
296
|
prompt: string;
|
293
297
|
content: string;
|
294
|
-
stop: true;
|
295
298
|
generation_settings: {
|
299
|
+
stop: string[];
|
296
300
|
model: string;
|
297
301
|
stream: boolean;
|
298
|
-
mirostat: number;
|
299
302
|
seed: number;
|
300
|
-
|
303
|
+
mirostat: number;
|
301
304
|
frequency_penalty: number;
|
302
305
|
ignore_eos: boolean;
|
303
306
|
logit_bias: number[];
|
@@ -353,16 +356,16 @@ export declare const LlamaCppTextGenerationResponseFormat: {
|
|
353
356
|
json: {
|
354
357
|
stream: false;
|
355
358
|
handler: ResponseHandler<{
|
359
|
+
stop: true;
|
356
360
|
model: string;
|
357
361
|
prompt: string;
|
358
362
|
content: string;
|
359
|
-
stop: true;
|
360
363
|
generation_settings: {
|
364
|
+
stop: string[];
|
361
365
|
model: string;
|
362
366
|
stream: boolean;
|
363
|
-
mirostat: number;
|
364
367
|
seed: number;
|
365
|
-
|
368
|
+
mirostat: number;
|
366
369
|
frequency_penalty: number;
|
367
370
|
ignore_eos: boolean;
|
368
371
|
logit_bias: number[];
|