modelfusion 0.101.0 → 0.103.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (77) hide show
  1. package/CHANGELOG.md +41 -0
  2. package/model-function/generate-image/ImageGenerationModel.d.ts +12 -2
  3. package/model-function/generate-image/PromptTemplateImageGenerationModel.cjs +3 -3
  4. package/model-function/generate-image/PromptTemplateImageGenerationModel.d.ts +2 -2
  5. package/model-function/generate-image/PromptTemplateImageGenerationModel.js +3 -3
  6. package/model-function/generate-image/generateImage.cjs +9 -7
  7. package/model-function/generate-image/generateImage.d.ts +2 -0
  8. package/model-function/generate-image/generateImage.js +9 -7
  9. package/model-function/generate-text/PromptTemplateTextGenerationModel.d.ts +1 -1
  10. package/model-function/generate-text/TextGenerationModel.cjs +7 -0
  11. package/model-function/generate-text/TextGenerationModel.d.ts +3 -1
  12. package/model-function/generate-text/TextGenerationModel.js +6 -1
  13. package/model-function/generate-text/TextGenerationResult.cjs +2 -0
  14. package/model-function/generate-text/TextGenerationResult.d.ts +11 -0
  15. package/model-function/generate-text/TextGenerationResult.js +1 -0
  16. package/model-function/generate-text/generateText.cjs +14 -9
  17. package/model-function/generate-text/generateText.d.ts +3 -0
  18. package/model-function/generate-text/generateText.js +14 -9
  19. package/model-function/generate-text/index.cjs +1 -0
  20. package/model-function/generate-text/index.d.ts +1 -0
  21. package/model-function/generate-text/index.js +1 -0
  22. package/model-provider/anthropic/AnthropicTextGenerationModel.cjs +23 -8
  23. package/model-provider/anthropic/AnthropicTextGenerationModel.d.ts +6 -1
  24. package/model-provider/anthropic/AnthropicTextGenerationModel.js +24 -9
  25. package/model-provider/automatic1111/Automatic1111ImageGenerationModel.cjs +7 -5
  26. package/model-provider/automatic1111/Automatic1111ImageGenerationModel.d.ts +16 -2
  27. package/model-provider/automatic1111/Automatic1111ImageGenerationModel.js +7 -5
  28. package/model-provider/automatic1111/Automatic1111ImageGenerationPrompt.d.ts +0 -1
  29. package/model-provider/cohere/CohereTextGenerationModel.cjs +22 -6
  30. package/model-provider/cohere/CohereTextGenerationModel.d.ts +6 -1
  31. package/model-provider/cohere/CohereTextGenerationModel.js +22 -6
  32. package/model-provider/elevenlabs/ElevenLabsSpeechModel.cjs +2 -2
  33. package/model-provider/elevenlabs/ElevenLabsSpeechModel.js +2 -2
  34. package/model-provider/huggingface/HuggingFaceTextGenerationModel.cjs +9 -8
  35. package/model-provider/huggingface/HuggingFaceTextGenerationModel.d.ts +4 -5
  36. package/model-provider/huggingface/HuggingFaceTextGenerationModel.js +9 -8
  37. package/model-provider/llamacpp/LlamaCppTextGenerationModel.cjs +51 -51
  38. package/model-provider/llamacpp/LlamaCppTextGenerationModel.d.ts +20 -17
  39. package/model-provider/llamacpp/LlamaCppTextGenerationModel.js +51 -51
  40. package/model-provider/mistral/MistralChatModel.cjs +19 -2
  41. package/model-provider/mistral/MistralChatModel.d.ts +6 -1
  42. package/model-provider/mistral/MistralChatModel.js +19 -2
  43. package/model-provider/ollama/OllamaChatModel.cjs +8 -3
  44. package/model-provider/ollama/OllamaChatModel.d.ts +4 -1
  45. package/model-provider/ollama/OllamaChatModel.js +8 -3
  46. package/model-provider/ollama/OllamaCompletionModel.cjs +8 -3
  47. package/model-provider/ollama/OllamaCompletionModel.d.ts +4 -1
  48. package/model-provider/ollama/OllamaCompletionModel.js +8 -3
  49. package/model-provider/openai/OpenAICompletionModel.cjs +20 -4
  50. package/model-provider/openai/OpenAICompletionModel.d.ts +6 -1
  51. package/model-provider/openai/OpenAICompletionModel.js +20 -4
  52. package/model-provider/openai/OpenAIImageGenerationModel.cjs +25 -31
  53. package/model-provider/openai/OpenAIImageGenerationModel.d.ts +2 -3
  54. package/model-provider/openai/OpenAIImageGenerationModel.js +25 -31
  55. package/model-provider/openai/chat/AbstractOpenAIChatModel.cjs +19 -1
  56. package/model-provider/openai/chat/AbstractOpenAIChatModel.d.ts +6 -1
  57. package/model-provider/openai/chat/AbstractOpenAIChatModel.js +19 -1
  58. package/model-provider/openai/chat/OpenAIChatModel.cjs +2 -3
  59. package/model-provider/openai/chat/OpenAIChatModel.js +2 -3
  60. package/model-provider/openai-compatible/OpenAICompatibleChatModel.cjs +2 -3
  61. package/model-provider/openai-compatible/OpenAICompatibleChatModel.js +2 -3
  62. package/model-provider/stability/StabilityApiConfiguration.cjs +12 -5
  63. package/model-provider/stability/StabilityApiConfiguration.d.ts +7 -8
  64. package/model-provider/stability/StabilityApiConfiguration.js +12 -5
  65. package/model-provider/stability/StabilityError.cjs +7 -31
  66. package/model-provider/stability/StabilityError.d.ts +2 -11
  67. package/model-provider/stability/StabilityError.js +6 -28
  68. package/model-provider/stability/StabilityFacade.cjs +11 -3
  69. package/model-provider/stability/StabilityFacade.d.ts +10 -2
  70. package/model-provider/stability/StabilityFacade.js +9 -2
  71. package/model-provider/stability/StabilityImageGenerationModel.cjs +39 -50
  72. package/model-provider/stability/StabilityImageGenerationModel.d.ts +42 -27
  73. package/model-provider/stability/StabilityImageGenerationModel.js +39 -50
  74. package/model-provider/stability/index.cjs +1 -3
  75. package/model-provider/stability/index.d.ts +1 -1
  76. package/model-provider/stability/index.js +0 -1
  77. package/package.json +2 -2
@@ -34,13 +34,15 @@ export class Automatic1111ImageGenerationModel extends AbstractModel {
34
34
  url: api.assembleUrl(`/txt2img`),
35
35
  headers: api.headers,
36
36
  body: {
37
- height: this.settings.height,
38
- width: this.settings.width,
39
37
  prompt: input.prompt,
40
38
  negative_prompt: input.negativePrompt,
39
+ seed: this.settings.seed,
40
+ batch_size: this.settings.numberOfGenerations,
41
+ height: this.settings.height,
42
+ width: this.settings.width,
43
+ cfg_scale: this.settings.cfgScale,
41
44
  sampler_index: this.settings.sampler,
42
45
  steps: this.settings.steps,
43
- seed: input.seed,
44
46
  override_settings: {
45
47
  sd_model_checkpoint: this.settings.model,
46
48
  },
@@ -59,11 +61,11 @@ export class Automatic1111ImageGenerationModel extends AbstractModel {
59
61
  steps: this.settings.steps,
60
62
  };
61
63
  }
62
- async doGenerateImage(prompt, options) {
64
+ async doGenerateImages(prompt, options) {
63
65
  const response = await this.callAPI(prompt, options);
64
66
  return {
65
67
  response,
66
- base64Image: response.images[0],
68
+ base64Images: response.images,
67
69
  };
68
70
  }
69
71
  withTextPrompt() {
@@ -2,7 +2,6 @@ import { PromptTemplate } from "../../model-function/PromptTemplate.js";
2
2
  export type Automatic1111ImageGenerationPrompt = {
3
3
  prompt: string;
4
4
  negativePrompt?: string;
5
- seed?: number;
6
5
  };
7
6
  /**
8
7
  * Formats a basic text prompt as an Automatic1111 prompt.
@@ -7,6 +7,7 @@ const postToApi_js_1 = require("../../core/api/postToApi.cjs");
7
7
  const ZodSchema_js_1 = require("../../core/schema/ZodSchema.cjs");
8
8
  const AbstractModel_js_1 = require("../../model-function/AbstractModel.cjs");
9
9
  const PromptTemplateTextStreamingModel_js_1 = require("../../model-function/generate-text/PromptTemplateTextStreamingModel.cjs");
10
+ const TextGenerationModel_js_1 = require("../../model-function/generate-text/TextGenerationModel.cjs");
10
11
  const TextPromptTemplate_js_1 = require("../../model-function/generate-text/prompt-template/TextPromptTemplate.cjs");
11
12
  const countTokens_js_1 = require("../../model-function/tokenize-text/countTokens.cjs");
12
13
  const AsyncQueue_js_1 = require("../../util/AsyncQueue.cjs");
@@ -84,8 +85,8 @@ class CohereTextGenerationModel extends AbstractModel_js_1.AbstractModel {
84
85
  const responseFormat = options.responseFormat;
85
86
  const abortSignal = options.run?.abortSignal;
86
87
  return (0, callWithRetryAndThrottle_js_1.callWithRetryAndThrottle)({
87
- retry: this.settings.api?.retry,
88
- throttle: this.settings.api?.throttle,
88
+ retry: api.retry,
89
+ throttle: api.throttle,
89
90
  call: async () => {
90
91
  return (0, postToApi_js_1.postJsonToApi)({
91
92
  url: api.assembleUrl(`/generate`),
@@ -116,9 +117,7 @@ class CohereTextGenerationModel extends AbstractModel_js_1.AbstractModel {
116
117
  }
117
118
  get settingsForEvent() {
118
119
  const eventSettingProperties = [
119
- "maxGenerationTokens",
120
- "stopSequences",
121
- "numberOfGenerations",
120
+ ...TextGenerationModel_js_1.textGenerationModelProperties,
122
121
  "temperature",
123
122
  "k",
124
123
  "p",
@@ -138,9 +137,26 @@ class CohereTextGenerationModel extends AbstractModel_js_1.AbstractModel {
138
137
  });
139
138
  return {
140
139
  response,
141
- texts: response.generations.map((generation) => generation.text),
140
+ textGenerationResults: response.generations.map((generation) => ({
141
+ text: generation.text,
142
+ finishReason: this.translateFinishReason(generation.finish_reason),
143
+ })),
142
144
  };
143
145
  }
146
+ translateFinishReason(finishReason) {
147
+ switch (finishReason) {
148
+ case "COMPLETE":
149
+ return "stop";
150
+ case "MAX_TOKENS":
151
+ return "length";
152
+ case "ERROR_TOXIC":
153
+ return "content-filter";
154
+ case "ERROR":
155
+ return "error";
156
+ default:
157
+ return "unknown";
158
+ }
159
+ }
144
160
  doStreamText(prompt, options) {
145
161
  return this.callAPI(prompt, {
146
162
  ...options,
@@ -7,6 +7,7 @@ import { Delta } from "../../model-function/Delta.js";
7
7
  import { PromptTemplateTextStreamingModel } from "../../model-function/generate-text/PromptTemplateTextStreamingModel.js";
8
8
  import { TextGenerationModelSettings, TextStreamingModel } from "../../model-function/generate-text/TextGenerationModel.js";
9
9
  import { TextGenerationPromptTemplate } from "../../model-function/generate-text/TextGenerationPromptTemplate.js";
10
+ import { TextGenerationFinishReason } from "../../model-function/generate-text/TextGenerationResult.js";
10
11
  import { CohereTokenizer } from "./CohereTokenizer.js";
11
12
  export declare const COHERE_TEXT_GENERATION_MODELS: {
12
13
  command: {
@@ -79,8 +80,12 @@ export declare class CohereTextGenerationModel extends AbstractModel<CohereTextG
79
80
  };
80
81
  } | undefined;
81
82
  };
82
- texts: string[];
83
+ textGenerationResults: {
84
+ text: string;
85
+ finishReason: TextGenerationFinishReason;
86
+ }[];
83
87
  }>;
88
+ private translateFinishReason;
84
89
  doStreamText(prompt: string, options?: FunctionOptions): Promise<AsyncIterable<Delta<string>>>;
85
90
  extractTextDelta(fullDelta: CohereTextGenerationDelta): string | undefined;
86
91
  /**
@@ -4,6 +4,7 @@ import { createJsonResponseHandler, postJsonToApi, } from "../../core/api/postTo
4
4
  import { ZodSchema } from "../../core/schema/ZodSchema.js";
5
5
  import { AbstractModel } from "../../model-function/AbstractModel.js";
6
6
  import { PromptTemplateTextStreamingModel } from "../../model-function/generate-text/PromptTemplateTextStreamingModel.js";
7
+ import { textGenerationModelProperties, } from "../../model-function/generate-text/TextGenerationModel.js";
7
8
  import { chat, instruction, } from "../../model-function/generate-text/prompt-template/TextPromptTemplate.js";
8
9
  import { countTokens } from "../../model-function/tokenize-text/countTokens.js";
9
10
  import { AsyncQueue } from "../../util/AsyncQueue.js";
@@ -81,8 +82,8 @@ export class CohereTextGenerationModel extends AbstractModel {
81
82
  const responseFormat = options.responseFormat;
82
83
  const abortSignal = options.run?.abortSignal;
83
84
  return callWithRetryAndThrottle({
84
- retry: this.settings.api?.retry,
85
- throttle: this.settings.api?.throttle,
85
+ retry: api.retry,
86
+ throttle: api.throttle,
86
87
  call: async () => {
87
88
  return postJsonToApi({
88
89
  url: api.assembleUrl(`/generate`),
@@ -113,9 +114,7 @@ export class CohereTextGenerationModel extends AbstractModel {
113
114
  }
114
115
  get settingsForEvent() {
115
116
  const eventSettingProperties = [
116
- "maxGenerationTokens",
117
- "stopSequences",
118
- "numberOfGenerations",
117
+ ...textGenerationModelProperties,
119
118
  "temperature",
120
119
  "k",
121
120
  "p",
@@ -135,9 +134,26 @@ export class CohereTextGenerationModel extends AbstractModel {
135
134
  });
136
135
  return {
137
136
  response,
138
- texts: response.generations.map((generation) => generation.text),
137
+ textGenerationResults: response.generations.map((generation) => ({
138
+ text: generation.text,
139
+ finishReason: this.translateFinishReason(generation.finish_reason),
140
+ })),
139
141
  };
140
142
  }
143
+ translateFinishReason(finishReason) {
144
+ switch (finishReason) {
145
+ case "COMPLETE":
146
+ return "stop";
147
+ case "MAX_TOKENS":
148
+ return "length";
149
+ case "ERROR_TOXIC":
150
+ return "content-filter";
151
+ case "ERROR":
152
+ return "error";
153
+ default:
154
+ return "unknown";
155
+ }
156
+ }
141
157
  doStreamText(prompt, options) {
142
158
  return this.callAPI(prompt, {
143
159
  ...options,
@@ -101,7 +101,7 @@ class ElevenLabsSpeechModel extends AbstractModel_js_1.AbstractModel {
101
101
  // The JS WebSocket API does not support authorization headers, so we send the API key in the BOS message.
102
102
  // See https://stackoverflow.com/questions/4361173/http-headers-in-websockets-client-api
103
103
  xi_api_key: api.apiKey,
104
- text: " ",
104
+ text: " ", // first message
105
105
  voice_settings: toApiVoiceSettings(this.settings.voiceSettings),
106
106
  generation_config: toGenerationConfig(this.settings.generationConfig),
107
107
  }));
@@ -125,7 +125,7 @@ class ElevenLabsSpeechModel extends AbstractModel_js_1.AbstractModel {
125
125
  // send remaining text:
126
126
  if (textBuffer.length > 0) {
127
127
  socket.send(JSON.stringify({
128
- text: `${textBuffer} `,
128
+ text: `${textBuffer} `, // append space
129
129
  try_trigger_generation: true,
130
130
  }));
131
131
  }
@@ -98,7 +98,7 @@ export class ElevenLabsSpeechModel extends AbstractModel {
98
98
  // The JS WebSocket API does not support authorization headers, so we send the API key in the BOS message.
99
99
  // See https://stackoverflow.com/questions/4361173/http-headers-in-websockets-client-api
100
100
  xi_api_key: api.apiKey,
101
- text: " ",
101
+ text: " ", // first message
102
102
  voice_settings: toApiVoiceSettings(this.settings.voiceSettings),
103
103
  generation_config: toGenerationConfig(this.settings.generationConfig),
104
104
  }));
@@ -122,7 +122,7 @@ export class ElevenLabsSpeechModel extends AbstractModel {
122
122
  // send remaining text:
123
123
  if (textBuffer.length > 0) {
124
124
  socket.send(JSON.stringify({
125
- text: `${textBuffer} `,
125
+ text: `${textBuffer} `, // append space
126
126
  try_trigger_generation: true,
127
127
  }));
128
128
  }
@@ -6,6 +6,7 @@ const callWithRetryAndThrottle_js_1 = require("../../core/api/callWithRetryAndTh
6
6
  const postToApi_js_1 = require("../../core/api/postToApi.cjs");
7
7
  const AbstractModel_js_1 = require("../../model-function/AbstractModel.cjs");
8
8
  const PromptTemplateTextGenerationModel_js_1 = require("../../model-function/generate-text/PromptTemplateTextGenerationModel.cjs");
9
+ const TextGenerationModel_js_1 = require("../../model-function/generate-text/TextGenerationModel.cjs");
9
10
  const HuggingFaceApiConfiguration_js_1 = require("./HuggingFaceApiConfiguration.cjs");
10
11
  const HuggingFaceError_js_1 = require("./HuggingFaceError.cjs");
11
12
  /**
@@ -61,8 +62,8 @@ class HuggingFaceTextGenerationModel extends AbstractModel_js_1.AbstractModel {
61
62
  const api = this.settings.api ?? new HuggingFaceApiConfiguration_js_1.HuggingFaceApiConfiguration();
62
63
  const abortSignal = options?.run?.abortSignal;
63
64
  return (0, callWithRetryAndThrottle_js_1.callWithRetryAndThrottle)({
64
- retry: this.settings.api?.retry,
65
- throttle: this.settings.api?.throttle,
65
+ retry: api.retry,
66
+ throttle: api.throttle,
66
67
  call: async () => {
67
68
  return (0, postToApi_js_1.postJsonToApi)({
68
69
  url: api.assembleUrl(`/${this.settings.model}`),
@@ -91,16 +92,13 @@ class HuggingFaceTextGenerationModel extends AbstractModel_js_1.AbstractModel {
91
92
  }
92
93
  get settingsForEvent() {
93
94
  const eventSettingProperties = [
94
- "stopSequences",
95
- "maxGenerationTokens",
96
- "numberOfGenerations",
95
+ ...TextGenerationModel_js_1.textGenerationModelProperties,
97
96
  "topK",
98
97
  "topP",
99
98
  "temperature",
100
99
  "repetitionPenalty",
101
100
  "maxTime",
102
101
  "doSample",
103
- "options",
104
102
  ];
105
103
  return Object.fromEntries(Object.entries(this.settings).filter(([key]) => eventSettingProperties.includes(key)));
106
104
  }
@@ -108,12 +106,15 @@ class HuggingFaceTextGenerationModel extends AbstractModel_js_1.AbstractModel {
108
106
  const response = await this.callAPI(prompt, options);
109
107
  return {
110
108
  response,
111
- texts: response.map((response) => response.generated_text),
109
+ textGenerationResults: response.map((response) => ({
110
+ text: response.generated_text,
111
+ finishReason: "unknown",
112
+ })),
112
113
  };
113
114
  }
114
115
  withPromptTemplate(promptTemplate) {
115
116
  return new PromptTemplateTextGenerationModel_js_1.PromptTemplateTextGenerationModel({
116
- model: this,
117
+ model: this, // stop tokens are not supported by this model
117
118
  promptTemplate,
118
119
  });
119
120
  }
@@ -14,10 +14,6 @@ export interface HuggingFaceTextGenerationModelSettings extends TextGenerationMo
14
14
  repetitionPenalty?: number;
15
15
  maxTime?: number;
16
16
  doSample?: boolean;
17
- options?: {
18
- useCache?: boolean;
19
- waitForModel?: boolean;
20
- };
21
17
  }
22
18
  /**
23
19
  * Create a text generation model that calls a Hugging Face Inference API Text Generation Task.
@@ -50,7 +46,10 @@ export declare class HuggingFaceTextGenerationModel extends AbstractModel<Huggin
50
46
  response: {
51
47
  generated_text: string;
52
48
  }[];
53
- texts: string[];
49
+ textGenerationResults: {
50
+ text: string;
51
+ finishReason: "unknown";
52
+ }[];
54
53
  }>;
55
54
  withPromptTemplate<INPUT_PROMPT>(promptTemplate: TextGenerationPromptTemplate<INPUT_PROMPT, string>): PromptTemplateTextGenerationModel<INPUT_PROMPT, string, HuggingFaceTextGenerationModelSettings, this>;
56
55
  withSettings(additionalSettings: Partial<HuggingFaceTextGenerationModelSettings>): this;
@@ -3,6 +3,7 @@ import { callWithRetryAndThrottle } from "../../core/api/callWithRetryAndThrottl
3
3
  import { createJsonResponseHandler, postJsonToApi, } from "../../core/api/postToApi.js";
4
4
  import { AbstractModel } from "../../model-function/AbstractModel.js";
5
5
  import { PromptTemplateTextGenerationModel } from "../../model-function/generate-text/PromptTemplateTextGenerationModel.js";
6
+ import { textGenerationModelProperties, } from "../../model-function/generate-text/TextGenerationModel.js";
6
7
  import { HuggingFaceApiConfiguration } from "./HuggingFaceApiConfiguration.js";
7
8
  import { failedHuggingFaceCallResponseHandler } from "./HuggingFaceError.js";
8
9
  /**
@@ -58,8 +59,8 @@ export class HuggingFaceTextGenerationModel extends AbstractModel {
58
59
  const api = this.settings.api ?? new HuggingFaceApiConfiguration();
59
60
  const abortSignal = options?.run?.abortSignal;
60
61
  return callWithRetryAndThrottle({
61
- retry: this.settings.api?.retry,
62
- throttle: this.settings.api?.throttle,
62
+ retry: api.retry,
63
+ throttle: api.throttle,
63
64
  call: async () => {
64
65
  return postJsonToApi({
65
66
  url: api.assembleUrl(`/${this.settings.model}`),
@@ -88,16 +89,13 @@ export class HuggingFaceTextGenerationModel extends AbstractModel {
88
89
  }
89
90
  get settingsForEvent() {
90
91
  const eventSettingProperties = [
91
- "stopSequences",
92
- "maxGenerationTokens",
93
- "numberOfGenerations",
92
+ ...textGenerationModelProperties,
94
93
  "topK",
95
94
  "topP",
96
95
  "temperature",
97
96
  "repetitionPenalty",
98
97
  "maxTime",
99
98
  "doSample",
100
- "options",
101
99
  ];
102
100
  return Object.fromEntries(Object.entries(this.settings).filter(([key]) => eventSettingProperties.includes(key)));
103
101
  }
@@ -105,12 +103,15 @@ export class HuggingFaceTextGenerationModel extends AbstractModel {
105
103
  const response = await this.callAPI(prompt, options);
106
104
  return {
107
105
  response,
108
- texts: response.map((response) => response.generated_text),
106
+ textGenerationResults: response.map((response) => ({
107
+ text: response.generated_text,
108
+ finishReason: "unknown",
109
+ })),
109
110
  };
110
111
  }
111
112
  withPromptTemplate(promptTemplate) {
112
113
  return new PromptTemplateTextGenerationModel({
113
- model: this,
114
+ model: this, // stop tokens are not supported by this model
114
115
  promptTemplate,
115
116
  });
116
117
  }
@@ -7,6 +7,7 @@ const postToApi_js_1 = require("../../core/api/postToApi.cjs");
7
7
  const ZodSchema_js_1 = require("../../core/schema/ZodSchema.cjs");
8
8
  const AbstractModel_js_1 = require("../../model-function/AbstractModel.cjs");
9
9
  const PromptTemplateTextStreamingModel_js_1 = require("../../model-function/generate-text/PromptTemplateTextStreamingModel.cjs");
10
+ const TextGenerationModel_js_1 = require("../../model-function/generate-text/TextGenerationModel.cjs");
10
11
  const AsyncQueue_js_1 = require("../../util/AsyncQueue.cjs");
11
12
  const parseJSON_js_1 = require("../../core/schema/parseJSON.cjs");
12
13
  const parseEventSourceStream_js_1 = require("../../util/streaming/parseEventSourceStream.cjs");
@@ -37,25 +38,52 @@ class LlamaCppTextGenerationModel extends AbstractModel_js_1.AbstractModel {
37
38
  return this.settings.contextWindowSize;
38
39
  }
39
40
  async callAPI(prompt, options) {
41
+ const api = this.settings.api ?? new LlamaCppApiConfiguration_js_1.LlamaCppApiConfiguration();
42
+ const responseFormat = options.responseFormat;
43
+ const abortSignal = options.run?.abortSignal;
40
44
  return (0, callWithRetryAndThrottle_js_1.callWithRetryAndThrottle)({
41
- retry: this.settings.api?.retry,
42
- throttle: this.settings.api?.throttle,
43
- call: async () => callLlamaCppTextGenerationAPI({
44
- ...this.settings,
45
- // mapping
46
- nPredict: this.settings.maxGenerationTokens,
47
- stop: this.settings.stopSequences,
48
- // other
49
- abortSignal: options.run?.abortSignal,
50
- prompt,
51
- responseFormat: options.responseFormat,
45
+ retry: api.retry,
46
+ throttle: api.throttle,
47
+ call: async () => (0, postToApi_js_1.postJsonToApi)({
48
+ url: api.assembleUrl(`/completion`),
49
+ headers: api.headers,
50
+ body: {
51
+ stream: responseFormat.stream,
52
+ prompt: prompt.text,
53
+ image_data: prompt.images != null
54
+ ? Object.entries(prompt.images).map(([id, data]) => ({
55
+ id: +id,
56
+ data,
57
+ }))
58
+ : undefined,
59
+ cache_prompt: this.settings.cachePrompt,
60
+ temperature: this.settings.temperature,
61
+ top_k: this.settings.topK,
62
+ top_p: this.settings.topP,
63
+ n_predict: this.settings.maxGenerationTokens,
64
+ n_keep: this.settings.nKeep,
65
+ stop: this.settings.stopSequences,
66
+ tfs_z: this.settings.tfsZ,
67
+ typical_p: this.settings.typicalP,
68
+ repeat_penalty: this.settings.repeatPenalty,
69
+ repeat_last_n: this.settings.repeatLastN,
70
+ penalize_nl: this.settings.penalizeNl,
71
+ mirostat: this.settings.mirostat,
72
+ mirostat_tau: this.settings.mirostatTau,
73
+ mirostat_eta: this.settings.mirostatEta,
74
+ seed: this.settings.seed,
75
+ ignore_eos: this.settings.ignoreEos,
76
+ logit_bias: this.settings.logitBias,
77
+ },
78
+ failedResponseHandler: LlamaCppError_js_1.failedLlamaCppCallResponseHandler,
79
+ successfulResponseHandler: responseFormat.handler,
80
+ abortSignal,
52
81
  }),
53
82
  });
54
83
  }
55
84
  get settingsForEvent() {
56
85
  const eventSettingProperties = [
57
- "maxGenerationTokens",
58
- "stopSequences",
86
+ ...TextGenerationModel_js_1.textGenerationModelProperties,
59
87
  "contextWindowSize",
60
88
  "cachePrompt",
61
89
  "temperature",
@@ -87,7 +115,16 @@ class LlamaCppTextGenerationModel extends AbstractModel_js_1.AbstractModel {
87
115
  });
88
116
  return {
89
117
  response,
90
- texts: [response.content],
118
+ textGenerationResults: [
119
+ {
120
+ text: response.content,
121
+ finishReason: response.stopped_eos || response.stopped_word
122
+ ? "stop"
123
+ : response.stopped_limit
124
+ ? "length"
125
+ : "unknown",
126
+ },
127
+ ],
91
128
  usage: {
92
129
  promptTokens: response.tokens_evaluated,
93
130
  completionTokens: response.tokens_predicted,
@@ -198,43 +235,6 @@ const llamaCppTextStreamingResponseSchema = new ZodSchema_js_1.ZodSchema(zod_1.z
198
235
  }),
199
236
  llamaCppTextGenerationResponseSchema,
200
237
  ]));
201
- async function callLlamaCppTextGenerationAPI({ api = new LlamaCppApiConfiguration_js_1.LlamaCppApiConfiguration(), abortSignal, responseFormat, prompt, cachePrompt, temperature, topK, topP, nPredict, nKeep, stop, tfsZ, typicalP, repeatPenalty, repeatLastN, penalizeNl, mirostat, mirostatTau, mirostatEta, seed, ignoreEos, logitBias, }) {
202
- return (0, postToApi_js_1.postJsonToApi)({
203
- url: api.assembleUrl(`/completion`),
204
- headers: api.headers,
205
- body: {
206
- stream: responseFormat.stream,
207
- prompt: prompt.text,
208
- cache_prompt: cachePrompt,
209
- temperature,
210
- top_k: topK,
211
- top_p: topP,
212
- n_predict: nPredict,
213
- n_keep: nKeep,
214
- stop,
215
- tfs_z: tfsZ,
216
- typical_p: typicalP,
217
- repeat_penalty: repeatPenalty,
218
- repeat_last_n: repeatLastN,
219
- penalize_nl: penalizeNl,
220
- mirostat,
221
- mirostat_tau: mirostatTau,
222
- mirostat_eta: mirostatEta,
223
- seed,
224
- ignore_eos: ignoreEos,
225
- logit_bias: logitBias,
226
- image_data: prompt.images != null
227
- ? Object.entries(prompt.images).map(([id, data]) => ({
228
- id: +id,
229
- data,
230
- }))
231
- : undefined,
232
- },
233
- failedResponseHandler: LlamaCppError_js_1.failedLlamaCppCallResponseHandler,
234
- successfulResponseHandler: responseFormat.handler,
235
- abortSignal,
236
- });
237
- }
238
238
  async function createLlamaCppFullDeltaIterableQueue(stream) {
239
239
  const queue = new AsyncQueue_js_1.AsyncQueue();
240
240
  let content = "";
@@ -58,16 +58,16 @@ export declare class LlamaCppTextGenerationModel<CONTEXT_WINDOW_SIZE extends num
58
58
  countPromptTokens(prompt: LlamaCppTextGenerationPrompt): Promise<number>;
59
59
  doGenerateTexts(prompt: LlamaCppTextGenerationPrompt, options?: FunctionOptions): Promise<{
60
60
  response: {
61
+ stop: true;
61
62
  model: string;
62
63
  prompt: string;
63
64
  content: string;
64
- stop: true;
65
65
  generation_settings: {
66
+ stop: string[];
66
67
  model: string;
67
68
  stream: boolean;
68
- mirostat: number;
69
69
  seed: number;
70
- stop: string[];
70
+ mirostat: number;
71
71
  frequency_penalty: number;
72
72
  ignore_eos: boolean;
73
73
  logit_bias: number[];
@@ -106,7 +106,10 @@ export declare class LlamaCppTextGenerationModel<CONTEXT_WINDOW_SIZE extends num
106
106
  tokens_predicted: number;
107
107
  truncated: boolean;
108
108
  };
109
- texts: string[];
109
+ textGenerationResults: {
110
+ text: string;
111
+ finishReason: "length" | "stop" | "unknown";
112
+ }[];
110
113
  usage: {
111
114
  promptTokens: number;
112
115
  completionTokens: number;
@@ -153,11 +156,11 @@ declare const llamaCppTextGenerationResponseSchema: z.ZodObject<{
153
156
  top_p: z.ZodNumber;
154
157
  typical_p: z.ZodNumber;
155
158
  }, "strip", z.ZodTypeAny, {
159
+ stop: string[];
156
160
  model: string;
157
161
  stream: boolean;
158
- mirostat: number;
159
162
  seed: number;
160
- stop: string[];
163
+ mirostat: number;
161
164
  frequency_penalty: number;
162
165
  ignore_eos: boolean;
163
166
  logit_bias: number[];
@@ -177,11 +180,11 @@ declare const llamaCppTextGenerationResponseSchema: z.ZodObject<{
177
180
  top_p: number;
178
181
  typical_p: number;
179
182
  }, {
183
+ stop: string[];
180
184
  model: string;
181
185
  stream: boolean;
182
- mirostat: number;
183
186
  seed: number;
184
- stop: string[];
187
+ mirostat: number;
185
188
  frequency_penalty: number;
186
189
  ignore_eos: boolean;
187
190
  logit_bias: number[];
@@ -240,16 +243,16 @@ declare const llamaCppTextGenerationResponseSchema: z.ZodObject<{
240
243
  tokens_predicted: z.ZodNumber;
241
244
  truncated: z.ZodBoolean;
242
245
  }, "strip", z.ZodTypeAny, {
246
+ stop: true;
243
247
  model: string;
244
248
  prompt: string;
245
249
  content: string;
246
- stop: true;
247
250
  generation_settings: {
251
+ stop: string[];
248
252
  model: string;
249
253
  stream: boolean;
250
- mirostat: number;
251
254
  seed: number;
252
- stop: string[];
255
+ mirostat: number;
253
256
  frequency_penalty: number;
254
257
  ignore_eos: boolean;
255
258
  logit_bias: number[];
@@ -288,16 +291,16 @@ declare const llamaCppTextGenerationResponseSchema: z.ZodObject<{
288
291
  tokens_predicted: number;
289
292
  truncated: boolean;
290
293
  }, {
294
+ stop: true;
291
295
  model: string;
292
296
  prompt: string;
293
297
  content: string;
294
- stop: true;
295
298
  generation_settings: {
299
+ stop: string[];
296
300
  model: string;
297
301
  stream: boolean;
298
- mirostat: number;
299
302
  seed: number;
300
- stop: string[];
303
+ mirostat: number;
301
304
  frequency_penalty: number;
302
305
  ignore_eos: boolean;
303
306
  logit_bias: number[];
@@ -353,16 +356,16 @@ export declare const LlamaCppTextGenerationResponseFormat: {
353
356
  json: {
354
357
  stream: false;
355
358
  handler: ResponseHandler<{
359
+ stop: true;
356
360
  model: string;
357
361
  prompt: string;
358
362
  content: string;
359
- stop: true;
360
363
  generation_settings: {
364
+ stop: string[];
361
365
  model: string;
362
366
  stream: boolean;
363
- mirostat: number;
364
367
  seed: number;
365
- stop: string[];
368
+ mirostat: number;
366
369
  frequency_penalty: number;
367
370
  ignore_eos: boolean;
368
371
  logit_bias: number[];