modelfusion 0.117.0 → 0.118.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +26 -0
- package/README.md +10 -9
- package/core/getFunctionCallLogger.cjs +6 -6
- package/core/getFunctionCallLogger.js +6 -6
- package/model-function/ModelCallEvent.d.ts +1 -1
- package/model-function/embed/EmbeddingEvent.d.ts +1 -1
- package/model-function/embed/EmbeddingModel.d.ts +1 -1
- package/model-function/embed/embed.cjs +5 -5
- package/model-function/embed/embed.d.ts +2 -2
- package/model-function/embed/embed.js +5 -5
- package/model-function/executeStandardCall.cjs +3 -3
- package/model-function/executeStandardCall.d.ts +2 -2
- package/model-function/executeStandardCall.js +3 -3
- package/model-function/generate-image/ImageGenerationEvent.d.ts +1 -1
- package/model-function/generate-image/ImageGenerationModel.d.ts +1 -1
- package/model-function/generate-image/PromptTemplateImageGenerationModel.d.ts +1 -1
- package/model-function/generate-image/generateImage.cjs +2 -2
- package/model-function/generate-image/generateImage.d.ts +1 -1
- package/model-function/generate-image/generateImage.js +2 -2
- package/model-function/generate-speech/SpeechGenerationEvent.d.ts +1 -1
- package/model-function/generate-speech/generateSpeech.cjs +2 -2
- package/model-function/generate-speech/generateSpeech.d.ts +1 -1
- package/model-function/generate-speech/generateSpeech.js +2 -2
- package/model-function/generate-structure/StructureFromTextGenerationModel.cjs +1 -1
- package/model-function/generate-structure/StructureFromTextGenerationModel.js +1 -1
- package/model-function/generate-structure/StructureFromTextStreamingModel.cjs +1 -1
- package/model-function/generate-structure/StructureFromTextStreamingModel.js +1 -1
- package/model-function/generate-structure/StructureGenerationEvent.d.ts +1 -1
- package/model-function/generate-structure/generateStructure.cjs +2 -2
- package/model-function/generate-structure/generateStructure.d.ts +1 -1
- package/model-function/generate-structure/generateStructure.js +2 -2
- package/model-function/generate-text/PromptTemplateFullTextModel.d.ts +2 -2
- package/model-function/generate-text/PromptTemplateTextGenerationModel.d.ts +2 -2
- package/model-function/generate-text/TextGenerationEvent.d.ts +1 -1
- package/model-function/generate-text/TextGenerationModel.d.ts +2 -2
- package/model-function/generate-text/generateText.cjs +3 -3
- package/model-function/generate-text/generateText.d.ts +1 -1
- package/model-function/generate-text/generateText.js +3 -3
- package/model-function/generate-text/prompt-template/AlpacaPromptTemplate.cjs +8 -1
- package/model-function/generate-text/prompt-template/AlpacaPromptTemplate.d.ts +5 -0
- package/model-function/generate-text/prompt-template/AlpacaPromptTemplate.js +6 -0
- package/model-function/generate-text/prompt-template/PromptTemplateProvider.cjs +2 -0
- package/model-function/generate-text/prompt-template/PromptTemplateProvider.d.ts +8 -0
- package/model-function/generate-text/prompt-template/PromptTemplateProvider.js +1 -0
- package/model-function/generate-text/prompt-template/VicunaPromptTemplate.cjs +34 -1
- package/model-function/generate-text/prompt-template/VicunaPromptTemplate.d.ts +9 -0
- package/model-function/generate-text/prompt-template/VicunaPromptTemplate.js +31 -0
- package/model-function/generate-text/prompt-template/VicunaPromptTemplate.test.cjs +28 -0
- package/model-function/generate-text/prompt-template/VicunaPromptTemplate.test.js +29 -1
- package/model-function/generate-text/prompt-template/index.cjs +1 -0
- package/model-function/generate-text/prompt-template/index.d.ts +1 -0
- package/model-function/generate-text/prompt-template/index.js +1 -0
- package/model-function/generate-transcription/TranscriptionEvent.d.ts +1 -1
- package/model-function/generate-transcription/TranscriptionModel.d.ts +1 -1
- package/model-function/generate-transcription/generateTranscription.cjs +1 -1
- package/model-function/generate-transcription/generateTranscription.d.ts +1 -1
- package/model-function/generate-transcription/generateTranscription.js +1 -1
- package/model-provider/automatic1111/Automatic1111ImageGenerationModel.cjs +3 -3
- package/model-provider/automatic1111/Automatic1111ImageGenerationModel.d.ts +1 -1
- package/model-provider/automatic1111/Automatic1111ImageGenerationModel.js +3 -3
- package/model-provider/cohere/CohereTextEmbeddingModel.cjs +3 -3
- package/model-provider/cohere/CohereTextEmbeddingModel.d.ts +1 -1
- package/model-provider/cohere/CohereTextEmbeddingModel.js +3 -3
- package/model-provider/cohere/CohereTextGenerationModel.cjs +3 -3
- package/model-provider/cohere/CohereTextGenerationModel.d.ts +4 -4
- package/model-provider/cohere/CohereTextGenerationModel.js +3 -3
- package/model-provider/huggingface/HuggingFaceTextEmbeddingModel.cjs +3 -3
- package/model-provider/huggingface/HuggingFaceTextEmbeddingModel.d.ts +1 -1
- package/model-provider/huggingface/HuggingFaceTextEmbeddingModel.js +3 -3
- package/model-provider/huggingface/HuggingFaceTextGenerationModel.cjs +3 -3
- package/model-provider/huggingface/HuggingFaceTextGenerationModel.d.ts +4 -4
- package/model-provider/huggingface/HuggingFaceTextGenerationModel.js +3 -3
- package/model-provider/llamacpp/LlamaCppBakLLaVA1PromptTemplate.cjs +15 -1
- package/model-provider/llamacpp/LlamaCppBakLLaVA1PromptTemplate.d.ts +4 -0
- package/model-provider/llamacpp/LlamaCppBakLLaVA1PromptTemplate.js +13 -0
- package/model-provider/llamacpp/LlamaCppCompletionModel.cjs +31 -28
- package/model-provider/llamacpp/LlamaCppCompletionModel.d.ts +17 -8
- package/model-provider/llamacpp/LlamaCppCompletionModel.js +31 -28
- package/model-provider/llamacpp/LlamaCppFacade.cjs +4 -3
- package/model-provider/llamacpp/LlamaCppFacade.d.ts +2 -1
- package/model-provider/llamacpp/LlamaCppFacade.js +2 -1
- package/model-provider/llamacpp/LlamaCppPrompt.cjs +59 -0
- package/model-provider/llamacpp/LlamaCppPrompt.d.ts +14 -0
- package/model-provider/llamacpp/LlamaCppPrompt.js +31 -0
- package/model-provider/llamacpp/LlamaCppTextEmbeddingModel.cjs +3 -3
- package/model-provider/llamacpp/LlamaCppTextEmbeddingModel.d.ts +1 -1
- package/model-provider/llamacpp/LlamaCppTextEmbeddingModel.js +3 -3
- package/model-provider/llamacpp/index.cjs +2 -3
- package/model-provider/llamacpp/index.d.ts +1 -2
- package/model-provider/llamacpp/index.js +1 -2
- package/model-provider/mistral/MistralChatModel.cjs +3 -3
- package/model-provider/mistral/MistralChatModel.d.ts +4 -4
- package/model-provider/mistral/MistralChatModel.js +3 -3
- package/model-provider/mistral/MistralTextEmbeddingModel.cjs +3 -3
- package/model-provider/mistral/MistralTextEmbeddingModel.d.ts +1 -1
- package/model-provider/mistral/MistralTextEmbeddingModel.js +3 -3
- package/model-provider/ollama/OllamaChatModel.cjs +3 -3
- package/model-provider/ollama/OllamaChatModel.d.ts +2 -2
- package/model-provider/ollama/OllamaChatModel.js +3 -3
- package/model-provider/ollama/OllamaCompletionModel.cjs +3 -3
- package/model-provider/ollama/OllamaCompletionModel.d.ts +14 -14
- package/model-provider/ollama/OllamaCompletionModel.js +3 -3
- package/model-provider/ollama/OllamaTextEmbeddingModel.cjs +3 -3
- package/model-provider/ollama/OllamaTextEmbeddingModel.d.ts +1 -1
- package/model-provider/ollama/OllamaTextEmbeddingModel.js +3 -3
- package/model-provider/openai/AbstractOpenAIChatModel.cjs +12 -12
- package/model-provider/openai/AbstractOpenAIChatModel.d.ts +6 -6
- package/model-provider/openai/AbstractOpenAIChatModel.js +12 -12
- package/model-provider/openai/AbstractOpenAICompletionModel.cjs +6 -6
- package/model-provider/openai/AbstractOpenAICompletionModel.d.ts +2 -2
- package/model-provider/openai/AbstractOpenAICompletionModel.js +6 -6
- package/model-provider/openai/OpenAIImageGenerationModel.cjs +3 -3
- package/model-provider/openai/OpenAIImageGenerationModel.d.ts +1 -1
- package/model-provider/openai/OpenAIImageGenerationModel.js +3 -3
- package/model-provider/openai/OpenAITextEmbeddingModel.cjs +3 -3
- package/model-provider/openai/OpenAITextEmbeddingModel.d.ts +1 -1
- package/model-provider/openai/OpenAITextEmbeddingModel.js +3 -3
- package/model-provider/openai/OpenAITranscriptionModel.cjs +3 -3
- package/model-provider/openai/OpenAITranscriptionModel.d.ts +1 -1
- package/model-provider/openai/OpenAITranscriptionModel.js +3 -3
- package/model-provider/stability/StabilityImageGenerationModel.cjs +3 -3
- package/model-provider/stability/StabilityImageGenerationModel.d.ts +1 -1
- package/model-provider/stability/StabilityImageGenerationModel.js +3 -3
- package/model-provider/whispercpp/WhisperCppTranscriptionModel.cjs +3 -3
- package/model-provider/whispercpp/WhisperCppTranscriptionModel.d.ts +1 -1
- package/model-provider/whispercpp/WhisperCppTranscriptionModel.js +3 -3
- package/package.json +1 -1
- package/tool/generate-tool-call/TextGenerationToolCallModel.cjs +2 -2
- package/tool/generate-tool-call/TextGenerationToolCallModel.d.ts +1 -1
- package/tool/generate-tool-call/TextGenerationToolCallModel.js +2 -2
- package/tool/generate-tool-call/ToolCallGenerationEvent.d.ts +1 -1
- package/tool/generate-tool-call/ToolCallGenerationModel.d.ts +1 -1
- package/tool/generate-tool-call/generateToolCall.cjs +2 -2
- package/tool/generate-tool-call/generateToolCall.js +2 -2
- package/tool/generate-tool-calls/TextGenerationToolCallsModel.cjs +2 -2
- package/tool/generate-tool-calls/TextGenerationToolCallsModel.d.ts +1 -1
- package/tool/generate-tool-calls/TextGenerationToolCallsModel.js +2 -2
- package/tool/generate-tool-calls/ToolCallsGenerationEvent.d.ts +1 -1
- package/tool/generate-tool-calls/ToolCallsGenerationModel.d.ts +1 -1
- package/tool/generate-tool-calls/generateToolCalls.cjs +2 -2
- package/tool/generate-tool-calls/generateToolCalls.d.ts +1 -1
- package/tool/generate-tool-calls/generateToolCalls.js +2 -2
@@ -76,10 +76,10 @@ export class MistralTextEmbeddingModel extends AbstractModel {
|
|
76
76
|
};
|
77
77
|
}
|
78
78
|
async doEmbedValues(texts, options) {
|
79
|
-
const
|
79
|
+
const rawResponse = await this.callAPI(texts, options);
|
80
80
|
return {
|
81
|
-
|
82
|
-
embeddings:
|
81
|
+
rawResponse,
|
82
|
+
embeddings: rawResponse.data.map((entry) => entry.embedding),
|
83
83
|
};
|
84
84
|
}
|
85
85
|
withSettings(additionalSettings) {
|
@@ -129,12 +129,12 @@ class OllamaChatModel extends AbstractModel_js_1.AbstractModel {
|
|
129
129
|
schema: (0, ZodSchema_js_1.zodSchema)(ollamaChatResponseSchema),
|
130
130
|
}));
|
131
131
|
}
|
132
|
-
processTextGenerationResponse(
|
132
|
+
processTextGenerationResponse(rawResponse) {
|
133
133
|
return {
|
134
|
-
|
134
|
+
rawResponse,
|
135
135
|
textGenerationResults: [
|
136
136
|
{
|
137
|
-
text:
|
137
|
+
text: rawResponse.message.content,
|
138
138
|
finishReason: "unknown",
|
139
139
|
},
|
140
140
|
],
|
@@ -39,7 +39,7 @@ export declare class OllamaChatModel extends AbstractModel<OllamaChatModelSettin
|
|
39
39
|
}): Promise<RESPONSE>;
|
40
40
|
get settingsForEvent(): Partial<OllamaChatModelSettings>;
|
41
41
|
doGenerateTexts(prompt: OllamaChatPrompt, options: FunctionCallOptions): Promise<{
|
42
|
-
|
42
|
+
rawResponse: {
|
43
43
|
model: string;
|
44
44
|
message: {
|
45
45
|
role: string;
|
@@ -60,7 +60,7 @@ export declare class OllamaChatModel extends AbstractModel<OllamaChatModelSettin
|
|
60
60
|
}[];
|
61
61
|
}>;
|
62
62
|
restoreGeneratedTexts(rawResponse: unknown): {
|
63
|
-
|
63
|
+
rawResponse: {
|
64
64
|
model: string;
|
65
65
|
message: {
|
66
66
|
role: string;
|
@@ -126,12 +126,12 @@ export class OllamaChatModel extends AbstractModel {
|
|
126
126
|
schema: zodSchema(ollamaChatResponseSchema),
|
127
127
|
}));
|
128
128
|
}
|
129
|
-
processTextGenerationResponse(
|
129
|
+
processTextGenerationResponse(rawResponse) {
|
130
130
|
return {
|
131
|
-
|
131
|
+
rawResponse,
|
132
132
|
textGenerationResults: [
|
133
133
|
{
|
134
|
-
text:
|
134
|
+
text: rawResponse.message.content,
|
135
135
|
finishReason: "unknown",
|
136
136
|
},
|
137
137
|
],
|
@@ -130,12 +130,12 @@ class OllamaCompletionModel extends AbstractModel_js_1.AbstractModel {
|
|
130
130
|
schema: (0, ZodSchema_js_1.zodSchema)(ollamaCompletionResponseSchema),
|
131
131
|
}));
|
132
132
|
}
|
133
|
-
processTextGenerationResponse(
|
133
|
+
processTextGenerationResponse(rawResponse) {
|
134
134
|
return {
|
135
|
-
|
135
|
+
rawResponse,
|
136
136
|
textGenerationResults: [
|
137
137
|
{
|
138
|
-
text:
|
138
|
+
text: rawResponse.response,
|
139
139
|
finishReason: "unknown",
|
140
140
|
},
|
141
141
|
],
|
@@ -52,10 +52,10 @@ export declare class OllamaCompletionModel<CONTEXT_WINDOW_SIZE extends number |
|
|
52
52
|
}): Promise<RESPONSE>;
|
53
53
|
get settingsForEvent(): Partial<OllamaCompletionModelSettings<CONTEXT_WINDOW_SIZE>>;
|
54
54
|
doGenerateTexts(prompt: OllamaCompletionPrompt, options: FunctionCallOptions): Promise<{
|
55
|
-
|
56
|
-
response: string;
|
55
|
+
rawResponse: {
|
57
56
|
model: string;
|
58
57
|
done: true;
|
58
|
+
response: string;
|
59
59
|
created_at: string;
|
60
60
|
total_duration: number;
|
61
61
|
prompt_eval_count: number;
|
@@ -71,10 +71,10 @@ export declare class OllamaCompletionModel<CONTEXT_WINDOW_SIZE extends number |
|
|
71
71
|
}[];
|
72
72
|
}>;
|
73
73
|
restoreGeneratedTexts(rawResponse: unknown): {
|
74
|
-
|
75
|
-
response: string;
|
74
|
+
rawResponse: {
|
76
75
|
model: string;
|
77
76
|
done: true;
|
77
|
+
response: string;
|
78
78
|
created_at: string;
|
79
79
|
total_duration: number;
|
80
80
|
prompt_eval_count: number;
|
@@ -89,11 +89,11 @@ export declare class OllamaCompletionModel<CONTEXT_WINDOW_SIZE extends number |
|
|
89
89
|
finishReason: "unknown";
|
90
90
|
}[];
|
91
91
|
};
|
92
|
-
processTextGenerationResponse(
|
93
|
-
|
94
|
-
response: string;
|
92
|
+
processTextGenerationResponse(rawResponse: OllamaCompletionResponse): {
|
93
|
+
rawResponse: {
|
95
94
|
model: string;
|
96
95
|
done: true;
|
96
|
+
response: string;
|
97
97
|
created_at: string;
|
98
98
|
total_duration: number;
|
99
99
|
prompt_eval_count: number;
|
@@ -109,9 +109,9 @@ export declare class OllamaCompletionModel<CONTEXT_WINDOW_SIZE extends number |
|
|
109
109
|
}[];
|
110
110
|
};
|
111
111
|
doStreamText(prompt: OllamaCompletionPrompt, options: FunctionCallOptions): Promise<AsyncIterable<import("../../index.js").Delta<{
|
112
|
-
response: string;
|
113
112
|
model: string;
|
114
113
|
done: false;
|
114
|
+
response: string;
|
115
115
|
created_at: string;
|
116
116
|
} | {
|
117
117
|
model: string;
|
@@ -151,9 +151,9 @@ declare const ollamaCompletionResponseSchema: z.ZodObject<{
|
|
151
151
|
eval_duration: z.ZodNumber;
|
152
152
|
context: z.ZodOptional<z.ZodArray<z.ZodNumber, "many">>;
|
153
153
|
}, "strip", z.ZodTypeAny, {
|
154
|
-
response: string;
|
155
154
|
model: string;
|
156
155
|
done: true;
|
156
|
+
response: string;
|
157
157
|
created_at: string;
|
158
158
|
total_duration: number;
|
159
159
|
prompt_eval_count: number;
|
@@ -163,9 +163,9 @@ declare const ollamaCompletionResponseSchema: z.ZodObject<{
|
|
163
163
|
prompt_eval_duration?: number | undefined;
|
164
164
|
context?: number[] | undefined;
|
165
165
|
}, {
|
166
|
-
response: string;
|
167
166
|
model: string;
|
168
167
|
done: true;
|
168
|
+
response: string;
|
169
169
|
created_at: string;
|
170
170
|
total_duration: number;
|
171
171
|
prompt_eval_count: number;
|
@@ -182,14 +182,14 @@ declare const ollamaCompletionStreamChunkSchema: z.ZodDiscriminatedUnion<"done",
|
|
182
182
|
created_at: z.ZodString;
|
183
183
|
response: z.ZodString;
|
184
184
|
}, "strip", z.ZodTypeAny, {
|
185
|
-
response: string;
|
186
185
|
model: string;
|
187
186
|
done: false;
|
187
|
+
response: string;
|
188
188
|
created_at: string;
|
189
189
|
}, {
|
190
|
-
response: string;
|
191
190
|
model: string;
|
192
191
|
done: false;
|
192
|
+
response: string;
|
193
193
|
created_at: string;
|
194
194
|
}>, z.ZodObject<{
|
195
195
|
done: z.ZodLiteral<true>;
|
@@ -247,9 +247,9 @@ export declare const OllamaCompletionResponseFormat: {
|
|
247
247
|
requestBodyValues: unknown;
|
248
248
|
response: Response;
|
249
249
|
}) => Promise<{
|
250
|
-
response: string;
|
251
250
|
model: string;
|
252
251
|
done: true;
|
252
|
+
response: string;
|
253
253
|
created_at: string;
|
254
254
|
total_duration: number;
|
255
255
|
prompt_eval_count: number;
|
@@ -269,9 +269,9 @@ export declare const OllamaCompletionResponseFormat: {
|
|
269
269
|
handler: ({ response }: {
|
270
270
|
response: Response;
|
271
271
|
}) => Promise<AsyncIterable<import("../../index.js").Delta<{
|
272
|
-
response: string;
|
273
272
|
model: string;
|
274
273
|
done: false;
|
274
|
+
response: string;
|
275
275
|
created_at: string;
|
276
276
|
} | {
|
277
277
|
model: string;
|
@@ -127,12 +127,12 @@ export class OllamaCompletionModel extends AbstractModel {
|
|
127
127
|
schema: zodSchema(ollamaCompletionResponseSchema),
|
128
128
|
}));
|
129
129
|
}
|
130
|
-
processTextGenerationResponse(
|
130
|
+
processTextGenerationResponse(rawResponse) {
|
131
131
|
return {
|
132
|
-
|
132
|
+
rawResponse,
|
133
133
|
textGenerationResults: [
|
134
134
|
{
|
135
|
-
text:
|
135
|
+
text: rawResponse.response,
|
136
136
|
finishReason: "unknown",
|
137
137
|
},
|
138
138
|
],
|
@@ -66,10 +66,10 @@ class OllamaTextEmbeddingModel extends AbstractModel_js_1.AbstractModel {
|
|
66
66
|
};
|
67
67
|
}
|
68
68
|
async doEmbedValues(texts, options) {
|
69
|
-
const
|
69
|
+
const rawResponse = await this.callAPI(texts, options);
|
70
70
|
return {
|
71
|
-
|
72
|
-
embeddings: [
|
71
|
+
rawResponse,
|
72
|
+
embeddings: [rawResponse.embedding],
|
73
73
|
};
|
74
74
|
}
|
75
75
|
withSettings(additionalSettings) {
|
@@ -19,7 +19,7 @@ export declare class OllamaTextEmbeddingModel extends AbstractModel<OllamaTextEm
|
|
19
19
|
callAPI(texts: Array<string>, callOptions: FunctionCallOptions): Promise<OllamaTextEmbeddingResponse>;
|
20
20
|
get settingsForEvent(): Partial<OllamaTextEmbeddingModelSettings>;
|
21
21
|
doEmbedValues(texts: string[], options: FunctionCallOptions): Promise<{
|
22
|
-
|
22
|
+
rawResponse: {
|
23
23
|
embedding: number[];
|
24
24
|
};
|
25
25
|
embeddings: number[][];
|
@@ -63,10 +63,10 @@ export class OllamaTextEmbeddingModel extends AbstractModel {
|
|
63
63
|
};
|
64
64
|
}
|
65
65
|
async doEmbedValues(texts, options) {
|
66
|
-
const
|
66
|
+
const rawResponse = await this.callAPI(texts, options);
|
67
67
|
return {
|
68
|
-
|
69
|
-
embeddings: [
|
68
|
+
rawResponse,
|
69
|
+
embeddings: [rawResponse.embedding],
|
70
70
|
};
|
71
71
|
}
|
72
72
|
withSettings(additionalSettings) {
|
@@ -90,14 +90,14 @@ class AbstractOpenAIChatModel extends AbstractModel_js_1.AbstractModel {
|
|
90
90
|
schema: (0, ZodSchema_js_1.zodSchema)(openAIChatResponseSchema),
|
91
91
|
}));
|
92
92
|
}
|
93
|
-
processTextGenerationResponse(
|
93
|
+
processTextGenerationResponse(rawResponse) {
|
94
94
|
return {
|
95
|
-
|
96
|
-
textGenerationResults:
|
95
|
+
rawResponse,
|
96
|
+
textGenerationResults: rawResponse.choices.map((choice) => ({
|
97
97
|
text: choice.message.content ?? "",
|
98
98
|
finishReason: this.translateFinishReason(choice.finish_reason),
|
99
99
|
})),
|
100
|
-
usage: this.extractUsage(
|
100
|
+
usage: this.extractUsage(rawResponse),
|
101
101
|
};
|
102
102
|
}
|
103
103
|
translateFinishReason(finishReason) {
|
@@ -133,7 +133,7 @@ class AbstractOpenAIChatModel extends AbstractModel_js_1.AbstractModel {
|
|
133
133
|
return firstChoice.delta.content ?? undefined;
|
134
134
|
}
|
135
135
|
async doGenerateToolCall(tool, prompt, options) {
|
136
|
-
const
|
136
|
+
const rawResponse = await this.callAPI(prompt, options, {
|
137
137
|
responseFormat: exports.OpenAIChatResponseFormat.json,
|
138
138
|
toolChoice: {
|
139
139
|
type: "function",
|
@@ -150,20 +150,20 @@ class AbstractOpenAIChatModel extends AbstractModel_js_1.AbstractModel {
|
|
150
150
|
},
|
151
151
|
],
|
152
152
|
});
|
153
|
-
const toolCalls =
|
153
|
+
const toolCalls = rawResponse.choices[0]?.message.tool_calls;
|
154
154
|
return {
|
155
|
-
|
155
|
+
rawResponse,
|
156
156
|
toolCall: toolCalls == null || toolCalls.length === 0
|
157
157
|
? null
|
158
158
|
: {
|
159
159
|
id: toolCalls[0].id,
|
160
160
|
args: (0, parseJSON_js_1.parseJSON)({ text: toolCalls[0].function.arguments }),
|
161
161
|
},
|
162
|
-
usage: this.extractUsage(
|
162
|
+
usage: this.extractUsage(rawResponse),
|
163
163
|
};
|
164
164
|
}
|
165
165
|
async doGenerateToolCalls(tools, prompt, options) {
|
166
|
-
const
|
166
|
+
const rawResponse = await this.callAPI(prompt, options, {
|
167
167
|
responseFormat: exports.OpenAIChatResponseFormat.json,
|
168
168
|
toolChoice: "auto",
|
169
169
|
tools: tools.map((tool) => ({
|
@@ -175,16 +175,16 @@ class AbstractOpenAIChatModel extends AbstractModel_js_1.AbstractModel {
|
|
175
175
|
},
|
176
176
|
})),
|
177
177
|
});
|
178
|
-
const message =
|
178
|
+
const message = rawResponse.choices[0]?.message;
|
179
179
|
return {
|
180
|
-
|
180
|
+
rawResponse,
|
181
181
|
text: message.content ?? null,
|
182
182
|
toolCalls: message.tool_calls?.map((toolCall) => ({
|
183
183
|
id: toolCall.id,
|
184
184
|
name: toolCall.function.name,
|
185
185
|
args: (0, parseJSON_js_1.parseJSON)({ text: toolCall.function.arguments }),
|
186
186
|
})) ?? null,
|
187
|
-
usage: this.extractUsage(
|
187
|
+
usage: this.extractUsage(rawResponse),
|
188
188
|
};
|
189
189
|
}
|
190
190
|
extractUsage(response) {
|
@@ -90,7 +90,7 @@ export declare abstract class AbstractOpenAIChatModel<SETTINGS extends AbstractO
|
|
90
90
|
toolChoice?: AbstractOpenAIChatSettings["toolChoice"];
|
91
91
|
}): Promise<RESULT>;
|
92
92
|
doGenerateTexts(prompt: OpenAIChatPrompt, options: FunctionCallOptions): Promise<{
|
93
|
-
|
93
|
+
rawResponse: {
|
94
94
|
object: "chat.completion";
|
95
95
|
model: string;
|
96
96
|
usage: {
|
@@ -134,7 +134,7 @@ export declare abstract class AbstractOpenAIChatModel<SETTINGS extends AbstractO
|
|
134
134
|
};
|
135
135
|
}>;
|
136
136
|
restoreGeneratedTexts(rawResponse: unknown): {
|
137
|
-
|
137
|
+
rawResponse: {
|
138
138
|
object: "chat.completion";
|
139
139
|
model: string;
|
140
140
|
usage: {
|
@@ -177,8 +177,8 @@ export declare abstract class AbstractOpenAIChatModel<SETTINGS extends AbstractO
|
|
177
177
|
totalTokens: number;
|
178
178
|
};
|
179
179
|
};
|
180
|
-
processTextGenerationResponse(
|
181
|
-
|
180
|
+
processTextGenerationResponse(rawResponse: OpenAIChatResponse): {
|
181
|
+
rawResponse: {
|
182
182
|
object: "chat.completion";
|
183
183
|
model: string;
|
184
184
|
usage: {
|
@@ -251,7 +251,7 @@ export declare abstract class AbstractOpenAIChatModel<SETTINGS extends AbstractO
|
|
251
251
|
}>>>;
|
252
252
|
extractTextDelta(delta: unknown): string | undefined;
|
253
253
|
doGenerateToolCall(tool: ToolDefinition<string, unknown>, prompt: OpenAIChatPrompt, options: FunctionCallOptions): Promise<{
|
254
|
-
|
254
|
+
rawResponse: {
|
255
255
|
object: "chat.completion";
|
256
256
|
model: string;
|
257
257
|
usage: {
|
@@ -295,7 +295,7 @@ export declare abstract class AbstractOpenAIChatModel<SETTINGS extends AbstractO
|
|
295
295
|
};
|
296
296
|
}>;
|
297
297
|
doGenerateToolCalls(tools: Array<ToolDefinition<string, unknown>>, prompt: OpenAIChatPrompt, options: FunctionCallOptions): Promise<{
|
298
|
-
|
298
|
+
rawResponse: {
|
299
299
|
object: "chat.completion";
|
300
300
|
model: string;
|
301
301
|
usage: {
|
@@ -87,14 +87,14 @@ export class AbstractOpenAIChatModel extends AbstractModel {
|
|
87
87
|
schema: zodSchema(openAIChatResponseSchema),
|
88
88
|
}));
|
89
89
|
}
|
90
|
-
processTextGenerationResponse(
|
90
|
+
processTextGenerationResponse(rawResponse) {
|
91
91
|
return {
|
92
|
-
|
93
|
-
textGenerationResults:
|
92
|
+
rawResponse,
|
93
|
+
textGenerationResults: rawResponse.choices.map((choice) => ({
|
94
94
|
text: choice.message.content ?? "",
|
95
95
|
finishReason: this.translateFinishReason(choice.finish_reason),
|
96
96
|
})),
|
97
|
-
usage: this.extractUsage(
|
97
|
+
usage: this.extractUsage(rawResponse),
|
98
98
|
};
|
99
99
|
}
|
100
100
|
translateFinishReason(finishReason) {
|
@@ -130,7 +130,7 @@ export class AbstractOpenAIChatModel extends AbstractModel {
|
|
130
130
|
return firstChoice.delta.content ?? undefined;
|
131
131
|
}
|
132
132
|
async doGenerateToolCall(tool, prompt, options) {
|
133
|
-
const
|
133
|
+
const rawResponse = await this.callAPI(prompt, options, {
|
134
134
|
responseFormat: OpenAIChatResponseFormat.json,
|
135
135
|
toolChoice: {
|
136
136
|
type: "function",
|
@@ -147,20 +147,20 @@ export class AbstractOpenAIChatModel extends AbstractModel {
|
|
147
147
|
},
|
148
148
|
],
|
149
149
|
});
|
150
|
-
const toolCalls =
|
150
|
+
const toolCalls = rawResponse.choices[0]?.message.tool_calls;
|
151
151
|
return {
|
152
|
-
|
152
|
+
rawResponse,
|
153
153
|
toolCall: toolCalls == null || toolCalls.length === 0
|
154
154
|
? null
|
155
155
|
: {
|
156
156
|
id: toolCalls[0].id,
|
157
157
|
args: parseJSON({ text: toolCalls[0].function.arguments }),
|
158
158
|
},
|
159
|
-
usage: this.extractUsage(
|
159
|
+
usage: this.extractUsage(rawResponse),
|
160
160
|
};
|
161
161
|
}
|
162
162
|
async doGenerateToolCalls(tools, prompt, options) {
|
163
|
-
const
|
163
|
+
const rawResponse = await this.callAPI(prompt, options, {
|
164
164
|
responseFormat: OpenAIChatResponseFormat.json,
|
165
165
|
toolChoice: "auto",
|
166
166
|
tools: tools.map((tool) => ({
|
@@ -172,16 +172,16 @@ export class AbstractOpenAIChatModel extends AbstractModel {
|
|
172
172
|
},
|
173
173
|
})),
|
174
174
|
});
|
175
|
-
const message =
|
175
|
+
const message = rawResponse.choices[0]?.message;
|
176
176
|
return {
|
177
|
-
|
177
|
+
rawResponse,
|
178
178
|
text: message.content ?? null,
|
179
179
|
toolCalls: message.tool_calls?.map((toolCall) => ({
|
180
180
|
id: toolCall.id,
|
181
181
|
name: toolCall.function.name,
|
182
182
|
args: parseJSON({ text: toolCall.function.arguments }),
|
183
183
|
})) ?? null,
|
184
|
-
usage: this.extractUsage(
|
184
|
+
usage: this.extractUsage(rawResponse),
|
185
185
|
};
|
186
186
|
}
|
187
187
|
extractUsage(response) {
|
@@ -79,19 +79,19 @@ class AbstractOpenAICompletionModel extends AbstractModel_js_1.AbstractModel {
|
|
79
79
|
schema: (0, ZodSchema_js_1.zodSchema)(OpenAICompletionResponseSchema),
|
80
80
|
}));
|
81
81
|
}
|
82
|
-
processTextGenerationResponse(
|
82
|
+
processTextGenerationResponse(rawResponse) {
|
83
83
|
return {
|
84
|
-
|
85
|
-
textGenerationResults:
|
84
|
+
rawResponse,
|
85
|
+
textGenerationResults: rawResponse.choices.map((choice) => {
|
86
86
|
return {
|
87
87
|
finishReason: this.translateFinishReason(choice.finish_reason),
|
88
88
|
text: choice.text,
|
89
89
|
};
|
90
90
|
}),
|
91
91
|
usage: {
|
92
|
-
promptTokens:
|
93
|
-
completionTokens:
|
94
|
-
totalTokens:
|
92
|
+
promptTokens: rawResponse.usage.prompt_tokens,
|
93
|
+
completionTokens: rawResponse.usage.completion_tokens,
|
94
|
+
totalTokens: rawResponse.usage.total_tokens,
|
95
95
|
},
|
96
96
|
};
|
97
97
|
}
|
@@ -31,7 +31,7 @@ export declare abstract class AbstractOpenAICompletionModel<SETTINGS extends Abs
|
|
31
31
|
responseFormat: OpenAITextResponseFormatType<RESULT>;
|
32
32
|
}): Promise<RESULT>;
|
33
33
|
doGenerateTexts(prompt: string, options: FunctionCallOptions): Promise<{
|
34
|
-
|
34
|
+
rawResponse: {
|
35
35
|
object: "text_completion";
|
36
36
|
model: string;
|
37
37
|
usage: {
|
@@ -60,7 +60,7 @@ export declare abstract class AbstractOpenAICompletionModel<SETTINGS extends Abs
|
|
60
60
|
};
|
61
61
|
}>;
|
62
62
|
restoreGeneratedTexts(rawResponse: unknown): {
|
63
|
-
|
63
|
+
rawResponse: {
|
64
64
|
object: "text_completion";
|
65
65
|
model: string;
|
66
66
|
usage: {
|
@@ -76,19 +76,19 @@ export class AbstractOpenAICompletionModel extends AbstractModel {
|
|
76
76
|
schema: zodSchema(OpenAICompletionResponseSchema),
|
77
77
|
}));
|
78
78
|
}
|
79
|
-
processTextGenerationResponse(
|
79
|
+
processTextGenerationResponse(rawResponse) {
|
80
80
|
return {
|
81
|
-
|
82
|
-
textGenerationResults:
|
81
|
+
rawResponse,
|
82
|
+
textGenerationResults: rawResponse.choices.map((choice) => {
|
83
83
|
return {
|
84
84
|
finishReason: this.translateFinishReason(choice.finish_reason),
|
85
85
|
text: choice.text,
|
86
86
|
};
|
87
87
|
}),
|
88
88
|
usage: {
|
89
|
-
promptTokens:
|
90
|
-
completionTokens:
|
91
|
-
totalTokens:
|
89
|
+
promptTokens: rawResponse.usage.prompt_tokens,
|
90
|
+
completionTokens: rawResponse.usage.completion_tokens,
|
91
|
+
totalTokens: rawResponse.usage.total_tokens,
|
92
92
|
},
|
93
93
|
};
|
94
94
|
}
|
@@ -127,12 +127,12 @@ class OpenAIImageGenerationModel extends AbstractModel_js_1.AbstractModel {
|
|
127
127
|
return Object.fromEntries(Object.entries(this.settings).filter(([key]) => eventSettingProperties.includes(key)));
|
128
128
|
}
|
129
129
|
async doGenerateImages(prompt, options) {
|
130
|
-
const
|
130
|
+
const rawResponse = await this.callAPI(prompt, options, {
|
131
131
|
responseFormat: exports.OpenAIImageGenerationResponseFormat.base64Json,
|
132
132
|
});
|
133
133
|
return {
|
134
|
-
|
135
|
-
base64Images:
|
134
|
+
rawResponse,
|
135
|
+
base64Images: rawResponse.data.map((item) => item.b64_json),
|
136
136
|
};
|
137
137
|
}
|
138
138
|
withPromptTemplate(promptTemplate) {
|
@@ -52,7 +52,7 @@ export declare class OpenAIImageGenerationModel extends AbstractModel<OpenAIImag
|
|
52
52
|
}): Promise<RESULT>;
|
53
53
|
get settingsForEvent(): Partial<OpenAIImageGenerationSettings>;
|
54
54
|
doGenerateImages(prompt: string, options: FunctionCallOptions): Promise<{
|
55
|
-
|
55
|
+
rawResponse: {
|
56
56
|
data: {
|
57
57
|
b64_json: string;
|
58
58
|
}[];
|
@@ -123,12 +123,12 @@ export class OpenAIImageGenerationModel extends AbstractModel {
|
|
123
123
|
return Object.fromEntries(Object.entries(this.settings).filter(([key]) => eventSettingProperties.includes(key)));
|
124
124
|
}
|
125
125
|
async doGenerateImages(prompt, options) {
|
126
|
-
const
|
126
|
+
const rawResponse = await this.callAPI(prompt, options, {
|
127
127
|
responseFormat: OpenAIImageGenerationResponseFormat.base64Json,
|
128
128
|
});
|
129
129
|
return {
|
130
|
-
|
131
|
-
base64Images:
|
130
|
+
rawResponse,
|
131
|
+
base64Images: rawResponse.data.map((item) => item.b64_json),
|
132
132
|
};
|
133
133
|
}
|
134
134
|
withPromptTemplate(promptTemplate) {
|
@@ -125,10 +125,10 @@ class OpenAITextEmbeddingModel extends AbstractModel_js_1.AbstractModel {
|
|
125
125
|
if (texts.length > this.maxValuesPerCall) {
|
126
126
|
throw new Error(`The OpenAI embedding API only supports ${this.maxValuesPerCall} texts per API call.`);
|
127
127
|
}
|
128
|
-
const
|
128
|
+
const rawResponse = await this.callAPI(texts, callOptions);
|
129
129
|
return {
|
130
|
-
|
131
|
-
embeddings:
|
130
|
+
rawResponse,
|
131
|
+
embeddings: rawResponse.data.map((data) => data.embedding),
|
132
132
|
};
|
133
133
|
}
|
134
134
|
withSettings(additionalSettings) {
|
@@ -50,7 +50,7 @@ export declare class OpenAITextEmbeddingModel extends AbstractModel<OpenAITextEm
|
|
50
50
|
callAPI(texts: Array<string>, callOptions: FunctionCallOptions): Promise<OpenAITextEmbeddingResponse>;
|
51
51
|
get settingsForEvent(): Partial<OpenAITextEmbeddingModelSettings>;
|
52
52
|
doEmbedValues(texts: string[], callOptions: FunctionCallOptions): Promise<{
|
53
|
-
|
53
|
+
rawResponse: {
|
54
54
|
object: "list";
|
55
55
|
model: string;
|
56
56
|
usage: {
|
@@ -120,10 +120,10 @@ export class OpenAITextEmbeddingModel extends AbstractModel {
|
|
120
120
|
if (texts.length > this.maxValuesPerCall) {
|
121
121
|
throw new Error(`The OpenAI embedding API only supports ${this.maxValuesPerCall} texts per API call.`);
|
122
122
|
}
|
123
|
-
const
|
123
|
+
const rawResponse = await this.callAPI(texts, callOptions);
|
124
124
|
return {
|
125
|
-
|
126
|
-
embeddings:
|
125
|
+
rawResponse,
|
126
|
+
embeddings: rawResponse.data.map((data) => data.embedding),
|
127
127
|
};
|
128
128
|
}
|
129
129
|
withSettings(additionalSettings) {
|
@@ -55,12 +55,12 @@ class OpenAITranscriptionModel extends AbstractModel_js_1.AbstractModel {
|
|
55
55
|
return this.settings.model;
|
56
56
|
}
|
57
57
|
async doTranscribe(data, options) {
|
58
|
-
const
|
58
|
+
const rawResponse = await this.callAPI(data, options, {
|
59
59
|
responseFormat: exports.OpenAITranscriptionResponseFormat.verboseJson,
|
60
60
|
});
|
61
61
|
return {
|
62
|
-
|
63
|
-
transcription:
|
62
|
+
rawResponse,
|
63
|
+
transcription: rawResponse.text,
|
64
64
|
};
|
65
65
|
}
|
66
66
|
async callAPI(data, callOptions, options) {
|
@@ -66,7 +66,7 @@ export declare class OpenAITranscriptionModel extends AbstractModel<OpenAITransc
|
|
66
66
|
readonly provider: "openai";
|
67
67
|
get modelName(): "whisper-1";
|
68
68
|
doTranscribe(data: OpenAITranscriptionInput, options: FunctionCallOptions): Promise<{
|
69
|
-
|
69
|
+
rawResponse: {
|
70
70
|
text: string;
|
71
71
|
duration: number;
|
72
72
|
task: "transcribe";
|
@@ -51,12 +51,12 @@ export class OpenAITranscriptionModel extends AbstractModel {
|
|
51
51
|
return this.settings.model;
|
52
52
|
}
|
53
53
|
async doTranscribe(data, options) {
|
54
|
-
const
|
54
|
+
const rawResponse = await this.callAPI(data, options, {
|
55
55
|
responseFormat: OpenAITranscriptionResponseFormat.verboseJson,
|
56
56
|
});
|
57
57
|
return {
|
58
|
-
|
59
|
-
transcription:
|
58
|
+
rawResponse,
|
59
|
+
transcription: rawResponse.text,
|
60
60
|
};
|
61
61
|
}
|
62
62
|
async callAPI(data, callOptions, options) {
|