modelfusion 0.103.0 → 0.104.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +7 -0
- package/model-function/generate-text/prompt-template/AlpacaPromptTemplate.cjs +2 -1
- package/model-function/generate-text/prompt-template/AlpacaPromptTemplate.d.ts +2 -2
- package/model-function/generate-text/prompt-template/AlpacaPromptTemplate.js +2 -1
- package/model-function/generate-text/prompt-template/ChatMLPromptTemplate.cjs +5 -4
- package/model-function/generate-text/prompt-template/ChatMLPromptTemplate.d.ts +4 -4
- package/model-function/generate-text/prompt-template/ChatMLPromptTemplate.js +5 -4
- package/model-function/generate-text/prompt-template/ChatPrompt.cjs +0 -24
- package/model-function/generate-text/prompt-template/ChatPrompt.d.ts +11 -34
- package/model-function/generate-text/prompt-template/ChatPrompt.js +1 -22
- package/model-function/generate-text/prompt-template/Content.cjs +9 -0
- package/model-function/generate-text/prompt-template/Content.d.ts +9 -4
- package/model-function/generate-text/prompt-template/Content.js +7 -1
- package/model-function/generate-text/prompt-template/InstructionPrompt.d.ts +6 -22
- package/model-function/generate-text/prompt-template/Llama2PromptTemplate.cjs +36 -5
- package/model-function/generate-text/prompt-template/Llama2PromptTemplate.d.ts +16 -4
- package/model-function/generate-text/prompt-template/Llama2PromptTemplate.js +34 -4
- package/model-function/generate-text/prompt-template/NeuralChatPromptTemplate.cjs +5 -4
- package/model-function/generate-text/prompt-template/NeuralChatPromptTemplate.d.ts +4 -4
- package/model-function/generate-text/prompt-template/NeuralChatPromptTemplate.js +5 -4
- package/model-function/generate-text/prompt-template/TextPromptTemplate.cjs +3 -4
- package/model-function/generate-text/prompt-template/TextPromptTemplate.d.ts +4 -4
- package/model-function/generate-text/prompt-template/TextPromptTemplate.js +3 -4
- package/model-function/generate-text/prompt-template/VicunaPromptTemplate.cjs +3 -3
- package/model-function/generate-text/prompt-template/VicunaPromptTemplate.d.ts +2 -2
- package/model-function/generate-text/prompt-template/VicunaPromptTemplate.js +3 -3
- package/model-function/generate-text/prompt-template/trimChatPrompt.cjs +0 -2
- package/model-function/generate-text/prompt-template/trimChatPrompt.d.ts +4 -4
- package/model-function/generate-text/prompt-template/trimChatPrompt.js +0 -2
- package/model-provider/anthropic/AnthropicPromptTemplate.cjs +5 -4
- package/model-provider/anthropic/AnthropicPromptTemplate.d.ts +4 -4
- package/model-provider/anthropic/AnthropicPromptTemplate.js +5 -4
- package/model-provider/anthropic/AnthropicTextGenerationModel.d.ts +2 -2
- package/model-provider/automatic1111/Automatic1111ImageGenerationModel.d.ts +1 -1
- package/model-provider/cohere/CohereTextGenerationModel.d.ts +2 -2
- package/model-provider/llamacpp/LlamaCppBakLLaVA1PromptTemplate.cjs +23 -16
- package/model-provider/llamacpp/LlamaCppBakLLaVA1PromptTemplate.d.ts +4 -4
- package/model-provider/llamacpp/LlamaCppBakLLaVA1PromptTemplate.js +23 -16
- package/model-provider/mistral/MistralChatModel.d.ts +2 -2
- package/model-provider/mistral/MistralPromptTemplate.cjs +5 -4
- package/model-provider/mistral/MistralPromptTemplate.d.ts +4 -4
- package/model-provider/mistral/MistralPromptTemplate.js +5 -4
- package/model-provider/ollama/OllamaChatModel.d.ts +2 -2
- package/model-provider/ollama/OllamaChatPromptTemplate.cjs +9 -13
- package/model-provider/ollama/OllamaChatPromptTemplate.d.ts +4 -4
- package/model-provider/ollama/OllamaChatPromptTemplate.js +9 -13
- package/model-provider/openai/OpenAICompletionModel.d.ts +2 -2
- package/model-provider/openai/chat/OpenAIChatFunctionCallStructureGenerationModel.d.ts +2 -2
- package/model-provider/openai/chat/OpenAIChatMessage.d.ts +2 -2
- package/model-provider/openai/chat/OpenAIChatModel.d.ts +2 -2
- package/model-provider/openai/chat/OpenAIChatPromptTemplate.cjs +0 -2
- package/model-provider/openai/chat/OpenAIChatPromptTemplate.d.ts +4 -4
- package/model-provider/openai/chat/OpenAIChatPromptTemplate.js +0 -2
- package/model-provider/openai-compatible/OpenAICompatibleChatModel.d.ts +2 -2
- package/model-provider/stability/StabilityImageGenerationModel.d.ts +1 -1
- package/package.json +1 -1
| @@ -1,7 +1,6 @@ | |
| 1 1 | 
             
            "use strict";
         | 
| 2 2 | 
             
            Object.defineProperty(exports, "__esModule", { value: true });
         | 
| 3 3 | 
             
            exports.chat = exports.instruction = exports.text = exports.identity = void 0;
         | 
| 4 | 
            -
            const ChatPrompt_js_1 = require("../../model-function/generate-text/prompt-template/ChatPrompt.cjs");
         | 
| 5 4 | 
             
            /**
         | 
| 6 5 | 
             
             * OllamaChatPrompt identity chat format.
         | 
| 7 6 | 
             
             */
         | 
| @@ -42,7 +41,6 @@ exports.instruction = instruction; | |
| 42 41 | 
             
            function chat() {
         | 
| 43 42 | 
             
                return {
         | 
| 44 43 | 
             
                    format(prompt) {
         | 
| 45 | 
            -
                        (0, ChatPrompt_js_1.validateChatPrompt)(prompt);
         | 
| 46 44 | 
             
                        const messages = [];
         | 
| 47 45 | 
             
                        if (prompt.system != null) {
         | 
| 48 46 | 
             
                            messages.push({ role: "system", content: prompt.system });
         | 
| @@ -60,17 +58,15 @@ function extractContent(input) { | |
| 60 58 | 
             
                if (typeof input === "string") {
         | 
| 61 59 | 
             
                    return { content: input, images: undefined };
         | 
| 62 60 | 
             
                }
         | 
| 63 | 
            -
                 | 
| 64 | 
            -
             | 
| 65 | 
            -
             | 
| 66 | 
            -
                     | 
| 67 | 
            -
                         | 
| 68 | 
            -
             | 
| 69 | 
            -
             | 
| 70 | 
            -
                         | 
| 71 | 
            -
                            images.push(part.base64Image);
         | 
| 72 | 
            -
                        }
         | 
| 61 | 
            +
                const images = [];
         | 
| 62 | 
            +
                let content = "";
         | 
| 63 | 
            +
                for (const part of input) {
         | 
| 64 | 
            +
                    if (part.type === "text") {
         | 
| 65 | 
            +
                        content += part.text;
         | 
| 66 | 
            +
                    }
         | 
| 67 | 
            +
                    else {
         | 
| 68 | 
            +
                        images.push(part.base64Image);
         | 
| 73 69 | 
             
                    }
         | 
| 74 | 
            -
                    return { content, images };
         | 
| 75 70 | 
             
                }
         | 
| 71 | 
            +
                return { content, images };
         | 
| 76 72 | 
             
            }
         | 
| @@ -1,6 +1,6 @@ | |
| 1 1 | 
             
            import { TextGenerationPromptTemplate } from "../../model-function/generate-text/TextGenerationPromptTemplate.js";
         | 
| 2 | 
            -
            import {  | 
| 3 | 
            -
            import {  | 
| 2 | 
            +
            import { ChatPrompt } from "../../model-function/generate-text/prompt-template/ChatPrompt.js";
         | 
| 3 | 
            +
            import { InstructionPrompt } from "../../model-function/generate-text/prompt-template/InstructionPrompt.js";
         | 
| 4 4 | 
             
            import { OllamaChatPrompt } from "./OllamaChatModel.js";
         | 
| 5 5 | 
             
            /**
         | 
| 6 6 | 
             
             * OllamaChatPrompt identity chat format.
         | 
| @@ -13,8 +13,8 @@ export declare function text(): TextGenerationPromptTemplate<string, OllamaChatP | |
| 13 13 | 
             
            /**
         | 
| 14 14 | 
             
             * Formats an instruction prompt as an Ollama chat prompt.
         | 
| 15 15 | 
             
             */
         | 
| 16 | 
            -
            export declare function instruction(): TextGenerationPromptTemplate< | 
| 16 | 
            +
            export declare function instruction(): TextGenerationPromptTemplate<InstructionPrompt, OllamaChatPrompt>;
         | 
| 17 17 | 
             
            /**
         | 
| 18 18 | 
             
             * Formats a chat prompt as an Ollama chat prompt.
         | 
| 19 19 | 
             
             */
         | 
| 20 | 
            -
            export declare function chat(): TextGenerationPromptTemplate< | 
| 20 | 
            +
            export declare function chat(): TextGenerationPromptTemplate<ChatPrompt, OllamaChatPrompt>;
         | 
| @@ -1,4 +1,3 @@ | |
| 1 | 
            -
            import { validateChatPrompt, } from "../../model-function/generate-text/prompt-template/ChatPrompt.js";
         | 
| 2 1 | 
             
            /**
         | 
| 3 2 | 
             
             * OllamaChatPrompt identity chat format.
         | 
| 4 3 | 
             
             */
         | 
| @@ -36,7 +35,6 @@ export function instruction() { | |
| 36 35 | 
             
            export function chat() {
         | 
| 37 36 | 
             
                return {
         | 
| 38 37 | 
             
                    format(prompt) {
         | 
| 39 | 
            -
                        validateChatPrompt(prompt);
         | 
| 40 38 | 
             
                        const messages = [];
         | 
| 41 39 | 
             
                        if (prompt.system != null) {
         | 
| 42 40 | 
             
                            messages.push({ role: "system", content: prompt.system });
         | 
| @@ -53,17 +51,15 @@ function extractContent(input) { | |
| 53 51 | 
             
                if (typeof input === "string") {
         | 
| 54 52 | 
             
                    return { content: input, images: undefined };
         | 
| 55 53 | 
             
                }
         | 
| 56 | 
            -
                 | 
| 57 | 
            -
             | 
| 58 | 
            -
             | 
| 59 | 
            -
                     | 
| 60 | 
            -
                         | 
| 61 | 
            -
             | 
| 62 | 
            -
             | 
| 63 | 
            -
                         | 
| 64 | 
            -
                            images.push(part.base64Image);
         | 
| 65 | 
            -
                        }
         | 
| 54 | 
            +
                const images = [];
         | 
| 55 | 
            +
                let content = "";
         | 
| 56 | 
            +
                for (const part of input) {
         | 
| 57 | 
            +
                    if (part.type === "text") {
         | 
| 58 | 
            +
                        content += part.text;
         | 
| 59 | 
            +
                    }
         | 
| 60 | 
            +
                    else {
         | 
| 61 | 
            +
                        images.push(part.base64Image);
         | 
| 66 62 | 
             
                    }
         | 
| 67 | 
            -
                    return { content, images };
         | 
| 68 63 | 
             
                }
         | 
| 64 | 
            +
                return { content, images };
         | 
| 69 65 | 
             
            }
         | 
| @@ -178,14 +178,14 @@ export declare class OpenAICompletionModel extends AbstractModel<OpenAICompletio | |
| 178 178 | 
             
                /**
         | 
| 179 179 | 
             
                 * Returns this model with an instruction prompt template.
         | 
| 180 180 | 
             
                 */
         | 
| 181 | 
            -
                withInstructionPrompt(): PromptTemplateTextStreamingModel<import("../../index.js"). | 
| 181 | 
            +
                withInstructionPrompt(): PromptTemplateTextStreamingModel<import("../../index.js").InstructionPrompt, string, OpenAICompletionModelSettings, this>;
         | 
| 182 182 | 
             
                /**
         | 
| 183 183 | 
             
                 * Returns this model with a chat prompt template.
         | 
| 184 184 | 
             
                 */
         | 
| 185 185 | 
             
                withChatPrompt(options?: {
         | 
| 186 186 | 
             
                    user?: string;
         | 
| 187 187 | 
             
                    assistant?: string;
         | 
| 188 | 
            -
                }): PromptTemplateTextStreamingModel<import("../../index.js"). | 
| 188 | 
            +
                }): PromptTemplateTextStreamingModel<import("../../index.js").ChatPrompt, string, OpenAICompletionModelSettings, this>;
         | 
| 189 189 | 
             
                withPromptTemplate<INPUT_PROMPT>(promptTemplate: TextGenerationPromptTemplate<INPUT_PROMPT, string>): PromptTemplateTextStreamingModel<INPUT_PROMPT, string, OpenAICompletionModelSettings, this>;
         | 
| 190 190 | 
             
                withSettings(additionalSettings: Partial<OpenAICompletionModelSettings>): this;
         | 
| 191 191 | 
             
            }
         | 
| @@ -27,11 +27,11 @@ OpenAIChatSettings> { | |
| 27 27 | 
             
                /**
         | 
| 28 28 | 
             
                 * Returns this model with an instruction prompt template.
         | 
| 29 29 | 
             
                 */
         | 
| 30 | 
            -
                withInstructionPrompt(): OpenAIChatFunctionCallStructureGenerationModel<TextGenerationPromptTemplate<import("../../../index.js"). | 
| 30 | 
            +
                withInstructionPrompt(): OpenAIChatFunctionCallStructureGenerationModel<TextGenerationPromptTemplate<import("../../../index.js").InstructionPrompt, OpenAIChatPrompt>>;
         | 
| 31 31 | 
             
                /**
         | 
| 32 32 | 
             
                 * Returns this model with a chat prompt template.
         | 
| 33 33 | 
             
                 */
         | 
| 34 | 
            -
                withChatPrompt(): OpenAIChatFunctionCallStructureGenerationModel<TextGenerationPromptTemplate<import("../../../index.js"). | 
| 34 | 
            +
                withChatPrompt(): OpenAIChatFunctionCallStructureGenerationModel<TextGenerationPromptTemplate<import("../../../index.js").ChatPrompt, OpenAIChatPrompt>>;
         | 
| 35 35 | 
             
                withPromptTemplate<TARGET_PROMPT_FORMAT extends TextGenerationPromptTemplate<unknown, OpenAIChatPrompt>>(promptTemplate: TARGET_PROMPT_FORMAT): OpenAIChatFunctionCallStructureGenerationModel<TARGET_PROMPT_FORMAT>;
         | 
| 36 36 | 
             
                withSettings(additionalSettings: Partial<OpenAIChatSettings>): this;
         | 
| 37 37 | 
             
                /**
         | 
| @@ -1,4 +1,4 @@ | |
| 1 | 
            -
            import {  | 
| 1 | 
            +
            import { Content } from "../../../model-function/generate-text/prompt-template/Content.js";
         | 
| 2 2 | 
             
            import { ToolCall } from "../../../tool/ToolCall.js";
         | 
| 3 3 | 
             
            export type OpenAIChatMessage = {
         | 
| 4 4 | 
             
                role: "system";
         | 
| @@ -50,7 +50,7 @@ export declare const OpenAIChatMessage: { | |
| 50 50 | 
             
                /**
         | 
| 51 51 | 
             
                 * Creates a user chat message. The message can be a string or a multi-modal input.
         | 
| 52 52 | 
             
                 */
         | 
| 53 | 
            -
                user(content:  | 
| 53 | 
            +
                user(content: Content, options?: {
         | 
| 54 54 | 
             
                    name?: string;
         | 
| 55 55 | 
             
                }): OpenAIChatMessage;
         | 
| 56 56 | 
             
                /**
         | 
| @@ -151,11 +151,11 @@ export declare class OpenAIChatModel extends AbstractOpenAIChatModel<OpenAIChatS | |
| 151 151 | 
             
                /**
         | 
| 152 152 | 
             
                 * Returns this model with an instruction prompt template.
         | 
| 153 153 | 
             
                 */
         | 
| 154 | 
            -
                withInstructionPrompt(): PromptTemplateTextStreamingModel<import("../../../index.js"). | 
| 154 | 
            +
                withInstructionPrompt(): PromptTemplateTextStreamingModel<import("../../../index.js").InstructionPrompt, OpenAIChatPrompt, OpenAIChatSettings, this>;
         | 
| 155 155 | 
             
                /**
         | 
| 156 156 | 
             
                 * Returns this model with a chat prompt template.
         | 
| 157 157 | 
             
                 */
         | 
| 158 | 
            -
                withChatPrompt(): PromptTemplateTextStreamingModel<import("../../../index.js"). | 
| 158 | 
            +
                withChatPrompt(): PromptTemplateTextStreamingModel<import("../../../index.js").ChatPrompt, OpenAIChatPrompt, OpenAIChatSettings, this>;
         | 
| 159 159 | 
             
                withPromptTemplate<INPUT_PROMPT>(promptTemplate: TextGenerationPromptTemplate<INPUT_PROMPT, OpenAIChatPrompt>): PromptTemplateTextStreamingModel<INPUT_PROMPT, OpenAIChatPrompt, OpenAIChatSettings, this>;
         | 
| 160 160 | 
             
                withSettings(additionalSettings: Partial<OpenAIChatSettings>): this;
         | 
| 161 161 | 
             
            }
         | 
| @@ -1,7 +1,6 @@ | |
| 1 1 | 
             
            "use strict";
         | 
| 2 2 | 
             
            Object.defineProperty(exports, "__esModule", { value: true });
         | 
| 3 3 | 
             
            exports.chat = exports.instruction = exports.text = exports.identity = void 0;
         | 
| 4 | 
            -
            const ChatPrompt_js_1 = require("../../../model-function/generate-text/prompt-template/ChatPrompt.cjs");
         | 
| 5 4 | 
             
            const OpenAIChatMessage_js_1 = require("./OpenAIChatMessage.cjs");
         | 
| 6 5 | 
             
            /**
         | 
| 7 6 | 
             
             * OpenAIMessage[] identity chat format.
         | 
| @@ -43,7 +42,6 @@ exports.instruction = instruction; | |
| 43 42 | 
             
            function chat() {
         | 
| 44 43 | 
             
                return {
         | 
| 45 44 | 
             
                    format(prompt) {
         | 
| 46 | 
            -
                        (0, ChatPrompt_js_1.validateChatPrompt)(prompt);
         | 
| 47 45 | 
             
                        const messages = [];
         | 
| 48 46 | 
             
                        if (prompt.system != null) {
         | 
| 49 47 | 
             
                            messages.push(OpenAIChatMessage_js_1.OpenAIChatMessage.system(prompt.system));
         | 
| @@ -1,6 +1,6 @@ | |
| 1 1 | 
             
            import { TextGenerationPromptTemplate } from "../../../model-function/generate-text/TextGenerationPromptTemplate.js";
         | 
| 2 | 
            -
            import {  | 
| 3 | 
            -
            import {  | 
| 2 | 
            +
            import { ChatPrompt } from "../../../model-function/generate-text/prompt-template/ChatPrompt.js";
         | 
| 3 | 
            +
            import { InstructionPrompt } from "../../../model-function/generate-text/prompt-template/InstructionPrompt.js";
         | 
| 4 4 | 
             
            import { OpenAIChatPrompt } from "./AbstractOpenAIChatModel.js";
         | 
| 5 5 | 
             
            /**
         | 
| 6 6 | 
             
             * OpenAIMessage[] identity chat format.
         | 
| @@ -13,8 +13,8 @@ export declare function text(): TextGenerationPromptTemplate<string, OpenAIChatP | |
| 13 13 | 
             
            /**
         | 
| 14 14 | 
             
             * Formats an instruction prompt as an OpenAI chat prompt.
         | 
| 15 15 | 
             
             */
         | 
| 16 | 
            -
            export declare function instruction(): TextGenerationPromptTemplate< | 
| 16 | 
            +
            export declare function instruction(): TextGenerationPromptTemplate<InstructionPrompt, OpenAIChatPrompt>;
         | 
| 17 17 | 
             
            /**
         | 
| 18 18 | 
             
             * Formats a chat prompt as an OpenAI chat prompt.
         | 
| 19 19 | 
             
             */
         | 
| 20 | 
            -
            export declare function chat(): TextGenerationPromptTemplate< | 
| 20 | 
            +
            export declare function chat(): TextGenerationPromptTemplate<ChatPrompt, OpenAIChatPrompt>;
         | 
| @@ -1,4 +1,3 @@ | |
| 1 | 
            -
            import { validateChatPrompt, } from "../../../model-function/generate-text/prompt-template/ChatPrompt.js";
         | 
| 2 1 | 
             
            import { OpenAIChatMessage } from "./OpenAIChatMessage.js";
         | 
| 3 2 | 
             
            /**
         | 
| 4 3 | 
             
             * OpenAIMessage[] identity chat format.
         | 
| @@ -37,7 +36,6 @@ export function instruction() { | |
| 37 36 | 
             
            export function chat() {
         | 
| 38 37 | 
             
                return {
         | 
| 39 38 | 
             
                    format(prompt) {
         | 
| 40 | 
            -
                        validateChatPrompt(prompt);
         | 
| 41 39 | 
             
                        const messages = [];
         | 
| 42 40 | 
             
                        if (prompt.system != null) {
         | 
| 43 41 | 
             
                            messages.push(OpenAIChatMessage.system(prompt.system));
         | 
| @@ -35,11 +35,11 @@ export declare class OpenAICompatibleChatModel extends AbstractOpenAIChatModel<O | |
| 35 35 | 
             
                /**
         | 
| 36 36 | 
             
                 * Returns this model with an instruction prompt template.
         | 
| 37 37 | 
             
                 */
         | 
| 38 | 
            -
                withInstructionPrompt(): PromptTemplateTextStreamingModel<import("../../index.js"). | 
| 38 | 
            +
                withInstructionPrompt(): PromptTemplateTextStreamingModel<import("../../index.js").InstructionPrompt, OpenAIChatPrompt, OpenAICompatibleChatSettings, this>;
         | 
| 39 39 | 
             
                /**
         | 
| 40 40 | 
             
                 * Returns this model with a chat prompt template.
         | 
| 41 41 | 
             
                 */
         | 
| 42 | 
            -
                withChatPrompt(): PromptTemplateTextStreamingModel<import("../../index.js"). | 
| 42 | 
            +
                withChatPrompt(): PromptTemplateTextStreamingModel<import("../../index.js").ChatPrompt, OpenAIChatPrompt, OpenAICompatibleChatSettings, this>;
         | 
| 43 43 | 
             
                withPromptTemplate<INPUT_PROMPT>(promptTemplate: TextGenerationPromptTemplate<INPUT_PROMPT, OpenAIChatPrompt>): PromptTemplateTextStreamingModel<INPUT_PROMPT, OpenAIChatPrompt, OpenAICompatibleChatSettings, this>;
         | 
| 44 44 | 
             
                withSettings(additionalSettings: Partial<OpenAICompatibleChatSettings>): this;
         | 
| 45 45 | 
             
            }
         | 
| @@ -78,7 +78,7 @@ export declare class StabilityImageGenerationModel extends AbstractModel<Stabili | |
| 78 78 | 
             
                }>;
         | 
| 79 79 | 
             
                withTextPrompt(): PromptTemplateImageGenerationModel<string, StabilityImageGenerationPrompt, StabilityImageGenerationSettings, this>;
         | 
| 80 80 | 
             
                withPromptTemplate<INPUT_PROMPT>(promptTemplate: PromptTemplate<INPUT_PROMPT, StabilityImageGenerationPrompt>): PromptTemplateImageGenerationModel<INPUT_PROMPT, StabilityImageGenerationPrompt, StabilityImageGenerationSettings, this>;
         | 
| 81 | 
            -
                withSettings(additionalSettings: StabilityImageGenerationSettings): this;
         | 
| 81 | 
            +
                withSettings(additionalSettings: Partial<StabilityImageGenerationSettings>): this;
         | 
| 82 82 | 
             
            }
         | 
| 83 83 | 
             
            declare const stabilityImageGenerationResponseSchema: z.ZodObject<{
         | 
| 84 84 | 
             
                artifacts: z.ZodArray<z.ZodObject<{
         |