modelfusion 0.110.0 → 0.112.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +26 -0
- package/README.md +41 -65
- package/core/FunctionEvent.d.ts +3 -3
- package/core/api/ApiFacade.cjs +20 -0
- package/core/api/ApiFacade.d.ts +4 -0
- package/core/api/ApiFacade.js +4 -0
- package/core/api/BaseUrlApiConfiguration.cjs +47 -5
- package/core/api/BaseUrlApiConfiguration.d.ts +23 -6
- package/core/api/BaseUrlApiConfiguration.js +45 -4
- package/core/api/BaseUrlApiConfiguration.test.cjs +11 -0
- package/core/api/BaseUrlApiConfiguration.test.js +9 -0
- package/core/api/callWithRetryAndThrottle.cjs +3 -3
- package/core/api/callWithRetryAndThrottle.js +3 -3
- package/core/api/index.cjs +15 -2
- package/core/api/index.d.ts +2 -2
- package/core/api/index.js +2 -2
- package/core/api/postToApi.cjs +28 -5
- package/core/api/postToApi.d.ts +5 -4
- package/core/api/postToApi.js +26 -4
- package/core/api/throttleOff.cjs +8 -0
- package/core/api/throttleOff.d.ts +5 -0
- package/core/api/throttleOff.js +4 -0
- package/internal/index.cjs +2 -8
- package/internal/index.d.ts +1 -1
- package/internal/index.js +1 -1
- package/model-function/ModelCallEvent.d.ts +3 -3
- package/model-function/generate-structure/generateStructure.d.ts +1 -1
- package/model-function/generate-text/PromptTemplateFullTextModel.cjs +2 -2
- package/model-function/generate-text/PromptTemplateFullTextModel.d.ts +3 -3
- package/model-function/generate-text/PromptTemplateFullTextModel.js +2 -2
- package/model-function/generate-text/PromptTemplateTextGenerationModel.cjs +2 -2
- package/model-function/generate-text/PromptTemplateTextGenerationModel.d.ts +3 -3
- package/model-function/generate-text/PromptTemplateTextGenerationModel.js +2 -2
- package/model-provider/anthropic/AnthropicApiConfiguration.cjs +14 -6
- package/model-provider/anthropic/AnthropicApiConfiguration.d.ts +7 -8
- package/model-provider/anthropic/AnthropicApiConfiguration.js +15 -7
- package/model-provider/anthropic/AnthropicError.cjs +7 -27
- package/model-provider/anthropic/AnthropicError.d.ts +21 -16
- package/model-provider/anthropic/AnthropicError.js +7 -25
- package/model-provider/anthropic/AnthropicFacade.cjs +10 -1
- package/model-provider/anthropic/AnthropicFacade.d.ts +9 -0
- package/model-provider/anthropic/AnthropicFacade.js +8 -0
- package/model-provider/anthropic/AnthropicTextGenerationModel.cjs +22 -24
- package/model-provider/anthropic/AnthropicTextGenerationModel.js +23 -25
- package/model-provider/anthropic/index.cjs +1 -4
- package/model-provider/anthropic/index.d.ts +1 -1
- package/model-provider/anthropic/index.js +0 -1
- package/model-provider/automatic1111/Automatic1111ApiConfiguration.cjs +12 -4
- package/model-provider/automatic1111/Automatic1111ApiConfiguration.d.ts +3 -3
- package/model-provider/automatic1111/Automatic1111ApiConfiguration.js +12 -4
- package/model-provider/automatic1111/Automatic1111Error.cjs +3 -3
- package/model-provider/automatic1111/Automatic1111Error.d.ts +13 -3
- package/model-provider/automatic1111/Automatic1111Error.js +4 -4
- package/model-provider/automatic1111/Automatic1111Facade.cjs +9 -9
- package/model-provider/automatic1111/Automatic1111Facade.d.ts +6 -6
- package/model-provider/automatic1111/Automatic1111Facade.js +7 -7
- package/model-provider/automatic1111/Automatic1111ImageGenerationModel.cjs +11 -7
- package/model-provider/automatic1111/Automatic1111ImageGenerationModel.js +11 -7
- package/model-provider/cohere/CohereApiConfiguration.cjs +14 -6
- package/model-provider/cohere/CohereApiConfiguration.d.ts +7 -8
- package/model-provider/cohere/CohereApiConfiguration.js +15 -7
- package/model-provider/cohere/CohereError.cjs +8 -43
- package/model-provider/cohere/CohereError.d.ts +9 -16
- package/model-provider/cohere/CohereError.js +8 -41
- package/model-provider/cohere/CohereFacade.cjs +12 -3
- package/model-provider/cohere/CohereFacade.d.ts +11 -2
- package/model-provider/cohere/CohereFacade.js +10 -2
- package/model-provider/cohere/CohereTextEmbeddingModel.cjs +18 -22
- package/model-provider/cohere/CohereTextEmbeddingModel.js +18 -22
- package/model-provider/cohere/CohereTextGenerationModel.cjs +31 -39
- package/model-provider/cohere/CohereTextGenerationModel.d.ts +110 -8
- package/model-provider/cohere/CohereTextGenerationModel.js +31 -39
- package/model-provider/cohere/CohereTokenizer.cjs +32 -41
- package/model-provider/cohere/CohereTokenizer.d.ts +2 -2
- package/model-provider/cohere/CohereTokenizer.js +32 -41
- package/model-provider/cohere/index.cjs +1 -3
- package/model-provider/cohere/index.d.ts +1 -1
- package/model-provider/cohere/index.js +0 -1
- package/model-provider/elevenlabs/ElevenLabsApiConfiguration.cjs +14 -6
- package/model-provider/elevenlabs/ElevenLabsApiConfiguration.d.ts +7 -8
- package/model-provider/elevenlabs/ElevenLabsApiConfiguration.js +15 -7
- package/model-provider/elevenlabs/ElevenLabsFacade.cjs +10 -1
- package/model-provider/elevenlabs/ElevenLabsFacade.d.ts +9 -0
- package/model-provider/elevenlabs/ElevenLabsFacade.js +8 -0
- package/model-provider/elevenlabs/ElevenLabsSpeechModel.cjs +42 -53
- package/model-provider/elevenlabs/ElevenLabsSpeechModel.js +44 -55
- package/model-provider/huggingface/HuggingFaceApiConfiguration.cjs +14 -6
- package/model-provider/huggingface/HuggingFaceApiConfiguration.d.ts +7 -8
- package/model-provider/huggingface/HuggingFaceApiConfiguration.js +15 -7
- package/model-provider/huggingface/HuggingFaceError.cjs +7 -29
- package/model-provider/huggingface/HuggingFaceError.d.ts +9 -16
- package/model-provider/huggingface/HuggingFaceError.js +7 -27
- package/model-provider/huggingface/HuggingFaceFacade.cjs +10 -1
- package/model-provider/huggingface/HuggingFaceFacade.d.ts +9 -0
- package/model-provider/huggingface/HuggingFaceFacade.js +8 -0
- package/model-provider/huggingface/HuggingFaceTextEmbeddingModel.cjs +17 -27
- package/model-provider/huggingface/HuggingFaceTextEmbeddingModel.js +17 -27
- package/model-provider/huggingface/HuggingFaceTextGenerationModel.cjs +22 -23
- package/model-provider/huggingface/HuggingFaceTextGenerationModel.js +22 -23
- package/model-provider/huggingface/index.cjs +1 -3
- package/model-provider/huggingface/index.d.ts +1 -1
- package/model-provider/huggingface/index.js +0 -1
- package/model-provider/llamacpp/LlamaCppApiConfiguration.cjs +13 -6
- package/model-provider/llamacpp/LlamaCppApiConfiguration.d.ts +7 -9
- package/model-provider/llamacpp/LlamaCppApiConfiguration.js +14 -7
- package/model-provider/llamacpp/LlamaCppCompletionModel.cjs +4 -4
- package/model-provider/llamacpp/LlamaCppCompletionModel.d.ts +173 -5
- package/model-provider/llamacpp/LlamaCppCompletionModel.js +4 -4
- package/model-provider/llamacpp/LlamaCppError.cjs +7 -27
- package/model-provider/llamacpp/LlamaCppError.d.ts +9 -16
- package/model-provider/llamacpp/LlamaCppError.js +7 -25
- package/model-provider/llamacpp/LlamaCppFacade.cjs +10 -2
- package/model-provider/llamacpp/LlamaCppFacade.d.ts +8 -1
- package/model-provider/llamacpp/LlamaCppFacade.js +8 -1
- package/model-provider/llamacpp/LlamaCppTextEmbeddingModel.cjs +10 -14
- package/model-provider/llamacpp/LlamaCppTextEmbeddingModel.js +10 -14
- package/model-provider/llamacpp/LlamaCppTokenizer.cjs +14 -18
- package/model-provider/llamacpp/LlamaCppTokenizer.js +14 -18
- package/model-provider/llamacpp/index.cjs +1 -3
- package/model-provider/llamacpp/index.d.ts +1 -1
- package/model-provider/llamacpp/index.js +0 -1
- package/model-provider/lmnt/LmntApiConfiguration.cjs +14 -6
- package/model-provider/lmnt/LmntApiConfiguration.d.ts +7 -8
- package/model-provider/lmnt/LmntApiConfiguration.js +15 -7
- package/model-provider/lmnt/LmntFacade.cjs +11 -2
- package/model-provider/lmnt/LmntFacade.d.ts +10 -1
- package/model-provider/lmnt/LmntFacade.js +9 -1
- package/model-provider/lmnt/LmntSpeechModel.cjs +53 -41
- package/model-provider/lmnt/LmntSpeechModel.d.ts +51 -3
- package/model-provider/lmnt/LmntSpeechModel.js +54 -42
- package/model-provider/mistral/MistralApiConfiguration.cjs +14 -6
- package/model-provider/mistral/MistralApiConfiguration.d.ts +9 -11
- package/model-provider/mistral/MistralApiConfiguration.js +15 -7
- package/model-provider/mistral/MistralChatModel.cjs +4 -4
- package/model-provider/mistral/MistralChatModel.d.ts +52 -7
- package/model-provider/mistral/MistralChatModel.js +5 -5
- package/model-provider/mistral/MistralError.cjs +3 -3
- package/model-provider/mistral/MistralError.d.ts +15 -3
- package/model-provider/mistral/MistralError.js +4 -4
- package/model-provider/mistral/MistralFacade.cjs +5 -1
- package/model-provider/mistral/MistralFacade.d.ts +10 -3
- package/model-provider/mistral/MistralFacade.js +6 -2
- package/model-provider/mistral/MistralTextEmbeddingModel.cjs +2 -1
- package/model-provider/mistral/MistralTextEmbeddingModel.d.ts +3 -3
- package/model-provider/mistral/MistralTextEmbeddingModel.js +2 -1
- package/model-provider/ollama/OllamaApiConfiguration.cjs +13 -6
- package/model-provider/ollama/OllamaApiConfiguration.d.ts +7 -10
- package/model-provider/ollama/OllamaApiConfiguration.js +14 -7
- package/model-provider/ollama/OllamaChatModel.cjs +6 -6
- package/model-provider/ollama/OllamaChatModel.d.ts +49 -8
- package/model-provider/ollama/OllamaChatModel.js +7 -7
- package/model-provider/ollama/OllamaCompletionModel.cjs +6 -6
- package/model-provider/ollama/OllamaCompletionModel.d.ts +43 -8
- package/model-provider/ollama/OllamaCompletionModel.js +7 -7
- package/model-provider/ollama/OllamaError.cjs +3 -3
- package/model-provider/ollama/OllamaError.d.ts +7 -3
- package/model-provider/ollama/OllamaError.js +4 -4
- package/model-provider/ollama/OllamaFacade.cjs +6 -2
- package/model-provider/ollama/OllamaFacade.d.ts +8 -3
- package/model-provider/ollama/OllamaFacade.js +6 -2
- package/model-provider/ollama/OllamaTextEmbeddingModel.cjs +15 -16
- package/model-provider/ollama/OllamaTextEmbeddingModel.js +15 -16
- package/model-provider/openai/AbstractOpenAIChatModel.cjs +5 -5
- package/model-provider/openai/AbstractOpenAIChatModel.d.ts +155 -9
- package/model-provider/openai/AbstractOpenAIChatModel.js +5 -5
- package/model-provider/openai/AbstractOpenAICompletionModel.cjs +30 -32
- package/model-provider/openai/AbstractOpenAICompletionModel.d.ts +4 -4
- package/model-provider/openai/AbstractOpenAICompletionModel.js +30 -32
- package/model-provider/openai/AzureOpenAIApiConfiguration.d.ts +9 -8
- package/model-provider/openai/OpenAIApiConfiguration.cjs +14 -6
- package/model-provider/openai/OpenAIApiConfiguration.d.ts +7 -8
- package/model-provider/openai/OpenAIApiConfiguration.js +15 -7
- package/model-provider/openai/OpenAIChatFunctionCallStructureGenerationModel.d.ts +1 -1
- package/model-provider/openai/OpenAIChatModel.d.ts +2 -2
- package/model-provider/openai/OpenAIError.cjs +8 -8
- package/model-provider/openai/OpenAIError.d.ts +27 -3
- package/model-provider/openai/OpenAIError.js +9 -9
- package/model-provider/openai/OpenAIFacade.cjs +23 -2
- package/model-provider/openai/OpenAIFacade.d.ts +20 -2
- package/model-provider/openai/OpenAIFacade.js +20 -1
- package/model-provider/openai/OpenAIImageGenerationModel.cjs +20 -21
- package/model-provider/openai/OpenAIImageGenerationModel.js +20 -21
- package/model-provider/openai/OpenAISpeechModel.cjs +17 -22
- package/model-provider/openai/OpenAISpeechModel.js +17 -22
- package/model-provider/openai/OpenAITextEmbeddingModel.cjs +18 -23
- package/model-provider/openai/OpenAITextEmbeddingModel.d.ts +3 -3
- package/model-provider/openai/OpenAITextEmbeddingModel.js +18 -23
- package/model-provider/openai/OpenAITranscriptionModel.cjs +42 -48
- package/model-provider/openai/OpenAITranscriptionModel.d.ts +14 -10
- package/model-provider/openai/OpenAITranscriptionModel.js +42 -48
- package/model-provider/openai-compatible/FireworksAIApiConfiguration.cjs +11 -7
- package/model-provider/openai-compatible/FireworksAIApiConfiguration.d.ts +4 -9
- package/model-provider/openai-compatible/FireworksAIApiConfiguration.js +12 -8
- package/model-provider/openai-compatible/OpenAICompatibleChatModel.d.ts +2 -2
- package/model-provider/openai-compatible/OpenAICompatibleFacade.cjs +25 -1
- package/model-provider/openai-compatible/OpenAICompatibleFacade.d.ts +23 -0
- package/model-provider/openai-compatible/OpenAICompatibleFacade.js +22 -0
- package/model-provider/openai-compatible/TogetherAIApiConfiguration.cjs +11 -7
- package/model-provider/openai-compatible/TogetherAIApiConfiguration.d.ts +4 -9
- package/model-provider/openai-compatible/TogetherAIApiConfiguration.js +12 -8
- package/model-provider/stability/StabilityApiConfiguration.cjs +13 -12
- package/model-provider/stability/StabilityApiConfiguration.d.ts +4 -4
- package/model-provider/stability/StabilityApiConfiguration.js +13 -12
- package/model-provider/stability/StabilityError.cjs +3 -3
- package/model-provider/stability/StabilityError.d.ts +7 -3
- package/model-provider/stability/StabilityError.js +4 -4
- package/model-provider/stability/StabilityFacade.cjs +9 -9
- package/model-provider/stability/StabilityFacade.d.ts +8 -8
- package/model-provider/stability/StabilityFacade.js +7 -7
- package/model-provider/stability/StabilityImageGenerationModel.cjs +2 -1
- package/model-provider/stability/StabilityImageGenerationModel.js +2 -1
- package/model-provider/whispercpp/WhisperCppApiConfiguration.cjs +13 -6
- package/model-provider/whispercpp/WhisperCppApiConfiguration.d.ts +7 -10
- package/model-provider/whispercpp/WhisperCppApiConfiguration.js +14 -7
- package/model-provider/whispercpp/WhisperCppFacade.cjs +9 -5
- package/model-provider/whispercpp/WhisperCppFacade.d.ts +7 -2
- package/model-provider/whispercpp/WhisperCppFacade.js +8 -4
- package/model-provider/whispercpp/WhisperCppTranscriptionModel.cjs +5 -2
- package/model-provider/whispercpp/WhisperCppTranscriptionModel.js +6 -3
- package/package.json +1 -1
- package/tool/Tool.cjs +0 -10
- package/tool/Tool.d.ts +0 -1
- package/tool/Tool.js +0 -10
- package/tool/WebSearchTool.cjs +2 -2
- package/tool/WebSearchTool.d.ts +1 -2
- package/tool/WebSearchTool.js +3 -3
- package/tool/{generate-tool-calls-or-text/TextGenerationToolCallsOrGenerateTextModel.cjs → generate-tool-calls/TextGenerationToolCallsModel.cjs} +7 -7
- package/tool/{generate-tool-calls-or-text/TextGenerationToolCallsOrGenerateTextModel.d.ts → generate-tool-calls/TextGenerationToolCallsModel.d.ts} +6 -6
- package/tool/{generate-tool-calls-or-text/TextGenerationToolCallsOrGenerateTextModel.js → generate-tool-calls/TextGenerationToolCallsModel.js} +5 -5
- package/tool/generate-tool-calls/ToolCallsGenerationEvent.d.ts +23 -0
- package/tool/{generate-tool-calls-or-text/ToolCallsOrTextGenerationModel.d.ts → generate-tool-calls/ToolCallsGenerationModel.d.ts} +3 -3
- package/tool/{generate-tool-calls-or-text/ToolCallsOrTextParseError.cjs → generate-tool-calls/ToolCallsParseError.cjs} +5 -5
- package/tool/{generate-tool-calls-or-text/ToolCallsOrTextParseError.d.ts → generate-tool-calls/ToolCallsParseError.d.ts} +1 -1
- package/tool/{generate-tool-calls-or-text/ToolCallsOrTextParseError.js → generate-tool-calls/ToolCallsParseError.js} +3 -3
- package/tool/{generate-tool-calls-or-text/ToolCallsOrGenerateTextPromptTemplate.d.ts → generate-tool-calls/ToolCallsPromptTemplate.d.ts} +1 -1
- package/tool/{generate-tool-calls-or-text/generateToolCallsOrText.cjs → generate-tool-calls/generateToolCalls.cjs} +5 -5
- package/tool/{generate-tool-calls-or-text/generateToolCallsOrText.d.ts → generate-tool-calls/generateToolCalls.d.ts} +3 -3
- package/tool/{generate-tool-calls-or-text/generateToolCallsOrText.js → generate-tool-calls/generateToolCalls.js} +3 -3
- package/tool/{generate-tool-calls-or-text → generate-tool-calls}/index.cjs +6 -6
- package/tool/generate-tool-calls/index.d.ts +6 -0
- package/tool/generate-tool-calls/index.js +6 -0
- package/tool/index.cjs +2 -3
- package/tool/index.d.ts +2 -3
- package/tool/index.js +2 -3
- package/tool/use-tools/UseToolsEvent.d.ts +7 -0
- package/tool/use-tools/UseToolsEvent.js +1 -0
- package/tool/{use-tools-or-generate-text → use-tools}/index.cjs +2 -2
- package/tool/use-tools/index.d.ts +2 -0
- package/tool/use-tools/index.js +2 -0
- package/tool/{use-tools-or-generate-text/useToolsOrGenerateText.cjs → use-tools/useTools.cjs} +6 -6
- package/tool/{use-tools-or-generate-text/useToolsOrGenerateText.d.ts → use-tools/useTools.d.ts} +2 -2
- package/tool/{use-tools-or-generate-text/useToolsOrGenerateText.js → use-tools/useTools.js} +4 -4
- package/vector-index/memory/MemoryVectorIndex.cjs +2 -2
- package/vector-index/memory/MemoryVectorIndex.js +3 -3
- package/core/api/BaseUrlPartsApiConfiguration.cjs +0 -53
- package/core/api/BaseUrlPartsApiConfiguration.d.ts +0 -26
- package/core/api/BaseUrlPartsApiConfiguration.js +0 -49
- package/core/api/throttleUnlimitedConcurrency.cjs +0 -8
- package/core/api/throttleUnlimitedConcurrency.d.ts +0 -5
- package/core/api/throttleUnlimitedConcurrency.js +0 -4
- package/model-provider/elevenlabs/ElevenLabsError.cjs +0 -30
- package/model-provider/elevenlabs/ElevenLabsError.d.ts +0 -3
- package/model-provider/elevenlabs/ElevenLabsError.js +0 -26
- package/model-provider/lmnt/LmntError.cjs +0 -30
- package/model-provider/lmnt/LmntError.d.ts +0 -3
- package/model-provider/lmnt/LmntError.js +0 -26
- package/tool/InvalidToolNameError.cjs +0 -17
- package/tool/InvalidToolNameError.d.ts +0 -7
- package/tool/InvalidToolNameError.js +0 -13
- package/tool/generate-tool-calls-or-text/ToolCallsOrTextGenerationEvent.d.ts +0 -23
- package/tool/generate-tool-calls-or-text/index.d.ts +0 -6
- package/tool/generate-tool-calls-or-text/index.js +0 -6
- package/tool/use-tools-or-generate-text/UseToolsOrGenerateTextEvent.d.ts +0 -7
- package/tool/use-tools-or-generate-text/index.d.ts +0 -2
- package/tool/use-tools-or-generate-text/index.js +0 -2
- /package/{tool/generate-tool-calls-or-text/ToolCallsOrGenerateTextPromptTemplate.js → core/api/BaseUrlApiConfiguration.test.d.ts} +0 -0
- /package/tool/{generate-tool-calls-or-text/ToolCallsOrGenerateTextPromptTemplate.cjs → generate-tool-calls/ToolCallsGenerationEvent.cjs} +0 -0
- /package/tool/{generate-tool-calls-or-text/ToolCallsOrTextGenerationEvent.js → generate-tool-calls/ToolCallsGenerationEvent.js} +0 -0
- /package/tool/{generate-tool-calls-or-text/ToolCallsOrTextGenerationEvent.cjs → generate-tool-calls/ToolCallsGenerationModel.cjs} +0 -0
- /package/tool/{generate-tool-calls-or-text/ToolCallsOrTextGenerationModel.js → generate-tool-calls/ToolCallsGenerationModel.js} +0 -0
- /package/tool/{generate-tool-calls-or-text/ToolCallsOrTextGenerationModel.cjs → generate-tool-calls/ToolCallsPromptTemplate.cjs} +0 -0
- /package/tool/{use-tools-or-generate-text/UseToolsOrGenerateTextEvent.js → generate-tool-calls/ToolCallsPromptTemplate.js} +0 -0
- /package/tool/{use-tools-or-generate-text/UseToolsOrGenerateTextEvent.cjs → use-tools/UseToolsEvent.cjs} +0 -0
@@ -26,10 +26,8 @@ var __importStar = (this && this.__importStar) || function (mod) {
|
|
26
26
|
return result;
|
27
27
|
};
|
28
28
|
Object.defineProperty(exports, "__esModule", { value: true });
|
29
|
-
exports.huggingface =
|
29
|
+
exports.huggingface = void 0;
|
30
30
|
__exportStar(require("./HuggingFaceApiConfiguration.cjs"), exports);
|
31
|
-
var HuggingFaceError_js_1 = require("./HuggingFaceError.cjs");
|
32
|
-
Object.defineProperty(exports, "HuggingFaceError", { enumerable: true, get: function () { return HuggingFaceError_js_1.HuggingFaceError; } });
|
33
31
|
exports.huggingface = __importStar(require("./HuggingFaceFacade.cjs"));
|
34
32
|
__exportStar(require("./HuggingFaceTextEmbeddingModel.cjs"), exports);
|
35
33
|
__exportStar(require("./HuggingFaceTextGenerationModel.cjs"), exports);
|
@@ -1,5 +1,5 @@
|
|
1
1
|
export * from "./HuggingFaceApiConfiguration.js";
|
2
|
-
export {
|
2
|
+
export { HuggingFaceErrorData } from "./HuggingFaceError.js";
|
3
3
|
export * as huggingface from "./HuggingFaceFacade.js";
|
4
4
|
export * from "./HuggingFaceTextEmbeddingModel.js";
|
5
5
|
export * from "./HuggingFaceTextGenerationModel.js";
|
@@ -1,5 +1,4 @@
|
|
1
1
|
export * from "./HuggingFaceApiConfiguration.js";
|
2
|
-
export { HuggingFaceError } from "./HuggingFaceError.js";
|
3
2
|
export * as huggingface from "./HuggingFaceFacade.js";
|
4
3
|
export * from "./HuggingFaceTextEmbeddingModel.js";
|
5
4
|
export * from "./HuggingFaceTextGenerationModel.js";
|
@@ -2,13 +2,20 @@
|
|
2
2
|
Object.defineProperty(exports, "__esModule", { value: true });
|
3
3
|
exports.LlamaCppApiConfiguration = void 0;
|
4
4
|
const BaseUrlApiConfiguration_js_1 = require("../../core/api/BaseUrlApiConfiguration.cjs");
|
5
|
-
|
6
|
-
|
5
|
+
/**
|
6
|
+
* Creates an API configuration for the Llama.cpp server.
|
7
|
+
* It calls the API at http://127.0.0.1:8080 by default.
|
8
|
+
*/
|
9
|
+
class LlamaCppApiConfiguration extends BaseUrlApiConfiguration_js_1.BaseUrlApiConfigurationWithDefaults {
|
10
|
+
constructor(settings = {}) {
|
7
11
|
super({
|
8
|
-
|
9
|
-
|
10
|
-
|
11
|
-
|
12
|
+
...settings,
|
13
|
+
baseUrlDefaults: {
|
14
|
+
protocol: "http",
|
15
|
+
host: "127.0.0.1",
|
16
|
+
port: "8080",
|
17
|
+
path: "",
|
18
|
+
},
|
12
19
|
});
|
13
20
|
}
|
14
21
|
}
|
@@ -1,10 +1,8 @@
|
|
1
|
-
import {
|
2
|
-
|
3
|
-
|
4
|
-
|
5
|
-
|
6
|
-
|
7
|
-
|
8
|
-
throttle?: ThrottleFunction;
|
9
|
-
});
|
1
|
+
import { BaseUrlApiConfigurationWithDefaults, PartialBaseUrlPartsApiConfigurationOptions } from "../../core/api/BaseUrlApiConfiguration.js";
|
2
|
+
/**
|
3
|
+
* Creates an API configuration for the Llama.cpp server.
|
4
|
+
* It calls the API at http://127.0.0.1:8080 by default.
|
5
|
+
*/
|
6
|
+
export declare class LlamaCppApiConfiguration extends BaseUrlApiConfigurationWithDefaults {
|
7
|
+
constructor(settings?: PartialBaseUrlPartsApiConfigurationOptions);
|
10
8
|
}
|
@@ -1,11 +1,18 @@
|
|
1
|
-
import {
|
2
|
-
|
3
|
-
|
1
|
+
import { BaseUrlApiConfigurationWithDefaults, } from "../../core/api/BaseUrlApiConfiguration.js";
|
2
|
+
/**
|
3
|
+
* Creates an API configuration for the Llama.cpp server.
|
4
|
+
* It calls the API at http://127.0.0.1:8080 by default.
|
5
|
+
*/
|
6
|
+
export class LlamaCppApiConfiguration extends BaseUrlApiConfigurationWithDefaults {
|
7
|
+
constructor(settings = {}) {
|
4
8
|
super({
|
5
|
-
|
6
|
-
|
7
|
-
|
8
|
-
|
9
|
+
...settings,
|
10
|
+
baseUrlDefaults: {
|
11
|
+
protocol: "http",
|
12
|
+
host: "127.0.0.1",
|
13
|
+
port: "8080",
|
14
|
+
path: "",
|
15
|
+
},
|
9
16
|
});
|
10
17
|
}
|
11
18
|
}
|
@@ -245,13 +245,13 @@ const llamaCppTextGenerationResponseSchema = zod_1.z.object({
|
|
245
245
|
tokens_predicted: zod_1.z.number(),
|
246
246
|
truncated: zod_1.z.boolean(),
|
247
247
|
});
|
248
|
-
const llamaCppTextStreamChunkSchema =
|
248
|
+
const llamaCppTextStreamChunkSchema = zod_1.z.discriminatedUnion("stop", [
|
249
249
|
zod_1.z.object({
|
250
250
|
content: zod_1.z.string(),
|
251
251
|
stop: zod_1.z.literal(false),
|
252
252
|
}),
|
253
253
|
llamaCppTextGenerationResponseSchema,
|
254
|
-
])
|
254
|
+
]);
|
255
255
|
async function createLlamaCppFullDeltaIterableQueue(stream) {
|
256
256
|
const queue = new AsyncQueue_js_1.AsyncQueue();
|
257
257
|
// process the stream asynchonously (no 'await' on purpose):
|
@@ -262,7 +262,7 @@ async function createLlamaCppFullDeltaIterableQueue(stream) {
|
|
262
262
|
const data = event.data;
|
263
263
|
const eventData = (0, parseJSON_js_1.parseJSON)({
|
264
264
|
text: data,
|
265
|
-
schema: llamaCppTextStreamChunkSchema,
|
265
|
+
schema: (0, ZodSchema_js_1.zodSchema)(llamaCppTextStreamChunkSchema),
|
266
266
|
});
|
267
267
|
queue.push({ type: "delta", deltaValue: eventData });
|
268
268
|
if (eventData.stop) {
|
@@ -287,7 +287,7 @@ exports.LlamaCppCompletionResponseFormat = {
|
|
287
287
|
*/
|
288
288
|
json: {
|
289
289
|
stream: false,
|
290
|
-
handler: (0, postToApi_js_1.createJsonResponseHandler)(llamaCppTextGenerationResponseSchema),
|
290
|
+
handler: (0, postToApi_js_1.createJsonResponseHandler)((0, ZodSchema_js_1.zodSchema)(llamaCppTextGenerationResponseSchema)),
|
291
291
|
},
|
292
292
|
/**
|
293
293
|
* Returns an async iterable over the full deltas (all choices, including full current state at time of event)
|
@@ -479,7 +479,130 @@ declare const llamaCppTextGenerationResponseSchema: z.ZodObject<{
|
|
479
479
|
truncated: boolean;
|
480
480
|
}>;
|
481
481
|
export type LlamaCppTextGenerationResponse = z.infer<typeof llamaCppTextGenerationResponseSchema>;
|
482
|
-
declare const llamaCppTextStreamChunkSchema:
|
482
|
+
declare const llamaCppTextStreamChunkSchema: z.ZodDiscriminatedUnion<"stop", [z.ZodObject<{
|
483
|
+
content: z.ZodString;
|
484
|
+
stop: z.ZodLiteral<false>;
|
485
|
+
}, "strip", z.ZodTypeAny, {
|
486
|
+
stop: false;
|
487
|
+
content: string;
|
488
|
+
}, {
|
489
|
+
stop: false;
|
490
|
+
content: string;
|
491
|
+
}>, z.ZodObject<{
|
492
|
+
content: z.ZodString;
|
493
|
+
stop: z.ZodLiteral<true>;
|
494
|
+
generation_settings: z.ZodObject<{
|
495
|
+
frequency_penalty: z.ZodNumber;
|
496
|
+
ignore_eos: z.ZodBoolean;
|
497
|
+
logit_bias: z.ZodArray<z.ZodNumber, "many">;
|
498
|
+
mirostat: z.ZodNumber;
|
499
|
+
mirostat_eta: z.ZodNumber;
|
500
|
+
mirostat_tau: z.ZodNumber;
|
501
|
+
model: z.ZodString;
|
502
|
+
n_ctx: z.ZodNumber;
|
503
|
+
n_keep: z.ZodNumber;
|
504
|
+
n_predict: z.ZodNumber;
|
505
|
+
n_probs: z.ZodNumber;
|
506
|
+
penalize_nl: z.ZodBoolean;
|
507
|
+
presence_penalty: z.ZodNumber;
|
508
|
+
repeat_last_n: z.ZodNumber;
|
509
|
+
repeat_penalty: z.ZodNumber;
|
510
|
+
seed: z.ZodNumber;
|
511
|
+
stop: z.ZodArray<z.ZodString, "many">;
|
512
|
+
stream: z.ZodBoolean;
|
513
|
+
temperature: z.ZodOptional<z.ZodNumber>;
|
514
|
+
tfs_z: z.ZodNumber;
|
515
|
+
top_k: z.ZodNumber;
|
516
|
+
top_p: z.ZodNumber;
|
517
|
+
typical_p: z.ZodNumber;
|
518
|
+
}, "strip", z.ZodTypeAny, {
|
519
|
+
model: string;
|
520
|
+
stream: boolean;
|
521
|
+
stop: string[];
|
522
|
+
seed: number;
|
523
|
+
mirostat: number;
|
524
|
+
frequency_penalty: number;
|
525
|
+
ignore_eos: boolean;
|
526
|
+
logit_bias: number[];
|
527
|
+
mirostat_eta: number;
|
528
|
+
mirostat_tau: number;
|
529
|
+
n_ctx: number;
|
530
|
+
n_keep: number;
|
531
|
+
n_predict: number;
|
532
|
+
n_probs: number;
|
533
|
+
penalize_nl: boolean;
|
534
|
+
presence_penalty: number;
|
535
|
+
repeat_last_n: number;
|
536
|
+
repeat_penalty: number;
|
537
|
+
tfs_z: number;
|
538
|
+
top_k: number;
|
539
|
+
top_p: number;
|
540
|
+
typical_p: number;
|
541
|
+
temperature?: number | undefined;
|
542
|
+
}, {
|
543
|
+
model: string;
|
544
|
+
stream: boolean;
|
545
|
+
stop: string[];
|
546
|
+
seed: number;
|
547
|
+
mirostat: number;
|
548
|
+
frequency_penalty: number;
|
549
|
+
ignore_eos: boolean;
|
550
|
+
logit_bias: number[];
|
551
|
+
mirostat_eta: number;
|
552
|
+
mirostat_tau: number;
|
553
|
+
n_ctx: number;
|
554
|
+
n_keep: number;
|
555
|
+
n_predict: number;
|
556
|
+
n_probs: number;
|
557
|
+
penalize_nl: boolean;
|
558
|
+
presence_penalty: number;
|
559
|
+
repeat_last_n: number;
|
560
|
+
repeat_penalty: number;
|
561
|
+
tfs_z: number;
|
562
|
+
top_k: number;
|
563
|
+
top_p: number;
|
564
|
+
typical_p: number;
|
565
|
+
temperature?: number | undefined;
|
566
|
+
}>;
|
567
|
+
model: z.ZodString;
|
568
|
+
prompt: z.ZodString;
|
569
|
+
stopped_eos: z.ZodBoolean;
|
570
|
+
stopped_limit: z.ZodBoolean;
|
571
|
+
stopped_word: z.ZodBoolean;
|
572
|
+
stopping_word: z.ZodString;
|
573
|
+
timings: z.ZodObject<{
|
574
|
+
predicted_ms: z.ZodNumber;
|
575
|
+
predicted_n: z.ZodNumber;
|
576
|
+
predicted_per_second: z.ZodNullable<z.ZodNumber>;
|
577
|
+
predicted_per_token_ms: z.ZodNullable<z.ZodNumber>;
|
578
|
+
prompt_ms: z.ZodOptional<z.ZodNullable<z.ZodNumber>>;
|
579
|
+
prompt_n: z.ZodNumber;
|
580
|
+
prompt_per_second: z.ZodNullable<z.ZodNumber>;
|
581
|
+
prompt_per_token_ms: z.ZodNullable<z.ZodNumber>;
|
582
|
+
}, "strip", z.ZodTypeAny, {
|
583
|
+
predicted_ms: number;
|
584
|
+
predicted_n: number;
|
585
|
+
predicted_per_second: number | null;
|
586
|
+
predicted_per_token_ms: number | null;
|
587
|
+
prompt_n: number;
|
588
|
+
prompt_per_second: number | null;
|
589
|
+
prompt_per_token_ms: number | null;
|
590
|
+
prompt_ms?: number | null | undefined;
|
591
|
+
}, {
|
592
|
+
predicted_ms: number;
|
593
|
+
predicted_n: number;
|
594
|
+
predicted_per_second: number | null;
|
595
|
+
predicted_per_token_ms: number | null;
|
596
|
+
prompt_n: number;
|
597
|
+
prompt_per_second: number | null;
|
598
|
+
prompt_per_token_ms: number | null;
|
599
|
+
prompt_ms?: number | null | undefined;
|
600
|
+
}>;
|
601
|
+
tokens_cached: z.ZodNumber;
|
602
|
+
tokens_evaluated: z.ZodNumber;
|
603
|
+
tokens_predicted: z.ZodNumber;
|
604
|
+
truncated: z.ZodBoolean;
|
605
|
+
}, "strip", z.ZodTypeAny, {
|
483
606
|
model: string;
|
484
607
|
stop: true;
|
485
608
|
content: string;
|
@@ -527,11 +650,56 @@ declare const llamaCppTextStreamChunkSchema: import("../../core/schema/ZodSchema
|
|
527
650
|
tokens_evaluated: number;
|
528
651
|
tokens_predicted: number;
|
529
652
|
truncated: boolean;
|
530
|
-
}
|
531
|
-
|
653
|
+
}, {
|
654
|
+
model: string;
|
655
|
+
stop: true;
|
532
656
|
content: string;
|
533
|
-
|
534
|
-
|
657
|
+
prompt: string;
|
658
|
+
generation_settings: {
|
659
|
+
model: string;
|
660
|
+
stream: boolean;
|
661
|
+
stop: string[];
|
662
|
+
seed: number;
|
663
|
+
mirostat: number;
|
664
|
+
frequency_penalty: number;
|
665
|
+
ignore_eos: boolean;
|
666
|
+
logit_bias: number[];
|
667
|
+
mirostat_eta: number;
|
668
|
+
mirostat_tau: number;
|
669
|
+
n_ctx: number;
|
670
|
+
n_keep: number;
|
671
|
+
n_predict: number;
|
672
|
+
n_probs: number;
|
673
|
+
penalize_nl: boolean;
|
674
|
+
presence_penalty: number;
|
675
|
+
repeat_last_n: number;
|
676
|
+
repeat_penalty: number;
|
677
|
+
tfs_z: number;
|
678
|
+
top_k: number;
|
679
|
+
top_p: number;
|
680
|
+
typical_p: number;
|
681
|
+
temperature?: number | undefined;
|
682
|
+
};
|
683
|
+
stopped_eos: boolean;
|
684
|
+
stopped_limit: boolean;
|
685
|
+
stopped_word: boolean;
|
686
|
+
stopping_word: string;
|
687
|
+
timings: {
|
688
|
+
predicted_ms: number;
|
689
|
+
predicted_n: number;
|
690
|
+
predicted_per_second: number | null;
|
691
|
+
predicted_per_token_ms: number | null;
|
692
|
+
prompt_n: number;
|
693
|
+
prompt_per_second: number | null;
|
694
|
+
prompt_per_token_ms: number | null;
|
695
|
+
prompt_ms?: number | null | undefined;
|
696
|
+
};
|
697
|
+
tokens_cached: number;
|
698
|
+
tokens_evaluated: number;
|
699
|
+
tokens_predicted: number;
|
700
|
+
truncated: boolean;
|
701
|
+
}>]>;
|
702
|
+
export type LlamaCppTextStreamChunk = z.infer<typeof llamaCppTextStreamChunkSchema>;
|
535
703
|
export type LlamaCppCompletionResponseFormatType<T> = {
|
536
704
|
stream: boolean;
|
537
705
|
handler: ResponseHandler<T>;
|
@@ -241,13 +241,13 @@ const llamaCppTextGenerationResponseSchema = z.object({
|
|
241
241
|
tokens_predicted: z.number(),
|
242
242
|
truncated: z.boolean(),
|
243
243
|
});
|
244
|
-
const llamaCppTextStreamChunkSchema =
|
244
|
+
const llamaCppTextStreamChunkSchema = z.discriminatedUnion("stop", [
|
245
245
|
z.object({
|
246
246
|
content: z.string(),
|
247
247
|
stop: z.literal(false),
|
248
248
|
}),
|
249
249
|
llamaCppTextGenerationResponseSchema,
|
250
|
-
])
|
250
|
+
]);
|
251
251
|
async function createLlamaCppFullDeltaIterableQueue(stream) {
|
252
252
|
const queue = new AsyncQueue();
|
253
253
|
// process the stream asynchonously (no 'await' on purpose):
|
@@ -258,7 +258,7 @@ async function createLlamaCppFullDeltaIterableQueue(stream) {
|
|
258
258
|
const data = event.data;
|
259
259
|
const eventData = parseJSON({
|
260
260
|
text: data,
|
261
|
-
schema: llamaCppTextStreamChunkSchema,
|
261
|
+
schema: zodSchema(llamaCppTextStreamChunkSchema),
|
262
262
|
});
|
263
263
|
queue.push({ type: "delta", deltaValue: eventData });
|
264
264
|
if (eventData.stop) {
|
@@ -283,7 +283,7 @@ export const LlamaCppCompletionResponseFormat = {
|
|
283
283
|
*/
|
284
284
|
json: {
|
285
285
|
stream: false,
|
286
|
-
handler: createJsonResponseHandler(llamaCppTextGenerationResponseSchema),
|
286
|
+
handler: createJsonResponseHandler(zodSchema(llamaCppTextGenerationResponseSchema)),
|
287
287
|
},
|
288
288
|
/**
|
289
289
|
* Returns an async iterable over the full deltas (all choices, including full current state at time of event)
|
@@ -1,33 +1,13 @@
|
|
1
1
|
"use strict";
|
2
2
|
Object.defineProperty(exports, "__esModule", { value: true });
|
3
|
-
exports.failedLlamaCppCallResponseHandler =
|
3
|
+
exports.failedLlamaCppCallResponseHandler = void 0;
|
4
4
|
const zod_1 = require("zod");
|
5
|
-
const
|
6
|
-
const parseJSON_js_1 = require("../../core/schema/parseJSON.cjs");
|
5
|
+
const postToApi_js_1 = require("../../core/api/postToApi.cjs");
|
7
6
|
const ZodSchema_js_1 = require("../../core/schema/ZodSchema.cjs");
|
8
|
-
|
7
|
+
const llamaCppErrorDataSchema = zod_1.z.object({
|
9
8
|
error: zod_1.z.string(),
|
10
|
-
}));
|
11
|
-
class LlamaCppError extends ApiCallError_js_1.ApiCallError {
|
12
|
-
constructor({ data, statusCode, url, requestBodyValues, message = data.error, }) {
|
13
|
-
super({ message, statusCode, requestBodyValues, url });
|
14
|
-
Object.defineProperty(this, "data", {
|
15
|
-
enumerable: true,
|
16
|
-
configurable: true,
|
17
|
-
writable: true,
|
18
|
-
value: void 0
|
19
|
-
});
|
20
|
-
this.data = data;
|
21
|
-
}
|
22
|
-
}
|
23
|
-
exports.LlamaCppError = LlamaCppError;
|
24
|
-
const failedLlamaCppCallResponseHandler = async ({ response, url, requestBodyValues }) => new LlamaCppError({
|
25
|
-
url,
|
26
|
-
requestBodyValues,
|
27
|
-
statusCode: response.status,
|
28
|
-
data: (0, parseJSON_js_1.parseJSON)({
|
29
|
-
text: await response.text(),
|
30
|
-
schema: exports.llamaCppErrorDataSchema,
|
31
|
-
}),
|
32
9
|
});
|
33
|
-
exports.failedLlamaCppCallResponseHandler =
|
10
|
+
exports.failedLlamaCppCallResponseHandler = (0, postToApi_js_1.createJsonErrorResponseHandler)({
|
11
|
+
errorSchema: (0, ZodSchema_js_1.zodSchema)(llamaCppErrorDataSchema),
|
12
|
+
errorToMessage: (error) => error.error,
|
13
|
+
});
|
@@ -1,18 +1,11 @@
|
|
1
|
-
import {
|
2
|
-
|
3
|
-
|
4
|
-
|
1
|
+
import { z } from "zod";
|
2
|
+
declare const llamaCppErrorDataSchema: z.ZodObject<{
|
3
|
+
error: z.ZodString;
|
4
|
+
}, "strip", z.ZodTypeAny, {
|
5
|
+
error: string;
|
6
|
+
}, {
|
5
7
|
error: string;
|
6
8
|
}>;
|
7
|
-
export type LlamaCppErrorData =
|
8
|
-
export declare
|
9
|
-
|
10
|
-
constructor({ data, statusCode, url, requestBodyValues, message, }: {
|
11
|
-
message?: string;
|
12
|
-
statusCode: number;
|
13
|
-
url: string;
|
14
|
-
requestBodyValues: unknown;
|
15
|
-
data: LlamaCppErrorData;
|
16
|
-
});
|
17
|
-
}
|
18
|
-
export declare const failedLlamaCppCallResponseHandler: ResponseHandler<ApiCallError>;
|
9
|
+
export type LlamaCppErrorData = z.infer<typeof llamaCppErrorDataSchema>;
|
10
|
+
export declare const failedLlamaCppCallResponseHandler: import("../../core/api/postToApi.js").ResponseHandler<import("../../index.js").ApiCallError>;
|
11
|
+
export {};
|
@@ -1,28 +1,10 @@
|
|
1
1
|
import { z } from "zod";
|
2
|
-
import {
|
3
|
-
import {
|
4
|
-
|
5
|
-
export const llamaCppErrorDataSchema = new ZodSchema(z.object({
|
2
|
+
import { createJsonErrorResponseHandler } from "../../core/api/postToApi.js";
|
3
|
+
import { zodSchema } from "../../core/schema/ZodSchema.js";
|
4
|
+
const llamaCppErrorDataSchema = z.object({
|
6
5
|
error: z.string(),
|
7
|
-
})
|
8
|
-
export
|
9
|
-
|
10
|
-
|
11
|
-
Object.defineProperty(this, "data", {
|
12
|
-
enumerable: true,
|
13
|
-
configurable: true,
|
14
|
-
writable: true,
|
15
|
-
value: void 0
|
16
|
-
});
|
17
|
-
this.data = data;
|
18
|
-
}
|
19
|
-
}
|
20
|
-
export const failedLlamaCppCallResponseHandler = async ({ response, url, requestBodyValues }) => new LlamaCppError({
|
21
|
-
url,
|
22
|
-
requestBodyValues,
|
23
|
-
statusCode: response.status,
|
24
|
-
data: parseJSON({
|
25
|
-
text: await response.text(),
|
26
|
-
schema: llamaCppErrorDataSchema,
|
27
|
-
}),
|
6
|
+
});
|
7
|
+
export const failedLlamaCppCallResponseHandler = createJsonErrorResponseHandler({
|
8
|
+
errorSchema: zodSchema(llamaCppErrorDataSchema),
|
9
|
+
errorToMessage: (error) => error.error,
|
28
10
|
});
|
@@ -1,10 +1,18 @@
|
|
1
1
|
"use strict";
|
2
2
|
Object.defineProperty(exports, "__esModule", { value: true });
|
3
|
-
exports.Tokenizer = exports.TextEmbedder = exports.TextGenerator = void 0;
|
3
|
+
exports.Tokenizer = exports.TextEmbedder = exports.TextGenerator = exports.Api = void 0;
|
4
4
|
const LlamaCppApiConfiguration_js_1 = require("./LlamaCppApiConfiguration.cjs");
|
5
|
-
const LlamaCppTextEmbeddingModel_js_1 = require("./LlamaCppTextEmbeddingModel.cjs");
|
6
5
|
const LlamaCppCompletionModel_js_1 = require("./LlamaCppCompletionModel.cjs");
|
6
|
+
const LlamaCppTextEmbeddingModel_js_1 = require("./LlamaCppTextEmbeddingModel.cjs");
|
7
7
|
const LlamaCppTokenizer_js_1 = require("./LlamaCppTokenizer.cjs");
|
8
|
+
/**
|
9
|
+
* Creates an API configuration for the Llama.cpp server.
|
10
|
+
* It calls the API at http://127.0.0.1:8080 by default.
|
11
|
+
*/
|
12
|
+
function Api(settings) {
|
13
|
+
return new LlamaCppApiConfiguration_js_1.LlamaCppApiConfiguration(settings);
|
14
|
+
}
|
15
|
+
exports.Api = Api;
|
8
16
|
function TextGenerator(settings = {}) {
|
9
17
|
return new LlamaCppCompletionModel_js_1.LlamaCppCompletionModel(settings);
|
10
18
|
}
|
@@ -1,7 +1,14 @@
|
|
1
1
|
import { ApiConfiguration } from "../../core/api/ApiConfiguration.js";
|
2
|
-
import {
|
2
|
+
import { PartialBaseUrlPartsApiConfigurationOptions } from "../../core/api/BaseUrlApiConfiguration.js";
|
3
|
+
import { LlamaCppApiConfiguration } from "./LlamaCppApiConfiguration.js";
|
3
4
|
import { LlamaCppCompletionModel, LlamaCppCompletionModelSettings } from "./LlamaCppCompletionModel.js";
|
5
|
+
import { LlamaCppTextEmbeddingModel, LlamaCppTextEmbeddingModelSettings } from "./LlamaCppTextEmbeddingModel.js";
|
4
6
|
import { LlamaCppTokenizer } from "./LlamaCppTokenizer.js";
|
7
|
+
/**
|
8
|
+
* Creates an API configuration for the Llama.cpp server.
|
9
|
+
* It calls the API at http://127.0.0.1:8080 by default.
|
10
|
+
*/
|
11
|
+
export declare function Api(settings: PartialBaseUrlPartsApiConfigurationOptions): LlamaCppApiConfiguration;
|
5
12
|
export declare function TextGenerator<CONTEXT_WINDOW_SIZE extends number>(settings?: LlamaCppCompletionModelSettings<CONTEXT_WINDOW_SIZE>): LlamaCppCompletionModel<CONTEXT_WINDOW_SIZE>;
|
6
13
|
export declare function TextEmbedder(settings?: LlamaCppTextEmbeddingModelSettings): LlamaCppTextEmbeddingModel;
|
7
14
|
export declare function Tokenizer(api?: ApiConfiguration): LlamaCppTokenizer;
|
@@ -1,7 +1,14 @@
|
|
1
1
|
import { LlamaCppApiConfiguration } from "./LlamaCppApiConfiguration.js";
|
2
|
-
import { LlamaCppTextEmbeddingModel, } from "./LlamaCppTextEmbeddingModel.js";
|
3
2
|
import { LlamaCppCompletionModel, } from "./LlamaCppCompletionModel.js";
|
3
|
+
import { LlamaCppTextEmbeddingModel, } from "./LlamaCppTextEmbeddingModel.js";
|
4
4
|
import { LlamaCppTokenizer } from "./LlamaCppTokenizer.js";
|
5
|
+
/**
|
6
|
+
* Creates an API configuration for the Llama.cpp server.
|
7
|
+
* It calls the API at http://127.0.0.1:8080 by default.
|
8
|
+
*/
|
9
|
+
export function Api(settings) {
|
10
|
+
return new LlamaCppApiConfiguration(settings);
|
11
|
+
}
|
5
12
|
export function TextGenerator(settings = {}) {
|
6
13
|
return new LlamaCppCompletionModel(settings);
|
7
14
|
}
|
@@ -4,6 +4,7 @@ exports.LlamaCppTextEmbeddingModel = void 0;
|
|
4
4
|
const zod_1 = require("zod");
|
5
5
|
const callWithRetryAndThrottle_js_1 = require("../../core/api/callWithRetryAndThrottle.cjs");
|
6
6
|
const postToApi_js_1 = require("../../core/api/postToApi.cjs");
|
7
|
+
const ZodSchema_js_1 = require("../../core/schema/ZodSchema.cjs");
|
7
8
|
const AbstractModel_js_1 = require("../../model-function/AbstractModel.cjs");
|
8
9
|
const LlamaCppApiConfiguration_js_1 = require("./LlamaCppApiConfiguration.cjs");
|
9
10
|
const LlamaCppError_js_1 = require("./LlamaCppError.cjs");
|
@@ -53,13 +54,18 @@ class LlamaCppTextEmbeddingModel extends AbstractModel_js_1.AbstractModel {
|
|
53
54
|
if (texts.length > this.maxValuesPerCall) {
|
54
55
|
throw new Error(`The Llama.cpp embedding API only supports ${this.maxValuesPerCall} texts per API call.`);
|
55
56
|
}
|
57
|
+
const api = this.settings.api ?? new LlamaCppApiConfiguration_js_1.LlamaCppApiConfiguration();
|
58
|
+
const abortSignal = options?.run?.abortSignal;
|
56
59
|
return (0, callWithRetryAndThrottle_js_1.callWithRetryAndThrottle)({
|
57
60
|
retry: this.settings.api?.retry,
|
58
61
|
throttle: this.settings.api?.throttle,
|
59
|
-
call: async () =>
|
60
|
-
|
61
|
-
|
62
|
-
content: texts[0],
|
62
|
+
call: async () => (0, postToApi_js_1.postJsonToApi)({
|
63
|
+
url: api.assembleUrl(`/embedding`),
|
64
|
+
headers: api.headers,
|
65
|
+
body: { content: texts[0] },
|
66
|
+
failedResponseHandler: LlamaCppError_js_1.failedLlamaCppCallResponseHandler,
|
67
|
+
successfulResponseHandler: (0, postToApi_js_1.createJsonResponseHandler)((0, ZodSchema_js_1.zodSchema)(llamaCppTextEmbeddingResponseSchema)),
|
68
|
+
abortSignal,
|
63
69
|
}),
|
64
70
|
});
|
65
71
|
}
|
@@ -83,13 +89,3 @@ exports.LlamaCppTextEmbeddingModel = LlamaCppTextEmbeddingModel;
|
|
83
89
|
const llamaCppTextEmbeddingResponseSchema = zod_1.z.object({
|
84
90
|
embedding: zod_1.z.array(zod_1.z.number()),
|
85
91
|
});
|
86
|
-
async function callLlamaCppEmbeddingAPI({ api = new LlamaCppApiConfiguration_js_1.LlamaCppApiConfiguration(), abortSignal, content, }) {
|
87
|
-
return (0, postToApi_js_1.postJsonToApi)({
|
88
|
-
url: api.assembleUrl(`/embedding`),
|
89
|
-
headers: api.headers,
|
90
|
-
body: { content },
|
91
|
-
failedResponseHandler: LlamaCppError_js_1.failedLlamaCppCallResponseHandler,
|
92
|
-
successfulResponseHandler: (0, postToApi_js_1.createJsonResponseHandler)(llamaCppTextEmbeddingResponseSchema),
|
93
|
-
abortSignal,
|
94
|
-
});
|
95
|
-
}
|
@@ -1,6 +1,7 @@
|
|
1
1
|
import { z } from "zod";
|
2
2
|
import { callWithRetryAndThrottle } from "../../core/api/callWithRetryAndThrottle.js";
|
3
3
|
import { createJsonResponseHandler, postJsonToApi, } from "../../core/api/postToApi.js";
|
4
|
+
import { zodSchema } from "../../core/schema/ZodSchema.js";
|
4
5
|
import { AbstractModel } from "../../model-function/AbstractModel.js";
|
5
6
|
import { LlamaCppApiConfiguration } from "./LlamaCppApiConfiguration.js";
|
6
7
|
import { failedLlamaCppCallResponseHandler } from "./LlamaCppError.js";
|
@@ -50,13 +51,18 @@ export class LlamaCppTextEmbeddingModel extends AbstractModel {
|
|
50
51
|
if (texts.length > this.maxValuesPerCall) {
|
51
52
|
throw new Error(`The Llama.cpp embedding API only supports ${this.maxValuesPerCall} texts per API call.`);
|
52
53
|
}
|
54
|
+
const api = this.settings.api ?? new LlamaCppApiConfiguration();
|
55
|
+
const abortSignal = options?.run?.abortSignal;
|
53
56
|
return callWithRetryAndThrottle({
|
54
57
|
retry: this.settings.api?.retry,
|
55
58
|
throttle: this.settings.api?.throttle,
|
56
|
-
call: async () =>
|
57
|
-
|
58
|
-
|
59
|
-
content: texts[0],
|
59
|
+
call: async () => postJsonToApi({
|
60
|
+
url: api.assembleUrl(`/embedding`),
|
61
|
+
headers: api.headers,
|
62
|
+
body: { content: texts[0] },
|
63
|
+
failedResponseHandler: failedLlamaCppCallResponseHandler,
|
64
|
+
successfulResponseHandler: createJsonResponseHandler(zodSchema(llamaCppTextEmbeddingResponseSchema)),
|
65
|
+
abortSignal,
|
60
66
|
}),
|
61
67
|
});
|
62
68
|
}
|
@@ -79,13 +85,3 @@ export class LlamaCppTextEmbeddingModel extends AbstractModel {
|
|
79
85
|
const llamaCppTextEmbeddingResponseSchema = z.object({
|
80
86
|
embedding: z.array(z.number()),
|
81
87
|
});
|
82
|
-
async function callLlamaCppEmbeddingAPI({ api = new LlamaCppApiConfiguration(), abortSignal, content, }) {
|
83
|
-
return postJsonToApi({
|
84
|
-
url: api.assembleUrl(`/embedding`),
|
85
|
-
headers: api.headers,
|
86
|
-
body: { content },
|
87
|
-
failedResponseHandler: failedLlamaCppCallResponseHandler,
|
88
|
-
successfulResponseHandler: createJsonResponseHandler(llamaCppTextEmbeddingResponseSchema),
|
89
|
-
abortSignal,
|
90
|
-
});
|
91
|
-
}
|