@workglow/ai-provider 0.0.121 → 0.0.122
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +30 -67
- package/dist/common/HfModelSearch.d.ts +32 -0
- package/dist/common/HfModelSearch.d.ts.map +1 -0
- package/dist/common/PipelineTaskMapping.d.ts +12 -0
- package/dist/common/PipelineTaskMapping.d.ts.map +1 -0
- package/dist/{anthropic → provider-anthropic}/AnthropicProvider.d.ts +2 -14
- package/dist/provider-anthropic/AnthropicProvider.d.ts.map +1 -0
- package/dist/provider-anthropic/AnthropicQueuedProvider.d.ts +16 -0
- package/dist/provider-anthropic/AnthropicQueuedProvider.d.ts.map +1 -0
- package/dist/provider-anthropic/common/Anthropic_Client.d.ts +13 -0
- package/dist/provider-anthropic/common/Anthropic_Client.d.ts.map +1 -0
- package/dist/provider-anthropic/common/Anthropic_Constants.d.ts.map +1 -0
- package/dist/provider-anthropic/common/Anthropic_CountTokens.d.ts +10 -0
- package/dist/provider-anthropic/common/Anthropic_CountTokens.d.ts.map +1 -0
- package/dist/provider-anthropic/common/Anthropic_JobRunFns.d.ts +12 -0
- package/dist/provider-anthropic/common/Anthropic_JobRunFns.d.ts.map +1 -0
- package/dist/provider-anthropic/common/Anthropic_ModelInfo.d.ts +9 -0
- package/dist/provider-anthropic/common/Anthropic_ModelInfo.d.ts.map +1 -0
- package/dist/{anthropic → provider-anthropic}/common/Anthropic_ModelSchema.d.ts +31 -31
- package/dist/provider-anthropic/common/Anthropic_ModelSchema.d.ts.map +1 -0
- package/dist/provider-anthropic/common/Anthropic_ModelSearch.d.ts +8 -0
- package/dist/provider-anthropic/common/Anthropic_ModelSearch.d.ts.map +1 -0
- package/dist/provider-anthropic/common/Anthropic_StructuredGeneration.d.ts +10 -0
- package/dist/provider-anthropic/common/Anthropic_StructuredGeneration.d.ts.map +1 -0
- package/dist/provider-anthropic/common/Anthropic_TextGeneration.d.ts +10 -0
- package/dist/provider-anthropic/common/Anthropic_TextGeneration.d.ts.map +1 -0
- package/dist/provider-anthropic/common/Anthropic_TextRewriter.d.ts +10 -0
- package/dist/provider-anthropic/common/Anthropic_TextRewriter.d.ts.map +1 -0
- package/dist/provider-anthropic/common/Anthropic_TextSummary.d.ts +10 -0
- package/dist/provider-anthropic/common/Anthropic_TextSummary.d.ts.map +1 -0
- package/dist/provider-anthropic/common/Anthropic_ToolCalling.d.ts +10 -0
- package/dist/provider-anthropic/common/Anthropic_ToolCalling.d.ts.map +1 -0
- package/dist/{anthropic → provider-anthropic}/index.d.ts +1 -3
- package/dist/provider-anthropic/index.d.ts.map +1 -0
- package/dist/{index-60ev6k93.js → provider-anthropic/index.js} +43 -11
- package/dist/provider-anthropic/index.js.map +13 -0
- package/dist/provider-anthropic/registerAnthropic.d.ts +10 -0
- package/dist/provider-anthropic/registerAnthropic.d.ts.map +1 -0
- package/dist/provider-anthropic/registerAnthropicInline.d.ts +8 -0
- package/dist/provider-anthropic/registerAnthropicInline.d.ts.map +1 -0
- package/dist/provider-anthropic/registerAnthropicWorker.d.ts +7 -0
- package/dist/provider-anthropic/registerAnthropicWorker.d.ts.map +1 -0
- package/dist/provider-anthropic/runtime.d.ts +16 -0
- package/dist/provider-anthropic/runtime.d.ts.map +1 -0
- package/dist/{anthropic/index.js → provider-anthropic/runtime.js} +291 -177
- package/dist/provider-anthropic/runtime.js.map +24 -0
- package/dist/{web-browser → provider-chrome}/WebBrowserProvider.d.ts +2 -15
- package/dist/provider-chrome/WebBrowserProvider.d.ts.map +1 -0
- package/dist/provider-chrome/WebBrowserQueuedProvider.d.ts +16 -0
- package/dist/provider-chrome/WebBrowserQueuedProvider.d.ts.map +1 -0
- package/dist/provider-chrome/common/WebBrowser_ChromeHelpers.d.ts +31 -0
- package/dist/provider-chrome/common/WebBrowser_ChromeHelpers.d.ts.map +1 -0
- package/dist/provider-chrome/common/WebBrowser_Constants.d.ts.map +1 -0
- package/dist/provider-chrome/common/WebBrowser_JobRunFns.d.ts +10 -0
- package/dist/provider-chrome/common/WebBrowser_JobRunFns.d.ts.map +1 -0
- package/dist/provider-chrome/common/WebBrowser_ModelInfo.d.ts +9 -0
- package/dist/provider-chrome/common/WebBrowser_ModelInfo.d.ts.map +1 -0
- package/dist/{web-browser → provider-chrome}/common/WebBrowser_ModelSchema.d.ts +31 -31
- package/dist/provider-chrome/common/WebBrowser_ModelSchema.d.ts.map +1 -0
- package/dist/provider-chrome/common/WebBrowser_ModelSearch.d.ts +8 -0
- package/dist/provider-chrome/common/WebBrowser_ModelSearch.d.ts.map +1 -0
- package/dist/provider-chrome/common/WebBrowser_TextGeneration.d.ts +10 -0
- package/dist/provider-chrome/common/WebBrowser_TextGeneration.d.ts.map +1 -0
- package/dist/provider-chrome/common/WebBrowser_TextLanguageDetection.d.ts +9 -0
- package/dist/provider-chrome/common/WebBrowser_TextLanguageDetection.d.ts.map +1 -0
- package/dist/provider-chrome/common/WebBrowser_TextRewriter.d.ts +10 -0
- package/dist/provider-chrome/common/WebBrowser_TextRewriter.d.ts.map +1 -0
- package/dist/provider-chrome/common/WebBrowser_TextSummary.d.ts +10 -0
- package/dist/provider-chrome/common/WebBrowser_TextSummary.d.ts.map +1 -0
- package/dist/provider-chrome/common/WebBrowser_TextTranslation.d.ts +10 -0
- package/dist/provider-chrome/common/WebBrowser_TextTranslation.d.ts.map +1 -0
- package/dist/{web-browser → provider-chrome}/index.d.ts +1 -3
- package/dist/provider-chrome/index.d.ts.map +1 -0
- package/dist/provider-chrome/index.js +132 -0
- package/dist/provider-chrome/index.js.map +13 -0
- package/dist/provider-chrome/registerWebBrowser.d.ts +10 -0
- package/dist/provider-chrome/registerWebBrowser.d.ts.map +1 -0
- package/dist/provider-chrome/registerWebBrowserInline.d.ts +8 -0
- package/dist/provider-chrome/registerWebBrowserInline.d.ts.map +1 -0
- package/dist/provider-chrome/registerWebBrowserWorker.d.ts +7 -0
- package/dist/provider-chrome/registerWebBrowserWorker.d.ts.map +1 -0
- package/dist/provider-chrome/runtime.d.ts +14 -0
- package/dist/provider-chrome/runtime.d.ts.map +1 -0
- package/dist/{web-browser/index.js → provider-chrome/runtime.js} +260 -235
- package/dist/provider-chrome/runtime.js.map +23 -0
- package/dist/{google-gemini → provider-gemini}/GoogleGeminiProvider.d.ts +2 -15
- package/dist/provider-gemini/GoogleGeminiProvider.d.ts.map +1 -0
- package/dist/provider-gemini/GoogleGeminiQueuedProvider.d.ts +16 -0
- package/dist/provider-gemini/GoogleGeminiQueuedProvider.d.ts.map +1 -0
- package/dist/provider-gemini/common/Gemini_Client.d.ts +10 -0
- package/dist/provider-gemini/common/Gemini_Client.d.ts.map +1 -0
- package/dist/provider-gemini/common/Gemini_Constants.d.ts.map +1 -0
- package/dist/provider-gemini/common/Gemini_CountTokens.d.ts +10 -0
- package/dist/provider-gemini/common/Gemini_CountTokens.d.ts.map +1 -0
- package/dist/provider-gemini/common/Gemini_JobRunFns.d.ts +13 -0
- package/dist/provider-gemini/common/Gemini_JobRunFns.d.ts.map +1 -0
- package/dist/provider-gemini/common/Gemini_ModelInfo.d.ts +9 -0
- package/dist/provider-gemini/common/Gemini_ModelInfo.d.ts.map +1 -0
- package/dist/{google-gemini → provider-gemini}/common/Gemini_ModelSchema.d.ts +31 -31
- package/dist/provider-gemini/common/Gemini_ModelSchema.d.ts.map +1 -0
- package/dist/provider-gemini/common/Gemini_ModelSearch.d.ts +8 -0
- package/dist/provider-gemini/common/Gemini_ModelSearch.d.ts.map +1 -0
- package/dist/provider-gemini/common/Gemini_Schema.d.ts +11 -0
- package/dist/provider-gemini/common/Gemini_Schema.d.ts.map +1 -0
- package/dist/provider-gemini/common/Gemini_StructuredGeneration.d.ts +10 -0
- package/dist/provider-gemini/common/Gemini_StructuredGeneration.d.ts.map +1 -0
- package/dist/provider-gemini/common/Gemini_TextEmbedding.d.ts +9 -0
- package/dist/provider-gemini/common/Gemini_TextEmbedding.d.ts.map +1 -0
- package/dist/provider-gemini/common/Gemini_TextGeneration.d.ts +10 -0
- package/dist/provider-gemini/common/Gemini_TextGeneration.d.ts.map +1 -0
- package/dist/provider-gemini/common/Gemini_TextRewriter.d.ts +10 -0
- package/dist/provider-gemini/common/Gemini_TextRewriter.d.ts.map +1 -0
- package/dist/provider-gemini/common/Gemini_TextSummary.d.ts +10 -0
- package/dist/provider-gemini/common/Gemini_TextSummary.d.ts.map +1 -0
- package/dist/provider-gemini/common/Gemini_ToolCalling.d.ts +10 -0
- package/dist/provider-gemini/common/Gemini_ToolCalling.d.ts.map +1 -0
- package/dist/{google-gemini → provider-gemini}/index.d.ts +1 -3
- package/dist/provider-gemini/index.d.ts.map +1 -0
- package/dist/{index-8651nz8y.js → provider-gemini/index.js} +43 -11
- package/dist/provider-gemini/index.js.map +13 -0
- package/dist/provider-gemini/registerGemini.d.ts +10 -0
- package/dist/provider-gemini/registerGemini.d.ts.map +1 -0
- package/dist/provider-gemini/registerGeminiInline.d.ts +8 -0
- package/dist/provider-gemini/registerGeminiInline.d.ts.map +1 -0
- package/dist/{anthropic/Anthropic_Worker.d.ts → provider-gemini/registerGeminiWorker.d.ts} +2 -2
- package/dist/provider-gemini/registerGeminiWorker.d.ts.map +1 -0
- package/dist/provider-gemini/runtime.d.ts +16 -0
- package/dist/provider-gemini/runtime.d.ts.map +1 -0
- package/dist/{google-gemini/index.js → provider-gemini/runtime.js} +281 -173
- package/dist/provider-gemini/runtime.js.map +26 -0
- package/dist/provider-hf-inference/HfInferenceProvider.d.ts +2 -19
- package/dist/provider-hf-inference/HfInferenceProvider.d.ts.map +1 -1
- package/dist/provider-hf-inference/HfInferenceQueuedProvider.d.ts +16 -0
- package/dist/provider-hf-inference/HfInferenceQueuedProvider.d.ts.map +1 -0
- package/dist/provider-hf-inference/common/HFI_Client.d.ts +12 -0
- package/dist/provider-hf-inference/common/HFI_Client.d.ts.map +1 -0
- package/dist/provider-hf-inference/common/HFI_JobRunFns.d.ts +2 -11
- package/dist/provider-hf-inference/common/HFI_JobRunFns.d.ts.map +1 -1
- package/dist/provider-hf-inference/common/HFI_ModelInfo.d.ts +9 -0
- package/dist/provider-hf-inference/common/HFI_ModelInfo.d.ts.map +1 -0
- package/dist/provider-hf-inference/common/HFI_ModelSchema.d.ts +31 -31
- package/dist/provider-hf-inference/common/HFI_ModelSchema.d.ts.map +1 -1
- package/dist/provider-hf-inference/common/HFI_ModelSearch.d.ts +8 -0
- package/dist/provider-hf-inference/common/HFI_ModelSearch.d.ts.map +1 -0
- package/dist/provider-hf-inference/common/HFI_TextEmbedding.d.ts +9 -0
- package/dist/provider-hf-inference/common/HFI_TextEmbedding.d.ts.map +1 -0
- package/dist/provider-hf-inference/common/HFI_TextGeneration.d.ts +10 -0
- package/dist/provider-hf-inference/common/HFI_TextGeneration.d.ts.map +1 -0
- package/dist/provider-hf-inference/common/HFI_TextRewriter.d.ts +10 -0
- package/dist/provider-hf-inference/common/HFI_TextRewriter.d.ts.map +1 -0
- package/dist/provider-hf-inference/common/HFI_TextSummary.d.ts +10 -0
- package/dist/provider-hf-inference/common/HFI_TextSummary.d.ts.map +1 -0
- package/dist/provider-hf-inference/common/HFI_ToolCalling.d.ts +10 -0
- package/dist/provider-hf-inference/common/HFI_ToolCalling.d.ts.map +1 -0
- package/dist/provider-hf-inference/index.d.ts +1 -3
- package/dist/provider-hf-inference/index.d.ts.map +1 -1
- package/dist/provider-hf-inference/index.js +98 -411
- package/dist/provider-hf-inference/index.js.map +7 -5
- package/dist/provider-hf-inference/registerHfInference.d.ts +10 -0
- package/dist/provider-hf-inference/registerHfInference.d.ts.map +1 -0
- package/dist/provider-hf-inference/registerHfInferenceInline.d.ts +8 -0
- package/dist/provider-hf-inference/registerHfInferenceInline.d.ts.map +1 -0
- package/dist/provider-hf-inference/registerHfInferenceWorker.d.ts +7 -0
- package/dist/provider-hf-inference/registerHfInferenceWorker.d.ts.map +1 -0
- package/dist/provider-hf-inference/runtime.d.ts +16 -0
- package/dist/provider-hf-inference/runtime.d.ts.map +1 -0
- package/dist/provider-hf-inference/runtime.js +592 -0
- package/dist/provider-hf-inference/runtime.js.map +25 -0
- package/dist/{hf-transformers → provider-hf-transformers}/HuggingFaceTransformersProvider.d.ts +2 -21
- package/dist/provider-hf-transformers/HuggingFaceTransformersProvider.d.ts.map +1 -0
- package/dist/provider-hf-transformers/HuggingFaceTransformersQueuedProvider.d.ts +16 -0
- package/dist/provider-hf-transformers/HuggingFaceTransformersQueuedProvider.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_BackgroundRemoval.d.ts +12 -0
- package/dist/provider-hf-transformers/common/HFT_BackgroundRemoval.d.ts.map +1 -0
- package/dist/{hf-transformers → provider-hf-transformers}/common/HFT_Constants.d.ts +25 -23
- package/dist/provider-hf-transformers/common/HFT_Constants.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_CountTokens.d.ts +10 -0
- package/dist/provider-hf-transformers/common/HFT_CountTokens.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_Download.d.ts +13 -0
- package/dist/provider-hf-transformers/common/HFT_Download.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_ImageClassification.d.ts +13 -0
- package/dist/provider-hf-transformers/common/HFT_ImageClassification.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_ImageEmbedding.d.ts +12 -0
- package/dist/provider-hf-transformers/common/HFT_ImageEmbedding.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_ImageHelpers.d.ts +11 -0
- package/dist/provider-hf-transformers/common/HFT_ImageHelpers.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_ImageSegmentation.d.ts +12 -0
- package/dist/provider-hf-transformers/common/HFT_ImageSegmentation.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_ImageToText.d.ts +12 -0
- package/dist/provider-hf-transformers/common/HFT_ImageToText.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_InlineLifecycle.d.ts +7 -0
- package/dist/provider-hf-transformers/common/HFT_InlineLifecycle.d.ts.map +1 -0
- package/dist/{hf-transformers → provider-hf-transformers}/common/HFT_JobRunFns.d.ts +673 -789
- package/dist/provider-hf-transformers/common/HFT_JobRunFns.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_ModelInfo.d.ts +9 -0
- package/dist/provider-hf-transformers/common/HFT_ModelInfo.d.ts.map +1 -0
- package/dist/{hf-transformers → provider-hf-transformers}/common/HFT_ModelSchema.d.ts +37 -37
- package/dist/provider-hf-transformers/common/HFT_ModelSchema.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_ModelSearch.d.ts +8 -0
- package/dist/provider-hf-transformers/common/HFT_ModelSearch.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_ObjectDetection.d.ts +13 -0
- package/dist/provider-hf-transformers/common/HFT_ObjectDetection.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_OnnxDtypes.d.ts +23 -0
- package/dist/provider-hf-transformers/common/HFT_OnnxDtypes.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_Pipeline.d.ts +32 -0
- package/dist/provider-hf-transformers/common/HFT_Pipeline.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_Streaming.d.ts +24 -0
- package/dist/provider-hf-transformers/common/HFT_Streaming.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_StructuredGeneration.d.ts +10 -0
- package/dist/provider-hf-transformers/common/HFT_StructuredGeneration.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_TextClassification.d.ts +9 -0
- package/dist/provider-hf-transformers/common/HFT_TextClassification.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_TextEmbedding.d.ts +13 -0
- package/dist/provider-hf-transformers/common/HFT_TextEmbedding.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_TextFillMask.d.ts +9 -0
- package/dist/provider-hf-transformers/common/HFT_TextFillMask.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_TextGeneration.d.ts +14 -0
- package/dist/provider-hf-transformers/common/HFT_TextGeneration.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_TextLanguageDetection.d.ts +9 -0
- package/dist/provider-hf-transformers/common/HFT_TextLanguageDetection.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_TextNamedEntityRecognition.d.ts +9 -0
- package/dist/provider-hf-transformers/common/HFT_TextNamedEntityRecognition.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_TextOutput.d.ts +8 -0
- package/dist/provider-hf-transformers/common/HFT_TextOutput.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_TextQuestionAnswer.d.ts +14 -0
- package/dist/provider-hf-transformers/common/HFT_TextQuestionAnswer.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_TextRewriter.d.ts +14 -0
- package/dist/provider-hf-transformers/common/HFT_TextRewriter.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_TextSummary.d.ts +14 -0
- package/dist/provider-hf-transformers/common/HFT_TextSummary.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_TextTranslation.d.ts +14 -0
- package/dist/provider-hf-transformers/common/HFT_TextTranslation.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_ToolCalling.d.ts +10 -0
- package/dist/provider-hf-transformers/common/HFT_ToolCalling.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_ToolMarkup.d.ts +40 -0
- package/dist/provider-hf-transformers/common/HFT_ToolMarkup.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_Unload.d.ts +13 -0
- package/dist/provider-hf-transformers/common/HFT_Unload.d.ts.map +1 -0
- package/dist/{hf-transformers → provider-hf-transformers}/index.d.ts +4 -2
- package/dist/provider-hf-transformers/index.d.ts.map +1 -0
- package/dist/provider-hf-transformers/index.js +513 -0
- package/dist/provider-hf-transformers/index.js.map +16 -0
- package/dist/provider-hf-transformers/registerHuggingFaceTransformers.d.ts +14 -0
- package/dist/provider-hf-transformers/registerHuggingFaceTransformers.d.ts.map +1 -0
- package/dist/provider-hf-transformers/registerHuggingFaceTransformersInline.d.ts +15 -0
- package/dist/provider-hf-transformers/registerHuggingFaceTransformersInline.d.ts.map +1 -0
- package/dist/provider-hf-transformers/registerHuggingFaceTransformersWorker.d.ts +7 -0
- package/dist/provider-hf-transformers/registerHuggingFaceTransformersWorker.d.ts.map +1 -0
- package/dist/provider-hf-transformers/runtime.d.ts +21 -0
- package/dist/provider-hf-transformers/runtime.d.ts.map +1 -0
- package/dist/{index-j4g81r4k.js → provider-hf-transformers/runtime.js} +1561 -927
- package/dist/provider-hf-transformers/runtime.js.map +49 -0
- package/dist/provider-llamacpp/LlamaCppProvider.d.ts +2 -15
- package/dist/provider-llamacpp/LlamaCppProvider.d.ts.map +1 -1
- package/dist/provider-llamacpp/LlamaCppQueuedProvider.d.ts +16 -0
- package/dist/provider-llamacpp/LlamaCppQueuedProvider.d.ts.map +1 -0
- package/dist/provider-llamacpp/common/LlamaCpp_CountTokens.d.ts +10 -0
- package/dist/provider-llamacpp/common/LlamaCpp_CountTokens.d.ts.map +1 -0
- package/dist/provider-llamacpp/common/LlamaCpp_Download.d.ts +9 -0
- package/dist/provider-llamacpp/common/LlamaCpp_Download.d.ts.map +1 -0
- package/dist/provider-llamacpp/common/LlamaCpp_JobRunFns.d.ts +2 -18
- package/dist/provider-llamacpp/common/LlamaCpp_JobRunFns.d.ts.map +1 -1
- package/dist/provider-llamacpp/common/LlamaCpp_ModelInfo.d.ts +9 -0
- package/dist/provider-llamacpp/common/LlamaCpp_ModelInfo.d.ts.map +1 -0
- package/dist/provider-llamacpp/common/LlamaCpp_ModelSchema.d.ts +31 -31
- package/dist/provider-llamacpp/common/LlamaCpp_ModelSchema.d.ts.map +1 -1
- package/dist/provider-llamacpp/common/LlamaCpp_ModelSearch.d.ts +8 -0
- package/dist/provider-llamacpp/common/LlamaCpp_ModelSearch.d.ts.map +1 -0
- package/dist/provider-llamacpp/common/LlamaCpp_Runtime.d.ts +31 -0
- package/dist/provider-llamacpp/common/LlamaCpp_Runtime.d.ts.map +1 -0
- package/dist/provider-llamacpp/common/LlamaCpp_StructuredGeneration.d.ts +10 -0
- package/dist/provider-llamacpp/common/LlamaCpp_StructuredGeneration.d.ts.map +1 -0
- package/dist/provider-llamacpp/common/LlamaCpp_TextEmbedding.d.ts +9 -0
- package/dist/provider-llamacpp/common/LlamaCpp_TextEmbedding.d.ts.map +1 -0
- package/dist/provider-llamacpp/common/LlamaCpp_TextGeneration.d.ts +10 -0
- package/dist/provider-llamacpp/common/LlamaCpp_TextGeneration.d.ts.map +1 -0
- package/dist/provider-llamacpp/common/LlamaCpp_TextRewriter.d.ts +10 -0
- package/dist/provider-llamacpp/common/LlamaCpp_TextRewriter.d.ts.map +1 -0
- package/dist/provider-llamacpp/common/LlamaCpp_TextSummary.d.ts +10 -0
- package/dist/provider-llamacpp/common/LlamaCpp_TextSummary.d.ts.map +1 -0
- package/dist/provider-llamacpp/common/LlamaCpp_ToolCalling.d.ts +10 -0
- package/dist/provider-llamacpp/common/LlamaCpp_ToolCalling.d.ts.map +1 -0
- package/dist/provider-llamacpp/common/LlamaCpp_Unload.d.ts +9 -0
- package/dist/provider-llamacpp/common/LlamaCpp_Unload.d.ts.map +1 -0
- package/dist/provider-llamacpp/index.d.ts +1 -3
- package/dist/provider-llamacpp/index.d.ts.map +1 -1
- package/dist/provider-llamacpp/index.js +121 -725
- package/dist/provider-llamacpp/index.js.map +7 -5
- package/dist/provider-llamacpp/registerLlamaCpp.d.ts +10 -0
- package/dist/provider-llamacpp/registerLlamaCpp.d.ts.map +1 -0
- package/dist/provider-llamacpp/registerLlamaCppInline.d.ts +8 -0
- package/dist/provider-llamacpp/registerLlamaCppInline.d.ts.map +1 -0
- package/dist/provider-llamacpp/registerLlamaCppWorker.d.ts +7 -0
- package/dist/provider-llamacpp/registerLlamaCppWorker.d.ts.map +1 -0
- package/dist/provider-llamacpp/runtime.d.ts +16 -0
- package/dist/provider-llamacpp/runtime.d.ts.map +1 -0
- package/dist/provider-llamacpp/runtime.js +929 -0
- package/dist/provider-llamacpp/runtime.js.map +29 -0
- package/dist/provider-ollama/OllamaProvider.d.ts +2 -15
- package/dist/provider-ollama/OllamaProvider.d.ts.map +1 -1
- package/dist/provider-ollama/OllamaQueuedProvider.d.ts +16 -0
- package/dist/provider-ollama/OllamaQueuedProvider.d.ts.map +1 -0
- package/dist/provider-ollama/common/Ollama_Client.browser.d.ts +13 -0
- package/dist/provider-ollama/common/Ollama_Client.browser.d.ts.map +1 -0
- package/dist/provider-ollama/common/Ollama_Client.d.ts +13 -0
- package/dist/provider-ollama/common/Ollama_Client.d.ts.map +1 -0
- package/dist/provider-ollama/common/Ollama_JobRunFns.browser.d.ts +362 -11
- package/dist/provider-ollama/common/Ollama_JobRunFns.browser.d.ts.map +1 -1
- package/dist/provider-ollama/common/Ollama_JobRunFns.d.ts +361 -11
- package/dist/provider-ollama/common/Ollama_JobRunFns.d.ts.map +1 -1
- package/dist/provider-ollama/common/Ollama_ModelInfo.d.ts +11 -0
- package/dist/provider-ollama/common/Ollama_ModelInfo.d.ts.map +1 -0
- package/dist/provider-ollama/common/Ollama_ModelSchema.d.ts +30 -30
- package/dist/provider-ollama/common/Ollama_ModelSchema.d.ts.map +1 -1
- package/dist/provider-ollama/common/Ollama_ModelSearch.d.ts +11 -0
- package/dist/provider-ollama/common/Ollama_ModelSearch.d.ts.map +1 -0
- package/dist/provider-ollama/common/Ollama_ModelUtil.d.ts +8 -0
- package/dist/provider-ollama/common/Ollama_ModelUtil.d.ts.map +1 -0
- package/dist/provider-ollama/common/Ollama_TextEmbedding.d.ts +11 -0
- package/dist/provider-ollama/common/Ollama_TextEmbedding.d.ts.map +1 -0
- package/dist/provider-ollama/common/Ollama_TextGeneration.d.ts +12 -0
- package/dist/provider-ollama/common/Ollama_TextGeneration.d.ts.map +1 -0
- package/dist/provider-ollama/common/Ollama_TextRewriter.d.ts +12 -0
- package/dist/provider-ollama/common/Ollama_TextRewriter.d.ts.map +1 -0
- package/dist/provider-ollama/common/Ollama_TextSummary.d.ts +12 -0
- package/dist/provider-ollama/common/Ollama_TextSummary.d.ts.map +1 -0
- package/dist/provider-ollama/common/Ollama_ToolCalling.d.ts +16 -0
- package/dist/provider-ollama/common/Ollama_ToolCalling.d.ts.map +1 -0
- package/dist/provider-ollama/index.browser.d.ts +1 -3
- package/dist/provider-ollama/index.browser.d.ts.map +1 -1
- package/dist/provider-ollama/index.browser.js +18 -396
- package/dist/provider-ollama/index.browser.js.map +6 -7
- package/dist/provider-ollama/index.d.ts +1 -3
- package/dist/provider-ollama/index.d.ts.map +1 -1
- package/dist/provider-ollama/index.js +93 -382
- package/dist/provider-ollama/index.js.map +7 -5
- package/dist/provider-ollama/registerOllama.d.ts +10 -0
- package/dist/provider-ollama/registerOllama.d.ts.map +1 -0
- package/dist/provider-ollama/registerOllamaInline.browser.d.ts +8 -0
- package/dist/provider-ollama/registerOllamaInline.browser.d.ts.map +1 -0
- package/dist/provider-ollama/registerOllamaInline.d.ts +8 -0
- package/dist/provider-ollama/registerOllamaInline.d.ts.map +1 -0
- package/dist/provider-ollama/registerOllamaWorker.browser.d.ts +7 -0
- package/dist/provider-ollama/registerOllamaWorker.browser.d.ts.map +1 -0
- package/dist/{google-gemini/Gemini_Worker.d.ts → provider-ollama/registerOllamaWorker.d.ts} +2 -2
- package/dist/provider-ollama/registerOllamaWorker.d.ts.map +1 -0
- package/dist/provider-ollama/runtime.browser.d.ts +16 -0
- package/dist/provider-ollama/runtime.browser.d.ts.map +1 -0
- package/dist/provider-ollama/runtime.browser.js +528 -0
- package/dist/provider-ollama/runtime.browser.js.map +24 -0
- package/dist/provider-ollama/runtime.d.ts +16 -0
- package/dist/provider-ollama/runtime.d.ts.map +1 -0
- package/dist/provider-ollama/runtime.js +538 -0
- package/dist/provider-ollama/runtime.js.map +24 -0
- package/dist/provider-openai/OpenAiProvider.d.ts +2 -19
- package/dist/provider-openai/OpenAiProvider.d.ts.map +1 -1
- package/dist/provider-openai/OpenAiQueuedProvider.d.ts +16 -0
- package/dist/provider-openai/OpenAiQueuedProvider.d.ts.map +1 -0
- package/dist/provider-openai/common/OpenAI_Client.d.ts +10 -0
- package/dist/provider-openai/common/OpenAI_Client.d.ts.map +1 -0
- package/dist/provider-openai/common/OpenAI_CountTokens.d.ts +10 -0
- package/dist/provider-openai/common/OpenAI_CountTokens.d.ts.map +1 -0
- package/dist/provider-openai/common/OpenAI_JobRunFns.d.ts +2 -15
- package/dist/provider-openai/common/OpenAI_JobRunFns.d.ts.map +1 -1
- package/dist/provider-openai/common/OpenAI_ModelInfo.d.ts +9 -0
- package/dist/provider-openai/common/OpenAI_ModelInfo.d.ts.map +1 -0
- package/dist/provider-openai/common/OpenAI_ModelSchema.d.ts +31 -31
- package/dist/provider-openai/common/OpenAI_ModelSchema.d.ts.map +1 -1
- package/dist/provider-openai/common/OpenAI_ModelSearch.d.ts +8 -0
- package/dist/provider-openai/common/OpenAI_ModelSearch.d.ts.map +1 -0
- package/dist/provider-openai/common/OpenAI_StructuredGeneration.d.ts +10 -0
- package/dist/provider-openai/common/OpenAI_StructuredGeneration.d.ts.map +1 -0
- package/dist/provider-openai/common/OpenAI_TextEmbedding.d.ts +9 -0
- package/dist/provider-openai/common/OpenAI_TextEmbedding.d.ts.map +1 -0
- package/dist/provider-openai/common/OpenAI_TextGeneration.d.ts +10 -0
- package/dist/provider-openai/common/OpenAI_TextGeneration.d.ts.map +1 -0
- package/dist/provider-openai/common/OpenAI_TextRewriter.d.ts +10 -0
- package/dist/provider-openai/common/OpenAI_TextRewriter.d.ts.map +1 -0
- package/dist/provider-openai/common/OpenAI_TextSummary.d.ts +10 -0
- package/dist/provider-openai/common/OpenAI_TextSummary.d.ts.map +1 -0
- package/dist/provider-openai/common/OpenAI_ToolCalling.d.ts +10 -0
- package/dist/provider-openai/common/OpenAI_ToolCalling.d.ts.map +1 -0
- package/dist/provider-openai/index.d.ts +1 -3
- package/dist/provider-openai/index.d.ts.map +1 -1
- package/dist/provider-openai/index.js +108 -519
- package/dist/provider-openai/index.js.map +7 -5
- package/dist/provider-openai/registerOpenAi.d.ts +10 -0
- package/dist/provider-openai/registerOpenAi.d.ts.map +1 -0
- package/dist/provider-openai/registerOpenAiInline.d.ts +8 -0
- package/dist/provider-openai/registerOpenAiInline.d.ts.map +1 -0
- package/dist/{ggml/model/GgmlLocalModel.d.ts → provider-openai/registerOpenAiWorker.d.ts} +2 -2
- package/dist/provider-openai/registerOpenAiWorker.d.ts.map +1 -0
- package/dist/provider-openai/runtime.d.ts +16 -0
- package/dist/provider-openai/runtime.d.ts.map +1 -0
- package/dist/provider-openai/runtime.js +662 -0
- package/dist/provider-openai/runtime.js.map +25 -0
- package/dist/provider-tf-mediapipe/TensorFlowMediaPipeProvider.d.ts +24 -0
- package/dist/provider-tf-mediapipe/TensorFlowMediaPipeProvider.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/TensorFlowMediaPipeQueuedProvider.d.ts +16 -0
- package/dist/provider-tf-mediapipe/TensorFlowMediaPipeQueuedProvider.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_Client.d.ts +8 -0
- package/dist/provider-tf-mediapipe/common/TFMP_Client.d.ts.map +1 -0
- package/dist/{tf-mediapipe → provider-tf-mediapipe}/common/TFMP_Constants.d.ts +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_Constants.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_Download.d.ts +9 -0
- package/dist/provider-tf-mediapipe/common/TFMP_Download.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_FaceDetector.d.ts +9 -0
- package/dist/provider-tf-mediapipe/common/TFMP_FaceDetector.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_FaceLandmarker.d.ts +9 -0
- package/dist/provider-tf-mediapipe/common/TFMP_FaceLandmarker.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_GestureRecognizer.d.ts +9 -0
- package/dist/provider-tf-mediapipe/common/TFMP_GestureRecognizer.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_HandLandmarker.d.ts +9 -0
- package/dist/provider-tf-mediapipe/common/TFMP_HandLandmarker.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_ImageClassification.d.ts +9 -0
- package/dist/provider-tf-mediapipe/common/TFMP_ImageClassification.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_ImageEmbedding.d.ts +9 -0
- package/dist/provider-tf-mediapipe/common/TFMP_ImageEmbedding.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_ImageSegmentation.d.ts +9 -0
- package/dist/provider-tf-mediapipe/common/TFMP_ImageSegmentation.d.ts.map +1 -0
- package/dist/{tf-mediapipe → provider-tf-mediapipe}/common/TFMP_JobRunFns.d.ts +340 -408
- package/dist/provider-tf-mediapipe/common/TFMP_JobRunFns.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_ModelInfo.d.ts +9 -0
- package/dist/provider-tf-mediapipe/common/TFMP_ModelInfo.d.ts.map +1 -0
- package/dist/{tf-mediapipe → provider-tf-mediapipe}/common/TFMP_ModelSchema.d.ts +40 -37
- package/dist/provider-tf-mediapipe/common/TFMP_ModelSchema.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_ModelSearch.d.ts +9 -0
- package/dist/provider-tf-mediapipe/common/TFMP_ModelSearch.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_ObjectDetection.d.ts +9 -0
- package/dist/provider-tf-mediapipe/common/TFMP_ObjectDetection.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_PoseLandmarker.d.ts +9 -0
- package/dist/provider-tf-mediapipe/common/TFMP_PoseLandmarker.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_Runtime.d.ts +43 -0
- package/dist/provider-tf-mediapipe/common/TFMP_Runtime.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_TextClassification.d.ts +9 -0
- package/dist/provider-tf-mediapipe/common/TFMP_TextClassification.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_TextEmbedding.d.ts +9 -0
- package/dist/provider-tf-mediapipe/common/TFMP_TextEmbedding.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_TextLanguageDetection.d.ts +9 -0
- package/dist/provider-tf-mediapipe/common/TFMP_TextLanguageDetection.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_Unload.d.ts +9 -0
- package/dist/provider-tf-mediapipe/common/TFMP_Unload.d.ts.map +1 -0
- package/dist/{tf-mediapipe → provider-tf-mediapipe}/index.d.ts +1 -3
- package/dist/provider-tf-mediapipe/index.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/index.js +129 -0
- package/dist/provider-tf-mediapipe/index.js.map +13 -0
- package/dist/provider-tf-mediapipe/registerTensorFlowMediaPipe.d.ts +10 -0
- package/dist/provider-tf-mediapipe/registerTensorFlowMediaPipe.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/registerTensorFlowMediaPipeInline.d.ts +8 -0
- package/dist/provider-tf-mediapipe/registerTensorFlowMediaPipeInline.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/registerTensorFlowMediaPipeWorker.d.ts +7 -0
- package/dist/provider-tf-mediapipe/registerTensorFlowMediaPipeWorker.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/runtime.d.ts +16 -0
- package/dist/provider-tf-mediapipe/runtime.d.ts.map +1 -0
- package/dist/{tf-mediapipe/index.js → provider-tf-mediapipe/runtime.js} +417 -380
- package/dist/provider-tf-mediapipe/runtime.js.map +33 -0
- package/package.json +66 -40
- package/dist/HFT_JobRunFns-8hcpea4c.js +0 -80
- package/dist/HFT_JobRunFns-8hcpea4c.js.map +0 -9
- package/dist/anthropic/AnthropicProvider.d.ts.map +0 -1
- package/dist/anthropic/Anthropic_Worker.d.ts.map +0 -1
- package/dist/anthropic/common/Anthropic_Constants.d.ts.map +0 -1
- package/dist/anthropic/common/Anthropic_JobRunFns.d.ts +0 -24
- package/dist/anthropic/common/Anthropic_JobRunFns.d.ts.map +0 -1
- package/dist/anthropic/common/Anthropic_ModelSchema.d.ts.map +0 -1
- package/dist/anthropic/index.d.ts.map +0 -1
- package/dist/anthropic/index.js.map +0 -11
- package/dist/ggml/model/GgmlLocalModel.d.ts.map +0 -1
- package/dist/google-gemini/Gemini_Worker.d.ts.map +0 -1
- package/dist/google-gemini/GoogleGeminiProvider.d.ts.map +0 -1
- package/dist/google-gemini/common/Gemini_Constants.d.ts.map +0 -1
- package/dist/google-gemini/common/Gemini_JobRunFns.d.ts +0 -25
- package/dist/google-gemini/common/Gemini_JobRunFns.d.ts.map +0 -1
- package/dist/google-gemini/common/Gemini_ModelSchema.d.ts.map +0 -1
- package/dist/google-gemini/index.d.ts.map +0 -1
- package/dist/google-gemini/index.js.map +0 -11
- package/dist/hf-transformers/HFT_Worker.d.ts +0 -7
- package/dist/hf-transformers/HFT_Worker.d.ts.map +0 -1
- package/dist/hf-transformers/HuggingFaceTransformersProvider.d.ts.map +0 -1
- package/dist/hf-transformers/common/HFT_Constants.d.ts.map +0 -1
- package/dist/hf-transformers/common/HFT_JobRunFns.d.ts.map +0 -1
- package/dist/hf-transformers/common/HFT_ModelSchema.d.ts.map +0 -1
- package/dist/hf-transformers/index.d.ts.map +0 -1
- package/dist/hf-transformers/index.js +0 -116
- package/dist/hf-transformers/index.js.map +0 -10
- package/dist/index-60ev6k93.js.map +0 -12
- package/dist/index-6j5pq722.js +0 -11
- package/dist/index-6j5pq722.js.map +0 -9
- package/dist/index-8651nz8y.js.map +0 -12
- package/dist/index-j4g81r4k.js.map +0 -10
- package/dist/index-pkd79j8b.js +0 -58
- package/dist/index-pkd79j8b.js.map +0 -10
- package/dist/index-q2t627d5.js +0 -88
- package/dist/index-q2t627d5.js.map +0 -12
- package/dist/index-tp5s7355.js +0 -77
- package/dist/index-tp5s7355.js.map +0 -12
- package/dist/index-v72vr07f.js +0 -81
- package/dist/index-v72vr07f.js.map +0 -12
- package/dist/index-wr57rwyx.js +0 -104
- package/dist/index-wr57rwyx.js.map +0 -12
- package/dist/index-zqq3kw0n.js +0 -171
- package/dist/index-zqq3kw0n.js.map +0 -11
- package/dist/index.browser-6j5pq722.js +0 -11
- package/dist/index.browser-6j5pq722.js.map +0 -9
- package/dist/index.d.ts +0 -33
- package/dist/index.d.ts.map +0 -1
- package/dist/index.js +0 -316
- package/dist/index.js.map +0 -15
- package/dist/provider-hf-inference/HFI_Worker.d.ts +0 -7
- package/dist/provider-hf-inference/HFI_Worker.d.ts.map +0 -1
- package/dist/provider-llamacpp/LlamaCpp_Worker.d.ts +0 -7
- package/dist/provider-llamacpp/LlamaCpp_Worker.d.ts.map +0 -1
- package/dist/provider-ollama/Ollama_Worker.browser.d.ts +0 -7
- package/dist/provider-ollama/Ollama_Worker.browser.d.ts.map +0 -1
- package/dist/provider-ollama/Ollama_Worker.d.ts +0 -7
- package/dist/provider-ollama/Ollama_Worker.d.ts.map +0 -1
- package/dist/provider-openai/OpenAI_Worker.d.ts +0 -7
- package/dist/provider-openai/OpenAI_Worker.d.ts.map +0 -1
- package/dist/tf-mediapipe/TFMP_Worker.d.ts +0 -7
- package/dist/tf-mediapipe/TFMP_Worker.d.ts.map +0 -1
- package/dist/tf-mediapipe/TensorFlowMediaPipeProvider.d.ts +0 -41
- package/dist/tf-mediapipe/TensorFlowMediaPipeProvider.d.ts.map +0 -1
- package/dist/tf-mediapipe/common/TFMP_Constants.d.ts.map +0 -1
- package/dist/tf-mediapipe/common/TFMP_JobRunFns.d.ts.map +0 -1
- package/dist/tf-mediapipe/common/TFMP_ModelSchema.d.ts.map +0 -1
- package/dist/tf-mediapipe/index.d.ts.map +0 -1
- package/dist/tf-mediapipe/index.js.map +0 -14
- package/dist/types.d.ts +0 -7
- package/dist/types.d.ts.map +0 -1
- package/dist/web-browser/WebBrowserProvider.d.ts.map +0 -1
- package/dist/web-browser/WebBrowser_Worker.d.ts +0 -7
- package/dist/web-browser/WebBrowser_Worker.d.ts.map +0 -1
- package/dist/web-browser/common/WebBrowser_Constants.d.ts.map +0 -1
- package/dist/web-browser/common/WebBrowser_JobRunFns.d.ts +0 -20
- package/dist/web-browser/common/WebBrowser_JobRunFns.d.ts.map +0 -1
- package/dist/web-browser/common/WebBrowser_ModelSchema.d.ts.map +0 -1
- package/dist/web-browser/index.d.ts.map +0 -1
- package/dist/web-browser/index.js.map +0 -14
- /package/dist/{anthropic → provider-anthropic}/common/Anthropic_Constants.d.ts +0 -0
- /package/dist/{web-browser → provider-chrome}/common/WebBrowser_Constants.d.ts +0 -0
- /package/dist/{google-gemini → provider-gemini}/common/Gemini_Constants.d.ts +0 -0
|
@@ -1,741 +1,137 @@
|
|
|
1
|
-
|
|
2
|
-
|
|
3
|
-
|
|
4
|
-
|
|
5
|
-
LlamaCppModelRecordSchema,
|
|
6
|
-
LlamaCppModelSchema,
|
|
7
|
-
LlamaCppProvider
|
|
8
|
-
} from "../index-wr57rwyx.js";
|
|
9
|
-
import {
|
|
10
|
-
__require
|
|
11
|
-
} from "../index-6j5pq722.js";
|
|
12
|
-
// src/provider-llamacpp/common/LlamaCpp_JobRunFns.ts
|
|
13
|
-
import { filterValidToolCalls } from "@workglow/ai";
|
|
14
|
-
import { getLogger, parsePartialJson } from "@workglow/util";
|
|
15
|
-
var _sdk;
|
|
16
|
-
async function loadSdk() {
|
|
17
|
-
if (!_sdk) {
|
|
18
|
-
try {
|
|
19
|
-
_sdk = await import("node-llama-cpp");
|
|
20
|
-
} catch (err) {
|
|
21
|
-
throw new Error("node-llama-cpp is required for LOCAL_LLAMACPP tasks. Install it with: bun add node-llama-cpp");
|
|
22
|
-
}
|
|
23
|
-
}
|
|
24
|
-
return _sdk;
|
|
25
|
-
}
|
|
26
|
-
var llamaInstance;
|
|
27
|
-
var models = new Map;
|
|
28
|
-
var textContexts = new Map;
|
|
29
|
-
var embeddingContexts = new Map;
|
|
30
|
-
var resolvedPaths = new Map;
|
|
31
|
-
async function getLlamaInstance() {
|
|
32
|
-
if (!llamaInstance) {
|
|
33
|
-
const { getLlama } = await loadSdk();
|
|
34
|
-
llamaInstance = await getLlama();
|
|
35
|
-
}
|
|
36
|
-
return llamaInstance;
|
|
37
|
-
}
|
|
38
|
-
function getConfigKey(model) {
|
|
39
|
-
return model.provider_config.model_url ?? model.provider_config.model_path;
|
|
40
|
-
}
|
|
41
|
-
function getActualModelPath(model) {
|
|
42
|
-
const key = getConfigKey(model);
|
|
43
|
-
const resolved = resolvedPaths.get(key);
|
|
44
|
-
return resolved ?? model.provider_config.model_path;
|
|
1
|
+
var __defProp = Object.defineProperty;
|
|
2
|
+
var __returnValue = (v) => v;
|
|
3
|
+
function __exportSetter(name, newValue) {
|
|
4
|
+
this[name] = __returnValue.bind(null, newValue);
|
|
45
5
|
}
|
|
46
|
-
|
|
47
|
-
|
|
48
|
-
|
|
49
|
-
|
|
50
|
-
|
|
51
|
-
|
|
52
|
-
|
|
53
|
-
const loadedModel = await llama.loadModel({
|
|
54
|
-
modelPath,
|
|
55
|
-
...config.gpu_layers !== undefined && { gpuLayers: config.gpu_layers }
|
|
56
|
-
});
|
|
57
|
-
models.set(modelPath, loadedModel);
|
|
58
|
-
return loadedModel;
|
|
59
|
-
}
|
|
60
|
-
async function getOrCreateTextContext(model) {
|
|
61
|
-
const modelPath = getActualModelPath(model);
|
|
62
|
-
const cached = textContexts.get(modelPath);
|
|
63
|
-
if (cached)
|
|
64
|
-
return cached;
|
|
65
|
-
const loadedModel = await getOrLoadModel(model);
|
|
66
|
-
const config = model.provider_config;
|
|
67
|
-
const context = await loadedModel.createContext({
|
|
68
|
-
...config.context_size && { contextSize: config.context_size },
|
|
69
|
-
...config.flash_attention !== undefined && { flashAttention: config.flash_attention }
|
|
70
|
-
});
|
|
71
|
-
textContexts.set(modelPath, context);
|
|
72
|
-
return context;
|
|
73
|
-
}
|
|
74
|
-
async function getOrCreateEmbeddingContext(model) {
|
|
75
|
-
const modelPath = getActualModelPath(model);
|
|
76
|
-
const cached = embeddingContexts.get(modelPath);
|
|
77
|
-
if (cached)
|
|
78
|
-
return cached;
|
|
79
|
-
const loadedModel = await getOrLoadModel(model);
|
|
80
|
-
const context = await loadedModel.createEmbeddingContext();
|
|
81
|
-
embeddingContexts.set(modelPath, context);
|
|
82
|
-
return context;
|
|
83
|
-
}
|
|
84
|
-
async function* streamFromSession(promptFn, signal) {
|
|
85
|
-
const queue = [];
|
|
86
|
-
let isComplete = false;
|
|
87
|
-
let completionError;
|
|
88
|
-
let resolveWait = null;
|
|
89
|
-
const notifyWaiter = () => {
|
|
90
|
-
resolveWait?.();
|
|
91
|
-
resolveWait = null;
|
|
92
|
-
};
|
|
93
|
-
const promptPromise = promptFn((chunk) => {
|
|
94
|
-
queue.push(chunk);
|
|
95
|
-
notifyWaiter();
|
|
96
|
-
}).then(() => {
|
|
97
|
-
isComplete = true;
|
|
98
|
-
notifyWaiter();
|
|
99
|
-
}).catch((err) => {
|
|
100
|
-
completionError = err;
|
|
101
|
-
isComplete = true;
|
|
102
|
-
notifyWaiter();
|
|
103
|
-
});
|
|
104
|
-
try {
|
|
105
|
-
while (true) {
|
|
106
|
-
while (queue.length > 0) {
|
|
107
|
-
yield { type: "text-delta", port: "text", textDelta: queue.shift() };
|
|
108
|
-
}
|
|
109
|
-
if (isComplete)
|
|
110
|
-
break;
|
|
111
|
-
await new Promise((r) => {
|
|
112
|
-
resolveWait = r;
|
|
113
|
-
});
|
|
114
|
-
}
|
|
115
|
-
while (queue.length > 0) {
|
|
116
|
-
yield { type: "text-delta", port: "text", textDelta: queue.shift() };
|
|
117
|
-
}
|
|
118
|
-
} finally {
|
|
119
|
-
await promptPromise.catch(() => {});
|
|
120
|
-
}
|
|
121
|
-
if (completionError) {
|
|
122
|
-
if (signal.aborted)
|
|
123
|
-
return;
|
|
124
|
-
throw completionError;
|
|
125
|
-
}
|
|
126
|
-
yield { type: "finish", data: {} };
|
|
127
|
-
}
|
|
128
|
-
var LlamaCpp_Download = async (input, model, update_progress, signal) => {
|
|
129
|
-
if (!model)
|
|
130
|
-
throw new Error("Model config is required for DownloadModelTask.");
|
|
131
|
-
const { createModelDownloader } = await loadSdk();
|
|
132
|
-
const config = model.provider_config;
|
|
133
|
-
const modelUri = config.model_url ?? config.model_path;
|
|
134
|
-
const dirPath = config.models_dir ?? LLAMACPP_DEFAULT_MODELS_DIR;
|
|
135
|
-
update_progress(0, "Creating model downloader");
|
|
136
|
-
const downloader = await createModelDownloader({ modelUri, dirPath });
|
|
137
|
-
const progressInterval = setInterval(() => {
|
|
138
|
-
const total = downloader.totalSize;
|
|
139
|
-
const downloaded = downloader.downloadedSize;
|
|
140
|
-
if (total && total > 0 && downloaded !== undefined) {
|
|
141
|
-
const pct = Math.min(99, Math.round(downloaded / total * 100));
|
|
142
|
-
update_progress(pct, "Downloading model", { file: modelUri, progress: pct / 100 });
|
|
143
|
-
}
|
|
144
|
-
}, 500);
|
|
145
|
-
let modelPath;
|
|
146
|
-
try {
|
|
147
|
-
modelPath = await downloader.download();
|
|
148
|
-
} finally {
|
|
149
|
-
clearInterval(progressInterval);
|
|
150
|
-
}
|
|
151
|
-
resolvedPaths.set(getConfigKey(model), modelPath);
|
|
152
|
-
update_progress(100, "Model downloaded", { file: modelUri, progress: 1 });
|
|
153
|
-
return { model: input.model };
|
|
154
|
-
};
|
|
155
|
-
var LlamaCpp_Unload = async (input, model, update_progress, _signal) => {
|
|
156
|
-
if (!model)
|
|
157
|
-
throw new Error("Model config is required for UnloadModelTask.");
|
|
158
|
-
const modelPath = getActualModelPath(model);
|
|
159
|
-
const ctx = textContexts.get(modelPath);
|
|
160
|
-
if (ctx) {
|
|
161
|
-
await ctx.dispose();
|
|
162
|
-
textContexts.delete(modelPath);
|
|
163
|
-
update_progress(33, "Text context disposed");
|
|
164
|
-
}
|
|
165
|
-
const embCtx = embeddingContexts.get(modelPath);
|
|
166
|
-
if (embCtx) {
|
|
167
|
-
await embCtx.dispose();
|
|
168
|
-
embeddingContexts.delete(modelPath);
|
|
169
|
-
update_progress(66, "Embedding context disposed");
|
|
170
|
-
}
|
|
171
|
-
const cachedModel = models.get(modelPath);
|
|
172
|
-
if (cachedModel) {
|
|
173
|
-
await cachedModel.dispose();
|
|
174
|
-
models.delete(modelPath);
|
|
175
|
-
update_progress(100, "Model unloaded from memory");
|
|
176
|
-
} else {
|
|
177
|
-
update_progress(100, "Model was not loaded");
|
|
178
|
-
}
|
|
179
|
-
return { model: input.model };
|
|
180
|
-
};
|
|
181
|
-
var LlamaCpp_TextGeneration = async (input, model, update_progress, signal) => {
|
|
182
|
-
if (Array.isArray(input.prompt)) {
|
|
183
|
-
getLogger().warn("LlamaCpp_TextGeneration: array input received; processing sequentially (no native batch support)");
|
|
184
|
-
const prompts = input.prompt;
|
|
185
|
-
const results = [];
|
|
186
|
-
for (const item of prompts) {
|
|
187
|
-
const r = await LlamaCpp_TextGeneration({ ...input, prompt: item }, model, update_progress, signal);
|
|
188
|
-
results.push(r.text);
|
|
189
|
-
}
|
|
190
|
-
return { text: results };
|
|
191
|
-
}
|
|
192
|
-
if (!model)
|
|
193
|
-
throw new Error("Model config is required for TextGenerationTask.");
|
|
194
|
-
const { LlamaChatSession } = await loadSdk();
|
|
195
|
-
update_progress(0, "Loading model");
|
|
196
|
-
const context = await getOrCreateTextContext(model);
|
|
197
|
-
update_progress(10, "Generating text");
|
|
198
|
-
const sequence = context.getSequence();
|
|
199
|
-
const session = new LlamaChatSession({ contextSequence: sequence });
|
|
200
|
-
try {
|
|
201
|
-
const text = await session.prompt(input.prompt, {
|
|
202
|
-
signal,
|
|
203
|
-
...input.temperature !== undefined && { temperature: input.temperature },
|
|
204
|
-
...input.maxTokens !== undefined && { maxTokens: input.maxTokens },
|
|
205
|
-
...input.topP !== undefined && { topP: input.topP }
|
|
6
|
+
var __export = (target, all) => {
|
|
7
|
+
for (var name in all)
|
|
8
|
+
__defProp(target, name, {
|
|
9
|
+
get: all[name],
|
|
10
|
+
enumerable: true,
|
|
11
|
+
configurable: true,
|
|
12
|
+
set: __exportSetter.bind(all, name)
|
|
206
13
|
});
|
|
207
|
-
update_progress(100, "Text generation complete");
|
|
208
|
-
return { text };
|
|
209
|
-
} finally {
|
|
210
|
-
sequence.dispose();
|
|
211
|
-
}
|
|
212
|
-
};
|
|
213
|
-
var LlamaCpp_TextGeneration_Stream = async function* (input, model, signal) {
|
|
214
|
-
if (!model)
|
|
215
|
-
throw new Error("Model config is required for TextGenerationTask.");
|
|
216
|
-
const { LlamaChatSession } = await loadSdk();
|
|
217
|
-
const context = await getOrCreateTextContext(model);
|
|
218
|
-
const sequence = context.getSequence();
|
|
219
|
-
const session = new LlamaChatSession({ contextSequence: sequence });
|
|
220
|
-
try {
|
|
221
|
-
yield* streamFromSession((onTextChunk) => {
|
|
222
|
-
return session.prompt(input.prompt, {
|
|
223
|
-
signal,
|
|
224
|
-
onTextChunk,
|
|
225
|
-
...input.temperature !== undefined && { temperature: input.temperature },
|
|
226
|
-
...input.maxTokens !== undefined && { maxTokens: input.maxTokens },
|
|
227
|
-
...input.topP !== undefined && { topP: input.topP }
|
|
228
|
-
});
|
|
229
|
-
}, signal);
|
|
230
|
-
} finally {
|
|
231
|
-
sequence.dispose();
|
|
232
|
-
}
|
|
233
|
-
};
|
|
234
|
-
var LlamaCpp_TextEmbedding = async (input, model, update_progress, _signal) => {
|
|
235
|
-
if (!model)
|
|
236
|
-
throw new Error("Model config is required for TextEmbeddingTask.");
|
|
237
|
-
update_progress(0, "Loading embedding model");
|
|
238
|
-
const context = await getOrCreateEmbeddingContext(model);
|
|
239
|
-
const texts = Array.isArray(input.text) ? input.text : [input.text];
|
|
240
|
-
update_progress(10, "Computing embeddings");
|
|
241
|
-
const embeddings = await Promise.all(texts.map((text) => context.getEmbeddingFor(text).then((e) => new Float32Array(e.vector))));
|
|
242
|
-
update_progress(100, "Embeddings complete");
|
|
243
|
-
if (Array.isArray(input.text)) {
|
|
244
|
-
return { vector: embeddings };
|
|
245
|
-
}
|
|
246
|
-
return { vector: embeddings[0] };
|
|
247
|
-
};
|
|
248
|
-
var LlamaCpp_TextRewriter = async (input, model, update_progress, signal) => {
|
|
249
|
-
if (Array.isArray(input.text)) {
|
|
250
|
-
getLogger().warn("LlamaCpp_TextRewriter: array input received; processing sequentially (no native batch support)");
|
|
251
|
-
const texts = input.text;
|
|
252
|
-
const results = [];
|
|
253
|
-
for (const item of texts) {
|
|
254
|
-
const r = await LlamaCpp_TextRewriter({ ...input, text: item }, model, update_progress, signal);
|
|
255
|
-
results.push(r.text);
|
|
256
|
-
}
|
|
257
|
-
return { text: results };
|
|
258
|
-
}
|
|
259
|
-
if (!model)
|
|
260
|
-
throw new Error("Model config is required for TextRewriterTask.");
|
|
261
|
-
const { LlamaChatSession } = await loadSdk();
|
|
262
|
-
update_progress(0, "Loading model");
|
|
263
|
-
const context = await getOrCreateTextContext(model);
|
|
264
|
-
update_progress(10, "Rewriting text");
|
|
265
|
-
const sequence = context.getSequence();
|
|
266
|
-
const session = new LlamaChatSession({
|
|
267
|
-
contextSequence: sequence,
|
|
268
|
-
systemPrompt: input.prompt
|
|
269
|
-
});
|
|
270
|
-
try {
|
|
271
|
-
const text = await session.prompt(input.text, { signal });
|
|
272
|
-
update_progress(100, "Text rewriting complete");
|
|
273
|
-
return { text };
|
|
274
|
-
} finally {
|
|
275
|
-
sequence.dispose();
|
|
276
|
-
}
|
|
277
|
-
};
|
|
278
|
-
var LlamaCpp_TextRewriter_Stream = async function* (input, model, signal) {
|
|
279
|
-
if (!model)
|
|
280
|
-
throw new Error("Model config is required for TextRewriterTask.");
|
|
281
|
-
const { LlamaChatSession } = await loadSdk();
|
|
282
|
-
const context = await getOrCreateTextContext(model);
|
|
283
|
-
const sequence = context.getSequence();
|
|
284
|
-
const session = new LlamaChatSession({
|
|
285
|
-
contextSequence: sequence,
|
|
286
|
-
systemPrompt: input.prompt
|
|
287
|
-
});
|
|
288
|
-
try {
|
|
289
|
-
yield* streamFromSession((onTextChunk) => {
|
|
290
|
-
return session.prompt(input.text, { signal, onTextChunk });
|
|
291
|
-
}, signal);
|
|
292
|
-
} finally {
|
|
293
|
-
sequence.dispose();
|
|
294
|
-
}
|
|
295
|
-
};
|
|
296
|
-
var LlamaCpp_TextSummary = async (input, model, update_progress, signal) => {
|
|
297
|
-
if (Array.isArray(input.text)) {
|
|
298
|
-
getLogger().warn("LlamaCpp_TextSummary: array input received; processing sequentially (no native batch support)");
|
|
299
|
-
const texts = input.text;
|
|
300
|
-
const results = [];
|
|
301
|
-
for (const item of texts) {
|
|
302
|
-
const r = await LlamaCpp_TextSummary({ ...input, text: item }, model, update_progress, signal);
|
|
303
|
-
results.push(r.text);
|
|
304
|
-
}
|
|
305
|
-
return { text: results };
|
|
306
|
-
}
|
|
307
|
-
if (!model)
|
|
308
|
-
throw new Error("Model config is required for TextSummaryTask.");
|
|
309
|
-
const { LlamaChatSession } = await loadSdk();
|
|
310
|
-
update_progress(0, "Loading model");
|
|
311
|
-
const context = await getOrCreateTextContext(model);
|
|
312
|
-
update_progress(10, "Summarizing text");
|
|
313
|
-
const sequence = context.getSequence();
|
|
314
|
-
const session = new LlamaChatSession({
|
|
315
|
-
contextSequence: sequence,
|
|
316
|
-
systemPrompt: "Summarize the following text concisely, preserving the key points."
|
|
317
|
-
});
|
|
318
|
-
try {
|
|
319
|
-
const text = await session.prompt(input.text, { signal });
|
|
320
|
-
update_progress(100, "Summarization complete");
|
|
321
|
-
return { text };
|
|
322
|
-
} finally {
|
|
323
|
-
sequence.dispose();
|
|
324
|
-
}
|
|
325
|
-
};
|
|
326
|
-
var LlamaCpp_TextSummary_Stream = async function* (input, model, signal) {
|
|
327
|
-
if (!model)
|
|
328
|
-
throw new Error("Model config is required for TextSummaryTask.");
|
|
329
|
-
const { LlamaChatSession } = await loadSdk();
|
|
330
|
-
const context = await getOrCreateTextContext(model);
|
|
331
|
-
const sequence = context.getSequence();
|
|
332
|
-
const session = new LlamaChatSession({
|
|
333
|
-
contextSequence: sequence,
|
|
334
|
-
systemPrompt: "Summarize the following text concisely, preserving the key points."
|
|
335
|
-
});
|
|
336
|
-
try {
|
|
337
|
-
yield* streamFromSession((onTextChunk) => {
|
|
338
|
-
return session.prompt(input.text, { signal, onTextChunk });
|
|
339
|
-
}, signal);
|
|
340
|
-
} finally {
|
|
341
|
-
sequence.dispose();
|
|
342
|
-
}
|
|
343
14
|
};
|
|
344
|
-
|
|
345
|
-
|
|
346
|
-
|
|
347
|
-
|
|
348
|
-
|
|
349
|
-
|
|
350
|
-
|
|
351
|
-
|
|
352
|
-
await disposeAll(embeddingContexts);
|
|
353
|
-
await disposeAll(models);
|
|
354
|
-
if (llamaInstance) {
|
|
355
|
-
await llamaInstance.dispose?.().catch(() => {});
|
|
356
|
-
llamaInstance = undefined;
|
|
357
|
-
}
|
|
358
|
-
resolvedPaths.clear();
|
|
359
|
-
}
|
|
360
|
-
var LlamaCpp_CountTokens = async (input, model, onProgress, signal) => {
|
|
361
|
-
if (Array.isArray(input.text)) {
|
|
362
|
-
getLogger().warn("LlamaCpp_CountTokens: array input received; processing sequentially (no native batch support)");
|
|
363
|
-
const texts = input.text;
|
|
364
|
-
const counts = [];
|
|
365
|
-
for (const item of texts) {
|
|
366
|
-
const r = await LlamaCpp_CountTokens({ ...input, text: item }, model, onProgress, signal);
|
|
367
|
-
counts.push(r.count);
|
|
368
|
-
}
|
|
369
|
-
return { count: counts };
|
|
370
|
-
}
|
|
371
|
-
const loadedModel = await getOrLoadModel(model);
|
|
372
|
-
const tokens = loadedModel.tokenizer(input.text);
|
|
373
|
-
return { count: tokens.length };
|
|
374
|
-
};
|
|
375
|
-
var LlamaCpp_CountTokens_Reactive = async (input, _output, model) => {
|
|
376
|
-
return LlamaCpp_CountTokens(input, model, () => {}, new AbortController().signal);
|
|
377
|
-
};
|
|
378
|
-
function buildLlamaCppPrompt(input) {
|
|
379
|
-
const inputMessages = input.messages;
|
|
380
|
-
if (!inputMessages || inputMessages.length === 0) {
|
|
381
|
-
return Array.isArray(input.prompt) ? input.prompt.join(`
|
|
382
|
-
`) : input.prompt;
|
|
383
|
-
}
|
|
384
|
-
const parts = [];
|
|
385
|
-
for (const msg of inputMessages) {
|
|
386
|
-
if (msg.role === "user") {
|
|
387
|
-
parts.push(`User: ${msg.content}`);
|
|
388
|
-
} else if (msg.role === "assistant" && Array.isArray(msg.content)) {
|
|
389
|
-
const text = msg.content.filter((b) => b.type === "text").map((b) => b.text).join("");
|
|
390
|
-
if (text)
|
|
391
|
-
parts.push(`Assistant: ${text}`);
|
|
392
|
-
} else if (msg.role === "tool" && Array.isArray(msg.content)) {
|
|
393
|
-
for (const block of msg.content) {
|
|
394
|
-
parts.push(`Tool Result: ${block.content}`);
|
|
395
|
-
}
|
|
396
|
-
}
|
|
397
|
-
}
|
|
398
|
-
return parts.join(`
|
|
15
|
+
var __esm = (fn, res) => () => (fn && (res = fn(fn = 0)), res);
|
|
16
|
+
var __require = /* @__PURE__ */ ((x) => typeof require !== "undefined" ? require : typeof Proxy !== "undefined" ? new Proxy(x, {
|
|
17
|
+
get: (a, b) => (typeof require !== "undefined" ? require : a)[b]
|
|
18
|
+
}) : x)(function(x) {
|
|
19
|
+
if (typeof require !== "undefined")
|
|
20
|
+
return require.apply(this, arguments);
|
|
21
|
+
throw Error('Dynamic require of "' + x + '" is not supported');
|
|
22
|
+
});
|
|
399
23
|
|
|
400
|
-
|
|
401
|
-
|
|
402
|
-
|
|
403
|
-
|
|
404
|
-
|
|
405
|
-
|
|
406
|
-
|
|
407
|
-
|
|
408
|
-
|
|
409
|
-
|
|
410
|
-
|
|
411
|
-
capturedCalls.push({ name: toolName, input: params ?? {} });
|
|
412
|
-
return "OK";
|
|
413
|
-
}
|
|
414
|
-
});
|
|
415
|
-
}
|
|
416
|
-
return functions;
|
|
417
|
-
}
|
|
418
|
-
var LlamaCpp_ToolCalling = async (input, model, update_progress, signal) => {
|
|
419
|
-
if (Array.isArray(input.prompt)) {
|
|
420
|
-
getLogger().warn("LlamaCpp_ToolCalling: array input received; processing sequentially (no native batch support)");
|
|
421
|
-
const prompts = input.prompt;
|
|
422
|
-
const texts = [];
|
|
423
|
-
const toolCallsList = [];
|
|
424
|
-
for (const item of prompts) {
|
|
425
|
-
const r = await LlamaCpp_ToolCalling({ ...input, prompt: item }, model, update_progress, signal);
|
|
426
|
-
texts.push(r.text);
|
|
427
|
-
toolCallsList.push(r.toolCalls);
|
|
428
|
-
}
|
|
429
|
-
return { text: texts, toolCalls: toolCallsList };
|
|
430
|
-
}
|
|
431
|
-
if (!model)
|
|
432
|
-
throw new Error("Model config is required for ToolCallingTask.");
|
|
433
|
-
await loadSdk();
|
|
434
|
-
update_progress(0, "Loading model");
|
|
435
|
-
const context = await getOrCreateTextContext(model);
|
|
436
|
-
const capturedCalls = [];
|
|
437
|
-
const functions = input.toolChoice === "none" ? undefined : buildLlamaCppFunctions(input.tools, capturedCalls);
|
|
438
|
-
update_progress(10, "Running tool calling");
|
|
439
|
-
const sequence = context.getSequence();
|
|
440
|
-
const { LlamaChatSession } = _sdk;
|
|
441
|
-
const promptText = buildLlamaCppPrompt(input);
|
|
442
|
-
const session = new LlamaChatSession({
|
|
443
|
-
contextSequence: sequence,
|
|
444
|
-
...input.systemPrompt && { systemPrompt: input.systemPrompt }
|
|
445
|
-
});
|
|
446
|
-
try {
|
|
447
|
-
const text = await session.prompt(promptText, {
|
|
448
|
-
signal,
|
|
449
|
-
...functions && { functions },
|
|
450
|
-
...input.temperature !== undefined && { temperature: input.temperature },
|
|
451
|
-
...input.maxTokens !== undefined && { maxTokens: input.maxTokens }
|
|
452
|
-
});
|
|
453
|
-
const toolCalls = [];
|
|
454
|
-
capturedCalls.forEach((call, index) => {
|
|
455
|
-
const id = `call_${index}`;
|
|
456
|
-
toolCalls.push({ id, name: call.name, input: call.input });
|
|
457
|
-
});
|
|
458
|
-
update_progress(100, "Tool calling complete");
|
|
459
|
-
return { text, toolCalls: filterValidToolCalls(toolCalls, input.tools) };
|
|
460
|
-
} finally {
|
|
461
|
-
sequence.dispose();
|
|
462
|
-
}
|
|
463
|
-
};
|
|
464
|
-
var LlamaCpp_ToolCalling_Stream = async function* (input, model, signal) {
|
|
465
|
-
if (!model)
|
|
466
|
-
throw new Error("Model config is required for ToolCallingTask.");
|
|
467
|
-
await loadSdk();
|
|
468
|
-
const context = await getOrCreateTextContext(model);
|
|
469
|
-
const capturedCalls = [];
|
|
470
|
-
const functions = input.toolChoice === "none" ? undefined : buildLlamaCppFunctions(input.tools, capturedCalls);
|
|
471
|
-
const sequence = context.getSequence();
|
|
472
|
-
const { LlamaChatSession } = _sdk;
|
|
473
|
-
const promptText = buildLlamaCppPrompt(input);
|
|
474
|
-
const session = new LlamaChatSession({
|
|
475
|
-
contextSequence: sequence,
|
|
476
|
-
...input.systemPrompt && { systemPrompt: input.systemPrompt }
|
|
477
|
-
});
|
|
478
|
-
const queue = [];
|
|
479
|
-
let isComplete = false;
|
|
480
|
-
let completionError;
|
|
481
|
-
let resolveWait = null;
|
|
482
|
-
const notifyWaiter = () => {
|
|
483
|
-
resolveWait?.();
|
|
484
|
-
resolveWait = null;
|
|
485
|
-
};
|
|
486
|
-
let accumulatedText = "";
|
|
487
|
-
const promptPromise = session.prompt(promptText, {
|
|
488
|
-
signal,
|
|
489
|
-
...functions && { functions },
|
|
490
|
-
onTextChunk: (chunk) => {
|
|
491
|
-
queue.push(chunk);
|
|
492
|
-
notifyWaiter();
|
|
24
|
+
// src/provider-llamacpp/common/LlamaCpp_Constants.ts
|
|
25
|
+
var LOCAL_LLAMACPP = "LOCAL_LLAMACPP";
|
|
26
|
+
var LLAMACPP_DEFAULT_MODELS_DIR = "./models";
|
|
27
|
+
// src/provider-llamacpp/common/LlamaCpp_ModelSchema.ts
|
|
28
|
+
import { ModelConfigSchema, ModelRecordSchema } from "@workglow/ai/worker";
|
|
29
|
+
var LlamaCppModelSchema = {
|
|
30
|
+
type: "object",
|
|
31
|
+
properties: {
|
|
32
|
+
provider: {
|
|
33
|
+
const: LOCAL_LLAMACPP,
|
|
34
|
+
description: "Discriminator: local node-llama-cpp (GGUF) model."
|
|
493
35
|
},
|
|
494
|
-
|
|
495
|
-
|
|
496
|
-
|
|
497
|
-
|
|
498
|
-
|
|
499
|
-
|
|
500
|
-
|
|
501
|
-
|
|
502
|
-
|
|
503
|
-
|
|
504
|
-
|
|
505
|
-
|
|
506
|
-
|
|
507
|
-
|
|
508
|
-
|
|
509
|
-
|
|
510
|
-
|
|
511
|
-
|
|
512
|
-
|
|
513
|
-
|
|
514
|
-
|
|
515
|
-
|
|
516
|
-
|
|
517
|
-
|
|
518
|
-
|
|
519
|
-
|
|
520
|
-
|
|
521
|
-
|
|
522
|
-
|
|
523
|
-
|
|
524
|
-
|
|
525
|
-
|
|
526
|
-
|
|
527
|
-
|
|
528
|
-
|
|
529
|
-
|
|
530
|
-
}
|
|
531
|
-
const toolCalls = [];
|
|
532
|
-
capturedCalls.forEach((call, index) => {
|
|
533
|
-
const id = `call_${index}`;
|
|
534
|
-
toolCalls.push({ id, name: call.name, input: call.input });
|
|
535
|
-
});
|
|
536
|
-
const validToolCalls = filterValidToolCalls(toolCalls, input.tools);
|
|
537
|
-
if (validToolCalls.length > 0) {
|
|
538
|
-
yield { type: "object-delta", port: "toolCalls", objectDelta: [...validToolCalls] };
|
|
539
|
-
}
|
|
540
|
-
yield {
|
|
541
|
-
type: "finish",
|
|
542
|
-
data: { text: accumulatedText, toolCalls: validToolCalls }
|
|
543
|
-
};
|
|
544
|
-
};
|
|
545
|
-
var LlamaCpp_StructuredGeneration = async (input, model, update_progress, signal) => {
|
|
546
|
-
if (!model)
|
|
547
|
-
throw new Error("Model config is required for StructuredGenerationTask.");
|
|
548
|
-
await loadSdk();
|
|
549
|
-
update_progress(0, "Loading model");
|
|
550
|
-
const llama = await getLlamaInstance();
|
|
551
|
-
const context = await getOrCreateTextContext(model);
|
|
552
|
-
update_progress(10, "Running structured generation");
|
|
553
|
-
const grammar = await llama.createGrammarForJsonSchema(input.outputSchema);
|
|
554
|
-
const sequence = context.getSequence();
|
|
555
|
-
const { LlamaChatSession } = _sdk;
|
|
556
|
-
const session = new LlamaChatSession({ contextSequence: sequence });
|
|
557
|
-
try {
|
|
558
|
-
const text = await session.prompt(input.prompt, {
|
|
559
|
-
signal,
|
|
560
|
-
grammar,
|
|
561
|
-
...input.temperature !== undefined && { temperature: input.temperature },
|
|
562
|
-
...input.maxTokens !== undefined && { maxTokens: input.maxTokens }
|
|
563
|
-
});
|
|
564
|
-
let object;
|
|
565
|
-
try {
|
|
566
|
-
object = JSON.parse(text);
|
|
567
|
-
} catch {
|
|
568
|
-
object = {};
|
|
569
|
-
}
|
|
570
|
-
update_progress(100, "Structured generation complete");
|
|
571
|
-
return { object };
|
|
572
|
-
} finally {
|
|
573
|
-
sequence.dispose();
|
|
574
|
-
}
|
|
575
|
-
};
|
|
576
|
-
var LlamaCpp_StructuredGeneration_Stream = async function* (input, model, signal) {
|
|
577
|
-
if (!model)
|
|
578
|
-
throw new Error("Model config is required for StructuredGenerationTask.");
|
|
579
|
-
await loadSdk();
|
|
580
|
-
const llama = await getLlamaInstance();
|
|
581
|
-
const context = await getOrCreateTextContext(model);
|
|
582
|
-
const grammar = await llama.createGrammarForJsonSchema(input.outputSchema);
|
|
583
|
-
const sequence = context.getSequence();
|
|
584
|
-
const { LlamaChatSession } = _sdk;
|
|
585
|
-
const session = new LlamaChatSession({ contextSequence: sequence });
|
|
586
|
-
const queue = [];
|
|
587
|
-
let isComplete = false;
|
|
588
|
-
let completionError;
|
|
589
|
-
let resolveWait = null;
|
|
590
|
-
const notifyWaiter = () => {
|
|
591
|
-
resolveWait?.();
|
|
592
|
-
resolveWait = null;
|
|
593
|
-
};
|
|
594
|
-
let accumulatedText = "";
|
|
595
|
-
const promptPromise = session.prompt(input.prompt, {
|
|
596
|
-
signal,
|
|
597
|
-
grammar,
|
|
598
|
-
onTextChunk: (chunk) => {
|
|
599
|
-
queue.push(chunk);
|
|
600
|
-
notifyWaiter();
|
|
601
|
-
},
|
|
602
|
-
...input.temperature !== undefined && { temperature: input.temperature },
|
|
603
|
-
...input.maxTokens !== undefined && { maxTokens: input.maxTokens }
|
|
604
|
-
}).then(() => {
|
|
605
|
-
isComplete = true;
|
|
606
|
-
notifyWaiter();
|
|
607
|
-
}).catch((err) => {
|
|
608
|
-
completionError = err;
|
|
609
|
-
isComplete = true;
|
|
610
|
-
notifyWaiter();
|
|
611
|
-
});
|
|
612
|
-
try {
|
|
613
|
-
while (true) {
|
|
614
|
-
while (queue.length > 0) {
|
|
615
|
-
const chunk = queue.shift();
|
|
616
|
-
accumulatedText += chunk;
|
|
617
|
-
const partial = parsePartialJson(accumulatedText);
|
|
618
|
-
if (partial !== undefined) {
|
|
619
|
-
yield {
|
|
620
|
-
type: "object-delta",
|
|
621
|
-
port: "object",
|
|
622
|
-
objectDelta: partial
|
|
623
|
-
};
|
|
36
|
+
provider_config: {
|
|
37
|
+
type: "object",
|
|
38
|
+
description: "node-llama-cpp specific configuration.",
|
|
39
|
+
properties: {
|
|
40
|
+
model_path: {
|
|
41
|
+
type: "string",
|
|
42
|
+
description: "Filesystem path to the .gguf model file."
|
|
43
|
+
},
|
|
44
|
+
model_url: {
|
|
45
|
+
type: "string",
|
|
46
|
+
description: "URI or URL to download the model from (e.g. 'hf:user/repo:quant' or an https URL). Used by DownloadModelTask."
|
|
47
|
+
},
|
|
48
|
+
models_dir: {
|
|
49
|
+
type: "string",
|
|
50
|
+
description: "Directory to download models into.",
|
|
51
|
+
default: LLAMACPP_DEFAULT_MODELS_DIR
|
|
52
|
+
},
|
|
53
|
+
gpu_layers: {
|
|
54
|
+
type: "integer",
|
|
55
|
+
description: "Number of model layers to offload to GPU. Use -1 for auto-detection.",
|
|
56
|
+
minimum: -1
|
|
57
|
+
},
|
|
58
|
+
context_size: {
|
|
59
|
+
type: "integer",
|
|
60
|
+
description: "Context window size in tokens.",
|
|
61
|
+
minimum: 1
|
|
62
|
+
},
|
|
63
|
+
flash_attention: {
|
|
64
|
+
type: "boolean",
|
|
65
|
+
description: "Enable flash attention for improved performance where supported.",
|
|
66
|
+
default: true
|
|
67
|
+
},
|
|
68
|
+
embedding: {
|
|
69
|
+
type: "boolean",
|
|
70
|
+
description: "Whether this model is used for text embedding (vs text generation).",
|
|
71
|
+
default: false
|
|
624
72
|
}
|
|
625
|
-
}
|
|
626
|
-
|
|
627
|
-
|
|
628
|
-
|
|
629
|
-
|
|
630
|
-
|
|
631
|
-
|
|
632
|
-
|
|
633
|
-
|
|
634
|
-
|
|
635
|
-
|
|
636
|
-
|
|
637
|
-
|
|
638
|
-
|
|
639
|
-
|
|
640
|
-
|
|
641
|
-
|
|
642
|
-
|
|
643
|
-
|
|
644
|
-
|
|
645
|
-
|
|
646
|
-
|
|
647
|
-
|
|
648
|
-
|
|
649
|
-
|
|
650
|
-
|
|
651
|
-
|
|
652
|
-
|
|
653
|
-
|
|
654
|
-
|
|
655
|
-
|
|
656
|
-
|
|
657
|
-
|
|
658
|
-
|
|
659
|
-
|
|
660
|
-
|
|
661
|
-
|
|
662
|
-
|
|
663
|
-
|
|
664
|
-
|
|
665
|
-
|
|
666
|
-
|
|
667
|
-
|
|
668
|
-
|
|
73
|
+
},
|
|
74
|
+
required: ["model_path"],
|
|
75
|
+
additionalProperties: false
|
|
76
|
+
}
|
|
77
|
+
},
|
|
78
|
+
required: ["provider", "provider_config"],
|
|
79
|
+
additionalProperties: true
|
|
80
|
+
};
|
|
81
|
+
var LlamaCppModelRecordSchema = {
|
|
82
|
+
type: "object",
|
|
83
|
+
properties: {
|
|
84
|
+
...ModelRecordSchema.properties,
|
|
85
|
+
...LlamaCppModelSchema.properties
|
|
86
|
+
},
|
|
87
|
+
required: [...ModelRecordSchema.required, ...LlamaCppModelSchema.required],
|
|
88
|
+
additionalProperties: false
|
|
89
|
+
};
|
|
90
|
+
var LlamaCppModelConfigSchema = {
|
|
91
|
+
type: "object",
|
|
92
|
+
properties: {
|
|
93
|
+
...ModelConfigSchema.properties,
|
|
94
|
+
...LlamaCppModelSchema.properties
|
|
95
|
+
},
|
|
96
|
+
required: [...ModelConfigSchema.required, ...LlamaCppModelSchema.required],
|
|
97
|
+
additionalProperties: false
|
|
98
|
+
};
|
|
99
|
+
// src/provider-llamacpp/LlamaCppQueuedProvider.ts
|
|
100
|
+
import {
|
|
101
|
+
QueuedAiProvider
|
|
102
|
+
} from "@workglow/ai";
|
|
103
|
+
class LlamaCppQueuedProvider extends QueuedAiProvider {
|
|
104
|
+
name = LOCAL_LLAMACPP;
|
|
105
|
+
isLocal = true;
|
|
106
|
+
supportsBrowser = false;
|
|
107
|
+
taskTypes = [
|
|
108
|
+
"DownloadModelTask",
|
|
109
|
+
"UnloadModelTask",
|
|
110
|
+
"ModelInfoTask",
|
|
111
|
+
"CountTokensTask",
|
|
112
|
+
"TextGenerationTask",
|
|
113
|
+
"TextEmbeddingTask",
|
|
114
|
+
"TextRewriterTask",
|
|
115
|
+
"TextSummaryTask",
|
|
116
|
+
"ToolCallingTask",
|
|
117
|
+
"ModelSearchTask"
|
|
118
|
+
];
|
|
119
|
+
constructor(tasks, streamTasks, reactiveTasks) {
|
|
120
|
+
super(tasks, streamTasks, reactiveTasks);
|
|
669
121
|
}
|
|
670
|
-
|
|
671
|
-
|
|
672
|
-
|
|
673
|
-
|
|
674
|
-
|
|
675
|
-
supports_node: true,
|
|
676
|
-
is_cached,
|
|
677
|
-
is_loaded,
|
|
678
|
-
file_sizes
|
|
679
|
-
};
|
|
680
|
-
};
|
|
681
|
-
var LLAMACPP_TASKS = {
|
|
682
|
-
DownloadModelTask: LlamaCpp_Download,
|
|
683
|
-
UnloadModelTask: LlamaCpp_Unload,
|
|
684
|
-
ModelInfoTask: LlamaCpp_ModelInfo,
|
|
685
|
-
CountTokensTask: LlamaCpp_CountTokens,
|
|
686
|
-
TextGenerationTask: LlamaCpp_TextGeneration,
|
|
687
|
-
TextEmbeddingTask: LlamaCpp_TextEmbedding,
|
|
688
|
-
TextRewriterTask: LlamaCpp_TextRewriter,
|
|
689
|
-
TextSummaryTask: LlamaCpp_TextSummary,
|
|
690
|
-
ToolCallingTask: LlamaCpp_ToolCalling,
|
|
691
|
-
StructuredGenerationTask: LlamaCpp_StructuredGeneration
|
|
692
|
-
};
|
|
693
|
-
var LLAMACPP_STREAM_TASKS = {
|
|
694
|
-
TextGenerationTask: LlamaCpp_TextGeneration_Stream,
|
|
695
|
-
TextRewriterTask: LlamaCpp_TextRewriter_Stream,
|
|
696
|
-
TextSummaryTask: LlamaCpp_TextSummary_Stream,
|
|
697
|
-
ToolCallingTask: LlamaCpp_ToolCalling_Stream,
|
|
698
|
-
StructuredGenerationTask: LlamaCpp_StructuredGeneration_Stream
|
|
699
|
-
};
|
|
700
|
-
var LLAMACPP_REACTIVE_TASKS = {
|
|
701
|
-
CountTokensTask: LlamaCpp_CountTokens_Reactive
|
|
702
|
-
};
|
|
703
|
-
// src/provider-llamacpp/LlamaCpp_Worker.ts
|
|
704
|
-
import { getLogger as getLogger2, globalServiceRegistry, WORKER_SERVER } from "@workglow/util";
|
|
705
|
-
function LLAMACPP_WORKER_JOBRUN_REGISTER() {
|
|
706
|
-
const workerServer = globalServiceRegistry.get(WORKER_SERVER);
|
|
707
|
-
new LlamaCppProvider(LLAMACPP_TASKS, LLAMACPP_STREAM_TASKS, LLAMACPP_REACTIVE_TASKS).registerOnWorkerServer(workerServer);
|
|
708
|
-
workerServer.sendReady();
|
|
709
|
-
getLogger2().info("LLAMACPP_WORKER_JOBRUN registered");
|
|
122
|
+
}
|
|
123
|
+
|
|
124
|
+
// src/provider-llamacpp/registerLlamaCpp.ts
|
|
125
|
+
async function registerLlamaCpp(options) {
|
|
126
|
+
await new LlamaCppQueuedProvider().register(options);
|
|
710
127
|
}
|
|
711
128
|
export {
|
|
712
|
-
|
|
713
|
-
LlamaCpp_Unload,
|
|
714
|
-
LlamaCpp_ToolCalling_Stream,
|
|
715
|
-
LlamaCpp_ToolCalling,
|
|
716
|
-
LlamaCpp_TextSummary_Stream,
|
|
717
|
-
LlamaCpp_TextSummary,
|
|
718
|
-
LlamaCpp_TextRewriter_Stream,
|
|
719
|
-
LlamaCpp_TextRewriter,
|
|
720
|
-
LlamaCpp_TextGeneration_Stream,
|
|
721
|
-
LlamaCpp_TextGeneration,
|
|
722
|
-
LlamaCpp_TextEmbedding,
|
|
723
|
-
LlamaCpp_StructuredGeneration_Stream,
|
|
724
|
-
LlamaCpp_StructuredGeneration,
|
|
725
|
-
LlamaCpp_ModelInfo,
|
|
726
|
-
LlamaCpp_Download,
|
|
727
|
-
LlamaCpp_CountTokens_Reactive,
|
|
728
|
-
LlamaCpp_CountTokens,
|
|
729
|
-
LlamaCppProvider,
|
|
129
|
+
registerLlamaCpp,
|
|
730
130
|
LlamaCppModelSchema,
|
|
731
131
|
LlamaCppModelRecordSchema,
|
|
732
132
|
LlamaCppModelConfigSchema,
|
|
733
133
|
LOCAL_LLAMACPP,
|
|
734
|
-
LLAMACPP_WORKER_JOBRUN_REGISTER,
|
|
735
|
-
LLAMACPP_TASKS,
|
|
736
|
-
LLAMACPP_STREAM_TASKS,
|
|
737
|
-
LLAMACPP_REACTIVE_TASKS,
|
|
738
134
|
LLAMACPP_DEFAULT_MODELS_DIR
|
|
739
135
|
};
|
|
740
136
|
|
|
741
|
-
//# debugId=
|
|
137
|
+
//# debugId=5DCB59F869EA8BFD64756E2164756E21
|