@workglow/ai-provider 0.0.121 → 0.0.122
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +30 -67
- package/dist/common/HfModelSearch.d.ts +32 -0
- package/dist/common/HfModelSearch.d.ts.map +1 -0
- package/dist/common/PipelineTaskMapping.d.ts +12 -0
- package/dist/common/PipelineTaskMapping.d.ts.map +1 -0
- package/dist/{anthropic → provider-anthropic}/AnthropicProvider.d.ts +2 -14
- package/dist/provider-anthropic/AnthropicProvider.d.ts.map +1 -0
- package/dist/provider-anthropic/AnthropicQueuedProvider.d.ts +16 -0
- package/dist/provider-anthropic/AnthropicQueuedProvider.d.ts.map +1 -0
- package/dist/provider-anthropic/common/Anthropic_Client.d.ts +13 -0
- package/dist/provider-anthropic/common/Anthropic_Client.d.ts.map +1 -0
- package/dist/provider-anthropic/common/Anthropic_Constants.d.ts.map +1 -0
- package/dist/provider-anthropic/common/Anthropic_CountTokens.d.ts +10 -0
- package/dist/provider-anthropic/common/Anthropic_CountTokens.d.ts.map +1 -0
- package/dist/provider-anthropic/common/Anthropic_JobRunFns.d.ts +12 -0
- package/dist/provider-anthropic/common/Anthropic_JobRunFns.d.ts.map +1 -0
- package/dist/provider-anthropic/common/Anthropic_ModelInfo.d.ts +9 -0
- package/dist/provider-anthropic/common/Anthropic_ModelInfo.d.ts.map +1 -0
- package/dist/{anthropic → provider-anthropic}/common/Anthropic_ModelSchema.d.ts +31 -31
- package/dist/provider-anthropic/common/Anthropic_ModelSchema.d.ts.map +1 -0
- package/dist/provider-anthropic/common/Anthropic_ModelSearch.d.ts +8 -0
- package/dist/provider-anthropic/common/Anthropic_ModelSearch.d.ts.map +1 -0
- package/dist/provider-anthropic/common/Anthropic_StructuredGeneration.d.ts +10 -0
- package/dist/provider-anthropic/common/Anthropic_StructuredGeneration.d.ts.map +1 -0
- package/dist/provider-anthropic/common/Anthropic_TextGeneration.d.ts +10 -0
- package/dist/provider-anthropic/common/Anthropic_TextGeneration.d.ts.map +1 -0
- package/dist/provider-anthropic/common/Anthropic_TextRewriter.d.ts +10 -0
- package/dist/provider-anthropic/common/Anthropic_TextRewriter.d.ts.map +1 -0
- package/dist/provider-anthropic/common/Anthropic_TextSummary.d.ts +10 -0
- package/dist/provider-anthropic/common/Anthropic_TextSummary.d.ts.map +1 -0
- package/dist/provider-anthropic/common/Anthropic_ToolCalling.d.ts +10 -0
- package/dist/provider-anthropic/common/Anthropic_ToolCalling.d.ts.map +1 -0
- package/dist/{anthropic → provider-anthropic}/index.d.ts +1 -3
- package/dist/provider-anthropic/index.d.ts.map +1 -0
- package/dist/{index-60ev6k93.js → provider-anthropic/index.js} +43 -11
- package/dist/provider-anthropic/index.js.map +13 -0
- package/dist/provider-anthropic/registerAnthropic.d.ts +10 -0
- package/dist/provider-anthropic/registerAnthropic.d.ts.map +1 -0
- package/dist/provider-anthropic/registerAnthropicInline.d.ts +8 -0
- package/dist/provider-anthropic/registerAnthropicInline.d.ts.map +1 -0
- package/dist/provider-anthropic/registerAnthropicWorker.d.ts +7 -0
- package/dist/provider-anthropic/registerAnthropicWorker.d.ts.map +1 -0
- package/dist/provider-anthropic/runtime.d.ts +16 -0
- package/dist/provider-anthropic/runtime.d.ts.map +1 -0
- package/dist/{anthropic/index.js → provider-anthropic/runtime.js} +291 -177
- package/dist/provider-anthropic/runtime.js.map +24 -0
- package/dist/{web-browser → provider-chrome}/WebBrowserProvider.d.ts +2 -15
- package/dist/provider-chrome/WebBrowserProvider.d.ts.map +1 -0
- package/dist/provider-chrome/WebBrowserQueuedProvider.d.ts +16 -0
- package/dist/provider-chrome/WebBrowserQueuedProvider.d.ts.map +1 -0
- package/dist/provider-chrome/common/WebBrowser_ChromeHelpers.d.ts +31 -0
- package/dist/provider-chrome/common/WebBrowser_ChromeHelpers.d.ts.map +1 -0
- package/dist/provider-chrome/common/WebBrowser_Constants.d.ts.map +1 -0
- package/dist/provider-chrome/common/WebBrowser_JobRunFns.d.ts +10 -0
- package/dist/provider-chrome/common/WebBrowser_JobRunFns.d.ts.map +1 -0
- package/dist/provider-chrome/common/WebBrowser_ModelInfo.d.ts +9 -0
- package/dist/provider-chrome/common/WebBrowser_ModelInfo.d.ts.map +1 -0
- package/dist/{web-browser → provider-chrome}/common/WebBrowser_ModelSchema.d.ts +31 -31
- package/dist/provider-chrome/common/WebBrowser_ModelSchema.d.ts.map +1 -0
- package/dist/provider-chrome/common/WebBrowser_ModelSearch.d.ts +8 -0
- package/dist/provider-chrome/common/WebBrowser_ModelSearch.d.ts.map +1 -0
- package/dist/provider-chrome/common/WebBrowser_TextGeneration.d.ts +10 -0
- package/dist/provider-chrome/common/WebBrowser_TextGeneration.d.ts.map +1 -0
- package/dist/provider-chrome/common/WebBrowser_TextLanguageDetection.d.ts +9 -0
- package/dist/provider-chrome/common/WebBrowser_TextLanguageDetection.d.ts.map +1 -0
- package/dist/provider-chrome/common/WebBrowser_TextRewriter.d.ts +10 -0
- package/dist/provider-chrome/common/WebBrowser_TextRewriter.d.ts.map +1 -0
- package/dist/provider-chrome/common/WebBrowser_TextSummary.d.ts +10 -0
- package/dist/provider-chrome/common/WebBrowser_TextSummary.d.ts.map +1 -0
- package/dist/provider-chrome/common/WebBrowser_TextTranslation.d.ts +10 -0
- package/dist/provider-chrome/common/WebBrowser_TextTranslation.d.ts.map +1 -0
- package/dist/{web-browser → provider-chrome}/index.d.ts +1 -3
- package/dist/provider-chrome/index.d.ts.map +1 -0
- package/dist/provider-chrome/index.js +132 -0
- package/dist/provider-chrome/index.js.map +13 -0
- package/dist/provider-chrome/registerWebBrowser.d.ts +10 -0
- package/dist/provider-chrome/registerWebBrowser.d.ts.map +1 -0
- package/dist/provider-chrome/registerWebBrowserInline.d.ts +8 -0
- package/dist/provider-chrome/registerWebBrowserInline.d.ts.map +1 -0
- package/dist/provider-chrome/registerWebBrowserWorker.d.ts +7 -0
- package/dist/provider-chrome/registerWebBrowserWorker.d.ts.map +1 -0
- package/dist/provider-chrome/runtime.d.ts +14 -0
- package/dist/provider-chrome/runtime.d.ts.map +1 -0
- package/dist/{web-browser/index.js → provider-chrome/runtime.js} +260 -235
- package/dist/provider-chrome/runtime.js.map +23 -0
- package/dist/{google-gemini → provider-gemini}/GoogleGeminiProvider.d.ts +2 -15
- package/dist/provider-gemini/GoogleGeminiProvider.d.ts.map +1 -0
- package/dist/provider-gemini/GoogleGeminiQueuedProvider.d.ts +16 -0
- package/dist/provider-gemini/GoogleGeminiQueuedProvider.d.ts.map +1 -0
- package/dist/provider-gemini/common/Gemini_Client.d.ts +10 -0
- package/dist/provider-gemini/common/Gemini_Client.d.ts.map +1 -0
- package/dist/provider-gemini/common/Gemini_Constants.d.ts.map +1 -0
- package/dist/provider-gemini/common/Gemini_CountTokens.d.ts +10 -0
- package/dist/provider-gemini/common/Gemini_CountTokens.d.ts.map +1 -0
- package/dist/provider-gemini/common/Gemini_JobRunFns.d.ts +13 -0
- package/dist/provider-gemini/common/Gemini_JobRunFns.d.ts.map +1 -0
- package/dist/provider-gemini/common/Gemini_ModelInfo.d.ts +9 -0
- package/dist/provider-gemini/common/Gemini_ModelInfo.d.ts.map +1 -0
- package/dist/{google-gemini → provider-gemini}/common/Gemini_ModelSchema.d.ts +31 -31
- package/dist/provider-gemini/common/Gemini_ModelSchema.d.ts.map +1 -0
- package/dist/provider-gemini/common/Gemini_ModelSearch.d.ts +8 -0
- package/dist/provider-gemini/common/Gemini_ModelSearch.d.ts.map +1 -0
- package/dist/provider-gemini/common/Gemini_Schema.d.ts +11 -0
- package/dist/provider-gemini/common/Gemini_Schema.d.ts.map +1 -0
- package/dist/provider-gemini/common/Gemini_StructuredGeneration.d.ts +10 -0
- package/dist/provider-gemini/common/Gemini_StructuredGeneration.d.ts.map +1 -0
- package/dist/provider-gemini/common/Gemini_TextEmbedding.d.ts +9 -0
- package/dist/provider-gemini/common/Gemini_TextEmbedding.d.ts.map +1 -0
- package/dist/provider-gemini/common/Gemini_TextGeneration.d.ts +10 -0
- package/dist/provider-gemini/common/Gemini_TextGeneration.d.ts.map +1 -0
- package/dist/provider-gemini/common/Gemini_TextRewriter.d.ts +10 -0
- package/dist/provider-gemini/common/Gemini_TextRewriter.d.ts.map +1 -0
- package/dist/provider-gemini/common/Gemini_TextSummary.d.ts +10 -0
- package/dist/provider-gemini/common/Gemini_TextSummary.d.ts.map +1 -0
- package/dist/provider-gemini/common/Gemini_ToolCalling.d.ts +10 -0
- package/dist/provider-gemini/common/Gemini_ToolCalling.d.ts.map +1 -0
- package/dist/{google-gemini → provider-gemini}/index.d.ts +1 -3
- package/dist/provider-gemini/index.d.ts.map +1 -0
- package/dist/{index-8651nz8y.js → provider-gemini/index.js} +43 -11
- package/dist/provider-gemini/index.js.map +13 -0
- package/dist/provider-gemini/registerGemini.d.ts +10 -0
- package/dist/provider-gemini/registerGemini.d.ts.map +1 -0
- package/dist/provider-gemini/registerGeminiInline.d.ts +8 -0
- package/dist/provider-gemini/registerGeminiInline.d.ts.map +1 -0
- package/dist/{anthropic/Anthropic_Worker.d.ts → provider-gemini/registerGeminiWorker.d.ts} +2 -2
- package/dist/provider-gemini/registerGeminiWorker.d.ts.map +1 -0
- package/dist/provider-gemini/runtime.d.ts +16 -0
- package/dist/provider-gemini/runtime.d.ts.map +1 -0
- package/dist/{google-gemini/index.js → provider-gemini/runtime.js} +281 -173
- package/dist/provider-gemini/runtime.js.map +26 -0
- package/dist/provider-hf-inference/HfInferenceProvider.d.ts +2 -19
- package/dist/provider-hf-inference/HfInferenceProvider.d.ts.map +1 -1
- package/dist/provider-hf-inference/HfInferenceQueuedProvider.d.ts +16 -0
- package/dist/provider-hf-inference/HfInferenceQueuedProvider.d.ts.map +1 -0
- package/dist/provider-hf-inference/common/HFI_Client.d.ts +12 -0
- package/dist/provider-hf-inference/common/HFI_Client.d.ts.map +1 -0
- package/dist/provider-hf-inference/common/HFI_JobRunFns.d.ts +2 -11
- package/dist/provider-hf-inference/common/HFI_JobRunFns.d.ts.map +1 -1
- package/dist/provider-hf-inference/common/HFI_ModelInfo.d.ts +9 -0
- package/dist/provider-hf-inference/common/HFI_ModelInfo.d.ts.map +1 -0
- package/dist/provider-hf-inference/common/HFI_ModelSchema.d.ts +31 -31
- package/dist/provider-hf-inference/common/HFI_ModelSchema.d.ts.map +1 -1
- package/dist/provider-hf-inference/common/HFI_ModelSearch.d.ts +8 -0
- package/dist/provider-hf-inference/common/HFI_ModelSearch.d.ts.map +1 -0
- package/dist/provider-hf-inference/common/HFI_TextEmbedding.d.ts +9 -0
- package/dist/provider-hf-inference/common/HFI_TextEmbedding.d.ts.map +1 -0
- package/dist/provider-hf-inference/common/HFI_TextGeneration.d.ts +10 -0
- package/dist/provider-hf-inference/common/HFI_TextGeneration.d.ts.map +1 -0
- package/dist/provider-hf-inference/common/HFI_TextRewriter.d.ts +10 -0
- package/dist/provider-hf-inference/common/HFI_TextRewriter.d.ts.map +1 -0
- package/dist/provider-hf-inference/common/HFI_TextSummary.d.ts +10 -0
- package/dist/provider-hf-inference/common/HFI_TextSummary.d.ts.map +1 -0
- package/dist/provider-hf-inference/common/HFI_ToolCalling.d.ts +10 -0
- package/dist/provider-hf-inference/common/HFI_ToolCalling.d.ts.map +1 -0
- package/dist/provider-hf-inference/index.d.ts +1 -3
- package/dist/provider-hf-inference/index.d.ts.map +1 -1
- package/dist/provider-hf-inference/index.js +98 -411
- package/dist/provider-hf-inference/index.js.map +7 -5
- package/dist/provider-hf-inference/registerHfInference.d.ts +10 -0
- package/dist/provider-hf-inference/registerHfInference.d.ts.map +1 -0
- package/dist/provider-hf-inference/registerHfInferenceInline.d.ts +8 -0
- package/dist/provider-hf-inference/registerHfInferenceInline.d.ts.map +1 -0
- package/dist/provider-hf-inference/registerHfInferenceWorker.d.ts +7 -0
- package/dist/provider-hf-inference/registerHfInferenceWorker.d.ts.map +1 -0
- package/dist/provider-hf-inference/runtime.d.ts +16 -0
- package/dist/provider-hf-inference/runtime.d.ts.map +1 -0
- package/dist/provider-hf-inference/runtime.js +592 -0
- package/dist/provider-hf-inference/runtime.js.map +25 -0
- package/dist/{hf-transformers → provider-hf-transformers}/HuggingFaceTransformersProvider.d.ts +2 -21
- package/dist/provider-hf-transformers/HuggingFaceTransformersProvider.d.ts.map +1 -0
- package/dist/provider-hf-transformers/HuggingFaceTransformersQueuedProvider.d.ts +16 -0
- package/dist/provider-hf-transformers/HuggingFaceTransformersQueuedProvider.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_BackgroundRemoval.d.ts +12 -0
- package/dist/provider-hf-transformers/common/HFT_BackgroundRemoval.d.ts.map +1 -0
- package/dist/{hf-transformers → provider-hf-transformers}/common/HFT_Constants.d.ts +25 -23
- package/dist/provider-hf-transformers/common/HFT_Constants.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_CountTokens.d.ts +10 -0
- package/dist/provider-hf-transformers/common/HFT_CountTokens.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_Download.d.ts +13 -0
- package/dist/provider-hf-transformers/common/HFT_Download.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_ImageClassification.d.ts +13 -0
- package/dist/provider-hf-transformers/common/HFT_ImageClassification.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_ImageEmbedding.d.ts +12 -0
- package/dist/provider-hf-transformers/common/HFT_ImageEmbedding.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_ImageHelpers.d.ts +11 -0
- package/dist/provider-hf-transformers/common/HFT_ImageHelpers.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_ImageSegmentation.d.ts +12 -0
- package/dist/provider-hf-transformers/common/HFT_ImageSegmentation.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_ImageToText.d.ts +12 -0
- package/dist/provider-hf-transformers/common/HFT_ImageToText.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_InlineLifecycle.d.ts +7 -0
- package/dist/provider-hf-transformers/common/HFT_InlineLifecycle.d.ts.map +1 -0
- package/dist/{hf-transformers → provider-hf-transformers}/common/HFT_JobRunFns.d.ts +673 -789
- package/dist/provider-hf-transformers/common/HFT_JobRunFns.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_ModelInfo.d.ts +9 -0
- package/dist/provider-hf-transformers/common/HFT_ModelInfo.d.ts.map +1 -0
- package/dist/{hf-transformers → provider-hf-transformers}/common/HFT_ModelSchema.d.ts +37 -37
- package/dist/provider-hf-transformers/common/HFT_ModelSchema.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_ModelSearch.d.ts +8 -0
- package/dist/provider-hf-transformers/common/HFT_ModelSearch.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_ObjectDetection.d.ts +13 -0
- package/dist/provider-hf-transformers/common/HFT_ObjectDetection.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_OnnxDtypes.d.ts +23 -0
- package/dist/provider-hf-transformers/common/HFT_OnnxDtypes.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_Pipeline.d.ts +32 -0
- package/dist/provider-hf-transformers/common/HFT_Pipeline.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_Streaming.d.ts +24 -0
- package/dist/provider-hf-transformers/common/HFT_Streaming.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_StructuredGeneration.d.ts +10 -0
- package/dist/provider-hf-transformers/common/HFT_StructuredGeneration.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_TextClassification.d.ts +9 -0
- package/dist/provider-hf-transformers/common/HFT_TextClassification.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_TextEmbedding.d.ts +13 -0
- package/dist/provider-hf-transformers/common/HFT_TextEmbedding.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_TextFillMask.d.ts +9 -0
- package/dist/provider-hf-transformers/common/HFT_TextFillMask.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_TextGeneration.d.ts +14 -0
- package/dist/provider-hf-transformers/common/HFT_TextGeneration.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_TextLanguageDetection.d.ts +9 -0
- package/dist/provider-hf-transformers/common/HFT_TextLanguageDetection.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_TextNamedEntityRecognition.d.ts +9 -0
- package/dist/provider-hf-transformers/common/HFT_TextNamedEntityRecognition.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_TextOutput.d.ts +8 -0
- package/dist/provider-hf-transformers/common/HFT_TextOutput.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_TextQuestionAnswer.d.ts +14 -0
- package/dist/provider-hf-transformers/common/HFT_TextQuestionAnswer.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_TextRewriter.d.ts +14 -0
- package/dist/provider-hf-transformers/common/HFT_TextRewriter.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_TextSummary.d.ts +14 -0
- package/dist/provider-hf-transformers/common/HFT_TextSummary.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_TextTranslation.d.ts +14 -0
- package/dist/provider-hf-transformers/common/HFT_TextTranslation.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_ToolCalling.d.ts +10 -0
- package/dist/provider-hf-transformers/common/HFT_ToolCalling.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_ToolMarkup.d.ts +40 -0
- package/dist/provider-hf-transformers/common/HFT_ToolMarkup.d.ts.map +1 -0
- package/dist/provider-hf-transformers/common/HFT_Unload.d.ts +13 -0
- package/dist/provider-hf-transformers/common/HFT_Unload.d.ts.map +1 -0
- package/dist/{hf-transformers → provider-hf-transformers}/index.d.ts +4 -2
- package/dist/provider-hf-transformers/index.d.ts.map +1 -0
- package/dist/provider-hf-transformers/index.js +513 -0
- package/dist/provider-hf-transformers/index.js.map +16 -0
- package/dist/provider-hf-transformers/registerHuggingFaceTransformers.d.ts +14 -0
- package/dist/provider-hf-transformers/registerHuggingFaceTransformers.d.ts.map +1 -0
- package/dist/provider-hf-transformers/registerHuggingFaceTransformersInline.d.ts +15 -0
- package/dist/provider-hf-transformers/registerHuggingFaceTransformersInline.d.ts.map +1 -0
- package/dist/provider-hf-transformers/registerHuggingFaceTransformersWorker.d.ts +7 -0
- package/dist/provider-hf-transformers/registerHuggingFaceTransformersWorker.d.ts.map +1 -0
- package/dist/provider-hf-transformers/runtime.d.ts +21 -0
- package/dist/provider-hf-transformers/runtime.d.ts.map +1 -0
- package/dist/{index-j4g81r4k.js → provider-hf-transformers/runtime.js} +1561 -927
- package/dist/provider-hf-transformers/runtime.js.map +49 -0
- package/dist/provider-llamacpp/LlamaCppProvider.d.ts +2 -15
- package/dist/provider-llamacpp/LlamaCppProvider.d.ts.map +1 -1
- package/dist/provider-llamacpp/LlamaCppQueuedProvider.d.ts +16 -0
- package/dist/provider-llamacpp/LlamaCppQueuedProvider.d.ts.map +1 -0
- package/dist/provider-llamacpp/common/LlamaCpp_CountTokens.d.ts +10 -0
- package/dist/provider-llamacpp/common/LlamaCpp_CountTokens.d.ts.map +1 -0
- package/dist/provider-llamacpp/common/LlamaCpp_Download.d.ts +9 -0
- package/dist/provider-llamacpp/common/LlamaCpp_Download.d.ts.map +1 -0
- package/dist/provider-llamacpp/common/LlamaCpp_JobRunFns.d.ts +2 -18
- package/dist/provider-llamacpp/common/LlamaCpp_JobRunFns.d.ts.map +1 -1
- package/dist/provider-llamacpp/common/LlamaCpp_ModelInfo.d.ts +9 -0
- package/dist/provider-llamacpp/common/LlamaCpp_ModelInfo.d.ts.map +1 -0
- package/dist/provider-llamacpp/common/LlamaCpp_ModelSchema.d.ts +31 -31
- package/dist/provider-llamacpp/common/LlamaCpp_ModelSchema.d.ts.map +1 -1
- package/dist/provider-llamacpp/common/LlamaCpp_ModelSearch.d.ts +8 -0
- package/dist/provider-llamacpp/common/LlamaCpp_ModelSearch.d.ts.map +1 -0
- package/dist/provider-llamacpp/common/LlamaCpp_Runtime.d.ts +31 -0
- package/dist/provider-llamacpp/common/LlamaCpp_Runtime.d.ts.map +1 -0
- package/dist/provider-llamacpp/common/LlamaCpp_StructuredGeneration.d.ts +10 -0
- package/dist/provider-llamacpp/common/LlamaCpp_StructuredGeneration.d.ts.map +1 -0
- package/dist/provider-llamacpp/common/LlamaCpp_TextEmbedding.d.ts +9 -0
- package/dist/provider-llamacpp/common/LlamaCpp_TextEmbedding.d.ts.map +1 -0
- package/dist/provider-llamacpp/common/LlamaCpp_TextGeneration.d.ts +10 -0
- package/dist/provider-llamacpp/common/LlamaCpp_TextGeneration.d.ts.map +1 -0
- package/dist/provider-llamacpp/common/LlamaCpp_TextRewriter.d.ts +10 -0
- package/dist/provider-llamacpp/common/LlamaCpp_TextRewriter.d.ts.map +1 -0
- package/dist/provider-llamacpp/common/LlamaCpp_TextSummary.d.ts +10 -0
- package/dist/provider-llamacpp/common/LlamaCpp_TextSummary.d.ts.map +1 -0
- package/dist/provider-llamacpp/common/LlamaCpp_ToolCalling.d.ts +10 -0
- package/dist/provider-llamacpp/common/LlamaCpp_ToolCalling.d.ts.map +1 -0
- package/dist/provider-llamacpp/common/LlamaCpp_Unload.d.ts +9 -0
- package/dist/provider-llamacpp/common/LlamaCpp_Unload.d.ts.map +1 -0
- package/dist/provider-llamacpp/index.d.ts +1 -3
- package/dist/provider-llamacpp/index.d.ts.map +1 -1
- package/dist/provider-llamacpp/index.js +121 -725
- package/dist/provider-llamacpp/index.js.map +7 -5
- package/dist/provider-llamacpp/registerLlamaCpp.d.ts +10 -0
- package/dist/provider-llamacpp/registerLlamaCpp.d.ts.map +1 -0
- package/dist/provider-llamacpp/registerLlamaCppInline.d.ts +8 -0
- package/dist/provider-llamacpp/registerLlamaCppInline.d.ts.map +1 -0
- package/dist/provider-llamacpp/registerLlamaCppWorker.d.ts +7 -0
- package/dist/provider-llamacpp/registerLlamaCppWorker.d.ts.map +1 -0
- package/dist/provider-llamacpp/runtime.d.ts +16 -0
- package/dist/provider-llamacpp/runtime.d.ts.map +1 -0
- package/dist/provider-llamacpp/runtime.js +929 -0
- package/dist/provider-llamacpp/runtime.js.map +29 -0
- package/dist/provider-ollama/OllamaProvider.d.ts +2 -15
- package/dist/provider-ollama/OllamaProvider.d.ts.map +1 -1
- package/dist/provider-ollama/OllamaQueuedProvider.d.ts +16 -0
- package/dist/provider-ollama/OllamaQueuedProvider.d.ts.map +1 -0
- package/dist/provider-ollama/common/Ollama_Client.browser.d.ts +13 -0
- package/dist/provider-ollama/common/Ollama_Client.browser.d.ts.map +1 -0
- package/dist/provider-ollama/common/Ollama_Client.d.ts +13 -0
- package/dist/provider-ollama/common/Ollama_Client.d.ts.map +1 -0
- package/dist/provider-ollama/common/Ollama_JobRunFns.browser.d.ts +362 -11
- package/dist/provider-ollama/common/Ollama_JobRunFns.browser.d.ts.map +1 -1
- package/dist/provider-ollama/common/Ollama_JobRunFns.d.ts +361 -11
- package/dist/provider-ollama/common/Ollama_JobRunFns.d.ts.map +1 -1
- package/dist/provider-ollama/common/Ollama_ModelInfo.d.ts +11 -0
- package/dist/provider-ollama/common/Ollama_ModelInfo.d.ts.map +1 -0
- package/dist/provider-ollama/common/Ollama_ModelSchema.d.ts +30 -30
- package/dist/provider-ollama/common/Ollama_ModelSchema.d.ts.map +1 -1
- package/dist/provider-ollama/common/Ollama_ModelSearch.d.ts +11 -0
- package/dist/provider-ollama/common/Ollama_ModelSearch.d.ts.map +1 -0
- package/dist/provider-ollama/common/Ollama_ModelUtil.d.ts +8 -0
- package/dist/provider-ollama/common/Ollama_ModelUtil.d.ts.map +1 -0
- package/dist/provider-ollama/common/Ollama_TextEmbedding.d.ts +11 -0
- package/dist/provider-ollama/common/Ollama_TextEmbedding.d.ts.map +1 -0
- package/dist/provider-ollama/common/Ollama_TextGeneration.d.ts +12 -0
- package/dist/provider-ollama/common/Ollama_TextGeneration.d.ts.map +1 -0
- package/dist/provider-ollama/common/Ollama_TextRewriter.d.ts +12 -0
- package/dist/provider-ollama/common/Ollama_TextRewriter.d.ts.map +1 -0
- package/dist/provider-ollama/common/Ollama_TextSummary.d.ts +12 -0
- package/dist/provider-ollama/common/Ollama_TextSummary.d.ts.map +1 -0
- package/dist/provider-ollama/common/Ollama_ToolCalling.d.ts +16 -0
- package/dist/provider-ollama/common/Ollama_ToolCalling.d.ts.map +1 -0
- package/dist/provider-ollama/index.browser.d.ts +1 -3
- package/dist/provider-ollama/index.browser.d.ts.map +1 -1
- package/dist/provider-ollama/index.browser.js +18 -396
- package/dist/provider-ollama/index.browser.js.map +6 -7
- package/dist/provider-ollama/index.d.ts +1 -3
- package/dist/provider-ollama/index.d.ts.map +1 -1
- package/dist/provider-ollama/index.js +93 -382
- package/dist/provider-ollama/index.js.map +7 -5
- package/dist/provider-ollama/registerOllama.d.ts +10 -0
- package/dist/provider-ollama/registerOllama.d.ts.map +1 -0
- package/dist/provider-ollama/registerOllamaInline.browser.d.ts +8 -0
- package/dist/provider-ollama/registerOllamaInline.browser.d.ts.map +1 -0
- package/dist/provider-ollama/registerOllamaInline.d.ts +8 -0
- package/dist/provider-ollama/registerOllamaInline.d.ts.map +1 -0
- package/dist/provider-ollama/registerOllamaWorker.browser.d.ts +7 -0
- package/dist/provider-ollama/registerOllamaWorker.browser.d.ts.map +1 -0
- package/dist/{google-gemini/Gemini_Worker.d.ts → provider-ollama/registerOllamaWorker.d.ts} +2 -2
- package/dist/provider-ollama/registerOllamaWorker.d.ts.map +1 -0
- package/dist/provider-ollama/runtime.browser.d.ts +16 -0
- package/dist/provider-ollama/runtime.browser.d.ts.map +1 -0
- package/dist/provider-ollama/runtime.browser.js +528 -0
- package/dist/provider-ollama/runtime.browser.js.map +24 -0
- package/dist/provider-ollama/runtime.d.ts +16 -0
- package/dist/provider-ollama/runtime.d.ts.map +1 -0
- package/dist/provider-ollama/runtime.js +538 -0
- package/dist/provider-ollama/runtime.js.map +24 -0
- package/dist/provider-openai/OpenAiProvider.d.ts +2 -19
- package/dist/provider-openai/OpenAiProvider.d.ts.map +1 -1
- package/dist/provider-openai/OpenAiQueuedProvider.d.ts +16 -0
- package/dist/provider-openai/OpenAiQueuedProvider.d.ts.map +1 -0
- package/dist/provider-openai/common/OpenAI_Client.d.ts +10 -0
- package/dist/provider-openai/common/OpenAI_Client.d.ts.map +1 -0
- package/dist/provider-openai/common/OpenAI_CountTokens.d.ts +10 -0
- package/dist/provider-openai/common/OpenAI_CountTokens.d.ts.map +1 -0
- package/dist/provider-openai/common/OpenAI_JobRunFns.d.ts +2 -15
- package/dist/provider-openai/common/OpenAI_JobRunFns.d.ts.map +1 -1
- package/dist/provider-openai/common/OpenAI_ModelInfo.d.ts +9 -0
- package/dist/provider-openai/common/OpenAI_ModelInfo.d.ts.map +1 -0
- package/dist/provider-openai/common/OpenAI_ModelSchema.d.ts +31 -31
- package/dist/provider-openai/common/OpenAI_ModelSchema.d.ts.map +1 -1
- package/dist/provider-openai/common/OpenAI_ModelSearch.d.ts +8 -0
- package/dist/provider-openai/common/OpenAI_ModelSearch.d.ts.map +1 -0
- package/dist/provider-openai/common/OpenAI_StructuredGeneration.d.ts +10 -0
- package/dist/provider-openai/common/OpenAI_StructuredGeneration.d.ts.map +1 -0
- package/dist/provider-openai/common/OpenAI_TextEmbedding.d.ts +9 -0
- package/dist/provider-openai/common/OpenAI_TextEmbedding.d.ts.map +1 -0
- package/dist/provider-openai/common/OpenAI_TextGeneration.d.ts +10 -0
- package/dist/provider-openai/common/OpenAI_TextGeneration.d.ts.map +1 -0
- package/dist/provider-openai/common/OpenAI_TextRewriter.d.ts +10 -0
- package/dist/provider-openai/common/OpenAI_TextRewriter.d.ts.map +1 -0
- package/dist/provider-openai/common/OpenAI_TextSummary.d.ts +10 -0
- package/dist/provider-openai/common/OpenAI_TextSummary.d.ts.map +1 -0
- package/dist/provider-openai/common/OpenAI_ToolCalling.d.ts +10 -0
- package/dist/provider-openai/common/OpenAI_ToolCalling.d.ts.map +1 -0
- package/dist/provider-openai/index.d.ts +1 -3
- package/dist/provider-openai/index.d.ts.map +1 -1
- package/dist/provider-openai/index.js +108 -519
- package/dist/provider-openai/index.js.map +7 -5
- package/dist/provider-openai/registerOpenAi.d.ts +10 -0
- package/dist/provider-openai/registerOpenAi.d.ts.map +1 -0
- package/dist/provider-openai/registerOpenAiInline.d.ts +8 -0
- package/dist/provider-openai/registerOpenAiInline.d.ts.map +1 -0
- package/dist/{ggml/model/GgmlLocalModel.d.ts → provider-openai/registerOpenAiWorker.d.ts} +2 -2
- package/dist/provider-openai/registerOpenAiWorker.d.ts.map +1 -0
- package/dist/provider-openai/runtime.d.ts +16 -0
- package/dist/provider-openai/runtime.d.ts.map +1 -0
- package/dist/provider-openai/runtime.js +662 -0
- package/dist/provider-openai/runtime.js.map +25 -0
- package/dist/provider-tf-mediapipe/TensorFlowMediaPipeProvider.d.ts +24 -0
- package/dist/provider-tf-mediapipe/TensorFlowMediaPipeProvider.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/TensorFlowMediaPipeQueuedProvider.d.ts +16 -0
- package/dist/provider-tf-mediapipe/TensorFlowMediaPipeQueuedProvider.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_Client.d.ts +8 -0
- package/dist/provider-tf-mediapipe/common/TFMP_Client.d.ts.map +1 -0
- package/dist/{tf-mediapipe → provider-tf-mediapipe}/common/TFMP_Constants.d.ts +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_Constants.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_Download.d.ts +9 -0
- package/dist/provider-tf-mediapipe/common/TFMP_Download.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_FaceDetector.d.ts +9 -0
- package/dist/provider-tf-mediapipe/common/TFMP_FaceDetector.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_FaceLandmarker.d.ts +9 -0
- package/dist/provider-tf-mediapipe/common/TFMP_FaceLandmarker.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_GestureRecognizer.d.ts +9 -0
- package/dist/provider-tf-mediapipe/common/TFMP_GestureRecognizer.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_HandLandmarker.d.ts +9 -0
- package/dist/provider-tf-mediapipe/common/TFMP_HandLandmarker.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_ImageClassification.d.ts +9 -0
- package/dist/provider-tf-mediapipe/common/TFMP_ImageClassification.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_ImageEmbedding.d.ts +9 -0
- package/dist/provider-tf-mediapipe/common/TFMP_ImageEmbedding.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_ImageSegmentation.d.ts +9 -0
- package/dist/provider-tf-mediapipe/common/TFMP_ImageSegmentation.d.ts.map +1 -0
- package/dist/{tf-mediapipe → provider-tf-mediapipe}/common/TFMP_JobRunFns.d.ts +340 -408
- package/dist/provider-tf-mediapipe/common/TFMP_JobRunFns.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_ModelInfo.d.ts +9 -0
- package/dist/provider-tf-mediapipe/common/TFMP_ModelInfo.d.ts.map +1 -0
- package/dist/{tf-mediapipe → provider-tf-mediapipe}/common/TFMP_ModelSchema.d.ts +40 -37
- package/dist/provider-tf-mediapipe/common/TFMP_ModelSchema.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_ModelSearch.d.ts +9 -0
- package/dist/provider-tf-mediapipe/common/TFMP_ModelSearch.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_ObjectDetection.d.ts +9 -0
- package/dist/provider-tf-mediapipe/common/TFMP_ObjectDetection.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_PoseLandmarker.d.ts +9 -0
- package/dist/provider-tf-mediapipe/common/TFMP_PoseLandmarker.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_Runtime.d.ts +43 -0
- package/dist/provider-tf-mediapipe/common/TFMP_Runtime.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_TextClassification.d.ts +9 -0
- package/dist/provider-tf-mediapipe/common/TFMP_TextClassification.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_TextEmbedding.d.ts +9 -0
- package/dist/provider-tf-mediapipe/common/TFMP_TextEmbedding.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_TextLanguageDetection.d.ts +9 -0
- package/dist/provider-tf-mediapipe/common/TFMP_TextLanguageDetection.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/common/TFMP_Unload.d.ts +9 -0
- package/dist/provider-tf-mediapipe/common/TFMP_Unload.d.ts.map +1 -0
- package/dist/{tf-mediapipe → provider-tf-mediapipe}/index.d.ts +1 -3
- package/dist/provider-tf-mediapipe/index.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/index.js +129 -0
- package/dist/provider-tf-mediapipe/index.js.map +13 -0
- package/dist/provider-tf-mediapipe/registerTensorFlowMediaPipe.d.ts +10 -0
- package/dist/provider-tf-mediapipe/registerTensorFlowMediaPipe.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/registerTensorFlowMediaPipeInline.d.ts +8 -0
- package/dist/provider-tf-mediapipe/registerTensorFlowMediaPipeInline.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/registerTensorFlowMediaPipeWorker.d.ts +7 -0
- package/dist/provider-tf-mediapipe/registerTensorFlowMediaPipeWorker.d.ts.map +1 -0
- package/dist/provider-tf-mediapipe/runtime.d.ts +16 -0
- package/dist/provider-tf-mediapipe/runtime.d.ts.map +1 -0
- package/dist/{tf-mediapipe/index.js → provider-tf-mediapipe/runtime.js} +417 -380
- package/dist/provider-tf-mediapipe/runtime.js.map +33 -0
- package/package.json +66 -40
- package/dist/HFT_JobRunFns-8hcpea4c.js +0 -80
- package/dist/HFT_JobRunFns-8hcpea4c.js.map +0 -9
- package/dist/anthropic/AnthropicProvider.d.ts.map +0 -1
- package/dist/anthropic/Anthropic_Worker.d.ts.map +0 -1
- package/dist/anthropic/common/Anthropic_Constants.d.ts.map +0 -1
- package/dist/anthropic/common/Anthropic_JobRunFns.d.ts +0 -24
- package/dist/anthropic/common/Anthropic_JobRunFns.d.ts.map +0 -1
- package/dist/anthropic/common/Anthropic_ModelSchema.d.ts.map +0 -1
- package/dist/anthropic/index.d.ts.map +0 -1
- package/dist/anthropic/index.js.map +0 -11
- package/dist/ggml/model/GgmlLocalModel.d.ts.map +0 -1
- package/dist/google-gemini/Gemini_Worker.d.ts.map +0 -1
- package/dist/google-gemini/GoogleGeminiProvider.d.ts.map +0 -1
- package/dist/google-gemini/common/Gemini_Constants.d.ts.map +0 -1
- package/dist/google-gemini/common/Gemini_JobRunFns.d.ts +0 -25
- package/dist/google-gemini/common/Gemini_JobRunFns.d.ts.map +0 -1
- package/dist/google-gemini/common/Gemini_ModelSchema.d.ts.map +0 -1
- package/dist/google-gemini/index.d.ts.map +0 -1
- package/dist/google-gemini/index.js.map +0 -11
- package/dist/hf-transformers/HFT_Worker.d.ts +0 -7
- package/dist/hf-transformers/HFT_Worker.d.ts.map +0 -1
- package/dist/hf-transformers/HuggingFaceTransformersProvider.d.ts.map +0 -1
- package/dist/hf-transformers/common/HFT_Constants.d.ts.map +0 -1
- package/dist/hf-transformers/common/HFT_JobRunFns.d.ts.map +0 -1
- package/dist/hf-transformers/common/HFT_ModelSchema.d.ts.map +0 -1
- package/dist/hf-transformers/index.d.ts.map +0 -1
- package/dist/hf-transformers/index.js +0 -116
- package/dist/hf-transformers/index.js.map +0 -10
- package/dist/index-60ev6k93.js.map +0 -12
- package/dist/index-6j5pq722.js +0 -11
- package/dist/index-6j5pq722.js.map +0 -9
- package/dist/index-8651nz8y.js.map +0 -12
- package/dist/index-j4g81r4k.js.map +0 -10
- package/dist/index-pkd79j8b.js +0 -58
- package/dist/index-pkd79j8b.js.map +0 -10
- package/dist/index-q2t627d5.js +0 -88
- package/dist/index-q2t627d5.js.map +0 -12
- package/dist/index-tp5s7355.js +0 -77
- package/dist/index-tp5s7355.js.map +0 -12
- package/dist/index-v72vr07f.js +0 -81
- package/dist/index-v72vr07f.js.map +0 -12
- package/dist/index-wr57rwyx.js +0 -104
- package/dist/index-wr57rwyx.js.map +0 -12
- package/dist/index-zqq3kw0n.js +0 -171
- package/dist/index-zqq3kw0n.js.map +0 -11
- package/dist/index.browser-6j5pq722.js +0 -11
- package/dist/index.browser-6j5pq722.js.map +0 -9
- package/dist/index.d.ts +0 -33
- package/dist/index.d.ts.map +0 -1
- package/dist/index.js +0 -316
- package/dist/index.js.map +0 -15
- package/dist/provider-hf-inference/HFI_Worker.d.ts +0 -7
- package/dist/provider-hf-inference/HFI_Worker.d.ts.map +0 -1
- package/dist/provider-llamacpp/LlamaCpp_Worker.d.ts +0 -7
- package/dist/provider-llamacpp/LlamaCpp_Worker.d.ts.map +0 -1
- package/dist/provider-ollama/Ollama_Worker.browser.d.ts +0 -7
- package/dist/provider-ollama/Ollama_Worker.browser.d.ts.map +0 -1
- package/dist/provider-ollama/Ollama_Worker.d.ts +0 -7
- package/dist/provider-ollama/Ollama_Worker.d.ts.map +0 -1
- package/dist/provider-openai/OpenAI_Worker.d.ts +0 -7
- package/dist/provider-openai/OpenAI_Worker.d.ts.map +0 -1
- package/dist/tf-mediapipe/TFMP_Worker.d.ts +0 -7
- package/dist/tf-mediapipe/TFMP_Worker.d.ts.map +0 -1
- package/dist/tf-mediapipe/TensorFlowMediaPipeProvider.d.ts +0 -41
- package/dist/tf-mediapipe/TensorFlowMediaPipeProvider.d.ts.map +0 -1
- package/dist/tf-mediapipe/common/TFMP_Constants.d.ts.map +0 -1
- package/dist/tf-mediapipe/common/TFMP_JobRunFns.d.ts.map +0 -1
- package/dist/tf-mediapipe/common/TFMP_ModelSchema.d.ts.map +0 -1
- package/dist/tf-mediapipe/index.d.ts.map +0 -1
- package/dist/tf-mediapipe/index.js.map +0 -14
- package/dist/types.d.ts +0 -7
- package/dist/types.d.ts.map +0 -1
- package/dist/web-browser/WebBrowserProvider.d.ts.map +0 -1
- package/dist/web-browser/WebBrowser_Worker.d.ts +0 -7
- package/dist/web-browser/WebBrowser_Worker.d.ts.map +0 -1
- package/dist/web-browser/common/WebBrowser_Constants.d.ts.map +0 -1
- package/dist/web-browser/common/WebBrowser_JobRunFns.d.ts +0 -20
- package/dist/web-browser/common/WebBrowser_JobRunFns.d.ts.map +0 -1
- package/dist/web-browser/common/WebBrowser_ModelSchema.d.ts.map +0 -1
- package/dist/web-browser/index.d.ts.map +0 -1
- package/dist/web-browser/index.js.map +0 -14
- /package/dist/{anthropic → provider-anthropic}/common/Anthropic_Constants.d.ts +0 -0
- /package/dist/{web-browser → provider-chrome}/common/WebBrowser_Constants.d.ts +0 -0
- /package/dist/{google-gemini → provider-gemini}/common/Gemini_Constants.d.ts +0 -0
|
@@ -1,11 +1,13 @@
|
|
|
1
1
|
{
|
|
2
2
|
"version": 3,
|
|
3
|
-
"sources": ["../src/provider-llamacpp/common/
|
|
3
|
+
"sources": ["../src/provider-llamacpp/common/LlamaCpp_Constants.ts", "../src/provider-llamacpp/common/LlamaCpp_ModelSchema.ts", "../src/provider-llamacpp/LlamaCppQueuedProvider.ts", "../src/provider-llamacpp/registerLlamaCpp.ts"],
|
|
4
4
|
"sourcesContent": [
|
|
5
|
-
"/**\n * @license\n * Copyright 2025 Steven Roussey <sroussey@gmail.com>\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { filterValidToolCalls } from \"@workglow/ai\";\nimport type {\n AiProviderReactiveRunFn,\n AiProviderRunFn,\n AiProviderStreamFn,\n CountTokensTaskInput,\n CountTokensTaskOutput,\n DownloadModelTaskRunInput,\n DownloadModelTaskRunOutput,\n ModelInfoTaskInput,\n ModelInfoTaskOutput,\n StructuredGenerationTaskInput,\n StructuredGenerationTaskOutput,\n TextEmbeddingTaskInput,\n TextEmbeddingTaskOutput,\n TextGenerationTaskInput,\n TextGenerationTaskOutput,\n TextRewriterTaskInput,\n TextRewriterTaskOutput,\n TextSummaryTaskInput,\n TextSummaryTaskOutput,\n ToolCallingTaskInput,\n ToolCallingTaskOutput,\n ToolCalls,\n ToolDefinition,\n UnloadModelTaskRunInput,\n UnloadModelTaskRunOutput,\n} from \"@workglow/ai\";\nimport type { StreamEvent } from \"@workglow/task-graph\";\nimport { getLogger, parsePartialJson } from \"@workglow/util\";\nimport { LLAMACPP_DEFAULT_MODELS_DIR } from \"./LlamaCpp_Constants\";\nimport type { LlamaCppModelConfig } from \"./LlamaCpp_ModelSchema\";\n\n// ========================================================================\n// Lazy SDK loading\n// ========================================================================\n\nlet _sdk: typeof import(\"node-llama-cpp\") | undefined;\n\nasync function loadSdk() {\n if (!_sdk) {\n try {\n _sdk = await import(\"node-llama-cpp\");\n } catch (err) {\n throw new Error(\n \"node-llama-cpp is required for LOCAL_LLAMACPP tasks. Install it with: bun add node-llama-cpp\"\n );\n }\n }\n return _sdk;\n}\n\n// ========================================================================\n// Module-level caches (shared across all task invocations)\n// ========================================================================\n\ntype LlamaInstance = Awaited<ReturnType<(typeof import(\"node-llama-cpp\"))[\"getLlama\"]>>;\ntype LlamaModel = Awaited<ReturnType<LlamaInstance[\"loadModel\"]>>;\ntype LlamaContext = Awaited<ReturnType<LlamaModel[\"createContext\"]>>;\ntype LlamaEmbeddingContext = Awaited<ReturnType<LlamaModel[\"createEmbeddingContext\"]>>;\n\nlet llamaInstance: LlamaInstance | undefined;\nconst models = new Map<string, LlamaModel>();\nconst textContexts = new Map<string, LlamaContext>();\nconst embeddingContexts = new Map<string, LlamaEmbeddingContext>();\n\n/** Maps model_url (or model_path when used as URI) to the actual downloaded filesystem path. */\nconst resolvedPaths = new Map<string, string>();\n\n// ========================================================================\n// Helpers\n// ========================================================================\n\nasync function getLlamaInstance(): Promise<LlamaInstance> {\n if (!llamaInstance) {\n const { getLlama } = await loadSdk();\n llamaInstance = await getLlama();\n }\n return llamaInstance;\n}\n\nfunction getConfigKey(model: LlamaCppModelConfig): string {\n return model.provider_config.model_url ?? model.provider_config.model_path;\n}\n\nfunction getActualModelPath(model: LlamaCppModelConfig): string {\n const key = getConfigKey(model);\n const resolved = resolvedPaths.get(key);\n return resolved ?? model.provider_config.model_path;\n}\n\nasync function getOrLoadModel(model: LlamaCppModelConfig): Promise<LlamaModel> {\n const modelPath = getActualModelPath(model);\n const cached = models.get(modelPath);\n if (cached) return cached;\n\n const llama = await getLlamaInstance();\n const config = model.provider_config;\n\n const loadedModel = await llama.loadModel({\n modelPath,\n ...(config.gpu_layers !== undefined && { gpuLayers: config.gpu_layers }),\n });\n\n models.set(modelPath, loadedModel);\n return loadedModel;\n}\n\nasync function getOrCreateTextContext(model: LlamaCppModelConfig): Promise<LlamaContext> {\n const modelPath = getActualModelPath(model);\n const cached = textContexts.get(modelPath);\n if (cached) return cached;\n\n const loadedModel = await getOrLoadModel(model);\n const config = model.provider_config;\n\n const context = await loadedModel.createContext({\n ...(config.context_size && { contextSize: config.context_size }),\n ...(config.flash_attention !== undefined && { flashAttention: config.flash_attention }),\n });\n\n textContexts.set(modelPath, context);\n return context;\n}\n\nasync function getOrCreateEmbeddingContext(\n model: LlamaCppModelConfig\n): Promise<LlamaEmbeddingContext> {\n const modelPath = getActualModelPath(model);\n const cached = embeddingContexts.get(modelPath);\n if (cached) return cached;\n\n const loadedModel = await getOrLoadModel(model);\n\n const context = await loadedModel.createEmbeddingContext();\n\n embeddingContexts.set(modelPath, context);\n return context;\n}\n\n/**\n * Bridges a node-llama-cpp session.prompt() call (callback-based) to an AsyncGenerator\n * of StreamEvents. Handles abort signals and errors cleanly.\n */\nasync function* streamFromSession<T extends Record<string, unknown>>(\n promptFn: (onTextChunk: (chunk: string) => void) => Promise<string>,\n signal: AbortSignal\n): AsyncGenerator<StreamEvent<T>> {\n const queue: string[] = [];\n let isComplete = false;\n let completionError: unknown;\n let resolveWait: (() => void) | null = null;\n\n const notifyWaiter = () => {\n resolveWait?.();\n resolveWait = null;\n };\n\n const promptPromise = promptFn((chunk: string) => {\n queue.push(chunk);\n notifyWaiter();\n })\n .then(() => {\n isComplete = true;\n notifyWaiter();\n })\n .catch((err: unknown) => {\n completionError = err;\n isComplete = true;\n notifyWaiter();\n });\n\n try {\n while (true) {\n while (queue.length > 0) {\n yield { type: \"text-delta\", port: \"text\", textDelta: queue.shift()! };\n }\n if (isComplete) break;\n await new Promise<void>((r) => {\n resolveWait = r;\n });\n }\n // Drain any remaining chunks after completion signal\n while (queue.length > 0) {\n yield { type: \"text-delta\", port: \"text\", textDelta: queue.shift()! };\n }\n } finally {\n await promptPromise.catch(() => {});\n }\n\n if (completionError) {\n if (signal.aborted) return;\n throw completionError;\n }\n\n yield { type: \"finish\", data: {} as T };\n}\n\n// ========================================================================\n// DownloadModelTask\n// ========================================================================\n\nexport const LlamaCpp_Download: AiProviderRunFn<\n DownloadModelTaskRunInput,\n DownloadModelTaskRunOutput,\n LlamaCppModelConfig\n> = async (input, model, update_progress, signal) => {\n if (!model) throw new Error(\"Model config is required for DownloadModelTask.\");\n\n const { createModelDownloader } = await loadSdk();\n const config = model.provider_config;\n const modelUri = config.model_url ?? config.model_path;\n const dirPath = config.models_dir ?? LLAMACPP_DEFAULT_MODELS_DIR;\n\n update_progress(0, \"Creating model downloader\");\n\n const downloader = await createModelDownloader({ modelUri, dirPath });\n\n // Poll download progress via interval (ModelDownloader exposes downloadedSize / totalSize)\n const progressInterval = setInterval(() => {\n const total = (downloader as any).totalSize as number | undefined;\n const downloaded = (downloader as any).downloadedSize as number | undefined;\n if (total && total > 0 && downloaded !== undefined) {\n const pct = Math.min(99, Math.round((downloaded / total) * 100));\n update_progress(pct, \"Downloading model\", { file: modelUri, progress: pct / 100 });\n }\n }, 500);\n\n let modelPath: string;\n try {\n modelPath = await downloader.download();\n } finally {\n clearInterval(progressInterval);\n }\n\n // Store the resolved filesystem path for subsequent inference tasks\n resolvedPaths.set(getConfigKey(model), modelPath);\n\n update_progress(100, \"Model downloaded\", { file: modelUri, progress: 1 });\n\n return { model: input.model! };\n};\n\n// ========================================================================\n// UnloadModelTask\n// ========================================================================\n\nexport const LlamaCpp_Unload: AiProviderRunFn<\n UnloadModelTaskRunInput,\n UnloadModelTaskRunOutput,\n LlamaCppModelConfig\n> = async (input, model, update_progress, _signal) => {\n if (!model) throw new Error(\"Model config is required for UnloadModelTask.\");\n\n const modelPath = getActualModelPath(model);\n\n // Dispose and remove the text context\n const ctx = textContexts.get(modelPath);\n if (ctx) {\n await ctx.dispose();\n textContexts.delete(modelPath);\n update_progress(33, \"Text context disposed\");\n }\n\n // Dispose and remove the embedding context\n const embCtx = embeddingContexts.get(modelPath);\n if (embCtx) {\n await embCtx.dispose();\n embeddingContexts.delete(modelPath);\n update_progress(66, \"Embedding context disposed\");\n }\n\n // Dispose and remove the model\n const cachedModel = models.get(modelPath);\n if (cachedModel) {\n await cachedModel.dispose();\n models.delete(modelPath);\n update_progress(100, \"Model unloaded from memory\");\n } else {\n update_progress(100, \"Model was not loaded\");\n }\n\n return { model: input.model! };\n};\n\n// ========================================================================\n// TextGenerationTask (non-streaming)\n// ========================================================================\n\nexport const LlamaCpp_TextGeneration: AiProviderRunFn<\n TextGenerationTaskInput,\n TextGenerationTaskOutput,\n LlamaCppModelConfig\n> = async (input, model, update_progress, signal) => {\n if (Array.isArray(input.prompt)) {\n getLogger().warn(\n \"LlamaCpp_TextGeneration: array input received; processing sequentially (no native batch support)\"\n );\n const prompts = input.prompt as string[];\n const results: string[] = [];\n for (const item of prompts) {\n const r = await LlamaCpp_TextGeneration(\n { ...input, prompt: item },\n model,\n update_progress,\n signal\n );\n results.push(r.text as string);\n }\n return { text: results };\n }\n\n if (!model) throw new Error(\"Model config is required for TextGenerationTask.\");\n\n const { LlamaChatSession } = await loadSdk();\n\n update_progress(0, \"Loading model\");\n const context = await getOrCreateTextContext(model);\n\n update_progress(10, \"Generating text\");\n const sequence = context.getSequence();\n const session = new LlamaChatSession({ contextSequence: sequence });\n try {\n const text = await session.prompt(input.prompt as string, {\n signal,\n ...(input.temperature !== undefined && { temperature: input.temperature }),\n ...(input.maxTokens !== undefined && { maxTokens: input.maxTokens }),\n ...(input.topP !== undefined && { topP: input.topP }),\n });\n update_progress(100, \"Text generation complete\");\n return { text };\n } finally {\n sequence.dispose();\n }\n};\n\n// ========================================================================\n// TextGenerationTask (streaming)\n// ========================================================================\n\nexport const LlamaCpp_TextGeneration_Stream: AiProviderStreamFn<\n TextGenerationTaskInput,\n TextGenerationTaskOutput,\n LlamaCppModelConfig\n> = async function* (input, model, signal): AsyncIterable<StreamEvent<TextGenerationTaskOutput>> {\n if (!model) throw new Error(\"Model config is required for TextGenerationTask.\");\n\n const { LlamaChatSession } = await loadSdk();\n\n const context = await getOrCreateTextContext(model);\n const sequence = context.getSequence();\n const session = new LlamaChatSession({ contextSequence: sequence });\n try {\n yield* streamFromSession<TextGenerationTaskOutput>((onTextChunk) => {\n return session.prompt(input.prompt as string, {\n signal,\n onTextChunk,\n ...(input.temperature !== undefined && { temperature: input.temperature }),\n ...(input.maxTokens !== undefined && { maxTokens: input.maxTokens }),\n ...(input.topP !== undefined && { topP: input.topP }),\n });\n }, signal);\n } finally {\n sequence.dispose();\n }\n};\n\n// ========================================================================\n// TextEmbeddingTask\n// ========================================================================\n\nexport const LlamaCpp_TextEmbedding: AiProviderRunFn<\n TextEmbeddingTaskInput,\n TextEmbeddingTaskOutput,\n LlamaCppModelConfig\n> = async (input, model, update_progress, _signal) => {\n if (!model) throw new Error(\"Model config is required for TextEmbeddingTask.\");\n\n update_progress(0, \"Loading embedding model\");\n const context = await getOrCreateEmbeddingContext(model);\n\n const texts = Array.isArray(input.text) ? input.text : [input.text];\n update_progress(10, \"Computing embeddings\");\n\n const embeddings = await Promise.all(\n texts.map((text) => context.getEmbeddingFor(text).then((e) => new Float32Array(e.vector)))\n );\n\n update_progress(100, \"Embeddings complete\");\n\n if (Array.isArray(input.text)) {\n return { vector: embeddings };\n }\n return { vector: embeddings[0] };\n};\n\n// ========================================================================\n// TextRewriterTask (non-streaming)\n// ========================================================================\n\nexport const LlamaCpp_TextRewriter: AiProviderRunFn<\n TextRewriterTaskInput,\n TextRewriterTaskOutput,\n LlamaCppModelConfig\n> = async (input, model, update_progress, signal) => {\n if (Array.isArray(input.text)) {\n getLogger().warn(\n \"LlamaCpp_TextRewriter: array input received; processing sequentially (no native batch support)\"\n );\n const texts = input.text as string[];\n const results: string[] = [];\n for (const item of texts) {\n const r = await LlamaCpp_TextRewriter(\n { ...input, text: item },\n model,\n update_progress,\n signal\n );\n results.push(r.text as string);\n }\n return { text: results };\n }\n\n if (!model) throw new Error(\"Model config is required for TextRewriterTask.\");\n\n const { LlamaChatSession } = await loadSdk();\n\n update_progress(0, \"Loading model\");\n const context = await getOrCreateTextContext(model);\n\n update_progress(10, \"Rewriting text\");\n const sequence = context.getSequence();\n const session = new LlamaChatSession({\n contextSequence: sequence,\n systemPrompt: input.prompt as string,\n });\n try {\n const text = await session.prompt(input.text as string, { signal });\n update_progress(100, \"Text rewriting complete\");\n return { text };\n } finally {\n sequence.dispose();\n }\n};\n\n// ========================================================================\n// TextRewriterTask (streaming)\n// ========================================================================\n\nexport const LlamaCpp_TextRewriter_Stream: AiProviderStreamFn<\n TextRewriterTaskInput,\n TextRewriterTaskOutput,\n LlamaCppModelConfig\n> = async function* (input, model, signal): AsyncIterable<StreamEvent<TextRewriterTaskOutput>> {\n if (!model) throw new Error(\"Model config is required for TextRewriterTask.\");\n\n const { LlamaChatSession } = await loadSdk();\n\n const context = await getOrCreateTextContext(model);\n const sequence = context.getSequence();\n const session = new LlamaChatSession({\n contextSequence: sequence,\n systemPrompt: input.prompt as string,\n });\n try {\n yield* streamFromSession<TextRewriterTaskOutput>((onTextChunk) => {\n return session.prompt(input.text as string, { signal, onTextChunk });\n }, signal);\n } finally {\n sequence.dispose();\n }\n};\n\n// ========================================================================\n// TextSummaryTask (non-streaming)\n// ========================================================================\n\nexport const LlamaCpp_TextSummary: AiProviderRunFn<\n TextSummaryTaskInput,\n TextSummaryTaskOutput,\n LlamaCppModelConfig\n> = async (input, model, update_progress, signal) => {\n if (Array.isArray(input.text)) {\n getLogger().warn(\n \"LlamaCpp_TextSummary: array input received; processing sequentially (no native batch support)\"\n );\n const texts = input.text as string[];\n const results: string[] = [];\n for (const item of texts) {\n const r = await LlamaCpp_TextSummary(\n { ...input, text: item },\n model,\n update_progress,\n signal\n );\n results.push(r.text as string);\n }\n return { text: results };\n }\n\n if (!model) throw new Error(\"Model config is required for TextSummaryTask.\");\n\n const { LlamaChatSession } = await loadSdk();\n\n update_progress(0, \"Loading model\");\n const context = await getOrCreateTextContext(model);\n\n update_progress(10, \"Summarizing text\");\n const sequence = context.getSequence();\n const session = new LlamaChatSession({\n contextSequence: sequence,\n systemPrompt: \"Summarize the following text concisely, preserving the key points.\",\n });\n try {\n const text = await session.prompt(input.text as string, { signal });\n update_progress(100, \"Summarization complete\");\n return { text };\n } finally {\n sequence.dispose();\n }\n};\n\n// ========================================================================\n// TextSummaryTask (streaming)\n// ========================================================================\n\nexport const LlamaCpp_TextSummary_Stream: AiProviderStreamFn<\n TextSummaryTaskInput,\n TextSummaryTaskOutput,\n LlamaCppModelConfig\n> = async function* (input, model, signal): AsyncIterable<StreamEvent<TextSummaryTaskOutput>> {\n if (!model) throw new Error(\"Model config is required for TextSummaryTask.\");\n\n const { LlamaChatSession } = await loadSdk();\n\n const context = await getOrCreateTextContext(model);\n const sequence = context.getSequence();\n const session = new LlamaChatSession({\n contextSequence: sequence,\n systemPrompt: \"Summarize the following text concisely, preserving the key points.\",\n });\n try {\n yield* streamFromSession<TextSummaryTaskOutput>((onTextChunk) => {\n return session.prompt(input.text as string, { signal, onTextChunk });\n }, signal);\n } finally {\n sequence.dispose();\n }\n};\n\n// ========================================================================\n// Dispose helper (called from LlamaCppProvider.dispose())\n// ========================================================================\n\nexport async function disposeLlamaCppResources(): Promise<void> {\n const disposeAll = async (map: Map<string, { dispose(): Promise<void> }>) => {\n for (const resource of map.values()) {\n await resource.dispose().catch(() => {});\n }\n map.clear();\n };\n\n await disposeAll(textContexts as Map<string, { dispose(): Promise<void> }>);\n await disposeAll(embeddingContexts as Map<string, { dispose(): Promise<void> }>);\n await disposeAll(models as Map<string, { dispose(): Promise<void> }>);\n\n if (llamaInstance) {\n await (llamaInstance as any).dispose?.().catch(() => {});\n llamaInstance = undefined;\n }\n\n resolvedPaths.clear();\n}\n\nexport const LlamaCpp_CountTokens: AiProviderRunFn<\n CountTokensTaskInput,\n CountTokensTaskOutput,\n LlamaCppModelConfig\n> = async (input, model, onProgress, signal) => {\n if (Array.isArray(input.text)) {\n getLogger().warn(\n \"LlamaCpp_CountTokens: array input received; processing sequentially (no native batch support)\"\n );\n const texts = input.text as string[];\n const counts: number[] = [];\n for (const item of texts) {\n const r = await LlamaCpp_CountTokens({ ...input, text: item }, model, onProgress, signal);\n counts.push(r.count as number);\n }\n return { count: counts };\n }\n\n const loadedModel = await getOrLoadModel(model!);\n // model.tokenizer is itself the tokenize function (Tokenizer = tokenize[\"tokenize\"])\n const tokens = loadedModel.tokenizer(input.text as string);\n return { count: tokens.length };\n};\n\nexport const LlamaCpp_CountTokens_Reactive: AiProviderReactiveRunFn<\n CountTokensTaskInput,\n CountTokensTaskOutput,\n LlamaCppModelConfig\n> = async (input, _output, model) => {\n return LlamaCpp_CountTokens(input, model, () => {}, new AbortController().signal);\n};\n\n// ========================================================================\n// ToolCallingTask (non-streaming)\n// ========================================================================\n\n/**\n * Build a prompt string from the task input.\n * When `input.messages` is present (multi-turn agent loop), concatenates\n * the conversation history into a single prompt string since LlamaCpp\n * uses a session-based approach that doesn't support external message arrays.\n */\nfunction buildLlamaCppPrompt(input: ToolCallingTaskInput): string {\n const inputMessages = input.messages;\n if (!inputMessages || inputMessages.length === 0) {\n return Array.isArray(input.prompt) ? input.prompt.join(\"\\n\") : input.prompt;\n }\n\n // Concatenate messages into a single prompt for the session\n const parts: string[] = [];\n for (const msg of inputMessages) {\n if (msg.role === \"user\") {\n parts.push(`User: ${msg.content}`);\n } else if (msg.role === \"assistant\" && Array.isArray(msg.content)) {\n const text = msg.content\n .filter((b: any) => b.type === \"text\")\n .map((b: any) => b.text)\n .join(\"\");\n if (text) parts.push(`Assistant: ${text}`);\n } else if (msg.role === \"tool\" && Array.isArray(msg.content)) {\n for (const block of msg.content) {\n parts.push(`Tool Result: ${block.content}`);\n }\n }\n }\n return parts.join(\"\\n\\n\");\n}\n\n/**\n * Builds function definitions for node-llama-cpp from ToolDefinition inputs.\n * Each function handler captures the call arguments and returns a simple\n * acknowledgment, allowing us to collect tool calls without side-effects.\n */\nfunction buildLlamaCppFunctions(\n tools: ReadonlyArray<ToolDefinition>,\n capturedCalls: Array<{ name: string; input: Record<string, unknown> }>\n) {\n const { defineChatSessionFunction } = _sdk!;\n const functions: Record<string, any> = {};\n for (const tool of tools) {\n const toolName = tool.name;\n functions[toolName] = defineChatSessionFunction({\n description: tool.description,\n params: tool.inputSchema as any,\n handler(params: any) {\n capturedCalls.push({ name: toolName, input: (params ?? {}) as Record<string, unknown> });\n return \"OK\";\n },\n } as any);\n }\n return functions;\n}\n\nexport const LlamaCpp_ToolCalling: AiProviderRunFn<\n ToolCallingTaskInput,\n ToolCallingTaskOutput,\n LlamaCppModelConfig\n> = async (input, model, update_progress, signal) => {\n if (Array.isArray(input.prompt)) {\n getLogger().warn(\n \"LlamaCpp_ToolCalling: array input received; processing sequentially (no native batch support)\"\n );\n const prompts = input.prompt as string[];\n const texts: string[] = [];\n const toolCallsList: ToolCalls[] = [];\n for (const item of prompts) {\n const r = await LlamaCpp_ToolCalling(\n { ...input, prompt: item },\n model,\n update_progress,\n signal\n );\n texts.push(r.text as string);\n toolCallsList.push(r.toolCalls as ToolCalls);\n }\n return { text: texts, toolCalls: toolCallsList } as unknown as ToolCallingTaskOutput;\n }\n\n if (!model) throw new Error(\"Model config is required for ToolCallingTask.\");\n\n await loadSdk();\n\n update_progress(0, \"Loading model\");\n const context = await getOrCreateTextContext(model);\n\n const capturedCalls: Array<{ name: string; input: Record<string, unknown> }> = [];\n const functions =\n input.toolChoice === \"none\" ? undefined : buildLlamaCppFunctions(input.tools, capturedCalls);\n\n update_progress(10, \"Running tool calling\");\n const sequence = context.getSequence();\n const { LlamaChatSession } = _sdk!;\n const promptText = buildLlamaCppPrompt(input);\n const session = new LlamaChatSession({\n contextSequence: sequence,\n ...(input.systemPrompt && { systemPrompt: input.systemPrompt }),\n });\n\n try {\n const text = await session.prompt(promptText, {\n signal,\n ...(functions && { functions }),\n ...(input.temperature !== undefined && { temperature: input.temperature }),\n ...(input.maxTokens !== undefined && { maxTokens: input.maxTokens }),\n });\n\n const toolCalls: ToolCalls = [];\n capturedCalls.forEach((call, index) => {\n const id = `call_${index}`;\n toolCalls.push({ id, name: call.name, input: call.input });\n });\n\n update_progress(100, \"Tool calling complete\");\n return { text, toolCalls: filterValidToolCalls(toolCalls, input.tools) };\n } finally {\n sequence.dispose();\n }\n};\n\n// ========================================================================\n// ToolCallingTask (streaming)\n// ========================================================================\n\nexport const LlamaCpp_ToolCalling_Stream: AiProviderStreamFn<\n ToolCallingTaskInput,\n ToolCallingTaskOutput,\n LlamaCppModelConfig\n> = async function* (input, model, signal): AsyncIterable<StreamEvent<ToolCallingTaskOutput>> {\n if (!model) throw new Error(\"Model config is required for ToolCallingTask.\");\n\n await loadSdk();\n\n const context = await getOrCreateTextContext(model);\n\n const capturedCalls: Array<{ name: string; input: Record<string, unknown> }> = [];\n const functions =\n input.toolChoice === \"none\" ? undefined : buildLlamaCppFunctions(input.tools, capturedCalls);\n\n const sequence = context.getSequence();\n const { LlamaChatSession } = _sdk!;\n const promptText = buildLlamaCppPrompt(input);\n const session = new LlamaChatSession({\n contextSequence: sequence,\n ...(input.systemPrompt && { systemPrompt: input.systemPrompt }),\n });\n\n const queue: string[] = [];\n let isComplete = false;\n let completionError: unknown;\n let resolveWait: (() => void) | null = null;\n\n const notifyWaiter = () => {\n resolveWait?.();\n resolveWait = null;\n };\n\n let accumulatedText = \"\";\n const promptPromise = session\n .prompt(promptText, {\n signal,\n ...(functions && { functions }),\n onTextChunk: (chunk: string) => {\n queue.push(chunk);\n notifyWaiter();\n },\n ...(input.temperature !== undefined && { temperature: input.temperature }),\n ...(input.maxTokens !== undefined && { maxTokens: input.maxTokens }),\n })\n .then(() => {\n isComplete = true;\n notifyWaiter();\n })\n .catch((err: unknown) => {\n completionError = err;\n isComplete = true;\n notifyWaiter();\n });\n\n try {\n while (true) {\n while (queue.length > 0) {\n const chunk = queue.shift()!;\n accumulatedText += chunk;\n yield { type: \"text-delta\", port: \"text\", textDelta: chunk };\n }\n if (isComplete) break;\n await new Promise<void>((r) => {\n resolveWait = r;\n });\n }\n // Drain any remaining chunks after completion signal\n while (queue.length > 0) {\n const chunk = queue.shift()!;\n accumulatedText += chunk;\n yield { type: \"text-delta\", port: \"text\", textDelta: chunk };\n }\n } finally {\n await promptPromise.catch(() => {});\n sequence.dispose();\n }\n\n if (completionError) {\n if (!signal.aborted) throw completionError;\n return;\n }\n\n const toolCalls: ToolCalls = [];\n capturedCalls.forEach((call, index) => {\n const id = `call_${index}`;\n toolCalls.push({ id, name: call.name, input: call.input });\n });\n const validToolCalls = filterValidToolCalls(toolCalls, input.tools);\n\n if (validToolCalls.length > 0) {\n yield { type: \"object-delta\", port: \"toolCalls\", objectDelta: [...validToolCalls] };\n }\n\n yield {\n type: \"finish\",\n data: { text: accumulatedText, toolCalls: validToolCalls } as ToolCallingTaskOutput,\n };\n};\n\n// ========================================================================\n// StructuredGenerationTask\n// ========================================================================\n\nexport const LlamaCpp_StructuredGeneration: AiProviderRunFn<\n StructuredGenerationTaskInput,\n StructuredGenerationTaskOutput,\n LlamaCppModelConfig\n> = async (input, model, update_progress, signal) => {\n if (!model) throw new Error(\"Model config is required for StructuredGenerationTask.\");\n\n await loadSdk();\n\n update_progress(0, \"Loading model\");\n const llama = await getLlamaInstance();\n const context = await getOrCreateTextContext(model);\n\n update_progress(10, \"Running structured generation\");\n const grammar = await llama.createGrammarForJsonSchema(input.outputSchema as any);\n const sequence = context.getSequence();\n const { LlamaChatSession } = _sdk!;\n const session = new LlamaChatSession({ contextSequence: sequence });\n\n try {\n const text = await session.prompt(input.prompt as string, {\n signal,\n grammar,\n ...(input.temperature !== undefined && { temperature: input.temperature }),\n ...(input.maxTokens !== undefined && { maxTokens: input.maxTokens }),\n });\n\n let object: Record<string, unknown>;\n try {\n object = JSON.parse(text);\n } catch {\n object = {};\n }\n\n update_progress(100, \"Structured generation complete\");\n return { object };\n } finally {\n sequence.dispose();\n }\n};\n\n// ========================================================================\n// StructuredGenerationTask (streaming)\n// ========================================================================\n\nexport const LlamaCpp_StructuredGeneration_Stream: AiProviderStreamFn<\n StructuredGenerationTaskInput,\n StructuredGenerationTaskOutput,\n LlamaCppModelConfig\n> = async function* (\n input,\n model,\n signal\n): AsyncIterable<StreamEvent<StructuredGenerationTaskOutput>> {\n if (!model) throw new Error(\"Model config is required for StructuredGenerationTask.\");\n\n await loadSdk();\n\n const llama = await getLlamaInstance();\n const context = await getOrCreateTextContext(model);\n const grammar = await llama.createGrammarForJsonSchema(input.outputSchema as any);\n\n const sequence = context.getSequence();\n const { LlamaChatSession } = _sdk!;\n const session = new LlamaChatSession({ contextSequence: sequence });\n\n const queue: string[] = [];\n let isComplete = false;\n let completionError: unknown;\n let resolveWait: (() => void) | null = null;\n\n const notifyWaiter = () => {\n resolveWait?.();\n resolveWait = null;\n };\n\n let accumulatedText = \"\";\n const promptPromise = session\n .prompt(input.prompt as string, {\n signal,\n grammar,\n onTextChunk: (chunk: string) => {\n queue.push(chunk);\n notifyWaiter();\n },\n ...(input.temperature !== undefined && { temperature: input.temperature }),\n ...(input.maxTokens !== undefined && { maxTokens: input.maxTokens }),\n })\n .then(() => {\n isComplete = true;\n notifyWaiter();\n })\n .catch((err: unknown) => {\n completionError = err;\n isComplete = true;\n notifyWaiter();\n });\n\n try {\n while (true) {\n while (queue.length > 0) {\n const chunk = queue.shift()!;\n accumulatedText += chunk;\n // Try to parse partial JSON and emit object-delta\n const partial = parsePartialJson(accumulatedText);\n if (partial !== undefined) {\n yield {\n type: \"object-delta\",\n port: \"object\",\n objectDelta: partial as Record<string, unknown>,\n };\n }\n }\n if (isComplete) break;\n await new Promise<void>((r) => {\n resolveWait = r;\n });\n }\n // Drain remaining chunks\n while (queue.length > 0) {\n const chunk = queue.shift()!;\n accumulatedText += chunk;\n }\n } finally {\n await promptPromise.catch(() => {});\n sequence.dispose();\n }\n\n if (completionError) {\n if (signal.aborted) return;\n throw completionError;\n }\n\n let finalObject: Record<string, unknown>;\n try {\n finalObject = JSON.parse(accumulatedText);\n } catch {\n finalObject = (parsePartialJson(accumulatedText) as Record<string, unknown>) ?? {};\n }\n\n yield { type: \"finish\", data: { object: finalObject } as StructuredGenerationTaskOutput };\n};\n\n// ========================================================================\n// Model info\n// ========================================================================\n\nexport const LlamaCpp_ModelInfo: AiProviderRunFn<\n ModelInfoTaskInput,\n ModelInfoTaskOutput,\n LlamaCppModelConfig\n> = async (input, model) => {\n if (!model) throw new Error(\"Model config is required for ModelInfoTask.\");\n\n const modelPath = getActualModelPath(model);\n const is_loaded = models.has(modelPath);\n\n let is_cached = is_loaded;\n let file_sizes: Record<string, number> | null = null;\n\n // Check if model file exists on disk\n try {\n const fs = await import(\"node:fs/promises\");\n const stat = await fs.stat(modelPath);\n is_cached = true;\n file_sizes = { model: stat.size };\n } catch {\n // File does not exist or fs not available\n // Fall back to checking if the path is in resolvedPaths\n if (resolvedPaths.has(getConfigKey(model))) {\n is_cached = true;\n }\n }\n\n return {\n model: input.model,\n is_local: true,\n is_remote: false,\n supports_browser: false,\n supports_node: true,\n is_cached,\n is_loaded,\n file_sizes,\n };\n};\n\n// ========================================================================\n// Task registries\n// ========================================================================\n\nexport const LLAMACPP_TASKS: Record<string, AiProviderRunFn<any, any, LlamaCppModelConfig>> = {\n DownloadModelTask: LlamaCpp_Download,\n UnloadModelTask: LlamaCpp_Unload,\n ModelInfoTask: LlamaCpp_ModelInfo,\n CountTokensTask: LlamaCpp_CountTokens,\n TextGenerationTask: LlamaCpp_TextGeneration,\n TextEmbeddingTask: LlamaCpp_TextEmbedding,\n TextRewriterTask: LlamaCpp_TextRewriter,\n TextSummaryTask: LlamaCpp_TextSummary,\n ToolCallingTask: LlamaCpp_ToolCalling,\n StructuredGenerationTask: LlamaCpp_StructuredGeneration,\n};\n\nexport const LLAMACPP_STREAM_TASKS: Record<\n string,\n AiProviderStreamFn<any, any, LlamaCppModelConfig>\n> = {\n TextGenerationTask: LlamaCpp_TextGeneration_Stream,\n TextRewriterTask: LlamaCpp_TextRewriter_Stream,\n TextSummaryTask: LlamaCpp_TextSummary_Stream,\n ToolCallingTask: LlamaCpp_ToolCalling_Stream,\n StructuredGenerationTask: LlamaCpp_StructuredGeneration_Stream,\n};\n\nexport const LLAMACPP_REACTIVE_TASKS: Record<\n string,\n AiProviderReactiveRunFn<any, any, LlamaCppModelConfig>\n> = {\n CountTokensTask: LlamaCpp_CountTokens_Reactive,\n};\n",
|
|
6
|
-
"/**\n * @license\n * Copyright 2025 Steven Roussey <sroussey@gmail.com>\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport {
|
|
5
|
+
"/**\n * @license\n * Copyright 2025 Steven Roussey <sroussey@gmail.com>\n * SPDX-License-Identifier: Apache-2.0\n */\n\nexport const LOCAL_LLAMACPP = \"LOCAL_LLAMACPP\";\nexport const LLAMACPP_DEFAULT_MODELS_DIR = \"./models\";\n",
|
|
6
|
+
"/**\n * @license\n * Copyright 2025 Steven Roussey <sroussey@gmail.com>\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { ModelConfigSchema, ModelRecordSchema } from \"@workglow/ai/worker\";\nimport { DataPortSchemaObject, FromSchema } from \"@workglow/util/worker\";\nimport { LLAMACPP_DEFAULT_MODELS_DIR, LOCAL_LLAMACPP } from \"./LlamaCpp_Constants\";\n\nexport const LlamaCppModelSchema = {\n type: \"object\",\n properties: {\n provider: {\n const: LOCAL_LLAMACPP,\n description: \"Discriminator: local node-llama-cpp (GGUF) model.\",\n },\n provider_config: {\n type: \"object\",\n description: \"node-llama-cpp specific configuration.\",\n properties: {\n model_path: {\n type: \"string\",\n description: \"Filesystem path to the .gguf model file.\",\n },\n model_url: {\n type: \"string\",\n description:\n \"URI or URL to download the model from (e.g. 'hf:user/repo:quant' or an https URL). Used by DownloadModelTask.\",\n },\n models_dir: {\n type: \"string\",\n description: \"Directory to download models into.\",\n default: LLAMACPP_DEFAULT_MODELS_DIR,\n },\n gpu_layers: {\n type: \"integer\",\n description: \"Number of model layers to offload to GPU. Use -1 for auto-detection.\",\n minimum: -1,\n },\n context_size: {\n type: \"integer\",\n description: \"Context window size in tokens.\",\n minimum: 1,\n },\n flash_attention: {\n type: \"boolean\",\n description: \"Enable flash attention for improved performance where supported.\",\n default: true,\n },\n embedding: {\n type: \"boolean\",\n description: \"Whether this model is used for text embedding (vs text generation).\",\n default: false,\n },\n },\n required: [\"model_path\"],\n additionalProperties: false,\n },\n },\n required: [\"provider\", \"provider_config\"],\n additionalProperties: true,\n} as const satisfies DataPortSchemaObject;\n\nexport const LlamaCppModelRecordSchema = {\n type: \"object\",\n properties: {\n ...ModelRecordSchema.properties,\n ...LlamaCppModelSchema.properties,\n },\n required: [...ModelRecordSchema.required, ...LlamaCppModelSchema.required],\n additionalProperties: false,\n} as const satisfies DataPortSchemaObject;\n\nexport type LlamaCppModelRecord = FromSchema<typeof LlamaCppModelRecordSchema>;\n\nexport const LlamaCppModelConfigSchema = {\n type: \"object\",\n properties: {\n ...ModelConfigSchema.properties,\n ...LlamaCppModelSchema.properties,\n },\n required: [...ModelConfigSchema.required, ...LlamaCppModelSchema.required],\n additionalProperties: false,\n} as const satisfies DataPortSchemaObject;\n\nexport type LlamaCppModelConfig = FromSchema<typeof LlamaCppModelConfigSchema>;\n",
|
|
7
|
+
"/**\n * @license\n * Copyright 2025 Steven Roussey <sroussey@gmail.com>\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport {\n QueuedAiProvider,\n type AiProviderReactiveRunFn,\n type AiProviderRunFn,\n type AiProviderStreamFn,\n} from \"@workglow/ai\";\nimport { LOCAL_LLAMACPP } from \"./common/LlamaCpp_Constants\";\nimport type { LlamaCppModelConfig } from \"./common/LlamaCpp_ModelSchema\";\n\n/** Main-thread registration (inline or worker-backed); creates the default job queue. */\nexport class LlamaCppQueuedProvider extends QueuedAiProvider<LlamaCppModelConfig> {\n readonly name = LOCAL_LLAMACPP;\n readonly isLocal = true;\n readonly supportsBrowser = false;\n\n readonly taskTypes = [\n \"DownloadModelTask\",\n \"UnloadModelTask\",\n \"ModelInfoTask\",\n \"CountTokensTask\",\n \"TextGenerationTask\",\n \"TextEmbeddingTask\",\n \"TextRewriterTask\",\n \"TextSummaryTask\",\n \"ToolCallingTask\",\n \"ModelSearchTask\",\n ] as const;\n\n constructor(\n tasks?: Record<string, AiProviderRunFn<any, any, LlamaCppModelConfig>>,\n streamTasks?: Record<string, AiProviderStreamFn<any, any, LlamaCppModelConfig>>,\n reactiveTasks?: Record<string, AiProviderReactiveRunFn<any, any, LlamaCppModelConfig>>\n ) {\n super(tasks, streamTasks, reactiveTasks);\n }\n}\n",
|
|
8
|
+
"/**\n * @license\n * Copyright 2025 Steven Roussey <sroussey@gmail.com>\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport type { AiProviderRegisterOptions } from \"@workglow/ai\";\nimport { LlamaCppQueuedProvider } from \"./LlamaCppQueuedProvider\";\n\nexport async function registerLlamaCpp(\n options: AiProviderRegisterOptions & {\n worker: Worker | (() => Worker);\n }\n): Promise<void> {\n await new LlamaCppQueuedProvider().register(options);\n}\n"
|
|
7
9
|
],
|
|
8
|
-
"mappings": ";;;;;;;;;;;;AAMA;AA6BA;AAQA,IAAI;AAEJ,eAAe,OAAO,GAAG;AAAA,EACvB,IAAI,CAAC,MAAM;AAAA,IACT,IAAI;AAAA,MACF,OAAO,MAAa;AAAA,MACpB,OAAO,KAAK;AAAA,MACZ,MAAM,IAAI,MACR,8FACF;AAAA;AAAA,EAEJ;AAAA,EACA,OAAO;AAAA;AAYT,IAAI;AACJ,IAAM,SAAS,IAAI;AACnB,IAAM,eAAe,IAAI;AACzB,IAAM,oBAAoB,IAAI;AAG9B,IAAM,gBAAgB,IAAI;AAM1B,eAAe,gBAAgB,GAA2B;AAAA,EACxD,IAAI,CAAC,eAAe;AAAA,IAClB,QAAQ,aAAa,MAAM,QAAQ;AAAA,IACnC,gBAAgB,MAAM,SAAS;AAAA,EACjC;AAAA,EACA,OAAO;AAAA;AAGT,SAAS,YAAY,CAAC,OAAoC;AAAA,EACxD,OAAO,MAAM,gBAAgB,aAAa,MAAM,gBAAgB;AAAA;AAGlE,SAAS,kBAAkB,CAAC,OAAoC;AAAA,EAC9D,MAAM,MAAM,aAAa,KAAK;AAAA,EAC9B,MAAM,WAAW,cAAc,IAAI,GAAG;AAAA,EACtC,OAAO,YAAY,MAAM,gBAAgB;AAAA;AAG3C,eAAe,cAAc,CAAC,OAAiD;AAAA,EAC7E,MAAM,YAAY,mBAAmB,KAAK;AAAA,EAC1C,MAAM,SAAS,OAAO,IAAI,SAAS;AAAA,EACnC,IAAI;AAAA,IAAQ,OAAO;AAAA,EAEnB,MAAM,QAAQ,MAAM,iBAAiB;AAAA,EACrC,MAAM,SAAS,MAAM;AAAA,EAErB,MAAM,cAAc,MAAM,MAAM,UAAU;AAAA,IACxC;AAAA,OACI,OAAO,eAAe,aAAa,EAAE,WAAW,OAAO,WAAW;AAAA,EACxE,CAAC;AAAA,EAED,OAAO,IAAI,WAAW,WAAW;AAAA,EACjC,OAAO;AAAA;AAGT,eAAe,sBAAsB,CAAC,OAAmD;AAAA,EACvF,MAAM,YAAY,mBAAmB,KAAK;AAAA,EAC1C,MAAM,SAAS,aAAa,IAAI,SAAS;AAAA,EACzC,IAAI;AAAA,IAAQ,OAAO;AAAA,EAEnB,MAAM,cAAc,MAAM,eAAe,KAAK;AAAA,EAC9C,MAAM,SAAS,MAAM;AAAA,EAErB,MAAM,UAAU,MAAM,YAAY,cAAc;AAAA,OAC1C,OAAO,gBAAgB,EAAE,aAAa,OAAO,aAAa;AAAA,OAC1D,OAAO,oBAAoB,aAAa,EAAE,gBAAgB,OAAO,gBAAgB;AAAA,EACvF,CAAC;AAAA,EAED,aAAa,IAAI,WAAW,OAAO;AAAA,EACnC,OAAO;AAAA;AAGT,eAAe,2BAA2B,CACxC,OACgC;AAAA,EAChC,MAAM,YAAY,mBAAmB,KAAK;AAAA,EAC1C,MAAM,SAAS,kBAAkB,IAAI,SAAS;AAAA,EAC9C,IAAI;AAAA,IAAQ,OAAO;AAAA,EAEnB,MAAM,cAAc,MAAM,eAAe,KAAK;AAAA,EAE9C,MAAM,UAAU,MAAM,YAAY,uBAAuB;AAAA,EAEzD,kBAAkB,IAAI,WAAW,OAAO;AAAA,EACxC,OAAO;AAAA;AAOT,gBAAgB,iBAAoD,CAClE,UACA,QACgC;AAAA,EAChC,MAAM,QAAkB,CAAC;AAAA,EACzB,IAAI,aAAa;AAAA,EACjB,IAAI;AAAA,EACJ,IAAI,cAAmC;AAAA,EAEvC,MAAM,eAAe,MAAM;AAAA,IACzB,cAAc;AAAA,IACd,cAAc;AAAA;AAAA,EAGhB,MAAM,gBAAgB,SAAS,CAAC,UAAkB;AAAA,IAChD,MAAM,KAAK,KAAK;AAAA,IAChB,aAAa;AAAA,GACd,EACE,KAAK,MAAM;AAAA,IACV,aAAa;AAAA,IACb,aAAa;AAAA,GACd,EACA,MAAM,CAAC,QAAiB;AAAA,IACvB,kBAAkB;AAAA,IAClB,aAAa;AAAA,IACb,aAAa;AAAA,GACd;AAAA,EAEH,IAAI;AAAA,IACF,OAAO,MAAM;AAAA,MACX,OAAO,MAAM,SAAS,GAAG;AAAA,QACvB,MAAM,EAAE,MAAM,cAAc,MAAM,QAAQ,WAAW,MAAM,MAAM,EAAG;AAAA,MACtE;AAAA,MACA,IAAI;AAAA,QAAY;AAAA,MAChB,MAAM,IAAI,QAAc,CAAC,MAAM;AAAA,QAC7B,cAAc;AAAA,OACf;AAAA,IACH;AAAA,IAEA,OAAO,MAAM,SAAS,GAAG;AAAA,MACvB,MAAM,EAAE,MAAM,cAAc,MAAM,QAAQ,WAAW,MAAM,MAAM,EAAG;AAAA,IACtE;AAAA,YACA;AAAA,IACA,MAAM,cAAc,MAAM,MAAM,EAAE;AAAA;AAAA,EAGpC,IAAI,iBAAiB;AAAA,IACnB,IAAI,OAAO;AAAA,MAAS;AAAA,IACpB,MAAM;AAAA,EACR;AAAA,EAEA,MAAM,EAAE,MAAM,UAAU,MAAM,CAAC,EAAO;AAAA;AAOjC,IAAM,oBAIT,OAAO,OAAO,OAAO,iBAAiB,WAAW;AAAA,EACnD,IAAI,CAAC;AAAA,IAAO,MAAM,IAAI,MAAM,iDAAiD;AAAA,EAE7E,QAAQ,0BAA0B,MAAM,QAAQ;AAAA,EAChD,MAAM,SAAS,MAAM;AAAA,EACrB,MAAM,WAAW,OAAO,aAAa,OAAO;AAAA,EAC5C,MAAM,UAAU,OAAO,cAAc;AAAA,EAErC,gBAAgB,GAAG,2BAA2B;AAAA,EAE9C,MAAM,aAAa,MAAM,sBAAsB,EAAE,UAAU,QAAQ,CAAC;AAAA,EAGpE,MAAM,mBAAmB,YAAY,MAAM;AAAA,IACzC,MAAM,QAAS,WAAmB;AAAA,IAClC,MAAM,aAAc,WAAmB;AAAA,IACvC,IAAI,SAAS,QAAQ,KAAK,eAAe,WAAW;AAAA,MAClD,MAAM,MAAM,KAAK,IAAI,IAAI,KAAK,MAAO,aAAa,QAAS,GAAG,CAAC;AAAA,MAC/D,gBAAgB,KAAK,qBAAqB,EAAE,MAAM,UAAU,UAAU,MAAM,IAAI,CAAC;AAAA,IACnF;AAAA,KACC,GAAG;AAAA,EAEN,IAAI;AAAA,EACJ,IAAI;AAAA,IACF,YAAY,MAAM,WAAW,SAAS;AAAA,YACtC;AAAA,IACA,cAAc,gBAAgB;AAAA;AAAA,EAIhC,cAAc,IAAI,aAAa,KAAK,GAAG,SAAS;AAAA,EAEhD,gBAAgB,KAAK,oBAAoB,EAAE,MAAM,UAAU,UAAU,EAAE,CAAC;AAAA,EAExE,OAAO,EAAE,OAAO,MAAM,MAAO;AAAA;AAOxB,IAAM,kBAIT,OAAO,OAAO,OAAO,iBAAiB,YAAY;AAAA,EACpD,IAAI,CAAC;AAAA,IAAO,MAAM,IAAI,MAAM,+CAA+C;AAAA,EAE3E,MAAM,YAAY,mBAAmB,KAAK;AAAA,EAG1C,MAAM,MAAM,aAAa,IAAI,SAAS;AAAA,EACtC,IAAI,KAAK;AAAA,IACP,MAAM,IAAI,QAAQ;AAAA,IAClB,aAAa,OAAO,SAAS;AAAA,IAC7B,gBAAgB,IAAI,uBAAuB;AAAA,EAC7C;AAAA,EAGA,MAAM,SAAS,kBAAkB,IAAI,SAAS;AAAA,EAC9C,IAAI,QAAQ;AAAA,IACV,MAAM,OAAO,QAAQ;AAAA,IACrB,kBAAkB,OAAO,SAAS;AAAA,IAClC,gBAAgB,IAAI,4BAA4B;AAAA,EAClD;AAAA,EAGA,MAAM,cAAc,OAAO,IAAI,SAAS;AAAA,EACxC,IAAI,aAAa;AAAA,IACf,MAAM,YAAY,QAAQ;AAAA,IAC1B,OAAO,OAAO,SAAS;AAAA,IACvB,gBAAgB,KAAK,4BAA4B;AAAA,EACnD,EAAO;AAAA,IACL,gBAAgB,KAAK,sBAAsB;AAAA;AAAA,EAG7C,OAAO,EAAE,OAAO,MAAM,MAAO;AAAA;AAOxB,IAAM,0BAIT,OAAO,OAAO,OAAO,iBAAiB,WAAW;AAAA,EACnD,IAAI,MAAM,QAAQ,MAAM,MAAM,GAAG;AAAA,IAC/B,UAAU,EAAE,KACV,kGACF;AAAA,IACA,MAAM,UAAU,MAAM;AAAA,IACtB,MAAM,UAAoB,CAAC;AAAA,IAC3B,WAAW,QAAQ,SAAS;AAAA,MAC1B,MAAM,IAAI,MAAM,wBACd,KAAK,OAAO,QAAQ,KAAK,GACzB,OACA,iBACA,MACF;AAAA,MACA,QAAQ,KAAK,EAAE,IAAc;AAAA,IAC/B;AAAA,IACA,OAAO,EAAE,MAAM,QAAQ;AAAA,EACzB;AAAA,EAEA,IAAI,CAAC;AAAA,IAAO,MAAM,IAAI,MAAM,kDAAkD;AAAA,EAE9E,QAAQ,qBAAqB,MAAM,QAAQ;AAAA,EAE3C,gBAAgB,GAAG,eAAe;AAAA,EAClC,MAAM,UAAU,MAAM,uBAAuB,KAAK;AAAA,EAElD,gBAAgB,IAAI,iBAAiB;AAAA,EACrC,MAAM,WAAW,QAAQ,YAAY;AAAA,EACrC,MAAM,UAAU,IAAI,iBAAiB,EAAE,iBAAiB,SAAS,CAAC;AAAA,EAClE,IAAI;AAAA,IACF,MAAM,OAAO,MAAM,QAAQ,OAAO,MAAM,QAAkB;AAAA,MACxD;AAAA,SACI,MAAM,gBAAgB,aAAa,EAAE,aAAa,MAAM,YAAY;AAAA,SACpE,MAAM,cAAc,aAAa,EAAE,WAAW,MAAM,UAAU;AAAA,SAC9D,MAAM,SAAS,aAAa,EAAE,MAAM,MAAM,KAAK;AAAA,IACrD,CAAC;AAAA,IACD,gBAAgB,KAAK,0BAA0B;AAAA,IAC/C,OAAO,EAAE,KAAK;AAAA,YACd;AAAA,IACA,SAAS,QAAQ;AAAA;AAAA;AAQd,IAAM,iCAIT,gBAAgB,CAAC,OAAO,OAAO,QAA8D;AAAA,EAC/F,IAAI,CAAC;AAAA,IAAO,MAAM,IAAI,MAAM,kDAAkD;AAAA,EAE9E,QAAQ,qBAAqB,MAAM,QAAQ;AAAA,EAE3C,MAAM,UAAU,MAAM,uBAAuB,KAAK;AAAA,EAClD,MAAM,WAAW,QAAQ,YAAY;AAAA,EACrC,MAAM,UAAU,IAAI,iBAAiB,EAAE,iBAAiB,SAAS,CAAC;AAAA,EAClE,IAAI;AAAA,IACF,OAAO,kBAA4C,CAAC,gBAAgB;AAAA,MAClE,OAAO,QAAQ,OAAO,MAAM,QAAkB;AAAA,QAC5C;AAAA,QACA;AAAA,WACI,MAAM,gBAAgB,aAAa,EAAE,aAAa,MAAM,YAAY;AAAA,WACpE,MAAM,cAAc,aAAa,EAAE,WAAW,MAAM,UAAU;AAAA,WAC9D,MAAM,SAAS,aAAa,EAAE,MAAM,MAAM,KAAK;AAAA,MACrD,CAAC;AAAA,OACA,MAAM;AAAA,YACT;AAAA,IACA,SAAS,QAAQ;AAAA;AAAA;AAQd,IAAM,yBAIT,OAAO,OAAO,OAAO,iBAAiB,YAAY;AAAA,EACpD,IAAI,CAAC;AAAA,IAAO,MAAM,IAAI,MAAM,iDAAiD;AAAA,EAE7E,gBAAgB,GAAG,yBAAyB;AAAA,EAC5C,MAAM,UAAU,MAAM,4BAA4B,KAAK;AAAA,EAEvD,MAAM,QAAQ,MAAM,QAAQ,MAAM,IAAI,IAAI,MAAM,OAAO,CAAC,MAAM,IAAI;AAAA,EAClE,gBAAgB,IAAI,sBAAsB;AAAA,EAE1C,MAAM,aAAa,MAAM,QAAQ,IAC/B,MAAM,IAAI,CAAC,SAAS,QAAQ,gBAAgB,IAAI,EAAE,KAAK,CAAC,MAAM,IAAI,aAAa,EAAE,MAAM,CAAC,CAAC,CAC3F;AAAA,EAEA,gBAAgB,KAAK,qBAAqB;AAAA,EAE1C,IAAI,MAAM,QAAQ,MAAM,IAAI,GAAG;AAAA,IAC7B,OAAO,EAAE,QAAQ,WAAW;AAAA,EAC9B;AAAA,EACA,OAAO,EAAE,QAAQ,WAAW,GAAG;AAAA;AAO1B,IAAM,wBAIT,OAAO,OAAO,OAAO,iBAAiB,WAAW;AAAA,EACnD,IAAI,MAAM,QAAQ,MAAM,IAAI,GAAG;AAAA,IAC7B,UAAU,EAAE,KACV,gGACF;AAAA,IACA,MAAM,QAAQ,MAAM;AAAA,IACpB,MAAM,UAAoB,CAAC;AAAA,IAC3B,WAAW,QAAQ,OAAO;AAAA,MACxB,MAAM,IAAI,MAAM,sBACd,KAAK,OAAO,MAAM,KAAK,GACvB,OACA,iBACA,MACF;AAAA,MACA,QAAQ,KAAK,EAAE,IAAc;AAAA,IAC/B;AAAA,IACA,OAAO,EAAE,MAAM,QAAQ;AAAA,EACzB;AAAA,EAEA,IAAI,CAAC;AAAA,IAAO,MAAM,IAAI,MAAM,gDAAgD;AAAA,EAE5E,QAAQ,qBAAqB,MAAM,QAAQ;AAAA,EAE3C,gBAAgB,GAAG,eAAe;AAAA,EAClC,MAAM,UAAU,MAAM,uBAAuB,KAAK;AAAA,EAElD,gBAAgB,IAAI,gBAAgB;AAAA,EACpC,MAAM,WAAW,QAAQ,YAAY;AAAA,EACrC,MAAM,UAAU,IAAI,iBAAiB;AAAA,IACnC,iBAAiB;AAAA,IACjB,cAAc,MAAM;AAAA,EACtB,CAAC;AAAA,EACD,IAAI;AAAA,IACF,MAAM,OAAO,MAAM,QAAQ,OAAO,MAAM,MAAgB,EAAE,OAAO,CAAC;AAAA,IAClE,gBAAgB,KAAK,yBAAyB;AAAA,IAC9C,OAAO,EAAE,KAAK;AAAA,YACd;AAAA,IACA,SAAS,QAAQ;AAAA;AAAA;AAQd,IAAM,+BAIT,gBAAgB,CAAC,OAAO,OAAO,QAA4D;AAAA,EAC7F,IAAI,CAAC;AAAA,IAAO,MAAM,IAAI,MAAM,gDAAgD;AAAA,EAE5E,QAAQ,qBAAqB,MAAM,QAAQ;AAAA,EAE3C,MAAM,UAAU,MAAM,uBAAuB,KAAK;AAAA,EAClD,MAAM,WAAW,QAAQ,YAAY;AAAA,EACrC,MAAM,UAAU,IAAI,iBAAiB;AAAA,IACnC,iBAAiB;AAAA,IACjB,cAAc,MAAM;AAAA,EACtB,CAAC;AAAA,EACD,IAAI;AAAA,IACF,OAAO,kBAA0C,CAAC,gBAAgB;AAAA,MAChE,OAAO,QAAQ,OAAO,MAAM,MAAgB,EAAE,QAAQ,YAAY,CAAC;AAAA,OAClE,MAAM;AAAA,YACT;AAAA,IACA,SAAS,QAAQ;AAAA;AAAA;AAQd,IAAM,uBAIT,OAAO,OAAO,OAAO,iBAAiB,WAAW;AAAA,EACnD,IAAI,MAAM,QAAQ,MAAM,IAAI,GAAG;AAAA,IAC7B,UAAU,EAAE,KACV,+FACF;AAAA,IACA,MAAM,QAAQ,MAAM;AAAA,IACpB,MAAM,UAAoB,CAAC;AAAA,IAC3B,WAAW,QAAQ,OAAO;AAAA,MACxB,MAAM,IAAI,MAAM,qBACd,KAAK,OAAO,MAAM,KAAK,GACvB,OACA,iBACA,MACF;AAAA,MACA,QAAQ,KAAK,EAAE,IAAc;AAAA,IAC/B;AAAA,IACA,OAAO,EAAE,MAAM,QAAQ;AAAA,EACzB;AAAA,EAEA,IAAI,CAAC;AAAA,IAAO,MAAM,IAAI,MAAM,+CAA+C;AAAA,EAE3E,QAAQ,qBAAqB,MAAM,QAAQ;AAAA,EAE3C,gBAAgB,GAAG,eAAe;AAAA,EAClC,MAAM,UAAU,MAAM,uBAAuB,KAAK;AAAA,EAElD,gBAAgB,IAAI,kBAAkB;AAAA,EACtC,MAAM,WAAW,QAAQ,YAAY;AAAA,EACrC,MAAM,UAAU,IAAI,iBAAiB;AAAA,IACnC,iBAAiB;AAAA,IACjB,cAAc;AAAA,EAChB,CAAC;AAAA,EACD,IAAI;AAAA,IACF,MAAM,OAAO,MAAM,QAAQ,OAAO,MAAM,MAAgB,EAAE,OAAO,CAAC;AAAA,IAClE,gBAAgB,KAAK,wBAAwB;AAAA,IAC7C,OAAO,EAAE,KAAK;AAAA,YACd;AAAA,IACA,SAAS,QAAQ;AAAA;AAAA;AAQd,IAAM,8BAIT,gBAAgB,CAAC,OAAO,OAAO,QAA2D;AAAA,EAC5F,IAAI,CAAC;AAAA,IAAO,MAAM,IAAI,MAAM,+CAA+C;AAAA,EAE3E,QAAQ,qBAAqB,MAAM,QAAQ;AAAA,EAE3C,MAAM,UAAU,MAAM,uBAAuB,KAAK;AAAA,EAClD,MAAM,WAAW,QAAQ,YAAY;AAAA,EACrC,MAAM,UAAU,IAAI,iBAAiB;AAAA,IACnC,iBAAiB;AAAA,IACjB,cAAc;AAAA,EAChB,CAAC;AAAA,EACD,IAAI;AAAA,IACF,OAAO,kBAAyC,CAAC,gBAAgB;AAAA,MAC/D,OAAO,QAAQ,OAAO,MAAM,MAAgB,EAAE,QAAQ,YAAY,CAAC;AAAA,OAClE,MAAM;AAAA,YACT;AAAA,IACA,SAAS,QAAQ;AAAA;AAAA;AAQrB,eAAsB,wBAAwB,GAAkB;AAAA,EAC9D,MAAM,aAAa,OAAO,QAAmD;AAAA,IAC3E,WAAW,YAAY,IAAI,OAAO,GAAG;AAAA,MACnC,MAAM,SAAS,QAAQ,EAAE,MAAM,MAAM,EAAE;AAAA,IACzC;AAAA,IACA,IAAI,MAAM;AAAA;AAAA,EAGZ,MAAM,WAAW,YAAyD;AAAA,EAC1E,MAAM,WAAW,iBAA8D;AAAA,EAC/E,MAAM,WAAW,MAAmD;AAAA,EAEpE,IAAI,eAAe;AAAA,IACjB,MAAO,cAAsB,UAAU,EAAE,MAAM,MAAM,EAAE;AAAA,IACvD,gBAAgB;AAAA,EAClB;AAAA,EAEA,cAAc,MAAM;AAAA;AAGf,IAAM,uBAIT,OAAO,OAAO,OAAO,YAAY,WAAW;AAAA,EAC9C,IAAI,MAAM,QAAQ,MAAM,IAAI,GAAG;AAAA,IAC7B,UAAU,EAAE,KACV,+FACF;AAAA,IACA,MAAM,QAAQ,MAAM;AAAA,IACpB,MAAM,SAAmB,CAAC;AAAA,IAC1B,WAAW,QAAQ,OAAO;AAAA,MACxB,MAAM,IAAI,MAAM,qBAAqB,KAAK,OAAO,MAAM,KAAK,GAAG,OAAO,YAAY,MAAM;AAAA,MACxF,OAAO,KAAK,EAAE,KAAe;AAAA,IAC/B;AAAA,IACA,OAAO,EAAE,OAAO,OAAO;AAAA,EACzB;AAAA,EAEA,MAAM,cAAc,MAAM,eAAe,KAAM;AAAA,EAE/C,MAAM,SAAS,YAAY,UAAU,MAAM,IAAc;AAAA,EACzD,OAAO,EAAE,OAAO,OAAO,OAAO;AAAA;AAGzB,IAAM,gCAIT,OAAO,OAAO,SAAS,UAAU;AAAA,EACnC,OAAO,qBAAqB,OAAO,OAAO,MAAM,IAAI,IAAI,gBAAgB,EAAE,MAAM;AAAA;AAalF,SAAS,mBAAmB,CAAC,OAAqC;AAAA,EAChE,MAAM,gBAAgB,MAAM;AAAA,EAC5B,IAAI,CAAC,iBAAiB,cAAc,WAAW,GAAG;AAAA,IAChD,OAAO,MAAM,QAAQ,MAAM,MAAM,IAAI,MAAM,OAAO,KAAK;AAAA,CAAI,IAAI,MAAM;AAAA,EACvE;AAAA,EAGA,MAAM,QAAkB,CAAC;AAAA,EACzB,WAAW,OAAO,eAAe;AAAA,IAC/B,IAAI,IAAI,SAAS,QAAQ;AAAA,MACvB,MAAM,KAAK,SAAS,IAAI,SAAS;AAAA,IACnC,EAAO,SAAI,IAAI,SAAS,eAAe,MAAM,QAAQ,IAAI,OAAO,GAAG;AAAA,MACjE,MAAM,OAAO,IAAI,QACd,OAAO,CAAC,MAAW,EAAE,SAAS,MAAM,EACpC,IAAI,CAAC,MAAW,EAAE,IAAI,EACtB,KAAK,EAAE;AAAA,MACV,IAAI;AAAA,QAAM,MAAM,KAAK,cAAc,MAAM;AAAA,IAC3C,EAAO,SAAI,IAAI,SAAS,UAAU,MAAM,QAAQ,IAAI,OAAO,GAAG;AAAA,MAC5D,WAAW,SAAS,IAAI,SAAS;AAAA,QAC/B,MAAM,KAAK,gBAAgB,MAAM,SAAS;AAAA,MAC5C;AAAA,IACF;AAAA,EACF;AAAA,EACA,OAAO,MAAM,KAAK;AAAA;AAAA,CAAM;AAAA;AAQ1B,SAAS,sBAAsB,CAC7B,OACA,eACA;AAAA,EACA,QAAQ,8BAA8B;AAAA,EACtC,MAAM,YAAiC,CAAC;AAAA,EACxC,WAAW,QAAQ,OAAO;AAAA,IACxB,MAAM,WAAW,KAAK;AAAA,IACtB,UAAU,YAAY,0BAA0B;AAAA,MAC9C,aAAa,KAAK;AAAA,MAClB,QAAQ,KAAK;AAAA,MACb,OAAO,CAAC,QAAa;AAAA,QACnB,cAAc,KAAK,EAAE,MAAM,UAAU,OAAQ,UAAU,CAAC,EAA8B,CAAC;AAAA,QACvF,OAAO;AAAA;AAAA,IAEX,CAAQ;AAAA,EACV;AAAA,EACA,OAAO;AAAA;AAGF,IAAM,uBAIT,OAAO,OAAO,OAAO,iBAAiB,WAAW;AAAA,EACnD,IAAI,MAAM,QAAQ,MAAM,MAAM,GAAG;AAAA,IAC/B,UAAU,EAAE,KACV,+FACF;AAAA,IACA,MAAM,UAAU,MAAM;AAAA,IACtB,MAAM,QAAkB,CAAC;AAAA,IACzB,MAAM,gBAA6B,CAAC;AAAA,IACpC,WAAW,QAAQ,SAAS;AAAA,MAC1B,MAAM,IAAI,MAAM,qBACd,KAAK,OAAO,QAAQ,KAAK,GACzB,OACA,iBACA,MACF;AAAA,MACA,MAAM,KAAK,EAAE,IAAc;AAAA,MAC3B,cAAc,KAAK,EAAE,SAAsB;AAAA,IAC7C;AAAA,IACA,OAAO,EAAE,MAAM,OAAO,WAAW,cAAc;AAAA,EACjD;AAAA,EAEA,IAAI,CAAC;AAAA,IAAO,MAAM,IAAI,MAAM,+CAA+C;AAAA,EAE3E,MAAM,QAAQ;AAAA,EAEd,gBAAgB,GAAG,eAAe;AAAA,EAClC,MAAM,UAAU,MAAM,uBAAuB,KAAK;AAAA,EAElD,MAAM,gBAAyE,CAAC;AAAA,EAChF,MAAM,YACJ,MAAM,eAAe,SAAS,YAAY,uBAAuB,MAAM,OAAO,aAAa;AAAA,EAE7F,gBAAgB,IAAI,sBAAsB;AAAA,EAC1C,MAAM,WAAW,QAAQ,YAAY;AAAA,EACrC,QAAQ,qBAAqB;AAAA,EAC7B,MAAM,aAAa,oBAAoB,KAAK;AAAA,EAC5C,MAAM,UAAU,IAAI,iBAAiB;AAAA,IACnC,iBAAiB;AAAA,OACb,MAAM,gBAAgB,EAAE,cAAc,MAAM,aAAa;AAAA,EAC/D,CAAC;AAAA,EAED,IAAI;AAAA,IACF,MAAM,OAAO,MAAM,QAAQ,OAAO,YAAY;AAAA,MAC5C;AAAA,SACI,aAAa,EAAE,UAAU;AAAA,SACzB,MAAM,gBAAgB,aAAa,EAAE,aAAa,MAAM,YAAY;AAAA,SACpE,MAAM,cAAc,aAAa,EAAE,WAAW,MAAM,UAAU;AAAA,IACpE,CAAC;AAAA,IAED,MAAM,YAAuB,CAAC;AAAA,IAC9B,cAAc,QAAQ,CAAC,MAAM,UAAU;AAAA,MACrC,MAAM,KAAK,QAAQ;AAAA,MACnB,UAAU,KAAK,EAAE,IAAI,MAAM,KAAK,MAAM,OAAO,KAAK,MAAM,CAAC;AAAA,KAC1D;AAAA,IAED,gBAAgB,KAAK,uBAAuB;AAAA,IAC5C,OAAO,EAAE,MAAM,WAAW,qBAAqB,WAAW,MAAM,KAAK,EAAE;AAAA,YACvE;AAAA,IACA,SAAS,QAAQ;AAAA;AAAA;AAQd,IAAM,8BAIT,gBAAgB,CAAC,OAAO,OAAO,QAA2D;AAAA,EAC5F,IAAI,CAAC;AAAA,IAAO,MAAM,IAAI,MAAM,+CAA+C;AAAA,EAE3E,MAAM,QAAQ;AAAA,EAEd,MAAM,UAAU,MAAM,uBAAuB,KAAK;AAAA,EAElD,MAAM,gBAAyE,CAAC;AAAA,EAChF,MAAM,YACJ,MAAM,eAAe,SAAS,YAAY,uBAAuB,MAAM,OAAO,aAAa;AAAA,EAE7F,MAAM,WAAW,QAAQ,YAAY;AAAA,EACrC,QAAQ,qBAAqB;AAAA,EAC7B,MAAM,aAAa,oBAAoB,KAAK;AAAA,EAC5C,MAAM,UAAU,IAAI,iBAAiB;AAAA,IACnC,iBAAiB;AAAA,OACb,MAAM,gBAAgB,EAAE,cAAc,MAAM,aAAa;AAAA,EAC/D,CAAC;AAAA,EAED,MAAM,QAAkB,CAAC;AAAA,EACzB,IAAI,aAAa;AAAA,EACjB,IAAI;AAAA,EACJ,IAAI,cAAmC;AAAA,EAEvC,MAAM,eAAe,MAAM;AAAA,IACzB,cAAc;AAAA,IACd,cAAc;AAAA;AAAA,EAGhB,IAAI,kBAAkB;AAAA,EACtB,MAAM,gBAAgB,QACnB,OAAO,YAAY;AAAA,IAClB;AAAA,OACI,aAAa,EAAE,UAAU;AAAA,IAC7B,aAAa,CAAC,UAAkB;AAAA,MAC9B,MAAM,KAAK,KAAK;AAAA,MAChB,aAAa;AAAA;AAAA,OAEX,MAAM,gBAAgB,aAAa,EAAE,aAAa,MAAM,YAAY;AAAA,OACpE,MAAM,cAAc,aAAa,EAAE,WAAW,MAAM,UAAU;AAAA,EACpE,CAAC,EACA,KAAK,MAAM;AAAA,IACV,aAAa;AAAA,IACb,aAAa;AAAA,GACd,EACA,MAAM,CAAC,QAAiB;AAAA,IACvB,kBAAkB;AAAA,IAClB,aAAa;AAAA,IACb,aAAa;AAAA,GACd;AAAA,EAEH,IAAI;AAAA,IACF,OAAO,MAAM;AAAA,MACX,OAAO,MAAM,SAAS,GAAG;AAAA,QACvB,MAAM,QAAQ,MAAM,MAAM;AAAA,QAC1B,mBAAmB;AAAA,QACnB,MAAM,EAAE,MAAM,cAAc,MAAM,QAAQ,WAAW,MAAM;AAAA,MAC7D;AAAA,MACA,IAAI;AAAA,QAAY;AAAA,MAChB,MAAM,IAAI,QAAc,CAAC,MAAM;AAAA,QAC7B,cAAc;AAAA,OACf;AAAA,IACH;AAAA,IAEA,OAAO,MAAM,SAAS,GAAG;AAAA,MACvB,MAAM,QAAQ,MAAM,MAAM;AAAA,MAC1B,mBAAmB;AAAA,MACnB,MAAM,EAAE,MAAM,cAAc,MAAM,QAAQ,WAAW,MAAM;AAAA,IAC7D;AAAA,YACA;AAAA,IACA,MAAM,cAAc,MAAM,MAAM,EAAE;AAAA,IAClC,SAAS,QAAQ;AAAA;AAAA,EAGnB,IAAI,iBAAiB;AAAA,IACnB,IAAI,CAAC,OAAO;AAAA,MAAS,MAAM;AAAA,IAC3B;AAAA,EACF;AAAA,EAEA,MAAM,YAAuB,CAAC;AAAA,EAC9B,cAAc,QAAQ,CAAC,MAAM,UAAU;AAAA,IACrC,MAAM,KAAK,QAAQ;AAAA,IACnB,UAAU,KAAK,EAAE,IAAI,MAAM,KAAK,MAAM,OAAO,KAAK,MAAM,CAAC;AAAA,GAC1D;AAAA,EACD,MAAM,iBAAiB,qBAAqB,WAAW,MAAM,KAAK;AAAA,EAElE,IAAI,eAAe,SAAS,GAAG;AAAA,IAC7B,MAAM,EAAE,MAAM,gBAAgB,MAAM,aAAa,aAAa,CAAC,GAAG,cAAc,EAAE;AAAA,EACpF;AAAA,EAEA,MAAM;AAAA,IACJ,MAAM;AAAA,IACN,MAAM,EAAE,MAAM,iBAAiB,WAAW,eAAe;AAAA,EAC3D;AAAA;AAOK,IAAM,gCAIT,OAAO,OAAO,OAAO,iBAAiB,WAAW;AAAA,EACnD,IAAI,CAAC;AAAA,IAAO,MAAM,IAAI,MAAM,wDAAwD;AAAA,EAEpF,MAAM,QAAQ;AAAA,EAEd,gBAAgB,GAAG,eAAe;AAAA,EAClC,MAAM,QAAQ,MAAM,iBAAiB;AAAA,EACrC,MAAM,UAAU,MAAM,uBAAuB,KAAK;AAAA,EAElD,gBAAgB,IAAI,+BAA+B;AAAA,EACnD,MAAM,UAAU,MAAM,MAAM,2BAA2B,MAAM,YAAmB;AAAA,EAChF,MAAM,WAAW,QAAQ,YAAY;AAAA,EACrC,QAAQ,qBAAqB;AAAA,EAC7B,MAAM,UAAU,IAAI,iBAAiB,EAAE,iBAAiB,SAAS,CAAC;AAAA,EAElE,IAAI;AAAA,IACF,MAAM,OAAO,MAAM,QAAQ,OAAO,MAAM,QAAkB;AAAA,MACxD;AAAA,MACA;AAAA,SACI,MAAM,gBAAgB,aAAa,EAAE,aAAa,MAAM,YAAY;AAAA,SACpE,MAAM,cAAc,aAAa,EAAE,WAAW,MAAM,UAAU;AAAA,IACpE,CAAC;AAAA,IAED,IAAI;AAAA,IACJ,IAAI;AAAA,MACF,SAAS,KAAK,MAAM,IAAI;AAAA,MACxB,MAAM;AAAA,MACN,SAAS,CAAC;AAAA;AAAA,IAGZ,gBAAgB,KAAK,gCAAgC;AAAA,IACrD,OAAO,EAAE,OAAO;AAAA,YAChB;AAAA,IACA,SAAS,QAAQ;AAAA;AAAA;AAQd,IAAM,uCAIT,gBAAgB,CAClB,OACA,OACA,QAC4D;AAAA,EAC5D,IAAI,CAAC;AAAA,IAAO,MAAM,IAAI,MAAM,wDAAwD;AAAA,EAEpF,MAAM,QAAQ;AAAA,EAEd,MAAM,QAAQ,MAAM,iBAAiB;AAAA,EACrC,MAAM,UAAU,MAAM,uBAAuB,KAAK;AAAA,EAClD,MAAM,UAAU,MAAM,MAAM,2BAA2B,MAAM,YAAmB;AAAA,EAEhF,MAAM,WAAW,QAAQ,YAAY;AAAA,EACrC,QAAQ,qBAAqB;AAAA,EAC7B,MAAM,UAAU,IAAI,iBAAiB,EAAE,iBAAiB,SAAS,CAAC;AAAA,EAElE,MAAM,QAAkB,CAAC;AAAA,EACzB,IAAI,aAAa;AAAA,EACjB,IAAI;AAAA,EACJ,IAAI,cAAmC;AAAA,EAEvC,MAAM,eAAe,MAAM;AAAA,IACzB,cAAc;AAAA,IACd,cAAc;AAAA;AAAA,EAGhB,IAAI,kBAAkB;AAAA,EACtB,MAAM,gBAAgB,QACnB,OAAO,MAAM,QAAkB;AAAA,IAC9B;AAAA,IACA;AAAA,IACA,aAAa,CAAC,UAAkB;AAAA,MAC9B,MAAM,KAAK,KAAK;AAAA,MAChB,aAAa;AAAA;AAAA,OAEX,MAAM,gBAAgB,aAAa,EAAE,aAAa,MAAM,YAAY;AAAA,OACpE,MAAM,cAAc,aAAa,EAAE,WAAW,MAAM,UAAU;AAAA,EACpE,CAAC,EACA,KAAK,MAAM;AAAA,IACV,aAAa;AAAA,IACb,aAAa;AAAA,GACd,EACA,MAAM,CAAC,QAAiB;AAAA,IACvB,kBAAkB;AAAA,IAClB,aAAa;AAAA,IACb,aAAa;AAAA,GACd;AAAA,EAEH,IAAI;AAAA,IACF,OAAO,MAAM;AAAA,MACX,OAAO,MAAM,SAAS,GAAG;AAAA,QACvB,MAAM,QAAQ,MAAM,MAAM;AAAA,QAC1B,mBAAmB;AAAA,QAEnB,MAAM,UAAU,iBAAiB,eAAe;AAAA,QAChD,IAAI,YAAY,WAAW;AAAA,UACzB,MAAM;AAAA,YACJ,MAAM;AAAA,YACN,MAAM;AAAA,YACN,aAAa;AAAA,UACf;AAAA,QACF;AAAA,MACF;AAAA,MACA,IAAI;AAAA,QAAY;AAAA,MAChB,MAAM,IAAI,QAAc,CAAC,MAAM;AAAA,QAC7B,cAAc;AAAA,OACf;AAAA,IACH;AAAA,IAEA,OAAO,MAAM,SAAS,GAAG;AAAA,MACvB,MAAM,QAAQ,MAAM,MAAM;AAAA,MAC1B,mBAAmB;AAAA,IACrB;AAAA,YACA;AAAA,IACA,MAAM,cAAc,MAAM,MAAM,EAAE;AAAA,IAClC,SAAS,QAAQ;AAAA;AAAA,EAGnB,IAAI,iBAAiB;AAAA,IACnB,IAAI,OAAO;AAAA,MAAS;AAAA,IACpB,MAAM;AAAA,EACR;AAAA,EAEA,IAAI;AAAA,EACJ,IAAI;AAAA,IACF,cAAc,KAAK,MAAM,eAAe;AAAA,IACxC,MAAM;AAAA,IACN,cAAe,iBAAiB,eAAe,KAAiC,CAAC;AAAA;AAAA,EAGnF,MAAM,EAAE,MAAM,UAAU,MAAM,EAAE,QAAQ,YAAY,EAAoC;AAAA;AAOnF,IAAM,qBAIT,OAAO,OAAO,UAAU;AAAA,EAC1B,IAAI,CAAC;AAAA,IAAO,MAAM,IAAI,MAAM,6CAA6C;AAAA,EAEzE,MAAM,YAAY,mBAAmB,KAAK;AAAA,EAC1C,MAAM,YAAY,OAAO,IAAI,SAAS;AAAA,EAEtC,IAAI,YAAY;AAAA,EAChB,IAAI,aAA4C;AAAA,EAGhD,IAAI;AAAA,IACF,MAAM,KAAK,MAAa;AAAA,IACxB,MAAM,OAAO,MAAM,GAAG,KAAK,SAAS;AAAA,IACpC,YAAY;AAAA,IACZ,aAAa,EAAE,OAAO,KAAK,KAAK;AAAA,IAChC,MAAM;AAAA,IAGN,IAAI,cAAc,IAAI,aAAa,KAAK,CAAC,GAAG;AAAA,MAC1C,YAAY;AAAA,IACd;AAAA;AAAA,EAGF,OAAO;AAAA,IACL,OAAO,MAAM;AAAA,IACb,UAAU;AAAA,IACV,WAAW;AAAA,IACX,kBAAkB;AAAA,IAClB,eAAe;AAAA,IACf;AAAA,IACA;AAAA,IACA;AAAA,EACF;AAAA;AAOK,IAAM,iBAAiF;AAAA,EAC5F,mBAAmB;AAAA,EACnB,iBAAiB;AAAA,EACjB,eAAe;AAAA,EACf,iBAAiB;AAAA,EACjB,oBAAoB;AAAA,EACpB,mBAAmB;AAAA,EACnB,kBAAkB;AAAA,EAClB,iBAAiB;AAAA,EACjB,iBAAiB;AAAA,EACjB,0BAA0B;AAC5B;AAEO,IAAM,wBAGT;AAAA,EACF,oBAAoB;AAAA,EACpB,kBAAkB;AAAA,EAClB,iBAAiB;AAAA,EACjB,iBAAiB;AAAA,EACjB,0BAA0B;AAC5B;AAEO,IAAM,0BAGT;AAAA,EACF,iBAAiB;AACnB;;ACpiCA,sBAAS;AAQF,SAAS,+BAA+B,GAAG;AAAA,EAChD,MAAM,eAAe,sBAAsB,IAAI,aAAa;AAAA,EAC5D,IAAI,iBACF,gBACA,uBACA,uBACF,EAAE,uBAAuB,YAAY;AAAA,EACrC,aAAa,UAAU;AAAA,EACvB,WAAU,EAAE,KAAK,mCAAmC;AAAA;",
|
|
9
|
-
"debugId": "
|
|
10
|
+
"mappings": ";;;;;;;;;;;;;;;;;;;;;;;;AAMO,IAAM,iBAAiB;AACvB,IAAM,8BAA8B;;ACD3C;AAIO,IAAM,sBAAsB;AAAA,EACjC,MAAM;AAAA,EACN,YAAY;AAAA,IACV,UAAU;AAAA,MACR,OAAO;AAAA,MACP,aAAa;AAAA,IACf;AAAA,IACA,iBAAiB;AAAA,MACf,MAAM;AAAA,MACN,aAAa;AAAA,MACb,YAAY;AAAA,QACV,YAAY;AAAA,UACV,MAAM;AAAA,UACN,aAAa;AAAA,QACf;AAAA,QACA,WAAW;AAAA,UACT,MAAM;AAAA,UACN,aACE;AAAA,QACJ;AAAA,QACA,YAAY;AAAA,UACV,MAAM;AAAA,UACN,aAAa;AAAA,UACb,SAAS;AAAA,QACX;AAAA,QACA,YAAY;AAAA,UACV,MAAM;AAAA,UACN,aAAa;AAAA,UACb,SAAS;AAAA,QACX;AAAA,QACA,cAAc;AAAA,UACZ,MAAM;AAAA,UACN,aAAa;AAAA,UACb,SAAS;AAAA,QACX;AAAA,QACA,iBAAiB;AAAA,UACf,MAAM;AAAA,UACN,aAAa;AAAA,UACb,SAAS;AAAA,QACX;AAAA,QACA,WAAW;AAAA,UACT,MAAM;AAAA,UACN,aAAa;AAAA,UACb,SAAS;AAAA,QACX;AAAA,MACF;AAAA,MACA,UAAU,CAAC,YAAY;AAAA,MACvB,sBAAsB;AAAA,IACxB;AAAA,EACF;AAAA,EACA,UAAU,CAAC,YAAY,iBAAiB;AAAA,EACxC,sBAAsB;AACxB;AAEO,IAAM,4BAA4B;AAAA,EACvC,MAAM;AAAA,EACN,YAAY;AAAA,OACP,kBAAkB;AAAA,OAClB,oBAAoB;AAAA,EACzB;AAAA,EACA,UAAU,CAAC,GAAG,kBAAkB,UAAU,GAAG,oBAAoB,QAAQ;AAAA,EACzE,sBAAsB;AACxB;AAIO,IAAM,4BAA4B;AAAA,EACvC,MAAM;AAAA,EACN,YAAY;AAAA,OACP,kBAAkB;AAAA,OAClB,oBAAoB;AAAA,EACzB;AAAA,EACA,UAAU,CAAC,GAAG,kBAAkB,UAAU,GAAG,oBAAoB,QAAQ;AAAA,EACzE,sBAAsB;AACxB;;AC9EA;AAAA;AAAA;AAUO,MAAM,+BAA+B,iBAAsC;AAAA,EACvE,OAAO;AAAA,EACP,UAAU;AAAA,EACV,kBAAkB;AAAA,EAElB,YAAY;AAAA,IACnB;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,EACF;AAAA,EAEA,WAAW,CACT,OACA,aACA,eACA;AAAA,IACA,MAAM,OAAO,aAAa,aAAa;AAAA;AAE3C;;;AChCA,eAAsB,gBAAgB,CACpC,SAGe;AAAA,EACf,MAAM,IAAI,uBAAuB,EAAE,SAAS,OAAO;AAAA;",
|
|
11
|
+
"debugId": "5DCB59F869EA8BFD64756E2164756E21",
|
|
10
12
|
"names": []
|
|
11
13
|
}
|
|
@@ -0,0 +1,10 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* @license
|
|
3
|
+
* Copyright 2025 Steven Roussey <sroussey@gmail.com>
|
|
4
|
+
* SPDX-License-Identifier: Apache-2.0
|
|
5
|
+
*/
|
|
6
|
+
import type { AiProviderRegisterOptions } from "@workglow/ai";
|
|
7
|
+
export declare function registerLlamaCpp(options: AiProviderRegisterOptions & {
|
|
8
|
+
worker: Worker | (() => Worker);
|
|
9
|
+
}): Promise<void>;
|
|
10
|
+
//# sourceMappingURL=registerLlamaCpp.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"registerLlamaCpp.d.ts","sourceRoot":"","sources":["../../src/provider-llamacpp/registerLlamaCpp.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AAEH,OAAO,KAAK,EAAE,yBAAyB,EAAE,MAAM,cAAc,CAAC;AAG9D,wBAAsB,gBAAgB,CACpC,OAAO,EAAE,yBAAyB,GAAG;IACnC,MAAM,EAAE,MAAM,GAAG,CAAC,MAAM,MAAM,CAAC,CAAC;CACjC,GACA,OAAO,CAAC,IAAI,CAAC,CAEf"}
|
|
@@ -0,0 +1,8 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* @license
|
|
3
|
+
* Copyright 2025 Steven Roussey <sroussey@gmail.com>
|
|
4
|
+
* SPDX-License-Identifier: Apache-2.0
|
|
5
|
+
*/
|
|
6
|
+
import type { AiProviderRegisterOptions } from "@workglow/ai";
|
|
7
|
+
export declare function registerLlamaCppInline(options?: AiProviderRegisterOptions): Promise<void>;
|
|
8
|
+
//# sourceMappingURL=registerLlamaCppInline.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"registerLlamaCppInline.d.ts","sourceRoot":"","sources":["../../src/provider-llamacpp/registerLlamaCppInline.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AAEH,OAAO,KAAK,EAAE,yBAAyB,EAAE,MAAM,cAAc,CAAC;AAQ9D,wBAAsB,sBAAsB,CAAC,OAAO,CAAC,EAAE,yBAAyB,GAAG,OAAO,CAAC,IAAI,CAAC,CAM/F"}
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"registerLlamaCppWorker.d.ts","sourceRoot":"","sources":["../../src/provider-llamacpp/registerLlamaCppWorker.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AAUH,wBAAsB,sBAAsB,IAAI,OAAO,CAAC,IAAI,CAAC,CAS5D"}
|
|
@@ -0,0 +1,16 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* @license
|
|
3
|
+
* Copyright 2025 Steven Roussey <sroussey@gmail.com>
|
|
4
|
+
* SPDX-License-Identifier: Apache-2.0
|
|
5
|
+
*/
|
|
6
|
+
/**
|
|
7
|
+
* Worker server and main-thread inline LlamaCpp registration (pulls in `LlamaCpp_JobRunFns`),
|
|
8
|
+
* plus runtime helpers (`LlamaCpp_Runtime`, e.g. `disposeLlamaCppResources`).
|
|
9
|
+
* Import from `@workglow/ai-provider/llamacpp/runtime` — not from the main `llamacpp` barrel.
|
|
10
|
+
*
|
|
11
|
+
* Use `export *` (not `export { … } from "…"`) so the Bun bundler keeps the module graph.
|
|
12
|
+
*/
|
|
13
|
+
export * from "./common/LlamaCpp_Runtime";
|
|
14
|
+
export * from "./registerLlamaCppInline";
|
|
15
|
+
export * from "./registerLlamaCppWorker";
|
|
16
|
+
//# sourceMappingURL=runtime.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"runtime.d.ts","sourceRoot":"","sources":["../../src/provider-llamacpp/runtime.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AAEH;;;;;;GAMG;AACH,cAAc,2BAA2B,CAAC;AAC1C,cAAc,0BAA0B,CAAC;AACzC,cAAc,0BAA0B,CAAC"}
|