@workglow/ai-provider 0.0.125 → 0.1.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/common/HfModelSearch.d.ts +1 -0
- package/dist/common/HfModelSearch.d.ts.map +1 -1
- package/dist/common/modelSearchQuery.d.ts +25 -0
- package/dist/common/modelSearchQuery.d.ts.map +1 -0
- package/dist/provider-anthropic/AnthropicProvider.d.ts +1 -1
- package/dist/provider-anthropic/AnthropicProvider.d.ts.map +1 -1
- package/dist/provider-anthropic/AnthropicQueuedProvider.d.ts +4 -4
- package/dist/provider-anthropic/AnthropicQueuedProvider.d.ts.map +1 -1
- package/dist/provider-anthropic/common/Anthropic_Client.d.ts.map +1 -1
- package/dist/provider-anthropic/common/Anthropic_CountTokens.d.ts.map +1 -1
- package/dist/provider-anthropic/common/Anthropic_JobRunFns.d.ts.map +1 -1
- package/dist/provider-anthropic/common/Anthropic_ModelSearch.d.ts.map +1 -1
- package/dist/provider-anthropic/common/Anthropic_TextGeneration.d.ts.map +1 -1
- package/dist/provider-anthropic/common/Anthropic_TextRewriter.d.ts.map +1 -1
- package/dist/provider-anthropic/common/Anthropic_TextSummary.d.ts.map +1 -1
- package/dist/provider-anthropic/index.js +3 -4
- package/dist/provider-anthropic/index.js.map +3 -3
- package/dist/provider-anthropic/runtime.js +39 -292
- package/dist/provider-anthropic/runtime.js.map +13 -13
- package/dist/provider-chrome/common/WebBrowser_ModelSearch.d.ts.map +1 -1
- package/dist/provider-chrome/common/WebBrowser_TextGeneration.d.ts.map +1 -1
- package/dist/provider-chrome/common/WebBrowser_TextLanguageDetection.d.ts.map +1 -1
- package/dist/provider-chrome/common/WebBrowser_TextRewriter.d.ts.map +1 -1
- package/dist/provider-chrome/common/WebBrowser_TextSummary.d.ts.map +1 -1
- package/dist/provider-chrome/common/WebBrowser_TextTranslation.d.ts.map +1 -1
- package/dist/provider-chrome/runtime.js +24 -99
- package/dist/provider-chrome/runtime.js.map +10 -9
- package/dist/provider-gemini/GoogleGeminiProvider.d.ts +1 -1
- package/dist/provider-gemini/GoogleGeminiProvider.d.ts.map +1 -1
- package/dist/provider-gemini/GoogleGeminiQueuedProvider.d.ts +4 -4
- package/dist/provider-gemini/GoogleGeminiQueuedProvider.d.ts.map +1 -1
- package/dist/provider-gemini/common/Gemini_CountTokens.d.ts.map +1 -1
- package/dist/provider-gemini/common/Gemini_JobRunFns.d.ts.map +1 -1
- package/dist/provider-gemini/common/Gemini_ModelSearch.d.ts.map +1 -1
- package/dist/provider-gemini/common/Gemini_TextEmbedding.d.ts.map +1 -1
- package/dist/provider-gemini/common/Gemini_TextGeneration.d.ts.map +1 -1
- package/dist/provider-gemini/common/Gemini_TextRewriter.d.ts.map +1 -1
- package/dist/provider-gemini/common/Gemini_TextSummary.d.ts.map +1 -1
- package/dist/provider-gemini/index.js +3 -4
- package/dist/provider-gemini/index.js.map +3 -3
- package/dist/provider-gemini/runtime.js +33 -259
- package/dist/provider-gemini/runtime.js.map +13 -13
- package/dist/provider-hf-inference/HfInferenceProvider.d.ts +1 -1
- package/dist/provider-hf-inference/HfInferenceProvider.d.ts.map +1 -1
- package/dist/provider-hf-inference/HfInferenceQueuedProvider.d.ts +4 -4
- package/dist/provider-hf-inference/HfInferenceQueuedProvider.d.ts.map +1 -1
- package/dist/provider-hf-inference/common/HFI_Client.d.ts.map +1 -1
- package/dist/provider-hf-inference/common/HFI_JobRunFns.d.ts.map +1 -1
- package/dist/provider-hf-inference/common/HFI_ModelSearch.d.ts.map +1 -1
- package/dist/provider-hf-inference/common/HFI_TextEmbedding.d.ts.map +1 -1
- package/dist/provider-hf-inference/common/HFI_TextGeneration.d.ts.map +1 -1
- package/dist/provider-hf-inference/common/HFI_TextRewriter.d.ts.map +1 -1
- package/dist/provider-hf-inference/common/HFI_TextSummary.d.ts.map +1 -1
- package/dist/provider-hf-inference/index.js +3 -4
- package/dist/provider-hf-inference/index.js.map +3 -3
- package/dist/provider-hf-inference/runtime.js +14 -207
- package/dist/provider-hf-inference/runtime.js.map +13 -14
- package/dist/provider-hf-transformers/HuggingFaceTransformersProvider.d.ts +1 -1
- package/dist/provider-hf-transformers/HuggingFaceTransformersProvider.d.ts.map +1 -1
- package/dist/provider-hf-transformers/HuggingFaceTransformersQueuedProvider.d.ts +13 -3
- package/dist/provider-hf-transformers/HuggingFaceTransformersQueuedProvider.d.ts.map +1 -1
- package/dist/provider-hf-transformers/common/HFT_Constants.d.ts +4 -0
- package/dist/provider-hf-transformers/common/HFT_Constants.d.ts.map +1 -1
- package/dist/provider-hf-transformers/common/HFT_CountTokens.d.ts.map +1 -1
- package/dist/provider-hf-transformers/common/HFT_Download.d.ts.map +1 -1
- package/dist/provider-hf-transformers/common/HFT_ImageEmbedding.d.ts.map +1 -1
- package/dist/provider-hf-transformers/common/HFT_JobRunFns.d.ts +117 -88
- package/dist/provider-hf-transformers/common/HFT_JobRunFns.d.ts.map +1 -1
- package/dist/provider-hf-transformers/common/HFT_ModelInfo.d.ts.map +1 -1
- package/dist/provider-hf-transformers/common/HFT_ModelSchema.d.ts +30 -0
- package/dist/provider-hf-transformers/common/HFT_ModelSchema.d.ts.map +1 -1
- package/dist/provider-hf-transformers/common/HFT_ModelSearch.d.ts.map +1 -1
- package/dist/provider-hf-transformers/common/HFT_OnnxDtypes.d.ts.map +1 -1
- package/dist/provider-hf-transformers/common/HFT_Pipeline.d.ts +9 -2
- package/dist/provider-hf-transformers/common/HFT_Pipeline.d.ts.map +1 -1
- package/dist/provider-hf-transformers/common/HFT_Streaming.d.ts +2 -2
- package/dist/provider-hf-transformers/common/HFT_Streaming.d.ts.map +1 -1
- package/dist/provider-hf-transformers/common/HFT_TextClassification.d.ts.map +1 -1
- package/dist/provider-hf-transformers/common/HFT_TextFillMask.d.ts.map +1 -1
- package/dist/provider-hf-transformers/common/HFT_TextGeneration.d.ts.map +1 -1
- package/dist/provider-hf-transformers/common/HFT_TextLanguageDetection.d.ts.map +1 -1
- package/dist/provider-hf-transformers/common/HFT_TextNamedEntityRecognition.d.ts.map +1 -1
- package/dist/provider-hf-transformers/common/HFT_TextQuestionAnswer.d.ts.map +1 -1
- package/dist/provider-hf-transformers/common/HFT_TextRewriter.d.ts.map +1 -1
- package/dist/provider-hf-transformers/common/HFT_TextSummary.d.ts.map +1 -1
- package/dist/provider-hf-transformers/common/HFT_TextTranslation.d.ts.map +1 -1
- package/dist/provider-hf-transformers/index.d.ts +0 -1
- package/dist/provider-hf-transformers/index.d.ts.map +1 -1
- package/dist/provider-hf-transformers/index.js +49 -177
- package/dist/provider-hf-transformers/index.js.map +8 -9
- package/dist/provider-hf-transformers/registerHuggingFaceTransformersWorker.d.ts.map +1 -1
- package/dist/provider-hf-transformers/runtime.d.ts +0 -1
- package/dist/provider-hf-transformers/runtime.d.ts.map +1 -1
- package/dist/provider-hf-transformers/runtime.js +209 -514
- package/dist/provider-hf-transformers/runtime.js.map +29 -31
- package/dist/provider-llamacpp/LlamaCppProvider.d.ts +1 -1
- package/dist/provider-llamacpp/LlamaCppProvider.d.ts.map +1 -1
- package/dist/provider-llamacpp/LlamaCppQueuedProvider.d.ts +1 -1
- package/dist/provider-llamacpp/LlamaCppQueuedProvider.d.ts.map +1 -1
- package/dist/provider-llamacpp/common/LlamaCpp_CountTokens.d.ts.map +1 -1
- package/dist/provider-llamacpp/common/LlamaCpp_JobRunFns.d.ts.map +1 -1
- package/dist/provider-llamacpp/common/LlamaCpp_ModelSchema.d.ts +15 -0
- package/dist/provider-llamacpp/common/LlamaCpp_ModelSchema.d.ts.map +1 -1
- package/dist/provider-llamacpp/common/LlamaCpp_ModelSearch.d.ts.map +1 -1
- package/dist/provider-llamacpp/common/LlamaCpp_Runtime.d.ts +10 -0
- package/dist/provider-llamacpp/common/LlamaCpp_Runtime.d.ts.map +1 -1
- package/dist/provider-llamacpp/common/LlamaCpp_StructuredGeneration.d.ts.map +1 -1
- package/dist/provider-llamacpp/common/LlamaCpp_TextEmbedding.d.ts.map +1 -1
- package/dist/provider-llamacpp/common/LlamaCpp_TextGeneration.d.ts.map +1 -1
- package/dist/provider-llamacpp/common/LlamaCpp_TextRewriter.d.ts.map +1 -1
- package/dist/provider-llamacpp/common/LlamaCpp_TextSummary.d.ts.map +1 -1
- package/dist/provider-llamacpp/index.js +6 -2
- package/dist/provider-llamacpp/index.js.map +4 -4
- package/dist/provider-llamacpp/runtime.js +83 -231
- package/dist/provider-llamacpp/runtime.js.map +15 -16
- package/dist/provider-ollama/OllamaProvider.d.ts +1 -1
- package/dist/provider-ollama/OllamaProvider.d.ts.map +1 -1
- package/dist/provider-ollama/OllamaQueuedProvider.d.ts +4 -4
- package/dist/provider-ollama/OllamaQueuedProvider.d.ts.map +1 -1
- package/dist/provider-ollama/common/Ollama_JobRunFns.browser.d.ts +14 -72
- package/dist/provider-ollama/common/Ollama_JobRunFns.browser.d.ts.map +1 -1
- package/dist/provider-ollama/common/Ollama_JobRunFns.d.ts +14 -72
- package/dist/provider-ollama/common/Ollama_JobRunFns.d.ts.map +1 -1
- package/dist/provider-ollama/common/Ollama_ModelSearch.d.ts.map +1 -1
- package/dist/provider-ollama/common/Ollama_TextGeneration.d.ts.map +1 -1
- package/dist/provider-ollama/common/Ollama_TextRewriter.d.ts.map +1 -1
- package/dist/provider-ollama/common/Ollama_TextSummary.d.ts.map +1 -1
- package/dist/provider-ollama/index.browser.js +3 -4
- package/dist/provider-ollama/index.browser.js.map +3 -3
- package/dist/provider-ollama/index.js +3 -4
- package/dist/provider-ollama/index.js.map +3 -3
- package/dist/provider-ollama/runtime.browser.js +28 -181
- package/dist/provider-ollama/runtime.browser.js.map +11 -11
- package/dist/provider-ollama/runtime.js +28 -176
- package/dist/provider-ollama/runtime.js.map +11 -11
- package/dist/provider-openai/OpenAiProvider.d.ts +1 -1
- package/dist/provider-openai/OpenAiProvider.d.ts.map +1 -1
- package/dist/provider-openai/OpenAiQueuedProvider.d.ts +4 -4
- package/dist/provider-openai/OpenAiQueuedProvider.d.ts.map +1 -1
- package/dist/provider-openai/common/OpenAI_Client.d.ts.map +1 -1
- package/dist/provider-openai/common/OpenAI_CountTokens.browser.d.ts.map +1 -1
- package/dist/provider-openai/common/OpenAI_CountTokens.d.ts.map +1 -1
- package/dist/provider-openai/common/OpenAI_JobRunFns.browser.d.ts.map +1 -1
- package/dist/provider-openai/common/OpenAI_JobRunFns.d.ts.map +1 -1
- package/dist/provider-openai/common/OpenAI_ModelSearch.d.ts.map +1 -1
- package/dist/provider-openai/common/OpenAI_TextEmbedding.d.ts.map +1 -1
- package/dist/provider-openai/common/OpenAI_TextGeneration.d.ts.map +1 -1
- package/dist/provider-openai/common/OpenAI_TextRewriter.d.ts.map +1 -1
- package/dist/provider-openai/common/OpenAI_TextSummary.d.ts.map +1 -1
- package/dist/provider-openai/index.browser.js +3 -4
- package/dist/provider-openai/index.browser.js.map +3 -3
- package/dist/provider-openai/index.js +3 -4
- package/dist/provider-openai/index.js.map +3 -3
- package/dist/provider-openai/runtime.browser.js +42 -225
- package/dist/provider-openai/runtime.browser.js.map +14 -14
- package/dist/provider-openai/runtime.js +42 -225
- package/dist/provider-openai/runtime.js.map +14 -14
- package/dist/provider-tf-mediapipe/TensorFlowMediaPipeQueuedProvider.d.ts +3 -3
- package/dist/provider-tf-mediapipe/TensorFlowMediaPipeQueuedProvider.d.ts.map +1 -1
- package/dist/provider-tf-mediapipe/common/TFMP_ImageEmbedding.d.ts.map +1 -1
- package/dist/provider-tf-mediapipe/common/TFMP_JobRunFns.d.ts +18 -11
- package/dist/provider-tf-mediapipe/common/TFMP_JobRunFns.d.ts.map +1 -1
- package/dist/provider-tf-mediapipe/common/TFMP_ModelSearch.d.ts.map +1 -1
- package/dist/provider-tf-mediapipe/common/TFMP_Unload.d.ts.map +1 -1
- package/dist/provider-tf-mediapipe/index.js +3 -3
- package/dist/provider-tf-mediapipe/index.js.map +3 -3
- package/dist/provider-tf-mediapipe/runtime.js +37 -7
- package/dist/provider-tf-mediapipe/runtime.js.map +8 -7
- package/package.json +31 -19
- package/dist/provider-anthropic/common/Anthropic_ToolCalling.d.ts +0 -10
- package/dist/provider-anthropic/common/Anthropic_ToolCalling.d.ts.map +0 -1
- package/dist/provider-gemini/common/Gemini_ToolCalling.d.ts +0 -10
- package/dist/provider-gemini/common/Gemini_ToolCalling.d.ts.map +0 -1
- package/dist/provider-hf-inference/common/HFI_ToolCalling.d.ts +0 -10
- package/dist/provider-hf-inference/common/HFI_ToolCalling.d.ts.map +0 -1
- package/dist/provider-hf-transformers/common/HFT_ToolCalling.d.ts +0 -10
- package/dist/provider-hf-transformers/common/HFT_ToolCalling.d.ts.map +0 -1
- package/dist/provider-hf-transformers/common/HFT_ToolMarkup.d.ts +0 -40
- package/dist/provider-hf-transformers/common/HFT_ToolMarkup.d.ts.map +0 -1
- package/dist/provider-llamacpp/common/LlamaCpp_ToolCalling.d.ts +0 -10
- package/dist/provider-llamacpp/common/LlamaCpp_ToolCalling.d.ts.map +0 -1
- package/dist/provider-ollama/common/Ollama_ToolCalling.d.ts +0 -16
- package/dist/provider-ollama/common/Ollama_ToolCalling.d.ts.map +0 -1
- package/dist/provider-openai/common/OpenAI_ToolCalling.d.ts +0 -10
- package/dist/provider-openai/common/OpenAI_ToolCalling.d.ts.map +0 -1
|
@@ -3,7 +3,7 @@
|
|
|
3
3
|
* Copyright 2025 Steven Roussey <sroussey@gmail.com>
|
|
4
4
|
* SPDX-License-Identifier: Apache-2.0
|
|
5
5
|
*/
|
|
6
|
-
import {
|
|
6
|
+
import type { AiProviderRunFn, AiProviderStreamFn } from "@workglow/ai/worker";
|
|
7
7
|
import type { OllamaModelConfig } from "./Ollama_ModelSchema";
|
|
8
8
|
export declare const Ollama_TextGeneration: AiProviderRunFn<{
|
|
9
9
|
frequencyPenalty?: number | undefined;
|
|
@@ -24,11 +24,11 @@ export declare const Ollama_TextGeneration: AiProviderRunFn<{
|
|
|
24
24
|
title?: string | undefined;
|
|
25
25
|
};
|
|
26
26
|
presencePenalty?: number | undefined;
|
|
27
|
-
prompt: string
|
|
27
|
+
prompt: string;
|
|
28
28
|
temperature?: number | undefined;
|
|
29
29
|
topP?: number | undefined;
|
|
30
30
|
}, {
|
|
31
|
-
text: string
|
|
31
|
+
text: string;
|
|
32
32
|
}, {
|
|
33
33
|
description?: string | undefined;
|
|
34
34
|
metadata?: {
|
|
@@ -93,9 +93,9 @@ export declare const Ollama_TextRewriter: AiProviderRunFn<{
|
|
|
93
93
|
title?: string | undefined;
|
|
94
94
|
};
|
|
95
95
|
prompt: string;
|
|
96
|
-
text: string
|
|
96
|
+
text: string;
|
|
97
97
|
}, {
|
|
98
|
-
text: string
|
|
98
|
+
text: string;
|
|
99
99
|
}, {
|
|
100
100
|
description?: string | undefined;
|
|
101
101
|
metadata?: {
|
|
@@ -126,9 +126,9 @@ export declare const Ollama_TextSummary: AiProviderRunFn<{
|
|
|
126
126
|
tasks?: string[] | undefined;
|
|
127
127
|
title?: string | undefined;
|
|
128
128
|
};
|
|
129
|
-
text: string
|
|
129
|
+
text: string;
|
|
130
130
|
}, {
|
|
131
|
-
text: string
|
|
131
|
+
text: string;
|
|
132
132
|
}, {
|
|
133
133
|
description?: string | undefined;
|
|
134
134
|
metadata?: {
|
|
@@ -162,11 +162,11 @@ export declare const Ollama_TextGeneration_Stream: AiProviderStreamFn<{
|
|
|
162
162
|
title?: string | undefined;
|
|
163
163
|
};
|
|
164
164
|
presencePenalty?: number | undefined;
|
|
165
|
-
prompt: string
|
|
165
|
+
prompt: string;
|
|
166
166
|
temperature?: number | undefined;
|
|
167
167
|
topP?: number | undefined;
|
|
168
168
|
}, {
|
|
169
|
-
text: string
|
|
169
|
+
text: string;
|
|
170
170
|
}, {
|
|
171
171
|
description?: string | undefined;
|
|
172
172
|
metadata?: {
|
|
@@ -198,9 +198,9 @@ export declare const Ollama_TextRewriter_Stream: AiProviderStreamFn<{
|
|
|
198
198
|
title?: string | undefined;
|
|
199
199
|
};
|
|
200
200
|
prompt: string;
|
|
201
|
-
text: string
|
|
201
|
+
text: string;
|
|
202
202
|
}, {
|
|
203
|
-
text: string
|
|
203
|
+
text: string;
|
|
204
204
|
}, {
|
|
205
205
|
description?: string | undefined;
|
|
206
206
|
metadata?: {
|
|
@@ -231,67 +231,9 @@ export declare const Ollama_TextSummary_Stream: AiProviderStreamFn<{
|
|
|
231
231
|
tasks?: string[] | undefined;
|
|
232
232
|
title?: string | undefined;
|
|
233
233
|
};
|
|
234
|
-
text: string
|
|
235
|
-
}, {
|
|
236
|
-
text: string | string[];
|
|
237
|
-
}, {
|
|
238
|
-
description?: string | undefined;
|
|
239
|
-
metadata?: {
|
|
240
|
-
[x: string]: unknown;
|
|
241
|
-
} | undefined;
|
|
242
|
-
model_id?: string | undefined;
|
|
243
|
-
provider: "OLLAMA";
|
|
244
|
-
provider_config: {
|
|
245
|
-
base_url?: string | undefined;
|
|
246
|
-
model_name: string;
|
|
247
|
-
};
|
|
248
|
-
tasks?: string[] | undefined;
|
|
249
|
-
title?: string | undefined;
|
|
250
|
-
}>;
|
|
251
|
-
export declare const Ollama_ToolCalling: AiProviderRunFn<import("@workglow/ai").ToolCallingTaskInput, {
|
|
252
|
-
text: string | string[];
|
|
253
|
-
toolCalls: {
|
|
254
|
-
id: string;
|
|
255
|
-
input: {
|
|
256
|
-
[x: string]: unknown;
|
|
257
|
-
};
|
|
258
|
-
name: string;
|
|
259
|
-
}[] | {
|
|
260
|
-
id: string;
|
|
261
|
-
input: {
|
|
262
|
-
[x: string]: unknown;
|
|
263
|
-
};
|
|
264
|
-
name: string;
|
|
265
|
-
};
|
|
234
|
+
text: string;
|
|
266
235
|
}, {
|
|
267
|
-
|
|
268
|
-
metadata?: {
|
|
269
|
-
[x: string]: unknown;
|
|
270
|
-
} | undefined;
|
|
271
|
-
model_id?: string | undefined;
|
|
272
|
-
provider: "OLLAMA";
|
|
273
|
-
provider_config: {
|
|
274
|
-
base_url?: string | undefined;
|
|
275
|
-
model_name: string;
|
|
276
|
-
};
|
|
277
|
-
tasks?: string[] | undefined;
|
|
278
|
-
title?: string | undefined;
|
|
279
|
-
}>;
|
|
280
|
-
export declare const Ollama_ToolCalling_Stream: AiProviderStreamFn<import("@workglow/ai").ToolCallingTaskInput, {
|
|
281
|
-
text: string | string[];
|
|
282
|
-
toolCalls: {
|
|
283
|
-
id: string;
|
|
284
|
-
input: {
|
|
285
|
-
[x: string]: unknown;
|
|
286
|
-
};
|
|
287
|
-
name: string;
|
|
288
|
-
}[] | {
|
|
289
|
-
id: string;
|
|
290
|
-
input: {
|
|
291
|
-
[x: string]: unknown;
|
|
292
|
-
};
|
|
293
|
-
name: string;
|
|
294
|
-
};
|
|
236
|
+
text: string;
|
|
295
237
|
}, {
|
|
296
238
|
description?: string | undefined;
|
|
297
239
|
metadata?: {
|
|
@@ -363,7 +305,7 @@ export declare const Ollama_ModelInfo: AiProviderRunFn<{
|
|
|
363
305
|
}>;
|
|
364
306
|
export declare const Ollama_ModelSearch: AiProviderRunFn<{
|
|
365
307
|
provider: string;
|
|
366
|
-
query
|
|
308
|
+
query?: string | undefined;
|
|
367
309
|
}, import("@workglow/ai").ModelSearchTaskOutput>;
|
|
368
310
|
export declare const OLLAMA_TASKS: Record<string, AiProviderRunFn<any, any, OllamaModelConfig>>;
|
|
369
311
|
export declare const OLLAMA_STREAM_TASKS: Record<string, AiProviderStreamFn<any, any, OllamaModelConfig>>;
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"Ollama_JobRunFns.d.ts","sourceRoot":"","sources":["../../../src/provider-ollama/common/Ollama_JobRunFns.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AAEH,OAAO,
|
|
1
|
+
{"version":3,"file":"Ollama_JobRunFns.d.ts","sourceRoot":"","sources":["../../../src/provider-ollama/common/Ollama_JobRunFns.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AAEH,OAAO,KAAK,EAAE,eAAe,EAAE,kBAAkB,EAAE,MAAM,qBAAqB,CAAC;AAC/E,OAAO,KAAK,EAAE,iBAAiB,EAAE,MAAM,sBAAsB,CAAC;AAW9D,eAAO,MAAM,qBAAqB;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;EAAwC,CAAC;AAC3E,eAAO,MAAM,oBAAoB;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;EAAuC,CAAC;AACzE,eAAO,MAAM,mBAAmB;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;EAAsC,CAAC;AACvE,eAAO,MAAM,kBAAkB;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;EAAqC,CAAC;AAErE,eAAO,MAAM,4BAA4B;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;EAA8C,CAAC;AACxF,eAAO,MAAM,0BAA0B;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;EAA4C,CAAC;AACpF,eAAO,MAAM,yBAAyB;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;EAA2C,CAAC;AAElF,eAAO,MAAM,gBAAgB;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;EAAmC,CAAC;AACjE,eAAO,MAAM,kBAAkB;;;gDAAqC,CAAC;AAErE,eAAO,MAAM,YAAY,EAAE,MAAM,CAAC,MAAM,EAAE,eAAe,CAAC,GAAG,EAAE,GAAG,EAAE,iBAAiB,CAAC,CAOrF,CAAC;AAEF,eAAO,MAAM,mBAAmB,EAAE,MAAM,CACtC,MAAM,EACN,kBAAkB,CAAC,GAAG,EAAE,GAAG,EAAE,iBAAiB,CAAC,CAKhD,CAAC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"Ollama_ModelSearch.d.ts","sourceRoot":"","sources":["../../../src/provider-ollama/common/Ollama_ModelSearch.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AAEH,OAAO,KAAK,EAAE,eAAe,EAAE,oBAAoB,EAAE,qBAAqB,EAAE,MAAM,cAAc,CAAC;
|
|
1
|
+
{"version":3,"file":"Ollama_ModelSearch.d.ts","sourceRoot":"","sources":["../../../src/provider-ollama/common/Ollama_ModelSearch.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AAEH,OAAO,KAAK,EAAE,eAAe,EAAE,oBAAoB,EAAE,qBAAqB,EAAE,MAAM,cAAc,CAAC;AAGjG,OAAO,KAAK,EAAE,iBAAiB,EAAE,MAAM,sBAAsB,CAAC;AAE9D,KAAK,SAAS,GAAG,CAAC,KAAK,EAAE,iBAAiB,GAAG,SAAS,KAAK,OAAO,CAAC,GAAG,CAAC,CAAC;AAExE,wBAAgB,uBAAuB,CACrC,SAAS,EAAE,SAAS,GACnB,eAAe,CAAC,oBAAoB,EAAE,qBAAqB,CAAC,CAyB9D"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"Ollama_TextGeneration.d.ts","sourceRoot":"","sources":["../../../src/provider-ollama/common/Ollama_TextGeneration.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AAEH,OAAO,KAAK,EACV,eAAe,EACf,kBAAkB,EAClB,uBAAuB,EACvB,wBAAwB,EACzB,MAAM,cAAc,CAAC;
|
|
1
|
+
{"version":3,"file":"Ollama_TextGeneration.d.ts","sourceRoot":"","sources":["../../../src/provider-ollama/common/Ollama_TextGeneration.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AAEH,OAAO,KAAK,EACV,eAAe,EACf,kBAAkB,EAClB,uBAAuB,EACvB,wBAAwB,EACzB,MAAM,cAAc,CAAC;AAEtB,OAAO,KAAK,EAAE,iBAAiB,EAAE,MAAM,sBAAsB,CAAC;AAG9D,KAAK,SAAS,GAAG,CAAC,KAAK,EAAE,iBAAiB,GAAG,SAAS,KAAK,OAAO,CAAC,GAAG,CAAC,CAAC;AAExE,wBAAgB,0BAA0B,CACxC,SAAS,EAAE,SAAS,GACnB,eAAe,CAAC,uBAAuB,EAAE,wBAAwB,EAAE,iBAAiB,CAAC,CA0BvF;AAED,wBAAgB,gCAAgC,CAC9C,SAAS,EAAE,SAAS,GACnB,kBAAkB,CAAC,uBAAuB,EAAE,wBAAwB,EAAE,iBAAiB,CAAC,CAoC1F"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"Ollama_TextRewriter.d.ts","sourceRoot":"","sources":["../../../src/provider-ollama/common/Ollama_TextRewriter.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AAEH,OAAO,KAAK,EACV,eAAe,EACf,kBAAkB,EAClB,qBAAqB,EACrB,sBAAsB,EACvB,MAAM,cAAc,CAAC;
|
|
1
|
+
{"version":3,"file":"Ollama_TextRewriter.d.ts","sourceRoot":"","sources":["../../../src/provider-ollama/common/Ollama_TextRewriter.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AAEH,OAAO,KAAK,EACV,eAAe,EACf,kBAAkB,EAClB,qBAAqB,EACrB,sBAAsB,EACvB,MAAM,cAAc,CAAC;AAEtB,OAAO,KAAK,EAAE,iBAAiB,EAAE,MAAM,sBAAsB,CAAC;AAG9D,KAAK,SAAS,GAAG,CAAC,KAAK,EAAE,iBAAiB,GAAG,SAAS,KAAK,OAAO,CAAC,GAAG,CAAC,CAAC;AAExE,wBAAgB,wBAAwB,CACtC,SAAS,EAAE,SAAS,GACnB,eAAe,CAAC,qBAAqB,EAAE,sBAAsB,EAAE,iBAAiB,CAAC,CAsBnF;AAED,wBAAgB,8BAA8B,CAC5C,SAAS,EAAE,SAAS,GACnB,kBAAkB,CAAC,qBAAqB,EAAE,sBAAsB,EAAE,iBAAiB,CAAC,CAgCtF"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"Ollama_TextSummary.d.ts","sourceRoot":"","sources":["../../../src/provider-ollama/common/Ollama_TextSummary.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AAEH,OAAO,KAAK,EACV,eAAe,EACf,kBAAkB,EAClB,oBAAoB,EACpB,qBAAqB,EACtB,MAAM,cAAc,CAAC;
|
|
1
|
+
{"version":3,"file":"Ollama_TextSummary.d.ts","sourceRoot":"","sources":["../../../src/provider-ollama/common/Ollama_TextSummary.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AAEH,OAAO,KAAK,EACV,eAAe,EACf,kBAAkB,EAClB,oBAAoB,EACpB,qBAAqB,EACtB,MAAM,cAAc,CAAC;AAEtB,OAAO,KAAK,EAAE,iBAAiB,EAAE,MAAM,sBAAsB,CAAC;AAG9D,KAAK,SAAS,GAAG,CAAC,KAAK,EAAE,iBAAiB,GAAG,SAAS,KAAK,OAAO,CAAC,GAAG,CAAC,CAAC;AAExE,wBAAgB,uBAAuB,CACrC,SAAS,EAAE,SAAS,GACnB,eAAe,CAAC,oBAAoB,EAAE,qBAAqB,EAAE,iBAAiB,CAAC,CAsBjF;AAED,wBAAgB,6BAA6B,CAC3C,SAAS,EAAE,SAAS,GACnB,kBAAkB,CAAC,oBAAoB,EAAE,qBAAqB,EAAE,iBAAiB,CAAC,CA4BpF"}
|
|
@@ -59,9 +59,9 @@ var OllamaModelConfigSchema = {
|
|
|
59
59
|
};
|
|
60
60
|
// src/provider-ollama/OllamaQueuedProvider.ts
|
|
61
61
|
import {
|
|
62
|
-
|
|
62
|
+
AiProvider
|
|
63
63
|
} from "@workglow/ai";
|
|
64
|
-
class OllamaQueuedProvider extends
|
|
64
|
+
class OllamaQueuedProvider extends AiProvider {
|
|
65
65
|
name = OLLAMA;
|
|
66
66
|
displayName = "Ollama";
|
|
67
67
|
isLocal = true;
|
|
@@ -72,7 +72,6 @@ class OllamaQueuedProvider extends QueuedAiProvider {
|
|
|
72
72
|
"TextEmbeddingTask",
|
|
73
73
|
"TextRewriterTask",
|
|
74
74
|
"TextSummaryTask",
|
|
75
|
-
"ToolCallingTask",
|
|
76
75
|
"ModelSearchTask"
|
|
77
76
|
];
|
|
78
77
|
constructor(tasks, streamTasks, reactiveTasks) {
|
|
@@ -93,4 +92,4 @@ export {
|
|
|
93
92
|
OLLAMA
|
|
94
93
|
};
|
|
95
94
|
|
|
96
|
-
//# debugId=
|
|
95
|
+
//# debugId=898D2F534251CE0C64756E2164756E21
|
|
@@ -4,10 +4,10 @@
|
|
|
4
4
|
"sourcesContent": [
|
|
5
5
|
"/**\n * @license\n * Copyright 2025 Steven Roussey <sroussey@gmail.com>\n * SPDX-License-Identifier: Apache-2.0\n */\n\nexport const OLLAMA = \"OLLAMA\";\nexport const OLLAMA_DEFAULT_BASE_URL = \"http://localhost:11434\";\n",
|
|
6
6
|
"/**\n * @license\n * Copyright 2025 Steven Roussey <sroussey@gmail.com>\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { ModelConfigSchema, ModelRecordSchema } from \"@workglow/ai/worker\";\nimport { DataPortSchemaObject, FromSchema } from \"@workglow/util/worker\";\nimport { OLLAMA, OLLAMA_DEFAULT_BASE_URL } from \"./Ollama_Constants\";\n\nexport const OllamaModelSchema = {\n type: \"object\",\n properties: {\n provider: {\n const: OLLAMA,\n description: \"Discriminator: Ollama local LLM server.\",\n },\n provider_config: {\n type: \"object\",\n description: \"Ollama-specific configuration.\",\n properties: {\n model_name: {\n type: \"string\",\n description: \"The Ollama model identifier (e.g., 'llama3.2', 'nomic-embed-text').\",\n },\n base_url: {\n type: \"string\",\n description: \"Base URL for the Ollama server.\",\n default: OLLAMA_DEFAULT_BASE_URL,\n },\n },\n required: [\"model_name\"],\n additionalProperties: false,\n },\n },\n required: [\"provider\", \"provider_config\"],\n additionalProperties: true,\n} as const satisfies DataPortSchemaObject;\n\nexport const OllamaModelRecordSchema = {\n type: \"object\",\n properties: {\n ...ModelRecordSchema.properties,\n ...OllamaModelSchema.properties,\n },\n required: [...ModelRecordSchema.required, ...OllamaModelSchema.required],\n additionalProperties: false,\n} as const satisfies DataPortSchemaObject;\n\nexport type OllamaModelRecord = FromSchema<typeof OllamaModelRecordSchema>;\n\nexport const OllamaModelConfigSchema = {\n type: \"object\",\n properties: {\n ...ModelConfigSchema.properties,\n ...OllamaModelSchema.properties,\n },\n required: [...ModelConfigSchema.required, ...OllamaModelSchema.required],\n additionalProperties: false,\n} as const satisfies DataPortSchemaObject;\n\nexport type OllamaModelConfig = FromSchema<typeof OllamaModelConfigSchema>;\n",
|
|
7
|
-
"/**\n * @license\n * Copyright 2025 Steven Roussey <sroussey@gmail.com>\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport {\n
|
|
7
|
+
"/**\n * @license\n * Copyright 2025 Steven Roussey <sroussey@gmail.com>\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport {\n AiProvider,\n type AiProviderReactiveRunFn,\n type AiProviderRunFn,\n type AiProviderStreamFn,\n} from \"@workglow/ai\";\nimport { OLLAMA } from \"./common/Ollama_Constants\";\nimport type { OllamaModelConfig } from \"./common/Ollama_ModelSchema\";\n\n/** Main-thread registration (inline or worker-backed). No queue — uses direct execution. */\nexport class OllamaQueuedProvider extends AiProvider<OllamaModelConfig> {\n readonly name = OLLAMA;\n readonly displayName = \"Ollama\";\n readonly isLocal = true;\n readonly supportsBrowser = true;\n\n readonly taskTypes = [\n \"ModelInfoTask\",\n \"TextGenerationTask\",\n \"TextEmbeddingTask\",\n \"TextRewriterTask\",\n \"TextSummaryTask\",\n \"ModelSearchTask\",\n ] as const;\n\n constructor(\n tasks?: Record<string, AiProviderRunFn<any, any, OllamaModelConfig>>,\n streamTasks?: Record<string, AiProviderStreamFn<any, any, OllamaModelConfig>>,\n reactiveTasks?: Record<string, AiProviderReactiveRunFn<any, any, OllamaModelConfig>>\n ) {\n super(tasks, streamTasks, reactiveTasks);\n }\n}\n",
|
|
8
8
|
"/**\n * @license\n * Copyright 2025 Steven Roussey <sroussey@gmail.com>\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport type { AiProviderRegisterOptions } from \"@workglow/ai\";\nimport { OllamaQueuedProvider } from \"./OllamaQueuedProvider\";\n\nexport async function registerOllama(\n options: AiProviderRegisterOptions & {\n worker: Worker | (() => Worker);\n }\n): Promise<void> {\n await new OllamaQueuedProvider().register(options);\n}\n"
|
|
9
9
|
],
|
|
10
|
-
"mappings": ";;;;;;;;;AAMO,IAAM,SAAS;AACf,IAAM,0BAA0B;;ACDvC;AAIO,IAAM,oBAAoB;AAAA,EAC/B,MAAM;AAAA,EACN,YAAY;AAAA,IACV,UAAU;AAAA,MACR,OAAO;AAAA,MACP,aAAa;AAAA,IACf;AAAA,IACA,iBAAiB;AAAA,MACf,MAAM;AAAA,MACN,aAAa;AAAA,MACb,YAAY;AAAA,QACV,YAAY;AAAA,UACV,MAAM;AAAA,UACN,aAAa;AAAA,QACf;AAAA,QACA,UAAU;AAAA,UACR,MAAM;AAAA,UACN,aAAa;AAAA,UACb,SAAS;AAAA,QACX;AAAA,MACF;AAAA,MACA,UAAU,CAAC,YAAY;AAAA,MACvB,sBAAsB;AAAA,IACxB;AAAA,EACF;AAAA,EACA,UAAU,CAAC,YAAY,iBAAiB;AAAA,EACxC,sBAAsB;AACxB;AAEO,IAAM,0BAA0B;AAAA,EACrC,MAAM;AAAA,EACN,YAAY;AAAA,OACP,kBAAkB;AAAA,OAClB,kBAAkB;AAAA,EACvB;AAAA,EACA,UAAU,CAAC,GAAG,kBAAkB,UAAU,GAAG,kBAAkB,QAAQ;AAAA,EACvE,sBAAsB;AACxB;AAIO,IAAM,0BAA0B;AAAA,EACrC,MAAM;AAAA,EACN,YAAY;AAAA,OACP,kBAAkB;AAAA,OAClB,kBAAkB;AAAA,EACvB;AAAA,EACA,UAAU,CAAC,GAAG,kBAAkB,UAAU,GAAG,kBAAkB,QAAQ;AAAA,EACvE,sBAAsB;AACxB;;ACrDA;AAAA;AAAA;AAUO,MAAM,6BAA6B,
|
|
11
|
-
"debugId": "
|
|
10
|
+
"mappings": ";;;;;;;;;AAMO,IAAM,SAAS;AACf,IAAM,0BAA0B;;ACDvC;AAIO,IAAM,oBAAoB;AAAA,EAC/B,MAAM;AAAA,EACN,YAAY;AAAA,IACV,UAAU;AAAA,MACR,OAAO;AAAA,MACP,aAAa;AAAA,IACf;AAAA,IACA,iBAAiB;AAAA,MACf,MAAM;AAAA,MACN,aAAa;AAAA,MACb,YAAY;AAAA,QACV,YAAY;AAAA,UACV,MAAM;AAAA,UACN,aAAa;AAAA,QACf;AAAA,QACA,UAAU;AAAA,UACR,MAAM;AAAA,UACN,aAAa;AAAA,UACb,SAAS;AAAA,QACX;AAAA,MACF;AAAA,MACA,UAAU,CAAC,YAAY;AAAA,MACvB,sBAAsB;AAAA,IACxB;AAAA,EACF;AAAA,EACA,UAAU,CAAC,YAAY,iBAAiB;AAAA,EACxC,sBAAsB;AACxB;AAEO,IAAM,0BAA0B;AAAA,EACrC,MAAM;AAAA,EACN,YAAY;AAAA,OACP,kBAAkB;AAAA,OAClB,kBAAkB;AAAA,EACvB;AAAA,EACA,UAAU,CAAC,GAAG,kBAAkB,UAAU,GAAG,kBAAkB,QAAQ;AAAA,EACvE,sBAAsB;AACxB;AAIO,IAAM,0BAA0B;AAAA,EACrC,MAAM;AAAA,EACN,YAAY;AAAA,OACP,kBAAkB;AAAA,OAClB,kBAAkB;AAAA,EACvB;AAAA,EACA,UAAU,CAAC,GAAG,kBAAkB,UAAU,GAAG,kBAAkB,QAAQ;AAAA,EACvE,sBAAsB;AACxB;;ACrDA;AAAA;AAAA;AAUO,MAAM,6BAA6B,WAA8B;AAAA,EAC7D,OAAO;AAAA,EACP,cAAc;AAAA,EACd,UAAU;AAAA,EACV,kBAAkB;AAAA,EAElB,YAAY;AAAA,IACnB;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,EACF;AAAA,EAEA,WAAW,CACT,OACA,aACA,eACA;AAAA,IACA,MAAM,OAAO,aAAa,aAAa;AAAA;AAE3C;;;AC7BA,eAAsB,cAAc,CAClC,SAGe;AAAA,EACf,MAAM,IAAI,qBAAqB,EAAE,SAAS,OAAO;AAAA;",
|
|
11
|
+
"debugId": "898D2F534251CE0C64756E2164756E21",
|
|
12
12
|
"names": []
|
|
13
13
|
}
|
|
@@ -74,9 +74,9 @@ var OllamaModelConfigSchema = {
|
|
|
74
74
|
};
|
|
75
75
|
// src/provider-ollama/OllamaQueuedProvider.ts
|
|
76
76
|
import {
|
|
77
|
-
|
|
77
|
+
AiProvider
|
|
78
78
|
} from "@workglow/ai";
|
|
79
|
-
class OllamaQueuedProvider extends
|
|
79
|
+
class OllamaQueuedProvider extends AiProvider {
|
|
80
80
|
name = OLLAMA;
|
|
81
81
|
displayName = "Ollama";
|
|
82
82
|
isLocal = true;
|
|
@@ -87,7 +87,6 @@ class OllamaQueuedProvider extends QueuedAiProvider {
|
|
|
87
87
|
"TextEmbeddingTask",
|
|
88
88
|
"TextRewriterTask",
|
|
89
89
|
"TextSummaryTask",
|
|
90
|
-
"ToolCallingTask",
|
|
91
90
|
"ModelSearchTask"
|
|
92
91
|
];
|
|
93
92
|
constructor(tasks, streamTasks, reactiveTasks) {
|
|
@@ -108,4 +107,4 @@ export {
|
|
|
108
107
|
OLLAMA
|
|
109
108
|
};
|
|
110
109
|
|
|
111
|
-
//# debugId=
|
|
110
|
+
//# debugId=4FCE45B627413BD864756E2164756E21
|
|
@@ -4,10 +4,10 @@
|
|
|
4
4
|
"sourcesContent": [
|
|
5
5
|
"/**\n * @license\n * Copyright 2025 Steven Roussey <sroussey@gmail.com>\n * SPDX-License-Identifier: Apache-2.0\n */\n\nexport const OLLAMA = \"OLLAMA\";\nexport const OLLAMA_DEFAULT_BASE_URL = \"http://localhost:11434\";\n",
|
|
6
6
|
"/**\n * @license\n * Copyright 2025 Steven Roussey <sroussey@gmail.com>\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { ModelConfigSchema, ModelRecordSchema } from \"@workglow/ai/worker\";\nimport { DataPortSchemaObject, FromSchema } from \"@workglow/util/worker\";\nimport { OLLAMA, OLLAMA_DEFAULT_BASE_URL } from \"./Ollama_Constants\";\n\nexport const OllamaModelSchema = {\n type: \"object\",\n properties: {\n provider: {\n const: OLLAMA,\n description: \"Discriminator: Ollama local LLM server.\",\n },\n provider_config: {\n type: \"object\",\n description: \"Ollama-specific configuration.\",\n properties: {\n model_name: {\n type: \"string\",\n description: \"The Ollama model identifier (e.g., 'llama3.2', 'nomic-embed-text').\",\n },\n base_url: {\n type: \"string\",\n description: \"Base URL for the Ollama server.\",\n default: OLLAMA_DEFAULT_BASE_URL,\n },\n },\n required: [\"model_name\"],\n additionalProperties: false,\n },\n },\n required: [\"provider\", \"provider_config\"],\n additionalProperties: true,\n} as const satisfies DataPortSchemaObject;\n\nexport const OllamaModelRecordSchema = {\n type: \"object\",\n properties: {\n ...ModelRecordSchema.properties,\n ...OllamaModelSchema.properties,\n },\n required: [...ModelRecordSchema.required, ...OllamaModelSchema.required],\n additionalProperties: false,\n} as const satisfies DataPortSchemaObject;\n\nexport type OllamaModelRecord = FromSchema<typeof OllamaModelRecordSchema>;\n\nexport const OllamaModelConfigSchema = {\n type: \"object\",\n properties: {\n ...ModelConfigSchema.properties,\n ...OllamaModelSchema.properties,\n },\n required: [...ModelConfigSchema.required, ...OllamaModelSchema.required],\n additionalProperties: false,\n} as const satisfies DataPortSchemaObject;\n\nexport type OllamaModelConfig = FromSchema<typeof OllamaModelConfigSchema>;\n",
|
|
7
|
-
"/**\n * @license\n * Copyright 2025 Steven Roussey <sroussey@gmail.com>\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport {\n
|
|
7
|
+
"/**\n * @license\n * Copyright 2025 Steven Roussey <sroussey@gmail.com>\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport {\n AiProvider,\n type AiProviderReactiveRunFn,\n type AiProviderRunFn,\n type AiProviderStreamFn,\n} from \"@workglow/ai\";\nimport { OLLAMA } from \"./common/Ollama_Constants\";\nimport type { OllamaModelConfig } from \"./common/Ollama_ModelSchema\";\n\n/** Main-thread registration (inline or worker-backed). No queue — uses direct execution. */\nexport class OllamaQueuedProvider extends AiProvider<OllamaModelConfig> {\n readonly name = OLLAMA;\n readonly displayName = \"Ollama\";\n readonly isLocal = true;\n readonly supportsBrowser = true;\n\n readonly taskTypes = [\n \"ModelInfoTask\",\n \"TextGenerationTask\",\n \"TextEmbeddingTask\",\n \"TextRewriterTask\",\n \"TextSummaryTask\",\n \"ModelSearchTask\",\n ] as const;\n\n constructor(\n tasks?: Record<string, AiProviderRunFn<any, any, OllamaModelConfig>>,\n streamTasks?: Record<string, AiProviderStreamFn<any, any, OllamaModelConfig>>,\n reactiveTasks?: Record<string, AiProviderReactiveRunFn<any, any, OllamaModelConfig>>\n ) {\n super(tasks, streamTasks, reactiveTasks);\n }\n}\n",
|
|
8
8
|
"/**\n * @license\n * Copyright 2025 Steven Roussey <sroussey@gmail.com>\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport type { AiProviderRegisterOptions } from \"@workglow/ai\";\nimport { OllamaQueuedProvider } from \"./OllamaQueuedProvider\";\n\nexport async function registerOllama(\n options: AiProviderRegisterOptions & {\n worker: Worker | (() => Worker);\n }\n): Promise<void> {\n await new OllamaQueuedProvider().register(options);\n}\n"
|
|
9
9
|
],
|
|
10
|
-
"mappings": ";;;;;;;;;;;;;;;;;;;;;;;;AAMO,IAAM,SAAS;AACf,IAAM,0BAA0B;;ACDvC;AAIO,IAAM,oBAAoB;AAAA,EAC/B,MAAM;AAAA,EACN,YAAY;AAAA,IACV,UAAU;AAAA,MACR,OAAO;AAAA,MACP,aAAa;AAAA,IACf;AAAA,IACA,iBAAiB;AAAA,MACf,MAAM;AAAA,MACN,aAAa;AAAA,MACb,YAAY;AAAA,QACV,YAAY;AAAA,UACV,MAAM;AAAA,UACN,aAAa;AAAA,QACf;AAAA,QACA,UAAU;AAAA,UACR,MAAM;AAAA,UACN,aAAa;AAAA,UACb,SAAS;AAAA,QACX;AAAA,MACF;AAAA,MACA,UAAU,CAAC,YAAY;AAAA,MACvB,sBAAsB;AAAA,IACxB;AAAA,EACF;AAAA,EACA,UAAU,CAAC,YAAY,iBAAiB;AAAA,EACxC,sBAAsB;AACxB;AAEO,IAAM,0BAA0B;AAAA,EACrC,MAAM;AAAA,EACN,YAAY;AAAA,OACP,kBAAkB;AAAA,OAClB,kBAAkB;AAAA,EACvB;AAAA,EACA,UAAU,CAAC,GAAG,kBAAkB,UAAU,GAAG,kBAAkB,QAAQ;AAAA,EACvE,sBAAsB;AACxB;AAIO,IAAM,0BAA0B;AAAA,EACrC,MAAM;AAAA,EACN,YAAY;AAAA,OACP,kBAAkB;AAAA,OAClB,kBAAkB;AAAA,EACvB;AAAA,EACA,UAAU,CAAC,GAAG,kBAAkB,UAAU,GAAG,kBAAkB,QAAQ;AAAA,EACvE,sBAAsB;AACxB;;ACrDA;AAAA;AAAA;AAUO,MAAM,6BAA6B,
|
|
11
|
-
"debugId": "
|
|
10
|
+
"mappings": ";;;;;;;;;;;;;;;;;;;;;;;;AAMO,IAAM,SAAS;AACf,IAAM,0BAA0B;;ACDvC;AAIO,IAAM,oBAAoB;AAAA,EAC/B,MAAM;AAAA,EACN,YAAY;AAAA,IACV,UAAU;AAAA,MACR,OAAO;AAAA,MACP,aAAa;AAAA,IACf;AAAA,IACA,iBAAiB;AAAA,MACf,MAAM;AAAA,MACN,aAAa;AAAA,MACb,YAAY;AAAA,QACV,YAAY;AAAA,UACV,MAAM;AAAA,UACN,aAAa;AAAA,QACf;AAAA,QACA,UAAU;AAAA,UACR,MAAM;AAAA,UACN,aAAa;AAAA,UACb,SAAS;AAAA,QACX;AAAA,MACF;AAAA,MACA,UAAU,CAAC,YAAY;AAAA,MACvB,sBAAsB;AAAA,IACxB;AAAA,EACF;AAAA,EACA,UAAU,CAAC,YAAY,iBAAiB;AAAA,EACxC,sBAAsB;AACxB;AAEO,IAAM,0BAA0B;AAAA,EACrC,MAAM;AAAA,EACN,YAAY;AAAA,OACP,kBAAkB;AAAA,OAClB,kBAAkB;AAAA,EACvB;AAAA,EACA,UAAU,CAAC,GAAG,kBAAkB,UAAU,GAAG,kBAAkB,QAAQ;AAAA,EACvE,sBAAsB;AACxB;AAIO,IAAM,0BAA0B;AAAA,EACrC,MAAM;AAAA,EACN,YAAY;AAAA,OACP,kBAAkB;AAAA,OAClB,kBAAkB;AAAA,EACvB;AAAA,EACA,UAAU,CAAC,GAAG,kBAAkB,UAAU,GAAG,kBAAkB,QAAQ;AAAA,EACvE,sBAAsB;AACxB;;ACrDA;AAAA;AAAA;AAUO,MAAM,6BAA6B,WAA8B;AAAA,EAC7D,OAAO;AAAA,EACP,cAAc;AAAA,EACd,UAAU;AAAA,EACV,kBAAkB;AAAA,EAElB,YAAY;AAAA,IACnB;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,EACF;AAAA,EAEA,WAAW,CACT,OACA,aACA,eACA;AAAA,IACA,MAAM,OAAO,aAAa,aAAa;AAAA;AAE3C;;;AC7BA,eAAsB,cAAc,CAClC,SAGe;AAAA,EACf,MAAM,IAAI,qBAAqB,EAAE,SAAS,OAAO;AAAA;",
|
|
11
|
+
"debugId": "4FCE45B627413BD864756E2164756E21",
|
|
12
12
|
"names": []
|
|
13
13
|
}
|
|
@@ -71,9 +71,27 @@ function createOllamaModelInfo(getClient2) {
|
|
|
71
71
|
};
|
|
72
72
|
}
|
|
73
73
|
|
|
74
|
+
// src/common/modelSearchQuery.ts
|
|
75
|
+
function normalizedModelSearchQuery(query) {
|
|
76
|
+
const t = query?.trim();
|
|
77
|
+
return t ? t.toLowerCase() : undefined;
|
|
78
|
+
}
|
|
79
|
+
function filterLabeledModelsByQuery(models, query) {
|
|
80
|
+
const q = normalizedModelSearchQuery(query);
|
|
81
|
+
if (!q)
|
|
82
|
+
return [...models];
|
|
83
|
+
return models.filter((m) => m.value.toLowerCase().includes(q) || m.label.toLowerCase().includes(q));
|
|
84
|
+
}
|
|
85
|
+
function filterModelSearchResultsByQuery(results, query) {
|
|
86
|
+
const q = normalizedModelSearchQuery(query);
|
|
87
|
+
if (!q)
|
|
88
|
+
return results;
|
|
89
|
+
return results.filter((m) => m.id.toLowerCase().includes(q) || m.label.toLowerCase().includes(q) || m.description.toLowerCase().includes(q));
|
|
90
|
+
}
|
|
91
|
+
|
|
74
92
|
// src/provider-ollama/common/Ollama_ModelSearch.ts
|
|
75
93
|
function createOllamaModelSearch(getClient2) {
|
|
76
|
-
return async () => {
|
|
94
|
+
return async (input) => {
|
|
77
95
|
try {
|
|
78
96
|
const client = await getClient2(undefined);
|
|
79
97
|
const response = await client.list();
|
|
@@ -92,7 +110,7 @@ function createOllamaModelSearch(getClient2) {
|
|
|
92
110
|
},
|
|
93
111
|
raw: m
|
|
94
112
|
}));
|
|
95
|
-
return { results };
|
|
113
|
+
return { results: filterModelSearchResultsByQuery(results, input.query) };
|
|
96
114
|
} catch {
|
|
97
115
|
return { results: [] };
|
|
98
116
|
}
|
|
@@ -121,19 +139,8 @@ function createOllamaTextEmbedding(getClient2) {
|
|
|
121
139
|
}
|
|
122
140
|
|
|
123
141
|
// src/provider-ollama/common/Ollama_TextGeneration.ts
|
|
124
|
-
import { getLogger } from "@workglow/util/worker";
|
|
125
142
|
function createOllamaTextGeneration(getClient2) {
|
|
126
143
|
const run = async (input, model, update_progress, _signal) => {
|
|
127
|
-
if (Array.isArray(input.prompt)) {
|
|
128
|
-
getLogger().warn("Ollama_TextGeneration: array input received; processing sequentially (no native batch support)");
|
|
129
|
-
const prompts = input.prompt;
|
|
130
|
-
const results = [];
|
|
131
|
-
for (const item of prompts) {
|
|
132
|
-
const r = await run({ ...input, prompt: item }, model, update_progress, _signal);
|
|
133
|
-
results.push(r.text);
|
|
134
|
-
}
|
|
135
|
-
return { text: results };
|
|
136
|
-
}
|
|
137
144
|
update_progress(0, "Starting Ollama text generation");
|
|
138
145
|
const client = await getClient2(model);
|
|
139
146
|
const modelName = getOllamaModelName(model);
|
|
@@ -186,19 +193,8 @@ function createOllamaTextGenerationStream(getClient2) {
|
|
|
186
193
|
}
|
|
187
194
|
|
|
188
195
|
// src/provider-ollama/common/Ollama_TextRewriter.ts
|
|
189
|
-
import { getLogger as getLogger2 } from "@workglow/util/worker";
|
|
190
196
|
function createOllamaTextRewriter(getClient2) {
|
|
191
197
|
const run = async (input, model, update_progress, _signal) => {
|
|
192
|
-
if (Array.isArray(input.text)) {
|
|
193
|
-
getLogger2().warn("Ollama_TextRewriter: array input received; processing sequentially (no native batch support)");
|
|
194
|
-
const texts = input.text;
|
|
195
|
-
const results = [];
|
|
196
|
-
for (const item of texts) {
|
|
197
|
-
const r = await run({ ...input, text: item }, model, update_progress, _signal);
|
|
198
|
-
results.push(r.text);
|
|
199
|
-
}
|
|
200
|
-
return { text: results };
|
|
201
|
-
}
|
|
202
198
|
update_progress(0, "Starting Ollama text rewriting");
|
|
203
199
|
const client = await getClient2(model);
|
|
204
200
|
const modelName = getOllamaModelName(model);
|
|
@@ -243,19 +239,8 @@ function createOllamaTextRewriterStream(getClient2) {
|
|
|
243
239
|
}
|
|
244
240
|
|
|
245
241
|
// src/provider-ollama/common/Ollama_TextSummary.ts
|
|
246
|
-
import { getLogger as getLogger3 } from "@workglow/util/worker";
|
|
247
242
|
function createOllamaTextSummary(getClient2) {
|
|
248
243
|
const run = async (input, model, update_progress, _signal) => {
|
|
249
|
-
if (Array.isArray(input.text)) {
|
|
250
|
-
getLogger3().warn("Ollama_TextSummary: array input received; processing sequentially (no native batch support)");
|
|
251
|
-
const texts = input.text;
|
|
252
|
-
const results = [];
|
|
253
|
-
for (const item of texts) {
|
|
254
|
-
const r = await run({ ...input, text: item }, model, update_progress, _signal);
|
|
255
|
-
results.push(r.text);
|
|
256
|
-
}
|
|
257
|
-
return { text: results };
|
|
258
|
-
}
|
|
259
244
|
update_progress(0, "Starting Ollama text summarization");
|
|
260
245
|
const client = await getClient2(model);
|
|
261
246
|
const modelName = getOllamaModelName(model);
|
|
@@ -299,139 +284,7 @@ function createOllamaTextSummaryStream(getClient2) {
|
|
|
299
284
|
};
|
|
300
285
|
}
|
|
301
286
|
|
|
302
|
-
// src/provider-ollama/common/Ollama_ToolCalling.ts
|
|
303
|
-
import { buildToolDescription, filterValidToolCalls } from "@workglow/ai/worker";
|
|
304
|
-
import { getLogger as getLogger4, parsePartialJson } from "@workglow/util/worker";
|
|
305
|
-
function mapOllamaTools(tools) {
|
|
306
|
-
return tools.map((t) => ({
|
|
307
|
-
type: "function",
|
|
308
|
-
function: {
|
|
309
|
-
name: t.name,
|
|
310
|
-
description: buildToolDescription(t),
|
|
311
|
-
parameters: t.inputSchema
|
|
312
|
-
}
|
|
313
|
-
}));
|
|
314
|
-
}
|
|
315
|
-
function createOllamaToolCalling(getClient2, buildMessages) {
|
|
316
|
-
const run = async (input, model, update_progress, _signal) => {
|
|
317
|
-
if (Array.isArray(input.prompt)) {
|
|
318
|
-
getLogger4().warn("Ollama_ToolCalling: array input received; processing sequentially (no native batch support)");
|
|
319
|
-
const prompts = input.prompt;
|
|
320
|
-
const texts = [];
|
|
321
|
-
const toolCallsList = [];
|
|
322
|
-
for (const item of prompts) {
|
|
323
|
-
const r = await run({ ...input, prompt: item }, model, update_progress, _signal);
|
|
324
|
-
texts.push(r.text);
|
|
325
|
-
toolCallsList.push(r.toolCalls);
|
|
326
|
-
}
|
|
327
|
-
return { text: texts, toolCalls: toolCallsList };
|
|
328
|
-
}
|
|
329
|
-
update_progress(0, "Starting Ollama tool calling");
|
|
330
|
-
const client = await getClient2(model);
|
|
331
|
-
const modelName = getOllamaModelName(model);
|
|
332
|
-
const messages = buildMessages(input);
|
|
333
|
-
const tools = input.toolChoice === "none" ? undefined : mapOllamaTools(input.tools);
|
|
334
|
-
const response = await client.chat({
|
|
335
|
-
model: modelName,
|
|
336
|
-
messages,
|
|
337
|
-
tools,
|
|
338
|
-
options: {
|
|
339
|
-
temperature: input.temperature,
|
|
340
|
-
num_predict: input.maxTokens
|
|
341
|
-
}
|
|
342
|
-
});
|
|
343
|
-
const text = response.message.content ?? "";
|
|
344
|
-
const toolCalls = [];
|
|
345
|
-
(response.message.tool_calls ?? []).forEach((tc, index) => {
|
|
346
|
-
let parsedInput = {};
|
|
347
|
-
const fnArgs = tc.function.arguments;
|
|
348
|
-
if (typeof fnArgs === "string") {
|
|
349
|
-
try {
|
|
350
|
-
parsedInput = JSON.parse(fnArgs);
|
|
351
|
-
} catch {
|
|
352
|
-
const partial = parsePartialJson(fnArgs);
|
|
353
|
-
parsedInput = partial ?? {};
|
|
354
|
-
}
|
|
355
|
-
} else if (fnArgs != null) {
|
|
356
|
-
parsedInput = fnArgs;
|
|
357
|
-
}
|
|
358
|
-
const id = `call_${index}`;
|
|
359
|
-
toolCalls.push({ id, name: tc.function.name, input: parsedInput });
|
|
360
|
-
});
|
|
361
|
-
update_progress(100, "Completed Ollama tool calling");
|
|
362
|
-
return { text, toolCalls: filterValidToolCalls(toolCalls, input.tools) };
|
|
363
|
-
};
|
|
364
|
-
return run;
|
|
365
|
-
}
|
|
366
|
-
function createOllamaToolCallingStream(getClient2, buildMessages) {
|
|
367
|
-
return async function* (input, model, signal) {
|
|
368
|
-
const client = await getClient2(model);
|
|
369
|
-
const modelName = getOllamaModelName(model);
|
|
370
|
-
const messages = buildMessages(input);
|
|
371
|
-
const tools = input.toolChoice === "none" ? undefined : mapOllamaTools(input.tools);
|
|
372
|
-
const stream = await client.chat({
|
|
373
|
-
model: modelName,
|
|
374
|
-
messages,
|
|
375
|
-
tools,
|
|
376
|
-
options: {
|
|
377
|
-
temperature: input.temperature,
|
|
378
|
-
num_predict: input.maxTokens
|
|
379
|
-
},
|
|
380
|
-
stream: true
|
|
381
|
-
});
|
|
382
|
-
const onAbort = () => stream.abort();
|
|
383
|
-
signal.addEventListener("abort", onAbort, { once: true });
|
|
384
|
-
let accumulatedText = "";
|
|
385
|
-
const toolCalls = [];
|
|
386
|
-
let callIndex = 0;
|
|
387
|
-
try {
|
|
388
|
-
for await (const chunk of stream) {
|
|
389
|
-
const delta = chunk.message.content;
|
|
390
|
-
if (delta) {
|
|
391
|
-
accumulatedText += delta;
|
|
392
|
-
yield { type: "text-delta", port: "text", textDelta: delta };
|
|
393
|
-
}
|
|
394
|
-
const chunkToolCalls = chunk.message.tool_calls;
|
|
395
|
-
if (Array.isArray(chunkToolCalls) && chunkToolCalls.length > 0) {
|
|
396
|
-
for (const tc of chunkToolCalls) {
|
|
397
|
-
let parsedInput = {};
|
|
398
|
-
const fnArgs = tc.function.arguments;
|
|
399
|
-
if (typeof fnArgs === "string") {
|
|
400
|
-
try {
|
|
401
|
-
parsedInput = JSON.parse(fnArgs);
|
|
402
|
-
} catch {
|
|
403
|
-
const partial = parsePartialJson(fnArgs);
|
|
404
|
-
parsedInput = partial ?? {};
|
|
405
|
-
}
|
|
406
|
-
} else if (fnArgs != null) {
|
|
407
|
-
parsedInput = fnArgs;
|
|
408
|
-
}
|
|
409
|
-
const id = `call_${callIndex++}`;
|
|
410
|
-
toolCalls.push({ id, name: tc.function.name, input: parsedInput });
|
|
411
|
-
}
|
|
412
|
-
yield { type: "object-delta", port: "toolCalls", objectDelta: [...toolCalls] };
|
|
413
|
-
}
|
|
414
|
-
}
|
|
415
|
-
const validToolCalls = filterValidToolCalls(toolCalls, input.tools);
|
|
416
|
-
yield {
|
|
417
|
-
type: "finish",
|
|
418
|
-
data: { text: accumulatedText, toolCalls: validToolCalls }
|
|
419
|
-
};
|
|
420
|
-
} finally {
|
|
421
|
-
signal.removeEventListener("abort", onAbort);
|
|
422
|
-
}
|
|
423
|
-
};
|
|
424
|
-
}
|
|
425
|
-
|
|
426
287
|
// src/provider-ollama/common/Ollama_JobRunFns.browser.ts
|
|
427
|
-
function buildBrowserToolCallingMessages(input) {
|
|
428
|
-
const messages = [];
|
|
429
|
-
if (input.systemPrompt) {
|
|
430
|
-
messages.push({ role: "system", content: input.systemPrompt });
|
|
431
|
-
}
|
|
432
|
-
messages.push({ role: "user", content: input.prompt });
|
|
433
|
-
return messages;
|
|
434
|
-
}
|
|
435
288
|
var Ollama_TextGeneration = createOllamaTextGeneration(getClient);
|
|
436
289
|
var Ollama_TextEmbedding = createOllamaTextEmbedding(getClient);
|
|
437
290
|
var Ollama_TextRewriter = createOllamaTextRewriter(getClient);
|
|
@@ -439,8 +292,6 @@ var Ollama_TextSummary = createOllamaTextSummary(getClient);
|
|
|
439
292
|
var Ollama_TextGeneration_Stream = createOllamaTextGenerationStream(getClient);
|
|
440
293
|
var Ollama_TextRewriter_Stream = createOllamaTextRewriterStream(getClient);
|
|
441
294
|
var Ollama_TextSummary_Stream = createOllamaTextSummaryStream(getClient);
|
|
442
|
-
var Ollama_ToolCalling = createOllamaToolCalling(getClient, buildBrowserToolCallingMessages);
|
|
443
|
-
var Ollama_ToolCalling_Stream = createOllamaToolCallingStream(getClient, buildBrowserToolCallingMessages);
|
|
444
295
|
var Ollama_ModelInfo = createOllamaModelInfo(getClient);
|
|
445
296
|
var Ollama_ModelSearch = createOllamaModelSearch(getClient);
|
|
446
297
|
var OLLAMA_TASKS = {
|
|
@@ -449,21 +300,19 @@ var OLLAMA_TASKS = {
|
|
|
449
300
|
TextEmbeddingTask: Ollama_TextEmbedding,
|
|
450
301
|
TextRewriterTask: Ollama_TextRewriter,
|
|
451
302
|
TextSummaryTask: Ollama_TextSummary,
|
|
452
|
-
ToolCallingTask: Ollama_ToolCalling,
|
|
453
303
|
ModelSearchTask: Ollama_ModelSearch
|
|
454
304
|
};
|
|
455
305
|
var OLLAMA_STREAM_TASKS = {
|
|
456
306
|
TextGenerationTask: Ollama_TextGeneration_Stream,
|
|
457
307
|
TextRewriterTask: Ollama_TextRewriter_Stream,
|
|
458
|
-
TextSummaryTask: Ollama_TextSummary_Stream
|
|
459
|
-
ToolCallingTask: Ollama_ToolCalling_Stream
|
|
308
|
+
TextSummaryTask: Ollama_TextSummary_Stream
|
|
460
309
|
};
|
|
461
310
|
|
|
462
311
|
// src/provider-ollama/OllamaQueuedProvider.ts
|
|
463
312
|
import {
|
|
464
|
-
|
|
313
|
+
AiProvider
|
|
465
314
|
} from "@workglow/ai";
|
|
466
|
-
class OllamaQueuedProvider extends
|
|
315
|
+
class OllamaQueuedProvider extends AiProvider {
|
|
467
316
|
name = OLLAMA;
|
|
468
317
|
displayName = "Ollama";
|
|
469
318
|
isLocal = true;
|
|
@@ -474,7 +323,6 @@ class OllamaQueuedProvider extends QueuedAiProvider {
|
|
|
474
323
|
"TextEmbeddingTask",
|
|
475
324
|
"TextRewriterTask",
|
|
476
325
|
"TextSummaryTask",
|
|
477
|
-
"ToolCallingTask",
|
|
478
326
|
"ModelSearchTask"
|
|
479
327
|
];
|
|
480
328
|
constructor(tasks, streamTasks, reactiveTasks) {
|
|
@@ -487,13 +335,13 @@ async function registerOllamaInline(options) {
|
|
|
487
335
|
await new OllamaQueuedProvider(OLLAMA_TASKS, OLLAMA_STREAM_TASKS).register(options ?? {});
|
|
488
336
|
}
|
|
489
337
|
// src/provider-ollama/registerOllamaWorker.browser.ts
|
|
490
|
-
import { getLogger
|
|
338
|
+
import { getLogger, globalServiceRegistry, WORKER_SERVER } from "@workglow/util/worker";
|
|
491
339
|
|
|
492
340
|
// src/provider-ollama/OllamaProvider.ts
|
|
493
341
|
import {
|
|
494
|
-
AiProvider
|
|
342
|
+
AiProvider as AiProvider2
|
|
495
343
|
} from "@workglow/ai/worker";
|
|
496
|
-
class OllamaProvider extends
|
|
344
|
+
class OllamaProvider extends AiProvider2 {
|
|
497
345
|
name = OLLAMA;
|
|
498
346
|
displayName = "Ollama";
|
|
499
347
|
isLocal = true;
|
|
@@ -504,7 +352,6 @@ class OllamaProvider extends AiProvider {
|
|
|
504
352
|
"TextEmbeddingTask",
|
|
505
353
|
"TextRewriterTask",
|
|
506
354
|
"TextSummaryTask",
|
|
507
|
-
"ToolCallingTask",
|
|
508
355
|
"ModelSearchTask"
|
|
509
356
|
];
|
|
510
357
|
constructor(tasks, streamTasks, reactiveTasks) {
|
|
@@ -517,7 +364,7 @@ async function registerOllamaWorker() {
|
|
|
517
364
|
const workerServer = globalServiceRegistry.get(WORKER_SERVER);
|
|
518
365
|
new OllamaProvider(OLLAMA_TASKS, OLLAMA_STREAM_TASKS).registerOnWorkerServer(workerServer);
|
|
519
366
|
workerServer.sendReady();
|
|
520
|
-
|
|
367
|
+
getLogger().info("Ollama worker job run functions registered");
|
|
521
368
|
}
|
|
522
369
|
export {
|
|
523
370
|
registerOllamaWorker,
|
|
@@ -527,4 +374,4 @@ export {
|
|
|
527
374
|
getClient
|
|
528
375
|
};
|
|
529
376
|
|
|
530
|
-
//# debugId=
|
|
377
|
+
//# debugId=6912A4E1147FEC9A64756E2164756E21
|