@workglow/huggingface-transformers 0.2.28
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/ai-provider/HuggingFaceTransformersProvider.d.ts +28 -0
- package/dist/ai-provider/HuggingFaceTransformersProvider.d.ts.map +1 -0
- package/dist/ai-provider/HuggingFaceTransformersQueuedProvider.d.ts +29 -0
- package/dist/ai-provider/HuggingFaceTransformersQueuedProvider.d.ts.map +1 -0
- package/dist/ai-provider/common/HFT_BackgroundRemoval.d.ts +12 -0
- package/dist/ai-provider/common/HFT_BackgroundRemoval.d.ts.map +1 -0
- package/dist/ai-provider/common/HFT_Chat.d.ts +10 -0
- package/dist/ai-provider/common/HFT_Chat.d.ts.map +1 -0
- package/dist/ai-provider/common/HFT_Constants.d.ts +95 -0
- package/dist/ai-provider/common/HFT_Constants.d.ts.map +1 -0
- package/dist/ai-provider/common/HFT_CountTokens.d.ts +10 -0
- package/dist/ai-provider/common/HFT_CountTokens.d.ts.map +1 -0
- package/dist/ai-provider/common/HFT_Download.d.ts +13 -0
- package/dist/ai-provider/common/HFT_Download.d.ts.map +1 -0
- package/dist/ai-provider/common/HFT_ImageClassification.d.ts +13 -0
- package/dist/ai-provider/common/HFT_ImageClassification.d.ts.map +1 -0
- package/dist/ai-provider/common/HFT_ImageEmbedding.d.ts +12 -0
- package/dist/ai-provider/common/HFT_ImageEmbedding.d.ts.map +1 -0
- package/dist/ai-provider/common/HFT_ImageSegmentation.d.ts +12 -0
- package/dist/ai-provider/common/HFT_ImageSegmentation.d.ts.map +1 -0
- package/dist/ai-provider/common/HFT_ImageToText.d.ts +12 -0
- package/dist/ai-provider/common/HFT_ImageToText.d.ts.map +1 -0
- package/dist/ai-provider/common/HFT_InlineLifecycle.d.ts +7 -0
- package/dist/ai-provider/common/HFT_InlineLifecycle.d.ts.map +1 -0
- package/dist/ai-provider/common/HFT_JobRunFns.d.ts +19 -0
- package/dist/ai-provider/common/HFT_JobRunFns.d.ts.map +1 -0
- package/dist/ai-provider/common/HFT_ModelInfo.d.ts +9 -0
- package/dist/ai-provider/common/HFT_ModelInfo.d.ts.map +1 -0
- package/dist/ai-provider/common/HFT_ModelSchema.d.ts +368 -0
- package/dist/ai-provider/common/HFT_ModelSchema.d.ts.map +1 -0
- package/dist/ai-provider/common/HFT_ModelSearch.d.ts +8 -0
- package/dist/ai-provider/common/HFT_ModelSearch.d.ts.map +1 -0
- package/dist/ai-provider/common/HFT_ObjectDetection.d.ts +13 -0
- package/dist/ai-provider/common/HFT_ObjectDetection.d.ts.map +1 -0
- package/dist/ai-provider/common/HFT_OnnxDtypes.d.ts +27 -0
- package/dist/ai-provider/common/HFT_OnnxDtypes.d.ts.map +1 -0
- package/dist/ai-provider/common/HFT_Pipeline.d.ts +62 -0
- package/dist/ai-provider/common/HFT_Pipeline.d.ts.map +1 -0
- package/dist/ai-provider/common/HFT_Streaming.d.ts +25 -0
- package/dist/ai-provider/common/HFT_Streaming.d.ts.map +1 -0
- package/dist/ai-provider/common/HFT_StructuredGeneration.d.ts +10 -0
- package/dist/ai-provider/common/HFT_StructuredGeneration.d.ts.map +1 -0
- package/dist/ai-provider/common/HFT_TextClassification.d.ts +9 -0
- package/dist/ai-provider/common/HFT_TextClassification.d.ts.map +1 -0
- package/dist/ai-provider/common/HFT_TextEmbedding.d.ts +13 -0
- package/dist/ai-provider/common/HFT_TextEmbedding.d.ts.map +1 -0
- package/dist/ai-provider/common/HFT_TextFillMask.d.ts +9 -0
- package/dist/ai-provider/common/HFT_TextFillMask.d.ts.map +1 -0
- package/dist/ai-provider/common/HFT_TextGeneration.d.ts +14 -0
- package/dist/ai-provider/common/HFT_TextGeneration.d.ts.map +1 -0
- package/dist/ai-provider/common/HFT_TextLanguageDetection.d.ts +9 -0
- package/dist/ai-provider/common/HFT_TextLanguageDetection.d.ts.map +1 -0
- package/dist/ai-provider/common/HFT_TextNamedEntityRecognition.d.ts +9 -0
- package/dist/ai-provider/common/HFT_TextNamedEntityRecognition.d.ts.map +1 -0
- package/dist/ai-provider/common/HFT_TextOutput.d.ts +8 -0
- package/dist/ai-provider/common/HFT_TextOutput.d.ts.map +1 -0
- package/dist/ai-provider/common/HFT_TextQuestionAnswer.d.ts +14 -0
- package/dist/ai-provider/common/HFT_TextQuestionAnswer.d.ts.map +1 -0
- package/dist/ai-provider/common/HFT_TextRewriter.d.ts +14 -0
- package/dist/ai-provider/common/HFT_TextRewriter.d.ts.map +1 -0
- package/dist/ai-provider/common/HFT_TextSummary.d.ts +14 -0
- package/dist/ai-provider/common/HFT_TextSummary.d.ts.map +1 -0
- package/dist/ai-provider/common/HFT_TextTranslation.d.ts +14 -0
- package/dist/ai-provider/common/HFT_TextTranslation.d.ts.map +1 -0
- package/dist/ai-provider/common/HFT_ToolCalling.d.ts +19 -0
- package/dist/ai-provider/common/HFT_ToolCalling.d.ts.map +1 -0
- package/dist/ai-provider/common/HFT_ToolMarkup.d.ts +20 -0
- package/dist/ai-provider/common/HFT_ToolMarkup.d.ts.map +1 -0
- package/dist/ai-provider/common/HFT_Unload.d.ts +13 -0
- package/dist/ai-provider/common/HFT_Unload.d.ts.map +1 -0
- package/dist/ai-provider/index.d.ts +13 -0
- package/dist/ai-provider/index.d.ts.map +1 -0
- package/dist/ai-provider/registerHuggingFaceTransformers.d.ts +14 -0
- package/dist/ai-provider/registerHuggingFaceTransformers.d.ts.map +1 -0
- package/dist/ai-provider/registerHuggingFaceTransformersInline.d.ts +15 -0
- package/dist/ai-provider/registerHuggingFaceTransformersInline.d.ts.map +1 -0
- package/dist/ai-provider/registerHuggingFaceTransformersWorker.d.ts +7 -0
- package/dist/ai-provider/registerHuggingFaceTransformersWorker.d.ts.map +1 -0
- package/dist/ai-provider/runtime.d.ts +21 -0
- package/dist/ai-provider/runtime.d.ts.map +1 -0
- package/dist/ai-provider-runtime.d.ts +7 -0
- package/dist/ai-provider-runtime.d.ts.map +1 -0
- package/dist/ai-provider-runtime.js +2367 -0
- package/dist/ai-provider-runtime.js.map +46 -0
- package/dist/ai-provider.d.ts +7 -0
- package/dist/ai-provider.d.ts.map +1 -0
- package/dist/ai-provider.js +879 -0
- package/dist/ai-provider.js.map +17 -0
- package/package.json +60 -0
|
@@ -0,0 +1,17 @@
|
|
|
1
|
+
{
|
|
2
|
+
"version": 3,
|
|
3
|
+
"sources": ["../src/ai-provider/common/HFT_Pipeline.ts", "../src/ai-provider/common/HFT_Constants.ts", "../src/ai-provider/common/HFT_ModelSchema.ts", "../src/ai-provider/HuggingFaceTransformersProvider.ts", "../src/ai-provider/HuggingFaceTransformersQueuedProvider.ts", "../src/ai-provider/registerHuggingFaceTransformers.ts", "../src/ai-provider/common/HFT_OnnxDtypes.ts", "../src/ai-provider/common/HFT_ToolMarkup.ts"],
|
|
4
|
+
"sourcesContent": [
|
|
5
|
+
"/**\n * @license\n * Copyright 2025 Steven Roussey <sroussey@gmail.com>\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport type { DynamicCache, PretrainedModelOptions, ProgressInfo } from \"@huggingface/transformers\";\nimport { getLogger } from \"@workglow/util/worker\";\nimport type { HfTransformersOnnxModelConfig } from \"./HFT_ModelSchema\";\n\nlet _transformersSdk: typeof import(\"@huggingface/transformers\") | undefined;\nlet _cacheDir: string | undefined;\n\n/**\n * Set the filesystem cache directory for downloaded transformers.js models.\n * Must be called before any model is loaded.\n */\nexport function setHftCacheDir(dir: string): void {\n _cacheDir = dir;\n if (_transformersSdk) {\n _transformersSdk.env.cacheDir = dir;\n }\n}\n\nexport async function loadTransformersSDK() {\n if (!_transformersSdk) {\n try {\n _transformersSdk = await import(\"@huggingface/transformers\");\n _transformersSdk.env.fetch = abortableFetch as typeof fetch;\n if (_cacheDir) {\n _transformersSdk.env.cacheDir = _cacheDir;\n }\n } catch {\n throw new Error(\n \"@huggingface/transformers is required for HuggingFace Transformers tasks. Install it with: bun add @huggingface/transformers\"\n );\n }\n }\n return _transformersSdk;\n}\n\n/** Per-model AbortControllers used by abortableFetch; keyed by model_path. */\nconst modelAbortControllers = new Map<string, AbortController>();\n\nfunction combineAbortSignals(\n existingSignal: AbortSignal | null | undefined,\n modelSignal: AbortSignal | undefined\n): AbortSignal | undefined {\n if (!existingSignal) {\n return modelSignal;\n }\n if (!modelSignal) {\n return existingSignal;\n }\n if (existingSignal.aborted || modelSignal.aborted) {\n return AbortSignal.abort(existingSignal.reason ?? modelSignal.reason);\n }\n if (typeof AbortSignal.any === \"function\") {\n return AbortSignal.any([existingSignal, modelSignal]);\n }\n\n const controller = new AbortController();\n const abort = (event: Event) => {\n const signal = event.target as AbortSignal;\n controller.abort(signal.reason);\n };\n existingSignal.addEventListener(\"abort\", abort, { once: true });\n modelSignal.addEventListener(\"abort\", abort, { once: true });\n return controller.signal;\n}\n\nfunction createAbortError(signal: AbortSignal): Error {\n const reason = signal.reason;\n if (reason instanceof Error) {\n return reason;\n }\n return new Error(String(reason ?? \"Fetch aborted\"));\n}\n\nfunction wrapAbortableResponse(response: Response, signal: AbortSignal | undefined): Response {\n if (!signal || !response.body) {\n return response;\n }\n\n const contentLengthHeader = response.headers.get(\"content-length\");\n const expectedSize =\n contentLengthHeader && /^\\d+$/.test(contentLengthHeader)\n ? Number.parseInt(contentLengthHeader, 10)\n : undefined;\n const sourceBody = response.body;\n\n // Use pull-based reading to maintain backpressure. The previous start()-based\n // loop eagerly drained the source into the internal queue without waiting for\n // the consumer, which could buffer the entire response body in memory — a\n // problem for large model files (hundreds of MB to several GB).\n let reader: ReadableStreamDefaultReader<Uint8Array>;\n let abortHandler: (() => void) | undefined;\n let loaded = 0;\n\n const cleanup = () => {\n if (abortHandler) {\n signal.removeEventListener(\"abort\", abortHandler);\n abortHandler = undefined;\n }\n reader?.releaseLock();\n };\n\n const body = new ReadableStream<Uint8Array>({\n start(controller) {\n reader = sourceBody.getReader();\n if (signal.aborted) {\n controller.error(createAbortError(signal));\n return;\n }\n abortHandler = () => controller.error(createAbortError(signal));\n signal.addEventListener(\"abort\", abortHandler, { once: true });\n },\n async pull(controller) {\n try {\n if (signal.aborted) {\n throw createAbortError(signal);\n }\n\n const { done, value } = await reader.read();\n if (done) {\n if (signal.aborted) {\n throw createAbortError(signal);\n }\n if (expectedSize !== undefined && loaded < expectedSize) {\n throw new Error(\n `Fetch ended before reading the full response body (${loaded}/${expectedSize} bytes)`\n );\n }\n cleanup();\n controller.close();\n return;\n }\n\n loaded += value.length;\n controller.enqueue(value);\n } catch (error) {\n cleanup();\n controller.error(error);\n }\n },\n cancel(reason) {\n cleanup();\n return sourceBody.cancel(reason);\n },\n });\n\n return new Response(body, {\n headers: new Headers(response.headers),\n status: response.status,\n statusText: response.statusText,\n });\n}\n\nfunction abortableFetch(url: string, options?: RequestInit): Promise<Response> {\n let modelSignal: AbortSignal | undefined;\n try {\n const pathname = new URL(url).pathname;\n for (const [modelPath, controller] of modelAbortControllers) {\n if (pathname.includes(`/${modelPath}/`)) {\n modelSignal = controller.signal;\n break;\n }\n }\n } catch {\n /* not a parseable URL, proceed without abort */\n }\n const combinedSignal = options?.signal\n ? combineAbortSignals(options.signal, modelSignal)\n : modelSignal;\n return fetch(url, { ...options, ...(combinedSignal ? { signal: combinedSignal } : {}) }).then(\n (response) => wrapAbortableResponse(response, combinedSignal)\n );\n}\n\nconst pipelines = new Map<string, any>();\n\n// ============================================================================\n// Session cache for multi-turn conversations\n// ============================================================================\n\ninterface HftSessionBase {\n readonly modelPath: string;\n}\n\nexport interface HftPrefixRewindSession extends HftSessionBase {\n readonly mode: \"prefix-rewind\";\n /** Snapshot of prefix KV entries. On each call, a fresh DynamicCache is\n * created from these entries so generation doesn't pollute the base prefix.\n * Safe for WASM/CPU tensors; WebGPU would need cloning since update()\n * disposes replaced GPU tensors. */\n readonly baseEntries: Record<string, any>;\n readonly baseSeqLength: number;\n}\n\nexport interface HftProgressiveSession extends HftSessionBase {\n readonly mode: \"progressive\";\n /** Live DynamicCache that grows with the conversation. */\n readonly cache: DynamicCache;\n}\n\nexport type HftSessionState = HftPrefixRewindSession | HftProgressiveSession;\n\nconst hftSessions = new Map<string, HftSessionState>();\n\nexport function getHftSession(sessionId: string): HftSessionState | undefined {\n return hftSessions.get(sessionId);\n}\n\nexport function setHftSession(sessionId: string, state: HftSessionState): void {\n hftSessions.set(sessionId, state);\n}\n\nfunction disposeSessionResources(session: HftSessionState): void {\n if (session.mode === \"progressive\") {\n if (session.cache?.dispose) {\n session.cache.dispose();\n }\n } else {\n for (const tensor of Object.values(session.baseEntries)) {\n if (tensor?.location === \"gpu-buffer\" && typeof tensor.dispose === \"function\") {\n tensor.dispose();\n }\n }\n }\n}\n\nexport function deleteHftSession(sessionId: string): boolean {\n const session = hftSessions.get(sessionId);\n if (session) {\n disposeSessionResources(session);\n }\n return hftSessions.delete(sessionId);\n}\n\nexport function disposeHftSessionsForModel(modelPath: string): void {\n for (const [id, state] of hftSessions) {\n if (state.modelPath === modelPath) {\n disposeSessionResources(state);\n hftSessions.delete(id);\n }\n }\n}\n\n/** In-flight pipeline loads by cache key. Ensures only one load per model at a time to avoid corrupt ONNX files (Protobuf parsing failed). */\nconst pipelineLoadPromises = new Map<string, Promise<any>>();\n\n/**\n * Vision/image pipeline types that require an image processor to be loaded.\n * If the processor is null after pipeline creation the model cache is incomplete\n * (e.g. `preprocessor_config.json` was not downloaded) and the load should be\n * treated as a retriable failure so the missing files are re-fetched.\n */\nconst IMAGE_PIPELINE_TYPES = new Set([\n \"image-classification\",\n \"image-segmentation\",\n \"object-detection\",\n \"image-to-text\",\n \"image-feature-extraction\",\n \"zero-shot-image-classification\",\n \"depth-estimation\",\n \"mask-generation\",\n]);\n\n/**\n * Error message prefix used when an image pipeline's processor failed to\n * initialize (null processor after load). The prefix is checked in\n * `AiJob.classifyProviderError()` to produce a `RetryableJobError` so the\n * queue re-downloads missing processor config files.\n */\nexport const HFT_NULL_PROCESSOR_PREFIX = \"HFT_NULL_PROCESSOR:\";\n\n/**\n * Clear all cached pipelines\n */\nexport function clearPipelineCache(): void {\n pipelines.clear();\n}\n\nexport function hasCachedPipeline(cacheKey: string): boolean {\n return pipelines.has(cacheKey);\n}\n\nexport function removeCachedPipeline(cacheKey: string): boolean {\n return pipelines.delete(cacheKey);\n}\n\n/** True when running in a browser or Web Worker. Transformers.js only accepts device \"wasm\" or \"webgpu\" in the browser build. */\nfunction isBrowserEnv(): boolean {\n if (typeof globalThis === \"undefined\") return false;\n // Main thread\n if (typeof (globalThis as any).window !== \"undefined\") return true;\n // Web Worker (has self but no window)\n if (typeof (globalThis as any).WorkerGlobalScope !== \"undefined\") return true;\n return false;\n}\n\n/**\n * Generate a cache key for a pipeline that includes all configuration options\n * that affect pipeline creation (model_path, pipeline, dtype, device)\n */\nexport function getPipelineCacheKey(model: HfTransformersOnnxModelConfig): string {\n const dtype = model.provider_config.dtype || \"\";\n const device = model.provider_config.device || \"\";\n const revision = model.provider_config.revision || \"main\";\n return `${model.provider_config.model_path}:${model.provider_config.pipeline}:${dtype}:${device}:${revision}`;\n}\n\n/**\n * Helper function to get a pipeline for a model\n * @param progressScaleMax - Maximum progress value for download phase (100 for download-only, 10 for download+run)\n *\n * Explicit `Promise<any>` return avoids TS2883 (inferred type not portable across package boundaries).\n */\nexport async function getPipeline(\n model: HfTransformersOnnxModelConfig,\n onProgress: (progress: number, message?: string, details?: any) => void,\n options: PretrainedModelOptions = {},\n signal?: AbortSignal,\n progressScaleMax: number = 10\n): Promise<any> {\n const cacheKey = getPipelineCacheKey(model);\n if (pipelines.has(cacheKey)) {\n getLogger().debug(\"HFT pipeline cache hit\", { cacheKey });\n return pipelines.get(cacheKey);\n }\n\n // In-flight: only one load per model at a time to avoid concurrent writes to the same\n // ONNX cache path (which can yield \"Protobuf parsing failed\" when one process reads while another writes).\n const inFlight = pipelineLoadPromises.get(cacheKey);\n if (inFlight) {\n try {\n await inFlight;\n } catch {\n // First load failed (e.g. aborted) — fall through to retry below.\n }\n const cached = pipelines.get(cacheKey);\n if (cached) return cached;\n // Load failed for the other caller; fall through to retry (we remove from map in finally).\n }\n\n const loadPromise = doGetPipeline(\n model,\n onProgress,\n options,\n progressScaleMax,\n cacheKey,\n signal\n ).finally(() => {\n pipelineLoadPromises.delete(cacheKey);\n });\n pipelineLoadPromises.set(cacheKey, loadPromise);\n return loadPromise;\n}\n\nconst doGetPipeline = async (\n model: HfTransformersOnnxModelConfig,\n onProgress: (progress: number, message?: string, details?: any) => void,\n options: PretrainedModelOptions,\n progressScaleMax: number,\n cacheKey: string,\n signal?: AbortSignal\n) => {\n // Throttle state for progress events\n let lastProgressTime = 0;\n type FilesByteMap = Record<string, { loaded: number; total: number }>;\n let pendingProgress: {\n progress: number;\n file: string;\n fileProgress: number;\n filesMap?: FilesByteMap;\n } | null = null;\n let throttleTimer: ReturnType<typeof setTimeout> | null = null;\n const THROTTLE_MS = 160;\n\n const buildProgressDetails = (\n file: string,\n fileProgress: number,\n filesMap?: FilesByteMap\n ): { file: string; progress: number; files?: FilesByteMap } => {\n const details: { file: string; progress: number; files?: FilesByteMap } = {\n file,\n progress: fileProgress,\n };\n if (filesMap && Object.keys(filesMap).length > 0) {\n details.files = filesMap;\n }\n return details;\n };\n\n /**\n * Sends a progress event, throttled to avoid flooding the worker channel.\n * Always sends first event and final (>=progressScaleMax) immediately.\n */\n const sendProgress = (\n progress: number,\n file: string,\n fileProgress: number,\n filesMap?: FilesByteMap\n ): void => {\n const now = Date.now();\n const timeSinceLastEvent = now - lastProgressTime;\n const isFirst = lastProgressTime === 0;\n const isFinal = progress >= progressScaleMax;\n\n if (isFirst || isFinal) {\n if (throttleTimer) {\n clearTimeout(throttleTimer);\n throttleTimer = null;\n }\n pendingProgress = null;\n onProgress(\n Math.round(progress),\n \"Downloading model\",\n buildProgressDetails(file, fileProgress, filesMap)\n );\n lastProgressTime = now;\n return;\n }\n\n if (timeSinceLastEvent < THROTTLE_MS) {\n pendingProgress = { progress, file, fileProgress, filesMap };\n if (!throttleTimer) {\n const timeRemaining = Math.max(1, THROTTLE_MS - timeSinceLastEvent);\n throttleTimer = setTimeout(() => {\n throttleTimer = null;\n if (pendingProgress) {\n const p = pendingProgress;\n onProgress(\n Math.round(p.progress),\n \"Downloading model\",\n buildProgressDetails(p.file, p.fileProgress, p.filesMap)\n );\n lastProgressTime = Date.now();\n pendingProgress = null;\n }\n }, timeRemaining);\n }\n return;\n }\n\n onProgress(\n Math.round(progress),\n \"Downloading model\",\n buildProgressDetails(file, fileProgress, filesMap)\n );\n lastProgressTime = now;\n pendingProgress = null;\n };\n\n // Get the abort signal from the signal parameter\n const abortSignal = signal;\n\n // Register a per-model AbortController so abortableFetch can cancel in-flight fetches\n const modelPath = model.provider_config.model_path;\n const modelController = new AbortController();\n modelAbortControllers.set(modelPath, modelController);\n if (abortSignal) {\n if (abortSignal.aborted) {\n modelController.abort();\n } else {\n abortSignal.addEventListener(\"abort\", () => modelController.abort(), { once: true });\n }\n }\n\n // Use aggregate progress_total event from @huggingface/transformers v4 pipeline()\n const progressCallback = (status: ProgressInfo) => {\n if (abortSignal?.aborted) return;\n\n if (status.status === \"progress_total\") {\n const totalStatus = status;\n const scaledProgress = (totalStatus.progress * progressScaleMax) / 100;\n\n // Find the currently active file (one still downloading)\n let activeFile = \"\";\n let activeFileProgress = 0;\n const files: Record<string, { loaded: number; total: number }> | undefined =\n totalStatus.files;\n if (files) {\n for (const [file, info] of Object.entries(files)) {\n if (info.loaded < info.total) {\n activeFile = file;\n activeFileProgress = info.total > 0 ? (info.loaded / info.total) * 100 : 0;\n break;\n }\n }\n if (!activeFile) {\n const fileNames = Object.keys(files);\n if (fileNames.length > 0) {\n activeFile = fileNames[fileNames.length - 1];\n activeFileProgress = 100;\n }\n }\n }\n\n sendProgress(scaledProgress, activeFile, activeFileProgress, files);\n }\n };\n\n let device = model.provider_config.device as string | undefined;\n if (isBrowserEnv()) {\n // we must make a choice for the device in the browser\n if (device === \"gpu\") {\n device = \"webgpu\";\n }\n if (device === \"cpu\") {\n device = \"wasm\";\n }\n if (device !== \"wasm\" && device !== \"webgpu\") {\n device = \"wasm\";\n }\n } else {\n // we can trust the lib to make a choice for the device on the server\n if (device === \"wasm\" || device === \"webgpu\") {\n device = undefined;\n }\n }\n\n const dtype = model.provider_config.dtype || \"\";\n const pipelineOptions: PretrainedModelOptions = {\n revision: model.provider_config.revision || \"main\",\n ...(model.provider_config.use_external_data_format\n ? { useExternalDataFormat: model.provider_config.use_external_data_format }\n : {}),\n ...(dtype ? { dtype: dtype as any } : {}),\n ...(device ? { device: device as any } : {}),\n ...options,\n progress_callback: progressCallback,\n };\n\n // Check if already aborted before starting\n if (abortSignal?.aborted) {\n modelAbortControllers.delete(modelPath);\n throw new Error(\"Operation aborted before pipeline creation\");\n }\n\n const pipelineType = model.provider_config.pipeline;\n\n const { pipeline } = await loadTransformersSDK();\n\n const logger = getLogger();\n const pipelineTimerLabel = `hft:pipeline:${cacheKey}`;\n logger.time(pipelineTimerLabel, { pipelineType, modelPath });\n\n try {\n const result = await pipeline(pipelineType, model.provider_config.model_path, pipelineOptions);\n\n // Flush pending throttled progress and clean up timer\n if (throttleTimer) {\n clearTimeout(throttleTimer);\n throttleTimer = null;\n }\n // pendingProgress may have been set by progressCallback during the pipeline() await\n const finalPending = pendingProgress as {\n progress: number;\n file: string;\n fileProgress: number;\n filesMap?: FilesByteMap;\n } | null;\n if (finalPending) {\n onProgress(\n Math.round(finalPending.progress),\n \"Downloading model\",\n buildProgressDetails(finalPending.file, finalPending.fileProgress, finalPending.filesMap)\n );\n pendingProgress = null;\n }\n\n // Check if aborted after pipeline creation\n if (abortSignal?.aborted) {\n logger.timeEnd(pipelineTimerLabel, { status: \"aborted\" });\n throw new Error(\"Operation aborted after pipeline creation\");\n }\n\n // For image/vision pipelines the processor must be initialized. A null processor\n // means the model cache is incomplete (e.g. preprocessor_config.json was not\n // downloaded, likely because a previous download was aborted). Throw a specific\n // error so the job queue can retry and re-fetch the missing files.\n if (IMAGE_PIPELINE_TYPES.has(pipelineType) && (result as any).processor == null) {\n throw new Error(\n `${HFT_NULL_PROCESSOR_PREFIX} Image processor not initialized for ` +\n `${pipelineType}/${modelPath}. Model cache may be incomplete.`\n );\n }\n\n logger.timeEnd(pipelineTimerLabel, { status: \"loaded\" });\n pipelines.set(cacheKey, result);\n return result;\n } catch (error: any) {\n logger.timeEnd(pipelineTimerLabel, { status: \"error\", error: String(error) });\n // If aborted, throw a clean abort error rather than internal stream errors.\n // Preserve processor-initialization errors so they propagate with their original message.\n if (\n !error?.message?.startsWith(HFT_NULL_PROCESSOR_PREFIX) &&\n (abortSignal?.aborted || modelController.signal.aborted)\n ) {\n throw new Error(\"Pipeline download aborted\");\n }\n throw error;\n } finally {\n modelAbortControllers.delete(modelPath);\n const { random } = await loadTransformersSDK();\n random.seed(model.provider_config.seed ?? undefined);\n }\n};\n",
|
|
6
|
+
"/**\n * @license\n * Copyright 2025 Steven Roussey <sroussey@gmail.com>\n * SPDX-License-Identifier: Apache-2.0\n */\n\nexport const HF_TRANSFORMERS_ONNX = \"HF_TRANSFORMERS_ONNX\";\n/** Job queue for WebGPU/GPU/Metal HFT models (serialized, concurrency 1). */\nexport const HF_TRANSFORMERS_ONNX_GPU = `${HF_TRANSFORMERS_ONNX}_gpu`;\n/** Job queue for WASM/CPU HFT models. */\nexport const HF_TRANSFORMERS_ONNX_CPU = `${HF_TRANSFORMERS_ONNX}_cpu`;\nexport const HTF_CACHE_NAME = \"transformers-cache\";\n\nexport type QuantizationDataType =\n | \"auto\" // Auto-detect based on environment\n | \"fp32\"\n | \"fp16\"\n | \"q8\"\n | \"int8\"\n | \"uint8\"\n | \"q4\"\n | \"bnb4\"\n | \"q4f16\"\n | \"q2\"\n | \"q2f16\"\n | \"q1\"\n | \"q1f16\";\n\nexport const QuantizationDataType = {\n auto: \"auto\",\n fp32: \"fp32\",\n fp16: \"fp16\",\n q8: \"q8\",\n int8: \"int8\",\n uint8: \"uint8\",\n q4: \"q4\",\n bnb4: \"bnb4\",\n q4f16: \"q4f16\",\n q2: \"q2\",\n q2f16: \"q2f16\",\n q1: \"q1\",\n q1f16: \"q1f16\",\n} as const satisfies Record<QuantizationDataType, QuantizationDataType>;\n\ntype TextPipelineUseCase =\n | \"fill-mask\" // https://huggingface.co/tasks/fill-mask\n | \"token-classification\" // https://huggingface.co/tasks/token-classification\n | \"text-generation\" // https://huggingface.co/tasks/text-generation#completion-generation-models\n | \"text2text-generation\" // https://huggingface.co/tasks/text-generation#text-to-text-generation-models\n | \"text-classification\" // https://huggingface.co/tasks/text-classification\n | \"summarization\" // https://huggingface.co/tasks/sentence-similarity\n | \"translation\" // https://huggingface.co/tasks/translation\n | \"feature-extraction\" // https://huggingface.co/tasks/feature-extraction\n | \"zero-shot-classification\" // https://huggingface.co/tasks/zero-shot-classification\n | \"question-answering\"; // https://huggingface.co/tasks/question-answering\n\nconst TextPipelineUseCase = {\n \"fill-mask\": \"fill-mask\",\n \"token-classification\": \"token-classification\",\n \"text-generation\": \"text-generation\",\n \"text2text-generation\": \"text2text-generation\",\n \"text-classification\": \"text-classification\",\n summarization: \"summarization\",\n translation: \"translation\",\n \"feature-extraction\": \"feature-extraction\",\n \"zero-shot-classification\": \"zero-shot-classification\",\n \"question-answering\": \"question-answering\",\n} as const satisfies Record<TextPipelineUseCase, TextPipelineUseCase>;\n\ntype VisionPipelineUseCase =\n | \"background-removal\" // https://huggingface.co/tasks/image-segmentation#background-removal\n | \"image-segmentation\" // https://huggingface.co/tasks/image-segmentation\n | \"depth-estimation\" // https://huggingface.co/tasks/depth-estimation\n | \"image-classification\" // https://huggingface.co/tasks/image-classification\n | \"image-to-image\" // https://huggingface.co/tasks/image-to-image\n // | \"text-to-image\" // https://huggingface.co/tasks/text-to-image\n | \"image-to-text\" // https://huggingface.co/tasks/image-to-text\n // | \"image-text-to-text\" // https://huggingface.co/tasks/image-text-to-text\n | \"object-detection\" // https://huggingface.co/tasks/object-detection\n | \"image-feature-extraction\"; // https://huggingface.co/tasks/image-feature-extraction\n\nconst VisionPipelineUseCase = {\n \"background-removal\": \"background-removal\",\n \"image-segmentation\": \"image-segmentation\",\n \"depth-estimation\": \"depth-estimation\",\n \"image-classification\": \"image-classification\",\n \"image-to-image\": \"image-to-image\",\n // \"text-to-image\": \"text-to-image\",\n \"image-to-text\": \"image-to-text\",\n // \"image-text-to-text\": \"image-text-to-text\",\n \"object-detection\": \"object-detection\",\n \"image-feature-extraction\": \"image-feature-extraction\",\n} as const satisfies Record<VisionPipelineUseCase, VisionPipelineUseCase>;\n\ntype AudioPipelineUseCase =\n | \"audio-classification\" // https://huggingface.co/tasks/audio-classification\n | \"automatic-speech-recognition\" // https://huggingface.co/tasks/automatic-speech-recognition\n | \"text-to-speech\"; // https://huggingface.co/tasks/text-to-speech\n\nconst AudioPipelineUseCase = {\n \"audio-classification\": \"audio-classification\",\n \"automatic-speech-recognition\": \"automatic-speech-recognition\",\n \"text-to-speech\": \"text-to-speech\",\n} as const satisfies Record<AudioPipelineUseCase, AudioPipelineUseCase>;\n\ntype MultimodalPipelineUseCase =\n | \"document-question-answering\" // https://huggingface.co/tasks/document-question-answering\n | \"image-to-text\" // https://huggingface.co/tasks/image-to-text\n | \"zero-shot-audio-classification\" // https://huggingface.co/tasks/zero-shot-audio-classification\n | \"zero-shot-image-classification\" // https://huggingface.co/tasks/zero-shot-image-classification\n | \"zero-shot-object-detection\"; // https://huggingface.co/tasks/zero-shot-object-detection\n\nconst MultimodalPipelineUseCase = {\n \"document-question-answering\": \"document-question-answering\",\n \"image-to-text\": \"image-to-text\",\n \"zero-shot-audio-classification\": \"zero-shot-audio-classification\",\n \"zero-shot-image-classification\": \"zero-shot-image-classification\",\n \"zero-shot-object-detection\": \"zero-shot-object-detection\",\n} as const satisfies Record<MultimodalPipelineUseCase, MultimodalPipelineUseCase>;\n\nexport type PipelineUseCase =\n | TextPipelineUseCase\n | VisionPipelineUseCase\n | AudioPipelineUseCase\n | MultimodalPipelineUseCase;\n\nexport const PipelineUseCase = {\n ...TextPipelineUseCase,\n ...VisionPipelineUseCase,\n ...AudioPipelineUseCase,\n ...MultimodalPipelineUseCase,\n} as const satisfies Record<PipelineUseCase, PipelineUseCase>;\n",
|
|
7
|
+
"/**\n * @license\n * Copyright 2025 Steven Roussey <sroussey@gmail.com>\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { ModelConfigSchema, ModelRecordSchema } from \"@workglow/ai/worker\";\nimport { DataPortSchemaObject, FromSchema } from \"@workglow/util/worker\";\nimport { HF_TRANSFORMERS_ONNX, PipelineUseCase, QuantizationDataType } from \"./HFT_Constants\";\n\nexport const HfTransformersOnnxModelSchema = {\n type: \"object\",\n properties: {\n provider: {\n const: HF_TRANSFORMERS_ONNX,\n description: \"Discriminator: ONNX runtime backend.\",\n },\n provider_config: {\n type: \"object\",\n description: \"ONNX runtime-specific options.\",\n properties: {\n pipeline: {\n type: \"string\",\n enum: Object.values(PipelineUseCase),\n description: \"Pipeline type for the ONNX model.\",\n default: \"text-generation\",\n },\n model_path: {\n type: \"string\",\n description: \"Filesystem path or URI for the ONNX model.\",\n },\n revision: {\n type: \"string\",\n description: \"Git revision (branch, tag, or commit hash) of the model repository.\",\n default: \"main\",\n },\n dtype: {\n type: \"string\",\n enum: Object.values(QuantizationDataType),\n description: \"Data type for the ONNX model.\",\n default: \"auto\",\n },\n device: {\n type: \"string\",\n enum: [\"cpu\", \"gpu\", \"webgpu\", \"wasm\", \"metal\"],\n description: \"High-level device selection.\",\n default: \"webgpu\",\n },\n execution_providers: {\n type: \"array\",\n items: { type: \"string\" },\n description: \"Raw ONNX Runtime execution provider identifiers.\",\n \"x-ui-hidden\": true,\n },\n intra_op_num_threads: {\n type: \"integer\",\n minimum: 1,\n },\n inter_op_num_threads: {\n type: \"integer\",\n minimum: 1,\n },\n use_external_data_format: {\n type: \"boolean\",\n description: \"Whether the model uses external data format.\",\n },\n native_dimensions: {\n type: \"integer\",\n description: \"The native dimensions of the model.\",\n },\n pooling: {\n type: \"string\",\n enum: [\"mean\", \"last_token\", \"cls\"],\n description: \"The pooling strategy to use for the model.\",\n default: \"mean\",\n },\n normalize: {\n type: \"boolean\",\n description: \"Whether the model uses normalization.\",\n default: true,\n },\n language_style: {\n type: \"string\",\n description: \"The language style of the model.\",\n },\n seed: {\n type: \"integer\",\n description:\n \"RNG seed passed to transformers.js sampling. Omit for time-based seeding; set for reproducible generation.\",\n minimum: 0,\n },\n mrl: {\n type: \"boolean\",\n description: \"Whether the model uses matryoshka.\",\n default: false,\n },\n },\n required: [\"model_path\", \"pipeline\"],\n additionalProperties: false,\n if: {\n properties: {\n pipeline: {\n const: \"feature-extraction\",\n },\n },\n },\n then: {\n required: [\"native_dimensions\"],\n },\n },\n },\n required: [\"provider\", \"provider_config\"],\n additionalProperties: true,\n} as const satisfies DataPortSchemaObject;\n\nexport const HfTransformersOnnxModelRecordSchema = {\n type: \"object\",\n properties: {\n ...ModelRecordSchema.properties,\n ...HfTransformersOnnxModelSchema.properties,\n },\n required: [...ModelRecordSchema.required, ...HfTransformersOnnxModelSchema.required],\n additionalProperties: false,\n} as const satisfies DataPortSchemaObject;\n\nexport type HfTransformersOnnxModelRecord = FromSchema<typeof HfTransformersOnnxModelRecordSchema>;\n\nexport const HfTransformersOnnxModelConfigSchema = {\n type: \"object\",\n properties: {\n ...ModelConfigSchema.properties,\n ...HfTransformersOnnxModelSchema.properties,\n },\n required: [...ModelConfigSchema.required, ...HfTransformersOnnxModelSchema.required],\n additionalProperties: false,\n} as const satisfies DataPortSchemaObject;\n\nexport type HfTransformersOnnxModelConfig = FromSchema<typeof HfTransformersOnnxModelConfigSchema>;\n",
|
|
8
|
+
"/**\n * @license\n * Copyright 2025 Steven Roussey <sroussey@gmail.com>\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { AiProvider } from \"@workglow/ai/worker\";\nimport type {\n AiProviderPreviewRunFn,\n AiProviderRunFn,\n AiProviderStreamFn,\n ModelConfig,\n} from \"@workglow/ai/worker\";\nimport { HF_TRANSFORMERS_ONNX } from \"./common/HFT_Constants\";\nimport type { HfTransformersOnnxModelConfig } from \"./common/HFT_ModelSchema\";\nimport { deleteHftSession } from \"./common/HFT_Pipeline\";\n\n/**\n * AI provider for HuggingFace Transformers ONNX models.\n *\n * Supports text, vision, and multimodal tasks via the @huggingface/transformers library.\n *\n * Task run functions are injected via the constructor so that the heavy\n * `@huggingface/transformers` library is only imported where actually needed\n * (inline mode, worker server), not on the main thread in worker mode.\n */\nexport class HuggingFaceTransformersProvider extends AiProvider<HfTransformersOnnxModelConfig> {\n readonly name = HF_TRANSFORMERS_ONNX;\n readonly displayName = \"Hugging Face Transformers (ONNX)\";\n readonly isLocal = true;\n readonly supportsBrowser = true;\n\n readonly taskTypes = [\n \"AiChatTask\",\n \"DownloadModelTask\",\n \"UnloadModelTask\",\n \"ModelInfoTask\",\n \"CountTokensTask\",\n \"TextEmbeddingTask\",\n \"TextGenerationTask\",\n \"TextQuestionAnswerTask\",\n \"TextLanguageDetectionTask\",\n \"TextClassificationTask\",\n \"TextFillMaskTask\",\n \"TextNamedEntityRecognitionTask\",\n \"TextRewriterTask\",\n \"TextSummaryTask\",\n \"TextTranslationTask\",\n \"ImageSegmentationTask\",\n \"ImageToTextTask\",\n \"BackgroundRemovalTask\",\n \"ImageEmbeddingTask\",\n \"ImageClassificationTask\",\n \"ObjectDetectionTask\",\n \"ToolCallingTask\",\n \"StructuredGenerationTask\",\n \"ModelSearchTask\",\n ] as const;\n\n constructor(\n tasks?: Record<string, AiProviderRunFn<any, any, HfTransformersOnnxModelConfig>>,\n streamTasks?: Record<string, AiProviderStreamFn<any, any, HfTransformersOnnxModelConfig>>,\n previewTasks?: Record<string, AiProviderPreviewRunFn<any, any, HfTransformersOnnxModelConfig>>\n ) {\n super(tasks, streamTasks, previewTasks);\n }\n\n override createSession(_model: ModelConfig): string {\n return crypto.randomUUID();\n }\n\n override async disposeSession(sessionId: string): Promise<void> {\n deleteHftSession(sessionId);\n }\n}\n",
|
|
9
|
+
"/**\n * @license\n * Copyright 2025 Steven Roussey <sroussey@gmail.com>\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { QueuedAiProvider } from \"@workglow/ai\";\nimport type {\n IAiExecutionStrategy,\n AiProviderPreviewRunFn,\n AiProviderRunFn,\n AiProviderStreamFn,\n AiProviderRegisterOptions,\n AiProviderQueueConcurrency,\n ModelConfig,\n} from \"@workglow/ai\";\nimport { HF_TRANSFORMERS_ONNX, HF_TRANSFORMERS_ONNX_CPU } from \"./common/HFT_Constants\";\nimport type { HfTransformersOnnxModelConfig } from \"./common/HFT_ModelSchema\";\nimport { deleteHftSession } from \"./common/HFT_Pipeline\";\n\nconst GPU_DEVICES = new Set([\"webgpu\", \"gpu\", \"metal\"]);\n\n/** Default concurrent WASM/CPU ONNX jobs in production (ONNX Runtime / wasm backend). */\nconst HFT_CPU_QUEUE_CONCURRENCY_PRODUCTION = 4;\n\n/**\n * When true, use a single worker for the CPU/WASM queue so tests do not contend on the\n * shared HF cache and ONNX wasm (Vitest, Jest, NODE_ENV=test, Bun test).\n */\nfunction hftIsAutomatedTestEnvironment(): boolean {\n if (typeof process === \"undefined\") {\n return false;\n }\n const e = process.env;\n return (\n e.VITEST === \"true\" ||\n e.NODE_ENV === \"test\" ||\n e.BUN_TEST === \"1\" ||\n e.JEST_WORKER_ID !== undefined\n );\n}\n\nfunction hftDefaultCpuQueueConcurrency(): number {\n return hftIsAutomatedTestEnvironment() ? 1 : HFT_CPU_QUEUE_CONCURRENCY_PRODUCTION;\n}\n\nfunction resolveHftCpuQueueConcurrency(\n concurrency: AiProviderQueueConcurrency | undefined,\n defaultCpu: () => number\n): number {\n if (concurrency === undefined) {\n return defaultCpu();\n }\n if (typeof concurrency === \"number\") {\n return defaultCpu();\n }\n return concurrency.cpu ?? defaultCpu();\n}\n\n/**\n * Main-thread registration (inline or worker-backed).\n * WebGPU/GPU/Metal models use the `gpu` slot (or a numeric `queue.concurrency`, default 1).\n * WASM/CPU models use a separate {@link HF_TRANSFORMERS_ONNX_CPU} queue with higher\n * concurrency (4 in production, 1 under automated tests) to limit ONNX worker contention.\n * Set `cpu` in `queue.concurrency` to override the default.\n */\nexport class HuggingFaceTransformersQueuedProvider extends QueuedAiProvider<HfTransformersOnnxModelConfig> {\n readonly name = HF_TRANSFORMERS_ONNX;\n readonly displayName = \"Hugging Face Transformers (ONNX)\";\n readonly isLocal = true;\n readonly supportsBrowser = true;\n\n private cpuStrategy: IAiExecutionStrategy | undefined;\n\n readonly taskTypes = [\n \"AiChatTask\",\n \"DownloadModelTask\",\n \"UnloadModelTask\",\n \"ModelInfoTask\",\n \"CountTokensTask\",\n \"TextEmbeddingTask\",\n \"TextGenerationTask\",\n \"TextQuestionAnswerTask\",\n \"TextLanguageDetectionTask\",\n \"TextClassificationTask\",\n \"TextFillMaskTask\",\n \"TextNamedEntityRecognitionTask\",\n \"TextRewriterTask\",\n \"TextSummaryTask\",\n \"TextTranslationTask\",\n \"ImageSegmentationTask\",\n \"ImageToTextTask\",\n \"BackgroundRemovalTask\",\n \"ImageEmbeddingTask\",\n \"ImageClassificationTask\",\n \"ObjectDetectionTask\",\n \"ToolCallingTask\",\n \"StructuredGenerationTask\",\n \"ModelSearchTask\",\n ] as const;\n\n constructor(\n tasks?: Record<string, AiProviderRunFn<any, any, HfTransformersOnnxModelConfig>>,\n streamTasks?: Record<string, AiProviderStreamFn<any, any, HfTransformersOnnxModelConfig>>,\n previewTasks?: Record<string, AiProviderPreviewRunFn<any, any, HfTransformersOnnxModelConfig>>\n ) {\n super(tasks, streamTasks, previewTasks);\n }\n\n override createSession(_model: ModelConfig): string {\n return crypto.randomUUID();\n }\n\n override async disposeSession(sessionId: string): Promise<void> {\n deleteHftSession(sessionId);\n }\n\n protected override async afterRegister(options: AiProviderRegisterOptions): Promise<void> {\n await super.afterRegister(options);\n this.cpuStrategy = this.createQueuedStrategy(\n HF_TRANSFORMERS_ONNX_CPU,\n resolveHftCpuQueueConcurrency(options.queue?.concurrency, hftDefaultCpuQueueConcurrency),\n options\n );\n }\n\n protected override getStrategyForModel(model: ModelConfig): IAiExecutionStrategy {\n const device = (model as HfTransformersOnnxModelConfig).provider_config?.device;\n if (device && GPU_DEVICES.has(device)) {\n return this.queuedStrategy!;\n }\n return this.cpuStrategy!;\n }\n}\n",
|
|
10
|
+
"/**\n * @license\n * Copyright 2025 Steven Roussey <sroussey@gmail.com>\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport type { AiProviderRegisterOptions } from \"@workglow/ai\";\nimport { registerProviderWithWorker } from \"@workglow/ai-provider/common\";\nimport { HuggingFaceTransformersQueuedProvider } from \"./HuggingFaceTransformersQueuedProvider\";\n\n/**\n * Register HuggingFace Transformers ONNX on the **main thread** with worker-backed execution\n * (lightweight proxy; heavy work in the worker).\n */\nexport async function registerHuggingFaceTransformers(\n options: AiProviderRegisterOptions & {\n worker: Worker | (() => Worker);\n }\n): Promise<void> {\n await registerProviderWithWorker(\n new HuggingFaceTransformersQueuedProvider(),\n \"HuggingFaceTransformers\",\n options\n );\n}\n",
|
|
11
|
+
"/**\n * @license\n * Copyright 2025 Steven Roussey <sroussey@gmail.com>\n * SPDX-License-Identifier: Apache-2.0\n */\n\nexport const ONNX_QUANTIZATION_SUFFIX_MAPPING = {\n fp32: \"\",\n fp16: \"_fp16\",\n int8: \"_int8\",\n uint8: \"_uint8\",\n q8: \"_quantized\",\n q4: \"_q4\",\n q4f16: \"_q4f16\",\n bnb4: \"_bnb4\",\n q2: \"_q2\",\n q2f16: \"_q2f16\",\n q1: \"_q1\",\n q1f16: \"_q1f16\",\n} as const;\n\nexport type OnnxQuantization = keyof typeof ONNX_QUANTIZATION_SUFFIX_MAPPING;\n\nconst SUFFIXES_LONGEST_FIRST = (\n Object.entries(ONNX_QUANTIZATION_SUFFIX_MAPPING) as [OnnxQuantization, string][]\n)\n .filter(([, suffix]) => suffix !== \"\")\n .sort((a, b) => b[1].length - a[1].length);\n\nexport function parseOnnxQuantizations(params: {\n /** File paths, e.g. from listModels/modelInfo with additionalFields: [\"filePaths\"] */\n filePaths: string[];\n /** Subdirectory containing ONNX files. @default \"onnx\" */\n subfolder?: string;\n}): OnnxQuantization[] {\n const subfolder = params.subfolder ?? \"onnx\";\n const prefix = subfolder + \"/\";\n\n const stems: string[] = [];\n for (const fp of params.filePaths) {\n if (!fp.startsWith(prefix)) continue;\n if (!fp.endsWith(\".onnx\")) continue;\n if (fp.endsWith(\".onnx_data\")) continue;\n stems.push(fp.slice(prefix.length, -\".onnx\".length));\n }\n\n if (stems.length === 0) return [];\n\n const parsed: Array<{ baseName: string; dtype: OnnxQuantization }> = [];\n for (const stem of stems) {\n let matched = false;\n for (const [dtype, suffix] of SUFFIXES_LONGEST_FIRST) {\n if (stem.endsWith(suffix)) {\n parsed.push({ baseName: stem.slice(0, -suffix.length), dtype });\n matched = true;\n break;\n }\n }\n if (!matched) {\n parsed.push({ baseName: stem, dtype: \"fp32\" });\n }\n }\n\n const allBaseNames = new Set(parsed.map((p) => p.baseName));\n const byDtype = new Map<OnnxQuantization, Set<string>>();\n for (const { baseName, dtype } of parsed) {\n let set = byDtype.get(dtype);\n if (!set) {\n set = new Set();\n byDtype.set(dtype, set);\n }\n set.add(baseName);\n }\n\n const allDtypes = Object.keys(ONNX_QUANTIZATION_SUFFIX_MAPPING) as OnnxQuantization[];\n return allDtypes.filter((dtype) => {\n const set = byDtype.get(dtype);\n return set !== undefined && set.size === allBaseNames.size;\n });\n}\n",
|
|
12
|
+
"/**\n * @license\n * Copyright 2025 Steven Roussey <sroussey@gmail.com>\n * SPDX-License-Identifier: Apache-2.0\n */\n\n/**\n * State machine that filters `<tool_call>…</tool_call>` markup out of a\n * stream of text-delta tokens. Tokens that are clearly outside markup are\n * flushed immediately; tokens that *might* be the start of a tag are held\n * in a lookahead buffer until they can be disambiguated.\n *\n * This only handles the XML-tag pattern. Bare-JSON tool calls cannot be\n * reliably detected token-by-token and are still cleaned up via the\n * post-hoc `parseToolCallsFromText` pass on the finish event.\n */\nexport function createToolCallMarkupFilter(emit: (text: string) => void) {\n const OPEN_TAG = \"<tool_call>\";\n const CLOSE_TAG = \"</tool_call>\";\n\n /** \"text\" = normal output, \"tag\" = inside a tool_call block */\n let state: \"text\" | \"tag\" = \"text\";\n /** Buffered text that might be a partial tag prefix */\n let pending = \"\";\n\n function feed(token: string) {\n if (state === \"tag\") {\n // Inside a tool_call block — suppress everything until we see the close tag\n pending += token;\n const closeIdx = pending.indexOf(CLOSE_TAG);\n if (closeIdx !== -1) {\n // End of the tool_call block; resume normal output after the close tag\n const afterClose = pending.slice(closeIdx + CLOSE_TAG.length);\n pending = \"\";\n state = \"text\";\n if (afterClose.length > 0) {\n feed(afterClose);\n }\n }\n // else: still inside the tag block, keep suppressing\n return;\n }\n\n // state === \"text\"\n const combined = pending + token;\n\n // Check for a complete open tag\n const openIdx = combined.indexOf(OPEN_TAG);\n if (openIdx !== -1) {\n // Emit everything before the tag\n const before = combined.slice(0, openIdx);\n if (before.length > 0) {\n emit(before);\n }\n // Switch to tag state; feed the remainder (after the open tag) back through\n pending = \"\";\n state = \"tag\";\n const afterOpen = combined.slice(openIdx + OPEN_TAG.length);\n if (afterOpen.length > 0) {\n feed(afterOpen);\n }\n return;\n }\n\n // Check if the tail of `combined` could be the start of \"<tool_call>\"\n // e.g. combined ends with \"<\", \"<t\", \"<to\", ..., \"<tool_call\"\n let prefixLen = 0;\n for (let len = Math.min(combined.length, OPEN_TAG.length - 1); len >= 1; len--) {\n if (combined.endsWith(OPEN_TAG.slice(0, len))) {\n prefixLen = len;\n break;\n }\n }\n\n if (prefixLen > 0) {\n // The tail is ambiguous — hold it back, flush the rest\n const safe = combined.slice(0, combined.length - prefixLen);\n if (safe.length > 0) {\n emit(safe);\n }\n pending = combined.slice(combined.length - prefixLen);\n } else {\n // No ambiguity — flush everything\n if (combined.length > 0) {\n emit(combined);\n }\n pending = \"\";\n }\n }\n\n /** Flush any remaining buffered text (called when the stream ends). */\n function flush() {\n if (pending.length > 0 && state === \"text\") {\n emit(pending);\n pending = \"\";\n }\n // If state === \"tag\", the pending content is suppressed tool-call markup\n pending = \"\";\n state = \"text\";\n }\n\n return { feed, flush };\n}\n"
|
|
13
|
+
],
|
|
14
|
+
"mappings": ";;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AAOA;AAUO,SAAS,cAAc,CAAC,KAAmB;AAAA,EAChD,YAAY;AAAA,EACZ,IAAI,kBAAkB;AAAA,IACpB,iBAAiB,IAAI,WAAW;AAAA,EAClC;AAAA;AAGF,eAAsB,mBAAmB,GAAG;AAAA,EAC1C,IAAI,CAAC,kBAAkB;AAAA,IACrB,IAAI;AAAA,MACF,mBAAmB,MAAa;AAAA,MAChC,iBAAiB,IAAI,QAAQ;AAAA,MAC7B,IAAI,WAAW;AAAA,QACb,iBAAiB,IAAI,WAAW;AAAA,MAClC;AAAA,MACA,MAAM;AAAA,MACN,MAAM,IAAI,MACR,8HACF;AAAA;AAAA,EAEJ;AAAA,EACA,OAAO;AAAA;AAMT,SAAS,mBAAmB,CAC1B,gBACA,aACyB;AAAA,EACzB,IAAI,CAAC,gBAAgB;AAAA,IACnB,OAAO;AAAA,EACT;AAAA,EACA,IAAI,CAAC,aAAa;AAAA,IAChB,OAAO;AAAA,EACT;AAAA,EACA,IAAI,eAAe,WAAW,YAAY,SAAS;AAAA,IACjD,OAAO,YAAY,MAAM,eAAe,UAAU,YAAY,MAAM;AAAA,EACtE;AAAA,EACA,IAAI,OAAO,YAAY,QAAQ,YAAY;AAAA,IACzC,OAAO,YAAY,IAAI,CAAC,gBAAgB,WAAW,CAAC;AAAA,EACtD;AAAA,EAEA,MAAM,aAAa,IAAI;AAAA,EACvB,MAAM,QAAQ,CAAC,UAAiB;AAAA,IAC9B,MAAM,SAAS,MAAM;AAAA,IACrB,WAAW,MAAM,OAAO,MAAM;AAAA;AAAA,EAEhC,eAAe,iBAAiB,SAAS,OAAO,EAAE,MAAM,KAAK,CAAC;AAAA,EAC9D,YAAY,iBAAiB,SAAS,OAAO,EAAE,MAAM,KAAK,CAAC;AAAA,EAC3D,OAAO,WAAW;AAAA;AAGpB,SAAS,gBAAgB,CAAC,QAA4B;AAAA,EACpD,MAAM,SAAS,OAAO;AAAA,EACtB,IAAI,kBAAkB,OAAO;AAAA,IAC3B,OAAO;AAAA,EACT;AAAA,EACA,OAAO,IAAI,MAAM,OAAO,UAAU,eAAe,CAAC;AAAA;AAGpD,SAAS,qBAAqB,CAAC,UAAoB,QAA2C;AAAA,EAC5F,IAAI,CAAC,UAAU,CAAC,SAAS,MAAM;AAAA,IAC7B,OAAO;AAAA,EACT;AAAA,EAEA,MAAM,sBAAsB,SAAS,QAAQ,IAAI,gBAAgB;AAAA,EACjE,MAAM,eACJ,uBAAuB,QAAQ,KAAK,mBAAmB,IACnD,OAAO,SAAS,qBAAqB,EAAE,IACvC;AAAA,EACN,MAAM,aAAa,SAAS;AAAA,EAM5B,IAAI;AAAA,EACJ,IAAI;AAAA,EACJ,IAAI,SAAS;AAAA,EAEb,MAAM,UAAU,MAAM;AAAA,IACpB,IAAI,cAAc;AAAA,MAChB,OAAO,oBAAoB,SAAS,YAAY;AAAA,MAChD,eAAe;AAAA,IACjB;AAAA,IACA,QAAQ,YAAY;AAAA;AAAA,EAGtB,MAAM,OAAO,IAAI,eAA2B;AAAA,IAC1C,KAAK,CAAC,YAAY;AAAA,MAChB,SAAS,WAAW,UAAU;AAAA,MAC9B,IAAI,OAAO,SAAS;AAAA,QAClB,WAAW,MAAM,iBAAiB,MAAM,CAAC;AAAA,QACzC;AAAA,MACF;AAAA,MACA,eAAe,MAAM,WAAW,MAAM,iBAAiB,MAAM,CAAC;AAAA,MAC9D,OAAO,iBAAiB,SAAS,cAAc,EAAE,MAAM,KAAK,CAAC;AAAA;AAAA,SAEzD,KAAI,CAAC,YAAY;AAAA,MACrB,IAAI;AAAA,QACF,IAAI,OAAO,SAAS;AAAA,UAClB,MAAM,iBAAiB,MAAM;AAAA,QAC/B;AAAA,QAEA,QAAQ,MAAM,UAAU,MAAM,OAAO,KAAK;AAAA,QAC1C,IAAI,MAAM;AAAA,UACR,IAAI,OAAO,SAAS;AAAA,YAClB,MAAM,iBAAiB,MAAM;AAAA,UAC/B;AAAA,UACA,IAAI,iBAAiB,aAAa,SAAS,cAAc;AAAA,YACvD,MAAM,IAAI,MACR,sDAAsD,UAAU,qBAClE;AAAA,UACF;AAAA,UACA,QAAQ;AAAA,UACR,WAAW,MAAM;AAAA,UACjB;AAAA,QACF;AAAA,QAEA,UAAU,MAAM;AAAA,QAChB,WAAW,QAAQ,KAAK;AAAA,QACxB,OAAO,OAAO;AAAA,QACd,QAAQ;AAAA,QACR,WAAW,MAAM,KAAK;AAAA;AAAA;AAAA,IAG1B,MAAM,CAAC,QAAQ;AAAA,MACb,QAAQ;AAAA,MACR,OAAO,WAAW,OAAO,MAAM;AAAA;AAAA,EAEnC,CAAC;AAAA,EAED,OAAO,IAAI,SAAS,MAAM;AAAA,IACxB,SAAS,IAAI,QAAQ,SAAS,OAAO;AAAA,IACrC,QAAQ,SAAS;AAAA,IACjB,YAAY,SAAS;AAAA,EACvB,CAAC;AAAA;AAGH,SAAS,cAAc,CAAC,KAAa,SAA0C;AAAA,EAC7E,IAAI;AAAA,EACJ,IAAI;AAAA,IACF,MAAM,WAAW,IAAI,IAAI,GAAG,EAAE;AAAA,IAC9B,YAAY,WAAW,eAAe,uBAAuB;AAAA,MAC3D,IAAI,SAAS,SAAS,IAAI,YAAY,GAAG;AAAA,QACvC,cAAc,WAAW;AAAA,QACzB;AAAA,MACF;AAAA,IACF;AAAA,IACA,MAAM;AAAA,EAGR,MAAM,iBAAiB,SAAS,SAC5B,oBAAoB,QAAQ,QAAQ,WAAW,IAC/C;AAAA,EACJ,OAAO,MAAM,KAAK,KAAK,YAAa,iBAAiB,EAAE,QAAQ,eAAe,IAAI,CAAC,EAAG,CAAC,EAAE,KACvF,CAAC,aAAa,sBAAsB,UAAU,cAAc,CAC9D;AAAA;AAiCK,SAAS,aAAa,CAAC,WAAgD;AAAA,EAC5E,OAAO,YAAY,IAAI,SAAS;AAAA;AAG3B,SAAS,aAAa,CAAC,WAAmB,OAA8B;AAAA,EAC7E,YAAY,IAAI,WAAW,KAAK;AAAA;AAGlC,SAAS,uBAAuB,CAAC,SAAgC;AAAA,EAC/D,IAAI,QAAQ,SAAS,eAAe;AAAA,IAClC,IAAI,QAAQ,OAAO,SAAS;AAAA,MAC1B,QAAQ,MAAM,QAAQ;AAAA,IACxB;AAAA,EACF,EAAO;AAAA,IACL,WAAW,UAAU,OAAO,OAAO,QAAQ,WAAW,GAAG;AAAA,MACvD,IAAI,QAAQ,aAAa,gBAAgB,OAAO,OAAO,YAAY,YAAY;AAAA,QAC7E,OAAO,QAAQ;AAAA,MACjB;AAAA,IACF;AAAA;AAAA;AAIG,SAAS,gBAAgB,CAAC,WAA4B;AAAA,EAC3D,MAAM,UAAU,YAAY,IAAI,SAAS;AAAA,EACzC,IAAI,SAAS;AAAA,IACX,wBAAwB,OAAO;AAAA,EACjC;AAAA,EACA,OAAO,YAAY,OAAO,SAAS;AAAA;AAG9B,SAAS,0BAA0B,CAAC,WAAyB;AAAA,EAClE,YAAY,IAAI,UAAU,aAAa;AAAA,IACrC,IAAI,MAAM,cAAc,WAAW;AAAA,MACjC,wBAAwB,KAAK;AAAA,MAC7B,YAAY,OAAO,EAAE;AAAA,IACvB;AAAA,EACF;AAAA;AAkCK,SAAS,kBAAkB,GAAS;AAAA,EACzC,UAAU,MAAM;AAAA;AAGX,SAAS,iBAAiB,CAAC,UAA2B;AAAA,EAC3D,OAAO,UAAU,IAAI,QAAQ;AAAA;AAGxB,SAAS,oBAAoB,CAAC,UAA2B;AAAA,EAC9D,OAAO,UAAU,OAAO,QAAQ;AAAA;AAIlC,SAAS,YAAY,GAAY;AAAA,EAC/B,IAAI,OAAO,eAAe;AAAA,IAAa,OAAO;AAAA,EAE9C,IAAI,OAAQ,WAAmB,WAAW;AAAA,IAAa,OAAO;AAAA,EAE9D,IAAI,OAAQ,WAAmB,sBAAsB;AAAA,IAAa,OAAO;AAAA,EACzE,OAAO;AAAA;AAOF,SAAS,mBAAmB,CAAC,OAA8C;AAAA,EAChF,MAAM,QAAQ,MAAM,gBAAgB,SAAS;AAAA,EAC7C,MAAM,SAAS,MAAM,gBAAgB,UAAU;AAAA,EAC/C,MAAM,WAAW,MAAM,gBAAgB,YAAY;AAAA,EACnD,OAAO,GAAG,MAAM,gBAAgB,cAAc,MAAM,gBAAgB,YAAY,SAAS,UAAU;AAAA;AASrG,eAAsB,WAAW,CAC/B,OACA,YACA,UAAkC,CAAC,GACnC,QACA,mBAA2B,IACb;AAAA,EACd,MAAM,WAAW,oBAAoB,KAAK;AAAA,EAC1C,IAAI,UAAU,IAAI,QAAQ,GAAG;AAAA,IAC3B,UAAU,EAAE,MAAM,0BAA0B,EAAE,SAAS,CAAC;AAAA,IACxD,OAAO,UAAU,IAAI,QAAQ;AAAA,EAC/B;AAAA,EAIA,MAAM,WAAW,qBAAqB,IAAI,QAAQ;AAAA,EAClD,IAAI,UAAU;AAAA,IACZ,IAAI;AAAA,MACF,MAAM;AAAA,MACN,MAAM;AAAA,IAGR,MAAM,SAAS,UAAU,IAAI,QAAQ;AAAA,IACrC,IAAI;AAAA,MAAQ,OAAO;AAAA,EAErB;AAAA,EAEA,MAAM,cAAc,cAClB,OACA,YACA,SACA,kBACA,UACA,MACF,EAAE,QAAQ,MAAM;AAAA,IACd,qBAAqB,OAAO,QAAQ;AAAA,GACrC;AAAA,EACD,qBAAqB,IAAI,UAAU,WAAW;AAAA,EAC9C,OAAO;AAAA;AAAA,IA1VL,kBACA,WA+BE,uBAyIA,WA4BA,aA0CA,sBAQA,sBAiBO,4BAA4B,uBAqFnC,gBAAgB,OACpB,OACA,YACA,SACA,kBACA,UACA,WACG;AAAA,EAEH,IAAI,mBAAmB;AAAA,EAEvB,IAAI,kBAKO;AAAA,EACX,IAAI,gBAAsD;AAAA,EAC1D,MAAM,cAAc;AAAA,EAEpB,MAAM,uBAAuB,CAC3B,MACA,cACA,aAC6D;AAAA,IAC7D,MAAM,UAAoE;AAAA,MACxE;AAAA,MACA,UAAU;AAAA,IACZ;AAAA,IACA,IAAI,YAAY,OAAO,KAAK,QAAQ,EAAE,SAAS,GAAG;AAAA,MAChD,QAAQ,QAAQ;AAAA,IAClB;AAAA,IACA,OAAO;AAAA;AAAA,EAOT,MAAM,eAAe,CACnB,UACA,MACA,cACA,aACS;AAAA,IACT,MAAM,MAAM,KAAK,IAAI;AAAA,IACrB,MAAM,qBAAqB,MAAM;AAAA,IACjC,MAAM,UAAU,qBAAqB;AAAA,IACrC,MAAM,UAAU,YAAY;AAAA,IAE5B,IAAI,WAAW,SAAS;AAAA,MACtB,IAAI,eAAe;AAAA,QACjB,aAAa,aAAa;AAAA,QAC1B,gBAAgB;AAAA,MAClB;AAAA,MACA,kBAAkB;AAAA,MAClB,WACE,KAAK,MAAM,QAAQ,GACnB,qBACA,qBAAqB,MAAM,cAAc,QAAQ,CACnD;AAAA,MACA,mBAAmB;AAAA,MACnB;AAAA,IACF;AAAA,IAEA,IAAI,qBAAqB,aAAa;AAAA,MACpC,kBAAkB,EAAE,UAAU,MAAM,cAAc,SAAS;AAAA,MAC3D,IAAI,CAAC,eAAe;AAAA,QAClB,MAAM,gBAAgB,KAAK,IAAI,GAAG,cAAc,kBAAkB;AAAA,QAClE,gBAAgB,WAAW,MAAM;AAAA,UAC/B,gBAAgB;AAAA,UAChB,IAAI,iBAAiB;AAAA,YACnB,MAAM,IAAI;AAAA,YACV,WACE,KAAK,MAAM,EAAE,QAAQ,GACrB,qBACA,qBAAqB,EAAE,MAAM,EAAE,cAAc,EAAE,QAAQ,CACzD;AAAA,YACA,mBAAmB,KAAK,IAAI;AAAA,YAC5B,kBAAkB;AAAA,UACpB;AAAA,WACC,aAAa;AAAA,MAClB;AAAA,MACA;AAAA,IACF;AAAA,IAEA,WACE,KAAK,MAAM,QAAQ,GACnB,qBACA,qBAAqB,MAAM,cAAc,QAAQ,CACnD;AAAA,IACA,mBAAmB;AAAA,IACnB,kBAAkB;AAAA;AAAA,EAIpB,MAAM,cAAc;AAAA,EAGpB,MAAM,YAAY,MAAM,gBAAgB;AAAA,EACxC,MAAM,kBAAkB,IAAI;AAAA,EAC5B,sBAAsB,IAAI,WAAW,eAAe;AAAA,EACpD,IAAI,aAAa;AAAA,IACf,IAAI,YAAY,SAAS;AAAA,MACvB,gBAAgB,MAAM;AAAA,IACxB,EAAO;AAAA,MACL,YAAY,iBAAiB,SAAS,MAAM,gBAAgB,MAAM,GAAG,EAAE,MAAM,KAAK,CAAC;AAAA;AAAA,EAEvF;AAAA,EAGA,MAAM,mBAAmB,CAAC,WAAyB;AAAA,IACjD,IAAI,aAAa;AAAA,MAAS;AAAA,IAE1B,IAAI,OAAO,WAAW,kBAAkB;AAAA,MACtC,MAAM,cAAc;AAAA,MACpB,MAAM,iBAAkB,YAAY,WAAW,mBAAoB;AAAA,MAGnE,IAAI,aAAa;AAAA,MACjB,IAAI,qBAAqB;AAAA,MACzB,MAAM,QACJ,YAAY;AAAA,MACd,IAAI,OAAO;AAAA,QACT,YAAY,MAAM,SAAS,OAAO,QAAQ,KAAK,GAAG;AAAA,UAChD,IAAI,KAAK,SAAS,KAAK,OAAO;AAAA,YAC5B,aAAa;AAAA,YACb,qBAAqB,KAAK,QAAQ,IAAK,KAAK,SAAS,KAAK,QAAS,MAAM;AAAA,YACzE;AAAA,UACF;AAAA,QACF;AAAA,QACA,IAAI,CAAC,YAAY;AAAA,UACf,MAAM,YAAY,OAAO,KAAK,KAAK;AAAA,UACnC,IAAI,UAAU,SAAS,GAAG;AAAA,YACxB,aAAa,UAAU,UAAU,SAAS;AAAA,YAC1C,qBAAqB;AAAA,UACvB;AAAA,QACF;AAAA,MACF;AAAA,MAEA,aAAa,gBAAgB,YAAY,oBAAoB,KAAK;AAAA,IACpE;AAAA;AAAA,EAGF,IAAI,SAAS,MAAM,gBAAgB;AAAA,EACnC,IAAI,aAAa,GAAG;AAAA,IAElB,IAAI,WAAW,OAAO;AAAA,MACpB,SAAS;AAAA,IACX;AAAA,IACA,IAAI,WAAW,OAAO;AAAA,MACpB,SAAS;AAAA,IACX;AAAA,IACA,IAAI,WAAW,UAAU,WAAW,UAAU;AAAA,MAC5C,SAAS;AAAA,IACX;AAAA,EACF,EAAO;AAAA,IAEL,IAAI,WAAW,UAAU,WAAW,UAAU;AAAA,MAC5C,SAAS;AAAA,IACX;AAAA;AAAA,EAGF,MAAM,QAAQ,MAAM,gBAAgB,SAAS;AAAA,EAC7C,MAAM,kBAA0C;AAAA,IAC9C,UAAU,MAAM,gBAAgB,YAAY;AAAA,OACxC,MAAM,gBAAgB,2BACtB,EAAE,uBAAuB,MAAM,gBAAgB,yBAAyB,IACxE,CAAC;AAAA,OACD,QAAQ,EAAE,MAAoB,IAAI,CAAC;AAAA,OACnC,SAAS,EAAE,OAAsB,IAAI,CAAC;AAAA,OACvC;AAAA,IACH,mBAAmB;AAAA,EACrB;AAAA,EAGA,IAAI,aAAa,SAAS;AAAA,IACxB,sBAAsB,OAAO,SAAS;AAAA,IACtC,MAAM,IAAI,MAAM,4CAA4C;AAAA,EAC9D;AAAA,EAEA,MAAM,eAAe,MAAM,gBAAgB;AAAA,EAE3C,QAAQ,aAAa,MAAM,oBAAoB;AAAA,EAE/C,MAAM,SAAS,UAAU;AAAA,EACzB,MAAM,qBAAqB,gBAAgB;AAAA,EAC3C,OAAO,KAAK,oBAAoB,EAAE,cAAc,UAAU,CAAC;AAAA,EAE3D,IAAI;AAAA,IACF,MAAM,SAAS,MAAM,SAAS,cAAc,MAAM,gBAAgB,YAAY,eAAe;AAAA,IAG7F,IAAI,eAAe;AAAA,MACjB,aAAa,aAAa;AAAA,MAC1B,gBAAgB;AAAA,IAClB;AAAA,IAEA,MAAM,eAAe;AAAA,IAMrB,IAAI,cAAc;AAAA,MAChB,WACE,KAAK,MAAM,aAAa,QAAQ,GAChC,qBACA,qBAAqB,aAAa,MAAM,aAAa,cAAc,aAAa,QAAQ,CAC1F;AAAA,MACA,kBAAkB;AAAA,IACpB;AAAA,IAGA,IAAI,aAAa,SAAS;AAAA,MACxB,OAAO,QAAQ,oBAAoB,EAAE,QAAQ,UAAU,CAAC;AAAA,MACxD,MAAM,IAAI,MAAM,2CAA2C;AAAA,IAC7D;AAAA,IAMA,IAAI,qBAAqB,IAAI,YAAY,KAAM,OAAe,aAAa,MAAM;AAAA,MAC/E,MAAM,IAAI,MACR,GAAG,iEACE,gBAAgB,2CACvB;AAAA,IACF;AAAA,IAEA,OAAO,QAAQ,oBAAoB,EAAE,QAAQ,SAAS,CAAC;AAAA,IACvD,UAAU,IAAI,UAAU,MAAM;AAAA,IAC9B,OAAO;AAAA,IACP,OAAO,OAAY;AAAA,IACnB,OAAO,QAAQ,oBAAoB,EAAE,QAAQ,SAAS,OAAO,OAAO,KAAK,EAAE,CAAC;AAAA,IAG5E,IACE,CAAC,OAAO,SAAS,WAAW,yBAAyB,MACpD,aAAa,WAAW,gBAAgB,OAAO,UAChD;AAAA,MACA,MAAM,IAAI,MAAM,2BAA2B;AAAA,IAC7C;AAAA,IACA,MAAM;AAAA,YACN;AAAA,IACA,sBAAsB,OAAO,SAAS;AAAA,IACtC,QAAQ,WAAW,MAAM,oBAAoB;AAAA,IAC7C,OAAO,KAAK,MAAM,gBAAgB,QAAQ,SAAS;AAAA;AAAA;AAAA;AAAA,EApjBjD,wBAAwB,IAAI;AAAA,EAyI5B,YAAY,IAAI;AAAA,EA4BhB,cAAc,IAAI;AAAA,EA0ClB,uBAAuB,IAAI;AAAA,EAQ3B,uBAAuB,IAAI,IAAI;AAAA,IACnC;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,EACF,CAAC;AAAA;;;ACpQM,IAAM,uBAAuB;AAE7B,IAAM,2BAA2B,GAAG;AAEpC,IAAM,2BAA2B,GAAG;AACpC,IAAM,iBAAiB;AAiBvB,IAAM,uBAAuB;AAAA,EAClC,MAAM;AAAA,EACN,MAAM;AAAA,EACN,MAAM;AAAA,EACN,IAAI;AAAA,EACJ,MAAM;AAAA,EACN,OAAO;AAAA,EACP,IAAI;AAAA,EACJ,MAAM;AAAA,EACN,OAAO;AAAA,EACP,IAAI;AAAA,EACJ,OAAO;AAAA,EACP,IAAI;AAAA,EACJ,OAAO;AACT;AAcA,IAAM,sBAAsB;AAAA,EAC1B,aAAa;AAAA,EACb,wBAAwB;AAAA,EACxB,mBAAmB;AAAA,EACnB,wBAAwB;AAAA,EACxB,uBAAuB;AAAA,EACvB,eAAe;AAAA,EACf,aAAa;AAAA,EACb,sBAAsB;AAAA,EACtB,4BAA4B;AAAA,EAC5B,sBAAsB;AACxB;AAcA,IAAM,wBAAwB;AAAA,EAC5B,sBAAsB;AAAA,EACtB,sBAAsB;AAAA,EACtB,oBAAoB;AAAA,EACpB,wBAAwB;AAAA,EACxB,kBAAkB;AAAA,EAElB,iBAAiB;AAAA,EAEjB,oBAAoB;AAAA,EACpB,4BAA4B;AAC9B;AAOA,IAAM,uBAAuB;AAAA,EAC3B,wBAAwB;AAAA,EACxB,gCAAgC;AAAA,EAChC,kBAAkB;AACpB;AASA,IAAM,4BAA4B;AAAA,EAChC,+BAA+B;AAAA,EAC/B,iBAAiB;AAAA,EACjB,kCAAkC;AAAA,EAClC,kCAAkC;AAAA,EAClC,8BAA8B;AAChC;AAQO,IAAM,kBAAkB;AAAA,KAC1B;AAAA,KACA;AAAA,KACA;AAAA,KACA;AACL;;;AC7HA;AAIO,IAAM,gCAAgC;AAAA,EAC3C,MAAM;AAAA,EACN,YAAY;AAAA,IACV,UAAU;AAAA,MACR,OAAO;AAAA,MACP,aAAa;AAAA,IACf;AAAA,IACA,iBAAiB;AAAA,MACf,MAAM;AAAA,MACN,aAAa;AAAA,MACb,YAAY;AAAA,QACV,UAAU;AAAA,UACR,MAAM;AAAA,UACN,MAAM,OAAO,OAAO,eAAe;AAAA,UACnC,aAAa;AAAA,UACb,SAAS;AAAA,QACX;AAAA,QACA,YAAY;AAAA,UACV,MAAM;AAAA,UACN,aAAa;AAAA,QACf;AAAA,QACA,UAAU;AAAA,UACR,MAAM;AAAA,UACN,aAAa;AAAA,UACb,SAAS;AAAA,QACX;AAAA,QACA,OAAO;AAAA,UACL,MAAM;AAAA,UACN,MAAM,OAAO,OAAO,oBAAoB;AAAA,UACxC,aAAa;AAAA,UACb,SAAS;AAAA,QACX;AAAA,QACA,QAAQ;AAAA,UACN,MAAM;AAAA,UACN,MAAM,CAAC,OAAO,OAAO,UAAU,QAAQ,OAAO;AAAA,UAC9C,aAAa;AAAA,UACb,SAAS;AAAA,QACX;AAAA,QACA,qBAAqB;AAAA,UACnB,MAAM;AAAA,UACN,OAAO,EAAE,MAAM,SAAS;AAAA,UACxB,aAAa;AAAA,UACb,eAAe;AAAA,QACjB;AAAA,QACA,sBAAsB;AAAA,UACpB,MAAM;AAAA,UACN,SAAS;AAAA,QACX;AAAA,QACA,sBAAsB;AAAA,UACpB,MAAM;AAAA,UACN,SAAS;AAAA,QACX;AAAA,QACA,0BAA0B;AAAA,UACxB,MAAM;AAAA,UACN,aAAa;AAAA,QACf;AAAA,QACA,mBAAmB;AAAA,UACjB,MAAM;AAAA,UACN,aAAa;AAAA,QACf;AAAA,QACA,SAAS;AAAA,UACP,MAAM;AAAA,UACN,MAAM,CAAC,QAAQ,cAAc,KAAK;AAAA,UAClC,aAAa;AAAA,UACb,SAAS;AAAA,QACX;AAAA,QACA,WAAW;AAAA,UACT,MAAM;AAAA,UACN,aAAa;AAAA,UACb,SAAS;AAAA,QACX;AAAA,QACA,gBAAgB;AAAA,UACd,MAAM;AAAA,UACN,aAAa;AAAA,QACf;AAAA,QACA,MAAM;AAAA,UACJ,MAAM;AAAA,UACN,aACE;AAAA,UACF,SAAS;AAAA,QACX;AAAA,QACA,KAAK;AAAA,UACH,MAAM;AAAA,UACN,aAAa;AAAA,UACb,SAAS;AAAA,QACX;AAAA,MACF;AAAA,MACA,UAAU,CAAC,cAAc,UAAU;AAAA,MACnC,sBAAsB;AAAA,MACtB,IAAI;AAAA,QACF,YAAY;AAAA,UACV,UAAU;AAAA,YACR,OAAO;AAAA,UACT;AAAA,QACF;AAAA,MACF;AAAA,MACA,MAAM;AAAA,QACJ,UAAU,CAAC,mBAAmB;AAAA,MAChC;AAAA,IACF;AAAA,EACF;AAAA,EACA,UAAU,CAAC,YAAY,iBAAiB;AAAA,EACxC,sBAAsB;AACxB;AAEO,IAAM,sCAAsC;AAAA,EACjD,MAAM;AAAA,EACN,YAAY;AAAA,OACP,kBAAkB;AAAA,OAClB,8BAA8B;AAAA,EACnC;AAAA,EACA,UAAU,CAAC,GAAG,kBAAkB,UAAU,GAAG,8BAA8B,QAAQ;AAAA,EACnF,sBAAsB;AACxB;AAIO,IAAM,sCAAsC;AAAA,EACjD,MAAM;AAAA,EACN,YAAY;AAAA,OACP,kBAAkB;AAAA,OAClB,8BAA8B;AAAA,EACnC;AAAA,EACA,UAAU,CAAC,GAAG,kBAAkB,UAAU,GAAG,8BAA8B,QAAQ;AAAA,EACnF,sBAAsB;AACxB;;;ACjIA;AASA;AAAA;AAWO,MAAM,wCAAwC,WAA0C;AAAA,EACpF,OAAO;AAAA,EACP,cAAc;AAAA,EACd,UAAU;AAAA,EACV,kBAAkB;AAAA,EAElB,YAAY;AAAA,IACnB;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,EACF;AAAA,EAEA,WAAW,CACT,OACA,aACA,cACA;AAAA,IACA,MAAM,OAAO,aAAa,YAAY;AAAA;AAAA,EAG/B,aAAa,CAAC,QAA6B;AAAA,IAClD,OAAO,OAAO,WAAW;AAAA;AAAA,OAGZ,eAAc,CAAC,WAAkC;AAAA,IAC9D,iBAAiB,SAAS;AAAA;AAE9B;;;ACpEA;AAYA;AAEA,IAAM,cAAc,IAAI,IAAI,CAAC,UAAU,OAAO,OAAO,CAAC;AAGtD,IAAM,uCAAuC;AAM7C,SAAS,6BAA6B,GAAY;AAAA,EAChD,IAAI,OAAO,YAAY,aAAa;AAAA,IAClC,OAAO;AAAA,EACT;AAAA,EACA,MAAM,IAAI,QAAQ;AAAA,EAClB,OACE,EAAE,WAAW,UACb,EAAE,aAAa,UACf,EAAE,aAAa,OACf,EAAE,mBAAmB;AAAA;AAIzB,SAAS,6BAA6B,GAAW;AAAA,EAC/C,OAAO,8BAA8B,IAAI,IAAI;AAAA;AAG/C,SAAS,6BAA6B,CACpC,aACA,YACQ;AAAA,EACR,IAAI,gBAAgB,WAAW;AAAA,IAC7B,OAAO,WAAW;AAAA,EACpB;AAAA,EACA,IAAI,OAAO,gBAAgB,UAAU;AAAA,IACnC,OAAO,WAAW;AAAA,EACpB;AAAA,EACA,OAAO,YAAY,OAAO,WAAW;AAAA;AAAA;AAUhC,MAAM,8CAA8C,iBAAgD;AAAA,EAChG,OAAO;AAAA,EACP,cAAc;AAAA,EACd,UAAU;AAAA,EACV,kBAAkB;AAAA,EAEnB;AAAA,EAEC,YAAY;AAAA,IACnB;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,EACF;AAAA,EAEA,WAAW,CACT,OACA,aACA,cACA;AAAA,IACA,MAAM,OAAO,aAAa,YAAY;AAAA;AAAA,EAG/B,aAAa,CAAC,QAA6B;AAAA,IAClD,OAAO,OAAO,WAAW;AAAA;AAAA,OAGZ,eAAc,CAAC,WAAkC;AAAA,IAC9D,iBAAiB,SAAS;AAAA;AAAA,OAGH,cAAa,CAAC,SAAmD;AAAA,IACxF,MAAM,MAAM,cAAc,OAAO;AAAA,IACjC,KAAK,cAAc,KAAK,qBACtB,0BACA,8BAA8B,QAAQ,OAAO,aAAa,6BAA6B,GACvF,OACF;AAAA;AAAA,EAGiB,mBAAmB,CAAC,OAA0C;AAAA,IAC/E,MAAM,SAAU,MAAwC,iBAAiB;AAAA,IACzE,IAAI,UAAU,YAAY,IAAI,MAAM,GAAG;AAAA,MACrC,OAAO,KAAK;AAAA,IACd;AAAA,IACA,OAAO,KAAK;AAAA;AAEhB;;;AC9HA;AAOA,eAAsB,+BAA+B,CACnD,SAGe;AAAA,EACf,MAAM,2BACJ,IAAI,uCACJ,2BACA,OACF;AAAA;;;ACjBK,IAAM,mCAAmC;AAAA,EAC9C,MAAM;AAAA,EACN,MAAM;AAAA,EACN,MAAM;AAAA,EACN,OAAO;AAAA,EACP,IAAI;AAAA,EACJ,IAAI;AAAA,EACJ,OAAO;AAAA,EACP,MAAM;AAAA,EACN,IAAI;AAAA,EACJ,OAAO;AAAA,EACP,IAAI;AAAA,EACJ,OAAO;AACT;AAIA,IAAM,yBACJ,OAAO,QAAQ,gCAAgC,EAE9C,OAAO,IAAI,YAAY,WAAW,EAAE,EACpC,KAAK,CAAC,GAAG,MAAM,EAAE,GAAG,SAAS,EAAE,GAAG,MAAM;AAEpC,SAAS,sBAAsB,CAAC,QAKhB;AAAA,EACrB,MAAM,YAAY,OAAO,aAAa;AAAA,EACtC,MAAM,SAAS,YAAY;AAAA,EAE3B,MAAM,QAAkB,CAAC;AAAA,EACzB,WAAW,MAAM,OAAO,WAAW;AAAA,IACjC,IAAI,CAAC,GAAG,WAAW,MAAM;AAAA,MAAG;AAAA,IAC5B,IAAI,CAAC,GAAG,SAAS,OAAO;AAAA,MAAG;AAAA,IAC3B,IAAI,GAAG,SAAS,YAAY;AAAA,MAAG;AAAA,IAC/B,MAAM,KAAK,GAAG,MAAM,OAAO,QAAQ,CAAC,QAAQ,MAAM,CAAC;AAAA,EACrD;AAAA,EAEA,IAAI,MAAM,WAAW;AAAA,IAAG,OAAO,CAAC;AAAA,EAEhC,MAAM,SAA+D,CAAC;AAAA,EACtE,WAAW,QAAQ,OAAO;AAAA,IACxB,IAAI,UAAU;AAAA,IACd,YAAY,OAAO,WAAW,wBAAwB;AAAA,MACpD,IAAI,KAAK,SAAS,MAAM,GAAG;AAAA,QACzB,OAAO,KAAK,EAAE,UAAU,KAAK,MAAM,GAAG,CAAC,OAAO,MAAM,GAAG,MAAM,CAAC;AAAA,QAC9D,UAAU;AAAA,QACV;AAAA,MACF;AAAA,IACF;AAAA,IACA,IAAI,CAAC,SAAS;AAAA,MACZ,OAAO,KAAK,EAAE,UAAU,MAAM,OAAO,OAAO,CAAC;AAAA,IAC/C;AAAA,EACF;AAAA,EAEA,MAAM,eAAe,IAAI,IAAI,OAAO,IAAI,CAAC,MAAM,EAAE,QAAQ,CAAC;AAAA,EAC1D,MAAM,UAAU,IAAI;AAAA,EACpB,aAAa,UAAU,WAAW,QAAQ;AAAA,IACxC,IAAI,MAAM,QAAQ,IAAI,KAAK;AAAA,IAC3B,IAAI,CAAC,KAAK;AAAA,MACR,MAAM,IAAI;AAAA,MACV,QAAQ,IAAI,OAAO,GAAG;AAAA,IACxB;AAAA,IACA,IAAI,IAAI,QAAQ;AAAA,EAClB;AAAA,EAEA,MAAM,YAAY,OAAO,KAAK,gCAAgC;AAAA,EAC9D,OAAO,UAAU,OAAO,CAAC,UAAU;AAAA,IACjC,MAAM,MAAM,QAAQ,IAAI,KAAK;AAAA,IAC7B,OAAO,QAAQ,aAAa,IAAI,SAAS,aAAa;AAAA,GACvD;AAAA;;;AC9DI,SAAS,0BAA0B,CAAC,MAA8B;AAAA,EACvE,MAAM,WAAW;AAAA,EACjB,MAAM,YAAY;AAAA,EAGlB,IAAI,QAAwB;AAAA,EAE5B,IAAI,UAAU;AAAA,EAEd,SAAS,IAAI,CAAC,OAAe;AAAA,IAC3B,IAAI,UAAU,OAAO;AAAA,MAEnB,WAAW;AAAA,MACX,MAAM,WAAW,QAAQ,QAAQ,SAAS;AAAA,MAC1C,IAAI,aAAa,IAAI;AAAA,QAEnB,MAAM,aAAa,QAAQ,MAAM,WAAW,UAAU,MAAM;AAAA,QAC5D,UAAU;AAAA,QACV,QAAQ;AAAA,QACR,IAAI,WAAW,SAAS,GAAG;AAAA,UACzB,KAAK,UAAU;AAAA,QACjB;AAAA,MACF;AAAA,MAEA;AAAA,IACF;AAAA,IAGA,MAAM,WAAW,UAAU;AAAA,IAG3B,MAAM,UAAU,SAAS,QAAQ,QAAQ;AAAA,IACzC,IAAI,YAAY,IAAI;AAAA,MAElB,MAAM,SAAS,SAAS,MAAM,GAAG,OAAO;AAAA,MACxC,IAAI,OAAO,SAAS,GAAG;AAAA,QACrB,KAAK,MAAM;AAAA,MACb;AAAA,MAEA,UAAU;AAAA,MACV,QAAQ;AAAA,MACR,MAAM,YAAY,SAAS,MAAM,UAAU,SAAS,MAAM;AAAA,MAC1D,IAAI,UAAU,SAAS,GAAG;AAAA,QACxB,KAAK,SAAS;AAAA,MAChB;AAAA,MACA;AAAA,IACF;AAAA,IAIA,IAAI,YAAY;AAAA,IAChB,SAAS,MAAM,KAAK,IAAI,SAAS,QAAQ,SAAS,SAAS,CAAC,EAAG,OAAO,GAAG,OAAO;AAAA,MAC9E,IAAI,SAAS,SAAS,SAAS,MAAM,GAAG,GAAG,CAAC,GAAG;AAAA,QAC7C,YAAY;AAAA,QACZ;AAAA,MACF;AAAA,IACF;AAAA,IAEA,IAAI,YAAY,GAAG;AAAA,MAEjB,MAAM,OAAO,SAAS,MAAM,GAAG,SAAS,SAAS,SAAS;AAAA,MAC1D,IAAI,KAAK,SAAS,GAAG;AAAA,QACnB,KAAK,IAAI;AAAA,MACX;AAAA,MACA,UAAU,SAAS,MAAM,SAAS,SAAS,SAAS;AAAA,IACtD,EAAO;AAAA,MAEL,IAAI,SAAS,SAAS,GAAG;AAAA,QACvB,KAAK,QAAQ;AAAA,MACf;AAAA,MACA,UAAU;AAAA;AAAA;AAAA,EAKd,SAAS,KAAK,GAAG;AAAA,IACf,IAAI,QAAQ,SAAS,KAAK,UAAU,QAAQ;AAAA,MAC1C,KAAK,OAAO;AAAA,MACZ,UAAU;AAAA,IACZ;AAAA,IAEA,UAAU;AAAA,IACV,QAAQ;AAAA;AAAA,EAGV,OAAO,EAAE,MAAM,MAAM;AAAA;",
|
|
15
|
+
"debugId": "158703D87AE0C17864756E2164756E21",
|
|
16
|
+
"names": []
|
|
17
|
+
}
|
package/package.json
ADDED
|
@@ -0,0 +1,60 @@
|
|
|
1
|
+
{
|
|
2
|
+
"name": "@workglow/huggingface-transformers",
|
|
3
|
+
"type": "module",
|
|
4
|
+
"sideEffects": false,
|
|
5
|
+
"version": "0.2.28",
|
|
6
|
+
"repository": {
|
|
7
|
+
"type": "git",
|
|
8
|
+
"url": "https://github.com/workglow-dev/workglow.git",
|
|
9
|
+
"directory": "packages/huggingface-transformers"
|
|
10
|
+
},
|
|
11
|
+
"description": "HuggingFace Transformers.js provider for @workglow/ai-provider.",
|
|
12
|
+
"scripts": {
|
|
13
|
+
"watch": "concurrently -c 'auto' 'bun:watch-*'",
|
|
14
|
+
"watch-code": "bun build --watch --no-clear-screen --sourcemap=external --packages=external --root ./src --outdir ./dist ./src/ai-provider.ts ./src/ai-provider-runtime.ts",
|
|
15
|
+
"watch-types": "tsc --watch --preserveWatchOutput",
|
|
16
|
+
"build-package": "concurrently -c 'auto' -n 'code,types' 'bun run build-code' 'bun run build-types'",
|
|
17
|
+
"build-js": "bun run build-code",
|
|
18
|
+
"build-clean": "rm -fr dist/* tsconfig.tsbuildinfo",
|
|
19
|
+
"build-code": "bun build --sourcemap=external --packages=external --root ./src --outdir ./dist ./src/ai-provider.ts ./src/ai-provider-runtime.ts",
|
|
20
|
+
"build-types": "rm -f tsconfig.tsbuildinfo && tsgo",
|
|
21
|
+
"lint": "eslint . --ext ts,tsx --report-unused-disable-directives --max-warnings 0",
|
|
22
|
+
"test": "bun test"
|
|
23
|
+
},
|
|
24
|
+
"exports": {
|
|
25
|
+
"./ai-provider": {
|
|
26
|
+
"types": "./dist/ai-provider.d.ts",
|
|
27
|
+
"import": "./dist/ai-provider.js"
|
|
28
|
+
},
|
|
29
|
+
"./ai-provider-runtime": {
|
|
30
|
+
"types": "./dist/ai-provider-runtime.d.ts",
|
|
31
|
+
"import": "./dist/ai-provider-runtime.js"
|
|
32
|
+
}
|
|
33
|
+
},
|
|
34
|
+
"dependencies": {
|
|
35
|
+
"@huggingface/transformers": "^4.2.0"
|
|
36
|
+
},
|
|
37
|
+
"peerDependencies": {
|
|
38
|
+
"@workglow/ai": "0.2.28",
|
|
39
|
+
"@workglow/ai-provider": "0.2.28",
|
|
40
|
+
"@workglow/job-queue": "0.2.28",
|
|
41
|
+
"@workglow/storage": "0.2.28",
|
|
42
|
+
"@workglow/task-graph": "0.2.28",
|
|
43
|
+
"@workglow/util": "0.2.28"
|
|
44
|
+
},
|
|
45
|
+
"devDependencies": {
|
|
46
|
+
"@workglow/ai": "0.2.28",
|
|
47
|
+
"@workglow/ai-provider": "0.2.28",
|
|
48
|
+
"@workglow/job-queue": "0.2.28",
|
|
49
|
+
"@workglow/storage": "0.2.28",
|
|
50
|
+
"@workglow/task-graph": "0.2.28",
|
|
51
|
+
"@workglow/util": "0.2.28"
|
|
52
|
+
},
|
|
53
|
+
"files": [
|
|
54
|
+
"dist",
|
|
55
|
+
"src/**/*.md"
|
|
56
|
+
],
|
|
57
|
+
"publishConfig": {
|
|
58
|
+
"access": "public"
|
|
59
|
+
}
|
|
60
|
+
}
|