@tryhamster/gerbil 1.0.0-rc.5 → 1.0.0-rc.7
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/browser/index.js +6 -6
- package/dist/cli.mjs +7 -7
- package/dist/cli.mjs.map +1 -1
- package/dist/frameworks/express.d.mts +1 -1
- package/dist/frameworks/express.mjs +1 -1
- package/dist/frameworks/fastify.d.mts +1 -1
- package/dist/frameworks/fastify.mjs +1 -1
- package/dist/frameworks/hono.d.mts +1 -1
- package/dist/frameworks/hono.mjs +1 -1
- package/dist/frameworks/next.d.mts +2 -2
- package/dist/frameworks/next.mjs +1 -1
- package/dist/frameworks/react.d.mts +1 -1
- package/dist/frameworks/trpc.d.mts +1 -1
- package/dist/frameworks/trpc.mjs +1 -1
- package/dist/gerbil-CJ3ifloF.mjs +4 -0
- package/dist/{gerbil-CocYaJhE.mjs → gerbil-Dw4Qj77e.mjs} +5 -5
- package/dist/{gerbil-CocYaJhE.mjs.map → gerbil-Dw4Qj77e.mjs.map} +1 -1
- package/dist/{gerbil-CnncBh38.d.mts → gerbil-qOTe1nl2.d.mts} +2 -2
- package/dist/{gerbil-CnncBh38.d.mts.map → gerbil-qOTe1nl2.d.mts.map} +1 -1
- package/dist/index.d.mts +2 -2
- package/dist/index.d.mts.map +1 -1
- package/dist/index.mjs +4 -4
- package/dist/integrations/ai-sdk.d.mts +1 -1
- package/dist/integrations/ai-sdk.mjs +3 -3
- package/dist/integrations/langchain.d.mts +1 -1
- package/dist/integrations/langchain.mjs +1 -1
- package/dist/integrations/llamaindex.d.mts +1 -1
- package/dist/integrations/llamaindex.mjs +1 -1
- package/dist/integrations/mcp.d.mts +2 -2
- package/dist/integrations/mcp.mjs +4 -4
- package/dist/kokoro-BNTb6egA.mjs +20210 -0
- package/dist/kokoro-BNTb6egA.mjs.map +1 -0
- package/dist/{kokoro-BSHqCbNW.js → kokoro-CMOGDSgT.js} +2 -2
- package/dist/{kokoro-BSHqCbNW.js.map → kokoro-CMOGDSgT.js.map} +1 -1
- package/dist/{mcp-m9BkEngp.mjs → mcp-BvbriaBy.mjs} +3 -3
- package/dist/{mcp-m9BkEngp.mjs.map → mcp-BvbriaBy.mjs.map} +1 -1
- package/dist/{one-liner-BBbeCywj.mjs → one-liner-s-lD8rCC.mjs} +2 -2
- package/dist/{one-liner-BBbeCywj.mjs.map → one-liner-s-lD8rCC.mjs.map} +1 -1
- package/dist/{repl-DX6JI6fo.mjs → repl-DveXw36T.mjs} +3 -3
- package/dist/skills/index.d.mts +2 -2
- package/dist/skills/index.d.mts.map +1 -1
- package/dist/skills/index.mjs +3 -3
- package/dist/{skills-CBpAWeKy.mjs → skills-CD3Orlex.mjs} +12 -12
- package/dist/skills-CD3Orlex.mjs.map +1 -0
- package/dist/{stt-BN46nKJd.js → stt-Bu-E23Sc.js} +2 -8
- package/dist/stt-Bu-E23Sc.js.map +1 -0
- package/dist/{stt-CkfJswka.mjs → stt-CpLYbGFd.mjs} +2 -8
- package/dist/stt-CpLYbGFd.mjs.map +1 -0
- package/dist/stt-DRPLEEHB.mjs +3 -0
- package/dist/{transformers.web-BschYzqZ.js → transformers.web-DiD1gTwk.js} +14484 -171
- package/dist/transformers.web-DiD1gTwk.js.map +1 -0
- package/dist/transformers.web-u34VxRFM.js +3 -0
- package/dist/{tts-Bn_zwjoM.js → tts-CqroPaSK.js} +3 -13
- package/dist/tts-CqroPaSK.js.map +1 -0
- package/dist/{tts-C8NaF1jo.mjs → tts-DXgsKGCe.mjs} +1 -1
- package/dist/{tts-CmaC0yA9.mjs → tts-DeGANMNV.mjs} +3 -13
- package/dist/tts-DeGANMNV.mjs.map +1 -0
- package/dist/{types-DJhOZ6Ct.d.mts → types-CiTc7ez3.d.mts} +1 -1
- package/dist/{types-DJhOZ6Ct.d.mts.map → types-CiTc7ez3.d.mts.map} +1 -1
- package/package.json +4 -4
- package/dist/gerbil-CjnM4Pra.mjs +0 -4
- package/dist/skills-CBpAWeKy.mjs.map +0 -1
- package/dist/stt-BN46nKJd.js.map +0 -1
- package/dist/stt-BT4Rt49f.mjs +0 -3
- package/dist/stt-CkfJswka.mjs.map +0 -1
- package/dist/transformers.web-BFh4CnBR.js +0 -3
- package/dist/transformers.web-BschYzqZ.js.map +0 -1
- package/dist/tts-Bn_zwjoM.js.map +0 -1
- package/dist/tts-CmaC0yA9.mjs.map +0 -1
package/dist/stt-BN46nKJd.js.map
DELETED
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"file":"stt-BN46nKJd.js","names":["WHISPER_MODELS: STTModelConfig[]","transformers: any","tfDevice: \"webgpu\" | \"cpu\" | \"wasm\"","audioData: Float32Array","pipelineOptions: any","transcribeResult: TranscribeResult","audioBuffer: Float32Array[]","intervalId: ReturnType<typeof setInterval> | null","e: any"],"sources":["../src/core/stt.ts"],"sourcesContent":["/**\n * Speech-to-Text with Whisper\n *\n * Provides local speech recognition using Whisper ONNX models via transformers.js.\n * Supports multiple model sizes and languages.\n *\n * @example\n * ```ts\n * const stt = new WhisperSTT();\n * await stt.load({ onProgress: (p) => console.log(p.status) });\n *\n * // Transcribe audio (Float32Array at 16kHz)\n * const result = await stt.transcribe(audioData);\n * console.log(result.text);\n *\n * // With timestamps\n * const result = await stt.transcribe(audioData, { timestamps: true });\n * for (const seg of result.segments) {\n * console.log(`[${seg.start.toFixed(1)}s] ${seg.text}`);\n * }\n * ```\n */\n\nimport type {\n LoadSTTOptions,\n ProgressInfo,\n STTModelConfig,\n StreamingTranscriptionOptions,\n StreamingTranscriptionSession,\n TranscribeOptions,\n TranscribeResult,\n TranscribeSegment,\n} from \"./types.js\";\n\n// ============================================\n// Model Registry\n// ============================================\n\n/**\n * Available Whisper models\n * Ordered by size (smallest first)\n */\nexport const WHISPER_MODELS: STTModelConfig[] = [\n {\n id: \"whisper-tiny.en\",\n repo: \"onnx-community/whisper-tiny.en\",\n description: \"Tiny English-only model, fastest\",\n size: \"39M\",\n multilingual: false,\n languages: [\"en\"],\n sampleRate: 16000,\n },\n {\n id: \"whisper-tiny\",\n repo: \"onnx-community/whisper-tiny\",\n description: \"Tiny multilingual model\",\n size: \"39M\",\n multilingual: true,\n languages: [\"en\", \"es\", \"fr\", \"de\", \"it\", \"pt\", \"nl\", \"ru\", \"zh\", \"ja\", \"ko\"],\n sampleRate: 16000,\n },\n {\n id: \"whisper-base.en\",\n repo: \"onnx-community/whisper-base.en\",\n description: \"Base English-only model, good balance\",\n size: \"74M\",\n multilingual: false,\n languages: [\"en\"],\n sampleRate: 16000,\n },\n {\n id: \"whisper-base\",\n repo: \"onnx-community/whisper-base\",\n description: \"Base multilingual model\",\n size: \"74M\",\n multilingual: true,\n languages: [\"en\", \"es\", \"fr\", \"de\", \"it\", \"pt\", \"nl\", \"ru\", \"zh\", \"ja\", \"ko\"],\n sampleRate: 16000,\n },\n {\n id: \"whisper-small.en\",\n repo: \"onnx-community/whisper-small.en\",\n description: \"Small English-only model, high quality\",\n size: \"244M\",\n multilingual: false,\n languages: [\"en\"],\n sampleRate: 16000,\n },\n {\n id: \"whisper-small\",\n repo: \"onnx-community/whisper-small\",\n description: \"Small multilingual model\",\n size: \"244M\",\n multilingual: true,\n languages: [\"en\", \"es\", \"fr\", \"de\", \"it\", \"pt\", \"nl\", \"ru\", \"zh\", \"ja\", \"ko\"],\n sampleRate: 16000,\n },\n {\n id: \"whisper-large-v3-turbo\",\n repo: \"onnx-community/whisper-large-v3-turbo\",\n description: \"Large Turbo model, 5.4x faster, 80+ languages\",\n size: \"809M\",\n multilingual: true,\n languages: [\n \"en\",\n \"es\",\n \"fr\",\n \"de\",\n \"it\",\n \"pt\",\n \"nl\",\n \"ru\",\n \"zh\",\n \"ja\",\n \"ko\",\n \"ar\",\n \"hi\",\n \"vi\",\n \"th\",\n ],\n sampleRate: 16000,\n },\n];\n\n// Default model\nconst DEFAULT_MODEL = \"whisper-tiny.en\";\n\n// ============================================\n// Audio Utilities\n// ============================================\n\n/**\n * Decode WAV file to Float32Array\n * Handles stereo to mono conversion\n */\nexport function decodeWav(buffer: Uint8Array): { audio: Float32Array; sampleRate: number } {\n const view = new DataView(buffer.buffer, buffer.byteOffset, buffer.byteLength);\n\n // Validate RIFF header\n const riff = String.fromCharCode(buffer[0], buffer[1], buffer[2], buffer[3]);\n if (riff !== \"RIFF\") {\n throw new Error(\"Invalid WAV file: missing RIFF header\");\n }\n\n // Get format details\n const numChannels = view.getUint16(22, true);\n const sampleRate = view.getUint32(24, true);\n const bitsPerSample = view.getUint16(34, true);\n\n if (bitsPerSample !== 16) {\n throw new Error(`Unsupported bit depth: ${bitsPerSample}. Only 16-bit WAV is supported.`);\n }\n\n // Find data chunk\n let dataOffset = 12;\n while (dataOffset < buffer.length - 8) {\n const chunkId = String.fromCharCode(\n buffer[dataOffset],\n buffer[dataOffset + 1],\n buffer[dataOffset + 2],\n buffer[dataOffset + 3],\n );\n const chunkSize = view.getUint32(dataOffset + 4, true);\n if (chunkId === \"data\") {\n dataOffset += 8;\n break;\n }\n dataOffset += 8 + chunkSize;\n }\n\n const dataSize = buffer.length - dataOffset;\n const bytesPerSample = bitsPerSample / 8;\n const totalSamples = Math.floor(dataSize / bytesPerSample);\n const samplesPerChannel = Math.floor(totalSamples / numChannels);\n\n // Convert to mono Float32Array\n const audio = new Float32Array(samplesPerChannel);\n\n for (let i = 0; i < samplesPerChannel; i++) {\n if (numChannels === 2) {\n const left = view.getInt16(dataOffset + i * 4, true);\n const right = view.getInt16(dataOffset + i * 4 + 2, true);\n audio[i] = (left + right) / 2 / 32768;\n } else {\n const sample = view.getInt16(dataOffset + i * 2, true);\n audio[i] = sample / 32768;\n }\n }\n\n return { audio, sampleRate };\n}\n\n/**\n * Resample audio to target sample rate using linear interpolation\n */\nexport function resampleAudio(audio: Float32Array, fromRate: number, toRate: number): Float32Array {\n if (fromRate === toRate) return audio;\n\n const ratio = toRate / fromRate;\n const newLength = Math.round(audio.length * ratio);\n const result = new Float32Array(newLength);\n\n for (let i = 0; i < newLength; i++) {\n const srcIndex = i / ratio;\n const floor = Math.floor(srcIndex);\n const ceil = Math.min(floor + 1, audio.length - 1);\n const t = srcIndex - floor;\n result[i] = audio[floor] * (1 - t) + audio[ceil] * t;\n }\n\n return result;\n}\n\n// ============================================\n// WhisperSTT Class\n// ============================================\n\n/**\n * Speech-to-Text using Whisper ONNX models\n */\nexport class WhisperSTT {\n private modelConfig: STTModelConfig;\n private pipeline: any = null;\n private loadPromise: Promise<void> | null = null;\n private _isLoaded = false;\n private _deviceMode: \"webgpu\" | \"cpu\" = \"cpu\";\n\n constructor(modelId: string = DEFAULT_MODEL) {\n const config = WHISPER_MODELS.find((m) => m.id === modelId);\n if (!config) {\n const available = WHISPER_MODELS.map((m) => m.id).join(\", \");\n throw new Error(`Unknown STT model: ${modelId}. Available: ${available}`);\n }\n this.modelConfig = config;\n }\n\n /**\n * Check if model is loaded\n */\n isLoaded(): boolean {\n return this._isLoaded;\n }\n\n /**\n * Get model configuration\n */\n getModelConfig(): STTModelConfig {\n return this.modelConfig;\n }\n\n /**\n * Get model info (alias for getModelConfig)\n */\n getModelInfo(): STTModelConfig {\n return this.modelConfig;\n }\n\n /**\n * Get current device mode\n */\n getDeviceMode(): \"webgpu\" | \"cpu\" {\n return this._deviceMode;\n }\n\n /**\n * List available models\n */\n static listModels(): STTModelConfig[] {\n return [...WHISPER_MODELS];\n }\n\n /**\n * Load the STT model\n */\n async load(options: LoadSTTOptions = {}): Promise<void> {\n if (this._isLoaded) return;\n if (this.loadPromise) {\n await this.loadPromise;\n return;\n }\n\n this.loadPromise = this._load(options);\n await this.loadPromise;\n }\n\n private async _load(options: LoadSTTOptions = {}): Promise<void> {\n const { onProgress, device = \"auto\" } = options;\n\n onProgress?.({ status: \"Loading transformers.js...\" });\n\n // Check if we're in Node.js or browser\n const isNode = typeof process !== \"undefined\" && process.versions?.node;\n\n // Import transformers.js dynamically\n // Both paths use webpackIgnore to prevent bundler static analysis\n // Node.js: uses npm package, Browser: uses CDN\n let transformers: any;\n if (isNode) {\n transformers = await import(/* webpackIgnore: true */ \"@huggingface/transformers\");\n } else {\n const cdnUrl = \"https://cdn.jsdelivr.net/npm/@huggingface/transformers@3.8.1\";\n transformers = await import(/* webpackIgnore: true */ cdnUrl);\n }\n const { pipeline, env } = transformers;\n\n // Configure environment based on runtime\n if (isNode) {\n // Node.js: allow local models (for CLI/server use)\n env.allowLocalModels = true;\n env.allowRemoteModels = true;\n } else {\n // Browser: use IndexedDB cache, fetch from HuggingFace CDN\n env.useBrowserCache = true;\n env.allowLocalModels = false;\n }\n\n // Determine device\n // Note: Whisper ONNX models work best with fp32 on CPU/WASM\n // WebGPU support for ASR is limited, so we use CPU for reliability\n let tfDevice: \"webgpu\" | \"cpu\" | \"wasm\" = \"cpu\";\n\n // In browser, use WASM for better compatibility\n if (!isNode) {\n tfDevice = \"wasm\";\n }\n\n // Store device mode\n this._deviceMode = \"cpu\"; // STT always reports as CPU since WASM is CPU-based\n\n onProgress?.({ status: `Loading ${this.modelConfig.id}...` });\n\n // Load the ASR pipeline\n // Always use fp32 for Whisper models (fp16 not available for ONNX ASR)\n this.pipeline = await pipeline(\"automatic-speech-recognition\", this.modelConfig.repo, {\n dtype: \"fp32\",\n device: tfDevice,\n progress_callback: (progress: any) => {\n if (progress.status === \"progress\" && progress.file) {\n onProgress?.({\n status: `Downloading ${progress.file}`,\n progress: Math.round(progress.progress || 0),\n file: progress.file,\n });\n }\n },\n });\n\n this._isLoaded = true;\n onProgress?.({ status: `Ready (${tfDevice.toUpperCase()})!` });\n }\n\n /**\n * Transcribe audio to text\n *\n * @param audio - Audio data as Float32Array (mono, 16kHz preferred) or Uint8Array (WAV file)\n * @param options - Transcription options\n * @returns Transcription result with text and optional timestamps\n */\n async transcribe(\n audio: Float32Array | Uint8Array,\n options: TranscribeOptions = {},\n ): Promise<TranscribeResult> {\n if (!this._isLoaded) {\n throw new Error(\"STT model not loaded. Call load() first.\");\n }\n\n const { language, timestamps = false, onProgress } = options;\n const startTime = performance.now();\n\n // Convert Uint8Array (WAV) to Float32Array\n let audioData: Float32Array;\n let inputSampleRate = 16000;\n\n if (audio instanceof Uint8Array) {\n onProgress?.({ status: \"Decoding audio...\" });\n const decoded = decodeWav(audio);\n audioData = decoded.audio;\n inputSampleRate = decoded.sampleRate;\n } else {\n audioData = audio;\n }\n\n // Resample to 16kHz if needed\n if (inputSampleRate !== 16000) {\n onProgress?.({ status: \"Resampling to 16kHz...\" });\n audioData = resampleAudio(audioData, inputSampleRate, 16000);\n }\n\n const audioDuration = audioData.length / 16000;\n onProgress?.({ status: `Transcribing ${audioDuration.toFixed(1)}s of audio...` });\n\n // Build pipeline options\n const pipelineOptions: any = {};\n\n // Only set language for multilingual models\n if (language && this.modelConfig.multilingual) {\n pipelineOptions.language = language;\n pipelineOptions.task = \"transcribe\";\n }\n\n // Enable timestamps if requested\n if (timestamps) {\n pipelineOptions.return_timestamps = true;\n }\n\n // Run transcription\n const result = await this.pipeline(audioData, pipelineOptions);\n\n const totalTime = performance.now() - startTime;\n\n // Build result\n let text = result.text?.trim() || \"\";\n\n // Filter out Whisper artifacts\n if (text === \"[BLANK_AUDIO]\" || text === \"(blank audio)\" || text === \"[BLANK AUDIO]\") {\n text = \"\";\n }\n\n const transcribeResult: TranscribeResult = {\n text,\n language: language || (this.modelConfig.multilingual ? \"auto\" : \"en\"),\n duration: audioDuration,\n totalTime,\n };\n\n // Add segments if timestamps were requested\n if (timestamps && result.chunks) {\n transcribeResult.segments = result.chunks.map(\n (chunk: any): TranscribeSegment => ({\n text: chunk.text?.trim() || \"\",\n start: chunk.timestamp?.[0] || 0,\n end: chunk.timestamp?.[1] || 0,\n }),\n );\n }\n\n onProgress?.({ status: \"Done!\" });\n\n return transcribeResult;\n }\n\n /**\n * Create a streaming transcription session\n *\n * Transcribes audio in real-time by processing chunks at regular intervals.\n * Perfect for live captioning, call transcription, or real-time subtitles.\n *\n * @param options - Streaming options\n * @returns Streaming session controller\n *\n * @example\n * ```ts\n * const session = stt.createStreamingSession({\n * chunkDuration: 3000, // Transcribe every 3 seconds\n * onChunk: (text, idx) => console.log(`Chunk ${idx}: ${text}`),\n * onTranscript: (fullText) => console.log(\"Full:\", fullText),\n * });\n *\n * // Feed audio data as it comes in (Float32Array at 16kHz)\n * session.feedAudio(audioChunk);\n *\n * // Or manually trigger transcription\n * await session.flush();\n *\n * // Stop and get final transcript\n * const finalText = await session.stop();\n * ```\n */\n createStreamingSession(\n options: StreamingTranscriptionOptions = {},\n ): StreamingTranscriptionSession {\n const {\n chunkDuration = 3000,\n minChunkSize = 8000, // ~0.5 seconds at 16kHz\n onChunk,\n onTranscript,\n onError,\n language,\n } = options;\n\n let audioBuffer: Float32Array[] = [];\n let fullTranscript = \"\";\n let chunkIndex = 0;\n let intervalId: ReturnType<typeof setInterval> | null = null;\n let isRunning = false;\n\n const getBufferSize = (): number => {\n return audioBuffer.reduce((sum, chunk) => sum + chunk.length, 0);\n };\n\n const mergeBuffer = (): Float32Array => {\n const totalLength = getBufferSize();\n const merged = new Float32Array(totalLength);\n let offset = 0;\n for (const chunk of audioBuffer) {\n merged.set(chunk, offset);\n offset += chunk.length;\n }\n return merged;\n };\n\n const transcribeBuffer = async (): Promise<string> => {\n if (!this._isLoaded || getBufferSize() < minChunkSize) {\n return \"\";\n }\n\n const audio = mergeBuffer();\n audioBuffer = []; // Clear buffer\n\n try {\n const result = await this.transcribe(audio, { language });\n const text = result.text.trim();\n\n if (text) {\n chunkIndex++;\n onChunk?.(text, chunkIndex);\n\n // Append to full transcript\n fullTranscript = fullTranscript + (fullTranscript ? \" \" : \"\") + text;\n onTranscript?.(fullTranscript);\n }\n\n return text;\n } catch (e: any) {\n onError?.(e.message || \"Transcription failed\");\n return \"\";\n }\n };\n\n let aborted = false;\n\n const session: StreamingTranscriptionSession = {\n feedAudio: (audio: Float32Array) => {\n if (!aborted) {\n audioBuffer.push(audio);\n }\n },\n\n flush: async () => {\n if (aborted) return \"\";\n return transcribeBuffer();\n },\n\n start: () => {\n if (isRunning || aborted) return;\n isRunning = true;\n\n intervalId = setInterval(async () => {\n if (isRunning && !aborted) {\n await transcribeBuffer();\n }\n }, chunkDuration);\n },\n\n stop: async () => {\n isRunning = false;\n\n if (intervalId) {\n clearInterval(intervalId);\n intervalId = null;\n }\n\n // Transcribe any remaining audio (unless aborted)\n if (!aborted && getBufferSize() >= minChunkSize) {\n await transcribeBuffer();\n }\n\n return fullTranscript;\n },\n\n abort: () => {\n // Immediately stop without final transcription\n aborted = true;\n isRunning = false;\n\n if (intervalId) {\n clearInterval(intervalId);\n intervalId = null;\n }\n\n audioBuffer = [];\n },\n\n isRunning: () => isRunning,\n\n getTranscript: () => fullTranscript,\n\n getChunkCount: () => chunkIndex,\n\n reset: () => {\n audioBuffer = [];\n fullTranscript = \"\";\n chunkIndex = 0;\n },\n };\n\n return session;\n }\n\n /**\n * Dispose of resources\n */\n dispose(): void {\n this.pipeline = null;\n this._isLoaded = false;\n this.loadPromise = null;\n }\n}\n"],"mappings":";;;;;AA0CA,MAAaA,iBAAmC;CAC9C;EACE,IAAI;EACJ,MAAM;EACN,aAAa;EACb,MAAM;EACN,cAAc;EACd,WAAW,CAAC,KAAK;EACjB,YAAY;EACb;CACD;EACE,IAAI;EACJ,MAAM;EACN,aAAa;EACb,MAAM;EACN,cAAc;EACd,WAAW;GAAC;GAAM;GAAM;GAAM;GAAM;GAAM;GAAM;GAAM;GAAM;GAAM;GAAM;GAAK;EAC7E,YAAY;EACb;CACD;EACE,IAAI;EACJ,MAAM;EACN,aAAa;EACb,MAAM;EACN,cAAc;EACd,WAAW,CAAC,KAAK;EACjB,YAAY;EACb;CACD;EACE,IAAI;EACJ,MAAM;EACN,aAAa;EACb,MAAM;EACN,cAAc;EACd,WAAW;GAAC;GAAM;GAAM;GAAM;GAAM;GAAM;GAAM;GAAM;GAAM;GAAM;GAAM;GAAK;EAC7E,YAAY;EACb;CACD;EACE,IAAI;EACJ,MAAM;EACN,aAAa;EACb,MAAM;EACN,cAAc;EACd,WAAW,CAAC,KAAK;EACjB,YAAY;EACb;CACD;EACE,IAAI;EACJ,MAAM;EACN,aAAa;EACb,MAAM;EACN,cAAc;EACd,WAAW;GAAC;GAAM;GAAM;GAAM;GAAM;GAAM;GAAM;GAAM;GAAM;GAAM;GAAM;GAAK;EAC7E,YAAY;EACb;CACD;EACE,IAAI;EACJ,MAAM;EACN,aAAa;EACb,MAAM;EACN,cAAc;EACd,WAAW;GACT;GACA;GACA;GACA;GACA;GACA;GACA;GACA;GACA;GACA;GACA;GACA;GACA;GACA;GACA;GACD;EACD,YAAY;EACb;CACF;AAGD,MAAM,gBAAgB;;;;;AAUtB,SAAgB,UAAU,QAAiE;CACzF,MAAM,OAAO,IAAI,SAAS,OAAO,QAAQ,OAAO,YAAY,OAAO,WAAW;AAI9E,KADa,OAAO,aAAa,OAAO,IAAI,OAAO,IAAI,OAAO,IAAI,OAAO,GAAG,KAC/D,OACX,OAAM,IAAI,MAAM,wCAAwC;CAI1D,MAAM,cAAc,KAAK,UAAU,IAAI,KAAK;CAC5C,MAAM,aAAa,KAAK,UAAU,IAAI,KAAK;CAC3C,MAAM,gBAAgB,KAAK,UAAU,IAAI,KAAK;AAE9C,KAAI,kBAAkB,GACpB,OAAM,IAAI,MAAM,0BAA0B,cAAc,iCAAiC;CAI3F,IAAI,aAAa;AACjB,QAAO,aAAa,OAAO,SAAS,GAAG;EACrC,MAAM,UAAU,OAAO,aACrB,OAAO,aACP,OAAO,aAAa,IACpB,OAAO,aAAa,IACpB,OAAO,aAAa,GACrB;EACD,MAAM,YAAY,KAAK,UAAU,aAAa,GAAG,KAAK;AACtD,MAAI,YAAY,QAAQ;AACtB,iBAAc;AACd;;AAEF,gBAAc,IAAI;;CAGpB,MAAM,WAAW,OAAO,SAAS;CACjC,MAAM,iBAAiB,gBAAgB;CACvC,MAAM,eAAe,KAAK,MAAM,WAAW,eAAe;CAC1D,MAAM,oBAAoB,KAAK,MAAM,eAAe,YAAY;CAGhE,MAAM,QAAQ,IAAI,aAAa,kBAAkB;AAEjD,MAAK,IAAI,IAAI,GAAG,IAAI,mBAAmB,IACrC,KAAI,gBAAgB,EAGlB,OAAM,MAFO,KAAK,SAAS,aAAa,IAAI,GAAG,KAAK,GACtC,KAAK,SAAS,aAAa,IAAI,IAAI,GAAG,KAAK,IAC7B,IAAI;KAGhC,OAAM,KADS,KAAK,SAAS,aAAa,IAAI,GAAG,KAAK,GAClC;AAIxB,QAAO;EAAE;EAAO;EAAY;;;;;AAM9B,SAAgB,cAAc,OAAqB,UAAkB,QAA8B;AACjG,KAAI,aAAa,OAAQ,QAAO;CAEhC,MAAM,QAAQ,SAAS;CACvB,MAAM,YAAY,KAAK,MAAM,MAAM,SAAS,MAAM;CAClD,MAAM,SAAS,IAAI,aAAa,UAAU;AAE1C,MAAK,IAAI,IAAI,GAAG,IAAI,WAAW,KAAK;EAClC,MAAM,WAAW,IAAI;EACrB,MAAM,QAAQ,KAAK,MAAM,SAAS;EAClC,MAAM,OAAO,KAAK,IAAI,QAAQ,GAAG,MAAM,SAAS,EAAE;EAClD,MAAM,IAAI,WAAW;AACrB,SAAO,KAAK,MAAM,UAAU,IAAI,KAAK,MAAM,QAAQ;;AAGrD,QAAO;;;;;AAUT,IAAa,aAAb,MAAwB;CACtB,AAAQ;CACR,AAAQ,WAAgB;CACxB,AAAQ,cAAoC;CAC5C,AAAQ,YAAY;CACpB,AAAQ,cAAgC;CAExC,YAAY,UAAkB,eAAe;EAC3C,MAAM,SAAS,eAAe,MAAM,MAAM,EAAE,OAAO,QAAQ;AAC3D,MAAI,CAAC,QAAQ;GACX,MAAM,YAAY,eAAe,KAAK,MAAM,EAAE,GAAG,CAAC,KAAK,KAAK;AAC5D,SAAM,IAAI,MAAM,sBAAsB,QAAQ,eAAe,YAAY;;AAE3E,OAAK,cAAc;;;;;CAMrB,WAAoB;AAClB,SAAO,KAAK;;;;;CAMd,iBAAiC;AAC/B,SAAO,KAAK;;;;;CAMd,eAA+B;AAC7B,SAAO,KAAK;;;;;CAMd,gBAAkC;AAChC,SAAO,KAAK;;;;;CAMd,OAAO,aAA+B;AACpC,SAAO,CAAC,GAAG,eAAe;;;;;CAM5B,MAAM,KAAK,UAA0B,EAAE,EAAiB;AACtD,MAAI,KAAK,UAAW;AACpB,MAAI,KAAK,aAAa;AACpB,SAAM,KAAK;AACX;;AAGF,OAAK,cAAc,KAAK,MAAM,QAAQ;AACtC,QAAM,KAAK;;CAGb,MAAc,MAAM,UAA0B,EAAE,EAAiB;EAC/D,MAAM,EAAE,YAAY,SAAS,WAAW;AAExC,eAAa,EAAE,QAAQ,8BAA8B,CAAC;EAGtD,MAAM,SAAS,OAAO,YAAY,eAAe,QAAQ,UAAU;EAKnE,IAAIC;AACJ,MAAI,OACF,gBAAe,MAAM;;GAAiC;;MAGtD,gBAAe,MAAM,OADN;EAGjB,MAAM,EAAE,UAAU,QAAQ;AAG1B,MAAI,QAAQ;AAEV,OAAI,mBAAmB;AACvB,OAAI,oBAAoB;SACnB;AAEL,OAAI,kBAAkB;AACtB,OAAI,mBAAmB;;EAMzB,IAAIC,WAAsC;AAG1C,MAAI,CAAC,OACH,YAAW;AAIb,OAAK,cAAc;AAEnB,eAAa,EAAE,QAAQ,WAAW,KAAK,YAAY,GAAG,MAAM,CAAC;AAI7D,OAAK,WAAW,MAAM,SAAS,gCAAgC,KAAK,YAAY,MAAM;GACpF,OAAO;GACP,QAAQ;GACR,oBAAoB,aAAkB;AACpC,QAAI,SAAS,WAAW,cAAc,SAAS,KAC7C,cAAa;KACX,QAAQ,eAAe,SAAS;KAChC,UAAU,KAAK,MAAM,SAAS,YAAY,EAAE;KAC5C,MAAM,SAAS;KAChB,CAAC;;GAGP,CAAC;AAEF,OAAK,YAAY;AACjB,eAAa,EAAE,QAAQ,UAAU,SAAS,aAAa,CAAC,KAAK,CAAC;;;;;;;;;CAUhE,MAAM,WACJ,OACA,UAA6B,EAAE,EACJ;AAC3B,MAAI,CAAC,KAAK,UACR,OAAM,IAAI,MAAM,2CAA2C;EAG7D,MAAM,EAAE,UAAU,aAAa,OAAO,eAAe;EACrD,MAAM,YAAY,YAAY,KAAK;EAGnC,IAAIC;EACJ,IAAI,kBAAkB;AAEtB,MAAI,iBAAiB,YAAY;AAC/B,gBAAa,EAAE,QAAQ,qBAAqB,CAAC;GAC7C,MAAM,UAAU,UAAU,MAAM;AAChC,eAAY,QAAQ;AACpB,qBAAkB,QAAQ;QAE1B,aAAY;AAId,MAAI,oBAAoB,MAAO;AAC7B,gBAAa,EAAE,QAAQ,0BAA0B,CAAC;AAClD,eAAY,cAAc,WAAW,iBAAiB,KAAM;;EAG9D,MAAM,gBAAgB,UAAU,SAAS;AACzC,eAAa,EAAE,QAAQ,gBAAgB,cAAc,QAAQ,EAAE,CAAC,gBAAgB,CAAC;EAGjF,MAAMC,kBAAuB,EAAE;AAG/B,MAAI,YAAY,KAAK,YAAY,cAAc;AAC7C,mBAAgB,WAAW;AAC3B,mBAAgB,OAAO;;AAIzB,MAAI,WACF,iBAAgB,oBAAoB;EAItC,MAAM,SAAS,MAAM,KAAK,SAAS,WAAW,gBAAgB;EAE9D,MAAM,YAAY,YAAY,KAAK,GAAG;EAGtC,IAAI,OAAO,OAAO,MAAM,MAAM,IAAI;AAGlC,MAAI,SAAS,mBAAmB,SAAS,mBAAmB,SAAS,gBACnE,QAAO;EAGT,MAAMC,mBAAqC;GACzC;GACA,UAAU,aAAa,KAAK,YAAY,eAAe,SAAS;GAChE,UAAU;GACV;GACD;AAGD,MAAI,cAAc,OAAO,OACvB,kBAAiB,WAAW,OAAO,OAAO,KACvC,WAAmC;GAClC,MAAM,MAAM,MAAM,MAAM,IAAI;GAC5B,OAAO,MAAM,YAAY,MAAM;GAC/B,KAAK,MAAM,YAAY,MAAM;GAC9B,EACF;AAGH,eAAa,EAAE,QAAQ,SAAS,CAAC;AAEjC,SAAO;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CA8BT,uBACE,UAAyC,EAAE,EACZ;EAC/B,MAAM,EACJ,gBAAgB,KAChB,eAAe,KACf,SACA,cACA,SACA,aACE;EAEJ,IAAIC,cAA8B,EAAE;EACpC,IAAI,iBAAiB;EACrB,IAAI,aAAa;EACjB,IAAIC,aAAoD;EACxD,IAAI,YAAY;EAEhB,MAAM,sBAA8B;AAClC,UAAO,YAAY,QAAQ,KAAK,UAAU,MAAM,MAAM,QAAQ,EAAE;;EAGlE,MAAM,oBAAkC;GACtC,MAAM,cAAc,eAAe;GACnC,MAAM,SAAS,IAAI,aAAa,YAAY;GAC5C,IAAI,SAAS;AACb,QAAK,MAAM,SAAS,aAAa;AAC/B,WAAO,IAAI,OAAO,OAAO;AACzB,cAAU,MAAM;;AAElB,UAAO;;EAGT,MAAM,mBAAmB,YAA6B;AACpD,OAAI,CAAC,KAAK,aAAa,eAAe,GAAG,aACvC,QAAO;GAGT,MAAM,QAAQ,aAAa;AAC3B,iBAAc,EAAE;AAEhB,OAAI;IAEF,MAAM,QADS,MAAM,KAAK,WAAW,OAAO,EAAE,UAAU,CAAC,EACrC,KAAK,MAAM;AAE/B,QAAI,MAAM;AACR;AACA,eAAU,MAAM,WAAW;AAG3B,sBAAiB,kBAAkB,iBAAiB,MAAM,MAAM;AAChE,oBAAe,eAAe;;AAGhC,WAAO;YACAC,GAAQ;AACf,cAAU,EAAE,WAAW,uBAAuB;AAC9C,WAAO;;;EAIX,IAAI,UAAU;AAmEd,SAjE+C;GAC7C,YAAY,UAAwB;AAClC,QAAI,CAAC,QACH,aAAY,KAAK,MAAM;;GAI3B,OAAO,YAAY;AACjB,QAAI,QAAS,QAAO;AACpB,WAAO,kBAAkB;;GAG3B,aAAa;AACX,QAAI,aAAa,QAAS;AAC1B,gBAAY;AAEZ,iBAAa,YAAY,YAAY;AACnC,SAAI,aAAa,CAAC,QAChB,OAAM,kBAAkB;OAEzB,cAAc;;GAGnB,MAAM,YAAY;AAChB,gBAAY;AAEZ,QAAI,YAAY;AACd,mBAAc,WAAW;AACzB,kBAAa;;AAIf,QAAI,CAAC,WAAW,eAAe,IAAI,aACjC,OAAM,kBAAkB;AAG1B,WAAO;;GAGT,aAAa;AAEX,cAAU;AACV,gBAAY;AAEZ,QAAI,YAAY;AACd,mBAAc,WAAW;AACzB,kBAAa;;AAGf,kBAAc,EAAE;;GAGlB,iBAAiB;GAEjB,qBAAqB;GAErB,qBAAqB;GAErB,aAAa;AACX,kBAAc,EAAE;AAChB,qBAAiB;AACjB,iBAAa;;GAEhB;;;;;CAQH,UAAgB;AACd,OAAK,WAAW;AAChB,OAAK,YAAY;AACjB,OAAK,cAAc"}
|
package/dist/stt-BT4Rt49f.mjs
DELETED
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"file":"stt-CkfJswka.mjs","names":["WHISPER_MODELS: STTModelConfig[]","transformers: any","tfDevice: \"webgpu\" | \"cpu\" | \"wasm\"","audioData: Float32Array","pipelineOptions: any","transcribeResult: TranscribeResult","audioBuffer: Float32Array[]","intervalId: ReturnType<typeof setInterval> | null","e: any"],"sources":["../src/core/stt.ts"],"sourcesContent":["/**\n * Speech-to-Text with Whisper\n *\n * Provides local speech recognition using Whisper ONNX models via transformers.js.\n * Supports multiple model sizes and languages.\n *\n * @example\n * ```ts\n * const stt = new WhisperSTT();\n * await stt.load({ onProgress: (p) => console.log(p.status) });\n *\n * // Transcribe audio (Float32Array at 16kHz)\n * const result = await stt.transcribe(audioData);\n * console.log(result.text);\n *\n * // With timestamps\n * const result = await stt.transcribe(audioData, { timestamps: true });\n * for (const seg of result.segments) {\n * console.log(`[${seg.start.toFixed(1)}s] ${seg.text}`);\n * }\n * ```\n */\n\nimport type {\n LoadSTTOptions,\n ProgressInfo,\n STTModelConfig,\n StreamingTranscriptionOptions,\n StreamingTranscriptionSession,\n TranscribeOptions,\n TranscribeResult,\n TranscribeSegment,\n} from \"./types.js\";\n\n// ============================================\n// Model Registry\n// ============================================\n\n/**\n * Available Whisper models\n * Ordered by size (smallest first)\n */\nexport const WHISPER_MODELS: STTModelConfig[] = [\n {\n id: \"whisper-tiny.en\",\n repo: \"onnx-community/whisper-tiny.en\",\n description: \"Tiny English-only model, fastest\",\n size: \"39M\",\n multilingual: false,\n languages: [\"en\"],\n sampleRate: 16000,\n },\n {\n id: \"whisper-tiny\",\n repo: \"onnx-community/whisper-tiny\",\n description: \"Tiny multilingual model\",\n size: \"39M\",\n multilingual: true,\n languages: [\"en\", \"es\", \"fr\", \"de\", \"it\", \"pt\", \"nl\", \"ru\", \"zh\", \"ja\", \"ko\"],\n sampleRate: 16000,\n },\n {\n id: \"whisper-base.en\",\n repo: \"onnx-community/whisper-base.en\",\n description: \"Base English-only model, good balance\",\n size: \"74M\",\n multilingual: false,\n languages: [\"en\"],\n sampleRate: 16000,\n },\n {\n id: \"whisper-base\",\n repo: \"onnx-community/whisper-base\",\n description: \"Base multilingual model\",\n size: \"74M\",\n multilingual: true,\n languages: [\"en\", \"es\", \"fr\", \"de\", \"it\", \"pt\", \"nl\", \"ru\", \"zh\", \"ja\", \"ko\"],\n sampleRate: 16000,\n },\n {\n id: \"whisper-small.en\",\n repo: \"onnx-community/whisper-small.en\",\n description: \"Small English-only model, high quality\",\n size: \"244M\",\n multilingual: false,\n languages: [\"en\"],\n sampleRate: 16000,\n },\n {\n id: \"whisper-small\",\n repo: \"onnx-community/whisper-small\",\n description: \"Small multilingual model\",\n size: \"244M\",\n multilingual: true,\n languages: [\"en\", \"es\", \"fr\", \"de\", \"it\", \"pt\", \"nl\", \"ru\", \"zh\", \"ja\", \"ko\"],\n sampleRate: 16000,\n },\n {\n id: \"whisper-large-v3-turbo\",\n repo: \"onnx-community/whisper-large-v3-turbo\",\n description: \"Large Turbo model, 5.4x faster, 80+ languages\",\n size: \"809M\",\n multilingual: true,\n languages: [\n \"en\",\n \"es\",\n \"fr\",\n \"de\",\n \"it\",\n \"pt\",\n \"nl\",\n \"ru\",\n \"zh\",\n \"ja\",\n \"ko\",\n \"ar\",\n \"hi\",\n \"vi\",\n \"th\",\n ],\n sampleRate: 16000,\n },\n];\n\n// Default model\nconst DEFAULT_MODEL = \"whisper-tiny.en\";\n\n// ============================================\n// Audio Utilities\n// ============================================\n\n/**\n * Decode WAV file to Float32Array\n * Handles stereo to mono conversion\n */\nexport function decodeWav(buffer: Uint8Array): { audio: Float32Array; sampleRate: number } {\n const view = new DataView(buffer.buffer, buffer.byteOffset, buffer.byteLength);\n\n // Validate RIFF header\n const riff = String.fromCharCode(buffer[0], buffer[1], buffer[2], buffer[3]);\n if (riff !== \"RIFF\") {\n throw new Error(\"Invalid WAV file: missing RIFF header\");\n }\n\n // Get format details\n const numChannels = view.getUint16(22, true);\n const sampleRate = view.getUint32(24, true);\n const bitsPerSample = view.getUint16(34, true);\n\n if (bitsPerSample !== 16) {\n throw new Error(`Unsupported bit depth: ${bitsPerSample}. Only 16-bit WAV is supported.`);\n }\n\n // Find data chunk\n let dataOffset = 12;\n while (dataOffset < buffer.length - 8) {\n const chunkId = String.fromCharCode(\n buffer[dataOffset],\n buffer[dataOffset + 1],\n buffer[dataOffset + 2],\n buffer[dataOffset + 3],\n );\n const chunkSize = view.getUint32(dataOffset + 4, true);\n if (chunkId === \"data\") {\n dataOffset += 8;\n break;\n }\n dataOffset += 8 + chunkSize;\n }\n\n const dataSize = buffer.length - dataOffset;\n const bytesPerSample = bitsPerSample / 8;\n const totalSamples = Math.floor(dataSize / bytesPerSample);\n const samplesPerChannel = Math.floor(totalSamples / numChannels);\n\n // Convert to mono Float32Array\n const audio = new Float32Array(samplesPerChannel);\n\n for (let i = 0; i < samplesPerChannel; i++) {\n if (numChannels === 2) {\n const left = view.getInt16(dataOffset + i * 4, true);\n const right = view.getInt16(dataOffset + i * 4 + 2, true);\n audio[i] = (left + right) / 2 / 32768;\n } else {\n const sample = view.getInt16(dataOffset + i * 2, true);\n audio[i] = sample / 32768;\n }\n }\n\n return { audio, sampleRate };\n}\n\n/**\n * Resample audio to target sample rate using linear interpolation\n */\nexport function resampleAudio(audio: Float32Array, fromRate: number, toRate: number): Float32Array {\n if (fromRate === toRate) return audio;\n\n const ratio = toRate / fromRate;\n const newLength = Math.round(audio.length * ratio);\n const result = new Float32Array(newLength);\n\n for (let i = 0; i < newLength; i++) {\n const srcIndex = i / ratio;\n const floor = Math.floor(srcIndex);\n const ceil = Math.min(floor + 1, audio.length - 1);\n const t = srcIndex - floor;\n result[i] = audio[floor] * (1 - t) + audio[ceil] * t;\n }\n\n return result;\n}\n\n// ============================================\n// WhisperSTT Class\n// ============================================\n\n/**\n * Speech-to-Text using Whisper ONNX models\n */\nexport class WhisperSTT {\n private modelConfig: STTModelConfig;\n private pipeline: any = null;\n private loadPromise: Promise<void> | null = null;\n private _isLoaded = false;\n private _deviceMode: \"webgpu\" | \"cpu\" = \"cpu\";\n\n constructor(modelId: string = DEFAULT_MODEL) {\n const config = WHISPER_MODELS.find((m) => m.id === modelId);\n if (!config) {\n const available = WHISPER_MODELS.map((m) => m.id).join(\", \");\n throw new Error(`Unknown STT model: ${modelId}. Available: ${available}`);\n }\n this.modelConfig = config;\n }\n\n /**\n * Check if model is loaded\n */\n isLoaded(): boolean {\n return this._isLoaded;\n }\n\n /**\n * Get model configuration\n */\n getModelConfig(): STTModelConfig {\n return this.modelConfig;\n }\n\n /**\n * Get model info (alias for getModelConfig)\n */\n getModelInfo(): STTModelConfig {\n return this.modelConfig;\n }\n\n /**\n * Get current device mode\n */\n getDeviceMode(): \"webgpu\" | \"cpu\" {\n return this._deviceMode;\n }\n\n /**\n * List available models\n */\n static listModels(): STTModelConfig[] {\n return [...WHISPER_MODELS];\n }\n\n /**\n * Load the STT model\n */\n async load(options: LoadSTTOptions = {}): Promise<void> {\n if (this._isLoaded) return;\n if (this.loadPromise) {\n await this.loadPromise;\n return;\n }\n\n this.loadPromise = this._load(options);\n await this.loadPromise;\n }\n\n private async _load(options: LoadSTTOptions = {}): Promise<void> {\n const { onProgress, device = \"auto\" } = options;\n\n onProgress?.({ status: \"Loading transformers.js...\" });\n\n // Check if we're in Node.js or browser\n const isNode = typeof process !== \"undefined\" && process.versions?.node;\n\n // Import transformers.js dynamically\n // Both paths use webpackIgnore to prevent bundler static analysis\n // Node.js: uses npm package, Browser: uses CDN\n let transformers: any;\n if (isNode) {\n transformers = await import(/* webpackIgnore: true */ \"@huggingface/transformers\");\n } else {\n const cdnUrl = \"https://cdn.jsdelivr.net/npm/@huggingface/transformers@3.8.1\";\n transformers = await import(/* webpackIgnore: true */ cdnUrl);\n }\n const { pipeline, env } = transformers;\n\n // Configure environment based on runtime\n if (isNode) {\n // Node.js: allow local models (for CLI/server use)\n env.allowLocalModels = true;\n env.allowRemoteModels = true;\n } else {\n // Browser: use IndexedDB cache, fetch from HuggingFace CDN\n env.useBrowserCache = true;\n env.allowLocalModels = false;\n }\n\n // Determine device\n // Note: Whisper ONNX models work best with fp32 on CPU/WASM\n // WebGPU support for ASR is limited, so we use CPU for reliability\n let tfDevice: \"webgpu\" | \"cpu\" | \"wasm\" = \"cpu\";\n\n // In browser, use WASM for better compatibility\n if (!isNode) {\n tfDevice = \"wasm\";\n }\n\n // Store device mode\n this._deviceMode = \"cpu\"; // STT always reports as CPU since WASM is CPU-based\n\n onProgress?.({ status: `Loading ${this.modelConfig.id}...` });\n\n // Load the ASR pipeline\n // Always use fp32 for Whisper models (fp16 not available for ONNX ASR)\n this.pipeline = await pipeline(\"automatic-speech-recognition\", this.modelConfig.repo, {\n dtype: \"fp32\",\n device: tfDevice,\n progress_callback: (progress: any) => {\n if (progress.status === \"progress\" && progress.file) {\n onProgress?.({\n status: `Downloading ${progress.file}`,\n progress: Math.round(progress.progress || 0),\n file: progress.file,\n });\n }\n },\n });\n\n this._isLoaded = true;\n onProgress?.({ status: `Ready (${tfDevice.toUpperCase()})!` });\n }\n\n /**\n * Transcribe audio to text\n *\n * @param audio - Audio data as Float32Array (mono, 16kHz preferred) or Uint8Array (WAV file)\n * @param options - Transcription options\n * @returns Transcription result with text and optional timestamps\n */\n async transcribe(\n audio: Float32Array | Uint8Array,\n options: TranscribeOptions = {},\n ): Promise<TranscribeResult> {\n if (!this._isLoaded) {\n throw new Error(\"STT model not loaded. Call load() first.\");\n }\n\n const { language, timestamps = false, onProgress } = options;\n const startTime = performance.now();\n\n // Convert Uint8Array (WAV) to Float32Array\n let audioData: Float32Array;\n let inputSampleRate = 16000;\n\n if (audio instanceof Uint8Array) {\n onProgress?.({ status: \"Decoding audio...\" });\n const decoded = decodeWav(audio);\n audioData = decoded.audio;\n inputSampleRate = decoded.sampleRate;\n } else {\n audioData = audio;\n }\n\n // Resample to 16kHz if needed\n if (inputSampleRate !== 16000) {\n onProgress?.({ status: \"Resampling to 16kHz...\" });\n audioData = resampleAudio(audioData, inputSampleRate, 16000);\n }\n\n const audioDuration = audioData.length / 16000;\n onProgress?.({ status: `Transcribing ${audioDuration.toFixed(1)}s of audio...` });\n\n // Build pipeline options\n const pipelineOptions: any = {};\n\n // Only set language for multilingual models\n if (language && this.modelConfig.multilingual) {\n pipelineOptions.language = language;\n pipelineOptions.task = \"transcribe\";\n }\n\n // Enable timestamps if requested\n if (timestamps) {\n pipelineOptions.return_timestamps = true;\n }\n\n // Run transcription\n const result = await this.pipeline(audioData, pipelineOptions);\n\n const totalTime = performance.now() - startTime;\n\n // Build result\n let text = result.text?.trim() || \"\";\n\n // Filter out Whisper artifacts\n if (text === \"[BLANK_AUDIO]\" || text === \"(blank audio)\" || text === \"[BLANK AUDIO]\") {\n text = \"\";\n }\n\n const transcribeResult: TranscribeResult = {\n text,\n language: language || (this.modelConfig.multilingual ? \"auto\" : \"en\"),\n duration: audioDuration,\n totalTime,\n };\n\n // Add segments if timestamps were requested\n if (timestamps && result.chunks) {\n transcribeResult.segments = result.chunks.map(\n (chunk: any): TranscribeSegment => ({\n text: chunk.text?.trim() || \"\",\n start: chunk.timestamp?.[0] || 0,\n end: chunk.timestamp?.[1] || 0,\n }),\n );\n }\n\n onProgress?.({ status: \"Done!\" });\n\n return transcribeResult;\n }\n\n /**\n * Create a streaming transcription session\n *\n * Transcribes audio in real-time by processing chunks at regular intervals.\n * Perfect for live captioning, call transcription, or real-time subtitles.\n *\n * @param options - Streaming options\n * @returns Streaming session controller\n *\n * @example\n * ```ts\n * const session = stt.createStreamingSession({\n * chunkDuration: 3000, // Transcribe every 3 seconds\n * onChunk: (text, idx) => console.log(`Chunk ${idx}: ${text}`),\n * onTranscript: (fullText) => console.log(\"Full:\", fullText),\n * });\n *\n * // Feed audio data as it comes in (Float32Array at 16kHz)\n * session.feedAudio(audioChunk);\n *\n * // Or manually trigger transcription\n * await session.flush();\n *\n * // Stop and get final transcript\n * const finalText = await session.stop();\n * ```\n */\n createStreamingSession(\n options: StreamingTranscriptionOptions = {},\n ): StreamingTranscriptionSession {\n const {\n chunkDuration = 3000,\n minChunkSize = 8000, // ~0.5 seconds at 16kHz\n onChunk,\n onTranscript,\n onError,\n language,\n } = options;\n\n let audioBuffer: Float32Array[] = [];\n let fullTranscript = \"\";\n let chunkIndex = 0;\n let intervalId: ReturnType<typeof setInterval> | null = null;\n let isRunning = false;\n\n const getBufferSize = (): number => {\n return audioBuffer.reduce((sum, chunk) => sum + chunk.length, 0);\n };\n\n const mergeBuffer = (): Float32Array => {\n const totalLength = getBufferSize();\n const merged = new Float32Array(totalLength);\n let offset = 0;\n for (const chunk of audioBuffer) {\n merged.set(chunk, offset);\n offset += chunk.length;\n }\n return merged;\n };\n\n const transcribeBuffer = async (): Promise<string> => {\n if (!this._isLoaded || getBufferSize() < minChunkSize) {\n return \"\";\n }\n\n const audio = mergeBuffer();\n audioBuffer = []; // Clear buffer\n\n try {\n const result = await this.transcribe(audio, { language });\n const text = result.text.trim();\n\n if (text) {\n chunkIndex++;\n onChunk?.(text, chunkIndex);\n\n // Append to full transcript\n fullTranscript = fullTranscript + (fullTranscript ? \" \" : \"\") + text;\n onTranscript?.(fullTranscript);\n }\n\n return text;\n } catch (e: any) {\n onError?.(e.message || \"Transcription failed\");\n return \"\";\n }\n };\n\n let aborted = false;\n\n const session: StreamingTranscriptionSession = {\n feedAudio: (audio: Float32Array) => {\n if (!aborted) {\n audioBuffer.push(audio);\n }\n },\n\n flush: async () => {\n if (aborted) return \"\";\n return transcribeBuffer();\n },\n\n start: () => {\n if (isRunning || aborted) return;\n isRunning = true;\n\n intervalId = setInterval(async () => {\n if (isRunning && !aborted) {\n await transcribeBuffer();\n }\n }, chunkDuration);\n },\n\n stop: async () => {\n isRunning = false;\n\n if (intervalId) {\n clearInterval(intervalId);\n intervalId = null;\n }\n\n // Transcribe any remaining audio (unless aborted)\n if (!aborted && getBufferSize() >= minChunkSize) {\n await transcribeBuffer();\n }\n\n return fullTranscript;\n },\n\n abort: () => {\n // Immediately stop without final transcription\n aborted = true;\n isRunning = false;\n\n if (intervalId) {\n clearInterval(intervalId);\n intervalId = null;\n }\n\n audioBuffer = [];\n },\n\n isRunning: () => isRunning,\n\n getTranscript: () => fullTranscript,\n\n getChunkCount: () => chunkIndex,\n\n reset: () => {\n audioBuffer = [];\n fullTranscript = \"\";\n chunkIndex = 0;\n },\n };\n\n return session;\n }\n\n /**\n * Dispose of resources\n */\n dispose(): void {\n this.pipeline = null;\n this._isLoaded = false;\n this.loadPromise = null;\n }\n}\n"],"mappings":";;;;;AA0CA,MAAaA,iBAAmC;CAC9C;EACE,IAAI;EACJ,MAAM;EACN,aAAa;EACb,MAAM;EACN,cAAc;EACd,WAAW,CAAC,KAAK;EACjB,YAAY;EACb;CACD;EACE,IAAI;EACJ,MAAM;EACN,aAAa;EACb,MAAM;EACN,cAAc;EACd,WAAW;GAAC;GAAM;GAAM;GAAM;GAAM;GAAM;GAAM;GAAM;GAAM;GAAM;GAAM;GAAK;EAC7E,YAAY;EACb;CACD;EACE,IAAI;EACJ,MAAM;EACN,aAAa;EACb,MAAM;EACN,cAAc;EACd,WAAW,CAAC,KAAK;EACjB,YAAY;EACb;CACD;EACE,IAAI;EACJ,MAAM;EACN,aAAa;EACb,MAAM;EACN,cAAc;EACd,WAAW;GAAC;GAAM;GAAM;GAAM;GAAM;GAAM;GAAM;GAAM;GAAM;GAAM;GAAM;GAAK;EAC7E,YAAY;EACb;CACD;EACE,IAAI;EACJ,MAAM;EACN,aAAa;EACb,MAAM;EACN,cAAc;EACd,WAAW,CAAC,KAAK;EACjB,YAAY;EACb;CACD;EACE,IAAI;EACJ,MAAM;EACN,aAAa;EACb,MAAM;EACN,cAAc;EACd,WAAW;GAAC;GAAM;GAAM;GAAM;GAAM;GAAM;GAAM;GAAM;GAAM;GAAM;GAAM;GAAK;EAC7E,YAAY;EACb;CACD;EACE,IAAI;EACJ,MAAM;EACN,aAAa;EACb,MAAM;EACN,cAAc;EACd,WAAW;GACT;GACA;GACA;GACA;GACA;GACA;GACA;GACA;GACA;GACA;GACA;GACA;GACA;GACA;GACA;GACD;EACD,YAAY;EACb;CACF;AAGD,MAAM,gBAAgB;;;;;AAUtB,SAAgB,UAAU,QAAiE;CACzF,MAAM,OAAO,IAAI,SAAS,OAAO,QAAQ,OAAO,YAAY,OAAO,WAAW;AAI9E,KADa,OAAO,aAAa,OAAO,IAAI,OAAO,IAAI,OAAO,IAAI,OAAO,GAAG,KAC/D,OACX,OAAM,IAAI,MAAM,wCAAwC;CAI1D,MAAM,cAAc,KAAK,UAAU,IAAI,KAAK;CAC5C,MAAM,aAAa,KAAK,UAAU,IAAI,KAAK;CAC3C,MAAM,gBAAgB,KAAK,UAAU,IAAI,KAAK;AAE9C,KAAI,kBAAkB,GACpB,OAAM,IAAI,MAAM,0BAA0B,cAAc,iCAAiC;CAI3F,IAAI,aAAa;AACjB,QAAO,aAAa,OAAO,SAAS,GAAG;EACrC,MAAM,UAAU,OAAO,aACrB,OAAO,aACP,OAAO,aAAa,IACpB,OAAO,aAAa,IACpB,OAAO,aAAa,GACrB;EACD,MAAM,YAAY,KAAK,UAAU,aAAa,GAAG,KAAK;AACtD,MAAI,YAAY,QAAQ;AACtB,iBAAc;AACd;;AAEF,gBAAc,IAAI;;CAGpB,MAAM,WAAW,OAAO,SAAS;CACjC,MAAM,iBAAiB,gBAAgB;CACvC,MAAM,eAAe,KAAK,MAAM,WAAW,eAAe;CAC1D,MAAM,oBAAoB,KAAK,MAAM,eAAe,YAAY;CAGhE,MAAM,QAAQ,IAAI,aAAa,kBAAkB;AAEjD,MAAK,IAAI,IAAI,GAAG,IAAI,mBAAmB,IACrC,KAAI,gBAAgB,EAGlB,OAAM,MAFO,KAAK,SAAS,aAAa,IAAI,GAAG,KAAK,GACtC,KAAK,SAAS,aAAa,IAAI,IAAI,GAAG,KAAK,IAC7B,IAAI;KAGhC,OAAM,KADS,KAAK,SAAS,aAAa,IAAI,GAAG,KAAK,GAClC;AAIxB,QAAO;EAAE;EAAO;EAAY;;;;;AAM9B,SAAgB,cAAc,OAAqB,UAAkB,QAA8B;AACjG,KAAI,aAAa,OAAQ,QAAO;CAEhC,MAAM,QAAQ,SAAS;CACvB,MAAM,YAAY,KAAK,MAAM,MAAM,SAAS,MAAM;CAClD,MAAM,SAAS,IAAI,aAAa,UAAU;AAE1C,MAAK,IAAI,IAAI,GAAG,IAAI,WAAW,KAAK;EAClC,MAAM,WAAW,IAAI;EACrB,MAAM,QAAQ,KAAK,MAAM,SAAS;EAClC,MAAM,OAAO,KAAK,IAAI,QAAQ,GAAG,MAAM,SAAS,EAAE;EAClD,MAAM,IAAI,WAAW;AACrB,SAAO,KAAK,MAAM,UAAU,IAAI,KAAK,MAAM,QAAQ;;AAGrD,QAAO;;;;;AAUT,IAAa,aAAb,MAAwB;CACtB,AAAQ;CACR,AAAQ,WAAgB;CACxB,AAAQ,cAAoC;CAC5C,AAAQ,YAAY;CACpB,AAAQ,cAAgC;CAExC,YAAY,UAAkB,eAAe;EAC3C,MAAM,SAAS,eAAe,MAAM,MAAM,EAAE,OAAO,QAAQ;AAC3D,MAAI,CAAC,QAAQ;GACX,MAAM,YAAY,eAAe,KAAK,MAAM,EAAE,GAAG,CAAC,KAAK,KAAK;AAC5D,SAAM,IAAI,MAAM,sBAAsB,QAAQ,eAAe,YAAY;;AAE3E,OAAK,cAAc;;;;;CAMrB,WAAoB;AAClB,SAAO,KAAK;;;;;CAMd,iBAAiC;AAC/B,SAAO,KAAK;;;;;CAMd,eAA+B;AAC7B,SAAO,KAAK;;;;;CAMd,gBAAkC;AAChC,SAAO,KAAK;;;;;CAMd,OAAO,aAA+B;AACpC,SAAO,CAAC,GAAG,eAAe;;;;;CAM5B,MAAM,KAAK,UAA0B,EAAE,EAAiB;AACtD,MAAI,KAAK,UAAW;AACpB,MAAI,KAAK,aAAa;AACpB,SAAM,KAAK;AACX;;AAGF,OAAK,cAAc,KAAK,MAAM,QAAQ;AACtC,QAAM,KAAK;;CAGb,MAAc,MAAM,UAA0B,EAAE,EAAiB;EAC/D,MAAM,EAAE,YAAY,SAAS,WAAW;AAExC,eAAa,EAAE,QAAQ,8BAA8B,CAAC;EAGtD,MAAM,SAAS,OAAO,YAAY,eAAe,QAAQ,UAAU;EAKnE,IAAIC;AACJ,MAAI,OACF,gBAAe,MAAM;;GAAiC;;MAGtD,gBAAe,MAAM,OADN;EAGjB,MAAM,EAAE,UAAU,QAAQ;AAG1B,MAAI,QAAQ;AAEV,OAAI,mBAAmB;AACvB,OAAI,oBAAoB;SACnB;AAEL,OAAI,kBAAkB;AACtB,OAAI,mBAAmB;;EAMzB,IAAIC,WAAsC;AAG1C,MAAI,CAAC,OACH,YAAW;AAIb,OAAK,cAAc;AAEnB,eAAa,EAAE,QAAQ,WAAW,KAAK,YAAY,GAAG,MAAM,CAAC;AAI7D,OAAK,WAAW,MAAM,SAAS,gCAAgC,KAAK,YAAY,MAAM;GACpF,OAAO;GACP,QAAQ;GACR,oBAAoB,aAAkB;AACpC,QAAI,SAAS,WAAW,cAAc,SAAS,KAC7C,cAAa;KACX,QAAQ,eAAe,SAAS;KAChC,UAAU,KAAK,MAAM,SAAS,YAAY,EAAE;KAC5C,MAAM,SAAS;KAChB,CAAC;;GAGP,CAAC;AAEF,OAAK,YAAY;AACjB,eAAa,EAAE,QAAQ,UAAU,SAAS,aAAa,CAAC,KAAK,CAAC;;;;;;;;;CAUhE,MAAM,WACJ,OACA,UAA6B,EAAE,EACJ;AAC3B,MAAI,CAAC,KAAK,UACR,OAAM,IAAI,MAAM,2CAA2C;EAG7D,MAAM,EAAE,UAAU,aAAa,OAAO,eAAe;EACrD,MAAM,YAAY,YAAY,KAAK;EAGnC,IAAIC;EACJ,IAAI,kBAAkB;AAEtB,MAAI,iBAAiB,YAAY;AAC/B,gBAAa,EAAE,QAAQ,qBAAqB,CAAC;GAC7C,MAAM,UAAU,UAAU,MAAM;AAChC,eAAY,QAAQ;AACpB,qBAAkB,QAAQ;QAE1B,aAAY;AAId,MAAI,oBAAoB,MAAO;AAC7B,gBAAa,EAAE,QAAQ,0BAA0B,CAAC;AAClD,eAAY,cAAc,WAAW,iBAAiB,KAAM;;EAG9D,MAAM,gBAAgB,UAAU,SAAS;AACzC,eAAa,EAAE,QAAQ,gBAAgB,cAAc,QAAQ,EAAE,CAAC,gBAAgB,CAAC;EAGjF,MAAMC,kBAAuB,EAAE;AAG/B,MAAI,YAAY,KAAK,YAAY,cAAc;AAC7C,mBAAgB,WAAW;AAC3B,mBAAgB,OAAO;;AAIzB,MAAI,WACF,iBAAgB,oBAAoB;EAItC,MAAM,SAAS,MAAM,KAAK,SAAS,WAAW,gBAAgB;EAE9D,MAAM,YAAY,YAAY,KAAK,GAAG;EAGtC,IAAI,OAAO,OAAO,MAAM,MAAM,IAAI;AAGlC,MAAI,SAAS,mBAAmB,SAAS,mBAAmB,SAAS,gBACnE,QAAO;EAGT,MAAMC,mBAAqC;GACzC;GACA,UAAU,aAAa,KAAK,YAAY,eAAe,SAAS;GAChE,UAAU;GACV;GACD;AAGD,MAAI,cAAc,OAAO,OACvB,kBAAiB,WAAW,OAAO,OAAO,KACvC,WAAmC;GAClC,MAAM,MAAM,MAAM,MAAM,IAAI;GAC5B,OAAO,MAAM,YAAY,MAAM;GAC/B,KAAK,MAAM,YAAY,MAAM;GAC9B,EACF;AAGH,eAAa,EAAE,QAAQ,SAAS,CAAC;AAEjC,SAAO;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CA8BT,uBACE,UAAyC,EAAE,EACZ;EAC/B,MAAM,EACJ,gBAAgB,KAChB,eAAe,KACf,SACA,cACA,SACA,aACE;EAEJ,IAAIC,cAA8B,EAAE;EACpC,IAAI,iBAAiB;EACrB,IAAI,aAAa;EACjB,IAAIC,aAAoD;EACxD,IAAI,YAAY;EAEhB,MAAM,sBAA8B;AAClC,UAAO,YAAY,QAAQ,KAAK,UAAU,MAAM,MAAM,QAAQ,EAAE;;EAGlE,MAAM,oBAAkC;GACtC,MAAM,cAAc,eAAe;GACnC,MAAM,SAAS,IAAI,aAAa,YAAY;GAC5C,IAAI,SAAS;AACb,QAAK,MAAM,SAAS,aAAa;AAC/B,WAAO,IAAI,OAAO,OAAO;AACzB,cAAU,MAAM;;AAElB,UAAO;;EAGT,MAAM,mBAAmB,YAA6B;AACpD,OAAI,CAAC,KAAK,aAAa,eAAe,GAAG,aACvC,QAAO;GAGT,MAAM,QAAQ,aAAa;AAC3B,iBAAc,EAAE;AAEhB,OAAI;IAEF,MAAM,QADS,MAAM,KAAK,WAAW,OAAO,EAAE,UAAU,CAAC,EACrC,KAAK,MAAM;AAE/B,QAAI,MAAM;AACR;AACA,eAAU,MAAM,WAAW;AAG3B,sBAAiB,kBAAkB,iBAAiB,MAAM,MAAM;AAChE,oBAAe,eAAe;;AAGhC,WAAO;YACAC,GAAQ;AACf,cAAU,EAAE,WAAW,uBAAuB;AAC9C,WAAO;;;EAIX,IAAI,UAAU;AAmEd,SAjE+C;GAC7C,YAAY,UAAwB;AAClC,QAAI,CAAC,QACH,aAAY,KAAK,MAAM;;GAI3B,OAAO,YAAY;AACjB,QAAI,QAAS,QAAO;AACpB,WAAO,kBAAkB;;GAG3B,aAAa;AACX,QAAI,aAAa,QAAS;AAC1B,gBAAY;AAEZ,iBAAa,YAAY,YAAY;AACnC,SAAI,aAAa,CAAC,QAChB,OAAM,kBAAkB;OAEzB,cAAc;;GAGnB,MAAM,YAAY;AAChB,gBAAY;AAEZ,QAAI,YAAY;AACd,mBAAc,WAAW;AACzB,kBAAa;;AAIf,QAAI,CAAC,WAAW,eAAe,IAAI,aACjC,OAAM,kBAAkB;AAG1B,WAAO;;GAGT,aAAa;AAEX,cAAU;AACV,gBAAY;AAEZ,QAAI,YAAY;AACd,mBAAc,WAAW;AACzB,kBAAa;;AAGf,kBAAc,EAAE;;GAGlB,iBAAiB;GAEjB,qBAAqB;GAErB,qBAAqB;GAErB,aAAa;AACX,kBAAc,EAAE;AAChB,qBAAiB;AACjB,iBAAa;;GAEhB;;;;;CAQH,UAAgB;AACd,OAAK,WAAW;AAChB,OAAK,YAAY;AACjB,OAAK,cAAc"}
|
|
@@ -1,3 +0,0 @@
|
|
|
1
|
-
import { $ as __webpack_exports__BartForConditionalGeneration, $a as __webpack_exports__ImageClassificationPipeline, $c as __webpack_exports__NeoBertForQuestionAnswering, $d as __webpack_exports__SnacPreTrainedModel, $f as __webpack_exports__TokenizerModel, $i as __webpack_exports__GPTNeoModel, $l as __webpack_exports__PhiForCausalLM, $m as __webpack_exports__full_like, $n as __webpack_exports__DebertaForSequenceClassification, $o as __webpack_exports__MPNetForTokenClassification, $p as __webpack_exports__Wav2Vec2ForSequenceClassification, $r as __webpack_exports__EfficientNetPreTrainedModel, $s as __webpack_exports__MobileNetV2FeatureExtractor, $t as __webpack_exports__ChineseCLIPPreTrainedModel, $u as __webpack_exports__RoFormerForSequenceClassification, A as __webpack_exports__AutoModelForImageTextToText, Aa as __webpack_exports__GroundingDinoImageProcessor, Ac as __webpack_exports__ModernBertForSequenceClassification, Ad as __webpack_exports__SegformerForImageClassification, Af as __webpack_exports__Swin2SRModel, Ah as __webpack_exports__slice, Ai as __webpack_exports__FeatureExtractionPipeline, Al as __webpack_exports__Owlv2ForObjectDetection, Am as __webpack_exports__XLMRobertaModel, An as __webpack_exports__DFineForObjectDetection, Ao as __webpack_exports__LlavaPreTrainedModel, Ap as __webpack_exports__ViTModel, Ar as __webpack_exports__DetrSegmentationOutput, As as __webpack_exports__MimiEncoderOutput, At as __webpack_exports__CLIPFeatureExtractor, Au as __webpack_exports__RFDetrForObjectDetection, B as __webpack_exports__AutoModelForSequenceClassification, Ba as __webpack_exports__HieraModel, Bc as __webpack_exports__MptForCausalLM, Bd as __webpack_exports__SiglipTextModel, Bf as __webpack_exports__TableTransformerForObjectDetection, Bi as __webpack_exports__GLPNModel, Bl as __webpack_exports__PatchTSMixerForPrediction, Bm as __webpack_exports__YolosObjectDetectionOutput, Bn as __webpack_exports__DPTImageProcessor, Bo as __webpack_exports__M2M100Model, Bp as __webpack_exports__VitsModelOutput, Br as __webpack_exports__DistilBertForTokenClassification, Bs as __webpack_exports__MobileBertForSequenceClassification, Bt as __webpack_exports__CLIPVisionModel, Bu as __webpack_exports__RTDetrV2Model, C as __webpack_exports__AutoModelForCausalLM, Ca as __webpack_exports__GraniteModel, Cc as __webpack_exports__MobileViTV2Model, Cd as __webpack_exports__SamProcessor, Cf as __webpack_exports__StyleTextToSpeech2PreTrainedModel, Ch as __webpack_exports__pipeline, Ci as __webpack_exports__FalconForCausalLM, Cl as __webpack_exports__OpenELMPreTrainedModel, Cm as __webpack_exports__XLMForTokenClassification, Cn as __webpack_exports__ConvNextForImageClassification, Co as __webpack_exports__LlamaForCausalLM, Cp as __webpack_exports__ViTForImageClassification, Cr as __webpack_exports__DetrFeatureExtractor, Cs as __webpack_exports__MgpstrModelOutput, Ct as __webpack_exports__BlenderbotSmallPreTrainedModel, Cu as __webpack_exports__Qwen2VLForConditionalGeneration, D as __webpack_exports__AutoModelForImageFeatureExtraction, Da as __webpack_exports__GranitePreTrainedModel, Dc as __webpack_exports__ModernBertDecoderModel, Dd as __webpack_exports__SapiensPreTrainedModel, Df as __webpack_exports__SuppressTokensAtBeginLogitsProcessor, Dh as __webpack_exports__read_audio, Di as __webpack_exports__FastViTForImageClassification, Dl as __webpack_exports__OwlViTModel, Dm as __webpack_exports__XLMRobertaForQuestionAnswering, Dn as __webpack_exports__ConvNextV2ForImageClassification, Do as __webpack_exports__LlavaForConditionalGeneration, Dp as __webpack_exports__ViTMSNForImageClassification, Dr as __webpack_exports__DetrModel, Ds as __webpack_exports__MimiDecoderModel, Dt as __webpack_exports__BloomModel, Du as __webpack_exports__Qwen3ForCausalLM, E as __webpack_exports__AutoModelForImageClassification, Ea as __webpack_exports__GraniteMoeHybridPreTrainedModel, Ec as __webpack_exports__ModernBertDecoderForCausalLM, Ed as __webpack_exports__SapiensForSemanticSegmentation, Ef as __webpack_exports__SupertonicPreTrainedModel, Eh as __webpack_exports__randn, Ei as __webpack_exports__FalconTokenizer, El as __webpack_exports__OwlViTImageProcessor, Em as __webpack_exports__XLMRobertaForMaskedLM, En as __webpack_exports__ConvNextPreTrainedModel, Eo as __webpack_exports__LlamaTokenizer, Ep as __webpack_exports__ViTMAEPreTrainedModel, Er as __webpack_exports__DetrImageProcessor, Es as __webpack_exports__MgpstrTokenizer, Et as __webpack_exports__BloomForCausalLM, Eu as __webpack_exports__Qwen2VLProcessor, F as __webpack_exports__AutoModelForObjectDetection, Fa as __webpack_exports__HeliumForCausalLM, Fc as __webpack_exports__MoonshineFeatureExtractor, Fd as __webpack_exports__Seq2SeqLMOutput, Ff as __webpack_exports__SwinPreTrainedModel, Fh as __webpack_exports__topk, Fi as __webpack_exports__Florence2Processor, Fl as __webpack_exports__PaliGemmaPreTrainedModel, Fm as __webpack_exports__XVectorOutput, Fn as __webpack_exports__DINOv3ViTImageProcessor, Fo as __webpack_exports__LogitsWarper, Fp as __webpack_exports__VitMattePreTrainedModel, Fr as __webpack_exports__Dinov2WithRegistersModel, Fs as __webpack_exports__MistralForCausalLM, Ft as __webpack_exports__CLIPSegModel, Fu as __webpack_exports__RTDetrImageProcessor, G as __webpack_exports__AutoModelForUniversalSegmentation, Ga as __webpack_exports__HubertPreTrainedModel, Gc as __webpack_exports__MusicgenForCausalLM, Gd as __webpack_exports__SmolLM3PreTrainedModel, Gf as __webpack_exports__Tensor, Gi as __webpack_exports__GPT2Tokenizer, Gl as __webpack_exports__PatchTSTPreTrainedModel, Gm as __webpack_exports__ZeroShotObjectDetectionPipeline, Gn as __webpack_exports__DacEncoderModel, Go as __webpack_exports__MBartForConditionalGeneration, Gp as __webpack_exports__Wav2Vec2BertForCTC, Gr as __webpack_exports__DonutFeatureExtractor, Gs as __webpack_exports__MobileLLMModel, Gt as __webpack_exports__CamembertForTokenClassification, Gu as __webpack_exports__RawVideo, H as __webpack_exports__AutoModelForTextToSpectrogram, Ha as __webpack_exports__HubertForCTC, Hc as __webpack_exports__MptPreTrainedModel, Hd as __webpack_exports__SiglipVisionModel, Hf as __webpack_exports__TableTransformerObjectDetectionOutput, Hi as __webpack_exports__GPT2LMHeadModel, Hl as __webpack_exports__PatchTSMixerPreTrainedModel, Hm as __webpack_exports__ZeroShotAudioClassificationPipeline, Hn as __webpack_exports__DPTPreTrainedModel, Ho as __webpack_exports__M2M100Tokenizer, Hp as __webpack_exports__VitsTokenizer, Hr as __webpack_exports__DistilBertPreTrainedModel, Hs as __webpack_exports__MobileBertPreTrainedModel, Ht as __webpack_exports__CamembertForMaskedLM, Hu as __webpack_exports__RTDetrV2PreTrainedModel, I as __webpack_exports__AutoModelForPoseEstimation, Ia as __webpack_exports__HeliumModel, Ic as __webpack_exports__MoonshineForConditionalGeneration, Id as __webpack_exports__SequenceClassifierOutput, If as __webpack_exports__T5ForConditionalGeneration, Ih as __webpack_exports__window_function, Ii as __webpack_exports__ForcedBOSTokenLogitsProcessor, Il as __webpack_exports__PaliGemmaProcessor, Im as __webpack_exports__YolosFeatureExtractor, In as __webpack_exports__DINOv3ViTModel, Io as __webpack_exports__LongT5ForConditionalGeneration, Ip as __webpack_exports__VitPoseForPoseEstimation, Ir as __webpack_exports__Dinov2WithRegistersPreTrainedModel, Is as __webpack_exports__MistralModel, It as __webpack_exports__CLIPSegPreTrainedModel, Iu as __webpack_exports__RTDetrModel, J as __webpack_exports__AutoModelForZeroShotObjectDetection, Ja as __webpack_exports__IJepaPreTrainedModel, Jc as __webpack_exports__MusicgenPreTrainedModel, Jd as __webpack_exports__SmolVLMProcessor, Jf as __webpack_exports__TextGenerationPipeline, Ji as __webpack_exports__GPTBigCodePreTrainedModel, Jl as __webpack_exports__Phi3PreTrainedModel, Jm as __webpack_exports__cos_sim, Jn as __webpack_exports__DacModel, Jo as __webpack_exports__MBartPreTrainedModel, Jp as __webpack_exports__Wav2Vec2BertPreTrainedModel, Jr as __webpack_exports__DonutSwinPreTrainedModel, Js as __webpack_exports__MobileNetV1ForImageClassification, Jt as __webpack_exports__CamembertTokenizer, Ju as __webpack_exports__ResNetForImageClassification, K as __webpack_exports__AutoModelForVision2Seq, Ka as __webpack_exports__IJepaForImageClassification, Kc as __webpack_exports__MusicgenForConditionalGeneration, Kd as __webpack_exports__SmolVLMForConditionalGeneration, Kf as __webpack_exports__Text2TextGenerationPipeline, Ki as __webpack_exports__GPTBigCodeForCausalLM, Kl as __webpack_exports__Phi3ForCausalLM, Km as __webpack_exports__bankers_round, Kn as __webpack_exports__DacEncoderOutput, Ko as __webpack_exports__MBartForSequenceClassification, Kp as __webpack_exports__Wav2Vec2BertForSequenceClassification, Kr as __webpack_exports__DonutImageProcessor, Ks as __webpack_exports__MobileLLMPreTrainedModel, Kt as __webpack_exports__CamembertModel, Ku as __webpack_exports__RawVideoFrame, L as __webpack_exports__AutoModelForQuestionAnswering, La as __webpack_exports__HeliumPreTrainedModel, Lc as __webpack_exports__MoonshineModel, Ld as __webpack_exports__SiglipImageProcessor, Lf as __webpack_exports__T5Model, Lh as __webpack_exports__zeros, Li as __webpack_exports__ForcedEOSTokenLogitsProcessor, Ll as __webpack_exports__ParakeetFeatureExtractor, Lm as __webpack_exports__YolosForObjectDetection, Ln as __webpack_exports__DINOv3ViTPreTrainedModel, Lo as __webpack_exports__LongT5Model, Lp as __webpack_exports__VitPoseImageProcessor, Lr as __webpack_exports__DistilBertForMaskedLM, Ls as __webpack_exports__MistralPreTrainedModel, Lt as __webpack_exports__CLIPTextModel, Lu as __webpack_exports__RTDetrObjectDetectionOutput, M as __webpack_exports__AutoModelForMaskGeneration, Ma as __webpack_exports__GroundingDinoProcessor, Mc as __webpack_exports__ModernBertModel, Md as __webpack_exports__SegformerImageProcessor, Mf as __webpack_exports__SwinForImageClassification, Mh as __webpack_exports__spectrogram, Mi as __webpack_exports__FillMaskPipeline, Ml as __webpack_exports__Owlv2Model, Mm as __webpack_exports__XLMRobertaTokenizer, Mn as __webpack_exports__DFinePreTrainedModel, Mo as __webpack_exports__LlavaQwen2ForCausalLM, Mp as __webpack_exports__VisionEncoderDecoderModel, Mr as __webpack_exports__Dinov2Model, Ms as __webpack_exports__MimiPreTrainedModel, Mt as __webpack_exports__CLIPModel, Mu as __webpack_exports__RFDetrObjectDetectionOutput, N as __webpack_exports__AutoModelForMaskedLM, Na as __webpack_exports__GroupViTModel, Nc as __webpack_exports__ModernBertPreTrainedModel, Nd as __webpack_exports__SegformerModel, Nf as __webpack_exports__SwinForSemanticSegmentation, Nh as __webpack_exports__stack, Ni as __webpack_exports__Florence2ForConditionalGeneration, Nl as __webpack_exports__Owlv2PreTrainedModel, Nm as __webpack_exports__XLMTokenizer, Nn as __webpack_exports__DINOv3ConvNextModel, No as __webpack_exports__LogitsProcessor, Np as __webpack_exports__VitMatteForImageMatting, Nr as __webpack_exports__Dinov2PreTrainedModel, Ns as __webpack_exports__MinLengthLogitsProcessor, Nt as __webpack_exports__CLIPPreTrainedModel, Nu as __webpack_exports__RFDetrPreTrainedModel, O as __webpack_exports__AutoModelForImageMatting, Oa as __webpack_exports__Grok1Tokenizer, Oc as __webpack_exports__ModernBertDecoderPreTrainedModel, Od as __webpack_exports__SeamlessM4TFeatureExtractor, Of as __webpack_exports__Swin2SRForImageSuperResolution, Oh as __webpack_exports__rfft, Oi as __webpack_exports__FastViTModel, Ol as __webpack_exports__OwlViTPreTrainedModel, Om as __webpack_exports__XLMRobertaForSequenceClassification, On as __webpack_exports__ConvNextV2Model, Oo as __webpack_exports__LlavaOnevisionForConditionalGeneration, Op as __webpack_exports__ViTMSNModel, Or as __webpack_exports__DetrObjectDetectionOutput, Os as __webpack_exports__MimiDecoderOutput, Ot as __webpack_exports__BloomPreTrainedModel, Ou as __webpack_exports__Qwen3Model, P as __webpack_exports__AutoModelForNormalEstimation, Pa as __webpack_exports__GroupViTPreTrainedModel, Pc as __webpack_exports__Moondream1ForConditionalGeneration, Pd as __webpack_exports__SegformerPreTrainedModel, Pf as __webpack_exports__SwinModel, Ph as __webpack_exports__std_mean, Pi as __webpack_exports__Florence2PreTrainedModel, Pl as __webpack_exports__PaliGemmaForConditionalGeneration, Pm as __webpack_exports__XLMWithLMHeadModel, Pn as __webpack_exports__DINOv3ConvNextPreTrainedModel, Po as __webpack_exports__LogitsProcessorList, Pp as __webpack_exports__VitMatteImageProcessor, Pr as __webpack_exports__Dinov2WithRegistersForImageClassification, Ps as __webpack_exports__MinNewTokensLengthLogitsProcessor, Pt as __webpack_exports__CLIPSegForImageSegmentation, Pu as __webpack_exports__RTDetrForObjectDetection, Q as __webpack_exports__BackgroundRemovalPipeline, Qa as __webpack_exports__Idefics3Processor, Qc as __webpack_exports__NeoBertForMaskedLM, Qd as __webpack_exports__SnacModel, Qf as __webpack_exports__TokenClassifierOutput, Qi as __webpack_exports__GPTNeoForCausalLM, Ql as __webpack_exports__Phi3VProcessor, Qm as __webpack_exports__full, Qn as __webpack_exports__DebertaForQuestionAnswering, Qo as __webpack_exports__MPNetForSequenceClassification, Qp as __webpack_exports__Wav2Vec2ForCTC, Qr as __webpack_exports__EfficientNetModel, Qs as __webpack_exports__MobileNetV1PreTrainedModel, Qt as __webpack_exports__ChineseCLIPModel, Qu as __webpack_exports__RoFormerForQuestionAnswering, R as __webpack_exports__AutoModelForSemanticSegmentation, Ra as __webpack_exports__HerbertTokenizer, Rc as __webpack_exports__MoonshinePreTrainedModel, Rd as __webpack_exports__SiglipModel, Rf as __webpack_exports__T5PreTrainedModel, Rh as __webpack_exports__zeros_like, Ri as __webpack_exports__GLPNFeatureExtractor, Rl as __webpack_exports__ParakeetForCTC, Rm as __webpack_exports__YolosImageProcessor, Rn as __webpack_exports__DPTFeatureExtractor, Ro as __webpack_exports__LongT5PreTrainedModel, Rp as __webpack_exports__VitPosePreTrainedModel, Rr as __webpack_exports__DistilBertForQuestionAnswering, Rs as __webpack_exports__MobileBertForMaskedLM, Rt as __webpack_exports__CLIPTextModelWithProjection, Ru as __webpack_exports__RTDetrPreTrainedModel, S as __webpack_exports__AutoModelForCTC, Sa as __webpack_exports__GraniteForCausalLM, Sc as __webpack_exports__MobileViTV2ForImageClassification, Sd as __webpack_exports__SamPreTrainedModel, Sf as __webpack_exports__StyleTextToSpeech2Model, Sh as __webpack_exports__permute_data, Si as __webpack_exports__FFT, Sl as __webpack_exports__OpenELMModel, Sm as __webpack_exports__XLMForSequenceClassification, Sn as __webpack_exports__ConvNextFeatureExtractor, So as __webpack_exports__Llama4PreTrainedModel, Sp as __webpack_exports__ViTFeatureExtractor, Sr as __webpack_exports__DepthProPreTrainedModel, Ss as __webpack_exports__MgpstrForSceneTextRecognition, St as __webpack_exports__BlenderbotSmallModel, Su as __webpack_exports__Qwen2Tokenizer, T as __webpack_exports__AutoModelForDocumentQuestionAnswering, Ta as __webpack_exports__GraniteMoeHybridModel, Tc as __webpack_exports__ModelOutput, Td as __webpack_exports__SapiensForNormalEstimation, Tf as __webpack_exports__SupertonicForConditionalGeneration, Th as __webpack_exports__rand, Ti as __webpack_exports__FalconPreTrainedModel, Tl as __webpack_exports__OwlViTForObjectDetection, Tm as __webpack_exports__XLMPreTrainedModel, Tn as __webpack_exports__ConvNextModel, To as __webpack_exports__LlamaPreTrainedModel, Tp as __webpack_exports__ViTMAEModel, Tr as __webpack_exports__DetrForSegmentation, Ts as __webpack_exports__MgpstrProcessor, Tt as __webpack_exports__BlenderbotTokenizer, Tu as __webpack_exports__Qwen2VLPreTrainedModel, U as __webpack_exports__AutoModelForTextToWaveform, Ua as __webpack_exports__HubertForSequenceClassification, Uc as __webpack_exports__MultiModalityCausalLM, Ud as __webpack_exports__SmolLM3ForCausalLM, Uf as __webpack_exports__TableTransformerPreTrainedModel, Ui as __webpack_exports__GPT2Model, Ul as __webpack_exports__PatchTSTForPrediction, Um as __webpack_exports__ZeroShotClassificationPipeline, Un as __webpack_exports__DacDecoderModel, Uo as __webpack_exports__MBart50Tokenizer, Up as __webpack_exports__VoxtralForConditionalGeneration, Ur as __webpack_exports__DistilBertTokenizer, Us as __webpack_exports__MobileBertTokenizer, Ut as __webpack_exports__CamembertForQuestionAnswering, Uu as __webpack_exports__RawAudio, V as __webpack_exports__AutoModelForSpeechSeq2Seq, Va as __webpack_exports__HieraPreTrainedModel, Vc as __webpack_exports__MptModel, Vd as __webpack_exports__SiglipTokenizer, Vf as __webpack_exports__TableTransformerModel, Vi as __webpack_exports__GLPNPreTrainedModel, Vl as __webpack_exports__PatchTSMixerModel, Vm as __webpack_exports__YolosPreTrainedModel, Vn as __webpack_exports__DPTModel, Vo as __webpack_exports__M2M100PreTrainedModel, Vp as __webpack_exports__VitsPreTrainedModel, Vr as __webpack_exports__DistilBertModel, Vs as __webpack_exports__MobileBertModel, Vt as __webpack_exports__CLIPVisionModelWithProjection, Vu as __webpack_exports__RTDetrV2ObjectDetectionOutput, W as __webpack_exports__AutoModelForTokenClassification, Wa as __webpack_exports__HubertModel, Wc as __webpack_exports__MultiModalityPreTrainedModel, Wd as __webpack_exports__SmolLM3Model, Wf as __webpack_exports__TemperatureLogitsWarper, Wi as __webpack_exports__GPT2PreTrainedModel, Wl as __webpack_exports__PatchTSTModel, Wm as __webpack_exports__ZeroShotImageClassificationPipeline, Wn as __webpack_exports__DacDecoderOutput, Wo as __webpack_exports__MBartForCausalLM, Wp as __webpack_exports__VoxtralProcessor, Wr as __webpack_exports__DocumentQuestionAnsweringPipeline, Ws as __webpack_exports__MobileLLMForCausalLM, Wt as __webpack_exports__CamembertForSequenceClassification, Wu as __webpack_exports__RawImage, X as __webpack_exports__AutoTokenizer, Xa as __webpack_exports__Idefics3ImageProcessor, Xc as __webpack_exports__NanoChatModel, Xd as __webpack_exports__SnacEncoderModel, Xf as __webpack_exports__TextToAudioPipeline, Xi as __webpack_exports__GPTJModel, Xl as __webpack_exports__Phi3VImageProcessor, Xm as __webpack_exports__dynamic_time_warping, Xn as __webpack_exports__DataTypeMap, Xo as __webpack_exports__MPNetForMaskedLM, Xp as __webpack_exports__Wav2Vec2FeatureExtractor, Xr as __webpack_exports__EfficientNetForImageClassification, Xs as __webpack_exports__MobileNetV1ImageProcessor, Xt as __webpack_exports__CausalLMOutputWithPast, Xu as __webpack_exports__ResNetPreTrainedModel, Y as __webpack_exports__AutoProcessor, Ya as __webpack_exports__Idefics3ForConditionalGeneration, Yc as __webpack_exports__NanoChatForCausalLM, Yd as __webpack_exports__SnacDecoderModel, Yf as __webpack_exports__TextStreamer, Yi as __webpack_exports__GPTJForCausalLM, Yl as __webpack_exports__Phi3VForCausalLM, Ym as __webpack_exports__dot, Yn as __webpack_exports__DacPreTrainedModel, Yo as __webpack_exports__MBartTokenizer, Yp as __webpack_exports__Wav2Vec2CTCTokenizer, Yr as __webpack_exports__EdgeTamModel, Ys as __webpack_exports__MobileNetV1ForSemanticSegmentation, Yt as __webpack_exports__CausalLMOutput, Yu as __webpack_exports__ResNetModel, Z as __webpack_exports__AutomaticSpeechRecognitionPipeline, Za as __webpack_exports__Idefics3PreTrainedModel, Zc as __webpack_exports__NanoChatPreTrainedModel, Zd as __webpack_exports__SnacFeatureExtractor, Zf as __webpack_exports__TokenClassificationPipeline, Zi as __webpack_exports__GPTJPreTrainedModel, Zl as __webpack_exports__Phi3VPreTrainedModel, Zm as __webpack_exports__env, Zn as __webpack_exports__DebertaForMaskedLM, Zo as __webpack_exports__MPNetForQuestionAnswering, Zp as __webpack_exports__Wav2Vec2ForAudioFrameClassification, Zr as __webpack_exports__EfficientNetImageProcessor, Zs as __webpack_exports__MobileNetV1Model, Zt as __webpack_exports__ChineseCLIPFeatureExtractor, Zu as __webpack_exports__RoFormerForMaskedLM, _ as __webpack_exports__AutoImageProcessor, _a as __webpack_exports__GemmaPreTrainedModel, _c as __webpack_exports__MobileViTFeatureExtractor, _d as __webpack_exports__Sam3ImageProcessor, _f as __webpack_exports__Starcoder2ForCausalLM, _h as __webpack_exports__mel_filter_bank, _i as __webpack_exports__EsmPreTrainedModel, _l as __webpack_exports__Olmo2PreTrainedModel, _m as __webpack_exports__WhisperProcessor, _n as __webpack_exports__ConvBertForSequenceClassification, _o as __webpack_exports__Lfm2ForCausalLM, _p as __webpack_exports__VLChatProcessor, _r as __webpack_exports__DeiTPreTrainedModel, _s as __webpack_exports__MaxLengthCriteria, _t as __webpack_exports__BitImageProcessor, _u as __webpack_exports__QuestionAnsweringModelOutput, a as __webpack_exports__AlbertForMaskedLM, aa as __webpack_exports__Gemma2ForCausalLM, ac as __webpack_exports__MobileNetV3FeatureExtractor, ad as __webpack_exports__RobertaForQuestionAnswering, af as __webpack_exports__SpeechT5PreTrainedModel, ah as __webpack_exports__interpolate_data, ai as __webpack_exports__ElectraPreTrainedModel, al as __webpack_exports__NoBadWordsLogitsProcessor, am as __webpack_exports__WavLMForCTC, an as __webpack_exports__ClassifierFreeGuidanceLogitsProcessor, ao as __webpack_exports__ImageToImagePipeline, ap as __webpack_exports__UltravoxModel, ar as __webpack_exports__DebertaV2ForQuestionAnswering, as as __webpack_exports__MT5PreTrainedModel, at as __webpack_exports__BaseStreamer, au as __webpack_exports__PretrainedConfig, b as __webpack_exports__AutoModelForAudioFrameClassification, ba as __webpack_exports__GlmModel, bc as __webpack_exports__MobileViTModel, bd as __webpack_exports__SamImageSegmentationOutput, bf as __webpack_exports__StoppingCriteria, bh as __webpack_exports__ones_like, bi as __webpack_exports__ExaoneModel, bl as __webpack_exports__OlmoPreTrainedModel, bm as __webpack_exports__WhisperTokenizer, bn as __webpack_exports__ConvBertPreTrainedModel, bo as __webpack_exports__LiteWhisperForConditionalGeneration, bp as __webpack_exports__VaultGemmaModel, br as __webpack_exports__DepthEstimationPipeline, bs as __webpack_exports__Metric3Dv2ForDepthEstimation, bt as __webpack_exports__BlenderbotPreTrainedModel, bu as __webpack_exports__Qwen2Model, c as __webpack_exports__AlbertModel, ca as __webpack_exports__Gemma3ForCausalLM, cc as __webpack_exports__MobileNetV3ImageProcessor, cd as __webpack_exports__RobertaModel, cf as __webpack_exports__SqueezeBertForMaskedLM, ch as __webpack_exports__load_image, ci as __webpack_exports__EosTokenCriteria, cl as __webpack_exports__NomicBertPreTrainedModel, cm as __webpack_exports__WavLMModel, cn as __webpack_exports__CodeGenPreTrainedModel, co as __webpack_exports__JAISLMHeadModel, cp as __webpack_exports__UniSpeechForCTC, cr as __webpack_exports__DebertaV2Model, cs as __webpack_exports__MarianPreTrainedModel, ct as __webpack_exports__BeitModel, cu as __webpack_exports__PvtForImageClassification, d as __webpack_exports__ArceeForCausalLM, da as __webpack_exports__Gemma3nAudioFeatureExtractor, dc as __webpack_exports__MobileNetV4FeatureExtractor, dd as __webpack_exports__Sam2ImageProcessor, df as __webpack_exports__SqueezeBertModel, dh as __webpack_exports__magnitude, di as __webpack_exports__Ernie4_5_PretrainedModel, dl as __webpack_exports__OPTForCausalLM, dm as __webpack_exports__WeSpeakerResNetModel, dn as __webpack_exports__CohereForCausalLM, do as __webpack_exports__JinaCLIPImageProcessor, dp as __webpack_exports__UniSpeechPreTrainedModel, dr as __webpack_exports__DecisionTransformerModel, ds as __webpack_exports__MaskFormerFeatureExtractor, dt as __webpack_exports__BertForQuestionAnswering, du as __webpack_exports__PvtPreTrainedModel, ea as __webpack_exports__GPTNeoPreTrainedModel, ec as __webpack_exports__MobileNetV2ForImageClassification, ed as __webpack_exports__RoFormerForTokenClassification, ef as __webpack_exports__SpeechT5FeatureExtractor, eh as __webpack_exports__getCacheShapes, ei as __webpack_exports__ElectraForMaskedLM, el as __webpack_exports__NeoBertForSequenceClassification, em as __webpack_exports__Wav2Vec2Model, en as __webpack_exports__ClapAudioModelWithProjection, eo as __webpack_exports__ImageFeatureExtractionPipeline, ep as __webpack_exports__TopKLogitsWarper, er as __webpack_exports__DebertaForTokenClassification, es as __webpack_exports__MPNetModel, et as __webpack_exports__BartForSequenceClassification, eu as __webpack_exports__PhiModel, f as __webpack_exports__ArceeModel, fa as __webpack_exports__Gemma3nForConditionalGeneration, fc as __webpack_exports__MobileNetV4ForImageClassification, fd as __webpack_exports__Sam2ImageSegmentationOutput, ff as __webpack_exports__SqueezeBertPreTrainedModel, fh as __webpack_exports__matmul, fi as __webpack_exports__Ernie4_5_Tokenizer, fl as __webpack_exports__OPTModel, fm as __webpack_exports__WeSpeakerResNetPreTrainedModel, fn as __webpack_exports__CohereModel, fo as __webpack_exports__JinaCLIPModel, fp as __webpack_exports__UniSpeechSatForAudioFrameClassification, fr as __webpack_exports__DecisionTransformerPreTrainedModel, fs as __webpack_exports__MaskFormerForInstanceSegmentation, ft as __webpack_exports__BertForSequenceClassification, fu as __webpack_exports__PyAnnoteFeatureExtractor, g as __webpack_exports__AutoFeatureExtractor, ga as __webpack_exports__GemmaModel, gc as __webpack_exports__MobileNetV4PreTrainedModel, gd as __webpack_exports__Sam2VideoProcessor, gf as __webpack_exports__StableLmPreTrainedModel, gh as __webpack_exports__medianFilter, gi as __webpack_exports__EsmModel, gl as __webpack_exports__Olmo2Model, gm as __webpack_exports__WhisperPreTrainedModel, gn as __webpack_exports__ConvBertForQuestionAnswering, go as __webpack_exports__JinaCLIPVisionModel, gp as __webpack_exports__UniSpeechSatPreTrainedModel, gr as __webpack_exports__DeiTModel, gs as __webpack_exports__MaskedLMOutput, gt as __webpack_exports__BertTokenizer, gu as __webpack_exports__PyAnnoteProcessor, h as __webpack_exports__AutoConfig, ha as __webpack_exports__GemmaForCausalLM, hc as __webpack_exports__MobileNetV4Model, hd as __webpack_exports__Sam2Processor, hf as __webpack_exports__StableLmModel, hh as __webpack_exports__mean_pooling, hi as __webpack_exports__EsmForTokenClassification, hl as __webpack_exports__Olmo2ForCausalLM, hm as __webpack_exports__WhisperModel, hn as __webpack_exports__ConvBertForMaskedLM, ho as __webpack_exports__JinaCLIPTextModel, hp as __webpack_exports__UniSpeechSatModel, hr as __webpack_exports__DeiTImageProcessor, hs as __webpack_exports__MaskFormerPreTrainedModel, ht as __webpack_exports__BertPreTrainedModel, hu as __webpack_exports__PyAnnotePreTrainedModel, i as __webpack_exports__ASTPreTrainedModel, ia as __webpack_exports__GPTNeoXTokenizer, ic as __webpack_exports__MobileNetV2PreTrainedModel, id as __webpack_exports__RobertaForMaskedLM, if as __webpack_exports__SpeechT5Model, ih as __webpack_exports__interpolate_4d, ii as __webpack_exports__ElectraModel, il as __webpack_exports__NllbTokenizer, im as __webpack_exports__WavLMForAudioFrameClassification, in as __webpack_exports__ClapTextModelWithProjection, io as __webpack_exports__ImageSegmentationPipeline, ip as __webpack_exports__TranslationPipeline, ir as __webpack_exports__DebertaV2ForMaskedLM, is as __webpack_exports__MT5Model, it as __webpack_exports__BaseModelOutput, iu as __webpack_exports__PreTrainedTokenizer, j as __webpack_exports__AutoModelForImageToImage, ja as __webpack_exports__GroundingDinoPreTrainedModel, jc as __webpack_exports__ModernBertForTokenClassification, jd as __webpack_exports__SegformerForSemanticSegmentation, jf as __webpack_exports__Swin2SRPreTrainedModel, jh as __webpack_exports__softmax, ji as __webpack_exports__FeatureExtractor, jl as __webpack_exports__Owlv2ImageProcessor, jm as __webpack_exports__XLMRobertaPreTrainedModel, jn as __webpack_exports__DFineModel, jo as __webpack_exports__LlavaProcessor, jp as __webpack_exports__ViTPreTrainedModel, jr as __webpack_exports__Dinov2ForImageClassification, js as __webpack_exports__MimiModel, jt as __webpack_exports__CLIPImageProcessor, ju as __webpack_exports__RFDetrModel, k as __webpack_exports__AutoModelForImageSegmentation, ka as __webpack_exports__GroundingDinoForObjectDetection, kc as __webpack_exports__ModernBertForMaskedLM, kd as __webpack_exports__SegformerFeatureExtractor, kf as __webpack_exports__Swin2SRImageProcessor, kh as __webpack_exports__round, ki as __webpack_exports__FastViTPreTrainedModel, kl as __webpack_exports__OwlViTProcessor, km as __webpack_exports__XLMRobertaForTokenClassification, kn as __webpack_exports__ConvNextV2PreTrainedModel, ko as __webpack_exports__LlavaOnevisionImageProcessor, kp as __webpack_exports__ViTMSNPreTrainedModel, kr as __webpack_exports__DetrPreTrainedModel, ks as __webpack_exports__MimiEncoderModel, kt as __webpack_exports__BloomTokenizer, ku as __webpack_exports__Qwen3PreTrainedModel, l as __webpack_exports__AlbertPreTrainedModel, la as __webpack_exports__Gemma3Model, lc as __webpack_exports__MobileNetV3Model, ld as __webpack_exports__RobertaPreTrainedModel, lf as __webpack_exports__SqueezeBertForQuestionAnswering, lh as __webpack_exports__load_video, li as __webpack_exports__Ernie4_5_ForCausalLM, ll as __webpack_exports__NougatImageProcessor, lm as __webpack_exports__WavLMPreTrainedModel, ln as __webpack_exports__CodeGenTokenizer, lo as __webpack_exports__JAISModel, lp as __webpack_exports__UniSpeechForSequenceClassification, lr as __webpack_exports__DebertaV2PreTrainedModel, ls as __webpack_exports__MarianTokenizer, lt as __webpack_exports__BeitPreTrainedModel, lu as __webpack_exports__PvtImageProcessor, m as __webpack_exports__AudioClassificationPipeline, ma as __webpack_exports__Gemma3nProcessor, mc as __webpack_exports__MobileNetV4ImageProcessor, md as __webpack_exports__Sam2PreTrainedModel, mf as __webpack_exports__StableLmForCausalLM, mh as __webpack_exports__mean, mi as __webpack_exports__EsmForSequenceClassification, ml as __webpack_exports__ObjectDetectionPipeline, mm as __webpack_exports__WhisperForConditionalGeneration, mn as __webpack_exports__CohereTokenizer, mo as __webpack_exports__JinaCLIPProcessor, mp as __webpack_exports__UniSpeechSatForSequenceClassification, mr as __webpack_exports__DeiTForImageClassification, ms as __webpack_exports__MaskFormerModel, mt as __webpack_exports__BertModel, mu as __webpack_exports__PyAnnoteModel, n as __webpack_exports__ASTForAudioClassification, na as __webpack_exports__GPTNeoXModel, nc as __webpack_exports__MobileNetV2ImageProcessor, nd as __webpack_exports__RoFormerPreTrainedModel, nf as __webpack_exports__SpeechT5ForTextToSpeech, nh as __webpack_exports__hanning, ni as __webpack_exports__ElectraForSequenceClassification, nl as __webpack_exports__NeoBertModel, nm as __webpack_exports__Wav2Vec2Processor, nn as __webpack_exports__ClapModel, no as __webpack_exports__ImageMattingOutput, np as __webpack_exports__TrOCRForCausalLM, nr as __webpack_exports__DebertaPreTrainedModel, ns as __webpack_exports__MPNetTokenizer, nt as __webpack_exports__BartPretrainedModel, nu as __webpack_exports__Pipeline, o as __webpack_exports__AlbertForQuestionAnswering, oa as __webpack_exports__Gemma2Model, oc as __webpack_exports__MobileNetV3ForImageClassification, od as __webpack_exports__RobertaForSequenceClassification, of as __webpack_exports__SpeechT5Processor, oh as __webpack_exports__is_chinese_char, oi as __webpack_exports__ElectraTokenizer, ol as __webpack_exports__NoRepeatNGramLogitsProcessor, om as __webpack_exports__WavLMForSequenceClassification, on as __webpack_exports__CodeGenForCausalLM, oo as __webpack_exports__ImageToTextPipeline, op as __webpack_exports__UltravoxPreTrainedModel, or as __webpack_exports__DebertaV2ForSequenceClassification, os as __webpack_exports__MarianMTModel, ot as __webpack_exports__BeitFeatureExtractor, ou as __webpack_exports__PretrainedMixin, p as __webpack_exports__ArceePreTrainedModel, pa as __webpack_exports__Gemma3nPreTrainedModel, pc as __webpack_exports__MobileNetV4ForSemanticSegmentation, pd as __webpack_exports__Sam2Model, pf as __webpack_exports__SqueezeBertTokenizer, ph as __webpack_exports__max, pi as __webpack_exports__EsmForMaskedLM, pl as __webpack_exports__OPTPreTrainedModel, pm as __webpack_exports__WhisperFeatureExtractor, pn as __webpack_exports__CoherePreTrainedModel, po as __webpack_exports__JinaCLIPPreTrainedModel, pp as __webpack_exports__UniSpeechSatForCTC, pr as __webpack_exports__DeiTFeatureExtractor, ps as __webpack_exports__MaskFormerImageProcessor, pt as __webpack_exports__BertForTokenClassification, pu as __webpack_exports__PyAnnoteForAudioFrameClassification, q as __webpack_exports__AutoModelForXVector, qa as __webpack_exports__IJepaModel, qc as __webpack_exports__MusicgenModel, qd as __webpack_exports__SmolVLMImageProcessor, qf as __webpack_exports__TextClassificationPipeline, qi as __webpack_exports__GPTBigCodeModel, ql as __webpack_exports__Phi3Model, qm as __webpack_exports__cat, qn as __webpack_exports__DacFeatureExtractor, qo as __webpack_exports__MBartModel, qp as __webpack_exports__Wav2Vec2BertModel, qr as __webpack_exports__DonutSwinModel, qs as __webpack_exports__MobileNetV1FeatureExtractor, qt as __webpack_exports__CamembertPreTrainedModel, qu as __webpack_exports__RepetitionPenaltyLogitsProcessor, r as __webpack_exports__ASTModel, ra as __webpack_exports__GPTNeoXPreTrainedModel, rc as __webpack_exports__MobileNetV2Model, rd as __webpack_exports__RoFormerTokenizer, rf as __webpack_exports__SpeechT5HifiGan, rh as __webpack_exports__interpolate, ri as __webpack_exports__ElectraForTokenClassification, rl as __webpack_exports__NeoBertPreTrainedModel, rm as __webpack_exports__Wav2Vec2ProcessorWithLM, rn as __webpack_exports__ClapPreTrainedModel, ro as __webpack_exports__ImageProcessor, rp as __webpack_exports__TrOCRPreTrainedModel, rr as __webpack_exports__DebertaTokenizer, rs as __webpack_exports__MT5ForConditionalGeneration, rt as __webpack_exports__BartTokenizer, ru as __webpack_exports__PreTrainedModel, s as __webpack_exports__AlbertForSequenceClassification, sa as __webpack_exports__Gemma2PreTrainedModel, sc as __webpack_exports__MobileNetV3ForSemanticSegmentation, sd as __webpack_exports__RobertaForTokenClassification, sf as __webpack_exports__SpeechT5Tokenizer, sh as __webpack_exports__layer_norm, si as __webpack_exports__EncodecFeatureExtractor, sl as __webpack_exports__NomicBertModel, sm as __webpack_exports__WavLMForXVector, sn as __webpack_exports__CodeGenModel, so as __webpack_exports__InterruptableStoppingCriteria, sp as __webpack_exports__UltravoxProcessor, sr as __webpack_exports__DebertaV2ForTokenClassification, ss as __webpack_exports__MarianModel, st as __webpack_exports__BeitForImageClassification, su as __webpack_exports__Processor, t as __webpack_exports__ASTFeatureExtractor, ta as __webpack_exports__GPTNeoXForCausalLM, tc as __webpack_exports__MobileNetV2ForSemanticSegmentation, td as __webpack_exports__RoFormerModel, tf as __webpack_exports__SpeechT5ForSpeechToText, th as __webpack_exports__hamming, ti as __webpack_exports__ElectraForQuestionAnswering, tl as __webpack_exports__NeoBertForTokenClassification, tm as __webpack_exports__Wav2Vec2PreTrainedModel, tn as __webpack_exports__ClapFeatureExtractor, to as __webpack_exports__ImageFeatureExtractor, tp as __webpack_exports__TopPLogitsWarper, tr as __webpack_exports__DebertaModel, ts as __webpack_exports__MPNetPreTrainedModel, tt as __webpack_exports__BartModel, tu as __webpack_exports__PhiPreTrainedModel, u as __webpack_exports__AlbertTokenizer, ua as __webpack_exports__Gemma3PreTrainedModel, uc as __webpack_exports__MobileNetV3PreTrainedModel, ud as __webpack_exports__RobertaTokenizer, uf as __webpack_exports__SqueezeBertForSequenceClassification, uh as __webpack_exports__log_softmax, ui as __webpack_exports__Ernie4_5_Model, ul as __webpack_exports__NougatTokenizer, um as __webpack_exports__WeSpeakerFeatureExtractor, un as __webpack_exports__CodeLlamaTokenizer, uo as __webpack_exports__JAISPreTrainedModel, up as __webpack_exports__UniSpeechModel, ur as __webpack_exports__DebertaV2Tokenizer, us as __webpack_exports__Mask2FormerImageProcessor, ut as __webpack_exports__BertForMaskedLM, uu as __webpack_exports__PvtModel, v as __webpack_exports__AutoModel, va as __webpack_exports__GemmaTokenizer, vc as __webpack_exports__MobileViTForImageClassification, vd as __webpack_exports__Sam3TrackerModel, vf as __webpack_exports__Starcoder2Model, vh as __webpack_exports__min, vi as __webpack_exports__EsmTokenizer, vl as __webpack_exports__OlmoForCausalLM, vm as __webpack_exports__WhisperTextStreamer, vn as __webpack_exports__ConvBertForTokenClassification, vo as __webpack_exports__Lfm2Model, vp as __webpack_exports__VLMImageProcessor, vr as __webpack_exports__DepthAnythingForDepthEstimation, vs as __webpack_exports__Metric3DForDepthEstimation, vt as __webpack_exports__BlenderbotForConditionalGeneration, vu as __webpack_exports__QuestionAnsweringPipeline, w as __webpack_exports__AutoModelForDepthEstimation, wa as __webpack_exports__GraniteMoeHybridForCausalLM, wc as __webpack_exports__MobileViTV2PreTrainedModel, wd as __webpack_exports__SapiensForDepthEstimation, wf as __webpack_exports__SummarizationPipeline, wh as __webpack_exports__quantize_embeddings, wi as __webpack_exports__FalconModel, wl as __webpack_exports__OwlViTFeatureExtractor, wm as __webpack_exports__XLMModel, wn as __webpack_exports__ConvNextImageProcessor, wo as __webpack_exports__LlamaModel, wp as __webpack_exports__ViTImageProcessor, wr as __webpack_exports__DetrForObjectDetection, ws as __webpack_exports__MgpstrPreTrainedModel, wt as __webpack_exports__BlenderbotSmallTokenizer, wu as __webpack_exports__Qwen2VLImageProcessor, x as __webpack_exports__AutoModelForAudioTextToText, xa as __webpack_exports__GlmPreTrainedModel, xc as __webpack_exports__MobileViTPreTrainedModel, xd as __webpack_exports__SamModel, xf as __webpack_exports__StoppingCriteriaList, xh as __webpack_exports__permute, xi as __webpack_exports__ExaonePreTrainedModel, xl as __webpack_exports__OpenELMForCausalLM, xm as __webpack_exports__XLMForQuestionAnswering, xn as __webpack_exports__ConvBertTokenizer, xo as __webpack_exports__Llama4ForCausalLM, xp as __webpack_exports__VaultGemmaPreTrainedModel, xr as __webpack_exports__DepthProForDepthEstimation, xs as __webpack_exports__Metric3Dv2PreTrainedModel, xt as __webpack_exports__BlenderbotSmallForConditionalGeneration, xu as __webpack_exports__Qwen2PreTrainedModel, y as __webpack_exports__AutoModelForAudioClassification, ya as __webpack_exports__GlmForCausalLM, yc as __webpack_exports__MobileViTImageProcessor, yd as __webpack_exports__SamImageProcessor, yf as __webpack_exports__Starcoder2PreTrainedModel, yh as __webpack_exports__ones, yi as __webpack_exports__ExaoneForCausalLM, yl as __webpack_exports__OlmoModel, ym as __webpack_exports__WhisperTimeStampLogitsProcessor, yn as __webpack_exports__ConvBertModel, yo as __webpack_exports__Lfm2PreTrainedModel, yp as __webpack_exports__VaultGemmaForCausalLM, yr as __webpack_exports__DepthAnythingPreTrainedModel, ys as __webpack_exports__Metric3DPreTrainedModel, yt as __webpack_exports__BlenderbotModel, yu as __webpack_exports__Qwen2ForCausalLM, z as __webpack_exports__AutoModelForSeq2SeqLM, za as __webpack_exports__HieraForImageClassification, zc as __webpack_exports__MoonshineProcessor, zd as __webpack_exports__SiglipPreTrainedModel, zf as __webpack_exports__T5Tokenizer, zi as __webpack_exports__GLPNForDepthEstimation, zl as __webpack_exports__ParakeetPreTrainedModel, zm as __webpack_exports__YolosModel, zn as __webpack_exports__DPTForDepthEstimation, zo as __webpack_exports__M2M100ForConditionalGeneration, zp as __webpack_exports__VitsModel, zr as __webpack_exports__DistilBertForSequenceClassification, zs as __webpack_exports__MobileBertForQuestionAnswering, zt as __webpack_exports__CLIPTokenizer, zu as __webpack_exports__RTDetrV2ForObjectDetection } from "./transformers.web-BschYzqZ.js";
|
|
2
|
-
|
|
3
|
-
export { __webpack_exports__ASTFeatureExtractor as ASTFeatureExtractor, __webpack_exports__ASTForAudioClassification as ASTForAudioClassification, __webpack_exports__ASTModel as ASTModel, __webpack_exports__ASTPreTrainedModel as ASTPreTrainedModel, __webpack_exports__AlbertForMaskedLM as AlbertForMaskedLM, __webpack_exports__AlbertForQuestionAnswering as AlbertForQuestionAnswering, __webpack_exports__AlbertForSequenceClassification as AlbertForSequenceClassification, __webpack_exports__AlbertModel as AlbertModel, __webpack_exports__AlbertPreTrainedModel as AlbertPreTrainedModel, __webpack_exports__AlbertTokenizer as AlbertTokenizer, __webpack_exports__ArceeForCausalLM as ArceeForCausalLM, __webpack_exports__ArceeModel as ArceeModel, __webpack_exports__ArceePreTrainedModel as ArceePreTrainedModel, __webpack_exports__AudioClassificationPipeline as AudioClassificationPipeline, __webpack_exports__AutoConfig as AutoConfig, __webpack_exports__AutoFeatureExtractor as AutoFeatureExtractor, __webpack_exports__AutoImageProcessor as AutoImageProcessor, __webpack_exports__AutoModel as AutoModel, __webpack_exports__AutoModelForAudioClassification as AutoModelForAudioClassification, __webpack_exports__AutoModelForAudioFrameClassification as AutoModelForAudioFrameClassification, __webpack_exports__AutoModelForAudioTextToText as AutoModelForAudioTextToText, __webpack_exports__AutoModelForCTC as AutoModelForCTC, __webpack_exports__AutoModelForCausalLM as AutoModelForCausalLM, __webpack_exports__AutoModelForDepthEstimation as AutoModelForDepthEstimation, __webpack_exports__AutoModelForDocumentQuestionAnswering as AutoModelForDocumentQuestionAnswering, __webpack_exports__AutoModelForImageClassification as AutoModelForImageClassification, __webpack_exports__AutoModelForImageFeatureExtraction as AutoModelForImageFeatureExtraction, __webpack_exports__AutoModelForImageMatting as AutoModelForImageMatting, __webpack_exports__AutoModelForImageSegmentation as AutoModelForImageSegmentation, __webpack_exports__AutoModelForImageTextToText as AutoModelForImageTextToText, __webpack_exports__AutoModelForImageToImage as AutoModelForImageToImage, __webpack_exports__AutoModelForMaskGeneration as AutoModelForMaskGeneration, __webpack_exports__AutoModelForMaskedLM as AutoModelForMaskedLM, __webpack_exports__AutoModelForNormalEstimation as AutoModelForNormalEstimation, __webpack_exports__AutoModelForObjectDetection as AutoModelForObjectDetection, __webpack_exports__AutoModelForPoseEstimation as AutoModelForPoseEstimation, __webpack_exports__AutoModelForQuestionAnswering as AutoModelForQuestionAnswering, __webpack_exports__AutoModelForSemanticSegmentation as AutoModelForSemanticSegmentation, __webpack_exports__AutoModelForSeq2SeqLM as AutoModelForSeq2SeqLM, __webpack_exports__AutoModelForSequenceClassification as AutoModelForSequenceClassification, __webpack_exports__AutoModelForSpeechSeq2Seq as AutoModelForSpeechSeq2Seq, __webpack_exports__AutoModelForTextToSpectrogram as AutoModelForTextToSpectrogram, __webpack_exports__AutoModelForTextToWaveform as AutoModelForTextToWaveform, __webpack_exports__AutoModelForTokenClassification as AutoModelForTokenClassification, __webpack_exports__AutoModelForUniversalSegmentation as AutoModelForUniversalSegmentation, __webpack_exports__AutoModelForVision2Seq as AutoModelForVision2Seq, __webpack_exports__AutoModelForXVector as AutoModelForXVector, __webpack_exports__AutoModelForZeroShotObjectDetection as AutoModelForZeroShotObjectDetection, __webpack_exports__AutoProcessor as AutoProcessor, __webpack_exports__AutoTokenizer as AutoTokenizer, __webpack_exports__AutomaticSpeechRecognitionPipeline as AutomaticSpeechRecognitionPipeline, __webpack_exports__BackgroundRemovalPipeline as BackgroundRemovalPipeline, __webpack_exports__BartForConditionalGeneration as BartForConditionalGeneration, __webpack_exports__BartForSequenceClassification as BartForSequenceClassification, __webpack_exports__BartModel as BartModel, __webpack_exports__BartPretrainedModel as BartPretrainedModel, __webpack_exports__BartTokenizer as BartTokenizer, __webpack_exports__BaseModelOutput as BaseModelOutput, __webpack_exports__BaseStreamer as BaseStreamer, __webpack_exports__BeitFeatureExtractor as BeitFeatureExtractor, __webpack_exports__BeitForImageClassification as BeitForImageClassification, __webpack_exports__BeitModel as BeitModel, __webpack_exports__BeitPreTrainedModel as BeitPreTrainedModel, __webpack_exports__BertForMaskedLM as BertForMaskedLM, __webpack_exports__BertForQuestionAnswering as BertForQuestionAnswering, __webpack_exports__BertForSequenceClassification as BertForSequenceClassification, __webpack_exports__BertForTokenClassification as BertForTokenClassification, __webpack_exports__BertModel as BertModel, __webpack_exports__BertPreTrainedModel as BertPreTrainedModel, __webpack_exports__BertTokenizer as BertTokenizer, __webpack_exports__BitImageProcessor as BitImageProcessor, __webpack_exports__BlenderbotForConditionalGeneration as BlenderbotForConditionalGeneration, __webpack_exports__BlenderbotModel as BlenderbotModel, __webpack_exports__BlenderbotPreTrainedModel as BlenderbotPreTrainedModel, __webpack_exports__BlenderbotSmallForConditionalGeneration as BlenderbotSmallForConditionalGeneration, __webpack_exports__BlenderbotSmallModel as BlenderbotSmallModel, __webpack_exports__BlenderbotSmallPreTrainedModel as BlenderbotSmallPreTrainedModel, __webpack_exports__BlenderbotSmallTokenizer as BlenderbotSmallTokenizer, __webpack_exports__BlenderbotTokenizer as BlenderbotTokenizer, __webpack_exports__BloomForCausalLM as BloomForCausalLM, __webpack_exports__BloomModel as BloomModel, __webpack_exports__BloomPreTrainedModel as BloomPreTrainedModel, __webpack_exports__BloomTokenizer as BloomTokenizer, __webpack_exports__CLIPFeatureExtractor as CLIPFeatureExtractor, __webpack_exports__CLIPImageProcessor as CLIPImageProcessor, __webpack_exports__CLIPModel as CLIPModel, __webpack_exports__CLIPPreTrainedModel as CLIPPreTrainedModel, __webpack_exports__CLIPSegForImageSegmentation as CLIPSegForImageSegmentation, __webpack_exports__CLIPSegModel as CLIPSegModel, __webpack_exports__CLIPSegPreTrainedModel as CLIPSegPreTrainedModel, __webpack_exports__CLIPTextModel as CLIPTextModel, __webpack_exports__CLIPTextModelWithProjection as CLIPTextModelWithProjection, __webpack_exports__CLIPTokenizer as CLIPTokenizer, __webpack_exports__CLIPVisionModel as CLIPVisionModel, __webpack_exports__CLIPVisionModelWithProjection as CLIPVisionModelWithProjection, __webpack_exports__CamembertForMaskedLM as CamembertForMaskedLM, __webpack_exports__CamembertForQuestionAnswering as CamembertForQuestionAnswering, __webpack_exports__CamembertForSequenceClassification as CamembertForSequenceClassification, __webpack_exports__CamembertForTokenClassification as CamembertForTokenClassification, __webpack_exports__CamembertModel as CamembertModel, __webpack_exports__CamembertPreTrainedModel as CamembertPreTrainedModel, __webpack_exports__CamembertTokenizer as CamembertTokenizer, __webpack_exports__CausalLMOutput as CausalLMOutput, __webpack_exports__CausalLMOutputWithPast as CausalLMOutputWithPast, __webpack_exports__ChineseCLIPFeatureExtractor as ChineseCLIPFeatureExtractor, __webpack_exports__ChineseCLIPModel as ChineseCLIPModel, __webpack_exports__ChineseCLIPPreTrainedModel as ChineseCLIPPreTrainedModel, __webpack_exports__ClapAudioModelWithProjection as ClapAudioModelWithProjection, __webpack_exports__ClapFeatureExtractor as ClapFeatureExtractor, __webpack_exports__ClapModel as ClapModel, __webpack_exports__ClapPreTrainedModel as ClapPreTrainedModel, __webpack_exports__ClapTextModelWithProjection as ClapTextModelWithProjection, __webpack_exports__ClassifierFreeGuidanceLogitsProcessor as ClassifierFreeGuidanceLogitsProcessor, __webpack_exports__CodeGenForCausalLM as CodeGenForCausalLM, __webpack_exports__CodeGenModel as CodeGenModel, __webpack_exports__CodeGenPreTrainedModel as CodeGenPreTrainedModel, __webpack_exports__CodeGenTokenizer as CodeGenTokenizer, __webpack_exports__CodeLlamaTokenizer as CodeLlamaTokenizer, __webpack_exports__CohereForCausalLM as CohereForCausalLM, __webpack_exports__CohereModel as CohereModel, __webpack_exports__CoherePreTrainedModel as CoherePreTrainedModel, __webpack_exports__CohereTokenizer as CohereTokenizer, __webpack_exports__ConvBertForMaskedLM as ConvBertForMaskedLM, __webpack_exports__ConvBertForQuestionAnswering as ConvBertForQuestionAnswering, __webpack_exports__ConvBertForSequenceClassification as ConvBertForSequenceClassification, __webpack_exports__ConvBertForTokenClassification as ConvBertForTokenClassification, __webpack_exports__ConvBertModel as ConvBertModel, __webpack_exports__ConvBertPreTrainedModel as ConvBertPreTrainedModel, __webpack_exports__ConvBertTokenizer as ConvBertTokenizer, __webpack_exports__ConvNextFeatureExtractor as ConvNextFeatureExtractor, __webpack_exports__ConvNextForImageClassification as ConvNextForImageClassification, __webpack_exports__ConvNextImageProcessor as ConvNextImageProcessor, __webpack_exports__ConvNextModel as ConvNextModel, __webpack_exports__ConvNextPreTrainedModel as ConvNextPreTrainedModel, __webpack_exports__ConvNextV2ForImageClassification as ConvNextV2ForImageClassification, __webpack_exports__ConvNextV2Model as ConvNextV2Model, __webpack_exports__ConvNextV2PreTrainedModel as ConvNextV2PreTrainedModel, __webpack_exports__DFineForObjectDetection as DFineForObjectDetection, __webpack_exports__DFineModel as DFineModel, __webpack_exports__DFinePreTrainedModel as DFinePreTrainedModel, __webpack_exports__DINOv3ConvNextModel as DINOv3ConvNextModel, __webpack_exports__DINOv3ConvNextPreTrainedModel as DINOv3ConvNextPreTrainedModel, __webpack_exports__DINOv3ViTImageProcessor as DINOv3ViTImageProcessor, __webpack_exports__DINOv3ViTModel as DINOv3ViTModel, __webpack_exports__DINOv3ViTPreTrainedModel as DINOv3ViTPreTrainedModel, __webpack_exports__DPTFeatureExtractor as DPTFeatureExtractor, __webpack_exports__DPTForDepthEstimation as DPTForDepthEstimation, __webpack_exports__DPTImageProcessor as DPTImageProcessor, __webpack_exports__DPTModel as DPTModel, __webpack_exports__DPTPreTrainedModel as DPTPreTrainedModel, __webpack_exports__DacDecoderModel as DacDecoderModel, __webpack_exports__DacDecoderOutput as DacDecoderOutput, __webpack_exports__DacEncoderModel as DacEncoderModel, __webpack_exports__DacEncoderOutput as DacEncoderOutput, __webpack_exports__DacFeatureExtractor as DacFeatureExtractor, __webpack_exports__DacModel as DacModel, __webpack_exports__DacPreTrainedModel as DacPreTrainedModel, __webpack_exports__DataTypeMap as DataTypeMap, __webpack_exports__DebertaForMaskedLM as DebertaForMaskedLM, __webpack_exports__DebertaForQuestionAnswering as DebertaForQuestionAnswering, __webpack_exports__DebertaForSequenceClassification as DebertaForSequenceClassification, __webpack_exports__DebertaForTokenClassification as DebertaForTokenClassification, __webpack_exports__DebertaModel as DebertaModel, __webpack_exports__DebertaPreTrainedModel as DebertaPreTrainedModel, __webpack_exports__DebertaTokenizer as DebertaTokenizer, __webpack_exports__DebertaV2ForMaskedLM as DebertaV2ForMaskedLM, __webpack_exports__DebertaV2ForQuestionAnswering as DebertaV2ForQuestionAnswering, __webpack_exports__DebertaV2ForSequenceClassification as DebertaV2ForSequenceClassification, __webpack_exports__DebertaV2ForTokenClassification as DebertaV2ForTokenClassification, __webpack_exports__DebertaV2Model as DebertaV2Model, __webpack_exports__DebertaV2PreTrainedModel as DebertaV2PreTrainedModel, __webpack_exports__DebertaV2Tokenizer as DebertaV2Tokenizer, __webpack_exports__DecisionTransformerModel as DecisionTransformerModel, __webpack_exports__DecisionTransformerPreTrainedModel as DecisionTransformerPreTrainedModel, __webpack_exports__DeiTFeatureExtractor as DeiTFeatureExtractor, __webpack_exports__DeiTForImageClassification as DeiTForImageClassification, __webpack_exports__DeiTImageProcessor as DeiTImageProcessor, __webpack_exports__DeiTModel as DeiTModel, __webpack_exports__DeiTPreTrainedModel as DeiTPreTrainedModel, __webpack_exports__DepthAnythingForDepthEstimation as DepthAnythingForDepthEstimation, __webpack_exports__DepthAnythingPreTrainedModel as DepthAnythingPreTrainedModel, __webpack_exports__DepthEstimationPipeline as DepthEstimationPipeline, __webpack_exports__DepthProForDepthEstimation as DepthProForDepthEstimation, __webpack_exports__DepthProPreTrainedModel as DepthProPreTrainedModel, __webpack_exports__DetrFeatureExtractor as DetrFeatureExtractor, __webpack_exports__DetrForObjectDetection as DetrForObjectDetection, __webpack_exports__DetrForSegmentation as DetrForSegmentation, __webpack_exports__DetrImageProcessor as DetrImageProcessor, __webpack_exports__DetrModel as DetrModel, __webpack_exports__DetrObjectDetectionOutput as DetrObjectDetectionOutput, __webpack_exports__DetrPreTrainedModel as DetrPreTrainedModel, __webpack_exports__DetrSegmentationOutput as DetrSegmentationOutput, __webpack_exports__Dinov2ForImageClassification as Dinov2ForImageClassification, __webpack_exports__Dinov2Model as Dinov2Model, __webpack_exports__Dinov2PreTrainedModel as Dinov2PreTrainedModel, __webpack_exports__Dinov2WithRegistersForImageClassification as Dinov2WithRegistersForImageClassification, __webpack_exports__Dinov2WithRegistersModel as Dinov2WithRegistersModel, __webpack_exports__Dinov2WithRegistersPreTrainedModel as Dinov2WithRegistersPreTrainedModel, __webpack_exports__DistilBertForMaskedLM as DistilBertForMaskedLM, __webpack_exports__DistilBertForQuestionAnswering as DistilBertForQuestionAnswering, __webpack_exports__DistilBertForSequenceClassification as DistilBertForSequenceClassification, __webpack_exports__DistilBertForTokenClassification as DistilBertForTokenClassification, __webpack_exports__DistilBertModel as DistilBertModel, __webpack_exports__DistilBertPreTrainedModel as DistilBertPreTrainedModel, __webpack_exports__DistilBertTokenizer as DistilBertTokenizer, __webpack_exports__DocumentQuestionAnsweringPipeline as DocumentQuestionAnsweringPipeline, __webpack_exports__DonutFeatureExtractor as DonutFeatureExtractor, __webpack_exports__DonutImageProcessor as DonutImageProcessor, __webpack_exports__DonutSwinModel as DonutSwinModel, __webpack_exports__DonutSwinPreTrainedModel as DonutSwinPreTrainedModel, __webpack_exports__EdgeTamModel as EdgeTamModel, __webpack_exports__EfficientNetForImageClassification as EfficientNetForImageClassification, __webpack_exports__EfficientNetImageProcessor as EfficientNetImageProcessor, __webpack_exports__EfficientNetModel as EfficientNetModel, __webpack_exports__EfficientNetPreTrainedModel as EfficientNetPreTrainedModel, __webpack_exports__ElectraForMaskedLM as ElectraForMaskedLM, __webpack_exports__ElectraForQuestionAnswering as ElectraForQuestionAnswering, __webpack_exports__ElectraForSequenceClassification as ElectraForSequenceClassification, __webpack_exports__ElectraForTokenClassification as ElectraForTokenClassification, __webpack_exports__ElectraModel as ElectraModel, __webpack_exports__ElectraPreTrainedModel as ElectraPreTrainedModel, __webpack_exports__ElectraTokenizer as ElectraTokenizer, __webpack_exports__EncodecFeatureExtractor as EncodecFeatureExtractor, __webpack_exports__EosTokenCriteria as EosTokenCriteria, __webpack_exports__Ernie4_5_ForCausalLM as Ernie4_5_ForCausalLM, __webpack_exports__Ernie4_5_Model as Ernie4_5_Model, __webpack_exports__Ernie4_5_PretrainedModel as Ernie4_5_PretrainedModel, __webpack_exports__Ernie4_5_Tokenizer as Ernie4_5_Tokenizer, __webpack_exports__EsmForMaskedLM as EsmForMaskedLM, __webpack_exports__EsmForSequenceClassification as EsmForSequenceClassification, __webpack_exports__EsmForTokenClassification as EsmForTokenClassification, __webpack_exports__EsmModel as EsmModel, __webpack_exports__EsmPreTrainedModel as EsmPreTrainedModel, __webpack_exports__EsmTokenizer as EsmTokenizer, __webpack_exports__ExaoneForCausalLM as ExaoneForCausalLM, __webpack_exports__ExaoneModel as ExaoneModel, __webpack_exports__ExaonePreTrainedModel as ExaonePreTrainedModel, __webpack_exports__FFT as FFT, __webpack_exports__FalconForCausalLM as FalconForCausalLM, __webpack_exports__FalconModel as FalconModel, __webpack_exports__FalconPreTrainedModel as FalconPreTrainedModel, __webpack_exports__FalconTokenizer as FalconTokenizer, __webpack_exports__FastViTForImageClassification as FastViTForImageClassification, __webpack_exports__FastViTModel as FastViTModel, __webpack_exports__FastViTPreTrainedModel as FastViTPreTrainedModel, __webpack_exports__FeatureExtractionPipeline as FeatureExtractionPipeline, __webpack_exports__FeatureExtractor as FeatureExtractor, __webpack_exports__FillMaskPipeline as FillMaskPipeline, __webpack_exports__Florence2ForConditionalGeneration as Florence2ForConditionalGeneration, __webpack_exports__Florence2PreTrainedModel as Florence2PreTrainedModel, __webpack_exports__Florence2Processor as Florence2Processor, __webpack_exports__ForcedBOSTokenLogitsProcessor as ForcedBOSTokenLogitsProcessor, __webpack_exports__ForcedEOSTokenLogitsProcessor as ForcedEOSTokenLogitsProcessor, __webpack_exports__GLPNFeatureExtractor as GLPNFeatureExtractor, __webpack_exports__GLPNForDepthEstimation as GLPNForDepthEstimation, __webpack_exports__GLPNModel as GLPNModel, __webpack_exports__GLPNPreTrainedModel as GLPNPreTrainedModel, __webpack_exports__GPT2LMHeadModel as GPT2LMHeadModel, __webpack_exports__GPT2Model as GPT2Model, __webpack_exports__GPT2PreTrainedModel as GPT2PreTrainedModel, __webpack_exports__GPT2Tokenizer as GPT2Tokenizer, __webpack_exports__GPTBigCodeForCausalLM as GPTBigCodeForCausalLM, __webpack_exports__GPTBigCodeModel as GPTBigCodeModel, __webpack_exports__GPTBigCodePreTrainedModel as GPTBigCodePreTrainedModel, __webpack_exports__GPTJForCausalLM as GPTJForCausalLM, __webpack_exports__GPTJModel as GPTJModel, __webpack_exports__GPTJPreTrainedModel as GPTJPreTrainedModel, __webpack_exports__GPTNeoForCausalLM as GPTNeoForCausalLM, __webpack_exports__GPTNeoModel as GPTNeoModel, __webpack_exports__GPTNeoPreTrainedModel as GPTNeoPreTrainedModel, __webpack_exports__GPTNeoXForCausalLM as GPTNeoXForCausalLM, __webpack_exports__GPTNeoXModel as GPTNeoXModel, __webpack_exports__GPTNeoXPreTrainedModel as GPTNeoXPreTrainedModel, __webpack_exports__GPTNeoXTokenizer as GPTNeoXTokenizer, __webpack_exports__Gemma2ForCausalLM as Gemma2ForCausalLM, __webpack_exports__Gemma2Model as Gemma2Model, __webpack_exports__Gemma2PreTrainedModel as Gemma2PreTrainedModel, __webpack_exports__Gemma3ForCausalLM as Gemma3ForCausalLM, __webpack_exports__Gemma3Model as Gemma3Model, __webpack_exports__Gemma3PreTrainedModel as Gemma3PreTrainedModel, __webpack_exports__Gemma3nAudioFeatureExtractor as Gemma3nAudioFeatureExtractor, __webpack_exports__Gemma3nForConditionalGeneration as Gemma3nForConditionalGeneration, __webpack_exports__Gemma3nPreTrainedModel as Gemma3nPreTrainedModel, __webpack_exports__Gemma3nProcessor as Gemma3nProcessor, __webpack_exports__GemmaForCausalLM as GemmaForCausalLM, __webpack_exports__GemmaModel as GemmaModel, __webpack_exports__GemmaPreTrainedModel as GemmaPreTrainedModel, __webpack_exports__GemmaTokenizer as GemmaTokenizer, __webpack_exports__GlmForCausalLM as GlmForCausalLM, __webpack_exports__GlmModel as GlmModel, __webpack_exports__GlmPreTrainedModel as GlmPreTrainedModel, __webpack_exports__GraniteForCausalLM as GraniteForCausalLM, __webpack_exports__GraniteModel as GraniteModel, __webpack_exports__GraniteMoeHybridForCausalLM as GraniteMoeHybridForCausalLM, __webpack_exports__GraniteMoeHybridModel as GraniteMoeHybridModel, __webpack_exports__GraniteMoeHybridPreTrainedModel as GraniteMoeHybridPreTrainedModel, __webpack_exports__GranitePreTrainedModel as GranitePreTrainedModel, __webpack_exports__Grok1Tokenizer as Grok1Tokenizer, __webpack_exports__GroundingDinoForObjectDetection as GroundingDinoForObjectDetection, __webpack_exports__GroundingDinoImageProcessor as GroundingDinoImageProcessor, __webpack_exports__GroundingDinoPreTrainedModel as GroundingDinoPreTrainedModel, __webpack_exports__GroundingDinoProcessor as GroundingDinoProcessor, __webpack_exports__GroupViTModel as GroupViTModel, __webpack_exports__GroupViTPreTrainedModel as GroupViTPreTrainedModel, __webpack_exports__HeliumForCausalLM as HeliumForCausalLM, __webpack_exports__HeliumModel as HeliumModel, __webpack_exports__HeliumPreTrainedModel as HeliumPreTrainedModel, __webpack_exports__HerbertTokenizer as HerbertTokenizer, __webpack_exports__HieraForImageClassification as HieraForImageClassification, __webpack_exports__HieraModel as HieraModel, __webpack_exports__HieraPreTrainedModel as HieraPreTrainedModel, __webpack_exports__HubertForCTC as HubertForCTC, __webpack_exports__HubertForSequenceClassification as HubertForSequenceClassification, __webpack_exports__HubertModel as HubertModel, __webpack_exports__HubertPreTrainedModel as HubertPreTrainedModel, __webpack_exports__IJepaForImageClassification as IJepaForImageClassification, __webpack_exports__IJepaModel as IJepaModel, __webpack_exports__IJepaPreTrainedModel as IJepaPreTrainedModel, __webpack_exports__Idefics3ForConditionalGeneration as Idefics3ForConditionalGeneration, __webpack_exports__Idefics3ImageProcessor as Idefics3ImageProcessor, __webpack_exports__Idefics3PreTrainedModel as Idefics3PreTrainedModel, __webpack_exports__Idefics3Processor as Idefics3Processor, __webpack_exports__ImageClassificationPipeline as ImageClassificationPipeline, __webpack_exports__ImageFeatureExtractionPipeline as ImageFeatureExtractionPipeline, __webpack_exports__ImageFeatureExtractor as ImageFeatureExtractor, __webpack_exports__ImageMattingOutput as ImageMattingOutput, __webpack_exports__ImageProcessor as ImageProcessor, __webpack_exports__ImageSegmentationPipeline as ImageSegmentationPipeline, __webpack_exports__ImageToImagePipeline as ImageToImagePipeline, __webpack_exports__ImageToTextPipeline as ImageToTextPipeline, __webpack_exports__InterruptableStoppingCriteria as InterruptableStoppingCriteria, __webpack_exports__JAISLMHeadModel as JAISLMHeadModel, __webpack_exports__JAISModel as JAISModel, __webpack_exports__JAISPreTrainedModel as JAISPreTrainedModel, __webpack_exports__JinaCLIPImageProcessor as JinaCLIPImageProcessor, __webpack_exports__JinaCLIPModel as JinaCLIPModel, __webpack_exports__JinaCLIPPreTrainedModel as JinaCLIPPreTrainedModel, __webpack_exports__JinaCLIPProcessor as JinaCLIPProcessor, __webpack_exports__JinaCLIPTextModel as JinaCLIPTextModel, __webpack_exports__JinaCLIPVisionModel as JinaCLIPVisionModel, __webpack_exports__Lfm2ForCausalLM as Lfm2ForCausalLM, __webpack_exports__Lfm2Model as Lfm2Model, __webpack_exports__Lfm2PreTrainedModel as Lfm2PreTrainedModel, __webpack_exports__LiteWhisperForConditionalGeneration as LiteWhisperForConditionalGeneration, __webpack_exports__Llama4ForCausalLM as Llama4ForCausalLM, __webpack_exports__Llama4PreTrainedModel as Llama4PreTrainedModel, __webpack_exports__LlamaForCausalLM as LlamaForCausalLM, __webpack_exports__LlamaModel as LlamaModel, __webpack_exports__LlamaPreTrainedModel as LlamaPreTrainedModel, __webpack_exports__LlamaTokenizer as LlamaTokenizer, __webpack_exports__LlavaForConditionalGeneration as LlavaForConditionalGeneration, __webpack_exports__LlavaOnevisionForConditionalGeneration as LlavaOnevisionForConditionalGeneration, __webpack_exports__LlavaOnevisionImageProcessor as LlavaOnevisionImageProcessor, __webpack_exports__LlavaPreTrainedModel as LlavaPreTrainedModel, __webpack_exports__LlavaProcessor as LlavaProcessor, __webpack_exports__LlavaQwen2ForCausalLM as LlavaQwen2ForCausalLM, __webpack_exports__LogitsProcessor as LogitsProcessor, __webpack_exports__LogitsProcessorList as LogitsProcessorList, __webpack_exports__LogitsWarper as LogitsWarper, __webpack_exports__LongT5ForConditionalGeneration as LongT5ForConditionalGeneration, __webpack_exports__LongT5Model as LongT5Model, __webpack_exports__LongT5PreTrainedModel as LongT5PreTrainedModel, __webpack_exports__M2M100ForConditionalGeneration as M2M100ForConditionalGeneration, __webpack_exports__M2M100Model as M2M100Model, __webpack_exports__M2M100PreTrainedModel as M2M100PreTrainedModel, __webpack_exports__M2M100Tokenizer as M2M100Tokenizer, __webpack_exports__MBart50Tokenizer as MBart50Tokenizer, __webpack_exports__MBartForCausalLM as MBartForCausalLM, __webpack_exports__MBartForConditionalGeneration as MBartForConditionalGeneration, __webpack_exports__MBartForSequenceClassification as MBartForSequenceClassification, __webpack_exports__MBartModel as MBartModel, __webpack_exports__MBartPreTrainedModel as MBartPreTrainedModel, __webpack_exports__MBartTokenizer as MBartTokenizer, __webpack_exports__MPNetForMaskedLM as MPNetForMaskedLM, __webpack_exports__MPNetForQuestionAnswering as MPNetForQuestionAnswering, __webpack_exports__MPNetForSequenceClassification as MPNetForSequenceClassification, __webpack_exports__MPNetForTokenClassification as MPNetForTokenClassification, __webpack_exports__MPNetModel as MPNetModel, __webpack_exports__MPNetPreTrainedModel as MPNetPreTrainedModel, __webpack_exports__MPNetTokenizer as MPNetTokenizer, __webpack_exports__MT5ForConditionalGeneration as MT5ForConditionalGeneration, __webpack_exports__MT5Model as MT5Model, __webpack_exports__MT5PreTrainedModel as MT5PreTrainedModel, __webpack_exports__MarianMTModel as MarianMTModel, __webpack_exports__MarianModel as MarianModel, __webpack_exports__MarianPreTrainedModel as MarianPreTrainedModel, __webpack_exports__MarianTokenizer as MarianTokenizer, __webpack_exports__Mask2FormerImageProcessor as Mask2FormerImageProcessor, __webpack_exports__MaskFormerFeatureExtractor as MaskFormerFeatureExtractor, __webpack_exports__MaskFormerForInstanceSegmentation as MaskFormerForInstanceSegmentation, __webpack_exports__MaskFormerImageProcessor as MaskFormerImageProcessor, __webpack_exports__MaskFormerModel as MaskFormerModel, __webpack_exports__MaskFormerPreTrainedModel as MaskFormerPreTrainedModel, __webpack_exports__MaskedLMOutput as MaskedLMOutput, __webpack_exports__MaxLengthCriteria as MaxLengthCriteria, __webpack_exports__Metric3DForDepthEstimation as Metric3DForDepthEstimation, __webpack_exports__Metric3DPreTrainedModel as Metric3DPreTrainedModel, __webpack_exports__Metric3Dv2ForDepthEstimation as Metric3Dv2ForDepthEstimation, __webpack_exports__Metric3Dv2PreTrainedModel as Metric3Dv2PreTrainedModel, __webpack_exports__MgpstrForSceneTextRecognition as MgpstrForSceneTextRecognition, __webpack_exports__MgpstrModelOutput as MgpstrModelOutput, __webpack_exports__MgpstrPreTrainedModel as MgpstrPreTrainedModel, __webpack_exports__MgpstrProcessor as MgpstrProcessor, __webpack_exports__MgpstrTokenizer as MgpstrTokenizer, __webpack_exports__MimiDecoderModel as MimiDecoderModel, __webpack_exports__MimiDecoderOutput as MimiDecoderOutput, __webpack_exports__MimiEncoderModel as MimiEncoderModel, __webpack_exports__MimiEncoderOutput as MimiEncoderOutput, __webpack_exports__MimiModel as MimiModel, __webpack_exports__MimiPreTrainedModel as MimiPreTrainedModel, __webpack_exports__MinLengthLogitsProcessor as MinLengthLogitsProcessor, __webpack_exports__MinNewTokensLengthLogitsProcessor as MinNewTokensLengthLogitsProcessor, __webpack_exports__MistralForCausalLM as MistralForCausalLM, __webpack_exports__MistralModel as MistralModel, __webpack_exports__MistralPreTrainedModel as MistralPreTrainedModel, __webpack_exports__MobileBertForMaskedLM as MobileBertForMaskedLM, __webpack_exports__MobileBertForQuestionAnswering as MobileBertForQuestionAnswering, __webpack_exports__MobileBertForSequenceClassification as MobileBertForSequenceClassification, __webpack_exports__MobileBertModel as MobileBertModel, __webpack_exports__MobileBertPreTrainedModel as MobileBertPreTrainedModel, __webpack_exports__MobileBertTokenizer as MobileBertTokenizer, __webpack_exports__MobileLLMForCausalLM as MobileLLMForCausalLM, __webpack_exports__MobileLLMModel as MobileLLMModel, __webpack_exports__MobileLLMPreTrainedModel as MobileLLMPreTrainedModel, __webpack_exports__MobileNetV1FeatureExtractor as MobileNetV1FeatureExtractor, __webpack_exports__MobileNetV1ForImageClassification as MobileNetV1ForImageClassification, __webpack_exports__MobileNetV1ForSemanticSegmentation as MobileNetV1ForSemanticSegmentation, __webpack_exports__MobileNetV1ImageProcessor as MobileNetV1ImageProcessor, __webpack_exports__MobileNetV1Model as MobileNetV1Model, __webpack_exports__MobileNetV1PreTrainedModel as MobileNetV1PreTrainedModel, __webpack_exports__MobileNetV2FeatureExtractor as MobileNetV2FeatureExtractor, __webpack_exports__MobileNetV2ForImageClassification as MobileNetV2ForImageClassification, __webpack_exports__MobileNetV2ForSemanticSegmentation as MobileNetV2ForSemanticSegmentation, __webpack_exports__MobileNetV2ImageProcessor as MobileNetV2ImageProcessor, __webpack_exports__MobileNetV2Model as MobileNetV2Model, __webpack_exports__MobileNetV2PreTrainedModel as MobileNetV2PreTrainedModel, __webpack_exports__MobileNetV3FeatureExtractor as MobileNetV3FeatureExtractor, __webpack_exports__MobileNetV3ForImageClassification as MobileNetV3ForImageClassification, __webpack_exports__MobileNetV3ForSemanticSegmentation as MobileNetV3ForSemanticSegmentation, __webpack_exports__MobileNetV3ImageProcessor as MobileNetV3ImageProcessor, __webpack_exports__MobileNetV3Model as MobileNetV3Model, __webpack_exports__MobileNetV3PreTrainedModel as MobileNetV3PreTrainedModel, __webpack_exports__MobileNetV4FeatureExtractor as MobileNetV4FeatureExtractor, __webpack_exports__MobileNetV4ForImageClassification as MobileNetV4ForImageClassification, __webpack_exports__MobileNetV4ForSemanticSegmentation as MobileNetV4ForSemanticSegmentation, __webpack_exports__MobileNetV4ImageProcessor as MobileNetV4ImageProcessor, __webpack_exports__MobileNetV4Model as MobileNetV4Model, __webpack_exports__MobileNetV4PreTrainedModel as MobileNetV4PreTrainedModel, __webpack_exports__MobileViTFeatureExtractor as MobileViTFeatureExtractor, __webpack_exports__MobileViTForImageClassification as MobileViTForImageClassification, __webpack_exports__MobileViTImageProcessor as MobileViTImageProcessor, __webpack_exports__MobileViTModel as MobileViTModel, __webpack_exports__MobileViTPreTrainedModel as MobileViTPreTrainedModel, __webpack_exports__MobileViTV2ForImageClassification as MobileViTV2ForImageClassification, __webpack_exports__MobileViTV2Model as MobileViTV2Model, __webpack_exports__MobileViTV2PreTrainedModel as MobileViTV2PreTrainedModel, __webpack_exports__ModelOutput as ModelOutput, __webpack_exports__ModernBertDecoderForCausalLM as ModernBertDecoderForCausalLM, __webpack_exports__ModernBertDecoderModel as ModernBertDecoderModel, __webpack_exports__ModernBertDecoderPreTrainedModel as ModernBertDecoderPreTrainedModel, __webpack_exports__ModernBertForMaskedLM as ModernBertForMaskedLM, __webpack_exports__ModernBertForSequenceClassification as ModernBertForSequenceClassification, __webpack_exports__ModernBertForTokenClassification as ModernBertForTokenClassification, __webpack_exports__ModernBertModel as ModernBertModel, __webpack_exports__ModernBertPreTrainedModel as ModernBertPreTrainedModel, __webpack_exports__Moondream1ForConditionalGeneration as Moondream1ForConditionalGeneration, __webpack_exports__MoonshineFeatureExtractor as MoonshineFeatureExtractor, __webpack_exports__MoonshineForConditionalGeneration as MoonshineForConditionalGeneration, __webpack_exports__MoonshineModel as MoonshineModel, __webpack_exports__MoonshinePreTrainedModel as MoonshinePreTrainedModel, __webpack_exports__MoonshineProcessor as MoonshineProcessor, __webpack_exports__MptForCausalLM as MptForCausalLM, __webpack_exports__MptModel as MptModel, __webpack_exports__MptPreTrainedModel as MptPreTrainedModel, __webpack_exports__MultiModalityCausalLM as MultiModalityCausalLM, __webpack_exports__MultiModalityPreTrainedModel as MultiModalityPreTrainedModel, __webpack_exports__MusicgenForCausalLM as MusicgenForCausalLM, __webpack_exports__MusicgenForConditionalGeneration as MusicgenForConditionalGeneration, __webpack_exports__MusicgenModel as MusicgenModel, __webpack_exports__MusicgenPreTrainedModel as MusicgenPreTrainedModel, __webpack_exports__NanoChatForCausalLM as NanoChatForCausalLM, __webpack_exports__NanoChatModel as NanoChatModel, __webpack_exports__NanoChatPreTrainedModel as NanoChatPreTrainedModel, __webpack_exports__NeoBertForMaskedLM as NeoBertForMaskedLM, __webpack_exports__NeoBertForQuestionAnswering as NeoBertForQuestionAnswering, __webpack_exports__NeoBertForSequenceClassification as NeoBertForSequenceClassification, __webpack_exports__NeoBertForTokenClassification as NeoBertForTokenClassification, __webpack_exports__NeoBertModel as NeoBertModel, __webpack_exports__NeoBertPreTrainedModel as NeoBertPreTrainedModel, __webpack_exports__NllbTokenizer as NllbTokenizer, __webpack_exports__NoBadWordsLogitsProcessor as NoBadWordsLogitsProcessor, __webpack_exports__NoRepeatNGramLogitsProcessor as NoRepeatNGramLogitsProcessor, __webpack_exports__NomicBertModel as NomicBertModel, __webpack_exports__NomicBertPreTrainedModel as NomicBertPreTrainedModel, __webpack_exports__NougatImageProcessor as NougatImageProcessor, __webpack_exports__NougatTokenizer as NougatTokenizer, __webpack_exports__OPTForCausalLM as OPTForCausalLM, __webpack_exports__OPTModel as OPTModel, __webpack_exports__OPTPreTrainedModel as OPTPreTrainedModel, __webpack_exports__ObjectDetectionPipeline as ObjectDetectionPipeline, __webpack_exports__Olmo2ForCausalLM as Olmo2ForCausalLM, __webpack_exports__Olmo2Model as Olmo2Model, __webpack_exports__Olmo2PreTrainedModel as Olmo2PreTrainedModel, __webpack_exports__OlmoForCausalLM as OlmoForCausalLM, __webpack_exports__OlmoModel as OlmoModel, __webpack_exports__OlmoPreTrainedModel as OlmoPreTrainedModel, __webpack_exports__OpenELMForCausalLM as OpenELMForCausalLM, __webpack_exports__OpenELMModel as OpenELMModel, __webpack_exports__OpenELMPreTrainedModel as OpenELMPreTrainedModel, __webpack_exports__OwlViTFeatureExtractor as OwlViTFeatureExtractor, __webpack_exports__OwlViTForObjectDetection as OwlViTForObjectDetection, __webpack_exports__OwlViTImageProcessor as OwlViTImageProcessor, __webpack_exports__OwlViTModel as OwlViTModel, __webpack_exports__OwlViTPreTrainedModel as OwlViTPreTrainedModel, __webpack_exports__OwlViTProcessor as OwlViTProcessor, __webpack_exports__Owlv2ForObjectDetection as Owlv2ForObjectDetection, __webpack_exports__Owlv2ImageProcessor as Owlv2ImageProcessor, __webpack_exports__Owlv2Model as Owlv2Model, __webpack_exports__Owlv2PreTrainedModel as Owlv2PreTrainedModel, __webpack_exports__PaliGemmaForConditionalGeneration as PaliGemmaForConditionalGeneration, __webpack_exports__PaliGemmaPreTrainedModel as PaliGemmaPreTrainedModel, __webpack_exports__PaliGemmaProcessor as PaliGemmaProcessor, __webpack_exports__ParakeetFeatureExtractor as ParakeetFeatureExtractor, __webpack_exports__ParakeetForCTC as ParakeetForCTC, __webpack_exports__ParakeetPreTrainedModel as ParakeetPreTrainedModel, __webpack_exports__PatchTSMixerForPrediction as PatchTSMixerForPrediction, __webpack_exports__PatchTSMixerModel as PatchTSMixerModel, __webpack_exports__PatchTSMixerPreTrainedModel as PatchTSMixerPreTrainedModel, __webpack_exports__PatchTSTForPrediction as PatchTSTForPrediction, __webpack_exports__PatchTSTModel as PatchTSTModel, __webpack_exports__PatchTSTPreTrainedModel as PatchTSTPreTrainedModel, __webpack_exports__Phi3ForCausalLM as Phi3ForCausalLM, __webpack_exports__Phi3Model as Phi3Model, __webpack_exports__Phi3PreTrainedModel as Phi3PreTrainedModel, __webpack_exports__Phi3VForCausalLM as Phi3VForCausalLM, __webpack_exports__Phi3VImageProcessor as Phi3VImageProcessor, __webpack_exports__Phi3VPreTrainedModel as Phi3VPreTrainedModel, __webpack_exports__Phi3VProcessor as Phi3VProcessor, __webpack_exports__PhiForCausalLM as PhiForCausalLM, __webpack_exports__PhiModel as PhiModel, __webpack_exports__PhiPreTrainedModel as PhiPreTrainedModel, __webpack_exports__Pipeline as Pipeline, __webpack_exports__PreTrainedModel as PreTrainedModel, __webpack_exports__PreTrainedTokenizer as PreTrainedTokenizer, __webpack_exports__PretrainedConfig as PretrainedConfig, __webpack_exports__PretrainedMixin as PretrainedMixin, __webpack_exports__Processor as Processor, __webpack_exports__PvtForImageClassification as PvtForImageClassification, __webpack_exports__PvtImageProcessor as PvtImageProcessor, __webpack_exports__PvtModel as PvtModel, __webpack_exports__PvtPreTrainedModel as PvtPreTrainedModel, __webpack_exports__PyAnnoteFeatureExtractor as PyAnnoteFeatureExtractor, __webpack_exports__PyAnnoteForAudioFrameClassification as PyAnnoteForAudioFrameClassification, __webpack_exports__PyAnnoteModel as PyAnnoteModel, __webpack_exports__PyAnnotePreTrainedModel as PyAnnotePreTrainedModel, __webpack_exports__PyAnnoteProcessor as PyAnnoteProcessor, __webpack_exports__QuestionAnsweringModelOutput as QuestionAnsweringModelOutput, __webpack_exports__QuestionAnsweringPipeline as QuestionAnsweringPipeline, __webpack_exports__Qwen2ForCausalLM as Qwen2ForCausalLM, __webpack_exports__Qwen2Model as Qwen2Model, __webpack_exports__Qwen2PreTrainedModel as Qwen2PreTrainedModel, __webpack_exports__Qwen2Tokenizer as Qwen2Tokenizer, __webpack_exports__Qwen2VLForConditionalGeneration as Qwen2VLForConditionalGeneration, __webpack_exports__Qwen2VLImageProcessor as Qwen2VLImageProcessor, __webpack_exports__Qwen2VLPreTrainedModel as Qwen2VLPreTrainedModel, __webpack_exports__Qwen2VLProcessor as Qwen2VLProcessor, __webpack_exports__Qwen3ForCausalLM as Qwen3ForCausalLM, __webpack_exports__Qwen3Model as Qwen3Model, __webpack_exports__Qwen3PreTrainedModel as Qwen3PreTrainedModel, __webpack_exports__RFDetrForObjectDetection as RFDetrForObjectDetection, __webpack_exports__RFDetrModel as RFDetrModel, __webpack_exports__RFDetrObjectDetectionOutput as RFDetrObjectDetectionOutput, __webpack_exports__RFDetrPreTrainedModel as RFDetrPreTrainedModel, __webpack_exports__RTDetrForObjectDetection as RTDetrForObjectDetection, __webpack_exports__RTDetrImageProcessor as RTDetrImageProcessor, __webpack_exports__RTDetrModel as RTDetrModel, __webpack_exports__RTDetrObjectDetectionOutput as RTDetrObjectDetectionOutput, __webpack_exports__RTDetrPreTrainedModel as RTDetrPreTrainedModel, __webpack_exports__RTDetrV2ForObjectDetection as RTDetrV2ForObjectDetection, __webpack_exports__RTDetrV2Model as RTDetrV2Model, __webpack_exports__RTDetrV2ObjectDetectionOutput as RTDetrV2ObjectDetectionOutput, __webpack_exports__RTDetrV2PreTrainedModel as RTDetrV2PreTrainedModel, __webpack_exports__RawAudio as RawAudio, __webpack_exports__RawImage as RawImage, __webpack_exports__RawVideo as RawVideo, __webpack_exports__RawVideoFrame as RawVideoFrame, __webpack_exports__RepetitionPenaltyLogitsProcessor as RepetitionPenaltyLogitsProcessor, __webpack_exports__ResNetForImageClassification as ResNetForImageClassification, __webpack_exports__ResNetModel as ResNetModel, __webpack_exports__ResNetPreTrainedModel as ResNetPreTrainedModel, __webpack_exports__RoFormerForMaskedLM as RoFormerForMaskedLM, __webpack_exports__RoFormerForQuestionAnswering as RoFormerForQuestionAnswering, __webpack_exports__RoFormerForSequenceClassification as RoFormerForSequenceClassification, __webpack_exports__RoFormerForTokenClassification as RoFormerForTokenClassification, __webpack_exports__RoFormerModel as RoFormerModel, __webpack_exports__RoFormerPreTrainedModel as RoFormerPreTrainedModel, __webpack_exports__RoFormerTokenizer as RoFormerTokenizer, __webpack_exports__RobertaForMaskedLM as RobertaForMaskedLM, __webpack_exports__RobertaForQuestionAnswering as RobertaForQuestionAnswering, __webpack_exports__RobertaForSequenceClassification as RobertaForSequenceClassification, __webpack_exports__RobertaForTokenClassification as RobertaForTokenClassification, __webpack_exports__RobertaModel as RobertaModel, __webpack_exports__RobertaPreTrainedModel as RobertaPreTrainedModel, __webpack_exports__RobertaTokenizer as RobertaTokenizer, __webpack_exports__Sam2ImageProcessor as Sam2ImageProcessor, __webpack_exports__Sam2ImageSegmentationOutput as Sam2ImageSegmentationOutput, __webpack_exports__Sam2Model as Sam2Model, __webpack_exports__Sam2PreTrainedModel as Sam2PreTrainedModel, __webpack_exports__Sam2Processor as Sam2Processor, __webpack_exports__Sam2VideoProcessor as Sam2VideoProcessor, __webpack_exports__Sam3ImageProcessor as Sam3ImageProcessor, __webpack_exports__Sam3TrackerModel as Sam3TrackerModel, __webpack_exports__SamImageProcessor as SamImageProcessor, __webpack_exports__SamImageSegmentationOutput as SamImageSegmentationOutput, __webpack_exports__SamModel as SamModel, __webpack_exports__SamPreTrainedModel as SamPreTrainedModel, __webpack_exports__SamProcessor as SamProcessor, __webpack_exports__SapiensForDepthEstimation as SapiensForDepthEstimation, __webpack_exports__SapiensForNormalEstimation as SapiensForNormalEstimation, __webpack_exports__SapiensForSemanticSegmentation as SapiensForSemanticSegmentation, __webpack_exports__SapiensPreTrainedModel as SapiensPreTrainedModel, __webpack_exports__SeamlessM4TFeatureExtractor as SeamlessM4TFeatureExtractor, __webpack_exports__SegformerFeatureExtractor as SegformerFeatureExtractor, __webpack_exports__SegformerForImageClassification as SegformerForImageClassification, __webpack_exports__SegformerForSemanticSegmentation as SegformerForSemanticSegmentation, __webpack_exports__SegformerImageProcessor as SegformerImageProcessor, __webpack_exports__SegformerModel as SegformerModel, __webpack_exports__SegformerPreTrainedModel as SegformerPreTrainedModel, __webpack_exports__Seq2SeqLMOutput as Seq2SeqLMOutput, __webpack_exports__SequenceClassifierOutput as SequenceClassifierOutput, __webpack_exports__SiglipImageProcessor as SiglipImageProcessor, __webpack_exports__SiglipModel as SiglipModel, __webpack_exports__SiglipPreTrainedModel as SiglipPreTrainedModel, __webpack_exports__SiglipTextModel as SiglipTextModel, __webpack_exports__SiglipTokenizer as SiglipTokenizer, __webpack_exports__SiglipVisionModel as SiglipVisionModel, __webpack_exports__SmolLM3ForCausalLM as SmolLM3ForCausalLM, __webpack_exports__SmolLM3Model as SmolLM3Model, __webpack_exports__SmolLM3PreTrainedModel as SmolLM3PreTrainedModel, __webpack_exports__SmolVLMForConditionalGeneration as SmolVLMForConditionalGeneration, __webpack_exports__SmolVLMImageProcessor as SmolVLMImageProcessor, __webpack_exports__SmolVLMProcessor as SmolVLMProcessor, __webpack_exports__SnacDecoderModel as SnacDecoderModel, __webpack_exports__SnacEncoderModel as SnacEncoderModel, __webpack_exports__SnacFeatureExtractor as SnacFeatureExtractor, __webpack_exports__SnacModel as SnacModel, __webpack_exports__SnacPreTrainedModel as SnacPreTrainedModel, __webpack_exports__SpeechT5FeatureExtractor as SpeechT5FeatureExtractor, __webpack_exports__SpeechT5ForSpeechToText as SpeechT5ForSpeechToText, __webpack_exports__SpeechT5ForTextToSpeech as SpeechT5ForTextToSpeech, __webpack_exports__SpeechT5HifiGan as SpeechT5HifiGan, __webpack_exports__SpeechT5Model as SpeechT5Model, __webpack_exports__SpeechT5PreTrainedModel as SpeechT5PreTrainedModel, __webpack_exports__SpeechT5Processor as SpeechT5Processor, __webpack_exports__SpeechT5Tokenizer as SpeechT5Tokenizer, __webpack_exports__SqueezeBertForMaskedLM as SqueezeBertForMaskedLM, __webpack_exports__SqueezeBertForQuestionAnswering as SqueezeBertForQuestionAnswering, __webpack_exports__SqueezeBertForSequenceClassification as SqueezeBertForSequenceClassification, __webpack_exports__SqueezeBertModel as SqueezeBertModel, __webpack_exports__SqueezeBertPreTrainedModel as SqueezeBertPreTrainedModel, __webpack_exports__SqueezeBertTokenizer as SqueezeBertTokenizer, __webpack_exports__StableLmForCausalLM as StableLmForCausalLM, __webpack_exports__StableLmModel as StableLmModel, __webpack_exports__StableLmPreTrainedModel as StableLmPreTrainedModel, __webpack_exports__Starcoder2ForCausalLM as Starcoder2ForCausalLM, __webpack_exports__Starcoder2Model as Starcoder2Model, __webpack_exports__Starcoder2PreTrainedModel as Starcoder2PreTrainedModel, __webpack_exports__StoppingCriteria as StoppingCriteria, __webpack_exports__StoppingCriteriaList as StoppingCriteriaList, __webpack_exports__StyleTextToSpeech2Model as StyleTextToSpeech2Model, __webpack_exports__StyleTextToSpeech2PreTrainedModel as StyleTextToSpeech2PreTrainedModel, __webpack_exports__SummarizationPipeline as SummarizationPipeline, __webpack_exports__SupertonicForConditionalGeneration as SupertonicForConditionalGeneration, __webpack_exports__SupertonicPreTrainedModel as SupertonicPreTrainedModel, __webpack_exports__SuppressTokensAtBeginLogitsProcessor as SuppressTokensAtBeginLogitsProcessor, __webpack_exports__Swin2SRForImageSuperResolution as Swin2SRForImageSuperResolution, __webpack_exports__Swin2SRImageProcessor as Swin2SRImageProcessor, __webpack_exports__Swin2SRModel as Swin2SRModel, __webpack_exports__Swin2SRPreTrainedModel as Swin2SRPreTrainedModel, __webpack_exports__SwinForImageClassification as SwinForImageClassification, __webpack_exports__SwinForSemanticSegmentation as SwinForSemanticSegmentation, __webpack_exports__SwinModel as SwinModel, __webpack_exports__SwinPreTrainedModel as SwinPreTrainedModel, __webpack_exports__T5ForConditionalGeneration as T5ForConditionalGeneration, __webpack_exports__T5Model as T5Model, __webpack_exports__T5PreTrainedModel as T5PreTrainedModel, __webpack_exports__T5Tokenizer as T5Tokenizer, __webpack_exports__TableTransformerForObjectDetection as TableTransformerForObjectDetection, __webpack_exports__TableTransformerModel as TableTransformerModel, __webpack_exports__TableTransformerObjectDetectionOutput as TableTransformerObjectDetectionOutput, __webpack_exports__TableTransformerPreTrainedModel as TableTransformerPreTrainedModel, __webpack_exports__TemperatureLogitsWarper as TemperatureLogitsWarper, __webpack_exports__Tensor as Tensor, __webpack_exports__Text2TextGenerationPipeline as Text2TextGenerationPipeline, __webpack_exports__TextClassificationPipeline as TextClassificationPipeline, __webpack_exports__TextGenerationPipeline as TextGenerationPipeline, __webpack_exports__TextStreamer as TextStreamer, __webpack_exports__TextToAudioPipeline as TextToAudioPipeline, __webpack_exports__TokenClassificationPipeline as TokenClassificationPipeline, __webpack_exports__TokenClassifierOutput as TokenClassifierOutput, __webpack_exports__TokenizerModel as TokenizerModel, __webpack_exports__TopKLogitsWarper as TopKLogitsWarper, __webpack_exports__TopPLogitsWarper as TopPLogitsWarper, __webpack_exports__TrOCRForCausalLM as TrOCRForCausalLM, __webpack_exports__TrOCRPreTrainedModel as TrOCRPreTrainedModel, __webpack_exports__TranslationPipeline as TranslationPipeline, __webpack_exports__UltravoxModel as UltravoxModel, __webpack_exports__UltravoxPreTrainedModel as UltravoxPreTrainedModel, __webpack_exports__UltravoxProcessor as UltravoxProcessor, __webpack_exports__UniSpeechForCTC as UniSpeechForCTC, __webpack_exports__UniSpeechForSequenceClassification as UniSpeechForSequenceClassification, __webpack_exports__UniSpeechModel as UniSpeechModel, __webpack_exports__UniSpeechPreTrainedModel as UniSpeechPreTrainedModel, __webpack_exports__UniSpeechSatForAudioFrameClassification as UniSpeechSatForAudioFrameClassification, __webpack_exports__UniSpeechSatForCTC as UniSpeechSatForCTC, __webpack_exports__UniSpeechSatForSequenceClassification as UniSpeechSatForSequenceClassification, __webpack_exports__UniSpeechSatModel as UniSpeechSatModel, __webpack_exports__UniSpeechSatPreTrainedModel as UniSpeechSatPreTrainedModel, __webpack_exports__VLChatProcessor as VLChatProcessor, __webpack_exports__VLMImageProcessor as VLMImageProcessor, __webpack_exports__VaultGemmaForCausalLM as VaultGemmaForCausalLM, __webpack_exports__VaultGemmaModel as VaultGemmaModel, __webpack_exports__VaultGemmaPreTrainedModel as VaultGemmaPreTrainedModel, __webpack_exports__ViTFeatureExtractor as ViTFeatureExtractor, __webpack_exports__ViTForImageClassification as ViTForImageClassification, __webpack_exports__ViTImageProcessor as ViTImageProcessor, __webpack_exports__ViTMAEModel as ViTMAEModel, __webpack_exports__ViTMAEPreTrainedModel as ViTMAEPreTrainedModel, __webpack_exports__ViTMSNForImageClassification as ViTMSNForImageClassification, __webpack_exports__ViTMSNModel as ViTMSNModel, __webpack_exports__ViTMSNPreTrainedModel as ViTMSNPreTrainedModel, __webpack_exports__ViTModel as ViTModel, __webpack_exports__ViTPreTrainedModel as ViTPreTrainedModel, __webpack_exports__VisionEncoderDecoderModel as VisionEncoderDecoderModel, __webpack_exports__VitMatteForImageMatting as VitMatteForImageMatting, __webpack_exports__VitMatteImageProcessor as VitMatteImageProcessor, __webpack_exports__VitMattePreTrainedModel as VitMattePreTrainedModel, __webpack_exports__VitPoseForPoseEstimation as VitPoseForPoseEstimation, __webpack_exports__VitPoseImageProcessor as VitPoseImageProcessor, __webpack_exports__VitPosePreTrainedModel as VitPosePreTrainedModel, __webpack_exports__VitsModel as VitsModel, __webpack_exports__VitsModelOutput as VitsModelOutput, __webpack_exports__VitsPreTrainedModel as VitsPreTrainedModel, __webpack_exports__VitsTokenizer as VitsTokenizer, __webpack_exports__VoxtralForConditionalGeneration as VoxtralForConditionalGeneration, __webpack_exports__VoxtralProcessor as VoxtralProcessor, __webpack_exports__Wav2Vec2BertForCTC as Wav2Vec2BertForCTC, __webpack_exports__Wav2Vec2BertForSequenceClassification as Wav2Vec2BertForSequenceClassification, __webpack_exports__Wav2Vec2BertModel as Wav2Vec2BertModel, __webpack_exports__Wav2Vec2BertPreTrainedModel as Wav2Vec2BertPreTrainedModel, __webpack_exports__Wav2Vec2CTCTokenizer as Wav2Vec2CTCTokenizer, __webpack_exports__Wav2Vec2FeatureExtractor as Wav2Vec2FeatureExtractor, __webpack_exports__Wav2Vec2ForAudioFrameClassification as Wav2Vec2ForAudioFrameClassification, __webpack_exports__Wav2Vec2ForCTC as Wav2Vec2ForCTC, __webpack_exports__Wav2Vec2ForSequenceClassification as Wav2Vec2ForSequenceClassification, __webpack_exports__Wav2Vec2Model as Wav2Vec2Model, __webpack_exports__Wav2Vec2PreTrainedModel as Wav2Vec2PreTrainedModel, __webpack_exports__Wav2Vec2Processor as Wav2Vec2Processor, __webpack_exports__Wav2Vec2ProcessorWithLM as Wav2Vec2ProcessorWithLM, __webpack_exports__WavLMForAudioFrameClassification as WavLMForAudioFrameClassification, __webpack_exports__WavLMForCTC as WavLMForCTC, __webpack_exports__WavLMForSequenceClassification as WavLMForSequenceClassification, __webpack_exports__WavLMForXVector as WavLMForXVector, __webpack_exports__WavLMModel as WavLMModel, __webpack_exports__WavLMPreTrainedModel as WavLMPreTrainedModel, __webpack_exports__WeSpeakerFeatureExtractor as WeSpeakerFeatureExtractor, __webpack_exports__WeSpeakerResNetModel as WeSpeakerResNetModel, __webpack_exports__WeSpeakerResNetPreTrainedModel as WeSpeakerResNetPreTrainedModel, __webpack_exports__WhisperFeatureExtractor as WhisperFeatureExtractor, __webpack_exports__WhisperForConditionalGeneration as WhisperForConditionalGeneration, __webpack_exports__WhisperModel as WhisperModel, __webpack_exports__WhisperPreTrainedModel as WhisperPreTrainedModel, __webpack_exports__WhisperProcessor as WhisperProcessor, __webpack_exports__WhisperTextStreamer as WhisperTextStreamer, __webpack_exports__WhisperTimeStampLogitsProcessor as WhisperTimeStampLogitsProcessor, __webpack_exports__WhisperTokenizer as WhisperTokenizer, __webpack_exports__XLMForQuestionAnswering as XLMForQuestionAnswering, __webpack_exports__XLMForSequenceClassification as XLMForSequenceClassification, __webpack_exports__XLMForTokenClassification as XLMForTokenClassification, __webpack_exports__XLMModel as XLMModel, __webpack_exports__XLMPreTrainedModel as XLMPreTrainedModel, __webpack_exports__XLMRobertaForMaskedLM as XLMRobertaForMaskedLM, __webpack_exports__XLMRobertaForQuestionAnswering as XLMRobertaForQuestionAnswering, __webpack_exports__XLMRobertaForSequenceClassification as XLMRobertaForSequenceClassification, __webpack_exports__XLMRobertaForTokenClassification as XLMRobertaForTokenClassification, __webpack_exports__XLMRobertaModel as XLMRobertaModel, __webpack_exports__XLMRobertaPreTrainedModel as XLMRobertaPreTrainedModel, __webpack_exports__XLMRobertaTokenizer as XLMRobertaTokenizer, __webpack_exports__XLMTokenizer as XLMTokenizer, __webpack_exports__XLMWithLMHeadModel as XLMWithLMHeadModel, __webpack_exports__XVectorOutput as XVectorOutput, __webpack_exports__YolosFeatureExtractor as YolosFeatureExtractor, __webpack_exports__YolosForObjectDetection as YolosForObjectDetection, __webpack_exports__YolosImageProcessor as YolosImageProcessor, __webpack_exports__YolosModel as YolosModel, __webpack_exports__YolosObjectDetectionOutput as YolosObjectDetectionOutput, __webpack_exports__YolosPreTrainedModel as YolosPreTrainedModel, __webpack_exports__ZeroShotAudioClassificationPipeline as ZeroShotAudioClassificationPipeline, __webpack_exports__ZeroShotClassificationPipeline as ZeroShotClassificationPipeline, __webpack_exports__ZeroShotImageClassificationPipeline as ZeroShotImageClassificationPipeline, __webpack_exports__ZeroShotObjectDetectionPipeline as ZeroShotObjectDetectionPipeline, __webpack_exports__bankers_round as bankers_round, __webpack_exports__cat as cat, __webpack_exports__cos_sim as cos_sim, __webpack_exports__dot as dot, __webpack_exports__dynamic_time_warping as dynamic_time_warping, __webpack_exports__env as env, __webpack_exports__full as full, __webpack_exports__full_like as full_like, __webpack_exports__getCacheShapes as getCacheShapes, __webpack_exports__hamming as hamming, __webpack_exports__hanning as hanning, __webpack_exports__interpolate as interpolate, __webpack_exports__interpolate_4d as interpolate_4d, __webpack_exports__interpolate_data as interpolate_data, __webpack_exports__is_chinese_char as is_chinese_char, __webpack_exports__layer_norm as layer_norm, __webpack_exports__load_image as load_image, __webpack_exports__load_video as load_video, __webpack_exports__log_softmax as log_softmax, __webpack_exports__magnitude as magnitude, __webpack_exports__matmul as matmul, __webpack_exports__max as max, __webpack_exports__mean as mean, __webpack_exports__mean_pooling as mean_pooling, __webpack_exports__medianFilter as medianFilter, __webpack_exports__mel_filter_bank as mel_filter_bank, __webpack_exports__min as min, __webpack_exports__ones as ones, __webpack_exports__ones_like as ones_like, __webpack_exports__permute as permute, __webpack_exports__permute_data as permute_data, __webpack_exports__pipeline as pipeline, __webpack_exports__quantize_embeddings as quantize_embeddings, __webpack_exports__rand as rand, __webpack_exports__randn as randn, __webpack_exports__read_audio as read_audio, __webpack_exports__rfft as rfft, __webpack_exports__round as round, __webpack_exports__slice as slice, __webpack_exports__softmax as softmax, __webpack_exports__spectrogram as spectrogram, __webpack_exports__stack as stack, __webpack_exports__std_mean as std_mean, __webpack_exports__topk as topk, __webpack_exports__window_function as window_function, __webpack_exports__zeros as zeros, __webpack_exports__zeros_like as zeros_like };
|