@huggingface/transformers 4.0.0-next.0 → 4.0.0-next.10
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +32 -6
- package/dist/ort-wasm-simd-threaded.jsep.mjs +31 -31
- package/dist/transformers.js +9261 -1709
- package/dist/transformers.min.js +25 -18
- package/dist/transformers.node.cjs +6119 -3174
- package/dist/transformers.node.min.cjs +25 -23
- package/dist/transformers.node.min.mjs +25 -23
- package/dist/transformers.node.mjs +6034 -3168
- package/dist/transformers.web.js +4255 -1381
- package/dist/transformers.web.min.js +23 -19
- package/package.json +6 -6
- package/src/backends/onnx.js +128 -53
- package/src/backends/utils/cacheWasm.js +28 -46
- package/src/cache_utils.js +62 -0
- package/src/configs.js +123 -23
- package/src/env.js +100 -11
- package/src/generation/logits_sampler.js +3 -15
- package/src/generation/parameters.js +1 -1
- package/src/generation/streamers.js +21 -0
- package/src/image_processors_utils.js +29 -23
- package/src/models/afmoe/modeling_afmoe.js +5 -0
- package/src/models/auto/image_processing_auto.js +2 -1
- package/src/models/auto/modeling_auto.js +16 -2
- package/src/models/auto/tokenization_auto.js +2 -1
- package/src/models/chatterbox/modeling_chatterbox.js +1 -1
- package/src/models/chmv2/image_processing_chmv2.js +3 -0
- package/src/models/chmv2/modeling_chmv2.js +4 -0
- package/src/models/clap/feature_extraction_clap.js +2 -1
- package/src/models/cohere2/modeling_cohere2.js +5 -0
- package/src/models/cohere_asr/feature_extraction_cohere_asr.js +117 -0
- package/src/models/cohere_asr/modeling_cohere_asr.js +11 -0
- package/src/models/cohere_asr/processing_cohere_asr.js +55 -0
- package/src/models/cohere_asr/tokenization_cohere_asr.js +3 -0
- package/src/models/deepseek_v3/modeling_deepseek_v3.js +5 -0
- package/src/models/detr/image_processing_detr.js +1 -1
- package/src/models/eurobert/modeling_eurobert.js +41 -0
- package/src/models/feature_extractors.js +3 -0
- package/src/models/gemma3/image_processing_gemma3.js +3 -0
- package/src/models/gemma3/modeling_gemma3.js +4 -1
- package/src/models/gemma3/processing_gemma3.js +45 -0
- package/src/models/gemma3n/modeling_gemma3n.js +2 -0
- package/src/models/glm46v/image_processing_glm46v.js +12 -0
- package/src/models/glm46v/processing_glm46v.js +5 -0
- package/src/models/glm_moe_dsa/modeling_glm_moe_dsa.js +5 -0
- package/src/models/glm_ocr/modeling_glm_ocr.js +78 -0
- package/src/models/granite_speech/feature_extraction_granite_speech.js +58 -0
- package/src/models/granite_speech/modeling_granite_speech.js +5 -0
- package/src/models/granite_speech/processing_granite_speech.js +62 -0
- package/src/models/grounding_dino/image_processing_grounding_dino.js +1 -1
- package/src/models/idefics3/modeling_idefics3.js +5 -32
- package/src/models/image_processors.js +4 -0
- package/src/models/lfm2_vl/image_processing_lfm2_vl.js +305 -0
- package/src/models/lfm2_vl/modeling_lfm2_vl.js +13 -0
- package/src/models/lfm2_vl/processing_lfm2_vl.js +77 -0
- package/src/models/lighton_ocr/modeling_lighton_ocr.js +3 -0
- package/src/models/llava/modeling_llava.js +1 -1
- package/src/models/marian/tokenization_marian.js +3 -2
- package/src/models/mistral3/modeling_mistral3.js +2 -2
- package/src/models/mistral4/modeling_mistral4.js +5 -0
- package/src/models/modeling_utils.js +283 -300
- package/src/models/models.js +26 -1
- package/src/models/nemotron_h/modeling_nemotron_h.js +5 -0
- package/src/models/olmo_hybrid/modeling_olmo_hybrid.js +5 -0
- package/src/models/paligemma/modeling_paligemma.js +2 -25
- package/src/models/paligemma/processing_paligemma.js +3 -2
- package/src/models/processors.js +8 -0
- package/src/models/qwen2_5_vl/modeling_qwen2_5_vl.js +9 -0
- package/src/models/qwen2_5_vl/processing_qwen2_5_vl.js +3 -0
- package/src/models/qwen2_moe/modeling_qwen2_moe.js +5 -0
- package/src/models/qwen2_vl/image_processing_qwen2_vl.js +15 -1
- package/src/models/qwen2_vl/modeling_qwen2_vl.js +240 -143
- package/src/models/qwen2_vl/processing_qwen2_vl.js +5 -4
- package/src/models/qwen3_5/modeling_qwen3_5.js +4 -0
- package/src/models/qwen3_5_moe/modeling_qwen3_5_moe.js +4 -0
- package/src/models/qwen3_moe/modeling_qwen3_moe.js +5 -0
- package/src/models/qwen3_next/modeling_qwen3_next.js +5 -0
- package/src/models/qwen3_vl/modeling_qwen3_vl.js +4 -0
- package/src/models/qwen3_vl/processing_qwen3_vl.js +3 -0
- package/src/models/qwen3_vl_moe/modeling_qwen3_vl_moe.js +4 -0
- package/src/models/registry.js +61 -5
- package/src/models/sam/image_processing_sam.js +1 -1
- package/src/models/session.js +33 -56
- package/src/models/smolvlm/modeling_smolvlm.js +7 -0
- package/src/models/solar_open/modeling_solar_open.js +5 -0
- package/src/models/tokenizers.js +1 -0
- package/src/models/ultravox/modeling_ultravox.js +1 -3
- package/src/models/voxtral/modeling_voxtral.js +3 -0
- package/src/models/voxtral_realtime/feature_extraction_voxtral_realtime.js +71 -0
- package/src/models/voxtral_realtime/modeling_voxtral_realtime.js +239 -0
- package/src/models/voxtral_realtime/processing_voxtral_realtime.js +113 -0
- package/src/models/whisper/feature_extraction_whisper.js +4 -13
- package/src/models/whisper/modeling_whisper.js +6 -5
- package/src/models/xlm/tokenization_xlm.js +2 -1
- package/src/pipelines/automatic-speech-recognition.js +47 -3
- package/src/pipelines/document-question-answering.js +1 -1
- package/src/pipelines/image-to-text.js +2 -2
- package/src/pipelines/index.js +313 -0
- package/src/pipelines/summarization.js +1 -1
- package/src/pipelines/text-generation.js +5 -1
- package/src/pipelines/text-to-audio.js +4 -2
- package/src/pipelines/text2text-generation.js +1 -1
- package/src/pipelines/translation.js +1 -1
- package/src/pipelines/zero-shot-classification.js +3 -2
- package/src/pipelines.js +140 -428
- package/src/tokenization_utils.js +42 -21
- package/src/transformers.js +10 -1
- package/src/utils/audio.js +20 -3
- package/src/utils/cache/CrossOriginStorageCache.js +251 -0
- package/src/utils/cache/FileCache.js +128 -0
- package/src/utils/cache/cross-origin-storage.d.ts +38 -0
- package/src/utils/cache.js +12 -4
- package/src/utils/core.js +23 -1
- package/src/utils/devices.js +22 -0
- package/src/utils/dtypes.js +55 -0
- package/src/utils/hub/{files.js → FileResponse.js} +0 -90
- package/src/utils/hub/utils.js +45 -5
- package/src/utils/hub.js +67 -23
- package/src/utils/image.js +14 -14
- package/src/utils/logger.js +67 -0
- package/src/utils/lru_cache.js +67 -0
- package/src/utils/memoize_promise.js +45 -0
- package/src/utils/model-loader.js +35 -17
- package/src/utils/model_registry/ModelRegistry.js +382 -0
- package/src/utils/model_registry/clear_cache.js +128 -0
- package/src/utils/model_registry/get_available_dtypes.js +68 -0
- package/src/utils/model_registry/get_file_metadata.js +162 -0
- package/src/utils/model_registry/get_files.js +42 -0
- package/src/utils/model_registry/get_model_files.js +114 -0
- package/src/utils/model_registry/get_pipeline_files.js +44 -0
- package/src/utils/model_registry/get_processor_files.js +20 -0
- package/src/utils/model_registry/get_tokenizer_files.js +21 -0
- package/src/utils/model_registry/is_cached.js +169 -0
- package/src/utils/model_registry/resolve_model_type.js +66 -0
- package/src/utils/random.js +225 -0
- package/src/utils/tensor.js +26 -23
- package/src/utils/video.js +2 -2
- package/types/backends/onnx.d.ts.map +1 -1
- package/types/backends/utils/cacheWasm.d.ts +3 -17
- package/types/backends/utils/cacheWasm.d.ts.map +1 -1
- package/types/cache_utils.d.ts +29 -0
- package/types/cache_utils.d.ts.map +1 -0
- package/types/configs.d.ts.map +1 -1
- package/types/env.d.ts +60 -27
- package/types/env.d.ts.map +1 -1
- package/types/generation/logits_sampler.d.ts +2 -2
- package/types/generation/logits_sampler.d.ts.map +1 -1
- package/types/generation/parameters.d.ts +1 -1
- package/types/generation/parameters.d.ts.map +1 -1
- package/types/generation/streamers.d.ts +1 -0
- package/types/generation/streamers.d.ts.map +1 -1
- package/types/image_processors_utils.d.ts +18 -1
- package/types/image_processors_utils.d.ts.map +1 -1
- package/types/models/afmoe/modeling_afmoe.d.ts +8 -0
- package/types/models/afmoe/modeling_afmoe.d.ts.map +1 -0
- package/types/models/{ast/modeling_ast.d.ts → audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.d.ts} +1 -1
- package/types/models/audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.d.ts.map +1 -0
- package/types/models/auto/image_processing_auto.d.ts.map +1 -1
- package/types/models/auto/modeling_auto.d.ts +6 -0
- package/types/models/auto/modeling_auto.d.ts.map +1 -1
- package/types/models/auto/tokenization_auto.d.ts.map +1 -1
- package/types/models/chmv2/image_processing_chmv2.d.ts +4 -0
- package/types/models/chmv2/image_processing_chmv2.d.ts.map +1 -0
- package/types/models/chmv2/modeling_chmv2.d.ts +6 -0
- package/types/models/chmv2/modeling_chmv2.d.ts.map +1 -0
- package/types/models/clap/feature_extraction_clap.d.ts.map +1 -1
- package/types/models/cohere2/modeling_cohere2.d.ts +8 -0
- package/types/models/cohere2/modeling_cohere2.d.ts.map +1 -0
- package/types/models/cohere_asr/feature_extraction_cohere_asr.d.ts +25 -0
- package/types/models/cohere_asr/feature_extraction_cohere_asr.d.ts.map +1 -0
- package/types/models/cohere_asr/modeling_cohere_asr.d.ts +9 -0
- package/types/models/cohere_asr/modeling_cohere_asr.d.ts.map +1 -0
- package/types/models/cohere_asr/processing_cohere_asr.d.ts +27 -0
- package/types/models/cohere_asr/processing_cohere_asr.d.ts.map +1 -0
- package/types/models/cohere_asr/tokenization_cohere_asr.d.ts +4 -0
- package/types/models/cohere_asr/tokenization_cohere_asr.d.ts.map +1 -0
- package/types/models/deepseek_v3/modeling_deepseek_v3.d.ts +8 -0
- package/types/models/deepseek_v3/modeling_deepseek_v3.d.ts.map +1 -0
- package/types/models/detr/image_processing_detr.d.ts +1 -1
- package/types/models/eurobert/modeling_eurobert.d.ts +36 -0
- package/types/models/eurobert/modeling_eurobert.d.ts.map +1 -0
- package/types/models/feature_extractors.d.ts +3 -0
- package/types/models/gemma3/image_processing_gemma3.d.ts +4 -0
- package/types/models/gemma3/image_processing_gemma3.d.ts.map +1 -0
- package/types/models/gemma3/modeling_gemma3.d.ts +4 -1
- package/types/models/gemma3/modeling_gemma3.d.ts.map +1 -1
- package/types/models/gemma3/processing_gemma3.d.ts +20 -0
- package/types/models/gemma3/processing_gemma3.d.ts.map +1 -0
- package/types/models/gemma3n/modeling_gemma3n.d.ts +2 -0
- package/types/models/gemma3n/modeling_gemma3n.d.ts.map +1 -1
- package/types/models/glm46v/image_processing_glm46v.d.ts +4 -0
- package/types/models/glm46v/image_processing_glm46v.d.ts.map +1 -0
- package/types/models/glm46v/processing_glm46v.d.ts +4 -0
- package/types/models/glm46v/processing_glm46v.d.ts.map +1 -0
- package/types/models/glm_moe_dsa/modeling_glm_moe_dsa.d.ts +8 -0
- package/types/models/glm_moe_dsa/modeling_glm_moe_dsa.d.ts.map +1 -0
- package/types/models/glm_ocr/modeling_glm_ocr.d.ts +26 -0
- package/types/models/glm_ocr/modeling_glm_ocr.d.ts.map +1 -0
- package/types/models/granite_speech/feature_extraction_granite_speech.d.ts +16 -0
- package/types/models/granite_speech/feature_extraction_granite_speech.d.ts.map +1 -0
- package/types/models/granite_speech/modeling_granite_speech.d.ts +4 -0
- package/types/models/granite_speech/modeling_granite_speech.d.ts.map +1 -0
- package/types/models/granite_speech/processing_granite_speech.d.ts +19 -0
- package/types/models/granite_speech/processing_granite_speech.d.ts.map +1 -0
- package/types/models/grounding_dino/image_processing_grounding_dino.d.ts +1 -1
- package/types/models/idefics3/modeling_idefics3.d.ts +2 -18
- package/types/models/idefics3/modeling_idefics3.d.ts.map +1 -1
- package/types/models/image_processors.d.ts +4 -0
- package/types/models/lfm2_vl/image_processing_lfm2_vl.d.ts +41 -0
- package/types/models/lfm2_vl/image_processing_lfm2_vl.d.ts.map +1 -0
- package/types/models/lfm2_vl/modeling_lfm2_vl.d.ts +4 -0
- package/types/models/lfm2_vl/modeling_lfm2_vl.d.ts.map +1 -0
- package/types/models/lfm2_vl/processing_lfm2_vl.d.ts +18 -0
- package/types/models/lfm2_vl/processing_lfm2_vl.d.ts.map +1 -0
- package/types/models/lighton_ocr/modeling_lighton_ocr.d.ts +4 -0
- package/types/models/lighton_ocr/modeling_lighton_ocr.d.ts.map +1 -0
- package/types/models/marian/tokenization_marian.d.ts.map +1 -1
- package/types/models/mistral3/modeling_mistral3.d.ts +2 -2
- package/types/models/mistral3/modeling_mistral3.d.ts.map +1 -1
- package/types/models/mistral4/modeling_mistral4.d.ts +8 -0
- package/types/models/mistral4/modeling_mistral4.d.ts.map +1 -0
- package/types/models/modeling_utils.d.ts +46 -27
- package/types/models/modeling_utils.d.ts.map +1 -1
- package/types/models/models.d.ts +26 -1
- package/types/models/nemotron_h/modeling_nemotron_h.d.ts +8 -0
- package/types/models/nemotron_h/modeling_nemotron_h.d.ts.map +1 -0
- package/types/models/olmo_hybrid/modeling_olmo_hybrid.d.ts +8 -0
- package/types/models/olmo_hybrid/modeling_olmo_hybrid.d.ts.map +1 -0
- package/types/models/paligemma/modeling_paligemma.d.ts +2 -8
- package/types/models/paligemma/modeling_paligemma.d.ts.map +1 -1
- package/types/models/paligemma/processing_paligemma.d.ts.map +1 -1
- package/types/models/processors.d.ts +8 -0
- package/types/models/qwen2_5_vl/modeling_qwen2_5_vl.d.ts +7 -0
- package/types/models/qwen2_5_vl/modeling_qwen2_5_vl.d.ts.map +1 -0
- package/types/models/qwen2_5_vl/processing_qwen2_5_vl.d.ts +4 -0
- package/types/models/qwen2_5_vl/processing_qwen2_5_vl.d.ts.map +1 -0
- package/types/models/qwen2_moe/modeling_qwen2_moe.d.ts +8 -0
- package/types/models/qwen2_moe/modeling_qwen2_moe.d.ts.map +1 -0
- package/types/models/qwen2_vl/image_processing_qwen2_vl.d.ts +3 -0
- package/types/models/qwen2_vl/image_processing_qwen2_vl.d.ts.map +1 -1
- package/types/models/qwen2_vl/modeling_qwen2_vl.d.ts +44 -6
- package/types/models/qwen2_vl/modeling_qwen2_vl.d.ts.map +1 -1
- package/types/models/qwen2_vl/processing_qwen2_vl.d.ts +1 -0
- package/types/models/qwen2_vl/processing_qwen2_vl.d.ts.map +1 -1
- package/types/models/qwen3_5/modeling_qwen3_5.d.ts +6 -0
- package/types/models/qwen3_5/modeling_qwen3_5.d.ts.map +1 -0
- package/types/models/qwen3_5_moe/modeling_qwen3_5_moe.d.ts +7 -0
- package/types/models/qwen3_5_moe/modeling_qwen3_5_moe.d.ts.map +1 -0
- package/types/models/qwen3_moe/modeling_qwen3_moe.d.ts +8 -0
- package/types/models/qwen3_moe/modeling_qwen3_moe.d.ts.map +1 -0
- package/types/models/qwen3_next/modeling_qwen3_next.d.ts +8 -0
- package/types/models/qwen3_next/modeling_qwen3_next.d.ts.map +1 -0
- package/types/models/qwen3_vl/modeling_qwen3_vl.d.ts +7 -0
- package/types/models/qwen3_vl/modeling_qwen3_vl.d.ts.map +1 -0
- package/types/models/qwen3_vl/processing_qwen3_vl.d.ts +4 -0
- package/types/models/qwen3_vl/processing_qwen3_vl.d.ts.map +1 -0
- package/types/models/qwen3_vl_moe/modeling_qwen3_vl_moe.d.ts +7 -0
- package/types/models/qwen3_vl_moe/modeling_qwen3_vl_moe.d.ts.map +1 -0
- package/types/models/registry.d.ts +2 -1
- package/types/models/registry.d.ts.map +1 -1
- package/types/models/sam/image_processing_sam.d.ts +1 -1
- package/types/models/session.d.ts +3 -2
- package/types/models/session.d.ts.map +1 -1
- package/types/models/smolvlm/modeling_smolvlm.d.ts +8 -0
- package/types/models/smolvlm/modeling_smolvlm.d.ts.map +1 -0
- package/types/models/solar_open/modeling_solar_open.d.ts +8 -0
- package/types/models/solar_open/modeling_solar_open.d.ts.map +1 -0
- package/types/models/tokenizers.d.ts +1 -0
- package/types/models/ultravox/modeling_ultravox.d.ts +0 -2
- package/types/models/ultravox/modeling_ultravox.d.ts.map +1 -1
- package/types/models/voxtral/modeling_voxtral.d.ts +4 -0
- package/types/models/voxtral/modeling_voxtral.d.ts.map +1 -0
- package/types/models/voxtral_realtime/feature_extraction_voxtral_realtime.d.ts +28 -0
- package/types/models/voxtral_realtime/feature_extraction_voxtral_realtime.d.ts.map +1 -0
- package/types/models/voxtral_realtime/modeling_voxtral_realtime.d.ts +17 -0
- package/types/models/voxtral_realtime/modeling_voxtral_realtime.d.ts.map +1 -0
- package/types/models/voxtral_realtime/processing_voxtral_realtime.d.ts +44 -0
- package/types/models/voxtral_realtime/processing_voxtral_realtime.d.ts.map +1 -0
- package/types/models/whisper/feature_extraction_whisper.d.ts.map +1 -1
- package/types/models/whisper/modeling_whisper.d.ts.map +1 -1
- package/types/models/xlm/tokenization_xlm.d.ts.map +1 -1
- package/types/pipelines/automatic-speech-recognition.d.ts +7 -2
- package/types/pipelines/automatic-speech-recognition.d.ts.map +1 -1
- package/types/pipelines/document-question-answering.d.ts +2 -2
- package/types/pipelines/document-question-answering.d.ts.map +1 -1
- package/types/pipelines/image-to-text.d.ts +4 -4
- package/types/pipelines/image-to-text.d.ts.map +1 -1
- package/types/pipelines/index.d.ts +265 -0
- package/types/pipelines/index.d.ts.map +1 -0
- package/types/pipelines/summarization.d.ts +2 -2
- package/types/pipelines/summarization.d.ts.map +1 -1
- package/types/pipelines/text-generation.d.ts +7 -3
- package/types/pipelines/text-generation.d.ts.map +1 -1
- package/types/pipelines/text-to-audio.d.ts.map +1 -1
- package/types/pipelines/text2text-generation.d.ts +3 -3
- package/types/pipelines/text2text-generation.d.ts.map +1 -1
- package/types/pipelines/translation.d.ts +2 -2
- package/types/pipelines/translation.d.ts.map +1 -1
- package/types/pipelines/zero-shot-classification.d.ts.map +1 -1
- package/types/pipelines.d.ts +51 -291
- package/types/pipelines.d.ts.map +1 -1
- package/types/tokenization_utils.d.ts +44 -26
- package/types/tokenization_utils.d.ts.map +1 -1
- package/types/transformers.d.ts +7 -1
- package/types/transformers.d.ts.map +1 -1
- package/types/utils/audio.d.ts +5 -2
- package/types/utils/audio.d.ts.map +1 -1
- package/types/utils/cache/CrossOriginStorageCache.d.ts +120 -0
- package/types/utils/cache/CrossOriginStorageCache.d.ts.map +1 -0
- package/types/utils/cache/FileCache.d.ts +39 -0
- package/types/utils/cache/FileCache.d.ts.map +1 -0
- package/types/utils/cache.d.ts +10 -4
- package/types/utils/cache.d.ts.map +1 -1
- package/types/utils/core.d.ts +59 -2
- package/types/utils/core.d.ts.map +1 -1
- package/types/utils/devices.d.ts +15 -0
- package/types/utils/devices.d.ts.map +1 -1
- package/types/utils/dtypes.d.ts +17 -1
- package/types/utils/dtypes.d.ts.map +1 -1
- package/types/utils/hub/{files.d.ts → FileResponse.d.ts} +1 -32
- package/types/utils/hub/FileResponse.d.ts.map +1 -0
- package/types/utils/hub/utils.d.ts +19 -3
- package/types/utils/hub/utils.d.ts.map +1 -1
- package/types/utils/hub.d.ts +36 -7
- package/types/utils/hub.d.ts.map +1 -1
- package/types/utils/image.d.ts +1 -1
- package/types/utils/logger.d.ts +28 -0
- package/types/utils/logger.d.ts.map +1 -0
- package/types/utils/lru_cache.d.ts +38 -0
- package/types/utils/lru_cache.d.ts.map +1 -0
- package/types/utils/memoize_promise.d.ts +14 -0
- package/types/utils/memoize_promise.d.ts.map +1 -0
- package/types/utils/model-loader.d.ts +15 -0
- package/types/utils/model-loader.d.ts.map +1 -1
- package/types/utils/model_registry/ModelRegistry.d.ts +298 -0
- package/types/utils/model_registry/ModelRegistry.d.ts.map +1 -0
- package/types/utils/model_registry/clear_cache.d.ts +74 -0
- package/types/utils/model_registry/clear_cache.d.ts.map +1 -0
- package/types/utils/model_registry/get_available_dtypes.d.ts +26 -0
- package/types/utils/model_registry/get_available_dtypes.d.ts.map +1 -0
- package/types/utils/model_registry/get_file_metadata.d.ts +20 -0
- package/types/utils/model_registry/get_file_metadata.d.ts.map +1 -0
- package/types/utils/model_registry/get_files.d.ts +23 -0
- package/types/utils/model_registry/get_files.d.ts.map +1 -0
- package/types/utils/model_registry/get_model_files.d.ts +48 -0
- package/types/utils/model_registry/get_model_files.d.ts.map +1 -0
- package/types/utils/model_registry/get_pipeline_files.d.ts +22 -0
- package/types/utils/model_registry/get_pipeline_files.d.ts.map +1 -0
- package/types/utils/model_registry/get_processor_files.d.ts +9 -0
- package/types/utils/model_registry/get_processor_files.d.ts.map +1 -0
- package/types/utils/model_registry/get_tokenizer_files.d.ts +9 -0
- package/types/utils/model_registry/get_tokenizer_files.d.ts.map +1 -0
- package/types/utils/model_registry/is_cached.d.ts +105 -0
- package/types/utils/model_registry/is_cached.d.ts.map +1 -0
- package/types/utils/model_registry/resolve_model_type.d.ts +24 -0
- package/types/utils/model_registry/resolve_model_type.d.ts.map +1 -0
- package/types/utils/random.d.ts +86 -0
- package/types/utils/random.d.ts.map +1 -0
- package/types/utils/tensor.d.ts.map +1 -1
- package/src/utils/data-structures.js +0 -572
- package/types/models/ast/modeling_ast.d.ts.map +0 -1
- package/types/utils/data-structures.d.ts +0 -294
- package/types/utils/data-structures.d.ts.map +0 -1
- package/types/utils/hub/files.d.ts.map +0 -1
- /package/src/models/{ast/modeling_ast.js → audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.js} +0 -0
|
@@ -0,0 +1,8 @@
|
|
|
1
|
+
export class GlmMoeDsaPreTrainedModel extends PreTrainedModel {
|
|
2
|
+
}
|
|
3
|
+
export class GlmMoeDsaModel extends GlmMoeDsaPreTrainedModel {
|
|
4
|
+
}
|
|
5
|
+
export class GlmMoeDsaForCausalLM extends GlmMoeDsaPreTrainedModel {
|
|
6
|
+
}
|
|
7
|
+
import { PreTrainedModel } from '../modeling_utils.js';
|
|
8
|
+
//# sourceMappingURL=modeling_glm_moe_dsa.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"modeling_glm_moe_dsa.d.ts","sourceRoot":"","sources":["../../../src/models/glm_moe_dsa/modeling_glm_moe_dsa.js"],"names":[],"mappings":"AAEA;CAAgE;AAChE;CAA+D;AAC/D;CAAqE;gCAJrC,sBAAsB"}
|
|
@@ -0,0 +1,26 @@
|
|
|
1
|
+
export class GlmOcrForConditionalGeneration extends Qwen2_5_VLForConditionalGeneration {
|
|
2
|
+
/**
|
|
3
|
+
* Compute 3D positional indices for vision tokens.
|
|
4
|
+
* Temporal is constant, height is repeat-interleaved, width tiles.
|
|
5
|
+
* @param {number} start_position
|
|
6
|
+
* @param {number[]} grid_thw [T, H, W]
|
|
7
|
+
* @param {number} temp_merge_size
|
|
8
|
+
* @param {number} spatial_merge_size
|
|
9
|
+
* @returns {number[]} Flat array of length 3 * seq_len: [temporal..., height..., width...]
|
|
10
|
+
*/
|
|
11
|
+
get_vision_position_ids(start_position: number, grid_thw: number[], temp_merge_size: number, spatial_merge_size: number): number[];
|
|
12
|
+
/**
|
|
13
|
+
* GlmOcr uses mm_token_type_ids-style grouping (image tokens identified by image_token_id)
|
|
14
|
+
* instead of vision_start_token_id scanning used by Qwen2VL.
|
|
15
|
+
* After a vision segment, position advances by max(h, w) / spatial_merge_size.
|
|
16
|
+
*/
|
|
17
|
+
_get_multimodal_rope_positions({ filtered_ids, image_grid_thw_list, video_grid_thw_list, spatial_merge_size, state, }: {
|
|
18
|
+
filtered_ids: any;
|
|
19
|
+
image_grid_thw_list: any;
|
|
20
|
+
video_grid_thw_list: any;
|
|
21
|
+
spatial_merge_size: any;
|
|
22
|
+
state: any;
|
|
23
|
+
}): number[][];
|
|
24
|
+
}
|
|
25
|
+
import { Qwen2_5_VLForConditionalGeneration } from '../qwen2_5_vl/modeling_qwen2_5_vl.js';
|
|
26
|
+
//# sourceMappingURL=modeling_glm_ocr.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"modeling_glm_ocr.d.ts","sourceRoot":"","sources":["../../../src/models/glm_ocr/modeling_glm_ocr.js"],"names":[],"mappings":"AAEA;IACI;;;;;;;;OAQG;IACH,wCANW,MAAM,YACN,MAAM,EAAE,mBACR,MAAM,sBACN,MAAM,GACJ,MAAM,EAAE,CAgBpB;IAED;;;;OAIG;IACH;;;;;;mBA2CC;CACJ;mDA7EkD,sCAAsC"}
|
|
@@ -0,0 +1,16 @@
|
|
|
1
|
+
export class GraniteSpeechFeatureExtractor extends FeatureExtractor {
|
|
2
|
+
constructor(config: any);
|
|
3
|
+
mel_filters: number[][];
|
|
4
|
+
window: Float64Array<any>;
|
|
5
|
+
/**
|
|
6
|
+
* Extract mel spectrogram features from audio, matching the Python GraniteSpeechFeatureExtractor.
|
|
7
|
+
* @param {Float32Array|Float64Array} audio The audio waveform.
|
|
8
|
+
* @returns {Promise<{input_features: Tensor}>}
|
|
9
|
+
*/
|
|
10
|
+
_call(audio: Float32Array | Float64Array): Promise<{
|
|
11
|
+
input_features: Tensor;
|
|
12
|
+
}>;
|
|
13
|
+
}
|
|
14
|
+
import { FeatureExtractor } from '../../feature_extraction_utils.js';
|
|
15
|
+
import { Tensor } from '../../utils/tensor.js';
|
|
16
|
+
//# sourceMappingURL=feature_extraction_granite_speech.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"feature_extraction_granite_speech.d.ts","sourceRoot":"","sources":["../../../src/models/granite_speech/feature_extraction_granite_speech.js"],"names":[],"mappings":"AAIA;IACI,yBAsBC;IAhBG,wBAQC;IAKD,0BAAqC;IAKzC;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY,GACvB,OAAO,CAAC;QAAC,cAAc,EAAE,MAAM,CAAA;KAAC,CAAC,CAwB7C;CACJ;iCAzDuD,mCAAmC;uBAEpE,uBAAuB"}
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"modeling_granite_speech.d.ts","sourceRoot":"","sources":["../../../src/models/granite_speech/modeling_granite_speech.js"],"names":[],"mappings":"AAEA;CAEC;8BAJ6B,kCAAkC"}
|
|
@@ -0,0 +1,19 @@
|
|
|
1
|
+
export class GraniteSpeechProcessor extends Processor {
|
|
2
|
+
static tokenizer_class: typeof AutoTokenizer;
|
|
3
|
+
static feature_extractor_class: typeof AutoFeatureExtractor;
|
|
4
|
+
/**
|
|
5
|
+
* Compute the number of audio tokens for a given raw audio length.
|
|
6
|
+
* @param {number} audioLength Raw audio sample count.
|
|
7
|
+
* @returns {number} Number of projector output tokens.
|
|
8
|
+
*/
|
|
9
|
+
_get_num_audio_features(audioLength: number): number;
|
|
10
|
+
/**
|
|
11
|
+
* @param {string} text The text input to process.
|
|
12
|
+
* @param {Float32Array} audio The audio input to process.
|
|
13
|
+
*/
|
|
14
|
+
_call(text: string, audio?: Float32Array, kwargs?: {}): Promise<any>;
|
|
15
|
+
}
|
|
16
|
+
import { Processor } from '../../processing_utils.js';
|
|
17
|
+
import { AutoTokenizer } from '../auto/tokenization_auto.js';
|
|
18
|
+
import { AutoFeatureExtractor } from '../auto/feature_extraction_auto.js';
|
|
19
|
+
//# sourceMappingURL=processing_granite_speech.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"processing_granite_speech.d.ts","sourceRoot":"","sources":["../../../src/models/granite_speech/processing_granite_speech.js"],"names":[],"mappings":"AAKA;IACI,6CAAuC;IACvC,4DAAsD;IAGtD;;;;OAIG;IACH,qCAHW,MAAM,GACJ,MAAM,CAUlB;IAED;;;OAGG;IACH,YAHW,MAAM,UACN,YAAY,6BAiCtB;CACJ;0BA3DyB,2BAA2B;8BADvB,8BAA8B;qCADvB,oCAAoC"}
|
|
@@ -1,5 +1,5 @@
|
|
|
1
1
|
/**
|
|
2
|
-
* @typedef {
|
|
2
|
+
* @typedef {Object} GroundingDinoFeatureExtractorResultProps
|
|
3
3
|
* @property {import('../../utils/tensor.js').Tensor} pixel_mask
|
|
4
4
|
* @typedef {import('../../image_processors_utils.js').ImageProcessorResult & GroundingDinoFeatureExtractorResultProps} GroundingDinoFeatureExtractorResult
|
|
5
5
|
*/
|
|
@@ -1,23 +1,7 @@
|
|
|
1
|
-
export class Idefics3PreTrainedModel extends PreTrainedModel {
|
|
2
|
-
}
|
|
3
1
|
/**
|
|
4
2
|
* The Idefics3 model which consists of a vision backbone and a language model.
|
|
5
3
|
*/
|
|
6
|
-
export class Idefics3ForConditionalGeneration extends
|
|
7
|
-
encode_image({ pixel_values, pixel_attention_mask }: {
|
|
8
|
-
pixel_values: any;
|
|
9
|
-
pixel_attention_mask: any;
|
|
10
|
-
}): Promise<any>;
|
|
11
|
-
_merge_input_ids_with_image_features(kwargs: any): {
|
|
12
|
-
inputs_embeds: any;
|
|
13
|
-
attention_mask: any;
|
|
14
|
-
};
|
|
15
|
-
}
|
|
16
|
-
/**
|
|
17
|
-
* The SmolVLM Model with a language modeling head.
|
|
18
|
-
* It is made up a SigLIP vision encoder, with a language modeling head on top.
|
|
19
|
-
*/
|
|
20
|
-
export class SmolVLMForConditionalGeneration extends Idefics3ForConditionalGeneration {
|
|
4
|
+
export class Idefics3ForConditionalGeneration extends LlavaForConditionalGeneration {
|
|
21
5
|
}
|
|
22
|
-
import {
|
|
6
|
+
import { LlavaForConditionalGeneration } from '../llava/modeling_llava.js';
|
|
23
7
|
//# sourceMappingURL=modeling_idefics3.d.ts.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"modeling_idefics3.d.ts","sourceRoot":"","sources":["../../../src/models/idefics3/modeling_idefics3.js"],"names":[],"mappings":"
|
|
1
|
+
{"version":3,"file":"modeling_idefics3.d.ts","sourceRoot":"","sources":["../../../src/models/idefics3/modeling_idefics3.js"],"names":[],"mappings":"AAEA;;GAEG;AACH;CASC;8CAd6C,4BAA4B"}
|
|
@@ -1,6 +1,7 @@
|
|
|
1
1
|
export * from "./beit/image_processing_beit.js";
|
|
2
2
|
export * from "./bit/image_processing_bit.js";
|
|
3
3
|
export * from "./chinese_clip/image_processing_chinese_clip.js";
|
|
4
|
+
export * from "./chmv2/image_processing_chmv2.js";
|
|
4
5
|
export * from "./clip/image_processing_clip.js";
|
|
5
6
|
export * from "./convnext/image_processing_convnext.js";
|
|
6
7
|
export * from "./deit/image_processing_deit.js";
|
|
@@ -9,11 +10,14 @@ export * from "./dinov3_vit/image_processing_dinov3_vit.js";
|
|
|
9
10
|
export * from "./donut/image_processing_donut.js";
|
|
10
11
|
export * from "./dpt/image_processing_dpt.js";
|
|
11
12
|
export * from "./efficientnet/image_processing_efficientnet.js";
|
|
13
|
+
export * from "./gemma3/image_processing_gemma3.js";
|
|
14
|
+
export * from "./glm46v/image_processing_glm46v.js";
|
|
12
15
|
export * from "./glpn/image_processing_glpn.js";
|
|
13
16
|
export * from "./grounding_dino/image_processing_grounding_dino.js";
|
|
14
17
|
export * from "./idefics3/image_processing_idefics3.js";
|
|
15
18
|
export * from "./janus/image_processing_janus.js";
|
|
16
19
|
export * from "./jina_clip/image_processing_jina_clip.js";
|
|
20
|
+
export * from "./lfm2_vl/image_processing_lfm2_vl.js";
|
|
17
21
|
export * from "./llava_onevision/image_processing_llava_onevision.js";
|
|
18
22
|
export * from "./mask2former/image_processing_mask2former.js";
|
|
19
23
|
export * from "./maskformer/image_processing_maskformer.js";
|
|
@@ -0,0 +1,41 @@
|
|
|
1
|
+
export class Lfm2VlImageProcessor extends ImageProcessor {
|
|
2
|
+
constructor(config: Record<string, any>);
|
|
3
|
+
downsample_factor: any;
|
|
4
|
+
do_image_splitting: any;
|
|
5
|
+
min_tiles: any;
|
|
6
|
+
max_tiles: any;
|
|
7
|
+
use_thumbnail: any;
|
|
8
|
+
min_image_tokens: any;
|
|
9
|
+
max_image_tokens: any;
|
|
10
|
+
encoder_patch_size: any;
|
|
11
|
+
tile_size: any;
|
|
12
|
+
max_pixels_tolerance: any;
|
|
13
|
+
return_row_col_info: any;
|
|
14
|
+
max_num_patches: number;
|
|
15
|
+
/**
|
|
16
|
+
* Check if the image is too large to be processed as a single tile.
|
|
17
|
+
* @param {number} height
|
|
18
|
+
* @param {number} width
|
|
19
|
+
* @returns {boolean}
|
|
20
|
+
*/
|
|
21
|
+
_is_image_too_large(height: number, width: number): boolean;
|
|
22
|
+
/**
|
|
23
|
+
* Get the grid layout for tiling a large image.
|
|
24
|
+
* @param {number} height
|
|
25
|
+
* @param {number} width
|
|
26
|
+
* @returns {{ grid_width: number, grid_height: number, target_width: number, target_height: number }}
|
|
27
|
+
*/
|
|
28
|
+
_get_grid_layout(height: number, width: number): {
|
|
29
|
+
grid_width: number;
|
|
30
|
+
grid_height: number;
|
|
31
|
+
target_width: number;
|
|
32
|
+
target_height: number;
|
|
33
|
+
};
|
|
34
|
+
/** @param {RawImage|RawImage[]|RawImage[][]} images */
|
|
35
|
+
_call(images: RawImage | RawImage[] | RawImage[][], { return_row_col_info }?: {
|
|
36
|
+
return_row_col_info?: any;
|
|
37
|
+
}): Promise<Record<string, any>>;
|
|
38
|
+
}
|
|
39
|
+
export type RawImage = import("../../utils/image.js").RawImage;
|
|
40
|
+
import { ImageProcessor } from '../../image_processors_utils.js';
|
|
41
|
+
//# sourceMappingURL=image_processing_lfm2_vl.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"image_processing_lfm2_vl.d.ts","sourceRoot":"","sources":["../../../src/models/lfm2_vl/image_processing_lfm2_vl.js"],"names":[],"mappings":"AAiIA;IACI,oBAAuB,MAAM,CAAC,MAAM,EAAE,GAAG,CAAC,EAiBzC;IAfG,uBAAsD;IACtD,wBAA2D;IAC3D,eAAsC;IACtC,eAAuC;IACvC,mBAAiD;IACjD,sBAAqD;IACrD,sBAAsD;IACtD,wBAA8E;IAC9E,eAAwC;IACxC,0BAA8D;IAC9D,yBAA8D;IAI9D,wBAAyE;IAG7E;;;;;OAKG;IACH,4BAJW,MAAM,SACN,MAAM,GACJ,OAAO,CAUnB;IAED;;;;;OAKG;IACH,yBAJW,MAAM,SACN,MAAM,GACJ;QAAE,UAAU,EAAE,MAAM,CAAC;QAAC,WAAW,EAAE,MAAM,CAAC;QAAC,YAAY,EAAE,MAAM,CAAC;QAAC,aAAa,EAAE,MAAM,CAAA;KAAE,CAiBpG;IAED,uDAAuD;IAEvD,cAFY,QAAQ,GAAC,QAAQ,EAAE,GAAC,QAAQ,EAAE,EAAE;;qCAmH3C;CACJ;uBA5SY,OAAO,sBAAsB,EAAE,QAAQ;+BAJP,iCAAiC"}
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"modeling_lfm2_vl.d.ts","sourceRoot":"","sources":["../../../src/models/lfm2_vl/modeling_lfm2_vl.js"],"names":[],"mappings":"AAEA;CAUC;8CAZ6C,4BAA4B"}
|
|
@@ -0,0 +1,18 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* @typedef {import('../../utils/image.js').RawImage} RawImage
|
|
3
|
+
*/
|
|
4
|
+
export class Lfm2VlProcessor extends Processor {
|
|
5
|
+
static tokenizer_class: typeof AutoTokenizer;
|
|
6
|
+
static image_processor_class: typeof AutoImageProcessor;
|
|
7
|
+
/**
|
|
8
|
+
* @param {RawImage|RawImage[]} images
|
|
9
|
+
* @param {string|string[]|null} [text]
|
|
10
|
+
* @param {Record<string, any>} [kwargs]
|
|
11
|
+
*/
|
|
12
|
+
_call(images: RawImage | RawImage[], text?: string | string[] | null, kwargs?: Record<string, any>): Promise<any>;
|
|
13
|
+
}
|
|
14
|
+
export type RawImage = import("../../utils/image.js").RawImage;
|
|
15
|
+
import { Processor } from '../../processing_utils.js';
|
|
16
|
+
import { AutoTokenizer } from '../auto/tokenization_auto.js';
|
|
17
|
+
import { AutoImageProcessor } from '../auto/image_processing_auto.js';
|
|
18
|
+
//# sourceMappingURL=processing_lfm2_vl.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"processing_lfm2_vl.d.ts","sourceRoot":"","sources":["../../../src/models/lfm2_vl/processing_lfm2_vl.js"],"names":[],"mappings":"AAIA;;GAEG;AAEH;IACI,6CAAuC;IACvC,wDAAkD;IAElD;;;;OAIG;IACH,cAJW,QAAQ,GAAC,QAAQ,EAAE,SACnB,MAAM,GAAC,MAAM,EAAE,GAAC,IAAI,WACpB,MAAM,CAAC,MAAM,EAAE,GAAG,CAAC,gBA4D7B;CACJ;uBAvEY,OAAO,sBAAsB,EAAE,QAAQ;0BAL1B,2BAA2B;8BAEvB,8BAA8B;mCADzB,kCAAkC"}
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"modeling_lighton_ocr.d.ts","sourceRoot":"","sources":["../../../src/models/lighton_ocr/modeling_lighton_ocr.js"],"names":[],"mappings":"AAEA;CAAwF;8CAF1C,4BAA4B"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"tokenization_marian.d.ts","sourceRoot":"","sources":["../../../src/models/marian/tokenization_marian.js"],"names":[],"mappings":"
|
|
1
|
+
{"version":3,"file":"tokenization_marian.d.ts","sourceRoot":"","sources":["../../../src/models/marian/tokenization_marian.js"],"names":[],"mappings":"AAIA;;;GAGG;AACH;IASQ,sBAAqC;IAErC,mCAA6G;CAoCpH;oCAvDmC,6BAA6B"}
|
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
export class Mistral3ForConditionalGeneration extends
|
|
1
|
+
export class Mistral3ForConditionalGeneration extends LlavaForConditionalGeneration {
|
|
2
2
|
}
|
|
3
|
-
import {
|
|
3
|
+
import { LlavaForConditionalGeneration } from '../llava/modeling_llava.js';
|
|
4
4
|
//# sourceMappingURL=modeling_mistral3.d.ts.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"modeling_mistral3.d.ts","sourceRoot":"","sources":["../../../src/models/mistral3/modeling_mistral3.js"],"names":[],"mappings":"AAEA;
|
|
1
|
+
{"version":3,"file":"modeling_mistral3.d.ts","sourceRoot":"","sources":["../../../src/models/mistral3/modeling_mistral3.js"],"names":[],"mappings":"AAEA;CAAsF;8CAFxC,4BAA4B"}
|
|
@@ -0,0 +1,8 @@
|
|
|
1
|
+
export class Mistral4PreTrainedModel extends PreTrainedModel {
|
|
2
|
+
}
|
|
3
|
+
export class Mistral4Model extends Mistral4PreTrainedModel {
|
|
4
|
+
}
|
|
5
|
+
export class Mistral4ForCausalLM extends Mistral4PreTrainedModel {
|
|
6
|
+
}
|
|
7
|
+
import { PreTrainedModel } from '../modeling_utils.js';
|
|
8
|
+
//# sourceMappingURL=modeling_mistral4.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"modeling_mistral4.d.ts","sourceRoot":"","sources":["../../../src/models/mistral4/modeling_mistral4.js"],"names":[],"mappings":"AAEA;CAA+D;AAC/D;CAA6D;AAC7D;CAAmE;gCAJnC,sBAAsB"}
|
|
@@ -10,6 +10,18 @@ export function registerTaskMappings(mappings: any): void;
|
|
|
10
10
|
* @private
|
|
11
11
|
*/
|
|
12
12
|
export function boolTensor(value: boolean): Tensor;
|
|
13
|
+
/**
|
|
14
|
+
* Get the session configuration for a given model type.
|
|
15
|
+
* @param {number} modelType The model type enum value.
|
|
16
|
+
* @param {Object} config The model config.
|
|
17
|
+
* @param {Object} [options] Loading options.
|
|
18
|
+
* @returns {{ sessions: Record<string, string>, cache_sessions?: Record<string, true>, optional_configs?: Record<string, string> }}
|
|
19
|
+
*/
|
|
20
|
+
export function getSessionsConfig(modelType: number, config: any, options?: any): {
|
|
21
|
+
sessions: Record<string, string>;
|
|
22
|
+
cache_sessions?: Record<string, true>;
|
|
23
|
+
optional_configs?: Record<string, string>;
|
|
24
|
+
};
|
|
13
25
|
/**
|
|
14
26
|
* Perform forward pass on the seq2seq model (both encoder and decoder).
|
|
15
27
|
* @param {Object} self The seq2seq model object.
|
|
@@ -41,28 +53,28 @@ export function decoder_forward(self: any, model_inputs: any, is_encoder_decoder
|
|
|
41
53
|
* @param {Object} params Additional parameters.
|
|
42
54
|
* @param {Function} [params.encode_function] The function to encode the modality values.
|
|
43
55
|
* @param {Function} [params.merge_function] The function to merge the modality features with the input embeddings.
|
|
44
|
-
* @param {string} [params.
|
|
56
|
+
* @param {string[]} [params.modality_input_names] The modality input name.
|
|
45
57
|
* @param {string} [params.modality_output_name] The modality output name.
|
|
46
58
|
* @param {Tensor} [params.input_ids=null]
|
|
47
59
|
* @param {Tensor} [params.attention_mask=null]
|
|
48
60
|
* @param {Tensor} [params.position_ids=null]
|
|
49
61
|
* @param {Tensor} [params.inputs_embeds=null]
|
|
50
|
-
* @param {
|
|
62
|
+
* @param {DynamicCache} [params.past_key_values=null]
|
|
51
63
|
* @param {Object} [params.generation_config=null]
|
|
52
64
|
* @param {Object} [params.logits_processor=null]
|
|
53
65
|
* @returns {Promise<Tensor>} The model's output tensor
|
|
54
66
|
* @private
|
|
55
67
|
*/
|
|
56
|
-
export function generic_text_to_text_forward(self: any, { encode_function, merge_function,
|
|
68
|
+
export function generic_text_to_text_forward(self: any, { encode_function, merge_function, modality_input_names, modality_output_name, input_ids, attention_mask, position_ids, inputs_embeds, past_key_values, generation_config, logits_processor, ...kwargs }: {
|
|
57
69
|
encode_function?: Function;
|
|
58
70
|
merge_function?: Function;
|
|
59
|
-
|
|
71
|
+
modality_input_names?: string[];
|
|
60
72
|
modality_output_name?: string;
|
|
61
73
|
input_ids?: Tensor;
|
|
62
74
|
attention_mask?: Tensor;
|
|
63
75
|
position_ids?: Tensor;
|
|
64
76
|
inputs_embeds?: Tensor;
|
|
65
|
-
past_key_values?:
|
|
77
|
+
past_key_values?: DynamicCache;
|
|
66
78
|
generation_config?: any;
|
|
67
79
|
logits_processor?: any;
|
|
68
80
|
}): Promise<Tensor>;
|
|
@@ -168,6 +180,7 @@ export namespace MODEL_TYPES {
|
|
|
168
180
|
let ImageAudioTextToText: number;
|
|
169
181
|
let Supertonic: number;
|
|
170
182
|
let Chatterbox: number;
|
|
183
|
+
let VoxtralRealtime: number;
|
|
171
184
|
}
|
|
172
185
|
export const MODEL_TYPE_MAPPING: Map<any, any>;
|
|
173
186
|
export const MODEL_NAME_TO_CLASS_MAPPING: Map<any, any>;
|
|
@@ -209,8 +222,8 @@ export class PreTrainedModel extends PreTrainedModel_base {
|
|
|
209
222
|
config: import("../configs.js").PretrainedConfig;
|
|
210
223
|
sessions: Record<string, any>;
|
|
211
224
|
configs: Record<string, any>;
|
|
212
|
-
can_generate:
|
|
213
|
-
_forward:
|
|
225
|
+
can_generate: any;
|
|
226
|
+
_forward: any;
|
|
214
227
|
_prepare_inputs_for_generation: any;
|
|
215
228
|
/** @type {import('../configs.js').TransformersJSConfig} */
|
|
216
229
|
custom_config: import("../configs.js").TransformersJSConfig;
|
|
@@ -257,9 +270,9 @@ export class PreTrainedModel extends PreTrainedModel_base {
|
|
|
257
270
|
/**
|
|
258
271
|
*
|
|
259
272
|
* @param {GenerationConfig} generation_config
|
|
260
|
-
* @param {StoppingCriteriaList} [stopping_criteria=null]
|
|
273
|
+
* @param {import('../generation/stopping_criteria.js').StoppingCriteria|import('../generation/stopping_criteria.js').StoppingCriteria[]|StoppingCriteriaList} [stopping_criteria=null]
|
|
261
274
|
*/
|
|
262
|
-
_get_stopping_criteria(generation_config: GenerationConfig, stopping_criteria?: StoppingCriteriaList): StoppingCriteriaList;
|
|
275
|
+
_get_stopping_criteria(generation_config: GenerationConfig, stopping_criteria?: import("../generation/stopping_criteria.js").StoppingCriteria | import("../generation/stopping_criteria.js").StoppingCriteria[] | StoppingCriteriaList): StoppingCriteriaList;
|
|
263
276
|
/**
|
|
264
277
|
* Confirms that the model class is compatible with generation.
|
|
265
278
|
* If not, raises an exception that points to the right class to use.
|
|
@@ -287,7 +300,7 @@ export class PreTrainedModel extends PreTrainedModel_base {
|
|
|
287
300
|
* @param {Tensor} [params.inputs=null]
|
|
288
301
|
* @param {number} [params.bos_token_id=null]
|
|
289
302
|
* @param {Record<string, Tensor|number[]>} [params.model_kwargs]
|
|
290
|
-
* @returns {{inputs_tensor: Tensor, model_inputs: Record<string, Tensor
|
|
303
|
+
* @returns {{inputs_tensor: Tensor, model_inputs: Record<string, Tensor> & {past_key_values?: DynamicCache}, model_input_name: string}} The model-specific inputs for generation.
|
|
291
304
|
*/
|
|
292
305
|
_prepare_model_inputs({ inputs, bos_token_id, model_kwargs }: {
|
|
293
306
|
inputs?: Tensor;
|
|
@@ -295,7 +308,9 @@ export class PreTrainedModel extends PreTrainedModel_base {
|
|
|
295
308
|
model_kwargs?: Record<string, Tensor | number[]>;
|
|
296
309
|
}): {
|
|
297
310
|
inputs_tensor: Tensor;
|
|
298
|
-
model_inputs: Record<string, Tensor
|
|
311
|
+
model_inputs: Record<string, Tensor> & {
|
|
312
|
+
past_key_values?: DynamicCache;
|
|
313
|
+
};
|
|
299
314
|
model_input_name: string;
|
|
300
315
|
};
|
|
301
316
|
_prepare_encoder_decoder_kwargs_for_generation({ inputs_tensor, model_inputs, model_input_name, generation_config, }: {
|
|
@@ -319,13 +334,14 @@ export class PreTrainedModel extends PreTrainedModel_base {
|
|
|
319
334
|
*/
|
|
320
335
|
generate({ inputs, generation_config, logits_processor, stopping_criteria, streamer, ...kwargs }: import("../generation/parameters.js").GenerationFunctionParameters): Promise<ModelOutput | Tensor>;
|
|
321
336
|
/**
|
|
322
|
-
* Returns
|
|
337
|
+
* Returns a DynamicCache containing past key values from the given decoder results object.
|
|
323
338
|
*
|
|
324
339
|
* @param {Object} decoderResults The decoder results object.
|
|
325
|
-
* @param {
|
|
326
|
-
* @
|
|
340
|
+
* @param {DynamicCache} pastKeyValues The previous past key values.
|
|
341
|
+
* @param {boolean} [disposeEncoderPKVs=false] Whether to dispose encoder past key values.
|
|
342
|
+
* @returns {DynamicCache} A new DynamicCache containing the updated past key values.
|
|
327
343
|
*/
|
|
328
|
-
getPastKeyValues(decoderResults: any, pastKeyValues:
|
|
344
|
+
getPastKeyValues(decoderResults: any, pastKeyValues: DynamicCache, disposeEncoderPKVs?: boolean): DynamicCache;
|
|
329
345
|
/**
|
|
330
346
|
* Returns an object containing attentions from the given model output object.
|
|
331
347
|
*
|
|
@@ -338,22 +354,25 @@ export class PreTrainedModel extends PreTrainedModel_base {
|
|
|
338
354
|
/**
|
|
339
355
|
* Adds past key values to the decoder feeds object. If pastKeyValues is null, creates new tensors for past key values.
|
|
340
356
|
*
|
|
341
|
-
* @param {
|
|
342
|
-
* @param {
|
|
357
|
+
* @param {Record<string, any>} decoderFeeds The decoder feeds object to add past key values to.
|
|
358
|
+
* @param {DynamicCache|null} pastKeyValues The cache containing past key values.
|
|
343
359
|
*/
|
|
344
|
-
addPastKeyValues(decoderFeeds:
|
|
345
|
-
|
|
346
|
-
|
|
347
|
-
|
|
348
|
-
|
|
349
|
-
|
|
350
|
-
|
|
351
|
-
|
|
352
|
-
|
|
353
|
-
|
|
360
|
+
addPastKeyValues(decoderFeeds: Record<string, any>, pastKeyValues: DynamicCache | null): void;
|
|
361
|
+
/**
|
|
362
|
+
* Helper function to select valid inputs and run through the appropriate encoder (vision, text, audio) based on the input type.
|
|
363
|
+
* @param {string} sessionName
|
|
364
|
+
* @param {Record<string, Tensor>} inputs
|
|
365
|
+
* @param {string} outputName
|
|
366
|
+
* @private
|
|
367
|
+
*/
|
|
368
|
+
private _encode_input;
|
|
369
|
+
encode_image(inputs: any): Promise<any>;
|
|
370
|
+
encode_text(inputs: any): Promise<any>;
|
|
371
|
+
encode_audio(inputs: any): Promise<any>;
|
|
354
372
|
}
|
|
355
373
|
import { Tensor } from '../utils/tensor.js';
|
|
356
374
|
import { Seq2SeqLMOutput } from './modeling_outputs.js';
|
|
375
|
+
import { DynamicCache } from '../cache_utils.js';
|
|
357
376
|
import { GenerationConfig } from '../generation/configuration_utils.js';
|
|
358
377
|
import { StoppingCriteriaList } from '../generation/stopping_criteria.js';
|
|
359
378
|
import { ModelOutput } from './modeling_outputs.js';
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"modeling_utils.d.ts","sourceRoot":"","sources":["../../src/models/modeling_utils.js"],"names":[],"mappings":"AASA;;;GAGG;AACH,0DAEC;
|
|
1
|
+
{"version":3,"file":"modeling_utils.d.ts","sourceRoot":"","sources":["../../src/models/modeling_utils.js"],"names":[],"mappings":"AASA;;;GAGG;AACH,0DAEC;AA2DD;;;;;GAKG;AACH,kCAJW,OAAO,GACL,MAAM,CAKlB;AAuLD;;;;;;GAMG;AACH,6CALW,MAAM,+BAGJ;IAAE,QAAQ,EAAE,MAAM,CAAC,MAAM,EAAE,MAAM,CAAC,CAAC;IAAC,cAAc,CAAC,EAAE,MAAM,CAAC,MAAM,EAAE,IAAI,CAAC,CAAC;IAAC,gBAAgB,CAAC,EAAE,MAAM,CAAC,MAAM,EAAE,MAAM,CAAC,CAAA;CAAE,CASlI;AAg9BD;;;;;;GAMG;AACH,+DAHa,OAAO,CAAC,eAAe,CAAC,CAyBpC;AAED;;;;;;GAMG;AACH,+DAHa,OAAO,KAAQ,CAgC3B;AAED,iFAIC;AAED;;;;;;GAMG;AACH,6FAHa,OAAO,KAAQ,CAoC3B;AAED;;;;;;;;;;;;;;;;;GAiBG;AACH,kQAdG;IAA0B,eAAe;IACf,cAAc;IACd,oBAAoB,GAAtC,MAAM,EAAE;IACQ,oBAAoB,GAApC,MAAM;IACU,SAAS,GAAzB,MAAM;IACU,cAAc,GAA9B,MAAM;IACU,YAAY,GAA5B,MAAM;IACU,aAAa,GAA7B,MAAM;IACgB,eAAe,GAArC,YAAY;IACI,iBAAiB;IACjB,gBAAgB;CACxC,GAAU,OAAO,CAAC,MAAM,CAAC,CAyG3B;AAED;;;;;;GAMG;AACH,oEAHa,OAAO,CAAC,MAAM,CAAC,CAW3B;AAED;;;;;;GAMG;AACH,oEAHa,OAAO,CAAC,MAAM,CAAC,CAW3B;AAED;;;;;;;;GAQG;AACH,mDAHW,MAAM,yBACJ;IAAC,IAAI,EAAE,aAAa,CAAC;IAAC,IAAI,EAAE,MAAM,EAAE,CAAA;CAAC,CAsBjD;AAED;;;;;;;;;;;GAWG;AACH,4GAUC;AAED,iIAkDC;AAED,yIASC;AAED,sGAMC;AAED;;;;;;;;;EAiCC;AAED;;;;;;;;;EAcC;AAED;;;;;;;;;EAcC;AAED;;;;;;;GAOG;AACH,oEANW,MAAM,SACN,MAAM,CAAC,MAAM,EAAE,MAAM,CAAC,WACtB,OAAO,iBAAiB,EAAE,sBAAsB,GAC9C,OAAO,CAAC,MAAM,CAAC,MAAM,EAAE,GAAG,CAAC,CAAC,CAYxC;AAnsDD,oCAAsC;;;;;;;;;;;;;;;;;;;;AA+StC,+CAA4C;AAC5C,wDAAqD;AACrD,wDAAqD;;KAtTb,GAAG;UAClC,GAAE;;AAuTX;;GAEG;AACH;IA8CI;;;;;;;;;;;;;;OAcG;IACH,sDATW,MAAM,gKAKN,OAAO,iBAAiB,EAAE,sBAAsB,GAE9C,OAAO,CAAC,eAAe,CAAC,CA0DpC;IAhHD;;;;;OAKG;IACH,oBAJW,OAAO,eAAe,EAAE,gBAAgB,YACxC,MAAM,CAAC,MAAM,EAAE,GAAG,CAAC,WACnB,MAAM,CAAC,MAAM,MAAS,EAsBhC;IA9BD,wBAA8B;IAC9B,yBAAiD;IAEjD,mCAAqC;IAUjC,iDAAoB;IACpB,8BAAwB;IACxB,6BAAsB;IAKtB,kBAA2C;IAC3C,cAAkC;IAClC,oCAA+D;IAM/D,2DAA2D;IAC3D,eADW,OAAO,eAAe,EAAE,oBAAoB,CACS;IAGpE;;;;OAIG;IACH,WAHa,OAAO,CAAC,OAAO,EAAE,CAAC,CAS9B;IA2ED;;;;OAIG;IACH,0BAFa,OAAO,KAAQ,CAI3B;IAED;;;;;;OAMG;IACH,4BAHa,OAAO,KAAQ,CAK3B;IAED;;;OAGG;IACH,yBAFa,gBAAgB,GAAC,IAAI,CAIjC;IAED;;;;;OAKG;IACH,8BA2JC;IAED;;;;;;OAMG;IACH,8CAJW,gBAAgB,GAAC,IAAI,+CAEnB,gBAAgB,CA+B5B;IAED;;;;OAIG;IACH,0CAHW,gBAAgB,sBAChB,OAAO,oCAAoC,EAAE,gBAAgB,GAAC,OAAO,oCAAoC,EAAE,gBAAgB,EAAE,GAAC,oBAAoB,wBAqB5J;IAED;;;OAGG;IACH,8BA2BC;IAED,mDAKC;IAED;;;;;;;;OAQG;IACH,wGANG;QAA2B,mBAAmB,EAAtC,MAAM,EAAE,EAAE;QACK,OAAO;QACP,YAAY;QACX,kBAAkB,EAAlC,OAAO;KACf,OA0BF;IAED;;;;;;;OAOG;IACH,8DALG;QAAwB,MAAM,GAAtB,MAAM;QACU,YAAY,GAA5B,MAAM;QACmC,YAAY,GAArD,MAAM,CAAC,MAAM,EAAE,MAAM,GAAC,MAAM,EAAE,CAAC;KACvC,GAAU;QAAC,aAAa,EAAE,MAAM,CAAC;QAAC,YAAY,EAAE,MAAM,CAAC,MAAM,EAAE,MAAM,CAAC,GAAG;YAAC,eAAe,CAAC,EAAE,YAAY,CAAA;SAAC,CAAC;QAAC,gBAAgB,EAAE,MAAM,CAAA;KAAC,CAmBtI;IAED;;;;;qBAmDC;IAED;;;OAGG;IACH,oJAFW,GAAC;;;MAwDX;IAED;;;;OAIG;IACH,kGAHW,OAAO,6BAA6B,EAAE,4BAA4B,GAChE,OAAO,CAAC,WAAW,GAAC,MAAM,CAAC,CAgOvC;IAED;;;;;;;OAOG;IACH,qDAJW,YAAY,uBACZ,OAAO,GACL,YAAY,CAsCxB;IAED;;;;;OAKG;IACH,kCAFa;QAAC,gBAAgB,CAAC,EAAE,MAAM,EAAE,CAAA;KAAC,CAgBzC;IAED;;;;;OAKG;IACH,+BAHW,MAAM,CAAC,MAAM,EAAE,GAAG,CAAC,iBACnB,YAAY,GAAC,IAAI,QAiB3B;IAED;;;;;;OAMG;IACH,sBAOC;IAED,wCAEC;IAED,uCAEC;IAED,wCAEC;CACJ;uBAluCmE,oBAAoB;gCAexD,uBAAuB;6BAoB1B,mBAAmB;iCANf,sCAAsC;qCACG,oCAAoC;4BAGlF,uBAAuB"}
|
package/types/models/models.d.ts
CHANGED
|
@@ -1,7 +1,8 @@
|
|
|
1
1
|
export * from "./albert/modeling_albert.js";
|
|
2
2
|
export * from "./apertus/modeling_apertus.js";
|
|
3
|
+
export * from "./afmoe/modeling_afmoe.js";
|
|
3
4
|
export * from "./arcee/modeling_arcee.js";
|
|
4
|
-
export * from "./
|
|
5
|
+
export * from "./audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.js";
|
|
5
6
|
export * from "./bart/modeling_bart.js";
|
|
6
7
|
export * from "./beit/modeling_beit.js";
|
|
7
8
|
export * from "./bert/modeling_bert.js";
|
|
@@ -11,17 +12,21 @@ export * from "./bloom/modeling_bloom.js";
|
|
|
11
12
|
export * from "./camembert/modeling_camembert.js";
|
|
12
13
|
export * from "./chatterbox/modeling_chatterbox.js";
|
|
13
14
|
export * from "./chinese_clip/modeling_chinese_clip.js";
|
|
15
|
+
export * from "./chmv2/modeling_chmv2.js";
|
|
14
16
|
export * from "./clap/modeling_clap.js";
|
|
15
17
|
export * from "./clip/modeling_clip.js";
|
|
16
18
|
export * from "./clipseg/modeling_clipseg.js";
|
|
17
19
|
export * from "./codegen/modeling_codegen.js";
|
|
18
20
|
export * from "./cohere/modeling_cohere.js";
|
|
21
|
+
export * from "./cohere2/modeling_cohere2.js";
|
|
22
|
+
export * from "./cohere_asr/modeling_cohere_asr.js";
|
|
19
23
|
export * from "./convbert/modeling_convbert.js";
|
|
20
24
|
export * from "./convnext/modeling_convnext.js";
|
|
21
25
|
export * from "./convnextv2/modeling_convnextv2.js";
|
|
22
26
|
export * from "./d_fine/modeling_d_fine.js";
|
|
23
27
|
export * from "./dac/modeling_dac.js";
|
|
24
28
|
export * from "./deberta/modeling_deberta.js";
|
|
29
|
+
export * from "./deepseek_v3/modeling_deepseek_v3.js";
|
|
25
30
|
export * from "./deberta_v2/modeling_deberta_v2.js";
|
|
26
31
|
export * from "./decision_transformer/modeling_decision_transformer.js";
|
|
27
32
|
export * from "./deit/modeling_deit.js";
|
|
@@ -39,6 +44,7 @@ export * from "./efficientnet/modeling_efficientnet.js";
|
|
|
39
44
|
export * from "./electra/modeling_electra.js";
|
|
40
45
|
export * from "./ernie4_5/modeling_ernie4_5.js";
|
|
41
46
|
export * from "./esm/modeling_esm.js";
|
|
47
|
+
export * from "./eurobert/modeling_eurobert.js";
|
|
42
48
|
export * from "./exaone/modeling_exaone.js";
|
|
43
49
|
export * from "./falcon/modeling_falcon.js";
|
|
44
50
|
export * from "./falcon_h1/modeling_falcon_h1.js";
|
|
@@ -49,6 +55,8 @@ export * from "./gemma2/modeling_gemma2.js";
|
|
|
49
55
|
export * from "./gemma3/modeling_gemma3.js";
|
|
50
56
|
export * from "./gemma3n/modeling_gemma3n.js";
|
|
51
57
|
export * from "./glm/modeling_glm.js";
|
|
58
|
+
export * from "./glm_moe_dsa/modeling_glm_moe_dsa.js";
|
|
59
|
+
export * from "./glm_ocr/modeling_glm_ocr.js";
|
|
52
60
|
export * from "./glpn/modeling_glpn.js";
|
|
53
61
|
export * from "./gpt_bigcode/modeling_gpt_bigcode.js";
|
|
54
62
|
export * from "./gpt_neo/modeling_gpt_neo.js";
|
|
@@ -58,6 +66,7 @@ export * from "./gpt2/modeling_gpt2.js";
|
|
|
58
66
|
export * from "./gptj/modeling_gptj.js";
|
|
59
67
|
export * from "./granite/modeling_granite.js";
|
|
60
68
|
export * from "./granitemoehybrid/modeling_granitemoehybrid.js";
|
|
69
|
+
export * from "./granite_speech/modeling_granite_speech.js";
|
|
61
70
|
export * from "./grounding_dino/modeling_grounding_dino.js";
|
|
62
71
|
export * from "./groupvit/modeling_groupvit.js";
|
|
63
72
|
export * from "./helium/modeling_helium.js";
|
|
@@ -69,7 +78,9 @@ export * from "./ijepa/modeling_ijepa.js";
|
|
|
69
78
|
export * from "./jais/modeling_jais.js";
|
|
70
79
|
export * from "./jina_clip/modeling_jina_clip.js";
|
|
71
80
|
export * from "./lfm2/modeling_lfm2.js";
|
|
81
|
+
export * from "./lighton_ocr/modeling_lighton_ocr.js";
|
|
72
82
|
export * from "./lfm2_moe/modeling_lfm2_moe.js";
|
|
83
|
+
export * from "./lfm2_vl/modeling_lfm2_vl.js";
|
|
73
84
|
export * from "./llama/modeling_llama.js";
|
|
74
85
|
export * from "./llama4/modeling_llama4.js";
|
|
75
86
|
export * from "./llava/modeling_llava.js";
|
|
@@ -84,6 +95,7 @@ export * from "./metric3dv2/modeling_metric3dv2.js";
|
|
|
84
95
|
export * from "./mgp_str/modeling_mgp_str.js";
|
|
85
96
|
export * from "./mimi/modeling_mimi.js";
|
|
86
97
|
export * from "./mistral/modeling_mistral.js";
|
|
98
|
+
export * from "./mistral4/modeling_mistral4.js";
|
|
87
99
|
export * from "./mobilebert/modeling_mobilebert.js";
|
|
88
100
|
export * from "./mobilellm/modeling_mobilellm.js";
|
|
89
101
|
export * from "./mobilenet_v1/modeling_mobilenet_v1.js";
|
|
@@ -101,11 +113,13 @@ export * from "./mt5/modeling_mt5.js";
|
|
|
101
113
|
export * from "./multi_modality/modeling_multi_modality.js";
|
|
102
114
|
export * from "./musicgen/modeling_musicgen.js";
|
|
103
115
|
export * from "./nanochat/modeling_nanochat.js";
|
|
116
|
+
export * from "./nemotron_h/modeling_nemotron_h.js";
|
|
104
117
|
export * from "./neobert/modeling_neobert.js";
|
|
105
118
|
export * from "./nomic_bert/modeling_nomic_bert.js";
|
|
106
119
|
export * from "./olmo/modeling_olmo.js";
|
|
107
120
|
export * from "./olmo2/modeling_olmo2.js";
|
|
108
121
|
export * from "./olmo3/modeling_olmo3.js";
|
|
122
|
+
export * from "./olmo_hybrid/modeling_olmo_hybrid.js";
|
|
109
123
|
export * from "./openelm/modeling_openelm.js";
|
|
110
124
|
export * from "./opt/modeling_opt.js";
|
|
111
125
|
export * from "./owlv2/modeling_owlv2.js";
|
|
@@ -120,8 +134,16 @@ export * from "./phi3_v/modeling_phi3_v.js";
|
|
|
120
134
|
export * from "./pvt/modeling_pvt.js";
|
|
121
135
|
export * from "./pyannote/modeling_pyannote.js";
|
|
122
136
|
export * from "./qwen2/modeling_qwen2.js";
|
|
137
|
+
export * from "./qwen2_moe/modeling_qwen2_moe.js";
|
|
123
138
|
export * from "./qwen2_vl/modeling_qwen2_vl.js";
|
|
139
|
+
export * from "./qwen2_5_vl/modeling_qwen2_5_vl.js";
|
|
124
140
|
export * from "./qwen3/modeling_qwen3.js";
|
|
141
|
+
export * from "./qwen3_moe/modeling_qwen3_moe.js";
|
|
142
|
+
export * from "./qwen3_next/modeling_qwen3_next.js";
|
|
143
|
+
export * from "./qwen3_vl/modeling_qwen3_vl.js";
|
|
144
|
+
export * from "./qwen3_vl_moe/modeling_qwen3_vl_moe.js";
|
|
145
|
+
export * from "./qwen3_5/modeling_qwen3_5.js";
|
|
146
|
+
export * from "./qwen3_5_moe/modeling_qwen3_5_moe.js";
|
|
125
147
|
export * from "./resnet/modeling_resnet.js";
|
|
126
148
|
export * from "./rf_detr/modeling_rf_detr.js";
|
|
127
149
|
export * from "./roberta/modeling_roberta.js";
|
|
@@ -135,6 +157,7 @@ export * from "./segformer/modeling_segformer.js";
|
|
|
135
157
|
export * from "./siglip/modeling_siglip.js";
|
|
136
158
|
export * from "./smollm3/modeling_smollm3.js";
|
|
137
159
|
export * from "./snac/modeling_snac.js";
|
|
160
|
+
export * from "./solar_open/modeling_solar_open.js";
|
|
138
161
|
export * from "./speecht5/modeling_speecht5.js";
|
|
139
162
|
export * from "./squeezebert/modeling_squeezebert.js";
|
|
140
163
|
export * from "./stablelm/modeling_stablelm.js";
|
|
@@ -157,6 +180,8 @@ export * from "./vit_msn/modeling_vit_msn.js";
|
|
|
157
180
|
export * from "./vitmatte/modeling_vitmatte.js";
|
|
158
181
|
export * from "./vitpose/modeling_vitpose.js";
|
|
159
182
|
export * from "./vits/modeling_vits.js";
|
|
183
|
+
export * from "./voxtral/modeling_voxtral.js";
|
|
184
|
+
export * from "./voxtral_realtime/modeling_voxtral_realtime.js";
|
|
160
185
|
export * from "./wav2vec2/modeling_wav2vec2.js";
|
|
161
186
|
export * from "./wav2vec2_bert/modeling_wav2vec2_bert.js";
|
|
162
187
|
export * from "./wavlm/modeling_wavlm.js";
|
|
@@ -0,0 +1,8 @@
|
|
|
1
|
+
export class NemotronHPreTrainedModel extends PreTrainedModel {
|
|
2
|
+
}
|
|
3
|
+
export class NemotronHModel extends NemotronHPreTrainedModel {
|
|
4
|
+
}
|
|
5
|
+
export class NemotronHForCausalLM extends NemotronHPreTrainedModel {
|
|
6
|
+
}
|
|
7
|
+
import { PreTrainedModel } from '../modeling_utils.js';
|
|
8
|
+
//# sourceMappingURL=modeling_nemotron_h.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"modeling_nemotron_h.d.ts","sourceRoot":"","sources":["../../../src/models/nemotron_h/modeling_nemotron_h.js"],"names":[],"mappings":"AAEA;CAAgE;AAChE;CAA+D;AAC/D;CAAqE;gCAJrC,sBAAsB"}
|
|
@@ -0,0 +1,8 @@
|
|
|
1
|
+
export class OlmoHybridPreTrainedModel extends PreTrainedModel {
|
|
2
|
+
}
|
|
3
|
+
export class OlmoHybridModel extends OlmoHybridPreTrainedModel {
|
|
4
|
+
}
|
|
5
|
+
export class OlmoHybridForCausalLM extends OlmoHybridPreTrainedModel {
|
|
6
|
+
}
|
|
7
|
+
import { PreTrainedModel } from '../modeling_utils.js';
|
|
8
|
+
//# sourceMappingURL=modeling_olmo_hybrid.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"modeling_olmo_hybrid.d.ts","sourceRoot":"","sources":["../../../src/models/olmo_hybrid/modeling_olmo_hybrid.js"],"names":[],"mappings":"AAEA;CAAiE;AACjE;CAAiE;AACjE;CAAuE;gCAJvC,sBAAsB"}
|
|
@@ -1,10 +1,4 @@
|
|
|
1
|
-
export class
|
|
1
|
+
export class PaliGemmaForConditionalGeneration extends LlavaForConditionalGeneration {
|
|
2
2
|
}
|
|
3
|
-
|
|
4
|
-
_merge_input_ids_with_image_features(kwargs: any): {
|
|
5
|
-
inputs_embeds: any;
|
|
6
|
-
attention_mask: any;
|
|
7
|
-
};
|
|
8
|
-
}
|
|
9
|
-
import { PreTrainedModel } from '../modeling_utils.js';
|
|
3
|
+
import { LlavaForConditionalGeneration } from '../llava/modeling_llava.js';
|
|
10
4
|
//# sourceMappingURL=modeling_paligemma.d.ts.map
|