@huggingface/transformers 4.0.0-next.0 → 4.0.0-next.10

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (364) hide show
  1. package/README.md +32 -6
  2. package/dist/ort-wasm-simd-threaded.jsep.mjs +31 -31
  3. package/dist/transformers.js +9261 -1709
  4. package/dist/transformers.min.js +25 -18
  5. package/dist/transformers.node.cjs +6119 -3174
  6. package/dist/transformers.node.min.cjs +25 -23
  7. package/dist/transformers.node.min.mjs +25 -23
  8. package/dist/transformers.node.mjs +6034 -3168
  9. package/dist/transformers.web.js +4255 -1381
  10. package/dist/transformers.web.min.js +23 -19
  11. package/package.json +6 -6
  12. package/src/backends/onnx.js +128 -53
  13. package/src/backends/utils/cacheWasm.js +28 -46
  14. package/src/cache_utils.js +62 -0
  15. package/src/configs.js +123 -23
  16. package/src/env.js +100 -11
  17. package/src/generation/logits_sampler.js +3 -15
  18. package/src/generation/parameters.js +1 -1
  19. package/src/generation/streamers.js +21 -0
  20. package/src/image_processors_utils.js +29 -23
  21. package/src/models/afmoe/modeling_afmoe.js +5 -0
  22. package/src/models/auto/image_processing_auto.js +2 -1
  23. package/src/models/auto/modeling_auto.js +16 -2
  24. package/src/models/auto/tokenization_auto.js +2 -1
  25. package/src/models/chatterbox/modeling_chatterbox.js +1 -1
  26. package/src/models/chmv2/image_processing_chmv2.js +3 -0
  27. package/src/models/chmv2/modeling_chmv2.js +4 -0
  28. package/src/models/clap/feature_extraction_clap.js +2 -1
  29. package/src/models/cohere2/modeling_cohere2.js +5 -0
  30. package/src/models/cohere_asr/feature_extraction_cohere_asr.js +117 -0
  31. package/src/models/cohere_asr/modeling_cohere_asr.js +11 -0
  32. package/src/models/cohere_asr/processing_cohere_asr.js +55 -0
  33. package/src/models/cohere_asr/tokenization_cohere_asr.js +3 -0
  34. package/src/models/deepseek_v3/modeling_deepseek_v3.js +5 -0
  35. package/src/models/detr/image_processing_detr.js +1 -1
  36. package/src/models/eurobert/modeling_eurobert.js +41 -0
  37. package/src/models/feature_extractors.js +3 -0
  38. package/src/models/gemma3/image_processing_gemma3.js +3 -0
  39. package/src/models/gemma3/modeling_gemma3.js +4 -1
  40. package/src/models/gemma3/processing_gemma3.js +45 -0
  41. package/src/models/gemma3n/modeling_gemma3n.js +2 -0
  42. package/src/models/glm46v/image_processing_glm46v.js +12 -0
  43. package/src/models/glm46v/processing_glm46v.js +5 -0
  44. package/src/models/glm_moe_dsa/modeling_glm_moe_dsa.js +5 -0
  45. package/src/models/glm_ocr/modeling_glm_ocr.js +78 -0
  46. package/src/models/granite_speech/feature_extraction_granite_speech.js +58 -0
  47. package/src/models/granite_speech/modeling_granite_speech.js +5 -0
  48. package/src/models/granite_speech/processing_granite_speech.js +62 -0
  49. package/src/models/grounding_dino/image_processing_grounding_dino.js +1 -1
  50. package/src/models/idefics3/modeling_idefics3.js +5 -32
  51. package/src/models/image_processors.js +4 -0
  52. package/src/models/lfm2_vl/image_processing_lfm2_vl.js +305 -0
  53. package/src/models/lfm2_vl/modeling_lfm2_vl.js +13 -0
  54. package/src/models/lfm2_vl/processing_lfm2_vl.js +77 -0
  55. package/src/models/lighton_ocr/modeling_lighton_ocr.js +3 -0
  56. package/src/models/llava/modeling_llava.js +1 -1
  57. package/src/models/marian/tokenization_marian.js +3 -2
  58. package/src/models/mistral3/modeling_mistral3.js +2 -2
  59. package/src/models/mistral4/modeling_mistral4.js +5 -0
  60. package/src/models/modeling_utils.js +283 -300
  61. package/src/models/models.js +26 -1
  62. package/src/models/nemotron_h/modeling_nemotron_h.js +5 -0
  63. package/src/models/olmo_hybrid/modeling_olmo_hybrid.js +5 -0
  64. package/src/models/paligemma/modeling_paligemma.js +2 -25
  65. package/src/models/paligemma/processing_paligemma.js +3 -2
  66. package/src/models/processors.js +8 -0
  67. package/src/models/qwen2_5_vl/modeling_qwen2_5_vl.js +9 -0
  68. package/src/models/qwen2_5_vl/processing_qwen2_5_vl.js +3 -0
  69. package/src/models/qwen2_moe/modeling_qwen2_moe.js +5 -0
  70. package/src/models/qwen2_vl/image_processing_qwen2_vl.js +15 -1
  71. package/src/models/qwen2_vl/modeling_qwen2_vl.js +240 -143
  72. package/src/models/qwen2_vl/processing_qwen2_vl.js +5 -4
  73. package/src/models/qwen3_5/modeling_qwen3_5.js +4 -0
  74. package/src/models/qwen3_5_moe/modeling_qwen3_5_moe.js +4 -0
  75. package/src/models/qwen3_moe/modeling_qwen3_moe.js +5 -0
  76. package/src/models/qwen3_next/modeling_qwen3_next.js +5 -0
  77. package/src/models/qwen3_vl/modeling_qwen3_vl.js +4 -0
  78. package/src/models/qwen3_vl/processing_qwen3_vl.js +3 -0
  79. package/src/models/qwen3_vl_moe/modeling_qwen3_vl_moe.js +4 -0
  80. package/src/models/registry.js +61 -5
  81. package/src/models/sam/image_processing_sam.js +1 -1
  82. package/src/models/session.js +33 -56
  83. package/src/models/smolvlm/modeling_smolvlm.js +7 -0
  84. package/src/models/solar_open/modeling_solar_open.js +5 -0
  85. package/src/models/tokenizers.js +1 -0
  86. package/src/models/ultravox/modeling_ultravox.js +1 -3
  87. package/src/models/voxtral/modeling_voxtral.js +3 -0
  88. package/src/models/voxtral_realtime/feature_extraction_voxtral_realtime.js +71 -0
  89. package/src/models/voxtral_realtime/modeling_voxtral_realtime.js +239 -0
  90. package/src/models/voxtral_realtime/processing_voxtral_realtime.js +113 -0
  91. package/src/models/whisper/feature_extraction_whisper.js +4 -13
  92. package/src/models/whisper/modeling_whisper.js +6 -5
  93. package/src/models/xlm/tokenization_xlm.js +2 -1
  94. package/src/pipelines/automatic-speech-recognition.js +47 -3
  95. package/src/pipelines/document-question-answering.js +1 -1
  96. package/src/pipelines/image-to-text.js +2 -2
  97. package/src/pipelines/index.js +313 -0
  98. package/src/pipelines/summarization.js +1 -1
  99. package/src/pipelines/text-generation.js +5 -1
  100. package/src/pipelines/text-to-audio.js +4 -2
  101. package/src/pipelines/text2text-generation.js +1 -1
  102. package/src/pipelines/translation.js +1 -1
  103. package/src/pipelines/zero-shot-classification.js +3 -2
  104. package/src/pipelines.js +140 -428
  105. package/src/tokenization_utils.js +42 -21
  106. package/src/transformers.js +10 -1
  107. package/src/utils/audio.js +20 -3
  108. package/src/utils/cache/CrossOriginStorageCache.js +251 -0
  109. package/src/utils/cache/FileCache.js +128 -0
  110. package/src/utils/cache/cross-origin-storage.d.ts +38 -0
  111. package/src/utils/cache.js +12 -4
  112. package/src/utils/core.js +23 -1
  113. package/src/utils/devices.js +22 -0
  114. package/src/utils/dtypes.js +55 -0
  115. package/src/utils/hub/{files.js → FileResponse.js} +0 -90
  116. package/src/utils/hub/utils.js +45 -5
  117. package/src/utils/hub.js +67 -23
  118. package/src/utils/image.js +14 -14
  119. package/src/utils/logger.js +67 -0
  120. package/src/utils/lru_cache.js +67 -0
  121. package/src/utils/memoize_promise.js +45 -0
  122. package/src/utils/model-loader.js +35 -17
  123. package/src/utils/model_registry/ModelRegistry.js +382 -0
  124. package/src/utils/model_registry/clear_cache.js +128 -0
  125. package/src/utils/model_registry/get_available_dtypes.js +68 -0
  126. package/src/utils/model_registry/get_file_metadata.js +162 -0
  127. package/src/utils/model_registry/get_files.js +42 -0
  128. package/src/utils/model_registry/get_model_files.js +114 -0
  129. package/src/utils/model_registry/get_pipeline_files.js +44 -0
  130. package/src/utils/model_registry/get_processor_files.js +20 -0
  131. package/src/utils/model_registry/get_tokenizer_files.js +21 -0
  132. package/src/utils/model_registry/is_cached.js +169 -0
  133. package/src/utils/model_registry/resolve_model_type.js +66 -0
  134. package/src/utils/random.js +225 -0
  135. package/src/utils/tensor.js +26 -23
  136. package/src/utils/video.js +2 -2
  137. package/types/backends/onnx.d.ts.map +1 -1
  138. package/types/backends/utils/cacheWasm.d.ts +3 -17
  139. package/types/backends/utils/cacheWasm.d.ts.map +1 -1
  140. package/types/cache_utils.d.ts +29 -0
  141. package/types/cache_utils.d.ts.map +1 -0
  142. package/types/configs.d.ts.map +1 -1
  143. package/types/env.d.ts +60 -27
  144. package/types/env.d.ts.map +1 -1
  145. package/types/generation/logits_sampler.d.ts +2 -2
  146. package/types/generation/logits_sampler.d.ts.map +1 -1
  147. package/types/generation/parameters.d.ts +1 -1
  148. package/types/generation/parameters.d.ts.map +1 -1
  149. package/types/generation/streamers.d.ts +1 -0
  150. package/types/generation/streamers.d.ts.map +1 -1
  151. package/types/image_processors_utils.d.ts +18 -1
  152. package/types/image_processors_utils.d.ts.map +1 -1
  153. package/types/models/afmoe/modeling_afmoe.d.ts +8 -0
  154. package/types/models/afmoe/modeling_afmoe.d.ts.map +1 -0
  155. package/types/models/{ast/modeling_ast.d.ts → audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.d.ts} +1 -1
  156. package/types/models/audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.d.ts.map +1 -0
  157. package/types/models/auto/image_processing_auto.d.ts.map +1 -1
  158. package/types/models/auto/modeling_auto.d.ts +6 -0
  159. package/types/models/auto/modeling_auto.d.ts.map +1 -1
  160. package/types/models/auto/tokenization_auto.d.ts.map +1 -1
  161. package/types/models/chmv2/image_processing_chmv2.d.ts +4 -0
  162. package/types/models/chmv2/image_processing_chmv2.d.ts.map +1 -0
  163. package/types/models/chmv2/modeling_chmv2.d.ts +6 -0
  164. package/types/models/chmv2/modeling_chmv2.d.ts.map +1 -0
  165. package/types/models/clap/feature_extraction_clap.d.ts.map +1 -1
  166. package/types/models/cohere2/modeling_cohere2.d.ts +8 -0
  167. package/types/models/cohere2/modeling_cohere2.d.ts.map +1 -0
  168. package/types/models/cohere_asr/feature_extraction_cohere_asr.d.ts +25 -0
  169. package/types/models/cohere_asr/feature_extraction_cohere_asr.d.ts.map +1 -0
  170. package/types/models/cohere_asr/modeling_cohere_asr.d.ts +9 -0
  171. package/types/models/cohere_asr/modeling_cohere_asr.d.ts.map +1 -0
  172. package/types/models/cohere_asr/processing_cohere_asr.d.ts +27 -0
  173. package/types/models/cohere_asr/processing_cohere_asr.d.ts.map +1 -0
  174. package/types/models/cohere_asr/tokenization_cohere_asr.d.ts +4 -0
  175. package/types/models/cohere_asr/tokenization_cohere_asr.d.ts.map +1 -0
  176. package/types/models/deepseek_v3/modeling_deepseek_v3.d.ts +8 -0
  177. package/types/models/deepseek_v3/modeling_deepseek_v3.d.ts.map +1 -0
  178. package/types/models/detr/image_processing_detr.d.ts +1 -1
  179. package/types/models/eurobert/modeling_eurobert.d.ts +36 -0
  180. package/types/models/eurobert/modeling_eurobert.d.ts.map +1 -0
  181. package/types/models/feature_extractors.d.ts +3 -0
  182. package/types/models/gemma3/image_processing_gemma3.d.ts +4 -0
  183. package/types/models/gemma3/image_processing_gemma3.d.ts.map +1 -0
  184. package/types/models/gemma3/modeling_gemma3.d.ts +4 -1
  185. package/types/models/gemma3/modeling_gemma3.d.ts.map +1 -1
  186. package/types/models/gemma3/processing_gemma3.d.ts +20 -0
  187. package/types/models/gemma3/processing_gemma3.d.ts.map +1 -0
  188. package/types/models/gemma3n/modeling_gemma3n.d.ts +2 -0
  189. package/types/models/gemma3n/modeling_gemma3n.d.ts.map +1 -1
  190. package/types/models/glm46v/image_processing_glm46v.d.ts +4 -0
  191. package/types/models/glm46v/image_processing_glm46v.d.ts.map +1 -0
  192. package/types/models/glm46v/processing_glm46v.d.ts +4 -0
  193. package/types/models/glm46v/processing_glm46v.d.ts.map +1 -0
  194. package/types/models/glm_moe_dsa/modeling_glm_moe_dsa.d.ts +8 -0
  195. package/types/models/glm_moe_dsa/modeling_glm_moe_dsa.d.ts.map +1 -0
  196. package/types/models/glm_ocr/modeling_glm_ocr.d.ts +26 -0
  197. package/types/models/glm_ocr/modeling_glm_ocr.d.ts.map +1 -0
  198. package/types/models/granite_speech/feature_extraction_granite_speech.d.ts +16 -0
  199. package/types/models/granite_speech/feature_extraction_granite_speech.d.ts.map +1 -0
  200. package/types/models/granite_speech/modeling_granite_speech.d.ts +4 -0
  201. package/types/models/granite_speech/modeling_granite_speech.d.ts.map +1 -0
  202. package/types/models/granite_speech/processing_granite_speech.d.ts +19 -0
  203. package/types/models/granite_speech/processing_granite_speech.d.ts.map +1 -0
  204. package/types/models/grounding_dino/image_processing_grounding_dino.d.ts +1 -1
  205. package/types/models/idefics3/modeling_idefics3.d.ts +2 -18
  206. package/types/models/idefics3/modeling_idefics3.d.ts.map +1 -1
  207. package/types/models/image_processors.d.ts +4 -0
  208. package/types/models/lfm2_vl/image_processing_lfm2_vl.d.ts +41 -0
  209. package/types/models/lfm2_vl/image_processing_lfm2_vl.d.ts.map +1 -0
  210. package/types/models/lfm2_vl/modeling_lfm2_vl.d.ts +4 -0
  211. package/types/models/lfm2_vl/modeling_lfm2_vl.d.ts.map +1 -0
  212. package/types/models/lfm2_vl/processing_lfm2_vl.d.ts +18 -0
  213. package/types/models/lfm2_vl/processing_lfm2_vl.d.ts.map +1 -0
  214. package/types/models/lighton_ocr/modeling_lighton_ocr.d.ts +4 -0
  215. package/types/models/lighton_ocr/modeling_lighton_ocr.d.ts.map +1 -0
  216. package/types/models/marian/tokenization_marian.d.ts.map +1 -1
  217. package/types/models/mistral3/modeling_mistral3.d.ts +2 -2
  218. package/types/models/mistral3/modeling_mistral3.d.ts.map +1 -1
  219. package/types/models/mistral4/modeling_mistral4.d.ts +8 -0
  220. package/types/models/mistral4/modeling_mistral4.d.ts.map +1 -0
  221. package/types/models/modeling_utils.d.ts +46 -27
  222. package/types/models/modeling_utils.d.ts.map +1 -1
  223. package/types/models/models.d.ts +26 -1
  224. package/types/models/nemotron_h/modeling_nemotron_h.d.ts +8 -0
  225. package/types/models/nemotron_h/modeling_nemotron_h.d.ts.map +1 -0
  226. package/types/models/olmo_hybrid/modeling_olmo_hybrid.d.ts +8 -0
  227. package/types/models/olmo_hybrid/modeling_olmo_hybrid.d.ts.map +1 -0
  228. package/types/models/paligemma/modeling_paligemma.d.ts +2 -8
  229. package/types/models/paligemma/modeling_paligemma.d.ts.map +1 -1
  230. package/types/models/paligemma/processing_paligemma.d.ts.map +1 -1
  231. package/types/models/processors.d.ts +8 -0
  232. package/types/models/qwen2_5_vl/modeling_qwen2_5_vl.d.ts +7 -0
  233. package/types/models/qwen2_5_vl/modeling_qwen2_5_vl.d.ts.map +1 -0
  234. package/types/models/qwen2_5_vl/processing_qwen2_5_vl.d.ts +4 -0
  235. package/types/models/qwen2_5_vl/processing_qwen2_5_vl.d.ts.map +1 -0
  236. package/types/models/qwen2_moe/modeling_qwen2_moe.d.ts +8 -0
  237. package/types/models/qwen2_moe/modeling_qwen2_moe.d.ts.map +1 -0
  238. package/types/models/qwen2_vl/image_processing_qwen2_vl.d.ts +3 -0
  239. package/types/models/qwen2_vl/image_processing_qwen2_vl.d.ts.map +1 -1
  240. package/types/models/qwen2_vl/modeling_qwen2_vl.d.ts +44 -6
  241. package/types/models/qwen2_vl/modeling_qwen2_vl.d.ts.map +1 -1
  242. package/types/models/qwen2_vl/processing_qwen2_vl.d.ts +1 -0
  243. package/types/models/qwen2_vl/processing_qwen2_vl.d.ts.map +1 -1
  244. package/types/models/qwen3_5/modeling_qwen3_5.d.ts +6 -0
  245. package/types/models/qwen3_5/modeling_qwen3_5.d.ts.map +1 -0
  246. package/types/models/qwen3_5_moe/modeling_qwen3_5_moe.d.ts +7 -0
  247. package/types/models/qwen3_5_moe/modeling_qwen3_5_moe.d.ts.map +1 -0
  248. package/types/models/qwen3_moe/modeling_qwen3_moe.d.ts +8 -0
  249. package/types/models/qwen3_moe/modeling_qwen3_moe.d.ts.map +1 -0
  250. package/types/models/qwen3_next/modeling_qwen3_next.d.ts +8 -0
  251. package/types/models/qwen3_next/modeling_qwen3_next.d.ts.map +1 -0
  252. package/types/models/qwen3_vl/modeling_qwen3_vl.d.ts +7 -0
  253. package/types/models/qwen3_vl/modeling_qwen3_vl.d.ts.map +1 -0
  254. package/types/models/qwen3_vl/processing_qwen3_vl.d.ts +4 -0
  255. package/types/models/qwen3_vl/processing_qwen3_vl.d.ts.map +1 -0
  256. package/types/models/qwen3_vl_moe/modeling_qwen3_vl_moe.d.ts +7 -0
  257. package/types/models/qwen3_vl_moe/modeling_qwen3_vl_moe.d.ts.map +1 -0
  258. package/types/models/registry.d.ts +2 -1
  259. package/types/models/registry.d.ts.map +1 -1
  260. package/types/models/sam/image_processing_sam.d.ts +1 -1
  261. package/types/models/session.d.ts +3 -2
  262. package/types/models/session.d.ts.map +1 -1
  263. package/types/models/smolvlm/modeling_smolvlm.d.ts +8 -0
  264. package/types/models/smolvlm/modeling_smolvlm.d.ts.map +1 -0
  265. package/types/models/solar_open/modeling_solar_open.d.ts +8 -0
  266. package/types/models/solar_open/modeling_solar_open.d.ts.map +1 -0
  267. package/types/models/tokenizers.d.ts +1 -0
  268. package/types/models/ultravox/modeling_ultravox.d.ts +0 -2
  269. package/types/models/ultravox/modeling_ultravox.d.ts.map +1 -1
  270. package/types/models/voxtral/modeling_voxtral.d.ts +4 -0
  271. package/types/models/voxtral/modeling_voxtral.d.ts.map +1 -0
  272. package/types/models/voxtral_realtime/feature_extraction_voxtral_realtime.d.ts +28 -0
  273. package/types/models/voxtral_realtime/feature_extraction_voxtral_realtime.d.ts.map +1 -0
  274. package/types/models/voxtral_realtime/modeling_voxtral_realtime.d.ts +17 -0
  275. package/types/models/voxtral_realtime/modeling_voxtral_realtime.d.ts.map +1 -0
  276. package/types/models/voxtral_realtime/processing_voxtral_realtime.d.ts +44 -0
  277. package/types/models/voxtral_realtime/processing_voxtral_realtime.d.ts.map +1 -0
  278. package/types/models/whisper/feature_extraction_whisper.d.ts.map +1 -1
  279. package/types/models/whisper/modeling_whisper.d.ts.map +1 -1
  280. package/types/models/xlm/tokenization_xlm.d.ts.map +1 -1
  281. package/types/pipelines/automatic-speech-recognition.d.ts +7 -2
  282. package/types/pipelines/automatic-speech-recognition.d.ts.map +1 -1
  283. package/types/pipelines/document-question-answering.d.ts +2 -2
  284. package/types/pipelines/document-question-answering.d.ts.map +1 -1
  285. package/types/pipelines/image-to-text.d.ts +4 -4
  286. package/types/pipelines/image-to-text.d.ts.map +1 -1
  287. package/types/pipelines/index.d.ts +265 -0
  288. package/types/pipelines/index.d.ts.map +1 -0
  289. package/types/pipelines/summarization.d.ts +2 -2
  290. package/types/pipelines/summarization.d.ts.map +1 -1
  291. package/types/pipelines/text-generation.d.ts +7 -3
  292. package/types/pipelines/text-generation.d.ts.map +1 -1
  293. package/types/pipelines/text-to-audio.d.ts.map +1 -1
  294. package/types/pipelines/text2text-generation.d.ts +3 -3
  295. package/types/pipelines/text2text-generation.d.ts.map +1 -1
  296. package/types/pipelines/translation.d.ts +2 -2
  297. package/types/pipelines/translation.d.ts.map +1 -1
  298. package/types/pipelines/zero-shot-classification.d.ts.map +1 -1
  299. package/types/pipelines.d.ts +51 -291
  300. package/types/pipelines.d.ts.map +1 -1
  301. package/types/tokenization_utils.d.ts +44 -26
  302. package/types/tokenization_utils.d.ts.map +1 -1
  303. package/types/transformers.d.ts +7 -1
  304. package/types/transformers.d.ts.map +1 -1
  305. package/types/utils/audio.d.ts +5 -2
  306. package/types/utils/audio.d.ts.map +1 -1
  307. package/types/utils/cache/CrossOriginStorageCache.d.ts +120 -0
  308. package/types/utils/cache/CrossOriginStorageCache.d.ts.map +1 -0
  309. package/types/utils/cache/FileCache.d.ts +39 -0
  310. package/types/utils/cache/FileCache.d.ts.map +1 -0
  311. package/types/utils/cache.d.ts +10 -4
  312. package/types/utils/cache.d.ts.map +1 -1
  313. package/types/utils/core.d.ts +59 -2
  314. package/types/utils/core.d.ts.map +1 -1
  315. package/types/utils/devices.d.ts +15 -0
  316. package/types/utils/devices.d.ts.map +1 -1
  317. package/types/utils/dtypes.d.ts +17 -1
  318. package/types/utils/dtypes.d.ts.map +1 -1
  319. package/types/utils/hub/{files.d.ts → FileResponse.d.ts} +1 -32
  320. package/types/utils/hub/FileResponse.d.ts.map +1 -0
  321. package/types/utils/hub/utils.d.ts +19 -3
  322. package/types/utils/hub/utils.d.ts.map +1 -1
  323. package/types/utils/hub.d.ts +36 -7
  324. package/types/utils/hub.d.ts.map +1 -1
  325. package/types/utils/image.d.ts +1 -1
  326. package/types/utils/logger.d.ts +28 -0
  327. package/types/utils/logger.d.ts.map +1 -0
  328. package/types/utils/lru_cache.d.ts +38 -0
  329. package/types/utils/lru_cache.d.ts.map +1 -0
  330. package/types/utils/memoize_promise.d.ts +14 -0
  331. package/types/utils/memoize_promise.d.ts.map +1 -0
  332. package/types/utils/model-loader.d.ts +15 -0
  333. package/types/utils/model-loader.d.ts.map +1 -1
  334. package/types/utils/model_registry/ModelRegistry.d.ts +298 -0
  335. package/types/utils/model_registry/ModelRegistry.d.ts.map +1 -0
  336. package/types/utils/model_registry/clear_cache.d.ts +74 -0
  337. package/types/utils/model_registry/clear_cache.d.ts.map +1 -0
  338. package/types/utils/model_registry/get_available_dtypes.d.ts +26 -0
  339. package/types/utils/model_registry/get_available_dtypes.d.ts.map +1 -0
  340. package/types/utils/model_registry/get_file_metadata.d.ts +20 -0
  341. package/types/utils/model_registry/get_file_metadata.d.ts.map +1 -0
  342. package/types/utils/model_registry/get_files.d.ts +23 -0
  343. package/types/utils/model_registry/get_files.d.ts.map +1 -0
  344. package/types/utils/model_registry/get_model_files.d.ts +48 -0
  345. package/types/utils/model_registry/get_model_files.d.ts.map +1 -0
  346. package/types/utils/model_registry/get_pipeline_files.d.ts +22 -0
  347. package/types/utils/model_registry/get_pipeline_files.d.ts.map +1 -0
  348. package/types/utils/model_registry/get_processor_files.d.ts +9 -0
  349. package/types/utils/model_registry/get_processor_files.d.ts.map +1 -0
  350. package/types/utils/model_registry/get_tokenizer_files.d.ts +9 -0
  351. package/types/utils/model_registry/get_tokenizer_files.d.ts.map +1 -0
  352. package/types/utils/model_registry/is_cached.d.ts +105 -0
  353. package/types/utils/model_registry/is_cached.d.ts.map +1 -0
  354. package/types/utils/model_registry/resolve_model_type.d.ts +24 -0
  355. package/types/utils/model_registry/resolve_model_type.d.ts.map +1 -0
  356. package/types/utils/random.d.ts +86 -0
  357. package/types/utils/random.d.ts.map +1 -0
  358. package/types/utils/tensor.d.ts.map +1 -1
  359. package/src/utils/data-structures.js +0 -572
  360. package/types/models/ast/modeling_ast.d.ts.map +0 -1
  361. package/types/utils/data-structures.d.ts +0 -294
  362. package/types/utils/data-structures.d.ts.map +0 -1
  363. package/types/utils/hub/files.d.ts.map +0 -1
  364. /package/src/models/{ast/modeling_ast.js → audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.js} +0 -0
@@ -0,0 +1,8 @@
1
+ export class GlmMoeDsaPreTrainedModel extends PreTrainedModel {
2
+ }
3
+ export class GlmMoeDsaModel extends GlmMoeDsaPreTrainedModel {
4
+ }
5
+ export class GlmMoeDsaForCausalLM extends GlmMoeDsaPreTrainedModel {
6
+ }
7
+ import { PreTrainedModel } from '../modeling_utils.js';
8
+ //# sourceMappingURL=modeling_glm_moe_dsa.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"modeling_glm_moe_dsa.d.ts","sourceRoot":"","sources":["../../../src/models/glm_moe_dsa/modeling_glm_moe_dsa.js"],"names":[],"mappings":"AAEA;CAAgE;AAChE;CAA+D;AAC/D;CAAqE;gCAJrC,sBAAsB"}
@@ -0,0 +1,26 @@
1
+ export class GlmOcrForConditionalGeneration extends Qwen2_5_VLForConditionalGeneration {
2
+ /**
3
+ * Compute 3D positional indices for vision tokens.
4
+ * Temporal is constant, height is repeat-interleaved, width tiles.
5
+ * @param {number} start_position
6
+ * @param {number[]} grid_thw [T, H, W]
7
+ * @param {number} temp_merge_size
8
+ * @param {number} spatial_merge_size
9
+ * @returns {number[]} Flat array of length 3 * seq_len: [temporal..., height..., width...]
10
+ */
11
+ get_vision_position_ids(start_position: number, grid_thw: number[], temp_merge_size: number, spatial_merge_size: number): number[];
12
+ /**
13
+ * GlmOcr uses mm_token_type_ids-style grouping (image tokens identified by image_token_id)
14
+ * instead of vision_start_token_id scanning used by Qwen2VL.
15
+ * After a vision segment, position advances by max(h, w) / spatial_merge_size.
16
+ */
17
+ _get_multimodal_rope_positions({ filtered_ids, image_grid_thw_list, video_grid_thw_list, spatial_merge_size, state, }: {
18
+ filtered_ids: any;
19
+ image_grid_thw_list: any;
20
+ video_grid_thw_list: any;
21
+ spatial_merge_size: any;
22
+ state: any;
23
+ }): number[][];
24
+ }
25
+ import { Qwen2_5_VLForConditionalGeneration } from '../qwen2_5_vl/modeling_qwen2_5_vl.js';
26
+ //# sourceMappingURL=modeling_glm_ocr.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"modeling_glm_ocr.d.ts","sourceRoot":"","sources":["../../../src/models/glm_ocr/modeling_glm_ocr.js"],"names":[],"mappings":"AAEA;IACI;;;;;;;;OAQG;IACH,wCANW,MAAM,YACN,MAAM,EAAE,mBACR,MAAM,sBACN,MAAM,GACJ,MAAM,EAAE,CAgBpB;IAED;;;;OAIG;IACH;;;;;;mBA2CC;CACJ;mDA7EkD,sCAAsC"}
@@ -0,0 +1,16 @@
1
+ export class GraniteSpeechFeatureExtractor extends FeatureExtractor {
2
+ constructor(config: any);
3
+ mel_filters: number[][];
4
+ window: Float64Array<any>;
5
+ /**
6
+ * Extract mel spectrogram features from audio, matching the Python GraniteSpeechFeatureExtractor.
7
+ * @param {Float32Array|Float64Array} audio The audio waveform.
8
+ * @returns {Promise<{input_features: Tensor}>}
9
+ */
10
+ _call(audio: Float32Array | Float64Array): Promise<{
11
+ input_features: Tensor;
12
+ }>;
13
+ }
14
+ import { FeatureExtractor } from '../../feature_extraction_utils.js';
15
+ import { Tensor } from '../../utils/tensor.js';
16
+ //# sourceMappingURL=feature_extraction_granite_speech.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"feature_extraction_granite_speech.d.ts","sourceRoot":"","sources":["../../../src/models/granite_speech/feature_extraction_granite_speech.js"],"names":[],"mappings":"AAIA;IACI,yBAsBC;IAhBG,wBAQC;IAKD,0BAAqC;IAKzC;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY,GACvB,OAAO,CAAC;QAAC,cAAc,EAAE,MAAM,CAAA;KAAC,CAAC,CAwB7C;CACJ;iCAzDuD,mCAAmC;uBAEpE,uBAAuB"}
@@ -0,0 +1,4 @@
1
+ export class GraniteSpeechForConditionalGeneration extends UltravoxModel {
2
+ }
3
+ import { UltravoxModel } from '../ultravox/modeling_ultravox.js';
4
+ //# sourceMappingURL=modeling_granite_speech.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"modeling_granite_speech.d.ts","sourceRoot":"","sources":["../../../src/models/granite_speech/modeling_granite_speech.js"],"names":[],"mappings":"AAEA;CAEC;8BAJ6B,kCAAkC"}
@@ -0,0 +1,19 @@
1
+ export class GraniteSpeechProcessor extends Processor {
2
+ static tokenizer_class: typeof AutoTokenizer;
3
+ static feature_extractor_class: typeof AutoFeatureExtractor;
4
+ /**
5
+ * Compute the number of audio tokens for a given raw audio length.
6
+ * @param {number} audioLength Raw audio sample count.
7
+ * @returns {number} Number of projector output tokens.
8
+ */
9
+ _get_num_audio_features(audioLength: number): number;
10
+ /**
11
+ * @param {string} text The text input to process.
12
+ * @param {Float32Array} audio The audio input to process.
13
+ */
14
+ _call(text: string, audio?: Float32Array, kwargs?: {}): Promise<any>;
15
+ }
16
+ import { Processor } from '../../processing_utils.js';
17
+ import { AutoTokenizer } from '../auto/tokenization_auto.js';
18
+ import { AutoFeatureExtractor } from '../auto/feature_extraction_auto.js';
19
+ //# sourceMappingURL=processing_granite_speech.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"processing_granite_speech.d.ts","sourceRoot":"","sources":["../../../src/models/granite_speech/processing_granite_speech.js"],"names":[],"mappings":"AAKA;IACI,6CAAuC;IACvC,4DAAsD;IAGtD;;;;OAIG;IACH,qCAHW,MAAM,GACJ,MAAM,CAUlB;IAED;;;OAGG;IACH,YAHW,MAAM,UACN,YAAY,6BAiCtB;CACJ;0BA3DyB,2BAA2B;8BADvB,8BAA8B;qCADvB,oCAAoC"}
@@ -1,5 +1,5 @@
1
1
  /**
2
- * @typedef {object} GroundingDinoFeatureExtractorResultProps
2
+ * @typedef {Object} GroundingDinoFeatureExtractorResultProps
3
3
  * @property {import('../../utils/tensor.js').Tensor} pixel_mask
4
4
  * @typedef {import('../../image_processors_utils.js').ImageProcessorResult & GroundingDinoFeatureExtractorResultProps} GroundingDinoFeatureExtractorResult
5
5
  */
@@ -1,23 +1,7 @@
1
- export class Idefics3PreTrainedModel extends PreTrainedModel {
2
- }
3
1
  /**
4
2
  * The Idefics3 model which consists of a vision backbone and a language model.
5
3
  */
6
- export class Idefics3ForConditionalGeneration extends Idefics3PreTrainedModel {
7
- encode_image({ pixel_values, pixel_attention_mask }: {
8
- pixel_values: any;
9
- pixel_attention_mask: any;
10
- }): Promise<any>;
11
- _merge_input_ids_with_image_features(kwargs: any): {
12
- inputs_embeds: any;
13
- attention_mask: any;
14
- };
15
- }
16
- /**
17
- * The SmolVLM Model with a language modeling head.
18
- * It is made up a SigLIP vision encoder, with a language modeling head on top.
19
- */
20
- export class SmolVLMForConditionalGeneration extends Idefics3ForConditionalGeneration {
4
+ export class Idefics3ForConditionalGeneration extends LlavaForConditionalGeneration {
21
5
  }
22
- import { PreTrainedModel } from '../modeling_utils.js';
6
+ import { LlavaForConditionalGeneration } from '../llava/modeling_llava.js';
23
7
  //# sourceMappingURL=modeling_idefics3.d.ts.map
@@ -1 +1 @@
1
- {"version":3,"file":"modeling_idefics3.d.ts","sourceRoot":"","sources":["../../../src/models/idefics3/modeling_idefics3.js"],"names":[],"mappings":"AAGA;CASC;AAED;;GAEG;AACH;IACI;;;qBAIC;IAED;;;MAUC;CACJ;AAED;;;GAGG;AACH;CAAwF;gCAzCX,sBAAsB"}
1
+ {"version":3,"file":"modeling_idefics3.d.ts","sourceRoot":"","sources":["../../../src/models/idefics3/modeling_idefics3.js"],"names":[],"mappings":"AAEA;;GAEG;AACH;CASC;8CAd6C,4BAA4B"}
@@ -1,6 +1,7 @@
1
1
  export * from "./beit/image_processing_beit.js";
2
2
  export * from "./bit/image_processing_bit.js";
3
3
  export * from "./chinese_clip/image_processing_chinese_clip.js";
4
+ export * from "./chmv2/image_processing_chmv2.js";
4
5
  export * from "./clip/image_processing_clip.js";
5
6
  export * from "./convnext/image_processing_convnext.js";
6
7
  export * from "./deit/image_processing_deit.js";
@@ -9,11 +10,14 @@ export * from "./dinov3_vit/image_processing_dinov3_vit.js";
9
10
  export * from "./donut/image_processing_donut.js";
10
11
  export * from "./dpt/image_processing_dpt.js";
11
12
  export * from "./efficientnet/image_processing_efficientnet.js";
13
+ export * from "./gemma3/image_processing_gemma3.js";
14
+ export * from "./glm46v/image_processing_glm46v.js";
12
15
  export * from "./glpn/image_processing_glpn.js";
13
16
  export * from "./grounding_dino/image_processing_grounding_dino.js";
14
17
  export * from "./idefics3/image_processing_idefics3.js";
15
18
  export * from "./janus/image_processing_janus.js";
16
19
  export * from "./jina_clip/image_processing_jina_clip.js";
20
+ export * from "./lfm2_vl/image_processing_lfm2_vl.js";
17
21
  export * from "./llava_onevision/image_processing_llava_onevision.js";
18
22
  export * from "./mask2former/image_processing_mask2former.js";
19
23
  export * from "./maskformer/image_processing_maskformer.js";
@@ -0,0 +1,41 @@
1
+ export class Lfm2VlImageProcessor extends ImageProcessor {
2
+ constructor(config: Record<string, any>);
3
+ downsample_factor: any;
4
+ do_image_splitting: any;
5
+ min_tiles: any;
6
+ max_tiles: any;
7
+ use_thumbnail: any;
8
+ min_image_tokens: any;
9
+ max_image_tokens: any;
10
+ encoder_patch_size: any;
11
+ tile_size: any;
12
+ max_pixels_tolerance: any;
13
+ return_row_col_info: any;
14
+ max_num_patches: number;
15
+ /**
16
+ * Check if the image is too large to be processed as a single tile.
17
+ * @param {number} height
18
+ * @param {number} width
19
+ * @returns {boolean}
20
+ */
21
+ _is_image_too_large(height: number, width: number): boolean;
22
+ /**
23
+ * Get the grid layout for tiling a large image.
24
+ * @param {number} height
25
+ * @param {number} width
26
+ * @returns {{ grid_width: number, grid_height: number, target_width: number, target_height: number }}
27
+ */
28
+ _get_grid_layout(height: number, width: number): {
29
+ grid_width: number;
30
+ grid_height: number;
31
+ target_width: number;
32
+ target_height: number;
33
+ };
34
+ /** @param {RawImage|RawImage[]|RawImage[][]} images */
35
+ _call(images: RawImage | RawImage[] | RawImage[][], { return_row_col_info }?: {
36
+ return_row_col_info?: any;
37
+ }): Promise<Record<string, any>>;
38
+ }
39
+ export type RawImage = import("../../utils/image.js").RawImage;
40
+ import { ImageProcessor } from '../../image_processors_utils.js';
41
+ //# sourceMappingURL=image_processing_lfm2_vl.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"image_processing_lfm2_vl.d.ts","sourceRoot":"","sources":["../../../src/models/lfm2_vl/image_processing_lfm2_vl.js"],"names":[],"mappings":"AAiIA;IACI,oBAAuB,MAAM,CAAC,MAAM,EAAE,GAAG,CAAC,EAiBzC;IAfG,uBAAsD;IACtD,wBAA2D;IAC3D,eAAsC;IACtC,eAAuC;IACvC,mBAAiD;IACjD,sBAAqD;IACrD,sBAAsD;IACtD,wBAA8E;IAC9E,eAAwC;IACxC,0BAA8D;IAC9D,yBAA8D;IAI9D,wBAAyE;IAG7E;;;;;OAKG;IACH,4BAJW,MAAM,SACN,MAAM,GACJ,OAAO,CAUnB;IAED;;;;;OAKG;IACH,yBAJW,MAAM,SACN,MAAM,GACJ;QAAE,UAAU,EAAE,MAAM,CAAC;QAAC,WAAW,EAAE,MAAM,CAAC;QAAC,YAAY,EAAE,MAAM,CAAC;QAAC,aAAa,EAAE,MAAM,CAAA;KAAE,CAiBpG;IAED,uDAAuD;IAEvD,cAFY,QAAQ,GAAC,QAAQ,EAAE,GAAC,QAAQ,EAAE,EAAE;;qCAmH3C;CACJ;uBA5SY,OAAO,sBAAsB,EAAE,QAAQ;+BAJP,iCAAiC"}
@@ -0,0 +1,4 @@
1
+ export class Lfm2VlForConditionalGeneration extends LlavaForConditionalGeneration {
2
+ }
3
+ import { LlavaForConditionalGeneration } from '../llava/modeling_llava.js';
4
+ //# sourceMappingURL=modeling_lfm2_vl.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"modeling_lfm2_vl.d.ts","sourceRoot":"","sources":["../../../src/models/lfm2_vl/modeling_lfm2_vl.js"],"names":[],"mappings":"AAEA;CAUC;8CAZ6C,4BAA4B"}
@@ -0,0 +1,18 @@
1
+ /**
2
+ * @typedef {import('../../utils/image.js').RawImage} RawImage
3
+ */
4
+ export class Lfm2VlProcessor extends Processor {
5
+ static tokenizer_class: typeof AutoTokenizer;
6
+ static image_processor_class: typeof AutoImageProcessor;
7
+ /**
8
+ * @param {RawImage|RawImage[]} images
9
+ * @param {string|string[]|null} [text]
10
+ * @param {Record<string, any>} [kwargs]
11
+ */
12
+ _call(images: RawImage | RawImage[], text?: string | string[] | null, kwargs?: Record<string, any>): Promise<any>;
13
+ }
14
+ export type RawImage = import("../../utils/image.js").RawImage;
15
+ import { Processor } from '../../processing_utils.js';
16
+ import { AutoTokenizer } from '../auto/tokenization_auto.js';
17
+ import { AutoImageProcessor } from '../auto/image_processing_auto.js';
18
+ //# sourceMappingURL=processing_lfm2_vl.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"processing_lfm2_vl.d.ts","sourceRoot":"","sources":["../../../src/models/lfm2_vl/processing_lfm2_vl.js"],"names":[],"mappings":"AAIA;;GAEG;AAEH;IACI,6CAAuC;IACvC,wDAAkD;IAElD;;;;OAIG;IACH,cAJW,QAAQ,GAAC,QAAQ,EAAE,SACnB,MAAM,GAAC,MAAM,EAAE,GAAC,IAAI,WACpB,MAAM,CAAC,MAAM,EAAE,GAAG,CAAC,gBA4D7B;CACJ;uBAvEY,OAAO,sBAAsB,EAAE,QAAQ;0BAL1B,2BAA2B;8BAEvB,8BAA8B;mCADzB,kCAAkC"}
@@ -0,0 +1,4 @@
1
+ export class LightOnOcrForConditionalGeneration extends LlavaForConditionalGeneration {
2
+ }
3
+ import { LlavaForConditionalGeneration } from '../llava/modeling_llava.js';
4
+ //# sourceMappingURL=modeling_lighton_ocr.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"modeling_lighton_ocr.d.ts","sourceRoot":"","sources":["../../../src/models/lighton_ocr/modeling_lighton_ocr.js"],"names":[],"mappings":"AAEA;CAAwF;8CAF1C,4BAA4B"}
@@ -1 +1 @@
1
- {"version":3,"file":"tokenization_marian.d.ts","sourceRoot":"","sources":["../../../src/models/marian/tokenization_marian.js"],"names":[],"mappings":"AAGA;;;GAGG;AACH;IASQ,sBAAqC;IAErC,mCAA6G;CAoCpH;oCAtDmC,6BAA6B"}
1
+ {"version":3,"file":"tokenization_marian.d.ts","sourceRoot":"","sources":["../../../src/models/marian/tokenization_marian.js"],"names":[],"mappings":"AAIA;;;GAGG;AACH;IASQ,sBAAqC;IAErC,mCAA6G;CAoCpH;oCAvDmC,6BAA6B"}
@@ -1,4 +1,4 @@
1
- export class Mistral3ForConditionalGeneration extends LlavaQwen2ForCausalLM {
1
+ export class Mistral3ForConditionalGeneration extends LlavaForConditionalGeneration {
2
2
  }
3
- import { LlavaQwen2ForCausalLM } from '../llava/modeling_llava.js';
3
+ import { LlavaForConditionalGeneration } from '../llava/modeling_llava.js';
4
4
  //# sourceMappingURL=modeling_mistral3.d.ts.map
@@ -1 +1 @@
1
- {"version":3,"file":"modeling_mistral3.d.ts","sourceRoot":"","sources":["../../../src/models/mistral3/modeling_mistral3.js"],"names":[],"mappings":"AAEA;CAA8E;sCAFxC,4BAA4B"}
1
+ {"version":3,"file":"modeling_mistral3.d.ts","sourceRoot":"","sources":["../../../src/models/mistral3/modeling_mistral3.js"],"names":[],"mappings":"AAEA;CAAsF;8CAFxC,4BAA4B"}
@@ -0,0 +1,8 @@
1
+ export class Mistral4PreTrainedModel extends PreTrainedModel {
2
+ }
3
+ export class Mistral4Model extends Mistral4PreTrainedModel {
4
+ }
5
+ export class Mistral4ForCausalLM extends Mistral4PreTrainedModel {
6
+ }
7
+ import { PreTrainedModel } from '../modeling_utils.js';
8
+ //# sourceMappingURL=modeling_mistral4.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"modeling_mistral4.d.ts","sourceRoot":"","sources":["../../../src/models/mistral4/modeling_mistral4.js"],"names":[],"mappings":"AAEA;CAA+D;AAC/D;CAA6D;AAC7D;CAAmE;gCAJnC,sBAAsB"}
@@ -10,6 +10,18 @@ export function registerTaskMappings(mappings: any): void;
10
10
  * @private
11
11
  */
12
12
  export function boolTensor(value: boolean): Tensor;
13
+ /**
14
+ * Get the session configuration for a given model type.
15
+ * @param {number} modelType The model type enum value.
16
+ * @param {Object} config The model config.
17
+ * @param {Object} [options] Loading options.
18
+ * @returns {{ sessions: Record<string, string>, cache_sessions?: Record<string, true>, optional_configs?: Record<string, string> }}
19
+ */
20
+ export function getSessionsConfig(modelType: number, config: any, options?: any): {
21
+ sessions: Record<string, string>;
22
+ cache_sessions?: Record<string, true>;
23
+ optional_configs?: Record<string, string>;
24
+ };
13
25
  /**
14
26
  * Perform forward pass on the seq2seq model (both encoder and decoder).
15
27
  * @param {Object} self The seq2seq model object.
@@ -41,28 +53,28 @@ export function decoder_forward(self: any, model_inputs: any, is_encoder_decoder
41
53
  * @param {Object} params Additional parameters.
42
54
  * @param {Function} [params.encode_function] The function to encode the modality values.
43
55
  * @param {Function} [params.merge_function] The function to merge the modality features with the input embeddings.
44
- * @param {string} [params.modality_input_name] The modality input name.
56
+ * @param {string[]} [params.modality_input_names] The modality input name.
45
57
  * @param {string} [params.modality_output_name] The modality output name.
46
58
  * @param {Tensor} [params.input_ids=null]
47
59
  * @param {Tensor} [params.attention_mask=null]
48
60
  * @param {Tensor} [params.position_ids=null]
49
61
  * @param {Tensor} [params.inputs_embeds=null]
50
- * @param {Tensor} [params.past_key_values=null]
62
+ * @param {DynamicCache} [params.past_key_values=null]
51
63
  * @param {Object} [params.generation_config=null]
52
64
  * @param {Object} [params.logits_processor=null]
53
65
  * @returns {Promise<Tensor>} The model's output tensor
54
66
  * @private
55
67
  */
56
- export function generic_text_to_text_forward(self: any, { encode_function, merge_function, modality_input_name, modality_output_name, input_ids, attention_mask, position_ids, inputs_embeds, past_key_values, generation_config, logits_processor, ...kwargs }: {
68
+ export function generic_text_to_text_forward(self: any, { encode_function, merge_function, modality_input_names, modality_output_name, input_ids, attention_mask, position_ids, inputs_embeds, past_key_values, generation_config, logits_processor, ...kwargs }: {
57
69
  encode_function?: Function;
58
70
  merge_function?: Function;
59
- modality_input_name?: string;
71
+ modality_input_names?: string[];
60
72
  modality_output_name?: string;
61
73
  input_ids?: Tensor;
62
74
  attention_mask?: Tensor;
63
75
  position_ids?: Tensor;
64
76
  inputs_embeds?: Tensor;
65
- past_key_values?: Tensor;
77
+ past_key_values?: DynamicCache;
66
78
  generation_config?: any;
67
79
  logits_processor?: any;
68
80
  }): Promise<Tensor>;
@@ -168,6 +180,7 @@ export namespace MODEL_TYPES {
168
180
  let ImageAudioTextToText: number;
169
181
  let Supertonic: number;
170
182
  let Chatterbox: number;
183
+ let VoxtralRealtime: number;
171
184
  }
172
185
  export const MODEL_TYPE_MAPPING: Map<any, any>;
173
186
  export const MODEL_NAME_TO_CLASS_MAPPING: Map<any, any>;
@@ -209,8 +222,8 @@ export class PreTrainedModel extends PreTrainedModel_base {
209
222
  config: import("../configs.js").PretrainedConfig;
210
223
  sessions: Record<string, any>;
211
224
  configs: Record<string, any>;
212
- can_generate: boolean;
213
- _forward: typeof decoder_forward | typeof auto_encoder_forward | typeof encoder_forward;
225
+ can_generate: any;
226
+ _forward: any;
214
227
  _prepare_inputs_for_generation: any;
215
228
  /** @type {import('../configs.js').TransformersJSConfig} */
216
229
  custom_config: import("../configs.js").TransformersJSConfig;
@@ -257,9 +270,9 @@ export class PreTrainedModel extends PreTrainedModel_base {
257
270
  /**
258
271
  *
259
272
  * @param {GenerationConfig} generation_config
260
- * @param {StoppingCriteriaList} [stopping_criteria=null]
273
+ * @param {import('../generation/stopping_criteria.js').StoppingCriteria|import('../generation/stopping_criteria.js').StoppingCriteria[]|StoppingCriteriaList} [stopping_criteria=null]
261
274
  */
262
- _get_stopping_criteria(generation_config: GenerationConfig, stopping_criteria?: StoppingCriteriaList): StoppingCriteriaList;
275
+ _get_stopping_criteria(generation_config: GenerationConfig, stopping_criteria?: import("../generation/stopping_criteria.js").StoppingCriteria | import("../generation/stopping_criteria.js").StoppingCriteria[] | StoppingCriteriaList): StoppingCriteriaList;
263
276
  /**
264
277
  * Confirms that the model class is compatible with generation.
265
278
  * If not, raises an exception that points to the right class to use.
@@ -287,7 +300,7 @@ export class PreTrainedModel extends PreTrainedModel_base {
287
300
  * @param {Tensor} [params.inputs=null]
288
301
  * @param {number} [params.bos_token_id=null]
289
302
  * @param {Record<string, Tensor|number[]>} [params.model_kwargs]
290
- * @returns {{inputs_tensor: Tensor, model_inputs: Record<string, Tensor>, model_input_name: string}} The model-specific inputs for generation.
303
+ * @returns {{inputs_tensor: Tensor, model_inputs: Record<string, Tensor> & {past_key_values?: DynamicCache}, model_input_name: string}} The model-specific inputs for generation.
291
304
  */
292
305
  _prepare_model_inputs({ inputs, bos_token_id, model_kwargs }: {
293
306
  inputs?: Tensor;
@@ -295,7 +308,9 @@ export class PreTrainedModel extends PreTrainedModel_base {
295
308
  model_kwargs?: Record<string, Tensor | number[]>;
296
309
  }): {
297
310
  inputs_tensor: Tensor;
298
- model_inputs: Record<string, Tensor>;
311
+ model_inputs: Record<string, Tensor> & {
312
+ past_key_values?: DynamicCache;
313
+ };
299
314
  model_input_name: string;
300
315
  };
301
316
  _prepare_encoder_decoder_kwargs_for_generation({ inputs_tensor, model_inputs, model_input_name, generation_config, }: {
@@ -319,13 +334,14 @@ export class PreTrainedModel extends PreTrainedModel_base {
319
334
  */
320
335
  generate({ inputs, generation_config, logits_processor, stopping_criteria, streamer, ...kwargs }: import("../generation/parameters.js").GenerationFunctionParameters): Promise<ModelOutput | Tensor>;
321
336
  /**
322
- * Returns an object containing past key values from the given decoder results object.
337
+ * Returns a DynamicCache containing past key values from the given decoder results object.
323
338
  *
324
339
  * @param {Object} decoderResults The decoder results object.
325
- * @param {Object} pastKeyValues The previous past key values.
326
- * @returns {Object} An object containing past key values.
340
+ * @param {DynamicCache} pastKeyValues The previous past key values.
341
+ * @param {boolean} [disposeEncoderPKVs=false] Whether to dispose encoder past key values.
342
+ * @returns {DynamicCache} A new DynamicCache containing the updated past key values.
327
343
  */
328
- getPastKeyValues(decoderResults: any, pastKeyValues: any, disposeEncoderPKVs?: boolean): any;
344
+ getPastKeyValues(decoderResults: any, pastKeyValues: DynamicCache, disposeEncoderPKVs?: boolean): DynamicCache;
329
345
  /**
330
346
  * Returns an object containing attentions from the given model output object.
331
347
  *
@@ -338,22 +354,25 @@ export class PreTrainedModel extends PreTrainedModel_base {
338
354
  /**
339
355
  * Adds past key values to the decoder feeds object. If pastKeyValues is null, creates new tensors for past key values.
340
356
  *
341
- * @param {Object} decoderFeeds The decoder feeds object to add past key values to.
342
- * @param {Object} pastKeyValues An object containing past key values.
357
+ * @param {Record<string, any>} decoderFeeds The decoder feeds object to add past key values to.
358
+ * @param {DynamicCache|null} pastKeyValues The cache containing past key values.
343
359
  */
344
- addPastKeyValues(decoderFeeds: any, pastKeyValues: any): void;
345
- encode_image({ pixel_values }: {
346
- pixel_values: any;
347
- }): Promise<any>;
348
- encode_text({ input_ids }: {
349
- input_ids: any;
350
- }): Promise<any>;
351
- encode_audio({ audio_values }: {
352
- audio_values: any;
353
- }): Promise<any>;
360
+ addPastKeyValues(decoderFeeds: Record<string, any>, pastKeyValues: DynamicCache | null): void;
361
+ /**
362
+ * Helper function to select valid inputs and run through the appropriate encoder (vision, text, audio) based on the input type.
363
+ * @param {string} sessionName
364
+ * @param {Record<string, Tensor>} inputs
365
+ * @param {string} outputName
366
+ * @private
367
+ */
368
+ private _encode_input;
369
+ encode_image(inputs: any): Promise<any>;
370
+ encode_text(inputs: any): Promise<any>;
371
+ encode_audio(inputs: any): Promise<any>;
354
372
  }
355
373
  import { Tensor } from '../utils/tensor.js';
356
374
  import { Seq2SeqLMOutput } from './modeling_outputs.js';
375
+ import { DynamicCache } from '../cache_utils.js';
357
376
  import { GenerationConfig } from '../generation/configuration_utils.js';
358
377
  import { StoppingCriteriaList } from '../generation/stopping_criteria.js';
359
378
  import { ModelOutput } from './modeling_outputs.js';
@@ -1 +1 @@
1
- {"version":3,"file":"modeling_utils.d.ts","sourceRoot":"","sources":["../../src/models/modeling_utils.js"],"names":[],"mappings":"AASA;;;GAGG;AACH,0DAEC;AAyDD;;;;;GAKG;AACH,kCAJW,OAAO,GACL,MAAM,CAKlB;AAotCD;;;;;;GAMG;AACH,+DAHa,OAAO,CAAC,eAAe,CAAC,CAoBpC;AAED;;;;;;GAMG;AACH,+DAHa,OAAO,KAAQ,CAgC3B;AAED,iFAIC;AAED;;;;;;GAMG;AACH,6FAHa,OAAO,KAAQ,CA2B3B;AAED;;;;;;;;;;;;;;;;;GAiBG;AACH,iQAdG;IAA0B,eAAe;IACf,cAAc;IAChB,mBAAmB,GAAnC,MAAM;IACU,oBAAoB,GAApC,MAAM;IACU,SAAS,GAAzB,MAAM;IACU,cAAc,GAA9B,MAAM;IACU,YAAY,GAA5B,MAAM;IACU,aAAa,GAA7B,MAAM;IACU,eAAe,GAA/B,MAAM;IACU,iBAAiB;IACjB,gBAAgB;CACxC,GAAU,OAAO,CAAC,MAAM,CAAC,CAsF3B;AAED;;;;;;GAMG;AACH,oEAHa,OAAO,CAAC,MAAM,CAAC,CAW3B;AAED;;;;;;GAMG;AACH,oEAHa,OAAO,CAAC,MAAM,CAAC,CAW3B;AAED;;;;;;;;GAQG;AACH,mDAHW,MAAM,yBACJ;IAAC,IAAI,EAAE,aAAa,CAAC;IAAC,IAAI,EAAE,MAAM,EAAE,CAAA;CAAC,CAsBjD;AAED;;;;;;;;;;;GAWG;AACH,4GAUC;AAED,iIA2CC;AAED,yIASC;AAED,sGAMC;AAED;;;;;;;;;EA6BC;AAED;;;;;;;;;EAcC;AAED;;;;;;;;;EAcC;AAED;;;;;;;GAOG;AACH,oEANW,MAAM,SACN,MAAM,CAAC,MAAM,EAAE,MAAM,CAAC,WACtB,OAAO,iBAAiB,EAAE,sBAAsB,GAC9C,OAAO,CAAC,MAAM,CAAC,MAAM,EAAE,GAAG,CAAC,CAAC,CAYxC;AAptDD,oCAAsC;;;;;;;;;;;;;;;;;;;AA8JtC,+CAA4C;AAC5C,wDAAqD;AACrD,wDAAqD;;KArKb,GAAG;UAClC,GAAE;;AAsKX;;GAEG;AACH;IAiDI;;;;;;;;;;;;;;OAcG;IACH,sDATW,MAAM,gKAKN,OAAO,iBAAiB,EAAE,sBAAsB,GAE9C,OAAO,CAAC,eAAe,CAAC,CAwRpC;IAjVD;;;;;OAKG;IACH,oBAJW,OAAO,eAAe,EAAE,gBAAgB,YACxC,MAAM,CAAC,MAAM,EAAE,GAAG,CAAC,WACnB,MAAM,CAAC,MAAM,MAAS,EAyBhC;IAjCD,wBAA8B;IAC9B,yBAAiD;IAEjD,mCAAqC;IAUjC,iDAAoB;IACpB,8BAAwB;IACxB,6BAAsB;IAQtB,sBAA2C;IAC3C,wFAAkC;IAClC,oCAA+D;IAM/D,2DAA2D;IAC3D,eADW,OAAO,eAAe,EAAE,oBAAoB,CACS;IAGpE;;;;OAIG;IACH,WAHa,OAAO,CAAC,OAAO,EAAE,CAAC,CAS9B;IAySD;;;;OAIG;IACH,0BAFa,OAAO,KAAQ,CAI3B;IAED;;;;;;OAMG;IACH,4BAHa,OAAO,KAAQ,CAK3B;IAED;;;OAGG;IACH,yBAFa,gBAAgB,GAAC,IAAI,CAIjC;IAED;;;;;OAKG;IACH,8BA2JC;IAED;;;;;;OAMG;IACH,8CAJW,gBAAgB,GAAC,IAAI,+CAEnB,gBAAgB,CA+B5B;IAED;;;;OAIG;IACH,0CAHW,gBAAgB,sBAChB,oBAAoB,wBAqB9B;IAED;;;OAGG;IACH,8BA2BC;IAED,mDAKC;IAED;;;;;;;;OAQG;IACH,wGANG;QAA2B,mBAAmB,EAAtC,MAAM,EAAE,EAAE;QACK,OAAO;QACP,YAAY;QACX,kBAAkB,EAAlC,OAAO;KACf,OAuBF;IAED;;;;;;;OAOG;IACH,8DALG;QAAwB,MAAM,GAAtB,MAAM;QACU,YAAY,GAA5B,MAAM;QACmC,YAAY,GAArD,MAAM,CAAC,MAAM,EAAE,MAAM,GAAC,MAAM,EAAE,CAAC;KACvC,GAAU;QAAC,aAAa,EAAE,MAAM,CAAC;QAAC,YAAY,EAAE,MAAM,CAAC,MAAM,EAAE,MAAM,CAAC,CAAC;QAAC,gBAAgB,EAAE,MAAM,CAAA;KAAC,CAmBnG;IAED;;;;;qBAmDC;IAED;;;OAGG;IACH,oJAFW,GAAC;;;MAwDX;IAED;;;;OAIG;IACH,kGAHW,OAAO,6BAA6B,EAAE,4BAA4B,GAChE,OAAO,CAAC,WAAW,GAAC,MAAM,CAAC,CAgOvC;IAED;;;;;;OAMG;IACH,6FAkCC;IAED;;;;;OAKG;IACH,kCAFa;QAAC,gBAAgB,CAAC,EAAE,MAAM,EAAE,CAAA;KAAC,CAgBzC;IAED;;;;;OAKG;IACH,8DAeC;IAED;;qBAGC;IAED;;qBAGC;IAED;;qBAGC;CACJ;uBA/xCmE,oBAAoB;gCAexD,uBAAuB;iCActB,sCAAsC;qCACG,oCAAoC;4BAGlF,uBAAuB"}
1
+ {"version":3,"file":"modeling_utils.d.ts","sourceRoot":"","sources":["../../src/models/modeling_utils.js"],"names":[],"mappings":"AASA;;;GAGG;AACH,0DAEC;AA2DD;;;;;GAKG;AACH,kCAJW,OAAO,GACL,MAAM,CAKlB;AAuLD;;;;;;GAMG;AACH,6CALW,MAAM,+BAGJ;IAAE,QAAQ,EAAE,MAAM,CAAC,MAAM,EAAE,MAAM,CAAC,CAAC;IAAC,cAAc,CAAC,EAAE,MAAM,CAAC,MAAM,EAAE,IAAI,CAAC,CAAC;IAAC,gBAAgB,CAAC,EAAE,MAAM,CAAC,MAAM,EAAE,MAAM,CAAC,CAAA;CAAE,CASlI;AAg9BD;;;;;;GAMG;AACH,+DAHa,OAAO,CAAC,eAAe,CAAC,CAyBpC;AAED;;;;;;GAMG;AACH,+DAHa,OAAO,KAAQ,CAgC3B;AAED,iFAIC;AAED;;;;;;GAMG;AACH,6FAHa,OAAO,KAAQ,CAoC3B;AAED;;;;;;;;;;;;;;;;;GAiBG;AACH,kQAdG;IAA0B,eAAe;IACf,cAAc;IACd,oBAAoB,GAAtC,MAAM,EAAE;IACQ,oBAAoB,GAApC,MAAM;IACU,SAAS,GAAzB,MAAM;IACU,cAAc,GAA9B,MAAM;IACU,YAAY,GAA5B,MAAM;IACU,aAAa,GAA7B,MAAM;IACgB,eAAe,GAArC,YAAY;IACI,iBAAiB;IACjB,gBAAgB;CACxC,GAAU,OAAO,CAAC,MAAM,CAAC,CAyG3B;AAED;;;;;;GAMG;AACH,oEAHa,OAAO,CAAC,MAAM,CAAC,CAW3B;AAED;;;;;;GAMG;AACH,oEAHa,OAAO,CAAC,MAAM,CAAC,CAW3B;AAED;;;;;;;;GAQG;AACH,mDAHW,MAAM,yBACJ;IAAC,IAAI,EAAE,aAAa,CAAC;IAAC,IAAI,EAAE,MAAM,EAAE,CAAA;CAAC,CAsBjD;AAED;;;;;;;;;;;GAWG;AACH,4GAUC;AAED,iIAkDC;AAED,yIASC;AAED,sGAMC;AAED;;;;;;;;;EAiCC;AAED;;;;;;;;;EAcC;AAED;;;;;;;;;EAcC;AAED;;;;;;;GAOG;AACH,oEANW,MAAM,SACN,MAAM,CAAC,MAAM,EAAE,MAAM,CAAC,WACtB,OAAO,iBAAiB,EAAE,sBAAsB,GAC9C,OAAO,CAAC,MAAM,CAAC,MAAM,EAAE,GAAG,CAAC,CAAC,CAYxC;AAnsDD,oCAAsC;;;;;;;;;;;;;;;;;;;;AA+StC,+CAA4C;AAC5C,wDAAqD;AACrD,wDAAqD;;KAtTb,GAAG;UAClC,GAAE;;AAuTX;;GAEG;AACH;IA8CI;;;;;;;;;;;;;;OAcG;IACH,sDATW,MAAM,gKAKN,OAAO,iBAAiB,EAAE,sBAAsB,GAE9C,OAAO,CAAC,eAAe,CAAC,CA0DpC;IAhHD;;;;;OAKG;IACH,oBAJW,OAAO,eAAe,EAAE,gBAAgB,YACxC,MAAM,CAAC,MAAM,EAAE,GAAG,CAAC,WACnB,MAAM,CAAC,MAAM,MAAS,EAsBhC;IA9BD,wBAA8B;IAC9B,yBAAiD;IAEjD,mCAAqC;IAUjC,iDAAoB;IACpB,8BAAwB;IACxB,6BAAsB;IAKtB,kBAA2C;IAC3C,cAAkC;IAClC,oCAA+D;IAM/D,2DAA2D;IAC3D,eADW,OAAO,eAAe,EAAE,oBAAoB,CACS;IAGpE;;;;OAIG;IACH,WAHa,OAAO,CAAC,OAAO,EAAE,CAAC,CAS9B;IA2ED;;;;OAIG;IACH,0BAFa,OAAO,KAAQ,CAI3B;IAED;;;;;;OAMG;IACH,4BAHa,OAAO,KAAQ,CAK3B;IAED;;;OAGG;IACH,yBAFa,gBAAgB,GAAC,IAAI,CAIjC;IAED;;;;;OAKG;IACH,8BA2JC;IAED;;;;;;OAMG;IACH,8CAJW,gBAAgB,GAAC,IAAI,+CAEnB,gBAAgB,CA+B5B;IAED;;;;OAIG;IACH,0CAHW,gBAAgB,sBAChB,OAAO,oCAAoC,EAAE,gBAAgB,GAAC,OAAO,oCAAoC,EAAE,gBAAgB,EAAE,GAAC,oBAAoB,wBAqB5J;IAED;;;OAGG;IACH,8BA2BC;IAED,mDAKC;IAED;;;;;;;;OAQG;IACH,wGANG;QAA2B,mBAAmB,EAAtC,MAAM,EAAE,EAAE;QACK,OAAO;QACP,YAAY;QACX,kBAAkB,EAAlC,OAAO;KACf,OA0BF;IAED;;;;;;;OAOG;IACH,8DALG;QAAwB,MAAM,GAAtB,MAAM;QACU,YAAY,GAA5B,MAAM;QACmC,YAAY,GAArD,MAAM,CAAC,MAAM,EAAE,MAAM,GAAC,MAAM,EAAE,CAAC;KACvC,GAAU;QAAC,aAAa,EAAE,MAAM,CAAC;QAAC,YAAY,EAAE,MAAM,CAAC,MAAM,EAAE,MAAM,CAAC,GAAG;YAAC,eAAe,CAAC,EAAE,YAAY,CAAA;SAAC,CAAC;QAAC,gBAAgB,EAAE,MAAM,CAAA;KAAC,CAmBtI;IAED;;;;;qBAmDC;IAED;;;OAGG;IACH,oJAFW,GAAC;;;MAwDX;IAED;;;;OAIG;IACH,kGAHW,OAAO,6BAA6B,EAAE,4BAA4B,GAChE,OAAO,CAAC,WAAW,GAAC,MAAM,CAAC,CAgOvC;IAED;;;;;;;OAOG;IACH,qDAJW,YAAY,uBACZ,OAAO,GACL,YAAY,CAsCxB;IAED;;;;;OAKG;IACH,kCAFa;QAAC,gBAAgB,CAAC,EAAE,MAAM,EAAE,CAAA;KAAC,CAgBzC;IAED;;;;;OAKG;IACH,+BAHW,MAAM,CAAC,MAAM,EAAE,GAAG,CAAC,iBACnB,YAAY,GAAC,IAAI,QAiB3B;IAED;;;;;;OAMG;IACH,sBAOC;IAED,wCAEC;IAED,uCAEC;IAED,wCAEC;CACJ;uBAluCmE,oBAAoB;gCAexD,uBAAuB;6BAoB1B,mBAAmB;iCANf,sCAAsC;qCACG,oCAAoC;4BAGlF,uBAAuB"}
@@ -1,7 +1,8 @@
1
1
  export * from "./albert/modeling_albert.js";
2
2
  export * from "./apertus/modeling_apertus.js";
3
+ export * from "./afmoe/modeling_afmoe.js";
3
4
  export * from "./arcee/modeling_arcee.js";
4
- export * from "./ast/modeling_ast.js";
5
+ export * from "./audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.js";
5
6
  export * from "./bart/modeling_bart.js";
6
7
  export * from "./beit/modeling_beit.js";
7
8
  export * from "./bert/modeling_bert.js";
@@ -11,17 +12,21 @@ export * from "./bloom/modeling_bloom.js";
11
12
  export * from "./camembert/modeling_camembert.js";
12
13
  export * from "./chatterbox/modeling_chatterbox.js";
13
14
  export * from "./chinese_clip/modeling_chinese_clip.js";
15
+ export * from "./chmv2/modeling_chmv2.js";
14
16
  export * from "./clap/modeling_clap.js";
15
17
  export * from "./clip/modeling_clip.js";
16
18
  export * from "./clipseg/modeling_clipseg.js";
17
19
  export * from "./codegen/modeling_codegen.js";
18
20
  export * from "./cohere/modeling_cohere.js";
21
+ export * from "./cohere2/modeling_cohere2.js";
22
+ export * from "./cohere_asr/modeling_cohere_asr.js";
19
23
  export * from "./convbert/modeling_convbert.js";
20
24
  export * from "./convnext/modeling_convnext.js";
21
25
  export * from "./convnextv2/modeling_convnextv2.js";
22
26
  export * from "./d_fine/modeling_d_fine.js";
23
27
  export * from "./dac/modeling_dac.js";
24
28
  export * from "./deberta/modeling_deberta.js";
29
+ export * from "./deepseek_v3/modeling_deepseek_v3.js";
25
30
  export * from "./deberta_v2/modeling_deberta_v2.js";
26
31
  export * from "./decision_transformer/modeling_decision_transformer.js";
27
32
  export * from "./deit/modeling_deit.js";
@@ -39,6 +44,7 @@ export * from "./efficientnet/modeling_efficientnet.js";
39
44
  export * from "./electra/modeling_electra.js";
40
45
  export * from "./ernie4_5/modeling_ernie4_5.js";
41
46
  export * from "./esm/modeling_esm.js";
47
+ export * from "./eurobert/modeling_eurobert.js";
42
48
  export * from "./exaone/modeling_exaone.js";
43
49
  export * from "./falcon/modeling_falcon.js";
44
50
  export * from "./falcon_h1/modeling_falcon_h1.js";
@@ -49,6 +55,8 @@ export * from "./gemma2/modeling_gemma2.js";
49
55
  export * from "./gemma3/modeling_gemma3.js";
50
56
  export * from "./gemma3n/modeling_gemma3n.js";
51
57
  export * from "./glm/modeling_glm.js";
58
+ export * from "./glm_moe_dsa/modeling_glm_moe_dsa.js";
59
+ export * from "./glm_ocr/modeling_glm_ocr.js";
52
60
  export * from "./glpn/modeling_glpn.js";
53
61
  export * from "./gpt_bigcode/modeling_gpt_bigcode.js";
54
62
  export * from "./gpt_neo/modeling_gpt_neo.js";
@@ -58,6 +66,7 @@ export * from "./gpt2/modeling_gpt2.js";
58
66
  export * from "./gptj/modeling_gptj.js";
59
67
  export * from "./granite/modeling_granite.js";
60
68
  export * from "./granitemoehybrid/modeling_granitemoehybrid.js";
69
+ export * from "./granite_speech/modeling_granite_speech.js";
61
70
  export * from "./grounding_dino/modeling_grounding_dino.js";
62
71
  export * from "./groupvit/modeling_groupvit.js";
63
72
  export * from "./helium/modeling_helium.js";
@@ -69,7 +78,9 @@ export * from "./ijepa/modeling_ijepa.js";
69
78
  export * from "./jais/modeling_jais.js";
70
79
  export * from "./jina_clip/modeling_jina_clip.js";
71
80
  export * from "./lfm2/modeling_lfm2.js";
81
+ export * from "./lighton_ocr/modeling_lighton_ocr.js";
72
82
  export * from "./lfm2_moe/modeling_lfm2_moe.js";
83
+ export * from "./lfm2_vl/modeling_lfm2_vl.js";
73
84
  export * from "./llama/modeling_llama.js";
74
85
  export * from "./llama4/modeling_llama4.js";
75
86
  export * from "./llava/modeling_llava.js";
@@ -84,6 +95,7 @@ export * from "./metric3dv2/modeling_metric3dv2.js";
84
95
  export * from "./mgp_str/modeling_mgp_str.js";
85
96
  export * from "./mimi/modeling_mimi.js";
86
97
  export * from "./mistral/modeling_mistral.js";
98
+ export * from "./mistral4/modeling_mistral4.js";
87
99
  export * from "./mobilebert/modeling_mobilebert.js";
88
100
  export * from "./mobilellm/modeling_mobilellm.js";
89
101
  export * from "./mobilenet_v1/modeling_mobilenet_v1.js";
@@ -101,11 +113,13 @@ export * from "./mt5/modeling_mt5.js";
101
113
  export * from "./multi_modality/modeling_multi_modality.js";
102
114
  export * from "./musicgen/modeling_musicgen.js";
103
115
  export * from "./nanochat/modeling_nanochat.js";
116
+ export * from "./nemotron_h/modeling_nemotron_h.js";
104
117
  export * from "./neobert/modeling_neobert.js";
105
118
  export * from "./nomic_bert/modeling_nomic_bert.js";
106
119
  export * from "./olmo/modeling_olmo.js";
107
120
  export * from "./olmo2/modeling_olmo2.js";
108
121
  export * from "./olmo3/modeling_olmo3.js";
122
+ export * from "./olmo_hybrid/modeling_olmo_hybrid.js";
109
123
  export * from "./openelm/modeling_openelm.js";
110
124
  export * from "./opt/modeling_opt.js";
111
125
  export * from "./owlv2/modeling_owlv2.js";
@@ -120,8 +134,16 @@ export * from "./phi3_v/modeling_phi3_v.js";
120
134
  export * from "./pvt/modeling_pvt.js";
121
135
  export * from "./pyannote/modeling_pyannote.js";
122
136
  export * from "./qwen2/modeling_qwen2.js";
137
+ export * from "./qwen2_moe/modeling_qwen2_moe.js";
123
138
  export * from "./qwen2_vl/modeling_qwen2_vl.js";
139
+ export * from "./qwen2_5_vl/modeling_qwen2_5_vl.js";
124
140
  export * from "./qwen3/modeling_qwen3.js";
141
+ export * from "./qwen3_moe/modeling_qwen3_moe.js";
142
+ export * from "./qwen3_next/modeling_qwen3_next.js";
143
+ export * from "./qwen3_vl/modeling_qwen3_vl.js";
144
+ export * from "./qwen3_vl_moe/modeling_qwen3_vl_moe.js";
145
+ export * from "./qwen3_5/modeling_qwen3_5.js";
146
+ export * from "./qwen3_5_moe/modeling_qwen3_5_moe.js";
125
147
  export * from "./resnet/modeling_resnet.js";
126
148
  export * from "./rf_detr/modeling_rf_detr.js";
127
149
  export * from "./roberta/modeling_roberta.js";
@@ -135,6 +157,7 @@ export * from "./segformer/modeling_segformer.js";
135
157
  export * from "./siglip/modeling_siglip.js";
136
158
  export * from "./smollm3/modeling_smollm3.js";
137
159
  export * from "./snac/modeling_snac.js";
160
+ export * from "./solar_open/modeling_solar_open.js";
138
161
  export * from "./speecht5/modeling_speecht5.js";
139
162
  export * from "./squeezebert/modeling_squeezebert.js";
140
163
  export * from "./stablelm/modeling_stablelm.js";
@@ -157,6 +180,8 @@ export * from "./vit_msn/modeling_vit_msn.js";
157
180
  export * from "./vitmatte/modeling_vitmatte.js";
158
181
  export * from "./vitpose/modeling_vitpose.js";
159
182
  export * from "./vits/modeling_vits.js";
183
+ export * from "./voxtral/modeling_voxtral.js";
184
+ export * from "./voxtral_realtime/modeling_voxtral_realtime.js";
160
185
  export * from "./wav2vec2/modeling_wav2vec2.js";
161
186
  export * from "./wav2vec2_bert/modeling_wav2vec2_bert.js";
162
187
  export * from "./wavlm/modeling_wavlm.js";
@@ -0,0 +1,8 @@
1
+ export class NemotronHPreTrainedModel extends PreTrainedModel {
2
+ }
3
+ export class NemotronHModel extends NemotronHPreTrainedModel {
4
+ }
5
+ export class NemotronHForCausalLM extends NemotronHPreTrainedModel {
6
+ }
7
+ import { PreTrainedModel } from '../modeling_utils.js';
8
+ //# sourceMappingURL=modeling_nemotron_h.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"modeling_nemotron_h.d.ts","sourceRoot":"","sources":["../../../src/models/nemotron_h/modeling_nemotron_h.js"],"names":[],"mappings":"AAEA;CAAgE;AAChE;CAA+D;AAC/D;CAAqE;gCAJrC,sBAAsB"}
@@ -0,0 +1,8 @@
1
+ export class OlmoHybridPreTrainedModel extends PreTrainedModel {
2
+ }
3
+ export class OlmoHybridModel extends OlmoHybridPreTrainedModel {
4
+ }
5
+ export class OlmoHybridForCausalLM extends OlmoHybridPreTrainedModel {
6
+ }
7
+ import { PreTrainedModel } from '../modeling_utils.js';
8
+ //# sourceMappingURL=modeling_olmo_hybrid.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"modeling_olmo_hybrid.d.ts","sourceRoot":"","sources":["../../../src/models/olmo_hybrid/modeling_olmo_hybrid.js"],"names":[],"mappings":"AAEA;CAAiE;AACjE;CAAiE;AACjE;CAAuE;gCAJvC,sBAAsB"}
@@ -1,10 +1,4 @@
1
- export class PaliGemmaPreTrainedModel extends PreTrainedModel {
1
+ export class PaliGemmaForConditionalGeneration extends LlavaForConditionalGeneration {
2
2
  }
3
- export class PaliGemmaForConditionalGeneration extends PaliGemmaPreTrainedModel {
4
- _merge_input_ids_with_image_features(kwargs: any): {
5
- inputs_embeds: any;
6
- attention_mask: any;
7
- };
8
- }
9
- import { PreTrainedModel } from '../modeling_utils.js';
3
+ import { LlavaForConditionalGeneration } from '../llava/modeling_llava.js';
10
4
  //# sourceMappingURL=modeling_paligemma.d.ts.map