@huggingface/transformers 3.4.0 → 3.4.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (38) hide show
  1. package/README.md +6 -2
  2. package/dist/transformers.js +315 -152
  3. package/dist/transformers.js.map +1 -1
  4. package/dist/transformers.min.js +1 -1
  5. package/dist/transformers.min.js.map +1 -1
  6. package/dist/transformers.node.cjs +303 -151
  7. package/dist/transformers.node.cjs.map +1 -1
  8. package/dist/transformers.node.min.cjs +1 -1
  9. package/dist/transformers.node.min.cjs.map +1 -1
  10. package/dist/transformers.node.min.mjs +1 -1
  11. package/dist/transformers.node.min.mjs.map +1 -1
  12. package/dist/transformers.node.mjs +315 -152
  13. package/dist/transformers.node.mjs.map +1 -1
  14. package/dist/transformers.web.js +315 -152
  15. package/dist/transformers.web.js.map +1 -1
  16. package/dist/transformers.web.min.js +1 -1
  17. package/dist/transformers.web.min.js.map +1 -1
  18. package/package.json +1 -1
  19. package/src/configs.js +2 -0
  20. package/src/env.js +1 -1
  21. package/src/models/feature_extractors.js +1 -0
  22. package/src/models/snac/feature_extraction_snac.js +3 -0
  23. package/src/models.js +90 -2
  24. package/src/pipelines.js +140 -135
  25. package/src/utils/image.js +9 -1
  26. package/src/utils/tensor.js +6 -2
  27. package/types/configs.d.ts.map +1 -1
  28. package/types/models/feature_extractors.d.ts +1 -0
  29. package/types/models/snac/feature_extraction_snac.d.ts +4 -0
  30. package/types/models/snac/feature_extraction_snac.d.ts.map +1 -0
  31. package/types/models.d.ts +48 -0
  32. package/types/models.d.ts.map +1 -1
  33. package/types/pipelines.d.ts +2 -2
  34. package/types/pipelines.d.ts.map +1 -1
  35. package/types/tsconfig.tsbuildinfo +1 -1
  36. package/types/utils/image.d.ts +2 -2
  37. package/types/utils/image.d.ts.map +1 -1
  38. package/types/utils/tensor.d.ts.map +1 -1
@@ -3676,6 +3676,7 @@ function getNormalizedConfig(config) {
3676
3676
  // Sub-configs
3677
3677
  case 'llava':
3678
3678
  case 'paligemma':
3679
+ case 'gemma3':
3679
3680
  case 'florence2':
3680
3681
  case 'llava_onevision':
3681
3682
  case 'idefics3':
@@ -3735,6 +3736,7 @@ function getNormalizedConfig(config) {
3735
3736
  break;
3736
3737
  case 'gemma':
3737
3738
  case 'gemma2':
3739
+ case 'gemma3_text':
3738
3740
  case 'glm':
3739
3741
  case 'helium':
3740
3742
  mapping['num_heads'] = 'num_key_value_heads';
@@ -4065,7 +4067,7 @@ __webpack_require__.r(__webpack_exports__);
4065
4067
 
4066
4068
 
4067
4069
 
4068
- const VERSION = '3.4.0';
4070
+ const VERSION = '3.4.1';
4069
4071
 
4070
4072
  // Check if various APIs are available (depends on environment)
4071
4073
  const IS_BROWSER_ENV = typeof window !== "undefined" && typeof window.document !== "undefined";
@@ -6203,6 +6205,9 @@ __webpack_require__.r(__webpack_exports__);
6203
6205
  /* harmony export */ Gemma2ForCausalLM: () => (/* binding */ Gemma2ForCausalLM),
6204
6206
  /* harmony export */ Gemma2Model: () => (/* binding */ Gemma2Model),
6205
6207
  /* harmony export */ Gemma2PreTrainedModel: () => (/* binding */ Gemma2PreTrainedModel),
6208
+ /* harmony export */ Gemma3ForCausalLM: () => (/* binding */ Gemma3ForCausalLM),
6209
+ /* harmony export */ Gemma3Model: () => (/* binding */ Gemma3Model),
6210
+ /* harmony export */ Gemma3PreTrainedModel: () => (/* binding */ Gemma3PreTrainedModel),
6206
6211
  /* harmony export */ GemmaForCausalLM: () => (/* binding */ GemmaForCausalLM),
6207
6212
  /* harmony export */ GemmaModel: () => (/* binding */ GemmaModel),
6208
6213
  /* harmony export */ GemmaPreTrainedModel: () => (/* binding */ GemmaPreTrainedModel),
@@ -6273,6 +6278,10 @@ __webpack_require__.r(__webpack_exports__);
6273
6278
  /* harmony export */ MaskFormerModel: () => (/* binding */ MaskFormerModel),
6274
6279
  /* harmony export */ MaskFormerPreTrainedModel: () => (/* binding */ MaskFormerPreTrainedModel),
6275
6280
  /* harmony export */ MaskedLMOutput: () => (/* binding */ MaskedLMOutput),
6281
+ /* harmony export */ Metric3DForDepthEstimation: () => (/* binding */ Metric3DForDepthEstimation),
6282
+ /* harmony export */ Metric3DPreTrainedModel: () => (/* binding */ Metric3DPreTrainedModel),
6283
+ /* harmony export */ Metric3Dv2ForDepthEstimation: () => (/* binding */ Metric3Dv2ForDepthEstimation),
6284
+ /* harmony export */ Metric3Dv2PreTrainedModel: () => (/* binding */ Metric3Dv2PreTrainedModel),
6276
6285
  /* harmony export */ MgpstrForSceneTextRecognition: () => (/* binding */ MgpstrForSceneTextRecognition),
6277
6286
  /* harmony export */ MgpstrModelOutput: () => (/* binding */ MgpstrModelOutput),
6278
6287
  /* harmony export */ MgpstrPreTrainedModel: () => (/* binding */ MgpstrPreTrainedModel),
@@ -6421,6 +6430,10 @@ __webpack_require__.r(__webpack_exports__);
6421
6430
  /* harmony export */ SiglipTextModel: () => (/* binding */ SiglipTextModel),
6422
6431
  /* harmony export */ SiglipVisionModel: () => (/* binding */ SiglipVisionModel),
6423
6432
  /* harmony export */ SmolVLMForConditionalGeneration: () => (/* binding */ SmolVLMForConditionalGeneration),
6433
+ /* harmony export */ SnacDecoderModel: () => (/* binding */ SnacDecoderModel),
6434
+ /* harmony export */ SnacEncoderModel: () => (/* binding */ SnacEncoderModel),
6435
+ /* harmony export */ SnacModel: () => (/* binding */ SnacModel),
6436
+ /* harmony export */ SnacPreTrainedModel: () => (/* binding */ SnacPreTrainedModel),
6424
6437
  /* harmony export */ SpeechT5ForSpeechToText: () => (/* binding */ SpeechT5ForSpeechToText),
6425
6438
  /* harmony export */ SpeechT5ForTextToSpeech: () => (/* binding */ SpeechT5ForTextToSpeech),
6426
6439
  /* harmony export */ SpeechT5HifiGan: () => (/* binding */ SpeechT5HifiGan),
@@ -7079,8 +7092,8 @@ async function decoderForward(self, model_inputs, is_encoder_decoder = false) {
7079
7092
  new_model_inputs.use_cache_branch = boolTensor(!!past_key_values);
7080
7093
  }
7081
7094
  if (session.inputNames.includes('position_ids') && new_model_inputs.attention_mask && !new_model_inputs.position_ids) {
7082
- // NOTE: Handle a special case for paligemma models, where positions are 1-indexed
7083
- const start_index = self.config.model_type === 'paligemma' ? 1 : 0;
7095
+ // NOTE: Handle a special case for paligemma/gemma3 models, where positions are 1-indexed
7096
+ const start_index = ['paligemma', 'gemma3_text', 'gemma3'].includes(self.config.model_type) ? 1 : 0;
7084
7097
  new_model_inputs.position_ids = createPositionIds(new_model_inputs, past_key_values, start_index);
7085
7098
  }
7086
7099
 
@@ -11005,6 +11018,23 @@ class Gemma2Model extends Gemma2PreTrainedModel { }
11005
11018
  class Gemma2ForCausalLM extends Gemma2PreTrainedModel { }
11006
11019
  //////////////////////////////////////////////////
11007
11020
 
11021
+
11022
+ //////////////////////////////////////////////////
11023
+ // Gemma3 models
11024
+
11025
+ /**
11026
+ * The bare Gemma3 Model outputting raw hidden-states without any specific head on top.
11027
+ */
11028
+ class Gemma3PreTrainedModel extends PreTrainedModel { }
11029
+ /**
11030
+ * The bare Gemma3 Model outputting raw hidden-states without any specific head on top.
11031
+ */
11032
+ class Gemma3Model extends Gemma3PreTrainedModel { }
11033
+
11034
+ class Gemma3ForCausalLM extends Gemma3PreTrainedModel { }
11035
+ //////////////////////////////////////////////////
11036
+
11037
+
11008
11038
  //////////////////////////////////////////////////
11009
11039
  class OpenELMPreTrainedModel extends PreTrainedModel { }
11010
11040
  class OpenELMModel extends OpenELMPreTrainedModel { }
@@ -11857,6 +11887,16 @@ class DepthProPreTrainedModel extends PreTrainedModel { }
11857
11887
  class DepthProForDepthEstimation extends DepthProPreTrainedModel { }
11858
11888
  //////////////////////////////////////////////////
11859
11889
 
11890
+ //////////////////////////////////////////////////
11891
+ class Metric3DPreTrainedModel extends PreTrainedModel { }
11892
+ class Metric3DForDepthEstimation extends Metric3DPreTrainedModel { }
11893
+ //////////////////////////////////////////////////
11894
+
11895
+ //////////////////////////////////////////////////
11896
+ class Metric3Dv2PreTrainedModel extends PreTrainedModel { }
11897
+ class Metric3Dv2ForDepthEstimation extends Metric3Dv2PreTrainedModel { }
11898
+ //////////////////////////////////////////////////
11899
+
11860
11900
  //////////////////////////////////////////////////
11861
11901
  class MaskFormerPreTrainedModel extends PreTrainedModel { }
11862
11902
  class MaskFormerModel extends MaskFormerPreTrainedModel { }
@@ -13772,6 +13812,60 @@ class DacDecoderModel extends DacPreTrainedModel {
13772
13812
  }
13773
13813
  //////////////////////////////////////////////////
13774
13814
 
13815
+
13816
+ //////////////////////////////////////////////////
13817
+ // Snac models
13818
+ class SnacPreTrainedModel extends PreTrainedModel {
13819
+ main_input_name = 'input_values';
13820
+ forward_params = ['input_values'];
13821
+ }
13822
+
13823
+ /**
13824
+ * The SNAC (Multi-Scale Neural Audio Codec) model.
13825
+ */
13826
+ class SnacModel extends SnacPreTrainedModel {
13827
+ /**
13828
+ * Encodes the input audio waveform into discrete codes.
13829
+ * @param {Object} inputs Model inputs
13830
+ * @param {Tensor} [inputs.input_values] Float values of the input audio waveform, of shape `(batch_size, channels, sequence_length)`).
13831
+ * @returns {Promise<Record<string, Tensor>>} The output tensors of shape `(batch_size, num_codebooks, sequence_length)`.
13832
+ */
13833
+ async encode(inputs) {
13834
+ return await sessionRun(this.sessions['encoder_model'], inputs);
13835
+ }
13836
+
13837
+ /**
13838
+ * Decodes the given frames into an output audio waveform.
13839
+ * @param {Record<string, Tensor>} inputs The encoded audio codes.
13840
+ * @returns {Promise<{audio_values: Tensor}>} The output tensor of shape `(batch_size, num_channels, sequence_length)`.
13841
+ */
13842
+ async decode(inputs) {
13843
+ return await sessionRun(this.sessions['decoder_model'], inputs);
13844
+ }
13845
+ }
13846
+
13847
+ class SnacEncoderModel extends SnacPreTrainedModel {
13848
+ /** @type {typeof PreTrainedModel.from_pretrained} */
13849
+ static async from_pretrained(pretrained_model_name_or_path, options = {}) {
13850
+ return super.from_pretrained(pretrained_model_name_or_path, {
13851
+ ...options,
13852
+ // Update default model file name if not provided
13853
+ model_file_name: options.model_file_name ?? 'encoder_model',
13854
+ });
13855
+ }
13856
+ }
13857
+ class SnacDecoderModel extends SnacPreTrainedModel {
13858
+ /** @type {typeof PreTrainedModel.from_pretrained} */
13859
+ static async from_pretrained(pretrained_model_name_or_path, options = {}) {
13860
+ return super.from_pretrained(pretrained_model_name_or_path, {
13861
+ ...options,
13862
+ // Update default model file name if not provided
13863
+ model_file_name: options.model_file_name ?? 'decoder_model',
13864
+ });
13865
+ }
13866
+ }
13867
+ //////////////////////////////////////////////////
13868
+
13775
13869
  //////////////////////////////////////////////////
13776
13870
  // AutoModels, used to simplify construction of PreTrainedModels
13777
13871
  // (uses config to instantiate correct class)
@@ -13953,6 +14047,7 @@ const MODEL_MAPPING_NAMES_ENCODER_DECODER = new Map([
13953
14047
  const MODEL_MAPPING_NAMES_AUTO_ENCODER = new Map([
13954
14048
  ['mimi', ['MimiModel', MimiModel]],
13955
14049
  ['dac', ['DacModel', DacModel]],
14050
+ ['snac', ['SnacModel', SnacModel]],
13956
14051
  ]);
13957
14052
 
13958
14053
  const MODEL_MAPPING_NAMES_DECODER_ONLY = new Map([
@@ -13973,6 +14068,7 @@ const MODEL_MAPPING_NAMES_DECODER_ONLY = new Map([
13973
14068
  ['cohere', ['CohereModel', CohereModel]],
13974
14069
  ['gemma', ['GemmaModel', GemmaModel]],
13975
14070
  ['gemma2', ['Gemma2Model', Gemma2Model]],
14071
+ ['gemma3_text', ['Gemma3Model', Gemma3Model]],
13976
14072
  ['helium', ['HeliumModel', HeliumModel]],
13977
14073
  ['glm', ['GlmModel', GlmModel]],
13978
14074
  ['openelm', ['OpenELMModel', OpenELMModel]],
@@ -14072,6 +14168,7 @@ const MODEL_FOR_CAUSAL_LM_MAPPING_NAMES = new Map([
14072
14168
  ['cohere', ['CohereForCausalLM', CohereForCausalLM]],
14073
14169
  ['gemma', ['GemmaForCausalLM', GemmaForCausalLM]],
14074
14170
  ['gemma2', ['Gemma2ForCausalLM', Gemma2ForCausalLM]],
14171
+ ['gemma3_text', ['Gemma3ForCausalLM', Gemma3ForCausalLM]],
14075
14172
  ['helium', ['HeliumForCausalLM', HeliumForCausalLM]],
14076
14173
  ['glm', ['GlmForCausalLM', GlmForCausalLM]],
14077
14174
  ['openelm', ['OpenELMForCausalLM', OpenELMForCausalLM]],
@@ -14273,6 +14370,8 @@ const MODEL_FOR_DEPTH_ESTIMATION_MAPPING_NAMES = new Map([
14273
14370
  ['glpn', ['GLPNForDepthEstimation', GLPNForDepthEstimation]],
14274
14371
  ['sapiens', ['SapiensForDepthEstimation', SapiensForDepthEstimation]],
14275
14372
  ['depth_pro', ['DepthProForDepthEstimation', DepthProForDepthEstimation]],
14373
+ ['metric3d', ['Metric3DForDepthEstimation', Metric3DForDepthEstimation]],
14374
+ ['metric3dv2', ['Metric3Dv2ForDepthEstimation', Metric3Dv2ForDepthEstimation]],
14276
14375
  ])
14277
14376
 
14278
14377
  const MODEL_FOR_NORMAL_ESTIMATION_MAPPING_NAMES = new Map([
@@ -14358,6 +14457,8 @@ const CUSTOM_MAPPING = [
14358
14457
  ['DacDecoderModel', DacDecoderModel, MODEL_TYPES.EncoderOnly],
14359
14458
  ['MimiEncoderModel', MimiEncoderModel, MODEL_TYPES.EncoderOnly],
14360
14459
  ['MimiDecoderModel', MimiDecoderModel, MODEL_TYPES.EncoderOnly],
14460
+ ['SnacEncoderModel', SnacEncoderModel, MODEL_TYPES.EncoderOnly],
14461
+ ['SnacDecoderModel', SnacDecoderModel, MODEL_TYPES.EncoderOnly],
14361
14462
  ]
14362
14463
  for (const [name, model, type] of CUSTOM_MAPPING) {
14363
14464
  MODEL_TYPE_MAPPING.set(name, type);
@@ -15656,14 +15757,15 @@ __webpack_require__.r(__webpack_exports__);
15656
15757
  /* harmony export */ ClapFeatureExtractor: () => (/* reexport safe */ _clap_feature_extraction_clap_js__WEBPACK_IMPORTED_MODULE_2__.ClapFeatureExtractor),
15657
15758
  /* harmony export */ DacFeatureExtractor: () => (/* reexport safe */ _dac_feature_extraction_dac_js__WEBPACK_IMPORTED_MODULE_3__.DacFeatureExtractor),
15658
15759
  /* harmony export */ EncodecFeatureExtractor: () => (/* reexport safe */ _encodec_feature_extraction_encodec_js__WEBPACK_IMPORTED_MODULE_1__.EncodecFeatureExtractor),
15659
- /* harmony export */ ImageFeatureExtractor: () => (/* reexport safe */ _base_image_processors_utils_js__WEBPACK_IMPORTED_MODULE_11__.ImageProcessor),
15760
+ /* harmony export */ ImageFeatureExtractor: () => (/* reexport safe */ _base_image_processors_utils_js__WEBPACK_IMPORTED_MODULE_12__.ImageProcessor),
15660
15761
  /* harmony export */ MoonshineFeatureExtractor: () => (/* reexport safe */ _moonshine_feature_extraction_moonshine_js__WEBPACK_IMPORTED_MODULE_4__.MoonshineFeatureExtractor),
15661
15762
  /* harmony export */ PyAnnoteFeatureExtractor: () => (/* reexport safe */ _pyannote_feature_extraction_pyannote_js__WEBPACK_IMPORTED_MODULE_5__.PyAnnoteFeatureExtractor),
15662
15763
  /* harmony export */ SeamlessM4TFeatureExtractor: () => (/* reexport safe */ _seamless_m4t_feature_extraction_seamless_m4t_js__WEBPACK_IMPORTED_MODULE_6__.SeamlessM4TFeatureExtractor),
15663
- /* harmony export */ SpeechT5FeatureExtractor: () => (/* reexport safe */ _speecht5_feature_extraction_speecht5_js__WEBPACK_IMPORTED_MODULE_7__.SpeechT5FeatureExtractor),
15664
- /* harmony export */ Wav2Vec2FeatureExtractor: () => (/* reexport safe */ _wav2vec2_feature_extraction_wav2vec2_js__WEBPACK_IMPORTED_MODULE_8__.Wav2Vec2FeatureExtractor),
15665
- /* harmony export */ WeSpeakerFeatureExtractor: () => (/* reexport safe */ _wespeaker_feature_extraction_wespeaker_js__WEBPACK_IMPORTED_MODULE_9__.WeSpeakerFeatureExtractor),
15666
- /* harmony export */ WhisperFeatureExtractor: () => (/* reexport safe */ _whisper_feature_extraction_whisper_js__WEBPACK_IMPORTED_MODULE_10__.WhisperFeatureExtractor)
15764
+ /* harmony export */ SnacFeatureExtractor: () => (/* reexport safe */ _snac_feature_extraction_snac_js__WEBPACK_IMPORTED_MODULE_7__.SnacFeatureExtractor),
15765
+ /* harmony export */ SpeechT5FeatureExtractor: () => (/* reexport safe */ _speecht5_feature_extraction_speecht5_js__WEBPACK_IMPORTED_MODULE_8__.SpeechT5FeatureExtractor),
15766
+ /* harmony export */ Wav2Vec2FeatureExtractor: () => (/* reexport safe */ _wav2vec2_feature_extraction_wav2vec2_js__WEBPACK_IMPORTED_MODULE_9__.Wav2Vec2FeatureExtractor),
15767
+ /* harmony export */ WeSpeakerFeatureExtractor: () => (/* reexport safe */ _wespeaker_feature_extraction_wespeaker_js__WEBPACK_IMPORTED_MODULE_10__.WeSpeakerFeatureExtractor),
15768
+ /* harmony export */ WhisperFeatureExtractor: () => (/* reexport safe */ _whisper_feature_extraction_whisper_js__WEBPACK_IMPORTED_MODULE_11__.WhisperFeatureExtractor)
15667
15769
  /* harmony export */ });
15668
15770
  /* harmony import */ var _audio_spectrogram_transformer_feature_extraction_audio_spectrogram_transformer_js__WEBPACK_IMPORTED_MODULE_0__ = __webpack_require__(/*! ./audio_spectrogram_transformer/feature_extraction_audio_spectrogram_transformer.js */ "./src/models/audio_spectrogram_transformer/feature_extraction_audio_spectrogram_transformer.js");
15669
15771
  /* harmony import */ var _encodec_feature_extraction_encodec_js__WEBPACK_IMPORTED_MODULE_1__ = __webpack_require__(/*! ./encodec/feature_extraction_encodec.js */ "./src/models/encodec/feature_extraction_encodec.js");
@@ -15672,11 +15774,13 @@ __webpack_require__.r(__webpack_exports__);
15672
15774
  /* harmony import */ var _moonshine_feature_extraction_moonshine_js__WEBPACK_IMPORTED_MODULE_4__ = __webpack_require__(/*! ./moonshine/feature_extraction_moonshine.js */ "./src/models/moonshine/feature_extraction_moonshine.js");
15673
15775
  /* harmony import */ var _pyannote_feature_extraction_pyannote_js__WEBPACK_IMPORTED_MODULE_5__ = __webpack_require__(/*! ./pyannote/feature_extraction_pyannote.js */ "./src/models/pyannote/feature_extraction_pyannote.js");
15674
15776
  /* harmony import */ var _seamless_m4t_feature_extraction_seamless_m4t_js__WEBPACK_IMPORTED_MODULE_6__ = __webpack_require__(/*! ./seamless_m4t/feature_extraction_seamless_m4t.js */ "./src/models/seamless_m4t/feature_extraction_seamless_m4t.js");
15675
- /* harmony import */ var _speecht5_feature_extraction_speecht5_js__WEBPACK_IMPORTED_MODULE_7__ = __webpack_require__(/*! ./speecht5/feature_extraction_speecht5.js */ "./src/models/speecht5/feature_extraction_speecht5.js");
15676
- /* harmony import */ var _wav2vec2_feature_extraction_wav2vec2_js__WEBPACK_IMPORTED_MODULE_8__ = __webpack_require__(/*! ./wav2vec2/feature_extraction_wav2vec2.js */ "./src/models/wav2vec2/feature_extraction_wav2vec2.js");
15677
- /* harmony import */ var _wespeaker_feature_extraction_wespeaker_js__WEBPACK_IMPORTED_MODULE_9__ = __webpack_require__(/*! ./wespeaker/feature_extraction_wespeaker.js */ "./src/models/wespeaker/feature_extraction_wespeaker.js");
15678
- /* harmony import */ var _whisper_feature_extraction_whisper_js__WEBPACK_IMPORTED_MODULE_10__ = __webpack_require__(/*! ./whisper/feature_extraction_whisper.js */ "./src/models/whisper/feature_extraction_whisper.js");
15679
- /* harmony import */ var _base_image_processors_utils_js__WEBPACK_IMPORTED_MODULE_11__ = __webpack_require__(/*! ../base/image_processors_utils.js */ "./src/base/image_processors_utils.js");
15777
+ /* harmony import */ var _snac_feature_extraction_snac_js__WEBPACK_IMPORTED_MODULE_7__ = __webpack_require__(/*! ./snac/feature_extraction_snac.js */ "./src/models/snac/feature_extraction_snac.js");
15778
+ /* harmony import */ var _speecht5_feature_extraction_speecht5_js__WEBPACK_IMPORTED_MODULE_8__ = __webpack_require__(/*! ./speecht5/feature_extraction_speecht5.js */ "./src/models/speecht5/feature_extraction_speecht5.js");
15779
+ /* harmony import */ var _wav2vec2_feature_extraction_wav2vec2_js__WEBPACK_IMPORTED_MODULE_9__ = __webpack_require__(/*! ./wav2vec2/feature_extraction_wav2vec2.js */ "./src/models/wav2vec2/feature_extraction_wav2vec2.js");
15780
+ /* harmony import */ var _wespeaker_feature_extraction_wespeaker_js__WEBPACK_IMPORTED_MODULE_10__ = __webpack_require__(/*! ./wespeaker/feature_extraction_wespeaker.js */ "./src/models/wespeaker/feature_extraction_wespeaker.js");
15781
+ /* harmony import */ var _whisper_feature_extraction_whisper_js__WEBPACK_IMPORTED_MODULE_11__ = __webpack_require__(/*! ./whisper/feature_extraction_whisper.js */ "./src/models/whisper/feature_extraction_whisper.js");
15782
+ /* harmony import */ var _base_image_processors_utils_js__WEBPACK_IMPORTED_MODULE_12__ = __webpack_require__(/*! ../base/image_processors_utils.js */ "./src/base/image_processors_utils.js");
15783
+
15680
15784
 
15681
15785
 
15682
15786
 
@@ -18666,6 +18770,24 @@ __webpack_require__.r(__webpack_exports__);
18666
18770
 
18667
18771
 
18668
18772
 
18773
+ /***/ }),
18774
+
18775
+ /***/ "./src/models/snac/feature_extraction_snac.js":
18776
+ /*!****************************************************!*\
18777
+ !*** ./src/models/snac/feature_extraction_snac.js ***!
18778
+ \****************************************************/
18779
+ /***/ ((__unused_webpack___webpack_module__, __webpack_exports__, __webpack_require__) => {
18780
+
18781
+ __webpack_require__.r(__webpack_exports__);
18782
+ /* harmony export */ __webpack_require__.d(__webpack_exports__, {
18783
+ /* harmony export */ SnacFeatureExtractor: () => (/* binding */ SnacFeatureExtractor)
18784
+ /* harmony export */ });
18785
+ /* harmony import */ var _dac_feature_extraction_dac_js__WEBPACK_IMPORTED_MODULE_0__ = __webpack_require__(/*! ../dac/feature_extraction_dac.js */ "./src/models/dac/feature_extraction_dac.js");
18786
+
18787
+
18788
+ class SnacFeatureExtractor extends _dac_feature_extraction_dac_js__WEBPACK_IMPORTED_MODULE_0__.DacFeatureExtractor { }
18789
+
18790
+
18669
18791
  /***/ }),
18670
18792
 
18671
18793
  /***/ "./src/models/speecht5/feature_extraction_speecht5.js":
@@ -19905,16 +20027,16 @@ __webpack_require__.r(__webpack_exports__);
19905
20027
  /* harmony import */ var _utils_image_js__WEBPACK_IMPORTED_MODULE_9__ = __webpack_require__(/*! ./utils/image.js */ "./src/utils/image.js");
19906
20028
  /**
19907
20029
  * @file Pipelines provide a high-level, easy to use, API for running machine learning models.
19908
- *
20030
+ *
19909
20031
  * **Example:** Instantiate pipeline using the `pipeline` function.
19910
20032
  * ```javascript
19911
20033
  * import { pipeline } from '@huggingface/transformers';
19912
- *
20034
+ *
19913
20035
  * const classifier = await pipeline('sentiment-analysis');
19914
20036
  * const output = await classifier('I love transformers!');
19915
20037
  * // [{'label': 'POSITIVE', 'score': 0.999817686}]
19916
20038
  * ```
19917
- *
20039
+ *
19918
20040
  * @module pipelines
19919
20041
  */
19920
20042
 
@@ -19933,7 +20055,7 @@ __webpack_require__.r(__webpack_exports__);
19933
20055
 
19934
20056
 
19935
20057
  /**
19936
- * @typedef {string | RawImage | URL} ImageInput
20058
+ * @typedef {string | RawImage | URL | Blob | HTMLCanvasElement | OffscreenCanvas} ImageInput
19937
20059
  * @typedef {ImageInput|ImageInput[]} ImagePipelineInputs
19938
20060
  */
19939
20061
 
@@ -20007,7 +20129,7 @@ function get_bounding_box(box, asInteger) {
20007
20129
  /**
20008
20130
  * @callback DisposeType Disposes the item.
20009
20131
  * @returns {Promise<void>} A promise that resolves when the item has been disposed.
20010
- *
20132
+ *
20011
20133
  * @typedef {Object} Disposable
20012
20134
  * @property {DisposeType} dispose A promise that resolves when the pipeline has been disposed.
20013
20135
  */
@@ -20044,7 +20166,7 @@ class Pipeline extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_4__.Callable {
20044
20166
  * @property {string} task The task of the pipeline. Useful for specifying subtasks.
20045
20167
  * @property {PreTrainedModel} model The model used by the pipeline.
20046
20168
  * @property {PreTrainedTokenizer} tokenizer The tokenizer used by the pipeline.
20047
- *
20169
+ *
20048
20170
  * @typedef {ModelTokenizerConstructorArgs} TextPipelineConstructorArgs An object used to instantiate a text-based pipeline.
20049
20171
  */
20050
20172
 
@@ -20053,7 +20175,7 @@ class Pipeline extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_4__.Callable {
20053
20175
  * @property {string} task The task of the pipeline. Useful for specifying subtasks.
20054
20176
  * @property {PreTrainedModel} model The model used by the pipeline.
20055
20177
  * @property {Processor} processor The processor used by the pipeline.
20056
- *
20178
+ *
20057
20179
  * @typedef {ModelProcessorConstructorArgs} AudioPipelineConstructorArgs An object used to instantiate an audio-based pipeline.
20058
20180
  * @typedef {ModelProcessorConstructorArgs} ImagePipelineConstructorArgs An object used to instantiate an image-based pipeline.
20059
20181
  */
@@ -20065,7 +20187,7 @@ class Pipeline extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_4__.Callable {
20065
20187
  * @property {PreTrainedModel} model The model used by the pipeline.
20066
20188
  * @property {PreTrainedTokenizer} tokenizer The tokenizer used by the pipeline.
20067
20189
  * @property {Processor} processor The processor used by the pipeline.
20068
- *
20190
+ *
20069
20191
  * @typedef {ModelTokenizerProcessorConstructorArgs} TextAudioPipelineConstructorArgs An object used to instantiate a text- and audio-based pipeline.
20070
20192
  * @typedef {ModelTokenizerProcessorConstructorArgs} TextImagePipelineConstructorArgs An object used to instantiate a text- and image-based pipeline.
20071
20193
  */
@@ -20075,15 +20197,15 @@ class Pipeline extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_4__.Callable {
20075
20197
  * @property {string} label The label predicted.
20076
20198
  * @property {number} score The corresponding probability.
20077
20199
  * @typedef {TextClassificationSingle[]} TextClassificationOutput
20078
- *
20200
+ *
20079
20201
  * @typedef {Object} TextClassificationPipelineOptions Parameters specific to text classification pipelines.
20080
20202
  * @property {number} [top_k=1] The number of top predictions to be returned.
20081
- *
20203
+ *
20082
20204
  * @callback TextClassificationPipelineCallback Classify the text(s) given as inputs.
20083
20205
  * @param {string|string[]} texts The input text(s) to be classified.
20084
20206
  * @param {TextClassificationPipelineOptions} [options] The options to use for text classification.
20085
20207
  * @returns {Promise<TextClassificationOutput|TextClassificationOutput[]>} An array or object containing the predicted labels and scores.
20086
- *
20208
+ *
20087
20209
  * @typedef {TextPipelineConstructorArgs & TextClassificationPipelineCallback & Disposable} TextClassificationPipelineType
20088
20210
  */
20089
20211
 
@@ -20096,7 +20218,7 @@ class Pipeline extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_4__.Callable {
20096
20218
  * const output = await classifier('I love transformers!');
20097
20219
  * // [{ label: 'POSITIVE', score: 0.999788761138916 }]
20098
20220
  * ```
20099
- *
20221
+ *
20100
20222
  * **Example:** Multilingual sentiment-analysis w/ `Xenova/bert-base-multilingual-uncased-sentiment` (and return top 5 classes).
20101
20223
  * ```javascript
20102
20224
  * const classifier = await pipeline('sentiment-analysis', 'Xenova/bert-base-multilingual-uncased-sentiment');
@@ -20109,7 +20231,7 @@ class Pipeline extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_4__.Callable {
20109
20231
  * // { label: '2 stars', score: 0.0009423971059732139 }
20110
20232
  * // ]
20111
20233
  * ```
20112
- *
20234
+ *
20113
20235
  * **Example:** Toxic comment classification w/ `Xenova/toxic-bert` (and return all classes).
20114
20236
  * ```javascript
20115
20237
  * const classifier = await pipeline('text-classification', 'Xenova/toxic-bert');
@@ -20194,21 +20316,21 @@ class TextClassificationPipeline extends (/** @type {new (options: TextPipelineC
20194
20316
  * @property {number} [start] The index of the start of the corresponding entity in the sentence.
20195
20317
  * @property {number} [end] The index of the end of the corresponding entity in the sentence.
20196
20318
  * @typedef {TokenClassificationSingle[]} TokenClassificationOutput
20197
- *
20319
+ *
20198
20320
  * @typedef {Object} TokenClassificationPipelineOptions Parameters specific to token classification pipelines.
20199
20321
  * @property {string[]} [ignore_labels] A list of labels to ignore.
20200
- *
20322
+ *
20201
20323
  * @callback TokenClassificationPipelineCallback Classify each token of the text(s) given as inputs.
20202
20324
  * @param {string|string[]} texts One or several texts (or one list of texts) for token classification.
20203
20325
  * @param {TokenClassificationPipelineOptions} [options] The options to use for token classification.
20204
20326
  * @returns {Promise<TokenClassificationOutput|TokenClassificationOutput[]>} The result.
20205
- *
20327
+ *
20206
20328
  * @typedef {TextPipelineConstructorArgs & TokenClassificationPipelineCallback & Disposable} TokenClassificationPipelineType
20207
20329
  */
20208
20330
 
20209
20331
  /**
20210
20332
  * Named Entity Recognition pipeline using any `ModelForTokenClassification`.
20211
- *
20333
+ *
20212
20334
  * **Example:** Perform named entity recognition with `Xenova/bert-base-NER`.
20213
20335
  * ```javascript
20214
20336
  * const classifier = await pipeline('token-classification', 'Xenova/bert-base-NER');
@@ -20218,7 +20340,7 @@ class TextClassificationPipeline extends (/** @type {new (options: TextPipelineC
20218
20340
  * // { entity: 'B-LOC', score: 0.9994474053382874, index: 9, word: 'London' }
20219
20341
  * // ]
20220
20342
  * ```
20221
- *
20343
+ *
20222
20344
  * **Example:** Perform named entity recognition with `Xenova/bert-base-NER` (and return all labels).
20223
20345
  * ```javascript
20224
20346
  * const classifier = await pipeline('token-classification', 'Xenova/bert-base-NER');
@@ -20314,22 +20436,22 @@ class TokenClassificationPipeline extends (/** @type {new (options: TextPipeline
20314
20436
  * @property {number} [start] The character start index of the answer (in the tokenized version of the input).
20315
20437
  * @property {number} [end] The character end index of the answer (in the tokenized version of the input).
20316
20438
  * @property {string} answer The answer to the question.
20317
- *
20439
+ *
20318
20440
  * @typedef {Object} QuestionAnsweringPipelineOptions Parameters specific to question answering pipelines.
20319
20441
  * @property {number} [top_k=1] The number of top answer predictions to be returned.
20320
- *
20442
+ *
20321
20443
  * @callback QuestionAnsweringPipelineCallback Answer the question(s) given as inputs by using the context(s).
20322
20444
  * @param {string|string[]} question One or several question(s) (must be used in conjunction with the `context` argument).
20323
20445
  * @param {string|string[]} context One or several context(s) associated with the question(s) (must be used in conjunction with the `question` argument).
20324
20446
  * @param {QuestionAnsweringPipelineOptions} [options] The options to use for question answering.
20325
20447
  * @returns {Promise<QuestionAnsweringOutput|QuestionAnsweringOutput[]>} An array or object containing the predicted answers and scores.
20326
- *
20448
+ *
20327
20449
  * @typedef {TextPipelineConstructorArgs & QuestionAnsweringPipelineCallback & Disposable} QuestionAnsweringPipelineType
20328
20450
  */
20329
20451
 
20330
20452
  /**
20331
20453
  * Question Answering pipeline using any `ModelForQuestionAnswering`.
20332
- *
20454
+ *
20333
20455
  * **Example:** Run question answering with `Xenova/distilbert-base-uncased-distilled-squad`.
20334
20456
  * ```javascript
20335
20457
  * const answerer = await pipeline('question-answering', 'Xenova/distilbert-base-uncased-distilled-squad');
@@ -20454,10 +20576,10 @@ class QuestionAnsweringPipeline extends (/** @type {new (options: TextPipelineCo
20454
20576
  * @property {number} token The predicted token id (to replace the masked one).
20455
20577
  * @property {string} token_str The predicted token (to replace the masked one).
20456
20578
  * @typedef {FillMaskSingle[]} FillMaskOutput
20457
- *
20579
+ *
20458
20580
  * @typedef {Object} FillMaskPipelineOptions Parameters specific to fill mask pipelines.
20459
20581
  * @property {number} [top_k=5] When passed, overrides the number of predictions to return.
20460
- *
20582
+ *
20461
20583
  * @callback FillMaskPipelineCallback Fill the masked token in the text(s) given as inputs.
20462
20584
  * @param {string|string[]} texts One or several texts (or one list of prompts) with masked tokens.
20463
20585
  * @param {FillMaskPipelineOptions} [options] The options to use for masked language modelling.
@@ -20465,13 +20587,13 @@ class QuestionAnsweringPipeline extends (/** @type {new (options: TextPipelineCo
20465
20587
  * and the sequence with the predicted token filled in, or an array of such arrays (one for each input text).
20466
20588
  * If only one input text is given, the output will be an array of objects.
20467
20589
  * @throws {Error} When the mask token is not found in the input text.
20468
- *
20590
+ *
20469
20591
  * @typedef {TextPipelineConstructorArgs & FillMaskPipelineCallback & Disposable} FillMaskPipelineType
20470
20592
  */
20471
20593
 
20472
20594
  /**
20473
20595
  * Masked language modeling prediction pipeline using any `ModelWithLMHead`.
20474
- *
20596
+ *
20475
20597
  * **Example:** Perform masked language modelling (a.k.a. "fill-mask") with `Xenova/bert-base-uncased`.
20476
20598
  * ```javascript
20477
20599
  * const unmasker = await pipeline('fill-mask', 'Xenova/bert-base-cased');
@@ -20484,7 +20606,7 @@ class QuestionAnsweringPipeline extends (/** @type {new (options: TextPipelineCo
20484
20606
  * // { token_str: 'life', score: 0.01859794743359089, token: 1297, sequence: 'The goal of life is life.' }
20485
20607
  * // ]
20486
20608
  * ```
20487
- *
20609
+ *
20488
20610
  * **Example:** Perform masked language modelling (a.k.a. "fill-mask") with `Xenova/bert-base-cased` (and return top result).
20489
20611
  * ```javascript
20490
20612
  * const unmasker = await pipeline('fill-mask', 'Xenova/bert-base-cased');
@@ -20561,18 +20683,18 @@ class FillMaskPipeline extends (/** @type {new (options: TextPipelineConstructor
20561
20683
  * @typedef {Object} Text2TextGenerationSingle
20562
20684
  * @property {string} generated_text The generated text.
20563
20685
  * @typedef {Text2TextGenerationSingle[]} Text2TextGenerationOutput
20564
- *
20686
+ *
20565
20687
  * @callback Text2TextGenerationPipelineCallback Generate the output text(s) using text(s) given as inputs.
20566
20688
  * @param {string|string[]} texts Input text for the encoder.
20567
20689
  * @param {Partial<import('./generation/configuration_utils.js').GenerationConfig>} [options] Additional keyword arguments to pass along to the generate method of the model.
20568
20690
  * @returns {Promise<Text2TextGenerationOutput|Text2TextGenerationOutput[]>}
20569
- *
20691
+ *
20570
20692
  * @typedef {TextPipelineConstructorArgs & Text2TextGenerationPipelineCallback & Disposable} Text2TextGenerationPipelineType
20571
20693
  */
20572
20694
 
20573
20695
  /**
20574
20696
  * Text2TextGenerationPipeline class for generating text using a model that performs text-to-text generation tasks.
20575
- *
20697
+ *
20576
20698
  * **Example:** Text-to-text generation w/ `Xenova/LaMini-Flan-T5-783M`.
20577
20699
  * ```javascript
20578
20700
  * const generator = await pipeline('text2text-generation', 'Xenova/LaMini-Flan-T5-783M');
@@ -20648,18 +20770,18 @@ class Text2TextGenerationPipeline extends (/** @type {new (options: TextPipeline
20648
20770
  * @typedef {Object} SummarizationSingle
20649
20771
  * @property {string} summary_text The summary text.
20650
20772
  * @typedef {SummarizationSingle[]} SummarizationOutput
20651
- *
20773
+ *
20652
20774
  * @callback SummarizationPipelineCallback Summarize the text(s) given as inputs.
20653
20775
  * @param {string|string[]} texts One or several articles (or one list of articles) to summarize.
20654
20776
  * @param {import('./generation/configuration_utils.js').GenerationConfig} [options] Additional keyword arguments to pass along to the generate method of the model.
20655
20777
  * @returns {Promise<SummarizationOutput|SummarizationOutput[]>}
20656
- *
20778
+ *
20657
20779
  * @typedef {TextPipelineConstructorArgs & SummarizationPipelineCallback & Disposable} SummarizationPipelineType
20658
20780
  */
20659
20781
 
20660
20782
  /**
20661
20783
  * A pipeline for summarization tasks, inheriting from Text2TextGenerationPipeline.
20662
- *
20784
+ *
20663
20785
  * **Example:** Summarization w/ `Xenova/distilbart-cnn-6-6`.
20664
20786
  * ```javascript
20665
20787
  * const generator = await pipeline('summarization', 'Xenova/distilbart-cnn-6-6');
@@ -20695,23 +20817,23 @@ class SummarizationPipeline extends (/** @type {new (options: TextPipelineConstr
20695
20817
  * @typedef {Object} TranslationSingle
20696
20818
  * @property {string} translation_text The translated text.
20697
20819
  * @typedef {TranslationSingle[]} TranslationOutput
20698
- *
20820
+ *
20699
20821
  * @callback TranslationPipelineCallback Translate the text(s) given as inputs.
20700
20822
  * @param {string|string[]} texts Texts to be translated.
20701
20823
  * @param {import('./generation/configuration_utils.js').GenerationConfig} [options] Additional keyword arguments to pass along to the generate method of the model.
20702
20824
  * @returns {Promise<TranslationOutput|TranslationOutput[]>}
20703
- *
20825
+ *
20704
20826
  * @typedef {TextPipelineConstructorArgs & TranslationPipelineCallback & Disposable} TranslationPipelineType
20705
20827
  */
20706
20828
 
20707
20829
  /**
20708
20830
  * Translates text from one language to another.
20709
- *
20831
+ *
20710
20832
  * **Example:** Multilingual translation w/ `Xenova/nllb-200-distilled-600M`.
20711
- *
20833
+ *
20712
20834
  * See [here](https://github.com/facebookresearch/flores/blob/main/flores200/README.md#languages-in-flores-200)
20713
20835
  * for the full list of languages and their corresponding codes.
20714
- *
20836
+ *
20715
20837
  * ```javascript
20716
20838
  * const translator = await pipeline('translation', 'Xenova/nllb-200-distilled-600M');
20717
20839
  * const output = await translator('जीवन एक चॉकलेट बॉक्स की तरह है।', {
@@ -20720,12 +20842,12 @@ class SummarizationPipeline extends (/** @type {new (options: TextPipelineConstr
20720
20842
  * });
20721
20843
  * // [{ translation_text: 'La vie est comme une boîte à chocolat.' }]
20722
20844
  * ```
20723
- *
20845
+ *
20724
20846
  * **Example:** Multilingual translation w/ `Xenova/m2m100_418M`.
20725
- *
20847
+ *
20726
20848
  * See [here](https://huggingface.co/facebook/m2m100_418M#languages-covered)
20727
20849
  * for the full list of languages and their corresponding codes.
20728
- *
20850
+ *
20729
20851
  * ```javascript
20730
20852
  * const translator = await pipeline('translation', 'Xenova/m2m100_418M');
20731
20853
  * const output = await translator('生活就像一盒巧克力。', {
@@ -20734,12 +20856,12 @@ class SummarizationPipeline extends (/** @type {new (options: TextPipelineConstr
20734
20856
  * });
20735
20857
  * // [{ translation_text: 'Life is like a box of chocolate.' }]
20736
20858
  * ```
20737
- *
20859
+ *
20738
20860
  * **Example:** Multilingual translation w/ `Xenova/mbart-large-50-many-to-many-mmt`.
20739
- *
20861
+ *
20740
20862
  * See [here](https://huggingface.co/facebook/mbart-large-50-many-to-many-mmt#languages-covered)
20741
20863
  * for the full list of languages and their corresponding codes.
20742
- *
20864
+ *
20743
20865
  * ```javascript
20744
20866
  * const translator = await pipeline('translation', 'Xenova/mbart-large-50-many-to-many-mmt');
20745
20867
  * const output = await translator('संयुक्त राष्ट्र के प्रमुख का कहना है कि सीरिया में कोई सैन्य समाधान नहीं है', {
@@ -20768,21 +20890,21 @@ function isChat(x) {
20768
20890
 
20769
20891
  /**
20770
20892
  * @typedef {import('./tokenizers.js').Message[]} Chat
20771
- *
20893
+ *
20772
20894
  * @typedef {Object} TextGenerationSingle
20773
20895
  * @property {string|Chat} generated_text The generated text.
20774
20896
  * @typedef {TextGenerationSingle[]} TextGenerationOutput
20775
- *
20897
+ *
20776
20898
  * @typedef {Object} TextGenerationSpecificParams Parameters specific to text-generation pipelines.
20777
20899
  * @property {boolean} [add_special_tokens] Whether or not to add special tokens when tokenizing the sequences.
20778
20900
  * @property {boolean} [return_full_text=true] If set to `false` only added text is returned, otherwise the full text is returned.
20779
20901
  * @typedef {import('./generation/configuration_utils.js').GenerationConfig & TextGenerationSpecificParams} TextGenerationConfig
20780
- *
20902
+ *
20781
20903
  * @callback TextGenerationPipelineCallback Complete the prompt(s) given as inputs.
20782
20904
  * @param {string|string[]|Chat|Chat[]} texts One or several prompts (or one list of prompts) to complete.
20783
20905
  * @param {Partial<TextGenerationConfig>} [options] Additional keyword arguments to pass along to the generate method of the model.
20784
20906
  * @returns {Promise<TextGenerationOutput|TextGenerationOutput[]>} An array or object containing the generated texts.
20785
- *
20907
+ *
20786
20908
  * @typedef {TextPipelineConstructorArgs & TextGenerationPipelineCallback & Disposable} TextGenerationPipelineType
20787
20909
  */
20788
20910
 
@@ -20790,7 +20912,7 @@ function isChat(x) {
20790
20912
  * Language generation pipeline using any `ModelWithLMHead` or `ModelForCausalLM`.
20791
20913
  * This pipeline predicts the words that will follow a specified text prompt.
20792
20914
  * NOTE: For the full list of generation parameters, see [`GenerationConfig`](./utils/generation#module_utils/generation.GenerationConfig).
20793
- *
20915
+ *
20794
20916
  * **Example:** Text generation with `Xenova/distilgpt2` (default settings).
20795
20917
  * ```javascript
20796
20918
  * const generator = await pipeline('text-generation', 'Xenova/distilgpt2');
@@ -20798,7 +20920,7 @@ function isChat(x) {
20798
20920
  * const output = await generator(text);
20799
20921
  * // [{ generated_text: "I enjoy walking with my cute dog, and I love to play with the other dogs." }]
20800
20922
  * ```
20801
- *
20923
+ *
20802
20924
  * **Example:** Text generation with `Xenova/distilgpt2` (custom settings).
20803
20925
  * ```javascript
20804
20926
  * const generator = await pipeline('text-generation', 'Xenova/distilgpt2');
@@ -20817,7 +20939,7 @@ function isChat(x) {
20817
20939
  * // "generated_text": "Once upon a time, there was an abundance of information about the most important and influential"
20818
20940
  * // }]
20819
20941
  * ```
20820
- *
20942
+ *
20821
20943
  * **Example:** Run code generation with `Xenova/codegen-350M-mono`.
20822
20944
  * ```javascript
20823
20945
  * const generator = await pipeline('text-generation', 'Xenova/codegen-350M-mono');
@@ -20936,7 +21058,7 @@ class TextGenerationPipeline extends (/** @type {new (options: TextPipelineConst
20936
21058
  * @property {string} sequence The sequence for which this is the output.
20937
21059
  * @property {string[]} labels The labels sorted by order of likelihood.
20938
21060
  * @property {number[]} scores The probabilities for each of the labels.
20939
- *
21061
+ *
20940
21062
  * @typedef {Object} ZeroShotClassificationPipelineOptions Parameters specific to zero-shot classification pipelines.
20941
21063
  * @property {string} [hypothesis_template="This example is {}."] The template used to turn each
20942
21064
  * candidate label into an NLI-style hypothesis. The candidate label will replace the {} placeholder.
@@ -20944,14 +21066,14 @@ class TextGenerationPipeline extends (/** @type {new (options: TextPipelineConst
20944
21066
  * If `false`, the scores are normalized such that the sum of the label likelihoods for each sequence
20945
21067
  * is 1. If `true`, the labels are considered independent and probabilities are normalized for each
20946
21068
  * candidate by doing a softmax of the entailment score vs. the contradiction score.
20947
- *
21069
+ *
20948
21070
  * @callback ZeroShotClassificationPipelineCallback Classify the sequence(s) given as inputs.
20949
21071
  * @param {string|string[]} texts The sequence(s) to classify, will be truncated if the model input is too large.
20950
21072
  * @param {string|string[]} candidate_labels The set of possible class labels to classify each sequence into.
20951
21073
  * Can be a single label, a string of comma-separated labels, or a list of labels.
20952
21074
  * @param {ZeroShotClassificationPipelineOptions} [options] The options to use for zero-shot classification.
20953
21075
  * @returns {Promise<ZeroShotClassificationOutput|ZeroShotClassificationOutput[]>} An array or object containing the predicted labels and scores.
20954
- *
21076
+ *
20955
21077
  * @typedef {TextPipelineConstructorArgs & ZeroShotClassificationPipelineCallback & Disposable} ZeroShotClassificationPipelineType
20956
21078
  */
20957
21079
 
@@ -20960,7 +21082,7 @@ class TextGenerationPipeline extends (/** @type {new (options: TextPipelineConst
20960
21082
  * trained on NLI (natural language inference) tasks. Equivalent of `text-classification`
20961
21083
  * pipelines, but these models don't require a hardcoded number of potential classes, they
20962
21084
  * can be chosen at runtime. It usually means it's slower but it is **much** more flexible.
20963
- *
21085
+ *
20964
21086
  * **Example:** Zero shot classification with `Xenova/mobilebert-uncased-mnli`.
20965
21087
  * ```javascript
20966
21088
  * const classifier = await pipeline('zero-shot-classification', 'Xenova/mobilebert-uncased-mnli');
@@ -20973,7 +21095,7 @@ class TextGenerationPipeline extends (/** @type {new (options: TextPipelineConst
20973
21095
  * // scores: [ 0.5562091040482018, 0.1843621307860853, 0.13942646639336376, 0.12000229877234923 ]
20974
21096
  * // }
20975
21097
  * ```
20976
- *
21098
+ *
20977
21099
  * **Example:** Zero shot classification with `Xenova/nli-deberta-v3-xsmall` (multi-label).
20978
21100
  * ```javascript
20979
21101
  * const classifier = await pipeline('zero-shot-classification', 'Xenova/nli-deberta-v3-xsmall');
@@ -21087,20 +21209,20 @@ class ZeroShotClassificationPipeline extends (/** @type {new (options: TextPipel
21087
21209
  * @property {'none'|'mean'|'cls'} [pooling="none"] The pooling method to use.
21088
21210
  * @property {boolean} [normalize=false] Whether or not to normalize the embeddings in the last dimension.
21089
21211
  * @property {boolean} [quantize=false] Whether or not to quantize the embeddings.
21090
- * @property {'binary'|'ubinary'} [precision='binary'] The precision to use for quantization.
21091
- *
21212
+ * @property {'binary'|'ubinary'} [precision='binary'] The precision to use for quantization.
21213
+ *
21092
21214
  * @callback FeatureExtractionPipelineCallback Extract the features of the input(s).
21093
21215
  * @param {string|string[]} texts One or several texts (or one list of texts) to get the features of.
21094
21216
  * @param {FeatureExtractionPipelineOptions} [options] The options to use for feature extraction.
21095
21217
  * @returns {Promise<Tensor>} The features computed by the model.
21096
- *
21218
+ *
21097
21219
  * @typedef {TextPipelineConstructorArgs & FeatureExtractionPipelineCallback & Disposable} FeatureExtractionPipelineType
21098
21220
  */
21099
21221
 
21100
21222
  /**
21101
21223
  * Feature extraction pipeline using no model head. This pipeline extracts the hidden
21102
21224
  * states from the base transformer, which can be used as features in downstream tasks.
21103
- *
21225
+ *
21104
21226
  * **Example:** Run feature extraction with `bert-base-uncased` (without pooling/normalization).
21105
21227
  * ```javascript
21106
21228
  * const extractor = await pipeline('feature-extraction', 'Xenova/bert-base-uncased', { revision: 'default' });
@@ -21111,7 +21233,7 @@ class ZeroShotClassificationPipeline extends (/** @type {new (options: TextPipel
21111
21233
  * // dims: [1, 8, 768]
21112
21234
  * // }
21113
21235
  * ```
21114
- *
21236
+ *
21115
21237
  * **Example:** Run feature extraction with `bert-base-uncased` (with pooling/normalization).
21116
21238
  * ```javascript
21117
21239
  * const extractor = await pipeline('feature-extraction', 'Xenova/bert-base-uncased', { revision: 'default' });
@@ -21122,7 +21244,7 @@ class ZeroShotClassificationPipeline extends (/** @type {new (options: TextPipel
21122
21244
  * // dims: [1, 768]
21123
21245
  * // }
21124
21246
  * ```
21125
- *
21247
+ *
21126
21248
  * **Example:** Calculating embeddings with `sentence-transformers` models.
21127
21249
  * ```javascript
21128
21250
  * const extractor = await pipeline('feature-extraction', 'Xenova/all-MiniLM-L6-v2');
@@ -21203,19 +21325,19 @@ class FeatureExtractionPipeline extends (/** @type {new (options: TextPipelineCo
21203
21325
  /**
21204
21326
  * @typedef {Object} ImageFeatureExtractionPipelineOptions Parameters specific to image feature extraction pipelines.
21205
21327
  * @property {boolean} [pool=null] Whether or not to return the pooled output. If set to `false`, the model will return the raw hidden states.
21206
- *
21328
+ *
21207
21329
  * @callback ImageFeatureExtractionPipelineCallback Extract the features of the input(s).
21208
21330
  * @param {ImagePipelineInputs} images One or several images (or one list of images) to get the features of.
21209
21331
  * @param {ImageFeatureExtractionPipelineOptions} [options] The options to use for image feature extraction.
21210
21332
  * @returns {Promise<Tensor>} The image features computed by the model.
21211
- *
21333
+ *
21212
21334
  * @typedef {ImagePipelineConstructorArgs & ImageFeatureExtractionPipelineCallback & Disposable} ImageFeatureExtractionPipelineType
21213
21335
  */
21214
21336
 
21215
21337
  /**
21216
21338
  * Image feature extraction pipeline using no model head. This pipeline extracts the hidden
21217
21339
  * states from the base transformer, which can be used as features in downstream tasks.
21218
- *
21340
+ *
21219
21341
  * **Example:** Perform image feature extraction with `Xenova/vit-base-patch16-224-in21k`.
21220
21342
  * ```javascript
21221
21343
  * const image_feature_extractor = await pipeline('image-feature-extraction', 'Xenova/vit-base-patch16-224-in21k');
@@ -21228,7 +21350,7 @@ class FeatureExtractionPipeline extends (/** @type {new (options: TextPipelineCo
21228
21350
  * // size: 151296
21229
21351
  * // }
21230
21352
  * ```
21231
- *
21353
+ *
21232
21354
  * **Example:** Compute image embeddings with `Xenova/clip-vit-base-patch32`.
21233
21355
  * ```javascript
21234
21356
  * const image_feature_extractor = await pipeline('image-feature-extraction', 'Xenova/clip-vit-base-patch32');
@@ -21284,12 +21406,12 @@ class ImageFeatureExtractionPipeline extends (/** @type {new (options: ImagePipe
21284
21406
  * @property {string} label The label predicted.
21285
21407
  * @property {number} score The corresponding probability.
21286
21408
  * @typedef {AudioClassificationSingle[]} AudioClassificationOutput
21287
- *
21409
+ *
21288
21410
  * @typedef {Object} AudioClassificationPipelineOptions Parameters specific to audio classification pipelines.
21289
21411
  * @property {number} [top_k=5] The number of top labels that will be returned by the pipeline.
21290
21412
  * If the provided number is `null` or higher than the number of labels available in the model configuration,
21291
21413
  * it will default to the number of labels.
21292
- *
21414
+ *
21293
21415
  * @callback AudioClassificationPipelineCallback Classify the sequence(s) given as inputs.
21294
21416
  * @param {AudioPipelineInputs} audio The input audio file(s) to be classified. The input is either:
21295
21417
  * - `string` or `URL` that is the filename/URL of the audio file, the file will be read at the processor's sampling rate
@@ -21298,14 +21420,14 @@ class ImageFeatureExtractionPipeline extends (/** @type {new (options: ImagePipe
21298
21420
  * - `Float32Array` or `Float64Array` of shape `(n, )`, representing the raw audio at the correct sampling rate (no further check will be done).
21299
21421
  * @param {AudioClassificationPipelineOptions} [options] The options to use for audio classification.
21300
21422
  * @returns {Promise<AudioClassificationOutput|AudioClassificationOutput[]>} An array or object containing the predicted labels and scores.
21301
- *
21423
+ *
21302
21424
  * @typedef {AudioPipelineConstructorArgs & AudioClassificationPipelineCallback & Disposable} AudioClassificationPipelineType
21303
21425
  */
21304
21426
 
21305
21427
  /**
21306
21428
  * Audio classification pipeline using any `AutoModelForAudioClassification`.
21307
21429
  * This pipeline predicts the class of a raw waveform or an audio file.
21308
- *
21430
+ *
21309
21431
  * **Example:** Perform audio classification with `Xenova/wav2vec2-large-xlsr-53-gender-recognition-librispeech`.
21310
21432
  * ```javascript
21311
21433
  * const classifier = await pipeline('audio-classification', 'Xenova/wav2vec2-large-xlsr-53-gender-recognition-librispeech');
@@ -21316,7 +21438,7 @@ class ImageFeatureExtractionPipeline extends (/** @type {new (options: ImagePipe
21316
21438
  * // { label: 'female', score: 0.001845747814513743 }
21317
21439
  * // ]
21318
21440
  * ```
21319
- *
21441
+ *
21320
21442
  * **Example:** Perform audio classification with `Xenova/ast-finetuned-audioset-10-10-0.4593` and return top 4 results.
21321
21443
  * ```javascript
21322
21444
  * const classifier = await pipeline('audio-classification', 'Xenova/ast-finetuned-audioset-10-10-0.4593');
@@ -21381,12 +21503,12 @@ class AudioClassificationPipeline extends (/** @type {new (options: AudioPipelin
21381
21503
  * @typedef {Object} ZeroShotAudioClassificationOutput
21382
21504
  * @property {string} label The label identified by the model. It is one of the suggested `candidate_label`.
21383
21505
  * @property {number} score The score attributed by the model for that label (between 0 and 1).
21384
- *
21506
+ *
21385
21507
  * @typedef {Object} ZeroShotAudioClassificationPipelineOptions Parameters specific to zero-shot audio classification pipelines.
21386
21508
  * @property {string} [hypothesis_template="This is a sound of {}."] The sentence used in conjunction with `candidate_labels`
21387
21509
  * to attempt the audio classification by replacing the placeholder with the candidate_labels.
21388
21510
  * Then likelihood is estimated by using `logits_per_audio`.
21389
- *
21511
+ *
21390
21512
  * @callback ZeroShotAudioClassificationPipelineCallback Classify the sequence(s) given as inputs.
21391
21513
  * @param {AudioPipelineInputs} audio The input audio file(s) to be classified. The input is either:
21392
21514
  * - `string` or `URL` that is the filename/URL of the audio file, the file will be read at the processor's sampling rate
@@ -21396,14 +21518,14 @@ class AudioClassificationPipeline extends (/** @type {new (options: AudioPipelin
21396
21518
  * @param {string[]} candidate_labels The candidate labels for this audio.
21397
21519
  * @param {ZeroShotAudioClassificationPipelineOptions} [options] The options to use for zero-shot audio classification.
21398
21520
  * @returns {Promise<ZeroShotAudioClassificationOutput[]|ZeroShotAudioClassificationOutput[][]>} An array of objects containing the predicted labels and scores.
21399
- *
21521
+ *
21400
21522
  * @typedef {TextAudioPipelineConstructorArgs & ZeroShotAudioClassificationPipelineCallback & Disposable} ZeroShotAudioClassificationPipelineType
21401
21523
  */
21402
21524
 
21403
21525
  /**
21404
21526
  * Zero shot audio classification pipeline using `ClapModel`. This pipeline predicts the class of an audio when you
21405
21527
  * provide an audio and a set of `candidate_labels`.
21406
- *
21528
+ *
21407
21529
  * **Example**: Perform zero-shot audio classification with `Xenova/clap-htsat-unfused`.
21408
21530
  * ```javascript
21409
21531
  * const classifier = await pipeline('zero-shot-audio-classification', 'Xenova/clap-htsat-unfused');
@@ -21436,7 +21558,7 @@ class ZeroShotAudioClassificationPipeline extends (/** @type {new (options: Text
21436
21558
  audio = [/** @type {AudioInput} */ (audio)];
21437
21559
  }
21438
21560
 
21439
- // Insert label into hypothesis template
21561
+ // Insert label into hypothesis template
21440
21562
  const texts = candidate_labels.map(
21441
21563
  x => hypothesis_template.replace('{}', x)
21442
21564
  );
@@ -21480,7 +21602,7 @@ class ZeroShotAudioClassificationPipeline extends (/** @type {new (options: Text
21480
21602
  * @property {string} text The recognized text.
21481
21603
  * @property {Chunk[]} [chunks] When using `return_timestamps`, the `chunks` will become a list
21482
21604
  * containing all the various text chunks identified by the model.
21483
- *
21605
+ *
21484
21606
  * @typedef {Object} AutomaticSpeechRecognitionSpecificParams Parameters specific to automatic-speech-recognition pipelines.
21485
21607
  * @property {boolean|'word'} [return_timestamps] Whether to return timestamps or not. Default is `false`.
21486
21608
  * @property {number} [chunk_length_s] The length of audio chunks to process in seconds. Default is 0 (no chunking).
@@ -21490,7 +21612,7 @@ class ZeroShotAudioClassificationPipeline extends (/** @type {new (options: Text
21490
21612
  * @property {string} [task] The task to perform. Default is `null`, meaning it should be auto-detected.
21491
21613
  * @property {number} [num_frames] The number of frames in the input audio.
21492
21614
  * @typedef {import('./generation/configuration_utils.js').GenerationConfig & AutomaticSpeechRecognitionSpecificParams} AutomaticSpeechRecognitionConfig
21493
- *
21615
+ *
21494
21616
  * @callback AutomaticSpeechRecognitionPipelineCallback Transcribe the audio sequence(s) given as inputs to text.
21495
21617
  * @param {AudioPipelineInputs} audio The input audio file(s) to be transcribed. The input is either:
21496
21618
  * - `string` or `URL` that is the filename/URL of the audio file, the file will be read at the processor's sampling rate
@@ -21499,7 +21621,7 @@ class ZeroShotAudioClassificationPipeline extends (/** @type {new (options: Text
21499
21621
  * - `Float32Array` or `Float64Array` of shape `(n, )`, representing the raw audio at the correct sampling rate (no further check will be done).
21500
21622
  * @param {Partial<AutomaticSpeechRecognitionConfig>} [options] Additional keyword arguments to pass along to the generate method of the model.
21501
21623
  * @returns {Promise<AutomaticSpeechRecognitionOutput|AutomaticSpeechRecognitionOutput[]>} An object containing the transcription text and optionally timestamps if `return_timestamps` is `true`.
21502
- *
21624
+ *
21503
21625
  * @typedef {TextAudioPipelineConstructorArgs & AutomaticSpeechRecognitionPipelineCallback & Disposable} AutomaticSpeechRecognitionPipelineType
21504
21626
  */
21505
21627
 
@@ -21513,7 +21635,7 @@ class ZeroShotAudioClassificationPipeline extends (/** @type {new (options: Text
21513
21635
  * const output = await transcriber(url);
21514
21636
  * // { text: " And so my fellow Americans ask not what your country can do for you, ask what you can do for your country." }
21515
21637
  * ```
21516
- *
21638
+ *
21517
21639
  * **Example:** Transcribe English w/ timestamps.
21518
21640
  * ```javascript
21519
21641
  * const transcriber = await pipeline('automatic-speech-recognition', 'Xenova/whisper-tiny.en');
@@ -21527,7 +21649,7 @@ class ZeroShotAudioClassificationPipeline extends (/** @type {new (options: Text
21527
21649
  * // ]
21528
21650
  * // }
21529
21651
  * ```
21530
- *
21652
+ *
21531
21653
  * **Example:** Transcribe English w/ word-level timestamps.
21532
21654
  * ```javascript
21533
21655
  * const transcriber = await pipeline('automatic-speech-recognition', 'Xenova/whisper-tiny.en');
@@ -21546,7 +21668,7 @@ class ZeroShotAudioClassificationPipeline extends (/** @type {new (options: Text
21546
21668
  * // ]
21547
21669
  * // }
21548
21670
  * ```
21549
- *
21671
+ *
21550
21672
  * **Example:** Transcribe French.
21551
21673
  * ```javascript
21552
21674
  * const transcriber = await pipeline('automatic-speech-recognition', 'Xenova/whisper-small');
@@ -21554,7 +21676,7 @@ class ZeroShotAudioClassificationPipeline extends (/** @type {new (options: Text
21554
21676
  * const output = await transcriber(url, { language: 'french', task: 'transcribe' });
21555
21677
  * // { text: " J'adore, j'aime, je n'aime pas, je déteste." }
21556
21678
  * ```
21557
- *
21679
+ *
21558
21680
  * **Example:** Translate French to English.
21559
21681
  * ```javascript
21560
21682
  * const transcriber = await pipeline('automatic-speech-recognition', 'Xenova/whisper-small');
@@ -21562,7 +21684,7 @@ class ZeroShotAudioClassificationPipeline extends (/** @type {new (options: Text
21562
21684
  * const output = await transcriber(url, { language: 'french', task: 'translate' });
21563
21685
  * // { text: " I love, I like, I don't like, I hate." }
21564
21686
  * ```
21565
- *
21687
+ *
21566
21688
  * **Example:** Transcribe/translate audio longer than 30 seconds.
21567
21689
  * ```javascript
21568
21690
  * const transcriber = await pipeline('automatic-speech-recognition', 'Xenova/whisper-tiny.en');
@@ -21785,18 +21907,18 @@ class AutomaticSpeechRecognitionPipeline extends (/** @type {new (options: TextA
21785
21907
  * @typedef {Object} ImageToTextSingle
21786
21908
  * @property {string} generated_text The generated text.
21787
21909
  * @typedef {ImageToTextSingle[]} ImageToTextOutput
21788
- *
21910
+ *
21789
21911
  * @callback ImageToTextPipelineCallback Assign labels to the image(s) passed as inputs.
21790
21912
  * @param {ImagePipelineInputs} texts The images to be captioned.
21791
21913
  * @param {Partial<import('./generation/configuration_utils.js').GenerationConfig>} [options] Additional keyword arguments to pass along to the generate method of the model.
21792
21914
  * @returns {Promise<ImageToTextOutput|ImageToTextOutput[]>} An object (or array of objects) containing the generated text(s).
21793
- *
21915
+ *
21794
21916
  * @typedef {TextImagePipelineConstructorArgs & ImageToTextPipelineCallback & Disposable} ImageToTextPipelineType
21795
21917
  */
21796
21918
 
21797
21919
  /**
21798
21920
  * Image To Text pipeline using a `AutoModelForVision2Seq`. This pipeline predicts a caption for a given image.
21799
- *
21921
+ *
21800
21922
  * **Example:** Generate a caption for an image w/ `Xenova/vit-gpt2-image-captioning`.
21801
21923
  * ```javascript
21802
21924
  * const captioner = await pipeline('image-to-text', 'Xenova/vit-gpt2-image-captioning');
@@ -21804,7 +21926,7 @@ class AutomaticSpeechRecognitionPipeline extends (/** @type {new (options: TextA
21804
21926
  * const output = await captioner(url);
21805
21927
  * // [{ generated_text: 'a cat laying on a couch with another cat' }]
21806
21928
  * ```
21807
- *
21929
+ *
21808
21930
  * **Example:** Optical Character Recognition (OCR) w/ `Xenova/trocr-small-handwritten`.
21809
21931
  * ```javascript
21810
21932
  * const captioner = await pipeline('image-to-text', 'Xenova/trocr-small-handwritten');
@@ -21850,22 +21972,22 @@ class ImageToTextPipeline extends (/** @type {new (options: TextImagePipelineCon
21850
21972
  * @property {string} label The label identified by the model.
21851
21973
  * @property {number} score The score attributed by the model for that label.
21852
21974
  * @typedef {ImageClassificationSingle[]} ImageClassificationOutput
21853
- *
21975
+ *
21854
21976
  * @typedef {Object} ImageClassificationPipelineOptions Parameters specific to image classification pipelines.
21855
- * @property {number} [top_k=1] The number of top labels that will be returned by the pipeline.
21856
- *
21977
+ * @property {number} [top_k=1] The number of top labels that will be returned by the pipeline.
21978
+ *
21857
21979
  * @callback ImageClassificationPipelineCallback Assign labels to the image(s) passed as inputs.
21858
21980
  * @param {ImagePipelineInputs} images The input images(s) to be classified.
21859
21981
  * @param {ImageClassificationPipelineOptions} [options] The options to use for image classification.
21860
21982
  * @returns {Promise<ImageClassificationOutput|ImageClassificationOutput[]>} An array or object containing the predicted labels and scores.
21861
- *
21983
+ *
21862
21984
  * @typedef {ImagePipelineConstructorArgs & ImageClassificationPipelineCallback & Disposable} ImageClassificationPipelineType
21863
21985
  */
21864
21986
 
21865
21987
  /**
21866
21988
  * Image classification pipeline using any `AutoModelForImageClassification`.
21867
21989
  * This pipeline predicts the class of an image.
21868
- *
21990
+ *
21869
21991
  * **Example:** Classify an image.
21870
21992
  * ```javascript
21871
21993
  * const classifier = await pipeline('image-classification', 'Xenova/vit-base-patch16-224');
@@ -21875,7 +21997,7 @@ class ImageToTextPipeline extends (/** @type {new (options: TextImagePipelineCon
21875
21997
  * // { label: 'tiger, Panthera tigris', score: 0.632695734500885 },
21876
21998
  * // ]
21877
21999
  * ```
21878
- *
22000
+ *
21879
22001
  * **Example:** Classify an image and return top `n` classes.
21880
22002
  * ```javascript
21881
22003
  * const classifier = await pipeline('image-classification', 'Xenova/vit-base-patch16-224');
@@ -21887,7 +22009,7 @@ class ImageToTextPipeline extends (/** @type {new (options: TextImagePipelineCon
21887
22009
  * // { label: 'lion, king of beasts, Panthera leo', score: 0.00045060308184474707 },
21888
22010
  * // ]
21889
22011
  * ```
21890
- *
22012
+ *
21891
22013
  * **Example:** Classify an image and return all classes.
21892
22014
  * ```javascript
21893
22015
  * const classifier = await pipeline('image-classification', 'Xenova/vit-base-patch16-224');
@@ -21954,7 +22076,7 @@ class ImageClassificationPipeline extends (/** @type {new (options: ImagePipelin
21954
22076
  * @property {string|null} label The label of the segment.
21955
22077
  * @property {number|null} score The score of the segment.
21956
22078
  * @property {RawImage} mask The mask of the segment.
21957
- *
22079
+ *
21958
22080
  * @typedef {Object} ImageSegmentationPipelineOptions Parameters specific to image segmentation pipelines.
21959
22081
  * @property {number} [threshold=0.5] Probability threshold to filter out predicted masks.
21960
22082
  * @property {number} [mask_threshold=0.5] Threshold to use when turning the predicted masks into binary values.
@@ -21963,19 +22085,19 @@ class ImageClassificationPipeline extends (/** @type {new (options: ImagePipelin
21963
22085
  * depending on model capabilities. If not set, the pipeline will attempt to resolve (in that order).
21964
22086
  * @property {number[]} [label_ids_to_fuse=null] List of label ids to fuse. If not set, do not fuse any labels.
21965
22087
  * @property {number[][]} [target_sizes=null] List of target sizes for the input images. If not set, use the original image sizes.
21966
- *
22088
+ *
21967
22089
  * @callback ImageSegmentationPipelineCallback Segment the input images.
21968
22090
  * @param {ImagePipelineInputs} images The input images.
21969
22091
  * @param {ImageSegmentationPipelineOptions} [options] The options to use for image segmentation.
21970
22092
  * @returns {Promise<ImageSegmentationPipelineOutput[]>} The annotated segments.
21971
- *
22093
+ *
21972
22094
  * @typedef {ImagePipelineConstructorArgs & ImageSegmentationPipelineCallback & Disposable} ImageSegmentationPipelineType
21973
22095
  */
21974
22096
 
21975
22097
  /**
21976
22098
  * Image segmentation pipeline using any `AutoModelForXXXSegmentation`.
21977
22099
  * This pipeline predicts masks of objects and their classes.
21978
- *
22100
+ *
21979
22101
  * **Example:** Perform image segmentation with `Xenova/detr-resnet-50-panoptic`.
21980
22102
  * ```javascript
21981
22103
  * const segmenter = await pipeline('image-segmentation', 'Xenova/detr-resnet-50-panoptic');
@@ -22059,12 +22181,17 @@ class ImageSegmentationPipeline extends (/** @type {new (options: ImagePipelineC
22059
22181
  /** @type {ImageSegmentationPipelineOutput[]} */
22060
22182
  const annotation = [];
22061
22183
  if (!subtask) {
22184
+ // We define an epsilon to safeguard against numerical/precision issues when detecting
22185
+ // the normalization mode of the output (i.e., sigmoid already applied, or not).
22186
+ // See https://github.com/microsoft/onnxruntime/issues/23943 for more information.
22187
+ const epsilon = 1e-5;
22188
+
22062
22189
  // Perform standard image segmentation
22063
22190
  const result = output[outputNames[0]];
22064
22191
  for (let i = 0; i < imageSizes.length; ++i) {
22065
22192
  const size = imageSizes[i];
22066
22193
  const item = result[i];
22067
- if (item.data.some(x => x < 0 || x > 1)) {
22194
+ if (item.data.some(x => x < -epsilon || x > 1 + epsilon)) {
22068
22195
  item.sigmoid_();
22069
22196
  }
22070
22197
  const mask = await _utils_image_js__WEBPACK_IMPORTED_MODULE_9__.RawImage.fromTensor(item.mul_(255).to('uint8')).resize(size[1], size[0]);
@@ -22133,19 +22260,19 @@ class ImageSegmentationPipeline extends (/** @type {new (options: ImagePipelineC
22133
22260
 
22134
22261
  /**
22135
22262
  * @typedef {Object} BackgroundRemovalPipelineOptions Parameters specific to image segmentation pipelines.
22136
- *
22263
+ *
22137
22264
  * @callback BackgroundRemovalPipelineCallback Segment the input images.
22138
22265
  * @param {ImagePipelineInputs} images The input images.
22139
22266
  * @param {BackgroundRemovalPipelineOptions} [options] The options to use for image segmentation.
22140
22267
  * @returns {Promise<RawImage[]>} The images with the background removed.
22141
- *
22268
+ *
22142
22269
  * @typedef {ImagePipelineConstructorArgs & BackgroundRemovalPipelineCallback & Disposable} BackgroundRemovalPipelineType
22143
22270
  */
22144
22271
 
22145
22272
  /**
22146
22273
  * Background removal pipeline using certain `AutoModelForXXXSegmentation`.
22147
22274
  * This pipeline removes the backgrounds of images.
22148
- *
22275
+ *
22149
22276
  * **Example:** Perform background removal with `Xenova/modnet`.
22150
22277
  * ```javascript
22151
22278
  * const segmenter = await pipeline('background-removal', 'Xenova/modnet');
@@ -22156,7 +22283,7 @@ class ImageSegmentationPipeline extends (/** @type {new (options: ImagePipelineC
22156
22283
  * // ]
22157
22284
  * ```
22158
22285
  */
22159
- class BackgroundRemovalPipeline extends (/** @type {new (options: ImagePipelineConstructorArgs) => ImageSegmentationPipelineType} */ (ImageSegmentationPipeline)) {
22286
+ class BackgroundRemovalPipeline extends (/** @type {new (options: ImagePipelineConstructorArgs) => BackgroundRemovalPipelineType} */ (/** @type {any} */(ImageSegmentationPipeline))) {
22160
22287
  /**
22161
22288
  * Create a new BackgroundRemovalPipeline.
22162
22289
  * @param {ImagePipelineConstructorArgs} options An object used to instantiate the pipeline.
@@ -22191,25 +22318,25 @@ class BackgroundRemovalPipeline extends (/** @type {new (options: ImagePipelineC
22191
22318
  * @typedef {Object} ZeroShotImageClassificationOutput
22192
22319
  * @property {string} label The label identified by the model. It is one of the suggested `candidate_label`.
22193
22320
  * @property {number} score The score attributed by the model for that label (between 0 and 1).
22194
- *
22321
+ *
22195
22322
  * @typedef {Object} ZeroShotImageClassificationPipelineOptions Parameters specific to zero-shot image classification pipelines.
22196
22323
  * @property {string} [hypothesis_template="This is a photo of {}"] The sentence used in conjunction with `candidate_labels`
22197
22324
  * to attempt the image classification by replacing the placeholder with the candidate_labels.
22198
22325
  * Then likelihood is estimated by using `logits_per_image`.
22199
- *
22326
+ *
22200
22327
  * @callback ZeroShotImageClassificationPipelineCallback Assign labels to the image(s) passed as inputs.
22201
22328
  * @param {ImagePipelineInputs} images The input images.
22202
22329
  * @param {string[]} candidate_labels The candidate labels for this image.
22203
22330
  * @param {ZeroShotImageClassificationPipelineOptions} [options] The options to use for zero-shot image classification.
22204
22331
  * @returns {Promise<ZeroShotImageClassificationOutput[]|ZeroShotImageClassificationOutput[][]>} An array of objects containing the predicted labels and scores.
22205
- *
22332
+ *
22206
22333
  * @typedef {TextImagePipelineConstructorArgs & ZeroShotImageClassificationPipelineCallback & Disposable} ZeroShotImageClassificationPipelineType
22207
22334
  */
22208
22335
 
22209
22336
  /**
22210
22337
  * Zero shot image classification pipeline. This pipeline predicts the class of
22211
22338
  * an image when you provide an image and a set of `candidate_labels`.
22212
- *
22339
+ *
22213
22340
  * **Example:** Zero shot image classification w/ `Xenova/clip-vit-base-patch32`.
22214
22341
  * ```javascript
22215
22342
  * const classifier = await pipeline('zero-shot-image-classification', 'Xenova/clip-vit-base-patch32');
@@ -22239,7 +22366,7 @@ class ZeroShotImageClassificationPipeline extends (/** @type {new (options: Text
22239
22366
  const isBatched = Array.isArray(images);
22240
22367
  const preparedImages = await prepareImages(images);
22241
22368
 
22242
- // Insert label into hypothesis template
22369
+ // Insert label into hypothesis template
22243
22370
  const texts = candidate_labels.map(
22244
22371
  x => hypothesis_template.replace('{}', x)
22245
22372
  );
@@ -22286,23 +22413,23 @@ class ZeroShotImageClassificationPipeline extends (/** @type {new (options: Text
22286
22413
  * @property {number} score The score attributed by the model for that label.
22287
22414
  * @property {BoundingBox} box The bounding box of detected object in image's original size, or as a percentage if `percentage` is set to true.
22288
22415
  * @typedef {ObjectDetectionPipelineSingle[]} ObjectDetectionPipelineOutput
22289
- *
22416
+ *
22290
22417
  * @typedef {Object} ObjectDetectionPipelineOptions Parameters specific to object detection pipelines.
22291
22418
  * @property {number} [threshold=0.9] The threshold used to filter boxes by score.
22292
22419
  * @property {boolean} [percentage=false] Whether to return the boxes coordinates in percentage (true) or in pixels (false).
22293
- *
22420
+ *
22294
22421
  * @callback ObjectDetectionPipelineCallback Detect objects (bounding boxes & classes) in the image(s) passed as inputs.
22295
22422
  * @param {ImagePipelineInputs} images The input images.
22296
22423
  * @param {ObjectDetectionPipelineOptions} [options] The options to use for object detection.
22297
- * @returns {Promise<ObjectDetectionPipelineOutput|ObjectDetectionPipelineOutput[]>} A list of objects or a list of list of objects.
22298
- *
22424
+ * @returns {Promise<ObjectDetectionPipelineOutput|ObjectDetectionPipelineOutput[]>} A list of objects or a list of list of objects.
22425
+ *
22299
22426
  * @typedef {ImagePipelineConstructorArgs & ObjectDetectionPipelineCallback & Disposable} ObjectDetectionPipelineType
22300
22427
  */
22301
22428
 
22302
22429
  /**
22303
22430
  * Object detection pipeline using any `AutoModelForObjectDetection`.
22304
22431
  * This pipeline predicts bounding boxes of objects and their classes.
22305
- *
22432
+ *
22306
22433
  * **Example:** Run object-detection with `Xenova/detr-resnet-50`.
22307
22434
  * ```javascript
22308
22435
  * const detector = await pipeline('object-detection', 'Xenova/detr-resnet-50');
@@ -22376,27 +22503,27 @@ class ObjectDetectionPipeline extends (/** @type {new (options: ImagePipelineCon
22376
22503
  * @property {string} label Text query corresponding to the found object.
22377
22504
  * @property {number} score Score corresponding to the object (between 0 and 1).
22378
22505
  * @property {BoundingBox} box Bounding box of the detected object in image's original size, or as a percentage if `percentage` is set to true.
22379
- *
22506
+ *
22380
22507
  * @typedef {Object} ZeroShotObjectDetectionPipelineOptions Parameters specific to zero-shot object detection pipelines.
22381
22508
  * @property {number} [threshold=0.1] The probability necessary to make a prediction.
22382
22509
  * @property {number} [top_k=null] The number of top predictions that will be returned by the pipeline.
22383
22510
  * If the provided number is `null` or higher than the number of predictions available, it will default
22384
22511
  * to the number of predictions.
22385
22512
  * @property {boolean} [percentage=false] Whether to return the boxes coordinates in percentage (true) or in pixels (false).
22386
- *
22513
+ *
22387
22514
  * @callback ZeroShotObjectDetectionPipelineCallback Detect objects (bounding boxes & classes) in the image(s) passed as inputs.
22388
22515
  * @param {ImagePipelineInputs} images The input images.
22389
22516
  * @param {string[]} candidate_labels What the model should recognize in the image.
22390
22517
  * @param {ZeroShotObjectDetectionPipelineOptions} [options] The options to use for zero-shot object detection.
22391
22518
  * @returns {Promise<ZeroShotObjectDetectionOutput[]|ZeroShotObjectDetectionOutput[][]>} An array of objects containing the predicted labels, scores, and bounding boxes.
22392
- *
22519
+ *
22393
22520
  * @typedef {TextImagePipelineConstructorArgs & ZeroShotObjectDetectionPipelineCallback & Disposable} ZeroShotObjectDetectionPipelineType
22394
22521
  */
22395
22522
 
22396
22523
  /**
22397
22524
  * Zero-shot object detection pipeline. This pipeline predicts bounding boxes of
22398
22525
  * objects when you provide an image and a set of `candidate_labels`.
22399
- *
22526
+ *
22400
22527
  * **Example:** Zero-shot object detection w/ `Xenova/owlvit-base-patch32`.
22401
22528
  * ```javascript
22402
22529
  * const detector = await pipeline('zero-shot-object-detection', 'Xenova/owlvit-base-patch32');
@@ -22426,7 +22553,7 @@ class ObjectDetectionPipeline extends (/** @type {new (options: ImagePipelineCon
22426
22553
  * // }
22427
22554
  * // ]
22428
22555
  * ```
22429
- *
22556
+ *
22430
22557
  * **Example:** Zero-shot object detection w/ `Xenova/owlvit-base-patch32` (returning top 4 matches and setting a threshold).
22431
22558
  * ```javascript
22432
22559
  * const detector = await pipeline('zero-shot-object-detection', 'Xenova/owlvit-base-patch32');
@@ -22541,13 +22668,13 @@ class ZeroShotObjectDetectionPipeline extends (/** @type {new (options: TextImag
22541
22668
  * @typedef {Object} DocumentQuestionAnsweringSingle
22542
22669
  * @property {string} answer The generated text.
22543
22670
  * @typedef {DocumentQuestionAnsweringSingle[]} DocumentQuestionAnsweringOutput
22544
- *
22671
+ *
22545
22672
  * @callback DocumentQuestionAnsweringPipelineCallback Answer the question given as input by using the document.
22546
22673
  * @param {ImageInput} image The image of the document to use.
22547
22674
  * @param {string} question A question to ask of the document.
22548
22675
  * @param {Partial<import('./generation/configuration_utils.js').GenerationConfig>} [options] Additional keyword arguments to pass along to the generate method of the model.
22549
22676
  * @returns {Promise<DocumentQuestionAnsweringOutput|DocumentQuestionAnsweringOutput[]>} An object (or array of objects) containing the answer(s).
22550
- *
22677
+ *
22551
22678
  * @typedef {TextImagePipelineConstructorArgs & DocumentQuestionAnsweringPipelineCallback & Disposable} DocumentQuestionAnsweringPipelineType
22552
22679
  */
22553
22680
 
@@ -22555,7 +22682,7 @@ class ZeroShotObjectDetectionPipeline extends (/** @type {new (options: TextImag
22555
22682
  * Document Question Answering pipeline using any `AutoModelForDocumentQuestionAnswering`.
22556
22683
  * The inputs/outputs are similar to the (extractive) question answering pipeline; however,
22557
22684
  * the pipeline takes an image (and optional OCR'd words/boxes) as input instead of text context.
22558
- *
22685
+ *
22559
22686
  * **Example:** Answer questions about a document with `Xenova/donut-base-finetuned-docvqa`.
22560
22687
  * ```javascript
22561
22688
  * const qa_pipeline = await pipeline('document-question-answering', 'Xenova/donut-base-finetuned-docvqa');
@@ -22625,22 +22752,22 @@ class DocumentQuestionAnsweringPipeline extends (/** @type {new (options: TextIm
22625
22752
  * @typedef {Object} TextToAudioOutput
22626
22753
  * @property {Float32Array} audio The generated audio waveform.
22627
22754
  * @property {number} sampling_rate The sampling rate of the generated audio waveform.
22628
- *
22755
+ *
22629
22756
  * @typedef {Object} TextToAudioPipelineOptions Parameters specific to text-to-audio pipelines.
22630
22757
  * @property {Tensor|Float32Array|string|URL} [speaker_embeddings=null] The speaker embeddings (if the model requires it).
22631
- *
22758
+ *
22632
22759
  * @callback TextToAudioPipelineCallback Generates speech/audio from the inputs.
22633
22760
  * @param {string|string[]} texts The text(s) to generate.
22634
22761
  * @param {TextToAudioPipelineOptions} options Parameters passed to the model generation/forward method.
22635
22762
  * @returns {Promise<TextToAudioOutput>} An object containing the generated audio and sampling rate.
22636
- *
22763
+ *
22637
22764
  * @typedef {TextToAudioPipelineConstructorArgs & TextToAudioPipelineCallback & Disposable} TextToAudioPipelineType
22638
22765
  */
22639
22766
 
22640
22767
  /**
22641
22768
  * Text-to-audio generation pipeline using any `AutoModelForTextToWaveform` or `AutoModelForTextToSpectrogram`.
22642
22769
  * This pipeline generates an audio file from an input text and optional other conditional inputs.
22643
- *
22770
+ *
22644
22771
  * **Example:** Generate audio from text with `Xenova/speecht5_tts`.
22645
22772
  * ```javascript
22646
22773
  * const synthesizer = await pipeline('text-to-speech', 'Xenova/speecht5_tts', { quantized: false });
@@ -22651,17 +22778,17 @@ class DocumentQuestionAnsweringPipeline extends (/** @type {new (options: TextIm
22651
22778
  * // sampling_rate: 16000
22652
22779
  * // }
22653
22780
  * ```
22654
- *
22781
+ *
22655
22782
  * You can then save the audio to a .wav file with the `wavefile` package:
22656
22783
  * ```javascript
22657
22784
  * import wavefile from 'wavefile';
22658
22785
  * import fs from 'fs';
22659
- *
22786
+ *
22660
22787
  * const wav = new wavefile.WaveFile();
22661
22788
  * wav.fromScratch(1, out.sampling_rate, '32f', out.audio);
22662
22789
  * fs.writeFileSync('out.wav', wav.toBuffer());
22663
22790
  * ```
22664
- *
22791
+ *
22665
22792
  * **Example:** Multilingual speech generation with `Xenova/mms-tts-fra`. See [here](https://huggingface.co/models?pipeline_tag=text-to-speech&other=vits&sort=trending) for the full list of available languages (1107).
22666
22793
  * ```javascript
22667
22794
  * const synthesizer = await pipeline('text-to-speech', 'Xenova/mms-tts-fra');
@@ -22767,13 +22894,13 @@ class TextToAudioPipeline extends (/** @type {new (options: TextToAudioPipelineC
22767
22894
  * @callback ImageToImagePipelineCallback Transform the image(s) passed as inputs.
22768
22895
  * @param {ImagePipelineInputs} images The images to transform.
22769
22896
  * @returns {Promise<RawImage|RawImage[]>} The transformed image or list of images.
22770
- *
22897
+ *
22771
22898
  * @typedef {ImagePipelineConstructorArgs & ImageToImagePipelineCallback & Disposable} ImageToImagePipelineType
22772
22899
  */
22773
22900
 
22774
22901
  /**
22775
22902
  * Image to Image pipeline using any `AutoModelForImageToImage`. This pipeline generates an image based on a previous image input.
22776
- *
22903
+ *
22777
22904
  * **Example:** Super-resolution w/ `Xenova/swin2SR-classical-sr-x2-64`
22778
22905
  * ```javascript
22779
22906
  * const upscaler = await pipeline('image-to-image', 'Xenova/swin2SR-classical-sr-x2-64');
@@ -22818,17 +22945,17 @@ class ImageToImagePipeline extends (/** @type {new (options: ImagePipelineConstr
22818
22945
  * @typedef {Object} DepthEstimationPipelineOutput
22819
22946
  * @property {Tensor} predicted_depth The raw depth map predicted by the model.
22820
22947
  * @property {RawImage} depth The processed depth map as an image (with the same size as the input image).
22821
- *
22948
+ *
22822
22949
  * @callback DepthEstimationPipelineCallback Predicts the depth for the image(s) passed as inputs.
22823
22950
  * @param {ImagePipelineInputs} images The images to compute depth for.
22824
22951
  * @returns {Promise<DepthEstimationPipelineOutput|DepthEstimationPipelineOutput[]>} An image or a list of images containing result(s).
22825
- *
22952
+ *
22826
22953
  * @typedef {ImagePipelineConstructorArgs & DepthEstimationPipelineCallback & Disposable} DepthEstimationPipelineType
22827
22954
  */
22828
22955
 
22829
22956
  /**
22830
22957
  * Depth estimation pipeline using any `AutoModelForDepthEstimation`. This pipeline predicts the depth of an image.
22831
- *
22958
+ *
22832
22959
  * **Example:** Depth estimation w/ `Xenova/dpt-hybrid-midas`
22833
22960
  * ```javascript
22834
22961
  * const depth_estimator = await pipeline('depth-estimation', 'Xenova/dpt-hybrid-midas');
@@ -23213,7 +23340,7 @@ const TASK_ALIASES = Object.freeze({
23213
23340
 
23214
23341
  /**
23215
23342
  * Utility factory method to build a `Pipeline` object.
23216
- *
23343
+ *
23217
23344
  * @template {PipelineType} T The type of pipeline to return.
23218
23345
  * @param {T} task The task defining which pipeline will be returned. Currently accepted tasks are:
23219
23346
  * - `"audio-classification"`: will return a `AudioClassificationPipeline`.
@@ -30479,7 +30606,7 @@ class RawImage {
30479
30606
 
30480
30607
  /**
30481
30608
  * Helper method for reading an image from a variety of input types.
30482
- * @param {RawImage|string|URL} input
30609
+ * @param {RawImage|string|URL|Blob|HTMLCanvasElement|OffscreenCanvas} input
30483
30610
  * @returns The image object.
30484
30611
  *
30485
30612
  * **Example:** Read image from a URL.
@@ -30498,6 +30625,14 @@ class RawImage {
30498
30625
  return input;
30499
30626
  } else if (typeof input === 'string' || input instanceof URL) {
30500
30627
  return await this.fromURL(input);
30628
+ } else if (input instanceof Blob) {
30629
+ return await this.fromBlob(input);
30630
+ } else if (
30631
+ (typeof HTMLCanvasElement !== "undefined" && input instanceof HTMLCanvasElement)
30632
+ ||
30633
+ (typeof OffscreenCanvas !== "undefined" && input instanceof OffscreenCanvas)
30634
+ ) {
30635
+ return this.fromCanvas(input);
30501
30636
  } else {
30502
30637
  throw new Error(`Unsupported input type: ${typeof input}`);
30503
30638
  }
@@ -33507,8 +33642,12 @@ function calc_unsqueeze_dims(dims, dim) {
33507
33642
  * @private
33508
33643
  */
33509
33644
  function safeIndex(index, size, dimension = null, boundsCheck = true) {
33510
- if (boundsCheck && (index < -size || index >= size)) {
33511
- throw new Error(`IndexError: index ${index} is out of bounds for dimension${dimension === null ? '' : ' ' + dimension} with size ${size}`);
33645
+ if (index < -size || index >= size) {
33646
+ if (boundsCheck) {
33647
+ throw new Error(`IndexError: index ${index} is out of bounds for dimension${dimension === null ? '' : ' ' + dimension} with size ${size}`);
33648
+ } else {
33649
+ return index < -size ? 0 : size;
33650
+ }
33512
33651
  }
33513
33652
 
33514
33653
  if (index < 0) {
@@ -34363,6 +34502,9 @@ __webpack_require__.r(__webpack_exports__);
34363
34502
  /* harmony export */ Gemma2ForCausalLM: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Gemma2ForCausalLM),
34364
34503
  /* harmony export */ Gemma2Model: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Gemma2Model),
34365
34504
  /* harmony export */ Gemma2PreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Gemma2PreTrainedModel),
34505
+ /* harmony export */ Gemma3ForCausalLM: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Gemma3ForCausalLM),
34506
+ /* harmony export */ Gemma3Model: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Gemma3Model),
34507
+ /* harmony export */ Gemma3PreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Gemma3PreTrainedModel),
34366
34508
  /* harmony export */ GemmaForCausalLM: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.GemmaForCausalLM),
34367
34509
  /* harmony export */ GemmaModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.GemmaModel),
34368
34510
  /* harmony export */ GemmaPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.GemmaPreTrainedModel),
@@ -34464,6 +34606,10 @@ __webpack_require__.r(__webpack_exports__);
34464
34606
  /* harmony export */ MaskFormerPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.MaskFormerPreTrainedModel),
34465
34607
  /* harmony export */ MaskedLMOutput: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.MaskedLMOutput),
34466
34608
  /* harmony export */ MaxLengthCriteria: () => (/* reexport safe */ _generation_stopping_criteria_js__WEBPACK_IMPORTED_MODULE_20__.MaxLengthCriteria),
34609
+ /* harmony export */ Metric3DForDepthEstimation: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Metric3DForDepthEstimation),
34610
+ /* harmony export */ Metric3DPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Metric3DPreTrainedModel),
34611
+ /* harmony export */ Metric3Dv2ForDepthEstimation: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Metric3Dv2ForDepthEstimation),
34612
+ /* harmony export */ Metric3Dv2PreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Metric3Dv2PreTrainedModel),
34467
34613
  /* harmony export */ MgpstrForSceneTextRecognition: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.MgpstrForSceneTextRecognition),
34468
34614
  /* harmony export */ MgpstrModelOutput: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.MgpstrModelOutput),
34469
34615
  /* harmony export */ MgpstrPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.MgpstrPreTrainedModel),
@@ -34670,6 +34816,11 @@ __webpack_require__.r(__webpack_exports__);
34670
34816
  /* harmony export */ SmolVLMForConditionalGeneration: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.SmolVLMForConditionalGeneration),
34671
34817
  /* harmony export */ SmolVLMImageProcessor: () => (/* reexport safe */ _models_image_processors_js__WEBPACK_IMPORTED_MODULE_14__.SmolVLMImageProcessor),
34672
34818
  /* harmony export */ SmolVLMProcessor: () => (/* reexport safe */ _models_processors_js__WEBPACK_IMPORTED_MODULE_17__.SmolVLMProcessor),
34819
+ /* harmony export */ SnacDecoderModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.SnacDecoderModel),
34820
+ /* harmony export */ SnacEncoderModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.SnacEncoderModel),
34821
+ /* harmony export */ SnacFeatureExtractor: () => (/* reexport safe */ _models_feature_extractors_js__WEBPACK_IMPORTED_MODULE_11__.SnacFeatureExtractor),
34822
+ /* harmony export */ SnacModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.SnacModel),
34823
+ /* harmony export */ SnacPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.SnacPreTrainedModel),
34673
34824
  /* harmony export */ SpeechT5FeatureExtractor: () => (/* reexport safe */ _models_feature_extractors_js__WEBPACK_IMPORTED_MODULE_11__.SpeechT5FeatureExtractor),
34674
34825
  /* harmony export */ SpeechT5ForSpeechToText: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.SpeechT5ForSpeechToText),
34675
34826
  /* harmony export */ SpeechT5ForTextToSpeech: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.SpeechT5ForTextToSpeech),
@@ -35192,6 +35343,9 @@ var __webpack_exports__GPTNeoXTokenizer = __webpack_exports__.GPTNeoXTokenizer;
35192
35343
  var __webpack_exports__Gemma2ForCausalLM = __webpack_exports__.Gemma2ForCausalLM;
35193
35344
  var __webpack_exports__Gemma2Model = __webpack_exports__.Gemma2Model;
35194
35345
  var __webpack_exports__Gemma2PreTrainedModel = __webpack_exports__.Gemma2PreTrainedModel;
35346
+ var __webpack_exports__Gemma3ForCausalLM = __webpack_exports__.Gemma3ForCausalLM;
35347
+ var __webpack_exports__Gemma3Model = __webpack_exports__.Gemma3Model;
35348
+ var __webpack_exports__Gemma3PreTrainedModel = __webpack_exports__.Gemma3PreTrainedModel;
35195
35349
  var __webpack_exports__GemmaForCausalLM = __webpack_exports__.GemmaForCausalLM;
35196
35350
  var __webpack_exports__GemmaModel = __webpack_exports__.GemmaModel;
35197
35351
  var __webpack_exports__GemmaPreTrainedModel = __webpack_exports__.GemmaPreTrainedModel;
@@ -35293,6 +35447,10 @@ var __webpack_exports__MaskFormerModel = __webpack_exports__.MaskFormerModel;
35293
35447
  var __webpack_exports__MaskFormerPreTrainedModel = __webpack_exports__.MaskFormerPreTrainedModel;
35294
35448
  var __webpack_exports__MaskedLMOutput = __webpack_exports__.MaskedLMOutput;
35295
35449
  var __webpack_exports__MaxLengthCriteria = __webpack_exports__.MaxLengthCriteria;
35450
+ var __webpack_exports__Metric3DForDepthEstimation = __webpack_exports__.Metric3DForDepthEstimation;
35451
+ var __webpack_exports__Metric3DPreTrainedModel = __webpack_exports__.Metric3DPreTrainedModel;
35452
+ var __webpack_exports__Metric3Dv2ForDepthEstimation = __webpack_exports__.Metric3Dv2ForDepthEstimation;
35453
+ var __webpack_exports__Metric3Dv2PreTrainedModel = __webpack_exports__.Metric3Dv2PreTrainedModel;
35296
35454
  var __webpack_exports__MgpstrForSceneTextRecognition = __webpack_exports__.MgpstrForSceneTextRecognition;
35297
35455
  var __webpack_exports__MgpstrModelOutput = __webpack_exports__.MgpstrModelOutput;
35298
35456
  var __webpack_exports__MgpstrPreTrainedModel = __webpack_exports__.MgpstrPreTrainedModel;
@@ -35499,6 +35657,11 @@ var __webpack_exports__SiglipVisionModel = __webpack_exports__.SiglipVisionModel
35499
35657
  var __webpack_exports__SmolVLMForConditionalGeneration = __webpack_exports__.SmolVLMForConditionalGeneration;
35500
35658
  var __webpack_exports__SmolVLMImageProcessor = __webpack_exports__.SmolVLMImageProcessor;
35501
35659
  var __webpack_exports__SmolVLMProcessor = __webpack_exports__.SmolVLMProcessor;
35660
+ var __webpack_exports__SnacDecoderModel = __webpack_exports__.SnacDecoderModel;
35661
+ var __webpack_exports__SnacEncoderModel = __webpack_exports__.SnacEncoderModel;
35662
+ var __webpack_exports__SnacFeatureExtractor = __webpack_exports__.SnacFeatureExtractor;
35663
+ var __webpack_exports__SnacModel = __webpack_exports__.SnacModel;
35664
+ var __webpack_exports__SnacPreTrainedModel = __webpack_exports__.SnacPreTrainedModel;
35502
35665
  var __webpack_exports__SpeechT5FeatureExtractor = __webpack_exports__.SpeechT5FeatureExtractor;
35503
35666
  var __webpack_exports__SpeechT5ForSpeechToText = __webpack_exports__.SpeechT5ForSpeechToText;
35504
35667
  var __webpack_exports__SpeechT5ForTextToSpeech = __webpack_exports__.SpeechT5ForTextToSpeech;
@@ -35692,6 +35855,6 @@ var __webpack_exports__topk = __webpack_exports__.topk;
35692
35855
  var __webpack_exports__window_function = __webpack_exports__.window_function;
35693
35856
  var __webpack_exports__zeros = __webpack_exports__.zeros;
35694
35857
  var __webpack_exports__zeros_like = __webpack_exports__.zeros_like;
35695
- export { __webpack_exports__ASTFeatureExtractor as ASTFeatureExtractor, __webpack_exports__ASTForAudioClassification as ASTForAudioClassification, __webpack_exports__ASTModel as ASTModel, __webpack_exports__ASTPreTrainedModel as ASTPreTrainedModel, __webpack_exports__AlbertForMaskedLM as AlbertForMaskedLM, __webpack_exports__AlbertForQuestionAnswering as AlbertForQuestionAnswering, __webpack_exports__AlbertForSequenceClassification as AlbertForSequenceClassification, __webpack_exports__AlbertModel as AlbertModel, __webpack_exports__AlbertPreTrainedModel as AlbertPreTrainedModel, __webpack_exports__AlbertTokenizer as AlbertTokenizer, __webpack_exports__AudioClassificationPipeline as AudioClassificationPipeline, __webpack_exports__AutoConfig as AutoConfig, __webpack_exports__AutoFeatureExtractor as AutoFeatureExtractor, __webpack_exports__AutoImageProcessor as AutoImageProcessor, __webpack_exports__AutoModel as AutoModel, __webpack_exports__AutoModelForAudioClassification as AutoModelForAudioClassification, __webpack_exports__AutoModelForAudioFrameClassification as AutoModelForAudioFrameClassification, __webpack_exports__AutoModelForAudioTextToText as AutoModelForAudioTextToText, __webpack_exports__AutoModelForCTC as AutoModelForCTC, __webpack_exports__AutoModelForCausalLM as AutoModelForCausalLM, __webpack_exports__AutoModelForDepthEstimation as AutoModelForDepthEstimation, __webpack_exports__AutoModelForDocumentQuestionAnswering as AutoModelForDocumentQuestionAnswering, __webpack_exports__AutoModelForImageClassification as AutoModelForImageClassification, __webpack_exports__AutoModelForImageFeatureExtraction as AutoModelForImageFeatureExtraction, __webpack_exports__AutoModelForImageMatting as AutoModelForImageMatting, __webpack_exports__AutoModelForImageSegmentation as AutoModelForImageSegmentation, __webpack_exports__AutoModelForImageTextToText as AutoModelForImageTextToText, __webpack_exports__AutoModelForImageToImage as AutoModelForImageToImage, __webpack_exports__AutoModelForMaskGeneration as AutoModelForMaskGeneration, __webpack_exports__AutoModelForMaskedLM as AutoModelForMaskedLM, __webpack_exports__AutoModelForNormalEstimation as AutoModelForNormalEstimation, __webpack_exports__AutoModelForObjectDetection as AutoModelForObjectDetection, __webpack_exports__AutoModelForPoseEstimation as AutoModelForPoseEstimation, __webpack_exports__AutoModelForQuestionAnswering as AutoModelForQuestionAnswering, __webpack_exports__AutoModelForSemanticSegmentation as AutoModelForSemanticSegmentation, __webpack_exports__AutoModelForSeq2SeqLM as AutoModelForSeq2SeqLM, __webpack_exports__AutoModelForSequenceClassification as AutoModelForSequenceClassification, __webpack_exports__AutoModelForSpeechSeq2Seq as AutoModelForSpeechSeq2Seq, __webpack_exports__AutoModelForTextToSpectrogram as AutoModelForTextToSpectrogram, __webpack_exports__AutoModelForTextToWaveform as AutoModelForTextToWaveform, __webpack_exports__AutoModelForTokenClassification as AutoModelForTokenClassification, __webpack_exports__AutoModelForUniversalSegmentation as AutoModelForUniversalSegmentation, __webpack_exports__AutoModelForVision2Seq as AutoModelForVision2Seq, __webpack_exports__AutoModelForXVector as AutoModelForXVector, __webpack_exports__AutoModelForZeroShotObjectDetection as AutoModelForZeroShotObjectDetection, __webpack_exports__AutoProcessor as AutoProcessor, __webpack_exports__AutoTokenizer as AutoTokenizer, __webpack_exports__AutomaticSpeechRecognitionPipeline as AutomaticSpeechRecognitionPipeline, __webpack_exports__BackgroundRemovalPipeline as BackgroundRemovalPipeline, __webpack_exports__BartForConditionalGeneration as BartForConditionalGeneration, __webpack_exports__BartForSequenceClassification as BartForSequenceClassification, __webpack_exports__BartModel as BartModel, __webpack_exports__BartPretrainedModel as BartPretrainedModel, __webpack_exports__BartTokenizer as BartTokenizer, __webpack_exports__BaseModelOutput as BaseModelOutput, __webpack_exports__BaseStreamer as BaseStreamer, __webpack_exports__BeitFeatureExtractor as BeitFeatureExtractor, __webpack_exports__BeitForImageClassification as BeitForImageClassification, __webpack_exports__BeitModel as BeitModel, __webpack_exports__BeitPreTrainedModel as BeitPreTrainedModel, __webpack_exports__BertForMaskedLM as BertForMaskedLM, __webpack_exports__BertForQuestionAnswering as BertForQuestionAnswering, __webpack_exports__BertForSequenceClassification as BertForSequenceClassification, __webpack_exports__BertForTokenClassification as BertForTokenClassification, __webpack_exports__BertModel as BertModel, __webpack_exports__BertPreTrainedModel as BertPreTrainedModel, __webpack_exports__BertTokenizer as BertTokenizer, __webpack_exports__BitImageProcessor as BitImageProcessor, __webpack_exports__BlenderbotForConditionalGeneration as BlenderbotForConditionalGeneration, __webpack_exports__BlenderbotModel as BlenderbotModel, __webpack_exports__BlenderbotPreTrainedModel as BlenderbotPreTrainedModel, __webpack_exports__BlenderbotSmallForConditionalGeneration as BlenderbotSmallForConditionalGeneration, __webpack_exports__BlenderbotSmallModel as BlenderbotSmallModel, __webpack_exports__BlenderbotSmallPreTrainedModel as BlenderbotSmallPreTrainedModel, __webpack_exports__BlenderbotSmallTokenizer as BlenderbotSmallTokenizer, __webpack_exports__BlenderbotTokenizer as BlenderbotTokenizer, __webpack_exports__BloomForCausalLM as BloomForCausalLM, __webpack_exports__BloomModel as BloomModel, __webpack_exports__BloomPreTrainedModel as BloomPreTrainedModel, __webpack_exports__BloomTokenizer as BloomTokenizer, __webpack_exports__CLIPFeatureExtractor as CLIPFeatureExtractor, __webpack_exports__CLIPImageProcessor as CLIPImageProcessor, __webpack_exports__CLIPModel as CLIPModel, __webpack_exports__CLIPPreTrainedModel as CLIPPreTrainedModel, __webpack_exports__CLIPSegForImageSegmentation as CLIPSegForImageSegmentation, __webpack_exports__CLIPSegModel as CLIPSegModel, __webpack_exports__CLIPSegPreTrainedModel as CLIPSegPreTrainedModel, __webpack_exports__CLIPTextModel as CLIPTextModel, __webpack_exports__CLIPTextModelWithProjection as CLIPTextModelWithProjection, __webpack_exports__CLIPTokenizer as CLIPTokenizer, __webpack_exports__CLIPVisionModel as CLIPVisionModel, __webpack_exports__CLIPVisionModelWithProjection as CLIPVisionModelWithProjection, __webpack_exports__CamembertForMaskedLM as CamembertForMaskedLM, __webpack_exports__CamembertForQuestionAnswering as CamembertForQuestionAnswering, __webpack_exports__CamembertForSequenceClassification as CamembertForSequenceClassification, __webpack_exports__CamembertForTokenClassification as CamembertForTokenClassification, __webpack_exports__CamembertModel as CamembertModel, __webpack_exports__CamembertPreTrainedModel as CamembertPreTrainedModel, __webpack_exports__CamembertTokenizer as CamembertTokenizer, __webpack_exports__CausalLMOutput as CausalLMOutput, __webpack_exports__CausalLMOutputWithPast as CausalLMOutputWithPast, __webpack_exports__ChineseCLIPFeatureExtractor as ChineseCLIPFeatureExtractor, __webpack_exports__ChineseCLIPModel as ChineseCLIPModel, __webpack_exports__ChineseCLIPPreTrainedModel as ChineseCLIPPreTrainedModel, __webpack_exports__ClapAudioModelWithProjection as ClapAudioModelWithProjection, __webpack_exports__ClapFeatureExtractor as ClapFeatureExtractor, __webpack_exports__ClapModel as ClapModel, __webpack_exports__ClapPreTrainedModel as ClapPreTrainedModel, __webpack_exports__ClapTextModelWithProjection as ClapTextModelWithProjection, __webpack_exports__ClassifierFreeGuidanceLogitsProcessor as ClassifierFreeGuidanceLogitsProcessor, __webpack_exports__CodeGenForCausalLM as CodeGenForCausalLM, __webpack_exports__CodeGenModel as CodeGenModel, __webpack_exports__CodeGenPreTrainedModel as CodeGenPreTrainedModel, __webpack_exports__CodeGenTokenizer as CodeGenTokenizer, __webpack_exports__CodeLlamaTokenizer as CodeLlamaTokenizer, __webpack_exports__CohereForCausalLM as CohereForCausalLM, __webpack_exports__CohereModel as CohereModel, __webpack_exports__CoherePreTrainedModel as CoherePreTrainedModel, __webpack_exports__CohereTokenizer as CohereTokenizer, __webpack_exports__ConvBertForMaskedLM as ConvBertForMaskedLM, __webpack_exports__ConvBertForQuestionAnswering as ConvBertForQuestionAnswering, __webpack_exports__ConvBertForSequenceClassification as ConvBertForSequenceClassification, __webpack_exports__ConvBertForTokenClassification as ConvBertForTokenClassification, __webpack_exports__ConvBertModel as ConvBertModel, __webpack_exports__ConvBertPreTrainedModel as ConvBertPreTrainedModel, __webpack_exports__ConvBertTokenizer as ConvBertTokenizer, __webpack_exports__ConvNextFeatureExtractor as ConvNextFeatureExtractor, __webpack_exports__ConvNextForImageClassification as ConvNextForImageClassification, __webpack_exports__ConvNextImageProcessor as ConvNextImageProcessor, __webpack_exports__ConvNextModel as ConvNextModel, __webpack_exports__ConvNextPreTrainedModel as ConvNextPreTrainedModel, __webpack_exports__ConvNextV2ForImageClassification as ConvNextV2ForImageClassification, __webpack_exports__ConvNextV2Model as ConvNextV2Model, __webpack_exports__ConvNextV2PreTrainedModel as ConvNextV2PreTrainedModel, __webpack_exports__DPTFeatureExtractor as DPTFeatureExtractor, __webpack_exports__DPTForDepthEstimation as DPTForDepthEstimation, __webpack_exports__DPTImageProcessor as DPTImageProcessor, __webpack_exports__DPTModel as DPTModel, __webpack_exports__DPTPreTrainedModel as DPTPreTrainedModel, __webpack_exports__DacDecoderModel as DacDecoderModel, __webpack_exports__DacDecoderOutput as DacDecoderOutput, __webpack_exports__DacEncoderModel as DacEncoderModel, __webpack_exports__DacEncoderOutput as DacEncoderOutput, __webpack_exports__DacFeatureExtractor as DacFeatureExtractor, __webpack_exports__DacModel as DacModel, __webpack_exports__DacPreTrainedModel as DacPreTrainedModel, __webpack_exports__DataTypeMap as DataTypeMap, __webpack_exports__DebertaForMaskedLM as DebertaForMaskedLM, __webpack_exports__DebertaForQuestionAnswering as DebertaForQuestionAnswering, __webpack_exports__DebertaForSequenceClassification as DebertaForSequenceClassification, __webpack_exports__DebertaForTokenClassification as DebertaForTokenClassification, __webpack_exports__DebertaModel as DebertaModel, __webpack_exports__DebertaPreTrainedModel as DebertaPreTrainedModel, __webpack_exports__DebertaTokenizer as DebertaTokenizer, __webpack_exports__DebertaV2ForMaskedLM as DebertaV2ForMaskedLM, __webpack_exports__DebertaV2ForQuestionAnswering as DebertaV2ForQuestionAnswering, __webpack_exports__DebertaV2ForSequenceClassification as DebertaV2ForSequenceClassification, __webpack_exports__DebertaV2ForTokenClassification as DebertaV2ForTokenClassification, __webpack_exports__DebertaV2Model as DebertaV2Model, __webpack_exports__DebertaV2PreTrainedModel as DebertaV2PreTrainedModel, __webpack_exports__DebertaV2Tokenizer as DebertaV2Tokenizer, __webpack_exports__DecisionTransformerModel as DecisionTransformerModel, __webpack_exports__DecisionTransformerPreTrainedModel as DecisionTransformerPreTrainedModel, __webpack_exports__DeiTFeatureExtractor as DeiTFeatureExtractor, __webpack_exports__DeiTForImageClassification as DeiTForImageClassification, __webpack_exports__DeiTImageProcessor as DeiTImageProcessor, __webpack_exports__DeiTModel as DeiTModel, __webpack_exports__DeiTPreTrainedModel as DeiTPreTrainedModel, __webpack_exports__DepthAnythingForDepthEstimation as DepthAnythingForDepthEstimation, __webpack_exports__DepthAnythingPreTrainedModel as DepthAnythingPreTrainedModel, __webpack_exports__DepthEstimationPipeline as DepthEstimationPipeline, __webpack_exports__DepthProForDepthEstimation as DepthProForDepthEstimation, __webpack_exports__DepthProPreTrainedModel as DepthProPreTrainedModel, __webpack_exports__DetrFeatureExtractor as DetrFeatureExtractor, __webpack_exports__DetrForObjectDetection as DetrForObjectDetection, __webpack_exports__DetrForSegmentation as DetrForSegmentation, __webpack_exports__DetrImageProcessor as DetrImageProcessor, __webpack_exports__DetrModel as DetrModel, __webpack_exports__DetrObjectDetectionOutput as DetrObjectDetectionOutput, __webpack_exports__DetrPreTrainedModel as DetrPreTrainedModel, __webpack_exports__DetrSegmentationOutput as DetrSegmentationOutput, __webpack_exports__Dinov2ForImageClassification as Dinov2ForImageClassification, __webpack_exports__Dinov2Model as Dinov2Model, __webpack_exports__Dinov2PreTrainedModel as Dinov2PreTrainedModel, __webpack_exports__Dinov2WithRegistersForImageClassification as Dinov2WithRegistersForImageClassification, __webpack_exports__Dinov2WithRegistersModel as Dinov2WithRegistersModel, __webpack_exports__Dinov2WithRegistersPreTrainedModel as Dinov2WithRegistersPreTrainedModel, __webpack_exports__DistilBertForMaskedLM as DistilBertForMaskedLM, __webpack_exports__DistilBertForQuestionAnswering as DistilBertForQuestionAnswering, __webpack_exports__DistilBertForSequenceClassification as DistilBertForSequenceClassification, __webpack_exports__DistilBertForTokenClassification as DistilBertForTokenClassification, __webpack_exports__DistilBertModel as DistilBertModel, __webpack_exports__DistilBertPreTrainedModel as DistilBertPreTrainedModel, __webpack_exports__DistilBertTokenizer as DistilBertTokenizer, __webpack_exports__DocumentQuestionAnsweringPipeline as DocumentQuestionAnsweringPipeline, __webpack_exports__DonutFeatureExtractor as DonutFeatureExtractor, __webpack_exports__DonutImageProcessor as DonutImageProcessor, __webpack_exports__DonutSwinModel as DonutSwinModel, __webpack_exports__DonutSwinPreTrainedModel as DonutSwinPreTrainedModel, __webpack_exports__EfficientNetForImageClassification as EfficientNetForImageClassification, __webpack_exports__EfficientNetImageProcessor as EfficientNetImageProcessor, __webpack_exports__EfficientNetModel as EfficientNetModel, __webpack_exports__EfficientNetPreTrainedModel as EfficientNetPreTrainedModel, __webpack_exports__ElectraForMaskedLM as ElectraForMaskedLM, __webpack_exports__ElectraForQuestionAnswering as ElectraForQuestionAnswering, __webpack_exports__ElectraForSequenceClassification as ElectraForSequenceClassification, __webpack_exports__ElectraForTokenClassification as ElectraForTokenClassification, __webpack_exports__ElectraModel as ElectraModel, __webpack_exports__ElectraPreTrainedModel as ElectraPreTrainedModel, __webpack_exports__ElectraTokenizer as ElectraTokenizer, __webpack_exports__EncodecFeatureExtractor as EncodecFeatureExtractor, __webpack_exports__EosTokenCriteria as EosTokenCriteria, __webpack_exports__EsmForMaskedLM as EsmForMaskedLM, __webpack_exports__EsmForSequenceClassification as EsmForSequenceClassification, __webpack_exports__EsmForTokenClassification as EsmForTokenClassification, __webpack_exports__EsmModel as EsmModel, __webpack_exports__EsmPreTrainedModel as EsmPreTrainedModel, __webpack_exports__EsmTokenizer as EsmTokenizer, __webpack_exports__ExaoneForCausalLM as ExaoneForCausalLM, __webpack_exports__ExaoneModel as ExaoneModel, __webpack_exports__ExaonePreTrainedModel as ExaonePreTrainedModel, __webpack_exports__FFT as FFT, __webpack_exports__FalconForCausalLM as FalconForCausalLM, __webpack_exports__FalconModel as FalconModel, __webpack_exports__FalconPreTrainedModel as FalconPreTrainedModel, __webpack_exports__FalconTokenizer as FalconTokenizer, __webpack_exports__FastViTForImageClassification as FastViTForImageClassification, __webpack_exports__FastViTModel as FastViTModel, __webpack_exports__FastViTPreTrainedModel as FastViTPreTrainedModel, __webpack_exports__FeatureExtractionPipeline as FeatureExtractionPipeline, __webpack_exports__FeatureExtractor as FeatureExtractor, __webpack_exports__FillMaskPipeline as FillMaskPipeline, __webpack_exports__Florence2ForConditionalGeneration as Florence2ForConditionalGeneration, __webpack_exports__Florence2PreTrainedModel as Florence2PreTrainedModel, __webpack_exports__Florence2Processor as Florence2Processor, __webpack_exports__ForcedBOSTokenLogitsProcessor as ForcedBOSTokenLogitsProcessor, __webpack_exports__ForcedEOSTokenLogitsProcessor as ForcedEOSTokenLogitsProcessor, __webpack_exports__GLPNFeatureExtractor as GLPNFeatureExtractor, __webpack_exports__GLPNForDepthEstimation as GLPNForDepthEstimation, __webpack_exports__GLPNModel as GLPNModel, __webpack_exports__GLPNPreTrainedModel as GLPNPreTrainedModel, __webpack_exports__GPT2LMHeadModel as GPT2LMHeadModel, __webpack_exports__GPT2Model as GPT2Model, __webpack_exports__GPT2PreTrainedModel as GPT2PreTrainedModel, __webpack_exports__GPT2Tokenizer as GPT2Tokenizer, __webpack_exports__GPTBigCodeForCausalLM as GPTBigCodeForCausalLM, __webpack_exports__GPTBigCodeModel as GPTBigCodeModel, __webpack_exports__GPTBigCodePreTrainedModel as GPTBigCodePreTrainedModel, __webpack_exports__GPTJForCausalLM as GPTJForCausalLM, __webpack_exports__GPTJModel as GPTJModel, __webpack_exports__GPTJPreTrainedModel as GPTJPreTrainedModel, __webpack_exports__GPTNeoForCausalLM as GPTNeoForCausalLM, __webpack_exports__GPTNeoModel as GPTNeoModel, __webpack_exports__GPTNeoPreTrainedModel as GPTNeoPreTrainedModel, __webpack_exports__GPTNeoXForCausalLM as GPTNeoXForCausalLM, __webpack_exports__GPTNeoXModel as GPTNeoXModel, __webpack_exports__GPTNeoXPreTrainedModel as GPTNeoXPreTrainedModel, __webpack_exports__GPTNeoXTokenizer as GPTNeoXTokenizer, __webpack_exports__Gemma2ForCausalLM as Gemma2ForCausalLM, __webpack_exports__Gemma2Model as Gemma2Model, __webpack_exports__Gemma2PreTrainedModel as Gemma2PreTrainedModel, __webpack_exports__GemmaForCausalLM as GemmaForCausalLM, __webpack_exports__GemmaModel as GemmaModel, __webpack_exports__GemmaPreTrainedModel as GemmaPreTrainedModel, __webpack_exports__GemmaTokenizer as GemmaTokenizer, __webpack_exports__GlmForCausalLM as GlmForCausalLM, __webpack_exports__GlmModel as GlmModel, __webpack_exports__GlmPreTrainedModel as GlmPreTrainedModel, __webpack_exports__GraniteForCausalLM as GraniteForCausalLM, __webpack_exports__GraniteModel as GraniteModel, __webpack_exports__GranitePreTrainedModel as GranitePreTrainedModel, __webpack_exports__Grok1Tokenizer as Grok1Tokenizer, __webpack_exports__GroundingDinoForObjectDetection as GroundingDinoForObjectDetection, __webpack_exports__GroundingDinoImageProcessor as GroundingDinoImageProcessor, __webpack_exports__GroundingDinoPreTrainedModel as GroundingDinoPreTrainedModel, __webpack_exports__GroundingDinoProcessor as GroundingDinoProcessor, __webpack_exports__GroupViTModel as GroupViTModel, __webpack_exports__GroupViTPreTrainedModel as GroupViTPreTrainedModel, __webpack_exports__HeliumForCausalLM as HeliumForCausalLM, __webpack_exports__HeliumModel as HeliumModel, __webpack_exports__HeliumPreTrainedModel as HeliumPreTrainedModel, __webpack_exports__HerbertTokenizer as HerbertTokenizer, __webpack_exports__HieraForImageClassification as HieraForImageClassification, __webpack_exports__HieraModel as HieraModel, __webpack_exports__HieraPreTrainedModel as HieraPreTrainedModel, __webpack_exports__HubertForCTC as HubertForCTC, __webpack_exports__HubertForSequenceClassification as HubertForSequenceClassification, __webpack_exports__HubertModel as HubertModel, __webpack_exports__HubertPreTrainedModel as HubertPreTrainedModel, __webpack_exports__IJepaForImageClassification as IJepaForImageClassification, __webpack_exports__IJepaModel as IJepaModel, __webpack_exports__IJepaPreTrainedModel as IJepaPreTrainedModel, __webpack_exports__Idefics3ForConditionalGeneration as Idefics3ForConditionalGeneration, __webpack_exports__Idefics3ImageProcessor as Idefics3ImageProcessor, __webpack_exports__Idefics3PreTrainedModel as Idefics3PreTrainedModel, __webpack_exports__Idefics3Processor as Idefics3Processor, __webpack_exports__ImageClassificationPipeline as ImageClassificationPipeline, __webpack_exports__ImageFeatureExtractionPipeline as ImageFeatureExtractionPipeline, __webpack_exports__ImageFeatureExtractor as ImageFeatureExtractor, __webpack_exports__ImageMattingOutput as ImageMattingOutput, __webpack_exports__ImageProcessor as ImageProcessor, __webpack_exports__ImageSegmentationPipeline as ImageSegmentationPipeline, __webpack_exports__ImageToImagePipeline as ImageToImagePipeline, __webpack_exports__ImageToTextPipeline as ImageToTextPipeline, __webpack_exports__InterruptableStoppingCriteria as InterruptableStoppingCriteria, __webpack_exports__JAISLMHeadModel as JAISLMHeadModel, __webpack_exports__JAISModel as JAISModel, __webpack_exports__JAISPreTrainedModel as JAISPreTrainedModel, __webpack_exports__JinaCLIPImageProcessor as JinaCLIPImageProcessor, __webpack_exports__JinaCLIPModel as JinaCLIPModel, __webpack_exports__JinaCLIPPreTrainedModel as JinaCLIPPreTrainedModel, __webpack_exports__JinaCLIPProcessor as JinaCLIPProcessor, __webpack_exports__JinaCLIPTextModel as JinaCLIPTextModel, __webpack_exports__JinaCLIPVisionModel as JinaCLIPVisionModel, __webpack_exports__LiteWhisperForConditionalGeneration as LiteWhisperForConditionalGeneration, __webpack_exports__LlamaForCausalLM as LlamaForCausalLM, __webpack_exports__LlamaModel as LlamaModel, __webpack_exports__LlamaPreTrainedModel as LlamaPreTrainedModel, __webpack_exports__LlamaTokenizer as LlamaTokenizer, __webpack_exports__LlavaForConditionalGeneration as LlavaForConditionalGeneration, __webpack_exports__LlavaOnevisionForConditionalGeneration as LlavaOnevisionForConditionalGeneration, __webpack_exports__LlavaOnevisionImageProcessor as LlavaOnevisionImageProcessor, __webpack_exports__LlavaPreTrainedModel as LlavaPreTrainedModel, __webpack_exports__LogitsProcessor as LogitsProcessor, __webpack_exports__LogitsProcessorList as LogitsProcessorList, __webpack_exports__LogitsWarper as LogitsWarper, __webpack_exports__LongT5ForConditionalGeneration as LongT5ForConditionalGeneration, __webpack_exports__LongT5Model as LongT5Model, __webpack_exports__LongT5PreTrainedModel as LongT5PreTrainedModel, __webpack_exports__M2M100ForConditionalGeneration as M2M100ForConditionalGeneration, __webpack_exports__M2M100Model as M2M100Model, __webpack_exports__M2M100PreTrainedModel as M2M100PreTrainedModel, __webpack_exports__M2M100Tokenizer as M2M100Tokenizer, __webpack_exports__MBart50Tokenizer as MBart50Tokenizer, __webpack_exports__MBartForCausalLM as MBartForCausalLM, __webpack_exports__MBartForConditionalGeneration as MBartForConditionalGeneration, __webpack_exports__MBartForSequenceClassification as MBartForSequenceClassification, __webpack_exports__MBartModel as MBartModel, __webpack_exports__MBartPreTrainedModel as MBartPreTrainedModel, __webpack_exports__MBartTokenizer as MBartTokenizer, __webpack_exports__MPNetForMaskedLM as MPNetForMaskedLM, __webpack_exports__MPNetForQuestionAnswering as MPNetForQuestionAnswering, __webpack_exports__MPNetForSequenceClassification as MPNetForSequenceClassification, __webpack_exports__MPNetForTokenClassification as MPNetForTokenClassification, __webpack_exports__MPNetModel as MPNetModel, __webpack_exports__MPNetPreTrainedModel as MPNetPreTrainedModel, __webpack_exports__MPNetTokenizer as MPNetTokenizer, __webpack_exports__MT5ForConditionalGeneration as MT5ForConditionalGeneration, __webpack_exports__MT5Model as MT5Model, __webpack_exports__MT5PreTrainedModel as MT5PreTrainedModel, __webpack_exports__MarianMTModel as MarianMTModel, __webpack_exports__MarianModel as MarianModel, __webpack_exports__MarianPreTrainedModel as MarianPreTrainedModel, __webpack_exports__MarianTokenizer as MarianTokenizer, __webpack_exports__Mask2FormerImageProcessor as Mask2FormerImageProcessor, __webpack_exports__MaskFormerFeatureExtractor as MaskFormerFeatureExtractor, __webpack_exports__MaskFormerForInstanceSegmentation as MaskFormerForInstanceSegmentation, __webpack_exports__MaskFormerImageProcessor as MaskFormerImageProcessor, __webpack_exports__MaskFormerModel as MaskFormerModel, __webpack_exports__MaskFormerPreTrainedModel as MaskFormerPreTrainedModel, __webpack_exports__MaskedLMOutput as MaskedLMOutput, __webpack_exports__MaxLengthCriteria as MaxLengthCriteria, __webpack_exports__MgpstrForSceneTextRecognition as MgpstrForSceneTextRecognition, __webpack_exports__MgpstrModelOutput as MgpstrModelOutput, __webpack_exports__MgpstrPreTrainedModel as MgpstrPreTrainedModel, __webpack_exports__MgpstrProcessor as MgpstrProcessor, __webpack_exports__MgpstrTokenizer as MgpstrTokenizer, __webpack_exports__MimiDecoderModel as MimiDecoderModel, __webpack_exports__MimiDecoderOutput as MimiDecoderOutput, __webpack_exports__MimiEncoderModel as MimiEncoderModel, __webpack_exports__MimiEncoderOutput as MimiEncoderOutput, __webpack_exports__MimiModel as MimiModel, __webpack_exports__MimiPreTrainedModel as MimiPreTrainedModel, __webpack_exports__MinLengthLogitsProcessor as MinLengthLogitsProcessor, __webpack_exports__MinNewTokensLengthLogitsProcessor as MinNewTokensLengthLogitsProcessor, __webpack_exports__MistralForCausalLM as MistralForCausalLM, __webpack_exports__MistralModel as MistralModel, __webpack_exports__MistralPreTrainedModel as MistralPreTrainedModel, __webpack_exports__MobileBertForMaskedLM as MobileBertForMaskedLM, __webpack_exports__MobileBertForQuestionAnswering as MobileBertForQuestionAnswering, __webpack_exports__MobileBertForSequenceClassification as MobileBertForSequenceClassification, __webpack_exports__MobileBertModel as MobileBertModel, __webpack_exports__MobileBertPreTrainedModel as MobileBertPreTrainedModel, __webpack_exports__MobileBertTokenizer as MobileBertTokenizer, __webpack_exports__MobileLLMForCausalLM as MobileLLMForCausalLM, __webpack_exports__MobileLLMModel as MobileLLMModel, __webpack_exports__MobileLLMPreTrainedModel as MobileLLMPreTrainedModel, __webpack_exports__MobileNetV1FeatureExtractor as MobileNetV1FeatureExtractor, __webpack_exports__MobileNetV1ForImageClassification as MobileNetV1ForImageClassification, __webpack_exports__MobileNetV1ForSemanticSegmentation as MobileNetV1ForSemanticSegmentation, __webpack_exports__MobileNetV1ImageProcessor as MobileNetV1ImageProcessor, __webpack_exports__MobileNetV1Model as MobileNetV1Model, __webpack_exports__MobileNetV1PreTrainedModel as MobileNetV1PreTrainedModel, __webpack_exports__MobileNetV2FeatureExtractor as MobileNetV2FeatureExtractor, __webpack_exports__MobileNetV2ForImageClassification as MobileNetV2ForImageClassification, __webpack_exports__MobileNetV2ForSemanticSegmentation as MobileNetV2ForSemanticSegmentation, __webpack_exports__MobileNetV2ImageProcessor as MobileNetV2ImageProcessor, __webpack_exports__MobileNetV2Model as MobileNetV2Model, __webpack_exports__MobileNetV2PreTrainedModel as MobileNetV2PreTrainedModel, __webpack_exports__MobileNetV3FeatureExtractor as MobileNetV3FeatureExtractor, __webpack_exports__MobileNetV3ForImageClassification as MobileNetV3ForImageClassification, __webpack_exports__MobileNetV3ForSemanticSegmentation as MobileNetV3ForSemanticSegmentation, __webpack_exports__MobileNetV3ImageProcessor as MobileNetV3ImageProcessor, __webpack_exports__MobileNetV3Model as MobileNetV3Model, __webpack_exports__MobileNetV3PreTrainedModel as MobileNetV3PreTrainedModel, __webpack_exports__MobileNetV4FeatureExtractor as MobileNetV4FeatureExtractor, __webpack_exports__MobileNetV4ForImageClassification as MobileNetV4ForImageClassification, __webpack_exports__MobileNetV4ForSemanticSegmentation as MobileNetV4ForSemanticSegmentation, __webpack_exports__MobileNetV4ImageProcessor as MobileNetV4ImageProcessor, __webpack_exports__MobileNetV4Model as MobileNetV4Model, __webpack_exports__MobileNetV4PreTrainedModel as MobileNetV4PreTrainedModel, __webpack_exports__MobileViTFeatureExtractor as MobileViTFeatureExtractor, __webpack_exports__MobileViTForImageClassification as MobileViTForImageClassification, __webpack_exports__MobileViTImageProcessor as MobileViTImageProcessor, __webpack_exports__MobileViTModel as MobileViTModel, __webpack_exports__MobileViTPreTrainedModel as MobileViTPreTrainedModel, __webpack_exports__MobileViTV2ForImageClassification as MobileViTV2ForImageClassification, __webpack_exports__MobileViTV2Model as MobileViTV2Model, __webpack_exports__MobileViTV2PreTrainedModel as MobileViTV2PreTrainedModel, __webpack_exports__ModelOutput as ModelOutput, __webpack_exports__ModernBertForMaskedLM as ModernBertForMaskedLM, __webpack_exports__ModernBertForSequenceClassification as ModernBertForSequenceClassification, __webpack_exports__ModernBertForTokenClassification as ModernBertForTokenClassification, __webpack_exports__ModernBertModel as ModernBertModel, __webpack_exports__ModernBertPreTrainedModel as ModernBertPreTrainedModel, __webpack_exports__Moondream1ForConditionalGeneration as Moondream1ForConditionalGeneration, __webpack_exports__MoonshineFeatureExtractor as MoonshineFeatureExtractor, __webpack_exports__MoonshineForConditionalGeneration as MoonshineForConditionalGeneration, __webpack_exports__MoonshineModel as MoonshineModel, __webpack_exports__MoonshinePreTrainedModel as MoonshinePreTrainedModel, __webpack_exports__MoonshineProcessor as MoonshineProcessor, __webpack_exports__MptForCausalLM as MptForCausalLM, __webpack_exports__MptModel as MptModel, __webpack_exports__MptPreTrainedModel as MptPreTrainedModel, __webpack_exports__MultiModalityCausalLM as MultiModalityCausalLM, __webpack_exports__MultiModalityPreTrainedModel as MultiModalityPreTrainedModel, __webpack_exports__MusicgenForCausalLM as MusicgenForCausalLM, __webpack_exports__MusicgenForConditionalGeneration as MusicgenForConditionalGeneration, __webpack_exports__MusicgenModel as MusicgenModel, __webpack_exports__MusicgenPreTrainedModel as MusicgenPreTrainedModel, __webpack_exports__NllbTokenizer as NllbTokenizer, __webpack_exports__NoBadWordsLogitsProcessor as NoBadWordsLogitsProcessor, __webpack_exports__NoRepeatNGramLogitsProcessor as NoRepeatNGramLogitsProcessor, __webpack_exports__NomicBertModel as NomicBertModel, __webpack_exports__NomicBertPreTrainedModel as NomicBertPreTrainedModel, __webpack_exports__NougatImageProcessor as NougatImageProcessor, __webpack_exports__NougatTokenizer as NougatTokenizer, __webpack_exports__OPTForCausalLM as OPTForCausalLM, __webpack_exports__OPTModel as OPTModel, __webpack_exports__OPTPreTrainedModel as OPTPreTrainedModel, __webpack_exports__ObjectDetectionPipeline as ObjectDetectionPipeline, __webpack_exports__Olmo2ForCausalLM as Olmo2ForCausalLM, __webpack_exports__Olmo2Model as Olmo2Model, __webpack_exports__Olmo2PreTrainedModel as Olmo2PreTrainedModel, __webpack_exports__OlmoForCausalLM as OlmoForCausalLM, __webpack_exports__OlmoModel as OlmoModel, __webpack_exports__OlmoPreTrainedModel as OlmoPreTrainedModel, __webpack_exports__OpenELMForCausalLM as OpenELMForCausalLM, __webpack_exports__OpenELMModel as OpenELMModel, __webpack_exports__OpenELMPreTrainedModel as OpenELMPreTrainedModel, __webpack_exports__OwlViTFeatureExtractor as OwlViTFeatureExtractor, __webpack_exports__OwlViTForObjectDetection as OwlViTForObjectDetection, __webpack_exports__OwlViTImageProcessor as OwlViTImageProcessor, __webpack_exports__OwlViTModel as OwlViTModel, __webpack_exports__OwlViTPreTrainedModel as OwlViTPreTrainedModel, __webpack_exports__OwlViTProcessor as OwlViTProcessor, __webpack_exports__Owlv2ForObjectDetection as Owlv2ForObjectDetection, __webpack_exports__Owlv2ImageProcessor as Owlv2ImageProcessor, __webpack_exports__Owlv2Model as Owlv2Model, __webpack_exports__Owlv2PreTrainedModel as Owlv2PreTrainedModel, __webpack_exports__PaliGemmaForConditionalGeneration as PaliGemmaForConditionalGeneration, __webpack_exports__PaliGemmaPreTrainedModel as PaliGemmaPreTrainedModel, __webpack_exports__PaliGemmaProcessor as PaliGemmaProcessor, __webpack_exports__PatchTSMixerForPrediction as PatchTSMixerForPrediction, __webpack_exports__PatchTSMixerModel as PatchTSMixerModel, __webpack_exports__PatchTSMixerPreTrainedModel as PatchTSMixerPreTrainedModel, __webpack_exports__PatchTSTForPrediction as PatchTSTForPrediction, __webpack_exports__PatchTSTModel as PatchTSTModel, __webpack_exports__PatchTSTPreTrainedModel as PatchTSTPreTrainedModel, __webpack_exports__Phi3ForCausalLM as Phi3ForCausalLM, __webpack_exports__Phi3Model as Phi3Model, __webpack_exports__Phi3PreTrainedModel as Phi3PreTrainedModel, __webpack_exports__Phi3VForCausalLM as Phi3VForCausalLM, __webpack_exports__Phi3VImageProcessor as Phi3VImageProcessor, __webpack_exports__Phi3VPreTrainedModel as Phi3VPreTrainedModel, __webpack_exports__Phi3VProcessor as Phi3VProcessor, __webpack_exports__PhiForCausalLM as PhiForCausalLM, __webpack_exports__PhiModel as PhiModel, __webpack_exports__PhiPreTrainedModel as PhiPreTrainedModel, __webpack_exports__Pipeline as Pipeline, __webpack_exports__PreTrainedModel as PreTrainedModel, __webpack_exports__PreTrainedTokenizer as PreTrainedTokenizer, __webpack_exports__PretrainedConfig as PretrainedConfig, __webpack_exports__PretrainedMixin as PretrainedMixin, __webpack_exports__Processor as Processor, __webpack_exports__PvtForImageClassification as PvtForImageClassification, __webpack_exports__PvtImageProcessor as PvtImageProcessor, __webpack_exports__PvtModel as PvtModel, __webpack_exports__PvtPreTrainedModel as PvtPreTrainedModel, __webpack_exports__PyAnnoteFeatureExtractor as PyAnnoteFeatureExtractor, __webpack_exports__PyAnnoteForAudioFrameClassification as PyAnnoteForAudioFrameClassification, __webpack_exports__PyAnnoteModel as PyAnnoteModel, __webpack_exports__PyAnnotePreTrainedModel as PyAnnotePreTrainedModel, __webpack_exports__PyAnnoteProcessor as PyAnnoteProcessor, __webpack_exports__QuestionAnsweringModelOutput as QuestionAnsweringModelOutput, __webpack_exports__QuestionAnsweringPipeline as QuestionAnsweringPipeline, __webpack_exports__Qwen2ForCausalLM as Qwen2ForCausalLM, __webpack_exports__Qwen2Model as Qwen2Model, __webpack_exports__Qwen2PreTrainedModel as Qwen2PreTrainedModel, __webpack_exports__Qwen2Tokenizer as Qwen2Tokenizer, __webpack_exports__Qwen2VLForConditionalGeneration as Qwen2VLForConditionalGeneration, __webpack_exports__Qwen2VLImageProcessor as Qwen2VLImageProcessor, __webpack_exports__Qwen2VLPreTrainedModel as Qwen2VLPreTrainedModel, __webpack_exports__Qwen2VLProcessor as Qwen2VLProcessor, __webpack_exports__RTDetrForObjectDetection as RTDetrForObjectDetection, __webpack_exports__RTDetrImageProcessor as RTDetrImageProcessor, __webpack_exports__RTDetrModel as RTDetrModel, __webpack_exports__RTDetrObjectDetectionOutput as RTDetrObjectDetectionOutput, __webpack_exports__RTDetrPreTrainedModel as RTDetrPreTrainedModel, __webpack_exports__RawAudio as RawAudio, __webpack_exports__RawImage as RawImage, __webpack_exports__RawVideo as RawVideo, __webpack_exports__RawVideoFrame as RawVideoFrame, __webpack_exports__RepetitionPenaltyLogitsProcessor as RepetitionPenaltyLogitsProcessor, __webpack_exports__ResNetForImageClassification as ResNetForImageClassification, __webpack_exports__ResNetModel as ResNetModel, __webpack_exports__ResNetPreTrainedModel as ResNetPreTrainedModel, __webpack_exports__RoFormerForMaskedLM as RoFormerForMaskedLM, __webpack_exports__RoFormerForQuestionAnswering as RoFormerForQuestionAnswering, __webpack_exports__RoFormerForSequenceClassification as RoFormerForSequenceClassification, __webpack_exports__RoFormerForTokenClassification as RoFormerForTokenClassification, __webpack_exports__RoFormerModel as RoFormerModel, __webpack_exports__RoFormerPreTrainedModel as RoFormerPreTrainedModel, __webpack_exports__RoFormerTokenizer as RoFormerTokenizer, __webpack_exports__RobertaForMaskedLM as RobertaForMaskedLM, __webpack_exports__RobertaForQuestionAnswering as RobertaForQuestionAnswering, __webpack_exports__RobertaForSequenceClassification as RobertaForSequenceClassification, __webpack_exports__RobertaForTokenClassification as RobertaForTokenClassification, __webpack_exports__RobertaModel as RobertaModel, __webpack_exports__RobertaPreTrainedModel as RobertaPreTrainedModel, __webpack_exports__RobertaTokenizer as RobertaTokenizer, __webpack_exports__SamImageProcessor as SamImageProcessor, __webpack_exports__SamImageSegmentationOutput as SamImageSegmentationOutput, __webpack_exports__SamModel as SamModel, __webpack_exports__SamPreTrainedModel as SamPreTrainedModel, __webpack_exports__SamProcessor as SamProcessor, __webpack_exports__SapiensForDepthEstimation as SapiensForDepthEstimation, __webpack_exports__SapiensForNormalEstimation as SapiensForNormalEstimation, __webpack_exports__SapiensForSemanticSegmentation as SapiensForSemanticSegmentation, __webpack_exports__SapiensPreTrainedModel as SapiensPreTrainedModel, __webpack_exports__SeamlessM4TFeatureExtractor as SeamlessM4TFeatureExtractor, __webpack_exports__SegformerFeatureExtractor as SegformerFeatureExtractor, __webpack_exports__SegformerForImageClassification as SegformerForImageClassification, __webpack_exports__SegformerForSemanticSegmentation as SegformerForSemanticSegmentation, __webpack_exports__SegformerImageProcessor as SegformerImageProcessor, __webpack_exports__SegformerModel as SegformerModel, __webpack_exports__SegformerPreTrainedModel as SegformerPreTrainedModel, __webpack_exports__Seq2SeqLMOutput as Seq2SeqLMOutput, __webpack_exports__SequenceClassifierOutput as SequenceClassifierOutput, __webpack_exports__SiglipImageProcessor as SiglipImageProcessor, __webpack_exports__SiglipModel as SiglipModel, __webpack_exports__SiglipPreTrainedModel as SiglipPreTrainedModel, __webpack_exports__SiglipTextModel as SiglipTextModel, __webpack_exports__SiglipTokenizer as SiglipTokenizer, __webpack_exports__SiglipVisionModel as SiglipVisionModel, __webpack_exports__SmolVLMForConditionalGeneration as SmolVLMForConditionalGeneration, __webpack_exports__SmolVLMImageProcessor as SmolVLMImageProcessor, __webpack_exports__SmolVLMProcessor as SmolVLMProcessor, __webpack_exports__SpeechT5FeatureExtractor as SpeechT5FeatureExtractor, __webpack_exports__SpeechT5ForSpeechToText as SpeechT5ForSpeechToText, __webpack_exports__SpeechT5ForTextToSpeech as SpeechT5ForTextToSpeech, __webpack_exports__SpeechT5HifiGan as SpeechT5HifiGan, __webpack_exports__SpeechT5Model as SpeechT5Model, __webpack_exports__SpeechT5PreTrainedModel as SpeechT5PreTrainedModel, __webpack_exports__SpeechT5Processor as SpeechT5Processor, __webpack_exports__SpeechT5Tokenizer as SpeechT5Tokenizer, __webpack_exports__SqueezeBertForMaskedLM as SqueezeBertForMaskedLM, __webpack_exports__SqueezeBertForQuestionAnswering as SqueezeBertForQuestionAnswering, __webpack_exports__SqueezeBertForSequenceClassification as SqueezeBertForSequenceClassification, __webpack_exports__SqueezeBertModel as SqueezeBertModel, __webpack_exports__SqueezeBertPreTrainedModel as SqueezeBertPreTrainedModel, __webpack_exports__SqueezeBertTokenizer as SqueezeBertTokenizer, __webpack_exports__StableLmForCausalLM as StableLmForCausalLM, __webpack_exports__StableLmModel as StableLmModel, __webpack_exports__StableLmPreTrainedModel as StableLmPreTrainedModel, __webpack_exports__Starcoder2ForCausalLM as Starcoder2ForCausalLM, __webpack_exports__Starcoder2Model as Starcoder2Model, __webpack_exports__Starcoder2PreTrainedModel as Starcoder2PreTrainedModel, __webpack_exports__StoppingCriteria as StoppingCriteria, __webpack_exports__StoppingCriteriaList as StoppingCriteriaList, __webpack_exports__StyleTextToSpeech2Model as StyleTextToSpeech2Model, __webpack_exports__StyleTextToSpeech2PreTrainedModel as StyleTextToSpeech2PreTrainedModel, __webpack_exports__SummarizationPipeline as SummarizationPipeline, __webpack_exports__SuppressTokensAtBeginLogitsProcessor as SuppressTokensAtBeginLogitsProcessor, __webpack_exports__Swin2SRForImageSuperResolution as Swin2SRForImageSuperResolution, __webpack_exports__Swin2SRImageProcessor as Swin2SRImageProcessor, __webpack_exports__Swin2SRModel as Swin2SRModel, __webpack_exports__Swin2SRPreTrainedModel as Swin2SRPreTrainedModel, __webpack_exports__SwinForImageClassification as SwinForImageClassification, __webpack_exports__SwinForSemanticSegmentation as SwinForSemanticSegmentation, __webpack_exports__SwinModel as SwinModel, __webpack_exports__SwinPreTrainedModel as SwinPreTrainedModel, __webpack_exports__T5ForConditionalGeneration as T5ForConditionalGeneration, __webpack_exports__T5Model as T5Model, __webpack_exports__T5PreTrainedModel as T5PreTrainedModel, __webpack_exports__T5Tokenizer as T5Tokenizer, __webpack_exports__TableTransformerForObjectDetection as TableTransformerForObjectDetection, __webpack_exports__TableTransformerModel as TableTransformerModel, __webpack_exports__TableTransformerObjectDetectionOutput as TableTransformerObjectDetectionOutput, __webpack_exports__TableTransformerPreTrainedModel as TableTransformerPreTrainedModel, __webpack_exports__TemperatureLogitsWarper as TemperatureLogitsWarper, __webpack_exports__Tensor as Tensor, __webpack_exports__Text2TextGenerationPipeline as Text2TextGenerationPipeline, __webpack_exports__TextClassificationPipeline as TextClassificationPipeline, __webpack_exports__TextGenerationPipeline as TextGenerationPipeline, __webpack_exports__TextStreamer as TextStreamer, __webpack_exports__TextToAudioPipeline as TextToAudioPipeline, __webpack_exports__TokenClassificationPipeline as TokenClassificationPipeline, __webpack_exports__TokenClassifierOutput as TokenClassifierOutput, __webpack_exports__TokenizerModel as TokenizerModel, __webpack_exports__TopKLogitsWarper as TopKLogitsWarper, __webpack_exports__TopPLogitsWarper as TopPLogitsWarper, __webpack_exports__TrOCRForCausalLM as TrOCRForCausalLM, __webpack_exports__TrOCRPreTrainedModel as TrOCRPreTrainedModel, __webpack_exports__TranslationPipeline as TranslationPipeline, __webpack_exports__UltravoxModel as UltravoxModel, __webpack_exports__UltravoxPreTrainedModel as UltravoxPreTrainedModel, __webpack_exports__UltravoxProcessor as UltravoxProcessor, __webpack_exports__UniSpeechForCTC as UniSpeechForCTC, __webpack_exports__UniSpeechForSequenceClassification as UniSpeechForSequenceClassification, __webpack_exports__UniSpeechModel as UniSpeechModel, __webpack_exports__UniSpeechPreTrainedModel as UniSpeechPreTrainedModel, __webpack_exports__UniSpeechSatForAudioFrameClassification as UniSpeechSatForAudioFrameClassification, __webpack_exports__UniSpeechSatForCTC as UniSpeechSatForCTC, __webpack_exports__UniSpeechSatForSequenceClassification as UniSpeechSatForSequenceClassification, __webpack_exports__UniSpeechSatModel as UniSpeechSatModel, __webpack_exports__UniSpeechSatPreTrainedModel as UniSpeechSatPreTrainedModel, __webpack_exports__VLChatProcessor as VLChatProcessor, __webpack_exports__VLMImageProcessor as VLMImageProcessor, __webpack_exports__ViTFeatureExtractor as ViTFeatureExtractor, __webpack_exports__ViTForImageClassification as ViTForImageClassification, __webpack_exports__ViTImageProcessor as ViTImageProcessor, __webpack_exports__ViTMAEModel as ViTMAEModel, __webpack_exports__ViTMAEPreTrainedModel as ViTMAEPreTrainedModel, __webpack_exports__ViTMSNForImageClassification as ViTMSNForImageClassification, __webpack_exports__ViTMSNModel as ViTMSNModel, __webpack_exports__ViTMSNPreTrainedModel as ViTMSNPreTrainedModel, __webpack_exports__ViTModel as ViTModel, __webpack_exports__ViTPreTrainedModel as ViTPreTrainedModel, __webpack_exports__VisionEncoderDecoderModel as VisionEncoderDecoderModel, __webpack_exports__VitMatteForImageMatting as VitMatteForImageMatting, __webpack_exports__VitMatteImageProcessor as VitMatteImageProcessor, __webpack_exports__VitMattePreTrainedModel as VitMattePreTrainedModel, __webpack_exports__VitPoseForPoseEstimation as VitPoseForPoseEstimation, __webpack_exports__VitPoseImageProcessor as VitPoseImageProcessor, __webpack_exports__VitPosePreTrainedModel as VitPosePreTrainedModel, __webpack_exports__VitsModel as VitsModel, __webpack_exports__VitsModelOutput as VitsModelOutput, __webpack_exports__VitsPreTrainedModel as VitsPreTrainedModel, __webpack_exports__VitsTokenizer as VitsTokenizer, __webpack_exports__Wav2Vec2BertForCTC as Wav2Vec2BertForCTC, __webpack_exports__Wav2Vec2BertForSequenceClassification as Wav2Vec2BertForSequenceClassification, __webpack_exports__Wav2Vec2BertModel as Wav2Vec2BertModel, __webpack_exports__Wav2Vec2BertPreTrainedModel as Wav2Vec2BertPreTrainedModel, __webpack_exports__Wav2Vec2CTCTokenizer as Wav2Vec2CTCTokenizer, __webpack_exports__Wav2Vec2FeatureExtractor as Wav2Vec2FeatureExtractor, __webpack_exports__Wav2Vec2ForAudioFrameClassification as Wav2Vec2ForAudioFrameClassification, __webpack_exports__Wav2Vec2ForCTC as Wav2Vec2ForCTC, __webpack_exports__Wav2Vec2ForSequenceClassification as Wav2Vec2ForSequenceClassification, __webpack_exports__Wav2Vec2Model as Wav2Vec2Model, __webpack_exports__Wav2Vec2PreTrainedModel as Wav2Vec2PreTrainedModel, __webpack_exports__Wav2Vec2Processor as Wav2Vec2Processor, __webpack_exports__Wav2Vec2ProcessorWithLM as Wav2Vec2ProcessorWithLM, __webpack_exports__WavLMForAudioFrameClassification as WavLMForAudioFrameClassification, __webpack_exports__WavLMForCTC as WavLMForCTC, __webpack_exports__WavLMForSequenceClassification as WavLMForSequenceClassification, __webpack_exports__WavLMForXVector as WavLMForXVector, __webpack_exports__WavLMModel as WavLMModel, __webpack_exports__WavLMPreTrainedModel as WavLMPreTrainedModel, __webpack_exports__WeSpeakerFeatureExtractor as WeSpeakerFeatureExtractor, __webpack_exports__WeSpeakerResNetModel as WeSpeakerResNetModel, __webpack_exports__WeSpeakerResNetPreTrainedModel as WeSpeakerResNetPreTrainedModel, __webpack_exports__WhisperFeatureExtractor as WhisperFeatureExtractor, __webpack_exports__WhisperForConditionalGeneration as WhisperForConditionalGeneration, __webpack_exports__WhisperModel as WhisperModel, __webpack_exports__WhisperPreTrainedModel as WhisperPreTrainedModel, __webpack_exports__WhisperProcessor as WhisperProcessor, __webpack_exports__WhisperTextStreamer as WhisperTextStreamer, __webpack_exports__WhisperTimeStampLogitsProcessor as WhisperTimeStampLogitsProcessor, __webpack_exports__WhisperTokenizer as WhisperTokenizer, __webpack_exports__XLMForQuestionAnswering as XLMForQuestionAnswering, __webpack_exports__XLMForSequenceClassification as XLMForSequenceClassification, __webpack_exports__XLMForTokenClassification as XLMForTokenClassification, __webpack_exports__XLMModel as XLMModel, __webpack_exports__XLMPreTrainedModel as XLMPreTrainedModel, __webpack_exports__XLMRobertaForMaskedLM as XLMRobertaForMaskedLM, __webpack_exports__XLMRobertaForQuestionAnswering as XLMRobertaForQuestionAnswering, __webpack_exports__XLMRobertaForSequenceClassification as XLMRobertaForSequenceClassification, __webpack_exports__XLMRobertaForTokenClassification as XLMRobertaForTokenClassification, __webpack_exports__XLMRobertaModel as XLMRobertaModel, __webpack_exports__XLMRobertaPreTrainedModel as XLMRobertaPreTrainedModel, __webpack_exports__XLMRobertaTokenizer as XLMRobertaTokenizer, __webpack_exports__XLMTokenizer as XLMTokenizer, __webpack_exports__XLMWithLMHeadModel as XLMWithLMHeadModel, __webpack_exports__XVectorOutput as XVectorOutput, __webpack_exports__YolosFeatureExtractor as YolosFeatureExtractor, __webpack_exports__YolosForObjectDetection as YolosForObjectDetection, __webpack_exports__YolosImageProcessor as YolosImageProcessor, __webpack_exports__YolosModel as YolosModel, __webpack_exports__YolosObjectDetectionOutput as YolosObjectDetectionOutput, __webpack_exports__YolosPreTrainedModel as YolosPreTrainedModel, __webpack_exports__ZeroShotAudioClassificationPipeline as ZeroShotAudioClassificationPipeline, __webpack_exports__ZeroShotClassificationPipeline as ZeroShotClassificationPipeline, __webpack_exports__ZeroShotImageClassificationPipeline as ZeroShotImageClassificationPipeline, __webpack_exports__ZeroShotObjectDetectionPipeline as ZeroShotObjectDetectionPipeline, __webpack_exports__bankers_round as bankers_round, __webpack_exports__cat as cat, __webpack_exports__cos_sim as cos_sim, __webpack_exports__dot as dot, __webpack_exports__dynamic_time_warping as dynamic_time_warping, __webpack_exports__env as env, __webpack_exports__full as full, __webpack_exports__full_like as full_like, __webpack_exports__getKeyValueShapes as getKeyValueShapes, __webpack_exports__hamming as hamming, __webpack_exports__hanning as hanning, __webpack_exports__interpolate as interpolate, __webpack_exports__interpolate_4d as interpolate_4d, __webpack_exports__interpolate_data as interpolate_data, __webpack_exports__is_chinese_char as is_chinese_char, __webpack_exports__layer_norm as layer_norm, __webpack_exports__load_image as load_image, __webpack_exports__load_video as load_video, __webpack_exports__log_softmax as log_softmax, __webpack_exports__magnitude as magnitude, __webpack_exports__matmul as matmul, __webpack_exports__max as max, __webpack_exports__mean as mean, __webpack_exports__mean_pooling as mean_pooling, __webpack_exports__medianFilter as medianFilter, __webpack_exports__mel_filter_bank as mel_filter_bank, __webpack_exports__min as min, __webpack_exports__ones as ones, __webpack_exports__ones_like as ones_like, __webpack_exports__permute as permute, __webpack_exports__permute_data as permute_data, __webpack_exports__pipeline as pipeline, __webpack_exports__quantize_embeddings as quantize_embeddings, __webpack_exports__rand as rand, __webpack_exports__read_audio as read_audio, __webpack_exports__rfft as rfft, __webpack_exports__round as round, __webpack_exports__slice as slice, __webpack_exports__softmax as softmax, __webpack_exports__spectrogram as spectrogram, __webpack_exports__stack as stack, __webpack_exports__std_mean as std_mean, __webpack_exports__topk as topk, __webpack_exports__window_function as window_function, __webpack_exports__zeros as zeros, __webpack_exports__zeros_like as zeros_like };
35858
+ export { __webpack_exports__ASTFeatureExtractor as ASTFeatureExtractor, __webpack_exports__ASTForAudioClassification as ASTForAudioClassification, __webpack_exports__ASTModel as ASTModel, __webpack_exports__ASTPreTrainedModel as ASTPreTrainedModel, __webpack_exports__AlbertForMaskedLM as AlbertForMaskedLM, __webpack_exports__AlbertForQuestionAnswering as AlbertForQuestionAnswering, __webpack_exports__AlbertForSequenceClassification as AlbertForSequenceClassification, __webpack_exports__AlbertModel as AlbertModel, __webpack_exports__AlbertPreTrainedModel as AlbertPreTrainedModel, __webpack_exports__AlbertTokenizer as AlbertTokenizer, __webpack_exports__AudioClassificationPipeline as AudioClassificationPipeline, __webpack_exports__AutoConfig as AutoConfig, __webpack_exports__AutoFeatureExtractor as AutoFeatureExtractor, __webpack_exports__AutoImageProcessor as AutoImageProcessor, __webpack_exports__AutoModel as AutoModel, __webpack_exports__AutoModelForAudioClassification as AutoModelForAudioClassification, __webpack_exports__AutoModelForAudioFrameClassification as AutoModelForAudioFrameClassification, __webpack_exports__AutoModelForAudioTextToText as AutoModelForAudioTextToText, __webpack_exports__AutoModelForCTC as AutoModelForCTC, __webpack_exports__AutoModelForCausalLM as AutoModelForCausalLM, __webpack_exports__AutoModelForDepthEstimation as AutoModelForDepthEstimation, __webpack_exports__AutoModelForDocumentQuestionAnswering as AutoModelForDocumentQuestionAnswering, __webpack_exports__AutoModelForImageClassification as AutoModelForImageClassification, __webpack_exports__AutoModelForImageFeatureExtraction as AutoModelForImageFeatureExtraction, __webpack_exports__AutoModelForImageMatting as AutoModelForImageMatting, __webpack_exports__AutoModelForImageSegmentation as AutoModelForImageSegmentation, __webpack_exports__AutoModelForImageTextToText as AutoModelForImageTextToText, __webpack_exports__AutoModelForImageToImage as AutoModelForImageToImage, __webpack_exports__AutoModelForMaskGeneration as AutoModelForMaskGeneration, __webpack_exports__AutoModelForMaskedLM as AutoModelForMaskedLM, __webpack_exports__AutoModelForNormalEstimation as AutoModelForNormalEstimation, __webpack_exports__AutoModelForObjectDetection as AutoModelForObjectDetection, __webpack_exports__AutoModelForPoseEstimation as AutoModelForPoseEstimation, __webpack_exports__AutoModelForQuestionAnswering as AutoModelForQuestionAnswering, __webpack_exports__AutoModelForSemanticSegmentation as AutoModelForSemanticSegmentation, __webpack_exports__AutoModelForSeq2SeqLM as AutoModelForSeq2SeqLM, __webpack_exports__AutoModelForSequenceClassification as AutoModelForSequenceClassification, __webpack_exports__AutoModelForSpeechSeq2Seq as AutoModelForSpeechSeq2Seq, __webpack_exports__AutoModelForTextToSpectrogram as AutoModelForTextToSpectrogram, __webpack_exports__AutoModelForTextToWaveform as AutoModelForTextToWaveform, __webpack_exports__AutoModelForTokenClassification as AutoModelForTokenClassification, __webpack_exports__AutoModelForUniversalSegmentation as AutoModelForUniversalSegmentation, __webpack_exports__AutoModelForVision2Seq as AutoModelForVision2Seq, __webpack_exports__AutoModelForXVector as AutoModelForXVector, __webpack_exports__AutoModelForZeroShotObjectDetection as AutoModelForZeroShotObjectDetection, __webpack_exports__AutoProcessor as AutoProcessor, __webpack_exports__AutoTokenizer as AutoTokenizer, __webpack_exports__AutomaticSpeechRecognitionPipeline as AutomaticSpeechRecognitionPipeline, __webpack_exports__BackgroundRemovalPipeline as BackgroundRemovalPipeline, __webpack_exports__BartForConditionalGeneration as BartForConditionalGeneration, __webpack_exports__BartForSequenceClassification as BartForSequenceClassification, __webpack_exports__BartModel as BartModel, __webpack_exports__BartPretrainedModel as BartPretrainedModel, __webpack_exports__BartTokenizer as BartTokenizer, __webpack_exports__BaseModelOutput as BaseModelOutput, __webpack_exports__BaseStreamer as BaseStreamer, __webpack_exports__BeitFeatureExtractor as BeitFeatureExtractor, __webpack_exports__BeitForImageClassification as BeitForImageClassification, __webpack_exports__BeitModel as BeitModel, __webpack_exports__BeitPreTrainedModel as BeitPreTrainedModel, __webpack_exports__BertForMaskedLM as BertForMaskedLM, __webpack_exports__BertForQuestionAnswering as BertForQuestionAnswering, __webpack_exports__BertForSequenceClassification as BertForSequenceClassification, __webpack_exports__BertForTokenClassification as BertForTokenClassification, __webpack_exports__BertModel as BertModel, __webpack_exports__BertPreTrainedModel as BertPreTrainedModel, __webpack_exports__BertTokenizer as BertTokenizer, __webpack_exports__BitImageProcessor as BitImageProcessor, __webpack_exports__BlenderbotForConditionalGeneration as BlenderbotForConditionalGeneration, __webpack_exports__BlenderbotModel as BlenderbotModel, __webpack_exports__BlenderbotPreTrainedModel as BlenderbotPreTrainedModel, __webpack_exports__BlenderbotSmallForConditionalGeneration as BlenderbotSmallForConditionalGeneration, __webpack_exports__BlenderbotSmallModel as BlenderbotSmallModel, __webpack_exports__BlenderbotSmallPreTrainedModel as BlenderbotSmallPreTrainedModel, __webpack_exports__BlenderbotSmallTokenizer as BlenderbotSmallTokenizer, __webpack_exports__BlenderbotTokenizer as BlenderbotTokenizer, __webpack_exports__BloomForCausalLM as BloomForCausalLM, __webpack_exports__BloomModel as BloomModel, __webpack_exports__BloomPreTrainedModel as BloomPreTrainedModel, __webpack_exports__BloomTokenizer as BloomTokenizer, __webpack_exports__CLIPFeatureExtractor as CLIPFeatureExtractor, __webpack_exports__CLIPImageProcessor as CLIPImageProcessor, __webpack_exports__CLIPModel as CLIPModel, __webpack_exports__CLIPPreTrainedModel as CLIPPreTrainedModel, __webpack_exports__CLIPSegForImageSegmentation as CLIPSegForImageSegmentation, __webpack_exports__CLIPSegModel as CLIPSegModel, __webpack_exports__CLIPSegPreTrainedModel as CLIPSegPreTrainedModel, __webpack_exports__CLIPTextModel as CLIPTextModel, __webpack_exports__CLIPTextModelWithProjection as CLIPTextModelWithProjection, __webpack_exports__CLIPTokenizer as CLIPTokenizer, __webpack_exports__CLIPVisionModel as CLIPVisionModel, __webpack_exports__CLIPVisionModelWithProjection as CLIPVisionModelWithProjection, __webpack_exports__CamembertForMaskedLM as CamembertForMaskedLM, __webpack_exports__CamembertForQuestionAnswering as CamembertForQuestionAnswering, __webpack_exports__CamembertForSequenceClassification as CamembertForSequenceClassification, __webpack_exports__CamembertForTokenClassification as CamembertForTokenClassification, __webpack_exports__CamembertModel as CamembertModel, __webpack_exports__CamembertPreTrainedModel as CamembertPreTrainedModel, __webpack_exports__CamembertTokenizer as CamembertTokenizer, __webpack_exports__CausalLMOutput as CausalLMOutput, __webpack_exports__CausalLMOutputWithPast as CausalLMOutputWithPast, __webpack_exports__ChineseCLIPFeatureExtractor as ChineseCLIPFeatureExtractor, __webpack_exports__ChineseCLIPModel as ChineseCLIPModel, __webpack_exports__ChineseCLIPPreTrainedModel as ChineseCLIPPreTrainedModel, __webpack_exports__ClapAudioModelWithProjection as ClapAudioModelWithProjection, __webpack_exports__ClapFeatureExtractor as ClapFeatureExtractor, __webpack_exports__ClapModel as ClapModel, __webpack_exports__ClapPreTrainedModel as ClapPreTrainedModel, __webpack_exports__ClapTextModelWithProjection as ClapTextModelWithProjection, __webpack_exports__ClassifierFreeGuidanceLogitsProcessor as ClassifierFreeGuidanceLogitsProcessor, __webpack_exports__CodeGenForCausalLM as CodeGenForCausalLM, __webpack_exports__CodeGenModel as CodeGenModel, __webpack_exports__CodeGenPreTrainedModel as CodeGenPreTrainedModel, __webpack_exports__CodeGenTokenizer as CodeGenTokenizer, __webpack_exports__CodeLlamaTokenizer as CodeLlamaTokenizer, __webpack_exports__CohereForCausalLM as CohereForCausalLM, __webpack_exports__CohereModel as CohereModel, __webpack_exports__CoherePreTrainedModel as CoherePreTrainedModel, __webpack_exports__CohereTokenizer as CohereTokenizer, __webpack_exports__ConvBertForMaskedLM as ConvBertForMaskedLM, __webpack_exports__ConvBertForQuestionAnswering as ConvBertForQuestionAnswering, __webpack_exports__ConvBertForSequenceClassification as ConvBertForSequenceClassification, __webpack_exports__ConvBertForTokenClassification as ConvBertForTokenClassification, __webpack_exports__ConvBertModel as ConvBertModel, __webpack_exports__ConvBertPreTrainedModel as ConvBertPreTrainedModel, __webpack_exports__ConvBertTokenizer as ConvBertTokenizer, __webpack_exports__ConvNextFeatureExtractor as ConvNextFeatureExtractor, __webpack_exports__ConvNextForImageClassification as ConvNextForImageClassification, __webpack_exports__ConvNextImageProcessor as ConvNextImageProcessor, __webpack_exports__ConvNextModel as ConvNextModel, __webpack_exports__ConvNextPreTrainedModel as ConvNextPreTrainedModel, __webpack_exports__ConvNextV2ForImageClassification as ConvNextV2ForImageClassification, __webpack_exports__ConvNextV2Model as ConvNextV2Model, __webpack_exports__ConvNextV2PreTrainedModel as ConvNextV2PreTrainedModel, __webpack_exports__DPTFeatureExtractor as DPTFeatureExtractor, __webpack_exports__DPTForDepthEstimation as DPTForDepthEstimation, __webpack_exports__DPTImageProcessor as DPTImageProcessor, __webpack_exports__DPTModel as DPTModel, __webpack_exports__DPTPreTrainedModel as DPTPreTrainedModel, __webpack_exports__DacDecoderModel as DacDecoderModel, __webpack_exports__DacDecoderOutput as DacDecoderOutput, __webpack_exports__DacEncoderModel as DacEncoderModel, __webpack_exports__DacEncoderOutput as DacEncoderOutput, __webpack_exports__DacFeatureExtractor as DacFeatureExtractor, __webpack_exports__DacModel as DacModel, __webpack_exports__DacPreTrainedModel as DacPreTrainedModel, __webpack_exports__DataTypeMap as DataTypeMap, __webpack_exports__DebertaForMaskedLM as DebertaForMaskedLM, __webpack_exports__DebertaForQuestionAnswering as DebertaForQuestionAnswering, __webpack_exports__DebertaForSequenceClassification as DebertaForSequenceClassification, __webpack_exports__DebertaForTokenClassification as DebertaForTokenClassification, __webpack_exports__DebertaModel as DebertaModel, __webpack_exports__DebertaPreTrainedModel as DebertaPreTrainedModel, __webpack_exports__DebertaTokenizer as DebertaTokenizer, __webpack_exports__DebertaV2ForMaskedLM as DebertaV2ForMaskedLM, __webpack_exports__DebertaV2ForQuestionAnswering as DebertaV2ForQuestionAnswering, __webpack_exports__DebertaV2ForSequenceClassification as DebertaV2ForSequenceClassification, __webpack_exports__DebertaV2ForTokenClassification as DebertaV2ForTokenClassification, __webpack_exports__DebertaV2Model as DebertaV2Model, __webpack_exports__DebertaV2PreTrainedModel as DebertaV2PreTrainedModel, __webpack_exports__DebertaV2Tokenizer as DebertaV2Tokenizer, __webpack_exports__DecisionTransformerModel as DecisionTransformerModel, __webpack_exports__DecisionTransformerPreTrainedModel as DecisionTransformerPreTrainedModel, __webpack_exports__DeiTFeatureExtractor as DeiTFeatureExtractor, __webpack_exports__DeiTForImageClassification as DeiTForImageClassification, __webpack_exports__DeiTImageProcessor as DeiTImageProcessor, __webpack_exports__DeiTModel as DeiTModel, __webpack_exports__DeiTPreTrainedModel as DeiTPreTrainedModel, __webpack_exports__DepthAnythingForDepthEstimation as DepthAnythingForDepthEstimation, __webpack_exports__DepthAnythingPreTrainedModel as DepthAnythingPreTrainedModel, __webpack_exports__DepthEstimationPipeline as DepthEstimationPipeline, __webpack_exports__DepthProForDepthEstimation as DepthProForDepthEstimation, __webpack_exports__DepthProPreTrainedModel as DepthProPreTrainedModel, __webpack_exports__DetrFeatureExtractor as DetrFeatureExtractor, __webpack_exports__DetrForObjectDetection as DetrForObjectDetection, __webpack_exports__DetrForSegmentation as DetrForSegmentation, __webpack_exports__DetrImageProcessor as DetrImageProcessor, __webpack_exports__DetrModel as DetrModel, __webpack_exports__DetrObjectDetectionOutput as DetrObjectDetectionOutput, __webpack_exports__DetrPreTrainedModel as DetrPreTrainedModel, __webpack_exports__DetrSegmentationOutput as DetrSegmentationOutput, __webpack_exports__Dinov2ForImageClassification as Dinov2ForImageClassification, __webpack_exports__Dinov2Model as Dinov2Model, __webpack_exports__Dinov2PreTrainedModel as Dinov2PreTrainedModel, __webpack_exports__Dinov2WithRegistersForImageClassification as Dinov2WithRegistersForImageClassification, __webpack_exports__Dinov2WithRegistersModel as Dinov2WithRegistersModel, __webpack_exports__Dinov2WithRegistersPreTrainedModel as Dinov2WithRegistersPreTrainedModel, __webpack_exports__DistilBertForMaskedLM as DistilBertForMaskedLM, __webpack_exports__DistilBertForQuestionAnswering as DistilBertForQuestionAnswering, __webpack_exports__DistilBertForSequenceClassification as DistilBertForSequenceClassification, __webpack_exports__DistilBertForTokenClassification as DistilBertForTokenClassification, __webpack_exports__DistilBertModel as DistilBertModel, __webpack_exports__DistilBertPreTrainedModel as DistilBertPreTrainedModel, __webpack_exports__DistilBertTokenizer as DistilBertTokenizer, __webpack_exports__DocumentQuestionAnsweringPipeline as DocumentQuestionAnsweringPipeline, __webpack_exports__DonutFeatureExtractor as DonutFeatureExtractor, __webpack_exports__DonutImageProcessor as DonutImageProcessor, __webpack_exports__DonutSwinModel as DonutSwinModel, __webpack_exports__DonutSwinPreTrainedModel as DonutSwinPreTrainedModel, __webpack_exports__EfficientNetForImageClassification as EfficientNetForImageClassification, __webpack_exports__EfficientNetImageProcessor as EfficientNetImageProcessor, __webpack_exports__EfficientNetModel as EfficientNetModel, __webpack_exports__EfficientNetPreTrainedModel as EfficientNetPreTrainedModel, __webpack_exports__ElectraForMaskedLM as ElectraForMaskedLM, __webpack_exports__ElectraForQuestionAnswering as ElectraForQuestionAnswering, __webpack_exports__ElectraForSequenceClassification as ElectraForSequenceClassification, __webpack_exports__ElectraForTokenClassification as ElectraForTokenClassification, __webpack_exports__ElectraModel as ElectraModel, __webpack_exports__ElectraPreTrainedModel as ElectraPreTrainedModel, __webpack_exports__ElectraTokenizer as ElectraTokenizer, __webpack_exports__EncodecFeatureExtractor as EncodecFeatureExtractor, __webpack_exports__EosTokenCriteria as EosTokenCriteria, __webpack_exports__EsmForMaskedLM as EsmForMaskedLM, __webpack_exports__EsmForSequenceClassification as EsmForSequenceClassification, __webpack_exports__EsmForTokenClassification as EsmForTokenClassification, __webpack_exports__EsmModel as EsmModel, __webpack_exports__EsmPreTrainedModel as EsmPreTrainedModel, __webpack_exports__EsmTokenizer as EsmTokenizer, __webpack_exports__ExaoneForCausalLM as ExaoneForCausalLM, __webpack_exports__ExaoneModel as ExaoneModel, __webpack_exports__ExaonePreTrainedModel as ExaonePreTrainedModel, __webpack_exports__FFT as FFT, __webpack_exports__FalconForCausalLM as FalconForCausalLM, __webpack_exports__FalconModel as FalconModel, __webpack_exports__FalconPreTrainedModel as FalconPreTrainedModel, __webpack_exports__FalconTokenizer as FalconTokenizer, __webpack_exports__FastViTForImageClassification as FastViTForImageClassification, __webpack_exports__FastViTModel as FastViTModel, __webpack_exports__FastViTPreTrainedModel as FastViTPreTrainedModel, __webpack_exports__FeatureExtractionPipeline as FeatureExtractionPipeline, __webpack_exports__FeatureExtractor as FeatureExtractor, __webpack_exports__FillMaskPipeline as FillMaskPipeline, __webpack_exports__Florence2ForConditionalGeneration as Florence2ForConditionalGeneration, __webpack_exports__Florence2PreTrainedModel as Florence2PreTrainedModel, __webpack_exports__Florence2Processor as Florence2Processor, __webpack_exports__ForcedBOSTokenLogitsProcessor as ForcedBOSTokenLogitsProcessor, __webpack_exports__ForcedEOSTokenLogitsProcessor as ForcedEOSTokenLogitsProcessor, __webpack_exports__GLPNFeatureExtractor as GLPNFeatureExtractor, __webpack_exports__GLPNForDepthEstimation as GLPNForDepthEstimation, __webpack_exports__GLPNModel as GLPNModel, __webpack_exports__GLPNPreTrainedModel as GLPNPreTrainedModel, __webpack_exports__GPT2LMHeadModel as GPT2LMHeadModel, __webpack_exports__GPT2Model as GPT2Model, __webpack_exports__GPT2PreTrainedModel as GPT2PreTrainedModel, __webpack_exports__GPT2Tokenizer as GPT2Tokenizer, __webpack_exports__GPTBigCodeForCausalLM as GPTBigCodeForCausalLM, __webpack_exports__GPTBigCodeModel as GPTBigCodeModel, __webpack_exports__GPTBigCodePreTrainedModel as GPTBigCodePreTrainedModel, __webpack_exports__GPTJForCausalLM as GPTJForCausalLM, __webpack_exports__GPTJModel as GPTJModel, __webpack_exports__GPTJPreTrainedModel as GPTJPreTrainedModel, __webpack_exports__GPTNeoForCausalLM as GPTNeoForCausalLM, __webpack_exports__GPTNeoModel as GPTNeoModel, __webpack_exports__GPTNeoPreTrainedModel as GPTNeoPreTrainedModel, __webpack_exports__GPTNeoXForCausalLM as GPTNeoXForCausalLM, __webpack_exports__GPTNeoXModel as GPTNeoXModel, __webpack_exports__GPTNeoXPreTrainedModel as GPTNeoXPreTrainedModel, __webpack_exports__GPTNeoXTokenizer as GPTNeoXTokenizer, __webpack_exports__Gemma2ForCausalLM as Gemma2ForCausalLM, __webpack_exports__Gemma2Model as Gemma2Model, __webpack_exports__Gemma2PreTrainedModel as Gemma2PreTrainedModel, __webpack_exports__Gemma3ForCausalLM as Gemma3ForCausalLM, __webpack_exports__Gemma3Model as Gemma3Model, __webpack_exports__Gemma3PreTrainedModel as Gemma3PreTrainedModel, __webpack_exports__GemmaForCausalLM as GemmaForCausalLM, __webpack_exports__GemmaModel as GemmaModel, __webpack_exports__GemmaPreTrainedModel as GemmaPreTrainedModel, __webpack_exports__GemmaTokenizer as GemmaTokenizer, __webpack_exports__GlmForCausalLM as GlmForCausalLM, __webpack_exports__GlmModel as GlmModel, __webpack_exports__GlmPreTrainedModel as GlmPreTrainedModel, __webpack_exports__GraniteForCausalLM as GraniteForCausalLM, __webpack_exports__GraniteModel as GraniteModel, __webpack_exports__GranitePreTrainedModel as GranitePreTrainedModel, __webpack_exports__Grok1Tokenizer as Grok1Tokenizer, __webpack_exports__GroundingDinoForObjectDetection as GroundingDinoForObjectDetection, __webpack_exports__GroundingDinoImageProcessor as GroundingDinoImageProcessor, __webpack_exports__GroundingDinoPreTrainedModel as GroundingDinoPreTrainedModel, __webpack_exports__GroundingDinoProcessor as GroundingDinoProcessor, __webpack_exports__GroupViTModel as GroupViTModel, __webpack_exports__GroupViTPreTrainedModel as GroupViTPreTrainedModel, __webpack_exports__HeliumForCausalLM as HeliumForCausalLM, __webpack_exports__HeliumModel as HeliumModel, __webpack_exports__HeliumPreTrainedModel as HeliumPreTrainedModel, __webpack_exports__HerbertTokenizer as HerbertTokenizer, __webpack_exports__HieraForImageClassification as HieraForImageClassification, __webpack_exports__HieraModel as HieraModel, __webpack_exports__HieraPreTrainedModel as HieraPreTrainedModel, __webpack_exports__HubertForCTC as HubertForCTC, __webpack_exports__HubertForSequenceClassification as HubertForSequenceClassification, __webpack_exports__HubertModel as HubertModel, __webpack_exports__HubertPreTrainedModel as HubertPreTrainedModel, __webpack_exports__IJepaForImageClassification as IJepaForImageClassification, __webpack_exports__IJepaModel as IJepaModel, __webpack_exports__IJepaPreTrainedModel as IJepaPreTrainedModel, __webpack_exports__Idefics3ForConditionalGeneration as Idefics3ForConditionalGeneration, __webpack_exports__Idefics3ImageProcessor as Idefics3ImageProcessor, __webpack_exports__Idefics3PreTrainedModel as Idefics3PreTrainedModel, __webpack_exports__Idefics3Processor as Idefics3Processor, __webpack_exports__ImageClassificationPipeline as ImageClassificationPipeline, __webpack_exports__ImageFeatureExtractionPipeline as ImageFeatureExtractionPipeline, __webpack_exports__ImageFeatureExtractor as ImageFeatureExtractor, __webpack_exports__ImageMattingOutput as ImageMattingOutput, __webpack_exports__ImageProcessor as ImageProcessor, __webpack_exports__ImageSegmentationPipeline as ImageSegmentationPipeline, __webpack_exports__ImageToImagePipeline as ImageToImagePipeline, __webpack_exports__ImageToTextPipeline as ImageToTextPipeline, __webpack_exports__InterruptableStoppingCriteria as InterruptableStoppingCriteria, __webpack_exports__JAISLMHeadModel as JAISLMHeadModel, __webpack_exports__JAISModel as JAISModel, __webpack_exports__JAISPreTrainedModel as JAISPreTrainedModel, __webpack_exports__JinaCLIPImageProcessor as JinaCLIPImageProcessor, __webpack_exports__JinaCLIPModel as JinaCLIPModel, __webpack_exports__JinaCLIPPreTrainedModel as JinaCLIPPreTrainedModel, __webpack_exports__JinaCLIPProcessor as JinaCLIPProcessor, __webpack_exports__JinaCLIPTextModel as JinaCLIPTextModel, __webpack_exports__JinaCLIPVisionModel as JinaCLIPVisionModel, __webpack_exports__LiteWhisperForConditionalGeneration as LiteWhisperForConditionalGeneration, __webpack_exports__LlamaForCausalLM as LlamaForCausalLM, __webpack_exports__LlamaModel as LlamaModel, __webpack_exports__LlamaPreTrainedModel as LlamaPreTrainedModel, __webpack_exports__LlamaTokenizer as LlamaTokenizer, __webpack_exports__LlavaForConditionalGeneration as LlavaForConditionalGeneration, __webpack_exports__LlavaOnevisionForConditionalGeneration as LlavaOnevisionForConditionalGeneration, __webpack_exports__LlavaOnevisionImageProcessor as LlavaOnevisionImageProcessor, __webpack_exports__LlavaPreTrainedModel as LlavaPreTrainedModel, __webpack_exports__LogitsProcessor as LogitsProcessor, __webpack_exports__LogitsProcessorList as LogitsProcessorList, __webpack_exports__LogitsWarper as LogitsWarper, __webpack_exports__LongT5ForConditionalGeneration as LongT5ForConditionalGeneration, __webpack_exports__LongT5Model as LongT5Model, __webpack_exports__LongT5PreTrainedModel as LongT5PreTrainedModel, __webpack_exports__M2M100ForConditionalGeneration as M2M100ForConditionalGeneration, __webpack_exports__M2M100Model as M2M100Model, __webpack_exports__M2M100PreTrainedModel as M2M100PreTrainedModel, __webpack_exports__M2M100Tokenizer as M2M100Tokenizer, __webpack_exports__MBart50Tokenizer as MBart50Tokenizer, __webpack_exports__MBartForCausalLM as MBartForCausalLM, __webpack_exports__MBartForConditionalGeneration as MBartForConditionalGeneration, __webpack_exports__MBartForSequenceClassification as MBartForSequenceClassification, __webpack_exports__MBartModel as MBartModel, __webpack_exports__MBartPreTrainedModel as MBartPreTrainedModel, __webpack_exports__MBartTokenizer as MBartTokenizer, __webpack_exports__MPNetForMaskedLM as MPNetForMaskedLM, __webpack_exports__MPNetForQuestionAnswering as MPNetForQuestionAnswering, __webpack_exports__MPNetForSequenceClassification as MPNetForSequenceClassification, __webpack_exports__MPNetForTokenClassification as MPNetForTokenClassification, __webpack_exports__MPNetModel as MPNetModel, __webpack_exports__MPNetPreTrainedModel as MPNetPreTrainedModel, __webpack_exports__MPNetTokenizer as MPNetTokenizer, __webpack_exports__MT5ForConditionalGeneration as MT5ForConditionalGeneration, __webpack_exports__MT5Model as MT5Model, __webpack_exports__MT5PreTrainedModel as MT5PreTrainedModel, __webpack_exports__MarianMTModel as MarianMTModel, __webpack_exports__MarianModel as MarianModel, __webpack_exports__MarianPreTrainedModel as MarianPreTrainedModel, __webpack_exports__MarianTokenizer as MarianTokenizer, __webpack_exports__Mask2FormerImageProcessor as Mask2FormerImageProcessor, __webpack_exports__MaskFormerFeatureExtractor as MaskFormerFeatureExtractor, __webpack_exports__MaskFormerForInstanceSegmentation as MaskFormerForInstanceSegmentation, __webpack_exports__MaskFormerImageProcessor as MaskFormerImageProcessor, __webpack_exports__MaskFormerModel as MaskFormerModel, __webpack_exports__MaskFormerPreTrainedModel as MaskFormerPreTrainedModel, __webpack_exports__MaskedLMOutput as MaskedLMOutput, __webpack_exports__MaxLengthCriteria as MaxLengthCriteria, __webpack_exports__Metric3DForDepthEstimation as Metric3DForDepthEstimation, __webpack_exports__Metric3DPreTrainedModel as Metric3DPreTrainedModel, __webpack_exports__Metric3Dv2ForDepthEstimation as Metric3Dv2ForDepthEstimation, __webpack_exports__Metric3Dv2PreTrainedModel as Metric3Dv2PreTrainedModel, __webpack_exports__MgpstrForSceneTextRecognition as MgpstrForSceneTextRecognition, __webpack_exports__MgpstrModelOutput as MgpstrModelOutput, __webpack_exports__MgpstrPreTrainedModel as MgpstrPreTrainedModel, __webpack_exports__MgpstrProcessor as MgpstrProcessor, __webpack_exports__MgpstrTokenizer as MgpstrTokenizer, __webpack_exports__MimiDecoderModel as MimiDecoderModel, __webpack_exports__MimiDecoderOutput as MimiDecoderOutput, __webpack_exports__MimiEncoderModel as MimiEncoderModel, __webpack_exports__MimiEncoderOutput as MimiEncoderOutput, __webpack_exports__MimiModel as MimiModel, __webpack_exports__MimiPreTrainedModel as MimiPreTrainedModel, __webpack_exports__MinLengthLogitsProcessor as MinLengthLogitsProcessor, __webpack_exports__MinNewTokensLengthLogitsProcessor as MinNewTokensLengthLogitsProcessor, __webpack_exports__MistralForCausalLM as MistralForCausalLM, __webpack_exports__MistralModel as MistralModel, __webpack_exports__MistralPreTrainedModel as MistralPreTrainedModel, __webpack_exports__MobileBertForMaskedLM as MobileBertForMaskedLM, __webpack_exports__MobileBertForQuestionAnswering as MobileBertForQuestionAnswering, __webpack_exports__MobileBertForSequenceClassification as MobileBertForSequenceClassification, __webpack_exports__MobileBertModel as MobileBertModel, __webpack_exports__MobileBertPreTrainedModel as MobileBertPreTrainedModel, __webpack_exports__MobileBertTokenizer as MobileBertTokenizer, __webpack_exports__MobileLLMForCausalLM as MobileLLMForCausalLM, __webpack_exports__MobileLLMModel as MobileLLMModel, __webpack_exports__MobileLLMPreTrainedModel as MobileLLMPreTrainedModel, __webpack_exports__MobileNetV1FeatureExtractor as MobileNetV1FeatureExtractor, __webpack_exports__MobileNetV1ForImageClassification as MobileNetV1ForImageClassification, __webpack_exports__MobileNetV1ForSemanticSegmentation as MobileNetV1ForSemanticSegmentation, __webpack_exports__MobileNetV1ImageProcessor as MobileNetV1ImageProcessor, __webpack_exports__MobileNetV1Model as MobileNetV1Model, __webpack_exports__MobileNetV1PreTrainedModel as MobileNetV1PreTrainedModel, __webpack_exports__MobileNetV2FeatureExtractor as MobileNetV2FeatureExtractor, __webpack_exports__MobileNetV2ForImageClassification as MobileNetV2ForImageClassification, __webpack_exports__MobileNetV2ForSemanticSegmentation as MobileNetV2ForSemanticSegmentation, __webpack_exports__MobileNetV2ImageProcessor as MobileNetV2ImageProcessor, __webpack_exports__MobileNetV2Model as MobileNetV2Model, __webpack_exports__MobileNetV2PreTrainedModel as MobileNetV2PreTrainedModel, __webpack_exports__MobileNetV3FeatureExtractor as MobileNetV3FeatureExtractor, __webpack_exports__MobileNetV3ForImageClassification as MobileNetV3ForImageClassification, __webpack_exports__MobileNetV3ForSemanticSegmentation as MobileNetV3ForSemanticSegmentation, __webpack_exports__MobileNetV3ImageProcessor as MobileNetV3ImageProcessor, __webpack_exports__MobileNetV3Model as MobileNetV3Model, __webpack_exports__MobileNetV3PreTrainedModel as MobileNetV3PreTrainedModel, __webpack_exports__MobileNetV4FeatureExtractor as MobileNetV4FeatureExtractor, __webpack_exports__MobileNetV4ForImageClassification as MobileNetV4ForImageClassification, __webpack_exports__MobileNetV4ForSemanticSegmentation as MobileNetV4ForSemanticSegmentation, __webpack_exports__MobileNetV4ImageProcessor as MobileNetV4ImageProcessor, __webpack_exports__MobileNetV4Model as MobileNetV4Model, __webpack_exports__MobileNetV4PreTrainedModel as MobileNetV4PreTrainedModel, __webpack_exports__MobileViTFeatureExtractor as MobileViTFeatureExtractor, __webpack_exports__MobileViTForImageClassification as MobileViTForImageClassification, __webpack_exports__MobileViTImageProcessor as MobileViTImageProcessor, __webpack_exports__MobileViTModel as MobileViTModel, __webpack_exports__MobileViTPreTrainedModel as MobileViTPreTrainedModel, __webpack_exports__MobileViTV2ForImageClassification as MobileViTV2ForImageClassification, __webpack_exports__MobileViTV2Model as MobileViTV2Model, __webpack_exports__MobileViTV2PreTrainedModel as MobileViTV2PreTrainedModel, __webpack_exports__ModelOutput as ModelOutput, __webpack_exports__ModernBertForMaskedLM as ModernBertForMaskedLM, __webpack_exports__ModernBertForSequenceClassification as ModernBertForSequenceClassification, __webpack_exports__ModernBertForTokenClassification as ModernBertForTokenClassification, __webpack_exports__ModernBertModel as ModernBertModel, __webpack_exports__ModernBertPreTrainedModel as ModernBertPreTrainedModel, __webpack_exports__Moondream1ForConditionalGeneration as Moondream1ForConditionalGeneration, __webpack_exports__MoonshineFeatureExtractor as MoonshineFeatureExtractor, __webpack_exports__MoonshineForConditionalGeneration as MoonshineForConditionalGeneration, __webpack_exports__MoonshineModel as MoonshineModel, __webpack_exports__MoonshinePreTrainedModel as MoonshinePreTrainedModel, __webpack_exports__MoonshineProcessor as MoonshineProcessor, __webpack_exports__MptForCausalLM as MptForCausalLM, __webpack_exports__MptModel as MptModel, __webpack_exports__MptPreTrainedModel as MptPreTrainedModel, __webpack_exports__MultiModalityCausalLM as MultiModalityCausalLM, __webpack_exports__MultiModalityPreTrainedModel as MultiModalityPreTrainedModel, __webpack_exports__MusicgenForCausalLM as MusicgenForCausalLM, __webpack_exports__MusicgenForConditionalGeneration as MusicgenForConditionalGeneration, __webpack_exports__MusicgenModel as MusicgenModel, __webpack_exports__MusicgenPreTrainedModel as MusicgenPreTrainedModel, __webpack_exports__NllbTokenizer as NllbTokenizer, __webpack_exports__NoBadWordsLogitsProcessor as NoBadWordsLogitsProcessor, __webpack_exports__NoRepeatNGramLogitsProcessor as NoRepeatNGramLogitsProcessor, __webpack_exports__NomicBertModel as NomicBertModel, __webpack_exports__NomicBertPreTrainedModel as NomicBertPreTrainedModel, __webpack_exports__NougatImageProcessor as NougatImageProcessor, __webpack_exports__NougatTokenizer as NougatTokenizer, __webpack_exports__OPTForCausalLM as OPTForCausalLM, __webpack_exports__OPTModel as OPTModel, __webpack_exports__OPTPreTrainedModel as OPTPreTrainedModel, __webpack_exports__ObjectDetectionPipeline as ObjectDetectionPipeline, __webpack_exports__Olmo2ForCausalLM as Olmo2ForCausalLM, __webpack_exports__Olmo2Model as Olmo2Model, __webpack_exports__Olmo2PreTrainedModel as Olmo2PreTrainedModel, __webpack_exports__OlmoForCausalLM as OlmoForCausalLM, __webpack_exports__OlmoModel as OlmoModel, __webpack_exports__OlmoPreTrainedModel as OlmoPreTrainedModel, __webpack_exports__OpenELMForCausalLM as OpenELMForCausalLM, __webpack_exports__OpenELMModel as OpenELMModel, __webpack_exports__OpenELMPreTrainedModel as OpenELMPreTrainedModel, __webpack_exports__OwlViTFeatureExtractor as OwlViTFeatureExtractor, __webpack_exports__OwlViTForObjectDetection as OwlViTForObjectDetection, __webpack_exports__OwlViTImageProcessor as OwlViTImageProcessor, __webpack_exports__OwlViTModel as OwlViTModel, __webpack_exports__OwlViTPreTrainedModel as OwlViTPreTrainedModel, __webpack_exports__OwlViTProcessor as OwlViTProcessor, __webpack_exports__Owlv2ForObjectDetection as Owlv2ForObjectDetection, __webpack_exports__Owlv2ImageProcessor as Owlv2ImageProcessor, __webpack_exports__Owlv2Model as Owlv2Model, __webpack_exports__Owlv2PreTrainedModel as Owlv2PreTrainedModel, __webpack_exports__PaliGemmaForConditionalGeneration as PaliGemmaForConditionalGeneration, __webpack_exports__PaliGemmaPreTrainedModel as PaliGemmaPreTrainedModel, __webpack_exports__PaliGemmaProcessor as PaliGemmaProcessor, __webpack_exports__PatchTSMixerForPrediction as PatchTSMixerForPrediction, __webpack_exports__PatchTSMixerModel as PatchTSMixerModel, __webpack_exports__PatchTSMixerPreTrainedModel as PatchTSMixerPreTrainedModel, __webpack_exports__PatchTSTForPrediction as PatchTSTForPrediction, __webpack_exports__PatchTSTModel as PatchTSTModel, __webpack_exports__PatchTSTPreTrainedModel as PatchTSTPreTrainedModel, __webpack_exports__Phi3ForCausalLM as Phi3ForCausalLM, __webpack_exports__Phi3Model as Phi3Model, __webpack_exports__Phi3PreTrainedModel as Phi3PreTrainedModel, __webpack_exports__Phi3VForCausalLM as Phi3VForCausalLM, __webpack_exports__Phi3VImageProcessor as Phi3VImageProcessor, __webpack_exports__Phi3VPreTrainedModel as Phi3VPreTrainedModel, __webpack_exports__Phi3VProcessor as Phi3VProcessor, __webpack_exports__PhiForCausalLM as PhiForCausalLM, __webpack_exports__PhiModel as PhiModel, __webpack_exports__PhiPreTrainedModel as PhiPreTrainedModel, __webpack_exports__Pipeline as Pipeline, __webpack_exports__PreTrainedModel as PreTrainedModel, __webpack_exports__PreTrainedTokenizer as PreTrainedTokenizer, __webpack_exports__PretrainedConfig as PretrainedConfig, __webpack_exports__PretrainedMixin as PretrainedMixin, __webpack_exports__Processor as Processor, __webpack_exports__PvtForImageClassification as PvtForImageClassification, __webpack_exports__PvtImageProcessor as PvtImageProcessor, __webpack_exports__PvtModel as PvtModel, __webpack_exports__PvtPreTrainedModel as PvtPreTrainedModel, __webpack_exports__PyAnnoteFeatureExtractor as PyAnnoteFeatureExtractor, __webpack_exports__PyAnnoteForAudioFrameClassification as PyAnnoteForAudioFrameClassification, __webpack_exports__PyAnnoteModel as PyAnnoteModel, __webpack_exports__PyAnnotePreTrainedModel as PyAnnotePreTrainedModel, __webpack_exports__PyAnnoteProcessor as PyAnnoteProcessor, __webpack_exports__QuestionAnsweringModelOutput as QuestionAnsweringModelOutput, __webpack_exports__QuestionAnsweringPipeline as QuestionAnsweringPipeline, __webpack_exports__Qwen2ForCausalLM as Qwen2ForCausalLM, __webpack_exports__Qwen2Model as Qwen2Model, __webpack_exports__Qwen2PreTrainedModel as Qwen2PreTrainedModel, __webpack_exports__Qwen2Tokenizer as Qwen2Tokenizer, __webpack_exports__Qwen2VLForConditionalGeneration as Qwen2VLForConditionalGeneration, __webpack_exports__Qwen2VLImageProcessor as Qwen2VLImageProcessor, __webpack_exports__Qwen2VLPreTrainedModel as Qwen2VLPreTrainedModel, __webpack_exports__Qwen2VLProcessor as Qwen2VLProcessor, __webpack_exports__RTDetrForObjectDetection as RTDetrForObjectDetection, __webpack_exports__RTDetrImageProcessor as RTDetrImageProcessor, __webpack_exports__RTDetrModel as RTDetrModel, __webpack_exports__RTDetrObjectDetectionOutput as RTDetrObjectDetectionOutput, __webpack_exports__RTDetrPreTrainedModel as RTDetrPreTrainedModel, __webpack_exports__RawAudio as RawAudio, __webpack_exports__RawImage as RawImage, __webpack_exports__RawVideo as RawVideo, __webpack_exports__RawVideoFrame as RawVideoFrame, __webpack_exports__RepetitionPenaltyLogitsProcessor as RepetitionPenaltyLogitsProcessor, __webpack_exports__ResNetForImageClassification as ResNetForImageClassification, __webpack_exports__ResNetModel as ResNetModel, __webpack_exports__ResNetPreTrainedModel as ResNetPreTrainedModel, __webpack_exports__RoFormerForMaskedLM as RoFormerForMaskedLM, __webpack_exports__RoFormerForQuestionAnswering as RoFormerForQuestionAnswering, __webpack_exports__RoFormerForSequenceClassification as RoFormerForSequenceClassification, __webpack_exports__RoFormerForTokenClassification as RoFormerForTokenClassification, __webpack_exports__RoFormerModel as RoFormerModel, __webpack_exports__RoFormerPreTrainedModel as RoFormerPreTrainedModel, __webpack_exports__RoFormerTokenizer as RoFormerTokenizer, __webpack_exports__RobertaForMaskedLM as RobertaForMaskedLM, __webpack_exports__RobertaForQuestionAnswering as RobertaForQuestionAnswering, __webpack_exports__RobertaForSequenceClassification as RobertaForSequenceClassification, __webpack_exports__RobertaForTokenClassification as RobertaForTokenClassification, __webpack_exports__RobertaModel as RobertaModel, __webpack_exports__RobertaPreTrainedModel as RobertaPreTrainedModel, __webpack_exports__RobertaTokenizer as RobertaTokenizer, __webpack_exports__SamImageProcessor as SamImageProcessor, __webpack_exports__SamImageSegmentationOutput as SamImageSegmentationOutput, __webpack_exports__SamModel as SamModel, __webpack_exports__SamPreTrainedModel as SamPreTrainedModel, __webpack_exports__SamProcessor as SamProcessor, __webpack_exports__SapiensForDepthEstimation as SapiensForDepthEstimation, __webpack_exports__SapiensForNormalEstimation as SapiensForNormalEstimation, __webpack_exports__SapiensForSemanticSegmentation as SapiensForSemanticSegmentation, __webpack_exports__SapiensPreTrainedModel as SapiensPreTrainedModel, __webpack_exports__SeamlessM4TFeatureExtractor as SeamlessM4TFeatureExtractor, __webpack_exports__SegformerFeatureExtractor as SegformerFeatureExtractor, __webpack_exports__SegformerForImageClassification as SegformerForImageClassification, __webpack_exports__SegformerForSemanticSegmentation as SegformerForSemanticSegmentation, __webpack_exports__SegformerImageProcessor as SegformerImageProcessor, __webpack_exports__SegformerModel as SegformerModel, __webpack_exports__SegformerPreTrainedModel as SegformerPreTrainedModel, __webpack_exports__Seq2SeqLMOutput as Seq2SeqLMOutput, __webpack_exports__SequenceClassifierOutput as SequenceClassifierOutput, __webpack_exports__SiglipImageProcessor as SiglipImageProcessor, __webpack_exports__SiglipModel as SiglipModel, __webpack_exports__SiglipPreTrainedModel as SiglipPreTrainedModel, __webpack_exports__SiglipTextModel as SiglipTextModel, __webpack_exports__SiglipTokenizer as SiglipTokenizer, __webpack_exports__SiglipVisionModel as SiglipVisionModel, __webpack_exports__SmolVLMForConditionalGeneration as SmolVLMForConditionalGeneration, __webpack_exports__SmolVLMImageProcessor as SmolVLMImageProcessor, __webpack_exports__SmolVLMProcessor as SmolVLMProcessor, __webpack_exports__SnacDecoderModel as SnacDecoderModel, __webpack_exports__SnacEncoderModel as SnacEncoderModel, __webpack_exports__SnacFeatureExtractor as SnacFeatureExtractor, __webpack_exports__SnacModel as SnacModel, __webpack_exports__SnacPreTrainedModel as SnacPreTrainedModel, __webpack_exports__SpeechT5FeatureExtractor as SpeechT5FeatureExtractor, __webpack_exports__SpeechT5ForSpeechToText as SpeechT5ForSpeechToText, __webpack_exports__SpeechT5ForTextToSpeech as SpeechT5ForTextToSpeech, __webpack_exports__SpeechT5HifiGan as SpeechT5HifiGan, __webpack_exports__SpeechT5Model as SpeechT5Model, __webpack_exports__SpeechT5PreTrainedModel as SpeechT5PreTrainedModel, __webpack_exports__SpeechT5Processor as SpeechT5Processor, __webpack_exports__SpeechT5Tokenizer as SpeechT5Tokenizer, __webpack_exports__SqueezeBertForMaskedLM as SqueezeBertForMaskedLM, __webpack_exports__SqueezeBertForQuestionAnswering as SqueezeBertForQuestionAnswering, __webpack_exports__SqueezeBertForSequenceClassification as SqueezeBertForSequenceClassification, __webpack_exports__SqueezeBertModel as SqueezeBertModel, __webpack_exports__SqueezeBertPreTrainedModel as SqueezeBertPreTrainedModel, __webpack_exports__SqueezeBertTokenizer as SqueezeBertTokenizer, __webpack_exports__StableLmForCausalLM as StableLmForCausalLM, __webpack_exports__StableLmModel as StableLmModel, __webpack_exports__StableLmPreTrainedModel as StableLmPreTrainedModel, __webpack_exports__Starcoder2ForCausalLM as Starcoder2ForCausalLM, __webpack_exports__Starcoder2Model as Starcoder2Model, __webpack_exports__Starcoder2PreTrainedModel as Starcoder2PreTrainedModel, __webpack_exports__StoppingCriteria as StoppingCriteria, __webpack_exports__StoppingCriteriaList as StoppingCriteriaList, __webpack_exports__StyleTextToSpeech2Model as StyleTextToSpeech2Model, __webpack_exports__StyleTextToSpeech2PreTrainedModel as StyleTextToSpeech2PreTrainedModel, __webpack_exports__SummarizationPipeline as SummarizationPipeline, __webpack_exports__SuppressTokensAtBeginLogitsProcessor as SuppressTokensAtBeginLogitsProcessor, __webpack_exports__Swin2SRForImageSuperResolution as Swin2SRForImageSuperResolution, __webpack_exports__Swin2SRImageProcessor as Swin2SRImageProcessor, __webpack_exports__Swin2SRModel as Swin2SRModel, __webpack_exports__Swin2SRPreTrainedModel as Swin2SRPreTrainedModel, __webpack_exports__SwinForImageClassification as SwinForImageClassification, __webpack_exports__SwinForSemanticSegmentation as SwinForSemanticSegmentation, __webpack_exports__SwinModel as SwinModel, __webpack_exports__SwinPreTrainedModel as SwinPreTrainedModel, __webpack_exports__T5ForConditionalGeneration as T5ForConditionalGeneration, __webpack_exports__T5Model as T5Model, __webpack_exports__T5PreTrainedModel as T5PreTrainedModel, __webpack_exports__T5Tokenizer as T5Tokenizer, __webpack_exports__TableTransformerForObjectDetection as TableTransformerForObjectDetection, __webpack_exports__TableTransformerModel as TableTransformerModel, __webpack_exports__TableTransformerObjectDetectionOutput as TableTransformerObjectDetectionOutput, __webpack_exports__TableTransformerPreTrainedModel as TableTransformerPreTrainedModel, __webpack_exports__TemperatureLogitsWarper as TemperatureLogitsWarper, __webpack_exports__Tensor as Tensor, __webpack_exports__Text2TextGenerationPipeline as Text2TextGenerationPipeline, __webpack_exports__TextClassificationPipeline as TextClassificationPipeline, __webpack_exports__TextGenerationPipeline as TextGenerationPipeline, __webpack_exports__TextStreamer as TextStreamer, __webpack_exports__TextToAudioPipeline as TextToAudioPipeline, __webpack_exports__TokenClassificationPipeline as TokenClassificationPipeline, __webpack_exports__TokenClassifierOutput as TokenClassifierOutput, __webpack_exports__TokenizerModel as TokenizerModel, __webpack_exports__TopKLogitsWarper as TopKLogitsWarper, __webpack_exports__TopPLogitsWarper as TopPLogitsWarper, __webpack_exports__TrOCRForCausalLM as TrOCRForCausalLM, __webpack_exports__TrOCRPreTrainedModel as TrOCRPreTrainedModel, __webpack_exports__TranslationPipeline as TranslationPipeline, __webpack_exports__UltravoxModel as UltravoxModel, __webpack_exports__UltravoxPreTrainedModel as UltravoxPreTrainedModel, __webpack_exports__UltravoxProcessor as UltravoxProcessor, __webpack_exports__UniSpeechForCTC as UniSpeechForCTC, __webpack_exports__UniSpeechForSequenceClassification as UniSpeechForSequenceClassification, __webpack_exports__UniSpeechModel as UniSpeechModel, __webpack_exports__UniSpeechPreTrainedModel as UniSpeechPreTrainedModel, __webpack_exports__UniSpeechSatForAudioFrameClassification as UniSpeechSatForAudioFrameClassification, __webpack_exports__UniSpeechSatForCTC as UniSpeechSatForCTC, __webpack_exports__UniSpeechSatForSequenceClassification as UniSpeechSatForSequenceClassification, __webpack_exports__UniSpeechSatModel as UniSpeechSatModel, __webpack_exports__UniSpeechSatPreTrainedModel as UniSpeechSatPreTrainedModel, __webpack_exports__VLChatProcessor as VLChatProcessor, __webpack_exports__VLMImageProcessor as VLMImageProcessor, __webpack_exports__ViTFeatureExtractor as ViTFeatureExtractor, __webpack_exports__ViTForImageClassification as ViTForImageClassification, __webpack_exports__ViTImageProcessor as ViTImageProcessor, __webpack_exports__ViTMAEModel as ViTMAEModel, __webpack_exports__ViTMAEPreTrainedModel as ViTMAEPreTrainedModel, __webpack_exports__ViTMSNForImageClassification as ViTMSNForImageClassification, __webpack_exports__ViTMSNModel as ViTMSNModel, __webpack_exports__ViTMSNPreTrainedModel as ViTMSNPreTrainedModel, __webpack_exports__ViTModel as ViTModel, __webpack_exports__ViTPreTrainedModel as ViTPreTrainedModel, __webpack_exports__VisionEncoderDecoderModel as VisionEncoderDecoderModel, __webpack_exports__VitMatteForImageMatting as VitMatteForImageMatting, __webpack_exports__VitMatteImageProcessor as VitMatteImageProcessor, __webpack_exports__VitMattePreTrainedModel as VitMattePreTrainedModel, __webpack_exports__VitPoseForPoseEstimation as VitPoseForPoseEstimation, __webpack_exports__VitPoseImageProcessor as VitPoseImageProcessor, __webpack_exports__VitPosePreTrainedModel as VitPosePreTrainedModel, __webpack_exports__VitsModel as VitsModel, __webpack_exports__VitsModelOutput as VitsModelOutput, __webpack_exports__VitsPreTrainedModel as VitsPreTrainedModel, __webpack_exports__VitsTokenizer as VitsTokenizer, __webpack_exports__Wav2Vec2BertForCTC as Wav2Vec2BertForCTC, __webpack_exports__Wav2Vec2BertForSequenceClassification as Wav2Vec2BertForSequenceClassification, __webpack_exports__Wav2Vec2BertModel as Wav2Vec2BertModel, __webpack_exports__Wav2Vec2BertPreTrainedModel as Wav2Vec2BertPreTrainedModel, __webpack_exports__Wav2Vec2CTCTokenizer as Wav2Vec2CTCTokenizer, __webpack_exports__Wav2Vec2FeatureExtractor as Wav2Vec2FeatureExtractor, __webpack_exports__Wav2Vec2ForAudioFrameClassification as Wav2Vec2ForAudioFrameClassification, __webpack_exports__Wav2Vec2ForCTC as Wav2Vec2ForCTC, __webpack_exports__Wav2Vec2ForSequenceClassification as Wav2Vec2ForSequenceClassification, __webpack_exports__Wav2Vec2Model as Wav2Vec2Model, __webpack_exports__Wav2Vec2PreTrainedModel as Wav2Vec2PreTrainedModel, __webpack_exports__Wav2Vec2Processor as Wav2Vec2Processor, __webpack_exports__Wav2Vec2ProcessorWithLM as Wav2Vec2ProcessorWithLM, __webpack_exports__WavLMForAudioFrameClassification as WavLMForAudioFrameClassification, __webpack_exports__WavLMForCTC as WavLMForCTC, __webpack_exports__WavLMForSequenceClassification as WavLMForSequenceClassification, __webpack_exports__WavLMForXVector as WavLMForXVector, __webpack_exports__WavLMModel as WavLMModel, __webpack_exports__WavLMPreTrainedModel as WavLMPreTrainedModel, __webpack_exports__WeSpeakerFeatureExtractor as WeSpeakerFeatureExtractor, __webpack_exports__WeSpeakerResNetModel as WeSpeakerResNetModel, __webpack_exports__WeSpeakerResNetPreTrainedModel as WeSpeakerResNetPreTrainedModel, __webpack_exports__WhisperFeatureExtractor as WhisperFeatureExtractor, __webpack_exports__WhisperForConditionalGeneration as WhisperForConditionalGeneration, __webpack_exports__WhisperModel as WhisperModel, __webpack_exports__WhisperPreTrainedModel as WhisperPreTrainedModel, __webpack_exports__WhisperProcessor as WhisperProcessor, __webpack_exports__WhisperTextStreamer as WhisperTextStreamer, __webpack_exports__WhisperTimeStampLogitsProcessor as WhisperTimeStampLogitsProcessor, __webpack_exports__WhisperTokenizer as WhisperTokenizer, __webpack_exports__XLMForQuestionAnswering as XLMForQuestionAnswering, __webpack_exports__XLMForSequenceClassification as XLMForSequenceClassification, __webpack_exports__XLMForTokenClassification as XLMForTokenClassification, __webpack_exports__XLMModel as XLMModel, __webpack_exports__XLMPreTrainedModel as XLMPreTrainedModel, __webpack_exports__XLMRobertaForMaskedLM as XLMRobertaForMaskedLM, __webpack_exports__XLMRobertaForQuestionAnswering as XLMRobertaForQuestionAnswering, __webpack_exports__XLMRobertaForSequenceClassification as XLMRobertaForSequenceClassification, __webpack_exports__XLMRobertaForTokenClassification as XLMRobertaForTokenClassification, __webpack_exports__XLMRobertaModel as XLMRobertaModel, __webpack_exports__XLMRobertaPreTrainedModel as XLMRobertaPreTrainedModel, __webpack_exports__XLMRobertaTokenizer as XLMRobertaTokenizer, __webpack_exports__XLMTokenizer as XLMTokenizer, __webpack_exports__XLMWithLMHeadModel as XLMWithLMHeadModel, __webpack_exports__XVectorOutput as XVectorOutput, __webpack_exports__YolosFeatureExtractor as YolosFeatureExtractor, __webpack_exports__YolosForObjectDetection as YolosForObjectDetection, __webpack_exports__YolosImageProcessor as YolosImageProcessor, __webpack_exports__YolosModel as YolosModel, __webpack_exports__YolosObjectDetectionOutput as YolosObjectDetectionOutput, __webpack_exports__YolosPreTrainedModel as YolosPreTrainedModel, __webpack_exports__ZeroShotAudioClassificationPipeline as ZeroShotAudioClassificationPipeline, __webpack_exports__ZeroShotClassificationPipeline as ZeroShotClassificationPipeline, __webpack_exports__ZeroShotImageClassificationPipeline as ZeroShotImageClassificationPipeline, __webpack_exports__ZeroShotObjectDetectionPipeline as ZeroShotObjectDetectionPipeline, __webpack_exports__bankers_round as bankers_round, __webpack_exports__cat as cat, __webpack_exports__cos_sim as cos_sim, __webpack_exports__dot as dot, __webpack_exports__dynamic_time_warping as dynamic_time_warping, __webpack_exports__env as env, __webpack_exports__full as full, __webpack_exports__full_like as full_like, __webpack_exports__getKeyValueShapes as getKeyValueShapes, __webpack_exports__hamming as hamming, __webpack_exports__hanning as hanning, __webpack_exports__interpolate as interpolate, __webpack_exports__interpolate_4d as interpolate_4d, __webpack_exports__interpolate_data as interpolate_data, __webpack_exports__is_chinese_char as is_chinese_char, __webpack_exports__layer_norm as layer_norm, __webpack_exports__load_image as load_image, __webpack_exports__load_video as load_video, __webpack_exports__log_softmax as log_softmax, __webpack_exports__magnitude as magnitude, __webpack_exports__matmul as matmul, __webpack_exports__max as max, __webpack_exports__mean as mean, __webpack_exports__mean_pooling as mean_pooling, __webpack_exports__medianFilter as medianFilter, __webpack_exports__mel_filter_bank as mel_filter_bank, __webpack_exports__min as min, __webpack_exports__ones as ones, __webpack_exports__ones_like as ones_like, __webpack_exports__permute as permute, __webpack_exports__permute_data as permute_data, __webpack_exports__pipeline as pipeline, __webpack_exports__quantize_embeddings as quantize_embeddings, __webpack_exports__rand as rand, __webpack_exports__read_audio as read_audio, __webpack_exports__rfft as rfft, __webpack_exports__round as round, __webpack_exports__slice as slice, __webpack_exports__softmax as softmax, __webpack_exports__spectrogram as spectrogram, __webpack_exports__stack as stack, __webpack_exports__std_mean as std_mean, __webpack_exports__topk as topk, __webpack_exports__window_function as window_function, __webpack_exports__zeros as zeros, __webpack_exports__zeros_like as zeros_like };
35696
35859
 
35697
35860
  //# sourceMappingURL=transformers.node.mjs.map