@huggingface/transformers 3.4.0 → 3.4.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (38) hide show
  1. package/README.md +6 -2
  2. package/dist/transformers.js +315 -152
  3. package/dist/transformers.js.map +1 -1
  4. package/dist/transformers.min.js +1 -1
  5. package/dist/transformers.min.js.map +1 -1
  6. package/dist/transformers.node.cjs +303 -151
  7. package/dist/transformers.node.cjs.map +1 -1
  8. package/dist/transformers.node.min.cjs +1 -1
  9. package/dist/transformers.node.min.cjs.map +1 -1
  10. package/dist/transformers.node.min.mjs +1 -1
  11. package/dist/transformers.node.min.mjs.map +1 -1
  12. package/dist/transformers.node.mjs +315 -152
  13. package/dist/transformers.node.mjs.map +1 -1
  14. package/dist/transformers.web.js +315 -152
  15. package/dist/transformers.web.js.map +1 -1
  16. package/dist/transformers.web.min.js +1 -1
  17. package/dist/transformers.web.min.js.map +1 -1
  18. package/package.json +1 -1
  19. package/src/configs.js +2 -0
  20. package/src/env.js +1 -1
  21. package/src/models/feature_extractors.js +1 -0
  22. package/src/models/snac/feature_extraction_snac.js +3 -0
  23. package/src/models.js +90 -2
  24. package/src/pipelines.js +140 -135
  25. package/src/utils/image.js +9 -1
  26. package/src/utils/tensor.js +6 -2
  27. package/types/configs.d.ts.map +1 -1
  28. package/types/models/feature_extractors.d.ts +1 -0
  29. package/types/models/snac/feature_extraction_snac.d.ts +4 -0
  30. package/types/models/snac/feature_extraction_snac.d.ts.map +1 -0
  31. package/types/models.d.ts +48 -0
  32. package/types/models.d.ts.map +1 -1
  33. package/types/pipelines.d.ts +2 -2
  34. package/types/pipelines.d.ts.map +1 -1
  35. package/types/tsconfig.tsbuildinfo +1 -1
  36. package/types/utils/image.d.ts +2 -2
  37. package/types/utils/image.d.ts.map +1 -1
  38. package/types/utils/tensor.d.ts.map +1 -1
@@ -8496,6 +8496,7 @@ function getNormalizedConfig(config) {
8496
8496
  // Sub-configs
8497
8497
  case 'llava':
8498
8498
  case 'paligemma':
8499
+ case 'gemma3':
8499
8500
  case 'florence2':
8500
8501
  case 'llava_onevision':
8501
8502
  case 'idefics3':
@@ -8555,6 +8556,7 @@ function getNormalizedConfig(config) {
8555
8556
  break;
8556
8557
  case 'gemma':
8557
8558
  case 'gemma2':
8559
+ case 'gemma3_text':
8558
8560
  case 'glm':
8559
8561
  case 'helium':
8560
8562
  mapping['num_heads'] = 'num_key_value_heads';
@@ -8885,7 +8887,7 @@ __webpack_require__.r(__webpack_exports__);
8885
8887
 
8886
8888
 
8887
8889
 
8888
- const VERSION = '3.4.0';
8890
+ const VERSION = '3.4.1';
8889
8891
 
8890
8892
  // Check if various APIs are available (depends on environment)
8891
8893
  const IS_BROWSER_ENV = typeof window !== "undefined" && typeof window.document !== "undefined";
@@ -11023,6 +11025,9 @@ __webpack_require__.r(__webpack_exports__);
11023
11025
  /* harmony export */ Gemma2ForCausalLM: () => (/* binding */ Gemma2ForCausalLM),
11024
11026
  /* harmony export */ Gemma2Model: () => (/* binding */ Gemma2Model),
11025
11027
  /* harmony export */ Gemma2PreTrainedModel: () => (/* binding */ Gemma2PreTrainedModel),
11028
+ /* harmony export */ Gemma3ForCausalLM: () => (/* binding */ Gemma3ForCausalLM),
11029
+ /* harmony export */ Gemma3Model: () => (/* binding */ Gemma3Model),
11030
+ /* harmony export */ Gemma3PreTrainedModel: () => (/* binding */ Gemma3PreTrainedModel),
11026
11031
  /* harmony export */ GemmaForCausalLM: () => (/* binding */ GemmaForCausalLM),
11027
11032
  /* harmony export */ GemmaModel: () => (/* binding */ GemmaModel),
11028
11033
  /* harmony export */ GemmaPreTrainedModel: () => (/* binding */ GemmaPreTrainedModel),
@@ -11093,6 +11098,10 @@ __webpack_require__.r(__webpack_exports__);
11093
11098
  /* harmony export */ MaskFormerModel: () => (/* binding */ MaskFormerModel),
11094
11099
  /* harmony export */ MaskFormerPreTrainedModel: () => (/* binding */ MaskFormerPreTrainedModel),
11095
11100
  /* harmony export */ MaskedLMOutput: () => (/* binding */ MaskedLMOutput),
11101
+ /* harmony export */ Metric3DForDepthEstimation: () => (/* binding */ Metric3DForDepthEstimation),
11102
+ /* harmony export */ Metric3DPreTrainedModel: () => (/* binding */ Metric3DPreTrainedModel),
11103
+ /* harmony export */ Metric3Dv2ForDepthEstimation: () => (/* binding */ Metric3Dv2ForDepthEstimation),
11104
+ /* harmony export */ Metric3Dv2PreTrainedModel: () => (/* binding */ Metric3Dv2PreTrainedModel),
11096
11105
  /* harmony export */ MgpstrForSceneTextRecognition: () => (/* binding */ MgpstrForSceneTextRecognition),
11097
11106
  /* harmony export */ MgpstrModelOutput: () => (/* binding */ MgpstrModelOutput),
11098
11107
  /* harmony export */ MgpstrPreTrainedModel: () => (/* binding */ MgpstrPreTrainedModel),
@@ -11241,6 +11250,10 @@ __webpack_require__.r(__webpack_exports__);
11241
11250
  /* harmony export */ SiglipTextModel: () => (/* binding */ SiglipTextModel),
11242
11251
  /* harmony export */ SiglipVisionModel: () => (/* binding */ SiglipVisionModel),
11243
11252
  /* harmony export */ SmolVLMForConditionalGeneration: () => (/* binding */ SmolVLMForConditionalGeneration),
11253
+ /* harmony export */ SnacDecoderModel: () => (/* binding */ SnacDecoderModel),
11254
+ /* harmony export */ SnacEncoderModel: () => (/* binding */ SnacEncoderModel),
11255
+ /* harmony export */ SnacModel: () => (/* binding */ SnacModel),
11256
+ /* harmony export */ SnacPreTrainedModel: () => (/* binding */ SnacPreTrainedModel),
11244
11257
  /* harmony export */ SpeechT5ForSpeechToText: () => (/* binding */ SpeechT5ForSpeechToText),
11245
11258
  /* harmony export */ SpeechT5ForTextToSpeech: () => (/* binding */ SpeechT5ForTextToSpeech),
11246
11259
  /* harmony export */ SpeechT5HifiGan: () => (/* binding */ SpeechT5HifiGan),
@@ -11899,8 +11912,8 @@ async function decoderForward(self, model_inputs, is_encoder_decoder = false) {
11899
11912
  new_model_inputs.use_cache_branch = boolTensor(!!past_key_values);
11900
11913
  }
11901
11914
  if (session.inputNames.includes('position_ids') && new_model_inputs.attention_mask && !new_model_inputs.position_ids) {
11902
- // NOTE: Handle a special case for paligemma models, where positions are 1-indexed
11903
- const start_index = self.config.model_type === 'paligemma' ? 1 : 0;
11915
+ // NOTE: Handle a special case for paligemma/gemma3 models, where positions are 1-indexed
11916
+ const start_index = ['paligemma', 'gemma3_text', 'gemma3'].includes(self.config.model_type) ? 1 : 0;
11904
11917
  new_model_inputs.position_ids = createPositionIds(new_model_inputs, past_key_values, start_index);
11905
11918
  }
11906
11919
 
@@ -15825,6 +15838,23 @@ class Gemma2Model extends Gemma2PreTrainedModel { }
15825
15838
  class Gemma2ForCausalLM extends Gemma2PreTrainedModel { }
15826
15839
  //////////////////////////////////////////////////
15827
15840
 
15841
+
15842
+ //////////////////////////////////////////////////
15843
+ // Gemma3 models
15844
+
15845
+ /**
15846
+ * The bare Gemma3 Model outputting raw hidden-states without any specific head on top.
15847
+ */
15848
+ class Gemma3PreTrainedModel extends PreTrainedModel { }
15849
+ /**
15850
+ * The bare Gemma3 Model outputting raw hidden-states without any specific head on top.
15851
+ */
15852
+ class Gemma3Model extends Gemma3PreTrainedModel { }
15853
+
15854
+ class Gemma3ForCausalLM extends Gemma3PreTrainedModel { }
15855
+ //////////////////////////////////////////////////
15856
+
15857
+
15828
15858
  //////////////////////////////////////////////////
15829
15859
  class OpenELMPreTrainedModel extends PreTrainedModel { }
15830
15860
  class OpenELMModel extends OpenELMPreTrainedModel { }
@@ -16677,6 +16707,16 @@ class DepthProPreTrainedModel extends PreTrainedModel { }
16677
16707
  class DepthProForDepthEstimation extends DepthProPreTrainedModel { }
16678
16708
  //////////////////////////////////////////////////
16679
16709
 
16710
+ //////////////////////////////////////////////////
16711
+ class Metric3DPreTrainedModel extends PreTrainedModel { }
16712
+ class Metric3DForDepthEstimation extends Metric3DPreTrainedModel { }
16713
+ //////////////////////////////////////////////////
16714
+
16715
+ //////////////////////////////////////////////////
16716
+ class Metric3Dv2PreTrainedModel extends PreTrainedModel { }
16717
+ class Metric3Dv2ForDepthEstimation extends Metric3Dv2PreTrainedModel { }
16718
+ //////////////////////////////////////////////////
16719
+
16680
16720
  //////////////////////////////////////////////////
16681
16721
  class MaskFormerPreTrainedModel extends PreTrainedModel { }
16682
16722
  class MaskFormerModel extends MaskFormerPreTrainedModel { }
@@ -18592,6 +18632,60 @@ class DacDecoderModel extends DacPreTrainedModel {
18592
18632
  }
18593
18633
  //////////////////////////////////////////////////
18594
18634
 
18635
+
18636
+ //////////////////////////////////////////////////
18637
+ // Snac models
18638
+ class SnacPreTrainedModel extends PreTrainedModel {
18639
+ main_input_name = 'input_values';
18640
+ forward_params = ['input_values'];
18641
+ }
18642
+
18643
+ /**
18644
+ * The SNAC (Multi-Scale Neural Audio Codec) model.
18645
+ */
18646
+ class SnacModel extends SnacPreTrainedModel {
18647
+ /**
18648
+ * Encodes the input audio waveform into discrete codes.
18649
+ * @param {Object} inputs Model inputs
18650
+ * @param {Tensor} [inputs.input_values] Float values of the input audio waveform, of shape `(batch_size, channels, sequence_length)`).
18651
+ * @returns {Promise<Record<string, Tensor>>} The output tensors of shape `(batch_size, num_codebooks, sequence_length)`.
18652
+ */
18653
+ async encode(inputs) {
18654
+ return await sessionRun(this.sessions['encoder_model'], inputs);
18655
+ }
18656
+
18657
+ /**
18658
+ * Decodes the given frames into an output audio waveform.
18659
+ * @param {Record<string, Tensor>} inputs The encoded audio codes.
18660
+ * @returns {Promise<{audio_values: Tensor}>} The output tensor of shape `(batch_size, num_channels, sequence_length)`.
18661
+ */
18662
+ async decode(inputs) {
18663
+ return await sessionRun(this.sessions['decoder_model'], inputs);
18664
+ }
18665
+ }
18666
+
18667
+ class SnacEncoderModel extends SnacPreTrainedModel {
18668
+ /** @type {typeof PreTrainedModel.from_pretrained} */
18669
+ static async from_pretrained(pretrained_model_name_or_path, options = {}) {
18670
+ return super.from_pretrained(pretrained_model_name_or_path, {
18671
+ ...options,
18672
+ // Update default model file name if not provided
18673
+ model_file_name: options.model_file_name ?? 'encoder_model',
18674
+ });
18675
+ }
18676
+ }
18677
+ class SnacDecoderModel extends SnacPreTrainedModel {
18678
+ /** @type {typeof PreTrainedModel.from_pretrained} */
18679
+ static async from_pretrained(pretrained_model_name_or_path, options = {}) {
18680
+ return super.from_pretrained(pretrained_model_name_or_path, {
18681
+ ...options,
18682
+ // Update default model file name if not provided
18683
+ model_file_name: options.model_file_name ?? 'decoder_model',
18684
+ });
18685
+ }
18686
+ }
18687
+ //////////////////////////////////////////////////
18688
+
18595
18689
  //////////////////////////////////////////////////
18596
18690
  // AutoModels, used to simplify construction of PreTrainedModels
18597
18691
  // (uses config to instantiate correct class)
@@ -18773,6 +18867,7 @@ const MODEL_MAPPING_NAMES_ENCODER_DECODER = new Map([
18773
18867
  const MODEL_MAPPING_NAMES_AUTO_ENCODER = new Map([
18774
18868
  ['mimi', ['MimiModel', MimiModel]],
18775
18869
  ['dac', ['DacModel', DacModel]],
18870
+ ['snac', ['SnacModel', SnacModel]],
18776
18871
  ]);
18777
18872
 
18778
18873
  const MODEL_MAPPING_NAMES_DECODER_ONLY = new Map([
@@ -18793,6 +18888,7 @@ const MODEL_MAPPING_NAMES_DECODER_ONLY = new Map([
18793
18888
  ['cohere', ['CohereModel', CohereModel]],
18794
18889
  ['gemma', ['GemmaModel', GemmaModel]],
18795
18890
  ['gemma2', ['Gemma2Model', Gemma2Model]],
18891
+ ['gemma3_text', ['Gemma3Model', Gemma3Model]],
18796
18892
  ['helium', ['HeliumModel', HeliumModel]],
18797
18893
  ['glm', ['GlmModel', GlmModel]],
18798
18894
  ['openelm', ['OpenELMModel', OpenELMModel]],
@@ -18892,6 +18988,7 @@ const MODEL_FOR_CAUSAL_LM_MAPPING_NAMES = new Map([
18892
18988
  ['cohere', ['CohereForCausalLM', CohereForCausalLM]],
18893
18989
  ['gemma', ['GemmaForCausalLM', GemmaForCausalLM]],
18894
18990
  ['gemma2', ['Gemma2ForCausalLM', Gemma2ForCausalLM]],
18991
+ ['gemma3_text', ['Gemma3ForCausalLM', Gemma3ForCausalLM]],
18895
18992
  ['helium', ['HeliumForCausalLM', HeliumForCausalLM]],
18896
18993
  ['glm', ['GlmForCausalLM', GlmForCausalLM]],
18897
18994
  ['openelm', ['OpenELMForCausalLM', OpenELMForCausalLM]],
@@ -19093,6 +19190,8 @@ const MODEL_FOR_DEPTH_ESTIMATION_MAPPING_NAMES = new Map([
19093
19190
  ['glpn', ['GLPNForDepthEstimation', GLPNForDepthEstimation]],
19094
19191
  ['sapiens', ['SapiensForDepthEstimation', SapiensForDepthEstimation]],
19095
19192
  ['depth_pro', ['DepthProForDepthEstimation', DepthProForDepthEstimation]],
19193
+ ['metric3d', ['Metric3DForDepthEstimation', Metric3DForDepthEstimation]],
19194
+ ['metric3dv2', ['Metric3Dv2ForDepthEstimation', Metric3Dv2ForDepthEstimation]],
19096
19195
  ])
19097
19196
 
19098
19197
  const MODEL_FOR_NORMAL_ESTIMATION_MAPPING_NAMES = new Map([
@@ -19178,6 +19277,8 @@ const CUSTOM_MAPPING = [
19178
19277
  ['DacDecoderModel', DacDecoderModel, MODEL_TYPES.EncoderOnly],
19179
19278
  ['MimiEncoderModel', MimiEncoderModel, MODEL_TYPES.EncoderOnly],
19180
19279
  ['MimiDecoderModel', MimiDecoderModel, MODEL_TYPES.EncoderOnly],
19280
+ ['SnacEncoderModel', SnacEncoderModel, MODEL_TYPES.EncoderOnly],
19281
+ ['SnacDecoderModel', SnacDecoderModel, MODEL_TYPES.EncoderOnly],
19181
19282
  ]
19182
19283
  for (const [name, model, type] of CUSTOM_MAPPING) {
19183
19284
  MODEL_TYPE_MAPPING.set(name, type);
@@ -20476,14 +20577,15 @@ __webpack_require__.r(__webpack_exports__);
20476
20577
  /* harmony export */ ClapFeatureExtractor: () => (/* reexport safe */ _clap_feature_extraction_clap_js__WEBPACK_IMPORTED_MODULE_2__.ClapFeatureExtractor),
20477
20578
  /* harmony export */ DacFeatureExtractor: () => (/* reexport safe */ _dac_feature_extraction_dac_js__WEBPACK_IMPORTED_MODULE_3__.DacFeatureExtractor),
20478
20579
  /* harmony export */ EncodecFeatureExtractor: () => (/* reexport safe */ _encodec_feature_extraction_encodec_js__WEBPACK_IMPORTED_MODULE_1__.EncodecFeatureExtractor),
20479
- /* harmony export */ ImageFeatureExtractor: () => (/* reexport safe */ _base_image_processors_utils_js__WEBPACK_IMPORTED_MODULE_11__.ImageProcessor),
20580
+ /* harmony export */ ImageFeatureExtractor: () => (/* reexport safe */ _base_image_processors_utils_js__WEBPACK_IMPORTED_MODULE_12__.ImageProcessor),
20480
20581
  /* harmony export */ MoonshineFeatureExtractor: () => (/* reexport safe */ _moonshine_feature_extraction_moonshine_js__WEBPACK_IMPORTED_MODULE_4__.MoonshineFeatureExtractor),
20481
20582
  /* harmony export */ PyAnnoteFeatureExtractor: () => (/* reexport safe */ _pyannote_feature_extraction_pyannote_js__WEBPACK_IMPORTED_MODULE_5__.PyAnnoteFeatureExtractor),
20482
20583
  /* harmony export */ SeamlessM4TFeatureExtractor: () => (/* reexport safe */ _seamless_m4t_feature_extraction_seamless_m4t_js__WEBPACK_IMPORTED_MODULE_6__.SeamlessM4TFeatureExtractor),
20483
- /* harmony export */ SpeechT5FeatureExtractor: () => (/* reexport safe */ _speecht5_feature_extraction_speecht5_js__WEBPACK_IMPORTED_MODULE_7__.SpeechT5FeatureExtractor),
20484
- /* harmony export */ Wav2Vec2FeatureExtractor: () => (/* reexport safe */ _wav2vec2_feature_extraction_wav2vec2_js__WEBPACK_IMPORTED_MODULE_8__.Wav2Vec2FeatureExtractor),
20485
- /* harmony export */ WeSpeakerFeatureExtractor: () => (/* reexport safe */ _wespeaker_feature_extraction_wespeaker_js__WEBPACK_IMPORTED_MODULE_9__.WeSpeakerFeatureExtractor),
20486
- /* harmony export */ WhisperFeatureExtractor: () => (/* reexport safe */ _whisper_feature_extraction_whisper_js__WEBPACK_IMPORTED_MODULE_10__.WhisperFeatureExtractor)
20584
+ /* harmony export */ SnacFeatureExtractor: () => (/* reexport safe */ _snac_feature_extraction_snac_js__WEBPACK_IMPORTED_MODULE_7__.SnacFeatureExtractor),
20585
+ /* harmony export */ SpeechT5FeatureExtractor: () => (/* reexport safe */ _speecht5_feature_extraction_speecht5_js__WEBPACK_IMPORTED_MODULE_8__.SpeechT5FeatureExtractor),
20586
+ /* harmony export */ Wav2Vec2FeatureExtractor: () => (/* reexport safe */ _wav2vec2_feature_extraction_wav2vec2_js__WEBPACK_IMPORTED_MODULE_9__.Wav2Vec2FeatureExtractor),
20587
+ /* harmony export */ WeSpeakerFeatureExtractor: () => (/* reexport safe */ _wespeaker_feature_extraction_wespeaker_js__WEBPACK_IMPORTED_MODULE_10__.WeSpeakerFeatureExtractor),
20588
+ /* harmony export */ WhisperFeatureExtractor: () => (/* reexport safe */ _whisper_feature_extraction_whisper_js__WEBPACK_IMPORTED_MODULE_11__.WhisperFeatureExtractor)
20487
20589
  /* harmony export */ });
20488
20590
  /* harmony import */ var _audio_spectrogram_transformer_feature_extraction_audio_spectrogram_transformer_js__WEBPACK_IMPORTED_MODULE_0__ = __webpack_require__(/*! ./audio_spectrogram_transformer/feature_extraction_audio_spectrogram_transformer.js */ "./src/models/audio_spectrogram_transformer/feature_extraction_audio_spectrogram_transformer.js");
20489
20591
  /* harmony import */ var _encodec_feature_extraction_encodec_js__WEBPACK_IMPORTED_MODULE_1__ = __webpack_require__(/*! ./encodec/feature_extraction_encodec.js */ "./src/models/encodec/feature_extraction_encodec.js");
@@ -20492,11 +20594,13 @@ __webpack_require__.r(__webpack_exports__);
20492
20594
  /* harmony import */ var _moonshine_feature_extraction_moonshine_js__WEBPACK_IMPORTED_MODULE_4__ = __webpack_require__(/*! ./moonshine/feature_extraction_moonshine.js */ "./src/models/moonshine/feature_extraction_moonshine.js");
20493
20595
  /* harmony import */ var _pyannote_feature_extraction_pyannote_js__WEBPACK_IMPORTED_MODULE_5__ = __webpack_require__(/*! ./pyannote/feature_extraction_pyannote.js */ "./src/models/pyannote/feature_extraction_pyannote.js");
20494
20596
  /* harmony import */ var _seamless_m4t_feature_extraction_seamless_m4t_js__WEBPACK_IMPORTED_MODULE_6__ = __webpack_require__(/*! ./seamless_m4t/feature_extraction_seamless_m4t.js */ "./src/models/seamless_m4t/feature_extraction_seamless_m4t.js");
20495
- /* harmony import */ var _speecht5_feature_extraction_speecht5_js__WEBPACK_IMPORTED_MODULE_7__ = __webpack_require__(/*! ./speecht5/feature_extraction_speecht5.js */ "./src/models/speecht5/feature_extraction_speecht5.js");
20496
- /* harmony import */ var _wav2vec2_feature_extraction_wav2vec2_js__WEBPACK_IMPORTED_MODULE_8__ = __webpack_require__(/*! ./wav2vec2/feature_extraction_wav2vec2.js */ "./src/models/wav2vec2/feature_extraction_wav2vec2.js");
20497
- /* harmony import */ var _wespeaker_feature_extraction_wespeaker_js__WEBPACK_IMPORTED_MODULE_9__ = __webpack_require__(/*! ./wespeaker/feature_extraction_wespeaker.js */ "./src/models/wespeaker/feature_extraction_wespeaker.js");
20498
- /* harmony import */ var _whisper_feature_extraction_whisper_js__WEBPACK_IMPORTED_MODULE_10__ = __webpack_require__(/*! ./whisper/feature_extraction_whisper.js */ "./src/models/whisper/feature_extraction_whisper.js");
20499
- /* harmony import */ var _base_image_processors_utils_js__WEBPACK_IMPORTED_MODULE_11__ = __webpack_require__(/*! ../base/image_processors_utils.js */ "./src/base/image_processors_utils.js");
20597
+ /* harmony import */ var _snac_feature_extraction_snac_js__WEBPACK_IMPORTED_MODULE_7__ = __webpack_require__(/*! ./snac/feature_extraction_snac.js */ "./src/models/snac/feature_extraction_snac.js");
20598
+ /* harmony import */ var _speecht5_feature_extraction_speecht5_js__WEBPACK_IMPORTED_MODULE_8__ = __webpack_require__(/*! ./speecht5/feature_extraction_speecht5.js */ "./src/models/speecht5/feature_extraction_speecht5.js");
20599
+ /* harmony import */ var _wav2vec2_feature_extraction_wav2vec2_js__WEBPACK_IMPORTED_MODULE_9__ = __webpack_require__(/*! ./wav2vec2/feature_extraction_wav2vec2.js */ "./src/models/wav2vec2/feature_extraction_wav2vec2.js");
20600
+ /* harmony import */ var _wespeaker_feature_extraction_wespeaker_js__WEBPACK_IMPORTED_MODULE_10__ = __webpack_require__(/*! ./wespeaker/feature_extraction_wespeaker.js */ "./src/models/wespeaker/feature_extraction_wespeaker.js");
20601
+ /* harmony import */ var _whisper_feature_extraction_whisper_js__WEBPACK_IMPORTED_MODULE_11__ = __webpack_require__(/*! ./whisper/feature_extraction_whisper.js */ "./src/models/whisper/feature_extraction_whisper.js");
20602
+ /* harmony import */ var _base_image_processors_utils_js__WEBPACK_IMPORTED_MODULE_12__ = __webpack_require__(/*! ../base/image_processors_utils.js */ "./src/base/image_processors_utils.js");
20603
+
20500
20604
 
20501
20605
 
20502
20606
 
@@ -23486,6 +23590,24 @@ __webpack_require__.r(__webpack_exports__);
23486
23590
 
23487
23591
 
23488
23592
 
23593
+ /***/ }),
23594
+
23595
+ /***/ "./src/models/snac/feature_extraction_snac.js":
23596
+ /*!****************************************************!*\
23597
+ !*** ./src/models/snac/feature_extraction_snac.js ***!
23598
+ \****************************************************/
23599
+ /***/ ((__unused_webpack___webpack_module__, __webpack_exports__, __webpack_require__) => {
23600
+
23601
+ __webpack_require__.r(__webpack_exports__);
23602
+ /* harmony export */ __webpack_require__.d(__webpack_exports__, {
23603
+ /* harmony export */ SnacFeatureExtractor: () => (/* binding */ SnacFeatureExtractor)
23604
+ /* harmony export */ });
23605
+ /* harmony import */ var _dac_feature_extraction_dac_js__WEBPACK_IMPORTED_MODULE_0__ = __webpack_require__(/*! ../dac/feature_extraction_dac.js */ "./src/models/dac/feature_extraction_dac.js");
23606
+
23607
+
23608
+ class SnacFeatureExtractor extends _dac_feature_extraction_dac_js__WEBPACK_IMPORTED_MODULE_0__.DacFeatureExtractor { }
23609
+
23610
+
23489
23611
  /***/ }),
23490
23612
 
23491
23613
  /***/ "./src/models/speecht5/feature_extraction_speecht5.js":
@@ -24725,16 +24847,16 @@ __webpack_require__.r(__webpack_exports__);
24725
24847
  /* harmony import */ var _utils_image_js__WEBPACK_IMPORTED_MODULE_9__ = __webpack_require__(/*! ./utils/image.js */ "./src/utils/image.js");
24726
24848
  /**
24727
24849
  * @file Pipelines provide a high-level, easy to use, API for running machine learning models.
24728
- *
24850
+ *
24729
24851
  * **Example:** Instantiate pipeline using the `pipeline` function.
24730
24852
  * ```javascript
24731
24853
  * import { pipeline } from '@huggingface/transformers';
24732
- *
24854
+ *
24733
24855
  * const classifier = await pipeline('sentiment-analysis');
24734
24856
  * const output = await classifier('I love transformers!');
24735
24857
  * // [{'label': 'POSITIVE', 'score': 0.999817686}]
24736
24858
  * ```
24737
- *
24859
+ *
24738
24860
  * @module pipelines
24739
24861
  */
24740
24862
 
@@ -24753,7 +24875,7 @@ __webpack_require__.r(__webpack_exports__);
24753
24875
 
24754
24876
 
24755
24877
  /**
24756
- * @typedef {string | RawImage | URL} ImageInput
24878
+ * @typedef {string | RawImage | URL | Blob | HTMLCanvasElement | OffscreenCanvas} ImageInput
24757
24879
  * @typedef {ImageInput|ImageInput[]} ImagePipelineInputs
24758
24880
  */
24759
24881
 
@@ -24827,7 +24949,7 @@ function get_bounding_box(box, asInteger) {
24827
24949
  /**
24828
24950
  * @callback DisposeType Disposes the item.
24829
24951
  * @returns {Promise<void>} A promise that resolves when the item has been disposed.
24830
- *
24952
+ *
24831
24953
  * @typedef {Object} Disposable
24832
24954
  * @property {DisposeType} dispose A promise that resolves when the pipeline has been disposed.
24833
24955
  */
@@ -24864,7 +24986,7 @@ class Pipeline extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_4__.Callable {
24864
24986
  * @property {string} task The task of the pipeline. Useful for specifying subtasks.
24865
24987
  * @property {PreTrainedModel} model The model used by the pipeline.
24866
24988
  * @property {PreTrainedTokenizer} tokenizer The tokenizer used by the pipeline.
24867
- *
24989
+ *
24868
24990
  * @typedef {ModelTokenizerConstructorArgs} TextPipelineConstructorArgs An object used to instantiate a text-based pipeline.
24869
24991
  */
24870
24992
 
@@ -24873,7 +24995,7 @@ class Pipeline extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_4__.Callable {
24873
24995
  * @property {string} task The task of the pipeline. Useful for specifying subtasks.
24874
24996
  * @property {PreTrainedModel} model The model used by the pipeline.
24875
24997
  * @property {Processor} processor The processor used by the pipeline.
24876
- *
24998
+ *
24877
24999
  * @typedef {ModelProcessorConstructorArgs} AudioPipelineConstructorArgs An object used to instantiate an audio-based pipeline.
24878
25000
  * @typedef {ModelProcessorConstructorArgs} ImagePipelineConstructorArgs An object used to instantiate an image-based pipeline.
24879
25001
  */
@@ -24885,7 +25007,7 @@ class Pipeline extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_4__.Callable {
24885
25007
  * @property {PreTrainedModel} model The model used by the pipeline.
24886
25008
  * @property {PreTrainedTokenizer} tokenizer The tokenizer used by the pipeline.
24887
25009
  * @property {Processor} processor The processor used by the pipeline.
24888
- *
25010
+ *
24889
25011
  * @typedef {ModelTokenizerProcessorConstructorArgs} TextAudioPipelineConstructorArgs An object used to instantiate a text- and audio-based pipeline.
24890
25012
  * @typedef {ModelTokenizerProcessorConstructorArgs} TextImagePipelineConstructorArgs An object used to instantiate a text- and image-based pipeline.
24891
25013
  */
@@ -24895,15 +25017,15 @@ class Pipeline extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_4__.Callable {
24895
25017
  * @property {string} label The label predicted.
24896
25018
  * @property {number} score The corresponding probability.
24897
25019
  * @typedef {TextClassificationSingle[]} TextClassificationOutput
24898
- *
25020
+ *
24899
25021
  * @typedef {Object} TextClassificationPipelineOptions Parameters specific to text classification pipelines.
24900
25022
  * @property {number} [top_k=1] The number of top predictions to be returned.
24901
- *
25023
+ *
24902
25024
  * @callback TextClassificationPipelineCallback Classify the text(s) given as inputs.
24903
25025
  * @param {string|string[]} texts The input text(s) to be classified.
24904
25026
  * @param {TextClassificationPipelineOptions} [options] The options to use for text classification.
24905
25027
  * @returns {Promise<TextClassificationOutput|TextClassificationOutput[]>} An array or object containing the predicted labels and scores.
24906
- *
25028
+ *
24907
25029
  * @typedef {TextPipelineConstructorArgs & TextClassificationPipelineCallback & Disposable} TextClassificationPipelineType
24908
25030
  */
24909
25031
 
@@ -24916,7 +25038,7 @@ class Pipeline extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_4__.Callable {
24916
25038
  * const output = await classifier('I love transformers!');
24917
25039
  * // [{ label: 'POSITIVE', score: 0.999788761138916 }]
24918
25040
  * ```
24919
- *
25041
+ *
24920
25042
  * **Example:** Multilingual sentiment-analysis w/ `Xenova/bert-base-multilingual-uncased-sentiment` (and return top 5 classes).
24921
25043
  * ```javascript
24922
25044
  * const classifier = await pipeline('sentiment-analysis', 'Xenova/bert-base-multilingual-uncased-sentiment');
@@ -24929,7 +25051,7 @@ class Pipeline extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_4__.Callable {
24929
25051
  * // { label: '2 stars', score: 0.0009423971059732139 }
24930
25052
  * // ]
24931
25053
  * ```
24932
- *
25054
+ *
24933
25055
  * **Example:** Toxic comment classification w/ `Xenova/toxic-bert` (and return all classes).
24934
25056
  * ```javascript
24935
25057
  * const classifier = await pipeline('text-classification', 'Xenova/toxic-bert');
@@ -25014,21 +25136,21 @@ class TextClassificationPipeline extends (/** @type {new (options: TextPipelineC
25014
25136
  * @property {number} [start] The index of the start of the corresponding entity in the sentence.
25015
25137
  * @property {number} [end] The index of the end of the corresponding entity in the sentence.
25016
25138
  * @typedef {TokenClassificationSingle[]} TokenClassificationOutput
25017
- *
25139
+ *
25018
25140
  * @typedef {Object} TokenClassificationPipelineOptions Parameters specific to token classification pipelines.
25019
25141
  * @property {string[]} [ignore_labels] A list of labels to ignore.
25020
- *
25142
+ *
25021
25143
  * @callback TokenClassificationPipelineCallback Classify each token of the text(s) given as inputs.
25022
25144
  * @param {string|string[]} texts One or several texts (or one list of texts) for token classification.
25023
25145
  * @param {TokenClassificationPipelineOptions} [options] The options to use for token classification.
25024
25146
  * @returns {Promise<TokenClassificationOutput|TokenClassificationOutput[]>} The result.
25025
- *
25147
+ *
25026
25148
  * @typedef {TextPipelineConstructorArgs & TokenClassificationPipelineCallback & Disposable} TokenClassificationPipelineType
25027
25149
  */
25028
25150
 
25029
25151
  /**
25030
25152
  * Named Entity Recognition pipeline using any `ModelForTokenClassification`.
25031
- *
25153
+ *
25032
25154
  * **Example:** Perform named entity recognition with `Xenova/bert-base-NER`.
25033
25155
  * ```javascript
25034
25156
  * const classifier = await pipeline('token-classification', 'Xenova/bert-base-NER');
@@ -25038,7 +25160,7 @@ class TextClassificationPipeline extends (/** @type {new (options: TextPipelineC
25038
25160
  * // { entity: 'B-LOC', score: 0.9994474053382874, index: 9, word: 'London' }
25039
25161
  * // ]
25040
25162
  * ```
25041
- *
25163
+ *
25042
25164
  * **Example:** Perform named entity recognition with `Xenova/bert-base-NER` (and return all labels).
25043
25165
  * ```javascript
25044
25166
  * const classifier = await pipeline('token-classification', 'Xenova/bert-base-NER');
@@ -25134,22 +25256,22 @@ class TokenClassificationPipeline extends (/** @type {new (options: TextPipeline
25134
25256
  * @property {number} [start] The character start index of the answer (in the tokenized version of the input).
25135
25257
  * @property {number} [end] The character end index of the answer (in the tokenized version of the input).
25136
25258
  * @property {string} answer The answer to the question.
25137
- *
25259
+ *
25138
25260
  * @typedef {Object} QuestionAnsweringPipelineOptions Parameters specific to question answering pipelines.
25139
25261
  * @property {number} [top_k=1] The number of top answer predictions to be returned.
25140
- *
25262
+ *
25141
25263
  * @callback QuestionAnsweringPipelineCallback Answer the question(s) given as inputs by using the context(s).
25142
25264
  * @param {string|string[]} question One or several question(s) (must be used in conjunction with the `context` argument).
25143
25265
  * @param {string|string[]} context One or several context(s) associated with the question(s) (must be used in conjunction with the `question` argument).
25144
25266
  * @param {QuestionAnsweringPipelineOptions} [options] The options to use for question answering.
25145
25267
  * @returns {Promise<QuestionAnsweringOutput|QuestionAnsweringOutput[]>} An array or object containing the predicted answers and scores.
25146
- *
25268
+ *
25147
25269
  * @typedef {TextPipelineConstructorArgs & QuestionAnsweringPipelineCallback & Disposable} QuestionAnsweringPipelineType
25148
25270
  */
25149
25271
 
25150
25272
  /**
25151
25273
  * Question Answering pipeline using any `ModelForQuestionAnswering`.
25152
- *
25274
+ *
25153
25275
  * **Example:** Run question answering with `Xenova/distilbert-base-uncased-distilled-squad`.
25154
25276
  * ```javascript
25155
25277
  * const answerer = await pipeline('question-answering', 'Xenova/distilbert-base-uncased-distilled-squad');
@@ -25274,10 +25396,10 @@ class QuestionAnsweringPipeline extends (/** @type {new (options: TextPipelineCo
25274
25396
  * @property {number} token The predicted token id (to replace the masked one).
25275
25397
  * @property {string} token_str The predicted token (to replace the masked one).
25276
25398
  * @typedef {FillMaskSingle[]} FillMaskOutput
25277
- *
25399
+ *
25278
25400
  * @typedef {Object} FillMaskPipelineOptions Parameters specific to fill mask pipelines.
25279
25401
  * @property {number} [top_k=5] When passed, overrides the number of predictions to return.
25280
- *
25402
+ *
25281
25403
  * @callback FillMaskPipelineCallback Fill the masked token in the text(s) given as inputs.
25282
25404
  * @param {string|string[]} texts One or several texts (or one list of prompts) with masked tokens.
25283
25405
  * @param {FillMaskPipelineOptions} [options] The options to use for masked language modelling.
@@ -25285,13 +25407,13 @@ class QuestionAnsweringPipeline extends (/** @type {new (options: TextPipelineCo
25285
25407
  * and the sequence with the predicted token filled in, or an array of such arrays (one for each input text).
25286
25408
  * If only one input text is given, the output will be an array of objects.
25287
25409
  * @throws {Error} When the mask token is not found in the input text.
25288
- *
25410
+ *
25289
25411
  * @typedef {TextPipelineConstructorArgs & FillMaskPipelineCallback & Disposable} FillMaskPipelineType
25290
25412
  */
25291
25413
 
25292
25414
  /**
25293
25415
  * Masked language modeling prediction pipeline using any `ModelWithLMHead`.
25294
- *
25416
+ *
25295
25417
  * **Example:** Perform masked language modelling (a.k.a. "fill-mask") with `Xenova/bert-base-uncased`.
25296
25418
  * ```javascript
25297
25419
  * const unmasker = await pipeline('fill-mask', 'Xenova/bert-base-cased');
@@ -25304,7 +25426,7 @@ class QuestionAnsweringPipeline extends (/** @type {new (options: TextPipelineCo
25304
25426
  * // { token_str: 'life', score: 0.01859794743359089, token: 1297, sequence: 'The goal of life is life.' }
25305
25427
  * // ]
25306
25428
  * ```
25307
- *
25429
+ *
25308
25430
  * **Example:** Perform masked language modelling (a.k.a. "fill-mask") with `Xenova/bert-base-cased` (and return top result).
25309
25431
  * ```javascript
25310
25432
  * const unmasker = await pipeline('fill-mask', 'Xenova/bert-base-cased');
@@ -25381,18 +25503,18 @@ class FillMaskPipeline extends (/** @type {new (options: TextPipelineConstructor
25381
25503
  * @typedef {Object} Text2TextGenerationSingle
25382
25504
  * @property {string} generated_text The generated text.
25383
25505
  * @typedef {Text2TextGenerationSingle[]} Text2TextGenerationOutput
25384
- *
25506
+ *
25385
25507
  * @callback Text2TextGenerationPipelineCallback Generate the output text(s) using text(s) given as inputs.
25386
25508
  * @param {string|string[]} texts Input text for the encoder.
25387
25509
  * @param {Partial<import('./generation/configuration_utils.js').GenerationConfig>} [options] Additional keyword arguments to pass along to the generate method of the model.
25388
25510
  * @returns {Promise<Text2TextGenerationOutput|Text2TextGenerationOutput[]>}
25389
- *
25511
+ *
25390
25512
  * @typedef {TextPipelineConstructorArgs & Text2TextGenerationPipelineCallback & Disposable} Text2TextGenerationPipelineType
25391
25513
  */
25392
25514
 
25393
25515
  /**
25394
25516
  * Text2TextGenerationPipeline class for generating text using a model that performs text-to-text generation tasks.
25395
- *
25517
+ *
25396
25518
  * **Example:** Text-to-text generation w/ `Xenova/LaMini-Flan-T5-783M`.
25397
25519
  * ```javascript
25398
25520
  * const generator = await pipeline('text2text-generation', 'Xenova/LaMini-Flan-T5-783M');
@@ -25468,18 +25590,18 @@ class Text2TextGenerationPipeline extends (/** @type {new (options: TextPipeline
25468
25590
  * @typedef {Object} SummarizationSingle
25469
25591
  * @property {string} summary_text The summary text.
25470
25592
  * @typedef {SummarizationSingle[]} SummarizationOutput
25471
- *
25593
+ *
25472
25594
  * @callback SummarizationPipelineCallback Summarize the text(s) given as inputs.
25473
25595
  * @param {string|string[]} texts One or several articles (or one list of articles) to summarize.
25474
25596
  * @param {import('./generation/configuration_utils.js').GenerationConfig} [options] Additional keyword arguments to pass along to the generate method of the model.
25475
25597
  * @returns {Promise<SummarizationOutput|SummarizationOutput[]>}
25476
- *
25598
+ *
25477
25599
  * @typedef {TextPipelineConstructorArgs & SummarizationPipelineCallback & Disposable} SummarizationPipelineType
25478
25600
  */
25479
25601
 
25480
25602
  /**
25481
25603
  * A pipeline for summarization tasks, inheriting from Text2TextGenerationPipeline.
25482
- *
25604
+ *
25483
25605
  * **Example:** Summarization w/ `Xenova/distilbart-cnn-6-6`.
25484
25606
  * ```javascript
25485
25607
  * const generator = await pipeline('summarization', 'Xenova/distilbart-cnn-6-6');
@@ -25515,23 +25637,23 @@ class SummarizationPipeline extends (/** @type {new (options: TextPipelineConstr
25515
25637
  * @typedef {Object} TranslationSingle
25516
25638
  * @property {string} translation_text The translated text.
25517
25639
  * @typedef {TranslationSingle[]} TranslationOutput
25518
- *
25640
+ *
25519
25641
  * @callback TranslationPipelineCallback Translate the text(s) given as inputs.
25520
25642
  * @param {string|string[]} texts Texts to be translated.
25521
25643
  * @param {import('./generation/configuration_utils.js').GenerationConfig} [options] Additional keyword arguments to pass along to the generate method of the model.
25522
25644
  * @returns {Promise<TranslationOutput|TranslationOutput[]>}
25523
- *
25645
+ *
25524
25646
  * @typedef {TextPipelineConstructorArgs & TranslationPipelineCallback & Disposable} TranslationPipelineType
25525
25647
  */
25526
25648
 
25527
25649
  /**
25528
25650
  * Translates text from one language to another.
25529
- *
25651
+ *
25530
25652
  * **Example:** Multilingual translation w/ `Xenova/nllb-200-distilled-600M`.
25531
- *
25653
+ *
25532
25654
  * See [here](https://github.com/facebookresearch/flores/blob/main/flores200/README.md#languages-in-flores-200)
25533
25655
  * for the full list of languages and their corresponding codes.
25534
- *
25656
+ *
25535
25657
  * ```javascript
25536
25658
  * const translator = await pipeline('translation', 'Xenova/nllb-200-distilled-600M');
25537
25659
  * const output = await translator('जीवन एक चॉकलेट बॉक्स की तरह है।', {
@@ -25540,12 +25662,12 @@ class SummarizationPipeline extends (/** @type {new (options: TextPipelineConstr
25540
25662
  * });
25541
25663
  * // [{ translation_text: 'La vie est comme une boîte à chocolat.' }]
25542
25664
  * ```
25543
- *
25665
+ *
25544
25666
  * **Example:** Multilingual translation w/ `Xenova/m2m100_418M`.
25545
- *
25667
+ *
25546
25668
  * See [here](https://huggingface.co/facebook/m2m100_418M#languages-covered)
25547
25669
  * for the full list of languages and their corresponding codes.
25548
- *
25670
+ *
25549
25671
  * ```javascript
25550
25672
  * const translator = await pipeline('translation', 'Xenova/m2m100_418M');
25551
25673
  * const output = await translator('生活就像一盒巧克力。', {
@@ -25554,12 +25676,12 @@ class SummarizationPipeline extends (/** @type {new (options: TextPipelineConstr
25554
25676
  * });
25555
25677
  * // [{ translation_text: 'Life is like a box of chocolate.' }]
25556
25678
  * ```
25557
- *
25679
+ *
25558
25680
  * **Example:** Multilingual translation w/ `Xenova/mbart-large-50-many-to-many-mmt`.
25559
- *
25681
+ *
25560
25682
  * See [here](https://huggingface.co/facebook/mbart-large-50-many-to-many-mmt#languages-covered)
25561
25683
  * for the full list of languages and their corresponding codes.
25562
- *
25684
+ *
25563
25685
  * ```javascript
25564
25686
  * const translator = await pipeline('translation', 'Xenova/mbart-large-50-many-to-many-mmt');
25565
25687
  * const output = await translator('संयुक्त राष्ट्र के प्रमुख का कहना है कि सीरिया में कोई सैन्य समाधान नहीं है', {
@@ -25588,21 +25710,21 @@ function isChat(x) {
25588
25710
 
25589
25711
  /**
25590
25712
  * @typedef {import('./tokenizers.js').Message[]} Chat
25591
- *
25713
+ *
25592
25714
  * @typedef {Object} TextGenerationSingle
25593
25715
  * @property {string|Chat} generated_text The generated text.
25594
25716
  * @typedef {TextGenerationSingle[]} TextGenerationOutput
25595
- *
25717
+ *
25596
25718
  * @typedef {Object} TextGenerationSpecificParams Parameters specific to text-generation pipelines.
25597
25719
  * @property {boolean} [add_special_tokens] Whether or not to add special tokens when tokenizing the sequences.
25598
25720
  * @property {boolean} [return_full_text=true] If set to `false` only added text is returned, otherwise the full text is returned.
25599
25721
  * @typedef {import('./generation/configuration_utils.js').GenerationConfig & TextGenerationSpecificParams} TextGenerationConfig
25600
- *
25722
+ *
25601
25723
  * @callback TextGenerationPipelineCallback Complete the prompt(s) given as inputs.
25602
25724
  * @param {string|string[]|Chat|Chat[]} texts One or several prompts (or one list of prompts) to complete.
25603
25725
  * @param {Partial<TextGenerationConfig>} [options] Additional keyword arguments to pass along to the generate method of the model.
25604
25726
  * @returns {Promise<TextGenerationOutput|TextGenerationOutput[]>} An array or object containing the generated texts.
25605
- *
25727
+ *
25606
25728
  * @typedef {TextPipelineConstructorArgs & TextGenerationPipelineCallback & Disposable} TextGenerationPipelineType
25607
25729
  */
25608
25730
 
@@ -25610,7 +25732,7 @@ function isChat(x) {
25610
25732
  * Language generation pipeline using any `ModelWithLMHead` or `ModelForCausalLM`.
25611
25733
  * This pipeline predicts the words that will follow a specified text prompt.
25612
25734
  * NOTE: For the full list of generation parameters, see [`GenerationConfig`](./utils/generation#module_utils/generation.GenerationConfig).
25613
- *
25735
+ *
25614
25736
  * **Example:** Text generation with `Xenova/distilgpt2` (default settings).
25615
25737
  * ```javascript
25616
25738
  * const generator = await pipeline('text-generation', 'Xenova/distilgpt2');
@@ -25618,7 +25740,7 @@ function isChat(x) {
25618
25740
  * const output = await generator(text);
25619
25741
  * // [{ generated_text: "I enjoy walking with my cute dog, and I love to play with the other dogs." }]
25620
25742
  * ```
25621
- *
25743
+ *
25622
25744
  * **Example:** Text generation with `Xenova/distilgpt2` (custom settings).
25623
25745
  * ```javascript
25624
25746
  * const generator = await pipeline('text-generation', 'Xenova/distilgpt2');
@@ -25637,7 +25759,7 @@ function isChat(x) {
25637
25759
  * // "generated_text": "Once upon a time, there was an abundance of information about the most important and influential"
25638
25760
  * // }]
25639
25761
  * ```
25640
- *
25762
+ *
25641
25763
  * **Example:** Run code generation with `Xenova/codegen-350M-mono`.
25642
25764
  * ```javascript
25643
25765
  * const generator = await pipeline('text-generation', 'Xenova/codegen-350M-mono');
@@ -25756,7 +25878,7 @@ class TextGenerationPipeline extends (/** @type {new (options: TextPipelineConst
25756
25878
  * @property {string} sequence The sequence for which this is the output.
25757
25879
  * @property {string[]} labels The labels sorted by order of likelihood.
25758
25880
  * @property {number[]} scores The probabilities for each of the labels.
25759
- *
25881
+ *
25760
25882
  * @typedef {Object} ZeroShotClassificationPipelineOptions Parameters specific to zero-shot classification pipelines.
25761
25883
  * @property {string} [hypothesis_template="This example is {}."] The template used to turn each
25762
25884
  * candidate label into an NLI-style hypothesis. The candidate label will replace the {} placeholder.
@@ -25764,14 +25886,14 @@ class TextGenerationPipeline extends (/** @type {new (options: TextPipelineConst
25764
25886
  * If `false`, the scores are normalized such that the sum of the label likelihoods for each sequence
25765
25887
  * is 1. If `true`, the labels are considered independent and probabilities are normalized for each
25766
25888
  * candidate by doing a softmax of the entailment score vs. the contradiction score.
25767
- *
25889
+ *
25768
25890
  * @callback ZeroShotClassificationPipelineCallback Classify the sequence(s) given as inputs.
25769
25891
  * @param {string|string[]} texts The sequence(s) to classify, will be truncated if the model input is too large.
25770
25892
  * @param {string|string[]} candidate_labels The set of possible class labels to classify each sequence into.
25771
25893
  * Can be a single label, a string of comma-separated labels, or a list of labels.
25772
25894
  * @param {ZeroShotClassificationPipelineOptions} [options] The options to use for zero-shot classification.
25773
25895
  * @returns {Promise<ZeroShotClassificationOutput|ZeroShotClassificationOutput[]>} An array or object containing the predicted labels and scores.
25774
- *
25896
+ *
25775
25897
  * @typedef {TextPipelineConstructorArgs & ZeroShotClassificationPipelineCallback & Disposable} ZeroShotClassificationPipelineType
25776
25898
  */
25777
25899
 
@@ -25780,7 +25902,7 @@ class TextGenerationPipeline extends (/** @type {new (options: TextPipelineConst
25780
25902
  * trained on NLI (natural language inference) tasks. Equivalent of `text-classification`
25781
25903
  * pipelines, but these models don't require a hardcoded number of potential classes, they
25782
25904
  * can be chosen at runtime. It usually means it's slower but it is **much** more flexible.
25783
- *
25905
+ *
25784
25906
  * **Example:** Zero shot classification with `Xenova/mobilebert-uncased-mnli`.
25785
25907
  * ```javascript
25786
25908
  * const classifier = await pipeline('zero-shot-classification', 'Xenova/mobilebert-uncased-mnli');
@@ -25793,7 +25915,7 @@ class TextGenerationPipeline extends (/** @type {new (options: TextPipelineConst
25793
25915
  * // scores: [ 0.5562091040482018, 0.1843621307860853, 0.13942646639336376, 0.12000229877234923 ]
25794
25916
  * // }
25795
25917
  * ```
25796
- *
25918
+ *
25797
25919
  * **Example:** Zero shot classification with `Xenova/nli-deberta-v3-xsmall` (multi-label).
25798
25920
  * ```javascript
25799
25921
  * const classifier = await pipeline('zero-shot-classification', 'Xenova/nli-deberta-v3-xsmall');
@@ -25907,20 +26029,20 @@ class ZeroShotClassificationPipeline extends (/** @type {new (options: TextPipel
25907
26029
  * @property {'none'|'mean'|'cls'} [pooling="none"] The pooling method to use.
25908
26030
  * @property {boolean} [normalize=false] Whether or not to normalize the embeddings in the last dimension.
25909
26031
  * @property {boolean} [quantize=false] Whether or not to quantize the embeddings.
25910
- * @property {'binary'|'ubinary'} [precision='binary'] The precision to use for quantization.
25911
- *
26032
+ * @property {'binary'|'ubinary'} [precision='binary'] The precision to use for quantization.
26033
+ *
25912
26034
  * @callback FeatureExtractionPipelineCallback Extract the features of the input(s).
25913
26035
  * @param {string|string[]} texts One or several texts (or one list of texts) to get the features of.
25914
26036
  * @param {FeatureExtractionPipelineOptions} [options] The options to use for feature extraction.
25915
26037
  * @returns {Promise<Tensor>} The features computed by the model.
25916
- *
26038
+ *
25917
26039
  * @typedef {TextPipelineConstructorArgs & FeatureExtractionPipelineCallback & Disposable} FeatureExtractionPipelineType
25918
26040
  */
25919
26041
 
25920
26042
  /**
25921
26043
  * Feature extraction pipeline using no model head. This pipeline extracts the hidden
25922
26044
  * states from the base transformer, which can be used as features in downstream tasks.
25923
- *
26045
+ *
25924
26046
  * **Example:** Run feature extraction with `bert-base-uncased` (without pooling/normalization).
25925
26047
  * ```javascript
25926
26048
  * const extractor = await pipeline('feature-extraction', 'Xenova/bert-base-uncased', { revision: 'default' });
@@ -25931,7 +26053,7 @@ class ZeroShotClassificationPipeline extends (/** @type {new (options: TextPipel
25931
26053
  * // dims: [1, 8, 768]
25932
26054
  * // }
25933
26055
  * ```
25934
- *
26056
+ *
25935
26057
  * **Example:** Run feature extraction with `bert-base-uncased` (with pooling/normalization).
25936
26058
  * ```javascript
25937
26059
  * const extractor = await pipeline('feature-extraction', 'Xenova/bert-base-uncased', { revision: 'default' });
@@ -25942,7 +26064,7 @@ class ZeroShotClassificationPipeline extends (/** @type {new (options: TextPipel
25942
26064
  * // dims: [1, 768]
25943
26065
  * // }
25944
26066
  * ```
25945
- *
26067
+ *
25946
26068
  * **Example:** Calculating embeddings with `sentence-transformers` models.
25947
26069
  * ```javascript
25948
26070
  * const extractor = await pipeline('feature-extraction', 'Xenova/all-MiniLM-L6-v2');
@@ -26023,19 +26145,19 @@ class FeatureExtractionPipeline extends (/** @type {new (options: TextPipelineCo
26023
26145
  /**
26024
26146
  * @typedef {Object} ImageFeatureExtractionPipelineOptions Parameters specific to image feature extraction pipelines.
26025
26147
  * @property {boolean} [pool=null] Whether or not to return the pooled output. If set to `false`, the model will return the raw hidden states.
26026
- *
26148
+ *
26027
26149
  * @callback ImageFeatureExtractionPipelineCallback Extract the features of the input(s).
26028
26150
  * @param {ImagePipelineInputs} images One or several images (or one list of images) to get the features of.
26029
26151
  * @param {ImageFeatureExtractionPipelineOptions} [options] The options to use for image feature extraction.
26030
26152
  * @returns {Promise<Tensor>} The image features computed by the model.
26031
- *
26153
+ *
26032
26154
  * @typedef {ImagePipelineConstructorArgs & ImageFeatureExtractionPipelineCallback & Disposable} ImageFeatureExtractionPipelineType
26033
26155
  */
26034
26156
 
26035
26157
  /**
26036
26158
  * Image feature extraction pipeline using no model head. This pipeline extracts the hidden
26037
26159
  * states from the base transformer, which can be used as features in downstream tasks.
26038
- *
26160
+ *
26039
26161
  * **Example:** Perform image feature extraction with `Xenova/vit-base-patch16-224-in21k`.
26040
26162
  * ```javascript
26041
26163
  * const image_feature_extractor = await pipeline('image-feature-extraction', 'Xenova/vit-base-patch16-224-in21k');
@@ -26048,7 +26170,7 @@ class FeatureExtractionPipeline extends (/** @type {new (options: TextPipelineCo
26048
26170
  * // size: 151296
26049
26171
  * // }
26050
26172
  * ```
26051
- *
26173
+ *
26052
26174
  * **Example:** Compute image embeddings with `Xenova/clip-vit-base-patch32`.
26053
26175
  * ```javascript
26054
26176
  * const image_feature_extractor = await pipeline('image-feature-extraction', 'Xenova/clip-vit-base-patch32');
@@ -26104,12 +26226,12 @@ class ImageFeatureExtractionPipeline extends (/** @type {new (options: ImagePipe
26104
26226
  * @property {string} label The label predicted.
26105
26227
  * @property {number} score The corresponding probability.
26106
26228
  * @typedef {AudioClassificationSingle[]} AudioClassificationOutput
26107
- *
26229
+ *
26108
26230
  * @typedef {Object} AudioClassificationPipelineOptions Parameters specific to audio classification pipelines.
26109
26231
  * @property {number} [top_k=5] The number of top labels that will be returned by the pipeline.
26110
26232
  * If the provided number is `null` or higher than the number of labels available in the model configuration,
26111
26233
  * it will default to the number of labels.
26112
- *
26234
+ *
26113
26235
  * @callback AudioClassificationPipelineCallback Classify the sequence(s) given as inputs.
26114
26236
  * @param {AudioPipelineInputs} audio The input audio file(s) to be classified. The input is either:
26115
26237
  * - `string` or `URL` that is the filename/URL of the audio file, the file will be read at the processor's sampling rate
@@ -26118,14 +26240,14 @@ class ImageFeatureExtractionPipeline extends (/** @type {new (options: ImagePipe
26118
26240
  * - `Float32Array` or `Float64Array` of shape `(n, )`, representing the raw audio at the correct sampling rate (no further check will be done).
26119
26241
  * @param {AudioClassificationPipelineOptions} [options] The options to use for audio classification.
26120
26242
  * @returns {Promise<AudioClassificationOutput|AudioClassificationOutput[]>} An array or object containing the predicted labels and scores.
26121
- *
26243
+ *
26122
26244
  * @typedef {AudioPipelineConstructorArgs & AudioClassificationPipelineCallback & Disposable} AudioClassificationPipelineType
26123
26245
  */
26124
26246
 
26125
26247
  /**
26126
26248
  * Audio classification pipeline using any `AutoModelForAudioClassification`.
26127
26249
  * This pipeline predicts the class of a raw waveform or an audio file.
26128
- *
26250
+ *
26129
26251
  * **Example:** Perform audio classification with `Xenova/wav2vec2-large-xlsr-53-gender-recognition-librispeech`.
26130
26252
  * ```javascript
26131
26253
  * const classifier = await pipeline('audio-classification', 'Xenova/wav2vec2-large-xlsr-53-gender-recognition-librispeech');
@@ -26136,7 +26258,7 @@ class ImageFeatureExtractionPipeline extends (/** @type {new (options: ImagePipe
26136
26258
  * // { label: 'female', score: 0.001845747814513743 }
26137
26259
  * // ]
26138
26260
  * ```
26139
- *
26261
+ *
26140
26262
  * **Example:** Perform audio classification with `Xenova/ast-finetuned-audioset-10-10-0.4593` and return top 4 results.
26141
26263
  * ```javascript
26142
26264
  * const classifier = await pipeline('audio-classification', 'Xenova/ast-finetuned-audioset-10-10-0.4593');
@@ -26201,12 +26323,12 @@ class AudioClassificationPipeline extends (/** @type {new (options: AudioPipelin
26201
26323
  * @typedef {Object} ZeroShotAudioClassificationOutput
26202
26324
  * @property {string} label The label identified by the model. It is one of the suggested `candidate_label`.
26203
26325
  * @property {number} score The score attributed by the model for that label (between 0 and 1).
26204
- *
26326
+ *
26205
26327
  * @typedef {Object} ZeroShotAudioClassificationPipelineOptions Parameters specific to zero-shot audio classification pipelines.
26206
26328
  * @property {string} [hypothesis_template="This is a sound of {}."] The sentence used in conjunction with `candidate_labels`
26207
26329
  * to attempt the audio classification by replacing the placeholder with the candidate_labels.
26208
26330
  * Then likelihood is estimated by using `logits_per_audio`.
26209
- *
26331
+ *
26210
26332
  * @callback ZeroShotAudioClassificationPipelineCallback Classify the sequence(s) given as inputs.
26211
26333
  * @param {AudioPipelineInputs} audio The input audio file(s) to be classified. The input is either:
26212
26334
  * - `string` or `URL` that is the filename/URL of the audio file, the file will be read at the processor's sampling rate
@@ -26216,14 +26338,14 @@ class AudioClassificationPipeline extends (/** @type {new (options: AudioPipelin
26216
26338
  * @param {string[]} candidate_labels The candidate labels for this audio.
26217
26339
  * @param {ZeroShotAudioClassificationPipelineOptions} [options] The options to use for zero-shot audio classification.
26218
26340
  * @returns {Promise<ZeroShotAudioClassificationOutput[]|ZeroShotAudioClassificationOutput[][]>} An array of objects containing the predicted labels and scores.
26219
- *
26341
+ *
26220
26342
  * @typedef {TextAudioPipelineConstructorArgs & ZeroShotAudioClassificationPipelineCallback & Disposable} ZeroShotAudioClassificationPipelineType
26221
26343
  */
26222
26344
 
26223
26345
  /**
26224
26346
  * Zero shot audio classification pipeline using `ClapModel`. This pipeline predicts the class of an audio when you
26225
26347
  * provide an audio and a set of `candidate_labels`.
26226
- *
26348
+ *
26227
26349
  * **Example**: Perform zero-shot audio classification with `Xenova/clap-htsat-unfused`.
26228
26350
  * ```javascript
26229
26351
  * const classifier = await pipeline('zero-shot-audio-classification', 'Xenova/clap-htsat-unfused');
@@ -26256,7 +26378,7 @@ class ZeroShotAudioClassificationPipeline extends (/** @type {new (options: Text
26256
26378
  audio = [/** @type {AudioInput} */ (audio)];
26257
26379
  }
26258
26380
 
26259
- // Insert label into hypothesis template
26381
+ // Insert label into hypothesis template
26260
26382
  const texts = candidate_labels.map(
26261
26383
  x => hypothesis_template.replace('{}', x)
26262
26384
  );
@@ -26300,7 +26422,7 @@ class ZeroShotAudioClassificationPipeline extends (/** @type {new (options: Text
26300
26422
  * @property {string} text The recognized text.
26301
26423
  * @property {Chunk[]} [chunks] When using `return_timestamps`, the `chunks` will become a list
26302
26424
  * containing all the various text chunks identified by the model.
26303
- *
26425
+ *
26304
26426
  * @typedef {Object} AutomaticSpeechRecognitionSpecificParams Parameters specific to automatic-speech-recognition pipelines.
26305
26427
  * @property {boolean|'word'} [return_timestamps] Whether to return timestamps or not. Default is `false`.
26306
26428
  * @property {number} [chunk_length_s] The length of audio chunks to process in seconds. Default is 0 (no chunking).
@@ -26310,7 +26432,7 @@ class ZeroShotAudioClassificationPipeline extends (/** @type {new (options: Text
26310
26432
  * @property {string} [task] The task to perform. Default is `null`, meaning it should be auto-detected.
26311
26433
  * @property {number} [num_frames] The number of frames in the input audio.
26312
26434
  * @typedef {import('./generation/configuration_utils.js').GenerationConfig & AutomaticSpeechRecognitionSpecificParams} AutomaticSpeechRecognitionConfig
26313
- *
26435
+ *
26314
26436
  * @callback AutomaticSpeechRecognitionPipelineCallback Transcribe the audio sequence(s) given as inputs to text.
26315
26437
  * @param {AudioPipelineInputs} audio The input audio file(s) to be transcribed. The input is either:
26316
26438
  * - `string` or `URL` that is the filename/URL of the audio file, the file will be read at the processor's sampling rate
@@ -26319,7 +26441,7 @@ class ZeroShotAudioClassificationPipeline extends (/** @type {new (options: Text
26319
26441
  * - `Float32Array` or `Float64Array` of shape `(n, )`, representing the raw audio at the correct sampling rate (no further check will be done).
26320
26442
  * @param {Partial<AutomaticSpeechRecognitionConfig>} [options] Additional keyword arguments to pass along to the generate method of the model.
26321
26443
  * @returns {Promise<AutomaticSpeechRecognitionOutput|AutomaticSpeechRecognitionOutput[]>} An object containing the transcription text and optionally timestamps if `return_timestamps` is `true`.
26322
- *
26444
+ *
26323
26445
  * @typedef {TextAudioPipelineConstructorArgs & AutomaticSpeechRecognitionPipelineCallback & Disposable} AutomaticSpeechRecognitionPipelineType
26324
26446
  */
26325
26447
 
@@ -26333,7 +26455,7 @@ class ZeroShotAudioClassificationPipeline extends (/** @type {new (options: Text
26333
26455
  * const output = await transcriber(url);
26334
26456
  * // { text: " And so my fellow Americans ask not what your country can do for you, ask what you can do for your country." }
26335
26457
  * ```
26336
- *
26458
+ *
26337
26459
  * **Example:** Transcribe English w/ timestamps.
26338
26460
  * ```javascript
26339
26461
  * const transcriber = await pipeline('automatic-speech-recognition', 'Xenova/whisper-tiny.en');
@@ -26347,7 +26469,7 @@ class ZeroShotAudioClassificationPipeline extends (/** @type {new (options: Text
26347
26469
  * // ]
26348
26470
  * // }
26349
26471
  * ```
26350
- *
26472
+ *
26351
26473
  * **Example:** Transcribe English w/ word-level timestamps.
26352
26474
  * ```javascript
26353
26475
  * const transcriber = await pipeline('automatic-speech-recognition', 'Xenova/whisper-tiny.en');
@@ -26366,7 +26488,7 @@ class ZeroShotAudioClassificationPipeline extends (/** @type {new (options: Text
26366
26488
  * // ]
26367
26489
  * // }
26368
26490
  * ```
26369
- *
26491
+ *
26370
26492
  * **Example:** Transcribe French.
26371
26493
  * ```javascript
26372
26494
  * const transcriber = await pipeline('automatic-speech-recognition', 'Xenova/whisper-small');
@@ -26374,7 +26496,7 @@ class ZeroShotAudioClassificationPipeline extends (/** @type {new (options: Text
26374
26496
  * const output = await transcriber(url, { language: 'french', task: 'transcribe' });
26375
26497
  * // { text: " J'adore, j'aime, je n'aime pas, je déteste." }
26376
26498
  * ```
26377
- *
26499
+ *
26378
26500
  * **Example:** Translate French to English.
26379
26501
  * ```javascript
26380
26502
  * const transcriber = await pipeline('automatic-speech-recognition', 'Xenova/whisper-small');
@@ -26382,7 +26504,7 @@ class ZeroShotAudioClassificationPipeline extends (/** @type {new (options: Text
26382
26504
  * const output = await transcriber(url, { language: 'french', task: 'translate' });
26383
26505
  * // { text: " I love, I like, I don't like, I hate." }
26384
26506
  * ```
26385
- *
26507
+ *
26386
26508
  * **Example:** Transcribe/translate audio longer than 30 seconds.
26387
26509
  * ```javascript
26388
26510
  * const transcriber = await pipeline('automatic-speech-recognition', 'Xenova/whisper-tiny.en');
@@ -26605,18 +26727,18 @@ class AutomaticSpeechRecognitionPipeline extends (/** @type {new (options: TextA
26605
26727
  * @typedef {Object} ImageToTextSingle
26606
26728
  * @property {string} generated_text The generated text.
26607
26729
  * @typedef {ImageToTextSingle[]} ImageToTextOutput
26608
- *
26730
+ *
26609
26731
  * @callback ImageToTextPipelineCallback Assign labels to the image(s) passed as inputs.
26610
26732
  * @param {ImagePipelineInputs} texts The images to be captioned.
26611
26733
  * @param {Partial<import('./generation/configuration_utils.js').GenerationConfig>} [options] Additional keyword arguments to pass along to the generate method of the model.
26612
26734
  * @returns {Promise<ImageToTextOutput|ImageToTextOutput[]>} An object (or array of objects) containing the generated text(s).
26613
- *
26735
+ *
26614
26736
  * @typedef {TextImagePipelineConstructorArgs & ImageToTextPipelineCallback & Disposable} ImageToTextPipelineType
26615
26737
  */
26616
26738
 
26617
26739
  /**
26618
26740
  * Image To Text pipeline using a `AutoModelForVision2Seq`. This pipeline predicts a caption for a given image.
26619
- *
26741
+ *
26620
26742
  * **Example:** Generate a caption for an image w/ `Xenova/vit-gpt2-image-captioning`.
26621
26743
  * ```javascript
26622
26744
  * const captioner = await pipeline('image-to-text', 'Xenova/vit-gpt2-image-captioning');
@@ -26624,7 +26746,7 @@ class AutomaticSpeechRecognitionPipeline extends (/** @type {new (options: TextA
26624
26746
  * const output = await captioner(url);
26625
26747
  * // [{ generated_text: 'a cat laying on a couch with another cat' }]
26626
26748
  * ```
26627
- *
26749
+ *
26628
26750
  * **Example:** Optical Character Recognition (OCR) w/ `Xenova/trocr-small-handwritten`.
26629
26751
  * ```javascript
26630
26752
  * const captioner = await pipeline('image-to-text', 'Xenova/trocr-small-handwritten');
@@ -26670,22 +26792,22 @@ class ImageToTextPipeline extends (/** @type {new (options: TextImagePipelineCon
26670
26792
  * @property {string} label The label identified by the model.
26671
26793
  * @property {number} score The score attributed by the model for that label.
26672
26794
  * @typedef {ImageClassificationSingle[]} ImageClassificationOutput
26673
- *
26795
+ *
26674
26796
  * @typedef {Object} ImageClassificationPipelineOptions Parameters specific to image classification pipelines.
26675
- * @property {number} [top_k=1] The number of top labels that will be returned by the pipeline.
26676
- *
26797
+ * @property {number} [top_k=1] The number of top labels that will be returned by the pipeline.
26798
+ *
26677
26799
  * @callback ImageClassificationPipelineCallback Assign labels to the image(s) passed as inputs.
26678
26800
  * @param {ImagePipelineInputs} images The input images(s) to be classified.
26679
26801
  * @param {ImageClassificationPipelineOptions} [options] The options to use for image classification.
26680
26802
  * @returns {Promise<ImageClassificationOutput|ImageClassificationOutput[]>} An array or object containing the predicted labels and scores.
26681
- *
26803
+ *
26682
26804
  * @typedef {ImagePipelineConstructorArgs & ImageClassificationPipelineCallback & Disposable} ImageClassificationPipelineType
26683
26805
  */
26684
26806
 
26685
26807
  /**
26686
26808
  * Image classification pipeline using any `AutoModelForImageClassification`.
26687
26809
  * This pipeline predicts the class of an image.
26688
- *
26810
+ *
26689
26811
  * **Example:** Classify an image.
26690
26812
  * ```javascript
26691
26813
  * const classifier = await pipeline('image-classification', 'Xenova/vit-base-patch16-224');
@@ -26695,7 +26817,7 @@ class ImageToTextPipeline extends (/** @type {new (options: TextImagePipelineCon
26695
26817
  * // { label: 'tiger, Panthera tigris', score: 0.632695734500885 },
26696
26818
  * // ]
26697
26819
  * ```
26698
- *
26820
+ *
26699
26821
  * **Example:** Classify an image and return top `n` classes.
26700
26822
  * ```javascript
26701
26823
  * const classifier = await pipeline('image-classification', 'Xenova/vit-base-patch16-224');
@@ -26707,7 +26829,7 @@ class ImageToTextPipeline extends (/** @type {new (options: TextImagePipelineCon
26707
26829
  * // { label: 'lion, king of beasts, Panthera leo', score: 0.00045060308184474707 },
26708
26830
  * // ]
26709
26831
  * ```
26710
- *
26832
+ *
26711
26833
  * **Example:** Classify an image and return all classes.
26712
26834
  * ```javascript
26713
26835
  * const classifier = await pipeline('image-classification', 'Xenova/vit-base-patch16-224');
@@ -26774,7 +26896,7 @@ class ImageClassificationPipeline extends (/** @type {new (options: ImagePipelin
26774
26896
  * @property {string|null} label The label of the segment.
26775
26897
  * @property {number|null} score The score of the segment.
26776
26898
  * @property {RawImage} mask The mask of the segment.
26777
- *
26899
+ *
26778
26900
  * @typedef {Object} ImageSegmentationPipelineOptions Parameters specific to image segmentation pipelines.
26779
26901
  * @property {number} [threshold=0.5] Probability threshold to filter out predicted masks.
26780
26902
  * @property {number} [mask_threshold=0.5] Threshold to use when turning the predicted masks into binary values.
@@ -26783,19 +26905,19 @@ class ImageClassificationPipeline extends (/** @type {new (options: ImagePipelin
26783
26905
  * depending on model capabilities. If not set, the pipeline will attempt to resolve (in that order).
26784
26906
  * @property {number[]} [label_ids_to_fuse=null] List of label ids to fuse. If not set, do not fuse any labels.
26785
26907
  * @property {number[][]} [target_sizes=null] List of target sizes for the input images. If not set, use the original image sizes.
26786
- *
26908
+ *
26787
26909
  * @callback ImageSegmentationPipelineCallback Segment the input images.
26788
26910
  * @param {ImagePipelineInputs} images The input images.
26789
26911
  * @param {ImageSegmentationPipelineOptions} [options] The options to use for image segmentation.
26790
26912
  * @returns {Promise<ImageSegmentationPipelineOutput[]>} The annotated segments.
26791
- *
26913
+ *
26792
26914
  * @typedef {ImagePipelineConstructorArgs & ImageSegmentationPipelineCallback & Disposable} ImageSegmentationPipelineType
26793
26915
  */
26794
26916
 
26795
26917
  /**
26796
26918
  * Image segmentation pipeline using any `AutoModelForXXXSegmentation`.
26797
26919
  * This pipeline predicts masks of objects and their classes.
26798
- *
26920
+ *
26799
26921
  * **Example:** Perform image segmentation with `Xenova/detr-resnet-50-panoptic`.
26800
26922
  * ```javascript
26801
26923
  * const segmenter = await pipeline('image-segmentation', 'Xenova/detr-resnet-50-panoptic');
@@ -26879,12 +27001,17 @@ class ImageSegmentationPipeline extends (/** @type {new (options: ImagePipelineC
26879
27001
  /** @type {ImageSegmentationPipelineOutput[]} */
26880
27002
  const annotation = [];
26881
27003
  if (!subtask) {
27004
+ // We define an epsilon to safeguard against numerical/precision issues when detecting
27005
+ // the normalization mode of the output (i.e., sigmoid already applied, or not).
27006
+ // See https://github.com/microsoft/onnxruntime/issues/23943 for more information.
27007
+ const epsilon = 1e-5;
27008
+
26882
27009
  // Perform standard image segmentation
26883
27010
  const result = output[outputNames[0]];
26884
27011
  for (let i = 0; i < imageSizes.length; ++i) {
26885
27012
  const size = imageSizes[i];
26886
27013
  const item = result[i];
26887
- if (item.data.some(x => x < 0 || x > 1)) {
27014
+ if (item.data.some(x => x < -epsilon || x > 1 + epsilon)) {
26888
27015
  item.sigmoid_();
26889
27016
  }
26890
27017
  const mask = await _utils_image_js__WEBPACK_IMPORTED_MODULE_9__.RawImage.fromTensor(item.mul_(255).to('uint8')).resize(size[1], size[0]);
@@ -26953,19 +27080,19 @@ class ImageSegmentationPipeline extends (/** @type {new (options: ImagePipelineC
26953
27080
 
26954
27081
  /**
26955
27082
  * @typedef {Object} BackgroundRemovalPipelineOptions Parameters specific to image segmentation pipelines.
26956
- *
27083
+ *
26957
27084
  * @callback BackgroundRemovalPipelineCallback Segment the input images.
26958
27085
  * @param {ImagePipelineInputs} images The input images.
26959
27086
  * @param {BackgroundRemovalPipelineOptions} [options] The options to use for image segmentation.
26960
27087
  * @returns {Promise<RawImage[]>} The images with the background removed.
26961
- *
27088
+ *
26962
27089
  * @typedef {ImagePipelineConstructorArgs & BackgroundRemovalPipelineCallback & Disposable} BackgroundRemovalPipelineType
26963
27090
  */
26964
27091
 
26965
27092
  /**
26966
27093
  * Background removal pipeline using certain `AutoModelForXXXSegmentation`.
26967
27094
  * This pipeline removes the backgrounds of images.
26968
- *
27095
+ *
26969
27096
  * **Example:** Perform background removal with `Xenova/modnet`.
26970
27097
  * ```javascript
26971
27098
  * const segmenter = await pipeline('background-removal', 'Xenova/modnet');
@@ -26976,7 +27103,7 @@ class ImageSegmentationPipeline extends (/** @type {new (options: ImagePipelineC
26976
27103
  * // ]
26977
27104
  * ```
26978
27105
  */
26979
- class BackgroundRemovalPipeline extends (/** @type {new (options: ImagePipelineConstructorArgs) => ImageSegmentationPipelineType} */ (ImageSegmentationPipeline)) {
27106
+ class BackgroundRemovalPipeline extends (/** @type {new (options: ImagePipelineConstructorArgs) => BackgroundRemovalPipelineType} */ (/** @type {any} */(ImageSegmentationPipeline))) {
26980
27107
  /**
26981
27108
  * Create a new BackgroundRemovalPipeline.
26982
27109
  * @param {ImagePipelineConstructorArgs} options An object used to instantiate the pipeline.
@@ -27011,25 +27138,25 @@ class BackgroundRemovalPipeline extends (/** @type {new (options: ImagePipelineC
27011
27138
  * @typedef {Object} ZeroShotImageClassificationOutput
27012
27139
  * @property {string} label The label identified by the model. It is one of the suggested `candidate_label`.
27013
27140
  * @property {number} score The score attributed by the model for that label (between 0 and 1).
27014
- *
27141
+ *
27015
27142
  * @typedef {Object} ZeroShotImageClassificationPipelineOptions Parameters specific to zero-shot image classification pipelines.
27016
27143
  * @property {string} [hypothesis_template="This is a photo of {}"] The sentence used in conjunction with `candidate_labels`
27017
27144
  * to attempt the image classification by replacing the placeholder with the candidate_labels.
27018
27145
  * Then likelihood is estimated by using `logits_per_image`.
27019
- *
27146
+ *
27020
27147
  * @callback ZeroShotImageClassificationPipelineCallback Assign labels to the image(s) passed as inputs.
27021
27148
  * @param {ImagePipelineInputs} images The input images.
27022
27149
  * @param {string[]} candidate_labels The candidate labels for this image.
27023
27150
  * @param {ZeroShotImageClassificationPipelineOptions} [options] The options to use for zero-shot image classification.
27024
27151
  * @returns {Promise<ZeroShotImageClassificationOutput[]|ZeroShotImageClassificationOutput[][]>} An array of objects containing the predicted labels and scores.
27025
- *
27152
+ *
27026
27153
  * @typedef {TextImagePipelineConstructorArgs & ZeroShotImageClassificationPipelineCallback & Disposable} ZeroShotImageClassificationPipelineType
27027
27154
  */
27028
27155
 
27029
27156
  /**
27030
27157
  * Zero shot image classification pipeline. This pipeline predicts the class of
27031
27158
  * an image when you provide an image and a set of `candidate_labels`.
27032
- *
27159
+ *
27033
27160
  * **Example:** Zero shot image classification w/ `Xenova/clip-vit-base-patch32`.
27034
27161
  * ```javascript
27035
27162
  * const classifier = await pipeline('zero-shot-image-classification', 'Xenova/clip-vit-base-patch32');
@@ -27059,7 +27186,7 @@ class ZeroShotImageClassificationPipeline extends (/** @type {new (options: Text
27059
27186
  const isBatched = Array.isArray(images);
27060
27187
  const preparedImages = await prepareImages(images);
27061
27188
 
27062
- // Insert label into hypothesis template
27189
+ // Insert label into hypothesis template
27063
27190
  const texts = candidate_labels.map(
27064
27191
  x => hypothesis_template.replace('{}', x)
27065
27192
  );
@@ -27106,23 +27233,23 @@ class ZeroShotImageClassificationPipeline extends (/** @type {new (options: Text
27106
27233
  * @property {number} score The score attributed by the model for that label.
27107
27234
  * @property {BoundingBox} box The bounding box of detected object in image's original size, or as a percentage if `percentage` is set to true.
27108
27235
  * @typedef {ObjectDetectionPipelineSingle[]} ObjectDetectionPipelineOutput
27109
- *
27236
+ *
27110
27237
  * @typedef {Object} ObjectDetectionPipelineOptions Parameters specific to object detection pipelines.
27111
27238
  * @property {number} [threshold=0.9] The threshold used to filter boxes by score.
27112
27239
  * @property {boolean} [percentage=false] Whether to return the boxes coordinates in percentage (true) or in pixels (false).
27113
- *
27240
+ *
27114
27241
  * @callback ObjectDetectionPipelineCallback Detect objects (bounding boxes & classes) in the image(s) passed as inputs.
27115
27242
  * @param {ImagePipelineInputs} images The input images.
27116
27243
  * @param {ObjectDetectionPipelineOptions} [options] The options to use for object detection.
27117
- * @returns {Promise<ObjectDetectionPipelineOutput|ObjectDetectionPipelineOutput[]>} A list of objects or a list of list of objects.
27118
- *
27244
+ * @returns {Promise<ObjectDetectionPipelineOutput|ObjectDetectionPipelineOutput[]>} A list of objects or a list of list of objects.
27245
+ *
27119
27246
  * @typedef {ImagePipelineConstructorArgs & ObjectDetectionPipelineCallback & Disposable} ObjectDetectionPipelineType
27120
27247
  */
27121
27248
 
27122
27249
  /**
27123
27250
  * Object detection pipeline using any `AutoModelForObjectDetection`.
27124
27251
  * This pipeline predicts bounding boxes of objects and their classes.
27125
- *
27252
+ *
27126
27253
  * **Example:** Run object-detection with `Xenova/detr-resnet-50`.
27127
27254
  * ```javascript
27128
27255
  * const detector = await pipeline('object-detection', 'Xenova/detr-resnet-50');
@@ -27196,27 +27323,27 @@ class ObjectDetectionPipeline extends (/** @type {new (options: ImagePipelineCon
27196
27323
  * @property {string} label Text query corresponding to the found object.
27197
27324
  * @property {number} score Score corresponding to the object (between 0 and 1).
27198
27325
  * @property {BoundingBox} box Bounding box of the detected object in image's original size, or as a percentage if `percentage` is set to true.
27199
- *
27326
+ *
27200
27327
  * @typedef {Object} ZeroShotObjectDetectionPipelineOptions Parameters specific to zero-shot object detection pipelines.
27201
27328
  * @property {number} [threshold=0.1] The probability necessary to make a prediction.
27202
27329
  * @property {number} [top_k=null] The number of top predictions that will be returned by the pipeline.
27203
27330
  * If the provided number is `null` or higher than the number of predictions available, it will default
27204
27331
  * to the number of predictions.
27205
27332
  * @property {boolean} [percentage=false] Whether to return the boxes coordinates in percentage (true) or in pixels (false).
27206
- *
27333
+ *
27207
27334
  * @callback ZeroShotObjectDetectionPipelineCallback Detect objects (bounding boxes & classes) in the image(s) passed as inputs.
27208
27335
  * @param {ImagePipelineInputs} images The input images.
27209
27336
  * @param {string[]} candidate_labels What the model should recognize in the image.
27210
27337
  * @param {ZeroShotObjectDetectionPipelineOptions} [options] The options to use for zero-shot object detection.
27211
27338
  * @returns {Promise<ZeroShotObjectDetectionOutput[]|ZeroShotObjectDetectionOutput[][]>} An array of objects containing the predicted labels, scores, and bounding boxes.
27212
- *
27339
+ *
27213
27340
  * @typedef {TextImagePipelineConstructorArgs & ZeroShotObjectDetectionPipelineCallback & Disposable} ZeroShotObjectDetectionPipelineType
27214
27341
  */
27215
27342
 
27216
27343
  /**
27217
27344
  * Zero-shot object detection pipeline. This pipeline predicts bounding boxes of
27218
27345
  * objects when you provide an image and a set of `candidate_labels`.
27219
- *
27346
+ *
27220
27347
  * **Example:** Zero-shot object detection w/ `Xenova/owlvit-base-patch32`.
27221
27348
  * ```javascript
27222
27349
  * const detector = await pipeline('zero-shot-object-detection', 'Xenova/owlvit-base-patch32');
@@ -27246,7 +27373,7 @@ class ObjectDetectionPipeline extends (/** @type {new (options: ImagePipelineCon
27246
27373
  * // }
27247
27374
  * // ]
27248
27375
  * ```
27249
- *
27376
+ *
27250
27377
  * **Example:** Zero-shot object detection w/ `Xenova/owlvit-base-patch32` (returning top 4 matches and setting a threshold).
27251
27378
  * ```javascript
27252
27379
  * const detector = await pipeline('zero-shot-object-detection', 'Xenova/owlvit-base-patch32');
@@ -27361,13 +27488,13 @@ class ZeroShotObjectDetectionPipeline extends (/** @type {new (options: TextImag
27361
27488
  * @typedef {Object} DocumentQuestionAnsweringSingle
27362
27489
  * @property {string} answer The generated text.
27363
27490
  * @typedef {DocumentQuestionAnsweringSingle[]} DocumentQuestionAnsweringOutput
27364
- *
27491
+ *
27365
27492
  * @callback DocumentQuestionAnsweringPipelineCallback Answer the question given as input by using the document.
27366
27493
  * @param {ImageInput} image The image of the document to use.
27367
27494
  * @param {string} question A question to ask of the document.
27368
27495
  * @param {Partial<import('./generation/configuration_utils.js').GenerationConfig>} [options] Additional keyword arguments to pass along to the generate method of the model.
27369
27496
  * @returns {Promise<DocumentQuestionAnsweringOutput|DocumentQuestionAnsweringOutput[]>} An object (or array of objects) containing the answer(s).
27370
- *
27497
+ *
27371
27498
  * @typedef {TextImagePipelineConstructorArgs & DocumentQuestionAnsweringPipelineCallback & Disposable} DocumentQuestionAnsweringPipelineType
27372
27499
  */
27373
27500
 
@@ -27375,7 +27502,7 @@ class ZeroShotObjectDetectionPipeline extends (/** @type {new (options: TextImag
27375
27502
  * Document Question Answering pipeline using any `AutoModelForDocumentQuestionAnswering`.
27376
27503
  * The inputs/outputs are similar to the (extractive) question answering pipeline; however,
27377
27504
  * the pipeline takes an image (and optional OCR'd words/boxes) as input instead of text context.
27378
- *
27505
+ *
27379
27506
  * **Example:** Answer questions about a document with `Xenova/donut-base-finetuned-docvqa`.
27380
27507
  * ```javascript
27381
27508
  * const qa_pipeline = await pipeline('document-question-answering', 'Xenova/donut-base-finetuned-docvqa');
@@ -27445,22 +27572,22 @@ class DocumentQuestionAnsweringPipeline extends (/** @type {new (options: TextIm
27445
27572
  * @typedef {Object} TextToAudioOutput
27446
27573
  * @property {Float32Array} audio The generated audio waveform.
27447
27574
  * @property {number} sampling_rate The sampling rate of the generated audio waveform.
27448
- *
27575
+ *
27449
27576
  * @typedef {Object} TextToAudioPipelineOptions Parameters specific to text-to-audio pipelines.
27450
27577
  * @property {Tensor|Float32Array|string|URL} [speaker_embeddings=null] The speaker embeddings (if the model requires it).
27451
- *
27578
+ *
27452
27579
  * @callback TextToAudioPipelineCallback Generates speech/audio from the inputs.
27453
27580
  * @param {string|string[]} texts The text(s) to generate.
27454
27581
  * @param {TextToAudioPipelineOptions} options Parameters passed to the model generation/forward method.
27455
27582
  * @returns {Promise<TextToAudioOutput>} An object containing the generated audio and sampling rate.
27456
- *
27583
+ *
27457
27584
  * @typedef {TextToAudioPipelineConstructorArgs & TextToAudioPipelineCallback & Disposable} TextToAudioPipelineType
27458
27585
  */
27459
27586
 
27460
27587
  /**
27461
27588
  * Text-to-audio generation pipeline using any `AutoModelForTextToWaveform` or `AutoModelForTextToSpectrogram`.
27462
27589
  * This pipeline generates an audio file from an input text and optional other conditional inputs.
27463
- *
27590
+ *
27464
27591
  * **Example:** Generate audio from text with `Xenova/speecht5_tts`.
27465
27592
  * ```javascript
27466
27593
  * const synthesizer = await pipeline('text-to-speech', 'Xenova/speecht5_tts', { quantized: false });
@@ -27471,17 +27598,17 @@ class DocumentQuestionAnsweringPipeline extends (/** @type {new (options: TextIm
27471
27598
  * // sampling_rate: 16000
27472
27599
  * // }
27473
27600
  * ```
27474
- *
27601
+ *
27475
27602
  * You can then save the audio to a .wav file with the `wavefile` package:
27476
27603
  * ```javascript
27477
27604
  * import wavefile from 'wavefile';
27478
27605
  * import fs from 'fs';
27479
- *
27606
+ *
27480
27607
  * const wav = new wavefile.WaveFile();
27481
27608
  * wav.fromScratch(1, out.sampling_rate, '32f', out.audio);
27482
27609
  * fs.writeFileSync('out.wav', wav.toBuffer());
27483
27610
  * ```
27484
- *
27611
+ *
27485
27612
  * **Example:** Multilingual speech generation with `Xenova/mms-tts-fra`. See [here](https://huggingface.co/models?pipeline_tag=text-to-speech&other=vits&sort=trending) for the full list of available languages (1107).
27486
27613
  * ```javascript
27487
27614
  * const synthesizer = await pipeline('text-to-speech', 'Xenova/mms-tts-fra');
@@ -27587,13 +27714,13 @@ class TextToAudioPipeline extends (/** @type {new (options: TextToAudioPipelineC
27587
27714
  * @callback ImageToImagePipelineCallback Transform the image(s) passed as inputs.
27588
27715
  * @param {ImagePipelineInputs} images The images to transform.
27589
27716
  * @returns {Promise<RawImage|RawImage[]>} The transformed image or list of images.
27590
- *
27717
+ *
27591
27718
  * @typedef {ImagePipelineConstructorArgs & ImageToImagePipelineCallback & Disposable} ImageToImagePipelineType
27592
27719
  */
27593
27720
 
27594
27721
  /**
27595
27722
  * Image to Image pipeline using any `AutoModelForImageToImage`. This pipeline generates an image based on a previous image input.
27596
- *
27723
+ *
27597
27724
  * **Example:** Super-resolution w/ `Xenova/swin2SR-classical-sr-x2-64`
27598
27725
  * ```javascript
27599
27726
  * const upscaler = await pipeline('image-to-image', 'Xenova/swin2SR-classical-sr-x2-64');
@@ -27638,17 +27765,17 @@ class ImageToImagePipeline extends (/** @type {new (options: ImagePipelineConstr
27638
27765
  * @typedef {Object} DepthEstimationPipelineOutput
27639
27766
  * @property {Tensor} predicted_depth The raw depth map predicted by the model.
27640
27767
  * @property {RawImage} depth The processed depth map as an image (with the same size as the input image).
27641
- *
27768
+ *
27642
27769
  * @callback DepthEstimationPipelineCallback Predicts the depth for the image(s) passed as inputs.
27643
27770
  * @param {ImagePipelineInputs} images The images to compute depth for.
27644
27771
  * @returns {Promise<DepthEstimationPipelineOutput|DepthEstimationPipelineOutput[]>} An image or a list of images containing result(s).
27645
- *
27772
+ *
27646
27773
  * @typedef {ImagePipelineConstructorArgs & DepthEstimationPipelineCallback & Disposable} DepthEstimationPipelineType
27647
27774
  */
27648
27775
 
27649
27776
  /**
27650
27777
  * Depth estimation pipeline using any `AutoModelForDepthEstimation`. This pipeline predicts the depth of an image.
27651
- *
27778
+ *
27652
27779
  * **Example:** Depth estimation w/ `Xenova/dpt-hybrid-midas`
27653
27780
  * ```javascript
27654
27781
  * const depth_estimator = await pipeline('depth-estimation', 'Xenova/dpt-hybrid-midas');
@@ -28033,7 +28160,7 @@ const TASK_ALIASES = Object.freeze({
28033
28160
 
28034
28161
  /**
28035
28162
  * Utility factory method to build a `Pipeline` object.
28036
- *
28163
+ *
28037
28164
  * @template {PipelineType} T The type of pipeline to return.
28038
28165
  * @param {T} task The task defining which pipeline will be returned. Currently accepted tasks are:
28039
28166
  * - `"audio-classification"`: will return a `AudioClassificationPipeline`.
@@ -35299,7 +35426,7 @@ class RawImage {
35299
35426
 
35300
35427
  /**
35301
35428
  * Helper method for reading an image from a variety of input types.
35302
- * @param {RawImage|string|URL} input
35429
+ * @param {RawImage|string|URL|Blob|HTMLCanvasElement|OffscreenCanvas} input
35303
35430
  * @returns The image object.
35304
35431
  *
35305
35432
  * **Example:** Read image from a URL.
@@ -35318,6 +35445,14 @@ class RawImage {
35318
35445
  return input;
35319
35446
  } else if (typeof input === 'string' || input instanceof URL) {
35320
35447
  return await this.fromURL(input);
35448
+ } else if (input instanceof Blob) {
35449
+ return await this.fromBlob(input);
35450
+ } else if (
35451
+ (typeof HTMLCanvasElement !== "undefined" && input instanceof HTMLCanvasElement)
35452
+ ||
35453
+ (typeof OffscreenCanvas !== "undefined" && input instanceof OffscreenCanvas)
35454
+ ) {
35455
+ return this.fromCanvas(input);
35321
35456
  } else {
35322
35457
  throw new Error(`Unsupported input type: ${typeof input}`);
35323
35458
  }
@@ -38327,8 +38462,12 @@ function calc_unsqueeze_dims(dims, dim) {
38327
38462
  * @private
38328
38463
  */
38329
38464
  function safeIndex(index, size, dimension = null, boundsCheck = true) {
38330
- if (boundsCheck && (index < -size || index >= size)) {
38331
- throw new Error(`IndexError: index ${index} is out of bounds for dimension${dimension === null ? '' : ' ' + dimension} with size ${size}`);
38465
+ if (index < -size || index >= size) {
38466
+ if (boundsCheck) {
38467
+ throw new Error(`IndexError: index ${index} is out of bounds for dimension${dimension === null ? '' : ' ' + dimension} with size ${size}`);
38468
+ } else {
38469
+ return index < -size ? 0 : size;
38470
+ }
38332
38471
  }
38333
38472
 
38334
38473
  if (index < 0) {
@@ -39199,6 +39338,9 @@ __webpack_require__.r(__webpack_exports__);
39199
39338
  /* harmony export */ Gemma2ForCausalLM: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Gemma2ForCausalLM),
39200
39339
  /* harmony export */ Gemma2Model: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Gemma2Model),
39201
39340
  /* harmony export */ Gemma2PreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Gemma2PreTrainedModel),
39341
+ /* harmony export */ Gemma3ForCausalLM: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Gemma3ForCausalLM),
39342
+ /* harmony export */ Gemma3Model: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Gemma3Model),
39343
+ /* harmony export */ Gemma3PreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Gemma3PreTrainedModel),
39202
39344
  /* harmony export */ GemmaForCausalLM: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.GemmaForCausalLM),
39203
39345
  /* harmony export */ GemmaModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.GemmaModel),
39204
39346
  /* harmony export */ GemmaPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.GemmaPreTrainedModel),
@@ -39300,6 +39442,10 @@ __webpack_require__.r(__webpack_exports__);
39300
39442
  /* harmony export */ MaskFormerPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.MaskFormerPreTrainedModel),
39301
39443
  /* harmony export */ MaskedLMOutput: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.MaskedLMOutput),
39302
39444
  /* harmony export */ MaxLengthCriteria: () => (/* reexport safe */ _generation_stopping_criteria_js__WEBPACK_IMPORTED_MODULE_20__.MaxLengthCriteria),
39445
+ /* harmony export */ Metric3DForDepthEstimation: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Metric3DForDepthEstimation),
39446
+ /* harmony export */ Metric3DPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Metric3DPreTrainedModel),
39447
+ /* harmony export */ Metric3Dv2ForDepthEstimation: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Metric3Dv2ForDepthEstimation),
39448
+ /* harmony export */ Metric3Dv2PreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Metric3Dv2PreTrainedModel),
39303
39449
  /* harmony export */ MgpstrForSceneTextRecognition: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.MgpstrForSceneTextRecognition),
39304
39450
  /* harmony export */ MgpstrModelOutput: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.MgpstrModelOutput),
39305
39451
  /* harmony export */ MgpstrPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.MgpstrPreTrainedModel),
@@ -39506,6 +39652,11 @@ __webpack_require__.r(__webpack_exports__);
39506
39652
  /* harmony export */ SmolVLMForConditionalGeneration: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.SmolVLMForConditionalGeneration),
39507
39653
  /* harmony export */ SmolVLMImageProcessor: () => (/* reexport safe */ _models_image_processors_js__WEBPACK_IMPORTED_MODULE_14__.SmolVLMImageProcessor),
39508
39654
  /* harmony export */ SmolVLMProcessor: () => (/* reexport safe */ _models_processors_js__WEBPACK_IMPORTED_MODULE_17__.SmolVLMProcessor),
39655
+ /* harmony export */ SnacDecoderModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.SnacDecoderModel),
39656
+ /* harmony export */ SnacEncoderModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.SnacEncoderModel),
39657
+ /* harmony export */ SnacFeatureExtractor: () => (/* reexport safe */ _models_feature_extractors_js__WEBPACK_IMPORTED_MODULE_11__.SnacFeatureExtractor),
39658
+ /* harmony export */ SnacModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.SnacModel),
39659
+ /* harmony export */ SnacPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.SnacPreTrainedModel),
39509
39660
  /* harmony export */ SpeechT5FeatureExtractor: () => (/* reexport safe */ _models_feature_extractors_js__WEBPACK_IMPORTED_MODULE_11__.SpeechT5FeatureExtractor),
39510
39661
  /* harmony export */ SpeechT5ForSpeechToText: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.SpeechT5ForSpeechToText),
39511
39662
  /* harmony export */ SpeechT5ForTextToSpeech: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.SpeechT5ForTextToSpeech),
@@ -40028,6 +40179,9 @@ var __webpack_exports__GPTNeoXTokenizer = __webpack_exports__.GPTNeoXTokenizer;
40028
40179
  var __webpack_exports__Gemma2ForCausalLM = __webpack_exports__.Gemma2ForCausalLM;
40029
40180
  var __webpack_exports__Gemma2Model = __webpack_exports__.Gemma2Model;
40030
40181
  var __webpack_exports__Gemma2PreTrainedModel = __webpack_exports__.Gemma2PreTrainedModel;
40182
+ var __webpack_exports__Gemma3ForCausalLM = __webpack_exports__.Gemma3ForCausalLM;
40183
+ var __webpack_exports__Gemma3Model = __webpack_exports__.Gemma3Model;
40184
+ var __webpack_exports__Gemma3PreTrainedModel = __webpack_exports__.Gemma3PreTrainedModel;
40031
40185
  var __webpack_exports__GemmaForCausalLM = __webpack_exports__.GemmaForCausalLM;
40032
40186
  var __webpack_exports__GemmaModel = __webpack_exports__.GemmaModel;
40033
40187
  var __webpack_exports__GemmaPreTrainedModel = __webpack_exports__.GemmaPreTrainedModel;
@@ -40129,6 +40283,10 @@ var __webpack_exports__MaskFormerModel = __webpack_exports__.MaskFormerModel;
40129
40283
  var __webpack_exports__MaskFormerPreTrainedModel = __webpack_exports__.MaskFormerPreTrainedModel;
40130
40284
  var __webpack_exports__MaskedLMOutput = __webpack_exports__.MaskedLMOutput;
40131
40285
  var __webpack_exports__MaxLengthCriteria = __webpack_exports__.MaxLengthCriteria;
40286
+ var __webpack_exports__Metric3DForDepthEstimation = __webpack_exports__.Metric3DForDepthEstimation;
40287
+ var __webpack_exports__Metric3DPreTrainedModel = __webpack_exports__.Metric3DPreTrainedModel;
40288
+ var __webpack_exports__Metric3Dv2ForDepthEstimation = __webpack_exports__.Metric3Dv2ForDepthEstimation;
40289
+ var __webpack_exports__Metric3Dv2PreTrainedModel = __webpack_exports__.Metric3Dv2PreTrainedModel;
40132
40290
  var __webpack_exports__MgpstrForSceneTextRecognition = __webpack_exports__.MgpstrForSceneTextRecognition;
40133
40291
  var __webpack_exports__MgpstrModelOutput = __webpack_exports__.MgpstrModelOutput;
40134
40292
  var __webpack_exports__MgpstrPreTrainedModel = __webpack_exports__.MgpstrPreTrainedModel;
@@ -40335,6 +40493,11 @@ var __webpack_exports__SiglipVisionModel = __webpack_exports__.SiglipVisionModel
40335
40493
  var __webpack_exports__SmolVLMForConditionalGeneration = __webpack_exports__.SmolVLMForConditionalGeneration;
40336
40494
  var __webpack_exports__SmolVLMImageProcessor = __webpack_exports__.SmolVLMImageProcessor;
40337
40495
  var __webpack_exports__SmolVLMProcessor = __webpack_exports__.SmolVLMProcessor;
40496
+ var __webpack_exports__SnacDecoderModel = __webpack_exports__.SnacDecoderModel;
40497
+ var __webpack_exports__SnacEncoderModel = __webpack_exports__.SnacEncoderModel;
40498
+ var __webpack_exports__SnacFeatureExtractor = __webpack_exports__.SnacFeatureExtractor;
40499
+ var __webpack_exports__SnacModel = __webpack_exports__.SnacModel;
40500
+ var __webpack_exports__SnacPreTrainedModel = __webpack_exports__.SnacPreTrainedModel;
40338
40501
  var __webpack_exports__SpeechT5FeatureExtractor = __webpack_exports__.SpeechT5FeatureExtractor;
40339
40502
  var __webpack_exports__SpeechT5ForSpeechToText = __webpack_exports__.SpeechT5ForSpeechToText;
40340
40503
  var __webpack_exports__SpeechT5ForTextToSpeech = __webpack_exports__.SpeechT5ForTextToSpeech;
@@ -40528,6 +40691,6 @@ var __webpack_exports__topk = __webpack_exports__.topk;
40528
40691
  var __webpack_exports__window_function = __webpack_exports__.window_function;
40529
40692
  var __webpack_exports__zeros = __webpack_exports__.zeros;
40530
40693
  var __webpack_exports__zeros_like = __webpack_exports__.zeros_like;
40531
- export { __webpack_exports__ASTFeatureExtractor as ASTFeatureExtractor, __webpack_exports__ASTForAudioClassification as ASTForAudioClassification, __webpack_exports__ASTModel as ASTModel, __webpack_exports__ASTPreTrainedModel as ASTPreTrainedModel, __webpack_exports__AlbertForMaskedLM as AlbertForMaskedLM, __webpack_exports__AlbertForQuestionAnswering as AlbertForQuestionAnswering, __webpack_exports__AlbertForSequenceClassification as AlbertForSequenceClassification, __webpack_exports__AlbertModel as AlbertModel, __webpack_exports__AlbertPreTrainedModel as AlbertPreTrainedModel, __webpack_exports__AlbertTokenizer as AlbertTokenizer, __webpack_exports__AudioClassificationPipeline as AudioClassificationPipeline, __webpack_exports__AutoConfig as AutoConfig, __webpack_exports__AutoFeatureExtractor as AutoFeatureExtractor, __webpack_exports__AutoImageProcessor as AutoImageProcessor, __webpack_exports__AutoModel as AutoModel, __webpack_exports__AutoModelForAudioClassification as AutoModelForAudioClassification, __webpack_exports__AutoModelForAudioFrameClassification as AutoModelForAudioFrameClassification, __webpack_exports__AutoModelForAudioTextToText as AutoModelForAudioTextToText, __webpack_exports__AutoModelForCTC as AutoModelForCTC, __webpack_exports__AutoModelForCausalLM as AutoModelForCausalLM, __webpack_exports__AutoModelForDepthEstimation as AutoModelForDepthEstimation, __webpack_exports__AutoModelForDocumentQuestionAnswering as AutoModelForDocumentQuestionAnswering, __webpack_exports__AutoModelForImageClassification as AutoModelForImageClassification, __webpack_exports__AutoModelForImageFeatureExtraction as AutoModelForImageFeatureExtraction, __webpack_exports__AutoModelForImageMatting as AutoModelForImageMatting, __webpack_exports__AutoModelForImageSegmentation as AutoModelForImageSegmentation, __webpack_exports__AutoModelForImageTextToText as AutoModelForImageTextToText, __webpack_exports__AutoModelForImageToImage as AutoModelForImageToImage, __webpack_exports__AutoModelForMaskGeneration as AutoModelForMaskGeneration, __webpack_exports__AutoModelForMaskedLM as AutoModelForMaskedLM, __webpack_exports__AutoModelForNormalEstimation as AutoModelForNormalEstimation, __webpack_exports__AutoModelForObjectDetection as AutoModelForObjectDetection, __webpack_exports__AutoModelForPoseEstimation as AutoModelForPoseEstimation, __webpack_exports__AutoModelForQuestionAnswering as AutoModelForQuestionAnswering, __webpack_exports__AutoModelForSemanticSegmentation as AutoModelForSemanticSegmentation, __webpack_exports__AutoModelForSeq2SeqLM as AutoModelForSeq2SeqLM, __webpack_exports__AutoModelForSequenceClassification as AutoModelForSequenceClassification, __webpack_exports__AutoModelForSpeechSeq2Seq as AutoModelForSpeechSeq2Seq, __webpack_exports__AutoModelForTextToSpectrogram as AutoModelForTextToSpectrogram, __webpack_exports__AutoModelForTextToWaveform as AutoModelForTextToWaveform, __webpack_exports__AutoModelForTokenClassification as AutoModelForTokenClassification, __webpack_exports__AutoModelForUniversalSegmentation as AutoModelForUniversalSegmentation, __webpack_exports__AutoModelForVision2Seq as AutoModelForVision2Seq, __webpack_exports__AutoModelForXVector as AutoModelForXVector, __webpack_exports__AutoModelForZeroShotObjectDetection as AutoModelForZeroShotObjectDetection, __webpack_exports__AutoProcessor as AutoProcessor, __webpack_exports__AutoTokenizer as AutoTokenizer, __webpack_exports__AutomaticSpeechRecognitionPipeline as AutomaticSpeechRecognitionPipeline, __webpack_exports__BackgroundRemovalPipeline as BackgroundRemovalPipeline, __webpack_exports__BartForConditionalGeneration as BartForConditionalGeneration, __webpack_exports__BartForSequenceClassification as BartForSequenceClassification, __webpack_exports__BartModel as BartModel, __webpack_exports__BartPretrainedModel as BartPretrainedModel, __webpack_exports__BartTokenizer as BartTokenizer, __webpack_exports__BaseModelOutput as BaseModelOutput, __webpack_exports__BaseStreamer as BaseStreamer, __webpack_exports__BeitFeatureExtractor as BeitFeatureExtractor, __webpack_exports__BeitForImageClassification as BeitForImageClassification, __webpack_exports__BeitModel as BeitModel, __webpack_exports__BeitPreTrainedModel as BeitPreTrainedModel, __webpack_exports__BertForMaskedLM as BertForMaskedLM, __webpack_exports__BertForQuestionAnswering as BertForQuestionAnswering, __webpack_exports__BertForSequenceClassification as BertForSequenceClassification, __webpack_exports__BertForTokenClassification as BertForTokenClassification, __webpack_exports__BertModel as BertModel, __webpack_exports__BertPreTrainedModel as BertPreTrainedModel, __webpack_exports__BertTokenizer as BertTokenizer, __webpack_exports__BitImageProcessor as BitImageProcessor, __webpack_exports__BlenderbotForConditionalGeneration as BlenderbotForConditionalGeneration, __webpack_exports__BlenderbotModel as BlenderbotModel, __webpack_exports__BlenderbotPreTrainedModel as BlenderbotPreTrainedModel, __webpack_exports__BlenderbotSmallForConditionalGeneration as BlenderbotSmallForConditionalGeneration, __webpack_exports__BlenderbotSmallModel as BlenderbotSmallModel, __webpack_exports__BlenderbotSmallPreTrainedModel as BlenderbotSmallPreTrainedModel, __webpack_exports__BlenderbotSmallTokenizer as BlenderbotSmallTokenizer, __webpack_exports__BlenderbotTokenizer as BlenderbotTokenizer, __webpack_exports__BloomForCausalLM as BloomForCausalLM, __webpack_exports__BloomModel as BloomModel, __webpack_exports__BloomPreTrainedModel as BloomPreTrainedModel, __webpack_exports__BloomTokenizer as BloomTokenizer, __webpack_exports__CLIPFeatureExtractor as CLIPFeatureExtractor, __webpack_exports__CLIPImageProcessor as CLIPImageProcessor, __webpack_exports__CLIPModel as CLIPModel, __webpack_exports__CLIPPreTrainedModel as CLIPPreTrainedModel, __webpack_exports__CLIPSegForImageSegmentation as CLIPSegForImageSegmentation, __webpack_exports__CLIPSegModel as CLIPSegModel, __webpack_exports__CLIPSegPreTrainedModel as CLIPSegPreTrainedModel, __webpack_exports__CLIPTextModel as CLIPTextModel, __webpack_exports__CLIPTextModelWithProjection as CLIPTextModelWithProjection, __webpack_exports__CLIPTokenizer as CLIPTokenizer, __webpack_exports__CLIPVisionModel as CLIPVisionModel, __webpack_exports__CLIPVisionModelWithProjection as CLIPVisionModelWithProjection, __webpack_exports__CamembertForMaskedLM as CamembertForMaskedLM, __webpack_exports__CamembertForQuestionAnswering as CamembertForQuestionAnswering, __webpack_exports__CamembertForSequenceClassification as CamembertForSequenceClassification, __webpack_exports__CamembertForTokenClassification as CamembertForTokenClassification, __webpack_exports__CamembertModel as CamembertModel, __webpack_exports__CamembertPreTrainedModel as CamembertPreTrainedModel, __webpack_exports__CamembertTokenizer as CamembertTokenizer, __webpack_exports__CausalLMOutput as CausalLMOutput, __webpack_exports__CausalLMOutputWithPast as CausalLMOutputWithPast, __webpack_exports__ChineseCLIPFeatureExtractor as ChineseCLIPFeatureExtractor, __webpack_exports__ChineseCLIPModel as ChineseCLIPModel, __webpack_exports__ChineseCLIPPreTrainedModel as ChineseCLIPPreTrainedModel, __webpack_exports__ClapAudioModelWithProjection as ClapAudioModelWithProjection, __webpack_exports__ClapFeatureExtractor as ClapFeatureExtractor, __webpack_exports__ClapModel as ClapModel, __webpack_exports__ClapPreTrainedModel as ClapPreTrainedModel, __webpack_exports__ClapTextModelWithProjection as ClapTextModelWithProjection, __webpack_exports__ClassifierFreeGuidanceLogitsProcessor as ClassifierFreeGuidanceLogitsProcessor, __webpack_exports__CodeGenForCausalLM as CodeGenForCausalLM, __webpack_exports__CodeGenModel as CodeGenModel, __webpack_exports__CodeGenPreTrainedModel as CodeGenPreTrainedModel, __webpack_exports__CodeGenTokenizer as CodeGenTokenizer, __webpack_exports__CodeLlamaTokenizer as CodeLlamaTokenizer, __webpack_exports__CohereForCausalLM as CohereForCausalLM, __webpack_exports__CohereModel as CohereModel, __webpack_exports__CoherePreTrainedModel as CoherePreTrainedModel, __webpack_exports__CohereTokenizer as CohereTokenizer, __webpack_exports__ConvBertForMaskedLM as ConvBertForMaskedLM, __webpack_exports__ConvBertForQuestionAnswering as ConvBertForQuestionAnswering, __webpack_exports__ConvBertForSequenceClassification as ConvBertForSequenceClassification, __webpack_exports__ConvBertForTokenClassification as ConvBertForTokenClassification, __webpack_exports__ConvBertModel as ConvBertModel, __webpack_exports__ConvBertPreTrainedModel as ConvBertPreTrainedModel, __webpack_exports__ConvBertTokenizer as ConvBertTokenizer, __webpack_exports__ConvNextFeatureExtractor as ConvNextFeatureExtractor, __webpack_exports__ConvNextForImageClassification as ConvNextForImageClassification, __webpack_exports__ConvNextImageProcessor as ConvNextImageProcessor, __webpack_exports__ConvNextModel as ConvNextModel, __webpack_exports__ConvNextPreTrainedModel as ConvNextPreTrainedModel, __webpack_exports__ConvNextV2ForImageClassification as ConvNextV2ForImageClassification, __webpack_exports__ConvNextV2Model as ConvNextV2Model, __webpack_exports__ConvNextV2PreTrainedModel as ConvNextV2PreTrainedModel, __webpack_exports__DPTFeatureExtractor as DPTFeatureExtractor, __webpack_exports__DPTForDepthEstimation as DPTForDepthEstimation, __webpack_exports__DPTImageProcessor as DPTImageProcessor, __webpack_exports__DPTModel as DPTModel, __webpack_exports__DPTPreTrainedModel as DPTPreTrainedModel, __webpack_exports__DacDecoderModel as DacDecoderModel, __webpack_exports__DacDecoderOutput as DacDecoderOutput, __webpack_exports__DacEncoderModel as DacEncoderModel, __webpack_exports__DacEncoderOutput as DacEncoderOutput, __webpack_exports__DacFeatureExtractor as DacFeatureExtractor, __webpack_exports__DacModel as DacModel, __webpack_exports__DacPreTrainedModel as DacPreTrainedModel, __webpack_exports__DataTypeMap as DataTypeMap, __webpack_exports__DebertaForMaskedLM as DebertaForMaskedLM, __webpack_exports__DebertaForQuestionAnswering as DebertaForQuestionAnswering, __webpack_exports__DebertaForSequenceClassification as DebertaForSequenceClassification, __webpack_exports__DebertaForTokenClassification as DebertaForTokenClassification, __webpack_exports__DebertaModel as DebertaModel, __webpack_exports__DebertaPreTrainedModel as DebertaPreTrainedModel, __webpack_exports__DebertaTokenizer as DebertaTokenizer, __webpack_exports__DebertaV2ForMaskedLM as DebertaV2ForMaskedLM, __webpack_exports__DebertaV2ForQuestionAnswering as DebertaV2ForQuestionAnswering, __webpack_exports__DebertaV2ForSequenceClassification as DebertaV2ForSequenceClassification, __webpack_exports__DebertaV2ForTokenClassification as DebertaV2ForTokenClassification, __webpack_exports__DebertaV2Model as DebertaV2Model, __webpack_exports__DebertaV2PreTrainedModel as DebertaV2PreTrainedModel, __webpack_exports__DebertaV2Tokenizer as DebertaV2Tokenizer, __webpack_exports__DecisionTransformerModel as DecisionTransformerModel, __webpack_exports__DecisionTransformerPreTrainedModel as DecisionTransformerPreTrainedModel, __webpack_exports__DeiTFeatureExtractor as DeiTFeatureExtractor, __webpack_exports__DeiTForImageClassification as DeiTForImageClassification, __webpack_exports__DeiTImageProcessor as DeiTImageProcessor, __webpack_exports__DeiTModel as DeiTModel, __webpack_exports__DeiTPreTrainedModel as DeiTPreTrainedModel, __webpack_exports__DepthAnythingForDepthEstimation as DepthAnythingForDepthEstimation, __webpack_exports__DepthAnythingPreTrainedModel as DepthAnythingPreTrainedModel, __webpack_exports__DepthEstimationPipeline as DepthEstimationPipeline, __webpack_exports__DepthProForDepthEstimation as DepthProForDepthEstimation, __webpack_exports__DepthProPreTrainedModel as DepthProPreTrainedModel, __webpack_exports__DetrFeatureExtractor as DetrFeatureExtractor, __webpack_exports__DetrForObjectDetection as DetrForObjectDetection, __webpack_exports__DetrForSegmentation as DetrForSegmentation, __webpack_exports__DetrImageProcessor as DetrImageProcessor, __webpack_exports__DetrModel as DetrModel, __webpack_exports__DetrObjectDetectionOutput as DetrObjectDetectionOutput, __webpack_exports__DetrPreTrainedModel as DetrPreTrainedModel, __webpack_exports__DetrSegmentationOutput as DetrSegmentationOutput, __webpack_exports__Dinov2ForImageClassification as Dinov2ForImageClassification, __webpack_exports__Dinov2Model as Dinov2Model, __webpack_exports__Dinov2PreTrainedModel as Dinov2PreTrainedModel, __webpack_exports__Dinov2WithRegistersForImageClassification as Dinov2WithRegistersForImageClassification, __webpack_exports__Dinov2WithRegistersModel as Dinov2WithRegistersModel, __webpack_exports__Dinov2WithRegistersPreTrainedModel as Dinov2WithRegistersPreTrainedModel, __webpack_exports__DistilBertForMaskedLM as DistilBertForMaskedLM, __webpack_exports__DistilBertForQuestionAnswering as DistilBertForQuestionAnswering, __webpack_exports__DistilBertForSequenceClassification as DistilBertForSequenceClassification, __webpack_exports__DistilBertForTokenClassification as DistilBertForTokenClassification, __webpack_exports__DistilBertModel as DistilBertModel, __webpack_exports__DistilBertPreTrainedModel as DistilBertPreTrainedModel, __webpack_exports__DistilBertTokenizer as DistilBertTokenizer, __webpack_exports__DocumentQuestionAnsweringPipeline as DocumentQuestionAnsweringPipeline, __webpack_exports__DonutFeatureExtractor as DonutFeatureExtractor, __webpack_exports__DonutImageProcessor as DonutImageProcessor, __webpack_exports__DonutSwinModel as DonutSwinModel, __webpack_exports__DonutSwinPreTrainedModel as DonutSwinPreTrainedModel, __webpack_exports__EfficientNetForImageClassification as EfficientNetForImageClassification, __webpack_exports__EfficientNetImageProcessor as EfficientNetImageProcessor, __webpack_exports__EfficientNetModel as EfficientNetModel, __webpack_exports__EfficientNetPreTrainedModel as EfficientNetPreTrainedModel, __webpack_exports__ElectraForMaskedLM as ElectraForMaskedLM, __webpack_exports__ElectraForQuestionAnswering as ElectraForQuestionAnswering, __webpack_exports__ElectraForSequenceClassification as ElectraForSequenceClassification, __webpack_exports__ElectraForTokenClassification as ElectraForTokenClassification, __webpack_exports__ElectraModel as ElectraModel, __webpack_exports__ElectraPreTrainedModel as ElectraPreTrainedModel, __webpack_exports__ElectraTokenizer as ElectraTokenizer, __webpack_exports__EncodecFeatureExtractor as EncodecFeatureExtractor, __webpack_exports__EosTokenCriteria as EosTokenCriteria, __webpack_exports__EsmForMaskedLM as EsmForMaskedLM, __webpack_exports__EsmForSequenceClassification as EsmForSequenceClassification, __webpack_exports__EsmForTokenClassification as EsmForTokenClassification, __webpack_exports__EsmModel as EsmModel, __webpack_exports__EsmPreTrainedModel as EsmPreTrainedModel, __webpack_exports__EsmTokenizer as EsmTokenizer, __webpack_exports__ExaoneForCausalLM as ExaoneForCausalLM, __webpack_exports__ExaoneModel as ExaoneModel, __webpack_exports__ExaonePreTrainedModel as ExaonePreTrainedModel, __webpack_exports__FFT as FFT, __webpack_exports__FalconForCausalLM as FalconForCausalLM, __webpack_exports__FalconModel as FalconModel, __webpack_exports__FalconPreTrainedModel as FalconPreTrainedModel, __webpack_exports__FalconTokenizer as FalconTokenizer, __webpack_exports__FastViTForImageClassification as FastViTForImageClassification, __webpack_exports__FastViTModel as FastViTModel, __webpack_exports__FastViTPreTrainedModel as FastViTPreTrainedModel, __webpack_exports__FeatureExtractionPipeline as FeatureExtractionPipeline, __webpack_exports__FeatureExtractor as FeatureExtractor, __webpack_exports__FillMaskPipeline as FillMaskPipeline, __webpack_exports__Florence2ForConditionalGeneration as Florence2ForConditionalGeneration, __webpack_exports__Florence2PreTrainedModel as Florence2PreTrainedModel, __webpack_exports__Florence2Processor as Florence2Processor, __webpack_exports__ForcedBOSTokenLogitsProcessor as ForcedBOSTokenLogitsProcessor, __webpack_exports__ForcedEOSTokenLogitsProcessor as ForcedEOSTokenLogitsProcessor, __webpack_exports__GLPNFeatureExtractor as GLPNFeatureExtractor, __webpack_exports__GLPNForDepthEstimation as GLPNForDepthEstimation, __webpack_exports__GLPNModel as GLPNModel, __webpack_exports__GLPNPreTrainedModel as GLPNPreTrainedModel, __webpack_exports__GPT2LMHeadModel as GPT2LMHeadModel, __webpack_exports__GPT2Model as GPT2Model, __webpack_exports__GPT2PreTrainedModel as GPT2PreTrainedModel, __webpack_exports__GPT2Tokenizer as GPT2Tokenizer, __webpack_exports__GPTBigCodeForCausalLM as GPTBigCodeForCausalLM, __webpack_exports__GPTBigCodeModel as GPTBigCodeModel, __webpack_exports__GPTBigCodePreTrainedModel as GPTBigCodePreTrainedModel, __webpack_exports__GPTJForCausalLM as GPTJForCausalLM, __webpack_exports__GPTJModel as GPTJModel, __webpack_exports__GPTJPreTrainedModel as GPTJPreTrainedModel, __webpack_exports__GPTNeoForCausalLM as GPTNeoForCausalLM, __webpack_exports__GPTNeoModel as GPTNeoModel, __webpack_exports__GPTNeoPreTrainedModel as GPTNeoPreTrainedModel, __webpack_exports__GPTNeoXForCausalLM as GPTNeoXForCausalLM, __webpack_exports__GPTNeoXModel as GPTNeoXModel, __webpack_exports__GPTNeoXPreTrainedModel as GPTNeoXPreTrainedModel, __webpack_exports__GPTNeoXTokenizer as GPTNeoXTokenizer, __webpack_exports__Gemma2ForCausalLM as Gemma2ForCausalLM, __webpack_exports__Gemma2Model as Gemma2Model, __webpack_exports__Gemma2PreTrainedModel as Gemma2PreTrainedModel, __webpack_exports__GemmaForCausalLM as GemmaForCausalLM, __webpack_exports__GemmaModel as GemmaModel, __webpack_exports__GemmaPreTrainedModel as GemmaPreTrainedModel, __webpack_exports__GemmaTokenizer as GemmaTokenizer, __webpack_exports__GlmForCausalLM as GlmForCausalLM, __webpack_exports__GlmModel as GlmModel, __webpack_exports__GlmPreTrainedModel as GlmPreTrainedModel, __webpack_exports__GraniteForCausalLM as GraniteForCausalLM, __webpack_exports__GraniteModel as GraniteModel, __webpack_exports__GranitePreTrainedModel as GranitePreTrainedModel, __webpack_exports__Grok1Tokenizer as Grok1Tokenizer, __webpack_exports__GroundingDinoForObjectDetection as GroundingDinoForObjectDetection, __webpack_exports__GroundingDinoImageProcessor as GroundingDinoImageProcessor, __webpack_exports__GroundingDinoPreTrainedModel as GroundingDinoPreTrainedModel, __webpack_exports__GroundingDinoProcessor as GroundingDinoProcessor, __webpack_exports__GroupViTModel as GroupViTModel, __webpack_exports__GroupViTPreTrainedModel as GroupViTPreTrainedModel, __webpack_exports__HeliumForCausalLM as HeliumForCausalLM, __webpack_exports__HeliumModel as HeliumModel, __webpack_exports__HeliumPreTrainedModel as HeliumPreTrainedModel, __webpack_exports__HerbertTokenizer as HerbertTokenizer, __webpack_exports__HieraForImageClassification as HieraForImageClassification, __webpack_exports__HieraModel as HieraModel, __webpack_exports__HieraPreTrainedModel as HieraPreTrainedModel, __webpack_exports__HubertForCTC as HubertForCTC, __webpack_exports__HubertForSequenceClassification as HubertForSequenceClassification, __webpack_exports__HubertModel as HubertModel, __webpack_exports__HubertPreTrainedModel as HubertPreTrainedModel, __webpack_exports__IJepaForImageClassification as IJepaForImageClassification, __webpack_exports__IJepaModel as IJepaModel, __webpack_exports__IJepaPreTrainedModel as IJepaPreTrainedModel, __webpack_exports__Idefics3ForConditionalGeneration as Idefics3ForConditionalGeneration, __webpack_exports__Idefics3ImageProcessor as Idefics3ImageProcessor, __webpack_exports__Idefics3PreTrainedModel as Idefics3PreTrainedModel, __webpack_exports__Idefics3Processor as Idefics3Processor, __webpack_exports__ImageClassificationPipeline as ImageClassificationPipeline, __webpack_exports__ImageFeatureExtractionPipeline as ImageFeatureExtractionPipeline, __webpack_exports__ImageFeatureExtractor as ImageFeatureExtractor, __webpack_exports__ImageMattingOutput as ImageMattingOutput, __webpack_exports__ImageProcessor as ImageProcessor, __webpack_exports__ImageSegmentationPipeline as ImageSegmentationPipeline, __webpack_exports__ImageToImagePipeline as ImageToImagePipeline, __webpack_exports__ImageToTextPipeline as ImageToTextPipeline, __webpack_exports__InterruptableStoppingCriteria as InterruptableStoppingCriteria, __webpack_exports__JAISLMHeadModel as JAISLMHeadModel, __webpack_exports__JAISModel as JAISModel, __webpack_exports__JAISPreTrainedModel as JAISPreTrainedModel, __webpack_exports__JinaCLIPImageProcessor as JinaCLIPImageProcessor, __webpack_exports__JinaCLIPModel as JinaCLIPModel, __webpack_exports__JinaCLIPPreTrainedModel as JinaCLIPPreTrainedModel, __webpack_exports__JinaCLIPProcessor as JinaCLIPProcessor, __webpack_exports__JinaCLIPTextModel as JinaCLIPTextModel, __webpack_exports__JinaCLIPVisionModel as JinaCLIPVisionModel, __webpack_exports__LiteWhisperForConditionalGeneration as LiteWhisperForConditionalGeneration, __webpack_exports__LlamaForCausalLM as LlamaForCausalLM, __webpack_exports__LlamaModel as LlamaModel, __webpack_exports__LlamaPreTrainedModel as LlamaPreTrainedModel, __webpack_exports__LlamaTokenizer as LlamaTokenizer, __webpack_exports__LlavaForConditionalGeneration as LlavaForConditionalGeneration, __webpack_exports__LlavaOnevisionForConditionalGeneration as LlavaOnevisionForConditionalGeneration, __webpack_exports__LlavaOnevisionImageProcessor as LlavaOnevisionImageProcessor, __webpack_exports__LlavaPreTrainedModel as LlavaPreTrainedModel, __webpack_exports__LogitsProcessor as LogitsProcessor, __webpack_exports__LogitsProcessorList as LogitsProcessorList, __webpack_exports__LogitsWarper as LogitsWarper, __webpack_exports__LongT5ForConditionalGeneration as LongT5ForConditionalGeneration, __webpack_exports__LongT5Model as LongT5Model, __webpack_exports__LongT5PreTrainedModel as LongT5PreTrainedModel, __webpack_exports__M2M100ForConditionalGeneration as M2M100ForConditionalGeneration, __webpack_exports__M2M100Model as M2M100Model, __webpack_exports__M2M100PreTrainedModel as M2M100PreTrainedModel, __webpack_exports__M2M100Tokenizer as M2M100Tokenizer, __webpack_exports__MBart50Tokenizer as MBart50Tokenizer, __webpack_exports__MBartForCausalLM as MBartForCausalLM, __webpack_exports__MBartForConditionalGeneration as MBartForConditionalGeneration, __webpack_exports__MBartForSequenceClassification as MBartForSequenceClassification, __webpack_exports__MBartModel as MBartModel, __webpack_exports__MBartPreTrainedModel as MBartPreTrainedModel, __webpack_exports__MBartTokenizer as MBartTokenizer, __webpack_exports__MPNetForMaskedLM as MPNetForMaskedLM, __webpack_exports__MPNetForQuestionAnswering as MPNetForQuestionAnswering, __webpack_exports__MPNetForSequenceClassification as MPNetForSequenceClassification, __webpack_exports__MPNetForTokenClassification as MPNetForTokenClassification, __webpack_exports__MPNetModel as MPNetModel, __webpack_exports__MPNetPreTrainedModel as MPNetPreTrainedModel, __webpack_exports__MPNetTokenizer as MPNetTokenizer, __webpack_exports__MT5ForConditionalGeneration as MT5ForConditionalGeneration, __webpack_exports__MT5Model as MT5Model, __webpack_exports__MT5PreTrainedModel as MT5PreTrainedModel, __webpack_exports__MarianMTModel as MarianMTModel, __webpack_exports__MarianModel as MarianModel, __webpack_exports__MarianPreTrainedModel as MarianPreTrainedModel, __webpack_exports__MarianTokenizer as MarianTokenizer, __webpack_exports__Mask2FormerImageProcessor as Mask2FormerImageProcessor, __webpack_exports__MaskFormerFeatureExtractor as MaskFormerFeatureExtractor, __webpack_exports__MaskFormerForInstanceSegmentation as MaskFormerForInstanceSegmentation, __webpack_exports__MaskFormerImageProcessor as MaskFormerImageProcessor, __webpack_exports__MaskFormerModel as MaskFormerModel, __webpack_exports__MaskFormerPreTrainedModel as MaskFormerPreTrainedModel, __webpack_exports__MaskedLMOutput as MaskedLMOutput, __webpack_exports__MaxLengthCriteria as MaxLengthCriteria, __webpack_exports__MgpstrForSceneTextRecognition as MgpstrForSceneTextRecognition, __webpack_exports__MgpstrModelOutput as MgpstrModelOutput, __webpack_exports__MgpstrPreTrainedModel as MgpstrPreTrainedModel, __webpack_exports__MgpstrProcessor as MgpstrProcessor, __webpack_exports__MgpstrTokenizer as MgpstrTokenizer, __webpack_exports__MimiDecoderModel as MimiDecoderModel, __webpack_exports__MimiDecoderOutput as MimiDecoderOutput, __webpack_exports__MimiEncoderModel as MimiEncoderModel, __webpack_exports__MimiEncoderOutput as MimiEncoderOutput, __webpack_exports__MimiModel as MimiModel, __webpack_exports__MimiPreTrainedModel as MimiPreTrainedModel, __webpack_exports__MinLengthLogitsProcessor as MinLengthLogitsProcessor, __webpack_exports__MinNewTokensLengthLogitsProcessor as MinNewTokensLengthLogitsProcessor, __webpack_exports__MistralForCausalLM as MistralForCausalLM, __webpack_exports__MistralModel as MistralModel, __webpack_exports__MistralPreTrainedModel as MistralPreTrainedModel, __webpack_exports__MobileBertForMaskedLM as MobileBertForMaskedLM, __webpack_exports__MobileBertForQuestionAnswering as MobileBertForQuestionAnswering, __webpack_exports__MobileBertForSequenceClassification as MobileBertForSequenceClassification, __webpack_exports__MobileBertModel as MobileBertModel, __webpack_exports__MobileBertPreTrainedModel as MobileBertPreTrainedModel, __webpack_exports__MobileBertTokenizer as MobileBertTokenizer, __webpack_exports__MobileLLMForCausalLM as MobileLLMForCausalLM, __webpack_exports__MobileLLMModel as MobileLLMModel, __webpack_exports__MobileLLMPreTrainedModel as MobileLLMPreTrainedModel, __webpack_exports__MobileNetV1FeatureExtractor as MobileNetV1FeatureExtractor, __webpack_exports__MobileNetV1ForImageClassification as MobileNetV1ForImageClassification, __webpack_exports__MobileNetV1ForSemanticSegmentation as MobileNetV1ForSemanticSegmentation, __webpack_exports__MobileNetV1ImageProcessor as MobileNetV1ImageProcessor, __webpack_exports__MobileNetV1Model as MobileNetV1Model, __webpack_exports__MobileNetV1PreTrainedModel as MobileNetV1PreTrainedModel, __webpack_exports__MobileNetV2FeatureExtractor as MobileNetV2FeatureExtractor, __webpack_exports__MobileNetV2ForImageClassification as MobileNetV2ForImageClassification, __webpack_exports__MobileNetV2ForSemanticSegmentation as MobileNetV2ForSemanticSegmentation, __webpack_exports__MobileNetV2ImageProcessor as MobileNetV2ImageProcessor, __webpack_exports__MobileNetV2Model as MobileNetV2Model, __webpack_exports__MobileNetV2PreTrainedModel as MobileNetV2PreTrainedModel, __webpack_exports__MobileNetV3FeatureExtractor as MobileNetV3FeatureExtractor, __webpack_exports__MobileNetV3ForImageClassification as MobileNetV3ForImageClassification, __webpack_exports__MobileNetV3ForSemanticSegmentation as MobileNetV3ForSemanticSegmentation, __webpack_exports__MobileNetV3ImageProcessor as MobileNetV3ImageProcessor, __webpack_exports__MobileNetV3Model as MobileNetV3Model, __webpack_exports__MobileNetV3PreTrainedModel as MobileNetV3PreTrainedModel, __webpack_exports__MobileNetV4FeatureExtractor as MobileNetV4FeatureExtractor, __webpack_exports__MobileNetV4ForImageClassification as MobileNetV4ForImageClassification, __webpack_exports__MobileNetV4ForSemanticSegmentation as MobileNetV4ForSemanticSegmentation, __webpack_exports__MobileNetV4ImageProcessor as MobileNetV4ImageProcessor, __webpack_exports__MobileNetV4Model as MobileNetV4Model, __webpack_exports__MobileNetV4PreTrainedModel as MobileNetV4PreTrainedModel, __webpack_exports__MobileViTFeatureExtractor as MobileViTFeatureExtractor, __webpack_exports__MobileViTForImageClassification as MobileViTForImageClassification, __webpack_exports__MobileViTImageProcessor as MobileViTImageProcessor, __webpack_exports__MobileViTModel as MobileViTModel, __webpack_exports__MobileViTPreTrainedModel as MobileViTPreTrainedModel, __webpack_exports__MobileViTV2ForImageClassification as MobileViTV2ForImageClassification, __webpack_exports__MobileViTV2Model as MobileViTV2Model, __webpack_exports__MobileViTV2PreTrainedModel as MobileViTV2PreTrainedModel, __webpack_exports__ModelOutput as ModelOutput, __webpack_exports__ModernBertForMaskedLM as ModernBertForMaskedLM, __webpack_exports__ModernBertForSequenceClassification as ModernBertForSequenceClassification, __webpack_exports__ModernBertForTokenClassification as ModernBertForTokenClassification, __webpack_exports__ModernBertModel as ModernBertModel, __webpack_exports__ModernBertPreTrainedModel as ModernBertPreTrainedModel, __webpack_exports__Moondream1ForConditionalGeneration as Moondream1ForConditionalGeneration, __webpack_exports__MoonshineFeatureExtractor as MoonshineFeatureExtractor, __webpack_exports__MoonshineForConditionalGeneration as MoonshineForConditionalGeneration, __webpack_exports__MoonshineModel as MoonshineModel, __webpack_exports__MoonshinePreTrainedModel as MoonshinePreTrainedModel, __webpack_exports__MoonshineProcessor as MoonshineProcessor, __webpack_exports__MptForCausalLM as MptForCausalLM, __webpack_exports__MptModel as MptModel, __webpack_exports__MptPreTrainedModel as MptPreTrainedModel, __webpack_exports__MultiModalityCausalLM as MultiModalityCausalLM, __webpack_exports__MultiModalityPreTrainedModel as MultiModalityPreTrainedModel, __webpack_exports__MusicgenForCausalLM as MusicgenForCausalLM, __webpack_exports__MusicgenForConditionalGeneration as MusicgenForConditionalGeneration, __webpack_exports__MusicgenModel as MusicgenModel, __webpack_exports__MusicgenPreTrainedModel as MusicgenPreTrainedModel, __webpack_exports__NllbTokenizer as NllbTokenizer, __webpack_exports__NoBadWordsLogitsProcessor as NoBadWordsLogitsProcessor, __webpack_exports__NoRepeatNGramLogitsProcessor as NoRepeatNGramLogitsProcessor, __webpack_exports__NomicBertModel as NomicBertModel, __webpack_exports__NomicBertPreTrainedModel as NomicBertPreTrainedModel, __webpack_exports__NougatImageProcessor as NougatImageProcessor, __webpack_exports__NougatTokenizer as NougatTokenizer, __webpack_exports__OPTForCausalLM as OPTForCausalLM, __webpack_exports__OPTModel as OPTModel, __webpack_exports__OPTPreTrainedModel as OPTPreTrainedModel, __webpack_exports__ObjectDetectionPipeline as ObjectDetectionPipeline, __webpack_exports__Olmo2ForCausalLM as Olmo2ForCausalLM, __webpack_exports__Olmo2Model as Olmo2Model, __webpack_exports__Olmo2PreTrainedModel as Olmo2PreTrainedModel, __webpack_exports__OlmoForCausalLM as OlmoForCausalLM, __webpack_exports__OlmoModel as OlmoModel, __webpack_exports__OlmoPreTrainedModel as OlmoPreTrainedModel, __webpack_exports__OpenELMForCausalLM as OpenELMForCausalLM, __webpack_exports__OpenELMModel as OpenELMModel, __webpack_exports__OpenELMPreTrainedModel as OpenELMPreTrainedModel, __webpack_exports__OwlViTFeatureExtractor as OwlViTFeatureExtractor, __webpack_exports__OwlViTForObjectDetection as OwlViTForObjectDetection, __webpack_exports__OwlViTImageProcessor as OwlViTImageProcessor, __webpack_exports__OwlViTModel as OwlViTModel, __webpack_exports__OwlViTPreTrainedModel as OwlViTPreTrainedModel, __webpack_exports__OwlViTProcessor as OwlViTProcessor, __webpack_exports__Owlv2ForObjectDetection as Owlv2ForObjectDetection, __webpack_exports__Owlv2ImageProcessor as Owlv2ImageProcessor, __webpack_exports__Owlv2Model as Owlv2Model, __webpack_exports__Owlv2PreTrainedModel as Owlv2PreTrainedModel, __webpack_exports__PaliGemmaForConditionalGeneration as PaliGemmaForConditionalGeneration, __webpack_exports__PaliGemmaPreTrainedModel as PaliGemmaPreTrainedModel, __webpack_exports__PaliGemmaProcessor as PaliGemmaProcessor, __webpack_exports__PatchTSMixerForPrediction as PatchTSMixerForPrediction, __webpack_exports__PatchTSMixerModel as PatchTSMixerModel, __webpack_exports__PatchTSMixerPreTrainedModel as PatchTSMixerPreTrainedModel, __webpack_exports__PatchTSTForPrediction as PatchTSTForPrediction, __webpack_exports__PatchTSTModel as PatchTSTModel, __webpack_exports__PatchTSTPreTrainedModel as PatchTSTPreTrainedModel, __webpack_exports__Phi3ForCausalLM as Phi3ForCausalLM, __webpack_exports__Phi3Model as Phi3Model, __webpack_exports__Phi3PreTrainedModel as Phi3PreTrainedModel, __webpack_exports__Phi3VForCausalLM as Phi3VForCausalLM, __webpack_exports__Phi3VImageProcessor as Phi3VImageProcessor, __webpack_exports__Phi3VPreTrainedModel as Phi3VPreTrainedModel, __webpack_exports__Phi3VProcessor as Phi3VProcessor, __webpack_exports__PhiForCausalLM as PhiForCausalLM, __webpack_exports__PhiModel as PhiModel, __webpack_exports__PhiPreTrainedModel as PhiPreTrainedModel, __webpack_exports__Pipeline as Pipeline, __webpack_exports__PreTrainedModel as PreTrainedModel, __webpack_exports__PreTrainedTokenizer as PreTrainedTokenizer, __webpack_exports__PretrainedConfig as PretrainedConfig, __webpack_exports__PretrainedMixin as PretrainedMixin, __webpack_exports__Processor as Processor, __webpack_exports__PvtForImageClassification as PvtForImageClassification, __webpack_exports__PvtImageProcessor as PvtImageProcessor, __webpack_exports__PvtModel as PvtModel, __webpack_exports__PvtPreTrainedModel as PvtPreTrainedModel, __webpack_exports__PyAnnoteFeatureExtractor as PyAnnoteFeatureExtractor, __webpack_exports__PyAnnoteForAudioFrameClassification as PyAnnoteForAudioFrameClassification, __webpack_exports__PyAnnoteModel as PyAnnoteModel, __webpack_exports__PyAnnotePreTrainedModel as PyAnnotePreTrainedModel, __webpack_exports__PyAnnoteProcessor as PyAnnoteProcessor, __webpack_exports__QuestionAnsweringModelOutput as QuestionAnsweringModelOutput, __webpack_exports__QuestionAnsweringPipeline as QuestionAnsweringPipeline, __webpack_exports__Qwen2ForCausalLM as Qwen2ForCausalLM, __webpack_exports__Qwen2Model as Qwen2Model, __webpack_exports__Qwen2PreTrainedModel as Qwen2PreTrainedModel, __webpack_exports__Qwen2Tokenizer as Qwen2Tokenizer, __webpack_exports__Qwen2VLForConditionalGeneration as Qwen2VLForConditionalGeneration, __webpack_exports__Qwen2VLImageProcessor as Qwen2VLImageProcessor, __webpack_exports__Qwen2VLPreTrainedModel as Qwen2VLPreTrainedModel, __webpack_exports__Qwen2VLProcessor as Qwen2VLProcessor, __webpack_exports__RTDetrForObjectDetection as RTDetrForObjectDetection, __webpack_exports__RTDetrImageProcessor as RTDetrImageProcessor, __webpack_exports__RTDetrModel as RTDetrModel, __webpack_exports__RTDetrObjectDetectionOutput as RTDetrObjectDetectionOutput, __webpack_exports__RTDetrPreTrainedModel as RTDetrPreTrainedModel, __webpack_exports__RawAudio as RawAudio, __webpack_exports__RawImage as RawImage, __webpack_exports__RawVideo as RawVideo, __webpack_exports__RawVideoFrame as RawVideoFrame, __webpack_exports__RepetitionPenaltyLogitsProcessor as RepetitionPenaltyLogitsProcessor, __webpack_exports__ResNetForImageClassification as ResNetForImageClassification, __webpack_exports__ResNetModel as ResNetModel, __webpack_exports__ResNetPreTrainedModel as ResNetPreTrainedModel, __webpack_exports__RoFormerForMaskedLM as RoFormerForMaskedLM, __webpack_exports__RoFormerForQuestionAnswering as RoFormerForQuestionAnswering, __webpack_exports__RoFormerForSequenceClassification as RoFormerForSequenceClassification, __webpack_exports__RoFormerForTokenClassification as RoFormerForTokenClassification, __webpack_exports__RoFormerModel as RoFormerModel, __webpack_exports__RoFormerPreTrainedModel as RoFormerPreTrainedModel, __webpack_exports__RoFormerTokenizer as RoFormerTokenizer, __webpack_exports__RobertaForMaskedLM as RobertaForMaskedLM, __webpack_exports__RobertaForQuestionAnswering as RobertaForQuestionAnswering, __webpack_exports__RobertaForSequenceClassification as RobertaForSequenceClassification, __webpack_exports__RobertaForTokenClassification as RobertaForTokenClassification, __webpack_exports__RobertaModel as RobertaModel, __webpack_exports__RobertaPreTrainedModel as RobertaPreTrainedModel, __webpack_exports__RobertaTokenizer as RobertaTokenizer, __webpack_exports__SamImageProcessor as SamImageProcessor, __webpack_exports__SamImageSegmentationOutput as SamImageSegmentationOutput, __webpack_exports__SamModel as SamModel, __webpack_exports__SamPreTrainedModel as SamPreTrainedModel, __webpack_exports__SamProcessor as SamProcessor, __webpack_exports__SapiensForDepthEstimation as SapiensForDepthEstimation, __webpack_exports__SapiensForNormalEstimation as SapiensForNormalEstimation, __webpack_exports__SapiensForSemanticSegmentation as SapiensForSemanticSegmentation, __webpack_exports__SapiensPreTrainedModel as SapiensPreTrainedModel, __webpack_exports__SeamlessM4TFeatureExtractor as SeamlessM4TFeatureExtractor, __webpack_exports__SegformerFeatureExtractor as SegformerFeatureExtractor, __webpack_exports__SegformerForImageClassification as SegformerForImageClassification, __webpack_exports__SegformerForSemanticSegmentation as SegformerForSemanticSegmentation, __webpack_exports__SegformerImageProcessor as SegformerImageProcessor, __webpack_exports__SegformerModel as SegformerModel, __webpack_exports__SegformerPreTrainedModel as SegformerPreTrainedModel, __webpack_exports__Seq2SeqLMOutput as Seq2SeqLMOutput, __webpack_exports__SequenceClassifierOutput as SequenceClassifierOutput, __webpack_exports__SiglipImageProcessor as SiglipImageProcessor, __webpack_exports__SiglipModel as SiglipModel, __webpack_exports__SiglipPreTrainedModel as SiglipPreTrainedModel, __webpack_exports__SiglipTextModel as SiglipTextModel, __webpack_exports__SiglipTokenizer as SiglipTokenizer, __webpack_exports__SiglipVisionModel as SiglipVisionModel, __webpack_exports__SmolVLMForConditionalGeneration as SmolVLMForConditionalGeneration, __webpack_exports__SmolVLMImageProcessor as SmolVLMImageProcessor, __webpack_exports__SmolVLMProcessor as SmolVLMProcessor, __webpack_exports__SpeechT5FeatureExtractor as SpeechT5FeatureExtractor, __webpack_exports__SpeechT5ForSpeechToText as SpeechT5ForSpeechToText, __webpack_exports__SpeechT5ForTextToSpeech as SpeechT5ForTextToSpeech, __webpack_exports__SpeechT5HifiGan as SpeechT5HifiGan, __webpack_exports__SpeechT5Model as SpeechT5Model, __webpack_exports__SpeechT5PreTrainedModel as SpeechT5PreTrainedModel, __webpack_exports__SpeechT5Processor as SpeechT5Processor, __webpack_exports__SpeechT5Tokenizer as SpeechT5Tokenizer, __webpack_exports__SqueezeBertForMaskedLM as SqueezeBertForMaskedLM, __webpack_exports__SqueezeBertForQuestionAnswering as SqueezeBertForQuestionAnswering, __webpack_exports__SqueezeBertForSequenceClassification as SqueezeBertForSequenceClassification, __webpack_exports__SqueezeBertModel as SqueezeBertModel, __webpack_exports__SqueezeBertPreTrainedModel as SqueezeBertPreTrainedModel, __webpack_exports__SqueezeBertTokenizer as SqueezeBertTokenizer, __webpack_exports__StableLmForCausalLM as StableLmForCausalLM, __webpack_exports__StableLmModel as StableLmModel, __webpack_exports__StableLmPreTrainedModel as StableLmPreTrainedModel, __webpack_exports__Starcoder2ForCausalLM as Starcoder2ForCausalLM, __webpack_exports__Starcoder2Model as Starcoder2Model, __webpack_exports__Starcoder2PreTrainedModel as Starcoder2PreTrainedModel, __webpack_exports__StoppingCriteria as StoppingCriteria, __webpack_exports__StoppingCriteriaList as StoppingCriteriaList, __webpack_exports__StyleTextToSpeech2Model as StyleTextToSpeech2Model, __webpack_exports__StyleTextToSpeech2PreTrainedModel as StyleTextToSpeech2PreTrainedModel, __webpack_exports__SummarizationPipeline as SummarizationPipeline, __webpack_exports__SuppressTokensAtBeginLogitsProcessor as SuppressTokensAtBeginLogitsProcessor, __webpack_exports__Swin2SRForImageSuperResolution as Swin2SRForImageSuperResolution, __webpack_exports__Swin2SRImageProcessor as Swin2SRImageProcessor, __webpack_exports__Swin2SRModel as Swin2SRModel, __webpack_exports__Swin2SRPreTrainedModel as Swin2SRPreTrainedModel, __webpack_exports__SwinForImageClassification as SwinForImageClassification, __webpack_exports__SwinForSemanticSegmentation as SwinForSemanticSegmentation, __webpack_exports__SwinModel as SwinModel, __webpack_exports__SwinPreTrainedModel as SwinPreTrainedModel, __webpack_exports__T5ForConditionalGeneration as T5ForConditionalGeneration, __webpack_exports__T5Model as T5Model, __webpack_exports__T5PreTrainedModel as T5PreTrainedModel, __webpack_exports__T5Tokenizer as T5Tokenizer, __webpack_exports__TableTransformerForObjectDetection as TableTransformerForObjectDetection, __webpack_exports__TableTransformerModel as TableTransformerModel, __webpack_exports__TableTransformerObjectDetectionOutput as TableTransformerObjectDetectionOutput, __webpack_exports__TableTransformerPreTrainedModel as TableTransformerPreTrainedModel, __webpack_exports__TemperatureLogitsWarper as TemperatureLogitsWarper, __webpack_exports__Tensor as Tensor, __webpack_exports__Text2TextGenerationPipeline as Text2TextGenerationPipeline, __webpack_exports__TextClassificationPipeline as TextClassificationPipeline, __webpack_exports__TextGenerationPipeline as TextGenerationPipeline, __webpack_exports__TextStreamer as TextStreamer, __webpack_exports__TextToAudioPipeline as TextToAudioPipeline, __webpack_exports__TokenClassificationPipeline as TokenClassificationPipeline, __webpack_exports__TokenClassifierOutput as TokenClassifierOutput, __webpack_exports__TokenizerModel as TokenizerModel, __webpack_exports__TopKLogitsWarper as TopKLogitsWarper, __webpack_exports__TopPLogitsWarper as TopPLogitsWarper, __webpack_exports__TrOCRForCausalLM as TrOCRForCausalLM, __webpack_exports__TrOCRPreTrainedModel as TrOCRPreTrainedModel, __webpack_exports__TranslationPipeline as TranslationPipeline, __webpack_exports__UltravoxModel as UltravoxModel, __webpack_exports__UltravoxPreTrainedModel as UltravoxPreTrainedModel, __webpack_exports__UltravoxProcessor as UltravoxProcessor, __webpack_exports__UniSpeechForCTC as UniSpeechForCTC, __webpack_exports__UniSpeechForSequenceClassification as UniSpeechForSequenceClassification, __webpack_exports__UniSpeechModel as UniSpeechModel, __webpack_exports__UniSpeechPreTrainedModel as UniSpeechPreTrainedModel, __webpack_exports__UniSpeechSatForAudioFrameClassification as UniSpeechSatForAudioFrameClassification, __webpack_exports__UniSpeechSatForCTC as UniSpeechSatForCTC, __webpack_exports__UniSpeechSatForSequenceClassification as UniSpeechSatForSequenceClassification, __webpack_exports__UniSpeechSatModel as UniSpeechSatModel, __webpack_exports__UniSpeechSatPreTrainedModel as UniSpeechSatPreTrainedModel, __webpack_exports__VLChatProcessor as VLChatProcessor, __webpack_exports__VLMImageProcessor as VLMImageProcessor, __webpack_exports__ViTFeatureExtractor as ViTFeatureExtractor, __webpack_exports__ViTForImageClassification as ViTForImageClassification, __webpack_exports__ViTImageProcessor as ViTImageProcessor, __webpack_exports__ViTMAEModel as ViTMAEModel, __webpack_exports__ViTMAEPreTrainedModel as ViTMAEPreTrainedModel, __webpack_exports__ViTMSNForImageClassification as ViTMSNForImageClassification, __webpack_exports__ViTMSNModel as ViTMSNModel, __webpack_exports__ViTMSNPreTrainedModel as ViTMSNPreTrainedModel, __webpack_exports__ViTModel as ViTModel, __webpack_exports__ViTPreTrainedModel as ViTPreTrainedModel, __webpack_exports__VisionEncoderDecoderModel as VisionEncoderDecoderModel, __webpack_exports__VitMatteForImageMatting as VitMatteForImageMatting, __webpack_exports__VitMatteImageProcessor as VitMatteImageProcessor, __webpack_exports__VitMattePreTrainedModel as VitMattePreTrainedModel, __webpack_exports__VitPoseForPoseEstimation as VitPoseForPoseEstimation, __webpack_exports__VitPoseImageProcessor as VitPoseImageProcessor, __webpack_exports__VitPosePreTrainedModel as VitPosePreTrainedModel, __webpack_exports__VitsModel as VitsModel, __webpack_exports__VitsModelOutput as VitsModelOutput, __webpack_exports__VitsPreTrainedModel as VitsPreTrainedModel, __webpack_exports__VitsTokenizer as VitsTokenizer, __webpack_exports__Wav2Vec2BertForCTC as Wav2Vec2BertForCTC, __webpack_exports__Wav2Vec2BertForSequenceClassification as Wav2Vec2BertForSequenceClassification, __webpack_exports__Wav2Vec2BertModel as Wav2Vec2BertModel, __webpack_exports__Wav2Vec2BertPreTrainedModel as Wav2Vec2BertPreTrainedModel, __webpack_exports__Wav2Vec2CTCTokenizer as Wav2Vec2CTCTokenizer, __webpack_exports__Wav2Vec2FeatureExtractor as Wav2Vec2FeatureExtractor, __webpack_exports__Wav2Vec2ForAudioFrameClassification as Wav2Vec2ForAudioFrameClassification, __webpack_exports__Wav2Vec2ForCTC as Wav2Vec2ForCTC, __webpack_exports__Wav2Vec2ForSequenceClassification as Wav2Vec2ForSequenceClassification, __webpack_exports__Wav2Vec2Model as Wav2Vec2Model, __webpack_exports__Wav2Vec2PreTrainedModel as Wav2Vec2PreTrainedModel, __webpack_exports__Wav2Vec2Processor as Wav2Vec2Processor, __webpack_exports__Wav2Vec2ProcessorWithLM as Wav2Vec2ProcessorWithLM, __webpack_exports__WavLMForAudioFrameClassification as WavLMForAudioFrameClassification, __webpack_exports__WavLMForCTC as WavLMForCTC, __webpack_exports__WavLMForSequenceClassification as WavLMForSequenceClassification, __webpack_exports__WavLMForXVector as WavLMForXVector, __webpack_exports__WavLMModel as WavLMModel, __webpack_exports__WavLMPreTrainedModel as WavLMPreTrainedModel, __webpack_exports__WeSpeakerFeatureExtractor as WeSpeakerFeatureExtractor, __webpack_exports__WeSpeakerResNetModel as WeSpeakerResNetModel, __webpack_exports__WeSpeakerResNetPreTrainedModel as WeSpeakerResNetPreTrainedModel, __webpack_exports__WhisperFeatureExtractor as WhisperFeatureExtractor, __webpack_exports__WhisperForConditionalGeneration as WhisperForConditionalGeneration, __webpack_exports__WhisperModel as WhisperModel, __webpack_exports__WhisperPreTrainedModel as WhisperPreTrainedModel, __webpack_exports__WhisperProcessor as WhisperProcessor, __webpack_exports__WhisperTextStreamer as WhisperTextStreamer, __webpack_exports__WhisperTimeStampLogitsProcessor as WhisperTimeStampLogitsProcessor, __webpack_exports__WhisperTokenizer as WhisperTokenizer, __webpack_exports__XLMForQuestionAnswering as XLMForQuestionAnswering, __webpack_exports__XLMForSequenceClassification as XLMForSequenceClassification, __webpack_exports__XLMForTokenClassification as XLMForTokenClassification, __webpack_exports__XLMModel as XLMModel, __webpack_exports__XLMPreTrainedModel as XLMPreTrainedModel, __webpack_exports__XLMRobertaForMaskedLM as XLMRobertaForMaskedLM, __webpack_exports__XLMRobertaForQuestionAnswering as XLMRobertaForQuestionAnswering, __webpack_exports__XLMRobertaForSequenceClassification as XLMRobertaForSequenceClassification, __webpack_exports__XLMRobertaForTokenClassification as XLMRobertaForTokenClassification, __webpack_exports__XLMRobertaModel as XLMRobertaModel, __webpack_exports__XLMRobertaPreTrainedModel as XLMRobertaPreTrainedModel, __webpack_exports__XLMRobertaTokenizer as XLMRobertaTokenizer, __webpack_exports__XLMTokenizer as XLMTokenizer, __webpack_exports__XLMWithLMHeadModel as XLMWithLMHeadModel, __webpack_exports__XVectorOutput as XVectorOutput, __webpack_exports__YolosFeatureExtractor as YolosFeatureExtractor, __webpack_exports__YolosForObjectDetection as YolosForObjectDetection, __webpack_exports__YolosImageProcessor as YolosImageProcessor, __webpack_exports__YolosModel as YolosModel, __webpack_exports__YolosObjectDetectionOutput as YolosObjectDetectionOutput, __webpack_exports__YolosPreTrainedModel as YolosPreTrainedModel, __webpack_exports__ZeroShotAudioClassificationPipeline as ZeroShotAudioClassificationPipeline, __webpack_exports__ZeroShotClassificationPipeline as ZeroShotClassificationPipeline, __webpack_exports__ZeroShotImageClassificationPipeline as ZeroShotImageClassificationPipeline, __webpack_exports__ZeroShotObjectDetectionPipeline as ZeroShotObjectDetectionPipeline, __webpack_exports__bankers_round as bankers_round, __webpack_exports__cat as cat, __webpack_exports__cos_sim as cos_sim, __webpack_exports__dot as dot, __webpack_exports__dynamic_time_warping as dynamic_time_warping, __webpack_exports__env as env, __webpack_exports__full as full, __webpack_exports__full_like as full_like, __webpack_exports__getKeyValueShapes as getKeyValueShapes, __webpack_exports__hamming as hamming, __webpack_exports__hanning as hanning, __webpack_exports__interpolate as interpolate, __webpack_exports__interpolate_4d as interpolate_4d, __webpack_exports__interpolate_data as interpolate_data, __webpack_exports__is_chinese_char as is_chinese_char, __webpack_exports__layer_norm as layer_norm, __webpack_exports__load_image as load_image, __webpack_exports__load_video as load_video, __webpack_exports__log_softmax as log_softmax, __webpack_exports__magnitude as magnitude, __webpack_exports__matmul as matmul, __webpack_exports__max as max, __webpack_exports__mean as mean, __webpack_exports__mean_pooling as mean_pooling, __webpack_exports__medianFilter as medianFilter, __webpack_exports__mel_filter_bank as mel_filter_bank, __webpack_exports__min as min, __webpack_exports__ones as ones, __webpack_exports__ones_like as ones_like, __webpack_exports__permute as permute, __webpack_exports__permute_data as permute_data, __webpack_exports__pipeline as pipeline, __webpack_exports__quantize_embeddings as quantize_embeddings, __webpack_exports__rand as rand, __webpack_exports__read_audio as read_audio, __webpack_exports__rfft as rfft, __webpack_exports__round as round, __webpack_exports__slice as slice, __webpack_exports__softmax as softmax, __webpack_exports__spectrogram as spectrogram, __webpack_exports__stack as stack, __webpack_exports__std_mean as std_mean, __webpack_exports__topk as topk, __webpack_exports__window_function as window_function, __webpack_exports__zeros as zeros, __webpack_exports__zeros_like as zeros_like };
40694
+ export { __webpack_exports__ASTFeatureExtractor as ASTFeatureExtractor, __webpack_exports__ASTForAudioClassification as ASTForAudioClassification, __webpack_exports__ASTModel as ASTModel, __webpack_exports__ASTPreTrainedModel as ASTPreTrainedModel, __webpack_exports__AlbertForMaskedLM as AlbertForMaskedLM, __webpack_exports__AlbertForQuestionAnswering as AlbertForQuestionAnswering, __webpack_exports__AlbertForSequenceClassification as AlbertForSequenceClassification, __webpack_exports__AlbertModel as AlbertModel, __webpack_exports__AlbertPreTrainedModel as AlbertPreTrainedModel, __webpack_exports__AlbertTokenizer as AlbertTokenizer, __webpack_exports__AudioClassificationPipeline as AudioClassificationPipeline, __webpack_exports__AutoConfig as AutoConfig, __webpack_exports__AutoFeatureExtractor as AutoFeatureExtractor, __webpack_exports__AutoImageProcessor as AutoImageProcessor, __webpack_exports__AutoModel as AutoModel, __webpack_exports__AutoModelForAudioClassification as AutoModelForAudioClassification, __webpack_exports__AutoModelForAudioFrameClassification as AutoModelForAudioFrameClassification, __webpack_exports__AutoModelForAudioTextToText as AutoModelForAudioTextToText, __webpack_exports__AutoModelForCTC as AutoModelForCTC, __webpack_exports__AutoModelForCausalLM as AutoModelForCausalLM, __webpack_exports__AutoModelForDepthEstimation as AutoModelForDepthEstimation, __webpack_exports__AutoModelForDocumentQuestionAnswering as AutoModelForDocumentQuestionAnswering, __webpack_exports__AutoModelForImageClassification as AutoModelForImageClassification, __webpack_exports__AutoModelForImageFeatureExtraction as AutoModelForImageFeatureExtraction, __webpack_exports__AutoModelForImageMatting as AutoModelForImageMatting, __webpack_exports__AutoModelForImageSegmentation as AutoModelForImageSegmentation, __webpack_exports__AutoModelForImageTextToText as AutoModelForImageTextToText, __webpack_exports__AutoModelForImageToImage as AutoModelForImageToImage, __webpack_exports__AutoModelForMaskGeneration as AutoModelForMaskGeneration, __webpack_exports__AutoModelForMaskedLM as AutoModelForMaskedLM, __webpack_exports__AutoModelForNormalEstimation as AutoModelForNormalEstimation, __webpack_exports__AutoModelForObjectDetection as AutoModelForObjectDetection, __webpack_exports__AutoModelForPoseEstimation as AutoModelForPoseEstimation, __webpack_exports__AutoModelForQuestionAnswering as AutoModelForQuestionAnswering, __webpack_exports__AutoModelForSemanticSegmentation as AutoModelForSemanticSegmentation, __webpack_exports__AutoModelForSeq2SeqLM as AutoModelForSeq2SeqLM, __webpack_exports__AutoModelForSequenceClassification as AutoModelForSequenceClassification, __webpack_exports__AutoModelForSpeechSeq2Seq as AutoModelForSpeechSeq2Seq, __webpack_exports__AutoModelForTextToSpectrogram as AutoModelForTextToSpectrogram, __webpack_exports__AutoModelForTextToWaveform as AutoModelForTextToWaveform, __webpack_exports__AutoModelForTokenClassification as AutoModelForTokenClassification, __webpack_exports__AutoModelForUniversalSegmentation as AutoModelForUniversalSegmentation, __webpack_exports__AutoModelForVision2Seq as AutoModelForVision2Seq, __webpack_exports__AutoModelForXVector as AutoModelForXVector, __webpack_exports__AutoModelForZeroShotObjectDetection as AutoModelForZeroShotObjectDetection, __webpack_exports__AutoProcessor as AutoProcessor, __webpack_exports__AutoTokenizer as AutoTokenizer, __webpack_exports__AutomaticSpeechRecognitionPipeline as AutomaticSpeechRecognitionPipeline, __webpack_exports__BackgroundRemovalPipeline as BackgroundRemovalPipeline, __webpack_exports__BartForConditionalGeneration as BartForConditionalGeneration, __webpack_exports__BartForSequenceClassification as BartForSequenceClassification, __webpack_exports__BartModel as BartModel, __webpack_exports__BartPretrainedModel as BartPretrainedModel, __webpack_exports__BartTokenizer as BartTokenizer, __webpack_exports__BaseModelOutput as BaseModelOutput, __webpack_exports__BaseStreamer as BaseStreamer, __webpack_exports__BeitFeatureExtractor as BeitFeatureExtractor, __webpack_exports__BeitForImageClassification as BeitForImageClassification, __webpack_exports__BeitModel as BeitModel, __webpack_exports__BeitPreTrainedModel as BeitPreTrainedModel, __webpack_exports__BertForMaskedLM as BertForMaskedLM, __webpack_exports__BertForQuestionAnswering as BertForQuestionAnswering, __webpack_exports__BertForSequenceClassification as BertForSequenceClassification, __webpack_exports__BertForTokenClassification as BertForTokenClassification, __webpack_exports__BertModel as BertModel, __webpack_exports__BertPreTrainedModel as BertPreTrainedModel, __webpack_exports__BertTokenizer as BertTokenizer, __webpack_exports__BitImageProcessor as BitImageProcessor, __webpack_exports__BlenderbotForConditionalGeneration as BlenderbotForConditionalGeneration, __webpack_exports__BlenderbotModel as BlenderbotModel, __webpack_exports__BlenderbotPreTrainedModel as BlenderbotPreTrainedModel, __webpack_exports__BlenderbotSmallForConditionalGeneration as BlenderbotSmallForConditionalGeneration, __webpack_exports__BlenderbotSmallModel as BlenderbotSmallModel, __webpack_exports__BlenderbotSmallPreTrainedModel as BlenderbotSmallPreTrainedModel, __webpack_exports__BlenderbotSmallTokenizer as BlenderbotSmallTokenizer, __webpack_exports__BlenderbotTokenizer as BlenderbotTokenizer, __webpack_exports__BloomForCausalLM as BloomForCausalLM, __webpack_exports__BloomModel as BloomModel, __webpack_exports__BloomPreTrainedModel as BloomPreTrainedModel, __webpack_exports__BloomTokenizer as BloomTokenizer, __webpack_exports__CLIPFeatureExtractor as CLIPFeatureExtractor, __webpack_exports__CLIPImageProcessor as CLIPImageProcessor, __webpack_exports__CLIPModel as CLIPModel, __webpack_exports__CLIPPreTrainedModel as CLIPPreTrainedModel, __webpack_exports__CLIPSegForImageSegmentation as CLIPSegForImageSegmentation, __webpack_exports__CLIPSegModel as CLIPSegModel, __webpack_exports__CLIPSegPreTrainedModel as CLIPSegPreTrainedModel, __webpack_exports__CLIPTextModel as CLIPTextModel, __webpack_exports__CLIPTextModelWithProjection as CLIPTextModelWithProjection, __webpack_exports__CLIPTokenizer as CLIPTokenizer, __webpack_exports__CLIPVisionModel as CLIPVisionModel, __webpack_exports__CLIPVisionModelWithProjection as CLIPVisionModelWithProjection, __webpack_exports__CamembertForMaskedLM as CamembertForMaskedLM, __webpack_exports__CamembertForQuestionAnswering as CamembertForQuestionAnswering, __webpack_exports__CamembertForSequenceClassification as CamembertForSequenceClassification, __webpack_exports__CamembertForTokenClassification as CamembertForTokenClassification, __webpack_exports__CamembertModel as CamembertModel, __webpack_exports__CamembertPreTrainedModel as CamembertPreTrainedModel, __webpack_exports__CamembertTokenizer as CamembertTokenizer, __webpack_exports__CausalLMOutput as CausalLMOutput, __webpack_exports__CausalLMOutputWithPast as CausalLMOutputWithPast, __webpack_exports__ChineseCLIPFeatureExtractor as ChineseCLIPFeatureExtractor, __webpack_exports__ChineseCLIPModel as ChineseCLIPModel, __webpack_exports__ChineseCLIPPreTrainedModel as ChineseCLIPPreTrainedModel, __webpack_exports__ClapAudioModelWithProjection as ClapAudioModelWithProjection, __webpack_exports__ClapFeatureExtractor as ClapFeatureExtractor, __webpack_exports__ClapModel as ClapModel, __webpack_exports__ClapPreTrainedModel as ClapPreTrainedModel, __webpack_exports__ClapTextModelWithProjection as ClapTextModelWithProjection, __webpack_exports__ClassifierFreeGuidanceLogitsProcessor as ClassifierFreeGuidanceLogitsProcessor, __webpack_exports__CodeGenForCausalLM as CodeGenForCausalLM, __webpack_exports__CodeGenModel as CodeGenModel, __webpack_exports__CodeGenPreTrainedModel as CodeGenPreTrainedModel, __webpack_exports__CodeGenTokenizer as CodeGenTokenizer, __webpack_exports__CodeLlamaTokenizer as CodeLlamaTokenizer, __webpack_exports__CohereForCausalLM as CohereForCausalLM, __webpack_exports__CohereModel as CohereModel, __webpack_exports__CoherePreTrainedModel as CoherePreTrainedModel, __webpack_exports__CohereTokenizer as CohereTokenizer, __webpack_exports__ConvBertForMaskedLM as ConvBertForMaskedLM, __webpack_exports__ConvBertForQuestionAnswering as ConvBertForQuestionAnswering, __webpack_exports__ConvBertForSequenceClassification as ConvBertForSequenceClassification, __webpack_exports__ConvBertForTokenClassification as ConvBertForTokenClassification, __webpack_exports__ConvBertModel as ConvBertModel, __webpack_exports__ConvBertPreTrainedModel as ConvBertPreTrainedModel, __webpack_exports__ConvBertTokenizer as ConvBertTokenizer, __webpack_exports__ConvNextFeatureExtractor as ConvNextFeatureExtractor, __webpack_exports__ConvNextForImageClassification as ConvNextForImageClassification, __webpack_exports__ConvNextImageProcessor as ConvNextImageProcessor, __webpack_exports__ConvNextModel as ConvNextModel, __webpack_exports__ConvNextPreTrainedModel as ConvNextPreTrainedModel, __webpack_exports__ConvNextV2ForImageClassification as ConvNextV2ForImageClassification, __webpack_exports__ConvNextV2Model as ConvNextV2Model, __webpack_exports__ConvNextV2PreTrainedModel as ConvNextV2PreTrainedModel, __webpack_exports__DPTFeatureExtractor as DPTFeatureExtractor, __webpack_exports__DPTForDepthEstimation as DPTForDepthEstimation, __webpack_exports__DPTImageProcessor as DPTImageProcessor, __webpack_exports__DPTModel as DPTModel, __webpack_exports__DPTPreTrainedModel as DPTPreTrainedModel, __webpack_exports__DacDecoderModel as DacDecoderModel, __webpack_exports__DacDecoderOutput as DacDecoderOutput, __webpack_exports__DacEncoderModel as DacEncoderModel, __webpack_exports__DacEncoderOutput as DacEncoderOutput, __webpack_exports__DacFeatureExtractor as DacFeatureExtractor, __webpack_exports__DacModel as DacModel, __webpack_exports__DacPreTrainedModel as DacPreTrainedModel, __webpack_exports__DataTypeMap as DataTypeMap, __webpack_exports__DebertaForMaskedLM as DebertaForMaskedLM, __webpack_exports__DebertaForQuestionAnswering as DebertaForQuestionAnswering, __webpack_exports__DebertaForSequenceClassification as DebertaForSequenceClassification, __webpack_exports__DebertaForTokenClassification as DebertaForTokenClassification, __webpack_exports__DebertaModel as DebertaModel, __webpack_exports__DebertaPreTrainedModel as DebertaPreTrainedModel, __webpack_exports__DebertaTokenizer as DebertaTokenizer, __webpack_exports__DebertaV2ForMaskedLM as DebertaV2ForMaskedLM, __webpack_exports__DebertaV2ForQuestionAnswering as DebertaV2ForQuestionAnswering, __webpack_exports__DebertaV2ForSequenceClassification as DebertaV2ForSequenceClassification, __webpack_exports__DebertaV2ForTokenClassification as DebertaV2ForTokenClassification, __webpack_exports__DebertaV2Model as DebertaV2Model, __webpack_exports__DebertaV2PreTrainedModel as DebertaV2PreTrainedModel, __webpack_exports__DebertaV2Tokenizer as DebertaV2Tokenizer, __webpack_exports__DecisionTransformerModel as DecisionTransformerModel, __webpack_exports__DecisionTransformerPreTrainedModel as DecisionTransformerPreTrainedModel, __webpack_exports__DeiTFeatureExtractor as DeiTFeatureExtractor, __webpack_exports__DeiTForImageClassification as DeiTForImageClassification, __webpack_exports__DeiTImageProcessor as DeiTImageProcessor, __webpack_exports__DeiTModel as DeiTModel, __webpack_exports__DeiTPreTrainedModel as DeiTPreTrainedModel, __webpack_exports__DepthAnythingForDepthEstimation as DepthAnythingForDepthEstimation, __webpack_exports__DepthAnythingPreTrainedModel as DepthAnythingPreTrainedModel, __webpack_exports__DepthEstimationPipeline as DepthEstimationPipeline, __webpack_exports__DepthProForDepthEstimation as DepthProForDepthEstimation, __webpack_exports__DepthProPreTrainedModel as DepthProPreTrainedModel, __webpack_exports__DetrFeatureExtractor as DetrFeatureExtractor, __webpack_exports__DetrForObjectDetection as DetrForObjectDetection, __webpack_exports__DetrForSegmentation as DetrForSegmentation, __webpack_exports__DetrImageProcessor as DetrImageProcessor, __webpack_exports__DetrModel as DetrModel, __webpack_exports__DetrObjectDetectionOutput as DetrObjectDetectionOutput, __webpack_exports__DetrPreTrainedModel as DetrPreTrainedModel, __webpack_exports__DetrSegmentationOutput as DetrSegmentationOutput, __webpack_exports__Dinov2ForImageClassification as Dinov2ForImageClassification, __webpack_exports__Dinov2Model as Dinov2Model, __webpack_exports__Dinov2PreTrainedModel as Dinov2PreTrainedModel, __webpack_exports__Dinov2WithRegistersForImageClassification as Dinov2WithRegistersForImageClassification, __webpack_exports__Dinov2WithRegistersModel as Dinov2WithRegistersModel, __webpack_exports__Dinov2WithRegistersPreTrainedModel as Dinov2WithRegistersPreTrainedModel, __webpack_exports__DistilBertForMaskedLM as DistilBertForMaskedLM, __webpack_exports__DistilBertForQuestionAnswering as DistilBertForQuestionAnswering, __webpack_exports__DistilBertForSequenceClassification as DistilBertForSequenceClassification, __webpack_exports__DistilBertForTokenClassification as DistilBertForTokenClassification, __webpack_exports__DistilBertModel as DistilBertModel, __webpack_exports__DistilBertPreTrainedModel as DistilBertPreTrainedModel, __webpack_exports__DistilBertTokenizer as DistilBertTokenizer, __webpack_exports__DocumentQuestionAnsweringPipeline as DocumentQuestionAnsweringPipeline, __webpack_exports__DonutFeatureExtractor as DonutFeatureExtractor, __webpack_exports__DonutImageProcessor as DonutImageProcessor, __webpack_exports__DonutSwinModel as DonutSwinModel, __webpack_exports__DonutSwinPreTrainedModel as DonutSwinPreTrainedModel, __webpack_exports__EfficientNetForImageClassification as EfficientNetForImageClassification, __webpack_exports__EfficientNetImageProcessor as EfficientNetImageProcessor, __webpack_exports__EfficientNetModel as EfficientNetModel, __webpack_exports__EfficientNetPreTrainedModel as EfficientNetPreTrainedModel, __webpack_exports__ElectraForMaskedLM as ElectraForMaskedLM, __webpack_exports__ElectraForQuestionAnswering as ElectraForQuestionAnswering, __webpack_exports__ElectraForSequenceClassification as ElectraForSequenceClassification, __webpack_exports__ElectraForTokenClassification as ElectraForTokenClassification, __webpack_exports__ElectraModel as ElectraModel, __webpack_exports__ElectraPreTrainedModel as ElectraPreTrainedModel, __webpack_exports__ElectraTokenizer as ElectraTokenizer, __webpack_exports__EncodecFeatureExtractor as EncodecFeatureExtractor, __webpack_exports__EosTokenCriteria as EosTokenCriteria, __webpack_exports__EsmForMaskedLM as EsmForMaskedLM, __webpack_exports__EsmForSequenceClassification as EsmForSequenceClassification, __webpack_exports__EsmForTokenClassification as EsmForTokenClassification, __webpack_exports__EsmModel as EsmModel, __webpack_exports__EsmPreTrainedModel as EsmPreTrainedModel, __webpack_exports__EsmTokenizer as EsmTokenizer, __webpack_exports__ExaoneForCausalLM as ExaoneForCausalLM, __webpack_exports__ExaoneModel as ExaoneModel, __webpack_exports__ExaonePreTrainedModel as ExaonePreTrainedModel, __webpack_exports__FFT as FFT, __webpack_exports__FalconForCausalLM as FalconForCausalLM, __webpack_exports__FalconModel as FalconModel, __webpack_exports__FalconPreTrainedModel as FalconPreTrainedModel, __webpack_exports__FalconTokenizer as FalconTokenizer, __webpack_exports__FastViTForImageClassification as FastViTForImageClassification, __webpack_exports__FastViTModel as FastViTModel, __webpack_exports__FastViTPreTrainedModel as FastViTPreTrainedModel, __webpack_exports__FeatureExtractionPipeline as FeatureExtractionPipeline, __webpack_exports__FeatureExtractor as FeatureExtractor, __webpack_exports__FillMaskPipeline as FillMaskPipeline, __webpack_exports__Florence2ForConditionalGeneration as Florence2ForConditionalGeneration, __webpack_exports__Florence2PreTrainedModel as Florence2PreTrainedModel, __webpack_exports__Florence2Processor as Florence2Processor, __webpack_exports__ForcedBOSTokenLogitsProcessor as ForcedBOSTokenLogitsProcessor, __webpack_exports__ForcedEOSTokenLogitsProcessor as ForcedEOSTokenLogitsProcessor, __webpack_exports__GLPNFeatureExtractor as GLPNFeatureExtractor, __webpack_exports__GLPNForDepthEstimation as GLPNForDepthEstimation, __webpack_exports__GLPNModel as GLPNModel, __webpack_exports__GLPNPreTrainedModel as GLPNPreTrainedModel, __webpack_exports__GPT2LMHeadModel as GPT2LMHeadModel, __webpack_exports__GPT2Model as GPT2Model, __webpack_exports__GPT2PreTrainedModel as GPT2PreTrainedModel, __webpack_exports__GPT2Tokenizer as GPT2Tokenizer, __webpack_exports__GPTBigCodeForCausalLM as GPTBigCodeForCausalLM, __webpack_exports__GPTBigCodeModel as GPTBigCodeModel, __webpack_exports__GPTBigCodePreTrainedModel as GPTBigCodePreTrainedModel, __webpack_exports__GPTJForCausalLM as GPTJForCausalLM, __webpack_exports__GPTJModel as GPTJModel, __webpack_exports__GPTJPreTrainedModel as GPTJPreTrainedModel, __webpack_exports__GPTNeoForCausalLM as GPTNeoForCausalLM, __webpack_exports__GPTNeoModel as GPTNeoModel, __webpack_exports__GPTNeoPreTrainedModel as GPTNeoPreTrainedModel, __webpack_exports__GPTNeoXForCausalLM as GPTNeoXForCausalLM, __webpack_exports__GPTNeoXModel as GPTNeoXModel, __webpack_exports__GPTNeoXPreTrainedModel as GPTNeoXPreTrainedModel, __webpack_exports__GPTNeoXTokenizer as GPTNeoXTokenizer, __webpack_exports__Gemma2ForCausalLM as Gemma2ForCausalLM, __webpack_exports__Gemma2Model as Gemma2Model, __webpack_exports__Gemma2PreTrainedModel as Gemma2PreTrainedModel, __webpack_exports__Gemma3ForCausalLM as Gemma3ForCausalLM, __webpack_exports__Gemma3Model as Gemma3Model, __webpack_exports__Gemma3PreTrainedModel as Gemma3PreTrainedModel, __webpack_exports__GemmaForCausalLM as GemmaForCausalLM, __webpack_exports__GemmaModel as GemmaModel, __webpack_exports__GemmaPreTrainedModel as GemmaPreTrainedModel, __webpack_exports__GemmaTokenizer as GemmaTokenizer, __webpack_exports__GlmForCausalLM as GlmForCausalLM, __webpack_exports__GlmModel as GlmModel, __webpack_exports__GlmPreTrainedModel as GlmPreTrainedModel, __webpack_exports__GraniteForCausalLM as GraniteForCausalLM, __webpack_exports__GraniteModel as GraniteModel, __webpack_exports__GranitePreTrainedModel as GranitePreTrainedModel, __webpack_exports__Grok1Tokenizer as Grok1Tokenizer, __webpack_exports__GroundingDinoForObjectDetection as GroundingDinoForObjectDetection, __webpack_exports__GroundingDinoImageProcessor as GroundingDinoImageProcessor, __webpack_exports__GroundingDinoPreTrainedModel as GroundingDinoPreTrainedModel, __webpack_exports__GroundingDinoProcessor as GroundingDinoProcessor, __webpack_exports__GroupViTModel as GroupViTModel, __webpack_exports__GroupViTPreTrainedModel as GroupViTPreTrainedModel, __webpack_exports__HeliumForCausalLM as HeliumForCausalLM, __webpack_exports__HeliumModel as HeliumModel, __webpack_exports__HeliumPreTrainedModel as HeliumPreTrainedModel, __webpack_exports__HerbertTokenizer as HerbertTokenizer, __webpack_exports__HieraForImageClassification as HieraForImageClassification, __webpack_exports__HieraModel as HieraModel, __webpack_exports__HieraPreTrainedModel as HieraPreTrainedModel, __webpack_exports__HubertForCTC as HubertForCTC, __webpack_exports__HubertForSequenceClassification as HubertForSequenceClassification, __webpack_exports__HubertModel as HubertModel, __webpack_exports__HubertPreTrainedModel as HubertPreTrainedModel, __webpack_exports__IJepaForImageClassification as IJepaForImageClassification, __webpack_exports__IJepaModel as IJepaModel, __webpack_exports__IJepaPreTrainedModel as IJepaPreTrainedModel, __webpack_exports__Idefics3ForConditionalGeneration as Idefics3ForConditionalGeneration, __webpack_exports__Idefics3ImageProcessor as Idefics3ImageProcessor, __webpack_exports__Idefics3PreTrainedModel as Idefics3PreTrainedModel, __webpack_exports__Idefics3Processor as Idefics3Processor, __webpack_exports__ImageClassificationPipeline as ImageClassificationPipeline, __webpack_exports__ImageFeatureExtractionPipeline as ImageFeatureExtractionPipeline, __webpack_exports__ImageFeatureExtractor as ImageFeatureExtractor, __webpack_exports__ImageMattingOutput as ImageMattingOutput, __webpack_exports__ImageProcessor as ImageProcessor, __webpack_exports__ImageSegmentationPipeline as ImageSegmentationPipeline, __webpack_exports__ImageToImagePipeline as ImageToImagePipeline, __webpack_exports__ImageToTextPipeline as ImageToTextPipeline, __webpack_exports__InterruptableStoppingCriteria as InterruptableStoppingCriteria, __webpack_exports__JAISLMHeadModel as JAISLMHeadModel, __webpack_exports__JAISModel as JAISModel, __webpack_exports__JAISPreTrainedModel as JAISPreTrainedModel, __webpack_exports__JinaCLIPImageProcessor as JinaCLIPImageProcessor, __webpack_exports__JinaCLIPModel as JinaCLIPModel, __webpack_exports__JinaCLIPPreTrainedModel as JinaCLIPPreTrainedModel, __webpack_exports__JinaCLIPProcessor as JinaCLIPProcessor, __webpack_exports__JinaCLIPTextModel as JinaCLIPTextModel, __webpack_exports__JinaCLIPVisionModel as JinaCLIPVisionModel, __webpack_exports__LiteWhisperForConditionalGeneration as LiteWhisperForConditionalGeneration, __webpack_exports__LlamaForCausalLM as LlamaForCausalLM, __webpack_exports__LlamaModel as LlamaModel, __webpack_exports__LlamaPreTrainedModel as LlamaPreTrainedModel, __webpack_exports__LlamaTokenizer as LlamaTokenizer, __webpack_exports__LlavaForConditionalGeneration as LlavaForConditionalGeneration, __webpack_exports__LlavaOnevisionForConditionalGeneration as LlavaOnevisionForConditionalGeneration, __webpack_exports__LlavaOnevisionImageProcessor as LlavaOnevisionImageProcessor, __webpack_exports__LlavaPreTrainedModel as LlavaPreTrainedModel, __webpack_exports__LogitsProcessor as LogitsProcessor, __webpack_exports__LogitsProcessorList as LogitsProcessorList, __webpack_exports__LogitsWarper as LogitsWarper, __webpack_exports__LongT5ForConditionalGeneration as LongT5ForConditionalGeneration, __webpack_exports__LongT5Model as LongT5Model, __webpack_exports__LongT5PreTrainedModel as LongT5PreTrainedModel, __webpack_exports__M2M100ForConditionalGeneration as M2M100ForConditionalGeneration, __webpack_exports__M2M100Model as M2M100Model, __webpack_exports__M2M100PreTrainedModel as M2M100PreTrainedModel, __webpack_exports__M2M100Tokenizer as M2M100Tokenizer, __webpack_exports__MBart50Tokenizer as MBart50Tokenizer, __webpack_exports__MBartForCausalLM as MBartForCausalLM, __webpack_exports__MBartForConditionalGeneration as MBartForConditionalGeneration, __webpack_exports__MBartForSequenceClassification as MBartForSequenceClassification, __webpack_exports__MBartModel as MBartModel, __webpack_exports__MBartPreTrainedModel as MBartPreTrainedModel, __webpack_exports__MBartTokenizer as MBartTokenizer, __webpack_exports__MPNetForMaskedLM as MPNetForMaskedLM, __webpack_exports__MPNetForQuestionAnswering as MPNetForQuestionAnswering, __webpack_exports__MPNetForSequenceClassification as MPNetForSequenceClassification, __webpack_exports__MPNetForTokenClassification as MPNetForTokenClassification, __webpack_exports__MPNetModel as MPNetModel, __webpack_exports__MPNetPreTrainedModel as MPNetPreTrainedModel, __webpack_exports__MPNetTokenizer as MPNetTokenizer, __webpack_exports__MT5ForConditionalGeneration as MT5ForConditionalGeneration, __webpack_exports__MT5Model as MT5Model, __webpack_exports__MT5PreTrainedModel as MT5PreTrainedModel, __webpack_exports__MarianMTModel as MarianMTModel, __webpack_exports__MarianModel as MarianModel, __webpack_exports__MarianPreTrainedModel as MarianPreTrainedModel, __webpack_exports__MarianTokenizer as MarianTokenizer, __webpack_exports__Mask2FormerImageProcessor as Mask2FormerImageProcessor, __webpack_exports__MaskFormerFeatureExtractor as MaskFormerFeatureExtractor, __webpack_exports__MaskFormerForInstanceSegmentation as MaskFormerForInstanceSegmentation, __webpack_exports__MaskFormerImageProcessor as MaskFormerImageProcessor, __webpack_exports__MaskFormerModel as MaskFormerModel, __webpack_exports__MaskFormerPreTrainedModel as MaskFormerPreTrainedModel, __webpack_exports__MaskedLMOutput as MaskedLMOutput, __webpack_exports__MaxLengthCriteria as MaxLengthCriteria, __webpack_exports__Metric3DForDepthEstimation as Metric3DForDepthEstimation, __webpack_exports__Metric3DPreTrainedModel as Metric3DPreTrainedModel, __webpack_exports__Metric3Dv2ForDepthEstimation as Metric3Dv2ForDepthEstimation, __webpack_exports__Metric3Dv2PreTrainedModel as Metric3Dv2PreTrainedModel, __webpack_exports__MgpstrForSceneTextRecognition as MgpstrForSceneTextRecognition, __webpack_exports__MgpstrModelOutput as MgpstrModelOutput, __webpack_exports__MgpstrPreTrainedModel as MgpstrPreTrainedModel, __webpack_exports__MgpstrProcessor as MgpstrProcessor, __webpack_exports__MgpstrTokenizer as MgpstrTokenizer, __webpack_exports__MimiDecoderModel as MimiDecoderModel, __webpack_exports__MimiDecoderOutput as MimiDecoderOutput, __webpack_exports__MimiEncoderModel as MimiEncoderModel, __webpack_exports__MimiEncoderOutput as MimiEncoderOutput, __webpack_exports__MimiModel as MimiModel, __webpack_exports__MimiPreTrainedModel as MimiPreTrainedModel, __webpack_exports__MinLengthLogitsProcessor as MinLengthLogitsProcessor, __webpack_exports__MinNewTokensLengthLogitsProcessor as MinNewTokensLengthLogitsProcessor, __webpack_exports__MistralForCausalLM as MistralForCausalLM, __webpack_exports__MistralModel as MistralModel, __webpack_exports__MistralPreTrainedModel as MistralPreTrainedModel, __webpack_exports__MobileBertForMaskedLM as MobileBertForMaskedLM, __webpack_exports__MobileBertForQuestionAnswering as MobileBertForQuestionAnswering, __webpack_exports__MobileBertForSequenceClassification as MobileBertForSequenceClassification, __webpack_exports__MobileBertModel as MobileBertModel, __webpack_exports__MobileBertPreTrainedModel as MobileBertPreTrainedModel, __webpack_exports__MobileBertTokenizer as MobileBertTokenizer, __webpack_exports__MobileLLMForCausalLM as MobileLLMForCausalLM, __webpack_exports__MobileLLMModel as MobileLLMModel, __webpack_exports__MobileLLMPreTrainedModel as MobileLLMPreTrainedModel, __webpack_exports__MobileNetV1FeatureExtractor as MobileNetV1FeatureExtractor, __webpack_exports__MobileNetV1ForImageClassification as MobileNetV1ForImageClassification, __webpack_exports__MobileNetV1ForSemanticSegmentation as MobileNetV1ForSemanticSegmentation, __webpack_exports__MobileNetV1ImageProcessor as MobileNetV1ImageProcessor, __webpack_exports__MobileNetV1Model as MobileNetV1Model, __webpack_exports__MobileNetV1PreTrainedModel as MobileNetV1PreTrainedModel, __webpack_exports__MobileNetV2FeatureExtractor as MobileNetV2FeatureExtractor, __webpack_exports__MobileNetV2ForImageClassification as MobileNetV2ForImageClassification, __webpack_exports__MobileNetV2ForSemanticSegmentation as MobileNetV2ForSemanticSegmentation, __webpack_exports__MobileNetV2ImageProcessor as MobileNetV2ImageProcessor, __webpack_exports__MobileNetV2Model as MobileNetV2Model, __webpack_exports__MobileNetV2PreTrainedModel as MobileNetV2PreTrainedModel, __webpack_exports__MobileNetV3FeatureExtractor as MobileNetV3FeatureExtractor, __webpack_exports__MobileNetV3ForImageClassification as MobileNetV3ForImageClassification, __webpack_exports__MobileNetV3ForSemanticSegmentation as MobileNetV3ForSemanticSegmentation, __webpack_exports__MobileNetV3ImageProcessor as MobileNetV3ImageProcessor, __webpack_exports__MobileNetV3Model as MobileNetV3Model, __webpack_exports__MobileNetV3PreTrainedModel as MobileNetV3PreTrainedModel, __webpack_exports__MobileNetV4FeatureExtractor as MobileNetV4FeatureExtractor, __webpack_exports__MobileNetV4ForImageClassification as MobileNetV4ForImageClassification, __webpack_exports__MobileNetV4ForSemanticSegmentation as MobileNetV4ForSemanticSegmentation, __webpack_exports__MobileNetV4ImageProcessor as MobileNetV4ImageProcessor, __webpack_exports__MobileNetV4Model as MobileNetV4Model, __webpack_exports__MobileNetV4PreTrainedModel as MobileNetV4PreTrainedModel, __webpack_exports__MobileViTFeatureExtractor as MobileViTFeatureExtractor, __webpack_exports__MobileViTForImageClassification as MobileViTForImageClassification, __webpack_exports__MobileViTImageProcessor as MobileViTImageProcessor, __webpack_exports__MobileViTModel as MobileViTModel, __webpack_exports__MobileViTPreTrainedModel as MobileViTPreTrainedModel, __webpack_exports__MobileViTV2ForImageClassification as MobileViTV2ForImageClassification, __webpack_exports__MobileViTV2Model as MobileViTV2Model, __webpack_exports__MobileViTV2PreTrainedModel as MobileViTV2PreTrainedModel, __webpack_exports__ModelOutput as ModelOutput, __webpack_exports__ModernBertForMaskedLM as ModernBertForMaskedLM, __webpack_exports__ModernBertForSequenceClassification as ModernBertForSequenceClassification, __webpack_exports__ModernBertForTokenClassification as ModernBertForTokenClassification, __webpack_exports__ModernBertModel as ModernBertModel, __webpack_exports__ModernBertPreTrainedModel as ModernBertPreTrainedModel, __webpack_exports__Moondream1ForConditionalGeneration as Moondream1ForConditionalGeneration, __webpack_exports__MoonshineFeatureExtractor as MoonshineFeatureExtractor, __webpack_exports__MoonshineForConditionalGeneration as MoonshineForConditionalGeneration, __webpack_exports__MoonshineModel as MoonshineModel, __webpack_exports__MoonshinePreTrainedModel as MoonshinePreTrainedModel, __webpack_exports__MoonshineProcessor as MoonshineProcessor, __webpack_exports__MptForCausalLM as MptForCausalLM, __webpack_exports__MptModel as MptModel, __webpack_exports__MptPreTrainedModel as MptPreTrainedModel, __webpack_exports__MultiModalityCausalLM as MultiModalityCausalLM, __webpack_exports__MultiModalityPreTrainedModel as MultiModalityPreTrainedModel, __webpack_exports__MusicgenForCausalLM as MusicgenForCausalLM, __webpack_exports__MusicgenForConditionalGeneration as MusicgenForConditionalGeneration, __webpack_exports__MusicgenModel as MusicgenModel, __webpack_exports__MusicgenPreTrainedModel as MusicgenPreTrainedModel, __webpack_exports__NllbTokenizer as NllbTokenizer, __webpack_exports__NoBadWordsLogitsProcessor as NoBadWordsLogitsProcessor, __webpack_exports__NoRepeatNGramLogitsProcessor as NoRepeatNGramLogitsProcessor, __webpack_exports__NomicBertModel as NomicBertModel, __webpack_exports__NomicBertPreTrainedModel as NomicBertPreTrainedModel, __webpack_exports__NougatImageProcessor as NougatImageProcessor, __webpack_exports__NougatTokenizer as NougatTokenizer, __webpack_exports__OPTForCausalLM as OPTForCausalLM, __webpack_exports__OPTModel as OPTModel, __webpack_exports__OPTPreTrainedModel as OPTPreTrainedModel, __webpack_exports__ObjectDetectionPipeline as ObjectDetectionPipeline, __webpack_exports__Olmo2ForCausalLM as Olmo2ForCausalLM, __webpack_exports__Olmo2Model as Olmo2Model, __webpack_exports__Olmo2PreTrainedModel as Olmo2PreTrainedModel, __webpack_exports__OlmoForCausalLM as OlmoForCausalLM, __webpack_exports__OlmoModel as OlmoModel, __webpack_exports__OlmoPreTrainedModel as OlmoPreTrainedModel, __webpack_exports__OpenELMForCausalLM as OpenELMForCausalLM, __webpack_exports__OpenELMModel as OpenELMModel, __webpack_exports__OpenELMPreTrainedModel as OpenELMPreTrainedModel, __webpack_exports__OwlViTFeatureExtractor as OwlViTFeatureExtractor, __webpack_exports__OwlViTForObjectDetection as OwlViTForObjectDetection, __webpack_exports__OwlViTImageProcessor as OwlViTImageProcessor, __webpack_exports__OwlViTModel as OwlViTModel, __webpack_exports__OwlViTPreTrainedModel as OwlViTPreTrainedModel, __webpack_exports__OwlViTProcessor as OwlViTProcessor, __webpack_exports__Owlv2ForObjectDetection as Owlv2ForObjectDetection, __webpack_exports__Owlv2ImageProcessor as Owlv2ImageProcessor, __webpack_exports__Owlv2Model as Owlv2Model, __webpack_exports__Owlv2PreTrainedModel as Owlv2PreTrainedModel, __webpack_exports__PaliGemmaForConditionalGeneration as PaliGemmaForConditionalGeneration, __webpack_exports__PaliGemmaPreTrainedModel as PaliGemmaPreTrainedModel, __webpack_exports__PaliGemmaProcessor as PaliGemmaProcessor, __webpack_exports__PatchTSMixerForPrediction as PatchTSMixerForPrediction, __webpack_exports__PatchTSMixerModel as PatchTSMixerModel, __webpack_exports__PatchTSMixerPreTrainedModel as PatchTSMixerPreTrainedModel, __webpack_exports__PatchTSTForPrediction as PatchTSTForPrediction, __webpack_exports__PatchTSTModel as PatchTSTModel, __webpack_exports__PatchTSTPreTrainedModel as PatchTSTPreTrainedModel, __webpack_exports__Phi3ForCausalLM as Phi3ForCausalLM, __webpack_exports__Phi3Model as Phi3Model, __webpack_exports__Phi3PreTrainedModel as Phi3PreTrainedModel, __webpack_exports__Phi3VForCausalLM as Phi3VForCausalLM, __webpack_exports__Phi3VImageProcessor as Phi3VImageProcessor, __webpack_exports__Phi3VPreTrainedModel as Phi3VPreTrainedModel, __webpack_exports__Phi3VProcessor as Phi3VProcessor, __webpack_exports__PhiForCausalLM as PhiForCausalLM, __webpack_exports__PhiModel as PhiModel, __webpack_exports__PhiPreTrainedModel as PhiPreTrainedModel, __webpack_exports__Pipeline as Pipeline, __webpack_exports__PreTrainedModel as PreTrainedModel, __webpack_exports__PreTrainedTokenizer as PreTrainedTokenizer, __webpack_exports__PretrainedConfig as PretrainedConfig, __webpack_exports__PretrainedMixin as PretrainedMixin, __webpack_exports__Processor as Processor, __webpack_exports__PvtForImageClassification as PvtForImageClassification, __webpack_exports__PvtImageProcessor as PvtImageProcessor, __webpack_exports__PvtModel as PvtModel, __webpack_exports__PvtPreTrainedModel as PvtPreTrainedModel, __webpack_exports__PyAnnoteFeatureExtractor as PyAnnoteFeatureExtractor, __webpack_exports__PyAnnoteForAudioFrameClassification as PyAnnoteForAudioFrameClassification, __webpack_exports__PyAnnoteModel as PyAnnoteModel, __webpack_exports__PyAnnotePreTrainedModel as PyAnnotePreTrainedModel, __webpack_exports__PyAnnoteProcessor as PyAnnoteProcessor, __webpack_exports__QuestionAnsweringModelOutput as QuestionAnsweringModelOutput, __webpack_exports__QuestionAnsweringPipeline as QuestionAnsweringPipeline, __webpack_exports__Qwen2ForCausalLM as Qwen2ForCausalLM, __webpack_exports__Qwen2Model as Qwen2Model, __webpack_exports__Qwen2PreTrainedModel as Qwen2PreTrainedModel, __webpack_exports__Qwen2Tokenizer as Qwen2Tokenizer, __webpack_exports__Qwen2VLForConditionalGeneration as Qwen2VLForConditionalGeneration, __webpack_exports__Qwen2VLImageProcessor as Qwen2VLImageProcessor, __webpack_exports__Qwen2VLPreTrainedModel as Qwen2VLPreTrainedModel, __webpack_exports__Qwen2VLProcessor as Qwen2VLProcessor, __webpack_exports__RTDetrForObjectDetection as RTDetrForObjectDetection, __webpack_exports__RTDetrImageProcessor as RTDetrImageProcessor, __webpack_exports__RTDetrModel as RTDetrModel, __webpack_exports__RTDetrObjectDetectionOutput as RTDetrObjectDetectionOutput, __webpack_exports__RTDetrPreTrainedModel as RTDetrPreTrainedModel, __webpack_exports__RawAudio as RawAudio, __webpack_exports__RawImage as RawImage, __webpack_exports__RawVideo as RawVideo, __webpack_exports__RawVideoFrame as RawVideoFrame, __webpack_exports__RepetitionPenaltyLogitsProcessor as RepetitionPenaltyLogitsProcessor, __webpack_exports__ResNetForImageClassification as ResNetForImageClassification, __webpack_exports__ResNetModel as ResNetModel, __webpack_exports__ResNetPreTrainedModel as ResNetPreTrainedModel, __webpack_exports__RoFormerForMaskedLM as RoFormerForMaskedLM, __webpack_exports__RoFormerForQuestionAnswering as RoFormerForQuestionAnswering, __webpack_exports__RoFormerForSequenceClassification as RoFormerForSequenceClassification, __webpack_exports__RoFormerForTokenClassification as RoFormerForTokenClassification, __webpack_exports__RoFormerModel as RoFormerModel, __webpack_exports__RoFormerPreTrainedModel as RoFormerPreTrainedModel, __webpack_exports__RoFormerTokenizer as RoFormerTokenizer, __webpack_exports__RobertaForMaskedLM as RobertaForMaskedLM, __webpack_exports__RobertaForQuestionAnswering as RobertaForQuestionAnswering, __webpack_exports__RobertaForSequenceClassification as RobertaForSequenceClassification, __webpack_exports__RobertaForTokenClassification as RobertaForTokenClassification, __webpack_exports__RobertaModel as RobertaModel, __webpack_exports__RobertaPreTrainedModel as RobertaPreTrainedModel, __webpack_exports__RobertaTokenizer as RobertaTokenizer, __webpack_exports__SamImageProcessor as SamImageProcessor, __webpack_exports__SamImageSegmentationOutput as SamImageSegmentationOutput, __webpack_exports__SamModel as SamModel, __webpack_exports__SamPreTrainedModel as SamPreTrainedModel, __webpack_exports__SamProcessor as SamProcessor, __webpack_exports__SapiensForDepthEstimation as SapiensForDepthEstimation, __webpack_exports__SapiensForNormalEstimation as SapiensForNormalEstimation, __webpack_exports__SapiensForSemanticSegmentation as SapiensForSemanticSegmentation, __webpack_exports__SapiensPreTrainedModel as SapiensPreTrainedModel, __webpack_exports__SeamlessM4TFeatureExtractor as SeamlessM4TFeatureExtractor, __webpack_exports__SegformerFeatureExtractor as SegformerFeatureExtractor, __webpack_exports__SegformerForImageClassification as SegformerForImageClassification, __webpack_exports__SegformerForSemanticSegmentation as SegformerForSemanticSegmentation, __webpack_exports__SegformerImageProcessor as SegformerImageProcessor, __webpack_exports__SegformerModel as SegformerModel, __webpack_exports__SegformerPreTrainedModel as SegformerPreTrainedModel, __webpack_exports__Seq2SeqLMOutput as Seq2SeqLMOutput, __webpack_exports__SequenceClassifierOutput as SequenceClassifierOutput, __webpack_exports__SiglipImageProcessor as SiglipImageProcessor, __webpack_exports__SiglipModel as SiglipModel, __webpack_exports__SiglipPreTrainedModel as SiglipPreTrainedModel, __webpack_exports__SiglipTextModel as SiglipTextModel, __webpack_exports__SiglipTokenizer as SiglipTokenizer, __webpack_exports__SiglipVisionModel as SiglipVisionModel, __webpack_exports__SmolVLMForConditionalGeneration as SmolVLMForConditionalGeneration, __webpack_exports__SmolVLMImageProcessor as SmolVLMImageProcessor, __webpack_exports__SmolVLMProcessor as SmolVLMProcessor, __webpack_exports__SnacDecoderModel as SnacDecoderModel, __webpack_exports__SnacEncoderModel as SnacEncoderModel, __webpack_exports__SnacFeatureExtractor as SnacFeatureExtractor, __webpack_exports__SnacModel as SnacModel, __webpack_exports__SnacPreTrainedModel as SnacPreTrainedModel, __webpack_exports__SpeechT5FeatureExtractor as SpeechT5FeatureExtractor, __webpack_exports__SpeechT5ForSpeechToText as SpeechT5ForSpeechToText, __webpack_exports__SpeechT5ForTextToSpeech as SpeechT5ForTextToSpeech, __webpack_exports__SpeechT5HifiGan as SpeechT5HifiGan, __webpack_exports__SpeechT5Model as SpeechT5Model, __webpack_exports__SpeechT5PreTrainedModel as SpeechT5PreTrainedModel, __webpack_exports__SpeechT5Processor as SpeechT5Processor, __webpack_exports__SpeechT5Tokenizer as SpeechT5Tokenizer, __webpack_exports__SqueezeBertForMaskedLM as SqueezeBertForMaskedLM, __webpack_exports__SqueezeBertForQuestionAnswering as SqueezeBertForQuestionAnswering, __webpack_exports__SqueezeBertForSequenceClassification as SqueezeBertForSequenceClassification, __webpack_exports__SqueezeBertModel as SqueezeBertModel, __webpack_exports__SqueezeBertPreTrainedModel as SqueezeBertPreTrainedModel, __webpack_exports__SqueezeBertTokenizer as SqueezeBertTokenizer, __webpack_exports__StableLmForCausalLM as StableLmForCausalLM, __webpack_exports__StableLmModel as StableLmModel, __webpack_exports__StableLmPreTrainedModel as StableLmPreTrainedModel, __webpack_exports__Starcoder2ForCausalLM as Starcoder2ForCausalLM, __webpack_exports__Starcoder2Model as Starcoder2Model, __webpack_exports__Starcoder2PreTrainedModel as Starcoder2PreTrainedModel, __webpack_exports__StoppingCriteria as StoppingCriteria, __webpack_exports__StoppingCriteriaList as StoppingCriteriaList, __webpack_exports__StyleTextToSpeech2Model as StyleTextToSpeech2Model, __webpack_exports__StyleTextToSpeech2PreTrainedModel as StyleTextToSpeech2PreTrainedModel, __webpack_exports__SummarizationPipeline as SummarizationPipeline, __webpack_exports__SuppressTokensAtBeginLogitsProcessor as SuppressTokensAtBeginLogitsProcessor, __webpack_exports__Swin2SRForImageSuperResolution as Swin2SRForImageSuperResolution, __webpack_exports__Swin2SRImageProcessor as Swin2SRImageProcessor, __webpack_exports__Swin2SRModel as Swin2SRModel, __webpack_exports__Swin2SRPreTrainedModel as Swin2SRPreTrainedModel, __webpack_exports__SwinForImageClassification as SwinForImageClassification, __webpack_exports__SwinForSemanticSegmentation as SwinForSemanticSegmentation, __webpack_exports__SwinModel as SwinModel, __webpack_exports__SwinPreTrainedModel as SwinPreTrainedModel, __webpack_exports__T5ForConditionalGeneration as T5ForConditionalGeneration, __webpack_exports__T5Model as T5Model, __webpack_exports__T5PreTrainedModel as T5PreTrainedModel, __webpack_exports__T5Tokenizer as T5Tokenizer, __webpack_exports__TableTransformerForObjectDetection as TableTransformerForObjectDetection, __webpack_exports__TableTransformerModel as TableTransformerModel, __webpack_exports__TableTransformerObjectDetectionOutput as TableTransformerObjectDetectionOutput, __webpack_exports__TableTransformerPreTrainedModel as TableTransformerPreTrainedModel, __webpack_exports__TemperatureLogitsWarper as TemperatureLogitsWarper, __webpack_exports__Tensor as Tensor, __webpack_exports__Text2TextGenerationPipeline as Text2TextGenerationPipeline, __webpack_exports__TextClassificationPipeline as TextClassificationPipeline, __webpack_exports__TextGenerationPipeline as TextGenerationPipeline, __webpack_exports__TextStreamer as TextStreamer, __webpack_exports__TextToAudioPipeline as TextToAudioPipeline, __webpack_exports__TokenClassificationPipeline as TokenClassificationPipeline, __webpack_exports__TokenClassifierOutput as TokenClassifierOutput, __webpack_exports__TokenizerModel as TokenizerModel, __webpack_exports__TopKLogitsWarper as TopKLogitsWarper, __webpack_exports__TopPLogitsWarper as TopPLogitsWarper, __webpack_exports__TrOCRForCausalLM as TrOCRForCausalLM, __webpack_exports__TrOCRPreTrainedModel as TrOCRPreTrainedModel, __webpack_exports__TranslationPipeline as TranslationPipeline, __webpack_exports__UltravoxModel as UltravoxModel, __webpack_exports__UltravoxPreTrainedModel as UltravoxPreTrainedModel, __webpack_exports__UltravoxProcessor as UltravoxProcessor, __webpack_exports__UniSpeechForCTC as UniSpeechForCTC, __webpack_exports__UniSpeechForSequenceClassification as UniSpeechForSequenceClassification, __webpack_exports__UniSpeechModel as UniSpeechModel, __webpack_exports__UniSpeechPreTrainedModel as UniSpeechPreTrainedModel, __webpack_exports__UniSpeechSatForAudioFrameClassification as UniSpeechSatForAudioFrameClassification, __webpack_exports__UniSpeechSatForCTC as UniSpeechSatForCTC, __webpack_exports__UniSpeechSatForSequenceClassification as UniSpeechSatForSequenceClassification, __webpack_exports__UniSpeechSatModel as UniSpeechSatModel, __webpack_exports__UniSpeechSatPreTrainedModel as UniSpeechSatPreTrainedModel, __webpack_exports__VLChatProcessor as VLChatProcessor, __webpack_exports__VLMImageProcessor as VLMImageProcessor, __webpack_exports__ViTFeatureExtractor as ViTFeatureExtractor, __webpack_exports__ViTForImageClassification as ViTForImageClassification, __webpack_exports__ViTImageProcessor as ViTImageProcessor, __webpack_exports__ViTMAEModel as ViTMAEModel, __webpack_exports__ViTMAEPreTrainedModel as ViTMAEPreTrainedModel, __webpack_exports__ViTMSNForImageClassification as ViTMSNForImageClassification, __webpack_exports__ViTMSNModel as ViTMSNModel, __webpack_exports__ViTMSNPreTrainedModel as ViTMSNPreTrainedModel, __webpack_exports__ViTModel as ViTModel, __webpack_exports__ViTPreTrainedModel as ViTPreTrainedModel, __webpack_exports__VisionEncoderDecoderModel as VisionEncoderDecoderModel, __webpack_exports__VitMatteForImageMatting as VitMatteForImageMatting, __webpack_exports__VitMatteImageProcessor as VitMatteImageProcessor, __webpack_exports__VitMattePreTrainedModel as VitMattePreTrainedModel, __webpack_exports__VitPoseForPoseEstimation as VitPoseForPoseEstimation, __webpack_exports__VitPoseImageProcessor as VitPoseImageProcessor, __webpack_exports__VitPosePreTrainedModel as VitPosePreTrainedModel, __webpack_exports__VitsModel as VitsModel, __webpack_exports__VitsModelOutput as VitsModelOutput, __webpack_exports__VitsPreTrainedModel as VitsPreTrainedModel, __webpack_exports__VitsTokenizer as VitsTokenizer, __webpack_exports__Wav2Vec2BertForCTC as Wav2Vec2BertForCTC, __webpack_exports__Wav2Vec2BertForSequenceClassification as Wav2Vec2BertForSequenceClassification, __webpack_exports__Wav2Vec2BertModel as Wav2Vec2BertModel, __webpack_exports__Wav2Vec2BertPreTrainedModel as Wav2Vec2BertPreTrainedModel, __webpack_exports__Wav2Vec2CTCTokenizer as Wav2Vec2CTCTokenizer, __webpack_exports__Wav2Vec2FeatureExtractor as Wav2Vec2FeatureExtractor, __webpack_exports__Wav2Vec2ForAudioFrameClassification as Wav2Vec2ForAudioFrameClassification, __webpack_exports__Wav2Vec2ForCTC as Wav2Vec2ForCTC, __webpack_exports__Wav2Vec2ForSequenceClassification as Wav2Vec2ForSequenceClassification, __webpack_exports__Wav2Vec2Model as Wav2Vec2Model, __webpack_exports__Wav2Vec2PreTrainedModel as Wav2Vec2PreTrainedModel, __webpack_exports__Wav2Vec2Processor as Wav2Vec2Processor, __webpack_exports__Wav2Vec2ProcessorWithLM as Wav2Vec2ProcessorWithLM, __webpack_exports__WavLMForAudioFrameClassification as WavLMForAudioFrameClassification, __webpack_exports__WavLMForCTC as WavLMForCTC, __webpack_exports__WavLMForSequenceClassification as WavLMForSequenceClassification, __webpack_exports__WavLMForXVector as WavLMForXVector, __webpack_exports__WavLMModel as WavLMModel, __webpack_exports__WavLMPreTrainedModel as WavLMPreTrainedModel, __webpack_exports__WeSpeakerFeatureExtractor as WeSpeakerFeatureExtractor, __webpack_exports__WeSpeakerResNetModel as WeSpeakerResNetModel, __webpack_exports__WeSpeakerResNetPreTrainedModel as WeSpeakerResNetPreTrainedModel, __webpack_exports__WhisperFeatureExtractor as WhisperFeatureExtractor, __webpack_exports__WhisperForConditionalGeneration as WhisperForConditionalGeneration, __webpack_exports__WhisperModel as WhisperModel, __webpack_exports__WhisperPreTrainedModel as WhisperPreTrainedModel, __webpack_exports__WhisperProcessor as WhisperProcessor, __webpack_exports__WhisperTextStreamer as WhisperTextStreamer, __webpack_exports__WhisperTimeStampLogitsProcessor as WhisperTimeStampLogitsProcessor, __webpack_exports__WhisperTokenizer as WhisperTokenizer, __webpack_exports__XLMForQuestionAnswering as XLMForQuestionAnswering, __webpack_exports__XLMForSequenceClassification as XLMForSequenceClassification, __webpack_exports__XLMForTokenClassification as XLMForTokenClassification, __webpack_exports__XLMModel as XLMModel, __webpack_exports__XLMPreTrainedModel as XLMPreTrainedModel, __webpack_exports__XLMRobertaForMaskedLM as XLMRobertaForMaskedLM, __webpack_exports__XLMRobertaForQuestionAnswering as XLMRobertaForQuestionAnswering, __webpack_exports__XLMRobertaForSequenceClassification as XLMRobertaForSequenceClassification, __webpack_exports__XLMRobertaForTokenClassification as XLMRobertaForTokenClassification, __webpack_exports__XLMRobertaModel as XLMRobertaModel, __webpack_exports__XLMRobertaPreTrainedModel as XLMRobertaPreTrainedModel, __webpack_exports__XLMRobertaTokenizer as XLMRobertaTokenizer, __webpack_exports__XLMTokenizer as XLMTokenizer, __webpack_exports__XLMWithLMHeadModel as XLMWithLMHeadModel, __webpack_exports__XVectorOutput as XVectorOutput, __webpack_exports__YolosFeatureExtractor as YolosFeatureExtractor, __webpack_exports__YolosForObjectDetection as YolosForObjectDetection, __webpack_exports__YolosImageProcessor as YolosImageProcessor, __webpack_exports__YolosModel as YolosModel, __webpack_exports__YolosObjectDetectionOutput as YolosObjectDetectionOutput, __webpack_exports__YolosPreTrainedModel as YolosPreTrainedModel, __webpack_exports__ZeroShotAudioClassificationPipeline as ZeroShotAudioClassificationPipeline, __webpack_exports__ZeroShotClassificationPipeline as ZeroShotClassificationPipeline, __webpack_exports__ZeroShotImageClassificationPipeline as ZeroShotImageClassificationPipeline, __webpack_exports__ZeroShotObjectDetectionPipeline as ZeroShotObjectDetectionPipeline, __webpack_exports__bankers_round as bankers_round, __webpack_exports__cat as cat, __webpack_exports__cos_sim as cos_sim, __webpack_exports__dot as dot, __webpack_exports__dynamic_time_warping as dynamic_time_warping, __webpack_exports__env as env, __webpack_exports__full as full, __webpack_exports__full_like as full_like, __webpack_exports__getKeyValueShapes as getKeyValueShapes, __webpack_exports__hamming as hamming, __webpack_exports__hanning as hanning, __webpack_exports__interpolate as interpolate, __webpack_exports__interpolate_4d as interpolate_4d, __webpack_exports__interpolate_data as interpolate_data, __webpack_exports__is_chinese_char as is_chinese_char, __webpack_exports__layer_norm as layer_norm, __webpack_exports__load_image as load_image, __webpack_exports__load_video as load_video, __webpack_exports__log_softmax as log_softmax, __webpack_exports__magnitude as magnitude, __webpack_exports__matmul as matmul, __webpack_exports__max as max, __webpack_exports__mean as mean, __webpack_exports__mean_pooling as mean_pooling, __webpack_exports__medianFilter as medianFilter, __webpack_exports__mel_filter_bank as mel_filter_bank, __webpack_exports__min as min, __webpack_exports__ones as ones, __webpack_exports__ones_like as ones_like, __webpack_exports__permute as permute, __webpack_exports__permute_data as permute_data, __webpack_exports__pipeline as pipeline, __webpack_exports__quantize_embeddings as quantize_embeddings, __webpack_exports__rand as rand, __webpack_exports__read_audio as read_audio, __webpack_exports__rfft as rfft, __webpack_exports__round as round, __webpack_exports__slice as slice, __webpack_exports__softmax as softmax, __webpack_exports__spectrogram as spectrogram, __webpack_exports__stack as stack, __webpack_exports__std_mean as std_mean, __webpack_exports__topk as topk, __webpack_exports__window_function as window_function, __webpack_exports__zeros as zeros, __webpack_exports__zeros_like as zeros_like };
40532
40695
 
40533
40696
  //# sourceMappingURL=transformers.js.map