@huggingface/transformers 3.0.0-alpha.13 → 3.0.0-alpha.15

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -3756,7 +3756,7 @@ __webpack_require__.r(__webpack_exports__);
3756
3756
  // Licensed under the MIT License.
3757
3757
  // This file is generated by /js/scripts/update-version.ts
3758
3758
  // Do not modify file content manually.
3759
- const version = '1.19.0';
3759
+ const version = '1.19.2';
3760
3760
  //# sourceMappingURL=version.js.map
3761
3761
 
3762
3762
  /***/ }),
@@ -4410,7 +4410,7 @@ __webpack_require__.r(__webpack_exports__);
4410
4410
 
4411
4411
 
4412
4412
 
4413
- const VERSION = '3.0.0-alpha.13';
4413
+ const VERSION = '3.0.0-alpha.15';
4414
4414
 
4415
4415
  // Check if various APIs are available (depends on environment)
4416
4416
  const IS_BROWSER_ENV = typeof self !== 'undefined';
@@ -4457,19 +4457,19 @@ const apis = Object.freeze({
4457
4457
  });
4458
4458
 
4459
4459
  const RUNNING_LOCALLY = IS_FS_AVAILABLE && IS_PATH_AVAILABLE;
4460
- const __dirname = RUNNING_LOCALLY
4460
+ const dirname__ = RUNNING_LOCALLY
4461
4461
  ? path__WEBPACK_IMPORTED_MODULE_1__["default"].dirname(path__WEBPACK_IMPORTED_MODULE_1__["default"].dirname(url__WEBPACK_IMPORTED_MODULE_2__["default"].fileURLToPath(import.meta.url)))
4462
4462
  : './';
4463
4463
 
4464
4464
  // Only used for environments with access to file system
4465
4465
  const DEFAULT_CACHE_DIR = RUNNING_LOCALLY
4466
- ? path__WEBPACK_IMPORTED_MODULE_1__["default"].join(__dirname, '/.cache/')
4466
+ ? path__WEBPACK_IMPORTED_MODULE_1__["default"].join(dirname__, '/.cache/')
4467
4467
  : null;
4468
4468
 
4469
4469
  // Set local model path, based on available APIs
4470
4470
  const DEFAULT_LOCAL_MODEL_PATH = '/models/';
4471
4471
  const localModelPath = RUNNING_LOCALLY
4472
- ? path__WEBPACK_IMPORTED_MODULE_1__["default"].join(__dirname, DEFAULT_LOCAL_MODEL_PATH)
4472
+ ? path__WEBPACK_IMPORTED_MODULE_1__["default"].join(dirname__, DEFAULT_LOCAL_MODEL_PATH)
4473
4473
  : DEFAULT_LOCAL_MODEL_PATH;
4474
4474
 
4475
4475
  /**
@@ -5504,18 +5504,18 @@ class NoBadWordsLogitsProcessor extends LogitsProcessor {
5504
5504
  _call(input_ids, logits) {
5505
5505
  for (let i = 0; i < input_ids.length; ++i) {
5506
5506
  const batch_logits_data = /** @type {Float32Array} */(logits[i].data);
5507
-
5507
+ const ids = input_ids[i];
5508
5508
  for (const bad_word_ids of this.bad_words_ids) {
5509
5509
  // Whether to modify the logits of the last token in the bad word id sequence
5510
5510
  let mark = true;
5511
5511
 
5512
5512
  // For each bad word in the list, if the current sequence of input ids ends with this sequence (excluding the last),
5513
5513
  // then we set the logits of the last bad word id to -Infinity.
5514
- for (let i = 1; i <= bad_word_ids.length - 1 && bad_word_ids.length < input_ids[i].length; ++i) {
5514
+ for (let j = 1; j <= bad_word_ids.length - 1 && bad_word_ids.length < ids.length; ++j) {
5515
5515
 
5516
5516
  // NOTE: We use != instead of !== to compare bigint and number
5517
5517
  // @ts-ignore
5518
- if (bad_word_ids.at(-i - 1) != input_ids[i].at(-i)) {
5518
+ if (bad_word_ids.at(-j - 1) != ids.at(-j)) {
5519
5519
  // We have found a mismatch
5520
5520
  mark = false;
5521
5521
  break;
@@ -6349,6 +6349,7 @@ __webpack_require__.r(__webpack_exports__);
6349
6349
  /* harmony export */ AutoModelForTextToSpectrogram: () => (/* binding */ AutoModelForTextToSpectrogram),
6350
6350
  /* harmony export */ AutoModelForTextToWaveform: () => (/* binding */ AutoModelForTextToWaveform),
6351
6351
  /* harmony export */ AutoModelForTokenClassification: () => (/* binding */ AutoModelForTokenClassification),
6352
+ /* harmony export */ AutoModelForUniversalSegmentation: () => (/* binding */ AutoModelForUniversalSegmentation),
6352
6353
  /* harmony export */ AutoModelForVision2Seq: () => (/* binding */ AutoModelForVision2Seq),
6353
6354
  /* harmony export */ AutoModelForXVector: () => (/* binding */ AutoModelForXVector),
6354
6355
  /* harmony export */ AutoModelForZeroShotObjectDetection: () => (/* binding */ AutoModelForZeroShotObjectDetection),
@@ -6380,7 +6381,9 @@ __webpack_require__.r(__webpack_exports__);
6380
6381
  /* harmony export */ CLIPSegForImageSegmentation: () => (/* binding */ CLIPSegForImageSegmentation),
6381
6382
  /* harmony export */ CLIPSegModel: () => (/* binding */ CLIPSegModel),
6382
6383
  /* harmony export */ CLIPSegPreTrainedModel: () => (/* binding */ CLIPSegPreTrainedModel),
6384
+ /* harmony export */ CLIPTextModel: () => (/* binding */ CLIPTextModel),
6383
6385
  /* harmony export */ CLIPTextModelWithProjection: () => (/* binding */ CLIPTextModelWithProjection),
6386
+ /* harmony export */ CLIPVisionModel: () => (/* binding */ CLIPVisionModel),
6384
6387
  /* harmony export */ CLIPVisionModelWithProjection: () => (/* binding */ CLIPVisionModelWithProjection),
6385
6388
  /* harmony export */ CamembertForMaskedLM: () => (/* binding */ CamembertForMaskedLM),
6386
6389
  /* harmony export */ CamembertForQuestionAnswering: () => (/* binding */ CamembertForQuestionAnswering),
@@ -6429,6 +6432,8 @@ __webpack_require__.r(__webpack_exports__);
6429
6432
  /* harmony export */ DebertaV2ForTokenClassification: () => (/* binding */ DebertaV2ForTokenClassification),
6430
6433
  /* harmony export */ DebertaV2Model: () => (/* binding */ DebertaV2Model),
6431
6434
  /* harmony export */ DebertaV2PreTrainedModel: () => (/* binding */ DebertaV2PreTrainedModel),
6435
+ /* harmony export */ DecisionTransformerModel: () => (/* binding */ DecisionTransformerModel),
6436
+ /* harmony export */ DecisionTransformerPreTrainedModel: () => (/* binding */ DecisionTransformerPreTrainedModel),
6432
6437
  /* harmony export */ DeiTForImageClassification: () => (/* binding */ DeiTForImageClassification),
6433
6438
  /* harmony export */ DeiTModel: () => (/* binding */ DeiTModel),
6434
6439
  /* harmony export */ DeiTPreTrainedModel: () => (/* binding */ DeiTPreTrainedModel),
@@ -6497,6 +6502,11 @@ __webpack_require__.r(__webpack_exports__);
6497
6502
  /* harmony export */ GemmaForCausalLM: () => (/* binding */ GemmaForCausalLM),
6498
6503
  /* harmony export */ GemmaModel: () => (/* binding */ GemmaModel),
6499
6504
  /* harmony export */ GemmaPreTrainedModel: () => (/* binding */ GemmaPreTrainedModel),
6505
+ /* harmony export */ GroupViTModel: () => (/* binding */ GroupViTModel),
6506
+ /* harmony export */ GroupViTPreTrainedModel: () => (/* binding */ GroupViTPreTrainedModel),
6507
+ /* harmony export */ HieraForImageClassification: () => (/* binding */ HieraForImageClassification),
6508
+ /* harmony export */ HieraModel: () => (/* binding */ HieraModel),
6509
+ /* harmony export */ HieraPreTrainedModel: () => (/* binding */ HieraPreTrainedModel),
6500
6510
  /* harmony export */ HubertForCTC: () => (/* binding */ HubertForCTC),
6501
6511
  /* harmony export */ HubertForSequenceClassification: () => (/* binding */ HubertForSequenceClassification),
6502
6512
  /* harmony export */ HubertModel: () => (/* binding */ HubertModel),
@@ -6533,6 +6543,9 @@ __webpack_require__.r(__webpack_exports__);
6533
6543
  /* harmony export */ MarianMTModel: () => (/* binding */ MarianMTModel),
6534
6544
  /* harmony export */ MarianModel: () => (/* binding */ MarianModel),
6535
6545
  /* harmony export */ MarianPreTrainedModel: () => (/* binding */ MarianPreTrainedModel),
6546
+ /* harmony export */ MaskFormerForInstanceSegmentation: () => (/* binding */ MaskFormerForInstanceSegmentation),
6547
+ /* harmony export */ MaskFormerModel: () => (/* binding */ MaskFormerModel),
6548
+ /* harmony export */ MaskFormerPreTrainedModel: () => (/* binding */ MaskFormerPreTrainedModel),
6536
6549
  /* harmony export */ MaskedLMOutput: () => (/* binding */ MaskedLMOutput),
6537
6550
  /* harmony export */ MistralForCausalLM: () => (/* binding */ MistralForCausalLM),
6538
6551
  /* harmony export */ MistralModel: () => (/* binding */ MistralModel),
@@ -6591,6 +6604,9 @@ __webpack_require__.r(__webpack_exports__);
6591
6604
  /* harmony export */ PhiPreTrainedModel: () => (/* binding */ PhiPreTrainedModel),
6592
6605
  /* harmony export */ PreTrainedModel: () => (/* binding */ PreTrainedModel),
6593
6606
  /* harmony export */ PretrainedMixin: () => (/* binding */ PretrainedMixin),
6607
+ /* harmony export */ PvtForImageClassification: () => (/* binding */ PvtForImageClassification),
6608
+ /* harmony export */ PvtModel: () => (/* binding */ PvtModel),
6609
+ /* harmony export */ PvtPreTrainedModel: () => (/* binding */ PvtPreTrainedModel),
6594
6610
  /* harmony export */ PyAnnoteForAudioFrameClassification: () => (/* binding */ PyAnnoteForAudioFrameClassification),
6595
6611
  /* harmony export */ PyAnnoteModel: () => (/* binding */ PyAnnoteModel),
6596
6612
  /* harmony export */ PyAnnotePreTrainedModel: () => (/* binding */ PyAnnotePreTrainedModel),
@@ -6676,6 +6692,11 @@ __webpack_require__.r(__webpack_exports__);
6676
6692
  /* harmony export */ UniSpeechSatModel: () => (/* binding */ UniSpeechSatModel),
6677
6693
  /* harmony export */ UniSpeechSatPreTrainedModel: () => (/* binding */ UniSpeechSatPreTrainedModel),
6678
6694
  /* harmony export */ ViTForImageClassification: () => (/* binding */ ViTForImageClassification),
6695
+ /* harmony export */ ViTMAEModel: () => (/* binding */ ViTMAEModel),
6696
+ /* harmony export */ ViTMAEPreTrainedModel: () => (/* binding */ ViTMAEPreTrainedModel),
6697
+ /* harmony export */ ViTMSNForImageClassification: () => (/* binding */ ViTMSNForImageClassification),
6698
+ /* harmony export */ ViTMSNModel: () => (/* binding */ ViTMSNModel),
6699
+ /* harmony export */ ViTMSNPreTrainedModel: () => (/* binding */ ViTMSNPreTrainedModel),
6679
6700
  /* harmony export */ ViTModel: () => (/* binding */ ViTModel),
6680
6701
  /* harmony export */ ViTPreTrainedModel: () => (/* binding */ ViTPreTrainedModel),
6681
6702
  /* harmony export */ VisionEncoderDecoderModel: () => (/* binding */ VisionEncoderDecoderModel),
@@ -6925,6 +6946,7 @@ async function getSession(pretrained_model_name_or_path, fileName, options) {
6925
6946
  });
6926
6947
  if (Object.keys(shapes).length > 0 && !(0,_backends_onnx_js__WEBPACK_IMPORTED_MODULE_1__.isONNXProxy)()) {
6927
6948
  // Only set preferredOutputLocation if shapes are present and we aren't proxying ONNX
6949
+ /** @type {Record<string, import('onnxruntime-common').Tensor.DataLocation>} */
6928
6950
  const preferredOutputLocation = {};
6929
6951
  for (const key in shapes) {
6930
6952
  preferredOutputLocation[key] = 'gpu-buffer';
@@ -10190,6 +10212,18 @@ class CLIPPreTrainedModel extends PreTrainedModel { }
10190
10212
  */
10191
10213
  class CLIPModel extends CLIPPreTrainedModel { }
10192
10214
 
10215
+ /**
10216
+ * The text model from CLIP without any head or projection on top.
10217
+ */
10218
+ class CLIPTextModel extends CLIPPreTrainedModel {
10219
+ /** @type {PreTrainedModel.from_pretrained} */
10220
+ static async from_pretrained(pretrained_model_name_or_path, options = {}) {
10221
+ // Update default model file name if not provided
10222
+ options.model_file_name ??= 'text_model';
10223
+ return super.from_pretrained(pretrained_model_name_or_path, options);
10224
+ }
10225
+ }
10226
+
10193
10227
  /**
10194
10228
  * CLIP Text Model with a projection layer on top (a linear layer on top of the pooled output)
10195
10229
  *
@@ -10217,7 +10251,6 @@ class CLIPModel extends CLIPPreTrainedModel { }
10217
10251
  * ```
10218
10252
  */
10219
10253
  class CLIPTextModelWithProjection extends CLIPPreTrainedModel {
10220
-
10221
10254
  /** @type {PreTrainedModel.from_pretrained} */
10222
10255
  static async from_pretrained(pretrained_model_name_or_path, options = {}) {
10223
10256
  // Update default model file name if not provided
@@ -10226,6 +10259,18 @@ class CLIPTextModelWithProjection extends CLIPPreTrainedModel {
10226
10259
  }
10227
10260
  }
10228
10261
 
10262
+ /**
10263
+ * The vision model from CLIP without any head or projection on top.
10264
+ */
10265
+ class CLIPVisionModel extends CLIPPreTrainedModel {
10266
+ /** @type {PreTrainedModel.from_pretrained} */
10267
+ static async from_pretrained(pretrained_model_name_or_path, options = {}) {
10268
+ // Update default model file name if not provided
10269
+ options.model_file_name ??= 'vision_model';
10270
+ return super.from_pretrained(pretrained_model_name_or_path, options);
10271
+ }
10272
+ }
10273
+
10229
10274
  /**
10230
10275
  * CLIP Vision Model with a projection layer on top (a linear layer on top of the pooled output)
10231
10276
  *
@@ -10892,6 +10937,43 @@ class ViTForImageClassification extends ViTPreTrainedModel {
10892
10937
  }
10893
10938
  //////////////////////////////////////////////////
10894
10939
 
10940
+ //////////////////////////////////////////////////
10941
+ class PvtPreTrainedModel extends PreTrainedModel { }
10942
+ class PvtModel extends PvtPreTrainedModel { }
10943
+ class PvtForImageClassification extends PvtPreTrainedModel {
10944
+ /**
10945
+ * @param {any} model_inputs
10946
+ */
10947
+ async _call(model_inputs) {
10948
+ return new SequenceClassifierOutput(await super._call(model_inputs));
10949
+ }
10950
+ }
10951
+ //////////////////////////////////////////////////
10952
+
10953
+ //////////////////////////////////////////////////
10954
+ class ViTMAEPreTrainedModel extends PreTrainedModel { }
10955
+ class ViTMAEModel extends ViTMAEPreTrainedModel { }
10956
+ //////////////////////////////////////////////////
10957
+
10958
+
10959
+ //////////////////////////////////////////////////
10960
+ class ViTMSNPreTrainedModel extends PreTrainedModel { }
10961
+ class ViTMSNModel extends ViTMSNPreTrainedModel { }
10962
+ class ViTMSNForImageClassification extends ViTMSNPreTrainedModel {
10963
+ /**
10964
+ * @param {any} model_inputs
10965
+ */
10966
+ async _call(model_inputs) {
10967
+ return new SequenceClassifierOutput(await super._call(model_inputs));
10968
+ }
10969
+ }
10970
+ //////////////////////////////////////////////////
10971
+
10972
+ //////////////////////////////////////////////////
10973
+ class GroupViTPreTrainedModel extends PreTrainedModel { }
10974
+ class GroupViTModel extends GroupViTPreTrainedModel { }
10975
+ //////////////////////////////////////////////////
10976
+
10895
10977
 
10896
10978
  //////////////////////////////////////////////////
10897
10979
  class FastViTPreTrainedModel extends PreTrainedModel { }
@@ -11145,6 +11227,19 @@ class DeiTForImageClassification extends DeiTPreTrainedModel {
11145
11227
  }
11146
11228
  //////////////////////////////////////////////////
11147
11229
 
11230
+ //////////////////////////////////////////////////
11231
+ class HieraPreTrainedModel extends PreTrainedModel { }
11232
+ class HieraModel extends HieraPreTrainedModel { }
11233
+ class HieraForImageClassification extends HieraPreTrainedModel {
11234
+ /**
11235
+ * @param {any} model_inputs
11236
+ */
11237
+ async _call(model_inputs) {
11238
+ return new SequenceClassifierOutput(await super._call(model_inputs));
11239
+ }
11240
+ }
11241
+ //////////////////////////////////////////////////
11242
+
11148
11243
 
11149
11244
  //////////////////////////////////////////////////
11150
11245
  /**
@@ -11291,6 +11386,11 @@ class SapiensForDepthEstimation extends SapiensPreTrainedModel { }
11291
11386
  class SapiensForNormalEstimation extends SapiensPreTrainedModel { }
11292
11387
  //////////////////////////////////////////////////
11293
11388
 
11389
+ //////////////////////////////////////////////////
11390
+ class MaskFormerPreTrainedModel extends PreTrainedModel { }
11391
+ class MaskFormerModel extends MaskFormerPreTrainedModel { }
11392
+ class MaskFormerForInstanceSegmentation extends MaskFormerPreTrainedModel { }
11393
+ //////////////////////////////////////////////////
11294
11394
 
11295
11395
  //////////////////////////////////////////////////
11296
11396
  class GLPNPreTrainedModel extends PreTrainedModel { }
@@ -12813,6 +12913,7 @@ class MusicgenForConditionalGeneration extends PreTrainedModel { // NOTE: not Mu
12813
12913
  return audio_values;
12814
12914
  }
12815
12915
  }
12916
+ //////////////////////////////////////////////////
12816
12917
 
12817
12918
  //////////////////////////////////////////////////
12818
12919
  // MobileNetV1 models
@@ -12906,6 +13007,17 @@ class MobileNetV4ForImageClassification extends MobileNetV4PreTrainedModel {
12906
13007
  }
12907
13008
  //////////////////////////////////////////////////
12908
13009
 
13010
+ //////////////////////////////////////////////////
13011
+ // Decision Transformer models
13012
+ class DecisionTransformerPreTrainedModel extends PreTrainedModel { }
13013
+
13014
+ /**
13015
+ * The model builds upon the GPT2 architecture to perform autoregressive prediction of actions in an offline RL setting.
13016
+ * Refer to the paper for more details: https://arxiv.org/abs/2106.01345
13017
+ */
13018
+ class DecisionTransformerModel extends DecisionTransformerPreTrainedModel { }
13019
+
13020
+ //////////////////////////////////////////////////
12909
13021
 
12910
13022
  //////////////////////////////////////////////////
12911
13023
  // AutoModels, used to simplify construction of PreTrainedModels
@@ -12944,7 +13056,7 @@ class PretrainedMixin {
12944
13056
  session_options = {},
12945
13057
  } = {}) {
12946
13058
 
12947
- let options = {
13059
+ const options = {
12948
13060
  progress_callback,
12949
13061
  config,
12950
13062
  cache_dir,
@@ -12963,7 +13075,7 @@ class PretrainedMixin {
12963
13075
  throw new Error("`MODEL_CLASS_MAPPINGS` not implemented for this type of `AutoClass`: " + this.name);
12964
13076
  }
12965
13077
 
12966
- for (let MODEL_CLASS_MAPPING of this.MODEL_CLASS_MAPPINGS) {
13078
+ for (const MODEL_CLASS_MAPPING of this.MODEL_CLASS_MAPPINGS) {
12967
13079
  const modelInfo = MODEL_CLASS_MAPPING.get(options.config.model_type);
12968
13080
  if (!modelInfo) {
12969
13081
  continue; // Item not found in this mapping
@@ -13018,6 +13130,10 @@ const MODEL_MAPPING_NAMES_ENCODER_ONLY = new Map([
13018
13130
  ['rt_detr', ['RTDetrModel', RTDetrModel]],
13019
13131
  ['table-transformer', ['TableTransformerModel', TableTransformerModel]],
13020
13132
  ['vit', ['ViTModel', ViTModel]],
13133
+ ['pvt', ['PvtModel', PvtModel]],
13134
+ ['vit_msn', ['ViTMSNModel', ViTMSNModel]],
13135
+ ['vit_mae', ['ViTMAEModel', ViTMAEModel]],
13136
+ ['groupvit', ['GroupViTModel', GroupViTModel]],
13021
13137
  ['fastvit', ['FastViTModel', FastViTModel]],
13022
13138
  ['mobilevit', ['MobileViTModel', MobileViTModel]],
13023
13139
  ['mobilevitv2', ['MobileViTV2Model', MobileViTV2Model]],
@@ -13025,6 +13141,7 @@ const MODEL_MAPPING_NAMES_ENCODER_ONLY = new Map([
13025
13141
  ['owlv2', ['Owlv2Model', Owlv2Model]],
13026
13142
  ['beit', ['BeitModel', BeitModel]],
13027
13143
  ['deit', ['DeiTModel', DeiTModel]],
13144
+ ['hiera', ['HieraModel', HieraModel]],
13028
13145
  ['convnext', ['ConvNextModel', ConvNextModel]],
13029
13146
  ['convnextv2', ['ConvNextV2Model', ConvNextV2Model]],
13030
13147
  ['dinov2', ['Dinov2Model', Dinov2Model]],
@@ -13039,10 +13156,14 @@ const MODEL_MAPPING_NAMES_ENCODER_ONLY = new Map([
13039
13156
  ['hifigan', ['SpeechT5HifiGan', SpeechT5HifiGan]],
13040
13157
  ['efficientnet', ['EfficientNetModel', EfficientNetModel]],
13041
13158
 
13159
+ ['decision_transformer', ['DecisionTransformerModel', DecisionTransformerModel]],
13160
+
13042
13161
  ['mobilenet_v1', ['MobileNetV1Model', MobileNetV1Model]],
13043
13162
  ['mobilenet_v2', ['MobileNetV2Model', MobileNetV2Model]],
13044
13163
  ['mobilenet_v3', ['MobileNetV3Model', MobileNetV3Model]],
13045
13164
  ['mobilenet_v4', ['MobileNetV4Model', MobileNetV4Model]],
13165
+
13166
+ ['maskformer', ['MaskFormerModel', MaskFormerModel]],
13046
13167
  ]);
13047
13168
 
13048
13169
  const MODEL_MAPPING_NAMES_ENCODER_DECODER = new Map([
@@ -13227,11 +13348,14 @@ const MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING_NAMES = new Map([
13227
13348
 
13228
13349
  const MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES = new Map([
13229
13350
  ['vit', ['ViTForImageClassification', ViTForImageClassification]],
13351
+ ['pvt', ['PvtForImageClassification', PvtForImageClassification]],
13352
+ ['vit_msn', ['ViTMSNForImageClassification', ViTMSNForImageClassification]],
13230
13353
  ['fastvit', ['FastViTForImageClassification', FastViTForImageClassification]],
13231
13354
  ['mobilevit', ['MobileViTForImageClassification', MobileViTForImageClassification]],
13232
13355
  ['mobilevitv2', ['MobileViTV2ForImageClassification', MobileViTV2ForImageClassification]],
13233
13356
  ['beit', ['BeitForImageClassification', BeitForImageClassification]],
13234
13357
  ['deit', ['DeiTForImageClassification', DeiTForImageClassification]],
13358
+ ['hiera', ['HieraForImageClassification', HieraForImageClassification]],
13235
13359
  ['convnext', ['ConvNextForImageClassification', ConvNextForImageClassification]],
13236
13360
  ['convnextv2', ['ConvNextV2ForImageClassification', ConvNextV2ForImageClassification]],
13237
13361
  ['dinov2', ['Dinov2ForImageClassification', Dinov2ForImageClassification]],
@@ -13258,6 +13382,7 @@ const MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING_NAMES = new Map([
13258
13382
  ]);
13259
13383
 
13260
13384
  const MODEL_FOR_IMAGE_SEGMENTATION_MAPPING_NAMES = new Map([
13385
+ // TODO: Do not add new models here
13261
13386
  ['detr', ['DetrForSegmentation', DetrForSegmentation]],
13262
13387
  ['clipseg', ['CLIPSegForImageSegmentation', CLIPSegForImageSegmentation]],
13263
13388
  ]);
@@ -13267,6 +13392,11 @@ const MODEL_FOR_SEMANTIC_SEGMENTATION_MAPPING_NAMES = new Map([
13267
13392
  ['sapiens', ['SapiensForSemanticSegmentation', SapiensForSemanticSegmentation]],
13268
13393
  ]);
13269
13394
 
13395
+ const MODEL_FOR_UNIVERSAL_SEGMENTATION_MAPPING_NAMES = new Map([
13396
+ ['detr', ['DetrForSegmentation', DetrForSegmentation]],
13397
+ ['maskformer', ['MaskFormerForInstanceSegmentation', MaskFormerForInstanceSegmentation]],
13398
+ ]);
13399
+
13270
13400
  const MODEL_FOR_MASK_GENERATION_MAPPING_NAMES = new Map([
13271
13401
  ['sam', ['SamModel', SamModel]],
13272
13402
  ]);
@@ -13342,6 +13472,7 @@ const MODEL_CLASS_TYPE_MAPPING = [
13342
13472
  [MODEL_FOR_IMAGE_TEXT_TO_TEXT_MAPPING_NAMES, MODEL_TYPES.ImageTextToText],
13343
13473
  [MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES, MODEL_TYPES.EncoderOnly],
13344
13474
  [MODEL_FOR_IMAGE_SEGMENTATION_MAPPING_NAMES, MODEL_TYPES.EncoderOnly],
13475
+ [MODEL_FOR_UNIVERSAL_SEGMENTATION_MAPPING_NAMES, MODEL_TYPES.EncoderOnly],
13345
13476
  [MODEL_FOR_SEMANTIC_SEGMENTATION_MAPPING_NAMES, MODEL_TYPES.EncoderOnly],
13346
13477
  [MODEL_FOR_IMAGE_MATTING_MAPPING_NAMES, MODEL_TYPES.EncoderOnly],
13347
13478
  [MODEL_FOR_IMAGE_TO_IMAGE_MAPPING_NAMES, MODEL_TYPES.EncoderOnly],
@@ -13544,6 +13675,17 @@ class AutoModelForSemanticSegmentation extends PretrainedMixin {
13544
13675
  static MODEL_CLASS_MAPPINGS = [MODEL_FOR_SEMANTIC_SEGMENTATION_MAPPING_NAMES];
13545
13676
  }
13546
13677
 
13678
+ /**
13679
+ * Helper class which is used to instantiate pretrained universal image segmentation models with the `from_pretrained` function.
13680
+ * The chosen model class is determined by the type specified in the model config.
13681
+ *
13682
+ * @example
13683
+ * let model = await AutoModelForUniversalSegmentation.from_pretrained('hf-internal-testing/tiny-random-MaskFormerForInstanceSegmentation');
13684
+ */
13685
+ class AutoModelForUniversalSegmentation extends PretrainedMixin {
13686
+ static MODEL_CLASS_MAPPINGS = [MODEL_FOR_UNIVERSAL_SEGMENTATION_MAPPING_NAMES];
13687
+ }
13688
+
13547
13689
  /**
13548
13690
  * Helper class which is used to instantiate pretrained object detection models with the `from_pretrained` function.
13549
13691
  * The chosen model class is determined by the type specified in the model config.
@@ -14056,20 +14198,31 @@ __webpack_require__.r(__webpack_exports__);
14056
14198
 
14057
14199
 
14058
14200
 
14201
+ /**
14202
+ * Asynchronously creates a wrapper function for running an ONNX inference session.
14203
+ *
14204
+ * @param {number[]} session_bytes The session data in bytes.
14205
+ * @param {import('onnxruntime-common').InferenceSession.SessionOptions} session_options The options for the ONNX session.
14206
+ * @template {string | [string] | string[]} T
14207
+ * @param {T} names The name(s) of the output tensor(s).
14208
+ *
14209
+ * @returns {Promise<function(Record<string, Tensor>): Promise<T extends string ? Tensor : T extends string[] ? { [K in keyof T]: Tensor } : never>>}
14210
+ * The wrapper function for running the ONNX inference session.
14211
+ */
14059
14212
  const wrap = async (session_bytes, session_options, names) => {
14060
14213
  const session = await (0,_backends_onnx_js__WEBPACK_IMPORTED_MODULE_0__.createInferenceSession)(
14061
14214
  new Uint8Array(session_bytes), session_options,
14062
14215
  );
14063
- return async (inputs) => {
14216
+ return /** @type {any} */(async (/** @type {Record<string, Tensor>} */ inputs) => {
14064
14217
  const ortFeed = Object.fromEntries(Object.entries(inputs).map(([k, v]) => [k, v.ort_tensor]));
14065
14218
  const outputs = await session.run(ortFeed);
14066
14219
 
14067
14220
  if (Array.isArray(names)) {
14068
14221
  return names.map((n) => new _utils_tensor_js__WEBPACK_IMPORTED_MODULE_1__.Tensor(outputs[n]));
14069
14222
  } else {
14070
- return new _utils_tensor_js__WEBPACK_IMPORTED_MODULE_1__.Tensor(outputs[names]);
14223
+ return new _utils_tensor_js__WEBPACK_IMPORTED_MODULE_1__.Tensor(outputs[/** @type {string} */(names)]);
14071
14224
  }
14072
- }
14225
+ })
14073
14226
  }
14074
14227
 
14075
14228
  // In-memory registry of initialized ONNX operators
@@ -17192,7 +17345,7 @@ const SUPPORTED_TASKS = Object.freeze({
17192
17345
  "image-segmentation": {
17193
17346
  // no tokenizer
17194
17347
  "pipeline": ImageSegmentationPipeline,
17195
- "model": [_models_js__WEBPACK_IMPORTED_MODULE_1__.AutoModelForImageSegmentation, _models_js__WEBPACK_IMPORTED_MODULE_1__.AutoModelForSemanticSegmentation],
17348
+ "model": [_models_js__WEBPACK_IMPORTED_MODULE_1__.AutoModelForImageSegmentation, _models_js__WEBPACK_IMPORTED_MODULE_1__.AutoModelForSemanticSegmentation, _models_js__WEBPACK_IMPORTED_MODULE_1__.AutoModelForUniversalSegmentation],
17196
17349
  "processor": _processors_js__WEBPACK_IMPORTED_MODULE_2__.AutoProcessor,
17197
17350
  "default": {
17198
17351
  // TODO: replace with original
@@ -17434,7 +17587,7 @@ async function loadItems(mapping, model, pretrainedOptions) {
17434
17587
 
17435
17588
  /**@type {Promise[]} */
17436
17589
  const promises = [];
17437
- for (let [name, cls] of mapping.entries()) {
17590
+ for (const [name, cls] of mapping.entries()) {
17438
17591
  if (!cls) continue;
17439
17592
 
17440
17593
  /**@type {Promise} */
@@ -17442,7 +17595,7 @@ async function loadItems(mapping, model, pretrainedOptions) {
17442
17595
  if (Array.isArray(cls)) {
17443
17596
  promise = new Promise(async (resolve, reject) => {
17444
17597
  let e;
17445
- for (let c of cls) {
17598
+ for (const c of cls) {
17446
17599
  if (c === null) {
17447
17600
  // If null, we resolve it immediately, meaning the relevant
17448
17601
  // class was not found, but it is optional.
@@ -17480,7 +17633,7 @@ async function loadItems(mapping, model, pretrainedOptions) {
17480
17633
  await Promise.all(promises);
17481
17634
 
17482
17635
  // Then assign to result
17483
- for (let [name, promise] of Object.entries(result)) {
17636
+ for (const [name, promise] of Object.entries(result)) {
17484
17637
  result[name] = await promise;
17485
17638
  }
17486
17639
 
@@ -17517,6 +17670,7 @@ __webpack_require__.r(__webpack_exports__);
17517
17670
  /* harmony export */ Florence2Processor: () => (/* binding */ Florence2Processor),
17518
17671
  /* harmony export */ GLPNFeatureExtractor: () => (/* binding */ GLPNFeatureExtractor),
17519
17672
  /* harmony export */ ImageFeatureExtractor: () => (/* binding */ ImageFeatureExtractor),
17673
+ /* harmony export */ MaskFormerFeatureExtractor: () => (/* binding */ MaskFormerFeatureExtractor),
17520
17674
  /* harmony export */ MobileNetV1FeatureExtractor: () => (/* binding */ MobileNetV1FeatureExtractor),
17521
17675
  /* harmony export */ MobileNetV2FeatureExtractor: () => (/* binding */ MobileNetV2FeatureExtractor),
17522
17676
  /* harmony export */ MobileNetV3FeatureExtractor: () => (/* binding */ MobileNetV3FeatureExtractor),
@@ -17528,6 +17682,7 @@ __webpack_require__.r(__webpack_exports__);
17528
17682
  /* harmony export */ OwlViTProcessor: () => (/* binding */ OwlViTProcessor),
17529
17683
  /* harmony export */ Owlv2ImageProcessor: () => (/* binding */ Owlv2ImageProcessor),
17530
17684
  /* harmony export */ Processor: () => (/* binding */ Processor),
17685
+ /* harmony export */ PvtImageProcessor: () => (/* binding */ PvtImageProcessor),
17531
17686
  /* harmony export */ PyAnnoteFeatureExtractor: () => (/* binding */ PyAnnoteFeatureExtractor),
17532
17687
  /* harmony export */ PyAnnoteProcessor: () => (/* binding */ PyAnnoteProcessor),
17533
17688
  /* harmony export */ RTDetrImageProcessor: () => (/* binding */ RTDetrImageProcessor),
@@ -17616,7 +17771,7 @@ function center_to_corners_format([centerX, centerY, width, height]) {
17616
17771
  * @param {Tensor} outputs.logits The logits
17617
17772
  * @param {Tensor} outputs.pred_boxes The predicted boxes.
17618
17773
  * @param {number} [threshold=0.5] The threshold to use for the scores.
17619
- * @param {number[][]} [target_sizes=null] The sizes of the original images.
17774
+ * @param {[number, number][]} [target_sizes=null] The sizes of the original images.
17620
17775
  * @param {boolean} [is_zero_shot=false] Whether zero-shot object detection was performed.
17621
17776
  * @return {Object[]} An array of objects containing the post-processed outputs.
17622
17777
  * @private
@@ -17697,7 +17852,7 @@ function post_process_object_detection(outputs, threshold = 0.5, target_sizes =
17697
17852
  /**
17698
17853
  * Post-processes the outputs of the model (for semantic segmentation).
17699
17854
  * @param {*} outputs Raw outputs of the model.
17700
- * @param {number[][]} [target_sizes=null] List of tuples corresponding to the requested final size
17855
+ * @param {[number, number][]} [target_sizes=null] List of tuples corresponding to the requested final size
17701
17856
  * (height, width) of each prediction. If unset, predictions will not be resized.
17702
17857
  * @returns {{segmentation: Tensor; labels: number[]}[]} The semantic segmentation maps.
17703
17858
  */
@@ -17745,9 +17900,8 @@ function post_process_semantic_segmentation(outputs, target_sizes = null) {
17745
17900
  // Store which objects have labels
17746
17901
  // This is much more efficient that creating a set of the final values
17747
17902
  const hasLabel = new Array(data.dims[0]);
17748
- const out = segmentation.data;
17749
- for (let j = 0; j < out.length; ++j) {
17750
- const index = out[j];
17903
+ for (let j = 0; j < segmentation_data.length; ++j) {
17904
+ const index = segmentation_data[j];
17751
17905
  hasLabel[index] = index;
17752
17906
  }
17753
17907
  /** @type {number[]} The unique list of labels that were detected */
@@ -17758,6 +17912,300 @@ function post_process_semantic_segmentation(outputs, target_sizes = null) {
17758
17912
  return toReturn;
17759
17913
  }
17760
17914
 
17915
+
17916
+ /**
17917
+ * Binarize the given masks using `object_mask_threshold`, it returns the associated values of `masks`, `scores` and `labels`.
17918
+ * @param {Tensor} class_logits The class logits.
17919
+ * @param {Tensor} mask_logits The mask logits.
17920
+ * @param {number} object_mask_threshold A number between 0 and 1 used to binarize the masks.
17921
+ * @param {number} num_labels The number of labels.
17922
+ * @returns {[Tensor[], number[], number[]]} The binarized masks, the scores, and the labels.
17923
+ * @private
17924
+ */
17925
+ function remove_low_and_no_objects(class_logits, mask_logits, object_mask_threshold, num_labels) {
17926
+
17927
+ const mask_probs_item = [];
17928
+ const pred_scores_item = [];
17929
+ const pred_labels_item = [];
17930
+
17931
+ for (let j = 0; j < class_logits.dims[0]; ++j) {
17932
+ const cls = class_logits[j];
17933
+ const mask = mask_logits[j];
17934
+
17935
+ const pred_label = (0,_utils_maths_js__WEBPACK_IMPORTED_MODULE_3__.max)(cls.data)[1];
17936
+ if (pred_label === num_labels) {
17937
+ // Is the background, so we ignore it
17938
+ continue;
17939
+ }
17940
+
17941
+ const scores = (0,_utils_maths_js__WEBPACK_IMPORTED_MODULE_3__.softmax)(cls.data);
17942
+ const pred_score = scores[pred_label];
17943
+ if (pred_score > object_mask_threshold) {
17944
+ mask_probs_item.push(mask);
17945
+ pred_scores_item.push(pred_score);
17946
+ pred_labels_item.push(pred_label);
17947
+ }
17948
+ }
17949
+
17950
+ return [mask_probs_item, pred_scores_item, pred_labels_item];
17951
+ }
17952
+
17953
+ /**
17954
+ * Checks whether the segment is valid or not.
17955
+ * @param {Int32Array} mask_labels Labels for each pixel in the mask.
17956
+ * @param {Tensor[]} mask_probs Probabilities for each pixel in the masks.
17957
+ * @param {number} k The class id of the segment.
17958
+ * @param {number} mask_threshold The mask threshold.
17959
+ * @param {number} overlap_mask_area_threshold The overlap mask area threshold.
17960
+ * @returns {[boolean, number[]]} Whether the segment is valid or not, and the indices of the valid labels.
17961
+ * @private
17962
+ */
17963
+ function check_segment_validity(
17964
+ mask_labels,
17965
+ mask_probs,
17966
+ k,
17967
+ mask_threshold = 0.5,
17968
+ overlap_mask_area_threshold = 0.8
17969
+ ) {
17970
+ // mask_k is a 1D array of indices, indicating where the mask is equal to k
17971
+ const mask_k = [];
17972
+ let mask_k_area = 0;
17973
+ let original_area = 0;
17974
+
17975
+ const mask_probs_k_data = mask_probs[k].data;
17976
+
17977
+ // Compute the area of all the stuff in query k
17978
+ for (let i = 0; i < mask_labels.length; ++i) {
17979
+ if (mask_labels[i] === k) {
17980
+ mask_k.push(i);
17981
+ ++mask_k_area;
17982
+ }
17983
+
17984
+ if (mask_probs_k_data[i] >= mask_threshold) {
17985
+ ++original_area;
17986
+ }
17987
+ }
17988
+ let mask_exists = mask_k_area > 0 && original_area > 0;
17989
+
17990
+ // Eliminate disconnected tiny segments
17991
+ if (mask_exists) {
17992
+ // Perform additional check
17993
+ let area_ratio = mask_k_area / original_area;
17994
+ mask_exists = area_ratio > overlap_mask_area_threshold;
17995
+ }
17996
+
17997
+ return [mask_exists, mask_k]
17998
+ }
17999
+
18000
+ /**
18001
+ * Computes the segments.
18002
+ * @param {Tensor[]} mask_probs The mask probabilities.
18003
+ * @param {number[]} pred_scores The predicted scores.
18004
+ * @param {number[]} pred_labels The predicted labels.
18005
+ * @param {number} mask_threshold The mask threshold.
18006
+ * @param {number} overlap_mask_area_threshold The overlap mask area threshold.
18007
+ * @param {Set<number>} label_ids_to_fuse The label ids to fuse.
18008
+ * @param {number[]} target_size The target size of the image.
18009
+ * @returns {[Tensor, Array<{id: number, label_id: number, score: number}>]} The computed segments.
18010
+ * @private
18011
+ */
18012
+ function compute_segments(
18013
+ mask_probs,
18014
+ pred_scores,
18015
+ pred_labels,
18016
+ mask_threshold,
18017
+ overlap_mask_area_threshold,
18018
+ label_ids_to_fuse = null,
18019
+ target_size = null,
18020
+ ) {
18021
+ const [height, width] = target_size ?? mask_probs[0].dims;
18022
+
18023
+ const segmentation = new _utils_tensor_js__WEBPACK_IMPORTED_MODULE_4__.Tensor(
18024
+ 'int32',
18025
+ new Int32Array(height * width),
18026
+ [height, width]
18027
+ );
18028
+ const segments = [];
18029
+
18030
+ // 1. If target_size is not null, we need to resize the masks to the target size
18031
+ if (target_size !== null) {
18032
+ // resize the masks to the target size
18033
+ for (let i = 0; i < mask_probs.length; ++i) {
18034
+ mask_probs[i] = (0,_utils_tensor_js__WEBPACK_IMPORTED_MODULE_4__.interpolate)(mask_probs[i], target_size, 'bilinear', false);
18035
+ }
18036
+ }
18037
+
18038
+ // 2. Weigh each mask by its prediction score
18039
+ // NOTE: `mask_probs` is updated in-place
18040
+ //
18041
+ // Temporary storage for the best label/scores for each pixel ([height, width]):
18042
+ const mask_labels = new Int32Array(mask_probs[0].data.length);
18043
+ const bestScores = new Float32Array(mask_probs[0].data.length);
18044
+
18045
+ for (let i = 0; i < mask_probs.length; ++i) {
18046
+ let score = pred_scores[i];
18047
+
18048
+ const mask_probs_i_data = mask_probs[i].data;
18049
+
18050
+ for (let j = 0; j < mask_probs_i_data.length; ++j) {
18051
+ mask_probs_i_data[j] *= score
18052
+ if (mask_probs_i_data[j] > bestScores[j]) {
18053
+ mask_labels[j] = i;
18054
+ bestScores[j] = mask_probs_i_data[j];
18055
+ }
18056
+ }
18057
+ }
18058
+
18059
+ let current_segment_id = 0;
18060
+
18061
+ // let stuff_memory_list = {}
18062
+ const segmentation_data = segmentation.data;
18063
+ for (let k = 0; k < pred_labels.length; ++k) {
18064
+ const pred_class = pred_labels[k];
18065
+
18066
+ // TODO add `should_fuse`
18067
+ // let should_fuse = pred_class in label_ids_to_fuse
18068
+
18069
+ // Check if mask exists and large enough to be a segment
18070
+ const [mask_exists, mask_k] = check_segment_validity(
18071
+ mask_labels,
18072
+ mask_probs,
18073
+ k,
18074
+ mask_threshold,
18075
+ overlap_mask_area_threshold
18076
+ )
18077
+
18078
+ if (!mask_exists) {
18079
+ // Nothing to see here
18080
+ continue;
18081
+ }
18082
+
18083
+ // TODO
18084
+ // if (pred_class in stuff_memory_list) {
18085
+ // current_segment_id = stuff_memory_list[pred_class]
18086
+ // } else {
18087
+ // current_segment_id += 1;
18088
+ // }
18089
+ ++current_segment_id;
18090
+
18091
+
18092
+ // Add current object segment to final segmentation map
18093
+ for (const index of mask_k) {
18094
+ segmentation_data[index] = current_segment_id;
18095
+ }
18096
+
18097
+ segments.push({
18098
+ id: current_segment_id,
18099
+ label_id: pred_class,
18100
+ // was_fused: should_fuse, TODO
18101
+ score: pred_scores[k],
18102
+ })
18103
+
18104
+ // TODO
18105
+ // if(should_fuse){
18106
+ // stuff_memory_list[pred_class] = current_segment_id
18107
+ // }
18108
+ }
18109
+
18110
+ return [segmentation, segments];
18111
+ }
18112
+
18113
+
18114
+ /**
18115
+ * Post-process the model output to generate the final panoptic segmentation.
18116
+ * @param {*} outputs The model output to post process
18117
+ * @param {number} [threshold=0.5] The probability score threshold to keep predicted instance masks.
18118
+ * @param {number} [mask_threshold=0.5] Threshold to use when turning the predicted masks into binary values.
18119
+ * @param {number} [overlap_mask_area_threshold=0.8] The overlap mask area threshold to merge or discard small disconnected parts within each binary instance mask.
18120
+ * @param {Set<number>} [label_ids_to_fuse=null] The labels in this state will have all their instances be fused together.
18121
+ * @param {[number, number][]} [target_sizes=null] The target sizes to resize the masks to.
18122
+ * @returns {Array<{ segmentation: Tensor, segments_info: Array<{id: number, label_id: number, score: number}>}>}
18123
+ */
18124
+ function post_process_panoptic_segmentation(
18125
+ outputs,
18126
+ threshold = 0.5,
18127
+ mask_threshold = 0.5,
18128
+ overlap_mask_area_threshold = 0.8,
18129
+ label_ids_to_fuse = null,
18130
+ target_sizes = null,
18131
+ ) {
18132
+ if (label_ids_to_fuse === null) {
18133
+ console.warn("`label_ids_to_fuse` unset. No instance will be fused.")
18134
+ label_ids_to_fuse = new Set();
18135
+ }
18136
+
18137
+ const class_queries_logits = outputs.class_queries_logits ?? outputs.logits; // [batch_size, num_queries, num_classes+1]
18138
+ const masks_queries_logits = outputs.masks_queries_logits ?? outputs.pred_masks; // [batch_size, num_queries, height, width]
18139
+
18140
+ const mask_probs = masks_queries_logits.sigmoid() // [batch_size, num_queries, height, width]
18141
+
18142
+ let [batch_size, num_queries, num_labels] = class_queries_logits.dims;
18143
+ num_labels -= 1; // Remove last class (background)
18144
+
18145
+ if (target_sizes !== null && target_sizes.length !== batch_size) {
18146
+ throw Error("Make sure that you pass in as many target sizes as the batch dimension of the logits")
18147
+ }
18148
+
18149
+ let toReturn = [];
18150
+ for (let i = 0; i < batch_size; ++i) {
18151
+ let target_size = target_sizes !== null ? target_sizes[i] : null;
18152
+
18153
+ let class_logits = class_queries_logits[i];
18154
+ let mask_logits = mask_probs[i];
18155
+
18156
+ let [mask_probs_item, pred_scores_item, pred_labels_item] = remove_low_and_no_objects(class_logits, mask_logits, threshold, num_labels);
18157
+
18158
+ if (pred_labels_item.length === 0) {
18159
+ // No mask found
18160
+ let [height, width] = target_size ?? mask_logits.dims.slice(-2);
18161
+
18162
+ let segmentation = new _utils_tensor_js__WEBPACK_IMPORTED_MODULE_4__.Tensor(
18163
+ 'int32',
18164
+ new Int32Array(height * width).fill(-1),
18165
+ [height, width]
18166
+ )
18167
+ toReturn.push({
18168
+ segmentation: segmentation,
18169
+ segments_info: []
18170
+ });
18171
+ continue;
18172
+ }
18173
+
18174
+
18175
+ // Get segmentation map and segment information of batch item
18176
+ let [segmentation, segments] = compute_segments(
18177
+ mask_probs_item,
18178
+ pred_scores_item,
18179
+ pred_labels_item,
18180
+ mask_threshold,
18181
+ overlap_mask_area_threshold,
18182
+ label_ids_to_fuse,
18183
+ target_size,
18184
+ )
18185
+
18186
+ toReturn.push({
18187
+ segmentation: segmentation,
18188
+ segments_info: segments
18189
+ })
18190
+ }
18191
+
18192
+ return toReturn;
18193
+ }
18194
+
18195
+
18196
+ /**
18197
+ * Post-processes the outputs of the model (for instance segmentation).
18198
+ * @param {*} outputs Raw outputs of the model.
18199
+ * @param {number} [threshold=0.5] The probability score threshold to keep predicted instance masks.
18200
+ * @param {[number, number][]} [target_sizes=null] List of tuples corresponding to the requested final size
18201
+ * (height, width) of each prediction. If unset, predictions will not be resized.
18202
+ * @returns {Array<{ segmentation: Tensor, segments_info: Array<{id: number, label_id: number, score: number}>}>}
18203
+ */
18204
+ function post_process_instance_segmentation(outputs, threshold = 0.5, target_sizes = null) {
18205
+ throw new Error('Not implemented yet');
18206
+ return [];
18207
+ }
18208
+
17761
18209
  /**
17762
18210
  * Named tuple to indicate the order we are using is (height x width), even though
17763
18211
  * the Graphics’ industry standard is (width x height).
@@ -18346,6 +18794,7 @@ class SegformerFeatureExtractor extends ImageFeatureExtractor {
18346
18794
  return post_process_semantic_segmentation(...args);
18347
18795
  }
18348
18796
  }
18797
+ class PvtImageProcessor extends ImageFeatureExtractor { }
18349
18798
  class DPTFeatureExtractor extends ImageFeatureExtractor { }
18350
18799
  class DPTImageProcessor extends DPTFeatureExtractor { } // NOTE: extends DPTFeatureExtractor
18351
18800
  class BitImageProcessor extends ImageFeatureExtractor { }
@@ -18485,302 +18934,32 @@ class DetrFeatureExtractor extends ImageFeatureExtractor {
18485
18934
  // TODO support different mask sizes (not just 64x64)
18486
18935
  // Currently, just fill pixel mask with 1s
18487
18936
  const maskSize = [result.pixel_values.dims[0], 64, 64];
18488
- const pixel_mask = new _utils_tensor_js__WEBPACK_IMPORTED_MODULE_4__.Tensor(
18489
- 'int64',
18490
- new BigInt64Array(maskSize.reduce((a, b) => a * b)).fill(1n),
18491
- maskSize
18492
- );
18937
+ const pixel_mask = (0,_utils_tensor_js__WEBPACK_IMPORTED_MODULE_4__.full)(maskSize, 1n);
18493
18938
 
18494
18939
  return { ...result, pixel_mask };
18495
18940
  }
18496
18941
 
18497
- /**
18498
- * Post-processes the outputs of the model (for object detection).
18499
- * @param {Object} outputs The outputs of the model that must be post-processed
18500
- * @param {Tensor} outputs.logits The logits
18501
- * @param {Tensor} outputs.pred_boxes The predicted boxes.
18502
- * @return {Object[]} An array of objects containing the post-processed outputs.
18503
- */
18504
-
18505
18942
  /** @type {typeof post_process_object_detection} */
18506
18943
  post_process_object_detection(...args) {
18507
18944
  return post_process_object_detection(...args);
18508
18945
  }
18509
18946
 
18510
- /**
18511
- * Binarize the given masks using `object_mask_threshold`, it returns the associated values of `masks`, `scores` and `labels`.
18512
- * @param {Tensor} class_logits The class logits.
18513
- * @param {Tensor} mask_logits The mask logits.
18514
- * @param {number} object_mask_threshold A number between 0 and 1 used to binarize the masks.
18515
- * @param {number} num_labels The number of labels.
18516
- * @returns {[Tensor[], number[], number[]]} The binarized masks, the scores, and the labels.
18517
- */
18518
- remove_low_and_no_objects(class_logits, mask_logits, object_mask_threshold, num_labels) {
18519
-
18520
- let mask_probs_item = [];
18521
- let pred_scores_item = [];
18522
- let pred_labels_item = [];
18523
-
18524
- for (let j = 0; j < class_logits.dims[0]; ++j) {
18525
- let cls = class_logits[j];
18526
- let mask = mask_logits[j];
18527
-
18528
- let pred_label = (0,_utils_maths_js__WEBPACK_IMPORTED_MODULE_3__.max)(cls.data)[1];
18529
- if (pred_label === num_labels) {
18530
- // Is the background, so we ignore it
18531
- continue;
18532
- }
18533
-
18534
- let scores = (0,_utils_maths_js__WEBPACK_IMPORTED_MODULE_3__.softmax)(cls.data);
18535
- let pred_score = scores[pred_label];
18536
- if (pred_score > object_mask_threshold) {
18537
- mask_probs_item.push(mask);
18538
- pred_scores_item.push(pred_score);
18539
- pred_labels_item.push(pred_label);
18540
- }
18541
- }
18542
-
18543
- return [mask_probs_item, pred_scores_item, pred_labels_item];
18544
-
18947
+ /** @type {typeof post_process_panoptic_segmentation} */
18948
+ post_process_panoptic_segmentation(...args) {
18949
+ return post_process_panoptic_segmentation(...args);
18545
18950
  }
18546
18951
 
18547
- /**
18548
- * Checks whether the segment is valid or not.
18549
- * @param {Int32Array} mask_labels Labels for each pixel in the mask.
18550
- * @param {Tensor[]} mask_probs Probabilities for each pixel in the masks.
18551
- * @param {number} k The class id of the segment.
18552
- * @param {number} mask_threshold The mask threshold.
18553
- * @param {number} overlap_mask_area_threshold The overlap mask area threshold.
18554
- * @returns {[boolean, number[]]} Whether the segment is valid or not, and the indices of the valid labels.
18555
- */
18556
- check_segment_validity(
18557
- mask_labels,
18558
- mask_probs,
18559
- k,
18560
- mask_threshold = 0.5,
18561
- overlap_mask_area_threshold = 0.8
18562
- ) {
18563
- // mask_k is a 1D array of indices, indicating where the mask is equal to k
18564
- let mask_k = [];
18565
- let mask_k_area = 0;
18566
- let original_area = 0;
18567
-
18568
- const mask_probs_k_data = mask_probs[k].data;
18569
-
18570
- // Compute the area of all the stuff in query k
18571
- for (let i = 0; i < mask_labels.length; ++i) {
18572
- if (mask_labels[i] === k) {
18573
- mask_k.push(i);
18574
- ++mask_k_area;
18575
- }
18576
-
18577
- if (mask_probs_k_data[i] >= mask_threshold) {
18578
- ++original_area;
18579
- }
18580
- }
18581
- let mask_exists = mask_k_area > 0 && original_area > 0;
18582
-
18583
- // Eliminate disconnected tiny segments
18584
- if (mask_exists) {
18585
- // Perform additional check
18586
- let area_ratio = mask_k_area / original_area;
18587
- mask_exists = area_ratio > overlap_mask_area_threshold;
18588
- }
18589
-
18590
- return [mask_exists, mask_k]
18591
- }
18592
-
18593
- /**
18594
- * Computes the segments.
18595
- * @param {Tensor[]} mask_probs The mask probabilities.
18596
- * @param {number[]} pred_scores The predicted scores.
18597
- * @param {number[]} pred_labels The predicted labels.
18598
- * @param {number} mask_threshold The mask threshold.
18599
- * @param {number} overlap_mask_area_threshold The overlap mask area threshold.
18600
- * @param {Set<number>} label_ids_to_fuse The label ids to fuse.
18601
- * @param {number[]} target_size The target size of the image.
18602
- * @returns {[Tensor, Array<{id: number, label_id: number, score: number}>]} The computed segments.
18603
- */
18604
- compute_segments(
18605
- mask_probs,
18606
- pred_scores,
18607
- pred_labels,
18608
- mask_threshold,
18609
- overlap_mask_area_threshold,
18610
- label_ids_to_fuse = null,
18611
- target_size = null,
18612
- ) {
18613
- let [height, width] = target_size ?? mask_probs[0].dims;
18614
-
18615
- let segmentation = new _utils_tensor_js__WEBPACK_IMPORTED_MODULE_4__.Tensor(
18616
- 'int32',
18617
- new Int32Array(height * width),
18618
- [height, width]
18619
- );
18620
- let segments = [];
18621
-
18622
- // 1. If target_size is not null, we need to resize the masks to the target size
18623
- if (target_size !== null) {
18624
- // resize the masks to the target size
18625
- for (let i = 0; i < mask_probs.length; ++i) {
18626
- mask_probs[i] = (0,_utils_tensor_js__WEBPACK_IMPORTED_MODULE_4__.interpolate)(mask_probs[i], target_size, 'bilinear', false);
18627
- }
18628
- }
18629
-
18630
- // 2. Weigh each mask by its prediction score
18631
- // NOTE: `mask_probs` is updated in-place
18632
- //
18633
- // Temporary storage for the best label/scores for each pixel ([height, width]):
18634
- let mask_labels = new Int32Array(mask_probs[0].data.length);
18635
- let bestScores = new Float32Array(mask_probs[0].data.length);
18636
-
18637
- for (let i = 0; i < mask_probs.length; ++i) {
18638
- let score = pred_scores[i];
18639
-
18640
- const mask_probs_i_data = mask_probs[i].data;
18641
-
18642
- for (let j = 0; j < mask_probs_i_data.length; ++j) {
18643
- mask_probs_i_data[j] *= score
18644
- if (mask_probs_i_data[j] > bestScores[j]) {
18645
- mask_labels[j] = i;
18646
- bestScores[j] = mask_probs_i_data[j];
18647
- }
18648
- }
18649
- }
18650
-
18651
- let current_segment_id = 0;
18652
-
18653
- // let stuff_memory_list = {}
18654
- const segmentation_data = segmentation.data;
18655
- for (let k = 0; k < pred_labels.length; ++k) {
18656
- let pred_class = pred_labels[k];
18657
-
18658
- // TODO add `should_fuse`
18659
- // let should_fuse = pred_class in label_ids_to_fuse
18660
-
18661
- // Check if mask exists and large enough to be a segment
18662
- let [mask_exists, mask_k] = this.check_segment_validity(
18663
- mask_labels,
18664
- mask_probs,
18665
- k,
18666
- mask_threshold,
18667
- overlap_mask_area_threshold
18668
- )
18669
-
18670
- if (!mask_exists) {
18671
- // Nothing to see here
18672
- continue;
18673
- }
18674
-
18675
- // TODO
18676
- // if (pred_class in stuff_memory_list) {
18677
- // current_segment_id = stuff_memory_list[pred_class]
18678
- // } else {
18679
- // current_segment_id += 1;
18680
- // }
18681
- ++current_segment_id;
18682
-
18683
-
18684
- // Add current object segment to final segmentation map
18685
- for (let index of mask_k) {
18686
- segmentation_data[index] = current_segment_id;
18687
- }
18688
-
18689
- segments.push({
18690
- id: current_segment_id,
18691
- label_id: pred_class,
18692
- // was_fused: should_fuse, TODO
18693
- score: pred_scores[k],
18694
- })
18695
-
18696
- // TODO
18697
- // if(should_fuse){
18698
- // stuff_memory_list[pred_class] = current_segment_id
18699
- // }
18700
- }
18701
-
18702
- return [segmentation, segments];
18952
+ post_process_instance_segmentation() {
18953
+ // TODO
18954
+ throw Error("Not implemented yet");
18703
18955
  }
18956
+ }
18704
18957
 
18705
- /**
18706
- * Post-process the model output to generate the final panoptic segmentation.
18707
- * @param {*} outputs The model output to post process
18708
- * @param {number} [threshold=0.5] The probability score threshold to keep predicted instance masks.
18709
- * @param {number} [mask_threshold=0.5] Threshold to use when turning the predicted masks into binary values.
18710
- * @param {number} [overlap_mask_area_threshold=0.8] The overlap mask area threshold to merge or discard small disconnected parts within each binary instance mask.
18711
- * @param {Set<number>} [label_ids_to_fuse=null] The labels in this state will have all their instances be fused together.
18712
- * @param {number[][]} [target_sizes=null] The target sizes to resize the masks to.
18713
- * @returns {Array<{ segmentation: Tensor, segments_info: Array<{id: number, label_id: number, score: number}>}>}
18714
- */
18715
- post_process_panoptic_segmentation(
18716
- outputs,
18717
- threshold = 0.5,
18718
- mask_threshold = 0.5,
18719
- overlap_mask_area_threshold = 0.8,
18720
- label_ids_to_fuse = null,
18721
- target_sizes = null,
18722
- ) {
18723
- if (label_ids_to_fuse === null) {
18724
- console.warn("`label_ids_to_fuse` unset. No instance will be fused.")
18725
- label_ids_to_fuse = new Set();
18726
- }
18727
-
18728
- const class_queries_logits = outputs.logits; // [batch_size, num_queries, num_classes+1]
18729
- const masks_queries_logits = outputs.pred_masks; // [batch_size, num_queries, height, width]
18730
-
18731
- const mask_probs = masks_queries_logits.sigmoid() // [batch_size, num_queries, height, width]
18732
-
18733
- let [batch_size, num_queries, num_labels] = class_queries_logits.dims;
18734
- num_labels -= 1; // Remove last class (background)
18735
-
18736
- if (target_sizes !== null && target_sizes.length !== batch_size) {
18737
- throw Error("Make sure that you pass in as many target sizes as the batch dimension of the logits")
18738
- }
18739
-
18740
- let toReturn = [];
18741
- for (let i = 0; i < batch_size; ++i) {
18742
- let target_size = target_sizes !== null ? target_sizes[i] : null;
18743
-
18744
- let class_logits = class_queries_logits[i];
18745
- let mask_logits = mask_probs[i];
18746
-
18747
- let [mask_probs_item, pred_scores_item, pred_labels_item] = this.remove_low_and_no_objects(class_logits, mask_logits, threshold, num_labels);
18748
-
18749
- if (pred_labels_item.length === 0) {
18750
- // No mask found
18751
- let [height, width] = target_size ?? mask_logits.dims.slice(-2);
18752
-
18753
- let segmentation = new _utils_tensor_js__WEBPACK_IMPORTED_MODULE_4__.Tensor(
18754
- 'int32',
18755
- new Int32Array(height * width).fill(-1),
18756
- [height, width]
18757
- )
18758
- toReturn.push({
18759
- segmentation: segmentation,
18760
- segments_info: []
18761
- });
18762
- continue;
18763
- }
18764
-
18765
-
18766
- // Get segmentation map and segment information of batch item
18767
- let [segmentation, segments] = this.compute_segments(
18768
- mask_probs_item,
18769
- pred_scores_item,
18770
- pred_labels_item,
18771
- mask_threshold,
18772
- overlap_mask_area_threshold,
18773
- label_ids_to_fuse,
18774
- target_size,
18775
- )
18776
-
18777
- toReturn.push({
18778
- segmentation: segmentation,
18779
- segments_info: segments
18780
- })
18781
- }
18958
+ class MaskFormerFeatureExtractor extends ImageFeatureExtractor {
18782
18959
 
18783
- return toReturn;
18960
+ /** @type {typeof post_process_panoptic_segmentation} */
18961
+ post_process_panoptic_segmentation(...args) {
18962
+ return post_process_panoptic_segmentation(...args);
18784
18963
  }
18785
18964
 
18786
18965
  post_process_instance_segmentation() {
@@ -18789,6 +18968,7 @@ class DetrFeatureExtractor extends ImageFeatureExtractor {
18789
18968
  }
18790
18969
  }
18791
18970
 
18971
+
18792
18972
  class YolosFeatureExtractor extends ImageFeatureExtractor {
18793
18973
  /** @type {typeof post_process_object_detection} */
18794
18974
  post_process_object_detection(...args) {
@@ -20078,11 +20258,13 @@ class AutoProcessor {
20078
20258
  BitImageProcessor,
20079
20259
  DPTImageProcessor,
20080
20260
  DPTFeatureExtractor,
20261
+ PvtImageProcessor,
20081
20262
  GLPNFeatureExtractor,
20082
20263
  BeitFeatureExtractor,
20083
20264
  DeiTFeatureExtractor,
20084
20265
  DetrFeatureExtractor,
20085
20266
  RTDetrImageProcessor,
20267
+ MaskFormerFeatureExtractor,
20086
20268
  YolosFeatureExtractor,
20087
20269
  DonutFeatureExtractor,
20088
20270
  NougatImageProcessor,
@@ -20459,23 +20641,26 @@ function is_chinese_char(cp) {
20459
20641
  }
20460
20642
 
20461
20643
  /**
20462
- * Helper function to fuse consecutive values in an array equal to the specified value.
20463
- * @param {string[]} arr The input array
20464
- * @param {any} value The value to fuse on.
20465
- * @param {Map<string, any>} mapping The mapping from input domain to value.
20644
+ * Helper function to fuse consecutive unknown tokens.
20645
+ * @param {string[]} arr The list of input tokens
20646
+ * @param {Map<string, any>} tokens_to_ids The mapping from tokens to token ids.
20647
+ * @param {number} unk_token_id The value to fuse on.
20648
+ * @private
20466
20649
  */
20467
- function fuse(arr, value, mapping) {
20650
+ function fuse_unk(arr, tokens_to_ids, unk_token_id) {
20468
20651
  const fused = [];
20469
20652
  let i = 0;
20470
20653
  while (i < arr.length) {
20471
20654
  fused.push(arr[i])
20472
- if ((mapping.get(arr[i]) ?? value) !== value) {
20655
+ if ((tokens_to_ids.get(arr[i]) ?? unk_token_id) !== unk_token_id) {
20473
20656
  ++i;
20474
20657
  continue;
20475
20658
  }
20476
20659
 
20477
- while (i < arr.length && (mapping.get(arr[i]) ?? value) === value) {
20478
- ++i;
20660
+ while (++i < arr.length && (tokens_to_ids.get(arr[i]) ?? unk_token_id) === unk_token_id) {
20661
+ if (tokens_to_ids.get(fused.at(-1)) !== unk_token_id) {
20662
+ fused[fused.length - 1] += arr[i];
20663
+ }
20479
20664
  }
20480
20665
  }
20481
20666
 
@@ -20592,15 +20777,15 @@ class TokenizerModel extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_0__.Call
20592
20777
  /**
20593
20778
  * Internal function to call the TokenizerModel instance.
20594
20779
  * @param {string[]} tokens The tokens to encode.
20595
- * @returns {string[]} The encoded token IDs.
20780
+ * @returns {string[]} The encoded tokens.
20596
20781
  */
20597
20782
  _call(tokens) {
20598
- let ids = this.encode(tokens);
20783
+ tokens = this.encode(tokens);
20599
20784
  if (this.fuse_unk) {
20600
20785
  // Fuse unknown tokens
20601
- ids = fuse(ids, this.unk_token_id, this.tokens_to_ids);
20786
+ tokens = fuse_unk(tokens, this.tokens_to_ids, this.unk_token_id);
20602
20787
  }
20603
- return ids;
20788
+ return tokens;
20604
20789
  }
20605
20790
 
20606
20791
  /**
@@ -21098,15 +21283,19 @@ class BPE extends TokenizerModel {
21098
21283
  for (const t of bpe_token_list) {
21099
21284
  if (this.tokens_to_ids.has(t)) {
21100
21285
  outputTokens.push(t);
21101
- } else {
21102
- if (this.byte_fallback) {
21103
- outputTokens.push(
21104
- ...Array.from(this.text_encoder.encode(t))
21105
- .map(x => `<0x${x.toString(16).toUpperCase().padStart(2, '0')}>`)
21106
- );
21286
+ } else if (this.byte_fallback) {
21287
+ const byteTokens = Array.from(this.text_encoder.encode(t))
21288
+ .map(x => `<0x${x.toString(16).toUpperCase().padStart(2, '0')}>`);
21289
+ if (byteTokens.every(x => this.tokens_to_ids.has(x))) {
21290
+ // Ensure the byte tokens are actually in the vocabulary, otherwise
21291
+ // we fall back to the unknown token. For more information, see
21292
+ // https://github.com/huggingface/transformers/issues/28096.
21293
+ outputTokens.push(...byteTokens);
21107
21294
  } else {
21108
21295
  outputTokens.push(this.unk_token);
21109
21296
  }
21297
+ } else {
21298
+ outputTokens.push(this.unk_token);
21110
21299
  }
21111
21300
  }
21112
21301
  }
@@ -22778,11 +22967,10 @@ class PreTrainedTokenizer extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_0__
22778
22967
  this.decoder.end_of_word_suffix = this.model.end_of_word_suffix;
22779
22968
  }
22780
22969
 
22781
-
22782
22970
  this.added_tokens_regex = this.added_tokens.length > 0 ? new RegExp(
22783
- this.added_tokens
22971
+ this.added_tokens.slice()
22784
22972
  // Sort by length (desc) to avoid early partial matches
22785
- .toSorted((a, b) => b.content.length - a.content.length)
22973
+ .sort((a, b) => b.content.length - a.content.length)
22786
22974
  .map(x => `${x.lstrip ? '\\s*' : ''}(${(0,_utils_core_js__WEBPACK_IMPORTED_MODULE_1__.escapeRegExp)(x.content)})${x.rstrip ? '\\s*' : ''}`)
22787
22975
  .join('|')
22788
22976
  ) : null;
@@ -23280,6 +23468,67 @@ class PreTrainedTokenizer extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_0__
23280
23468
 
23281
23469
  return decoded;
23282
23470
  }
23471
+
23472
+ /**
23473
+ * Retrieve the chat template string used for tokenizing chat messages. This template is used
23474
+ * internally by the `apply_chat_template` method and can also be used externally to retrieve the model's chat
23475
+ * template for better generation tracking.
23476
+ *
23477
+ * @param {Object} options An optional object containing the following properties:
23478
+ * @param {string} [options.chat_template=null]
23479
+ * A Jinja template or the name of a template to use for this conversion.
23480
+ * It is usually not necessary to pass anything to this argument,
23481
+ * as the model's template will be used by default.
23482
+ * @param {Object[]} [options.tools=null]
23483
+ * A list of tools (callable functions) that will be accessible to the model. If the template does not
23484
+ * support function calling, this argument will have no effect. Each tool should be passed as a JSON Schema,
23485
+ * giving the name, description and argument types for the tool. See our
23486
+ * [chat templating guide](https://huggingface.co/docs/transformers/main/en/chat_templating#automated-function-conversion-for-tool-use)
23487
+ * for more information.
23488
+ * @returns {string} The chat template string.
23489
+ */
23490
+ get_chat_template({
23491
+ chat_template = null,
23492
+ tools = null,
23493
+ } = {}) {
23494
+
23495
+ // First, handle the cases when the model has a dict of multiple templates
23496
+ if (this.chat_template && typeof this.chat_template === 'object') {
23497
+ const template_dict = this.chat_template;
23498
+
23499
+ if (chat_template !== null && Object.hasOwn(template_dict, chat_template)) {
23500
+ // The user can pass the name of a template to the chat template argument instead of an entire template
23501
+ chat_template = template_dict[chat_template];
23502
+ } else if (chat_template === null) {
23503
+ if (tools !== null && 'tool_use' in template_dict) {
23504
+ chat_template = template_dict['tool_use'];
23505
+ } else if ('default' in template_dict) {
23506
+ chat_template = template_dict['default'];
23507
+ } else {
23508
+ throw Error(
23509
+ `This model has multiple chat templates with no default specified! Please either pass a chat ` +
23510
+ `template or the name of the template you wish to use to the 'chat_template' argument. Available ` +
23511
+ `template names are ${Object.keys(template_dict).sort()}.`
23512
+ )
23513
+ }
23514
+ }
23515
+ } else if (chat_template === null) {
23516
+ // These are the cases when the model has a single template
23517
+ // priority: `chat_template` argument > `tokenizer.chat_template`
23518
+ if (this.chat_template) {
23519
+ chat_template = this.chat_template;
23520
+ } else {
23521
+ throw Error(
23522
+ "Cannot use apply_chat_template() because tokenizer.chat_template is not set and no template " +
23523
+ "argument was passed! For information about writing templates and setting the " +
23524
+ "tokenizer.chat_template attribute, please see the documentation at " +
23525
+ "https://huggingface.co/docs/transformers/main/en/chat_templating"
23526
+ )
23527
+ }
23528
+ }
23529
+ return chat_template;
23530
+ }
23531
+
23283
23532
  /**
23284
23533
  * Converts a list of message objects with `"role"` and `"content"` keys to a list of token
23285
23534
  * ids. This method is intended for use with chat models, and will read the tokenizer's chat_template attribute to
@@ -23353,39 +23602,8 @@ class PreTrainedTokenizer extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_0__
23353
23602
  ...kwargs
23354
23603
  } = {}) {
23355
23604
 
23356
- // First, handle the cases when the model has a dict of multiple templates
23357
- if (
23358
- (this.chat_template && typeof this.chat_template === 'object')
23359
- || this.chat_template === null
23360
- ) {
23361
- const template_dict = this.chat_template;
23605
+ chat_template = this.get_chat_template({ chat_template, tools });
23362
23606
 
23363
- if (chat_template !== null && Object.hasOwn(template_dict, chat_template)) {
23364
- // The user can pass the name of a template to the chat template argument instead of an entire template
23365
- chat_template = template_dict[chat_template];
23366
- } else if (chat_template === null && 'default' in template_dict) {
23367
- chat_template = template_dict['default'];
23368
- } else if (chat_template === null) {
23369
- throw Error(
23370
- `This model has multiple chat templates with no default specified! Please either pass a chat ` +
23371
- `template or the name of the template you wish to use to the 'chat_template' argument. Available ` +
23372
- `template names are ${Object.keys(template_dict).sort()}.`
23373
- )
23374
- }
23375
- } else {
23376
- // These are the cases when the model has a single template
23377
- // priority: `chat_template` argument > `tokenizer.chat_template`
23378
- if (this.chat_template) {
23379
- chat_template = this.chat_template;
23380
- } else {
23381
- throw Error(
23382
- "Cannot use apply_chat_template() because tokenizer.chat_template is not set and no template " +
23383
- "argument was passed! For information about writing templates and setting the " +
23384
- "tokenizer.chat_template attribute, please see the documentation at " +
23385
- "https://huggingface.co/docs/transformers/main/en/chat_templating"
23386
- )
23387
- }
23388
- }
23389
23607
  if (typeof chat_template !== 'string') {
23390
23608
  throw Error(`chat_template must be a string, but got ${typeof chat_template}`);
23391
23609
  }
@@ -27793,11 +28011,20 @@ function softmax(arr) {
27793
28011
  * @returns {T} The resulting log_softmax array.
27794
28012
  */
27795
28013
  function log_softmax(arr) {
27796
- // Compute the softmax values
27797
- const softmaxArr = softmax(arr);
28014
+ // Compute the maximum value in the array
28015
+ const maxVal = max(arr)[0];
28016
+
28017
+ // Compute the sum of the exponentials
28018
+ let sumExps = 0;
28019
+ for(let i = 0; i < arr.length; ++i) {
28020
+ sumExps += Math.exp(arr[i] - maxVal);
28021
+ }
28022
+
28023
+ // Compute the log of the sum
28024
+ const logSum = Math.log(sumExps);
27798
28025
 
27799
- // Apply log formula to each element
27800
- const logSoftmaxArr = softmaxArr.map(x => Math.log(x));
28026
+ // Compute the softmax values
28027
+ const logSoftmaxArr = arr.map(x => x - maxVal - logSum);
27801
28028
 
27802
28029
  return /** @type {T} */(logSoftmaxArr);
27803
28030
  }
@@ -27852,7 +28079,7 @@ function magnitude(arr) {
27852
28079
  /**
27853
28080
  * Returns the value and index of the minimum element in an array.
27854
28081
  * @param {number[]|TypedArray} arr array of numbers.
27855
- * @returns {number[]} the value and index of the minimum element, of the form: [valueOfMin, indexOfMin]
28082
+ * @returns {[number, number]} the value and index of the minimum element, of the form: [valueOfMin, indexOfMin]
27856
28083
  * @throws {Error} If array is empty.
27857
28084
  */
27858
28085
  function min(arr) {
@@ -30247,6 +30474,7 @@ __webpack_require__.r(__webpack_exports__);
30247
30474
  /* harmony export */ AutoModelForTextToSpectrogram: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.AutoModelForTextToSpectrogram),
30248
30475
  /* harmony export */ AutoModelForTextToWaveform: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.AutoModelForTextToWaveform),
30249
30476
  /* harmony export */ AutoModelForTokenClassification: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.AutoModelForTokenClassification),
30477
+ /* harmony export */ AutoModelForUniversalSegmentation: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.AutoModelForUniversalSegmentation),
30250
30478
  /* harmony export */ AutoModelForVision2Seq: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.AutoModelForVision2Seq),
30251
30479
  /* harmony export */ AutoModelForXVector: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.AutoModelForXVector),
30252
30480
  /* harmony export */ AutoModelForZeroShotObjectDetection: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.AutoModelForZeroShotObjectDetection),
@@ -30291,8 +30519,10 @@ __webpack_require__.r(__webpack_exports__);
30291
30519
  /* harmony export */ CLIPSegForImageSegmentation: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.CLIPSegForImageSegmentation),
30292
30520
  /* harmony export */ CLIPSegModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.CLIPSegModel),
30293
30521
  /* harmony export */ CLIPSegPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.CLIPSegPreTrainedModel),
30522
+ /* harmony export */ CLIPTextModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.CLIPTextModel),
30294
30523
  /* harmony export */ CLIPTextModelWithProjection: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.CLIPTextModelWithProjection),
30295
30524
  /* harmony export */ CLIPTokenizer: () => (/* reexport safe */ _tokenizers_js__WEBPACK_IMPORTED_MODULE_3__.CLIPTokenizer),
30525
+ /* harmony export */ CLIPVisionModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.CLIPVisionModel),
30296
30526
  /* harmony export */ CLIPVisionModelWithProjection: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.CLIPVisionModelWithProjection),
30297
30527
  /* harmony export */ CamembertForMaskedLM: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.CamembertForMaskedLM),
30298
30528
  /* harmony export */ CamembertForQuestionAnswering: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.CamembertForQuestionAnswering),
@@ -30354,6 +30584,8 @@ __webpack_require__.r(__webpack_exports__);
30354
30584
  /* harmony export */ DebertaV2Model: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.DebertaV2Model),
30355
30585
  /* harmony export */ DebertaV2PreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.DebertaV2PreTrainedModel),
30356
30586
  /* harmony export */ DebertaV2Tokenizer: () => (/* reexport safe */ _tokenizers_js__WEBPACK_IMPORTED_MODULE_3__.DebertaV2Tokenizer),
30587
+ /* harmony export */ DecisionTransformerModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.DecisionTransformerModel),
30588
+ /* harmony export */ DecisionTransformerPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.DecisionTransformerPreTrainedModel),
30357
30589
  /* harmony export */ DeiTFeatureExtractor: () => (/* reexport safe */ _processors_js__WEBPACK_IMPORTED_MODULE_4__.DeiTFeatureExtractor),
30358
30590
  /* harmony export */ DeiTForImageClassification: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.DeiTForImageClassification),
30359
30591
  /* harmony export */ DeiTModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.DeiTModel),
@@ -30443,7 +30675,12 @@ __webpack_require__.r(__webpack_exports__);
30443
30675
  /* harmony export */ GemmaPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.GemmaPreTrainedModel),
30444
30676
  /* harmony export */ GemmaTokenizer: () => (/* reexport safe */ _tokenizers_js__WEBPACK_IMPORTED_MODULE_3__.GemmaTokenizer),
30445
30677
  /* harmony export */ Grok1Tokenizer: () => (/* reexport safe */ _tokenizers_js__WEBPACK_IMPORTED_MODULE_3__.Grok1Tokenizer),
30678
+ /* harmony export */ GroupViTModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.GroupViTModel),
30679
+ /* harmony export */ GroupViTPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.GroupViTPreTrainedModel),
30446
30680
  /* harmony export */ HerbertTokenizer: () => (/* reexport safe */ _tokenizers_js__WEBPACK_IMPORTED_MODULE_3__.HerbertTokenizer),
30681
+ /* harmony export */ HieraForImageClassification: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.HieraForImageClassification),
30682
+ /* harmony export */ HieraModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.HieraModel),
30683
+ /* harmony export */ HieraPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.HieraPreTrainedModel),
30447
30684
  /* harmony export */ HubertForCTC: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.HubertForCTC),
30448
30685
  /* harmony export */ HubertForSequenceClassification: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.HubertForSequenceClassification),
30449
30686
  /* harmony export */ HubertModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.HubertModel),
@@ -30493,6 +30730,10 @@ __webpack_require__.r(__webpack_exports__);
30493
30730
  /* harmony export */ MarianModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.MarianModel),
30494
30731
  /* harmony export */ MarianPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.MarianPreTrainedModel),
30495
30732
  /* harmony export */ MarianTokenizer: () => (/* reexport safe */ _tokenizers_js__WEBPACK_IMPORTED_MODULE_3__.MarianTokenizer),
30733
+ /* harmony export */ MaskFormerFeatureExtractor: () => (/* reexport safe */ _processors_js__WEBPACK_IMPORTED_MODULE_4__.MaskFormerFeatureExtractor),
30734
+ /* harmony export */ MaskFormerForInstanceSegmentation: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.MaskFormerForInstanceSegmentation),
30735
+ /* harmony export */ MaskFormerModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.MaskFormerModel),
30736
+ /* harmony export */ MaskFormerPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.MaskFormerPreTrainedModel),
30496
30737
  /* harmony export */ MaskedLMOutput: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.MaskedLMOutput),
30497
30738
  /* harmony export */ MaxLengthCriteria: () => (/* reexport safe */ _generation_stopping_criteria_js__WEBPACK_IMPORTED_MODULE_11__.MaxLengthCriteria),
30498
30739
  /* harmony export */ MistralForCausalLM: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.MistralForCausalLM),
@@ -30570,6 +30811,10 @@ __webpack_require__.r(__webpack_exports__);
30570
30811
  /* harmony export */ PretrainedConfig: () => (/* reexport safe */ _configs_js__WEBPACK_IMPORTED_MODULE_5__.PretrainedConfig),
30571
30812
  /* harmony export */ PretrainedMixin: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.PretrainedMixin),
30572
30813
  /* harmony export */ Processor: () => (/* reexport safe */ _processors_js__WEBPACK_IMPORTED_MODULE_4__.Processor),
30814
+ /* harmony export */ PvtForImageClassification: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.PvtForImageClassification),
30815
+ /* harmony export */ PvtImageProcessor: () => (/* reexport safe */ _processors_js__WEBPACK_IMPORTED_MODULE_4__.PvtImageProcessor),
30816
+ /* harmony export */ PvtModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.PvtModel),
30817
+ /* harmony export */ PvtPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.PvtPreTrainedModel),
30573
30818
  /* harmony export */ PyAnnoteFeatureExtractor: () => (/* reexport safe */ _processors_js__WEBPACK_IMPORTED_MODULE_4__.PyAnnoteFeatureExtractor),
30574
30819
  /* harmony export */ PyAnnoteForAudioFrameClassification: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.PyAnnoteForAudioFrameClassification),
30575
30820
  /* harmony export */ PyAnnoteModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.PyAnnoteModel),
@@ -30690,6 +30935,11 @@ __webpack_require__.r(__webpack_exports__);
30690
30935
  /* harmony export */ ViTFeatureExtractor: () => (/* reexport safe */ _processors_js__WEBPACK_IMPORTED_MODULE_4__.ViTFeatureExtractor),
30691
30936
  /* harmony export */ ViTForImageClassification: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.ViTForImageClassification),
30692
30937
  /* harmony export */ ViTImageProcessor: () => (/* reexport safe */ _processors_js__WEBPACK_IMPORTED_MODULE_4__.ViTImageProcessor),
30938
+ /* harmony export */ ViTMAEModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.ViTMAEModel),
30939
+ /* harmony export */ ViTMAEPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.ViTMAEPreTrainedModel),
30940
+ /* harmony export */ ViTMSNForImageClassification: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.ViTMSNForImageClassification),
30941
+ /* harmony export */ ViTMSNModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.ViTMSNModel),
30942
+ /* harmony export */ ViTMSNPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.ViTMSNPreTrainedModel),
30693
30943
  /* harmony export */ ViTModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.ViTModel),
30694
30944
  /* harmony export */ ViTPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.ViTPreTrainedModel),
30695
30945
  /* harmony export */ VisionEncoderDecoderModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.VisionEncoderDecoderModel),
@@ -30872,6 +31122,7 @@ var __webpack_exports__AutoModelForSpeechSeq2Seq = __webpack_exports__.AutoModel
30872
31122
  var __webpack_exports__AutoModelForTextToSpectrogram = __webpack_exports__.AutoModelForTextToSpectrogram;
30873
31123
  var __webpack_exports__AutoModelForTextToWaveform = __webpack_exports__.AutoModelForTextToWaveform;
30874
31124
  var __webpack_exports__AutoModelForTokenClassification = __webpack_exports__.AutoModelForTokenClassification;
31125
+ var __webpack_exports__AutoModelForUniversalSegmentation = __webpack_exports__.AutoModelForUniversalSegmentation;
30875
31126
  var __webpack_exports__AutoModelForVision2Seq = __webpack_exports__.AutoModelForVision2Seq;
30876
31127
  var __webpack_exports__AutoModelForXVector = __webpack_exports__.AutoModelForXVector;
30877
31128
  var __webpack_exports__AutoModelForZeroShotObjectDetection = __webpack_exports__.AutoModelForZeroShotObjectDetection;
@@ -30916,8 +31167,10 @@ var __webpack_exports__CLIPPreTrainedModel = __webpack_exports__.CLIPPreTrainedM
30916
31167
  var __webpack_exports__CLIPSegForImageSegmentation = __webpack_exports__.CLIPSegForImageSegmentation;
30917
31168
  var __webpack_exports__CLIPSegModel = __webpack_exports__.CLIPSegModel;
30918
31169
  var __webpack_exports__CLIPSegPreTrainedModel = __webpack_exports__.CLIPSegPreTrainedModel;
31170
+ var __webpack_exports__CLIPTextModel = __webpack_exports__.CLIPTextModel;
30919
31171
  var __webpack_exports__CLIPTextModelWithProjection = __webpack_exports__.CLIPTextModelWithProjection;
30920
31172
  var __webpack_exports__CLIPTokenizer = __webpack_exports__.CLIPTokenizer;
31173
+ var __webpack_exports__CLIPVisionModel = __webpack_exports__.CLIPVisionModel;
30921
31174
  var __webpack_exports__CLIPVisionModelWithProjection = __webpack_exports__.CLIPVisionModelWithProjection;
30922
31175
  var __webpack_exports__CamembertForMaskedLM = __webpack_exports__.CamembertForMaskedLM;
30923
31176
  var __webpack_exports__CamembertForQuestionAnswering = __webpack_exports__.CamembertForQuestionAnswering;
@@ -30979,6 +31232,8 @@ var __webpack_exports__DebertaV2ForTokenClassification = __webpack_exports__.Deb
30979
31232
  var __webpack_exports__DebertaV2Model = __webpack_exports__.DebertaV2Model;
30980
31233
  var __webpack_exports__DebertaV2PreTrainedModel = __webpack_exports__.DebertaV2PreTrainedModel;
30981
31234
  var __webpack_exports__DebertaV2Tokenizer = __webpack_exports__.DebertaV2Tokenizer;
31235
+ var __webpack_exports__DecisionTransformerModel = __webpack_exports__.DecisionTransformerModel;
31236
+ var __webpack_exports__DecisionTransformerPreTrainedModel = __webpack_exports__.DecisionTransformerPreTrainedModel;
30982
31237
  var __webpack_exports__DeiTFeatureExtractor = __webpack_exports__.DeiTFeatureExtractor;
30983
31238
  var __webpack_exports__DeiTForImageClassification = __webpack_exports__.DeiTForImageClassification;
30984
31239
  var __webpack_exports__DeiTModel = __webpack_exports__.DeiTModel;
@@ -31068,7 +31323,12 @@ var __webpack_exports__GemmaModel = __webpack_exports__.GemmaModel;
31068
31323
  var __webpack_exports__GemmaPreTrainedModel = __webpack_exports__.GemmaPreTrainedModel;
31069
31324
  var __webpack_exports__GemmaTokenizer = __webpack_exports__.GemmaTokenizer;
31070
31325
  var __webpack_exports__Grok1Tokenizer = __webpack_exports__.Grok1Tokenizer;
31326
+ var __webpack_exports__GroupViTModel = __webpack_exports__.GroupViTModel;
31327
+ var __webpack_exports__GroupViTPreTrainedModel = __webpack_exports__.GroupViTPreTrainedModel;
31071
31328
  var __webpack_exports__HerbertTokenizer = __webpack_exports__.HerbertTokenizer;
31329
+ var __webpack_exports__HieraForImageClassification = __webpack_exports__.HieraForImageClassification;
31330
+ var __webpack_exports__HieraModel = __webpack_exports__.HieraModel;
31331
+ var __webpack_exports__HieraPreTrainedModel = __webpack_exports__.HieraPreTrainedModel;
31072
31332
  var __webpack_exports__HubertForCTC = __webpack_exports__.HubertForCTC;
31073
31333
  var __webpack_exports__HubertForSequenceClassification = __webpack_exports__.HubertForSequenceClassification;
31074
31334
  var __webpack_exports__HubertModel = __webpack_exports__.HubertModel;
@@ -31118,6 +31378,10 @@ var __webpack_exports__MarianMTModel = __webpack_exports__.MarianMTModel;
31118
31378
  var __webpack_exports__MarianModel = __webpack_exports__.MarianModel;
31119
31379
  var __webpack_exports__MarianPreTrainedModel = __webpack_exports__.MarianPreTrainedModel;
31120
31380
  var __webpack_exports__MarianTokenizer = __webpack_exports__.MarianTokenizer;
31381
+ var __webpack_exports__MaskFormerFeatureExtractor = __webpack_exports__.MaskFormerFeatureExtractor;
31382
+ var __webpack_exports__MaskFormerForInstanceSegmentation = __webpack_exports__.MaskFormerForInstanceSegmentation;
31383
+ var __webpack_exports__MaskFormerModel = __webpack_exports__.MaskFormerModel;
31384
+ var __webpack_exports__MaskFormerPreTrainedModel = __webpack_exports__.MaskFormerPreTrainedModel;
31121
31385
  var __webpack_exports__MaskedLMOutput = __webpack_exports__.MaskedLMOutput;
31122
31386
  var __webpack_exports__MaxLengthCriteria = __webpack_exports__.MaxLengthCriteria;
31123
31387
  var __webpack_exports__MistralForCausalLM = __webpack_exports__.MistralForCausalLM;
@@ -31195,6 +31459,10 @@ var __webpack_exports__PreTrainedTokenizer = __webpack_exports__.PreTrainedToken
31195
31459
  var __webpack_exports__PretrainedConfig = __webpack_exports__.PretrainedConfig;
31196
31460
  var __webpack_exports__PretrainedMixin = __webpack_exports__.PretrainedMixin;
31197
31461
  var __webpack_exports__Processor = __webpack_exports__.Processor;
31462
+ var __webpack_exports__PvtForImageClassification = __webpack_exports__.PvtForImageClassification;
31463
+ var __webpack_exports__PvtImageProcessor = __webpack_exports__.PvtImageProcessor;
31464
+ var __webpack_exports__PvtModel = __webpack_exports__.PvtModel;
31465
+ var __webpack_exports__PvtPreTrainedModel = __webpack_exports__.PvtPreTrainedModel;
31198
31466
  var __webpack_exports__PyAnnoteFeatureExtractor = __webpack_exports__.PyAnnoteFeatureExtractor;
31199
31467
  var __webpack_exports__PyAnnoteForAudioFrameClassification = __webpack_exports__.PyAnnoteForAudioFrameClassification;
31200
31468
  var __webpack_exports__PyAnnoteModel = __webpack_exports__.PyAnnoteModel;
@@ -31315,6 +31583,11 @@ var __webpack_exports__UniSpeechSatPreTrainedModel = __webpack_exports__.UniSpee
31315
31583
  var __webpack_exports__ViTFeatureExtractor = __webpack_exports__.ViTFeatureExtractor;
31316
31584
  var __webpack_exports__ViTForImageClassification = __webpack_exports__.ViTForImageClassification;
31317
31585
  var __webpack_exports__ViTImageProcessor = __webpack_exports__.ViTImageProcessor;
31586
+ var __webpack_exports__ViTMAEModel = __webpack_exports__.ViTMAEModel;
31587
+ var __webpack_exports__ViTMAEPreTrainedModel = __webpack_exports__.ViTMAEPreTrainedModel;
31588
+ var __webpack_exports__ViTMSNForImageClassification = __webpack_exports__.ViTMSNForImageClassification;
31589
+ var __webpack_exports__ViTMSNModel = __webpack_exports__.ViTMSNModel;
31590
+ var __webpack_exports__ViTMSNPreTrainedModel = __webpack_exports__.ViTMSNPreTrainedModel;
31318
31591
  var __webpack_exports__ViTModel = __webpack_exports__.ViTModel;
31319
31592
  var __webpack_exports__ViTPreTrainedModel = __webpack_exports__.ViTPreTrainedModel;
31320
31593
  var __webpack_exports__VisionEncoderDecoderModel = __webpack_exports__.VisionEncoderDecoderModel;
@@ -31419,6 +31692,6 @@ var __webpack_exports__topk = __webpack_exports__.topk;
31419
31692
  var __webpack_exports__window_function = __webpack_exports__.window_function;
31420
31693
  var __webpack_exports__zeros = __webpack_exports__.zeros;
31421
31694
  var __webpack_exports__zeros_like = __webpack_exports__.zeros_like;
31422
- export { __webpack_exports__ASTFeatureExtractor as ASTFeatureExtractor, __webpack_exports__ASTForAudioClassification as ASTForAudioClassification, __webpack_exports__ASTModel as ASTModel, __webpack_exports__ASTPreTrainedModel as ASTPreTrainedModel, __webpack_exports__AlbertForMaskedLM as AlbertForMaskedLM, __webpack_exports__AlbertForQuestionAnswering as AlbertForQuestionAnswering, __webpack_exports__AlbertForSequenceClassification as AlbertForSequenceClassification, __webpack_exports__AlbertModel as AlbertModel, __webpack_exports__AlbertPreTrainedModel as AlbertPreTrainedModel, __webpack_exports__AlbertTokenizer as AlbertTokenizer, __webpack_exports__AudioClassificationPipeline as AudioClassificationPipeline, __webpack_exports__AutoConfig as AutoConfig, __webpack_exports__AutoModel as AutoModel, __webpack_exports__AutoModelForAudioClassification as AutoModelForAudioClassification, __webpack_exports__AutoModelForAudioFrameClassification as AutoModelForAudioFrameClassification, __webpack_exports__AutoModelForCTC as AutoModelForCTC, __webpack_exports__AutoModelForCausalLM as AutoModelForCausalLM, __webpack_exports__AutoModelForDepthEstimation as AutoModelForDepthEstimation, __webpack_exports__AutoModelForDocumentQuestionAnswering as AutoModelForDocumentQuestionAnswering, __webpack_exports__AutoModelForImageClassification as AutoModelForImageClassification, __webpack_exports__AutoModelForImageFeatureExtraction as AutoModelForImageFeatureExtraction, __webpack_exports__AutoModelForImageMatting as AutoModelForImageMatting, __webpack_exports__AutoModelForImageSegmentation as AutoModelForImageSegmentation, __webpack_exports__AutoModelForImageToImage as AutoModelForImageToImage, __webpack_exports__AutoModelForMaskGeneration as AutoModelForMaskGeneration, __webpack_exports__AutoModelForMaskedLM as AutoModelForMaskedLM, __webpack_exports__AutoModelForNormalEstimation as AutoModelForNormalEstimation, __webpack_exports__AutoModelForObjectDetection as AutoModelForObjectDetection, __webpack_exports__AutoModelForQuestionAnswering as AutoModelForQuestionAnswering, __webpack_exports__AutoModelForSemanticSegmentation as AutoModelForSemanticSegmentation, __webpack_exports__AutoModelForSeq2SeqLM as AutoModelForSeq2SeqLM, __webpack_exports__AutoModelForSequenceClassification as AutoModelForSequenceClassification, __webpack_exports__AutoModelForSpeechSeq2Seq as AutoModelForSpeechSeq2Seq, __webpack_exports__AutoModelForTextToSpectrogram as AutoModelForTextToSpectrogram, __webpack_exports__AutoModelForTextToWaveform as AutoModelForTextToWaveform, __webpack_exports__AutoModelForTokenClassification as AutoModelForTokenClassification, __webpack_exports__AutoModelForVision2Seq as AutoModelForVision2Seq, __webpack_exports__AutoModelForXVector as AutoModelForXVector, __webpack_exports__AutoModelForZeroShotObjectDetection as AutoModelForZeroShotObjectDetection, __webpack_exports__AutoProcessor as AutoProcessor, __webpack_exports__AutoTokenizer as AutoTokenizer, __webpack_exports__AutomaticSpeechRecognitionPipeline as AutomaticSpeechRecognitionPipeline, __webpack_exports__BartForConditionalGeneration as BartForConditionalGeneration, __webpack_exports__BartForSequenceClassification as BartForSequenceClassification, __webpack_exports__BartModel as BartModel, __webpack_exports__BartPretrainedModel as BartPretrainedModel, __webpack_exports__BartTokenizer as BartTokenizer, __webpack_exports__BaseModelOutput as BaseModelOutput, __webpack_exports__BaseStreamer as BaseStreamer, __webpack_exports__BeitFeatureExtractor as BeitFeatureExtractor, __webpack_exports__BeitForImageClassification as BeitForImageClassification, __webpack_exports__BeitModel as BeitModel, __webpack_exports__BeitPreTrainedModel as BeitPreTrainedModel, __webpack_exports__BertForMaskedLM as BertForMaskedLM, __webpack_exports__BertForQuestionAnswering as BertForQuestionAnswering, __webpack_exports__BertForSequenceClassification as BertForSequenceClassification, __webpack_exports__BertForTokenClassification as BertForTokenClassification, __webpack_exports__BertModel as BertModel, __webpack_exports__BertPreTrainedModel as BertPreTrainedModel, __webpack_exports__BertTokenizer as BertTokenizer, __webpack_exports__BitImageProcessor as BitImageProcessor, __webpack_exports__BlenderbotForConditionalGeneration as BlenderbotForConditionalGeneration, __webpack_exports__BlenderbotModel as BlenderbotModel, __webpack_exports__BlenderbotPreTrainedModel as BlenderbotPreTrainedModel, __webpack_exports__BlenderbotSmallForConditionalGeneration as BlenderbotSmallForConditionalGeneration, __webpack_exports__BlenderbotSmallModel as BlenderbotSmallModel, __webpack_exports__BlenderbotSmallPreTrainedModel as BlenderbotSmallPreTrainedModel, __webpack_exports__BlenderbotSmallTokenizer as BlenderbotSmallTokenizer, __webpack_exports__BlenderbotTokenizer as BlenderbotTokenizer, __webpack_exports__BloomForCausalLM as BloomForCausalLM, __webpack_exports__BloomModel as BloomModel, __webpack_exports__BloomPreTrainedModel as BloomPreTrainedModel, __webpack_exports__BloomTokenizer as BloomTokenizer, __webpack_exports__CLIPFeatureExtractor as CLIPFeatureExtractor, __webpack_exports__CLIPImageProcessor as CLIPImageProcessor, __webpack_exports__CLIPModel as CLIPModel, __webpack_exports__CLIPPreTrainedModel as CLIPPreTrainedModel, __webpack_exports__CLIPSegForImageSegmentation as CLIPSegForImageSegmentation, __webpack_exports__CLIPSegModel as CLIPSegModel, __webpack_exports__CLIPSegPreTrainedModel as CLIPSegPreTrainedModel, __webpack_exports__CLIPTextModelWithProjection as CLIPTextModelWithProjection, __webpack_exports__CLIPTokenizer as CLIPTokenizer, __webpack_exports__CLIPVisionModelWithProjection as CLIPVisionModelWithProjection, __webpack_exports__CamembertForMaskedLM as CamembertForMaskedLM, __webpack_exports__CamembertForQuestionAnswering as CamembertForQuestionAnswering, __webpack_exports__CamembertForSequenceClassification as CamembertForSequenceClassification, __webpack_exports__CamembertForTokenClassification as CamembertForTokenClassification, __webpack_exports__CamembertModel as CamembertModel, __webpack_exports__CamembertPreTrainedModel as CamembertPreTrainedModel, __webpack_exports__CamembertTokenizer as CamembertTokenizer, __webpack_exports__CausalLMOutput as CausalLMOutput, __webpack_exports__CausalLMOutputWithPast as CausalLMOutputWithPast, __webpack_exports__ChineseCLIPFeatureExtractor as ChineseCLIPFeatureExtractor, __webpack_exports__ChineseCLIPModel as ChineseCLIPModel, __webpack_exports__ChineseCLIPPreTrainedModel as ChineseCLIPPreTrainedModel, __webpack_exports__ClapAudioModelWithProjection as ClapAudioModelWithProjection, __webpack_exports__ClapFeatureExtractor as ClapFeatureExtractor, __webpack_exports__ClapModel as ClapModel, __webpack_exports__ClapPreTrainedModel as ClapPreTrainedModel, __webpack_exports__ClapTextModelWithProjection as ClapTextModelWithProjection, __webpack_exports__CodeGenForCausalLM as CodeGenForCausalLM, __webpack_exports__CodeGenModel as CodeGenModel, __webpack_exports__CodeGenPreTrainedModel as CodeGenPreTrainedModel, __webpack_exports__CodeGenTokenizer as CodeGenTokenizer, __webpack_exports__CodeLlamaTokenizer as CodeLlamaTokenizer, __webpack_exports__CohereForCausalLM as CohereForCausalLM, __webpack_exports__CohereModel as CohereModel, __webpack_exports__CoherePreTrainedModel as CoherePreTrainedModel, __webpack_exports__CohereTokenizer as CohereTokenizer, __webpack_exports__ConvBertForMaskedLM as ConvBertForMaskedLM, __webpack_exports__ConvBertForQuestionAnswering as ConvBertForQuestionAnswering, __webpack_exports__ConvBertForSequenceClassification as ConvBertForSequenceClassification, __webpack_exports__ConvBertForTokenClassification as ConvBertForTokenClassification, __webpack_exports__ConvBertModel as ConvBertModel, __webpack_exports__ConvBertPreTrainedModel as ConvBertPreTrainedModel, __webpack_exports__ConvBertTokenizer as ConvBertTokenizer, __webpack_exports__ConvNextFeatureExtractor as ConvNextFeatureExtractor, __webpack_exports__ConvNextForImageClassification as ConvNextForImageClassification, __webpack_exports__ConvNextImageProcessor as ConvNextImageProcessor, __webpack_exports__ConvNextModel as ConvNextModel, __webpack_exports__ConvNextPreTrainedModel as ConvNextPreTrainedModel, __webpack_exports__ConvNextV2ForImageClassification as ConvNextV2ForImageClassification, __webpack_exports__ConvNextV2Model as ConvNextV2Model, __webpack_exports__ConvNextV2PreTrainedModel as ConvNextV2PreTrainedModel, __webpack_exports__DPTFeatureExtractor as DPTFeatureExtractor, __webpack_exports__DPTForDepthEstimation as DPTForDepthEstimation, __webpack_exports__DPTImageProcessor as DPTImageProcessor, __webpack_exports__DPTModel as DPTModel, __webpack_exports__DPTPreTrainedModel as DPTPreTrainedModel, __webpack_exports__DebertaForMaskedLM as DebertaForMaskedLM, __webpack_exports__DebertaForQuestionAnswering as DebertaForQuestionAnswering, __webpack_exports__DebertaForSequenceClassification as DebertaForSequenceClassification, __webpack_exports__DebertaForTokenClassification as DebertaForTokenClassification, __webpack_exports__DebertaModel as DebertaModel, __webpack_exports__DebertaPreTrainedModel as DebertaPreTrainedModel, __webpack_exports__DebertaTokenizer as DebertaTokenizer, __webpack_exports__DebertaV2ForMaskedLM as DebertaV2ForMaskedLM, __webpack_exports__DebertaV2ForQuestionAnswering as DebertaV2ForQuestionAnswering, __webpack_exports__DebertaV2ForSequenceClassification as DebertaV2ForSequenceClassification, __webpack_exports__DebertaV2ForTokenClassification as DebertaV2ForTokenClassification, __webpack_exports__DebertaV2Model as DebertaV2Model, __webpack_exports__DebertaV2PreTrainedModel as DebertaV2PreTrainedModel, __webpack_exports__DebertaV2Tokenizer as DebertaV2Tokenizer, __webpack_exports__DeiTFeatureExtractor as DeiTFeatureExtractor, __webpack_exports__DeiTForImageClassification as DeiTForImageClassification, __webpack_exports__DeiTModel as DeiTModel, __webpack_exports__DeiTPreTrainedModel as DeiTPreTrainedModel, __webpack_exports__DepthAnythingForDepthEstimation as DepthAnythingForDepthEstimation, __webpack_exports__DepthAnythingPreTrainedModel as DepthAnythingPreTrainedModel, __webpack_exports__DepthEstimationPipeline as DepthEstimationPipeline, __webpack_exports__DetrFeatureExtractor as DetrFeatureExtractor, __webpack_exports__DetrForObjectDetection as DetrForObjectDetection, __webpack_exports__DetrForSegmentation as DetrForSegmentation, __webpack_exports__DetrModel as DetrModel, __webpack_exports__DetrObjectDetectionOutput as DetrObjectDetectionOutput, __webpack_exports__DetrPreTrainedModel as DetrPreTrainedModel, __webpack_exports__DetrSegmentationOutput as DetrSegmentationOutput, __webpack_exports__Dinov2ForImageClassification as Dinov2ForImageClassification, __webpack_exports__Dinov2Model as Dinov2Model, __webpack_exports__Dinov2PreTrainedModel as Dinov2PreTrainedModel, __webpack_exports__DistilBertForMaskedLM as DistilBertForMaskedLM, __webpack_exports__DistilBertForQuestionAnswering as DistilBertForQuestionAnswering, __webpack_exports__DistilBertForSequenceClassification as DistilBertForSequenceClassification, __webpack_exports__DistilBertForTokenClassification as DistilBertForTokenClassification, __webpack_exports__DistilBertModel as DistilBertModel, __webpack_exports__DistilBertPreTrainedModel as DistilBertPreTrainedModel, __webpack_exports__DistilBertTokenizer as DistilBertTokenizer, __webpack_exports__DocumentQuestionAnsweringPipeline as DocumentQuestionAnsweringPipeline, __webpack_exports__DonutFeatureExtractor as DonutFeatureExtractor, __webpack_exports__DonutSwinModel as DonutSwinModel, __webpack_exports__DonutSwinPreTrainedModel as DonutSwinPreTrainedModel, __webpack_exports__EfficientNetForImageClassification as EfficientNetForImageClassification, __webpack_exports__EfficientNetImageProcessor as EfficientNetImageProcessor, __webpack_exports__EfficientNetModel as EfficientNetModel, __webpack_exports__EfficientNetPreTrainedModel as EfficientNetPreTrainedModel, __webpack_exports__ElectraForMaskedLM as ElectraForMaskedLM, __webpack_exports__ElectraForQuestionAnswering as ElectraForQuestionAnswering, __webpack_exports__ElectraForSequenceClassification as ElectraForSequenceClassification, __webpack_exports__ElectraForTokenClassification as ElectraForTokenClassification, __webpack_exports__ElectraModel as ElectraModel, __webpack_exports__ElectraPreTrainedModel as ElectraPreTrainedModel, __webpack_exports__ElectraTokenizer as ElectraTokenizer, __webpack_exports__EosTokenCriteria as EosTokenCriteria, __webpack_exports__EsmForMaskedLM as EsmForMaskedLM, __webpack_exports__EsmForSequenceClassification as EsmForSequenceClassification, __webpack_exports__EsmForTokenClassification as EsmForTokenClassification, __webpack_exports__EsmModel as EsmModel, __webpack_exports__EsmPreTrainedModel as EsmPreTrainedModel, __webpack_exports__EsmTokenizer as EsmTokenizer, __webpack_exports__FFT as FFT, __webpack_exports__FalconForCausalLM as FalconForCausalLM, __webpack_exports__FalconModel as FalconModel, __webpack_exports__FalconPreTrainedModel as FalconPreTrainedModel, __webpack_exports__FalconTokenizer as FalconTokenizer, __webpack_exports__FastViTForImageClassification as FastViTForImageClassification, __webpack_exports__FastViTModel as FastViTModel, __webpack_exports__FastViTPreTrainedModel as FastViTPreTrainedModel, __webpack_exports__FeatureExtractionPipeline as FeatureExtractionPipeline, __webpack_exports__FeatureExtractor as FeatureExtractor, __webpack_exports__FillMaskPipeline as FillMaskPipeline, __webpack_exports__Florence2ForConditionalGeneration as Florence2ForConditionalGeneration, __webpack_exports__Florence2PreTrainedModel as Florence2PreTrainedModel, __webpack_exports__Florence2Processor as Florence2Processor, __webpack_exports__GLPNFeatureExtractor as GLPNFeatureExtractor, __webpack_exports__GLPNForDepthEstimation as GLPNForDepthEstimation, __webpack_exports__GLPNModel as GLPNModel, __webpack_exports__GLPNPreTrainedModel as GLPNPreTrainedModel, __webpack_exports__GPT2LMHeadModel as GPT2LMHeadModel, __webpack_exports__GPT2Model as GPT2Model, __webpack_exports__GPT2PreTrainedModel as GPT2PreTrainedModel, __webpack_exports__GPT2Tokenizer as GPT2Tokenizer, __webpack_exports__GPTBigCodeForCausalLM as GPTBigCodeForCausalLM, __webpack_exports__GPTBigCodeModel as GPTBigCodeModel, __webpack_exports__GPTBigCodePreTrainedModel as GPTBigCodePreTrainedModel, __webpack_exports__GPTJForCausalLM as GPTJForCausalLM, __webpack_exports__GPTJModel as GPTJModel, __webpack_exports__GPTJPreTrainedModel as GPTJPreTrainedModel, __webpack_exports__GPTNeoForCausalLM as GPTNeoForCausalLM, __webpack_exports__GPTNeoModel as GPTNeoModel, __webpack_exports__GPTNeoPreTrainedModel as GPTNeoPreTrainedModel, __webpack_exports__GPTNeoXForCausalLM as GPTNeoXForCausalLM, __webpack_exports__GPTNeoXModel as GPTNeoXModel, __webpack_exports__GPTNeoXPreTrainedModel as GPTNeoXPreTrainedModel, __webpack_exports__GPTNeoXTokenizer as GPTNeoXTokenizer, __webpack_exports__Gemma2ForCausalLM as Gemma2ForCausalLM, __webpack_exports__Gemma2Model as Gemma2Model, __webpack_exports__Gemma2PreTrainedModel as Gemma2PreTrainedModel, __webpack_exports__GemmaForCausalLM as GemmaForCausalLM, __webpack_exports__GemmaModel as GemmaModel, __webpack_exports__GemmaPreTrainedModel as GemmaPreTrainedModel, __webpack_exports__GemmaTokenizer as GemmaTokenizer, __webpack_exports__Grok1Tokenizer as Grok1Tokenizer, __webpack_exports__HerbertTokenizer as HerbertTokenizer, __webpack_exports__HubertForCTC as HubertForCTC, __webpack_exports__HubertForSequenceClassification as HubertForSequenceClassification, __webpack_exports__HubertModel as HubertModel, __webpack_exports__HubertPreTrainedModel as HubertPreTrainedModel, __webpack_exports__ImageClassificationPipeline as ImageClassificationPipeline, __webpack_exports__ImageFeatureExtractionPipeline as ImageFeatureExtractionPipeline, __webpack_exports__ImageFeatureExtractor as ImageFeatureExtractor, __webpack_exports__ImageMattingOutput as ImageMattingOutput, __webpack_exports__ImageSegmentationPipeline as ImageSegmentationPipeline, __webpack_exports__ImageToImagePipeline as ImageToImagePipeline, __webpack_exports__ImageToTextPipeline as ImageToTextPipeline, __webpack_exports__InterruptableStoppingCriteria as InterruptableStoppingCriteria, __webpack_exports__JAISLMHeadModel as JAISLMHeadModel, __webpack_exports__JAISModel as JAISModel, __webpack_exports__JAISPreTrainedModel as JAISPreTrainedModel, __webpack_exports__LlamaForCausalLM as LlamaForCausalLM, __webpack_exports__LlamaModel as LlamaModel, __webpack_exports__LlamaPreTrainedModel as LlamaPreTrainedModel, __webpack_exports__LlamaTokenizer as LlamaTokenizer, __webpack_exports__LlavaForConditionalGeneration as LlavaForConditionalGeneration, __webpack_exports__LlavaPreTrainedModel as LlavaPreTrainedModel, __webpack_exports__LongT5ForConditionalGeneration as LongT5ForConditionalGeneration, __webpack_exports__LongT5Model as LongT5Model, __webpack_exports__LongT5PreTrainedModel as LongT5PreTrainedModel, __webpack_exports__M2M100ForConditionalGeneration as M2M100ForConditionalGeneration, __webpack_exports__M2M100Model as M2M100Model, __webpack_exports__M2M100PreTrainedModel as M2M100PreTrainedModel, __webpack_exports__M2M100Tokenizer as M2M100Tokenizer, __webpack_exports__MBart50Tokenizer as MBart50Tokenizer, __webpack_exports__MBartForCausalLM as MBartForCausalLM, __webpack_exports__MBartForConditionalGeneration as MBartForConditionalGeneration, __webpack_exports__MBartForSequenceClassification as MBartForSequenceClassification, __webpack_exports__MBartModel as MBartModel, __webpack_exports__MBartPreTrainedModel as MBartPreTrainedModel, __webpack_exports__MBartTokenizer as MBartTokenizer, __webpack_exports__MPNetForMaskedLM as MPNetForMaskedLM, __webpack_exports__MPNetForQuestionAnswering as MPNetForQuestionAnswering, __webpack_exports__MPNetForSequenceClassification as MPNetForSequenceClassification, __webpack_exports__MPNetForTokenClassification as MPNetForTokenClassification, __webpack_exports__MPNetModel as MPNetModel, __webpack_exports__MPNetPreTrainedModel as MPNetPreTrainedModel, __webpack_exports__MPNetTokenizer as MPNetTokenizer, __webpack_exports__MT5ForConditionalGeneration as MT5ForConditionalGeneration, __webpack_exports__MT5Model as MT5Model, __webpack_exports__MT5PreTrainedModel as MT5PreTrainedModel, __webpack_exports__MarianMTModel as MarianMTModel, __webpack_exports__MarianModel as MarianModel, __webpack_exports__MarianPreTrainedModel as MarianPreTrainedModel, __webpack_exports__MarianTokenizer as MarianTokenizer, __webpack_exports__MaskedLMOutput as MaskedLMOutput, __webpack_exports__MaxLengthCriteria as MaxLengthCriteria, __webpack_exports__MistralForCausalLM as MistralForCausalLM, __webpack_exports__MistralModel as MistralModel, __webpack_exports__MistralPreTrainedModel as MistralPreTrainedModel, __webpack_exports__MobileBertForMaskedLM as MobileBertForMaskedLM, __webpack_exports__MobileBertForQuestionAnswering as MobileBertForQuestionAnswering, __webpack_exports__MobileBertForSequenceClassification as MobileBertForSequenceClassification, __webpack_exports__MobileBertModel as MobileBertModel, __webpack_exports__MobileBertPreTrainedModel as MobileBertPreTrainedModel, __webpack_exports__MobileBertTokenizer as MobileBertTokenizer, __webpack_exports__MobileNetV1FeatureExtractor as MobileNetV1FeatureExtractor, __webpack_exports__MobileNetV1ForImageClassification as MobileNetV1ForImageClassification, __webpack_exports__MobileNetV1Model as MobileNetV1Model, __webpack_exports__MobileNetV1PreTrainedModel as MobileNetV1PreTrainedModel, __webpack_exports__MobileNetV2FeatureExtractor as MobileNetV2FeatureExtractor, __webpack_exports__MobileNetV2ForImageClassification as MobileNetV2ForImageClassification, __webpack_exports__MobileNetV2Model as MobileNetV2Model, __webpack_exports__MobileNetV2PreTrainedModel as MobileNetV2PreTrainedModel, __webpack_exports__MobileNetV3FeatureExtractor as MobileNetV3FeatureExtractor, __webpack_exports__MobileNetV3ForImageClassification as MobileNetV3ForImageClassification, __webpack_exports__MobileNetV3Model as MobileNetV3Model, __webpack_exports__MobileNetV3PreTrainedModel as MobileNetV3PreTrainedModel, __webpack_exports__MobileNetV4FeatureExtractor as MobileNetV4FeatureExtractor, __webpack_exports__MobileNetV4ForImageClassification as MobileNetV4ForImageClassification, __webpack_exports__MobileNetV4Model as MobileNetV4Model, __webpack_exports__MobileNetV4PreTrainedModel as MobileNetV4PreTrainedModel, __webpack_exports__MobileViTFeatureExtractor as MobileViTFeatureExtractor, __webpack_exports__MobileViTForImageClassification as MobileViTForImageClassification, __webpack_exports__MobileViTImageProcessor as MobileViTImageProcessor, __webpack_exports__MobileViTModel as MobileViTModel, __webpack_exports__MobileViTPreTrainedModel as MobileViTPreTrainedModel, __webpack_exports__MobileViTV2ForImageClassification as MobileViTV2ForImageClassification, __webpack_exports__MobileViTV2Model as MobileViTV2Model, __webpack_exports__MobileViTV2PreTrainedModel as MobileViTV2PreTrainedModel, __webpack_exports__ModelOutput as ModelOutput, __webpack_exports__Moondream1ForConditionalGeneration as Moondream1ForConditionalGeneration, __webpack_exports__MptForCausalLM as MptForCausalLM, __webpack_exports__MptModel as MptModel, __webpack_exports__MptPreTrainedModel as MptPreTrainedModel, __webpack_exports__MusicgenForCausalLM as MusicgenForCausalLM, __webpack_exports__MusicgenForConditionalGeneration as MusicgenForConditionalGeneration, __webpack_exports__MusicgenModel as MusicgenModel, __webpack_exports__MusicgenPreTrainedModel as MusicgenPreTrainedModel, __webpack_exports__NllbTokenizer as NllbTokenizer, __webpack_exports__NomicBertModel as NomicBertModel, __webpack_exports__NomicBertPreTrainedModel as NomicBertPreTrainedModel, __webpack_exports__NougatImageProcessor as NougatImageProcessor, __webpack_exports__NougatTokenizer as NougatTokenizer, __webpack_exports__OPTForCausalLM as OPTForCausalLM, __webpack_exports__OPTModel as OPTModel, __webpack_exports__OPTPreTrainedModel as OPTPreTrainedModel, __webpack_exports__ObjectDetectionPipeline as ObjectDetectionPipeline, __webpack_exports__OpenELMForCausalLM as OpenELMForCausalLM, __webpack_exports__OpenELMModel as OpenELMModel, __webpack_exports__OpenELMPreTrainedModel as OpenELMPreTrainedModel, __webpack_exports__OwlViTFeatureExtractor as OwlViTFeatureExtractor, __webpack_exports__OwlViTForObjectDetection as OwlViTForObjectDetection, __webpack_exports__OwlViTModel as OwlViTModel, __webpack_exports__OwlViTPreTrainedModel as OwlViTPreTrainedModel, __webpack_exports__OwlViTProcessor as OwlViTProcessor, __webpack_exports__Owlv2ForObjectDetection as Owlv2ForObjectDetection, __webpack_exports__Owlv2ImageProcessor as Owlv2ImageProcessor, __webpack_exports__Owlv2Model as Owlv2Model, __webpack_exports__Owlv2PreTrainedModel as Owlv2PreTrainedModel, __webpack_exports__Phi3ForCausalLM as Phi3ForCausalLM, __webpack_exports__Phi3Model as Phi3Model, __webpack_exports__Phi3PreTrainedModel as Phi3PreTrainedModel, __webpack_exports__PhiForCausalLM as PhiForCausalLM, __webpack_exports__PhiModel as PhiModel, __webpack_exports__PhiPreTrainedModel as PhiPreTrainedModel, __webpack_exports__Pipeline as Pipeline, __webpack_exports__PreTrainedModel as PreTrainedModel, __webpack_exports__PreTrainedTokenizer as PreTrainedTokenizer, __webpack_exports__PretrainedConfig as PretrainedConfig, __webpack_exports__PretrainedMixin as PretrainedMixin, __webpack_exports__Processor as Processor, __webpack_exports__PyAnnoteFeatureExtractor as PyAnnoteFeatureExtractor, __webpack_exports__PyAnnoteForAudioFrameClassification as PyAnnoteForAudioFrameClassification, __webpack_exports__PyAnnoteModel as PyAnnoteModel, __webpack_exports__PyAnnotePreTrainedModel as PyAnnotePreTrainedModel, __webpack_exports__PyAnnoteProcessor as PyAnnoteProcessor, __webpack_exports__QuestionAnsweringModelOutput as QuestionAnsweringModelOutput, __webpack_exports__QuestionAnsweringPipeline as QuestionAnsweringPipeline, __webpack_exports__Qwen2ForCausalLM as Qwen2ForCausalLM, __webpack_exports__Qwen2Model as Qwen2Model, __webpack_exports__Qwen2PreTrainedModel as Qwen2PreTrainedModel, __webpack_exports__Qwen2Tokenizer as Qwen2Tokenizer, __webpack_exports__RTDetrForObjectDetection as RTDetrForObjectDetection, __webpack_exports__RTDetrImageProcessor as RTDetrImageProcessor, __webpack_exports__RTDetrModel as RTDetrModel, __webpack_exports__RTDetrObjectDetectionOutput as RTDetrObjectDetectionOutput, __webpack_exports__RTDetrPreTrainedModel as RTDetrPreTrainedModel, __webpack_exports__RawImage as RawImage, __webpack_exports__ResNetForImageClassification as ResNetForImageClassification, __webpack_exports__ResNetModel as ResNetModel, __webpack_exports__ResNetPreTrainedModel as ResNetPreTrainedModel, __webpack_exports__RoFormerForMaskedLM as RoFormerForMaskedLM, __webpack_exports__RoFormerForQuestionAnswering as RoFormerForQuestionAnswering, __webpack_exports__RoFormerForSequenceClassification as RoFormerForSequenceClassification, __webpack_exports__RoFormerForTokenClassification as RoFormerForTokenClassification, __webpack_exports__RoFormerModel as RoFormerModel, __webpack_exports__RoFormerPreTrainedModel as RoFormerPreTrainedModel, __webpack_exports__RoFormerTokenizer as RoFormerTokenizer, __webpack_exports__RobertaForMaskedLM as RobertaForMaskedLM, __webpack_exports__RobertaForQuestionAnswering as RobertaForQuestionAnswering, __webpack_exports__RobertaForSequenceClassification as RobertaForSequenceClassification, __webpack_exports__RobertaForTokenClassification as RobertaForTokenClassification, __webpack_exports__RobertaModel as RobertaModel, __webpack_exports__RobertaPreTrainedModel as RobertaPreTrainedModel, __webpack_exports__RobertaTokenizer as RobertaTokenizer, __webpack_exports__SamImageProcessor as SamImageProcessor, __webpack_exports__SamImageSegmentationOutput as SamImageSegmentationOutput, __webpack_exports__SamModel as SamModel, __webpack_exports__SamPreTrainedModel as SamPreTrainedModel, __webpack_exports__SamProcessor as SamProcessor, __webpack_exports__SapiensFeatureExtractor as SapiensFeatureExtractor, __webpack_exports__SapiensForDepthEstimation as SapiensForDepthEstimation, __webpack_exports__SapiensForNormalEstimation as SapiensForNormalEstimation, __webpack_exports__SapiensForSemanticSegmentation as SapiensForSemanticSegmentation, __webpack_exports__SapiensPreTrainedModel as SapiensPreTrainedModel, __webpack_exports__SeamlessM4TFeatureExtractor as SeamlessM4TFeatureExtractor, __webpack_exports__SegformerFeatureExtractor as SegformerFeatureExtractor, __webpack_exports__SegformerForImageClassification as SegformerForImageClassification, __webpack_exports__SegformerForSemanticSegmentation as SegformerForSemanticSegmentation, __webpack_exports__SegformerModel as SegformerModel, __webpack_exports__SegformerPreTrainedModel as SegformerPreTrainedModel, __webpack_exports__Seq2SeqLMOutput as Seq2SeqLMOutput, __webpack_exports__SequenceClassifierOutput as SequenceClassifierOutput, __webpack_exports__SiglipImageProcessor as SiglipImageProcessor, __webpack_exports__SiglipModel as SiglipModel, __webpack_exports__SiglipPreTrainedModel as SiglipPreTrainedModel, __webpack_exports__SiglipTextModel as SiglipTextModel, __webpack_exports__SiglipTokenizer as SiglipTokenizer, __webpack_exports__SiglipVisionModel as SiglipVisionModel, __webpack_exports__SpeechT5FeatureExtractor as SpeechT5FeatureExtractor, __webpack_exports__SpeechT5ForSpeechToText as SpeechT5ForSpeechToText, __webpack_exports__SpeechT5ForTextToSpeech as SpeechT5ForTextToSpeech, __webpack_exports__SpeechT5HifiGan as SpeechT5HifiGan, __webpack_exports__SpeechT5Model as SpeechT5Model, __webpack_exports__SpeechT5PreTrainedModel as SpeechT5PreTrainedModel, __webpack_exports__SpeechT5Processor as SpeechT5Processor, __webpack_exports__SpeechT5Tokenizer as SpeechT5Tokenizer, __webpack_exports__SqueezeBertForMaskedLM as SqueezeBertForMaskedLM, __webpack_exports__SqueezeBertForQuestionAnswering as SqueezeBertForQuestionAnswering, __webpack_exports__SqueezeBertForSequenceClassification as SqueezeBertForSequenceClassification, __webpack_exports__SqueezeBertModel as SqueezeBertModel, __webpack_exports__SqueezeBertPreTrainedModel as SqueezeBertPreTrainedModel, __webpack_exports__SqueezeBertTokenizer as SqueezeBertTokenizer, __webpack_exports__StableLmForCausalLM as StableLmForCausalLM, __webpack_exports__StableLmModel as StableLmModel, __webpack_exports__StableLmPreTrainedModel as StableLmPreTrainedModel, __webpack_exports__Starcoder2ForCausalLM as Starcoder2ForCausalLM, __webpack_exports__Starcoder2Model as Starcoder2Model, __webpack_exports__Starcoder2PreTrainedModel as Starcoder2PreTrainedModel, __webpack_exports__StoppingCriteria as StoppingCriteria, __webpack_exports__StoppingCriteriaList as StoppingCriteriaList, __webpack_exports__SummarizationPipeline as SummarizationPipeline, __webpack_exports__Swin2SRForImageSuperResolution as Swin2SRForImageSuperResolution, __webpack_exports__Swin2SRImageProcessor as Swin2SRImageProcessor, __webpack_exports__Swin2SRModel as Swin2SRModel, __webpack_exports__Swin2SRPreTrainedModel as Swin2SRPreTrainedModel, __webpack_exports__SwinForImageClassification as SwinForImageClassification, __webpack_exports__SwinModel as SwinModel, __webpack_exports__SwinPreTrainedModel as SwinPreTrainedModel, __webpack_exports__T5ForConditionalGeneration as T5ForConditionalGeneration, __webpack_exports__T5Model as T5Model, __webpack_exports__T5PreTrainedModel as T5PreTrainedModel, __webpack_exports__T5Tokenizer as T5Tokenizer, __webpack_exports__TableTransformerForObjectDetection as TableTransformerForObjectDetection, __webpack_exports__TableTransformerModel as TableTransformerModel, __webpack_exports__TableTransformerObjectDetectionOutput as TableTransformerObjectDetectionOutput, __webpack_exports__TableTransformerPreTrainedModel as TableTransformerPreTrainedModel, __webpack_exports__Tensor as Tensor, __webpack_exports__Text2TextGenerationPipeline as Text2TextGenerationPipeline, __webpack_exports__TextClassificationPipeline as TextClassificationPipeline, __webpack_exports__TextGenerationPipeline as TextGenerationPipeline, __webpack_exports__TextStreamer as TextStreamer, __webpack_exports__TextToAudioPipeline as TextToAudioPipeline, __webpack_exports__TokenClassificationPipeline as TokenClassificationPipeline, __webpack_exports__TokenClassifierOutput as TokenClassifierOutput, __webpack_exports__TokenizerModel as TokenizerModel, __webpack_exports__TrOCRForCausalLM as TrOCRForCausalLM, __webpack_exports__TrOCRPreTrainedModel as TrOCRPreTrainedModel, __webpack_exports__TranslationPipeline as TranslationPipeline, __webpack_exports__UniSpeechForCTC as UniSpeechForCTC, __webpack_exports__UniSpeechForSequenceClassification as UniSpeechForSequenceClassification, __webpack_exports__UniSpeechModel as UniSpeechModel, __webpack_exports__UniSpeechPreTrainedModel as UniSpeechPreTrainedModel, __webpack_exports__UniSpeechSatForAudioFrameClassification as UniSpeechSatForAudioFrameClassification, __webpack_exports__UniSpeechSatForCTC as UniSpeechSatForCTC, __webpack_exports__UniSpeechSatForSequenceClassification as UniSpeechSatForSequenceClassification, __webpack_exports__UniSpeechSatModel as UniSpeechSatModel, __webpack_exports__UniSpeechSatPreTrainedModel as UniSpeechSatPreTrainedModel, __webpack_exports__ViTFeatureExtractor as ViTFeatureExtractor, __webpack_exports__ViTForImageClassification as ViTForImageClassification, __webpack_exports__ViTImageProcessor as ViTImageProcessor, __webpack_exports__ViTModel as ViTModel, __webpack_exports__ViTPreTrainedModel as ViTPreTrainedModel, __webpack_exports__VisionEncoderDecoderModel as VisionEncoderDecoderModel, __webpack_exports__VitMatteForImageMatting as VitMatteForImageMatting, __webpack_exports__VitMatteImageProcessor as VitMatteImageProcessor, __webpack_exports__VitMattePreTrainedModel as VitMattePreTrainedModel, __webpack_exports__VitsModel as VitsModel, __webpack_exports__VitsModelOutput as VitsModelOutput, __webpack_exports__VitsPreTrainedModel as VitsPreTrainedModel, __webpack_exports__VitsTokenizer as VitsTokenizer, __webpack_exports__Wav2Vec2BertForCTC as Wav2Vec2BertForCTC, __webpack_exports__Wav2Vec2BertForSequenceClassification as Wav2Vec2BertForSequenceClassification, __webpack_exports__Wav2Vec2BertModel as Wav2Vec2BertModel, __webpack_exports__Wav2Vec2BertPreTrainedModel as Wav2Vec2BertPreTrainedModel, __webpack_exports__Wav2Vec2CTCTokenizer as Wav2Vec2CTCTokenizer, __webpack_exports__Wav2Vec2FeatureExtractor as Wav2Vec2FeatureExtractor, __webpack_exports__Wav2Vec2ForAudioFrameClassification as Wav2Vec2ForAudioFrameClassification, __webpack_exports__Wav2Vec2ForCTC as Wav2Vec2ForCTC, __webpack_exports__Wav2Vec2ForSequenceClassification as Wav2Vec2ForSequenceClassification, __webpack_exports__Wav2Vec2Model as Wav2Vec2Model, __webpack_exports__Wav2Vec2PreTrainedModel as Wav2Vec2PreTrainedModel, __webpack_exports__Wav2Vec2ProcessorWithLM as Wav2Vec2ProcessorWithLM, __webpack_exports__WavLMForAudioFrameClassification as WavLMForAudioFrameClassification, __webpack_exports__WavLMForCTC as WavLMForCTC, __webpack_exports__WavLMForSequenceClassification as WavLMForSequenceClassification, __webpack_exports__WavLMForXVector as WavLMForXVector, __webpack_exports__WavLMModel as WavLMModel, __webpack_exports__WavLMPreTrainedModel as WavLMPreTrainedModel, __webpack_exports__WeSpeakerFeatureExtractor as WeSpeakerFeatureExtractor, __webpack_exports__WeSpeakerResNetModel as WeSpeakerResNetModel, __webpack_exports__WeSpeakerResNetPreTrainedModel as WeSpeakerResNetPreTrainedModel, __webpack_exports__WhisperFeatureExtractor as WhisperFeatureExtractor, __webpack_exports__WhisperForConditionalGeneration as WhisperForConditionalGeneration, __webpack_exports__WhisperModel as WhisperModel, __webpack_exports__WhisperPreTrainedModel as WhisperPreTrainedModel, __webpack_exports__WhisperProcessor as WhisperProcessor, __webpack_exports__WhisperTextStreamer as WhisperTextStreamer, __webpack_exports__WhisperTokenizer as WhisperTokenizer, __webpack_exports__XLMForQuestionAnswering as XLMForQuestionAnswering, __webpack_exports__XLMForSequenceClassification as XLMForSequenceClassification, __webpack_exports__XLMForTokenClassification as XLMForTokenClassification, __webpack_exports__XLMModel as XLMModel, __webpack_exports__XLMPreTrainedModel as XLMPreTrainedModel, __webpack_exports__XLMRobertaForMaskedLM as XLMRobertaForMaskedLM, __webpack_exports__XLMRobertaForQuestionAnswering as XLMRobertaForQuestionAnswering, __webpack_exports__XLMRobertaForSequenceClassification as XLMRobertaForSequenceClassification, __webpack_exports__XLMRobertaForTokenClassification as XLMRobertaForTokenClassification, __webpack_exports__XLMRobertaModel as XLMRobertaModel, __webpack_exports__XLMRobertaPreTrainedModel as XLMRobertaPreTrainedModel, __webpack_exports__XLMRobertaTokenizer as XLMRobertaTokenizer, __webpack_exports__XLMTokenizer as XLMTokenizer, __webpack_exports__XLMWithLMHeadModel as XLMWithLMHeadModel, __webpack_exports__XVectorOutput as XVectorOutput, __webpack_exports__YolosFeatureExtractor as YolosFeatureExtractor, __webpack_exports__YolosForObjectDetection as YolosForObjectDetection, __webpack_exports__YolosModel as YolosModel, __webpack_exports__YolosObjectDetectionOutput as YolosObjectDetectionOutput, __webpack_exports__YolosPreTrainedModel as YolosPreTrainedModel, __webpack_exports__ZeroShotAudioClassificationPipeline as ZeroShotAudioClassificationPipeline, __webpack_exports__ZeroShotClassificationPipeline as ZeroShotClassificationPipeline, __webpack_exports__ZeroShotImageClassificationPipeline as ZeroShotImageClassificationPipeline, __webpack_exports__ZeroShotObjectDetectionPipeline as ZeroShotObjectDetectionPipeline, __webpack_exports__bankers_round as bankers_round, __webpack_exports__cat as cat, __webpack_exports__cos_sim as cos_sim, __webpack_exports__dot as dot, __webpack_exports__dynamic_time_warping as dynamic_time_warping, __webpack_exports__env as env, __webpack_exports__full as full, __webpack_exports__full_like as full_like, __webpack_exports__getKeyValueShapes as getKeyValueShapes, __webpack_exports__hamming as hamming, __webpack_exports__hanning as hanning, __webpack_exports__interpolate as interpolate, __webpack_exports__interpolate_4d as interpolate_4d, __webpack_exports__interpolate_data as interpolate_data, __webpack_exports__is_chinese_char as is_chinese_char, __webpack_exports__layer_norm as layer_norm, __webpack_exports__log_softmax as log_softmax, __webpack_exports__magnitude as magnitude, __webpack_exports__matmul as matmul, __webpack_exports__max as max, __webpack_exports__mean as mean, __webpack_exports__mean_pooling as mean_pooling, __webpack_exports__medianFilter as medianFilter, __webpack_exports__mel_filter_bank as mel_filter_bank, __webpack_exports__min as min, __webpack_exports__ones as ones, __webpack_exports__ones_like as ones_like, __webpack_exports__permute as permute, __webpack_exports__permute_data as permute_data, __webpack_exports__pipeline as pipeline, __webpack_exports__quantize_embeddings as quantize_embeddings, __webpack_exports__read_audio as read_audio, __webpack_exports__rfft as rfft, __webpack_exports__round as round, __webpack_exports__softmax as softmax, __webpack_exports__spectrogram as spectrogram, __webpack_exports__stack as stack, __webpack_exports__std_mean as std_mean, __webpack_exports__topk as topk, __webpack_exports__window_function as window_function, __webpack_exports__zeros as zeros, __webpack_exports__zeros_like as zeros_like };
31695
+ export { __webpack_exports__ASTFeatureExtractor as ASTFeatureExtractor, __webpack_exports__ASTForAudioClassification as ASTForAudioClassification, __webpack_exports__ASTModel as ASTModel, __webpack_exports__ASTPreTrainedModel as ASTPreTrainedModel, __webpack_exports__AlbertForMaskedLM as AlbertForMaskedLM, __webpack_exports__AlbertForQuestionAnswering as AlbertForQuestionAnswering, __webpack_exports__AlbertForSequenceClassification as AlbertForSequenceClassification, __webpack_exports__AlbertModel as AlbertModel, __webpack_exports__AlbertPreTrainedModel as AlbertPreTrainedModel, __webpack_exports__AlbertTokenizer as AlbertTokenizer, __webpack_exports__AudioClassificationPipeline as AudioClassificationPipeline, __webpack_exports__AutoConfig as AutoConfig, __webpack_exports__AutoModel as AutoModel, __webpack_exports__AutoModelForAudioClassification as AutoModelForAudioClassification, __webpack_exports__AutoModelForAudioFrameClassification as AutoModelForAudioFrameClassification, __webpack_exports__AutoModelForCTC as AutoModelForCTC, __webpack_exports__AutoModelForCausalLM as AutoModelForCausalLM, __webpack_exports__AutoModelForDepthEstimation as AutoModelForDepthEstimation, __webpack_exports__AutoModelForDocumentQuestionAnswering as AutoModelForDocumentQuestionAnswering, __webpack_exports__AutoModelForImageClassification as AutoModelForImageClassification, __webpack_exports__AutoModelForImageFeatureExtraction as AutoModelForImageFeatureExtraction, __webpack_exports__AutoModelForImageMatting as AutoModelForImageMatting, __webpack_exports__AutoModelForImageSegmentation as AutoModelForImageSegmentation, __webpack_exports__AutoModelForImageToImage as AutoModelForImageToImage, __webpack_exports__AutoModelForMaskGeneration as AutoModelForMaskGeneration, __webpack_exports__AutoModelForMaskedLM as AutoModelForMaskedLM, __webpack_exports__AutoModelForNormalEstimation as AutoModelForNormalEstimation, __webpack_exports__AutoModelForObjectDetection as AutoModelForObjectDetection, __webpack_exports__AutoModelForQuestionAnswering as AutoModelForQuestionAnswering, __webpack_exports__AutoModelForSemanticSegmentation as AutoModelForSemanticSegmentation, __webpack_exports__AutoModelForSeq2SeqLM as AutoModelForSeq2SeqLM, __webpack_exports__AutoModelForSequenceClassification as AutoModelForSequenceClassification, __webpack_exports__AutoModelForSpeechSeq2Seq as AutoModelForSpeechSeq2Seq, __webpack_exports__AutoModelForTextToSpectrogram as AutoModelForTextToSpectrogram, __webpack_exports__AutoModelForTextToWaveform as AutoModelForTextToWaveform, __webpack_exports__AutoModelForTokenClassification as AutoModelForTokenClassification, __webpack_exports__AutoModelForUniversalSegmentation as AutoModelForUniversalSegmentation, __webpack_exports__AutoModelForVision2Seq as AutoModelForVision2Seq, __webpack_exports__AutoModelForXVector as AutoModelForXVector, __webpack_exports__AutoModelForZeroShotObjectDetection as AutoModelForZeroShotObjectDetection, __webpack_exports__AutoProcessor as AutoProcessor, __webpack_exports__AutoTokenizer as AutoTokenizer, __webpack_exports__AutomaticSpeechRecognitionPipeline as AutomaticSpeechRecognitionPipeline, __webpack_exports__BartForConditionalGeneration as BartForConditionalGeneration, __webpack_exports__BartForSequenceClassification as BartForSequenceClassification, __webpack_exports__BartModel as BartModel, __webpack_exports__BartPretrainedModel as BartPretrainedModel, __webpack_exports__BartTokenizer as BartTokenizer, __webpack_exports__BaseModelOutput as BaseModelOutput, __webpack_exports__BaseStreamer as BaseStreamer, __webpack_exports__BeitFeatureExtractor as BeitFeatureExtractor, __webpack_exports__BeitForImageClassification as BeitForImageClassification, __webpack_exports__BeitModel as BeitModel, __webpack_exports__BeitPreTrainedModel as BeitPreTrainedModel, __webpack_exports__BertForMaskedLM as BertForMaskedLM, __webpack_exports__BertForQuestionAnswering as BertForQuestionAnswering, __webpack_exports__BertForSequenceClassification as BertForSequenceClassification, __webpack_exports__BertForTokenClassification as BertForTokenClassification, __webpack_exports__BertModel as BertModel, __webpack_exports__BertPreTrainedModel as BertPreTrainedModel, __webpack_exports__BertTokenizer as BertTokenizer, __webpack_exports__BitImageProcessor as BitImageProcessor, __webpack_exports__BlenderbotForConditionalGeneration as BlenderbotForConditionalGeneration, __webpack_exports__BlenderbotModel as BlenderbotModel, __webpack_exports__BlenderbotPreTrainedModel as BlenderbotPreTrainedModel, __webpack_exports__BlenderbotSmallForConditionalGeneration as BlenderbotSmallForConditionalGeneration, __webpack_exports__BlenderbotSmallModel as BlenderbotSmallModel, __webpack_exports__BlenderbotSmallPreTrainedModel as BlenderbotSmallPreTrainedModel, __webpack_exports__BlenderbotSmallTokenizer as BlenderbotSmallTokenizer, __webpack_exports__BlenderbotTokenizer as BlenderbotTokenizer, __webpack_exports__BloomForCausalLM as BloomForCausalLM, __webpack_exports__BloomModel as BloomModel, __webpack_exports__BloomPreTrainedModel as BloomPreTrainedModel, __webpack_exports__BloomTokenizer as BloomTokenizer, __webpack_exports__CLIPFeatureExtractor as CLIPFeatureExtractor, __webpack_exports__CLIPImageProcessor as CLIPImageProcessor, __webpack_exports__CLIPModel as CLIPModel, __webpack_exports__CLIPPreTrainedModel as CLIPPreTrainedModel, __webpack_exports__CLIPSegForImageSegmentation as CLIPSegForImageSegmentation, __webpack_exports__CLIPSegModel as CLIPSegModel, __webpack_exports__CLIPSegPreTrainedModel as CLIPSegPreTrainedModel, __webpack_exports__CLIPTextModel as CLIPTextModel, __webpack_exports__CLIPTextModelWithProjection as CLIPTextModelWithProjection, __webpack_exports__CLIPTokenizer as CLIPTokenizer, __webpack_exports__CLIPVisionModel as CLIPVisionModel, __webpack_exports__CLIPVisionModelWithProjection as CLIPVisionModelWithProjection, __webpack_exports__CamembertForMaskedLM as CamembertForMaskedLM, __webpack_exports__CamembertForQuestionAnswering as CamembertForQuestionAnswering, __webpack_exports__CamembertForSequenceClassification as CamembertForSequenceClassification, __webpack_exports__CamembertForTokenClassification as CamembertForTokenClassification, __webpack_exports__CamembertModel as CamembertModel, __webpack_exports__CamembertPreTrainedModel as CamembertPreTrainedModel, __webpack_exports__CamembertTokenizer as CamembertTokenizer, __webpack_exports__CausalLMOutput as CausalLMOutput, __webpack_exports__CausalLMOutputWithPast as CausalLMOutputWithPast, __webpack_exports__ChineseCLIPFeatureExtractor as ChineseCLIPFeatureExtractor, __webpack_exports__ChineseCLIPModel as ChineseCLIPModel, __webpack_exports__ChineseCLIPPreTrainedModel as ChineseCLIPPreTrainedModel, __webpack_exports__ClapAudioModelWithProjection as ClapAudioModelWithProjection, __webpack_exports__ClapFeatureExtractor as ClapFeatureExtractor, __webpack_exports__ClapModel as ClapModel, __webpack_exports__ClapPreTrainedModel as ClapPreTrainedModel, __webpack_exports__ClapTextModelWithProjection as ClapTextModelWithProjection, __webpack_exports__CodeGenForCausalLM as CodeGenForCausalLM, __webpack_exports__CodeGenModel as CodeGenModel, __webpack_exports__CodeGenPreTrainedModel as CodeGenPreTrainedModel, __webpack_exports__CodeGenTokenizer as CodeGenTokenizer, __webpack_exports__CodeLlamaTokenizer as CodeLlamaTokenizer, __webpack_exports__CohereForCausalLM as CohereForCausalLM, __webpack_exports__CohereModel as CohereModel, __webpack_exports__CoherePreTrainedModel as CoherePreTrainedModel, __webpack_exports__CohereTokenizer as CohereTokenizer, __webpack_exports__ConvBertForMaskedLM as ConvBertForMaskedLM, __webpack_exports__ConvBertForQuestionAnswering as ConvBertForQuestionAnswering, __webpack_exports__ConvBertForSequenceClassification as ConvBertForSequenceClassification, __webpack_exports__ConvBertForTokenClassification as ConvBertForTokenClassification, __webpack_exports__ConvBertModel as ConvBertModel, __webpack_exports__ConvBertPreTrainedModel as ConvBertPreTrainedModel, __webpack_exports__ConvBertTokenizer as ConvBertTokenizer, __webpack_exports__ConvNextFeatureExtractor as ConvNextFeatureExtractor, __webpack_exports__ConvNextForImageClassification as ConvNextForImageClassification, __webpack_exports__ConvNextImageProcessor as ConvNextImageProcessor, __webpack_exports__ConvNextModel as ConvNextModel, __webpack_exports__ConvNextPreTrainedModel as ConvNextPreTrainedModel, __webpack_exports__ConvNextV2ForImageClassification as ConvNextV2ForImageClassification, __webpack_exports__ConvNextV2Model as ConvNextV2Model, __webpack_exports__ConvNextV2PreTrainedModel as ConvNextV2PreTrainedModel, __webpack_exports__DPTFeatureExtractor as DPTFeatureExtractor, __webpack_exports__DPTForDepthEstimation as DPTForDepthEstimation, __webpack_exports__DPTImageProcessor as DPTImageProcessor, __webpack_exports__DPTModel as DPTModel, __webpack_exports__DPTPreTrainedModel as DPTPreTrainedModel, __webpack_exports__DebertaForMaskedLM as DebertaForMaskedLM, __webpack_exports__DebertaForQuestionAnswering as DebertaForQuestionAnswering, __webpack_exports__DebertaForSequenceClassification as DebertaForSequenceClassification, __webpack_exports__DebertaForTokenClassification as DebertaForTokenClassification, __webpack_exports__DebertaModel as DebertaModel, __webpack_exports__DebertaPreTrainedModel as DebertaPreTrainedModel, __webpack_exports__DebertaTokenizer as DebertaTokenizer, __webpack_exports__DebertaV2ForMaskedLM as DebertaV2ForMaskedLM, __webpack_exports__DebertaV2ForQuestionAnswering as DebertaV2ForQuestionAnswering, __webpack_exports__DebertaV2ForSequenceClassification as DebertaV2ForSequenceClassification, __webpack_exports__DebertaV2ForTokenClassification as DebertaV2ForTokenClassification, __webpack_exports__DebertaV2Model as DebertaV2Model, __webpack_exports__DebertaV2PreTrainedModel as DebertaV2PreTrainedModel, __webpack_exports__DebertaV2Tokenizer as DebertaV2Tokenizer, __webpack_exports__DecisionTransformerModel as DecisionTransformerModel, __webpack_exports__DecisionTransformerPreTrainedModel as DecisionTransformerPreTrainedModel, __webpack_exports__DeiTFeatureExtractor as DeiTFeatureExtractor, __webpack_exports__DeiTForImageClassification as DeiTForImageClassification, __webpack_exports__DeiTModel as DeiTModel, __webpack_exports__DeiTPreTrainedModel as DeiTPreTrainedModel, __webpack_exports__DepthAnythingForDepthEstimation as DepthAnythingForDepthEstimation, __webpack_exports__DepthAnythingPreTrainedModel as DepthAnythingPreTrainedModel, __webpack_exports__DepthEstimationPipeline as DepthEstimationPipeline, __webpack_exports__DetrFeatureExtractor as DetrFeatureExtractor, __webpack_exports__DetrForObjectDetection as DetrForObjectDetection, __webpack_exports__DetrForSegmentation as DetrForSegmentation, __webpack_exports__DetrModel as DetrModel, __webpack_exports__DetrObjectDetectionOutput as DetrObjectDetectionOutput, __webpack_exports__DetrPreTrainedModel as DetrPreTrainedModel, __webpack_exports__DetrSegmentationOutput as DetrSegmentationOutput, __webpack_exports__Dinov2ForImageClassification as Dinov2ForImageClassification, __webpack_exports__Dinov2Model as Dinov2Model, __webpack_exports__Dinov2PreTrainedModel as Dinov2PreTrainedModel, __webpack_exports__DistilBertForMaskedLM as DistilBertForMaskedLM, __webpack_exports__DistilBertForQuestionAnswering as DistilBertForQuestionAnswering, __webpack_exports__DistilBertForSequenceClassification as DistilBertForSequenceClassification, __webpack_exports__DistilBertForTokenClassification as DistilBertForTokenClassification, __webpack_exports__DistilBertModel as DistilBertModel, __webpack_exports__DistilBertPreTrainedModel as DistilBertPreTrainedModel, __webpack_exports__DistilBertTokenizer as DistilBertTokenizer, __webpack_exports__DocumentQuestionAnsweringPipeline as DocumentQuestionAnsweringPipeline, __webpack_exports__DonutFeatureExtractor as DonutFeatureExtractor, __webpack_exports__DonutSwinModel as DonutSwinModel, __webpack_exports__DonutSwinPreTrainedModel as DonutSwinPreTrainedModel, __webpack_exports__EfficientNetForImageClassification as EfficientNetForImageClassification, __webpack_exports__EfficientNetImageProcessor as EfficientNetImageProcessor, __webpack_exports__EfficientNetModel as EfficientNetModel, __webpack_exports__EfficientNetPreTrainedModel as EfficientNetPreTrainedModel, __webpack_exports__ElectraForMaskedLM as ElectraForMaskedLM, __webpack_exports__ElectraForQuestionAnswering as ElectraForQuestionAnswering, __webpack_exports__ElectraForSequenceClassification as ElectraForSequenceClassification, __webpack_exports__ElectraForTokenClassification as ElectraForTokenClassification, __webpack_exports__ElectraModel as ElectraModel, __webpack_exports__ElectraPreTrainedModel as ElectraPreTrainedModel, __webpack_exports__ElectraTokenizer as ElectraTokenizer, __webpack_exports__EosTokenCriteria as EosTokenCriteria, __webpack_exports__EsmForMaskedLM as EsmForMaskedLM, __webpack_exports__EsmForSequenceClassification as EsmForSequenceClassification, __webpack_exports__EsmForTokenClassification as EsmForTokenClassification, __webpack_exports__EsmModel as EsmModel, __webpack_exports__EsmPreTrainedModel as EsmPreTrainedModel, __webpack_exports__EsmTokenizer as EsmTokenizer, __webpack_exports__FFT as FFT, __webpack_exports__FalconForCausalLM as FalconForCausalLM, __webpack_exports__FalconModel as FalconModel, __webpack_exports__FalconPreTrainedModel as FalconPreTrainedModel, __webpack_exports__FalconTokenizer as FalconTokenizer, __webpack_exports__FastViTForImageClassification as FastViTForImageClassification, __webpack_exports__FastViTModel as FastViTModel, __webpack_exports__FastViTPreTrainedModel as FastViTPreTrainedModel, __webpack_exports__FeatureExtractionPipeline as FeatureExtractionPipeline, __webpack_exports__FeatureExtractor as FeatureExtractor, __webpack_exports__FillMaskPipeline as FillMaskPipeline, __webpack_exports__Florence2ForConditionalGeneration as Florence2ForConditionalGeneration, __webpack_exports__Florence2PreTrainedModel as Florence2PreTrainedModel, __webpack_exports__Florence2Processor as Florence2Processor, __webpack_exports__GLPNFeatureExtractor as GLPNFeatureExtractor, __webpack_exports__GLPNForDepthEstimation as GLPNForDepthEstimation, __webpack_exports__GLPNModel as GLPNModel, __webpack_exports__GLPNPreTrainedModel as GLPNPreTrainedModel, __webpack_exports__GPT2LMHeadModel as GPT2LMHeadModel, __webpack_exports__GPT2Model as GPT2Model, __webpack_exports__GPT2PreTrainedModel as GPT2PreTrainedModel, __webpack_exports__GPT2Tokenizer as GPT2Tokenizer, __webpack_exports__GPTBigCodeForCausalLM as GPTBigCodeForCausalLM, __webpack_exports__GPTBigCodeModel as GPTBigCodeModel, __webpack_exports__GPTBigCodePreTrainedModel as GPTBigCodePreTrainedModel, __webpack_exports__GPTJForCausalLM as GPTJForCausalLM, __webpack_exports__GPTJModel as GPTJModel, __webpack_exports__GPTJPreTrainedModel as GPTJPreTrainedModel, __webpack_exports__GPTNeoForCausalLM as GPTNeoForCausalLM, __webpack_exports__GPTNeoModel as GPTNeoModel, __webpack_exports__GPTNeoPreTrainedModel as GPTNeoPreTrainedModel, __webpack_exports__GPTNeoXForCausalLM as GPTNeoXForCausalLM, __webpack_exports__GPTNeoXModel as GPTNeoXModel, __webpack_exports__GPTNeoXPreTrainedModel as GPTNeoXPreTrainedModel, __webpack_exports__GPTNeoXTokenizer as GPTNeoXTokenizer, __webpack_exports__Gemma2ForCausalLM as Gemma2ForCausalLM, __webpack_exports__Gemma2Model as Gemma2Model, __webpack_exports__Gemma2PreTrainedModel as Gemma2PreTrainedModel, __webpack_exports__GemmaForCausalLM as GemmaForCausalLM, __webpack_exports__GemmaModel as GemmaModel, __webpack_exports__GemmaPreTrainedModel as GemmaPreTrainedModel, __webpack_exports__GemmaTokenizer as GemmaTokenizer, __webpack_exports__Grok1Tokenizer as Grok1Tokenizer, __webpack_exports__GroupViTModel as GroupViTModel, __webpack_exports__GroupViTPreTrainedModel as GroupViTPreTrainedModel, __webpack_exports__HerbertTokenizer as HerbertTokenizer, __webpack_exports__HieraForImageClassification as HieraForImageClassification, __webpack_exports__HieraModel as HieraModel, __webpack_exports__HieraPreTrainedModel as HieraPreTrainedModel, __webpack_exports__HubertForCTC as HubertForCTC, __webpack_exports__HubertForSequenceClassification as HubertForSequenceClassification, __webpack_exports__HubertModel as HubertModel, __webpack_exports__HubertPreTrainedModel as HubertPreTrainedModel, __webpack_exports__ImageClassificationPipeline as ImageClassificationPipeline, __webpack_exports__ImageFeatureExtractionPipeline as ImageFeatureExtractionPipeline, __webpack_exports__ImageFeatureExtractor as ImageFeatureExtractor, __webpack_exports__ImageMattingOutput as ImageMattingOutput, __webpack_exports__ImageSegmentationPipeline as ImageSegmentationPipeline, __webpack_exports__ImageToImagePipeline as ImageToImagePipeline, __webpack_exports__ImageToTextPipeline as ImageToTextPipeline, __webpack_exports__InterruptableStoppingCriteria as InterruptableStoppingCriteria, __webpack_exports__JAISLMHeadModel as JAISLMHeadModel, __webpack_exports__JAISModel as JAISModel, __webpack_exports__JAISPreTrainedModel as JAISPreTrainedModel, __webpack_exports__LlamaForCausalLM as LlamaForCausalLM, __webpack_exports__LlamaModel as LlamaModel, __webpack_exports__LlamaPreTrainedModel as LlamaPreTrainedModel, __webpack_exports__LlamaTokenizer as LlamaTokenizer, __webpack_exports__LlavaForConditionalGeneration as LlavaForConditionalGeneration, __webpack_exports__LlavaPreTrainedModel as LlavaPreTrainedModel, __webpack_exports__LongT5ForConditionalGeneration as LongT5ForConditionalGeneration, __webpack_exports__LongT5Model as LongT5Model, __webpack_exports__LongT5PreTrainedModel as LongT5PreTrainedModel, __webpack_exports__M2M100ForConditionalGeneration as M2M100ForConditionalGeneration, __webpack_exports__M2M100Model as M2M100Model, __webpack_exports__M2M100PreTrainedModel as M2M100PreTrainedModel, __webpack_exports__M2M100Tokenizer as M2M100Tokenizer, __webpack_exports__MBart50Tokenizer as MBart50Tokenizer, __webpack_exports__MBartForCausalLM as MBartForCausalLM, __webpack_exports__MBartForConditionalGeneration as MBartForConditionalGeneration, __webpack_exports__MBartForSequenceClassification as MBartForSequenceClassification, __webpack_exports__MBartModel as MBartModel, __webpack_exports__MBartPreTrainedModel as MBartPreTrainedModel, __webpack_exports__MBartTokenizer as MBartTokenizer, __webpack_exports__MPNetForMaskedLM as MPNetForMaskedLM, __webpack_exports__MPNetForQuestionAnswering as MPNetForQuestionAnswering, __webpack_exports__MPNetForSequenceClassification as MPNetForSequenceClassification, __webpack_exports__MPNetForTokenClassification as MPNetForTokenClassification, __webpack_exports__MPNetModel as MPNetModel, __webpack_exports__MPNetPreTrainedModel as MPNetPreTrainedModel, __webpack_exports__MPNetTokenizer as MPNetTokenizer, __webpack_exports__MT5ForConditionalGeneration as MT5ForConditionalGeneration, __webpack_exports__MT5Model as MT5Model, __webpack_exports__MT5PreTrainedModel as MT5PreTrainedModel, __webpack_exports__MarianMTModel as MarianMTModel, __webpack_exports__MarianModel as MarianModel, __webpack_exports__MarianPreTrainedModel as MarianPreTrainedModel, __webpack_exports__MarianTokenizer as MarianTokenizer, __webpack_exports__MaskFormerFeatureExtractor as MaskFormerFeatureExtractor, __webpack_exports__MaskFormerForInstanceSegmentation as MaskFormerForInstanceSegmentation, __webpack_exports__MaskFormerModel as MaskFormerModel, __webpack_exports__MaskFormerPreTrainedModel as MaskFormerPreTrainedModel, __webpack_exports__MaskedLMOutput as MaskedLMOutput, __webpack_exports__MaxLengthCriteria as MaxLengthCriteria, __webpack_exports__MistralForCausalLM as MistralForCausalLM, __webpack_exports__MistralModel as MistralModel, __webpack_exports__MistralPreTrainedModel as MistralPreTrainedModel, __webpack_exports__MobileBertForMaskedLM as MobileBertForMaskedLM, __webpack_exports__MobileBertForQuestionAnswering as MobileBertForQuestionAnswering, __webpack_exports__MobileBertForSequenceClassification as MobileBertForSequenceClassification, __webpack_exports__MobileBertModel as MobileBertModel, __webpack_exports__MobileBertPreTrainedModel as MobileBertPreTrainedModel, __webpack_exports__MobileBertTokenizer as MobileBertTokenizer, __webpack_exports__MobileNetV1FeatureExtractor as MobileNetV1FeatureExtractor, __webpack_exports__MobileNetV1ForImageClassification as MobileNetV1ForImageClassification, __webpack_exports__MobileNetV1Model as MobileNetV1Model, __webpack_exports__MobileNetV1PreTrainedModel as MobileNetV1PreTrainedModel, __webpack_exports__MobileNetV2FeatureExtractor as MobileNetV2FeatureExtractor, __webpack_exports__MobileNetV2ForImageClassification as MobileNetV2ForImageClassification, __webpack_exports__MobileNetV2Model as MobileNetV2Model, __webpack_exports__MobileNetV2PreTrainedModel as MobileNetV2PreTrainedModel, __webpack_exports__MobileNetV3FeatureExtractor as MobileNetV3FeatureExtractor, __webpack_exports__MobileNetV3ForImageClassification as MobileNetV3ForImageClassification, __webpack_exports__MobileNetV3Model as MobileNetV3Model, __webpack_exports__MobileNetV3PreTrainedModel as MobileNetV3PreTrainedModel, __webpack_exports__MobileNetV4FeatureExtractor as MobileNetV4FeatureExtractor, __webpack_exports__MobileNetV4ForImageClassification as MobileNetV4ForImageClassification, __webpack_exports__MobileNetV4Model as MobileNetV4Model, __webpack_exports__MobileNetV4PreTrainedModel as MobileNetV4PreTrainedModel, __webpack_exports__MobileViTFeatureExtractor as MobileViTFeatureExtractor, __webpack_exports__MobileViTForImageClassification as MobileViTForImageClassification, __webpack_exports__MobileViTImageProcessor as MobileViTImageProcessor, __webpack_exports__MobileViTModel as MobileViTModel, __webpack_exports__MobileViTPreTrainedModel as MobileViTPreTrainedModel, __webpack_exports__MobileViTV2ForImageClassification as MobileViTV2ForImageClassification, __webpack_exports__MobileViTV2Model as MobileViTV2Model, __webpack_exports__MobileViTV2PreTrainedModel as MobileViTV2PreTrainedModel, __webpack_exports__ModelOutput as ModelOutput, __webpack_exports__Moondream1ForConditionalGeneration as Moondream1ForConditionalGeneration, __webpack_exports__MptForCausalLM as MptForCausalLM, __webpack_exports__MptModel as MptModel, __webpack_exports__MptPreTrainedModel as MptPreTrainedModel, __webpack_exports__MusicgenForCausalLM as MusicgenForCausalLM, __webpack_exports__MusicgenForConditionalGeneration as MusicgenForConditionalGeneration, __webpack_exports__MusicgenModel as MusicgenModel, __webpack_exports__MusicgenPreTrainedModel as MusicgenPreTrainedModel, __webpack_exports__NllbTokenizer as NllbTokenizer, __webpack_exports__NomicBertModel as NomicBertModel, __webpack_exports__NomicBertPreTrainedModel as NomicBertPreTrainedModel, __webpack_exports__NougatImageProcessor as NougatImageProcessor, __webpack_exports__NougatTokenizer as NougatTokenizer, __webpack_exports__OPTForCausalLM as OPTForCausalLM, __webpack_exports__OPTModel as OPTModel, __webpack_exports__OPTPreTrainedModel as OPTPreTrainedModel, __webpack_exports__ObjectDetectionPipeline as ObjectDetectionPipeline, __webpack_exports__OpenELMForCausalLM as OpenELMForCausalLM, __webpack_exports__OpenELMModel as OpenELMModel, __webpack_exports__OpenELMPreTrainedModel as OpenELMPreTrainedModel, __webpack_exports__OwlViTFeatureExtractor as OwlViTFeatureExtractor, __webpack_exports__OwlViTForObjectDetection as OwlViTForObjectDetection, __webpack_exports__OwlViTModel as OwlViTModel, __webpack_exports__OwlViTPreTrainedModel as OwlViTPreTrainedModel, __webpack_exports__OwlViTProcessor as OwlViTProcessor, __webpack_exports__Owlv2ForObjectDetection as Owlv2ForObjectDetection, __webpack_exports__Owlv2ImageProcessor as Owlv2ImageProcessor, __webpack_exports__Owlv2Model as Owlv2Model, __webpack_exports__Owlv2PreTrainedModel as Owlv2PreTrainedModel, __webpack_exports__Phi3ForCausalLM as Phi3ForCausalLM, __webpack_exports__Phi3Model as Phi3Model, __webpack_exports__Phi3PreTrainedModel as Phi3PreTrainedModel, __webpack_exports__PhiForCausalLM as PhiForCausalLM, __webpack_exports__PhiModel as PhiModel, __webpack_exports__PhiPreTrainedModel as PhiPreTrainedModel, __webpack_exports__Pipeline as Pipeline, __webpack_exports__PreTrainedModel as PreTrainedModel, __webpack_exports__PreTrainedTokenizer as PreTrainedTokenizer, __webpack_exports__PretrainedConfig as PretrainedConfig, __webpack_exports__PretrainedMixin as PretrainedMixin, __webpack_exports__Processor as Processor, __webpack_exports__PvtForImageClassification as PvtForImageClassification, __webpack_exports__PvtImageProcessor as PvtImageProcessor, __webpack_exports__PvtModel as PvtModel, __webpack_exports__PvtPreTrainedModel as PvtPreTrainedModel, __webpack_exports__PyAnnoteFeatureExtractor as PyAnnoteFeatureExtractor, __webpack_exports__PyAnnoteForAudioFrameClassification as PyAnnoteForAudioFrameClassification, __webpack_exports__PyAnnoteModel as PyAnnoteModel, __webpack_exports__PyAnnotePreTrainedModel as PyAnnotePreTrainedModel, __webpack_exports__PyAnnoteProcessor as PyAnnoteProcessor, __webpack_exports__QuestionAnsweringModelOutput as QuestionAnsweringModelOutput, __webpack_exports__QuestionAnsweringPipeline as QuestionAnsweringPipeline, __webpack_exports__Qwen2ForCausalLM as Qwen2ForCausalLM, __webpack_exports__Qwen2Model as Qwen2Model, __webpack_exports__Qwen2PreTrainedModel as Qwen2PreTrainedModel, __webpack_exports__Qwen2Tokenizer as Qwen2Tokenizer, __webpack_exports__RTDetrForObjectDetection as RTDetrForObjectDetection, __webpack_exports__RTDetrImageProcessor as RTDetrImageProcessor, __webpack_exports__RTDetrModel as RTDetrModel, __webpack_exports__RTDetrObjectDetectionOutput as RTDetrObjectDetectionOutput, __webpack_exports__RTDetrPreTrainedModel as RTDetrPreTrainedModel, __webpack_exports__RawImage as RawImage, __webpack_exports__ResNetForImageClassification as ResNetForImageClassification, __webpack_exports__ResNetModel as ResNetModel, __webpack_exports__ResNetPreTrainedModel as ResNetPreTrainedModel, __webpack_exports__RoFormerForMaskedLM as RoFormerForMaskedLM, __webpack_exports__RoFormerForQuestionAnswering as RoFormerForQuestionAnswering, __webpack_exports__RoFormerForSequenceClassification as RoFormerForSequenceClassification, __webpack_exports__RoFormerForTokenClassification as RoFormerForTokenClassification, __webpack_exports__RoFormerModel as RoFormerModel, __webpack_exports__RoFormerPreTrainedModel as RoFormerPreTrainedModel, __webpack_exports__RoFormerTokenizer as RoFormerTokenizer, __webpack_exports__RobertaForMaskedLM as RobertaForMaskedLM, __webpack_exports__RobertaForQuestionAnswering as RobertaForQuestionAnswering, __webpack_exports__RobertaForSequenceClassification as RobertaForSequenceClassification, __webpack_exports__RobertaForTokenClassification as RobertaForTokenClassification, __webpack_exports__RobertaModel as RobertaModel, __webpack_exports__RobertaPreTrainedModel as RobertaPreTrainedModel, __webpack_exports__RobertaTokenizer as RobertaTokenizer, __webpack_exports__SamImageProcessor as SamImageProcessor, __webpack_exports__SamImageSegmentationOutput as SamImageSegmentationOutput, __webpack_exports__SamModel as SamModel, __webpack_exports__SamPreTrainedModel as SamPreTrainedModel, __webpack_exports__SamProcessor as SamProcessor, __webpack_exports__SapiensFeatureExtractor as SapiensFeatureExtractor, __webpack_exports__SapiensForDepthEstimation as SapiensForDepthEstimation, __webpack_exports__SapiensForNormalEstimation as SapiensForNormalEstimation, __webpack_exports__SapiensForSemanticSegmentation as SapiensForSemanticSegmentation, __webpack_exports__SapiensPreTrainedModel as SapiensPreTrainedModel, __webpack_exports__SeamlessM4TFeatureExtractor as SeamlessM4TFeatureExtractor, __webpack_exports__SegformerFeatureExtractor as SegformerFeatureExtractor, __webpack_exports__SegformerForImageClassification as SegformerForImageClassification, __webpack_exports__SegformerForSemanticSegmentation as SegformerForSemanticSegmentation, __webpack_exports__SegformerModel as SegformerModel, __webpack_exports__SegformerPreTrainedModel as SegformerPreTrainedModel, __webpack_exports__Seq2SeqLMOutput as Seq2SeqLMOutput, __webpack_exports__SequenceClassifierOutput as SequenceClassifierOutput, __webpack_exports__SiglipImageProcessor as SiglipImageProcessor, __webpack_exports__SiglipModel as SiglipModel, __webpack_exports__SiglipPreTrainedModel as SiglipPreTrainedModel, __webpack_exports__SiglipTextModel as SiglipTextModel, __webpack_exports__SiglipTokenizer as SiglipTokenizer, __webpack_exports__SiglipVisionModel as SiglipVisionModel, __webpack_exports__SpeechT5FeatureExtractor as SpeechT5FeatureExtractor, __webpack_exports__SpeechT5ForSpeechToText as SpeechT5ForSpeechToText, __webpack_exports__SpeechT5ForTextToSpeech as SpeechT5ForTextToSpeech, __webpack_exports__SpeechT5HifiGan as SpeechT5HifiGan, __webpack_exports__SpeechT5Model as SpeechT5Model, __webpack_exports__SpeechT5PreTrainedModel as SpeechT5PreTrainedModel, __webpack_exports__SpeechT5Processor as SpeechT5Processor, __webpack_exports__SpeechT5Tokenizer as SpeechT5Tokenizer, __webpack_exports__SqueezeBertForMaskedLM as SqueezeBertForMaskedLM, __webpack_exports__SqueezeBertForQuestionAnswering as SqueezeBertForQuestionAnswering, __webpack_exports__SqueezeBertForSequenceClassification as SqueezeBertForSequenceClassification, __webpack_exports__SqueezeBertModel as SqueezeBertModel, __webpack_exports__SqueezeBertPreTrainedModel as SqueezeBertPreTrainedModel, __webpack_exports__SqueezeBertTokenizer as SqueezeBertTokenizer, __webpack_exports__StableLmForCausalLM as StableLmForCausalLM, __webpack_exports__StableLmModel as StableLmModel, __webpack_exports__StableLmPreTrainedModel as StableLmPreTrainedModel, __webpack_exports__Starcoder2ForCausalLM as Starcoder2ForCausalLM, __webpack_exports__Starcoder2Model as Starcoder2Model, __webpack_exports__Starcoder2PreTrainedModel as Starcoder2PreTrainedModel, __webpack_exports__StoppingCriteria as StoppingCriteria, __webpack_exports__StoppingCriteriaList as StoppingCriteriaList, __webpack_exports__SummarizationPipeline as SummarizationPipeline, __webpack_exports__Swin2SRForImageSuperResolution as Swin2SRForImageSuperResolution, __webpack_exports__Swin2SRImageProcessor as Swin2SRImageProcessor, __webpack_exports__Swin2SRModel as Swin2SRModel, __webpack_exports__Swin2SRPreTrainedModel as Swin2SRPreTrainedModel, __webpack_exports__SwinForImageClassification as SwinForImageClassification, __webpack_exports__SwinModel as SwinModel, __webpack_exports__SwinPreTrainedModel as SwinPreTrainedModel, __webpack_exports__T5ForConditionalGeneration as T5ForConditionalGeneration, __webpack_exports__T5Model as T5Model, __webpack_exports__T5PreTrainedModel as T5PreTrainedModel, __webpack_exports__T5Tokenizer as T5Tokenizer, __webpack_exports__TableTransformerForObjectDetection as TableTransformerForObjectDetection, __webpack_exports__TableTransformerModel as TableTransformerModel, __webpack_exports__TableTransformerObjectDetectionOutput as TableTransformerObjectDetectionOutput, __webpack_exports__TableTransformerPreTrainedModel as TableTransformerPreTrainedModel, __webpack_exports__Tensor as Tensor, __webpack_exports__Text2TextGenerationPipeline as Text2TextGenerationPipeline, __webpack_exports__TextClassificationPipeline as TextClassificationPipeline, __webpack_exports__TextGenerationPipeline as TextGenerationPipeline, __webpack_exports__TextStreamer as TextStreamer, __webpack_exports__TextToAudioPipeline as TextToAudioPipeline, __webpack_exports__TokenClassificationPipeline as TokenClassificationPipeline, __webpack_exports__TokenClassifierOutput as TokenClassifierOutput, __webpack_exports__TokenizerModel as TokenizerModel, __webpack_exports__TrOCRForCausalLM as TrOCRForCausalLM, __webpack_exports__TrOCRPreTrainedModel as TrOCRPreTrainedModel, __webpack_exports__TranslationPipeline as TranslationPipeline, __webpack_exports__UniSpeechForCTC as UniSpeechForCTC, __webpack_exports__UniSpeechForSequenceClassification as UniSpeechForSequenceClassification, __webpack_exports__UniSpeechModel as UniSpeechModel, __webpack_exports__UniSpeechPreTrainedModel as UniSpeechPreTrainedModel, __webpack_exports__UniSpeechSatForAudioFrameClassification as UniSpeechSatForAudioFrameClassification, __webpack_exports__UniSpeechSatForCTC as UniSpeechSatForCTC, __webpack_exports__UniSpeechSatForSequenceClassification as UniSpeechSatForSequenceClassification, __webpack_exports__UniSpeechSatModel as UniSpeechSatModel, __webpack_exports__UniSpeechSatPreTrainedModel as UniSpeechSatPreTrainedModel, __webpack_exports__ViTFeatureExtractor as ViTFeatureExtractor, __webpack_exports__ViTForImageClassification as ViTForImageClassification, __webpack_exports__ViTImageProcessor as ViTImageProcessor, __webpack_exports__ViTMAEModel as ViTMAEModel, __webpack_exports__ViTMAEPreTrainedModel as ViTMAEPreTrainedModel, __webpack_exports__ViTMSNForImageClassification as ViTMSNForImageClassification, __webpack_exports__ViTMSNModel as ViTMSNModel, __webpack_exports__ViTMSNPreTrainedModel as ViTMSNPreTrainedModel, __webpack_exports__ViTModel as ViTModel, __webpack_exports__ViTPreTrainedModel as ViTPreTrainedModel, __webpack_exports__VisionEncoderDecoderModel as VisionEncoderDecoderModel, __webpack_exports__VitMatteForImageMatting as VitMatteForImageMatting, __webpack_exports__VitMatteImageProcessor as VitMatteImageProcessor, __webpack_exports__VitMattePreTrainedModel as VitMattePreTrainedModel, __webpack_exports__VitsModel as VitsModel, __webpack_exports__VitsModelOutput as VitsModelOutput, __webpack_exports__VitsPreTrainedModel as VitsPreTrainedModel, __webpack_exports__VitsTokenizer as VitsTokenizer, __webpack_exports__Wav2Vec2BertForCTC as Wav2Vec2BertForCTC, __webpack_exports__Wav2Vec2BertForSequenceClassification as Wav2Vec2BertForSequenceClassification, __webpack_exports__Wav2Vec2BertModel as Wav2Vec2BertModel, __webpack_exports__Wav2Vec2BertPreTrainedModel as Wav2Vec2BertPreTrainedModel, __webpack_exports__Wav2Vec2CTCTokenizer as Wav2Vec2CTCTokenizer, __webpack_exports__Wav2Vec2FeatureExtractor as Wav2Vec2FeatureExtractor, __webpack_exports__Wav2Vec2ForAudioFrameClassification as Wav2Vec2ForAudioFrameClassification, __webpack_exports__Wav2Vec2ForCTC as Wav2Vec2ForCTC, __webpack_exports__Wav2Vec2ForSequenceClassification as Wav2Vec2ForSequenceClassification, __webpack_exports__Wav2Vec2Model as Wav2Vec2Model, __webpack_exports__Wav2Vec2PreTrainedModel as Wav2Vec2PreTrainedModel, __webpack_exports__Wav2Vec2ProcessorWithLM as Wav2Vec2ProcessorWithLM, __webpack_exports__WavLMForAudioFrameClassification as WavLMForAudioFrameClassification, __webpack_exports__WavLMForCTC as WavLMForCTC, __webpack_exports__WavLMForSequenceClassification as WavLMForSequenceClassification, __webpack_exports__WavLMForXVector as WavLMForXVector, __webpack_exports__WavLMModel as WavLMModel, __webpack_exports__WavLMPreTrainedModel as WavLMPreTrainedModel, __webpack_exports__WeSpeakerFeatureExtractor as WeSpeakerFeatureExtractor, __webpack_exports__WeSpeakerResNetModel as WeSpeakerResNetModel, __webpack_exports__WeSpeakerResNetPreTrainedModel as WeSpeakerResNetPreTrainedModel, __webpack_exports__WhisperFeatureExtractor as WhisperFeatureExtractor, __webpack_exports__WhisperForConditionalGeneration as WhisperForConditionalGeneration, __webpack_exports__WhisperModel as WhisperModel, __webpack_exports__WhisperPreTrainedModel as WhisperPreTrainedModel, __webpack_exports__WhisperProcessor as WhisperProcessor, __webpack_exports__WhisperTextStreamer as WhisperTextStreamer, __webpack_exports__WhisperTokenizer as WhisperTokenizer, __webpack_exports__XLMForQuestionAnswering as XLMForQuestionAnswering, __webpack_exports__XLMForSequenceClassification as XLMForSequenceClassification, __webpack_exports__XLMForTokenClassification as XLMForTokenClassification, __webpack_exports__XLMModel as XLMModel, __webpack_exports__XLMPreTrainedModel as XLMPreTrainedModel, __webpack_exports__XLMRobertaForMaskedLM as XLMRobertaForMaskedLM, __webpack_exports__XLMRobertaForQuestionAnswering as XLMRobertaForQuestionAnswering, __webpack_exports__XLMRobertaForSequenceClassification as XLMRobertaForSequenceClassification, __webpack_exports__XLMRobertaForTokenClassification as XLMRobertaForTokenClassification, __webpack_exports__XLMRobertaModel as XLMRobertaModel, __webpack_exports__XLMRobertaPreTrainedModel as XLMRobertaPreTrainedModel, __webpack_exports__XLMRobertaTokenizer as XLMRobertaTokenizer, __webpack_exports__XLMTokenizer as XLMTokenizer, __webpack_exports__XLMWithLMHeadModel as XLMWithLMHeadModel, __webpack_exports__XVectorOutput as XVectorOutput, __webpack_exports__YolosFeatureExtractor as YolosFeatureExtractor, __webpack_exports__YolosForObjectDetection as YolosForObjectDetection, __webpack_exports__YolosModel as YolosModel, __webpack_exports__YolosObjectDetectionOutput as YolosObjectDetectionOutput, __webpack_exports__YolosPreTrainedModel as YolosPreTrainedModel, __webpack_exports__ZeroShotAudioClassificationPipeline as ZeroShotAudioClassificationPipeline, __webpack_exports__ZeroShotClassificationPipeline as ZeroShotClassificationPipeline, __webpack_exports__ZeroShotImageClassificationPipeline as ZeroShotImageClassificationPipeline, __webpack_exports__ZeroShotObjectDetectionPipeline as ZeroShotObjectDetectionPipeline, __webpack_exports__bankers_round as bankers_round, __webpack_exports__cat as cat, __webpack_exports__cos_sim as cos_sim, __webpack_exports__dot as dot, __webpack_exports__dynamic_time_warping as dynamic_time_warping, __webpack_exports__env as env, __webpack_exports__full as full, __webpack_exports__full_like as full_like, __webpack_exports__getKeyValueShapes as getKeyValueShapes, __webpack_exports__hamming as hamming, __webpack_exports__hanning as hanning, __webpack_exports__interpolate as interpolate, __webpack_exports__interpolate_4d as interpolate_4d, __webpack_exports__interpolate_data as interpolate_data, __webpack_exports__is_chinese_char as is_chinese_char, __webpack_exports__layer_norm as layer_norm, __webpack_exports__log_softmax as log_softmax, __webpack_exports__magnitude as magnitude, __webpack_exports__matmul as matmul, __webpack_exports__max as max, __webpack_exports__mean as mean, __webpack_exports__mean_pooling as mean_pooling, __webpack_exports__medianFilter as medianFilter, __webpack_exports__mel_filter_bank as mel_filter_bank, __webpack_exports__min as min, __webpack_exports__ones as ones, __webpack_exports__ones_like as ones_like, __webpack_exports__permute as permute, __webpack_exports__permute_data as permute_data, __webpack_exports__pipeline as pipeline, __webpack_exports__quantize_embeddings as quantize_embeddings, __webpack_exports__read_audio as read_audio, __webpack_exports__rfft as rfft, __webpack_exports__round as round, __webpack_exports__softmax as softmax, __webpack_exports__spectrogram as spectrogram, __webpack_exports__stack as stack, __webpack_exports__std_mean as std_mean, __webpack_exports__topk as topk, __webpack_exports__window_function as window_function, __webpack_exports__zeros as zeros, __webpack_exports__zeros_like as zeros_like };
31423
31696
 
31424
31697
  //# sourceMappingURL=transformers.mjs.map