@huggingface/transformers 3.2.3 → 3.3.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +5 -3
- package/dist/ort-wasm-simd-threaded.jsep.wasm +0 -0
- package/dist/ort.bundle.min.mjs +2776 -0
- package/dist/transformers.cjs +792 -330
- package/dist/transformers.cjs.map +1 -1
- package/dist/transformers.js +1150 -656
- package/dist/transformers.js.map +1 -1
- package/dist/transformers.min.cjs +1 -1
- package/dist/transformers.min.cjs.map +1 -1
- package/dist/transformers.min.js +1 -1
- package/dist/transformers.min.js.map +1 -1
- package/dist/transformers.min.mjs +1 -1
- package/dist/transformers.min.mjs.map +1 -1
- package/dist/transformers.mjs +798 -331
- package/dist/transformers.mjs.map +1 -1
- package/package.json +3 -3
- package/src/base/feature_extraction_utils.js +9 -9
- package/src/base/image_processors_utils.js +12 -1
- package/src/base/processing_utils.js +24 -3
- package/src/configs.js +5 -0
- package/src/env.js +1 -2
- package/src/generation/streamers.js +5 -2
- package/src/models/auto/feature_extraction_auto.js +0 -16
- package/src/models/auto/processing_auto.js +0 -16
- package/src/models/convnext/image_processing_convnext.js +1 -0
- package/src/models/efficientnet/image_processing_efficientnet.js +1 -0
- package/src/models/florence2/processing_florence2.js +3 -0
- package/src/models/grounding_dino/image_processing_grounding_dino.js +29 -0
- package/src/models/grounding_dino/processing_grounding_dino.js +101 -0
- package/src/models/idefics3/image_processing_idefics3.js +2 -0
- package/src/models/image_processors.js +1 -0
- package/src/models/janus/image_processing_janus.js +1 -0
- package/src/models/mgp_str/processing_mgp_str.js +2 -0
- package/src/models/paligemma/processing_paligemma.js +1 -0
- package/src/models/phi3_v/processing_phi3_v.js +1 -1
- package/src/models/processors.js +3 -2
- package/src/models/pyannote/feature_extraction_pyannote.js +1 -0
- package/src/models/qwen2_vl/processing_qwen2_vl.js +1 -0
- package/src/models/seamless_m4t/feature_extraction_seamless_m4t.js +2 -2
- package/src/models/whisper/feature_extraction_whisper.js +1 -1
- package/src/models.js +72 -20
- package/src/ops/registry.js +10 -0
- package/src/pipelines.js +73 -23
- package/src/tokenizers.js +4 -7
- package/src/utils/audio.js +113 -1
- package/src/utils/core.js +26 -0
- package/src/utils/dtypes.js +2 -0
- package/src/utils/hub.js +1 -1
- package/src/utils/image.js +5 -18
- package/src/utils/maths.js +8 -6
- package/src/utils/tensor.js +134 -114
- package/types/base/feature_extraction_utils.d.ts +7 -7
- package/types/base/image_processors_utils.d.ts +7 -0
- package/types/base/image_processors_utils.d.ts.map +1 -1
- package/types/base/processing_utils.d.ts +25 -19
- package/types/base/processing_utils.d.ts.map +1 -1
- package/types/configs.d.ts.map +1 -1
- package/types/generation/parameters.d.ts +1 -1
- package/types/generation/streamers.d.ts +3 -1
- package/types/generation/streamers.d.ts.map +1 -1
- package/types/models/auto/feature_extraction_auto.d.ts.map +1 -1
- package/types/models/auto/image_processing_auto.d.ts.map +1 -1
- package/types/models/auto/processing_auto.d.ts.map +1 -1
- package/types/models/convnext/image_processing_convnext.d.ts.map +1 -1
- package/types/models/efficientnet/image_processing_efficientnet.d.ts.map +1 -1
- package/types/models/florence2/processing_florence2.d.ts.map +1 -1
- package/types/models/grounding_dino/image_processing_grounding_dino.d.ts +20 -0
- package/types/models/grounding_dino/image_processing_grounding_dino.d.ts.map +1 -0
- package/types/models/grounding_dino/processing_grounding_dino.d.ts +27 -0
- package/types/models/grounding_dino/processing_grounding_dino.d.ts.map +1 -0
- package/types/models/idefics3/image_processing_idefics3.d.ts.map +1 -1
- package/types/models/image_processors.d.ts +1 -0
- package/types/models/janus/image_processing_janus.d.ts.map +1 -1
- package/types/models/mgp_str/processing_mgp_str.d.ts.map +1 -1
- package/types/models/paligemma/processing_paligemma.d.ts.map +1 -1
- package/types/models/phi3_v/processing_phi3_v.d.ts +6 -2
- package/types/models/phi3_v/processing_phi3_v.d.ts.map +1 -1
- package/types/models/processors.d.ts +3 -2
- package/types/models/pyannote/feature_extraction_pyannote.d.ts.map +1 -1
- package/types/models/qwen2_vl/processing_qwen2_vl.d.ts.map +1 -1
- package/types/models/sapiens/image_processing_sapiens.d.ts +10 -0
- package/types/models/sapiens/image_processing_sapiens.d.ts.map +1 -0
- package/types/models/whisper/generation_whisper.d.ts +1 -1
- package/types/models/whisper/generation_whisper.d.ts.map +1 -1
- package/types/models.d.ts +40 -17
- package/types/models.d.ts.map +1 -1
- package/types/ops/registry.d.ts +1 -0
- package/types/ops/registry.d.ts.map +1 -1
- package/types/pipelines.d.ts +7 -12
- package/types/pipelines.d.ts.map +1 -1
- package/types/tokenizers.d.ts.map +1 -1
- package/types/tsconfig.tsbuildinfo +1 -0
- package/types/utils/audio.d.ts +25 -0
- package/types/utils/audio.d.ts.map +1 -1
- package/types/utils/core.d.ts +6 -0
- package/types/utils/core.d.ts.map +1 -1
- package/types/utils/dtypes.d.ts.map +1 -1
- package/types/utils/hub.d.ts +1 -1
- package/types/utils/hub.d.ts.map +1 -1
- package/types/utils/image.d.ts +3 -2
- package/types/utils/image.d.ts.map +1 -1
- package/types/utils/maths.d.ts +8 -6
- package/types/utils/maths.d.ts.map +1 -1
- package/types/utils/tensor.d.ts +22 -6
- package/types/utils/tensor.d.ts.map +1 -1
package/dist/transformers.mjs
CHANGED
|
@@ -4132,23 +4132,23 @@ class FeatureExtractor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_1__.Ca
|
|
|
4132
4132
|
}
|
|
4133
4133
|
|
|
4134
4134
|
/**
|
|
4135
|
-
* Instantiate one of the
|
|
4135
|
+
* Instantiate one of the feature extractor classes of the library from a pretrained model.
|
|
4136
4136
|
*
|
|
4137
|
-
* The
|
|
4138
|
-
*
|
|
4137
|
+
* The feature extractor class to instantiate is selected based on the `feature_extractor_type` property of
|
|
4138
|
+
* the config object (either passed as an argument or loaded from `pretrained_model_name_or_path` if possible)
|
|
4139
4139
|
*
|
|
4140
4140
|
* @param {string} pretrained_model_name_or_path The name or path of the pretrained model. Can be either:
|
|
4141
|
-
* - A string, the *model id* of a pretrained
|
|
4141
|
+
* - A string, the *model id* of a pretrained feature_extractor hosted inside a model repo on huggingface.co.
|
|
4142
4142
|
* Valid model ids can be located at the root-level, like `bert-base-uncased`, or namespaced under a
|
|
4143
4143
|
* user or organization name, like `dbmdz/bert-base-german-cased`.
|
|
4144
|
-
* - A path to a *directory* containing
|
|
4145
|
-
* @param {import('../utils/hub.js').PretrainedOptions} options Additional options for loading the
|
|
4144
|
+
* - A path to a *directory* containing feature_extractor files, e.g., `./my_model_directory/`.
|
|
4145
|
+
* @param {import('../utils/hub.js').PretrainedOptions} options Additional options for loading the feature_extractor.
|
|
4146
4146
|
*
|
|
4147
|
-
* @returns {Promise<FeatureExtractor>} A new instance of the
|
|
4147
|
+
* @returns {Promise<FeatureExtractor>} A new instance of the Feature Extractor class.
|
|
4148
4148
|
*/
|
|
4149
4149
|
static async from_pretrained(pretrained_model_name_or_path, options) {
|
|
4150
|
-
const
|
|
4151
|
-
return new this(
|
|
4150
|
+
const config = await (0,_utils_hub_js__WEBPACK_IMPORTED_MODULE_2__.getModelJSON)(pretrained_model_name_or_path, _utils_constants_js__WEBPACK_IMPORTED_MODULE_0__.FEATURE_EXTRACTOR_NAME, true, options);
|
|
4151
|
+
return new this(config);
|
|
4152
4152
|
}
|
|
4153
4153
|
}
|
|
4154
4154
|
|
|
@@ -4180,6 +4180,7 @@ function validate_audio_inputs(audio, feature_extractor) {
|
|
|
4180
4180
|
__webpack_require__.r(__webpack_exports__);
|
|
4181
4181
|
/* harmony export */ __webpack_require__.d(__webpack_exports__, {
|
|
4182
4182
|
/* harmony export */ ImageProcessor: () => (/* binding */ ImageProcessor),
|
|
4183
|
+
/* harmony export */ center_to_corners_format: () => (/* binding */ center_to_corners_format),
|
|
4183
4184
|
/* harmony export */ post_process_instance_segmentation: () => (/* binding */ post_process_instance_segmentation),
|
|
4184
4185
|
/* harmony export */ post_process_object_detection: () => (/* binding */ post_process_object_detection),
|
|
4185
4186
|
/* harmony export */ post_process_panoptic_segmentation: () => (/* binding */ post_process_panoptic_segmentation),
|
|
@@ -4798,14 +4799,20 @@ class ImageProcessor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_0__.Call
|
|
|
4798
4799
|
this.do_thumbnail = config.do_thumbnail;
|
|
4799
4800
|
this.size = config.size ?? config.image_size;
|
|
4800
4801
|
this.do_resize = config.do_resize ?? (this.size !== undefined);
|
|
4802
|
+
// @ts-expect-error TS2339
|
|
4801
4803
|
this.size_divisibility = config.size_divisibility ?? config.size_divisor;
|
|
4802
4804
|
|
|
4803
4805
|
this.do_center_crop = config.do_center_crop;
|
|
4806
|
+
// @ts-expect-error TS2339
|
|
4804
4807
|
this.crop_size = config.crop_size;
|
|
4808
|
+
// @ts-expect-error TS2339
|
|
4805
4809
|
this.do_convert_rgb = config.do_convert_rgb ?? true;
|
|
4810
|
+
// @ts-expect-error TS2339
|
|
4806
4811
|
this.do_crop_margin = config.do_crop_margin;
|
|
4807
4812
|
|
|
4813
|
+
// @ts-expect-error TS2339
|
|
4808
4814
|
this.pad_size = config.pad_size;
|
|
4815
|
+
// @ts-expect-error TS2339
|
|
4809
4816
|
this.do_pad = config.do_pad;
|
|
4810
4817
|
|
|
4811
4818
|
if (this.do_pad && !this.pad_size && this.size && this.size.width !== undefined && this.size.height !== undefined) {
|
|
@@ -5014,6 +5021,7 @@ class ImageProcessor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_0__.Call
|
|
|
5014
5021
|
// Support both formats for backwards compatibility
|
|
5015
5022
|
else if (Number.isInteger(size)) {
|
|
5016
5023
|
shortest_edge = size;
|
|
5024
|
+
// @ts-expect-error TS2339
|
|
5017
5025
|
longest_edge = this.config.max_size ?? shortest_edge;
|
|
5018
5026
|
|
|
5019
5027
|
} else if (size !== undefined) {
|
|
@@ -5082,6 +5090,7 @@ class ImageProcessor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_0__.Call
|
|
|
5082
5090
|
} else if (size.min_pixels !== undefined && size.max_pixels !== undefined) {
|
|
5083
5091
|
// Custom resize logic for Qwen2-VL models
|
|
5084
5092
|
const { min_pixels, max_pixels } = size;
|
|
5093
|
+
// @ts-expect-error TS2339
|
|
5085
5094
|
const factor = this.config.patch_size * this.config.merge_size;
|
|
5086
5095
|
return smart_resize(srcHeight, srcWidth, factor, min_pixels, max_pixels);
|
|
5087
5096
|
} else {
|
|
@@ -5097,6 +5106,7 @@ class ImageProcessor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_0__.Call
|
|
|
5097
5106
|
async resize(image) {
|
|
5098
5107
|
const [newWidth, newHeight] = this.get_resize_output_image_size(image, this.size);
|
|
5099
5108
|
return await image.resize(newWidth, newHeight, {
|
|
5109
|
+
// @ts-expect-error TS2322
|
|
5100
5110
|
resample: this.resample,
|
|
5101
5111
|
});
|
|
5102
5112
|
}
|
|
@@ -5147,6 +5157,7 @@ class ImageProcessor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_0__.Call
|
|
|
5147
5157
|
|
|
5148
5158
|
// Resize the image using thumbnail method.
|
|
5149
5159
|
if (this.do_thumbnail) {
|
|
5160
|
+
// @ts-expect-error TS2345
|
|
5150
5161
|
image = await this.thumbnail(image, this.size, this.resample);
|
|
5151
5162
|
}
|
|
5152
5163
|
|
|
@@ -5171,6 +5182,7 @@ class ImageProcessor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_0__.Call
|
|
|
5171
5182
|
// NOTE: All pixel-level manipulation (i.e., modifying `pixelData`)
|
|
5172
5183
|
// occurs with data in the hwc format (height, width, channels),
|
|
5173
5184
|
// to emulate the behavior of the original Python code (w/ numpy).
|
|
5185
|
+
/** @type {Float32Array} */
|
|
5174
5186
|
let pixelData = Float32Array.from(image.data);
|
|
5175
5187
|
let imgDims = [image.height, image.width, image.channels];
|
|
5176
5188
|
|
|
@@ -5328,6 +5340,7 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
5328
5340
|
/**
|
|
5329
5341
|
* @typedef {Object} ProcessorProperties Additional processor-specific properties.
|
|
5330
5342
|
* @typedef {import('../utils/hub.js').PretrainedOptions & ProcessorProperties} PretrainedProcessorOptions
|
|
5343
|
+
* @typedef {import('../tokenizers.js').PreTrainedTokenizer} PreTrainedTokenizer
|
|
5331
5344
|
*/
|
|
5332
5345
|
|
|
5333
5346
|
|
|
@@ -5361,7 +5374,7 @@ class Processor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_1__.Callable
|
|
|
5361
5374
|
}
|
|
5362
5375
|
|
|
5363
5376
|
/**
|
|
5364
|
-
* @returns {
|
|
5377
|
+
* @returns {PreTrainedTokenizer|undefined} The tokenizer of the processor, if it exists.
|
|
5365
5378
|
*/
|
|
5366
5379
|
get tokenizer() {
|
|
5367
5380
|
return this.components.tokenizer;
|
|
@@ -5374,6 +5387,11 @@ class Processor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_1__.Callable
|
|
|
5374
5387
|
return this.components.feature_extractor;
|
|
5375
5388
|
}
|
|
5376
5389
|
|
|
5390
|
+
/**
|
|
5391
|
+
* @param {Parameters<PreTrainedTokenizer['apply_chat_template']>[0]} messages
|
|
5392
|
+
* @param {Parameters<PreTrainedTokenizer['apply_chat_template']>[1]} options
|
|
5393
|
+
* @returns {ReturnType<PreTrainedTokenizer['apply_chat_template']>}
|
|
5394
|
+
*/
|
|
5377
5395
|
apply_chat_template(messages, options = {}) {
|
|
5378
5396
|
if (!this.tokenizer) {
|
|
5379
5397
|
throw new Error('Unable to apply chat template without a tokenizer.');
|
|
@@ -5384,6 +5402,10 @@ class Processor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_1__.Callable
|
|
|
5384
5402
|
});
|
|
5385
5403
|
}
|
|
5386
5404
|
|
|
5405
|
+
/**
|
|
5406
|
+
* @param {Parameters<PreTrainedTokenizer['batch_decode']>} args
|
|
5407
|
+
* @returns {ReturnType<PreTrainedTokenizer['batch_decode']>}
|
|
5408
|
+
*/
|
|
5387
5409
|
batch_decode(...args) {
|
|
5388
5410
|
if (!this.tokenizer) {
|
|
5389
5411
|
throw new Error('Unable to decode without a tokenizer.');
|
|
@@ -5391,6 +5413,17 @@ class Processor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_1__.Callable
|
|
|
5391
5413
|
return this.tokenizer.batch_decode(...args);
|
|
5392
5414
|
}
|
|
5393
5415
|
|
|
5416
|
+
/**
|
|
5417
|
+
* @param {Parameters<PreTrainedTokenizer['decode']>} args
|
|
5418
|
+
* @returns {ReturnType<PreTrainedTokenizer['decode']>}
|
|
5419
|
+
*/
|
|
5420
|
+
decode(...args) {
|
|
5421
|
+
if (!this.tokenizer) {
|
|
5422
|
+
throw new Error('Unable to decode without a tokenizer.');
|
|
5423
|
+
}
|
|
5424
|
+
return this.tokenizer.decode(...args);
|
|
5425
|
+
}
|
|
5426
|
+
|
|
5394
5427
|
|
|
5395
5428
|
/**
|
|
5396
5429
|
* Calls the feature_extractor function with the given input.
|
|
@@ -5411,8 +5444,8 @@ class Processor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_1__.Callable
|
|
|
5411
5444
|
/**
|
|
5412
5445
|
* Instantiate one of the processor classes of the library from a pretrained model.
|
|
5413
5446
|
*
|
|
5414
|
-
* The processor class to instantiate is selected based on the `
|
|
5415
|
-
* (either passed as an argument or loaded from `pretrained_model_name_or_path` if possible)
|
|
5447
|
+
* The processor class to instantiate is selected based on the `image_processor_type` (or `feature_extractor_type`; legacy)
|
|
5448
|
+
* property of the config object (either passed as an argument or loaded from `pretrained_model_name_or_path` if possible)
|
|
5416
5449
|
*
|
|
5417
5450
|
* @param {string} pretrained_model_name_or_path The name or path of the pretrained model. Can be either:
|
|
5418
5451
|
* - A string, the *model id* of a pretrained processor hosted inside a model repo on huggingface.co.
|
|
@@ -5531,15 +5564,19 @@ function getNormalizedConfig(config) {
|
|
|
5531
5564
|
case 'florence2':
|
|
5532
5565
|
case 'llava_onevision':
|
|
5533
5566
|
case 'idefics3':
|
|
5567
|
+
// @ts-expect-error TS2339
|
|
5534
5568
|
init_normalized_config = getNormalizedConfig(config.text_config);
|
|
5535
5569
|
break;
|
|
5536
5570
|
case 'moondream1':
|
|
5571
|
+
// @ts-expect-error TS2339
|
|
5537
5572
|
init_normalized_config = getNormalizedConfig(config.phi_config);
|
|
5538
5573
|
break;
|
|
5539
5574
|
case 'musicgen':
|
|
5575
|
+
// @ts-expect-error TS2339
|
|
5540
5576
|
init_normalized_config = getNormalizedConfig(config.decoder);
|
|
5541
5577
|
break;
|
|
5542
5578
|
case 'multi_modality':
|
|
5579
|
+
// @ts-expect-error TS2339
|
|
5543
5580
|
init_normalized_config = getNormalizedConfig(config.language_config);
|
|
5544
5581
|
break;
|
|
5545
5582
|
|
|
@@ -5660,6 +5697,7 @@ function getNormalizedConfig(config) {
|
|
|
5660
5697
|
break;
|
|
5661
5698
|
|
|
5662
5699
|
case 'vision-encoder-decoder':
|
|
5700
|
+
// @ts-expect-error TS2339
|
|
5663
5701
|
const decoderConfig = getNormalizedConfig(config.decoder);
|
|
5664
5702
|
|
|
5665
5703
|
const add_encoder_pkv = 'num_decoder_layers' in decoderConfig;
|
|
@@ -5902,7 +5940,7 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
5902
5940
|
|
|
5903
5941
|
|
|
5904
5942
|
|
|
5905
|
-
const VERSION = '3.
|
|
5943
|
+
const VERSION = '3.3.0';
|
|
5906
5944
|
|
|
5907
5945
|
// Check if various APIs are available (depends on environment)
|
|
5908
5946
|
const IS_BROWSER_ENV = typeof window !== "undefined" && typeof window.document !== "undefined";
|
|
@@ -6038,7 +6076,6 @@ function isEmpty(obj) {
|
|
|
6038
6076
|
}
|
|
6039
6077
|
|
|
6040
6078
|
|
|
6041
|
-
|
|
6042
6079
|
/***/ }),
|
|
6043
6080
|
|
|
6044
6081
|
/***/ "./src/generation/configuration_utils.js":
|
|
@@ -7648,6 +7685,7 @@ class TextStreamer extends BaseStreamer {
|
|
|
7648
7685
|
* @param {import('../tokenizers.js').PreTrainedTokenizer} tokenizer
|
|
7649
7686
|
* @param {Object} options
|
|
7650
7687
|
* @param {boolean} [options.skip_prompt=false] Whether to skip the prompt tokens
|
|
7688
|
+
* @param {boolean} [options.skip_special_tokens=true] Whether to skip special tokens when decoding
|
|
7651
7689
|
* @param {function(string): void} [options.callback_function=null] Function to call when a piece of text is ready to display
|
|
7652
7690
|
* @param {function(bigint[]): void} [options.token_callback_function=null] Function to call when a new token is generated
|
|
7653
7691
|
* @param {Object} [options.decode_kwargs={}] Additional keyword arguments to pass to the tokenizer's decode method
|
|
@@ -7656,6 +7694,7 @@ class TextStreamer extends BaseStreamer {
|
|
|
7656
7694
|
skip_prompt = false,
|
|
7657
7695
|
callback_function = null,
|
|
7658
7696
|
token_callback_function = null,
|
|
7697
|
+
skip_special_tokens = true,
|
|
7659
7698
|
decode_kwargs = {},
|
|
7660
7699
|
...kwargs
|
|
7661
7700
|
} = {}) {
|
|
@@ -7664,7 +7703,7 @@ class TextStreamer extends BaseStreamer {
|
|
|
7664
7703
|
this.skip_prompt = skip_prompt;
|
|
7665
7704
|
this.callback_function = callback_function ?? stdout_write;
|
|
7666
7705
|
this.token_callback_function = token_callback_function;
|
|
7667
|
-
this.decode_kwargs = { ...decode_kwargs, ...kwargs };
|
|
7706
|
+
this.decode_kwargs = { skip_special_tokens, ...decode_kwargs, ...kwargs };
|
|
7668
7707
|
|
|
7669
7708
|
// variables used in the streaming process
|
|
7670
7709
|
this.token_cache = [];
|
|
@@ -7780,9 +7819,10 @@ class WhisperTextStreamer extends TextStreamer {
|
|
|
7780
7819
|
} = {}) {
|
|
7781
7820
|
super(tokenizer, {
|
|
7782
7821
|
skip_prompt,
|
|
7822
|
+
skip_special_tokens,
|
|
7783
7823
|
callback_function,
|
|
7784
7824
|
token_callback_function,
|
|
7785
|
-
decode_kwargs
|
|
7825
|
+
decode_kwargs,
|
|
7786
7826
|
});
|
|
7787
7827
|
this.timestamp_begin = tokenizer.timestamp_begin;
|
|
7788
7828
|
|
|
@@ -8035,6 +8075,8 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
8035
8075
|
/* harmony export */ GraniteForCausalLM: () => (/* binding */ GraniteForCausalLM),
|
|
8036
8076
|
/* harmony export */ GraniteModel: () => (/* binding */ GraniteModel),
|
|
8037
8077
|
/* harmony export */ GranitePreTrainedModel: () => (/* binding */ GranitePreTrainedModel),
|
|
8078
|
+
/* harmony export */ GroundingDinoForObjectDetection: () => (/* binding */ GroundingDinoForObjectDetection),
|
|
8079
|
+
/* harmony export */ GroundingDinoPreTrainedModel: () => (/* binding */ GroundingDinoPreTrainedModel),
|
|
8038
8080
|
/* harmony export */ GroupViTModel: () => (/* binding */ GroupViTModel),
|
|
8039
8081
|
/* harmony export */ GroupViTPreTrainedModel: () => (/* binding */ GroupViTPreTrainedModel),
|
|
8040
8082
|
/* harmony export */ HieraForImageClassification: () => (/* binding */ HieraForImageClassification),
|
|
@@ -8243,6 +8285,8 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
8243
8285
|
/* harmony export */ Starcoder2ForCausalLM: () => (/* binding */ Starcoder2ForCausalLM),
|
|
8244
8286
|
/* harmony export */ Starcoder2Model: () => (/* binding */ Starcoder2Model),
|
|
8245
8287
|
/* harmony export */ Starcoder2PreTrainedModel: () => (/* binding */ Starcoder2PreTrainedModel),
|
|
8288
|
+
/* harmony export */ StyleTextToSpeech2Model: () => (/* binding */ StyleTextToSpeech2Model),
|
|
8289
|
+
/* harmony export */ StyleTextToSpeech2PreTrainedModel: () => (/* binding */ StyleTextToSpeech2PreTrainedModel),
|
|
8246
8290
|
/* harmony export */ Swin2SRForImageSuperResolution: () => (/* binding */ Swin2SRForImageSuperResolution),
|
|
8247
8291
|
/* harmony export */ Swin2SRModel: () => (/* binding */ Swin2SRModel),
|
|
8248
8292
|
/* harmony export */ Swin2SRPreTrainedModel: () => (/* binding */ Swin2SRPreTrainedModel),
|
|
@@ -8558,8 +8602,11 @@ async function getSession(pretrained_model_name_or_path, fileName, options) {
|
|
|
8558
8602
|
} else if (session_options.externalData !== undefined) {
|
|
8559
8603
|
externalDataPromises = session_options.externalData.map(async (ext) => {
|
|
8560
8604
|
// if the external data is a string, fetch the file and replace the string with its content
|
|
8605
|
+
// @ts-expect-error TS2339
|
|
8561
8606
|
if (typeof ext.data === "string") {
|
|
8607
|
+
// @ts-expect-error TS2339
|
|
8562
8608
|
const ext_buffer = await (0,_utils_hub_js__WEBPACK_IMPORTED_MODULE_5__.getModelFile)(pretrained_model_name_or_path, ext.data, true, options);
|
|
8609
|
+
// @ts-expect-error TS2698
|
|
8563
8610
|
return { ...ext, data: ext_buffer };
|
|
8564
8611
|
}
|
|
8565
8612
|
return ext;
|
|
@@ -8817,14 +8864,23 @@ async function encoderForward(self, model_inputs) {
|
|
|
8817
8864
|
encoderFeeds.inputs_embeds = await self.encode_text({ input_ids: model_inputs.input_ids });
|
|
8818
8865
|
}
|
|
8819
8866
|
if (session.inputNames.includes('token_type_ids') && !encoderFeeds.token_type_ids) {
|
|
8867
|
+
if (!encoderFeeds.input_ids) {
|
|
8868
|
+
throw new Error('Both `input_ids` and `token_type_ids` are missing in the model inputs.');
|
|
8869
|
+
}
|
|
8820
8870
|
// Assign default `token_type_ids` (all zeroes) to the `encoderFeeds` if the model expects it,
|
|
8821
8871
|
// but they weren't created by the tokenizer.
|
|
8822
|
-
encoderFeeds.token_type_ids =
|
|
8823
|
-
'int64',
|
|
8824
|
-
new BigInt64Array(encoderFeeds.input_ids.data.length),
|
|
8825
|
-
encoderFeeds.input_ids.dims
|
|
8826
|
-
)
|
|
8872
|
+
encoderFeeds.token_type_ids = (0,_utils_tensor_js__WEBPACK_IMPORTED_MODULE_9__.zeros_like)(encoderFeeds.input_ids);
|
|
8827
8873
|
}
|
|
8874
|
+
if (session.inputNames.includes('pixel_mask') && !encoderFeeds.pixel_mask) {
|
|
8875
|
+
if (!encoderFeeds.pixel_values) {
|
|
8876
|
+
throw new Error('Both `pixel_values` and `pixel_mask` are missing in the model inputs.');
|
|
8877
|
+
}
|
|
8878
|
+
// Assign default `pixel_mask` (all ones) to the `encoderFeeds` if the model expects it,
|
|
8879
|
+
// but they weren't created by the processor.
|
|
8880
|
+
const dims = encoderFeeds.pixel_values.dims;
|
|
8881
|
+
encoderFeeds.pixel_mask = (0,_utils_tensor_js__WEBPACK_IMPORTED_MODULE_9__.ones)([dims[0], dims[2], dims[3]]);
|
|
8882
|
+
}
|
|
8883
|
+
|
|
8828
8884
|
return await sessionRun(session, encoderFeeds);
|
|
8829
8885
|
}
|
|
8830
8886
|
|
|
@@ -9807,6 +9863,7 @@ class PreTrainedModel extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_3__.Cal
|
|
|
9807
9863
|
if (this.config.model_type === 'musicgen') {
|
|
9808
9864
|
// Custom logic (TODO: move to Musicgen class)
|
|
9809
9865
|
decoder_input_ids = Array.from({
|
|
9866
|
+
// @ts-expect-error TS2339
|
|
9810
9867
|
length: batch_size * this.config.decoder.num_codebooks
|
|
9811
9868
|
}, () => [decoder_start_token_id]);
|
|
9812
9869
|
|
|
@@ -10136,11 +10193,13 @@ class PreTrainedModel extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_3__.Cal
|
|
|
10136
10193
|
async encode_image({ pixel_values }) {
|
|
10137
10194
|
// image_inputs === { pixel_values }
|
|
10138
10195
|
const features = (await sessionRun(this.sessions['vision_encoder'], { pixel_values })).image_features;
|
|
10196
|
+
// @ts-expect-error TS2339
|
|
10139
10197
|
if (!this.config.num_image_tokens) {
|
|
10140
10198
|
console.warn(
|
|
10141
10199
|
'The number of image tokens was not set in the model configuration. ' +
|
|
10142
10200
|
`Setting it to the number of features detected by the vision encoder (${features.dims[1]}).`
|
|
10143
10201
|
)
|
|
10202
|
+
// @ts-expect-error TS2339
|
|
10144
10203
|
this.config.num_image_tokens = features.dims[1];
|
|
10145
10204
|
}
|
|
10146
10205
|
return features;
|
|
@@ -11568,6 +11627,7 @@ class WhisperForConditionalGeneration extends WhisperPreTrainedModel {
|
|
|
11568
11627
|
|
|
11569
11628
|
if (generation_config.return_token_timestamps) {
|
|
11570
11629
|
outputs["token_timestamps"] = this._extract_token_timestamps(
|
|
11630
|
+
// @ts-expect-error TS2345
|
|
11571
11631
|
outputs,
|
|
11572
11632
|
generation_config.alignment_heads,
|
|
11573
11633
|
generation_config.num_frames,
|
|
@@ -11603,6 +11663,7 @@ class WhisperForConditionalGeneration extends WhisperPreTrainedModel {
|
|
|
11603
11663
|
);
|
|
11604
11664
|
}
|
|
11605
11665
|
|
|
11666
|
+
// @ts-expect-error TS2339
|
|
11606
11667
|
let median_filter_width = this.config.median_filter_width;
|
|
11607
11668
|
if (median_filter_width === undefined) {
|
|
11608
11669
|
console.warn("Model config has no `median_filter_width`, using default value of 7.")
|
|
@@ -11613,6 +11674,7 @@ class WhisperForConditionalGeneration extends WhisperPreTrainedModel {
|
|
|
11613
11674
|
const batch = generate_outputs.cross_attentions;
|
|
11614
11675
|
// Create a list with `decoder_layers` elements, each a tensor of shape
|
|
11615
11676
|
// (batch size, attention_heads, output length, input length).
|
|
11677
|
+
// @ts-expect-error TS2339
|
|
11616
11678
|
const cross_attentions = Array.from({ length: this.config.decoder_layers },
|
|
11617
11679
|
// Concatenate the cross attentions for each layer across sequence length dimension.
|
|
11618
11680
|
(_, i) => (0,_utils_tensor_js__WEBPACK_IMPORTED_MODULE_9__.cat)(batch.map(x => x[i]), 2)
|
|
@@ -11756,6 +11818,7 @@ class LlavaForConditionalGeneration extends LlavaPreTrainedModel {
|
|
|
11756
11818
|
attention_mask,
|
|
11757
11819
|
}) {
|
|
11758
11820
|
|
|
11821
|
+
// @ts-expect-error TS2339
|
|
11759
11822
|
const image_token_index = this.config.image_token_index;
|
|
11760
11823
|
|
|
11761
11824
|
const idsList = input_ids.tolist();
|
|
@@ -12741,6 +12804,7 @@ class Qwen2VLForConditionalGeneration extends Qwen2VLPreTrainedModel {
|
|
|
12741
12804
|
const image_nums = vision_tokens.filter(x => x == image_token_id).length;
|
|
12742
12805
|
const video_nums = vision_tokens.filter(x => x == video_token_id).length;
|
|
12743
12806
|
|
|
12807
|
+
/** @type {number[][]} */
|
|
12744
12808
|
let llm_pos_ids_list = [];
|
|
12745
12809
|
let st = 0;
|
|
12746
12810
|
let remain_images = image_nums;
|
|
@@ -12810,6 +12874,7 @@ class Qwen2VLForConditionalGeneration extends Qwen2VLPreTrainedModel {
|
|
|
12810
12874
|
// NOTE: Each item in llm_pos_ids_list is an array of shape (3, text_len),
|
|
12811
12875
|
// meaning to perform concatenation along dim=1, we can do the following:
|
|
12812
12876
|
const num_items = llm_pos_ids_list.reduce((acc, x) => acc + x.length, 0);
|
|
12877
|
+
/** @type {number[]} */
|
|
12813
12878
|
const llm_positions = new Array(num_items);
|
|
12814
12879
|
let index = 0;
|
|
12815
12880
|
for (let x = 0; x < 3; ++x) {
|
|
@@ -12850,9 +12915,10 @@ class Qwen2VLForConditionalGeneration extends Qwen2VLPreTrainedModel {
|
|
|
12850
12915
|
{ length: 3 * data.length },
|
|
12851
12916
|
(_, i) => data[i % data.length]
|
|
12852
12917
|
);
|
|
12918
|
+
/** @type {bigint[]} */
|
|
12853
12919
|
const mrope_position_deltas = Array.from(
|
|
12854
12920
|
{ length: dims[0] },
|
|
12855
|
-
(_, i) => (0,_utils_maths_js__WEBPACK_IMPORTED_MODULE_11__.max)(data.subarray(dims[1] * i, dims[1] * (i + 1)))[0] +
|
|
12921
|
+
(_, i) => (0,_utils_maths_js__WEBPACK_IMPORTED_MODULE_11__.max)(data.subarray(dims[1] * i, dims[1] * (i + 1)))[0] + 1n + BigInt(dims[1])
|
|
12856
12922
|
);
|
|
12857
12923
|
|
|
12858
12924
|
return [
|
|
@@ -13423,7 +13489,7 @@ class DPTModel extends DPTPreTrainedModel { }
|
|
|
13423
13489
|
*
|
|
13424
13490
|
* **Example:** Depth estimation w/ `Xenova/dpt-hybrid-midas`.
|
|
13425
13491
|
* ```javascript
|
|
13426
|
-
* import { DPTForDepthEstimation, AutoProcessor, RawImage,
|
|
13492
|
+
* import { DPTForDepthEstimation, AutoProcessor, RawImage, interpolate_4d } from '@huggingface/transformers';
|
|
13427
13493
|
*
|
|
13428
13494
|
* // Load model and processor
|
|
13429
13495
|
* const model_id = 'Xenova/dpt-hybrid-midas';
|
|
@@ -13432,7 +13498,7 @@ class DPTModel extends DPTPreTrainedModel { }
|
|
|
13432
13498
|
*
|
|
13433
13499
|
* // Load image from URL
|
|
13434
13500
|
* const url = 'http://images.cocodataset.org/val2017/000000039769.jpg';
|
|
13435
|
-
* const image = await RawImage.
|
|
13501
|
+
* const image = await RawImage.read(url);
|
|
13436
13502
|
*
|
|
13437
13503
|
* // Prepare image for the model
|
|
13438
13504
|
* const inputs = await processor(image);
|
|
@@ -13441,10 +13507,15 @@ class DPTModel extends DPTPreTrainedModel { }
|
|
|
13441
13507
|
* const { predicted_depth } = await model(inputs);
|
|
13442
13508
|
*
|
|
13443
13509
|
* // Interpolate to original size
|
|
13444
|
-
* const prediction =
|
|
13510
|
+
* const prediction = (await interpolate_4d(predicted_depth.unsqueeze(1), {
|
|
13511
|
+
* size: image.size.reverse(),
|
|
13512
|
+
* mode: 'bilinear',
|
|
13513
|
+
* })).squeeze(1);
|
|
13445
13514
|
*
|
|
13446
13515
|
* // Visualize the prediction
|
|
13447
|
-
* const
|
|
13516
|
+
* const min = prediction.min().item();
|
|
13517
|
+
* const max = prediction.max().item();
|
|
13518
|
+
* const formatted = prediction.sub_(min).div_(max - min).mul_(255).to('uint8');
|
|
13448
13519
|
* const depth = RawImage.fromTensor(formatted);
|
|
13449
13520
|
* // RawImage {
|
|
13450
13521
|
* // data: Uint8Array(307200) [ 85, 85, 84, ... ],
|
|
@@ -13494,11 +13565,7 @@ class GLPNPreTrainedModel extends PreTrainedModel { }
|
|
|
13494
13565
|
class GLPNModel extends GLPNPreTrainedModel { }
|
|
13495
13566
|
|
|
13496
13567
|
/**
|
|
13497
|
-
*
|
|
13498
|
-
*
|
|
13499
|
-
* **Example:** Depth estimation w/ `Xenova/glpn-kitti`.
|
|
13500
|
-
* ```javascript
|
|
13501
|
-
* import { GLPNForDepthEstimation, AutoProcessor, RawImage, interpolate, max } from '@huggingface/transformers';
|
|
13568
|
+
* import { GLPNForDepthEstimation, AutoProcessor, RawImage, interpolate_4d } from '@huggingface/transformers';
|
|
13502
13569
|
*
|
|
13503
13570
|
* // Load model and processor
|
|
13504
13571
|
* const model_id = 'Xenova/glpn-kitti';
|
|
@@ -13507,7 +13574,7 @@ class GLPNModel extends GLPNPreTrainedModel { }
|
|
|
13507
13574
|
*
|
|
13508
13575
|
* // Load image from URL
|
|
13509
13576
|
* const url = 'http://images.cocodataset.org/val2017/000000039769.jpg';
|
|
13510
|
-
* const image = await RawImage.
|
|
13577
|
+
* const image = await RawImage.read(url);
|
|
13511
13578
|
*
|
|
13512
13579
|
* // Prepare image for the model
|
|
13513
13580
|
* const inputs = await processor(image);
|
|
@@ -13516,13 +13583,18 @@ class GLPNModel extends GLPNPreTrainedModel { }
|
|
|
13516
13583
|
* const { predicted_depth } = await model(inputs);
|
|
13517
13584
|
*
|
|
13518
13585
|
* // Interpolate to original size
|
|
13519
|
-
* const prediction =
|
|
13586
|
+
* const prediction = (await interpolate_4d(predicted_depth.unsqueeze(1), {
|
|
13587
|
+
* size: image.size.reverse(),
|
|
13588
|
+
* mode: 'bilinear',
|
|
13589
|
+
* })).squeeze(1);
|
|
13520
13590
|
*
|
|
13521
13591
|
* // Visualize the prediction
|
|
13522
|
-
* const
|
|
13592
|
+
* const min = prediction.min().item();
|
|
13593
|
+
* const max = prediction.max().item();
|
|
13594
|
+
* const formatted = prediction.sub_(min).div_(max - min).mul_(255).to('uint8');
|
|
13523
13595
|
* const depth = RawImage.fromTensor(formatted);
|
|
13524
13596
|
* // RawImage {
|
|
13525
|
-
* // data: Uint8Array(307200) [
|
|
13597
|
+
* // data: Uint8Array(307200) [ 85, 85, 84, ... ],
|
|
13526
13598
|
* // width: 640,
|
|
13527
13599
|
* // height: 480,
|
|
13528
13600
|
* // channels: 1
|
|
@@ -13697,6 +13769,8 @@ class Dinov2WithRegistersForImageClassification extends Dinov2WithRegistersPreTr
|
|
|
13697
13769
|
}
|
|
13698
13770
|
}
|
|
13699
13771
|
//////////////////////////////////////////////////
|
|
13772
|
+
class GroundingDinoPreTrainedModel extends PreTrainedModel { }
|
|
13773
|
+
class GroundingDinoForObjectDetection extends GroundingDinoPreTrainedModel { }
|
|
13700
13774
|
|
|
13701
13775
|
//////////////////////////////////////////////////
|
|
13702
13776
|
class YolosPreTrainedModel extends PreTrainedModel { }
|
|
@@ -14395,6 +14469,9 @@ class WavLMForAudioFrameClassification extends WavLMPreTrainedModel {
|
|
|
14395
14469
|
}
|
|
14396
14470
|
}
|
|
14397
14471
|
|
|
14472
|
+
class StyleTextToSpeech2PreTrainedModel extends PreTrainedModel { }
|
|
14473
|
+
class StyleTextToSpeech2Model extends StyleTextToSpeech2PreTrainedModel { }
|
|
14474
|
+
|
|
14398
14475
|
//////////////////////////////////////////////////
|
|
14399
14476
|
// SpeechT5 models
|
|
14400
14477
|
/**
|
|
@@ -14489,10 +14566,12 @@ class SpeechT5ForTextToSpeech extends SpeechT5PreTrainedModel {
|
|
|
14489
14566
|
|
|
14490
14567
|
const { encoder_outputs, encoder_attention_mask } = await encoderForward(this, model_inputs);
|
|
14491
14568
|
|
|
14569
|
+
// @ts-expect-error TS2339
|
|
14492
14570
|
const r = encoder_outputs.dims[1] / this.config.reduction_factor;
|
|
14493
14571
|
const maxlen = Math.floor(r * maxlenratio);
|
|
14494
14572
|
const minlen = Math.floor(r * minlenratio);
|
|
14495
14573
|
|
|
14574
|
+
// @ts-expect-error TS2339
|
|
14496
14575
|
const num_mel_bins = this.config.num_mel_bins;
|
|
14497
14576
|
|
|
14498
14577
|
let spectrogramParts = [];
|
|
@@ -14857,11 +14936,13 @@ class MusicgenForConditionalGeneration extends PreTrainedModel { // NOTE: not Mu
|
|
|
14857
14936
|
*/
|
|
14858
14937
|
_apply_and_filter_by_delay_pattern_mask(outputs) {
|
|
14859
14938
|
const [bs_x_codebooks, seqLength] = outputs.dims;
|
|
14939
|
+
// @ts-expect-error TS2339
|
|
14860
14940
|
const num_codebooks = this.config.decoder.num_codebooks;
|
|
14861
14941
|
const upperBound = (seqLength - num_codebooks);
|
|
14862
14942
|
|
|
14863
14943
|
let newDataSize = 0;
|
|
14864
14944
|
for (let i = 0; i < outputs.size; ++i) {
|
|
14945
|
+
// @ts-expect-error TS2339
|
|
14865
14946
|
if (outputs.data[i] === this.config.decoder.pad_token_id) {
|
|
14866
14947
|
continue;
|
|
14867
14948
|
}
|
|
@@ -14891,7 +14972,9 @@ class MusicgenForConditionalGeneration extends PreTrainedModel { // NOTE: not Mu
|
|
|
14891
14972
|
let clonedInputIds = structuredClone(input_ids);
|
|
14892
14973
|
for (let i = 0; i < clonedInputIds.length; ++i) {
|
|
14893
14974
|
for (let j = 0; j < clonedInputIds[i].length; ++j) {
|
|
14975
|
+
// @ts-expect-error TS2339
|
|
14894
14976
|
if ((i % this.config.decoder.num_codebooks) >= j) {
|
|
14977
|
+
// @ts-expect-error TS2339
|
|
14895
14978
|
clonedInputIds[i][j] = BigInt(this.config.decoder.pad_token_id);
|
|
14896
14979
|
}
|
|
14897
14980
|
}
|
|
@@ -15048,6 +15131,9 @@ class MultiModalityCausalLM extends MultiModalityPreTrainedModel {
|
|
|
15048
15131
|
'past_key_values',
|
|
15049
15132
|
];
|
|
15050
15133
|
|
|
15134
|
+
/**
|
|
15135
|
+
* @param {ConstructorParameters<typeof MultiModalityPreTrainedModel>} args
|
|
15136
|
+
*/
|
|
15051
15137
|
constructor(...args) {
|
|
15052
15138
|
super(...args);
|
|
15053
15139
|
|
|
@@ -15349,6 +15435,8 @@ const MODEL_MAPPING_NAMES_ENCODER_ONLY = new Map([
|
|
|
15349
15435
|
|
|
15350
15436
|
['maskformer', ['MaskFormerModel', MaskFormerModel]],
|
|
15351
15437
|
['mgp-str', ['MgpstrForSceneTextRecognition', MgpstrForSceneTextRecognition]],
|
|
15438
|
+
|
|
15439
|
+
['style_text_to_speech_2', ['StyleTextToSpeech2Model', StyleTextToSpeech2Model]],
|
|
15352
15440
|
]);
|
|
15353
15441
|
|
|
15354
15442
|
const MODEL_MAPPING_NAMES_ENCODER_DECODER = new Map([
|
|
@@ -15593,6 +15681,7 @@ const MODEL_FOR_OBJECT_DETECTION_MAPPING_NAMES = new Map([
|
|
|
15593
15681
|
const MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING_NAMES = new Map([
|
|
15594
15682
|
['owlvit', ['OwlViTForObjectDetection', OwlViTForObjectDetection]],
|
|
15595
15683
|
['owlv2', ['Owlv2ForObjectDetection', Owlv2ForObjectDetection]],
|
|
15684
|
+
['grounding-dino', ['GroundingDinoForObjectDetection', GroundingDinoForObjectDetection]],
|
|
15596
15685
|
]);
|
|
15597
15686
|
|
|
15598
15687
|
const MODEL_FOR_IMAGE_SEGMENTATION_MAPPING_NAMES = new Map([
|
|
@@ -16016,10 +16105,17 @@ class SequenceClassifierOutput extends ModelOutput {
|
|
|
16016
16105
|
/**
|
|
16017
16106
|
* @param {Object} output The output of the model.
|
|
16018
16107
|
* @param {Tensor} output.logits classification (or regression if config.num_labels==1) scores (before SoftMax).
|
|
16108
|
+
* @param {Record<string, Tensor>} [output.attentions] Object of `torch.FloatTensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, sequence_length)`.
|
|
16109
|
+
* Attentions weights after the attention softmax, used to compute the weighted average in the self-attention heads.
|
|
16019
16110
|
*/
|
|
16020
|
-
constructor({ logits }) {
|
|
16111
|
+
constructor({ logits, ...attentions }) {
|
|
16021
16112
|
super();
|
|
16022
16113
|
this.logits = logits;
|
|
16114
|
+
const attentions_list = Object.values(attentions);
|
|
16115
|
+
if (attentions_list.length > 0) {
|
|
16116
|
+
// Only set attentions if they are not empty
|
|
16117
|
+
this.attentions = attentions_list;
|
|
16118
|
+
}
|
|
16023
16119
|
}
|
|
16024
16120
|
}
|
|
16025
16121
|
|
|
@@ -16275,22 +16371,6 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
16275
16371
|
|
|
16276
16372
|
class AutoFeatureExtractor {
|
|
16277
16373
|
|
|
16278
|
-
/**
|
|
16279
|
-
* Instantiate one of the feature extractor classes of the library from a pretrained model.
|
|
16280
|
-
*
|
|
16281
|
-
* The processor class to instantiate is selected based on the `feature_extractor_type` property of
|
|
16282
|
-
* the config object (either passed as an argument or loaded from `pretrained_model_name_or_path` if possible)
|
|
16283
|
-
*
|
|
16284
|
-
* @param {string} pretrained_model_name_or_path The name or path of the pretrained model. Can be either:
|
|
16285
|
-
* - A string, the *model id* of a pretrained processor hosted inside a model repo on huggingface.co.
|
|
16286
|
-
* Valid model ids can be located at the root-level, like `bert-base-uncased`, or namespaced under a
|
|
16287
|
-
* user or organization name, like `dbmdz/bert-base-german-cased`.
|
|
16288
|
-
* - A path to a *directory* containing processor files, e.g., `./my_model_directory/`.
|
|
16289
|
-
* @param {import('../../utils/hub.js').PretrainedOptions} options Additional options for loading the processor.
|
|
16290
|
-
*
|
|
16291
|
-
* @returns {Promise<AllFeatureExtractors.ImageProcessor>} A new instance of the Processor class.
|
|
16292
|
-
*/
|
|
16293
|
-
|
|
16294
16374
|
/** @type {typeof FeatureExtractor.from_pretrained} */
|
|
16295
16375
|
static async from_pretrained(pretrained_model_name_or_path, options={}) {
|
|
16296
16376
|
|
|
@@ -16417,22 +16497,6 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
16417
16497
|
*/
|
|
16418
16498
|
class AutoProcessor {
|
|
16419
16499
|
|
|
16420
|
-
/**
|
|
16421
|
-
* Instantiate one of the processor classes of the library from a pretrained model.
|
|
16422
|
-
*
|
|
16423
|
-
* The processor class to instantiate is selected based on the `image_processor_type` (or `feature_extractor_type`; legacy)
|
|
16424
|
-
* property of the config object (either passed as an argument or loaded from `pretrained_model_name_or_path` if possible)
|
|
16425
|
-
*
|
|
16426
|
-
* @param {string} pretrained_model_name_or_path The name or path of the pretrained model. Can be either:
|
|
16427
|
-
* - A string, the *model id* of a pretrained processor hosted inside a model repo on huggingface.co.
|
|
16428
|
-
* Valid model ids can be located at the root-level, like `bert-base-uncased`, or namespaced under a
|
|
16429
|
-
* user or organization name, like `dbmdz/bert-base-german-cased`.
|
|
16430
|
-
* - A path to a *directory* containing processor files, e.g., `./my_model_directory/`.
|
|
16431
|
-
* @param {import('../../utils/hub.js').PretrainedOptions} options Additional options for loading the processor.
|
|
16432
|
-
*
|
|
16433
|
-
* @returns {Promise<Processor>} A new instance of the Processor class.
|
|
16434
|
-
*/
|
|
16435
|
-
|
|
16436
16500
|
/** @type {typeof Processor.from_pretrained} */
|
|
16437
16501
|
static async from_pretrained(pretrained_model_name_or_path, options={}) {
|
|
16438
16502
|
|
|
@@ -16750,6 +16814,7 @@ class ConvNextImageProcessor extends _base_image_processors_utils_js__WEBPACK_IM
|
|
|
16750
16814
|
/**
|
|
16751
16815
|
* Percentage of the image to crop. Only has an effect if this.size < 384.
|
|
16752
16816
|
*/
|
|
16817
|
+
// @ts-expect-error TS2339
|
|
16753
16818
|
this.crop_pct = this.config.crop_pct ?? (224 / 256);
|
|
16754
16819
|
}
|
|
16755
16820
|
|
|
@@ -16952,6 +17017,7 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
16952
17017
|
class EfficientNetImageProcessor extends _base_image_processors_utils_js__WEBPACK_IMPORTED_MODULE_0__.ImageProcessor {
|
|
16953
17018
|
constructor(config) {
|
|
16954
17019
|
super(config);
|
|
17020
|
+
// @ts-expect-error TS2339
|
|
16955
17021
|
this.include_top = this.config.include_top ?? true;
|
|
16956
17022
|
if (this.include_top) {
|
|
16957
17023
|
this.image_std = this.image_std.map(x => x * x);
|
|
@@ -17033,8 +17099,11 @@ class Florence2Processor extends _base_processing_utils_js__WEBPACK_IMPORTED_MOD
|
|
|
17033
17099
|
super(config, components);
|
|
17034
17100
|
|
|
17035
17101
|
const {
|
|
17102
|
+
// @ts-expect-error TS2339
|
|
17036
17103
|
tasks_answer_post_processing_type,
|
|
17104
|
+
// @ts-expect-error TS2339
|
|
17037
17105
|
task_prompts_without_inputs,
|
|
17106
|
+
// @ts-expect-error TS2339
|
|
17038
17107
|
task_prompts_with_input,
|
|
17039
17108
|
} = this.image_processor.config;
|
|
17040
17109
|
|
|
@@ -17169,6 +17238,168 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
17169
17238
|
class GLPNFeatureExtractor extends _base_image_processors_utils_js__WEBPACK_IMPORTED_MODULE_0__.ImageProcessor { }
|
|
17170
17239
|
|
|
17171
17240
|
|
|
17241
|
+
/***/ }),
|
|
17242
|
+
|
|
17243
|
+
/***/ "./src/models/grounding_dino/image_processing_grounding_dino.js":
|
|
17244
|
+
/*!**********************************************************************!*\
|
|
17245
|
+
!*** ./src/models/grounding_dino/image_processing_grounding_dino.js ***!
|
|
17246
|
+
\**********************************************************************/
|
|
17247
|
+
/***/ ((__unused_webpack___webpack_module__, __webpack_exports__, __webpack_require__) => {
|
|
17248
|
+
|
|
17249
|
+
__webpack_require__.r(__webpack_exports__);
|
|
17250
|
+
/* harmony export */ __webpack_require__.d(__webpack_exports__, {
|
|
17251
|
+
/* harmony export */ GroundingDinoImageProcessor: () => (/* binding */ GroundingDinoImageProcessor)
|
|
17252
|
+
/* harmony export */ });
|
|
17253
|
+
/* harmony import */ var _base_image_processors_utils_js__WEBPACK_IMPORTED_MODULE_0__ = __webpack_require__(/*! ../../base/image_processors_utils.js */ "./src/base/image_processors_utils.js");
|
|
17254
|
+
/* harmony import */ var _utils_tensor_js__WEBPACK_IMPORTED_MODULE_1__ = __webpack_require__(/*! ../../utils/tensor.js */ "./src/utils/tensor.js");
|
|
17255
|
+
|
|
17256
|
+
|
|
17257
|
+
|
|
17258
|
+
|
|
17259
|
+
|
|
17260
|
+
/**
|
|
17261
|
+
* @typedef {object} GroundingDinoFeatureExtractorResultProps
|
|
17262
|
+
* @property {import('../../utils/tensor.js').Tensor} pixel_mask
|
|
17263
|
+
* @typedef {import('../../base/image_processors_utils.js').ImageProcessorResult & GroundingDinoFeatureExtractorResultProps} GroundingDinoFeatureExtractorResult
|
|
17264
|
+
*/
|
|
17265
|
+
|
|
17266
|
+
class GroundingDinoImageProcessor extends _base_image_processors_utils_js__WEBPACK_IMPORTED_MODULE_0__.ImageProcessor {
|
|
17267
|
+
/**
|
|
17268
|
+
* Calls the feature extraction process on an array of images, preprocesses
|
|
17269
|
+
* each image, and concatenates the resulting features into a single Tensor.
|
|
17270
|
+
* @param {import('../../utils/image.js').RawImage[]} images The image(s) to extract features from.
|
|
17271
|
+
* @returns {Promise<GroundingDinoFeatureExtractorResult>} An object containing the concatenated pixel values of the preprocessed images.
|
|
17272
|
+
*/
|
|
17273
|
+
async _call(images) {
|
|
17274
|
+
const result = await super._call(images);
|
|
17275
|
+
|
|
17276
|
+
const dims = result.pixel_values.dims;
|
|
17277
|
+
const pixel_mask = (0,_utils_tensor_js__WEBPACK_IMPORTED_MODULE_1__.ones)([dims[0], dims[2], dims[3]]);
|
|
17278
|
+
|
|
17279
|
+
return { ...result, pixel_mask };
|
|
17280
|
+
}
|
|
17281
|
+
}
|
|
17282
|
+
|
|
17283
|
+
|
|
17284
|
+
/***/ }),
|
|
17285
|
+
|
|
17286
|
+
/***/ "./src/models/grounding_dino/processing_grounding_dino.js":
|
|
17287
|
+
/*!****************************************************************!*\
|
|
17288
|
+
!*** ./src/models/grounding_dino/processing_grounding_dino.js ***!
|
|
17289
|
+
\****************************************************************/
|
|
17290
|
+
/***/ ((__unused_webpack___webpack_module__, __webpack_exports__, __webpack_require__) => {
|
|
17291
|
+
|
|
17292
|
+
__webpack_require__.r(__webpack_exports__);
|
|
17293
|
+
/* harmony export */ __webpack_require__.d(__webpack_exports__, {
|
|
17294
|
+
/* harmony export */ GroundingDinoProcessor: () => (/* binding */ GroundingDinoProcessor)
|
|
17295
|
+
/* harmony export */ });
|
|
17296
|
+
/* harmony import */ var _base_processing_utils_js__WEBPACK_IMPORTED_MODULE_0__ = __webpack_require__(/*! ../../base/processing_utils.js */ "./src/base/processing_utils.js");
|
|
17297
|
+
/* harmony import */ var _auto_image_processing_auto_js__WEBPACK_IMPORTED_MODULE_1__ = __webpack_require__(/*! ../auto/image_processing_auto.js */ "./src/models/auto/image_processing_auto.js");
|
|
17298
|
+
/* harmony import */ var _tokenizers_js__WEBPACK_IMPORTED_MODULE_2__ = __webpack_require__(/*! ../../tokenizers.js */ "./src/tokenizers.js");
|
|
17299
|
+
/* harmony import */ var _base_image_processors_utils_js__WEBPACK_IMPORTED_MODULE_3__ = __webpack_require__(/*! ../../base/image_processors_utils.js */ "./src/base/image_processors_utils.js");
|
|
17300
|
+
|
|
17301
|
+
|
|
17302
|
+
|
|
17303
|
+
|
|
17304
|
+
|
|
17305
|
+
/**
|
|
17306
|
+
* Get token ids of phrases from posmaps and input_ids.
|
|
17307
|
+
* @param {import('../../utils/tensor.js').Tensor} posmaps A boolean tensor of unbatched text-thresholded logits related to the detected bounding boxes of shape `(hidden_size, )`.
|
|
17308
|
+
* @param {import('../../utils/tensor.js').Tensor} input_ids A tensor of token ids of shape `(sequence_length, )`.
|
|
17309
|
+
*/
|
|
17310
|
+
function get_phrases_from_posmap(posmaps, input_ids) {
|
|
17311
|
+
|
|
17312
|
+
const left_idx = 0;
|
|
17313
|
+
const right_idx = posmaps.dims.at(-1) - 1;
|
|
17314
|
+
|
|
17315
|
+
const posmaps_list = posmaps.tolist();
|
|
17316
|
+
posmaps_list.fill(false, 0, left_idx + 1);
|
|
17317
|
+
posmaps_list.fill(false, right_idx);
|
|
17318
|
+
|
|
17319
|
+
const input_ids_list = input_ids.tolist();
|
|
17320
|
+
return posmaps_list
|
|
17321
|
+
.map((val, idx) => val ? idx : null)
|
|
17322
|
+
.filter(idx => idx !== null)
|
|
17323
|
+
.map(i => input_ids_list[i]);
|
|
17324
|
+
}
|
|
17325
|
+
|
|
17326
|
+
class GroundingDinoProcessor extends _base_processing_utils_js__WEBPACK_IMPORTED_MODULE_0__.Processor {
|
|
17327
|
+
static tokenizer_class = _tokenizers_js__WEBPACK_IMPORTED_MODULE_2__.AutoTokenizer
|
|
17328
|
+
static image_processor_class = _auto_image_processing_auto_js__WEBPACK_IMPORTED_MODULE_1__.AutoImageProcessor
|
|
17329
|
+
|
|
17330
|
+
/**
|
|
17331
|
+
* @typedef {import('../../utils/image.js').RawImage} RawImage
|
|
17332
|
+
*/
|
|
17333
|
+
/**
|
|
17334
|
+
*
|
|
17335
|
+
* @param {RawImage|RawImage[]|RawImage[][]} images
|
|
17336
|
+
* @param {string|string[]} text
|
|
17337
|
+
* @returns {Promise<any>}
|
|
17338
|
+
*/
|
|
17339
|
+
async _call(images, text, options = {}) {
|
|
17340
|
+
|
|
17341
|
+
const image_inputs = images ? await this.image_processor(images, options) : {};
|
|
17342
|
+
const text_inputs = text ? this.tokenizer(text, options) : {};
|
|
17343
|
+
|
|
17344
|
+
return {
|
|
17345
|
+
...text_inputs,
|
|
17346
|
+
...image_inputs,
|
|
17347
|
+
}
|
|
17348
|
+
}
|
|
17349
|
+
post_process_grounded_object_detection(outputs, input_ids, {
|
|
17350
|
+
box_threshold = 0.25,
|
|
17351
|
+
text_threshold = 0.25,
|
|
17352
|
+
target_sizes = null
|
|
17353
|
+
} = {}) {
|
|
17354
|
+
const { logits, pred_boxes } = outputs;
|
|
17355
|
+
const batch_size = logits.dims[0];
|
|
17356
|
+
|
|
17357
|
+
if (target_sizes !== null && target_sizes.length !== batch_size) {
|
|
17358
|
+
throw Error("Make sure that you pass in as many target sizes as the batch dimension of the logits")
|
|
17359
|
+
}
|
|
17360
|
+
const num_queries = logits.dims.at(1);
|
|
17361
|
+
|
|
17362
|
+
const probs = logits.sigmoid(); // (batch_size, num_queries, 256)
|
|
17363
|
+
const scores = probs.max(-1).tolist(); // (batch_size, num_queries)
|
|
17364
|
+
|
|
17365
|
+
// Convert to [x0, y0, x1, y1] format
|
|
17366
|
+
const boxes = pred_boxes.tolist() // (batch_size, num_queries, 4)
|
|
17367
|
+
.map(batch => batch.map(box => (0,_base_image_processors_utils_js__WEBPACK_IMPORTED_MODULE_3__.center_to_corners_format)(box)));
|
|
17368
|
+
|
|
17369
|
+
const results = [];
|
|
17370
|
+
for (let i = 0; i < batch_size; ++i) {
|
|
17371
|
+
const target_size = target_sizes !== null ? target_sizes[i] : null;
|
|
17372
|
+
|
|
17373
|
+
// Convert from relative [0, 1] to absolute [0, height] coordinates
|
|
17374
|
+
if (target_size !== null) {
|
|
17375
|
+
boxes[i] = boxes[i].map(box => box.map((x, j) => x * target_size[(j + 1) % 2]));
|
|
17376
|
+
}
|
|
17377
|
+
|
|
17378
|
+
const batch_scores = scores[i];
|
|
17379
|
+
const final_scores = [];
|
|
17380
|
+
const final_phrases = [];
|
|
17381
|
+
const final_boxes = [];
|
|
17382
|
+
for (let j = 0; j < num_queries; ++j) {
|
|
17383
|
+
const score = batch_scores[j];
|
|
17384
|
+
if (score <= box_threshold) {
|
|
17385
|
+
continue;
|
|
17386
|
+
}
|
|
17387
|
+
const box = boxes[i][j];
|
|
17388
|
+
const prob = probs[i][j];
|
|
17389
|
+
|
|
17390
|
+
final_scores.push(score);
|
|
17391
|
+
final_boxes.push(box);
|
|
17392
|
+
|
|
17393
|
+
const phrases = get_phrases_from_posmap(prob.gt(text_threshold), input_ids[i]);
|
|
17394
|
+
final_phrases.push(phrases);
|
|
17395
|
+
}
|
|
17396
|
+
results.push({ scores: final_scores, boxes: final_boxes, labels: this.batch_decode(final_phrases) });
|
|
17397
|
+
}
|
|
17398
|
+
return results;
|
|
17399
|
+
}
|
|
17400
|
+
}
|
|
17401
|
+
|
|
17402
|
+
|
|
17172
17403
|
/***/ }),
|
|
17173
17404
|
|
|
17174
17405
|
/***/ "./src/models/idefics3/image_processing_idefics3.js":
|
|
@@ -17329,6 +17560,8 @@ class Idefics3ImageProcessor extends _base_image_processors_utils_js__WEBPACK_IM
|
|
|
17329
17560
|
|
|
17330
17561
|
const start_offset = i * pixel_attention_mask_stride + num_patches * h * w;
|
|
17331
17562
|
const end_offset = (i + 1) * pixel_attention_mask_stride;
|
|
17563
|
+
|
|
17564
|
+
// @ts-expect-error
|
|
17332
17565
|
pixel_attention_mask_data.fill(false, start_offset, end_offset);
|
|
17333
17566
|
}
|
|
17334
17567
|
}
|
|
@@ -17595,42 +17828,43 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
17595
17828
|
/* harmony export */ DonutImageProcessor: () => (/* reexport safe */ _donut_image_processing_donut_js__WEBPACK_IMPORTED_MODULE_7__.DonutImageProcessor),
|
|
17596
17829
|
/* harmony export */ EfficientNetImageProcessor: () => (/* reexport safe */ _efficientnet_image_processing_efficientnet_js__WEBPACK_IMPORTED_MODULE_9__.EfficientNetImageProcessor),
|
|
17597
17830
|
/* harmony export */ GLPNFeatureExtractor: () => (/* reexport safe */ _glpn_image_processing_glpn_js__WEBPACK_IMPORTED_MODULE_10__.GLPNFeatureExtractor),
|
|
17598
|
-
/* harmony export */
|
|
17599
|
-
/* harmony export */
|
|
17600
|
-
/* harmony export */
|
|
17601
|
-
/* harmony export */
|
|
17602
|
-
/* harmony export */
|
|
17603
|
-
/* harmony export */
|
|
17604
|
-
/* harmony export */
|
|
17605
|
-
/* harmony export */
|
|
17606
|
-
/* harmony export */
|
|
17607
|
-
/* harmony export */
|
|
17608
|
-
/* harmony export */
|
|
17609
|
-
/* harmony export */
|
|
17610
|
-
/* harmony export */
|
|
17611
|
-
/* harmony export */
|
|
17612
|
-
/* harmony export */
|
|
17613
|
-
/* harmony export */
|
|
17614
|
-
/* harmony export */
|
|
17615
|
-
/* harmony export */
|
|
17616
|
-
/* harmony export */
|
|
17617
|
-
/* harmony export */
|
|
17618
|
-
/* harmony export */
|
|
17619
|
-
/* harmony export */
|
|
17620
|
-
/* harmony export */
|
|
17621
|
-
/* harmony export */
|
|
17622
|
-
/* harmony export */
|
|
17623
|
-
/* harmony export */
|
|
17624
|
-
/* harmony export */
|
|
17625
|
-
/* harmony export */
|
|
17626
|
-
/* harmony export */
|
|
17627
|
-
/* harmony export */
|
|
17628
|
-
/* harmony export */
|
|
17629
|
-
/* harmony export */
|
|
17630
|
-
/* harmony export */
|
|
17631
|
-
/* harmony export */
|
|
17632
|
-
/* harmony export */
|
|
17633
|
-
/* harmony export */
|
|
17831
|
+
/* harmony export */ GroundingDinoImageProcessor: () => (/* reexport safe */ _grounding_dino_image_processing_grounding_dino_js__WEBPACK_IMPORTED_MODULE_11__.GroundingDinoImageProcessor),
|
|
17832
|
+
/* harmony export */ Idefics3ImageProcessor: () => (/* reexport safe */ _idefics3_image_processing_idefics3_js__WEBPACK_IMPORTED_MODULE_12__.Idefics3ImageProcessor),
|
|
17833
|
+
/* harmony export */ JinaCLIPImageProcessor: () => (/* reexport safe */ _jina_clip_image_processing_jina_clip_js__WEBPACK_IMPORTED_MODULE_14__.JinaCLIPImageProcessor),
|
|
17834
|
+
/* harmony export */ LlavaOnevisionImageProcessor: () => (/* reexport safe */ _llava_onevision_image_processing_llava_onevision_js__WEBPACK_IMPORTED_MODULE_15__.LlavaOnevisionImageProcessor),
|
|
17835
|
+
/* harmony export */ Mask2FormerImageProcessor: () => (/* reexport safe */ _mask2former_image_processing_mask2former_js__WEBPACK_IMPORTED_MODULE_16__.Mask2FormerImageProcessor),
|
|
17836
|
+
/* harmony export */ MaskFormerFeatureExtractor: () => (/* reexport safe */ _maskformer_image_processing_maskformer_js__WEBPACK_IMPORTED_MODULE_17__.MaskFormerFeatureExtractor),
|
|
17837
|
+
/* harmony export */ MaskFormerImageProcessor: () => (/* reexport safe */ _maskformer_image_processing_maskformer_js__WEBPACK_IMPORTED_MODULE_17__.MaskFormerImageProcessor),
|
|
17838
|
+
/* harmony export */ MobileNetV1FeatureExtractor: () => (/* reexport safe */ _mobilenet_v1_image_processing_mobilenet_v1_js__WEBPACK_IMPORTED_MODULE_18__.MobileNetV1FeatureExtractor),
|
|
17839
|
+
/* harmony export */ MobileNetV1ImageProcessor: () => (/* reexport safe */ _mobilenet_v1_image_processing_mobilenet_v1_js__WEBPACK_IMPORTED_MODULE_18__.MobileNetV1ImageProcessor),
|
|
17840
|
+
/* harmony export */ MobileNetV2FeatureExtractor: () => (/* reexport safe */ _mobilenet_v2_image_processing_mobilenet_v2_js__WEBPACK_IMPORTED_MODULE_19__.MobileNetV2FeatureExtractor),
|
|
17841
|
+
/* harmony export */ MobileNetV2ImageProcessor: () => (/* reexport safe */ _mobilenet_v2_image_processing_mobilenet_v2_js__WEBPACK_IMPORTED_MODULE_19__.MobileNetV2ImageProcessor),
|
|
17842
|
+
/* harmony export */ MobileNetV3FeatureExtractor: () => (/* reexport safe */ _mobilenet_v3_image_processing_mobilenet_v3_js__WEBPACK_IMPORTED_MODULE_20__.MobileNetV3FeatureExtractor),
|
|
17843
|
+
/* harmony export */ MobileNetV3ImageProcessor: () => (/* reexport safe */ _mobilenet_v3_image_processing_mobilenet_v3_js__WEBPACK_IMPORTED_MODULE_20__.MobileNetV3ImageProcessor),
|
|
17844
|
+
/* harmony export */ MobileNetV4FeatureExtractor: () => (/* reexport safe */ _mobilenet_v4_image_processing_mobilenet_v4_js__WEBPACK_IMPORTED_MODULE_21__.MobileNetV4FeatureExtractor),
|
|
17845
|
+
/* harmony export */ MobileNetV4ImageProcessor: () => (/* reexport safe */ _mobilenet_v4_image_processing_mobilenet_v4_js__WEBPACK_IMPORTED_MODULE_21__.MobileNetV4ImageProcessor),
|
|
17846
|
+
/* harmony export */ MobileViTFeatureExtractor: () => (/* reexport safe */ _mobilevit_image_processing_mobilevit_js__WEBPACK_IMPORTED_MODULE_22__.MobileViTFeatureExtractor),
|
|
17847
|
+
/* harmony export */ MobileViTImageProcessor: () => (/* reexport safe */ _mobilevit_image_processing_mobilevit_js__WEBPACK_IMPORTED_MODULE_22__.MobileViTImageProcessor),
|
|
17848
|
+
/* harmony export */ NougatImageProcessor: () => (/* reexport safe */ _nougat_image_processing_nougat_js__WEBPACK_IMPORTED_MODULE_23__.NougatImageProcessor),
|
|
17849
|
+
/* harmony export */ OwlViTFeatureExtractor: () => (/* reexport safe */ _owlvit_image_processing_owlvit_js__WEBPACK_IMPORTED_MODULE_25__.OwlViTFeatureExtractor),
|
|
17850
|
+
/* harmony export */ OwlViTImageProcessor: () => (/* reexport safe */ _owlvit_image_processing_owlvit_js__WEBPACK_IMPORTED_MODULE_25__.OwlViTImageProcessor),
|
|
17851
|
+
/* harmony export */ Owlv2ImageProcessor: () => (/* reexport safe */ _owlv2_image_processing_owlv2_js__WEBPACK_IMPORTED_MODULE_24__.Owlv2ImageProcessor),
|
|
17852
|
+
/* harmony export */ Phi3VImageProcessor: () => (/* reexport safe */ _phi3_v_image_processing_phi3_v_js__WEBPACK_IMPORTED_MODULE_26__.Phi3VImageProcessor),
|
|
17853
|
+
/* harmony export */ PvtImageProcessor: () => (/* reexport safe */ _pvt_image_processing_pvt_js__WEBPACK_IMPORTED_MODULE_27__.PvtImageProcessor),
|
|
17854
|
+
/* harmony export */ Qwen2VLImageProcessor: () => (/* reexport safe */ _qwen2_vl_image_processing_qwen2_vl_js__WEBPACK_IMPORTED_MODULE_28__.Qwen2VLImageProcessor),
|
|
17855
|
+
/* harmony export */ RTDetrImageProcessor: () => (/* reexport safe */ _rt_detr_image_processing_rt_detr_js__WEBPACK_IMPORTED_MODULE_29__.RTDetrImageProcessor),
|
|
17856
|
+
/* harmony export */ SamImageProcessor: () => (/* reexport safe */ _sam_image_processing_sam_js__WEBPACK_IMPORTED_MODULE_30__.SamImageProcessor),
|
|
17857
|
+
/* harmony export */ SegformerFeatureExtractor: () => (/* reexport safe */ _segformer_image_processing_segformer_js__WEBPACK_IMPORTED_MODULE_31__.SegformerFeatureExtractor),
|
|
17858
|
+
/* harmony export */ SegformerImageProcessor: () => (/* reexport safe */ _segformer_image_processing_segformer_js__WEBPACK_IMPORTED_MODULE_31__.SegformerImageProcessor),
|
|
17859
|
+
/* harmony export */ SiglipImageProcessor: () => (/* reexport safe */ _siglip_image_processing_siglip_js__WEBPACK_IMPORTED_MODULE_32__.SiglipImageProcessor),
|
|
17860
|
+
/* harmony export */ Swin2SRImageProcessor: () => (/* reexport safe */ _swin2sr_image_processing_swin2sr_js__WEBPACK_IMPORTED_MODULE_33__.Swin2SRImageProcessor),
|
|
17861
|
+
/* harmony export */ VLMImageProcessor: () => (/* reexport safe */ _janus_image_processing_janus_js__WEBPACK_IMPORTED_MODULE_13__.VLMImageProcessor),
|
|
17862
|
+
/* harmony export */ ViTFeatureExtractor: () => (/* reexport safe */ _vit_image_processing_vit_js__WEBPACK_IMPORTED_MODULE_34__.ViTFeatureExtractor),
|
|
17863
|
+
/* harmony export */ ViTImageProcessor: () => (/* reexport safe */ _vit_image_processing_vit_js__WEBPACK_IMPORTED_MODULE_34__.ViTImageProcessor),
|
|
17864
|
+
/* harmony export */ VitMatteImageProcessor: () => (/* reexport safe */ _vitmatte_image_processing_vitmatte_js__WEBPACK_IMPORTED_MODULE_35__.VitMatteImageProcessor),
|
|
17865
|
+
/* harmony export */ VitPoseImageProcessor: () => (/* reexport safe */ _vitpose_image_processing_vitpose_js__WEBPACK_IMPORTED_MODULE_36__.VitPoseImageProcessor),
|
|
17866
|
+
/* harmony export */ YolosFeatureExtractor: () => (/* reexport safe */ _yolos_image_processing_yolos_js__WEBPACK_IMPORTED_MODULE_37__.YolosFeatureExtractor),
|
|
17867
|
+
/* harmony export */ YolosImageProcessor: () => (/* reexport safe */ _yolos_image_processing_yolos_js__WEBPACK_IMPORTED_MODULE_37__.YolosImageProcessor)
|
|
17634
17868
|
/* harmony export */ });
|
|
17635
17869
|
/* harmony import */ var _beit_image_processing_beit_js__WEBPACK_IMPORTED_MODULE_0__ = __webpack_require__(/*! ./beit/image_processing_beit.js */ "./src/models/beit/image_processing_beit.js");
|
|
17636
17870
|
/* harmony import */ var _bit_image_processing_bit_js__WEBPACK_IMPORTED_MODULE_1__ = __webpack_require__(/*! ./bit/image_processing_bit.js */ "./src/models/bit/image_processing_bit.js");
|
|
@@ -17643,32 +17877,34 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
17643
17877
|
/* harmony import */ var _dpt_image_processing_dpt_js__WEBPACK_IMPORTED_MODULE_8__ = __webpack_require__(/*! ./dpt/image_processing_dpt.js */ "./src/models/dpt/image_processing_dpt.js");
|
|
17644
17878
|
/* harmony import */ var _efficientnet_image_processing_efficientnet_js__WEBPACK_IMPORTED_MODULE_9__ = __webpack_require__(/*! ./efficientnet/image_processing_efficientnet.js */ "./src/models/efficientnet/image_processing_efficientnet.js");
|
|
17645
17879
|
/* harmony import */ var _glpn_image_processing_glpn_js__WEBPACK_IMPORTED_MODULE_10__ = __webpack_require__(/*! ./glpn/image_processing_glpn.js */ "./src/models/glpn/image_processing_glpn.js");
|
|
17646
|
-
/* harmony import */ var
|
|
17647
|
-
/* harmony import */ var
|
|
17648
|
-
/* harmony import */ var
|
|
17649
|
-
/* harmony import */ var
|
|
17650
|
-
/* harmony import */ var
|
|
17651
|
-
/* harmony import */ var
|
|
17652
|
-
/* harmony import */ var
|
|
17653
|
-
/* harmony import */ var
|
|
17654
|
-
/* harmony import */ var
|
|
17655
|
-
/* harmony import */ var
|
|
17656
|
-
/* harmony import */ var
|
|
17657
|
-
/* harmony import */ var
|
|
17658
|
-
/* harmony import */ var
|
|
17659
|
-
/* harmony import */ var
|
|
17660
|
-
/* harmony import */ var
|
|
17661
|
-
/* harmony import */ var
|
|
17662
|
-
/* harmony import */ var
|
|
17663
|
-
/* harmony import */ var
|
|
17664
|
-
/* harmony import */ var
|
|
17665
|
-
/* harmony import */ var
|
|
17666
|
-
/* harmony import */ var
|
|
17667
|
-
/* harmony import */ var
|
|
17668
|
-
/* harmony import */ var
|
|
17669
|
-
/* harmony import */ var
|
|
17670
|
-
/* harmony import */ var
|
|
17671
|
-
/* harmony import */ var
|
|
17880
|
+
/* harmony import */ var _grounding_dino_image_processing_grounding_dino_js__WEBPACK_IMPORTED_MODULE_11__ = __webpack_require__(/*! ./grounding_dino/image_processing_grounding_dino.js */ "./src/models/grounding_dino/image_processing_grounding_dino.js");
|
|
17881
|
+
/* harmony import */ var _idefics3_image_processing_idefics3_js__WEBPACK_IMPORTED_MODULE_12__ = __webpack_require__(/*! ./idefics3/image_processing_idefics3.js */ "./src/models/idefics3/image_processing_idefics3.js");
|
|
17882
|
+
/* harmony import */ var _janus_image_processing_janus_js__WEBPACK_IMPORTED_MODULE_13__ = __webpack_require__(/*! ./janus/image_processing_janus.js */ "./src/models/janus/image_processing_janus.js");
|
|
17883
|
+
/* harmony import */ var _jina_clip_image_processing_jina_clip_js__WEBPACK_IMPORTED_MODULE_14__ = __webpack_require__(/*! ./jina_clip/image_processing_jina_clip.js */ "./src/models/jina_clip/image_processing_jina_clip.js");
|
|
17884
|
+
/* harmony import */ var _llava_onevision_image_processing_llava_onevision_js__WEBPACK_IMPORTED_MODULE_15__ = __webpack_require__(/*! ./llava_onevision/image_processing_llava_onevision.js */ "./src/models/llava_onevision/image_processing_llava_onevision.js");
|
|
17885
|
+
/* harmony import */ var _mask2former_image_processing_mask2former_js__WEBPACK_IMPORTED_MODULE_16__ = __webpack_require__(/*! ./mask2former/image_processing_mask2former.js */ "./src/models/mask2former/image_processing_mask2former.js");
|
|
17886
|
+
/* harmony import */ var _maskformer_image_processing_maskformer_js__WEBPACK_IMPORTED_MODULE_17__ = __webpack_require__(/*! ./maskformer/image_processing_maskformer.js */ "./src/models/maskformer/image_processing_maskformer.js");
|
|
17887
|
+
/* harmony import */ var _mobilenet_v1_image_processing_mobilenet_v1_js__WEBPACK_IMPORTED_MODULE_18__ = __webpack_require__(/*! ./mobilenet_v1/image_processing_mobilenet_v1.js */ "./src/models/mobilenet_v1/image_processing_mobilenet_v1.js");
|
|
17888
|
+
/* harmony import */ var _mobilenet_v2_image_processing_mobilenet_v2_js__WEBPACK_IMPORTED_MODULE_19__ = __webpack_require__(/*! ./mobilenet_v2/image_processing_mobilenet_v2.js */ "./src/models/mobilenet_v2/image_processing_mobilenet_v2.js");
|
|
17889
|
+
/* harmony import */ var _mobilenet_v3_image_processing_mobilenet_v3_js__WEBPACK_IMPORTED_MODULE_20__ = __webpack_require__(/*! ./mobilenet_v3/image_processing_mobilenet_v3.js */ "./src/models/mobilenet_v3/image_processing_mobilenet_v3.js");
|
|
17890
|
+
/* harmony import */ var _mobilenet_v4_image_processing_mobilenet_v4_js__WEBPACK_IMPORTED_MODULE_21__ = __webpack_require__(/*! ./mobilenet_v4/image_processing_mobilenet_v4.js */ "./src/models/mobilenet_v4/image_processing_mobilenet_v4.js");
|
|
17891
|
+
/* harmony import */ var _mobilevit_image_processing_mobilevit_js__WEBPACK_IMPORTED_MODULE_22__ = __webpack_require__(/*! ./mobilevit/image_processing_mobilevit.js */ "./src/models/mobilevit/image_processing_mobilevit.js");
|
|
17892
|
+
/* harmony import */ var _nougat_image_processing_nougat_js__WEBPACK_IMPORTED_MODULE_23__ = __webpack_require__(/*! ./nougat/image_processing_nougat.js */ "./src/models/nougat/image_processing_nougat.js");
|
|
17893
|
+
/* harmony import */ var _owlv2_image_processing_owlv2_js__WEBPACK_IMPORTED_MODULE_24__ = __webpack_require__(/*! ./owlv2/image_processing_owlv2.js */ "./src/models/owlv2/image_processing_owlv2.js");
|
|
17894
|
+
/* harmony import */ var _owlvit_image_processing_owlvit_js__WEBPACK_IMPORTED_MODULE_25__ = __webpack_require__(/*! ./owlvit/image_processing_owlvit.js */ "./src/models/owlvit/image_processing_owlvit.js");
|
|
17895
|
+
/* harmony import */ var _phi3_v_image_processing_phi3_v_js__WEBPACK_IMPORTED_MODULE_26__ = __webpack_require__(/*! ./phi3_v/image_processing_phi3_v.js */ "./src/models/phi3_v/image_processing_phi3_v.js");
|
|
17896
|
+
/* harmony import */ var _pvt_image_processing_pvt_js__WEBPACK_IMPORTED_MODULE_27__ = __webpack_require__(/*! ./pvt/image_processing_pvt.js */ "./src/models/pvt/image_processing_pvt.js");
|
|
17897
|
+
/* harmony import */ var _qwen2_vl_image_processing_qwen2_vl_js__WEBPACK_IMPORTED_MODULE_28__ = __webpack_require__(/*! ./qwen2_vl/image_processing_qwen2_vl.js */ "./src/models/qwen2_vl/image_processing_qwen2_vl.js");
|
|
17898
|
+
/* harmony import */ var _rt_detr_image_processing_rt_detr_js__WEBPACK_IMPORTED_MODULE_29__ = __webpack_require__(/*! ./rt_detr/image_processing_rt_detr.js */ "./src/models/rt_detr/image_processing_rt_detr.js");
|
|
17899
|
+
/* harmony import */ var _sam_image_processing_sam_js__WEBPACK_IMPORTED_MODULE_30__ = __webpack_require__(/*! ./sam/image_processing_sam.js */ "./src/models/sam/image_processing_sam.js");
|
|
17900
|
+
/* harmony import */ var _segformer_image_processing_segformer_js__WEBPACK_IMPORTED_MODULE_31__ = __webpack_require__(/*! ./segformer/image_processing_segformer.js */ "./src/models/segformer/image_processing_segformer.js");
|
|
17901
|
+
/* harmony import */ var _siglip_image_processing_siglip_js__WEBPACK_IMPORTED_MODULE_32__ = __webpack_require__(/*! ./siglip/image_processing_siglip.js */ "./src/models/siglip/image_processing_siglip.js");
|
|
17902
|
+
/* harmony import */ var _swin2sr_image_processing_swin2sr_js__WEBPACK_IMPORTED_MODULE_33__ = __webpack_require__(/*! ./swin2sr/image_processing_swin2sr.js */ "./src/models/swin2sr/image_processing_swin2sr.js");
|
|
17903
|
+
/* harmony import */ var _vit_image_processing_vit_js__WEBPACK_IMPORTED_MODULE_34__ = __webpack_require__(/*! ./vit/image_processing_vit.js */ "./src/models/vit/image_processing_vit.js");
|
|
17904
|
+
/* harmony import */ var _vitmatte_image_processing_vitmatte_js__WEBPACK_IMPORTED_MODULE_35__ = __webpack_require__(/*! ./vitmatte/image_processing_vitmatte.js */ "./src/models/vitmatte/image_processing_vitmatte.js");
|
|
17905
|
+
/* harmony import */ var _vitpose_image_processing_vitpose_js__WEBPACK_IMPORTED_MODULE_36__ = __webpack_require__(/*! ./vitpose/image_processing_vitpose.js */ "./src/models/vitpose/image_processing_vitpose.js");
|
|
17906
|
+
/* harmony import */ var _yolos_image_processing_yolos_js__WEBPACK_IMPORTED_MODULE_37__ = __webpack_require__(/*! ./yolos/image_processing_yolos.js */ "./src/models/yolos/image_processing_yolos.js");
|
|
17907
|
+
|
|
17672
17908
|
|
|
17673
17909
|
|
|
17674
17910
|
|
|
@@ -17735,6 +17971,7 @@ class VLMImageProcessor extends _base_image_processors_utils_js__WEBPACK_IMPORTE
|
|
|
17735
17971
|
},
|
|
17736
17972
|
...config,
|
|
17737
17973
|
});
|
|
17974
|
+
// @ts-expect-error TS2339
|
|
17738
17975
|
this.constant_values = this.config.background_color.map(x => x * this.rescale_factor)
|
|
17739
17976
|
}
|
|
17740
17977
|
|
|
@@ -18176,6 +18413,8 @@ class MgpstrProcessor extends _base_processing_utils_js__WEBPACK_IMPORTED_MODULE
|
|
|
18176
18413
|
* - bpe_preds: The list of BPE decoded sentences.
|
|
18177
18414
|
* - wp_preds: The list of wp decoded sentences.
|
|
18178
18415
|
*/
|
|
18416
|
+
// @ts-expect-error The type of this method is not compatible with the one
|
|
18417
|
+
// in the base class. It might be a good idea to fix this.
|
|
18179
18418
|
batch_decode([char_logits, bpe_logits, wp_logits]) {
|
|
18180
18419
|
const [char_preds, char_scores] = this._decode_helper(char_logits, 'char');
|
|
18181
18420
|
const [bpe_preds, bpe_scores] = this._decode_helper(bpe_logits, 'bpe');
|
|
@@ -18557,6 +18796,7 @@ class PaliGemmaProcessor extends _base_processing_utils_js__WEBPACK_IMPORTED_MOD
|
|
|
18557
18796
|
}
|
|
18558
18797
|
|
|
18559
18798
|
const bos_token = this.tokenizer.bos_token;
|
|
18799
|
+
// @ts-expect-error TS2339
|
|
18560
18800
|
const image_seq_length = this.image_processor.config.image_seq_length;
|
|
18561
18801
|
let input_strings;
|
|
18562
18802
|
if (text.some((t) => t.includes(IMAGE_TOKEN))) {
|
|
@@ -18807,7 +19047,7 @@ class Phi3VProcessor extends _base_processing_utils_js__WEBPACK_IMPORTED_MODULE_
|
|
|
18807
19047
|
*
|
|
18808
19048
|
* @param {string|string[]} text
|
|
18809
19049
|
* @param {RawImage|RawImage[]} images
|
|
18810
|
-
* @param {
|
|
19050
|
+
* @param { { padding?: boolean, truncation?: boolean, num_crops?: number } | undefined } options
|
|
18811
19051
|
* @returns {Promise<any>}
|
|
18812
19052
|
*/
|
|
18813
19053
|
async _call(text, images = null, {
|
|
@@ -18857,36 +19097,39 @@ class Phi3VProcessor extends _base_processing_utils_js__WEBPACK_IMPORTED_MODULE_
|
|
|
18857
19097
|
__webpack_require__.r(__webpack_exports__);
|
|
18858
19098
|
/* harmony export */ __webpack_require__.d(__webpack_exports__, {
|
|
18859
19099
|
/* harmony export */ Florence2Processor: () => (/* reexport safe */ _florence2_processing_florence2_js__WEBPACK_IMPORTED_MODULE_0__.Florence2Processor),
|
|
18860
|
-
/* harmony export */
|
|
18861
|
-
/* harmony export */
|
|
18862
|
-
/* harmony export */
|
|
18863
|
-
/* harmony export */
|
|
18864
|
-
/* harmony export */
|
|
18865
|
-
/* harmony export */
|
|
18866
|
-
/* harmony export */
|
|
18867
|
-
/* harmony export */
|
|
18868
|
-
/* harmony export */
|
|
18869
|
-
/* harmony export */
|
|
18870
|
-
/* harmony export */
|
|
18871
|
-
/* harmony export */
|
|
18872
|
-
/* harmony export */
|
|
18873
|
-
/* harmony export */
|
|
19100
|
+
/* harmony export */ GroundingDinoProcessor: () => (/* reexport safe */ _grounding_dino_processing_grounding_dino_js__WEBPACK_IMPORTED_MODULE_1__.GroundingDinoProcessor),
|
|
19101
|
+
/* harmony export */ Idefics3Processor: () => (/* reexport safe */ _idefics3_processing_idefics3_js__WEBPACK_IMPORTED_MODULE_2__.Idefics3Processor),
|
|
19102
|
+
/* harmony export */ JinaCLIPProcessor: () => (/* reexport safe */ _jina_clip_processing_jina_clip_js__WEBPACK_IMPORTED_MODULE_4__.JinaCLIPProcessor),
|
|
19103
|
+
/* harmony export */ MgpstrProcessor: () => (/* reexport safe */ _mgp_str_processing_mgp_str_js__WEBPACK_IMPORTED_MODULE_5__.MgpstrProcessor),
|
|
19104
|
+
/* harmony export */ MoonshineProcessor: () => (/* reexport safe */ _moonshine_processing_moonshine_js__WEBPACK_IMPORTED_MODULE_6__.MoonshineProcessor),
|
|
19105
|
+
/* harmony export */ OwlViTProcessor: () => (/* reexport safe */ _owlvit_processing_owlvit_js__WEBPACK_IMPORTED_MODULE_7__.OwlViTProcessor),
|
|
19106
|
+
/* harmony export */ PaliGemmaProcessor: () => (/* reexport safe */ _paligemma_processing_paligemma_js__WEBPACK_IMPORTED_MODULE_9__.PaliGemmaProcessor),
|
|
19107
|
+
/* harmony export */ Phi3VProcessor: () => (/* reexport safe */ _phi3_v_processing_phi3_v_js__WEBPACK_IMPORTED_MODULE_8__.Phi3VProcessor),
|
|
19108
|
+
/* harmony export */ PyAnnoteProcessor: () => (/* reexport safe */ _pyannote_processing_pyannote_js__WEBPACK_IMPORTED_MODULE_10__.PyAnnoteProcessor),
|
|
19109
|
+
/* harmony export */ Qwen2VLProcessor: () => (/* reexport safe */ _qwen2_vl_processing_qwen2_vl_js__WEBPACK_IMPORTED_MODULE_11__.Qwen2VLProcessor),
|
|
19110
|
+
/* harmony export */ SamProcessor: () => (/* reexport safe */ _sam_processing_sam_js__WEBPACK_IMPORTED_MODULE_12__.SamProcessor),
|
|
19111
|
+
/* harmony export */ SpeechT5Processor: () => (/* reexport safe */ _speecht5_processing_speecht5_js__WEBPACK_IMPORTED_MODULE_13__.SpeechT5Processor),
|
|
19112
|
+
/* harmony export */ VLChatProcessor: () => (/* reexport safe */ _janus_processing_janus_js__WEBPACK_IMPORTED_MODULE_3__.VLChatProcessor),
|
|
19113
|
+
/* harmony export */ Wav2Vec2ProcessorWithLM: () => (/* reexport safe */ _wav2vec2_processing_wav2vec2_js__WEBPACK_IMPORTED_MODULE_14__.Wav2Vec2ProcessorWithLM),
|
|
19114
|
+
/* harmony export */ WhisperProcessor: () => (/* reexport safe */ _whisper_processing_whisper_js__WEBPACK_IMPORTED_MODULE_15__.WhisperProcessor)
|
|
18874
19115
|
/* harmony export */ });
|
|
18875
19116
|
/* harmony import */ var _florence2_processing_florence2_js__WEBPACK_IMPORTED_MODULE_0__ = __webpack_require__(/*! ./florence2/processing_florence2.js */ "./src/models/florence2/processing_florence2.js");
|
|
18876
|
-
/* harmony import */ var
|
|
18877
|
-
/* harmony import */ var
|
|
18878
|
-
/* harmony import */ var
|
|
18879
|
-
/* harmony import */ var
|
|
18880
|
-
/* harmony import */ var
|
|
18881
|
-
/* harmony import */ var
|
|
18882
|
-
/* harmony import */ var
|
|
18883
|
-
/* harmony import */ var
|
|
18884
|
-
/* harmony import */ var
|
|
18885
|
-
/* harmony import */ var
|
|
18886
|
-
/* harmony import */ var
|
|
18887
|
-
/* harmony import */ var
|
|
18888
|
-
/* harmony import */ var
|
|
18889
|
-
/* harmony import */ var
|
|
19117
|
+
/* harmony import */ var _grounding_dino_processing_grounding_dino_js__WEBPACK_IMPORTED_MODULE_1__ = __webpack_require__(/*! ./grounding_dino/processing_grounding_dino.js */ "./src/models/grounding_dino/processing_grounding_dino.js");
|
|
19118
|
+
/* harmony import */ var _idefics3_processing_idefics3_js__WEBPACK_IMPORTED_MODULE_2__ = __webpack_require__(/*! ./idefics3/processing_idefics3.js */ "./src/models/idefics3/processing_idefics3.js");
|
|
19119
|
+
/* harmony import */ var _janus_processing_janus_js__WEBPACK_IMPORTED_MODULE_3__ = __webpack_require__(/*! ./janus/processing_janus.js */ "./src/models/janus/processing_janus.js");
|
|
19120
|
+
/* harmony import */ var _jina_clip_processing_jina_clip_js__WEBPACK_IMPORTED_MODULE_4__ = __webpack_require__(/*! ./jina_clip/processing_jina_clip.js */ "./src/models/jina_clip/processing_jina_clip.js");
|
|
19121
|
+
/* harmony import */ var _mgp_str_processing_mgp_str_js__WEBPACK_IMPORTED_MODULE_5__ = __webpack_require__(/*! ./mgp_str/processing_mgp_str.js */ "./src/models/mgp_str/processing_mgp_str.js");
|
|
19122
|
+
/* harmony import */ var _moonshine_processing_moonshine_js__WEBPACK_IMPORTED_MODULE_6__ = __webpack_require__(/*! ./moonshine/processing_moonshine.js */ "./src/models/moonshine/processing_moonshine.js");
|
|
19123
|
+
/* harmony import */ var _owlvit_processing_owlvit_js__WEBPACK_IMPORTED_MODULE_7__ = __webpack_require__(/*! ./owlvit/processing_owlvit.js */ "./src/models/owlvit/processing_owlvit.js");
|
|
19124
|
+
/* harmony import */ var _phi3_v_processing_phi3_v_js__WEBPACK_IMPORTED_MODULE_8__ = __webpack_require__(/*! ./phi3_v/processing_phi3_v.js */ "./src/models/phi3_v/processing_phi3_v.js");
|
|
19125
|
+
/* harmony import */ var _paligemma_processing_paligemma_js__WEBPACK_IMPORTED_MODULE_9__ = __webpack_require__(/*! ./paligemma/processing_paligemma.js */ "./src/models/paligemma/processing_paligemma.js");
|
|
19126
|
+
/* harmony import */ var _pyannote_processing_pyannote_js__WEBPACK_IMPORTED_MODULE_10__ = __webpack_require__(/*! ./pyannote/processing_pyannote.js */ "./src/models/pyannote/processing_pyannote.js");
|
|
19127
|
+
/* harmony import */ var _qwen2_vl_processing_qwen2_vl_js__WEBPACK_IMPORTED_MODULE_11__ = __webpack_require__(/*! ./qwen2_vl/processing_qwen2_vl.js */ "./src/models/qwen2_vl/processing_qwen2_vl.js");
|
|
19128
|
+
/* harmony import */ var _sam_processing_sam_js__WEBPACK_IMPORTED_MODULE_12__ = __webpack_require__(/*! ./sam/processing_sam.js */ "./src/models/sam/processing_sam.js");
|
|
19129
|
+
/* harmony import */ var _speecht5_processing_speecht5_js__WEBPACK_IMPORTED_MODULE_13__ = __webpack_require__(/*! ./speecht5/processing_speecht5.js */ "./src/models/speecht5/processing_speecht5.js");
|
|
19130
|
+
/* harmony import */ var _wav2vec2_processing_wav2vec2_js__WEBPACK_IMPORTED_MODULE_14__ = __webpack_require__(/*! ./wav2vec2/processing_wav2vec2.js */ "./src/models/wav2vec2/processing_wav2vec2.js");
|
|
19131
|
+
/* harmony import */ var _whisper_processing_whisper_js__WEBPACK_IMPORTED_MODULE_15__ = __webpack_require__(/*! ./whisper/processing_whisper.js */ "./src/models/whisper/processing_whisper.js");
|
|
19132
|
+
|
|
18890
19133
|
|
|
18891
19134
|
|
|
18892
19135
|
|
|
@@ -18991,6 +19234,7 @@ class PyAnnoteFeatureExtractor extends _base_feature_extraction_utils_js__WEBPAC
|
|
|
18991
19234
|
|
|
18992
19235
|
let current_speaker = -1;
|
|
18993
19236
|
for (let i = 0; i < scores.length; ++i) {
|
|
19237
|
+
/** @type {number[]} */
|
|
18994
19238
|
const probabilities = (0,_utils_maths_js__WEBPACK_IMPORTED_MODULE_2__.softmax)(scores[i]);
|
|
18995
19239
|
const [score, id] = (0,_utils_maths_js__WEBPACK_IMPORTED_MODULE_2__.max)(probabilities);
|
|
18996
19240
|
const [start, end] = [i, i + 1];
|
|
@@ -19175,6 +19419,7 @@ class Qwen2VLProcessor extends _base_processing_utils_js__WEBPACK_IMPORTED_MODUL
|
|
|
19175
19419
|
}
|
|
19176
19420
|
|
|
19177
19421
|
if (image_grid_thw) {
|
|
19422
|
+
// @ts-expect-error TS2551
|
|
19178
19423
|
let merge_length = this.image_processor.config.merge_size ** 2;
|
|
19179
19424
|
let index = 0;
|
|
19180
19425
|
|
|
@@ -19662,8 +19907,8 @@ class SeamlessM4TFeatureExtractor extends _base_feature_extraction_utils_js__WEB
|
|
|
19662
19907
|
'int64',
|
|
19663
19908
|
new BigInt64Array(numPaddedFrames),
|
|
19664
19909
|
[1, numPaddedFrames],
|
|
19665
|
-
)
|
|
19666
|
-
padded_attention_mask.data.fill(1n, 0, num_frames);
|
|
19910
|
+
);
|
|
19911
|
+
/** @type {BigInt64Array} */ (padded_attention_mask.data).fill(1n, 0, num_frames);
|
|
19667
19912
|
}
|
|
19668
19913
|
}
|
|
19669
19914
|
}
|
|
@@ -20463,7 +20708,7 @@ class WhisperFeatureExtractor extends _base_feature_extraction_utils_js__WEBPACK
|
|
|
20463
20708
|
)
|
|
20464
20709
|
|
|
20465
20710
|
const data = features.data;
|
|
20466
|
-
const maxValue = (0,_utils_maths_js__WEBPACK_IMPORTED_MODULE_3__.max)(data)[0];
|
|
20711
|
+
const maxValue = (0,_utils_maths_js__WEBPACK_IMPORTED_MODULE_3__.max)(/** @type {Float32Array} */(data))[0];
|
|
20467
20712
|
|
|
20468
20713
|
for (let i = 0; i < data.length; ++i) {
|
|
20469
20714
|
data[i] = (Math.max(data[i], maxValue - 8.0) + 4.0) / 4.0;
|
|
@@ -20722,6 +20967,16 @@ class TensorOpRegistry {
|
|
|
20722
20967
|
// executionProviders: ['webgpu'],
|
|
20723
20968
|
};
|
|
20724
20969
|
|
|
20970
|
+
static get nearest_interpolate_4d() {
|
|
20971
|
+
if (!this._nearest_interpolate_4d) {
|
|
20972
|
+
this._nearest_interpolate_4d = wrap(
|
|
20973
|
+
[8, 10, 18, 0, 58, 129, 1, 10, 41, 10, 1, 120, 10, 0, 10, 0, 10, 1, 115, 18, 1, 121, 34, 6, 82, 101, 115, 105, 122, 101, 42, 18, 10, 4, 109, 111, 100, 101, 34, 7, 110, 101, 97, 114, 101, 115, 116, 160, 1, 3, 18, 1, 114, 90, 31, 10, 1, 120, 18, 26, 10, 24, 8, 1, 18, 20, 10, 3, 18, 1, 98, 10, 3, 18, 1, 99, 10, 3, 18, 1, 104, 10, 3, 18, 1, 119, 90, 15, 10, 1, 115, 18, 10, 10, 8, 8, 7, 18, 4, 10, 2, 8, 4, 98, 31, 10, 1, 121, 18, 26, 10, 24, 8, 1, 18, 20, 10, 3, 18, 1, 98, 10, 3, 18, 1, 99, 10, 3, 18, 1, 104, 10, 3, 18, 1, 119, 66, 2, 16, 21],
|
|
20974
|
+
this.session_options,
|
|
20975
|
+
'y',
|
|
20976
|
+
);
|
|
20977
|
+
}
|
|
20978
|
+
return this._nearest_interpolate_4d;
|
|
20979
|
+
}
|
|
20725
20980
|
static get bilinear_interpolate_4d() {
|
|
20726
20981
|
if (!this._bilinear_interpolate_4d) {
|
|
20727
20982
|
this._bilinear_interpolate_4d = wrap(
|
|
@@ -21095,6 +21350,7 @@ class TextClassificationPipeline extends (/** @type {new (options: TextPipelineC
|
|
|
21095
21350
|
|
|
21096
21351
|
// TODO: Use softmax tensor function
|
|
21097
21352
|
const function_to_apply =
|
|
21353
|
+
// @ts-expect-error TS2339
|
|
21098
21354
|
this.model.config.problem_type === 'multi_label_classification'
|
|
21099
21355
|
? batch => batch.sigmoid()
|
|
21100
21356
|
: batch => new _utils_tensor_js__WEBPACK_IMPORTED_MODULE_8__.Tensor(
|
|
@@ -21103,6 +21359,7 @@ class TextClassificationPipeline extends (/** @type {new (options: TextPipelineC
|
|
|
21103
21359
|
batch.dims,
|
|
21104
21360
|
); // single_label_classification (default)
|
|
21105
21361
|
|
|
21362
|
+
// @ts-expect-error TS2339
|
|
21106
21363
|
const id2label = this.model.config.id2label;
|
|
21107
21364
|
|
|
21108
21365
|
const toReturn = [];
|
|
@@ -21205,6 +21462,7 @@ class TokenClassificationPipeline extends (/** @type {new (options: TextPipeline
|
|
|
21205
21462
|
const outputs = await this.model(model_inputs)
|
|
21206
21463
|
|
|
21207
21464
|
const logits = outputs.logits;
|
|
21465
|
+
// @ts-expect-error TS2339
|
|
21208
21466
|
const id2label = this.model.config.id2label;
|
|
21209
21467
|
|
|
21210
21468
|
const toReturn = [];
|
|
@@ -21544,11 +21802,14 @@ class Text2TextGenerationPipeline extends (/** @type {new (options: TextPipeline
|
|
|
21544
21802
|
|
|
21545
21803
|
|
|
21546
21804
|
// Add global prefix, if present
|
|
21805
|
+
// @ts-expect-error TS2339
|
|
21547
21806
|
if (this.model.config.prefix) {
|
|
21807
|
+
// @ts-expect-error TS2339
|
|
21548
21808
|
texts = texts.map(x => this.model.config.prefix + x)
|
|
21549
21809
|
}
|
|
21550
21810
|
|
|
21551
21811
|
// Handle task specific params:
|
|
21812
|
+
// @ts-expect-error TS2339
|
|
21552
21813
|
const task_specific_params = this.model.config.task_specific_params
|
|
21553
21814
|
if (task_specific_params && task_specific_params[this.task]) {
|
|
21554
21815
|
// Add prefixes, if present
|
|
@@ -22287,6 +22548,7 @@ class AudioClassificationPipeline extends (/** @type {new (options: AudioPipelin
|
|
|
22287
22548
|
const sampling_rate = this.processor.feature_extractor.config.sampling_rate;
|
|
22288
22549
|
const preparedAudios = await prepareAudios(audio, sampling_rate);
|
|
22289
22550
|
|
|
22551
|
+
// @ts-expect-error TS2339
|
|
22290
22552
|
const id2label = this.model.config.id2label;
|
|
22291
22553
|
|
|
22292
22554
|
const toReturn = [];
|
|
@@ -22597,6 +22859,7 @@ class AutomaticSpeechRecognitionPipeline extends (/** @type {new (options: TextA
|
|
|
22597
22859
|
audio = [/** @type {AudioInput} */ (audio)];
|
|
22598
22860
|
}
|
|
22599
22861
|
|
|
22862
|
+
// @ts-expect-error TS2339
|
|
22600
22863
|
const time_precision = this.processor.feature_extractor.config.chunk_length / this.model.config.max_source_positions;
|
|
22601
22864
|
const hop_length = this.processor.feature_extractor.config.hop_length;
|
|
22602
22865
|
|
|
@@ -22662,7 +22925,9 @@ class AutomaticSpeechRecognitionPipeline extends (/** @type {new (options: TextA
|
|
|
22662
22925
|
|
|
22663
22926
|
// TODO: Right now we only get top beam
|
|
22664
22927
|
if (return_timestamps === 'word') {
|
|
22928
|
+
// @ts-expect-error TS2339
|
|
22665
22929
|
chunk.tokens = data.sequences.tolist()[0];
|
|
22930
|
+
// @ts-expect-error TS2339
|
|
22666
22931
|
chunk.token_timestamps = data.token_timestamps.tolist()[0].map(
|
|
22667
22932
|
(/** @type {number} */ x) => (0,_utils_maths_js__WEBPACK_IMPORTED_MODULE_6__.round)(x, 2)
|
|
22668
22933
|
);
|
|
@@ -22707,7 +22972,7 @@ class AutomaticSpeechRecognitionPipeline extends (/** @type {new (options: TextA
|
|
|
22707
22972
|
const max_new_tokens = Math.floor(aud.length / sampling_rate) * 6;
|
|
22708
22973
|
const outputs = await this.model.generate({ max_new_tokens, ...kwargs, ...inputs });
|
|
22709
22974
|
|
|
22710
|
-
const text = this.processor.batch_decode(outputs, { skip_special_tokens: true })[0];
|
|
22975
|
+
const text = this.processor.batch_decode(/** @type {Tensor} */(outputs), { skip_special_tokens: true })[0];
|
|
22711
22976
|
toReturn.push({ text });
|
|
22712
22977
|
}
|
|
22713
22978
|
return single ? toReturn[0] : toReturn;
|
|
@@ -22856,6 +23121,7 @@ class ImageClassificationPipeline extends (/** @type {new (options: ImagePipelin
|
|
|
22856
23121
|
const { pixel_values } = await this.processor(preparedImages);
|
|
22857
23122
|
const output = await this.model({ pixel_values });
|
|
22858
23123
|
|
|
23124
|
+
// @ts-expect-error TS2339
|
|
22859
23125
|
const id2label = this.model.config.id2label;
|
|
22860
23126
|
|
|
22861
23127
|
/** @type {ImageClassificationOutput[]} */
|
|
@@ -22970,6 +23236,7 @@ class ImageSegmentationPipeline extends (/** @type {new (options: ImagePipelineC
|
|
|
22970
23236
|
}
|
|
22971
23237
|
}
|
|
22972
23238
|
|
|
23239
|
+
// @ts-expect-error TS2339
|
|
22973
23240
|
const id2label = this.model.config.id2label;
|
|
22974
23241
|
|
|
22975
23242
|
/** @type {ImageSegmentationPipelineOutput[]} */
|
|
@@ -23196,6 +23463,7 @@ class ObjectDetectionPipeline extends (/** @type {new (options: ImagePipelineCon
|
|
|
23196
23463
|
const processed = this.processor.image_processor.post_process_object_detection(output, threshold, imageSizes);
|
|
23197
23464
|
|
|
23198
23465
|
// Add labels
|
|
23466
|
+
// @ts-expect-error TS2339
|
|
23199
23467
|
const id2label = this.model.config.id2label;
|
|
23200
23468
|
|
|
23201
23469
|
// Format output
|
|
@@ -23340,13 +23608,35 @@ class ZeroShotObjectDetectionPipeline extends (/** @type {new (options: TextImag
|
|
|
23340
23608
|
// Run model with both text and pixel inputs
|
|
23341
23609
|
const output = await this.model({ ...text_inputs, pixel_values });
|
|
23342
23610
|
|
|
23343
|
-
|
|
23344
|
-
|
|
23345
|
-
|
|
23346
|
-
|
|
23347
|
-
|
|
23348
|
-
|
|
23349
|
-
|
|
23611
|
+
let result;
|
|
23612
|
+
if('post_process_grounded_object_detection' in this.processor) {
|
|
23613
|
+
// @ts-ignore
|
|
23614
|
+
const processed = this.processor.post_process_grounded_object_detection(
|
|
23615
|
+
output,
|
|
23616
|
+
text_inputs.input_ids,
|
|
23617
|
+
{
|
|
23618
|
+
// TODO: support separate threshold values
|
|
23619
|
+
box_threshold: threshold,
|
|
23620
|
+
text_threshold: threshold,
|
|
23621
|
+
target_sizes: imageSize,
|
|
23622
|
+
},
|
|
23623
|
+
)[0];
|
|
23624
|
+
result = processed.boxes.map((box, i) => ({
|
|
23625
|
+
score: processed.scores[i],
|
|
23626
|
+
label: processed.labels[i],
|
|
23627
|
+
box: get_bounding_box(box, !percentage),
|
|
23628
|
+
}))
|
|
23629
|
+
} else {
|
|
23630
|
+
// @ts-ignore
|
|
23631
|
+
const processed = this.processor.image_processor.post_process_object_detection(output, threshold, imageSize, true)[0];
|
|
23632
|
+
result = processed.boxes.map((box, i) => ({
|
|
23633
|
+
score: processed.scores[i],
|
|
23634
|
+
label: candidate_labels[processed.classes[i]],
|
|
23635
|
+
box: get_bounding_box(box, !percentage),
|
|
23636
|
+
}))
|
|
23637
|
+
}
|
|
23638
|
+
result.sort((a, b) => b.score - a.score);
|
|
23639
|
+
|
|
23350
23640
|
if (top_k !== null) {
|
|
23351
23641
|
result = result.slice(0, top_k);
|
|
23352
23642
|
}
|
|
@@ -23415,6 +23705,7 @@ class DocumentQuestionAnsweringPipeline extends (/** @type {new (options: TextIm
|
|
|
23415
23705
|
// Run model
|
|
23416
23706
|
const output = await this.model.generate({
|
|
23417
23707
|
inputs: pixel_values,
|
|
23708
|
+
// @ts-expect-error TS2339
|
|
23418
23709
|
max_length: this.model.config.decoder.max_position_embeddings,
|
|
23419
23710
|
decoder_input_ids,
|
|
23420
23711
|
...generate_kwargs,
|
|
@@ -23465,7 +23756,7 @@ class DocumentQuestionAnsweringPipeline extends (/** @type {new (options: TextIm
|
|
|
23465
23756
|
* const synthesizer = await pipeline('text-to-speech', 'Xenova/speecht5_tts', { quantized: false });
|
|
23466
23757
|
* const speaker_embeddings = 'https://huggingface.co/datasets/Xenova/transformers.js-docs/resolve/main/speaker_embeddings.bin';
|
|
23467
23758
|
* const out = await synthesizer('Hello, my dog is cute', { speaker_embeddings });
|
|
23468
|
-
* // {
|
|
23759
|
+
* // RawAudio {
|
|
23469
23760
|
* // audio: Float32Array(26112) [-0.00005657337896991521, 0.00020583874720614403, ...],
|
|
23470
23761
|
* // sampling_rate: 16000
|
|
23471
23762
|
* // }
|
|
@@ -23485,7 +23776,7 @@ class DocumentQuestionAnsweringPipeline extends (/** @type {new (options: TextIm
|
|
|
23485
23776
|
* ```javascript
|
|
23486
23777
|
* const synthesizer = await pipeline('text-to-speech', 'Xenova/mms-tts-fra');
|
|
23487
23778
|
* const out = await synthesizer('Bonjour');
|
|
23488
|
-
* // {
|
|
23779
|
+
* // RawAudio {
|
|
23489
23780
|
* // audio: Float32Array(23808) [-0.00037693005288019776, 0.0003325853613205254, ...],
|
|
23490
23781
|
* // sampling_rate: 16000
|
|
23491
23782
|
* // }
|
|
@@ -23530,11 +23821,12 @@ class TextToAudioPipeline extends (/** @type {new (options: TextToAudioPipelineC
|
|
|
23530
23821
|
// Generate waveform
|
|
23531
23822
|
const { waveform } = await this.model(inputs);
|
|
23532
23823
|
|
|
23824
|
+
// @ts-expect-error TS2339
|
|
23533
23825
|
const sampling_rate = this.model.config.sampling_rate;
|
|
23534
|
-
return
|
|
23535
|
-
|
|
23826
|
+
return new _utils_audio_js__WEBPACK_IMPORTED_MODULE_7__.RawAudio(
|
|
23827
|
+
waveform.data,
|
|
23536
23828
|
sampling_rate,
|
|
23537
|
-
|
|
23829
|
+
)
|
|
23538
23830
|
}
|
|
23539
23831
|
|
|
23540
23832
|
async _call_text_to_spectrogram(text_inputs, { speaker_embeddings }) {
|
|
@@ -23574,10 +23866,10 @@ class TextToAudioPipeline extends (/** @type {new (options: TextToAudioPipelineC
|
|
|
23574
23866
|
const { waveform } = await this.model.generate_speech(input_ids, speaker_embeddings, { vocoder: this.vocoder });
|
|
23575
23867
|
|
|
23576
23868
|
const sampling_rate = this.processor.feature_extractor.config.sampling_rate;
|
|
23577
|
-
return
|
|
23578
|
-
|
|
23869
|
+
return new _utils_audio_js__WEBPACK_IMPORTED_MODULE_7__.RawAudio(
|
|
23870
|
+
waveform.data,
|
|
23579
23871
|
sampling_rate,
|
|
23580
|
-
|
|
23872
|
+
)
|
|
23581
23873
|
}
|
|
23582
23874
|
}
|
|
23583
23875
|
|
|
@@ -23687,11 +23979,23 @@ class DepthEstimationPipeline extends (/** @type {new (options: ImagePipelineCon
|
|
|
23687
23979
|
|
|
23688
23980
|
const toReturn = [];
|
|
23689
23981
|
for (let i = 0; i < preparedImages.length; ++i) {
|
|
23690
|
-
const
|
|
23691
|
-
const
|
|
23982
|
+
const batch = predicted_depth[i];
|
|
23983
|
+
const [height, width] = batch.dims.slice(-2);
|
|
23984
|
+
const [new_width, new_height] = preparedImages[i].size;
|
|
23985
|
+
|
|
23986
|
+
// Interpolate to original size
|
|
23987
|
+
const prediction = (await (0,_utils_tensor_js__WEBPACK_IMPORTED_MODULE_8__.interpolate_4d)(batch.view(1, 1, height, width), {
|
|
23988
|
+
size: [new_height, new_width],
|
|
23989
|
+
mode: 'bilinear',
|
|
23990
|
+
})).view(new_height, new_width);
|
|
23991
|
+
|
|
23992
|
+
const minval = /** @type {number} */(prediction.min().item());
|
|
23993
|
+
const maxval = /** @type {number} */(prediction.max().item());
|
|
23994
|
+
const formatted = prediction.sub(minval).div_(maxval - minval).mul_(255).to('uint8').unsqueeze(0);
|
|
23995
|
+
const depth = _utils_image_js__WEBPACK_IMPORTED_MODULE_9__.RawImage.fromTensor(formatted);
|
|
23692
23996
|
toReturn.push({
|
|
23693
|
-
predicted_depth:
|
|
23694
|
-
depth
|
|
23997
|
+
predicted_depth: prediction,
|
|
23998
|
+
depth,
|
|
23695
23999
|
});
|
|
23696
24000
|
}
|
|
23697
24001
|
|
|
@@ -24171,6 +24475,7 @@ async function loadItems(mapping, model, pretrainedOptions) {
|
|
|
24171
24475
|
return result;
|
|
24172
24476
|
}
|
|
24173
24477
|
|
|
24478
|
+
|
|
24174
24479
|
/***/ }),
|
|
24175
24480
|
|
|
24176
24481
|
/***/ "./src/tokenizers.js":
|
|
@@ -24239,7 +24544,6 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
24239
24544
|
/* harmony import */ var _utils_data_structures_js__WEBPACK_IMPORTED_MODULE_5__ = __webpack_require__(/*! ./utils/data-structures.js */ "./src/utils/data-structures.js");
|
|
24240
24545
|
/* harmony import */ var _huggingface_jinja__WEBPACK_IMPORTED_MODULE_6__ = __webpack_require__(/*! @huggingface/jinja */ "./node_modules/@huggingface/jinja/dist/index.js");
|
|
24241
24546
|
/* harmony import */ var _models_whisper_common_whisper_js__WEBPACK_IMPORTED_MODULE_7__ = __webpack_require__(/*! ./models/whisper/common_whisper.js */ "./src/models/whisper/common_whisper.js");
|
|
24242
|
-
/* harmony import */ var _utils_constants_js__WEBPACK_IMPORTED_MODULE_8__ = __webpack_require__(/*! ./utils/constants.js */ "./src/utils/constants.js");
|
|
24243
24547
|
|
|
24244
24548
|
/**
|
|
24245
24549
|
* @file Tokenizers are used to prepare textual inputs for a model.
|
|
@@ -24276,7 +24580,6 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
24276
24580
|
|
|
24277
24581
|
|
|
24278
24582
|
|
|
24279
|
-
|
|
24280
24583
|
/**
|
|
24281
24584
|
* @typedef {Object} TokenizerProperties Additional tokenizer-specific properties.
|
|
24282
24585
|
* @property {boolean} [legacy=false] Whether or not the `legacy` behavior of the tokenizer should be used.
|
|
@@ -24760,7 +25063,7 @@ class Unigram extends TokenizerModel {
|
|
|
24760
25063
|
* Create a new Unigram tokenizer model.
|
|
24761
25064
|
* @param {Object} config The configuration object for the Unigram model.
|
|
24762
25065
|
* @param {number} config.unk_id The ID of the unknown token
|
|
24763
|
-
* @param {
|
|
25066
|
+
* @param {[string, number][]} config.vocab A 2D array representing a mapping of tokens to scores.
|
|
24764
25067
|
* @param {Object} moreConfig Additional configuration object for the Unigram model.
|
|
24765
25068
|
*/
|
|
24766
25069
|
constructor(config, moreConfig) {
|
|
@@ -24768,11 +25071,10 @@ class Unigram extends TokenizerModel {
|
|
|
24768
25071
|
|
|
24769
25072
|
const vocabSize = config.vocab.length;
|
|
24770
25073
|
this.vocab = new Array(vocabSize);
|
|
25074
|
+
/** @type {number[]} */
|
|
24771
25075
|
this.scores = new Array(vocabSize);
|
|
24772
25076
|
for (let i = 0; i < vocabSize; ++i) {
|
|
24773
|
-
|
|
24774
|
-
this.vocab[i] = piece[0];
|
|
24775
|
-
this.scores[i] = piece[1];
|
|
25077
|
+
[this.vocab[i], this.scores[i]] = config.vocab[i];
|
|
24776
25078
|
}
|
|
24777
25079
|
|
|
24778
25080
|
this.unk_token_id = config.unk_id;
|
|
@@ -28611,6 +28913,7 @@ class AutoTokenizer {
|
|
|
28611
28913
|
|
|
28612
28914
|
__webpack_require__.r(__webpack_exports__);
|
|
28613
28915
|
/* harmony export */ __webpack_require__.d(__webpack_exports__, {
|
|
28916
|
+
/* harmony export */ RawAudio: () => (/* binding */ RawAudio),
|
|
28614
28917
|
/* harmony export */ hamming: () => (/* binding */ hamming),
|
|
28615
28918
|
/* harmony export */ hanning: () => (/* binding */ hanning),
|
|
28616
28919
|
/* harmony export */ mel_filter_bank: () => (/* binding */ mel_filter_bank),
|
|
@@ -28621,7 +28924,9 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
28621
28924
|
/* harmony import */ var _hub_js__WEBPACK_IMPORTED_MODULE_0__ = __webpack_require__(/*! ./hub.js */ "./src/utils/hub.js");
|
|
28622
28925
|
/* harmony import */ var _maths_js__WEBPACK_IMPORTED_MODULE_1__ = __webpack_require__(/*! ./maths.js */ "./src/utils/maths.js");
|
|
28623
28926
|
/* harmony import */ var _core_js__WEBPACK_IMPORTED_MODULE_2__ = __webpack_require__(/*! ./core.js */ "./src/utils/core.js");
|
|
28624
|
-
/* harmony import */ var
|
|
28927
|
+
/* harmony import */ var _env_js__WEBPACK_IMPORTED_MODULE_3__ = __webpack_require__(/*! ../env.js */ "./src/env.js");
|
|
28928
|
+
/* harmony import */ var fs__WEBPACK_IMPORTED_MODULE_4__ = __webpack_require__(/*! fs */ "fs");
|
|
28929
|
+
/* harmony import */ var _tensor_js__WEBPACK_IMPORTED_MODULE_5__ = __webpack_require__(/*! ./tensor.js */ "./src/utils/tensor.js");
|
|
28625
28930
|
/**
|
|
28626
28931
|
* @file Helper module for audio processing.
|
|
28627
28932
|
*
|
|
@@ -28637,6 +28942,8 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
28637
28942
|
|
|
28638
28943
|
|
|
28639
28944
|
|
|
28945
|
+
|
|
28946
|
+
|
|
28640
28947
|
/**
|
|
28641
28948
|
* Helper function to read audio from a path/URL.
|
|
28642
28949
|
* @param {string|URL} url The path/URL to load the audio from.
|
|
@@ -29230,10 +29537,10 @@ async function spectrogram(
|
|
|
29230
29537
|
// - mel_filters.shape=(80, 201)
|
|
29231
29538
|
// - magnitudes.shape=(3000, 201) => magnitudes.T.shape=(201, 3000)
|
|
29232
29539
|
// - mel_spec.shape=(80, 3000)
|
|
29233
|
-
let mel_spec = await (0,
|
|
29540
|
+
let mel_spec = await (0,_tensor_js__WEBPACK_IMPORTED_MODULE_5__.matmul)(
|
|
29234
29541
|
// TODO: Make `mel_filters` a Tensor during initialization
|
|
29235
|
-
new
|
|
29236
|
-
new
|
|
29542
|
+
new _tensor_js__WEBPACK_IMPORTED_MODULE_5__.Tensor('float32', mel_filters.flat(), [num_mel_filters, num_frequency_bins]),
|
|
29543
|
+
new _tensor_js__WEBPACK_IMPORTED_MODULE_5__.Tensor('float32', transposedMagnitudeData, [num_frequency_bins, d1Max]),
|
|
29237
29544
|
);
|
|
29238
29545
|
if (transpose) {
|
|
29239
29546
|
mel_spec = mel_spec.transpose(1, 0);
|
|
@@ -29323,6 +29630,116 @@ function window_function(window_length, name, {
|
|
|
29323
29630
|
return window;
|
|
29324
29631
|
}
|
|
29325
29632
|
|
|
29633
|
+
/**
|
|
29634
|
+
* Encode audio data to a WAV file.
|
|
29635
|
+
* WAV file specs : https://en.wikipedia.org/wiki/WAV#WAV_File_header
|
|
29636
|
+
*
|
|
29637
|
+
* Adapted from https://www.npmjs.com/package/audiobuffer-to-wav
|
|
29638
|
+
* @param {Float32Array} samples The audio samples.
|
|
29639
|
+
* @param {number} rate The sample rate.
|
|
29640
|
+
* @returns {ArrayBuffer} The WAV audio buffer.
|
|
29641
|
+
*/
|
|
29642
|
+
function encodeWAV(samples, rate) {
|
|
29643
|
+
let offset = 44;
|
|
29644
|
+
const buffer = new ArrayBuffer(offset + samples.length * 4);
|
|
29645
|
+
const view = new DataView(buffer);
|
|
29646
|
+
|
|
29647
|
+
/* RIFF identifier */
|
|
29648
|
+
writeString(view, 0, "RIFF");
|
|
29649
|
+
/* RIFF chunk length */
|
|
29650
|
+
view.setUint32(4, 36 + samples.length * 4, true);
|
|
29651
|
+
/* RIFF type */
|
|
29652
|
+
writeString(view, 8, "WAVE");
|
|
29653
|
+
/* format chunk identifier */
|
|
29654
|
+
writeString(view, 12, "fmt ");
|
|
29655
|
+
/* format chunk length */
|
|
29656
|
+
view.setUint32(16, 16, true);
|
|
29657
|
+
/* sample format (raw) */
|
|
29658
|
+
view.setUint16(20, 3, true);
|
|
29659
|
+
/* channel count */
|
|
29660
|
+
view.setUint16(22, 1, true);
|
|
29661
|
+
/* sample rate */
|
|
29662
|
+
view.setUint32(24, rate, true);
|
|
29663
|
+
/* byte rate (sample rate * block align) */
|
|
29664
|
+
view.setUint32(28, rate * 4, true);
|
|
29665
|
+
/* block align (channel count * bytes per sample) */
|
|
29666
|
+
view.setUint16(32, 4, true);
|
|
29667
|
+
/* bits per sample */
|
|
29668
|
+
view.setUint16(34, 32, true);
|
|
29669
|
+
/* data chunk identifier */
|
|
29670
|
+
writeString(view, 36, "data");
|
|
29671
|
+
/* data chunk length */
|
|
29672
|
+
view.setUint32(40, samples.length * 4, true);
|
|
29673
|
+
|
|
29674
|
+
for (let i = 0; i < samples.length; ++i, offset += 4) {
|
|
29675
|
+
view.setFloat32(offset, samples[i], true);
|
|
29676
|
+
}
|
|
29677
|
+
|
|
29678
|
+
return buffer;
|
|
29679
|
+
}
|
|
29680
|
+
|
|
29681
|
+
function writeString(view, offset, string) {
|
|
29682
|
+
for (let i = 0; i < string.length; ++i) {
|
|
29683
|
+
view.setUint8(offset + i, string.charCodeAt(i));
|
|
29684
|
+
}
|
|
29685
|
+
}
|
|
29686
|
+
|
|
29687
|
+
|
|
29688
|
+
class RawAudio {
|
|
29689
|
+
|
|
29690
|
+
/**
|
|
29691
|
+
* Create a new `RawAudio` object.
|
|
29692
|
+
* @param {Float32Array} audio Audio data
|
|
29693
|
+
* @param {number} sampling_rate Sampling rate of the audio data
|
|
29694
|
+
*/
|
|
29695
|
+
constructor(audio, sampling_rate) {
|
|
29696
|
+
this.audio = audio
|
|
29697
|
+
this.sampling_rate = sampling_rate
|
|
29698
|
+
}
|
|
29699
|
+
|
|
29700
|
+
/**
|
|
29701
|
+
* Convert the audio to a wav file buffer.
|
|
29702
|
+
* @returns {ArrayBuffer} The WAV file.
|
|
29703
|
+
*/
|
|
29704
|
+
toWav() {
|
|
29705
|
+
return encodeWAV(this.audio, this.sampling_rate)
|
|
29706
|
+
}
|
|
29707
|
+
|
|
29708
|
+
/**
|
|
29709
|
+
* Convert the audio to a blob.
|
|
29710
|
+
* @returns {Blob}
|
|
29711
|
+
*/
|
|
29712
|
+
toBlob() {
|
|
29713
|
+
const wav = this.toWav();
|
|
29714
|
+
const blob = new Blob([wav], { type: 'audio/wav' });
|
|
29715
|
+
return blob;
|
|
29716
|
+
}
|
|
29717
|
+
|
|
29718
|
+
/**
|
|
29719
|
+
* Save the audio to a wav file.
|
|
29720
|
+
* @param {string} path
|
|
29721
|
+
*/
|
|
29722
|
+
async save(path) {
|
|
29723
|
+
let fn;
|
|
29724
|
+
|
|
29725
|
+
if (_env_js__WEBPACK_IMPORTED_MODULE_3__.apis.IS_BROWSER_ENV) {
|
|
29726
|
+
if (_env_js__WEBPACK_IMPORTED_MODULE_3__.apis.IS_WEBWORKER_ENV) {
|
|
29727
|
+
throw new Error('Unable to save a file from a Web Worker.')
|
|
29728
|
+
}
|
|
29729
|
+
fn = _core_js__WEBPACK_IMPORTED_MODULE_2__.saveBlob;
|
|
29730
|
+
} else if (_env_js__WEBPACK_IMPORTED_MODULE_3__.apis.IS_FS_AVAILABLE) {
|
|
29731
|
+
fn = async (/** @type {string} */ path, /** @type {Blob} */ blob) => {
|
|
29732
|
+
let buffer = await blob.arrayBuffer();
|
|
29733
|
+
fs__WEBPACK_IMPORTED_MODULE_4__["default"].writeFileSync(path, Buffer.from(buffer));
|
|
29734
|
+
}
|
|
29735
|
+
} else {
|
|
29736
|
+
throw new Error('Unable to save because filesystem is disabled in this environment.')
|
|
29737
|
+
}
|
|
29738
|
+
|
|
29739
|
+
await fn(path, this.toBlob())
|
|
29740
|
+
}
|
|
29741
|
+
}
|
|
29742
|
+
|
|
29326
29743
|
|
|
29327
29744
|
/***/ }),
|
|
29328
29745
|
|
|
@@ -29376,7 +29793,8 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
29376
29793
|
/* harmony export */ pick: () => (/* binding */ pick),
|
|
29377
29794
|
/* harmony export */ pop: () => (/* binding */ pop),
|
|
29378
29795
|
/* harmony export */ product: () => (/* binding */ product),
|
|
29379
|
-
/* harmony export */ reverseDictionary: () => (/* binding */ reverseDictionary)
|
|
29796
|
+
/* harmony export */ reverseDictionary: () => (/* binding */ reverseDictionary),
|
|
29797
|
+
/* harmony export */ saveBlob: () => (/* binding */ saveBlob)
|
|
29380
29798
|
/* harmony export */ });
|
|
29381
29799
|
|
|
29382
29800
|
/**
|
|
@@ -29569,6 +29987,32 @@ function calculateReflectOffset(i, w) {
|
|
|
29569
29987
|
return Math.abs((i + w) % (2 * w) - w);
|
|
29570
29988
|
}
|
|
29571
29989
|
|
|
29990
|
+
/**
|
|
29991
|
+
* Save blob file on the web.
|
|
29992
|
+
* @param {string} path The path to save the blob to
|
|
29993
|
+
* @param {Blob} blob The blob to save
|
|
29994
|
+
*/
|
|
29995
|
+
function saveBlob(path, blob){
|
|
29996
|
+
// Convert the canvas content to a data URL
|
|
29997
|
+
const dataURL = URL.createObjectURL(blob);
|
|
29998
|
+
|
|
29999
|
+
// Create an anchor element with the data URL as the href attribute
|
|
30000
|
+
const downloadLink = document.createElement('a');
|
|
30001
|
+
downloadLink.href = dataURL;
|
|
30002
|
+
|
|
30003
|
+
// Set the download attribute to specify the desired filename for the downloaded image
|
|
30004
|
+
downloadLink.download = path;
|
|
30005
|
+
|
|
30006
|
+
// Trigger the download
|
|
30007
|
+
downloadLink.click();
|
|
30008
|
+
|
|
30009
|
+
// Clean up: remove the anchor element from the DOM
|
|
30010
|
+
downloadLink.remove();
|
|
30011
|
+
|
|
30012
|
+
// Revoke the Object URL to free up memory
|
|
30013
|
+
URL.revokeObjectURL(dataURL);
|
|
30014
|
+
}
|
|
30015
|
+
|
|
29572
30016
|
/**
|
|
29573
30017
|
*
|
|
29574
30018
|
* @param {Object} o
|
|
@@ -30129,6 +30573,8 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
30129
30573
|
/* harmony export */ });
|
|
30130
30574
|
/* harmony import */ var _env_js__WEBPACK_IMPORTED_MODULE_0__ = __webpack_require__(/*! ../env.js */ "./src/env.js");
|
|
30131
30575
|
/* harmony import */ var _devices_js__WEBPACK_IMPORTED_MODULE_1__ = __webpack_require__(/*! ./devices.js */ "./src/utils/devices.js");
|
|
30576
|
+
/// <reference types="@webgpu/types" />
|
|
30577
|
+
|
|
30132
30578
|
|
|
30133
30579
|
|
|
30134
30580
|
|
|
@@ -30382,7 +30828,7 @@ class FileResponse {
|
|
|
30382
30828
|
*/
|
|
30383
30829
|
async arrayBuffer() {
|
|
30384
30830
|
const data = await fs__WEBPACK_IMPORTED_MODULE_0__["default"].promises.readFile(this.filePath);
|
|
30385
|
-
return data.buffer;
|
|
30831
|
+
return /** @type {ArrayBuffer} */ (data.buffer);
|
|
30386
30832
|
}
|
|
30387
30833
|
|
|
30388
30834
|
/**
|
|
@@ -31743,23 +32189,9 @@ class RawImage {
|
|
|
31743
32189
|
// Convert image to Blob
|
|
31744
32190
|
const blob = await this.toBlob(mime);
|
|
31745
32191
|
|
|
31746
|
-
|
|
31747
|
-
const dataURL = URL.createObjectURL(blob);
|
|
31748
|
-
|
|
31749
|
-
// Create an anchor element with the data URL as the href attribute
|
|
31750
|
-
const downloadLink = document.createElement('a');
|
|
31751
|
-
downloadLink.href = dataURL;
|
|
32192
|
+
(0,_core_js__WEBPACK_IMPORTED_MODULE_0__.saveBlob)(path, blob)
|
|
31752
32193
|
|
|
31753
|
-
|
|
31754
|
-
downloadLink.download = path;
|
|
31755
|
-
|
|
31756
|
-
// Trigger the download
|
|
31757
|
-
downloadLink.click();
|
|
31758
|
-
|
|
31759
|
-
// Clean up: remove the anchor element from the DOM
|
|
31760
|
-
downloadLink.remove();
|
|
31761
|
-
|
|
31762
|
-
} else if (!_env_js__WEBPACK_IMPORTED_MODULE_2__.env.useFS) {
|
|
32194
|
+
} else if (!_env_js__WEBPACK_IMPORTED_MODULE_2__.apis.IS_FS_AVAILABLE) {
|
|
31763
32195
|
throw new Error('Unable to save the image because filesystem is disabled in this environment.')
|
|
31764
32196
|
|
|
31765
32197
|
} else {
|
|
@@ -31789,6 +32221,7 @@ class RawImage {
|
|
|
31789
32221
|
const load_image = RawImage.read.bind(RawImage);
|
|
31790
32222
|
|
|
31791
32223
|
|
|
32224
|
+
|
|
31792
32225
|
/***/ }),
|
|
31793
32226
|
|
|
31794
32227
|
/***/ "./src/utils/maths.js":
|
|
@@ -32041,8 +32474,9 @@ function magnitude(arr) {
|
|
|
32041
32474
|
|
|
32042
32475
|
/**
|
|
32043
32476
|
* Returns the value and index of the minimum element in an array.
|
|
32044
|
-
* @
|
|
32045
|
-
* @
|
|
32477
|
+
* @template {number[]|bigint[]|AnyTypedArray} T
|
|
32478
|
+
* @param {T} arr array of numbers.
|
|
32479
|
+
* @returns {T extends bigint[]|BigTypedArray ? [bigint, number] : [number, number]} the value and index of the minimum element, of the form: [valueOfMin, indexOfMin]
|
|
32046
32480
|
* @throws {Error} If array is empty.
|
|
32047
32481
|
*/
|
|
32048
32482
|
function min(arr) {
|
|
@@ -32055,14 +32489,15 @@ function min(arr) {
|
|
|
32055
32489
|
indexOfMin = i;
|
|
32056
32490
|
}
|
|
32057
32491
|
}
|
|
32058
|
-
return [min, indexOfMin];
|
|
32492
|
+
return /** @type {T extends bigint[]|BigTypedArray ? [bigint, number] : [number, number]} */([min, indexOfMin]);
|
|
32059
32493
|
}
|
|
32060
32494
|
|
|
32061
32495
|
|
|
32062
32496
|
/**
|
|
32063
32497
|
* Returns the value and index of the maximum element in an array.
|
|
32064
|
-
* @
|
|
32065
|
-
* @
|
|
32498
|
+
* @template {number[]|bigint[]|AnyTypedArray} T
|
|
32499
|
+
* @param {T} arr array of numbers.
|
|
32500
|
+
* @returns {T extends bigint[]|BigTypedArray ? [bigint, number] : [number, number]} the value and index of the maximum element, of the form: [valueOfMax, indexOfMax]
|
|
32066
32501
|
* @throws {Error} If array is empty.
|
|
32067
32502
|
*/
|
|
32068
32503
|
function max(arr) {
|
|
@@ -32075,7 +32510,7 @@ function max(arr) {
|
|
|
32075
32510
|
indexOfMax = i;
|
|
32076
32511
|
}
|
|
32077
32512
|
}
|
|
32078
|
-
return [
|
|
32513
|
+
return /** @type {T extends bigint[]|BigTypedArray ? [bigint, number] : [number, number]} */([max, indexOfMax]);
|
|
32079
32514
|
}
|
|
32080
32515
|
|
|
32081
32516
|
function isPowerOfTwo(number) {
|
|
@@ -33372,8 +33807,6 @@ class Tensor {
|
|
|
33372
33807
|
return this.permute(...dims);
|
|
33373
33808
|
}
|
|
33374
33809
|
|
|
33375
|
-
// TODO add .max() and .min() methods
|
|
33376
|
-
|
|
33377
33810
|
/**
|
|
33378
33811
|
* Returns the sum of each row of the input tensor in the given dimension dim.
|
|
33379
33812
|
*
|
|
@@ -33402,55 +33835,22 @@ class Tensor {
|
|
|
33402
33835
|
}
|
|
33403
33836
|
|
|
33404
33837
|
const this_data = this.data;
|
|
33838
|
+
const fn = (a, b) => a + (b ** p);
|
|
33405
33839
|
|
|
33406
33840
|
if (dim === null) {
|
|
33407
33841
|
// @ts-ignore
|
|
33408
|
-
|
|
33842
|
+
const val = this_data.reduce(fn, 0) ** (1 / p);
|
|
33409
33843
|
return new Tensor(this.type, [val], []);
|
|
33410
33844
|
}
|
|
33411
33845
|
|
|
33412
|
-
|
|
33413
|
-
dim = safeIndex(dim, this.dims.length);
|
|
33414
|
-
|
|
33415
|
-
// Calculate the shape of the resulting array after summation
|
|
33416
|
-
const resultDims = this.dims.slice(); // Copy the original dimensions
|
|
33417
|
-
resultDims[dim] = 1; // Remove the specified axis
|
|
33418
|
-
|
|
33419
|
-
// Create a new array to store the accumulated values
|
|
33420
|
-
// @ts-ignore
|
|
33421
|
-
const result = new this_data.constructor(this_data.length / this.dims[dim]);
|
|
33422
|
-
|
|
33423
|
-
// Iterate over the data array
|
|
33424
|
-
for (let i = 0; i < this_data.length; ++i) {
|
|
33425
|
-
|
|
33426
|
-
// Calculate the index in the resulting array
|
|
33427
|
-
let resultIndex = 0;
|
|
33428
|
-
|
|
33429
|
-
for (let j = this.dims.length - 1, num = i, resultMultiplier = 1; j >= 0; --j) {
|
|
33430
|
-
const size = this.dims[j];
|
|
33431
|
-
if (j !== dim) {
|
|
33432
|
-
const index = num % size;
|
|
33433
|
-
resultIndex += index * resultMultiplier;
|
|
33434
|
-
resultMultiplier *= resultDims[j];
|
|
33435
|
-
}
|
|
33436
|
-
num = Math.floor(num / size);
|
|
33437
|
-
}
|
|
33438
|
-
|
|
33439
|
-
// Accumulate the value at the current index
|
|
33440
|
-
result[resultIndex] += (this_data[i]) ** p;
|
|
33441
|
-
}
|
|
33846
|
+
const [type, result, resultDims] = reduce_helper(fn, this, dim, keepdim);
|
|
33442
33847
|
|
|
33443
33848
|
if (p !== 1) {
|
|
33444
33849
|
for (let i = 0; i < result.length; ++i) {
|
|
33445
33850
|
result[i] = result[i] ** (1 / p);
|
|
33446
33851
|
}
|
|
33447
33852
|
}
|
|
33448
|
-
|
|
33449
|
-
if (!keepdim) {
|
|
33450
|
-
resultDims.splice(dim, 1);
|
|
33451
|
-
}
|
|
33452
|
-
|
|
33453
|
-
return new Tensor(this.type, result, resultDims);
|
|
33853
|
+
return new Tensor(type, result, resultDims);
|
|
33454
33854
|
}
|
|
33455
33855
|
|
|
33456
33856
|
/**
|
|
@@ -33513,7 +33913,7 @@ class Tensor {
|
|
|
33513
33913
|
* NOTE: The returned tensor shares the storage with the input tensor, so changing the contents of one will change the contents of the other.
|
|
33514
33914
|
* If you would like a copy, use `tensor.clone()` before squeezing.
|
|
33515
33915
|
*
|
|
33516
|
-
* @param {number} [dim=null] If given, the input will be squeezed only in the specified dimensions.
|
|
33916
|
+
* @param {number|number[]} [dim=null] If given, the input will be squeezed only in the specified dimensions.
|
|
33517
33917
|
* @returns {Tensor} The squeezed tensor
|
|
33518
33918
|
*/
|
|
33519
33919
|
squeeze(dim = null) {
|
|
@@ -33623,6 +34023,34 @@ class Tensor {
|
|
|
33623
34023
|
return this.clone().neg_();
|
|
33624
34024
|
}
|
|
33625
34025
|
|
|
34026
|
+
/**
|
|
34027
|
+
* Computes input > val element-wise.
|
|
34028
|
+
* @param {number} val The value to compare with.
|
|
34029
|
+
* @returns {Tensor} A boolean tensor that is `true` where input is greater than other and `false` elsewhere.
|
|
34030
|
+
*/
|
|
34031
|
+
gt(val) {
|
|
34032
|
+
const mask = new Uint8Array(this.data.length);
|
|
34033
|
+
const this_data = this.data;
|
|
34034
|
+
for (let i = 0; i < this_data.length; ++i) {
|
|
34035
|
+
mask[i] = this_data[i] > val ? 1 : 0;
|
|
34036
|
+
}
|
|
34037
|
+
return new Tensor('bool', mask, this.dims);
|
|
34038
|
+
}
|
|
34039
|
+
|
|
34040
|
+
/**
|
|
34041
|
+
* Computes input < val element-wise.
|
|
34042
|
+
* @param {number} val The value to compare with.
|
|
34043
|
+
* @returns {Tensor} A boolean tensor that is `true` where input is less than other and `false` elsewhere.
|
|
34044
|
+
*/
|
|
34045
|
+
lt(val) {
|
|
34046
|
+
const mask = new Uint8Array(this.data.length);
|
|
34047
|
+
const this_data = this.data;
|
|
34048
|
+
for (let i = 0; i < this_data.length; ++i) {
|
|
34049
|
+
mask[i] = this_data[i] < val ? 1 : 0;
|
|
34050
|
+
}
|
|
34051
|
+
return new Tensor('bool', mask, this.dims);
|
|
34052
|
+
}
|
|
34053
|
+
|
|
33626
34054
|
/**
|
|
33627
34055
|
* In-place version of @see {@link Tensor.clamp}
|
|
33628
34056
|
*/
|
|
@@ -33667,6 +34095,41 @@ class Tensor {
|
|
|
33667
34095
|
return mean(this, dim, keepdim);
|
|
33668
34096
|
}
|
|
33669
34097
|
|
|
34098
|
+
min(dim = null, keepdim = false) {
|
|
34099
|
+
if (dim === null) {
|
|
34100
|
+
// None to reduce over all dimensions.
|
|
34101
|
+
const val = (0,_maths_js__WEBPACK_IMPORTED_MODULE_0__.min)(this.data)[0];
|
|
34102
|
+
return new Tensor(this.type, [val], [/* scalar */]);
|
|
34103
|
+
}
|
|
34104
|
+
const [type, result, resultDims] = reduce_helper((a, b) => Math.min(a, b), this, dim, keepdim, Infinity);
|
|
34105
|
+
return new Tensor(type, result, resultDims);
|
|
34106
|
+
}
|
|
34107
|
+
|
|
34108
|
+
max(dim = null, keepdim = false) {
|
|
34109
|
+
if (dim === null) {
|
|
34110
|
+
// None to reduce over all dimensions.
|
|
34111
|
+
const val = (0,_maths_js__WEBPACK_IMPORTED_MODULE_0__.max)(this.data)[0];
|
|
34112
|
+
return new Tensor(this.type, [val], [/* scalar */]);
|
|
34113
|
+
}
|
|
34114
|
+
const [type, result, resultDims] = reduce_helper((a, b) => Math.max(a, b), this, dim, keepdim, -Infinity);
|
|
34115
|
+
return new Tensor(type, result, resultDims);
|
|
34116
|
+
}
|
|
34117
|
+
|
|
34118
|
+
argmin(dim = null, keepdim = false) {
|
|
34119
|
+
if (dim !== null) {
|
|
34120
|
+
throw new Error("`dim !== null` not yet implemented.");
|
|
34121
|
+
}
|
|
34122
|
+
const index = (0,_maths_js__WEBPACK_IMPORTED_MODULE_0__.min)(this.data)[1];
|
|
34123
|
+
return new Tensor('int64', [BigInt(index)], []);
|
|
34124
|
+
}
|
|
34125
|
+
argmax(dim = null, keepdim = false) {
|
|
34126
|
+
if (dim !== null) {
|
|
34127
|
+
throw new Error("`dim !== null` not yet implemented.");
|
|
34128
|
+
}
|
|
34129
|
+
const index = (0,_maths_js__WEBPACK_IMPORTED_MODULE_0__.max)(this.data)[1];
|
|
34130
|
+
return new Tensor('int64', [BigInt(index)], []);
|
|
34131
|
+
}
|
|
34132
|
+
|
|
33670
34133
|
/**
|
|
33671
34134
|
* Performs Tensor dtype conversion.
|
|
33672
34135
|
* @param {DataType} type The desired data type.
|
|
@@ -33800,7 +34263,7 @@ function interpolate(input, [out_height, out_width], mode = 'bilinear', align_co
|
|
|
33800
34263
|
* @param {Tensor} input the input tensor
|
|
33801
34264
|
* @param {Object} options the options for the interpolation
|
|
33802
34265
|
* @param {[number, number]|[number, number, number]|[number, number, number, number]} [options.size=null] output spatial size.
|
|
33803
|
-
* @param {"bilinear"|"bicubic"} [options.mode='bilinear'] algorithm used for upsampling
|
|
34266
|
+
* @param {"nearest"|"bilinear"|"bicubic"} [options.mode='bilinear'] algorithm used for upsampling
|
|
33804
34267
|
* @returns {Promise<Tensor>} The interpolated tensor.
|
|
33805
34268
|
*/
|
|
33806
34269
|
async function interpolate_4d(input, {
|
|
@@ -33830,7 +34293,9 @@ async function interpolate_4d(input, {
|
|
|
33830
34293
|
}
|
|
33831
34294
|
|
|
33832
34295
|
let op;
|
|
33833
|
-
if (mode === '
|
|
34296
|
+
if (mode === 'nearest') {
|
|
34297
|
+
op = await _ops_registry_js__WEBPACK_IMPORTED_MODULE_2__.TensorOpRegistry.nearest_interpolate_4d;
|
|
34298
|
+
} else if (mode === 'bilinear') {
|
|
33834
34299
|
op = await _ops_registry_js__WEBPACK_IMPORTED_MODULE_2__.TensorOpRegistry.bilinear_interpolate_4d;
|
|
33835
34300
|
} else if (mode === 'bicubic') {
|
|
33836
34301
|
op = await _ops_registry_js__WEBPACK_IMPORTED_MODULE_2__.TensorOpRegistry.bicubic_interpolate_4d;
|
|
@@ -33871,13 +34336,13 @@ async function rfft(x, a) {
|
|
|
33871
34336
|
* Returns the k largest elements of the given input tensor.
|
|
33872
34337
|
* Inspired by https://pytorch.org/docs/stable/generated/torch.topk.html
|
|
33873
34338
|
* @param {Tensor} x the input tensor
|
|
33874
|
-
* @param {number} k the k in "top-k"
|
|
34339
|
+
* @param {number} [k] the k in "top-k"
|
|
33875
34340
|
* @returns {Promise<[Tensor, Tensor]>} the output tuple of (Tensor, LongTensor) of top-k elements and their indices.
|
|
33876
34341
|
*/
|
|
33877
34342
|
async function topk(x, k) {
|
|
33878
34343
|
const op = await _ops_registry_js__WEBPACK_IMPORTED_MODULE_2__.TensorOpRegistry.top_k;
|
|
33879
34344
|
|
|
33880
|
-
if (k
|
|
34345
|
+
if (k == null) {
|
|
33881
34346
|
k = x.dims.at(-1);
|
|
33882
34347
|
} else {
|
|
33883
34348
|
k = Math.min(k, x.dims.at(-1));
|
|
@@ -33906,10 +34371,10 @@ const arrayToIndexTensor = (array) => new Tensor('int64', array, [array.length])
|
|
|
33906
34371
|
async function slice(data, starts, ends, axes, steps) {
|
|
33907
34372
|
const op = await _ops_registry_js__WEBPACK_IMPORTED_MODULE_2__.TensorOpRegistry.slice;
|
|
33908
34373
|
return await op({
|
|
33909
|
-
x: data,
|
|
33910
|
-
s: arrayToIndexTensor(starts),
|
|
33911
|
-
e: arrayToIndexTensor(ends),
|
|
33912
|
-
a: arrayToIndexTensor(axes),
|
|
34374
|
+
x: data,
|
|
34375
|
+
s: arrayToIndexTensor(starts),
|
|
34376
|
+
e: arrayToIndexTensor(ends),
|
|
34377
|
+
a: arrayToIndexTensor(axes),
|
|
33913
34378
|
t: arrayToIndexTensor(steps ?? new Array(axes.length).fill(1)),
|
|
33914
34379
|
});
|
|
33915
34380
|
}
|
|
@@ -34144,35 +34609,19 @@ function stack(tensors, dim = 0) {
|
|
|
34144
34609
|
|
|
34145
34610
|
|
|
34146
34611
|
/**
|
|
34147
|
-
*
|
|
34148
|
-
* @param {Tensor} input the input
|
|
34149
|
-
* @param {number|null} dim the dimension to reduce.
|
|
34150
|
-
* @param {number} correction difference between the sample size and sample degrees of freedom. Defaults to Bessel's correction, correction=1.
|
|
34612
|
+
* @param {(previousValue: any, currentValue: any, currentIndex?: number, resultIndex?: number) => any} callbackfn
|
|
34613
|
+
* @param {Tensor} input the input tensor.
|
|
34614
|
+
* @param {number|null} dim the dimension to reduce.
|
|
34151
34615
|
* @param {boolean} keepdim whether the output tensor has dim retained or not.
|
|
34152
|
-
* @returns {
|
|
34616
|
+
* @returns {[DataType, any, number[]]} The reduced tensor data.
|
|
34153
34617
|
*/
|
|
34154
|
-
function
|
|
34155
|
-
const inputData =
|
|
34618
|
+
function reduce_helper(callbackfn, input, dim = null, keepdim = false, initialValue = null) {
|
|
34619
|
+
const inputData = input.data;
|
|
34156
34620
|
const inputDims = input.dims;
|
|
34157
34621
|
|
|
34158
|
-
if (dim === null) {
|
|
34159
|
-
// None to reduce over all dimensions.
|
|
34160
|
-
const sum = inputData.reduce((a, b) => a + b, 0);
|
|
34161
|
-
const mean = sum / inputData.length;
|
|
34162
|
-
const std = Math.sqrt(inputData.reduce((a, b) => a + (b - mean) ** 2, 0) / (inputData.length - correction));
|
|
34163
|
-
|
|
34164
|
-
const meanTensor = new Tensor(input.type, [mean], [/* scalar */]);
|
|
34165
|
-
const stdTensor = new Tensor(input.type, [std], [/* scalar */]);
|
|
34166
|
-
|
|
34167
|
-
return [stdTensor, meanTensor];
|
|
34168
|
-
}
|
|
34169
|
-
|
|
34170
34622
|
// Negative indexing
|
|
34171
34623
|
dim = safeIndex(dim, inputDims.length);
|
|
34172
34624
|
|
|
34173
|
-
const meanTensor = mean(input, dim, keepdim);
|
|
34174
|
-
const meanTensorData = meanTensor.data;
|
|
34175
|
-
|
|
34176
34625
|
// Calculate the shape of the resulting array after summation
|
|
34177
34626
|
const resultDims = inputDims.slice(); // Copy the original dimensions
|
|
34178
34627
|
resultDims[dim] = 1; // Remove the specified axis
|
|
@@ -34180,6 +34629,9 @@ function std_mean(input, dim = null, correction = 1, keepdim = false) {
|
|
|
34180
34629
|
// Create a new array to store the accumulated values
|
|
34181
34630
|
// @ts-ignore
|
|
34182
34631
|
const result = new inputData.constructor(inputData.length / inputDims[dim]);
|
|
34632
|
+
if (initialValue !== null) {
|
|
34633
|
+
result.fill(initialValue);
|
|
34634
|
+
}
|
|
34183
34635
|
|
|
34184
34636
|
// Iterate over the data array
|
|
34185
34637
|
for (let i = 0; i < inputData.length; ++i) {
|
|
@@ -34198,23 +34650,55 @@ function std_mean(input, dim = null, correction = 1, keepdim = false) {
|
|
|
34198
34650
|
}
|
|
34199
34651
|
|
|
34200
34652
|
// Accumulate the value at the current index
|
|
34201
|
-
result[resultIndex]
|
|
34653
|
+
result[resultIndex] = callbackfn(result[resultIndex], inputData[i], i, resultIndex);
|
|
34202
34654
|
}
|
|
34203
34655
|
|
|
34204
|
-
|
|
34205
|
-
|
|
34656
|
+
if (!keepdim) resultDims.splice(dim, 1);
|
|
34657
|
+
|
|
34658
|
+
return [input.type, result, resultDims];
|
|
34659
|
+
}
|
|
34660
|
+
|
|
34661
|
+
|
|
34662
|
+
/**
|
|
34663
|
+
* Calculates the standard deviation and mean over the dimensions specified by dim. dim can be a single dimension or `null` to reduce over all dimensions.
|
|
34664
|
+
* @param {Tensor} input the input tenso
|
|
34665
|
+
* @param {number|null} dim the dimension to reduce. If None, all dimensions are reduced.
|
|
34666
|
+
* @param {number} correction difference between the sample size and sample degrees of freedom. Defaults to Bessel's correction, correction=1.
|
|
34667
|
+
* @param {boolean} keepdim whether the output tensor has dim retained or not.
|
|
34668
|
+
* @returns {Tensor[]} A tuple of (std, mean) tensors.
|
|
34669
|
+
*/
|
|
34670
|
+
function std_mean(input, dim = null, correction = 1, keepdim = false) {
|
|
34671
|
+
const inputData = /** @type {Float32Array} */(input.data);
|
|
34672
|
+
const inputDims = input.dims;
|
|
34673
|
+
|
|
34674
|
+
if (dim === null) {
|
|
34675
|
+
// None to reduce over all dimensions.
|
|
34676
|
+
const sum = inputData.reduce((a, b) => a + b, 0);
|
|
34677
|
+
const mean = sum / inputData.length;
|
|
34678
|
+
const std = Math.sqrt(inputData.reduce((a, b) => a + (b - mean) ** 2, 0) / (inputData.length - correction));
|
|
34679
|
+
|
|
34680
|
+
const meanTensor = new Tensor(input.type, [mean], [/* scalar */]);
|
|
34681
|
+
const stdTensor = new Tensor(input.type, [std], [/* scalar */]);
|
|
34682
|
+
|
|
34683
|
+
return [stdTensor, meanTensor];
|
|
34206
34684
|
}
|
|
34685
|
+
dim = safeIndex(dim, inputDims.length);
|
|
34686
|
+
const meanTensor = mean(input, dim, keepdim);
|
|
34687
|
+
const meanTensorData = meanTensor.data;
|
|
34688
|
+
|
|
34689
|
+
// Compute squared sum
|
|
34690
|
+
const [type, result, resultDims] = reduce_helper((a, b, i, j) => a + (b - meanTensorData[j]) ** 2, input, dim, keepdim);
|
|
34207
34691
|
|
|
34208
|
-
|
|
34209
|
-
|
|
34692
|
+
// Square root of the squared sum
|
|
34693
|
+
for (let i = 0; i < result.length; ++i) {
|
|
34694
|
+
result[i] = Math.sqrt(result[i] / (inputDims[dim] - correction));
|
|
34210
34695
|
}
|
|
34211
34696
|
|
|
34212
|
-
const stdTensor = new Tensor(
|
|
34697
|
+
const stdTensor = new Tensor(type, result, resultDims);
|
|
34213
34698
|
|
|
34214
34699
|
return [stdTensor, meanTensor];
|
|
34215
34700
|
}
|
|
34216
34701
|
|
|
34217
|
-
|
|
34218
34702
|
/**
|
|
34219
34703
|
* Returns the mean value of each row of the input tensor in the given dimension dim.
|
|
34220
34704
|
* @param {Tensor} input the input tensor.
|
|
@@ -34223,58 +34707,27 @@ function std_mean(input, dim = null, correction = 1, keepdim = false) {
|
|
|
34223
34707
|
* @returns {Tensor} A new tensor with means taken along the specified dimension.
|
|
34224
34708
|
*/
|
|
34225
34709
|
function mean(input, dim = null, keepdim = false) {
|
|
34710
|
+
const inputDims = input.dims;
|
|
34226
34711
|
const inputData = /** @type {Float32Array} */(input.data);
|
|
34227
34712
|
|
|
34228
34713
|
if (dim === null) {
|
|
34229
34714
|
// None to reduce over all dimensions.
|
|
34230
|
-
// @ts-ignore
|
|
34231
34715
|
const val = inputData.reduce((a, b) => a + b, 0);
|
|
34232
34716
|
return new Tensor(input.type, [val / inputData.length], [/* scalar */]);
|
|
34233
34717
|
}
|
|
34234
|
-
const inputDims = input.dims;
|
|
34235
|
-
|
|
34236
|
-
// Negative indexing
|
|
34237
34718
|
dim = safeIndex(dim, inputDims.length);
|
|
34238
34719
|
|
|
34239
|
-
//
|
|
34240
|
-
const resultDims =
|
|
34241
|
-
resultDims[dim] = 1; // Remove the specified axis
|
|
34242
|
-
|
|
34243
|
-
// Create a new array to store the accumulated values
|
|
34244
|
-
// @ts-ignore
|
|
34245
|
-
const result = new inputData.constructor(inputData.length / inputDims[dim]);
|
|
34246
|
-
|
|
34247
|
-
// Iterate over the data array
|
|
34248
|
-
for (let i = 0; i < inputData.length; ++i) {
|
|
34249
|
-
|
|
34250
|
-
// Calculate the index in the resulting array
|
|
34251
|
-
let resultIndex = 0;
|
|
34252
|
-
|
|
34253
|
-
for (let j = inputDims.length - 1, num = i, resultMultiplier = 1; j >= 0; --j) {
|
|
34254
|
-
const size = inputDims[j];
|
|
34255
|
-
if (j !== dim) {
|
|
34256
|
-
const index = num % size;
|
|
34257
|
-
resultIndex += index * resultMultiplier;
|
|
34258
|
-
resultMultiplier *= resultDims[j];
|
|
34259
|
-
}
|
|
34260
|
-
num = Math.floor(num / size);
|
|
34261
|
-
}
|
|
34262
|
-
|
|
34263
|
-
// Accumulate the value at the current index
|
|
34264
|
-
result[resultIndex] += inputData[i];
|
|
34265
|
-
}
|
|
34720
|
+
// Compute sum
|
|
34721
|
+
const [type, result, resultDims] = reduce_helper((a, b) => a + b, input, dim, keepdim);
|
|
34266
34722
|
|
|
34723
|
+
// Divide by number of elements in the dimension
|
|
34267
34724
|
if (inputDims[dim] !== 1) {
|
|
34268
34725
|
for (let i = 0; i < result.length; ++i) {
|
|
34269
|
-
result[i]
|
|
34726
|
+
result[i] /= inputDims[dim];
|
|
34270
34727
|
}
|
|
34271
34728
|
}
|
|
34272
34729
|
|
|
34273
|
-
|
|
34274
|
-
resultDims.splice(dim, 1);
|
|
34275
|
-
}
|
|
34276
|
-
|
|
34277
|
-
return new Tensor(input.type, result, resultDims);
|
|
34730
|
+
return new Tensor(type, result, resultDims);
|
|
34278
34731
|
}
|
|
34279
34732
|
|
|
34280
34733
|
|
|
@@ -34773,6 +35226,10 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
34773
35226
|
/* harmony export */ GraniteModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.GraniteModel),
|
|
34774
35227
|
/* harmony export */ GranitePreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.GranitePreTrainedModel),
|
|
34775
35228
|
/* harmony export */ Grok1Tokenizer: () => (/* reexport safe */ _tokenizers_js__WEBPACK_IMPORTED_MODULE_3__.Grok1Tokenizer),
|
|
35229
|
+
/* harmony export */ GroundingDinoForObjectDetection: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.GroundingDinoForObjectDetection),
|
|
35230
|
+
/* harmony export */ GroundingDinoImageProcessor: () => (/* reexport safe */ _models_image_processors_js__WEBPACK_IMPORTED_MODULE_13__.GroundingDinoImageProcessor),
|
|
35231
|
+
/* harmony export */ GroundingDinoPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.GroundingDinoPreTrainedModel),
|
|
35232
|
+
/* harmony export */ GroundingDinoProcessor: () => (/* reexport safe */ _models_processors_js__WEBPACK_IMPORTED_MODULE_16__.GroundingDinoProcessor),
|
|
34776
35233
|
/* harmony export */ GroupViTModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.GroupViTModel),
|
|
34777
35234
|
/* harmony export */ GroupViTPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.GroupViTPreTrainedModel),
|
|
34778
35235
|
/* harmony export */ HerbertTokenizer: () => (/* reexport safe */ _tokenizers_js__WEBPACK_IMPORTED_MODULE_3__.HerbertTokenizer),
|
|
@@ -35002,6 +35459,7 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
35002
35459
|
/* harmony export */ RTDetrModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.RTDetrModel),
|
|
35003
35460
|
/* harmony export */ RTDetrObjectDetectionOutput: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.RTDetrObjectDetectionOutput),
|
|
35004
35461
|
/* harmony export */ RTDetrPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.RTDetrPreTrainedModel),
|
|
35462
|
+
/* harmony export */ RawAudio: () => (/* reexport safe */ _utils_audio_js__WEBPACK_IMPORTED_MODULE_5__.RawAudio),
|
|
35005
35463
|
/* harmony export */ RawImage: () => (/* reexport safe */ _utils_image_js__WEBPACK_IMPORTED_MODULE_6__.RawImage),
|
|
35006
35464
|
/* harmony export */ RepetitionPenaltyLogitsProcessor: () => (/* reexport safe */ _generation_logits_process_js__WEBPACK_IMPORTED_MODULE_20__.RepetitionPenaltyLogitsProcessor),
|
|
35007
35465
|
/* harmony export */ ResNetForImageClassification: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.ResNetForImageClassification),
|
|
@@ -35067,6 +35525,8 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
35067
35525
|
/* harmony export */ Starcoder2PreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Starcoder2PreTrainedModel),
|
|
35068
35526
|
/* harmony export */ StoppingCriteria: () => (/* reexport safe */ _generation_stopping_criteria_js__WEBPACK_IMPORTED_MODULE_19__.StoppingCriteria),
|
|
35069
35527
|
/* harmony export */ StoppingCriteriaList: () => (/* reexport safe */ _generation_stopping_criteria_js__WEBPACK_IMPORTED_MODULE_19__.StoppingCriteriaList),
|
|
35528
|
+
/* harmony export */ StyleTextToSpeech2Model: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.StyleTextToSpeech2Model),
|
|
35529
|
+
/* harmony export */ StyleTextToSpeech2PreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.StyleTextToSpeech2PreTrainedModel),
|
|
35070
35530
|
/* harmony export */ SummarizationPipeline: () => (/* reexport safe */ _pipelines_js__WEBPACK_IMPORTED_MODULE_1__.SummarizationPipeline),
|
|
35071
35531
|
/* harmony export */ SuppressTokensAtBeginLogitsProcessor: () => (/* reexport safe */ _generation_logits_process_js__WEBPACK_IMPORTED_MODULE_20__.SuppressTokensAtBeginLogitsProcessor),
|
|
35072
35532
|
/* harmony export */ Swin2SRForImageSuperResolution: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Swin2SRForImageSuperResolution),
|
|
@@ -35553,6 +36013,10 @@ var __webpack_exports__GraniteForCausalLM = __webpack_exports__.GraniteForCausal
|
|
|
35553
36013
|
var __webpack_exports__GraniteModel = __webpack_exports__.GraniteModel;
|
|
35554
36014
|
var __webpack_exports__GranitePreTrainedModel = __webpack_exports__.GranitePreTrainedModel;
|
|
35555
36015
|
var __webpack_exports__Grok1Tokenizer = __webpack_exports__.Grok1Tokenizer;
|
|
36016
|
+
var __webpack_exports__GroundingDinoForObjectDetection = __webpack_exports__.GroundingDinoForObjectDetection;
|
|
36017
|
+
var __webpack_exports__GroundingDinoImageProcessor = __webpack_exports__.GroundingDinoImageProcessor;
|
|
36018
|
+
var __webpack_exports__GroundingDinoPreTrainedModel = __webpack_exports__.GroundingDinoPreTrainedModel;
|
|
36019
|
+
var __webpack_exports__GroundingDinoProcessor = __webpack_exports__.GroundingDinoProcessor;
|
|
35556
36020
|
var __webpack_exports__GroupViTModel = __webpack_exports__.GroupViTModel;
|
|
35557
36021
|
var __webpack_exports__GroupViTPreTrainedModel = __webpack_exports__.GroupViTPreTrainedModel;
|
|
35558
36022
|
var __webpack_exports__HerbertTokenizer = __webpack_exports__.HerbertTokenizer;
|
|
@@ -35782,6 +36246,7 @@ var __webpack_exports__RTDetrImageProcessor = __webpack_exports__.RTDetrImagePro
|
|
|
35782
36246
|
var __webpack_exports__RTDetrModel = __webpack_exports__.RTDetrModel;
|
|
35783
36247
|
var __webpack_exports__RTDetrObjectDetectionOutput = __webpack_exports__.RTDetrObjectDetectionOutput;
|
|
35784
36248
|
var __webpack_exports__RTDetrPreTrainedModel = __webpack_exports__.RTDetrPreTrainedModel;
|
|
36249
|
+
var __webpack_exports__RawAudio = __webpack_exports__.RawAudio;
|
|
35785
36250
|
var __webpack_exports__RawImage = __webpack_exports__.RawImage;
|
|
35786
36251
|
var __webpack_exports__RepetitionPenaltyLogitsProcessor = __webpack_exports__.RepetitionPenaltyLogitsProcessor;
|
|
35787
36252
|
var __webpack_exports__ResNetForImageClassification = __webpack_exports__.ResNetForImageClassification;
|
|
@@ -35847,6 +36312,8 @@ var __webpack_exports__Starcoder2Model = __webpack_exports__.Starcoder2Model;
|
|
|
35847
36312
|
var __webpack_exports__Starcoder2PreTrainedModel = __webpack_exports__.Starcoder2PreTrainedModel;
|
|
35848
36313
|
var __webpack_exports__StoppingCriteria = __webpack_exports__.StoppingCriteria;
|
|
35849
36314
|
var __webpack_exports__StoppingCriteriaList = __webpack_exports__.StoppingCriteriaList;
|
|
36315
|
+
var __webpack_exports__StyleTextToSpeech2Model = __webpack_exports__.StyleTextToSpeech2Model;
|
|
36316
|
+
var __webpack_exports__StyleTextToSpeech2PreTrainedModel = __webpack_exports__.StyleTextToSpeech2PreTrainedModel;
|
|
35850
36317
|
var __webpack_exports__SummarizationPipeline = __webpack_exports__.SummarizationPipeline;
|
|
35851
36318
|
var __webpack_exports__SuppressTokensAtBeginLogitsProcessor = __webpack_exports__.SuppressTokensAtBeginLogitsProcessor;
|
|
35852
36319
|
var __webpack_exports__Swin2SRForImageSuperResolution = __webpack_exports__.Swin2SRForImageSuperResolution;
|
|
@@ -36010,6 +36477,6 @@ var __webpack_exports__topk = __webpack_exports__.topk;
|
|
|
36010
36477
|
var __webpack_exports__window_function = __webpack_exports__.window_function;
|
|
36011
36478
|
var __webpack_exports__zeros = __webpack_exports__.zeros;
|
|
36012
36479
|
var __webpack_exports__zeros_like = __webpack_exports__.zeros_like;
|
|
36013
|
-
export { __webpack_exports__ASTFeatureExtractor as ASTFeatureExtractor, __webpack_exports__ASTForAudioClassification as ASTForAudioClassification, __webpack_exports__ASTModel as ASTModel, __webpack_exports__ASTPreTrainedModel as ASTPreTrainedModel, __webpack_exports__AlbertForMaskedLM as AlbertForMaskedLM, __webpack_exports__AlbertForQuestionAnswering as AlbertForQuestionAnswering, __webpack_exports__AlbertForSequenceClassification as AlbertForSequenceClassification, __webpack_exports__AlbertModel as AlbertModel, __webpack_exports__AlbertPreTrainedModel as AlbertPreTrainedModel, __webpack_exports__AlbertTokenizer as AlbertTokenizer, __webpack_exports__AudioClassificationPipeline as AudioClassificationPipeline, __webpack_exports__AutoConfig as AutoConfig, __webpack_exports__AutoFeatureExtractor as AutoFeatureExtractor, __webpack_exports__AutoImageProcessor as AutoImageProcessor, __webpack_exports__AutoModel as AutoModel, __webpack_exports__AutoModelForAudioClassification as AutoModelForAudioClassification, __webpack_exports__AutoModelForAudioFrameClassification as AutoModelForAudioFrameClassification, __webpack_exports__AutoModelForCTC as AutoModelForCTC, __webpack_exports__AutoModelForCausalLM as AutoModelForCausalLM, __webpack_exports__AutoModelForDepthEstimation as AutoModelForDepthEstimation, __webpack_exports__AutoModelForDocumentQuestionAnswering as AutoModelForDocumentQuestionAnswering, __webpack_exports__AutoModelForImageClassification as AutoModelForImageClassification, __webpack_exports__AutoModelForImageFeatureExtraction as AutoModelForImageFeatureExtraction, __webpack_exports__AutoModelForImageMatting as AutoModelForImageMatting, __webpack_exports__AutoModelForImageSegmentation as AutoModelForImageSegmentation, __webpack_exports__AutoModelForImageToImage as AutoModelForImageToImage, __webpack_exports__AutoModelForMaskGeneration as AutoModelForMaskGeneration, __webpack_exports__AutoModelForMaskedLM as AutoModelForMaskedLM, __webpack_exports__AutoModelForNormalEstimation as AutoModelForNormalEstimation, __webpack_exports__AutoModelForObjectDetection as AutoModelForObjectDetection, __webpack_exports__AutoModelForPoseEstimation as AutoModelForPoseEstimation, __webpack_exports__AutoModelForQuestionAnswering as AutoModelForQuestionAnswering, __webpack_exports__AutoModelForSemanticSegmentation as AutoModelForSemanticSegmentation, __webpack_exports__AutoModelForSeq2SeqLM as AutoModelForSeq2SeqLM, __webpack_exports__AutoModelForSequenceClassification as AutoModelForSequenceClassification, __webpack_exports__AutoModelForSpeechSeq2Seq as AutoModelForSpeechSeq2Seq, __webpack_exports__AutoModelForTextToSpectrogram as AutoModelForTextToSpectrogram, __webpack_exports__AutoModelForTextToWaveform as AutoModelForTextToWaveform, __webpack_exports__AutoModelForTokenClassification as AutoModelForTokenClassification, __webpack_exports__AutoModelForUniversalSegmentation as AutoModelForUniversalSegmentation, __webpack_exports__AutoModelForVision2Seq as AutoModelForVision2Seq, __webpack_exports__AutoModelForXVector as AutoModelForXVector, __webpack_exports__AutoModelForZeroShotObjectDetection as AutoModelForZeroShotObjectDetection, __webpack_exports__AutoProcessor as AutoProcessor, __webpack_exports__AutoTokenizer as AutoTokenizer, __webpack_exports__AutomaticSpeechRecognitionPipeline as AutomaticSpeechRecognitionPipeline, __webpack_exports__BartForConditionalGeneration as BartForConditionalGeneration, __webpack_exports__BartForSequenceClassification as BartForSequenceClassification, __webpack_exports__BartModel as BartModel, __webpack_exports__BartPretrainedModel as BartPretrainedModel, __webpack_exports__BartTokenizer as BartTokenizer, __webpack_exports__BaseModelOutput as BaseModelOutput, __webpack_exports__BaseStreamer as BaseStreamer, __webpack_exports__BeitFeatureExtractor as BeitFeatureExtractor, __webpack_exports__BeitForImageClassification as BeitForImageClassification, __webpack_exports__BeitModel as BeitModel, __webpack_exports__BeitPreTrainedModel as BeitPreTrainedModel, __webpack_exports__BertForMaskedLM as BertForMaskedLM, __webpack_exports__BertForQuestionAnswering as BertForQuestionAnswering, __webpack_exports__BertForSequenceClassification as BertForSequenceClassification, __webpack_exports__BertForTokenClassification as BertForTokenClassification, __webpack_exports__BertModel as BertModel, __webpack_exports__BertPreTrainedModel as BertPreTrainedModel, __webpack_exports__BertTokenizer as BertTokenizer, __webpack_exports__BitImageProcessor as BitImageProcessor, __webpack_exports__BlenderbotForConditionalGeneration as BlenderbotForConditionalGeneration, __webpack_exports__BlenderbotModel as BlenderbotModel, __webpack_exports__BlenderbotPreTrainedModel as BlenderbotPreTrainedModel, __webpack_exports__BlenderbotSmallForConditionalGeneration as BlenderbotSmallForConditionalGeneration, __webpack_exports__BlenderbotSmallModel as BlenderbotSmallModel, __webpack_exports__BlenderbotSmallPreTrainedModel as BlenderbotSmallPreTrainedModel, __webpack_exports__BlenderbotSmallTokenizer as BlenderbotSmallTokenizer, __webpack_exports__BlenderbotTokenizer as BlenderbotTokenizer, __webpack_exports__BloomForCausalLM as BloomForCausalLM, __webpack_exports__BloomModel as BloomModel, __webpack_exports__BloomPreTrainedModel as BloomPreTrainedModel, __webpack_exports__BloomTokenizer as BloomTokenizer, __webpack_exports__CLIPFeatureExtractor as CLIPFeatureExtractor, __webpack_exports__CLIPImageProcessor as CLIPImageProcessor, __webpack_exports__CLIPModel as CLIPModel, __webpack_exports__CLIPPreTrainedModel as CLIPPreTrainedModel, __webpack_exports__CLIPSegForImageSegmentation as CLIPSegForImageSegmentation, __webpack_exports__CLIPSegModel as CLIPSegModel, __webpack_exports__CLIPSegPreTrainedModel as CLIPSegPreTrainedModel, __webpack_exports__CLIPTextModel as CLIPTextModel, __webpack_exports__CLIPTextModelWithProjection as CLIPTextModelWithProjection, __webpack_exports__CLIPTokenizer as CLIPTokenizer, __webpack_exports__CLIPVisionModel as CLIPVisionModel, __webpack_exports__CLIPVisionModelWithProjection as CLIPVisionModelWithProjection, __webpack_exports__CamembertForMaskedLM as CamembertForMaskedLM, __webpack_exports__CamembertForQuestionAnswering as CamembertForQuestionAnswering, __webpack_exports__CamembertForSequenceClassification as CamembertForSequenceClassification, __webpack_exports__CamembertForTokenClassification as CamembertForTokenClassification, __webpack_exports__CamembertModel as CamembertModel, __webpack_exports__CamembertPreTrainedModel as CamembertPreTrainedModel, __webpack_exports__CamembertTokenizer as CamembertTokenizer, __webpack_exports__CausalLMOutput as CausalLMOutput, __webpack_exports__CausalLMOutputWithPast as CausalLMOutputWithPast, __webpack_exports__ChineseCLIPFeatureExtractor as ChineseCLIPFeatureExtractor, __webpack_exports__ChineseCLIPModel as ChineseCLIPModel, __webpack_exports__ChineseCLIPPreTrainedModel as ChineseCLIPPreTrainedModel, __webpack_exports__ClapAudioModelWithProjection as ClapAudioModelWithProjection, __webpack_exports__ClapFeatureExtractor as ClapFeatureExtractor, __webpack_exports__ClapModel as ClapModel, __webpack_exports__ClapPreTrainedModel as ClapPreTrainedModel, __webpack_exports__ClapTextModelWithProjection as ClapTextModelWithProjection, __webpack_exports__ClassifierFreeGuidanceLogitsProcessor as ClassifierFreeGuidanceLogitsProcessor, __webpack_exports__CodeGenForCausalLM as CodeGenForCausalLM, __webpack_exports__CodeGenModel as CodeGenModel, __webpack_exports__CodeGenPreTrainedModel as CodeGenPreTrainedModel, __webpack_exports__CodeGenTokenizer as CodeGenTokenizer, __webpack_exports__CodeLlamaTokenizer as CodeLlamaTokenizer, __webpack_exports__CohereForCausalLM as CohereForCausalLM, __webpack_exports__CohereModel as CohereModel, __webpack_exports__CoherePreTrainedModel as CoherePreTrainedModel, __webpack_exports__CohereTokenizer as CohereTokenizer, __webpack_exports__ConvBertForMaskedLM as ConvBertForMaskedLM, __webpack_exports__ConvBertForQuestionAnswering as ConvBertForQuestionAnswering, __webpack_exports__ConvBertForSequenceClassification as ConvBertForSequenceClassification, __webpack_exports__ConvBertForTokenClassification as ConvBertForTokenClassification, __webpack_exports__ConvBertModel as ConvBertModel, __webpack_exports__ConvBertPreTrainedModel as ConvBertPreTrainedModel, __webpack_exports__ConvBertTokenizer as ConvBertTokenizer, __webpack_exports__ConvNextFeatureExtractor as ConvNextFeatureExtractor, __webpack_exports__ConvNextForImageClassification as ConvNextForImageClassification, __webpack_exports__ConvNextImageProcessor as ConvNextImageProcessor, __webpack_exports__ConvNextModel as ConvNextModel, __webpack_exports__ConvNextPreTrainedModel as ConvNextPreTrainedModel, __webpack_exports__ConvNextV2ForImageClassification as ConvNextV2ForImageClassification, __webpack_exports__ConvNextV2Model as ConvNextV2Model, __webpack_exports__ConvNextV2PreTrainedModel as ConvNextV2PreTrainedModel, __webpack_exports__DPTFeatureExtractor as DPTFeatureExtractor, __webpack_exports__DPTForDepthEstimation as DPTForDepthEstimation, __webpack_exports__DPTImageProcessor as DPTImageProcessor, __webpack_exports__DPTModel as DPTModel, __webpack_exports__DPTPreTrainedModel as DPTPreTrainedModel, __webpack_exports__DebertaForMaskedLM as DebertaForMaskedLM, __webpack_exports__DebertaForQuestionAnswering as DebertaForQuestionAnswering, __webpack_exports__DebertaForSequenceClassification as DebertaForSequenceClassification, __webpack_exports__DebertaForTokenClassification as DebertaForTokenClassification, __webpack_exports__DebertaModel as DebertaModel, __webpack_exports__DebertaPreTrainedModel as DebertaPreTrainedModel, __webpack_exports__DebertaTokenizer as DebertaTokenizer, __webpack_exports__DebertaV2ForMaskedLM as DebertaV2ForMaskedLM, __webpack_exports__DebertaV2ForQuestionAnswering as DebertaV2ForQuestionAnswering, __webpack_exports__DebertaV2ForSequenceClassification as DebertaV2ForSequenceClassification, __webpack_exports__DebertaV2ForTokenClassification as DebertaV2ForTokenClassification, __webpack_exports__DebertaV2Model as DebertaV2Model, __webpack_exports__DebertaV2PreTrainedModel as DebertaV2PreTrainedModel, __webpack_exports__DebertaV2Tokenizer as DebertaV2Tokenizer, __webpack_exports__DecisionTransformerModel as DecisionTransformerModel, __webpack_exports__DecisionTransformerPreTrainedModel as DecisionTransformerPreTrainedModel, __webpack_exports__DeiTFeatureExtractor as DeiTFeatureExtractor, __webpack_exports__DeiTForImageClassification as DeiTForImageClassification, __webpack_exports__DeiTImageProcessor as DeiTImageProcessor, __webpack_exports__DeiTModel as DeiTModel, __webpack_exports__DeiTPreTrainedModel as DeiTPreTrainedModel, __webpack_exports__DepthAnythingForDepthEstimation as DepthAnythingForDepthEstimation, __webpack_exports__DepthAnythingPreTrainedModel as DepthAnythingPreTrainedModel, __webpack_exports__DepthEstimationPipeline as DepthEstimationPipeline, __webpack_exports__DepthProForDepthEstimation as DepthProForDepthEstimation, __webpack_exports__DepthProPreTrainedModel as DepthProPreTrainedModel, __webpack_exports__DetrFeatureExtractor as DetrFeatureExtractor, __webpack_exports__DetrForObjectDetection as DetrForObjectDetection, __webpack_exports__DetrForSegmentation as DetrForSegmentation, __webpack_exports__DetrImageProcessor as DetrImageProcessor, __webpack_exports__DetrModel as DetrModel, __webpack_exports__DetrObjectDetectionOutput as DetrObjectDetectionOutput, __webpack_exports__DetrPreTrainedModel as DetrPreTrainedModel, __webpack_exports__DetrSegmentationOutput as DetrSegmentationOutput, __webpack_exports__Dinov2ForImageClassification as Dinov2ForImageClassification, __webpack_exports__Dinov2Model as Dinov2Model, __webpack_exports__Dinov2PreTrainedModel as Dinov2PreTrainedModel, __webpack_exports__Dinov2WithRegistersForImageClassification as Dinov2WithRegistersForImageClassification, __webpack_exports__Dinov2WithRegistersModel as Dinov2WithRegistersModel, __webpack_exports__Dinov2WithRegistersPreTrainedModel as Dinov2WithRegistersPreTrainedModel, __webpack_exports__DistilBertForMaskedLM as DistilBertForMaskedLM, __webpack_exports__DistilBertForQuestionAnswering as DistilBertForQuestionAnswering, __webpack_exports__DistilBertForSequenceClassification as DistilBertForSequenceClassification, __webpack_exports__DistilBertForTokenClassification as DistilBertForTokenClassification, __webpack_exports__DistilBertModel as DistilBertModel, __webpack_exports__DistilBertPreTrainedModel as DistilBertPreTrainedModel, __webpack_exports__DistilBertTokenizer as DistilBertTokenizer, __webpack_exports__DocumentQuestionAnsweringPipeline as DocumentQuestionAnsweringPipeline, __webpack_exports__DonutFeatureExtractor as DonutFeatureExtractor, __webpack_exports__DonutImageProcessor as DonutImageProcessor, __webpack_exports__DonutSwinModel as DonutSwinModel, __webpack_exports__DonutSwinPreTrainedModel as DonutSwinPreTrainedModel, __webpack_exports__EfficientNetForImageClassification as EfficientNetForImageClassification, __webpack_exports__EfficientNetImageProcessor as EfficientNetImageProcessor, __webpack_exports__EfficientNetModel as EfficientNetModel, __webpack_exports__EfficientNetPreTrainedModel as EfficientNetPreTrainedModel, __webpack_exports__ElectraForMaskedLM as ElectraForMaskedLM, __webpack_exports__ElectraForQuestionAnswering as ElectraForQuestionAnswering, __webpack_exports__ElectraForSequenceClassification as ElectraForSequenceClassification, __webpack_exports__ElectraForTokenClassification as ElectraForTokenClassification, __webpack_exports__ElectraModel as ElectraModel, __webpack_exports__ElectraPreTrainedModel as ElectraPreTrainedModel, __webpack_exports__ElectraTokenizer as ElectraTokenizer, __webpack_exports__EosTokenCriteria as EosTokenCriteria, __webpack_exports__EsmForMaskedLM as EsmForMaskedLM, __webpack_exports__EsmForSequenceClassification as EsmForSequenceClassification, __webpack_exports__EsmForTokenClassification as EsmForTokenClassification, __webpack_exports__EsmModel as EsmModel, __webpack_exports__EsmPreTrainedModel as EsmPreTrainedModel, __webpack_exports__EsmTokenizer as EsmTokenizer, __webpack_exports__ExaoneForCausalLM as ExaoneForCausalLM, __webpack_exports__ExaoneModel as ExaoneModel, __webpack_exports__ExaonePreTrainedModel as ExaonePreTrainedModel, __webpack_exports__FFT as FFT, __webpack_exports__FalconForCausalLM as FalconForCausalLM, __webpack_exports__FalconModel as FalconModel, __webpack_exports__FalconPreTrainedModel as FalconPreTrainedModel, __webpack_exports__FalconTokenizer as FalconTokenizer, __webpack_exports__FastViTForImageClassification as FastViTForImageClassification, __webpack_exports__FastViTModel as FastViTModel, __webpack_exports__FastViTPreTrainedModel as FastViTPreTrainedModel, __webpack_exports__FeatureExtractionPipeline as FeatureExtractionPipeline, __webpack_exports__FeatureExtractor as FeatureExtractor, __webpack_exports__FillMaskPipeline as FillMaskPipeline, __webpack_exports__Florence2ForConditionalGeneration as Florence2ForConditionalGeneration, __webpack_exports__Florence2PreTrainedModel as Florence2PreTrainedModel, __webpack_exports__Florence2Processor as Florence2Processor, __webpack_exports__ForcedBOSTokenLogitsProcessor as ForcedBOSTokenLogitsProcessor, __webpack_exports__ForcedEOSTokenLogitsProcessor as ForcedEOSTokenLogitsProcessor, __webpack_exports__GLPNFeatureExtractor as GLPNFeatureExtractor, __webpack_exports__GLPNForDepthEstimation as GLPNForDepthEstimation, __webpack_exports__GLPNModel as GLPNModel, __webpack_exports__GLPNPreTrainedModel as GLPNPreTrainedModel, __webpack_exports__GPT2LMHeadModel as GPT2LMHeadModel, __webpack_exports__GPT2Model as GPT2Model, __webpack_exports__GPT2PreTrainedModel as GPT2PreTrainedModel, __webpack_exports__GPT2Tokenizer as GPT2Tokenizer, __webpack_exports__GPTBigCodeForCausalLM as GPTBigCodeForCausalLM, __webpack_exports__GPTBigCodeModel as GPTBigCodeModel, __webpack_exports__GPTBigCodePreTrainedModel as GPTBigCodePreTrainedModel, __webpack_exports__GPTJForCausalLM as GPTJForCausalLM, __webpack_exports__GPTJModel as GPTJModel, __webpack_exports__GPTJPreTrainedModel as GPTJPreTrainedModel, __webpack_exports__GPTNeoForCausalLM as GPTNeoForCausalLM, __webpack_exports__GPTNeoModel as GPTNeoModel, __webpack_exports__GPTNeoPreTrainedModel as GPTNeoPreTrainedModel, __webpack_exports__GPTNeoXForCausalLM as GPTNeoXForCausalLM, __webpack_exports__GPTNeoXModel as GPTNeoXModel, __webpack_exports__GPTNeoXPreTrainedModel as GPTNeoXPreTrainedModel, __webpack_exports__GPTNeoXTokenizer as GPTNeoXTokenizer, __webpack_exports__Gemma2ForCausalLM as Gemma2ForCausalLM, __webpack_exports__Gemma2Model as Gemma2Model, __webpack_exports__Gemma2PreTrainedModel as Gemma2PreTrainedModel, __webpack_exports__GemmaForCausalLM as GemmaForCausalLM, __webpack_exports__GemmaModel as GemmaModel, __webpack_exports__GemmaPreTrainedModel as GemmaPreTrainedModel, __webpack_exports__GemmaTokenizer as GemmaTokenizer, __webpack_exports__GraniteForCausalLM as GraniteForCausalLM, __webpack_exports__GraniteModel as GraniteModel, __webpack_exports__GranitePreTrainedModel as GranitePreTrainedModel, __webpack_exports__Grok1Tokenizer as Grok1Tokenizer, __webpack_exports__GroupViTModel as GroupViTModel, __webpack_exports__GroupViTPreTrainedModel as GroupViTPreTrainedModel, __webpack_exports__HerbertTokenizer as HerbertTokenizer, __webpack_exports__HieraForImageClassification as HieraForImageClassification, __webpack_exports__HieraModel as HieraModel, __webpack_exports__HieraPreTrainedModel as HieraPreTrainedModel, __webpack_exports__HubertForCTC as HubertForCTC, __webpack_exports__HubertForSequenceClassification as HubertForSequenceClassification, __webpack_exports__HubertModel as HubertModel, __webpack_exports__HubertPreTrainedModel as HubertPreTrainedModel, __webpack_exports__IJepaForImageClassification as IJepaForImageClassification, __webpack_exports__IJepaModel as IJepaModel, __webpack_exports__IJepaPreTrainedModel as IJepaPreTrainedModel, __webpack_exports__Idefics3ForConditionalGeneration as Idefics3ForConditionalGeneration, __webpack_exports__Idefics3ImageProcessor as Idefics3ImageProcessor, __webpack_exports__Idefics3PreTrainedModel as Idefics3PreTrainedModel, __webpack_exports__Idefics3Processor as Idefics3Processor, __webpack_exports__ImageClassificationPipeline as ImageClassificationPipeline, __webpack_exports__ImageFeatureExtractionPipeline as ImageFeatureExtractionPipeline, __webpack_exports__ImageFeatureExtractor as ImageFeatureExtractor, __webpack_exports__ImageMattingOutput as ImageMattingOutput, __webpack_exports__ImageProcessor as ImageProcessor, __webpack_exports__ImageSegmentationPipeline as ImageSegmentationPipeline, __webpack_exports__ImageToImagePipeline as ImageToImagePipeline, __webpack_exports__ImageToTextPipeline as ImageToTextPipeline, __webpack_exports__InterruptableStoppingCriteria as InterruptableStoppingCriteria, __webpack_exports__JAISLMHeadModel as JAISLMHeadModel, __webpack_exports__JAISModel as JAISModel, __webpack_exports__JAISPreTrainedModel as JAISPreTrainedModel, __webpack_exports__JinaCLIPImageProcessor as JinaCLIPImageProcessor, __webpack_exports__JinaCLIPModel as JinaCLIPModel, __webpack_exports__JinaCLIPPreTrainedModel as JinaCLIPPreTrainedModel, __webpack_exports__JinaCLIPProcessor as JinaCLIPProcessor, __webpack_exports__JinaCLIPTextModel as JinaCLIPTextModel, __webpack_exports__JinaCLIPVisionModel as JinaCLIPVisionModel, __webpack_exports__LlamaForCausalLM as LlamaForCausalLM, __webpack_exports__LlamaModel as LlamaModel, __webpack_exports__LlamaPreTrainedModel as LlamaPreTrainedModel, __webpack_exports__LlamaTokenizer as LlamaTokenizer, __webpack_exports__LlavaForConditionalGeneration as LlavaForConditionalGeneration, __webpack_exports__LlavaOnevisionForConditionalGeneration as LlavaOnevisionForConditionalGeneration, __webpack_exports__LlavaOnevisionImageProcessor as LlavaOnevisionImageProcessor, __webpack_exports__LlavaPreTrainedModel as LlavaPreTrainedModel, __webpack_exports__LogitsProcessor as LogitsProcessor, __webpack_exports__LogitsProcessorList as LogitsProcessorList, __webpack_exports__LogitsWarper as LogitsWarper, __webpack_exports__LongT5ForConditionalGeneration as LongT5ForConditionalGeneration, __webpack_exports__LongT5Model as LongT5Model, __webpack_exports__LongT5PreTrainedModel as LongT5PreTrainedModel, __webpack_exports__M2M100ForConditionalGeneration as M2M100ForConditionalGeneration, __webpack_exports__M2M100Model as M2M100Model, __webpack_exports__M2M100PreTrainedModel as M2M100PreTrainedModel, __webpack_exports__M2M100Tokenizer as M2M100Tokenizer, __webpack_exports__MBart50Tokenizer as MBart50Tokenizer, __webpack_exports__MBartForCausalLM as MBartForCausalLM, __webpack_exports__MBartForConditionalGeneration as MBartForConditionalGeneration, __webpack_exports__MBartForSequenceClassification as MBartForSequenceClassification, __webpack_exports__MBartModel as MBartModel, __webpack_exports__MBartPreTrainedModel as MBartPreTrainedModel, __webpack_exports__MBartTokenizer as MBartTokenizer, __webpack_exports__MPNetForMaskedLM as MPNetForMaskedLM, __webpack_exports__MPNetForQuestionAnswering as MPNetForQuestionAnswering, __webpack_exports__MPNetForSequenceClassification as MPNetForSequenceClassification, __webpack_exports__MPNetForTokenClassification as MPNetForTokenClassification, __webpack_exports__MPNetModel as MPNetModel, __webpack_exports__MPNetPreTrainedModel as MPNetPreTrainedModel, __webpack_exports__MPNetTokenizer as MPNetTokenizer, __webpack_exports__MT5ForConditionalGeneration as MT5ForConditionalGeneration, __webpack_exports__MT5Model as MT5Model, __webpack_exports__MT5PreTrainedModel as MT5PreTrainedModel, __webpack_exports__MarianMTModel as MarianMTModel, __webpack_exports__MarianModel as MarianModel, __webpack_exports__MarianPreTrainedModel as MarianPreTrainedModel, __webpack_exports__MarianTokenizer as MarianTokenizer, __webpack_exports__Mask2FormerImageProcessor as Mask2FormerImageProcessor, __webpack_exports__MaskFormerFeatureExtractor as MaskFormerFeatureExtractor, __webpack_exports__MaskFormerForInstanceSegmentation as MaskFormerForInstanceSegmentation, __webpack_exports__MaskFormerImageProcessor as MaskFormerImageProcessor, __webpack_exports__MaskFormerModel as MaskFormerModel, __webpack_exports__MaskFormerPreTrainedModel as MaskFormerPreTrainedModel, __webpack_exports__MaskedLMOutput as MaskedLMOutput, __webpack_exports__MaxLengthCriteria as MaxLengthCriteria, __webpack_exports__MgpstrForSceneTextRecognition as MgpstrForSceneTextRecognition, __webpack_exports__MgpstrModelOutput as MgpstrModelOutput, __webpack_exports__MgpstrPreTrainedModel as MgpstrPreTrainedModel, __webpack_exports__MgpstrProcessor as MgpstrProcessor, __webpack_exports__MgpstrTokenizer as MgpstrTokenizer, __webpack_exports__MinLengthLogitsProcessor as MinLengthLogitsProcessor, __webpack_exports__MinNewTokensLengthLogitsProcessor as MinNewTokensLengthLogitsProcessor, __webpack_exports__MistralForCausalLM as MistralForCausalLM, __webpack_exports__MistralModel as MistralModel, __webpack_exports__MistralPreTrainedModel as MistralPreTrainedModel, __webpack_exports__MobileBertForMaskedLM as MobileBertForMaskedLM, __webpack_exports__MobileBertForQuestionAnswering as MobileBertForQuestionAnswering, __webpack_exports__MobileBertForSequenceClassification as MobileBertForSequenceClassification, __webpack_exports__MobileBertModel as MobileBertModel, __webpack_exports__MobileBertPreTrainedModel as MobileBertPreTrainedModel, __webpack_exports__MobileBertTokenizer as MobileBertTokenizer, __webpack_exports__MobileLLMForCausalLM as MobileLLMForCausalLM, __webpack_exports__MobileLLMModel as MobileLLMModel, __webpack_exports__MobileLLMPreTrainedModel as MobileLLMPreTrainedModel, __webpack_exports__MobileNetV1FeatureExtractor as MobileNetV1FeatureExtractor, __webpack_exports__MobileNetV1ForImageClassification as MobileNetV1ForImageClassification, __webpack_exports__MobileNetV1ImageProcessor as MobileNetV1ImageProcessor, __webpack_exports__MobileNetV1Model as MobileNetV1Model, __webpack_exports__MobileNetV1PreTrainedModel as MobileNetV1PreTrainedModel, __webpack_exports__MobileNetV2FeatureExtractor as MobileNetV2FeatureExtractor, __webpack_exports__MobileNetV2ForImageClassification as MobileNetV2ForImageClassification, __webpack_exports__MobileNetV2ImageProcessor as MobileNetV2ImageProcessor, __webpack_exports__MobileNetV2Model as MobileNetV2Model, __webpack_exports__MobileNetV2PreTrainedModel as MobileNetV2PreTrainedModel, __webpack_exports__MobileNetV3FeatureExtractor as MobileNetV3FeatureExtractor, __webpack_exports__MobileNetV3ForImageClassification as MobileNetV3ForImageClassification, __webpack_exports__MobileNetV3ImageProcessor as MobileNetV3ImageProcessor, __webpack_exports__MobileNetV3Model as MobileNetV3Model, __webpack_exports__MobileNetV3PreTrainedModel as MobileNetV3PreTrainedModel, __webpack_exports__MobileNetV4FeatureExtractor as MobileNetV4FeatureExtractor, __webpack_exports__MobileNetV4ForImageClassification as MobileNetV4ForImageClassification, __webpack_exports__MobileNetV4ImageProcessor as MobileNetV4ImageProcessor, __webpack_exports__MobileNetV4Model as MobileNetV4Model, __webpack_exports__MobileNetV4PreTrainedModel as MobileNetV4PreTrainedModel, __webpack_exports__MobileViTFeatureExtractor as MobileViTFeatureExtractor, __webpack_exports__MobileViTForImageClassification as MobileViTForImageClassification, __webpack_exports__MobileViTImageProcessor as MobileViTImageProcessor, __webpack_exports__MobileViTModel as MobileViTModel, __webpack_exports__MobileViTPreTrainedModel as MobileViTPreTrainedModel, __webpack_exports__MobileViTV2ForImageClassification as MobileViTV2ForImageClassification, __webpack_exports__MobileViTV2Model as MobileViTV2Model, __webpack_exports__MobileViTV2PreTrainedModel as MobileViTV2PreTrainedModel, __webpack_exports__ModelOutput as ModelOutput, __webpack_exports__ModernBertForMaskedLM as ModernBertForMaskedLM, __webpack_exports__ModernBertForSequenceClassification as ModernBertForSequenceClassification, __webpack_exports__ModernBertForTokenClassification as ModernBertForTokenClassification, __webpack_exports__ModernBertModel as ModernBertModel, __webpack_exports__ModernBertPreTrainedModel as ModernBertPreTrainedModel, __webpack_exports__Moondream1ForConditionalGeneration as Moondream1ForConditionalGeneration, __webpack_exports__MoonshineFeatureExtractor as MoonshineFeatureExtractor, __webpack_exports__MoonshineForConditionalGeneration as MoonshineForConditionalGeneration, __webpack_exports__MoonshineModel as MoonshineModel, __webpack_exports__MoonshinePreTrainedModel as MoonshinePreTrainedModel, __webpack_exports__MoonshineProcessor as MoonshineProcessor, __webpack_exports__MptForCausalLM as MptForCausalLM, __webpack_exports__MptModel as MptModel, __webpack_exports__MptPreTrainedModel as MptPreTrainedModel, __webpack_exports__MultiModalityCausalLM as MultiModalityCausalLM, __webpack_exports__MultiModalityPreTrainedModel as MultiModalityPreTrainedModel, __webpack_exports__MusicgenForCausalLM as MusicgenForCausalLM, __webpack_exports__MusicgenForConditionalGeneration as MusicgenForConditionalGeneration, __webpack_exports__MusicgenModel as MusicgenModel, __webpack_exports__MusicgenPreTrainedModel as MusicgenPreTrainedModel, __webpack_exports__NllbTokenizer as NllbTokenizer, __webpack_exports__NoBadWordsLogitsProcessor as NoBadWordsLogitsProcessor, __webpack_exports__NoRepeatNGramLogitsProcessor as NoRepeatNGramLogitsProcessor, __webpack_exports__NomicBertModel as NomicBertModel, __webpack_exports__NomicBertPreTrainedModel as NomicBertPreTrainedModel, __webpack_exports__NougatImageProcessor as NougatImageProcessor, __webpack_exports__NougatTokenizer as NougatTokenizer, __webpack_exports__OPTForCausalLM as OPTForCausalLM, __webpack_exports__OPTModel as OPTModel, __webpack_exports__OPTPreTrainedModel as OPTPreTrainedModel, __webpack_exports__ObjectDetectionPipeline as ObjectDetectionPipeline, __webpack_exports__Olmo2ForCausalLM as Olmo2ForCausalLM, __webpack_exports__Olmo2Model as Olmo2Model, __webpack_exports__Olmo2PreTrainedModel as Olmo2PreTrainedModel, __webpack_exports__OlmoForCausalLM as OlmoForCausalLM, __webpack_exports__OlmoModel as OlmoModel, __webpack_exports__OlmoPreTrainedModel as OlmoPreTrainedModel, __webpack_exports__OpenELMForCausalLM as OpenELMForCausalLM, __webpack_exports__OpenELMModel as OpenELMModel, __webpack_exports__OpenELMPreTrainedModel as OpenELMPreTrainedModel, __webpack_exports__OwlViTFeatureExtractor as OwlViTFeatureExtractor, __webpack_exports__OwlViTForObjectDetection as OwlViTForObjectDetection, __webpack_exports__OwlViTImageProcessor as OwlViTImageProcessor, __webpack_exports__OwlViTModel as OwlViTModel, __webpack_exports__OwlViTPreTrainedModel as OwlViTPreTrainedModel, __webpack_exports__OwlViTProcessor as OwlViTProcessor, __webpack_exports__Owlv2ForObjectDetection as Owlv2ForObjectDetection, __webpack_exports__Owlv2ImageProcessor as Owlv2ImageProcessor, __webpack_exports__Owlv2Model as Owlv2Model, __webpack_exports__Owlv2PreTrainedModel as Owlv2PreTrainedModel, __webpack_exports__PaliGemmaForConditionalGeneration as PaliGemmaForConditionalGeneration, __webpack_exports__PaliGemmaPreTrainedModel as PaliGemmaPreTrainedModel, __webpack_exports__PaliGemmaProcessor as PaliGemmaProcessor, __webpack_exports__PatchTSMixerForPrediction as PatchTSMixerForPrediction, __webpack_exports__PatchTSMixerModel as PatchTSMixerModel, __webpack_exports__PatchTSMixerPreTrainedModel as PatchTSMixerPreTrainedModel, __webpack_exports__PatchTSTForPrediction as PatchTSTForPrediction, __webpack_exports__PatchTSTModel as PatchTSTModel, __webpack_exports__PatchTSTPreTrainedModel as PatchTSTPreTrainedModel, __webpack_exports__Phi3ForCausalLM as Phi3ForCausalLM, __webpack_exports__Phi3Model as Phi3Model, __webpack_exports__Phi3PreTrainedModel as Phi3PreTrainedModel, __webpack_exports__Phi3VForCausalLM as Phi3VForCausalLM, __webpack_exports__Phi3VImageProcessor as Phi3VImageProcessor, __webpack_exports__Phi3VPreTrainedModel as Phi3VPreTrainedModel, __webpack_exports__Phi3VProcessor as Phi3VProcessor, __webpack_exports__PhiForCausalLM as PhiForCausalLM, __webpack_exports__PhiModel as PhiModel, __webpack_exports__PhiPreTrainedModel as PhiPreTrainedModel, __webpack_exports__Pipeline as Pipeline, __webpack_exports__PreTrainedModel as PreTrainedModel, __webpack_exports__PreTrainedTokenizer as PreTrainedTokenizer, __webpack_exports__PretrainedConfig as PretrainedConfig, __webpack_exports__PretrainedMixin as PretrainedMixin, __webpack_exports__Processor as Processor, __webpack_exports__PvtForImageClassification as PvtForImageClassification, __webpack_exports__PvtImageProcessor as PvtImageProcessor, __webpack_exports__PvtModel as PvtModel, __webpack_exports__PvtPreTrainedModel as PvtPreTrainedModel, __webpack_exports__PyAnnoteFeatureExtractor as PyAnnoteFeatureExtractor, __webpack_exports__PyAnnoteForAudioFrameClassification as PyAnnoteForAudioFrameClassification, __webpack_exports__PyAnnoteModel as PyAnnoteModel, __webpack_exports__PyAnnotePreTrainedModel as PyAnnotePreTrainedModel, __webpack_exports__PyAnnoteProcessor as PyAnnoteProcessor, __webpack_exports__QuestionAnsweringModelOutput as QuestionAnsweringModelOutput, __webpack_exports__QuestionAnsweringPipeline as QuestionAnsweringPipeline, __webpack_exports__Qwen2ForCausalLM as Qwen2ForCausalLM, __webpack_exports__Qwen2Model as Qwen2Model, __webpack_exports__Qwen2PreTrainedModel as Qwen2PreTrainedModel, __webpack_exports__Qwen2Tokenizer as Qwen2Tokenizer, __webpack_exports__Qwen2VLForConditionalGeneration as Qwen2VLForConditionalGeneration, __webpack_exports__Qwen2VLImageProcessor as Qwen2VLImageProcessor, __webpack_exports__Qwen2VLPreTrainedModel as Qwen2VLPreTrainedModel, __webpack_exports__Qwen2VLProcessor as Qwen2VLProcessor, __webpack_exports__RTDetrForObjectDetection as RTDetrForObjectDetection, __webpack_exports__RTDetrImageProcessor as RTDetrImageProcessor, __webpack_exports__RTDetrModel as RTDetrModel, __webpack_exports__RTDetrObjectDetectionOutput as RTDetrObjectDetectionOutput, __webpack_exports__RTDetrPreTrainedModel as RTDetrPreTrainedModel, __webpack_exports__RawImage as RawImage, __webpack_exports__RepetitionPenaltyLogitsProcessor as RepetitionPenaltyLogitsProcessor, __webpack_exports__ResNetForImageClassification as ResNetForImageClassification, __webpack_exports__ResNetModel as ResNetModel, __webpack_exports__ResNetPreTrainedModel as ResNetPreTrainedModel, __webpack_exports__RoFormerForMaskedLM as RoFormerForMaskedLM, __webpack_exports__RoFormerForQuestionAnswering as RoFormerForQuestionAnswering, __webpack_exports__RoFormerForSequenceClassification as RoFormerForSequenceClassification, __webpack_exports__RoFormerForTokenClassification as RoFormerForTokenClassification, __webpack_exports__RoFormerModel as RoFormerModel, __webpack_exports__RoFormerPreTrainedModel as RoFormerPreTrainedModel, __webpack_exports__RoFormerTokenizer as RoFormerTokenizer, __webpack_exports__RobertaForMaskedLM as RobertaForMaskedLM, __webpack_exports__RobertaForQuestionAnswering as RobertaForQuestionAnswering, __webpack_exports__RobertaForSequenceClassification as RobertaForSequenceClassification, __webpack_exports__RobertaForTokenClassification as RobertaForTokenClassification, __webpack_exports__RobertaModel as RobertaModel, __webpack_exports__RobertaPreTrainedModel as RobertaPreTrainedModel, __webpack_exports__RobertaTokenizer as RobertaTokenizer, __webpack_exports__SamImageProcessor as SamImageProcessor, __webpack_exports__SamImageSegmentationOutput as SamImageSegmentationOutput, __webpack_exports__SamModel as SamModel, __webpack_exports__SamPreTrainedModel as SamPreTrainedModel, __webpack_exports__SamProcessor as SamProcessor, __webpack_exports__SapiensForDepthEstimation as SapiensForDepthEstimation, __webpack_exports__SapiensForNormalEstimation as SapiensForNormalEstimation, __webpack_exports__SapiensForSemanticSegmentation as SapiensForSemanticSegmentation, __webpack_exports__SapiensPreTrainedModel as SapiensPreTrainedModel, __webpack_exports__SeamlessM4TFeatureExtractor as SeamlessM4TFeatureExtractor, __webpack_exports__SegformerFeatureExtractor as SegformerFeatureExtractor, __webpack_exports__SegformerForImageClassification as SegformerForImageClassification, __webpack_exports__SegformerForSemanticSegmentation as SegformerForSemanticSegmentation, __webpack_exports__SegformerImageProcessor as SegformerImageProcessor, __webpack_exports__SegformerModel as SegformerModel, __webpack_exports__SegformerPreTrainedModel as SegformerPreTrainedModel, __webpack_exports__Seq2SeqLMOutput as Seq2SeqLMOutput, __webpack_exports__SequenceClassifierOutput as SequenceClassifierOutput, __webpack_exports__SiglipImageProcessor as SiglipImageProcessor, __webpack_exports__SiglipModel as SiglipModel, __webpack_exports__SiglipPreTrainedModel as SiglipPreTrainedModel, __webpack_exports__SiglipTextModel as SiglipTextModel, __webpack_exports__SiglipTokenizer as SiglipTokenizer, __webpack_exports__SiglipVisionModel as SiglipVisionModel, __webpack_exports__SpeechT5FeatureExtractor as SpeechT5FeatureExtractor, __webpack_exports__SpeechT5ForSpeechToText as SpeechT5ForSpeechToText, __webpack_exports__SpeechT5ForTextToSpeech as SpeechT5ForTextToSpeech, __webpack_exports__SpeechT5HifiGan as SpeechT5HifiGan, __webpack_exports__SpeechT5Model as SpeechT5Model, __webpack_exports__SpeechT5PreTrainedModel as SpeechT5PreTrainedModel, __webpack_exports__SpeechT5Processor as SpeechT5Processor, __webpack_exports__SpeechT5Tokenizer as SpeechT5Tokenizer, __webpack_exports__SqueezeBertForMaskedLM as SqueezeBertForMaskedLM, __webpack_exports__SqueezeBertForQuestionAnswering as SqueezeBertForQuestionAnswering, __webpack_exports__SqueezeBertForSequenceClassification as SqueezeBertForSequenceClassification, __webpack_exports__SqueezeBertModel as SqueezeBertModel, __webpack_exports__SqueezeBertPreTrainedModel as SqueezeBertPreTrainedModel, __webpack_exports__SqueezeBertTokenizer as SqueezeBertTokenizer, __webpack_exports__StableLmForCausalLM as StableLmForCausalLM, __webpack_exports__StableLmModel as StableLmModel, __webpack_exports__StableLmPreTrainedModel as StableLmPreTrainedModel, __webpack_exports__Starcoder2ForCausalLM as Starcoder2ForCausalLM, __webpack_exports__Starcoder2Model as Starcoder2Model, __webpack_exports__Starcoder2PreTrainedModel as Starcoder2PreTrainedModel, __webpack_exports__StoppingCriteria as StoppingCriteria, __webpack_exports__StoppingCriteriaList as StoppingCriteriaList, __webpack_exports__SummarizationPipeline as SummarizationPipeline, __webpack_exports__SuppressTokensAtBeginLogitsProcessor as SuppressTokensAtBeginLogitsProcessor, __webpack_exports__Swin2SRForImageSuperResolution as Swin2SRForImageSuperResolution, __webpack_exports__Swin2SRImageProcessor as Swin2SRImageProcessor, __webpack_exports__Swin2SRModel as Swin2SRModel, __webpack_exports__Swin2SRPreTrainedModel as Swin2SRPreTrainedModel, __webpack_exports__SwinForImageClassification as SwinForImageClassification, __webpack_exports__SwinModel as SwinModel, __webpack_exports__SwinPreTrainedModel as SwinPreTrainedModel, __webpack_exports__T5ForConditionalGeneration as T5ForConditionalGeneration, __webpack_exports__T5Model as T5Model, __webpack_exports__T5PreTrainedModel as T5PreTrainedModel, __webpack_exports__T5Tokenizer as T5Tokenizer, __webpack_exports__TableTransformerForObjectDetection as TableTransformerForObjectDetection, __webpack_exports__TableTransformerModel as TableTransformerModel, __webpack_exports__TableTransformerObjectDetectionOutput as TableTransformerObjectDetectionOutput, __webpack_exports__TableTransformerPreTrainedModel as TableTransformerPreTrainedModel, __webpack_exports__TemperatureLogitsWarper as TemperatureLogitsWarper, __webpack_exports__Tensor as Tensor, __webpack_exports__Text2TextGenerationPipeline as Text2TextGenerationPipeline, __webpack_exports__TextClassificationPipeline as TextClassificationPipeline, __webpack_exports__TextGenerationPipeline as TextGenerationPipeline, __webpack_exports__TextStreamer as TextStreamer, __webpack_exports__TextToAudioPipeline as TextToAudioPipeline, __webpack_exports__TokenClassificationPipeline as TokenClassificationPipeline, __webpack_exports__TokenClassifierOutput as TokenClassifierOutput, __webpack_exports__TokenizerModel as TokenizerModel, __webpack_exports__TopKLogitsWarper as TopKLogitsWarper, __webpack_exports__TopPLogitsWarper as TopPLogitsWarper, __webpack_exports__TrOCRForCausalLM as TrOCRForCausalLM, __webpack_exports__TrOCRPreTrainedModel as TrOCRPreTrainedModel, __webpack_exports__TranslationPipeline as TranslationPipeline, __webpack_exports__UniSpeechForCTC as UniSpeechForCTC, __webpack_exports__UniSpeechForSequenceClassification as UniSpeechForSequenceClassification, __webpack_exports__UniSpeechModel as UniSpeechModel, __webpack_exports__UniSpeechPreTrainedModel as UniSpeechPreTrainedModel, __webpack_exports__UniSpeechSatForAudioFrameClassification as UniSpeechSatForAudioFrameClassification, __webpack_exports__UniSpeechSatForCTC as UniSpeechSatForCTC, __webpack_exports__UniSpeechSatForSequenceClassification as UniSpeechSatForSequenceClassification, __webpack_exports__UniSpeechSatModel as UniSpeechSatModel, __webpack_exports__UniSpeechSatPreTrainedModel as UniSpeechSatPreTrainedModel, __webpack_exports__VLChatProcessor as VLChatProcessor, __webpack_exports__VLMImageProcessor as VLMImageProcessor, __webpack_exports__ViTFeatureExtractor as ViTFeatureExtractor, __webpack_exports__ViTForImageClassification as ViTForImageClassification, __webpack_exports__ViTImageProcessor as ViTImageProcessor, __webpack_exports__ViTMAEModel as ViTMAEModel, __webpack_exports__ViTMAEPreTrainedModel as ViTMAEPreTrainedModel, __webpack_exports__ViTMSNForImageClassification as ViTMSNForImageClassification, __webpack_exports__ViTMSNModel as ViTMSNModel, __webpack_exports__ViTMSNPreTrainedModel as ViTMSNPreTrainedModel, __webpack_exports__ViTModel as ViTModel, __webpack_exports__ViTPreTrainedModel as ViTPreTrainedModel, __webpack_exports__VisionEncoderDecoderModel as VisionEncoderDecoderModel, __webpack_exports__VitMatteForImageMatting as VitMatteForImageMatting, __webpack_exports__VitMatteImageProcessor as VitMatteImageProcessor, __webpack_exports__VitMattePreTrainedModel as VitMattePreTrainedModel, __webpack_exports__VitPoseForPoseEstimation as VitPoseForPoseEstimation, __webpack_exports__VitPoseImageProcessor as VitPoseImageProcessor, __webpack_exports__VitPosePreTrainedModel as VitPosePreTrainedModel, __webpack_exports__VitsModel as VitsModel, __webpack_exports__VitsModelOutput as VitsModelOutput, __webpack_exports__VitsPreTrainedModel as VitsPreTrainedModel, __webpack_exports__VitsTokenizer as VitsTokenizer, __webpack_exports__Wav2Vec2BertForCTC as Wav2Vec2BertForCTC, __webpack_exports__Wav2Vec2BertForSequenceClassification as Wav2Vec2BertForSequenceClassification, __webpack_exports__Wav2Vec2BertModel as Wav2Vec2BertModel, __webpack_exports__Wav2Vec2BertPreTrainedModel as Wav2Vec2BertPreTrainedModel, __webpack_exports__Wav2Vec2CTCTokenizer as Wav2Vec2CTCTokenizer, __webpack_exports__Wav2Vec2FeatureExtractor as Wav2Vec2FeatureExtractor, __webpack_exports__Wav2Vec2ForAudioFrameClassification as Wav2Vec2ForAudioFrameClassification, __webpack_exports__Wav2Vec2ForCTC as Wav2Vec2ForCTC, __webpack_exports__Wav2Vec2ForSequenceClassification as Wav2Vec2ForSequenceClassification, __webpack_exports__Wav2Vec2Model as Wav2Vec2Model, __webpack_exports__Wav2Vec2PreTrainedModel as Wav2Vec2PreTrainedModel, __webpack_exports__Wav2Vec2ProcessorWithLM as Wav2Vec2ProcessorWithLM, __webpack_exports__WavLMForAudioFrameClassification as WavLMForAudioFrameClassification, __webpack_exports__WavLMForCTC as WavLMForCTC, __webpack_exports__WavLMForSequenceClassification as WavLMForSequenceClassification, __webpack_exports__WavLMForXVector as WavLMForXVector, __webpack_exports__WavLMModel as WavLMModel, __webpack_exports__WavLMPreTrainedModel as WavLMPreTrainedModel, __webpack_exports__WeSpeakerFeatureExtractor as WeSpeakerFeatureExtractor, __webpack_exports__WeSpeakerResNetModel as WeSpeakerResNetModel, __webpack_exports__WeSpeakerResNetPreTrainedModel as WeSpeakerResNetPreTrainedModel, __webpack_exports__WhisperFeatureExtractor as WhisperFeatureExtractor, __webpack_exports__WhisperForConditionalGeneration as WhisperForConditionalGeneration, __webpack_exports__WhisperModel as WhisperModel, __webpack_exports__WhisperPreTrainedModel as WhisperPreTrainedModel, __webpack_exports__WhisperProcessor as WhisperProcessor, __webpack_exports__WhisperTextStreamer as WhisperTextStreamer, __webpack_exports__WhisperTimeStampLogitsProcessor as WhisperTimeStampLogitsProcessor, __webpack_exports__WhisperTokenizer as WhisperTokenizer, __webpack_exports__XLMForQuestionAnswering as XLMForQuestionAnswering, __webpack_exports__XLMForSequenceClassification as XLMForSequenceClassification, __webpack_exports__XLMForTokenClassification as XLMForTokenClassification, __webpack_exports__XLMModel as XLMModel, __webpack_exports__XLMPreTrainedModel as XLMPreTrainedModel, __webpack_exports__XLMRobertaForMaskedLM as XLMRobertaForMaskedLM, __webpack_exports__XLMRobertaForQuestionAnswering as XLMRobertaForQuestionAnswering, __webpack_exports__XLMRobertaForSequenceClassification as XLMRobertaForSequenceClassification, __webpack_exports__XLMRobertaForTokenClassification as XLMRobertaForTokenClassification, __webpack_exports__XLMRobertaModel as XLMRobertaModel, __webpack_exports__XLMRobertaPreTrainedModel as XLMRobertaPreTrainedModel, __webpack_exports__XLMRobertaTokenizer as XLMRobertaTokenizer, __webpack_exports__XLMTokenizer as XLMTokenizer, __webpack_exports__XLMWithLMHeadModel as XLMWithLMHeadModel, __webpack_exports__XVectorOutput as XVectorOutput, __webpack_exports__YolosFeatureExtractor as YolosFeatureExtractor, __webpack_exports__YolosForObjectDetection as YolosForObjectDetection, __webpack_exports__YolosImageProcessor as YolosImageProcessor, __webpack_exports__YolosModel as YolosModel, __webpack_exports__YolosObjectDetectionOutput as YolosObjectDetectionOutput, __webpack_exports__YolosPreTrainedModel as YolosPreTrainedModel, __webpack_exports__ZeroShotAudioClassificationPipeline as ZeroShotAudioClassificationPipeline, __webpack_exports__ZeroShotClassificationPipeline as ZeroShotClassificationPipeline, __webpack_exports__ZeroShotImageClassificationPipeline as ZeroShotImageClassificationPipeline, __webpack_exports__ZeroShotObjectDetectionPipeline as ZeroShotObjectDetectionPipeline, __webpack_exports__bankers_round as bankers_round, __webpack_exports__cat as cat, __webpack_exports__cos_sim as cos_sim, __webpack_exports__dot as dot, __webpack_exports__dynamic_time_warping as dynamic_time_warping, __webpack_exports__env as env, __webpack_exports__full as full, __webpack_exports__full_like as full_like, __webpack_exports__getKeyValueShapes as getKeyValueShapes, __webpack_exports__hamming as hamming, __webpack_exports__hanning as hanning, __webpack_exports__interpolate as interpolate, __webpack_exports__interpolate_4d as interpolate_4d, __webpack_exports__interpolate_data as interpolate_data, __webpack_exports__is_chinese_char as is_chinese_char, __webpack_exports__layer_norm as layer_norm, __webpack_exports__load_image as load_image, __webpack_exports__log_softmax as log_softmax, __webpack_exports__magnitude as magnitude, __webpack_exports__matmul as matmul, __webpack_exports__max as max, __webpack_exports__mean as mean, __webpack_exports__mean_pooling as mean_pooling, __webpack_exports__medianFilter as medianFilter, __webpack_exports__mel_filter_bank as mel_filter_bank, __webpack_exports__min as min, __webpack_exports__ones as ones, __webpack_exports__ones_like as ones_like, __webpack_exports__permute as permute, __webpack_exports__permute_data as permute_data, __webpack_exports__pipeline as pipeline, __webpack_exports__quantize_embeddings as quantize_embeddings, __webpack_exports__rand as rand, __webpack_exports__read_audio as read_audio, __webpack_exports__rfft as rfft, __webpack_exports__round as round, __webpack_exports__slice as slice, __webpack_exports__softmax as softmax, __webpack_exports__spectrogram as spectrogram, __webpack_exports__stack as stack, __webpack_exports__std_mean as std_mean, __webpack_exports__topk as topk, __webpack_exports__window_function as window_function, __webpack_exports__zeros as zeros, __webpack_exports__zeros_like as zeros_like };
|
|
36480
|
+
export { __webpack_exports__ASTFeatureExtractor as ASTFeatureExtractor, __webpack_exports__ASTForAudioClassification as ASTForAudioClassification, __webpack_exports__ASTModel as ASTModel, __webpack_exports__ASTPreTrainedModel as ASTPreTrainedModel, __webpack_exports__AlbertForMaskedLM as AlbertForMaskedLM, __webpack_exports__AlbertForQuestionAnswering as AlbertForQuestionAnswering, __webpack_exports__AlbertForSequenceClassification as AlbertForSequenceClassification, __webpack_exports__AlbertModel as AlbertModel, __webpack_exports__AlbertPreTrainedModel as AlbertPreTrainedModel, __webpack_exports__AlbertTokenizer as AlbertTokenizer, __webpack_exports__AudioClassificationPipeline as AudioClassificationPipeline, __webpack_exports__AutoConfig as AutoConfig, __webpack_exports__AutoFeatureExtractor as AutoFeatureExtractor, __webpack_exports__AutoImageProcessor as AutoImageProcessor, __webpack_exports__AutoModel as AutoModel, __webpack_exports__AutoModelForAudioClassification as AutoModelForAudioClassification, __webpack_exports__AutoModelForAudioFrameClassification as AutoModelForAudioFrameClassification, __webpack_exports__AutoModelForCTC as AutoModelForCTC, __webpack_exports__AutoModelForCausalLM as AutoModelForCausalLM, __webpack_exports__AutoModelForDepthEstimation as AutoModelForDepthEstimation, __webpack_exports__AutoModelForDocumentQuestionAnswering as AutoModelForDocumentQuestionAnswering, __webpack_exports__AutoModelForImageClassification as AutoModelForImageClassification, __webpack_exports__AutoModelForImageFeatureExtraction as AutoModelForImageFeatureExtraction, __webpack_exports__AutoModelForImageMatting as AutoModelForImageMatting, __webpack_exports__AutoModelForImageSegmentation as AutoModelForImageSegmentation, __webpack_exports__AutoModelForImageToImage as AutoModelForImageToImage, __webpack_exports__AutoModelForMaskGeneration as AutoModelForMaskGeneration, __webpack_exports__AutoModelForMaskedLM as AutoModelForMaskedLM, __webpack_exports__AutoModelForNormalEstimation as AutoModelForNormalEstimation, __webpack_exports__AutoModelForObjectDetection as AutoModelForObjectDetection, __webpack_exports__AutoModelForPoseEstimation as AutoModelForPoseEstimation, __webpack_exports__AutoModelForQuestionAnswering as AutoModelForQuestionAnswering, __webpack_exports__AutoModelForSemanticSegmentation as AutoModelForSemanticSegmentation, __webpack_exports__AutoModelForSeq2SeqLM as AutoModelForSeq2SeqLM, __webpack_exports__AutoModelForSequenceClassification as AutoModelForSequenceClassification, __webpack_exports__AutoModelForSpeechSeq2Seq as AutoModelForSpeechSeq2Seq, __webpack_exports__AutoModelForTextToSpectrogram as AutoModelForTextToSpectrogram, __webpack_exports__AutoModelForTextToWaveform as AutoModelForTextToWaveform, __webpack_exports__AutoModelForTokenClassification as AutoModelForTokenClassification, __webpack_exports__AutoModelForUniversalSegmentation as AutoModelForUniversalSegmentation, __webpack_exports__AutoModelForVision2Seq as AutoModelForVision2Seq, __webpack_exports__AutoModelForXVector as AutoModelForXVector, __webpack_exports__AutoModelForZeroShotObjectDetection as AutoModelForZeroShotObjectDetection, __webpack_exports__AutoProcessor as AutoProcessor, __webpack_exports__AutoTokenizer as AutoTokenizer, __webpack_exports__AutomaticSpeechRecognitionPipeline as AutomaticSpeechRecognitionPipeline, __webpack_exports__BartForConditionalGeneration as BartForConditionalGeneration, __webpack_exports__BartForSequenceClassification as BartForSequenceClassification, __webpack_exports__BartModel as BartModel, __webpack_exports__BartPretrainedModel as BartPretrainedModel, __webpack_exports__BartTokenizer as BartTokenizer, __webpack_exports__BaseModelOutput as BaseModelOutput, __webpack_exports__BaseStreamer as BaseStreamer, __webpack_exports__BeitFeatureExtractor as BeitFeatureExtractor, __webpack_exports__BeitForImageClassification as BeitForImageClassification, __webpack_exports__BeitModel as BeitModel, __webpack_exports__BeitPreTrainedModel as BeitPreTrainedModel, __webpack_exports__BertForMaskedLM as BertForMaskedLM, __webpack_exports__BertForQuestionAnswering as BertForQuestionAnswering, __webpack_exports__BertForSequenceClassification as BertForSequenceClassification, __webpack_exports__BertForTokenClassification as BertForTokenClassification, __webpack_exports__BertModel as BertModel, __webpack_exports__BertPreTrainedModel as BertPreTrainedModel, __webpack_exports__BertTokenizer as BertTokenizer, __webpack_exports__BitImageProcessor as BitImageProcessor, __webpack_exports__BlenderbotForConditionalGeneration as BlenderbotForConditionalGeneration, __webpack_exports__BlenderbotModel as BlenderbotModel, __webpack_exports__BlenderbotPreTrainedModel as BlenderbotPreTrainedModel, __webpack_exports__BlenderbotSmallForConditionalGeneration as BlenderbotSmallForConditionalGeneration, __webpack_exports__BlenderbotSmallModel as BlenderbotSmallModel, __webpack_exports__BlenderbotSmallPreTrainedModel as BlenderbotSmallPreTrainedModel, __webpack_exports__BlenderbotSmallTokenizer as BlenderbotSmallTokenizer, __webpack_exports__BlenderbotTokenizer as BlenderbotTokenizer, __webpack_exports__BloomForCausalLM as BloomForCausalLM, __webpack_exports__BloomModel as BloomModel, __webpack_exports__BloomPreTrainedModel as BloomPreTrainedModel, __webpack_exports__BloomTokenizer as BloomTokenizer, __webpack_exports__CLIPFeatureExtractor as CLIPFeatureExtractor, __webpack_exports__CLIPImageProcessor as CLIPImageProcessor, __webpack_exports__CLIPModel as CLIPModel, __webpack_exports__CLIPPreTrainedModel as CLIPPreTrainedModel, __webpack_exports__CLIPSegForImageSegmentation as CLIPSegForImageSegmentation, __webpack_exports__CLIPSegModel as CLIPSegModel, __webpack_exports__CLIPSegPreTrainedModel as CLIPSegPreTrainedModel, __webpack_exports__CLIPTextModel as CLIPTextModel, __webpack_exports__CLIPTextModelWithProjection as CLIPTextModelWithProjection, __webpack_exports__CLIPTokenizer as CLIPTokenizer, __webpack_exports__CLIPVisionModel as CLIPVisionModel, __webpack_exports__CLIPVisionModelWithProjection as CLIPVisionModelWithProjection, __webpack_exports__CamembertForMaskedLM as CamembertForMaskedLM, __webpack_exports__CamembertForQuestionAnswering as CamembertForQuestionAnswering, __webpack_exports__CamembertForSequenceClassification as CamembertForSequenceClassification, __webpack_exports__CamembertForTokenClassification as CamembertForTokenClassification, __webpack_exports__CamembertModel as CamembertModel, __webpack_exports__CamembertPreTrainedModel as CamembertPreTrainedModel, __webpack_exports__CamembertTokenizer as CamembertTokenizer, __webpack_exports__CausalLMOutput as CausalLMOutput, __webpack_exports__CausalLMOutputWithPast as CausalLMOutputWithPast, __webpack_exports__ChineseCLIPFeatureExtractor as ChineseCLIPFeatureExtractor, __webpack_exports__ChineseCLIPModel as ChineseCLIPModel, __webpack_exports__ChineseCLIPPreTrainedModel as ChineseCLIPPreTrainedModel, __webpack_exports__ClapAudioModelWithProjection as ClapAudioModelWithProjection, __webpack_exports__ClapFeatureExtractor as ClapFeatureExtractor, __webpack_exports__ClapModel as ClapModel, __webpack_exports__ClapPreTrainedModel as ClapPreTrainedModel, __webpack_exports__ClapTextModelWithProjection as ClapTextModelWithProjection, __webpack_exports__ClassifierFreeGuidanceLogitsProcessor as ClassifierFreeGuidanceLogitsProcessor, __webpack_exports__CodeGenForCausalLM as CodeGenForCausalLM, __webpack_exports__CodeGenModel as CodeGenModel, __webpack_exports__CodeGenPreTrainedModel as CodeGenPreTrainedModel, __webpack_exports__CodeGenTokenizer as CodeGenTokenizer, __webpack_exports__CodeLlamaTokenizer as CodeLlamaTokenizer, __webpack_exports__CohereForCausalLM as CohereForCausalLM, __webpack_exports__CohereModel as CohereModel, __webpack_exports__CoherePreTrainedModel as CoherePreTrainedModel, __webpack_exports__CohereTokenizer as CohereTokenizer, __webpack_exports__ConvBertForMaskedLM as ConvBertForMaskedLM, __webpack_exports__ConvBertForQuestionAnswering as ConvBertForQuestionAnswering, __webpack_exports__ConvBertForSequenceClassification as ConvBertForSequenceClassification, __webpack_exports__ConvBertForTokenClassification as ConvBertForTokenClassification, __webpack_exports__ConvBertModel as ConvBertModel, __webpack_exports__ConvBertPreTrainedModel as ConvBertPreTrainedModel, __webpack_exports__ConvBertTokenizer as ConvBertTokenizer, __webpack_exports__ConvNextFeatureExtractor as ConvNextFeatureExtractor, __webpack_exports__ConvNextForImageClassification as ConvNextForImageClassification, __webpack_exports__ConvNextImageProcessor as ConvNextImageProcessor, __webpack_exports__ConvNextModel as ConvNextModel, __webpack_exports__ConvNextPreTrainedModel as ConvNextPreTrainedModel, __webpack_exports__ConvNextV2ForImageClassification as ConvNextV2ForImageClassification, __webpack_exports__ConvNextV2Model as ConvNextV2Model, __webpack_exports__ConvNextV2PreTrainedModel as ConvNextV2PreTrainedModel, __webpack_exports__DPTFeatureExtractor as DPTFeatureExtractor, __webpack_exports__DPTForDepthEstimation as DPTForDepthEstimation, __webpack_exports__DPTImageProcessor as DPTImageProcessor, __webpack_exports__DPTModel as DPTModel, __webpack_exports__DPTPreTrainedModel as DPTPreTrainedModel, __webpack_exports__DebertaForMaskedLM as DebertaForMaskedLM, __webpack_exports__DebertaForQuestionAnswering as DebertaForQuestionAnswering, __webpack_exports__DebertaForSequenceClassification as DebertaForSequenceClassification, __webpack_exports__DebertaForTokenClassification as DebertaForTokenClassification, __webpack_exports__DebertaModel as DebertaModel, __webpack_exports__DebertaPreTrainedModel as DebertaPreTrainedModel, __webpack_exports__DebertaTokenizer as DebertaTokenizer, __webpack_exports__DebertaV2ForMaskedLM as DebertaV2ForMaskedLM, __webpack_exports__DebertaV2ForQuestionAnswering as DebertaV2ForQuestionAnswering, __webpack_exports__DebertaV2ForSequenceClassification as DebertaV2ForSequenceClassification, __webpack_exports__DebertaV2ForTokenClassification as DebertaV2ForTokenClassification, __webpack_exports__DebertaV2Model as DebertaV2Model, __webpack_exports__DebertaV2PreTrainedModel as DebertaV2PreTrainedModel, __webpack_exports__DebertaV2Tokenizer as DebertaV2Tokenizer, __webpack_exports__DecisionTransformerModel as DecisionTransformerModel, __webpack_exports__DecisionTransformerPreTrainedModel as DecisionTransformerPreTrainedModel, __webpack_exports__DeiTFeatureExtractor as DeiTFeatureExtractor, __webpack_exports__DeiTForImageClassification as DeiTForImageClassification, __webpack_exports__DeiTImageProcessor as DeiTImageProcessor, __webpack_exports__DeiTModel as DeiTModel, __webpack_exports__DeiTPreTrainedModel as DeiTPreTrainedModel, __webpack_exports__DepthAnythingForDepthEstimation as DepthAnythingForDepthEstimation, __webpack_exports__DepthAnythingPreTrainedModel as DepthAnythingPreTrainedModel, __webpack_exports__DepthEstimationPipeline as DepthEstimationPipeline, __webpack_exports__DepthProForDepthEstimation as DepthProForDepthEstimation, __webpack_exports__DepthProPreTrainedModel as DepthProPreTrainedModel, __webpack_exports__DetrFeatureExtractor as DetrFeatureExtractor, __webpack_exports__DetrForObjectDetection as DetrForObjectDetection, __webpack_exports__DetrForSegmentation as DetrForSegmentation, __webpack_exports__DetrImageProcessor as DetrImageProcessor, __webpack_exports__DetrModel as DetrModel, __webpack_exports__DetrObjectDetectionOutput as DetrObjectDetectionOutput, __webpack_exports__DetrPreTrainedModel as DetrPreTrainedModel, __webpack_exports__DetrSegmentationOutput as DetrSegmentationOutput, __webpack_exports__Dinov2ForImageClassification as Dinov2ForImageClassification, __webpack_exports__Dinov2Model as Dinov2Model, __webpack_exports__Dinov2PreTrainedModel as Dinov2PreTrainedModel, __webpack_exports__Dinov2WithRegistersForImageClassification as Dinov2WithRegistersForImageClassification, __webpack_exports__Dinov2WithRegistersModel as Dinov2WithRegistersModel, __webpack_exports__Dinov2WithRegistersPreTrainedModel as Dinov2WithRegistersPreTrainedModel, __webpack_exports__DistilBertForMaskedLM as DistilBertForMaskedLM, __webpack_exports__DistilBertForQuestionAnswering as DistilBertForQuestionAnswering, __webpack_exports__DistilBertForSequenceClassification as DistilBertForSequenceClassification, __webpack_exports__DistilBertForTokenClassification as DistilBertForTokenClassification, __webpack_exports__DistilBertModel as DistilBertModel, __webpack_exports__DistilBertPreTrainedModel as DistilBertPreTrainedModel, __webpack_exports__DistilBertTokenizer as DistilBertTokenizer, __webpack_exports__DocumentQuestionAnsweringPipeline as DocumentQuestionAnsweringPipeline, __webpack_exports__DonutFeatureExtractor as DonutFeatureExtractor, __webpack_exports__DonutImageProcessor as DonutImageProcessor, __webpack_exports__DonutSwinModel as DonutSwinModel, __webpack_exports__DonutSwinPreTrainedModel as DonutSwinPreTrainedModel, __webpack_exports__EfficientNetForImageClassification as EfficientNetForImageClassification, __webpack_exports__EfficientNetImageProcessor as EfficientNetImageProcessor, __webpack_exports__EfficientNetModel as EfficientNetModel, __webpack_exports__EfficientNetPreTrainedModel as EfficientNetPreTrainedModel, __webpack_exports__ElectraForMaskedLM as ElectraForMaskedLM, __webpack_exports__ElectraForQuestionAnswering as ElectraForQuestionAnswering, __webpack_exports__ElectraForSequenceClassification as ElectraForSequenceClassification, __webpack_exports__ElectraForTokenClassification as ElectraForTokenClassification, __webpack_exports__ElectraModel as ElectraModel, __webpack_exports__ElectraPreTrainedModel as ElectraPreTrainedModel, __webpack_exports__ElectraTokenizer as ElectraTokenizer, __webpack_exports__EosTokenCriteria as EosTokenCriteria, __webpack_exports__EsmForMaskedLM as EsmForMaskedLM, __webpack_exports__EsmForSequenceClassification as EsmForSequenceClassification, __webpack_exports__EsmForTokenClassification as EsmForTokenClassification, __webpack_exports__EsmModel as EsmModel, __webpack_exports__EsmPreTrainedModel as EsmPreTrainedModel, __webpack_exports__EsmTokenizer as EsmTokenizer, __webpack_exports__ExaoneForCausalLM as ExaoneForCausalLM, __webpack_exports__ExaoneModel as ExaoneModel, __webpack_exports__ExaonePreTrainedModel as ExaonePreTrainedModel, __webpack_exports__FFT as FFT, __webpack_exports__FalconForCausalLM as FalconForCausalLM, __webpack_exports__FalconModel as FalconModel, __webpack_exports__FalconPreTrainedModel as FalconPreTrainedModel, __webpack_exports__FalconTokenizer as FalconTokenizer, __webpack_exports__FastViTForImageClassification as FastViTForImageClassification, __webpack_exports__FastViTModel as FastViTModel, __webpack_exports__FastViTPreTrainedModel as FastViTPreTrainedModel, __webpack_exports__FeatureExtractionPipeline as FeatureExtractionPipeline, __webpack_exports__FeatureExtractor as FeatureExtractor, __webpack_exports__FillMaskPipeline as FillMaskPipeline, __webpack_exports__Florence2ForConditionalGeneration as Florence2ForConditionalGeneration, __webpack_exports__Florence2PreTrainedModel as Florence2PreTrainedModel, __webpack_exports__Florence2Processor as Florence2Processor, __webpack_exports__ForcedBOSTokenLogitsProcessor as ForcedBOSTokenLogitsProcessor, __webpack_exports__ForcedEOSTokenLogitsProcessor as ForcedEOSTokenLogitsProcessor, __webpack_exports__GLPNFeatureExtractor as GLPNFeatureExtractor, __webpack_exports__GLPNForDepthEstimation as GLPNForDepthEstimation, __webpack_exports__GLPNModel as GLPNModel, __webpack_exports__GLPNPreTrainedModel as GLPNPreTrainedModel, __webpack_exports__GPT2LMHeadModel as GPT2LMHeadModel, __webpack_exports__GPT2Model as GPT2Model, __webpack_exports__GPT2PreTrainedModel as GPT2PreTrainedModel, __webpack_exports__GPT2Tokenizer as GPT2Tokenizer, __webpack_exports__GPTBigCodeForCausalLM as GPTBigCodeForCausalLM, __webpack_exports__GPTBigCodeModel as GPTBigCodeModel, __webpack_exports__GPTBigCodePreTrainedModel as GPTBigCodePreTrainedModel, __webpack_exports__GPTJForCausalLM as GPTJForCausalLM, __webpack_exports__GPTJModel as GPTJModel, __webpack_exports__GPTJPreTrainedModel as GPTJPreTrainedModel, __webpack_exports__GPTNeoForCausalLM as GPTNeoForCausalLM, __webpack_exports__GPTNeoModel as GPTNeoModel, __webpack_exports__GPTNeoPreTrainedModel as GPTNeoPreTrainedModel, __webpack_exports__GPTNeoXForCausalLM as GPTNeoXForCausalLM, __webpack_exports__GPTNeoXModel as GPTNeoXModel, __webpack_exports__GPTNeoXPreTrainedModel as GPTNeoXPreTrainedModel, __webpack_exports__GPTNeoXTokenizer as GPTNeoXTokenizer, __webpack_exports__Gemma2ForCausalLM as Gemma2ForCausalLM, __webpack_exports__Gemma2Model as Gemma2Model, __webpack_exports__Gemma2PreTrainedModel as Gemma2PreTrainedModel, __webpack_exports__GemmaForCausalLM as GemmaForCausalLM, __webpack_exports__GemmaModel as GemmaModel, __webpack_exports__GemmaPreTrainedModel as GemmaPreTrainedModel, __webpack_exports__GemmaTokenizer as GemmaTokenizer, __webpack_exports__GraniteForCausalLM as GraniteForCausalLM, __webpack_exports__GraniteModel as GraniteModel, __webpack_exports__GranitePreTrainedModel as GranitePreTrainedModel, __webpack_exports__Grok1Tokenizer as Grok1Tokenizer, __webpack_exports__GroundingDinoForObjectDetection as GroundingDinoForObjectDetection, __webpack_exports__GroundingDinoImageProcessor as GroundingDinoImageProcessor, __webpack_exports__GroundingDinoPreTrainedModel as GroundingDinoPreTrainedModel, __webpack_exports__GroundingDinoProcessor as GroundingDinoProcessor, __webpack_exports__GroupViTModel as GroupViTModel, __webpack_exports__GroupViTPreTrainedModel as GroupViTPreTrainedModel, __webpack_exports__HerbertTokenizer as HerbertTokenizer, __webpack_exports__HieraForImageClassification as HieraForImageClassification, __webpack_exports__HieraModel as HieraModel, __webpack_exports__HieraPreTrainedModel as HieraPreTrainedModel, __webpack_exports__HubertForCTC as HubertForCTC, __webpack_exports__HubertForSequenceClassification as HubertForSequenceClassification, __webpack_exports__HubertModel as HubertModel, __webpack_exports__HubertPreTrainedModel as HubertPreTrainedModel, __webpack_exports__IJepaForImageClassification as IJepaForImageClassification, __webpack_exports__IJepaModel as IJepaModel, __webpack_exports__IJepaPreTrainedModel as IJepaPreTrainedModel, __webpack_exports__Idefics3ForConditionalGeneration as Idefics3ForConditionalGeneration, __webpack_exports__Idefics3ImageProcessor as Idefics3ImageProcessor, __webpack_exports__Idefics3PreTrainedModel as Idefics3PreTrainedModel, __webpack_exports__Idefics3Processor as Idefics3Processor, __webpack_exports__ImageClassificationPipeline as ImageClassificationPipeline, __webpack_exports__ImageFeatureExtractionPipeline as ImageFeatureExtractionPipeline, __webpack_exports__ImageFeatureExtractor as ImageFeatureExtractor, __webpack_exports__ImageMattingOutput as ImageMattingOutput, __webpack_exports__ImageProcessor as ImageProcessor, __webpack_exports__ImageSegmentationPipeline as ImageSegmentationPipeline, __webpack_exports__ImageToImagePipeline as ImageToImagePipeline, __webpack_exports__ImageToTextPipeline as ImageToTextPipeline, __webpack_exports__InterruptableStoppingCriteria as InterruptableStoppingCriteria, __webpack_exports__JAISLMHeadModel as JAISLMHeadModel, __webpack_exports__JAISModel as JAISModel, __webpack_exports__JAISPreTrainedModel as JAISPreTrainedModel, __webpack_exports__JinaCLIPImageProcessor as JinaCLIPImageProcessor, __webpack_exports__JinaCLIPModel as JinaCLIPModel, __webpack_exports__JinaCLIPPreTrainedModel as JinaCLIPPreTrainedModel, __webpack_exports__JinaCLIPProcessor as JinaCLIPProcessor, __webpack_exports__JinaCLIPTextModel as JinaCLIPTextModel, __webpack_exports__JinaCLIPVisionModel as JinaCLIPVisionModel, __webpack_exports__LlamaForCausalLM as LlamaForCausalLM, __webpack_exports__LlamaModel as LlamaModel, __webpack_exports__LlamaPreTrainedModel as LlamaPreTrainedModel, __webpack_exports__LlamaTokenizer as LlamaTokenizer, __webpack_exports__LlavaForConditionalGeneration as LlavaForConditionalGeneration, __webpack_exports__LlavaOnevisionForConditionalGeneration as LlavaOnevisionForConditionalGeneration, __webpack_exports__LlavaOnevisionImageProcessor as LlavaOnevisionImageProcessor, __webpack_exports__LlavaPreTrainedModel as LlavaPreTrainedModel, __webpack_exports__LogitsProcessor as LogitsProcessor, __webpack_exports__LogitsProcessorList as LogitsProcessorList, __webpack_exports__LogitsWarper as LogitsWarper, __webpack_exports__LongT5ForConditionalGeneration as LongT5ForConditionalGeneration, __webpack_exports__LongT5Model as LongT5Model, __webpack_exports__LongT5PreTrainedModel as LongT5PreTrainedModel, __webpack_exports__M2M100ForConditionalGeneration as M2M100ForConditionalGeneration, __webpack_exports__M2M100Model as M2M100Model, __webpack_exports__M2M100PreTrainedModel as M2M100PreTrainedModel, __webpack_exports__M2M100Tokenizer as M2M100Tokenizer, __webpack_exports__MBart50Tokenizer as MBart50Tokenizer, __webpack_exports__MBartForCausalLM as MBartForCausalLM, __webpack_exports__MBartForConditionalGeneration as MBartForConditionalGeneration, __webpack_exports__MBartForSequenceClassification as MBartForSequenceClassification, __webpack_exports__MBartModel as MBartModel, __webpack_exports__MBartPreTrainedModel as MBartPreTrainedModel, __webpack_exports__MBartTokenizer as MBartTokenizer, __webpack_exports__MPNetForMaskedLM as MPNetForMaskedLM, __webpack_exports__MPNetForQuestionAnswering as MPNetForQuestionAnswering, __webpack_exports__MPNetForSequenceClassification as MPNetForSequenceClassification, __webpack_exports__MPNetForTokenClassification as MPNetForTokenClassification, __webpack_exports__MPNetModel as MPNetModel, __webpack_exports__MPNetPreTrainedModel as MPNetPreTrainedModel, __webpack_exports__MPNetTokenizer as MPNetTokenizer, __webpack_exports__MT5ForConditionalGeneration as MT5ForConditionalGeneration, __webpack_exports__MT5Model as MT5Model, __webpack_exports__MT5PreTrainedModel as MT5PreTrainedModel, __webpack_exports__MarianMTModel as MarianMTModel, __webpack_exports__MarianModel as MarianModel, __webpack_exports__MarianPreTrainedModel as MarianPreTrainedModel, __webpack_exports__MarianTokenizer as MarianTokenizer, __webpack_exports__Mask2FormerImageProcessor as Mask2FormerImageProcessor, __webpack_exports__MaskFormerFeatureExtractor as MaskFormerFeatureExtractor, __webpack_exports__MaskFormerForInstanceSegmentation as MaskFormerForInstanceSegmentation, __webpack_exports__MaskFormerImageProcessor as MaskFormerImageProcessor, __webpack_exports__MaskFormerModel as MaskFormerModel, __webpack_exports__MaskFormerPreTrainedModel as MaskFormerPreTrainedModel, __webpack_exports__MaskedLMOutput as MaskedLMOutput, __webpack_exports__MaxLengthCriteria as MaxLengthCriteria, __webpack_exports__MgpstrForSceneTextRecognition as MgpstrForSceneTextRecognition, __webpack_exports__MgpstrModelOutput as MgpstrModelOutput, __webpack_exports__MgpstrPreTrainedModel as MgpstrPreTrainedModel, __webpack_exports__MgpstrProcessor as MgpstrProcessor, __webpack_exports__MgpstrTokenizer as MgpstrTokenizer, __webpack_exports__MinLengthLogitsProcessor as MinLengthLogitsProcessor, __webpack_exports__MinNewTokensLengthLogitsProcessor as MinNewTokensLengthLogitsProcessor, __webpack_exports__MistralForCausalLM as MistralForCausalLM, __webpack_exports__MistralModel as MistralModel, __webpack_exports__MistralPreTrainedModel as MistralPreTrainedModel, __webpack_exports__MobileBertForMaskedLM as MobileBertForMaskedLM, __webpack_exports__MobileBertForQuestionAnswering as MobileBertForQuestionAnswering, __webpack_exports__MobileBertForSequenceClassification as MobileBertForSequenceClassification, __webpack_exports__MobileBertModel as MobileBertModel, __webpack_exports__MobileBertPreTrainedModel as MobileBertPreTrainedModel, __webpack_exports__MobileBertTokenizer as MobileBertTokenizer, __webpack_exports__MobileLLMForCausalLM as MobileLLMForCausalLM, __webpack_exports__MobileLLMModel as MobileLLMModel, __webpack_exports__MobileLLMPreTrainedModel as MobileLLMPreTrainedModel, __webpack_exports__MobileNetV1FeatureExtractor as MobileNetV1FeatureExtractor, __webpack_exports__MobileNetV1ForImageClassification as MobileNetV1ForImageClassification, __webpack_exports__MobileNetV1ImageProcessor as MobileNetV1ImageProcessor, __webpack_exports__MobileNetV1Model as MobileNetV1Model, __webpack_exports__MobileNetV1PreTrainedModel as MobileNetV1PreTrainedModel, __webpack_exports__MobileNetV2FeatureExtractor as MobileNetV2FeatureExtractor, __webpack_exports__MobileNetV2ForImageClassification as MobileNetV2ForImageClassification, __webpack_exports__MobileNetV2ImageProcessor as MobileNetV2ImageProcessor, __webpack_exports__MobileNetV2Model as MobileNetV2Model, __webpack_exports__MobileNetV2PreTrainedModel as MobileNetV2PreTrainedModel, __webpack_exports__MobileNetV3FeatureExtractor as MobileNetV3FeatureExtractor, __webpack_exports__MobileNetV3ForImageClassification as MobileNetV3ForImageClassification, __webpack_exports__MobileNetV3ImageProcessor as MobileNetV3ImageProcessor, __webpack_exports__MobileNetV3Model as MobileNetV3Model, __webpack_exports__MobileNetV3PreTrainedModel as MobileNetV3PreTrainedModel, __webpack_exports__MobileNetV4FeatureExtractor as MobileNetV4FeatureExtractor, __webpack_exports__MobileNetV4ForImageClassification as MobileNetV4ForImageClassification, __webpack_exports__MobileNetV4ImageProcessor as MobileNetV4ImageProcessor, __webpack_exports__MobileNetV4Model as MobileNetV4Model, __webpack_exports__MobileNetV4PreTrainedModel as MobileNetV4PreTrainedModel, __webpack_exports__MobileViTFeatureExtractor as MobileViTFeatureExtractor, __webpack_exports__MobileViTForImageClassification as MobileViTForImageClassification, __webpack_exports__MobileViTImageProcessor as MobileViTImageProcessor, __webpack_exports__MobileViTModel as MobileViTModel, __webpack_exports__MobileViTPreTrainedModel as MobileViTPreTrainedModel, __webpack_exports__MobileViTV2ForImageClassification as MobileViTV2ForImageClassification, __webpack_exports__MobileViTV2Model as MobileViTV2Model, __webpack_exports__MobileViTV2PreTrainedModel as MobileViTV2PreTrainedModel, __webpack_exports__ModelOutput as ModelOutput, __webpack_exports__ModernBertForMaskedLM as ModernBertForMaskedLM, __webpack_exports__ModernBertForSequenceClassification as ModernBertForSequenceClassification, __webpack_exports__ModernBertForTokenClassification as ModernBertForTokenClassification, __webpack_exports__ModernBertModel as ModernBertModel, __webpack_exports__ModernBertPreTrainedModel as ModernBertPreTrainedModel, __webpack_exports__Moondream1ForConditionalGeneration as Moondream1ForConditionalGeneration, __webpack_exports__MoonshineFeatureExtractor as MoonshineFeatureExtractor, __webpack_exports__MoonshineForConditionalGeneration as MoonshineForConditionalGeneration, __webpack_exports__MoonshineModel as MoonshineModel, __webpack_exports__MoonshinePreTrainedModel as MoonshinePreTrainedModel, __webpack_exports__MoonshineProcessor as MoonshineProcessor, __webpack_exports__MptForCausalLM as MptForCausalLM, __webpack_exports__MptModel as MptModel, __webpack_exports__MptPreTrainedModel as MptPreTrainedModel, __webpack_exports__MultiModalityCausalLM as MultiModalityCausalLM, __webpack_exports__MultiModalityPreTrainedModel as MultiModalityPreTrainedModel, __webpack_exports__MusicgenForCausalLM as MusicgenForCausalLM, __webpack_exports__MusicgenForConditionalGeneration as MusicgenForConditionalGeneration, __webpack_exports__MusicgenModel as MusicgenModel, __webpack_exports__MusicgenPreTrainedModel as MusicgenPreTrainedModel, __webpack_exports__NllbTokenizer as NllbTokenizer, __webpack_exports__NoBadWordsLogitsProcessor as NoBadWordsLogitsProcessor, __webpack_exports__NoRepeatNGramLogitsProcessor as NoRepeatNGramLogitsProcessor, __webpack_exports__NomicBertModel as NomicBertModel, __webpack_exports__NomicBertPreTrainedModel as NomicBertPreTrainedModel, __webpack_exports__NougatImageProcessor as NougatImageProcessor, __webpack_exports__NougatTokenizer as NougatTokenizer, __webpack_exports__OPTForCausalLM as OPTForCausalLM, __webpack_exports__OPTModel as OPTModel, __webpack_exports__OPTPreTrainedModel as OPTPreTrainedModel, __webpack_exports__ObjectDetectionPipeline as ObjectDetectionPipeline, __webpack_exports__Olmo2ForCausalLM as Olmo2ForCausalLM, __webpack_exports__Olmo2Model as Olmo2Model, __webpack_exports__Olmo2PreTrainedModel as Olmo2PreTrainedModel, __webpack_exports__OlmoForCausalLM as OlmoForCausalLM, __webpack_exports__OlmoModel as OlmoModel, __webpack_exports__OlmoPreTrainedModel as OlmoPreTrainedModel, __webpack_exports__OpenELMForCausalLM as OpenELMForCausalLM, __webpack_exports__OpenELMModel as OpenELMModel, __webpack_exports__OpenELMPreTrainedModel as OpenELMPreTrainedModel, __webpack_exports__OwlViTFeatureExtractor as OwlViTFeatureExtractor, __webpack_exports__OwlViTForObjectDetection as OwlViTForObjectDetection, __webpack_exports__OwlViTImageProcessor as OwlViTImageProcessor, __webpack_exports__OwlViTModel as OwlViTModel, __webpack_exports__OwlViTPreTrainedModel as OwlViTPreTrainedModel, __webpack_exports__OwlViTProcessor as OwlViTProcessor, __webpack_exports__Owlv2ForObjectDetection as Owlv2ForObjectDetection, __webpack_exports__Owlv2ImageProcessor as Owlv2ImageProcessor, __webpack_exports__Owlv2Model as Owlv2Model, __webpack_exports__Owlv2PreTrainedModel as Owlv2PreTrainedModel, __webpack_exports__PaliGemmaForConditionalGeneration as PaliGemmaForConditionalGeneration, __webpack_exports__PaliGemmaPreTrainedModel as PaliGemmaPreTrainedModel, __webpack_exports__PaliGemmaProcessor as PaliGemmaProcessor, __webpack_exports__PatchTSMixerForPrediction as PatchTSMixerForPrediction, __webpack_exports__PatchTSMixerModel as PatchTSMixerModel, __webpack_exports__PatchTSMixerPreTrainedModel as PatchTSMixerPreTrainedModel, __webpack_exports__PatchTSTForPrediction as PatchTSTForPrediction, __webpack_exports__PatchTSTModel as PatchTSTModel, __webpack_exports__PatchTSTPreTrainedModel as PatchTSTPreTrainedModel, __webpack_exports__Phi3ForCausalLM as Phi3ForCausalLM, __webpack_exports__Phi3Model as Phi3Model, __webpack_exports__Phi3PreTrainedModel as Phi3PreTrainedModel, __webpack_exports__Phi3VForCausalLM as Phi3VForCausalLM, __webpack_exports__Phi3VImageProcessor as Phi3VImageProcessor, __webpack_exports__Phi3VPreTrainedModel as Phi3VPreTrainedModel, __webpack_exports__Phi3VProcessor as Phi3VProcessor, __webpack_exports__PhiForCausalLM as PhiForCausalLM, __webpack_exports__PhiModel as PhiModel, __webpack_exports__PhiPreTrainedModel as PhiPreTrainedModel, __webpack_exports__Pipeline as Pipeline, __webpack_exports__PreTrainedModel as PreTrainedModel, __webpack_exports__PreTrainedTokenizer as PreTrainedTokenizer, __webpack_exports__PretrainedConfig as PretrainedConfig, __webpack_exports__PretrainedMixin as PretrainedMixin, __webpack_exports__Processor as Processor, __webpack_exports__PvtForImageClassification as PvtForImageClassification, __webpack_exports__PvtImageProcessor as PvtImageProcessor, __webpack_exports__PvtModel as PvtModel, __webpack_exports__PvtPreTrainedModel as PvtPreTrainedModel, __webpack_exports__PyAnnoteFeatureExtractor as PyAnnoteFeatureExtractor, __webpack_exports__PyAnnoteForAudioFrameClassification as PyAnnoteForAudioFrameClassification, __webpack_exports__PyAnnoteModel as PyAnnoteModel, __webpack_exports__PyAnnotePreTrainedModel as PyAnnotePreTrainedModel, __webpack_exports__PyAnnoteProcessor as PyAnnoteProcessor, __webpack_exports__QuestionAnsweringModelOutput as QuestionAnsweringModelOutput, __webpack_exports__QuestionAnsweringPipeline as QuestionAnsweringPipeline, __webpack_exports__Qwen2ForCausalLM as Qwen2ForCausalLM, __webpack_exports__Qwen2Model as Qwen2Model, __webpack_exports__Qwen2PreTrainedModel as Qwen2PreTrainedModel, __webpack_exports__Qwen2Tokenizer as Qwen2Tokenizer, __webpack_exports__Qwen2VLForConditionalGeneration as Qwen2VLForConditionalGeneration, __webpack_exports__Qwen2VLImageProcessor as Qwen2VLImageProcessor, __webpack_exports__Qwen2VLPreTrainedModel as Qwen2VLPreTrainedModel, __webpack_exports__Qwen2VLProcessor as Qwen2VLProcessor, __webpack_exports__RTDetrForObjectDetection as RTDetrForObjectDetection, __webpack_exports__RTDetrImageProcessor as RTDetrImageProcessor, __webpack_exports__RTDetrModel as RTDetrModel, __webpack_exports__RTDetrObjectDetectionOutput as RTDetrObjectDetectionOutput, __webpack_exports__RTDetrPreTrainedModel as RTDetrPreTrainedModel, __webpack_exports__RawAudio as RawAudio, __webpack_exports__RawImage as RawImage, __webpack_exports__RepetitionPenaltyLogitsProcessor as RepetitionPenaltyLogitsProcessor, __webpack_exports__ResNetForImageClassification as ResNetForImageClassification, __webpack_exports__ResNetModel as ResNetModel, __webpack_exports__ResNetPreTrainedModel as ResNetPreTrainedModel, __webpack_exports__RoFormerForMaskedLM as RoFormerForMaskedLM, __webpack_exports__RoFormerForQuestionAnswering as RoFormerForQuestionAnswering, __webpack_exports__RoFormerForSequenceClassification as RoFormerForSequenceClassification, __webpack_exports__RoFormerForTokenClassification as RoFormerForTokenClassification, __webpack_exports__RoFormerModel as RoFormerModel, __webpack_exports__RoFormerPreTrainedModel as RoFormerPreTrainedModel, __webpack_exports__RoFormerTokenizer as RoFormerTokenizer, __webpack_exports__RobertaForMaskedLM as RobertaForMaskedLM, __webpack_exports__RobertaForQuestionAnswering as RobertaForQuestionAnswering, __webpack_exports__RobertaForSequenceClassification as RobertaForSequenceClassification, __webpack_exports__RobertaForTokenClassification as RobertaForTokenClassification, __webpack_exports__RobertaModel as RobertaModel, __webpack_exports__RobertaPreTrainedModel as RobertaPreTrainedModel, __webpack_exports__RobertaTokenizer as RobertaTokenizer, __webpack_exports__SamImageProcessor as SamImageProcessor, __webpack_exports__SamImageSegmentationOutput as SamImageSegmentationOutput, __webpack_exports__SamModel as SamModel, __webpack_exports__SamPreTrainedModel as SamPreTrainedModel, __webpack_exports__SamProcessor as SamProcessor, __webpack_exports__SapiensForDepthEstimation as SapiensForDepthEstimation, __webpack_exports__SapiensForNormalEstimation as SapiensForNormalEstimation, __webpack_exports__SapiensForSemanticSegmentation as SapiensForSemanticSegmentation, __webpack_exports__SapiensPreTrainedModel as SapiensPreTrainedModel, __webpack_exports__SeamlessM4TFeatureExtractor as SeamlessM4TFeatureExtractor, __webpack_exports__SegformerFeatureExtractor as SegformerFeatureExtractor, __webpack_exports__SegformerForImageClassification as SegformerForImageClassification, __webpack_exports__SegformerForSemanticSegmentation as SegformerForSemanticSegmentation, __webpack_exports__SegformerImageProcessor as SegformerImageProcessor, __webpack_exports__SegformerModel as SegformerModel, __webpack_exports__SegformerPreTrainedModel as SegformerPreTrainedModel, __webpack_exports__Seq2SeqLMOutput as Seq2SeqLMOutput, __webpack_exports__SequenceClassifierOutput as SequenceClassifierOutput, __webpack_exports__SiglipImageProcessor as SiglipImageProcessor, __webpack_exports__SiglipModel as SiglipModel, __webpack_exports__SiglipPreTrainedModel as SiglipPreTrainedModel, __webpack_exports__SiglipTextModel as SiglipTextModel, __webpack_exports__SiglipTokenizer as SiglipTokenizer, __webpack_exports__SiglipVisionModel as SiglipVisionModel, __webpack_exports__SpeechT5FeatureExtractor as SpeechT5FeatureExtractor, __webpack_exports__SpeechT5ForSpeechToText as SpeechT5ForSpeechToText, __webpack_exports__SpeechT5ForTextToSpeech as SpeechT5ForTextToSpeech, __webpack_exports__SpeechT5HifiGan as SpeechT5HifiGan, __webpack_exports__SpeechT5Model as SpeechT5Model, __webpack_exports__SpeechT5PreTrainedModel as SpeechT5PreTrainedModel, __webpack_exports__SpeechT5Processor as SpeechT5Processor, __webpack_exports__SpeechT5Tokenizer as SpeechT5Tokenizer, __webpack_exports__SqueezeBertForMaskedLM as SqueezeBertForMaskedLM, __webpack_exports__SqueezeBertForQuestionAnswering as SqueezeBertForQuestionAnswering, __webpack_exports__SqueezeBertForSequenceClassification as SqueezeBertForSequenceClassification, __webpack_exports__SqueezeBertModel as SqueezeBertModel, __webpack_exports__SqueezeBertPreTrainedModel as SqueezeBertPreTrainedModel, __webpack_exports__SqueezeBertTokenizer as SqueezeBertTokenizer, __webpack_exports__StableLmForCausalLM as StableLmForCausalLM, __webpack_exports__StableLmModel as StableLmModel, __webpack_exports__StableLmPreTrainedModel as StableLmPreTrainedModel, __webpack_exports__Starcoder2ForCausalLM as Starcoder2ForCausalLM, __webpack_exports__Starcoder2Model as Starcoder2Model, __webpack_exports__Starcoder2PreTrainedModel as Starcoder2PreTrainedModel, __webpack_exports__StoppingCriteria as StoppingCriteria, __webpack_exports__StoppingCriteriaList as StoppingCriteriaList, __webpack_exports__StyleTextToSpeech2Model as StyleTextToSpeech2Model, __webpack_exports__StyleTextToSpeech2PreTrainedModel as StyleTextToSpeech2PreTrainedModel, __webpack_exports__SummarizationPipeline as SummarizationPipeline, __webpack_exports__SuppressTokensAtBeginLogitsProcessor as SuppressTokensAtBeginLogitsProcessor, __webpack_exports__Swin2SRForImageSuperResolution as Swin2SRForImageSuperResolution, __webpack_exports__Swin2SRImageProcessor as Swin2SRImageProcessor, __webpack_exports__Swin2SRModel as Swin2SRModel, __webpack_exports__Swin2SRPreTrainedModel as Swin2SRPreTrainedModel, __webpack_exports__SwinForImageClassification as SwinForImageClassification, __webpack_exports__SwinModel as SwinModel, __webpack_exports__SwinPreTrainedModel as SwinPreTrainedModel, __webpack_exports__T5ForConditionalGeneration as T5ForConditionalGeneration, __webpack_exports__T5Model as T5Model, __webpack_exports__T5PreTrainedModel as T5PreTrainedModel, __webpack_exports__T5Tokenizer as T5Tokenizer, __webpack_exports__TableTransformerForObjectDetection as TableTransformerForObjectDetection, __webpack_exports__TableTransformerModel as TableTransformerModel, __webpack_exports__TableTransformerObjectDetectionOutput as TableTransformerObjectDetectionOutput, __webpack_exports__TableTransformerPreTrainedModel as TableTransformerPreTrainedModel, __webpack_exports__TemperatureLogitsWarper as TemperatureLogitsWarper, __webpack_exports__Tensor as Tensor, __webpack_exports__Text2TextGenerationPipeline as Text2TextGenerationPipeline, __webpack_exports__TextClassificationPipeline as TextClassificationPipeline, __webpack_exports__TextGenerationPipeline as TextGenerationPipeline, __webpack_exports__TextStreamer as TextStreamer, __webpack_exports__TextToAudioPipeline as TextToAudioPipeline, __webpack_exports__TokenClassificationPipeline as TokenClassificationPipeline, __webpack_exports__TokenClassifierOutput as TokenClassifierOutput, __webpack_exports__TokenizerModel as TokenizerModel, __webpack_exports__TopKLogitsWarper as TopKLogitsWarper, __webpack_exports__TopPLogitsWarper as TopPLogitsWarper, __webpack_exports__TrOCRForCausalLM as TrOCRForCausalLM, __webpack_exports__TrOCRPreTrainedModel as TrOCRPreTrainedModel, __webpack_exports__TranslationPipeline as TranslationPipeline, __webpack_exports__UniSpeechForCTC as UniSpeechForCTC, __webpack_exports__UniSpeechForSequenceClassification as UniSpeechForSequenceClassification, __webpack_exports__UniSpeechModel as UniSpeechModel, __webpack_exports__UniSpeechPreTrainedModel as UniSpeechPreTrainedModel, __webpack_exports__UniSpeechSatForAudioFrameClassification as UniSpeechSatForAudioFrameClassification, __webpack_exports__UniSpeechSatForCTC as UniSpeechSatForCTC, __webpack_exports__UniSpeechSatForSequenceClassification as UniSpeechSatForSequenceClassification, __webpack_exports__UniSpeechSatModel as UniSpeechSatModel, __webpack_exports__UniSpeechSatPreTrainedModel as UniSpeechSatPreTrainedModel, __webpack_exports__VLChatProcessor as VLChatProcessor, __webpack_exports__VLMImageProcessor as VLMImageProcessor, __webpack_exports__ViTFeatureExtractor as ViTFeatureExtractor, __webpack_exports__ViTForImageClassification as ViTForImageClassification, __webpack_exports__ViTImageProcessor as ViTImageProcessor, __webpack_exports__ViTMAEModel as ViTMAEModel, __webpack_exports__ViTMAEPreTrainedModel as ViTMAEPreTrainedModel, __webpack_exports__ViTMSNForImageClassification as ViTMSNForImageClassification, __webpack_exports__ViTMSNModel as ViTMSNModel, __webpack_exports__ViTMSNPreTrainedModel as ViTMSNPreTrainedModel, __webpack_exports__ViTModel as ViTModel, __webpack_exports__ViTPreTrainedModel as ViTPreTrainedModel, __webpack_exports__VisionEncoderDecoderModel as VisionEncoderDecoderModel, __webpack_exports__VitMatteForImageMatting as VitMatteForImageMatting, __webpack_exports__VitMatteImageProcessor as VitMatteImageProcessor, __webpack_exports__VitMattePreTrainedModel as VitMattePreTrainedModel, __webpack_exports__VitPoseForPoseEstimation as VitPoseForPoseEstimation, __webpack_exports__VitPoseImageProcessor as VitPoseImageProcessor, __webpack_exports__VitPosePreTrainedModel as VitPosePreTrainedModel, __webpack_exports__VitsModel as VitsModel, __webpack_exports__VitsModelOutput as VitsModelOutput, __webpack_exports__VitsPreTrainedModel as VitsPreTrainedModel, __webpack_exports__VitsTokenizer as VitsTokenizer, __webpack_exports__Wav2Vec2BertForCTC as Wav2Vec2BertForCTC, __webpack_exports__Wav2Vec2BertForSequenceClassification as Wav2Vec2BertForSequenceClassification, __webpack_exports__Wav2Vec2BertModel as Wav2Vec2BertModel, __webpack_exports__Wav2Vec2BertPreTrainedModel as Wav2Vec2BertPreTrainedModel, __webpack_exports__Wav2Vec2CTCTokenizer as Wav2Vec2CTCTokenizer, __webpack_exports__Wav2Vec2FeatureExtractor as Wav2Vec2FeatureExtractor, __webpack_exports__Wav2Vec2ForAudioFrameClassification as Wav2Vec2ForAudioFrameClassification, __webpack_exports__Wav2Vec2ForCTC as Wav2Vec2ForCTC, __webpack_exports__Wav2Vec2ForSequenceClassification as Wav2Vec2ForSequenceClassification, __webpack_exports__Wav2Vec2Model as Wav2Vec2Model, __webpack_exports__Wav2Vec2PreTrainedModel as Wav2Vec2PreTrainedModel, __webpack_exports__Wav2Vec2ProcessorWithLM as Wav2Vec2ProcessorWithLM, __webpack_exports__WavLMForAudioFrameClassification as WavLMForAudioFrameClassification, __webpack_exports__WavLMForCTC as WavLMForCTC, __webpack_exports__WavLMForSequenceClassification as WavLMForSequenceClassification, __webpack_exports__WavLMForXVector as WavLMForXVector, __webpack_exports__WavLMModel as WavLMModel, __webpack_exports__WavLMPreTrainedModel as WavLMPreTrainedModel, __webpack_exports__WeSpeakerFeatureExtractor as WeSpeakerFeatureExtractor, __webpack_exports__WeSpeakerResNetModel as WeSpeakerResNetModel, __webpack_exports__WeSpeakerResNetPreTrainedModel as WeSpeakerResNetPreTrainedModel, __webpack_exports__WhisperFeatureExtractor as WhisperFeatureExtractor, __webpack_exports__WhisperForConditionalGeneration as WhisperForConditionalGeneration, __webpack_exports__WhisperModel as WhisperModel, __webpack_exports__WhisperPreTrainedModel as WhisperPreTrainedModel, __webpack_exports__WhisperProcessor as WhisperProcessor, __webpack_exports__WhisperTextStreamer as WhisperTextStreamer, __webpack_exports__WhisperTimeStampLogitsProcessor as WhisperTimeStampLogitsProcessor, __webpack_exports__WhisperTokenizer as WhisperTokenizer, __webpack_exports__XLMForQuestionAnswering as XLMForQuestionAnswering, __webpack_exports__XLMForSequenceClassification as XLMForSequenceClassification, __webpack_exports__XLMForTokenClassification as XLMForTokenClassification, __webpack_exports__XLMModel as XLMModel, __webpack_exports__XLMPreTrainedModel as XLMPreTrainedModel, __webpack_exports__XLMRobertaForMaskedLM as XLMRobertaForMaskedLM, __webpack_exports__XLMRobertaForQuestionAnswering as XLMRobertaForQuestionAnswering, __webpack_exports__XLMRobertaForSequenceClassification as XLMRobertaForSequenceClassification, __webpack_exports__XLMRobertaForTokenClassification as XLMRobertaForTokenClassification, __webpack_exports__XLMRobertaModel as XLMRobertaModel, __webpack_exports__XLMRobertaPreTrainedModel as XLMRobertaPreTrainedModel, __webpack_exports__XLMRobertaTokenizer as XLMRobertaTokenizer, __webpack_exports__XLMTokenizer as XLMTokenizer, __webpack_exports__XLMWithLMHeadModel as XLMWithLMHeadModel, __webpack_exports__XVectorOutput as XVectorOutput, __webpack_exports__YolosFeatureExtractor as YolosFeatureExtractor, __webpack_exports__YolosForObjectDetection as YolosForObjectDetection, __webpack_exports__YolosImageProcessor as YolosImageProcessor, __webpack_exports__YolosModel as YolosModel, __webpack_exports__YolosObjectDetectionOutput as YolosObjectDetectionOutput, __webpack_exports__YolosPreTrainedModel as YolosPreTrainedModel, __webpack_exports__ZeroShotAudioClassificationPipeline as ZeroShotAudioClassificationPipeline, __webpack_exports__ZeroShotClassificationPipeline as ZeroShotClassificationPipeline, __webpack_exports__ZeroShotImageClassificationPipeline as ZeroShotImageClassificationPipeline, __webpack_exports__ZeroShotObjectDetectionPipeline as ZeroShotObjectDetectionPipeline, __webpack_exports__bankers_round as bankers_round, __webpack_exports__cat as cat, __webpack_exports__cos_sim as cos_sim, __webpack_exports__dot as dot, __webpack_exports__dynamic_time_warping as dynamic_time_warping, __webpack_exports__env as env, __webpack_exports__full as full, __webpack_exports__full_like as full_like, __webpack_exports__getKeyValueShapes as getKeyValueShapes, __webpack_exports__hamming as hamming, __webpack_exports__hanning as hanning, __webpack_exports__interpolate as interpolate, __webpack_exports__interpolate_4d as interpolate_4d, __webpack_exports__interpolate_data as interpolate_data, __webpack_exports__is_chinese_char as is_chinese_char, __webpack_exports__layer_norm as layer_norm, __webpack_exports__load_image as load_image, __webpack_exports__log_softmax as log_softmax, __webpack_exports__magnitude as magnitude, __webpack_exports__matmul as matmul, __webpack_exports__max as max, __webpack_exports__mean as mean, __webpack_exports__mean_pooling as mean_pooling, __webpack_exports__medianFilter as medianFilter, __webpack_exports__mel_filter_bank as mel_filter_bank, __webpack_exports__min as min, __webpack_exports__ones as ones, __webpack_exports__ones_like as ones_like, __webpack_exports__permute as permute, __webpack_exports__permute_data as permute_data, __webpack_exports__pipeline as pipeline, __webpack_exports__quantize_embeddings as quantize_embeddings, __webpack_exports__rand as rand, __webpack_exports__read_audio as read_audio, __webpack_exports__rfft as rfft, __webpack_exports__round as round, __webpack_exports__slice as slice, __webpack_exports__softmax as softmax, __webpack_exports__spectrogram as spectrogram, __webpack_exports__stack as stack, __webpack_exports__std_mean as std_mean, __webpack_exports__topk as topk, __webpack_exports__window_function as window_function, __webpack_exports__zeros as zeros, __webpack_exports__zeros_like as zeros_like };
|
|
36014
36481
|
|
|
36015
36482
|
//# sourceMappingURL=transformers.mjs.map
|