@huggingface/transformers 3.2.2 → 3.2.4
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +3 -2
- package/dist/transformers.cjs +252 -113
- package/dist/transformers.cjs.map +1 -1
- package/dist/transformers.js +256 -114
- package/dist/transformers.js.map +1 -1
- package/dist/transformers.min.cjs +1 -1
- package/dist/transformers.min.cjs.map +1 -1
- package/dist/transformers.min.js +1 -1
- package/dist/transformers.min.js.map +1 -1
- package/dist/transformers.min.mjs +1 -1
- package/dist/transformers.min.mjs.map +1 -1
- package/dist/transformers.mjs +256 -114
- package/dist/transformers.mjs.map +1 -1
- package/package.json +2 -2
- package/src/base/feature_extraction_utils.js +9 -9
- package/src/base/image_processors_utils.js +11 -0
- package/src/base/processing_utils.js +13 -3
- package/src/configs.js +5 -0
- package/src/env.js +1 -1
- package/src/models/auto/feature_extraction_auto.js +0 -16
- package/src/models/auto/processing_auto.js +0 -16
- package/src/models/convnext/image_processing_convnext.js +1 -0
- package/src/models/efficientnet/image_processing_efficientnet.js +1 -0
- package/src/models/florence2/processing_florence2.js +3 -0
- package/src/models/idefics3/image_processing_idefics3.js +2 -0
- package/src/models/janus/image_processing_janus.js +1 -0
- package/src/models/mgp_str/processing_mgp_str.js +2 -0
- package/src/models/paligemma/processing_paligemma.js +1 -0
- package/src/models/phi3_v/processing_phi3_v.js +1 -1
- package/src/models/pyannote/feature_extraction_pyannote.js +1 -0
- package/src/models/qwen2_vl/processing_qwen2_vl.js +1 -0
- package/src/models/seamless_m4t/feature_extraction_seamless_m4t.js +2 -2
- package/src/models/whisper/feature_extraction_whisper.js +1 -1
- package/src/models.js +93 -36
- package/src/ops/registry.js +10 -0
- package/src/pipelines.js +34 -7
- package/src/tokenizers.js +4 -7
- package/src/utils/dtypes.js +2 -0
- package/src/utils/hub.js +1 -1
- package/src/utils/maths.js +8 -6
- package/src/utils/tensor.js +42 -10
- package/types/base/feature_extraction_utils.d.ts +7 -7
- package/types/base/image_processors_utils.d.ts.map +1 -1
- package/types/base/processing_utils.d.ts +17 -19
- package/types/base/processing_utils.d.ts.map +1 -1
- package/types/configs.d.ts.map +1 -1
- package/types/generation/parameters.d.ts +1 -1
- package/types/models/auto/feature_extraction_auto.d.ts.map +1 -1
- package/types/models/auto/image_processing_auto.d.ts.map +1 -1
- package/types/models/auto/processing_auto.d.ts.map +1 -1
- package/types/models/convnext/image_processing_convnext.d.ts.map +1 -1
- package/types/models/efficientnet/image_processing_efficientnet.d.ts.map +1 -1
- package/types/models/florence2/processing_florence2.d.ts.map +1 -1
- package/types/models/idefics3/image_processing_idefics3.d.ts.map +1 -1
- package/types/models/janus/image_processing_janus.d.ts.map +1 -1
- package/types/models/mgp_str/processing_mgp_str.d.ts.map +1 -1
- package/types/models/paligemma/processing_paligemma.d.ts.map +1 -1
- package/types/models/phi3_v/processing_phi3_v.d.ts +6 -2
- package/types/models/phi3_v/processing_phi3_v.d.ts.map +1 -1
- package/types/models/pyannote/feature_extraction_pyannote.d.ts.map +1 -1
- package/types/models/qwen2_vl/processing_qwen2_vl.d.ts.map +1 -1
- package/types/models/sapiens/image_processing_sapiens.d.ts +10 -0
- package/types/models/sapiens/image_processing_sapiens.d.ts.map +1 -0
- package/types/models/whisper/generation_whisper.d.ts +1 -1
- package/types/models/whisper/generation_whisper.d.ts.map +1 -1
- package/types/models.d.ts +48 -17
- package/types/models.d.ts.map +1 -1
- package/types/ops/registry.d.ts +1 -0
- package/types/ops/registry.d.ts.map +1 -1
- package/types/pipelines.d.ts +2 -2
- package/types/pipelines.d.ts.map +1 -1
- package/types/tokenizers.d.ts.map +1 -1
- package/types/tsconfig.tsbuildinfo +1 -0
- package/types/utils/dtypes.d.ts.map +1 -1
- package/types/utils/hub.d.ts +1 -1
- package/types/utils/hub.d.ts.map +1 -1
- package/types/utils/image.d.ts +3 -2
- package/types/utils/image.d.ts.map +1 -1
- package/types/utils/maths.d.ts +8 -6
- package/types/utils/maths.d.ts.map +1 -1
- package/types/utils/tensor.d.ts +8 -4
- package/types/utils/tensor.d.ts.map +1 -1
package/dist/transformers.mjs
CHANGED
|
@@ -4132,23 +4132,23 @@ class FeatureExtractor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_1__.Ca
|
|
|
4132
4132
|
}
|
|
4133
4133
|
|
|
4134
4134
|
/**
|
|
4135
|
-
* Instantiate one of the
|
|
4135
|
+
* Instantiate one of the feature extractor classes of the library from a pretrained model.
|
|
4136
4136
|
*
|
|
4137
|
-
* The
|
|
4138
|
-
*
|
|
4137
|
+
* The feature extractor class to instantiate is selected based on the `feature_extractor_type` property of
|
|
4138
|
+
* the config object (either passed as an argument or loaded from `pretrained_model_name_or_path` if possible)
|
|
4139
4139
|
*
|
|
4140
4140
|
* @param {string} pretrained_model_name_or_path The name or path of the pretrained model. Can be either:
|
|
4141
|
-
* - A string, the *model id* of a pretrained
|
|
4141
|
+
* - A string, the *model id* of a pretrained feature_extractor hosted inside a model repo on huggingface.co.
|
|
4142
4142
|
* Valid model ids can be located at the root-level, like `bert-base-uncased`, or namespaced under a
|
|
4143
4143
|
* user or organization name, like `dbmdz/bert-base-german-cased`.
|
|
4144
|
-
* - A path to a *directory* containing
|
|
4145
|
-
* @param {import('../utils/hub.js').PretrainedOptions} options Additional options for loading the
|
|
4144
|
+
* - A path to a *directory* containing feature_extractor files, e.g., `./my_model_directory/`.
|
|
4145
|
+
* @param {import('../utils/hub.js').PretrainedOptions} options Additional options for loading the feature_extractor.
|
|
4146
4146
|
*
|
|
4147
|
-
* @returns {Promise<FeatureExtractor>} A new instance of the
|
|
4147
|
+
* @returns {Promise<FeatureExtractor>} A new instance of the Feature Extractor class.
|
|
4148
4148
|
*/
|
|
4149
4149
|
static async from_pretrained(pretrained_model_name_or_path, options) {
|
|
4150
|
-
const
|
|
4151
|
-
return new this(
|
|
4150
|
+
const config = await (0,_utils_hub_js__WEBPACK_IMPORTED_MODULE_2__.getModelJSON)(pretrained_model_name_or_path, _utils_constants_js__WEBPACK_IMPORTED_MODULE_0__.FEATURE_EXTRACTOR_NAME, true, options);
|
|
4151
|
+
return new this(config);
|
|
4152
4152
|
}
|
|
4153
4153
|
}
|
|
4154
4154
|
|
|
@@ -4798,14 +4798,20 @@ class ImageProcessor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_0__.Call
|
|
|
4798
4798
|
this.do_thumbnail = config.do_thumbnail;
|
|
4799
4799
|
this.size = config.size ?? config.image_size;
|
|
4800
4800
|
this.do_resize = config.do_resize ?? (this.size !== undefined);
|
|
4801
|
+
// @ts-expect-error TS2339
|
|
4801
4802
|
this.size_divisibility = config.size_divisibility ?? config.size_divisor;
|
|
4802
4803
|
|
|
4803
4804
|
this.do_center_crop = config.do_center_crop;
|
|
4805
|
+
// @ts-expect-error TS2339
|
|
4804
4806
|
this.crop_size = config.crop_size;
|
|
4807
|
+
// @ts-expect-error TS2339
|
|
4805
4808
|
this.do_convert_rgb = config.do_convert_rgb ?? true;
|
|
4809
|
+
// @ts-expect-error TS2339
|
|
4806
4810
|
this.do_crop_margin = config.do_crop_margin;
|
|
4807
4811
|
|
|
4812
|
+
// @ts-expect-error TS2339
|
|
4808
4813
|
this.pad_size = config.pad_size;
|
|
4814
|
+
// @ts-expect-error TS2339
|
|
4809
4815
|
this.do_pad = config.do_pad;
|
|
4810
4816
|
|
|
4811
4817
|
if (this.do_pad && !this.pad_size && this.size && this.size.width !== undefined && this.size.height !== undefined) {
|
|
@@ -5014,6 +5020,7 @@ class ImageProcessor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_0__.Call
|
|
|
5014
5020
|
// Support both formats for backwards compatibility
|
|
5015
5021
|
else if (Number.isInteger(size)) {
|
|
5016
5022
|
shortest_edge = size;
|
|
5023
|
+
// @ts-expect-error TS2339
|
|
5017
5024
|
longest_edge = this.config.max_size ?? shortest_edge;
|
|
5018
5025
|
|
|
5019
5026
|
} else if (size !== undefined) {
|
|
@@ -5082,6 +5089,7 @@ class ImageProcessor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_0__.Call
|
|
|
5082
5089
|
} else if (size.min_pixels !== undefined && size.max_pixels !== undefined) {
|
|
5083
5090
|
// Custom resize logic for Qwen2-VL models
|
|
5084
5091
|
const { min_pixels, max_pixels } = size;
|
|
5092
|
+
// @ts-expect-error TS2339
|
|
5085
5093
|
const factor = this.config.patch_size * this.config.merge_size;
|
|
5086
5094
|
return smart_resize(srcHeight, srcWidth, factor, min_pixels, max_pixels);
|
|
5087
5095
|
} else {
|
|
@@ -5097,6 +5105,7 @@ class ImageProcessor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_0__.Call
|
|
|
5097
5105
|
async resize(image) {
|
|
5098
5106
|
const [newWidth, newHeight] = this.get_resize_output_image_size(image, this.size);
|
|
5099
5107
|
return await image.resize(newWidth, newHeight, {
|
|
5108
|
+
// @ts-expect-error TS2322
|
|
5100
5109
|
resample: this.resample,
|
|
5101
5110
|
});
|
|
5102
5111
|
}
|
|
@@ -5147,6 +5156,7 @@ class ImageProcessor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_0__.Call
|
|
|
5147
5156
|
|
|
5148
5157
|
// Resize the image using thumbnail method.
|
|
5149
5158
|
if (this.do_thumbnail) {
|
|
5159
|
+
// @ts-expect-error TS2345
|
|
5150
5160
|
image = await this.thumbnail(image, this.size, this.resample);
|
|
5151
5161
|
}
|
|
5152
5162
|
|
|
@@ -5171,6 +5181,7 @@ class ImageProcessor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_0__.Call
|
|
|
5171
5181
|
// NOTE: All pixel-level manipulation (i.e., modifying `pixelData`)
|
|
5172
5182
|
// occurs with data in the hwc format (height, width, channels),
|
|
5173
5183
|
// to emulate the behavior of the original Python code (w/ numpy).
|
|
5184
|
+
/** @type {Float32Array} */
|
|
5174
5185
|
let pixelData = Float32Array.from(image.data);
|
|
5175
5186
|
let imgDims = [image.height, image.width, image.channels];
|
|
5176
5187
|
|
|
@@ -5328,6 +5339,7 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
5328
5339
|
/**
|
|
5329
5340
|
* @typedef {Object} ProcessorProperties Additional processor-specific properties.
|
|
5330
5341
|
* @typedef {import('../utils/hub.js').PretrainedOptions & ProcessorProperties} PretrainedProcessorOptions
|
|
5342
|
+
* @typedef {import('../tokenizers.js').PreTrainedTokenizer} PreTrainedTokenizer
|
|
5331
5343
|
*/
|
|
5332
5344
|
|
|
5333
5345
|
|
|
@@ -5361,7 +5373,7 @@ class Processor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_1__.Callable
|
|
|
5361
5373
|
}
|
|
5362
5374
|
|
|
5363
5375
|
/**
|
|
5364
|
-
* @returns {
|
|
5376
|
+
* @returns {PreTrainedTokenizer|undefined} The tokenizer of the processor, if it exists.
|
|
5365
5377
|
*/
|
|
5366
5378
|
get tokenizer() {
|
|
5367
5379
|
return this.components.tokenizer;
|
|
@@ -5374,6 +5386,11 @@ class Processor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_1__.Callable
|
|
|
5374
5386
|
return this.components.feature_extractor;
|
|
5375
5387
|
}
|
|
5376
5388
|
|
|
5389
|
+
/**
|
|
5390
|
+
* @param {Parameters<PreTrainedTokenizer['apply_chat_template']>[0]} messages
|
|
5391
|
+
* @param {Parameters<PreTrainedTokenizer['apply_chat_template']>[1]} options
|
|
5392
|
+
* @returns {ReturnType<PreTrainedTokenizer['apply_chat_template']>}
|
|
5393
|
+
*/
|
|
5377
5394
|
apply_chat_template(messages, options = {}) {
|
|
5378
5395
|
if (!this.tokenizer) {
|
|
5379
5396
|
throw new Error('Unable to apply chat template without a tokenizer.');
|
|
@@ -5384,6 +5401,10 @@ class Processor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_1__.Callable
|
|
|
5384
5401
|
});
|
|
5385
5402
|
}
|
|
5386
5403
|
|
|
5404
|
+
/**
|
|
5405
|
+
* @param {Parameters<PreTrainedTokenizer['batch_decode']>} args
|
|
5406
|
+
* @returns {ReturnType<PreTrainedTokenizer['batch_decode']>}
|
|
5407
|
+
*/
|
|
5387
5408
|
batch_decode(...args) {
|
|
5388
5409
|
if (!this.tokenizer) {
|
|
5389
5410
|
throw new Error('Unable to decode without a tokenizer.');
|
|
@@ -5411,8 +5432,8 @@ class Processor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_1__.Callable
|
|
|
5411
5432
|
/**
|
|
5412
5433
|
* Instantiate one of the processor classes of the library from a pretrained model.
|
|
5413
5434
|
*
|
|
5414
|
-
* The processor class to instantiate is selected based on the `
|
|
5415
|
-
* (either passed as an argument or loaded from `pretrained_model_name_or_path` if possible)
|
|
5435
|
+
* The processor class to instantiate is selected based on the `image_processor_type` (or `feature_extractor_type`; legacy)
|
|
5436
|
+
* property of the config object (either passed as an argument or loaded from `pretrained_model_name_or_path` if possible)
|
|
5416
5437
|
*
|
|
5417
5438
|
* @param {string} pretrained_model_name_or_path The name or path of the pretrained model. Can be either:
|
|
5418
5439
|
* - A string, the *model id* of a pretrained processor hosted inside a model repo on huggingface.co.
|
|
@@ -5531,15 +5552,19 @@ function getNormalizedConfig(config) {
|
|
|
5531
5552
|
case 'florence2':
|
|
5532
5553
|
case 'llava_onevision':
|
|
5533
5554
|
case 'idefics3':
|
|
5555
|
+
// @ts-expect-error TS2339
|
|
5534
5556
|
init_normalized_config = getNormalizedConfig(config.text_config);
|
|
5535
5557
|
break;
|
|
5536
5558
|
case 'moondream1':
|
|
5559
|
+
// @ts-expect-error TS2339
|
|
5537
5560
|
init_normalized_config = getNormalizedConfig(config.phi_config);
|
|
5538
5561
|
break;
|
|
5539
5562
|
case 'musicgen':
|
|
5563
|
+
// @ts-expect-error TS2339
|
|
5540
5564
|
init_normalized_config = getNormalizedConfig(config.decoder);
|
|
5541
5565
|
break;
|
|
5542
5566
|
case 'multi_modality':
|
|
5567
|
+
// @ts-expect-error TS2339
|
|
5543
5568
|
init_normalized_config = getNormalizedConfig(config.language_config);
|
|
5544
5569
|
break;
|
|
5545
5570
|
|
|
@@ -5660,6 +5685,7 @@ function getNormalizedConfig(config) {
|
|
|
5660
5685
|
break;
|
|
5661
5686
|
|
|
5662
5687
|
case 'vision-encoder-decoder':
|
|
5688
|
+
// @ts-expect-error TS2339
|
|
5663
5689
|
const decoderConfig = getNormalizedConfig(config.decoder);
|
|
5664
5690
|
|
|
5665
5691
|
const add_encoder_pkv = 'num_decoder_layers' in decoderConfig;
|
|
@@ -5902,7 +5928,7 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
5902
5928
|
|
|
5903
5929
|
|
|
5904
5930
|
|
|
5905
|
-
const VERSION = '3.2.
|
|
5931
|
+
const VERSION = '3.2.4';
|
|
5906
5932
|
|
|
5907
5933
|
// Check if various APIs are available (depends on environment)
|
|
5908
5934
|
const IS_BROWSER_ENV = typeof window !== "undefined" && typeof window.document !== "undefined";
|
|
@@ -7972,6 +7998,9 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
7972
7998
|
/* harmony export */ Dinov2ForImageClassification: () => (/* binding */ Dinov2ForImageClassification),
|
|
7973
7999
|
/* harmony export */ Dinov2Model: () => (/* binding */ Dinov2Model),
|
|
7974
8000
|
/* harmony export */ Dinov2PreTrainedModel: () => (/* binding */ Dinov2PreTrainedModel),
|
|
8001
|
+
/* harmony export */ Dinov2WithRegistersForImageClassification: () => (/* binding */ Dinov2WithRegistersForImageClassification),
|
|
8002
|
+
/* harmony export */ Dinov2WithRegistersModel: () => (/* binding */ Dinov2WithRegistersModel),
|
|
8003
|
+
/* harmony export */ Dinov2WithRegistersPreTrainedModel: () => (/* binding */ Dinov2WithRegistersPreTrainedModel),
|
|
7975
8004
|
/* harmony export */ DistilBertForMaskedLM: () => (/* binding */ DistilBertForMaskedLM),
|
|
7976
8005
|
/* harmony export */ DistilBertForQuestionAnswering: () => (/* binding */ DistilBertForQuestionAnswering),
|
|
7977
8006
|
/* harmony export */ DistilBertForSequenceClassification: () => (/* binding */ DistilBertForSequenceClassification),
|
|
@@ -8555,8 +8584,11 @@ async function getSession(pretrained_model_name_or_path, fileName, options) {
|
|
|
8555
8584
|
} else if (session_options.externalData !== undefined) {
|
|
8556
8585
|
externalDataPromises = session_options.externalData.map(async (ext) => {
|
|
8557
8586
|
// if the external data is a string, fetch the file and replace the string with its content
|
|
8587
|
+
// @ts-expect-error TS2339
|
|
8558
8588
|
if (typeof ext.data === "string") {
|
|
8589
|
+
// @ts-expect-error TS2339
|
|
8559
8590
|
const ext_buffer = await (0,_utils_hub_js__WEBPACK_IMPORTED_MODULE_5__.getModelFile)(pretrained_model_name_or_path, ext.data, true, options);
|
|
8591
|
+
// @ts-expect-error TS2698
|
|
8560
8592
|
return { ...ext, data: ext_buffer };
|
|
8561
8593
|
}
|
|
8562
8594
|
return ext;
|
|
@@ -9804,6 +9836,7 @@ class PreTrainedModel extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_3__.Cal
|
|
|
9804
9836
|
if (this.config.model_type === 'musicgen') {
|
|
9805
9837
|
// Custom logic (TODO: move to Musicgen class)
|
|
9806
9838
|
decoder_input_ids = Array.from({
|
|
9839
|
+
// @ts-expect-error TS2339
|
|
9807
9840
|
length: batch_size * this.config.decoder.num_codebooks
|
|
9808
9841
|
}, () => [decoder_start_token_id]);
|
|
9809
9842
|
|
|
@@ -10133,11 +10166,13 @@ class PreTrainedModel extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_3__.Cal
|
|
|
10133
10166
|
async encode_image({ pixel_values }) {
|
|
10134
10167
|
// image_inputs === { pixel_values }
|
|
10135
10168
|
const features = (await sessionRun(this.sessions['vision_encoder'], { pixel_values })).image_features;
|
|
10169
|
+
// @ts-expect-error TS2339
|
|
10136
10170
|
if (!this.config.num_image_tokens) {
|
|
10137
10171
|
console.warn(
|
|
10138
10172
|
'The number of image tokens was not set in the model configuration. ' +
|
|
10139
10173
|
`Setting it to the number of features detected by the vision encoder (${features.dims[1]}).`
|
|
10140
10174
|
)
|
|
10175
|
+
// @ts-expect-error TS2339
|
|
10141
10176
|
this.config.num_image_tokens = features.dims[1];
|
|
10142
10177
|
}
|
|
10143
10178
|
return features;
|
|
@@ -11565,6 +11600,7 @@ class WhisperForConditionalGeneration extends WhisperPreTrainedModel {
|
|
|
11565
11600
|
|
|
11566
11601
|
if (generation_config.return_token_timestamps) {
|
|
11567
11602
|
outputs["token_timestamps"] = this._extract_token_timestamps(
|
|
11603
|
+
// @ts-expect-error TS2345
|
|
11568
11604
|
outputs,
|
|
11569
11605
|
generation_config.alignment_heads,
|
|
11570
11606
|
generation_config.num_frames,
|
|
@@ -11600,6 +11636,7 @@ class WhisperForConditionalGeneration extends WhisperPreTrainedModel {
|
|
|
11600
11636
|
);
|
|
11601
11637
|
}
|
|
11602
11638
|
|
|
11639
|
+
// @ts-expect-error TS2339
|
|
11603
11640
|
let median_filter_width = this.config.median_filter_width;
|
|
11604
11641
|
if (median_filter_width === undefined) {
|
|
11605
11642
|
console.warn("Model config has no `median_filter_width`, using default value of 7.")
|
|
@@ -11610,6 +11647,7 @@ class WhisperForConditionalGeneration extends WhisperPreTrainedModel {
|
|
|
11610
11647
|
const batch = generate_outputs.cross_attentions;
|
|
11611
11648
|
// Create a list with `decoder_layers` elements, each a tensor of shape
|
|
11612
11649
|
// (batch size, attention_heads, output length, input length).
|
|
11650
|
+
// @ts-expect-error TS2339
|
|
11613
11651
|
const cross_attentions = Array.from({ length: this.config.decoder_layers },
|
|
11614
11652
|
// Concatenate the cross attentions for each layer across sequence length dimension.
|
|
11615
11653
|
(_, i) => (0,_utils_tensor_js__WEBPACK_IMPORTED_MODULE_9__.cat)(batch.map(x => x[i]), 2)
|
|
@@ -11706,7 +11744,7 @@ class MoonshinePreTrainedModel extends PreTrainedModel {
|
|
|
11706
11744
|
*/
|
|
11707
11745
|
class MoonshineModel extends MoonshinePreTrainedModel { }
|
|
11708
11746
|
|
|
11709
|
-
class MoonshineForConditionalGeneration extends MoonshinePreTrainedModel { }
|
|
11747
|
+
class MoonshineForConditionalGeneration extends MoonshinePreTrainedModel { }
|
|
11710
11748
|
//////////////////////////////////////////////////
|
|
11711
11749
|
|
|
11712
11750
|
|
|
@@ -11753,6 +11791,7 @@ class LlavaForConditionalGeneration extends LlavaPreTrainedModel {
|
|
|
11753
11791
|
attention_mask,
|
|
11754
11792
|
}) {
|
|
11755
11793
|
|
|
11794
|
+
// @ts-expect-error TS2339
|
|
11756
11795
|
const image_token_index = this.config.image_token_index;
|
|
11757
11796
|
|
|
11758
11797
|
const idsList = input_ids.tolist();
|
|
@@ -12106,9 +12145,9 @@ class CLIPTextModel extends CLIPPreTrainedModel {
|
|
|
12106
12145
|
/** @type {typeof PreTrainedModel.from_pretrained} */
|
|
12107
12146
|
static async from_pretrained(pretrained_model_name_or_path, options = {}) {
|
|
12108
12147
|
return super.from_pretrained(pretrained_model_name_or_path, {
|
|
12109
|
-
// Update default model file name if not provided
|
|
12110
|
-
model_file_name: 'text_model',
|
|
12111
12148
|
...options,
|
|
12149
|
+
// Update default model file name if not provided
|
|
12150
|
+
model_file_name: options.model_file_name ?? 'text_model',
|
|
12112
12151
|
});
|
|
12113
12152
|
}
|
|
12114
12153
|
}
|
|
@@ -12143,9 +12182,9 @@ class CLIPTextModelWithProjection extends CLIPPreTrainedModel {
|
|
|
12143
12182
|
/** @type {typeof PreTrainedModel.from_pretrained} */
|
|
12144
12183
|
static async from_pretrained(pretrained_model_name_or_path, options = {}) {
|
|
12145
12184
|
return super.from_pretrained(pretrained_model_name_or_path, {
|
|
12146
|
-
// Update default model file name if not provided
|
|
12147
|
-
model_file_name: 'text_model',
|
|
12148
12185
|
...options,
|
|
12186
|
+
// Update default model file name if not provided
|
|
12187
|
+
model_file_name: options.model_file_name ?? 'text_model',
|
|
12149
12188
|
});
|
|
12150
12189
|
}
|
|
12151
12190
|
}
|
|
@@ -12157,9 +12196,9 @@ class CLIPVisionModel extends CLIPPreTrainedModel {
|
|
|
12157
12196
|
/** @type {typeof PreTrainedModel.from_pretrained} */
|
|
12158
12197
|
static async from_pretrained(pretrained_model_name_or_path, options = {}) {
|
|
12159
12198
|
return super.from_pretrained(pretrained_model_name_or_path, {
|
|
12160
|
-
// Update default model file name if not provided
|
|
12161
|
-
model_file_name: 'vision_model',
|
|
12162
12199
|
...options,
|
|
12200
|
+
// Update default model file name if not provided
|
|
12201
|
+
model_file_name: options.model_file_name ?? 'vision_model',
|
|
12163
12202
|
});
|
|
12164
12203
|
}
|
|
12165
12204
|
}
|
|
@@ -12194,9 +12233,9 @@ class CLIPVisionModelWithProjection extends CLIPPreTrainedModel {
|
|
|
12194
12233
|
/** @type {typeof PreTrainedModel.from_pretrained} */
|
|
12195
12234
|
static async from_pretrained(pretrained_model_name_or_path, options = {}) {
|
|
12196
12235
|
return super.from_pretrained(pretrained_model_name_or_path, {
|
|
12197
|
-
// Update default model file name if not provided
|
|
12198
|
-
model_file_name: 'vision_model',
|
|
12199
12236
|
...options,
|
|
12237
|
+
// Update default model file name if not provided
|
|
12238
|
+
model_file_name: options.model_file_name ?? 'vision_model',
|
|
12200
12239
|
});
|
|
12201
12240
|
}
|
|
12202
12241
|
}
|
|
@@ -12282,9 +12321,9 @@ class SiglipTextModel extends SiglipPreTrainedModel {
|
|
|
12282
12321
|
/** @type {typeof PreTrainedModel.from_pretrained} */
|
|
12283
12322
|
static async from_pretrained(pretrained_model_name_or_path, options = {}) {
|
|
12284
12323
|
return super.from_pretrained(pretrained_model_name_or_path, {
|
|
12285
|
-
// Update default model file name if not provided
|
|
12286
|
-
model_file_name: 'text_model',
|
|
12287
12324
|
...options,
|
|
12325
|
+
// Update default model file name if not provided
|
|
12326
|
+
model_file_name: options.model_file_name ?? 'text_model',
|
|
12288
12327
|
});
|
|
12289
12328
|
}
|
|
12290
12329
|
}
|
|
@@ -12319,9 +12358,9 @@ class SiglipVisionModel extends CLIPPreTrainedModel {
|
|
|
12319
12358
|
/** @type {typeof PreTrainedModel.from_pretrained} */
|
|
12320
12359
|
static async from_pretrained(pretrained_model_name_or_path, options = {}) {
|
|
12321
12360
|
return super.from_pretrained(pretrained_model_name_or_path, {
|
|
12322
|
-
// Update default model file name if not provided
|
|
12323
|
-
model_file_name: 'vision_model',
|
|
12324
12361
|
...options,
|
|
12362
|
+
// Update default model file name if not provided
|
|
12363
|
+
model_file_name: options.model_file_name ?? 'vision_model',
|
|
12325
12364
|
});
|
|
12326
12365
|
}
|
|
12327
12366
|
}
|
|
@@ -12378,9 +12417,9 @@ class JinaCLIPTextModel extends JinaCLIPPreTrainedModel {
|
|
|
12378
12417
|
/** @type {typeof PreTrainedModel.from_pretrained} */
|
|
12379
12418
|
static async from_pretrained(pretrained_model_name_or_path, options = {}) {
|
|
12380
12419
|
return super.from_pretrained(pretrained_model_name_or_path, {
|
|
12381
|
-
// Update default model file name if not provided
|
|
12382
|
-
model_file_name: 'text_model',
|
|
12383
12420
|
...options,
|
|
12421
|
+
// Update default model file name if not provided
|
|
12422
|
+
model_file_name: options.model_file_name ?? 'text_model',
|
|
12384
12423
|
});
|
|
12385
12424
|
}
|
|
12386
12425
|
}
|
|
@@ -12389,9 +12428,9 @@ class JinaCLIPVisionModel extends JinaCLIPPreTrainedModel {
|
|
|
12389
12428
|
/** @type {typeof PreTrainedModel.from_pretrained} */
|
|
12390
12429
|
static async from_pretrained(pretrained_model_name_or_path, options = {}) {
|
|
12391
12430
|
return super.from_pretrained(pretrained_model_name_or_path, {
|
|
12392
|
-
// Update default model file name if not provided
|
|
12393
|
-
model_file_name: 'vision_model',
|
|
12394
12431
|
...options,
|
|
12432
|
+
// Update default model file name if not provided
|
|
12433
|
+
model_file_name: options.model_file_name ?? 'vision_model',
|
|
12395
12434
|
});
|
|
12396
12435
|
}
|
|
12397
12436
|
}
|
|
@@ -12738,6 +12777,7 @@ class Qwen2VLForConditionalGeneration extends Qwen2VLPreTrainedModel {
|
|
|
12738
12777
|
const image_nums = vision_tokens.filter(x => x == image_token_id).length;
|
|
12739
12778
|
const video_nums = vision_tokens.filter(x => x == video_token_id).length;
|
|
12740
12779
|
|
|
12780
|
+
/** @type {number[][]} */
|
|
12741
12781
|
let llm_pos_ids_list = [];
|
|
12742
12782
|
let st = 0;
|
|
12743
12783
|
let remain_images = image_nums;
|
|
@@ -12807,6 +12847,7 @@ class Qwen2VLForConditionalGeneration extends Qwen2VLPreTrainedModel {
|
|
|
12807
12847
|
// NOTE: Each item in llm_pos_ids_list is an array of shape (3, text_len),
|
|
12808
12848
|
// meaning to perform concatenation along dim=1, we can do the following:
|
|
12809
12849
|
const num_items = llm_pos_ids_list.reduce((acc, x) => acc + x.length, 0);
|
|
12850
|
+
/** @type {number[]} */
|
|
12810
12851
|
const llm_positions = new Array(num_items);
|
|
12811
12852
|
let index = 0;
|
|
12812
12853
|
for (let x = 0; x < 3; ++x) {
|
|
@@ -12847,9 +12888,10 @@ class Qwen2VLForConditionalGeneration extends Qwen2VLPreTrainedModel {
|
|
|
12847
12888
|
{ length: 3 * data.length },
|
|
12848
12889
|
(_, i) => data[i % data.length]
|
|
12849
12890
|
);
|
|
12891
|
+
/** @type {bigint[]} */
|
|
12850
12892
|
const mrope_position_deltas = Array.from(
|
|
12851
12893
|
{ length: dims[0] },
|
|
12852
|
-
(_, i) => (0,_utils_maths_js__WEBPACK_IMPORTED_MODULE_11__.max)(data.subarray(dims[1] * i, dims[1] * (i + 1)))[0] +
|
|
12894
|
+
(_, i) => (0,_utils_maths_js__WEBPACK_IMPORTED_MODULE_11__.max)(data.subarray(dims[1] * i, dims[1] * (i + 1)))[0] + 1n + BigInt(dims[1])
|
|
12853
12895
|
);
|
|
12854
12896
|
|
|
12855
12897
|
return [
|
|
@@ -13420,7 +13462,7 @@ class DPTModel extends DPTPreTrainedModel { }
|
|
|
13420
13462
|
*
|
|
13421
13463
|
* **Example:** Depth estimation w/ `Xenova/dpt-hybrid-midas`.
|
|
13422
13464
|
* ```javascript
|
|
13423
|
-
* import { DPTForDepthEstimation, AutoProcessor, RawImage,
|
|
13465
|
+
* import { DPTForDepthEstimation, AutoProcessor, RawImage, interpolate_4d } from '@huggingface/transformers';
|
|
13424
13466
|
*
|
|
13425
13467
|
* // Load model and processor
|
|
13426
13468
|
* const model_id = 'Xenova/dpt-hybrid-midas';
|
|
@@ -13429,7 +13471,7 @@ class DPTModel extends DPTPreTrainedModel { }
|
|
|
13429
13471
|
*
|
|
13430
13472
|
* // Load image from URL
|
|
13431
13473
|
* const url = 'http://images.cocodataset.org/val2017/000000039769.jpg';
|
|
13432
|
-
* const image = await RawImage.
|
|
13474
|
+
* const image = await RawImage.read(url);
|
|
13433
13475
|
*
|
|
13434
13476
|
* // Prepare image for the model
|
|
13435
13477
|
* const inputs = await processor(image);
|
|
@@ -13438,10 +13480,15 @@ class DPTModel extends DPTPreTrainedModel { }
|
|
|
13438
13480
|
* const { predicted_depth } = await model(inputs);
|
|
13439
13481
|
*
|
|
13440
13482
|
* // Interpolate to original size
|
|
13441
|
-
* const prediction =
|
|
13483
|
+
* const prediction = (await interpolate_4d(predicted_depth.unsqueeze(1), {
|
|
13484
|
+
* size: image.size.reverse(),
|
|
13485
|
+
* mode: 'bilinear',
|
|
13486
|
+
* })).squeeze(1);
|
|
13442
13487
|
*
|
|
13443
13488
|
* // Visualize the prediction
|
|
13444
|
-
* const
|
|
13489
|
+
* const min = prediction.min().item();
|
|
13490
|
+
* const max = prediction.max().item();
|
|
13491
|
+
* const formatted = prediction.sub_(min).div_(max - min).mul_(255).to('uint8');
|
|
13445
13492
|
* const depth = RawImage.fromTensor(formatted);
|
|
13446
13493
|
* // RawImage {
|
|
13447
13494
|
* // data: Uint8Array(307200) [ 85, 85, 84, ... ],
|
|
@@ -13491,11 +13538,7 @@ class GLPNPreTrainedModel extends PreTrainedModel { }
|
|
|
13491
13538
|
class GLPNModel extends GLPNPreTrainedModel { }
|
|
13492
13539
|
|
|
13493
13540
|
/**
|
|
13494
|
-
*
|
|
13495
|
-
*
|
|
13496
|
-
* **Example:** Depth estimation w/ `Xenova/glpn-kitti`.
|
|
13497
|
-
* ```javascript
|
|
13498
|
-
* import { GLPNForDepthEstimation, AutoProcessor, RawImage, interpolate, max } from '@huggingface/transformers';
|
|
13541
|
+
* import { GLPNForDepthEstimation, AutoProcessor, RawImage, interpolate_4d } from '@huggingface/transformers';
|
|
13499
13542
|
*
|
|
13500
13543
|
* // Load model and processor
|
|
13501
13544
|
* const model_id = 'Xenova/glpn-kitti';
|
|
@@ -13504,7 +13547,7 @@ class GLPNModel extends GLPNPreTrainedModel { }
|
|
|
13504
13547
|
*
|
|
13505
13548
|
* // Load image from URL
|
|
13506
13549
|
* const url = 'http://images.cocodataset.org/val2017/000000039769.jpg';
|
|
13507
|
-
* const image = await RawImage.
|
|
13550
|
+
* const image = await RawImage.read(url);
|
|
13508
13551
|
*
|
|
13509
13552
|
* // Prepare image for the model
|
|
13510
13553
|
* const inputs = await processor(image);
|
|
@@ -13513,13 +13556,18 @@ class GLPNModel extends GLPNPreTrainedModel { }
|
|
|
13513
13556
|
* const { predicted_depth } = await model(inputs);
|
|
13514
13557
|
*
|
|
13515
13558
|
* // Interpolate to original size
|
|
13516
|
-
* const prediction =
|
|
13559
|
+
* const prediction = (await interpolate_4d(predicted_depth.unsqueeze(1), {
|
|
13560
|
+
* size: image.size.reverse(),
|
|
13561
|
+
* mode: 'bilinear',
|
|
13562
|
+
* })).squeeze(1);
|
|
13517
13563
|
*
|
|
13518
13564
|
* // Visualize the prediction
|
|
13519
|
-
* const
|
|
13565
|
+
* const min = prediction.min().item();
|
|
13566
|
+
* const max = prediction.max().item();
|
|
13567
|
+
* const formatted = prediction.sub_(min).div_(max - min).mul_(255).to('uint8');
|
|
13520
13568
|
* const depth = RawImage.fromTensor(formatted);
|
|
13521
13569
|
* // RawImage {
|
|
13522
|
-
* // data: Uint8Array(307200) [
|
|
13570
|
+
* // data: Uint8Array(307200) [ 85, 85, 84, ... ],
|
|
13523
13571
|
* // width: 640,
|
|
13524
13572
|
* // height: 480,
|
|
13525
13573
|
* // channels: 1
|
|
@@ -13674,6 +13722,26 @@ class Dinov2ForImageClassification extends Dinov2PreTrainedModel {
|
|
|
13674
13722
|
}
|
|
13675
13723
|
//////////////////////////////////////////////////
|
|
13676
13724
|
|
|
13725
|
+
//////////////////////////////////////////////////
|
|
13726
|
+
class Dinov2WithRegistersPreTrainedModel extends PreTrainedModel { }
|
|
13727
|
+
|
|
13728
|
+
/**
|
|
13729
|
+
* The bare Dinov2WithRegisters Model transformer outputting raw hidden-states without any specific head on top.
|
|
13730
|
+
*/
|
|
13731
|
+
class Dinov2WithRegistersModel extends Dinov2WithRegistersPreTrainedModel { }
|
|
13732
|
+
|
|
13733
|
+
/**
|
|
13734
|
+
* Dinov2WithRegisters Model transformer with an image classification head on top (a linear layer on top of the final hidden state of the [CLS] token) e.g. for ImageNet.
|
|
13735
|
+
*/
|
|
13736
|
+
class Dinov2WithRegistersForImageClassification extends Dinov2WithRegistersPreTrainedModel {
|
|
13737
|
+
/**
|
|
13738
|
+
* @param {any} model_inputs
|
|
13739
|
+
*/
|
|
13740
|
+
async _call(model_inputs) {
|
|
13741
|
+
return new SequenceClassifierOutput(await super._call(model_inputs));
|
|
13742
|
+
}
|
|
13743
|
+
}
|
|
13744
|
+
//////////////////////////////////////////////////
|
|
13677
13745
|
|
|
13678
13746
|
//////////////////////////////////////////////////
|
|
13679
13747
|
class YolosPreTrainedModel extends PreTrainedModel { }
|
|
@@ -14466,10 +14534,12 @@ class SpeechT5ForTextToSpeech extends SpeechT5PreTrainedModel {
|
|
|
14466
14534
|
|
|
14467
14535
|
const { encoder_outputs, encoder_attention_mask } = await encoderForward(this, model_inputs);
|
|
14468
14536
|
|
|
14537
|
+
// @ts-expect-error TS2339
|
|
14469
14538
|
const r = encoder_outputs.dims[1] / this.config.reduction_factor;
|
|
14470
14539
|
const maxlen = Math.floor(r * maxlenratio);
|
|
14471
14540
|
const minlen = Math.floor(r * minlenratio);
|
|
14472
14541
|
|
|
14542
|
+
// @ts-expect-error TS2339
|
|
14473
14543
|
const num_mel_bins = this.config.num_mel_bins;
|
|
14474
14544
|
|
|
14475
14545
|
let spectrogramParts = [];
|
|
@@ -14623,9 +14693,9 @@ class ClapTextModelWithProjection extends ClapPreTrainedModel {
|
|
|
14623
14693
|
/** @type {typeof PreTrainedModel.from_pretrained} */
|
|
14624
14694
|
static async from_pretrained(pretrained_model_name_or_path, options = {}) {
|
|
14625
14695
|
return super.from_pretrained(pretrained_model_name_or_path, {
|
|
14626
|
-
// Update default model file name if not provided
|
|
14627
|
-
model_file_name: 'text_model',
|
|
14628
14696
|
...options,
|
|
14697
|
+
// Update default model file name if not provided
|
|
14698
|
+
model_file_name: options.model_file_name ?? 'text_model',
|
|
14629
14699
|
});
|
|
14630
14700
|
}
|
|
14631
14701
|
}
|
|
@@ -14660,9 +14730,9 @@ class ClapAudioModelWithProjection extends ClapPreTrainedModel {
|
|
|
14660
14730
|
/** @type {typeof PreTrainedModel.from_pretrained} */
|
|
14661
14731
|
static async from_pretrained(pretrained_model_name_or_path, options = {}) {
|
|
14662
14732
|
return super.from_pretrained(pretrained_model_name_or_path, {
|
|
14663
|
-
// Update default model file name if not provided
|
|
14664
|
-
model_file_name: 'audio_model',
|
|
14665
14733
|
...options,
|
|
14734
|
+
// Update default model file name if not provided
|
|
14735
|
+
model_file_name: options.model_file_name ?? 'audio_model',
|
|
14666
14736
|
});
|
|
14667
14737
|
}
|
|
14668
14738
|
}
|
|
@@ -14834,11 +14904,13 @@ class MusicgenForConditionalGeneration extends PreTrainedModel { // NOTE: not Mu
|
|
|
14834
14904
|
*/
|
|
14835
14905
|
_apply_and_filter_by_delay_pattern_mask(outputs) {
|
|
14836
14906
|
const [bs_x_codebooks, seqLength] = outputs.dims;
|
|
14907
|
+
// @ts-expect-error TS2339
|
|
14837
14908
|
const num_codebooks = this.config.decoder.num_codebooks;
|
|
14838
14909
|
const upperBound = (seqLength - num_codebooks);
|
|
14839
14910
|
|
|
14840
14911
|
let newDataSize = 0;
|
|
14841
14912
|
for (let i = 0; i < outputs.size; ++i) {
|
|
14913
|
+
// @ts-expect-error TS2339
|
|
14842
14914
|
if (outputs.data[i] === this.config.decoder.pad_token_id) {
|
|
14843
14915
|
continue;
|
|
14844
14916
|
}
|
|
@@ -14868,7 +14940,9 @@ class MusicgenForConditionalGeneration extends PreTrainedModel { // NOTE: not Mu
|
|
|
14868
14940
|
let clonedInputIds = structuredClone(input_ids);
|
|
14869
14941
|
for (let i = 0; i < clonedInputIds.length; ++i) {
|
|
14870
14942
|
for (let j = 0; j < clonedInputIds[i].length; ++j) {
|
|
14943
|
+
// @ts-expect-error TS2339
|
|
14871
14944
|
if ((i % this.config.decoder.num_codebooks) >= j) {
|
|
14945
|
+
// @ts-expect-error TS2339
|
|
14872
14946
|
clonedInputIds[i][j] = BigInt(this.config.decoder.pad_token_id);
|
|
14873
14947
|
}
|
|
14874
14948
|
}
|
|
@@ -15025,6 +15099,9 @@ class MultiModalityCausalLM extends MultiModalityPreTrainedModel {
|
|
|
15025
15099
|
'past_key_values',
|
|
15026
15100
|
];
|
|
15027
15101
|
|
|
15102
|
+
/**
|
|
15103
|
+
* @param {ConstructorParameters<typeof MultiModalityPreTrainedModel>} args
|
|
15104
|
+
*/
|
|
15028
15105
|
constructor(...args) {
|
|
15029
15106
|
super(...args);
|
|
15030
15107
|
|
|
@@ -15303,6 +15380,7 @@ const MODEL_MAPPING_NAMES_ENCODER_ONLY = new Map([
|
|
|
15303
15380
|
['convnext', ['ConvNextModel', ConvNextModel]],
|
|
15304
15381
|
['convnextv2', ['ConvNextV2Model', ConvNextV2Model]],
|
|
15305
15382
|
['dinov2', ['Dinov2Model', Dinov2Model]],
|
|
15383
|
+
['dinov2_with_registers', ['Dinov2WithRegistersModel', Dinov2WithRegistersModel]],
|
|
15306
15384
|
['resnet', ['ResNetModel', ResNetModel]],
|
|
15307
15385
|
['swin', ['SwinModel', SwinModel]],
|
|
15308
15386
|
['swin2sr', ['Swin2SRModel', Swin2SRModel]],
|
|
@@ -15548,6 +15626,7 @@ const MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES = new Map([
|
|
|
15548
15626
|
['convnext', ['ConvNextForImageClassification', ConvNextForImageClassification]],
|
|
15549
15627
|
['convnextv2', ['ConvNextV2ForImageClassification', ConvNextV2ForImageClassification]],
|
|
15550
15628
|
['dinov2', ['Dinov2ForImageClassification', Dinov2ForImageClassification]],
|
|
15629
|
+
['dinov2_with_registers', ['Dinov2WithRegistersForImageClassification', Dinov2WithRegistersForImageClassification]],
|
|
15551
15630
|
['resnet', ['ResNetForImageClassification', ResNetForImageClassification]],
|
|
15552
15631
|
['swin', ['SwinForImageClassification', SwinForImageClassification]],
|
|
15553
15632
|
['segformer', ['SegformerForImageClassification', SegformerForImageClassification]],
|
|
@@ -15991,10 +16070,17 @@ class SequenceClassifierOutput extends ModelOutput {
|
|
|
15991
16070
|
/**
|
|
15992
16071
|
* @param {Object} output The output of the model.
|
|
15993
16072
|
* @param {Tensor} output.logits classification (or regression if config.num_labels==1) scores (before SoftMax).
|
|
16073
|
+
* @param {Record<string, Tensor>} [output.attentions] Object of `torch.FloatTensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, sequence_length)`.
|
|
16074
|
+
* Attentions weights after the attention softmax, used to compute the weighted average in the self-attention heads.
|
|
15994
16075
|
*/
|
|
15995
|
-
constructor({ logits }) {
|
|
16076
|
+
constructor({ logits, ...attentions }) {
|
|
15996
16077
|
super();
|
|
15997
16078
|
this.logits = logits;
|
|
16079
|
+
const attentions_list = Object.values(attentions);
|
|
16080
|
+
if (attentions_list.length > 0) {
|
|
16081
|
+
// Only set attentions if they are not empty
|
|
16082
|
+
this.attentions = attentions_list;
|
|
16083
|
+
}
|
|
15998
16084
|
}
|
|
15999
16085
|
}
|
|
16000
16086
|
|
|
@@ -16250,22 +16336,6 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
16250
16336
|
|
|
16251
16337
|
class AutoFeatureExtractor {
|
|
16252
16338
|
|
|
16253
|
-
/**
|
|
16254
|
-
* Instantiate one of the feature extractor classes of the library from a pretrained model.
|
|
16255
|
-
*
|
|
16256
|
-
* The processor class to instantiate is selected based on the `feature_extractor_type` property of
|
|
16257
|
-
* the config object (either passed as an argument or loaded from `pretrained_model_name_or_path` if possible)
|
|
16258
|
-
*
|
|
16259
|
-
* @param {string} pretrained_model_name_or_path The name or path of the pretrained model. Can be either:
|
|
16260
|
-
* - A string, the *model id* of a pretrained processor hosted inside a model repo on huggingface.co.
|
|
16261
|
-
* Valid model ids can be located at the root-level, like `bert-base-uncased`, or namespaced under a
|
|
16262
|
-
* user or organization name, like `dbmdz/bert-base-german-cased`.
|
|
16263
|
-
* - A path to a *directory* containing processor files, e.g., `./my_model_directory/`.
|
|
16264
|
-
* @param {import('../../utils/hub.js').PretrainedOptions} options Additional options for loading the processor.
|
|
16265
|
-
*
|
|
16266
|
-
* @returns {Promise<AllFeatureExtractors.ImageProcessor>} A new instance of the Processor class.
|
|
16267
|
-
*/
|
|
16268
|
-
|
|
16269
16339
|
/** @type {typeof FeatureExtractor.from_pretrained} */
|
|
16270
16340
|
static async from_pretrained(pretrained_model_name_or_path, options={}) {
|
|
16271
16341
|
|
|
@@ -16392,22 +16462,6 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
16392
16462
|
*/
|
|
16393
16463
|
class AutoProcessor {
|
|
16394
16464
|
|
|
16395
|
-
/**
|
|
16396
|
-
* Instantiate one of the processor classes of the library from a pretrained model.
|
|
16397
|
-
*
|
|
16398
|
-
* The processor class to instantiate is selected based on the `image_processor_type` (or `feature_extractor_type`; legacy)
|
|
16399
|
-
* property of the config object (either passed as an argument or loaded from `pretrained_model_name_or_path` if possible)
|
|
16400
|
-
*
|
|
16401
|
-
* @param {string} pretrained_model_name_or_path The name or path of the pretrained model. Can be either:
|
|
16402
|
-
* - A string, the *model id* of a pretrained processor hosted inside a model repo on huggingface.co.
|
|
16403
|
-
* Valid model ids can be located at the root-level, like `bert-base-uncased`, or namespaced under a
|
|
16404
|
-
* user or organization name, like `dbmdz/bert-base-german-cased`.
|
|
16405
|
-
* - A path to a *directory* containing processor files, e.g., `./my_model_directory/`.
|
|
16406
|
-
* @param {import('../../utils/hub.js').PretrainedOptions} options Additional options for loading the processor.
|
|
16407
|
-
*
|
|
16408
|
-
* @returns {Promise<Processor>} A new instance of the Processor class.
|
|
16409
|
-
*/
|
|
16410
|
-
|
|
16411
16465
|
/** @type {typeof Processor.from_pretrained} */
|
|
16412
16466
|
static async from_pretrained(pretrained_model_name_or_path, options={}) {
|
|
16413
16467
|
|
|
@@ -16725,6 +16779,7 @@ class ConvNextImageProcessor extends _base_image_processors_utils_js__WEBPACK_IM
|
|
|
16725
16779
|
/**
|
|
16726
16780
|
* Percentage of the image to crop. Only has an effect if this.size < 384.
|
|
16727
16781
|
*/
|
|
16782
|
+
// @ts-expect-error TS2339
|
|
16728
16783
|
this.crop_pct = this.config.crop_pct ?? (224 / 256);
|
|
16729
16784
|
}
|
|
16730
16785
|
|
|
@@ -16927,6 +16982,7 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
16927
16982
|
class EfficientNetImageProcessor extends _base_image_processors_utils_js__WEBPACK_IMPORTED_MODULE_0__.ImageProcessor {
|
|
16928
16983
|
constructor(config) {
|
|
16929
16984
|
super(config);
|
|
16985
|
+
// @ts-expect-error TS2339
|
|
16930
16986
|
this.include_top = this.config.include_top ?? true;
|
|
16931
16987
|
if (this.include_top) {
|
|
16932
16988
|
this.image_std = this.image_std.map(x => x * x);
|
|
@@ -17008,8 +17064,11 @@ class Florence2Processor extends _base_processing_utils_js__WEBPACK_IMPORTED_MOD
|
|
|
17008
17064
|
super(config, components);
|
|
17009
17065
|
|
|
17010
17066
|
const {
|
|
17067
|
+
// @ts-expect-error TS2339
|
|
17011
17068
|
tasks_answer_post_processing_type,
|
|
17069
|
+
// @ts-expect-error TS2339
|
|
17012
17070
|
task_prompts_without_inputs,
|
|
17071
|
+
// @ts-expect-error TS2339
|
|
17013
17072
|
task_prompts_with_input,
|
|
17014
17073
|
} = this.image_processor.config;
|
|
17015
17074
|
|
|
@@ -17304,6 +17363,8 @@ class Idefics3ImageProcessor extends _base_image_processors_utils_js__WEBPACK_IM
|
|
|
17304
17363
|
|
|
17305
17364
|
const start_offset = i * pixel_attention_mask_stride + num_patches * h * w;
|
|
17306
17365
|
const end_offset = (i + 1) * pixel_attention_mask_stride;
|
|
17366
|
+
|
|
17367
|
+
// @ts-expect-error
|
|
17307
17368
|
pixel_attention_mask_data.fill(false, start_offset, end_offset);
|
|
17308
17369
|
}
|
|
17309
17370
|
}
|
|
@@ -17710,6 +17771,7 @@ class VLMImageProcessor extends _base_image_processors_utils_js__WEBPACK_IMPORTE
|
|
|
17710
17771
|
},
|
|
17711
17772
|
...config,
|
|
17712
17773
|
});
|
|
17774
|
+
// @ts-expect-error TS2339
|
|
17713
17775
|
this.constant_values = this.config.background_color.map(x => x * this.rescale_factor)
|
|
17714
17776
|
}
|
|
17715
17777
|
|
|
@@ -18151,6 +18213,8 @@ class MgpstrProcessor extends _base_processing_utils_js__WEBPACK_IMPORTED_MODULE
|
|
|
18151
18213
|
* - bpe_preds: The list of BPE decoded sentences.
|
|
18152
18214
|
* - wp_preds: The list of wp decoded sentences.
|
|
18153
18215
|
*/
|
|
18216
|
+
// @ts-expect-error The type of this method is not compatible with the one
|
|
18217
|
+
// in the base class. It might be a good idea to fix this.
|
|
18154
18218
|
batch_decode([char_logits, bpe_logits, wp_logits]) {
|
|
18155
18219
|
const [char_preds, char_scores] = this._decode_helper(char_logits, 'char');
|
|
18156
18220
|
const [bpe_preds, bpe_scores] = this._decode_helper(bpe_logits, 'bpe');
|
|
@@ -18532,6 +18596,7 @@ class PaliGemmaProcessor extends _base_processing_utils_js__WEBPACK_IMPORTED_MOD
|
|
|
18532
18596
|
}
|
|
18533
18597
|
|
|
18534
18598
|
const bos_token = this.tokenizer.bos_token;
|
|
18599
|
+
// @ts-expect-error TS2339
|
|
18535
18600
|
const image_seq_length = this.image_processor.config.image_seq_length;
|
|
18536
18601
|
let input_strings;
|
|
18537
18602
|
if (text.some((t) => t.includes(IMAGE_TOKEN))) {
|
|
@@ -18782,7 +18847,7 @@ class Phi3VProcessor extends _base_processing_utils_js__WEBPACK_IMPORTED_MODULE_
|
|
|
18782
18847
|
*
|
|
18783
18848
|
* @param {string|string[]} text
|
|
18784
18849
|
* @param {RawImage|RawImage[]} images
|
|
18785
|
-
* @param {
|
|
18850
|
+
* @param { { padding?: boolean, truncation?: boolean, num_crops?: number } | undefined } options
|
|
18786
18851
|
* @returns {Promise<any>}
|
|
18787
18852
|
*/
|
|
18788
18853
|
async _call(text, images = null, {
|
|
@@ -18966,6 +19031,7 @@ class PyAnnoteFeatureExtractor extends _base_feature_extraction_utils_js__WEBPAC
|
|
|
18966
19031
|
|
|
18967
19032
|
let current_speaker = -1;
|
|
18968
19033
|
for (let i = 0; i < scores.length; ++i) {
|
|
19034
|
+
/** @type {number[]} */
|
|
18969
19035
|
const probabilities = (0,_utils_maths_js__WEBPACK_IMPORTED_MODULE_2__.softmax)(scores[i]);
|
|
18970
19036
|
const [score, id] = (0,_utils_maths_js__WEBPACK_IMPORTED_MODULE_2__.max)(probabilities);
|
|
18971
19037
|
const [start, end] = [i, i + 1];
|
|
@@ -19150,6 +19216,7 @@ class Qwen2VLProcessor extends _base_processing_utils_js__WEBPACK_IMPORTED_MODUL
|
|
|
19150
19216
|
}
|
|
19151
19217
|
|
|
19152
19218
|
if (image_grid_thw) {
|
|
19219
|
+
// @ts-expect-error TS2551
|
|
19153
19220
|
let merge_length = this.image_processor.config.merge_size ** 2;
|
|
19154
19221
|
let index = 0;
|
|
19155
19222
|
|
|
@@ -19637,8 +19704,8 @@ class SeamlessM4TFeatureExtractor extends _base_feature_extraction_utils_js__WEB
|
|
|
19637
19704
|
'int64',
|
|
19638
19705
|
new BigInt64Array(numPaddedFrames),
|
|
19639
19706
|
[1, numPaddedFrames],
|
|
19640
|
-
)
|
|
19641
|
-
padded_attention_mask.data.fill(1n, 0, num_frames);
|
|
19707
|
+
);
|
|
19708
|
+
/** @type {BigInt64Array} */ (padded_attention_mask.data).fill(1n, 0, num_frames);
|
|
19642
19709
|
}
|
|
19643
19710
|
}
|
|
19644
19711
|
}
|
|
@@ -20438,7 +20505,7 @@ class WhisperFeatureExtractor extends _base_feature_extraction_utils_js__WEBPACK
|
|
|
20438
20505
|
)
|
|
20439
20506
|
|
|
20440
20507
|
const data = features.data;
|
|
20441
|
-
const maxValue = (0,_utils_maths_js__WEBPACK_IMPORTED_MODULE_3__.max)(data)[0];
|
|
20508
|
+
const maxValue = (0,_utils_maths_js__WEBPACK_IMPORTED_MODULE_3__.max)(/** @type {Float32Array} */(data))[0];
|
|
20442
20509
|
|
|
20443
20510
|
for (let i = 0; i < data.length; ++i) {
|
|
20444
20511
|
data[i] = (Math.max(data[i], maxValue - 8.0) + 4.0) / 4.0;
|
|
@@ -20697,6 +20764,16 @@ class TensorOpRegistry {
|
|
|
20697
20764
|
// executionProviders: ['webgpu'],
|
|
20698
20765
|
};
|
|
20699
20766
|
|
|
20767
|
+
static get nearest_interpolate_4d() {
|
|
20768
|
+
if (!this._nearest_interpolate_4d) {
|
|
20769
|
+
this._nearest_interpolate_4d = wrap(
|
|
20770
|
+
[8, 10, 18, 0, 58, 129, 1, 10, 41, 10, 1, 120, 10, 0, 10, 0, 10, 1, 115, 18, 1, 121, 34, 6, 82, 101, 115, 105, 122, 101, 42, 18, 10, 4, 109, 111, 100, 101, 34, 7, 110, 101, 97, 114, 101, 115, 116, 160, 1, 3, 18, 1, 114, 90, 31, 10, 1, 120, 18, 26, 10, 24, 8, 1, 18, 20, 10, 3, 18, 1, 98, 10, 3, 18, 1, 99, 10, 3, 18, 1, 104, 10, 3, 18, 1, 119, 90, 15, 10, 1, 115, 18, 10, 10, 8, 8, 7, 18, 4, 10, 2, 8, 4, 98, 31, 10, 1, 121, 18, 26, 10, 24, 8, 1, 18, 20, 10, 3, 18, 1, 98, 10, 3, 18, 1, 99, 10, 3, 18, 1, 104, 10, 3, 18, 1, 119, 66, 2, 16, 21],
|
|
20771
|
+
this.session_options,
|
|
20772
|
+
'y',
|
|
20773
|
+
);
|
|
20774
|
+
}
|
|
20775
|
+
return this._nearest_interpolate_4d;
|
|
20776
|
+
}
|
|
20700
20777
|
static get bilinear_interpolate_4d() {
|
|
20701
20778
|
if (!this._bilinear_interpolate_4d) {
|
|
20702
20779
|
this._bilinear_interpolate_4d = wrap(
|
|
@@ -21070,6 +21147,7 @@ class TextClassificationPipeline extends (/** @type {new (options: TextPipelineC
|
|
|
21070
21147
|
|
|
21071
21148
|
// TODO: Use softmax tensor function
|
|
21072
21149
|
const function_to_apply =
|
|
21150
|
+
// @ts-expect-error TS2339
|
|
21073
21151
|
this.model.config.problem_type === 'multi_label_classification'
|
|
21074
21152
|
? batch => batch.sigmoid()
|
|
21075
21153
|
: batch => new _utils_tensor_js__WEBPACK_IMPORTED_MODULE_8__.Tensor(
|
|
@@ -21078,6 +21156,7 @@ class TextClassificationPipeline extends (/** @type {new (options: TextPipelineC
|
|
|
21078
21156
|
batch.dims,
|
|
21079
21157
|
); // single_label_classification (default)
|
|
21080
21158
|
|
|
21159
|
+
// @ts-expect-error TS2339
|
|
21081
21160
|
const id2label = this.model.config.id2label;
|
|
21082
21161
|
|
|
21083
21162
|
const toReturn = [];
|
|
@@ -21180,6 +21259,7 @@ class TokenClassificationPipeline extends (/** @type {new (options: TextPipeline
|
|
|
21180
21259
|
const outputs = await this.model(model_inputs)
|
|
21181
21260
|
|
|
21182
21261
|
const logits = outputs.logits;
|
|
21262
|
+
// @ts-expect-error TS2339
|
|
21183
21263
|
const id2label = this.model.config.id2label;
|
|
21184
21264
|
|
|
21185
21265
|
const toReturn = [];
|
|
@@ -21519,11 +21599,14 @@ class Text2TextGenerationPipeline extends (/** @type {new (options: TextPipeline
|
|
|
21519
21599
|
|
|
21520
21600
|
|
|
21521
21601
|
// Add global prefix, if present
|
|
21602
|
+
// @ts-expect-error TS2339
|
|
21522
21603
|
if (this.model.config.prefix) {
|
|
21604
|
+
// @ts-expect-error TS2339
|
|
21523
21605
|
texts = texts.map(x => this.model.config.prefix + x)
|
|
21524
21606
|
}
|
|
21525
21607
|
|
|
21526
21608
|
// Handle task specific params:
|
|
21609
|
+
// @ts-expect-error TS2339
|
|
21527
21610
|
const task_specific_params = this.model.config.task_specific_params
|
|
21528
21611
|
if (task_specific_params && task_specific_params[this.task]) {
|
|
21529
21612
|
// Add prefixes, if present
|
|
@@ -22262,6 +22345,7 @@ class AudioClassificationPipeline extends (/** @type {new (options: AudioPipelin
|
|
|
22262
22345
|
const sampling_rate = this.processor.feature_extractor.config.sampling_rate;
|
|
22263
22346
|
const preparedAudios = await prepareAudios(audio, sampling_rate);
|
|
22264
22347
|
|
|
22348
|
+
// @ts-expect-error TS2339
|
|
22265
22349
|
const id2label = this.model.config.id2label;
|
|
22266
22350
|
|
|
22267
22351
|
const toReturn = [];
|
|
@@ -22572,6 +22656,7 @@ class AutomaticSpeechRecognitionPipeline extends (/** @type {new (options: TextA
|
|
|
22572
22656
|
audio = [/** @type {AudioInput} */ (audio)];
|
|
22573
22657
|
}
|
|
22574
22658
|
|
|
22659
|
+
// @ts-expect-error TS2339
|
|
22575
22660
|
const time_precision = this.processor.feature_extractor.config.chunk_length / this.model.config.max_source_positions;
|
|
22576
22661
|
const hop_length = this.processor.feature_extractor.config.hop_length;
|
|
22577
22662
|
|
|
@@ -22637,7 +22722,9 @@ class AutomaticSpeechRecognitionPipeline extends (/** @type {new (options: TextA
|
|
|
22637
22722
|
|
|
22638
22723
|
// TODO: Right now we only get top beam
|
|
22639
22724
|
if (return_timestamps === 'word') {
|
|
22725
|
+
// @ts-expect-error TS2339
|
|
22640
22726
|
chunk.tokens = data.sequences.tolist()[0];
|
|
22727
|
+
// @ts-expect-error TS2339
|
|
22641
22728
|
chunk.token_timestamps = data.token_timestamps.tolist()[0].map(
|
|
22642
22729
|
(/** @type {number} */ x) => (0,_utils_maths_js__WEBPACK_IMPORTED_MODULE_6__.round)(x, 2)
|
|
22643
22730
|
);
|
|
@@ -22682,7 +22769,7 @@ class AutomaticSpeechRecognitionPipeline extends (/** @type {new (options: TextA
|
|
|
22682
22769
|
const max_new_tokens = Math.floor(aud.length / sampling_rate) * 6;
|
|
22683
22770
|
const outputs = await this.model.generate({ max_new_tokens, ...kwargs, ...inputs });
|
|
22684
22771
|
|
|
22685
|
-
const text = this.processor.batch_decode(outputs, { skip_special_tokens: true })[0];
|
|
22772
|
+
const text = this.processor.batch_decode(/** @type {Tensor} */(outputs), { skip_special_tokens: true })[0];
|
|
22686
22773
|
toReturn.push({ text });
|
|
22687
22774
|
}
|
|
22688
22775
|
return single ? toReturn[0] : toReturn;
|
|
@@ -22831,6 +22918,7 @@ class ImageClassificationPipeline extends (/** @type {new (options: ImagePipelin
|
|
|
22831
22918
|
const { pixel_values } = await this.processor(preparedImages);
|
|
22832
22919
|
const output = await this.model({ pixel_values });
|
|
22833
22920
|
|
|
22921
|
+
// @ts-expect-error TS2339
|
|
22834
22922
|
const id2label = this.model.config.id2label;
|
|
22835
22923
|
|
|
22836
22924
|
/** @type {ImageClassificationOutput[]} */
|
|
@@ -22945,6 +23033,7 @@ class ImageSegmentationPipeline extends (/** @type {new (options: ImagePipelineC
|
|
|
22945
23033
|
}
|
|
22946
23034
|
}
|
|
22947
23035
|
|
|
23036
|
+
// @ts-expect-error TS2339
|
|
22948
23037
|
const id2label = this.model.config.id2label;
|
|
22949
23038
|
|
|
22950
23039
|
/** @type {ImageSegmentationPipelineOutput[]} */
|
|
@@ -23171,6 +23260,7 @@ class ObjectDetectionPipeline extends (/** @type {new (options: ImagePipelineCon
|
|
|
23171
23260
|
const processed = this.processor.image_processor.post_process_object_detection(output, threshold, imageSizes);
|
|
23172
23261
|
|
|
23173
23262
|
// Add labels
|
|
23263
|
+
// @ts-expect-error TS2339
|
|
23174
23264
|
const id2label = this.model.config.id2label;
|
|
23175
23265
|
|
|
23176
23266
|
// Format output
|
|
@@ -23390,6 +23480,7 @@ class DocumentQuestionAnsweringPipeline extends (/** @type {new (options: TextIm
|
|
|
23390
23480
|
// Run model
|
|
23391
23481
|
const output = await this.model.generate({
|
|
23392
23482
|
inputs: pixel_values,
|
|
23483
|
+
// @ts-expect-error TS2339
|
|
23393
23484
|
max_length: this.model.config.decoder.max_position_embeddings,
|
|
23394
23485
|
decoder_input_ids,
|
|
23395
23486
|
...generate_kwargs,
|
|
@@ -23505,6 +23596,7 @@ class TextToAudioPipeline extends (/** @type {new (options: TextToAudioPipelineC
|
|
|
23505
23596
|
// Generate waveform
|
|
23506
23597
|
const { waveform } = await this.model(inputs);
|
|
23507
23598
|
|
|
23599
|
+
// @ts-expect-error TS2339
|
|
23508
23600
|
const sampling_rate = this.model.config.sampling_rate;
|
|
23509
23601
|
return {
|
|
23510
23602
|
audio: waveform.data,
|
|
@@ -23662,11 +23754,23 @@ class DepthEstimationPipeline extends (/** @type {new (options: ImagePipelineCon
|
|
|
23662
23754
|
|
|
23663
23755
|
const toReturn = [];
|
|
23664
23756
|
for (let i = 0; i < preparedImages.length; ++i) {
|
|
23665
|
-
const
|
|
23666
|
-
const
|
|
23757
|
+
const batch = predicted_depth[i];
|
|
23758
|
+
const [height, width] = batch.dims.slice(-2);
|
|
23759
|
+
const [new_width, new_height] = preparedImages[i].size;
|
|
23760
|
+
|
|
23761
|
+
// Interpolate to original size
|
|
23762
|
+
const prediction = (await (0,_utils_tensor_js__WEBPACK_IMPORTED_MODULE_8__.interpolate_4d)(batch.view(1, 1, height, width), {
|
|
23763
|
+
size: [new_height, new_width],
|
|
23764
|
+
mode: 'bilinear',
|
|
23765
|
+
})).view(new_height, new_width);
|
|
23766
|
+
|
|
23767
|
+
const minval = /** @type {number} */(prediction.min().item());
|
|
23768
|
+
const maxval = /** @type {number} */(prediction.max().item());
|
|
23769
|
+
const formatted = prediction.sub(minval).div_(maxval - minval).mul_(255).to('uint8').unsqueeze(0);
|
|
23770
|
+
const depth = _utils_image_js__WEBPACK_IMPORTED_MODULE_9__.RawImage.fromTensor(formatted);
|
|
23667
23771
|
toReturn.push({
|
|
23668
|
-
predicted_depth:
|
|
23669
|
-
depth
|
|
23772
|
+
predicted_depth: prediction,
|
|
23773
|
+
depth,
|
|
23670
23774
|
});
|
|
23671
23775
|
}
|
|
23672
23776
|
|
|
@@ -24146,6 +24250,7 @@ async function loadItems(mapping, model, pretrainedOptions) {
|
|
|
24146
24250
|
return result;
|
|
24147
24251
|
}
|
|
24148
24252
|
|
|
24253
|
+
|
|
24149
24254
|
/***/ }),
|
|
24150
24255
|
|
|
24151
24256
|
/***/ "./src/tokenizers.js":
|
|
@@ -24214,7 +24319,6 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
24214
24319
|
/* harmony import */ var _utils_data_structures_js__WEBPACK_IMPORTED_MODULE_5__ = __webpack_require__(/*! ./utils/data-structures.js */ "./src/utils/data-structures.js");
|
|
24215
24320
|
/* harmony import */ var _huggingface_jinja__WEBPACK_IMPORTED_MODULE_6__ = __webpack_require__(/*! @huggingface/jinja */ "./node_modules/@huggingface/jinja/dist/index.js");
|
|
24216
24321
|
/* harmony import */ var _models_whisper_common_whisper_js__WEBPACK_IMPORTED_MODULE_7__ = __webpack_require__(/*! ./models/whisper/common_whisper.js */ "./src/models/whisper/common_whisper.js");
|
|
24217
|
-
/* harmony import */ var _utils_constants_js__WEBPACK_IMPORTED_MODULE_8__ = __webpack_require__(/*! ./utils/constants.js */ "./src/utils/constants.js");
|
|
24218
24322
|
|
|
24219
24323
|
/**
|
|
24220
24324
|
* @file Tokenizers are used to prepare textual inputs for a model.
|
|
@@ -24251,7 +24355,6 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
24251
24355
|
|
|
24252
24356
|
|
|
24253
24357
|
|
|
24254
|
-
|
|
24255
24358
|
/**
|
|
24256
24359
|
* @typedef {Object} TokenizerProperties Additional tokenizer-specific properties.
|
|
24257
24360
|
* @property {boolean} [legacy=false] Whether or not the `legacy` behavior of the tokenizer should be used.
|
|
@@ -24735,7 +24838,7 @@ class Unigram extends TokenizerModel {
|
|
|
24735
24838
|
* Create a new Unigram tokenizer model.
|
|
24736
24839
|
* @param {Object} config The configuration object for the Unigram model.
|
|
24737
24840
|
* @param {number} config.unk_id The ID of the unknown token
|
|
24738
|
-
* @param {
|
|
24841
|
+
* @param {[string, number][]} config.vocab A 2D array representing a mapping of tokens to scores.
|
|
24739
24842
|
* @param {Object} moreConfig Additional configuration object for the Unigram model.
|
|
24740
24843
|
*/
|
|
24741
24844
|
constructor(config, moreConfig) {
|
|
@@ -24743,11 +24846,10 @@ class Unigram extends TokenizerModel {
|
|
|
24743
24846
|
|
|
24744
24847
|
const vocabSize = config.vocab.length;
|
|
24745
24848
|
this.vocab = new Array(vocabSize);
|
|
24849
|
+
/** @type {number[]} */
|
|
24746
24850
|
this.scores = new Array(vocabSize);
|
|
24747
24851
|
for (let i = 0; i < vocabSize; ++i) {
|
|
24748
|
-
|
|
24749
|
-
this.vocab[i] = piece[0];
|
|
24750
|
-
this.scores[i] = piece[1];
|
|
24852
|
+
[this.vocab[i], this.scores[i]] = config.vocab[i];
|
|
24751
24853
|
}
|
|
24752
24854
|
|
|
24753
24855
|
this.unk_token_id = config.unk_id;
|
|
@@ -30104,6 +30206,8 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
30104
30206
|
/* harmony export */ });
|
|
30105
30207
|
/* harmony import */ var _env_js__WEBPACK_IMPORTED_MODULE_0__ = __webpack_require__(/*! ../env.js */ "./src/env.js");
|
|
30106
30208
|
/* harmony import */ var _devices_js__WEBPACK_IMPORTED_MODULE_1__ = __webpack_require__(/*! ./devices.js */ "./src/utils/devices.js");
|
|
30209
|
+
/// <reference types="@webgpu/types" />
|
|
30210
|
+
|
|
30107
30211
|
|
|
30108
30212
|
|
|
30109
30213
|
|
|
@@ -30357,7 +30461,7 @@ class FileResponse {
|
|
|
30357
30461
|
*/
|
|
30358
30462
|
async arrayBuffer() {
|
|
30359
30463
|
const data = await fs__WEBPACK_IMPORTED_MODULE_0__["default"].promises.readFile(this.filePath);
|
|
30360
|
-
return data.buffer;
|
|
30464
|
+
return /** @type {ArrayBuffer} */ (data.buffer);
|
|
30361
30465
|
}
|
|
30362
30466
|
|
|
30363
30467
|
/**
|
|
@@ -32016,8 +32120,9 @@ function magnitude(arr) {
|
|
|
32016
32120
|
|
|
32017
32121
|
/**
|
|
32018
32122
|
* Returns the value and index of the minimum element in an array.
|
|
32019
|
-
* @
|
|
32020
|
-
* @
|
|
32123
|
+
* @template {number[]|bigint[]|AnyTypedArray} T
|
|
32124
|
+
* @param {T} arr array of numbers.
|
|
32125
|
+
* @returns {T extends bigint[]|BigTypedArray ? [bigint, number] : [number, number]} the value and index of the minimum element, of the form: [valueOfMin, indexOfMin]
|
|
32021
32126
|
* @throws {Error} If array is empty.
|
|
32022
32127
|
*/
|
|
32023
32128
|
function min(arr) {
|
|
@@ -32030,14 +32135,15 @@ function min(arr) {
|
|
|
32030
32135
|
indexOfMin = i;
|
|
32031
32136
|
}
|
|
32032
32137
|
}
|
|
32033
|
-
return [min, indexOfMin];
|
|
32138
|
+
return /** @type {T extends bigint[]|BigTypedArray ? [bigint, number] : [number, number]} */([min, indexOfMin]);
|
|
32034
32139
|
}
|
|
32035
32140
|
|
|
32036
32141
|
|
|
32037
32142
|
/**
|
|
32038
32143
|
* Returns the value and index of the maximum element in an array.
|
|
32039
|
-
* @
|
|
32040
|
-
* @
|
|
32144
|
+
* @template {number[]|bigint[]|AnyTypedArray} T
|
|
32145
|
+
* @param {T} arr array of numbers.
|
|
32146
|
+
* @returns {T extends bigint[]|BigTypedArray ? [bigint, number] : [number, number]} the value and index of the maximum element, of the form: [valueOfMax, indexOfMax]
|
|
32041
32147
|
* @throws {Error} If array is empty.
|
|
32042
32148
|
*/
|
|
32043
32149
|
function max(arr) {
|
|
@@ -32050,7 +32156,7 @@ function max(arr) {
|
|
|
32050
32156
|
indexOfMax = i;
|
|
32051
32157
|
}
|
|
32052
32158
|
}
|
|
32053
|
-
return [
|
|
32159
|
+
return /** @type {T extends bigint[]|BigTypedArray ? [bigint, number] : [number, number]} */([max, indexOfMax]);
|
|
32054
32160
|
}
|
|
32055
32161
|
|
|
32056
32162
|
function isPowerOfTwo(number) {
|
|
@@ -33347,8 +33453,6 @@ class Tensor {
|
|
|
33347
33453
|
return this.permute(...dims);
|
|
33348
33454
|
}
|
|
33349
33455
|
|
|
33350
|
-
// TODO add .max() and .min() methods
|
|
33351
|
-
|
|
33352
33456
|
/**
|
|
33353
33457
|
* Returns the sum of each row of the input tensor in the given dimension dim.
|
|
33354
33458
|
*
|
|
@@ -33642,6 +33746,36 @@ class Tensor {
|
|
|
33642
33746
|
return mean(this, dim, keepdim);
|
|
33643
33747
|
}
|
|
33644
33748
|
|
|
33749
|
+
min(dim = null, keepdim = false) {
|
|
33750
|
+
if (dim !== null) {
|
|
33751
|
+
throw new Error("`dim !== null` not yet implemented.");
|
|
33752
|
+
}
|
|
33753
|
+
const value = (0,_maths_js__WEBPACK_IMPORTED_MODULE_0__.min)(this.data)[0];
|
|
33754
|
+
return new Tensor(this.type, [value], []);
|
|
33755
|
+
}
|
|
33756
|
+
max(dim = null, keepdim = false) {
|
|
33757
|
+
if (dim !== null) {
|
|
33758
|
+
throw new Error("`dim !== null` not yet implemented.");
|
|
33759
|
+
}
|
|
33760
|
+
const value = (0,_maths_js__WEBPACK_IMPORTED_MODULE_0__.max)(this.data)[0];
|
|
33761
|
+
return new Tensor(this.type, [value], []);
|
|
33762
|
+
}
|
|
33763
|
+
|
|
33764
|
+
argmin(dim = null, keepdim = false) {
|
|
33765
|
+
if (dim !== null) {
|
|
33766
|
+
throw new Error("`dim !== null` not yet implemented.");
|
|
33767
|
+
}
|
|
33768
|
+
const index = (0,_maths_js__WEBPACK_IMPORTED_MODULE_0__.min)(this.data)[1];
|
|
33769
|
+
return new Tensor('int64', [BigInt(index)], []);
|
|
33770
|
+
}
|
|
33771
|
+
argmax(dim = null, keepdim = false) {
|
|
33772
|
+
if (dim !== null) {
|
|
33773
|
+
throw new Error("`dim !== null` not yet implemented.");
|
|
33774
|
+
}
|
|
33775
|
+
const index = (0,_maths_js__WEBPACK_IMPORTED_MODULE_0__.max)(this.data)[1];
|
|
33776
|
+
return new Tensor('int64', [BigInt(index)], []);
|
|
33777
|
+
}
|
|
33778
|
+
|
|
33645
33779
|
/**
|
|
33646
33780
|
* Performs Tensor dtype conversion.
|
|
33647
33781
|
* @param {DataType} type The desired data type.
|
|
@@ -33775,7 +33909,7 @@ function interpolate(input, [out_height, out_width], mode = 'bilinear', align_co
|
|
|
33775
33909
|
* @param {Tensor} input the input tensor
|
|
33776
33910
|
* @param {Object} options the options for the interpolation
|
|
33777
33911
|
* @param {[number, number]|[number, number, number]|[number, number, number, number]} [options.size=null] output spatial size.
|
|
33778
|
-
* @param {"bilinear"|"bicubic"} [options.mode='bilinear'] algorithm used for upsampling
|
|
33912
|
+
* @param {"nearest"|"bilinear"|"bicubic"} [options.mode='bilinear'] algorithm used for upsampling
|
|
33779
33913
|
* @returns {Promise<Tensor>} The interpolated tensor.
|
|
33780
33914
|
*/
|
|
33781
33915
|
async function interpolate_4d(input, {
|
|
@@ -33805,7 +33939,9 @@ async function interpolate_4d(input, {
|
|
|
33805
33939
|
}
|
|
33806
33940
|
|
|
33807
33941
|
let op;
|
|
33808
|
-
if (mode === '
|
|
33942
|
+
if (mode === 'nearest') {
|
|
33943
|
+
op = await _ops_registry_js__WEBPACK_IMPORTED_MODULE_2__.TensorOpRegistry.nearest_interpolate_4d;
|
|
33944
|
+
} else if (mode === 'bilinear') {
|
|
33809
33945
|
op = await _ops_registry_js__WEBPACK_IMPORTED_MODULE_2__.TensorOpRegistry.bilinear_interpolate_4d;
|
|
33810
33946
|
} else if (mode === 'bicubic') {
|
|
33811
33947
|
op = await _ops_registry_js__WEBPACK_IMPORTED_MODULE_2__.TensorOpRegistry.bicubic_interpolate_4d;
|
|
@@ -33846,13 +33982,13 @@ async function rfft(x, a) {
|
|
|
33846
33982
|
* Returns the k largest elements of the given input tensor.
|
|
33847
33983
|
* Inspired by https://pytorch.org/docs/stable/generated/torch.topk.html
|
|
33848
33984
|
* @param {Tensor} x the input tensor
|
|
33849
|
-
* @param {number} k the k in "top-k"
|
|
33985
|
+
* @param {number} [k] the k in "top-k"
|
|
33850
33986
|
* @returns {Promise<[Tensor, Tensor]>} the output tuple of (Tensor, LongTensor) of top-k elements and their indices.
|
|
33851
33987
|
*/
|
|
33852
33988
|
async function topk(x, k) {
|
|
33853
33989
|
const op = await _ops_registry_js__WEBPACK_IMPORTED_MODULE_2__.TensorOpRegistry.top_k;
|
|
33854
33990
|
|
|
33855
|
-
if (k
|
|
33991
|
+
if (k == null) {
|
|
33856
33992
|
k = x.dims.at(-1);
|
|
33857
33993
|
} else {
|
|
33858
33994
|
k = Math.min(k, x.dims.at(-1));
|
|
@@ -33881,10 +34017,10 @@ const arrayToIndexTensor = (array) => new Tensor('int64', array, [array.length])
|
|
|
33881
34017
|
async function slice(data, starts, ends, axes, steps) {
|
|
33882
34018
|
const op = await _ops_registry_js__WEBPACK_IMPORTED_MODULE_2__.TensorOpRegistry.slice;
|
|
33883
34019
|
return await op({
|
|
33884
|
-
x: data,
|
|
33885
|
-
s: arrayToIndexTensor(starts),
|
|
33886
|
-
e: arrayToIndexTensor(ends),
|
|
33887
|
-
a: arrayToIndexTensor(axes),
|
|
34020
|
+
x: data,
|
|
34021
|
+
s: arrayToIndexTensor(starts),
|
|
34022
|
+
e: arrayToIndexTensor(ends),
|
|
34023
|
+
a: arrayToIndexTensor(axes),
|
|
33888
34024
|
t: arrayToIndexTensor(steps ?? new Array(axes.length).fill(1)),
|
|
33889
34025
|
});
|
|
33890
34026
|
}
|
|
@@ -34664,6 +34800,9 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
34664
34800
|
/* harmony export */ Dinov2ForImageClassification: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Dinov2ForImageClassification),
|
|
34665
34801
|
/* harmony export */ Dinov2Model: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Dinov2Model),
|
|
34666
34802
|
/* harmony export */ Dinov2PreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Dinov2PreTrainedModel),
|
|
34803
|
+
/* harmony export */ Dinov2WithRegistersForImageClassification: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Dinov2WithRegistersForImageClassification),
|
|
34804
|
+
/* harmony export */ Dinov2WithRegistersModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Dinov2WithRegistersModel),
|
|
34805
|
+
/* harmony export */ Dinov2WithRegistersPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Dinov2WithRegistersPreTrainedModel),
|
|
34667
34806
|
/* harmony export */ DistilBertForMaskedLM: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.DistilBertForMaskedLM),
|
|
34668
34807
|
/* harmony export */ DistilBertForQuestionAnswering: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.DistilBertForQuestionAnswering),
|
|
34669
34808
|
/* harmony export */ DistilBertForSequenceClassification: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.DistilBertForSequenceClassification),
|
|
@@ -35441,6 +35580,9 @@ var __webpack_exports__DetrSegmentationOutput = __webpack_exports__.DetrSegmenta
|
|
|
35441
35580
|
var __webpack_exports__Dinov2ForImageClassification = __webpack_exports__.Dinov2ForImageClassification;
|
|
35442
35581
|
var __webpack_exports__Dinov2Model = __webpack_exports__.Dinov2Model;
|
|
35443
35582
|
var __webpack_exports__Dinov2PreTrainedModel = __webpack_exports__.Dinov2PreTrainedModel;
|
|
35583
|
+
var __webpack_exports__Dinov2WithRegistersForImageClassification = __webpack_exports__.Dinov2WithRegistersForImageClassification;
|
|
35584
|
+
var __webpack_exports__Dinov2WithRegistersModel = __webpack_exports__.Dinov2WithRegistersModel;
|
|
35585
|
+
var __webpack_exports__Dinov2WithRegistersPreTrainedModel = __webpack_exports__.Dinov2WithRegistersPreTrainedModel;
|
|
35444
35586
|
var __webpack_exports__DistilBertForMaskedLM = __webpack_exports__.DistilBertForMaskedLM;
|
|
35445
35587
|
var __webpack_exports__DistilBertForQuestionAnswering = __webpack_exports__.DistilBertForQuestionAnswering;
|
|
35446
35588
|
var __webpack_exports__DistilBertForSequenceClassification = __webpack_exports__.DistilBertForSequenceClassification;
|
|
@@ -35979,6 +36121,6 @@ var __webpack_exports__topk = __webpack_exports__.topk;
|
|
|
35979
36121
|
var __webpack_exports__window_function = __webpack_exports__.window_function;
|
|
35980
36122
|
var __webpack_exports__zeros = __webpack_exports__.zeros;
|
|
35981
36123
|
var __webpack_exports__zeros_like = __webpack_exports__.zeros_like;
|
|
35982
|
-
export { __webpack_exports__ASTFeatureExtractor as ASTFeatureExtractor, __webpack_exports__ASTForAudioClassification as ASTForAudioClassification, __webpack_exports__ASTModel as ASTModel, __webpack_exports__ASTPreTrainedModel as ASTPreTrainedModel, __webpack_exports__AlbertForMaskedLM as AlbertForMaskedLM, __webpack_exports__AlbertForQuestionAnswering as AlbertForQuestionAnswering, __webpack_exports__AlbertForSequenceClassification as AlbertForSequenceClassification, __webpack_exports__AlbertModel as AlbertModel, __webpack_exports__AlbertPreTrainedModel as AlbertPreTrainedModel, __webpack_exports__AlbertTokenizer as AlbertTokenizer, __webpack_exports__AudioClassificationPipeline as AudioClassificationPipeline, __webpack_exports__AutoConfig as AutoConfig, __webpack_exports__AutoFeatureExtractor as AutoFeatureExtractor, __webpack_exports__AutoImageProcessor as AutoImageProcessor, __webpack_exports__AutoModel as AutoModel, __webpack_exports__AutoModelForAudioClassification as AutoModelForAudioClassification, __webpack_exports__AutoModelForAudioFrameClassification as AutoModelForAudioFrameClassification, __webpack_exports__AutoModelForCTC as AutoModelForCTC, __webpack_exports__AutoModelForCausalLM as AutoModelForCausalLM, __webpack_exports__AutoModelForDepthEstimation as AutoModelForDepthEstimation, __webpack_exports__AutoModelForDocumentQuestionAnswering as AutoModelForDocumentQuestionAnswering, __webpack_exports__AutoModelForImageClassification as AutoModelForImageClassification, __webpack_exports__AutoModelForImageFeatureExtraction as AutoModelForImageFeatureExtraction, __webpack_exports__AutoModelForImageMatting as AutoModelForImageMatting, __webpack_exports__AutoModelForImageSegmentation as AutoModelForImageSegmentation, __webpack_exports__AutoModelForImageToImage as AutoModelForImageToImage, __webpack_exports__AutoModelForMaskGeneration as AutoModelForMaskGeneration, __webpack_exports__AutoModelForMaskedLM as AutoModelForMaskedLM, __webpack_exports__AutoModelForNormalEstimation as AutoModelForNormalEstimation, __webpack_exports__AutoModelForObjectDetection as AutoModelForObjectDetection, __webpack_exports__AutoModelForPoseEstimation as AutoModelForPoseEstimation, __webpack_exports__AutoModelForQuestionAnswering as AutoModelForQuestionAnswering, __webpack_exports__AutoModelForSemanticSegmentation as AutoModelForSemanticSegmentation, __webpack_exports__AutoModelForSeq2SeqLM as AutoModelForSeq2SeqLM, __webpack_exports__AutoModelForSequenceClassification as AutoModelForSequenceClassification, __webpack_exports__AutoModelForSpeechSeq2Seq as AutoModelForSpeechSeq2Seq, __webpack_exports__AutoModelForTextToSpectrogram as AutoModelForTextToSpectrogram, __webpack_exports__AutoModelForTextToWaveform as AutoModelForTextToWaveform, __webpack_exports__AutoModelForTokenClassification as AutoModelForTokenClassification, __webpack_exports__AutoModelForUniversalSegmentation as AutoModelForUniversalSegmentation, __webpack_exports__AutoModelForVision2Seq as AutoModelForVision2Seq, __webpack_exports__AutoModelForXVector as AutoModelForXVector, __webpack_exports__AutoModelForZeroShotObjectDetection as AutoModelForZeroShotObjectDetection, __webpack_exports__AutoProcessor as AutoProcessor, __webpack_exports__AutoTokenizer as AutoTokenizer, __webpack_exports__AutomaticSpeechRecognitionPipeline as AutomaticSpeechRecognitionPipeline, __webpack_exports__BartForConditionalGeneration as BartForConditionalGeneration, __webpack_exports__BartForSequenceClassification as BartForSequenceClassification, __webpack_exports__BartModel as BartModel, __webpack_exports__BartPretrainedModel as BartPretrainedModel, __webpack_exports__BartTokenizer as BartTokenizer, __webpack_exports__BaseModelOutput as BaseModelOutput, __webpack_exports__BaseStreamer as BaseStreamer, __webpack_exports__BeitFeatureExtractor as BeitFeatureExtractor, __webpack_exports__BeitForImageClassification as BeitForImageClassification, __webpack_exports__BeitModel as BeitModel, __webpack_exports__BeitPreTrainedModel as BeitPreTrainedModel, __webpack_exports__BertForMaskedLM as BertForMaskedLM, __webpack_exports__BertForQuestionAnswering as BertForQuestionAnswering, __webpack_exports__BertForSequenceClassification as BertForSequenceClassification, __webpack_exports__BertForTokenClassification as BertForTokenClassification, __webpack_exports__BertModel as BertModel, __webpack_exports__BertPreTrainedModel as BertPreTrainedModel, __webpack_exports__BertTokenizer as BertTokenizer, __webpack_exports__BitImageProcessor as BitImageProcessor, __webpack_exports__BlenderbotForConditionalGeneration as BlenderbotForConditionalGeneration, __webpack_exports__BlenderbotModel as BlenderbotModel, __webpack_exports__BlenderbotPreTrainedModel as BlenderbotPreTrainedModel, __webpack_exports__BlenderbotSmallForConditionalGeneration as BlenderbotSmallForConditionalGeneration, __webpack_exports__BlenderbotSmallModel as BlenderbotSmallModel, __webpack_exports__BlenderbotSmallPreTrainedModel as BlenderbotSmallPreTrainedModel, __webpack_exports__BlenderbotSmallTokenizer as BlenderbotSmallTokenizer, __webpack_exports__BlenderbotTokenizer as BlenderbotTokenizer, __webpack_exports__BloomForCausalLM as BloomForCausalLM, __webpack_exports__BloomModel as BloomModel, __webpack_exports__BloomPreTrainedModel as BloomPreTrainedModel, __webpack_exports__BloomTokenizer as BloomTokenizer, __webpack_exports__CLIPFeatureExtractor as CLIPFeatureExtractor, __webpack_exports__CLIPImageProcessor as CLIPImageProcessor, __webpack_exports__CLIPModel as CLIPModel, __webpack_exports__CLIPPreTrainedModel as CLIPPreTrainedModel, __webpack_exports__CLIPSegForImageSegmentation as CLIPSegForImageSegmentation, __webpack_exports__CLIPSegModel as CLIPSegModel, __webpack_exports__CLIPSegPreTrainedModel as CLIPSegPreTrainedModel, __webpack_exports__CLIPTextModel as CLIPTextModel, __webpack_exports__CLIPTextModelWithProjection as CLIPTextModelWithProjection, __webpack_exports__CLIPTokenizer as CLIPTokenizer, __webpack_exports__CLIPVisionModel as CLIPVisionModel, __webpack_exports__CLIPVisionModelWithProjection as CLIPVisionModelWithProjection, __webpack_exports__CamembertForMaskedLM as CamembertForMaskedLM, __webpack_exports__CamembertForQuestionAnswering as CamembertForQuestionAnswering, __webpack_exports__CamembertForSequenceClassification as CamembertForSequenceClassification, __webpack_exports__CamembertForTokenClassification as CamembertForTokenClassification, __webpack_exports__CamembertModel as CamembertModel, __webpack_exports__CamembertPreTrainedModel as CamembertPreTrainedModel, __webpack_exports__CamembertTokenizer as CamembertTokenizer, __webpack_exports__CausalLMOutput as CausalLMOutput, __webpack_exports__CausalLMOutputWithPast as CausalLMOutputWithPast, __webpack_exports__ChineseCLIPFeatureExtractor as ChineseCLIPFeatureExtractor, __webpack_exports__ChineseCLIPModel as ChineseCLIPModel, __webpack_exports__ChineseCLIPPreTrainedModel as ChineseCLIPPreTrainedModel, __webpack_exports__ClapAudioModelWithProjection as ClapAudioModelWithProjection, __webpack_exports__ClapFeatureExtractor as ClapFeatureExtractor, __webpack_exports__ClapModel as ClapModel, __webpack_exports__ClapPreTrainedModel as ClapPreTrainedModel, __webpack_exports__ClapTextModelWithProjection as ClapTextModelWithProjection, __webpack_exports__ClassifierFreeGuidanceLogitsProcessor as ClassifierFreeGuidanceLogitsProcessor, __webpack_exports__CodeGenForCausalLM as CodeGenForCausalLM, __webpack_exports__CodeGenModel as CodeGenModel, __webpack_exports__CodeGenPreTrainedModel as CodeGenPreTrainedModel, __webpack_exports__CodeGenTokenizer as CodeGenTokenizer, __webpack_exports__CodeLlamaTokenizer as CodeLlamaTokenizer, __webpack_exports__CohereForCausalLM as CohereForCausalLM, __webpack_exports__CohereModel as CohereModel, __webpack_exports__CoherePreTrainedModel as CoherePreTrainedModel, __webpack_exports__CohereTokenizer as CohereTokenizer, __webpack_exports__ConvBertForMaskedLM as ConvBertForMaskedLM, __webpack_exports__ConvBertForQuestionAnswering as ConvBertForQuestionAnswering, __webpack_exports__ConvBertForSequenceClassification as ConvBertForSequenceClassification, __webpack_exports__ConvBertForTokenClassification as ConvBertForTokenClassification, __webpack_exports__ConvBertModel as ConvBertModel, __webpack_exports__ConvBertPreTrainedModel as ConvBertPreTrainedModel, __webpack_exports__ConvBertTokenizer as ConvBertTokenizer, __webpack_exports__ConvNextFeatureExtractor as ConvNextFeatureExtractor, __webpack_exports__ConvNextForImageClassification as ConvNextForImageClassification, __webpack_exports__ConvNextImageProcessor as ConvNextImageProcessor, __webpack_exports__ConvNextModel as ConvNextModel, __webpack_exports__ConvNextPreTrainedModel as ConvNextPreTrainedModel, __webpack_exports__ConvNextV2ForImageClassification as ConvNextV2ForImageClassification, __webpack_exports__ConvNextV2Model as ConvNextV2Model, __webpack_exports__ConvNextV2PreTrainedModel as ConvNextV2PreTrainedModel, __webpack_exports__DPTFeatureExtractor as DPTFeatureExtractor, __webpack_exports__DPTForDepthEstimation as DPTForDepthEstimation, __webpack_exports__DPTImageProcessor as DPTImageProcessor, __webpack_exports__DPTModel as DPTModel, __webpack_exports__DPTPreTrainedModel as DPTPreTrainedModel, __webpack_exports__DebertaForMaskedLM as DebertaForMaskedLM, __webpack_exports__DebertaForQuestionAnswering as DebertaForQuestionAnswering, __webpack_exports__DebertaForSequenceClassification as DebertaForSequenceClassification, __webpack_exports__DebertaForTokenClassification as DebertaForTokenClassification, __webpack_exports__DebertaModel as DebertaModel, __webpack_exports__DebertaPreTrainedModel as DebertaPreTrainedModel, __webpack_exports__DebertaTokenizer as DebertaTokenizer, __webpack_exports__DebertaV2ForMaskedLM as DebertaV2ForMaskedLM, __webpack_exports__DebertaV2ForQuestionAnswering as DebertaV2ForQuestionAnswering, __webpack_exports__DebertaV2ForSequenceClassification as DebertaV2ForSequenceClassification, __webpack_exports__DebertaV2ForTokenClassification as DebertaV2ForTokenClassification, __webpack_exports__DebertaV2Model as DebertaV2Model, __webpack_exports__DebertaV2PreTrainedModel as DebertaV2PreTrainedModel, __webpack_exports__DebertaV2Tokenizer as DebertaV2Tokenizer, __webpack_exports__DecisionTransformerModel as DecisionTransformerModel, __webpack_exports__DecisionTransformerPreTrainedModel as DecisionTransformerPreTrainedModel, __webpack_exports__DeiTFeatureExtractor as DeiTFeatureExtractor, __webpack_exports__DeiTForImageClassification as DeiTForImageClassification, __webpack_exports__DeiTImageProcessor as DeiTImageProcessor, __webpack_exports__DeiTModel as DeiTModel, __webpack_exports__DeiTPreTrainedModel as DeiTPreTrainedModel, __webpack_exports__DepthAnythingForDepthEstimation as DepthAnythingForDepthEstimation, __webpack_exports__DepthAnythingPreTrainedModel as DepthAnythingPreTrainedModel, __webpack_exports__DepthEstimationPipeline as DepthEstimationPipeline, __webpack_exports__DepthProForDepthEstimation as DepthProForDepthEstimation, __webpack_exports__DepthProPreTrainedModel as DepthProPreTrainedModel, __webpack_exports__DetrFeatureExtractor as DetrFeatureExtractor, __webpack_exports__DetrForObjectDetection as DetrForObjectDetection, __webpack_exports__DetrForSegmentation as DetrForSegmentation, __webpack_exports__DetrImageProcessor as DetrImageProcessor, __webpack_exports__DetrModel as DetrModel, __webpack_exports__DetrObjectDetectionOutput as DetrObjectDetectionOutput, __webpack_exports__DetrPreTrainedModel as DetrPreTrainedModel, __webpack_exports__DetrSegmentationOutput as DetrSegmentationOutput, __webpack_exports__Dinov2ForImageClassification as Dinov2ForImageClassification, __webpack_exports__Dinov2Model as Dinov2Model, __webpack_exports__Dinov2PreTrainedModel as Dinov2PreTrainedModel, __webpack_exports__DistilBertForMaskedLM as DistilBertForMaskedLM, __webpack_exports__DistilBertForQuestionAnswering as DistilBertForQuestionAnswering, __webpack_exports__DistilBertForSequenceClassification as DistilBertForSequenceClassification, __webpack_exports__DistilBertForTokenClassification as DistilBertForTokenClassification, __webpack_exports__DistilBertModel as DistilBertModel, __webpack_exports__DistilBertPreTrainedModel as DistilBertPreTrainedModel, __webpack_exports__DistilBertTokenizer as DistilBertTokenizer, __webpack_exports__DocumentQuestionAnsweringPipeline as DocumentQuestionAnsweringPipeline, __webpack_exports__DonutFeatureExtractor as DonutFeatureExtractor, __webpack_exports__DonutImageProcessor as DonutImageProcessor, __webpack_exports__DonutSwinModel as DonutSwinModel, __webpack_exports__DonutSwinPreTrainedModel as DonutSwinPreTrainedModel, __webpack_exports__EfficientNetForImageClassification as EfficientNetForImageClassification, __webpack_exports__EfficientNetImageProcessor as EfficientNetImageProcessor, __webpack_exports__EfficientNetModel as EfficientNetModel, __webpack_exports__EfficientNetPreTrainedModel as EfficientNetPreTrainedModel, __webpack_exports__ElectraForMaskedLM as ElectraForMaskedLM, __webpack_exports__ElectraForQuestionAnswering as ElectraForQuestionAnswering, __webpack_exports__ElectraForSequenceClassification as ElectraForSequenceClassification, __webpack_exports__ElectraForTokenClassification as ElectraForTokenClassification, __webpack_exports__ElectraModel as ElectraModel, __webpack_exports__ElectraPreTrainedModel as ElectraPreTrainedModel, __webpack_exports__ElectraTokenizer as ElectraTokenizer, __webpack_exports__EosTokenCriteria as EosTokenCriteria, __webpack_exports__EsmForMaskedLM as EsmForMaskedLM, __webpack_exports__EsmForSequenceClassification as EsmForSequenceClassification, __webpack_exports__EsmForTokenClassification as EsmForTokenClassification, __webpack_exports__EsmModel as EsmModel, __webpack_exports__EsmPreTrainedModel as EsmPreTrainedModel, __webpack_exports__EsmTokenizer as EsmTokenizer, __webpack_exports__ExaoneForCausalLM as ExaoneForCausalLM, __webpack_exports__ExaoneModel as ExaoneModel, __webpack_exports__ExaonePreTrainedModel as ExaonePreTrainedModel, __webpack_exports__FFT as FFT, __webpack_exports__FalconForCausalLM as FalconForCausalLM, __webpack_exports__FalconModel as FalconModel, __webpack_exports__FalconPreTrainedModel as FalconPreTrainedModel, __webpack_exports__FalconTokenizer as FalconTokenizer, __webpack_exports__FastViTForImageClassification as FastViTForImageClassification, __webpack_exports__FastViTModel as FastViTModel, __webpack_exports__FastViTPreTrainedModel as FastViTPreTrainedModel, __webpack_exports__FeatureExtractionPipeline as FeatureExtractionPipeline, __webpack_exports__FeatureExtractor as FeatureExtractor, __webpack_exports__FillMaskPipeline as FillMaskPipeline, __webpack_exports__Florence2ForConditionalGeneration as Florence2ForConditionalGeneration, __webpack_exports__Florence2PreTrainedModel as Florence2PreTrainedModel, __webpack_exports__Florence2Processor as Florence2Processor, __webpack_exports__ForcedBOSTokenLogitsProcessor as ForcedBOSTokenLogitsProcessor, __webpack_exports__ForcedEOSTokenLogitsProcessor as ForcedEOSTokenLogitsProcessor, __webpack_exports__GLPNFeatureExtractor as GLPNFeatureExtractor, __webpack_exports__GLPNForDepthEstimation as GLPNForDepthEstimation, __webpack_exports__GLPNModel as GLPNModel, __webpack_exports__GLPNPreTrainedModel as GLPNPreTrainedModel, __webpack_exports__GPT2LMHeadModel as GPT2LMHeadModel, __webpack_exports__GPT2Model as GPT2Model, __webpack_exports__GPT2PreTrainedModel as GPT2PreTrainedModel, __webpack_exports__GPT2Tokenizer as GPT2Tokenizer, __webpack_exports__GPTBigCodeForCausalLM as GPTBigCodeForCausalLM, __webpack_exports__GPTBigCodeModel as GPTBigCodeModel, __webpack_exports__GPTBigCodePreTrainedModel as GPTBigCodePreTrainedModel, __webpack_exports__GPTJForCausalLM as GPTJForCausalLM, __webpack_exports__GPTJModel as GPTJModel, __webpack_exports__GPTJPreTrainedModel as GPTJPreTrainedModel, __webpack_exports__GPTNeoForCausalLM as GPTNeoForCausalLM, __webpack_exports__GPTNeoModel as GPTNeoModel, __webpack_exports__GPTNeoPreTrainedModel as GPTNeoPreTrainedModel, __webpack_exports__GPTNeoXForCausalLM as GPTNeoXForCausalLM, __webpack_exports__GPTNeoXModel as GPTNeoXModel, __webpack_exports__GPTNeoXPreTrainedModel as GPTNeoXPreTrainedModel, __webpack_exports__GPTNeoXTokenizer as GPTNeoXTokenizer, __webpack_exports__Gemma2ForCausalLM as Gemma2ForCausalLM, __webpack_exports__Gemma2Model as Gemma2Model, __webpack_exports__Gemma2PreTrainedModel as Gemma2PreTrainedModel, __webpack_exports__GemmaForCausalLM as GemmaForCausalLM, __webpack_exports__GemmaModel as GemmaModel, __webpack_exports__GemmaPreTrainedModel as GemmaPreTrainedModel, __webpack_exports__GemmaTokenizer as GemmaTokenizer, __webpack_exports__GraniteForCausalLM as GraniteForCausalLM, __webpack_exports__GraniteModel as GraniteModel, __webpack_exports__GranitePreTrainedModel as GranitePreTrainedModel, __webpack_exports__Grok1Tokenizer as Grok1Tokenizer, __webpack_exports__GroupViTModel as GroupViTModel, __webpack_exports__GroupViTPreTrainedModel as GroupViTPreTrainedModel, __webpack_exports__HerbertTokenizer as HerbertTokenizer, __webpack_exports__HieraForImageClassification as HieraForImageClassification, __webpack_exports__HieraModel as HieraModel, __webpack_exports__HieraPreTrainedModel as HieraPreTrainedModel, __webpack_exports__HubertForCTC as HubertForCTC, __webpack_exports__HubertForSequenceClassification as HubertForSequenceClassification, __webpack_exports__HubertModel as HubertModel, __webpack_exports__HubertPreTrainedModel as HubertPreTrainedModel, __webpack_exports__IJepaForImageClassification as IJepaForImageClassification, __webpack_exports__IJepaModel as IJepaModel, __webpack_exports__IJepaPreTrainedModel as IJepaPreTrainedModel, __webpack_exports__Idefics3ForConditionalGeneration as Idefics3ForConditionalGeneration, __webpack_exports__Idefics3ImageProcessor as Idefics3ImageProcessor, __webpack_exports__Idefics3PreTrainedModel as Idefics3PreTrainedModel, __webpack_exports__Idefics3Processor as Idefics3Processor, __webpack_exports__ImageClassificationPipeline as ImageClassificationPipeline, __webpack_exports__ImageFeatureExtractionPipeline as ImageFeatureExtractionPipeline, __webpack_exports__ImageFeatureExtractor as ImageFeatureExtractor, __webpack_exports__ImageMattingOutput as ImageMattingOutput, __webpack_exports__ImageProcessor as ImageProcessor, __webpack_exports__ImageSegmentationPipeline as ImageSegmentationPipeline, __webpack_exports__ImageToImagePipeline as ImageToImagePipeline, __webpack_exports__ImageToTextPipeline as ImageToTextPipeline, __webpack_exports__InterruptableStoppingCriteria as InterruptableStoppingCriteria, __webpack_exports__JAISLMHeadModel as JAISLMHeadModel, __webpack_exports__JAISModel as JAISModel, __webpack_exports__JAISPreTrainedModel as JAISPreTrainedModel, __webpack_exports__JinaCLIPImageProcessor as JinaCLIPImageProcessor, __webpack_exports__JinaCLIPModel as JinaCLIPModel, __webpack_exports__JinaCLIPPreTrainedModel as JinaCLIPPreTrainedModel, __webpack_exports__JinaCLIPProcessor as JinaCLIPProcessor, __webpack_exports__JinaCLIPTextModel as JinaCLIPTextModel, __webpack_exports__JinaCLIPVisionModel as JinaCLIPVisionModel, __webpack_exports__LlamaForCausalLM as LlamaForCausalLM, __webpack_exports__LlamaModel as LlamaModel, __webpack_exports__LlamaPreTrainedModel as LlamaPreTrainedModel, __webpack_exports__LlamaTokenizer as LlamaTokenizer, __webpack_exports__LlavaForConditionalGeneration as LlavaForConditionalGeneration, __webpack_exports__LlavaOnevisionForConditionalGeneration as LlavaOnevisionForConditionalGeneration, __webpack_exports__LlavaOnevisionImageProcessor as LlavaOnevisionImageProcessor, __webpack_exports__LlavaPreTrainedModel as LlavaPreTrainedModel, __webpack_exports__LogitsProcessor as LogitsProcessor, __webpack_exports__LogitsProcessorList as LogitsProcessorList, __webpack_exports__LogitsWarper as LogitsWarper, __webpack_exports__LongT5ForConditionalGeneration as LongT5ForConditionalGeneration, __webpack_exports__LongT5Model as LongT5Model, __webpack_exports__LongT5PreTrainedModel as LongT5PreTrainedModel, __webpack_exports__M2M100ForConditionalGeneration as M2M100ForConditionalGeneration, __webpack_exports__M2M100Model as M2M100Model, __webpack_exports__M2M100PreTrainedModel as M2M100PreTrainedModel, __webpack_exports__M2M100Tokenizer as M2M100Tokenizer, __webpack_exports__MBart50Tokenizer as MBart50Tokenizer, __webpack_exports__MBartForCausalLM as MBartForCausalLM, __webpack_exports__MBartForConditionalGeneration as MBartForConditionalGeneration, __webpack_exports__MBartForSequenceClassification as MBartForSequenceClassification, __webpack_exports__MBartModel as MBartModel, __webpack_exports__MBartPreTrainedModel as MBartPreTrainedModel, __webpack_exports__MBartTokenizer as MBartTokenizer, __webpack_exports__MPNetForMaskedLM as MPNetForMaskedLM, __webpack_exports__MPNetForQuestionAnswering as MPNetForQuestionAnswering, __webpack_exports__MPNetForSequenceClassification as MPNetForSequenceClassification, __webpack_exports__MPNetForTokenClassification as MPNetForTokenClassification, __webpack_exports__MPNetModel as MPNetModel, __webpack_exports__MPNetPreTrainedModel as MPNetPreTrainedModel, __webpack_exports__MPNetTokenizer as MPNetTokenizer, __webpack_exports__MT5ForConditionalGeneration as MT5ForConditionalGeneration, __webpack_exports__MT5Model as MT5Model, __webpack_exports__MT5PreTrainedModel as MT5PreTrainedModel, __webpack_exports__MarianMTModel as MarianMTModel, __webpack_exports__MarianModel as MarianModel, __webpack_exports__MarianPreTrainedModel as MarianPreTrainedModel, __webpack_exports__MarianTokenizer as MarianTokenizer, __webpack_exports__Mask2FormerImageProcessor as Mask2FormerImageProcessor, __webpack_exports__MaskFormerFeatureExtractor as MaskFormerFeatureExtractor, __webpack_exports__MaskFormerForInstanceSegmentation as MaskFormerForInstanceSegmentation, __webpack_exports__MaskFormerImageProcessor as MaskFormerImageProcessor, __webpack_exports__MaskFormerModel as MaskFormerModel, __webpack_exports__MaskFormerPreTrainedModel as MaskFormerPreTrainedModel, __webpack_exports__MaskedLMOutput as MaskedLMOutput, __webpack_exports__MaxLengthCriteria as MaxLengthCriteria, __webpack_exports__MgpstrForSceneTextRecognition as MgpstrForSceneTextRecognition, __webpack_exports__MgpstrModelOutput as MgpstrModelOutput, __webpack_exports__MgpstrPreTrainedModel as MgpstrPreTrainedModel, __webpack_exports__MgpstrProcessor as MgpstrProcessor, __webpack_exports__MgpstrTokenizer as MgpstrTokenizer, __webpack_exports__MinLengthLogitsProcessor as MinLengthLogitsProcessor, __webpack_exports__MinNewTokensLengthLogitsProcessor as MinNewTokensLengthLogitsProcessor, __webpack_exports__MistralForCausalLM as MistralForCausalLM, __webpack_exports__MistralModel as MistralModel, __webpack_exports__MistralPreTrainedModel as MistralPreTrainedModel, __webpack_exports__MobileBertForMaskedLM as MobileBertForMaskedLM, __webpack_exports__MobileBertForQuestionAnswering as MobileBertForQuestionAnswering, __webpack_exports__MobileBertForSequenceClassification as MobileBertForSequenceClassification, __webpack_exports__MobileBertModel as MobileBertModel, __webpack_exports__MobileBertPreTrainedModel as MobileBertPreTrainedModel, __webpack_exports__MobileBertTokenizer as MobileBertTokenizer, __webpack_exports__MobileLLMForCausalLM as MobileLLMForCausalLM, __webpack_exports__MobileLLMModel as MobileLLMModel, __webpack_exports__MobileLLMPreTrainedModel as MobileLLMPreTrainedModel, __webpack_exports__MobileNetV1FeatureExtractor as MobileNetV1FeatureExtractor, __webpack_exports__MobileNetV1ForImageClassification as MobileNetV1ForImageClassification, __webpack_exports__MobileNetV1ImageProcessor as MobileNetV1ImageProcessor, __webpack_exports__MobileNetV1Model as MobileNetV1Model, __webpack_exports__MobileNetV1PreTrainedModel as MobileNetV1PreTrainedModel, __webpack_exports__MobileNetV2FeatureExtractor as MobileNetV2FeatureExtractor, __webpack_exports__MobileNetV2ForImageClassification as MobileNetV2ForImageClassification, __webpack_exports__MobileNetV2ImageProcessor as MobileNetV2ImageProcessor, __webpack_exports__MobileNetV2Model as MobileNetV2Model, __webpack_exports__MobileNetV2PreTrainedModel as MobileNetV2PreTrainedModel, __webpack_exports__MobileNetV3FeatureExtractor as MobileNetV3FeatureExtractor, __webpack_exports__MobileNetV3ForImageClassification as MobileNetV3ForImageClassification, __webpack_exports__MobileNetV3ImageProcessor as MobileNetV3ImageProcessor, __webpack_exports__MobileNetV3Model as MobileNetV3Model, __webpack_exports__MobileNetV3PreTrainedModel as MobileNetV3PreTrainedModel, __webpack_exports__MobileNetV4FeatureExtractor as MobileNetV4FeatureExtractor, __webpack_exports__MobileNetV4ForImageClassification as MobileNetV4ForImageClassification, __webpack_exports__MobileNetV4ImageProcessor as MobileNetV4ImageProcessor, __webpack_exports__MobileNetV4Model as MobileNetV4Model, __webpack_exports__MobileNetV4PreTrainedModel as MobileNetV4PreTrainedModel, __webpack_exports__MobileViTFeatureExtractor as MobileViTFeatureExtractor, __webpack_exports__MobileViTForImageClassification as MobileViTForImageClassification, __webpack_exports__MobileViTImageProcessor as MobileViTImageProcessor, __webpack_exports__MobileViTModel as MobileViTModel, __webpack_exports__MobileViTPreTrainedModel as MobileViTPreTrainedModel, __webpack_exports__MobileViTV2ForImageClassification as MobileViTV2ForImageClassification, __webpack_exports__MobileViTV2Model as MobileViTV2Model, __webpack_exports__MobileViTV2PreTrainedModel as MobileViTV2PreTrainedModel, __webpack_exports__ModelOutput as ModelOutput, __webpack_exports__ModernBertForMaskedLM as ModernBertForMaskedLM, __webpack_exports__ModernBertForSequenceClassification as ModernBertForSequenceClassification, __webpack_exports__ModernBertForTokenClassification as ModernBertForTokenClassification, __webpack_exports__ModernBertModel as ModernBertModel, __webpack_exports__ModernBertPreTrainedModel as ModernBertPreTrainedModel, __webpack_exports__Moondream1ForConditionalGeneration as Moondream1ForConditionalGeneration, __webpack_exports__MoonshineFeatureExtractor as MoonshineFeatureExtractor, __webpack_exports__MoonshineForConditionalGeneration as MoonshineForConditionalGeneration, __webpack_exports__MoonshineModel as MoonshineModel, __webpack_exports__MoonshinePreTrainedModel as MoonshinePreTrainedModel, __webpack_exports__MoonshineProcessor as MoonshineProcessor, __webpack_exports__MptForCausalLM as MptForCausalLM, __webpack_exports__MptModel as MptModel, __webpack_exports__MptPreTrainedModel as MptPreTrainedModel, __webpack_exports__MultiModalityCausalLM as MultiModalityCausalLM, __webpack_exports__MultiModalityPreTrainedModel as MultiModalityPreTrainedModel, __webpack_exports__MusicgenForCausalLM as MusicgenForCausalLM, __webpack_exports__MusicgenForConditionalGeneration as MusicgenForConditionalGeneration, __webpack_exports__MusicgenModel as MusicgenModel, __webpack_exports__MusicgenPreTrainedModel as MusicgenPreTrainedModel, __webpack_exports__NllbTokenizer as NllbTokenizer, __webpack_exports__NoBadWordsLogitsProcessor as NoBadWordsLogitsProcessor, __webpack_exports__NoRepeatNGramLogitsProcessor as NoRepeatNGramLogitsProcessor, __webpack_exports__NomicBertModel as NomicBertModel, __webpack_exports__NomicBertPreTrainedModel as NomicBertPreTrainedModel, __webpack_exports__NougatImageProcessor as NougatImageProcessor, __webpack_exports__NougatTokenizer as NougatTokenizer, __webpack_exports__OPTForCausalLM as OPTForCausalLM, __webpack_exports__OPTModel as OPTModel, __webpack_exports__OPTPreTrainedModel as OPTPreTrainedModel, __webpack_exports__ObjectDetectionPipeline as ObjectDetectionPipeline, __webpack_exports__Olmo2ForCausalLM as Olmo2ForCausalLM, __webpack_exports__Olmo2Model as Olmo2Model, __webpack_exports__Olmo2PreTrainedModel as Olmo2PreTrainedModel, __webpack_exports__OlmoForCausalLM as OlmoForCausalLM, __webpack_exports__OlmoModel as OlmoModel, __webpack_exports__OlmoPreTrainedModel as OlmoPreTrainedModel, __webpack_exports__OpenELMForCausalLM as OpenELMForCausalLM, __webpack_exports__OpenELMModel as OpenELMModel, __webpack_exports__OpenELMPreTrainedModel as OpenELMPreTrainedModel, __webpack_exports__OwlViTFeatureExtractor as OwlViTFeatureExtractor, __webpack_exports__OwlViTForObjectDetection as OwlViTForObjectDetection, __webpack_exports__OwlViTImageProcessor as OwlViTImageProcessor, __webpack_exports__OwlViTModel as OwlViTModel, __webpack_exports__OwlViTPreTrainedModel as OwlViTPreTrainedModel, __webpack_exports__OwlViTProcessor as OwlViTProcessor, __webpack_exports__Owlv2ForObjectDetection as Owlv2ForObjectDetection, __webpack_exports__Owlv2ImageProcessor as Owlv2ImageProcessor, __webpack_exports__Owlv2Model as Owlv2Model, __webpack_exports__Owlv2PreTrainedModel as Owlv2PreTrainedModel, __webpack_exports__PaliGemmaForConditionalGeneration as PaliGemmaForConditionalGeneration, __webpack_exports__PaliGemmaPreTrainedModel as PaliGemmaPreTrainedModel, __webpack_exports__PaliGemmaProcessor as PaliGemmaProcessor, __webpack_exports__PatchTSMixerForPrediction as PatchTSMixerForPrediction, __webpack_exports__PatchTSMixerModel as PatchTSMixerModel, __webpack_exports__PatchTSMixerPreTrainedModel as PatchTSMixerPreTrainedModel, __webpack_exports__PatchTSTForPrediction as PatchTSTForPrediction, __webpack_exports__PatchTSTModel as PatchTSTModel, __webpack_exports__PatchTSTPreTrainedModel as PatchTSTPreTrainedModel, __webpack_exports__Phi3ForCausalLM as Phi3ForCausalLM, __webpack_exports__Phi3Model as Phi3Model, __webpack_exports__Phi3PreTrainedModel as Phi3PreTrainedModel, __webpack_exports__Phi3VForCausalLM as Phi3VForCausalLM, __webpack_exports__Phi3VImageProcessor as Phi3VImageProcessor, __webpack_exports__Phi3VPreTrainedModel as Phi3VPreTrainedModel, __webpack_exports__Phi3VProcessor as Phi3VProcessor, __webpack_exports__PhiForCausalLM as PhiForCausalLM, __webpack_exports__PhiModel as PhiModel, __webpack_exports__PhiPreTrainedModel as PhiPreTrainedModel, __webpack_exports__Pipeline as Pipeline, __webpack_exports__PreTrainedModel as PreTrainedModel, __webpack_exports__PreTrainedTokenizer as PreTrainedTokenizer, __webpack_exports__PretrainedConfig as PretrainedConfig, __webpack_exports__PretrainedMixin as PretrainedMixin, __webpack_exports__Processor as Processor, __webpack_exports__PvtForImageClassification as PvtForImageClassification, __webpack_exports__PvtImageProcessor as PvtImageProcessor, __webpack_exports__PvtModel as PvtModel, __webpack_exports__PvtPreTrainedModel as PvtPreTrainedModel, __webpack_exports__PyAnnoteFeatureExtractor as PyAnnoteFeatureExtractor, __webpack_exports__PyAnnoteForAudioFrameClassification as PyAnnoteForAudioFrameClassification, __webpack_exports__PyAnnoteModel as PyAnnoteModel, __webpack_exports__PyAnnotePreTrainedModel as PyAnnotePreTrainedModel, __webpack_exports__PyAnnoteProcessor as PyAnnoteProcessor, __webpack_exports__QuestionAnsweringModelOutput as QuestionAnsweringModelOutput, __webpack_exports__QuestionAnsweringPipeline as QuestionAnsweringPipeline, __webpack_exports__Qwen2ForCausalLM as Qwen2ForCausalLM, __webpack_exports__Qwen2Model as Qwen2Model, __webpack_exports__Qwen2PreTrainedModel as Qwen2PreTrainedModel, __webpack_exports__Qwen2Tokenizer as Qwen2Tokenizer, __webpack_exports__Qwen2VLForConditionalGeneration as Qwen2VLForConditionalGeneration, __webpack_exports__Qwen2VLImageProcessor as Qwen2VLImageProcessor, __webpack_exports__Qwen2VLPreTrainedModel as Qwen2VLPreTrainedModel, __webpack_exports__Qwen2VLProcessor as Qwen2VLProcessor, __webpack_exports__RTDetrForObjectDetection as RTDetrForObjectDetection, __webpack_exports__RTDetrImageProcessor as RTDetrImageProcessor, __webpack_exports__RTDetrModel as RTDetrModel, __webpack_exports__RTDetrObjectDetectionOutput as RTDetrObjectDetectionOutput, __webpack_exports__RTDetrPreTrainedModel as RTDetrPreTrainedModel, __webpack_exports__RawImage as RawImage, __webpack_exports__RepetitionPenaltyLogitsProcessor as RepetitionPenaltyLogitsProcessor, __webpack_exports__ResNetForImageClassification as ResNetForImageClassification, __webpack_exports__ResNetModel as ResNetModel, __webpack_exports__ResNetPreTrainedModel as ResNetPreTrainedModel, __webpack_exports__RoFormerForMaskedLM as RoFormerForMaskedLM, __webpack_exports__RoFormerForQuestionAnswering as RoFormerForQuestionAnswering, __webpack_exports__RoFormerForSequenceClassification as RoFormerForSequenceClassification, __webpack_exports__RoFormerForTokenClassification as RoFormerForTokenClassification, __webpack_exports__RoFormerModel as RoFormerModel, __webpack_exports__RoFormerPreTrainedModel as RoFormerPreTrainedModel, __webpack_exports__RoFormerTokenizer as RoFormerTokenizer, __webpack_exports__RobertaForMaskedLM as RobertaForMaskedLM, __webpack_exports__RobertaForQuestionAnswering as RobertaForQuestionAnswering, __webpack_exports__RobertaForSequenceClassification as RobertaForSequenceClassification, __webpack_exports__RobertaForTokenClassification as RobertaForTokenClassification, __webpack_exports__RobertaModel as RobertaModel, __webpack_exports__RobertaPreTrainedModel as RobertaPreTrainedModel, __webpack_exports__RobertaTokenizer as RobertaTokenizer, __webpack_exports__SamImageProcessor as SamImageProcessor, __webpack_exports__SamImageSegmentationOutput as SamImageSegmentationOutput, __webpack_exports__SamModel as SamModel, __webpack_exports__SamPreTrainedModel as SamPreTrainedModel, __webpack_exports__SamProcessor as SamProcessor, __webpack_exports__SapiensForDepthEstimation as SapiensForDepthEstimation, __webpack_exports__SapiensForNormalEstimation as SapiensForNormalEstimation, __webpack_exports__SapiensForSemanticSegmentation as SapiensForSemanticSegmentation, __webpack_exports__SapiensPreTrainedModel as SapiensPreTrainedModel, __webpack_exports__SeamlessM4TFeatureExtractor as SeamlessM4TFeatureExtractor, __webpack_exports__SegformerFeatureExtractor as SegformerFeatureExtractor, __webpack_exports__SegformerForImageClassification as SegformerForImageClassification, __webpack_exports__SegformerForSemanticSegmentation as SegformerForSemanticSegmentation, __webpack_exports__SegformerImageProcessor as SegformerImageProcessor, __webpack_exports__SegformerModel as SegformerModel, __webpack_exports__SegformerPreTrainedModel as SegformerPreTrainedModel, __webpack_exports__Seq2SeqLMOutput as Seq2SeqLMOutput, __webpack_exports__SequenceClassifierOutput as SequenceClassifierOutput, __webpack_exports__SiglipImageProcessor as SiglipImageProcessor, __webpack_exports__SiglipModel as SiglipModel, __webpack_exports__SiglipPreTrainedModel as SiglipPreTrainedModel, __webpack_exports__SiglipTextModel as SiglipTextModel, __webpack_exports__SiglipTokenizer as SiglipTokenizer, __webpack_exports__SiglipVisionModel as SiglipVisionModel, __webpack_exports__SpeechT5FeatureExtractor as SpeechT5FeatureExtractor, __webpack_exports__SpeechT5ForSpeechToText as SpeechT5ForSpeechToText, __webpack_exports__SpeechT5ForTextToSpeech as SpeechT5ForTextToSpeech, __webpack_exports__SpeechT5HifiGan as SpeechT5HifiGan, __webpack_exports__SpeechT5Model as SpeechT5Model, __webpack_exports__SpeechT5PreTrainedModel as SpeechT5PreTrainedModel, __webpack_exports__SpeechT5Processor as SpeechT5Processor, __webpack_exports__SpeechT5Tokenizer as SpeechT5Tokenizer, __webpack_exports__SqueezeBertForMaskedLM as SqueezeBertForMaskedLM, __webpack_exports__SqueezeBertForQuestionAnswering as SqueezeBertForQuestionAnswering, __webpack_exports__SqueezeBertForSequenceClassification as SqueezeBertForSequenceClassification, __webpack_exports__SqueezeBertModel as SqueezeBertModel, __webpack_exports__SqueezeBertPreTrainedModel as SqueezeBertPreTrainedModel, __webpack_exports__SqueezeBertTokenizer as SqueezeBertTokenizer, __webpack_exports__StableLmForCausalLM as StableLmForCausalLM, __webpack_exports__StableLmModel as StableLmModel, __webpack_exports__StableLmPreTrainedModel as StableLmPreTrainedModel, __webpack_exports__Starcoder2ForCausalLM as Starcoder2ForCausalLM, __webpack_exports__Starcoder2Model as Starcoder2Model, __webpack_exports__Starcoder2PreTrainedModel as Starcoder2PreTrainedModel, __webpack_exports__StoppingCriteria as StoppingCriteria, __webpack_exports__StoppingCriteriaList as StoppingCriteriaList, __webpack_exports__SummarizationPipeline as SummarizationPipeline, __webpack_exports__SuppressTokensAtBeginLogitsProcessor as SuppressTokensAtBeginLogitsProcessor, __webpack_exports__Swin2SRForImageSuperResolution as Swin2SRForImageSuperResolution, __webpack_exports__Swin2SRImageProcessor as Swin2SRImageProcessor, __webpack_exports__Swin2SRModel as Swin2SRModel, __webpack_exports__Swin2SRPreTrainedModel as Swin2SRPreTrainedModel, __webpack_exports__SwinForImageClassification as SwinForImageClassification, __webpack_exports__SwinModel as SwinModel, __webpack_exports__SwinPreTrainedModel as SwinPreTrainedModel, __webpack_exports__T5ForConditionalGeneration as T5ForConditionalGeneration, __webpack_exports__T5Model as T5Model, __webpack_exports__T5PreTrainedModel as T5PreTrainedModel, __webpack_exports__T5Tokenizer as T5Tokenizer, __webpack_exports__TableTransformerForObjectDetection as TableTransformerForObjectDetection, __webpack_exports__TableTransformerModel as TableTransformerModel, __webpack_exports__TableTransformerObjectDetectionOutput as TableTransformerObjectDetectionOutput, __webpack_exports__TableTransformerPreTrainedModel as TableTransformerPreTrainedModel, __webpack_exports__TemperatureLogitsWarper as TemperatureLogitsWarper, __webpack_exports__Tensor as Tensor, __webpack_exports__Text2TextGenerationPipeline as Text2TextGenerationPipeline, __webpack_exports__TextClassificationPipeline as TextClassificationPipeline, __webpack_exports__TextGenerationPipeline as TextGenerationPipeline, __webpack_exports__TextStreamer as TextStreamer, __webpack_exports__TextToAudioPipeline as TextToAudioPipeline, __webpack_exports__TokenClassificationPipeline as TokenClassificationPipeline, __webpack_exports__TokenClassifierOutput as TokenClassifierOutput, __webpack_exports__TokenizerModel as TokenizerModel, __webpack_exports__TopKLogitsWarper as TopKLogitsWarper, __webpack_exports__TopPLogitsWarper as TopPLogitsWarper, __webpack_exports__TrOCRForCausalLM as TrOCRForCausalLM, __webpack_exports__TrOCRPreTrainedModel as TrOCRPreTrainedModel, __webpack_exports__TranslationPipeline as TranslationPipeline, __webpack_exports__UniSpeechForCTC as UniSpeechForCTC, __webpack_exports__UniSpeechForSequenceClassification as UniSpeechForSequenceClassification, __webpack_exports__UniSpeechModel as UniSpeechModel, __webpack_exports__UniSpeechPreTrainedModel as UniSpeechPreTrainedModel, __webpack_exports__UniSpeechSatForAudioFrameClassification as UniSpeechSatForAudioFrameClassification, __webpack_exports__UniSpeechSatForCTC as UniSpeechSatForCTC, __webpack_exports__UniSpeechSatForSequenceClassification as UniSpeechSatForSequenceClassification, __webpack_exports__UniSpeechSatModel as UniSpeechSatModel, __webpack_exports__UniSpeechSatPreTrainedModel as UniSpeechSatPreTrainedModel, __webpack_exports__VLChatProcessor as VLChatProcessor, __webpack_exports__VLMImageProcessor as VLMImageProcessor, __webpack_exports__ViTFeatureExtractor as ViTFeatureExtractor, __webpack_exports__ViTForImageClassification as ViTForImageClassification, __webpack_exports__ViTImageProcessor as ViTImageProcessor, __webpack_exports__ViTMAEModel as ViTMAEModel, __webpack_exports__ViTMAEPreTrainedModel as ViTMAEPreTrainedModel, __webpack_exports__ViTMSNForImageClassification as ViTMSNForImageClassification, __webpack_exports__ViTMSNModel as ViTMSNModel, __webpack_exports__ViTMSNPreTrainedModel as ViTMSNPreTrainedModel, __webpack_exports__ViTModel as ViTModel, __webpack_exports__ViTPreTrainedModel as ViTPreTrainedModel, __webpack_exports__VisionEncoderDecoderModel as VisionEncoderDecoderModel, __webpack_exports__VitMatteForImageMatting as VitMatteForImageMatting, __webpack_exports__VitMatteImageProcessor as VitMatteImageProcessor, __webpack_exports__VitMattePreTrainedModel as VitMattePreTrainedModel, __webpack_exports__VitPoseForPoseEstimation as VitPoseForPoseEstimation, __webpack_exports__VitPoseImageProcessor as VitPoseImageProcessor, __webpack_exports__VitPosePreTrainedModel as VitPosePreTrainedModel, __webpack_exports__VitsModel as VitsModel, __webpack_exports__VitsModelOutput as VitsModelOutput, __webpack_exports__VitsPreTrainedModel as VitsPreTrainedModel, __webpack_exports__VitsTokenizer as VitsTokenizer, __webpack_exports__Wav2Vec2BertForCTC as Wav2Vec2BertForCTC, __webpack_exports__Wav2Vec2BertForSequenceClassification as Wav2Vec2BertForSequenceClassification, __webpack_exports__Wav2Vec2BertModel as Wav2Vec2BertModel, __webpack_exports__Wav2Vec2BertPreTrainedModel as Wav2Vec2BertPreTrainedModel, __webpack_exports__Wav2Vec2CTCTokenizer as Wav2Vec2CTCTokenizer, __webpack_exports__Wav2Vec2FeatureExtractor as Wav2Vec2FeatureExtractor, __webpack_exports__Wav2Vec2ForAudioFrameClassification as Wav2Vec2ForAudioFrameClassification, __webpack_exports__Wav2Vec2ForCTC as Wav2Vec2ForCTC, __webpack_exports__Wav2Vec2ForSequenceClassification as Wav2Vec2ForSequenceClassification, __webpack_exports__Wav2Vec2Model as Wav2Vec2Model, __webpack_exports__Wav2Vec2PreTrainedModel as Wav2Vec2PreTrainedModel, __webpack_exports__Wav2Vec2ProcessorWithLM as Wav2Vec2ProcessorWithLM, __webpack_exports__WavLMForAudioFrameClassification as WavLMForAudioFrameClassification, __webpack_exports__WavLMForCTC as WavLMForCTC, __webpack_exports__WavLMForSequenceClassification as WavLMForSequenceClassification, __webpack_exports__WavLMForXVector as WavLMForXVector, __webpack_exports__WavLMModel as WavLMModel, __webpack_exports__WavLMPreTrainedModel as WavLMPreTrainedModel, __webpack_exports__WeSpeakerFeatureExtractor as WeSpeakerFeatureExtractor, __webpack_exports__WeSpeakerResNetModel as WeSpeakerResNetModel, __webpack_exports__WeSpeakerResNetPreTrainedModel as WeSpeakerResNetPreTrainedModel, __webpack_exports__WhisperFeatureExtractor as WhisperFeatureExtractor, __webpack_exports__WhisperForConditionalGeneration as WhisperForConditionalGeneration, __webpack_exports__WhisperModel as WhisperModel, __webpack_exports__WhisperPreTrainedModel as WhisperPreTrainedModel, __webpack_exports__WhisperProcessor as WhisperProcessor, __webpack_exports__WhisperTextStreamer as WhisperTextStreamer, __webpack_exports__WhisperTimeStampLogitsProcessor as WhisperTimeStampLogitsProcessor, __webpack_exports__WhisperTokenizer as WhisperTokenizer, __webpack_exports__XLMForQuestionAnswering as XLMForQuestionAnswering, __webpack_exports__XLMForSequenceClassification as XLMForSequenceClassification, __webpack_exports__XLMForTokenClassification as XLMForTokenClassification, __webpack_exports__XLMModel as XLMModel, __webpack_exports__XLMPreTrainedModel as XLMPreTrainedModel, __webpack_exports__XLMRobertaForMaskedLM as XLMRobertaForMaskedLM, __webpack_exports__XLMRobertaForQuestionAnswering as XLMRobertaForQuestionAnswering, __webpack_exports__XLMRobertaForSequenceClassification as XLMRobertaForSequenceClassification, __webpack_exports__XLMRobertaForTokenClassification as XLMRobertaForTokenClassification, __webpack_exports__XLMRobertaModel as XLMRobertaModel, __webpack_exports__XLMRobertaPreTrainedModel as XLMRobertaPreTrainedModel, __webpack_exports__XLMRobertaTokenizer as XLMRobertaTokenizer, __webpack_exports__XLMTokenizer as XLMTokenizer, __webpack_exports__XLMWithLMHeadModel as XLMWithLMHeadModel, __webpack_exports__XVectorOutput as XVectorOutput, __webpack_exports__YolosFeatureExtractor as YolosFeatureExtractor, __webpack_exports__YolosForObjectDetection as YolosForObjectDetection, __webpack_exports__YolosImageProcessor as YolosImageProcessor, __webpack_exports__YolosModel as YolosModel, __webpack_exports__YolosObjectDetectionOutput as YolosObjectDetectionOutput, __webpack_exports__YolosPreTrainedModel as YolosPreTrainedModel, __webpack_exports__ZeroShotAudioClassificationPipeline as ZeroShotAudioClassificationPipeline, __webpack_exports__ZeroShotClassificationPipeline as ZeroShotClassificationPipeline, __webpack_exports__ZeroShotImageClassificationPipeline as ZeroShotImageClassificationPipeline, __webpack_exports__ZeroShotObjectDetectionPipeline as ZeroShotObjectDetectionPipeline, __webpack_exports__bankers_round as bankers_round, __webpack_exports__cat as cat, __webpack_exports__cos_sim as cos_sim, __webpack_exports__dot as dot, __webpack_exports__dynamic_time_warping as dynamic_time_warping, __webpack_exports__env as env, __webpack_exports__full as full, __webpack_exports__full_like as full_like, __webpack_exports__getKeyValueShapes as getKeyValueShapes, __webpack_exports__hamming as hamming, __webpack_exports__hanning as hanning, __webpack_exports__interpolate as interpolate, __webpack_exports__interpolate_4d as interpolate_4d, __webpack_exports__interpolate_data as interpolate_data, __webpack_exports__is_chinese_char as is_chinese_char, __webpack_exports__layer_norm as layer_norm, __webpack_exports__load_image as load_image, __webpack_exports__log_softmax as log_softmax, __webpack_exports__magnitude as magnitude, __webpack_exports__matmul as matmul, __webpack_exports__max as max, __webpack_exports__mean as mean, __webpack_exports__mean_pooling as mean_pooling, __webpack_exports__medianFilter as medianFilter, __webpack_exports__mel_filter_bank as mel_filter_bank, __webpack_exports__min as min, __webpack_exports__ones as ones, __webpack_exports__ones_like as ones_like, __webpack_exports__permute as permute, __webpack_exports__permute_data as permute_data, __webpack_exports__pipeline as pipeline, __webpack_exports__quantize_embeddings as quantize_embeddings, __webpack_exports__rand as rand, __webpack_exports__read_audio as read_audio, __webpack_exports__rfft as rfft, __webpack_exports__round as round, __webpack_exports__slice as slice, __webpack_exports__softmax as softmax, __webpack_exports__spectrogram as spectrogram, __webpack_exports__stack as stack, __webpack_exports__std_mean as std_mean, __webpack_exports__topk as topk, __webpack_exports__window_function as window_function, __webpack_exports__zeros as zeros, __webpack_exports__zeros_like as zeros_like };
|
|
36124
|
+
export { __webpack_exports__ASTFeatureExtractor as ASTFeatureExtractor, __webpack_exports__ASTForAudioClassification as ASTForAudioClassification, __webpack_exports__ASTModel as ASTModel, __webpack_exports__ASTPreTrainedModel as ASTPreTrainedModel, __webpack_exports__AlbertForMaskedLM as AlbertForMaskedLM, __webpack_exports__AlbertForQuestionAnswering as AlbertForQuestionAnswering, __webpack_exports__AlbertForSequenceClassification as AlbertForSequenceClassification, __webpack_exports__AlbertModel as AlbertModel, __webpack_exports__AlbertPreTrainedModel as AlbertPreTrainedModel, __webpack_exports__AlbertTokenizer as AlbertTokenizer, __webpack_exports__AudioClassificationPipeline as AudioClassificationPipeline, __webpack_exports__AutoConfig as AutoConfig, __webpack_exports__AutoFeatureExtractor as AutoFeatureExtractor, __webpack_exports__AutoImageProcessor as AutoImageProcessor, __webpack_exports__AutoModel as AutoModel, __webpack_exports__AutoModelForAudioClassification as AutoModelForAudioClassification, __webpack_exports__AutoModelForAudioFrameClassification as AutoModelForAudioFrameClassification, __webpack_exports__AutoModelForCTC as AutoModelForCTC, __webpack_exports__AutoModelForCausalLM as AutoModelForCausalLM, __webpack_exports__AutoModelForDepthEstimation as AutoModelForDepthEstimation, __webpack_exports__AutoModelForDocumentQuestionAnswering as AutoModelForDocumentQuestionAnswering, __webpack_exports__AutoModelForImageClassification as AutoModelForImageClassification, __webpack_exports__AutoModelForImageFeatureExtraction as AutoModelForImageFeatureExtraction, __webpack_exports__AutoModelForImageMatting as AutoModelForImageMatting, __webpack_exports__AutoModelForImageSegmentation as AutoModelForImageSegmentation, __webpack_exports__AutoModelForImageToImage as AutoModelForImageToImage, __webpack_exports__AutoModelForMaskGeneration as AutoModelForMaskGeneration, __webpack_exports__AutoModelForMaskedLM as AutoModelForMaskedLM, __webpack_exports__AutoModelForNormalEstimation as AutoModelForNormalEstimation, __webpack_exports__AutoModelForObjectDetection as AutoModelForObjectDetection, __webpack_exports__AutoModelForPoseEstimation as AutoModelForPoseEstimation, __webpack_exports__AutoModelForQuestionAnswering as AutoModelForQuestionAnswering, __webpack_exports__AutoModelForSemanticSegmentation as AutoModelForSemanticSegmentation, __webpack_exports__AutoModelForSeq2SeqLM as AutoModelForSeq2SeqLM, __webpack_exports__AutoModelForSequenceClassification as AutoModelForSequenceClassification, __webpack_exports__AutoModelForSpeechSeq2Seq as AutoModelForSpeechSeq2Seq, __webpack_exports__AutoModelForTextToSpectrogram as AutoModelForTextToSpectrogram, __webpack_exports__AutoModelForTextToWaveform as AutoModelForTextToWaveform, __webpack_exports__AutoModelForTokenClassification as AutoModelForTokenClassification, __webpack_exports__AutoModelForUniversalSegmentation as AutoModelForUniversalSegmentation, __webpack_exports__AutoModelForVision2Seq as AutoModelForVision2Seq, __webpack_exports__AutoModelForXVector as AutoModelForXVector, __webpack_exports__AutoModelForZeroShotObjectDetection as AutoModelForZeroShotObjectDetection, __webpack_exports__AutoProcessor as AutoProcessor, __webpack_exports__AutoTokenizer as AutoTokenizer, __webpack_exports__AutomaticSpeechRecognitionPipeline as AutomaticSpeechRecognitionPipeline, __webpack_exports__BartForConditionalGeneration as BartForConditionalGeneration, __webpack_exports__BartForSequenceClassification as BartForSequenceClassification, __webpack_exports__BartModel as BartModel, __webpack_exports__BartPretrainedModel as BartPretrainedModel, __webpack_exports__BartTokenizer as BartTokenizer, __webpack_exports__BaseModelOutput as BaseModelOutput, __webpack_exports__BaseStreamer as BaseStreamer, __webpack_exports__BeitFeatureExtractor as BeitFeatureExtractor, __webpack_exports__BeitForImageClassification as BeitForImageClassification, __webpack_exports__BeitModel as BeitModel, __webpack_exports__BeitPreTrainedModel as BeitPreTrainedModel, __webpack_exports__BertForMaskedLM as BertForMaskedLM, __webpack_exports__BertForQuestionAnswering as BertForQuestionAnswering, __webpack_exports__BertForSequenceClassification as BertForSequenceClassification, __webpack_exports__BertForTokenClassification as BertForTokenClassification, __webpack_exports__BertModel as BertModel, __webpack_exports__BertPreTrainedModel as BertPreTrainedModel, __webpack_exports__BertTokenizer as BertTokenizer, __webpack_exports__BitImageProcessor as BitImageProcessor, __webpack_exports__BlenderbotForConditionalGeneration as BlenderbotForConditionalGeneration, __webpack_exports__BlenderbotModel as BlenderbotModel, __webpack_exports__BlenderbotPreTrainedModel as BlenderbotPreTrainedModel, __webpack_exports__BlenderbotSmallForConditionalGeneration as BlenderbotSmallForConditionalGeneration, __webpack_exports__BlenderbotSmallModel as BlenderbotSmallModel, __webpack_exports__BlenderbotSmallPreTrainedModel as BlenderbotSmallPreTrainedModel, __webpack_exports__BlenderbotSmallTokenizer as BlenderbotSmallTokenizer, __webpack_exports__BlenderbotTokenizer as BlenderbotTokenizer, __webpack_exports__BloomForCausalLM as BloomForCausalLM, __webpack_exports__BloomModel as BloomModel, __webpack_exports__BloomPreTrainedModel as BloomPreTrainedModel, __webpack_exports__BloomTokenizer as BloomTokenizer, __webpack_exports__CLIPFeatureExtractor as CLIPFeatureExtractor, __webpack_exports__CLIPImageProcessor as CLIPImageProcessor, __webpack_exports__CLIPModel as CLIPModel, __webpack_exports__CLIPPreTrainedModel as CLIPPreTrainedModel, __webpack_exports__CLIPSegForImageSegmentation as CLIPSegForImageSegmentation, __webpack_exports__CLIPSegModel as CLIPSegModel, __webpack_exports__CLIPSegPreTrainedModel as CLIPSegPreTrainedModel, __webpack_exports__CLIPTextModel as CLIPTextModel, __webpack_exports__CLIPTextModelWithProjection as CLIPTextModelWithProjection, __webpack_exports__CLIPTokenizer as CLIPTokenizer, __webpack_exports__CLIPVisionModel as CLIPVisionModel, __webpack_exports__CLIPVisionModelWithProjection as CLIPVisionModelWithProjection, __webpack_exports__CamembertForMaskedLM as CamembertForMaskedLM, __webpack_exports__CamembertForQuestionAnswering as CamembertForQuestionAnswering, __webpack_exports__CamembertForSequenceClassification as CamembertForSequenceClassification, __webpack_exports__CamembertForTokenClassification as CamembertForTokenClassification, __webpack_exports__CamembertModel as CamembertModel, __webpack_exports__CamembertPreTrainedModel as CamembertPreTrainedModel, __webpack_exports__CamembertTokenizer as CamembertTokenizer, __webpack_exports__CausalLMOutput as CausalLMOutput, __webpack_exports__CausalLMOutputWithPast as CausalLMOutputWithPast, __webpack_exports__ChineseCLIPFeatureExtractor as ChineseCLIPFeatureExtractor, __webpack_exports__ChineseCLIPModel as ChineseCLIPModel, __webpack_exports__ChineseCLIPPreTrainedModel as ChineseCLIPPreTrainedModel, __webpack_exports__ClapAudioModelWithProjection as ClapAudioModelWithProjection, __webpack_exports__ClapFeatureExtractor as ClapFeatureExtractor, __webpack_exports__ClapModel as ClapModel, __webpack_exports__ClapPreTrainedModel as ClapPreTrainedModel, __webpack_exports__ClapTextModelWithProjection as ClapTextModelWithProjection, __webpack_exports__ClassifierFreeGuidanceLogitsProcessor as ClassifierFreeGuidanceLogitsProcessor, __webpack_exports__CodeGenForCausalLM as CodeGenForCausalLM, __webpack_exports__CodeGenModel as CodeGenModel, __webpack_exports__CodeGenPreTrainedModel as CodeGenPreTrainedModel, __webpack_exports__CodeGenTokenizer as CodeGenTokenizer, __webpack_exports__CodeLlamaTokenizer as CodeLlamaTokenizer, __webpack_exports__CohereForCausalLM as CohereForCausalLM, __webpack_exports__CohereModel as CohereModel, __webpack_exports__CoherePreTrainedModel as CoherePreTrainedModel, __webpack_exports__CohereTokenizer as CohereTokenizer, __webpack_exports__ConvBertForMaskedLM as ConvBertForMaskedLM, __webpack_exports__ConvBertForQuestionAnswering as ConvBertForQuestionAnswering, __webpack_exports__ConvBertForSequenceClassification as ConvBertForSequenceClassification, __webpack_exports__ConvBertForTokenClassification as ConvBertForTokenClassification, __webpack_exports__ConvBertModel as ConvBertModel, __webpack_exports__ConvBertPreTrainedModel as ConvBertPreTrainedModel, __webpack_exports__ConvBertTokenizer as ConvBertTokenizer, __webpack_exports__ConvNextFeatureExtractor as ConvNextFeatureExtractor, __webpack_exports__ConvNextForImageClassification as ConvNextForImageClassification, __webpack_exports__ConvNextImageProcessor as ConvNextImageProcessor, __webpack_exports__ConvNextModel as ConvNextModel, __webpack_exports__ConvNextPreTrainedModel as ConvNextPreTrainedModel, __webpack_exports__ConvNextV2ForImageClassification as ConvNextV2ForImageClassification, __webpack_exports__ConvNextV2Model as ConvNextV2Model, __webpack_exports__ConvNextV2PreTrainedModel as ConvNextV2PreTrainedModel, __webpack_exports__DPTFeatureExtractor as DPTFeatureExtractor, __webpack_exports__DPTForDepthEstimation as DPTForDepthEstimation, __webpack_exports__DPTImageProcessor as DPTImageProcessor, __webpack_exports__DPTModel as DPTModel, __webpack_exports__DPTPreTrainedModel as DPTPreTrainedModel, __webpack_exports__DebertaForMaskedLM as DebertaForMaskedLM, __webpack_exports__DebertaForQuestionAnswering as DebertaForQuestionAnswering, __webpack_exports__DebertaForSequenceClassification as DebertaForSequenceClassification, __webpack_exports__DebertaForTokenClassification as DebertaForTokenClassification, __webpack_exports__DebertaModel as DebertaModel, __webpack_exports__DebertaPreTrainedModel as DebertaPreTrainedModel, __webpack_exports__DebertaTokenizer as DebertaTokenizer, __webpack_exports__DebertaV2ForMaskedLM as DebertaV2ForMaskedLM, __webpack_exports__DebertaV2ForQuestionAnswering as DebertaV2ForQuestionAnswering, __webpack_exports__DebertaV2ForSequenceClassification as DebertaV2ForSequenceClassification, __webpack_exports__DebertaV2ForTokenClassification as DebertaV2ForTokenClassification, __webpack_exports__DebertaV2Model as DebertaV2Model, __webpack_exports__DebertaV2PreTrainedModel as DebertaV2PreTrainedModel, __webpack_exports__DebertaV2Tokenizer as DebertaV2Tokenizer, __webpack_exports__DecisionTransformerModel as DecisionTransformerModel, __webpack_exports__DecisionTransformerPreTrainedModel as DecisionTransformerPreTrainedModel, __webpack_exports__DeiTFeatureExtractor as DeiTFeatureExtractor, __webpack_exports__DeiTForImageClassification as DeiTForImageClassification, __webpack_exports__DeiTImageProcessor as DeiTImageProcessor, __webpack_exports__DeiTModel as DeiTModel, __webpack_exports__DeiTPreTrainedModel as DeiTPreTrainedModel, __webpack_exports__DepthAnythingForDepthEstimation as DepthAnythingForDepthEstimation, __webpack_exports__DepthAnythingPreTrainedModel as DepthAnythingPreTrainedModel, __webpack_exports__DepthEstimationPipeline as DepthEstimationPipeline, __webpack_exports__DepthProForDepthEstimation as DepthProForDepthEstimation, __webpack_exports__DepthProPreTrainedModel as DepthProPreTrainedModel, __webpack_exports__DetrFeatureExtractor as DetrFeatureExtractor, __webpack_exports__DetrForObjectDetection as DetrForObjectDetection, __webpack_exports__DetrForSegmentation as DetrForSegmentation, __webpack_exports__DetrImageProcessor as DetrImageProcessor, __webpack_exports__DetrModel as DetrModel, __webpack_exports__DetrObjectDetectionOutput as DetrObjectDetectionOutput, __webpack_exports__DetrPreTrainedModel as DetrPreTrainedModel, __webpack_exports__DetrSegmentationOutput as DetrSegmentationOutput, __webpack_exports__Dinov2ForImageClassification as Dinov2ForImageClassification, __webpack_exports__Dinov2Model as Dinov2Model, __webpack_exports__Dinov2PreTrainedModel as Dinov2PreTrainedModel, __webpack_exports__Dinov2WithRegistersForImageClassification as Dinov2WithRegistersForImageClassification, __webpack_exports__Dinov2WithRegistersModel as Dinov2WithRegistersModel, __webpack_exports__Dinov2WithRegistersPreTrainedModel as Dinov2WithRegistersPreTrainedModel, __webpack_exports__DistilBertForMaskedLM as DistilBertForMaskedLM, __webpack_exports__DistilBertForQuestionAnswering as DistilBertForQuestionAnswering, __webpack_exports__DistilBertForSequenceClassification as DistilBertForSequenceClassification, __webpack_exports__DistilBertForTokenClassification as DistilBertForTokenClassification, __webpack_exports__DistilBertModel as DistilBertModel, __webpack_exports__DistilBertPreTrainedModel as DistilBertPreTrainedModel, __webpack_exports__DistilBertTokenizer as DistilBertTokenizer, __webpack_exports__DocumentQuestionAnsweringPipeline as DocumentQuestionAnsweringPipeline, __webpack_exports__DonutFeatureExtractor as DonutFeatureExtractor, __webpack_exports__DonutImageProcessor as DonutImageProcessor, __webpack_exports__DonutSwinModel as DonutSwinModel, __webpack_exports__DonutSwinPreTrainedModel as DonutSwinPreTrainedModel, __webpack_exports__EfficientNetForImageClassification as EfficientNetForImageClassification, __webpack_exports__EfficientNetImageProcessor as EfficientNetImageProcessor, __webpack_exports__EfficientNetModel as EfficientNetModel, __webpack_exports__EfficientNetPreTrainedModel as EfficientNetPreTrainedModel, __webpack_exports__ElectraForMaskedLM as ElectraForMaskedLM, __webpack_exports__ElectraForQuestionAnswering as ElectraForQuestionAnswering, __webpack_exports__ElectraForSequenceClassification as ElectraForSequenceClassification, __webpack_exports__ElectraForTokenClassification as ElectraForTokenClassification, __webpack_exports__ElectraModel as ElectraModel, __webpack_exports__ElectraPreTrainedModel as ElectraPreTrainedModel, __webpack_exports__ElectraTokenizer as ElectraTokenizer, __webpack_exports__EosTokenCriteria as EosTokenCriteria, __webpack_exports__EsmForMaskedLM as EsmForMaskedLM, __webpack_exports__EsmForSequenceClassification as EsmForSequenceClassification, __webpack_exports__EsmForTokenClassification as EsmForTokenClassification, __webpack_exports__EsmModel as EsmModel, __webpack_exports__EsmPreTrainedModel as EsmPreTrainedModel, __webpack_exports__EsmTokenizer as EsmTokenizer, __webpack_exports__ExaoneForCausalLM as ExaoneForCausalLM, __webpack_exports__ExaoneModel as ExaoneModel, __webpack_exports__ExaonePreTrainedModel as ExaonePreTrainedModel, __webpack_exports__FFT as FFT, __webpack_exports__FalconForCausalLM as FalconForCausalLM, __webpack_exports__FalconModel as FalconModel, __webpack_exports__FalconPreTrainedModel as FalconPreTrainedModel, __webpack_exports__FalconTokenizer as FalconTokenizer, __webpack_exports__FastViTForImageClassification as FastViTForImageClassification, __webpack_exports__FastViTModel as FastViTModel, __webpack_exports__FastViTPreTrainedModel as FastViTPreTrainedModel, __webpack_exports__FeatureExtractionPipeline as FeatureExtractionPipeline, __webpack_exports__FeatureExtractor as FeatureExtractor, __webpack_exports__FillMaskPipeline as FillMaskPipeline, __webpack_exports__Florence2ForConditionalGeneration as Florence2ForConditionalGeneration, __webpack_exports__Florence2PreTrainedModel as Florence2PreTrainedModel, __webpack_exports__Florence2Processor as Florence2Processor, __webpack_exports__ForcedBOSTokenLogitsProcessor as ForcedBOSTokenLogitsProcessor, __webpack_exports__ForcedEOSTokenLogitsProcessor as ForcedEOSTokenLogitsProcessor, __webpack_exports__GLPNFeatureExtractor as GLPNFeatureExtractor, __webpack_exports__GLPNForDepthEstimation as GLPNForDepthEstimation, __webpack_exports__GLPNModel as GLPNModel, __webpack_exports__GLPNPreTrainedModel as GLPNPreTrainedModel, __webpack_exports__GPT2LMHeadModel as GPT2LMHeadModel, __webpack_exports__GPT2Model as GPT2Model, __webpack_exports__GPT2PreTrainedModel as GPT2PreTrainedModel, __webpack_exports__GPT2Tokenizer as GPT2Tokenizer, __webpack_exports__GPTBigCodeForCausalLM as GPTBigCodeForCausalLM, __webpack_exports__GPTBigCodeModel as GPTBigCodeModel, __webpack_exports__GPTBigCodePreTrainedModel as GPTBigCodePreTrainedModel, __webpack_exports__GPTJForCausalLM as GPTJForCausalLM, __webpack_exports__GPTJModel as GPTJModel, __webpack_exports__GPTJPreTrainedModel as GPTJPreTrainedModel, __webpack_exports__GPTNeoForCausalLM as GPTNeoForCausalLM, __webpack_exports__GPTNeoModel as GPTNeoModel, __webpack_exports__GPTNeoPreTrainedModel as GPTNeoPreTrainedModel, __webpack_exports__GPTNeoXForCausalLM as GPTNeoXForCausalLM, __webpack_exports__GPTNeoXModel as GPTNeoXModel, __webpack_exports__GPTNeoXPreTrainedModel as GPTNeoXPreTrainedModel, __webpack_exports__GPTNeoXTokenizer as GPTNeoXTokenizer, __webpack_exports__Gemma2ForCausalLM as Gemma2ForCausalLM, __webpack_exports__Gemma2Model as Gemma2Model, __webpack_exports__Gemma2PreTrainedModel as Gemma2PreTrainedModel, __webpack_exports__GemmaForCausalLM as GemmaForCausalLM, __webpack_exports__GemmaModel as GemmaModel, __webpack_exports__GemmaPreTrainedModel as GemmaPreTrainedModel, __webpack_exports__GemmaTokenizer as GemmaTokenizer, __webpack_exports__GraniteForCausalLM as GraniteForCausalLM, __webpack_exports__GraniteModel as GraniteModel, __webpack_exports__GranitePreTrainedModel as GranitePreTrainedModel, __webpack_exports__Grok1Tokenizer as Grok1Tokenizer, __webpack_exports__GroupViTModel as GroupViTModel, __webpack_exports__GroupViTPreTrainedModel as GroupViTPreTrainedModel, __webpack_exports__HerbertTokenizer as HerbertTokenizer, __webpack_exports__HieraForImageClassification as HieraForImageClassification, __webpack_exports__HieraModel as HieraModel, __webpack_exports__HieraPreTrainedModel as HieraPreTrainedModel, __webpack_exports__HubertForCTC as HubertForCTC, __webpack_exports__HubertForSequenceClassification as HubertForSequenceClassification, __webpack_exports__HubertModel as HubertModel, __webpack_exports__HubertPreTrainedModel as HubertPreTrainedModel, __webpack_exports__IJepaForImageClassification as IJepaForImageClassification, __webpack_exports__IJepaModel as IJepaModel, __webpack_exports__IJepaPreTrainedModel as IJepaPreTrainedModel, __webpack_exports__Idefics3ForConditionalGeneration as Idefics3ForConditionalGeneration, __webpack_exports__Idefics3ImageProcessor as Idefics3ImageProcessor, __webpack_exports__Idefics3PreTrainedModel as Idefics3PreTrainedModel, __webpack_exports__Idefics3Processor as Idefics3Processor, __webpack_exports__ImageClassificationPipeline as ImageClassificationPipeline, __webpack_exports__ImageFeatureExtractionPipeline as ImageFeatureExtractionPipeline, __webpack_exports__ImageFeatureExtractor as ImageFeatureExtractor, __webpack_exports__ImageMattingOutput as ImageMattingOutput, __webpack_exports__ImageProcessor as ImageProcessor, __webpack_exports__ImageSegmentationPipeline as ImageSegmentationPipeline, __webpack_exports__ImageToImagePipeline as ImageToImagePipeline, __webpack_exports__ImageToTextPipeline as ImageToTextPipeline, __webpack_exports__InterruptableStoppingCriteria as InterruptableStoppingCriteria, __webpack_exports__JAISLMHeadModel as JAISLMHeadModel, __webpack_exports__JAISModel as JAISModel, __webpack_exports__JAISPreTrainedModel as JAISPreTrainedModel, __webpack_exports__JinaCLIPImageProcessor as JinaCLIPImageProcessor, __webpack_exports__JinaCLIPModel as JinaCLIPModel, __webpack_exports__JinaCLIPPreTrainedModel as JinaCLIPPreTrainedModel, __webpack_exports__JinaCLIPProcessor as JinaCLIPProcessor, __webpack_exports__JinaCLIPTextModel as JinaCLIPTextModel, __webpack_exports__JinaCLIPVisionModel as JinaCLIPVisionModel, __webpack_exports__LlamaForCausalLM as LlamaForCausalLM, __webpack_exports__LlamaModel as LlamaModel, __webpack_exports__LlamaPreTrainedModel as LlamaPreTrainedModel, __webpack_exports__LlamaTokenizer as LlamaTokenizer, __webpack_exports__LlavaForConditionalGeneration as LlavaForConditionalGeneration, __webpack_exports__LlavaOnevisionForConditionalGeneration as LlavaOnevisionForConditionalGeneration, __webpack_exports__LlavaOnevisionImageProcessor as LlavaOnevisionImageProcessor, __webpack_exports__LlavaPreTrainedModel as LlavaPreTrainedModel, __webpack_exports__LogitsProcessor as LogitsProcessor, __webpack_exports__LogitsProcessorList as LogitsProcessorList, __webpack_exports__LogitsWarper as LogitsWarper, __webpack_exports__LongT5ForConditionalGeneration as LongT5ForConditionalGeneration, __webpack_exports__LongT5Model as LongT5Model, __webpack_exports__LongT5PreTrainedModel as LongT5PreTrainedModel, __webpack_exports__M2M100ForConditionalGeneration as M2M100ForConditionalGeneration, __webpack_exports__M2M100Model as M2M100Model, __webpack_exports__M2M100PreTrainedModel as M2M100PreTrainedModel, __webpack_exports__M2M100Tokenizer as M2M100Tokenizer, __webpack_exports__MBart50Tokenizer as MBart50Tokenizer, __webpack_exports__MBartForCausalLM as MBartForCausalLM, __webpack_exports__MBartForConditionalGeneration as MBartForConditionalGeneration, __webpack_exports__MBartForSequenceClassification as MBartForSequenceClassification, __webpack_exports__MBartModel as MBartModel, __webpack_exports__MBartPreTrainedModel as MBartPreTrainedModel, __webpack_exports__MBartTokenizer as MBartTokenizer, __webpack_exports__MPNetForMaskedLM as MPNetForMaskedLM, __webpack_exports__MPNetForQuestionAnswering as MPNetForQuestionAnswering, __webpack_exports__MPNetForSequenceClassification as MPNetForSequenceClassification, __webpack_exports__MPNetForTokenClassification as MPNetForTokenClassification, __webpack_exports__MPNetModel as MPNetModel, __webpack_exports__MPNetPreTrainedModel as MPNetPreTrainedModel, __webpack_exports__MPNetTokenizer as MPNetTokenizer, __webpack_exports__MT5ForConditionalGeneration as MT5ForConditionalGeneration, __webpack_exports__MT5Model as MT5Model, __webpack_exports__MT5PreTrainedModel as MT5PreTrainedModel, __webpack_exports__MarianMTModel as MarianMTModel, __webpack_exports__MarianModel as MarianModel, __webpack_exports__MarianPreTrainedModel as MarianPreTrainedModel, __webpack_exports__MarianTokenizer as MarianTokenizer, __webpack_exports__Mask2FormerImageProcessor as Mask2FormerImageProcessor, __webpack_exports__MaskFormerFeatureExtractor as MaskFormerFeatureExtractor, __webpack_exports__MaskFormerForInstanceSegmentation as MaskFormerForInstanceSegmentation, __webpack_exports__MaskFormerImageProcessor as MaskFormerImageProcessor, __webpack_exports__MaskFormerModel as MaskFormerModel, __webpack_exports__MaskFormerPreTrainedModel as MaskFormerPreTrainedModel, __webpack_exports__MaskedLMOutput as MaskedLMOutput, __webpack_exports__MaxLengthCriteria as MaxLengthCriteria, __webpack_exports__MgpstrForSceneTextRecognition as MgpstrForSceneTextRecognition, __webpack_exports__MgpstrModelOutput as MgpstrModelOutput, __webpack_exports__MgpstrPreTrainedModel as MgpstrPreTrainedModel, __webpack_exports__MgpstrProcessor as MgpstrProcessor, __webpack_exports__MgpstrTokenizer as MgpstrTokenizer, __webpack_exports__MinLengthLogitsProcessor as MinLengthLogitsProcessor, __webpack_exports__MinNewTokensLengthLogitsProcessor as MinNewTokensLengthLogitsProcessor, __webpack_exports__MistralForCausalLM as MistralForCausalLM, __webpack_exports__MistralModel as MistralModel, __webpack_exports__MistralPreTrainedModel as MistralPreTrainedModel, __webpack_exports__MobileBertForMaskedLM as MobileBertForMaskedLM, __webpack_exports__MobileBertForQuestionAnswering as MobileBertForQuestionAnswering, __webpack_exports__MobileBertForSequenceClassification as MobileBertForSequenceClassification, __webpack_exports__MobileBertModel as MobileBertModel, __webpack_exports__MobileBertPreTrainedModel as MobileBertPreTrainedModel, __webpack_exports__MobileBertTokenizer as MobileBertTokenizer, __webpack_exports__MobileLLMForCausalLM as MobileLLMForCausalLM, __webpack_exports__MobileLLMModel as MobileLLMModel, __webpack_exports__MobileLLMPreTrainedModel as MobileLLMPreTrainedModel, __webpack_exports__MobileNetV1FeatureExtractor as MobileNetV1FeatureExtractor, __webpack_exports__MobileNetV1ForImageClassification as MobileNetV1ForImageClassification, __webpack_exports__MobileNetV1ImageProcessor as MobileNetV1ImageProcessor, __webpack_exports__MobileNetV1Model as MobileNetV1Model, __webpack_exports__MobileNetV1PreTrainedModel as MobileNetV1PreTrainedModel, __webpack_exports__MobileNetV2FeatureExtractor as MobileNetV2FeatureExtractor, __webpack_exports__MobileNetV2ForImageClassification as MobileNetV2ForImageClassification, __webpack_exports__MobileNetV2ImageProcessor as MobileNetV2ImageProcessor, __webpack_exports__MobileNetV2Model as MobileNetV2Model, __webpack_exports__MobileNetV2PreTrainedModel as MobileNetV2PreTrainedModel, __webpack_exports__MobileNetV3FeatureExtractor as MobileNetV3FeatureExtractor, __webpack_exports__MobileNetV3ForImageClassification as MobileNetV3ForImageClassification, __webpack_exports__MobileNetV3ImageProcessor as MobileNetV3ImageProcessor, __webpack_exports__MobileNetV3Model as MobileNetV3Model, __webpack_exports__MobileNetV3PreTrainedModel as MobileNetV3PreTrainedModel, __webpack_exports__MobileNetV4FeatureExtractor as MobileNetV4FeatureExtractor, __webpack_exports__MobileNetV4ForImageClassification as MobileNetV4ForImageClassification, __webpack_exports__MobileNetV4ImageProcessor as MobileNetV4ImageProcessor, __webpack_exports__MobileNetV4Model as MobileNetV4Model, __webpack_exports__MobileNetV4PreTrainedModel as MobileNetV4PreTrainedModel, __webpack_exports__MobileViTFeatureExtractor as MobileViTFeatureExtractor, __webpack_exports__MobileViTForImageClassification as MobileViTForImageClassification, __webpack_exports__MobileViTImageProcessor as MobileViTImageProcessor, __webpack_exports__MobileViTModel as MobileViTModel, __webpack_exports__MobileViTPreTrainedModel as MobileViTPreTrainedModel, __webpack_exports__MobileViTV2ForImageClassification as MobileViTV2ForImageClassification, __webpack_exports__MobileViTV2Model as MobileViTV2Model, __webpack_exports__MobileViTV2PreTrainedModel as MobileViTV2PreTrainedModel, __webpack_exports__ModelOutput as ModelOutput, __webpack_exports__ModernBertForMaskedLM as ModernBertForMaskedLM, __webpack_exports__ModernBertForSequenceClassification as ModernBertForSequenceClassification, __webpack_exports__ModernBertForTokenClassification as ModernBertForTokenClassification, __webpack_exports__ModernBertModel as ModernBertModel, __webpack_exports__ModernBertPreTrainedModel as ModernBertPreTrainedModel, __webpack_exports__Moondream1ForConditionalGeneration as Moondream1ForConditionalGeneration, __webpack_exports__MoonshineFeatureExtractor as MoonshineFeatureExtractor, __webpack_exports__MoonshineForConditionalGeneration as MoonshineForConditionalGeneration, __webpack_exports__MoonshineModel as MoonshineModel, __webpack_exports__MoonshinePreTrainedModel as MoonshinePreTrainedModel, __webpack_exports__MoonshineProcessor as MoonshineProcessor, __webpack_exports__MptForCausalLM as MptForCausalLM, __webpack_exports__MptModel as MptModel, __webpack_exports__MptPreTrainedModel as MptPreTrainedModel, __webpack_exports__MultiModalityCausalLM as MultiModalityCausalLM, __webpack_exports__MultiModalityPreTrainedModel as MultiModalityPreTrainedModel, __webpack_exports__MusicgenForCausalLM as MusicgenForCausalLM, __webpack_exports__MusicgenForConditionalGeneration as MusicgenForConditionalGeneration, __webpack_exports__MusicgenModel as MusicgenModel, __webpack_exports__MusicgenPreTrainedModel as MusicgenPreTrainedModel, __webpack_exports__NllbTokenizer as NllbTokenizer, __webpack_exports__NoBadWordsLogitsProcessor as NoBadWordsLogitsProcessor, __webpack_exports__NoRepeatNGramLogitsProcessor as NoRepeatNGramLogitsProcessor, __webpack_exports__NomicBertModel as NomicBertModel, __webpack_exports__NomicBertPreTrainedModel as NomicBertPreTrainedModel, __webpack_exports__NougatImageProcessor as NougatImageProcessor, __webpack_exports__NougatTokenizer as NougatTokenizer, __webpack_exports__OPTForCausalLM as OPTForCausalLM, __webpack_exports__OPTModel as OPTModel, __webpack_exports__OPTPreTrainedModel as OPTPreTrainedModel, __webpack_exports__ObjectDetectionPipeline as ObjectDetectionPipeline, __webpack_exports__Olmo2ForCausalLM as Olmo2ForCausalLM, __webpack_exports__Olmo2Model as Olmo2Model, __webpack_exports__Olmo2PreTrainedModel as Olmo2PreTrainedModel, __webpack_exports__OlmoForCausalLM as OlmoForCausalLM, __webpack_exports__OlmoModel as OlmoModel, __webpack_exports__OlmoPreTrainedModel as OlmoPreTrainedModel, __webpack_exports__OpenELMForCausalLM as OpenELMForCausalLM, __webpack_exports__OpenELMModel as OpenELMModel, __webpack_exports__OpenELMPreTrainedModel as OpenELMPreTrainedModel, __webpack_exports__OwlViTFeatureExtractor as OwlViTFeatureExtractor, __webpack_exports__OwlViTForObjectDetection as OwlViTForObjectDetection, __webpack_exports__OwlViTImageProcessor as OwlViTImageProcessor, __webpack_exports__OwlViTModel as OwlViTModel, __webpack_exports__OwlViTPreTrainedModel as OwlViTPreTrainedModel, __webpack_exports__OwlViTProcessor as OwlViTProcessor, __webpack_exports__Owlv2ForObjectDetection as Owlv2ForObjectDetection, __webpack_exports__Owlv2ImageProcessor as Owlv2ImageProcessor, __webpack_exports__Owlv2Model as Owlv2Model, __webpack_exports__Owlv2PreTrainedModel as Owlv2PreTrainedModel, __webpack_exports__PaliGemmaForConditionalGeneration as PaliGemmaForConditionalGeneration, __webpack_exports__PaliGemmaPreTrainedModel as PaliGemmaPreTrainedModel, __webpack_exports__PaliGemmaProcessor as PaliGemmaProcessor, __webpack_exports__PatchTSMixerForPrediction as PatchTSMixerForPrediction, __webpack_exports__PatchTSMixerModel as PatchTSMixerModel, __webpack_exports__PatchTSMixerPreTrainedModel as PatchTSMixerPreTrainedModel, __webpack_exports__PatchTSTForPrediction as PatchTSTForPrediction, __webpack_exports__PatchTSTModel as PatchTSTModel, __webpack_exports__PatchTSTPreTrainedModel as PatchTSTPreTrainedModel, __webpack_exports__Phi3ForCausalLM as Phi3ForCausalLM, __webpack_exports__Phi3Model as Phi3Model, __webpack_exports__Phi3PreTrainedModel as Phi3PreTrainedModel, __webpack_exports__Phi3VForCausalLM as Phi3VForCausalLM, __webpack_exports__Phi3VImageProcessor as Phi3VImageProcessor, __webpack_exports__Phi3VPreTrainedModel as Phi3VPreTrainedModel, __webpack_exports__Phi3VProcessor as Phi3VProcessor, __webpack_exports__PhiForCausalLM as PhiForCausalLM, __webpack_exports__PhiModel as PhiModel, __webpack_exports__PhiPreTrainedModel as PhiPreTrainedModel, __webpack_exports__Pipeline as Pipeline, __webpack_exports__PreTrainedModel as PreTrainedModel, __webpack_exports__PreTrainedTokenizer as PreTrainedTokenizer, __webpack_exports__PretrainedConfig as PretrainedConfig, __webpack_exports__PretrainedMixin as PretrainedMixin, __webpack_exports__Processor as Processor, __webpack_exports__PvtForImageClassification as PvtForImageClassification, __webpack_exports__PvtImageProcessor as PvtImageProcessor, __webpack_exports__PvtModel as PvtModel, __webpack_exports__PvtPreTrainedModel as PvtPreTrainedModel, __webpack_exports__PyAnnoteFeatureExtractor as PyAnnoteFeatureExtractor, __webpack_exports__PyAnnoteForAudioFrameClassification as PyAnnoteForAudioFrameClassification, __webpack_exports__PyAnnoteModel as PyAnnoteModel, __webpack_exports__PyAnnotePreTrainedModel as PyAnnotePreTrainedModel, __webpack_exports__PyAnnoteProcessor as PyAnnoteProcessor, __webpack_exports__QuestionAnsweringModelOutput as QuestionAnsweringModelOutput, __webpack_exports__QuestionAnsweringPipeline as QuestionAnsweringPipeline, __webpack_exports__Qwen2ForCausalLM as Qwen2ForCausalLM, __webpack_exports__Qwen2Model as Qwen2Model, __webpack_exports__Qwen2PreTrainedModel as Qwen2PreTrainedModel, __webpack_exports__Qwen2Tokenizer as Qwen2Tokenizer, __webpack_exports__Qwen2VLForConditionalGeneration as Qwen2VLForConditionalGeneration, __webpack_exports__Qwen2VLImageProcessor as Qwen2VLImageProcessor, __webpack_exports__Qwen2VLPreTrainedModel as Qwen2VLPreTrainedModel, __webpack_exports__Qwen2VLProcessor as Qwen2VLProcessor, __webpack_exports__RTDetrForObjectDetection as RTDetrForObjectDetection, __webpack_exports__RTDetrImageProcessor as RTDetrImageProcessor, __webpack_exports__RTDetrModel as RTDetrModel, __webpack_exports__RTDetrObjectDetectionOutput as RTDetrObjectDetectionOutput, __webpack_exports__RTDetrPreTrainedModel as RTDetrPreTrainedModel, __webpack_exports__RawImage as RawImage, __webpack_exports__RepetitionPenaltyLogitsProcessor as RepetitionPenaltyLogitsProcessor, __webpack_exports__ResNetForImageClassification as ResNetForImageClassification, __webpack_exports__ResNetModel as ResNetModel, __webpack_exports__ResNetPreTrainedModel as ResNetPreTrainedModel, __webpack_exports__RoFormerForMaskedLM as RoFormerForMaskedLM, __webpack_exports__RoFormerForQuestionAnswering as RoFormerForQuestionAnswering, __webpack_exports__RoFormerForSequenceClassification as RoFormerForSequenceClassification, __webpack_exports__RoFormerForTokenClassification as RoFormerForTokenClassification, __webpack_exports__RoFormerModel as RoFormerModel, __webpack_exports__RoFormerPreTrainedModel as RoFormerPreTrainedModel, __webpack_exports__RoFormerTokenizer as RoFormerTokenizer, __webpack_exports__RobertaForMaskedLM as RobertaForMaskedLM, __webpack_exports__RobertaForQuestionAnswering as RobertaForQuestionAnswering, __webpack_exports__RobertaForSequenceClassification as RobertaForSequenceClassification, __webpack_exports__RobertaForTokenClassification as RobertaForTokenClassification, __webpack_exports__RobertaModel as RobertaModel, __webpack_exports__RobertaPreTrainedModel as RobertaPreTrainedModel, __webpack_exports__RobertaTokenizer as RobertaTokenizer, __webpack_exports__SamImageProcessor as SamImageProcessor, __webpack_exports__SamImageSegmentationOutput as SamImageSegmentationOutput, __webpack_exports__SamModel as SamModel, __webpack_exports__SamPreTrainedModel as SamPreTrainedModel, __webpack_exports__SamProcessor as SamProcessor, __webpack_exports__SapiensForDepthEstimation as SapiensForDepthEstimation, __webpack_exports__SapiensForNormalEstimation as SapiensForNormalEstimation, __webpack_exports__SapiensForSemanticSegmentation as SapiensForSemanticSegmentation, __webpack_exports__SapiensPreTrainedModel as SapiensPreTrainedModel, __webpack_exports__SeamlessM4TFeatureExtractor as SeamlessM4TFeatureExtractor, __webpack_exports__SegformerFeatureExtractor as SegformerFeatureExtractor, __webpack_exports__SegformerForImageClassification as SegformerForImageClassification, __webpack_exports__SegformerForSemanticSegmentation as SegformerForSemanticSegmentation, __webpack_exports__SegformerImageProcessor as SegformerImageProcessor, __webpack_exports__SegformerModel as SegformerModel, __webpack_exports__SegformerPreTrainedModel as SegformerPreTrainedModel, __webpack_exports__Seq2SeqLMOutput as Seq2SeqLMOutput, __webpack_exports__SequenceClassifierOutput as SequenceClassifierOutput, __webpack_exports__SiglipImageProcessor as SiglipImageProcessor, __webpack_exports__SiglipModel as SiglipModel, __webpack_exports__SiglipPreTrainedModel as SiglipPreTrainedModel, __webpack_exports__SiglipTextModel as SiglipTextModel, __webpack_exports__SiglipTokenizer as SiglipTokenizer, __webpack_exports__SiglipVisionModel as SiglipVisionModel, __webpack_exports__SpeechT5FeatureExtractor as SpeechT5FeatureExtractor, __webpack_exports__SpeechT5ForSpeechToText as SpeechT5ForSpeechToText, __webpack_exports__SpeechT5ForTextToSpeech as SpeechT5ForTextToSpeech, __webpack_exports__SpeechT5HifiGan as SpeechT5HifiGan, __webpack_exports__SpeechT5Model as SpeechT5Model, __webpack_exports__SpeechT5PreTrainedModel as SpeechT5PreTrainedModel, __webpack_exports__SpeechT5Processor as SpeechT5Processor, __webpack_exports__SpeechT5Tokenizer as SpeechT5Tokenizer, __webpack_exports__SqueezeBertForMaskedLM as SqueezeBertForMaskedLM, __webpack_exports__SqueezeBertForQuestionAnswering as SqueezeBertForQuestionAnswering, __webpack_exports__SqueezeBertForSequenceClassification as SqueezeBertForSequenceClassification, __webpack_exports__SqueezeBertModel as SqueezeBertModel, __webpack_exports__SqueezeBertPreTrainedModel as SqueezeBertPreTrainedModel, __webpack_exports__SqueezeBertTokenizer as SqueezeBertTokenizer, __webpack_exports__StableLmForCausalLM as StableLmForCausalLM, __webpack_exports__StableLmModel as StableLmModel, __webpack_exports__StableLmPreTrainedModel as StableLmPreTrainedModel, __webpack_exports__Starcoder2ForCausalLM as Starcoder2ForCausalLM, __webpack_exports__Starcoder2Model as Starcoder2Model, __webpack_exports__Starcoder2PreTrainedModel as Starcoder2PreTrainedModel, __webpack_exports__StoppingCriteria as StoppingCriteria, __webpack_exports__StoppingCriteriaList as StoppingCriteriaList, __webpack_exports__SummarizationPipeline as SummarizationPipeline, __webpack_exports__SuppressTokensAtBeginLogitsProcessor as SuppressTokensAtBeginLogitsProcessor, __webpack_exports__Swin2SRForImageSuperResolution as Swin2SRForImageSuperResolution, __webpack_exports__Swin2SRImageProcessor as Swin2SRImageProcessor, __webpack_exports__Swin2SRModel as Swin2SRModel, __webpack_exports__Swin2SRPreTrainedModel as Swin2SRPreTrainedModel, __webpack_exports__SwinForImageClassification as SwinForImageClassification, __webpack_exports__SwinModel as SwinModel, __webpack_exports__SwinPreTrainedModel as SwinPreTrainedModel, __webpack_exports__T5ForConditionalGeneration as T5ForConditionalGeneration, __webpack_exports__T5Model as T5Model, __webpack_exports__T5PreTrainedModel as T5PreTrainedModel, __webpack_exports__T5Tokenizer as T5Tokenizer, __webpack_exports__TableTransformerForObjectDetection as TableTransformerForObjectDetection, __webpack_exports__TableTransformerModel as TableTransformerModel, __webpack_exports__TableTransformerObjectDetectionOutput as TableTransformerObjectDetectionOutput, __webpack_exports__TableTransformerPreTrainedModel as TableTransformerPreTrainedModel, __webpack_exports__TemperatureLogitsWarper as TemperatureLogitsWarper, __webpack_exports__Tensor as Tensor, __webpack_exports__Text2TextGenerationPipeline as Text2TextGenerationPipeline, __webpack_exports__TextClassificationPipeline as TextClassificationPipeline, __webpack_exports__TextGenerationPipeline as TextGenerationPipeline, __webpack_exports__TextStreamer as TextStreamer, __webpack_exports__TextToAudioPipeline as TextToAudioPipeline, __webpack_exports__TokenClassificationPipeline as TokenClassificationPipeline, __webpack_exports__TokenClassifierOutput as TokenClassifierOutput, __webpack_exports__TokenizerModel as TokenizerModel, __webpack_exports__TopKLogitsWarper as TopKLogitsWarper, __webpack_exports__TopPLogitsWarper as TopPLogitsWarper, __webpack_exports__TrOCRForCausalLM as TrOCRForCausalLM, __webpack_exports__TrOCRPreTrainedModel as TrOCRPreTrainedModel, __webpack_exports__TranslationPipeline as TranslationPipeline, __webpack_exports__UniSpeechForCTC as UniSpeechForCTC, __webpack_exports__UniSpeechForSequenceClassification as UniSpeechForSequenceClassification, __webpack_exports__UniSpeechModel as UniSpeechModel, __webpack_exports__UniSpeechPreTrainedModel as UniSpeechPreTrainedModel, __webpack_exports__UniSpeechSatForAudioFrameClassification as UniSpeechSatForAudioFrameClassification, __webpack_exports__UniSpeechSatForCTC as UniSpeechSatForCTC, __webpack_exports__UniSpeechSatForSequenceClassification as UniSpeechSatForSequenceClassification, __webpack_exports__UniSpeechSatModel as UniSpeechSatModel, __webpack_exports__UniSpeechSatPreTrainedModel as UniSpeechSatPreTrainedModel, __webpack_exports__VLChatProcessor as VLChatProcessor, __webpack_exports__VLMImageProcessor as VLMImageProcessor, __webpack_exports__ViTFeatureExtractor as ViTFeatureExtractor, __webpack_exports__ViTForImageClassification as ViTForImageClassification, __webpack_exports__ViTImageProcessor as ViTImageProcessor, __webpack_exports__ViTMAEModel as ViTMAEModel, __webpack_exports__ViTMAEPreTrainedModel as ViTMAEPreTrainedModel, __webpack_exports__ViTMSNForImageClassification as ViTMSNForImageClassification, __webpack_exports__ViTMSNModel as ViTMSNModel, __webpack_exports__ViTMSNPreTrainedModel as ViTMSNPreTrainedModel, __webpack_exports__ViTModel as ViTModel, __webpack_exports__ViTPreTrainedModel as ViTPreTrainedModel, __webpack_exports__VisionEncoderDecoderModel as VisionEncoderDecoderModel, __webpack_exports__VitMatteForImageMatting as VitMatteForImageMatting, __webpack_exports__VitMatteImageProcessor as VitMatteImageProcessor, __webpack_exports__VitMattePreTrainedModel as VitMattePreTrainedModel, __webpack_exports__VitPoseForPoseEstimation as VitPoseForPoseEstimation, __webpack_exports__VitPoseImageProcessor as VitPoseImageProcessor, __webpack_exports__VitPosePreTrainedModel as VitPosePreTrainedModel, __webpack_exports__VitsModel as VitsModel, __webpack_exports__VitsModelOutput as VitsModelOutput, __webpack_exports__VitsPreTrainedModel as VitsPreTrainedModel, __webpack_exports__VitsTokenizer as VitsTokenizer, __webpack_exports__Wav2Vec2BertForCTC as Wav2Vec2BertForCTC, __webpack_exports__Wav2Vec2BertForSequenceClassification as Wav2Vec2BertForSequenceClassification, __webpack_exports__Wav2Vec2BertModel as Wav2Vec2BertModel, __webpack_exports__Wav2Vec2BertPreTrainedModel as Wav2Vec2BertPreTrainedModel, __webpack_exports__Wav2Vec2CTCTokenizer as Wav2Vec2CTCTokenizer, __webpack_exports__Wav2Vec2FeatureExtractor as Wav2Vec2FeatureExtractor, __webpack_exports__Wav2Vec2ForAudioFrameClassification as Wav2Vec2ForAudioFrameClassification, __webpack_exports__Wav2Vec2ForCTC as Wav2Vec2ForCTC, __webpack_exports__Wav2Vec2ForSequenceClassification as Wav2Vec2ForSequenceClassification, __webpack_exports__Wav2Vec2Model as Wav2Vec2Model, __webpack_exports__Wav2Vec2PreTrainedModel as Wav2Vec2PreTrainedModel, __webpack_exports__Wav2Vec2ProcessorWithLM as Wav2Vec2ProcessorWithLM, __webpack_exports__WavLMForAudioFrameClassification as WavLMForAudioFrameClassification, __webpack_exports__WavLMForCTC as WavLMForCTC, __webpack_exports__WavLMForSequenceClassification as WavLMForSequenceClassification, __webpack_exports__WavLMForXVector as WavLMForXVector, __webpack_exports__WavLMModel as WavLMModel, __webpack_exports__WavLMPreTrainedModel as WavLMPreTrainedModel, __webpack_exports__WeSpeakerFeatureExtractor as WeSpeakerFeatureExtractor, __webpack_exports__WeSpeakerResNetModel as WeSpeakerResNetModel, __webpack_exports__WeSpeakerResNetPreTrainedModel as WeSpeakerResNetPreTrainedModel, __webpack_exports__WhisperFeatureExtractor as WhisperFeatureExtractor, __webpack_exports__WhisperForConditionalGeneration as WhisperForConditionalGeneration, __webpack_exports__WhisperModel as WhisperModel, __webpack_exports__WhisperPreTrainedModel as WhisperPreTrainedModel, __webpack_exports__WhisperProcessor as WhisperProcessor, __webpack_exports__WhisperTextStreamer as WhisperTextStreamer, __webpack_exports__WhisperTimeStampLogitsProcessor as WhisperTimeStampLogitsProcessor, __webpack_exports__WhisperTokenizer as WhisperTokenizer, __webpack_exports__XLMForQuestionAnswering as XLMForQuestionAnswering, __webpack_exports__XLMForSequenceClassification as XLMForSequenceClassification, __webpack_exports__XLMForTokenClassification as XLMForTokenClassification, __webpack_exports__XLMModel as XLMModel, __webpack_exports__XLMPreTrainedModel as XLMPreTrainedModel, __webpack_exports__XLMRobertaForMaskedLM as XLMRobertaForMaskedLM, __webpack_exports__XLMRobertaForQuestionAnswering as XLMRobertaForQuestionAnswering, __webpack_exports__XLMRobertaForSequenceClassification as XLMRobertaForSequenceClassification, __webpack_exports__XLMRobertaForTokenClassification as XLMRobertaForTokenClassification, __webpack_exports__XLMRobertaModel as XLMRobertaModel, __webpack_exports__XLMRobertaPreTrainedModel as XLMRobertaPreTrainedModel, __webpack_exports__XLMRobertaTokenizer as XLMRobertaTokenizer, __webpack_exports__XLMTokenizer as XLMTokenizer, __webpack_exports__XLMWithLMHeadModel as XLMWithLMHeadModel, __webpack_exports__XVectorOutput as XVectorOutput, __webpack_exports__YolosFeatureExtractor as YolosFeatureExtractor, __webpack_exports__YolosForObjectDetection as YolosForObjectDetection, __webpack_exports__YolosImageProcessor as YolosImageProcessor, __webpack_exports__YolosModel as YolosModel, __webpack_exports__YolosObjectDetectionOutput as YolosObjectDetectionOutput, __webpack_exports__YolosPreTrainedModel as YolosPreTrainedModel, __webpack_exports__ZeroShotAudioClassificationPipeline as ZeroShotAudioClassificationPipeline, __webpack_exports__ZeroShotClassificationPipeline as ZeroShotClassificationPipeline, __webpack_exports__ZeroShotImageClassificationPipeline as ZeroShotImageClassificationPipeline, __webpack_exports__ZeroShotObjectDetectionPipeline as ZeroShotObjectDetectionPipeline, __webpack_exports__bankers_round as bankers_round, __webpack_exports__cat as cat, __webpack_exports__cos_sim as cos_sim, __webpack_exports__dot as dot, __webpack_exports__dynamic_time_warping as dynamic_time_warping, __webpack_exports__env as env, __webpack_exports__full as full, __webpack_exports__full_like as full_like, __webpack_exports__getKeyValueShapes as getKeyValueShapes, __webpack_exports__hamming as hamming, __webpack_exports__hanning as hanning, __webpack_exports__interpolate as interpolate, __webpack_exports__interpolate_4d as interpolate_4d, __webpack_exports__interpolate_data as interpolate_data, __webpack_exports__is_chinese_char as is_chinese_char, __webpack_exports__layer_norm as layer_norm, __webpack_exports__load_image as load_image, __webpack_exports__log_softmax as log_softmax, __webpack_exports__magnitude as magnitude, __webpack_exports__matmul as matmul, __webpack_exports__max as max, __webpack_exports__mean as mean, __webpack_exports__mean_pooling as mean_pooling, __webpack_exports__medianFilter as medianFilter, __webpack_exports__mel_filter_bank as mel_filter_bank, __webpack_exports__min as min, __webpack_exports__ones as ones, __webpack_exports__ones_like as ones_like, __webpack_exports__permute as permute, __webpack_exports__permute_data as permute_data, __webpack_exports__pipeline as pipeline, __webpack_exports__quantize_embeddings as quantize_embeddings, __webpack_exports__rand as rand, __webpack_exports__read_audio as read_audio, __webpack_exports__rfft as rfft, __webpack_exports__round as round, __webpack_exports__slice as slice, __webpack_exports__softmax as softmax, __webpack_exports__spectrogram as spectrogram, __webpack_exports__stack as stack, __webpack_exports__std_mean as std_mean, __webpack_exports__topk as topk, __webpack_exports__window_function as window_function, __webpack_exports__zeros as zeros, __webpack_exports__zeros_like as zeros_like };
|
|
35983
36125
|
|
|
35984
36126
|
//# sourceMappingURL=transformers.mjs.map
|