@huggingface/transformers 3.2.2 → 3.2.4
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +3 -2
- package/dist/transformers.cjs +252 -113
- package/dist/transformers.cjs.map +1 -1
- package/dist/transformers.js +256 -114
- package/dist/transformers.js.map +1 -1
- package/dist/transformers.min.cjs +1 -1
- package/dist/transformers.min.cjs.map +1 -1
- package/dist/transformers.min.js +1 -1
- package/dist/transformers.min.js.map +1 -1
- package/dist/transformers.min.mjs +1 -1
- package/dist/transformers.min.mjs.map +1 -1
- package/dist/transformers.mjs +256 -114
- package/dist/transformers.mjs.map +1 -1
- package/package.json +2 -2
- package/src/base/feature_extraction_utils.js +9 -9
- package/src/base/image_processors_utils.js +11 -0
- package/src/base/processing_utils.js +13 -3
- package/src/configs.js +5 -0
- package/src/env.js +1 -1
- package/src/models/auto/feature_extraction_auto.js +0 -16
- package/src/models/auto/processing_auto.js +0 -16
- package/src/models/convnext/image_processing_convnext.js +1 -0
- package/src/models/efficientnet/image_processing_efficientnet.js +1 -0
- package/src/models/florence2/processing_florence2.js +3 -0
- package/src/models/idefics3/image_processing_idefics3.js +2 -0
- package/src/models/janus/image_processing_janus.js +1 -0
- package/src/models/mgp_str/processing_mgp_str.js +2 -0
- package/src/models/paligemma/processing_paligemma.js +1 -0
- package/src/models/phi3_v/processing_phi3_v.js +1 -1
- package/src/models/pyannote/feature_extraction_pyannote.js +1 -0
- package/src/models/qwen2_vl/processing_qwen2_vl.js +1 -0
- package/src/models/seamless_m4t/feature_extraction_seamless_m4t.js +2 -2
- package/src/models/whisper/feature_extraction_whisper.js +1 -1
- package/src/models.js +93 -36
- package/src/ops/registry.js +10 -0
- package/src/pipelines.js +34 -7
- package/src/tokenizers.js +4 -7
- package/src/utils/dtypes.js +2 -0
- package/src/utils/hub.js +1 -1
- package/src/utils/maths.js +8 -6
- package/src/utils/tensor.js +42 -10
- package/types/base/feature_extraction_utils.d.ts +7 -7
- package/types/base/image_processors_utils.d.ts.map +1 -1
- package/types/base/processing_utils.d.ts +17 -19
- package/types/base/processing_utils.d.ts.map +1 -1
- package/types/configs.d.ts.map +1 -1
- package/types/generation/parameters.d.ts +1 -1
- package/types/models/auto/feature_extraction_auto.d.ts.map +1 -1
- package/types/models/auto/image_processing_auto.d.ts.map +1 -1
- package/types/models/auto/processing_auto.d.ts.map +1 -1
- package/types/models/convnext/image_processing_convnext.d.ts.map +1 -1
- package/types/models/efficientnet/image_processing_efficientnet.d.ts.map +1 -1
- package/types/models/florence2/processing_florence2.d.ts.map +1 -1
- package/types/models/idefics3/image_processing_idefics3.d.ts.map +1 -1
- package/types/models/janus/image_processing_janus.d.ts.map +1 -1
- package/types/models/mgp_str/processing_mgp_str.d.ts.map +1 -1
- package/types/models/paligemma/processing_paligemma.d.ts.map +1 -1
- package/types/models/phi3_v/processing_phi3_v.d.ts +6 -2
- package/types/models/phi3_v/processing_phi3_v.d.ts.map +1 -1
- package/types/models/pyannote/feature_extraction_pyannote.d.ts.map +1 -1
- package/types/models/qwen2_vl/processing_qwen2_vl.d.ts.map +1 -1
- package/types/models/sapiens/image_processing_sapiens.d.ts +10 -0
- package/types/models/sapiens/image_processing_sapiens.d.ts.map +1 -0
- package/types/models/whisper/generation_whisper.d.ts +1 -1
- package/types/models/whisper/generation_whisper.d.ts.map +1 -1
- package/types/models.d.ts +48 -17
- package/types/models.d.ts.map +1 -1
- package/types/ops/registry.d.ts +1 -0
- package/types/ops/registry.d.ts.map +1 -1
- package/types/pipelines.d.ts +2 -2
- package/types/pipelines.d.ts.map +1 -1
- package/types/tokenizers.d.ts.map +1 -1
- package/types/tsconfig.tsbuildinfo +1 -0
- package/types/utils/dtypes.d.ts.map +1 -1
- package/types/utils/hub.d.ts +1 -1
- package/types/utils/hub.d.ts.map +1 -1
- package/types/utils/image.d.ts +3 -2
- package/types/utils/image.d.ts.map +1 -1
- package/types/utils/maths.d.ts +8 -6
- package/types/utils/maths.d.ts.map +1 -1
- package/types/utils/tensor.d.ts +8 -4
- package/types/utils/tensor.d.ts.map +1 -1
package/dist/transformers.js
CHANGED
|
@@ -6927,23 +6927,23 @@ class FeatureExtractor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_1__.Ca
|
|
|
6927
6927
|
}
|
|
6928
6928
|
|
|
6929
6929
|
/**
|
|
6930
|
-
* Instantiate one of the
|
|
6930
|
+
* Instantiate one of the feature extractor classes of the library from a pretrained model.
|
|
6931
6931
|
*
|
|
6932
|
-
* The
|
|
6933
|
-
*
|
|
6932
|
+
* The feature extractor class to instantiate is selected based on the `feature_extractor_type` property of
|
|
6933
|
+
* the config object (either passed as an argument or loaded from `pretrained_model_name_or_path` if possible)
|
|
6934
6934
|
*
|
|
6935
6935
|
* @param {string} pretrained_model_name_or_path The name or path of the pretrained model. Can be either:
|
|
6936
|
-
* - A string, the *model id* of a pretrained
|
|
6936
|
+
* - A string, the *model id* of a pretrained feature_extractor hosted inside a model repo on huggingface.co.
|
|
6937
6937
|
* Valid model ids can be located at the root-level, like `bert-base-uncased`, or namespaced under a
|
|
6938
6938
|
* user or organization name, like `dbmdz/bert-base-german-cased`.
|
|
6939
|
-
* - A path to a *directory* containing
|
|
6940
|
-
* @param {import('../utils/hub.js').PretrainedOptions} options Additional options for loading the
|
|
6939
|
+
* - A path to a *directory* containing feature_extractor files, e.g., `./my_model_directory/`.
|
|
6940
|
+
* @param {import('../utils/hub.js').PretrainedOptions} options Additional options for loading the feature_extractor.
|
|
6941
6941
|
*
|
|
6942
|
-
* @returns {Promise<FeatureExtractor>} A new instance of the
|
|
6942
|
+
* @returns {Promise<FeatureExtractor>} A new instance of the Feature Extractor class.
|
|
6943
6943
|
*/
|
|
6944
6944
|
static async from_pretrained(pretrained_model_name_or_path, options) {
|
|
6945
|
-
const
|
|
6946
|
-
return new this(
|
|
6945
|
+
const config = await (0,_utils_hub_js__WEBPACK_IMPORTED_MODULE_2__.getModelJSON)(pretrained_model_name_or_path, _utils_constants_js__WEBPACK_IMPORTED_MODULE_0__.FEATURE_EXTRACTOR_NAME, true, options);
|
|
6946
|
+
return new this(config);
|
|
6947
6947
|
}
|
|
6948
6948
|
}
|
|
6949
6949
|
|
|
@@ -7593,14 +7593,20 @@ class ImageProcessor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_0__.Call
|
|
|
7593
7593
|
this.do_thumbnail = config.do_thumbnail;
|
|
7594
7594
|
this.size = config.size ?? config.image_size;
|
|
7595
7595
|
this.do_resize = config.do_resize ?? (this.size !== undefined);
|
|
7596
|
+
// @ts-expect-error TS2339
|
|
7596
7597
|
this.size_divisibility = config.size_divisibility ?? config.size_divisor;
|
|
7597
7598
|
|
|
7598
7599
|
this.do_center_crop = config.do_center_crop;
|
|
7600
|
+
// @ts-expect-error TS2339
|
|
7599
7601
|
this.crop_size = config.crop_size;
|
|
7602
|
+
// @ts-expect-error TS2339
|
|
7600
7603
|
this.do_convert_rgb = config.do_convert_rgb ?? true;
|
|
7604
|
+
// @ts-expect-error TS2339
|
|
7601
7605
|
this.do_crop_margin = config.do_crop_margin;
|
|
7602
7606
|
|
|
7607
|
+
// @ts-expect-error TS2339
|
|
7603
7608
|
this.pad_size = config.pad_size;
|
|
7609
|
+
// @ts-expect-error TS2339
|
|
7604
7610
|
this.do_pad = config.do_pad;
|
|
7605
7611
|
|
|
7606
7612
|
if (this.do_pad && !this.pad_size && this.size && this.size.width !== undefined && this.size.height !== undefined) {
|
|
@@ -7809,6 +7815,7 @@ class ImageProcessor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_0__.Call
|
|
|
7809
7815
|
// Support both formats for backwards compatibility
|
|
7810
7816
|
else if (Number.isInteger(size)) {
|
|
7811
7817
|
shortest_edge = size;
|
|
7818
|
+
// @ts-expect-error TS2339
|
|
7812
7819
|
longest_edge = this.config.max_size ?? shortest_edge;
|
|
7813
7820
|
|
|
7814
7821
|
} else if (size !== undefined) {
|
|
@@ -7877,6 +7884,7 @@ class ImageProcessor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_0__.Call
|
|
|
7877
7884
|
} else if (size.min_pixels !== undefined && size.max_pixels !== undefined) {
|
|
7878
7885
|
// Custom resize logic for Qwen2-VL models
|
|
7879
7886
|
const { min_pixels, max_pixels } = size;
|
|
7887
|
+
// @ts-expect-error TS2339
|
|
7880
7888
|
const factor = this.config.patch_size * this.config.merge_size;
|
|
7881
7889
|
return smart_resize(srcHeight, srcWidth, factor, min_pixels, max_pixels);
|
|
7882
7890
|
} else {
|
|
@@ -7892,6 +7900,7 @@ class ImageProcessor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_0__.Call
|
|
|
7892
7900
|
async resize(image) {
|
|
7893
7901
|
const [newWidth, newHeight] = this.get_resize_output_image_size(image, this.size);
|
|
7894
7902
|
return await image.resize(newWidth, newHeight, {
|
|
7903
|
+
// @ts-expect-error TS2322
|
|
7895
7904
|
resample: this.resample,
|
|
7896
7905
|
});
|
|
7897
7906
|
}
|
|
@@ -7942,6 +7951,7 @@ class ImageProcessor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_0__.Call
|
|
|
7942
7951
|
|
|
7943
7952
|
// Resize the image using thumbnail method.
|
|
7944
7953
|
if (this.do_thumbnail) {
|
|
7954
|
+
// @ts-expect-error TS2345
|
|
7945
7955
|
image = await this.thumbnail(image, this.size, this.resample);
|
|
7946
7956
|
}
|
|
7947
7957
|
|
|
@@ -7966,6 +7976,7 @@ class ImageProcessor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_0__.Call
|
|
|
7966
7976
|
// NOTE: All pixel-level manipulation (i.e., modifying `pixelData`)
|
|
7967
7977
|
// occurs with data in the hwc format (height, width, channels),
|
|
7968
7978
|
// to emulate the behavior of the original Python code (w/ numpy).
|
|
7979
|
+
/** @type {Float32Array} */
|
|
7969
7980
|
let pixelData = Float32Array.from(image.data);
|
|
7970
7981
|
let imgDims = [image.height, image.width, image.channels];
|
|
7971
7982
|
|
|
@@ -8123,6 +8134,7 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
8123
8134
|
/**
|
|
8124
8135
|
* @typedef {Object} ProcessorProperties Additional processor-specific properties.
|
|
8125
8136
|
* @typedef {import('../utils/hub.js').PretrainedOptions & ProcessorProperties} PretrainedProcessorOptions
|
|
8137
|
+
* @typedef {import('../tokenizers.js').PreTrainedTokenizer} PreTrainedTokenizer
|
|
8126
8138
|
*/
|
|
8127
8139
|
|
|
8128
8140
|
|
|
@@ -8156,7 +8168,7 @@ class Processor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_1__.Callable
|
|
|
8156
8168
|
}
|
|
8157
8169
|
|
|
8158
8170
|
/**
|
|
8159
|
-
* @returns {
|
|
8171
|
+
* @returns {PreTrainedTokenizer|undefined} The tokenizer of the processor, if it exists.
|
|
8160
8172
|
*/
|
|
8161
8173
|
get tokenizer() {
|
|
8162
8174
|
return this.components.tokenizer;
|
|
@@ -8169,6 +8181,11 @@ class Processor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_1__.Callable
|
|
|
8169
8181
|
return this.components.feature_extractor;
|
|
8170
8182
|
}
|
|
8171
8183
|
|
|
8184
|
+
/**
|
|
8185
|
+
* @param {Parameters<PreTrainedTokenizer['apply_chat_template']>[0]} messages
|
|
8186
|
+
* @param {Parameters<PreTrainedTokenizer['apply_chat_template']>[1]} options
|
|
8187
|
+
* @returns {ReturnType<PreTrainedTokenizer['apply_chat_template']>}
|
|
8188
|
+
*/
|
|
8172
8189
|
apply_chat_template(messages, options = {}) {
|
|
8173
8190
|
if (!this.tokenizer) {
|
|
8174
8191
|
throw new Error('Unable to apply chat template without a tokenizer.');
|
|
@@ -8179,6 +8196,10 @@ class Processor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_1__.Callable
|
|
|
8179
8196
|
});
|
|
8180
8197
|
}
|
|
8181
8198
|
|
|
8199
|
+
/**
|
|
8200
|
+
* @param {Parameters<PreTrainedTokenizer['batch_decode']>} args
|
|
8201
|
+
* @returns {ReturnType<PreTrainedTokenizer['batch_decode']>}
|
|
8202
|
+
*/
|
|
8182
8203
|
batch_decode(...args) {
|
|
8183
8204
|
if (!this.tokenizer) {
|
|
8184
8205
|
throw new Error('Unable to decode without a tokenizer.');
|
|
@@ -8206,8 +8227,8 @@ class Processor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_1__.Callable
|
|
|
8206
8227
|
/**
|
|
8207
8228
|
* Instantiate one of the processor classes of the library from a pretrained model.
|
|
8208
8229
|
*
|
|
8209
|
-
* The processor class to instantiate is selected based on the `
|
|
8210
|
-
* (either passed as an argument or loaded from `pretrained_model_name_or_path` if possible)
|
|
8230
|
+
* The processor class to instantiate is selected based on the `image_processor_type` (or `feature_extractor_type`; legacy)
|
|
8231
|
+
* property of the config object (either passed as an argument or loaded from `pretrained_model_name_or_path` if possible)
|
|
8211
8232
|
*
|
|
8212
8233
|
* @param {string} pretrained_model_name_or_path The name or path of the pretrained model. Can be either:
|
|
8213
8234
|
* - A string, the *model id* of a pretrained processor hosted inside a model repo on huggingface.co.
|
|
@@ -8326,15 +8347,19 @@ function getNormalizedConfig(config) {
|
|
|
8326
8347
|
case 'florence2':
|
|
8327
8348
|
case 'llava_onevision':
|
|
8328
8349
|
case 'idefics3':
|
|
8350
|
+
// @ts-expect-error TS2339
|
|
8329
8351
|
init_normalized_config = getNormalizedConfig(config.text_config);
|
|
8330
8352
|
break;
|
|
8331
8353
|
case 'moondream1':
|
|
8354
|
+
// @ts-expect-error TS2339
|
|
8332
8355
|
init_normalized_config = getNormalizedConfig(config.phi_config);
|
|
8333
8356
|
break;
|
|
8334
8357
|
case 'musicgen':
|
|
8358
|
+
// @ts-expect-error TS2339
|
|
8335
8359
|
init_normalized_config = getNormalizedConfig(config.decoder);
|
|
8336
8360
|
break;
|
|
8337
8361
|
case 'multi_modality':
|
|
8362
|
+
// @ts-expect-error TS2339
|
|
8338
8363
|
init_normalized_config = getNormalizedConfig(config.language_config);
|
|
8339
8364
|
break;
|
|
8340
8365
|
|
|
@@ -8455,6 +8480,7 @@ function getNormalizedConfig(config) {
|
|
|
8455
8480
|
break;
|
|
8456
8481
|
|
|
8457
8482
|
case 'vision-encoder-decoder':
|
|
8483
|
+
// @ts-expect-error TS2339
|
|
8458
8484
|
const decoderConfig = getNormalizedConfig(config.decoder);
|
|
8459
8485
|
|
|
8460
8486
|
const add_encoder_pkv = 'num_decoder_layers' in decoderConfig;
|
|
@@ -8697,7 +8723,7 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
8697
8723
|
|
|
8698
8724
|
|
|
8699
8725
|
|
|
8700
|
-
const VERSION = '3.2.
|
|
8726
|
+
const VERSION = '3.2.4';
|
|
8701
8727
|
|
|
8702
8728
|
// Check if various APIs are available (depends on environment)
|
|
8703
8729
|
const IS_BROWSER_ENV = typeof window !== "undefined" && typeof window.document !== "undefined";
|
|
@@ -10767,6 +10793,9 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
10767
10793
|
/* harmony export */ Dinov2ForImageClassification: () => (/* binding */ Dinov2ForImageClassification),
|
|
10768
10794
|
/* harmony export */ Dinov2Model: () => (/* binding */ Dinov2Model),
|
|
10769
10795
|
/* harmony export */ Dinov2PreTrainedModel: () => (/* binding */ Dinov2PreTrainedModel),
|
|
10796
|
+
/* harmony export */ Dinov2WithRegistersForImageClassification: () => (/* binding */ Dinov2WithRegistersForImageClassification),
|
|
10797
|
+
/* harmony export */ Dinov2WithRegistersModel: () => (/* binding */ Dinov2WithRegistersModel),
|
|
10798
|
+
/* harmony export */ Dinov2WithRegistersPreTrainedModel: () => (/* binding */ Dinov2WithRegistersPreTrainedModel),
|
|
10770
10799
|
/* harmony export */ DistilBertForMaskedLM: () => (/* binding */ DistilBertForMaskedLM),
|
|
10771
10800
|
/* harmony export */ DistilBertForQuestionAnswering: () => (/* binding */ DistilBertForQuestionAnswering),
|
|
10772
10801
|
/* harmony export */ DistilBertForSequenceClassification: () => (/* binding */ DistilBertForSequenceClassification),
|
|
@@ -11350,8 +11379,11 @@ async function getSession(pretrained_model_name_or_path, fileName, options) {
|
|
|
11350
11379
|
} else if (session_options.externalData !== undefined) {
|
|
11351
11380
|
externalDataPromises = session_options.externalData.map(async (ext) => {
|
|
11352
11381
|
// if the external data is a string, fetch the file and replace the string with its content
|
|
11382
|
+
// @ts-expect-error TS2339
|
|
11353
11383
|
if (typeof ext.data === "string") {
|
|
11384
|
+
// @ts-expect-error TS2339
|
|
11354
11385
|
const ext_buffer = await (0,_utils_hub_js__WEBPACK_IMPORTED_MODULE_5__.getModelFile)(pretrained_model_name_or_path, ext.data, true, options);
|
|
11386
|
+
// @ts-expect-error TS2698
|
|
11355
11387
|
return { ...ext, data: ext_buffer };
|
|
11356
11388
|
}
|
|
11357
11389
|
return ext;
|
|
@@ -12599,6 +12631,7 @@ class PreTrainedModel extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_3__.Cal
|
|
|
12599
12631
|
if (this.config.model_type === 'musicgen') {
|
|
12600
12632
|
// Custom logic (TODO: move to Musicgen class)
|
|
12601
12633
|
decoder_input_ids = Array.from({
|
|
12634
|
+
// @ts-expect-error TS2339
|
|
12602
12635
|
length: batch_size * this.config.decoder.num_codebooks
|
|
12603
12636
|
}, () => [decoder_start_token_id]);
|
|
12604
12637
|
|
|
@@ -12928,11 +12961,13 @@ class PreTrainedModel extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_3__.Cal
|
|
|
12928
12961
|
async encode_image({ pixel_values }) {
|
|
12929
12962
|
// image_inputs === { pixel_values }
|
|
12930
12963
|
const features = (await sessionRun(this.sessions['vision_encoder'], { pixel_values })).image_features;
|
|
12964
|
+
// @ts-expect-error TS2339
|
|
12931
12965
|
if (!this.config.num_image_tokens) {
|
|
12932
12966
|
console.warn(
|
|
12933
12967
|
'The number of image tokens was not set in the model configuration. ' +
|
|
12934
12968
|
`Setting it to the number of features detected by the vision encoder (${features.dims[1]}).`
|
|
12935
12969
|
)
|
|
12970
|
+
// @ts-expect-error TS2339
|
|
12936
12971
|
this.config.num_image_tokens = features.dims[1];
|
|
12937
12972
|
}
|
|
12938
12973
|
return features;
|
|
@@ -14360,6 +14395,7 @@ class WhisperForConditionalGeneration extends WhisperPreTrainedModel {
|
|
|
14360
14395
|
|
|
14361
14396
|
if (generation_config.return_token_timestamps) {
|
|
14362
14397
|
outputs["token_timestamps"] = this._extract_token_timestamps(
|
|
14398
|
+
// @ts-expect-error TS2345
|
|
14363
14399
|
outputs,
|
|
14364
14400
|
generation_config.alignment_heads,
|
|
14365
14401
|
generation_config.num_frames,
|
|
@@ -14395,6 +14431,7 @@ class WhisperForConditionalGeneration extends WhisperPreTrainedModel {
|
|
|
14395
14431
|
);
|
|
14396
14432
|
}
|
|
14397
14433
|
|
|
14434
|
+
// @ts-expect-error TS2339
|
|
14398
14435
|
let median_filter_width = this.config.median_filter_width;
|
|
14399
14436
|
if (median_filter_width === undefined) {
|
|
14400
14437
|
console.warn("Model config has no `median_filter_width`, using default value of 7.")
|
|
@@ -14405,6 +14442,7 @@ class WhisperForConditionalGeneration extends WhisperPreTrainedModel {
|
|
|
14405
14442
|
const batch = generate_outputs.cross_attentions;
|
|
14406
14443
|
// Create a list with `decoder_layers` elements, each a tensor of shape
|
|
14407
14444
|
// (batch size, attention_heads, output length, input length).
|
|
14445
|
+
// @ts-expect-error TS2339
|
|
14408
14446
|
const cross_attentions = Array.from({ length: this.config.decoder_layers },
|
|
14409
14447
|
// Concatenate the cross attentions for each layer across sequence length dimension.
|
|
14410
14448
|
(_, i) => (0,_utils_tensor_js__WEBPACK_IMPORTED_MODULE_9__.cat)(batch.map(x => x[i]), 2)
|
|
@@ -14501,7 +14539,7 @@ class MoonshinePreTrainedModel extends PreTrainedModel {
|
|
|
14501
14539
|
*/
|
|
14502
14540
|
class MoonshineModel extends MoonshinePreTrainedModel { }
|
|
14503
14541
|
|
|
14504
|
-
class MoonshineForConditionalGeneration extends MoonshinePreTrainedModel { }
|
|
14542
|
+
class MoonshineForConditionalGeneration extends MoonshinePreTrainedModel { }
|
|
14505
14543
|
//////////////////////////////////////////////////
|
|
14506
14544
|
|
|
14507
14545
|
|
|
@@ -14548,6 +14586,7 @@ class LlavaForConditionalGeneration extends LlavaPreTrainedModel {
|
|
|
14548
14586
|
attention_mask,
|
|
14549
14587
|
}) {
|
|
14550
14588
|
|
|
14589
|
+
// @ts-expect-error TS2339
|
|
14551
14590
|
const image_token_index = this.config.image_token_index;
|
|
14552
14591
|
|
|
14553
14592
|
const idsList = input_ids.tolist();
|
|
@@ -14901,9 +14940,9 @@ class CLIPTextModel extends CLIPPreTrainedModel {
|
|
|
14901
14940
|
/** @type {typeof PreTrainedModel.from_pretrained} */
|
|
14902
14941
|
static async from_pretrained(pretrained_model_name_or_path, options = {}) {
|
|
14903
14942
|
return super.from_pretrained(pretrained_model_name_or_path, {
|
|
14904
|
-
// Update default model file name if not provided
|
|
14905
|
-
model_file_name: 'text_model',
|
|
14906
14943
|
...options,
|
|
14944
|
+
// Update default model file name if not provided
|
|
14945
|
+
model_file_name: options.model_file_name ?? 'text_model',
|
|
14907
14946
|
});
|
|
14908
14947
|
}
|
|
14909
14948
|
}
|
|
@@ -14938,9 +14977,9 @@ class CLIPTextModelWithProjection extends CLIPPreTrainedModel {
|
|
|
14938
14977
|
/** @type {typeof PreTrainedModel.from_pretrained} */
|
|
14939
14978
|
static async from_pretrained(pretrained_model_name_or_path, options = {}) {
|
|
14940
14979
|
return super.from_pretrained(pretrained_model_name_or_path, {
|
|
14941
|
-
// Update default model file name if not provided
|
|
14942
|
-
model_file_name: 'text_model',
|
|
14943
14980
|
...options,
|
|
14981
|
+
// Update default model file name if not provided
|
|
14982
|
+
model_file_name: options.model_file_name ?? 'text_model',
|
|
14944
14983
|
});
|
|
14945
14984
|
}
|
|
14946
14985
|
}
|
|
@@ -14952,9 +14991,9 @@ class CLIPVisionModel extends CLIPPreTrainedModel {
|
|
|
14952
14991
|
/** @type {typeof PreTrainedModel.from_pretrained} */
|
|
14953
14992
|
static async from_pretrained(pretrained_model_name_or_path, options = {}) {
|
|
14954
14993
|
return super.from_pretrained(pretrained_model_name_or_path, {
|
|
14955
|
-
// Update default model file name if not provided
|
|
14956
|
-
model_file_name: 'vision_model',
|
|
14957
14994
|
...options,
|
|
14995
|
+
// Update default model file name if not provided
|
|
14996
|
+
model_file_name: options.model_file_name ?? 'vision_model',
|
|
14958
14997
|
});
|
|
14959
14998
|
}
|
|
14960
14999
|
}
|
|
@@ -14989,9 +15028,9 @@ class CLIPVisionModelWithProjection extends CLIPPreTrainedModel {
|
|
|
14989
15028
|
/** @type {typeof PreTrainedModel.from_pretrained} */
|
|
14990
15029
|
static async from_pretrained(pretrained_model_name_or_path, options = {}) {
|
|
14991
15030
|
return super.from_pretrained(pretrained_model_name_or_path, {
|
|
14992
|
-
// Update default model file name if not provided
|
|
14993
|
-
model_file_name: 'vision_model',
|
|
14994
15031
|
...options,
|
|
15032
|
+
// Update default model file name if not provided
|
|
15033
|
+
model_file_name: options.model_file_name ?? 'vision_model',
|
|
14995
15034
|
});
|
|
14996
15035
|
}
|
|
14997
15036
|
}
|
|
@@ -15077,9 +15116,9 @@ class SiglipTextModel extends SiglipPreTrainedModel {
|
|
|
15077
15116
|
/** @type {typeof PreTrainedModel.from_pretrained} */
|
|
15078
15117
|
static async from_pretrained(pretrained_model_name_or_path, options = {}) {
|
|
15079
15118
|
return super.from_pretrained(pretrained_model_name_or_path, {
|
|
15080
|
-
// Update default model file name if not provided
|
|
15081
|
-
model_file_name: 'text_model',
|
|
15082
15119
|
...options,
|
|
15120
|
+
// Update default model file name if not provided
|
|
15121
|
+
model_file_name: options.model_file_name ?? 'text_model',
|
|
15083
15122
|
});
|
|
15084
15123
|
}
|
|
15085
15124
|
}
|
|
@@ -15114,9 +15153,9 @@ class SiglipVisionModel extends CLIPPreTrainedModel {
|
|
|
15114
15153
|
/** @type {typeof PreTrainedModel.from_pretrained} */
|
|
15115
15154
|
static async from_pretrained(pretrained_model_name_or_path, options = {}) {
|
|
15116
15155
|
return super.from_pretrained(pretrained_model_name_or_path, {
|
|
15117
|
-
// Update default model file name if not provided
|
|
15118
|
-
model_file_name: 'vision_model',
|
|
15119
15156
|
...options,
|
|
15157
|
+
// Update default model file name if not provided
|
|
15158
|
+
model_file_name: options.model_file_name ?? 'vision_model',
|
|
15120
15159
|
});
|
|
15121
15160
|
}
|
|
15122
15161
|
}
|
|
@@ -15173,9 +15212,9 @@ class JinaCLIPTextModel extends JinaCLIPPreTrainedModel {
|
|
|
15173
15212
|
/** @type {typeof PreTrainedModel.from_pretrained} */
|
|
15174
15213
|
static async from_pretrained(pretrained_model_name_or_path, options = {}) {
|
|
15175
15214
|
return super.from_pretrained(pretrained_model_name_or_path, {
|
|
15176
|
-
// Update default model file name if not provided
|
|
15177
|
-
model_file_name: 'text_model',
|
|
15178
15215
|
...options,
|
|
15216
|
+
// Update default model file name if not provided
|
|
15217
|
+
model_file_name: options.model_file_name ?? 'text_model',
|
|
15179
15218
|
});
|
|
15180
15219
|
}
|
|
15181
15220
|
}
|
|
@@ -15184,9 +15223,9 @@ class JinaCLIPVisionModel extends JinaCLIPPreTrainedModel {
|
|
|
15184
15223
|
/** @type {typeof PreTrainedModel.from_pretrained} */
|
|
15185
15224
|
static async from_pretrained(pretrained_model_name_or_path, options = {}) {
|
|
15186
15225
|
return super.from_pretrained(pretrained_model_name_or_path, {
|
|
15187
|
-
// Update default model file name if not provided
|
|
15188
|
-
model_file_name: 'vision_model',
|
|
15189
15226
|
...options,
|
|
15227
|
+
// Update default model file name if not provided
|
|
15228
|
+
model_file_name: options.model_file_name ?? 'vision_model',
|
|
15190
15229
|
});
|
|
15191
15230
|
}
|
|
15192
15231
|
}
|
|
@@ -15533,6 +15572,7 @@ class Qwen2VLForConditionalGeneration extends Qwen2VLPreTrainedModel {
|
|
|
15533
15572
|
const image_nums = vision_tokens.filter(x => x == image_token_id).length;
|
|
15534
15573
|
const video_nums = vision_tokens.filter(x => x == video_token_id).length;
|
|
15535
15574
|
|
|
15575
|
+
/** @type {number[][]} */
|
|
15536
15576
|
let llm_pos_ids_list = [];
|
|
15537
15577
|
let st = 0;
|
|
15538
15578
|
let remain_images = image_nums;
|
|
@@ -15602,6 +15642,7 @@ class Qwen2VLForConditionalGeneration extends Qwen2VLPreTrainedModel {
|
|
|
15602
15642
|
// NOTE: Each item in llm_pos_ids_list is an array of shape (3, text_len),
|
|
15603
15643
|
// meaning to perform concatenation along dim=1, we can do the following:
|
|
15604
15644
|
const num_items = llm_pos_ids_list.reduce((acc, x) => acc + x.length, 0);
|
|
15645
|
+
/** @type {number[]} */
|
|
15605
15646
|
const llm_positions = new Array(num_items);
|
|
15606
15647
|
let index = 0;
|
|
15607
15648
|
for (let x = 0; x < 3; ++x) {
|
|
@@ -15642,9 +15683,10 @@ class Qwen2VLForConditionalGeneration extends Qwen2VLPreTrainedModel {
|
|
|
15642
15683
|
{ length: 3 * data.length },
|
|
15643
15684
|
(_, i) => data[i % data.length]
|
|
15644
15685
|
);
|
|
15686
|
+
/** @type {bigint[]} */
|
|
15645
15687
|
const mrope_position_deltas = Array.from(
|
|
15646
15688
|
{ length: dims[0] },
|
|
15647
|
-
(_, i) => (0,_utils_maths_js__WEBPACK_IMPORTED_MODULE_11__.max)(data.subarray(dims[1] * i, dims[1] * (i + 1)))[0] +
|
|
15689
|
+
(_, i) => (0,_utils_maths_js__WEBPACK_IMPORTED_MODULE_11__.max)(data.subarray(dims[1] * i, dims[1] * (i + 1)))[0] + 1n + BigInt(dims[1])
|
|
15648
15690
|
);
|
|
15649
15691
|
|
|
15650
15692
|
return [
|
|
@@ -16215,7 +16257,7 @@ class DPTModel extends DPTPreTrainedModel { }
|
|
|
16215
16257
|
*
|
|
16216
16258
|
* **Example:** Depth estimation w/ `Xenova/dpt-hybrid-midas`.
|
|
16217
16259
|
* ```javascript
|
|
16218
|
-
* import { DPTForDepthEstimation, AutoProcessor, RawImage,
|
|
16260
|
+
* import { DPTForDepthEstimation, AutoProcessor, RawImage, interpolate_4d } from '@huggingface/transformers';
|
|
16219
16261
|
*
|
|
16220
16262
|
* // Load model and processor
|
|
16221
16263
|
* const model_id = 'Xenova/dpt-hybrid-midas';
|
|
@@ -16224,7 +16266,7 @@ class DPTModel extends DPTPreTrainedModel { }
|
|
|
16224
16266
|
*
|
|
16225
16267
|
* // Load image from URL
|
|
16226
16268
|
* const url = 'http://images.cocodataset.org/val2017/000000039769.jpg';
|
|
16227
|
-
* const image = await RawImage.
|
|
16269
|
+
* const image = await RawImage.read(url);
|
|
16228
16270
|
*
|
|
16229
16271
|
* // Prepare image for the model
|
|
16230
16272
|
* const inputs = await processor(image);
|
|
@@ -16233,10 +16275,15 @@ class DPTModel extends DPTPreTrainedModel { }
|
|
|
16233
16275
|
* const { predicted_depth } = await model(inputs);
|
|
16234
16276
|
*
|
|
16235
16277
|
* // Interpolate to original size
|
|
16236
|
-
* const prediction =
|
|
16278
|
+
* const prediction = (await interpolate_4d(predicted_depth.unsqueeze(1), {
|
|
16279
|
+
* size: image.size.reverse(),
|
|
16280
|
+
* mode: 'bilinear',
|
|
16281
|
+
* })).squeeze(1);
|
|
16237
16282
|
*
|
|
16238
16283
|
* // Visualize the prediction
|
|
16239
|
-
* const
|
|
16284
|
+
* const min = prediction.min().item();
|
|
16285
|
+
* const max = prediction.max().item();
|
|
16286
|
+
* const formatted = prediction.sub_(min).div_(max - min).mul_(255).to('uint8');
|
|
16240
16287
|
* const depth = RawImage.fromTensor(formatted);
|
|
16241
16288
|
* // RawImage {
|
|
16242
16289
|
* // data: Uint8Array(307200) [ 85, 85, 84, ... ],
|
|
@@ -16286,11 +16333,7 @@ class GLPNPreTrainedModel extends PreTrainedModel { }
|
|
|
16286
16333
|
class GLPNModel extends GLPNPreTrainedModel { }
|
|
16287
16334
|
|
|
16288
16335
|
/**
|
|
16289
|
-
*
|
|
16290
|
-
*
|
|
16291
|
-
* **Example:** Depth estimation w/ `Xenova/glpn-kitti`.
|
|
16292
|
-
* ```javascript
|
|
16293
|
-
* import { GLPNForDepthEstimation, AutoProcessor, RawImage, interpolate, max } from '@huggingface/transformers';
|
|
16336
|
+
* import { GLPNForDepthEstimation, AutoProcessor, RawImage, interpolate_4d } from '@huggingface/transformers';
|
|
16294
16337
|
*
|
|
16295
16338
|
* // Load model and processor
|
|
16296
16339
|
* const model_id = 'Xenova/glpn-kitti';
|
|
@@ -16299,7 +16342,7 @@ class GLPNModel extends GLPNPreTrainedModel { }
|
|
|
16299
16342
|
*
|
|
16300
16343
|
* // Load image from URL
|
|
16301
16344
|
* const url = 'http://images.cocodataset.org/val2017/000000039769.jpg';
|
|
16302
|
-
* const image = await RawImage.
|
|
16345
|
+
* const image = await RawImage.read(url);
|
|
16303
16346
|
*
|
|
16304
16347
|
* // Prepare image for the model
|
|
16305
16348
|
* const inputs = await processor(image);
|
|
@@ -16308,13 +16351,18 @@ class GLPNModel extends GLPNPreTrainedModel { }
|
|
|
16308
16351
|
* const { predicted_depth } = await model(inputs);
|
|
16309
16352
|
*
|
|
16310
16353
|
* // Interpolate to original size
|
|
16311
|
-
* const prediction =
|
|
16354
|
+
* const prediction = (await interpolate_4d(predicted_depth.unsqueeze(1), {
|
|
16355
|
+
* size: image.size.reverse(),
|
|
16356
|
+
* mode: 'bilinear',
|
|
16357
|
+
* })).squeeze(1);
|
|
16312
16358
|
*
|
|
16313
16359
|
* // Visualize the prediction
|
|
16314
|
-
* const
|
|
16360
|
+
* const min = prediction.min().item();
|
|
16361
|
+
* const max = prediction.max().item();
|
|
16362
|
+
* const formatted = prediction.sub_(min).div_(max - min).mul_(255).to('uint8');
|
|
16315
16363
|
* const depth = RawImage.fromTensor(formatted);
|
|
16316
16364
|
* // RawImage {
|
|
16317
|
-
* // data: Uint8Array(307200) [
|
|
16365
|
+
* // data: Uint8Array(307200) [ 85, 85, 84, ... ],
|
|
16318
16366
|
* // width: 640,
|
|
16319
16367
|
* // height: 480,
|
|
16320
16368
|
* // channels: 1
|
|
@@ -16469,6 +16517,26 @@ class Dinov2ForImageClassification extends Dinov2PreTrainedModel {
|
|
|
16469
16517
|
}
|
|
16470
16518
|
//////////////////////////////////////////////////
|
|
16471
16519
|
|
|
16520
|
+
//////////////////////////////////////////////////
|
|
16521
|
+
class Dinov2WithRegistersPreTrainedModel extends PreTrainedModel { }
|
|
16522
|
+
|
|
16523
|
+
/**
|
|
16524
|
+
* The bare Dinov2WithRegisters Model transformer outputting raw hidden-states without any specific head on top.
|
|
16525
|
+
*/
|
|
16526
|
+
class Dinov2WithRegistersModel extends Dinov2WithRegistersPreTrainedModel { }
|
|
16527
|
+
|
|
16528
|
+
/**
|
|
16529
|
+
* Dinov2WithRegisters Model transformer with an image classification head on top (a linear layer on top of the final hidden state of the [CLS] token) e.g. for ImageNet.
|
|
16530
|
+
*/
|
|
16531
|
+
class Dinov2WithRegistersForImageClassification extends Dinov2WithRegistersPreTrainedModel {
|
|
16532
|
+
/**
|
|
16533
|
+
* @param {any} model_inputs
|
|
16534
|
+
*/
|
|
16535
|
+
async _call(model_inputs) {
|
|
16536
|
+
return new SequenceClassifierOutput(await super._call(model_inputs));
|
|
16537
|
+
}
|
|
16538
|
+
}
|
|
16539
|
+
//////////////////////////////////////////////////
|
|
16472
16540
|
|
|
16473
16541
|
//////////////////////////////////////////////////
|
|
16474
16542
|
class YolosPreTrainedModel extends PreTrainedModel { }
|
|
@@ -17261,10 +17329,12 @@ class SpeechT5ForTextToSpeech extends SpeechT5PreTrainedModel {
|
|
|
17261
17329
|
|
|
17262
17330
|
const { encoder_outputs, encoder_attention_mask } = await encoderForward(this, model_inputs);
|
|
17263
17331
|
|
|
17332
|
+
// @ts-expect-error TS2339
|
|
17264
17333
|
const r = encoder_outputs.dims[1] / this.config.reduction_factor;
|
|
17265
17334
|
const maxlen = Math.floor(r * maxlenratio);
|
|
17266
17335
|
const minlen = Math.floor(r * minlenratio);
|
|
17267
17336
|
|
|
17337
|
+
// @ts-expect-error TS2339
|
|
17268
17338
|
const num_mel_bins = this.config.num_mel_bins;
|
|
17269
17339
|
|
|
17270
17340
|
let spectrogramParts = [];
|
|
@@ -17418,9 +17488,9 @@ class ClapTextModelWithProjection extends ClapPreTrainedModel {
|
|
|
17418
17488
|
/** @type {typeof PreTrainedModel.from_pretrained} */
|
|
17419
17489
|
static async from_pretrained(pretrained_model_name_or_path, options = {}) {
|
|
17420
17490
|
return super.from_pretrained(pretrained_model_name_or_path, {
|
|
17421
|
-
// Update default model file name if not provided
|
|
17422
|
-
model_file_name: 'text_model',
|
|
17423
17491
|
...options,
|
|
17492
|
+
// Update default model file name if not provided
|
|
17493
|
+
model_file_name: options.model_file_name ?? 'text_model',
|
|
17424
17494
|
});
|
|
17425
17495
|
}
|
|
17426
17496
|
}
|
|
@@ -17455,9 +17525,9 @@ class ClapAudioModelWithProjection extends ClapPreTrainedModel {
|
|
|
17455
17525
|
/** @type {typeof PreTrainedModel.from_pretrained} */
|
|
17456
17526
|
static async from_pretrained(pretrained_model_name_or_path, options = {}) {
|
|
17457
17527
|
return super.from_pretrained(pretrained_model_name_or_path, {
|
|
17458
|
-
// Update default model file name if not provided
|
|
17459
|
-
model_file_name: 'audio_model',
|
|
17460
17528
|
...options,
|
|
17529
|
+
// Update default model file name if not provided
|
|
17530
|
+
model_file_name: options.model_file_name ?? 'audio_model',
|
|
17461
17531
|
});
|
|
17462
17532
|
}
|
|
17463
17533
|
}
|
|
@@ -17629,11 +17699,13 @@ class MusicgenForConditionalGeneration extends PreTrainedModel { // NOTE: not Mu
|
|
|
17629
17699
|
*/
|
|
17630
17700
|
_apply_and_filter_by_delay_pattern_mask(outputs) {
|
|
17631
17701
|
const [bs_x_codebooks, seqLength] = outputs.dims;
|
|
17702
|
+
// @ts-expect-error TS2339
|
|
17632
17703
|
const num_codebooks = this.config.decoder.num_codebooks;
|
|
17633
17704
|
const upperBound = (seqLength - num_codebooks);
|
|
17634
17705
|
|
|
17635
17706
|
let newDataSize = 0;
|
|
17636
17707
|
for (let i = 0; i < outputs.size; ++i) {
|
|
17708
|
+
// @ts-expect-error TS2339
|
|
17637
17709
|
if (outputs.data[i] === this.config.decoder.pad_token_id) {
|
|
17638
17710
|
continue;
|
|
17639
17711
|
}
|
|
@@ -17663,7 +17735,9 @@ class MusicgenForConditionalGeneration extends PreTrainedModel { // NOTE: not Mu
|
|
|
17663
17735
|
let clonedInputIds = structuredClone(input_ids);
|
|
17664
17736
|
for (let i = 0; i < clonedInputIds.length; ++i) {
|
|
17665
17737
|
for (let j = 0; j < clonedInputIds[i].length; ++j) {
|
|
17738
|
+
// @ts-expect-error TS2339
|
|
17666
17739
|
if ((i % this.config.decoder.num_codebooks) >= j) {
|
|
17740
|
+
// @ts-expect-error TS2339
|
|
17667
17741
|
clonedInputIds[i][j] = BigInt(this.config.decoder.pad_token_id);
|
|
17668
17742
|
}
|
|
17669
17743
|
}
|
|
@@ -17820,6 +17894,9 @@ class MultiModalityCausalLM extends MultiModalityPreTrainedModel {
|
|
|
17820
17894
|
'past_key_values',
|
|
17821
17895
|
];
|
|
17822
17896
|
|
|
17897
|
+
/**
|
|
17898
|
+
* @param {ConstructorParameters<typeof MultiModalityPreTrainedModel>} args
|
|
17899
|
+
*/
|
|
17823
17900
|
constructor(...args) {
|
|
17824
17901
|
super(...args);
|
|
17825
17902
|
|
|
@@ -18098,6 +18175,7 @@ const MODEL_MAPPING_NAMES_ENCODER_ONLY = new Map([
|
|
|
18098
18175
|
['convnext', ['ConvNextModel', ConvNextModel]],
|
|
18099
18176
|
['convnextv2', ['ConvNextV2Model', ConvNextV2Model]],
|
|
18100
18177
|
['dinov2', ['Dinov2Model', Dinov2Model]],
|
|
18178
|
+
['dinov2_with_registers', ['Dinov2WithRegistersModel', Dinov2WithRegistersModel]],
|
|
18101
18179
|
['resnet', ['ResNetModel', ResNetModel]],
|
|
18102
18180
|
['swin', ['SwinModel', SwinModel]],
|
|
18103
18181
|
['swin2sr', ['Swin2SRModel', Swin2SRModel]],
|
|
@@ -18343,6 +18421,7 @@ const MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES = new Map([
|
|
|
18343
18421
|
['convnext', ['ConvNextForImageClassification', ConvNextForImageClassification]],
|
|
18344
18422
|
['convnextv2', ['ConvNextV2ForImageClassification', ConvNextV2ForImageClassification]],
|
|
18345
18423
|
['dinov2', ['Dinov2ForImageClassification', Dinov2ForImageClassification]],
|
|
18424
|
+
['dinov2_with_registers', ['Dinov2WithRegistersForImageClassification', Dinov2WithRegistersForImageClassification]],
|
|
18346
18425
|
['resnet', ['ResNetForImageClassification', ResNetForImageClassification]],
|
|
18347
18426
|
['swin', ['SwinForImageClassification', SwinForImageClassification]],
|
|
18348
18427
|
['segformer', ['SegformerForImageClassification', SegformerForImageClassification]],
|
|
@@ -18786,10 +18865,17 @@ class SequenceClassifierOutput extends ModelOutput {
|
|
|
18786
18865
|
/**
|
|
18787
18866
|
* @param {Object} output The output of the model.
|
|
18788
18867
|
* @param {Tensor} output.logits classification (or regression if config.num_labels==1) scores (before SoftMax).
|
|
18868
|
+
* @param {Record<string, Tensor>} [output.attentions] Object of `torch.FloatTensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, sequence_length)`.
|
|
18869
|
+
* Attentions weights after the attention softmax, used to compute the weighted average in the self-attention heads.
|
|
18789
18870
|
*/
|
|
18790
|
-
constructor({ logits }) {
|
|
18871
|
+
constructor({ logits, ...attentions }) {
|
|
18791
18872
|
super();
|
|
18792
18873
|
this.logits = logits;
|
|
18874
|
+
const attentions_list = Object.values(attentions);
|
|
18875
|
+
if (attentions_list.length > 0) {
|
|
18876
|
+
// Only set attentions if they are not empty
|
|
18877
|
+
this.attentions = attentions_list;
|
|
18878
|
+
}
|
|
18793
18879
|
}
|
|
18794
18880
|
}
|
|
18795
18881
|
|
|
@@ -19045,22 +19131,6 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
19045
19131
|
|
|
19046
19132
|
class AutoFeatureExtractor {
|
|
19047
19133
|
|
|
19048
|
-
/**
|
|
19049
|
-
* Instantiate one of the feature extractor classes of the library from a pretrained model.
|
|
19050
|
-
*
|
|
19051
|
-
* The processor class to instantiate is selected based on the `feature_extractor_type` property of
|
|
19052
|
-
* the config object (either passed as an argument or loaded from `pretrained_model_name_or_path` if possible)
|
|
19053
|
-
*
|
|
19054
|
-
* @param {string} pretrained_model_name_or_path The name or path of the pretrained model. Can be either:
|
|
19055
|
-
* - A string, the *model id* of a pretrained processor hosted inside a model repo on huggingface.co.
|
|
19056
|
-
* Valid model ids can be located at the root-level, like `bert-base-uncased`, or namespaced under a
|
|
19057
|
-
* user or organization name, like `dbmdz/bert-base-german-cased`.
|
|
19058
|
-
* - A path to a *directory* containing processor files, e.g., `./my_model_directory/`.
|
|
19059
|
-
* @param {import('../../utils/hub.js').PretrainedOptions} options Additional options for loading the processor.
|
|
19060
|
-
*
|
|
19061
|
-
* @returns {Promise<AllFeatureExtractors.ImageProcessor>} A new instance of the Processor class.
|
|
19062
|
-
*/
|
|
19063
|
-
|
|
19064
19134
|
/** @type {typeof FeatureExtractor.from_pretrained} */
|
|
19065
19135
|
static async from_pretrained(pretrained_model_name_or_path, options={}) {
|
|
19066
19136
|
|
|
@@ -19187,22 +19257,6 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
19187
19257
|
*/
|
|
19188
19258
|
class AutoProcessor {
|
|
19189
19259
|
|
|
19190
|
-
/**
|
|
19191
|
-
* Instantiate one of the processor classes of the library from a pretrained model.
|
|
19192
|
-
*
|
|
19193
|
-
* The processor class to instantiate is selected based on the `image_processor_type` (or `feature_extractor_type`; legacy)
|
|
19194
|
-
* property of the config object (either passed as an argument or loaded from `pretrained_model_name_or_path` if possible)
|
|
19195
|
-
*
|
|
19196
|
-
* @param {string} pretrained_model_name_or_path The name or path of the pretrained model. Can be either:
|
|
19197
|
-
* - A string, the *model id* of a pretrained processor hosted inside a model repo on huggingface.co.
|
|
19198
|
-
* Valid model ids can be located at the root-level, like `bert-base-uncased`, or namespaced under a
|
|
19199
|
-
* user or organization name, like `dbmdz/bert-base-german-cased`.
|
|
19200
|
-
* - A path to a *directory* containing processor files, e.g., `./my_model_directory/`.
|
|
19201
|
-
* @param {import('../../utils/hub.js').PretrainedOptions} options Additional options for loading the processor.
|
|
19202
|
-
*
|
|
19203
|
-
* @returns {Promise<Processor>} A new instance of the Processor class.
|
|
19204
|
-
*/
|
|
19205
|
-
|
|
19206
19260
|
/** @type {typeof Processor.from_pretrained} */
|
|
19207
19261
|
static async from_pretrained(pretrained_model_name_or_path, options={}) {
|
|
19208
19262
|
|
|
@@ -19520,6 +19574,7 @@ class ConvNextImageProcessor extends _base_image_processors_utils_js__WEBPACK_IM
|
|
|
19520
19574
|
/**
|
|
19521
19575
|
* Percentage of the image to crop. Only has an effect if this.size < 384.
|
|
19522
19576
|
*/
|
|
19577
|
+
// @ts-expect-error TS2339
|
|
19523
19578
|
this.crop_pct = this.config.crop_pct ?? (224 / 256);
|
|
19524
19579
|
}
|
|
19525
19580
|
|
|
@@ -19722,6 +19777,7 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
19722
19777
|
class EfficientNetImageProcessor extends _base_image_processors_utils_js__WEBPACK_IMPORTED_MODULE_0__.ImageProcessor {
|
|
19723
19778
|
constructor(config) {
|
|
19724
19779
|
super(config);
|
|
19780
|
+
// @ts-expect-error TS2339
|
|
19725
19781
|
this.include_top = this.config.include_top ?? true;
|
|
19726
19782
|
if (this.include_top) {
|
|
19727
19783
|
this.image_std = this.image_std.map(x => x * x);
|
|
@@ -19803,8 +19859,11 @@ class Florence2Processor extends _base_processing_utils_js__WEBPACK_IMPORTED_MOD
|
|
|
19803
19859
|
super(config, components);
|
|
19804
19860
|
|
|
19805
19861
|
const {
|
|
19862
|
+
// @ts-expect-error TS2339
|
|
19806
19863
|
tasks_answer_post_processing_type,
|
|
19864
|
+
// @ts-expect-error TS2339
|
|
19807
19865
|
task_prompts_without_inputs,
|
|
19866
|
+
// @ts-expect-error TS2339
|
|
19808
19867
|
task_prompts_with_input,
|
|
19809
19868
|
} = this.image_processor.config;
|
|
19810
19869
|
|
|
@@ -20099,6 +20158,8 @@ class Idefics3ImageProcessor extends _base_image_processors_utils_js__WEBPACK_IM
|
|
|
20099
20158
|
|
|
20100
20159
|
const start_offset = i * pixel_attention_mask_stride + num_patches * h * w;
|
|
20101
20160
|
const end_offset = (i + 1) * pixel_attention_mask_stride;
|
|
20161
|
+
|
|
20162
|
+
// @ts-expect-error
|
|
20102
20163
|
pixel_attention_mask_data.fill(false, start_offset, end_offset);
|
|
20103
20164
|
}
|
|
20104
20165
|
}
|
|
@@ -20505,6 +20566,7 @@ class VLMImageProcessor extends _base_image_processors_utils_js__WEBPACK_IMPORTE
|
|
|
20505
20566
|
},
|
|
20506
20567
|
...config,
|
|
20507
20568
|
});
|
|
20569
|
+
// @ts-expect-error TS2339
|
|
20508
20570
|
this.constant_values = this.config.background_color.map(x => x * this.rescale_factor)
|
|
20509
20571
|
}
|
|
20510
20572
|
|
|
@@ -20946,6 +21008,8 @@ class MgpstrProcessor extends _base_processing_utils_js__WEBPACK_IMPORTED_MODULE
|
|
|
20946
21008
|
* - bpe_preds: The list of BPE decoded sentences.
|
|
20947
21009
|
* - wp_preds: The list of wp decoded sentences.
|
|
20948
21010
|
*/
|
|
21011
|
+
// @ts-expect-error The type of this method is not compatible with the one
|
|
21012
|
+
// in the base class. It might be a good idea to fix this.
|
|
20949
21013
|
batch_decode([char_logits, bpe_logits, wp_logits]) {
|
|
20950
21014
|
const [char_preds, char_scores] = this._decode_helper(char_logits, 'char');
|
|
20951
21015
|
const [bpe_preds, bpe_scores] = this._decode_helper(bpe_logits, 'bpe');
|
|
@@ -21327,6 +21391,7 @@ class PaliGemmaProcessor extends _base_processing_utils_js__WEBPACK_IMPORTED_MOD
|
|
|
21327
21391
|
}
|
|
21328
21392
|
|
|
21329
21393
|
const bos_token = this.tokenizer.bos_token;
|
|
21394
|
+
// @ts-expect-error TS2339
|
|
21330
21395
|
const image_seq_length = this.image_processor.config.image_seq_length;
|
|
21331
21396
|
let input_strings;
|
|
21332
21397
|
if (text.some((t) => t.includes(IMAGE_TOKEN))) {
|
|
@@ -21577,7 +21642,7 @@ class Phi3VProcessor extends _base_processing_utils_js__WEBPACK_IMPORTED_MODULE_
|
|
|
21577
21642
|
*
|
|
21578
21643
|
* @param {string|string[]} text
|
|
21579
21644
|
* @param {RawImage|RawImage[]} images
|
|
21580
|
-
* @param {
|
|
21645
|
+
* @param { { padding?: boolean, truncation?: boolean, num_crops?: number } | undefined } options
|
|
21581
21646
|
* @returns {Promise<any>}
|
|
21582
21647
|
*/
|
|
21583
21648
|
async _call(text, images = null, {
|
|
@@ -21761,6 +21826,7 @@ class PyAnnoteFeatureExtractor extends _base_feature_extraction_utils_js__WEBPAC
|
|
|
21761
21826
|
|
|
21762
21827
|
let current_speaker = -1;
|
|
21763
21828
|
for (let i = 0; i < scores.length; ++i) {
|
|
21829
|
+
/** @type {number[]} */
|
|
21764
21830
|
const probabilities = (0,_utils_maths_js__WEBPACK_IMPORTED_MODULE_2__.softmax)(scores[i]);
|
|
21765
21831
|
const [score, id] = (0,_utils_maths_js__WEBPACK_IMPORTED_MODULE_2__.max)(probabilities);
|
|
21766
21832
|
const [start, end] = [i, i + 1];
|
|
@@ -21945,6 +22011,7 @@ class Qwen2VLProcessor extends _base_processing_utils_js__WEBPACK_IMPORTED_MODUL
|
|
|
21945
22011
|
}
|
|
21946
22012
|
|
|
21947
22013
|
if (image_grid_thw) {
|
|
22014
|
+
// @ts-expect-error TS2551
|
|
21948
22015
|
let merge_length = this.image_processor.config.merge_size ** 2;
|
|
21949
22016
|
let index = 0;
|
|
21950
22017
|
|
|
@@ -22432,8 +22499,8 @@ class SeamlessM4TFeatureExtractor extends _base_feature_extraction_utils_js__WEB
|
|
|
22432
22499
|
'int64',
|
|
22433
22500
|
new BigInt64Array(numPaddedFrames),
|
|
22434
22501
|
[1, numPaddedFrames],
|
|
22435
|
-
)
|
|
22436
|
-
padded_attention_mask.data.fill(1n, 0, num_frames);
|
|
22502
|
+
);
|
|
22503
|
+
/** @type {BigInt64Array} */ (padded_attention_mask.data).fill(1n, 0, num_frames);
|
|
22437
22504
|
}
|
|
22438
22505
|
}
|
|
22439
22506
|
}
|
|
@@ -23233,7 +23300,7 @@ class WhisperFeatureExtractor extends _base_feature_extraction_utils_js__WEBPACK
|
|
|
23233
23300
|
)
|
|
23234
23301
|
|
|
23235
23302
|
const data = features.data;
|
|
23236
|
-
const maxValue = (0,_utils_maths_js__WEBPACK_IMPORTED_MODULE_3__.max)(data)[0];
|
|
23303
|
+
const maxValue = (0,_utils_maths_js__WEBPACK_IMPORTED_MODULE_3__.max)(/** @type {Float32Array} */(data))[0];
|
|
23237
23304
|
|
|
23238
23305
|
for (let i = 0; i < data.length; ++i) {
|
|
23239
23306
|
data[i] = (Math.max(data[i], maxValue - 8.0) + 4.0) / 4.0;
|
|
@@ -23492,6 +23559,16 @@ class TensorOpRegistry {
|
|
|
23492
23559
|
// executionProviders: ['webgpu'],
|
|
23493
23560
|
};
|
|
23494
23561
|
|
|
23562
|
+
static get nearest_interpolate_4d() {
|
|
23563
|
+
if (!this._nearest_interpolate_4d) {
|
|
23564
|
+
this._nearest_interpolate_4d = wrap(
|
|
23565
|
+
[8, 10, 18, 0, 58, 129, 1, 10, 41, 10, 1, 120, 10, 0, 10, 0, 10, 1, 115, 18, 1, 121, 34, 6, 82, 101, 115, 105, 122, 101, 42, 18, 10, 4, 109, 111, 100, 101, 34, 7, 110, 101, 97, 114, 101, 115, 116, 160, 1, 3, 18, 1, 114, 90, 31, 10, 1, 120, 18, 26, 10, 24, 8, 1, 18, 20, 10, 3, 18, 1, 98, 10, 3, 18, 1, 99, 10, 3, 18, 1, 104, 10, 3, 18, 1, 119, 90, 15, 10, 1, 115, 18, 10, 10, 8, 8, 7, 18, 4, 10, 2, 8, 4, 98, 31, 10, 1, 121, 18, 26, 10, 24, 8, 1, 18, 20, 10, 3, 18, 1, 98, 10, 3, 18, 1, 99, 10, 3, 18, 1, 104, 10, 3, 18, 1, 119, 66, 2, 16, 21],
|
|
23566
|
+
this.session_options,
|
|
23567
|
+
'y',
|
|
23568
|
+
);
|
|
23569
|
+
}
|
|
23570
|
+
return this._nearest_interpolate_4d;
|
|
23571
|
+
}
|
|
23495
23572
|
static get bilinear_interpolate_4d() {
|
|
23496
23573
|
if (!this._bilinear_interpolate_4d) {
|
|
23497
23574
|
this._bilinear_interpolate_4d = wrap(
|
|
@@ -23865,6 +23942,7 @@ class TextClassificationPipeline extends (/** @type {new (options: TextPipelineC
|
|
|
23865
23942
|
|
|
23866
23943
|
// TODO: Use softmax tensor function
|
|
23867
23944
|
const function_to_apply =
|
|
23945
|
+
// @ts-expect-error TS2339
|
|
23868
23946
|
this.model.config.problem_type === 'multi_label_classification'
|
|
23869
23947
|
? batch => batch.sigmoid()
|
|
23870
23948
|
: batch => new _utils_tensor_js__WEBPACK_IMPORTED_MODULE_8__.Tensor(
|
|
@@ -23873,6 +23951,7 @@ class TextClassificationPipeline extends (/** @type {new (options: TextPipelineC
|
|
|
23873
23951
|
batch.dims,
|
|
23874
23952
|
); // single_label_classification (default)
|
|
23875
23953
|
|
|
23954
|
+
// @ts-expect-error TS2339
|
|
23876
23955
|
const id2label = this.model.config.id2label;
|
|
23877
23956
|
|
|
23878
23957
|
const toReturn = [];
|
|
@@ -23975,6 +24054,7 @@ class TokenClassificationPipeline extends (/** @type {new (options: TextPipeline
|
|
|
23975
24054
|
const outputs = await this.model(model_inputs)
|
|
23976
24055
|
|
|
23977
24056
|
const logits = outputs.logits;
|
|
24057
|
+
// @ts-expect-error TS2339
|
|
23978
24058
|
const id2label = this.model.config.id2label;
|
|
23979
24059
|
|
|
23980
24060
|
const toReturn = [];
|
|
@@ -24314,11 +24394,14 @@ class Text2TextGenerationPipeline extends (/** @type {new (options: TextPipeline
|
|
|
24314
24394
|
|
|
24315
24395
|
|
|
24316
24396
|
// Add global prefix, if present
|
|
24397
|
+
// @ts-expect-error TS2339
|
|
24317
24398
|
if (this.model.config.prefix) {
|
|
24399
|
+
// @ts-expect-error TS2339
|
|
24318
24400
|
texts = texts.map(x => this.model.config.prefix + x)
|
|
24319
24401
|
}
|
|
24320
24402
|
|
|
24321
24403
|
// Handle task specific params:
|
|
24404
|
+
// @ts-expect-error TS2339
|
|
24322
24405
|
const task_specific_params = this.model.config.task_specific_params
|
|
24323
24406
|
if (task_specific_params && task_specific_params[this.task]) {
|
|
24324
24407
|
// Add prefixes, if present
|
|
@@ -25057,6 +25140,7 @@ class AudioClassificationPipeline extends (/** @type {new (options: AudioPipelin
|
|
|
25057
25140
|
const sampling_rate = this.processor.feature_extractor.config.sampling_rate;
|
|
25058
25141
|
const preparedAudios = await prepareAudios(audio, sampling_rate);
|
|
25059
25142
|
|
|
25143
|
+
// @ts-expect-error TS2339
|
|
25060
25144
|
const id2label = this.model.config.id2label;
|
|
25061
25145
|
|
|
25062
25146
|
const toReturn = [];
|
|
@@ -25367,6 +25451,7 @@ class AutomaticSpeechRecognitionPipeline extends (/** @type {new (options: TextA
|
|
|
25367
25451
|
audio = [/** @type {AudioInput} */ (audio)];
|
|
25368
25452
|
}
|
|
25369
25453
|
|
|
25454
|
+
// @ts-expect-error TS2339
|
|
25370
25455
|
const time_precision = this.processor.feature_extractor.config.chunk_length / this.model.config.max_source_positions;
|
|
25371
25456
|
const hop_length = this.processor.feature_extractor.config.hop_length;
|
|
25372
25457
|
|
|
@@ -25432,7 +25517,9 @@ class AutomaticSpeechRecognitionPipeline extends (/** @type {new (options: TextA
|
|
|
25432
25517
|
|
|
25433
25518
|
// TODO: Right now we only get top beam
|
|
25434
25519
|
if (return_timestamps === 'word') {
|
|
25520
|
+
// @ts-expect-error TS2339
|
|
25435
25521
|
chunk.tokens = data.sequences.tolist()[0];
|
|
25522
|
+
// @ts-expect-error TS2339
|
|
25436
25523
|
chunk.token_timestamps = data.token_timestamps.tolist()[0].map(
|
|
25437
25524
|
(/** @type {number} */ x) => (0,_utils_maths_js__WEBPACK_IMPORTED_MODULE_6__.round)(x, 2)
|
|
25438
25525
|
);
|
|
@@ -25477,7 +25564,7 @@ class AutomaticSpeechRecognitionPipeline extends (/** @type {new (options: TextA
|
|
|
25477
25564
|
const max_new_tokens = Math.floor(aud.length / sampling_rate) * 6;
|
|
25478
25565
|
const outputs = await this.model.generate({ max_new_tokens, ...kwargs, ...inputs });
|
|
25479
25566
|
|
|
25480
|
-
const text = this.processor.batch_decode(outputs, { skip_special_tokens: true })[0];
|
|
25567
|
+
const text = this.processor.batch_decode(/** @type {Tensor} */(outputs), { skip_special_tokens: true })[0];
|
|
25481
25568
|
toReturn.push({ text });
|
|
25482
25569
|
}
|
|
25483
25570
|
return single ? toReturn[0] : toReturn;
|
|
@@ -25626,6 +25713,7 @@ class ImageClassificationPipeline extends (/** @type {new (options: ImagePipelin
|
|
|
25626
25713
|
const { pixel_values } = await this.processor(preparedImages);
|
|
25627
25714
|
const output = await this.model({ pixel_values });
|
|
25628
25715
|
|
|
25716
|
+
// @ts-expect-error TS2339
|
|
25629
25717
|
const id2label = this.model.config.id2label;
|
|
25630
25718
|
|
|
25631
25719
|
/** @type {ImageClassificationOutput[]} */
|
|
@@ -25740,6 +25828,7 @@ class ImageSegmentationPipeline extends (/** @type {new (options: ImagePipelineC
|
|
|
25740
25828
|
}
|
|
25741
25829
|
}
|
|
25742
25830
|
|
|
25831
|
+
// @ts-expect-error TS2339
|
|
25743
25832
|
const id2label = this.model.config.id2label;
|
|
25744
25833
|
|
|
25745
25834
|
/** @type {ImageSegmentationPipelineOutput[]} */
|
|
@@ -25966,6 +26055,7 @@ class ObjectDetectionPipeline extends (/** @type {new (options: ImagePipelineCon
|
|
|
25966
26055
|
const processed = this.processor.image_processor.post_process_object_detection(output, threshold, imageSizes);
|
|
25967
26056
|
|
|
25968
26057
|
// Add labels
|
|
26058
|
+
// @ts-expect-error TS2339
|
|
25969
26059
|
const id2label = this.model.config.id2label;
|
|
25970
26060
|
|
|
25971
26061
|
// Format output
|
|
@@ -26185,6 +26275,7 @@ class DocumentQuestionAnsweringPipeline extends (/** @type {new (options: TextIm
|
|
|
26185
26275
|
// Run model
|
|
26186
26276
|
const output = await this.model.generate({
|
|
26187
26277
|
inputs: pixel_values,
|
|
26278
|
+
// @ts-expect-error TS2339
|
|
26188
26279
|
max_length: this.model.config.decoder.max_position_embeddings,
|
|
26189
26280
|
decoder_input_ids,
|
|
26190
26281
|
...generate_kwargs,
|
|
@@ -26300,6 +26391,7 @@ class TextToAudioPipeline extends (/** @type {new (options: TextToAudioPipelineC
|
|
|
26300
26391
|
// Generate waveform
|
|
26301
26392
|
const { waveform } = await this.model(inputs);
|
|
26302
26393
|
|
|
26394
|
+
// @ts-expect-error TS2339
|
|
26303
26395
|
const sampling_rate = this.model.config.sampling_rate;
|
|
26304
26396
|
return {
|
|
26305
26397
|
audio: waveform.data,
|
|
@@ -26457,11 +26549,23 @@ class DepthEstimationPipeline extends (/** @type {new (options: ImagePipelineCon
|
|
|
26457
26549
|
|
|
26458
26550
|
const toReturn = [];
|
|
26459
26551
|
for (let i = 0; i < preparedImages.length; ++i) {
|
|
26460
|
-
const
|
|
26461
|
-
const
|
|
26552
|
+
const batch = predicted_depth[i];
|
|
26553
|
+
const [height, width] = batch.dims.slice(-2);
|
|
26554
|
+
const [new_width, new_height] = preparedImages[i].size;
|
|
26555
|
+
|
|
26556
|
+
// Interpolate to original size
|
|
26557
|
+
const prediction = (await (0,_utils_tensor_js__WEBPACK_IMPORTED_MODULE_8__.interpolate_4d)(batch.view(1, 1, height, width), {
|
|
26558
|
+
size: [new_height, new_width],
|
|
26559
|
+
mode: 'bilinear',
|
|
26560
|
+
})).view(new_height, new_width);
|
|
26561
|
+
|
|
26562
|
+
const minval = /** @type {number} */(prediction.min().item());
|
|
26563
|
+
const maxval = /** @type {number} */(prediction.max().item());
|
|
26564
|
+
const formatted = prediction.sub(minval).div_(maxval - minval).mul_(255).to('uint8').unsqueeze(0);
|
|
26565
|
+
const depth = _utils_image_js__WEBPACK_IMPORTED_MODULE_9__.RawImage.fromTensor(formatted);
|
|
26462
26566
|
toReturn.push({
|
|
26463
|
-
predicted_depth:
|
|
26464
|
-
depth
|
|
26567
|
+
predicted_depth: prediction,
|
|
26568
|
+
depth,
|
|
26465
26569
|
});
|
|
26466
26570
|
}
|
|
26467
26571
|
|
|
@@ -26941,6 +27045,7 @@ async function loadItems(mapping, model, pretrainedOptions) {
|
|
|
26941
27045
|
return result;
|
|
26942
27046
|
}
|
|
26943
27047
|
|
|
27048
|
+
|
|
26944
27049
|
/***/ }),
|
|
26945
27050
|
|
|
26946
27051
|
/***/ "./src/tokenizers.js":
|
|
@@ -27009,7 +27114,6 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
27009
27114
|
/* harmony import */ var _utils_data_structures_js__WEBPACK_IMPORTED_MODULE_5__ = __webpack_require__(/*! ./utils/data-structures.js */ "./src/utils/data-structures.js");
|
|
27010
27115
|
/* harmony import */ var _huggingface_jinja__WEBPACK_IMPORTED_MODULE_6__ = __webpack_require__(/*! @huggingface/jinja */ "./node_modules/@huggingface/jinja/dist/index.js");
|
|
27011
27116
|
/* harmony import */ var _models_whisper_common_whisper_js__WEBPACK_IMPORTED_MODULE_7__ = __webpack_require__(/*! ./models/whisper/common_whisper.js */ "./src/models/whisper/common_whisper.js");
|
|
27012
|
-
/* harmony import */ var _utils_constants_js__WEBPACK_IMPORTED_MODULE_8__ = __webpack_require__(/*! ./utils/constants.js */ "./src/utils/constants.js");
|
|
27013
27117
|
|
|
27014
27118
|
/**
|
|
27015
27119
|
* @file Tokenizers are used to prepare textual inputs for a model.
|
|
@@ -27046,7 +27150,6 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
27046
27150
|
|
|
27047
27151
|
|
|
27048
27152
|
|
|
27049
|
-
|
|
27050
27153
|
/**
|
|
27051
27154
|
* @typedef {Object} TokenizerProperties Additional tokenizer-specific properties.
|
|
27052
27155
|
* @property {boolean} [legacy=false] Whether or not the `legacy` behavior of the tokenizer should be used.
|
|
@@ -27530,7 +27633,7 @@ class Unigram extends TokenizerModel {
|
|
|
27530
27633
|
* Create a new Unigram tokenizer model.
|
|
27531
27634
|
* @param {Object} config The configuration object for the Unigram model.
|
|
27532
27635
|
* @param {number} config.unk_id The ID of the unknown token
|
|
27533
|
-
* @param {
|
|
27636
|
+
* @param {[string, number][]} config.vocab A 2D array representing a mapping of tokens to scores.
|
|
27534
27637
|
* @param {Object} moreConfig Additional configuration object for the Unigram model.
|
|
27535
27638
|
*/
|
|
27536
27639
|
constructor(config, moreConfig) {
|
|
@@ -27538,11 +27641,10 @@ class Unigram extends TokenizerModel {
|
|
|
27538
27641
|
|
|
27539
27642
|
const vocabSize = config.vocab.length;
|
|
27540
27643
|
this.vocab = new Array(vocabSize);
|
|
27644
|
+
/** @type {number[]} */
|
|
27541
27645
|
this.scores = new Array(vocabSize);
|
|
27542
27646
|
for (let i = 0; i < vocabSize; ++i) {
|
|
27543
|
-
|
|
27544
|
-
this.vocab[i] = piece[0];
|
|
27545
|
-
this.scores[i] = piece[1];
|
|
27647
|
+
[this.vocab[i], this.scores[i]] = config.vocab[i];
|
|
27546
27648
|
}
|
|
27547
27649
|
|
|
27548
27650
|
this.unk_token_id = config.unk_id;
|
|
@@ -32899,6 +33001,8 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
32899
33001
|
/* harmony export */ });
|
|
32900
33002
|
/* harmony import */ var _env_js__WEBPACK_IMPORTED_MODULE_0__ = __webpack_require__(/*! ../env.js */ "./src/env.js");
|
|
32901
33003
|
/* harmony import */ var _devices_js__WEBPACK_IMPORTED_MODULE_1__ = __webpack_require__(/*! ./devices.js */ "./src/utils/devices.js");
|
|
33004
|
+
/// <reference types="@webgpu/types" />
|
|
33005
|
+
|
|
32902
33006
|
|
|
32903
33007
|
|
|
32904
33008
|
|
|
@@ -33152,7 +33256,7 @@ class FileResponse {
|
|
|
33152
33256
|
*/
|
|
33153
33257
|
async arrayBuffer() {
|
|
33154
33258
|
const data = await fs__WEBPACK_IMPORTED_MODULE_0__.promises.readFile(this.filePath);
|
|
33155
|
-
return data.buffer;
|
|
33259
|
+
return /** @type {ArrayBuffer} */ (data.buffer);
|
|
33156
33260
|
}
|
|
33157
33261
|
|
|
33158
33262
|
/**
|
|
@@ -34811,8 +34915,9 @@ function magnitude(arr) {
|
|
|
34811
34915
|
|
|
34812
34916
|
/**
|
|
34813
34917
|
* Returns the value and index of the minimum element in an array.
|
|
34814
|
-
* @
|
|
34815
|
-
* @
|
|
34918
|
+
* @template {number[]|bigint[]|AnyTypedArray} T
|
|
34919
|
+
* @param {T} arr array of numbers.
|
|
34920
|
+
* @returns {T extends bigint[]|BigTypedArray ? [bigint, number] : [number, number]} the value and index of the minimum element, of the form: [valueOfMin, indexOfMin]
|
|
34816
34921
|
* @throws {Error} If array is empty.
|
|
34817
34922
|
*/
|
|
34818
34923
|
function min(arr) {
|
|
@@ -34825,14 +34930,15 @@ function min(arr) {
|
|
|
34825
34930
|
indexOfMin = i;
|
|
34826
34931
|
}
|
|
34827
34932
|
}
|
|
34828
|
-
return [min, indexOfMin];
|
|
34933
|
+
return /** @type {T extends bigint[]|BigTypedArray ? [bigint, number] : [number, number]} */([min, indexOfMin]);
|
|
34829
34934
|
}
|
|
34830
34935
|
|
|
34831
34936
|
|
|
34832
34937
|
/**
|
|
34833
34938
|
* Returns the value and index of the maximum element in an array.
|
|
34834
|
-
* @
|
|
34835
|
-
* @
|
|
34939
|
+
* @template {number[]|bigint[]|AnyTypedArray} T
|
|
34940
|
+
* @param {T} arr array of numbers.
|
|
34941
|
+
* @returns {T extends bigint[]|BigTypedArray ? [bigint, number] : [number, number]} the value and index of the maximum element, of the form: [valueOfMax, indexOfMax]
|
|
34836
34942
|
* @throws {Error} If array is empty.
|
|
34837
34943
|
*/
|
|
34838
34944
|
function max(arr) {
|
|
@@ -34845,7 +34951,7 @@ function max(arr) {
|
|
|
34845
34951
|
indexOfMax = i;
|
|
34846
34952
|
}
|
|
34847
34953
|
}
|
|
34848
|
-
return [
|
|
34954
|
+
return /** @type {T extends bigint[]|BigTypedArray ? [bigint, number] : [number, number]} */([max, indexOfMax]);
|
|
34849
34955
|
}
|
|
34850
34956
|
|
|
34851
34957
|
function isPowerOfTwo(number) {
|
|
@@ -36142,8 +36248,6 @@ class Tensor {
|
|
|
36142
36248
|
return this.permute(...dims);
|
|
36143
36249
|
}
|
|
36144
36250
|
|
|
36145
|
-
// TODO add .max() and .min() methods
|
|
36146
|
-
|
|
36147
36251
|
/**
|
|
36148
36252
|
* Returns the sum of each row of the input tensor in the given dimension dim.
|
|
36149
36253
|
*
|
|
@@ -36437,6 +36541,36 @@ class Tensor {
|
|
|
36437
36541
|
return mean(this, dim, keepdim);
|
|
36438
36542
|
}
|
|
36439
36543
|
|
|
36544
|
+
min(dim = null, keepdim = false) {
|
|
36545
|
+
if (dim !== null) {
|
|
36546
|
+
throw new Error("`dim !== null` not yet implemented.");
|
|
36547
|
+
}
|
|
36548
|
+
const value = (0,_maths_js__WEBPACK_IMPORTED_MODULE_0__.min)(this.data)[0];
|
|
36549
|
+
return new Tensor(this.type, [value], []);
|
|
36550
|
+
}
|
|
36551
|
+
max(dim = null, keepdim = false) {
|
|
36552
|
+
if (dim !== null) {
|
|
36553
|
+
throw new Error("`dim !== null` not yet implemented.");
|
|
36554
|
+
}
|
|
36555
|
+
const value = (0,_maths_js__WEBPACK_IMPORTED_MODULE_0__.max)(this.data)[0];
|
|
36556
|
+
return new Tensor(this.type, [value], []);
|
|
36557
|
+
}
|
|
36558
|
+
|
|
36559
|
+
argmin(dim = null, keepdim = false) {
|
|
36560
|
+
if (dim !== null) {
|
|
36561
|
+
throw new Error("`dim !== null` not yet implemented.");
|
|
36562
|
+
}
|
|
36563
|
+
const index = (0,_maths_js__WEBPACK_IMPORTED_MODULE_0__.min)(this.data)[1];
|
|
36564
|
+
return new Tensor('int64', [BigInt(index)], []);
|
|
36565
|
+
}
|
|
36566
|
+
argmax(dim = null, keepdim = false) {
|
|
36567
|
+
if (dim !== null) {
|
|
36568
|
+
throw new Error("`dim !== null` not yet implemented.");
|
|
36569
|
+
}
|
|
36570
|
+
const index = (0,_maths_js__WEBPACK_IMPORTED_MODULE_0__.max)(this.data)[1];
|
|
36571
|
+
return new Tensor('int64', [BigInt(index)], []);
|
|
36572
|
+
}
|
|
36573
|
+
|
|
36440
36574
|
/**
|
|
36441
36575
|
* Performs Tensor dtype conversion.
|
|
36442
36576
|
* @param {DataType} type The desired data type.
|
|
@@ -36570,7 +36704,7 @@ function interpolate(input, [out_height, out_width], mode = 'bilinear', align_co
|
|
|
36570
36704
|
* @param {Tensor} input the input tensor
|
|
36571
36705
|
* @param {Object} options the options for the interpolation
|
|
36572
36706
|
* @param {[number, number]|[number, number, number]|[number, number, number, number]} [options.size=null] output spatial size.
|
|
36573
|
-
* @param {"bilinear"|"bicubic"} [options.mode='bilinear'] algorithm used for upsampling
|
|
36707
|
+
* @param {"nearest"|"bilinear"|"bicubic"} [options.mode='bilinear'] algorithm used for upsampling
|
|
36574
36708
|
* @returns {Promise<Tensor>} The interpolated tensor.
|
|
36575
36709
|
*/
|
|
36576
36710
|
async function interpolate_4d(input, {
|
|
@@ -36600,7 +36734,9 @@ async function interpolate_4d(input, {
|
|
|
36600
36734
|
}
|
|
36601
36735
|
|
|
36602
36736
|
let op;
|
|
36603
|
-
if (mode === '
|
|
36737
|
+
if (mode === 'nearest') {
|
|
36738
|
+
op = await _ops_registry_js__WEBPACK_IMPORTED_MODULE_2__.TensorOpRegistry.nearest_interpolate_4d;
|
|
36739
|
+
} else if (mode === 'bilinear') {
|
|
36604
36740
|
op = await _ops_registry_js__WEBPACK_IMPORTED_MODULE_2__.TensorOpRegistry.bilinear_interpolate_4d;
|
|
36605
36741
|
} else if (mode === 'bicubic') {
|
|
36606
36742
|
op = await _ops_registry_js__WEBPACK_IMPORTED_MODULE_2__.TensorOpRegistry.bicubic_interpolate_4d;
|
|
@@ -36641,13 +36777,13 @@ async function rfft(x, a) {
|
|
|
36641
36777
|
* Returns the k largest elements of the given input tensor.
|
|
36642
36778
|
* Inspired by https://pytorch.org/docs/stable/generated/torch.topk.html
|
|
36643
36779
|
* @param {Tensor} x the input tensor
|
|
36644
|
-
* @param {number} k the k in "top-k"
|
|
36780
|
+
* @param {number} [k] the k in "top-k"
|
|
36645
36781
|
* @returns {Promise<[Tensor, Tensor]>} the output tuple of (Tensor, LongTensor) of top-k elements and their indices.
|
|
36646
36782
|
*/
|
|
36647
36783
|
async function topk(x, k) {
|
|
36648
36784
|
const op = await _ops_registry_js__WEBPACK_IMPORTED_MODULE_2__.TensorOpRegistry.top_k;
|
|
36649
36785
|
|
|
36650
|
-
if (k
|
|
36786
|
+
if (k == null) {
|
|
36651
36787
|
k = x.dims.at(-1);
|
|
36652
36788
|
} else {
|
|
36653
36789
|
k = Math.min(k, x.dims.at(-1));
|
|
@@ -36676,10 +36812,10 @@ const arrayToIndexTensor = (array) => new Tensor('int64', array, [array.length])
|
|
|
36676
36812
|
async function slice(data, starts, ends, axes, steps) {
|
|
36677
36813
|
const op = await _ops_registry_js__WEBPACK_IMPORTED_MODULE_2__.TensorOpRegistry.slice;
|
|
36678
36814
|
return await op({
|
|
36679
|
-
x: data,
|
|
36680
|
-
s: arrayToIndexTensor(starts),
|
|
36681
|
-
e: arrayToIndexTensor(ends),
|
|
36682
|
-
a: arrayToIndexTensor(axes),
|
|
36815
|
+
x: data,
|
|
36816
|
+
s: arrayToIndexTensor(starts),
|
|
36817
|
+
e: arrayToIndexTensor(ends),
|
|
36818
|
+
a: arrayToIndexTensor(axes),
|
|
36683
36819
|
t: arrayToIndexTensor(steps ?? new Array(axes.length).fill(1)),
|
|
36684
36820
|
});
|
|
36685
36821
|
}
|
|
@@ -37497,6 +37633,9 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
37497
37633
|
/* harmony export */ Dinov2ForImageClassification: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Dinov2ForImageClassification),
|
|
37498
37634
|
/* harmony export */ Dinov2Model: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Dinov2Model),
|
|
37499
37635
|
/* harmony export */ Dinov2PreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Dinov2PreTrainedModel),
|
|
37636
|
+
/* harmony export */ Dinov2WithRegistersForImageClassification: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Dinov2WithRegistersForImageClassification),
|
|
37637
|
+
/* harmony export */ Dinov2WithRegistersModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Dinov2WithRegistersModel),
|
|
37638
|
+
/* harmony export */ Dinov2WithRegistersPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Dinov2WithRegistersPreTrainedModel),
|
|
37500
37639
|
/* harmony export */ DistilBertForMaskedLM: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.DistilBertForMaskedLM),
|
|
37501
37640
|
/* harmony export */ DistilBertForQuestionAnswering: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.DistilBertForQuestionAnswering),
|
|
37502
37641
|
/* harmony export */ DistilBertForSequenceClassification: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.DistilBertForSequenceClassification),
|
|
@@ -38274,6 +38413,9 @@ var __webpack_exports__DetrSegmentationOutput = __webpack_exports__.DetrSegmenta
|
|
|
38274
38413
|
var __webpack_exports__Dinov2ForImageClassification = __webpack_exports__.Dinov2ForImageClassification;
|
|
38275
38414
|
var __webpack_exports__Dinov2Model = __webpack_exports__.Dinov2Model;
|
|
38276
38415
|
var __webpack_exports__Dinov2PreTrainedModel = __webpack_exports__.Dinov2PreTrainedModel;
|
|
38416
|
+
var __webpack_exports__Dinov2WithRegistersForImageClassification = __webpack_exports__.Dinov2WithRegistersForImageClassification;
|
|
38417
|
+
var __webpack_exports__Dinov2WithRegistersModel = __webpack_exports__.Dinov2WithRegistersModel;
|
|
38418
|
+
var __webpack_exports__Dinov2WithRegistersPreTrainedModel = __webpack_exports__.Dinov2WithRegistersPreTrainedModel;
|
|
38277
38419
|
var __webpack_exports__DistilBertForMaskedLM = __webpack_exports__.DistilBertForMaskedLM;
|
|
38278
38420
|
var __webpack_exports__DistilBertForQuestionAnswering = __webpack_exports__.DistilBertForQuestionAnswering;
|
|
38279
38421
|
var __webpack_exports__DistilBertForSequenceClassification = __webpack_exports__.DistilBertForSequenceClassification;
|
|
@@ -38812,6 +38954,6 @@ var __webpack_exports__topk = __webpack_exports__.topk;
|
|
|
38812
38954
|
var __webpack_exports__window_function = __webpack_exports__.window_function;
|
|
38813
38955
|
var __webpack_exports__zeros = __webpack_exports__.zeros;
|
|
38814
38956
|
var __webpack_exports__zeros_like = __webpack_exports__.zeros_like;
|
|
38815
|
-
export { __webpack_exports__ASTFeatureExtractor as ASTFeatureExtractor, __webpack_exports__ASTForAudioClassification as ASTForAudioClassification, __webpack_exports__ASTModel as ASTModel, __webpack_exports__ASTPreTrainedModel as ASTPreTrainedModel, __webpack_exports__AlbertForMaskedLM as AlbertForMaskedLM, __webpack_exports__AlbertForQuestionAnswering as AlbertForQuestionAnswering, __webpack_exports__AlbertForSequenceClassification as AlbertForSequenceClassification, __webpack_exports__AlbertModel as AlbertModel, __webpack_exports__AlbertPreTrainedModel as AlbertPreTrainedModel, __webpack_exports__AlbertTokenizer as AlbertTokenizer, __webpack_exports__AudioClassificationPipeline as AudioClassificationPipeline, __webpack_exports__AutoConfig as AutoConfig, __webpack_exports__AutoFeatureExtractor as AutoFeatureExtractor, __webpack_exports__AutoImageProcessor as AutoImageProcessor, __webpack_exports__AutoModel as AutoModel, __webpack_exports__AutoModelForAudioClassification as AutoModelForAudioClassification, __webpack_exports__AutoModelForAudioFrameClassification as AutoModelForAudioFrameClassification, __webpack_exports__AutoModelForCTC as AutoModelForCTC, __webpack_exports__AutoModelForCausalLM as AutoModelForCausalLM, __webpack_exports__AutoModelForDepthEstimation as AutoModelForDepthEstimation, __webpack_exports__AutoModelForDocumentQuestionAnswering as AutoModelForDocumentQuestionAnswering, __webpack_exports__AutoModelForImageClassification as AutoModelForImageClassification, __webpack_exports__AutoModelForImageFeatureExtraction as AutoModelForImageFeatureExtraction, __webpack_exports__AutoModelForImageMatting as AutoModelForImageMatting, __webpack_exports__AutoModelForImageSegmentation as AutoModelForImageSegmentation, __webpack_exports__AutoModelForImageToImage as AutoModelForImageToImage, __webpack_exports__AutoModelForMaskGeneration as AutoModelForMaskGeneration, __webpack_exports__AutoModelForMaskedLM as AutoModelForMaskedLM, __webpack_exports__AutoModelForNormalEstimation as AutoModelForNormalEstimation, __webpack_exports__AutoModelForObjectDetection as AutoModelForObjectDetection, __webpack_exports__AutoModelForPoseEstimation as AutoModelForPoseEstimation, __webpack_exports__AutoModelForQuestionAnswering as AutoModelForQuestionAnswering, __webpack_exports__AutoModelForSemanticSegmentation as AutoModelForSemanticSegmentation, __webpack_exports__AutoModelForSeq2SeqLM as AutoModelForSeq2SeqLM, __webpack_exports__AutoModelForSequenceClassification as AutoModelForSequenceClassification, __webpack_exports__AutoModelForSpeechSeq2Seq as AutoModelForSpeechSeq2Seq, __webpack_exports__AutoModelForTextToSpectrogram as AutoModelForTextToSpectrogram, __webpack_exports__AutoModelForTextToWaveform as AutoModelForTextToWaveform, __webpack_exports__AutoModelForTokenClassification as AutoModelForTokenClassification, __webpack_exports__AutoModelForUniversalSegmentation as AutoModelForUniversalSegmentation, __webpack_exports__AutoModelForVision2Seq as AutoModelForVision2Seq, __webpack_exports__AutoModelForXVector as AutoModelForXVector, __webpack_exports__AutoModelForZeroShotObjectDetection as AutoModelForZeroShotObjectDetection, __webpack_exports__AutoProcessor as AutoProcessor, __webpack_exports__AutoTokenizer as AutoTokenizer, __webpack_exports__AutomaticSpeechRecognitionPipeline as AutomaticSpeechRecognitionPipeline, __webpack_exports__BartForConditionalGeneration as BartForConditionalGeneration, __webpack_exports__BartForSequenceClassification as BartForSequenceClassification, __webpack_exports__BartModel as BartModel, __webpack_exports__BartPretrainedModel as BartPretrainedModel, __webpack_exports__BartTokenizer as BartTokenizer, __webpack_exports__BaseModelOutput as BaseModelOutput, __webpack_exports__BaseStreamer as BaseStreamer, __webpack_exports__BeitFeatureExtractor as BeitFeatureExtractor, __webpack_exports__BeitForImageClassification as BeitForImageClassification, __webpack_exports__BeitModel as BeitModel, __webpack_exports__BeitPreTrainedModel as BeitPreTrainedModel, __webpack_exports__BertForMaskedLM as BertForMaskedLM, __webpack_exports__BertForQuestionAnswering as BertForQuestionAnswering, __webpack_exports__BertForSequenceClassification as BertForSequenceClassification, __webpack_exports__BertForTokenClassification as BertForTokenClassification, __webpack_exports__BertModel as BertModel, __webpack_exports__BertPreTrainedModel as BertPreTrainedModel, __webpack_exports__BertTokenizer as BertTokenizer, __webpack_exports__BitImageProcessor as BitImageProcessor, __webpack_exports__BlenderbotForConditionalGeneration as BlenderbotForConditionalGeneration, __webpack_exports__BlenderbotModel as BlenderbotModel, __webpack_exports__BlenderbotPreTrainedModel as BlenderbotPreTrainedModel, __webpack_exports__BlenderbotSmallForConditionalGeneration as BlenderbotSmallForConditionalGeneration, __webpack_exports__BlenderbotSmallModel as BlenderbotSmallModel, __webpack_exports__BlenderbotSmallPreTrainedModel as BlenderbotSmallPreTrainedModel, __webpack_exports__BlenderbotSmallTokenizer as BlenderbotSmallTokenizer, __webpack_exports__BlenderbotTokenizer as BlenderbotTokenizer, __webpack_exports__BloomForCausalLM as BloomForCausalLM, __webpack_exports__BloomModel as BloomModel, __webpack_exports__BloomPreTrainedModel as BloomPreTrainedModel, __webpack_exports__BloomTokenizer as BloomTokenizer, __webpack_exports__CLIPFeatureExtractor as CLIPFeatureExtractor, __webpack_exports__CLIPImageProcessor as CLIPImageProcessor, __webpack_exports__CLIPModel as CLIPModel, __webpack_exports__CLIPPreTrainedModel as CLIPPreTrainedModel, __webpack_exports__CLIPSegForImageSegmentation as CLIPSegForImageSegmentation, __webpack_exports__CLIPSegModel as CLIPSegModel, __webpack_exports__CLIPSegPreTrainedModel as CLIPSegPreTrainedModel, __webpack_exports__CLIPTextModel as CLIPTextModel, __webpack_exports__CLIPTextModelWithProjection as CLIPTextModelWithProjection, __webpack_exports__CLIPTokenizer as CLIPTokenizer, __webpack_exports__CLIPVisionModel as CLIPVisionModel, __webpack_exports__CLIPVisionModelWithProjection as CLIPVisionModelWithProjection, __webpack_exports__CamembertForMaskedLM as CamembertForMaskedLM, __webpack_exports__CamembertForQuestionAnswering as CamembertForQuestionAnswering, __webpack_exports__CamembertForSequenceClassification as CamembertForSequenceClassification, __webpack_exports__CamembertForTokenClassification as CamembertForTokenClassification, __webpack_exports__CamembertModel as CamembertModel, __webpack_exports__CamembertPreTrainedModel as CamembertPreTrainedModel, __webpack_exports__CamembertTokenizer as CamembertTokenizer, __webpack_exports__CausalLMOutput as CausalLMOutput, __webpack_exports__CausalLMOutputWithPast as CausalLMOutputWithPast, __webpack_exports__ChineseCLIPFeatureExtractor as ChineseCLIPFeatureExtractor, __webpack_exports__ChineseCLIPModel as ChineseCLIPModel, __webpack_exports__ChineseCLIPPreTrainedModel as ChineseCLIPPreTrainedModel, __webpack_exports__ClapAudioModelWithProjection as ClapAudioModelWithProjection, __webpack_exports__ClapFeatureExtractor as ClapFeatureExtractor, __webpack_exports__ClapModel as ClapModel, __webpack_exports__ClapPreTrainedModel as ClapPreTrainedModel, __webpack_exports__ClapTextModelWithProjection as ClapTextModelWithProjection, __webpack_exports__ClassifierFreeGuidanceLogitsProcessor as ClassifierFreeGuidanceLogitsProcessor, __webpack_exports__CodeGenForCausalLM as CodeGenForCausalLM, __webpack_exports__CodeGenModel as CodeGenModel, __webpack_exports__CodeGenPreTrainedModel as CodeGenPreTrainedModel, __webpack_exports__CodeGenTokenizer as CodeGenTokenizer, __webpack_exports__CodeLlamaTokenizer as CodeLlamaTokenizer, __webpack_exports__CohereForCausalLM as CohereForCausalLM, __webpack_exports__CohereModel as CohereModel, __webpack_exports__CoherePreTrainedModel as CoherePreTrainedModel, __webpack_exports__CohereTokenizer as CohereTokenizer, __webpack_exports__ConvBertForMaskedLM as ConvBertForMaskedLM, __webpack_exports__ConvBertForQuestionAnswering as ConvBertForQuestionAnswering, __webpack_exports__ConvBertForSequenceClassification as ConvBertForSequenceClassification, __webpack_exports__ConvBertForTokenClassification as ConvBertForTokenClassification, __webpack_exports__ConvBertModel as ConvBertModel, __webpack_exports__ConvBertPreTrainedModel as ConvBertPreTrainedModel, __webpack_exports__ConvBertTokenizer as ConvBertTokenizer, __webpack_exports__ConvNextFeatureExtractor as ConvNextFeatureExtractor, __webpack_exports__ConvNextForImageClassification as ConvNextForImageClassification, __webpack_exports__ConvNextImageProcessor as ConvNextImageProcessor, __webpack_exports__ConvNextModel as ConvNextModel, __webpack_exports__ConvNextPreTrainedModel as ConvNextPreTrainedModel, __webpack_exports__ConvNextV2ForImageClassification as ConvNextV2ForImageClassification, __webpack_exports__ConvNextV2Model as ConvNextV2Model, __webpack_exports__ConvNextV2PreTrainedModel as ConvNextV2PreTrainedModel, __webpack_exports__DPTFeatureExtractor as DPTFeatureExtractor, __webpack_exports__DPTForDepthEstimation as DPTForDepthEstimation, __webpack_exports__DPTImageProcessor as DPTImageProcessor, __webpack_exports__DPTModel as DPTModel, __webpack_exports__DPTPreTrainedModel as DPTPreTrainedModel, __webpack_exports__DebertaForMaskedLM as DebertaForMaskedLM, __webpack_exports__DebertaForQuestionAnswering as DebertaForQuestionAnswering, __webpack_exports__DebertaForSequenceClassification as DebertaForSequenceClassification, __webpack_exports__DebertaForTokenClassification as DebertaForTokenClassification, __webpack_exports__DebertaModel as DebertaModel, __webpack_exports__DebertaPreTrainedModel as DebertaPreTrainedModel, __webpack_exports__DebertaTokenizer as DebertaTokenizer, __webpack_exports__DebertaV2ForMaskedLM as DebertaV2ForMaskedLM, __webpack_exports__DebertaV2ForQuestionAnswering as DebertaV2ForQuestionAnswering, __webpack_exports__DebertaV2ForSequenceClassification as DebertaV2ForSequenceClassification, __webpack_exports__DebertaV2ForTokenClassification as DebertaV2ForTokenClassification, __webpack_exports__DebertaV2Model as DebertaV2Model, __webpack_exports__DebertaV2PreTrainedModel as DebertaV2PreTrainedModel, __webpack_exports__DebertaV2Tokenizer as DebertaV2Tokenizer, __webpack_exports__DecisionTransformerModel as DecisionTransformerModel, __webpack_exports__DecisionTransformerPreTrainedModel as DecisionTransformerPreTrainedModel, __webpack_exports__DeiTFeatureExtractor as DeiTFeatureExtractor, __webpack_exports__DeiTForImageClassification as DeiTForImageClassification, __webpack_exports__DeiTImageProcessor as DeiTImageProcessor, __webpack_exports__DeiTModel as DeiTModel, __webpack_exports__DeiTPreTrainedModel as DeiTPreTrainedModel, __webpack_exports__DepthAnythingForDepthEstimation as DepthAnythingForDepthEstimation, __webpack_exports__DepthAnythingPreTrainedModel as DepthAnythingPreTrainedModel, __webpack_exports__DepthEstimationPipeline as DepthEstimationPipeline, __webpack_exports__DepthProForDepthEstimation as DepthProForDepthEstimation, __webpack_exports__DepthProPreTrainedModel as DepthProPreTrainedModel, __webpack_exports__DetrFeatureExtractor as DetrFeatureExtractor, __webpack_exports__DetrForObjectDetection as DetrForObjectDetection, __webpack_exports__DetrForSegmentation as DetrForSegmentation, __webpack_exports__DetrImageProcessor as DetrImageProcessor, __webpack_exports__DetrModel as DetrModel, __webpack_exports__DetrObjectDetectionOutput as DetrObjectDetectionOutput, __webpack_exports__DetrPreTrainedModel as DetrPreTrainedModel, __webpack_exports__DetrSegmentationOutput as DetrSegmentationOutput, __webpack_exports__Dinov2ForImageClassification as Dinov2ForImageClassification, __webpack_exports__Dinov2Model as Dinov2Model, __webpack_exports__Dinov2PreTrainedModel as Dinov2PreTrainedModel, __webpack_exports__DistilBertForMaskedLM as DistilBertForMaskedLM, __webpack_exports__DistilBertForQuestionAnswering as DistilBertForQuestionAnswering, __webpack_exports__DistilBertForSequenceClassification as DistilBertForSequenceClassification, __webpack_exports__DistilBertForTokenClassification as DistilBertForTokenClassification, __webpack_exports__DistilBertModel as DistilBertModel, __webpack_exports__DistilBertPreTrainedModel as DistilBertPreTrainedModel, __webpack_exports__DistilBertTokenizer as DistilBertTokenizer, __webpack_exports__DocumentQuestionAnsweringPipeline as DocumentQuestionAnsweringPipeline, __webpack_exports__DonutFeatureExtractor as DonutFeatureExtractor, __webpack_exports__DonutImageProcessor as DonutImageProcessor, __webpack_exports__DonutSwinModel as DonutSwinModel, __webpack_exports__DonutSwinPreTrainedModel as DonutSwinPreTrainedModel, __webpack_exports__EfficientNetForImageClassification as EfficientNetForImageClassification, __webpack_exports__EfficientNetImageProcessor as EfficientNetImageProcessor, __webpack_exports__EfficientNetModel as EfficientNetModel, __webpack_exports__EfficientNetPreTrainedModel as EfficientNetPreTrainedModel, __webpack_exports__ElectraForMaskedLM as ElectraForMaskedLM, __webpack_exports__ElectraForQuestionAnswering as ElectraForQuestionAnswering, __webpack_exports__ElectraForSequenceClassification as ElectraForSequenceClassification, __webpack_exports__ElectraForTokenClassification as ElectraForTokenClassification, __webpack_exports__ElectraModel as ElectraModel, __webpack_exports__ElectraPreTrainedModel as ElectraPreTrainedModel, __webpack_exports__ElectraTokenizer as ElectraTokenizer, __webpack_exports__EosTokenCriteria as EosTokenCriteria, __webpack_exports__EsmForMaskedLM as EsmForMaskedLM, __webpack_exports__EsmForSequenceClassification as EsmForSequenceClassification, __webpack_exports__EsmForTokenClassification as EsmForTokenClassification, __webpack_exports__EsmModel as EsmModel, __webpack_exports__EsmPreTrainedModel as EsmPreTrainedModel, __webpack_exports__EsmTokenizer as EsmTokenizer, __webpack_exports__ExaoneForCausalLM as ExaoneForCausalLM, __webpack_exports__ExaoneModel as ExaoneModel, __webpack_exports__ExaonePreTrainedModel as ExaonePreTrainedModel, __webpack_exports__FFT as FFT, __webpack_exports__FalconForCausalLM as FalconForCausalLM, __webpack_exports__FalconModel as FalconModel, __webpack_exports__FalconPreTrainedModel as FalconPreTrainedModel, __webpack_exports__FalconTokenizer as FalconTokenizer, __webpack_exports__FastViTForImageClassification as FastViTForImageClassification, __webpack_exports__FastViTModel as FastViTModel, __webpack_exports__FastViTPreTrainedModel as FastViTPreTrainedModel, __webpack_exports__FeatureExtractionPipeline as FeatureExtractionPipeline, __webpack_exports__FeatureExtractor as FeatureExtractor, __webpack_exports__FillMaskPipeline as FillMaskPipeline, __webpack_exports__Florence2ForConditionalGeneration as Florence2ForConditionalGeneration, __webpack_exports__Florence2PreTrainedModel as Florence2PreTrainedModel, __webpack_exports__Florence2Processor as Florence2Processor, __webpack_exports__ForcedBOSTokenLogitsProcessor as ForcedBOSTokenLogitsProcessor, __webpack_exports__ForcedEOSTokenLogitsProcessor as ForcedEOSTokenLogitsProcessor, __webpack_exports__GLPNFeatureExtractor as GLPNFeatureExtractor, __webpack_exports__GLPNForDepthEstimation as GLPNForDepthEstimation, __webpack_exports__GLPNModel as GLPNModel, __webpack_exports__GLPNPreTrainedModel as GLPNPreTrainedModel, __webpack_exports__GPT2LMHeadModel as GPT2LMHeadModel, __webpack_exports__GPT2Model as GPT2Model, __webpack_exports__GPT2PreTrainedModel as GPT2PreTrainedModel, __webpack_exports__GPT2Tokenizer as GPT2Tokenizer, __webpack_exports__GPTBigCodeForCausalLM as GPTBigCodeForCausalLM, __webpack_exports__GPTBigCodeModel as GPTBigCodeModel, __webpack_exports__GPTBigCodePreTrainedModel as GPTBigCodePreTrainedModel, __webpack_exports__GPTJForCausalLM as GPTJForCausalLM, __webpack_exports__GPTJModel as GPTJModel, __webpack_exports__GPTJPreTrainedModel as GPTJPreTrainedModel, __webpack_exports__GPTNeoForCausalLM as GPTNeoForCausalLM, __webpack_exports__GPTNeoModel as GPTNeoModel, __webpack_exports__GPTNeoPreTrainedModel as GPTNeoPreTrainedModel, __webpack_exports__GPTNeoXForCausalLM as GPTNeoXForCausalLM, __webpack_exports__GPTNeoXModel as GPTNeoXModel, __webpack_exports__GPTNeoXPreTrainedModel as GPTNeoXPreTrainedModel, __webpack_exports__GPTNeoXTokenizer as GPTNeoXTokenizer, __webpack_exports__Gemma2ForCausalLM as Gemma2ForCausalLM, __webpack_exports__Gemma2Model as Gemma2Model, __webpack_exports__Gemma2PreTrainedModel as Gemma2PreTrainedModel, __webpack_exports__GemmaForCausalLM as GemmaForCausalLM, __webpack_exports__GemmaModel as GemmaModel, __webpack_exports__GemmaPreTrainedModel as GemmaPreTrainedModel, __webpack_exports__GemmaTokenizer as GemmaTokenizer, __webpack_exports__GraniteForCausalLM as GraniteForCausalLM, __webpack_exports__GraniteModel as GraniteModel, __webpack_exports__GranitePreTrainedModel as GranitePreTrainedModel, __webpack_exports__Grok1Tokenizer as Grok1Tokenizer, __webpack_exports__GroupViTModel as GroupViTModel, __webpack_exports__GroupViTPreTrainedModel as GroupViTPreTrainedModel, __webpack_exports__HerbertTokenizer as HerbertTokenizer, __webpack_exports__HieraForImageClassification as HieraForImageClassification, __webpack_exports__HieraModel as HieraModel, __webpack_exports__HieraPreTrainedModel as HieraPreTrainedModel, __webpack_exports__HubertForCTC as HubertForCTC, __webpack_exports__HubertForSequenceClassification as HubertForSequenceClassification, __webpack_exports__HubertModel as HubertModel, __webpack_exports__HubertPreTrainedModel as HubertPreTrainedModel, __webpack_exports__IJepaForImageClassification as IJepaForImageClassification, __webpack_exports__IJepaModel as IJepaModel, __webpack_exports__IJepaPreTrainedModel as IJepaPreTrainedModel, __webpack_exports__Idefics3ForConditionalGeneration as Idefics3ForConditionalGeneration, __webpack_exports__Idefics3ImageProcessor as Idefics3ImageProcessor, __webpack_exports__Idefics3PreTrainedModel as Idefics3PreTrainedModel, __webpack_exports__Idefics3Processor as Idefics3Processor, __webpack_exports__ImageClassificationPipeline as ImageClassificationPipeline, __webpack_exports__ImageFeatureExtractionPipeline as ImageFeatureExtractionPipeline, __webpack_exports__ImageFeatureExtractor as ImageFeatureExtractor, __webpack_exports__ImageMattingOutput as ImageMattingOutput, __webpack_exports__ImageProcessor as ImageProcessor, __webpack_exports__ImageSegmentationPipeline as ImageSegmentationPipeline, __webpack_exports__ImageToImagePipeline as ImageToImagePipeline, __webpack_exports__ImageToTextPipeline as ImageToTextPipeline, __webpack_exports__InterruptableStoppingCriteria as InterruptableStoppingCriteria, __webpack_exports__JAISLMHeadModel as JAISLMHeadModel, __webpack_exports__JAISModel as JAISModel, __webpack_exports__JAISPreTrainedModel as JAISPreTrainedModel, __webpack_exports__JinaCLIPImageProcessor as JinaCLIPImageProcessor, __webpack_exports__JinaCLIPModel as JinaCLIPModel, __webpack_exports__JinaCLIPPreTrainedModel as JinaCLIPPreTrainedModel, __webpack_exports__JinaCLIPProcessor as JinaCLIPProcessor, __webpack_exports__JinaCLIPTextModel as JinaCLIPTextModel, __webpack_exports__JinaCLIPVisionModel as JinaCLIPVisionModel, __webpack_exports__LlamaForCausalLM as LlamaForCausalLM, __webpack_exports__LlamaModel as LlamaModel, __webpack_exports__LlamaPreTrainedModel as LlamaPreTrainedModel, __webpack_exports__LlamaTokenizer as LlamaTokenizer, __webpack_exports__LlavaForConditionalGeneration as LlavaForConditionalGeneration, __webpack_exports__LlavaOnevisionForConditionalGeneration as LlavaOnevisionForConditionalGeneration, __webpack_exports__LlavaOnevisionImageProcessor as LlavaOnevisionImageProcessor, __webpack_exports__LlavaPreTrainedModel as LlavaPreTrainedModel, __webpack_exports__LogitsProcessor as LogitsProcessor, __webpack_exports__LogitsProcessorList as LogitsProcessorList, __webpack_exports__LogitsWarper as LogitsWarper, __webpack_exports__LongT5ForConditionalGeneration as LongT5ForConditionalGeneration, __webpack_exports__LongT5Model as LongT5Model, __webpack_exports__LongT5PreTrainedModel as LongT5PreTrainedModel, __webpack_exports__M2M100ForConditionalGeneration as M2M100ForConditionalGeneration, __webpack_exports__M2M100Model as M2M100Model, __webpack_exports__M2M100PreTrainedModel as M2M100PreTrainedModel, __webpack_exports__M2M100Tokenizer as M2M100Tokenizer, __webpack_exports__MBart50Tokenizer as MBart50Tokenizer, __webpack_exports__MBartForCausalLM as MBartForCausalLM, __webpack_exports__MBartForConditionalGeneration as MBartForConditionalGeneration, __webpack_exports__MBartForSequenceClassification as MBartForSequenceClassification, __webpack_exports__MBartModel as MBartModel, __webpack_exports__MBartPreTrainedModel as MBartPreTrainedModel, __webpack_exports__MBartTokenizer as MBartTokenizer, __webpack_exports__MPNetForMaskedLM as MPNetForMaskedLM, __webpack_exports__MPNetForQuestionAnswering as MPNetForQuestionAnswering, __webpack_exports__MPNetForSequenceClassification as MPNetForSequenceClassification, __webpack_exports__MPNetForTokenClassification as MPNetForTokenClassification, __webpack_exports__MPNetModel as MPNetModel, __webpack_exports__MPNetPreTrainedModel as MPNetPreTrainedModel, __webpack_exports__MPNetTokenizer as MPNetTokenizer, __webpack_exports__MT5ForConditionalGeneration as MT5ForConditionalGeneration, __webpack_exports__MT5Model as MT5Model, __webpack_exports__MT5PreTrainedModel as MT5PreTrainedModel, __webpack_exports__MarianMTModel as MarianMTModel, __webpack_exports__MarianModel as MarianModel, __webpack_exports__MarianPreTrainedModel as MarianPreTrainedModel, __webpack_exports__MarianTokenizer as MarianTokenizer, __webpack_exports__Mask2FormerImageProcessor as Mask2FormerImageProcessor, __webpack_exports__MaskFormerFeatureExtractor as MaskFormerFeatureExtractor, __webpack_exports__MaskFormerForInstanceSegmentation as MaskFormerForInstanceSegmentation, __webpack_exports__MaskFormerImageProcessor as MaskFormerImageProcessor, __webpack_exports__MaskFormerModel as MaskFormerModel, __webpack_exports__MaskFormerPreTrainedModel as MaskFormerPreTrainedModel, __webpack_exports__MaskedLMOutput as MaskedLMOutput, __webpack_exports__MaxLengthCriteria as MaxLengthCriteria, __webpack_exports__MgpstrForSceneTextRecognition as MgpstrForSceneTextRecognition, __webpack_exports__MgpstrModelOutput as MgpstrModelOutput, __webpack_exports__MgpstrPreTrainedModel as MgpstrPreTrainedModel, __webpack_exports__MgpstrProcessor as MgpstrProcessor, __webpack_exports__MgpstrTokenizer as MgpstrTokenizer, __webpack_exports__MinLengthLogitsProcessor as MinLengthLogitsProcessor, __webpack_exports__MinNewTokensLengthLogitsProcessor as MinNewTokensLengthLogitsProcessor, __webpack_exports__MistralForCausalLM as MistralForCausalLM, __webpack_exports__MistralModel as MistralModel, __webpack_exports__MistralPreTrainedModel as MistralPreTrainedModel, __webpack_exports__MobileBertForMaskedLM as MobileBertForMaskedLM, __webpack_exports__MobileBertForQuestionAnswering as MobileBertForQuestionAnswering, __webpack_exports__MobileBertForSequenceClassification as MobileBertForSequenceClassification, __webpack_exports__MobileBertModel as MobileBertModel, __webpack_exports__MobileBertPreTrainedModel as MobileBertPreTrainedModel, __webpack_exports__MobileBertTokenizer as MobileBertTokenizer, __webpack_exports__MobileLLMForCausalLM as MobileLLMForCausalLM, __webpack_exports__MobileLLMModel as MobileLLMModel, __webpack_exports__MobileLLMPreTrainedModel as MobileLLMPreTrainedModel, __webpack_exports__MobileNetV1FeatureExtractor as MobileNetV1FeatureExtractor, __webpack_exports__MobileNetV1ForImageClassification as MobileNetV1ForImageClassification, __webpack_exports__MobileNetV1ImageProcessor as MobileNetV1ImageProcessor, __webpack_exports__MobileNetV1Model as MobileNetV1Model, __webpack_exports__MobileNetV1PreTrainedModel as MobileNetV1PreTrainedModel, __webpack_exports__MobileNetV2FeatureExtractor as MobileNetV2FeatureExtractor, __webpack_exports__MobileNetV2ForImageClassification as MobileNetV2ForImageClassification, __webpack_exports__MobileNetV2ImageProcessor as MobileNetV2ImageProcessor, __webpack_exports__MobileNetV2Model as MobileNetV2Model, __webpack_exports__MobileNetV2PreTrainedModel as MobileNetV2PreTrainedModel, __webpack_exports__MobileNetV3FeatureExtractor as MobileNetV3FeatureExtractor, __webpack_exports__MobileNetV3ForImageClassification as MobileNetV3ForImageClassification, __webpack_exports__MobileNetV3ImageProcessor as MobileNetV3ImageProcessor, __webpack_exports__MobileNetV3Model as MobileNetV3Model, __webpack_exports__MobileNetV3PreTrainedModel as MobileNetV3PreTrainedModel, __webpack_exports__MobileNetV4FeatureExtractor as MobileNetV4FeatureExtractor, __webpack_exports__MobileNetV4ForImageClassification as MobileNetV4ForImageClassification, __webpack_exports__MobileNetV4ImageProcessor as MobileNetV4ImageProcessor, __webpack_exports__MobileNetV4Model as MobileNetV4Model, __webpack_exports__MobileNetV4PreTrainedModel as MobileNetV4PreTrainedModel, __webpack_exports__MobileViTFeatureExtractor as MobileViTFeatureExtractor, __webpack_exports__MobileViTForImageClassification as MobileViTForImageClassification, __webpack_exports__MobileViTImageProcessor as MobileViTImageProcessor, __webpack_exports__MobileViTModel as MobileViTModel, __webpack_exports__MobileViTPreTrainedModel as MobileViTPreTrainedModel, __webpack_exports__MobileViTV2ForImageClassification as MobileViTV2ForImageClassification, __webpack_exports__MobileViTV2Model as MobileViTV2Model, __webpack_exports__MobileViTV2PreTrainedModel as MobileViTV2PreTrainedModel, __webpack_exports__ModelOutput as ModelOutput, __webpack_exports__ModernBertForMaskedLM as ModernBertForMaskedLM, __webpack_exports__ModernBertForSequenceClassification as ModernBertForSequenceClassification, __webpack_exports__ModernBertForTokenClassification as ModernBertForTokenClassification, __webpack_exports__ModernBertModel as ModernBertModel, __webpack_exports__ModernBertPreTrainedModel as ModernBertPreTrainedModel, __webpack_exports__Moondream1ForConditionalGeneration as Moondream1ForConditionalGeneration, __webpack_exports__MoonshineFeatureExtractor as MoonshineFeatureExtractor, __webpack_exports__MoonshineForConditionalGeneration as MoonshineForConditionalGeneration, __webpack_exports__MoonshineModel as MoonshineModel, __webpack_exports__MoonshinePreTrainedModel as MoonshinePreTrainedModel, __webpack_exports__MoonshineProcessor as MoonshineProcessor, __webpack_exports__MptForCausalLM as MptForCausalLM, __webpack_exports__MptModel as MptModel, __webpack_exports__MptPreTrainedModel as MptPreTrainedModel, __webpack_exports__MultiModalityCausalLM as MultiModalityCausalLM, __webpack_exports__MultiModalityPreTrainedModel as MultiModalityPreTrainedModel, __webpack_exports__MusicgenForCausalLM as MusicgenForCausalLM, __webpack_exports__MusicgenForConditionalGeneration as MusicgenForConditionalGeneration, __webpack_exports__MusicgenModel as MusicgenModel, __webpack_exports__MusicgenPreTrainedModel as MusicgenPreTrainedModel, __webpack_exports__NllbTokenizer as NllbTokenizer, __webpack_exports__NoBadWordsLogitsProcessor as NoBadWordsLogitsProcessor, __webpack_exports__NoRepeatNGramLogitsProcessor as NoRepeatNGramLogitsProcessor, __webpack_exports__NomicBertModel as NomicBertModel, __webpack_exports__NomicBertPreTrainedModel as NomicBertPreTrainedModel, __webpack_exports__NougatImageProcessor as NougatImageProcessor, __webpack_exports__NougatTokenizer as NougatTokenizer, __webpack_exports__OPTForCausalLM as OPTForCausalLM, __webpack_exports__OPTModel as OPTModel, __webpack_exports__OPTPreTrainedModel as OPTPreTrainedModel, __webpack_exports__ObjectDetectionPipeline as ObjectDetectionPipeline, __webpack_exports__Olmo2ForCausalLM as Olmo2ForCausalLM, __webpack_exports__Olmo2Model as Olmo2Model, __webpack_exports__Olmo2PreTrainedModel as Olmo2PreTrainedModel, __webpack_exports__OlmoForCausalLM as OlmoForCausalLM, __webpack_exports__OlmoModel as OlmoModel, __webpack_exports__OlmoPreTrainedModel as OlmoPreTrainedModel, __webpack_exports__OpenELMForCausalLM as OpenELMForCausalLM, __webpack_exports__OpenELMModel as OpenELMModel, __webpack_exports__OpenELMPreTrainedModel as OpenELMPreTrainedModel, __webpack_exports__OwlViTFeatureExtractor as OwlViTFeatureExtractor, __webpack_exports__OwlViTForObjectDetection as OwlViTForObjectDetection, __webpack_exports__OwlViTImageProcessor as OwlViTImageProcessor, __webpack_exports__OwlViTModel as OwlViTModel, __webpack_exports__OwlViTPreTrainedModel as OwlViTPreTrainedModel, __webpack_exports__OwlViTProcessor as OwlViTProcessor, __webpack_exports__Owlv2ForObjectDetection as Owlv2ForObjectDetection, __webpack_exports__Owlv2ImageProcessor as Owlv2ImageProcessor, __webpack_exports__Owlv2Model as Owlv2Model, __webpack_exports__Owlv2PreTrainedModel as Owlv2PreTrainedModel, __webpack_exports__PaliGemmaForConditionalGeneration as PaliGemmaForConditionalGeneration, __webpack_exports__PaliGemmaPreTrainedModel as PaliGemmaPreTrainedModel, __webpack_exports__PaliGemmaProcessor as PaliGemmaProcessor, __webpack_exports__PatchTSMixerForPrediction as PatchTSMixerForPrediction, __webpack_exports__PatchTSMixerModel as PatchTSMixerModel, __webpack_exports__PatchTSMixerPreTrainedModel as PatchTSMixerPreTrainedModel, __webpack_exports__PatchTSTForPrediction as PatchTSTForPrediction, __webpack_exports__PatchTSTModel as PatchTSTModel, __webpack_exports__PatchTSTPreTrainedModel as PatchTSTPreTrainedModel, __webpack_exports__Phi3ForCausalLM as Phi3ForCausalLM, __webpack_exports__Phi3Model as Phi3Model, __webpack_exports__Phi3PreTrainedModel as Phi3PreTrainedModel, __webpack_exports__Phi3VForCausalLM as Phi3VForCausalLM, __webpack_exports__Phi3VImageProcessor as Phi3VImageProcessor, __webpack_exports__Phi3VPreTrainedModel as Phi3VPreTrainedModel, __webpack_exports__Phi3VProcessor as Phi3VProcessor, __webpack_exports__PhiForCausalLM as PhiForCausalLM, __webpack_exports__PhiModel as PhiModel, __webpack_exports__PhiPreTrainedModel as PhiPreTrainedModel, __webpack_exports__Pipeline as Pipeline, __webpack_exports__PreTrainedModel as PreTrainedModel, __webpack_exports__PreTrainedTokenizer as PreTrainedTokenizer, __webpack_exports__PretrainedConfig as PretrainedConfig, __webpack_exports__PretrainedMixin as PretrainedMixin, __webpack_exports__Processor as Processor, __webpack_exports__PvtForImageClassification as PvtForImageClassification, __webpack_exports__PvtImageProcessor as PvtImageProcessor, __webpack_exports__PvtModel as PvtModel, __webpack_exports__PvtPreTrainedModel as PvtPreTrainedModel, __webpack_exports__PyAnnoteFeatureExtractor as PyAnnoteFeatureExtractor, __webpack_exports__PyAnnoteForAudioFrameClassification as PyAnnoteForAudioFrameClassification, __webpack_exports__PyAnnoteModel as PyAnnoteModel, __webpack_exports__PyAnnotePreTrainedModel as PyAnnotePreTrainedModel, __webpack_exports__PyAnnoteProcessor as PyAnnoteProcessor, __webpack_exports__QuestionAnsweringModelOutput as QuestionAnsweringModelOutput, __webpack_exports__QuestionAnsweringPipeline as QuestionAnsweringPipeline, __webpack_exports__Qwen2ForCausalLM as Qwen2ForCausalLM, __webpack_exports__Qwen2Model as Qwen2Model, __webpack_exports__Qwen2PreTrainedModel as Qwen2PreTrainedModel, __webpack_exports__Qwen2Tokenizer as Qwen2Tokenizer, __webpack_exports__Qwen2VLForConditionalGeneration as Qwen2VLForConditionalGeneration, __webpack_exports__Qwen2VLImageProcessor as Qwen2VLImageProcessor, __webpack_exports__Qwen2VLPreTrainedModel as Qwen2VLPreTrainedModel, __webpack_exports__Qwen2VLProcessor as Qwen2VLProcessor, __webpack_exports__RTDetrForObjectDetection as RTDetrForObjectDetection, __webpack_exports__RTDetrImageProcessor as RTDetrImageProcessor, __webpack_exports__RTDetrModel as RTDetrModel, __webpack_exports__RTDetrObjectDetectionOutput as RTDetrObjectDetectionOutput, __webpack_exports__RTDetrPreTrainedModel as RTDetrPreTrainedModel, __webpack_exports__RawImage as RawImage, __webpack_exports__RepetitionPenaltyLogitsProcessor as RepetitionPenaltyLogitsProcessor, __webpack_exports__ResNetForImageClassification as ResNetForImageClassification, __webpack_exports__ResNetModel as ResNetModel, __webpack_exports__ResNetPreTrainedModel as ResNetPreTrainedModel, __webpack_exports__RoFormerForMaskedLM as RoFormerForMaskedLM, __webpack_exports__RoFormerForQuestionAnswering as RoFormerForQuestionAnswering, __webpack_exports__RoFormerForSequenceClassification as RoFormerForSequenceClassification, __webpack_exports__RoFormerForTokenClassification as RoFormerForTokenClassification, __webpack_exports__RoFormerModel as RoFormerModel, __webpack_exports__RoFormerPreTrainedModel as RoFormerPreTrainedModel, __webpack_exports__RoFormerTokenizer as RoFormerTokenizer, __webpack_exports__RobertaForMaskedLM as RobertaForMaskedLM, __webpack_exports__RobertaForQuestionAnswering as RobertaForQuestionAnswering, __webpack_exports__RobertaForSequenceClassification as RobertaForSequenceClassification, __webpack_exports__RobertaForTokenClassification as RobertaForTokenClassification, __webpack_exports__RobertaModel as RobertaModel, __webpack_exports__RobertaPreTrainedModel as RobertaPreTrainedModel, __webpack_exports__RobertaTokenizer as RobertaTokenizer, __webpack_exports__SamImageProcessor as SamImageProcessor, __webpack_exports__SamImageSegmentationOutput as SamImageSegmentationOutput, __webpack_exports__SamModel as SamModel, __webpack_exports__SamPreTrainedModel as SamPreTrainedModel, __webpack_exports__SamProcessor as SamProcessor, __webpack_exports__SapiensForDepthEstimation as SapiensForDepthEstimation, __webpack_exports__SapiensForNormalEstimation as SapiensForNormalEstimation, __webpack_exports__SapiensForSemanticSegmentation as SapiensForSemanticSegmentation, __webpack_exports__SapiensPreTrainedModel as SapiensPreTrainedModel, __webpack_exports__SeamlessM4TFeatureExtractor as SeamlessM4TFeatureExtractor, __webpack_exports__SegformerFeatureExtractor as SegformerFeatureExtractor, __webpack_exports__SegformerForImageClassification as SegformerForImageClassification, __webpack_exports__SegformerForSemanticSegmentation as SegformerForSemanticSegmentation, __webpack_exports__SegformerImageProcessor as SegformerImageProcessor, __webpack_exports__SegformerModel as SegformerModel, __webpack_exports__SegformerPreTrainedModel as SegformerPreTrainedModel, __webpack_exports__Seq2SeqLMOutput as Seq2SeqLMOutput, __webpack_exports__SequenceClassifierOutput as SequenceClassifierOutput, __webpack_exports__SiglipImageProcessor as SiglipImageProcessor, __webpack_exports__SiglipModel as SiglipModel, __webpack_exports__SiglipPreTrainedModel as SiglipPreTrainedModel, __webpack_exports__SiglipTextModel as SiglipTextModel, __webpack_exports__SiglipTokenizer as SiglipTokenizer, __webpack_exports__SiglipVisionModel as SiglipVisionModel, __webpack_exports__SpeechT5FeatureExtractor as SpeechT5FeatureExtractor, __webpack_exports__SpeechT5ForSpeechToText as SpeechT5ForSpeechToText, __webpack_exports__SpeechT5ForTextToSpeech as SpeechT5ForTextToSpeech, __webpack_exports__SpeechT5HifiGan as SpeechT5HifiGan, __webpack_exports__SpeechT5Model as SpeechT5Model, __webpack_exports__SpeechT5PreTrainedModel as SpeechT5PreTrainedModel, __webpack_exports__SpeechT5Processor as SpeechT5Processor, __webpack_exports__SpeechT5Tokenizer as SpeechT5Tokenizer, __webpack_exports__SqueezeBertForMaskedLM as SqueezeBertForMaskedLM, __webpack_exports__SqueezeBertForQuestionAnswering as SqueezeBertForQuestionAnswering, __webpack_exports__SqueezeBertForSequenceClassification as SqueezeBertForSequenceClassification, __webpack_exports__SqueezeBertModel as SqueezeBertModel, __webpack_exports__SqueezeBertPreTrainedModel as SqueezeBertPreTrainedModel, __webpack_exports__SqueezeBertTokenizer as SqueezeBertTokenizer, __webpack_exports__StableLmForCausalLM as StableLmForCausalLM, __webpack_exports__StableLmModel as StableLmModel, __webpack_exports__StableLmPreTrainedModel as StableLmPreTrainedModel, __webpack_exports__Starcoder2ForCausalLM as Starcoder2ForCausalLM, __webpack_exports__Starcoder2Model as Starcoder2Model, __webpack_exports__Starcoder2PreTrainedModel as Starcoder2PreTrainedModel, __webpack_exports__StoppingCriteria as StoppingCriteria, __webpack_exports__StoppingCriteriaList as StoppingCriteriaList, __webpack_exports__SummarizationPipeline as SummarizationPipeline, __webpack_exports__SuppressTokensAtBeginLogitsProcessor as SuppressTokensAtBeginLogitsProcessor, __webpack_exports__Swin2SRForImageSuperResolution as Swin2SRForImageSuperResolution, __webpack_exports__Swin2SRImageProcessor as Swin2SRImageProcessor, __webpack_exports__Swin2SRModel as Swin2SRModel, __webpack_exports__Swin2SRPreTrainedModel as Swin2SRPreTrainedModel, __webpack_exports__SwinForImageClassification as SwinForImageClassification, __webpack_exports__SwinModel as SwinModel, __webpack_exports__SwinPreTrainedModel as SwinPreTrainedModel, __webpack_exports__T5ForConditionalGeneration as T5ForConditionalGeneration, __webpack_exports__T5Model as T5Model, __webpack_exports__T5PreTrainedModel as T5PreTrainedModel, __webpack_exports__T5Tokenizer as T5Tokenizer, __webpack_exports__TableTransformerForObjectDetection as TableTransformerForObjectDetection, __webpack_exports__TableTransformerModel as TableTransformerModel, __webpack_exports__TableTransformerObjectDetectionOutput as TableTransformerObjectDetectionOutput, __webpack_exports__TableTransformerPreTrainedModel as TableTransformerPreTrainedModel, __webpack_exports__TemperatureLogitsWarper as TemperatureLogitsWarper, __webpack_exports__Tensor as Tensor, __webpack_exports__Text2TextGenerationPipeline as Text2TextGenerationPipeline, __webpack_exports__TextClassificationPipeline as TextClassificationPipeline, __webpack_exports__TextGenerationPipeline as TextGenerationPipeline, __webpack_exports__TextStreamer as TextStreamer, __webpack_exports__TextToAudioPipeline as TextToAudioPipeline, __webpack_exports__TokenClassificationPipeline as TokenClassificationPipeline, __webpack_exports__TokenClassifierOutput as TokenClassifierOutput, __webpack_exports__TokenizerModel as TokenizerModel, __webpack_exports__TopKLogitsWarper as TopKLogitsWarper, __webpack_exports__TopPLogitsWarper as TopPLogitsWarper, __webpack_exports__TrOCRForCausalLM as TrOCRForCausalLM, __webpack_exports__TrOCRPreTrainedModel as TrOCRPreTrainedModel, __webpack_exports__TranslationPipeline as TranslationPipeline, __webpack_exports__UniSpeechForCTC as UniSpeechForCTC, __webpack_exports__UniSpeechForSequenceClassification as UniSpeechForSequenceClassification, __webpack_exports__UniSpeechModel as UniSpeechModel, __webpack_exports__UniSpeechPreTrainedModel as UniSpeechPreTrainedModel, __webpack_exports__UniSpeechSatForAudioFrameClassification as UniSpeechSatForAudioFrameClassification, __webpack_exports__UniSpeechSatForCTC as UniSpeechSatForCTC, __webpack_exports__UniSpeechSatForSequenceClassification as UniSpeechSatForSequenceClassification, __webpack_exports__UniSpeechSatModel as UniSpeechSatModel, __webpack_exports__UniSpeechSatPreTrainedModel as UniSpeechSatPreTrainedModel, __webpack_exports__VLChatProcessor as VLChatProcessor, __webpack_exports__VLMImageProcessor as VLMImageProcessor, __webpack_exports__ViTFeatureExtractor as ViTFeatureExtractor, __webpack_exports__ViTForImageClassification as ViTForImageClassification, __webpack_exports__ViTImageProcessor as ViTImageProcessor, __webpack_exports__ViTMAEModel as ViTMAEModel, __webpack_exports__ViTMAEPreTrainedModel as ViTMAEPreTrainedModel, __webpack_exports__ViTMSNForImageClassification as ViTMSNForImageClassification, __webpack_exports__ViTMSNModel as ViTMSNModel, __webpack_exports__ViTMSNPreTrainedModel as ViTMSNPreTrainedModel, __webpack_exports__ViTModel as ViTModel, __webpack_exports__ViTPreTrainedModel as ViTPreTrainedModel, __webpack_exports__VisionEncoderDecoderModel as VisionEncoderDecoderModel, __webpack_exports__VitMatteForImageMatting as VitMatteForImageMatting, __webpack_exports__VitMatteImageProcessor as VitMatteImageProcessor, __webpack_exports__VitMattePreTrainedModel as VitMattePreTrainedModel, __webpack_exports__VitPoseForPoseEstimation as VitPoseForPoseEstimation, __webpack_exports__VitPoseImageProcessor as VitPoseImageProcessor, __webpack_exports__VitPosePreTrainedModel as VitPosePreTrainedModel, __webpack_exports__VitsModel as VitsModel, __webpack_exports__VitsModelOutput as VitsModelOutput, __webpack_exports__VitsPreTrainedModel as VitsPreTrainedModel, __webpack_exports__VitsTokenizer as VitsTokenizer, __webpack_exports__Wav2Vec2BertForCTC as Wav2Vec2BertForCTC, __webpack_exports__Wav2Vec2BertForSequenceClassification as Wav2Vec2BertForSequenceClassification, __webpack_exports__Wav2Vec2BertModel as Wav2Vec2BertModel, __webpack_exports__Wav2Vec2BertPreTrainedModel as Wav2Vec2BertPreTrainedModel, __webpack_exports__Wav2Vec2CTCTokenizer as Wav2Vec2CTCTokenizer, __webpack_exports__Wav2Vec2FeatureExtractor as Wav2Vec2FeatureExtractor, __webpack_exports__Wav2Vec2ForAudioFrameClassification as Wav2Vec2ForAudioFrameClassification, __webpack_exports__Wav2Vec2ForCTC as Wav2Vec2ForCTC, __webpack_exports__Wav2Vec2ForSequenceClassification as Wav2Vec2ForSequenceClassification, __webpack_exports__Wav2Vec2Model as Wav2Vec2Model, __webpack_exports__Wav2Vec2PreTrainedModel as Wav2Vec2PreTrainedModel, __webpack_exports__Wav2Vec2ProcessorWithLM as Wav2Vec2ProcessorWithLM, __webpack_exports__WavLMForAudioFrameClassification as WavLMForAudioFrameClassification, __webpack_exports__WavLMForCTC as WavLMForCTC, __webpack_exports__WavLMForSequenceClassification as WavLMForSequenceClassification, __webpack_exports__WavLMForXVector as WavLMForXVector, __webpack_exports__WavLMModel as WavLMModel, __webpack_exports__WavLMPreTrainedModel as WavLMPreTrainedModel, __webpack_exports__WeSpeakerFeatureExtractor as WeSpeakerFeatureExtractor, __webpack_exports__WeSpeakerResNetModel as WeSpeakerResNetModel, __webpack_exports__WeSpeakerResNetPreTrainedModel as WeSpeakerResNetPreTrainedModel, __webpack_exports__WhisperFeatureExtractor as WhisperFeatureExtractor, __webpack_exports__WhisperForConditionalGeneration as WhisperForConditionalGeneration, __webpack_exports__WhisperModel as WhisperModel, __webpack_exports__WhisperPreTrainedModel as WhisperPreTrainedModel, __webpack_exports__WhisperProcessor as WhisperProcessor, __webpack_exports__WhisperTextStreamer as WhisperTextStreamer, __webpack_exports__WhisperTimeStampLogitsProcessor as WhisperTimeStampLogitsProcessor, __webpack_exports__WhisperTokenizer as WhisperTokenizer, __webpack_exports__XLMForQuestionAnswering as XLMForQuestionAnswering, __webpack_exports__XLMForSequenceClassification as XLMForSequenceClassification, __webpack_exports__XLMForTokenClassification as XLMForTokenClassification, __webpack_exports__XLMModel as XLMModel, __webpack_exports__XLMPreTrainedModel as XLMPreTrainedModel, __webpack_exports__XLMRobertaForMaskedLM as XLMRobertaForMaskedLM, __webpack_exports__XLMRobertaForQuestionAnswering as XLMRobertaForQuestionAnswering, __webpack_exports__XLMRobertaForSequenceClassification as XLMRobertaForSequenceClassification, __webpack_exports__XLMRobertaForTokenClassification as XLMRobertaForTokenClassification, __webpack_exports__XLMRobertaModel as XLMRobertaModel, __webpack_exports__XLMRobertaPreTrainedModel as XLMRobertaPreTrainedModel, __webpack_exports__XLMRobertaTokenizer as XLMRobertaTokenizer, __webpack_exports__XLMTokenizer as XLMTokenizer, __webpack_exports__XLMWithLMHeadModel as XLMWithLMHeadModel, __webpack_exports__XVectorOutput as XVectorOutput, __webpack_exports__YolosFeatureExtractor as YolosFeatureExtractor, __webpack_exports__YolosForObjectDetection as YolosForObjectDetection, __webpack_exports__YolosImageProcessor as YolosImageProcessor, __webpack_exports__YolosModel as YolosModel, __webpack_exports__YolosObjectDetectionOutput as YolosObjectDetectionOutput, __webpack_exports__YolosPreTrainedModel as YolosPreTrainedModel, __webpack_exports__ZeroShotAudioClassificationPipeline as ZeroShotAudioClassificationPipeline, __webpack_exports__ZeroShotClassificationPipeline as ZeroShotClassificationPipeline, __webpack_exports__ZeroShotImageClassificationPipeline as ZeroShotImageClassificationPipeline, __webpack_exports__ZeroShotObjectDetectionPipeline as ZeroShotObjectDetectionPipeline, __webpack_exports__bankers_round as bankers_round, __webpack_exports__cat as cat, __webpack_exports__cos_sim as cos_sim, __webpack_exports__dot as dot, __webpack_exports__dynamic_time_warping as dynamic_time_warping, __webpack_exports__env as env, __webpack_exports__full as full, __webpack_exports__full_like as full_like, __webpack_exports__getKeyValueShapes as getKeyValueShapes, __webpack_exports__hamming as hamming, __webpack_exports__hanning as hanning, __webpack_exports__interpolate as interpolate, __webpack_exports__interpolate_4d as interpolate_4d, __webpack_exports__interpolate_data as interpolate_data, __webpack_exports__is_chinese_char as is_chinese_char, __webpack_exports__layer_norm as layer_norm, __webpack_exports__load_image as load_image, __webpack_exports__log_softmax as log_softmax, __webpack_exports__magnitude as magnitude, __webpack_exports__matmul as matmul, __webpack_exports__max as max, __webpack_exports__mean as mean, __webpack_exports__mean_pooling as mean_pooling, __webpack_exports__medianFilter as medianFilter, __webpack_exports__mel_filter_bank as mel_filter_bank, __webpack_exports__min as min, __webpack_exports__ones as ones, __webpack_exports__ones_like as ones_like, __webpack_exports__permute as permute, __webpack_exports__permute_data as permute_data, __webpack_exports__pipeline as pipeline, __webpack_exports__quantize_embeddings as quantize_embeddings, __webpack_exports__rand as rand, __webpack_exports__read_audio as read_audio, __webpack_exports__rfft as rfft, __webpack_exports__round as round, __webpack_exports__slice as slice, __webpack_exports__softmax as softmax, __webpack_exports__spectrogram as spectrogram, __webpack_exports__stack as stack, __webpack_exports__std_mean as std_mean, __webpack_exports__topk as topk, __webpack_exports__window_function as window_function, __webpack_exports__zeros as zeros, __webpack_exports__zeros_like as zeros_like };
|
|
38957
|
+
export { __webpack_exports__ASTFeatureExtractor as ASTFeatureExtractor, __webpack_exports__ASTForAudioClassification as ASTForAudioClassification, __webpack_exports__ASTModel as ASTModel, __webpack_exports__ASTPreTrainedModel as ASTPreTrainedModel, __webpack_exports__AlbertForMaskedLM as AlbertForMaskedLM, __webpack_exports__AlbertForQuestionAnswering as AlbertForQuestionAnswering, __webpack_exports__AlbertForSequenceClassification as AlbertForSequenceClassification, __webpack_exports__AlbertModel as AlbertModel, __webpack_exports__AlbertPreTrainedModel as AlbertPreTrainedModel, __webpack_exports__AlbertTokenizer as AlbertTokenizer, __webpack_exports__AudioClassificationPipeline as AudioClassificationPipeline, __webpack_exports__AutoConfig as AutoConfig, __webpack_exports__AutoFeatureExtractor as AutoFeatureExtractor, __webpack_exports__AutoImageProcessor as AutoImageProcessor, __webpack_exports__AutoModel as AutoModel, __webpack_exports__AutoModelForAudioClassification as AutoModelForAudioClassification, __webpack_exports__AutoModelForAudioFrameClassification as AutoModelForAudioFrameClassification, __webpack_exports__AutoModelForCTC as AutoModelForCTC, __webpack_exports__AutoModelForCausalLM as AutoModelForCausalLM, __webpack_exports__AutoModelForDepthEstimation as AutoModelForDepthEstimation, __webpack_exports__AutoModelForDocumentQuestionAnswering as AutoModelForDocumentQuestionAnswering, __webpack_exports__AutoModelForImageClassification as AutoModelForImageClassification, __webpack_exports__AutoModelForImageFeatureExtraction as AutoModelForImageFeatureExtraction, __webpack_exports__AutoModelForImageMatting as AutoModelForImageMatting, __webpack_exports__AutoModelForImageSegmentation as AutoModelForImageSegmentation, __webpack_exports__AutoModelForImageToImage as AutoModelForImageToImage, __webpack_exports__AutoModelForMaskGeneration as AutoModelForMaskGeneration, __webpack_exports__AutoModelForMaskedLM as AutoModelForMaskedLM, __webpack_exports__AutoModelForNormalEstimation as AutoModelForNormalEstimation, __webpack_exports__AutoModelForObjectDetection as AutoModelForObjectDetection, __webpack_exports__AutoModelForPoseEstimation as AutoModelForPoseEstimation, __webpack_exports__AutoModelForQuestionAnswering as AutoModelForQuestionAnswering, __webpack_exports__AutoModelForSemanticSegmentation as AutoModelForSemanticSegmentation, __webpack_exports__AutoModelForSeq2SeqLM as AutoModelForSeq2SeqLM, __webpack_exports__AutoModelForSequenceClassification as AutoModelForSequenceClassification, __webpack_exports__AutoModelForSpeechSeq2Seq as AutoModelForSpeechSeq2Seq, __webpack_exports__AutoModelForTextToSpectrogram as AutoModelForTextToSpectrogram, __webpack_exports__AutoModelForTextToWaveform as AutoModelForTextToWaveform, __webpack_exports__AutoModelForTokenClassification as AutoModelForTokenClassification, __webpack_exports__AutoModelForUniversalSegmentation as AutoModelForUniversalSegmentation, __webpack_exports__AutoModelForVision2Seq as AutoModelForVision2Seq, __webpack_exports__AutoModelForXVector as AutoModelForXVector, __webpack_exports__AutoModelForZeroShotObjectDetection as AutoModelForZeroShotObjectDetection, __webpack_exports__AutoProcessor as AutoProcessor, __webpack_exports__AutoTokenizer as AutoTokenizer, __webpack_exports__AutomaticSpeechRecognitionPipeline as AutomaticSpeechRecognitionPipeline, __webpack_exports__BartForConditionalGeneration as BartForConditionalGeneration, __webpack_exports__BartForSequenceClassification as BartForSequenceClassification, __webpack_exports__BartModel as BartModel, __webpack_exports__BartPretrainedModel as BartPretrainedModel, __webpack_exports__BartTokenizer as BartTokenizer, __webpack_exports__BaseModelOutput as BaseModelOutput, __webpack_exports__BaseStreamer as BaseStreamer, __webpack_exports__BeitFeatureExtractor as BeitFeatureExtractor, __webpack_exports__BeitForImageClassification as BeitForImageClassification, __webpack_exports__BeitModel as BeitModel, __webpack_exports__BeitPreTrainedModel as BeitPreTrainedModel, __webpack_exports__BertForMaskedLM as BertForMaskedLM, __webpack_exports__BertForQuestionAnswering as BertForQuestionAnswering, __webpack_exports__BertForSequenceClassification as BertForSequenceClassification, __webpack_exports__BertForTokenClassification as BertForTokenClassification, __webpack_exports__BertModel as BertModel, __webpack_exports__BertPreTrainedModel as BertPreTrainedModel, __webpack_exports__BertTokenizer as BertTokenizer, __webpack_exports__BitImageProcessor as BitImageProcessor, __webpack_exports__BlenderbotForConditionalGeneration as BlenderbotForConditionalGeneration, __webpack_exports__BlenderbotModel as BlenderbotModel, __webpack_exports__BlenderbotPreTrainedModel as BlenderbotPreTrainedModel, __webpack_exports__BlenderbotSmallForConditionalGeneration as BlenderbotSmallForConditionalGeneration, __webpack_exports__BlenderbotSmallModel as BlenderbotSmallModel, __webpack_exports__BlenderbotSmallPreTrainedModel as BlenderbotSmallPreTrainedModel, __webpack_exports__BlenderbotSmallTokenizer as BlenderbotSmallTokenizer, __webpack_exports__BlenderbotTokenizer as BlenderbotTokenizer, __webpack_exports__BloomForCausalLM as BloomForCausalLM, __webpack_exports__BloomModel as BloomModel, __webpack_exports__BloomPreTrainedModel as BloomPreTrainedModel, __webpack_exports__BloomTokenizer as BloomTokenizer, __webpack_exports__CLIPFeatureExtractor as CLIPFeatureExtractor, __webpack_exports__CLIPImageProcessor as CLIPImageProcessor, __webpack_exports__CLIPModel as CLIPModel, __webpack_exports__CLIPPreTrainedModel as CLIPPreTrainedModel, __webpack_exports__CLIPSegForImageSegmentation as CLIPSegForImageSegmentation, __webpack_exports__CLIPSegModel as CLIPSegModel, __webpack_exports__CLIPSegPreTrainedModel as CLIPSegPreTrainedModel, __webpack_exports__CLIPTextModel as CLIPTextModel, __webpack_exports__CLIPTextModelWithProjection as CLIPTextModelWithProjection, __webpack_exports__CLIPTokenizer as CLIPTokenizer, __webpack_exports__CLIPVisionModel as CLIPVisionModel, __webpack_exports__CLIPVisionModelWithProjection as CLIPVisionModelWithProjection, __webpack_exports__CamembertForMaskedLM as CamembertForMaskedLM, __webpack_exports__CamembertForQuestionAnswering as CamembertForQuestionAnswering, __webpack_exports__CamembertForSequenceClassification as CamembertForSequenceClassification, __webpack_exports__CamembertForTokenClassification as CamembertForTokenClassification, __webpack_exports__CamembertModel as CamembertModel, __webpack_exports__CamembertPreTrainedModel as CamembertPreTrainedModel, __webpack_exports__CamembertTokenizer as CamembertTokenizer, __webpack_exports__CausalLMOutput as CausalLMOutput, __webpack_exports__CausalLMOutputWithPast as CausalLMOutputWithPast, __webpack_exports__ChineseCLIPFeatureExtractor as ChineseCLIPFeatureExtractor, __webpack_exports__ChineseCLIPModel as ChineseCLIPModel, __webpack_exports__ChineseCLIPPreTrainedModel as ChineseCLIPPreTrainedModel, __webpack_exports__ClapAudioModelWithProjection as ClapAudioModelWithProjection, __webpack_exports__ClapFeatureExtractor as ClapFeatureExtractor, __webpack_exports__ClapModel as ClapModel, __webpack_exports__ClapPreTrainedModel as ClapPreTrainedModel, __webpack_exports__ClapTextModelWithProjection as ClapTextModelWithProjection, __webpack_exports__ClassifierFreeGuidanceLogitsProcessor as ClassifierFreeGuidanceLogitsProcessor, __webpack_exports__CodeGenForCausalLM as CodeGenForCausalLM, __webpack_exports__CodeGenModel as CodeGenModel, __webpack_exports__CodeGenPreTrainedModel as CodeGenPreTrainedModel, __webpack_exports__CodeGenTokenizer as CodeGenTokenizer, __webpack_exports__CodeLlamaTokenizer as CodeLlamaTokenizer, __webpack_exports__CohereForCausalLM as CohereForCausalLM, __webpack_exports__CohereModel as CohereModel, __webpack_exports__CoherePreTrainedModel as CoherePreTrainedModel, __webpack_exports__CohereTokenizer as CohereTokenizer, __webpack_exports__ConvBertForMaskedLM as ConvBertForMaskedLM, __webpack_exports__ConvBertForQuestionAnswering as ConvBertForQuestionAnswering, __webpack_exports__ConvBertForSequenceClassification as ConvBertForSequenceClassification, __webpack_exports__ConvBertForTokenClassification as ConvBertForTokenClassification, __webpack_exports__ConvBertModel as ConvBertModel, __webpack_exports__ConvBertPreTrainedModel as ConvBertPreTrainedModel, __webpack_exports__ConvBertTokenizer as ConvBertTokenizer, __webpack_exports__ConvNextFeatureExtractor as ConvNextFeatureExtractor, __webpack_exports__ConvNextForImageClassification as ConvNextForImageClassification, __webpack_exports__ConvNextImageProcessor as ConvNextImageProcessor, __webpack_exports__ConvNextModel as ConvNextModel, __webpack_exports__ConvNextPreTrainedModel as ConvNextPreTrainedModel, __webpack_exports__ConvNextV2ForImageClassification as ConvNextV2ForImageClassification, __webpack_exports__ConvNextV2Model as ConvNextV2Model, __webpack_exports__ConvNextV2PreTrainedModel as ConvNextV2PreTrainedModel, __webpack_exports__DPTFeatureExtractor as DPTFeatureExtractor, __webpack_exports__DPTForDepthEstimation as DPTForDepthEstimation, __webpack_exports__DPTImageProcessor as DPTImageProcessor, __webpack_exports__DPTModel as DPTModel, __webpack_exports__DPTPreTrainedModel as DPTPreTrainedModel, __webpack_exports__DebertaForMaskedLM as DebertaForMaskedLM, __webpack_exports__DebertaForQuestionAnswering as DebertaForQuestionAnswering, __webpack_exports__DebertaForSequenceClassification as DebertaForSequenceClassification, __webpack_exports__DebertaForTokenClassification as DebertaForTokenClassification, __webpack_exports__DebertaModel as DebertaModel, __webpack_exports__DebertaPreTrainedModel as DebertaPreTrainedModel, __webpack_exports__DebertaTokenizer as DebertaTokenizer, __webpack_exports__DebertaV2ForMaskedLM as DebertaV2ForMaskedLM, __webpack_exports__DebertaV2ForQuestionAnswering as DebertaV2ForQuestionAnswering, __webpack_exports__DebertaV2ForSequenceClassification as DebertaV2ForSequenceClassification, __webpack_exports__DebertaV2ForTokenClassification as DebertaV2ForTokenClassification, __webpack_exports__DebertaV2Model as DebertaV2Model, __webpack_exports__DebertaV2PreTrainedModel as DebertaV2PreTrainedModel, __webpack_exports__DebertaV2Tokenizer as DebertaV2Tokenizer, __webpack_exports__DecisionTransformerModel as DecisionTransformerModel, __webpack_exports__DecisionTransformerPreTrainedModel as DecisionTransformerPreTrainedModel, __webpack_exports__DeiTFeatureExtractor as DeiTFeatureExtractor, __webpack_exports__DeiTForImageClassification as DeiTForImageClassification, __webpack_exports__DeiTImageProcessor as DeiTImageProcessor, __webpack_exports__DeiTModel as DeiTModel, __webpack_exports__DeiTPreTrainedModel as DeiTPreTrainedModel, __webpack_exports__DepthAnythingForDepthEstimation as DepthAnythingForDepthEstimation, __webpack_exports__DepthAnythingPreTrainedModel as DepthAnythingPreTrainedModel, __webpack_exports__DepthEstimationPipeline as DepthEstimationPipeline, __webpack_exports__DepthProForDepthEstimation as DepthProForDepthEstimation, __webpack_exports__DepthProPreTrainedModel as DepthProPreTrainedModel, __webpack_exports__DetrFeatureExtractor as DetrFeatureExtractor, __webpack_exports__DetrForObjectDetection as DetrForObjectDetection, __webpack_exports__DetrForSegmentation as DetrForSegmentation, __webpack_exports__DetrImageProcessor as DetrImageProcessor, __webpack_exports__DetrModel as DetrModel, __webpack_exports__DetrObjectDetectionOutput as DetrObjectDetectionOutput, __webpack_exports__DetrPreTrainedModel as DetrPreTrainedModel, __webpack_exports__DetrSegmentationOutput as DetrSegmentationOutput, __webpack_exports__Dinov2ForImageClassification as Dinov2ForImageClassification, __webpack_exports__Dinov2Model as Dinov2Model, __webpack_exports__Dinov2PreTrainedModel as Dinov2PreTrainedModel, __webpack_exports__Dinov2WithRegistersForImageClassification as Dinov2WithRegistersForImageClassification, __webpack_exports__Dinov2WithRegistersModel as Dinov2WithRegistersModel, __webpack_exports__Dinov2WithRegistersPreTrainedModel as Dinov2WithRegistersPreTrainedModel, __webpack_exports__DistilBertForMaskedLM as DistilBertForMaskedLM, __webpack_exports__DistilBertForQuestionAnswering as DistilBertForQuestionAnswering, __webpack_exports__DistilBertForSequenceClassification as DistilBertForSequenceClassification, __webpack_exports__DistilBertForTokenClassification as DistilBertForTokenClassification, __webpack_exports__DistilBertModel as DistilBertModel, __webpack_exports__DistilBertPreTrainedModel as DistilBertPreTrainedModel, __webpack_exports__DistilBertTokenizer as DistilBertTokenizer, __webpack_exports__DocumentQuestionAnsweringPipeline as DocumentQuestionAnsweringPipeline, __webpack_exports__DonutFeatureExtractor as DonutFeatureExtractor, __webpack_exports__DonutImageProcessor as DonutImageProcessor, __webpack_exports__DonutSwinModel as DonutSwinModel, __webpack_exports__DonutSwinPreTrainedModel as DonutSwinPreTrainedModel, __webpack_exports__EfficientNetForImageClassification as EfficientNetForImageClassification, __webpack_exports__EfficientNetImageProcessor as EfficientNetImageProcessor, __webpack_exports__EfficientNetModel as EfficientNetModel, __webpack_exports__EfficientNetPreTrainedModel as EfficientNetPreTrainedModel, __webpack_exports__ElectraForMaskedLM as ElectraForMaskedLM, __webpack_exports__ElectraForQuestionAnswering as ElectraForQuestionAnswering, __webpack_exports__ElectraForSequenceClassification as ElectraForSequenceClassification, __webpack_exports__ElectraForTokenClassification as ElectraForTokenClassification, __webpack_exports__ElectraModel as ElectraModel, __webpack_exports__ElectraPreTrainedModel as ElectraPreTrainedModel, __webpack_exports__ElectraTokenizer as ElectraTokenizer, __webpack_exports__EosTokenCriteria as EosTokenCriteria, __webpack_exports__EsmForMaskedLM as EsmForMaskedLM, __webpack_exports__EsmForSequenceClassification as EsmForSequenceClassification, __webpack_exports__EsmForTokenClassification as EsmForTokenClassification, __webpack_exports__EsmModel as EsmModel, __webpack_exports__EsmPreTrainedModel as EsmPreTrainedModel, __webpack_exports__EsmTokenizer as EsmTokenizer, __webpack_exports__ExaoneForCausalLM as ExaoneForCausalLM, __webpack_exports__ExaoneModel as ExaoneModel, __webpack_exports__ExaonePreTrainedModel as ExaonePreTrainedModel, __webpack_exports__FFT as FFT, __webpack_exports__FalconForCausalLM as FalconForCausalLM, __webpack_exports__FalconModel as FalconModel, __webpack_exports__FalconPreTrainedModel as FalconPreTrainedModel, __webpack_exports__FalconTokenizer as FalconTokenizer, __webpack_exports__FastViTForImageClassification as FastViTForImageClassification, __webpack_exports__FastViTModel as FastViTModel, __webpack_exports__FastViTPreTrainedModel as FastViTPreTrainedModel, __webpack_exports__FeatureExtractionPipeline as FeatureExtractionPipeline, __webpack_exports__FeatureExtractor as FeatureExtractor, __webpack_exports__FillMaskPipeline as FillMaskPipeline, __webpack_exports__Florence2ForConditionalGeneration as Florence2ForConditionalGeneration, __webpack_exports__Florence2PreTrainedModel as Florence2PreTrainedModel, __webpack_exports__Florence2Processor as Florence2Processor, __webpack_exports__ForcedBOSTokenLogitsProcessor as ForcedBOSTokenLogitsProcessor, __webpack_exports__ForcedEOSTokenLogitsProcessor as ForcedEOSTokenLogitsProcessor, __webpack_exports__GLPNFeatureExtractor as GLPNFeatureExtractor, __webpack_exports__GLPNForDepthEstimation as GLPNForDepthEstimation, __webpack_exports__GLPNModel as GLPNModel, __webpack_exports__GLPNPreTrainedModel as GLPNPreTrainedModel, __webpack_exports__GPT2LMHeadModel as GPT2LMHeadModel, __webpack_exports__GPT2Model as GPT2Model, __webpack_exports__GPT2PreTrainedModel as GPT2PreTrainedModel, __webpack_exports__GPT2Tokenizer as GPT2Tokenizer, __webpack_exports__GPTBigCodeForCausalLM as GPTBigCodeForCausalLM, __webpack_exports__GPTBigCodeModel as GPTBigCodeModel, __webpack_exports__GPTBigCodePreTrainedModel as GPTBigCodePreTrainedModel, __webpack_exports__GPTJForCausalLM as GPTJForCausalLM, __webpack_exports__GPTJModel as GPTJModel, __webpack_exports__GPTJPreTrainedModel as GPTJPreTrainedModel, __webpack_exports__GPTNeoForCausalLM as GPTNeoForCausalLM, __webpack_exports__GPTNeoModel as GPTNeoModel, __webpack_exports__GPTNeoPreTrainedModel as GPTNeoPreTrainedModel, __webpack_exports__GPTNeoXForCausalLM as GPTNeoXForCausalLM, __webpack_exports__GPTNeoXModel as GPTNeoXModel, __webpack_exports__GPTNeoXPreTrainedModel as GPTNeoXPreTrainedModel, __webpack_exports__GPTNeoXTokenizer as GPTNeoXTokenizer, __webpack_exports__Gemma2ForCausalLM as Gemma2ForCausalLM, __webpack_exports__Gemma2Model as Gemma2Model, __webpack_exports__Gemma2PreTrainedModel as Gemma2PreTrainedModel, __webpack_exports__GemmaForCausalLM as GemmaForCausalLM, __webpack_exports__GemmaModel as GemmaModel, __webpack_exports__GemmaPreTrainedModel as GemmaPreTrainedModel, __webpack_exports__GemmaTokenizer as GemmaTokenizer, __webpack_exports__GraniteForCausalLM as GraniteForCausalLM, __webpack_exports__GraniteModel as GraniteModel, __webpack_exports__GranitePreTrainedModel as GranitePreTrainedModel, __webpack_exports__Grok1Tokenizer as Grok1Tokenizer, __webpack_exports__GroupViTModel as GroupViTModel, __webpack_exports__GroupViTPreTrainedModel as GroupViTPreTrainedModel, __webpack_exports__HerbertTokenizer as HerbertTokenizer, __webpack_exports__HieraForImageClassification as HieraForImageClassification, __webpack_exports__HieraModel as HieraModel, __webpack_exports__HieraPreTrainedModel as HieraPreTrainedModel, __webpack_exports__HubertForCTC as HubertForCTC, __webpack_exports__HubertForSequenceClassification as HubertForSequenceClassification, __webpack_exports__HubertModel as HubertModel, __webpack_exports__HubertPreTrainedModel as HubertPreTrainedModel, __webpack_exports__IJepaForImageClassification as IJepaForImageClassification, __webpack_exports__IJepaModel as IJepaModel, __webpack_exports__IJepaPreTrainedModel as IJepaPreTrainedModel, __webpack_exports__Idefics3ForConditionalGeneration as Idefics3ForConditionalGeneration, __webpack_exports__Idefics3ImageProcessor as Idefics3ImageProcessor, __webpack_exports__Idefics3PreTrainedModel as Idefics3PreTrainedModel, __webpack_exports__Idefics3Processor as Idefics3Processor, __webpack_exports__ImageClassificationPipeline as ImageClassificationPipeline, __webpack_exports__ImageFeatureExtractionPipeline as ImageFeatureExtractionPipeline, __webpack_exports__ImageFeatureExtractor as ImageFeatureExtractor, __webpack_exports__ImageMattingOutput as ImageMattingOutput, __webpack_exports__ImageProcessor as ImageProcessor, __webpack_exports__ImageSegmentationPipeline as ImageSegmentationPipeline, __webpack_exports__ImageToImagePipeline as ImageToImagePipeline, __webpack_exports__ImageToTextPipeline as ImageToTextPipeline, __webpack_exports__InterruptableStoppingCriteria as InterruptableStoppingCriteria, __webpack_exports__JAISLMHeadModel as JAISLMHeadModel, __webpack_exports__JAISModel as JAISModel, __webpack_exports__JAISPreTrainedModel as JAISPreTrainedModel, __webpack_exports__JinaCLIPImageProcessor as JinaCLIPImageProcessor, __webpack_exports__JinaCLIPModel as JinaCLIPModel, __webpack_exports__JinaCLIPPreTrainedModel as JinaCLIPPreTrainedModel, __webpack_exports__JinaCLIPProcessor as JinaCLIPProcessor, __webpack_exports__JinaCLIPTextModel as JinaCLIPTextModel, __webpack_exports__JinaCLIPVisionModel as JinaCLIPVisionModel, __webpack_exports__LlamaForCausalLM as LlamaForCausalLM, __webpack_exports__LlamaModel as LlamaModel, __webpack_exports__LlamaPreTrainedModel as LlamaPreTrainedModel, __webpack_exports__LlamaTokenizer as LlamaTokenizer, __webpack_exports__LlavaForConditionalGeneration as LlavaForConditionalGeneration, __webpack_exports__LlavaOnevisionForConditionalGeneration as LlavaOnevisionForConditionalGeneration, __webpack_exports__LlavaOnevisionImageProcessor as LlavaOnevisionImageProcessor, __webpack_exports__LlavaPreTrainedModel as LlavaPreTrainedModel, __webpack_exports__LogitsProcessor as LogitsProcessor, __webpack_exports__LogitsProcessorList as LogitsProcessorList, __webpack_exports__LogitsWarper as LogitsWarper, __webpack_exports__LongT5ForConditionalGeneration as LongT5ForConditionalGeneration, __webpack_exports__LongT5Model as LongT5Model, __webpack_exports__LongT5PreTrainedModel as LongT5PreTrainedModel, __webpack_exports__M2M100ForConditionalGeneration as M2M100ForConditionalGeneration, __webpack_exports__M2M100Model as M2M100Model, __webpack_exports__M2M100PreTrainedModel as M2M100PreTrainedModel, __webpack_exports__M2M100Tokenizer as M2M100Tokenizer, __webpack_exports__MBart50Tokenizer as MBart50Tokenizer, __webpack_exports__MBartForCausalLM as MBartForCausalLM, __webpack_exports__MBartForConditionalGeneration as MBartForConditionalGeneration, __webpack_exports__MBartForSequenceClassification as MBartForSequenceClassification, __webpack_exports__MBartModel as MBartModel, __webpack_exports__MBartPreTrainedModel as MBartPreTrainedModel, __webpack_exports__MBartTokenizer as MBartTokenizer, __webpack_exports__MPNetForMaskedLM as MPNetForMaskedLM, __webpack_exports__MPNetForQuestionAnswering as MPNetForQuestionAnswering, __webpack_exports__MPNetForSequenceClassification as MPNetForSequenceClassification, __webpack_exports__MPNetForTokenClassification as MPNetForTokenClassification, __webpack_exports__MPNetModel as MPNetModel, __webpack_exports__MPNetPreTrainedModel as MPNetPreTrainedModel, __webpack_exports__MPNetTokenizer as MPNetTokenizer, __webpack_exports__MT5ForConditionalGeneration as MT5ForConditionalGeneration, __webpack_exports__MT5Model as MT5Model, __webpack_exports__MT5PreTrainedModel as MT5PreTrainedModel, __webpack_exports__MarianMTModel as MarianMTModel, __webpack_exports__MarianModel as MarianModel, __webpack_exports__MarianPreTrainedModel as MarianPreTrainedModel, __webpack_exports__MarianTokenizer as MarianTokenizer, __webpack_exports__Mask2FormerImageProcessor as Mask2FormerImageProcessor, __webpack_exports__MaskFormerFeatureExtractor as MaskFormerFeatureExtractor, __webpack_exports__MaskFormerForInstanceSegmentation as MaskFormerForInstanceSegmentation, __webpack_exports__MaskFormerImageProcessor as MaskFormerImageProcessor, __webpack_exports__MaskFormerModel as MaskFormerModel, __webpack_exports__MaskFormerPreTrainedModel as MaskFormerPreTrainedModel, __webpack_exports__MaskedLMOutput as MaskedLMOutput, __webpack_exports__MaxLengthCriteria as MaxLengthCriteria, __webpack_exports__MgpstrForSceneTextRecognition as MgpstrForSceneTextRecognition, __webpack_exports__MgpstrModelOutput as MgpstrModelOutput, __webpack_exports__MgpstrPreTrainedModel as MgpstrPreTrainedModel, __webpack_exports__MgpstrProcessor as MgpstrProcessor, __webpack_exports__MgpstrTokenizer as MgpstrTokenizer, __webpack_exports__MinLengthLogitsProcessor as MinLengthLogitsProcessor, __webpack_exports__MinNewTokensLengthLogitsProcessor as MinNewTokensLengthLogitsProcessor, __webpack_exports__MistralForCausalLM as MistralForCausalLM, __webpack_exports__MistralModel as MistralModel, __webpack_exports__MistralPreTrainedModel as MistralPreTrainedModel, __webpack_exports__MobileBertForMaskedLM as MobileBertForMaskedLM, __webpack_exports__MobileBertForQuestionAnswering as MobileBertForQuestionAnswering, __webpack_exports__MobileBertForSequenceClassification as MobileBertForSequenceClassification, __webpack_exports__MobileBertModel as MobileBertModel, __webpack_exports__MobileBertPreTrainedModel as MobileBertPreTrainedModel, __webpack_exports__MobileBertTokenizer as MobileBertTokenizer, __webpack_exports__MobileLLMForCausalLM as MobileLLMForCausalLM, __webpack_exports__MobileLLMModel as MobileLLMModel, __webpack_exports__MobileLLMPreTrainedModel as MobileLLMPreTrainedModel, __webpack_exports__MobileNetV1FeatureExtractor as MobileNetV1FeatureExtractor, __webpack_exports__MobileNetV1ForImageClassification as MobileNetV1ForImageClassification, __webpack_exports__MobileNetV1ImageProcessor as MobileNetV1ImageProcessor, __webpack_exports__MobileNetV1Model as MobileNetV1Model, __webpack_exports__MobileNetV1PreTrainedModel as MobileNetV1PreTrainedModel, __webpack_exports__MobileNetV2FeatureExtractor as MobileNetV2FeatureExtractor, __webpack_exports__MobileNetV2ForImageClassification as MobileNetV2ForImageClassification, __webpack_exports__MobileNetV2ImageProcessor as MobileNetV2ImageProcessor, __webpack_exports__MobileNetV2Model as MobileNetV2Model, __webpack_exports__MobileNetV2PreTrainedModel as MobileNetV2PreTrainedModel, __webpack_exports__MobileNetV3FeatureExtractor as MobileNetV3FeatureExtractor, __webpack_exports__MobileNetV3ForImageClassification as MobileNetV3ForImageClassification, __webpack_exports__MobileNetV3ImageProcessor as MobileNetV3ImageProcessor, __webpack_exports__MobileNetV3Model as MobileNetV3Model, __webpack_exports__MobileNetV3PreTrainedModel as MobileNetV3PreTrainedModel, __webpack_exports__MobileNetV4FeatureExtractor as MobileNetV4FeatureExtractor, __webpack_exports__MobileNetV4ForImageClassification as MobileNetV4ForImageClassification, __webpack_exports__MobileNetV4ImageProcessor as MobileNetV4ImageProcessor, __webpack_exports__MobileNetV4Model as MobileNetV4Model, __webpack_exports__MobileNetV4PreTrainedModel as MobileNetV4PreTrainedModel, __webpack_exports__MobileViTFeatureExtractor as MobileViTFeatureExtractor, __webpack_exports__MobileViTForImageClassification as MobileViTForImageClassification, __webpack_exports__MobileViTImageProcessor as MobileViTImageProcessor, __webpack_exports__MobileViTModel as MobileViTModel, __webpack_exports__MobileViTPreTrainedModel as MobileViTPreTrainedModel, __webpack_exports__MobileViTV2ForImageClassification as MobileViTV2ForImageClassification, __webpack_exports__MobileViTV2Model as MobileViTV2Model, __webpack_exports__MobileViTV2PreTrainedModel as MobileViTV2PreTrainedModel, __webpack_exports__ModelOutput as ModelOutput, __webpack_exports__ModernBertForMaskedLM as ModernBertForMaskedLM, __webpack_exports__ModernBertForSequenceClassification as ModernBertForSequenceClassification, __webpack_exports__ModernBertForTokenClassification as ModernBertForTokenClassification, __webpack_exports__ModernBertModel as ModernBertModel, __webpack_exports__ModernBertPreTrainedModel as ModernBertPreTrainedModel, __webpack_exports__Moondream1ForConditionalGeneration as Moondream1ForConditionalGeneration, __webpack_exports__MoonshineFeatureExtractor as MoonshineFeatureExtractor, __webpack_exports__MoonshineForConditionalGeneration as MoonshineForConditionalGeneration, __webpack_exports__MoonshineModel as MoonshineModel, __webpack_exports__MoonshinePreTrainedModel as MoonshinePreTrainedModel, __webpack_exports__MoonshineProcessor as MoonshineProcessor, __webpack_exports__MptForCausalLM as MptForCausalLM, __webpack_exports__MptModel as MptModel, __webpack_exports__MptPreTrainedModel as MptPreTrainedModel, __webpack_exports__MultiModalityCausalLM as MultiModalityCausalLM, __webpack_exports__MultiModalityPreTrainedModel as MultiModalityPreTrainedModel, __webpack_exports__MusicgenForCausalLM as MusicgenForCausalLM, __webpack_exports__MusicgenForConditionalGeneration as MusicgenForConditionalGeneration, __webpack_exports__MusicgenModel as MusicgenModel, __webpack_exports__MusicgenPreTrainedModel as MusicgenPreTrainedModel, __webpack_exports__NllbTokenizer as NllbTokenizer, __webpack_exports__NoBadWordsLogitsProcessor as NoBadWordsLogitsProcessor, __webpack_exports__NoRepeatNGramLogitsProcessor as NoRepeatNGramLogitsProcessor, __webpack_exports__NomicBertModel as NomicBertModel, __webpack_exports__NomicBertPreTrainedModel as NomicBertPreTrainedModel, __webpack_exports__NougatImageProcessor as NougatImageProcessor, __webpack_exports__NougatTokenizer as NougatTokenizer, __webpack_exports__OPTForCausalLM as OPTForCausalLM, __webpack_exports__OPTModel as OPTModel, __webpack_exports__OPTPreTrainedModel as OPTPreTrainedModel, __webpack_exports__ObjectDetectionPipeline as ObjectDetectionPipeline, __webpack_exports__Olmo2ForCausalLM as Olmo2ForCausalLM, __webpack_exports__Olmo2Model as Olmo2Model, __webpack_exports__Olmo2PreTrainedModel as Olmo2PreTrainedModel, __webpack_exports__OlmoForCausalLM as OlmoForCausalLM, __webpack_exports__OlmoModel as OlmoModel, __webpack_exports__OlmoPreTrainedModel as OlmoPreTrainedModel, __webpack_exports__OpenELMForCausalLM as OpenELMForCausalLM, __webpack_exports__OpenELMModel as OpenELMModel, __webpack_exports__OpenELMPreTrainedModel as OpenELMPreTrainedModel, __webpack_exports__OwlViTFeatureExtractor as OwlViTFeatureExtractor, __webpack_exports__OwlViTForObjectDetection as OwlViTForObjectDetection, __webpack_exports__OwlViTImageProcessor as OwlViTImageProcessor, __webpack_exports__OwlViTModel as OwlViTModel, __webpack_exports__OwlViTPreTrainedModel as OwlViTPreTrainedModel, __webpack_exports__OwlViTProcessor as OwlViTProcessor, __webpack_exports__Owlv2ForObjectDetection as Owlv2ForObjectDetection, __webpack_exports__Owlv2ImageProcessor as Owlv2ImageProcessor, __webpack_exports__Owlv2Model as Owlv2Model, __webpack_exports__Owlv2PreTrainedModel as Owlv2PreTrainedModel, __webpack_exports__PaliGemmaForConditionalGeneration as PaliGemmaForConditionalGeneration, __webpack_exports__PaliGemmaPreTrainedModel as PaliGemmaPreTrainedModel, __webpack_exports__PaliGemmaProcessor as PaliGemmaProcessor, __webpack_exports__PatchTSMixerForPrediction as PatchTSMixerForPrediction, __webpack_exports__PatchTSMixerModel as PatchTSMixerModel, __webpack_exports__PatchTSMixerPreTrainedModel as PatchTSMixerPreTrainedModel, __webpack_exports__PatchTSTForPrediction as PatchTSTForPrediction, __webpack_exports__PatchTSTModel as PatchTSTModel, __webpack_exports__PatchTSTPreTrainedModel as PatchTSTPreTrainedModel, __webpack_exports__Phi3ForCausalLM as Phi3ForCausalLM, __webpack_exports__Phi3Model as Phi3Model, __webpack_exports__Phi3PreTrainedModel as Phi3PreTrainedModel, __webpack_exports__Phi3VForCausalLM as Phi3VForCausalLM, __webpack_exports__Phi3VImageProcessor as Phi3VImageProcessor, __webpack_exports__Phi3VPreTrainedModel as Phi3VPreTrainedModel, __webpack_exports__Phi3VProcessor as Phi3VProcessor, __webpack_exports__PhiForCausalLM as PhiForCausalLM, __webpack_exports__PhiModel as PhiModel, __webpack_exports__PhiPreTrainedModel as PhiPreTrainedModel, __webpack_exports__Pipeline as Pipeline, __webpack_exports__PreTrainedModel as PreTrainedModel, __webpack_exports__PreTrainedTokenizer as PreTrainedTokenizer, __webpack_exports__PretrainedConfig as PretrainedConfig, __webpack_exports__PretrainedMixin as PretrainedMixin, __webpack_exports__Processor as Processor, __webpack_exports__PvtForImageClassification as PvtForImageClassification, __webpack_exports__PvtImageProcessor as PvtImageProcessor, __webpack_exports__PvtModel as PvtModel, __webpack_exports__PvtPreTrainedModel as PvtPreTrainedModel, __webpack_exports__PyAnnoteFeatureExtractor as PyAnnoteFeatureExtractor, __webpack_exports__PyAnnoteForAudioFrameClassification as PyAnnoteForAudioFrameClassification, __webpack_exports__PyAnnoteModel as PyAnnoteModel, __webpack_exports__PyAnnotePreTrainedModel as PyAnnotePreTrainedModel, __webpack_exports__PyAnnoteProcessor as PyAnnoteProcessor, __webpack_exports__QuestionAnsweringModelOutput as QuestionAnsweringModelOutput, __webpack_exports__QuestionAnsweringPipeline as QuestionAnsweringPipeline, __webpack_exports__Qwen2ForCausalLM as Qwen2ForCausalLM, __webpack_exports__Qwen2Model as Qwen2Model, __webpack_exports__Qwen2PreTrainedModel as Qwen2PreTrainedModel, __webpack_exports__Qwen2Tokenizer as Qwen2Tokenizer, __webpack_exports__Qwen2VLForConditionalGeneration as Qwen2VLForConditionalGeneration, __webpack_exports__Qwen2VLImageProcessor as Qwen2VLImageProcessor, __webpack_exports__Qwen2VLPreTrainedModel as Qwen2VLPreTrainedModel, __webpack_exports__Qwen2VLProcessor as Qwen2VLProcessor, __webpack_exports__RTDetrForObjectDetection as RTDetrForObjectDetection, __webpack_exports__RTDetrImageProcessor as RTDetrImageProcessor, __webpack_exports__RTDetrModel as RTDetrModel, __webpack_exports__RTDetrObjectDetectionOutput as RTDetrObjectDetectionOutput, __webpack_exports__RTDetrPreTrainedModel as RTDetrPreTrainedModel, __webpack_exports__RawImage as RawImage, __webpack_exports__RepetitionPenaltyLogitsProcessor as RepetitionPenaltyLogitsProcessor, __webpack_exports__ResNetForImageClassification as ResNetForImageClassification, __webpack_exports__ResNetModel as ResNetModel, __webpack_exports__ResNetPreTrainedModel as ResNetPreTrainedModel, __webpack_exports__RoFormerForMaskedLM as RoFormerForMaskedLM, __webpack_exports__RoFormerForQuestionAnswering as RoFormerForQuestionAnswering, __webpack_exports__RoFormerForSequenceClassification as RoFormerForSequenceClassification, __webpack_exports__RoFormerForTokenClassification as RoFormerForTokenClassification, __webpack_exports__RoFormerModel as RoFormerModel, __webpack_exports__RoFormerPreTrainedModel as RoFormerPreTrainedModel, __webpack_exports__RoFormerTokenizer as RoFormerTokenizer, __webpack_exports__RobertaForMaskedLM as RobertaForMaskedLM, __webpack_exports__RobertaForQuestionAnswering as RobertaForQuestionAnswering, __webpack_exports__RobertaForSequenceClassification as RobertaForSequenceClassification, __webpack_exports__RobertaForTokenClassification as RobertaForTokenClassification, __webpack_exports__RobertaModel as RobertaModel, __webpack_exports__RobertaPreTrainedModel as RobertaPreTrainedModel, __webpack_exports__RobertaTokenizer as RobertaTokenizer, __webpack_exports__SamImageProcessor as SamImageProcessor, __webpack_exports__SamImageSegmentationOutput as SamImageSegmentationOutput, __webpack_exports__SamModel as SamModel, __webpack_exports__SamPreTrainedModel as SamPreTrainedModel, __webpack_exports__SamProcessor as SamProcessor, __webpack_exports__SapiensForDepthEstimation as SapiensForDepthEstimation, __webpack_exports__SapiensForNormalEstimation as SapiensForNormalEstimation, __webpack_exports__SapiensForSemanticSegmentation as SapiensForSemanticSegmentation, __webpack_exports__SapiensPreTrainedModel as SapiensPreTrainedModel, __webpack_exports__SeamlessM4TFeatureExtractor as SeamlessM4TFeatureExtractor, __webpack_exports__SegformerFeatureExtractor as SegformerFeatureExtractor, __webpack_exports__SegformerForImageClassification as SegformerForImageClassification, __webpack_exports__SegformerForSemanticSegmentation as SegformerForSemanticSegmentation, __webpack_exports__SegformerImageProcessor as SegformerImageProcessor, __webpack_exports__SegformerModel as SegformerModel, __webpack_exports__SegformerPreTrainedModel as SegformerPreTrainedModel, __webpack_exports__Seq2SeqLMOutput as Seq2SeqLMOutput, __webpack_exports__SequenceClassifierOutput as SequenceClassifierOutput, __webpack_exports__SiglipImageProcessor as SiglipImageProcessor, __webpack_exports__SiglipModel as SiglipModel, __webpack_exports__SiglipPreTrainedModel as SiglipPreTrainedModel, __webpack_exports__SiglipTextModel as SiglipTextModel, __webpack_exports__SiglipTokenizer as SiglipTokenizer, __webpack_exports__SiglipVisionModel as SiglipVisionModel, __webpack_exports__SpeechT5FeatureExtractor as SpeechT5FeatureExtractor, __webpack_exports__SpeechT5ForSpeechToText as SpeechT5ForSpeechToText, __webpack_exports__SpeechT5ForTextToSpeech as SpeechT5ForTextToSpeech, __webpack_exports__SpeechT5HifiGan as SpeechT5HifiGan, __webpack_exports__SpeechT5Model as SpeechT5Model, __webpack_exports__SpeechT5PreTrainedModel as SpeechT5PreTrainedModel, __webpack_exports__SpeechT5Processor as SpeechT5Processor, __webpack_exports__SpeechT5Tokenizer as SpeechT5Tokenizer, __webpack_exports__SqueezeBertForMaskedLM as SqueezeBertForMaskedLM, __webpack_exports__SqueezeBertForQuestionAnswering as SqueezeBertForQuestionAnswering, __webpack_exports__SqueezeBertForSequenceClassification as SqueezeBertForSequenceClassification, __webpack_exports__SqueezeBertModel as SqueezeBertModel, __webpack_exports__SqueezeBertPreTrainedModel as SqueezeBertPreTrainedModel, __webpack_exports__SqueezeBertTokenizer as SqueezeBertTokenizer, __webpack_exports__StableLmForCausalLM as StableLmForCausalLM, __webpack_exports__StableLmModel as StableLmModel, __webpack_exports__StableLmPreTrainedModel as StableLmPreTrainedModel, __webpack_exports__Starcoder2ForCausalLM as Starcoder2ForCausalLM, __webpack_exports__Starcoder2Model as Starcoder2Model, __webpack_exports__Starcoder2PreTrainedModel as Starcoder2PreTrainedModel, __webpack_exports__StoppingCriteria as StoppingCriteria, __webpack_exports__StoppingCriteriaList as StoppingCriteriaList, __webpack_exports__SummarizationPipeline as SummarizationPipeline, __webpack_exports__SuppressTokensAtBeginLogitsProcessor as SuppressTokensAtBeginLogitsProcessor, __webpack_exports__Swin2SRForImageSuperResolution as Swin2SRForImageSuperResolution, __webpack_exports__Swin2SRImageProcessor as Swin2SRImageProcessor, __webpack_exports__Swin2SRModel as Swin2SRModel, __webpack_exports__Swin2SRPreTrainedModel as Swin2SRPreTrainedModel, __webpack_exports__SwinForImageClassification as SwinForImageClassification, __webpack_exports__SwinModel as SwinModel, __webpack_exports__SwinPreTrainedModel as SwinPreTrainedModel, __webpack_exports__T5ForConditionalGeneration as T5ForConditionalGeneration, __webpack_exports__T5Model as T5Model, __webpack_exports__T5PreTrainedModel as T5PreTrainedModel, __webpack_exports__T5Tokenizer as T5Tokenizer, __webpack_exports__TableTransformerForObjectDetection as TableTransformerForObjectDetection, __webpack_exports__TableTransformerModel as TableTransformerModel, __webpack_exports__TableTransformerObjectDetectionOutput as TableTransformerObjectDetectionOutput, __webpack_exports__TableTransformerPreTrainedModel as TableTransformerPreTrainedModel, __webpack_exports__TemperatureLogitsWarper as TemperatureLogitsWarper, __webpack_exports__Tensor as Tensor, __webpack_exports__Text2TextGenerationPipeline as Text2TextGenerationPipeline, __webpack_exports__TextClassificationPipeline as TextClassificationPipeline, __webpack_exports__TextGenerationPipeline as TextGenerationPipeline, __webpack_exports__TextStreamer as TextStreamer, __webpack_exports__TextToAudioPipeline as TextToAudioPipeline, __webpack_exports__TokenClassificationPipeline as TokenClassificationPipeline, __webpack_exports__TokenClassifierOutput as TokenClassifierOutput, __webpack_exports__TokenizerModel as TokenizerModel, __webpack_exports__TopKLogitsWarper as TopKLogitsWarper, __webpack_exports__TopPLogitsWarper as TopPLogitsWarper, __webpack_exports__TrOCRForCausalLM as TrOCRForCausalLM, __webpack_exports__TrOCRPreTrainedModel as TrOCRPreTrainedModel, __webpack_exports__TranslationPipeline as TranslationPipeline, __webpack_exports__UniSpeechForCTC as UniSpeechForCTC, __webpack_exports__UniSpeechForSequenceClassification as UniSpeechForSequenceClassification, __webpack_exports__UniSpeechModel as UniSpeechModel, __webpack_exports__UniSpeechPreTrainedModel as UniSpeechPreTrainedModel, __webpack_exports__UniSpeechSatForAudioFrameClassification as UniSpeechSatForAudioFrameClassification, __webpack_exports__UniSpeechSatForCTC as UniSpeechSatForCTC, __webpack_exports__UniSpeechSatForSequenceClassification as UniSpeechSatForSequenceClassification, __webpack_exports__UniSpeechSatModel as UniSpeechSatModel, __webpack_exports__UniSpeechSatPreTrainedModel as UniSpeechSatPreTrainedModel, __webpack_exports__VLChatProcessor as VLChatProcessor, __webpack_exports__VLMImageProcessor as VLMImageProcessor, __webpack_exports__ViTFeatureExtractor as ViTFeatureExtractor, __webpack_exports__ViTForImageClassification as ViTForImageClassification, __webpack_exports__ViTImageProcessor as ViTImageProcessor, __webpack_exports__ViTMAEModel as ViTMAEModel, __webpack_exports__ViTMAEPreTrainedModel as ViTMAEPreTrainedModel, __webpack_exports__ViTMSNForImageClassification as ViTMSNForImageClassification, __webpack_exports__ViTMSNModel as ViTMSNModel, __webpack_exports__ViTMSNPreTrainedModel as ViTMSNPreTrainedModel, __webpack_exports__ViTModel as ViTModel, __webpack_exports__ViTPreTrainedModel as ViTPreTrainedModel, __webpack_exports__VisionEncoderDecoderModel as VisionEncoderDecoderModel, __webpack_exports__VitMatteForImageMatting as VitMatteForImageMatting, __webpack_exports__VitMatteImageProcessor as VitMatteImageProcessor, __webpack_exports__VitMattePreTrainedModel as VitMattePreTrainedModel, __webpack_exports__VitPoseForPoseEstimation as VitPoseForPoseEstimation, __webpack_exports__VitPoseImageProcessor as VitPoseImageProcessor, __webpack_exports__VitPosePreTrainedModel as VitPosePreTrainedModel, __webpack_exports__VitsModel as VitsModel, __webpack_exports__VitsModelOutput as VitsModelOutput, __webpack_exports__VitsPreTrainedModel as VitsPreTrainedModel, __webpack_exports__VitsTokenizer as VitsTokenizer, __webpack_exports__Wav2Vec2BertForCTC as Wav2Vec2BertForCTC, __webpack_exports__Wav2Vec2BertForSequenceClassification as Wav2Vec2BertForSequenceClassification, __webpack_exports__Wav2Vec2BertModel as Wav2Vec2BertModel, __webpack_exports__Wav2Vec2BertPreTrainedModel as Wav2Vec2BertPreTrainedModel, __webpack_exports__Wav2Vec2CTCTokenizer as Wav2Vec2CTCTokenizer, __webpack_exports__Wav2Vec2FeatureExtractor as Wav2Vec2FeatureExtractor, __webpack_exports__Wav2Vec2ForAudioFrameClassification as Wav2Vec2ForAudioFrameClassification, __webpack_exports__Wav2Vec2ForCTC as Wav2Vec2ForCTC, __webpack_exports__Wav2Vec2ForSequenceClassification as Wav2Vec2ForSequenceClassification, __webpack_exports__Wav2Vec2Model as Wav2Vec2Model, __webpack_exports__Wav2Vec2PreTrainedModel as Wav2Vec2PreTrainedModel, __webpack_exports__Wav2Vec2ProcessorWithLM as Wav2Vec2ProcessorWithLM, __webpack_exports__WavLMForAudioFrameClassification as WavLMForAudioFrameClassification, __webpack_exports__WavLMForCTC as WavLMForCTC, __webpack_exports__WavLMForSequenceClassification as WavLMForSequenceClassification, __webpack_exports__WavLMForXVector as WavLMForXVector, __webpack_exports__WavLMModel as WavLMModel, __webpack_exports__WavLMPreTrainedModel as WavLMPreTrainedModel, __webpack_exports__WeSpeakerFeatureExtractor as WeSpeakerFeatureExtractor, __webpack_exports__WeSpeakerResNetModel as WeSpeakerResNetModel, __webpack_exports__WeSpeakerResNetPreTrainedModel as WeSpeakerResNetPreTrainedModel, __webpack_exports__WhisperFeatureExtractor as WhisperFeatureExtractor, __webpack_exports__WhisperForConditionalGeneration as WhisperForConditionalGeneration, __webpack_exports__WhisperModel as WhisperModel, __webpack_exports__WhisperPreTrainedModel as WhisperPreTrainedModel, __webpack_exports__WhisperProcessor as WhisperProcessor, __webpack_exports__WhisperTextStreamer as WhisperTextStreamer, __webpack_exports__WhisperTimeStampLogitsProcessor as WhisperTimeStampLogitsProcessor, __webpack_exports__WhisperTokenizer as WhisperTokenizer, __webpack_exports__XLMForQuestionAnswering as XLMForQuestionAnswering, __webpack_exports__XLMForSequenceClassification as XLMForSequenceClassification, __webpack_exports__XLMForTokenClassification as XLMForTokenClassification, __webpack_exports__XLMModel as XLMModel, __webpack_exports__XLMPreTrainedModel as XLMPreTrainedModel, __webpack_exports__XLMRobertaForMaskedLM as XLMRobertaForMaskedLM, __webpack_exports__XLMRobertaForQuestionAnswering as XLMRobertaForQuestionAnswering, __webpack_exports__XLMRobertaForSequenceClassification as XLMRobertaForSequenceClassification, __webpack_exports__XLMRobertaForTokenClassification as XLMRobertaForTokenClassification, __webpack_exports__XLMRobertaModel as XLMRobertaModel, __webpack_exports__XLMRobertaPreTrainedModel as XLMRobertaPreTrainedModel, __webpack_exports__XLMRobertaTokenizer as XLMRobertaTokenizer, __webpack_exports__XLMTokenizer as XLMTokenizer, __webpack_exports__XLMWithLMHeadModel as XLMWithLMHeadModel, __webpack_exports__XVectorOutput as XVectorOutput, __webpack_exports__YolosFeatureExtractor as YolosFeatureExtractor, __webpack_exports__YolosForObjectDetection as YolosForObjectDetection, __webpack_exports__YolosImageProcessor as YolosImageProcessor, __webpack_exports__YolosModel as YolosModel, __webpack_exports__YolosObjectDetectionOutput as YolosObjectDetectionOutput, __webpack_exports__YolosPreTrainedModel as YolosPreTrainedModel, __webpack_exports__ZeroShotAudioClassificationPipeline as ZeroShotAudioClassificationPipeline, __webpack_exports__ZeroShotClassificationPipeline as ZeroShotClassificationPipeline, __webpack_exports__ZeroShotImageClassificationPipeline as ZeroShotImageClassificationPipeline, __webpack_exports__ZeroShotObjectDetectionPipeline as ZeroShotObjectDetectionPipeline, __webpack_exports__bankers_round as bankers_round, __webpack_exports__cat as cat, __webpack_exports__cos_sim as cos_sim, __webpack_exports__dot as dot, __webpack_exports__dynamic_time_warping as dynamic_time_warping, __webpack_exports__env as env, __webpack_exports__full as full, __webpack_exports__full_like as full_like, __webpack_exports__getKeyValueShapes as getKeyValueShapes, __webpack_exports__hamming as hamming, __webpack_exports__hanning as hanning, __webpack_exports__interpolate as interpolate, __webpack_exports__interpolate_4d as interpolate_4d, __webpack_exports__interpolate_data as interpolate_data, __webpack_exports__is_chinese_char as is_chinese_char, __webpack_exports__layer_norm as layer_norm, __webpack_exports__load_image as load_image, __webpack_exports__log_softmax as log_softmax, __webpack_exports__magnitude as magnitude, __webpack_exports__matmul as matmul, __webpack_exports__max as max, __webpack_exports__mean as mean, __webpack_exports__mean_pooling as mean_pooling, __webpack_exports__medianFilter as medianFilter, __webpack_exports__mel_filter_bank as mel_filter_bank, __webpack_exports__min as min, __webpack_exports__ones as ones, __webpack_exports__ones_like as ones_like, __webpack_exports__permute as permute, __webpack_exports__permute_data as permute_data, __webpack_exports__pipeline as pipeline, __webpack_exports__quantize_embeddings as quantize_embeddings, __webpack_exports__rand as rand, __webpack_exports__read_audio as read_audio, __webpack_exports__rfft as rfft, __webpack_exports__round as round, __webpack_exports__slice as slice, __webpack_exports__softmax as softmax, __webpack_exports__spectrogram as spectrogram, __webpack_exports__stack as stack, __webpack_exports__std_mean as std_mean, __webpack_exports__topk as topk, __webpack_exports__window_function as window_function, __webpack_exports__zeros as zeros, __webpack_exports__zeros_like as zeros_like };
|
|
38816
38958
|
|
|
38817
38959
|
//# sourceMappingURL=transformers.js.map
|