@huggingface/transformers 3.1.1 → 3.2.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +10 -4
- package/dist/ort-wasm-simd-threaded.jsep.wasm +0 -0
- package/dist/transformers.cjs +1062 -183
- package/dist/transformers.cjs.map +1 -1
- package/dist/transformers.js +2239 -1232
- package/dist/transformers.js.map +1 -1
- package/dist/transformers.min.cjs +1 -358
- package/dist/transformers.min.cjs.map +1 -1
- package/dist/transformers.min.js +1 -421
- package/dist/transformers.min.js.map +1 -1
- package/dist/transformers.min.mjs +1 -358
- package/dist/transformers.min.mjs.map +1 -1
- package/dist/transformers.mjs +1082 -181
- package/dist/transformers.mjs.map +1 -1
- package/package.json +11 -16
- package/src/backends/onnx.js +2 -7
- package/src/base/image_processors_utils.js +3 -1
- package/src/configs.js +11 -2
- package/src/env.js +1 -1
- package/src/models/feature_extractors.js +1 -0
- package/src/models/idefics3/image_processing_idefics3.js +24 -13
- package/src/models/image_processors.js +1 -0
- package/src/models/moonshine/feature_extraction_moonshine.js +26 -0
- package/src/models/moonshine/processing_moonshine.js +20 -0
- package/src/models/paligemma/processing_paligemma.js +82 -0
- package/src/models/phi3_v/image_processing_phi3_v.js +163 -0
- package/src/models/phi3_v/processing_phi3_v.js +53 -0
- package/src/models/processors.js +3 -0
- package/src/models/pyannote/feature_extraction_pyannote.js +56 -0
- package/src/models/pyannote/processing_pyannote.js +7 -54
- package/src/models.js +233 -35
- package/src/ops/registry.js +11 -0
- package/src/pipelines.js +30 -0
- package/src/tokenizers.js +12 -1
- package/src/utils/core.js +39 -9
- package/src/utils/hub.js +8 -12
- package/src/utils/image.js +40 -0
- package/src/utils/tensor.js +51 -1
- package/types/backends/onnx.d.ts +2 -2
- package/types/backends/onnx.d.ts.map +1 -1
- package/types/base/feature_extraction_utils.d.ts +1 -1
- package/types/base/feature_extraction_utils.d.ts.map +1 -1
- package/types/base/image_processors_utils.d.ts +4 -4
- package/types/base/image_processors_utils.d.ts.map +1 -1
- package/types/base/processing_utils.d.ts +4 -4
- package/types/base/processing_utils.d.ts.map +1 -1
- package/types/configs.d.ts +7 -7
- package/types/configs.d.ts.map +1 -1
- package/types/env.d.ts +1 -1
- package/types/env.d.ts.map +1 -1
- package/types/generation/configuration_utils.d.ts +2 -2
- package/types/generation/logits_process.d.ts +2 -2
- package/types/generation/logits_process.d.ts.map +1 -1
- package/types/generation/logits_sampler.d.ts.map +1 -1
- package/types/generation/parameters.d.ts +5 -5
- package/types/generation/stopping_criteria.d.ts +1 -1
- package/types/generation/stopping_criteria.d.ts.map +1 -1
- package/types/generation/streamers.d.ts +2 -2
- package/types/generation/streamers.d.ts.map +1 -1
- package/types/models/audio_spectrogram_transformer/feature_extraction_audio_spectrogram_transformer.d.ts +1 -1
- package/types/models/audio_spectrogram_transformer/feature_extraction_audio_spectrogram_transformer.d.ts.map +1 -1
- package/types/models/auto/feature_extraction_auto.d.ts.map +1 -1
- package/types/models/auto/image_processing_auto.d.ts.map +1 -1
- package/types/models/auto/processing_auto.d.ts +1 -1
- package/types/models/auto/processing_auto.d.ts.map +1 -1
- package/types/models/clap/feature_extraction_clap.d.ts +1 -1
- package/types/models/clap/feature_extraction_clap.d.ts.map +1 -1
- package/types/models/detr/image_processing_detr.d.ts +11 -11
- package/types/models/detr/image_processing_detr.d.ts.map +1 -1
- package/types/models/donut/image_processing_donut.d.ts +1 -1
- package/types/models/donut/image_processing_donut.d.ts.map +1 -1
- package/types/models/feature_extractors.d.ts +1 -0
- package/types/models/florence2/processing_florence2.d.ts.map +1 -1
- package/types/models/idefics3/image_processing_idefics3.d.ts.map +1 -1
- package/types/models/idefics3/processing_idefics3.d.ts.map +1 -1
- package/types/models/image_processors.d.ts +1 -0
- package/types/models/janus/image_processing_janus.d.ts +1 -1
- package/types/models/janus/image_processing_janus.d.ts.map +1 -1
- package/types/models/janus/processing_janus.d.ts.map +1 -1
- package/types/models/maskformer/image_processing_maskformer.d.ts +8 -8
- package/types/models/maskformer/image_processing_maskformer.d.ts.map +1 -1
- package/types/models/mgp_str/processing_mgp_str.d.ts +2 -2
- package/types/models/mgp_str/processing_mgp_str.d.ts.map +1 -1
- package/types/models/moonshine/feature_extraction_moonshine.d.ts +13 -0
- package/types/models/moonshine/feature_extraction_moonshine.d.ts.map +1 -0
- package/types/models/moonshine/processing_moonshine.d.ts +17 -0
- package/types/models/moonshine/processing_moonshine.d.ts.map +1 -0
- package/types/models/owlvit/image_processing_owlvit.d.ts.map +1 -1
- package/types/models/paligemma/processing_paligemma.d.ts +12 -0
- package/types/models/paligemma/processing_paligemma.d.ts.map +1 -0
- package/types/models/phi3_v/image_processing_phi3_v.d.ts +17 -0
- package/types/models/phi3_v/image_processing_phi3_v.d.ts.map +1 -0
- package/types/models/phi3_v/processing_phi3_v.d.ts +17 -0
- package/types/models/phi3_v/processing_phi3_v.d.ts.map +1 -0
- package/types/models/processors.d.ts +3 -0
- package/types/models/pyannote/feature_extraction_pyannote.d.ts +18 -0
- package/types/models/pyannote/feature_extraction_pyannote.d.ts.map +1 -1
- package/types/models/pyannote/processing_pyannote.d.ts +4 -15
- package/types/models/pyannote/processing_pyannote.d.ts.map +1 -1
- package/types/models/qwen2_vl/processing_qwen2_vl.d.ts.map +1 -1
- package/types/models/rt_detr/image_processing_rt_detr.d.ts.map +1 -1
- package/types/models/sam/image_processing_sam.d.ts.map +1 -1
- package/types/models/seamless_m4t/feature_extraction_seamless_m4t.d.ts +1 -1
- package/types/models/seamless_m4t/feature_extraction_seamless_m4t.d.ts.map +1 -1
- package/types/models/segformer/image_processing_segformer.d.ts.map +1 -1
- package/types/models/speecht5/processing_speecht5.d.ts.map +1 -1
- package/types/models/swin2sr/image_processing_swin2sr.d.ts +1 -1
- package/types/models/swin2sr/image_processing_swin2sr.d.ts.map +1 -1
- package/types/models/vitmatte/image_processing_vitmatte.d.ts.map +1 -1
- package/types/models/vitpose/image_processing_vitpose.d.ts +1 -1
- package/types/models/vitpose/image_processing_vitpose.d.ts.map +1 -1
- package/types/models/wav2vec2/feature_extraction_wav2vec2.d.ts.map +1 -1
- package/types/models/wav2vec2/processing_wav2vec2.d.ts.map +1 -1
- package/types/models/wespeaker/feature_extraction_wespeaker.d.ts +1 -1
- package/types/models/wespeaker/feature_extraction_wespeaker.d.ts.map +1 -1
- package/types/models/whisper/feature_extraction_whisper.d.ts +1 -1
- package/types/models/whisper/feature_extraction_whisper.d.ts.map +1 -1
- package/types/models/whisper/generation_whisper.d.ts.map +1 -1
- package/types/models/whisper/processing_whisper.d.ts.map +1 -1
- package/types/models/yolos/image_processing_yolos.d.ts.map +1 -1
- package/types/models.d.ts +61 -5
- package/types/models.d.ts.map +1 -1
- package/types/ops/registry.d.ts +1 -0
- package/types/ops/registry.d.ts.map +1 -1
- package/types/pipelines.d.ts +31 -51
- package/types/pipelines.d.ts.map +1 -1
- package/types/tokenizers.d.ts +10 -6
- package/types/tokenizers.d.ts.map +1 -1
- package/types/utils/audio.d.ts.map +1 -1
- package/types/utils/constants.d.ts.map +1 -1
- package/types/utils/core.d.ts +87 -22
- package/types/utils/core.d.ts.map +1 -1
- package/types/utils/data-structures.d.ts.map +1 -1
- package/types/utils/devices.d.ts.map +1 -1
- package/types/utils/dtypes.d.ts.map +1 -1
- package/types/utils/generic.d.ts.map +1 -1
- package/types/utils/hub.d.ts +3 -3
- package/types/utils/hub.d.ts.map +1 -1
- package/types/utils/image.d.ts +10 -1
- package/types/utils/image.d.ts.map +1 -1
- package/types/utils/maths.d.ts +10 -10
- package/types/utils/maths.d.ts.map +1 -1
- package/types/utils/tensor.d.ts +22 -6
- package/types/utils/tensor.d.ts.map +1 -1
package/dist/transformers.mjs
CHANGED
|
@@ -55,10 +55,10 @@ module.exports = __WEBPACK_EXTERNAL_MODULE_url__;
|
|
|
55
55
|
|
|
56
56
|
/***/ }),
|
|
57
57
|
|
|
58
|
-
/***/ "?
|
|
59
|
-
|
|
60
|
-
!***
|
|
61
|
-
|
|
58
|
+
/***/ "?8b6b":
|
|
59
|
+
/*!*********************************!*\
|
|
60
|
+
!*** onnxruntime-web (ignored) ***!
|
|
61
|
+
\*********************************/
|
|
62
62
|
/***/ (() => {
|
|
63
63
|
|
|
64
64
|
/* (ignored) */
|
|
@@ -3871,7 +3871,7 @@ const version = '1.20.1';
|
|
|
3871
3871
|
\******************************/
|
|
3872
3872
|
/***/ ((__unused_webpack___webpack_module__, __webpack_exports__, __webpack_require__) => {
|
|
3873
3873
|
|
|
3874
|
-
var
|
|
3874
|
+
var onnxruntime_web__WEBPACK_IMPORTED_MODULE_2___namespace_cache;
|
|
3875
3875
|
__webpack_require__.r(__webpack_exports__);
|
|
3876
3876
|
/* harmony export */ __webpack_require__.d(__webpack_exports__, {
|
|
3877
3877
|
/* harmony export */ Tensor: () => (/* reexport safe */ onnxruntime_common__WEBPACK_IMPORTED_MODULE_3__.Tensor),
|
|
@@ -3882,7 +3882,7 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
3882
3882
|
/* harmony export */ });
|
|
3883
3883
|
/* harmony import */ var _env_js__WEBPACK_IMPORTED_MODULE_0__ = __webpack_require__(/*! ../env.js */ "./src/env.js");
|
|
3884
3884
|
/* harmony import */ var onnxruntime_node__WEBPACK_IMPORTED_MODULE_1__ = __webpack_require__(/*! onnxruntime-node */ "onnxruntime-node");
|
|
3885
|
-
/* harmony import */ var
|
|
3885
|
+
/* harmony import */ var onnxruntime_web__WEBPACK_IMPORTED_MODULE_2__ = __webpack_require__(/*! onnxruntime-web */ "?8b6b");
|
|
3886
3886
|
/* harmony import */ var onnxruntime_common__WEBPACK_IMPORTED_MODULE_3__ = __webpack_require__(/*! onnxruntime-common */ "./node_modules/onnxruntime-common/dist/esm/index.js");
|
|
3887
3887
|
/**
|
|
3888
3888
|
* @file Handler file for choosing the correct version of ONNX Runtime, based on the environment.
|
|
@@ -3908,11 +3908,6 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
3908
3908
|
// In either case, we select the default export if it exists, otherwise we use the named export.
|
|
3909
3909
|
|
|
3910
3910
|
|
|
3911
|
-
// Use subpath-imports to ensure Node.js and browser interoperability.
|
|
3912
|
-
// See package.json and https://nodejs.org/api/packages.html#subpath-imports
|
|
3913
|
-
// for more information.
|
|
3914
|
-
// @ts-ignore
|
|
3915
|
-
|
|
3916
3911
|
|
|
3917
3912
|
|
|
3918
3913
|
|
|
@@ -3954,7 +3949,7 @@ if (ORT_SYMBOL in globalThis) {
|
|
|
3954
3949
|
} else if (_env_js__WEBPACK_IMPORTED_MODULE_0__.apis.IS_NODE_ENV) {
|
|
3955
3950
|
ONNX = onnxruntime_node__WEBPACK_IMPORTED_MODULE_1__["default"] ?? onnxruntime_node__WEBPACK_IMPORTED_MODULE_1__;
|
|
3956
3951
|
|
|
3957
|
-
// Updated as of ONNX Runtime 1.
|
|
3952
|
+
// Updated as of ONNX Runtime 1.20.1
|
|
3958
3953
|
// The following table lists the supported versions of ONNX Runtime Node.js binding provided with pre-built binaries.
|
|
3959
3954
|
// | EPs/Platforms | Windows x64 | Windows arm64 | Linux x64 | Linux arm64 | MacOS x64 | MacOS arm64 |
|
|
3960
3955
|
// | ------------- | ----------- | ------------- | ----------------- | ----------- | --------- | ----------- |
|
|
@@ -3977,7 +3972,7 @@ if (ORT_SYMBOL in globalThis) {
|
|
|
3977
3972
|
supportedDevices.push('cpu');
|
|
3978
3973
|
defaultDevices = ['cpu'];
|
|
3979
3974
|
} else {
|
|
3980
|
-
ONNX = /*#__PURE__*/ (
|
|
3975
|
+
ONNX = /*#__PURE__*/ (onnxruntime_web__WEBPACK_IMPORTED_MODULE_2___namespace_cache || (onnxruntime_web__WEBPACK_IMPORTED_MODULE_2___namespace_cache = __webpack_require__.t(onnxruntime_web__WEBPACK_IMPORTED_MODULE_2__, 2)));
|
|
3981
3976
|
|
|
3982
3977
|
if (_env_js__WEBPACK_IMPORTED_MODULE_0__.apis.IS_WEBNN_AVAILABLE) {
|
|
3983
3978
|
// TODO: Only push supported providers (depending on available hardware)
|
|
@@ -4898,7 +4893,7 @@ class ImageProcessor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_0__.Call
|
|
|
4898
4893
|
* Pad the image by a certain amount.
|
|
4899
4894
|
* @param {Float32Array} pixelData The pixel data to pad.
|
|
4900
4895
|
* @param {number[]} imgDims The dimensions of the image (height, width, channels).
|
|
4901
|
-
* @param {{width:number; height:number}|number} padSize The dimensions of the padded image.
|
|
4896
|
+
* @param {{width:number; height:number}|number|'square'} padSize The dimensions of the padded image.
|
|
4902
4897
|
* @param {Object} options The options for padding.
|
|
4903
4898
|
* @param {'constant'|'symmetric'} [options.mode='constant'] The type of padding to add.
|
|
4904
4899
|
* @param {boolean} [options.center=false] Whether to center the image.
|
|
@@ -4916,6 +4911,8 @@ class ImageProcessor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_0__.Call
|
|
|
4916
4911
|
if (typeof padSize === 'number') {
|
|
4917
4912
|
paddedImageWidth = padSize;
|
|
4918
4913
|
paddedImageHeight = padSize;
|
|
4914
|
+
} else if (padSize === 'square') {
|
|
4915
|
+
paddedImageWidth = paddedImageHeight = Math.max(imageHeight, imageWidth);
|
|
4919
4916
|
} else {
|
|
4920
4917
|
paddedImageWidth = padSize.width;
|
|
4921
4918
|
paddedImageHeight = padSize.height;
|
|
@@ -5559,8 +5556,6 @@ function getNormalizedConfig(config) {
|
|
|
5559
5556
|
case 'gpt_neox':
|
|
5560
5557
|
case 'stablelm':
|
|
5561
5558
|
case 'opt':
|
|
5562
|
-
case 'phi':
|
|
5563
|
-
case 'phi3':
|
|
5564
5559
|
case 'falcon':
|
|
5565
5560
|
mapping['num_heads'] = 'num_attention_heads';
|
|
5566
5561
|
mapping['num_layers'] = 'num_hidden_layers';
|
|
@@ -5568,6 +5563,7 @@ function getNormalizedConfig(config) {
|
|
|
5568
5563
|
break;
|
|
5569
5564
|
case 'llama':
|
|
5570
5565
|
case 'olmo':
|
|
5566
|
+
case 'olmo2':
|
|
5571
5567
|
case 'mobilellm':
|
|
5572
5568
|
case 'granite':
|
|
5573
5569
|
case 'cohere':
|
|
@@ -5575,6 +5571,9 @@ function getNormalizedConfig(config) {
|
|
|
5575
5571
|
case 'starcoder2':
|
|
5576
5572
|
case 'qwen2':
|
|
5577
5573
|
case 'qwen2_vl':
|
|
5574
|
+
case 'phi':
|
|
5575
|
+
case 'phi3':
|
|
5576
|
+
case 'phi3_v':
|
|
5578
5577
|
mapping['num_heads'] = 'num_key_value_heads';
|
|
5579
5578
|
mapping['num_layers'] = 'num_hidden_layers';
|
|
5580
5579
|
mapping['hidden_size'] = 'hidden_size';
|
|
@@ -5607,6 +5606,12 @@ function getNormalizedConfig(config) {
|
|
|
5607
5606
|
mapping['num_layers'] = 'n_layers';
|
|
5608
5607
|
mapping['hidden_size'] = 'd_model';
|
|
5609
5608
|
break;
|
|
5609
|
+
case 'exaone':
|
|
5610
|
+
mapping['num_heads'] = 'num_key_value_heads';
|
|
5611
|
+
mapping['num_layers'] = 'num_layers';
|
|
5612
|
+
mapping['dim_kv'] = 'head_dim';
|
|
5613
|
+
mapping['num_attention_heads'] = 'num_attention_heads';
|
|
5614
|
+
break;
|
|
5610
5615
|
|
|
5611
5616
|
// Encoder-decoder models
|
|
5612
5617
|
case 't5':
|
|
@@ -5648,6 +5653,7 @@ function getNormalizedConfig(config) {
|
|
|
5648
5653
|
mapping['encoder_hidden_size'] = mapping['decoder_hidden_size'] = 'd_model';
|
|
5649
5654
|
break;
|
|
5650
5655
|
case 'musicgen_decoder':
|
|
5656
|
+
case 'moonshine':
|
|
5651
5657
|
mapping['num_encoder_layers'] = mapping['num_decoder_layers'] = 'num_hidden_layers';
|
|
5652
5658
|
mapping['num_encoder_heads'] = mapping['num_decoder_heads'] = 'num_attention_heads';
|
|
5653
5659
|
mapping['encoder_hidden_size'] = mapping['decoder_hidden_size'] = 'hidden_size';
|
|
@@ -5896,7 +5902,7 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
5896
5902
|
|
|
5897
5903
|
|
|
5898
5904
|
|
|
5899
|
-
const VERSION = '3.
|
|
5905
|
+
const VERSION = '3.2.0';
|
|
5900
5906
|
|
|
5901
5907
|
// Check if various APIs are available (depends on environment)
|
|
5902
5908
|
const IS_BROWSER_ENV = typeof window !== "undefined" && typeof window.document !== "undefined";
|
|
@@ -7988,6 +7994,9 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
7988
7994
|
/* harmony export */ EsmForTokenClassification: () => (/* binding */ EsmForTokenClassification),
|
|
7989
7995
|
/* harmony export */ EsmModel: () => (/* binding */ EsmModel),
|
|
7990
7996
|
/* harmony export */ EsmPreTrainedModel: () => (/* binding */ EsmPreTrainedModel),
|
|
7997
|
+
/* harmony export */ ExaoneForCausalLM: () => (/* binding */ ExaoneForCausalLM),
|
|
7998
|
+
/* harmony export */ ExaoneModel: () => (/* binding */ ExaoneModel),
|
|
7999
|
+
/* harmony export */ ExaonePreTrainedModel: () => (/* binding */ ExaonePreTrainedModel),
|
|
7991
8000
|
/* harmony export */ FalconForCausalLM: () => (/* binding */ FalconForCausalLM),
|
|
7992
8001
|
/* harmony export */ FalconModel: () => (/* binding */ FalconModel),
|
|
7993
8002
|
/* harmony export */ FalconPreTrainedModel: () => (/* binding */ FalconPreTrainedModel),
|
|
@@ -8032,6 +8041,9 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
8032
8041
|
/* harmony export */ HubertForSequenceClassification: () => (/* binding */ HubertForSequenceClassification),
|
|
8033
8042
|
/* harmony export */ HubertModel: () => (/* binding */ HubertModel),
|
|
8034
8043
|
/* harmony export */ HubertPreTrainedModel: () => (/* binding */ HubertPreTrainedModel),
|
|
8044
|
+
/* harmony export */ IJepaForImageClassification: () => (/* binding */ IJepaForImageClassification),
|
|
8045
|
+
/* harmony export */ IJepaModel: () => (/* binding */ IJepaModel),
|
|
8046
|
+
/* harmony export */ IJepaPreTrainedModel: () => (/* binding */ IJepaPreTrainedModel),
|
|
8035
8047
|
/* harmony export */ Idefics3ForConditionalGeneration: () => (/* binding */ Idefics3ForConditionalGeneration),
|
|
8036
8048
|
/* harmony export */ Idefics3PreTrainedModel: () => (/* binding */ Idefics3PreTrainedModel),
|
|
8037
8049
|
/* harmony export */ ImageMattingOutput: () => (/* binding */ ImageMattingOutput),
|
|
@@ -8109,6 +8121,9 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
8109
8121
|
/* harmony export */ MobileViTV2PreTrainedModel: () => (/* binding */ MobileViTV2PreTrainedModel),
|
|
8110
8122
|
/* harmony export */ ModelOutput: () => (/* binding */ ModelOutput),
|
|
8111
8123
|
/* harmony export */ Moondream1ForConditionalGeneration: () => (/* binding */ Moondream1ForConditionalGeneration),
|
|
8124
|
+
/* harmony export */ MoonshineForConditionalGeneration: () => (/* binding */ MoonshineForConditionalGeneration),
|
|
8125
|
+
/* harmony export */ MoonshineModel: () => (/* binding */ MoonshineModel),
|
|
8126
|
+
/* harmony export */ MoonshinePreTrainedModel: () => (/* binding */ MoonshinePreTrainedModel),
|
|
8112
8127
|
/* harmony export */ MptForCausalLM: () => (/* binding */ MptForCausalLM),
|
|
8113
8128
|
/* harmony export */ MptModel: () => (/* binding */ MptModel),
|
|
8114
8129
|
/* harmony export */ MptPreTrainedModel: () => (/* binding */ MptPreTrainedModel),
|
|
@@ -8123,6 +8138,9 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
8123
8138
|
/* harmony export */ OPTForCausalLM: () => (/* binding */ OPTForCausalLM),
|
|
8124
8139
|
/* harmony export */ OPTModel: () => (/* binding */ OPTModel),
|
|
8125
8140
|
/* harmony export */ OPTPreTrainedModel: () => (/* binding */ OPTPreTrainedModel),
|
|
8141
|
+
/* harmony export */ Olmo2ForCausalLM: () => (/* binding */ Olmo2ForCausalLM),
|
|
8142
|
+
/* harmony export */ Olmo2Model: () => (/* binding */ Olmo2Model),
|
|
8143
|
+
/* harmony export */ Olmo2PreTrainedModel: () => (/* binding */ Olmo2PreTrainedModel),
|
|
8126
8144
|
/* harmony export */ OlmoForCausalLM: () => (/* binding */ OlmoForCausalLM),
|
|
8127
8145
|
/* harmony export */ OlmoModel: () => (/* binding */ OlmoModel),
|
|
8128
8146
|
/* harmony export */ OlmoPreTrainedModel: () => (/* binding */ OlmoPreTrainedModel),
|
|
@@ -8135,6 +8153,8 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
8135
8153
|
/* harmony export */ Owlv2ForObjectDetection: () => (/* binding */ Owlv2ForObjectDetection),
|
|
8136
8154
|
/* harmony export */ Owlv2Model: () => (/* binding */ Owlv2Model),
|
|
8137
8155
|
/* harmony export */ Owlv2PreTrainedModel: () => (/* binding */ Owlv2PreTrainedModel),
|
|
8156
|
+
/* harmony export */ PaliGemmaForConditionalGeneration: () => (/* binding */ PaliGemmaForConditionalGeneration),
|
|
8157
|
+
/* harmony export */ PaliGemmaPreTrainedModel: () => (/* binding */ PaliGemmaPreTrainedModel),
|
|
8138
8158
|
/* harmony export */ PatchTSMixerForPrediction: () => (/* binding */ PatchTSMixerForPrediction),
|
|
8139
8159
|
/* harmony export */ PatchTSMixerModel: () => (/* binding */ PatchTSMixerModel),
|
|
8140
8160
|
/* harmony export */ PatchTSMixerPreTrainedModel: () => (/* binding */ PatchTSMixerPreTrainedModel),
|
|
@@ -8144,6 +8164,8 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
8144
8164
|
/* harmony export */ Phi3ForCausalLM: () => (/* binding */ Phi3ForCausalLM),
|
|
8145
8165
|
/* harmony export */ Phi3Model: () => (/* binding */ Phi3Model),
|
|
8146
8166
|
/* harmony export */ Phi3PreTrainedModel: () => (/* binding */ Phi3PreTrainedModel),
|
|
8167
|
+
/* harmony export */ Phi3VForCausalLM: () => (/* binding */ Phi3VForCausalLM),
|
|
8168
|
+
/* harmony export */ Phi3VPreTrainedModel: () => (/* binding */ Phi3VPreTrainedModel),
|
|
8147
8169
|
/* harmony export */ PhiForCausalLM: () => (/* binding */ PhiForCausalLM),
|
|
8148
8170
|
/* harmony export */ PhiModel: () => (/* binding */ PhiModel),
|
|
8149
8171
|
/* harmony export */ PhiPreTrainedModel: () => (/* binding */ PhiPreTrainedModel),
|
|
@@ -8389,6 +8411,7 @@ const MODEL_TYPES = {
|
|
|
8389
8411
|
ImageTextToText: 6,
|
|
8390
8412
|
Musicgen: 7,
|
|
8391
8413
|
MultiModality: 8,
|
|
8414
|
+
Phi3V: 9,
|
|
8392
8415
|
}
|
|
8393
8416
|
//////////////////////////////////////////////////
|
|
8394
8417
|
|
|
@@ -8816,7 +8839,9 @@ async function decoderForward(self, model_inputs, is_encoder_decoder = false) {
|
|
|
8816
8839
|
new_model_inputs.use_cache_branch = boolTensor(!!past_key_values);
|
|
8817
8840
|
}
|
|
8818
8841
|
if (session.inputNames.includes('position_ids') && new_model_inputs.attention_mask && !new_model_inputs.position_ids) {
|
|
8819
|
-
|
|
8842
|
+
// NOTE: Handle a special case for paligemma models, where positions are 1-indexed
|
|
8843
|
+
const start_index = self.config.model_type === 'paligemma' ? 1 : 0;
|
|
8844
|
+
new_model_inputs.position_ids = createPositionIds(new_model_inputs, past_key_values, start_index);
|
|
8820
8845
|
}
|
|
8821
8846
|
|
|
8822
8847
|
// Unpack the `past_key_values` object into model inputs
|
|
@@ -8952,14 +8977,14 @@ async function imageTextToTextForward(self, {
|
|
|
8952
8977
|
* @param {Tensor} attention_mask
|
|
8953
8978
|
* @returns {{data: BigInt64Array, dims: number[]}}
|
|
8954
8979
|
*/
|
|
8955
|
-
function cumsum_masked_fill(attention_mask) {
|
|
8980
|
+
function cumsum_masked_fill(attention_mask, start_index = 0) {
|
|
8956
8981
|
const [bz, seq_len] = attention_mask.dims;
|
|
8957
8982
|
const attn_mask_data = attention_mask.data;
|
|
8958
8983
|
|
|
8959
8984
|
const data = new BigInt64Array(attn_mask_data.length);
|
|
8960
8985
|
for (let i = 0; i < bz; ++i) {
|
|
8961
8986
|
const start = i * seq_len;
|
|
8962
|
-
let sum = BigInt(
|
|
8987
|
+
let sum = BigInt(start_index);
|
|
8963
8988
|
for (let j = 0; j < seq_len; ++j) {
|
|
8964
8989
|
const index = start + j;
|
|
8965
8990
|
if (attn_mask_data[index] === 0n) {
|
|
@@ -8986,10 +9011,10 @@ function cumsum_masked_fill(attention_mask) {
|
|
|
8986
9011
|
* position_ids = position_ids[:, -input_ids.shape[1] :]
|
|
8987
9012
|
* ```
|
|
8988
9013
|
*/
|
|
8989
|
-
function createPositionIds(model_inputs, past_key_values = null) {
|
|
9014
|
+
function createPositionIds(model_inputs, past_key_values = null, start_index = 0) {
|
|
8990
9015
|
const { input_ids, inputs_embeds, attention_mask } = model_inputs;
|
|
8991
9016
|
|
|
8992
|
-
const { data, dims } = cumsum_masked_fill(attention_mask);
|
|
9017
|
+
const { data, dims } = cumsum_masked_fill(attention_mask, start_index);
|
|
8993
9018
|
let position_ids = new _utils_tensor_js__WEBPACK_IMPORTED_MODULE_9__.Tensor('int64', data, dims);
|
|
8994
9019
|
if (past_key_values) {
|
|
8995
9020
|
const offset = -(input_ids ?? inputs_embeds).dims.at(1);
|
|
@@ -9162,6 +9187,10 @@ class PreTrainedModel extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_3__.Cal
|
|
|
9162
9187
|
this._forward = imageTextToTextForward;
|
|
9163
9188
|
this._prepare_inputs_for_generation = image_text_to_text_prepare_inputs_for_generation;
|
|
9164
9189
|
break;
|
|
9190
|
+
case MODEL_TYPES.Phi3V:
|
|
9191
|
+
this.can_generate = true;
|
|
9192
|
+
this._prepare_inputs_for_generation = image_text_to_text_prepare_inputs_for_generation;
|
|
9193
|
+
break;
|
|
9165
9194
|
|
|
9166
9195
|
case MODEL_TYPES.MultiModality:
|
|
9167
9196
|
this.can_generate = true;
|
|
@@ -9326,6 +9355,18 @@ class PreTrainedModel extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_3__.Cal
|
|
|
9326
9355
|
}, options),
|
|
9327
9356
|
]);
|
|
9328
9357
|
|
|
9358
|
+
} else if (modelType === MODEL_TYPES.Phi3V) {
|
|
9359
|
+
info = await Promise.all([
|
|
9360
|
+
constructSessions(pretrained_model_name_or_path, {
|
|
9361
|
+
prepare_inputs_embeds: 'prepare_inputs_embeds',
|
|
9362
|
+
model: 'model',
|
|
9363
|
+
vision_encoder: 'vision_encoder',
|
|
9364
|
+
}, options),
|
|
9365
|
+
getOptionalConfigs(pretrained_model_name_or_path, {
|
|
9366
|
+
generation_config: 'generation_config.json',
|
|
9367
|
+
}, options),
|
|
9368
|
+
]);
|
|
9369
|
+
|
|
9329
9370
|
} else { // should be MODEL_TYPES.EncoderOnly
|
|
9330
9371
|
if (modelType !== MODEL_TYPES.EncoderOnly) {
|
|
9331
9372
|
const type = modelName ?? config?.model_type;
|
|
@@ -11598,6 +11639,29 @@ class WhisperForConditionalGeneration extends WhisperPreTrainedModel {
|
|
|
11598
11639
|
}
|
|
11599
11640
|
//////////////////////////////////////////////////
|
|
11600
11641
|
|
|
11642
|
+
|
|
11643
|
+
//////////////////////////////////////////////////
|
|
11644
|
+
// Moonshine models
|
|
11645
|
+
class MoonshinePreTrainedModel extends PreTrainedModel {
|
|
11646
|
+
|
|
11647
|
+
requires_attention_mask = false;
|
|
11648
|
+
main_input_name = 'input_values';
|
|
11649
|
+
forward_params = [
|
|
11650
|
+
'input_values',
|
|
11651
|
+
'decoder_input_ids',
|
|
11652
|
+
'past_key_values',
|
|
11653
|
+
];
|
|
11654
|
+
};
|
|
11655
|
+
|
|
11656
|
+
/**
|
|
11657
|
+
* MoonshineModel class for training Moonshine models without a language model head.
|
|
11658
|
+
*/
|
|
11659
|
+
class MoonshineModel extends MoonshinePreTrainedModel { }
|
|
11660
|
+
|
|
11661
|
+
class MoonshineForConditionalGeneration extends MoonshinePreTrainedModel { }
|
|
11662
|
+
//////////////////////////////////////////////////
|
|
11663
|
+
|
|
11664
|
+
|
|
11601
11665
|
//////////////////////////////////////////////////
|
|
11602
11666
|
/**
|
|
11603
11667
|
* Vision Encoder-Decoder model based on OpenAI's GPT architecture for image captioning and other vision tasks
|
|
@@ -11806,6 +11870,30 @@ class Florence2ForConditionalGeneration extends Florence2PreTrainedModel {
|
|
|
11806
11870
|
}
|
|
11807
11871
|
}
|
|
11808
11872
|
|
|
11873
|
+
class PaliGemmaPreTrainedModel extends PreTrainedModel {
|
|
11874
|
+
forward_params = [
|
|
11875
|
+
'input_ids',
|
|
11876
|
+
// 'inputs_embeds',
|
|
11877
|
+
'attention_mask',
|
|
11878
|
+
'pixel_values',
|
|
11879
|
+
'position_ids',
|
|
11880
|
+
'past_key_values',
|
|
11881
|
+
];
|
|
11882
|
+
}
|
|
11883
|
+
|
|
11884
|
+
class PaliGemmaForConditionalGeneration extends PaliGemmaPreTrainedModel {
|
|
11885
|
+
_merge_input_ids_with_image_features(kwargs) {
|
|
11886
|
+
const vision_hidden_size = kwargs.image_features.dims.at(-1);
|
|
11887
|
+
const reshaped_image_hidden_states = kwargs.image_features.view(-1, vision_hidden_size);
|
|
11888
|
+
|
|
11889
|
+
return default_merge_input_ids_with_image_features({
|
|
11890
|
+
// @ts-ignore
|
|
11891
|
+
image_token_id: this.config.image_token_index,
|
|
11892
|
+
...kwargs,
|
|
11893
|
+
image_features: reshaped_image_hidden_states,
|
|
11894
|
+
})
|
|
11895
|
+
}
|
|
11896
|
+
}
|
|
11809
11897
|
|
|
11810
11898
|
//////////////////////////////////////////////////
|
|
11811
11899
|
// Idefics3 Models
|
|
@@ -11844,6 +11932,77 @@ class Idefics3ForConditionalGeneration extends Idefics3PreTrainedModel {
|
|
|
11844
11932
|
}
|
|
11845
11933
|
//////////////////////////////////////////////////
|
|
11846
11934
|
|
|
11935
|
+
class Phi3VPreTrainedModel extends PreTrainedModel {
|
|
11936
|
+
forward_params = [
|
|
11937
|
+
'input_ids',
|
|
11938
|
+
'inputs_embeds',
|
|
11939
|
+
'attention_mask',
|
|
11940
|
+
'position_ids',
|
|
11941
|
+
'pixel_values',
|
|
11942
|
+
'image_sizes',
|
|
11943
|
+
'past_key_values',
|
|
11944
|
+
];
|
|
11945
|
+
}
|
|
11946
|
+
class Phi3VForCausalLM extends Phi3VPreTrainedModel {
|
|
11947
|
+
|
|
11948
|
+
async forward({
|
|
11949
|
+
// Produced by the tokenizer/processor:
|
|
11950
|
+
input_ids = null,
|
|
11951
|
+
attention_mask = null,
|
|
11952
|
+
pixel_values = null,
|
|
11953
|
+
image_sizes = null,
|
|
11954
|
+
|
|
11955
|
+
// Used during generation:
|
|
11956
|
+
position_ids = null,
|
|
11957
|
+
inputs_embeds = null,
|
|
11958
|
+
past_key_values = null,
|
|
11959
|
+
|
|
11960
|
+
// Generic generation parameters
|
|
11961
|
+
generation_config = null,
|
|
11962
|
+
logits_processor = null,
|
|
11963
|
+
|
|
11964
|
+
// TODO: needed?
|
|
11965
|
+
...kwargs
|
|
11966
|
+
}) {
|
|
11967
|
+
if (!inputs_embeds) {
|
|
11968
|
+
let image_features;
|
|
11969
|
+
if (pixel_values && input_ids.dims[1] !== 1) {
|
|
11970
|
+
if (!image_sizes) {
|
|
11971
|
+
throw new Error('`image_sizes` must be provided when `pixel_values` is provided.');
|
|
11972
|
+
}
|
|
11973
|
+
|
|
11974
|
+
// Encode the image
|
|
11975
|
+
({ image_features } = await sessionRun(this.sessions['vision_encoder'], {
|
|
11976
|
+
pixel_values,
|
|
11977
|
+
image_sizes,
|
|
11978
|
+
}));
|
|
11979
|
+
} else {
|
|
11980
|
+
const hidden_size = this.config.normalized_config.hidden_size;
|
|
11981
|
+
image_features = new _utils_tensor_js__WEBPACK_IMPORTED_MODULE_9__.Tensor(
|
|
11982
|
+
'float32',
|
|
11983
|
+
[],
|
|
11984
|
+
[0, hidden_size],
|
|
11985
|
+
);
|
|
11986
|
+
}
|
|
11987
|
+
|
|
11988
|
+
({ inputs_embeds } = await sessionRun(this.sessions['prepare_inputs_embeds'], {
|
|
11989
|
+
input_ids,
|
|
11990
|
+
image_features,
|
|
11991
|
+
}));
|
|
11992
|
+
}
|
|
11993
|
+
|
|
11994
|
+
const outputs = await decoderForward(this, {
|
|
11995
|
+
inputs_embeds,
|
|
11996
|
+
past_key_values,
|
|
11997
|
+
attention_mask,
|
|
11998
|
+
position_ids,
|
|
11999
|
+
generation_config,
|
|
12000
|
+
logits_processor,
|
|
12001
|
+
}, false);
|
|
12002
|
+
return outputs;
|
|
12003
|
+
}
|
|
12004
|
+
}
|
|
12005
|
+
|
|
11847
12006
|
//////////////////////////////////////////////////
|
|
11848
12007
|
class CLIPPreTrainedModel extends PreTrainedModel { }
|
|
11849
12008
|
|
|
@@ -11898,9 +12057,11 @@ class CLIPModel extends CLIPPreTrainedModel { }
|
|
|
11898
12057
|
class CLIPTextModel extends CLIPPreTrainedModel {
|
|
11899
12058
|
/** @type {typeof PreTrainedModel.from_pretrained} */
|
|
11900
12059
|
static async from_pretrained(pretrained_model_name_or_path, options = {}) {
|
|
11901
|
-
|
|
11902
|
-
|
|
11903
|
-
|
|
12060
|
+
return super.from_pretrained(pretrained_model_name_or_path, {
|
|
12061
|
+
// Update default model file name if not provided
|
|
12062
|
+
model_file_name: 'text_model',
|
|
12063
|
+
...options,
|
|
12064
|
+
});
|
|
11904
12065
|
}
|
|
11905
12066
|
}
|
|
11906
12067
|
|
|
@@ -11933,9 +12094,11 @@ class CLIPTextModel extends CLIPPreTrainedModel {
|
|
|
11933
12094
|
class CLIPTextModelWithProjection extends CLIPPreTrainedModel {
|
|
11934
12095
|
/** @type {typeof PreTrainedModel.from_pretrained} */
|
|
11935
12096
|
static async from_pretrained(pretrained_model_name_or_path, options = {}) {
|
|
11936
|
-
|
|
11937
|
-
|
|
11938
|
-
|
|
12097
|
+
return super.from_pretrained(pretrained_model_name_or_path, {
|
|
12098
|
+
// Update default model file name if not provided
|
|
12099
|
+
model_file_name: 'text_model',
|
|
12100
|
+
...options,
|
|
12101
|
+
});
|
|
11939
12102
|
}
|
|
11940
12103
|
}
|
|
11941
12104
|
|
|
@@ -11945,9 +12108,11 @@ class CLIPTextModelWithProjection extends CLIPPreTrainedModel {
|
|
|
11945
12108
|
class CLIPVisionModel extends CLIPPreTrainedModel {
|
|
11946
12109
|
/** @type {typeof PreTrainedModel.from_pretrained} */
|
|
11947
12110
|
static async from_pretrained(pretrained_model_name_or_path, options = {}) {
|
|
11948
|
-
|
|
11949
|
-
|
|
11950
|
-
|
|
12111
|
+
return super.from_pretrained(pretrained_model_name_or_path, {
|
|
12112
|
+
// Update default model file name if not provided
|
|
12113
|
+
model_file_name: 'vision_model',
|
|
12114
|
+
...options,
|
|
12115
|
+
});
|
|
11951
12116
|
}
|
|
11952
12117
|
}
|
|
11953
12118
|
|
|
@@ -11980,9 +12145,11 @@ class CLIPVisionModel extends CLIPPreTrainedModel {
|
|
|
11980
12145
|
class CLIPVisionModelWithProjection extends CLIPPreTrainedModel {
|
|
11981
12146
|
/** @type {typeof PreTrainedModel.from_pretrained} */
|
|
11982
12147
|
static async from_pretrained(pretrained_model_name_or_path, options = {}) {
|
|
11983
|
-
|
|
11984
|
-
|
|
11985
|
-
|
|
12148
|
+
return super.from_pretrained(pretrained_model_name_or_path, {
|
|
12149
|
+
// Update default model file name if not provided
|
|
12150
|
+
model_file_name: 'vision_model',
|
|
12151
|
+
...options,
|
|
12152
|
+
});
|
|
11986
12153
|
}
|
|
11987
12154
|
}
|
|
11988
12155
|
//////////////////////////////////////////////////
|
|
@@ -12066,9 +12233,11 @@ class SiglipModel extends SiglipPreTrainedModel { }
|
|
|
12066
12233
|
class SiglipTextModel extends SiglipPreTrainedModel {
|
|
12067
12234
|
/** @type {typeof PreTrainedModel.from_pretrained} */
|
|
12068
12235
|
static async from_pretrained(pretrained_model_name_or_path, options = {}) {
|
|
12069
|
-
|
|
12070
|
-
|
|
12071
|
-
|
|
12236
|
+
return super.from_pretrained(pretrained_model_name_or_path, {
|
|
12237
|
+
// Update default model file name if not provided
|
|
12238
|
+
model_file_name: 'text_model',
|
|
12239
|
+
...options,
|
|
12240
|
+
});
|
|
12072
12241
|
}
|
|
12073
12242
|
}
|
|
12074
12243
|
|
|
@@ -12101,9 +12270,11 @@ class SiglipTextModel extends SiglipPreTrainedModel {
|
|
|
12101
12270
|
class SiglipVisionModel extends CLIPPreTrainedModel {
|
|
12102
12271
|
/** @type {typeof PreTrainedModel.from_pretrained} */
|
|
12103
12272
|
static async from_pretrained(pretrained_model_name_or_path, options = {}) {
|
|
12104
|
-
|
|
12105
|
-
|
|
12106
|
-
|
|
12273
|
+
return super.from_pretrained(pretrained_model_name_or_path, {
|
|
12274
|
+
// Update default model file name if not provided
|
|
12275
|
+
model_file_name: 'vision_model',
|
|
12276
|
+
...options,
|
|
12277
|
+
});
|
|
12107
12278
|
}
|
|
12108
12279
|
}
|
|
12109
12280
|
//////////////////////////////////////////////////
|
|
@@ -12158,18 +12329,22 @@ class JinaCLIPModel extends JinaCLIPPreTrainedModel {
|
|
|
12158
12329
|
class JinaCLIPTextModel extends JinaCLIPPreTrainedModel {
|
|
12159
12330
|
/** @type {typeof PreTrainedModel.from_pretrained} */
|
|
12160
12331
|
static async from_pretrained(pretrained_model_name_or_path, options = {}) {
|
|
12161
|
-
|
|
12162
|
-
|
|
12163
|
-
|
|
12332
|
+
return super.from_pretrained(pretrained_model_name_or_path, {
|
|
12333
|
+
// Update default model file name if not provided
|
|
12334
|
+
model_file_name: 'text_model',
|
|
12335
|
+
...options,
|
|
12336
|
+
});
|
|
12164
12337
|
}
|
|
12165
12338
|
}
|
|
12166
12339
|
|
|
12167
12340
|
class JinaCLIPVisionModel extends JinaCLIPPreTrainedModel {
|
|
12168
12341
|
/** @type {typeof PreTrainedModel.from_pretrained} */
|
|
12169
12342
|
static async from_pretrained(pretrained_model_name_or_path, options = {}) {
|
|
12170
|
-
|
|
12171
|
-
|
|
12172
|
-
|
|
12343
|
+
return super.from_pretrained(pretrained_model_name_or_path, {
|
|
12344
|
+
// Update default model file name if not provided
|
|
12345
|
+
model_file_name: 'vision_model',
|
|
12346
|
+
...options,
|
|
12347
|
+
});
|
|
12173
12348
|
}
|
|
12174
12349
|
}
|
|
12175
12350
|
//////////////////////////////////////////////////
|
|
@@ -12329,6 +12504,14 @@ class LlamaForCausalLM extends LlamaPreTrainedModel { }
|
|
|
12329
12504
|
//////////////////////////////////////////////////
|
|
12330
12505
|
|
|
12331
12506
|
|
|
12507
|
+
//////////////////////////////////////////////////
|
|
12508
|
+
// EXAONE models
|
|
12509
|
+
class ExaonePreTrainedModel extends PreTrainedModel { }
|
|
12510
|
+
class ExaoneModel extends ExaonePreTrainedModel { }
|
|
12511
|
+
class ExaoneForCausalLM extends ExaonePreTrainedModel { }
|
|
12512
|
+
//////////////////////////////////////////////////
|
|
12513
|
+
|
|
12514
|
+
|
|
12332
12515
|
//////////////////////////////////////////////////
|
|
12333
12516
|
// MobileLLM models
|
|
12334
12517
|
class MobileLLMPreTrainedModel extends PreTrainedModel { }
|
|
@@ -12344,6 +12527,13 @@ class OlmoModel extends OlmoPreTrainedModel { }
|
|
|
12344
12527
|
class OlmoForCausalLM extends OlmoPreTrainedModel { }
|
|
12345
12528
|
//////////////////////////////////////////////////
|
|
12346
12529
|
|
|
12530
|
+
//////////////////////////////////////////////////
|
|
12531
|
+
// OLMo2 models
|
|
12532
|
+
class Olmo2PreTrainedModel extends PreTrainedModel { }
|
|
12533
|
+
class Olmo2Model extends Olmo2PreTrainedModel { }
|
|
12534
|
+
class Olmo2ForCausalLM extends Olmo2PreTrainedModel { }
|
|
12535
|
+
//////////////////////////////////////////////////
|
|
12536
|
+
|
|
12347
12537
|
|
|
12348
12538
|
//////////////////////////////////////////////////
|
|
12349
12539
|
// Granite models
|
|
@@ -12760,6 +12950,20 @@ class ViTForImageClassification extends ViTPreTrainedModel {
|
|
|
12760
12950
|
//////////////////////////////////////////////////
|
|
12761
12951
|
|
|
12762
12952
|
|
|
12953
|
+
//////////////////////////////////////////////////
|
|
12954
|
+
class IJepaPreTrainedModel extends PreTrainedModel { }
|
|
12955
|
+
class IJepaModel extends IJepaPreTrainedModel { }
|
|
12956
|
+
class IJepaForImageClassification extends IJepaPreTrainedModel {
|
|
12957
|
+
/**
|
|
12958
|
+
* @param {any} model_inputs
|
|
12959
|
+
*/
|
|
12960
|
+
async _call(model_inputs) {
|
|
12961
|
+
return new SequenceClassifierOutput(await super._call(model_inputs));
|
|
12962
|
+
}
|
|
12963
|
+
}
|
|
12964
|
+
//////////////////////////////////////////////////
|
|
12965
|
+
|
|
12966
|
+
|
|
12763
12967
|
//////////////////////////////////////////////////
|
|
12764
12968
|
class VitPosePreTrainedModel extends PreTrainedModel { }
|
|
12765
12969
|
|
|
@@ -14370,9 +14574,11 @@ class ClapModel extends ClapPreTrainedModel { }
|
|
|
14370
14574
|
class ClapTextModelWithProjection extends ClapPreTrainedModel {
|
|
14371
14575
|
/** @type {typeof PreTrainedModel.from_pretrained} */
|
|
14372
14576
|
static async from_pretrained(pretrained_model_name_or_path, options = {}) {
|
|
14373
|
-
|
|
14374
|
-
|
|
14375
|
-
|
|
14577
|
+
return super.from_pretrained(pretrained_model_name_or_path, {
|
|
14578
|
+
// Update default model file name if not provided
|
|
14579
|
+
model_file_name: 'text_model',
|
|
14580
|
+
...options,
|
|
14581
|
+
});
|
|
14376
14582
|
}
|
|
14377
14583
|
}
|
|
14378
14584
|
|
|
@@ -14405,9 +14611,11 @@ class ClapTextModelWithProjection extends ClapPreTrainedModel {
|
|
|
14405
14611
|
class ClapAudioModelWithProjection extends ClapPreTrainedModel {
|
|
14406
14612
|
/** @type {typeof PreTrainedModel.from_pretrained} */
|
|
14407
14613
|
static async from_pretrained(pretrained_model_name_or_path, options = {}) {
|
|
14408
|
-
|
|
14409
|
-
|
|
14410
|
-
|
|
14614
|
+
return super.from_pretrained(pretrained_model_name_or_path, {
|
|
14615
|
+
// Update default model file name if not provided
|
|
14616
|
+
model_file_name: 'audio_model',
|
|
14617
|
+
...options,
|
|
14618
|
+
});
|
|
14411
14619
|
}
|
|
14412
14620
|
}
|
|
14413
14621
|
//////////////////////////////////////////////////
|
|
@@ -15030,6 +15238,7 @@ const MODEL_MAPPING_NAMES_ENCODER_ONLY = new Map([
|
|
|
15030
15238
|
['rt_detr', ['RTDetrModel', RTDetrModel]],
|
|
15031
15239
|
['table-transformer', ['TableTransformerModel', TableTransformerModel]],
|
|
15032
15240
|
['vit', ['ViTModel', ViTModel]],
|
|
15241
|
+
['ijepa', ['IJepaModel', IJepaModel]],
|
|
15033
15242
|
['pvt', ['PvtModel', PvtModel]],
|
|
15034
15243
|
['vit_msn', ['ViTMSNModel', ViTMSNModel]],
|
|
15035
15244
|
['vit_mae', ['ViTMAEModel', ViTMAEModel]],
|
|
@@ -15093,7 +15302,9 @@ const MODEL_MAPPING_NAMES_DECODER_ONLY = new Map([
|
|
|
15093
15302
|
['gpt_neox', ['GPTNeoXModel', GPTNeoXModel]],
|
|
15094
15303
|
['codegen', ['CodeGenModel', CodeGenModel]],
|
|
15095
15304
|
['llama', ['LlamaModel', LlamaModel]],
|
|
15305
|
+
['exaone', ['ExaoneModel', ExaoneModel]],
|
|
15096
15306
|
['olmo', ['OlmoModel', OlmoModel]],
|
|
15307
|
+
['olmo2', ['Olmo2Model', Olmo2Model]],
|
|
15097
15308
|
['mobilellm', ['MobileLLMModel', MobileLLMModel]],
|
|
15098
15309
|
['granite', ['GraniteModel', GraniteModel]],
|
|
15099
15310
|
['cohere', ['CohereModel', CohereModel]],
|
|
@@ -15114,6 +15325,7 @@ const MODEL_MAPPING_NAMES_DECODER_ONLY = new Map([
|
|
|
15114
15325
|
const MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING_NAMES = new Map([
|
|
15115
15326
|
['speecht5', ['SpeechT5ForSpeechToText', SpeechT5ForSpeechToText]],
|
|
15116
15327
|
['whisper', ['WhisperForConditionalGeneration', WhisperForConditionalGeneration]],
|
|
15328
|
+
['moonshine', ['MoonshineForConditionalGeneration', MoonshineForConditionalGeneration]],
|
|
15117
15329
|
]);
|
|
15118
15330
|
|
|
15119
15331
|
const MODEL_FOR_TEXT_TO_SPECTROGRAM_MAPPING_NAMES = new Map([
|
|
@@ -15184,7 +15396,9 @@ const MODEL_FOR_CAUSAL_LM_MAPPING_NAMES = new Map([
|
|
|
15184
15396
|
['gpt_neox', ['GPTNeoXForCausalLM', GPTNeoXForCausalLM]],
|
|
15185
15397
|
['codegen', ['CodeGenForCausalLM', CodeGenForCausalLM]],
|
|
15186
15398
|
['llama', ['LlamaForCausalLM', LlamaForCausalLM]],
|
|
15399
|
+
['exaone', ['ExaoneForCausalLM', ExaoneForCausalLM]],
|
|
15187
15400
|
['olmo', ['OlmoForCausalLM', OlmoForCausalLM]],
|
|
15401
|
+
['olmo2', ['Olmo2ForCausalLM', Olmo2ForCausalLM]],
|
|
15188
15402
|
['mobilellm', ['MobileLLMForCausalLM', MobileLLMForCausalLM]],
|
|
15189
15403
|
['granite', ['GraniteForCausalLM', GraniteForCausalLM]],
|
|
15190
15404
|
['cohere', ['CohereForCausalLM', CohereForCausalLM]],
|
|
@@ -15202,6 +15416,9 @@ const MODEL_FOR_CAUSAL_LM_MAPPING_NAMES = new Map([
|
|
|
15202
15416
|
['falcon', ['FalconForCausalLM', FalconForCausalLM]],
|
|
15203
15417
|
['trocr', ['TrOCRForCausalLM', TrOCRForCausalLM]],
|
|
15204
15418
|
['stablelm', ['StableLmForCausalLM', StableLmForCausalLM]],
|
|
15419
|
+
|
|
15420
|
+
// Also image-text-to-text
|
|
15421
|
+
['phi3_v', ['Phi3VForCausalLM', Phi3VForCausalLM]],
|
|
15205
15422
|
]);
|
|
15206
15423
|
|
|
15207
15424
|
const MODEL_FOR_MULTIMODALITY_MAPPING_NAMES = new Map([
|
|
@@ -15258,6 +15475,7 @@ const MODEL_FOR_IMAGE_TEXT_TO_TEXT_MAPPING_NAMES = new Map([
|
|
|
15258
15475
|
['florence2', ['Florence2ForConditionalGeneration', Florence2ForConditionalGeneration]],
|
|
15259
15476
|
['qwen2-vl', ['Qwen2VLForConditionalGeneration', Qwen2VLForConditionalGeneration]],
|
|
15260
15477
|
['idefics3', ['Idefics3ForConditionalGeneration', Idefics3ForConditionalGeneration]],
|
|
15478
|
+
['paligemma', ['PaliGemmaForConditionalGeneration', PaliGemmaForConditionalGeneration]],
|
|
15261
15479
|
]);
|
|
15262
15480
|
|
|
15263
15481
|
const MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING_NAMES = new Map([
|
|
@@ -15266,6 +15484,7 @@ const MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING_NAMES = new Map([
|
|
|
15266
15484
|
|
|
15267
15485
|
const MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES = new Map([
|
|
15268
15486
|
['vit', ['ViTForImageClassification', ViTForImageClassification]],
|
|
15487
|
+
['ijepa', ['IJepaForImageClassification', IJepaForImageClassification]],
|
|
15269
15488
|
['pvt', ['PvtForImageClassification', PvtForImageClassification]],
|
|
15270
15489
|
['vit_msn', ['ViTMSNForImageClassification', ViTMSNForImageClassification]],
|
|
15271
15490
|
['fastvit', ['FastViTForImageClassification', FastViTForImageClassification]],
|
|
@@ -15437,6 +15656,7 @@ const CUSTOM_MAPPING = [
|
|
|
15437
15656
|
// OVERRIDE:
|
|
15438
15657
|
// TODO: Refactor to allow class to specify model
|
|
15439
15658
|
['MusicgenForConditionalGeneration', MusicgenForConditionalGeneration, MODEL_TYPES.Musicgen],
|
|
15659
|
+
['Phi3VForCausalLM', Phi3VForCausalLM, MODEL_TYPES.Phi3V],
|
|
15440
15660
|
|
|
15441
15661
|
['CLIPTextModelWithProjection', CLIPTextModelWithProjection, MODEL_TYPES.EncoderOnly],
|
|
15442
15662
|
['SiglipTextModel', SiglipTextModel, MODEL_TYPES.EncoderOnly],
|
|
@@ -16675,23 +16895,26 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
16675
16895
|
/* harmony export */ __webpack_require__.d(__webpack_exports__, {
|
|
16676
16896
|
/* harmony export */ ASTFeatureExtractor: () => (/* reexport safe */ _audio_spectrogram_transformer_feature_extraction_audio_spectrogram_transformer_js__WEBPACK_IMPORTED_MODULE_0__.ASTFeatureExtractor),
|
|
16677
16897
|
/* harmony export */ ClapFeatureExtractor: () => (/* reexport safe */ _clap_feature_extraction_clap_js__WEBPACK_IMPORTED_MODULE_1__.ClapFeatureExtractor),
|
|
16678
|
-
/* harmony export */ ImageFeatureExtractor: () => (/* reexport safe */
|
|
16679
|
-
/* harmony export */
|
|
16680
|
-
/* harmony export */
|
|
16681
|
-
/* harmony export */
|
|
16682
|
-
/* harmony export */
|
|
16683
|
-
/* harmony export */
|
|
16684
|
-
/* harmony export */
|
|
16898
|
+
/* harmony export */ ImageFeatureExtractor: () => (/* reexport safe */ _base_image_processors_utils_js__WEBPACK_IMPORTED_MODULE_9__.ImageProcessor),
|
|
16899
|
+
/* harmony export */ MoonshineFeatureExtractor: () => (/* reexport safe */ _moonshine_feature_extraction_moonshine_js__WEBPACK_IMPORTED_MODULE_2__.MoonshineFeatureExtractor),
|
|
16900
|
+
/* harmony export */ PyAnnoteFeatureExtractor: () => (/* reexport safe */ _pyannote_feature_extraction_pyannote_js__WEBPACK_IMPORTED_MODULE_3__.PyAnnoteFeatureExtractor),
|
|
16901
|
+
/* harmony export */ SeamlessM4TFeatureExtractor: () => (/* reexport safe */ _seamless_m4t_feature_extraction_seamless_m4t_js__WEBPACK_IMPORTED_MODULE_4__.SeamlessM4TFeatureExtractor),
|
|
16902
|
+
/* harmony export */ SpeechT5FeatureExtractor: () => (/* reexport safe */ _speecht5_feature_extraction_speecht5_js__WEBPACK_IMPORTED_MODULE_5__.SpeechT5FeatureExtractor),
|
|
16903
|
+
/* harmony export */ Wav2Vec2FeatureExtractor: () => (/* reexport safe */ _wav2vec2_feature_extraction_wav2vec2_js__WEBPACK_IMPORTED_MODULE_6__.Wav2Vec2FeatureExtractor),
|
|
16904
|
+
/* harmony export */ WeSpeakerFeatureExtractor: () => (/* reexport safe */ _wespeaker_feature_extraction_wespeaker_js__WEBPACK_IMPORTED_MODULE_7__.WeSpeakerFeatureExtractor),
|
|
16905
|
+
/* harmony export */ WhisperFeatureExtractor: () => (/* reexport safe */ _whisper_feature_extraction_whisper_js__WEBPACK_IMPORTED_MODULE_8__.WhisperFeatureExtractor)
|
|
16685
16906
|
/* harmony export */ });
|
|
16686
16907
|
/* harmony import */ var _audio_spectrogram_transformer_feature_extraction_audio_spectrogram_transformer_js__WEBPACK_IMPORTED_MODULE_0__ = __webpack_require__(/*! ./audio_spectrogram_transformer/feature_extraction_audio_spectrogram_transformer.js */ "./src/models/audio_spectrogram_transformer/feature_extraction_audio_spectrogram_transformer.js");
|
|
16687
16908
|
/* harmony import */ var _clap_feature_extraction_clap_js__WEBPACK_IMPORTED_MODULE_1__ = __webpack_require__(/*! ./clap/feature_extraction_clap.js */ "./src/models/clap/feature_extraction_clap.js");
|
|
16688
|
-
/* harmony import */ var
|
|
16689
|
-
/* harmony import */ var
|
|
16690
|
-
/* harmony import */ var
|
|
16691
|
-
/* harmony import */ var
|
|
16692
|
-
/* harmony import */ var
|
|
16693
|
-
/* harmony import */ var
|
|
16694
|
-
/* harmony import */ var
|
|
16909
|
+
/* harmony import */ var _moonshine_feature_extraction_moonshine_js__WEBPACK_IMPORTED_MODULE_2__ = __webpack_require__(/*! ./moonshine/feature_extraction_moonshine.js */ "./src/models/moonshine/feature_extraction_moonshine.js");
|
|
16910
|
+
/* harmony import */ var _pyannote_feature_extraction_pyannote_js__WEBPACK_IMPORTED_MODULE_3__ = __webpack_require__(/*! ./pyannote/feature_extraction_pyannote.js */ "./src/models/pyannote/feature_extraction_pyannote.js");
|
|
16911
|
+
/* harmony import */ var _seamless_m4t_feature_extraction_seamless_m4t_js__WEBPACK_IMPORTED_MODULE_4__ = __webpack_require__(/*! ./seamless_m4t/feature_extraction_seamless_m4t.js */ "./src/models/seamless_m4t/feature_extraction_seamless_m4t.js");
|
|
16912
|
+
/* harmony import */ var _speecht5_feature_extraction_speecht5_js__WEBPACK_IMPORTED_MODULE_5__ = __webpack_require__(/*! ./speecht5/feature_extraction_speecht5.js */ "./src/models/speecht5/feature_extraction_speecht5.js");
|
|
16913
|
+
/* harmony import */ var _wav2vec2_feature_extraction_wav2vec2_js__WEBPACK_IMPORTED_MODULE_6__ = __webpack_require__(/*! ./wav2vec2/feature_extraction_wav2vec2.js */ "./src/models/wav2vec2/feature_extraction_wav2vec2.js");
|
|
16914
|
+
/* harmony import */ var _wespeaker_feature_extraction_wespeaker_js__WEBPACK_IMPORTED_MODULE_7__ = __webpack_require__(/*! ./wespeaker/feature_extraction_wespeaker.js */ "./src/models/wespeaker/feature_extraction_wespeaker.js");
|
|
16915
|
+
/* harmony import */ var _whisper_feature_extraction_whisper_js__WEBPACK_IMPORTED_MODULE_8__ = __webpack_require__(/*! ./whisper/feature_extraction_whisper.js */ "./src/models/whisper/feature_extraction_whisper.js");
|
|
16916
|
+
/* harmony import */ var _base_image_processors_utils_js__WEBPACK_IMPORTED_MODULE_9__ = __webpack_require__(/*! ../base/image_processors_utils.js */ "./src/base/image_processors_utils.js");
|
|
16917
|
+
|
|
16695
16918
|
|
|
16696
16919
|
|
|
16697
16920
|
|
|
@@ -17069,18 +17292,29 @@ class Idefics3ImageProcessor extends _base_image_processors_utils_js__WEBPACK_IM
|
|
|
17069
17292
|
const optimal_width = Math.ceil(width / num_splits_w);
|
|
17070
17293
|
|
|
17071
17294
|
// Iterate through each row and column
|
|
17072
|
-
for (let r = 0; r < num_splits_h; r
|
|
17073
|
-
for (let c = 0; c < num_splits_w; c
|
|
17074
|
-
|
|
17075
|
-
|
|
17076
|
-
|
|
17077
|
-
|
|
17078
|
-
|
|
17079
|
-
|
|
17080
|
-
|
|
17081
|
-
|
|
17082
|
-
//
|
|
17083
|
-
|
|
17295
|
+
for (let r = 0; r < num_splits_h; ++r) {
|
|
17296
|
+
for (let c = 0; c < num_splits_w; ++c) {
|
|
17297
|
+
let start_x, start_y, end_x, end_y;
|
|
17298
|
+
if (r === num_splits_h - 1) { // At bottom
|
|
17299
|
+
start_y = height - optimal_height;
|
|
17300
|
+
end_y = height;
|
|
17301
|
+
} else {
|
|
17302
|
+
start_y = r * optimal_height;
|
|
17303
|
+
end_y = (r + 1) * optimal_height;
|
|
17304
|
+
}
|
|
17305
|
+
if (c === num_splits_w - 1) { // At right
|
|
17306
|
+
start_x = width - optimal_width;
|
|
17307
|
+
end_x = width;
|
|
17308
|
+
} else {
|
|
17309
|
+
start_x = c * optimal_width;
|
|
17310
|
+
end_x = (c + 1) * optimal_width;
|
|
17311
|
+
}
|
|
17312
|
+
|
|
17313
|
+
const starts = [start_y, start_x];
|
|
17314
|
+
const ends = [end_y, end_x];
|
|
17315
|
+
|
|
17316
|
+
const patch = await (0,_utils_tensor_js__WEBPACK_IMPORTED_MODULE_1__.slice)(pixel_values, starts, ends, [2, 3]);
|
|
17317
|
+
frames.push(patch);
|
|
17084
17318
|
}
|
|
17085
17319
|
}
|
|
17086
17320
|
|
|
@@ -17304,21 +17538,22 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
17304
17538
|
/* harmony export */ OwlViTFeatureExtractor: () => (/* reexport safe */ _owlvit_image_processing_owlvit_js__WEBPACK_IMPORTED_MODULE_24__.OwlViTFeatureExtractor),
|
|
17305
17539
|
/* harmony export */ OwlViTImageProcessor: () => (/* reexport safe */ _owlvit_image_processing_owlvit_js__WEBPACK_IMPORTED_MODULE_24__.OwlViTImageProcessor),
|
|
17306
17540
|
/* harmony export */ Owlv2ImageProcessor: () => (/* reexport safe */ _owlv2_image_processing_owlv2_js__WEBPACK_IMPORTED_MODULE_23__.Owlv2ImageProcessor),
|
|
17307
|
-
/* harmony export */
|
|
17308
|
-
/* harmony export */
|
|
17309
|
-
/* harmony export */
|
|
17310
|
-
/* harmony export */
|
|
17311
|
-
/* harmony export */
|
|
17312
|
-
/* harmony export */
|
|
17313
|
-
/* harmony export */
|
|
17314
|
-
/* harmony export */
|
|
17541
|
+
/* harmony export */ Phi3VImageProcessor: () => (/* reexport safe */ _phi3_v_image_processing_phi3_v_js__WEBPACK_IMPORTED_MODULE_25__.Phi3VImageProcessor),
|
|
17542
|
+
/* harmony export */ PvtImageProcessor: () => (/* reexport safe */ _pvt_image_processing_pvt_js__WEBPACK_IMPORTED_MODULE_26__.PvtImageProcessor),
|
|
17543
|
+
/* harmony export */ Qwen2VLImageProcessor: () => (/* reexport safe */ _qwen2_vl_image_processing_qwen2_vl_js__WEBPACK_IMPORTED_MODULE_27__.Qwen2VLImageProcessor),
|
|
17544
|
+
/* harmony export */ RTDetrImageProcessor: () => (/* reexport safe */ _rt_detr_image_processing_rt_detr_js__WEBPACK_IMPORTED_MODULE_28__.RTDetrImageProcessor),
|
|
17545
|
+
/* harmony export */ SamImageProcessor: () => (/* reexport safe */ _sam_image_processing_sam_js__WEBPACK_IMPORTED_MODULE_29__.SamImageProcessor),
|
|
17546
|
+
/* harmony export */ SegformerFeatureExtractor: () => (/* reexport safe */ _segformer_image_processing_segformer_js__WEBPACK_IMPORTED_MODULE_30__.SegformerFeatureExtractor),
|
|
17547
|
+
/* harmony export */ SegformerImageProcessor: () => (/* reexport safe */ _segformer_image_processing_segformer_js__WEBPACK_IMPORTED_MODULE_30__.SegformerImageProcessor),
|
|
17548
|
+
/* harmony export */ SiglipImageProcessor: () => (/* reexport safe */ _siglip_image_processing_siglip_js__WEBPACK_IMPORTED_MODULE_31__.SiglipImageProcessor),
|
|
17549
|
+
/* harmony export */ Swin2SRImageProcessor: () => (/* reexport safe */ _swin2sr_image_processing_swin2sr_js__WEBPACK_IMPORTED_MODULE_32__.Swin2SRImageProcessor),
|
|
17315
17550
|
/* harmony export */ VLMImageProcessor: () => (/* reexport safe */ _janus_image_processing_janus_js__WEBPACK_IMPORTED_MODULE_12__.VLMImageProcessor),
|
|
17316
|
-
/* harmony export */ ViTFeatureExtractor: () => (/* reexport safe */
|
|
17317
|
-
/* harmony export */ ViTImageProcessor: () => (/* reexport safe */
|
|
17318
|
-
/* harmony export */ VitMatteImageProcessor: () => (/* reexport safe */
|
|
17319
|
-
/* harmony export */ VitPoseImageProcessor: () => (/* reexport safe */
|
|
17320
|
-
/* harmony export */ YolosFeatureExtractor: () => (/* reexport safe */
|
|
17321
|
-
/* harmony export */ YolosImageProcessor: () => (/* reexport safe */
|
|
17551
|
+
/* harmony export */ ViTFeatureExtractor: () => (/* reexport safe */ _vit_image_processing_vit_js__WEBPACK_IMPORTED_MODULE_33__.ViTFeatureExtractor),
|
|
17552
|
+
/* harmony export */ ViTImageProcessor: () => (/* reexport safe */ _vit_image_processing_vit_js__WEBPACK_IMPORTED_MODULE_33__.ViTImageProcessor),
|
|
17553
|
+
/* harmony export */ VitMatteImageProcessor: () => (/* reexport safe */ _vitmatte_image_processing_vitmatte_js__WEBPACK_IMPORTED_MODULE_34__.VitMatteImageProcessor),
|
|
17554
|
+
/* harmony export */ VitPoseImageProcessor: () => (/* reexport safe */ _vitpose_image_processing_vitpose_js__WEBPACK_IMPORTED_MODULE_35__.VitPoseImageProcessor),
|
|
17555
|
+
/* harmony export */ YolosFeatureExtractor: () => (/* reexport safe */ _yolos_image_processing_yolos_js__WEBPACK_IMPORTED_MODULE_36__.YolosFeatureExtractor),
|
|
17556
|
+
/* harmony export */ YolosImageProcessor: () => (/* reexport safe */ _yolos_image_processing_yolos_js__WEBPACK_IMPORTED_MODULE_36__.YolosImageProcessor)
|
|
17322
17557
|
/* harmony export */ });
|
|
17323
17558
|
/* harmony import */ var _beit_image_processing_beit_js__WEBPACK_IMPORTED_MODULE_0__ = __webpack_require__(/*! ./beit/image_processing_beit.js */ "./src/models/beit/image_processing_beit.js");
|
|
17324
17559
|
/* harmony import */ var _bit_image_processing_bit_js__WEBPACK_IMPORTED_MODULE_1__ = __webpack_require__(/*! ./bit/image_processing_bit.js */ "./src/models/bit/image_processing_bit.js");
|
|
@@ -17345,17 +17580,19 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
17345
17580
|
/* harmony import */ var _nougat_image_processing_nougat_js__WEBPACK_IMPORTED_MODULE_22__ = __webpack_require__(/*! ./nougat/image_processing_nougat.js */ "./src/models/nougat/image_processing_nougat.js");
|
|
17346
17581
|
/* harmony import */ var _owlv2_image_processing_owlv2_js__WEBPACK_IMPORTED_MODULE_23__ = __webpack_require__(/*! ./owlv2/image_processing_owlv2.js */ "./src/models/owlv2/image_processing_owlv2.js");
|
|
17347
17582
|
/* harmony import */ var _owlvit_image_processing_owlvit_js__WEBPACK_IMPORTED_MODULE_24__ = __webpack_require__(/*! ./owlvit/image_processing_owlvit.js */ "./src/models/owlvit/image_processing_owlvit.js");
|
|
17348
|
-
/* harmony import */ var
|
|
17349
|
-
/* harmony import */ var
|
|
17350
|
-
/* harmony import */ var
|
|
17351
|
-
/* harmony import */ var
|
|
17352
|
-
/* harmony import */ var
|
|
17353
|
-
/* harmony import */ var
|
|
17354
|
-
/* harmony import */ var
|
|
17355
|
-
/* harmony import */ var
|
|
17356
|
-
/* harmony import */ var
|
|
17357
|
-
/* harmony import */ var
|
|
17358
|
-
/* harmony import */ var
|
|
17583
|
+
/* harmony import */ var _phi3_v_image_processing_phi3_v_js__WEBPACK_IMPORTED_MODULE_25__ = __webpack_require__(/*! ./phi3_v/image_processing_phi3_v.js */ "./src/models/phi3_v/image_processing_phi3_v.js");
|
|
17584
|
+
/* harmony import */ var _pvt_image_processing_pvt_js__WEBPACK_IMPORTED_MODULE_26__ = __webpack_require__(/*! ./pvt/image_processing_pvt.js */ "./src/models/pvt/image_processing_pvt.js");
|
|
17585
|
+
/* harmony import */ var _qwen2_vl_image_processing_qwen2_vl_js__WEBPACK_IMPORTED_MODULE_27__ = __webpack_require__(/*! ./qwen2_vl/image_processing_qwen2_vl.js */ "./src/models/qwen2_vl/image_processing_qwen2_vl.js");
|
|
17586
|
+
/* harmony import */ var _rt_detr_image_processing_rt_detr_js__WEBPACK_IMPORTED_MODULE_28__ = __webpack_require__(/*! ./rt_detr/image_processing_rt_detr.js */ "./src/models/rt_detr/image_processing_rt_detr.js");
|
|
17587
|
+
/* harmony import */ var _sam_image_processing_sam_js__WEBPACK_IMPORTED_MODULE_29__ = __webpack_require__(/*! ./sam/image_processing_sam.js */ "./src/models/sam/image_processing_sam.js");
|
|
17588
|
+
/* harmony import */ var _segformer_image_processing_segformer_js__WEBPACK_IMPORTED_MODULE_30__ = __webpack_require__(/*! ./segformer/image_processing_segformer.js */ "./src/models/segformer/image_processing_segformer.js");
|
|
17589
|
+
/* harmony import */ var _siglip_image_processing_siglip_js__WEBPACK_IMPORTED_MODULE_31__ = __webpack_require__(/*! ./siglip/image_processing_siglip.js */ "./src/models/siglip/image_processing_siglip.js");
|
|
17590
|
+
/* harmony import */ var _swin2sr_image_processing_swin2sr_js__WEBPACK_IMPORTED_MODULE_32__ = __webpack_require__(/*! ./swin2sr/image_processing_swin2sr.js */ "./src/models/swin2sr/image_processing_swin2sr.js");
|
|
17591
|
+
/* harmony import */ var _vit_image_processing_vit_js__WEBPACK_IMPORTED_MODULE_33__ = __webpack_require__(/*! ./vit/image_processing_vit.js */ "./src/models/vit/image_processing_vit.js");
|
|
17592
|
+
/* harmony import */ var _vitmatte_image_processing_vitmatte_js__WEBPACK_IMPORTED_MODULE_34__ = __webpack_require__(/*! ./vitmatte/image_processing_vitmatte.js */ "./src/models/vitmatte/image_processing_vitmatte.js");
|
|
17593
|
+
/* harmony import */ var _vitpose_image_processing_vitpose_js__WEBPACK_IMPORTED_MODULE_35__ = __webpack_require__(/*! ./vitpose/image_processing_vitpose.js */ "./src/models/vitpose/image_processing_vitpose.js");
|
|
17594
|
+
/* harmony import */ var _yolos_image_processing_yolos_js__WEBPACK_IMPORTED_MODULE_36__ = __webpack_require__(/*! ./yolos/image_processing_yolos.js */ "./src/models/yolos/image_processing_yolos.js");
|
|
17595
|
+
|
|
17359
17596
|
|
|
17360
17597
|
|
|
17361
17598
|
|
|
@@ -18017,6 +18254,85 @@ class MobileViTImageProcessor extends _base_image_processors_utils_js__WEBPACK_I
|
|
|
18017
18254
|
class MobileViTFeatureExtractor extends MobileViTImageProcessor { }
|
|
18018
18255
|
|
|
18019
18256
|
|
|
18257
|
+
/***/ }),
|
|
18258
|
+
|
|
18259
|
+
/***/ "./src/models/moonshine/feature_extraction_moonshine.js":
|
|
18260
|
+
/*!**************************************************************!*\
|
|
18261
|
+
!*** ./src/models/moonshine/feature_extraction_moonshine.js ***!
|
|
18262
|
+
\**************************************************************/
|
|
18263
|
+
/***/ ((__unused_webpack___webpack_module__, __webpack_exports__, __webpack_require__) => {
|
|
18264
|
+
|
|
18265
|
+
__webpack_require__.r(__webpack_exports__);
|
|
18266
|
+
/* harmony export */ __webpack_require__.d(__webpack_exports__, {
|
|
18267
|
+
/* harmony export */ MoonshineFeatureExtractor: () => (/* binding */ MoonshineFeatureExtractor)
|
|
18268
|
+
/* harmony export */ });
|
|
18269
|
+
/* harmony import */ var _base_feature_extraction_utils_js__WEBPACK_IMPORTED_MODULE_0__ = __webpack_require__(/*! ../../base/feature_extraction_utils.js */ "./src/base/feature_extraction_utils.js");
|
|
18270
|
+
/* harmony import */ var _utils_tensor_js__WEBPACK_IMPORTED_MODULE_1__ = __webpack_require__(/*! ../../utils/tensor.js */ "./src/utils/tensor.js");
|
|
18271
|
+
|
|
18272
|
+
|
|
18273
|
+
|
|
18274
|
+
|
|
18275
|
+
class MoonshineFeatureExtractor extends _base_feature_extraction_utils_js__WEBPACK_IMPORTED_MODULE_0__.FeatureExtractor {
|
|
18276
|
+
/**
|
|
18277
|
+
* Asynchronously extracts input values from a given audio using the provided configuration.
|
|
18278
|
+
* @param {Float32Array|Float64Array} audio The audio data as a Float32Array/Float64Array.
|
|
18279
|
+
* @returns {Promise<{ input_values: Tensor; }>} The extracted input values.
|
|
18280
|
+
*/
|
|
18281
|
+
async _call(audio) {
|
|
18282
|
+
(0,_base_feature_extraction_utils_js__WEBPACK_IMPORTED_MODULE_0__.validate_audio_inputs)(audio, 'MoonshineFeatureExtractor');
|
|
18283
|
+
|
|
18284
|
+
if (audio instanceof Float64Array) {
|
|
18285
|
+
audio = new Float32Array(audio);
|
|
18286
|
+
}
|
|
18287
|
+
|
|
18288
|
+
const shape = [
|
|
18289
|
+
1, /* batch_size */
|
|
18290
|
+
audio.length, /* num_samples */
|
|
18291
|
+
];
|
|
18292
|
+
return {
|
|
18293
|
+
input_values: new _utils_tensor_js__WEBPACK_IMPORTED_MODULE_1__.Tensor('float32', audio, shape),
|
|
18294
|
+
};
|
|
18295
|
+
}
|
|
18296
|
+
}
|
|
18297
|
+
|
|
18298
|
+
|
|
18299
|
+
/***/ }),
|
|
18300
|
+
|
|
18301
|
+
/***/ "./src/models/moonshine/processing_moonshine.js":
|
|
18302
|
+
/*!******************************************************!*\
|
|
18303
|
+
!*** ./src/models/moonshine/processing_moonshine.js ***!
|
|
18304
|
+
\******************************************************/
|
|
18305
|
+
/***/ ((__unused_webpack___webpack_module__, __webpack_exports__, __webpack_require__) => {
|
|
18306
|
+
|
|
18307
|
+
__webpack_require__.r(__webpack_exports__);
|
|
18308
|
+
/* harmony export */ __webpack_require__.d(__webpack_exports__, {
|
|
18309
|
+
/* harmony export */ MoonshineProcessor: () => (/* binding */ MoonshineProcessor)
|
|
18310
|
+
/* harmony export */ });
|
|
18311
|
+
/* harmony import */ var _auto_feature_extraction_auto_js__WEBPACK_IMPORTED_MODULE_0__ = __webpack_require__(/*! ../auto/feature_extraction_auto.js */ "./src/models/auto/feature_extraction_auto.js");
|
|
18312
|
+
/* harmony import */ var _tokenizers_js__WEBPACK_IMPORTED_MODULE_1__ = __webpack_require__(/*! ../../tokenizers.js */ "./src/tokenizers.js");
|
|
18313
|
+
/* harmony import */ var _base_processing_utils_js__WEBPACK_IMPORTED_MODULE_2__ = __webpack_require__(/*! ../../base/processing_utils.js */ "./src/base/processing_utils.js");
|
|
18314
|
+
|
|
18315
|
+
|
|
18316
|
+
|
|
18317
|
+
|
|
18318
|
+
/**
|
|
18319
|
+
* Represents a MoonshineProcessor that extracts features from an audio input.
|
|
18320
|
+
*/
|
|
18321
|
+
class MoonshineProcessor extends _base_processing_utils_js__WEBPACK_IMPORTED_MODULE_2__.Processor {
|
|
18322
|
+
static tokenizer_class = _tokenizers_js__WEBPACK_IMPORTED_MODULE_1__.AutoTokenizer
|
|
18323
|
+
static feature_extractor_class = _auto_feature_extraction_auto_js__WEBPACK_IMPORTED_MODULE_0__.AutoFeatureExtractor
|
|
18324
|
+
|
|
18325
|
+
/**
|
|
18326
|
+
* Calls the feature_extractor function with the given audio input.
|
|
18327
|
+
* @param {any} audio The audio input to extract features from.
|
|
18328
|
+
* @returns {Promise<any>} A Promise that resolves with the extracted features.
|
|
18329
|
+
*/
|
|
18330
|
+
async _call(audio) {
|
|
18331
|
+
return await this.feature_extractor(audio);
|
|
18332
|
+
}
|
|
18333
|
+
}
|
|
18334
|
+
|
|
18335
|
+
|
|
18020
18336
|
/***/ }),
|
|
18021
18337
|
|
|
18022
18338
|
/***/ "./src/models/nougat/image_processing_nougat.js":
|
|
@@ -18106,6 +18422,353 @@ class OwlViTProcessor extends _base_processing_utils_js__WEBPACK_IMPORTED_MODULE
|
|
|
18106
18422
|
}
|
|
18107
18423
|
|
|
18108
18424
|
|
|
18425
|
+
/***/ }),
|
|
18426
|
+
|
|
18427
|
+
/***/ "./src/models/paligemma/processing_paligemma.js":
|
|
18428
|
+
/*!******************************************************!*\
|
|
18429
|
+
!*** ./src/models/paligemma/processing_paligemma.js ***!
|
|
18430
|
+
\******************************************************/
|
|
18431
|
+
/***/ ((__unused_webpack___webpack_module__, __webpack_exports__, __webpack_require__) => {
|
|
18432
|
+
|
|
18433
|
+
__webpack_require__.r(__webpack_exports__);
|
|
18434
|
+
/* harmony export */ __webpack_require__.d(__webpack_exports__, {
|
|
18435
|
+
/* harmony export */ PaliGemmaProcessor: () => (/* binding */ PaliGemmaProcessor)
|
|
18436
|
+
/* harmony export */ });
|
|
18437
|
+
/* harmony import */ var _base_processing_utils_js__WEBPACK_IMPORTED_MODULE_0__ = __webpack_require__(/*! ../../base/processing_utils.js */ "./src/base/processing_utils.js");
|
|
18438
|
+
/* harmony import */ var _auto_image_processing_auto_js__WEBPACK_IMPORTED_MODULE_1__ = __webpack_require__(/*! ../auto/image_processing_auto.js */ "./src/models/auto/image_processing_auto.js");
|
|
18439
|
+
/* harmony import */ var _tokenizers_js__WEBPACK_IMPORTED_MODULE_2__ = __webpack_require__(/*! ../../tokenizers.js */ "./src/tokenizers.js");
|
|
18440
|
+
|
|
18441
|
+
|
|
18442
|
+
|
|
18443
|
+
|
|
18444
|
+
const IMAGE_TOKEN = "<image>";
|
|
18445
|
+
|
|
18446
|
+
function build_string_from_input(
|
|
18447
|
+
prompt,
|
|
18448
|
+
bos_token,
|
|
18449
|
+
image_seq_len,
|
|
18450
|
+
image_token,
|
|
18451
|
+
num_images,
|
|
18452
|
+
) {
|
|
18453
|
+
return `${image_token.repeat(image_seq_len * num_images)}${bos_token}${prompt}\n`
|
|
18454
|
+
}
|
|
18455
|
+
|
|
18456
|
+
class PaliGemmaProcessor extends _base_processing_utils_js__WEBPACK_IMPORTED_MODULE_0__.Processor {
|
|
18457
|
+
static tokenizer_class = _tokenizers_js__WEBPACK_IMPORTED_MODULE_2__.AutoTokenizer
|
|
18458
|
+
static image_processor_class = _auto_image_processing_auto_js__WEBPACK_IMPORTED_MODULE_1__.AutoImageProcessor
|
|
18459
|
+
static uses_processor_config = false;
|
|
18460
|
+
|
|
18461
|
+
/**
|
|
18462
|
+
* @typedef {import('../../utils/image.js').RawImage} RawImage
|
|
18463
|
+
*/
|
|
18464
|
+
|
|
18465
|
+
// `images` is required, `text` is optional
|
|
18466
|
+
async _call(/** @type {RawImage|RawImage[]} */ images, text = null, kwargs = {}) {
|
|
18467
|
+
if (!text) {
|
|
18468
|
+
console.warn(
|
|
18469
|
+
"You are using PaliGemma without a text prefix. It will perform as a picture-captioning model."
|
|
18470
|
+
)
|
|
18471
|
+
text = ""
|
|
18472
|
+
}
|
|
18473
|
+
|
|
18474
|
+
if (!Array.isArray(images)) {
|
|
18475
|
+
images = [images]
|
|
18476
|
+
}
|
|
18477
|
+
|
|
18478
|
+
if (!Array.isArray(text)) {
|
|
18479
|
+
text = [text]
|
|
18480
|
+
}
|
|
18481
|
+
|
|
18482
|
+
const bos_token = this.tokenizer.bos_token;
|
|
18483
|
+
const image_seq_length = this.image_processor.config.image_seq_length;
|
|
18484
|
+
let input_strings;
|
|
18485
|
+
if (text.some((t) => t.includes(IMAGE_TOKEN))) {
|
|
18486
|
+
input_strings = text.map(
|
|
18487
|
+
sample => {
|
|
18488
|
+
const expanded_sample = sample.replaceAll(IMAGE_TOKEN, IMAGE_TOKEN.repeat(image_seq_length));
|
|
18489
|
+
const bos_rfind_index = expanded_sample.lastIndexOf(IMAGE_TOKEN);
|
|
18490
|
+
const bos_index = bos_rfind_index === -1 ? 0 : bos_rfind_index + IMAGE_TOKEN.length;
|
|
18491
|
+
return expanded_sample.slice(0, bos_index) + bos_token + expanded_sample.slice(bos_index) + "\n";
|
|
18492
|
+
}
|
|
18493
|
+
)
|
|
18494
|
+
} else {
|
|
18495
|
+
console.warn(
|
|
18496
|
+
"You are passing both `text` and `images` to `PaliGemmaProcessor`. The processor expects special " +
|
|
18497
|
+
"image tokens in the text, as many tokens as there are images per each text. It is recommended to " +
|
|
18498
|
+
"add `<image>` tokens in the very beginning of your text. For this call, we will infer how many images " +
|
|
18499
|
+
"each text has and add special tokens."
|
|
18500
|
+
)
|
|
18501
|
+
|
|
18502
|
+
input_strings = text.map(
|
|
18503
|
+
sample => build_string_from_input(
|
|
18504
|
+
sample,
|
|
18505
|
+
bos_token,
|
|
18506
|
+
image_seq_length,
|
|
18507
|
+
IMAGE_TOKEN,
|
|
18508
|
+
images.length,
|
|
18509
|
+
)
|
|
18510
|
+
)
|
|
18511
|
+
}
|
|
18512
|
+
|
|
18513
|
+
const text_inputs = this.tokenizer(input_strings, kwargs);
|
|
18514
|
+
const image_inputs = await this.image_processor(images, kwargs);
|
|
18515
|
+
|
|
18516
|
+
return {
|
|
18517
|
+
...image_inputs,
|
|
18518
|
+
...text_inputs,
|
|
18519
|
+
}
|
|
18520
|
+
}
|
|
18521
|
+
}
|
|
18522
|
+
|
|
18523
|
+
|
|
18524
|
+
/***/ }),
|
|
18525
|
+
|
|
18526
|
+
/***/ "./src/models/phi3_v/image_processing_phi3_v.js":
|
|
18527
|
+
/*!******************************************************!*\
|
|
18528
|
+
!*** ./src/models/phi3_v/image_processing_phi3_v.js ***!
|
|
18529
|
+
\******************************************************/
|
|
18530
|
+
/***/ ((__unused_webpack___webpack_module__, __webpack_exports__, __webpack_require__) => {
|
|
18531
|
+
|
|
18532
|
+
__webpack_require__.r(__webpack_exports__);
|
|
18533
|
+
/* harmony export */ __webpack_require__.d(__webpack_exports__, {
|
|
18534
|
+
/* harmony export */ Phi3VImageProcessor: () => (/* binding */ Phi3VImageProcessor)
|
|
18535
|
+
/* harmony export */ });
|
|
18536
|
+
/* harmony import */ var _base_image_processors_utils_js__WEBPACK_IMPORTED_MODULE_0__ = __webpack_require__(/*! ../../base/image_processors_utils.js */ "./src/base/image_processors_utils.js");
|
|
18537
|
+
/* harmony import */ var _utils_tensor_js__WEBPACK_IMPORTED_MODULE_1__ = __webpack_require__(/*! ../../utils/tensor.js */ "./src/utils/tensor.js");
|
|
18538
|
+
|
|
18539
|
+
|
|
18540
|
+
|
|
18541
|
+
const IMAGE_SIZE = 336;
|
|
18542
|
+
const SLICE_AXES = [2, 3]; // axes to slice on
|
|
18543
|
+
const { ceil, floor, sqrt } = Math;
|
|
18544
|
+
|
|
18545
|
+
class Phi3VImageProcessor extends _base_image_processors_utils_js__WEBPACK_IMPORTED_MODULE_0__.ImageProcessor {
|
|
18546
|
+
constructor(config) {
|
|
18547
|
+
super({
|
|
18548
|
+
...config,
|
|
18549
|
+
do_normalize: true,
|
|
18550
|
+
do_pad: true,
|
|
18551
|
+
pad_size: 'custom',
|
|
18552
|
+
do_convert_rgb: true,
|
|
18553
|
+
do_resize: true, // Smart resizing "hd_transform"
|
|
18554
|
+
});
|
|
18555
|
+
|
|
18556
|
+
this._num_crops = config.num_crops;
|
|
18557
|
+
}
|
|
18558
|
+
calc_num_image_tokens_from_image_size(width, height) {
|
|
18559
|
+
// @ts-expect-error
|
|
18560
|
+
const { num_img_tokens } = this.config;
|
|
18561
|
+
return floor(((floor((height / IMAGE_SIZE)) * floor((width / IMAGE_SIZE)) + 1) * num_img_tokens) + 1 + (floor(height / IMAGE_SIZE) + 1) * sqrt(num_img_tokens));
|
|
18562
|
+
}
|
|
18563
|
+
|
|
18564
|
+
/** @type {ImageProcessor['get_resize_output_image_size']} */
|
|
18565
|
+
get_resize_output_image_size(image, size) {
|
|
18566
|
+
const hd_num = this._num_crops;
|
|
18567
|
+
const [width, height] = image.size
|
|
18568
|
+
|
|
18569
|
+
let ratio = width / height;
|
|
18570
|
+
let scale = 1;
|
|
18571
|
+
|
|
18572
|
+
// Calculate the scaling factor
|
|
18573
|
+
while (scale * Math.ceil(scale / ratio) <= hd_num) {
|
|
18574
|
+
scale += 1;
|
|
18575
|
+
}
|
|
18576
|
+
scale -= 1;
|
|
18577
|
+
|
|
18578
|
+
// Compute the new dimensions
|
|
18579
|
+
const new_w = Math.floor(scale * 336);
|
|
18580
|
+
const new_h = Math.floor(new_w / ratio);
|
|
18581
|
+
|
|
18582
|
+
return [new_w, new_h]
|
|
18583
|
+
}
|
|
18584
|
+
|
|
18585
|
+
|
|
18586
|
+
/** @type {ImageProcessor['pad_image']} */
|
|
18587
|
+
pad_image(pixelData, imgDims, padSize, options = {}) {
|
|
18588
|
+
// Phi3V uses a custom padding strategy:
|
|
18589
|
+
// - Pad to a multiple of 336
|
|
18590
|
+
// - Pad with white pixels
|
|
18591
|
+
const [imageHeight, imageWidth] = imgDims;
|
|
18592
|
+
const height = IMAGE_SIZE * ceil(imageHeight / IMAGE_SIZE);
|
|
18593
|
+
const width = IMAGE_SIZE * ceil(imageWidth / IMAGE_SIZE);
|
|
18594
|
+
|
|
18595
|
+
// NOTE: Since padding is done after normalization, we need to fill with the normalized values
|
|
18596
|
+
const constant_values = [1, 1, 1].map((x, i) => (x - this.image_mean[i]) / this.image_std[i]);
|
|
18597
|
+
return super.pad_image(pixelData, imgDims, { width, height }, {
|
|
18598
|
+
center: true,
|
|
18599
|
+
constant_values,
|
|
18600
|
+
...options,
|
|
18601
|
+
});
|
|
18602
|
+
}
|
|
18603
|
+
|
|
18604
|
+
async _call(images, {
|
|
18605
|
+
num_crops = null,
|
|
18606
|
+
} = {}) {
|
|
18607
|
+
// @ts-expect-error
|
|
18608
|
+
this._num_crops = num_crops ??= this.config.num_crops;
|
|
18609
|
+
if (num_crops < 4 || sqrt(num_crops) % 1 !== 0) {
|
|
18610
|
+
throw new Error("num_crops must be a square number >= 4");
|
|
18611
|
+
}
|
|
18612
|
+
|
|
18613
|
+
if (!Array.isArray(images)) {
|
|
18614
|
+
images = [images];
|
|
18615
|
+
}
|
|
18616
|
+
|
|
18617
|
+
const num_images = images.length;
|
|
18618
|
+
const imageData = await Promise.all(images.map(x => this.preprocess(x)));
|
|
18619
|
+
|
|
18620
|
+
const original_sizes = imageData.map(x => x.original_size);
|
|
18621
|
+
const reshaped_input_sizes = imageData.map(x => x.reshaped_input_size);
|
|
18622
|
+
|
|
18623
|
+
// Process each image in batch
|
|
18624
|
+
const all_pixel_values = [];
|
|
18625
|
+
for (const { pixel_values } of imageData) {
|
|
18626
|
+
pixel_values.unsqueeze_(0); // Easier processing as 4D tensor
|
|
18627
|
+
|
|
18628
|
+
const [height, width] = pixel_values.dims.slice(-2);
|
|
18629
|
+
|
|
18630
|
+
// Global image (Tensor of shape [num_channels, height, width])
|
|
18631
|
+
const batch_pixel_values = await (0,_utils_tensor_js__WEBPACK_IMPORTED_MODULE_1__.interpolate_4d)(pixel_values, {
|
|
18632
|
+
size: [IMAGE_SIZE, IMAGE_SIZE],
|
|
18633
|
+
mode: 'bicubic',
|
|
18634
|
+
});
|
|
18635
|
+
|
|
18636
|
+
if (num_crops > 0) {
|
|
18637
|
+
const patches = [];
|
|
18638
|
+
const sqrt_patches = sqrt(num_crops);
|
|
18639
|
+
const patch_width = floor(width / sqrt_patches);
|
|
18640
|
+
const patch_height = floor(height / sqrt_patches);
|
|
18641
|
+
for (let y = 0; y < sqrt_patches; ++y) {
|
|
18642
|
+
for (let x = 0; x < sqrt_patches; ++x) {
|
|
18643
|
+
let start_x, start_y, end_x, end_y;
|
|
18644
|
+
if (y === sqrt_patches - 1) { // At bottom
|
|
18645
|
+
start_y = height - patch_height;
|
|
18646
|
+
end_y = height;
|
|
18647
|
+
} else {
|
|
18648
|
+
start_y = y * patch_height;
|
|
18649
|
+
end_y = (y + 1) * patch_height;
|
|
18650
|
+
}
|
|
18651
|
+
if (x === sqrt_patches - 1) { // At right
|
|
18652
|
+
start_x = width - patch_width;
|
|
18653
|
+
end_x = width;
|
|
18654
|
+
} else {
|
|
18655
|
+
start_x = x * patch_width;
|
|
18656
|
+
end_x = (x + 1) * patch_width;
|
|
18657
|
+
}
|
|
18658
|
+
|
|
18659
|
+
const starts = [start_y, start_x];
|
|
18660
|
+
const ends = [end_y, end_x];
|
|
18661
|
+
const patch = await (0,_utils_tensor_js__WEBPACK_IMPORTED_MODULE_1__.slice)(pixel_values, starts, ends, SLICE_AXES);
|
|
18662
|
+
patches.push(patch);
|
|
18663
|
+
}
|
|
18664
|
+
}
|
|
18665
|
+
|
|
18666
|
+
const resized_tensors = await (0,_utils_tensor_js__WEBPACK_IMPORTED_MODULE_1__.interpolate_4d)((0,_utils_tensor_js__WEBPACK_IMPORTED_MODULE_1__.cat)(patches, 0), {
|
|
18667
|
+
size: [IMAGE_SIZE, IMAGE_SIZE],
|
|
18668
|
+
mode: 'bicubic',
|
|
18669
|
+
}); // [num_crops, 3, 336, 336]
|
|
18670
|
+
|
|
18671
|
+
// Concatenate the global image with the patches
|
|
18672
|
+
all_pixel_values.push((0,_utils_tensor_js__WEBPACK_IMPORTED_MODULE_1__.cat)([batch_pixel_values, resized_tensors], 0));
|
|
18673
|
+
} else {
|
|
18674
|
+
// Only use the global image
|
|
18675
|
+
// NOTE: Not currently supported in modelling code
|
|
18676
|
+
all_pixel_values.push(batch_pixel_values);
|
|
18677
|
+
}
|
|
18678
|
+
}
|
|
18679
|
+
|
|
18680
|
+
// [num_images, 1 + num_crops, num_channels=3, height, width]
|
|
18681
|
+
const pixel_values = (0,_utils_tensor_js__WEBPACK_IMPORTED_MODULE_1__.stack)(all_pixel_values, 0);
|
|
18682
|
+
|
|
18683
|
+
// Calculate padded image sizes
|
|
18684
|
+
const sizes = reshaped_input_sizes.map(x => x.map(y => IMAGE_SIZE * ceil(y / IMAGE_SIZE)));
|
|
18685
|
+
|
|
18686
|
+
const image_sizes = new _utils_tensor_js__WEBPACK_IMPORTED_MODULE_1__.Tensor(
|
|
18687
|
+
'int64',
|
|
18688
|
+
sizes.flat(),
|
|
18689
|
+
[num_images, 2],
|
|
18690
|
+
);
|
|
18691
|
+
|
|
18692
|
+
const num_img_tokens = sizes.map(
|
|
18693
|
+
([height, width]) => this.calc_num_image_tokens_from_image_size(width, height),
|
|
18694
|
+
);
|
|
18695
|
+
|
|
18696
|
+
return { pixel_values, original_sizes, reshaped_input_sizes, image_sizes, num_img_tokens };
|
|
18697
|
+
}
|
|
18698
|
+
}
|
|
18699
|
+
|
|
18700
|
+
|
|
18701
|
+
/***/ }),
|
|
18702
|
+
|
|
18703
|
+
/***/ "./src/models/phi3_v/processing_phi3_v.js":
|
|
18704
|
+
/*!************************************************!*\
|
|
18705
|
+
!*** ./src/models/phi3_v/processing_phi3_v.js ***!
|
|
18706
|
+
\************************************************/
|
|
18707
|
+
/***/ ((__unused_webpack___webpack_module__, __webpack_exports__, __webpack_require__) => {
|
|
18708
|
+
|
|
18709
|
+
__webpack_require__.r(__webpack_exports__);
|
|
18710
|
+
/* harmony export */ __webpack_require__.d(__webpack_exports__, {
|
|
18711
|
+
/* harmony export */ Phi3VProcessor: () => (/* binding */ Phi3VProcessor)
|
|
18712
|
+
/* harmony export */ });
|
|
18713
|
+
/* harmony import */ var _base_processing_utils_js__WEBPACK_IMPORTED_MODULE_0__ = __webpack_require__(/*! ../../base/processing_utils.js */ "./src/base/processing_utils.js");
|
|
18714
|
+
/* harmony import */ var _auto_image_processing_auto_js__WEBPACK_IMPORTED_MODULE_1__ = __webpack_require__(/*! ../auto/image_processing_auto.js */ "./src/models/auto/image_processing_auto.js");
|
|
18715
|
+
/* harmony import */ var _tokenizers_js__WEBPACK_IMPORTED_MODULE_2__ = __webpack_require__(/*! ../../tokenizers.js */ "./src/tokenizers.js");
|
|
18716
|
+
/* harmony import */ var _utils_image_js__WEBPACK_IMPORTED_MODULE_3__ = __webpack_require__(/*! ../../utils/image.js */ "./src/utils/image.js");
|
|
18717
|
+
|
|
18718
|
+
|
|
18719
|
+
|
|
18720
|
+
|
|
18721
|
+
|
|
18722
|
+
const IMAGE_TOKEN = "<|image|>";
|
|
18723
|
+
const IMAGE_TOKEN_PATTERN = /<\|image_\d+\|>/g;
|
|
18724
|
+
|
|
18725
|
+
class Phi3VProcessor extends _base_processing_utils_js__WEBPACK_IMPORTED_MODULE_0__.Processor {
|
|
18726
|
+
static image_processor_class = _auto_image_processing_auto_js__WEBPACK_IMPORTED_MODULE_1__.AutoImageProcessor
|
|
18727
|
+
static tokenizer_class = _tokenizers_js__WEBPACK_IMPORTED_MODULE_2__.AutoTokenizer
|
|
18728
|
+
|
|
18729
|
+
/**
|
|
18730
|
+
*
|
|
18731
|
+
* @param {string|string[]} text
|
|
18732
|
+
* @param {RawImage|RawImage[]} images
|
|
18733
|
+
* @param {...any} args
|
|
18734
|
+
* @returns {Promise<any>}
|
|
18735
|
+
*/
|
|
18736
|
+
async _call(text, images = null, {
|
|
18737
|
+
padding = true,
|
|
18738
|
+
truncation = true,
|
|
18739
|
+
num_crops = null,
|
|
18740
|
+
} = {}) {
|
|
18741
|
+
|
|
18742
|
+
if (!Array.isArray(text)) {
|
|
18743
|
+
text = [text];
|
|
18744
|
+
}
|
|
18745
|
+
|
|
18746
|
+
let text_inputs, image_inputs;
|
|
18747
|
+
if (images) {
|
|
18748
|
+
image_inputs = await this.image_processor(images, { num_crops });
|
|
18749
|
+
const { num_img_tokens } = image_inputs;
|
|
18750
|
+
|
|
18751
|
+
// The original implementation adds a bos_token before the image tokens
|
|
18752
|
+
// TODO: Check if this affects performance, since it looks like a bug in the original implementation
|
|
18753
|
+
const prompt_chunks = text.map((t, i) => t.split(IMAGE_TOKEN_PATTERN).join(IMAGE_TOKEN.repeat(num_img_tokens[i])));
|
|
18754
|
+
|
|
18755
|
+
text_inputs = this.tokenizer(prompt_chunks, { padding, truncation });
|
|
18756
|
+
|
|
18757
|
+
// The model expects image tokens to be negative, so we negate the image token ids
|
|
18758
|
+
const image_token_id = this.tokenizer.model.convert_tokens_to_ids([IMAGE_TOKEN])[0];
|
|
18759
|
+
text_inputs.input_ids.map_(id => (id == image_token_id) ? -id : id);
|
|
18760
|
+
} else {
|
|
18761
|
+
text_inputs = this.tokenizer(text);
|
|
18762
|
+
}
|
|
18763
|
+
|
|
18764
|
+
return {
|
|
18765
|
+
...text_inputs,
|
|
18766
|
+
...image_inputs,
|
|
18767
|
+
}
|
|
18768
|
+
}
|
|
18769
|
+
}
|
|
18770
|
+
|
|
18771
|
+
|
|
18109
18772
|
/***/ }),
|
|
18110
18773
|
|
|
18111
18774
|
/***/ "./src/models/processors.js":
|
|
@@ -18117,30 +18780,39 @@ class OwlViTProcessor extends _base_processing_utils_js__WEBPACK_IMPORTED_MODULE
|
|
|
18117
18780
|
__webpack_require__.r(__webpack_exports__);
|
|
18118
18781
|
/* harmony export */ __webpack_require__.d(__webpack_exports__, {
|
|
18119
18782
|
/* harmony export */ Florence2Processor: () => (/* reexport safe */ _florence2_processing_florence2_js__WEBPACK_IMPORTED_MODULE_0__.Florence2Processor),
|
|
18120
|
-
/* harmony export */ Idefics3Processor: () => (/* reexport safe */
|
|
18121
|
-
/* harmony export */ JinaCLIPProcessor: () => (/* reexport safe */
|
|
18783
|
+
/* harmony export */ Idefics3Processor: () => (/* reexport safe */ _idefics3_processing_idefics3_js__WEBPACK_IMPORTED_MODULE_3__.Idefics3Processor),
|
|
18784
|
+
/* harmony export */ JinaCLIPProcessor: () => (/* reexport safe */ _jina_clip_processing_jina_clip_js__WEBPACK_IMPORTED_MODULE_5__.JinaCLIPProcessor),
|
|
18122
18785
|
/* harmony export */ MgpstrProcessor: () => (/* reexport safe */ _mgp_str_processing_mgp_str_js__WEBPACK_IMPORTED_MODULE_1__.MgpstrProcessor),
|
|
18123
|
-
/* harmony export */
|
|
18124
|
-
/* harmony export */
|
|
18125
|
-
/* harmony export */
|
|
18126
|
-
/* harmony export */
|
|
18127
|
-
/* harmony export */
|
|
18128
|
-
/* harmony export */
|
|
18129
|
-
/* harmony export */
|
|
18130
|
-
/* harmony export */
|
|
18786
|
+
/* harmony export */ MoonshineProcessor: () => (/* reexport safe */ _moonshine_processing_moonshine_js__WEBPACK_IMPORTED_MODULE_2__.MoonshineProcessor),
|
|
18787
|
+
/* harmony export */ OwlViTProcessor: () => (/* reexport safe */ _owlvit_processing_owlvit_js__WEBPACK_IMPORTED_MODULE_6__.OwlViTProcessor),
|
|
18788
|
+
/* harmony export */ PaliGemmaProcessor: () => (/* reexport safe */ _paligemma_processing_paligemma_js__WEBPACK_IMPORTED_MODULE_8__.PaliGemmaProcessor),
|
|
18789
|
+
/* harmony export */ Phi3VProcessor: () => (/* reexport safe */ _phi3_v_processing_phi3_v_js__WEBPACK_IMPORTED_MODULE_7__.Phi3VProcessor),
|
|
18790
|
+
/* harmony export */ PyAnnoteProcessor: () => (/* reexport safe */ _pyannote_processing_pyannote_js__WEBPACK_IMPORTED_MODULE_9__.PyAnnoteProcessor),
|
|
18791
|
+
/* harmony export */ Qwen2VLProcessor: () => (/* reexport safe */ _qwen2_vl_processing_qwen2_vl_js__WEBPACK_IMPORTED_MODULE_10__.Qwen2VLProcessor),
|
|
18792
|
+
/* harmony export */ SamProcessor: () => (/* reexport safe */ _sam_processing_sam_js__WEBPACK_IMPORTED_MODULE_11__.SamProcessor),
|
|
18793
|
+
/* harmony export */ SpeechT5Processor: () => (/* reexport safe */ _speecht5_processing_speecht5_js__WEBPACK_IMPORTED_MODULE_12__.SpeechT5Processor),
|
|
18794
|
+
/* harmony export */ VLChatProcessor: () => (/* reexport safe */ _janus_processing_janus_js__WEBPACK_IMPORTED_MODULE_4__.VLChatProcessor),
|
|
18795
|
+
/* harmony export */ Wav2Vec2ProcessorWithLM: () => (/* reexport safe */ _wav2vec2_processing_wav2vec2_js__WEBPACK_IMPORTED_MODULE_13__.Wav2Vec2ProcessorWithLM),
|
|
18796
|
+
/* harmony export */ WhisperProcessor: () => (/* reexport safe */ _whisper_processing_whisper_js__WEBPACK_IMPORTED_MODULE_14__.WhisperProcessor)
|
|
18131
18797
|
/* harmony export */ });
|
|
18132
18798
|
/* harmony import */ var _florence2_processing_florence2_js__WEBPACK_IMPORTED_MODULE_0__ = __webpack_require__(/*! ./florence2/processing_florence2.js */ "./src/models/florence2/processing_florence2.js");
|
|
18133
18799
|
/* harmony import */ var _mgp_str_processing_mgp_str_js__WEBPACK_IMPORTED_MODULE_1__ = __webpack_require__(/*! ./mgp_str/processing_mgp_str.js */ "./src/models/mgp_str/processing_mgp_str.js");
|
|
18134
|
-
/* harmony import */ var
|
|
18135
|
-
/* harmony import */ var
|
|
18136
|
-
/* harmony import */ var
|
|
18137
|
-
/* harmony import */ var
|
|
18138
|
-
/* harmony import */ var
|
|
18139
|
-
/* harmony import */ var
|
|
18140
|
-
/* harmony import */ var
|
|
18141
|
-
/* harmony import */ var
|
|
18142
|
-
/* harmony import */ var
|
|
18143
|
-
/* harmony import */ var
|
|
18800
|
+
/* harmony import */ var _moonshine_processing_moonshine_js__WEBPACK_IMPORTED_MODULE_2__ = __webpack_require__(/*! ./moonshine/processing_moonshine.js */ "./src/models/moonshine/processing_moonshine.js");
|
|
18801
|
+
/* harmony import */ var _idefics3_processing_idefics3_js__WEBPACK_IMPORTED_MODULE_3__ = __webpack_require__(/*! ./idefics3/processing_idefics3.js */ "./src/models/idefics3/processing_idefics3.js");
|
|
18802
|
+
/* harmony import */ var _janus_processing_janus_js__WEBPACK_IMPORTED_MODULE_4__ = __webpack_require__(/*! ./janus/processing_janus.js */ "./src/models/janus/processing_janus.js");
|
|
18803
|
+
/* harmony import */ var _jina_clip_processing_jina_clip_js__WEBPACK_IMPORTED_MODULE_5__ = __webpack_require__(/*! ./jina_clip/processing_jina_clip.js */ "./src/models/jina_clip/processing_jina_clip.js");
|
|
18804
|
+
/* harmony import */ var _owlvit_processing_owlvit_js__WEBPACK_IMPORTED_MODULE_6__ = __webpack_require__(/*! ./owlvit/processing_owlvit.js */ "./src/models/owlvit/processing_owlvit.js");
|
|
18805
|
+
/* harmony import */ var _phi3_v_processing_phi3_v_js__WEBPACK_IMPORTED_MODULE_7__ = __webpack_require__(/*! ./phi3_v/processing_phi3_v.js */ "./src/models/phi3_v/processing_phi3_v.js");
|
|
18806
|
+
/* harmony import */ var _paligemma_processing_paligemma_js__WEBPACK_IMPORTED_MODULE_8__ = __webpack_require__(/*! ./paligemma/processing_paligemma.js */ "./src/models/paligemma/processing_paligemma.js");
|
|
18807
|
+
/* harmony import */ var _pyannote_processing_pyannote_js__WEBPACK_IMPORTED_MODULE_9__ = __webpack_require__(/*! ./pyannote/processing_pyannote.js */ "./src/models/pyannote/processing_pyannote.js");
|
|
18808
|
+
/* harmony import */ var _qwen2_vl_processing_qwen2_vl_js__WEBPACK_IMPORTED_MODULE_10__ = __webpack_require__(/*! ./qwen2_vl/processing_qwen2_vl.js */ "./src/models/qwen2_vl/processing_qwen2_vl.js");
|
|
18809
|
+
/* harmony import */ var _sam_processing_sam_js__WEBPACK_IMPORTED_MODULE_11__ = __webpack_require__(/*! ./sam/processing_sam.js */ "./src/models/sam/processing_sam.js");
|
|
18810
|
+
/* harmony import */ var _speecht5_processing_speecht5_js__WEBPACK_IMPORTED_MODULE_12__ = __webpack_require__(/*! ./speecht5/processing_speecht5.js */ "./src/models/speecht5/processing_speecht5.js");
|
|
18811
|
+
/* harmony import */ var _wav2vec2_processing_wav2vec2_js__WEBPACK_IMPORTED_MODULE_13__ = __webpack_require__(/*! ./wav2vec2/processing_wav2vec2.js */ "./src/models/wav2vec2/processing_wav2vec2.js");
|
|
18812
|
+
/* harmony import */ var _whisper_processing_whisper_js__WEBPACK_IMPORTED_MODULE_14__ = __webpack_require__(/*! ./whisper/processing_whisper.js */ "./src/models/whisper/processing_whisper.js");
|
|
18813
|
+
|
|
18814
|
+
|
|
18815
|
+
|
|
18144
18816
|
|
|
18145
18817
|
|
|
18146
18818
|
|
|
@@ -18187,6 +18859,8 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
18187
18859
|
/* harmony export */ });
|
|
18188
18860
|
/* harmony import */ var _base_feature_extraction_utils_js__WEBPACK_IMPORTED_MODULE_0__ = __webpack_require__(/*! ../../base/feature_extraction_utils.js */ "./src/base/feature_extraction_utils.js");
|
|
18189
18861
|
/* harmony import */ var _utils_tensor_js__WEBPACK_IMPORTED_MODULE_1__ = __webpack_require__(/*! ../../utils/tensor.js */ "./src/utils/tensor.js");
|
|
18862
|
+
/* harmony import */ var _utils_maths_js__WEBPACK_IMPORTED_MODULE_2__ = __webpack_require__(/*! ../../utils/maths.js */ "./src/utils/maths.js");
|
|
18863
|
+
|
|
18190
18864
|
|
|
18191
18865
|
|
|
18192
18866
|
|
|
@@ -18214,40 +18888,6 @@ class PyAnnoteFeatureExtractor extends _base_feature_extraction_utils_js__WEBPAC
|
|
|
18214
18888
|
};
|
|
18215
18889
|
}
|
|
18216
18890
|
|
|
18217
|
-
}
|
|
18218
|
-
|
|
18219
|
-
|
|
18220
|
-
/***/ }),
|
|
18221
|
-
|
|
18222
|
-
/***/ "./src/models/pyannote/processing_pyannote.js":
|
|
18223
|
-
/*!****************************************************!*\
|
|
18224
|
-
!*** ./src/models/pyannote/processing_pyannote.js ***!
|
|
18225
|
-
\****************************************************/
|
|
18226
|
-
/***/ ((__unused_webpack___webpack_module__, __webpack_exports__, __webpack_require__) => {
|
|
18227
|
-
|
|
18228
|
-
__webpack_require__.r(__webpack_exports__);
|
|
18229
|
-
/* harmony export */ __webpack_require__.d(__webpack_exports__, {
|
|
18230
|
-
/* harmony export */ PyAnnoteProcessor: () => (/* binding */ PyAnnoteProcessor)
|
|
18231
|
-
/* harmony export */ });
|
|
18232
|
-
/* harmony import */ var _base_processing_utils_js__WEBPACK_IMPORTED_MODULE_0__ = __webpack_require__(/*! ../../base/processing_utils.js */ "./src/base/processing_utils.js");
|
|
18233
|
-
/* harmony import */ var _auto_feature_extraction_auto_js__WEBPACK_IMPORTED_MODULE_1__ = __webpack_require__(/*! ../auto/feature_extraction_auto.js */ "./src/models/auto/feature_extraction_auto.js");
|
|
18234
|
-
/* harmony import */ var _utils_maths_js__WEBPACK_IMPORTED_MODULE_2__ = __webpack_require__(/*! ../../utils/maths.js */ "./src/utils/maths.js");
|
|
18235
|
-
|
|
18236
|
-
|
|
18237
|
-
|
|
18238
|
-
|
|
18239
|
-
class PyAnnoteProcessor extends _base_processing_utils_js__WEBPACK_IMPORTED_MODULE_0__.Processor {
|
|
18240
|
-
static feature_extractor_class = _auto_feature_extraction_auto_js__WEBPACK_IMPORTED_MODULE_1__.AutoFeatureExtractor
|
|
18241
|
-
|
|
18242
|
-
/**
|
|
18243
|
-
* Calls the feature_extractor function with the given audio input.
|
|
18244
|
-
* @param {any} audio The audio input to extract features from.
|
|
18245
|
-
* @returns {Promise<any>} A Promise that resolves with the extracted features.
|
|
18246
|
-
*/
|
|
18247
|
-
async _call(audio) {
|
|
18248
|
-
return await this.feature_extractor(audio)
|
|
18249
|
-
}
|
|
18250
|
-
|
|
18251
18891
|
/**
|
|
18252
18892
|
* NOTE: Can return fractional values. `Math.ceil` will ensure correct value.
|
|
18253
18893
|
* @param {number} samples The number of frames in the audio.
|
|
@@ -18302,6 +18942,47 @@ class PyAnnoteProcessor extends _base_processing_utils_js__WEBPACK_IMPORTED_MODU
|
|
|
18302
18942
|
}
|
|
18303
18943
|
return results;
|
|
18304
18944
|
}
|
|
18945
|
+
|
|
18946
|
+
}
|
|
18947
|
+
|
|
18948
|
+
|
|
18949
|
+
/***/ }),
|
|
18950
|
+
|
|
18951
|
+
/***/ "./src/models/pyannote/processing_pyannote.js":
|
|
18952
|
+
/*!****************************************************!*\
|
|
18953
|
+
!*** ./src/models/pyannote/processing_pyannote.js ***!
|
|
18954
|
+
\****************************************************/
|
|
18955
|
+
/***/ ((__unused_webpack___webpack_module__, __webpack_exports__, __webpack_require__) => {
|
|
18956
|
+
|
|
18957
|
+
__webpack_require__.r(__webpack_exports__);
|
|
18958
|
+
/* harmony export */ __webpack_require__.d(__webpack_exports__, {
|
|
18959
|
+
/* harmony export */ PyAnnoteProcessor: () => (/* binding */ PyAnnoteProcessor)
|
|
18960
|
+
/* harmony export */ });
|
|
18961
|
+
/* harmony import */ var _base_processing_utils_js__WEBPACK_IMPORTED_MODULE_0__ = __webpack_require__(/*! ../../base/processing_utils.js */ "./src/base/processing_utils.js");
|
|
18962
|
+
/* harmony import */ var _feature_extraction_pyannote_js__WEBPACK_IMPORTED_MODULE_1__ = __webpack_require__(/*! ./feature_extraction_pyannote.js */ "./src/models/pyannote/feature_extraction_pyannote.js");
|
|
18963
|
+
|
|
18964
|
+
|
|
18965
|
+
|
|
18966
|
+
class PyAnnoteProcessor extends _base_processing_utils_js__WEBPACK_IMPORTED_MODULE_0__.Processor {
|
|
18967
|
+
static feature_extractor_class = _feature_extraction_pyannote_js__WEBPACK_IMPORTED_MODULE_1__.PyAnnoteFeatureExtractor
|
|
18968
|
+
|
|
18969
|
+
/**
|
|
18970
|
+
* Calls the feature_extractor function with the given audio input.
|
|
18971
|
+
* @param {any} audio The audio input to extract features from.
|
|
18972
|
+
* @returns {Promise<any>} A Promise that resolves with the extracted features.
|
|
18973
|
+
*/
|
|
18974
|
+
async _call(audio) {
|
|
18975
|
+
return await this.feature_extractor(audio)
|
|
18976
|
+
}
|
|
18977
|
+
|
|
18978
|
+
/** @type {PyAnnoteFeatureExtractor['post_process_speaker_diarization']} */
|
|
18979
|
+
post_process_speaker_diarization(...args) {
|
|
18980
|
+
return /** @type {PyAnnoteFeatureExtractor} */(this.feature_extractor).post_process_speaker_diarization(...args);
|
|
18981
|
+
}
|
|
18982
|
+
|
|
18983
|
+
get sampling_rate() {
|
|
18984
|
+
return this.feature_extractor.config.sampling_rate;
|
|
18985
|
+
}
|
|
18305
18986
|
}
|
|
18306
18987
|
|
|
18307
18988
|
|
|
@@ -20028,6 +20709,17 @@ class TensorOpRegistry {
|
|
|
20028
20709
|
}
|
|
20029
20710
|
return this._top_k;
|
|
20030
20711
|
}
|
|
20712
|
+
|
|
20713
|
+
static get slice() {
|
|
20714
|
+
if (!this._slice) {
|
|
20715
|
+
this._slice = wrap(
|
|
20716
|
+
[8, 7, 18, 0, 58, 96, 10, 25, 10, 1, 120, 10, 1, 115, 10, 1, 101, 10, 1, 97, 10, 1, 116, 18, 1, 121, 34, 5, 83, 108, 105, 99, 101, 18, 1, 114, 90, 9, 10, 1, 120, 18, 4, 10, 2, 8, 1, 90, 9, 10, 1, 115, 18, 4, 10, 2, 8, 7, 90, 9, 10, 1, 101, 18, 4, 10, 2, 8, 7, 90, 9, 10, 1, 97, 18, 4, 10, 2, 8, 7, 90, 9, 10, 1, 116, 18, 4, 10, 2, 8, 7, 98, 9, 10, 1, 121, 18, 4, 10, 2, 8, 1, 66, 2, 16, 13],
|
|
20717
|
+
this.session_options,
|
|
20718
|
+
'y',
|
|
20719
|
+
)
|
|
20720
|
+
}
|
|
20721
|
+
return this._slice;
|
|
20722
|
+
}
|
|
20031
20723
|
}
|
|
20032
20724
|
|
|
20033
20725
|
|
|
@@ -21760,6 +22452,8 @@ class AutomaticSpeechRecognitionPipeline extends (/** @type {new (options: TextA
|
|
|
21760
22452
|
case 'unispeech-sat':
|
|
21761
22453
|
case 'hubert':
|
|
21762
22454
|
return this._call_wav2vec2(audio, kwargs)
|
|
22455
|
+
case 'moonshine':
|
|
22456
|
+
return this._call_moonshine(audio, kwargs)
|
|
21763
22457
|
default:
|
|
21764
22458
|
throw new Error(`AutomaticSpeechRecognitionPipeline does not support model type '${this.model.config.model_type}'.`)
|
|
21765
22459
|
}
|
|
@@ -21913,6 +22607,34 @@ class AutomaticSpeechRecognitionPipeline extends (/** @type {new (options: TextA
|
|
|
21913
22607
|
}
|
|
21914
22608
|
return single ? toReturn[0] : toReturn;
|
|
21915
22609
|
}
|
|
22610
|
+
|
|
22611
|
+
/**
|
|
22612
|
+
* @type {AutomaticSpeechRecognitionPipelineCallback}
|
|
22613
|
+
* @private
|
|
22614
|
+
*/
|
|
22615
|
+
async _call_moonshine(audio, kwargs) {
|
|
22616
|
+
const single = !Array.isArray(audio);
|
|
22617
|
+
if (single) {
|
|
22618
|
+
audio = [/** @type {AudioInput} */ (audio)];
|
|
22619
|
+
}
|
|
22620
|
+
const sampling_rate = this.processor.feature_extractor.config.sampling_rate;
|
|
22621
|
+
const preparedAudios = await prepareAudios(audio, sampling_rate);
|
|
22622
|
+
const toReturn = [];
|
|
22623
|
+
for (const aud of preparedAudios) {
|
|
22624
|
+
const inputs = await this.processor(aud);
|
|
22625
|
+
|
|
22626
|
+
// According to the [paper](https://arxiv.org/pdf/2410.15608):
|
|
22627
|
+
// "We use greedy decoding, with a heuristic limit of 6 output tokens
|
|
22628
|
+
// per second of audio to avoid repeated output sequences."
|
|
22629
|
+
const max_new_tokens = Math.floor(aud.length / sampling_rate) * 6;
|
|
22630
|
+
const outputs = await this.model.generate({ max_new_tokens, ...kwargs, ...inputs });
|
|
22631
|
+
|
|
22632
|
+
const text = this.processor.batch_decode(outputs, { skip_special_tokens: true })[0];
|
|
22633
|
+
toReturn.push({ text });
|
|
22634
|
+
}
|
|
22635
|
+
return single ? toReturn[0] : toReturn;
|
|
22636
|
+
}
|
|
22637
|
+
|
|
21916
22638
|
}
|
|
21917
22639
|
|
|
21918
22640
|
/**
|
|
@@ -26030,6 +26752,12 @@ class PreTrainedTokenizer extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_0__
|
|
|
26030
26752
|
this.unk_token = this.getToken('unk_token');
|
|
26031
26753
|
this.unk_token_id = this.model.tokens_to_ids.get(this.unk_token);
|
|
26032
26754
|
|
|
26755
|
+
this.bos_token = this.getToken('bos_token');
|
|
26756
|
+
this.bos_token_id = this.model.tokens_to_ids.get(this.bos_token);
|
|
26757
|
+
|
|
26758
|
+
this.eos_token = this.getToken('eos_token');
|
|
26759
|
+
this.eos_token_id = this.model.tokens_to_ids.get(this.eos_token);
|
|
26760
|
+
|
|
26033
26761
|
this.model_max_length = tokenizerConfig.model_max_length;
|
|
26034
26762
|
|
|
26035
26763
|
/** @type {boolean} Whether or not to strip the text when tokenizing (removing excess spaces before and after the string). */
|
|
@@ -27002,6 +27730,11 @@ class WhisperTokenizer extends PreTrainedTokenizer {
|
|
|
27002
27730
|
let chunk = new_chunk();
|
|
27003
27731
|
let time_offset = 0.0;
|
|
27004
27732
|
const timestamp_begin = this.timestamp_begin;
|
|
27733
|
+
// Whisper timestamp tokens start from 0.00 and go to timestamp 30.00 in 0.02 increments.
|
|
27734
|
+
// We can calculate the last time stamp token as timestamp_begin plus the number of tokens
|
|
27735
|
+
// tokens from 0.00 to 30.00 which is 1500.
|
|
27736
|
+
const total_timestamp_tokens = 1500; // (30.00 - 0.00) / 0.02
|
|
27737
|
+
const timestamp_end = timestamp_begin + total_timestamp_tokens;
|
|
27005
27738
|
|
|
27006
27739
|
let previous_tokens = [];
|
|
27007
27740
|
let previous_token_timestamps = [];
|
|
@@ -27089,7 +27822,7 @@ class WhisperTokenizer extends PreTrainedTokenizer {
|
|
|
27089
27822
|
} else {
|
|
27090
27823
|
// 2/ This is a regular special token, ignoring it
|
|
27091
27824
|
}
|
|
27092
|
-
} else if (token >= timestamp_begin) {
|
|
27825
|
+
} else if (token >= timestamp_begin && token <= timestamp_end) {
|
|
27093
27826
|
// 3/ Timestamp token
|
|
27094
27827
|
const time = (token - timestamp_begin) * time_precision + time_offset;
|
|
27095
27828
|
const rounded_time = (0,_utils_maths_js__WEBPACK_IMPORTED_MODULE_3__.round)(time, 2);
|
|
@@ -28578,15 +29311,45 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
28578
29311
|
*/
|
|
28579
29312
|
|
|
28580
29313
|
/**
|
|
28581
|
-
* @typedef {Object}
|
|
28582
|
-
* @property {'initiate'
|
|
28583
|
-
* @property {string} name
|
|
28584
|
-
*
|
|
28585
|
-
|
|
28586
|
-
|
|
28587
|
-
|
|
28588
|
-
* @
|
|
28589
|
-
* @property {
|
|
29314
|
+
* @typedef {Object} InitiateProgressInfo
|
|
29315
|
+
* @property {'initiate'} status
|
|
29316
|
+
* @property {string} name The model id or directory path.
|
|
29317
|
+
* @property {string} file The name of the file.
|
|
29318
|
+
*/
|
|
29319
|
+
|
|
29320
|
+
/**
|
|
29321
|
+
* @typedef {Object} DownloadProgressInfo
|
|
29322
|
+
* @property {'download'} status
|
|
29323
|
+
* @property {string} name The model id or directory path.
|
|
29324
|
+
* @property {string} file The name of the file.
|
|
29325
|
+
*/
|
|
29326
|
+
|
|
29327
|
+
/**
|
|
29328
|
+
* @typedef {Object} ProgressStatusInfo
|
|
29329
|
+
* @property {'progress'} status
|
|
29330
|
+
* @property {string} name The model id or directory path.
|
|
29331
|
+
* @property {string} file The name of the file.
|
|
29332
|
+
* @property {number} progress A number between 0 and 100.
|
|
29333
|
+
* @property {number} loaded The number of bytes loaded.
|
|
29334
|
+
* @property {number} total The total number of bytes to be loaded.
|
|
29335
|
+
*/
|
|
29336
|
+
|
|
29337
|
+
/**
|
|
29338
|
+
* @typedef {Object} DoneProgressInfo
|
|
29339
|
+
* @property {'done'} status
|
|
29340
|
+
* @property {string} name The model id or directory path.
|
|
29341
|
+
* @property {string} file The name of the file.
|
|
29342
|
+
*/
|
|
29343
|
+
|
|
29344
|
+
/**
|
|
29345
|
+
* @typedef {Object} ReadyProgressInfo
|
|
29346
|
+
* @property {'ready'} status
|
|
29347
|
+
* @property {string} task The loaded task.
|
|
29348
|
+
* @property {string} model The loaded model.
|
|
29349
|
+
*/
|
|
29350
|
+
|
|
29351
|
+
/**
|
|
29352
|
+
* @typedef {InitiateProgressInfo | DownloadProgressInfo | ProgressStatusInfo | DoneProgressInfo | ReadyProgressInfo} ProgressInfo
|
|
28590
29353
|
*/
|
|
28591
29354
|
|
|
28592
29355
|
/**
|
|
@@ -29924,13 +30687,6 @@ async function getModelFile(path_or_repo_id, filename, fatal = true, options = {
|
|
|
29924
30687
|
file: filename
|
|
29925
30688
|
})
|
|
29926
30689
|
|
|
29927
|
-
/** @type {import('./core.js').ProgressInfo} */
|
|
29928
|
-
const progressInfo = {
|
|
29929
|
-
status: 'progress',
|
|
29930
|
-
name: path_or_repo_id,
|
|
29931
|
-
file: filename
|
|
29932
|
-
}
|
|
29933
|
-
|
|
29934
30690
|
/** @type {Uint8Array} */
|
|
29935
30691
|
let buffer;
|
|
29936
30692
|
|
|
@@ -29950,7 +30706,9 @@ async function getModelFile(path_or_repo_id, filename, fatal = true, options = {
|
|
|
29950
30706
|
|
|
29951
30707
|
// For completeness, we still fire the final progress callback
|
|
29952
30708
|
(0,_core_js__WEBPACK_IMPORTED_MODULE_3__.dispatchCallback)(options.progress_callback, {
|
|
29953
|
-
|
|
30709
|
+
status: 'progress',
|
|
30710
|
+
name: path_or_repo_id,
|
|
30711
|
+
file: filename,
|
|
29954
30712
|
progress: 100,
|
|
29955
30713
|
loaded: buffer.length,
|
|
29956
30714
|
total: buffer.length,
|
|
@@ -29958,7 +30716,9 @@ async function getModelFile(path_or_repo_id, filename, fatal = true, options = {
|
|
|
29958
30716
|
} else {
|
|
29959
30717
|
buffer = await readResponse(response, data => {
|
|
29960
30718
|
(0,_core_js__WEBPACK_IMPORTED_MODULE_3__.dispatchCallback)(options.progress_callback, {
|
|
29961
|
-
|
|
30719
|
+
status: 'progress',
|
|
30720
|
+
name: path_or_repo_id,
|
|
30721
|
+
file: filename,
|
|
29962
30722
|
...data,
|
|
29963
30723
|
})
|
|
29964
30724
|
})
|
|
@@ -30015,12 +30775,11 @@ async function getModelJSON(modelPath, fileName, fatal = true, options = {}) {
|
|
|
30015
30775
|
|
|
30016
30776
|
return JSON.parse(jsonData);
|
|
30017
30777
|
}
|
|
30018
|
-
|
|
30019
30778
|
/**
|
|
30020
30779
|
* Read and track progress when reading a Response object
|
|
30021
30780
|
*
|
|
30022
|
-
* @param {
|
|
30023
|
-
* @param {
|
|
30781
|
+
* @param {Response|FileResponse} response The Response object to read
|
|
30782
|
+
* @param {(data: {progress: number, loaded: number, total: number}) => void} progress_callback The function to call with progress updates
|
|
30024
30783
|
* @returns {Promise<Uint8Array>} A Promise that resolves with the Uint8Array buffer
|
|
30025
30784
|
*/
|
|
30026
30785
|
async function readResponse(response, progress_callback) {
|
|
@@ -30416,6 +31175,46 @@ class RawImage {
|
|
|
30416
31175
|
return this._update(newData, this.width, this.height, 4);
|
|
30417
31176
|
}
|
|
30418
31177
|
|
|
31178
|
+
/**
|
|
31179
|
+
* Apply an alpha mask to the image. Operates in place.
|
|
31180
|
+
* @param {RawImage} mask The mask to apply. It should have a single channel.
|
|
31181
|
+
* @returns {RawImage} The masked image.
|
|
31182
|
+
* @throws {Error} If the mask is not the same size as the image.
|
|
31183
|
+
* @throws {Error} If the image does not have 4 channels.
|
|
31184
|
+
* @throws {Error} If the mask is not a single channel.
|
|
31185
|
+
*/
|
|
31186
|
+
putAlpha(mask) {
|
|
31187
|
+
if (mask.width !== this.width || mask.height !== this.height) {
|
|
31188
|
+
throw new Error(`Expected mask size to be ${this.width}x${this.height}, but got ${mask.width}x${mask.height}`);
|
|
31189
|
+
}
|
|
31190
|
+
if (mask.channels !== 1) {
|
|
31191
|
+
throw new Error(`Expected mask to have 1 channel, but got ${mask.channels}`);
|
|
31192
|
+
}
|
|
31193
|
+
|
|
31194
|
+
const this_data = this.data;
|
|
31195
|
+
const mask_data = mask.data;
|
|
31196
|
+
const num_pixels = this.width * this.height;
|
|
31197
|
+
if (this.channels === 3) {
|
|
31198
|
+
// Convert to RGBA and simultaneously apply mask to alpha channel
|
|
31199
|
+
const newData = new Uint8ClampedArray(num_pixels * 4);
|
|
31200
|
+
for (let i = 0, in_offset = 0, out_offset = 0; i < num_pixels; ++i) {
|
|
31201
|
+
newData[out_offset++] = this_data[in_offset++];
|
|
31202
|
+
newData[out_offset++] = this_data[in_offset++];
|
|
31203
|
+
newData[out_offset++] = this_data[in_offset++];
|
|
31204
|
+
newData[out_offset++] = mask_data[i];
|
|
31205
|
+
}
|
|
31206
|
+
return this._update(newData, this.width, this.height, 4);
|
|
31207
|
+
|
|
31208
|
+
} else if (this.channels === 4) {
|
|
31209
|
+
// Apply mask to alpha channel in place
|
|
31210
|
+
for (let i = 0; i < num_pixels; ++i) {
|
|
31211
|
+
this_data[4 * i + 3] = mask_data[i];
|
|
31212
|
+
}
|
|
31213
|
+
return this;
|
|
31214
|
+
}
|
|
31215
|
+
throw new Error(`Expected image to have 3 or 4 channels, but got ${this.channels}`);
|
|
31216
|
+
}
|
|
31217
|
+
|
|
30419
31218
|
/**
|
|
30420
31219
|
* Resize the image to the given dimensions. This method uses the canvas API to perform the resizing.
|
|
30421
31220
|
* @param {number} width The width of the new image. `null` or `-1` will preserve the aspect ratio.
|
|
@@ -32022,7 +32821,9 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
32022
32821
|
/* harmony export */ ones_like: () => (/* binding */ ones_like),
|
|
32023
32822
|
/* harmony export */ permute: () => (/* binding */ permute),
|
|
32024
32823
|
/* harmony export */ quantize_embeddings: () => (/* binding */ quantize_embeddings),
|
|
32824
|
+
/* harmony export */ rand: () => (/* binding */ rand),
|
|
32025
32825
|
/* harmony export */ rfft: () => (/* binding */ rfft),
|
|
32826
|
+
/* harmony export */ slice: () => (/* binding */ slice),
|
|
32026
32827
|
/* harmony export */ stack: () => (/* binding */ stack),
|
|
32027
32828
|
/* harmony export */ std_mean: () => (/* binding */ std_mean),
|
|
32028
32829
|
/* harmony export */ topk: () => (/* binding */ topk),
|
|
@@ -32801,8 +33602,21 @@ class Tensor {
|
|
|
32801
33602
|
if (!DataTypeMap.hasOwnProperty(type)) {
|
|
32802
33603
|
throw new Error(`Unsupported type: ${type}`);
|
|
32803
33604
|
}
|
|
33605
|
+
|
|
33606
|
+
// Handle special cases where a mapping function is needed (e.g., where one type is a bigint and the other is a number)
|
|
33607
|
+
let map_fn;
|
|
33608
|
+
const is_source_bigint = ['int64', 'uint64'].includes(this.type);
|
|
33609
|
+
const is_dest_bigint = ['int64', 'uint64'].includes(type);
|
|
33610
|
+
if (is_source_bigint && !is_dest_bigint) {
|
|
33611
|
+
// TypeError: Cannot convert a BigInt value to a number
|
|
33612
|
+
map_fn = Number;
|
|
33613
|
+
} else if (!is_source_bigint && is_dest_bigint) {
|
|
33614
|
+
// TypeError: Cannot convert [x] to a BigInt
|
|
33615
|
+
map_fn = BigInt;
|
|
33616
|
+
}
|
|
33617
|
+
|
|
32804
33618
|
// @ts-ignore
|
|
32805
|
-
return new Tensor(type, DataTypeMap[type].from(this.data), this.dims);
|
|
33619
|
+
return new Tensor(type, DataTypeMap[type].from(this.data, map_fn), this.dims);
|
|
32806
33620
|
}
|
|
32807
33621
|
}
|
|
32808
33622
|
|
|
@@ -33000,6 +33814,29 @@ async function topk(x, k) {
|
|
|
33000
33814
|
});
|
|
33001
33815
|
}
|
|
33002
33816
|
|
|
33817
|
+
|
|
33818
|
+
const arrayToIndexTensor = (array) => new Tensor('int64', array, [array.length]);
|
|
33819
|
+
/**
|
|
33820
|
+
* Slice a multidimensional float32 tensor.
|
|
33821
|
+
* @param {Tensor} data: Tensor of data to extract slices from
|
|
33822
|
+
* @param {number[]} starts: 1-D array of starting indices of corresponding axis in axes
|
|
33823
|
+
* @param {number[]} ends: 1-D array of ending indices (exclusive) of corresponding axis in axes
|
|
33824
|
+
* @param {number[]} axes: 1-D array of axes that starts and ends apply to
|
|
33825
|
+
* @param {number[]} [steps]: 1-D array of slice step of corresponding axis in axes.
|
|
33826
|
+
* @returns {Promise<Tensor>} Sliced data tensor.
|
|
33827
|
+
*/
|
|
33828
|
+
async function slice(data, starts, ends, axes, steps) {
|
|
33829
|
+
const op = await _ops_registry_js__WEBPACK_IMPORTED_MODULE_2__.TensorOpRegistry.slice;
|
|
33830
|
+
return await op({
|
|
33831
|
+
x: data,
|
|
33832
|
+
s: arrayToIndexTensor(starts),
|
|
33833
|
+
e: arrayToIndexTensor(ends),
|
|
33834
|
+
a: arrayToIndexTensor(axes),
|
|
33835
|
+
t: arrayToIndexTensor(steps ?? new Array(axes.length).fill(1)),
|
|
33836
|
+
});
|
|
33837
|
+
}
|
|
33838
|
+
|
|
33839
|
+
|
|
33003
33840
|
/**
|
|
33004
33841
|
* Perform mean pooling of the last hidden state followed by a normalization step.
|
|
33005
33842
|
* @param {Tensor} last_hidden_state Tensor of shape [batchSize, seqLength, embedDim]
|
|
@@ -33446,6 +34283,20 @@ function zeros_like(tensor) {
|
|
|
33446
34283
|
return zeros(tensor.dims);
|
|
33447
34284
|
}
|
|
33448
34285
|
|
|
34286
|
+
/**
|
|
34287
|
+
* Returns a tensor filled with random numbers from a uniform distribution on the interval [0, 1)
|
|
34288
|
+
* @param {number[]} size A sequence of integers defining the shape of the output tensor.
|
|
34289
|
+
* @returns {Tensor} The random tensor.
|
|
34290
|
+
*/
|
|
34291
|
+
function rand(size) {
|
|
34292
|
+
const length = size.reduce((a, b) => a * b, 1);
|
|
34293
|
+
return new Tensor(
|
|
34294
|
+
"float32",
|
|
34295
|
+
Float32Array.from({ length }, () => Math.random()),
|
|
34296
|
+
size,
|
|
34297
|
+
)
|
|
34298
|
+
}
|
|
34299
|
+
|
|
33449
34300
|
/**
|
|
33450
34301
|
* Quantizes the embeddings tensor to binary or unsigned binary precision.
|
|
33451
34302
|
* @param {Tensor} tensor The tensor to quantize.
|
|
@@ -33580,6 +34431,8 @@ function quantize_embeddings(tensor, precision) {
|
|
|
33580
34431
|
/******/
|
|
33581
34432
|
/************************************************************************/
|
|
33582
34433
|
var __webpack_exports__ = {};
|
|
34434
|
+
// This entry needs to be wrapped in an IIFE because it needs to be isolated against other modules in the chunk.
|
|
34435
|
+
(() => {
|
|
33583
34436
|
/*!*****************************!*\
|
|
33584
34437
|
!*** ./src/transformers.js ***!
|
|
33585
34438
|
\*****************************/
|
|
@@ -33788,6 +34641,9 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
33788
34641
|
/* harmony export */ EsmModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.EsmModel),
|
|
33789
34642
|
/* harmony export */ EsmPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.EsmPreTrainedModel),
|
|
33790
34643
|
/* harmony export */ EsmTokenizer: () => (/* reexport safe */ _tokenizers_js__WEBPACK_IMPORTED_MODULE_3__.EsmTokenizer),
|
|
34644
|
+
/* harmony export */ ExaoneForCausalLM: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.ExaoneForCausalLM),
|
|
34645
|
+
/* harmony export */ ExaoneModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.ExaoneModel),
|
|
34646
|
+
/* harmony export */ ExaonePreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.ExaonePreTrainedModel),
|
|
33791
34647
|
/* harmony export */ FFT: () => (/* reexport safe */ _utils_maths_js__WEBPACK_IMPORTED_MODULE_8__.FFT),
|
|
33792
34648
|
/* harmony export */ FalconForCausalLM: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.FalconForCausalLM),
|
|
33793
34649
|
/* harmony export */ FalconModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.FalconModel),
|
|
@@ -33846,6 +34702,9 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
33846
34702
|
/* harmony export */ HubertForSequenceClassification: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.HubertForSequenceClassification),
|
|
33847
34703
|
/* harmony export */ HubertModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.HubertModel),
|
|
33848
34704
|
/* harmony export */ HubertPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.HubertPreTrainedModel),
|
|
34705
|
+
/* harmony export */ IJepaForImageClassification: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.IJepaForImageClassification),
|
|
34706
|
+
/* harmony export */ IJepaModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.IJepaModel),
|
|
34707
|
+
/* harmony export */ IJepaPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.IJepaPreTrainedModel),
|
|
33849
34708
|
/* harmony export */ Idefics3ForConditionalGeneration: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Idefics3ForConditionalGeneration),
|
|
33850
34709
|
/* harmony export */ Idefics3ImageProcessor: () => (/* reexport safe */ _models_image_processors_js__WEBPACK_IMPORTED_MODULE_13__.Idefics3ImageProcessor),
|
|
33851
34710
|
/* harmony export */ Idefics3PreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Idefics3PreTrainedModel),
|
|
@@ -33964,6 +34823,11 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
33964
34823
|
/* harmony export */ MobileViTV2PreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.MobileViTV2PreTrainedModel),
|
|
33965
34824
|
/* harmony export */ ModelOutput: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.ModelOutput),
|
|
33966
34825
|
/* harmony export */ Moondream1ForConditionalGeneration: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Moondream1ForConditionalGeneration),
|
|
34826
|
+
/* harmony export */ MoonshineFeatureExtractor: () => (/* reexport safe */ _models_feature_extractors_js__WEBPACK_IMPORTED_MODULE_10__.MoonshineFeatureExtractor),
|
|
34827
|
+
/* harmony export */ MoonshineForConditionalGeneration: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.MoonshineForConditionalGeneration),
|
|
34828
|
+
/* harmony export */ MoonshineModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.MoonshineModel),
|
|
34829
|
+
/* harmony export */ MoonshinePreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.MoonshinePreTrainedModel),
|
|
34830
|
+
/* harmony export */ MoonshineProcessor: () => (/* reexport safe */ _models_processors_js__WEBPACK_IMPORTED_MODULE_16__.MoonshineProcessor),
|
|
33967
34831
|
/* harmony export */ MptForCausalLM: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.MptForCausalLM),
|
|
33968
34832
|
/* harmony export */ MptModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.MptModel),
|
|
33969
34833
|
/* harmony export */ MptPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.MptPreTrainedModel),
|
|
@@ -33984,6 +34848,9 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
33984
34848
|
/* harmony export */ OPTModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.OPTModel),
|
|
33985
34849
|
/* harmony export */ OPTPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.OPTPreTrainedModel),
|
|
33986
34850
|
/* harmony export */ ObjectDetectionPipeline: () => (/* reexport safe */ _pipelines_js__WEBPACK_IMPORTED_MODULE_1__.ObjectDetectionPipeline),
|
|
34851
|
+
/* harmony export */ Olmo2ForCausalLM: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Olmo2ForCausalLM),
|
|
34852
|
+
/* harmony export */ Olmo2Model: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Olmo2Model),
|
|
34853
|
+
/* harmony export */ Olmo2PreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Olmo2PreTrainedModel),
|
|
33987
34854
|
/* harmony export */ OlmoForCausalLM: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.OlmoForCausalLM),
|
|
33988
34855
|
/* harmony export */ OlmoModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.OlmoModel),
|
|
33989
34856
|
/* harmony export */ OlmoPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.OlmoPreTrainedModel),
|
|
@@ -34000,6 +34867,9 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
34000
34867
|
/* harmony export */ Owlv2ImageProcessor: () => (/* reexport safe */ _models_image_processors_js__WEBPACK_IMPORTED_MODULE_13__.Owlv2ImageProcessor),
|
|
34001
34868
|
/* harmony export */ Owlv2Model: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Owlv2Model),
|
|
34002
34869
|
/* harmony export */ Owlv2PreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Owlv2PreTrainedModel),
|
|
34870
|
+
/* harmony export */ PaliGemmaForConditionalGeneration: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.PaliGemmaForConditionalGeneration),
|
|
34871
|
+
/* harmony export */ PaliGemmaPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.PaliGemmaPreTrainedModel),
|
|
34872
|
+
/* harmony export */ PaliGemmaProcessor: () => (/* reexport safe */ _models_processors_js__WEBPACK_IMPORTED_MODULE_16__.PaliGemmaProcessor),
|
|
34003
34873
|
/* harmony export */ PatchTSMixerForPrediction: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.PatchTSMixerForPrediction),
|
|
34004
34874
|
/* harmony export */ PatchTSMixerModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.PatchTSMixerModel),
|
|
34005
34875
|
/* harmony export */ PatchTSMixerPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.PatchTSMixerPreTrainedModel),
|
|
@@ -34009,6 +34879,10 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
34009
34879
|
/* harmony export */ Phi3ForCausalLM: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Phi3ForCausalLM),
|
|
34010
34880
|
/* harmony export */ Phi3Model: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Phi3Model),
|
|
34011
34881
|
/* harmony export */ Phi3PreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Phi3PreTrainedModel),
|
|
34882
|
+
/* harmony export */ Phi3VForCausalLM: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Phi3VForCausalLM),
|
|
34883
|
+
/* harmony export */ Phi3VImageProcessor: () => (/* reexport safe */ _models_image_processors_js__WEBPACK_IMPORTED_MODULE_13__.Phi3VImageProcessor),
|
|
34884
|
+
/* harmony export */ Phi3VPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Phi3VPreTrainedModel),
|
|
34885
|
+
/* harmony export */ Phi3VProcessor: () => (/* reexport safe */ _models_processors_js__WEBPACK_IMPORTED_MODULE_16__.Phi3VProcessor),
|
|
34012
34886
|
/* harmony export */ PhiForCausalLM: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.PhiForCausalLM),
|
|
34013
34887
|
/* harmony export */ PhiModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.PhiModel),
|
|
34014
34888
|
/* harmony export */ PhiPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.PhiPreTrainedModel),
|
|
@@ -34257,9 +35131,11 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
34257
35131
|
/* harmony export */ permute_data: () => (/* reexport safe */ _utils_maths_js__WEBPACK_IMPORTED_MODULE_8__.permute_data),
|
|
34258
35132
|
/* harmony export */ pipeline: () => (/* reexport safe */ _pipelines_js__WEBPACK_IMPORTED_MODULE_1__.pipeline),
|
|
34259
35133
|
/* harmony export */ quantize_embeddings: () => (/* reexport safe */ _utils_tensor_js__WEBPACK_IMPORTED_MODULE_7__.quantize_embeddings),
|
|
35134
|
+
/* harmony export */ rand: () => (/* reexport safe */ _utils_tensor_js__WEBPACK_IMPORTED_MODULE_7__.rand),
|
|
34260
35135
|
/* harmony export */ read_audio: () => (/* reexport safe */ _utils_audio_js__WEBPACK_IMPORTED_MODULE_5__.read_audio),
|
|
34261
35136
|
/* harmony export */ rfft: () => (/* reexport safe */ _utils_tensor_js__WEBPACK_IMPORTED_MODULE_7__.rfft),
|
|
34262
35137
|
/* harmony export */ round: () => (/* reexport safe */ _utils_maths_js__WEBPACK_IMPORTED_MODULE_8__.round),
|
|
35138
|
+
/* harmony export */ slice: () => (/* reexport safe */ _utils_tensor_js__WEBPACK_IMPORTED_MODULE_7__.slice),
|
|
34263
35139
|
/* harmony export */ softmax: () => (/* reexport safe */ _utils_maths_js__WEBPACK_IMPORTED_MODULE_8__.softmax),
|
|
34264
35140
|
/* harmony export */ spectrogram: () => (/* reexport safe */ _utils_audio_js__WEBPACK_IMPORTED_MODULE_5__.spectrogram),
|
|
34265
35141
|
/* harmony export */ stack: () => (/* reexport safe */ _utils_tensor_js__WEBPACK_IMPORTED_MODULE_7__.stack),
|
|
@@ -34331,6 +35207,8 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
34331
35207
|
|
|
34332
35208
|
|
|
34333
35209
|
|
|
35210
|
+
|
|
35211
|
+
})();
|
|
34334
35212
|
|
|
34335
35213
|
var __webpack_exports__ASTFeatureExtractor = __webpack_exports__.ASTFeatureExtractor;
|
|
34336
35214
|
var __webpack_exports__ASTForAudioClassification = __webpack_exports__.ASTForAudioClassification;
|
|
@@ -34535,6 +35413,9 @@ var __webpack_exports__EsmForTokenClassification = __webpack_exports__.EsmForTok
|
|
|
34535
35413
|
var __webpack_exports__EsmModel = __webpack_exports__.EsmModel;
|
|
34536
35414
|
var __webpack_exports__EsmPreTrainedModel = __webpack_exports__.EsmPreTrainedModel;
|
|
34537
35415
|
var __webpack_exports__EsmTokenizer = __webpack_exports__.EsmTokenizer;
|
|
35416
|
+
var __webpack_exports__ExaoneForCausalLM = __webpack_exports__.ExaoneForCausalLM;
|
|
35417
|
+
var __webpack_exports__ExaoneModel = __webpack_exports__.ExaoneModel;
|
|
35418
|
+
var __webpack_exports__ExaonePreTrainedModel = __webpack_exports__.ExaonePreTrainedModel;
|
|
34538
35419
|
var __webpack_exports__FFT = __webpack_exports__.FFT;
|
|
34539
35420
|
var __webpack_exports__FalconForCausalLM = __webpack_exports__.FalconForCausalLM;
|
|
34540
35421
|
var __webpack_exports__FalconModel = __webpack_exports__.FalconModel;
|
|
@@ -34593,6 +35474,9 @@ var __webpack_exports__HubertForCTC = __webpack_exports__.HubertForCTC;
|
|
|
34593
35474
|
var __webpack_exports__HubertForSequenceClassification = __webpack_exports__.HubertForSequenceClassification;
|
|
34594
35475
|
var __webpack_exports__HubertModel = __webpack_exports__.HubertModel;
|
|
34595
35476
|
var __webpack_exports__HubertPreTrainedModel = __webpack_exports__.HubertPreTrainedModel;
|
|
35477
|
+
var __webpack_exports__IJepaForImageClassification = __webpack_exports__.IJepaForImageClassification;
|
|
35478
|
+
var __webpack_exports__IJepaModel = __webpack_exports__.IJepaModel;
|
|
35479
|
+
var __webpack_exports__IJepaPreTrainedModel = __webpack_exports__.IJepaPreTrainedModel;
|
|
34596
35480
|
var __webpack_exports__Idefics3ForConditionalGeneration = __webpack_exports__.Idefics3ForConditionalGeneration;
|
|
34597
35481
|
var __webpack_exports__Idefics3ImageProcessor = __webpack_exports__.Idefics3ImageProcessor;
|
|
34598
35482
|
var __webpack_exports__Idefics3PreTrainedModel = __webpack_exports__.Idefics3PreTrainedModel;
|
|
@@ -34711,6 +35595,11 @@ var __webpack_exports__MobileViTV2Model = __webpack_exports__.MobileViTV2Model;
|
|
|
34711
35595
|
var __webpack_exports__MobileViTV2PreTrainedModel = __webpack_exports__.MobileViTV2PreTrainedModel;
|
|
34712
35596
|
var __webpack_exports__ModelOutput = __webpack_exports__.ModelOutput;
|
|
34713
35597
|
var __webpack_exports__Moondream1ForConditionalGeneration = __webpack_exports__.Moondream1ForConditionalGeneration;
|
|
35598
|
+
var __webpack_exports__MoonshineFeatureExtractor = __webpack_exports__.MoonshineFeatureExtractor;
|
|
35599
|
+
var __webpack_exports__MoonshineForConditionalGeneration = __webpack_exports__.MoonshineForConditionalGeneration;
|
|
35600
|
+
var __webpack_exports__MoonshineModel = __webpack_exports__.MoonshineModel;
|
|
35601
|
+
var __webpack_exports__MoonshinePreTrainedModel = __webpack_exports__.MoonshinePreTrainedModel;
|
|
35602
|
+
var __webpack_exports__MoonshineProcessor = __webpack_exports__.MoonshineProcessor;
|
|
34714
35603
|
var __webpack_exports__MptForCausalLM = __webpack_exports__.MptForCausalLM;
|
|
34715
35604
|
var __webpack_exports__MptModel = __webpack_exports__.MptModel;
|
|
34716
35605
|
var __webpack_exports__MptPreTrainedModel = __webpack_exports__.MptPreTrainedModel;
|
|
@@ -34731,6 +35620,9 @@ var __webpack_exports__OPTForCausalLM = __webpack_exports__.OPTForCausalLM;
|
|
|
34731
35620
|
var __webpack_exports__OPTModel = __webpack_exports__.OPTModel;
|
|
34732
35621
|
var __webpack_exports__OPTPreTrainedModel = __webpack_exports__.OPTPreTrainedModel;
|
|
34733
35622
|
var __webpack_exports__ObjectDetectionPipeline = __webpack_exports__.ObjectDetectionPipeline;
|
|
35623
|
+
var __webpack_exports__Olmo2ForCausalLM = __webpack_exports__.Olmo2ForCausalLM;
|
|
35624
|
+
var __webpack_exports__Olmo2Model = __webpack_exports__.Olmo2Model;
|
|
35625
|
+
var __webpack_exports__Olmo2PreTrainedModel = __webpack_exports__.Olmo2PreTrainedModel;
|
|
34734
35626
|
var __webpack_exports__OlmoForCausalLM = __webpack_exports__.OlmoForCausalLM;
|
|
34735
35627
|
var __webpack_exports__OlmoModel = __webpack_exports__.OlmoModel;
|
|
34736
35628
|
var __webpack_exports__OlmoPreTrainedModel = __webpack_exports__.OlmoPreTrainedModel;
|
|
@@ -34747,6 +35639,9 @@ var __webpack_exports__Owlv2ForObjectDetection = __webpack_exports__.Owlv2ForObj
|
|
|
34747
35639
|
var __webpack_exports__Owlv2ImageProcessor = __webpack_exports__.Owlv2ImageProcessor;
|
|
34748
35640
|
var __webpack_exports__Owlv2Model = __webpack_exports__.Owlv2Model;
|
|
34749
35641
|
var __webpack_exports__Owlv2PreTrainedModel = __webpack_exports__.Owlv2PreTrainedModel;
|
|
35642
|
+
var __webpack_exports__PaliGemmaForConditionalGeneration = __webpack_exports__.PaliGemmaForConditionalGeneration;
|
|
35643
|
+
var __webpack_exports__PaliGemmaPreTrainedModel = __webpack_exports__.PaliGemmaPreTrainedModel;
|
|
35644
|
+
var __webpack_exports__PaliGemmaProcessor = __webpack_exports__.PaliGemmaProcessor;
|
|
34750
35645
|
var __webpack_exports__PatchTSMixerForPrediction = __webpack_exports__.PatchTSMixerForPrediction;
|
|
34751
35646
|
var __webpack_exports__PatchTSMixerModel = __webpack_exports__.PatchTSMixerModel;
|
|
34752
35647
|
var __webpack_exports__PatchTSMixerPreTrainedModel = __webpack_exports__.PatchTSMixerPreTrainedModel;
|
|
@@ -34756,6 +35651,10 @@ var __webpack_exports__PatchTSTPreTrainedModel = __webpack_exports__.PatchTSTPre
|
|
|
34756
35651
|
var __webpack_exports__Phi3ForCausalLM = __webpack_exports__.Phi3ForCausalLM;
|
|
34757
35652
|
var __webpack_exports__Phi3Model = __webpack_exports__.Phi3Model;
|
|
34758
35653
|
var __webpack_exports__Phi3PreTrainedModel = __webpack_exports__.Phi3PreTrainedModel;
|
|
35654
|
+
var __webpack_exports__Phi3VForCausalLM = __webpack_exports__.Phi3VForCausalLM;
|
|
35655
|
+
var __webpack_exports__Phi3VImageProcessor = __webpack_exports__.Phi3VImageProcessor;
|
|
35656
|
+
var __webpack_exports__Phi3VPreTrainedModel = __webpack_exports__.Phi3VPreTrainedModel;
|
|
35657
|
+
var __webpack_exports__Phi3VProcessor = __webpack_exports__.Phi3VProcessor;
|
|
34759
35658
|
var __webpack_exports__PhiForCausalLM = __webpack_exports__.PhiForCausalLM;
|
|
34760
35659
|
var __webpack_exports__PhiModel = __webpack_exports__.PhiModel;
|
|
34761
35660
|
var __webpack_exports__PhiPreTrainedModel = __webpack_exports__.PhiPreTrainedModel;
|
|
@@ -35004,9 +35903,11 @@ var __webpack_exports__permute = __webpack_exports__.permute;
|
|
|
35004
35903
|
var __webpack_exports__permute_data = __webpack_exports__.permute_data;
|
|
35005
35904
|
var __webpack_exports__pipeline = __webpack_exports__.pipeline;
|
|
35006
35905
|
var __webpack_exports__quantize_embeddings = __webpack_exports__.quantize_embeddings;
|
|
35906
|
+
var __webpack_exports__rand = __webpack_exports__.rand;
|
|
35007
35907
|
var __webpack_exports__read_audio = __webpack_exports__.read_audio;
|
|
35008
35908
|
var __webpack_exports__rfft = __webpack_exports__.rfft;
|
|
35009
35909
|
var __webpack_exports__round = __webpack_exports__.round;
|
|
35910
|
+
var __webpack_exports__slice = __webpack_exports__.slice;
|
|
35010
35911
|
var __webpack_exports__softmax = __webpack_exports__.softmax;
|
|
35011
35912
|
var __webpack_exports__spectrogram = __webpack_exports__.spectrogram;
|
|
35012
35913
|
var __webpack_exports__stack = __webpack_exports__.stack;
|
|
@@ -35015,6 +35916,6 @@ var __webpack_exports__topk = __webpack_exports__.topk;
|
|
|
35015
35916
|
var __webpack_exports__window_function = __webpack_exports__.window_function;
|
|
35016
35917
|
var __webpack_exports__zeros = __webpack_exports__.zeros;
|
|
35017
35918
|
var __webpack_exports__zeros_like = __webpack_exports__.zeros_like;
|
|
35018
|
-
export { __webpack_exports__ASTFeatureExtractor as ASTFeatureExtractor, __webpack_exports__ASTForAudioClassification as ASTForAudioClassification, __webpack_exports__ASTModel as ASTModel, __webpack_exports__ASTPreTrainedModel as ASTPreTrainedModel, __webpack_exports__AlbertForMaskedLM as AlbertForMaskedLM, __webpack_exports__AlbertForQuestionAnswering as AlbertForQuestionAnswering, __webpack_exports__AlbertForSequenceClassification as AlbertForSequenceClassification, __webpack_exports__AlbertModel as AlbertModel, __webpack_exports__AlbertPreTrainedModel as AlbertPreTrainedModel, __webpack_exports__AlbertTokenizer as AlbertTokenizer, __webpack_exports__AudioClassificationPipeline as AudioClassificationPipeline, __webpack_exports__AutoConfig as AutoConfig, __webpack_exports__AutoFeatureExtractor as AutoFeatureExtractor, __webpack_exports__AutoImageProcessor as AutoImageProcessor, __webpack_exports__AutoModel as AutoModel, __webpack_exports__AutoModelForAudioClassification as AutoModelForAudioClassification, __webpack_exports__AutoModelForAudioFrameClassification as AutoModelForAudioFrameClassification, __webpack_exports__AutoModelForCTC as AutoModelForCTC, __webpack_exports__AutoModelForCausalLM as AutoModelForCausalLM, __webpack_exports__AutoModelForDepthEstimation as AutoModelForDepthEstimation, __webpack_exports__AutoModelForDocumentQuestionAnswering as AutoModelForDocumentQuestionAnswering, __webpack_exports__AutoModelForImageClassification as AutoModelForImageClassification, __webpack_exports__AutoModelForImageFeatureExtraction as AutoModelForImageFeatureExtraction, __webpack_exports__AutoModelForImageMatting as AutoModelForImageMatting, __webpack_exports__AutoModelForImageSegmentation as AutoModelForImageSegmentation, __webpack_exports__AutoModelForImageToImage as AutoModelForImageToImage, __webpack_exports__AutoModelForMaskGeneration as AutoModelForMaskGeneration, __webpack_exports__AutoModelForMaskedLM as AutoModelForMaskedLM, __webpack_exports__AutoModelForNormalEstimation as AutoModelForNormalEstimation, __webpack_exports__AutoModelForObjectDetection as AutoModelForObjectDetection, __webpack_exports__AutoModelForPoseEstimation as AutoModelForPoseEstimation, __webpack_exports__AutoModelForQuestionAnswering as AutoModelForQuestionAnswering, __webpack_exports__AutoModelForSemanticSegmentation as AutoModelForSemanticSegmentation, __webpack_exports__AutoModelForSeq2SeqLM as AutoModelForSeq2SeqLM, __webpack_exports__AutoModelForSequenceClassification as AutoModelForSequenceClassification, __webpack_exports__AutoModelForSpeechSeq2Seq as AutoModelForSpeechSeq2Seq, __webpack_exports__AutoModelForTextToSpectrogram as AutoModelForTextToSpectrogram, __webpack_exports__AutoModelForTextToWaveform as AutoModelForTextToWaveform, __webpack_exports__AutoModelForTokenClassification as AutoModelForTokenClassification, __webpack_exports__AutoModelForUniversalSegmentation as AutoModelForUniversalSegmentation, __webpack_exports__AutoModelForVision2Seq as AutoModelForVision2Seq, __webpack_exports__AutoModelForXVector as AutoModelForXVector, __webpack_exports__AutoModelForZeroShotObjectDetection as AutoModelForZeroShotObjectDetection, __webpack_exports__AutoProcessor as AutoProcessor, __webpack_exports__AutoTokenizer as AutoTokenizer, __webpack_exports__AutomaticSpeechRecognitionPipeline as AutomaticSpeechRecognitionPipeline, __webpack_exports__BartForConditionalGeneration as BartForConditionalGeneration, __webpack_exports__BartForSequenceClassification as BartForSequenceClassification, __webpack_exports__BartModel as BartModel, __webpack_exports__BartPretrainedModel as BartPretrainedModel, __webpack_exports__BartTokenizer as BartTokenizer, __webpack_exports__BaseModelOutput as BaseModelOutput, __webpack_exports__BaseStreamer as BaseStreamer, __webpack_exports__BeitFeatureExtractor as BeitFeatureExtractor, __webpack_exports__BeitForImageClassification as BeitForImageClassification, __webpack_exports__BeitModel as BeitModel, __webpack_exports__BeitPreTrainedModel as BeitPreTrainedModel, __webpack_exports__BertForMaskedLM as BertForMaskedLM, __webpack_exports__BertForQuestionAnswering as BertForQuestionAnswering, __webpack_exports__BertForSequenceClassification as BertForSequenceClassification, __webpack_exports__BertForTokenClassification as BertForTokenClassification, __webpack_exports__BertModel as BertModel, __webpack_exports__BertPreTrainedModel as BertPreTrainedModel, __webpack_exports__BertTokenizer as BertTokenizer, __webpack_exports__BitImageProcessor as BitImageProcessor, __webpack_exports__BlenderbotForConditionalGeneration as BlenderbotForConditionalGeneration, __webpack_exports__BlenderbotModel as BlenderbotModel, __webpack_exports__BlenderbotPreTrainedModel as BlenderbotPreTrainedModel, __webpack_exports__BlenderbotSmallForConditionalGeneration as BlenderbotSmallForConditionalGeneration, __webpack_exports__BlenderbotSmallModel as BlenderbotSmallModel, __webpack_exports__BlenderbotSmallPreTrainedModel as BlenderbotSmallPreTrainedModel, __webpack_exports__BlenderbotSmallTokenizer as BlenderbotSmallTokenizer, __webpack_exports__BlenderbotTokenizer as BlenderbotTokenizer, __webpack_exports__BloomForCausalLM as BloomForCausalLM, __webpack_exports__BloomModel as BloomModel, __webpack_exports__BloomPreTrainedModel as BloomPreTrainedModel, __webpack_exports__BloomTokenizer as BloomTokenizer, __webpack_exports__CLIPFeatureExtractor as CLIPFeatureExtractor, __webpack_exports__CLIPImageProcessor as CLIPImageProcessor, __webpack_exports__CLIPModel as CLIPModel, __webpack_exports__CLIPPreTrainedModel as CLIPPreTrainedModel, __webpack_exports__CLIPSegForImageSegmentation as CLIPSegForImageSegmentation, __webpack_exports__CLIPSegModel as CLIPSegModel, __webpack_exports__CLIPSegPreTrainedModel as CLIPSegPreTrainedModel, __webpack_exports__CLIPTextModel as CLIPTextModel, __webpack_exports__CLIPTextModelWithProjection as CLIPTextModelWithProjection, __webpack_exports__CLIPTokenizer as CLIPTokenizer, __webpack_exports__CLIPVisionModel as CLIPVisionModel, __webpack_exports__CLIPVisionModelWithProjection as CLIPVisionModelWithProjection, __webpack_exports__CamembertForMaskedLM as CamembertForMaskedLM, __webpack_exports__CamembertForQuestionAnswering as CamembertForQuestionAnswering, __webpack_exports__CamembertForSequenceClassification as CamembertForSequenceClassification, __webpack_exports__CamembertForTokenClassification as CamembertForTokenClassification, __webpack_exports__CamembertModel as CamembertModel, __webpack_exports__CamembertPreTrainedModel as CamembertPreTrainedModel, __webpack_exports__CamembertTokenizer as CamembertTokenizer, __webpack_exports__CausalLMOutput as CausalLMOutput, __webpack_exports__CausalLMOutputWithPast as CausalLMOutputWithPast, __webpack_exports__ChineseCLIPFeatureExtractor as ChineseCLIPFeatureExtractor, __webpack_exports__ChineseCLIPModel as ChineseCLIPModel, __webpack_exports__ChineseCLIPPreTrainedModel as ChineseCLIPPreTrainedModel, __webpack_exports__ClapAudioModelWithProjection as ClapAudioModelWithProjection, __webpack_exports__ClapFeatureExtractor as ClapFeatureExtractor, __webpack_exports__ClapModel as ClapModel, __webpack_exports__ClapPreTrainedModel as ClapPreTrainedModel, __webpack_exports__ClapTextModelWithProjection as ClapTextModelWithProjection, __webpack_exports__ClassifierFreeGuidanceLogitsProcessor as ClassifierFreeGuidanceLogitsProcessor, __webpack_exports__CodeGenForCausalLM as CodeGenForCausalLM, __webpack_exports__CodeGenModel as CodeGenModel, __webpack_exports__CodeGenPreTrainedModel as CodeGenPreTrainedModel, __webpack_exports__CodeGenTokenizer as CodeGenTokenizer, __webpack_exports__CodeLlamaTokenizer as CodeLlamaTokenizer, __webpack_exports__CohereForCausalLM as CohereForCausalLM, __webpack_exports__CohereModel as CohereModel, __webpack_exports__CoherePreTrainedModel as CoherePreTrainedModel, __webpack_exports__CohereTokenizer as CohereTokenizer, __webpack_exports__ConvBertForMaskedLM as ConvBertForMaskedLM, __webpack_exports__ConvBertForQuestionAnswering as ConvBertForQuestionAnswering, __webpack_exports__ConvBertForSequenceClassification as ConvBertForSequenceClassification, __webpack_exports__ConvBertForTokenClassification as ConvBertForTokenClassification, __webpack_exports__ConvBertModel as ConvBertModel, __webpack_exports__ConvBertPreTrainedModel as ConvBertPreTrainedModel, __webpack_exports__ConvBertTokenizer as ConvBertTokenizer, __webpack_exports__ConvNextFeatureExtractor as ConvNextFeatureExtractor, __webpack_exports__ConvNextForImageClassification as ConvNextForImageClassification, __webpack_exports__ConvNextImageProcessor as ConvNextImageProcessor, __webpack_exports__ConvNextModel as ConvNextModel, __webpack_exports__ConvNextPreTrainedModel as ConvNextPreTrainedModel, __webpack_exports__ConvNextV2ForImageClassification as ConvNextV2ForImageClassification, __webpack_exports__ConvNextV2Model as ConvNextV2Model, __webpack_exports__ConvNextV2PreTrainedModel as ConvNextV2PreTrainedModel, __webpack_exports__DPTFeatureExtractor as DPTFeatureExtractor, __webpack_exports__DPTForDepthEstimation as DPTForDepthEstimation, __webpack_exports__DPTImageProcessor as DPTImageProcessor, __webpack_exports__DPTModel as DPTModel, __webpack_exports__DPTPreTrainedModel as DPTPreTrainedModel, __webpack_exports__DebertaForMaskedLM as DebertaForMaskedLM, __webpack_exports__DebertaForQuestionAnswering as DebertaForQuestionAnswering, __webpack_exports__DebertaForSequenceClassification as DebertaForSequenceClassification, __webpack_exports__DebertaForTokenClassification as DebertaForTokenClassification, __webpack_exports__DebertaModel as DebertaModel, __webpack_exports__DebertaPreTrainedModel as DebertaPreTrainedModel, __webpack_exports__DebertaTokenizer as DebertaTokenizer, __webpack_exports__DebertaV2ForMaskedLM as DebertaV2ForMaskedLM, __webpack_exports__DebertaV2ForQuestionAnswering as DebertaV2ForQuestionAnswering, __webpack_exports__DebertaV2ForSequenceClassification as DebertaV2ForSequenceClassification, __webpack_exports__DebertaV2ForTokenClassification as DebertaV2ForTokenClassification, __webpack_exports__DebertaV2Model as DebertaV2Model, __webpack_exports__DebertaV2PreTrainedModel as DebertaV2PreTrainedModel, __webpack_exports__DebertaV2Tokenizer as DebertaV2Tokenizer, __webpack_exports__DecisionTransformerModel as DecisionTransformerModel, __webpack_exports__DecisionTransformerPreTrainedModel as DecisionTransformerPreTrainedModel, __webpack_exports__DeiTFeatureExtractor as DeiTFeatureExtractor, __webpack_exports__DeiTForImageClassification as DeiTForImageClassification, __webpack_exports__DeiTImageProcessor as DeiTImageProcessor, __webpack_exports__DeiTModel as DeiTModel, __webpack_exports__DeiTPreTrainedModel as DeiTPreTrainedModel, __webpack_exports__DepthAnythingForDepthEstimation as DepthAnythingForDepthEstimation, __webpack_exports__DepthAnythingPreTrainedModel as DepthAnythingPreTrainedModel, __webpack_exports__DepthEstimationPipeline as DepthEstimationPipeline, __webpack_exports__DepthProForDepthEstimation as DepthProForDepthEstimation, __webpack_exports__DepthProPreTrainedModel as DepthProPreTrainedModel, __webpack_exports__DetrFeatureExtractor as DetrFeatureExtractor, __webpack_exports__DetrForObjectDetection as DetrForObjectDetection, __webpack_exports__DetrForSegmentation as DetrForSegmentation, __webpack_exports__DetrImageProcessor as DetrImageProcessor, __webpack_exports__DetrModel as DetrModel, __webpack_exports__DetrObjectDetectionOutput as DetrObjectDetectionOutput, __webpack_exports__DetrPreTrainedModel as DetrPreTrainedModel, __webpack_exports__DetrSegmentationOutput as DetrSegmentationOutput, __webpack_exports__Dinov2ForImageClassification as Dinov2ForImageClassification, __webpack_exports__Dinov2Model as Dinov2Model, __webpack_exports__Dinov2PreTrainedModel as Dinov2PreTrainedModel, __webpack_exports__DistilBertForMaskedLM as DistilBertForMaskedLM, __webpack_exports__DistilBertForQuestionAnswering as DistilBertForQuestionAnswering, __webpack_exports__DistilBertForSequenceClassification as DistilBertForSequenceClassification, __webpack_exports__DistilBertForTokenClassification as DistilBertForTokenClassification, __webpack_exports__DistilBertModel as DistilBertModel, __webpack_exports__DistilBertPreTrainedModel as DistilBertPreTrainedModel, __webpack_exports__DistilBertTokenizer as DistilBertTokenizer, __webpack_exports__DocumentQuestionAnsweringPipeline as DocumentQuestionAnsweringPipeline, __webpack_exports__DonutFeatureExtractor as DonutFeatureExtractor, __webpack_exports__DonutImageProcessor as DonutImageProcessor, __webpack_exports__DonutSwinModel as DonutSwinModel, __webpack_exports__DonutSwinPreTrainedModel as DonutSwinPreTrainedModel, __webpack_exports__EfficientNetForImageClassification as EfficientNetForImageClassification, __webpack_exports__EfficientNetImageProcessor as EfficientNetImageProcessor, __webpack_exports__EfficientNetModel as EfficientNetModel, __webpack_exports__EfficientNetPreTrainedModel as EfficientNetPreTrainedModel, __webpack_exports__ElectraForMaskedLM as ElectraForMaskedLM, __webpack_exports__ElectraForQuestionAnswering as ElectraForQuestionAnswering, __webpack_exports__ElectraForSequenceClassification as ElectraForSequenceClassification, __webpack_exports__ElectraForTokenClassification as ElectraForTokenClassification, __webpack_exports__ElectraModel as ElectraModel, __webpack_exports__ElectraPreTrainedModel as ElectraPreTrainedModel, __webpack_exports__ElectraTokenizer as ElectraTokenizer, __webpack_exports__EosTokenCriteria as EosTokenCriteria, __webpack_exports__EsmForMaskedLM as EsmForMaskedLM, __webpack_exports__EsmForSequenceClassification as EsmForSequenceClassification, __webpack_exports__EsmForTokenClassification as EsmForTokenClassification, __webpack_exports__EsmModel as EsmModel, __webpack_exports__EsmPreTrainedModel as EsmPreTrainedModel, __webpack_exports__EsmTokenizer as EsmTokenizer, __webpack_exports__FFT as FFT, __webpack_exports__FalconForCausalLM as FalconForCausalLM, __webpack_exports__FalconModel as FalconModel, __webpack_exports__FalconPreTrainedModel as FalconPreTrainedModel, __webpack_exports__FalconTokenizer as FalconTokenizer, __webpack_exports__FastViTForImageClassification as FastViTForImageClassification, __webpack_exports__FastViTModel as FastViTModel, __webpack_exports__FastViTPreTrainedModel as FastViTPreTrainedModel, __webpack_exports__FeatureExtractionPipeline as FeatureExtractionPipeline, __webpack_exports__FeatureExtractor as FeatureExtractor, __webpack_exports__FillMaskPipeline as FillMaskPipeline, __webpack_exports__Florence2ForConditionalGeneration as Florence2ForConditionalGeneration, __webpack_exports__Florence2PreTrainedModel as Florence2PreTrainedModel, __webpack_exports__Florence2Processor as Florence2Processor, __webpack_exports__ForcedBOSTokenLogitsProcessor as ForcedBOSTokenLogitsProcessor, __webpack_exports__ForcedEOSTokenLogitsProcessor as ForcedEOSTokenLogitsProcessor, __webpack_exports__GLPNFeatureExtractor as GLPNFeatureExtractor, __webpack_exports__GLPNForDepthEstimation as GLPNForDepthEstimation, __webpack_exports__GLPNModel as GLPNModel, __webpack_exports__GLPNPreTrainedModel as GLPNPreTrainedModel, __webpack_exports__GPT2LMHeadModel as GPT2LMHeadModel, __webpack_exports__GPT2Model as GPT2Model, __webpack_exports__GPT2PreTrainedModel as GPT2PreTrainedModel, __webpack_exports__GPT2Tokenizer as GPT2Tokenizer, __webpack_exports__GPTBigCodeForCausalLM as GPTBigCodeForCausalLM, __webpack_exports__GPTBigCodeModel as GPTBigCodeModel, __webpack_exports__GPTBigCodePreTrainedModel as GPTBigCodePreTrainedModel, __webpack_exports__GPTJForCausalLM as GPTJForCausalLM, __webpack_exports__GPTJModel as GPTJModel, __webpack_exports__GPTJPreTrainedModel as GPTJPreTrainedModel, __webpack_exports__GPTNeoForCausalLM as GPTNeoForCausalLM, __webpack_exports__GPTNeoModel as GPTNeoModel, __webpack_exports__GPTNeoPreTrainedModel as GPTNeoPreTrainedModel, __webpack_exports__GPTNeoXForCausalLM as GPTNeoXForCausalLM, __webpack_exports__GPTNeoXModel as GPTNeoXModel, __webpack_exports__GPTNeoXPreTrainedModel as GPTNeoXPreTrainedModel, __webpack_exports__GPTNeoXTokenizer as GPTNeoXTokenizer, __webpack_exports__Gemma2ForCausalLM as Gemma2ForCausalLM, __webpack_exports__Gemma2Model as Gemma2Model, __webpack_exports__Gemma2PreTrainedModel as Gemma2PreTrainedModel, __webpack_exports__GemmaForCausalLM as GemmaForCausalLM, __webpack_exports__GemmaModel as GemmaModel, __webpack_exports__GemmaPreTrainedModel as GemmaPreTrainedModel, __webpack_exports__GemmaTokenizer as GemmaTokenizer, __webpack_exports__GraniteForCausalLM as GraniteForCausalLM, __webpack_exports__GraniteModel as GraniteModel, __webpack_exports__GranitePreTrainedModel as GranitePreTrainedModel, __webpack_exports__Grok1Tokenizer as Grok1Tokenizer, __webpack_exports__GroupViTModel as GroupViTModel, __webpack_exports__GroupViTPreTrainedModel as GroupViTPreTrainedModel, __webpack_exports__HerbertTokenizer as HerbertTokenizer, __webpack_exports__HieraForImageClassification as HieraForImageClassification, __webpack_exports__HieraModel as HieraModel, __webpack_exports__HieraPreTrainedModel as HieraPreTrainedModel, __webpack_exports__HubertForCTC as HubertForCTC, __webpack_exports__HubertForSequenceClassification as HubertForSequenceClassification, __webpack_exports__HubertModel as HubertModel, __webpack_exports__HubertPreTrainedModel as HubertPreTrainedModel, __webpack_exports__Idefics3ForConditionalGeneration as Idefics3ForConditionalGeneration, __webpack_exports__Idefics3ImageProcessor as Idefics3ImageProcessor, __webpack_exports__Idefics3PreTrainedModel as Idefics3PreTrainedModel, __webpack_exports__Idefics3Processor as Idefics3Processor, __webpack_exports__ImageClassificationPipeline as ImageClassificationPipeline, __webpack_exports__ImageFeatureExtractionPipeline as ImageFeatureExtractionPipeline, __webpack_exports__ImageFeatureExtractor as ImageFeatureExtractor, __webpack_exports__ImageMattingOutput as ImageMattingOutput, __webpack_exports__ImageProcessor as ImageProcessor, __webpack_exports__ImageSegmentationPipeline as ImageSegmentationPipeline, __webpack_exports__ImageToImagePipeline as ImageToImagePipeline, __webpack_exports__ImageToTextPipeline as ImageToTextPipeline, __webpack_exports__InterruptableStoppingCriteria as InterruptableStoppingCriteria, __webpack_exports__JAISLMHeadModel as JAISLMHeadModel, __webpack_exports__JAISModel as JAISModel, __webpack_exports__JAISPreTrainedModel as JAISPreTrainedModel, __webpack_exports__JinaCLIPImageProcessor as JinaCLIPImageProcessor, __webpack_exports__JinaCLIPModel as JinaCLIPModel, __webpack_exports__JinaCLIPPreTrainedModel as JinaCLIPPreTrainedModel, __webpack_exports__JinaCLIPProcessor as JinaCLIPProcessor, __webpack_exports__JinaCLIPTextModel as JinaCLIPTextModel, __webpack_exports__JinaCLIPVisionModel as JinaCLIPVisionModel, __webpack_exports__LlamaForCausalLM as LlamaForCausalLM, __webpack_exports__LlamaModel as LlamaModel, __webpack_exports__LlamaPreTrainedModel as LlamaPreTrainedModel, __webpack_exports__LlamaTokenizer as LlamaTokenizer, __webpack_exports__LlavaForConditionalGeneration as LlavaForConditionalGeneration, __webpack_exports__LlavaOnevisionForConditionalGeneration as LlavaOnevisionForConditionalGeneration, __webpack_exports__LlavaOnevisionImageProcessor as LlavaOnevisionImageProcessor, __webpack_exports__LlavaPreTrainedModel as LlavaPreTrainedModel, __webpack_exports__LogitsProcessor as LogitsProcessor, __webpack_exports__LogitsProcessorList as LogitsProcessorList, __webpack_exports__LogitsWarper as LogitsWarper, __webpack_exports__LongT5ForConditionalGeneration as LongT5ForConditionalGeneration, __webpack_exports__LongT5Model as LongT5Model, __webpack_exports__LongT5PreTrainedModel as LongT5PreTrainedModel, __webpack_exports__M2M100ForConditionalGeneration as M2M100ForConditionalGeneration, __webpack_exports__M2M100Model as M2M100Model, __webpack_exports__M2M100PreTrainedModel as M2M100PreTrainedModel, __webpack_exports__M2M100Tokenizer as M2M100Tokenizer, __webpack_exports__MBart50Tokenizer as MBart50Tokenizer, __webpack_exports__MBartForCausalLM as MBartForCausalLM, __webpack_exports__MBartForConditionalGeneration as MBartForConditionalGeneration, __webpack_exports__MBartForSequenceClassification as MBartForSequenceClassification, __webpack_exports__MBartModel as MBartModel, __webpack_exports__MBartPreTrainedModel as MBartPreTrainedModel, __webpack_exports__MBartTokenizer as MBartTokenizer, __webpack_exports__MPNetForMaskedLM as MPNetForMaskedLM, __webpack_exports__MPNetForQuestionAnswering as MPNetForQuestionAnswering, __webpack_exports__MPNetForSequenceClassification as MPNetForSequenceClassification, __webpack_exports__MPNetForTokenClassification as MPNetForTokenClassification, __webpack_exports__MPNetModel as MPNetModel, __webpack_exports__MPNetPreTrainedModel as MPNetPreTrainedModel, __webpack_exports__MPNetTokenizer as MPNetTokenizer, __webpack_exports__MT5ForConditionalGeneration as MT5ForConditionalGeneration, __webpack_exports__MT5Model as MT5Model, __webpack_exports__MT5PreTrainedModel as MT5PreTrainedModel, __webpack_exports__MarianMTModel as MarianMTModel, __webpack_exports__MarianModel as MarianModel, __webpack_exports__MarianPreTrainedModel as MarianPreTrainedModel, __webpack_exports__MarianTokenizer as MarianTokenizer, __webpack_exports__Mask2FormerImageProcessor as Mask2FormerImageProcessor, __webpack_exports__MaskFormerFeatureExtractor as MaskFormerFeatureExtractor, __webpack_exports__MaskFormerForInstanceSegmentation as MaskFormerForInstanceSegmentation, __webpack_exports__MaskFormerImageProcessor as MaskFormerImageProcessor, __webpack_exports__MaskFormerModel as MaskFormerModel, __webpack_exports__MaskFormerPreTrainedModel as MaskFormerPreTrainedModel, __webpack_exports__MaskedLMOutput as MaskedLMOutput, __webpack_exports__MaxLengthCriteria as MaxLengthCriteria, __webpack_exports__MgpstrForSceneTextRecognition as MgpstrForSceneTextRecognition, __webpack_exports__MgpstrModelOutput as MgpstrModelOutput, __webpack_exports__MgpstrPreTrainedModel as MgpstrPreTrainedModel, __webpack_exports__MgpstrProcessor as MgpstrProcessor, __webpack_exports__MgpstrTokenizer as MgpstrTokenizer, __webpack_exports__MinLengthLogitsProcessor as MinLengthLogitsProcessor, __webpack_exports__MinNewTokensLengthLogitsProcessor as MinNewTokensLengthLogitsProcessor, __webpack_exports__MistralForCausalLM as MistralForCausalLM, __webpack_exports__MistralModel as MistralModel, __webpack_exports__MistralPreTrainedModel as MistralPreTrainedModel, __webpack_exports__MobileBertForMaskedLM as MobileBertForMaskedLM, __webpack_exports__MobileBertForQuestionAnswering as MobileBertForQuestionAnswering, __webpack_exports__MobileBertForSequenceClassification as MobileBertForSequenceClassification, __webpack_exports__MobileBertModel as MobileBertModel, __webpack_exports__MobileBertPreTrainedModel as MobileBertPreTrainedModel, __webpack_exports__MobileBertTokenizer as MobileBertTokenizer, __webpack_exports__MobileLLMForCausalLM as MobileLLMForCausalLM, __webpack_exports__MobileLLMModel as MobileLLMModel, __webpack_exports__MobileLLMPreTrainedModel as MobileLLMPreTrainedModel, __webpack_exports__MobileNetV1FeatureExtractor as MobileNetV1FeatureExtractor, __webpack_exports__MobileNetV1ForImageClassification as MobileNetV1ForImageClassification, __webpack_exports__MobileNetV1ImageProcessor as MobileNetV1ImageProcessor, __webpack_exports__MobileNetV1Model as MobileNetV1Model, __webpack_exports__MobileNetV1PreTrainedModel as MobileNetV1PreTrainedModel, __webpack_exports__MobileNetV2FeatureExtractor as MobileNetV2FeatureExtractor, __webpack_exports__MobileNetV2ForImageClassification as MobileNetV2ForImageClassification, __webpack_exports__MobileNetV2ImageProcessor as MobileNetV2ImageProcessor, __webpack_exports__MobileNetV2Model as MobileNetV2Model, __webpack_exports__MobileNetV2PreTrainedModel as MobileNetV2PreTrainedModel, __webpack_exports__MobileNetV3FeatureExtractor as MobileNetV3FeatureExtractor, __webpack_exports__MobileNetV3ForImageClassification as MobileNetV3ForImageClassification, __webpack_exports__MobileNetV3ImageProcessor as MobileNetV3ImageProcessor, __webpack_exports__MobileNetV3Model as MobileNetV3Model, __webpack_exports__MobileNetV3PreTrainedModel as MobileNetV3PreTrainedModel, __webpack_exports__MobileNetV4FeatureExtractor as MobileNetV4FeatureExtractor, __webpack_exports__MobileNetV4ForImageClassification as MobileNetV4ForImageClassification, __webpack_exports__MobileNetV4ImageProcessor as MobileNetV4ImageProcessor, __webpack_exports__MobileNetV4Model as MobileNetV4Model, __webpack_exports__MobileNetV4PreTrainedModel as MobileNetV4PreTrainedModel, __webpack_exports__MobileViTFeatureExtractor as MobileViTFeatureExtractor, __webpack_exports__MobileViTForImageClassification as MobileViTForImageClassification, __webpack_exports__MobileViTImageProcessor as MobileViTImageProcessor, __webpack_exports__MobileViTModel as MobileViTModel, __webpack_exports__MobileViTPreTrainedModel as MobileViTPreTrainedModel, __webpack_exports__MobileViTV2ForImageClassification as MobileViTV2ForImageClassification, __webpack_exports__MobileViTV2Model as MobileViTV2Model, __webpack_exports__MobileViTV2PreTrainedModel as MobileViTV2PreTrainedModel, __webpack_exports__ModelOutput as ModelOutput, __webpack_exports__Moondream1ForConditionalGeneration as Moondream1ForConditionalGeneration, __webpack_exports__MptForCausalLM as MptForCausalLM, __webpack_exports__MptModel as MptModel, __webpack_exports__MptPreTrainedModel as MptPreTrainedModel, __webpack_exports__MultiModalityCausalLM as MultiModalityCausalLM, __webpack_exports__MultiModalityPreTrainedModel as MultiModalityPreTrainedModel, __webpack_exports__MusicgenForCausalLM as MusicgenForCausalLM, __webpack_exports__MusicgenForConditionalGeneration as MusicgenForConditionalGeneration, __webpack_exports__MusicgenModel as MusicgenModel, __webpack_exports__MusicgenPreTrainedModel as MusicgenPreTrainedModel, __webpack_exports__NllbTokenizer as NllbTokenizer, __webpack_exports__NoBadWordsLogitsProcessor as NoBadWordsLogitsProcessor, __webpack_exports__NoRepeatNGramLogitsProcessor as NoRepeatNGramLogitsProcessor, __webpack_exports__NomicBertModel as NomicBertModel, __webpack_exports__NomicBertPreTrainedModel as NomicBertPreTrainedModel, __webpack_exports__NougatImageProcessor as NougatImageProcessor, __webpack_exports__NougatTokenizer as NougatTokenizer, __webpack_exports__OPTForCausalLM as OPTForCausalLM, __webpack_exports__OPTModel as OPTModel, __webpack_exports__OPTPreTrainedModel as OPTPreTrainedModel, __webpack_exports__ObjectDetectionPipeline as ObjectDetectionPipeline, __webpack_exports__OlmoForCausalLM as OlmoForCausalLM, __webpack_exports__OlmoModel as OlmoModel, __webpack_exports__OlmoPreTrainedModel as OlmoPreTrainedModel, __webpack_exports__OpenELMForCausalLM as OpenELMForCausalLM, __webpack_exports__OpenELMModel as OpenELMModel, __webpack_exports__OpenELMPreTrainedModel as OpenELMPreTrainedModel, __webpack_exports__OwlViTFeatureExtractor as OwlViTFeatureExtractor, __webpack_exports__OwlViTForObjectDetection as OwlViTForObjectDetection, __webpack_exports__OwlViTImageProcessor as OwlViTImageProcessor, __webpack_exports__OwlViTModel as OwlViTModel, __webpack_exports__OwlViTPreTrainedModel as OwlViTPreTrainedModel, __webpack_exports__OwlViTProcessor as OwlViTProcessor, __webpack_exports__Owlv2ForObjectDetection as Owlv2ForObjectDetection, __webpack_exports__Owlv2ImageProcessor as Owlv2ImageProcessor, __webpack_exports__Owlv2Model as Owlv2Model, __webpack_exports__Owlv2PreTrainedModel as Owlv2PreTrainedModel, __webpack_exports__PatchTSMixerForPrediction as PatchTSMixerForPrediction, __webpack_exports__PatchTSMixerModel as PatchTSMixerModel, __webpack_exports__PatchTSMixerPreTrainedModel as PatchTSMixerPreTrainedModel, __webpack_exports__PatchTSTForPrediction as PatchTSTForPrediction, __webpack_exports__PatchTSTModel as PatchTSTModel, __webpack_exports__PatchTSTPreTrainedModel as PatchTSTPreTrainedModel, __webpack_exports__Phi3ForCausalLM as Phi3ForCausalLM, __webpack_exports__Phi3Model as Phi3Model, __webpack_exports__Phi3PreTrainedModel as Phi3PreTrainedModel, __webpack_exports__PhiForCausalLM as PhiForCausalLM, __webpack_exports__PhiModel as PhiModel, __webpack_exports__PhiPreTrainedModel as PhiPreTrainedModel, __webpack_exports__Pipeline as Pipeline, __webpack_exports__PreTrainedModel as PreTrainedModel, __webpack_exports__PreTrainedTokenizer as PreTrainedTokenizer, __webpack_exports__PretrainedConfig as PretrainedConfig, __webpack_exports__PretrainedMixin as PretrainedMixin, __webpack_exports__Processor as Processor, __webpack_exports__PvtForImageClassification as PvtForImageClassification, __webpack_exports__PvtImageProcessor as PvtImageProcessor, __webpack_exports__PvtModel as PvtModel, __webpack_exports__PvtPreTrainedModel as PvtPreTrainedModel, __webpack_exports__PyAnnoteFeatureExtractor as PyAnnoteFeatureExtractor, __webpack_exports__PyAnnoteForAudioFrameClassification as PyAnnoteForAudioFrameClassification, __webpack_exports__PyAnnoteModel as PyAnnoteModel, __webpack_exports__PyAnnotePreTrainedModel as PyAnnotePreTrainedModel, __webpack_exports__PyAnnoteProcessor as PyAnnoteProcessor, __webpack_exports__QuestionAnsweringModelOutput as QuestionAnsweringModelOutput, __webpack_exports__QuestionAnsweringPipeline as QuestionAnsweringPipeline, __webpack_exports__Qwen2ForCausalLM as Qwen2ForCausalLM, __webpack_exports__Qwen2Model as Qwen2Model, __webpack_exports__Qwen2PreTrainedModel as Qwen2PreTrainedModel, __webpack_exports__Qwen2Tokenizer as Qwen2Tokenizer, __webpack_exports__Qwen2VLForConditionalGeneration as Qwen2VLForConditionalGeneration, __webpack_exports__Qwen2VLImageProcessor as Qwen2VLImageProcessor, __webpack_exports__Qwen2VLPreTrainedModel as Qwen2VLPreTrainedModel, __webpack_exports__Qwen2VLProcessor as Qwen2VLProcessor, __webpack_exports__RTDetrForObjectDetection as RTDetrForObjectDetection, __webpack_exports__RTDetrImageProcessor as RTDetrImageProcessor, __webpack_exports__RTDetrModel as RTDetrModel, __webpack_exports__RTDetrObjectDetectionOutput as RTDetrObjectDetectionOutput, __webpack_exports__RTDetrPreTrainedModel as RTDetrPreTrainedModel, __webpack_exports__RawImage as RawImage, __webpack_exports__RepetitionPenaltyLogitsProcessor as RepetitionPenaltyLogitsProcessor, __webpack_exports__ResNetForImageClassification as ResNetForImageClassification, __webpack_exports__ResNetModel as ResNetModel, __webpack_exports__ResNetPreTrainedModel as ResNetPreTrainedModel, __webpack_exports__RoFormerForMaskedLM as RoFormerForMaskedLM, __webpack_exports__RoFormerForQuestionAnswering as RoFormerForQuestionAnswering, __webpack_exports__RoFormerForSequenceClassification as RoFormerForSequenceClassification, __webpack_exports__RoFormerForTokenClassification as RoFormerForTokenClassification, __webpack_exports__RoFormerModel as RoFormerModel, __webpack_exports__RoFormerPreTrainedModel as RoFormerPreTrainedModel, __webpack_exports__RoFormerTokenizer as RoFormerTokenizer, __webpack_exports__RobertaForMaskedLM as RobertaForMaskedLM, __webpack_exports__RobertaForQuestionAnswering as RobertaForQuestionAnswering, __webpack_exports__RobertaForSequenceClassification as RobertaForSequenceClassification, __webpack_exports__RobertaForTokenClassification as RobertaForTokenClassification, __webpack_exports__RobertaModel as RobertaModel, __webpack_exports__RobertaPreTrainedModel as RobertaPreTrainedModel, __webpack_exports__RobertaTokenizer as RobertaTokenizer, __webpack_exports__SamImageProcessor as SamImageProcessor, __webpack_exports__SamImageSegmentationOutput as SamImageSegmentationOutput, __webpack_exports__SamModel as SamModel, __webpack_exports__SamPreTrainedModel as SamPreTrainedModel, __webpack_exports__SamProcessor as SamProcessor, __webpack_exports__SapiensForDepthEstimation as SapiensForDepthEstimation, __webpack_exports__SapiensForNormalEstimation as SapiensForNormalEstimation, __webpack_exports__SapiensForSemanticSegmentation as SapiensForSemanticSegmentation, __webpack_exports__SapiensPreTrainedModel as SapiensPreTrainedModel, __webpack_exports__SeamlessM4TFeatureExtractor as SeamlessM4TFeatureExtractor, __webpack_exports__SegformerFeatureExtractor as SegformerFeatureExtractor, __webpack_exports__SegformerForImageClassification as SegformerForImageClassification, __webpack_exports__SegformerForSemanticSegmentation as SegformerForSemanticSegmentation, __webpack_exports__SegformerImageProcessor as SegformerImageProcessor, __webpack_exports__SegformerModel as SegformerModel, __webpack_exports__SegformerPreTrainedModel as SegformerPreTrainedModel, __webpack_exports__Seq2SeqLMOutput as Seq2SeqLMOutput, __webpack_exports__SequenceClassifierOutput as SequenceClassifierOutput, __webpack_exports__SiglipImageProcessor as SiglipImageProcessor, __webpack_exports__SiglipModel as SiglipModel, __webpack_exports__SiglipPreTrainedModel as SiglipPreTrainedModel, __webpack_exports__SiglipTextModel as SiglipTextModel, __webpack_exports__SiglipTokenizer as SiglipTokenizer, __webpack_exports__SiglipVisionModel as SiglipVisionModel, __webpack_exports__SpeechT5FeatureExtractor as SpeechT5FeatureExtractor, __webpack_exports__SpeechT5ForSpeechToText as SpeechT5ForSpeechToText, __webpack_exports__SpeechT5ForTextToSpeech as SpeechT5ForTextToSpeech, __webpack_exports__SpeechT5HifiGan as SpeechT5HifiGan, __webpack_exports__SpeechT5Model as SpeechT5Model, __webpack_exports__SpeechT5PreTrainedModel as SpeechT5PreTrainedModel, __webpack_exports__SpeechT5Processor as SpeechT5Processor, __webpack_exports__SpeechT5Tokenizer as SpeechT5Tokenizer, __webpack_exports__SqueezeBertForMaskedLM as SqueezeBertForMaskedLM, __webpack_exports__SqueezeBertForQuestionAnswering as SqueezeBertForQuestionAnswering, __webpack_exports__SqueezeBertForSequenceClassification as SqueezeBertForSequenceClassification, __webpack_exports__SqueezeBertModel as SqueezeBertModel, __webpack_exports__SqueezeBertPreTrainedModel as SqueezeBertPreTrainedModel, __webpack_exports__SqueezeBertTokenizer as SqueezeBertTokenizer, __webpack_exports__StableLmForCausalLM as StableLmForCausalLM, __webpack_exports__StableLmModel as StableLmModel, __webpack_exports__StableLmPreTrainedModel as StableLmPreTrainedModel, __webpack_exports__Starcoder2ForCausalLM as Starcoder2ForCausalLM, __webpack_exports__Starcoder2Model as Starcoder2Model, __webpack_exports__Starcoder2PreTrainedModel as Starcoder2PreTrainedModel, __webpack_exports__StoppingCriteria as StoppingCriteria, __webpack_exports__StoppingCriteriaList as StoppingCriteriaList, __webpack_exports__SummarizationPipeline as SummarizationPipeline, __webpack_exports__SuppressTokensAtBeginLogitsProcessor as SuppressTokensAtBeginLogitsProcessor, __webpack_exports__Swin2SRForImageSuperResolution as Swin2SRForImageSuperResolution, __webpack_exports__Swin2SRImageProcessor as Swin2SRImageProcessor, __webpack_exports__Swin2SRModel as Swin2SRModel, __webpack_exports__Swin2SRPreTrainedModel as Swin2SRPreTrainedModel, __webpack_exports__SwinForImageClassification as SwinForImageClassification, __webpack_exports__SwinModel as SwinModel, __webpack_exports__SwinPreTrainedModel as SwinPreTrainedModel, __webpack_exports__T5ForConditionalGeneration as T5ForConditionalGeneration, __webpack_exports__T5Model as T5Model, __webpack_exports__T5PreTrainedModel as T5PreTrainedModel, __webpack_exports__T5Tokenizer as T5Tokenizer, __webpack_exports__TableTransformerForObjectDetection as TableTransformerForObjectDetection, __webpack_exports__TableTransformerModel as TableTransformerModel, __webpack_exports__TableTransformerObjectDetectionOutput as TableTransformerObjectDetectionOutput, __webpack_exports__TableTransformerPreTrainedModel as TableTransformerPreTrainedModel, __webpack_exports__TemperatureLogitsWarper as TemperatureLogitsWarper, __webpack_exports__Tensor as Tensor, __webpack_exports__Text2TextGenerationPipeline as Text2TextGenerationPipeline, __webpack_exports__TextClassificationPipeline as TextClassificationPipeline, __webpack_exports__TextGenerationPipeline as TextGenerationPipeline, __webpack_exports__TextStreamer as TextStreamer, __webpack_exports__TextToAudioPipeline as TextToAudioPipeline, __webpack_exports__TokenClassificationPipeline as TokenClassificationPipeline, __webpack_exports__TokenClassifierOutput as TokenClassifierOutput, __webpack_exports__TokenizerModel as TokenizerModel, __webpack_exports__TopKLogitsWarper as TopKLogitsWarper, __webpack_exports__TopPLogitsWarper as TopPLogitsWarper, __webpack_exports__TrOCRForCausalLM as TrOCRForCausalLM, __webpack_exports__TrOCRPreTrainedModel as TrOCRPreTrainedModel, __webpack_exports__TranslationPipeline as TranslationPipeline, __webpack_exports__UniSpeechForCTC as UniSpeechForCTC, __webpack_exports__UniSpeechForSequenceClassification as UniSpeechForSequenceClassification, __webpack_exports__UniSpeechModel as UniSpeechModel, __webpack_exports__UniSpeechPreTrainedModel as UniSpeechPreTrainedModel, __webpack_exports__UniSpeechSatForAudioFrameClassification as UniSpeechSatForAudioFrameClassification, __webpack_exports__UniSpeechSatForCTC as UniSpeechSatForCTC, __webpack_exports__UniSpeechSatForSequenceClassification as UniSpeechSatForSequenceClassification, __webpack_exports__UniSpeechSatModel as UniSpeechSatModel, __webpack_exports__UniSpeechSatPreTrainedModel as UniSpeechSatPreTrainedModel, __webpack_exports__VLChatProcessor as VLChatProcessor, __webpack_exports__VLMImageProcessor as VLMImageProcessor, __webpack_exports__ViTFeatureExtractor as ViTFeatureExtractor, __webpack_exports__ViTForImageClassification as ViTForImageClassification, __webpack_exports__ViTImageProcessor as ViTImageProcessor, __webpack_exports__ViTMAEModel as ViTMAEModel, __webpack_exports__ViTMAEPreTrainedModel as ViTMAEPreTrainedModel, __webpack_exports__ViTMSNForImageClassification as ViTMSNForImageClassification, __webpack_exports__ViTMSNModel as ViTMSNModel, __webpack_exports__ViTMSNPreTrainedModel as ViTMSNPreTrainedModel, __webpack_exports__ViTModel as ViTModel, __webpack_exports__ViTPreTrainedModel as ViTPreTrainedModel, __webpack_exports__VisionEncoderDecoderModel as VisionEncoderDecoderModel, __webpack_exports__VitMatteForImageMatting as VitMatteForImageMatting, __webpack_exports__VitMatteImageProcessor as VitMatteImageProcessor, __webpack_exports__VitMattePreTrainedModel as VitMattePreTrainedModel, __webpack_exports__VitPoseForPoseEstimation as VitPoseForPoseEstimation, __webpack_exports__VitPoseImageProcessor as VitPoseImageProcessor, __webpack_exports__VitPosePreTrainedModel as VitPosePreTrainedModel, __webpack_exports__VitsModel as VitsModel, __webpack_exports__VitsModelOutput as VitsModelOutput, __webpack_exports__VitsPreTrainedModel as VitsPreTrainedModel, __webpack_exports__VitsTokenizer as VitsTokenizer, __webpack_exports__Wav2Vec2BertForCTC as Wav2Vec2BertForCTC, __webpack_exports__Wav2Vec2BertForSequenceClassification as Wav2Vec2BertForSequenceClassification, __webpack_exports__Wav2Vec2BertModel as Wav2Vec2BertModel, __webpack_exports__Wav2Vec2BertPreTrainedModel as Wav2Vec2BertPreTrainedModel, __webpack_exports__Wav2Vec2CTCTokenizer as Wav2Vec2CTCTokenizer, __webpack_exports__Wav2Vec2FeatureExtractor as Wav2Vec2FeatureExtractor, __webpack_exports__Wav2Vec2ForAudioFrameClassification as Wav2Vec2ForAudioFrameClassification, __webpack_exports__Wav2Vec2ForCTC as Wav2Vec2ForCTC, __webpack_exports__Wav2Vec2ForSequenceClassification as Wav2Vec2ForSequenceClassification, __webpack_exports__Wav2Vec2Model as Wav2Vec2Model, __webpack_exports__Wav2Vec2PreTrainedModel as Wav2Vec2PreTrainedModel, __webpack_exports__Wav2Vec2ProcessorWithLM as Wav2Vec2ProcessorWithLM, __webpack_exports__WavLMForAudioFrameClassification as WavLMForAudioFrameClassification, __webpack_exports__WavLMForCTC as WavLMForCTC, __webpack_exports__WavLMForSequenceClassification as WavLMForSequenceClassification, __webpack_exports__WavLMForXVector as WavLMForXVector, __webpack_exports__WavLMModel as WavLMModel, __webpack_exports__WavLMPreTrainedModel as WavLMPreTrainedModel, __webpack_exports__WeSpeakerFeatureExtractor as WeSpeakerFeatureExtractor, __webpack_exports__WeSpeakerResNetModel as WeSpeakerResNetModel, __webpack_exports__WeSpeakerResNetPreTrainedModel as WeSpeakerResNetPreTrainedModel, __webpack_exports__WhisperFeatureExtractor as WhisperFeatureExtractor, __webpack_exports__WhisperForConditionalGeneration as WhisperForConditionalGeneration, __webpack_exports__WhisperModel as WhisperModel, __webpack_exports__WhisperPreTrainedModel as WhisperPreTrainedModel, __webpack_exports__WhisperProcessor as WhisperProcessor, __webpack_exports__WhisperTextStreamer as WhisperTextStreamer, __webpack_exports__WhisperTimeStampLogitsProcessor as WhisperTimeStampLogitsProcessor, __webpack_exports__WhisperTokenizer as WhisperTokenizer, __webpack_exports__XLMForQuestionAnswering as XLMForQuestionAnswering, __webpack_exports__XLMForSequenceClassification as XLMForSequenceClassification, __webpack_exports__XLMForTokenClassification as XLMForTokenClassification, __webpack_exports__XLMModel as XLMModel, __webpack_exports__XLMPreTrainedModel as XLMPreTrainedModel, __webpack_exports__XLMRobertaForMaskedLM as XLMRobertaForMaskedLM, __webpack_exports__XLMRobertaForQuestionAnswering as XLMRobertaForQuestionAnswering, __webpack_exports__XLMRobertaForSequenceClassification as XLMRobertaForSequenceClassification, __webpack_exports__XLMRobertaForTokenClassification as XLMRobertaForTokenClassification, __webpack_exports__XLMRobertaModel as XLMRobertaModel, __webpack_exports__XLMRobertaPreTrainedModel as XLMRobertaPreTrainedModel, __webpack_exports__XLMRobertaTokenizer as XLMRobertaTokenizer, __webpack_exports__XLMTokenizer as XLMTokenizer, __webpack_exports__XLMWithLMHeadModel as XLMWithLMHeadModel, __webpack_exports__XVectorOutput as XVectorOutput, __webpack_exports__YolosFeatureExtractor as YolosFeatureExtractor, __webpack_exports__YolosForObjectDetection as YolosForObjectDetection, __webpack_exports__YolosImageProcessor as YolosImageProcessor, __webpack_exports__YolosModel as YolosModel, __webpack_exports__YolosObjectDetectionOutput as YolosObjectDetectionOutput, __webpack_exports__YolosPreTrainedModel as YolosPreTrainedModel, __webpack_exports__ZeroShotAudioClassificationPipeline as ZeroShotAudioClassificationPipeline, __webpack_exports__ZeroShotClassificationPipeline as ZeroShotClassificationPipeline, __webpack_exports__ZeroShotImageClassificationPipeline as ZeroShotImageClassificationPipeline, __webpack_exports__ZeroShotObjectDetectionPipeline as ZeroShotObjectDetectionPipeline, __webpack_exports__bankers_round as bankers_round, __webpack_exports__cat as cat, __webpack_exports__cos_sim as cos_sim, __webpack_exports__dot as dot, __webpack_exports__dynamic_time_warping as dynamic_time_warping, __webpack_exports__env as env, __webpack_exports__full as full, __webpack_exports__full_like as full_like, __webpack_exports__getKeyValueShapes as getKeyValueShapes, __webpack_exports__hamming as hamming, __webpack_exports__hanning as hanning, __webpack_exports__interpolate as interpolate, __webpack_exports__interpolate_4d as interpolate_4d, __webpack_exports__interpolate_data as interpolate_data, __webpack_exports__is_chinese_char as is_chinese_char, __webpack_exports__layer_norm as layer_norm, __webpack_exports__load_image as load_image, __webpack_exports__log_softmax as log_softmax, __webpack_exports__magnitude as magnitude, __webpack_exports__matmul as matmul, __webpack_exports__max as max, __webpack_exports__mean as mean, __webpack_exports__mean_pooling as mean_pooling, __webpack_exports__medianFilter as medianFilter, __webpack_exports__mel_filter_bank as mel_filter_bank, __webpack_exports__min as min, __webpack_exports__ones as ones, __webpack_exports__ones_like as ones_like, __webpack_exports__permute as permute, __webpack_exports__permute_data as permute_data, __webpack_exports__pipeline as pipeline, __webpack_exports__quantize_embeddings as quantize_embeddings, __webpack_exports__read_audio as read_audio, __webpack_exports__rfft as rfft, __webpack_exports__round as round, __webpack_exports__softmax as softmax, __webpack_exports__spectrogram as spectrogram, __webpack_exports__stack as stack, __webpack_exports__std_mean as std_mean, __webpack_exports__topk as topk, __webpack_exports__window_function as window_function, __webpack_exports__zeros as zeros, __webpack_exports__zeros_like as zeros_like };
|
|
35919
|
+
export { __webpack_exports__ASTFeatureExtractor as ASTFeatureExtractor, __webpack_exports__ASTForAudioClassification as ASTForAudioClassification, __webpack_exports__ASTModel as ASTModel, __webpack_exports__ASTPreTrainedModel as ASTPreTrainedModel, __webpack_exports__AlbertForMaskedLM as AlbertForMaskedLM, __webpack_exports__AlbertForQuestionAnswering as AlbertForQuestionAnswering, __webpack_exports__AlbertForSequenceClassification as AlbertForSequenceClassification, __webpack_exports__AlbertModel as AlbertModel, __webpack_exports__AlbertPreTrainedModel as AlbertPreTrainedModel, __webpack_exports__AlbertTokenizer as AlbertTokenizer, __webpack_exports__AudioClassificationPipeline as AudioClassificationPipeline, __webpack_exports__AutoConfig as AutoConfig, __webpack_exports__AutoFeatureExtractor as AutoFeatureExtractor, __webpack_exports__AutoImageProcessor as AutoImageProcessor, __webpack_exports__AutoModel as AutoModel, __webpack_exports__AutoModelForAudioClassification as AutoModelForAudioClassification, __webpack_exports__AutoModelForAudioFrameClassification as AutoModelForAudioFrameClassification, __webpack_exports__AutoModelForCTC as AutoModelForCTC, __webpack_exports__AutoModelForCausalLM as AutoModelForCausalLM, __webpack_exports__AutoModelForDepthEstimation as AutoModelForDepthEstimation, __webpack_exports__AutoModelForDocumentQuestionAnswering as AutoModelForDocumentQuestionAnswering, __webpack_exports__AutoModelForImageClassification as AutoModelForImageClassification, __webpack_exports__AutoModelForImageFeatureExtraction as AutoModelForImageFeatureExtraction, __webpack_exports__AutoModelForImageMatting as AutoModelForImageMatting, __webpack_exports__AutoModelForImageSegmentation as AutoModelForImageSegmentation, __webpack_exports__AutoModelForImageToImage as AutoModelForImageToImage, __webpack_exports__AutoModelForMaskGeneration as AutoModelForMaskGeneration, __webpack_exports__AutoModelForMaskedLM as AutoModelForMaskedLM, __webpack_exports__AutoModelForNormalEstimation as AutoModelForNormalEstimation, __webpack_exports__AutoModelForObjectDetection as AutoModelForObjectDetection, __webpack_exports__AutoModelForPoseEstimation as AutoModelForPoseEstimation, __webpack_exports__AutoModelForQuestionAnswering as AutoModelForQuestionAnswering, __webpack_exports__AutoModelForSemanticSegmentation as AutoModelForSemanticSegmentation, __webpack_exports__AutoModelForSeq2SeqLM as AutoModelForSeq2SeqLM, __webpack_exports__AutoModelForSequenceClassification as AutoModelForSequenceClassification, __webpack_exports__AutoModelForSpeechSeq2Seq as AutoModelForSpeechSeq2Seq, __webpack_exports__AutoModelForTextToSpectrogram as AutoModelForTextToSpectrogram, __webpack_exports__AutoModelForTextToWaveform as AutoModelForTextToWaveform, __webpack_exports__AutoModelForTokenClassification as AutoModelForTokenClassification, __webpack_exports__AutoModelForUniversalSegmentation as AutoModelForUniversalSegmentation, __webpack_exports__AutoModelForVision2Seq as AutoModelForVision2Seq, __webpack_exports__AutoModelForXVector as AutoModelForXVector, __webpack_exports__AutoModelForZeroShotObjectDetection as AutoModelForZeroShotObjectDetection, __webpack_exports__AutoProcessor as AutoProcessor, __webpack_exports__AutoTokenizer as AutoTokenizer, __webpack_exports__AutomaticSpeechRecognitionPipeline as AutomaticSpeechRecognitionPipeline, __webpack_exports__BartForConditionalGeneration as BartForConditionalGeneration, __webpack_exports__BartForSequenceClassification as BartForSequenceClassification, __webpack_exports__BartModel as BartModel, __webpack_exports__BartPretrainedModel as BartPretrainedModel, __webpack_exports__BartTokenizer as BartTokenizer, __webpack_exports__BaseModelOutput as BaseModelOutput, __webpack_exports__BaseStreamer as BaseStreamer, __webpack_exports__BeitFeatureExtractor as BeitFeatureExtractor, __webpack_exports__BeitForImageClassification as BeitForImageClassification, __webpack_exports__BeitModel as BeitModel, __webpack_exports__BeitPreTrainedModel as BeitPreTrainedModel, __webpack_exports__BertForMaskedLM as BertForMaskedLM, __webpack_exports__BertForQuestionAnswering as BertForQuestionAnswering, __webpack_exports__BertForSequenceClassification as BertForSequenceClassification, __webpack_exports__BertForTokenClassification as BertForTokenClassification, __webpack_exports__BertModel as BertModel, __webpack_exports__BertPreTrainedModel as BertPreTrainedModel, __webpack_exports__BertTokenizer as BertTokenizer, __webpack_exports__BitImageProcessor as BitImageProcessor, __webpack_exports__BlenderbotForConditionalGeneration as BlenderbotForConditionalGeneration, __webpack_exports__BlenderbotModel as BlenderbotModel, __webpack_exports__BlenderbotPreTrainedModel as BlenderbotPreTrainedModel, __webpack_exports__BlenderbotSmallForConditionalGeneration as BlenderbotSmallForConditionalGeneration, __webpack_exports__BlenderbotSmallModel as BlenderbotSmallModel, __webpack_exports__BlenderbotSmallPreTrainedModel as BlenderbotSmallPreTrainedModel, __webpack_exports__BlenderbotSmallTokenizer as BlenderbotSmallTokenizer, __webpack_exports__BlenderbotTokenizer as BlenderbotTokenizer, __webpack_exports__BloomForCausalLM as BloomForCausalLM, __webpack_exports__BloomModel as BloomModel, __webpack_exports__BloomPreTrainedModel as BloomPreTrainedModel, __webpack_exports__BloomTokenizer as BloomTokenizer, __webpack_exports__CLIPFeatureExtractor as CLIPFeatureExtractor, __webpack_exports__CLIPImageProcessor as CLIPImageProcessor, __webpack_exports__CLIPModel as CLIPModel, __webpack_exports__CLIPPreTrainedModel as CLIPPreTrainedModel, __webpack_exports__CLIPSegForImageSegmentation as CLIPSegForImageSegmentation, __webpack_exports__CLIPSegModel as CLIPSegModel, __webpack_exports__CLIPSegPreTrainedModel as CLIPSegPreTrainedModel, __webpack_exports__CLIPTextModel as CLIPTextModel, __webpack_exports__CLIPTextModelWithProjection as CLIPTextModelWithProjection, __webpack_exports__CLIPTokenizer as CLIPTokenizer, __webpack_exports__CLIPVisionModel as CLIPVisionModel, __webpack_exports__CLIPVisionModelWithProjection as CLIPVisionModelWithProjection, __webpack_exports__CamembertForMaskedLM as CamembertForMaskedLM, __webpack_exports__CamembertForQuestionAnswering as CamembertForQuestionAnswering, __webpack_exports__CamembertForSequenceClassification as CamembertForSequenceClassification, __webpack_exports__CamembertForTokenClassification as CamembertForTokenClassification, __webpack_exports__CamembertModel as CamembertModel, __webpack_exports__CamembertPreTrainedModel as CamembertPreTrainedModel, __webpack_exports__CamembertTokenizer as CamembertTokenizer, __webpack_exports__CausalLMOutput as CausalLMOutput, __webpack_exports__CausalLMOutputWithPast as CausalLMOutputWithPast, __webpack_exports__ChineseCLIPFeatureExtractor as ChineseCLIPFeatureExtractor, __webpack_exports__ChineseCLIPModel as ChineseCLIPModel, __webpack_exports__ChineseCLIPPreTrainedModel as ChineseCLIPPreTrainedModel, __webpack_exports__ClapAudioModelWithProjection as ClapAudioModelWithProjection, __webpack_exports__ClapFeatureExtractor as ClapFeatureExtractor, __webpack_exports__ClapModel as ClapModel, __webpack_exports__ClapPreTrainedModel as ClapPreTrainedModel, __webpack_exports__ClapTextModelWithProjection as ClapTextModelWithProjection, __webpack_exports__ClassifierFreeGuidanceLogitsProcessor as ClassifierFreeGuidanceLogitsProcessor, __webpack_exports__CodeGenForCausalLM as CodeGenForCausalLM, __webpack_exports__CodeGenModel as CodeGenModel, __webpack_exports__CodeGenPreTrainedModel as CodeGenPreTrainedModel, __webpack_exports__CodeGenTokenizer as CodeGenTokenizer, __webpack_exports__CodeLlamaTokenizer as CodeLlamaTokenizer, __webpack_exports__CohereForCausalLM as CohereForCausalLM, __webpack_exports__CohereModel as CohereModel, __webpack_exports__CoherePreTrainedModel as CoherePreTrainedModel, __webpack_exports__CohereTokenizer as CohereTokenizer, __webpack_exports__ConvBertForMaskedLM as ConvBertForMaskedLM, __webpack_exports__ConvBertForQuestionAnswering as ConvBertForQuestionAnswering, __webpack_exports__ConvBertForSequenceClassification as ConvBertForSequenceClassification, __webpack_exports__ConvBertForTokenClassification as ConvBertForTokenClassification, __webpack_exports__ConvBertModel as ConvBertModel, __webpack_exports__ConvBertPreTrainedModel as ConvBertPreTrainedModel, __webpack_exports__ConvBertTokenizer as ConvBertTokenizer, __webpack_exports__ConvNextFeatureExtractor as ConvNextFeatureExtractor, __webpack_exports__ConvNextForImageClassification as ConvNextForImageClassification, __webpack_exports__ConvNextImageProcessor as ConvNextImageProcessor, __webpack_exports__ConvNextModel as ConvNextModel, __webpack_exports__ConvNextPreTrainedModel as ConvNextPreTrainedModel, __webpack_exports__ConvNextV2ForImageClassification as ConvNextV2ForImageClassification, __webpack_exports__ConvNextV2Model as ConvNextV2Model, __webpack_exports__ConvNextV2PreTrainedModel as ConvNextV2PreTrainedModel, __webpack_exports__DPTFeatureExtractor as DPTFeatureExtractor, __webpack_exports__DPTForDepthEstimation as DPTForDepthEstimation, __webpack_exports__DPTImageProcessor as DPTImageProcessor, __webpack_exports__DPTModel as DPTModel, __webpack_exports__DPTPreTrainedModel as DPTPreTrainedModel, __webpack_exports__DebertaForMaskedLM as DebertaForMaskedLM, __webpack_exports__DebertaForQuestionAnswering as DebertaForQuestionAnswering, __webpack_exports__DebertaForSequenceClassification as DebertaForSequenceClassification, __webpack_exports__DebertaForTokenClassification as DebertaForTokenClassification, __webpack_exports__DebertaModel as DebertaModel, __webpack_exports__DebertaPreTrainedModel as DebertaPreTrainedModel, __webpack_exports__DebertaTokenizer as DebertaTokenizer, __webpack_exports__DebertaV2ForMaskedLM as DebertaV2ForMaskedLM, __webpack_exports__DebertaV2ForQuestionAnswering as DebertaV2ForQuestionAnswering, __webpack_exports__DebertaV2ForSequenceClassification as DebertaV2ForSequenceClassification, __webpack_exports__DebertaV2ForTokenClassification as DebertaV2ForTokenClassification, __webpack_exports__DebertaV2Model as DebertaV2Model, __webpack_exports__DebertaV2PreTrainedModel as DebertaV2PreTrainedModel, __webpack_exports__DebertaV2Tokenizer as DebertaV2Tokenizer, __webpack_exports__DecisionTransformerModel as DecisionTransformerModel, __webpack_exports__DecisionTransformerPreTrainedModel as DecisionTransformerPreTrainedModel, __webpack_exports__DeiTFeatureExtractor as DeiTFeatureExtractor, __webpack_exports__DeiTForImageClassification as DeiTForImageClassification, __webpack_exports__DeiTImageProcessor as DeiTImageProcessor, __webpack_exports__DeiTModel as DeiTModel, __webpack_exports__DeiTPreTrainedModel as DeiTPreTrainedModel, __webpack_exports__DepthAnythingForDepthEstimation as DepthAnythingForDepthEstimation, __webpack_exports__DepthAnythingPreTrainedModel as DepthAnythingPreTrainedModel, __webpack_exports__DepthEstimationPipeline as DepthEstimationPipeline, __webpack_exports__DepthProForDepthEstimation as DepthProForDepthEstimation, __webpack_exports__DepthProPreTrainedModel as DepthProPreTrainedModel, __webpack_exports__DetrFeatureExtractor as DetrFeatureExtractor, __webpack_exports__DetrForObjectDetection as DetrForObjectDetection, __webpack_exports__DetrForSegmentation as DetrForSegmentation, __webpack_exports__DetrImageProcessor as DetrImageProcessor, __webpack_exports__DetrModel as DetrModel, __webpack_exports__DetrObjectDetectionOutput as DetrObjectDetectionOutput, __webpack_exports__DetrPreTrainedModel as DetrPreTrainedModel, __webpack_exports__DetrSegmentationOutput as DetrSegmentationOutput, __webpack_exports__Dinov2ForImageClassification as Dinov2ForImageClassification, __webpack_exports__Dinov2Model as Dinov2Model, __webpack_exports__Dinov2PreTrainedModel as Dinov2PreTrainedModel, __webpack_exports__DistilBertForMaskedLM as DistilBertForMaskedLM, __webpack_exports__DistilBertForQuestionAnswering as DistilBertForQuestionAnswering, __webpack_exports__DistilBertForSequenceClassification as DistilBertForSequenceClassification, __webpack_exports__DistilBertForTokenClassification as DistilBertForTokenClassification, __webpack_exports__DistilBertModel as DistilBertModel, __webpack_exports__DistilBertPreTrainedModel as DistilBertPreTrainedModel, __webpack_exports__DistilBertTokenizer as DistilBertTokenizer, __webpack_exports__DocumentQuestionAnsweringPipeline as DocumentQuestionAnsweringPipeline, __webpack_exports__DonutFeatureExtractor as DonutFeatureExtractor, __webpack_exports__DonutImageProcessor as DonutImageProcessor, __webpack_exports__DonutSwinModel as DonutSwinModel, __webpack_exports__DonutSwinPreTrainedModel as DonutSwinPreTrainedModel, __webpack_exports__EfficientNetForImageClassification as EfficientNetForImageClassification, __webpack_exports__EfficientNetImageProcessor as EfficientNetImageProcessor, __webpack_exports__EfficientNetModel as EfficientNetModel, __webpack_exports__EfficientNetPreTrainedModel as EfficientNetPreTrainedModel, __webpack_exports__ElectraForMaskedLM as ElectraForMaskedLM, __webpack_exports__ElectraForQuestionAnswering as ElectraForQuestionAnswering, __webpack_exports__ElectraForSequenceClassification as ElectraForSequenceClassification, __webpack_exports__ElectraForTokenClassification as ElectraForTokenClassification, __webpack_exports__ElectraModel as ElectraModel, __webpack_exports__ElectraPreTrainedModel as ElectraPreTrainedModel, __webpack_exports__ElectraTokenizer as ElectraTokenizer, __webpack_exports__EosTokenCriteria as EosTokenCriteria, __webpack_exports__EsmForMaskedLM as EsmForMaskedLM, __webpack_exports__EsmForSequenceClassification as EsmForSequenceClassification, __webpack_exports__EsmForTokenClassification as EsmForTokenClassification, __webpack_exports__EsmModel as EsmModel, __webpack_exports__EsmPreTrainedModel as EsmPreTrainedModel, __webpack_exports__EsmTokenizer as EsmTokenizer, __webpack_exports__ExaoneForCausalLM as ExaoneForCausalLM, __webpack_exports__ExaoneModel as ExaoneModel, __webpack_exports__ExaonePreTrainedModel as ExaonePreTrainedModel, __webpack_exports__FFT as FFT, __webpack_exports__FalconForCausalLM as FalconForCausalLM, __webpack_exports__FalconModel as FalconModel, __webpack_exports__FalconPreTrainedModel as FalconPreTrainedModel, __webpack_exports__FalconTokenizer as FalconTokenizer, __webpack_exports__FastViTForImageClassification as FastViTForImageClassification, __webpack_exports__FastViTModel as FastViTModel, __webpack_exports__FastViTPreTrainedModel as FastViTPreTrainedModel, __webpack_exports__FeatureExtractionPipeline as FeatureExtractionPipeline, __webpack_exports__FeatureExtractor as FeatureExtractor, __webpack_exports__FillMaskPipeline as FillMaskPipeline, __webpack_exports__Florence2ForConditionalGeneration as Florence2ForConditionalGeneration, __webpack_exports__Florence2PreTrainedModel as Florence2PreTrainedModel, __webpack_exports__Florence2Processor as Florence2Processor, __webpack_exports__ForcedBOSTokenLogitsProcessor as ForcedBOSTokenLogitsProcessor, __webpack_exports__ForcedEOSTokenLogitsProcessor as ForcedEOSTokenLogitsProcessor, __webpack_exports__GLPNFeatureExtractor as GLPNFeatureExtractor, __webpack_exports__GLPNForDepthEstimation as GLPNForDepthEstimation, __webpack_exports__GLPNModel as GLPNModel, __webpack_exports__GLPNPreTrainedModel as GLPNPreTrainedModel, __webpack_exports__GPT2LMHeadModel as GPT2LMHeadModel, __webpack_exports__GPT2Model as GPT2Model, __webpack_exports__GPT2PreTrainedModel as GPT2PreTrainedModel, __webpack_exports__GPT2Tokenizer as GPT2Tokenizer, __webpack_exports__GPTBigCodeForCausalLM as GPTBigCodeForCausalLM, __webpack_exports__GPTBigCodeModel as GPTBigCodeModel, __webpack_exports__GPTBigCodePreTrainedModel as GPTBigCodePreTrainedModel, __webpack_exports__GPTJForCausalLM as GPTJForCausalLM, __webpack_exports__GPTJModel as GPTJModel, __webpack_exports__GPTJPreTrainedModel as GPTJPreTrainedModel, __webpack_exports__GPTNeoForCausalLM as GPTNeoForCausalLM, __webpack_exports__GPTNeoModel as GPTNeoModel, __webpack_exports__GPTNeoPreTrainedModel as GPTNeoPreTrainedModel, __webpack_exports__GPTNeoXForCausalLM as GPTNeoXForCausalLM, __webpack_exports__GPTNeoXModel as GPTNeoXModel, __webpack_exports__GPTNeoXPreTrainedModel as GPTNeoXPreTrainedModel, __webpack_exports__GPTNeoXTokenizer as GPTNeoXTokenizer, __webpack_exports__Gemma2ForCausalLM as Gemma2ForCausalLM, __webpack_exports__Gemma2Model as Gemma2Model, __webpack_exports__Gemma2PreTrainedModel as Gemma2PreTrainedModel, __webpack_exports__GemmaForCausalLM as GemmaForCausalLM, __webpack_exports__GemmaModel as GemmaModel, __webpack_exports__GemmaPreTrainedModel as GemmaPreTrainedModel, __webpack_exports__GemmaTokenizer as GemmaTokenizer, __webpack_exports__GraniteForCausalLM as GraniteForCausalLM, __webpack_exports__GraniteModel as GraniteModel, __webpack_exports__GranitePreTrainedModel as GranitePreTrainedModel, __webpack_exports__Grok1Tokenizer as Grok1Tokenizer, __webpack_exports__GroupViTModel as GroupViTModel, __webpack_exports__GroupViTPreTrainedModel as GroupViTPreTrainedModel, __webpack_exports__HerbertTokenizer as HerbertTokenizer, __webpack_exports__HieraForImageClassification as HieraForImageClassification, __webpack_exports__HieraModel as HieraModel, __webpack_exports__HieraPreTrainedModel as HieraPreTrainedModel, __webpack_exports__HubertForCTC as HubertForCTC, __webpack_exports__HubertForSequenceClassification as HubertForSequenceClassification, __webpack_exports__HubertModel as HubertModel, __webpack_exports__HubertPreTrainedModel as HubertPreTrainedModel, __webpack_exports__IJepaForImageClassification as IJepaForImageClassification, __webpack_exports__IJepaModel as IJepaModel, __webpack_exports__IJepaPreTrainedModel as IJepaPreTrainedModel, __webpack_exports__Idefics3ForConditionalGeneration as Idefics3ForConditionalGeneration, __webpack_exports__Idefics3ImageProcessor as Idefics3ImageProcessor, __webpack_exports__Idefics3PreTrainedModel as Idefics3PreTrainedModel, __webpack_exports__Idefics3Processor as Idefics3Processor, __webpack_exports__ImageClassificationPipeline as ImageClassificationPipeline, __webpack_exports__ImageFeatureExtractionPipeline as ImageFeatureExtractionPipeline, __webpack_exports__ImageFeatureExtractor as ImageFeatureExtractor, __webpack_exports__ImageMattingOutput as ImageMattingOutput, __webpack_exports__ImageProcessor as ImageProcessor, __webpack_exports__ImageSegmentationPipeline as ImageSegmentationPipeline, __webpack_exports__ImageToImagePipeline as ImageToImagePipeline, __webpack_exports__ImageToTextPipeline as ImageToTextPipeline, __webpack_exports__InterruptableStoppingCriteria as InterruptableStoppingCriteria, __webpack_exports__JAISLMHeadModel as JAISLMHeadModel, __webpack_exports__JAISModel as JAISModel, __webpack_exports__JAISPreTrainedModel as JAISPreTrainedModel, __webpack_exports__JinaCLIPImageProcessor as JinaCLIPImageProcessor, __webpack_exports__JinaCLIPModel as JinaCLIPModel, __webpack_exports__JinaCLIPPreTrainedModel as JinaCLIPPreTrainedModel, __webpack_exports__JinaCLIPProcessor as JinaCLIPProcessor, __webpack_exports__JinaCLIPTextModel as JinaCLIPTextModel, __webpack_exports__JinaCLIPVisionModel as JinaCLIPVisionModel, __webpack_exports__LlamaForCausalLM as LlamaForCausalLM, __webpack_exports__LlamaModel as LlamaModel, __webpack_exports__LlamaPreTrainedModel as LlamaPreTrainedModel, __webpack_exports__LlamaTokenizer as LlamaTokenizer, __webpack_exports__LlavaForConditionalGeneration as LlavaForConditionalGeneration, __webpack_exports__LlavaOnevisionForConditionalGeneration as LlavaOnevisionForConditionalGeneration, __webpack_exports__LlavaOnevisionImageProcessor as LlavaOnevisionImageProcessor, __webpack_exports__LlavaPreTrainedModel as LlavaPreTrainedModel, __webpack_exports__LogitsProcessor as LogitsProcessor, __webpack_exports__LogitsProcessorList as LogitsProcessorList, __webpack_exports__LogitsWarper as LogitsWarper, __webpack_exports__LongT5ForConditionalGeneration as LongT5ForConditionalGeneration, __webpack_exports__LongT5Model as LongT5Model, __webpack_exports__LongT5PreTrainedModel as LongT5PreTrainedModel, __webpack_exports__M2M100ForConditionalGeneration as M2M100ForConditionalGeneration, __webpack_exports__M2M100Model as M2M100Model, __webpack_exports__M2M100PreTrainedModel as M2M100PreTrainedModel, __webpack_exports__M2M100Tokenizer as M2M100Tokenizer, __webpack_exports__MBart50Tokenizer as MBart50Tokenizer, __webpack_exports__MBartForCausalLM as MBartForCausalLM, __webpack_exports__MBartForConditionalGeneration as MBartForConditionalGeneration, __webpack_exports__MBartForSequenceClassification as MBartForSequenceClassification, __webpack_exports__MBartModel as MBartModel, __webpack_exports__MBartPreTrainedModel as MBartPreTrainedModel, __webpack_exports__MBartTokenizer as MBartTokenizer, __webpack_exports__MPNetForMaskedLM as MPNetForMaskedLM, __webpack_exports__MPNetForQuestionAnswering as MPNetForQuestionAnswering, __webpack_exports__MPNetForSequenceClassification as MPNetForSequenceClassification, __webpack_exports__MPNetForTokenClassification as MPNetForTokenClassification, __webpack_exports__MPNetModel as MPNetModel, __webpack_exports__MPNetPreTrainedModel as MPNetPreTrainedModel, __webpack_exports__MPNetTokenizer as MPNetTokenizer, __webpack_exports__MT5ForConditionalGeneration as MT5ForConditionalGeneration, __webpack_exports__MT5Model as MT5Model, __webpack_exports__MT5PreTrainedModel as MT5PreTrainedModel, __webpack_exports__MarianMTModel as MarianMTModel, __webpack_exports__MarianModel as MarianModel, __webpack_exports__MarianPreTrainedModel as MarianPreTrainedModel, __webpack_exports__MarianTokenizer as MarianTokenizer, __webpack_exports__Mask2FormerImageProcessor as Mask2FormerImageProcessor, __webpack_exports__MaskFormerFeatureExtractor as MaskFormerFeatureExtractor, __webpack_exports__MaskFormerForInstanceSegmentation as MaskFormerForInstanceSegmentation, __webpack_exports__MaskFormerImageProcessor as MaskFormerImageProcessor, __webpack_exports__MaskFormerModel as MaskFormerModel, __webpack_exports__MaskFormerPreTrainedModel as MaskFormerPreTrainedModel, __webpack_exports__MaskedLMOutput as MaskedLMOutput, __webpack_exports__MaxLengthCriteria as MaxLengthCriteria, __webpack_exports__MgpstrForSceneTextRecognition as MgpstrForSceneTextRecognition, __webpack_exports__MgpstrModelOutput as MgpstrModelOutput, __webpack_exports__MgpstrPreTrainedModel as MgpstrPreTrainedModel, __webpack_exports__MgpstrProcessor as MgpstrProcessor, __webpack_exports__MgpstrTokenizer as MgpstrTokenizer, __webpack_exports__MinLengthLogitsProcessor as MinLengthLogitsProcessor, __webpack_exports__MinNewTokensLengthLogitsProcessor as MinNewTokensLengthLogitsProcessor, __webpack_exports__MistralForCausalLM as MistralForCausalLM, __webpack_exports__MistralModel as MistralModel, __webpack_exports__MistralPreTrainedModel as MistralPreTrainedModel, __webpack_exports__MobileBertForMaskedLM as MobileBertForMaskedLM, __webpack_exports__MobileBertForQuestionAnswering as MobileBertForQuestionAnswering, __webpack_exports__MobileBertForSequenceClassification as MobileBertForSequenceClassification, __webpack_exports__MobileBertModel as MobileBertModel, __webpack_exports__MobileBertPreTrainedModel as MobileBertPreTrainedModel, __webpack_exports__MobileBertTokenizer as MobileBertTokenizer, __webpack_exports__MobileLLMForCausalLM as MobileLLMForCausalLM, __webpack_exports__MobileLLMModel as MobileLLMModel, __webpack_exports__MobileLLMPreTrainedModel as MobileLLMPreTrainedModel, __webpack_exports__MobileNetV1FeatureExtractor as MobileNetV1FeatureExtractor, __webpack_exports__MobileNetV1ForImageClassification as MobileNetV1ForImageClassification, __webpack_exports__MobileNetV1ImageProcessor as MobileNetV1ImageProcessor, __webpack_exports__MobileNetV1Model as MobileNetV1Model, __webpack_exports__MobileNetV1PreTrainedModel as MobileNetV1PreTrainedModel, __webpack_exports__MobileNetV2FeatureExtractor as MobileNetV2FeatureExtractor, __webpack_exports__MobileNetV2ForImageClassification as MobileNetV2ForImageClassification, __webpack_exports__MobileNetV2ImageProcessor as MobileNetV2ImageProcessor, __webpack_exports__MobileNetV2Model as MobileNetV2Model, __webpack_exports__MobileNetV2PreTrainedModel as MobileNetV2PreTrainedModel, __webpack_exports__MobileNetV3FeatureExtractor as MobileNetV3FeatureExtractor, __webpack_exports__MobileNetV3ForImageClassification as MobileNetV3ForImageClassification, __webpack_exports__MobileNetV3ImageProcessor as MobileNetV3ImageProcessor, __webpack_exports__MobileNetV3Model as MobileNetV3Model, __webpack_exports__MobileNetV3PreTrainedModel as MobileNetV3PreTrainedModel, __webpack_exports__MobileNetV4FeatureExtractor as MobileNetV4FeatureExtractor, __webpack_exports__MobileNetV4ForImageClassification as MobileNetV4ForImageClassification, __webpack_exports__MobileNetV4ImageProcessor as MobileNetV4ImageProcessor, __webpack_exports__MobileNetV4Model as MobileNetV4Model, __webpack_exports__MobileNetV4PreTrainedModel as MobileNetV4PreTrainedModel, __webpack_exports__MobileViTFeatureExtractor as MobileViTFeatureExtractor, __webpack_exports__MobileViTForImageClassification as MobileViTForImageClassification, __webpack_exports__MobileViTImageProcessor as MobileViTImageProcessor, __webpack_exports__MobileViTModel as MobileViTModel, __webpack_exports__MobileViTPreTrainedModel as MobileViTPreTrainedModel, __webpack_exports__MobileViTV2ForImageClassification as MobileViTV2ForImageClassification, __webpack_exports__MobileViTV2Model as MobileViTV2Model, __webpack_exports__MobileViTV2PreTrainedModel as MobileViTV2PreTrainedModel, __webpack_exports__ModelOutput as ModelOutput, __webpack_exports__Moondream1ForConditionalGeneration as Moondream1ForConditionalGeneration, __webpack_exports__MoonshineFeatureExtractor as MoonshineFeatureExtractor, __webpack_exports__MoonshineForConditionalGeneration as MoonshineForConditionalGeneration, __webpack_exports__MoonshineModel as MoonshineModel, __webpack_exports__MoonshinePreTrainedModel as MoonshinePreTrainedModel, __webpack_exports__MoonshineProcessor as MoonshineProcessor, __webpack_exports__MptForCausalLM as MptForCausalLM, __webpack_exports__MptModel as MptModel, __webpack_exports__MptPreTrainedModel as MptPreTrainedModel, __webpack_exports__MultiModalityCausalLM as MultiModalityCausalLM, __webpack_exports__MultiModalityPreTrainedModel as MultiModalityPreTrainedModel, __webpack_exports__MusicgenForCausalLM as MusicgenForCausalLM, __webpack_exports__MusicgenForConditionalGeneration as MusicgenForConditionalGeneration, __webpack_exports__MusicgenModel as MusicgenModel, __webpack_exports__MusicgenPreTrainedModel as MusicgenPreTrainedModel, __webpack_exports__NllbTokenizer as NllbTokenizer, __webpack_exports__NoBadWordsLogitsProcessor as NoBadWordsLogitsProcessor, __webpack_exports__NoRepeatNGramLogitsProcessor as NoRepeatNGramLogitsProcessor, __webpack_exports__NomicBertModel as NomicBertModel, __webpack_exports__NomicBertPreTrainedModel as NomicBertPreTrainedModel, __webpack_exports__NougatImageProcessor as NougatImageProcessor, __webpack_exports__NougatTokenizer as NougatTokenizer, __webpack_exports__OPTForCausalLM as OPTForCausalLM, __webpack_exports__OPTModel as OPTModel, __webpack_exports__OPTPreTrainedModel as OPTPreTrainedModel, __webpack_exports__ObjectDetectionPipeline as ObjectDetectionPipeline, __webpack_exports__Olmo2ForCausalLM as Olmo2ForCausalLM, __webpack_exports__Olmo2Model as Olmo2Model, __webpack_exports__Olmo2PreTrainedModel as Olmo2PreTrainedModel, __webpack_exports__OlmoForCausalLM as OlmoForCausalLM, __webpack_exports__OlmoModel as OlmoModel, __webpack_exports__OlmoPreTrainedModel as OlmoPreTrainedModel, __webpack_exports__OpenELMForCausalLM as OpenELMForCausalLM, __webpack_exports__OpenELMModel as OpenELMModel, __webpack_exports__OpenELMPreTrainedModel as OpenELMPreTrainedModel, __webpack_exports__OwlViTFeatureExtractor as OwlViTFeatureExtractor, __webpack_exports__OwlViTForObjectDetection as OwlViTForObjectDetection, __webpack_exports__OwlViTImageProcessor as OwlViTImageProcessor, __webpack_exports__OwlViTModel as OwlViTModel, __webpack_exports__OwlViTPreTrainedModel as OwlViTPreTrainedModel, __webpack_exports__OwlViTProcessor as OwlViTProcessor, __webpack_exports__Owlv2ForObjectDetection as Owlv2ForObjectDetection, __webpack_exports__Owlv2ImageProcessor as Owlv2ImageProcessor, __webpack_exports__Owlv2Model as Owlv2Model, __webpack_exports__Owlv2PreTrainedModel as Owlv2PreTrainedModel, __webpack_exports__PaliGemmaForConditionalGeneration as PaliGemmaForConditionalGeneration, __webpack_exports__PaliGemmaPreTrainedModel as PaliGemmaPreTrainedModel, __webpack_exports__PaliGemmaProcessor as PaliGemmaProcessor, __webpack_exports__PatchTSMixerForPrediction as PatchTSMixerForPrediction, __webpack_exports__PatchTSMixerModel as PatchTSMixerModel, __webpack_exports__PatchTSMixerPreTrainedModel as PatchTSMixerPreTrainedModel, __webpack_exports__PatchTSTForPrediction as PatchTSTForPrediction, __webpack_exports__PatchTSTModel as PatchTSTModel, __webpack_exports__PatchTSTPreTrainedModel as PatchTSTPreTrainedModel, __webpack_exports__Phi3ForCausalLM as Phi3ForCausalLM, __webpack_exports__Phi3Model as Phi3Model, __webpack_exports__Phi3PreTrainedModel as Phi3PreTrainedModel, __webpack_exports__Phi3VForCausalLM as Phi3VForCausalLM, __webpack_exports__Phi3VImageProcessor as Phi3VImageProcessor, __webpack_exports__Phi3VPreTrainedModel as Phi3VPreTrainedModel, __webpack_exports__Phi3VProcessor as Phi3VProcessor, __webpack_exports__PhiForCausalLM as PhiForCausalLM, __webpack_exports__PhiModel as PhiModel, __webpack_exports__PhiPreTrainedModel as PhiPreTrainedModel, __webpack_exports__Pipeline as Pipeline, __webpack_exports__PreTrainedModel as PreTrainedModel, __webpack_exports__PreTrainedTokenizer as PreTrainedTokenizer, __webpack_exports__PretrainedConfig as PretrainedConfig, __webpack_exports__PretrainedMixin as PretrainedMixin, __webpack_exports__Processor as Processor, __webpack_exports__PvtForImageClassification as PvtForImageClassification, __webpack_exports__PvtImageProcessor as PvtImageProcessor, __webpack_exports__PvtModel as PvtModel, __webpack_exports__PvtPreTrainedModel as PvtPreTrainedModel, __webpack_exports__PyAnnoteFeatureExtractor as PyAnnoteFeatureExtractor, __webpack_exports__PyAnnoteForAudioFrameClassification as PyAnnoteForAudioFrameClassification, __webpack_exports__PyAnnoteModel as PyAnnoteModel, __webpack_exports__PyAnnotePreTrainedModel as PyAnnotePreTrainedModel, __webpack_exports__PyAnnoteProcessor as PyAnnoteProcessor, __webpack_exports__QuestionAnsweringModelOutput as QuestionAnsweringModelOutput, __webpack_exports__QuestionAnsweringPipeline as QuestionAnsweringPipeline, __webpack_exports__Qwen2ForCausalLM as Qwen2ForCausalLM, __webpack_exports__Qwen2Model as Qwen2Model, __webpack_exports__Qwen2PreTrainedModel as Qwen2PreTrainedModel, __webpack_exports__Qwen2Tokenizer as Qwen2Tokenizer, __webpack_exports__Qwen2VLForConditionalGeneration as Qwen2VLForConditionalGeneration, __webpack_exports__Qwen2VLImageProcessor as Qwen2VLImageProcessor, __webpack_exports__Qwen2VLPreTrainedModel as Qwen2VLPreTrainedModel, __webpack_exports__Qwen2VLProcessor as Qwen2VLProcessor, __webpack_exports__RTDetrForObjectDetection as RTDetrForObjectDetection, __webpack_exports__RTDetrImageProcessor as RTDetrImageProcessor, __webpack_exports__RTDetrModel as RTDetrModel, __webpack_exports__RTDetrObjectDetectionOutput as RTDetrObjectDetectionOutput, __webpack_exports__RTDetrPreTrainedModel as RTDetrPreTrainedModel, __webpack_exports__RawImage as RawImage, __webpack_exports__RepetitionPenaltyLogitsProcessor as RepetitionPenaltyLogitsProcessor, __webpack_exports__ResNetForImageClassification as ResNetForImageClassification, __webpack_exports__ResNetModel as ResNetModel, __webpack_exports__ResNetPreTrainedModel as ResNetPreTrainedModel, __webpack_exports__RoFormerForMaskedLM as RoFormerForMaskedLM, __webpack_exports__RoFormerForQuestionAnswering as RoFormerForQuestionAnswering, __webpack_exports__RoFormerForSequenceClassification as RoFormerForSequenceClassification, __webpack_exports__RoFormerForTokenClassification as RoFormerForTokenClassification, __webpack_exports__RoFormerModel as RoFormerModel, __webpack_exports__RoFormerPreTrainedModel as RoFormerPreTrainedModel, __webpack_exports__RoFormerTokenizer as RoFormerTokenizer, __webpack_exports__RobertaForMaskedLM as RobertaForMaskedLM, __webpack_exports__RobertaForQuestionAnswering as RobertaForQuestionAnswering, __webpack_exports__RobertaForSequenceClassification as RobertaForSequenceClassification, __webpack_exports__RobertaForTokenClassification as RobertaForTokenClassification, __webpack_exports__RobertaModel as RobertaModel, __webpack_exports__RobertaPreTrainedModel as RobertaPreTrainedModel, __webpack_exports__RobertaTokenizer as RobertaTokenizer, __webpack_exports__SamImageProcessor as SamImageProcessor, __webpack_exports__SamImageSegmentationOutput as SamImageSegmentationOutput, __webpack_exports__SamModel as SamModel, __webpack_exports__SamPreTrainedModel as SamPreTrainedModel, __webpack_exports__SamProcessor as SamProcessor, __webpack_exports__SapiensForDepthEstimation as SapiensForDepthEstimation, __webpack_exports__SapiensForNormalEstimation as SapiensForNormalEstimation, __webpack_exports__SapiensForSemanticSegmentation as SapiensForSemanticSegmentation, __webpack_exports__SapiensPreTrainedModel as SapiensPreTrainedModel, __webpack_exports__SeamlessM4TFeatureExtractor as SeamlessM4TFeatureExtractor, __webpack_exports__SegformerFeatureExtractor as SegformerFeatureExtractor, __webpack_exports__SegformerForImageClassification as SegformerForImageClassification, __webpack_exports__SegformerForSemanticSegmentation as SegformerForSemanticSegmentation, __webpack_exports__SegformerImageProcessor as SegformerImageProcessor, __webpack_exports__SegformerModel as SegformerModel, __webpack_exports__SegformerPreTrainedModel as SegformerPreTrainedModel, __webpack_exports__Seq2SeqLMOutput as Seq2SeqLMOutput, __webpack_exports__SequenceClassifierOutput as SequenceClassifierOutput, __webpack_exports__SiglipImageProcessor as SiglipImageProcessor, __webpack_exports__SiglipModel as SiglipModel, __webpack_exports__SiglipPreTrainedModel as SiglipPreTrainedModel, __webpack_exports__SiglipTextModel as SiglipTextModel, __webpack_exports__SiglipTokenizer as SiglipTokenizer, __webpack_exports__SiglipVisionModel as SiglipVisionModel, __webpack_exports__SpeechT5FeatureExtractor as SpeechT5FeatureExtractor, __webpack_exports__SpeechT5ForSpeechToText as SpeechT5ForSpeechToText, __webpack_exports__SpeechT5ForTextToSpeech as SpeechT5ForTextToSpeech, __webpack_exports__SpeechT5HifiGan as SpeechT5HifiGan, __webpack_exports__SpeechT5Model as SpeechT5Model, __webpack_exports__SpeechT5PreTrainedModel as SpeechT5PreTrainedModel, __webpack_exports__SpeechT5Processor as SpeechT5Processor, __webpack_exports__SpeechT5Tokenizer as SpeechT5Tokenizer, __webpack_exports__SqueezeBertForMaskedLM as SqueezeBertForMaskedLM, __webpack_exports__SqueezeBertForQuestionAnswering as SqueezeBertForQuestionAnswering, __webpack_exports__SqueezeBertForSequenceClassification as SqueezeBertForSequenceClassification, __webpack_exports__SqueezeBertModel as SqueezeBertModel, __webpack_exports__SqueezeBertPreTrainedModel as SqueezeBertPreTrainedModel, __webpack_exports__SqueezeBertTokenizer as SqueezeBertTokenizer, __webpack_exports__StableLmForCausalLM as StableLmForCausalLM, __webpack_exports__StableLmModel as StableLmModel, __webpack_exports__StableLmPreTrainedModel as StableLmPreTrainedModel, __webpack_exports__Starcoder2ForCausalLM as Starcoder2ForCausalLM, __webpack_exports__Starcoder2Model as Starcoder2Model, __webpack_exports__Starcoder2PreTrainedModel as Starcoder2PreTrainedModel, __webpack_exports__StoppingCriteria as StoppingCriteria, __webpack_exports__StoppingCriteriaList as StoppingCriteriaList, __webpack_exports__SummarizationPipeline as SummarizationPipeline, __webpack_exports__SuppressTokensAtBeginLogitsProcessor as SuppressTokensAtBeginLogitsProcessor, __webpack_exports__Swin2SRForImageSuperResolution as Swin2SRForImageSuperResolution, __webpack_exports__Swin2SRImageProcessor as Swin2SRImageProcessor, __webpack_exports__Swin2SRModel as Swin2SRModel, __webpack_exports__Swin2SRPreTrainedModel as Swin2SRPreTrainedModel, __webpack_exports__SwinForImageClassification as SwinForImageClassification, __webpack_exports__SwinModel as SwinModel, __webpack_exports__SwinPreTrainedModel as SwinPreTrainedModel, __webpack_exports__T5ForConditionalGeneration as T5ForConditionalGeneration, __webpack_exports__T5Model as T5Model, __webpack_exports__T5PreTrainedModel as T5PreTrainedModel, __webpack_exports__T5Tokenizer as T5Tokenizer, __webpack_exports__TableTransformerForObjectDetection as TableTransformerForObjectDetection, __webpack_exports__TableTransformerModel as TableTransformerModel, __webpack_exports__TableTransformerObjectDetectionOutput as TableTransformerObjectDetectionOutput, __webpack_exports__TableTransformerPreTrainedModel as TableTransformerPreTrainedModel, __webpack_exports__TemperatureLogitsWarper as TemperatureLogitsWarper, __webpack_exports__Tensor as Tensor, __webpack_exports__Text2TextGenerationPipeline as Text2TextGenerationPipeline, __webpack_exports__TextClassificationPipeline as TextClassificationPipeline, __webpack_exports__TextGenerationPipeline as TextGenerationPipeline, __webpack_exports__TextStreamer as TextStreamer, __webpack_exports__TextToAudioPipeline as TextToAudioPipeline, __webpack_exports__TokenClassificationPipeline as TokenClassificationPipeline, __webpack_exports__TokenClassifierOutput as TokenClassifierOutput, __webpack_exports__TokenizerModel as TokenizerModel, __webpack_exports__TopKLogitsWarper as TopKLogitsWarper, __webpack_exports__TopPLogitsWarper as TopPLogitsWarper, __webpack_exports__TrOCRForCausalLM as TrOCRForCausalLM, __webpack_exports__TrOCRPreTrainedModel as TrOCRPreTrainedModel, __webpack_exports__TranslationPipeline as TranslationPipeline, __webpack_exports__UniSpeechForCTC as UniSpeechForCTC, __webpack_exports__UniSpeechForSequenceClassification as UniSpeechForSequenceClassification, __webpack_exports__UniSpeechModel as UniSpeechModel, __webpack_exports__UniSpeechPreTrainedModel as UniSpeechPreTrainedModel, __webpack_exports__UniSpeechSatForAudioFrameClassification as UniSpeechSatForAudioFrameClassification, __webpack_exports__UniSpeechSatForCTC as UniSpeechSatForCTC, __webpack_exports__UniSpeechSatForSequenceClassification as UniSpeechSatForSequenceClassification, __webpack_exports__UniSpeechSatModel as UniSpeechSatModel, __webpack_exports__UniSpeechSatPreTrainedModel as UniSpeechSatPreTrainedModel, __webpack_exports__VLChatProcessor as VLChatProcessor, __webpack_exports__VLMImageProcessor as VLMImageProcessor, __webpack_exports__ViTFeatureExtractor as ViTFeatureExtractor, __webpack_exports__ViTForImageClassification as ViTForImageClassification, __webpack_exports__ViTImageProcessor as ViTImageProcessor, __webpack_exports__ViTMAEModel as ViTMAEModel, __webpack_exports__ViTMAEPreTrainedModel as ViTMAEPreTrainedModel, __webpack_exports__ViTMSNForImageClassification as ViTMSNForImageClassification, __webpack_exports__ViTMSNModel as ViTMSNModel, __webpack_exports__ViTMSNPreTrainedModel as ViTMSNPreTrainedModel, __webpack_exports__ViTModel as ViTModel, __webpack_exports__ViTPreTrainedModel as ViTPreTrainedModel, __webpack_exports__VisionEncoderDecoderModel as VisionEncoderDecoderModel, __webpack_exports__VitMatteForImageMatting as VitMatteForImageMatting, __webpack_exports__VitMatteImageProcessor as VitMatteImageProcessor, __webpack_exports__VitMattePreTrainedModel as VitMattePreTrainedModel, __webpack_exports__VitPoseForPoseEstimation as VitPoseForPoseEstimation, __webpack_exports__VitPoseImageProcessor as VitPoseImageProcessor, __webpack_exports__VitPosePreTrainedModel as VitPosePreTrainedModel, __webpack_exports__VitsModel as VitsModel, __webpack_exports__VitsModelOutput as VitsModelOutput, __webpack_exports__VitsPreTrainedModel as VitsPreTrainedModel, __webpack_exports__VitsTokenizer as VitsTokenizer, __webpack_exports__Wav2Vec2BertForCTC as Wav2Vec2BertForCTC, __webpack_exports__Wav2Vec2BertForSequenceClassification as Wav2Vec2BertForSequenceClassification, __webpack_exports__Wav2Vec2BertModel as Wav2Vec2BertModel, __webpack_exports__Wav2Vec2BertPreTrainedModel as Wav2Vec2BertPreTrainedModel, __webpack_exports__Wav2Vec2CTCTokenizer as Wav2Vec2CTCTokenizer, __webpack_exports__Wav2Vec2FeatureExtractor as Wav2Vec2FeatureExtractor, __webpack_exports__Wav2Vec2ForAudioFrameClassification as Wav2Vec2ForAudioFrameClassification, __webpack_exports__Wav2Vec2ForCTC as Wav2Vec2ForCTC, __webpack_exports__Wav2Vec2ForSequenceClassification as Wav2Vec2ForSequenceClassification, __webpack_exports__Wav2Vec2Model as Wav2Vec2Model, __webpack_exports__Wav2Vec2PreTrainedModel as Wav2Vec2PreTrainedModel, __webpack_exports__Wav2Vec2ProcessorWithLM as Wav2Vec2ProcessorWithLM, __webpack_exports__WavLMForAudioFrameClassification as WavLMForAudioFrameClassification, __webpack_exports__WavLMForCTC as WavLMForCTC, __webpack_exports__WavLMForSequenceClassification as WavLMForSequenceClassification, __webpack_exports__WavLMForXVector as WavLMForXVector, __webpack_exports__WavLMModel as WavLMModel, __webpack_exports__WavLMPreTrainedModel as WavLMPreTrainedModel, __webpack_exports__WeSpeakerFeatureExtractor as WeSpeakerFeatureExtractor, __webpack_exports__WeSpeakerResNetModel as WeSpeakerResNetModel, __webpack_exports__WeSpeakerResNetPreTrainedModel as WeSpeakerResNetPreTrainedModel, __webpack_exports__WhisperFeatureExtractor as WhisperFeatureExtractor, __webpack_exports__WhisperForConditionalGeneration as WhisperForConditionalGeneration, __webpack_exports__WhisperModel as WhisperModel, __webpack_exports__WhisperPreTrainedModel as WhisperPreTrainedModel, __webpack_exports__WhisperProcessor as WhisperProcessor, __webpack_exports__WhisperTextStreamer as WhisperTextStreamer, __webpack_exports__WhisperTimeStampLogitsProcessor as WhisperTimeStampLogitsProcessor, __webpack_exports__WhisperTokenizer as WhisperTokenizer, __webpack_exports__XLMForQuestionAnswering as XLMForQuestionAnswering, __webpack_exports__XLMForSequenceClassification as XLMForSequenceClassification, __webpack_exports__XLMForTokenClassification as XLMForTokenClassification, __webpack_exports__XLMModel as XLMModel, __webpack_exports__XLMPreTrainedModel as XLMPreTrainedModel, __webpack_exports__XLMRobertaForMaskedLM as XLMRobertaForMaskedLM, __webpack_exports__XLMRobertaForQuestionAnswering as XLMRobertaForQuestionAnswering, __webpack_exports__XLMRobertaForSequenceClassification as XLMRobertaForSequenceClassification, __webpack_exports__XLMRobertaForTokenClassification as XLMRobertaForTokenClassification, __webpack_exports__XLMRobertaModel as XLMRobertaModel, __webpack_exports__XLMRobertaPreTrainedModel as XLMRobertaPreTrainedModel, __webpack_exports__XLMRobertaTokenizer as XLMRobertaTokenizer, __webpack_exports__XLMTokenizer as XLMTokenizer, __webpack_exports__XLMWithLMHeadModel as XLMWithLMHeadModel, __webpack_exports__XVectorOutput as XVectorOutput, __webpack_exports__YolosFeatureExtractor as YolosFeatureExtractor, __webpack_exports__YolosForObjectDetection as YolosForObjectDetection, __webpack_exports__YolosImageProcessor as YolosImageProcessor, __webpack_exports__YolosModel as YolosModel, __webpack_exports__YolosObjectDetectionOutput as YolosObjectDetectionOutput, __webpack_exports__YolosPreTrainedModel as YolosPreTrainedModel, __webpack_exports__ZeroShotAudioClassificationPipeline as ZeroShotAudioClassificationPipeline, __webpack_exports__ZeroShotClassificationPipeline as ZeroShotClassificationPipeline, __webpack_exports__ZeroShotImageClassificationPipeline as ZeroShotImageClassificationPipeline, __webpack_exports__ZeroShotObjectDetectionPipeline as ZeroShotObjectDetectionPipeline, __webpack_exports__bankers_round as bankers_round, __webpack_exports__cat as cat, __webpack_exports__cos_sim as cos_sim, __webpack_exports__dot as dot, __webpack_exports__dynamic_time_warping as dynamic_time_warping, __webpack_exports__env as env, __webpack_exports__full as full, __webpack_exports__full_like as full_like, __webpack_exports__getKeyValueShapes as getKeyValueShapes, __webpack_exports__hamming as hamming, __webpack_exports__hanning as hanning, __webpack_exports__interpolate as interpolate, __webpack_exports__interpolate_4d as interpolate_4d, __webpack_exports__interpolate_data as interpolate_data, __webpack_exports__is_chinese_char as is_chinese_char, __webpack_exports__layer_norm as layer_norm, __webpack_exports__load_image as load_image, __webpack_exports__log_softmax as log_softmax, __webpack_exports__magnitude as magnitude, __webpack_exports__matmul as matmul, __webpack_exports__max as max, __webpack_exports__mean as mean, __webpack_exports__mean_pooling as mean_pooling, __webpack_exports__medianFilter as medianFilter, __webpack_exports__mel_filter_bank as mel_filter_bank, __webpack_exports__min as min, __webpack_exports__ones as ones, __webpack_exports__ones_like as ones_like, __webpack_exports__permute as permute, __webpack_exports__permute_data as permute_data, __webpack_exports__pipeline as pipeline, __webpack_exports__quantize_embeddings as quantize_embeddings, __webpack_exports__rand as rand, __webpack_exports__read_audio as read_audio, __webpack_exports__rfft as rfft, __webpack_exports__round as round, __webpack_exports__slice as slice, __webpack_exports__softmax as softmax, __webpack_exports__spectrogram as spectrogram, __webpack_exports__stack as stack, __webpack_exports__std_mean as std_mean, __webpack_exports__topk as topk, __webpack_exports__window_function as window_function, __webpack_exports__zeros as zeros, __webpack_exports__zeros_like as zeros_like };
|
|
35019
35920
|
|
|
35020
35921
|
//# sourceMappingURL=transformers.mjs.map
|