@huggingface/transformers 3.4.0 → 3.4.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +8 -2
- package/dist/transformers.js +528 -201
- package/dist/transformers.js.map +1 -1
- package/dist/transformers.min.js +1 -1
- package/dist/transformers.min.js.map +1 -1
- package/dist/transformers.node.cjs +508 -200
- package/dist/transformers.node.cjs.map +1 -1
- package/dist/transformers.node.min.cjs +1 -1
- package/dist/transformers.node.min.cjs.map +1 -1
- package/dist/transformers.node.min.mjs +1 -1
- package/dist/transformers.node.min.mjs.map +1 -1
- package/dist/transformers.node.mjs +528 -201
- package/dist/transformers.node.mjs.map +1 -1
- package/dist/transformers.web.js +528 -201
- package/dist/transformers.web.js.map +1 -1
- package/dist/transformers.web.min.js +1 -1
- package/dist/transformers.web.min.js.map +1 -1
- package/package.json +1 -1
- package/src/configs.js +2 -0
- package/src/env.js +1 -1
- package/src/models/feature_extractors.js +1 -0
- package/src/models/snac/feature_extraction_snac.js +3 -0
- package/src/models.js +125 -2
- package/src/pipelines.js +140 -135
- package/src/tokenizers.js +44 -34
- package/src/utils/data-structures.js +74 -0
- package/src/utils/hub.js +36 -15
- package/src/utils/image.js +9 -1
- package/src/utils/tensor.js +6 -2
- package/types/configs.d.ts.map +1 -1
- package/types/models/feature_extractors.d.ts +1 -0
- package/types/models/snac/feature_extraction_snac.d.ts +4 -0
- package/types/models/snac/feature_extraction_snac.d.ts.map +1 -0
- package/types/models.d.ts +72 -0
- package/types/models.d.ts.map +1 -1
- package/types/pipelines.d.ts +2 -2
- package/types/pipelines.d.ts.map +1 -1
- package/types/tokenizers.d.ts +4 -1
- package/types/tokenizers.d.ts.map +1 -1
- package/types/tsconfig.tsbuildinfo +1 -1
- package/types/utils/data-structures.d.ts +26 -0
- package/types/utils/data-structures.d.ts.map +1 -1
- package/types/utils/hub.d.ts.map +1 -1
- package/types/utils/image.d.ts +2 -2
- package/types/utils/image.d.ts.map +1 -1
- package/types/utils/tensor.d.ts.map +1 -1
|
@@ -3676,6 +3676,7 @@ function getNormalizedConfig(config) {
|
|
|
3676
3676
|
// Sub-configs
|
|
3677
3677
|
case 'llava':
|
|
3678
3678
|
case 'paligemma':
|
|
3679
|
+
case 'gemma3':
|
|
3679
3680
|
case 'florence2':
|
|
3680
3681
|
case 'llava_onevision':
|
|
3681
3682
|
case 'idefics3':
|
|
@@ -3735,6 +3736,7 @@ function getNormalizedConfig(config) {
|
|
|
3735
3736
|
break;
|
|
3736
3737
|
case 'gemma':
|
|
3737
3738
|
case 'gemma2':
|
|
3739
|
+
case 'gemma3_text':
|
|
3738
3740
|
case 'glm':
|
|
3739
3741
|
case 'helium':
|
|
3740
3742
|
mapping['num_heads'] = 'num_key_value_heads';
|
|
@@ -4065,7 +4067,7 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
4065
4067
|
|
|
4066
4068
|
|
|
4067
4069
|
|
|
4068
|
-
const VERSION = '3.4.
|
|
4070
|
+
const VERSION = '3.4.2';
|
|
4069
4071
|
|
|
4070
4072
|
// Check if various APIs are available (depends on environment)
|
|
4071
4073
|
const IS_BROWSER_ENV = typeof window !== "undefined" && typeof window.document !== "undefined";
|
|
@@ -6203,6 +6205,9 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
6203
6205
|
/* harmony export */ Gemma2ForCausalLM: () => (/* binding */ Gemma2ForCausalLM),
|
|
6204
6206
|
/* harmony export */ Gemma2Model: () => (/* binding */ Gemma2Model),
|
|
6205
6207
|
/* harmony export */ Gemma2PreTrainedModel: () => (/* binding */ Gemma2PreTrainedModel),
|
|
6208
|
+
/* harmony export */ Gemma3ForCausalLM: () => (/* binding */ Gemma3ForCausalLM),
|
|
6209
|
+
/* harmony export */ Gemma3Model: () => (/* binding */ Gemma3Model),
|
|
6210
|
+
/* harmony export */ Gemma3PreTrainedModel: () => (/* binding */ Gemma3PreTrainedModel),
|
|
6206
6211
|
/* harmony export */ GemmaForCausalLM: () => (/* binding */ GemmaForCausalLM),
|
|
6207
6212
|
/* harmony export */ GemmaModel: () => (/* binding */ GemmaModel),
|
|
6208
6213
|
/* harmony export */ GemmaPreTrainedModel: () => (/* binding */ GemmaPreTrainedModel),
|
|
@@ -6273,6 +6278,10 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
6273
6278
|
/* harmony export */ MaskFormerModel: () => (/* binding */ MaskFormerModel),
|
|
6274
6279
|
/* harmony export */ MaskFormerPreTrainedModel: () => (/* binding */ MaskFormerPreTrainedModel),
|
|
6275
6280
|
/* harmony export */ MaskedLMOutput: () => (/* binding */ MaskedLMOutput),
|
|
6281
|
+
/* harmony export */ Metric3DForDepthEstimation: () => (/* binding */ Metric3DForDepthEstimation),
|
|
6282
|
+
/* harmony export */ Metric3DPreTrainedModel: () => (/* binding */ Metric3DPreTrainedModel),
|
|
6283
|
+
/* harmony export */ Metric3Dv2ForDepthEstimation: () => (/* binding */ Metric3Dv2ForDepthEstimation),
|
|
6284
|
+
/* harmony export */ Metric3Dv2PreTrainedModel: () => (/* binding */ Metric3Dv2PreTrainedModel),
|
|
6276
6285
|
/* harmony export */ MgpstrForSceneTextRecognition: () => (/* binding */ MgpstrForSceneTextRecognition),
|
|
6277
6286
|
/* harmony export */ MgpstrModelOutput: () => (/* binding */ MgpstrModelOutput),
|
|
6278
6287
|
/* harmony export */ MgpstrPreTrainedModel: () => (/* binding */ MgpstrPreTrainedModel),
|
|
@@ -6384,10 +6393,18 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
6384
6393
|
/* harmony export */ Qwen2PreTrainedModel: () => (/* binding */ Qwen2PreTrainedModel),
|
|
6385
6394
|
/* harmony export */ Qwen2VLForConditionalGeneration: () => (/* binding */ Qwen2VLForConditionalGeneration),
|
|
6386
6395
|
/* harmony export */ Qwen2VLPreTrainedModel: () => (/* binding */ Qwen2VLPreTrainedModel),
|
|
6396
|
+
/* harmony export */ RFDetrForObjectDetection: () => (/* binding */ RFDetrForObjectDetection),
|
|
6397
|
+
/* harmony export */ RFDetrModel: () => (/* binding */ RFDetrModel),
|
|
6398
|
+
/* harmony export */ RFDetrObjectDetectionOutput: () => (/* binding */ RFDetrObjectDetectionOutput),
|
|
6399
|
+
/* harmony export */ RFDetrPreTrainedModel: () => (/* binding */ RFDetrPreTrainedModel),
|
|
6387
6400
|
/* harmony export */ RTDetrForObjectDetection: () => (/* binding */ RTDetrForObjectDetection),
|
|
6388
6401
|
/* harmony export */ RTDetrModel: () => (/* binding */ RTDetrModel),
|
|
6389
6402
|
/* harmony export */ RTDetrObjectDetectionOutput: () => (/* binding */ RTDetrObjectDetectionOutput),
|
|
6390
6403
|
/* harmony export */ RTDetrPreTrainedModel: () => (/* binding */ RTDetrPreTrainedModel),
|
|
6404
|
+
/* harmony export */ RTDetrV2ForObjectDetection: () => (/* binding */ RTDetrV2ForObjectDetection),
|
|
6405
|
+
/* harmony export */ RTDetrV2Model: () => (/* binding */ RTDetrV2Model),
|
|
6406
|
+
/* harmony export */ RTDetrV2ObjectDetectionOutput: () => (/* binding */ RTDetrV2ObjectDetectionOutput),
|
|
6407
|
+
/* harmony export */ RTDetrV2PreTrainedModel: () => (/* binding */ RTDetrV2PreTrainedModel),
|
|
6391
6408
|
/* harmony export */ ResNetForImageClassification: () => (/* binding */ ResNetForImageClassification),
|
|
6392
6409
|
/* harmony export */ ResNetModel: () => (/* binding */ ResNetModel),
|
|
6393
6410
|
/* harmony export */ ResNetPreTrainedModel: () => (/* binding */ ResNetPreTrainedModel),
|
|
@@ -6421,6 +6438,10 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
6421
6438
|
/* harmony export */ SiglipTextModel: () => (/* binding */ SiglipTextModel),
|
|
6422
6439
|
/* harmony export */ SiglipVisionModel: () => (/* binding */ SiglipVisionModel),
|
|
6423
6440
|
/* harmony export */ SmolVLMForConditionalGeneration: () => (/* binding */ SmolVLMForConditionalGeneration),
|
|
6441
|
+
/* harmony export */ SnacDecoderModel: () => (/* binding */ SnacDecoderModel),
|
|
6442
|
+
/* harmony export */ SnacEncoderModel: () => (/* binding */ SnacEncoderModel),
|
|
6443
|
+
/* harmony export */ SnacModel: () => (/* binding */ SnacModel),
|
|
6444
|
+
/* harmony export */ SnacPreTrainedModel: () => (/* binding */ SnacPreTrainedModel),
|
|
6424
6445
|
/* harmony export */ SpeechT5ForSpeechToText: () => (/* binding */ SpeechT5ForSpeechToText),
|
|
6425
6446
|
/* harmony export */ SpeechT5ForTextToSpeech: () => (/* binding */ SpeechT5ForTextToSpeech),
|
|
6426
6447
|
/* harmony export */ SpeechT5HifiGan: () => (/* binding */ SpeechT5HifiGan),
|
|
@@ -7079,8 +7100,8 @@ async function decoderForward(self, model_inputs, is_encoder_decoder = false) {
|
|
|
7079
7100
|
new_model_inputs.use_cache_branch = boolTensor(!!past_key_values);
|
|
7080
7101
|
}
|
|
7081
7102
|
if (session.inputNames.includes('position_ids') && new_model_inputs.attention_mask && !new_model_inputs.position_ids) {
|
|
7082
|
-
// NOTE: Handle a special case for paligemma models, where positions are 1-indexed
|
|
7083
|
-
const start_index = self.config.model_type
|
|
7103
|
+
// NOTE: Handle a special case for paligemma/gemma3 models, where positions are 1-indexed
|
|
7104
|
+
const start_index = ['paligemma', 'gemma3_text', 'gemma3'].includes(self.config.model_type) ? 1 : 0;
|
|
7084
7105
|
new_model_inputs.position_ids = createPositionIds(new_model_inputs, past_key_values, start_index);
|
|
7085
7106
|
}
|
|
7086
7107
|
|
|
@@ -11005,6 +11026,23 @@ class Gemma2Model extends Gemma2PreTrainedModel { }
|
|
|
11005
11026
|
class Gemma2ForCausalLM extends Gemma2PreTrainedModel { }
|
|
11006
11027
|
//////////////////////////////////////////////////
|
|
11007
11028
|
|
|
11029
|
+
|
|
11030
|
+
//////////////////////////////////////////////////
|
|
11031
|
+
// Gemma3 models
|
|
11032
|
+
|
|
11033
|
+
/**
|
|
11034
|
+
* The bare Gemma3 Model outputting raw hidden-states without any specific head on top.
|
|
11035
|
+
*/
|
|
11036
|
+
class Gemma3PreTrainedModel extends PreTrainedModel { }
|
|
11037
|
+
/**
|
|
11038
|
+
* The bare Gemma3 Model outputting raw hidden-states without any specific head on top.
|
|
11039
|
+
*/
|
|
11040
|
+
class Gemma3Model extends Gemma3PreTrainedModel { }
|
|
11041
|
+
|
|
11042
|
+
class Gemma3ForCausalLM extends Gemma3PreTrainedModel { }
|
|
11043
|
+
//////////////////////////////////////////////////
|
|
11044
|
+
|
|
11045
|
+
|
|
11008
11046
|
//////////////////////////////////////////////////
|
|
11009
11047
|
class OpenELMPreTrainedModel extends PreTrainedModel { }
|
|
11010
11048
|
class OpenELMModel extends OpenELMPreTrainedModel { }
|
|
@@ -11649,6 +11687,37 @@ class RTDetrObjectDetectionOutput extends ModelOutput {
|
|
|
11649
11687
|
}
|
|
11650
11688
|
//////////////////////////////////////////////////
|
|
11651
11689
|
|
|
11690
|
+
|
|
11691
|
+
//////////////////////////////////////////////////
|
|
11692
|
+
class RTDetrV2PreTrainedModel extends PreTrainedModel { }
|
|
11693
|
+
class RTDetrV2Model extends RTDetrV2PreTrainedModel { }
|
|
11694
|
+
class RTDetrV2ForObjectDetection extends RTDetrV2PreTrainedModel {
|
|
11695
|
+
/**
|
|
11696
|
+
* @param {any} model_inputs
|
|
11697
|
+
*/
|
|
11698
|
+
async _call(model_inputs) {
|
|
11699
|
+
return new RTDetrV2ObjectDetectionOutput(await super._call(model_inputs));
|
|
11700
|
+
}
|
|
11701
|
+
}
|
|
11702
|
+
|
|
11703
|
+
class RTDetrV2ObjectDetectionOutput extends RTDetrObjectDetectionOutput {}
|
|
11704
|
+
//////////////////////////////////////////////////
|
|
11705
|
+
|
|
11706
|
+
//////////////////////////////////////////////////
|
|
11707
|
+
class RFDetrPreTrainedModel extends PreTrainedModel { }
|
|
11708
|
+
class RFDetrModel extends RFDetrPreTrainedModel { }
|
|
11709
|
+
class RFDetrForObjectDetection extends RFDetrPreTrainedModel {
|
|
11710
|
+
/**
|
|
11711
|
+
* @param {any} model_inputs
|
|
11712
|
+
*/
|
|
11713
|
+
async _call(model_inputs) {
|
|
11714
|
+
return new RFDetrObjectDetectionOutput(await super._call(model_inputs));
|
|
11715
|
+
}
|
|
11716
|
+
}
|
|
11717
|
+
|
|
11718
|
+
class RFDetrObjectDetectionOutput extends RTDetrObjectDetectionOutput {}
|
|
11719
|
+
//////////////////////////////////////////////////
|
|
11720
|
+
|
|
11652
11721
|
//////////////////////////////////////////////////
|
|
11653
11722
|
class TableTransformerPreTrainedModel extends PreTrainedModel { }
|
|
11654
11723
|
|
|
@@ -11857,6 +11926,16 @@ class DepthProPreTrainedModel extends PreTrainedModel { }
|
|
|
11857
11926
|
class DepthProForDepthEstimation extends DepthProPreTrainedModel { }
|
|
11858
11927
|
//////////////////////////////////////////////////
|
|
11859
11928
|
|
|
11929
|
+
//////////////////////////////////////////////////
|
|
11930
|
+
class Metric3DPreTrainedModel extends PreTrainedModel { }
|
|
11931
|
+
class Metric3DForDepthEstimation extends Metric3DPreTrainedModel { }
|
|
11932
|
+
//////////////////////////////////////////////////
|
|
11933
|
+
|
|
11934
|
+
//////////////////////////////////////////////////
|
|
11935
|
+
class Metric3Dv2PreTrainedModel extends PreTrainedModel { }
|
|
11936
|
+
class Metric3Dv2ForDepthEstimation extends Metric3Dv2PreTrainedModel { }
|
|
11937
|
+
//////////////////////////////////////////////////
|
|
11938
|
+
|
|
11860
11939
|
//////////////////////////////////////////////////
|
|
11861
11940
|
class MaskFormerPreTrainedModel extends PreTrainedModel { }
|
|
11862
11941
|
class MaskFormerModel extends MaskFormerPreTrainedModel { }
|
|
@@ -13772,6 +13851,60 @@ class DacDecoderModel extends DacPreTrainedModel {
|
|
|
13772
13851
|
}
|
|
13773
13852
|
//////////////////////////////////////////////////
|
|
13774
13853
|
|
|
13854
|
+
|
|
13855
|
+
//////////////////////////////////////////////////
|
|
13856
|
+
// Snac models
|
|
13857
|
+
class SnacPreTrainedModel extends PreTrainedModel {
|
|
13858
|
+
main_input_name = 'input_values';
|
|
13859
|
+
forward_params = ['input_values'];
|
|
13860
|
+
}
|
|
13861
|
+
|
|
13862
|
+
/**
|
|
13863
|
+
* The SNAC (Multi-Scale Neural Audio Codec) model.
|
|
13864
|
+
*/
|
|
13865
|
+
class SnacModel extends SnacPreTrainedModel {
|
|
13866
|
+
/**
|
|
13867
|
+
* Encodes the input audio waveform into discrete codes.
|
|
13868
|
+
* @param {Object} inputs Model inputs
|
|
13869
|
+
* @param {Tensor} [inputs.input_values] Float values of the input audio waveform, of shape `(batch_size, channels, sequence_length)`).
|
|
13870
|
+
* @returns {Promise<Record<string, Tensor>>} The output tensors of shape `(batch_size, num_codebooks, sequence_length)`.
|
|
13871
|
+
*/
|
|
13872
|
+
async encode(inputs) {
|
|
13873
|
+
return await sessionRun(this.sessions['encoder_model'], inputs);
|
|
13874
|
+
}
|
|
13875
|
+
|
|
13876
|
+
/**
|
|
13877
|
+
* Decodes the given frames into an output audio waveform.
|
|
13878
|
+
* @param {Record<string, Tensor>} inputs The encoded audio codes.
|
|
13879
|
+
* @returns {Promise<{audio_values: Tensor}>} The output tensor of shape `(batch_size, num_channels, sequence_length)`.
|
|
13880
|
+
*/
|
|
13881
|
+
async decode(inputs) {
|
|
13882
|
+
return await sessionRun(this.sessions['decoder_model'], inputs);
|
|
13883
|
+
}
|
|
13884
|
+
}
|
|
13885
|
+
|
|
13886
|
+
class SnacEncoderModel extends SnacPreTrainedModel {
|
|
13887
|
+
/** @type {typeof PreTrainedModel.from_pretrained} */
|
|
13888
|
+
static async from_pretrained(pretrained_model_name_or_path, options = {}) {
|
|
13889
|
+
return super.from_pretrained(pretrained_model_name_or_path, {
|
|
13890
|
+
...options,
|
|
13891
|
+
// Update default model file name if not provided
|
|
13892
|
+
model_file_name: options.model_file_name ?? 'encoder_model',
|
|
13893
|
+
});
|
|
13894
|
+
}
|
|
13895
|
+
}
|
|
13896
|
+
class SnacDecoderModel extends SnacPreTrainedModel {
|
|
13897
|
+
/** @type {typeof PreTrainedModel.from_pretrained} */
|
|
13898
|
+
static async from_pretrained(pretrained_model_name_or_path, options = {}) {
|
|
13899
|
+
return super.from_pretrained(pretrained_model_name_or_path, {
|
|
13900
|
+
...options,
|
|
13901
|
+
// Update default model file name if not provided
|
|
13902
|
+
model_file_name: options.model_file_name ?? 'decoder_model',
|
|
13903
|
+
});
|
|
13904
|
+
}
|
|
13905
|
+
}
|
|
13906
|
+
//////////////////////////////////////////////////
|
|
13907
|
+
|
|
13775
13908
|
//////////////////////////////////////////////////
|
|
13776
13909
|
// AutoModels, used to simplify construction of PreTrainedModels
|
|
13777
13910
|
// (uses config to instantiate correct class)
|
|
@@ -13892,6 +14025,8 @@ const MODEL_MAPPING_NAMES_ENCODER_ONLY = new Map([
|
|
|
13892
14025
|
|
|
13893
14026
|
['detr', ['DetrModel', DetrModel]],
|
|
13894
14027
|
['rt_detr', ['RTDetrModel', RTDetrModel]],
|
|
14028
|
+
['rt_detr_v2', ['RTDetrV2Model', RTDetrV2Model]],
|
|
14029
|
+
['rf_detr', ['RFDetrModel', RFDetrModel]],
|
|
13895
14030
|
['table-transformer', ['TableTransformerModel', TableTransformerModel]],
|
|
13896
14031
|
['vit', ['ViTModel', ViTModel]],
|
|
13897
14032
|
['ijepa', ['IJepaModel', IJepaModel]],
|
|
@@ -13953,6 +14088,7 @@ const MODEL_MAPPING_NAMES_ENCODER_DECODER = new Map([
|
|
|
13953
14088
|
const MODEL_MAPPING_NAMES_AUTO_ENCODER = new Map([
|
|
13954
14089
|
['mimi', ['MimiModel', MimiModel]],
|
|
13955
14090
|
['dac', ['DacModel', DacModel]],
|
|
14091
|
+
['snac', ['SnacModel', SnacModel]],
|
|
13956
14092
|
]);
|
|
13957
14093
|
|
|
13958
14094
|
const MODEL_MAPPING_NAMES_DECODER_ONLY = new Map([
|
|
@@ -13973,6 +14109,7 @@ const MODEL_MAPPING_NAMES_DECODER_ONLY = new Map([
|
|
|
13973
14109
|
['cohere', ['CohereModel', CohereModel]],
|
|
13974
14110
|
['gemma', ['GemmaModel', GemmaModel]],
|
|
13975
14111
|
['gemma2', ['Gemma2Model', Gemma2Model]],
|
|
14112
|
+
['gemma3_text', ['Gemma3Model', Gemma3Model]],
|
|
13976
14113
|
['helium', ['HeliumModel', HeliumModel]],
|
|
13977
14114
|
['glm', ['GlmModel', GlmModel]],
|
|
13978
14115
|
['openelm', ['OpenELMModel', OpenELMModel]],
|
|
@@ -14072,6 +14209,7 @@ const MODEL_FOR_CAUSAL_LM_MAPPING_NAMES = new Map([
|
|
|
14072
14209
|
['cohere', ['CohereForCausalLM', CohereForCausalLM]],
|
|
14073
14210
|
['gemma', ['GemmaForCausalLM', GemmaForCausalLM]],
|
|
14074
14211
|
['gemma2', ['Gemma2ForCausalLM', Gemma2ForCausalLM]],
|
|
14212
|
+
['gemma3_text', ['Gemma3ForCausalLM', Gemma3ForCausalLM]],
|
|
14075
14213
|
['helium', ['HeliumForCausalLM', HeliumForCausalLM]],
|
|
14076
14214
|
['glm', ['GlmForCausalLM', GlmForCausalLM]],
|
|
14077
14215
|
['openelm', ['OpenELMForCausalLM', OpenELMForCausalLM]],
|
|
@@ -14188,6 +14326,8 @@ const MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES = new Map([
|
|
|
14188
14326
|
const MODEL_FOR_OBJECT_DETECTION_MAPPING_NAMES = new Map([
|
|
14189
14327
|
['detr', ['DetrForObjectDetection', DetrForObjectDetection]],
|
|
14190
14328
|
['rt_detr', ['RTDetrForObjectDetection', RTDetrForObjectDetection]],
|
|
14329
|
+
['rt_detr_v2', ['RTDetrV2ForObjectDetection', RTDetrV2ForObjectDetection]],
|
|
14330
|
+
['rf_detr', ['RFDetrForObjectDetection', RFDetrForObjectDetection]],
|
|
14191
14331
|
['table-transformer', ['TableTransformerForObjectDetection', TableTransformerForObjectDetection]],
|
|
14192
14332
|
['yolos', ['YolosForObjectDetection', YolosForObjectDetection]],
|
|
14193
14333
|
]);
|
|
@@ -14273,6 +14413,8 @@ const MODEL_FOR_DEPTH_ESTIMATION_MAPPING_NAMES = new Map([
|
|
|
14273
14413
|
['glpn', ['GLPNForDepthEstimation', GLPNForDepthEstimation]],
|
|
14274
14414
|
['sapiens', ['SapiensForDepthEstimation', SapiensForDepthEstimation]],
|
|
14275
14415
|
['depth_pro', ['DepthProForDepthEstimation', DepthProForDepthEstimation]],
|
|
14416
|
+
['metric3d', ['Metric3DForDepthEstimation', Metric3DForDepthEstimation]],
|
|
14417
|
+
['metric3dv2', ['Metric3Dv2ForDepthEstimation', Metric3Dv2ForDepthEstimation]],
|
|
14276
14418
|
])
|
|
14277
14419
|
|
|
14278
14420
|
const MODEL_FOR_NORMAL_ESTIMATION_MAPPING_NAMES = new Map([
|
|
@@ -14358,6 +14500,8 @@ const CUSTOM_MAPPING = [
|
|
|
14358
14500
|
['DacDecoderModel', DacDecoderModel, MODEL_TYPES.EncoderOnly],
|
|
14359
14501
|
['MimiEncoderModel', MimiEncoderModel, MODEL_TYPES.EncoderOnly],
|
|
14360
14502
|
['MimiDecoderModel', MimiDecoderModel, MODEL_TYPES.EncoderOnly],
|
|
14503
|
+
['SnacEncoderModel', SnacEncoderModel, MODEL_TYPES.EncoderOnly],
|
|
14504
|
+
['SnacDecoderModel', SnacDecoderModel, MODEL_TYPES.EncoderOnly],
|
|
14361
14505
|
]
|
|
14362
14506
|
for (const [name, model, type] of CUSTOM_MAPPING) {
|
|
14363
14507
|
MODEL_TYPE_MAPPING.set(name, type);
|
|
@@ -15656,14 +15800,15 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
15656
15800
|
/* harmony export */ ClapFeatureExtractor: () => (/* reexport safe */ _clap_feature_extraction_clap_js__WEBPACK_IMPORTED_MODULE_2__.ClapFeatureExtractor),
|
|
15657
15801
|
/* harmony export */ DacFeatureExtractor: () => (/* reexport safe */ _dac_feature_extraction_dac_js__WEBPACK_IMPORTED_MODULE_3__.DacFeatureExtractor),
|
|
15658
15802
|
/* harmony export */ EncodecFeatureExtractor: () => (/* reexport safe */ _encodec_feature_extraction_encodec_js__WEBPACK_IMPORTED_MODULE_1__.EncodecFeatureExtractor),
|
|
15659
|
-
/* harmony export */ ImageFeatureExtractor: () => (/* reexport safe */
|
|
15803
|
+
/* harmony export */ ImageFeatureExtractor: () => (/* reexport safe */ _base_image_processors_utils_js__WEBPACK_IMPORTED_MODULE_12__.ImageProcessor),
|
|
15660
15804
|
/* harmony export */ MoonshineFeatureExtractor: () => (/* reexport safe */ _moonshine_feature_extraction_moonshine_js__WEBPACK_IMPORTED_MODULE_4__.MoonshineFeatureExtractor),
|
|
15661
15805
|
/* harmony export */ PyAnnoteFeatureExtractor: () => (/* reexport safe */ _pyannote_feature_extraction_pyannote_js__WEBPACK_IMPORTED_MODULE_5__.PyAnnoteFeatureExtractor),
|
|
15662
15806
|
/* harmony export */ SeamlessM4TFeatureExtractor: () => (/* reexport safe */ _seamless_m4t_feature_extraction_seamless_m4t_js__WEBPACK_IMPORTED_MODULE_6__.SeamlessM4TFeatureExtractor),
|
|
15663
|
-
/* harmony export */
|
|
15664
|
-
/* harmony export */
|
|
15665
|
-
/* harmony export */
|
|
15666
|
-
/* harmony export */
|
|
15807
|
+
/* harmony export */ SnacFeatureExtractor: () => (/* reexport safe */ _snac_feature_extraction_snac_js__WEBPACK_IMPORTED_MODULE_7__.SnacFeatureExtractor),
|
|
15808
|
+
/* harmony export */ SpeechT5FeatureExtractor: () => (/* reexport safe */ _speecht5_feature_extraction_speecht5_js__WEBPACK_IMPORTED_MODULE_8__.SpeechT5FeatureExtractor),
|
|
15809
|
+
/* harmony export */ Wav2Vec2FeatureExtractor: () => (/* reexport safe */ _wav2vec2_feature_extraction_wav2vec2_js__WEBPACK_IMPORTED_MODULE_9__.Wav2Vec2FeatureExtractor),
|
|
15810
|
+
/* harmony export */ WeSpeakerFeatureExtractor: () => (/* reexport safe */ _wespeaker_feature_extraction_wespeaker_js__WEBPACK_IMPORTED_MODULE_10__.WeSpeakerFeatureExtractor),
|
|
15811
|
+
/* harmony export */ WhisperFeatureExtractor: () => (/* reexport safe */ _whisper_feature_extraction_whisper_js__WEBPACK_IMPORTED_MODULE_11__.WhisperFeatureExtractor)
|
|
15667
15812
|
/* harmony export */ });
|
|
15668
15813
|
/* harmony import */ var _audio_spectrogram_transformer_feature_extraction_audio_spectrogram_transformer_js__WEBPACK_IMPORTED_MODULE_0__ = __webpack_require__(/*! ./audio_spectrogram_transformer/feature_extraction_audio_spectrogram_transformer.js */ "./src/models/audio_spectrogram_transformer/feature_extraction_audio_spectrogram_transformer.js");
|
|
15669
15814
|
/* harmony import */ var _encodec_feature_extraction_encodec_js__WEBPACK_IMPORTED_MODULE_1__ = __webpack_require__(/*! ./encodec/feature_extraction_encodec.js */ "./src/models/encodec/feature_extraction_encodec.js");
|
|
@@ -15672,11 +15817,13 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
15672
15817
|
/* harmony import */ var _moonshine_feature_extraction_moonshine_js__WEBPACK_IMPORTED_MODULE_4__ = __webpack_require__(/*! ./moonshine/feature_extraction_moonshine.js */ "./src/models/moonshine/feature_extraction_moonshine.js");
|
|
15673
15818
|
/* harmony import */ var _pyannote_feature_extraction_pyannote_js__WEBPACK_IMPORTED_MODULE_5__ = __webpack_require__(/*! ./pyannote/feature_extraction_pyannote.js */ "./src/models/pyannote/feature_extraction_pyannote.js");
|
|
15674
15819
|
/* harmony import */ var _seamless_m4t_feature_extraction_seamless_m4t_js__WEBPACK_IMPORTED_MODULE_6__ = __webpack_require__(/*! ./seamless_m4t/feature_extraction_seamless_m4t.js */ "./src/models/seamless_m4t/feature_extraction_seamless_m4t.js");
|
|
15675
|
-
/* harmony import */ var
|
|
15676
|
-
/* harmony import */ var
|
|
15677
|
-
/* harmony import */ var
|
|
15678
|
-
/* harmony import */ var
|
|
15679
|
-
/* harmony import */ var
|
|
15820
|
+
/* harmony import */ var _snac_feature_extraction_snac_js__WEBPACK_IMPORTED_MODULE_7__ = __webpack_require__(/*! ./snac/feature_extraction_snac.js */ "./src/models/snac/feature_extraction_snac.js");
|
|
15821
|
+
/* harmony import */ var _speecht5_feature_extraction_speecht5_js__WEBPACK_IMPORTED_MODULE_8__ = __webpack_require__(/*! ./speecht5/feature_extraction_speecht5.js */ "./src/models/speecht5/feature_extraction_speecht5.js");
|
|
15822
|
+
/* harmony import */ var _wav2vec2_feature_extraction_wav2vec2_js__WEBPACK_IMPORTED_MODULE_9__ = __webpack_require__(/*! ./wav2vec2/feature_extraction_wav2vec2.js */ "./src/models/wav2vec2/feature_extraction_wav2vec2.js");
|
|
15823
|
+
/* harmony import */ var _wespeaker_feature_extraction_wespeaker_js__WEBPACK_IMPORTED_MODULE_10__ = __webpack_require__(/*! ./wespeaker/feature_extraction_wespeaker.js */ "./src/models/wespeaker/feature_extraction_wespeaker.js");
|
|
15824
|
+
/* harmony import */ var _whisper_feature_extraction_whisper_js__WEBPACK_IMPORTED_MODULE_11__ = __webpack_require__(/*! ./whisper/feature_extraction_whisper.js */ "./src/models/whisper/feature_extraction_whisper.js");
|
|
15825
|
+
/* harmony import */ var _base_image_processors_utils_js__WEBPACK_IMPORTED_MODULE_12__ = __webpack_require__(/*! ../base/image_processors_utils.js */ "./src/base/image_processors_utils.js");
|
|
15826
|
+
|
|
15680
15827
|
|
|
15681
15828
|
|
|
15682
15829
|
|
|
@@ -18666,6 +18813,24 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
18666
18813
|
|
|
18667
18814
|
|
|
18668
18815
|
|
|
18816
|
+
/***/ }),
|
|
18817
|
+
|
|
18818
|
+
/***/ "./src/models/snac/feature_extraction_snac.js":
|
|
18819
|
+
/*!****************************************************!*\
|
|
18820
|
+
!*** ./src/models/snac/feature_extraction_snac.js ***!
|
|
18821
|
+
\****************************************************/
|
|
18822
|
+
/***/ ((__unused_webpack___webpack_module__, __webpack_exports__, __webpack_require__) => {
|
|
18823
|
+
|
|
18824
|
+
__webpack_require__.r(__webpack_exports__);
|
|
18825
|
+
/* harmony export */ __webpack_require__.d(__webpack_exports__, {
|
|
18826
|
+
/* harmony export */ SnacFeatureExtractor: () => (/* binding */ SnacFeatureExtractor)
|
|
18827
|
+
/* harmony export */ });
|
|
18828
|
+
/* harmony import */ var _dac_feature_extraction_dac_js__WEBPACK_IMPORTED_MODULE_0__ = __webpack_require__(/*! ../dac/feature_extraction_dac.js */ "./src/models/dac/feature_extraction_dac.js");
|
|
18829
|
+
|
|
18830
|
+
|
|
18831
|
+
class SnacFeatureExtractor extends _dac_feature_extraction_dac_js__WEBPACK_IMPORTED_MODULE_0__.DacFeatureExtractor { }
|
|
18832
|
+
|
|
18833
|
+
|
|
18669
18834
|
/***/ }),
|
|
18670
18835
|
|
|
18671
18836
|
/***/ "./src/models/speecht5/feature_extraction_speecht5.js":
|
|
@@ -19905,16 +20070,16 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
19905
20070
|
/* harmony import */ var _utils_image_js__WEBPACK_IMPORTED_MODULE_9__ = __webpack_require__(/*! ./utils/image.js */ "./src/utils/image.js");
|
|
19906
20071
|
/**
|
|
19907
20072
|
* @file Pipelines provide a high-level, easy to use, API for running machine learning models.
|
|
19908
|
-
*
|
|
20073
|
+
*
|
|
19909
20074
|
* **Example:** Instantiate pipeline using the `pipeline` function.
|
|
19910
20075
|
* ```javascript
|
|
19911
20076
|
* import { pipeline } from '@huggingface/transformers';
|
|
19912
|
-
*
|
|
20077
|
+
*
|
|
19913
20078
|
* const classifier = await pipeline('sentiment-analysis');
|
|
19914
20079
|
* const output = await classifier('I love transformers!');
|
|
19915
20080
|
* // [{'label': 'POSITIVE', 'score': 0.999817686}]
|
|
19916
20081
|
* ```
|
|
19917
|
-
*
|
|
20082
|
+
*
|
|
19918
20083
|
* @module pipelines
|
|
19919
20084
|
*/
|
|
19920
20085
|
|
|
@@ -19933,7 +20098,7 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
19933
20098
|
|
|
19934
20099
|
|
|
19935
20100
|
/**
|
|
19936
|
-
* @typedef {string | RawImage | URL} ImageInput
|
|
20101
|
+
* @typedef {string | RawImage | URL | Blob | HTMLCanvasElement | OffscreenCanvas} ImageInput
|
|
19937
20102
|
* @typedef {ImageInput|ImageInput[]} ImagePipelineInputs
|
|
19938
20103
|
*/
|
|
19939
20104
|
|
|
@@ -20007,7 +20172,7 @@ function get_bounding_box(box, asInteger) {
|
|
|
20007
20172
|
/**
|
|
20008
20173
|
* @callback DisposeType Disposes the item.
|
|
20009
20174
|
* @returns {Promise<void>} A promise that resolves when the item has been disposed.
|
|
20010
|
-
*
|
|
20175
|
+
*
|
|
20011
20176
|
* @typedef {Object} Disposable
|
|
20012
20177
|
* @property {DisposeType} dispose A promise that resolves when the pipeline has been disposed.
|
|
20013
20178
|
*/
|
|
@@ -20044,7 +20209,7 @@ class Pipeline extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_4__.Callable {
|
|
|
20044
20209
|
* @property {string} task The task of the pipeline. Useful for specifying subtasks.
|
|
20045
20210
|
* @property {PreTrainedModel} model The model used by the pipeline.
|
|
20046
20211
|
* @property {PreTrainedTokenizer} tokenizer The tokenizer used by the pipeline.
|
|
20047
|
-
*
|
|
20212
|
+
*
|
|
20048
20213
|
* @typedef {ModelTokenizerConstructorArgs} TextPipelineConstructorArgs An object used to instantiate a text-based pipeline.
|
|
20049
20214
|
*/
|
|
20050
20215
|
|
|
@@ -20053,7 +20218,7 @@ class Pipeline extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_4__.Callable {
|
|
|
20053
20218
|
* @property {string} task The task of the pipeline. Useful for specifying subtasks.
|
|
20054
20219
|
* @property {PreTrainedModel} model The model used by the pipeline.
|
|
20055
20220
|
* @property {Processor} processor The processor used by the pipeline.
|
|
20056
|
-
*
|
|
20221
|
+
*
|
|
20057
20222
|
* @typedef {ModelProcessorConstructorArgs} AudioPipelineConstructorArgs An object used to instantiate an audio-based pipeline.
|
|
20058
20223
|
* @typedef {ModelProcessorConstructorArgs} ImagePipelineConstructorArgs An object used to instantiate an image-based pipeline.
|
|
20059
20224
|
*/
|
|
@@ -20065,7 +20230,7 @@ class Pipeline extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_4__.Callable {
|
|
|
20065
20230
|
* @property {PreTrainedModel} model The model used by the pipeline.
|
|
20066
20231
|
* @property {PreTrainedTokenizer} tokenizer The tokenizer used by the pipeline.
|
|
20067
20232
|
* @property {Processor} processor The processor used by the pipeline.
|
|
20068
|
-
*
|
|
20233
|
+
*
|
|
20069
20234
|
* @typedef {ModelTokenizerProcessorConstructorArgs} TextAudioPipelineConstructorArgs An object used to instantiate a text- and audio-based pipeline.
|
|
20070
20235
|
* @typedef {ModelTokenizerProcessorConstructorArgs} TextImagePipelineConstructorArgs An object used to instantiate a text- and image-based pipeline.
|
|
20071
20236
|
*/
|
|
@@ -20075,15 +20240,15 @@ class Pipeline extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_4__.Callable {
|
|
|
20075
20240
|
* @property {string} label The label predicted.
|
|
20076
20241
|
* @property {number} score The corresponding probability.
|
|
20077
20242
|
* @typedef {TextClassificationSingle[]} TextClassificationOutput
|
|
20078
|
-
*
|
|
20243
|
+
*
|
|
20079
20244
|
* @typedef {Object} TextClassificationPipelineOptions Parameters specific to text classification pipelines.
|
|
20080
20245
|
* @property {number} [top_k=1] The number of top predictions to be returned.
|
|
20081
|
-
*
|
|
20246
|
+
*
|
|
20082
20247
|
* @callback TextClassificationPipelineCallback Classify the text(s) given as inputs.
|
|
20083
20248
|
* @param {string|string[]} texts The input text(s) to be classified.
|
|
20084
20249
|
* @param {TextClassificationPipelineOptions} [options] The options to use for text classification.
|
|
20085
20250
|
* @returns {Promise<TextClassificationOutput|TextClassificationOutput[]>} An array or object containing the predicted labels and scores.
|
|
20086
|
-
*
|
|
20251
|
+
*
|
|
20087
20252
|
* @typedef {TextPipelineConstructorArgs & TextClassificationPipelineCallback & Disposable} TextClassificationPipelineType
|
|
20088
20253
|
*/
|
|
20089
20254
|
|
|
@@ -20096,7 +20261,7 @@ class Pipeline extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_4__.Callable {
|
|
|
20096
20261
|
* const output = await classifier('I love transformers!');
|
|
20097
20262
|
* // [{ label: 'POSITIVE', score: 0.999788761138916 }]
|
|
20098
20263
|
* ```
|
|
20099
|
-
*
|
|
20264
|
+
*
|
|
20100
20265
|
* **Example:** Multilingual sentiment-analysis w/ `Xenova/bert-base-multilingual-uncased-sentiment` (and return top 5 classes).
|
|
20101
20266
|
* ```javascript
|
|
20102
20267
|
* const classifier = await pipeline('sentiment-analysis', 'Xenova/bert-base-multilingual-uncased-sentiment');
|
|
@@ -20109,7 +20274,7 @@ class Pipeline extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_4__.Callable {
|
|
|
20109
20274
|
* // { label: '2 stars', score: 0.0009423971059732139 }
|
|
20110
20275
|
* // ]
|
|
20111
20276
|
* ```
|
|
20112
|
-
*
|
|
20277
|
+
*
|
|
20113
20278
|
* **Example:** Toxic comment classification w/ `Xenova/toxic-bert` (and return all classes).
|
|
20114
20279
|
* ```javascript
|
|
20115
20280
|
* const classifier = await pipeline('text-classification', 'Xenova/toxic-bert');
|
|
@@ -20194,21 +20359,21 @@ class TextClassificationPipeline extends (/** @type {new (options: TextPipelineC
|
|
|
20194
20359
|
* @property {number} [start] The index of the start of the corresponding entity in the sentence.
|
|
20195
20360
|
* @property {number} [end] The index of the end of the corresponding entity in the sentence.
|
|
20196
20361
|
* @typedef {TokenClassificationSingle[]} TokenClassificationOutput
|
|
20197
|
-
*
|
|
20362
|
+
*
|
|
20198
20363
|
* @typedef {Object} TokenClassificationPipelineOptions Parameters specific to token classification pipelines.
|
|
20199
20364
|
* @property {string[]} [ignore_labels] A list of labels to ignore.
|
|
20200
|
-
*
|
|
20365
|
+
*
|
|
20201
20366
|
* @callback TokenClassificationPipelineCallback Classify each token of the text(s) given as inputs.
|
|
20202
20367
|
* @param {string|string[]} texts One or several texts (or one list of texts) for token classification.
|
|
20203
20368
|
* @param {TokenClassificationPipelineOptions} [options] The options to use for token classification.
|
|
20204
20369
|
* @returns {Promise<TokenClassificationOutput|TokenClassificationOutput[]>} The result.
|
|
20205
|
-
*
|
|
20370
|
+
*
|
|
20206
20371
|
* @typedef {TextPipelineConstructorArgs & TokenClassificationPipelineCallback & Disposable} TokenClassificationPipelineType
|
|
20207
20372
|
*/
|
|
20208
20373
|
|
|
20209
20374
|
/**
|
|
20210
20375
|
* Named Entity Recognition pipeline using any `ModelForTokenClassification`.
|
|
20211
|
-
*
|
|
20376
|
+
*
|
|
20212
20377
|
* **Example:** Perform named entity recognition with `Xenova/bert-base-NER`.
|
|
20213
20378
|
* ```javascript
|
|
20214
20379
|
* const classifier = await pipeline('token-classification', 'Xenova/bert-base-NER');
|
|
@@ -20218,7 +20383,7 @@ class TextClassificationPipeline extends (/** @type {new (options: TextPipelineC
|
|
|
20218
20383
|
* // { entity: 'B-LOC', score: 0.9994474053382874, index: 9, word: 'London' }
|
|
20219
20384
|
* // ]
|
|
20220
20385
|
* ```
|
|
20221
|
-
*
|
|
20386
|
+
*
|
|
20222
20387
|
* **Example:** Perform named entity recognition with `Xenova/bert-base-NER` (and return all labels).
|
|
20223
20388
|
* ```javascript
|
|
20224
20389
|
* const classifier = await pipeline('token-classification', 'Xenova/bert-base-NER');
|
|
@@ -20314,22 +20479,22 @@ class TokenClassificationPipeline extends (/** @type {new (options: TextPipeline
|
|
|
20314
20479
|
* @property {number} [start] The character start index of the answer (in the tokenized version of the input).
|
|
20315
20480
|
* @property {number} [end] The character end index of the answer (in the tokenized version of the input).
|
|
20316
20481
|
* @property {string} answer The answer to the question.
|
|
20317
|
-
*
|
|
20482
|
+
*
|
|
20318
20483
|
* @typedef {Object} QuestionAnsweringPipelineOptions Parameters specific to question answering pipelines.
|
|
20319
20484
|
* @property {number} [top_k=1] The number of top answer predictions to be returned.
|
|
20320
|
-
*
|
|
20485
|
+
*
|
|
20321
20486
|
* @callback QuestionAnsweringPipelineCallback Answer the question(s) given as inputs by using the context(s).
|
|
20322
20487
|
* @param {string|string[]} question One or several question(s) (must be used in conjunction with the `context` argument).
|
|
20323
20488
|
* @param {string|string[]} context One or several context(s) associated with the question(s) (must be used in conjunction with the `question` argument).
|
|
20324
20489
|
* @param {QuestionAnsweringPipelineOptions} [options] The options to use for question answering.
|
|
20325
20490
|
* @returns {Promise<QuestionAnsweringOutput|QuestionAnsweringOutput[]>} An array or object containing the predicted answers and scores.
|
|
20326
|
-
*
|
|
20491
|
+
*
|
|
20327
20492
|
* @typedef {TextPipelineConstructorArgs & QuestionAnsweringPipelineCallback & Disposable} QuestionAnsweringPipelineType
|
|
20328
20493
|
*/
|
|
20329
20494
|
|
|
20330
20495
|
/**
|
|
20331
20496
|
* Question Answering pipeline using any `ModelForQuestionAnswering`.
|
|
20332
|
-
*
|
|
20497
|
+
*
|
|
20333
20498
|
* **Example:** Run question answering with `Xenova/distilbert-base-uncased-distilled-squad`.
|
|
20334
20499
|
* ```javascript
|
|
20335
20500
|
* const answerer = await pipeline('question-answering', 'Xenova/distilbert-base-uncased-distilled-squad');
|
|
@@ -20454,10 +20619,10 @@ class QuestionAnsweringPipeline extends (/** @type {new (options: TextPipelineCo
|
|
|
20454
20619
|
* @property {number} token The predicted token id (to replace the masked one).
|
|
20455
20620
|
* @property {string} token_str The predicted token (to replace the masked one).
|
|
20456
20621
|
* @typedef {FillMaskSingle[]} FillMaskOutput
|
|
20457
|
-
*
|
|
20622
|
+
*
|
|
20458
20623
|
* @typedef {Object} FillMaskPipelineOptions Parameters specific to fill mask pipelines.
|
|
20459
20624
|
* @property {number} [top_k=5] When passed, overrides the number of predictions to return.
|
|
20460
|
-
*
|
|
20625
|
+
*
|
|
20461
20626
|
* @callback FillMaskPipelineCallback Fill the masked token in the text(s) given as inputs.
|
|
20462
20627
|
* @param {string|string[]} texts One or several texts (or one list of prompts) with masked tokens.
|
|
20463
20628
|
* @param {FillMaskPipelineOptions} [options] The options to use for masked language modelling.
|
|
@@ -20465,13 +20630,13 @@ class QuestionAnsweringPipeline extends (/** @type {new (options: TextPipelineCo
|
|
|
20465
20630
|
* and the sequence with the predicted token filled in, or an array of such arrays (one for each input text).
|
|
20466
20631
|
* If only one input text is given, the output will be an array of objects.
|
|
20467
20632
|
* @throws {Error} When the mask token is not found in the input text.
|
|
20468
|
-
*
|
|
20633
|
+
*
|
|
20469
20634
|
* @typedef {TextPipelineConstructorArgs & FillMaskPipelineCallback & Disposable} FillMaskPipelineType
|
|
20470
20635
|
*/
|
|
20471
20636
|
|
|
20472
20637
|
/**
|
|
20473
20638
|
* Masked language modeling prediction pipeline using any `ModelWithLMHead`.
|
|
20474
|
-
*
|
|
20639
|
+
*
|
|
20475
20640
|
* **Example:** Perform masked language modelling (a.k.a. "fill-mask") with `Xenova/bert-base-uncased`.
|
|
20476
20641
|
* ```javascript
|
|
20477
20642
|
* const unmasker = await pipeline('fill-mask', 'Xenova/bert-base-cased');
|
|
@@ -20484,7 +20649,7 @@ class QuestionAnsweringPipeline extends (/** @type {new (options: TextPipelineCo
|
|
|
20484
20649
|
* // { token_str: 'life', score: 0.01859794743359089, token: 1297, sequence: 'The goal of life is life.' }
|
|
20485
20650
|
* // ]
|
|
20486
20651
|
* ```
|
|
20487
|
-
*
|
|
20652
|
+
*
|
|
20488
20653
|
* **Example:** Perform masked language modelling (a.k.a. "fill-mask") with `Xenova/bert-base-cased` (and return top result).
|
|
20489
20654
|
* ```javascript
|
|
20490
20655
|
* const unmasker = await pipeline('fill-mask', 'Xenova/bert-base-cased');
|
|
@@ -20561,18 +20726,18 @@ class FillMaskPipeline extends (/** @type {new (options: TextPipelineConstructor
|
|
|
20561
20726
|
* @typedef {Object} Text2TextGenerationSingle
|
|
20562
20727
|
* @property {string} generated_text The generated text.
|
|
20563
20728
|
* @typedef {Text2TextGenerationSingle[]} Text2TextGenerationOutput
|
|
20564
|
-
*
|
|
20729
|
+
*
|
|
20565
20730
|
* @callback Text2TextGenerationPipelineCallback Generate the output text(s) using text(s) given as inputs.
|
|
20566
20731
|
* @param {string|string[]} texts Input text for the encoder.
|
|
20567
20732
|
* @param {Partial<import('./generation/configuration_utils.js').GenerationConfig>} [options] Additional keyword arguments to pass along to the generate method of the model.
|
|
20568
20733
|
* @returns {Promise<Text2TextGenerationOutput|Text2TextGenerationOutput[]>}
|
|
20569
|
-
*
|
|
20734
|
+
*
|
|
20570
20735
|
* @typedef {TextPipelineConstructorArgs & Text2TextGenerationPipelineCallback & Disposable} Text2TextGenerationPipelineType
|
|
20571
20736
|
*/
|
|
20572
20737
|
|
|
20573
20738
|
/**
|
|
20574
20739
|
* Text2TextGenerationPipeline class for generating text using a model that performs text-to-text generation tasks.
|
|
20575
|
-
*
|
|
20740
|
+
*
|
|
20576
20741
|
* **Example:** Text-to-text generation w/ `Xenova/LaMini-Flan-T5-783M`.
|
|
20577
20742
|
* ```javascript
|
|
20578
20743
|
* const generator = await pipeline('text2text-generation', 'Xenova/LaMini-Flan-T5-783M');
|
|
@@ -20648,18 +20813,18 @@ class Text2TextGenerationPipeline extends (/** @type {new (options: TextPipeline
|
|
|
20648
20813
|
* @typedef {Object} SummarizationSingle
|
|
20649
20814
|
* @property {string} summary_text The summary text.
|
|
20650
20815
|
* @typedef {SummarizationSingle[]} SummarizationOutput
|
|
20651
|
-
*
|
|
20816
|
+
*
|
|
20652
20817
|
* @callback SummarizationPipelineCallback Summarize the text(s) given as inputs.
|
|
20653
20818
|
* @param {string|string[]} texts One or several articles (or one list of articles) to summarize.
|
|
20654
20819
|
* @param {import('./generation/configuration_utils.js').GenerationConfig} [options] Additional keyword arguments to pass along to the generate method of the model.
|
|
20655
20820
|
* @returns {Promise<SummarizationOutput|SummarizationOutput[]>}
|
|
20656
|
-
*
|
|
20821
|
+
*
|
|
20657
20822
|
* @typedef {TextPipelineConstructorArgs & SummarizationPipelineCallback & Disposable} SummarizationPipelineType
|
|
20658
20823
|
*/
|
|
20659
20824
|
|
|
20660
20825
|
/**
|
|
20661
20826
|
* A pipeline for summarization tasks, inheriting from Text2TextGenerationPipeline.
|
|
20662
|
-
*
|
|
20827
|
+
*
|
|
20663
20828
|
* **Example:** Summarization w/ `Xenova/distilbart-cnn-6-6`.
|
|
20664
20829
|
* ```javascript
|
|
20665
20830
|
* const generator = await pipeline('summarization', 'Xenova/distilbart-cnn-6-6');
|
|
@@ -20695,23 +20860,23 @@ class SummarizationPipeline extends (/** @type {new (options: TextPipelineConstr
|
|
|
20695
20860
|
* @typedef {Object} TranslationSingle
|
|
20696
20861
|
* @property {string} translation_text The translated text.
|
|
20697
20862
|
* @typedef {TranslationSingle[]} TranslationOutput
|
|
20698
|
-
*
|
|
20863
|
+
*
|
|
20699
20864
|
* @callback TranslationPipelineCallback Translate the text(s) given as inputs.
|
|
20700
20865
|
* @param {string|string[]} texts Texts to be translated.
|
|
20701
20866
|
* @param {import('./generation/configuration_utils.js').GenerationConfig} [options] Additional keyword arguments to pass along to the generate method of the model.
|
|
20702
20867
|
* @returns {Promise<TranslationOutput|TranslationOutput[]>}
|
|
20703
|
-
*
|
|
20868
|
+
*
|
|
20704
20869
|
* @typedef {TextPipelineConstructorArgs & TranslationPipelineCallback & Disposable} TranslationPipelineType
|
|
20705
20870
|
*/
|
|
20706
20871
|
|
|
20707
20872
|
/**
|
|
20708
20873
|
* Translates text from one language to another.
|
|
20709
|
-
*
|
|
20874
|
+
*
|
|
20710
20875
|
* **Example:** Multilingual translation w/ `Xenova/nllb-200-distilled-600M`.
|
|
20711
|
-
*
|
|
20876
|
+
*
|
|
20712
20877
|
* See [here](https://github.com/facebookresearch/flores/blob/main/flores200/README.md#languages-in-flores-200)
|
|
20713
20878
|
* for the full list of languages and their corresponding codes.
|
|
20714
|
-
*
|
|
20879
|
+
*
|
|
20715
20880
|
* ```javascript
|
|
20716
20881
|
* const translator = await pipeline('translation', 'Xenova/nllb-200-distilled-600M');
|
|
20717
20882
|
* const output = await translator('जीवन एक चॉकलेट बॉक्स की तरह है।', {
|
|
@@ -20720,12 +20885,12 @@ class SummarizationPipeline extends (/** @type {new (options: TextPipelineConstr
|
|
|
20720
20885
|
* });
|
|
20721
20886
|
* // [{ translation_text: 'La vie est comme une boîte à chocolat.' }]
|
|
20722
20887
|
* ```
|
|
20723
|
-
*
|
|
20888
|
+
*
|
|
20724
20889
|
* **Example:** Multilingual translation w/ `Xenova/m2m100_418M`.
|
|
20725
|
-
*
|
|
20890
|
+
*
|
|
20726
20891
|
* See [here](https://huggingface.co/facebook/m2m100_418M#languages-covered)
|
|
20727
20892
|
* for the full list of languages and their corresponding codes.
|
|
20728
|
-
*
|
|
20893
|
+
*
|
|
20729
20894
|
* ```javascript
|
|
20730
20895
|
* const translator = await pipeline('translation', 'Xenova/m2m100_418M');
|
|
20731
20896
|
* const output = await translator('生活就像一盒巧克力。', {
|
|
@@ -20734,12 +20899,12 @@ class SummarizationPipeline extends (/** @type {new (options: TextPipelineConstr
|
|
|
20734
20899
|
* });
|
|
20735
20900
|
* // [{ translation_text: 'Life is like a box of chocolate.' }]
|
|
20736
20901
|
* ```
|
|
20737
|
-
*
|
|
20902
|
+
*
|
|
20738
20903
|
* **Example:** Multilingual translation w/ `Xenova/mbart-large-50-many-to-many-mmt`.
|
|
20739
|
-
*
|
|
20904
|
+
*
|
|
20740
20905
|
* See [here](https://huggingface.co/facebook/mbart-large-50-many-to-many-mmt#languages-covered)
|
|
20741
20906
|
* for the full list of languages and their corresponding codes.
|
|
20742
|
-
*
|
|
20907
|
+
*
|
|
20743
20908
|
* ```javascript
|
|
20744
20909
|
* const translator = await pipeline('translation', 'Xenova/mbart-large-50-many-to-many-mmt');
|
|
20745
20910
|
* const output = await translator('संयुक्त राष्ट्र के प्रमुख का कहना है कि सीरिया में कोई सैन्य समाधान नहीं है', {
|
|
@@ -20768,21 +20933,21 @@ function isChat(x) {
|
|
|
20768
20933
|
|
|
20769
20934
|
/**
|
|
20770
20935
|
* @typedef {import('./tokenizers.js').Message[]} Chat
|
|
20771
|
-
*
|
|
20936
|
+
*
|
|
20772
20937
|
* @typedef {Object} TextGenerationSingle
|
|
20773
20938
|
* @property {string|Chat} generated_text The generated text.
|
|
20774
20939
|
* @typedef {TextGenerationSingle[]} TextGenerationOutput
|
|
20775
|
-
*
|
|
20940
|
+
*
|
|
20776
20941
|
* @typedef {Object} TextGenerationSpecificParams Parameters specific to text-generation pipelines.
|
|
20777
20942
|
* @property {boolean} [add_special_tokens] Whether or not to add special tokens when tokenizing the sequences.
|
|
20778
20943
|
* @property {boolean} [return_full_text=true] If set to `false` only added text is returned, otherwise the full text is returned.
|
|
20779
20944
|
* @typedef {import('./generation/configuration_utils.js').GenerationConfig & TextGenerationSpecificParams} TextGenerationConfig
|
|
20780
|
-
*
|
|
20945
|
+
*
|
|
20781
20946
|
* @callback TextGenerationPipelineCallback Complete the prompt(s) given as inputs.
|
|
20782
20947
|
* @param {string|string[]|Chat|Chat[]} texts One or several prompts (or one list of prompts) to complete.
|
|
20783
20948
|
* @param {Partial<TextGenerationConfig>} [options] Additional keyword arguments to pass along to the generate method of the model.
|
|
20784
20949
|
* @returns {Promise<TextGenerationOutput|TextGenerationOutput[]>} An array or object containing the generated texts.
|
|
20785
|
-
*
|
|
20950
|
+
*
|
|
20786
20951
|
* @typedef {TextPipelineConstructorArgs & TextGenerationPipelineCallback & Disposable} TextGenerationPipelineType
|
|
20787
20952
|
*/
|
|
20788
20953
|
|
|
@@ -20790,7 +20955,7 @@ function isChat(x) {
|
|
|
20790
20955
|
* Language generation pipeline using any `ModelWithLMHead` or `ModelForCausalLM`.
|
|
20791
20956
|
* This pipeline predicts the words that will follow a specified text prompt.
|
|
20792
20957
|
* NOTE: For the full list of generation parameters, see [`GenerationConfig`](./utils/generation#module_utils/generation.GenerationConfig).
|
|
20793
|
-
*
|
|
20958
|
+
*
|
|
20794
20959
|
* **Example:** Text generation with `Xenova/distilgpt2` (default settings).
|
|
20795
20960
|
* ```javascript
|
|
20796
20961
|
* const generator = await pipeline('text-generation', 'Xenova/distilgpt2');
|
|
@@ -20798,7 +20963,7 @@ function isChat(x) {
|
|
|
20798
20963
|
* const output = await generator(text);
|
|
20799
20964
|
* // [{ generated_text: "I enjoy walking with my cute dog, and I love to play with the other dogs." }]
|
|
20800
20965
|
* ```
|
|
20801
|
-
*
|
|
20966
|
+
*
|
|
20802
20967
|
* **Example:** Text generation with `Xenova/distilgpt2` (custom settings).
|
|
20803
20968
|
* ```javascript
|
|
20804
20969
|
* const generator = await pipeline('text-generation', 'Xenova/distilgpt2');
|
|
@@ -20817,7 +20982,7 @@ function isChat(x) {
|
|
|
20817
20982
|
* // "generated_text": "Once upon a time, there was an abundance of information about the most important and influential"
|
|
20818
20983
|
* // }]
|
|
20819
20984
|
* ```
|
|
20820
|
-
*
|
|
20985
|
+
*
|
|
20821
20986
|
* **Example:** Run code generation with `Xenova/codegen-350M-mono`.
|
|
20822
20987
|
* ```javascript
|
|
20823
20988
|
* const generator = await pipeline('text-generation', 'Xenova/codegen-350M-mono');
|
|
@@ -20936,7 +21101,7 @@ class TextGenerationPipeline extends (/** @type {new (options: TextPipelineConst
|
|
|
20936
21101
|
* @property {string} sequence The sequence for which this is the output.
|
|
20937
21102
|
* @property {string[]} labels The labels sorted by order of likelihood.
|
|
20938
21103
|
* @property {number[]} scores The probabilities for each of the labels.
|
|
20939
|
-
*
|
|
21104
|
+
*
|
|
20940
21105
|
* @typedef {Object} ZeroShotClassificationPipelineOptions Parameters specific to zero-shot classification pipelines.
|
|
20941
21106
|
* @property {string} [hypothesis_template="This example is {}."] The template used to turn each
|
|
20942
21107
|
* candidate label into an NLI-style hypothesis. The candidate label will replace the {} placeholder.
|
|
@@ -20944,14 +21109,14 @@ class TextGenerationPipeline extends (/** @type {new (options: TextPipelineConst
|
|
|
20944
21109
|
* If `false`, the scores are normalized such that the sum of the label likelihoods for each sequence
|
|
20945
21110
|
* is 1. If `true`, the labels are considered independent and probabilities are normalized for each
|
|
20946
21111
|
* candidate by doing a softmax of the entailment score vs. the contradiction score.
|
|
20947
|
-
*
|
|
21112
|
+
*
|
|
20948
21113
|
* @callback ZeroShotClassificationPipelineCallback Classify the sequence(s) given as inputs.
|
|
20949
21114
|
* @param {string|string[]} texts The sequence(s) to classify, will be truncated if the model input is too large.
|
|
20950
21115
|
* @param {string|string[]} candidate_labels The set of possible class labels to classify each sequence into.
|
|
20951
21116
|
* Can be a single label, a string of comma-separated labels, or a list of labels.
|
|
20952
21117
|
* @param {ZeroShotClassificationPipelineOptions} [options] The options to use for zero-shot classification.
|
|
20953
21118
|
* @returns {Promise<ZeroShotClassificationOutput|ZeroShotClassificationOutput[]>} An array or object containing the predicted labels and scores.
|
|
20954
|
-
*
|
|
21119
|
+
*
|
|
20955
21120
|
* @typedef {TextPipelineConstructorArgs & ZeroShotClassificationPipelineCallback & Disposable} ZeroShotClassificationPipelineType
|
|
20956
21121
|
*/
|
|
20957
21122
|
|
|
@@ -20960,7 +21125,7 @@ class TextGenerationPipeline extends (/** @type {new (options: TextPipelineConst
|
|
|
20960
21125
|
* trained on NLI (natural language inference) tasks. Equivalent of `text-classification`
|
|
20961
21126
|
* pipelines, but these models don't require a hardcoded number of potential classes, they
|
|
20962
21127
|
* can be chosen at runtime. It usually means it's slower but it is **much** more flexible.
|
|
20963
|
-
*
|
|
21128
|
+
*
|
|
20964
21129
|
* **Example:** Zero shot classification with `Xenova/mobilebert-uncased-mnli`.
|
|
20965
21130
|
* ```javascript
|
|
20966
21131
|
* const classifier = await pipeline('zero-shot-classification', 'Xenova/mobilebert-uncased-mnli');
|
|
@@ -20973,7 +21138,7 @@ class TextGenerationPipeline extends (/** @type {new (options: TextPipelineConst
|
|
|
20973
21138
|
* // scores: [ 0.5562091040482018, 0.1843621307860853, 0.13942646639336376, 0.12000229877234923 ]
|
|
20974
21139
|
* // }
|
|
20975
21140
|
* ```
|
|
20976
|
-
*
|
|
21141
|
+
*
|
|
20977
21142
|
* **Example:** Zero shot classification with `Xenova/nli-deberta-v3-xsmall` (multi-label).
|
|
20978
21143
|
* ```javascript
|
|
20979
21144
|
* const classifier = await pipeline('zero-shot-classification', 'Xenova/nli-deberta-v3-xsmall');
|
|
@@ -21087,20 +21252,20 @@ class ZeroShotClassificationPipeline extends (/** @type {new (options: TextPipel
|
|
|
21087
21252
|
* @property {'none'|'mean'|'cls'} [pooling="none"] The pooling method to use.
|
|
21088
21253
|
* @property {boolean} [normalize=false] Whether or not to normalize the embeddings in the last dimension.
|
|
21089
21254
|
* @property {boolean} [quantize=false] Whether or not to quantize the embeddings.
|
|
21090
|
-
* @property {'binary'|'ubinary'} [precision='binary'] The precision to use for quantization.
|
|
21091
|
-
*
|
|
21255
|
+
* @property {'binary'|'ubinary'} [precision='binary'] The precision to use for quantization.
|
|
21256
|
+
*
|
|
21092
21257
|
* @callback FeatureExtractionPipelineCallback Extract the features of the input(s).
|
|
21093
21258
|
* @param {string|string[]} texts One or several texts (or one list of texts) to get the features of.
|
|
21094
21259
|
* @param {FeatureExtractionPipelineOptions} [options] The options to use for feature extraction.
|
|
21095
21260
|
* @returns {Promise<Tensor>} The features computed by the model.
|
|
21096
|
-
*
|
|
21261
|
+
*
|
|
21097
21262
|
* @typedef {TextPipelineConstructorArgs & FeatureExtractionPipelineCallback & Disposable} FeatureExtractionPipelineType
|
|
21098
21263
|
*/
|
|
21099
21264
|
|
|
21100
21265
|
/**
|
|
21101
21266
|
* Feature extraction pipeline using no model head. This pipeline extracts the hidden
|
|
21102
21267
|
* states from the base transformer, which can be used as features in downstream tasks.
|
|
21103
|
-
*
|
|
21268
|
+
*
|
|
21104
21269
|
* **Example:** Run feature extraction with `bert-base-uncased` (without pooling/normalization).
|
|
21105
21270
|
* ```javascript
|
|
21106
21271
|
* const extractor = await pipeline('feature-extraction', 'Xenova/bert-base-uncased', { revision: 'default' });
|
|
@@ -21111,7 +21276,7 @@ class ZeroShotClassificationPipeline extends (/** @type {new (options: TextPipel
|
|
|
21111
21276
|
* // dims: [1, 8, 768]
|
|
21112
21277
|
* // }
|
|
21113
21278
|
* ```
|
|
21114
|
-
*
|
|
21279
|
+
*
|
|
21115
21280
|
* **Example:** Run feature extraction with `bert-base-uncased` (with pooling/normalization).
|
|
21116
21281
|
* ```javascript
|
|
21117
21282
|
* const extractor = await pipeline('feature-extraction', 'Xenova/bert-base-uncased', { revision: 'default' });
|
|
@@ -21122,7 +21287,7 @@ class ZeroShotClassificationPipeline extends (/** @type {new (options: TextPipel
|
|
|
21122
21287
|
* // dims: [1, 768]
|
|
21123
21288
|
* // }
|
|
21124
21289
|
* ```
|
|
21125
|
-
*
|
|
21290
|
+
*
|
|
21126
21291
|
* **Example:** Calculating embeddings with `sentence-transformers` models.
|
|
21127
21292
|
* ```javascript
|
|
21128
21293
|
* const extractor = await pipeline('feature-extraction', 'Xenova/all-MiniLM-L6-v2');
|
|
@@ -21203,19 +21368,19 @@ class FeatureExtractionPipeline extends (/** @type {new (options: TextPipelineCo
|
|
|
21203
21368
|
/**
|
|
21204
21369
|
* @typedef {Object} ImageFeatureExtractionPipelineOptions Parameters specific to image feature extraction pipelines.
|
|
21205
21370
|
* @property {boolean} [pool=null] Whether or not to return the pooled output. If set to `false`, the model will return the raw hidden states.
|
|
21206
|
-
*
|
|
21371
|
+
*
|
|
21207
21372
|
* @callback ImageFeatureExtractionPipelineCallback Extract the features of the input(s).
|
|
21208
21373
|
* @param {ImagePipelineInputs} images One or several images (or one list of images) to get the features of.
|
|
21209
21374
|
* @param {ImageFeatureExtractionPipelineOptions} [options] The options to use for image feature extraction.
|
|
21210
21375
|
* @returns {Promise<Tensor>} The image features computed by the model.
|
|
21211
|
-
*
|
|
21376
|
+
*
|
|
21212
21377
|
* @typedef {ImagePipelineConstructorArgs & ImageFeatureExtractionPipelineCallback & Disposable} ImageFeatureExtractionPipelineType
|
|
21213
21378
|
*/
|
|
21214
21379
|
|
|
21215
21380
|
/**
|
|
21216
21381
|
* Image feature extraction pipeline using no model head. This pipeline extracts the hidden
|
|
21217
21382
|
* states from the base transformer, which can be used as features in downstream tasks.
|
|
21218
|
-
*
|
|
21383
|
+
*
|
|
21219
21384
|
* **Example:** Perform image feature extraction with `Xenova/vit-base-patch16-224-in21k`.
|
|
21220
21385
|
* ```javascript
|
|
21221
21386
|
* const image_feature_extractor = await pipeline('image-feature-extraction', 'Xenova/vit-base-patch16-224-in21k');
|
|
@@ -21228,7 +21393,7 @@ class FeatureExtractionPipeline extends (/** @type {new (options: TextPipelineCo
|
|
|
21228
21393
|
* // size: 151296
|
|
21229
21394
|
* // }
|
|
21230
21395
|
* ```
|
|
21231
|
-
*
|
|
21396
|
+
*
|
|
21232
21397
|
* **Example:** Compute image embeddings with `Xenova/clip-vit-base-patch32`.
|
|
21233
21398
|
* ```javascript
|
|
21234
21399
|
* const image_feature_extractor = await pipeline('image-feature-extraction', 'Xenova/clip-vit-base-patch32');
|
|
@@ -21284,12 +21449,12 @@ class ImageFeatureExtractionPipeline extends (/** @type {new (options: ImagePipe
|
|
|
21284
21449
|
* @property {string} label The label predicted.
|
|
21285
21450
|
* @property {number} score The corresponding probability.
|
|
21286
21451
|
* @typedef {AudioClassificationSingle[]} AudioClassificationOutput
|
|
21287
|
-
*
|
|
21452
|
+
*
|
|
21288
21453
|
* @typedef {Object} AudioClassificationPipelineOptions Parameters specific to audio classification pipelines.
|
|
21289
21454
|
* @property {number} [top_k=5] The number of top labels that will be returned by the pipeline.
|
|
21290
21455
|
* If the provided number is `null` or higher than the number of labels available in the model configuration,
|
|
21291
21456
|
* it will default to the number of labels.
|
|
21292
|
-
*
|
|
21457
|
+
*
|
|
21293
21458
|
* @callback AudioClassificationPipelineCallback Classify the sequence(s) given as inputs.
|
|
21294
21459
|
* @param {AudioPipelineInputs} audio The input audio file(s) to be classified. The input is either:
|
|
21295
21460
|
* - `string` or `URL` that is the filename/URL of the audio file, the file will be read at the processor's sampling rate
|
|
@@ -21298,14 +21463,14 @@ class ImageFeatureExtractionPipeline extends (/** @type {new (options: ImagePipe
|
|
|
21298
21463
|
* - `Float32Array` or `Float64Array` of shape `(n, )`, representing the raw audio at the correct sampling rate (no further check will be done).
|
|
21299
21464
|
* @param {AudioClassificationPipelineOptions} [options] The options to use for audio classification.
|
|
21300
21465
|
* @returns {Promise<AudioClassificationOutput|AudioClassificationOutput[]>} An array or object containing the predicted labels and scores.
|
|
21301
|
-
*
|
|
21466
|
+
*
|
|
21302
21467
|
* @typedef {AudioPipelineConstructorArgs & AudioClassificationPipelineCallback & Disposable} AudioClassificationPipelineType
|
|
21303
21468
|
*/
|
|
21304
21469
|
|
|
21305
21470
|
/**
|
|
21306
21471
|
* Audio classification pipeline using any `AutoModelForAudioClassification`.
|
|
21307
21472
|
* This pipeline predicts the class of a raw waveform or an audio file.
|
|
21308
|
-
*
|
|
21473
|
+
*
|
|
21309
21474
|
* **Example:** Perform audio classification with `Xenova/wav2vec2-large-xlsr-53-gender-recognition-librispeech`.
|
|
21310
21475
|
* ```javascript
|
|
21311
21476
|
* const classifier = await pipeline('audio-classification', 'Xenova/wav2vec2-large-xlsr-53-gender-recognition-librispeech');
|
|
@@ -21316,7 +21481,7 @@ class ImageFeatureExtractionPipeline extends (/** @type {new (options: ImagePipe
|
|
|
21316
21481
|
* // { label: 'female', score: 0.001845747814513743 }
|
|
21317
21482
|
* // ]
|
|
21318
21483
|
* ```
|
|
21319
|
-
*
|
|
21484
|
+
*
|
|
21320
21485
|
* **Example:** Perform audio classification with `Xenova/ast-finetuned-audioset-10-10-0.4593` and return top 4 results.
|
|
21321
21486
|
* ```javascript
|
|
21322
21487
|
* const classifier = await pipeline('audio-classification', 'Xenova/ast-finetuned-audioset-10-10-0.4593');
|
|
@@ -21381,12 +21546,12 @@ class AudioClassificationPipeline extends (/** @type {new (options: AudioPipelin
|
|
|
21381
21546
|
* @typedef {Object} ZeroShotAudioClassificationOutput
|
|
21382
21547
|
* @property {string} label The label identified by the model. It is one of the suggested `candidate_label`.
|
|
21383
21548
|
* @property {number} score The score attributed by the model for that label (between 0 and 1).
|
|
21384
|
-
*
|
|
21549
|
+
*
|
|
21385
21550
|
* @typedef {Object} ZeroShotAudioClassificationPipelineOptions Parameters specific to zero-shot audio classification pipelines.
|
|
21386
21551
|
* @property {string} [hypothesis_template="This is a sound of {}."] The sentence used in conjunction with `candidate_labels`
|
|
21387
21552
|
* to attempt the audio classification by replacing the placeholder with the candidate_labels.
|
|
21388
21553
|
* Then likelihood is estimated by using `logits_per_audio`.
|
|
21389
|
-
*
|
|
21554
|
+
*
|
|
21390
21555
|
* @callback ZeroShotAudioClassificationPipelineCallback Classify the sequence(s) given as inputs.
|
|
21391
21556
|
* @param {AudioPipelineInputs} audio The input audio file(s) to be classified. The input is either:
|
|
21392
21557
|
* - `string` or `URL` that is the filename/URL of the audio file, the file will be read at the processor's sampling rate
|
|
@@ -21396,14 +21561,14 @@ class AudioClassificationPipeline extends (/** @type {new (options: AudioPipelin
|
|
|
21396
21561
|
* @param {string[]} candidate_labels The candidate labels for this audio.
|
|
21397
21562
|
* @param {ZeroShotAudioClassificationPipelineOptions} [options] The options to use for zero-shot audio classification.
|
|
21398
21563
|
* @returns {Promise<ZeroShotAudioClassificationOutput[]|ZeroShotAudioClassificationOutput[][]>} An array of objects containing the predicted labels and scores.
|
|
21399
|
-
*
|
|
21564
|
+
*
|
|
21400
21565
|
* @typedef {TextAudioPipelineConstructorArgs & ZeroShotAudioClassificationPipelineCallback & Disposable} ZeroShotAudioClassificationPipelineType
|
|
21401
21566
|
*/
|
|
21402
21567
|
|
|
21403
21568
|
/**
|
|
21404
21569
|
* Zero shot audio classification pipeline using `ClapModel`. This pipeline predicts the class of an audio when you
|
|
21405
21570
|
* provide an audio and a set of `candidate_labels`.
|
|
21406
|
-
*
|
|
21571
|
+
*
|
|
21407
21572
|
* **Example**: Perform zero-shot audio classification with `Xenova/clap-htsat-unfused`.
|
|
21408
21573
|
* ```javascript
|
|
21409
21574
|
* const classifier = await pipeline('zero-shot-audio-classification', 'Xenova/clap-htsat-unfused');
|
|
@@ -21436,7 +21601,7 @@ class ZeroShotAudioClassificationPipeline extends (/** @type {new (options: Text
|
|
|
21436
21601
|
audio = [/** @type {AudioInput} */ (audio)];
|
|
21437
21602
|
}
|
|
21438
21603
|
|
|
21439
|
-
// Insert label into hypothesis template
|
|
21604
|
+
// Insert label into hypothesis template
|
|
21440
21605
|
const texts = candidate_labels.map(
|
|
21441
21606
|
x => hypothesis_template.replace('{}', x)
|
|
21442
21607
|
);
|
|
@@ -21480,7 +21645,7 @@ class ZeroShotAudioClassificationPipeline extends (/** @type {new (options: Text
|
|
|
21480
21645
|
* @property {string} text The recognized text.
|
|
21481
21646
|
* @property {Chunk[]} [chunks] When using `return_timestamps`, the `chunks` will become a list
|
|
21482
21647
|
* containing all the various text chunks identified by the model.
|
|
21483
|
-
*
|
|
21648
|
+
*
|
|
21484
21649
|
* @typedef {Object} AutomaticSpeechRecognitionSpecificParams Parameters specific to automatic-speech-recognition pipelines.
|
|
21485
21650
|
* @property {boolean|'word'} [return_timestamps] Whether to return timestamps or not. Default is `false`.
|
|
21486
21651
|
* @property {number} [chunk_length_s] The length of audio chunks to process in seconds. Default is 0 (no chunking).
|
|
@@ -21490,7 +21655,7 @@ class ZeroShotAudioClassificationPipeline extends (/** @type {new (options: Text
|
|
|
21490
21655
|
* @property {string} [task] The task to perform. Default is `null`, meaning it should be auto-detected.
|
|
21491
21656
|
* @property {number} [num_frames] The number of frames in the input audio.
|
|
21492
21657
|
* @typedef {import('./generation/configuration_utils.js').GenerationConfig & AutomaticSpeechRecognitionSpecificParams} AutomaticSpeechRecognitionConfig
|
|
21493
|
-
*
|
|
21658
|
+
*
|
|
21494
21659
|
* @callback AutomaticSpeechRecognitionPipelineCallback Transcribe the audio sequence(s) given as inputs to text.
|
|
21495
21660
|
* @param {AudioPipelineInputs} audio The input audio file(s) to be transcribed. The input is either:
|
|
21496
21661
|
* - `string` or `URL` that is the filename/URL of the audio file, the file will be read at the processor's sampling rate
|
|
@@ -21499,7 +21664,7 @@ class ZeroShotAudioClassificationPipeline extends (/** @type {new (options: Text
|
|
|
21499
21664
|
* - `Float32Array` or `Float64Array` of shape `(n, )`, representing the raw audio at the correct sampling rate (no further check will be done).
|
|
21500
21665
|
* @param {Partial<AutomaticSpeechRecognitionConfig>} [options] Additional keyword arguments to pass along to the generate method of the model.
|
|
21501
21666
|
* @returns {Promise<AutomaticSpeechRecognitionOutput|AutomaticSpeechRecognitionOutput[]>} An object containing the transcription text and optionally timestamps if `return_timestamps` is `true`.
|
|
21502
|
-
*
|
|
21667
|
+
*
|
|
21503
21668
|
* @typedef {TextAudioPipelineConstructorArgs & AutomaticSpeechRecognitionPipelineCallback & Disposable} AutomaticSpeechRecognitionPipelineType
|
|
21504
21669
|
*/
|
|
21505
21670
|
|
|
@@ -21513,7 +21678,7 @@ class ZeroShotAudioClassificationPipeline extends (/** @type {new (options: Text
|
|
|
21513
21678
|
* const output = await transcriber(url);
|
|
21514
21679
|
* // { text: " And so my fellow Americans ask not what your country can do for you, ask what you can do for your country." }
|
|
21515
21680
|
* ```
|
|
21516
|
-
*
|
|
21681
|
+
*
|
|
21517
21682
|
* **Example:** Transcribe English w/ timestamps.
|
|
21518
21683
|
* ```javascript
|
|
21519
21684
|
* const transcriber = await pipeline('automatic-speech-recognition', 'Xenova/whisper-tiny.en');
|
|
@@ -21527,7 +21692,7 @@ class ZeroShotAudioClassificationPipeline extends (/** @type {new (options: Text
|
|
|
21527
21692
|
* // ]
|
|
21528
21693
|
* // }
|
|
21529
21694
|
* ```
|
|
21530
|
-
*
|
|
21695
|
+
*
|
|
21531
21696
|
* **Example:** Transcribe English w/ word-level timestamps.
|
|
21532
21697
|
* ```javascript
|
|
21533
21698
|
* const transcriber = await pipeline('automatic-speech-recognition', 'Xenova/whisper-tiny.en');
|
|
@@ -21546,7 +21711,7 @@ class ZeroShotAudioClassificationPipeline extends (/** @type {new (options: Text
|
|
|
21546
21711
|
* // ]
|
|
21547
21712
|
* // }
|
|
21548
21713
|
* ```
|
|
21549
|
-
*
|
|
21714
|
+
*
|
|
21550
21715
|
* **Example:** Transcribe French.
|
|
21551
21716
|
* ```javascript
|
|
21552
21717
|
* const transcriber = await pipeline('automatic-speech-recognition', 'Xenova/whisper-small');
|
|
@@ -21554,7 +21719,7 @@ class ZeroShotAudioClassificationPipeline extends (/** @type {new (options: Text
|
|
|
21554
21719
|
* const output = await transcriber(url, { language: 'french', task: 'transcribe' });
|
|
21555
21720
|
* // { text: " J'adore, j'aime, je n'aime pas, je déteste." }
|
|
21556
21721
|
* ```
|
|
21557
|
-
*
|
|
21722
|
+
*
|
|
21558
21723
|
* **Example:** Translate French to English.
|
|
21559
21724
|
* ```javascript
|
|
21560
21725
|
* const transcriber = await pipeline('automatic-speech-recognition', 'Xenova/whisper-small');
|
|
@@ -21562,7 +21727,7 @@ class ZeroShotAudioClassificationPipeline extends (/** @type {new (options: Text
|
|
|
21562
21727
|
* const output = await transcriber(url, { language: 'french', task: 'translate' });
|
|
21563
21728
|
* // { text: " I love, I like, I don't like, I hate." }
|
|
21564
21729
|
* ```
|
|
21565
|
-
*
|
|
21730
|
+
*
|
|
21566
21731
|
* **Example:** Transcribe/translate audio longer than 30 seconds.
|
|
21567
21732
|
* ```javascript
|
|
21568
21733
|
* const transcriber = await pipeline('automatic-speech-recognition', 'Xenova/whisper-tiny.en');
|
|
@@ -21785,18 +21950,18 @@ class AutomaticSpeechRecognitionPipeline extends (/** @type {new (options: TextA
|
|
|
21785
21950
|
* @typedef {Object} ImageToTextSingle
|
|
21786
21951
|
* @property {string} generated_text The generated text.
|
|
21787
21952
|
* @typedef {ImageToTextSingle[]} ImageToTextOutput
|
|
21788
|
-
*
|
|
21953
|
+
*
|
|
21789
21954
|
* @callback ImageToTextPipelineCallback Assign labels to the image(s) passed as inputs.
|
|
21790
21955
|
* @param {ImagePipelineInputs} texts The images to be captioned.
|
|
21791
21956
|
* @param {Partial<import('./generation/configuration_utils.js').GenerationConfig>} [options] Additional keyword arguments to pass along to the generate method of the model.
|
|
21792
21957
|
* @returns {Promise<ImageToTextOutput|ImageToTextOutput[]>} An object (or array of objects) containing the generated text(s).
|
|
21793
|
-
*
|
|
21958
|
+
*
|
|
21794
21959
|
* @typedef {TextImagePipelineConstructorArgs & ImageToTextPipelineCallback & Disposable} ImageToTextPipelineType
|
|
21795
21960
|
*/
|
|
21796
21961
|
|
|
21797
21962
|
/**
|
|
21798
21963
|
* Image To Text pipeline using a `AutoModelForVision2Seq`. This pipeline predicts a caption for a given image.
|
|
21799
|
-
*
|
|
21964
|
+
*
|
|
21800
21965
|
* **Example:** Generate a caption for an image w/ `Xenova/vit-gpt2-image-captioning`.
|
|
21801
21966
|
* ```javascript
|
|
21802
21967
|
* const captioner = await pipeline('image-to-text', 'Xenova/vit-gpt2-image-captioning');
|
|
@@ -21804,7 +21969,7 @@ class AutomaticSpeechRecognitionPipeline extends (/** @type {new (options: TextA
|
|
|
21804
21969
|
* const output = await captioner(url);
|
|
21805
21970
|
* // [{ generated_text: 'a cat laying on a couch with another cat' }]
|
|
21806
21971
|
* ```
|
|
21807
|
-
*
|
|
21972
|
+
*
|
|
21808
21973
|
* **Example:** Optical Character Recognition (OCR) w/ `Xenova/trocr-small-handwritten`.
|
|
21809
21974
|
* ```javascript
|
|
21810
21975
|
* const captioner = await pipeline('image-to-text', 'Xenova/trocr-small-handwritten');
|
|
@@ -21850,22 +22015,22 @@ class ImageToTextPipeline extends (/** @type {new (options: TextImagePipelineCon
|
|
|
21850
22015
|
* @property {string} label The label identified by the model.
|
|
21851
22016
|
* @property {number} score The score attributed by the model for that label.
|
|
21852
22017
|
* @typedef {ImageClassificationSingle[]} ImageClassificationOutput
|
|
21853
|
-
*
|
|
22018
|
+
*
|
|
21854
22019
|
* @typedef {Object} ImageClassificationPipelineOptions Parameters specific to image classification pipelines.
|
|
21855
|
-
* @property {number} [top_k=1] The number of top labels that will be returned by the pipeline.
|
|
21856
|
-
*
|
|
22020
|
+
* @property {number} [top_k=1] The number of top labels that will be returned by the pipeline.
|
|
22021
|
+
*
|
|
21857
22022
|
* @callback ImageClassificationPipelineCallback Assign labels to the image(s) passed as inputs.
|
|
21858
22023
|
* @param {ImagePipelineInputs} images The input images(s) to be classified.
|
|
21859
22024
|
* @param {ImageClassificationPipelineOptions} [options] The options to use for image classification.
|
|
21860
22025
|
* @returns {Promise<ImageClassificationOutput|ImageClassificationOutput[]>} An array or object containing the predicted labels and scores.
|
|
21861
|
-
*
|
|
22026
|
+
*
|
|
21862
22027
|
* @typedef {ImagePipelineConstructorArgs & ImageClassificationPipelineCallback & Disposable} ImageClassificationPipelineType
|
|
21863
22028
|
*/
|
|
21864
22029
|
|
|
21865
22030
|
/**
|
|
21866
22031
|
* Image classification pipeline using any `AutoModelForImageClassification`.
|
|
21867
22032
|
* This pipeline predicts the class of an image.
|
|
21868
|
-
*
|
|
22033
|
+
*
|
|
21869
22034
|
* **Example:** Classify an image.
|
|
21870
22035
|
* ```javascript
|
|
21871
22036
|
* const classifier = await pipeline('image-classification', 'Xenova/vit-base-patch16-224');
|
|
@@ -21875,7 +22040,7 @@ class ImageToTextPipeline extends (/** @type {new (options: TextImagePipelineCon
|
|
|
21875
22040
|
* // { label: 'tiger, Panthera tigris', score: 0.632695734500885 },
|
|
21876
22041
|
* // ]
|
|
21877
22042
|
* ```
|
|
21878
|
-
*
|
|
22043
|
+
*
|
|
21879
22044
|
* **Example:** Classify an image and return top `n` classes.
|
|
21880
22045
|
* ```javascript
|
|
21881
22046
|
* const classifier = await pipeline('image-classification', 'Xenova/vit-base-patch16-224');
|
|
@@ -21887,7 +22052,7 @@ class ImageToTextPipeline extends (/** @type {new (options: TextImagePipelineCon
|
|
|
21887
22052
|
* // { label: 'lion, king of beasts, Panthera leo', score: 0.00045060308184474707 },
|
|
21888
22053
|
* // ]
|
|
21889
22054
|
* ```
|
|
21890
|
-
*
|
|
22055
|
+
*
|
|
21891
22056
|
* **Example:** Classify an image and return all classes.
|
|
21892
22057
|
* ```javascript
|
|
21893
22058
|
* const classifier = await pipeline('image-classification', 'Xenova/vit-base-patch16-224');
|
|
@@ -21954,7 +22119,7 @@ class ImageClassificationPipeline extends (/** @type {new (options: ImagePipelin
|
|
|
21954
22119
|
* @property {string|null} label The label of the segment.
|
|
21955
22120
|
* @property {number|null} score The score of the segment.
|
|
21956
22121
|
* @property {RawImage} mask The mask of the segment.
|
|
21957
|
-
*
|
|
22122
|
+
*
|
|
21958
22123
|
* @typedef {Object} ImageSegmentationPipelineOptions Parameters specific to image segmentation pipelines.
|
|
21959
22124
|
* @property {number} [threshold=0.5] Probability threshold to filter out predicted masks.
|
|
21960
22125
|
* @property {number} [mask_threshold=0.5] Threshold to use when turning the predicted masks into binary values.
|
|
@@ -21963,19 +22128,19 @@ class ImageClassificationPipeline extends (/** @type {new (options: ImagePipelin
|
|
|
21963
22128
|
* depending on model capabilities. If not set, the pipeline will attempt to resolve (in that order).
|
|
21964
22129
|
* @property {number[]} [label_ids_to_fuse=null] List of label ids to fuse. If not set, do not fuse any labels.
|
|
21965
22130
|
* @property {number[][]} [target_sizes=null] List of target sizes for the input images. If not set, use the original image sizes.
|
|
21966
|
-
*
|
|
22131
|
+
*
|
|
21967
22132
|
* @callback ImageSegmentationPipelineCallback Segment the input images.
|
|
21968
22133
|
* @param {ImagePipelineInputs} images The input images.
|
|
21969
22134
|
* @param {ImageSegmentationPipelineOptions} [options] The options to use for image segmentation.
|
|
21970
22135
|
* @returns {Promise<ImageSegmentationPipelineOutput[]>} The annotated segments.
|
|
21971
|
-
*
|
|
22136
|
+
*
|
|
21972
22137
|
* @typedef {ImagePipelineConstructorArgs & ImageSegmentationPipelineCallback & Disposable} ImageSegmentationPipelineType
|
|
21973
22138
|
*/
|
|
21974
22139
|
|
|
21975
22140
|
/**
|
|
21976
22141
|
* Image segmentation pipeline using any `AutoModelForXXXSegmentation`.
|
|
21977
22142
|
* This pipeline predicts masks of objects and their classes.
|
|
21978
|
-
*
|
|
22143
|
+
*
|
|
21979
22144
|
* **Example:** Perform image segmentation with `Xenova/detr-resnet-50-panoptic`.
|
|
21980
22145
|
* ```javascript
|
|
21981
22146
|
* const segmenter = await pipeline('image-segmentation', 'Xenova/detr-resnet-50-panoptic');
|
|
@@ -22059,12 +22224,17 @@ class ImageSegmentationPipeline extends (/** @type {new (options: ImagePipelineC
|
|
|
22059
22224
|
/** @type {ImageSegmentationPipelineOutput[]} */
|
|
22060
22225
|
const annotation = [];
|
|
22061
22226
|
if (!subtask) {
|
|
22227
|
+
// We define an epsilon to safeguard against numerical/precision issues when detecting
|
|
22228
|
+
// the normalization mode of the output (i.e., sigmoid already applied, or not).
|
|
22229
|
+
// See https://github.com/microsoft/onnxruntime/issues/23943 for more information.
|
|
22230
|
+
const epsilon = 1e-5;
|
|
22231
|
+
|
|
22062
22232
|
// Perform standard image segmentation
|
|
22063
22233
|
const result = output[outputNames[0]];
|
|
22064
22234
|
for (let i = 0; i < imageSizes.length; ++i) {
|
|
22065
22235
|
const size = imageSizes[i];
|
|
22066
22236
|
const item = result[i];
|
|
22067
|
-
if (item.data.some(x => x <
|
|
22237
|
+
if (item.data.some(x => x < -epsilon || x > 1 + epsilon)) {
|
|
22068
22238
|
item.sigmoid_();
|
|
22069
22239
|
}
|
|
22070
22240
|
const mask = await _utils_image_js__WEBPACK_IMPORTED_MODULE_9__.RawImage.fromTensor(item.mul_(255).to('uint8')).resize(size[1], size[0]);
|
|
@@ -22133,19 +22303,19 @@ class ImageSegmentationPipeline extends (/** @type {new (options: ImagePipelineC
|
|
|
22133
22303
|
|
|
22134
22304
|
/**
|
|
22135
22305
|
* @typedef {Object} BackgroundRemovalPipelineOptions Parameters specific to image segmentation pipelines.
|
|
22136
|
-
*
|
|
22306
|
+
*
|
|
22137
22307
|
* @callback BackgroundRemovalPipelineCallback Segment the input images.
|
|
22138
22308
|
* @param {ImagePipelineInputs} images The input images.
|
|
22139
22309
|
* @param {BackgroundRemovalPipelineOptions} [options] The options to use for image segmentation.
|
|
22140
22310
|
* @returns {Promise<RawImage[]>} The images with the background removed.
|
|
22141
|
-
*
|
|
22311
|
+
*
|
|
22142
22312
|
* @typedef {ImagePipelineConstructorArgs & BackgroundRemovalPipelineCallback & Disposable} BackgroundRemovalPipelineType
|
|
22143
22313
|
*/
|
|
22144
22314
|
|
|
22145
22315
|
/**
|
|
22146
22316
|
* Background removal pipeline using certain `AutoModelForXXXSegmentation`.
|
|
22147
22317
|
* This pipeline removes the backgrounds of images.
|
|
22148
|
-
*
|
|
22318
|
+
*
|
|
22149
22319
|
* **Example:** Perform background removal with `Xenova/modnet`.
|
|
22150
22320
|
* ```javascript
|
|
22151
22321
|
* const segmenter = await pipeline('background-removal', 'Xenova/modnet');
|
|
@@ -22156,7 +22326,7 @@ class ImageSegmentationPipeline extends (/** @type {new (options: ImagePipelineC
|
|
|
22156
22326
|
* // ]
|
|
22157
22327
|
* ```
|
|
22158
22328
|
*/
|
|
22159
|
-
class BackgroundRemovalPipeline extends (/** @type {new (options: ImagePipelineConstructorArgs) =>
|
|
22329
|
+
class BackgroundRemovalPipeline extends (/** @type {new (options: ImagePipelineConstructorArgs) => BackgroundRemovalPipelineType} */ (/** @type {any} */(ImageSegmentationPipeline))) {
|
|
22160
22330
|
/**
|
|
22161
22331
|
* Create a new BackgroundRemovalPipeline.
|
|
22162
22332
|
* @param {ImagePipelineConstructorArgs} options An object used to instantiate the pipeline.
|
|
@@ -22191,25 +22361,25 @@ class BackgroundRemovalPipeline extends (/** @type {new (options: ImagePipelineC
|
|
|
22191
22361
|
* @typedef {Object} ZeroShotImageClassificationOutput
|
|
22192
22362
|
* @property {string} label The label identified by the model. It is one of the suggested `candidate_label`.
|
|
22193
22363
|
* @property {number} score The score attributed by the model for that label (between 0 and 1).
|
|
22194
|
-
*
|
|
22364
|
+
*
|
|
22195
22365
|
* @typedef {Object} ZeroShotImageClassificationPipelineOptions Parameters specific to zero-shot image classification pipelines.
|
|
22196
22366
|
* @property {string} [hypothesis_template="This is a photo of {}"] The sentence used in conjunction with `candidate_labels`
|
|
22197
22367
|
* to attempt the image classification by replacing the placeholder with the candidate_labels.
|
|
22198
22368
|
* Then likelihood is estimated by using `logits_per_image`.
|
|
22199
|
-
*
|
|
22369
|
+
*
|
|
22200
22370
|
* @callback ZeroShotImageClassificationPipelineCallback Assign labels to the image(s) passed as inputs.
|
|
22201
22371
|
* @param {ImagePipelineInputs} images The input images.
|
|
22202
22372
|
* @param {string[]} candidate_labels The candidate labels for this image.
|
|
22203
22373
|
* @param {ZeroShotImageClassificationPipelineOptions} [options] The options to use for zero-shot image classification.
|
|
22204
22374
|
* @returns {Promise<ZeroShotImageClassificationOutput[]|ZeroShotImageClassificationOutput[][]>} An array of objects containing the predicted labels and scores.
|
|
22205
|
-
*
|
|
22375
|
+
*
|
|
22206
22376
|
* @typedef {TextImagePipelineConstructorArgs & ZeroShotImageClassificationPipelineCallback & Disposable} ZeroShotImageClassificationPipelineType
|
|
22207
22377
|
*/
|
|
22208
22378
|
|
|
22209
22379
|
/**
|
|
22210
22380
|
* Zero shot image classification pipeline. This pipeline predicts the class of
|
|
22211
22381
|
* an image when you provide an image and a set of `candidate_labels`.
|
|
22212
|
-
*
|
|
22382
|
+
*
|
|
22213
22383
|
* **Example:** Zero shot image classification w/ `Xenova/clip-vit-base-patch32`.
|
|
22214
22384
|
* ```javascript
|
|
22215
22385
|
* const classifier = await pipeline('zero-shot-image-classification', 'Xenova/clip-vit-base-patch32');
|
|
@@ -22239,7 +22409,7 @@ class ZeroShotImageClassificationPipeline extends (/** @type {new (options: Text
|
|
|
22239
22409
|
const isBatched = Array.isArray(images);
|
|
22240
22410
|
const preparedImages = await prepareImages(images);
|
|
22241
22411
|
|
|
22242
|
-
// Insert label into hypothesis template
|
|
22412
|
+
// Insert label into hypothesis template
|
|
22243
22413
|
const texts = candidate_labels.map(
|
|
22244
22414
|
x => hypothesis_template.replace('{}', x)
|
|
22245
22415
|
);
|
|
@@ -22286,23 +22456,23 @@ class ZeroShotImageClassificationPipeline extends (/** @type {new (options: Text
|
|
|
22286
22456
|
* @property {number} score The score attributed by the model for that label.
|
|
22287
22457
|
* @property {BoundingBox} box The bounding box of detected object in image's original size, or as a percentage if `percentage` is set to true.
|
|
22288
22458
|
* @typedef {ObjectDetectionPipelineSingle[]} ObjectDetectionPipelineOutput
|
|
22289
|
-
*
|
|
22459
|
+
*
|
|
22290
22460
|
* @typedef {Object} ObjectDetectionPipelineOptions Parameters specific to object detection pipelines.
|
|
22291
22461
|
* @property {number} [threshold=0.9] The threshold used to filter boxes by score.
|
|
22292
22462
|
* @property {boolean} [percentage=false] Whether to return the boxes coordinates in percentage (true) or in pixels (false).
|
|
22293
|
-
*
|
|
22463
|
+
*
|
|
22294
22464
|
* @callback ObjectDetectionPipelineCallback Detect objects (bounding boxes & classes) in the image(s) passed as inputs.
|
|
22295
22465
|
* @param {ImagePipelineInputs} images The input images.
|
|
22296
22466
|
* @param {ObjectDetectionPipelineOptions} [options] The options to use for object detection.
|
|
22297
|
-
* @returns {Promise<ObjectDetectionPipelineOutput|ObjectDetectionPipelineOutput[]>} A list of objects or a list of list of objects.
|
|
22298
|
-
*
|
|
22467
|
+
* @returns {Promise<ObjectDetectionPipelineOutput|ObjectDetectionPipelineOutput[]>} A list of objects or a list of list of objects.
|
|
22468
|
+
*
|
|
22299
22469
|
* @typedef {ImagePipelineConstructorArgs & ObjectDetectionPipelineCallback & Disposable} ObjectDetectionPipelineType
|
|
22300
22470
|
*/
|
|
22301
22471
|
|
|
22302
22472
|
/**
|
|
22303
22473
|
* Object detection pipeline using any `AutoModelForObjectDetection`.
|
|
22304
22474
|
* This pipeline predicts bounding boxes of objects and their classes.
|
|
22305
|
-
*
|
|
22475
|
+
*
|
|
22306
22476
|
* **Example:** Run object-detection with `Xenova/detr-resnet-50`.
|
|
22307
22477
|
* ```javascript
|
|
22308
22478
|
* const detector = await pipeline('object-detection', 'Xenova/detr-resnet-50');
|
|
@@ -22376,27 +22546,27 @@ class ObjectDetectionPipeline extends (/** @type {new (options: ImagePipelineCon
|
|
|
22376
22546
|
* @property {string} label Text query corresponding to the found object.
|
|
22377
22547
|
* @property {number} score Score corresponding to the object (between 0 and 1).
|
|
22378
22548
|
* @property {BoundingBox} box Bounding box of the detected object in image's original size, or as a percentage if `percentage` is set to true.
|
|
22379
|
-
*
|
|
22549
|
+
*
|
|
22380
22550
|
* @typedef {Object} ZeroShotObjectDetectionPipelineOptions Parameters specific to zero-shot object detection pipelines.
|
|
22381
22551
|
* @property {number} [threshold=0.1] The probability necessary to make a prediction.
|
|
22382
22552
|
* @property {number} [top_k=null] The number of top predictions that will be returned by the pipeline.
|
|
22383
22553
|
* If the provided number is `null` or higher than the number of predictions available, it will default
|
|
22384
22554
|
* to the number of predictions.
|
|
22385
22555
|
* @property {boolean} [percentage=false] Whether to return the boxes coordinates in percentage (true) or in pixels (false).
|
|
22386
|
-
*
|
|
22556
|
+
*
|
|
22387
22557
|
* @callback ZeroShotObjectDetectionPipelineCallback Detect objects (bounding boxes & classes) in the image(s) passed as inputs.
|
|
22388
22558
|
* @param {ImagePipelineInputs} images The input images.
|
|
22389
22559
|
* @param {string[]} candidate_labels What the model should recognize in the image.
|
|
22390
22560
|
* @param {ZeroShotObjectDetectionPipelineOptions} [options] The options to use for zero-shot object detection.
|
|
22391
22561
|
* @returns {Promise<ZeroShotObjectDetectionOutput[]|ZeroShotObjectDetectionOutput[][]>} An array of objects containing the predicted labels, scores, and bounding boxes.
|
|
22392
|
-
*
|
|
22562
|
+
*
|
|
22393
22563
|
* @typedef {TextImagePipelineConstructorArgs & ZeroShotObjectDetectionPipelineCallback & Disposable} ZeroShotObjectDetectionPipelineType
|
|
22394
22564
|
*/
|
|
22395
22565
|
|
|
22396
22566
|
/**
|
|
22397
22567
|
* Zero-shot object detection pipeline. This pipeline predicts bounding boxes of
|
|
22398
22568
|
* objects when you provide an image and a set of `candidate_labels`.
|
|
22399
|
-
*
|
|
22569
|
+
*
|
|
22400
22570
|
* **Example:** Zero-shot object detection w/ `Xenova/owlvit-base-patch32`.
|
|
22401
22571
|
* ```javascript
|
|
22402
22572
|
* const detector = await pipeline('zero-shot-object-detection', 'Xenova/owlvit-base-patch32');
|
|
@@ -22426,7 +22596,7 @@ class ObjectDetectionPipeline extends (/** @type {new (options: ImagePipelineCon
|
|
|
22426
22596
|
* // }
|
|
22427
22597
|
* // ]
|
|
22428
22598
|
* ```
|
|
22429
|
-
*
|
|
22599
|
+
*
|
|
22430
22600
|
* **Example:** Zero-shot object detection w/ `Xenova/owlvit-base-patch32` (returning top 4 matches and setting a threshold).
|
|
22431
22601
|
* ```javascript
|
|
22432
22602
|
* const detector = await pipeline('zero-shot-object-detection', 'Xenova/owlvit-base-patch32');
|
|
@@ -22541,13 +22711,13 @@ class ZeroShotObjectDetectionPipeline extends (/** @type {new (options: TextImag
|
|
|
22541
22711
|
* @typedef {Object} DocumentQuestionAnsweringSingle
|
|
22542
22712
|
* @property {string} answer The generated text.
|
|
22543
22713
|
* @typedef {DocumentQuestionAnsweringSingle[]} DocumentQuestionAnsweringOutput
|
|
22544
|
-
*
|
|
22714
|
+
*
|
|
22545
22715
|
* @callback DocumentQuestionAnsweringPipelineCallback Answer the question given as input by using the document.
|
|
22546
22716
|
* @param {ImageInput} image The image of the document to use.
|
|
22547
22717
|
* @param {string} question A question to ask of the document.
|
|
22548
22718
|
* @param {Partial<import('./generation/configuration_utils.js').GenerationConfig>} [options] Additional keyword arguments to pass along to the generate method of the model.
|
|
22549
22719
|
* @returns {Promise<DocumentQuestionAnsweringOutput|DocumentQuestionAnsweringOutput[]>} An object (or array of objects) containing the answer(s).
|
|
22550
|
-
*
|
|
22720
|
+
*
|
|
22551
22721
|
* @typedef {TextImagePipelineConstructorArgs & DocumentQuestionAnsweringPipelineCallback & Disposable} DocumentQuestionAnsweringPipelineType
|
|
22552
22722
|
*/
|
|
22553
22723
|
|
|
@@ -22555,7 +22725,7 @@ class ZeroShotObjectDetectionPipeline extends (/** @type {new (options: TextImag
|
|
|
22555
22725
|
* Document Question Answering pipeline using any `AutoModelForDocumentQuestionAnswering`.
|
|
22556
22726
|
* The inputs/outputs are similar to the (extractive) question answering pipeline; however,
|
|
22557
22727
|
* the pipeline takes an image (and optional OCR'd words/boxes) as input instead of text context.
|
|
22558
|
-
*
|
|
22728
|
+
*
|
|
22559
22729
|
* **Example:** Answer questions about a document with `Xenova/donut-base-finetuned-docvqa`.
|
|
22560
22730
|
* ```javascript
|
|
22561
22731
|
* const qa_pipeline = await pipeline('document-question-answering', 'Xenova/donut-base-finetuned-docvqa');
|
|
@@ -22625,22 +22795,22 @@ class DocumentQuestionAnsweringPipeline extends (/** @type {new (options: TextIm
|
|
|
22625
22795
|
* @typedef {Object} TextToAudioOutput
|
|
22626
22796
|
* @property {Float32Array} audio The generated audio waveform.
|
|
22627
22797
|
* @property {number} sampling_rate The sampling rate of the generated audio waveform.
|
|
22628
|
-
*
|
|
22798
|
+
*
|
|
22629
22799
|
* @typedef {Object} TextToAudioPipelineOptions Parameters specific to text-to-audio pipelines.
|
|
22630
22800
|
* @property {Tensor|Float32Array|string|URL} [speaker_embeddings=null] The speaker embeddings (if the model requires it).
|
|
22631
|
-
*
|
|
22801
|
+
*
|
|
22632
22802
|
* @callback TextToAudioPipelineCallback Generates speech/audio from the inputs.
|
|
22633
22803
|
* @param {string|string[]} texts The text(s) to generate.
|
|
22634
22804
|
* @param {TextToAudioPipelineOptions} options Parameters passed to the model generation/forward method.
|
|
22635
22805
|
* @returns {Promise<TextToAudioOutput>} An object containing the generated audio and sampling rate.
|
|
22636
|
-
*
|
|
22806
|
+
*
|
|
22637
22807
|
* @typedef {TextToAudioPipelineConstructorArgs & TextToAudioPipelineCallback & Disposable} TextToAudioPipelineType
|
|
22638
22808
|
*/
|
|
22639
22809
|
|
|
22640
22810
|
/**
|
|
22641
22811
|
* Text-to-audio generation pipeline using any `AutoModelForTextToWaveform` or `AutoModelForTextToSpectrogram`.
|
|
22642
22812
|
* This pipeline generates an audio file from an input text and optional other conditional inputs.
|
|
22643
|
-
*
|
|
22813
|
+
*
|
|
22644
22814
|
* **Example:** Generate audio from text with `Xenova/speecht5_tts`.
|
|
22645
22815
|
* ```javascript
|
|
22646
22816
|
* const synthesizer = await pipeline('text-to-speech', 'Xenova/speecht5_tts', { quantized: false });
|
|
@@ -22651,17 +22821,17 @@ class DocumentQuestionAnsweringPipeline extends (/** @type {new (options: TextIm
|
|
|
22651
22821
|
* // sampling_rate: 16000
|
|
22652
22822
|
* // }
|
|
22653
22823
|
* ```
|
|
22654
|
-
*
|
|
22824
|
+
*
|
|
22655
22825
|
* You can then save the audio to a .wav file with the `wavefile` package:
|
|
22656
22826
|
* ```javascript
|
|
22657
22827
|
* import wavefile from 'wavefile';
|
|
22658
22828
|
* import fs from 'fs';
|
|
22659
|
-
*
|
|
22829
|
+
*
|
|
22660
22830
|
* const wav = new wavefile.WaveFile();
|
|
22661
22831
|
* wav.fromScratch(1, out.sampling_rate, '32f', out.audio);
|
|
22662
22832
|
* fs.writeFileSync('out.wav', wav.toBuffer());
|
|
22663
22833
|
* ```
|
|
22664
|
-
*
|
|
22834
|
+
*
|
|
22665
22835
|
* **Example:** Multilingual speech generation with `Xenova/mms-tts-fra`. See [here](https://huggingface.co/models?pipeline_tag=text-to-speech&other=vits&sort=trending) for the full list of available languages (1107).
|
|
22666
22836
|
* ```javascript
|
|
22667
22837
|
* const synthesizer = await pipeline('text-to-speech', 'Xenova/mms-tts-fra');
|
|
@@ -22767,13 +22937,13 @@ class TextToAudioPipeline extends (/** @type {new (options: TextToAudioPipelineC
|
|
|
22767
22937
|
* @callback ImageToImagePipelineCallback Transform the image(s) passed as inputs.
|
|
22768
22938
|
* @param {ImagePipelineInputs} images The images to transform.
|
|
22769
22939
|
* @returns {Promise<RawImage|RawImage[]>} The transformed image or list of images.
|
|
22770
|
-
*
|
|
22940
|
+
*
|
|
22771
22941
|
* @typedef {ImagePipelineConstructorArgs & ImageToImagePipelineCallback & Disposable} ImageToImagePipelineType
|
|
22772
22942
|
*/
|
|
22773
22943
|
|
|
22774
22944
|
/**
|
|
22775
22945
|
* Image to Image pipeline using any `AutoModelForImageToImage`. This pipeline generates an image based on a previous image input.
|
|
22776
|
-
*
|
|
22946
|
+
*
|
|
22777
22947
|
* **Example:** Super-resolution w/ `Xenova/swin2SR-classical-sr-x2-64`
|
|
22778
22948
|
* ```javascript
|
|
22779
22949
|
* const upscaler = await pipeline('image-to-image', 'Xenova/swin2SR-classical-sr-x2-64');
|
|
@@ -22818,17 +22988,17 @@ class ImageToImagePipeline extends (/** @type {new (options: ImagePipelineConstr
|
|
|
22818
22988
|
* @typedef {Object} DepthEstimationPipelineOutput
|
|
22819
22989
|
* @property {Tensor} predicted_depth The raw depth map predicted by the model.
|
|
22820
22990
|
* @property {RawImage} depth The processed depth map as an image (with the same size as the input image).
|
|
22821
|
-
*
|
|
22991
|
+
*
|
|
22822
22992
|
* @callback DepthEstimationPipelineCallback Predicts the depth for the image(s) passed as inputs.
|
|
22823
22993
|
* @param {ImagePipelineInputs} images The images to compute depth for.
|
|
22824
22994
|
* @returns {Promise<DepthEstimationPipelineOutput|DepthEstimationPipelineOutput[]>} An image or a list of images containing result(s).
|
|
22825
|
-
*
|
|
22995
|
+
*
|
|
22826
22996
|
* @typedef {ImagePipelineConstructorArgs & DepthEstimationPipelineCallback & Disposable} DepthEstimationPipelineType
|
|
22827
22997
|
*/
|
|
22828
22998
|
|
|
22829
22999
|
/**
|
|
22830
23000
|
* Depth estimation pipeline using any `AutoModelForDepthEstimation`. This pipeline predicts the depth of an image.
|
|
22831
|
-
*
|
|
23001
|
+
*
|
|
22832
23002
|
* **Example:** Depth estimation w/ `Xenova/dpt-hybrid-midas`
|
|
22833
23003
|
* ```javascript
|
|
22834
23004
|
* const depth_estimator = await pipeline('depth-estimation', 'Xenova/dpt-hybrid-midas');
|
|
@@ -23213,7 +23383,7 @@ const TASK_ALIASES = Object.freeze({
|
|
|
23213
23383
|
|
|
23214
23384
|
/**
|
|
23215
23385
|
* Utility factory method to build a `Pipeline` object.
|
|
23216
|
-
*
|
|
23386
|
+
*
|
|
23217
23387
|
* @template {PipelineType} T The type of pipeline to return.
|
|
23218
23388
|
* @param {T} task The task defining which pipeline will be returned. Currently accepted tasks are:
|
|
23219
23389
|
* - `"audio-classification"`: will return a `AudioClassificationPipeline`.
|
|
@@ -26031,13 +26201,12 @@ class PreTrainedTokenizer extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_0__
|
|
|
26031
26201
|
this.decoder.end_of_word_suffix = this.model.end_of_word_suffix;
|
|
26032
26202
|
}
|
|
26033
26203
|
|
|
26034
|
-
this.
|
|
26035
|
-
this.added_tokens.
|
|
26036
|
-
|
|
26037
|
-
|
|
26038
|
-
|
|
26039
|
-
|
|
26040
|
-
) : null;
|
|
26204
|
+
this.added_tokens_splitter = new _utils_data_structures_js__WEBPACK_IMPORTED_MODULE_5__.DictionarySplitter(
|
|
26205
|
+
this.added_tokens.map(x => x.content),
|
|
26206
|
+
);
|
|
26207
|
+
|
|
26208
|
+
/** @type {Map<string, AddedToken>} */
|
|
26209
|
+
this.added_tokens_map = new Map(this.added_tokens.map(x => [x.content, x]))
|
|
26041
26210
|
|
|
26042
26211
|
// Set mask token if present (otherwise will be undefined, which is fine)
|
|
26043
26212
|
this.mask_token = this.getToken('mask_token');
|
|
@@ -26332,40 +26501,50 @@ class PreTrainedTokenizer extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_0__
|
|
|
26332
26501
|
// Actual function which does encoding, for a single text
|
|
26333
26502
|
// First, we take care of special tokens. Needed to avoid issues arising from
|
|
26334
26503
|
// normalization and/or pretokenization (which may not preserve special tokens)
|
|
26335
|
-
const sections = this.
|
|
26336
|
-
|
|
26337
|
-
|
|
26338
|
-
|
|
26339
|
-
|
|
26340
|
-
|
|
26341
|
-
|
|
26342
|
-
|
|
26343
|
-
if (this.remove_space === true) {
|
|
26344
|
-
x = x.trim().split(/\s+/).join(' ');
|
|
26345
|
-
}
|
|
26346
|
-
if (this.do_lowercase_and_remove_accent) {
|
|
26347
|
-
x = lowercase_and_remove_accent(x);
|
|
26504
|
+
const sections = this.added_tokens_splitter.split(text);
|
|
26505
|
+
|
|
26506
|
+
// Process left/right stripping of added tokens
|
|
26507
|
+
for (let i = 0; i < sections.length; ++i) {
|
|
26508
|
+
const addedToken = this.added_tokens_map.get(sections[i]);
|
|
26509
|
+
if (addedToken) {
|
|
26510
|
+
if (addedToken.lstrip && i > 0) {
|
|
26511
|
+
sections[i - 1] = sections[i - 1].trimEnd();
|
|
26348
26512
|
}
|
|
26349
|
-
|
|
26350
|
-
|
|
26351
|
-
x = this.normalizer(x);
|
|
26513
|
+
if (addedToken.rstrip && i < sections.length - 1) {
|
|
26514
|
+
sections[i + 1] = sections[i + 1].trimStart();
|
|
26352
26515
|
}
|
|
26516
|
+
}
|
|
26517
|
+
}
|
|
26353
26518
|
|
|
26354
|
-
|
|
26355
|
-
|
|
26356
|
-
|
|
26357
|
-
return [];
|
|
26358
|
-
}
|
|
26519
|
+
const tokens = sections.flatMap((x, section_index) => {
|
|
26520
|
+
if (x.length === 0) return [];
|
|
26521
|
+
if (this.added_tokens_map.has(x)) return [x]; // Return added tokens unchanged
|
|
26359
26522
|
|
|
26360
|
-
|
|
26361
|
-
|
|
26362
|
-
|
|
26523
|
+
if (this.remove_space === true) {
|
|
26524
|
+
x = x.trim().split(/\s+/).join(' ');
|
|
26525
|
+
}
|
|
26526
|
+
if (this.do_lowercase_and_remove_accent) {
|
|
26527
|
+
x = lowercase_and_remove_accent(x);
|
|
26528
|
+
}
|
|
26363
26529
|
|
|
26364
|
-
|
|
26530
|
+
if (this.normalizer !== null) {
|
|
26531
|
+
x = this.normalizer(x);
|
|
26532
|
+
}
|
|
26365
26533
|
|
|
26366
|
-
|
|
26534
|
+
// If, after normalization, this section is empty (e.g., trimming whitespace),
|
|
26535
|
+
// we return an empty array
|
|
26536
|
+
if (x.length === 0) {
|
|
26537
|
+
return [];
|
|
26367
26538
|
}
|
|
26368
|
-
|
|
26539
|
+
|
|
26540
|
+
const sectionTokens = (this.pre_tokenizer !== null) ? this.pre_tokenizer(x, {
|
|
26541
|
+
section_index,
|
|
26542
|
+
}) : [x];
|
|
26543
|
+
|
|
26544
|
+
const tokens = this.model(sectionTokens);
|
|
26545
|
+
|
|
26546
|
+
return tokens;
|
|
26547
|
+
});
|
|
26369
26548
|
|
|
26370
26549
|
return tokens;
|
|
26371
26550
|
}
|
|
@@ -28990,6 +29169,7 @@ function count(arr, value) {
|
|
|
28990
29169
|
__webpack_require__.r(__webpack_exports__);
|
|
28991
29170
|
/* harmony export */ __webpack_require__.d(__webpack_exports__, {
|
|
28992
29171
|
/* harmony export */ CharTrie: () => (/* binding */ CharTrie),
|
|
29172
|
+
/* harmony export */ DictionarySplitter: () => (/* binding */ DictionarySplitter),
|
|
28993
29173
|
/* harmony export */ PriorityQueue: () => (/* binding */ PriorityQueue),
|
|
28994
29174
|
/* harmony export */ TokenLattice: () => (/* binding */ TokenLattice)
|
|
28995
29175
|
/* harmony export */ });
|
|
@@ -29441,6 +29621,80 @@ class TokenLatticeNode {
|
|
|
29441
29621
|
}
|
|
29442
29622
|
}
|
|
29443
29623
|
|
|
29624
|
+
/**
|
|
29625
|
+
* A data structure which uses a trie to split a string into tokens based on a dictionary.
|
|
29626
|
+
* It can also use a regular expression to preprocess the input text before splitting.
|
|
29627
|
+
*
|
|
29628
|
+
* NOTE: To ensure multi-byte characters are handled correctly, we operate at byte-level instead of character-level.
|
|
29629
|
+
*/
|
|
29630
|
+
class DictionarySplitter {
|
|
29631
|
+
/**
|
|
29632
|
+
* @param {string[]} dictionary The dictionary of words to use for splitting.
|
|
29633
|
+
*/
|
|
29634
|
+
constructor(dictionary) {
|
|
29635
|
+
this.trie = this._buildTrie(dictionary);
|
|
29636
|
+
}
|
|
29637
|
+
|
|
29638
|
+
/**
|
|
29639
|
+
* Builds a trie from the given dictionary.
|
|
29640
|
+
* @param {string[]} dictionary The dictionary of words to build the trie from.
|
|
29641
|
+
* @returns {Object} The root node of the trie.
|
|
29642
|
+
* @private
|
|
29643
|
+
*/
|
|
29644
|
+
_buildTrie(dictionary) {
|
|
29645
|
+
const trie = Object.create(null);
|
|
29646
|
+
for (const word of dictionary) {
|
|
29647
|
+
let node = trie;
|
|
29648
|
+
for (let i = 0; i < word.length; ++i) {
|
|
29649
|
+
node = (node[word[i]] ??= Object.create(null));
|
|
29650
|
+
}
|
|
29651
|
+
node.end = word;
|
|
29652
|
+
}
|
|
29653
|
+
return trie;
|
|
29654
|
+
}
|
|
29655
|
+
|
|
29656
|
+
/**
|
|
29657
|
+
* Splits the input text into tokens based on the dictionary.
|
|
29658
|
+
* @param {string} text The input text to split.
|
|
29659
|
+
* @returns {string[]} An array of tokens.
|
|
29660
|
+
*/
|
|
29661
|
+
split(text) {
|
|
29662
|
+
const result = [];
|
|
29663
|
+
const n = text.length;
|
|
29664
|
+
let start = 0;
|
|
29665
|
+
let i = 0;
|
|
29666
|
+
|
|
29667
|
+
while (i < n) {
|
|
29668
|
+
let node = this.trie;
|
|
29669
|
+
let match = null;
|
|
29670
|
+
let j = i;
|
|
29671
|
+
|
|
29672
|
+
while (j < n && (node = node[text[j]])) {
|
|
29673
|
+
if (node.end) {
|
|
29674
|
+
// Always keep the last (i.e., longest) match.
|
|
29675
|
+
match = node.end;
|
|
29676
|
+
}
|
|
29677
|
+
++j;
|
|
29678
|
+
}
|
|
29679
|
+
|
|
29680
|
+
if (match) {
|
|
29681
|
+
if (i > start) {
|
|
29682
|
+
result.push(text.slice(start, i));
|
|
29683
|
+
}
|
|
29684
|
+
result.push(match);
|
|
29685
|
+
i += match.length;
|
|
29686
|
+
start = i;
|
|
29687
|
+
} else {
|
|
29688
|
+
++i;
|
|
29689
|
+
}
|
|
29690
|
+
}
|
|
29691
|
+
if (start < n) {
|
|
29692
|
+
result.push(text.slice(start));
|
|
29693
|
+
}
|
|
29694
|
+
return result;
|
|
29695
|
+
}
|
|
29696
|
+
}
|
|
29697
|
+
|
|
29444
29698
|
|
|
29445
29699
|
/***/ }),
|
|
29446
29700
|
|
|
@@ -29820,6 +30074,22 @@ function isValidUrl(string, protocols = null, validHosts = null) {
|
|
|
29820
30074
|
return true;
|
|
29821
30075
|
}
|
|
29822
30076
|
|
|
30077
|
+
const REPO_ID_REGEX = /^(\b[\w\-.]+\b\/)?\b[\w\-.]{1,96}\b$/;
|
|
30078
|
+
|
|
30079
|
+
/**
|
|
30080
|
+
* Tests whether a string is a valid Hugging Face model ID or not.
|
|
30081
|
+
* Adapted from https://github.com/huggingface/huggingface_hub/blob/6378820ebb03f071988a96c7f3268f5bdf8f9449/src/huggingface_hub/utils/_validators.py#L119-L170
|
|
30082
|
+
*
|
|
30083
|
+
* @param {string} string The string to test
|
|
30084
|
+
* @returns {boolean} True if the string is a valid model ID, false otherwise.
|
|
30085
|
+
*/
|
|
30086
|
+
function isValidHfModelId(string) {
|
|
30087
|
+
if (!REPO_ID_REGEX.test(string)) return false;
|
|
30088
|
+
if (string.includes("..") || string.includes("--")) return false;
|
|
30089
|
+
if (string.endsWith(".git") || string.endsWith(".ipynb")) return false;
|
|
30090
|
+
return true;
|
|
30091
|
+
}
|
|
30092
|
+
|
|
29823
30093
|
/**
|
|
29824
30094
|
* Helper function to get a file, using either the Fetch API or FileSystem API.
|
|
29825
30095
|
*
|
|
@@ -30072,12 +30342,13 @@ async function getModelFile(path_or_repo_id, filename, fatal = true, options = {
|
|
|
30072
30342
|
}
|
|
30073
30343
|
|
|
30074
30344
|
const revision = options.revision ?? 'main';
|
|
30345
|
+
const requestURL = pathJoin(path_or_repo_id, filename);
|
|
30075
30346
|
|
|
30076
|
-
|
|
30077
|
-
|
|
30078
|
-
|
|
30079
|
-
|
|
30080
|
-
|
|
30347
|
+
const validModelId = isValidHfModelId(path_or_repo_id);
|
|
30348
|
+
const localPath = validModelId
|
|
30349
|
+
? pathJoin(_env_js__WEBPACK_IMPORTED_MODULE_2__.env.localModelPath, requestURL)
|
|
30350
|
+
: requestURL;
|
|
30351
|
+
const remoteURL = pathJoin(
|
|
30081
30352
|
_env_js__WEBPACK_IMPORTED_MODULE_2__.env.remoteHost,
|
|
30082
30353
|
_env_js__WEBPACK_IMPORTED_MODULE_2__.env.remotePathTemplate
|
|
30083
30354
|
.replaceAll('{model}', path_or_repo_id)
|
|
@@ -30085,14 +30356,14 @@ async function getModelFile(path_or_repo_id, filename, fatal = true, options = {
|
|
|
30085
30356
|
filename
|
|
30086
30357
|
);
|
|
30087
30358
|
|
|
30088
|
-
// Choose cache key for filesystem cache
|
|
30089
|
-
// When using the main revision (default), we use the request URL as the cache key.
|
|
30090
|
-
// If a specific revision is requested, we account for this in the cache key.
|
|
30091
|
-
let fsCacheKey = revision === 'main' ? requestURL : pathJoin(path_or_repo_id, revision, filename);
|
|
30092
|
-
|
|
30093
30359
|
/** @type {string} */
|
|
30094
30360
|
let cacheKey;
|
|
30095
|
-
|
|
30361
|
+
const proposedCacheKey = cache instanceof FileCache
|
|
30362
|
+
// Choose cache key for filesystem cache
|
|
30363
|
+
// When using the main revision (default), we use the request URL as the cache key.
|
|
30364
|
+
// If a specific revision is requested, we account for this in the cache key.
|
|
30365
|
+
? revision === 'main' ? requestURL : pathJoin(path_or_repo_id, revision, filename)
|
|
30366
|
+
: remoteURL;
|
|
30096
30367
|
|
|
30097
30368
|
// Whether to cache the final response in the end.
|
|
30098
30369
|
let toCacheResponse = false;
|
|
@@ -30105,11 +30376,10 @@ async function getModelFile(path_or_repo_id, filename, fatal = true, options = {
|
|
|
30105
30376
|
// 1. We first try to get from cache using the local path. In some environments (like deno),
|
|
30106
30377
|
// non-URL cache keys are not allowed. In these cases, `response` will be undefined.
|
|
30107
30378
|
// 2. If no response is found, we try to get from cache using the remote URL or file system cache.
|
|
30108
|
-
response = await tryCache(cache,
|
|
30379
|
+
response = await tryCache(cache, localPath, proposedCacheKey);
|
|
30109
30380
|
}
|
|
30110
30381
|
|
|
30111
30382
|
const cacheHit = response !== undefined;
|
|
30112
|
-
|
|
30113
30383
|
if (response === undefined) {
|
|
30114
30384
|
// Caching not available, or file is not cached, so we perform the request
|
|
30115
30385
|
|
|
@@ -30127,9 +30397,9 @@ async function getModelFile(path_or_repo_id, filename, fatal = true, options = {
|
|
|
30127
30397
|
console.warn(`Unable to load from local path "${localPath}": "${e}"`);
|
|
30128
30398
|
}
|
|
30129
30399
|
} else if (options.local_files_only) {
|
|
30130
|
-
throw new Error(`\`local_files_only=true\`, but attempted to load a remote file from: ${
|
|
30400
|
+
throw new Error(`\`local_files_only=true\`, but attempted to load a remote file from: ${requestURL}.`);
|
|
30131
30401
|
} else if (!_env_js__WEBPACK_IMPORTED_MODULE_2__.env.allowRemoteModels) {
|
|
30132
|
-
throw new Error(`\`env.allowRemoteModels=false\`, but attempted to load a remote file from: ${
|
|
30402
|
+
throw new Error(`\`env.allowRemoteModels=false\`, but attempted to load a remote file from: ${requestURL}.`);
|
|
30133
30403
|
}
|
|
30134
30404
|
}
|
|
30135
30405
|
|
|
@@ -30149,6 +30419,11 @@ async function getModelFile(path_or_repo_id, filename, fatal = true, options = {
|
|
|
30149
30419
|
return null;
|
|
30150
30420
|
}
|
|
30151
30421
|
}
|
|
30422
|
+
if (!validModelId) {
|
|
30423
|
+
// Before making any requests to the remote server, we check if the model ID is valid.
|
|
30424
|
+
// This prevents unnecessary network requests for invalid model IDs.
|
|
30425
|
+
throw Error(`Local file missing at "${localPath}" and download aborted due to invalid model ID "${path_or_repo_id}".`);
|
|
30426
|
+
}
|
|
30152
30427
|
|
|
30153
30428
|
// File not found locally, so we try to download it from the remote server
|
|
30154
30429
|
response = await getFile(remoteURL);
|
|
@@ -30479,7 +30754,7 @@ class RawImage {
|
|
|
30479
30754
|
|
|
30480
30755
|
/**
|
|
30481
30756
|
* Helper method for reading an image from a variety of input types.
|
|
30482
|
-
* @param {RawImage|string|URL} input
|
|
30757
|
+
* @param {RawImage|string|URL|Blob|HTMLCanvasElement|OffscreenCanvas} input
|
|
30483
30758
|
* @returns The image object.
|
|
30484
30759
|
*
|
|
30485
30760
|
* **Example:** Read image from a URL.
|
|
@@ -30498,6 +30773,14 @@ class RawImage {
|
|
|
30498
30773
|
return input;
|
|
30499
30774
|
} else if (typeof input === 'string' || input instanceof URL) {
|
|
30500
30775
|
return await this.fromURL(input);
|
|
30776
|
+
} else if (input instanceof Blob) {
|
|
30777
|
+
return await this.fromBlob(input);
|
|
30778
|
+
} else if (
|
|
30779
|
+
(typeof HTMLCanvasElement !== "undefined" && input instanceof HTMLCanvasElement)
|
|
30780
|
+
||
|
|
30781
|
+
(typeof OffscreenCanvas !== "undefined" && input instanceof OffscreenCanvas)
|
|
30782
|
+
) {
|
|
30783
|
+
return this.fromCanvas(input);
|
|
30501
30784
|
} else {
|
|
30502
30785
|
throw new Error(`Unsupported input type: ${typeof input}`);
|
|
30503
30786
|
}
|
|
@@ -33507,8 +33790,12 @@ function calc_unsqueeze_dims(dims, dim) {
|
|
|
33507
33790
|
* @private
|
|
33508
33791
|
*/
|
|
33509
33792
|
function safeIndex(index, size, dimension = null, boundsCheck = true) {
|
|
33510
|
-
if (
|
|
33511
|
-
|
|
33793
|
+
if (index < -size || index >= size) {
|
|
33794
|
+
if (boundsCheck) {
|
|
33795
|
+
throw new Error(`IndexError: index ${index} is out of bounds for dimension${dimension === null ? '' : ' ' + dimension} with size ${size}`);
|
|
33796
|
+
} else {
|
|
33797
|
+
return index < -size ? 0 : size;
|
|
33798
|
+
}
|
|
33512
33799
|
}
|
|
33513
33800
|
|
|
33514
33801
|
if (index < 0) {
|
|
@@ -34363,6 +34650,9 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
34363
34650
|
/* harmony export */ Gemma2ForCausalLM: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Gemma2ForCausalLM),
|
|
34364
34651
|
/* harmony export */ Gemma2Model: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Gemma2Model),
|
|
34365
34652
|
/* harmony export */ Gemma2PreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Gemma2PreTrainedModel),
|
|
34653
|
+
/* harmony export */ Gemma3ForCausalLM: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Gemma3ForCausalLM),
|
|
34654
|
+
/* harmony export */ Gemma3Model: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Gemma3Model),
|
|
34655
|
+
/* harmony export */ Gemma3PreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Gemma3PreTrainedModel),
|
|
34366
34656
|
/* harmony export */ GemmaForCausalLM: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.GemmaForCausalLM),
|
|
34367
34657
|
/* harmony export */ GemmaModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.GemmaModel),
|
|
34368
34658
|
/* harmony export */ GemmaPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.GemmaPreTrainedModel),
|
|
@@ -34464,6 +34754,10 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
34464
34754
|
/* harmony export */ MaskFormerPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.MaskFormerPreTrainedModel),
|
|
34465
34755
|
/* harmony export */ MaskedLMOutput: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.MaskedLMOutput),
|
|
34466
34756
|
/* harmony export */ MaxLengthCriteria: () => (/* reexport safe */ _generation_stopping_criteria_js__WEBPACK_IMPORTED_MODULE_20__.MaxLengthCriteria),
|
|
34757
|
+
/* harmony export */ Metric3DForDepthEstimation: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Metric3DForDepthEstimation),
|
|
34758
|
+
/* harmony export */ Metric3DPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Metric3DPreTrainedModel),
|
|
34759
|
+
/* harmony export */ Metric3Dv2ForDepthEstimation: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Metric3Dv2ForDepthEstimation),
|
|
34760
|
+
/* harmony export */ Metric3Dv2PreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Metric3Dv2PreTrainedModel),
|
|
34467
34761
|
/* harmony export */ MgpstrForSceneTextRecognition: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.MgpstrForSceneTextRecognition),
|
|
34468
34762
|
/* harmony export */ MgpstrModelOutput: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.MgpstrModelOutput),
|
|
34469
34763
|
/* harmony export */ MgpstrPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.MgpstrPreTrainedModel),
|
|
@@ -34616,11 +34910,19 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
34616
34910
|
/* harmony export */ Qwen2VLImageProcessor: () => (/* reexport safe */ _models_image_processors_js__WEBPACK_IMPORTED_MODULE_14__.Qwen2VLImageProcessor),
|
|
34617
34911
|
/* harmony export */ Qwen2VLPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Qwen2VLPreTrainedModel),
|
|
34618
34912
|
/* harmony export */ Qwen2VLProcessor: () => (/* reexport safe */ _models_processors_js__WEBPACK_IMPORTED_MODULE_17__.Qwen2VLProcessor),
|
|
34913
|
+
/* harmony export */ RFDetrForObjectDetection: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.RFDetrForObjectDetection),
|
|
34914
|
+
/* harmony export */ RFDetrModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.RFDetrModel),
|
|
34915
|
+
/* harmony export */ RFDetrObjectDetectionOutput: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.RFDetrObjectDetectionOutput),
|
|
34916
|
+
/* harmony export */ RFDetrPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.RFDetrPreTrainedModel),
|
|
34619
34917
|
/* harmony export */ RTDetrForObjectDetection: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.RTDetrForObjectDetection),
|
|
34620
34918
|
/* harmony export */ RTDetrImageProcessor: () => (/* reexport safe */ _models_image_processors_js__WEBPACK_IMPORTED_MODULE_14__.RTDetrImageProcessor),
|
|
34621
34919
|
/* harmony export */ RTDetrModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.RTDetrModel),
|
|
34622
34920
|
/* harmony export */ RTDetrObjectDetectionOutput: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.RTDetrObjectDetectionOutput),
|
|
34623
34921
|
/* harmony export */ RTDetrPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.RTDetrPreTrainedModel),
|
|
34922
|
+
/* harmony export */ RTDetrV2ForObjectDetection: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.RTDetrV2ForObjectDetection),
|
|
34923
|
+
/* harmony export */ RTDetrV2Model: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.RTDetrV2Model),
|
|
34924
|
+
/* harmony export */ RTDetrV2ObjectDetectionOutput: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.RTDetrV2ObjectDetectionOutput),
|
|
34925
|
+
/* harmony export */ RTDetrV2PreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.RTDetrV2PreTrainedModel),
|
|
34624
34926
|
/* harmony export */ RawAudio: () => (/* reexport safe */ _utils_audio_js__WEBPACK_IMPORTED_MODULE_5__.RawAudio),
|
|
34625
34927
|
/* harmony export */ RawImage: () => (/* reexport safe */ _utils_image_js__WEBPACK_IMPORTED_MODULE_6__.RawImage),
|
|
34626
34928
|
/* harmony export */ RawVideo: () => (/* reexport safe */ _utils_video_js__WEBPACK_IMPORTED_MODULE_7__.RawVideo),
|
|
@@ -34670,6 +34972,11 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
34670
34972
|
/* harmony export */ SmolVLMForConditionalGeneration: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.SmolVLMForConditionalGeneration),
|
|
34671
34973
|
/* harmony export */ SmolVLMImageProcessor: () => (/* reexport safe */ _models_image_processors_js__WEBPACK_IMPORTED_MODULE_14__.SmolVLMImageProcessor),
|
|
34672
34974
|
/* harmony export */ SmolVLMProcessor: () => (/* reexport safe */ _models_processors_js__WEBPACK_IMPORTED_MODULE_17__.SmolVLMProcessor),
|
|
34975
|
+
/* harmony export */ SnacDecoderModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.SnacDecoderModel),
|
|
34976
|
+
/* harmony export */ SnacEncoderModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.SnacEncoderModel),
|
|
34977
|
+
/* harmony export */ SnacFeatureExtractor: () => (/* reexport safe */ _models_feature_extractors_js__WEBPACK_IMPORTED_MODULE_11__.SnacFeatureExtractor),
|
|
34978
|
+
/* harmony export */ SnacModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.SnacModel),
|
|
34979
|
+
/* harmony export */ SnacPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.SnacPreTrainedModel),
|
|
34673
34980
|
/* harmony export */ SpeechT5FeatureExtractor: () => (/* reexport safe */ _models_feature_extractors_js__WEBPACK_IMPORTED_MODULE_11__.SpeechT5FeatureExtractor),
|
|
34674
34981
|
/* harmony export */ SpeechT5ForSpeechToText: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.SpeechT5ForSpeechToText),
|
|
34675
34982
|
/* harmony export */ SpeechT5ForTextToSpeech: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.SpeechT5ForTextToSpeech),
|
|
@@ -35192,6 +35499,9 @@ var __webpack_exports__GPTNeoXTokenizer = __webpack_exports__.GPTNeoXTokenizer;
|
|
|
35192
35499
|
var __webpack_exports__Gemma2ForCausalLM = __webpack_exports__.Gemma2ForCausalLM;
|
|
35193
35500
|
var __webpack_exports__Gemma2Model = __webpack_exports__.Gemma2Model;
|
|
35194
35501
|
var __webpack_exports__Gemma2PreTrainedModel = __webpack_exports__.Gemma2PreTrainedModel;
|
|
35502
|
+
var __webpack_exports__Gemma3ForCausalLM = __webpack_exports__.Gemma3ForCausalLM;
|
|
35503
|
+
var __webpack_exports__Gemma3Model = __webpack_exports__.Gemma3Model;
|
|
35504
|
+
var __webpack_exports__Gemma3PreTrainedModel = __webpack_exports__.Gemma3PreTrainedModel;
|
|
35195
35505
|
var __webpack_exports__GemmaForCausalLM = __webpack_exports__.GemmaForCausalLM;
|
|
35196
35506
|
var __webpack_exports__GemmaModel = __webpack_exports__.GemmaModel;
|
|
35197
35507
|
var __webpack_exports__GemmaPreTrainedModel = __webpack_exports__.GemmaPreTrainedModel;
|
|
@@ -35293,6 +35603,10 @@ var __webpack_exports__MaskFormerModel = __webpack_exports__.MaskFormerModel;
|
|
|
35293
35603
|
var __webpack_exports__MaskFormerPreTrainedModel = __webpack_exports__.MaskFormerPreTrainedModel;
|
|
35294
35604
|
var __webpack_exports__MaskedLMOutput = __webpack_exports__.MaskedLMOutput;
|
|
35295
35605
|
var __webpack_exports__MaxLengthCriteria = __webpack_exports__.MaxLengthCriteria;
|
|
35606
|
+
var __webpack_exports__Metric3DForDepthEstimation = __webpack_exports__.Metric3DForDepthEstimation;
|
|
35607
|
+
var __webpack_exports__Metric3DPreTrainedModel = __webpack_exports__.Metric3DPreTrainedModel;
|
|
35608
|
+
var __webpack_exports__Metric3Dv2ForDepthEstimation = __webpack_exports__.Metric3Dv2ForDepthEstimation;
|
|
35609
|
+
var __webpack_exports__Metric3Dv2PreTrainedModel = __webpack_exports__.Metric3Dv2PreTrainedModel;
|
|
35296
35610
|
var __webpack_exports__MgpstrForSceneTextRecognition = __webpack_exports__.MgpstrForSceneTextRecognition;
|
|
35297
35611
|
var __webpack_exports__MgpstrModelOutput = __webpack_exports__.MgpstrModelOutput;
|
|
35298
35612
|
var __webpack_exports__MgpstrPreTrainedModel = __webpack_exports__.MgpstrPreTrainedModel;
|
|
@@ -35445,11 +35759,19 @@ var __webpack_exports__Qwen2VLForConditionalGeneration = __webpack_exports__.Qwe
|
|
|
35445
35759
|
var __webpack_exports__Qwen2VLImageProcessor = __webpack_exports__.Qwen2VLImageProcessor;
|
|
35446
35760
|
var __webpack_exports__Qwen2VLPreTrainedModel = __webpack_exports__.Qwen2VLPreTrainedModel;
|
|
35447
35761
|
var __webpack_exports__Qwen2VLProcessor = __webpack_exports__.Qwen2VLProcessor;
|
|
35762
|
+
var __webpack_exports__RFDetrForObjectDetection = __webpack_exports__.RFDetrForObjectDetection;
|
|
35763
|
+
var __webpack_exports__RFDetrModel = __webpack_exports__.RFDetrModel;
|
|
35764
|
+
var __webpack_exports__RFDetrObjectDetectionOutput = __webpack_exports__.RFDetrObjectDetectionOutput;
|
|
35765
|
+
var __webpack_exports__RFDetrPreTrainedModel = __webpack_exports__.RFDetrPreTrainedModel;
|
|
35448
35766
|
var __webpack_exports__RTDetrForObjectDetection = __webpack_exports__.RTDetrForObjectDetection;
|
|
35449
35767
|
var __webpack_exports__RTDetrImageProcessor = __webpack_exports__.RTDetrImageProcessor;
|
|
35450
35768
|
var __webpack_exports__RTDetrModel = __webpack_exports__.RTDetrModel;
|
|
35451
35769
|
var __webpack_exports__RTDetrObjectDetectionOutput = __webpack_exports__.RTDetrObjectDetectionOutput;
|
|
35452
35770
|
var __webpack_exports__RTDetrPreTrainedModel = __webpack_exports__.RTDetrPreTrainedModel;
|
|
35771
|
+
var __webpack_exports__RTDetrV2ForObjectDetection = __webpack_exports__.RTDetrV2ForObjectDetection;
|
|
35772
|
+
var __webpack_exports__RTDetrV2Model = __webpack_exports__.RTDetrV2Model;
|
|
35773
|
+
var __webpack_exports__RTDetrV2ObjectDetectionOutput = __webpack_exports__.RTDetrV2ObjectDetectionOutput;
|
|
35774
|
+
var __webpack_exports__RTDetrV2PreTrainedModel = __webpack_exports__.RTDetrV2PreTrainedModel;
|
|
35453
35775
|
var __webpack_exports__RawAudio = __webpack_exports__.RawAudio;
|
|
35454
35776
|
var __webpack_exports__RawImage = __webpack_exports__.RawImage;
|
|
35455
35777
|
var __webpack_exports__RawVideo = __webpack_exports__.RawVideo;
|
|
@@ -35499,6 +35821,11 @@ var __webpack_exports__SiglipVisionModel = __webpack_exports__.SiglipVisionModel
|
|
|
35499
35821
|
var __webpack_exports__SmolVLMForConditionalGeneration = __webpack_exports__.SmolVLMForConditionalGeneration;
|
|
35500
35822
|
var __webpack_exports__SmolVLMImageProcessor = __webpack_exports__.SmolVLMImageProcessor;
|
|
35501
35823
|
var __webpack_exports__SmolVLMProcessor = __webpack_exports__.SmolVLMProcessor;
|
|
35824
|
+
var __webpack_exports__SnacDecoderModel = __webpack_exports__.SnacDecoderModel;
|
|
35825
|
+
var __webpack_exports__SnacEncoderModel = __webpack_exports__.SnacEncoderModel;
|
|
35826
|
+
var __webpack_exports__SnacFeatureExtractor = __webpack_exports__.SnacFeatureExtractor;
|
|
35827
|
+
var __webpack_exports__SnacModel = __webpack_exports__.SnacModel;
|
|
35828
|
+
var __webpack_exports__SnacPreTrainedModel = __webpack_exports__.SnacPreTrainedModel;
|
|
35502
35829
|
var __webpack_exports__SpeechT5FeatureExtractor = __webpack_exports__.SpeechT5FeatureExtractor;
|
|
35503
35830
|
var __webpack_exports__SpeechT5ForSpeechToText = __webpack_exports__.SpeechT5ForSpeechToText;
|
|
35504
35831
|
var __webpack_exports__SpeechT5ForTextToSpeech = __webpack_exports__.SpeechT5ForTextToSpeech;
|
|
@@ -35692,6 +36019,6 @@ var __webpack_exports__topk = __webpack_exports__.topk;
|
|
|
35692
36019
|
var __webpack_exports__window_function = __webpack_exports__.window_function;
|
|
35693
36020
|
var __webpack_exports__zeros = __webpack_exports__.zeros;
|
|
35694
36021
|
var __webpack_exports__zeros_like = __webpack_exports__.zeros_like;
|
|
35695
|
-
export { __webpack_exports__ASTFeatureExtractor as ASTFeatureExtractor, __webpack_exports__ASTForAudioClassification as ASTForAudioClassification, __webpack_exports__ASTModel as ASTModel, __webpack_exports__ASTPreTrainedModel as ASTPreTrainedModel, __webpack_exports__AlbertForMaskedLM as AlbertForMaskedLM, __webpack_exports__AlbertForQuestionAnswering as AlbertForQuestionAnswering, __webpack_exports__AlbertForSequenceClassification as AlbertForSequenceClassification, __webpack_exports__AlbertModel as AlbertModel, __webpack_exports__AlbertPreTrainedModel as AlbertPreTrainedModel, __webpack_exports__AlbertTokenizer as AlbertTokenizer, __webpack_exports__AudioClassificationPipeline as AudioClassificationPipeline, __webpack_exports__AutoConfig as AutoConfig, __webpack_exports__AutoFeatureExtractor as AutoFeatureExtractor, __webpack_exports__AutoImageProcessor as AutoImageProcessor, __webpack_exports__AutoModel as AutoModel, __webpack_exports__AutoModelForAudioClassification as AutoModelForAudioClassification, __webpack_exports__AutoModelForAudioFrameClassification as AutoModelForAudioFrameClassification, __webpack_exports__AutoModelForAudioTextToText as AutoModelForAudioTextToText, __webpack_exports__AutoModelForCTC as AutoModelForCTC, __webpack_exports__AutoModelForCausalLM as AutoModelForCausalLM, __webpack_exports__AutoModelForDepthEstimation as AutoModelForDepthEstimation, __webpack_exports__AutoModelForDocumentQuestionAnswering as AutoModelForDocumentQuestionAnswering, __webpack_exports__AutoModelForImageClassification as AutoModelForImageClassification, __webpack_exports__AutoModelForImageFeatureExtraction as AutoModelForImageFeatureExtraction, __webpack_exports__AutoModelForImageMatting as AutoModelForImageMatting, __webpack_exports__AutoModelForImageSegmentation as AutoModelForImageSegmentation, __webpack_exports__AutoModelForImageTextToText as AutoModelForImageTextToText, __webpack_exports__AutoModelForImageToImage as AutoModelForImageToImage, __webpack_exports__AutoModelForMaskGeneration as AutoModelForMaskGeneration, __webpack_exports__AutoModelForMaskedLM as AutoModelForMaskedLM, __webpack_exports__AutoModelForNormalEstimation as AutoModelForNormalEstimation, __webpack_exports__AutoModelForObjectDetection as AutoModelForObjectDetection, __webpack_exports__AutoModelForPoseEstimation as AutoModelForPoseEstimation, __webpack_exports__AutoModelForQuestionAnswering as AutoModelForQuestionAnswering, __webpack_exports__AutoModelForSemanticSegmentation as AutoModelForSemanticSegmentation, __webpack_exports__AutoModelForSeq2SeqLM as AutoModelForSeq2SeqLM, __webpack_exports__AutoModelForSequenceClassification as AutoModelForSequenceClassification, __webpack_exports__AutoModelForSpeechSeq2Seq as AutoModelForSpeechSeq2Seq, __webpack_exports__AutoModelForTextToSpectrogram as AutoModelForTextToSpectrogram, __webpack_exports__AutoModelForTextToWaveform as AutoModelForTextToWaveform, __webpack_exports__AutoModelForTokenClassification as AutoModelForTokenClassification, __webpack_exports__AutoModelForUniversalSegmentation as AutoModelForUniversalSegmentation, __webpack_exports__AutoModelForVision2Seq as AutoModelForVision2Seq, __webpack_exports__AutoModelForXVector as AutoModelForXVector, __webpack_exports__AutoModelForZeroShotObjectDetection as AutoModelForZeroShotObjectDetection, __webpack_exports__AutoProcessor as AutoProcessor, __webpack_exports__AutoTokenizer as AutoTokenizer, __webpack_exports__AutomaticSpeechRecognitionPipeline as AutomaticSpeechRecognitionPipeline, __webpack_exports__BackgroundRemovalPipeline as BackgroundRemovalPipeline, __webpack_exports__BartForConditionalGeneration as BartForConditionalGeneration, __webpack_exports__BartForSequenceClassification as BartForSequenceClassification, __webpack_exports__BartModel as BartModel, __webpack_exports__BartPretrainedModel as BartPretrainedModel, __webpack_exports__BartTokenizer as BartTokenizer, __webpack_exports__BaseModelOutput as BaseModelOutput, __webpack_exports__BaseStreamer as BaseStreamer, __webpack_exports__BeitFeatureExtractor as BeitFeatureExtractor, __webpack_exports__BeitForImageClassification as BeitForImageClassification, __webpack_exports__BeitModel as BeitModel, __webpack_exports__BeitPreTrainedModel as BeitPreTrainedModel, __webpack_exports__BertForMaskedLM as BertForMaskedLM, __webpack_exports__BertForQuestionAnswering as BertForQuestionAnswering, __webpack_exports__BertForSequenceClassification as BertForSequenceClassification, __webpack_exports__BertForTokenClassification as BertForTokenClassification, __webpack_exports__BertModel as BertModel, __webpack_exports__BertPreTrainedModel as BertPreTrainedModel, __webpack_exports__BertTokenizer as BertTokenizer, __webpack_exports__BitImageProcessor as BitImageProcessor, __webpack_exports__BlenderbotForConditionalGeneration as BlenderbotForConditionalGeneration, __webpack_exports__BlenderbotModel as BlenderbotModel, __webpack_exports__BlenderbotPreTrainedModel as BlenderbotPreTrainedModel, __webpack_exports__BlenderbotSmallForConditionalGeneration as BlenderbotSmallForConditionalGeneration, __webpack_exports__BlenderbotSmallModel as BlenderbotSmallModel, __webpack_exports__BlenderbotSmallPreTrainedModel as BlenderbotSmallPreTrainedModel, __webpack_exports__BlenderbotSmallTokenizer as BlenderbotSmallTokenizer, __webpack_exports__BlenderbotTokenizer as BlenderbotTokenizer, __webpack_exports__BloomForCausalLM as BloomForCausalLM, __webpack_exports__BloomModel as BloomModel, __webpack_exports__BloomPreTrainedModel as BloomPreTrainedModel, __webpack_exports__BloomTokenizer as BloomTokenizer, __webpack_exports__CLIPFeatureExtractor as CLIPFeatureExtractor, __webpack_exports__CLIPImageProcessor as CLIPImageProcessor, __webpack_exports__CLIPModel as CLIPModel, __webpack_exports__CLIPPreTrainedModel as CLIPPreTrainedModel, __webpack_exports__CLIPSegForImageSegmentation as CLIPSegForImageSegmentation, __webpack_exports__CLIPSegModel as CLIPSegModel, __webpack_exports__CLIPSegPreTrainedModel as CLIPSegPreTrainedModel, __webpack_exports__CLIPTextModel as CLIPTextModel, __webpack_exports__CLIPTextModelWithProjection as CLIPTextModelWithProjection, __webpack_exports__CLIPTokenizer as CLIPTokenizer, __webpack_exports__CLIPVisionModel as CLIPVisionModel, __webpack_exports__CLIPVisionModelWithProjection as CLIPVisionModelWithProjection, __webpack_exports__CamembertForMaskedLM as CamembertForMaskedLM, __webpack_exports__CamembertForQuestionAnswering as CamembertForQuestionAnswering, __webpack_exports__CamembertForSequenceClassification as CamembertForSequenceClassification, __webpack_exports__CamembertForTokenClassification as CamembertForTokenClassification, __webpack_exports__CamembertModel as CamembertModel, __webpack_exports__CamembertPreTrainedModel as CamembertPreTrainedModel, __webpack_exports__CamembertTokenizer as CamembertTokenizer, __webpack_exports__CausalLMOutput as CausalLMOutput, __webpack_exports__CausalLMOutputWithPast as CausalLMOutputWithPast, __webpack_exports__ChineseCLIPFeatureExtractor as ChineseCLIPFeatureExtractor, __webpack_exports__ChineseCLIPModel as ChineseCLIPModel, __webpack_exports__ChineseCLIPPreTrainedModel as ChineseCLIPPreTrainedModel, __webpack_exports__ClapAudioModelWithProjection as ClapAudioModelWithProjection, __webpack_exports__ClapFeatureExtractor as ClapFeatureExtractor, __webpack_exports__ClapModel as ClapModel, __webpack_exports__ClapPreTrainedModel as ClapPreTrainedModel, __webpack_exports__ClapTextModelWithProjection as ClapTextModelWithProjection, __webpack_exports__ClassifierFreeGuidanceLogitsProcessor as ClassifierFreeGuidanceLogitsProcessor, __webpack_exports__CodeGenForCausalLM as CodeGenForCausalLM, __webpack_exports__CodeGenModel as CodeGenModel, __webpack_exports__CodeGenPreTrainedModel as CodeGenPreTrainedModel, __webpack_exports__CodeGenTokenizer as CodeGenTokenizer, __webpack_exports__CodeLlamaTokenizer as CodeLlamaTokenizer, __webpack_exports__CohereForCausalLM as CohereForCausalLM, __webpack_exports__CohereModel as CohereModel, __webpack_exports__CoherePreTrainedModel as CoherePreTrainedModel, __webpack_exports__CohereTokenizer as CohereTokenizer, __webpack_exports__ConvBertForMaskedLM as ConvBertForMaskedLM, __webpack_exports__ConvBertForQuestionAnswering as ConvBertForQuestionAnswering, __webpack_exports__ConvBertForSequenceClassification as ConvBertForSequenceClassification, __webpack_exports__ConvBertForTokenClassification as ConvBertForTokenClassification, __webpack_exports__ConvBertModel as ConvBertModel, __webpack_exports__ConvBertPreTrainedModel as ConvBertPreTrainedModel, __webpack_exports__ConvBertTokenizer as ConvBertTokenizer, __webpack_exports__ConvNextFeatureExtractor as ConvNextFeatureExtractor, __webpack_exports__ConvNextForImageClassification as ConvNextForImageClassification, __webpack_exports__ConvNextImageProcessor as ConvNextImageProcessor, __webpack_exports__ConvNextModel as ConvNextModel, __webpack_exports__ConvNextPreTrainedModel as ConvNextPreTrainedModel, __webpack_exports__ConvNextV2ForImageClassification as ConvNextV2ForImageClassification, __webpack_exports__ConvNextV2Model as ConvNextV2Model, __webpack_exports__ConvNextV2PreTrainedModel as ConvNextV2PreTrainedModel, __webpack_exports__DPTFeatureExtractor as DPTFeatureExtractor, __webpack_exports__DPTForDepthEstimation as DPTForDepthEstimation, __webpack_exports__DPTImageProcessor as DPTImageProcessor, __webpack_exports__DPTModel as DPTModel, __webpack_exports__DPTPreTrainedModel as DPTPreTrainedModel, __webpack_exports__DacDecoderModel as DacDecoderModel, __webpack_exports__DacDecoderOutput as DacDecoderOutput, __webpack_exports__DacEncoderModel as DacEncoderModel, __webpack_exports__DacEncoderOutput as DacEncoderOutput, __webpack_exports__DacFeatureExtractor as DacFeatureExtractor, __webpack_exports__DacModel as DacModel, __webpack_exports__DacPreTrainedModel as DacPreTrainedModel, __webpack_exports__DataTypeMap as DataTypeMap, __webpack_exports__DebertaForMaskedLM as DebertaForMaskedLM, __webpack_exports__DebertaForQuestionAnswering as DebertaForQuestionAnswering, __webpack_exports__DebertaForSequenceClassification as DebertaForSequenceClassification, __webpack_exports__DebertaForTokenClassification as DebertaForTokenClassification, __webpack_exports__DebertaModel as DebertaModel, __webpack_exports__DebertaPreTrainedModel as DebertaPreTrainedModel, __webpack_exports__DebertaTokenizer as DebertaTokenizer, __webpack_exports__DebertaV2ForMaskedLM as DebertaV2ForMaskedLM, __webpack_exports__DebertaV2ForQuestionAnswering as DebertaV2ForQuestionAnswering, __webpack_exports__DebertaV2ForSequenceClassification as DebertaV2ForSequenceClassification, __webpack_exports__DebertaV2ForTokenClassification as DebertaV2ForTokenClassification, __webpack_exports__DebertaV2Model as DebertaV2Model, __webpack_exports__DebertaV2PreTrainedModel as DebertaV2PreTrainedModel, __webpack_exports__DebertaV2Tokenizer as DebertaV2Tokenizer, __webpack_exports__DecisionTransformerModel as DecisionTransformerModel, __webpack_exports__DecisionTransformerPreTrainedModel as DecisionTransformerPreTrainedModel, __webpack_exports__DeiTFeatureExtractor as DeiTFeatureExtractor, __webpack_exports__DeiTForImageClassification as DeiTForImageClassification, __webpack_exports__DeiTImageProcessor as DeiTImageProcessor, __webpack_exports__DeiTModel as DeiTModel, __webpack_exports__DeiTPreTrainedModel as DeiTPreTrainedModel, __webpack_exports__DepthAnythingForDepthEstimation as DepthAnythingForDepthEstimation, __webpack_exports__DepthAnythingPreTrainedModel as DepthAnythingPreTrainedModel, __webpack_exports__DepthEstimationPipeline as DepthEstimationPipeline, __webpack_exports__DepthProForDepthEstimation as DepthProForDepthEstimation, __webpack_exports__DepthProPreTrainedModel as DepthProPreTrainedModel, __webpack_exports__DetrFeatureExtractor as DetrFeatureExtractor, __webpack_exports__DetrForObjectDetection as DetrForObjectDetection, __webpack_exports__DetrForSegmentation as DetrForSegmentation, __webpack_exports__DetrImageProcessor as DetrImageProcessor, __webpack_exports__DetrModel as DetrModel, __webpack_exports__DetrObjectDetectionOutput as DetrObjectDetectionOutput, __webpack_exports__DetrPreTrainedModel as DetrPreTrainedModel, __webpack_exports__DetrSegmentationOutput as DetrSegmentationOutput, __webpack_exports__Dinov2ForImageClassification as Dinov2ForImageClassification, __webpack_exports__Dinov2Model as Dinov2Model, __webpack_exports__Dinov2PreTrainedModel as Dinov2PreTrainedModel, __webpack_exports__Dinov2WithRegistersForImageClassification as Dinov2WithRegistersForImageClassification, __webpack_exports__Dinov2WithRegistersModel as Dinov2WithRegistersModel, __webpack_exports__Dinov2WithRegistersPreTrainedModel as Dinov2WithRegistersPreTrainedModel, __webpack_exports__DistilBertForMaskedLM as DistilBertForMaskedLM, __webpack_exports__DistilBertForQuestionAnswering as DistilBertForQuestionAnswering, __webpack_exports__DistilBertForSequenceClassification as DistilBertForSequenceClassification, __webpack_exports__DistilBertForTokenClassification as DistilBertForTokenClassification, __webpack_exports__DistilBertModel as DistilBertModel, __webpack_exports__DistilBertPreTrainedModel as DistilBertPreTrainedModel, __webpack_exports__DistilBertTokenizer as DistilBertTokenizer, __webpack_exports__DocumentQuestionAnsweringPipeline as DocumentQuestionAnsweringPipeline, __webpack_exports__DonutFeatureExtractor as DonutFeatureExtractor, __webpack_exports__DonutImageProcessor as DonutImageProcessor, __webpack_exports__DonutSwinModel as DonutSwinModel, __webpack_exports__DonutSwinPreTrainedModel as DonutSwinPreTrainedModel, __webpack_exports__EfficientNetForImageClassification as EfficientNetForImageClassification, __webpack_exports__EfficientNetImageProcessor as EfficientNetImageProcessor, __webpack_exports__EfficientNetModel as EfficientNetModel, __webpack_exports__EfficientNetPreTrainedModel as EfficientNetPreTrainedModel, __webpack_exports__ElectraForMaskedLM as ElectraForMaskedLM, __webpack_exports__ElectraForQuestionAnswering as ElectraForQuestionAnswering, __webpack_exports__ElectraForSequenceClassification as ElectraForSequenceClassification, __webpack_exports__ElectraForTokenClassification as ElectraForTokenClassification, __webpack_exports__ElectraModel as ElectraModel, __webpack_exports__ElectraPreTrainedModel as ElectraPreTrainedModel, __webpack_exports__ElectraTokenizer as ElectraTokenizer, __webpack_exports__EncodecFeatureExtractor as EncodecFeatureExtractor, __webpack_exports__EosTokenCriteria as EosTokenCriteria, __webpack_exports__EsmForMaskedLM as EsmForMaskedLM, __webpack_exports__EsmForSequenceClassification as EsmForSequenceClassification, __webpack_exports__EsmForTokenClassification as EsmForTokenClassification, __webpack_exports__EsmModel as EsmModel, __webpack_exports__EsmPreTrainedModel as EsmPreTrainedModel, __webpack_exports__EsmTokenizer as EsmTokenizer, __webpack_exports__ExaoneForCausalLM as ExaoneForCausalLM, __webpack_exports__ExaoneModel as ExaoneModel, __webpack_exports__ExaonePreTrainedModel as ExaonePreTrainedModel, __webpack_exports__FFT as FFT, __webpack_exports__FalconForCausalLM as FalconForCausalLM, __webpack_exports__FalconModel as FalconModel, __webpack_exports__FalconPreTrainedModel as FalconPreTrainedModel, __webpack_exports__FalconTokenizer as FalconTokenizer, __webpack_exports__FastViTForImageClassification as FastViTForImageClassification, __webpack_exports__FastViTModel as FastViTModel, __webpack_exports__FastViTPreTrainedModel as FastViTPreTrainedModel, __webpack_exports__FeatureExtractionPipeline as FeatureExtractionPipeline, __webpack_exports__FeatureExtractor as FeatureExtractor, __webpack_exports__FillMaskPipeline as FillMaskPipeline, __webpack_exports__Florence2ForConditionalGeneration as Florence2ForConditionalGeneration, __webpack_exports__Florence2PreTrainedModel as Florence2PreTrainedModel, __webpack_exports__Florence2Processor as Florence2Processor, __webpack_exports__ForcedBOSTokenLogitsProcessor as ForcedBOSTokenLogitsProcessor, __webpack_exports__ForcedEOSTokenLogitsProcessor as ForcedEOSTokenLogitsProcessor, __webpack_exports__GLPNFeatureExtractor as GLPNFeatureExtractor, __webpack_exports__GLPNForDepthEstimation as GLPNForDepthEstimation, __webpack_exports__GLPNModel as GLPNModel, __webpack_exports__GLPNPreTrainedModel as GLPNPreTrainedModel, __webpack_exports__GPT2LMHeadModel as GPT2LMHeadModel, __webpack_exports__GPT2Model as GPT2Model, __webpack_exports__GPT2PreTrainedModel as GPT2PreTrainedModel, __webpack_exports__GPT2Tokenizer as GPT2Tokenizer, __webpack_exports__GPTBigCodeForCausalLM as GPTBigCodeForCausalLM, __webpack_exports__GPTBigCodeModel as GPTBigCodeModel, __webpack_exports__GPTBigCodePreTrainedModel as GPTBigCodePreTrainedModel, __webpack_exports__GPTJForCausalLM as GPTJForCausalLM, __webpack_exports__GPTJModel as GPTJModel, __webpack_exports__GPTJPreTrainedModel as GPTJPreTrainedModel, __webpack_exports__GPTNeoForCausalLM as GPTNeoForCausalLM, __webpack_exports__GPTNeoModel as GPTNeoModel, __webpack_exports__GPTNeoPreTrainedModel as GPTNeoPreTrainedModel, __webpack_exports__GPTNeoXForCausalLM as GPTNeoXForCausalLM, __webpack_exports__GPTNeoXModel as GPTNeoXModel, __webpack_exports__GPTNeoXPreTrainedModel as GPTNeoXPreTrainedModel, __webpack_exports__GPTNeoXTokenizer as GPTNeoXTokenizer, __webpack_exports__Gemma2ForCausalLM as Gemma2ForCausalLM, __webpack_exports__Gemma2Model as Gemma2Model, __webpack_exports__Gemma2PreTrainedModel as Gemma2PreTrainedModel, __webpack_exports__GemmaForCausalLM as GemmaForCausalLM, __webpack_exports__GemmaModel as GemmaModel, __webpack_exports__GemmaPreTrainedModel as GemmaPreTrainedModel, __webpack_exports__GemmaTokenizer as GemmaTokenizer, __webpack_exports__GlmForCausalLM as GlmForCausalLM, __webpack_exports__GlmModel as GlmModel, __webpack_exports__GlmPreTrainedModel as GlmPreTrainedModel, __webpack_exports__GraniteForCausalLM as GraniteForCausalLM, __webpack_exports__GraniteModel as GraniteModel, __webpack_exports__GranitePreTrainedModel as GranitePreTrainedModel, __webpack_exports__Grok1Tokenizer as Grok1Tokenizer, __webpack_exports__GroundingDinoForObjectDetection as GroundingDinoForObjectDetection, __webpack_exports__GroundingDinoImageProcessor as GroundingDinoImageProcessor, __webpack_exports__GroundingDinoPreTrainedModel as GroundingDinoPreTrainedModel, __webpack_exports__GroundingDinoProcessor as GroundingDinoProcessor, __webpack_exports__GroupViTModel as GroupViTModel, __webpack_exports__GroupViTPreTrainedModel as GroupViTPreTrainedModel, __webpack_exports__HeliumForCausalLM as HeliumForCausalLM, __webpack_exports__HeliumModel as HeliumModel, __webpack_exports__HeliumPreTrainedModel as HeliumPreTrainedModel, __webpack_exports__HerbertTokenizer as HerbertTokenizer, __webpack_exports__HieraForImageClassification as HieraForImageClassification, __webpack_exports__HieraModel as HieraModel, __webpack_exports__HieraPreTrainedModel as HieraPreTrainedModel, __webpack_exports__HubertForCTC as HubertForCTC, __webpack_exports__HubertForSequenceClassification as HubertForSequenceClassification, __webpack_exports__HubertModel as HubertModel, __webpack_exports__HubertPreTrainedModel as HubertPreTrainedModel, __webpack_exports__IJepaForImageClassification as IJepaForImageClassification, __webpack_exports__IJepaModel as IJepaModel, __webpack_exports__IJepaPreTrainedModel as IJepaPreTrainedModel, __webpack_exports__Idefics3ForConditionalGeneration as Idefics3ForConditionalGeneration, __webpack_exports__Idefics3ImageProcessor as Idefics3ImageProcessor, __webpack_exports__Idefics3PreTrainedModel as Idefics3PreTrainedModel, __webpack_exports__Idefics3Processor as Idefics3Processor, __webpack_exports__ImageClassificationPipeline as ImageClassificationPipeline, __webpack_exports__ImageFeatureExtractionPipeline as ImageFeatureExtractionPipeline, __webpack_exports__ImageFeatureExtractor as ImageFeatureExtractor, __webpack_exports__ImageMattingOutput as ImageMattingOutput, __webpack_exports__ImageProcessor as ImageProcessor, __webpack_exports__ImageSegmentationPipeline as ImageSegmentationPipeline, __webpack_exports__ImageToImagePipeline as ImageToImagePipeline, __webpack_exports__ImageToTextPipeline as ImageToTextPipeline, __webpack_exports__InterruptableStoppingCriteria as InterruptableStoppingCriteria, __webpack_exports__JAISLMHeadModel as JAISLMHeadModel, __webpack_exports__JAISModel as JAISModel, __webpack_exports__JAISPreTrainedModel as JAISPreTrainedModel, __webpack_exports__JinaCLIPImageProcessor as JinaCLIPImageProcessor, __webpack_exports__JinaCLIPModel as JinaCLIPModel, __webpack_exports__JinaCLIPPreTrainedModel as JinaCLIPPreTrainedModel, __webpack_exports__JinaCLIPProcessor as JinaCLIPProcessor, __webpack_exports__JinaCLIPTextModel as JinaCLIPTextModel, __webpack_exports__JinaCLIPVisionModel as JinaCLIPVisionModel, __webpack_exports__LiteWhisperForConditionalGeneration as LiteWhisperForConditionalGeneration, __webpack_exports__LlamaForCausalLM as LlamaForCausalLM, __webpack_exports__LlamaModel as LlamaModel, __webpack_exports__LlamaPreTrainedModel as LlamaPreTrainedModel, __webpack_exports__LlamaTokenizer as LlamaTokenizer, __webpack_exports__LlavaForConditionalGeneration as LlavaForConditionalGeneration, __webpack_exports__LlavaOnevisionForConditionalGeneration as LlavaOnevisionForConditionalGeneration, __webpack_exports__LlavaOnevisionImageProcessor as LlavaOnevisionImageProcessor, __webpack_exports__LlavaPreTrainedModel as LlavaPreTrainedModel, __webpack_exports__LogitsProcessor as LogitsProcessor, __webpack_exports__LogitsProcessorList as LogitsProcessorList, __webpack_exports__LogitsWarper as LogitsWarper, __webpack_exports__LongT5ForConditionalGeneration as LongT5ForConditionalGeneration, __webpack_exports__LongT5Model as LongT5Model, __webpack_exports__LongT5PreTrainedModel as LongT5PreTrainedModel, __webpack_exports__M2M100ForConditionalGeneration as M2M100ForConditionalGeneration, __webpack_exports__M2M100Model as M2M100Model, __webpack_exports__M2M100PreTrainedModel as M2M100PreTrainedModel, __webpack_exports__M2M100Tokenizer as M2M100Tokenizer, __webpack_exports__MBart50Tokenizer as MBart50Tokenizer, __webpack_exports__MBartForCausalLM as MBartForCausalLM, __webpack_exports__MBartForConditionalGeneration as MBartForConditionalGeneration, __webpack_exports__MBartForSequenceClassification as MBartForSequenceClassification, __webpack_exports__MBartModel as MBartModel, __webpack_exports__MBartPreTrainedModel as MBartPreTrainedModel, __webpack_exports__MBartTokenizer as MBartTokenizer, __webpack_exports__MPNetForMaskedLM as MPNetForMaskedLM, __webpack_exports__MPNetForQuestionAnswering as MPNetForQuestionAnswering, __webpack_exports__MPNetForSequenceClassification as MPNetForSequenceClassification, __webpack_exports__MPNetForTokenClassification as MPNetForTokenClassification, __webpack_exports__MPNetModel as MPNetModel, __webpack_exports__MPNetPreTrainedModel as MPNetPreTrainedModel, __webpack_exports__MPNetTokenizer as MPNetTokenizer, __webpack_exports__MT5ForConditionalGeneration as MT5ForConditionalGeneration, __webpack_exports__MT5Model as MT5Model, __webpack_exports__MT5PreTrainedModel as MT5PreTrainedModel, __webpack_exports__MarianMTModel as MarianMTModel, __webpack_exports__MarianModel as MarianModel, __webpack_exports__MarianPreTrainedModel as MarianPreTrainedModel, __webpack_exports__MarianTokenizer as MarianTokenizer, __webpack_exports__Mask2FormerImageProcessor as Mask2FormerImageProcessor, __webpack_exports__MaskFormerFeatureExtractor as MaskFormerFeatureExtractor, __webpack_exports__MaskFormerForInstanceSegmentation as MaskFormerForInstanceSegmentation, __webpack_exports__MaskFormerImageProcessor as MaskFormerImageProcessor, __webpack_exports__MaskFormerModel as MaskFormerModel, __webpack_exports__MaskFormerPreTrainedModel as MaskFormerPreTrainedModel, __webpack_exports__MaskedLMOutput as MaskedLMOutput, __webpack_exports__MaxLengthCriteria as MaxLengthCriteria, __webpack_exports__MgpstrForSceneTextRecognition as MgpstrForSceneTextRecognition, __webpack_exports__MgpstrModelOutput as MgpstrModelOutput, __webpack_exports__MgpstrPreTrainedModel as MgpstrPreTrainedModel, __webpack_exports__MgpstrProcessor as MgpstrProcessor, __webpack_exports__MgpstrTokenizer as MgpstrTokenizer, __webpack_exports__MimiDecoderModel as MimiDecoderModel, __webpack_exports__MimiDecoderOutput as MimiDecoderOutput, __webpack_exports__MimiEncoderModel as MimiEncoderModel, __webpack_exports__MimiEncoderOutput as MimiEncoderOutput, __webpack_exports__MimiModel as MimiModel, __webpack_exports__MimiPreTrainedModel as MimiPreTrainedModel, __webpack_exports__MinLengthLogitsProcessor as MinLengthLogitsProcessor, __webpack_exports__MinNewTokensLengthLogitsProcessor as MinNewTokensLengthLogitsProcessor, __webpack_exports__MistralForCausalLM as MistralForCausalLM, __webpack_exports__MistralModel as MistralModel, __webpack_exports__MistralPreTrainedModel as MistralPreTrainedModel, __webpack_exports__MobileBertForMaskedLM as MobileBertForMaskedLM, __webpack_exports__MobileBertForQuestionAnswering as MobileBertForQuestionAnswering, __webpack_exports__MobileBertForSequenceClassification as MobileBertForSequenceClassification, __webpack_exports__MobileBertModel as MobileBertModel, __webpack_exports__MobileBertPreTrainedModel as MobileBertPreTrainedModel, __webpack_exports__MobileBertTokenizer as MobileBertTokenizer, __webpack_exports__MobileLLMForCausalLM as MobileLLMForCausalLM, __webpack_exports__MobileLLMModel as MobileLLMModel, __webpack_exports__MobileLLMPreTrainedModel as MobileLLMPreTrainedModel, __webpack_exports__MobileNetV1FeatureExtractor as MobileNetV1FeatureExtractor, __webpack_exports__MobileNetV1ForImageClassification as MobileNetV1ForImageClassification, __webpack_exports__MobileNetV1ForSemanticSegmentation as MobileNetV1ForSemanticSegmentation, __webpack_exports__MobileNetV1ImageProcessor as MobileNetV1ImageProcessor, __webpack_exports__MobileNetV1Model as MobileNetV1Model, __webpack_exports__MobileNetV1PreTrainedModel as MobileNetV1PreTrainedModel, __webpack_exports__MobileNetV2FeatureExtractor as MobileNetV2FeatureExtractor, __webpack_exports__MobileNetV2ForImageClassification as MobileNetV2ForImageClassification, __webpack_exports__MobileNetV2ForSemanticSegmentation as MobileNetV2ForSemanticSegmentation, __webpack_exports__MobileNetV2ImageProcessor as MobileNetV2ImageProcessor, __webpack_exports__MobileNetV2Model as MobileNetV2Model, __webpack_exports__MobileNetV2PreTrainedModel as MobileNetV2PreTrainedModel, __webpack_exports__MobileNetV3FeatureExtractor as MobileNetV3FeatureExtractor, __webpack_exports__MobileNetV3ForImageClassification as MobileNetV3ForImageClassification, __webpack_exports__MobileNetV3ForSemanticSegmentation as MobileNetV3ForSemanticSegmentation, __webpack_exports__MobileNetV3ImageProcessor as MobileNetV3ImageProcessor, __webpack_exports__MobileNetV3Model as MobileNetV3Model, __webpack_exports__MobileNetV3PreTrainedModel as MobileNetV3PreTrainedModel, __webpack_exports__MobileNetV4FeatureExtractor as MobileNetV4FeatureExtractor, __webpack_exports__MobileNetV4ForImageClassification as MobileNetV4ForImageClassification, __webpack_exports__MobileNetV4ForSemanticSegmentation as MobileNetV4ForSemanticSegmentation, __webpack_exports__MobileNetV4ImageProcessor as MobileNetV4ImageProcessor, __webpack_exports__MobileNetV4Model as MobileNetV4Model, __webpack_exports__MobileNetV4PreTrainedModel as MobileNetV4PreTrainedModel, __webpack_exports__MobileViTFeatureExtractor as MobileViTFeatureExtractor, __webpack_exports__MobileViTForImageClassification as MobileViTForImageClassification, __webpack_exports__MobileViTImageProcessor as MobileViTImageProcessor, __webpack_exports__MobileViTModel as MobileViTModel, __webpack_exports__MobileViTPreTrainedModel as MobileViTPreTrainedModel, __webpack_exports__MobileViTV2ForImageClassification as MobileViTV2ForImageClassification, __webpack_exports__MobileViTV2Model as MobileViTV2Model, __webpack_exports__MobileViTV2PreTrainedModel as MobileViTV2PreTrainedModel, __webpack_exports__ModelOutput as ModelOutput, __webpack_exports__ModernBertForMaskedLM as ModernBertForMaskedLM, __webpack_exports__ModernBertForSequenceClassification as ModernBertForSequenceClassification, __webpack_exports__ModernBertForTokenClassification as ModernBertForTokenClassification, __webpack_exports__ModernBertModel as ModernBertModel, __webpack_exports__ModernBertPreTrainedModel as ModernBertPreTrainedModel, __webpack_exports__Moondream1ForConditionalGeneration as Moondream1ForConditionalGeneration, __webpack_exports__MoonshineFeatureExtractor as MoonshineFeatureExtractor, __webpack_exports__MoonshineForConditionalGeneration as MoonshineForConditionalGeneration, __webpack_exports__MoonshineModel as MoonshineModel, __webpack_exports__MoonshinePreTrainedModel as MoonshinePreTrainedModel, __webpack_exports__MoonshineProcessor as MoonshineProcessor, __webpack_exports__MptForCausalLM as MptForCausalLM, __webpack_exports__MptModel as MptModel, __webpack_exports__MptPreTrainedModel as MptPreTrainedModel, __webpack_exports__MultiModalityCausalLM as MultiModalityCausalLM, __webpack_exports__MultiModalityPreTrainedModel as MultiModalityPreTrainedModel, __webpack_exports__MusicgenForCausalLM as MusicgenForCausalLM, __webpack_exports__MusicgenForConditionalGeneration as MusicgenForConditionalGeneration, __webpack_exports__MusicgenModel as MusicgenModel, __webpack_exports__MusicgenPreTrainedModel as MusicgenPreTrainedModel, __webpack_exports__NllbTokenizer as NllbTokenizer, __webpack_exports__NoBadWordsLogitsProcessor as NoBadWordsLogitsProcessor, __webpack_exports__NoRepeatNGramLogitsProcessor as NoRepeatNGramLogitsProcessor, __webpack_exports__NomicBertModel as NomicBertModel, __webpack_exports__NomicBertPreTrainedModel as NomicBertPreTrainedModel, __webpack_exports__NougatImageProcessor as NougatImageProcessor, __webpack_exports__NougatTokenizer as NougatTokenizer, __webpack_exports__OPTForCausalLM as OPTForCausalLM, __webpack_exports__OPTModel as OPTModel, __webpack_exports__OPTPreTrainedModel as OPTPreTrainedModel, __webpack_exports__ObjectDetectionPipeline as ObjectDetectionPipeline, __webpack_exports__Olmo2ForCausalLM as Olmo2ForCausalLM, __webpack_exports__Olmo2Model as Olmo2Model, __webpack_exports__Olmo2PreTrainedModel as Olmo2PreTrainedModel, __webpack_exports__OlmoForCausalLM as OlmoForCausalLM, __webpack_exports__OlmoModel as OlmoModel, __webpack_exports__OlmoPreTrainedModel as OlmoPreTrainedModel, __webpack_exports__OpenELMForCausalLM as OpenELMForCausalLM, __webpack_exports__OpenELMModel as OpenELMModel, __webpack_exports__OpenELMPreTrainedModel as OpenELMPreTrainedModel, __webpack_exports__OwlViTFeatureExtractor as OwlViTFeatureExtractor, __webpack_exports__OwlViTForObjectDetection as OwlViTForObjectDetection, __webpack_exports__OwlViTImageProcessor as OwlViTImageProcessor, __webpack_exports__OwlViTModel as OwlViTModel, __webpack_exports__OwlViTPreTrainedModel as OwlViTPreTrainedModel, __webpack_exports__OwlViTProcessor as OwlViTProcessor, __webpack_exports__Owlv2ForObjectDetection as Owlv2ForObjectDetection, __webpack_exports__Owlv2ImageProcessor as Owlv2ImageProcessor, __webpack_exports__Owlv2Model as Owlv2Model, __webpack_exports__Owlv2PreTrainedModel as Owlv2PreTrainedModel, __webpack_exports__PaliGemmaForConditionalGeneration as PaliGemmaForConditionalGeneration, __webpack_exports__PaliGemmaPreTrainedModel as PaliGemmaPreTrainedModel, __webpack_exports__PaliGemmaProcessor as PaliGemmaProcessor, __webpack_exports__PatchTSMixerForPrediction as PatchTSMixerForPrediction, __webpack_exports__PatchTSMixerModel as PatchTSMixerModel, __webpack_exports__PatchTSMixerPreTrainedModel as PatchTSMixerPreTrainedModel, __webpack_exports__PatchTSTForPrediction as PatchTSTForPrediction, __webpack_exports__PatchTSTModel as PatchTSTModel, __webpack_exports__PatchTSTPreTrainedModel as PatchTSTPreTrainedModel, __webpack_exports__Phi3ForCausalLM as Phi3ForCausalLM, __webpack_exports__Phi3Model as Phi3Model, __webpack_exports__Phi3PreTrainedModel as Phi3PreTrainedModel, __webpack_exports__Phi3VForCausalLM as Phi3VForCausalLM, __webpack_exports__Phi3VImageProcessor as Phi3VImageProcessor, __webpack_exports__Phi3VPreTrainedModel as Phi3VPreTrainedModel, __webpack_exports__Phi3VProcessor as Phi3VProcessor, __webpack_exports__PhiForCausalLM as PhiForCausalLM, __webpack_exports__PhiModel as PhiModel, __webpack_exports__PhiPreTrainedModel as PhiPreTrainedModel, __webpack_exports__Pipeline as Pipeline, __webpack_exports__PreTrainedModel as PreTrainedModel, __webpack_exports__PreTrainedTokenizer as PreTrainedTokenizer, __webpack_exports__PretrainedConfig as PretrainedConfig, __webpack_exports__PretrainedMixin as PretrainedMixin, __webpack_exports__Processor as Processor, __webpack_exports__PvtForImageClassification as PvtForImageClassification, __webpack_exports__PvtImageProcessor as PvtImageProcessor, __webpack_exports__PvtModel as PvtModel, __webpack_exports__PvtPreTrainedModel as PvtPreTrainedModel, __webpack_exports__PyAnnoteFeatureExtractor as PyAnnoteFeatureExtractor, __webpack_exports__PyAnnoteForAudioFrameClassification as PyAnnoteForAudioFrameClassification, __webpack_exports__PyAnnoteModel as PyAnnoteModel, __webpack_exports__PyAnnotePreTrainedModel as PyAnnotePreTrainedModel, __webpack_exports__PyAnnoteProcessor as PyAnnoteProcessor, __webpack_exports__QuestionAnsweringModelOutput as QuestionAnsweringModelOutput, __webpack_exports__QuestionAnsweringPipeline as QuestionAnsweringPipeline, __webpack_exports__Qwen2ForCausalLM as Qwen2ForCausalLM, __webpack_exports__Qwen2Model as Qwen2Model, __webpack_exports__Qwen2PreTrainedModel as Qwen2PreTrainedModel, __webpack_exports__Qwen2Tokenizer as Qwen2Tokenizer, __webpack_exports__Qwen2VLForConditionalGeneration as Qwen2VLForConditionalGeneration, __webpack_exports__Qwen2VLImageProcessor as Qwen2VLImageProcessor, __webpack_exports__Qwen2VLPreTrainedModel as Qwen2VLPreTrainedModel, __webpack_exports__Qwen2VLProcessor as Qwen2VLProcessor, __webpack_exports__RTDetrForObjectDetection as RTDetrForObjectDetection, __webpack_exports__RTDetrImageProcessor as RTDetrImageProcessor, __webpack_exports__RTDetrModel as RTDetrModel, __webpack_exports__RTDetrObjectDetectionOutput as RTDetrObjectDetectionOutput, __webpack_exports__RTDetrPreTrainedModel as RTDetrPreTrainedModel, __webpack_exports__RawAudio as RawAudio, __webpack_exports__RawImage as RawImage, __webpack_exports__RawVideo as RawVideo, __webpack_exports__RawVideoFrame as RawVideoFrame, __webpack_exports__RepetitionPenaltyLogitsProcessor as RepetitionPenaltyLogitsProcessor, __webpack_exports__ResNetForImageClassification as ResNetForImageClassification, __webpack_exports__ResNetModel as ResNetModel, __webpack_exports__ResNetPreTrainedModel as ResNetPreTrainedModel, __webpack_exports__RoFormerForMaskedLM as RoFormerForMaskedLM, __webpack_exports__RoFormerForQuestionAnswering as RoFormerForQuestionAnswering, __webpack_exports__RoFormerForSequenceClassification as RoFormerForSequenceClassification, __webpack_exports__RoFormerForTokenClassification as RoFormerForTokenClassification, __webpack_exports__RoFormerModel as RoFormerModel, __webpack_exports__RoFormerPreTrainedModel as RoFormerPreTrainedModel, __webpack_exports__RoFormerTokenizer as RoFormerTokenizer, __webpack_exports__RobertaForMaskedLM as RobertaForMaskedLM, __webpack_exports__RobertaForQuestionAnswering as RobertaForQuestionAnswering, __webpack_exports__RobertaForSequenceClassification as RobertaForSequenceClassification, __webpack_exports__RobertaForTokenClassification as RobertaForTokenClassification, __webpack_exports__RobertaModel as RobertaModel, __webpack_exports__RobertaPreTrainedModel as RobertaPreTrainedModel, __webpack_exports__RobertaTokenizer as RobertaTokenizer, __webpack_exports__SamImageProcessor as SamImageProcessor, __webpack_exports__SamImageSegmentationOutput as SamImageSegmentationOutput, __webpack_exports__SamModel as SamModel, __webpack_exports__SamPreTrainedModel as SamPreTrainedModel, __webpack_exports__SamProcessor as SamProcessor, __webpack_exports__SapiensForDepthEstimation as SapiensForDepthEstimation, __webpack_exports__SapiensForNormalEstimation as SapiensForNormalEstimation, __webpack_exports__SapiensForSemanticSegmentation as SapiensForSemanticSegmentation, __webpack_exports__SapiensPreTrainedModel as SapiensPreTrainedModel, __webpack_exports__SeamlessM4TFeatureExtractor as SeamlessM4TFeatureExtractor, __webpack_exports__SegformerFeatureExtractor as SegformerFeatureExtractor, __webpack_exports__SegformerForImageClassification as SegformerForImageClassification, __webpack_exports__SegformerForSemanticSegmentation as SegformerForSemanticSegmentation, __webpack_exports__SegformerImageProcessor as SegformerImageProcessor, __webpack_exports__SegformerModel as SegformerModel, __webpack_exports__SegformerPreTrainedModel as SegformerPreTrainedModel, __webpack_exports__Seq2SeqLMOutput as Seq2SeqLMOutput, __webpack_exports__SequenceClassifierOutput as SequenceClassifierOutput, __webpack_exports__SiglipImageProcessor as SiglipImageProcessor, __webpack_exports__SiglipModel as SiglipModel, __webpack_exports__SiglipPreTrainedModel as SiglipPreTrainedModel, __webpack_exports__SiglipTextModel as SiglipTextModel, __webpack_exports__SiglipTokenizer as SiglipTokenizer, __webpack_exports__SiglipVisionModel as SiglipVisionModel, __webpack_exports__SmolVLMForConditionalGeneration as SmolVLMForConditionalGeneration, __webpack_exports__SmolVLMImageProcessor as SmolVLMImageProcessor, __webpack_exports__SmolVLMProcessor as SmolVLMProcessor, __webpack_exports__SpeechT5FeatureExtractor as SpeechT5FeatureExtractor, __webpack_exports__SpeechT5ForSpeechToText as SpeechT5ForSpeechToText, __webpack_exports__SpeechT5ForTextToSpeech as SpeechT5ForTextToSpeech, __webpack_exports__SpeechT5HifiGan as SpeechT5HifiGan, __webpack_exports__SpeechT5Model as SpeechT5Model, __webpack_exports__SpeechT5PreTrainedModel as SpeechT5PreTrainedModel, __webpack_exports__SpeechT5Processor as SpeechT5Processor, __webpack_exports__SpeechT5Tokenizer as SpeechT5Tokenizer, __webpack_exports__SqueezeBertForMaskedLM as SqueezeBertForMaskedLM, __webpack_exports__SqueezeBertForQuestionAnswering as SqueezeBertForQuestionAnswering, __webpack_exports__SqueezeBertForSequenceClassification as SqueezeBertForSequenceClassification, __webpack_exports__SqueezeBertModel as SqueezeBertModel, __webpack_exports__SqueezeBertPreTrainedModel as SqueezeBertPreTrainedModel, __webpack_exports__SqueezeBertTokenizer as SqueezeBertTokenizer, __webpack_exports__StableLmForCausalLM as StableLmForCausalLM, __webpack_exports__StableLmModel as StableLmModel, __webpack_exports__StableLmPreTrainedModel as StableLmPreTrainedModel, __webpack_exports__Starcoder2ForCausalLM as Starcoder2ForCausalLM, __webpack_exports__Starcoder2Model as Starcoder2Model, __webpack_exports__Starcoder2PreTrainedModel as Starcoder2PreTrainedModel, __webpack_exports__StoppingCriteria as StoppingCriteria, __webpack_exports__StoppingCriteriaList as StoppingCriteriaList, __webpack_exports__StyleTextToSpeech2Model as StyleTextToSpeech2Model, __webpack_exports__StyleTextToSpeech2PreTrainedModel as StyleTextToSpeech2PreTrainedModel, __webpack_exports__SummarizationPipeline as SummarizationPipeline, __webpack_exports__SuppressTokensAtBeginLogitsProcessor as SuppressTokensAtBeginLogitsProcessor, __webpack_exports__Swin2SRForImageSuperResolution as Swin2SRForImageSuperResolution, __webpack_exports__Swin2SRImageProcessor as Swin2SRImageProcessor, __webpack_exports__Swin2SRModel as Swin2SRModel, __webpack_exports__Swin2SRPreTrainedModel as Swin2SRPreTrainedModel, __webpack_exports__SwinForImageClassification as SwinForImageClassification, __webpack_exports__SwinForSemanticSegmentation as SwinForSemanticSegmentation, __webpack_exports__SwinModel as SwinModel, __webpack_exports__SwinPreTrainedModel as SwinPreTrainedModel, __webpack_exports__T5ForConditionalGeneration as T5ForConditionalGeneration, __webpack_exports__T5Model as T5Model, __webpack_exports__T5PreTrainedModel as T5PreTrainedModel, __webpack_exports__T5Tokenizer as T5Tokenizer, __webpack_exports__TableTransformerForObjectDetection as TableTransformerForObjectDetection, __webpack_exports__TableTransformerModel as TableTransformerModel, __webpack_exports__TableTransformerObjectDetectionOutput as TableTransformerObjectDetectionOutput, __webpack_exports__TableTransformerPreTrainedModel as TableTransformerPreTrainedModel, __webpack_exports__TemperatureLogitsWarper as TemperatureLogitsWarper, __webpack_exports__Tensor as Tensor, __webpack_exports__Text2TextGenerationPipeline as Text2TextGenerationPipeline, __webpack_exports__TextClassificationPipeline as TextClassificationPipeline, __webpack_exports__TextGenerationPipeline as TextGenerationPipeline, __webpack_exports__TextStreamer as TextStreamer, __webpack_exports__TextToAudioPipeline as TextToAudioPipeline, __webpack_exports__TokenClassificationPipeline as TokenClassificationPipeline, __webpack_exports__TokenClassifierOutput as TokenClassifierOutput, __webpack_exports__TokenizerModel as TokenizerModel, __webpack_exports__TopKLogitsWarper as TopKLogitsWarper, __webpack_exports__TopPLogitsWarper as TopPLogitsWarper, __webpack_exports__TrOCRForCausalLM as TrOCRForCausalLM, __webpack_exports__TrOCRPreTrainedModel as TrOCRPreTrainedModel, __webpack_exports__TranslationPipeline as TranslationPipeline, __webpack_exports__UltravoxModel as UltravoxModel, __webpack_exports__UltravoxPreTrainedModel as UltravoxPreTrainedModel, __webpack_exports__UltravoxProcessor as UltravoxProcessor, __webpack_exports__UniSpeechForCTC as UniSpeechForCTC, __webpack_exports__UniSpeechForSequenceClassification as UniSpeechForSequenceClassification, __webpack_exports__UniSpeechModel as UniSpeechModel, __webpack_exports__UniSpeechPreTrainedModel as UniSpeechPreTrainedModel, __webpack_exports__UniSpeechSatForAudioFrameClassification as UniSpeechSatForAudioFrameClassification, __webpack_exports__UniSpeechSatForCTC as UniSpeechSatForCTC, __webpack_exports__UniSpeechSatForSequenceClassification as UniSpeechSatForSequenceClassification, __webpack_exports__UniSpeechSatModel as UniSpeechSatModel, __webpack_exports__UniSpeechSatPreTrainedModel as UniSpeechSatPreTrainedModel, __webpack_exports__VLChatProcessor as VLChatProcessor, __webpack_exports__VLMImageProcessor as VLMImageProcessor, __webpack_exports__ViTFeatureExtractor as ViTFeatureExtractor, __webpack_exports__ViTForImageClassification as ViTForImageClassification, __webpack_exports__ViTImageProcessor as ViTImageProcessor, __webpack_exports__ViTMAEModel as ViTMAEModel, __webpack_exports__ViTMAEPreTrainedModel as ViTMAEPreTrainedModel, __webpack_exports__ViTMSNForImageClassification as ViTMSNForImageClassification, __webpack_exports__ViTMSNModel as ViTMSNModel, __webpack_exports__ViTMSNPreTrainedModel as ViTMSNPreTrainedModel, __webpack_exports__ViTModel as ViTModel, __webpack_exports__ViTPreTrainedModel as ViTPreTrainedModel, __webpack_exports__VisionEncoderDecoderModel as VisionEncoderDecoderModel, __webpack_exports__VitMatteForImageMatting as VitMatteForImageMatting, __webpack_exports__VitMatteImageProcessor as VitMatteImageProcessor, __webpack_exports__VitMattePreTrainedModel as VitMattePreTrainedModel, __webpack_exports__VitPoseForPoseEstimation as VitPoseForPoseEstimation, __webpack_exports__VitPoseImageProcessor as VitPoseImageProcessor, __webpack_exports__VitPosePreTrainedModel as VitPosePreTrainedModel, __webpack_exports__VitsModel as VitsModel, __webpack_exports__VitsModelOutput as VitsModelOutput, __webpack_exports__VitsPreTrainedModel as VitsPreTrainedModel, __webpack_exports__VitsTokenizer as VitsTokenizer, __webpack_exports__Wav2Vec2BertForCTC as Wav2Vec2BertForCTC, __webpack_exports__Wav2Vec2BertForSequenceClassification as Wav2Vec2BertForSequenceClassification, __webpack_exports__Wav2Vec2BertModel as Wav2Vec2BertModel, __webpack_exports__Wav2Vec2BertPreTrainedModel as Wav2Vec2BertPreTrainedModel, __webpack_exports__Wav2Vec2CTCTokenizer as Wav2Vec2CTCTokenizer, __webpack_exports__Wav2Vec2FeatureExtractor as Wav2Vec2FeatureExtractor, __webpack_exports__Wav2Vec2ForAudioFrameClassification as Wav2Vec2ForAudioFrameClassification, __webpack_exports__Wav2Vec2ForCTC as Wav2Vec2ForCTC, __webpack_exports__Wav2Vec2ForSequenceClassification as Wav2Vec2ForSequenceClassification, __webpack_exports__Wav2Vec2Model as Wav2Vec2Model, __webpack_exports__Wav2Vec2PreTrainedModel as Wav2Vec2PreTrainedModel, __webpack_exports__Wav2Vec2Processor as Wav2Vec2Processor, __webpack_exports__Wav2Vec2ProcessorWithLM as Wav2Vec2ProcessorWithLM, __webpack_exports__WavLMForAudioFrameClassification as WavLMForAudioFrameClassification, __webpack_exports__WavLMForCTC as WavLMForCTC, __webpack_exports__WavLMForSequenceClassification as WavLMForSequenceClassification, __webpack_exports__WavLMForXVector as WavLMForXVector, __webpack_exports__WavLMModel as WavLMModel, __webpack_exports__WavLMPreTrainedModel as WavLMPreTrainedModel, __webpack_exports__WeSpeakerFeatureExtractor as WeSpeakerFeatureExtractor, __webpack_exports__WeSpeakerResNetModel as WeSpeakerResNetModel, __webpack_exports__WeSpeakerResNetPreTrainedModel as WeSpeakerResNetPreTrainedModel, __webpack_exports__WhisperFeatureExtractor as WhisperFeatureExtractor, __webpack_exports__WhisperForConditionalGeneration as WhisperForConditionalGeneration, __webpack_exports__WhisperModel as WhisperModel, __webpack_exports__WhisperPreTrainedModel as WhisperPreTrainedModel, __webpack_exports__WhisperProcessor as WhisperProcessor, __webpack_exports__WhisperTextStreamer as WhisperTextStreamer, __webpack_exports__WhisperTimeStampLogitsProcessor as WhisperTimeStampLogitsProcessor, __webpack_exports__WhisperTokenizer as WhisperTokenizer, __webpack_exports__XLMForQuestionAnswering as XLMForQuestionAnswering, __webpack_exports__XLMForSequenceClassification as XLMForSequenceClassification, __webpack_exports__XLMForTokenClassification as XLMForTokenClassification, __webpack_exports__XLMModel as XLMModel, __webpack_exports__XLMPreTrainedModel as XLMPreTrainedModel, __webpack_exports__XLMRobertaForMaskedLM as XLMRobertaForMaskedLM, __webpack_exports__XLMRobertaForQuestionAnswering as XLMRobertaForQuestionAnswering, __webpack_exports__XLMRobertaForSequenceClassification as XLMRobertaForSequenceClassification, __webpack_exports__XLMRobertaForTokenClassification as XLMRobertaForTokenClassification, __webpack_exports__XLMRobertaModel as XLMRobertaModel, __webpack_exports__XLMRobertaPreTrainedModel as XLMRobertaPreTrainedModel, __webpack_exports__XLMRobertaTokenizer as XLMRobertaTokenizer, __webpack_exports__XLMTokenizer as XLMTokenizer, __webpack_exports__XLMWithLMHeadModel as XLMWithLMHeadModel, __webpack_exports__XVectorOutput as XVectorOutput, __webpack_exports__YolosFeatureExtractor as YolosFeatureExtractor, __webpack_exports__YolosForObjectDetection as YolosForObjectDetection, __webpack_exports__YolosImageProcessor as YolosImageProcessor, __webpack_exports__YolosModel as YolosModel, __webpack_exports__YolosObjectDetectionOutput as YolosObjectDetectionOutput, __webpack_exports__YolosPreTrainedModel as YolosPreTrainedModel, __webpack_exports__ZeroShotAudioClassificationPipeline as ZeroShotAudioClassificationPipeline, __webpack_exports__ZeroShotClassificationPipeline as ZeroShotClassificationPipeline, __webpack_exports__ZeroShotImageClassificationPipeline as ZeroShotImageClassificationPipeline, __webpack_exports__ZeroShotObjectDetectionPipeline as ZeroShotObjectDetectionPipeline, __webpack_exports__bankers_round as bankers_round, __webpack_exports__cat as cat, __webpack_exports__cos_sim as cos_sim, __webpack_exports__dot as dot, __webpack_exports__dynamic_time_warping as dynamic_time_warping, __webpack_exports__env as env, __webpack_exports__full as full, __webpack_exports__full_like as full_like, __webpack_exports__getKeyValueShapes as getKeyValueShapes, __webpack_exports__hamming as hamming, __webpack_exports__hanning as hanning, __webpack_exports__interpolate as interpolate, __webpack_exports__interpolate_4d as interpolate_4d, __webpack_exports__interpolate_data as interpolate_data, __webpack_exports__is_chinese_char as is_chinese_char, __webpack_exports__layer_norm as layer_norm, __webpack_exports__load_image as load_image, __webpack_exports__load_video as load_video, __webpack_exports__log_softmax as log_softmax, __webpack_exports__magnitude as magnitude, __webpack_exports__matmul as matmul, __webpack_exports__max as max, __webpack_exports__mean as mean, __webpack_exports__mean_pooling as mean_pooling, __webpack_exports__medianFilter as medianFilter, __webpack_exports__mel_filter_bank as mel_filter_bank, __webpack_exports__min as min, __webpack_exports__ones as ones, __webpack_exports__ones_like as ones_like, __webpack_exports__permute as permute, __webpack_exports__permute_data as permute_data, __webpack_exports__pipeline as pipeline, __webpack_exports__quantize_embeddings as quantize_embeddings, __webpack_exports__rand as rand, __webpack_exports__read_audio as read_audio, __webpack_exports__rfft as rfft, __webpack_exports__round as round, __webpack_exports__slice as slice, __webpack_exports__softmax as softmax, __webpack_exports__spectrogram as spectrogram, __webpack_exports__stack as stack, __webpack_exports__std_mean as std_mean, __webpack_exports__topk as topk, __webpack_exports__window_function as window_function, __webpack_exports__zeros as zeros, __webpack_exports__zeros_like as zeros_like };
|
|
36022
|
+
export { __webpack_exports__ASTFeatureExtractor as ASTFeatureExtractor, __webpack_exports__ASTForAudioClassification as ASTForAudioClassification, __webpack_exports__ASTModel as ASTModel, __webpack_exports__ASTPreTrainedModel as ASTPreTrainedModel, __webpack_exports__AlbertForMaskedLM as AlbertForMaskedLM, __webpack_exports__AlbertForQuestionAnswering as AlbertForQuestionAnswering, __webpack_exports__AlbertForSequenceClassification as AlbertForSequenceClassification, __webpack_exports__AlbertModel as AlbertModel, __webpack_exports__AlbertPreTrainedModel as AlbertPreTrainedModel, __webpack_exports__AlbertTokenizer as AlbertTokenizer, __webpack_exports__AudioClassificationPipeline as AudioClassificationPipeline, __webpack_exports__AutoConfig as AutoConfig, __webpack_exports__AutoFeatureExtractor as AutoFeatureExtractor, __webpack_exports__AutoImageProcessor as AutoImageProcessor, __webpack_exports__AutoModel as AutoModel, __webpack_exports__AutoModelForAudioClassification as AutoModelForAudioClassification, __webpack_exports__AutoModelForAudioFrameClassification as AutoModelForAudioFrameClassification, __webpack_exports__AutoModelForAudioTextToText as AutoModelForAudioTextToText, __webpack_exports__AutoModelForCTC as AutoModelForCTC, __webpack_exports__AutoModelForCausalLM as AutoModelForCausalLM, __webpack_exports__AutoModelForDepthEstimation as AutoModelForDepthEstimation, __webpack_exports__AutoModelForDocumentQuestionAnswering as AutoModelForDocumentQuestionAnswering, __webpack_exports__AutoModelForImageClassification as AutoModelForImageClassification, __webpack_exports__AutoModelForImageFeatureExtraction as AutoModelForImageFeatureExtraction, __webpack_exports__AutoModelForImageMatting as AutoModelForImageMatting, __webpack_exports__AutoModelForImageSegmentation as AutoModelForImageSegmentation, __webpack_exports__AutoModelForImageTextToText as AutoModelForImageTextToText, __webpack_exports__AutoModelForImageToImage as AutoModelForImageToImage, __webpack_exports__AutoModelForMaskGeneration as AutoModelForMaskGeneration, __webpack_exports__AutoModelForMaskedLM as AutoModelForMaskedLM, __webpack_exports__AutoModelForNormalEstimation as AutoModelForNormalEstimation, __webpack_exports__AutoModelForObjectDetection as AutoModelForObjectDetection, __webpack_exports__AutoModelForPoseEstimation as AutoModelForPoseEstimation, __webpack_exports__AutoModelForQuestionAnswering as AutoModelForQuestionAnswering, __webpack_exports__AutoModelForSemanticSegmentation as AutoModelForSemanticSegmentation, __webpack_exports__AutoModelForSeq2SeqLM as AutoModelForSeq2SeqLM, __webpack_exports__AutoModelForSequenceClassification as AutoModelForSequenceClassification, __webpack_exports__AutoModelForSpeechSeq2Seq as AutoModelForSpeechSeq2Seq, __webpack_exports__AutoModelForTextToSpectrogram as AutoModelForTextToSpectrogram, __webpack_exports__AutoModelForTextToWaveform as AutoModelForTextToWaveform, __webpack_exports__AutoModelForTokenClassification as AutoModelForTokenClassification, __webpack_exports__AutoModelForUniversalSegmentation as AutoModelForUniversalSegmentation, __webpack_exports__AutoModelForVision2Seq as AutoModelForVision2Seq, __webpack_exports__AutoModelForXVector as AutoModelForXVector, __webpack_exports__AutoModelForZeroShotObjectDetection as AutoModelForZeroShotObjectDetection, __webpack_exports__AutoProcessor as AutoProcessor, __webpack_exports__AutoTokenizer as AutoTokenizer, __webpack_exports__AutomaticSpeechRecognitionPipeline as AutomaticSpeechRecognitionPipeline, __webpack_exports__BackgroundRemovalPipeline as BackgroundRemovalPipeline, __webpack_exports__BartForConditionalGeneration as BartForConditionalGeneration, __webpack_exports__BartForSequenceClassification as BartForSequenceClassification, __webpack_exports__BartModel as BartModel, __webpack_exports__BartPretrainedModel as BartPretrainedModel, __webpack_exports__BartTokenizer as BartTokenizer, __webpack_exports__BaseModelOutput as BaseModelOutput, __webpack_exports__BaseStreamer as BaseStreamer, __webpack_exports__BeitFeatureExtractor as BeitFeatureExtractor, __webpack_exports__BeitForImageClassification as BeitForImageClassification, __webpack_exports__BeitModel as BeitModel, __webpack_exports__BeitPreTrainedModel as BeitPreTrainedModel, __webpack_exports__BertForMaskedLM as BertForMaskedLM, __webpack_exports__BertForQuestionAnswering as BertForQuestionAnswering, __webpack_exports__BertForSequenceClassification as BertForSequenceClassification, __webpack_exports__BertForTokenClassification as BertForTokenClassification, __webpack_exports__BertModel as BertModel, __webpack_exports__BertPreTrainedModel as BertPreTrainedModel, __webpack_exports__BertTokenizer as BertTokenizer, __webpack_exports__BitImageProcessor as BitImageProcessor, __webpack_exports__BlenderbotForConditionalGeneration as BlenderbotForConditionalGeneration, __webpack_exports__BlenderbotModel as BlenderbotModel, __webpack_exports__BlenderbotPreTrainedModel as BlenderbotPreTrainedModel, __webpack_exports__BlenderbotSmallForConditionalGeneration as BlenderbotSmallForConditionalGeneration, __webpack_exports__BlenderbotSmallModel as BlenderbotSmallModel, __webpack_exports__BlenderbotSmallPreTrainedModel as BlenderbotSmallPreTrainedModel, __webpack_exports__BlenderbotSmallTokenizer as BlenderbotSmallTokenizer, __webpack_exports__BlenderbotTokenizer as BlenderbotTokenizer, __webpack_exports__BloomForCausalLM as BloomForCausalLM, __webpack_exports__BloomModel as BloomModel, __webpack_exports__BloomPreTrainedModel as BloomPreTrainedModel, __webpack_exports__BloomTokenizer as BloomTokenizer, __webpack_exports__CLIPFeatureExtractor as CLIPFeatureExtractor, __webpack_exports__CLIPImageProcessor as CLIPImageProcessor, __webpack_exports__CLIPModel as CLIPModel, __webpack_exports__CLIPPreTrainedModel as CLIPPreTrainedModel, __webpack_exports__CLIPSegForImageSegmentation as CLIPSegForImageSegmentation, __webpack_exports__CLIPSegModel as CLIPSegModel, __webpack_exports__CLIPSegPreTrainedModel as CLIPSegPreTrainedModel, __webpack_exports__CLIPTextModel as CLIPTextModel, __webpack_exports__CLIPTextModelWithProjection as CLIPTextModelWithProjection, __webpack_exports__CLIPTokenizer as CLIPTokenizer, __webpack_exports__CLIPVisionModel as CLIPVisionModel, __webpack_exports__CLIPVisionModelWithProjection as CLIPVisionModelWithProjection, __webpack_exports__CamembertForMaskedLM as CamembertForMaskedLM, __webpack_exports__CamembertForQuestionAnswering as CamembertForQuestionAnswering, __webpack_exports__CamembertForSequenceClassification as CamembertForSequenceClassification, __webpack_exports__CamembertForTokenClassification as CamembertForTokenClassification, __webpack_exports__CamembertModel as CamembertModel, __webpack_exports__CamembertPreTrainedModel as CamembertPreTrainedModel, __webpack_exports__CamembertTokenizer as CamembertTokenizer, __webpack_exports__CausalLMOutput as CausalLMOutput, __webpack_exports__CausalLMOutputWithPast as CausalLMOutputWithPast, __webpack_exports__ChineseCLIPFeatureExtractor as ChineseCLIPFeatureExtractor, __webpack_exports__ChineseCLIPModel as ChineseCLIPModel, __webpack_exports__ChineseCLIPPreTrainedModel as ChineseCLIPPreTrainedModel, __webpack_exports__ClapAudioModelWithProjection as ClapAudioModelWithProjection, __webpack_exports__ClapFeatureExtractor as ClapFeatureExtractor, __webpack_exports__ClapModel as ClapModel, __webpack_exports__ClapPreTrainedModel as ClapPreTrainedModel, __webpack_exports__ClapTextModelWithProjection as ClapTextModelWithProjection, __webpack_exports__ClassifierFreeGuidanceLogitsProcessor as ClassifierFreeGuidanceLogitsProcessor, __webpack_exports__CodeGenForCausalLM as CodeGenForCausalLM, __webpack_exports__CodeGenModel as CodeGenModel, __webpack_exports__CodeGenPreTrainedModel as CodeGenPreTrainedModel, __webpack_exports__CodeGenTokenizer as CodeGenTokenizer, __webpack_exports__CodeLlamaTokenizer as CodeLlamaTokenizer, __webpack_exports__CohereForCausalLM as CohereForCausalLM, __webpack_exports__CohereModel as CohereModel, __webpack_exports__CoherePreTrainedModel as CoherePreTrainedModel, __webpack_exports__CohereTokenizer as CohereTokenizer, __webpack_exports__ConvBertForMaskedLM as ConvBertForMaskedLM, __webpack_exports__ConvBertForQuestionAnswering as ConvBertForQuestionAnswering, __webpack_exports__ConvBertForSequenceClassification as ConvBertForSequenceClassification, __webpack_exports__ConvBertForTokenClassification as ConvBertForTokenClassification, __webpack_exports__ConvBertModel as ConvBertModel, __webpack_exports__ConvBertPreTrainedModel as ConvBertPreTrainedModel, __webpack_exports__ConvBertTokenizer as ConvBertTokenizer, __webpack_exports__ConvNextFeatureExtractor as ConvNextFeatureExtractor, __webpack_exports__ConvNextForImageClassification as ConvNextForImageClassification, __webpack_exports__ConvNextImageProcessor as ConvNextImageProcessor, __webpack_exports__ConvNextModel as ConvNextModel, __webpack_exports__ConvNextPreTrainedModel as ConvNextPreTrainedModel, __webpack_exports__ConvNextV2ForImageClassification as ConvNextV2ForImageClassification, __webpack_exports__ConvNextV2Model as ConvNextV2Model, __webpack_exports__ConvNextV2PreTrainedModel as ConvNextV2PreTrainedModel, __webpack_exports__DPTFeatureExtractor as DPTFeatureExtractor, __webpack_exports__DPTForDepthEstimation as DPTForDepthEstimation, __webpack_exports__DPTImageProcessor as DPTImageProcessor, __webpack_exports__DPTModel as DPTModel, __webpack_exports__DPTPreTrainedModel as DPTPreTrainedModel, __webpack_exports__DacDecoderModel as DacDecoderModel, __webpack_exports__DacDecoderOutput as DacDecoderOutput, __webpack_exports__DacEncoderModel as DacEncoderModel, __webpack_exports__DacEncoderOutput as DacEncoderOutput, __webpack_exports__DacFeatureExtractor as DacFeatureExtractor, __webpack_exports__DacModel as DacModel, __webpack_exports__DacPreTrainedModel as DacPreTrainedModel, __webpack_exports__DataTypeMap as DataTypeMap, __webpack_exports__DebertaForMaskedLM as DebertaForMaskedLM, __webpack_exports__DebertaForQuestionAnswering as DebertaForQuestionAnswering, __webpack_exports__DebertaForSequenceClassification as DebertaForSequenceClassification, __webpack_exports__DebertaForTokenClassification as DebertaForTokenClassification, __webpack_exports__DebertaModel as DebertaModel, __webpack_exports__DebertaPreTrainedModel as DebertaPreTrainedModel, __webpack_exports__DebertaTokenizer as DebertaTokenizer, __webpack_exports__DebertaV2ForMaskedLM as DebertaV2ForMaskedLM, __webpack_exports__DebertaV2ForQuestionAnswering as DebertaV2ForQuestionAnswering, __webpack_exports__DebertaV2ForSequenceClassification as DebertaV2ForSequenceClassification, __webpack_exports__DebertaV2ForTokenClassification as DebertaV2ForTokenClassification, __webpack_exports__DebertaV2Model as DebertaV2Model, __webpack_exports__DebertaV2PreTrainedModel as DebertaV2PreTrainedModel, __webpack_exports__DebertaV2Tokenizer as DebertaV2Tokenizer, __webpack_exports__DecisionTransformerModel as DecisionTransformerModel, __webpack_exports__DecisionTransformerPreTrainedModel as DecisionTransformerPreTrainedModel, __webpack_exports__DeiTFeatureExtractor as DeiTFeatureExtractor, __webpack_exports__DeiTForImageClassification as DeiTForImageClassification, __webpack_exports__DeiTImageProcessor as DeiTImageProcessor, __webpack_exports__DeiTModel as DeiTModel, __webpack_exports__DeiTPreTrainedModel as DeiTPreTrainedModel, __webpack_exports__DepthAnythingForDepthEstimation as DepthAnythingForDepthEstimation, __webpack_exports__DepthAnythingPreTrainedModel as DepthAnythingPreTrainedModel, __webpack_exports__DepthEstimationPipeline as DepthEstimationPipeline, __webpack_exports__DepthProForDepthEstimation as DepthProForDepthEstimation, __webpack_exports__DepthProPreTrainedModel as DepthProPreTrainedModel, __webpack_exports__DetrFeatureExtractor as DetrFeatureExtractor, __webpack_exports__DetrForObjectDetection as DetrForObjectDetection, __webpack_exports__DetrForSegmentation as DetrForSegmentation, __webpack_exports__DetrImageProcessor as DetrImageProcessor, __webpack_exports__DetrModel as DetrModel, __webpack_exports__DetrObjectDetectionOutput as DetrObjectDetectionOutput, __webpack_exports__DetrPreTrainedModel as DetrPreTrainedModel, __webpack_exports__DetrSegmentationOutput as DetrSegmentationOutput, __webpack_exports__Dinov2ForImageClassification as Dinov2ForImageClassification, __webpack_exports__Dinov2Model as Dinov2Model, __webpack_exports__Dinov2PreTrainedModel as Dinov2PreTrainedModel, __webpack_exports__Dinov2WithRegistersForImageClassification as Dinov2WithRegistersForImageClassification, __webpack_exports__Dinov2WithRegistersModel as Dinov2WithRegistersModel, __webpack_exports__Dinov2WithRegistersPreTrainedModel as Dinov2WithRegistersPreTrainedModel, __webpack_exports__DistilBertForMaskedLM as DistilBertForMaskedLM, __webpack_exports__DistilBertForQuestionAnswering as DistilBertForQuestionAnswering, __webpack_exports__DistilBertForSequenceClassification as DistilBertForSequenceClassification, __webpack_exports__DistilBertForTokenClassification as DistilBertForTokenClassification, __webpack_exports__DistilBertModel as DistilBertModel, __webpack_exports__DistilBertPreTrainedModel as DistilBertPreTrainedModel, __webpack_exports__DistilBertTokenizer as DistilBertTokenizer, __webpack_exports__DocumentQuestionAnsweringPipeline as DocumentQuestionAnsweringPipeline, __webpack_exports__DonutFeatureExtractor as DonutFeatureExtractor, __webpack_exports__DonutImageProcessor as DonutImageProcessor, __webpack_exports__DonutSwinModel as DonutSwinModel, __webpack_exports__DonutSwinPreTrainedModel as DonutSwinPreTrainedModel, __webpack_exports__EfficientNetForImageClassification as EfficientNetForImageClassification, __webpack_exports__EfficientNetImageProcessor as EfficientNetImageProcessor, __webpack_exports__EfficientNetModel as EfficientNetModel, __webpack_exports__EfficientNetPreTrainedModel as EfficientNetPreTrainedModel, __webpack_exports__ElectraForMaskedLM as ElectraForMaskedLM, __webpack_exports__ElectraForQuestionAnswering as ElectraForQuestionAnswering, __webpack_exports__ElectraForSequenceClassification as ElectraForSequenceClassification, __webpack_exports__ElectraForTokenClassification as ElectraForTokenClassification, __webpack_exports__ElectraModel as ElectraModel, __webpack_exports__ElectraPreTrainedModel as ElectraPreTrainedModel, __webpack_exports__ElectraTokenizer as ElectraTokenizer, __webpack_exports__EncodecFeatureExtractor as EncodecFeatureExtractor, __webpack_exports__EosTokenCriteria as EosTokenCriteria, __webpack_exports__EsmForMaskedLM as EsmForMaskedLM, __webpack_exports__EsmForSequenceClassification as EsmForSequenceClassification, __webpack_exports__EsmForTokenClassification as EsmForTokenClassification, __webpack_exports__EsmModel as EsmModel, __webpack_exports__EsmPreTrainedModel as EsmPreTrainedModel, __webpack_exports__EsmTokenizer as EsmTokenizer, __webpack_exports__ExaoneForCausalLM as ExaoneForCausalLM, __webpack_exports__ExaoneModel as ExaoneModel, __webpack_exports__ExaonePreTrainedModel as ExaonePreTrainedModel, __webpack_exports__FFT as FFT, __webpack_exports__FalconForCausalLM as FalconForCausalLM, __webpack_exports__FalconModel as FalconModel, __webpack_exports__FalconPreTrainedModel as FalconPreTrainedModel, __webpack_exports__FalconTokenizer as FalconTokenizer, __webpack_exports__FastViTForImageClassification as FastViTForImageClassification, __webpack_exports__FastViTModel as FastViTModel, __webpack_exports__FastViTPreTrainedModel as FastViTPreTrainedModel, __webpack_exports__FeatureExtractionPipeline as FeatureExtractionPipeline, __webpack_exports__FeatureExtractor as FeatureExtractor, __webpack_exports__FillMaskPipeline as FillMaskPipeline, __webpack_exports__Florence2ForConditionalGeneration as Florence2ForConditionalGeneration, __webpack_exports__Florence2PreTrainedModel as Florence2PreTrainedModel, __webpack_exports__Florence2Processor as Florence2Processor, __webpack_exports__ForcedBOSTokenLogitsProcessor as ForcedBOSTokenLogitsProcessor, __webpack_exports__ForcedEOSTokenLogitsProcessor as ForcedEOSTokenLogitsProcessor, __webpack_exports__GLPNFeatureExtractor as GLPNFeatureExtractor, __webpack_exports__GLPNForDepthEstimation as GLPNForDepthEstimation, __webpack_exports__GLPNModel as GLPNModel, __webpack_exports__GLPNPreTrainedModel as GLPNPreTrainedModel, __webpack_exports__GPT2LMHeadModel as GPT2LMHeadModel, __webpack_exports__GPT2Model as GPT2Model, __webpack_exports__GPT2PreTrainedModel as GPT2PreTrainedModel, __webpack_exports__GPT2Tokenizer as GPT2Tokenizer, __webpack_exports__GPTBigCodeForCausalLM as GPTBigCodeForCausalLM, __webpack_exports__GPTBigCodeModel as GPTBigCodeModel, __webpack_exports__GPTBigCodePreTrainedModel as GPTBigCodePreTrainedModel, __webpack_exports__GPTJForCausalLM as GPTJForCausalLM, __webpack_exports__GPTJModel as GPTJModel, __webpack_exports__GPTJPreTrainedModel as GPTJPreTrainedModel, __webpack_exports__GPTNeoForCausalLM as GPTNeoForCausalLM, __webpack_exports__GPTNeoModel as GPTNeoModel, __webpack_exports__GPTNeoPreTrainedModel as GPTNeoPreTrainedModel, __webpack_exports__GPTNeoXForCausalLM as GPTNeoXForCausalLM, __webpack_exports__GPTNeoXModel as GPTNeoXModel, __webpack_exports__GPTNeoXPreTrainedModel as GPTNeoXPreTrainedModel, __webpack_exports__GPTNeoXTokenizer as GPTNeoXTokenizer, __webpack_exports__Gemma2ForCausalLM as Gemma2ForCausalLM, __webpack_exports__Gemma2Model as Gemma2Model, __webpack_exports__Gemma2PreTrainedModel as Gemma2PreTrainedModel, __webpack_exports__Gemma3ForCausalLM as Gemma3ForCausalLM, __webpack_exports__Gemma3Model as Gemma3Model, __webpack_exports__Gemma3PreTrainedModel as Gemma3PreTrainedModel, __webpack_exports__GemmaForCausalLM as GemmaForCausalLM, __webpack_exports__GemmaModel as GemmaModel, __webpack_exports__GemmaPreTrainedModel as GemmaPreTrainedModel, __webpack_exports__GemmaTokenizer as GemmaTokenizer, __webpack_exports__GlmForCausalLM as GlmForCausalLM, __webpack_exports__GlmModel as GlmModel, __webpack_exports__GlmPreTrainedModel as GlmPreTrainedModel, __webpack_exports__GraniteForCausalLM as GraniteForCausalLM, __webpack_exports__GraniteModel as GraniteModel, __webpack_exports__GranitePreTrainedModel as GranitePreTrainedModel, __webpack_exports__Grok1Tokenizer as Grok1Tokenizer, __webpack_exports__GroundingDinoForObjectDetection as GroundingDinoForObjectDetection, __webpack_exports__GroundingDinoImageProcessor as GroundingDinoImageProcessor, __webpack_exports__GroundingDinoPreTrainedModel as GroundingDinoPreTrainedModel, __webpack_exports__GroundingDinoProcessor as GroundingDinoProcessor, __webpack_exports__GroupViTModel as GroupViTModel, __webpack_exports__GroupViTPreTrainedModel as GroupViTPreTrainedModel, __webpack_exports__HeliumForCausalLM as HeliumForCausalLM, __webpack_exports__HeliumModel as HeliumModel, __webpack_exports__HeliumPreTrainedModel as HeliumPreTrainedModel, __webpack_exports__HerbertTokenizer as HerbertTokenizer, __webpack_exports__HieraForImageClassification as HieraForImageClassification, __webpack_exports__HieraModel as HieraModel, __webpack_exports__HieraPreTrainedModel as HieraPreTrainedModel, __webpack_exports__HubertForCTC as HubertForCTC, __webpack_exports__HubertForSequenceClassification as HubertForSequenceClassification, __webpack_exports__HubertModel as HubertModel, __webpack_exports__HubertPreTrainedModel as HubertPreTrainedModel, __webpack_exports__IJepaForImageClassification as IJepaForImageClassification, __webpack_exports__IJepaModel as IJepaModel, __webpack_exports__IJepaPreTrainedModel as IJepaPreTrainedModel, __webpack_exports__Idefics3ForConditionalGeneration as Idefics3ForConditionalGeneration, __webpack_exports__Idefics3ImageProcessor as Idefics3ImageProcessor, __webpack_exports__Idefics3PreTrainedModel as Idefics3PreTrainedModel, __webpack_exports__Idefics3Processor as Idefics3Processor, __webpack_exports__ImageClassificationPipeline as ImageClassificationPipeline, __webpack_exports__ImageFeatureExtractionPipeline as ImageFeatureExtractionPipeline, __webpack_exports__ImageFeatureExtractor as ImageFeatureExtractor, __webpack_exports__ImageMattingOutput as ImageMattingOutput, __webpack_exports__ImageProcessor as ImageProcessor, __webpack_exports__ImageSegmentationPipeline as ImageSegmentationPipeline, __webpack_exports__ImageToImagePipeline as ImageToImagePipeline, __webpack_exports__ImageToTextPipeline as ImageToTextPipeline, __webpack_exports__InterruptableStoppingCriteria as InterruptableStoppingCriteria, __webpack_exports__JAISLMHeadModel as JAISLMHeadModel, __webpack_exports__JAISModel as JAISModel, __webpack_exports__JAISPreTrainedModel as JAISPreTrainedModel, __webpack_exports__JinaCLIPImageProcessor as JinaCLIPImageProcessor, __webpack_exports__JinaCLIPModel as JinaCLIPModel, __webpack_exports__JinaCLIPPreTrainedModel as JinaCLIPPreTrainedModel, __webpack_exports__JinaCLIPProcessor as JinaCLIPProcessor, __webpack_exports__JinaCLIPTextModel as JinaCLIPTextModel, __webpack_exports__JinaCLIPVisionModel as JinaCLIPVisionModel, __webpack_exports__LiteWhisperForConditionalGeneration as LiteWhisperForConditionalGeneration, __webpack_exports__LlamaForCausalLM as LlamaForCausalLM, __webpack_exports__LlamaModel as LlamaModel, __webpack_exports__LlamaPreTrainedModel as LlamaPreTrainedModel, __webpack_exports__LlamaTokenizer as LlamaTokenizer, __webpack_exports__LlavaForConditionalGeneration as LlavaForConditionalGeneration, __webpack_exports__LlavaOnevisionForConditionalGeneration as LlavaOnevisionForConditionalGeneration, __webpack_exports__LlavaOnevisionImageProcessor as LlavaOnevisionImageProcessor, __webpack_exports__LlavaPreTrainedModel as LlavaPreTrainedModel, __webpack_exports__LogitsProcessor as LogitsProcessor, __webpack_exports__LogitsProcessorList as LogitsProcessorList, __webpack_exports__LogitsWarper as LogitsWarper, __webpack_exports__LongT5ForConditionalGeneration as LongT5ForConditionalGeneration, __webpack_exports__LongT5Model as LongT5Model, __webpack_exports__LongT5PreTrainedModel as LongT5PreTrainedModel, __webpack_exports__M2M100ForConditionalGeneration as M2M100ForConditionalGeneration, __webpack_exports__M2M100Model as M2M100Model, __webpack_exports__M2M100PreTrainedModel as M2M100PreTrainedModel, __webpack_exports__M2M100Tokenizer as M2M100Tokenizer, __webpack_exports__MBart50Tokenizer as MBart50Tokenizer, __webpack_exports__MBartForCausalLM as MBartForCausalLM, __webpack_exports__MBartForConditionalGeneration as MBartForConditionalGeneration, __webpack_exports__MBartForSequenceClassification as MBartForSequenceClassification, __webpack_exports__MBartModel as MBartModel, __webpack_exports__MBartPreTrainedModel as MBartPreTrainedModel, __webpack_exports__MBartTokenizer as MBartTokenizer, __webpack_exports__MPNetForMaskedLM as MPNetForMaskedLM, __webpack_exports__MPNetForQuestionAnswering as MPNetForQuestionAnswering, __webpack_exports__MPNetForSequenceClassification as MPNetForSequenceClassification, __webpack_exports__MPNetForTokenClassification as MPNetForTokenClassification, __webpack_exports__MPNetModel as MPNetModel, __webpack_exports__MPNetPreTrainedModel as MPNetPreTrainedModel, __webpack_exports__MPNetTokenizer as MPNetTokenizer, __webpack_exports__MT5ForConditionalGeneration as MT5ForConditionalGeneration, __webpack_exports__MT5Model as MT5Model, __webpack_exports__MT5PreTrainedModel as MT5PreTrainedModel, __webpack_exports__MarianMTModel as MarianMTModel, __webpack_exports__MarianModel as MarianModel, __webpack_exports__MarianPreTrainedModel as MarianPreTrainedModel, __webpack_exports__MarianTokenizer as MarianTokenizer, __webpack_exports__Mask2FormerImageProcessor as Mask2FormerImageProcessor, __webpack_exports__MaskFormerFeatureExtractor as MaskFormerFeatureExtractor, __webpack_exports__MaskFormerForInstanceSegmentation as MaskFormerForInstanceSegmentation, __webpack_exports__MaskFormerImageProcessor as MaskFormerImageProcessor, __webpack_exports__MaskFormerModel as MaskFormerModel, __webpack_exports__MaskFormerPreTrainedModel as MaskFormerPreTrainedModel, __webpack_exports__MaskedLMOutput as MaskedLMOutput, __webpack_exports__MaxLengthCriteria as MaxLengthCriteria, __webpack_exports__Metric3DForDepthEstimation as Metric3DForDepthEstimation, __webpack_exports__Metric3DPreTrainedModel as Metric3DPreTrainedModel, __webpack_exports__Metric3Dv2ForDepthEstimation as Metric3Dv2ForDepthEstimation, __webpack_exports__Metric3Dv2PreTrainedModel as Metric3Dv2PreTrainedModel, __webpack_exports__MgpstrForSceneTextRecognition as MgpstrForSceneTextRecognition, __webpack_exports__MgpstrModelOutput as MgpstrModelOutput, __webpack_exports__MgpstrPreTrainedModel as MgpstrPreTrainedModel, __webpack_exports__MgpstrProcessor as MgpstrProcessor, __webpack_exports__MgpstrTokenizer as MgpstrTokenizer, __webpack_exports__MimiDecoderModel as MimiDecoderModel, __webpack_exports__MimiDecoderOutput as MimiDecoderOutput, __webpack_exports__MimiEncoderModel as MimiEncoderModel, __webpack_exports__MimiEncoderOutput as MimiEncoderOutput, __webpack_exports__MimiModel as MimiModel, __webpack_exports__MimiPreTrainedModel as MimiPreTrainedModel, __webpack_exports__MinLengthLogitsProcessor as MinLengthLogitsProcessor, __webpack_exports__MinNewTokensLengthLogitsProcessor as MinNewTokensLengthLogitsProcessor, __webpack_exports__MistralForCausalLM as MistralForCausalLM, __webpack_exports__MistralModel as MistralModel, __webpack_exports__MistralPreTrainedModel as MistralPreTrainedModel, __webpack_exports__MobileBertForMaskedLM as MobileBertForMaskedLM, __webpack_exports__MobileBertForQuestionAnswering as MobileBertForQuestionAnswering, __webpack_exports__MobileBertForSequenceClassification as MobileBertForSequenceClassification, __webpack_exports__MobileBertModel as MobileBertModel, __webpack_exports__MobileBertPreTrainedModel as MobileBertPreTrainedModel, __webpack_exports__MobileBertTokenizer as MobileBertTokenizer, __webpack_exports__MobileLLMForCausalLM as MobileLLMForCausalLM, __webpack_exports__MobileLLMModel as MobileLLMModel, __webpack_exports__MobileLLMPreTrainedModel as MobileLLMPreTrainedModel, __webpack_exports__MobileNetV1FeatureExtractor as MobileNetV1FeatureExtractor, __webpack_exports__MobileNetV1ForImageClassification as MobileNetV1ForImageClassification, __webpack_exports__MobileNetV1ForSemanticSegmentation as MobileNetV1ForSemanticSegmentation, __webpack_exports__MobileNetV1ImageProcessor as MobileNetV1ImageProcessor, __webpack_exports__MobileNetV1Model as MobileNetV1Model, __webpack_exports__MobileNetV1PreTrainedModel as MobileNetV1PreTrainedModel, __webpack_exports__MobileNetV2FeatureExtractor as MobileNetV2FeatureExtractor, __webpack_exports__MobileNetV2ForImageClassification as MobileNetV2ForImageClassification, __webpack_exports__MobileNetV2ForSemanticSegmentation as MobileNetV2ForSemanticSegmentation, __webpack_exports__MobileNetV2ImageProcessor as MobileNetV2ImageProcessor, __webpack_exports__MobileNetV2Model as MobileNetV2Model, __webpack_exports__MobileNetV2PreTrainedModel as MobileNetV2PreTrainedModel, __webpack_exports__MobileNetV3FeatureExtractor as MobileNetV3FeatureExtractor, __webpack_exports__MobileNetV3ForImageClassification as MobileNetV3ForImageClassification, __webpack_exports__MobileNetV3ForSemanticSegmentation as MobileNetV3ForSemanticSegmentation, __webpack_exports__MobileNetV3ImageProcessor as MobileNetV3ImageProcessor, __webpack_exports__MobileNetV3Model as MobileNetV3Model, __webpack_exports__MobileNetV3PreTrainedModel as MobileNetV3PreTrainedModel, __webpack_exports__MobileNetV4FeatureExtractor as MobileNetV4FeatureExtractor, __webpack_exports__MobileNetV4ForImageClassification as MobileNetV4ForImageClassification, __webpack_exports__MobileNetV4ForSemanticSegmentation as MobileNetV4ForSemanticSegmentation, __webpack_exports__MobileNetV4ImageProcessor as MobileNetV4ImageProcessor, __webpack_exports__MobileNetV4Model as MobileNetV4Model, __webpack_exports__MobileNetV4PreTrainedModel as MobileNetV4PreTrainedModel, __webpack_exports__MobileViTFeatureExtractor as MobileViTFeatureExtractor, __webpack_exports__MobileViTForImageClassification as MobileViTForImageClassification, __webpack_exports__MobileViTImageProcessor as MobileViTImageProcessor, __webpack_exports__MobileViTModel as MobileViTModel, __webpack_exports__MobileViTPreTrainedModel as MobileViTPreTrainedModel, __webpack_exports__MobileViTV2ForImageClassification as MobileViTV2ForImageClassification, __webpack_exports__MobileViTV2Model as MobileViTV2Model, __webpack_exports__MobileViTV2PreTrainedModel as MobileViTV2PreTrainedModel, __webpack_exports__ModelOutput as ModelOutput, __webpack_exports__ModernBertForMaskedLM as ModernBertForMaskedLM, __webpack_exports__ModernBertForSequenceClassification as ModernBertForSequenceClassification, __webpack_exports__ModernBertForTokenClassification as ModernBertForTokenClassification, __webpack_exports__ModernBertModel as ModernBertModel, __webpack_exports__ModernBertPreTrainedModel as ModernBertPreTrainedModel, __webpack_exports__Moondream1ForConditionalGeneration as Moondream1ForConditionalGeneration, __webpack_exports__MoonshineFeatureExtractor as MoonshineFeatureExtractor, __webpack_exports__MoonshineForConditionalGeneration as MoonshineForConditionalGeneration, __webpack_exports__MoonshineModel as MoonshineModel, __webpack_exports__MoonshinePreTrainedModel as MoonshinePreTrainedModel, __webpack_exports__MoonshineProcessor as MoonshineProcessor, __webpack_exports__MptForCausalLM as MptForCausalLM, __webpack_exports__MptModel as MptModel, __webpack_exports__MptPreTrainedModel as MptPreTrainedModel, __webpack_exports__MultiModalityCausalLM as MultiModalityCausalLM, __webpack_exports__MultiModalityPreTrainedModel as MultiModalityPreTrainedModel, __webpack_exports__MusicgenForCausalLM as MusicgenForCausalLM, __webpack_exports__MusicgenForConditionalGeneration as MusicgenForConditionalGeneration, __webpack_exports__MusicgenModel as MusicgenModel, __webpack_exports__MusicgenPreTrainedModel as MusicgenPreTrainedModel, __webpack_exports__NllbTokenizer as NllbTokenizer, __webpack_exports__NoBadWordsLogitsProcessor as NoBadWordsLogitsProcessor, __webpack_exports__NoRepeatNGramLogitsProcessor as NoRepeatNGramLogitsProcessor, __webpack_exports__NomicBertModel as NomicBertModel, __webpack_exports__NomicBertPreTrainedModel as NomicBertPreTrainedModel, __webpack_exports__NougatImageProcessor as NougatImageProcessor, __webpack_exports__NougatTokenizer as NougatTokenizer, __webpack_exports__OPTForCausalLM as OPTForCausalLM, __webpack_exports__OPTModel as OPTModel, __webpack_exports__OPTPreTrainedModel as OPTPreTrainedModel, __webpack_exports__ObjectDetectionPipeline as ObjectDetectionPipeline, __webpack_exports__Olmo2ForCausalLM as Olmo2ForCausalLM, __webpack_exports__Olmo2Model as Olmo2Model, __webpack_exports__Olmo2PreTrainedModel as Olmo2PreTrainedModel, __webpack_exports__OlmoForCausalLM as OlmoForCausalLM, __webpack_exports__OlmoModel as OlmoModel, __webpack_exports__OlmoPreTrainedModel as OlmoPreTrainedModel, __webpack_exports__OpenELMForCausalLM as OpenELMForCausalLM, __webpack_exports__OpenELMModel as OpenELMModel, __webpack_exports__OpenELMPreTrainedModel as OpenELMPreTrainedModel, __webpack_exports__OwlViTFeatureExtractor as OwlViTFeatureExtractor, __webpack_exports__OwlViTForObjectDetection as OwlViTForObjectDetection, __webpack_exports__OwlViTImageProcessor as OwlViTImageProcessor, __webpack_exports__OwlViTModel as OwlViTModel, __webpack_exports__OwlViTPreTrainedModel as OwlViTPreTrainedModel, __webpack_exports__OwlViTProcessor as OwlViTProcessor, __webpack_exports__Owlv2ForObjectDetection as Owlv2ForObjectDetection, __webpack_exports__Owlv2ImageProcessor as Owlv2ImageProcessor, __webpack_exports__Owlv2Model as Owlv2Model, __webpack_exports__Owlv2PreTrainedModel as Owlv2PreTrainedModel, __webpack_exports__PaliGemmaForConditionalGeneration as PaliGemmaForConditionalGeneration, __webpack_exports__PaliGemmaPreTrainedModel as PaliGemmaPreTrainedModel, __webpack_exports__PaliGemmaProcessor as PaliGemmaProcessor, __webpack_exports__PatchTSMixerForPrediction as PatchTSMixerForPrediction, __webpack_exports__PatchTSMixerModel as PatchTSMixerModel, __webpack_exports__PatchTSMixerPreTrainedModel as PatchTSMixerPreTrainedModel, __webpack_exports__PatchTSTForPrediction as PatchTSTForPrediction, __webpack_exports__PatchTSTModel as PatchTSTModel, __webpack_exports__PatchTSTPreTrainedModel as PatchTSTPreTrainedModel, __webpack_exports__Phi3ForCausalLM as Phi3ForCausalLM, __webpack_exports__Phi3Model as Phi3Model, __webpack_exports__Phi3PreTrainedModel as Phi3PreTrainedModel, __webpack_exports__Phi3VForCausalLM as Phi3VForCausalLM, __webpack_exports__Phi3VImageProcessor as Phi3VImageProcessor, __webpack_exports__Phi3VPreTrainedModel as Phi3VPreTrainedModel, __webpack_exports__Phi3VProcessor as Phi3VProcessor, __webpack_exports__PhiForCausalLM as PhiForCausalLM, __webpack_exports__PhiModel as PhiModel, __webpack_exports__PhiPreTrainedModel as PhiPreTrainedModel, __webpack_exports__Pipeline as Pipeline, __webpack_exports__PreTrainedModel as PreTrainedModel, __webpack_exports__PreTrainedTokenizer as PreTrainedTokenizer, __webpack_exports__PretrainedConfig as PretrainedConfig, __webpack_exports__PretrainedMixin as PretrainedMixin, __webpack_exports__Processor as Processor, __webpack_exports__PvtForImageClassification as PvtForImageClassification, __webpack_exports__PvtImageProcessor as PvtImageProcessor, __webpack_exports__PvtModel as PvtModel, __webpack_exports__PvtPreTrainedModel as PvtPreTrainedModel, __webpack_exports__PyAnnoteFeatureExtractor as PyAnnoteFeatureExtractor, __webpack_exports__PyAnnoteForAudioFrameClassification as PyAnnoteForAudioFrameClassification, __webpack_exports__PyAnnoteModel as PyAnnoteModel, __webpack_exports__PyAnnotePreTrainedModel as PyAnnotePreTrainedModel, __webpack_exports__PyAnnoteProcessor as PyAnnoteProcessor, __webpack_exports__QuestionAnsweringModelOutput as QuestionAnsweringModelOutput, __webpack_exports__QuestionAnsweringPipeline as QuestionAnsweringPipeline, __webpack_exports__Qwen2ForCausalLM as Qwen2ForCausalLM, __webpack_exports__Qwen2Model as Qwen2Model, __webpack_exports__Qwen2PreTrainedModel as Qwen2PreTrainedModel, __webpack_exports__Qwen2Tokenizer as Qwen2Tokenizer, __webpack_exports__Qwen2VLForConditionalGeneration as Qwen2VLForConditionalGeneration, __webpack_exports__Qwen2VLImageProcessor as Qwen2VLImageProcessor, __webpack_exports__Qwen2VLPreTrainedModel as Qwen2VLPreTrainedModel, __webpack_exports__Qwen2VLProcessor as Qwen2VLProcessor, __webpack_exports__RFDetrForObjectDetection as RFDetrForObjectDetection, __webpack_exports__RFDetrModel as RFDetrModel, __webpack_exports__RFDetrObjectDetectionOutput as RFDetrObjectDetectionOutput, __webpack_exports__RFDetrPreTrainedModel as RFDetrPreTrainedModel, __webpack_exports__RTDetrForObjectDetection as RTDetrForObjectDetection, __webpack_exports__RTDetrImageProcessor as RTDetrImageProcessor, __webpack_exports__RTDetrModel as RTDetrModel, __webpack_exports__RTDetrObjectDetectionOutput as RTDetrObjectDetectionOutput, __webpack_exports__RTDetrPreTrainedModel as RTDetrPreTrainedModel, __webpack_exports__RTDetrV2ForObjectDetection as RTDetrV2ForObjectDetection, __webpack_exports__RTDetrV2Model as RTDetrV2Model, __webpack_exports__RTDetrV2ObjectDetectionOutput as RTDetrV2ObjectDetectionOutput, __webpack_exports__RTDetrV2PreTrainedModel as RTDetrV2PreTrainedModel, __webpack_exports__RawAudio as RawAudio, __webpack_exports__RawImage as RawImage, __webpack_exports__RawVideo as RawVideo, __webpack_exports__RawVideoFrame as RawVideoFrame, __webpack_exports__RepetitionPenaltyLogitsProcessor as RepetitionPenaltyLogitsProcessor, __webpack_exports__ResNetForImageClassification as ResNetForImageClassification, __webpack_exports__ResNetModel as ResNetModel, __webpack_exports__ResNetPreTrainedModel as ResNetPreTrainedModel, __webpack_exports__RoFormerForMaskedLM as RoFormerForMaskedLM, __webpack_exports__RoFormerForQuestionAnswering as RoFormerForQuestionAnswering, __webpack_exports__RoFormerForSequenceClassification as RoFormerForSequenceClassification, __webpack_exports__RoFormerForTokenClassification as RoFormerForTokenClassification, __webpack_exports__RoFormerModel as RoFormerModel, __webpack_exports__RoFormerPreTrainedModel as RoFormerPreTrainedModel, __webpack_exports__RoFormerTokenizer as RoFormerTokenizer, __webpack_exports__RobertaForMaskedLM as RobertaForMaskedLM, __webpack_exports__RobertaForQuestionAnswering as RobertaForQuestionAnswering, __webpack_exports__RobertaForSequenceClassification as RobertaForSequenceClassification, __webpack_exports__RobertaForTokenClassification as RobertaForTokenClassification, __webpack_exports__RobertaModel as RobertaModel, __webpack_exports__RobertaPreTrainedModel as RobertaPreTrainedModel, __webpack_exports__RobertaTokenizer as RobertaTokenizer, __webpack_exports__SamImageProcessor as SamImageProcessor, __webpack_exports__SamImageSegmentationOutput as SamImageSegmentationOutput, __webpack_exports__SamModel as SamModel, __webpack_exports__SamPreTrainedModel as SamPreTrainedModel, __webpack_exports__SamProcessor as SamProcessor, __webpack_exports__SapiensForDepthEstimation as SapiensForDepthEstimation, __webpack_exports__SapiensForNormalEstimation as SapiensForNormalEstimation, __webpack_exports__SapiensForSemanticSegmentation as SapiensForSemanticSegmentation, __webpack_exports__SapiensPreTrainedModel as SapiensPreTrainedModel, __webpack_exports__SeamlessM4TFeatureExtractor as SeamlessM4TFeatureExtractor, __webpack_exports__SegformerFeatureExtractor as SegformerFeatureExtractor, __webpack_exports__SegformerForImageClassification as SegformerForImageClassification, __webpack_exports__SegformerForSemanticSegmentation as SegformerForSemanticSegmentation, __webpack_exports__SegformerImageProcessor as SegformerImageProcessor, __webpack_exports__SegformerModel as SegformerModel, __webpack_exports__SegformerPreTrainedModel as SegformerPreTrainedModel, __webpack_exports__Seq2SeqLMOutput as Seq2SeqLMOutput, __webpack_exports__SequenceClassifierOutput as SequenceClassifierOutput, __webpack_exports__SiglipImageProcessor as SiglipImageProcessor, __webpack_exports__SiglipModel as SiglipModel, __webpack_exports__SiglipPreTrainedModel as SiglipPreTrainedModel, __webpack_exports__SiglipTextModel as SiglipTextModel, __webpack_exports__SiglipTokenizer as SiglipTokenizer, __webpack_exports__SiglipVisionModel as SiglipVisionModel, __webpack_exports__SmolVLMForConditionalGeneration as SmolVLMForConditionalGeneration, __webpack_exports__SmolVLMImageProcessor as SmolVLMImageProcessor, __webpack_exports__SmolVLMProcessor as SmolVLMProcessor, __webpack_exports__SnacDecoderModel as SnacDecoderModel, __webpack_exports__SnacEncoderModel as SnacEncoderModel, __webpack_exports__SnacFeatureExtractor as SnacFeatureExtractor, __webpack_exports__SnacModel as SnacModel, __webpack_exports__SnacPreTrainedModel as SnacPreTrainedModel, __webpack_exports__SpeechT5FeatureExtractor as SpeechT5FeatureExtractor, __webpack_exports__SpeechT5ForSpeechToText as SpeechT5ForSpeechToText, __webpack_exports__SpeechT5ForTextToSpeech as SpeechT5ForTextToSpeech, __webpack_exports__SpeechT5HifiGan as SpeechT5HifiGan, __webpack_exports__SpeechT5Model as SpeechT5Model, __webpack_exports__SpeechT5PreTrainedModel as SpeechT5PreTrainedModel, __webpack_exports__SpeechT5Processor as SpeechT5Processor, __webpack_exports__SpeechT5Tokenizer as SpeechT5Tokenizer, __webpack_exports__SqueezeBertForMaskedLM as SqueezeBertForMaskedLM, __webpack_exports__SqueezeBertForQuestionAnswering as SqueezeBertForQuestionAnswering, __webpack_exports__SqueezeBertForSequenceClassification as SqueezeBertForSequenceClassification, __webpack_exports__SqueezeBertModel as SqueezeBertModel, __webpack_exports__SqueezeBertPreTrainedModel as SqueezeBertPreTrainedModel, __webpack_exports__SqueezeBertTokenizer as SqueezeBertTokenizer, __webpack_exports__StableLmForCausalLM as StableLmForCausalLM, __webpack_exports__StableLmModel as StableLmModel, __webpack_exports__StableLmPreTrainedModel as StableLmPreTrainedModel, __webpack_exports__Starcoder2ForCausalLM as Starcoder2ForCausalLM, __webpack_exports__Starcoder2Model as Starcoder2Model, __webpack_exports__Starcoder2PreTrainedModel as Starcoder2PreTrainedModel, __webpack_exports__StoppingCriteria as StoppingCriteria, __webpack_exports__StoppingCriteriaList as StoppingCriteriaList, __webpack_exports__StyleTextToSpeech2Model as StyleTextToSpeech2Model, __webpack_exports__StyleTextToSpeech2PreTrainedModel as StyleTextToSpeech2PreTrainedModel, __webpack_exports__SummarizationPipeline as SummarizationPipeline, __webpack_exports__SuppressTokensAtBeginLogitsProcessor as SuppressTokensAtBeginLogitsProcessor, __webpack_exports__Swin2SRForImageSuperResolution as Swin2SRForImageSuperResolution, __webpack_exports__Swin2SRImageProcessor as Swin2SRImageProcessor, __webpack_exports__Swin2SRModel as Swin2SRModel, __webpack_exports__Swin2SRPreTrainedModel as Swin2SRPreTrainedModel, __webpack_exports__SwinForImageClassification as SwinForImageClassification, __webpack_exports__SwinForSemanticSegmentation as SwinForSemanticSegmentation, __webpack_exports__SwinModel as SwinModel, __webpack_exports__SwinPreTrainedModel as SwinPreTrainedModel, __webpack_exports__T5ForConditionalGeneration as T5ForConditionalGeneration, __webpack_exports__T5Model as T5Model, __webpack_exports__T5PreTrainedModel as T5PreTrainedModel, __webpack_exports__T5Tokenizer as T5Tokenizer, __webpack_exports__TableTransformerForObjectDetection as TableTransformerForObjectDetection, __webpack_exports__TableTransformerModel as TableTransformerModel, __webpack_exports__TableTransformerObjectDetectionOutput as TableTransformerObjectDetectionOutput, __webpack_exports__TableTransformerPreTrainedModel as TableTransformerPreTrainedModel, __webpack_exports__TemperatureLogitsWarper as TemperatureLogitsWarper, __webpack_exports__Tensor as Tensor, __webpack_exports__Text2TextGenerationPipeline as Text2TextGenerationPipeline, __webpack_exports__TextClassificationPipeline as TextClassificationPipeline, __webpack_exports__TextGenerationPipeline as TextGenerationPipeline, __webpack_exports__TextStreamer as TextStreamer, __webpack_exports__TextToAudioPipeline as TextToAudioPipeline, __webpack_exports__TokenClassificationPipeline as TokenClassificationPipeline, __webpack_exports__TokenClassifierOutput as TokenClassifierOutput, __webpack_exports__TokenizerModel as TokenizerModel, __webpack_exports__TopKLogitsWarper as TopKLogitsWarper, __webpack_exports__TopPLogitsWarper as TopPLogitsWarper, __webpack_exports__TrOCRForCausalLM as TrOCRForCausalLM, __webpack_exports__TrOCRPreTrainedModel as TrOCRPreTrainedModel, __webpack_exports__TranslationPipeline as TranslationPipeline, __webpack_exports__UltravoxModel as UltravoxModel, __webpack_exports__UltravoxPreTrainedModel as UltravoxPreTrainedModel, __webpack_exports__UltravoxProcessor as UltravoxProcessor, __webpack_exports__UniSpeechForCTC as UniSpeechForCTC, __webpack_exports__UniSpeechForSequenceClassification as UniSpeechForSequenceClassification, __webpack_exports__UniSpeechModel as UniSpeechModel, __webpack_exports__UniSpeechPreTrainedModel as UniSpeechPreTrainedModel, __webpack_exports__UniSpeechSatForAudioFrameClassification as UniSpeechSatForAudioFrameClassification, __webpack_exports__UniSpeechSatForCTC as UniSpeechSatForCTC, __webpack_exports__UniSpeechSatForSequenceClassification as UniSpeechSatForSequenceClassification, __webpack_exports__UniSpeechSatModel as UniSpeechSatModel, __webpack_exports__UniSpeechSatPreTrainedModel as UniSpeechSatPreTrainedModel, __webpack_exports__VLChatProcessor as VLChatProcessor, __webpack_exports__VLMImageProcessor as VLMImageProcessor, __webpack_exports__ViTFeatureExtractor as ViTFeatureExtractor, __webpack_exports__ViTForImageClassification as ViTForImageClassification, __webpack_exports__ViTImageProcessor as ViTImageProcessor, __webpack_exports__ViTMAEModel as ViTMAEModel, __webpack_exports__ViTMAEPreTrainedModel as ViTMAEPreTrainedModel, __webpack_exports__ViTMSNForImageClassification as ViTMSNForImageClassification, __webpack_exports__ViTMSNModel as ViTMSNModel, __webpack_exports__ViTMSNPreTrainedModel as ViTMSNPreTrainedModel, __webpack_exports__ViTModel as ViTModel, __webpack_exports__ViTPreTrainedModel as ViTPreTrainedModel, __webpack_exports__VisionEncoderDecoderModel as VisionEncoderDecoderModel, __webpack_exports__VitMatteForImageMatting as VitMatteForImageMatting, __webpack_exports__VitMatteImageProcessor as VitMatteImageProcessor, __webpack_exports__VitMattePreTrainedModel as VitMattePreTrainedModel, __webpack_exports__VitPoseForPoseEstimation as VitPoseForPoseEstimation, __webpack_exports__VitPoseImageProcessor as VitPoseImageProcessor, __webpack_exports__VitPosePreTrainedModel as VitPosePreTrainedModel, __webpack_exports__VitsModel as VitsModel, __webpack_exports__VitsModelOutput as VitsModelOutput, __webpack_exports__VitsPreTrainedModel as VitsPreTrainedModel, __webpack_exports__VitsTokenizer as VitsTokenizer, __webpack_exports__Wav2Vec2BertForCTC as Wav2Vec2BertForCTC, __webpack_exports__Wav2Vec2BertForSequenceClassification as Wav2Vec2BertForSequenceClassification, __webpack_exports__Wav2Vec2BertModel as Wav2Vec2BertModel, __webpack_exports__Wav2Vec2BertPreTrainedModel as Wav2Vec2BertPreTrainedModel, __webpack_exports__Wav2Vec2CTCTokenizer as Wav2Vec2CTCTokenizer, __webpack_exports__Wav2Vec2FeatureExtractor as Wav2Vec2FeatureExtractor, __webpack_exports__Wav2Vec2ForAudioFrameClassification as Wav2Vec2ForAudioFrameClassification, __webpack_exports__Wav2Vec2ForCTC as Wav2Vec2ForCTC, __webpack_exports__Wav2Vec2ForSequenceClassification as Wav2Vec2ForSequenceClassification, __webpack_exports__Wav2Vec2Model as Wav2Vec2Model, __webpack_exports__Wav2Vec2PreTrainedModel as Wav2Vec2PreTrainedModel, __webpack_exports__Wav2Vec2Processor as Wav2Vec2Processor, __webpack_exports__Wav2Vec2ProcessorWithLM as Wav2Vec2ProcessorWithLM, __webpack_exports__WavLMForAudioFrameClassification as WavLMForAudioFrameClassification, __webpack_exports__WavLMForCTC as WavLMForCTC, __webpack_exports__WavLMForSequenceClassification as WavLMForSequenceClassification, __webpack_exports__WavLMForXVector as WavLMForXVector, __webpack_exports__WavLMModel as WavLMModel, __webpack_exports__WavLMPreTrainedModel as WavLMPreTrainedModel, __webpack_exports__WeSpeakerFeatureExtractor as WeSpeakerFeatureExtractor, __webpack_exports__WeSpeakerResNetModel as WeSpeakerResNetModel, __webpack_exports__WeSpeakerResNetPreTrainedModel as WeSpeakerResNetPreTrainedModel, __webpack_exports__WhisperFeatureExtractor as WhisperFeatureExtractor, __webpack_exports__WhisperForConditionalGeneration as WhisperForConditionalGeneration, __webpack_exports__WhisperModel as WhisperModel, __webpack_exports__WhisperPreTrainedModel as WhisperPreTrainedModel, __webpack_exports__WhisperProcessor as WhisperProcessor, __webpack_exports__WhisperTextStreamer as WhisperTextStreamer, __webpack_exports__WhisperTimeStampLogitsProcessor as WhisperTimeStampLogitsProcessor, __webpack_exports__WhisperTokenizer as WhisperTokenizer, __webpack_exports__XLMForQuestionAnswering as XLMForQuestionAnswering, __webpack_exports__XLMForSequenceClassification as XLMForSequenceClassification, __webpack_exports__XLMForTokenClassification as XLMForTokenClassification, __webpack_exports__XLMModel as XLMModel, __webpack_exports__XLMPreTrainedModel as XLMPreTrainedModel, __webpack_exports__XLMRobertaForMaskedLM as XLMRobertaForMaskedLM, __webpack_exports__XLMRobertaForQuestionAnswering as XLMRobertaForQuestionAnswering, __webpack_exports__XLMRobertaForSequenceClassification as XLMRobertaForSequenceClassification, __webpack_exports__XLMRobertaForTokenClassification as XLMRobertaForTokenClassification, __webpack_exports__XLMRobertaModel as XLMRobertaModel, __webpack_exports__XLMRobertaPreTrainedModel as XLMRobertaPreTrainedModel, __webpack_exports__XLMRobertaTokenizer as XLMRobertaTokenizer, __webpack_exports__XLMTokenizer as XLMTokenizer, __webpack_exports__XLMWithLMHeadModel as XLMWithLMHeadModel, __webpack_exports__XVectorOutput as XVectorOutput, __webpack_exports__YolosFeatureExtractor as YolosFeatureExtractor, __webpack_exports__YolosForObjectDetection as YolosForObjectDetection, __webpack_exports__YolosImageProcessor as YolosImageProcessor, __webpack_exports__YolosModel as YolosModel, __webpack_exports__YolosObjectDetectionOutput as YolosObjectDetectionOutput, __webpack_exports__YolosPreTrainedModel as YolosPreTrainedModel, __webpack_exports__ZeroShotAudioClassificationPipeline as ZeroShotAudioClassificationPipeline, __webpack_exports__ZeroShotClassificationPipeline as ZeroShotClassificationPipeline, __webpack_exports__ZeroShotImageClassificationPipeline as ZeroShotImageClassificationPipeline, __webpack_exports__ZeroShotObjectDetectionPipeline as ZeroShotObjectDetectionPipeline, __webpack_exports__bankers_round as bankers_round, __webpack_exports__cat as cat, __webpack_exports__cos_sim as cos_sim, __webpack_exports__dot as dot, __webpack_exports__dynamic_time_warping as dynamic_time_warping, __webpack_exports__env as env, __webpack_exports__full as full, __webpack_exports__full_like as full_like, __webpack_exports__getKeyValueShapes as getKeyValueShapes, __webpack_exports__hamming as hamming, __webpack_exports__hanning as hanning, __webpack_exports__interpolate as interpolate, __webpack_exports__interpolate_4d as interpolate_4d, __webpack_exports__interpolate_data as interpolate_data, __webpack_exports__is_chinese_char as is_chinese_char, __webpack_exports__layer_norm as layer_norm, __webpack_exports__load_image as load_image, __webpack_exports__load_video as load_video, __webpack_exports__log_softmax as log_softmax, __webpack_exports__magnitude as magnitude, __webpack_exports__matmul as matmul, __webpack_exports__max as max, __webpack_exports__mean as mean, __webpack_exports__mean_pooling as mean_pooling, __webpack_exports__medianFilter as medianFilter, __webpack_exports__mel_filter_bank as mel_filter_bank, __webpack_exports__min as min, __webpack_exports__ones as ones, __webpack_exports__ones_like as ones_like, __webpack_exports__permute as permute, __webpack_exports__permute_data as permute_data, __webpack_exports__pipeline as pipeline, __webpack_exports__quantize_embeddings as quantize_embeddings, __webpack_exports__rand as rand, __webpack_exports__read_audio as read_audio, __webpack_exports__rfft as rfft, __webpack_exports__round as round, __webpack_exports__slice as slice, __webpack_exports__softmax as softmax, __webpack_exports__spectrogram as spectrogram, __webpack_exports__stack as stack, __webpack_exports__std_mean as std_mean, __webpack_exports__topk as topk, __webpack_exports__window_function as window_function, __webpack_exports__zeros as zeros, __webpack_exports__zeros_like as zeros_like };
|
|
35696
36023
|
|
|
35697
36024
|
//# sourceMappingURL=transformers.node.mjs.map
|