@huggingface/transformers 3.4.0 → 3.4.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +8 -2
- package/dist/transformers.js +528 -201
- package/dist/transformers.js.map +1 -1
- package/dist/transformers.min.js +1 -1
- package/dist/transformers.min.js.map +1 -1
- package/dist/transformers.node.cjs +508 -200
- package/dist/transformers.node.cjs.map +1 -1
- package/dist/transformers.node.min.cjs +1 -1
- package/dist/transformers.node.min.cjs.map +1 -1
- package/dist/transformers.node.min.mjs +1 -1
- package/dist/transformers.node.min.mjs.map +1 -1
- package/dist/transformers.node.mjs +528 -201
- package/dist/transformers.node.mjs.map +1 -1
- package/dist/transformers.web.js +528 -201
- package/dist/transformers.web.js.map +1 -1
- package/dist/transformers.web.min.js +1 -1
- package/dist/transformers.web.min.js.map +1 -1
- package/package.json +1 -1
- package/src/configs.js +2 -0
- package/src/env.js +1 -1
- package/src/models/feature_extractors.js +1 -0
- package/src/models/snac/feature_extraction_snac.js +3 -0
- package/src/models.js +125 -2
- package/src/pipelines.js +140 -135
- package/src/tokenizers.js +44 -34
- package/src/utils/data-structures.js +74 -0
- package/src/utils/hub.js +36 -15
- package/src/utils/image.js +9 -1
- package/src/utils/tensor.js +6 -2
- package/types/configs.d.ts.map +1 -1
- package/types/models/feature_extractors.d.ts +1 -0
- package/types/models/snac/feature_extraction_snac.d.ts +4 -0
- package/types/models/snac/feature_extraction_snac.d.ts.map +1 -0
- package/types/models.d.ts +72 -0
- package/types/models.d.ts.map +1 -1
- package/types/pipelines.d.ts +2 -2
- package/types/pipelines.d.ts.map +1 -1
- package/types/tokenizers.d.ts +4 -1
- package/types/tokenizers.d.ts.map +1 -1
- package/types/tsconfig.tsbuildinfo +1 -1
- package/types/utils/data-structures.d.ts +26 -0
- package/types/utils/data-structures.d.ts.map +1 -1
- package/types/utils/hub.d.ts.map +1 -1
- package/types/utils/image.d.ts +2 -2
- package/types/utils/image.d.ts.map +1 -1
- package/types/utils/tensor.d.ts.map +1 -1
package/dist/transformers.js
CHANGED
|
@@ -8496,6 +8496,7 @@ function getNormalizedConfig(config) {
|
|
|
8496
8496
|
// Sub-configs
|
|
8497
8497
|
case 'llava':
|
|
8498
8498
|
case 'paligemma':
|
|
8499
|
+
case 'gemma3':
|
|
8499
8500
|
case 'florence2':
|
|
8500
8501
|
case 'llava_onevision':
|
|
8501
8502
|
case 'idefics3':
|
|
@@ -8555,6 +8556,7 @@ function getNormalizedConfig(config) {
|
|
|
8555
8556
|
break;
|
|
8556
8557
|
case 'gemma':
|
|
8557
8558
|
case 'gemma2':
|
|
8559
|
+
case 'gemma3_text':
|
|
8558
8560
|
case 'glm':
|
|
8559
8561
|
case 'helium':
|
|
8560
8562
|
mapping['num_heads'] = 'num_key_value_heads';
|
|
@@ -8885,7 +8887,7 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
8885
8887
|
|
|
8886
8888
|
|
|
8887
8889
|
|
|
8888
|
-
const VERSION = '3.4.
|
|
8890
|
+
const VERSION = '3.4.2';
|
|
8889
8891
|
|
|
8890
8892
|
// Check if various APIs are available (depends on environment)
|
|
8891
8893
|
const IS_BROWSER_ENV = typeof window !== "undefined" && typeof window.document !== "undefined";
|
|
@@ -11023,6 +11025,9 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
11023
11025
|
/* harmony export */ Gemma2ForCausalLM: () => (/* binding */ Gemma2ForCausalLM),
|
|
11024
11026
|
/* harmony export */ Gemma2Model: () => (/* binding */ Gemma2Model),
|
|
11025
11027
|
/* harmony export */ Gemma2PreTrainedModel: () => (/* binding */ Gemma2PreTrainedModel),
|
|
11028
|
+
/* harmony export */ Gemma3ForCausalLM: () => (/* binding */ Gemma3ForCausalLM),
|
|
11029
|
+
/* harmony export */ Gemma3Model: () => (/* binding */ Gemma3Model),
|
|
11030
|
+
/* harmony export */ Gemma3PreTrainedModel: () => (/* binding */ Gemma3PreTrainedModel),
|
|
11026
11031
|
/* harmony export */ GemmaForCausalLM: () => (/* binding */ GemmaForCausalLM),
|
|
11027
11032
|
/* harmony export */ GemmaModel: () => (/* binding */ GemmaModel),
|
|
11028
11033
|
/* harmony export */ GemmaPreTrainedModel: () => (/* binding */ GemmaPreTrainedModel),
|
|
@@ -11093,6 +11098,10 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
11093
11098
|
/* harmony export */ MaskFormerModel: () => (/* binding */ MaskFormerModel),
|
|
11094
11099
|
/* harmony export */ MaskFormerPreTrainedModel: () => (/* binding */ MaskFormerPreTrainedModel),
|
|
11095
11100
|
/* harmony export */ MaskedLMOutput: () => (/* binding */ MaskedLMOutput),
|
|
11101
|
+
/* harmony export */ Metric3DForDepthEstimation: () => (/* binding */ Metric3DForDepthEstimation),
|
|
11102
|
+
/* harmony export */ Metric3DPreTrainedModel: () => (/* binding */ Metric3DPreTrainedModel),
|
|
11103
|
+
/* harmony export */ Metric3Dv2ForDepthEstimation: () => (/* binding */ Metric3Dv2ForDepthEstimation),
|
|
11104
|
+
/* harmony export */ Metric3Dv2PreTrainedModel: () => (/* binding */ Metric3Dv2PreTrainedModel),
|
|
11096
11105
|
/* harmony export */ MgpstrForSceneTextRecognition: () => (/* binding */ MgpstrForSceneTextRecognition),
|
|
11097
11106
|
/* harmony export */ MgpstrModelOutput: () => (/* binding */ MgpstrModelOutput),
|
|
11098
11107
|
/* harmony export */ MgpstrPreTrainedModel: () => (/* binding */ MgpstrPreTrainedModel),
|
|
@@ -11204,10 +11213,18 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
11204
11213
|
/* harmony export */ Qwen2PreTrainedModel: () => (/* binding */ Qwen2PreTrainedModel),
|
|
11205
11214
|
/* harmony export */ Qwen2VLForConditionalGeneration: () => (/* binding */ Qwen2VLForConditionalGeneration),
|
|
11206
11215
|
/* harmony export */ Qwen2VLPreTrainedModel: () => (/* binding */ Qwen2VLPreTrainedModel),
|
|
11216
|
+
/* harmony export */ RFDetrForObjectDetection: () => (/* binding */ RFDetrForObjectDetection),
|
|
11217
|
+
/* harmony export */ RFDetrModel: () => (/* binding */ RFDetrModel),
|
|
11218
|
+
/* harmony export */ RFDetrObjectDetectionOutput: () => (/* binding */ RFDetrObjectDetectionOutput),
|
|
11219
|
+
/* harmony export */ RFDetrPreTrainedModel: () => (/* binding */ RFDetrPreTrainedModel),
|
|
11207
11220
|
/* harmony export */ RTDetrForObjectDetection: () => (/* binding */ RTDetrForObjectDetection),
|
|
11208
11221
|
/* harmony export */ RTDetrModel: () => (/* binding */ RTDetrModel),
|
|
11209
11222
|
/* harmony export */ RTDetrObjectDetectionOutput: () => (/* binding */ RTDetrObjectDetectionOutput),
|
|
11210
11223
|
/* harmony export */ RTDetrPreTrainedModel: () => (/* binding */ RTDetrPreTrainedModel),
|
|
11224
|
+
/* harmony export */ RTDetrV2ForObjectDetection: () => (/* binding */ RTDetrV2ForObjectDetection),
|
|
11225
|
+
/* harmony export */ RTDetrV2Model: () => (/* binding */ RTDetrV2Model),
|
|
11226
|
+
/* harmony export */ RTDetrV2ObjectDetectionOutput: () => (/* binding */ RTDetrV2ObjectDetectionOutput),
|
|
11227
|
+
/* harmony export */ RTDetrV2PreTrainedModel: () => (/* binding */ RTDetrV2PreTrainedModel),
|
|
11211
11228
|
/* harmony export */ ResNetForImageClassification: () => (/* binding */ ResNetForImageClassification),
|
|
11212
11229
|
/* harmony export */ ResNetModel: () => (/* binding */ ResNetModel),
|
|
11213
11230
|
/* harmony export */ ResNetPreTrainedModel: () => (/* binding */ ResNetPreTrainedModel),
|
|
@@ -11241,6 +11258,10 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
11241
11258
|
/* harmony export */ SiglipTextModel: () => (/* binding */ SiglipTextModel),
|
|
11242
11259
|
/* harmony export */ SiglipVisionModel: () => (/* binding */ SiglipVisionModel),
|
|
11243
11260
|
/* harmony export */ SmolVLMForConditionalGeneration: () => (/* binding */ SmolVLMForConditionalGeneration),
|
|
11261
|
+
/* harmony export */ SnacDecoderModel: () => (/* binding */ SnacDecoderModel),
|
|
11262
|
+
/* harmony export */ SnacEncoderModel: () => (/* binding */ SnacEncoderModel),
|
|
11263
|
+
/* harmony export */ SnacModel: () => (/* binding */ SnacModel),
|
|
11264
|
+
/* harmony export */ SnacPreTrainedModel: () => (/* binding */ SnacPreTrainedModel),
|
|
11244
11265
|
/* harmony export */ SpeechT5ForSpeechToText: () => (/* binding */ SpeechT5ForSpeechToText),
|
|
11245
11266
|
/* harmony export */ SpeechT5ForTextToSpeech: () => (/* binding */ SpeechT5ForTextToSpeech),
|
|
11246
11267
|
/* harmony export */ SpeechT5HifiGan: () => (/* binding */ SpeechT5HifiGan),
|
|
@@ -11899,8 +11920,8 @@ async function decoderForward(self, model_inputs, is_encoder_decoder = false) {
|
|
|
11899
11920
|
new_model_inputs.use_cache_branch = boolTensor(!!past_key_values);
|
|
11900
11921
|
}
|
|
11901
11922
|
if (session.inputNames.includes('position_ids') && new_model_inputs.attention_mask && !new_model_inputs.position_ids) {
|
|
11902
|
-
// NOTE: Handle a special case for paligemma models, where positions are 1-indexed
|
|
11903
|
-
const start_index = self.config.model_type
|
|
11923
|
+
// NOTE: Handle a special case for paligemma/gemma3 models, where positions are 1-indexed
|
|
11924
|
+
const start_index = ['paligemma', 'gemma3_text', 'gemma3'].includes(self.config.model_type) ? 1 : 0;
|
|
11904
11925
|
new_model_inputs.position_ids = createPositionIds(new_model_inputs, past_key_values, start_index);
|
|
11905
11926
|
}
|
|
11906
11927
|
|
|
@@ -15825,6 +15846,23 @@ class Gemma2Model extends Gemma2PreTrainedModel { }
|
|
|
15825
15846
|
class Gemma2ForCausalLM extends Gemma2PreTrainedModel { }
|
|
15826
15847
|
//////////////////////////////////////////////////
|
|
15827
15848
|
|
|
15849
|
+
|
|
15850
|
+
//////////////////////////////////////////////////
|
|
15851
|
+
// Gemma3 models
|
|
15852
|
+
|
|
15853
|
+
/**
|
|
15854
|
+
* The bare Gemma3 Model outputting raw hidden-states without any specific head on top.
|
|
15855
|
+
*/
|
|
15856
|
+
class Gemma3PreTrainedModel extends PreTrainedModel { }
|
|
15857
|
+
/**
|
|
15858
|
+
* The bare Gemma3 Model outputting raw hidden-states without any specific head on top.
|
|
15859
|
+
*/
|
|
15860
|
+
class Gemma3Model extends Gemma3PreTrainedModel { }
|
|
15861
|
+
|
|
15862
|
+
class Gemma3ForCausalLM extends Gemma3PreTrainedModel { }
|
|
15863
|
+
//////////////////////////////////////////////////
|
|
15864
|
+
|
|
15865
|
+
|
|
15828
15866
|
//////////////////////////////////////////////////
|
|
15829
15867
|
class OpenELMPreTrainedModel extends PreTrainedModel { }
|
|
15830
15868
|
class OpenELMModel extends OpenELMPreTrainedModel { }
|
|
@@ -16469,6 +16507,37 @@ class RTDetrObjectDetectionOutput extends ModelOutput {
|
|
|
16469
16507
|
}
|
|
16470
16508
|
//////////////////////////////////////////////////
|
|
16471
16509
|
|
|
16510
|
+
|
|
16511
|
+
//////////////////////////////////////////////////
|
|
16512
|
+
class RTDetrV2PreTrainedModel extends PreTrainedModel { }
|
|
16513
|
+
class RTDetrV2Model extends RTDetrV2PreTrainedModel { }
|
|
16514
|
+
class RTDetrV2ForObjectDetection extends RTDetrV2PreTrainedModel {
|
|
16515
|
+
/**
|
|
16516
|
+
* @param {any} model_inputs
|
|
16517
|
+
*/
|
|
16518
|
+
async _call(model_inputs) {
|
|
16519
|
+
return new RTDetrV2ObjectDetectionOutput(await super._call(model_inputs));
|
|
16520
|
+
}
|
|
16521
|
+
}
|
|
16522
|
+
|
|
16523
|
+
class RTDetrV2ObjectDetectionOutput extends RTDetrObjectDetectionOutput {}
|
|
16524
|
+
//////////////////////////////////////////////////
|
|
16525
|
+
|
|
16526
|
+
//////////////////////////////////////////////////
|
|
16527
|
+
class RFDetrPreTrainedModel extends PreTrainedModel { }
|
|
16528
|
+
class RFDetrModel extends RFDetrPreTrainedModel { }
|
|
16529
|
+
class RFDetrForObjectDetection extends RFDetrPreTrainedModel {
|
|
16530
|
+
/**
|
|
16531
|
+
* @param {any} model_inputs
|
|
16532
|
+
*/
|
|
16533
|
+
async _call(model_inputs) {
|
|
16534
|
+
return new RFDetrObjectDetectionOutput(await super._call(model_inputs));
|
|
16535
|
+
}
|
|
16536
|
+
}
|
|
16537
|
+
|
|
16538
|
+
class RFDetrObjectDetectionOutput extends RTDetrObjectDetectionOutput {}
|
|
16539
|
+
//////////////////////////////////////////////////
|
|
16540
|
+
|
|
16472
16541
|
//////////////////////////////////////////////////
|
|
16473
16542
|
class TableTransformerPreTrainedModel extends PreTrainedModel { }
|
|
16474
16543
|
|
|
@@ -16677,6 +16746,16 @@ class DepthProPreTrainedModel extends PreTrainedModel { }
|
|
|
16677
16746
|
class DepthProForDepthEstimation extends DepthProPreTrainedModel { }
|
|
16678
16747
|
//////////////////////////////////////////////////
|
|
16679
16748
|
|
|
16749
|
+
//////////////////////////////////////////////////
|
|
16750
|
+
class Metric3DPreTrainedModel extends PreTrainedModel { }
|
|
16751
|
+
class Metric3DForDepthEstimation extends Metric3DPreTrainedModel { }
|
|
16752
|
+
//////////////////////////////////////////////////
|
|
16753
|
+
|
|
16754
|
+
//////////////////////////////////////////////////
|
|
16755
|
+
class Metric3Dv2PreTrainedModel extends PreTrainedModel { }
|
|
16756
|
+
class Metric3Dv2ForDepthEstimation extends Metric3Dv2PreTrainedModel { }
|
|
16757
|
+
//////////////////////////////////////////////////
|
|
16758
|
+
|
|
16680
16759
|
//////////////////////////////////////////////////
|
|
16681
16760
|
class MaskFormerPreTrainedModel extends PreTrainedModel { }
|
|
16682
16761
|
class MaskFormerModel extends MaskFormerPreTrainedModel { }
|
|
@@ -18592,6 +18671,60 @@ class DacDecoderModel extends DacPreTrainedModel {
|
|
|
18592
18671
|
}
|
|
18593
18672
|
//////////////////////////////////////////////////
|
|
18594
18673
|
|
|
18674
|
+
|
|
18675
|
+
//////////////////////////////////////////////////
|
|
18676
|
+
// Snac models
|
|
18677
|
+
class SnacPreTrainedModel extends PreTrainedModel {
|
|
18678
|
+
main_input_name = 'input_values';
|
|
18679
|
+
forward_params = ['input_values'];
|
|
18680
|
+
}
|
|
18681
|
+
|
|
18682
|
+
/**
|
|
18683
|
+
* The SNAC (Multi-Scale Neural Audio Codec) model.
|
|
18684
|
+
*/
|
|
18685
|
+
class SnacModel extends SnacPreTrainedModel {
|
|
18686
|
+
/**
|
|
18687
|
+
* Encodes the input audio waveform into discrete codes.
|
|
18688
|
+
* @param {Object} inputs Model inputs
|
|
18689
|
+
* @param {Tensor} [inputs.input_values] Float values of the input audio waveform, of shape `(batch_size, channels, sequence_length)`).
|
|
18690
|
+
* @returns {Promise<Record<string, Tensor>>} The output tensors of shape `(batch_size, num_codebooks, sequence_length)`.
|
|
18691
|
+
*/
|
|
18692
|
+
async encode(inputs) {
|
|
18693
|
+
return await sessionRun(this.sessions['encoder_model'], inputs);
|
|
18694
|
+
}
|
|
18695
|
+
|
|
18696
|
+
/**
|
|
18697
|
+
* Decodes the given frames into an output audio waveform.
|
|
18698
|
+
* @param {Record<string, Tensor>} inputs The encoded audio codes.
|
|
18699
|
+
* @returns {Promise<{audio_values: Tensor}>} The output tensor of shape `(batch_size, num_channels, sequence_length)`.
|
|
18700
|
+
*/
|
|
18701
|
+
async decode(inputs) {
|
|
18702
|
+
return await sessionRun(this.sessions['decoder_model'], inputs);
|
|
18703
|
+
}
|
|
18704
|
+
}
|
|
18705
|
+
|
|
18706
|
+
class SnacEncoderModel extends SnacPreTrainedModel {
|
|
18707
|
+
/** @type {typeof PreTrainedModel.from_pretrained} */
|
|
18708
|
+
static async from_pretrained(pretrained_model_name_or_path, options = {}) {
|
|
18709
|
+
return super.from_pretrained(pretrained_model_name_or_path, {
|
|
18710
|
+
...options,
|
|
18711
|
+
// Update default model file name if not provided
|
|
18712
|
+
model_file_name: options.model_file_name ?? 'encoder_model',
|
|
18713
|
+
});
|
|
18714
|
+
}
|
|
18715
|
+
}
|
|
18716
|
+
class SnacDecoderModel extends SnacPreTrainedModel {
|
|
18717
|
+
/** @type {typeof PreTrainedModel.from_pretrained} */
|
|
18718
|
+
static async from_pretrained(pretrained_model_name_or_path, options = {}) {
|
|
18719
|
+
return super.from_pretrained(pretrained_model_name_or_path, {
|
|
18720
|
+
...options,
|
|
18721
|
+
// Update default model file name if not provided
|
|
18722
|
+
model_file_name: options.model_file_name ?? 'decoder_model',
|
|
18723
|
+
});
|
|
18724
|
+
}
|
|
18725
|
+
}
|
|
18726
|
+
//////////////////////////////////////////////////
|
|
18727
|
+
|
|
18595
18728
|
//////////////////////////////////////////////////
|
|
18596
18729
|
// AutoModels, used to simplify construction of PreTrainedModels
|
|
18597
18730
|
// (uses config to instantiate correct class)
|
|
@@ -18712,6 +18845,8 @@ const MODEL_MAPPING_NAMES_ENCODER_ONLY = new Map([
|
|
|
18712
18845
|
|
|
18713
18846
|
['detr', ['DetrModel', DetrModel]],
|
|
18714
18847
|
['rt_detr', ['RTDetrModel', RTDetrModel]],
|
|
18848
|
+
['rt_detr_v2', ['RTDetrV2Model', RTDetrV2Model]],
|
|
18849
|
+
['rf_detr', ['RFDetrModel', RFDetrModel]],
|
|
18715
18850
|
['table-transformer', ['TableTransformerModel', TableTransformerModel]],
|
|
18716
18851
|
['vit', ['ViTModel', ViTModel]],
|
|
18717
18852
|
['ijepa', ['IJepaModel', IJepaModel]],
|
|
@@ -18773,6 +18908,7 @@ const MODEL_MAPPING_NAMES_ENCODER_DECODER = new Map([
|
|
|
18773
18908
|
const MODEL_MAPPING_NAMES_AUTO_ENCODER = new Map([
|
|
18774
18909
|
['mimi', ['MimiModel', MimiModel]],
|
|
18775
18910
|
['dac', ['DacModel', DacModel]],
|
|
18911
|
+
['snac', ['SnacModel', SnacModel]],
|
|
18776
18912
|
]);
|
|
18777
18913
|
|
|
18778
18914
|
const MODEL_MAPPING_NAMES_DECODER_ONLY = new Map([
|
|
@@ -18793,6 +18929,7 @@ const MODEL_MAPPING_NAMES_DECODER_ONLY = new Map([
|
|
|
18793
18929
|
['cohere', ['CohereModel', CohereModel]],
|
|
18794
18930
|
['gemma', ['GemmaModel', GemmaModel]],
|
|
18795
18931
|
['gemma2', ['Gemma2Model', Gemma2Model]],
|
|
18932
|
+
['gemma3_text', ['Gemma3Model', Gemma3Model]],
|
|
18796
18933
|
['helium', ['HeliumModel', HeliumModel]],
|
|
18797
18934
|
['glm', ['GlmModel', GlmModel]],
|
|
18798
18935
|
['openelm', ['OpenELMModel', OpenELMModel]],
|
|
@@ -18892,6 +19029,7 @@ const MODEL_FOR_CAUSAL_LM_MAPPING_NAMES = new Map([
|
|
|
18892
19029
|
['cohere', ['CohereForCausalLM', CohereForCausalLM]],
|
|
18893
19030
|
['gemma', ['GemmaForCausalLM', GemmaForCausalLM]],
|
|
18894
19031
|
['gemma2', ['Gemma2ForCausalLM', Gemma2ForCausalLM]],
|
|
19032
|
+
['gemma3_text', ['Gemma3ForCausalLM', Gemma3ForCausalLM]],
|
|
18895
19033
|
['helium', ['HeliumForCausalLM', HeliumForCausalLM]],
|
|
18896
19034
|
['glm', ['GlmForCausalLM', GlmForCausalLM]],
|
|
18897
19035
|
['openelm', ['OpenELMForCausalLM', OpenELMForCausalLM]],
|
|
@@ -19008,6 +19146,8 @@ const MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES = new Map([
|
|
|
19008
19146
|
const MODEL_FOR_OBJECT_DETECTION_MAPPING_NAMES = new Map([
|
|
19009
19147
|
['detr', ['DetrForObjectDetection', DetrForObjectDetection]],
|
|
19010
19148
|
['rt_detr', ['RTDetrForObjectDetection', RTDetrForObjectDetection]],
|
|
19149
|
+
['rt_detr_v2', ['RTDetrV2ForObjectDetection', RTDetrV2ForObjectDetection]],
|
|
19150
|
+
['rf_detr', ['RFDetrForObjectDetection', RFDetrForObjectDetection]],
|
|
19011
19151
|
['table-transformer', ['TableTransformerForObjectDetection', TableTransformerForObjectDetection]],
|
|
19012
19152
|
['yolos', ['YolosForObjectDetection', YolosForObjectDetection]],
|
|
19013
19153
|
]);
|
|
@@ -19093,6 +19233,8 @@ const MODEL_FOR_DEPTH_ESTIMATION_MAPPING_NAMES = new Map([
|
|
|
19093
19233
|
['glpn', ['GLPNForDepthEstimation', GLPNForDepthEstimation]],
|
|
19094
19234
|
['sapiens', ['SapiensForDepthEstimation', SapiensForDepthEstimation]],
|
|
19095
19235
|
['depth_pro', ['DepthProForDepthEstimation', DepthProForDepthEstimation]],
|
|
19236
|
+
['metric3d', ['Metric3DForDepthEstimation', Metric3DForDepthEstimation]],
|
|
19237
|
+
['metric3dv2', ['Metric3Dv2ForDepthEstimation', Metric3Dv2ForDepthEstimation]],
|
|
19096
19238
|
])
|
|
19097
19239
|
|
|
19098
19240
|
const MODEL_FOR_NORMAL_ESTIMATION_MAPPING_NAMES = new Map([
|
|
@@ -19178,6 +19320,8 @@ const CUSTOM_MAPPING = [
|
|
|
19178
19320
|
['DacDecoderModel', DacDecoderModel, MODEL_TYPES.EncoderOnly],
|
|
19179
19321
|
['MimiEncoderModel', MimiEncoderModel, MODEL_TYPES.EncoderOnly],
|
|
19180
19322
|
['MimiDecoderModel', MimiDecoderModel, MODEL_TYPES.EncoderOnly],
|
|
19323
|
+
['SnacEncoderModel', SnacEncoderModel, MODEL_TYPES.EncoderOnly],
|
|
19324
|
+
['SnacDecoderModel', SnacDecoderModel, MODEL_TYPES.EncoderOnly],
|
|
19181
19325
|
]
|
|
19182
19326
|
for (const [name, model, type] of CUSTOM_MAPPING) {
|
|
19183
19327
|
MODEL_TYPE_MAPPING.set(name, type);
|
|
@@ -20476,14 +20620,15 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
20476
20620
|
/* harmony export */ ClapFeatureExtractor: () => (/* reexport safe */ _clap_feature_extraction_clap_js__WEBPACK_IMPORTED_MODULE_2__.ClapFeatureExtractor),
|
|
20477
20621
|
/* harmony export */ DacFeatureExtractor: () => (/* reexport safe */ _dac_feature_extraction_dac_js__WEBPACK_IMPORTED_MODULE_3__.DacFeatureExtractor),
|
|
20478
20622
|
/* harmony export */ EncodecFeatureExtractor: () => (/* reexport safe */ _encodec_feature_extraction_encodec_js__WEBPACK_IMPORTED_MODULE_1__.EncodecFeatureExtractor),
|
|
20479
|
-
/* harmony export */ ImageFeatureExtractor: () => (/* reexport safe */
|
|
20623
|
+
/* harmony export */ ImageFeatureExtractor: () => (/* reexport safe */ _base_image_processors_utils_js__WEBPACK_IMPORTED_MODULE_12__.ImageProcessor),
|
|
20480
20624
|
/* harmony export */ MoonshineFeatureExtractor: () => (/* reexport safe */ _moonshine_feature_extraction_moonshine_js__WEBPACK_IMPORTED_MODULE_4__.MoonshineFeatureExtractor),
|
|
20481
20625
|
/* harmony export */ PyAnnoteFeatureExtractor: () => (/* reexport safe */ _pyannote_feature_extraction_pyannote_js__WEBPACK_IMPORTED_MODULE_5__.PyAnnoteFeatureExtractor),
|
|
20482
20626
|
/* harmony export */ SeamlessM4TFeatureExtractor: () => (/* reexport safe */ _seamless_m4t_feature_extraction_seamless_m4t_js__WEBPACK_IMPORTED_MODULE_6__.SeamlessM4TFeatureExtractor),
|
|
20483
|
-
/* harmony export */
|
|
20484
|
-
/* harmony export */
|
|
20485
|
-
/* harmony export */
|
|
20486
|
-
/* harmony export */
|
|
20627
|
+
/* harmony export */ SnacFeatureExtractor: () => (/* reexport safe */ _snac_feature_extraction_snac_js__WEBPACK_IMPORTED_MODULE_7__.SnacFeatureExtractor),
|
|
20628
|
+
/* harmony export */ SpeechT5FeatureExtractor: () => (/* reexport safe */ _speecht5_feature_extraction_speecht5_js__WEBPACK_IMPORTED_MODULE_8__.SpeechT5FeatureExtractor),
|
|
20629
|
+
/* harmony export */ Wav2Vec2FeatureExtractor: () => (/* reexport safe */ _wav2vec2_feature_extraction_wav2vec2_js__WEBPACK_IMPORTED_MODULE_9__.Wav2Vec2FeatureExtractor),
|
|
20630
|
+
/* harmony export */ WeSpeakerFeatureExtractor: () => (/* reexport safe */ _wespeaker_feature_extraction_wespeaker_js__WEBPACK_IMPORTED_MODULE_10__.WeSpeakerFeatureExtractor),
|
|
20631
|
+
/* harmony export */ WhisperFeatureExtractor: () => (/* reexport safe */ _whisper_feature_extraction_whisper_js__WEBPACK_IMPORTED_MODULE_11__.WhisperFeatureExtractor)
|
|
20487
20632
|
/* harmony export */ });
|
|
20488
20633
|
/* harmony import */ var _audio_spectrogram_transformer_feature_extraction_audio_spectrogram_transformer_js__WEBPACK_IMPORTED_MODULE_0__ = __webpack_require__(/*! ./audio_spectrogram_transformer/feature_extraction_audio_spectrogram_transformer.js */ "./src/models/audio_spectrogram_transformer/feature_extraction_audio_spectrogram_transformer.js");
|
|
20489
20634
|
/* harmony import */ var _encodec_feature_extraction_encodec_js__WEBPACK_IMPORTED_MODULE_1__ = __webpack_require__(/*! ./encodec/feature_extraction_encodec.js */ "./src/models/encodec/feature_extraction_encodec.js");
|
|
@@ -20492,11 +20637,13 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
20492
20637
|
/* harmony import */ var _moonshine_feature_extraction_moonshine_js__WEBPACK_IMPORTED_MODULE_4__ = __webpack_require__(/*! ./moonshine/feature_extraction_moonshine.js */ "./src/models/moonshine/feature_extraction_moonshine.js");
|
|
20493
20638
|
/* harmony import */ var _pyannote_feature_extraction_pyannote_js__WEBPACK_IMPORTED_MODULE_5__ = __webpack_require__(/*! ./pyannote/feature_extraction_pyannote.js */ "./src/models/pyannote/feature_extraction_pyannote.js");
|
|
20494
20639
|
/* harmony import */ var _seamless_m4t_feature_extraction_seamless_m4t_js__WEBPACK_IMPORTED_MODULE_6__ = __webpack_require__(/*! ./seamless_m4t/feature_extraction_seamless_m4t.js */ "./src/models/seamless_m4t/feature_extraction_seamless_m4t.js");
|
|
20495
|
-
/* harmony import */ var
|
|
20496
|
-
/* harmony import */ var
|
|
20497
|
-
/* harmony import */ var
|
|
20498
|
-
/* harmony import */ var
|
|
20499
|
-
/* harmony import */ var
|
|
20640
|
+
/* harmony import */ var _snac_feature_extraction_snac_js__WEBPACK_IMPORTED_MODULE_7__ = __webpack_require__(/*! ./snac/feature_extraction_snac.js */ "./src/models/snac/feature_extraction_snac.js");
|
|
20641
|
+
/* harmony import */ var _speecht5_feature_extraction_speecht5_js__WEBPACK_IMPORTED_MODULE_8__ = __webpack_require__(/*! ./speecht5/feature_extraction_speecht5.js */ "./src/models/speecht5/feature_extraction_speecht5.js");
|
|
20642
|
+
/* harmony import */ var _wav2vec2_feature_extraction_wav2vec2_js__WEBPACK_IMPORTED_MODULE_9__ = __webpack_require__(/*! ./wav2vec2/feature_extraction_wav2vec2.js */ "./src/models/wav2vec2/feature_extraction_wav2vec2.js");
|
|
20643
|
+
/* harmony import */ var _wespeaker_feature_extraction_wespeaker_js__WEBPACK_IMPORTED_MODULE_10__ = __webpack_require__(/*! ./wespeaker/feature_extraction_wespeaker.js */ "./src/models/wespeaker/feature_extraction_wespeaker.js");
|
|
20644
|
+
/* harmony import */ var _whisper_feature_extraction_whisper_js__WEBPACK_IMPORTED_MODULE_11__ = __webpack_require__(/*! ./whisper/feature_extraction_whisper.js */ "./src/models/whisper/feature_extraction_whisper.js");
|
|
20645
|
+
/* harmony import */ var _base_image_processors_utils_js__WEBPACK_IMPORTED_MODULE_12__ = __webpack_require__(/*! ../base/image_processors_utils.js */ "./src/base/image_processors_utils.js");
|
|
20646
|
+
|
|
20500
20647
|
|
|
20501
20648
|
|
|
20502
20649
|
|
|
@@ -23486,6 +23633,24 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
23486
23633
|
|
|
23487
23634
|
|
|
23488
23635
|
|
|
23636
|
+
/***/ }),
|
|
23637
|
+
|
|
23638
|
+
/***/ "./src/models/snac/feature_extraction_snac.js":
|
|
23639
|
+
/*!****************************************************!*\
|
|
23640
|
+
!*** ./src/models/snac/feature_extraction_snac.js ***!
|
|
23641
|
+
\****************************************************/
|
|
23642
|
+
/***/ ((__unused_webpack___webpack_module__, __webpack_exports__, __webpack_require__) => {
|
|
23643
|
+
|
|
23644
|
+
__webpack_require__.r(__webpack_exports__);
|
|
23645
|
+
/* harmony export */ __webpack_require__.d(__webpack_exports__, {
|
|
23646
|
+
/* harmony export */ SnacFeatureExtractor: () => (/* binding */ SnacFeatureExtractor)
|
|
23647
|
+
/* harmony export */ });
|
|
23648
|
+
/* harmony import */ var _dac_feature_extraction_dac_js__WEBPACK_IMPORTED_MODULE_0__ = __webpack_require__(/*! ../dac/feature_extraction_dac.js */ "./src/models/dac/feature_extraction_dac.js");
|
|
23649
|
+
|
|
23650
|
+
|
|
23651
|
+
class SnacFeatureExtractor extends _dac_feature_extraction_dac_js__WEBPACK_IMPORTED_MODULE_0__.DacFeatureExtractor { }
|
|
23652
|
+
|
|
23653
|
+
|
|
23489
23654
|
/***/ }),
|
|
23490
23655
|
|
|
23491
23656
|
/***/ "./src/models/speecht5/feature_extraction_speecht5.js":
|
|
@@ -24725,16 +24890,16 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
24725
24890
|
/* harmony import */ var _utils_image_js__WEBPACK_IMPORTED_MODULE_9__ = __webpack_require__(/*! ./utils/image.js */ "./src/utils/image.js");
|
|
24726
24891
|
/**
|
|
24727
24892
|
* @file Pipelines provide a high-level, easy to use, API for running machine learning models.
|
|
24728
|
-
*
|
|
24893
|
+
*
|
|
24729
24894
|
* **Example:** Instantiate pipeline using the `pipeline` function.
|
|
24730
24895
|
* ```javascript
|
|
24731
24896
|
* import { pipeline } from '@huggingface/transformers';
|
|
24732
|
-
*
|
|
24897
|
+
*
|
|
24733
24898
|
* const classifier = await pipeline('sentiment-analysis');
|
|
24734
24899
|
* const output = await classifier('I love transformers!');
|
|
24735
24900
|
* // [{'label': 'POSITIVE', 'score': 0.999817686}]
|
|
24736
24901
|
* ```
|
|
24737
|
-
*
|
|
24902
|
+
*
|
|
24738
24903
|
* @module pipelines
|
|
24739
24904
|
*/
|
|
24740
24905
|
|
|
@@ -24753,7 +24918,7 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
24753
24918
|
|
|
24754
24919
|
|
|
24755
24920
|
/**
|
|
24756
|
-
* @typedef {string | RawImage | URL} ImageInput
|
|
24921
|
+
* @typedef {string | RawImage | URL | Blob | HTMLCanvasElement | OffscreenCanvas} ImageInput
|
|
24757
24922
|
* @typedef {ImageInput|ImageInput[]} ImagePipelineInputs
|
|
24758
24923
|
*/
|
|
24759
24924
|
|
|
@@ -24827,7 +24992,7 @@ function get_bounding_box(box, asInteger) {
|
|
|
24827
24992
|
/**
|
|
24828
24993
|
* @callback DisposeType Disposes the item.
|
|
24829
24994
|
* @returns {Promise<void>} A promise that resolves when the item has been disposed.
|
|
24830
|
-
*
|
|
24995
|
+
*
|
|
24831
24996
|
* @typedef {Object} Disposable
|
|
24832
24997
|
* @property {DisposeType} dispose A promise that resolves when the pipeline has been disposed.
|
|
24833
24998
|
*/
|
|
@@ -24864,7 +25029,7 @@ class Pipeline extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_4__.Callable {
|
|
|
24864
25029
|
* @property {string} task The task of the pipeline. Useful for specifying subtasks.
|
|
24865
25030
|
* @property {PreTrainedModel} model The model used by the pipeline.
|
|
24866
25031
|
* @property {PreTrainedTokenizer} tokenizer The tokenizer used by the pipeline.
|
|
24867
|
-
*
|
|
25032
|
+
*
|
|
24868
25033
|
* @typedef {ModelTokenizerConstructorArgs} TextPipelineConstructorArgs An object used to instantiate a text-based pipeline.
|
|
24869
25034
|
*/
|
|
24870
25035
|
|
|
@@ -24873,7 +25038,7 @@ class Pipeline extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_4__.Callable {
|
|
|
24873
25038
|
* @property {string} task The task of the pipeline. Useful for specifying subtasks.
|
|
24874
25039
|
* @property {PreTrainedModel} model The model used by the pipeline.
|
|
24875
25040
|
* @property {Processor} processor The processor used by the pipeline.
|
|
24876
|
-
*
|
|
25041
|
+
*
|
|
24877
25042
|
* @typedef {ModelProcessorConstructorArgs} AudioPipelineConstructorArgs An object used to instantiate an audio-based pipeline.
|
|
24878
25043
|
* @typedef {ModelProcessorConstructorArgs} ImagePipelineConstructorArgs An object used to instantiate an image-based pipeline.
|
|
24879
25044
|
*/
|
|
@@ -24885,7 +25050,7 @@ class Pipeline extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_4__.Callable {
|
|
|
24885
25050
|
* @property {PreTrainedModel} model The model used by the pipeline.
|
|
24886
25051
|
* @property {PreTrainedTokenizer} tokenizer The tokenizer used by the pipeline.
|
|
24887
25052
|
* @property {Processor} processor The processor used by the pipeline.
|
|
24888
|
-
*
|
|
25053
|
+
*
|
|
24889
25054
|
* @typedef {ModelTokenizerProcessorConstructorArgs} TextAudioPipelineConstructorArgs An object used to instantiate a text- and audio-based pipeline.
|
|
24890
25055
|
* @typedef {ModelTokenizerProcessorConstructorArgs} TextImagePipelineConstructorArgs An object used to instantiate a text- and image-based pipeline.
|
|
24891
25056
|
*/
|
|
@@ -24895,15 +25060,15 @@ class Pipeline extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_4__.Callable {
|
|
|
24895
25060
|
* @property {string} label The label predicted.
|
|
24896
25061
|
* @property {number} score The corresponding probability.
|
|
24897
25062
|
* @typedef {TextClassificationSingle[]} TextClassificationOutput
|
|
24898
|
-
*
|
|
25063
|
+
*
|
|
24899
25064
|
* @typedef {Object} TextClassificationPipelineOptions Parameters specific to text classification pipelines.
|
|
24900
25065
|
* @property {number} [top_k=1] The number of top predictions to be returned.
|
|
24901
|
-
*
|
|
25066
|
+
*
|
|
24902
25067
|
* @callback TextClassificationPipelineCallback Classify the text(s) given as inputs.
|
|
24903
25068
|
* @param {string|string[]} texts The input text(s) to be classified.
|
|
24904
25069
|
* @param {TextClassificationPipelineOptions} [options] The options to use for text classification.
|
|
24905
25070
|
* @returns {Promise<TextClassificationOutput|TextClassificationOutput[]>} An array or object containing the predicted labels and scores.
|
|
24906
|
-
*
|
|
25071
|
+
*
|
|
24907
25072
|
* @typedef {TextPipelineConstructorArgs & TextClassificationPipelineCallback & Disposable} TextClassificationPipelineType
|
|
24908
25073
|
*/
|
|
24909
25074
|
|
|
@@ -24916,7 +25081,7 @@ class Pipeline extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_4__.Callable {
|
|
|
24916
25081
|
* const output = await classifier('I love transformers!');
|
|
24917
25082
|
* // [{ label: 'POSITIVE', score: 0.999788761138916 }]
|
|
24918
25083
|
* ```
|
|
24919
|
-
*
|
|
25084
|
+
*
|
|
24920
25085
|
* **Example:** Multilingual sentiment-analysis w/ `Xenova/bert-base-multilingual-uncased-sentiment` (and return top 5 classes).
|
|
24921
25086
|
* ```javascript
|
|
24922
25087
|
* const classifier = await pipeline('sentiment-analysis', 'Xenova/bert-base-multilingual-uncased-sentiment');
|
|
@@ -24929,7 +25094,7 @@ class Pipeline extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_4__.Callable {
|
|
|
24929
25094
|
* // { label: '2 stars', score: 0.0009423971059732139 }
|
|
24930
25095
|
* // ]
|
|
24931
25096
|
* ```
|
|
24932
|
-
*
|
|
25097
|
+
*
|
|
24933
25098
|
* **Example:** Toxic comment classification w/ `Xenova/toxic-bert` (and return all classes).
|
|
24934
25099
|
* ```javascript
|
|
24935
25100
|
* const classifier = await pipeline('text-classification', 'Xenova/toxic-bert');
|
|
@@ -25014,21 +25179,21 @@ class TextClassificationPipeline extends (/** @type {new (options: TextPipelineC
|
|
|
25014
25179
|
* @property {number} [start] The index of the start of the corresponding entity in the sentence.
|
|
25015
25180
|
* @property {number} [end] The index of the end of the corresponding entity in the sentence.
|
|
25016
25181
|
* @typedef {TokenClassificationSingle[]} TokenClassificationOutput
|
|
25017
|
-
*
|
|
25182
|
+
*
|
|
25018
25183
|
* @typedef {Object} TokenClassificationPipelineOptions Parameters specific to token classification pipelines.
|
|
25019
25184
|
* @property {string[]} [ignore_labels] A list of labels to ignore.
|
|
25020
|
-
*
|
|
25185
|
+
*
|
|
25021
25186
|
* @callback TokenClassificationPipelineCallback Classify each token of the text(s) given as inputs.
|
|
25022
25187
|
* @param {string|string[]} texts One or several texts (or one list of texts) for token classification.
|
|
25023
25188
|
* @param {TokenClassificationPipelineOptions} [options] The options to use for token classification.
|
|
25024
25189
|
* @returns {Promise<TokenClassificationOutput|TokenClassificationOutput[]>} The result.
|
|
25025
|
-
*
|
|
25190
|
+
*
|
|
25026
25191
|
* @typedef {TextPipelineConstructorArgs & TokenClassificationPipelineCallback & Disposable} TokenClassificationPipelineType
|
|
25027
25192
|
*/
|
|
25028
25193
|
|
|
25029
25194
|
/**
|
|
25030
25195
|
* Named Entity Recognition pipeline using any `ModelForTokenClassification`.
|
|
25031
|
-
*
|
|
25196
|
+
*
|
|
25032
25197
|
* **Example:** Perform named entity recognition with `Xenova/bert-base-NER`.
|
|
25033
25198
|
* ```javascript
|
|
25034
25199
|
* const classifier = await pipeline('token-classification', 'Xenova/bert-base-NER');
|
|
@@ -25038,7 +25203,7 @@ class TextClassificationPipeline extends (/** @type {new (options: TextPipelineC
|
|
|
25038
25203
|
* // { entity: 'B-LOC', score: 0.9994474053382874, index: 9, word: 'London' }
|
|
25039
25204
|
* // ]
|
|
25040
25205
|
* ```
|
|
25041
|
-
*
|
|
25206
|
+
*
|
|
25042
25207
|
* **Example:** Perform named entity recognition with `Xenova/bert-base-NER` (and return all labels).
|
|
25043
25208
|
* ```javascript
|
|
25044
25209
|
* const classifier = await pipeline('token-classification', 'Xenova/bert-base-NER');
|
|
@@ -25134,22 +25299,22 @@ class TokenClassificationPipeline extends (/** @type {new (options: TextPipeline
|
|
|
25134
25299
|
* @property {number} [start] The character start index of the answer (in the tokenized version of the input).
|
|
25135
25300
|
* @property {number} [end] The character end index of the answer (in the tokenized version of the input).
|
|
25136
25301
|
* @property {string} answer The answer to the question.
|
|
25137
|
-
*
|
|
25302
|
+
*
|
|
25138
25303
|
* @typedef {Object} QuestionAnsweringPipelineOptions Parameters specific to question answering pipelines.
|
|
25139
25304
|
* @property {number} [top_k=1] The number of top answer predictions to be returned.
|
|
25140
|
-
*
|
|
25305
|
+
*
|
|
25141
25306
|
* @callback QuestionAnsweringPipelineCallback Answer the question(s) given as inputs by using the context(s).
|
|
25142
25307
|
* @param {string|string[]} question One or several question(s) (must be used in conjunction with the `context` argument).
|
|
25143
25308
|
* @param {string|string[]} context One or several context(s) associated with the question(s) (must be used in conjunction with the `question` argument).
|
|
25144
25309
|
* @param {QuestionAnsweringPipelineOptions} [options] The options to use for question answering.
|
|
25145
25310
|
* @returns {Promise<QuestionAnsweringOutput|QuestionAnsweringOutput[]>} An array or object containing the predicted answers and scores.
|
|
25146
|
-
*
|
|
25311
|
+
*
|
|
25147
25312
|
* @typedef {TextPipelineConstructorArgs & QuestionAnsweringPipelineCallback & Disposable} QuestionAnsweringPipelineType
|
|
25148
25313
|
*/
|
|
25149
25314
|
|
|
25150
25315
|
/**
|
|
25151
25316
|
* Question Answering pipeline using any `ModelForQuestionAnswering`.
|
|
25152
|
-
*
|
|
25317
|
+
*
|
|
25153
25318
|
* **Example:** Run question answering with `Xenova/distilbert-base-uncased-distilled-squad`.
|
|
25154
25319
|
* ```javascript
|
|
25155
25320
|
* const answerer = await pipeline('question-answering', 'Xenova/distilbert-base-uncased-distilled-squad');
|
|
@@ -25274,10 +25439,10 @@ class QuestionAnsweringPipeline extends (/** @type {new (options: TextPipelineCo
|
|
|
25274
25439
|
* @property {number} token The predicted token id (to replace the masked one).
|
|
25275
25440
|
* @property {string} token_str The predicted token (to replace the masked one).
|
|
25276
25441
|
* @typedef {FillMaskSingle[]} FillMaskOutput
|
|
25277
|
-
*
|
|
25442
|
+
*
|
|
25278
25443
|
* @typedef {Object} FillMaskPipelineOptions Parameters specific to fill mask pipelines.
|
|
25279
25444
|
* @property {number} [top_k=5] When passed, overrides the number of predictions to return.
|
|
25280
|
-
*
|
|
25445
|
+
*
|
|
25281
25446
|
* @callback FillMaskPipelineCallback Fill the masked token in the text(s) given as inputs.
|
|
25282
25447
|
* @param {string|string[]} texts One or several texts (or one list of prompts) with masked tokens.
|
|
25283
25448
|
* @param {FillMaskPipelineOptions} [options] The options to use for masked language modelling.
|
|
@@ -25285,13 +25450,13 @@ class QuestionAnsweringPipeline extends (/** @type {new (options: TextPipelineCo
|
|
|
25285
25450
|
* and the sequence with the predicted token filled in, or an array of such arrays (one for each input text).
|
|
25286
25451
|
* If only one input text is given, the output will be an array of objects.
|
|
25287
25452
|
* @throws {Error} When the mask token is not found in the input text.
|
|
25288
|
-
*
|
|
25453
|
+
*
|
|
25289
25454
|
* @typedef {TextPipelineConstructorArgs & FillMaskPipelineCallback & Disposable} FillMaskPipelineType
|
|
25290
25455
|
*/
|
|
25291
25456
|
|
|
25292
25457
|
/**
|
|
25293
25458
|
* Masked language modeling prediction pipeline using any `ModelWithLMHead`.
|
|
25294
|
-
*
|
|
25459
|
+
*
|
|
25295
25460
|
* **Example:** Perform masked language modelling (a.k.a. "fill-mask") with `Xenova/bert-base-uncased`.
|
|
25296
25461
|
* ```javascript
|
|
25297
25462
|
* const unmasker = await pipeline('fill-mask', 'Xenova/bert-base-cased');
|
|
@@ -25304,7 +25469,7 @@ class QuestionAnsweringPipeline extends (/** @type {new (options: TextPipelineCo
|
|
|
25304
25469
|
* // { token_str: 'life', score: 0.01859794743359089, token: 1297, sequence: 'The goal of life is life.' }
|
|
25305
25470
|
* // ]
|
|
25306
25471
|
* ```
|
|
25307
|
-
*
|
|
25472
|
+
*
|
|
25308
25473
|
* **Example:** Perform masked language modelling (a.k.a. "fill-mask") with `Xenova/bert-base-cased` (and return top result).
|
|
25309
25474
|
* ```javascript
|
|
25310
25475
|
* const unmasker = await pipeline('fill-mask', 'Xenova/bert-base-cased');
|
|
@@ -25381,18 +25546,18 @@ class FillMaskPipeline extends (/** @type {new (options: TextPipelineConstructor
|
|
|
25381
25546
|
* @typedef {Object} Text2TextGenerationSingle
|
|
25382
25547
|
* @property {string} generated_text The generated text.
|
|
25383
25548
|
* @typedef {Text2TextGenerationSingle[]} Text2TextGenerationOutput
|
|
25384
|
-
*
|
|
25549
|
+
*
|
|
25385
25550
|
* @callback Text2TextGenerationPipelineCallback Generate the output text(s) using text(s) given as inputs.
|
|
25386
25551
|
* @param {string|string[]} texts Input text for the encoder.
|
|
25387
25552
|
* @param {Partial<import('./generation/configuration_utils.js').GenerationConfig>} [options] Additional keyword arguments to pass along to the generate method of the model.
|
|
25388
25553
|
* @returns {Promise<Text2TextGenerationOutput|Text2TextGenerationOutput[]>}
|
|
25389
|
-
*
|
|
25554
|
+
*
|
|
25390
25555
|
* @typedef {TextPipelineConstructorArgs & Text2TextGenerationPipelineCallback & Disposable} Text2TextGenerationPipelineType
|
|
25391
25556
|
*/
|
|
25392
25557
|
|
|
25393
25558
|
/**
|
|
25394
25559
|
* Text2TextGenerationPipeline class for generating text using a model that performs text-to-text generation tasks.
|
|
25395
|
-
*
|
|
25560
|
+
*
|
|
25396
25561
|
* **Example:** Text-to-text generation w/ `Xenova/LaMini-Flan-T5-783M`.
|
|
25397
25562
|
* ```javascript
|
|
25398
25563
|
* const generator = await pipeline('text2text-generation', 'Xenova/LaMini-Flan-T5-783M');
|
|
@@ -25468,18 +25633,18 @@ class Text2TextGenerationPipeline extends (/** @type {new (options: TextPipeline
|
|
|
25468
25633
|
* @typedef {Object} SummarizationSingle
|
|
25469
25634
|
* @property {string} summary_text The summary text.
|
|
25470
25635
|
* @typedef {SummarizationSingle[]} SummarizationOutput
|
|
25471
|
-
*
|
|
25636
|
+
*
|
|
25472
25637
|
* @callback SummarizationPipelineCallback Summarize the text(s) given as inputs.
|
|
25473
25638
|
* @param {string|string[]} texts One or several articles (or one list of articles) to summarize.
|
|
25474
25639
|
* @param {import('./generation/configuration_utils.js').GenerationConfig} [options] Additional keyword arguments to pass along to the generate method of the model.
|
|
25475
25640
|
* @returns {Promise<SummarizationOutput|SummarizationOutput[]>}
|
|
25476
|
-
*
|
|
25641
|
+
*
|
|
25477
25642
|
* @typedef {TextPipelineConstructorArgs & SummarizationPipelineCallback & Disposable} SummarizationPipelineType
|
|
25478
25643
|
*/
|
|
25479
25644
|
|
|
25480
25645
|
/**
|
|
25481
25646
|
* A pipeline for summarization tasks, inheriting from Text2TextGenerationPipeline.
|
|
25482
|
-
*
|
|
25647
|
+
*
|
|
25483
25648
|
* **Example:** Summarization w/ `Xenova/distilbart-cnn-6-6`.
|
|
25484
25649
|
* ```javascript
|
|
25485
25650
|
* const generator = await pipeline('summarization', 'Xenova/distilbart-cnn-6-6');
|
|
@@ -25515,23 +25680,23 @@ class SummarizationPipeline extends (/** @type {new (options: TextPipelineConstr
|
|
|
25515
25680
|
* @typedef {Object} TranslationSingle
|
|
25516
25681
|
* @property {string} translation_text The translated text.
|
|
25517
25682
|
* @typedef {TranslationSingle[]} TranslationOutput
|
|
25518
|
-
*
|
|
25683
|
+
*
|
|
25519
25684
|
* @callback TranslationPipelineCallback Translate the text(s) given as inputs.
|
|
25520
25685
|
* @param {string|string[]} texts Texts to be translated.
|
|
25521
25686
|
* @param {import('./generation/configuration_utils.js').GenerationConfig} [options] Additional keyword arguments to pass along to the generate method of the model.
|
|
25522
25687
|
* @returns {Promise<TranslationOutput|TranslationOutput[]>}
|
|
25523
|
-
*
|
|
25688
|
+
*
|
|
25524
25689
|
* @typedef {TextPipelineConstructorArgs & TranslationPipelineCallback & Disposable} TranslationPipelineType
|
|
25525
25690
|
*/
|
|
25526
25691
|
|
|
25527
25692
|
/**
|
|
25528
25693
|
* Translates text from one language to another.
|
|
25529
|
-
*
|
|
25694
|
+
*
|
|
25530
25695
|
* **Example:** Multilingual translation w/ `Xenova/nllb-200-distilled-600M`.
|
|
25531
|
-
*
|
|
25696
|
+
*
|
|
25532
25697
|
* See [here](https://github.com/facebookresearch/flores/blob/main/flores200/README.md#languages-in-flores-200)
|
|
25533
25698
|
* for the full list of languages and their corresponding codes.
|
|
25534
|
-
*
|
|
25699
|
+
*
|
|
25535
25700
|
* ```javascript
|
|
25536
25701
|
* const translator = await pipeline('translation', 'Xenova/nllb-200-distilled-600M');
|
|
25537
25702
|
* const output = await translator('जीवन एक चॉकलेट बॉक्स की तरह है।', {
|
|
@@ -25540,12 +25705,12 @@ class SummarizationPipeline extends (/** @type {new (options: TextPipelineConstr
|
|
|
25540
25705
|
* });
|
|
25541
25706
|
* // [{ translation_text: 'La vie est comme une boîte à chocolat.' }]
|
|
25542
25707
|
* ```
|
|
25543
|
-
*
|
|
25708
|
+
*
|
|
25544
25709
|
* **Example:** Multilingual translation w/ `Xenova/m2m100_418M`.
|
|
25545
|
-
*
|
|
25710
|
+
*
|
|
25546
25711
|
* See [here](https://huggingface.co/facebook/m2m100_418M#languages-covered)
|
|
25547
25712
|
* for the full list of languages and their corresponding codes.
|
|
25548
|
-
*
|
|
25713
|
+
*
|
|
25549
25714
|
* ```javascript
|
|
25550
25715
|
* const translator = await pipeline('translation', 'Xenova/m2m100_418M');
|
|
25551
25716
|
* const output = await translator('生活就像一盒巧克力。', {
|
|
@@ -25554,12 +25719,12 @@ class SummarizationPipeline extends (/** @type {new (options: TextPipelineConstr
|
|
|
25554
25719
|
* });
|
|
25555
25720
|
* // [{ translation_text: 'Life is like a box of chocolate.' }]
|
|
25556
25721
|
* ```
|
|
25557
|
-
*
|
|
25722
|
+
*
|
|
25558
25723
|
* **Example:** Multilingual translation w/ `Xenova/mbart-large-50-many-to-many-mmt`.
|
|
25559
|
-
*
|
|
25724
|
+
*
|
|
25560
25725
|
* See [here](https://huggingface.co/facebook/mbart-large-50-many-to-many-mmt#languages-covered)
|
|
25561
25726
|
* for the full list of languages and their corresponding codes.
|
|
25562
|
-
*
|
|
25727
|
+
*
|
|
25563
25728
|
* ```javascript
|
|
25564
25729
|
* const translator = await pipeline('translation', 'Xenova/mbart-large-50-many-to-many-mmt');
|
|
25565
25730
|
* const output = await translator('संयुक्त राष्ट्र के प्रमुख का कहना है कि सीरिया में कोई सैन्य समाधान नहीं है', {
|
|
@@ -25588,21 +25753,21 @@ function isChat(x) {
|
|
|
25588
25753
|
|
|
25589
25754
|
/**
|
|
25590
25755
|
* @typedef {import('./tokenizers.js').Message[]} Chat
|
|
25591
|
-
*
|
|
25756
|
+
*
|
|
25592
25757
|
* @typedef {Object} TextGenerationSingle
|
|
25593
25758
|
* @property {string|Chat} generated_text The generated text.
|
|
25594
25759
|
* @typedef {TextGenerationSingle[]} TextGenerationOutput
|
|
25595
|
-
*
|
|
25760
|
+
*
|
|
25596
25761
|
* @typedef {Object} TextGenerationSpecificParams Parameters specific to text-generation pipelines.
|
|
25597
25762
|
* @property {boolean} [add_special_tokens] Whether or not to add special tokens when tokenizing the sequences.
|
|
25598
25763
|
* @property {boolean} [return_full_text=true] If set to `false` only added text is returned, otherwise the full text is returned.
|
|
25599
25764
|
* @typedef {import('./generation/configuration_utils.js').GenerationConfig & TextGenerationSpecificParams} TextGenerationConfig
|
|
25600
|
-
*
|
|
25765
|
+
*
|
|
25601
25766
|
* @callback TextGenerationPipelineCallback Complete the prompt(s) given as inputs.
|
|
25602
25767
|
* @param {string|string[]|Chat|Chat[]} texts One or several prompts (or one list of prompts) to complete.
|
|
25603
25768
|
* @param {Partial<TextGenerationConfig>} [options] Additional keyword arguments to pass along to the generate method of the model.
|
|
25604
25769
|
* @returns {Promise<TextGenerationOutput|TextGenerationOutput[]>} An array or object containing the generated texts.
|
|
25605
|
-
*
|
|
25770
|
+
*
|
|
25606
25771
|
* @typedef {TextPipelineConstructorArgs & TextGenerationPipelineCallback & Disposable} TextGenerationPipelineType
|
|
25607
25772
|
*/
|
|
25608
25773
|
|
|
@@ -25610,7 +25775,7 @@ function isChat(x) {
|
|
|
25610
25775
|
* Language generation pipeline using any `ModelWithLMHead` or `ModelForCausalLM`.
|
|
25611
25776
|
* This pipeline predicts the words that will follow a specified text prompt.
|
|
25612
25777
|
* NOTE: For the full list of generation parameters, see [`GenerationConfig`](./utils/generation#module_utils/generation.GenerationConfig).
|
|
25613
|
-
*
|
|
25778
|
+
*
|
|
25614
25779
|
* **Example:** Text generation with `Xenova/distilgpt2` (default settings).
|
|
25615
25780
|
* ```javascript
|
|
25616
25781
|
* const generator = await pipeline('text-generation', 'Xenova/distilgpt2');
|
|
@@ -25618,7 +25783,7 @@ function isChat(x) {
|
|
|
25618
25783
|
* const output = await generator(text);
|
|
25619
25784
|
* // [{ generated_text: "I enjoy walking with my cute dog, and I love to play with the other dogs." }]
|
|
25620
25785
|
* ```
|
|
25621
|
-
*
|
|
25786
|
+
*
|
|
25622
25787
|
* **Example:** Text generation with `Xenova/distilgpt2` (custom settings).
|
|
25623
25788
|
* ```javascript
|
|
25624
25789
|
* const generator = await pipeline('text-generation', 'Xenova/distilgpt2');
|
|
@@ -25637,7 +25802,7 @@ function isChat(x) {
|
|
|
25637
25802
|
* // "generated_text": "Once upon a time, there was an abundance of information about the most important and influential"
|
|
25638
25803
|
* // }]
|
|
25639
25804
|
* ```
|
|
25640
|
-
*
|
|
25805
|
+
*
|
|
25641
25806
|
* **Example:** Run code generation with `Xenova/codegen-350M-mono`.
|
|
25642
25807
|
* ```javascript
|
|
25643
25808
|
* const generator = await pipeline('text-generation', 'Xenova/codegen-350M-mono');
|
|
@@ -25756,7 +25921,7 @@ class TextGenerationPipeline extends (/** @type {new (options: TextPipelineConst
|
|
|
25756
25921
|
* @property {string} sequence The sequence for which this is the output.
|
|
25757
25922
|
* @property {string[]} labels The labels sorted by order of likelihood.
|
|
25758
25923
|
* @property {number[]} scores The probabilities for each of the labels.
|
|
25759
|
-
*
|
|
25924
|
+
*
|
|
25760
25925
|
* @typedef {Object} ZeroShotClassificationPipelineOptions Parameters specific to zero-shot classification pipelines.
|
|
25761
25926
|
* @property {string} [hypothesis_template="This example is {}."] The template used to turn each
|
|
25762
25927
|
* candidate label into an NLI-style hypothesis. The candidate label will replace the {} placeholder.
|
|
@@ -25764,14 +25929,14 @@ class TextGenerationPipeline extends (/** @type {new (options: TextPipelineConst
|
|
|
25764
25929
|
* If `false`, the scores are normalized such that the sum of the label likelihoods for each sequence
|
|
25765
25930
|
* is 1. If `true`, the labels are considered independent and probabilities are normalized for each
|
|
25766
25931
|
* candidate by doing a softmax of the entailment score vs. the contradiction score.
|
|
25767
|
-
*
|
|
25932
|
+
*
|
|
25768
25933
|
* @callback ZeroShotClassificationPipelineCallback Classify the sequence(s) given as inputs.
|
|
25769
25934
|
* @param {string|string[]} texts The sequence(s) to classify, will be truncated if the model input is too large.
|
|
25770
25935
|
* @param {string|string[]} candidate_labels The set of possible class labels to classify each sequence into.
|
|
25771
25936
|
* Can be a single label, a string of comma-separated labels, or a list of labels.
|
|
25772
25937
|
* @param {ZeroShotClassificationPipelineOptions} [options] The options to use for zero-shot classification.
|
|
25773
25938
|
* @returns {Promise<ZeroShotClassificationOutput|ZeroShotClassificationOutput[]>} An array or object containing the predicted labels and scores.
|
|
25774
|
-
*
|
|
25939
|
+
*
|
|
25775
25940
|
* @typedef {TextPipelineConstructorArgs & ZeroShotClassificationPipelineCallback & Disposable} ZeroShotClassificationPipelineType
|
|
25776
25941
|
*/
|
|
25777
25942
|
|
|
@@ -25780,7 +25945,7 @@ class TextGenerationPipeline extends (/** @type {new (options: TextPipelineConst
|
|
|
25780
25945
|
* trained on NLI (natural language inference) tasks. Equivalent of `text-classification`
|
|
25781
25946
|
* pipelines, but these models don't require a hardcoded number of potential classes, they
|
|
25782
25947
|
* can be chosen at runtime. It usually means it's slower but it is **much** more flexible.
|
|
25783
|
-
*
|
|
25948
|
+
*
|
|
25784
25949
|
* **Example:** Zero shot classification with `Xenova/mobilebert-uncased-mnli`.
|
|
25785
25950
|
* ```javascript
|
|
25786
25951
|
* const classifier = await pipeline('zero-shot-classification', 'Xenova/mobilebert-uncased-mnli');
|
|
@@ -25793,7 +25958,7 @@ class TextGenerationPipeline extends (/** @type {new (options: TextPipelineConst
|
|
|
25793
25958
|
* // scores: [ 0.5562091040482018, 0.1843621307860853, 0.13942646639336376, 0.12000229877234923 ]
|
|
25794
25959
|
* // }
|
|
25795
25960
|
* ```
|
|
25796
|
-
*
|
|
25961
|
+
*
|
|
25797
25962
|
* **Example:** Zero shot classification with `Xenova/nli-deberta-v3-xsmall` (multi-label).
|
|
25798
25963
|
* ```javascript
|
|
25799
25964
|
* const classifier = await pipeline('zero-shot-classification', 'Xenova/nli-deberta-v3-xsmall');
|
|
@@ -25907,20 +26072,20 @@ class ZeroShotClassificationPipeline extends (/** @type {new (options: TextPipel
|
|
|
25907
26072
|
* @property {'none'|'mean'|'cls'} [pooling="none"] The pooling method to use.
|
|
25908
26073
|
* @property {boolean} [normalize=false] Whether or not to normalize the embeddings in the last dimension.
|
|
25909
26074
|
* @property {boolean} [quantize=false] Whether or not to quantize the embeddings.
|
|
25910
|
-
* @property {'binary'|'ubinary'} [precision='binary'] The precision to use for quantization.
|
|
25911
|
-
*
|
|
26075
|
+
* @property {'binary'|'ubinary'} [precision='binary'] The precision to use for quantization.
|
|
26076
|
+
*
|
|
25912
26077
|
* @callback FeatureExtractionPipelineCallback Extract the features of the input(s).
|
|
25913
26078
|
* @param {string|string[]} texts One or several texts (or one list of texts) to get the features of.
|
|
25914
26079
|
* @param {FeatureExtractionPipelineOptions} [options] The options to use for feature extraction.
|
|
25915
26080
|
* @returns {Promise<Tensor>} The features computed by the model.
|
|
25916
|
-
*
|
|
26081
|
+
*
|
|
25917
26082
|
* @typedef {TextPipelineConstructorArgs & FeatureExtractionPipelineCallback & Disposable} FeatureExtractionPipelineType
|
|
25918
26083
|
*/
|
|
25919
26084
|
|
|
25920
26085
|
/**
|
|
25921
26086
|
* Feature extraction pipeline using no model head. This pipeline extracts the hidden
|
|
25922
26087
|
* states from the base transformer, which can be used as features in downstream tasks.
|
|
25923
|
-
*
|
|
26088
|
+
*
|
|
25924
26089
|
* **Example:** Run feature extraction with `bert-base-uncased` (without pooling/normalization).
|
|
25925
26090
|
* ```javascript
|
|
25926
26091
|
* const extractor = await pipeline('feature-extraction', 'Xenova/bert-base-uncased', { revision: 'default' });
|
|
@@ -25931,7 +26096,7 @@ class ZeroShotClassificationPipeline extends (/** @type {new (options: TextPipel
|
|
|
25931
26096
|
* // dims: [1, 8, 768]
|
|
25932
26097
|
* // }
|
|
25933
26098
|
* ```
|
|
25934
|
-
*
|
|
26099
|
+
*
|
|
25935
26100
|
* **Example:** Run feature extraction with `bert-base-uncased` (with pooling/normalization).
|
|
25936
26101
|
* ```javascript
|
|
25937
26102
|
* const extractor = await pipeline('feature-extraction', 'Xenova/bert-base-uncased', { revision: 'default' });
|
|
@@ -25942,7 +26107,7 @@ class ZeroShotClassificationPipeline extends (/** @type {new (options: TextPipel
|
|
|
25942
26107
|
* // dims: [1, 768]
|
|
25943
26108
|
* // }
|
|
25944
26109
|
* ```
|
|
25945
|
-
*
|
|
26110
|
+
*
|
|
25946
26111
|
* **Example:** Calculating embeddings with `sentence-transformers` models.
|
|
25947
26112
|
* ```javascript
|
|
25948
26113
|
* const extractor = await pipeline('feature-extraction', 'Xenova/all-MiniLM-L6-v2');
|
|
@@ -26023,19 +26188,19 @@ class FeatureExtractionPipeline extends (/** @type {new (options: TextPipelineCo
|
|
|
26023
26188
|
/**
|
|
26024
26189
|
* @typedef {Object} ImageFeatureExtractionPipelineOptions Parameters specific to image feature extraction pipelines.
|
|
26025
26190
|
* @property {boolean} [pool=null] Whether or not to return the pooled output. If set to `false`, the model will return the raw hidden states.
|
|
26026
|
-
*
|
|
26191
|
+
*
|
|
26027
26192
|
* @callback ImageFeatureExtractionPipelineCallback Extract the features of the input(s).
|
|
26028
26193
|
* @param {ImagePipelineInputs} images One or several images (or one list of images) to get the features of.
|
|
26029
26194
|
* @param {ImageFeatureExtractionPipelineOptions} [options] The options to use for image feature extraction.
|
|
26030
26195
|
* @returns {Promise<Tensor>} The image features computed by the model.
|
|
26031
|
-
*
|
|
26196
|
+
*
|
|
26032
26197
|
* @typedef {ImagePipelineConstructorArgs & ImageFeatureExtractionPipelineCallback & Disposable} ImageFeatureExtractionPipelineType
|
|
26033
26198
|
*/
|
|
26034
26199
|
|
|
26035
26200
|
/**
|
|
26036
26201
|
* Image feature extraction pipeline using no model head. This pipeline extracts the hidden
|
|
26037
26202
|
* states from the base transformer, which can be used as features in downstream tasks.
|
|
26038
|
-
*
|
|
26203
|
+
*
|
|
26039
26204
|
* **Example:** Perform image feature extraction with `Xenova/vit-base-patch16-224-in21k`.
|
|
26040
26205
|
* ```javascript
|
|
26041
26206
|
* const image_feature_extractor = await pipeline('image-feature-extraction', 'Xenova/vit-base-patch16-224-in21k');
|
|
@@ -26048,7 +26213,7 @@ class FeatureExtractionPipeline extends (/** @type {new (options: TextPipelineCo
|
|
|
26048
26213
|
* // size: 151296
|
|
26049
26214
|
* // }
|
|
26050
26215
|
* ```
|
|
26051
|
-
*
|
|
26216
|
+
*
|
|
26052
26217
|
* **Example:** Compute image embeddings with `Xenova/clip-vit-base-patch32`.
|
|
26053
26218
|
* ```javascript
|
|
26054
26219
|
* const image_feature_extractor = await pipeline('image-feature-extraction', 'Xenova/clip-vit-base-patch32');
|
|
@@ -26104,12 +26269,12 @@ class ImageFeatureExtractionPipeline extends (/** @type {new (options: ImagePipe
|
|
|
26104
26269
|
* @property {string} label The label predicted.
|
|
26105
26270
|
* @property {number} score The corresponding probability.
|
|
26106
26271
|
* @typedef {AudioClassificationSingle[]} AudioClassificationOutput
|
|
26107
|
-
*
|
|
26272
|
+
*
|
|
26108
26273
|
* @typedef {Object} AudioClassificationPipelineOptions Parameters specific to audio classification pipelines.
|
|
26109
26274
|
* @property {number} [top_k=5] The number of top labels that will be returned by the pipeline.
|
|
26110
26275
|
* If the provided number is `null` or higher than the number of labels available in the model configuration,
|
|
26111
26276
|
* it will default to the number of labels.
|
|
26112
|
-
*
|
|
26277
|
+
*
|
|
26113
26278
|
* @callback AudioClassificationPipelineCallback Classify the sequence(s) given as inputs.
|
|
26114
26279
|
* @param {AudioPipelineInputs} audio The input audio file(s) to be classified. The input is either:
|
|
26115
26280
|
* - `string` or `URL` that is the filename/URL of the audio file, the file will be read at the processor's sampling rate
|
|
@@ -26118,14 +26283,14 @@ class ImageFeatureExtractionPipeline extends (/** @type {new (options: ImagePipe
|
|
|
26118
26283
|
* - `Float32Array` or `Float64Array` of shape `(n, )`, representing the raw audio at the correct sampling rate (no further check will be done).
|
|
26119
26284
|
* @param {AudioClassificationPipelineOptions} [options] The options to use for audio classification.
|
|
26120
26285
|
* @returns {Promise<AudioClassificationOutput|AudioClassificationOutput[]>} An array or object containing the predicted labels and scores.
|
|
26121
|
-
*
|
|
26286
|
+
*
|
|
26122
26287
|
* @typedef {AudioPipelineConstructorArgs & AudioClassificationPipelineCallback & Disposable} AudioClassificationPipelineType
|
|
26123
26288
|
*/
|
|
26124
26289
|
|
|
26125
26290
|
/**
|
|
26126
26291
|
* Audio classification pipeline using any `AutoModelForAudioClassification`.
|
|
26127
26292
|
* This pipeline predicts the class of a raw waveform or an audio file.
|
|
26128
|
-
*
|
|
26293
|
+
*
|
|
26129
26294
|
* **Example:** Perform audio classification with `Xenova/wav2vec2-large-xlsr-53-gender-recognition-librispeech`.
|
|
26130
26295
|
* ```javascript
|
|
26131
26296
|
* const classifier = await pipeline('audio-classification', 'Xenova/wav2vec2-large-xlsr-53-gender-recognition-librispeech');
|
|
@@ -26136,7 +26301,7 @@ class ImageFeatureExtractionPipeline extends (/** @type {new (options: ImagePipe
|
|
|
26136
26301
|
* // { label: 'female', score: 0.001845747814513743 }
|
|
26137
26302
|
* // ]
|
|
26138
26303
|
* ```
|
|
26139
|
-
*
|
|
26304
|
+
*
|
|
26140
26305
|
* **Example:** Perform audio classification with `Xenova/ast-finetuned-audioset-10-10-0.4593` and return top 4 results.
|
|
26141
26306
|
* ```javascript
|
|
26142
26307
|
* const classifier = await pipeline('audio-classification', 'Xenova/ast-finetuned-audioset-10-10-0.4593');
|
|
@@ -26201,12 +26366,12 @@ class AudioClassificationPipeline extends (/** @type {new (options: AudioPipelin
|
|
|
26201
26366
|
* @typedef {Object} ZeroShotAudioClassificationOutput
|
|
26202
26367
|
* @property {string} label The label identified by the model. It is one of the suggested `candidate_label`.
|
|
26203
26368
|
* @property {number} score The score attributed by the model for that label (between 0 and 1).
|
|
26204
|
-
*
|
|
26369
|
+
*
|
|
26205
26370
|
* @typedef {Object} ZeroShotAudioClassificationPipelineOptions Parameters specific to zero-shot audio classification pipelines.
|
|
26206
26371
|
* @property {string} [hypothesis_template="This is a sound of {}."] The sentence used in conjunction with `candidate_labels`
|
|
26207
26372
|
* to attempt the audio classification by replacing the placeholder with the candidate_labels.
|
|
26208
26373
|
* Then likelihood is estimated by using `logits_per_audio`.
|
|
26209
|
-
*
|
|
26374
|
+
*
|
|
26210
26375
|
* @callback ZeroShotAudioClassificationPipelineCallback Classify the sequence(s) given as inputs.
|
|
26211
26376
|
* @param {AudioPipelineInputs} audio The input audio file(s) to be classified. The input is either:
|
|
26212
26377
|
* - `string` or `URL` that is the filename/URL of the audio file, the file will be read at the processor's sampling rate
|
|
@@ -26216,14 +26381,14 @@ class AudioClassificationPipeline extends (/** @type {new (options: AudioPipelin
|
|
|
26216
26381
|
* @param {string[]} candidate_labels The candidate labels for this audio.
|
|
26217
26382
|
* @param {ZeroShotAudioClassificationPipelineOptions} [options] The options to use for zero-shot audio classification.
|
|
26218
26383
|
* @returns {Promise<ZeroShotAudioClassificationOutput[]|ZeroShotAudioClassificationOutput[][]>} An array of objects containing the predicted labels and scores.
|
|
26219
|
-
*
|
|
26384
|
+
*
|
|
26220
26385
|
* @typedef {TextAudioPipelineConstructorArgs & ZeroShotAudioClassificationPipelineCallback & Disposable} ZeroShotAudioClassificationPipelineType
|
|
26221
26386
|
*/
|
|
26222
26387
|
|
|
26223
26388
|
/**
|
|
26224
26389
|
* Zero shot audio classification pipeline using `ClapModel`. This pipeline predicts the class of an audio when you
|
|
26225
26390
|
* provide an audio and a set of `candidate_labels`.
|
|
26226
|
-
*
|
|
26391
|
+
*
|
|
26227
26392
|
* **Example**: Perform zero-shot audio classification with `Xenova/clap-htsat-unfused`.
|
|
26228
26393
|
* ```javascript
|
|
26229
26394
|
* const classifier = await pipeline('zero-shot-audio-classification', 'Xenova/clap-htsat-unfused');
|
|
@@ -26256,7 +26421,7 @@ class ZeroShotAudioClassificationPipeline extends (/** @type {new (options: Text
|
|
|
26256
26421
|
audio = [/** @type {AudioInput} */ (audio)];
|
|
26257
26422
|
}
|
|
26258
26423
|
|
|
26259
|
-
// Insert label into hypothesis template
|
|
26424
|
+
// Insert label into hypothesis template
|
|
26260
26425
|
const texts = candidate_labels.map(
|
|
26261
26426
|
x => hypothesis_template.replace('{}', x)
|
|
26262
26427
|
);
|
|
@@ -26300,7 +26465,7 @@ class ZeroShotAudioClassificationPipeline extends (/** @type {new (options: Text
|
|
|
26300
26465
|
* @property {string} text The recognized text.
|
|
26301
26466
|
* @property {Chunk[]} [chunks] When using `return_timestamps`, the `chunks` will become a list
|
|
26302
26467
|
* containing all the various text chunks identified by the model.
|
|
26303
|
-
*
|
|
26468
|
+
*
|
|
26304
26469
|
* @typedef {Object} AutomaticSpeechRecognitionSpecificParams Parameters specific to automatic-speech-recognition pipelines.
|
|
26305
26470
|
* @property {boolean|'word'} [return_timestamps] Whether to return timestamps or not. Default is `false`.
|
|
26306
26471
|
* @property {number} [chunk_length_s] The length of audio chunks to process in seconds. Default is 0 (no chunking).
|
|
@@ -26310,7 +26475,7 @@ class ZeroShotAudioClassificationPipeline extends (/** @type {new (options: Text
|
|
|
26310
26475
|
* @property {string} [task] The task to perform. Default is `null`, meaning it should be auto-detected.
|
|
26311
26476
|
* @property {number} [num_frames] The number of frames in the input audio.
|
|
26312
26477
|
* @typedef {import('./generation/configuration_utils.js').GenerationConfig & AutomaticSpeechRecognitionSpecificParams} AutomaticSpeechRecognitionConfig
|
|
26313
|
-
*
|
|
26478
|
+
*
|
|
26314
26479
|
* @callback AutomaticSpeechRecognitionPipelineCallback Transcribe the audio sequence(s) given as inputs to text.
|
|
26315
26480
|
* @param {AudioPipelineInputs} audio The input audio file(s) to be transcribed. The input is either:
|
|
26316
26481
|
* - `string` or `URL` that is the filename/URL of the audio file, the file will be read at the processor's sampling rate
|
|
@@ -26319,7 +26484,7 @@ class ZeroShotAudioClassificationPipeline extends (/** @type {new (options: Text
|
|
|
26319
26484
|
* - `Float32Array` or `Float64Array` of shape `(n, )`, representing the raw audio at the correct sampling rate (no further check will be done).
|
|
26320
26485
|
* @param {Partial<AutomaticSpeechRecognitionConfig>} [options] Additional keyword arguments to pass along to the generate method of the model.
|
|
26321
26486
|
* @returns {Promise<AutomaticSpeechRecognitionOutput|AutomaticSpeechRecognitionOutput[]>} An object containing the transcription text and optionally timestamps if `return_timestamps` is `true`.
|
|
26322
|
-
*
|
|
26487
|
+
*
|
|
26323
26488
|
* @typedef {TextAudioPipelineConstructorArgs & AutomaticSpeechRecognitionPipelineCallback & Disposable} AutomaticSpeechRecognitionPipelineType
|
|
26324
26489
|
*/
|
|
26325
26490
|
|
|
@@ -26333,7 +26498,7 @@ class ZeroShotAudioClassificationPipeline extends (/** @type {new (options: Text
|
|
|
26333
26498
|
* const output = await transcriber(url);
|
|
26334
26499
|
* // { text: " And so my fellow Americans ask not what your country can do for you, ask what you can do for your country." }
|
|
26335
26500
|
* ```
|
|
26336
|
-
*
|
|
26501
|
+
*
|
|
26337
26502
|
* **Example:** Transcribe English w/ timestamps.
|
|
26338
26503
|
* ```javascript
|
|
26339
26504
|
* const transcriber = await pipeline('automatic-speech-recognition', 'Xenova/whisper-tiny.en');
|
|
@@ -26347,7 +26512,7 @@ class ZeroShotAudioClassificationPipeline extends (/** @type {new (options: Text
|
|
|
26347
26512
|
* // ]
|
|
26348
26513
|
* // }
|
|
26349
26514
|
* ```
|
|
26350
|
-
*
|
|
26515
|
+
*
|
|
26351
26516
|
* **Example:** Transcribe English w/ word-level timestamps.
|
|
26352
26517
|
* ```javascript
|
|
26353
26518
|
* const transcriber = await pipeline('automatic-speech-recognition', 'Xenova/whisper-tiny.en');
|
|
@@ -26366,7 +26531,7 @@ class ZeroShotAudioClassificationPipeline extends (/** @type {new (options: Text
|
|
|
26366
26531
|
* // ]
|
|
26367
26532
|
* // }
|
|
26368
26533
|
* ```
|
|
26369
|
-
*
|
|
26534
|
+
*
|
|
26370
26535
|
* **Example:** Transcribe French.
|
|
26371
26536
|
* ```javascript
|
|
26372
26537
|
* const transcriber = await pipeline('automatic-speech-recognition', 'Xenova/whisper-small');
|
|
@@ -26374,7 +26539,7 @@ class ZeroShotAudioClassificationPipeline extends (/** @type {new (options: Text
|
|
|
26374
26539
|
* const output = await transcriber(url, { language: 'french', task: 'transcribe' });
|
|
26375
26540
|
* // { text: " J'adore, j'aime, je n'aime pas, je déteste." }
|
|
26376
26541
|
* ```
|
|
26377
|
-
*
|
|
26542
|
+
*
|
|
26378
26543
|
* **Example:** Translate French to English.
|
|
26379
26544
|
* ```javascript
|
|
26380
26545
|
* const transcriber = await pipeline('automatic-speech-recognition', 'Xenova/whisper-small');
|
|
@@ -26382,7 +26547,7 @@ class ZeroShotAudioClassificationPipeline extends (/** @type {new (options: Text
|
|
|
26382
26547
|
* const output = await transcriber(url, { language: 'french', task: 'translate' });
|
|
26383
26548
|
* // { text: " I love, I like, I don't like, I hate." }
|
|
26384
26549
|
* ```
|
|
26385
|
-
*
|
|
26550
|
+
*
|
|
26386
26551
|
* **Example:** Transcribe/translate audio longer than 30 seconds.
|
|
26387
26552
|
* ```javascript
|
|
26388
26553
|
* const transcriber = await pipeline('automatic-speech-recognition', 'Xenova/whisper-tiny.en');
|
|
@@ -26605,18 +26770,18 @@ class AutomaticSpeechRecognitionPipeline extends (/** @type {new (options: TextA
|
|
|
26605
26770
|
* @typedef {Object} ImageToTextSingle
|
|
26606
26771
|
* @property {string} generated_text The generated text.
|
|
26607
26772
|
* @typedef {ImageToTextSingle[]} ImageToTextOutput
|
|
26608
|
-
*
|
|
26773
|
+
*
|
|
26609
26774
|
* @callback ImageToTextPipelineCallback Assign labels to the image(s) passed as inputs.
|
|
26610
26775
|
* @param {ImagePipelineInputs} texts The images to be captioned.
|
|
26611
26776
|
* @param {Partial<import('./generation/configuration_utils.js').GenerationConfig>} [options] Additional keyword arguments to pass along to the generate method of the model.
|
|
26612
26777
|
* @returns {Promise<ImageToTextOutput|ImageToTextOutput[]>} An object (or array of objects) containing the generated text(s).
|
|
26613
|
-
*
|
|
26778
|
+
*
|
|
26614
26779
|
* @typedef {TextImagePipelineConstructorArgs & ImageToTextPipelineCallback & Disposable} ImageToTextPipelineType
|
|
26615
26780
|
*/
|
|
26616
26781
|
|
|
26617
26782
|
/**
|
|
26618
26783
|
* Image To Text pipeline using a `AutoModelForVision2Seq`. This pipeline predicts a caption for a given image.
|
|
26619
|
-
*
|
|
26784
|
+
*
|
|
26620
26785
|
* **Example:** Generate a caption for an image w/ `Xenova/vit-gpt2-image-captioning`.
|
|
26621
26786
|
* ```javascript
|
|
26622
26787
|
* const captioner = await pipeline('image-to-text', 'Xenova/vit-gpt2-image-captioning');
|
|
@@ -26624,7 +26789,7 @@ class AutomaticSpeechRecognitionPipeline extends (/** @type {new (options: TextA
|
|
|
26624
26789
|
* const output = await captioner(url);
|
|
26625
26790
|
* // [{ generated_text: 'a cat laying on a couch with another cat' }]
|
|
26626
26791
|
* ```
|
|
26627
|
-
*
|
|
26792
|
+
*
|
|
26628
26793
|
* **Example:** Optical Character Recognition (OCR) w/ `Xenova/trocr-small-handwritten`.
|
|
26629
26794
|
* ```javascript
|
|
26630
26795
|
* const captioner = await pipeline('image-to-text', 'Xenova/trocr-small-handwritten');
|
|
@@ -26670,22 +26835,22 @@ class ImageToTextPipeline extends (/** @type {new (options: TextImagePipelineCon
|
|
|
26670
26835
|
* @property {string} label The label identified by the model.
|
|
26671
26836
|
* @property {number} score The score attributed by the model for that label.
|
|
26672
26837
|
* @typedef {ImageClassificationSingle[]} ImageClassificationOutput
|
|
26673
|
-
*
|
|
26838
|
+
*
|
|
26674
26839
|
* @typedef {Object} ImageClassificationPipelineOptions Parameters specific to image classification pipelines.
|
|
26675
|
-
* @property {number} [top_k=1] The number of top labels that will be returned by the pipeline.
|
|
26676
|
-
*
|
|
26840
|
+
* @property {number} [top_k=1] The number of top labels that will be returned by the pipeline.
|
|
26841
|
+
*
|
|
26677
26842
|
* @callback ImageClassificationPipelineCallback Assign labels to the image(s) passed as inputs.
|
|
26678
26843
|
* @param {ImagePipelineInputs} images The input images(s) to be classified.
|
|
26679
26844
|
* @param {ImageClassificationPipelineOptions} [options] The options to use for image classification.
|
|
26680
26845
|
* @returns {Promise<ImageClassificationOutput|ImageClassificationOutput[]>} An array or object containing the predicted labels and scores.
|
|
26681
|
-
*
|
|
26846
|
+
*
|
|
26682
26847
|
* @typedef {ImagePipelineConstructorArgs & ImageClassificationPipelineCallback & Disposable} ImageClassificationPipelineType
|
|
26683
26848
|
*/
|
|
26684
26849
|
|
|
26685
26850
|
/**
|
|
26686
26851
|
* Image classification pipeline using any `AutoModelForImageClassification`.
|
|
26687
26852
|
* This pipeline predicts the class of an image.
|
|
26688
|
-
*
|
|
26853
|
+
*
|
|
26689
26854
|
* **Example:** Classify an image.
|
|
26690
26855
|
* ```javascript
|
|
26691
26856
|
* const classifier = await pipeline('image-classification', 'Xenova/vit-base-patch16-224');
|
|
@@ -26695,7 +26860,7 @@ class ImageToTextPipeline extends (/** @type {new (options: TextImagePipelineCon
|
|
|
26695
26860
|
* // { label: 'tiger, Panthera tigris', score: 0.632695734500885 },
|
|
26696
26861
|
* // ]
|
|
26697
26862
|
* ```
|
|
26698
|
-
*
|
|
26863
|
+
*
|
|
26699
26864
|
* **Example:** Classify an image and return top `n` classes.
|
|
26700
26865
|
* ```javascript
|
|
26701
26866
|
* const classifier = await pipeline('image-classification', 'Xenova/vit-base-patch16-224');
|
|
@@ -26707,7 +26872,7 @@ class ImageToTextPipeline extends (/** @type {new (options: TextImagePipelineCon
|
|
|
26707
26872
|
* // { label: 'lion, king of beasts, Panthera leo', score: 0.00045060308184474707 },
|
|
26708
26873
|
* // ]
|
|
26709
26874
|
* ```
|
|
26710
|
-
*
|
|
26875
|
+
*
|
|
26711
26876
|
* **Example:** Classify an image and return all classes.
|
|
26712
26877
|
* ```javascript
|
|
26713
26878
|
* const classifier = await pipeline('image-classification', 'Xenova/vit-base-patch16-224');
|
|
@@ -26774,7 +26939,7 @@ class ImageClassificationPipeline extends (/** @type {new (options: ImagePipelin
|
|
|
26774
26939
|
* @property {string|null} label The label of the segment.
|
|
26775
26940
|
* @property {number|null} score The score of the segment.
|
|
26776
26941
|
* @property {RawImage} mask The mask of the segment.
|
|
26777
|
-
*
|
|
26942
|
+
*
|
|
26778
26943
|
* @typedef {Object} ImageSegmentationPipelineOptions Parameters specific to image segmentation pipelines.
|
|
26779
26944
|
* @property {number} [threshold=0.5] Probability threshold to filter out predicted masks.
|
|
26780
26945
|
* @property {number} [mask_threshold=0.5] Threshold to use when turning the predicted masks into binary values.
|
|
@@ -26783,19 +26948,19 @@ class ImageClassificationPipeline extends (/** @type {new (options: ImagePipelin
|
|
|
26783
26948
|
* depending on model capabilities. If not set, the pipeline will attempt to resolve (in that order).
|
|
26784
26949
|
* @property {number[]} [label_ids_to_fuse=null] List of label ids to fuse. If not set, do not fuse any labels.
|
|
26785
26950
|
* @property {number[][]} [target_sizes=null] List of target sizes for the input images. If not set, use the original image sizes.
|
|
26786
|
-
*
|
|
26951
|
+
*
|
|
26787
26952
|
* @callback ImageSegmentationPipelineCallback Segment the input images.
|
|
26788
26953
|
* @param {ImagePipelineInputs} images The input images.
|
|
26789
26954
|
* @param {ImageSegmentationPipelineOptions} [options] The options to use for image segmentation.
|
|
26790
26955
|
* @returns {Promise<ImageSegmentationPipelineOutput[]>} The annotated segments.
|
|
26791
|
-
*
|
|
26956
|
+
*
|
|
26792
26957
|
* @typedef {ImagePipelineConstructorArgs & ImageSegmentationPipelineCallback & Disposable} ImageSegmentationPipelineType
|
|
26793
26958
|
*/
|
|
26794
26959
|
|
|
26795
26960
|
/**
|
|
26796
26961
|
* Image segmentation pipeline using any `AutoModelForXXXSegmentation`.
|
|
26797
26962
|
* This pipeline predicts masks of objects and their classes.
|
|
26798
|
-
*
|
|
26963
|
+
*
|
|
26799
26964
|
* **Example:** Perform image segmentation with `Xenova/detr-resnet-50-panoptic`.
|
|
26800
26965
|
* ```javascript
|
|
26801
26966
|
* const segmenter = await pipeline('image-segmentation', 'Xenova/detr-resnet-50-panoptic');
|
|
@@ -26879,12 +27044,17 @@ class ImageSegmentationPipeline extends (/** @type {new (options: ImagePipelineC
|
|
|
26879
27044
|
/** @type {ImageSegmentationPipelineOutput[]} */
|
|
26880
27045
|
const annotation = [];
|
|
26881
27046
|
if (!subtask) {
|
|
27047
|
+
// We define an epsilon to safeguard against numerical/precision issues when detecting
|
|
27048
|
+
// the normalization mode of the output (i.e., sigmoid already applied, or not).
|
|
27049
|
+
// See https://github.com/microsoft/onnxruntime/issues/23943 for more information.
|
|
27050
|
+
const epsilon = 1e-5;
|
|
27051
|
+
|
|
26882
27052
|
// Perform standard image segmentation
|
|
26883
27053
|
const result = output[outputNames[0]];
|
|
26884
27054
|
for (let i = 0; i < imageSizes.length; ++i) {
|
|
26885
27055
|
const size = imageSizes[i];
|
|
26886
27056
|
const item = result[i];
|
|
26887
|
-
if (item.data.some(x => x <
|
|
27057
|
+
if (item.data.some(x => x < -epsilon || x > 1 + epsilon)) {
|
|
26888
27058
|
item.sigmoid_();
|
|
26889
27059
|
}
|
|
26890
27060
|
const mask = await _utils_image_js__WEBPACK_IMPORTED_MODULE_9__.RawImage.fromTensor(item.mul_(255).to('uint8')).resize(size[1], size[0]);
|
|
@@ -26953,19 +27123,19 @@ class ImageSegmentationPipeline extends (/** @type {new (options: ImagePipelineC
|
|
|
26953
27123
|
|
|
26954
27124
|
/**
|
|
26955
27125
|
* @typedef {Object} BackgroundRemovalPipelineOptions Parameters specific to image segmentation pipelines.
|
|
26956
|
-
*
|
|
27126
|
+
*
|
|
26957
27127
|
* @callback BackgroundRemovalPipelineCallback Segment the input images.
|
|
26958
27128
|
* @param {ImagePipelineInputs} images The input images.
|
|
26959
27129
|
* @param {BackgroundRemovalPipelineOptions} [options] The options to use for image segmentation.
|
|
26960
27130
|
* @returns {Promise<RawImage[]>} The images with the background removed.
|
|
26961
|
-
*
|
|
27131
|
+
*
|
|
26962
27132
|
* @typedef {ImagePipelineConstructorArgs & BackgroundRemovalPipelineCallback & Disposable} BackgroundRemovalPipelineType
|
|
26963
27133
|
*/
|
|
26964
27134
|
|
|
26965
27135
|
/**
|
|
26966
27136
|
* Background removal pipeline using certain `AutoModelForXXXSegmentation`.
|
|
26967
27137
|
* This pipeline removes the backgrounds of images.
|
|
26968
|
-
*
|
|
27138
|
+
*
|
|
26969
27139
|
* **Example:** Perform background removal with `Xenova/modnet`.
|
|
26970
27140
|
* ```javascript
|
|
26971
27141
|
* const segmenter = await pipeline('background-removal', 'Xenova/modnet');
|
|
@@ -26976,7 +27146,7 @@ class ImageSegmentationPipeline extends (/** @type {new (options: ImagePipelineC
|
|
|
26976
27146
|
* // ]
|
|
26977
27147
|
* ```
|
|
26978
27148
|
*/
|
|
26979
|
-
class BackgroundRemovalPipeline extends (/** @type {new (options: ImagePipelineConstructorArgs) =>
|
|
27149
|
+
class BackgroundRemovalPipeline extends (/** @type {new (options: ImagePipelineConstructorArgs) => BackgroundRemovalPipelineType} */ (/** @type {any} */(ImageSegmentationPipeline))) {
|
|
26980
27150
|
/**
|
|
26981
27151
|
* Create a new BackgroundRemovalPipeline.
|
|
26982
27152
|
* @param {ImagePipelineConstructorArgs} options An object used to instantiate the pipeline.
|
|
@@ -27011,25 +27181,25 @@ class BackgroundRemovalPipeline extends (/** @type {new (options: ImagePipelineC
|
|
|
27011
27181
|
* @typedef {Object} ZeroShotImageClassificationOutput
|
|
27012
27182
|
* @property {string} label The label identified by the model. It is one of the suggested `candidate_label`.
|
|
27013
27183
|
* @property {number} score The score attributed by the model for that label (between 0 and 1).
|
|
27014
|
-
*
|
|
27184
|
+
*
|
|
27015
27185
|
* @typedef {Object} ZeroShotImageClassificationPipelineOptions Parameters specific to zero-shot image classification pipelines.
|
|
27016
27186
|
* @property {string} [hypothesis_template="This is a photo of {}"] The sentence used in conjunction with `candidate_labels`
|
|
27017
27187
|
* to attempt the image classification by replacing the placeholder with the candidate_labels.
|
|
27018
27188
|
* Then likelihood is estimated by using `logits_per_image`.
|
|
27019
|
-
*
|
|
27189
|
+
*
|
|
27020
27190
|
* @callback ZeroShotImageClassificationPipelineCallback Assign labels to the image(s) passed as inputs.
|
|
27021
27191
|
* @param {ImagePipelineInputs} images The input images.
|
|
27022
27192
|
* @param {string[]} candidate_labels The candidate labels for this image.
|
|
27023
27193
|
* @param {ZeroShotImageClassificationPipelineOptions} [options] The options to use for zero-shot image classification.
|
|
27024
27194
|
* @returns {Promise<ZeroShotImageClassificationOutput[]|ZeroShotImageClassificationOutput[][]>} An array of objects containing the predicted labels and scores.
|
|
27025
|
-
*
|
|
27195
|
+
*
|
|
27026
27196
|
* @typedef {TextImagePipelineConstructorArgs & ZeroShotImageClassificationPipelineCallback & Disposable} ZeroShotImageClassificationPipelineType
|
|
27027
27197
|
*/
|
|
27028
27198
|
|
|
27029
27199
|
/**
|
|
27030
27200
|
* Zero shot image classification pipeline. This pipeline predicts the class of
|
|
27031
27201
|
* an image when you provide an image and a set of `candidate_labels`.
|
|
27032
|
-
*
|
|
27202
|
+
*
|
|
27033
27203
|
* **Example:** Zero shot image classification w/ `Xenova/clip-vit-base-patch32`.
|
|
27034
27204
|
* ```javascript
|
|
27035
27205
|
* const classifier = await pipeline('zero-shot-image-classification', 'Xenova/clip-vit-base-patch32');
|
|
@@ -27059,7 +27229,7 @@ class ZeroShotImageClassificationPipeline extends (/** @type {new (options: Text
|
|
|
27059
27229
|
const isBatched = Array.isArray(images);
|
|
27060
27230
|
const preparedImages = await prepareImages(images);
|
|
27061
27231
|
|
|
27062
|
-
// Insert label into hypothesis template
|
|
27232
|
+
// Insert label into hypothesis template
|
|
27063
27233
|
const texts = candidate_labels.map(
|
|
27064
27234
|
x => hypothesis_template.replace('{}', x)
|
|
27065
27235
|
);
|
|
@@ -27106,23 +27276,23 @@ class ZeroShotImageClassificationPipeline extends (/** @type {new (options: Text
|
|
|
27106
27276
|
* @property {number} score The score attributed by the model for that label.
|
|
27107
27277
|
* @property {BoundingBox} box The bounding box of detected object in image's original size, or as a percentage if `percentage` is set to true.
|
|
27108
27278
|
* @typedef {ObjectDetectionPipelineSingle[]} ObjectDetectionPipelineOutput
|
|
27109
|
-
*
|
|
27279
|
+
*
|
|
27110
27280
|
* @typedef {Object} ObjectDetectionPipelineOptions Parameters specific to object detection pipelines.
|
|
27111
27281
|
* @property {number} [threshold=0.9] The threshold used to filter boxes by score.
|
|
27112
27282
|
* @property {boolean} [percentage=false] Whether to return the boxes coordinates in percentage (true) or in pixels (false).
|
|
27113
|
-
*
|
|
27283
|
+
*
|
|
27114
27284
|
* @callback ObjectDetectionPipelineCallback Detect objects (bounding boxes & classes) in the image(s) passed as inputs.
|
|
27115
27285
|
* @param {ImagePipelineInputs} images The input images.
|
|
27116
27286
|
* @param {ObjectDetectionPipelineOptions} [options] The options to use for object detection.
|
|
27117
|
-
* @returns {Promise<ObjectDetectionPipelineOutput|ObjectDetectionPipelineOutput[]>} A list of objects or a list of list of objects.
|
|
27118
|
-
*
|
|
27287
|
+
* @returns {Promise<ObjectDetectionPipelineOutput|ObjectDetectionPipelineOutput[]>} A list of objects or a list of list of objects.
|
|
27288
|
+
*
|
|
27119
27289
|
* @typedef {ImagePipelineConstructorArgs & ObjectDetectionPipelineCallback & Disposable} ObjectDetectionPipelineType
|
|
27120
27290
|
*/
|
|
27121
27291
|
|
|
27122
27292
|
/**
|
|
27123
27293
|
* Object detection pipeline using any `AutoModelForObjectDetection`.
|
|
27124
27294
|
* This pipeline predicts bounding boxes of objects and their classes.
|
|
27125
|
-
*
|
|
27295
|
+
*
|
|
27126
27296
|
* **Example:** Run object-detection with `Xenova/detr-resnet-50`.
|
|
27127
27297
|
* ```javascript
|
|
27128
27298
|
* const detector = await pipeline('object-detection', 'Xenova/detr-resnet-50');
|
|
@@ -27196,27 +27366,27 @@ class ObjectDetectionPipeline extends (/** @type {new (options: ImagePipelineCon
|
|
|
27196
27366
|
* @property {string} label Text query corresponding to the found object.
|
|
27197
27367
|
* @property {number} score Score corresponding to the object (between 0 and 1).
|
|
27198
27368
|
* @property {BoundingBox} box Bounding box of the detected object in image's original size, or as a percentage if `percentage` is set to true.
|
|
27199
|
-
*
|
|
27369
|
+
*
|
|
27200
27370
|
* @typedef {Object} ZeroShotObjectDetectionPipelineOptions Parameters specific to zero-shot object detection pipelines.
|
|
27201
27371
|
* @property {number} [threshold=0.1] The probability necessary to make a prediction.
|
|
27202
27372
|
* @property {number} [top_k=null] The number of top predictions that will be returned by the pipeline.
|
|
27203
27373
|
* If the provided number is `null` or higher than the number of predictions available, it will default
|
|
27204
27374
|
* to the number of predictions.
|
|
27205
27375
|
* @property {boolean} [percentage=false] Whether to return the boxes coordinates in percentage (true) or in pixels (false).
|
|
27206
|
-
*
|
|
27376
|
+
*
|
|
27207
27377
|
* @callback ZeroShotObjectDetectionPipelineCallback Detect objects (bounding boxes & classes) in the image(s) passed as inputs.
|
|
27208
27378
|
* @param {ImagePipelineInputs} images The input images.
|
|
27209
27379
|
* @param {string[]} candidate_labels What the model should recognize in the image.
|
|
27210
27380
|
* @param {ZeroShotObjectDetectionPipelineOptions} [options] The options to use for zero-shot object detection.
|
|
27211
27381
|
* @returns {Promise<ZeroShotObjectDetectionOutput[]|ZeroShotObjectDetectionOutput[][]>} An array of objects containing the predicted labels, scores, and bounding boxes.
|
|
27212
|
-
*
|
|
27382
|
+
*
|
|
27213
27383
|
* @typedef {TextImagePipelineConstructorArgs & ZeroShotObjectDetectionPipelineCallback & Disposable} ZeroShotObjectDetectionPipelineType
|
|
27214
27384
|
*/
|
|
27215
27385
|
|
|
27216
27386
|
/**
|
|
27217
27387
|
* Zero-shot object detection pipeline. This pipeline predicts bounding boxes of
|
|
27218
27388
|
* objects when you provide an image and a set of `candidate_labels`.
|
|
27219
|
-
*
|
|
27389
|
+
*
|
|
27220
27390
|
* **Example:** Zero-shot object detection w/ `Xenova/owlvit-base-patch32`.
|
|
27221
27391
|
* ```javascript
|
|
27222
27392
|
* const detector = await pipeline('zero-shot-object-detection', 'Xenova/owlvit-base-patch32');
|
|
@@ -27246,7 +27416,7 @@ class ObjectDetectionPipeline extends (/** @type {new (options: ImagePipelineCon
|
|
|
27246
27416
|
* // }
|
|
27247
27417
|
* // ]
|
|
27248
27418
|
* ```
|
|
27249
|
-
*
|
|
27419
|
+
*
|
|
27250
27420
|
* **Example:** Zero-shot object detection w/ `Xenova/owlvit-base-patch32` (returning top 4 matches and setting a threshold).
|
|
27251
27421
|
* ```javascript
|
|
27252
27422
|
* const detector = await pipeline('zero-shot-object-detection', 'Xenova/owlvit-base-patch32');
|
|
@@ -27361,13 +27531,13 @@ class ZeroShotObjectDetectionPipeline extends (/** @type {new (options: TextImag
|
|
|
27361
27531
|
* @typedef {Object} DocumentQuestionAnsweringSingle
|
|
27362
27532
|
* @property {string} answer The generated text.
|
|
27363
27533
|
* @typedef {DocumentQuestionAnsweringSingle[]} DocumentQuestionAnsweringOutput
|
|
27364
|
-
*
|
|
27534
|
+
*
|
|
27365
27535
|
* @callback DocumentQuestionAnsweringPipelineCallback Answer the question given as input by using the document.
|
|
27366
27536
|
* @param {ImageInput} image The image of the document to use.
|
|
27367
27537
|
* @param {string} question A question to ask of the document.
|
|
27368
27538
|
* @param {Partial<import('./generation/configuration_utils.js').GenerationConfig>} [options] Additional keyword arguments to pass along to the generate method of the model.
|
|
27369
27539
|
* @returns {Promise<DocumentQuestionAnsweringOutput|DocumentQuestionAnsweringOutput[]>} An object (or array of objects) containing the answer(s).
|
|
27370
|
-
*
|
|
27540
|
+
*
|
|
27371
27541
|
* @typedef {TextImagePipelineConstructorArgs & DocumentQuestionAnsweringPipelineCallback & Disposable} DocumentQuestionAnsweringPipelineType
|
|
27372
27542
|
*/
|
|
27373
27543
|
|
|
@@ -27375,7 +27545,7 @@ class ZeroShotObjectDetectionPipeline extends (/** @type {new (options: TextImag
|
|
|
27375
27545
|
* Document Question Answering pipeline using any `AutoModelForDocumentQuestionAnswering`.
|
|
27376
27546
|
* The inputs/outputs are similar to the (extractive) question answering pipeline; however,
|
|
27377
27547
|
* the pipeline takes an image (and optional OCR'd words/boxes) as input instead of text context.
|
|
27378
|
-
*
|
|
27548
|
+
*
|
|
27379
27549
|
* **Example:** Answer questions about a document with `Xenova/donut-base-finetuned-docvqa`.
|
|
27380
27550
|
* ```javascript
|
|
27381
27551
|
* const qa_pipeline = await pipeline('document-question-answering', 'Xenova/donut-base-finetuned-docvqa');
|
|
@@ -27445,22 +27615,22 @@ class DocumentQuestionAnsweringPipeline extends (/** @type {new (options: TextIm
|
|
|
27445
27615
|
* @typedef {Object} TextToAudioOutput
|
|
27446
27616
|
* @property {Float32Array} audio The generated audio waveform.
|
|
27447
27617
|
* @property {number} sampling_rate The sampling rate of the generated audio waveform.
|
|
27448
|
-
*
|
|
27618
|
+
*
|
|
27449
27619
|
* @typedef {Object} TextToAudioPipelineOptions Parameters specific to text-to-audio pipelines.
|
|
27450
27620
|
* @property {Tensor|Float32Array|string|URL} [speaker_embeddings=null] The speaker embeddings (if the model requires it).
|
|
27451
|
-
*
|
|
27621
|
+
*
|
|
27452
27622
|
* @callback TextToAudioPipelineCallback Generates speech/audio from the inputs.
|
|
27453
27623
|
* @param {string|string[]} texts The text(s) to generate.
|
|
27454
27624
|
* @param {TextToAudioPipelineOptions} options Parameters passed to the model generation/forward method.
|
|
27455
27625
|
* @returns {Promise<TextToAudioOutput>} An object containing the generated audio and sampling rate.
|
|
27456
|
-
*
|
|
27626
|
+
*
|
|
27457
27627
|
* @typedef {TextToAudioPipelineConstructorArgs & TextToAudioPipelineCallback & Disposable} TextToAudioPipelineType
|
|
27458
27628
|
*/
|
|
27459
27629
|
|
|
27460
27630
|
/**
|
|
27461
27631
|
* Text-to-audio generation pipeline using any `AutoModelForTextToWaveform` or `AutoModelForTextToSpectrogram`.
|
|
27462
27632
|
* This pipeline generates an audio file from an input text and optional other conditional inputs.
|
|
27463
|
-
*
|
|
27633
|
+
*
|
|
27464
27634
|
* **Example:** Generate audio from text with `Xenova/speecht5_tts`.
|
|
27465
27635
|
* ```javascript
|
|
27466
27636
|
* const synthesizer = await pipeline('text-to-speech', 'Xenova/speecht5_tts', { quantized: false });
|
|
@@ -27471,17 +27641,17 @@ class DocumentQuestionAnsweringPipeline extends (/** @type {new (options: TextIm
|
|
|
27471
27641
|
* // sampling_rate: 16000
|
|
27472
27642
|
* // }
|
|
27473
27643
|
* ```
|
|
27474
|
-
*
|
|
27644
|
+
*
|
|
27475
27645
|
* You can then save the audio to a .wav file with the `wavefile` package:
|
|
27476
27646
|
* ```javascript
|
|
27477
27647
|
* import wavefile from 'wavefile';
|
|
27478
27648
|
* import fs from 'fs';
|
|
27479
|
-
*
|
|
27649
|
+
*
|
|
27480
27650
|
* const wav = new wavefile.WaveFile();
|
|
27481
27651
|
* wav.fromScratch(1, out.sampling_rate, '32f', out.audio);
|
|
27482
27652
|
* fs.writeFileSync('out.wav', wav.toBuffer());
|
|
27483
27653
|
* ```
|
|
27484
|
-
*
|
|
27654
|
+
*
|
|
27485
27655
|
* **Example:** Multilingual speech generation with `Xenova/mms-tts-fra`. See [here](https://huggingface.co/models?pipeline_tag=text-to-speech&other=vits&sort=trending) for the full list of available languages (1107).
|
|
27486
27656
|
* ```javascript
|
|
27487
27657
|
* const synthesizer = await pipeline('text-to-speech', 'Xenova/mms-tts-fra');
|
|
@@ -27587,13 +27757,13 @@ class TextToAudioPipeline extends (/** @type {new (options: TextToAudioPipelineC
|
|
|
27587
27757
|
* @callback ImageToImagePipelineCallback Transform the image(s) passed as inputs.
|
|
27588
27758
|
* @param {ImagePipelineInputs} images The images to transform.
|
|
27589
27759
|
* @returns {Promise<RawImage|RawImage[]>} The transformed image or list of images.
|
|
27590
|
-
*
|
|
27760
|
+
*
|
|
27591
27761
|
* @typedef {ImagePipelineConstructorArgs & ImageToImagePipelineCallback & Disposable} ImageToImagePipelineType
|
|
27592
27762
|
*/
|
|
27593
27763
|
|
|
27594
27764
|
/**
|
|
27595
27765
|
* Image to Image pipeline using any `AutoModelForImageToImage`. This pipeline generates an image based on a previous image input.
|
|
27596
|
-
*
|
|
27766
|
+
*
|
|
27597
27767
|
* **Example:** Super-resolution w/ `Xenova/swin2SR-classical-sr-x2-64`
|
|
27598
27768
|
* ```javascript
|
|
27599
27769
|
* const upscaler = await pipeline('image-to-image', 'Xenova/swin2SR-classical-sr-x2-64');
|
|
@@ -27638,17 +27808,17 @@ class ImageToImagePipeline extends (/** @type {new (options: ImagePipelineConstr
|
|
|
27638
27808
|
* @typedef {Object} DepthEstimationPipelineOutput
|
|
27639
27809
|
* @property {Tensor} predicted_depth The raw depth map predicted by the model.
|
|
27640
27810
|
* @property {RawImage} depth The processed depth map as an image (with the same size as the input image).
|
|
27641
|
-
*
|
|
27811
|
+
*
|
|
27642
27812
|
* @callback DepthEstimationPipelineCallback Predicts the depth for the image(s) passed as inputs.
|
|
27643
27813
|
* @param {ImagePipelineInputs} images The images to compute depth for.
|
|
27644
27814
|
* @returns {Promise<DepthEstimationPipelineOutput|DepthEstimationPipelineOutput[]>} An image or a list of images containing result(s).
|
|
27645
|
-
*
|
|
27815
|
+
*
|
|
27646
27816
|
* @typedef {ImagePipelineConstructorArgs & DepthEstimationPipelineCallback & Disposable} DepthEstimationPipelineType
|
|
27647
27817
|
*/
|
|
27648
27818
|
|
|
27649
27819
|
/**
|
|
27650
27820
|
* Depth estimation pipeline using any `AutoModelForDepthEstimation`. This pipeline predicts the depth of an image.
|
|
27651
|
-
*
|
|
27821
|
+
*
|
|
27652
27822
|
* **Example:** Depth estimation w/ `Xenova/dpt-hybrid-midas`
|
|
27653
27823
|
* ```javascript
|
|
27654
27824
|
* const depth_estimator = await pipeline('depth-estimation', 'Xenova/dpt-hybrid-midas');
|
|
@@ -28033,7 +28203,7 @@ const TASK_ALIASES = Object.freeze({
|
|
|
28033
28203
|
|
|
28034
28204
|
/**
|
|
28035
28205
|
* Utility factory method to build a `Pipeline` object.
|
|
28036
|
-
*
|
|
28206
|
+
*
|
|
28037
28207
|
* @template {PipelineType} T The type of pipeline to return.
|
|
28038
28208
|
* @param {T} task The task defining which pipeline will be returned. Currently accepted tasks are:
|
|
28039
28209
|
* - `"audio-classification"`: will return a `AudioClassificationPipeline`.
|
|
@@ -30851,13 +31021,12 @@ class PreTrainedTokenizer extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_0__
|
|
|
30851
31021
|
this.decoder.end_of_word_suffix = this.model.end_of_word_suffix;
|
|
30852
31022
|
}
|
|
30853
31023
|
|
|
30854
|
-
this.
|
|
30855
|
-
this.added_tokens.
|
|
30856
|
-
|
|
30857
|
-
|
|
30858
|
-
|
|
30859
|
-
|
|
30860
|
-
) : null;
|
|
31024
|
+
this.added_tokens_splitter = new _utils_data_structures_js__WEBPACK_IMPORTED_MODULE_5__.DictionarySplitter(
|
|
31025
|
+
this.added_tokens.map(x => x.content),
|
|
31026
|
+
);
|
|
31027
|
+
|
|
31028
|
+
/** @type {Map<string, AddedToken>} */
|
|
31029
|
+
this.added_tokens_map = new Map(this.added_tokens.map(x => [x.content, x]))
|
|
30861
31030
|
|
|
30862
31031
|
// Set mask token if present (otherwise will be undefined, which is fine)
|
|
30863
31032
|
this.mask_token = this.getToken('mask_token');
|
|
@@ -31152,40 +31321,50 @@ class PreTrainedTokenizer extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_0__
|
|
|
31152
31321
|
// Actual function which does encoding, for a single text
|
|
31153
31322
|
// First, we take care of special tokens. Needed to avoid issues arising from
|
|
31154
31323
|
// normalization and/or pretokenization (which may not preserve special tokens)
|
|
31155
|
-
const sections = this.
|
|
31156
|
-
|
|
31157
|
-
|
|
31158
|
-
|
|
31159
|
-
|
|
31160
|
-
|
|
31161
|
-
|
|
31162
|
-
|
|
31163
|
-
if (this.remove_space === true) {
|
|
31164
|
-
x = x.trim().split(/\s+/).join(' ');
|
|
31165
|
-
}
|
|
31166
|
-
if (this.do_lowercase_and_remove_accent) {
|
|
31167
|
-
x = lowercase_and_remove_accent(x);
|
|
31324
|
+
const sections = this.added_tokens_splitter.split(text);
|
|
31325
|
+
|
|
31326
|
+
// Process left/right stripping of added tokens
|
|
31327
|
+
for (let i = 0; i < sections.length; ++i) {
|
|
31328
|
+
const addedToken = this.added_tokens_map.get(sections[i]);
|
|
31329
|
+
if (addedToken) {
|
|
31330
|
+
if (addedToken.lstrip && i > 0) {
|
|
31331
|
+
sections[i - 1] = sections[i - 1].trimEnd();
|
|
31168
31332
|
}
|
|
31169
|
-
|
|
31170
|
-
|
|
31171
|
-
x = this.normalizer(x);
|
|
31333
|
+
if (addedToken.rstrip && i < sections.length - 1) {
|
|
31334
|
+
sections[i + 1] = sections[i + 1].trimStart();
|
|
31172
31335
|
}
|
|
31336
|
+
}
|
|
31337
|
+
}
|
|
31173
31338
|
|
|
31174
|
-
|
|
31175
|
-
|
|
31176
|
-
|
|
31177
|
-
return [];
|
|
31178
|
-
}
|
|
31339
|
+
const tokens = sections.flatMap((x, section_index) => {
|
|
31340
|
+
if (x.length === 0) return [];
|
|
31341
|
+
if (this.added_tokens_map.has(x)) return [x]; // Return added tokens unchanged
|
|
31179
31342
|
|
|
31180
|
-
|
|
31181
|
-
|
|
31182
|
-
|
|
31343
|
+
if (this.remove_space === true) {
|
|
31344
|
+
x = x.trim().split(/\s+/).join(' ');
|
|
31345
|
+
}
|
|
31346
|
+
if (this.do_lowercase_and_remove_accent) {
|
|
31347
|
+
x = lowercase_and_remove_accent(x);
|
|
31348
|
+
}
|
|
31183
31349
|
|
|
31184
|
-
|
|
31350
|
+
if (this.normalizer !== null) {
|
|
31351
|
+
x = this.normalizer(x);
|
|
31352
|
+
}
|
|
31185
31353
|
|
|
31186
|
-
|
|
31354
|
+
// If, after normalization, this section is empty (e.g., trimming whitespace),
|
|
31355
|
+
// we return an empty array
|
|
31356
|
+
if (x.length === 0) {
|
|
31357
|
+
return [];
|
|
31187
31358
|
}
|
|
31188
|
-
|
|
31359
|
+
|
|
31360
|
+
const sectionTokens = (this.pre_tokenizer !== null) ? this.pre_tokenizer(x, {
|
|
31361
|
+
section_index,
|
|
31362
|
+
}) : [x];
|
|
31363
|
+
|
|
31364
|
+
const tokens = this.model(sectionTokens);
|
|
31365
|
+
|
|
31366
|
+
return tokens;
|
|
31367
|
+
});
|
|
31189
31368
|
|
|
31190
31369
|
return tokens;
|
|
31191
31370
|
}
|
|
@@ -33810,6 +33989,7 @@ function count(arr, value) {
|
|
|
33810
33989
|
__webpack_require__.r(__webpack_exports__);
|
|
33811
33990
|
/* harmony export */ __webpack_require__.d(__webpack_exports__, {
|
|
33812
33991
|
/* harmony export */ CharTrie: () => (/* binding */ CharTrie),
|
|
33992
|
+
/* harmony export */ DictionarySplitter: () => (/* binding */ DictionarySplitter),
|
|
33813
33993
|
/* harmony export */ PriorityQueue: () => (/* binding */ PriorityQueue),
|
|
33814
33994
|
/* harmony export */ TokenLattice: () => (/* binding */ TokenLattice)
|
|
33815
33995
|
/* harmony export */ });
|
|
@@ -34261,6 +34441,80 @@ class TokenLatticeNode {
|
|
|
34261
34441
|
}
|
|
34262
34442
|
}
|
|
34263
34443
|
|
|
34444
|
+
/**
|
|
34445
|
+
* A data structure which uses a trie to split a string into tokens based on a dictionary.
|
|
34446
|
+
* It can also use a regular expression to preprocess the input text before splitting.
|
|
34447
|
+
*
|
|
34448
|
+
* NOTE: To ensure multi-byte characters are handled correctly, we operate at byte-level instead of character-level.
|
|
34449
|
+
*/
|
|
34450
|
+
class DictionarySplitter {
|
|
34451
|
+
/**
|
|
34452
|
+
* @param {string[]} dictionary The dictionary of words to use for splitting.
|
|
34453
|
+
*/
|
|
34454
|
+
constructor(dictionary) {
|
|
34455
|
+
this.trie = this._buildTrie(dictionary);
|
|
34456
|
+
}
|
|
34457
|
+
|
|
34458
|
+
/**
|
|
34459
|
+
* Builds a trie from the given dictionary.
|
|
34460
|
+
* @param {string[]} dictionary The dictionary of words to build the trie from.
|
|
34461
|
+
* @returns {Object} The root node of the trie.
|
|
34462
|
+
* @private
|
|
34463
|
+
*/
|
|
34464
|
+
_buildTrie(dictionary) {
|
|
34465
|
+
const trie = Object.create(null);
|
|
34466
|
+
for (const word of dictionary) {
|
|
34467
|
+
let node = trie;
|
|
34468
|
+
for (let i = 0; i < word.length; ++i) {
|
|
34469
|
+
node = (node[word[i]] ??= Object.create(null));
|
|
34470
|
+
}
|
|
34471
|
+
node.end = word;
|
|
34472
|
+
}
|
|
34473
|
+
return trie;
|
|
34474
|
+
}
|
|
34475
|
+
|
|
34476
|
+
/**
|
|
34477
|
+
* Splits the input text into tokens based on the dictionary.
|
|
34478
|
+
* @param {string} text The input text to split.
|
|
34479
|
+
* @returns {string[]} An array of tokens.
|
|
34480
|
+
*/
|
|
34481
|
+
split(text) {
|
|
34482
|
+
const result = [];
|
|
34483
|
+
const n = text.length;
|
|
34484
|
+
let start = 0;
|
|
34485
|
+
let i = 0;
|
|
34486
|
+
|
|
34487
|
+
while (i < n) {
|
|
34488
|
+
let node = this.trie;
|
|
34489
|
+
let match = null;
|
|
34490
|
+
let j = i;
|
|
34491
|
+
|
|
34492
|
+
while (j < n && (node = node[text[j]])) {
|
|
34493
|
+
if (node.end) {
|
|
34494
|
+
// Always keep the last (i.e., longest) match.
|
|
34495
|
+
match = node.end;
|
|
34496
|
+
}
|
|
34497
|
+
++j;
|
|
34498
|
+
}
|
|
34499
|
+
|
|
34500
|
+
if (match) {
|
|
34501
|
+
if (i > start) {
|
|
34502
|
+
result.push(text.slice(start, i));
|
|
34503
|
+
}
|
|
34504
|
+
result.push(match);
|
|
34505
|
+
i += match.length;
|
|
34506
|
+
start = i;
|
|
34507
|
+
} else {
|
|
34508
|
+
++i;
|
|
34509
|
+
}
|
|
34510
|
+
}
|
|
34511
|
+
if (start < n) {
|
|
34512
|
+
result.push(text.slice(start));
|
|
34513
|
+
}
|
|
34514
|
+
return result;
|
|
34515
|
+
}
|
|
34516
|
+
}
|
|
34517
|
+
|
|
34264
34518
|
|
|
34265
34519
|
/***/ }),
|
|
34266
34520
|
|
|
@@ -34640,6 +34894,22 @@ function isValidUrl(string, protocols = null, validHosts = null) {
|
|
|
34640
34894
|
return true;
|
|
34641
34895
|
}
|
|
34642
34896
|
|
|
34897
|
+
const REPO_ID_REGEX = /^(\b[\w\-.]+\b\/)?\b[\w\-.]{1,96}\b$/;
|
|
34898
|
+
|
|
34899
|
+
/**
|
|
34900
|
+
* Tests whether a string is a valid Hugging Face model ID or not.
|
|
34901
|
+
* Adapted from https://github.com/huggingface/huggingface_hub/blob/6378820ebb03f071988a96c7f3268f5bdf8f9449/src/huggingface_hub/utils/_validators.py#L119-L170
|
|
34902
|
+
*
|
|
34903
|
+
* @param {string} string The string to test
|
|
34904
|
+
* @returns {boolean} True if the string is a valid model ID, false otherwise.
|
|
34905
|
+
*/
|
|
34906
|
+
function isValidHfModelId(string) {
|
|
34907
|
+
if (!REPO_ID_REGEX.test(string)) return false;
|
|
34908
|
+
if (string.includes("..") || string.includes("--")) return false;
|
|
34909
|
+
if (string.endsWith(".git") || string.endsWith(".ipynb")) return false;
|
|
34910
|
+
return true;
|
|
34911
|
+
}
|
|
34912
|
+
|
|
34643
34913
|
/**
|
|
34644
34914
|
* Helper function to get a file, using either the Fetch API or FileSystem API.
|
|
34645
34915
|
*
|
|
@@ -34892,12 +35162,13 @@ async function getModelFile(path_or_repo_id, filename, fatal = true, options = {
|
|
|
34892
35162
|
}
|
|
34893
35163
|
|
|
34894
35164
|
const revision = options.revision ?? 'main';
|
|
35165
|
+
const requestURL = pathJoin(path_or_repo_id, filename);
|
|
34895
35166
|
|
|
34896
|
-
|
|
34897
|
-
|
|
34898
|
-
|
|
34899
|
-
|
|
34900
|
-
|
|
35167
|
+
const validModelId = isValidHfModelId(path_or_repo_id);
|
|
35168
|
+
const localPath = validModelId
|
|
35169
|
+
? pathJoin(_env_js__WEBPACK_IMPORTED_MODULE_2__.env.localModelPath, requestURL)
|
|
35170
|
+
: requestURL;
|
|
35171
|
+
const remoteURL = pathJoin(
|
|
34901
35172
|
_env_js__WEBPACK_IMPORTED_MODULE_2__.env.remoteHost,
|
|
34902
35173
|
_env_js__WEBPACK_IMPORTED_MODULE_2__.env.remotePathTemplate
|
|
34903
35174
|
.replaceAll('{model}', path_or_repo_id)
|
|
@@ -34905,14 +35176,14 @@ async function getModelFile(path_or_repo_id, filename, fatal = true, options = {
|
|
|
34905
35176
|
filename
|
|
34906
35177
|
);
|
|
34907
35178
|
|
|
34908
|
-
// Choose cache key for filesystem cache
|
|
34909
|
-
// When using the main revision (default), we use the request URL as the cache key.
|
|
34910
|
-
// If a specific revision is requested, we account for this in the cache key.
|
|
34911
|
-
let fsCacheKey = revision === 'main' ? requestURL : pathJoin(path_or_repo_id, revision, filename);
|
|
34912
|
-
|
|
34913
35179
|
/** @type {string} */
|
|
34914
35180
|
let cacheKey;
|
|
34915
|
-
|
|
35181
|
+
const proposedCacheKey = cache instanceof FileCache
|
|
35182
|
+
// Choose cache key for filesystem cache
|
|
35183
|
+
// When using the main revision (default), we use the request URL as the cache key.
|
|
35184
|
+
// If a specific revision is requested, we account for this in the cache key.
|
|
35185
|
+
? revision === 'main' ? requestURL : pathJoin(path_or_repo_id, revision, filename)
|
|
35186
|
+
: remoteURL;
|
|
34916
35187
|
|
|
34917
35188
|
// Whether to cache the final response in the end.
|
|
34918
35189
|
let toCacheResponse = false;
|
|
@@ -34925,11 +35196,10 @@ async function getModelFile(path_or_repo_id, filename, fatal = true, options = {
|
|
|
34925
35196
|
// 1. We first try to get from cache using the local path. In some environments (like deno),
|
|
34926
35197
|
// non-URL cache keys are not allowed. In these cases, `response` will be undefined.
|
|
34927
35198
|
// 2. If no response is found, we try to get from cache using the remote URL or file system cache.
|
|
34928
|
-
response = await tryCache(cache,
|
|
35199
|
+
response = await tryCache(cache, localPath, proposedCacheKey);
|
|
34929
35200
|
}
|
|
34930
35201
|
|
|
34931
35202
|
const cacheHit = response !== undefined;
|
|
34932
|
-
|
|
34933
35203
|
if (response === undefined) {
|
|
34934
35204
|
// Caching not available, or file is not cached, so we perform the request
|
|
34935
35205
|
|
|
@@ -34947,9 +35217,9 @@ async function getModelFile(path_or_repo_id, filename, fatal = true, options = {
|
|
|
34947
35217
|
console.warn(`Unable to load from local path "${localPath}": "${e}"`);
|
|
34948
35218
|
}
|
|
34949
35219
|
} else if (options.local_files_only) {
|
|
34950
|
-
throw new Error(`\`local_files_only=true\`, but attempted to load a remote file from: ${
|
|
35220
|
+
throw new Error(`\`local_files_only=true\`, but attempted to load a remote file from: ${requestURL}.`);
|
|
34951
35221
|
} else if (!_env_js__WEBPACK_IMPORTED_MODULE_2__.env.allowRemoteModels) {
|
|
34952
|
-
throw new Error(`\`env.allowRemoteModels=false\`, but attempted to load a remote file from: ${
|
|
35222
|
+
throw new Error(`\`env.allowRemoteModels=false\`, but attempted to load a remote file from: ${requestURL}.`);
|
|
34953
35223
|
}
|
|
34954
35224
|
}
|
|
34955
35225
|
|
|
@@ -34969,6 +35239,11 @@ async function getModelFile(path_or_repo_id, filename, fatal = true, options = {
|
|
|
34969
35239
|
return null;
|
|
34970
35240
|
}
|
|
34971
35241
|
}
|
|
35242
|
+
if (!validModelId) {
|
|
35243
|
+
// Before making any requests to the remote server, we check if the model ID is valid.
|
|
35244
|
+
// This prevents unnecessary network requests for invalid model IDs.
|
|
35245
|
+
throw Error(`Local file missing at "${localPath}" and download aborted due to invalid model ID "${path_or_repo_id}".`);
|
|
35246
|
+
}
|
|
34972
35247
|
|
|
34973
35248
|
// File not found locally, so we try to download it from the remote server
|
|
34974
35249
|
response = await getFile(remoteURL);
|
|
@@ -35299,7 +35574,7 @@ class RawImage {
|
|
|
35299
35574
|
|
|
35300
35575
|
/**
|
|
35301
35576
|
* Helper method for reading an image from a variety of input types.
|
|
35302
|
-
* @param {RawImage|string|URL} input
|
|
35577
|
+
* @param {RawImage|string|URL|Blob|HTMLCanvasElement|OffscreenCanvas} input
|
|
35303
35578
|
* @returns The image object.
|
|
35304
35579
|
*
|
|
35305
35580
|
* **Example:** Read image from a URL.
|
|
@@ -35318,6 +35593,14 @@ class RawImage {
|
|
|
35318
35593
|
return input;
|
|
35319
35594
|
} else if (typeof input === 'string' || input instanceof URL) {
|
|
35320
35595
|
return await this.fromURL(input);
|
|
35596
|
+
} else if (input instanceof Blob) {
|
|
35597
|
+
return await this.fromBlob(input);
|
|
35598
|
+
} else if (
|
|
35599
|
+
(typeof HTMLCanvasElement !== "undefined" && input instanceof HTMLCanvasElement)
|
|
35600
|
+
||
|
|
35601
|
+
(typeof OffscreenCanvas !== "undefined" && input instanceof OffscreenCanvas)
|
|
35602
|
+
) {
|
|
35603
|
+
return this.fromCanvas(input);
|
|
35321
35604
|
} else {
|
|
35322
35605
|
throw new Error(`Unsupported input type: ${typeof input}`);
|
|
35323
35606
|
}
|
|
@@ -38327,8 +38610,12 @@ function calc_unsqueeze_dims(dims, dim) {
|
|
|
38327
38610
|
* @private
|
|
38328
38611
|
*/
|
|
38329
38612
|
function safeIndex(index, size, dimension = null, boundsCheck = true) {
|
|
38330
|
-
if (
|
|
38331
|
-
|
|
38613
|
+
if (index < -size || index >= size) {
|
|
38614
|
+
if (boundsCheck) {
|
|
38615
|
+
throw new Error(`IndexError: index ${index} is out of bounds for dimension${dimension === null ? '' : ' ' + dimension} with size ${size}`);
|
|
38616
|
+
} else {
|
|
38617
|
+
return index < -size ? 0 : size;
|
|
38618
|
+
}
|
|
38332
38619
|
}
|
|
38333
38620
|
|
|
38334
38621
|
if (index < 0) {
|
|
@@ -39199,6 +39486,9 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
39199
39486
|
/* harmony export */ Gemma2ForCausalLM: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Gemma2ForCausalLM),
|
|
39200
39487
|
/* harmony export */ Gemma2Model: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Gemma2Model),
|
|
39201
39488
|
/* harmony export */ Gemma2PreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Gemma2PreTrainedModel),
|
|
39489
|
+
/* harmony export */ Gemma3ForCausalLM: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Gemma3ForCausalLM),
|
|
39490
|
+
/* harmony export */ Gemma3Model: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Gemma3Model),
|
|
39491
|
+
/* harmony export */ Gemma3PreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Gemma3PreTrainedModel),
|
|
39202
39492
|
/* harmony export */ GemmaForCausalLM: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.GemmaForCausalLM),
|
|
39203
39493
|
/* harmony export */ GemmaModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.GemmaModel),
|
|
39204
39494
|
/* harmony export */ GemmaPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.GemmaPreTrainedModel),
|
|
@@ -39300,6 +39590,10 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
39300
39590
|
/* harmony export */ MaskFormerPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.MaskFormerPreTrainedModel),
|
|
39301
39591
|
/* harmony export */ MaskedLMOutput: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.MaskedLMOutput),
|
|
39302
39592
|
/* harmony export */ MaxLengthCriteria: () => (/* reexport safe */ _generation_stopping_criteria_js__WEBPACK_IMPORTED_MODULE_20__.MaxLengthCriteria),
|
|
39593
|
+
/* harmony export */ Metric3DForDepthEstimation: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Metric3DForDepthEstimation),
|
|
39594
|
+
/* harmony export */ Metric3DPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Metric3DPreTrainedModel),
|
|
39595
|
+
/* harmony export */ Metric3Dv2ForDepthEstimation: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Metric3Dv2ForDepthEstimation),
|
|
39596
|
+
/* harmony export */ Metric3Dv2PreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Metric3Dv2PreTrainedModel),
|
|
39303
39597
|
/* harmony export */ MgpstrForSceneTextRecognition: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.MgpstrForSceneTextRecognition),
|
|
39304
39598
|
/* harmony export */ MgpstrModelOutput: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.MgpstrModelOutput),
|
|
39305
39599
|
/* harmony export */ MgpstrPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.MgpstrPreTrainedModel),
|
|
@@ -39452,11 +39746,19 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
39452
39746
|
/* harmony export */ Qwen2VLImageProcessor: () => (/* reexport safe */ _models_image_processors_js__WEBPACK_IMPORTED_MODULE_14__.Qwen2VLImageProcessor),
|
|
39453
39747
|
/* harmony export */ Qwen2VLPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Qwen2VLPreTrainedModel),
|
|
39454
39748
|
/* harmony export */ Qwen2VLProcessor: () => (/* reexport safe */ _models_processors_js__WEBPACK_IMPORTED_MODULE_17__.Qwen2VLProcessor),
|
|
39749
|
+
/* harmony export */ RFDetrForObjectDetection: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.RFDetrForObjectDetection),
|
|
39750
|
+
/* harmony export */ RFDetrModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.RFDetrModel),
|
|
39751
|
+
/* harmony export */ RFDetrObjectDetectionOutput: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.RFDetrObjectDetectionOutput),
|
|
39752
|
+
/* harmony export */ RFDetrPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.RFDetrPreTrainedModel),
|
|
39455
39753
|
/* harmony export */ RTDetrForObjectDetection: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.RTDetrForObjectDetection),
|
|
39456
39754
|
/* harmony export */ RTDetrImageProcessor: () => (/* reexport safe */ _models_image_processors_js__WEBPACK_IMPORTED_MODULE_14__.RTDetrImageProcessor),
|
|
39457
39755
|
/* harmony export */ RTDetrModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.RTDetrModel),
|
|
39458
39756
|
/* harmony export */ RTDetrObjectDetectionOutput: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.RTDetrObjectDetectionOutput),
|
|
39459
39757
|
/* harmony export */ RTDetrPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.RTDetrPreTrainedModel),
|
|
39758
|
+
/* harmony export */ RTDetrV2ForObjectDetection: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.RTDetrV2ForObjectDetection),
|
|
39759
|
+
/* harmony export */ RTDetrV2Model: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.RTDetrV2Model),
|
|
39760
|
+
/* harmony export */ RTDetrV2ObjectDetectionOutput: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.RTDetrV2ObjectDetectionOutput),
|
|
39761
|
+
/* harmony export */ RTDetrV2PreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.RTDetrV2PreTrainedModel),
|
|
39460
39762
|
/* harmony export */ RawAudio: () => (/* reexport safe */ _utils_audio_js__WEBPACK_IMPORTED_MODULE_5__.RawAudio),
|
|
39461
39763
|
/* harmony export */ RawImage: () => (/* reexport safe */ _utils_image_js__WEBPACK_IMPORTED_MODULE_6__.RawImage),
|
|
39462
39764
|
/* harmony export */ RawVideo: () => (/* reexport safe */ _utils_video_js__WEBPACK_IMPORTED_MODULE_7__.RawVideo),
|
|
@@ -39506,6 +39808,11 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
39506
39808
|
/* harmony export */ SmolVLMForConditionalGeneration: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.SmolVLMForConditionalGeneration),
|
|
39507
39809
|
/* harmony export */ SmolVLMImageProcessor: () => (/* reexport safe */ _models_image_processors_js__WEBPACK_IMPORTED_MODULE_14__.SmolVLMImageProcessor),
|
|
39508
39810
|
/* harmony export */ SmolVLMProcessor: () => (/* reexport safe */ _models_processors_js__WEBPACK_IMPORTED_MODULE_17__.SmolVLMProcessor),
|
|
39811
|
+
/* harmony export */ SnacDecoderModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.SnacDecoderModel),
|
|
39812
|
+
/* harmony export */ SnacEncoderModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.SnacEncoderModel),
|
|
39813
|
+
/* harmony export */ SnacFeatureExtractor: () => (/* reexport safe */ _models_feature_extractors_js__WEBPACK_IMPORTED_MODULE_11__.SnacFeatureExtractor),
|
|
39814
|
+
/* harmony export */ SnacModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.SnacModel),
|
|
39815
|
+
/* harmony export */ SnacPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.SnacPreTrainedModel),
|
|
39509
39816
|
/* harmony export */ SpeechT5FeatureExtractor: () => (/* reexport safe */ _models_feature_extractors_js__WEBPACK_IMPORTED_MODULE_11__.SpeechT5FeatureExtractor),
|
|
39510
39817
|
/* harmony export */ SpeechT5ForSpeechToText: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.SpeechT5ForSpeechToText),
|
|
39511
39818
|
/* harmony export */ SpeechT5ForTextToSpeech: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.SpeechT5ForTextToSpeech),
|
|
@@ -40028,6 +40335,9 @@ var __webpack_exports__GPTNeoXTokenizer = __webpack_exports__.GPTNeoXTokenizer;
|
|
|
40028
40335
|
var __webpack_exports__Gemma2ForCausalLM = __webpack_exports__.Gemma2ForCausalLM;
|
|
40029
40336
|
var __webpack_exports__Gemma2Model = __webpack_exports__.Gemma2Model;
|
|
40030
40337
|
var __webpack_exports__Gemma2PreTrainedModel = __webpack_exports__.Gemma2PreTrainedModel;
|
|
40338
|
+
var __webpack_exports__Gemma3ForCausalLM = __webpack_exports__.Gemma3ForCausalLM;
|
|
40339
|
+
var __webpack_exports__Gemma3Model = __webpack_exports__.Gemma3Model;
|
|
40340
|
+
var __webpack_exports__Gemma3PreTrainedModel = __webpack_exports__.Gemma3PreTrainedModel;
|
|
40031
40341
|
var __webpack_exports__GemmaForCausalLM = __webpack_exports__.GemmaForCausalLM;
|
|
40032
40342
|
var __webpack_exports__GemmaModel = __webpack_exports__.GemmaModel;
|
|
40033
40343
|
var __webpack_exports__GemmaPreTrainedModel = __webpack_exports__.GemmaPreTrainedModel;
|
|
@@ -40129,6 +40439,10 @@ var __webpack_exports__MaskFormerModel = __webpack_exports__.MaskFormerModel;
|
|
|
40129
40439
|
var __webpack_exports__MaskFormerPreTrainedModel = __webpack_exports__.MaskFormerPreTrainedModel;
|
|
40130
40440
|
var __webpack_exports__MaskedLMOutput = __webpack_exports__.MaskedLMOutput;
|
|
40131
40441
|
var __webpack_exports__MaxLengthCriteria = __webpack_exports__.MaxLengthCriteria;
|
|
40442
|
+
var __webpack_exports__Metric3DForDepthEstimation = __webpack_exports__.Metric3DForDepthEstimation;
|
|
40443
|
+
var __webpack_exports__Metric3DPreTrainedModel = __webpack_exports__.Metric3DPreTrainedModel;
|
|
40444
|
+
var __webpack_exports__Metric3Dv2ForDepthEstimation = __webpack_exports__.Metric3Dv2ForDepthEstimation;
|
|
40445
|
+
var __webpack_exports__Metric3Dv2PreTrainedModel = __webpack_exports__.Metric3Dv2PreTrainedModel;
|
|
40132
40446
|
var __webpack_exports__MgpstrForSceneTextRecognition = __webpack_exports__.MgpstrForSceneTextRecognition;
|
|
40133
40447
|
var __webpack_exports__MgpstrModelOutput = __webpack_exports__.MgpstrModelOutput;
|
|
40134
40448
|
var __webpack_exports__MgpstrPreTrainedModel = __webpack_exports__.MgpstrPreTrainedModel;
|
|
@@ -40281,11 +40595,19 @@ var __webpack_exports__Qwen2VLForConditionalGeneration = __webpack_exports__.Qwe
|
|
|
40281
40595
|
var __webpack_exports__Qwen2VLImageProcessor = __webpack_exports__.Qwen2VLImageProcessor;
|
|
40282
40596
|
var __webpack_exports__Qwen2VLPreTrainedModel = __webpack_exports__.Qwen2VLPreTrainedModel;
|
|
40283
40597
|
var __webpack_exports__Qwen2VLProcessor = __webpack_exports__.Qwen2VLProcessor;
|
|
40598
|
+
var __webpack_exports__RFDetrForObjectDetection = __webpack_exports__.RFDetrForObjectDetection;
|
|
40599
|
+
var __webpack_exports__RFDetrModel = __webpack_exports__.RFDetrModel;
|
|
40600
|
+
var __webpack_exports__RFDetrObjectDetectionOutput = __webpack_exports__.RFDetrObjectDetectionOutput;
|
|
40601
|
+
var __webpack_exports__RFDetrPreTrainedModel = __webpack_exports__.RFDetrPreTrainedModel;
|
|
40284
40602
|
var __webpack_exports__RTDetrForObjectDetection = __webpack_exports__.RTDetrForObjectDetection;
|
|
40285
40603
|
var __webpack_exports__RTDetrImageProcessor = __webpack_exports__.RTDetrImageProcessor;
|
|
40286
40604
|
var __webpack_exports__RTDetrModel = __webpack_exports__.RTDetrModel;
|
|
40287
40605
|
var __webpack_exports__RTDetrObjectDetectionOutput = __webpack_exports__.RTDetrObjectDetectionOutput;
|
|
40288
40606
|
var __webpack_exports__RTDetrPreTrainedModel = __webpack_exports__.RTDetrPreTrainedModel;
|
|
40607
|
+
var __webpack_exports__RTDetrV2ForObjectDetection = __webpack_exports__.RTDetrV2ForObjectDetection;
|
|
40608
|
+
var __webpack_exports__RTDetrV2Model = __webpack_exports__.RTDetrV2Model;
|
|
40609
|
+
var __webpack_exports__RTDetrV2ObjectDetectionOutput = __webpack_exports__.RTDetrV2ObjectDetectionOutput;
|
|
40610
|
+
var __webpack_exports__RTDetrV2PreTrainedModel = __webpack_exports__.RTDetrV2PreTrainedModel;
|
|
40289
40611
|
var __webpack_exports__RawAudio = __webpack_exports__.RawAudio;
|
|
40290
40612
|
var __webpack_exports__RawImage = __webpack_exports__.RawImage;
|
|
40291
40613
|
var __webpack_exports__RawVideo = __webpack_exports__.RawVideo;
|
|
@@ -40335,6 +40657,11 @@ var __webpack_exports__SiglipVisionModel = __webpack_exports__.SiglipVisionModel
|
|
|
40335
40657
|
var __webpack_exports__SmolVLMForConditionalGeneration = __webpack_exports__.SmolVLMForConditionalGeneration;
|
|
40336
40658
|
var __webpack_exports__SmolVLMImageProcessor = __webpack_exports__.SmolVLMImageProcessor;
|
|
40337
40659
|
var __webpack_exports__SmolVLMProcessor = __webpack_exports__.SmolVLMProcessor;
|
|
40660
|
+
var __webpack_exports__SnacDecoderModel = __webpack_exports__.SnacDecoderModel;
|
|
40661
|
+
var __webpack_exports__SnacEncoderModel = __webpack_exports__.SnacEncoderModel;
|
|
40662
|
+
var __webpack_exports__SnacFeatureExtractor = __webpack_exports__.SnacFeatureExtractor;
|
|
40663
|
+
var __webpack_exports__SnacModel = __webpack_exports__.SnacModel;
|
|
40664
|
+
var __webpack_exports__SnacPreTrainedModel = __webpack_exports__.SnacPreTrainedModel;
|
|
40338
40665
|
var __webpack_exports__SpeechT5FeatureExtractor = __webpack_exports__.SpeechT5FeatureExtractor;
|
|
40339
40666
|
var __webpack_exports__SpeechT5ForSpeechToText = __webpack_exports__.SpeechT5ForSpeechToText;
|
|
40340
40667
|
var __webpack_exports__SpeechT5ForTextToSpeech = __webpack_exports__.SpeechT5ForTextToSpeech;
|
|
@@ -40528,6 +40855,6 @@ var __webpack_exports__topk = __webpack_exports__.topk;
|
|
|
40528
40855
|
var __webpack_exports__window_function = __webpack_exports__.window_function;
|
|
40529
40856
|
var __webpack_exports__zeros = __webpack_exports__.zeros;
|
|
40530
40857
|
var __webpack_exports__zeros_like = __webpack_exports__.zeros_like;
|
|
40531
|
-
export { __webpack_exports__ASTFeatureExtractor as ASTFeatureExtractor, __webpack_exports__ASTForAudioClassification as ASTForAudioClassification, __webpack_exports__ASTModel as ASTModel, __webpack_exports__ASTPreTrainedModel as ASTPreTrainedModel, __webpack_exports__AlbertForMaskedLM as AlbertForMaskedLM, __webpack_exports__AlbertForQuestionAnswering as AlbertForQuestionAnswering, __webpack_exports__AlbertForSequenceClassification as AlbertForSequenceClassification, __webpack_exports__AlbertModel as AlbertModel, __webpack_exports__AlbertPreTrainedModel as AlbertPreTrainedModel, __webpack_exports__AlbertTokenizer as AlbertTokenizer, __webpack_exports__AudioClassificationPipeline as AudioClassificationPipeline, __webpack_exports__AutoConfig as AutoConfig, __webpack_exports__AutoFeatureExtractor as AutoFeatureExtractor, __webpack_exports__AutoImageProcessor as AutoImageProcessor, __webpack_exports__AutoModel as AutoModel, __webpack_exports__AutoModelForAudioClassification as AutoModelForAudioClassification, __webpack_exports__AutoModelForAudioFrameClassification as AutoModelForAudioFrameClassification, __webpack_exports__AutoModelForAudioTextToText as AutoModelForAudioTextToText, __webpack_exports__AutoModelForCTC as AutoModelForCTC, __webpack_exports__AutoModelForCausalLM as AutoModelForCausalLM, __webpack_exports__AutoModelForDepthEstimation as AutoModelForDepthEstimation, __webpack_exports__AutoModelForDocumentQuestionAnswering as AutoModelForDocumentQuestionAnswering, __webpack_exports__AutoModelForImageClassification as AutoModelForImageClassification, __webpack_exports__AutoModelForImageFeatureExtraction as AutoModelForImageFeatureExtraction, __webpack_exports__AutoModelForImageMatting as AutoModelForImageMatting, __webpack_exports__AutoModelForImageSegmentation as AutoModelForImageSegmentation, __webpack_exports__AutoModelForImageTextToText as AutoModelForImageTextToText, __webpack_exports__AutoModelForImageToImage as AutoModelForImageToImage, __webpack_exports__AutoModelForMaskGeneration as AutoModelForMaskGeneration, __webpack_exports__AutoModelForMaskedLM as AutoModelForMaskedLM, __webpack_exports__AutoModelForNormalEstimation as AutoModelForNormalEstimation, __webpack_exports__AutoModelForObjectDetection as AutoModelForObjectDetection, __webpack_exports__AutoModelForPoseEstimation as AutoModelForPoseEstimation, __webpack_exports__AutoModelForQuestionAnswering as AutoModelForQuestionAnswering, __webpack_exports__AutoModelForSemanticSegmentation as AutoModelForSemanticSegmentation, __webpack_exports__AutoModelForSeq2SeqLM as AutoModelForSeq2SeqLM, __webpack_exports__AutoModelForSequenceClassification as AutoModelForSequenceClassification, __webpack_exports__AutoModelForSpeechSeq2Seq as AutoModelForSpeechSeq2Seq, __webpack_exports__AutoModelForTextToSpectrogram as AutoModelForTextToSpectrogram, __webpack_exports__AutoModelForTextToWaveform as AutoModelForTextToWaveform, __webpack_exports__AutoModelForTokenClassification as AutoModelForTokenClassification, __webpack_exports__AutoModelForUniversalSegmentation as AutoModelForUniversalSegmentation, __webpack_exports__AutoModelForVision2Seq as AutoModelForVision2Seq, __webpack_exports__AutoModelForXVector as AutoModelForXVector, __webpack_exports__AutoModelForZeroShotObjectDetection as AutoModelForZeroShotObjectDetection, __webpack_exports__AutoProcessor as AutoProcessor, __webpack_exports__AutoTokenizer as AutoTokenizer, __webpack_exports__AutomaticSpeechRecognitionPipeline as AutomaticSpeechRecognitionPipeline, __webpack_exports__BackgroundRemovalPipeline as BackgroundRemovalPipeline, __webpack_exports__BartForConditionalGeneration as BartForConditionalGeneration, __webpack_exports__BartForSequenceClassification as BartForSequenceClassification, __webpack_exports__BartModel as BartModel, __webpack_exports__BartPretrainedModel as BartPretrainedModel, __webpack_exports__BartTokenizer as BartTokenizer, __webpack_exports__BaseModelOutput as BaseModelOutput, __webpack_exports__BaseStreamer as BaseStreamer, __webpack_exports__BeitFeatureExtractor as BeitFeatureExtractor, __webpack_exports__BeitForImageClassification as BeitForImageClassification, __webpack_exports__BeitModel as BeitModel, __webpack_exports__BeitPreTrainedModel as BeitPreTrainedModel, __webpack_exports__BertForMaskedLM as BertForMaskedLM, __webpack_exports__BertForQuestionAnswering as BertForQuestionAnswering, __webpack_exports__BertForSequenceClassification as BertForSequenceClassification, __webpack_exports__BertForTokenClassification as BertForTokenClassification, __webpack_exports__BertModel as BertModel, __webpack_exports__BertPreTrainedModel as BertPreTrainedModel, __webpack_exports__BertTokenizer as BertTokenizer, __webpack_exports__BitImageProcessor as BitImageProcessor, __webpack_exports__BlenderbotForConditionalGeneration as BlenderbotForConditionalGeneration, __webpack_exports__BlenderbotModel as BlenderbotModel, __webpack_exports__BlenderbotPreTrainedModel as BlenderbotPreTrainedModel, __webpack_exports__BlenderbotSmallForConditionalGeneration as BlenderbotSmallForConditionalGeneration, __webpack_exports__BlenderbotSmallModel as BlenderbotSmallModel, __webpack_exports__BlenderbotSmallPreTrainedModel as BlenderbotSmallPreTrainedModel, __webpack_exports__BlenderbotSmallTokenizer as BlenderbotSmallTokenizer, __webpack_exports__BlenderbotTokenizer as BlenderbotTokenizer, __webpack_exports__BloomForCausalLM as BloomForCausalLM, __webpack_exports__BloomModel as BloomModel, __webpack_exports__BloomPreTrainedModel as BloomPreTrainedModel, __webpack_exports__BloomTokenizer as BloomTokenizer, __webpack_exports__CLIPFeatureExtractor as CLIPFeatureExtractor, __webpack_exports__CLIPImageProcessor as CLIPImageProcessor, __webpack_exports__CLIPModel as CLIPModel, __webpack_exports__CLIPPreTrainedModel as CLIPPreTrainedModel, __webpack_exports__CLIPSegForImageSegmentation as CLIPSegForImageSegmentation, __webpack_exports__CLIPSegModel as CLIPSegModel, __webpack_exports__CLIPSegPreTrainedModel as CLIPSegPreTrainedModel, __webpack_exports__CLIPTextModel as CLIPTextModel, __webpack_exports__CLIPTextModelWithProjection as CLIPTextModelWithProjection, __webpack_exports__CLIPTokenizer as CLIPTokenizer, __webpack_exports__CLIPVisionModel as CLIPVisionModel, __webpack_exports__CLIPVisionModelWithProjection as CLIPVisionModelWithProjection, __webpack_exports__CamembertForMaskedLM as CamembertForMaskedLM, __webpack_exports__CamembertForQuestionAnswering as CamembertForQuestionAnswering, __webpack_exports__CamembertForSequenceClassification as CamembertForSequenceClassification, __webpack_exports__CamembertForTokenClassification as CamembertForTokenClassification, __webpack_exports__CamembertModel as CamembertModel, __webpack_exports__CamembertPreTrainedModel as CamembertPreTrainedModel, __webpack_exports__CamembertTokenizer as CamembertTokenizer, __webpack_exports__CausalLMOutput as CausalLMOutput, __webpack_exports__CausalLMOutputWithPast as CausalLMOutputWithPast, __webpack_exports__ChineseCLIPFeatureExtractor as ChineseCLIPFeatureExtractor, __webpack_exports__ChineseCLIPModel as ChineseCLIPModel, __webpack_exports__ChineseCLIPPreTrainedModel as ChineseCLIPPreTrainedModel, __webpack_exports__ClapAudioModelWithProjection as ClapAudioModelWithProjection, __webpack_exports__ClapFeatureExtractor as ClapFeatureExtractor, __webpack_exports__ClapModel as ClapModel, __webpack_exports__ClapPreTrainedModel as ClapPreTrainedModel, __webpack_exports__ClapTextModelWithProjection as ClapTextModelWithProjection, __webpack_exports__ClassifierFreeGuidanceLogitsProcessor as ClassifierFreeGuidanceLogitsProcessor, __webpack_exports__CodeGenForCausalLM as CodeGenForCausalLM, __webpack_exports__CodeGenModel as CodeGenModel, __webpack_exports__CodeGenPreTrainedModel as CodeGenPreTrainedModel, __webpack_exports__CodeGenTokenizer as CodeGenTokenizer, __webpack_exports__CodeLlamaTokenizer as CodeLlamaTokenizer, __webpack_exports__CohereForCausalLM as CohereForCausalLM, __webpack_exports__CohereModel as CohereModel, __webpack_exports__CoherePreTrainedModel as CoherePreTrainedModel, __webpack_exports__CohereTokenizer as CohereTokenizer, __webpack_exports__ConvBertForMaskedLM as ConvBertForMaskedLM, __webpack_exports__ConvBertForQuestionAnswering as ConvBertForQuestionAnswering, __webpack_exports__ConvBertForSequenceClassification as ConvBertForSequenceClassification, __webpack_exports__ConvBertForTokenClassification as ConvBertForTokenClassification, __webpack_exports__ConvBertModel as ConvBertModel, __webpack_exports__ConvBertPreTrainedModel as ConvBertPreTrainedModel, __webpack_exports__ConvBertTokenizer as ConvBertTokenizer, __webpack_exports__ConvNextFeatureExtractor as ConvNextFeatureExtractor, __webpack_exports__ConvNextForImageClassification as ConvNextForImageClassification, __webpack_exports__ConvNextImageProcessor as ConvNextImageProcessor, __webpack_exports__ConvNextModel as ConvNextModel, __webpack_exports__ConvNextPreTrainedModel as ConvNextPreTrainedModel, __webpack_exports__ConvNextV2ForImageClassification as ConvNextV2ForImageClassification, __webpack_exports__ConvNextV2Model as ConvNextV2Model, __webpack_exports__ConvNextV2PreTrainedModel as ConvNextV2PreTrainedModel, __webpack_exports__DPTFeatureExtractor as DPTFeatureExtractor, __webpack_exports__DPTForDepthEstimation as DPTForDepthEstimation, __webpack_exports__DPTImageProcessor as DPTImageProcessor, __webpack_exports__DPTModel as DPTModel, __webpack_exports__DPTPreTrainedModel as DPTPreTrainedModel, __webpack_exports__DacDecoderModel as DacDecoderModel, __webpack_exports__DacDecoderOutput as DacDecoderOutput, __webpack_exports__DacEncoderModel as DacEncoderModel, __webpack_exports__DacEncoderOutput as DacEncoderOutput, __webpack_exports__DacFeatureExtractor as DacFeatureExtractor, __webpack_exports__DacModel as DacModel, __webpack_exports__DacPreTrainedModel as DacPreTrainedModel, __webpack_exports__DataTypeMap as DataTypeMap, __webpack_exports__DebertaForMaskedLM as DebertaForMaskedLM, __webpack_exports__DebertaForQuestionAnswering as DebertaForQuestionAnswering, __webpack_exports__DebertaForSequenceClassification as DebertaForSequenceClassification, __webpack_exports__DebertaForTokenClassification as DebertaForTokenClassification, __webpack_exports__DebertaModel as DebertaModel, __webpack_exports__DebertaPreTrainedModel as DebertaPreTrainedModel, __webpack_exports__DebertaTokenizer as DebertaTokenizer, __webpack_exports__DebertaV2ForMaskedLM as DebertaV2ForMaskedLM, __webpack_exports__DebertaV2ForQuestionAnswering as DebertaV2ForQuestionAnswering, __webpack_exports__DebertaV2ForSequenceClassification as DebertaV2ForSequenceClassification, __webpack_exports__DebertaV2ForTokenClassification as DebertaV2ForTokenClassification, __webpack_exports__DebertaV2Model as DebertaV2Model, __webpack_exports__DebertaV2PreTrainedModel as DebertaV2PreTrainedModel, __webpack_exports__DebertaV2Tokenizer as DebertaV2Tokenizer, __webpack_exports__DecisionTransformerModel as DecisionTransformerModel, __webpack_exports__DecisionTransformerPreTrainedModel as DecisionTransformerPreTrainedModel, __webpack_exports__DeiTFeatureExtractor as DeiTFeatureExtractor, __webpack_exports__DeiTForImageClassification as DeiTForImageClassification, __webpack_exports__DeiTImageProcessor as DeiTImageProcessor, __webpack_exports__DeiTModel as DeiTModel, __webpack_exports__DeiTPreTrainedModel as DeiTPreTrainedModel, __webpack_exports__DepthAnythingForDepthEstimation as DepthAnythingForDepthEstimation, __webpack_exports__DepthAnythingPreTrainedModel as DepthAnythingPreTrainedModel, __webpack_exports__DepthEstimationPipeline as DepthEstimationPipeline, __webpack_exports__DepthProForDepthEstimation as DepthProForDepthEstimation, __webpack_exports__DepthProPreTrainedModel as DepthProPreTrainedModel, __webpack_exports__DetrFeatureExtractor as DetrFeatureExtractor, __webpack_exports__DetrForObjectDetection as DetrForObjectDetection, __webpack_exports__DetrForSegmentation as DetrForSegmentation, __webpack_exports__DetrImageProcessor as DetrImageProcessor, __webpack_exports__DetrModel as DetrModel, __webpack_exports__DetrObjectDetectionOutput as DetrObjectDetectionOutput, __webpack_exports__DetrPreTrainedModel as DetrPreTrainedModel, __webpack_exports__DetrSegmentationOutput as DetrSegmentationOutput, __webpack_exports__Dinov2ForImageClassification as Dinov2ForImageClassification, __webpack_exports__Dinov2Model as Dinov2Model, __webpack_exports__Dinov2PreTrainedModel as Dinov2PreTrainedModel, __webpack_exports__Dinov2WithRegistersForImageClassification as Dinov2WithRegistersForImageClassification, __webpack_exports__Dinov2WithRegistersModel as Dinov2WithRegistersModel, __webpack_exports__Dinov2WithRegistersPreTrainedModel as Dinov2WithRegistersPreTrainedModel, __webpack_exports__DistilBertForMaskedLM as DistilBertForMaskedLM, __webpack_exports__DistilBertForQuestionAnswering as DistilBertForQuestionAnswering, __webpack_exports__DistilBertForSequenceClassification as DistilBertForSequenceClassification, __webpack_exports__DistilBertForTokenClassification as DistilBertForTokenClassification, __webpack_exports__DistilBertModel as DistilBertModel, __webpack_exports__DistilBertPreTrainedModel as DistilBertPreTrainedModel, __webpack_exports__DistilBertTokenizer as DistilBertTokenizer, __webpack_exports__DocumentQuestionAnsweringPipeline as DocumentQuestionAnsweringPipeline, __webpack_exports__DonutFeatureExtractor as DonutFeatureExtractor, __webpack_exports__DonutImageProcessor as DonutImageProcessor, __webpack_exports__DonutSwinModel as DonutSwinModel, __webpack_exports__DonutSwinPreTrainedModel as DonutSwinPreTrainedModel, __webpack_exports__EfficientNetForImageClassification as EfficientNetForImageClassification, __webpack_exports__EfficientNetImageProcessor as EfficientNetImageProcessor, __webpack_exports__EfficientNetModel as EfficientNetModel, __webpack_exports__EfficientNetPreTrainedModel as EfficientNetPreTrainedModel, __webpack_exports__ElectraForMaskedLM as ElectraForMaskedLM, __webpack_exports__ElectraForQuestionAnswering as ElectraForQuestionAnswering, __webpack_exports__ElectraForSequenceClassification as ElectraForSequenceClassification, __webpack_exports__ElectraForTokenClassification as ElectraForTokenClassification, __webpack_exports__ElectraModel as ElectraModel, __webpack_exports__ElectraPreTrainedModel as ElectraPreTrainedModel, __webpack_exports__ElectraTokenizer as ElectraTokenizer, __webpack_exports__EncodecFeatureExtractor as EncodecFeatureExtractor, __webpack_exports__EosTokenCriteria as EosTokenCriteria, __webpack_exports__EsmForMaskedLM as EsmForMaskedLM, __webpack_exports__EsmForSequenceClassification as EsmForSequenceClassification, __webpack_exports__EsmForTokenClassification as EsmForTokenClassification, __webpack_exports__EsmModel as EsmModel, __webpack_exports__EsmPreTrainedModel as EsmPreTrainedModel, __webpack_exports__EsmTokenizer as EsmTokenizer, __webpack_exports__ExaoneForCausalLM as ExaoneForCausalLM, __webpack_exports__ExaoneModel as ExaoneModel, __webpack_exports__ExaonePreTrainedModel as ExaonePreTrainedModel, __webpack_exports__FFT as FFT, __webpack_exports__FalconForCausalLM as FalconForCausalLM, __webpack_exports__FalconModel as FalconModel, __webpack_exports__FalconPreTrainedModel as FalconPreTrainedModel, __webpack_exports__FalconTokenizer as FalconTokenizer, __webpack_exports__FastViTForImageClassification as FastViTForImageClassification, __webpack_exports__FastViTModel as FastViTModel, __webpack_exports__FastViTPreTrainedModel as FastViTPreTrainedModel, __webpack_exports__FeatureExtractionPipeline as FeatureExtractionPipeline, __webpack_exports__FeatureExtractor as FeatureExtractor, __webpack_exports__FillMaskPipeline as FillMaskPipeline, __webpack_exports__Florence2ForConditionalGeneration as Florence2ForConditionalGeneration, __webpack_exports__Florence2PreTrainedModel as Florence2PreTrainedModel, __webpack_exports__Florence2Processor as Florence2Processor, __webpack_exports__ForcedBOSTokenLogitsProcessor as ForcedBOSTokenLogitsProcessor, __webpack_exports__ForcedEOSTokenLogitsProcessor as ForcedEOSTokenLogitsProcessor, __webpack_exports__GLPNFeatureExtractor as GLPNFeatureExtractor, __webpack_exports__GLPNForDepthEstimation as GLPNForDepthEstimation, __webpack_exports__GLPNModel as GLPNModel, __webpack_exports__GLPNPreTrainedModel as GLPNPreTrainedModel, __webpack_exports__GPT2LMHeadModel as GPT2LMHeadModel, __webpack_exports__GPT2Model as GPT2Model, __webpack_exports__GPT2PreTrainedModel as GPT2PreTrainedModel, __webpack_exports__GPT2Tokenizer as GPT2Tokenizer, __webpack_exports__GPTBigCodeForCausalLM as GPTBigCodeForCausalLM, __webpack_exports__GPTBigCodeModel as GPTBigCodeModel, __webpack_exports__GPTBigCodePreTrainedModel as GPTBigCodePreTrainedModel, __webpack_exports__GPTJForCausalLM as GPTJForCausalLM, __webpack_exports__GPTJModel as GPTJModel, __webpack_exports__GPTJPreTrainedModel as GPTJPreTrainedModel, __webpack_exports__GPTNeoForCausalLM as GPTNeoForCausalLM, __webpack_exports__GPTNeoModel as GPTNeoModel, __webpack_exports__GPTNeoPreTrainedModel as GPTNeoPreTrainedModel, __webpack_exports__GPTNeoXForCausalLM as GPTNeoXForCausalLM, __webpack_exports__GPTNeoXModel as GPTNeoXModel, __webpack_exports__GPTNeoXPreTrainedModel as GPTNeoXPreTrainedModel, __webpack_exports__GPTNeoXTokenizer as GPTNeoXTokenizer, __webpack_exports__Gemma2ForCausalLM as Gemma2ForCausalLM, __webpack_exports__Gemma2Model as Gemma2Model, __webpack_exports__Gemma2PreTrainedModel as Gemma2PreTrainedModel, __webpack_exports__GemmaForCausalLM as GemmaForCausalLM, __webpack_exports__GemmaModel as GemmaModel, __webpack_exports__GemmaPreTrainedModel as GemmaPreTrainedModel, __webpack_exports__GemmaTokenizer as GemmaTokenizer, __webpack_exports__GlmForCausalLM as GlmForCausalLM, __webpack_exports__GlmModel as GlmModel, __webpack_exports__GlmPreTrainedModel as GlmPreTrainedModel, __webpack_exports__GraniteForCausalLM as GraniteForCausalLM, __webpack_exports__GraniteModel as GraniteModel, __webpack_exports__GranitePreTrainedModel as GranitePreTrainedModel, __webpack_exports__Grok1Tokenizer as Grok1Tokenizer, __webpack_exports__GroundingDinoForObjectDetection as GroundingDinoForObjectDetection, __webpack_exports__GroundingDinoImageProcessor as GroundingDinoImageProcessor, __webpack_exports__GroundingDinoPreTrainedModel as GroundingDinoPreTrainedModel, __webpack_exports__GroundingDinoProcessor as GroundingDinoProcessor, __webpack_exports__GroupViTModel as GroupViTModel, __webpack_exports__GroupViTPreTrainedModel as GroupViTPreTrainedModel, __webpack_exports__HeliumForCausalLM as HeliumForCausalLM, __webpack_exports__HeliumModel as HeliumModel, __webpack_exports__HeliumPreTrainedModel as HeliumPreTrainedModel, __webpack_exports__HerbertTokenizer as HerbertTokenizer, __webpack_exports__HieraForImageClassification as HieraForImageClassification, __webpack_exports__HieraModel as HieraModel, __webpack_exports__HieraPreTrainedModel as HieraPreTrainedModel, __webpack_exports__HubertForCTC as HubertForCTC, __webpack_exports__HubertForSequenceClassification as HubertForSequenceClassification, __webpack_exports__HubertModel as HubertModel, __webpack_exports__HubertPreTrainedModel as HubertPreTrainedModel, __webpack_exports__IJepaForImageClassification as IJepaForImageClassification, __webpack_exports__IJepaModel as IJepaModel, __webpack_exports__IJepaPreTrainedModel as IJepaPreTrainedModel, __webpack_exports__Idefics3ForConditionalGeneration as Idefics3ForConditionalGeneration, __webpack_exports__Idefics3ImageProcessor as Idefics3ImageProcessor, __webpack_exports__Idefics3PreTrainedModel as Idefics3PreTrainedModel, __webpack_exports__Idefics3Processor as Idefics3Processor, __webpack_exports__ImageClassificationPipeline as ImageClassificationPipeline, __webpack_exports__ImageFeatureExtractionPipeline as ImageFeatureExtractionPipeline, __webpack_exports__ImageFeatureExtractor as ImageFeatureExtractor, __webpack_exports__ImageMattingOutput as ImageMattingOutput, __webpack_exports__ImageProcessor as ImageProcessor, __webpack_exports__ImageSegmentationPipeline as ImageSegmentationPipeline, __webpack_exports__ImageToImagePipeline as ImageToImagePipeline, __webpack_exports__ImageToTextPipeline as ImageToTextPipeline, __webpack_exports__InterruptableStoppingCriteria as InterruptableStoppingCriteria, __webpack_exports__JAISLMHeadModel as JAISLMHeadModel, __webpack_exports__JAISModel as JAISModel, __webpack_exports__JAISPreTrainedModel as JAISPreTrainedModel, __webpack_exports__JinaCLIPImageProcessor as JinaCLIPImageProcessor, __webpack_exports__JinaCLIPModel as JinaCLIPModel, __webpack_exports__JinaCLIPPreTrainedModel as JinaCLIPPreTrainedModel, __webpack_exports__JinaCLIPProcessor as JinaCLIPProcessor, __webpack_exports__JinaCLIPTextModel as JinaCLIPTextModel, __webpack_exports__JinaCLIPVisionModel as JinaCLIPVisionModel, __webpack_exports__LiteWhisperForConditionalGeneration as LiteWhisperForConditionalGeneration, __webpack_exports__LlamaForCausalLM as LlamaForCausalLM, __webpack_exports__LlamaModel as LlamaModel, __webpack_exports__LlamaPreTrainedModel as LlamaPreTrainedModel, __webpack_exports__LlamaTokenizer as LlamaTokenizer, __webpack_exports__LlavaForConditionalGeneration as LlavaForConditionalGeneration, __webpack_exports__LlavaOnevisionForConditionalGeneration as LlavaOnevisionForConditionalGeneration, __webpack_exports__LlavaOnevisionImageProcessor as LlavaOnevisionImageProcessor, __webpack_exports__LlavaPreTrainedModel as LlavaPreTrainedModel, __webpack_exports__LogitsProcessor as LogitsProcessor, __webpack_exports__LogitsProcessorList as LogitsProcessorList, __webpack_exports__LogitsWarper as LogitsWarper, __webpack_exports__LongT5ForConditionalGeneration as LongT5ForConditionalGeneration, __webpack_exports__LongT5Model as LongT5Model, __webpack_exports__LongT5PreTrainedModel as LongT5PreTrainedModel, __webpack_exports__M2M100ForConditionalGeneration as M2M100ForConditionalGeneration, __webpack_exports__M2M100Model as M2M100Model, __webpack_exports__M2M100PreTrainedModel as M2M100PreTrainedModel, __webpack_exports__M2M100Tokenizer as M2M100Tokenizer, __webpack_exports__MBart50Tokenizer as MBart50Tokenizer, __webpack_exports__MBartForCausalLM as MBartForCausalLM, __webpack_exports__MBartForConditionalGeneration as MBartForConditionalGeneration, __webpack_exports__MBartForSequenceClassification as MBartForSequenceClassification, __webpack_exports__MBartModel as MBartModel, __webpack_exports__MBartPreTrainedModel as MBartPreTrainedModel, __webpack_exports__MBartTokenizer as MBartTokenizer, __webpack_exports__MPNetForMaskedLM as MPNetForMaskedLM, __webpack_exports__MPNetForQuestionAnswering as MPNetForQuestionAnswering, __webpack_exports__MPNetForSequenceClassification as MPNetForSequenceClassification, __webpack_exports__MPNetForTokenClassification as MPNetForTokenClassification, __webpack_exports__MPNetModel as MPNetModel, __webpack_exports__MPNetPreTrainedModel as MPNetPreTrainedModel, __webpack_exports__MPNetTokenizer as MPNetTokenizer, __webpack_exports__MT5ForConditionalGeneration as MT5ForConditionalGeneration, __webpack_exports__MT5Model as MT5Model, __webpack_exports__MT5PreTrainedModel as MT5PreTrainedModel, __webpack_exports__MarianMTModel as MarianMTModel, __webpack_exports__MarianModel as MarianModel, __webpack_exports__MarianPreTrainedModel as MarianPreTrainedModel, __webpack_exports__MarianTokenizer as MarianTokenizer, __webpack_exports__Mask2FormerImageProcessor as Mask2FormerImageProcessor, __webpack_exports__MaskFormerFeatureExtractor as MaskFormerFeatureExtractor, __webpack_exports__MaskFormerForInstanceSegmentation as MaskFormerForInstanceSegmentation, __webpack_exports__MaskFormerImageProcessor as MaskFormerImageProcessor, __webpack_exports__MaskFormerModel as MaskFormerModel, __webpack_exports__MaskFormerPreTrainedModel as MaskFormerPreTrainedModel, __webpack_exports__MaskedLMOutput as MaskedLMOutput, __webpack_exports__MaxLengthCriteria as MaxLengthCriteria, __webpack_exports__MgpstrForSceneTextRecognition as MgpstrForSceneTextRecognition, __webpack_exports__MgpstrModelOutput as MgpstrModelOutput, __webpack_exports__MgpstrPreTrainedModel as MgpstrPreTrainedModel, __webpack_exports__MgpstrProcessor as MgpstrProcessor, __webpack_exports__MgpstrTokenizer as MgpstrTokenizer, __webpack_exports__MimiDecoderModel as MimiDecoderModel, __webpack_exports__MimiDecoderOutput as MimiDecoderOutput, __webpack_exports__MimiEncoderModel as MimiEncoderModel, __webpack_exports__MimiEncoderOutput as MimiEncoderOutput, __webpack_exports__MimiModel as MimiModel, __webpack_exports__MimiPreTrainedModel as MimiPreTrainedModel, __webpack_exports__MinLengthLogitsProcessor as MinLengthLogitsProcessor, __webpack_exports__MinNewTokensLengthLogitsProcessor as MinNewTokensLengthLogitsProcessor, __webpack_exports__MistralForCausalLM as MistralForCausalLM, __webpack_exports__MistralModel as MistralModel, __webpack_exports__MistralPreTrainedModel as MistralPreTrainedModel, __webpack_exports__MobileBertForMaskedLM as MobileBertForMaskedLM, __webpack_exports__MobileBertForQuestionAnswering as MobileBertForQuestionAnswering, __webpack_exports__MobileBertForSequenceClassification as MobileBertForSequenceClassification, __webpack_exports__MobileBertModel as MobileBertModel, __webpack_exports__MobileBertPreTrainedModel as MobileBertPreTrainedModel, __webpack_exports__MobileBertTokenizer as MobileBertTokenizer, __webpack_exports__MobileLLMForCausalLM as MobileLLMForCausalLM, __webpack_exports__MobileLLMModel as MobileLLMModel, __webpack_exports__MobileLLMPreTrainedModel as MobileLLMPreTrainedModel, __webpack_exports__MobileNetV1FeatureExtractor as MobileNetV1FeatureExtractor, __webpack_exports__MobileNetV1ForImageClassification as MobileNetV1ForImageClassification, __webpack_exports__MobileNetV1ForSemanticSegmentation as MobileNetV1ForSemanticSegmentation, __webpack_exports__MobileNetV1ImageProcessor as MobileNetV1ImageProcessor, __webpack_exports__MobileNetV1Model as MobileNetV1Model, __webpack_exports__MobileNetV1PreTrainedModel as MobileNetV1PreTrainedModel, __webpack_exports__MobileNetV2FeatureExtractor as MobileNetV2FeatureExtractor, __webpack_exports__MobileNetV2ForImageClassification as MobileNetV2ForImageClassification, __webpack_exports__MobileNetV2ForSemanticSegmentation as MobileNetV2ForSemanticSegmentation, __webpack_exports__MobileNetV2ImageProcessor as MobileNetV2ImageProcessor, __webpack_exports__MobileNetV2Model as MobileNetV2Model, __webpack_exports__MobileNetV2PreTrainedModel as MobileNetV2PreTrainedModel, __webpack_exports__MobileNetV3FeatureExtractor as MobileNetV3FeatureExtractor, __webpack_exports__MobileNetV3ForImageClassification as MobileNetV3ForImageClassification, __webpack_exports__MobileNetV3ForSemanticSegmentation as MobileNetV3ForSemanticSegmentation, __webpack_exports__MobileNetV3ImageProcessor as MobileNetV3ImageProcessor, __webpack_exports__MobileNetV3Model as MobileNetV3Model, __webpack_exports__MobileNetV3PreTrainedModel as MobileNetV3PreTrainedModel, __webpack_exports__MobileNetV4FeatureExtractor as MobileNetV4FeatureExtractor, __webpack_exports__MobileNetV4ForImageClassification as MobileNetV4ForImageClassification, __webpack_exports__MobileNetV4ForSemanticSegmentation as MobileNetV4ForSemanticSegmentation, __webpack_exports__MobileNetV4ImageProcessor as MobileNetV4ImageProcessor, __webpack_exports__MobileNetV4Model as MobileNetV4Model, __webpack_exports__MobileNetV4PreTrainedModel as MobileNetV4PreTrainedModel, __webpack_exports__MobileViTFeatureExtractor as MobileViTFeatureExtractor, __webpack_exports__MobileViTForImageClassification as MobileViTForImageClassification, __webpack_exports__MobileViTImageProcessor as MobileViTImageProcessor, __webpack_exports__MobileViTModel as MobileViTModel, __webpack_exports__MobileViTPreTrainedModel as MobileViTPreTrainedModel, __webpack_exports__MobileViTV2ForImageClassification as MobileViTV2ForImageClassification, __webpack_exports__MobileViTV2Model as MobileViTV2Model, __webpack_exports__MobileViTV2PreTrainedModel as MobileViTV2PreTrainedModel, __webpack_exports__ModelOutput as ModelOutput, __webpack_exports__ModernBertForMaskedLM as ModernBertForMaskedLM, __webpack_exports__ModernBertForSequenceClassification as ModernBertForSequenceClassification, __webpack_exports__ModernBertForTokenClassification as ModernBertForTokenClassification, __webpack_exports__ModernBertModel as ModernBertModel, __webpack_exports__ModernBertPreTrainedModel as ModernBertPreTrainedModel, __webpack_exports__Moondream1ForConditionalGeneration as Moondream1ForConditionalGeneration, __webpack_exports__MoonshineFeatureExtractor as MoonshineFeatureExtractor, __webpack_exports__MoonshineForConditionalGeneration as MoonshineForConditionalGeneration, __webpack_exports__MoonshineModel as MoonshineModel, __webpack_exports__MoonshinePreTrainedModel as MoonshinePreTrainedModel, __webpack_exports__MoonshineProcessor as MoonshineProcessor, __webpack_exports__MptForCausalLM as MptForCausalLM, __webpack_exports__MptModel as MptModel, __webpack_exports__MptPreTrainedModel as MptPreTrainedModel, __webpack_exports__MultiModalityCausalLM as MultiModalityCausalLM, __webpack_exports__MultiModalityPreTrainedModel as MultiModalityPreTrainedModel, __webpack_exports__MusicgenForCausalLM as MusicgenForCausalLM, __webpack_exports__MusicgenForConditionalGeneration as MusicgenForConditionalGeneration, __webpack_exports__MusicgenModel as MusicgenModel, __webpack_exports__MusicgenPreTrainedModel as MusicgenPreTrainedModel, __webpack_exports__NllbTokenizer as NllbTokenizer, __webpack_exports__NoBadWordsLogitsProcessor as NoBadWordsLogitsProcessor, __webpack_exports__NoRepeatNGramLogitsProcessor as NoRepeatNGramLogitsProcessor, __webpack_exports__NomicBertModel as NomicBertModel, __webpack_exports__NomicBertPreTrainedModel as NomicBertPreTrainedModel, __webpack_exports__NougatImageProcessor as NougatImageProcessor, __webpack_exports__NougatTokenizer as NougatTokenizer, __webpack_exports__OPTForCausalLM as OPTForCausalLM, __webpack_exports__OPTModel as OPTModel, __webpack_exports__OPTPreTrainedModel as OPTPreTrainedModel, __webpack_exports__ObjectDetectionPipeline as ObjectDetectionPipeline, __webpack_exports__Olmo2ForCausalLM as Olmo2ForCausalLM, __webpack_exports__Olmo2Model as Olmo2Model, __webpack_exports__Olmo2PreTrainedModel as Olmo2PreTrainedModel, __webpack_exports__OlmoForCausalLM as OlmoForCausalLM, __webpack_exports__OlmoModel as OlmoModel, __webpack_exports__OlmoPreTrainedModel as OlmoPreTrainedModel, __webpack_exports__OpenELMForCausalLM as OpenELMForCausalLM, __webpack_exports__OpenELMModel as OpenELMModel, __webpack_exports__OpenELMPreTrainedModel as OpenELMPreTrainedModel, __webpack_exports__OwlViTFeatureExtractor as OwlViTFeatureExtractor, __webpack_exports__OwlViTForObjectDetection as OwlViTForObjectDetection, __webpack_exports__OwlViTImageProcessor as OwlViTImageProcessor, __webpack_exports__OwlViTModel as OwlViTModel, __webpack_exports__OwlViTPreTrainedModel as OwlViTPreTrainedModel, __webpack_exports__OwlViTProcessor as OwlViTProcessor, __webpack_exports__Owlv2ForObjectDetection as Owlv2ForObjectDetection, __webpack_exports__Owlv2ImageProcessor as Owlv2ImageProcessor, __webpack_exports__Owlv2Model as Owlv2Model, __webpack_exports__Owlv2PreTrainedModel as Owlv2PreTrainedModel, __webpack_exports__PaliGemmaForConditionalGeneration as PaliGemmaForConditionalGeneration, __webpack_exports__PaliGemmaPreTrainedModel as PaliGemmaPreTrainedModel, __webpack_exports__PaliGemmaProcessor as PaliGemmaProcessor, __webpack_exports__PatchTSMixerForPrediction as PatchTSMixerForPrediction, __webpack_exports__PatchTSMixerModel as PatchTSMixerModel, __webpack_exports__PatchTSMixerPreTrainedModel as PatchTSMixerPreTrainedModel, __webpack_exports__PatchTSTForPrediction as PatchTSTForPrediction, __webpack_exports__PatchTSTModel as PatchTSTModel, __webpack_exports__PatchTSTPreTrainedModel as PatchTSTPreTrainedModel, __webpack_exports__Phi3ForCausalLM as Phi3ForCausalLM, __webpack_exports__Phi3Model as Phi3Model, __webpack_exports__Phi3PreTrainedModel as Phi3PreTrainedModel, __webpack_exports__Phi3VForCausalLM as Phi3VForCausalLM, __webpack_exports__Phi3VImageProcessor as Phi3VImageProcessor, __webpack_exports__Phi3VPreTrainedModel as Phi3VPreTrainedModel, __webpack_exports__Phi3VProcessor as Phi3VProcessor, __webpack_exports__PhiForCausalLM as PhiForCausalLM, __webpack_exports__PhiModel as PhiModel, __webpack_exports__PhiPreTrainedModel as PhiPreTrainedModel, __webpack_exports__Pipeline as Pipeline, __webpack_exports__PreTrainedModel as PreTrainedModel, __webpack_exports__PreTrainedTokenizer as PreTrainedTokenizer, __webpack_exports__PretrainedConfig as PretrainedConfig, __webpack_exports__PretrainedMixin as PretrainedMixin, __webpack_exports__Processor as Processor, __webpack_exports__PvtForImageClassification as PvtForImageClassification, __webpack_exports__PvtImageProcessor as PvtImageProcessor, __webpack_exports__PvtModel as PvtModel, __webpack_exports__PvtPreTrainedModel as PvtPreTrainedModel, __webpack_exports__PyAnnoteFeatureExtractor as PyAnnoteFeatureExtractor, __webpack_exports__PyAnnoteForAudioFrameClassification as PyAnnoteForAudioFrameClassification, __webpack_exports__PyAnnoteModel as PyAnnoteModel, __webpack_exports__PyAnnotePreTrainedModel as PyAnnotePreTrainedModel, __webpack_exports__PyAnnoteProcessor as PyAnnoteProcessor, __webpack_exports__QuestionAnsweringModelOutput as QuestionAnsweringModelOutput, __webpack_exports__QuestionAnsweringPipeline as QuestionAnsweringPipeline, __webpack_exports__Qwen2ForCausalLM as Qwen2ForCausalLM, __webpack_exports__Qwen2Model as Qwen2Model, __webpack_exports__Qwen2PreTrainedModel as Qwen2PreTrainedModel, __webpack_exports__Qwen2Tokenizer as Qwen2Tokenizer, __webpack_exports__Qwen2VLForConditionalGeneration as Qwen2VLForConditionalGeneration, __webpack_exports__Qwen2VLImageProcessor as Qwen2VLImageProcessor, __webpack_exports__Qwen2VLPreTrainedModel as Qwen2VLPreTrainedModel, __webpack_exports__Qwen2VLProcessor as Qwen2VLProcessor, __webpack_exports__RTDetrForObjectDetection as RTDetrForObjectDetection, __webpack_exports__RTDetrImageProcessor as RTDetrImageProcessor, __webpack_exports__RTDetrModel as RTDetrModel, __webpack_exports__RTDetrObjectDetectionOutput as RTDetrObjectDetectionOutput, __webpack_exports__RTDetrPreTrainedModel as RTDetrPreTrainedModel, __webpack_exports__RawAudio as RawAudio, __webpack_exports__RawImage as RawImage, __webpack_exports__RawVideo as RawVideo, __webpack_exports__RawVideoFrame as RawVideoFrame, __webpack_exports__RepetitionPenaltyLogitsProcessor as RepetitionPenaltyLogitsProcessor, __webpack_exports__ResNetForImageClassification as ResNetForImageClassification, __webpack_exports__ResNetModel as ResNetModel, __webpack_exports__ResNetPreTrainedModel as ResNetPreTrainedModel, __webpack_exports__RoFormerForMaskedLM as RoFormerForMaskedLM, __webpack_exports__RoFormerForQuestionAnswering as RoFormerForQuestionAnswering, __webpack_exports__RoFormerForSequenceClassification as RoFormerForSequenceClassification, __webpack_exports__RoFormerForTokenClassification as RoFormerForTokenClassification, __webpack_exports__RoFormerModel as RoFormerModel, __webpack_exports__RoFormerPreTrainedModel as RoFormerPreTrainedModel, __webpack_exports__RoFormerTokenizer as RoFormerTokenizer, __webpack_exports__RobertaForMaskedLM as RobertaForMaskedLM, __webpack_exports__RobertaForQuestionAnswering as RobertaForQuestionAnswering, __webpack_exports__RobertaForSequenceClassification as RobertaForSequenceClassification, __webpack_exports__RobertaForTokenClassification as RobertaForTokenClassification, __webpack_exports__RobertaModel as RobertaModel, __webpack_exports__RobertaPreTrainedModel as RobertaPreTrainedModel, __webpack_exports__RobertaTokenizer as RobertaTokenizer, __webpack_exports__SamImageProcessor as SamImageProcessor, __webpack_exports__SamImageSegmentationOutput as SamImageSegmentationOutput, __webpack_exports__SamModel as SamModel, __webpack_exports__SamPreTrainedModel as SamPreTrainedModel, __webpack_exports__SamProcessor as SamProcessor, __webpack_exports__SapiensForDepthEstimation as SapiensForDepthEstimation, __webpack_exports__SapiensForNormalEstimation as SapiensForNormalEstimation, __webpack_exports__SapiensForSemanticSegmentation as SapiensForSemanticSegmentation, __webpack_exports__SapiensPreTrainedModel as SapiensPreTrainedModel, __webpack_exports__SeamlessM4TFeatureExtractor as SeamlessM4TFeatureExtractor, __webpack_exports__SegformerFeatureExtractor as SegformerFeatureExtractor, __webpack_exports__SegformerForImageClassification as SegformerForImageClassification, __webpack_exports__SegformerForSemanticSegmentation as SegformerForSemanticSegmentation, __webpack_exports__SegformerImageProcessor as SegformerImageProcessor, __webpack_exports__SegformerModel as SegformerModel, __webpack_exports__SegformerPreTrainedModel as SegformerPreTrainedModel, __webpack_exports__Seq2SeqLMOutput as Seq2SeqLMOutput, __webpack_exports__SequenceClassifierOutput as SequenceClassifierOutput, __webpack_exports__SiglipImageProcessor as SiglipImageProcessor, __webpack_exports__SiglipModel as SiglipModel, __webpack_exports__SiglipPreTrainedModel as SiglipPreTrainedModel, __webpack_exports__SiglipTextModel as SiglipTextModel, __webpack_exports__SiglipTokenizer as SiglipTokenizer, __webpack_exports__SiglipVisionModel as SiglipVisionModel, __webpack_exports__SmolVLMForConditionalGeneration as SmolVLMForConditionalGeneration, __webpack_exports__SmolVLMImageProcessor as SmolVLMImageProcessor, __webpack_exports__SmolVLMProcessor as SmolVLMProcessor, __webpack_exports__SpeechT5FeatureExtractor as SpeechT5FeatureExtractor, __webpack_exports__SpeechT5ForSpeechToText as SpeechT5ForSpeechToText, __webpack_exports__SpeechT5ForTextToSpeech as SpeechT5ForTextToSpeech, __webpack_exports__SpeechT5HifiGan as SpeechT5HifiGan, __webpack_exports__SpeechT5Model as SpeechT5Model, __webpack_exports__SpeechT5PreTrainedModel as SpeechT5PreTrainedModel, __webpack_exports__SpeechT5Processor as SpeechT5Processor, __webpack_exports__SpeechT5Tokenizer as SpeechT5Tokenizer, __webpack_exports__SqueezeBertForMaskedLM as SqueezeBertForMaskedLM, __webpack_exports__SqueezeBertForQuestionAnswering as SqueezeBertForQuestionAnswering, __webpack_exports__SqueezeBertForSequenceClassification as SqueezeBertForSequenceClassification, __webpack_exports__SqueezeBertModel as SqueezeBertModel, __webpack_exports__SqueezeBertPreTrainedModel as SqueezeBertPreTrainedModel, __webpack_exports__SqueezeBertTokenizer as SqueezeBertTokenizer, __webpack_exports__StableLmForCausalLM as StableLmForCausalLM, __webpack_exports__StableLmModel as StableLmModel, __webpack_exports__StableLmPreTrainedModel as StableLmPreTrainedModel, __webpack_exports__Starcoder2ForCausalLM as Starcoder2ForCausalLM, __webpack_exports__Starcoder2Model as Starcoder2Model, __webpack_exports__Starcoder2PreTrainedModel as Starcoder2PreTrainedModel, __webpack_exports__StoppingCriteria as StoppingCriteria, __webpack_exports__StoppingCriteriaList as StoppingCriteriaList, __webpack_exports__StyleTextToSpeech2Model as StyleTextToSpeech2Model, __webpack_exports__StyleTextToSpeech2PreTrainedModel as StyleTextToSpeech2PreTrainedModel, __webpack_exports__SummarizationPipeline as SummarizationPipeline, __webpack_exports__SuppressTokensAtBeginLogitsProcessor as SuppressTokensAtBeginLogitsProcessor, __webpack_exports__Swin2SRForImageSuperResolution as Swin2SRForImageSuperResolution, __webpack_exports__Swin2SRImageProcessor as Swin2SRImageProcessor, __webpack_exports__Swin2SRModel as Swin2SRModel, __webpack_exports__Swin2SRPreTrainedModel as Swin2SRPreTrainedModel, __webpack_exports__SwinForImageClassification as SwinForImageClassification, __webpack_exports__SwinForSemanticSegmentation as SwinForSemanticSegmentation, __webpack_exports__SwinModel as SwinModel, __webpack_exports__SwinPreTrainedModel as SwinPreTrainedModel, __webpack_exports__T5ForConditionalGeneration as T5ForConditionalGeneration, __webpack_exports__T5Model as T5Model, __webpack_exports__T5PreTrainedModel as T5PreTrainedModel, __webpack_exports__T5Tokenizer as T5Tokenizer, __webpack_exports__TableTransformerForObjectDetection as TableTransformerForObjectDetection, __webpack_exports__TableTransformerModel as TableTransformerModel, __webpack_exports__TableTransformerObjectDetectionOutput as TableTransformerObjectDetectionOutput, __webpack_exports__TableTransformerPreTrainedModel as TableTransformerPreTrainedModel, __webpack_exports__TemperatureLogitsWarper as TemperatureLogitsWarper, __webpack_exports__Tensor as Tensor, __webpack_exports__Text2TextGenerationPipeline as Text2TextGenerationPipeline, __webpack_exports__TextClassificationPipeline as TextClassificationPipeline, __webpack_exports__TextGenerationPipeline as TextGenerationPipeline, __webpack_exports__TextStreamer as TextStreamer, __webpack_exports__TextToAudioPipeline as TextToAudioPipeline, __webpack_exports__TokenClassificationPipeline as TokenClassificationPipeline, __webpack_exports__TokenClassifierOutput as TokenClassifierOutput, __webpack_exports__TokenizerModel as TokenizerModel, __webpack_exports__TopKLogitsWarper as TopKLogitsWarper, __webpack_exports__TopPLogitsWarper as TopPLogitsWarper, __webpack_exports__TrOCRForCausalLM as TrOCRForCausalLM, __webpack_exports__TrOCRPreTrainedModel as TrOCRPreTrainedModel, __webpack_exports__TranslationPipeline as TranslationPipeline, __webpack_exports__UltravoxModel as UltravoxModel, __webpack_exports__UltravoxPreTrainedModel as UltravoxPreTrainedModel, __webpack_exports__UltravoxProcessor as UltravoxProcessor, __webpack_exports__UniSpeechForCTC as UniSpeechForCTC, __webpack_exports__UniSpeechForSequenceClassification as UniSpeechForSequenceClassification, __webpack_exports__UniSpeechModel as UniSpeechModel, __webpack_exports__UniSpeechPreTrainedModel as UniSpeechPreTrainedModel, __webpack_exports__UniSpeechSatForAudioFrameClassification as UniSpeechSatForAudioFrameClassification, __webpack_exports__UniSpeechSatForCTC as UniSpeechSatForCTC, __webpack_exports__UniSpeechSatForSequenceClassification as UniSpeechSatForSequenceClassification, __webpack_exports__UniSpeechSatModel as UniSpeechSatModel, __webpack_exports__UniSpeechSatPreTrainedModel as UniSpeechSatPreTrainedModel, __webpack_exports__VLChatProcessor as VLChatProcessor, __webpack_exports__VLMImageProcessor as VLMImageProcessor, __webpack_exports__ViTFeatureExtractor as ViTFeatureExtractor, __webpack_exports__ViTForImageClassification as ViTForImageClassification, __webpack_exports__ViTImageProcessor as ViTImageProcessor, __webpack_exports__ViTMAEModel as ViTMAEModel, __webpack_exports__ViTMAEPreTrainedModel as ViTMAEPreTrainedModel, __webpack_exports__ViTMSNForImageClassification as ViTMSNForImageClassification, __webpack_exports__ViTMSNModel as ViTMSNModel, __webpack_exports__ViTMSNPreTrainedModel as ViTMSNPreTrainedModel, __webpack_exports__ViTModel as ViTModel, __webpack_exports__ViTPreTrainedModel as ViTPreTrainedModel, __webpack_exports__VisionEncoderDecoderModel as VisionEncoderDecoderModel, __webpack_exports__VitMatteForImageMatting as VitMatteForImageMatting, __webpack_exports__VitMatteImageProcessor as VitMatteImageProcessor, __webpack_exports__VitMattePreTrainedModel as VitMattePreTrainedModel, __webpack_exports__VitPoseForPoseEstimation as VitPoseForPoseEstimation, __webpack_exports__VitPoseImageProcessor as VitPoseImageProcessor, __webpack_exports__VitPosePreTrainedModel as VitPosePreTrainedModel, __webpack_exports__VitsModel as VitsModel, __webpack_exports__VitsModelOutput as VitsModelOutput, __webpack_exports__VitsPreTrainedModel as VitsPreTrainedModel, __webpack_exports__VitsTokenizer as VitsTokenizer, __webpack_exports__Wav2Vec2BertForCTC as Wav2Vec2BertForCTC, __webpack_exports__Wav2Vec2BertForSequenceClassification as Wav2Vec2BertForSequenceClassification, __webpack_exports__Wav2Vec2BertModel as Wav2Vec2BertModel, __webpack_exports__Wav2Vec2BertPreTrainedModel as Wav2Vec2BertPreTrainedModel, __webpack_exports__Wav2Vec2CTCTokenizer as Wav2Vec2CTCTokenizer, __webpack_exports__Wav2Vec2FeatureExtractor as Wav2Vec2FeatureExtractor, __webpack_exports__Wav2Vec2ForAudioFrameClassification as Wav2Vec2ForAudioFrameClassification, __webpack_exports__Wav2Vec2ForCTC as Wav2Vec2ForCTC, __webpack_exports__Wav2Vec2ForSequenceClassification as Wav2Vec2ForSequenceClassification, __webpack_exports__Wav2Vec2Model as Wav2Vec2Model, __webpack_exports__Wav2Vec2PreTrainedModel as Wav2Vec2PreTrainedModel, __webpack_exports__Wav2Vec2Processor as Wav2Vec2Processor, __webpack_exports__Wav2Vec2ProcessorWithLM as Wav2Vec2ProcessorWithLM, __webpack_exports__WavLMForAudioFrameClassification as WavLMForAudioFrameClassification, __webpack_exports__WavLMForCTC as WavLMForCTC, __webpack_exports__WavLMForSequenceClassification as WavLMForSequenceClassification, __webpack_exports__WavLMForXVector as WavLMForXVector, __webpack_exports__WavLMModel as WavLMModel, __webpack_exports__WavLMPreTrainedModel as WavLMPreTrainedModel, __webpack_exports__WeSpeakerFeatureExtractor as WeSpeakerFeatureExtractor, __webpack_exports__WeSpeakerResNetModel as WeSpeakerResNetModel, __webpack_exports__WeSpeakerResNetPreTrainedModel as WeSpeakerResNetPreTrainedModel, __webpack_exports__WhisperFeatureExtractor as WhisperFeatureExtractor, __webpack_exports__WhisperForConditionalGeneration as WhisperForConditionalGeneration, __webpack_exports__WhisperModel as WhisperModel, __webpack_exports__WhisperPreTrainedModel as WhisperPreTrainedModel, __webpack_exports__WhisperProcessor as WhisperProcessor, __webpack_exports__WhisperTextStreamer as WhisperTextStreamer, __webpack_exports__WhisperTimeStampLogitsProcessor as WhisperTimeStampLogitsProcessor, __webpack_exports__WhisperTokenizer as WhisperTokenizer, __webpack_exports__XLMForQuestionAnswering as XLMForQuestionAnswering, __webpack_exports__XLMForSequenceClassification as XLMForSequenceClassification, __webpack_exports__XLMForTokenClassification as XLMForTokenClassification, __webpack_exports__XLMModel as XLMModel, __webpack_exports__XLMPreTrainedModel as XLMPreTrainedModel, __webpack_exports__XLMRobertaForMaskedLM as XLMRobertaForMaskedLM, __webpack_exports__XLMRobertaForQuestionAnswering as XLMRobertaForQuestionAnswering, __webpack_exports__XLMRobertaForSequenceClassification as XLMRobertaForSequenceClassification, __webpack_exports__XLMRobertaForTokenClassification as XLMRobertaForTokenClassification, __webpack_exports__XLMRobertaModel as XLMRobertaModel, __webpack_exports__XLMRobertaPreTrainedModel as XLMRobertaPreTrainedModel, __webpack_exports__XLMRobertaTokenizer as XLMRobertaTokenizer, __webpack_exports__XLMTokenizer as XLMTokenizer, __webpack_exports__XLMWithLMHeadModel as XLMWithLMHeadModel, __webpack_exports__XVectorOutput as XVectorOutput, __webpack_exports__YolosFeatureExtractor as YolosFeatureExtractor, __webpack_exports__YolosForObjectDetection as YolosForObjectDetection, __webpack_exports__YolosImageProcessor as YolosImageProcessor, __webpack_exports__YolosModel as YolosModel, __webpack_exports__YolosObjectDetectionOutput as YolosObjectDetectionOutput, __webpack_exports__YolosPreTrainedModel as YolosPreTrainedModel, __webpack_exports__ZeroShotAudioClassificationPipeline as ZeroShotAudioClassificationPipeline, __webpack_exports__ZeroShotClassificationPipeline as ZeroShotClassificationPipeline, __webpack_exports__ZeroShotImageClassificationPipeline as ZeroShotImageClassificationPipeline, __webpack_exports__ZeroShotObjectDetectionPipeline as ZeroShotObjectDetectionPipeline, __webpack_exports__bankers_round as bankers_round, __webpack_exports__cat as cat, __webpack_exports__cos_sim as cos_sim, __webpack_exports__dot as dot, __webpack_exports__dynamic_time_warping as dynamic_time_warping, __webpack_exports__env as env, __webpack_exports__full as full, __webpack_exports__full_like as full_like, __webpack_exports__getKeyValueShapes as getKeyValueShapes, __webpack_exports__hamming as hamming, __webpack_exports__hanning as hanning, __webpack_exports__interpolate as interpolate, __webpack_exports__interpolate_4d as interpolate_4d, __webpack_exports__interpolate_data as interpolate_data, __webpack_exports__is_chinese_char as is_chinese_char, __webpack_exports__layer_norm as layer_norm, __webpack_exports__load_image as load_image, __webpack_exports__load_video as load_video, __webpack_exports__log_softmax as log_softmax, __webpack_exports__magnitude as magnitude, __webpack_exports__matmul as matmul, __webpack_exports__max as max, __webpack_exports__mean as mean, __webpack_exports__mean_pooling as mean_pooling, __webpack_exports__medianFilter as medianFilter, __webpack_exports__mel_filter_bank as mel_filter_bank, __webpack_exports__min as min, __webpack_exports__ones as ones, __webpack_exports__ones_like as ones_like, __webpack_exports__permute as permute, __webpack_exports__permute_data as permute_data, __webpack_exports__pipeline as pipeline, __webpack_exports__quantize_embeddings as quantize_embeddings, __webpack_exports__rand as rand, __webpack_exports__read_audio as read_audio, __webpack_exports__rfft as rfft, __webpack_exports__round as round, __webpack_exports__slice as slice, __webpack_exports__softmax as softmax, __webpack_exports__spectrogram as spectrogram, __webpack_exports__stack as stack, __webpack_exports__std_mean as std_mean, __webpack_exports__topk as topk, __webpack_exports__window_function as window_function, __webpack_exports__zeros as zeros, __webpack_exports__zeros_like as zeros_like };
|
|
40858
|
+
export { __webpack_exports__ASTFeatureExtractor as ASTFeatureExtractor, __webpack_exports__ASTForAudioClassification as ASTForAudioClassification, __webpack_exports__ASTModel as ASTModel, __webpack_exports__ASTPreTrainedModel as ASTPreTrainedModel, __webpack_exports__AlbertForMaskedLM as AlbertForMaskedLM, __webpack_exports__AlbertForQuestionAnswering as AlbertForQuestionAnswering, __webpack_exports__AlbertForSequenceClassification as AlbertForSequenceClassification, __webpack_exports__AlbertModel as AlbertModel, __webpack_exports__AlbertPreTrainedModel as AlbertPreTrainedModel, __webpack_exports__AlbertTokenizer as AlbertTokenizer, __webpack_exports__AudioClassificationPipeline as AudioClassificationPipeline, __webpack_exports__AutoConfig as AutoConfig, __webpack_exports__AutoFeatureExtractor as AutoFeatureExtractor, __webpack_exports__AutoImageProcessor as AutoImageProcessor, __webpack_exports__AutoModel as AutoModel, __webpack_exports__AutoModelForAudioClassification as AutoModelForAudioClassification, __webpack_exports__AutoModelForAudioFrameClassification as AutoModelForAudioFrameClassification, __webpack_exports__AutoModelForAudioTextToText as AutoModelForAudioTextToText, __webpack_exports__AutoModelForCTC as AutoModelForCTC, __webpack_exports__AutoModelForCausalLM as AutoModelForCausalLM, __webpack_exports__AutoModelForDepthEstimation as AutoModelForDepthEstimation, __webpack_exports__AutoModelForDocumentQuestionAnswering as AutoModelForDocumentQuestionAnswering, __webpack_exports__AutoModelForImageClassification as AutoModelForImageClassification, __webpack_exports__AutoModelForImageFeatureExtraction as AutoModelForImageFeatureExtraction, __webpack_exports__AutoModelForImageMatting as AutoModelForImageMatting, __webpack_exports__AutoModelForImageSegmentation as AutoModelForImageSegmentation, __webpack_exports__AutoModelForImageTextToText as AutoModelForImageTextToText, __webpack_exports__AutoModelForImageToImage as AutoModelForImageToImage, __webpack_exports__AutoModelForMaskGeneration as AutoModelForMaskGeneration, __webpack_exports__AutoModelForMaskedLM as AutoModelForMaskedLM, __webpack_exports__AutoModelForNormalEstimation as AutoModelForNormalEstimation, __webpack_exports__AutoModelForObjectDetection as AutoModelForObjectDetection, __webpack_exports__AutoModelForPoseEstimation as AutoModelForPoseEstimation, __webpack_exports__AutoModelForQuestionAnswering as AutoModelForQuestionAnswering, __webpack_exports__AutoModelForSemanticSegmentation as AutoModelForSemanticSegmentation, __webpack_exports__AutoModelForSeq2SeqLM as AutoModelForSeq2SeqLM, __webpack_exports__AutoModelForSequenceClassification as AutoModelForSequenceClassification, __webpack_exports__AutoModelForSpeechSeq2Seq as AutoModelForSpeechSeq2Seq, __webpack_exports__AutoModelForTextToSpectrogram as AutoModelForTextToSpectrogram, __webpack_exports__AutoModelForTextToWaveform as AutoModelForTextToWaveform, __webpack_exports__AutoModelForTokenClassification as AutoModelForTokenClassification, __webpack_exports__AutoModelForUniversalSegmentation as AutoModelForUniversalSegmentation, __webpack_exports__AutoModelForVision2Seq as AutoModelForVision2Seq, __webpack_exports__AutoModelForXVector as AutoModelForXVector, __webpack_exports__AutoModelForZeroShotObjectDetection as AutoModelForZeroShotObjectDetection, __webpack_exports__AutoProcessor as AutoProcessor, __webpack_exports__AutoTokenizer as AutoTokenizer, __webpack_exports__AutomaticSpeechRecognitionPipeline as AutomaticSpeechRecognitionPipeline, __webpack_exports__BackgroundRemovalPipeline as BackgroundRemovalPipeline, __webpack_exports__BartForConditionalGeneration as BartForConditionalGeneration, __webpack_exports__BartForSequenceClassification as BartForSequenceClassification, __webpack_exports__BartModel as BartModel, __webpack_exports__BartPretrainedModel as BartPretrainedModel, __webpack_exports__BartTokenizer as BartTokenizer, __webpack_exports__BaseModelOutput as BaseModelOutput, __webpack_exports__BaseStreamer as BaseStreamer, __webpack_exports__BeitFeatureExtractor as BeitFeatureExtractor, __webpack_exports__BeitForImageClassification as BeitForImageClassification, __webpack_exports__BeitModel as BeitModel, __webpack_exports__BeitPreTrainedModel as BeitPreTrainedModel, __webpack_exports__BertForMaskedLM as BertForMaskedLM, __webpack_exports__BertForQuestionAnswering as BertForQuestionAnswering, __webpack_exports__BertForSequenceClassification as BertForSequenceClassification, __webpack_exports__BertForTokenClassification as BertForTokenClassification, __webpack_exports__BertModel as BertModel, __webpack_exports__BertPreTrainedModel as BertPreTrainedModel, __webpack_exports__BertTokenizer as BertTokenizer, __webpack_exports__BitImageProcessor as BitImageProcessor, __webpack_exports__BlenderbotForConditionalGeneration as BlenderbotForConditionalGeneration, __webpack_exports__BlenderbotModel as BlenderbotModel, __webpack_exports__BlenderbotPreTrainedModel as BlenderbotPreTrainedModel, __webpack_exports__BlenderbotSmallForConditionalGeneration as BlenderbotSmallForConditionalGeneration, __webpack_exports__BlenderbotSmallModel as BlenderbotSmallModel, __webpack_exports__BlenderbotSmallPreTrainedModel as BlenderbotSmallPreTrainedModel, __webpack_exports__BlenderbotSmallTokenizer as BlenderbotSmallTokenizer, __webpack_exports__BlenderbotTokenizer as BlenderbotTokenizer, __webpack_exports__BloomForCausalLM as BloomForCausalLM, __webpack_exports__BloomModel as BloomModel, __webpack_exports__BloomPreTrainedModel as BloomPreTrainedModel, __webpack_exports__BloomTokenizer as BloomTokenizer, __webpack_exports__CLIPFeatureExtractor as CLIPFeatureExtractor, __webpack_exports__CLIPImageProcessor as CLIPImageProcessor, __webpack_exports__CLIPModel as CLIPModel, __webpack_exports__CLIPPreTrainedModel as CLIPPreTrainedModel, __webpack_exports__CLIPSegForImageSegmentation as CLIPSegForImageSegmentation, __webpack_exports__CLIPSegModel as CLIPSegModel, __webpack_exports__CLIPSegPreTrainedModel as CLIPSegPreTrainedModel, __webpack_exports__CLIPTextModel as CLIPTextModel, __webpack_exports__CLIPTextModelWithProjection as CLIPTextModelWithProjection, __webpack_exports__CLIPTokenizer as CLIPTokenizer, __webpack_exports__CLIPVisionModel as CLIPVisionModel, __webpack_exports__CLIPVisionModelWithProjection as CLIPVisionModelWithProjection, __webpack_exports__CamembertForMaskedLM as CamembertForMaskedLM, __webpack_exports__CamembertForQuestionAnswering as CamembertForQuestionAnswering, __webpack_exports__CamembertForSequenceClassification as CamembertForSequenceClassification, __webpack_exports__CamembertForTokenClassification as CamembertForTokenClassification, __webpack_exports__CamembertModel as CamembertModel, __webpack_exports__CamembertPreTrainedModel as CamembertPreTrainedModel, __webpack_exports__CamembertTokenizer as CamembertTokenizer, __webpack_exports__CausalLMOutput as CausalLMOutput, __webpack_exports__CausalLMOutputWithPast as CausalLMOutputWithPast, __webpack_exports__ChineseCLIPFeatureExtractor as ChineseCLIPFeatureExtractor, __webpack_exports__ChineseCLIPModel as ChineseCLIPModel, __webpack_exports__ChineseCLIPPreTrainedModel as ChineseCLIPPreTrainedModel, __webpack_exports__ClapAudioModelWithProjection as ClapAudioModelWithProjection, __webpack_exports__ClapFeatureExtractor as ClapFeatureExtractor, __webpack_exports__ClapModel as ClapModel, __webpack_exports__ClapPreTrainedModel as ClapPreTrainedModel, __webpack_exports__ClapTextModelWithProjection as ClapTextModelWithProjection, __webpack_exports__ClassifierFreeGuidanceLogitsProcessor as ClassifierFreeGuidanceLogitsProcessor, __webpack_exports__CodeGenForCausalLM as CodeGenForCausalLM, __webpack_exports__CodeGenModel as CodeGenModel, __webpack_exports__CodeGenPreTrainedModel as CodeGenPreTrainedModel, __webpack_exports__CodeGenTokenizer as CodeGenTokenizer, __webpack_exports__CodeLlamaTokenizer as CodeLlamaTokenizer, __webpack_exports__CohereForCausalLM as CohereForCausalLM, __webpack_exports__CohereModel as CohereModel, __webpack_exports__CoherePreTrainedModel as CoherePreTrainedModel, __webpack_exports__CohereTokenizer as CohereTokenizer, __webpack_exports__ConvBertForMaskedLM as ConvBertForMaskedLM, __webpack_exports__ConvBertForQuestionAnswering as ConvBertForQuestionAnswering, __webpack_exports__ConvBertForSequenceClassification as ConvBertForSequenceClassification, __webpack_exports__ConvBertForTokenClassification as ConvBertForTokenClassification, __webpack_exports__ConvBertModel as ConvBertModel, __webpack_exports__ConvBertPreTrainedModel as ConvBertPreTrainedModel, __webpack_exports__ConvBertTokenizer as ConvBertTokenizer, __webpack_exports__ConvNextFeatureExtractor as ConvNextFeatureExtractor, __webpack_exports__ConvNextForImageClassification as ConvNextForImageClassification, __webpack_exports__ConvNextImageProcessor as ConvNextImageProcessor, __webpack_exports__ConvNextModel as ConvNextModel, __webpack_exports__ConvNextPreTrainedModel as ConvNextPreTrainedModel, __webpack_exports__ConvNextV2ForImageClassification as ConvNextV2ForImageClassification, __webpack_exports__ConvNextV2Model as ConvNextV2Model, __webpack_exports__ConvNextV2PreTrainedModel as ConvNextV2PreTrainedModel, __webpack_exports__DPTFeatureExtractor as DPTFeatureExtractor, __webpack_exports__DPTForDepthEstimation as DPTForDepthEstimation, __webpack_exports__DPTImageProcessor as DPTImageProcessor, __webpack_exports__DPTModel as DPTModel, __webpack_exports__DPTPreTrainedModel as DPTPreTrainedModel, __webpack_exports__DacDecoderModel as DacDecoderModel, __webpack_exports__DacDecoderOutput as DacDecoderOutput, __webpack_exports__DacEncoderModel as DacEncoderModel, __webpack_exports__DacEncoderOutput as DacEncoderOutput, __webpack_exports__DacFeatureExtractor as DacFeatureExtractor, __webpack_exports__DacModel as DacModel, __webpack_exports__DacPreTrainedModel as DacPreTrainedModel, __webpack_exports__DataTypeMap as DataTypeMap, __webpack_exports__DebertaForMaskedLM as DebertaForMaskedLM, __webpack_exports__DebertaForQuestionAnswering as DebertaForQuestionAnswering, __webpack_exports__DebertaForSequenceClassification as DebertaForSequenceClassification, __webpack_exports__DebertaForTokenClassification as DebertaForTokenClassification, __webpack_exports__DebertaModel as DebertaModel, __webpack_exports__DebertaPreTrainedModel as DebertaPreTrainedModel, __webpack_exports__DebertaTokenizer as DebertaTokenizer, __webpack_exports__DebertaV2ForMaskedLM as DebertaV2ForMaskedLM, __webpack_exports__DebertaV2ForQuestionAnswering as DebertaV2ForQuestionAnswering, __webpack_exports__DebertaV2ForSequenceClassification as DebertaV2ForSequenceClassification, __webpack_exports__DebertaV2ForTokenClassification as DebertaV2ForTokenClassification, __webpack_exports__DebertaV2Model as DebertaV2Model, __webpack_exports__DebertaV2PreTrainedModel as DebertaV2PreTrainedModel, __webpack_exports__DebertaV2Tokenizer as DebertaV2Tokenizer, __webpack_exports__DecisionTransformerModel as DecisionTransformerModel, __webpack_exports__DecisionTransformerPreTrainedModel as DecisionTransformerPreTrainedModel, __webpack_exports__DeiTFeatureExtractor as DeiTFeatureExtractor, __webpack_exports__DeiTForImageClassification as DeiTForImageClassification, __webpack_exports__DeiTImageProcessor as DeiTImageProcessor, __webpack_exports__DeiTModel as DeiTModel, __webpack_exports__DeiTPreTrainedModel as DeiTPreTrainedModel, __webpack_exports__DepthAnythingForDepthEstimation as DepthAnythingForDepthEstimation, __webpack_exports__DepthAnythingPreTrainedModel as DepthAnythingPreTrainedModel, __webpack_exports__DepthEstimationPipeline as DepthEstimationPipeline, __webpack_exports__DepthProForDepthEstimation as DepthProForDepthEstimation, __webpack_exports__DepthProPreTrainedModel as DepthProPreTrainedModel, __webpack_exports__DetrFeatureExtractor as DetrFeatureExtractor, __webpack_exports__DetrForObjectDetection as DetrForObjectDetection, __webpack_exports__DetrForSegmentation as DetrForSegmentation, __webpack_exports__DetrImageProcessor as DetrImageProcessor, __webpack_exports__DetrModel as DetrModel, __webpack_exports__DetrObjectDetectionOutput as DetrObjectDetectionOutput, __webpack_exports__DetrPreTrainedModel as DetrPreTrainedModel, __webpack_exports__DetrSegmentationOutput as DetrSegmentationOutput, __webpack_exports__Dinov2ForImageClassification as Dinov2ForImageClassification, __webpack_exports__Dinov2Model as Dinov2Model, __webpack_exports__Dinov2PreTrainedModel as Dinov2PreTrainedModel, __webpack_exports__Dinov2WithRegistersForImageClassification as Dinov2WithRegistersForImageClassification, __webpack_exports__Dinov2WithRegistersModel as Dinov2WithRegistersModel, __webpack_exports__Dinov2WithRegistersPreTrainedModel as Dinov2WithRegistersPreTrainedModel, __webpack_exports__DistilBertForMaskedLM as DistilBertForMaskedLM, __webpack_exports__DistilBertForQuestionAnswering as DistilBertForQuestionAnswering, __webpack_exports__DistilBertForSequenceClassification as DistilBertForSequenceClassification, __webpack_exports__DistilBertForTokenClassification as DistilBertForTokenClassification, __webpack_exports__DistilBertModel as DistilBertModel, __webpack_exports__DistilBertPreTrainedModel as DistilBertPreTrainedModel, __webpack_exports__DistilBertTokenizer as DistilBertTokenizer, __webpack_exports__DocumentQuestionAnsweringPipeline as DocumentQuestionAnsweringPipeline, __webpack_exports__DonutFeatureExtractor as DonutFeatureExtractor, __webpack_exports__DonutImageProcessor as DonutImageProcessor, __webpack_exports__DonutSwinModel as DonutSwinModel, __webpack_exports__DonutSwinPreTrainedModel as DonutSwinPreTrainedModel, __webpack_exports__EfficientNetForImageClassification as EfficientNetForImageClassification, __webpack_exports__EfficientNetImageProcessor as EfficientNetImageProcessor, __webpack_exports__EfficientNetModel as EfficientNetModel, __webpack_exports__EfficientNetPreTrainedModel as EfficientNetPreTrainedModel, __webpack_exports__ElectraForMaskedLM as ElectraForMaskedLM, __webpack_exports__ElectraForQuestionAnswering as ElectraForQuestionAnswering, __webpack_exports__ElectraForSequenceClassification as ElectraForSequenceClassification, __webpack_exports__ElectraForTokenClassification as ElectraForTokenClassification, __webpack_exports__ElectraModel as ElectraModel, __webpack_exports__ElectraPreTrainedModel as ElectraPreTrainedModel, __webpack_exports__ElectraTokenizer as ElectraTokenizer, __webpack_exports__EncodecFeatureExtractor as EncodecFeatureExtractor, __webpack_exports__EosTokenCriteria as EosTokenCriteria, __webpack_exports__EsmForMaskedLM as EsmForMaskedLM, __webpack_exports__EsmForSequenceClassification as EsmForSequenceClassification, __webpack_exports__EsmForTokenClassification as EsmForTokenClassification, __webpack_exports__EsmModel as EsmModel, __webpack_exports__EsmPreTrainedModel as EsmPreTrainedModel, __webpack_exports__EsmTokenizer as EsmTokenizer, __webpack_exports__ExaoneForCausalLM as ExaoneForCausalLM, __webpack_exports__ExaoneModel as ExaoneModel, __webpack_exports__ExaonePreTrainedModel as ExaonePreTrainedModel, __webpack_exports__FFT as FFT, __webpack_exports__FalconForCausalLM as FalconForCausalLM, __webpack_exports__FalconModel as FalconModel, __webpack_exports__FalconPreTrainedModel as FalconPreTrainedModel, __webpack_exports__FalconTokenizer as FalconTokenizer, __webpack_exports__FastViTForImageClassification as FastViTForImageClassification, __webpack_exports__FastViTModel as FastViTModel, __webpack_exports__FastViTPreTrainedModel as FastViTPreTrainedModel, __webpack_exports__FeatureExtractionPipeline as FeatureExtractionPipeline, __webpack_exports__FeatureExtractor as FeatureExtractor, __webpack_exports__FillMaskPipeline as FillMaskPipeline, __webpack_exports__Florence2ForConditionalGeneration as Florence2ForConditionalGeneration, __webpack_exports__Florence2PreTrainedModel as Florence2PreTrainedModel, __webpack_exports__Florence2Processor as Florence2Processor, __webpack_exports__ForcedBOSTokenLogitsProcessor as ForcedBOSTokenLogitsProcessor, __webpack_exports__ForcedEOSTokenLogitsProcessor as ForcedEOSTokenLogitsProcessor, __webpack_exports__GLPNFeatureExtractor as GLPNFeatureExtractor, __webpack_exports__GLPNForDepthEstimation as GLPNForDepthEstimation, __webpack_exports__GLPNModel as GLPNModel, __webpack_exports__GLPNPreTrainedModel as GLPNPreTrainedModel, __webpack_exports__GPT2LMHeadModel as GPT2LMHeadModel, __webpack_exports__GPT2Model as GPT2Model, __webpack_exports__GPT2PreTrainedModel as GPT2PreTrainedModel, __webpack_exports__GPT2Tokenizer as GPT2Tokenizer, __webpack_exports__GPTBigCodeForCausalLM as GPTBigCodeForCausalLM, __webpack_exports__GPTBigCodeModel as GPTBigCodeModel, __webpack_exports__GPTBigCodePreTrainedModel as GPTBigCodePreTrainedModel, __webpack_exports__GPTJForCausalLM as GPTJForCausalLM, __webpack_exports__GPTJModel as GPTJModel, __webpack_exports__GPTJPreTrainedModel as GPTJPreTrainedModel, __webpack_exports__GPTNeoForCausalLM as GPTNeoForCausalLM, __webpack_exports__GPTNeoModel as GPTNeoModel, __webpack_exports__GPTNeoPreTrainedModel as GPTNeoPreTrainedModel, __webpack_exports__GPTNeoXForCausalLM as GPTNeoXForCausalLM, __webpack_exports__GPTNeoXModel as GPTNeoXModel, __webpack_exports__GPTNeoXPreTrainedModel as GPTNeoXPreTrainedModel, __webpack_exports__GPTNeoXTokenizer as GPTNeoXTokenizer, __webpack_exports__Gemma2ForCausalLM as Gemma2ForCausalLM, __webpack_exports__Gemma2Model as Gemma2Model, __webpack_exports__Gemma2PreTrainedModel as Gemma2PreTrainedModel, __webpack_exports__Gemma3ForCausalLM as Gemma3ForCausalLM, __webpack_exports__Gemma3Model as Gemma3Model, __webpack_exports__Gemma3PreTrainedModel as Gemma3PreTrainedModel, __webpack_exports__GemmaForCausalLM as GemmaForCausalLM, __webpack_exports__GemmaModel as GemmaModel, __webpack_exports__GemmaPreTrainedModel as GemmaPreTrainedModel, __webpack_exports__GemmaTokenizer as GemmaTokenizer, __webpack_exports__GlmForCausalLM as GlmForCausalLM, __webpack_exports__GlmModel as GlmModel, __webpack_exports__GlmPreTrainedModel as GlmPreTrainedModel, __webpack_exports__GraniteForCausalLM as GraniteForCausalLM, __webpack_exports__GraniteModel as GraniteModel, __webpack_exports__GranitePreTrainedModel as GranitePreTrainedModel, __webpack_exports__Grok1Tokenizer as Grok1Tokenizer, __webpack_exports__GroundingDinoForObjectDetection as GroundingDinoForObjectDetection, __webpack_exports__GroundingDinoImageProcessor as GroundingDinoImageProcessor, __webpack_exports__GroundingDinoPreTrainedModel as GroundingDinoPreTrainedModel, __webpack_exports__GroundingDinoProcessor as GroundingDinoProcessor, __webpack_exports__GroupViTModel as GroupViTModel, __webpack_exports__GroupViTPreTrainedModel as GroupViTPreTrainedModel, __webpack_exports__HeliumForCausalLM as HeliumForCausalLM, __webpack_exports__HeliumModel as HeliumModel, __webpack_exports__HeliumPreTrainedModel as HeliumPreTrainedModel, __webpack_exports__HerbertTokenizer as HerbertTokenizer, __webpack_exports__HieraForImageClassification as HieraForImageClassification, __webpack_exports__HieraModel as HieraModel, __webpack_exports__HieraPreTrainedModel as HieraPreTrainedModel, __webpack_exports__HubertForCTC as HubertForCTC, __webpack_exports__HubertForSequenceClassification as HubertForSequenceClassification, __webpack_exports__HubertModel as HubertModel, __webpack_exports__HubertPreTrainedModel as HubertPreTrainedModel, __webpack_exports__IJepaForImageClassification as IJepaForImageClassification, __webpack_exports__IJepaModel as IJepaModel, __webpack_exports__IJepaPreTrainedModel as IJepaPreTrainedModel, __webpack_exports__Idefics3ForConditionalGeneration as Idefics3ForConditionalGeneration, __webpack_exports__Idefics3ImageProcessor as Idefics3ImageProcessor, __webpack_exports__Idefics3PreTrainedModel as Idefics3PreTrainedModel, __webpack_exports__Idefics3Processor as Idefics3Processor, __webpack_exports__ImageClassificationPipeline as ImageClassificationPipeline, __webpack_exports__ImageFeatureExtractionPipeline as ImageFeatureExtractionPipeline, __webpack_exports__ImageFeatureExtractor as ImageFeatureExtractor, __webpack_exports__ImageMattingOutput as ImageMattingOutput, __webpack_exports__ImageProcessor as ImageProcessor, __webpack_exports__ImageSegmentationPipeline as ImageSegmentationPipeline, __webpack_exports__ImageToImagePipeline as ImageToImagePipeline, __webpack_exports__ImageToTextPipeline as ImageToTextPipeline, __webpack_exports__InterruptableStoppingCriteria as InterruptableStoppingCriteria, __webpack_exports__JAISLMHeadModel as JAISLMHeadModel, __webpack_exports__JAISModel as JAISModel, __webpack_exports__JAISPreTrainedModel as JAISPreTrainedModel, __webpack_exports__JinaCLIPImageProcessor as JinaCLIPImageProcessor, __webpack_exports__JinaCLIPModel as JinaCLIPModel, __webpack_exports__JinaCLIPPreTrainedModel as JinaCLIPPreTrainedModel, __webpack_exports__JinaCLIPProcessor as JinaCLIPProcessor, __webpack_exports__JinaCLIPTextModel as JinaCLIPTextModel, __webpack_exports__JinaCLIPVisionModel as JinaCLIPVisionModel, __webpack_exports__LiteWhisperForConditionalGeneration as LiteWhisperForConditionalGeneration, __webpack_exports__LlamaForCausalLM as LlamaForCausalLM, __webpack_exports__LlamaModel as LlamaModel, __webpack_exports__LlamaPreTrainedModel as LlamaPreTrainedModel, __webpack_exports__LlamaTokenizer as LlamaTokenizer, __webpack_exports__LlavaForConditionalGeneration as LlavaForConditionalGeneration, __webpack_exports__LlavaOnevisionForConditionalGeneration as LlavaOnevisionForConditionalGeneration, __webpack_exports__LlavaOnevisionImageProcessor as LlavaOnevisionImageProcessor, __webpack_exports__LlavaPreTrainedModel as LlavaPreTrainedModel, __webpack_exports__LogitsProcessor as LogitsProcessor, __webpack_exports__LogitsProcessorList as LogitsProcessorList, __webpack_exports__LogitsWarper as LogitsWarper, __webpack_exports__LongT5ForConditionalGeneration as LongT5ForConditionalGeneration, __webpack_exports__LongT5Model as LongT5Model, __webpack_exports__LongT5PreTrainedModel as LongT5PreTrainedModel, __webpack_exports__M2M100ForConditionalGeneration as M2M100ForConditionalGeneration, __webpack_exports__M2M100Model as M2M100Model, __webpack_exports__M2M100PreTrainedModel as M2M100PreTrainedModel, __webpack_exports__M2M100Tokenizer as M2M100Tokenizer, __webpack_exports__MBart50Tokenizer as MBart50Tokenizer, __webpack_exports__MBartForCausalLM as MBartForCausalLM, __webpack_exports__MBartForConditionalGeneration as MBartForConditionalGeneration, __webpack_exports__MBartForSequenceClassification as MBartForSequenceClassification, __webpack_exports__MBartModel as MBartModel, __webpack_exports__MBartPreTrainedModel as MBartPreTrainedModel, __webpack_exports__MBartTokenizer as MBartTokenizer, __webpack_exports__MPNetForMaskedLM as MPNetForMaskedLM, __webpack_exports__MPNetForQuestionAnswering as MPNetForQuestionAnswering, __webpack_exports__MPNetForSequenceClassification as MPNetForSequenceClassification, __webpack_exports__MPNetForTokenClassification as MPNetForTokenClassification, __webpack_exports__MPNetModel as MPNetModel, __webpack_exports__MPNetPreTrainedModel as MPNetPreTrainedModel, __webpack_exports__MPNetTokenizer as MPNetTokenizer, __webpack_exports__MT5ForConditionalGeneration as MT5ForConditionalGeneration, __webpack_exports__MT5Model as MT5Model, __webpack_exports__MT5PreTrainedModel as MT5PreTrainedModel, __webpack_exports__MarianMTModel as MarianMTModel, __webpack_exports__MarianModel as MarianModel, __webpack_exports__MarianPreTrainedModel as MarianPreTrainedModel, __webpack_exports__MarianTokenizer as MarianTokenizer, __webpack_exports__Mask2FormerImageProcessor as Mask2FormerImageProcessor, __webpack_exports__MaskFormerFeatureExtractor as MaskFormerFeatureExtractor, __webpack_exports__MaskFormerForInstanceSegmentation as MaskFormerForInstanceSegmentation, __webpack_exports__MaskFormerImageProcessor as MaskFormerImageProcessor, __webpack_exports__MaskFormerModel as MaskFormerModel, __webpack_exports__MaskFormerPreTrainedModel as MaskFormerPreTrainedModel, __webpack_exports__MaskedLMOutput as MaskedLMOutput, __webpack_exports__MaxLengthCriteria as MaxLengthCriteria, __webpack_exports__Metric3DForDepthEstimation as Metric3DForDepthEstimation, __webpack_exports__Metric3DPreTrainedModel as Metric3DPreTrainedModel, __webpack_exports__Metric3Dv2ForDepthEstimation as Metric3Dv2ForDepthEstimation, __webpack_exports__Metric3Dv2PreTrainedModel as Metric3Dv2PreTrainedModel, __webpack_exports__MgpstrForSceneTextRecognition as MgpstrForSceneTextRecognition, __webpack_exports__MgpstrModelOutput as MgpstrModelOutput, __webpack_exports__MgpstrPreTrainedModel as MgpstrPreTrainedModel, __webpack_exports__MgpstrProcessor as MgpstrProcessor, __webpack_exports__MgpstrTokenizer as MgpstrTokenizer, __webpack_exports__MimiDecoderModel as MimiDecoderModel, __webpack_exports__MimiDecoderOutput as MimiDecoderOutput, __webpack_exports__MimiEncoderModel as MimiEncoderModel, __webpack_exports__MimiEncoderOutput as MimiEncoderOutput, __webpack_exports__MimiModel as MimiModel, __webpack_exports__MimiPreTrainedModel as MimiPreTrainedModel, __webpack_exports__MinLengthLogitsProcessor as MinLengthLogitsProcessor, __webpack_exports__MinNewTokensLengthLogitsProcessor as MinNewTokensLengthLogitsProcessor, __webpack_exports__MistralForCausalLM as MistralForCausalLM, __webpack_exports__MistralModel as MistralModel, __webpack_exports__MistralPreTrainedModel as MistralPreTrainedModel, __webpack_exports__MobileBertForMaskedLM as MobileBertForMaskedLM, __webpack_exports__MobileBertForQuestionAnswering as MobileBertForQuestionAnswering, __webpack_exports__MobileBertForSequenceClassification as MobileBertForSequenceClassification, __webpack_exports__MobileBertModel as MobileBertModel, __webpack_exports__MobileBertPreTrainedModel as MobileBertPreTrainedModel, __webpack_exports__MobileBertTokenizer as MobileBertTokenizer, __webpack_exports__MobileLLMForCausalLM as MobileLLMForCausalLM, __webpack_exports__MobileLLMModel as MobileLLMModel, __webpack_exports__MobileLLMPreTrainedModel as MobileLLMPreTrainedModel, __webpack_exports__MobileNetV1FeatureExtractor as MobileNetV1FeatureExtractor, __webpack_exports__MobileNetV1ForImageClassification as MobileNetV1ForImageClassification, __webpack_exports__MobileNetV1ForSemanticSegmentation as MobileNetV1ForSemanticSegmentation, __webpack_exports__MobileNetV1ImageProcessor as MobileNetV1ImageProcessor, __webpack_exports__MobileNetV1Model as MobileNetV1Model, __webpack_exports__MobileNetV1PreTrainedModel as MobileNetV1PreTrainedModel, __webpack_exports__MobileNetV2FeatureExtractor as MobileNetV2FeatureExtractor, __webpack_exports__MobileNetV2ForImageClassification as MobileNetV2ForImageClassification, __webpack_exports__MobileNetV2ForSemanticSegmentation as MobileNetV2ForSemanticSegmentation, __webpack_exports__MobileNetV2ImageProcessor as MobileNetV2ImageProcessor, __webpack_exports__MobileNetV2Model as MobileNetV2Model, __webpack_exports__MobileNetV2PreTrainedModel as MobileNetV2PreTrainedModel, __webpack_exports__MobileNetV3FeatureExtractor as MobileNetV3FeatureExtractor, __webpack_exports__MobileNetV3ForImageClassification as MobileNetV3ForImageClassification, __webpack_exports__MobileNetV3ForSemanticSegmentation as MobileNetV3ForSemanticSegmentation, __webpack_exports__MobileNetV3ImageProcessor as MobileNetV3ImageProcessor, __webpack_exports__MobileNetV3Model as MobileNetV3Model, __webpack_exports__MobileNetV3PreTrainedModel as MobileNetV3PreTrainedModel, __webpack_exports__MobileNetV4FeatureExtractor as MobileNetV4FeatureExtractor, __webpack_exports__MobileNetV4ForImageClassification as MobileNetV4ForImageClassification, __webpack_exports__MobileNetV4ForSemanticSegmentation as MobileNetV4ForSemanticSegmentation, __webpack_exports__MobileNetV4ImageProcessor as MobileNetV4ImageProcessor, __webpack_exports__MobileNetV4Model as MobileNetV4Model, __webpack_exports__MobileNetV4PreTrainedModel as MobileNetV4PreTrainedModel, __webpack_exports__MobileViTFeatureExtractor as MobileViTFeatureExtractor, __webpack_exports__MobileViTForImageClassification as MobileViTForImageClassification, __webpack_exports__MobileViTImageProcessor as MobileViTImageProcessor, __webpack_exports__MobileViTModel as MobileViTModel, __webpack_exports__MobileViTPreTrainedModel as MobileViTPreTrainedModel, __webpack_exports__MobileViTV2ForImageClassification as MobileViTV2ForImageClassification, __webpack_exports__MobileViTV2Model as MobileViTV2Model, __webpack_exports__MobileViTV2PreTrainedModel as MobileViTV2PreTrainedModel, __webpack_exports__ModelOutput as ModelOutput, __webpack_exports__ModernBertForMaskedLM as ModernBertForMaskedLM, __webpack_exports__ModernBertForSequenceClassification as ModernBertForSequenceClassification, __webpack_exports__ModernBertForTokenClassification as ModernBertForTokenClassification, __webpack_exports__ModernBertModel as ModernBertModel, __webpack_exports__ModernBertPreTrainedModel as ModernBertPreTrainedModel, __webpack_exports__Moondream1ForConditionalGeneration as Moondream1ForConditionalGeneration, __webpack_exports__MoonshineFeatureExtractor as MoonshineFeatureExtractor, __webpack_exports__MoonshineForConditionalGeneration as MoonshineForConditionalGeneration, __webpack_exports__MoonshineModel as MoonshineModel, __webpack_exports__MoonshinePreTrainedModel as MoonshinePreTrainedModel, __webpack_exports__MoonshineProcessor as MoonshineProcessor, __webpack_exports__MptForCausalLM as MptForCausalLM, __webpack_exports__MptModel as MptModel, __webpack_exports__MptPreTrainedModel as MptPreTrainedModel, __webpack_exports__MultiModalityCausalLM as MultiModalityCausalLM, __webpack_exports__MultiModalityPreTrainedModel as MultiModalityPreTrainedModel, __webpack_exports__MusicgenForCausalLM as MusicgenForCausalLM, __webpack_exports__MusicgenForConditionalGeneration as MusicgenForConditionalGeneration, __webpack_exports__MusicgenModel as MusicgenModel, __webpack_exports__MusicgenPreTrainedModel as MusicgenPreTrainedModel, __webpack_exports__NllbTokenizer as NllbTokenizer, __webpack_exports__NoBadWordsLogitsProcessor as NoBadWordsLogitsProcessor, __webpack_exports__NoRepeatNGramLogitsProcessor as NoRepeatNGramLogitsProcessor, __webpack_exports__NomicBertModel as NomicBertModel, __webpack_exports__NomicBertPreTrainedModel as NomicBertPreTrainedModel, __webpack_exports__NougatImageProcessor as NougatImageProcessor, __webpack_exports__NougatTokenizer as NougatTokenizer, __webpack_exports__OPTForCausalLM as OPTForCausalLM, __webpack_exports__OPTModel as OPTModel, __webpack_exports__OPTPreTrainedModel as OPTPreTrainedModel, __webpack_exports__ObjectDetectionPipeline as ObjectDetectionPipeline, __webpack_exports__Olmo2ForCausalLM as Olmo2ForCausalLM, __webpack_exports__Olmo2Model as Olmo2Model, __webpack_exports__Olmo2PreTrainedModel as Olmo2PreTrainedModel, __webpack_exports__OlmoForCausalLM as OlmoForCausalLM, __webpack_exports__OlmoModel as OlmoModel, __webpack_exports__OlmoPreTrainedModel as OlmoPreTrainedModel, __webpack_exports__OpenELMForCausalLM as OpenELMForCausalLM, __webpack_exports__OpenELMModel as OpenELMModel, __webpack_exports__OpenELMPreTrainedModel as OpenELMPreTrainedModel, __webpack_exports__OwlViTFeatureExtractor as OwlViTFeatureExtractor, __webpack_exports__OwlViTForObjectDetection as OwlViTForObjectDetection, __webpack_exports__OwlViTImageProcessor as OwlViTImageProcessor, __webpack_exports__OwlViTModel as OwlViTModel, __webpack_exports__OwlViTPreTrainedModel as OwlViTPreTrainedModel, __webpack_exports__OwlViTProcessor as OwlViTProcessor, __webpack_exports__Owlv2ForObjectDetection as Owlv2ForObjectDetection, __webpack_exports__Owlv2ImageProcessor as Owlv2ImageProcessor, __webpack_exports__Owlv2Model as Owlv2Model, __webpack_exports__Owlv2PreTrainedModel as Owlv2PreTrainedModel, __webpack_exports__PaliGemmaForConditionalGeneration as PaliGemmaForConditionalGeneration, __webpack_exports__PaliGemmaPreTrainedModel as PaliGemmaPreTrainedModel, __webpack_exports__PaliGemmaProcessor as PaliGemmaProcessor, __webpack_exports__PatchTSMixerForPrediction as PatchTSMixerForPrediction, __webpack_exports__PatchTSMixerModel as PatchTSMixerModel, __webpack_exports__PatchTSMixerPreTrainedModel as PatchTSMixerPreTrainedModel, __webpack_exports__PatchTSTForPrediction as PatchTSTForPrediction, __webpack_exports__PatchTSTModel as PatchTSTModel, __webpack_exports__PatchTSTPreTrainedModel as PatchTSTPreTrainedModel, __webpack_exports__Phi3ForCausalLM as Phi3ForCausalLM, __webpack_exports__Phi3Model as Phi3Model, __webpack_exports__Phi3PreTrainedModel as Phi3PreTrainedModel, __webpack_exports__Phi3VForCausalLM as Phi3VForCausalLM, __webpack_exports__Phi3VImageProcessor as Phi3VImageProcessor, __webpack_exports__Phi3VPreTrainedModel as Phi3VPreTrainedModel, __webpack_exports__Phi3VProcessor as Phi3VProcessor, __webpack_exports__PhiForCausalLM as PhiForCausalLM, __webpack_exports__PhiModel as PhiModel, __webpack_exports__PhiPreTrainedModel as PhiPreTrainedModel, __webpack_exports__Pipeline as Pipeline, __webpack_exports__PreTrainedModel as PreTrainedModel, __webpack_exports__PreTrainedTokenizer as PreTrainedTokenizer, __webpack_exports__PretrainedConfig as PretrainedConfig, __webpack_exports__PretrainedMixin as PretrainedMixin, __webpack_exports__Processor as Processor, __webpack_exports__PvtForImageClassification as PvtForImageClassification, __webpack_exports__PvtImageProcessor as PvtImageProcessor, __webpack_exports__PvtModel as PvtModel, __webpack_exports__PvtPreTrainedModel as PvtPreTrainedModel, __webpack_exports__PyAnnoteFeatureExtractor as PyAnnoteFeatureExtractor, __webpack_exports__PyAnnoteForAudioFrameClassification as PyAnnoteForAudioFrameClassification, __webpack_exports__PyAnnoteModel as PyAnnoteModel, __webpack_exports__PyAnnotePreTrainedModel as PyAnnotePreTrainedModel, __webpack_exports__PyAnnoteProcessor as PyAnnoteProcessor, __webpack_exports__QuestionAnsweringModelOutput as QuestionAnsweringModelOutput, __webpack_exports__QuestionAnsweringPipeline as QuestionAnsweringPipeline, __webpack_exports__Qwen2ForCausalLM as Qwen2ForCausalLM, __webpack_exports__Qwen2Model as Qwen2Model, __webpack_exports__Qwen2PreTrainedModel as Qwen2PreTrainedModel, __webpack_exports__Qwen2Tokenizer as Qwen2Tokenizer, __webpack_exports__Qwen2VLForConditionalGeneration as Qwen2VLForConditionalGeneration, __webpack_exports__Qwen2VLImageProcessor as Qwen2VLImageProcessor, __webpack_exports__Qwen2VLPreTrainedModel as Qwen2VLPreTrainedModel, __webpack_exports__Qwen2VLProcessor as Qwen2VLProcessor, __webpack_exports__RFDetrForObjectDetection as RFDetrForObjectDetection, __webpack_exports__RFDetrModel as RFDetrModel, __webpack_exports__RFDetrObjectDetectionOutput as RFDetrObjectDetectionOutput, __webpack_exports__RFDetrPreTrainedModel as RFDetrPreTrainedModel, __webpack_exports__RTDetrForObjectDetection as RTDetrForObjectDetection, __webpack_exports__RTDetrImageProcessor as RTDetrImageProcessor, __webpack_exports__RTDetrModel as RTDetrModel, __webpack_exports__RTDetrObjectDetectionOutput as RTDetrObjectDetectionOutput, __webpack_exports__RTDetrPreTrainedModel as RTDetrPreTrainedModel, __webpack_exports__RTDetrV2ForObjectDetection as RTDetrV2ForObjectDetection, __webpack_exports__RTDetrV2Model as RTDetrV2Model, __webpack_exports__RTDetrV2ObjectDetectionOutput as RTDetrV2ObjectDetectionOutput, __webpack_exports__RTDetrV2PreTrainedModel as RTDetrV2PreTrainedModel, __webpack_exports__RawAudio as RawAudio, __webpack_exports__RawImage as RawImage, __webpack_exports__RawVideo as RawVideo, __webpack_exports__RawVideoFrame as RawVideoFrame, __webpack_exports__RepetitionPenaltyLogitsProcessor as RepetitionPenaltyLogitsProcessor, __webpack_exports__ResNetForImageClassification as ResNetForImageClassification, __webpack_exports__ResNetModel as ResNetModel, __webpack_exports__ResNetPreTrainedModel as ResNetPreTrainedModel, __webpack_exports__RoFormerForMaskedLM as RoFormerForMaskedLM, __webpack_exports__RoFormerForQuestionAnswering as RoFormerForQuestionAnswering, __webpack_exports__RoFormerForSequenceClassification as RoFormerForSequenceClassification, __webpack_exports__RoFormerForTokenClassification as RoFormerForTokenClassification, __webpack_exports__RoFormerModel as RoFormerModel, __webpack_exports__RoFormerPreTrainedModel as RoFormerPreTrainedModel, __webpack_exports__RoFormerTokenizer as RoFormerTokenizer, __webpack_exports__RobertaForMaskedLM as RobertaForMaskedLM, __webpack_exports__RobertaForQuestionAnswering as RobertaForQuestionAnswering, __webpack_exports__RobertaForSequenceClassification as RobertaForSequenceClassification, __webpack_exports__RobertaForTokenClassification as RobertaForTokenClassification, __webpack_exports__RobertaModel as RobertaModel, __webpack_exports__RobertaPreTrainedModel as RobertaPreTrainedModel, __webpack_exports__RobertaTokenizer as RobertaTokenizer, __webpack_exports__SamImageProcessor as SamImageProcessor, __webpack_exports__SamImageSegmentationOutput as SamImageSegmentationOutput, __webpack_exports__SamModel as SamModel, __webpack_exports__SamPreTrainedModel as SamPreTrainedModel, __webpack_exports__SamProcessor as SamProcessor, __webpack_exports__SapiensForDepthEstimation as SapiensForDepthEstimation, __webpack_exports__SapiensForNormalEstimation as SapiensForNormalEstimation, __webpack_exports__SapiensForSemanticSegmentation as SapiensForSemanticSegmentation, __webpack_exports__SapiensPreTrainedModel as SapiensPreTrainedModel, __webpack_exports__SeamlessM4TFeatureExtractor as SeamlessM4TFeatureExtractor, __webpack_exports__SegformerFeatureExtractor as SegformerFeatureExtractor, __webpack_exports__SegformerForImageClassification as SegformerForImageClassification, __webpack_exports__SegformerForSemanticSegmentation as SegformerForSemanticSegmentation, __webpack_exports__SegformerImageProcessor as SegformerImageProcessor, __webpack_exports__SegformerModel as SegformerModel, __webpack_exports__SegformerPreTrainedModel as SegformerPreTrainedModel, __webpack_exports__Seq2SeqLMOutput as Seq2SeqLMOutput, __webpack_exports__SequenceClassifierOutput as SequenceClassifierOutput, __webpack_exports__SiglipImageProcessor as SiglipImageProcessor, __webpack_exports__SiglipModel as SiglipModel, __webpack_exports__SiglipPreTrainedModel as SiglipPreTrainedModel, __webpack_exports__SiglipTextModel as SiglipTextModel, __webpack_exports__SiglipTokenizer as SiglipTokenizer, __webpack_exports__SiglipVisionModel as SiglipVisionModel, __webpack_exports__SmolVLMForConditionalGeneration as SmolVLMForConditionalGeneration, __webpack_exports__SmolVLMImageProcessor as SmolVLMImageProcessor, __webpack_exports__SmolVLMProcessor as SmolVLMProcessor, __webpack_exports__SnacDecoderModel as SnacDecoderModel, __webpack_exports__SnacEncoderModel as SnacEncoderModel, __webpack_exports__SnacFeatureExtractor as SnacFeatureExtractor, __webpack_exports__SnacModel as SnacModel, __webpack_exports__SnacPreTrainedModel as SnacPreTrainedModel, __webpack_exports__SpeechT5FeatureExtractor as SpeechT5FeatureExtractor, __webpack_exports__SpeechT5ForSpeechToText as SpeechT5ForSpeechToText, __webpack_exports__SpeechT5ForTextToSpeech as SpeechT5ForTextToSpeech, __webpack_exports__SpeechT5HifiGan as SpeechT5HifiGan, __webpack_exports__SpeechT5Model as SpeechT5Model, __webpack_exports__SpeechT5PreTrainedModel as SpeechT5PreTrainedModel, __webpack_exports__SpeechT5Processor as SpeechT5Processor, __webpack_exports__SpeechT5Tokenizer as SpeechT5Tokenizer, __webpack_exports__SqueezeBertForMaskedLM as SqueezeBertForMaskedLM, __webpack_exports__SqueezeBertForQuestionAnswering as SqueezeBertForQuestionAnswering, __webpack_exports__SqueezeBertForSequenceClassification as SqueezeBertForSequenceClassification, __webpack_exports__SqueezeBertModel as SqueezeBertModel, __webpack_exports__SqueezeBertPreTrainedModel as SqueezeBertPreTrainedModel, __webpack_exports__SqueezeBertTokenizer as SqueezeBertTokenizer, __webpack_exports__StableLmForCausalLM as StableLmForCausalLM, __webpack_exports__StableLmModel as StableLmModel, __webpack_exports__StableLmPreTrainedModel as StableLmPreTrainedModel, __webpack_exports__Starcoder2ForCausalLM as Starcoder2ForCausalLM, __webpack_exports__Starcoder2Model as Starcoder2Model, __webpack_exports__Starcoder2PreTrainedModel as Starcoder2PreTrainedModel, __webpack_exports__StoppingCriteria as StoppingCriteria, __webpack_exports__StoppingCriteriaList as StoppingCriteriaList, __webpack_exports__StyleTextToSpeech2Model as StyleTextToSpeech2Model, __webpack_exports__StyleTextToSpeech2PreTrainedModel as StyleTextToSpeech2PreTrainedModel, __webpack_exports__SummarizationPipeline as SummarizationPipeline, __webpack_exports__SuppressTokensAtBeginLogitsProcessor as SuppressTokensAtBeginLogitsProcessor, __webpack_exports__Swin2SRForImageSuperResolution as Swin2SRForImageSuperResolution, __webpack_exports__Swin2SRImageProcessor as Swin2SRImageProcessor, __webpack_exports__Swin2SRModel as Swin2SRModel, __webpack_exports__Swin2SRPreTrainedModel as Swin2SRPreTrainedModel, __webpack_exports__SwinForImageClassification as SwinForImageClassification, __webpack_exports__SwinForSemanticSegmentation as SwinForSemanticSegmentation, __webpack_exports__SwinModel as SwinModel, __webpack_exports__SwinPreTrainedModel as SwinPreTrainedModel, __webpack_exports__T5ForConditionalGeneration as T5ForConditionalGeneration, __webpack_exports__T5Model as T5Model, __webpack_exports__T5PreTrainedModel as T5PreTrainedModel, __webpack_exports__T5Tokenizer as T5Tokenizer, __webpack_exports__TableTransformerForObjectDetection as TableTransformerForObjectDetection, __webpack_exports__TableTransformerModel as TableTransformerModel, __webpack_exports__TableTransformerObjectDetectionOutput as TableTransformerObjectDetectionOutput, __webpack_exports__TableTransformerPreTrainedModel as TableTransformerPreTrainedModel, __webpack_exports__TemperatureLogitsWarper as TemperatureLogitsWarper, __webpack_exports__Tensor as Tensor, __webpack_exports__Text2TextGenerationPipeline as Text2TextGenerationPipeline, __webpack_exports__TextClassificationPipeline as TextClassificationPipeline, __webpack_exports__TextGenerationPipeline as TextGenerationPipeline, __webpack_exports__TextStreamer as TextStreamer, __webpack_exports__TextToAudioPipeline as TextToAudioPipeline, __webpack_exports__TokenClassificationPipeline as TokenClassificationPipeline, __webpack_exports__TokenClassifierOutput as TokenClassifierOutput, __webpack_exports__TokenizerModel as TokenizerModel, __webpack_exports__TopKLogitsWarper as TopKLogitsWarper, __webpack_exports__TopPLogitsWarper as TopPLogitsWarper, __webpack_exports__TrOCRForCausalLM as TrOCRForCausalLM, __webpack_exports__TrOCRPreTrainedModel as TrOCRPreTrainedModel, __webpack_exports__TranslationPipeline as TranslationPipeline, __webpack_exports__UltravoxModel as UltravoxModel, __webpack_exports__UltravoxPreTrainedModel as UltravoxPreTrainedModel, __webpack_exports__UltravoxProcessor as UltravoxProcessor, __webpack_exports__UniSpeechForCTC as UniSpeechForCTC, __webpack_exports__UniSpeechForSequenceClassification as UniSpeechForSequenceClassification, __webpack_exports__UniSpeechModel as UniSpeechModel, __webpack_exports__UniSpeechPreTrainedModel as UniSpeechPreTrainedModel, __webpack_exports__UniSpeechSatForAudioFrameClassification as UniSpeechSatForAudioFrameClassification, __webpack_exports__UniSpeechSatForCTC as UniSpeechSatForCTC, __webpack_exports__UniSpeechSatForSequenceClassification as UniSpeechSatForSequenceClassification, __webpack_exports__UniSpeechSatModel as UniSpeechSatModel, __webpack_exports__UniSpeechSatPreTrainedModel as UniSpeechSatPreTrainedModel, __webpack_exports__VLChatProcessor as VLChatProcessor, __webpack_exports__VLMImageProcessor as VLMImageProcessor, __webpack_exports__ViTFeatureExtractor as ViTFeatureExtractor, __webpack_exports__ViTForImageClassification as ViTForImageClassification, __webpack_exports__ViTImageProcessor as ViTImageProcessor, __webpack_exports__ViTMAEModel as ViTMAEModel, __webpack_exports__ViTMAEPreTrainedModel as ViTMAEPreTrainedModel, __webpack_exports__ViTMSNForImageClassification as ViTMSNForImageClassification, __webpack_exports__ViTMSNModel as ViTMSNModel, __webpack_exports__ViTMSNPreTrainedModel as ViTMSNPreTrainedModel, __webpack_exports__ViTModel as ViTModel, __webpack_exports__ViTPreTrainedModel as ViTPreTrainedModel, __webpack_exports__VisionEncoderDecoderModel as VisionEncoderDecoderModel, __webpack_exports__VitMatteForImageMatting as VitMatteForImageMatting, __webpack_exports__VitMatteImageProcessor as VitMatteImageProcessor, __webpack_exports__VitMattePreTrainedModel as VitMattePreTrainedModel, __webpack_exports__VitPoseForPoseEstimation as VitPoseForPoseEstimation, __webpack_exports__VitPoseImageProcessor as VitPoseImageProcessor, __webpack_exports__VitPosePreTrainedModel as VitPosePreTrainedModel, __webpack_exports__VitsModel as VitsModel, __webpack_exports__VitsModelOutput as VitsModelOutput, __webpack_exports__VitsPreTrainedModel as VitsPreTrainedModel, __webpack_exports__VitsTokenizer as VitsTokenizer, __webpack_exports__Wav2Vec2BertForCTC as Wav2Vec2BertForCTC, __webpack_exports__Wav2Vec2BertForSequenceClassification as Wav2Vec2BertForSequenceClassification, __webpack_exports__Wav2Vec2BertModel as Wav2Vec2BertModel, __webpack_exports__Wav2Vec2BertPreTrainedModel as Wav2Vec2BertPreTrainedModel, __webpack_exports__Wav2Vec2CTCTokenizer as Wav2Vec2CTCTokenizer, __webpack_exports__Wav2Vec2FeatureExtractor as Wav2Vec2FeatureExtractor, __webpack_exports__Wav2Vec2ForAudioFrameClassification as Wav2Vec2ForAudioFrameClassification, __webpack_exports__Wav2Vec2ForCTC as Wav2Vec2ForCTC, __webpack_exports__Wav2Vec2ForSequenceClassification as Wav2Vec2ForSequenceClassification, __webpack_exports__Wav2Vec2Model as Wav2Vec2Model, __webpack_exports__Wav2Vec2PreTrainedModel as Wav2Vec2PreTrainedModel, __webpack_exports__Wav2Vec2Processor as Wav2Vec2Processor, __webpack_exports__Wav2Vec2ProcessorWithLM as Wav2Vec2ProcessorWithLM, __webpack_exports__WavLMForAudioFrameClassification as WavLMForAudioFrameClassification, __webpack_exports__WavLMForCTC as WavLMForCTC, __webpack_exports__WavLMForSequenceClassification as WavLMForSequenceClassification, __webpack_exports__WavLMForXVector as WavLMForXVector, __webpack_exports__WavLMModel as WavLMModel, __webpack_exports__WavLMPreTrainedModel as WavLMPreTrainedModel, __webpack_exports__WeSpeakerFeatureExtractor as WeSpeakerFeatureExtractor, __webpack_exports__WeSpeakerResNetModel as WeSpeakerResNetModel, __webpack_exports__WeSpeakerResNetPreTrainedModel as WeSpeakerResNetPreTrainedModel, __webpack_exports__WhisperFeatureExtractor as WhisperFeatureExtractor, __webpack_exports__WhisperForConditionalGeneration as WhisperForConditionalGeneration, __webpack_exports__WhisperModel as WhisperModel, __webpack_exports__WhisperPreTrainedModel as WhisperPreTrainedModel, __webpack_exports__WhisperProcessor as WhisperProcessor, __webpack_exports__WhisperTextStreamer as WhisperTextStreamer, __webpack_exports__WhisperTimeStampLogitsProcessor as WhisperTimeStampLogitsProcessor, __webpack_exports__WhisperTokenizer as WhisperTokenizer, __webpack_exports__XLMForQuestionAnswering as XLMForQuestionAnswering, __webpack_exports__XLMForSequenceClassification as XLMForSequenceClassification, __webpack_exports__XLMForTokenClassification as XLMForTokenClassification, __webpack_exports__XLMModel as XLMModel, __webpack_exports__XLMPreTrainedModel as XLMPreTrainedModel, __webpack_exports__XLMRobertaForMaskedLM as XLMRobertaForMaskedLM, __webpack_exports__XLMRobertaForQuestionAnswering as XLMRobertaForQuestionAnswering, __webpack_exports__XLMRobertaForSequenceClassification as XLMRobertaForSequenceClassification, __webpack_exports__XLMRobertaForTokenClassification as XLMRobertaForTokenClassification, __webpack_exports__XLMRobertaModel as XLMRobertaModel, __webpack_exports__XLMRobertaPreTrainedModel as XLMRobertaPreTrainedModel, __webpack_exports__XLMRobertaTokenizer as XLMRobertaTokenizer, __webpack_exports__XLMTokenizer as XLMTokenizer, __webpack_exports__XLMWithLMHeadModel as XLMWithLMHeadModel, __webpack_exports__XVectorOutput as XVectorOutput, __webpack_exports__YolosFeatureExtractor as YolosFeatureExtractor, __webpack_exports__YolosForObjectDetection as YolosForObjectDetection, __webpack_exports__YolosImageProcessor as YolosImageProcessor, __webpack_exports__YolosModel as YolosModel, __webpack_exports__YolosObjectDetectionOutput as YolosObjectDetectionOutput, __webpack_exports__YolosPreTrainedModel as YolosPreTrainedModel, __webpack_exports__ZeroShotAudioClassificationPipeline as ZeroShotAudioClassificationPipeline, __webpack_exports__ZeroShotClassificationPipeline as ZeroShotClassificationPipeline, __webpack_exports__ZeroShotImageClassificationPipeline as ZeroShotImageClassificationPipeline, __webpack_exports__ZeroShotObjectDetectionPipeline as ZeroShotObjectDetectionPipeline, __webpack_exports__bankers_round as bankers_round, __webpack_exports__cat as cat, __webpack_exports__cos_sim as cos_sim, __webpack_exports__dot as dot, __webpack_exports__dynamic_time_warping as dynamic_time_warping, __webpack_exports__env as env, __webpack_exports__full as full, __webpack_exports__full_like as full_like, __webpack_exports__getKeyValueShapes as getKeyValueShapes, __webpack_exports__hamming as hamming, __webpack_exports__hanning as hanning, __webpack_exports__interpolate as interpolate, __webpack_exports__interpolate_4d as interpolate_4d, __webpack_exports__interpolate_data as interpolate_data, __webpack_exports__is_chinese_char as is_chinese_char, __webpack_exports__layer_norm as layer_norm, __webpack_exports__load_image as load_image, __webpack_exports__load_video as load_video, __webpack_exports__log_softmax as log_softmax, __webpack_exports__magnitude as magnitude, __webpack_exports__matmul as matmul, __webpack_exports__max as max, __webpack_exports__mean as mean, __webpack_exports__mean_pooling as mean_pooling, __webpack_exports__medianFilter as medianFilter, __webpack_exports__mel_filter_bank as mel_filter_bank, __webpack_exports__min as min, __webpack_exports__ones as ones, __webpack_exports__ones_like as ones_like, __webpack_exports__permute as permute, __webpack_exports__permute_data as permute_data, __webpack_exports__pipeline as pipeline, __webpack_exports__quantize_embeddings as quantize_embeddings, __webpack_exports__rand as rand, __webpack_exports__read_audio as read_audio, __webpack_exports__rfft as rfft, __webpack_exports__round as round, __webpack_exports__slice as slice, __webpack_exports__softmax as softmax, __webpack_exports__spectrogram as spectrogram, __webpack_exports__stack as stack, __webpack_exports__std_mean as std_mean, __webpack_exports__topk as topk, __webpack_exports__window_function as window_function, __webpack_exports__zeros as zeros, __webpack_exports__zeros_like as zeros_like };
|
|
40532
40859
|
|
|
40533
40860
|
//# sourceMappingURL=transformers.js.map
|