@huggingface/transformers 3.6.0 → 3.6.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +5 -2
- package/dist/transformers.js +130 -7
- package/dist/transformers.js.map +1 -1
- package/dist/transformers.min.js +1 -1
- package/dist/transformers.min.js.map +1 -1
- package/dist/transformers.node.cjs +116 -6
- package/dist/transformers.node.cjs.map +1 -1
- package/dist/transformers.node.min.cjs +1 -1
- package/dist/transformers.node.min.cjs.map +1 -1
- package/dist/transformers.node.min.mjs +1 -1
- package/dist/transformers.node.min.mjs.map +1 -1
- package/dist/transformers.node.mjs +130 -7
- package/dist/transformers.node.mjs.map +1 -1
- package/dist/transformers.web.js +130 -7
- package/dist/transformers.web.js.map +1 -1
- package/dist/transformers.web.min.js +1 -1
- package/dist/transformers.web.min.js.map +1 -1
- package/package.json +1 -1
- package/src/base/feature_extraction_utils.js +1 -1
- package/src/base/image_processors_utils.js +1 -1
- package/src/base/processing_utils.js +1 -1
- package/src/configs.js +2 -0
- package/src/env.js +2 -2
- package/src/models.js +79 -0
- package/src/tokenizers.js +3 -0
- package/src/utils/image.js +1 -1
- package/types/base/feature_extraction_utils.d.ts +1 -1
- package/types/base/feature_extraction_utils.d.ts.map +1 -1
- package/types/base/image_processors_utils.d.ts +1 -1
- package/types/base/image_processors_utils.d.ts.map +1 -1
- package/types/base/processing_utils.d.ts +1 -1
- package/types/base/processing_utils.d.ts.map +1 -1
- package/types/configs.d.ts.map +1 -1
- package/types/models/auto/feature_extraction_auto.d.ts +1 -1
- package/types/models/auto/feature_extraction_auto.d.ts.map +1 -1
- package/types/models/auto/image_processing_auto.d.ts +1 -1
- package/types/models/auto/image_processing_auto.d.ts.map +1 -1
- package/types/models/auto/processing_auto.d.ts +1 -1
- package/types/models.d.ts +52 -0
- package/types/models.d.ts.map +1 -1
- package/types/tokenizers.d.ts +3 -0
- package/types/tokenizers.d.ts.map +1 -1
|
@@ -2535,7 +2535,7 @@ class FeatureExtractor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_1__.Ca
|
|
|
2535
2535
|
*
|
|
2536
2536
|
* @returns {Promise<FeatureExtractor>} A new instance of the Feature Extractor class.
|
|
2537
2537
|
*/
|
|
2538
|
-
static async from_pretrained(pretrained_model_name_or_path, options) {
|
|
2538
|
+
static async from_pretrained(pretrained_model_name_or_path, options={}) {
|
|
2539
2539
|
const config = await (0,_utils_hub_js__WEBPACK_IMPORTED_MODULE_2__.getModelJSON)(pretrained_model_name_or_path, _utils_constants_js__WEBPACK_IMPORTED_MODULE_0__.FEATURE_EXTRACTOR_NAME, true, options);
|
|
2540
2540
|
return new this(config);
|
|
2541
2541
|
}
|
|
@@ -3683,7 +3683,7 @@ class ImageProcessor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_0__.Call
|
|
|
3683
3683
|
*
|
|
3684
3684
|
* @returns {Promise<ImageProcessor>} A new instance of the Processor class.
|
|
3685
3685
|
*/
|
|
3686
|
-
static async from_pretrained(pretrained_model_name_or_path, options) {
|
|
3686
|
+
static async from_pretrained(pretrained_model_name_or_path, options={}) {
|
|
3687
3687
|
const preprocessorConfig = await (0,_utils_hub_js__WEBPACK_IMPORTED_MODULE_5__.getModelJSON)(pretrained_model_name_or_path, _utils_constants_js__WEBPACK_IMPORTED_MODULE_6__.IMAGE_PROCESSOR_NAME, true, options);
|
|
3688
3688
|
return new this(preprocessorConfig);
|
|
3689
3689
|
}
|
|
@@ -3854,7 +3854,7 @@ class Processor extends _utils_generic_js__WEBPACK_IMPORTED_MODULE_1__.Callable
|
|
|
3854
3854
|
*
|
|
3855
3855
|
* @returns {Promise<Processor>} A new instance of the Processor class.
|
|
3856
3856
|
*/
|
|
3857
|
-
static async from_pretrained(pretrained_model_name_or_path, options) {
|
|
3857
|
+
static async from_pretrained(pretrained_model_name_or_path, options={}) {
|
|
3858
3858
|
|
|
3859
3859
|
const [config, components, chat_template] = await Promise.all([
|
|
3860
3860
|
// TODO:
|
|
@@ -4005,6 +4005,7 @@ function getNormalizedConfig(config) {
|
|
|
4005
4005
|
mapping['hidden_size'] = 'hidden_size';
|
|
4006
4006
|
break;
|
|
4007
4007
|
case 'llama':
|
|
4008
|
+
case 'smollm3':
|
|
4008
4009
|
case 'olmo':
|
|
4009
4010
|
case 'olmo2':
|
|
4010
4011
|
case 'mobilellm':
|
|
@@ -4030,6 +4031,7 @@ function getNormalizedConfig(config) {
|
|
|
4030
4031
|
case 'gemma3n_text':
|
|
4031
4032
|
case 'glm':
|
|
4032
4033
|
case 'helium':
|
|
4034
|
+
case 'ernie4_5':
|
|
4033
4035
|
mapping['num_heads'] = 'num_key_value_heads';
|
|
4034
4036
|
mapping['num_layers'] = 'num_hidden_layers';
|
|
4035
4037
|
mapping['dim_kv'] = 'head_dim';
|
|
@@ -4365,11 +4367,11 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
4365
4367
|
|
|
4366
4368
|
|
|
4367
4369
|
|
|
4368
|
-
const VERSION = '3.6.
|
|
4370
|
+
const VERSION = '3.6.2';
|
|
4369
4371
|
|
|
4370
4372
|
// Check if various APIs are available (depends on environment)
|
|
4371
4373
|
const IS_BROWSER_ENV = typeof window !== "undefined" && typeof window.document !== "undefined";
|
|
4372
|
-
const IS_WEBWORKER_ENV = typeof self !== "undefined"
|
|
4374
|
+
const IS_WEBWORKER_ENV = typeof self !== "undefined" && (['DedicatedWorkerGlobalScope', 'ServiceWorkerGlobalScope', 'SharedWorkerGlobalScope'].includes(self.constructor?.name));
|
|
4373
4375
|
const IS_WEB_CACHE_AVAILABLE = typeof self !== "undefined" && 'caches' in self;
|
|
4374
4376
|
const IS_WEBGPU_AVAILABLE = typeof navigator !== 'undefined' && 'gpu' in navigator;
|
|
4375
4377
|
const IS_WEBNN_AVAILABLE = typeof navigator !== 'undefined' && 'ml' in navigator;
|
|
@@ -6486,6 +6488,9 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
6486
6488
|
/* harmony export */ ElectraForTokenClassification: () => (/* binding */ ElectraForTokenClassification),
|
|
6487
6489
|
/* harmony export */ ElectraModel: () => (/* binding */ ElectraModel),
|
|
6488
6490
|
/* harmony export */ ElectraPreTrainedModel: () => (/* binding */ ElectraPreTrainedModel),
|
|
6491
|
+
/* harmony export */ Ernie4_5_ForCausalLM: () => (/* binding */ Ernie4_5_ForCausalLM),
|
|
6492
|
+
/* harmony export */ Ernie4_5_Model: () => (/* binding */ Ernie4_5_Model),
|
|
6493
|
+
/* harmony export */ Ernie4_5_PretrainedModel: () => (/* binding */ Ernie4_5_PretrainedModel),
|
|
6489
6494
|
/* harmony export */ EsmForMaskedLM: () => (/* binding */ EsmForMaskedLM),
|
|
6490
6495
|
/* harmony export */ EsmForSequenceClassification: () => (/* binding */ EsmForSequenceClassification),
|
|
6491
6496
|
/* harmony export */ EsmForTokenClassification: () => (/* binding */ EsmForTokenClassification),
|
|
@@ -6664,6 +6669,12 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
6664
6669
|
/* harmony export */ MusicgenForConditionalGeneration: () => (/* binding */ MusicgenForConditionalGeneration),
|
|
6665
6670
|
/* harmony export */ MusicgenModel: () => (/* binding */ MusicgenModel),
|
|
6666
6671
|
/* harmony export */ MusicgenPreTrainedModel: () => (/* binding */ MusicgenPreTrainedModel),
|
|
6672
|
+
/* harmony export */ NeoBertForMaskedLM: () => (/* binding */ NeoBertForMaskedLM),
|
|
6673
|
+
/* harmony export */ NeoBertForQuestionAnswering: () => (/* binding */ NeoBertForQuestionAnswering),
|
|
6674
|
+
/* harmony export */ NeoBertForSequenceClassification: () => (/* binding */ NeoBertForSequenceClassification),
|
|
6675
|
+
/* harmony export */ NeoBertForTokenClassification: () => (/* binding */ NeoBertForTokenClassification),
|
|
6676
|
+
/* harmony export */ NeoBertModel: () => (/* binding */ NeoBertModel),
|
|
6677
|
+
/* harmony export */ NeoBertPreTrainedModel: () => (/* binding */ NeoBertPreTrainedModel),
|
|
6667
6678
|
/* harmony export */ NomicBertModel: () => (/* binding */ NomicBertModel),
|
|
6668
6679
|
/* harmony export */ NomicBertPreTrainedModel: () => (/* binding */ NomicBertPreTrainedModel),
|
|
6669
6680
|
/* harmony export */ OPTForCausalLM: () => (/* binding */ OPTForCausalLM),
|
|
@@ -6761,6 +6772,9 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
6761
6772
|
/* harmony export */ SiglipPreTrainedModel: () => (/* binding */ SiglipPreTrainedModel),
|
|
6762
6773
|
/* harmony export */ SiglipTextModel: () => (/* binding */ SiglipTextModel),
|
|
6763
6774
|
/* harmony export */ SiglipVisionModel: () => (/* binding */ SiglipVisionModel),
|
|
6775
|
+
/* harmony export */ SmolLM3ForCausalLM: () => (/* binding */ SmolLM3ForCausalLM),
|
|
6776
|
+
/* harmony export */ SmolLM3Model: () => (/* binding */ SmolLM3Model),
|
|
6777
|
+
/* harmony export */ SmolLM3PreTrainedModel: () => (/* binding */ SmolLM3PreTrainedModel),
|
|
6764
6778
|
/* harmony export */ SmolVLMForConditionalGeneration: () => (/* binding */ SmolVLMForConditionalGeneration),
|
|
6765
6779
|
/* harmony export */ SnacDecoderModel: () => (/* binding */ SnacDecoderModel),
|
|
6766
6780
|
/* harmony export */ SnacEncoderModel: () => (/* binding */ SnacEncoderModel),
|
|
@@ -8962,6 +8976,60 @@ class BertForQuestionAnswering extends BertPreTrainedModel {
|
|
|
8962
8976
|
}
|
|
8963
8977
|
//////////////////////////////////////////////////
|
|
8964
8978
|
|
|
8979
|
+
//////////////////////////////////////////////////
|
|
8980
|
+
// NeoBert models
|
|
8981
|
+
class NeoBertPreTrainedModel extends PreTrainedModel { }
|
|
8982
|
+
class NeoBertModel extends NeoBertPreTrainedModel { }
|
|
8983
|
+
|
|
8984
|
+
class NeoBertForMaskedLM extends NeoBertPreTrainedModel {
|
|
8985
|
+
/**
|
|
8986
|
+
* Calls the model on new inputs.
|
|
8987
|
+
*
|
|
8988
|
+
* @param {Object} model_inputs The inputs to the model.
|
|
8989
|
+
* @returns {Promise<MaskedLMOutput>} An object containing the model's output logits for masked language modeling.
|
|
8990
|
+
*/
|
|
8991
|
+
async _call(model_inputs) {
|
|
8992
|
+
return new MaskedLMOutput(await super._call(model_inputs));
|
|
8993
|
+
}
|
|
8994
|
+
}
|
|
8995
|
+
|
|
8996
|
+
class NeoBertForSequenceClassification extends NeoBertPreTrainedModel {
|
|
8997
|
+
/**
|
|
8998
|
+
* Calls the model on new inputs.
|
|
8999
|
+
*
|
|
9000
|
+
* @param {Object} model_inputs The inputs to the model.
|
|
9001
|
+
* @returns {Promise<SequenceClassifierOutput>} An object containing the model's output logits for sequence classification.
|
|
9002
|
+
*/
|
|
9003
|
+
async _call(model_inputs) {
|
|
9004
|
+
return new SequenceClassifierOutput(await super._call(model_inputs));
|
|
9005
|
+
}
|
|
9006
|
+
}
|
|
9007
|
+
|
|
9008
|
+
class NeoBertForTokenClassification extends NeoBertPreTrainedModel {
|
|
9009
|
+
/**
|
|
9010
|
+
* Calls the model on new inputs.
|
|
9011
|
+
*
|
|
9012
|
+
* @param {Object} model_inputs The inputs to the model.
|
|
9013
|
+
* @returns {Promise<TokenClassifierOutput>} An object containing the model's output logits for token classification.
|
|
9014
|
+
*/
|
|
9015
|
+
async _call(model_inputs) {
|
|
9016
|
+
return new TokenClassifierOutput(await super._call(model_inputs));
|
|
9017
|
+
}
|
|
9018
|
+
}
|
|
9019
|
+
|
|
9020
|
+
class NeoBertForQuestionAnswering extends NeoBertPreTrainedModel {
|
|
9021
|
+
/**
|
|
9022
|
+
* Calls the model on new inputs.
|
|
9023
|
+
*
|
|
9024
|
+
* @param {Object} model_inputs The inputs to the model.
|
|
9025
|
+
* @returns {Promise<QuestionAnsweringModelOutput>} An object containing the model's output logits for question answering.
|
|
9026
|
+
*/
|
|
9027
|
+
async _call(model_inputs) {
|
|
9028
|
+
return new QuestionAnsweringModelOutput(await super._call(model_inputs));
|
|
9029
|
+
}
|
|
9030
|
+
}
|
|
9031
|
+
//////////////////////////////////////////////////
|
|
9032
|
+
|
|
8965
9033
|
//////////////////////////////////////////////////
|
|
8966
9034
|
// ModernBert models
|
|
8967
9035
|
class ModernBertPreTrainedModel extends PreTrainedModel { }
|
|
@@ -11362,6 +11430,13 @@ class LlamaModel extends LlamaPreTrainedModel { }
|
|
|
11362
11430
|
class LlamaForCausalLM extends LlamaPreTrainedModel { }
|
|
11363
11431
|
//////////////////////////////////////////////////
|
|
11364
11432
|
|
|
11433
|
+
//////////////////////////////////////////////////
|
|
11434
|
+
// SmolLM3 models
|
|
11435
|
+
class SmolLM3PreTrainedModel extends PreTrainedModel { }
|
|
11436
|
+
class SmolLM3Model extends SmolLM3PreTrainedModel { }
|
|
11437
|
+
class SmolLM3ForCausalLM extends SmolLM3PreTrainedModel { }
|
|
11438
|
+
//////////////////////////////////////////////////
|
|
11439
|
+
|
|
11365
11440
|
//////////////////////////////////////////////////
|
|
11366
11441
|
// Helium models
|
|
11367
11442
|
class HeliumPreTrainedModel extends PreTrainedModel { }
|
|
@@ -13508,6 +13583,15 @@ class MistralModel extends MistralPreTrainedModel { }
|
|
|
13508
13583
|
class MistralForCausalLM extends MistralPreTrainedModel { }
|
|
13509
13584
|
//////////////////////////////////////////////////
|
|
13510
13585
|
|
|
13586
|
+
//////////////////////////////////////////////////
|
|
13587
|
+
// ERNIE-4.5 models
|
|
13588
|
+
class Ernie4_5_PretrainedModel extends PreTrainedModel { }
|
|
13589
|
+
|
|
13590
|
+
class Ernie4_5_Model extends Ernie4_5_PretrainedModel { }
|
|
13591
|
+
|
|
13592
|
+
class Ernie4_5_ForCausalLM extends Ernie4_5_PretrainedModel { }
|
|
13593
|
+
//////////////////////////////////////////////////
|
|
13594
|
+
|
|
13511
13595
|
|
|
13512
13596
|
//////////////////////////////////////////////////
|
|
13513
13597
|
// Starcoder2 models
|
|
@@ -14449,6 +14533,7 @@ class PretrainedMixin {
|
|
|
14449
14533
|
|
|
14450
14534
|
const MODEL_MAPPING_NAMES_ENCODER_ONLY = new Map([
|
|
14451
14535
|
['bert', ['BertModel', BertModel]],
|
|
14536
|
+
['neobert', ['NeoBertModel', NeoBertModel]],
|
|
14452
14537
|
['modernbert', ['ModernBertModel', ModernBertModel]],
|
|
14453
14538
|
['nomic_bert', ['NomicBertModel', NomicBertModel]],
|
|
14454
14539
|
['roformer', ['RoFormerModel', RoFormerModel]],
|
|
@@ -14562,6 +14647,7 @@ const MODEL_MAPPING_NAMES_DECODER_ONLY = new Map([
|
|
|
14562
14647
|
['gpt_neox', ['GPTNeoXModel', GPTNeoXModel]],
|
|
14563
14648
|
['codegen', ['CodeGenModel', CodeGenModel]],
|
|
14564
14649
|
['llama', ['LlamaModel', LlamaModel]],
|
|
14650
|
+
['smollm3', ['SmolLM3Model', SmolLM3Model]],
|
|
14565
14651
|
['exaone', ['ExaoneModel', ExaoneModel]],
|
|
14566
14652
|
['olmo', ['OlmoModel', OlmoModel]],
|
|
14567
14653
|
['olmo2', ['Olmo2Model', Olmo2Model]],
|
|
@@ -14581,6 +14667,7 @@ const MODEL_MAPPING_NAMES_DECODER_ONLY = new Map([
|
|
|
14581
14667
|
['mpt', ['MptModel', MptModel]],
|
|
14582
14668
|
['opt', ['OPTModel', OPTModel]],
|
|
14583
14669
|
['mistral', ['MistralModel', MistralModel]],
|
|
14670
|
+
['ernie4_5', ['Ernie4_5_Model', Ernie4_5_Model]],
|
|
14584
14671
|
['starcoder2', ['Starcoder2Model', Starcoder2Model]],
|
|
14585
14672
|
['falcon', ['FalconModel', FalconModel]],
|
|
14586
14673
|
['stablelm', ['StableLmModel', StableLmModel]],
|
|
@@ -14604,6 +14691,7 @@ const MODEL_FOR_TEXT_TO_WAVEFORM_MAPPING_NAMES = new Map([
|
|
|
14604
14691
|
|
|
14605
14692
|
const MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES = new Map([
|
|
14606
14693
|
['bert', ['BertForSequenceClassification', BertForSequenceClassification]],
|
|
14694
|
+
['neobert', ['NeoBertForSequenceClassification', NeoBertForSequenceClassification]],
|
|
14607
14695
|
['modernbert', ['ModernBertForSequenceClassification', ModernBertForSequenceClassification]],
|
|
14608
14696
|
['roformer', ['RoFormerForSequenceClassification', RoFormerForSequenceClassification]],
|
|
14609
14697
|
['electra', ['ElectraForSequenceClassification', ElectraForSequenceClassification]],
|
|
@@ -14626,6 +14714,7 @@ const MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES = new Map([
|
|
|
14626
14714
|
|
|
14627
14715
|
const MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES = new Map([
|
|
14628
14716
|
['bert', ['BertForTokenClassification', BertForTokenClassification]],
|
|
14717
|
+
['neobert', ['NeoBertForTokenClassification', NeoBertForTokenClassification]],
|
|
14629
14718
|
['modernbert', ['ModernBertForTokenClassification', ModernBertForTokenClassification]],
|
|
14630
14719
|
['roformer', ['RoFormerForTokenClassification', RoFormerForTokenClassification]],
|
|
14631
14720
|
['electra', ['ElectraForTokenClassification', ElectraForTokenClassification]],
|
|
@@ -14663,6 +14752,7 @@ const MODEL_FOR_CAUSAL_LM_MAPPING_NAMES = new Map([
|
|
|
14663
14752
|
['gpt_neox', ['GPTNeoXForCausalLM', GPTNeoXForCausalLM]],
|
|
14664
14753
|
['codegen', ['CodeGenForCausalLM', CodeGenForCausalLM]],
|
|
14665
14754
|
['llama', ['LlamaForCausalLM', LlamaForCausalLM]],
|
|
14755
|
+
['smollm3', ['SmolLM3ForCausalLM', SmolLM3ForCausalLM]],
|
|
14666
14756
|
['exaone', ['ExaoneForCausalLM', ExaoneForCausalLM]],
|
|
14667
14757
|
['olmo', ['OlmoForCausalLM', OlmoForCausalLM]],
|
|
14668
14758
|
['olmo2', ['Olmo2ForCausalLM', Olmo2ForCausalLM]],
|
|
@@ -14683,6 +14773,7 @@ const MODEL_FOR_CAUSAL_LM_MAPPING_NAMES = new Map([
|
|
|
14683
14773
|
['opt', ['OPTForCausalLM', OPTForCausalLM]],
|
|
14684
14774
|
['mbart', ['MBartForCausalLM', MBartForCausalLM]],
|
|
14685
14775
|
['mistral', ['MistralForCausalLM', MistralForCausalLM]],
|
|
14776
|
+
['ernie4_5', ['Ernie4_5_ForCausalLM', Ernie4_5_ForCausalLM]],
|
|
14686
14777
|
['starcoder2', ['Starcoder2ForCausalLM', Starcoder2ForCausalLM]],
|
|
14687
14778
|
['falcon', ['FalconForCausalLM', FalconForCausalLM]],
|
|
14688
14779
|
['trocr', ['TrOCRForCausalLM', TrOCRForCausalLM]],
|
|
@@ -14699,6 +14790,7 @@ const MODEL_FOR_MULTIMODALITY_MAPPING_NAMES = new Map([
|
|
|
14699
14790
|
|
|
14700
14791
|
const MODEL_FOR_MASKED_LM_MAPPING_NAMES = new Map([
|
|
14701
14792
|
['bert', ['BertForMaskedLM', BertForMaskedLM]],
|
|
14793
|
+
['neobert', ['NeoBertForMaskedLM', NeoBertForMaskedLM]],
|
|
14702
14794
|
['modernbert', ['ModernBertForMaskedLM', ModernBertForMaskedLM]],
|
|
14703
14795
|
['roformer', ['RoFormerForMaskedLM', RoFormerForMaskedLM]],
|
|
14704
14796
|
['electra', ['ElectraForMaskedLM', ElectraForMaskedLM]],
|
|
@@ -14719,6 +14811,7 @@ const MODEL_FOR_MASKED_LM_MAPPING_NAMES = new Map([
|
|
|
14719
14811
|
|
|
14720
14812
|
const MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES = new Map([
|
|
14721
14813
|
['bert', ['BertForQuestionAnswering', BertForQuestionAnswering]],
|
|
14814
|
+
['neobert', ['NeoBertForQuestionAnswering', NeoBertForQuestionAnswering]],
|
|
14722
14815
|
['roformer', ['RoFormerForQuestionAnswering', RoFormerForQuestionAnswering]],
|
|
14723
14816
|
['electra', ['ElectraForQuestionAnswering', ElectraForQuestionAnswering]],
|
|
14724
14817
|
['convbert', ['ConvBertForQuestionAnswering', ConvBertForQuestionAnswering]],
|
|
@@ -24403,6 +24496,7 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
24403
24496
|
/* harmony export */ DebertaV2Tokenizer: () => (/* binding */ DebertaV2Tokenizer),
|
|
24404
24497
|
/* harmony export */ DistilBertTokenizer: () => (/* binding */ DistilBertTokenizer),
|
|
24405
24498
|
/* harmony export */ ElectraTokenizer: () => (/* binding */ ElectraTokenizer),
|
|
24499
|
+
/* harmony export */ Ernie4_5_Tokenizer: () => (/* binding */ Ernie4_5_Tokenizer),
|
|
24406
24500
|
/* harmony export */ EsmTokenizer: () => (/* binding */ EsmTokenizer),
|
|
24407
24501
|
/* harmony export */ FalconTokenizer: () => (/* binding */ FalconTokenizer),
|
|
24408
24502
|
/* harmony export */ GPT2Tokenizer: () => (/* binding */ GPT2Tokenizer),
|
|
@@ -28751,6 +28845,8 @@ class CohereTokenizer extends PreTrainedTokenizer { }
|
|
|
28751
28845
|
|
|
28752
28846
|
class MgpstrTokenizer extends PreTrainedTokenizer { }
|
|
28753
28847
|
|
|
28848
|
+
class Ernie4_5_Tokenizer extends PreTrainedTokenizer { }
|
|
28849
|
+
|
|
28754
28850
|
/**
|
|
28755
28851
|
* Helper class which is used to instantiate pretrained tokenizers with the `from_pretrained` function.
|
|
28756
28852
|
* The chosen tokenizer class is determined by the type specified in the tokenizer config.
|
|
@@ -28805,6 +28901,7 @@ class AutoTokenizer {
|
|
|
28805
28901
|
Grok1Tokenizer,
|
|
28806
28902
|
CohereTokenizer,
|
|
28807
28903
|
MgpstrTokenizer,
|
|
28904
|
+
Ernie4_5_Tokenizer,
|
|
28808
28905
|
|
|
28809
28906
|
// Base case:
|
|
28810
28907
|
PreTrainedTokenizer,
|
|
@@ -31761,7 +31858,7 @@ class RawImage {
|
|
|
31761
31858
|
throw new Error('fromCanvas() is only supported in browser environments.')
|
|
31762
31859
|
}
|
|
31763
31860
|
|
|
31764
|
-
const ctx = canvas.getContext('2d');
|
|
31861
|
+
const ctx = /** @type {CanvasRenderingContext2D | OffscreenCanvasRenderingContext2D} */ (canvas.getContext('2d'));
|
|
31765
31862
|
const data = ctx.getImageData(0, 0, canvas.width, canvas.height).data;
|
|
31766
31863
|
return new RawImage(data, canvas.width, canvas.height, 4);
|
|
31767
31864
|
}
|
|
@@ -35573,6 +35670,10 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
35573
35670
|
/* harmony export */ ElectraTokenizer: () => (/* reexport safe */ _tokenizers_js__WEBPACK_IMPORTED_MODULE_3__.ElectraTokenizer),
|
|
35574
35671
|
/* harmony export */ EncodecFeatureExtractor: () => (/* reexport safe */ _models_feature_extractors_js__WEBPACK_IMPORTED_MODULE_11__.EncodecFeatureExtractor),
|
|
35575
35672
|
/* harmony export */ EosTokenCriteria: () => (/* reexport safe */ _generation_stopping_criteria_js__WEBPACK_IMPORTED_MODULE_20__.EosTokenCriteria),
|
|
35673
|
+
/* harmony export */ Ernie4_5_ForCausalLM: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Ernie4_5_ForCausalLM),
|
|
35674
|
+
/* harmony export */ Ernie4_5_Model: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Ernie4_5_Model),
|
|
35675
|
+
/* harmony export */ Ernie4_5_PretrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.Ernie4_5_PretrainedModel),
|
|
35676
|
+
/* harmony export */ Ernie4_5_Tokenizer: () => (/* reexport safe */ _tokenizers_js__WEBPACK_IMPORTED_MODULE_3__.Ernie4_5_Tokenizer),
|
|
35576
35677
|
/* harmony export */ EsmForMaskedLM: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.EsmForMaskedLM),
|
|
35577
35678
|
/* harmony export */ EsmForSequenceClassification: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.EsmForSequenceClassification),
|
|
35578
35679
|
/* harmony export */ EsmForTokenClassification: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.EsmForTokenClassification),
|
|
@@ -35814,6 +35915,12 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
35814
35915
|
/* harmony export */ MusicgenForConditionalGeneration: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.MusicgenForConditionalGeneration),
|
|
35815
35916
|
/* harmony export */ MusicgenModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.MusicgenModel),
|
|
35816
35917
|
/* harmony export */ MusicgenPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.MusicgenPreTrainedModel),
|
|
35918
|
+
/* harmony export */ NeoBertForMaskedLM: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.NeoBertForMaskedLM),
|
|
35919
|
+
/* harmony export */ NeoBertForQuestionAnswering: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.NeoBertForQuestionAnswering),
|
|
35920
|
+
/* harmony export */ NeoBertForSequenceClassification: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.NeoBertForSequenceClassification),
|
|
35921
|
+
/* harmony export */ NeoBertForTokenClassification: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.NeoBertForTokenClassification),
|
|
35922
|
+
/* harmony export */ NeoBertModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.NeoBertModel),
|
|
35923
|
+
/* harmony export */ NeoBertPreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.NeoBertPreTrainedModel),
|
|
35817
35924
|
/* harmony export */ NllbTokenizer: () => (/* reexport safe */ _tokenizers_js__WEBPACK_IMPORTED_MODULE_3__.NllbTokenizer),
|
|
35818
35925
|
/* harmony export */ NoBadWordsLogitsProcessor: () => (/* reexport safe */ _generation_logits_process_js__WEBPACK_IMPORTED_MODULE_21__.NoBadWordsLogitsProcessor),
|
|
35819
35926
|
/* harmony export */ NoRepeatNGramLogitsProcessor: () => (/* reexport safe */ _generation_logits_process_js__WEBPACK_IMPORTED_MODULE_21__.NoRepeatNGramLogitsProcessor),
|
|
@@ -35950,6 +36057,9 @@ __webpack_require__.r(__webpack_exports__);
|
|
|
35950
36057
|
/* harmony export */ SiglipTextModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.SiglipTextModel),
|
|
35951
36058
|
/* harmony export */ SiglipTokenizer: () => (/* reexport safe */ _tokenizers_js__WEBPACK_IMPORTED_MODULE_3__.SiglipTokenizer),
|
|
35952
36059
|
/* harmony export */ SiglipVisionModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.SiglipVisionModel),
|
|
36060
|
+
/* harmony export */ SmolLM3ForCausalLM: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.SmolLM3ForCausalLM),
|
|
36061
|
+
/* harmony export */ SmolLM3Model: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.SmolLM3Model),
|
|
36062
|
+
/* harmony export */ SmolLM3PreTrainedModel: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.SmolLM3PreTrainedModel),
|
|
35953
36063
|
/* harmony export */ SmolVLMForConditionalGeneration: () => (/* reexport safe */ _models_js__WEBPACK_IMPORTED_MODULE_2__.SmolVLMForConditionalGeneration),
|
|
35954
36064
|
/* harmony export */ SmolVLMImageProcessor: () => (/* reexport safe */ _models_image_processors_js__WEBPACK_IMPORTED_MODULE_14__.SmolVLMImageProcessor),
|
|
35955
36065
|
/* harmony export */ SmolVLMProcessor: () => (/* reexport safe */ _models_processors_js__WEBPACK_IMPORTED_MODULE_17__.SmolVLMProcessor),
|