@huggingface/transformers 4.0.0-next.2 → 4.0.0-next.4
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +4 -3
- package/dist/ort-wasm-simd-threaded.jsep.mjs +23 -23
- package/dist/transformers.js +49 -18
- package/dist/transformers.min.js +20 -20
- package/dist/transformers.node.cjs +43 -9
- package/dist/transformers.node.min.cjs +18 -18
- package/dist/transformers.node.min.mjs +18 -18
- package/dist/transformers.node.mjs +40 -9
- package/dist/transformers.web.js +33 -2
- package/dist/transformers.web.min.js +17 -17
- package/package.json +2 -2
- package/src/configs.js +1 -0
- package/src/env.js +1 -1
- package/src/generation/parameters.js +1 -1
- package/src/generation/streamers.js +21 -0
- package/src/models/cohere2/modeling_cohere2.js +5 -0
- package/src/models/modeling_utils.js +1 -1
- package/src/models/models.js +1 -0
- package/src/models/registry.js +2 -0
- package/src/pipelines/automatic-speech-recognition.js +1 -1
- package/src/pipelines/document-question-answering.js +1 -1
- package/src/pipelines/image-to-text.js +2 -2
- package/src/pipelines/summarization.js +1 -1
- package/src/pipelines/text-generation.js +1 -1
- package/src/pipelines/text2text-generation.js +1 -1
- package/src/pipelines/translation.js +1 -1
- package/src/transformers.js +2 -0
- package/types/configs.d.ts.map +1 -1
- package/types/generation/parameters.d.ts +1 -1
- package/types/generation/parameters.d.ts.map +1 -1
- package/types/generation/streamers.d.ts +1 -0
- package/types/generation/streamers.d.ts.map +1 -1
- package/types/models/cohere2/modeling_cohere2.d.ts +8 -0
- package/types/models/cohere2/modeling_cohere2.d.ts.map +1 -0
- package/types/models/modeling_utils.d.ts +2 -2
- package/types/models/modeling_utils.d.ts.map +1 -1
- package/types/models/models.d.ts +1 -0
- package/types/models/registry.d.ts.map +1 -1
- package/types/pipelines/automatic-speech-recognition.d.ts +2 -2
- package/types/pipelines/automatic-speech-recognition.d.ts.map +1 -1
- package/types/pipelines/document-question-answering.d.ts +2 -2
- package/types/pipelines/document-question-answering.d.ts.map +1 -1
- package/types/pipelines/image-to-text.d.ts +4 -4
- package/types/pipelines/image-to-text.d.ts.map +1 -1
- package/types/pipelines/summarization.d.ts +2 -2
- package/types/pipelines/summarization.d.ts.map +1 -1
- package/types/pipelines/text-generation.d.ts +2 -2
- package/types/pipelines/text-generation.d.ts.map +1 -1
- package/types/pipelines/text2text-generation.d.ts +3 -3
- package/types/pipelines/text2text-generation.d.ts.map +1 -1
- package/types/pipelines/translation.d.ts +2 -2
- package/types/pipelines/translation.d.ts.map +1 -1
- package/types/transformers.d.ts +2 -0
- package/types/transformers.d.ts.map +1 -1
|
@@ -14,7 +14,7 @@ var __export = (target, all) => {
|
|
|
14
14
|
import fs from "fs";
|
|
15
15
|
import path from "path";
|
|
16
16
|
import url from "url";
|
|
17
|
-
var VERSION = "4.0.0-next.
|
|
17
|
+
var VERSION = "4.0.0-next.4";
|
|
18
18
|
var IS_PROCESS_AVAILABLE = typeof process !== "undefined";
|
|
19
19
|
var IS_NODE_ENV = IS_PROCESS_AVAILABLE && process?.release?.name === "node";
|
|
20
20
|
var IS_FS_AVAILABLE = !isEmpty(fs);
|
|
@@ -6759,7 +6759,7 @@ var uint16_to_float32 = /* @__PURE__ */ (function() {
|
|
|
6759
6759
|
// src/backends/onnx.js
|
|
6760
6760
|
import * as ONNX_NODE from "onnxruntime-node";
|
|
6761
6761
|
|
|
6762
|
-
// ../../node_modules/.pnpm/onnxruntime-web@1.25.0-dev.
|
|
6762
|
+
// ../../node_modules/.pnpm/onnxruntime-web@1.25.0-dev.20260212-1a71a5f46e/node_modules/onnxruntime-web/dist/ort.webgpu.bundle.min.mjs
|
|
6763
6763
|
var ort_webgpu_bundle_min_exports = {};
|
|
6764
6764
|
__export(ort_webgpu_bundle_min_exports, {
|
|
6765
6765
|
InferenceSession: () => nc,
|
|
@@ -8728,7 +8728,7 @@ async function is(a = {}) {
|
|
|
8728
8728
|
we(`invalid type for getValue: ${t}`);
|
|
8729
8729
|
}
|
|
8730
8730
|
}, r.UTF8ToString = ct, r.stringToUTF8 = Pe, r.lengthBytesUTF8 = _e;
|
|
8731
|
-
var bo, wo, Rr, $t, Te, mt, go, To, vo, Eo, So, Ao, Io, xo, Lo, Oo, Bo, Nr, Wr, kr, Fr, Et, Gr, Mo, $r, Uo, Co, Do, zr, Po, _o, Vr, k, St, Ro, C, zt, D, No, Hr, Wo, ko, Fo, jr, Go, $o, zo, Vo, Ho, jo, Yo, qo, Jo, Xo, Zo, Ko, Qo, ea, ta, ra, na, oa, aa, sa, ia, ua, fa, ca, da, la, pa, ma, ha, ya, ba, wa, ga, Ta, va, Ea, Sa, Aa, Ia, xa, Ne, df = [Ye, gr, In, Mn, Un, Cn, Dn, Pn, _n, Rn, Nn, Wn, kn, Fn, Gn, $n, eo, to, ro, co, lo, po, mo, ho, yo], Yr = {
|
|
8731
|
+
var bo, wo, Rr, $t, Te, mt, go, To, vo, Eo, So, Ao, Io, xo, Lo, Oo, Bo, Nr, Wr, kr, Fr, Et, Gr, Mo, $r, Uo, Co, Do, zr, Po, _o, Vr, k, St, Ro, C, zt, D, No, Hr, Wo, ko, Fo, jr, Go, $o, zo, Vo, Ho, jo, Yo, qo, Jo, Xo, Zo, Ko, Qo, ea, ta, ra, na, oa, aa, sa, ia, ua, fa, ca, da, la, pa, ma, ha, ya, ba, wa, ga, Ta, va, Ea, Sa, Aa, Ia, xa, Ne, df = [Ye, gr, In, Mn, Un, Cn, Dn, Pn, _n, Rn, Nn, Wn, kn, Fn, Gn, $n, eo, to, ro, co, lo, po, mo, ho, yo], Yr = { 1117404: (e, t, n, o, u) => {
|
|
8732
8732
|
if (r === void 0 || !r.Zc) return 1;
|
|
8733
8733
|
if ((e = ct(Number(e >>> 0))).startsWith("./") && (e = e.substring(2)), !(e = r.Zc.get(e))) return 2;
|
|
8734
8734
|
if (t = Number(t >>> 0), n = Number(n >>> 0), o = Number(o >>> 0), t + n > e.byteLength) return 3;
|
|
@@ -8748,11 +8748,11 @@ async function is(a = {}) {
|
|
|
8748
8748
|
} catch {
|
|
8749
8749
|
return 4;
|
|
8750
8750
|
}
|
|
8751
|
-
},
|
|
8751
|
+
}, 1118228: (e, t, n) => {
|
|
8752
8752
|
r.ke(e, (l(), J).subarray(t >>> 0, t + n >>> 0));
|
|
8753
|
-
},
|
|
8753
|
+
}, 1118292: () => r.Ie(), 1118334: (e) => {
|
|
8754
8754
|
r.je(e);
|
|
8755
|
-
},
|
|
8755
|
+
}, 1118371: () => typeof wasmOffsetConverter < "u" };
|
|
8756
8756
|
function lf() {
|
|
8757
8757
|
return typeof wasmOffsetConverter < "u";
|
|
8758
8758
|
}
|
|
@@ -10680,7 +10680,7 @@ var qs = F(() => {
|
|
|
10680
10680
|
ze();
|
|
10681
10681
|
ze();
|
|
10682
10682
|
ze();
|
|
10683
|
-
var rs = "1.25.0-dev.
|
|
10683
|
+
var rs = "1.25.0-dev.20260212-1a71a5f46e";
|
|
10684
10684
|
var Ol = tn;
|
|
10685
10685
|
{
|
|
10686
10686
|
let a = (qs(), Ht(Ys)).wasmBackend;
|
|
@@ -19979,6 +19979,7 @@ function getNormalizedConfig(config) {
|
|
|
19979
19979
|
case "granite":
|
|
19980
19980
|
case "granitemoehybrid":
|
|
19981
19981
|
case "cohere":
|
|
19982
|
+
case "cohere2":
|
|
19982
19983
|
case "mistral":
|
|
19983
19984
|
case "starcoder2":
|
|
19984
19985
|
case "qwen2":
|
|
@@ -22314,7 +22315,7 @@ var PreTrainedModel = class extends Callable2 {
|
|
|
22314
22315
|
/**
|
|
22315
22316
|
*
|
|
22316
22317
|
* @param {GenerationConfig} generation_config
|
|
22317
|
-
* @param {StoppingCriteriaList} [stopping_criteria=null]
|
|
22318
|
+
* @param {import('../generation/stopping_criteria.js').StoppingCriteria|import('../generation/stopping_criteria.js').StoppingCriteria[]|StoppingCriteriaList} [stopping_criteria=null]
|
|
22318
22319
|
*/
|
|
22319
22320
|
_get_stopping_criteria(generation_config, stopping_criteria = null) {
|
|
22320
22321
|
const criteria = new StoppingCriteriaList();
|
|
@@ -23058,6 +23059,9 @@ __export(models_exports, {
|
|
|
23058
23059
|
CodeGenForCausalLM: () => CodeGenForCausalLM,
|
|
23059
23060
|
CodeGenModel: () => CodeGenModel,
|
|
23060
23061
|
CodeGenPreTrainedModel: () => CodeGenPreTrainedModel,
|
|
23062
|
+
Cohere2ForCausalLM: () => Cohere2ForCausalLM,
|
|
23063
|
+
Cohere2Model: () => Cohere2Model,
|
|
23064
|
+
Cohere2PreTrainedModel: () => Cohere2PreTrainedModel,
|
|
23061
23065
|
CohereForCausalLM: () => CohereForCausalLM,
|
|
23062
23066
|
CohereModel: () => CohereModel,
|
|
23063
23067
|
CoherePreTrainedModel: () => CoherePreTrainedModel,
|
|
@@ -24055,6 +24059,14 @@ var CohereModel = class extends CoherePreTrainedModel {
|
|
|
24055
24059
|
var CohereForCausalLM = class extends CoherePreTrainedModel {
|
|
24056
24060
|
};
|
|
24057
24061
|
|
|
24062
|
+
// src/models/cohere2/modeling_cohere2.js
|
|
24063
|
+
var Cohere2PreTrainedModel = class extends PreTrainedModel {
|
|
24064
|
+
};
|
|
24065
|
+
var Cohere2Model = class extends Cohere2PreTrainedModel {
|
|
24066
|
+
};
|
|
24067
|
+
var Cohere2ForCausalLM = class extends Cohere2PreTrainedModel {
|
|
24068
|
+
};
|
|
24069
|
+
|
|
24058
24070
|
// src/models/convbert/modeling_convbert.js
|
|
24059
24071
|
var ConvBertPreTrainedModel = class extends PreTrainedModel {
|
|
24060
24072
|
};
|
|
@@ -27859,6 +27871,7 @@ var MODEL_MAPPING_NAMES_DECODER_ONLY = /* @__PURE__ */ new Map([
|
|
|
27859
27871
|
["granite", "GraniteModel"],
|
|
27860
27872
|
["granitemoehybrid", "GraniteMoeHybridModel"],
|
|
27861
27873
|
["cohere", "CohereModel"],
|
|
27874
|
+
["cohere2", "Cohere2Model"],
|
|
27862
27875
|
["gemma", "GemmaModel"],
|
|
27863
27876
|
["gemma2", "Gemma2Model"],
|
|
27864
27877
|
["vaultgemma", "VaultGemmaModel"],
|
|
@@ -27972,6 +27985,7 @@ var MODEL_FOR_CAUSAL_LM_MAPPING_NAMES = /* @__PURE__ */ new Map([
|
|
|
27972
27985
|
["granite", "GraniteForCausalLM"],
|
|
27973
27986
|
["granitemoehybrid", "GraniteMoeHybridForCausalLM"],
|
|
27974
27987
|
["cohere", "CohereForCausalLM"],
|
|
27988
|
+
["cohere2", "Cohere2ForCausalLM"],
|
|
27975
27989
|
["gemma", "GemmaForCausalLM"],
|
|
27976
27990
|
["gemma2", "Gemma2ForCausalLM"],
|
|
27977
27991
|
["vaultgemma", "VaultGemmaForCausalLM"],
|
|
@@ -30123,6 +30137,7 @@ var TextStreamer = class extends BaseStreamer {
|
|
|
30123
30137
|
this.token_cache = [];
|
|
30124
30138
|
this.print_len = 0;
|
|
30125
30139
|
this.next_tokens_are_prompt = true;
|
|
30140
|
+
this.special_ids = new Set(this.tokenizer.all_special_ids.map(BigInt));
|
|
30126
30141
|
}
|
|
30127
30142
|
/**
|
|
30128
30143
|
* Receives tokens, decodes them, and prints them to stdout as soon as they form entire words.
|
|
@@ -30139,6 +30154,19 @@ var TextStreamer = class extends BaseStreamer {
|
|
|
30139
30154
|
}
|
|
30140
30155
|
const tokens = value[0];
|
|
30141
30156
|
this.token_callback_function?.(tokens);
|
|
30157
|
+
if (tokens.length === 1 && this.special_ids.has(tokens[0])) {
|
|
30158
|
+
if (this.decode_kwargs.skip_special_tokens) return;
|
|
30159
|
+
if (this.token_cache.length > 0) {
|
|
30160
|
+
const text2 = this.tokenizer.decode(this.token_cache, this.decode_kwargs);
|
|
30161
|
+
const printable_text2 = text2.slice(this.print_len);
|
|
30162
|
+
this.on_finalized_text(printable_text2, false);
|
|
30163
|
+
this.token_cache = [];
|
|
30164
|
+
this.print_len = 0;
|
|
30165
|
+
}
|
|
30166
|
+
const special_text = this.tokenizer.decode(tokens, this.decode_kwargs);
|
|
30167
|
+
this.on_finalized_text(special_text, false);
|
|
30168
|
+
return;
|
|
30169
|
+
}
|
|
30142
30170
|
this.token_cache = mergeArrays(this.token_cache, tokens);
|
|
30143
30171
|
const text = this.tokenizer.decode(this.token_cache, this.decode_kwargs);
|
|
30144
30172
|
let printable_text;
|
|
@@ -30469,6 +30497,9 @@ export {
|
|
|
30469
30497
|
CodeGenPreTrainedModel,
|
|
30470
30498
|
CodeGenTokenizer,
|
|
30471
30499
|
CodeLlamaTokenizer,
|
|
30500
|
+
Cohere2ForCausalLM,
|
|
30501
|
+
Cohere2Model,
|
|
30502
|
+
Cohere2PreTrainedModel,
|
|
30472
30503
|
CohereForCausalLM,
|
|
30473
30504
|
CohereModel,
|
|
30474
30505
|
CoherePreTrainedModel,
|
|
@@ -31197,7 +31228,7 @@ export {
|
|
|
31197
31228
|
|
|
31198
31229
|
onnxruntime-web/dist/ort.webgpu.bundle.min.mjs:
|
|
31199
31230
|
(*!
|
|
31200
|
-
* ONNX Runtime Web v1.25.0-dev.
|
|
31231
|
+
* ONNX Runtime Web v1.25.0-dev.20260212-1a71a5f46e
|
|
31201
31232
|
* Copyright (c) Microsoft Corporation. All rights reserved.
|
|
31202
31233
|
* Licensed under the MIT License.
|
|
31203
31234
|
*)
|
package/dist/transformers.web.js
CHANGED
|
@@ -17,7 +17,7 @@ var emptyObj3 = {};
|
|
|
17
17
|
var node_url_default = emptyObj3;
|
|
18
18
|
|
|
19
19
|
// src/env.js
|
|
20
|
-
var VERSION = "4.0.0-next.
|
|
20
|
+
var VERSION = "4.0.0-next.4";
|
|
21
21
|
var IS_PROCESS_AVAILABLE = typeof process !== "undefined";
|
|
22
22
|
var IS_NODE_ENV = IS_PROCESS_AVAILABLE && process?.release?.name === "node";
|
|
23
23
|
var IS_FS_AVAILABLE = !isEmpty(node_fs_default);
|
|
@@ -16079,6 +16079,7 @@ function getNormalizedConfig(config) {
|
|
|
16079
16079
|
case "granite":
|
|
16080
16080
|
case "granitemoehybrid":
|
|
16081
16081
|
case "cohere":
|
|
16082
|
+
case "cohere2":
|
|
16082
16083
|
case "mistral":
|
|
16083
16084
|
case "starcoder2":
|
|
16084
16085
|
case "qwen2":
|
|
@@ -18414,7 +18415,7 @@ var PreTrainedModel = class extends Callable2 {
|
|
|
18414
18415
|
/**
|
|
18415
18416
|
*
|
|
18416
18417
|
* @param {GenerationConfig} generation_config
|
|
18417
|
-
* @param {StoppingCriteriaList} [stopping_criteria=null]
|
|
18418
|
+
* @param {import('../generation/stopping_criteria.js').StoppingCriteria|import('../generation/stopping_criteria.js').StoppingCriteria[]|StoppingCriteriaList} [stopping_criteria=null]
|
|
18418
18419
|
*/
|
|
18419
18420
|
_get_stopping_criteria(generation_config, stopping_criteria = null) {
|
|
18420
18421
|
const criteria = new StoppingCriteriaList();
|
|
@@ -19158,6 +19159,9 @@ __export(models_exports, {
|
|
|
19158
19159
|
CodeGenForCausalLM: () => CodeGenForCausalLM,
|
|
19159
19160
|
CodeGenModel: () => CodeGenModel,
|
|
19160
19161
|
CodeGenPreTrainedModel: () => CodeGenPreTrainedModel,
|
|
19162
|
+
Cohere2ForCausalLM: () => Cohere2ForCausalLM,
|
|
19163
|
+
Cohere2Model: () => Cohere2Model,
|
|
19164
|
+
Cohere2PreTrainedModel: () => Cohere2PreTrainedModel,
|
|
19161
19165
|
CohereForCausalLM: () => CohereForCausalLM,
|
|
19162
19166
|
CohereModel: () => CohereModel,
|
|
19163
19167
|
CoherePreTrainedModel: () => CoherePreTrainedModel,
|
|
@@ -20155,6 +20159,14 @@ var CohereModel = class extends CoherePreTrainedModel {
|
|
|
20155
20159
|
var CohereForCausalLM = class extends CoherePreTrainedModel {
|
|
20156
20160
|
};
|
|
20157
20161
|
|
|
20162
|
+
// src/models/cohere2/modeling_cohere2.js
|
|
20163
|
+
var Cohere2PreTrainedModel = class extends PreTrainedModel {
|
|
20164
|
+
};
|
|
20165
|
+
var Cohere2Model = class extends Cohere2PreTrainedModel {
|
|
20166
|
+
};
|
|
20167
|
+
var Cohere2ForCausalLM = class extends Cohere2PreTrainedModel {
|
|
20168
|
+
};
|
|
20169
|
+
|
|
20158
20170
|
// src/models/convbert/modeling_convbert.js
|
|
20159
20171
|
var ConvBertPreTrainedModel = class extends PreTrainedModel {
|
|
20160
20172
|
};
|
|
@@ -23959,6 +23971,7 @@ var MODEL_MAPPING_NAMES_DECODER_ONLY = /* @__PURE__ */ new Map([
|
|
|
23959
23971
|
["granite", "GraniteModel"],
|
|
23960
23972
|
["granitemoehybrid", "GraniteMoeHybridModel"],
|
|
23961
23973
|
["cohere", "CohereModel"],
|
|
23974
|
+
["cohere2", "Cohere2Model"],
|
|
23962
23975
|
["gemma", "GemmaModel"],
|
|
23963
23976
|
["gemma2", "Gemma2Model"],
|
|
23964
23977
|
["vaultgemma", "VaultGemmaModel"],
|
|
@@ -24072,6 +24085,7 @@ var MODEL_FOR_CAUSAL_LM_MAPPING_NAMES = /* @__PURE__ */ new Map([
|
|
|
24072
24085
|
["granite", "GraniteForCausalLM"],
|
|
24073
24086
|
["granitemoehybrid", "GraniteMoeHybridForCausalLM"],
|
|
24074
24087
|
["cohere", "CohereForCausalLM"],
|
|
24088
|
+
["cohere2", "Cohere2ForCausalLM"],
|
|
24075
24089
|
["gemma", "GemmaForCausalLM"],
|
|
24076
24090
|
["gemma2", "Gemma2ForCausalLM"],
|
|
24077
24091
|
["vaultgemma", "VaultGemmaForCausalLM"],
|
|
@@ -26223,6 +26237,7 @@ var TextStreamer = class extends BaseStreamer {
|
|
|
26223
26237
|
this.token_cache = [];
|
|
26224
26238
|
this.print_len = 0;
|
|
26225
26239
|
this.next_tokens_are_prompt = true;
|
|
26240
|
+
this.special_ids = new Set(this.tokenizer.all_special_ids.map(BigInt));
|
|
26226
26241
|
}
|
|
26227
26242
|
/**
|
|
26228
26243
|
* Receives tokens, decodes them, and prints them to stdout as soon as they form entire words.
|
|
@@ -26239,6 +26254,19 @@ var TextStreamer = class extends BaseStreamer {
|
|
|
26239
26254
|
}
|
|
26240
26255
|
const tokens = value[0];
|
|
26241
26256
|
this.token_callback_function?.(tokens);
|
|
26257
|
+
if (tokens.length === 1 && this.special_ids.has(tokens[0])) {
|
|
26258
|
+
if (this.decode_kwargs.skip_special_tokens) return;
|
|
26259
|
+
if (this.token_cache.length > 0) {
|
|
26260
|
+
const text2 = this.tokenizer.decode(this.token_cache, this.decode_kwargs);
|
|
26261
|
+
const printable_text2 = text2.slice(this.print_len);
|
|
26262
|
+
this.on_finalized_text(printable_text2, false);
|
|
26263
|
+
this.token_cache = [];
|
|
26264
|
+
this.print_len = 0;
|
|
26265
|
+
}
|
|
26266
|
+
const special_text = this.tokenizer.decode(tokens, this.decode_kwargs);
|
|
26267
|
+
this.on_finalized_text(special_text, false);
|
|
26268
|
+
return;
|
|
26269
|
+
}
|
|
26242
26270
|
this.token_cache = mergeArrays(this.token_cache, tokens);
|
|
26243
26271
|
const text = this.tokenizer.decode(this.token_cache, this.decode_kwargs);
|
|
26244
26272
|
let printable_text;
|
|
@@ -26569,6 +26597,9 @@ export {
|
|
|
26569
26597
|
CodeGenPreTrainedModel,
|
|
26570
26598
|
CodeGenTokenizer,
|
|
26571
26599
|
CodeLlamaTokenizer,
|
|
26600
|
+
Cohere2ForCausalLM,
|
|
26601
|
+
Cohere2Model,
|
|
26602
|
+
Cohere2PreTrainedModel,
|
|
26572
26603
|
CohereForCausalLM,
|
|
26573
26604
|
CohereModel,
|
|
26574
26605
|
CoherePreTrainedModel,
|