@huggingface/transformers 4.0.0-next.2 → 4.0.0-next.4

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (54) hide show
  1. package/README.md +4 -3
  2. package/dist/ort-wasm-simd-threaded.jsep.mjs +23 -23
  3. package/dist/transformers.js +49 -18
  4. package/dist/transformers.min.js +20 -20
  5. package/dist/transformers.node.cjs +43 -9
  6. package/dist/transformers.node.min.cjs +18 -18
  7. package/dist/transformers.node.min.mjs +18 -18
  8. package/dist/transformers.node.mjs +40 -9
  9. package/dist/transformers.web.js +33 -2
  10. package/dist/transformers.web.min.js +17 -17
  11. package/package.json +2 -2
  12. package/src/configs.js +1 -0
  13. package/src/env.js +1 -1
  14. package/src/generation/parameters.js +1 -1
  15. package/src/generation/streamers.js +21 -0
  16. package/src/models/cohere2/modeling_cohere2.js +5 -0
  17. package/src/models/modeling_utils.js +1 -1
  18. package/src/models/models.js +1 -0
  19. package/src/models/registry.js +2 -0
  20. package/src/pipelines/automatic-speech-recognition.js +1 -1
  21. package/src/pipelines/document-question-answering.js +1 -1
  22. package/src/pipelines/image-to-text.js +2 -2
  23. package/src/pipelines/summarization.js +1 -1
  24. package/src/pipelines/text-generation.js +1 -1
  25. package/src/pipelines/text2text-generation.js +1 -1
  26. package/src/pipelines/translation.js +1 -1
  27. package/src/transformers.js +2 -0
  28. package/types/configs.d.ts.map +1 -1
  29. package/types/generation/parameters.d.ts +1 -1
  30. package/types/generation/parameters.d.ts.map +1 -1
  31. package/types/generation/streamers.d.ts +1 -0
  32. package/types/generation/streamers.d.ts.map +1 -1
  33. package/types/models/cohere2/modeling_cohere2.d.ts +8 -0
  34. package/types/models/cohere2/modeling_cohere2.d.ts.map +1 -0
  35. package/types/models/modeling_utils.d.ts +2 -2
  36. package/types/models/modeling_utils.d.ts.map +1 -1
  37. package/types/models/models.d.ts +1 -0
  38. package/types/models/registry.d.ts.map +1 -1
  39. package/types/pipelines/automatic-speech-recognition.d.ts +2 -2
  40. package/types/pipelines/automatic-speech-recognition.d.ts.map +1 -1
  41. package/types/pipelines/document-question-answering.d.ts +2 -2
  42. package/types/pipelines/document-question-answering.d.ts.map +1 -1
  43. package/types/pipelines/image-to-text.d.ts +4 -4
  44. package/types/pipelines/image-to-text.d.ts.map +1 -1
  45. package/types/pipelines/summarization.d.ts +2 -2
  46. package/types/pipelines/summarization.d.ts.map +1 -1
  47. package/types/pipelines/text-generation.d.ts +2 -2
  48. package/types/pipelines/text-generation.d.ts.map +1 -1
  49. package/types/pipelines/text2text-generation.d.ts +3 -3
  50. package/types/pipelines/text2text-generation.d.ts.map +1 -1
  51. package/types/pipelines/translation.d.ts +2 -2
  52. package/types/pipelines/translation.d.ts.map +1 -1
  53. package/types/transformers.d.ts +2 -0
  54. package/types/transformers.d.ts.map +1 -1
@@ -151,6 +151,9 @@ __export(transformers_exports, {
151
151
  CodeGenPreTrainedModel: () => CodeGenPreTrainedModel,
152
152
  CodeGenTokenizer: () => CodeGenTokenizer,
153
153
  CodeLlamaTokenizer: () => CodeLlamaTokenizer,
154
+ Cohere2ForCausalLM: () => Cohere2ForCausalLM,
155
+ Cohere2Model: () => Cohere2Model,
156
+ Cohere2PreTrainedModel: () => Cohere2PreTrainedModel,
154
157
  CohereForCausalLM: () => CohereForCausalLM,
155
158
  CohereModel: () => CohereModel,
156
159
  CoherePreTrainedModel: () => CoherePreTrainedModel,
@@ -882,7 +885,7 @@ var import_node_fs = __toESM(require("fs"), 1);
882
885
  var import_node_path = __toESM(require("path"), 1);
883
886
  var import_node_url = __toESM(require("url"), 1);
884
887
  var import_meta = {};
885
- var VERSION = "4.0.0-next.2";
888
+ var VERSION = "4.0.0-next.4";
886
889
  var IS_PROCESS_AVAILABLE = typeof process !== "undefined";
887
890
  var IS_NODE_ENV = IS_PROCESS_AVAILABLE && process?.release?.name === "node";
888
891
  var IS_FS_AVAILABLE = !isEmpty(import_node_fs.default);
@@ -7627,7 +7630,7 @@ var uint16_to_float32 = /* @__PURE__ */ (function() {
7627
7630
  // src/backends/onnx.js
7628
7631
  var ONNX_NODE = __toESM(require("onnxruntime-node"), 1);
7629
7632
 
7630
- // ../../node_modules/.pnpm/onnxruntime-web@1.25.0-dev.20260209-a3749f1353/node_modules/onnxruntime-web/dist/ort.webgpu.bundle.min.mjs
7633
+ // ../../node_modules/.pnpm/onnxruntime-web@1.25.0-dev.20260212-1a71a5f46e/node_modules/onnxruntime-web/dist/ort.webgpu.bundle.min.mjs
7631
7634
  var ort_webgpu_bundle_min_exports = {};
7632
7635
  __export(ort_webgpu_bundle_min_exports, {
7633
7636
  InferenceSession: () => nc,
@@ -9597,7 +9600,7 @@ async function is(a = {}) {
9597
9600
  we(`invalid type for getValue: ${t}`);
9598
9601
  }
9599
9602
  }, r.UTF8ToString = ct, r.stringToUTF8 = Pe, r.lengthBytesUTF8 = _e;
9600
- var bo, wo, Rr, $t, Te, mt, go, To, vo, Eo, So, Ao, Io, xo, Lo, Oo, Bo, Nr, Wr, kr, Fr, Et, Gr, Mo, $r, Uo, Co, Do, zr, Po, _o, Vr, k, St, Ro, C, zt, D, No, Hr, Wo, ko, Fo, jr, Go, $o, zo, Vo, Ho, jo, Yo, qo, Jo, Xo, Zo, Ko, Qo, ea, ta, ra, na, oa, aa, sa, ia, ua, fa, ca, da, la, pa, ma, ha, ya, ba, wa, ga, Ta, va, Ea, Sa, Aa, Ia, xa, Ne, df = [Ye, gr, In, Mn, Un, Cn, Dn, Pn, _n, Rn, Nn, Wn, kn, Fn, Gn, $n, eo, to, ro, co, lo, po, mo, ho, yo], Yr = { 1116828: (e, t, n, o, u) => {
9603
+ var bo, wo, Rr, $t, Te, mt, go, To, vo, Eo, So, Ao, Io, xo, Lo, Oo, Bo, Nr, Wr, kr, Fr, Et, Gr, Mo, $r, Uo, Co, Do, zr, Po, _o, Vr, k, St, Ro, C, zt, D, No, Hr, Wo, ko, Fo, jr, Go, $o, zo, Vo, Ho, jo, Yo, qo, Jo, Xo, Zo, Ko, Qo, ea, ta, ra, na, oa, aa, sa, ia, ua, fa, ca, da, la, pa, ma, ha, ya, ba, wa, ga, Ta, va, Ea, Sa, Aa, Ia, xa, Ne, df = [Ye, gr, In, Mn, Un, Cn, Dn, Pn, _n, Rn, Nn, Wn, kn, Fn, Gn, $n, eo, to, ro, co, lo, po, mo, ho, yo], Yr = { 1117404: (e, t, n, o, u) => {
9601
9604
  if (r === void 0 || !r.Zc) return 1;
9602
9605
  if ((e = ct(Number(e >>> 0))).startsWith("./") && (e = e.substring(2)), !(e = r.Zc.get(e))) return 2;
9603
9606
  if (t = Number(t >>> 0), n = Number(n >>> 0), o = Number(o >>> 0), t + n > e.byteLength) return 3;
@@ -9617,11 +9620,11 @@ async function is(a = {}) {
9617
9620
  } catch {
9618
9621
  return 4;
9619
9622
  }
9620
- }, 1117652: (e, t, n) => {
9623
+ }, 1118228: (e, t, n) => {
9621
9624
  r.ke(e, (l(), J).subarray(t >>> 0, t + n >>> 0));
9622
- }, 1117716: () => r.Ie(), 1117758: (e) => {
9625
+ }, 1118292: () => r.Ie(), 1118334: (e) => {
9623
9626
  r.je(e);
9624
- }, 1117795: () => typeof wasmOffsetConverter < "u" };
9627
+ }, 1118371: () => typeof wasmOffsetConverter < "u" };
9625
9628
  function lf() {
9626
9629
  return typeof wasmOffsetConverter < "u";
9627
9630
  }
@@ -11549,7 +11552,7 @@ var qs = F(() => {
11549
11552
  ze();
11550
11553
  ze();
11551
11554
  ze();
11552
- var rs = "1.25.0-dev.20260209-a3749f1353";
11555
+ var rs = "1.25.0-dev.20260212-1a71a5f46e";
11553
11556
  var Ol = tn;
11554
11557
  {
11555
11558
  let a = (qs(), Ht(Ys)).wasmBackend;
@@ -20849,6 +20852,7 @@ function getNormalizedConfig(config) {
20849
20852
  case "granite":
20850
20853
  case "granitemoehybrid":
20851
20854
  case "cohere":
20855
+ case "cohere2":
20852
20856
  case "mistral":
20853
20857
  case "starcoder2":
20854
20858
  case "qwen2":
@@ -23184,7 +23188,7 @@ var PreTrainedModel = class extends Callable2 {
23184
23188
  /**
23185
23189
  *
23186
23190
  * @param {GenerationConfig} generation_config
23187
- * @param {StoppingCriteriaList} [stopping_criteria=null]
23191
+ * @param {import('../generation/stopping_criteria.js').StoppingCriteria|import('../generation/stopping_criteria.js').StoppingCriteria[]|StoppingCriteriaList} [stopping_criteria=null]
23188
23192
  */
23189
23193
  _get_stopping_criteria(generation_config, stopping_criteria = null) {
23190
23194
  const criteria = new StoppingCriteriaList();
@@ -23928,6 +23932,9 @@ __export(models_exports, {
23928
23932
  CodeGenForCausalLM: () => CodeGenForCausalLM,
23929
23933
  CodeGenModel: () => CodeGenModel,
23930
23934
  CodeGenPreTrainedModel: () => CodeGenPreTrainedModel,
23935
+ Cohere2ForCausalLM: () => Cohere2ForCausalLM,
23936
+ Cohere2Model: () => Cohere2Model,
23937
+ Cohere2PreTrainedModel: () => Cohere2PreTrainedModel,
23931
23938
  CohereForCausalLM: () => CohereForCausalLM,
23932
23939
  CohereModel: () => CohereModel,
23933
23940
  CoherePreTrainedModel: () => CoherePreTrainedModel,
@@ -24925,6 +24932,14 @@ var CohereModel = class extends CoherePreTrainedModel {
24925
24932
  var CohereForCausalLM = class extends CoherePreTrainedModel {
24926
24933
  };
24927
24934
 
24935
+ // src/models/cohere2/modeling_cohere2.js
24936
+ var Cohere2PreTrainedModel = class extends PreTrainedModel {
24937
+ };
24938
+ var Cohere2Model = class extends Cohere2PreTrainedModel {
24939
+ };
24940
+ var Cohere2ForCausalLM = class extends Cohere2PreTrainedModel {
24941
+ };
24942
+
24928
24943
  // src/models/convbert/modeling_convbert.js
24929
24944
  var ConvBertPreTrainedModel = class extends PreTrainedModel {
24930
24945
  };
@@ -28729,6 +28744,7 @@ var MODEL_MAPPING_NAMES_DECODER_ONLY = /* @__PURE__ */ new Map([
28729
28744
  ["granite", "GraniteModel"],
28730
28745
  ["granitemoehybrid", "GraniteMoeHybridModel"],
28731
28746
  ["cohere", "CohereModel"],
28747
+ ["cohere2", "Cohere2Model"],
28732
28748
  ["gemma", "GemmaModel"],
28733
28749
  ["gemma2", "Gemma2Model"],
28734
28750
  ["vaultgemma", "VaultGemmaModel"],
@@ -28842,6 +28858,7 @@ var MODEL_FOR_CAUSAL_LM_MAPPING_NAMES = /* @__PURE__ */ new Map([
28842
28858
  ["granite", "GraniteForCausalLM"],
28843
28859
  ["granitemoehybrid", "GraniteMoeHybridForCausalLM"],
28844
28860
  ["cohere", "CohereForCausalLM"],
28861
+ ["cohere2", "Cohere2ForCausalLM"],
28845
28862
  ["gemma", "GemmaForCausalLM"],
28846
28863
  ["gemma2", "Gemma2ForCausalLM"],
28847
28864
  ["vaultgemma", "VaultGemmaForCausalLM"],
@@ -30993,6 +31010,7 @@ var TextStreamer = class extends BaseStreamer {
30993
31010
  this.token_cache = [];
30994
31011
  this.print_len = 0;
30995
31012
  this.next_tokens_are_prompt = true;
31013
+ this.special_ids = new Set(this.tokenizer.all_special_ids.map(BigInt));
30996
31014
  }
30997
31015
  /**
30998
31016
  * Receives tokens, decodes them, and prints them to stdout as soon as they form entire words.
@@ -31009,6 +31027,19 @@ var TextStreamer = class extends BaseStreamer {
31009
31027
  }
31010
31028
  const tokens = value[0];
31011
31029
  this.token_callback_function?.(tokens);
31030
+ if (tokens.length === 1 && this.special_ids.has(tokens[0])) {
31031
+ if (this.decode_kwargs.skip_special_tokens) return;
31032
+ if (this.token_cache.length > 0) {
31033
+ const text2 = this.tokenizer.decode(this.token_cache, this.decode_kwargs);
31034
+ const printable_text2 = text2.slice(this.print_len);
31035
+ this.on_finalized_text(printable_text2, false);
31036
+ this.token_cache = [];
31037
+ this.print_len = 0;
31038
+ }
31039
+ const special_text = this.tokenizer.decode(tokens, this.decode_kwargs);
31040
+ this.on_finalized_text(special_text, false);
31041
+ return;
31042
+ }
31012
31043
  this.token_cache = mergeArrays(this.token_cache, tokens);
31013
31044
  const text = this.tokenizer.decode(this.token_cache, this.decode_kwargs);
31014
31045
  let printable_text;
@@ -31340,6 +31371,9 @@ async function load_video(src, { num_frames = null, fps = null } = {}) {
31340
31371
  CodeGenPreTrainedModel,
31341
31372
  CodeGenTokenizer,
31342
31373
  CodeLlamaTokenizer,
31374
+ Cohere2ForCausalLM,
31375
+ Cohere2Model,
31376
+ Cohere2PreTrainedModel,
31343
31377
  CohereForCausalLM,
31344
31378
  CohereModel,
31345
31379
  CoherePreTrainedModel,
@@ -32068,7 +32102,7 @@ async function load_video(src, { num_frames = null, fps = null } = {}) {
32068
32102
 
32069
32103
  onnxruntime-web/dist/ort.webgpu.bundle.min.mjs:
32070
32104
  (*!
32071
- * ONNX Runtime Web v1.25.0-dev.20260209-a3749f1353
32105
+ * ONNX Runtime Web v1.25.0-dev.20260212-1a71a5f46e
32072
32106
  * Copyright (c) Microsoft Corporation. All rights reserved.
32073
32107
  * Licensed under the MIT License.
32074
32108
  *)