@huggingface/transformers 4.0.0-next.2 → 4.0.0-next.4

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (54) hide show
  1. package/README.md +4 -3
  2. package/dist/ort-wasm-simd-threaded.jsep.mjs +23 -23
  3. package/dist/transformers.js +49 -18
  4. package/dist/transformers.min.js +20 -20
  5. package/dist/transformers.node.cjs +43 -9
  6. package/dist/transformers.node.min.cjs +18 -18
  7. package/dist/transformers.node.min.mjs +18 -18
  8. package/dist/transformers.node.mjs +40 -9
  9. package/dist/transformers.web.js +33 -2
  10. package/dist/transformers.web.min.js +17 -17
  11. package/package.json +2 -2
  12. package/src/configs.js +1 -0
  13. package/src/env.js +1 -1
  14. package/src/generation/parameters.js +1 -1
  15. package/src/generation/streamers.js +21 -0
  16. package/src/models/cohere2/modeling_cohere2.js +5 -0
  17. package/src/models/modeling_utils.js +1 -1
  18. package/src/models/models.js +1 -0
  19. package/src/models/registry.js +2 -0
  20. package/src/pipelines/automatic-speech-recognition.js +1 -1
  21. package/src/pipelines/document-question-answering.js +1 -1
  22. package/src/pipelines/image-to-text.js +2 -2
  23. package/src/pipelines/summarization.js +1 -1
  24. package/src/pipelines/text-generation.js +1 -1
  25. package/src/pipelines/text2text-generation.js +1 -1
  26. package/src/pipelines/translation.js +1 -1
  27. package/src/transformers.js +2 -0
  28. package/types/configs.d.ts.map +1 -1
  29. package/types/generation/parameters.d.ts +1 -1
  30. package/types/generation/parameters.d.ts.map +1 -1
  31. package/types/generation/streamers.d.ts +1 -0
  32. package/types/generation/streamers.d.ts.map +1 -1
  33. package/types/models/cohere2/modeling_cohere2.d.ts +8 -0
  34. package/types/models/cohere2/modeling_cohere2.d.ts.map +1 -0
  35. package/types/models/modeling_utils.d.ts +2 -2
  36. package/types/models/modeling_utils.d.ts.map +1 -1
  37. package/types/models/models.d.ts +1 -0
  38. package/types/models/registry.d.ts.map +1 -1
  39. package/types/pipelines/automatic-speech-recognition.d.ts +2 -2
  40. package/types/pipelines/automatic-speech-recognition.d.ts.map +1 -1
  41. package/types/pipelines/document-question-answering.d.ts +2 -2
  42. package/types/pipelines/document-question-answering.d.ts.map +1 -1
  43. package/types/pipelines/image-to-text.d.ts +4 -4
  44. package/types/pipelines/image-to-text.d.ts.map +1 -1
  45. package/types/pipelines/summarization.d.ts +2 -2
  46. package/types/pipelines/summarization.d.ts.map +1 -1
  47. package/types/pipelines/text-generation.d.ts +2 -2
  48. package/types/pipelines/text-generation.d.ts.map +1 -1
  49. package/types/pipelines/text2text-generation.d.ts +3 -3
  50. package/types/pipelines/text2text-generation.d.ts.map +1 -1
  51. package/types/pipelines/translation.d.ts +2 -2
  52. package/types/pipelines/translation.d.ts.map +1 -1
  53. package/types/transformers.d.ts +2 -0
  54. package/types/transformers.d.ts.map +1 -1
@@ -14,7 +14,7 @@ var __export = (target, all) => {
14
14
  import fs from "fs";
15
15
  import path from "path";
16
16
  import url from "url";
17
- var VERSION = "4.0.0-next.2";
17
+ var VERSION = "4.0.0-next.4";
18
18
  var IS_PROCESS_AVAILABLE = typeof process !== "undefined";
19
19
  var IS_NODE_ENV = IS_PROCESS_AVAILABLE && process?.release?.name === "node";
20
20
  var IS_FS_AVAILABLE = !isEmpty(fs);
@@ -6759,7 +6759,7 @@ var uint16_to_float32 = /* @__PURE__ */ (function() {
6759
6759
  // src/backends/onnx.js
6760
6760
  import * as ONNX_NODE from "onnxruntime-node";
6761
6761
 
6762
- // ../../node_modules/.pnpm/onnxruntime-web@1.25.0-dev.20260209-a3749f1353/node_modules/onnxruntime-web/dist/ort.webgpu.bundle.min.mjs
6762
+ // ../../node_modules/.pnpm/onnxruntime-web@1.25.0-dev.20260212-1a71a5f46e/node_modules/onnxruntime-web/dist/ort.webgpu.bundle.min.mjs
6763
6763
  var ort_webgpu_bundle_min_exports = {};
6764
6764
  __export(ort_webgpu_bundle_min_exports, {
6765
6765
  InferenceSession: () => nc,
@@ -8728,7 +8728,7 @@ async function is(a = {}) {
8728
8728
  we(`invalid type for getValue: ${t}`);
8729
8729
  }
8730
8730
  }, r.UTF8ToString = ct, r.stringToUTF8 = Pe, r.lengthBytesUTF8 = _e;
8731
- var bo, wo, Rr, $t, Te, mt, go, To, vo, Eo, So, Ao, Io, xo, Lo, Oo, Bo, Nr, Wr, kr, Fr, Et, Gr, Mo, $r, Uo, Co, Do, zr, Po, _o, Vr, k, St, Ro, C, zt, D, No, Hr, Wo, ko, Fo, jr, Go, $o, zo, Vo, Ho, jo, Yo, qo, Jo, Xo, Zo, Ko, Qo, ea, ta, ra, na, oa, aa, sa, ia, ua, fa, ca, da, la, pa, ma, ha, ya, ba, wa, ga, Ta, va, Ea, Sa, Aa, Ia, xa, Ne, df = [Ye, gr, In, Mn, Un, Cn, Dn, Pn, _n, Rn, Nn, Wn, kn, Fn, Gn, $n, eo, to, ro, co, lo, po, mo, ho, yo], Yr = { 1116828: (e, t, n, o, u) => {
8731
+ var bo, wo, Rr, $t, Te, mt, go, To, vo, Eo, So, Ao, Io, xo, Lo, Oo, Bo, Nr, Wr, kr, Fr, Et, Gr, Mo, $r, Uo, Co, Do, zr, Po, _o, Vr, k, St, Ro, C, zt, D, No, Hr, Wo, ko, Fo, jr, Go, $o, zo, Vo, Ho, jo, Yo, qo, Jo, Xo, Zo, Ko, Qo, ea, ta, ra, na, oa, aa, sa, ia, ua, fa, ca, da, la, pa, ma, ha, ya, ba, wa, ga, Ta, va, Ea, Sa, Aa, Ia, xa, Ne, df = [Ye, gr, In, Mn, Un, Cn, Dn, Pn, _n, Rn, Nn, Wn, kn, Fn, Gn, $n, eo, to, ro, co, lo, po, mo, ho, yo], Yr = { 1117404: (e, t, n, o, u) => {
8732
8732
  if (r === void 0 || !r.Zc) return 1;
8733
8733
  if ((e = ct(Number(e >>> 0))).startsWith("./") && (e = e.substring(2)), !(e = r.Zc.get(e))) return 2;
8734
8734
  if (t = Number(t >>> 0), n = Number(n >>> 0), o = Number(o >>> 0), t + n > e.byteLength) return 3;
@@ -8748,11 +8748,11 @@ async function is(a = {}) {
8748
8748
  } catch {
8749
8749
  return 4;
8750
8750
  }
8751
- }, 1117652: (e, t, n) => {
8751
+ }, 1118228: (e, t, n) => {
8752
8752
  r.ke(e, (l(), J).subarray(t >>> 0, t + n >>> 0));
8753
- }, 1117716: () => r.Ie(), 1117758: (e) => {
8753
+ }, 1118292: () => r.Ie(), 1118334: (e) => {
8754
8754
  r.je(e);
8755
- }, 1117795: () => typeof wasmOffsetConverter < "u" };
8755
+ }, 1118371: () => typeof wasmOffsetConverter < "u" };
8756
8756
  function lf() {
8757
8757
  return typeof wasmOffsetConverter < "u";
8758
8758
  }
@@ -10680,7 +10680,7 @@ var qs = F(() => {
10680
10680
  ze();
10681
10681
  ze();
10682
10682
  ze();
10683
- var rs = "1.25.0-dev.20260209-a3749f1353";
10683
+ var rs = "1.25.0-dev.20260212-1a71a5f46e";
10684
10684
  var Ol = tn;
10685
10685
  {
10686
10686
  let a = (qs(), Ht(Ys)).wasmBackend;
@@ -19979,6 +19979,7 @@ function getNormalizedConfig(config) {
19979
19979
  case "granite":
19980
19980
  case "granitemoehybrid":
19981
19981
  case "cohere":
19982
+ case "cohere2":
19982
19983
  case "mistral":
19983
19984
  case "starcoder2":
19984
19985
  case "qwen2":
@@ -22314,7 +22315,7 @@ var PreTrainedModel = class extends Callable2 {
22314
22315
  /**
22315
22316
  *
22316
22317
  * @param {GenerationConfig} generation_config
22317
- * @param {StoppingCriteriaList} [stopping_criteria=null]
22318
+ * @param {import('../generation/stopping_criteria.js').StoppingCriteria|import('../generation/stopping_criteria.js').StoppingCriteria[]|StoppingCriteriaList} [stopping_criteria=null]
22318
22319
  */
22319
22320
  _get_stopping_criteria(generation_config, stopping_criteria = null) {
22320
22321
  const criteria = new StoppingCriteriaList();
@@ -23058,6 +23059,9 @@ __export(models_exports, {
23058
23059
  CodeGenForCausalLM: () => CodeGenForCausalLM,
23059
23060
  CodeGenModel: () => CodeGenModel,
23060
23061
  CodeGenPreTrainedModel: () => CodeGenPreTrainedModel,
23062
+ Cohere2ForCausalLM: () => Cohere2ForCausalLM,
23063
+ Cohere2Model: () => Cohere2Model,
23064
+ Cohere2PreTrainedModel: () => Cohere2PreTrainedModel,
23061
23065
  CohereForCausalLM: () => CohereForCausalLM,
23062
23066
  CohereModel: () => CohereModel,
23063
23067
  CoherePreTrainedModel: () => CoherePreTrainedModel,
@@ -24055,6 +24059,14 @@ var CohereModel = class extends CoherePreTrainedModel {
24055
24059
  var CohereForCausalLM = class extends CoherePreTrainedModel {
24056
24060
  };
24057
24061
 
24062
+ // src/models/cohere2/modeling_cohere2.js
24063
+ var Cohere2PreTrainedModel = class extends PreTrainedModel {
24064
+ };
24065
+ var Cohere2Model = class extends Cohere2PreTrainedModel {
24066
+ };
24067
+ var Cohere2ForCausalLM = class extends Cohere2PreTrainedModel {
24068
+ };
24069
+
24058
24070
  // src/models/convbert/modeling_convbert.js
24059
24071
  var ConvBertPreTrainedModel = class extends PreTrainedModel {
24060
24072
  };
@@ -27859,6 +27871,7 @@ var MODEL_MAPPING_NAMES_DECODER_ONLY = /* @__PURE__ */ new Map([
27859
27871
  ["granite", "GraniteModel"],
27860
27872
  ["granitemoehybrid", "GraniteMoeHybridModel"],
27861
27873
  ["cohere", "CohereModel"],
27874
+ ["cohere2", "Cohere2Model"],
27862
27875
  ["gemma", "GemmaModel"],
27863
27876
  ["gemma2", "Gemma2Model"],
27864
27877
  ["vaultgemma", "VaultGemmaModel"],
@@ -27972,6 +27985,7 @@ var MODEL_FOR_CAUSAL_LM_MAPPING_NAMES = /* @__PURE__ */ new Map([
27972
27985
  ["granite", "GraniteForCausalLM"],
27973
27986
  ["granitemoehybrid", "GraniteMoeHybridForCausalLM"],
27974
27987
  ["cohere", "CohereForCausalLM"],
27988
+ ["cohere2", "Cohere2ForCausalLM"],
27975
27989
  ["gemma", "GemmaForCausalLM"],
27976
27990
  ["gemma2", "Gemma2ForCausalLM"],
27977
27991
  ["vaultgemma", "VaultGemmaForCausalLM"],
@@ -30123,6 +30137,7 @@ var TextStreamer = class extends BaseStreamer {
30123
30137
  this.token_cache = [];
30124
30138
  this.print_len = 0;
30125
30139
  this.next_tokens_are_prompt = true;
30140
+ this.special_ids = new Set(this.tokenizer.all_special_ids.map(BigInt));
30126
30141
  }
30127
30142
  /**
30128
30143
  * Receives tokens, decodes them, and prints them to stdout as soon as they form entire words.
@@ -30139,6 +30154,19 @@ var TextStreamer = class extends BaseStreamer {
30139
30154
  }
30140
30155
  const tokens = value[0];
30141
30156
  this.token_callback_function?.(tokens);
30157
+ if (tokens.length === 1 && this.special_ids.has(tokens[0])) {
30158
+ if (this.decode_kwargs.skip_special_tokens) return;
30159
+ if (this.token_cache.length > 0) {
30160
+ const text2 = this.tokenizer.decode(this.token_cache, this.decode_kwargs);
30161
+ const printable_text2 = text2.slice(this.print_len);
30162
+ this.on_finalized_text(printable_text2, false);
30163
+ this.token_cache = [];
30164
+ this.print_len = 0;
30165
+ }
30166
+ const special_text = this.tokenizer.decode(tokens, this.decode_kwargs);
30167
+ this.on_finalized_text(special_text, false);
30168
+ return;
30169
+ }
30142
30170
  this.token_cache = mergeArrays(this.token_cache, tokens);
30143
30171
  const text = this.tokenizer.decode(this.token_cache, this.decode_kwargs);
30144
30172
  let printable_text;
@@ -30469,6 +30497,9 @@ export {
30469
30497
  CodeGenPreTrainedModel,
30470
30498
  CodeGenTokenizer,
30471
30499
  CodeLlamaTokenizer,
30500
+ Cohere2ForCausalLM,
30501
+ Cohere2Model,
30502
+ Cohere2PreTrainedModel,
30472
30503
  CohereForCausalLM,
30473
30504
  CohereModel,
30474
30505
  CoherePreTrainedModel,
@@ -31197,7 +31228,7 @@ export {
31197
31228
 
31198
31229
  onnxruntime-web/dist/ort.webgpu.bundle.min.mjs:
31199
31230
  (*!
31200
- * ONNX Runtime Web v1.25.0-dev.20260209-a3749f1353
31231
+ * ONNX Runtime Web v1.25.0-dev.20260212-1a71a5f46e
31201
31232
  * Copyright (c) Microsoft Corporation. All rights reserved.
31202
31233
  * Licensed under the MIT License.
31203
31234
  *)
@@ -17,7 +17,7 @@ var emptyObj3 = {};
17
17
  var node_url_default = emptyObj3;
18
18
 
19
19
  // src/env.js
20
- var VERSION = "4.0.0-next.2";
20
+ var VERSION = "4.0.0-next.4";
21
21
  var IS_PROCESS_AVAILABLE = typeof process !== "undefined";
22
22
  var IS_NODE_ENV = IS_PROCESS_AVAILABLE && process?.release?.name === "node";
23
23
  var IS_FS_AVAILABLE = !isEmpty(node_fs_default);
@@ -16079,6 +16079,7 @@ function getNormalizedConfig(config) {
16079
16079
  case "granite":
16080
16080
  case "granitemoehybrid":
16081
16081
  case "cohere":
16082
+ case "cohere2":
16082
16083
  case "mistral":
16083
16084
  case "starcoder2":
16084
16085
  case "qwen2":
@@ -18414,7 +18415,7 @@ var PreTrainedModel = class extends Callable2 {
18414
18415
  /**
18415
18416
  *
18416
18417
  * @param {GenerationConfig} generation_config
18417
- * @param {StoppingCriteriaList} [stopping_criteria=null]
18418
+ * @param {import('../generation/stopping_criteria.js').StoppingCriteria|import('../generation/stopping_criteria.js').StoppingCriteria[]|StoppingCriteriaList} [stopping_criteria=null]
18418
18419
  */
18419
18420
  _get_stopping_criteria(generation_config, stopping_criteria = null) {
18420
18421
  const criteria = new StoppingCriteriaList();
@@ -19158,6 +19159,9 @@ __export(models_exports, {
19158
19159
  CodeGenForCausalLM: () => CodeGenForCausalLM,
19159
19160
  CodeGenModel: () => CodeGenModel,
19160
19161
  CodeGenPreTrainedModel: () => CodeGenPreTrainedModel,
19162
+ Cohere2ForCausalLM: () => Cohere2ForCausalLM,
19163
+ Cohere2Model: () => Cohere2Model,
19164
+ Cohere2PreTrainedModel: () => Cohere2PreTrainedModel,
19161
19165
  CohereForCausalLM: () => CohereForCausalLM,
19162
19166
  CohereModel: () => CohereModel,
19163
19167
  CoherePreTrainedModel: () => CoherePreTrainedModel,
@@ -20155,6 +20159,14 @@ var CohereModel = class extends CoherePreTrainedModel {
20155
20159
  var CohereForCausalLM = class extends CoherePreTrainedModel {
20156
20160
  };
20157
20161
 
20162
+ // src/models/cohere2/modeling_cohere2.js
20163
+ var Cohere2PreTrainedModel = class extends PreTrainedModel {
20164
+ };
20165
+ var Cohere2Model = class extends Cohere2PreTrainedModel {
20166
+ };
20167
+ var Cohere2ForCausalLM = class extends Cohere2PreTrainedModel {
20168
+ };
20169
+
20158
20170
  // src/models/convbert/modeling_convbert.js
20159
20171
  var ConvBertPreTrainedModel = class extends PreTrainedModel {
20160
20172
  };
@@ -23959,6 +23971,7 @@ var MODEL_MAPPING_NAMES_DECODER_ONLY = /* @__PURE__ */ new Map([
23959
23971
  ["granite", "GraniteModel"],
23960
23972
  ["granitemoehybrid", "GraniteMoeHybridModel"],
23961
23973
  ["cohere", "CohereModel"],
23974
+ ["cohere2", "Cohere2Model"],
23962
23975
  ["gemma", "GemmaModel"],
23963
23976
  ["gemma2", "Gemma2Model"],
23964
23977
  ["vaultgemma", "VaultGemmaModel"],
@@ -24072,6 +24085,7 @@ var MODEL_FOR_CAUSAL_LM_MAPPING_NAMES = /* @__PURE__ */ new Map([
24072
24085
  ["granite", "GraniteForCausalLM"],
24073
24086
  ["granitemoehybrid", "GraniteMoeHybridForCausalLM"],
24074
24087
  ["cohere", "CohereForCausalLM"],
24088
+ ["cohere2", "Cohere2ForCausalLM"],
24075
24089
  ["gemma", "GemmaForCausalLM"],
24076
24090
  ["gemma2", "Gemma2ForCausalLM"],
24077
24091
  ["vaultgemma", "VaultGemmaForCausalLM"],
@@ -26223,6 +26237,7 @@ var TextStreamer = class extends BaseStreamer {
26223
26237
  this.token_cache = [];
26224
26238
  this.print_len = 0;
26225
26239
  this.next_tokens_are_prompt = true;
26240
+ this.special_ids = new Set(this.tokenizer.all_special_ids.map(BigInt));
26226
26241
  }
26227
26242
  /**
26228
26243
  * Receives tokens, decodes them, and prints them to stdout as soon as they form entire words.
@@ -26239,6 +26254,19 @@ var TextStreamer = class extends BaseStreamer {
26239
26254
  }
26240
26255
  const tokens = value[0];
26241
26256
  this.token_callback_function?.(tokens);
26257
+ if (tokens.length === 1 && this.special_ids.has(tokens[0])) {
26258
+ if (this.decode_kwargs.skip_special_tokens) return;
26259
+ if (this.token_cache.length > 0) {
26260
+ const text2 = this.tokenizer.decode(this.token_cache, this.decode_kwargs);
26261
+ const printable_text2 = text2.slice(this.print_len);
26262
+ this.on_finalized_text(printable_text2, false);
26263
+ this.token_cache = [];
26264
+ this.print_len = 0;
26265
+ }
26266
+ const special_text = this.tokenizer.decode(tokens, this.decode_kwargs);
26267
+ this.on_finalized_text(special_text, false);
26268
+ return;
26269
+ }
26242
26270
  this.token_cache = mergeArrays(this.token_cache, tokens);
26243
26271
  const text = this.tokenizer.decode(this.token_cache, this.decode_kwargs);
26244
26272
  let printable_text;
@@ -26569,6 +26597,9 @@ export {
26569
26597
  CodeGenPreTrainedModel,
26570
26598
  CodeGenTokenizer,
26571
26599
  CodeLlamaTokenizer,
26600
+ Cohere2ForCausalLM,
26601
+ Cohere2Model,
26602
+ Cohere2PreTrainedModel,
26572
26603
  CohereForCausalLM,
26573
26604
  CohereModel,
26574
26605
  CoherePreTrainedModel,