@huggingface/transformers 3.1.0 → 3.1.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (132) hide show
  1. package/README.md +7 -3
  2. package/dist/ort-wasm-simd-threaded.jsep.wasm +0 -0
  3. package/dist/transformers.cjs +965 -195
  4. package/dist/transformers.cjs.map +1 -1
  5. package/dist/transformers.js +2251 -1360
  6. package/dist/transformers.js.map +1 -1
  7. package/dist/transformers.min.cjs +1 -352
  8. package/dist/transformers.min.cjs.map +1 -1
  9. package/dist/transformers.min.js +1 -415
  10. package/dist/transformers.min.js.map +1 -1
  11. package/dist/transformers.min.mjs +1 -352
  12. package/dist/transformers.min.mjs.map +1 -1
  13. package/dist/transformers.mjs +979 -194
  14. package/dist/transformers.mjs.map +1 -1
  15. package/package.json +11 -16
  16. package/src/backends/onnx.js +2 -7
  17. package/src/configs.js +3 -1
  18. package/src/env.js +6 -6
  19. package/src/generation/configuration_utils.js +7 -0
  20. package/src/generation/logits_process.js +22 -16
  21. package/src/generation/streamers.js +7 -2
  22. package/src/models/idefics3/image_processing_idefics3.js +219 -0
  23. package/src/models/idefics3/processing_idefics3.js +136 -0
  24. package/src/models/image_processors.js +1 -0
  25. package/src/models/paligemma/processing_paligemma.js +82 -0
  26. package/src/models/processors.js +2 -0
  27. package/src/models.js +169 -39
  28. package/src/tokenizers.js +12 -1
  29. package/src/utils/core.js +53 -9
  30. package/src/utils/dtypes.js +2 -1
  31. package/src/utils/hub.js +8 -12
  32. package/src/utils/image.js +59 -16
  33. package/src/utils/tensor.js +6 -1
  34. package/types/backends/onnx.d.ts +2 -2
  35. package/types/backends/onnx.d.ts.map +1 -1
  36. package/types/base/feature_extraction_utils.d.ts +1 -1
  37. package/types/base/feature_extraction_utils.d.ts.map +1 -1
  38. package/types/base/image_processors_utils.d.ts +2 -2
  39. package/types/base/image_processors_utils.d.ts.map +1 -1
  40. package/types/base/processing_utils.d.ts +4 -4
  41. package/types/base/processing_utils.d.ts.map +1 -1
  42. package/types/configs.d.ts +7 -7
  43. package/types/configs.d.ts.map +1 -1
  44. package/types/env.d.ts +2 -2
  45. package/types/env.d.ts.map +1 -1
  46. package/types/generation/configuration_utils.d.ts +7 -1
  47. package/types/generation/configuration_utils.d.ts.map +1 -1
  48. package/types/generation/logits_process.d.ts +32 -22
  49. package/types/generation/logits_process.d.ts.map +1 -1
  50. package/types/generation/logits_sampler.d.ts.map +1 -1
  51. package/types/generation/parameters.d.ts +5 -5
  52. package/types/generation/stopping_criteria.d.ts +1 -1
  53. package/types/generation/stopping_criteria.d.ts.map +1 -1
  54. package/types/generation/streamers.d.ts +15 -10
  55. package/types/generation/streamers.d.ts.map +1 -1
  56. package/types/models/audio_spectrogram_transformer/feature_extraction_audio_spectrogram_transformer.d.ts +1 -1
  57. package/types/models/audio_spectrogram_transformer/feature_extraction_audio_spectrogram_transformer.d.ts.map +1 -1
  58. package/types/models/auto/feature_extraction_auto.d.ts.map +1 -1
  59. package/types/models/auto/image_processing_auto.d.ts.map +1 -1
  60. package/types/models/auto/processing_auto.d.ts +1 -1
  61. package/types/models/auto/processing_auto.d.ts.map +1 -1
  62. package/types/models/clap/feature_extraction_clap.d.ts +1 -1
  63. package/types/models/clap/feature_extraction_clap.d.ts.map +1 -1
  64. package/types/models/detr/image_processing_detr.d.ts +11 -11
  65. package/types/models/detr/image_processing_detr.d.ts.map +1 -1
  66. package/types/models/donut/image_processing_donut.d.ts +1 -1
  67. package/types/models/donut/image_processing_donut.d.ts.map +1 -1
  68. package/types/models/florence2/processing_florence2.d.ts.map +1 -1
  69. package/types/models/idefics3/image_processing_idefics3.d.ts +40 -0
  70. package/types/models/idefics3/image_processing_idefics3.d.ts.map +1 -0
  71. package/types/models/idefics3/processing_idefics3.d.ts +19 -0
  72. package/types/models/idefics3/processing_idefics3.d.ts.map +1 -0
  73. package/types/models/image_processors.d.ts +1 -0
  74. package/types/models/janus/image_processing_janus.d.ts +1 -1
  75. package/types/models/janus/image_processing_janus.d.ts.map +1 -1
  76. package/types/models/janus/processing_janus.d.ts.map +1 -1
  77. package/types/models/maskformer/image_processing_maskformer.d.ts +8 -8
  78. package/types/models/maskformer/image_processing_maskformer.d.ts.map +1 -1
  79. package/types/models/mgp_str/processing_mgp_str.d.ts +2 -2
  80. package/types/models/mgp_str/processing_mgp_str.d.ts.map +1 -1
  81. package/types/models/owlvit/image_processing_owlvit.d.ts.map +1 -1
  82. package/types/models/paligemma/processing_paligemma.d.ts +12 -0
  83. package/types/models/paligemma/processing_paligemma.d.ts.map +1 -0
  84. package/types/models/processors.d.ts +2 -0
  85. package/types/models/pyannote/feature_extraction_pyannote.d.ts.map +1 -1
  86. package/types/models/pyannote/processing_pyannote.d.ts +1 -1
  87. package/types/models/pyannote/processing_pyannote.d.ts.map +1 -1
  88. package/types/models/qwen2_vl/processing_qwen2_vl.d.ts.map +1 -1
  89. package/types/models/rt_detr/image_processing_rt_detr.d.ts.map +1 -1
  90. package/types/models/sam/image_processing_sam.d.ts.map +1 -1
  91. package/types/models/seamless_m4t/feature_extraction_seamless_m4t.d.ts +1 -1
  92. package/types/models/seamless_m4t/feature_extraction_seamless_m4t.d.ts.map +1 -1
  93. package/types/models/segformer/image_processing_segformer.d.ts.map +1 -1
  94. package/types/models/speecht5/processing_speecht5.d.ts.map +1 -1
  95. package/types/models/swin2sr/image_processing_swin2sr.d.ts +1 -1
  96. package/types/models/swin2sr/image_processing_swin2sr.d.ts.map +1 -1
  97. package/types/models/vitmatte/image_processing_vitmatte.d.ts.map +1 -1
  98. package/types/models/vitpose/image_processing_vitpose.d.ts +1 -1
  99. package/types/models/vitpose/image_processing_vitpose.d.ts.map +1 -1
  100. package/types/models/wav2vec2/feature_extraction_wav2vec2.d.ts.map +1 -1
  101. package/types/models/wav2vec2/processing_wav2vec2.d.ts.map +1 -1
  102. package/types/models/wespeaker/feature_extraction_wespeaker.d.ts +1 -1
  103. package/types/models/wespeaker/feature_extraction_wespeaker.d.ts.map +1 -1
  104. package/types/models/whisper/feature_extraction_whisper.d.ts +1 -1
  105. package/types/models/whisper/feature_extraction_whisper.d.ts.map +1 -1
  106. package/types/models/whisper/generation_whisper.d.ts.map +1 -1
  107. package/types/models/whisper/processing_whisper.d.ts.map +1 -1
  108. package/types/models/yolos/image_processing_yolos.d.ts.map +1 -1
  109. package/types/models.d.ts +44 -10
  110. package/types/models.d.ts.map +1 -1
  111. package/types/ops/registry.d.ts.map +1 -1
  112. package/types/pipelines.d.ts +26 -51
  113. package/types/pipelines.d.ts.map +1 -1
  114. package/types/tokenizers.d.ts +10 -6
  115. package/types/tokenizers.d.ts.map +1 -1
  116. package/types/utils/audio.d.ts.map +1 -1
  117. package/types/utils/constants.d.ts.map +1 -1
  118. package/types/utils/core.d.ts +94 -22
  119. package/types/utils/core.d.ts.map +1 -1
  120. package/types/utils/data-structures.d.ts.map +1 -1
  121. package/types/utils/devices.d.ts.map +1 -1
  122. package/types/utils/dtypes.d.ts +3 -2
  123. package/types/utils/dtypes.d.ts.map +1 -1
  124. package/types/utils/generic.d.ts.map +1 -1
  125. package/types/utils/hub.d.ts +3 -3
  126. package/types/utils/hub.d.ts.map +1 -1
  127. package/types/utils/image.d.ts +14 -1
  128. package/types/utils/image.d.ts.map +1 -1
  129. package/types/utils/maths.d.ts +10 -10
  130. package/types/utils/maths.d.ts.map +1 -1
  131. package/types/utils/tensor.d.ts +10 -8
  132. package/types/utils/tensor.d.ts.map +1 -1
@@ -1 +1 @@
1
- {"version":3,"file":"stopping_criteria.d.ts","sourceRoot":"","sources":["../../src/generation/stopping_criteria.js"],"names":[],"mappings":";;;;AAUA;;GAEG;AACH;IACI;;;;;;;;OAQG;IACH,iBAPW,MAAM,EAAE,EAAE,UAEV,MAAM,EAAE,EAAE,GAGR,OAAO,EAAE,CAIrB;CACJ;;;;;AACD;GACG;AACH;IAMQ,gBAAkB;IAGtB;;;;OAIG;IACH,WAFW,gBAAgB,QAI1B;IAED;;;;OAIG;IACH,cAFW,gBAAgB,GAAC,oBAAoB,GAAC,gBAAgB,EAAE,QASlE;IAED,0CASC;IAED,2CAEC;CACJ;AAED;;;GAGG;AACH;IAEI;;;;OAIG;IACH,wBAHW,MAAM,4BACN,MAAM,EAMhB;IAFG,mBAA4B;IAC5B,gCAAsD;IAG1D,2BAEC;CACJ;AAID;;;GAGG;AACH;IAEI;;;;OAIG;IACH,0BAHW,MAAM,GAAC,MAAM,EAAE,EASzB;IADG,uBAAgC;CAgBvC;AAED;;GAEG;AACH;IAGQ,qBAAwB;IAG5B,kBAEC;IAED,cAEC;IAED,0CAEC;CACJ"}
1
+ {"version":3,"file":"stopping_criteria.d.ts","sourceRoot":"","sources":["../../src/generation/stopping_criteria.js"],"names":[],"mappings":";KAQ4C,GAAG;UAA0B,GAAE;;AAE3E;;GAEG;AACH;IACI;;;;;;;;OAQG;IACH,iBAPW,MAAM,EAAE,EAAE,UAEV,MAAM,EAAE,EAAE,GAGR,OAAO,EAAE,CAIrB;CACJ;;KAlB2C,GAAG;UAA0B,GAAE;;AAmB3E;GACG;AACH;IAMQ,gBAAkB;IAGtB;;;;OAIG;IACH,WAFW,gBAAgB,QAI1B;IAED;;;;OAIG;IACH,cAFW,gBAAgB,GAAC,oBAAoB,GAAC,gBAAgB,EAAE,QASlE;IAED,0CASC;IAED,wCAEC;CACJ;AAED;;;GAGG;AACH;IAEI;;;;OAIG;IACH,wBAHW,MAAM,4BACN,MAAM,EAMhB;IAFG,mBAA4B;IAC5B,gCAAsD;IAG1D,2BAEC;CACJ;AAID;;;GAGG;AACH;IAEI;;;;OAIG;IACH,0BAHW,MAAM,GAAC,MAAM,EAAE,EASzB;IADG,uBAAgC;CAgBvC;AAED;;GAEG;AACH;IAGQ,qBAAwB;IAG5B,kBAEC;IAED,cAEC;IAED,0CAEC;CACJ"}
@@ -16,18 +16,23 @@ export class TextStreamer extends BaseStreamer {
16
16
  /**
17
17
  *
18
18
  * @param {import('../tokenizers.js').PreTrainedTokenizer} tokenizer
19
+ * @param {Object} options
20
+ * @param {boolean} [options.skip_prompt=false] Whether to skip the prompt tokens
21
+ * @param {function(string): void} [options.callback_function=null] Function to call when a piece of text is ready to display
22
+ * @param {function(bigint[]): void} [options.token_callback_function=null] Function to call when a new token is generated
23
+ * @param {Object} [options.decode_kwargs={}] Additional keyword arguments to pass to the tokenizer's decode method
19
24
  */
20
- constructor(tokenizer: import('../tokenizers.js').PreTrainedTokenizer, { skip_prompt, callback_function, token_callback_function, decode_kwargs, ...kwargs }?: {
25
+ constructor(tokenizer: import("../tokenizers.js").PreTrainedTokenizer, { skip_prompt, callback_function, token_callback_function, decode_kwargs, ...kwargs }?: {
21
26
  skip_prompt?: boolean;
22
- callback_function?: any;
23
- token_callback_function?: any;
24
- decode_kwargs?: {};
27
+ callback_function?: (arg0: string) => void;
28
+ token_callback_function?: (arg0: bigint[]) => void;
29
+ decode_kwargs?: any;
25
30
  });
26
31
  tokenizer: import("../tokenizers.js").PreTrainedTokenizer;
27
32
  skip_prompt: boolean;
28
- callback_function: any;
29
- token_callback_function: any;
30
- decode_kwargs: {};
33
+ callback_function: (x: any) => void;
34
+ token_callback_function: (arg0: bigint[]) => void;
35
+ decode_kwargs: any;
31
36
  token_cache: any[];
32
37
  print_len: number;
33
38
  next_tokens_are_prompt: boolean;
@@ -52,7 +57,7 @@ export class WhisperTextStreamer extends TextStreamer {
52
57
  * @param {Object} options
53
58
  * @param {boolean} [options.skip_prompt=false] Whether to skip the prompt tokens
54
59
  * @param {function(string): void} [options.callback_function=null] Function to call when a piece of text is ready to display
55
- * @param {function(string): void} [options.token_callback_function=null] Function to call when a new token is generated
60
+ * @param {function(bigint[]): void} [options.token_callback_function=null] Function to call when a new token is generated
56
61
  * @param {function(number): void} [options.on_chunk_start=null] Function to call when a new chunk starts
57
62
  * @param {function(number): void} [options.on_chunk_end=null] Function to call when a chunk ends
58
63
  * @param {function(): void} [options.on_finalize=null] Function to call when the stream is finalized
@@ -60,10 +65,10 @@ export class WhisperTextStreamer extends TextStreamer {
60
65
  * @param {boolean} [options.skip_special_tokens=true] Whether to skip special tokens when decoding
61
66
  * @param {Object} [options.decode_kwargs={}] Additional keyword arguments to pass to the tokenizer's decode method
62
67
  */
63
- constructor(tokenizer: import('../tokenizers.js').WhisperTokenizer, { skip_prompt, callback_function, token_callback_function, on_chunk_start, on_chunk_end, on_finalize, time_precision, skip_special_tokens, decode_kwargs, }?: {
68
+ constructor(tokenizer: import("../tokenizers.js").WhisperTokenizer, { skip_prompt, callback_function, token_callback_function, on_chunk_start, on_chunk_end, on_finalize, time_precision, skip_special_tokens, decode_kwargs, }?: {
64
69
  skip_prompt?: boolean;
65
70
  callback_function?: (arg0: string) => void;
66
- token_callback_function?: (arg0: string) => void;
71
+ token_callback_function?: (arg0: bigint[]) => void;
67
72
  on_chunk_start?: (arg0: number) => void;
68
73
  on_chunk_end?: (arg0: number) => void;
69
74
  on_finalize?: () => void;
@@ -1 +1 @@
1
- {"version":3,"file":"streamers.d.ts","sourceRoot":"","sources":["../../src/generation/streamers.js"],"names":[],"mappings":"AASA;IACI;;;OAGG;IACH,WAFW,MAAM,EAAE,EAAE,QAIpB;IAED;;OAEG;IACH,YAEC;CACJ;AAMD;;GAEG;AACH;IACI;;;OAGG;IACH,uBAFW,OAAO,kBAAkB,EAAE,mBAAmB;;;;;OAoBxD;IAVG,0DAA0B;IAC1B,qBAA8B;IAC9B,uBAA0D;IAC1D,6BAAsD;IACtD,kBAAoD;IAGpD,mBAAqB;IACrB,kBAAkB;IAClB,gCAAkC;IA6DtC;;;;OAIG;IACH,wBAHW,MAAM,cACN,OAAO,QASjB;CACJ;AAED;;;;;;;GAOG;AACH;IACI;;;;;;;;;;;;OAYG;IACH,uBAZW,OAAO,kBAAkB,EAAE,gBAAgB;QAEzB,WAAW,GAA7B,OAAO;QAC0B,iBAAiB,UAAzC,MAAM,KAAG,IAAI;QACW,uBAAuB,UAA/C,MAAM,KAAG,IAAI;QACW,cAAc,UAAtC,MAAM,KAAG,IAAI;QACW,YAAY,UAApC,MAAM,KAAG,IAAI;QACK,WAAW,SAA1B,IAAI;QACC,cAAc,GAA/B,MAAM;QACY,mBAAmB,GAArC,OAAO;QACU,aAAa;OA4BxC;IATG,wBAAgD;IAEhD,uBA1BgB,MAAM,KAAG,IAAI,CA0BO;IACpC,qBA1BgB,MAAM,KAAG,IAAI,CA0BG;IAChC,mBA1BmB,IAAI,CA0BO;IAE9B,uBAAoC;IAEpC,+BAAkC;CAiCzC"}
1
+ {"version":3,"file":"streamers.d.ts","sourceRoot":"","sources":["../../src/generation/streamers.js"],"names":[],"mappings":"AASA;IACI;;;OAGG;IACH,WAFW,MAAM,EAAE,EAAE,QAIpB;IAED;;OAEG;IACH,YAEC;CACJ;AAMD;;GAEG;AACH;IACI;;;;;;;;OAQG;IACH,uBAPW,OAAO,kBAAkB,EAAE,mBAAmB,0FAEtD;QAA0B,WAAW,GAA7B,OAAO;QAC0B,iBAAiB,GAAlD,CAAS,IAAM,EAAN,MAAM,KAAG,IAAI;QACa,uBAAuB,GAA1D,CAAS,IAAQ,EAAR,MAAM,EAAE,KAAG,IAAI;QACP,aAAa;KACxC,EAmBA;IAVG,0DAA0B;IAC1B,qBAA8B;IAC9B,oCAA0D;IAC1D,gCAdgB,MAAM,EAAE,KAAG,IAAI,CAcuB;IACtD,mBAAoD;IAGpD,mBAAqB;IACrB,kBAAkB;IAClB,gCAAkC;IA6DtC;;;;OAIG;IACH,wBAHW,MAAM,cACN,OAAO,QASjB;CACJ;AAED;;;;;;;GAOG;AACH;IACI;;;;;;;;;;;;OAYG;IACH,uBAZW,OAAO,kBAAkB,EAAE,gBAAgB,gKAEnD;QAA0B,WAAW,GAA7B,OAAO;QAC0B,iBAAiB,GAAlD,CAAS,IAAM,EAAN,MAAM,KAAG,IAAI;QACa,uBAAuB,GAA1D,CAAS,IAAQ,EAAR,MAAM,EAAE,KAAG,IAAI;QACS,cAAc,GAA/C,CAAS,IAAM,EAAN,MAAM,KAAG,IAAI;QACW,YAAY,GAA7C,CAAS,IAAM,EAAN,MAAM,KAAG,IAAI;QACK,WAAW,GAAtC,MAAY,IAAI;QACC,cAAc,GAA/B,MAAM;QACY,mBAAmB,GAArC,OAAO;QACU,aAAa;KACxC,EA2BA;IATG,wBAAgD;IAEhD,uBA1BgB,MAAM,KAAG,IAAI,CA0BO;IACpC,qBA1BgB,MAAM,KAAG,IAAI,CA0BG;IAChC,mBA1BmB,IAAI,CA0BO;IAE9B,uBAAoC;IAEpC,+BAAkC;CAiCzC"}
@@ -1,7 +1,7 @@
1
1
  export class ASTFeatureExtractor extends FeatureExtractor {
2
2
  constructor(config: any);
3
3
  mel_filters: number[][];
4
- window: Float64Array;
4
+ window: Float64Array<ArrayBufferLike>;
5
5
  mean: any;
6
6
  std: any;
7
7
  /**
@@ -1 +1 @@
1
- {"version":3,"file":"feature_extraction_audio_spectrogram_transformer.d.ts","sourceRoot":"","sources":["../../../src/models/audio_spectrogram_transformer/feature_extraction_audio_spectrogram_transformer.js"],"names":[],"mappings":"AAKA;IAEI,yBA2BC;IARG,wBAA8B;IAE9B,qBAEE;IAEF,UAA4B;IAC5B,SAA0B;IAG9B;;;;;OAKG;IACH,kCAJW,YAAY,GAAC,YAAY,cACzB,MAAM,GACJ,QAAQ,MAAM,CAAC,CAwB3B;IAGD;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY,GACvB,QAAQ;QAAE,YAAY,EAAE,MAAM,CAAA;KAAE,CAAC,CAkB7C;CACJ;iCAzFuD,wCAAwC;uBACzE,uBAAuB"}
1
+ {"version":3,"file":"feature_extraction_audio_spectrogram_transformer.d.ts","sourceRoot":"","sources":["../../../src/models/audio_spectrogram_transformer/feature_extraction_audio_spectrogram_transformer.js"],"names":[],"mappings":"AAKA;IAEI,yBA2BC;IARG,wBAA8B;IAE9B,sCAEE;IAEF,UAA4B;IAC5B,SAA0B;IAG9B;;;;;OAKG;IACH,kCAJW,YAAY,GAAC,YAAY,cACzB,MAAM,GACJ,OAAO,CAAC,MAAM,CAAC,CAwB3B;IAGD;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY,GACvB,OAAO,CAAC;QAAE,YAAY,EAAE,MAAM,CAAA;KAAE,CAAC,CAkB7C;CACJ;iCAzFuD,wCAAwC;uBACzE,uBAAuB"}
@@ -1 +1 @@
1
- {"version":3,"file":"feature_extraction_auto.d.ts","sourceRoot":"","sources":["../../../src/models/auto/feature_extraction_auto.js"],"names":[],"mappings":"AAMA;IAKsD,kJAoBH;CASlD;iCArCgC,wCAAwC"}
1
+ {"version":3,"file":"feature_extraction_auto.d.ts","sourceRoot":"","sources":["../../../src/models/auto/feature_extraction_auto.js"],"names":[],"mappings":"AAMA;kFAe6F,oBAE7F;CAiBC;iCArCgC,wCAAwC"}
@@ -1 +1 @@
1
- {"version":3,"file":"image_processing_auto.d.ts","sourceRoot":"","sources":["../../../src/models/auto/image_processing_auto.js"],"names":[],"mappings":"AAMA;IAuBkkwC,gJAAqyC;CADt2yC;+BAzB8B,sCAAsC"}
1
+ {"version":3,"file":"image_processing_auto.d.ts","sourceRoot":"","sources":["../../../src/models/auto/image_processing_auto.js"],"names":[],"mappings":"AAMA;kFAuBk7xC,oBAAiB;CADl8xC;+BAzB8B,sCAAsC"}
@@ -29,7 +29,7 @@
29
29
  * ```
30
30
  */
31
31
  export class AutoProcessor {
32
- static from_pretrained(pretrained_model_name_or_path: string, options: any): Promise<Processor>;
32
+ static from_pretrained(pretrained_model_name_or_path: string, options: PretrainedProcessorOptions): Promise<Processor>;
33
33
  }
34
34
  import { Processor } from '../../base/processing_utils.js';
35
35
  //# sourceMappingURL=processing_auto.d.ts.map
@@ -1 +1 @@
1
- {"version":3,"file":"processing_auto.d.ts","sourceRoot":"","sources":["../../../src/models/auto/processing_auto.js"],"names":[],"mappings":"AAUA;;;;;;;;;;;;;;;;;;;;;;;;;;;;;GA6BG;AACH;IAoCuB,gGAwB8qB;CADpsB;0BA/FyB,gCAAgC"}
1
+ {"version":3,"file":"processing_auto.d.ts","sourceRoot":"","sources":["../../../src/models/auto/processing_auto.js"],"names":[],"mappings":"AAUA;;;;;;;;;;;;;;;;;;;;;;;;;;;;;GA6BG;AACH;;CA2DC;0BA/FyB,gCAAgC"}
@@ -2,7 +2,7 @@ export class ClapFeatureExtractor extends FeatureExtractor {
2
2
  constructor(config: any);
3
3
  mel_filters: number[][];
4
4
  mel_filters_slaney: number[][];
5
- window: Float64Array;
5
+ window: Float64Array<ArrayBufferLike>;
6
6
  /**
7
7
  * Extracts the mel spectrogram and prepares it for the mode based on the `truncation` and `padding` arguments.
8
8
  *
@@ -1 +1 @@
1
- {"version":3,"file":"feature_extraction_clap.d.ts","sourceRoot":"","sources":["../../../src/models/clap/feature_extraction_clap.js"],"names":[],"mappings":"AAKA;IAEI,yBAyBC;IAtBG,wBAQC;IAED,+BAQC;IAED,qBAAkE;IAKtE;;;;;;;;;;;;;;;;;;;;OAoBG;IACH,uBA0CC;IAED;;;;;;;;;;;;;;OAcG;IACH,kCALW,YAAY,GAAC,YAAY,eACzB,MAAM,EAAE,EAAE,eACV,MAAM,GACJ,QAAQ,MAAM,CAAC,CAoB3B;IAGD;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY;;QACvB,QAAQ;QAAE,cAAc,EAAE,MAAM,CAAA;KAAE,CAAC,CAkB/C;CACJ;iCA9JuD,wCAAwC;uBACzE,uBAAuB"}
1
+ {"version":3,"file":"feature_extraction_clap.d.ts","sourceRoot":"","sources":["../../../src/models/clap/feature_extraction_clap.js"],"names":[],"mappings":"AAKA;IAEI,yBAyBC;IAtBG,wBAQC;IAED,+BAQC;IAED,sCAAkE;IAKtE;;;;;;;;;;;;;;;;;;;;OAoBG;IACH,uBA0CC;IAED;;;;;;;;;;;;;;OAcG;IACH,kCALW,YAAY,GAAC,YAAY,eACzB,MAAM,EAAE,EAAE,eACV,MAAM,GACJ,OAAO,CAAC,MAAM,CAAC,CAoB3B;IAGD;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY;;QACvB,OAAO,CAAC;QAAE,cAAc,EAAE,MAAM,CAAA;KAAE,CAAC,CAkB/C;CACJ;iCA9JuD,wCAAwC;uBACzE,uBAAuB"}
@@ -10,33 +10,33 @@ export class DetrImageProcessor extends ImageProcessor {
10
10
  * @param {import('../../utils/image.js').RawImage[]} images The image(s) to extract features from.
11
11
  * @returns {Promise<DetrFeatureExtractorResult>} An object containing the concatenated pixel values of the preprocessed images.
12
12
  */
13
- _call(images: import('../../utils/image.js').RawImage[]): Promise<DetrFeatureExtractorResult>;
13
+ _call(images: import("../../utils/image.js").RawImage[]): Promise<DetrFeatureExtractorResult>;
14
14
  post_process_object_detection(outputs: {
15
15
  logits: import("../../utils/tensor.js").Tensor;
16
16
  pred_boxes: import("../../utils/tensor.js").Tensor;
17
17
  }, threshold?: number, target_sizes?: [number, number][], is_zero_shot?: boolean): any[];
18
- post_process_panoptic_segmentation(outputs: any, threshold?: number, mask_threshold?: number, overlap_mask_area_threshold?: number, label_ids_to_fuse?: Set<number>, target_sizes?: [number, number][]): {
18
+ post_process_panoptic_segmentation(outputs: any, threshold?: number, mask_threshold?: number, overlap_mask_area_threshold?: number, label_ids_to_fuse?: Set<number>, target_sizes?: [number, number][]): Array<{
19
19
  segmentation: import("../../utils/tensor.js").Tensor;
20
- segments_info: {
20
+ segments_info: Array<{
21
21
  id: number;
22
22
  label_id: number;
23
23
  score: number;
24
- }[];
25
- }[];
26
- post_process_instance_segmentation(outputs: any, threshold?: number, target_sizes?: [number, number][]): {
24
+ }>;
25
+ }>;
26
+ post_process_instance_segmentation(outputs: any, threshold?: number, target_sizes?: [number, number][]): Array<{
27
27
  segmentation: import("../../utils/tensor.js").Tensor;
28
- segments_info: {
28
+ segments_info: Array<{
29
29
  id: number;
30
30
  label_id: number;
31
31
  score: number;
32
- }[];
33
- }[];
32
+ }>;
33
+ }>;
34
34
  }
35
35
  export class DetrFeatureExtractor extends DetrImageProcessor {
36
36
  }
37
37
  export type DetrFeatureExtractorResultProps = {
38
- pixel_mask: import('../../utils/tensor.js').Tensor;
38
+ pixel_mask: import("../../utils/tensor.js").Tensor;
39
39
  };
40
- export type DetrFeatureExtractorResult = import('../../base/image_processors_utils.js').ImageProcessorResult & DetrFeatureExtractorResultProps;
40
+ export type DetrFeatureExtractorResult = import("../../base/image_processors_utils.js").ImageProcessorResult & DetrFeatureExtractorResultProps;
41
41
  import { ImageProcessor } from "../../base/image_processors_utils.js";
42
42
  //# sourceMappingURL=image_processing_detr.d.ts.map
@@ -1 +1 @@
1
- {"version":3,"file":"image_processing_detr.d.ts","sourceRoot":"","sources":["../../../src/models/detr/image_processing_detr.js"],"names":[],"mappings":"AAUA;;;;GAIG;AAEH;IACI;;;;;OAKG;IACH,cAHW,OAAO,sBAAsB,EAAE,QAAQ,EAAE,GACvC,QAAQ,0BAA0B,CAAC,CAY/C;IAmBwmB;;;6FAAu6F;IAA8yU;;;;;;;QAA4qG;IAAA;;;;;;;QAAktB;CAH/rjB;AAED;CAAgE;;gBAvClD,OAAO,uBAAuB,EAAE,MAAM;;yCACvC,OAAO,sCAAsC,EAAE,oBAAoB,GAAG,+BAA+B;+BAR3G,sCAAsC"}
1
+ {"version":3,"file":"image_processing_detr.d.ts","sourceRoot":"","sources":["../../../src/models/detr/image_processing_detr.js"],"names":[],"mappings":"AAUA;;;;GAIG;AAEH;IACI;;;;;OAKG;IACH,cAHW,OAAO,sBAAsB,EAAE,QAAQ,EAAE,GACvC,OAAO,CAAC,0BAA0B,CAAC,CAY/C;;;;;;;;;;;;;;;;;;;;;CAgBJ;AAED;CAAgE;;gBAvClD,OAAO,uBAAuB,EAAE,MAAM;;yCACvC,OAAO,sCAAsC,EAAE,oBAAoB,GAAG,+BAA+B;+BAR3G,sCAAsC"}
@@ -1,5 +1,5 @@
1
1
  export class DonutImageProcessor extends ImageProcessor {
2
- pad_image(pixelData: any, imgDims: any, padSize: any, options?: {}): [Float32Array, number[]];
2
+ pad_image(pixelData: any, imgDims: any, padSize: any, options?: {}): [Float32Array<ArrayBufferLike>, number[]];
3
3
  }
4
4
  export class DonutFeatureExtractor extends DonutImageProcessor {
5
5
  }
@@ -1 +1 @@
1
- {"version":3,"file":"image_processing_donut.d.ts","sourceRoot":"","sources":["../../../src/models/donut/image_processing_donut.js"],"names":[],"mappings":"AAIA;IACI,8FAuBC;CACJ;AACD;CAAkE;+BA5B3D,sCAAsC"}
1
+ {"version":3,"file":"image_processing_donut.d.ts","sourceRoot":"","sources":["../../../src/models/donut/image_processing_donut.js"],"names":[],"mappings":"AAIA;IACI,+GAuBC;CACJ;AACD;CAAkE;+BA5B3D,sCAAsC"}
@@ -1 +1 @@
1
- {"version":3,"file":"processing_florence2.d.ts","sourceRoot":"","sources":["../../../src/models/florence2/processing_florence2.js"],"names":[],"mappings":"AAIA;IACI,6CAAsC;IACtC,wDAAiD;IAEjD,0CAuBC;IAdG,kCAAkC;IAClC,mCADW,IAAI,MAAM,EAAE,MAAM,CAAC,CAC2E;IAEzG,kCAAkC;IAClC,6BADW,IAAI,MAAM,EAAE,MAAM,CAAC,CAC+D;IAE7F,kCAAkC;IAClC,yBADW,IAAI,MAAM,EAAE,MAAM,CAAC,CACuD;IAErF;;;MAGC;IACD,qBAAwB;IAG5B;;;;OAIG;IACH,wBAHW,MAAM,GAAC,MAAM,EAAE,GACb,MAAM,EAAE,CA6BpB;IAED;;;;;OAKG;IACH,8BAJW,MAAM,QACN,MAAM,cACN,CAAC,MAAM,EAAE,MAAM,CAAC;;;;;MAsC1B;IAID,0DAaC;CACJ;0BA/HyB,gCAAgC;8BAE5B,qBAAqB;mCADhB,kCAAkC"}
1
+ {"version":3,"file":"processing_florence2.d.ts","sourceRoot":"","sources":["../../../src/models/florence2/processing_florence2.js"],"names":[],"mappings":"AAIA;IACI,6CAAsC;IACtC,wDAAiD;IAEjD,0CAuBC;IAdG,kCAAkC;IAClC,mCADW,GAAG,CAAC,MAAM,EAAE,MAAM,CAAC,CAC2E;IAEzG,kCAAkC;IAClC,6BADW,GAAG,CAAC,MAAM,EAAE,MAAM,CAAC,CAC+D;IAE7F,kCAAkC;IAClC,yBADW,GAAG,CAAC,MAAM,EAAE,MAAM,CAAC,CACuD;IAErF;;;MAGC;IACD,qBAAwB;IAG5B;;;;OAIG;IACH,wBAHW,MAAM,GAAC,MAAM,EAAE,GACb,MAAM,EAAE,CA6BpB;IAED;;;;;OAKG;IACH,8BAJW,MAAM,QACN,MAAM,cACN,CAAC,MAAM,EAAE,MAAM,CAAC;;;;;MAsC1B;IAID,0DAaC;CACJ;0BA/HyB,gCAAgC;8BAE5B,qBAAqB;mCADhB,kCAAkC"}
@@ -0,0 +1,40 @@
1
+ export class Idefics3ImageProcessor extends ImageProcessor {
2
+ constructor(config: any);
3
+ do_image_splitting: any;
4
+ max_image_size: any;
5
+ /**
6
+ * @typedef {import('../../utils/image.js').RawImage} RawImage
7
+ * @typedef {import('../../utils/tensor.js').Tensor} Tensor
8
+ */
9
+ /**
10
+ * Calculate size to resize images to, to be multiples of `vision_encoder_max_size` while preserving the aspect ratio.
11
+ * @param {Tensor} pixel_values Tensor of the image to resize.
12
+ * @param {number} vision_encoder_max_size Maximum size of the output image. If the image is larger than this size,
13
+ * it will be split into patches of this size, and the original image will be concatenated with the patches, resized to max_size.
14
+ */
15
+ get_resize_for_vision_encoder(pixel_values: import("../../utils/tensor.js").Tensor, vision_encoder_max_size: number): {
16
+ height: number;
17
+ width: number;
18
+ };
19
+ /** @param {RawImage|RawImage[]|RawImage[][]} images */
20
+ _call(images: import("../../utils/image.js").RawImage | import("../../utils/image.js").RawImage[] | import("../../utils/image.js").RawImage[][], { do_image_splitting, return_row_col_info, }?: {
21
+ do_image_splitting?: any;
22
+ return_row_col_info?: boolean;
23
+ }): Promise<{
24
+ rows?: any[][];
25
+ cols?: any[][];
26
+ pixel_values: import("../../utils/tensor.js").Tensor;
27
+ pixel_attention_mask: import("../../utils/tensor.js").Tensor;
28
+ original_sizes: import("../../base/image_processors_utils.js").HeightWidth[];
29
+ reshaped_input_sizes: import("../../base/image_processors_utils.js").HeightWidth[];
30
+ }>;
31
+ split_image(pixel_values: any, { longest_edge }: {
32
+ longest_edge: any;
33
+ }): Promise<{
34
+ frames: any[];
35
+ num_splits_h: number;
36
+ num_splits_w: number;
37
+ }>;
38
+ }
39
+ import { ImageProcessor } from "../../base/image_processors_utils.js";
40
+ //# sourceMappingURL=image_processing_idefics3.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"image_processing_idefics3.d.ts","sourceRoot":"","sources":["../../../src/models/idefics3/image_processing_idefics3.js"],"names":[],"mappings":"AAOA;IACI,yBAKC;IAFG,wBAA2D;IAC3D,oBAA2C;IAG/C;;;OAGG;IAEH;;;;;OAKG;IACH,6GAHW,MAAM;;;MAiBhB;IAED,uDAAuD;IACvD,cADY,0CAAS,yCAAU,GAAC,yCAAU,EAAE;;;;;;;;;;OA4H3C;IAED;;;;;;OAiDC;CACJ;+BAtNM,sCAAsC"}
@@ -0,0 +1,19 @@
1
+ export class Idefics3Processor extends Processor {
2
+ static image_processor_class: typeof AutoImageProcessor;
3
+ static tokenizer_class: typeof AutoTokenizer;
4
+ fake_image_token: string;
5
+ image_token: string;
6
+ global_img_token: string;
7
+ /**
8
+ *
9
+ * @param {string|string[]} text
10
+ * @param {RawImage|RawImage[]|RawImage[][]} images
11
+ * @returns {Promise<any>}
12
+ */
13
+ _call(text: string | string[], images?: RawImage | RawImage[] | RawImage[][], options?: {}): Promise<any>;
14
+ }
15
+ import { Processor } from "../../base/processing_utils.js";
16
+ import { RawImage } from "../../utils/image.js";
17
+ import { AutoImageProcessor } from "../auto/image_processing_auto.js";
18
+ import { AutoTokenizer } from "../../tokenizers.js";
19
+ //# sourceMappingURL=processing_idefics3.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"processing_idefics3.d.ts","sourceRoot":"","sources":["../../../src/models/idefics3/processing_idefics3.js"],"names":[],"mappings":"AA6DA;IACI,wDAAiD;IACjD,6CAAsC;IAGtC,yBAA+C;IAC/C,oBAAwB;IACxB,yBAAkC;IAElC;;;;;OAKG;IACH,YAJW,MAAM,GAAC,MAAM,EAAE,WACf,QAAQ,GAAC,QAAQ,EAAE,GAAC,QAAQ,EAAE,EAAE,iBAC9B,OAAO,CAAC,GAAG,CAAC,CA4DxB;CACJ;0BAtIyB,gCAAgC;yBAGjC,sBAAsB;mCAFZ,kCAAkC;8BACvC,qBAAqB"}
@@ -9,6 +9,7 @@ export * from "./donut/image_processing_donut.js";
9
9
  export * from "./dpt/image_processing_dpt.js";
10
10
  export * from "./efficientnet/image_processing_efficientnet.js";
11
11
  export * from "./glpn/image_processing_glpn.js";
12
+ export * from "./idefics3/image_processing_idefics3.js";
12
13
  export * from "./janus/image_processing_janus.js";
13
14
  export * from "./jina_clip/image_processing_jina_clip.js";
14
15
  export * from "./llava_onevision/image_processing_llava_onevision.js";
@@ -1,7 +1,7 @@
1
1
  export class VLMImageProcessor extends ImageProcessor {
2
2
  constructor(config: any);
3
3
  constant_values: any;
4
- pad_image(pixelData: any, imgDims: any, padSize: any, options: any): [Float32Array, number[]];
4
+ pad_image(pixelData: any, imgDims: any, padSize: any, options: any): [Float32Array<ArrayBufferLike>, number[]];
5
5
  }
6
6
  import { ImageProcessor } from "../../base/image_processors_utils.js";
7
7
  //# sourceMappingURL=image_processing_janus.d.ts.map
@@ -1 +1 @@
1
- {"version":3,"file":"image_processing_janus.d.ts","sourceRoot":"","sources":["../../../src/models/janus/image_processing_janus.js"],"names":[],"mappings":"AAKA;IACI,yBAUC;IADG,qBAAqF;IAGzF,8FAMC;CACJ;+BAtBM,sCAAsC"}
1
+ {"version":3,"file":"image_processing_janus.d.ts","sourceRoot":"","sources":["../../../src/models/janus/image_processing_janus.js"],"names":[],"mappings":"AAKA;IACI,yBAUC;IADG,qBAAqF;IAGzF,+GAMC;CACJ;+BAtBM,sCAAsC"}
@@ -1 +1 @@
1
- {"version":3,"file":"processing_janus.d.ts","sourceRoot":"","sources":["../../../src/models/janus/processing_janus.js"],"names":[],"mappings":"AAQA;IACI,wDAAiD;IACjD,6CAAsC;IAGtC,0CAOC;IAJG,eAAsC;IACtC,qBAAkD;IAClD,mBAA8C;IAC9C,sBAAoD;IAGxD;;;;OAIG;IAEH;;;;;;OAMG;IAEH;;;;;;OAMG;IACH;;;;iBAnBc,CAAC,QAAQ,GAAG,MAAM,GAAG,GAAG,CAAC,EAAE;;QAeA,MAAM,GAApC,QAAQ,GAAC,QAAQ,EAAE;QACF,aAAa,GAA9B,MAAM;;;;;mBAVH,MAAM;;;;wBACN,MAAM;;;;yBACN,MAAM;;;;yBACN,MAAM;;;;;mBAHN,MAAM;;;;wBACN,MAAM;;;;yBACN,MAAM;;;;yBACN,MAAM;8EAwFnB;CACJ;0BAzHyB,gCAAgC;yBAKjC,sBAAsB;uBADxB,uBAAuB;mCAHX,kCAAkC;8BACvC,qBAAqB"}
1
+ {"version":3,"file":"processing_janus.d.ts","sourceRoot":"","sources":["../../../src/models/janus/processing_janus.js"],"names":[],"mappings":"AAQA;IACI,wDAAiD;IACjD,6CAAsC;IAGtC,0CAOC;IAJG,eAAsC;IACtC,qBAAkD;IAClD,mBAA8C;IAC9C,sBAAoD;IAGxD;;;;OAIG;IAEH;;;;;;OAMG;IAEH;;;;;;OAMG;IACH;;;;iBAnBc,CAAC,QAAQ,GAAG,MAAM,GAAG,GAAG,CAAC,EAAE;uCAetC;QAAsC,MAAM,GAApC,QAAQ,GAAC,QAAQ,EAAE;QACF,aAAa,GAA9B,MAAM;KACd,GAAU,OAAO,CAAC;;;;mBAXP,MAAM;;;;wBACN,MAAM;;;;yBACN,MAAM;;;;yBACN,MAAM;SAQ0B;;;;mBAXhC,MAAM;;;;wBACN,MAAM;;;;yBACN,MAAM;;;;yBACN,MAAM;QAQmD,OAAO,sCAAsC,EAAE,oBAAoB,CAAA,CAAC,CAgF1I;CACJ;0BAzHyB,gCAAgC;yBAKjC,sBAAsB;uBADxB,uBAAuB;mCAHX,kCAAkC;8BACvC,qBAAqB"}
@@ -1,20 +1,20 @@
1
1
  export class MaskFormerImageProcessor extends ImageProcessor {
2
- post_process_panoptic_segmentation(outputs: any, threshold?: number, mask_threshold?: number, overlap_mask_area_threshold?: number, label_ids_to_fuse?: Set<number>, target_sizes?: [number, number][]): {
2
+ post_process_panoptic_segmentation(outputs: any, threshold?: number, mask_threshold?: number, overlap_mask_area_threshold?: number, label_ids_to_fuse?: Set<number>, target_sizes?: [number, number][]): Array<{
3
3
  segmentation: import("../../transformers.js").Tensor;
4
- segments_info: {
4
+ segments_info: Array<{
5
5
  id: number;
6
6
  label_id: number;
7
7
  score: number;
8
- }[];
9
- }[];
10
- post_process_instance_segmentation(outputs: any, threshold?: number, target_sizes?: [number, number][]): {
8
+ }>;
9
+ }>;
10
+ post_process_instance_segmentation(outputs: any, threshold?: number, target_sizes?: [number, number][]): Array<{
11
11
  segmentation: import("../../transformers.js").Tensor;
12
- segments_info: {
12
+ segments_info: Array<{
13
13
  id: number;
14
14
  label_id: number;
15
15
  score: number;
16
- }[];
17
- }[];
16
+ }>;
17
+ }>;
18
18
  }
19
19
  export class MaskFormerFeatureExtractor extends MaskFormerImageProcessor {
20
20
  }
@@ -1 +1 @@
1
- {"version":3,"file":"image_processing_maskformer.d.ts","sourceRoot":"","sources":["../../../src/models/maskformer/image_processing_maskformer.js"],"names":[],"mappings":"AAMA;IAYure;;;;;;;QAA4qG;IAAA;;;;;;;QAAktB;CAFpjmB;AACD;CAA4E;+BAbrE,sCAAsC"}
1
+ {"version":3,"file":"image_processing_maskformer.d.ts","sourceRoot":"","sources":["../../../src/models/maskformer/image_processing_maskformer.js"],"names":[],"mappings":"AAMA;;;;;;;;;;;;;;;;;CAUC;AACD;CAA4E;+BAbrE,sCAAsC"}
@@ -19,7 +19,7 @@ export class MgpstrProcessor extends Processor {
19
19
  * @param {string} format Type of model prediction. Must be one of ['char', 'bpe', 'wp'].
20
20
  * @returns {[string[], number[]]} The decoded sentences and their confidence scores.
21
21
  */
22
- _decode_helper(pred_logits: import('../../utils/tensor.js').Tensor, format: string): [string[], number[]];
22
+ _decode_helper(pred_logits: import("../../utils/tensor.js").Tensor, format: string): [string[], number[]];
23
23
  /**
24
24
  * Convert a list of lists of char token ids into a list of strings by calling char tokenizer.
25
25
  * @param {number[][]} sequences List of tokenized input ids.
@@ -49,7 +49,7 @@ export class MgpstrProcessor extends Processor {
49
49
  * - bpe_preds: The list of BPE decoded sentences.
50
50
  * - wp_preds: The list of wp decoded sentences.
51
51
  */
52
- batch_decode([char_logits, bpe_logits, wp_logits]: import('../../utils/tensor.js').Tensor[]): {
52
+ batch_decode([char_logits, bpe_logits, wp_logits]: import("../../utils/tensor.js").Tensor[]): {
53
53
  generated_text: string[];
54
54
  scores: number[];
55
55
  char_preds: string[];
@@ -1 +1 @@
1
- {"version":3,"file":"processing_mgp_str.d.ts","sourceRoot":"","sources":["../../../src/models/mgp_str/processing_mgp_str.js"],"names":[],"mappings":"AAUA;IACI,6CAAsC;IACtC,wDAAiD;IAEjD;;OAEG;IACH,oEAEC;IAED;;OAEG;IACH,iEAEC;IAED;;OAEG;IACH,gEAEC;IAED;;;;;OAKG;IACH,4BAJW,OAAO,uBAAuB,EAAE,MAAM,UACtC,MAAM,GACJ,CAAC,MAAM,EAAE,EAAE,MAAM,EAAE,CAAC,CA0ChC;IAED;;;;OAIG;IACH,uBAHW,MAAM,EAAE,EAAE,GACR,MAAM,EAAE,CAIpB;IAED;;;;OAIG;IACH,sBAHW,MAAM,EAAE,EAAE,GACR,MAAM,EAAE,CAIpB;IAED;;;;OAIG;IACH,qBAHW,MAAM,EAAE,EAAE,GACR,MAAM,EAAE,CAIpB;IAED;;;;;;;;;;OAUG;IACH,mDATW,OAAO,uBAAuB,EAAE,MAAM,EAAE;wBACrB,MAAM,EAAE;gBAAU,MAAM,EAAE;oBAAc,MAAM,EAAE;mBAAa,MAAM,EAAE;kBAAY,MAAM,EAAE;MA4BtH;IAmBD,6CAQC;CACJ;0BAzKyB,gCAAgC;8BAE5B,qBAAqB;mCADhB,kCAAkC"}
1
+ {"version":3,"file":"processing_mgp_str.d.ts","sourceRoot":"","sources":["../../../src/models/mgp_str/processing_mgp_str.js"],"names":[],"mappings":"AAUA;IACI,6CAAsC;IACtC,wDAAiD;IAEjD;;OAEG;IACH,sBAFa,OAAO,qBAAqB,EAAE,eAAe,CAIzD;IAED;;OAEG;IACH,qBAFa,OAAO,qBAAqB,EAAE,aAAa,CAIvD;IAED;;OAEG;IACH,oBAFa,OAAO,qBAAqB,EAAE,aAAa,CAIvD;IAED;;;;;OAKG;IACH,4BAJW,OAAO,uBAAuB,EAAE,MAAM,UACtC,MAAM,GACJ,CAAC,MAAM,EAAE,EAAE,MAAM,EAAE,CAAC,CA0ChC;IAED;;;;OAIG;IACH,uBAHW,MAAM,EAAE,EAAE,GACR,MAAM,EAAE,CAIpB;IAED;;;;OAIG;IACH,sBAHW,MAAM,EAAE,EAAE,GACR,MAAM,EAAE,CAIpB;IAED;;;;OAIG;IACH,qBAHW,MAAM,EAAE,EAAE,GACR,MAAM,EAAE,CAIpB;IAED;;;;;;;;;;OAUG;IACH,mDATW,OAAO,uBAAuB,EAAE,MAAM,EAAE,GACtC;QAAC,cAAc,EAAE,MAAM,EAAE,CAAC;QAAC,MAAM,EAAE,MAAM,EAAE,CAAC;QAAC,UAAU,EAAE,MAAM,EAAE,CAAC;QAAC,SAAS,EAAE,MAAM,EAAE,CAAC;QAAC,QAAQ,EAAE,MAAM,EAAE,CAAA;KAAC,CA4BvH;IAmBD,6CAQC;CACJ;0BAzKyB,gCAAgC;8BAE5B,qBAAqB;mCADhB,kCAAkC"}
@@ -1 +1 @@
1
- {"version":3,"file":"image_processing_owlvit.d.ts","sourceRoot":"","sources":["../../../src/models/owlvit/image_processing_owlvit.js"],"names":[],"mappings":"AAKA;IAO4tE;;;6FAAu6F;CAFloK;AACD;CAAoE;+BAR7D,sCAAsC"}
1
+ {"version":3,"file":"image_processing_owlvit.d.ts","sourceRoot":"","sources":["../../../src/models/owlvit/image_processing_owlvit.js"],"names":[],"mappings":"AAKA;;;;;CAKC;AACD;CAAoE;+BAR7D,sCAAsC"}
@@ -0,0 +1,12 @@
1
+ export class PaliGemmaProcessor extends Processor {
2
+ static tokenizer_class: typeof AutoTokenizer;
3
+ static image_processor_class: typeof AutoImageProcessor;
4
+ /**
5
+ * @typedef {import('../../utils/image.js').RawImage} RawImage
6
+ */
7
+ _call(images: import("../../utils/image.js").RawImage | import("../../utils/image.js").RawImage[], text?: any, kwargs?: {}): Promise<any>;
8
+ }
9
+ import { Processor } from "../../base/processing_utils.js";
10
+ import { AutoTokenizer } from "../../tokenizers.js";
11
+ import { AutoImageProcessor } from "../auto/image_processing_auto.js";
12
+ //# sourceMappingURL=processing_paligemma.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"processing_paligemma.d.ts","sourceRoot":"","sources":["../../../src/models/paligemma/processing_paligemma.js"],"names":[],"mappings":"AAgBA;IACI,6CAAsC;IACtC,wDAAiD;IAGjD;;OAEG;IAGH,cAAuB,0CAAS,yCAAU,yCAsDzC;CACJ;0BAjFyB,gCAAgC;8BAE5B,qBAAqB;mCADhB,kCAAkC"}
@@ -1,8 +1,10 @@
1
1
  export * from "./florence2/processing_florence2.js";
2
2
  export * from "./mgp_str/processing_mgp_str.js";
3
+ export * from "./idefics3/processing_idefics3.js";
3
4
  export * from "./janus/processing_janus.js";
4
5
  export * from "./jina_clip/processing_jina_clip.js";
5
6
  export * from "./owlvit/processing_owlvit.js";
7
+ export * from "./paligemma/processing_paligemma.js";
6
8
  export * from "./pyannote/processing_pyannote.js";
7
9
  export * from "./qwen2_vl/processing_qwen2_vl.js";
8
10
  export * from "./sam/processing_sam.js";
@@ -1 +1 @@
1
- {"version":3,"file":"feature_extraction_pyannote.d.ts","sourceRoot":"","sources":["../../../src/models/pyannote/feature_extraction_pyannote.js"],"names":[],"mappings":"AAIA;IACI;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY,GACvB,QAAQ;QAAE,YAAY,EAAE,MAAM,CAAC;KAAE,CAAC,CAiB9C;CAEJ;iCA3BuD,wCAAwC;uBACzE,uBAAuB"}
1
+ {"version":3,"file":"feature_extraction_pyannote.d.ts","sourceRoot":"","sources":["../../../src/models/pyannote/feature_extraction_pyannote.js"],"names":[],"mappings":"AAIA;IACI;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY,GACvB,OAAO,CAAC;QAAE,YAAY,EAAE,MAAM,CAAC;KAAE,CAAC,CAiB9C;CAEJ;iCA3BuD,wCAAwC;uBACzE,uBAAuB"}
@@ -18,7 +18,7 @@ export class PyAnnoteProcessor extends Processor {
18
18
  * @param {number} num_samples Number of samples in the input audio.
19
19
  * @returns {Array<Array<{ id: number, start: number, end: number, confidence: number }>>} The post-processed speaker diarization results.
20
20
  */
21
- post_process_speaker_diarization(logits: import('../../utils/tensor.js').Tensor, num_samples: number): Array<Array<{
21
+ post_process_speaker_diarization(logits: import("../../utils/tensor.js").Tensor, num_samples: number): Array<Array<{
22
22
  id: number;
23
23
  start: number;
24
24
  end: number;
@@ -1 +1 @@
1
- {"version":3,"file":"processing_pyannote.d.ts","sourceRoot":"","sources":["../../../src/models/pyannote/processing_pyannote.js"],"names":[],"mappings":"AAIA;IACI,4DAAqD;IAErD;;;;OAIG;IACH,aAHW,GAAG,GACD,QAAQ,GAAG,CAAC,CAIxB;IAED;;;;OAIG;IACH,2BAHW,MAAM,GACJ,MAAM,CAIlB;IAED;;;;;OAKG;IACH,yCAJW,OAAO,uBAAuB,EAAE,MAAM,eACtC,MAAM,GACJ,MAAM,MAAM;QAAE,EAAE,EAAE,MAAM,CAAC;QAAC,KAAK,EAAE,MAAM,CAAC;QAAC,GAAG,EAAE,MAAM,CAAC;QAAC,UAAU,EAAE,MAAM,CAAA;KAAE,CAAC,CAAC,CAwCxF;CACJ;0BAtEyB,gCAAgC;qCACrB,oCAAoC"}
1
+ {"version":3,"file":"processing_pyannote.d.ts","sourceRoot":"","sources":["../../../src/models/pyannote/processing_pyannote.js"],"names":[],"mappings":"AAIA;IACI,4DAAqD;IAErD;;;;OAIG;IACH,aAHW,GAAG,GACD,OAAO,CAAC,GAAG,CAAC,CAIxB;IAED;;;;OAIG;IACH,2BAHW,MAAM,GACJ,MAAM,CAIlB;IAED;;;;;OAKG;IACH,yCAJW,OAAO,uBAAuB,EAAE,MAAM,eACtC,MAAM,GACJ,KAAK,CAAC,KAAK,CAAC;QAAE,EAAE,EAAE,MAAM,CAAC;QAAC,KAAK,EAAE,MAAM,CAAC;QAAC,GAAG,EAAE,MAAM,CAAC;QAAC,UAAU,EAAE,MAAM,CAAA;KAAE,CAAC,CAAC,CAwCxF;CACJ;0BAtEyB,gCAAgC;qCACrB,oCAAoC"}
@@ -1 +1 @@
1
- {"version":3,"file":"processing_qwen2_vl.d.ts","sourceRoot":"","sources":["../../../src/models/qwen2_vl/processing_qwen2_vl.js"],"names":[],"mappings":"AAKA;IACI,wDAAiD;IACjD,6CAAsC;IAEtC;;;;;;OAMG;IACH,YALW,MAAM,GAAC,MAAM,EAAE,WACf,QAAQ,GAAC,QAAQ,EAAE,WACf,GAAG,KACL,QAAQ,GAAG,CAAC,CAoCxB;CACJ;0BAnDyB,gCAAgC;yBAGjC,sBAAsB;mCAFZ,kCAAkC;8BACvC,qBAAqB"}
1
+ {"version":3,"file":"processing_qwen2_vl.d.ts","sourceRoot":"","sources":["../../../src/models/qwen2_vl/processing_qwen2_vl.js"],"names":[],"mappings":"AAKA;IACI,wDAAiD;IACjD,6CAAsC;IAEtC;;;;;;OAMG;IACH,YALW,MAAM,GAAC,MAAM,EAAE,WACf,QAAQ,GAAC,QAAQ,EAAE,WACf,GAAG,EAAA,GACL,OAAO,CAAC,GAAG,CAAC,CAoCxB;CACJ;0BAnDyB,gCAAgC;yBAGjC,sBAAsB;mCAFZ,kCAAkC;8BACvC,qBAAqB"}
@@ -1 +1 @@
1
- {"version":3,"file":"image_processing_rt_detr.d.ts","sourceRoot":"","sources":["../../../src/models/rt_detr/image_processing_rt_detr.js"],"names":[],"mappings":"AAMA;IAMgyE;;;6FAAu6F;CADtsK;+BARM,sCAAsC"}
1
+ {"version":3,"file":"image_processing_rt_detr.d.ts","sourceRoot":"","sources":["../../../src/models/rt_detr/image_processing_rt_detr.js"],"names":[],"mappings":"AAMA;;;;;CAKC;+BARM,sCAAsC"}
@@ -1 +1 @@
1
- {"version":3,"file":"image_processing_sam.d.ts","sourceRoot":"","sources":["../../../src/models/sam/image_processing_sam.js"],"names":[],"mappings":"AAWA;;;;;;;;GAQG;AAEH;IAEI;;;;;;OAMG;IACH,mCALW,GAAG,kBACH,OAAO,sCAAsC,EAAE,WAAW,EAAE,wBAC5D,OAAO,sCAAsC,EAAE,WAAW,EAAE,8BAC1D,MAAM,CA4ClB;IAED;;;;;OAKG;IACH,+BAJW,GAAG,gBACH,MAAM,GACJ,MAAM,CAoBlB;IACD;;;;;;;;;;;;;;;;;;;OAmBG;IACH,cAnBW,GAAG,EAAE;QAES,YAAY,GAA1B,GAAG;QAGW,YAAY,GAA1B,GAAG;QAGoB,WAAW,GAAlC,MAAM,EAAE,EAAE,EAAE;QASV,QAAQ,uBAAuB,CAAC,CA+B5C;IAED;;;;;;;;;;;;OAYG;IACH,0BAXW,MAAM,kBACN,CAAC,MAAM,EAAE,MAAM,CAAC,EAAE,wBAClB,CAAC,MAAM,EAAE,MAAM,CAAC,EAAE;QAED,cAAc,GAA/B,MAAM;QACY,QAAQ,GAA1B,OAAO;QACU,QAAQ;YACC,MAAM,GAAhC,MAAM;YACoB,KAAK,GAA/B,MAAM;;QACJ,QAAQ,MAAM,EAAE,CAAC,CAsD7B;IAED;;;;;;;;;;;;;OAaG;IACH,2BAZW,OAAO,sBAAsB,EAAE,QAAQ,eACvC,MAAM;QAEW,aAAa,GAA9B,MAAM;QAEW,aAAa,GAA9B,MAAM;QAEW,eAAe,GAAhC,MAAM;QACW,8BAA8B,GAA/C,MAAM;YAYhB;CACJ;;kBAnOa,MAAM;oBACN,OAAO,sCAAsC,EAAE,WAAW,EAAE;0BAC5D,OAAO,sCAAsC,EAAE,WAAW,EAAE;mBAC5D,MAAM;mBACN,MAAM;kBACN,MAAM;;+BAhBb,sCAAsC;uBAMtC,uBAAuB"}
1
+ {"version":3,"file":"image_processing_sam.d.ts","sourceRoot":"","sources":["../../../src/models/sam/image_processing_sam.js"],"names":[],"mappings":"AAWA;;;;;;;;GAQG;AAEH;IAEI;;;;;;OAMG;IACH,mCALW,GAAG,kBACH,OAAO,sCAAsC,EAAE,WAAW,EAAE,wBAC5D,OAAO,sCAAsC,EAAE,WAAW,EAAE,8BAC1D,MAAM,CA4ClB;IAED;;;;;OAKG;IACH,+BAJW,GAAG,gBACH,MAAM,GACJ,MAAM,CAoBlB;IACD;;;;;;;;;;;;;;;;;;;OAmBG;IACH,cAnBW,GAAG,EAAE,gDAEb;QAAsB,YAAY,GAA1B,GAAG;QAGW,YAAY,GAA1B,GAAG;QAGoB,WAAW,GAAlC,MAAM,EAAE,EAAE,EAAE;KASpB,GAAU,OAAO,CAAC,uBAAuB,CAAC,CA+B5C;IAED;;;;;;;;;;;;OAYG;IACH,0BAXW,MAAM,kBACN,CAAC,MAAM,EAAE,MAAM,CAAC,EAAE,wBAClB,CAAC,MAAM,EAAE,MAAM,CAAC,EAAE,4CAE1B;QAAyB,cAAc,GAA/B,MAAM;QACY,QAAQ,GAA1B,OAAO;QACU,QAAQ,GACjC;YAAkC,MAAM,GAAhC,MAAM;YACoB,KAAK,GAA/B,MAAM;SACd;KAAA,GAAU,OAAO,CAAC,MAAM,EAAE,CAAC,CAsD7B;IAED;;;;;;;;;;;;;OAaG;IACH,2BAZW,OAAO,sBAAsB,EAAE,QAAQ,eACvC,MAAM,uFAEd;QAAyB,aAAa,GAA9B,MAAM;QAEW,aAAa,GAA9B,MAAM;QAEW,eAAe,GAAhC,MAAM;QACW,8BAA8B,GAA/C,MAAM;KAEd,OAUF;CACJ;;kBAnOa,MAAM;oBACN,OAAO,sCAAsC,EAAE,WAAW,EAAE;0BAC5D,OAAO,sCAAsC,EAAE,WAAW,EAAE;mBAC5D,MAAM;mBACN,MAAM;kBACN,MAAM;;+BAhBb,sCAAsC;uBAMtC,uBAAuB"}
@@ -1,7 +1,7 @@
1
1
  export class SeamlessM4TFeatureExtractor extends FeatureExtractor {
2
2
  constructor(config: any);
3
3
  mel_filters: number[][];
4
- window: Float64Array;
4
+ window: Float64Array<ArrayBufferLike>;
5
5
  /**
6
6
  * Computes the log-Mel spectrogram of the provided audio waveform.
7
7
  * @param {Float32Array|Float64Array} waveform The audio waveform to process.
@@ -1 +1 @@
1
- {"version":3,"file":"feature_extraction_seamless_m4t.d.ts","sourceRoot":"","sources":["../../../src/models/seamless_m4t/feature_extraction_seamless_m4t.js"],"names":[],"mappings":"AAIA;IAEI,yBAwBC;IALG,wBAA8B;IAE9B,qBAEE;IAGN;;;;;OAKG;IACH,kCAJW,YAAY,GAAC,YAAY,cACzB,MAAM,GACJ,QAAQ,MAAM,CAAC,CA6B3B;IAED;;;;;;;;;OASG;IACH,aARW,YAAY,GAAC,YAAY;QAEP,OAAO,GAAzB,OAAO;QACU,kBAAkB,GAAnC,MAAM;QACY,yBAAyB,GAA3C,OAAO;QACW,qBAAqB,GAAvC,OAAO;;wBACqB,MAAM;yBAAmB,MAAM;OAuGrE;CACJ;iCAnLuD,wCAAwC;uBACzE,uBAAuB"}
1
+ {"version":3,"file":"feature_extraction_seamless_m4t.d.ts","sourceRoot":"","sources":["../../../src/models/seamless_m4t/feature_extraction_seamless_m4t.js"],"names":[],"mappings":"AAIA;IAEI,yBAwBC;IALG,wBAA8B;IAE9B,sCAEE;IAGN;;;;;OAKG;IACH,kCAJW,YAAY,GAAC,YAAY,cACzB,MAAM,GACJ,OAAO,CAAC,MAAM,CAAC,CA6B3B;IAED;;;;;;;;;OASG;IACH,aARW,YAAY,GAAC,YAAY,uFAEjC;QAA0B,OAAO,GAAzB,OAAO;QACU,kBAAkB,GAAnC,MAAM;QACY,yBAAyB,GAA3C,OAAO;QACW,qBAAqB,GAAvC,OAAO;KACf,GAAU,OAAO,CAAC;QAAE,cAAc,EAAE,MAAM,CAAC;QAAC,cAAc,CAAC,EAAE,MAAM,CAAA;KAAE,CAAC,CAuGxE;CACJ;iCAnLuD,wCAAwC;uBACzE,uBAAuB"}
@@ -1 +1 @@
1
- {"version":3,"file":"image_processing_segformer.d.ts","sourceRoot":"","sources":["../../../src/models/segformer/image_processing_segformer.js"],"names":[],"mappings":"AAMA;IAOqmK;;;QAA21E;CAF/7O;AACD;CAA0E;+BATnE,sCAAsC"}
1
+ {"version":3,"file":"image_processing_segformer.d.ts","sourceRoot":"","sources":["../../../src/models/segformer/image_processing_segformer.js"],"names":[],"mappings":"AAMA;;;;;CAKC;AACD;CAA0E;+BATnE,sCAAsC"}
@@ -1 +1 @@
1
- {"version":3,"file":"processing_speecht5.d.ts","sourceRoot":"","sources":["../../../src/models/speecht5/processing_speecht5.js"],"names":[],"mappings":"AAIA;IACI,6CAAsC;IACtC,4DAAqD;IAErD;;;;OAIG;IACH,aAHW,GAAG,GACD,QAAQ,GAAG,CAAC,CAIxB;CACJ;0BAhByB,gCAAgC;8BAC5B,qBAAqB;qCACd,oCAAoC"}
1
+ {"version":3,"file":"processing_speecht5.d.ts","sourceRoot":"","sources":["../../../src/models/speecht5/processing_speecht5.js"],"names":[],"mappings":"AAIA;IACI,6CAAsC;IACtC,4DAAqD;IAErD;;;;OAIG;IACH,aAHW,GAAG,GACD,OAAO,CAAC,GAAG,CAAC,CAIxB;CACJ;0BAhByB,gCAAgC;8BAC5B,qBAAqB;qCACd,oCAAoC"}
@@ -1,5 +1,5 @@
1
1
  export class Swin2SRImageProcessor extends ImageProcessor {
2
- pad_image(pixelData: any, imgDims: any, padSize: any, options?: {}): [Float32Array, number[]];
2
+ pad_image(pixelData: any, imgDims: any, padSize: any, options?: {}): [Float32Array<ArrayBufferLike>, number[]];
3
3
  }
4
4
  import { ImageProcessor } from "../../base/image_processors_utils.js";
5
5
  //# sourceMappingURL=image_processing_swin2sr.d.ts.map
@@ -1 +1 @@
1
- {"version":3,"file":"image_processing_swin2sr.d.ts","sourceRoot":"","sources":["../../../src/models/swin2sr/image_processing_swin2sr.js"],"names":[],"mappings":"AAIA;IACI,8FAiBC;CACJ;+BArBM,sCAAsC"}
1
+ {"version":3,"file":"image_processing_swin2sr.d.ts","sourceRoot":"","sources":["../../../src/models/swin2sr/image_processing_swin2sr.js"],"names":[],"mappings":"AAIA;IACI,+GAiBC;CACJ;+BArBM,sCAAsC"}
@@ -1 +1 @@
1
- {"version":3,"file":"image_processing_vitmatte.d.ts","sourceRoot":"","sources":["../../../src/models/vitmatte/image_processing_vitmatte.js"],"names":[],"mappings":"AASA;IACI;;;;;;OAMG;IACH,cAJW,OAAO,sBAAsB,EAAE,QAAQ,EAAE,WACzC,OAAO,sBAAsB,EAAE,QAAQ,EAAE,GACvC,QAAQ,OAAO,sCAAsC,EAAE,oBAAoB,CAAC,CAiCxF;CACJ;+BA/CM,sCAAsC"}
1
+ {"version":3,"file":"image_processing_vitmatte.d.ts","sourceRoot":"","sources":["../../../src/models/vitmatte/image_processing_vitmatte.js"],"names":[],"mappings":"AASA;IACI;;;;;;OAMG;IACH,cAJW,OAAO,sBAAsB,EAAE,QAAQ,EAAE,WACzC,OAAO,sBAAsB,EAAE,QAAQ,EAAE,GACvC,OAAO,CAAC,OAAO,sCAAsC,EAAE,oBAAoB,CAAC,CAiCxF;CACJ;+BA/CM,sCAAsC"}
@@ -13,7 +13,7 @@ export class VitPoseImageProcessor extends ImageProcessor {
13
13
  * keypoints: [number, number][]
14
14
  * }[][]} List of keypoints predictions for each image.
15
15
  */
16
- post_process_pose_estimation(outputs: import('../../utils/tensor.js').Tensor, boxes: [number, number, number, number][][], { threshold, }?: {
16
+ post_process_pose_estimation(outputs: import("../../utils/tensor.js").Tensor, boxes: [number, number, number, number][][], { threshold, }?: {
17
17
  threshold?: any;
18
18
  }): {
19
19
  bbox: [number, number, number, number];
@@ -1 +1 @@
1
- {"version":3,"file":"image_processing_vitpose.d.ts","sourceRoot":"","sources":["../../../src/models/vitpose/image_processing_vitpose.js"],"names":[],"mappings":"AAIA;IAEI;;;;;;;;;;;;;OAaG;IACH,sCAVW,OAAO,uBAAuB,EAAE,MAAM,SACtC,CAAC,MAAM,EAAE,MAAM,EAAE,MAAM,EAAE,MAAM,CAAC,EAAE,EAAE;;QAElC;QACZ,IAAQ,EAAE,CAAC,MAAM,EAAE,MAAM,EAAE,MAAM,EAAE,MAAM,CAAC,CAAC;QAC3C,MAAU,EAAE,MAAM,EAAE,CAAC;QACrB,MAAU,EAAE,MAAM,EAAE,CAAC;QACrB,SAAa,EAAE,CAAC,MAAM,EAAE,MAAM,CAAC,EAAE,CAAA;KAC9B,EAAE,EAAE,CAqEP;CACJ;+BAtFM,sCAAsC"}
1
+ {"version":3,"file":"image_processing_vitpose.d.ts","sourceRoot":"","sources":["../../../src/models/vitpose/image_processing_vitpose.js"],"names":[],"mappings":"AAIA;IAEI;;;;;;;;;;;;;OAaG;IACH,sCAVW,OAAO,uBAAuB,EAAE,MAAM,SACtC,CAAC,MAAM,EAAE,MAAM,EAAE,MAAM,EAAE,MAAM,CAAC,EAAE,EAAE;;QAElC;QACR,IAAI,EAAE,CAAC,MAAM,EAAE,MAAM,EAAE,MAAM,EAAE,MAAM,CAAC,CAAC;QACvC,MAAM,EAAE,MAAM,EAAE,CAAC;QACjB,MAAM,EAAE,MAAM,EAAE,CAAC;QACjB,SAAS,EAAE,CAAC,MAAM,EAAE,MAAM,CAAC,EAAE,CAAA;KAC9B,EAAE,EAAE,CAqEP;CACJ;+BAtFM,sCAAsC"}
@@ -1 +1 @@
1
- {"version":3,"file":"feature_extraction_wav2vec2.d.ts","sourceRoot":"","sources":["../../../src/models/wav2vec2/feature_extraction_wav2vec2.js"],"names":[],"mappings":"AAGA;IAEI;;;OAGG;IACH,uCAHW,YAAY,GACV,YAAY,CAQxB;IAED;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY;sBACC,MAAM;wBAAkB,MAAM;OAsBlE;CACJ;iCA3CuD,wCAAwC;uBACzE,uBAAuB"}
1
+ {"version":3,"file":"feature_extraction_wav2vec2.d.ts","sourceRoot":"","sources":["../../../src/models/wav2vec2/feature_extraction_wav2vec2.js"],"names":[],"mappings":"AAGA;IAEI;;;OAGG;IACH,uCAHW,YAAY,GACV,YAAY,CAQxB;IAED;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY,GACvB,OAAO,CAAC;QAAE,YAAY,EAAE,MAAM,CAAC;QAAC,cAAc,EAAE,MAAM,CAAA;KAAE,CAAC,CAsBrE;CACJ;iCA3CuD,wCAAwC;uBACzE,uBAAuB"}
@@ -1 +1 @@
1
- {"version":3,"file":"processing_wav2vec2.d.ts","sourceRoot":"","sources":["../../../src/models/wav2vec2/processing_wav2vec2.js"],"names":[],"mappings":"AAGA;IACI,4DAAqD;IAErD;;;;OAIG;IACH,aAHW,GAAG,GACD,QAAQ,GAAG,CAAC,CAIxB;CACJ;0BAdyB,gCAAgC;qCACrB,oCAAoC"}
1
+ {"version":3,"file":"processing_wav2vec2.d.ts","sourceRoot":"","sources":["../../../src/models/wav2vec2/processing_wav2vec2.js"],"names":[],"mappings":"AAGA;IACI,4DAAqD;IAErD;;;;OAIG;IACH,aAHW,GAAG,GACD,OAAO,CAAC,GAAG,CAAC,CAIxB;CACJ;0BAdyB,gCAAgC;qCACrB,oCAAoC"}
@@ -1,7 +1,7 @@
1
1
  export class WeSpeakerFeatureExtractor extends FeatureExtractor {
2
2
  constructor(config: any);
3
3
  mel_filters: number[][];
4
- window: Float64Array;
4
+ window: Float64Array<ArrayBufferLike>;
5
5
  min_num_frames: any;
6
6
  /**
7
7
  * Computes the log-Mel spectrogram of the provided audio waveform.
@@ -1 +1 @@
1
- {"version":3,"file":"feature_extraction_wespeaker.d.ts","sourceRoot":"","sources":["../../../src/models/wespeaker/feature_extraction_wespeaker.js"],"names":[],"mappings":"AAKA;IAEI,yBAyBC;IANG,wBAA8B;IAE9B,qBAEE;IACF,oBAAgD;IAGpD;;;;OAIG;IACH,kCAHW,YAAY,GAAC,YAAY,GACvB,QAAQ,MAAM,CAAC,CA2B3B;IAGD;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY,GACvB,QAAQ;QAAE,cAAc,EAAE,MAAM,CAAA;KAAE,CAAC,CA4B/C;CACJ;iCAnGuD,wCAAwC;uBACzE,uBAAuB"}
1
+ {"version":3,"file":"feature_extraction_wespeaker.d.ts","sourceRoot":"","sources":["../../../src/models/wespeaker/feature_extraction_wespeaker.js"],"names":[],"mappings":"AAKA;IAEI,yBAyBC;IANG,wBAA8B;IAE9B,sCAEE;IACF,oBAAgD;IAGpD;;;;OAIG;IACH,kCAHW,YAAY,GAAC,YAAY,GACvB,OAAO,CAAC,MAAM,CAAC,CA2B3B;IAGD;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY,GACvB,OAAO,CAAC;QAAE,cAAc,EAAE,MAAM,CAAA;KAAE,CAAC,CA4B/C;CACJ;iCAnGuD,wCAAwC;uBACzE,uBAAuB"}
@@ -1,6 +1,6 @@
1
1
  export class WhisperFeatureExtractor extends FeatureExtractor {
2
2
  constructor(config: any);
3
- window: Float64Array;
3
+ window: Float64Array<ArrayBufferLike>;
4
4
  /**
5
5
  * Computes the log-Mel spectrogram of the provided audio waveform.
6
6
  * @param {Float32Array|Float64Array} waveform The audio waveform to process.
@@ -1 +1 @@
1
- {"version":3,"file":"feature_extraction_whisper.d.ts","sourceRoot":"","sources":["../../../src/models/whisper/feature_extraction_whisper.js"],"names":[],"mappings":"AAKA;IAEI,yBAeC;IADG,qBAAwD;IAG5D;;;;OAIG;IACH,kCAHW,YAAY,GAAC,YAAY,GACvB,QAAQ,MAAM,CAAC,CA0B3B;IAED;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY,GACvB,QAAQ;QAAE,cAAc,EAAE,MAAM,CAAA;KAAE,CAAC,CAwB/C;CACJ;iCAnFuD,wCAAwC;uBACzE,uBAAuB"}
1
+ {"version":3,"file":"feature_extraction_whisper.d.ts","sourceRoot":"","sources":["../../../src/models/whisper/feature_extraction_whisper.js"],"names":[],"mappings":"AAKA;IAEI,yBAeC;IADG,sCAAwD;IAG5D;;;;OAIG;IACH,kCAHW,YAAY,GAAC,YAAY,GACvB,OAAO,CAAC,MAAM,CAAC,CA0B3B;IAED;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY,GACvB,OAAO,CAAC;QAAE,cAAc,EAAE,MAAM,CAAA;KAAE,CAAC,CAwB/C;CACJ;iCAnFuD,wCAAwC;uBACzE,uBAAuB"}
@@ -1 +1 @@
1
- {"version":3,"file":"generation_whisper.d.ts","sourceRoot":"","sources":["../../../src/models/whisper/generation_whisper.js"],"names":[],"mappings":"AAEA;IAEI;;;OAGG;IACH,mBAFU,OAAO,CAEQ;IAEzB;;;;;OAKG;IACH,yBAFU,OAAO,CAEc;IAE/B;;;OAGG;IACH,YAFU,MAAM,CAEE;IAElB;;;;OAIG;IACH,iBAFU,CAAC,MAAM,EAAE,MAAM,CAAC,EAAE,CAEL;IAEvB;;;OAGG;IACH,MAFU,MAAM,CAEJ;IAEZ;;;;OAIG;IACH,UAFU,MAAM,CAEA;IAEhB;;;OAGG;IACH,wBAFU,MAAM,CAEc;IAE9B;;;;;;OAMG;IACH,YAFU,MAAM,EAAE,CAEA;IAElB;;;OAGG;IACH,iBAFU,OAAO,CAEM;IAEvB;;;;OAIG;IACH,YAFU,OAAO,MAAM,EAAE,MAAM,CAAC,GAAC,IAAI,CAEnB;IAElB;;;OAGG;IACH,YAFU,OAAO,MAAM,EAAE,MAAM,CAAC,GAAC,IAAI,CAEnB;IAElB;;;;OAIG;IACH,6BAFU,MAAM,CAEgB;CACnC;kDAGY,MAAwE;IAAC,iBAAiB,EAAE,uBAAuB,CAAA;CAAC,GAAG,uBAAuB;iCAvF1H,yCAAyC"}
1
+ {"version":3,"file":"generation_whisper.d.ts","sourceRoot":"","sources":["../../../src/models/whisper/generation_whisper.js"],"names":[],"mappings":"AAEA;IAEI;;;OAGG;IACH,mBAFU,OAAO,CAEQ;IAEzB;;;;;OAKG;IACH,yBAFU,OAAO,CAEc;IAE/B;;;OAGG;IACH,YAFU,MAAM,CAEE;IAElB;;;;OAIG;IACH,iBAFU,CAAC,MAAM,EAAE,MAAM,CAAC,EAAE,CAEL;IAEvB;;;OAGG;IACH,MAFU,MAAM,CAEJ;IAEZ;;;;OAIG;IACH,UAFU,MAAM,CAEA;IAEhB;;;OAGG;IACH,wBAFU,MAAM,CAEc;IAE9B;;;;;;OAMG;IACH,YAFU,MAAM,EAAE,CAEA;IAElB;;;OAGG;IACH,iBAFU,OAAO,CAEM;IAEvB;;;;OAIG;IACH,YAFU,MAAM,CAAC,MAAM,EAAE,MAAM,CAAC,GAAC,IAAI,CAEnB;IAElB;;;OAGG;IACH,YAFU,MAAM,CAAC,MAAM,EAAE,MAAM,CAAC,GAAC,IAAI,CAEnB;IAElB;;;;OAIG;IACH,6BAFU,MAAM,CAEgB;CACnC;kDAGY,GAAqE,GAAG;IAAC,iBAAiB,EAAE,uBAAuB,CAAA;CAAC,GAAG,uBAAuB;iCAvF1H,yCAAyC"}
@@ -1 +1 @@
1
- {"version":3,"file":"processing_whisper.d.ts","sourceRoot":"","sources":["../../../src/models/whisper/processing_whisper.js"],"names":[],"mappings":"AAIA;;GAEG;AACH;IACI,6CAAsC;IACtC,4DAAqD;IAErD;;;;OAIG;IACH,aAHW,GAAG,GACD,QAAQ,GAAG,CAAC,CAIxB;CACJ;0BAjByB,gCAAgC;8BAD5B,qBAAqB;qCADd,oCAAoC"}
1
+ {"version":3,"file":"processing_whisper.d.ts","sourceRoot":"","sources":["../../../src/models/whisper/processing_whisper.js"],"names":[],"mappings":"AAIA;;GAEG;AACH;IACI,6CAAsC;IACtC,4DAAqD;IAErD;;;;OAIG;IACH,aAHW,GAAG,GACD,OAAO,CAAC,GAAG,CAAC,CAIxB;CACJ;0BAjByB,gCAAgC;8BAD5B,qBAAqB;qCADd,oCAAoC"}
@@ -1 +1 @@
1
- {"version":3,"file":"image_processing_yolos.d.ts","sourceRoot":"","sources":["../../../src/models/yolos/image_processing_yolos.js"],"names":[],"mappings":"AAKA;IAO+tE;;;6FAAu6F;CAFroK;AACD;CAAkE;+BAR3D,sCAAsC"}
1
+ {"version":3,"file":"image_processing_yolos.d.ts","sourceRoot":"","sources":["../../../src/models/yolos/image_processing_yolos.js"],"names":[],"mappings":"AAKA;;;;;CAKC;AACD;CAAkE;+BAR3D,sCAAsC"}