@huggingface/transformers 3.1.1 → 3.2.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (144) hide show
  1. package/README.md +10 -4
  2. package/dist/ort-wasm-simd-threaded.jsep.wasm +0 -0
  3. package/dist/transformers.cjs +1062 -183
  4. package/dist/transformers.cjs.map +1 -1
  5. package/dist/transformers.js +2239 -1232
  6. package/dist/transformers.js.map +1 -1
  7. package/dist/transformers.min.cjs +1 -358
  8. package/dist/transformers.min.cjs.map +1 -1
  9. package/dist/transformers.min.js +1 -421
  10. package/dist/transformers.min.js.map +1 -1
  11. package/dist/transformers.min.mjs +1 -358
  12. package/dist/transformers.min.mjs.map +1 -1
  13. package/dist/transformers.mjs +1082 -181
  14. package/dist/transformers.mjs.map +1 -1
  15. package/package.json +11 -16
  16. package/src/backends/onnx.js +2 -7
  17. package/src/base/image_processors_utils.js +3 -1
  18. package/src/configs.js +11 -2
  19. package/src/env.js +1 -1
  20. package/src/models/feature_extractors.js +1 -0
  21. package/src/models/idefics3/image_processing_idefics3.js +24 -13
  22. package/src/models/image_processors.js +1 -0
  23. package/src/models/moonshine/feature_extraction_moonshine.js +26 -0
  24. package/src/models/moonshine/processing_moonshine.js +20 -0
  25. package/src/models/paligemma/processing_paligemma.js +82 -0
  26. package/src/models/phi3_v/image_processing_phi3_v.js +163 -0
  27. package/src/models/phi3_v/processing_phi3_v.js +53 -0
  28. package/src/models/processors.js +3 -0
  29. package/src/models/pyannote/feature_extraction_pyannote.js +56 -0
  30. package/src/models/pyannote/processing_pyannote.js +7 -54
  31. package/src/models.js +233 -35
  32. package/src/ops/registry.js +11 -0
  33. package/src/pipelines.js +30 -0
  34. package/src/tokenizers.js +12 -1
  35. package/src/utils/core.js +39 -9
  36. package/src/utils/hub.js +8 -12
  37. package/src/utils/image.js +40 -0
  38. package/src/utils/tensor.js +51 -1
  39. package/types/backends/onnx.d.ts +2 -2
  40. package/types/backends/onnx.d.ts.map +1 -1
  41. package/types/base/feature_extraction_utils.d.ts +1 -1
  42. package/types/base/feature_extraction_utils.d.ts.map +1 -1
  43. package/types/base/image_processors_utils.d.ts +4 -4
  44. package/types/base/image_processors_utils.d.ts.map +1 -1
  45. package/types/base/processing_utils.d.ts +4 -4
  46. package/types/base/processing_utils.d.ts.map +1 -1
  47. package/types/configs.d.ts +7 -7
  48. package/types/configs.d.ts.map +1 -1
  49. package/types/env.d.ts +1 -1
  50. package/types/env.d.ts.map +1 -1
  51. package/types/generation/configuration_utils.d.ts +2 -2
  52. package/types/generation/logits_process.d.ts +2 -2
  53. package/types/generation/logits_process.d.ts.map +1 -1
  54. package/types/generation/logits_sampler.d.ts.map +1 -1
  55. package/types/generation/parameters.d.ts +5 -5
  56. package/types/generation/stopping_criteria.d.ts +1 -1
  57. package/types/generation/stopping_criteria.d.ts.map +1 -1
  58. package/types/generation/streamers.d.ts +2 -2
  59. package/types/generation/streamers.d.ts.map +1 -1
  60. package/types/models/audio_spectrogram_transformer/feature_extraction_audio_spectrogram_transformer.d.ts +1 -1
  61. package/types/models/audio_spectrogram_transformer/feature_extraction_audio_spectrogram_transformer.d.ts.map +1 -1
  62. package/types/models/auto/feature_extraction_auto.d.ts.map +1 -1
  63. package/types/models/auto/image_processing_auto.d.ts.map +1 -1
  64. package/types/models/auto/processing_auto.d.ts +1 -1
  65. package/types/models/auto/processing_auto.d.ts.map +1 -1
  66. package/types/models/clap/feature_extraction_clap.d.ts +1 -1
  67. package/types/models/clap/feature_extraction_clap.d.ts.map +1 -1
  68. package/types/models/detr/image_processing_detr.d.ts +11 -11
  69. package/types/models/detr/image_processing_detr.d.ts.map +1 -1
  70. package/types/models/donut/image_processing_donut.d.ts +1 -1
  71. package/types/models/donut/image_processing_donut.d.ts.map +1 -1
  72. package/types/models/feature_extractors.d.ts +1 -0
  73. package/types/models/florence2/processing_florence2.d.ts.map +1 -1
  74. package/types/models/idefics3/image_processing_idefics3.d.ts.map +1 -1
  75. package/types/models/idefics3/processing_idefics3.d.ts.map +1 -1
  76. package/types/models/image_processors.d.ts +1 -0
  77. package/types/models/janus/image_processing_janus.d.ts +1 -1
  78. package/types/models/janus/image_processing_janus.d.ts.map +1 -1
  79. package/types/models/janus/processing_janus.d.ts.map +1 -1
  80. package/types/models/maskformer/image_processing_maskformer.d.ts +8 -8
  81. package/types/models/maskformer/image_processing_maskformer.d.ts.map +1 -1
  82. package/types/models/mgp_str/processing_mgp_str.d.ts +2 -2
  83. package/types/models/mgp_str/processing_mgp_str.d.ts.map +1 -1
  84. package/types/models/moonshine/feature_extraction_moonshine.d.ts +13 -0
  85. package/types/models/moonshine/feature_extraction_moonshine.d.ts.map +1 -0
  86. package/types/models/moonshine/processing_moonshine.d.ts +17 -0
  87. package/types/models/moonshine/processing_moonshine.d.ts.map +1 -0
  88. package/types/models/owlvit/image_processing_owlvit.d.ts.map +1 -1
  89. package/types/models/paligemma/processing_paligemma.d.ts +12 -0
  90. package/types/models/paligemma/processing_paligemma.d.ts.map +1 -0
  91. package/types/models/phi3_v/image_processing_phi3_v.d.ts +17 -0
  92. package/types/models/phi3_v/image_processing_phi3_v.d.ts.map +1 -0
  93. package/types/models/phi3_v/processing_phi3_v.d.ts +17 -0
  94. package/types/models/phi3_v/processing_phi3_v.d.ts.map +1 -0
  95. package/types/models/processors.d.ts +3 -0
  96. package/types/models/pyannote/feature_extraction_pyannote.d.ts +18 -0
  97. package/types/models/pyannote/feature_extraction_pyannote.d.ts.map +1 -1
  98. package/types/models/pyannote/processing_pyannote.d.ts +4 -15
  99. package/types/models/pyannote/processing_pyannote.d.ts.map +1 -1
  100. package/types/models/qwen2_vl/processing_qwen2_vl.d.ts.map +1 -1
  101. package/types/models/rt_detr/image_processing_rt_detr.d.ts.map +1 -1
  102. package/types/models/sam/image_processing_sam.d.ts.map +1 -1
  103. package/types/models/seamless_m4t/feature_extraction_seamless_m4t.d.ts +1 -1
  104. package/types/models/seamless_m4t/feature_extraction_seamless_m4t.d.ts.map +1 -1
  105. package/types/models/segformer/image_processing_segformer.d.ts.map +1 -1
  106. package/types/models/speecht5/processing_speecht5.d.ts.map +1 -1
  107. package/types/models/swin2sr/image_processing_swin2sr.d.ts +1 -1
  108. package/types/models/swin2sr/image_processing_swin2sr.d.ts.map +1 -1
  109. package/types/models/vitmatte/image_processing_vitmatte.d.ts.map +1 -1
  110. package/types/models/vitpose/image_processing_vitpose.d.ts +1 -1
  111. package/types/models/vitpose/image_processing_vitpose.d.ts.map +1 -1
  112. package/types/models/wav2vec2/feature_extraction_wav2vec2.d.ts.map +1 -1
  113. package/types/models/wav2vec2/processing_wav2vec2.d.ts.map +1 -1
  114. package/types/models/wespeaker/feature_extraction_wespeaker.d.ts +1 -1
  115. package/types/models/wespeaker/feature_extraction_wespeaker.d.ts.map +1 -1
  116. package/types/models/whisper/feature_extraction_whisper.d.ts +1 -1
  117. package/types/models/whisper/feature_extraction_whisper.d.ts.map +1 -1
  118. package/types/models/whisper/generation_whisper.d.ts.map +1 -1
  119. package/types/models/whisper/processing_whisper.d.ts.map +1 -1
  120. package/types/models/yolos/image_processing_yolos.d.ts.map +1 -1
  121. package/types/models.d.ts +61 -5
  122. package/types/models.d.ts.map +1 -1
  123. package/types/ops/registry.d.ts +1 -0
  124. package/types/ops/registry.d.ts.map +1 -1
  125. package/types/pipelines.d.ts +31 -51
  126. package/types/pipelines.d.ts.map +1 -1
  127. package/types/tokenizers.d.ts +10 -6
  128. package/types/tokenizers.d.ts.map +1 -1
  129. package/types/utils/audio.d.ts.map +1 -1
  130. package/types/utils/constants.d.ts.map +1 -1
  131. package/types/utils/core.d.ts +87 -22
  132. package/types/utils/core.d.ts.map +1 -1
  133. package/types/utils/data-structures.d.ts.map +1 -1
  134. package/types/utils/devices.d.ts.map +1 -1
  135. package/types/utils/dtypes.d.ts.map +1 -1
  136. package/types/utils/generic.d.ts.map +1 -1
  137. package/types/utils/hub.d.ts +3 -3
  138. package/types/utils/hub.d.ts.map +1 -1
  139. package/types/utils/image.d.ts +10 -1
  140. package/types/utils/image.d.ts.map +1 -1
  141. package/types/utils/maths.d.ts +10 -10
  142. package/types/utils/maths.d.ts.map +1 -1
  143. package/types/utils/tensor.d.ts +22 -6
  144. package/types/utils/tensor.d.ts.map +1 -1
@@ -1 +1 @@
1
- {"version":3,"file":"feature_extraction_clap.d.ts","sourceRoot":"","sources":["../../../src/models/clap/feature_extraction_clap.js"],"names":[],"mappings":"AAKA;IAEI,yBAyBC;IAtBG,wBAQC;IAED,+BAQC;IAED,qBAAkE;IAKtE;;;;;;;;;;;;;;;;;;;;OAoBG;IACH,uBA0CC;IAED;;;;;;;;;;;;;;OAcG;IACH,kCALW,YAAY,GAAC,YAAY,eACzB,MAAM,EAAE,EAAE,eACV,MAAM,GACJ,QAAQ,MAAM,CAAC,CAoB3B;IAGD;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY;;QACvB,QAAQ;QAAE,cAAc,EAAE,MAAM,CAAA;KAAE,CAAC,CAkB/C;CACJ;iCA9JuD,wCAAwC;uBACzE,uBAAuB"}
1
+ {"version":3,"file":"feature_extraction_clap.d.ts","sourceRoot":"","sources":["../../../src/models/clap/feature_extraction_clap.js"],"names":[],"mappings":"AAKA;IAEI,yBAyBC;IAtBG,wBAQC;IAED,+BAQC;IAED,sCAAkE;IAKtE;;;;;;;;;;;;;;;;;;;;OAoBG;IACH,uBA0CC;IAED;;;;;;;;;;;;;;OAcG;IACH,kCALW,YAAY,GAAC,YAAY,eACzB,MAAM,EAAE,EAAE,eACV,MAAM,GACJ,OAAO,CAAC,MAAM,CAAC,CAoB3B;IAGD;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY;;QACvB,OAAO,CAAC;QAAE,cAAc,EAAE,MAAM,CAAA;KAAE,CAAC,CAkB/C;CACJ;iCA9JuD,wCAAwC;uBACzE,uBAAuB"}
@@ -10,33 +10,33 @@ export class DetrImageProcessor extends ImageProcessor {
10
10
  * @param {import('../../utils/image.js').RawImage[]} images The image(s) to extract features from.
11
11
  * @returns {Promise<DetrFeatureExtractorResult>} An object containing the concatenated pixel values of the preprocessed images.
12
12
  */
13
- _call(images: import('../../utils/image.js').RawImage[]): Promise<DetrFeatureExtractorResult>;
13
+ _call(images: import("../../utils/image.js").RawImage[]): Promise<DetrFeatureExtractorResult>;
14
14
  post_process_object_detection(outputs: {
15
15
  logits: import("../../utils/tensor.js").Tensor;
16
16
  pred_boxes: import("../../utils/tensor.js").Tensor;
17
17
  }, threshold?: number, target_sizes?: [number, number][], is_zero_shot?: boolean): any[];
18
- post_process_panoptic_segmentation(outputs: any, threshold?: number, mask_threshold?: number, overlap_mask_area_threshold?: number, label_ids_to_fuse?: Set<number>, target_sizes?: [number, number][]): {
18
+ post_process_panoptic_segmentation(outputs: any, threshold?: number, mask_threshold?: number, overlap_mask_area_threshold?: number, label_ids_to_fuse?: Set<number>, target_sizes?: [number, number][]): Array<{
19
19
  segmentation: import("../../utils/tensor.js").Tensor;
20
- segments_info: {
20
+ segments_info: Array<{
21
21
  id: number;
22
22
  label_id: number;
23
23
  score: number;
24
- }[];
25
- }[];
26
- post_process_instance_segmentation(outputs: any, threshold?: number, target_sizes?: [number, number][]): {
24
+ }>;
25
+ }>;
26
+ post_process_instance_segmentation(outputs: any, threshold?: number, target_sizes?: [number, number][]): Array<{
27
27
  segmentation: import("../../utils/tensor.js").Tensor;
28
- segments_info: {
28
+ segments_info: Array<{
29
29
  id: number;
30
30
  label_id: number;
31
31
  score: number;
32
- }[];
33
- }[];
32
+ }>;
33
+ }>;
34
34
  }
35
35
  export class DetrFeatureExtractor extends DetrImageProcessor {
36
36
  }
37
37
  export type DetrFeatureExtractorResultProps = {
38
- pixel_mask: import('../../utils/tensor.js').Tensor;
38
+ pixel_mask: import("../../utils/tensor.js").Tensor;
39
39
  };
40
- export type DetrFeatureExtractorResult = import('../../base/image_processors_utils.js').ImageProcessorResult & DetrFeatureExtractorResultProps;
40
+ export type DetrFeatureExtractorResult = import("../../base/image_processors_utils.js").ImageProcessorResult & DetrFeatureExtractorResultProps;
41
41
  import { ImageProcessor } from "../../base/image_processors_utils.js";
42
42
  //# sourceMappingURL=image_processing_detr.d.ts.map
@@ -1 +1 @@
1
- {"version":3,"file":"image_processing_detr.d.ts","sourceRoot":"","sources":["../../../src/models/detr/image_processing_detr.js"],"names":[],"mappings":"AAUA;;;;GAIG;AAEH;IACI;;;;;OAKG;IACH,cAHW,OAAO,sBAAsB,EAAE,QAAQ,EAAE,GACvC,QAAQ,0BAA0B,CAAC,CAY/C;IAmBwmB;;;6FAAu6F;IAA8yU;;;;;;;QAA4qG;IAAA;;;;;;;QAAktB;CAH/rjB;AAED;CAAgE;;gBAvClD,OAAO,uBAAuB,EAAE,MAAM;;yCACvC,OAAO,sCAAsC,EAAE,oBAAoB,GAAG,+BAA+B;+BAR3G,sCAAsC"}
1
+ {"version":3,"file":"image_processing_detr.d.ts","sourceRoot":"","sources":["../../../src/models/detr/image_processing_detr.js"],"names":[],"mappings":"AAUA;;;;GAIG;AAEH;IACI;;;;;OAKG;IACH,cAHW,OAAO,sBAAsB,EAAE,QAAQ,EAAE,GACvC,OAAO,CAAC,0BAA0B,CAAC,CAY/C;;;;;;;;;;;;;;;;;;;;;CAgBJ;AAED;CAAgE;;gBAvClD,OAAO,uBAAuB,EAAE,MAAM;;yCACvC,OAAO,sCAAsC,EAAE,oBAAoB,GAAG,+BAA+B;+BAR3G,sCAAsC"}
@@ -1,5 +1,5 @@
1
1
  export class DonutImageProcessor extends ImageProcessor {
2
- pad_image(pixelData: any, imgDims: any, padSize: any, options?: {}): [Float32Array, number[]];
2
+ pad_image(pixelData: any, imgDims: any, padSize: any, options?: {}): [Float32Array<ArrayBufferLike>, number[]];
3
3
  }
4
4
  export class DonutFeatureExtractor extends DonutImageProcessor {
5
5
  }
@@ -1 +1 @@
1
- {"version":3,"file":"image_processing_donut.d.ts","sourceRoot":"","sources":["../../../src/models/donut/image_processing_donut.js"],"names":[],"mappings":"AAIA;IACI,8FAuBC;CACJ;AACD;CAAkE;+BA5B3D,sCAAsC"}
1
+ {"version":3,"file":"image_processing_donut.d.ts","sourceRoot":"","sources":["../../../src/models/donut/image_processing_donut.js"],"names":[],"mappings":"AAIA;IACI,+GAuBC;CACJ;AACD;CAAkE;+BA5B3D,sCAAsC"}
@@ -1,5 +1,6 @@
1
1
  export * from "./audio_spectrogram_transformer/feature_extraction_audio_spectrogram_transformer.js";
2
2
  export * from "./clap/feature_extraction_clap.js";
3
+ export * from "./moonshine/feature_extraction_moonshine.js";
3
4
  export * from "./pyannote/feature_extraction_pyannote.js";
4
5
  export * from "./seamless_m4t/feature_extraction_seamless_m4t.js";
5
6
  export * from "./speecht5/feature_extraction_speecht5.js";
@@ -1 +1 @@
1
- {"version":3,"file":"processing_florence2.d.ts","sourceRoot":"","sources":["../../../src/models/florence2/processing_florence2.js"],"names":[],"mappings":"AAIA;IACI,6CAAsC;IACtC,wDAAiD;IAEjD,0CAuBC;IAdG,kCAAkC;IAClC,mCADW,IAAI,MAAM,EAAE,MAAM,CAAC,CAC2E;IAEzG,kCAAkC;IAClC,6BADW,IAAI,MAAM,EAAE,MAAM,CAAC,CAC+D;IAE7F,kCAAkC;IAClC,yBADW,IAAI,MAAM,EAAE,MAAM,CAAC,CACuD;IAErF;;;MAGC;IACD,qBAAwB;IAG5B;;;;OAIG;IACH,wBAHW,MAAM,GAAC,MAAM,EAAE,GACb,MAAM,EAAE,CA6BpB;IAED;;;;;OAKG;IACH,8BAJW,MAAM,QACN,MAAM,cACN,CAAC,MAAM,EAAE,MAAM,CAAC;;;;;MAsC1B;IAID,0DAaC;CACJ;0BA/HyB,gCAAgC;8BAE5B,qBAAqB;mCADhB,kCAAkC"}
1
+ {"version":3,"file":"processing_florence2.d.ts","sourceRoot":"","sources":["../../../src/models/florence2/processing_florence2.js"],"names":[],"mappings":"AAIA;IACI,6CAAsC;IACtC,wDAAiD;IAEjD,0CAuBC;IAdG,kCAAkC;IAClC,mCADW,GAAG,CAAC,MAAM,EAAE,MAAM,CAAC,CAC2E;IAEzG,kCAAkC;IAClC,6BADW,GAAG,CAAC,MAAM,EAAE,MAAM,CAAC,CAC+D;IAE7F,kCAAkC;IAClC,yBADW,GAAG,CAAC,MAAM,EAAE,MAAM,CAAC,CACuD;IAErF;;;MAGC;IACD,qBAAwB;IAG5B;;;;OAIG;IACH,wBAHW,MAAM,GAAC,MAAM,EAAE,GACb,MAAM,EAAE,CA6BpB;IAED;;;;;OAKG;IACH,8BAJW,MAAM,QACN,MAAM,cACN,CAAC,MAAM,EAAE,MAAM,CAAC;;;;;MAsC1B;IAID,0DAaC;CACJ;0BA/HyB,gCAAgC;8BAE5B,qBAAqB;mCADhB,kCAAkC"}
@@ -1 +1 @@
1
- {"version":3,"file":"image_processing_idefics3.d.ts","sourceRoot":"","sources":["../../../src/models/idefics3/image_processing_idefics3.js"],"names":[],"mappings":"AAOA;IACI,yBAKC;IAFG,wBAA2D;IAC3D,oBAA2C;IAG/C;;;OAGG;IAEH;;;;;OAKG;IACH,6GAHW,MAAM;;;MAiBhB;IAED,uDAAuD;IACvD;;;;;;;;;;OA2HC;IAED;;;;;;OAiDC;CACJ;+BAtNM,sCAAsC"}
1
+ {"version":3,"file":"image_processing_idefics3.d.ts","sourceRoot":"","sources":["../../../src/models/idefics3/image_processing_idefics3.js"],"names":[],"mappings":"AAOA;IACI,yBAKC;IAFG,wBAA2D;IAC3D,oBAA2C;IAG/C;;;OAGG;IAEH;;;;;OAKG;IACH,6GAHW,MAAM;;;MAiBhB;IAED,uDAAuD;IACvD,cADY,0CAAS,yCAAU,GAAC,yCAAU,EAAE;;;;;;;;;;OA4H3C;IAED;;;;;;OA4DC;CACJ;+BAjOM,sCAAsC"}
@@ -1 +1 @@
1
- {"version":3,"file":"processing_idefics3.d.ts","sourceRoot":"","sources":["../../../src/models/idefics3/processing_idefics3.js"],"names":[],"mappings":"AA6DA;IACI,wDAAiD;IACjD,6CAAsC;IAGtC,yBAA+C;IAC/C,oBAAwB;IACxB,yBAAkC;IAElC;;;;;OAKG;IACH,YAJW,MAAM,GAAC,MAAM,EAAE,WACf,QAAQ,GAAC,QAAQ,EAAE,GAAC,QAAQ,EAAE,EAAE,iBAC9B,QAAQ,GAAG,CAAC,CA4DxB;CACJ;0BAtIyB,gCAAgC;yBAGjC,sBAAsB;mCAFZ,kCAAkC;8BACvC,qBAAqB"}
1
+ {"version":3,"file":"processing_idefics3.d.ts","sourceRoot":"","sources":["../../../src/models/idefics3/processing_idefics3.js"],"names":[],"mappings":"AA6DA;IACI,wDAAiD;IACjD,6CAAsC;IAGtC,yBAA+C;IAC/C,oBAAwB;IACxB,yBAAkC;IAElC;;;;;OAKG;IACH,YAJW,MAAM,GAAC,MAAM,EAAE,WACf,QAAQ,GAAC,QAAQ,EAAE,GAAC,QAAQ,EAAE,EAAE,iBAC9B,OAAO,CAAC,GAAG,CAAC,CA4DxB;CACJ;0BAtIyB,gCAAgC;yBAGjC,sBAAsB;mCAFZ,kCAAkC;8BACvC,qBAAqB"}
@@ -23,6 +23,7 @@ export * from "./mobilevit/image_processing_mobilevit.js";
23
23
  export * from "./nougat/image_processing_nougat.js";
24
24
  export * from "./owlv2/image_processing_owlv2.js";
25
25
  export * from "./owlvit/image_processing_owlvit.js";
26
+ export * from "./phi3_v/image_processing_phi3_v.js";
26
27
  export * from "./pvt/image_processing_pvt.js";
27
28
  export * from "./qwen2_vl/image_processing_qwen2_vl.js";
28
29
  export * from "./rt_detr/image_processing_rt_detr.js";
@@ -1,7 +1,7 @@
1
1
  export class VLMImageProcessor extends ImageProcessor {
2
2
  constructor(config: any);
3
3
  constant_values: any;
4
- pad_image(pixelData: any, imgDims: any, padSize: any, options: any): [Float32Array, number[]];
4
+ pad_image(pixelData: any, imgDims: any, padSize: any, options: any): [Float32Array<ArrayBufferLike>, number[]];
5
5
  }
6
6
  import { ImageProcessor } from "../../base/image_processors_utils.js";
7
7
  //# sourceMappingURL=image_processing_janus.d.ts.map
@@ -1 +1 @@
1
- {"version":3,"file":"image_processing_janus.d.ts","sourceRoot":"","sources":["../../../src/models/janus/image_processing_janus.js"],"names":[],"mappings":"AAKA;IACI,yBAUC;IADG,qBAAqF;IAGzF,8FAMC;CACJ;+BAtBM,sCAAsC"}
1
+ {"version":3,"file":"image_processing_janus.d.ts","sourceRoot":"","sources":["../../../src/models/janus/image_processing_janus.js"],"names":[],"mappings":"AAKA;IACI,yBAUC;IADG,qBAAqF;IAGzF,+GAMC;CACJ;+BAtBM,sCAAsC"}
@@ -1 +1 @@
1
- {"version":3,"file":"processing_janus.d.ts","sourceRoot":"","sources":["../../../src/models/janus/processing_janus.js"],"names":[],"mappings":"AAQA;IACI,wDAAiD;IACjD,6CAAsC;IAGtC,0CAOC;IAJG,eAAsC;IACtC,qBAAkD;IAClD,mBAA8C;IAC9C,sBAAoD;IAGxD;;;;OAIG;IAEH;;;;;;OAMG;IAEH;;;;;;OAMG;IACH;;;;iBAnBc,CAAC,QAAQ,GAAG,MAAM,GAAG,GAAG,CAAC,EAAE;;QAeA,MAAM,GAApC,QAAQ,GAAC,QAAQ,EAAE;QACF,aAAa,GAA9B,MAAM;;;;;mBAVH,MAAM;;;;wBACN,MAAM;;;;yBACN,MAAM;;;;yBACN,MAAM;;;;;mBAHN,MAAM;;;;wBACN,MAAM;;;;yBACN,MAAM;;;;yBACN,MAAM;8EAwFnB;CACJ;0BAzHyB,gCAAgC;yBAKjC,sBAAsB;uBADxB,uBAAuB;mCAHX,kCAAkC;8BACvC,qBAAqB"}
1
+ {"version":3,"file":"processing_janus.d.ts","sourceRoot":"","sources":["../../../src/models/janus/processing_janus.js"],"names":[],"mappings":"AAQA;IACI,wDAAiD;IACjD,6CAAsC;IAGtC,0CAOC;IAJG,eAAsC;IACtC,qBAAkD;IAClD,mBAA8C;IAC9C,sBAAoD;IAGxD;;;;OAIG;IAEH;;;;;;OAMG;IAEH;;;;;;OAMG;IACH;;;;iBAnBc,CAAC,QAAQ,GAAG,MAAM,GAAG,GAAG,CAAC,EAAE;uCAetC;QAAsC,MAAM,GAApC,QAAQ,GAAC,QAAQ,EAAE;QACF,aAAa,GAA9B,MAAM;KACd,GAAU,OAAO,CAAC;;;;mBAXP,MAAM;;;;wBACN,MAAM;;;;yBACN,MAAM;;;;yBACN,MAAM;SAQ0B;;;;mBAXhC,MAAM;;;;wBACN,MAAM;;;;yBACN,MAAM;;;;yBACN,MAAM;QAQmD,OAAO,sCAAsC,EAAE,oBAAoB,CAAA,CAAC,CAgF1I;CACJ;0BAzHyB,gCAAgC;yBAKjC,sBAAsB;uBADxB,uBAAuB;mCAHX,kCAAkC;8BACvC,qBAAqB"}
@@ -1,20 +1,20 @@
1
1
  export class MaskFormerImageProcessor extends ImageProcessor {
2
- post_process_panoptic_segmentation(outputs: any, threshold?: number, mask_threshold?: number, overlap_mask_area_threshold?: number, label_ids_to_fuse?: Set<number>, target_sizes?: [number, number][]): {
2
+ post_process_panoptic_segmentation(outputs: any, threshold?: number, mask_threshold?: number, overlap_mask_area_threshold?: number, label_ids_to_fuse?: Set<number>, target_sizes?: [number, number][]): Array<{
3
3
  segmentation: import("../../transformers.js").Tensor;
4
- segments_info: {
4
+ segments_info: Array<{
5
5
  id: number;
6
6
  label_id: number;
7
7
  score: number;
8
- }[];
9
- }[];
10
- post_process_instance_segmentation(outputs: any, threshold?: number, target_sizes?: [number, number][]): {
8
+ }>;
9
+ }>;
10
+ post_process_instance_segmentation(outputs: any, threshold?: number, target_sizes?: [number, number][]): Array<{
11
11
  segmentation: import("../../transformers.js").Tensor;
12
- segments_info: {
12
+ segments_info: Array<{
13
13
  id: number;
14
14
  label_id: number;
15
15
  score: number;
16
- }[];
17
- }[];
16
+ }>;
17
+ }>;
18
18
  }
19
19
  export class MaskFormerFeatureExtractor extends MaskFormerImageProcessor {
20
20
  }
@@ -1 +1 @@
1
- {"version":3,"file":"image_processing_maskformer.d.ts","sourceRoot":"","sources":["../../../src/models/maskformer/image_processing_maskformer.js"],"names":[],"mappings":"AAMA;IAYure;;;;;;;QAA4qG;IAAA;;;;;;;QAAktB;CAFpjmB;AACD;CAA4E;+BAbrE,sCAAsC"}
1
+ {"version":3,"file":"image_processing_maskformer.d.ts","sourceRoot":"","sources":["../../../src/models/maskformer/image_processing_maskformer.js"],"names":[],"mappings":"AAMA;;;;;;;;;;;;;;;;;CAUC;AACD;CAA4E;+BAbrE,sCAAsC"}
@@ -19,7 +19,7 @@ export class MgpstrProcessor extends Processor {
19
19
  * @param {string} format Type of model prediction. Must be one of ['char', 'bpe', 'wp'].
20
20
  * @returns {[string[], number[]]} The decoded sentences and their confidence scores.
21
21
  */
22
- _decode_helper(pred_logits: import('../../utils/tensor.js').Tensor, format: string): [string[], number[]];
22
+ _decode_helper(pred_logits: import("../../utils/tensor.js").Tensor, format: string): [string[], number[]];
23
23
  /**
24
24
  * Convert a list of lists of char token ids into a list of strings by calling char tokenizer.
25
25
  * @param {number[][]} sequences List of tokenized input ids.
@@ -49,7 +49,7 @@ export class MgpstrProcessor extends Processor {
49
49
  * - bpe_preds: The list of BPE decoded sentences.
50
50
  * - wp_preds: The list of wp decoded sentences.
51
51
  */
52
- batch_decode([char_logits, bpe_logits, wp_logits]: import('../../utils/tensor.js').Tensor[]): {
52
+ batch_decode([char_logits, bpe_logits, wp_logits]: import("../../utils/tensor.js").Tensor[]): {
53
53
  generated_text: string[];
54
54
  scores: number[];
55
55
  char_preds: string[];
@@ -1 +1 @@
1
- {"version":3,"file":"processing_mgp_str.d.ts","sourceRoot":"","sources":["../../../src/models/mgp_str/processing_mgp_str.js"],"names":[],"mappings":"AAUA;IACI,6CAAsC;IACtC,wDAAiD;IAEjD;;OAEG;IACH,oEAEC;IAED;;OAEG;IACH,iEAEC;IAED;;OAEG;IACH,gEAEC;IAED;;;;;OAKG;IACH,4BAJW,OAAO,uBAAuB,EAAE,MAAM,UACtC,MAAM,GACJ,CAAC,MAAM,EAAE,EAAE,MAAM,EAAE,CAAC,CA0ChC;IAED;;;;OAIG;IACH,uBAHW,MAAM,EAAE,EAAE,GACR,MAAM,EAAE,CAIpB;IAED;;;;OAIG;IACH,sBAHW,MAAM,EAAE,EAAE,GACR,MAAM,EAAE,CAIpB;IAED;;;;OAIG;IACH,qBAHW,MAAM,EAAE,EAAE,GACR,MAAM,EAAE,CAIpB;IAED;;;;;;;;;;OAUG;IACH,mDATW,OAAO,uBAAuB,EAAE,MAAM,EAAE;wBACrB,MAAM,EAAE;gBAAU,MAAM,EAAE;oBAAc,MAAM,EAAE;mBAAa,MAAM,EAAE;kBAAY,MAAM,EAAE;MA4BtH;IAmBD,6CAQC;CACJ;0BAzKyB,gCAAgC;8BAE5B,qBAAqB;mCADhB,kCAAkC"}
1
+ {"version":3,"file":"processing_mgp_str.d.ts","sourceRoot":"","sources":["../../../src/models/mgp_str/processing_mgp_str.js"],"names":[],"mappings":"AAUA;IACI,6CAAsC;IACtC,wDAAiD;IAEjD;;OAEG;IACH,sBAFa,OAAO,qBAAqB,EAAE,eAAe,CAIzD;IAED;;OAEG;IACH,qBAFa,OAAO,qBAAqB,EAAE,aAAa,CAIvD;IAED;;OAEG;IACH,oBAFa,OAAO,qBAAqB,EAAE,aAAa,CAIvD;IAED;;;;;OAKG;IACH,4BAJW,OAAO,uBAAuB,EAAE,MAAM,UACtC,MAAM,GACJ,CAAC,MAAM,EAAE,EAAE,MAAM,EAAE,CAAC,CA0ChC;IAED;;;;OAIG;IACH,uBAHW,MAAM,EAAE,EAAE,GACR,MAAM,EAAE,CAIpB;IAED;;;;OAIG;IACH,sBAHW,MAAM,EAAE,EAAE,GACR,MAAM,EAAE,CAIpB;IAED;;;;OAIG;IACH,qBAHW,MAAM,EAAE,EAAE,GACR,MAAM,EAAE,CAIpB;IAED;;;;;;;;;;OAUG;IACH,mDATW,OAAO,uBAAuB,EAAE,MAAM,EAAE,GACtC;QAAC,cAAc,EAAE,MAAM,EAAE,CAAC;QAAC,MAAM,EAAE,MAAM,EAAE,CAAC;QAAC,UAAU,EAAE,MAAM,EAAE,CAAC;QAAC,SAAS,EAAE,MAAM,EAAE,CAAC;QAAC,QAAQ,EAAE,MAAM,EAAE,CAAA;KAAC,CA4BvH;IAmBD,6CAQC;CACJ;0BAzKyB,gCAAgC;8BAE5B,qBAAqB;mCADhB,kCAAkC"}
@@ -0,0 +1,13 @@
1
+ export class MoonshineFeatureExtractor extends FeatureExtractor {
2
+ /**
3
+ * Asynchronously extracts input values from a given audio using the provided configuration.
4
+ * @param {Float32Array|Float64Array} audio The audio data as a Float32Array/Float64Array.
5
+ * @returns {Promise<{ input_values: Tensor; }>} The extracted input values.
6
+ */
7
+ _call(audio: Float32Array | Float64Array): Promise<{
8
+ input_values: Tensor;
9
+ }>;
10
+ }
11
+ import { FeatureExtractor } from '../../base/feature_extraction_utils.js';
12
+ import { Tensor } from '../../utils/tensor.js';
13
+ //# sourceMappingURL=feature_extraction_moonshine.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"feature_extraction_moonshine.d.ts","sourceRoot":"","sources":["../../../src/models/moonshine/feature_extraction_moonshine.js"],"names":[],"mappings":"AAIA;IACI;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY,GACvB,OAAO,CAAC;QAAE,YAAY,EAAE,MAAM,CAAC;KAAE,CAAC,CAgB9C;CACJ;iCAzBuD,wCAAwC;uBACzE,uBAAuB"}
@@ -0,0 +1,17 @@
1
+ /**
2
+ * Represents a MoonshineProcessor that extracts features from an audio input.
3
+ */
4
+ export class MoonshineProcessor extends Processor {
5
+ static tokenizer_class: typeof AutoTokenizer;
6
+ static feature_extractor_class: typeof AutoFeatureExtractor;
7
+ /**
8
+ * Calls the feature_extractor function with the given audio input.
9
+ * @param {any} audio The audio input to extract features from.
10
+ * @returns {Promise<any>} A Promise that resolves with the extracted features.
11
+ */
12
+ _call(audio: any): Promise<any>;
13
+ }
14
+ import { Processor } from "../../base/processing_utils.js";
15
+ import { AutoTokenizer } from "../../tokenizers.js";
16
+ import { AutoFeatureExtractor } from "../auto/feature_extraction_auto.js";
17
+ //# sourceMappingURL=processing_moonshine.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"processing_moonshine.d.ts","sourceRoot":"","sources":["../../../src/models/moonshine/processing_moonshine.js"],"names":[],"mappings":"AAIA;;GAEG;AACH;IACI,6CAAsC;IACtC,4DAAqD;IAErD;;;;OAIG;IACH,aAHW,GAAG,GACD,OAAO,CAAC,GAAG,CAAC,CAIxB;CACJ;0BAjByB,gCAAgC;8BAD5B,qBAAqB;qCADd,oCAAoC"}
@@ -1 +1 @@
1
- {"version":3,"file":"image_processing_owlvit.d.ts","sourceRoot":"","sources":["../../../src/models/owlvit/image_processing_owlvit.js"],"names":[],"mappings":"AAKA;IAO4tE;;;6FAAu6F;CAFloK;AACD;CAAoE;+BAR7D,sCAAsC"}
1
+ {"version":3,"file":"image_processing_owlvit.d.ts","sourceRoot":"","sources":["../../../src/models/owlvit/image_processing_owlvit.js"],"names":[],"mappings":"AAKA;;;;;CAKC;AACD;CAAoE;+BAR7D,sCAAsC"}
@@ -0,0 +1,12 @@
1
+ export class PaliGemmaProcessor extends Processor {
2
+ static tokenizer_class: typeof AutoTokenizer;
3
+ static image_processor_class: typeof AutoImageProcessor;
4
+ /**
5
+ * @typedef {import('../../utils/image.js').RawImage} RawImage
6
+ */
7
+ _call(images: import("../../utils/image.js").RawImage | import("../../utils/image.js").RawImage[], text?: any, kwargs?: {}): Promise<any>;
8
+ }
9
+ import { Processor } from "../../base/processing_utils.js";
10
+ import { AutoTokenizer } from "../../tokenizers.js";
11
+ import { AutoImageProcessor } from "../auto/image_processing_auto.js";
12
+ //# sourceMappingURL=processing_paligemma.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"processing_paligemma.d.ts","sourceRoot":"","sources":["../../../src/models/paligemma/processing_paligemma.js"],"names":[],"mappings":"AAgBA;IACI,6CAAsC;IACtC,wDAAiD;IAGjD;;OAEG;IAGH,cAAuB,0CAAS,yCAAU,yCAsDzC;CACJ;0BAjFyB,gCAAgC;8BAE5B,qBAAqB;mCADhB,kCAAkC"}
@@ -0,0 +1,17 @@
1
+ export class Phi3VImageProcessor extends ImageProcessor {
2
+ constructor(config: any);
3
+ _num_crops: any;
4
+ calc_num_image_tokens_from_image_size(width: any, height: any): number;
5
+ _call(images: any, { num_crops, }?: {
6
+ num_crops?: any;
7
+ }): Promise<{
8
+ pixel_values: Tensor;
9
+ original_sizes: any[];
10
+ reshaped_input_sizes: any[];
11
+ image_sizes: Tensor;
12
+ num_img_tokens: number[];
13
+ }>;
14
+ }
15
+ import { ImageProcessor } from "../../base/image_processors_utils.js";
16
+ import { Tensor } from "../../utils/tensor.js";
17
+ //# sourceMappingURL=image_processing_phi3_v.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"image_processing_phi3_v.d.ts","sourceRoot":"","sources":["../../../src/models/phi3_v/image_processing_phi3_v.js"],"names":[],"mappings":"AASA;IACI,yBAWC;IADG,gBAAkC;IAEtC,uEAIC;IA0CD;;;;;;;;OA6FC;CACJ;+BAhKM,sCAAsC;uBACa,uBAAuB"}
@@ -0,0 +1,17 @@
1
+ export class Phi3VProcessor extends Processor {
2
+ static image_processor_class: typeof AutoImageProcessor;
3
+ static tokenizer_class: typeof AutoTokenizer;
4
+ /**
5
+ *
6
+ * @param {string|string[]} text
7
+ * @param {RawImage|RawImage[]} images
8
+ * @param {...any} args
9
+ * @returns {Promise<any>}
10
+ */
11
+ _call(text: string | string[], images?: RawImage | RawImage[], { padding, truncation, num_crops, }?: any[]): Promise<any>;
12
+ }
13
+ import { Processor } from "../../base/processing_utils.js";
14
+ import { RawImage } from "../../utils/image.js";
15
+ import { AutoImageProcessor } from "../auto/image_processing_auto.js";
16
+ import { AutoTokenizer } from "../../tokenizers.js";
17
+ //# sourceMappingURL=processing_phi3_v.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"processing_phi3_v.d.ts","sourceRoot":"","sources":["../../../src/models/phi3_v/processing_phi3_v.js"],"names":[],"mappings":"AAQA;IACI,wDAAiD;IACjD,6CAAsC;IAEtC;;;;;;OAMG;IACH,YALW,MAAM,GAAC,MAAM,EAAE,WACf,QAAQ,GAAC,QAAQ,EAAE,wCACf,GAAG,EAAA,GACL,OAAO,CAAC,GAAG,CAAC,CAkCxB;CACJ;0BApDyB,gCAAgC;yBAGjC,sBAAsB;mCAFZ,kCAAkC;8BACvC,qBAAqB"}
@@ -1,9 +1,12 @@
1
1
  export * from "./florence2/processing_florence2.js";
2
2
  export * from "./mgp_str/processing_mgp_str.js";
3
+ export * from "./moonshine/processing_moonshine.js";
3
4
  export * from "./idefics3/processing_idefics3.js";
4
5
  export * from "./janus/processing_janus.js";
5
6
  export * from "./jina_clip/processing_jina_clip.js";
6
7
  export * from "./owlvit/processing_owlvit.js";
8
+ export * from "./phi3_v/processing_phi3_v.js";
9
+ export * from "./paligemma/processing_paligemma.js";
7
10
  export * from "./pyannote/processing_pyannote.js";
8
11
  export * from "./qwen2_vl/processing_qwen2_vl.js";
9
12
  export * from "./sam/processing_sam.js";
@@ -7,6 +7,24 @@ export class PyAnnoteFeatureExtractor extends FeatureExtractor {
7
7
  _call(audio: Float32Array | Float64Array): Promise<{
8
8
  input_values: Tensor;
9
9
  }>;
10
+ /**
11
+ * NOTE: Can return fractional values. `Math.ceil` will ensure correct value.
12
+ * @param {number} samples The number of frames in the audio.
13
+ * @returns {number} The number of frames in the audio.
14
+ */
15
+ samples_to_frames(samples: number): number;
16
+ /**
17
+ * Post-processes the speaker diarization logits output by the model.
18
+ * @param {import('../../utils/tensor.js').Tensor} logits The speaker diarization logits output by the model.
19
+ * @param {number} num_samples Number of samples in the input audio.
20
+ * @returns {Array<Array<{ id: number, start: number, end: number, confidence: number }>>} The post-processed speaker diarization results.
21
+ */
22
+ post_process_speaker_diarization(logits: import("../../utils/tensor.js").Tensor, num_samples: number): Array<Array<{
23
+ id: number;
24
+ start: number;
25
+ end: number;
26
+ confidence: number;
27
+ }>>;
10
28
  }
11
29
  import { FeatureExtractor } from '../../base/feature_extraction_utils.js';
12
30
  import { Tensor } from '../../utils/tensor.js';
@@ -1 +1 @@
1
- {"version":3,"file":"feature_extraction_pyannote.d.ts","sourceRoot":"","sources":["../../../src/models/pyannote/feature_extraction_pyannote.js"],"names":[],"mappings":"AAIA;IACI;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY,GACvB,QAAQ;QAAE,YAAY,EAAE,MAAM,CAAC;KAAE,CAAC,CAiB9C;CAEJ;iCA3BuD,wCAAwC;uBACzE,uBAAuB"}
1
+ {"version":3,"file":"feature_extraction_pyannote.d.ts","sourceRoot":"","sources":["../../../src/models/pyannote/feature_extraction_pyannote.js"],"names":[],"mappings":"AAKA;IACI;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY,GACvB,OAAO,CAAC;QAAE,YAAY,EAAE,MAAM,CAAC;KAAE,CAAC,CAiB9C;IAED;;;;OAIG;IACH,2BAHW,MAAM,GACJ,MAAM,CAIlB;IAED;;;;;OAKG;IACH,yCAJW,OAAO,uBAAuB,EAAE,MAAM,eACtC,MAAM,GACJ,KAAK,CAAC,KAAK,CAAC;QAAE,EAAE,EAAE,MAAM,CAAC;QAAC,KAAK,EAAE,MAAM,CAAC;QAAC,GAAG,EAAE,MAAM,CAAC;QAAC,UAAU,EAAE,MAAM,CAAA;KAAE,CAAC,CAAC,CAwCxF;CAEJ;iCAnFuD,wCAAwC;uBACzE,uBAAuB"}
@@ -1,30 +1,19 @@
1
1
  export class PyAnnoteProcessor extends Processor {
2
- static feature_extractor_class: typeof AutoFeatureExtractor;
2
+ static feature_extractor_class: typeof PyAnnoteFeatureExtractor;
3
3
  /**
4
4
  * Calls the feature_extractor function with the given audio input.
5
5
  * @param {any} audio The audio input to extract features from.
6
6
  * @returns {Promise<any>} A Promise that resolves with the extracted features.
7
7
  */
8
8
  _call(audio: any): Promise<any>;
9
- /**
10
- * NOTE: Can return fractional values. `Math.ceil` will ensure correct value.
11
- * @param {number} samples The number of frames in the audio.
12
- * @returns {number} The number of frames in the audio.
13
- */
14
- samples_to_frames(samples: number): number;
15
- /**
16
- * Post-processes the speaker diarization logits output by the model.
17
- * @param {import('../../utils/tensor.js').Tensor} logits The speaker diarization logits output by the model.
18
- * @param {number} num_samples Number of samples in the input audio.
19
- * @returns {Array<Array<{ id: number, start: number, end: number, confidence: number }>>} The post-processed speaker diarization results.
20
- */
21
- post_process_speaker_diarization(logits: import('../../utils/tensor.js').Tensor, num_samples: number): Array<Array<{
9
+ post_process_speaker_diarization(logits: import("../../transformers.js").Tensor, num_samples: number): Array<Array<{
22
10
  id: number;
23
11
  start: number;
24
12
  end: number;
25
13
  confidence: number;
26
14
  }>>;
15
+ get sampling_rate(): any;
27
16
  }
28
17
  import { Processor } from '../../base/processing_utils.js';
29
- import { AutoFeatureExtractor } from '../auto/feature_extraction_auto.js';
18
+ import { PyAnnoteFeatureExtractor } from './feature_extraction_pyannote.js';
30
19
  //# sourceMappingURL=processing_pyannote.d.ts.map
@@ -1 +1 @@
1
- {"version":3,"file":"processing_pyannote.d.ts","sourceRoot":"","sources":["../../../src/models/pyannote/processing_pyannote.js"],"names":[],"mappings":"AAIA;IACI,4DAAqD;IAErD;;;;OAIG;IACH,aAHW,GAAG,GACD,QAAQ,GAAG,CAAC,CAIxB;IAED;;;;OAIG;IACH,2BAHW,MAAM,GACJ,MAAM,CAIlB;IAED;;;;;OAKG;IACH,yCAJW,OAAO,uBAAuB,EAAE,MAAM,eACtC,MAAM,GACJ,MAAM,MAAM;QAAE,EAAE,EAAE,MAAM,CAAC;QAAC,KAAK,EAAE,MAAM,CAAC;QAAC,GAAG,EAAE,MAAM,CAAC;QAAC,UAAU,EAAE,MAAM,CAAA;KAAE,CAAC,CAAC,CAwCxF;CACJ;0BAtEyB,gCAAgC;qCACrB,oCAAoC"}
1
+ {"version":3,"file":"processing_pyannote.d.ts","sourceRoot":"","sources":["../../../src/models/pyannote/processing_pyannote.js"],"names":[],"mappings":"AAGA;IACI,gEAAyD;IAEzD;;;;OAIG;IACH,aAHW,GAAG,GACD,OAAO,CAAC,GAAG,CAAC,CAIxB;oDAWogB,uBAAuB;;;;;;IAJ5hB,yBAEC;CACJ;0BAvByB,gCAAgC;yCACjB,kCAAkC"}
@@ -1 +1 @@
1
- {"version":3,"file":"processing_qwen2_vl.d.ts","sourceRoot":"","sources":["../../../src/models/qwen2_vl/processing_qwen2_vl.js"],"names":[],"mappings":"AAKA;IACI,wDAAiD;IACjD,6CAAsC;IAEtC;;;;;;OAMG;IACH,YALW,MAAM,GAAC,MAAM,EAAE,WACf,QAAQ,GAAC,QAAQ,EAAE,WACf,GAAG,KACL,QAAQ,GAAG,CAAC,CAoCxB;CACJ;0BAnDyB,gCAAgC;yBAGjC,sBAAsB;mCAFZ,kCAAkC;8BACvC,qBAAqB"}
1
+ {"version":3,"file":"processing_qwen2_vl.d.ts","sourceRoot":"","sources":["../../../src/models/qwen2_vl/processing_qwen2_vl.js"],"names":[],"mappings":"AAKA;IACI,wDAAiD;IACjD,6CAAsC;IAEtC;;;;;;OAMG;IACH,YALW,MAAM,GAAC,MAAM,EAAE,WACf,QAAQ,GAAC,QAAQ,EAAE,WACf,GAAG,EAAA,GACL,OAAO,CAAC,GAAG,CAAC,CAoCxB;CACJ;0BAnDyB,gCAAgC;yBAGjC,sBAAsB;mCAFZ,kCAAkC;8BACvC,qBAAqB"}
@@ -1 +1 @@
1
- {"version":3,"file":"image_processing_rt_detr.d.ts","sourceRoot":"","sources":["../../../src/models/rt_detr/image_processing_rt_detr.js"],"names":[],"mappings":"AAMA;IAMgyE;;;6FAAu6F;CADtsK;+BARM,sCAAsC"}
1
+ {"version":3,"file":"image_processing_rt_detr.d.ts","sourceRoot":"","sources":["../../../src/models/rt_detr/image_processing_rt_detr.js"],"names":[],"mappings":"AAMA;;;;;CAKC;+BARM,sCAAsC"}
@@ -1 +1 @@
1
- {"version":3,"file":"image_processing_sam.d.ts","sourceRoot":"","sources":["../../../src/models/sam/image_processing_sam.js"],"names":[],"mappings":"AAWA;;;;;;;;GAQG;AAEH;IAEI;;;;;;OAMG;IACH,mCALW,GAAG,kBACH,OAAO,sCAAsC,EAAE,WAAW,EAAE,wBAC5D,OAAO,sCAAsC,EAAE,WAAW,EAAE,8BAC1D,MAAM,CA4ClB;IAED;;;;;OAKG;IACH,+BAJW,GAAG,gBACH,MAAM,GACJ,MAAM,CAoBlB;IACD;;;;;;;;;;;;;;;;;;;OAmBG;IACH,cAnBW,GAAG,EAAE;QAES,YAAY,GAA1B,GAAG;QAGW,YAAY,GAA1B,GAAG;QAGoB,WAAW,GAAlC,MAAM,EAAE,EAAE,EAAE;QASV,QAAQ,uBAAuB,CAAC,CA+B5C;IAED;;;;;;;;;;;;OAYG;IACH,0BAXW,MAAM,kBACN,CAAC,MAAM,EAAE,MAAM,CAAC,EAAE,wBAClB,CAAC,MAAM,EAAE,MAAM,CAAC,EAAE;QAED,cAAc,GAA/B,MAAM;QACY,QAAQ,GAA1B,OAAO;QACU,QAAQ;YACC,MAAM,GAAhC,MAAM;YACoB,KAAK,GAA/B,MAAM;;QACJ,QAAQ,MAAM,EAAE,CAAC,CAsD7B;IAED;;;;;;;;;;;;;OAaG;IACH,2BAZW,OAAO,sBAAsB,EAAE,QAAQ,eACvC,MAAM;QAEW,aAAa,GAA9B,MAAM;QAEW,aAAa,GAA9B,MAAM;QAEW,eAAe,GAAhC,MAAM;QACW,8BAA8B,GAA/C,MAAM;YAYhB;CACJ;;kBAnOa,MAAM;oBACN,OAAO,sCAAsC,EAAE,WAAW,EAAE;0BAC5D,OAAO,sCAAsC,EAAE,WAAW,EAAE;mBAC5D,MAAM;mBACN,MAAM;kBACN,MAAM;;+BAhBb,sCAAsC;uBAMtC,uBAAuB"}
1
+ {"version":3,"file":"image_processing_sam.d.ts","sourceRoot":"","sources":["../../../src/models/sam/image_processing_sam.js"],"names":[],"mappings":"AAWA;;;;;;;;GAQG;AAEH;IAEI;;;;;;OAMG;IACH,mCALW,GAAG,kBACH,OAAO,sCAAsC,EAAE,WAAW,EAAE,wBAC5D,OAAO,sCAAsC,EAAE,WAAW,EAAE,8BAC1D,MAAM,CA4ClB;IAED;;;;;OAKG;IACH,+BAJW,GAAG,gBACH,MAAM,GACJ,MAAM,CAoBlB;IACD;;;;;;;;;;;;;;;;;;;OAmBG;IACH,cAnBW,GAAG,EAAE,gDAEb;QAAsB,YAAY,GAA1B,GAAG;QAGW,YAAY,GAA1B,GAAG;QAGoB,WAAW,GAAlC,MAAM,EAAE,EAAE,EAAE;KASpB,GAAU,OAAO,CAAC,uBAAuB,CAAC,CA+B5C;IAED;;;;;;;;;;;;OAYG;IACH,0BAXW,MAAM,kBACN,CAAC,MAAM,EAAE,MAAM,CAAC,EAAE,wBAClB,CAAC,MAAM,EAAE,MAAM,CAAC,EAAE,4CAE1B;QAAyB,cAAc,GAA/B,MAAM;QACY,QAAQ,GAA1B,OAAO;QACU,QAAQ,GACjC;YAAkC,MAAM,GAAhC,MAAM;YACoB,KAAK,GAA/B,MAAM;SACd;KAAA,GAAU,OAAO,CAAC,MAAM,EAAE,CAAC,CAsD7B;IAED;;;;;;;;;;;;;OAaG;IACH,2BAZW,OAAO,sBAAsB,EAAE,QAAQ,eACvC,MAAM,uFAEd;QAAyB,aAAa,GAA9B,MAAM;QAEW,aAAa,GAA9B,MAAM;QAEW,eAAe,GAAhC,MAAM;QACW,8BAA8B,GAA/C,MAAM;KAEd,OAUF;CACJ;;kBAnOa,MAAM;oBACN,OAAO,sCAAsC,EAAE,WAAW,EAAE;0BAC5D,OAAO,sCAAsC,EAAE,WAAW,EAAE;mBAC5D,MAAM;mBACN,MAAM;kBACN,MAAM;;+BAhBb,sCAAsC;uBAMtC,uBAAuB"}
@@ -1,7 +1,7 @@
1
1
  export class SeamlessM4TFeatureExtractor extends FeatureExtractor {
2
2
  constructor(config: any);
3
3
  mel_filters: number[][];
4
- window: Float64Array;
4
+ window: Float64Array<ArrayBufferLike>;
5
5
  /**
6
6
  * Computes the log-Mel spectrogram of the provided audio waveform.
7
7
  * @param {Float32Array|Float64Array} waveform The audio waveform to process.
@@ -1 +1 @@
1
- {"version":3,"file":"feature_extraction_seamless_m4t.d.ts","sourceRoot":"","sources":["../../../src/models/seamless_m4t/feature_extraction_seamless_m4t.js"],"names":[],"mappings":"AAIA;IAEI,yBAwBC;IALG,wBAA8B;IAE9B,qBAEE;IAGN;;;;;OAKG;IACH,kCAJW,YAAY,GAAC,YAAY,cACzB,MAAM,GACJ,QAAQ,MAAM,CAAC,CA6B3B;IAED;;;;;;;;;OASG;IACH,aARW,YAAY,GAAC,YAAY;QAEP,OAAO,GAAzB,OAAO;QACU,kBAAkB,GAAnC,MAAM;QACY,yBAAyB,GAA3C,OAAO;QACW,qBAAqB,GAAvC,OAAO;;wBACqB,MAAM;yBAAmB,MAAM;OAuGrE;CACJ;iCAnLuD,wCAAwC;uBACzE,uBAAuB"}
1
+ {"version":3,"file":"feature_extraction_seamless_m4t.d.ts","sourceRoot":"","sources":["../../../src/models/seamless_m4t/feature_extraction_seamless_m4t.js"],"names":[],"mappings":"AAIA;IAEI,yBAwBC;IALG,wBAA8B;IAE9B,sCAEE;IAGN;;;;;OAKG;IACH,kCAJW,YAAY,GAAC,YAAY,cACzB,MAAM,GACJ,OAAO,CAAC,MAAM,CAAC,CA6B3B;IAED;;;;;;;;;OASG;IACH,aARW,YAAY,GAAC,YAAY,uFAEjC;QAA0B,OAAO,GAAzB,OAAO;QACU,kBAAkB,GAAnC,MAAM;QACY,yBAAyB,GAA3C,OAAO;QACW,qBAAqB,GAAvC,OAAO;KACf,GAAU,OAAO,CAAC;QAAE,cAAc,EAAE,MAAM,CAAC;QAAC,cAAc,CAAC,EAAE,MAAM,CAAA;KAAE,CAAC,CAuGxE;CACJ;iCAnLuD,wCAAwC;uBACzE,uBAAuB"}
@@ -1 +1 @@
1
- {"version":3,"file":"image_processing_segformer.d.ts","sourceRoot":"","sources":["../../../src/models/segformer/image_processing_segformer.js"],"names":[],"mappings":"AAMA;IAOqmK;;;QAA21E;CAF/7O;AACD;CAA0E;+BATnE,sCAAsC"}
1
+ {"version":3,"file":"image_processing_segformer.d.ts","sourceRoot":"","sources":["../../../src/models/segformer/image_processing_segformer.js"],"names":[],"mappings":"AAMA;;;;;CAKC;AACD;CAA0E;+BATnE,sCAAsC"}
@@ -1 +1 @@
1
- {"version":3,"file":"processing_speecht5.d.ts","sourceRoot":"","sources":["../../../src/models/speecht5/processing_speecht5.js"],"names":[],"mappings":"AAIA;IACI,6CAAsC;IACtC,4DAAqD;IAErD;;;;OAIG;IACH,aAHW,GAAG,GACD,QAAQ,GAAG,CAAC,CAIxB;CACJ;0BAhByB,gCAAgC;8BAC5B,qBAAqB;qCACd,oCAAoC"}
1
+ {"version":3,"file":"processing_speecht5.d.ts","sourceRoot":"","sources":["../../../src/models/speecht5/processing_speecht5.js"],"names":[],"mappings":"AAIA;IACI,6CAAsC;IACtC,4DAAqD;IAErD;;;;OAIG;IACH,aAHW,GAAG,GACD,OAAO,CAAC,GAAG,CAAC,CAIxB;CACJ;0BAhByB,gCAAgC;8BAC5B,qBAAqB;qCACd,oCAAoC"}
@@ -1,5 +1,5 @@
1
1
  export class Swin2SRImageProcessor extends ImageProcessor {
2
- pad_image(pixelData: any, imgDims: any, padSize: any, options?: {}): [Float32Array, number[]];
2
+ pad_image(pixelData: any, imgDims: any, padSize: any, options?: {}): [Float32Array<ArrayBufferLike>, number[]];
3
3
  }
4
4
  import { ImageProcessor } from "../../base/image_processors_utils.js";
5
5
  //# sourceMappingURL=image_processing_swin2sr.d.ts.map
@@ -1 +1 @@
1
- {"version":3,"file":"image_processing_swin2sr.d.ts","sourceRoot":"","sources":["../../../src/models/swin2sr/image_processing_swin2sr.js"],"names":[],"mappings":"AAIA;IACI,8FAiBC;CACJ;+BArBM,sCAAsC"}
1
+ {"version":3,"file":"image_processing_swin2sr.d.ts","sourceRoot":"","sources":["../../../src/models/swin2sr/image_processing_swin2sr.js"],"names":[],"mappings":"AAIA;IACI,+GAiBC;CACJ;+BArBM,sCAAsC"}
@@ -1 +1 @@
1
- {"version":3,"file":"image_processing_vitmatte.d.ts","sourceRoot":"","sources":["../../../src/models/vitmatte/image_processing_vitmatte.js"],"names":[],"mappings":"AASA;IACI;;;;;;OAMG;IACH,cAJW,OAAO,sBAAsB,EAAE,QAAQ,EAAE,WACzC,OAAO,sBAAsB,EAAE,QAAQ,EAAE,GACvC,QAAQ,OAAO,sCAAsC,EAAE,oBAAoB,CAAC,CAiCxF;CACJ;+BA/CM,sCAAsC"}
1
+ {"version":3,"file":"image_processing_vitmatte.d.ts","sourceRoot":"","sources":["../../../src/models/vitmatte/image_processing_vitmatte.js"],"names":[],"mappings":"AASA;IACI;;;;;;OAMG;IACH,cAJW,OAAO,sBAAsB,EAAE,QAAQ,EAAE,WACzC,OAAO,sBAAsB,EAAE,QAAQ,EAAE,GACvC,OAAO,CAAC,OAAO,sCAAsC,EAAE,oBAAoB,CAAC,CAiCxF;CACJ;+BA/CM,sCAAsC"}
@@ -13,7 +13,7 @@ export class VitPoseImageProcessor extends ImageProcessor {
13
13
  * keypoints: [number, number][]
14
14
  * }[][]} List of keypoints predictions for each image.
15
15
  */
16
- post_process_pose_estimation(outputs: import('../../utils/tensor.js').Tensor, boxes: [number, number, number, number][][], { threshold, }?: {
16
+ post_process_pose_estimation(outputs: import("../../utils/tensor.js").Tensor, boxes: [number, number, number, number][][], { threshold, }?: {
17
17
  threshold?: any;
18
18
  }): {
19
19
  bbox: [number, number, number, number];
@@ -1 +1 @@
1
- {"version":3,"file":"image_processing_vitpose.d.ts","sourceRoot":"","sources":["../../../src/models/vitpose/image_processing_vitpose.js"],"names":[],"mappings":"AAIA;IAEI;;;;;;;;;;;;;OAaG;IACH,sCAVW,OAAO,uBAAuB,EAAE,MAAM,SACtC,CAAC,MAAM,EAAE,MAAM,EAAE,MAAM,EAAE,MAAM,CAAC,EAAE,EAAE;;QAElC;QACZ,IAAQ,EAAE,CAAC,MAAM,EAAE,MAAM,EAAE,MAAM,EAAE,MAAM,CAAC,CAAC;QAC3C,MAAU,EAAE,MAAM,EAAE,CAAC;QACrB,MAAU,EAAE,MAAM,EAAE,CAAC;QACrB,SAAa,EAAE,CAAC,MAAM,EAAE,MAAM,CAAC,EAAE,CAAA;KAC9B,EAAE,EAAE,CAqEP;CACJ;+BAtFM,sCAAsC"}
1
+ {"version":3,"file":"image_processing_vitpose.d.ts","sourceRoot":"","sources":["../../../src/models/vitpose/image_processing_vitpose.js"],"names":[],"mappings":"AAIA;IAEI;;;;;;;;;;;;;OAaG;IACH,sCAVW,OAAO,uBAAuB,EAAE,MAAM,SACtC,CAAC,MAAM,EAAE,MAAM,EAAE,MAAM,EAAE,MAAM,CAAC,EAAE,EAAE;;QAElC;QACR,IAAI,EAAE,CAAC,MAAM,EAAE,MAAM,EAAE,MAAM,EAAE,MAAM,CAAC,CAAC;QACvC,MAAM,EAAE,MAAM,EAAE,CAAC;QACjB,MAAM,EAAE,MAAM,EAAE,CAAC;QACjB,SAAS,EAAE,CAAC,MAAM,EAAE,MAAM,CAAC,EAAE,CAAA;KAC9B,EAAE,EAAE,CAqEP;CACJ;+BAtFM,sCAAsC"}
@@ -1 +1 @@
1
- {"version":3,"file":"feature_extraction_wav2vec2.d.ts","sourceRoot":"","sources":["../../../src/models/wav2vec2/feature_extraction_wav2vec2.js"],"names":[],"mappings":"AAGA;IAEI;;;OAGG;IACH,uCAHW,YAAY,GACV,YAAY,CAQxB;IAED;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY;sBACC,MAAM;wBAAkB,MAAM;OAsBlE;CACJ;iCA3CuD,wCAAwC;uBACzE,uBAAuB"}
1
+ {"version":3,"file":"feature_extraction_wav2vec2.d.ts","sourceRoot":"","sources":["../../../src/models/wav2vec2/feature_extraction_wav2vec2.js"],"names":[],"mappings":"AAGA;IAEI;;;OAGG;IACH,uCAHW,YAAY,GACV,YAAY,CAQxB;IAED;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY,GACvB,OAAO,CAAC;QAAE,YAAY,EAAE,MAAM,CAAC;QAAC,cAAc,EAAE,MAAM,CAAA;KAAE,CAAC,CAsBrE;CACJ;iCA3CuD,wCAAwC;uBACzE,uBAAuB"}
@@ -1 +1 @@
1
- {"version":3,"file":"processing_wav2vec2.d.ts","sourceRoot":"","sources":["../../../src/models/wav2vec2/processing_wav2vec2.js"],"names":[],"mappings":"AAGA;IACI,4DAAqD;IAErD;;;;OAIG;IACH,aAHW,GAAG,GACD,QAAQ,GAAG,CAAC,CAIxB;CACJ;0BAdyB,gCAAgC;qCACrB,oCAAoC"}
1
+ {"version":3,"file":"processing_wav2vec2.d.ts","sourceRoot":"","sources":["../../../src/models/wav2vec2/processing_wav2vec2.js"],"names":[],"mappings":"AAGA;IACI,4DAAqD;IAErD;;;;OAIG;IACH,aAHW,GAAG,GACD,OAAO,CAAC,GAAG,CAAC,CAIxB;CACJ;0BAdyB,gCAAgC;qCACrB,oCAAoC"}
@@ -1,7 +1,7 @@
1
1
  export class WeSpeakerFeatureExtractor extends FeatureExtractor {
2
2
  constructor(config: any);
3
3
  mel_filters: number[][];
4
- window: Float64Array;
4
+ window: Float64Array<ArrayBufferLike>;
5
5
  min_num_frames: any;
6
6
  /**
7
7
  * Computes the log-Mel spectrogram of the provided audio waveform.
@@ -1 +1 @@
1
- {"version":3,"file":"feature_extraction_wespeaker.d.ts","sourceRoot":"","sources":["../../../src/models/wespeaker/feature_extraction_wespeaker.js"],"names":[],"mappings":"AAKA;IAEI,yBAyBC;IANG,wBAA8B;IAE9B,qBAEE;IACF,oBAAgD;IAGpD;;;;OAIG;IACH,kCAHW,YAAY,GAAC,YAAY,GACvB,QAAQ,MAAM,CAAC,CA2B3B;IAGD;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY,GACvB,QAAQ;QAAE,cAAc,EAAE,MAAM,CAAA;KAAE,CAAC,CA4B/C;CACJ;iCAnGuD,wCAAwC;uBACzE,uBAAuB"}
1
+ {"version":3,"file":"feature_extraction_wespeaker.d.ts","sourceRoot":"","sources":["../../../src/models/wespeaker/feature_extraction_wespeaker.js"],"names":[],"mappings":"AAKA;IAEI,yBAyBC;IANG,wBAA8B;IAE9B,sCAEE;IACF,oBAAgD;IAGpD;;;;OAIG;IACH,kCAHW,YAAY,GAAC,YAAY,GACvB,OAAO,CAAC,MAAM,CAAC,CA2B3B;IAGD;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY,GACvB,OAAO,CAAC;QAAE,cAAc,EAAE,MAAM,CAAA;KAAE,CAAC,CA4B/C;CACJ;iCAnGuD,wCAAwC;uBACzE,uBAAuB"}
@@ -1,6 +1,6 @@
1
1
  export class WhisperFeatureExtractor extends FeatureExtractor {
2
2
  constructor(config: any);
3
- window: Float64Array;
3
+ window: Float64Array<ArrayBufferLike>;
4
4
  /**
5
5
  * Computes the log-Mel spectrogram of the provided audio waveform.
6
6
  * @param {Float32Array|Float64Array} waveform The audio waveform to process.
@@ -1 +1 @@
1
- {"version":3,"file":"feature_extraction_whisper.d.ts","sourceRoot":"","sources":["../../../src/models/whisper/feature_extraction_whisper.js"],"names":[],"mappings":"AAKA;IAEI,yBAeC;IADG,qBAAwD;IAG5D;;;;OAIG;IACH,kCAHW,YAAY,GAAC,YAAY,GACvB,QAAQ,MAAM,CAAC,CA0B3B;IAED;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY,GACvB,QAAQ;QAAE,cAAc,EAAE,MAAM,CAAA;KAAE,CAAC,CAwB/C;CACJ;iCAnFuD,wCAAwC;uBACzE,uBAAuB"}
1
+ {"version":3,"file":"feature_extraction_whisper.d.ts","sourceRoot":"","sources":["../../../src/models/whisper/feature_extraction_whisper.js"],"names":[],"mappings":"AAKA;IAEI,yBAeC;IADG,sCAAwD;IAG5D;;;;OAIG;IACH,kCAHW,YAAY,GAAC,YAAY,GACvB,OAAO,CAAC,MAAM,CAAC,CA0B3B;IAED;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY,GACvB,OAAO,CAAC;QAAE,cAAc,EAAE,MAAM,CAAA;KAAE,CAAC,CAwB/C;CACJ;iCAnFuD,wCAAwC;uBACzE,uBAAuB"}
@@ -1 +1 @@
1
- {"version":3,"file":"generation_whisper.d.ts","sourceRoot":"","sources":["../../../src/models/whisper/generation_whisper.js"],"names":[],"mappings":"AAEA;IAEI;;;OAGG;IACH,mBAFU,OAAO,CAEQ;IAEzB;;;;;OAKG;IACH,yBAFU,OAAO,CAEc;IAE/B;;;OAGG;IACH,YAFU,MAAM,CAEE;IAElB;;;;OAIG;IACH,iBAFU,CAAC,MAAM,EAAE,MAAM,CAAC,EAAE,CAEL;IAEvB;;;OAGG;IACH,MAFU,MAAM,CAEJ;IAEZ;;;;OAIG;IACH,UAFU,MAAM,CAEA;IAEhB;;;OAGG;IACH,wBAFU,MAAM,CAEc;IAE9B;;;;;;OAMG;IACH,YAFU,MAAM,EAAE,CAEA;IAElB;;;OAGG;IACH,iBAFU,OAAO,CAEM;IAEvB;;;;OAIG;IACH,YAFU,OAAO,MAAM,EAAE,MAAM,CAAC,GAAC,IAAI,CAEnB;IAElB;;;OAGG;IACH,YAFU,OAAO,MAAM,EAAE,MAAM,CAAC,GAAC,IAAI,CAEnB;IAElB;;;;OAIG;IACH,6BAFU,MAAM,CAEgB;CACnC;kDAGY,MAAwE;IAAC,iBAAiB,EAAE,uBAAuB,CAAA;CAAC,GAAG,uBAAuB;iCAvF1H,yCAAyC"}
1
+ {"version":3,"file":"generation_whisper.d.ts","sourceRoot":"","sources":["../../../src/models/whisper/generation_whisper.js"],"names":[],"mappings":"AAEA;IAEI;;;OAGG;IACH,mBAFU,OAAO,CAEQ;IAEzB;;;;;OAKG;IACH,yBAFU,OAAO,CAEc;IAE/B;;;OAGG;IACH,YAFU,MAAM,CAEE;IAElB;;;;OAIG;IACH,iBAFU,CAAC,MAAM,EAAE,MAAM,CAAC,EAAE,CAEL;IAEvB;;;OAGG;IACH,MAFU,MAAM,CAEJ;IAEZ;;;;OAIG;IACH,UAFU,MAAM,CAEA;IAEhB;;;OAGG;IACH,wBAFU,MAAM,CAEc;IAE9B;;;;;;OAMG;IACH,YAFU,MAAM,EAAE,CAEA;IAElB;;;OAGG;IACH,iBAFU,OAAO,CAEM;IAEvB;;;;OAIG;IACH,YAFU,MAAM,CAAC,MAAM,EAAE,MAAM,CAAC,GAAC,IAAI,CAEnB;IAElB;;;OAGG;IACH,YAFU,MAAM,CAAC,MAAM,EAAE,MAAM,CAAC,GAAC,IAAI,CAEnB;IAElB;;;;OAIG;IACH,6BAFU,MAAM,CAEgB;CACnC;kDAGY,GAAqE,GAAG;IAAC,iBAAiB,EAAE,uBAAuB,CAAA;CAAC,GAAG,uBAAuB;iCAvF1H,yCAAyC"}
@@ -1 +1 @@
1
- {"version":3,"file":"processing_whisper.d.ts","sourceRoot":"","sources":["../../../src/models/whisper/processing_whisper.js"],"names":[],"mappings":"AAIA;;GAEG;AACH;IACI,6CAAsC;IACtC,4DAAqD;IAErD;;;;OAIG;IACH,aAHW,GAAG,GACD,QAAQ,GAAG,CAAC,CAIxB;CACJ;0BAjByB,gCAAgC;8BAD5B,qBAAqB;qCADd,oCAAoC"}
1
+ {"version":3,"file":"processing_whisper.d.ts","sourceRoot":"","sources":["../../../src/models/whisper/processing_whisper.js"],"names":[],"mappings":"AAIA;;GAEG;AACH;IACI,6CAAsC;IACtC,4DAAqD;IAErD;;;;OAIG;IACH,aAHW,GAAG,GACD,OAAO,CAAC,GAAG,CAAC,CAIxB;CACJ;0BAjByB,gCAAgC;8BAD5B,qBAAqB;qCADd,oCAAoC"}
@@ -1 +1 @@
1
- {"version":3,"file":"image_processing_yolos.d.ts","sourceRoot":"","sources":["../../../src/models/yolos/image_processing_yolos.js"],"names":[],"mappings":"AAKA;IAO+tE;;;6FAAu6F;CAFroK;AACD;CAAkE;+BAR3D,sCAAsC"}
1
+ {"version":3,"file":"image_processing_yolos.d.ts","sourceRoot":"","sources":["../../../src/models/yolos/image_processing_yolos.js"],"names":[],"mappings":"AAKA;;;;;CAKC;AACD;CAAkE;+BAR3D,sCAAsC"}