@huggingface/transformers 3.1.1 → 3.2.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +10 -4
- package/dist/ort-wasm-simd-threaded.jsep.wasm +0 -0
- package/dist/transformers.cjs +1062 -183
- package/dist/transformers.cjs.map +1 -1
- package/dist/transformers.js +2239 -1232
- package/dist/transformers.js.map +1 -1
- package/dist/transformers.min.cjs +1 -358
- package/dist/transformers.min.cjs.map +1 -1
- package/dist/transformers.min.js +1 -421
- package/dist/transformers.min.js.map +1 -1
- package/dist/transformers.min.mjs +1 -358
- package/dist/transformers.min.mjs.map +1 -1
- package/dist/transformers.mjs +1082 -181
- package/dist/transformers.mjs.map +1 -1
- package/package.json +11 -16
- package/src/backends/onnx.js +2 -7
- package/src/base/image_processors_utils.js +3 -1
- package/src/configs.js +11 -2
- package/src/env.js +1 -1
- package/src/models/feature_extractors.js +1 -0
- package/src/models/idefics3/image_processing_idefics3.js +24 -13
- package/src/models/image_processors.js +1 -0
- package/src/models/moonshine/feature_extraction_moonshine.js +26 -0
- package/src/models/moonshine/processing_moonshine.js +20 -0
- package/src/models/paligemma/processing_paligemma.js +82 -0
- package/src/models/phi3_v/image_processing_phi3_v.js +163 -0
- package/src/models/phi3_v/processing_phi3_v.js +53 -0
- package/src/models/processors.js +3 -0
- package/src/models/pyannote/feature_extraction_pyannote.js +56 -0
- package/src/models/pyannote/processing_pyannote.js +7 -54
- package/src/models.js +233 -35
- package/src/ops/registry.js +11 -0
- package/src/pipelines.js +30 -0
- package/src/tokenizers.js +12 -1
- package/src/utils/core.js +39 -9
- package/src/utils/hub.js +8 -12
- package/src/utils/image.js +40 -0
- package/src/utils/tensor.js +51 -1
- package/types/backends/onnx.d.ts +2 -2
- package/types/backends/onnx.d.ts.map +1 -1
- package/types/base/feature_extraction_utils.d.ts +1 -1
- package/types/base/feature_extraction_utils.d.ts.map +1 -1
- package/types/base/image_processors_utils.d.ts +4 -4
- package/types/base/image_processors_utils.d.ts.map +1 -1
- package/types/base/processing_utils.d.ts +4 -4
- package/types/base/processing_utils.d.ts.map +1 -1
- package/types/configs.d.ts +7 -7
- package/types/configs.d.ts.map +1 -1
- package/types/env.d.ts +1 -1
- package/types/env.d.ts.map +1 -1
- package/types/generation/configuration_utils.d.ts +2 -2
- package/types/generation/logits_process.d.ts +2 -2
- package/types/generation/logits_process.d.ts.map +1 -1
- package/types/generation/logits_sampler.d.ts.map +1 -1
- package/types/generation/parameters.d.ts +5 -5
- package/types/generation/stopping_criteria.d.ts +1 -1
- package/types/generation/stopping_criteria.d.ts.map +1 -1
- package/types/generation/streamers.d.ts +2 -2
- package/types/generation/streamers.d.ts.map +1 -1
- package/types/models/audio_spectrogram_transformer/feature_extraction_audio_spectrogram_transformer.d.ts +1 -1
- package/types/models/audio_spectrogram_transformer/feature_extraction_audio_spectrogram_transformer.d.ts.map +1 -1
- package/types/models/auto/feature_extraction_auto.d.ts.map +1 -1
- package/types/models/auto/image_processing_auto.d.ts.map +1 -1
- package/types/models/auto/processing_auto.d.ts +1 -1
- package/types/models/auto/processing_auto.d.ts.map +1 -1
- package/types/models/clap/feature_extraction_clap.d.ts +1 -1
- package/types/models/clap/feature_extraction_clap.d.ts.map +1 -1
- package/types/models/detr/image_processing_detr.d.ts +11 -11
- package/types/models/detr/image_processing_detr.d.ts.map +1 -1
- package/types/models/donut/image_processing_donut.d.ts +1 -1
- package/types/models/donut/image_processing_donut.d.ts.map +1 -1
- package/types/models/feature_extractors.d.ts +1 -0
- package/types/models/florence2/processing_florence2.d.ts.map +1 -1
- package/types/models/idefics3/image_processing_idefics3.d.ts.map +1 -1
- package/types/models/idefics3/processing_idefics3.d.ts.map +1 -1
- package/types/models/image_processors.d.ts +1 -0
- package/types/models/janus/image_processing_janus.d.ts +1 -1
- package/types/models/janus/image_processing_janus.d.ts.map +1 -1
- package/types/models/janus/processing_janus.d.ts.map +1 -1
- package/types/models/maskformer/image_processing_maskformer.d.ts +8 -8
- package/types/models/maskformer/image_processing_maskformer.d.ts.map +1 -1
- package/types/models/mgp_str/processing_mgp_str.d.ts +2 -2
- package/types/models/mgp_str/processing_mgp_str.d.ts.map +1 -1
- package/types/models/moonshine/feature_extraction_moonshine.d.ts +13 -0
- package/types/models/moonshine/feature_extraction_moonshine.d.ts.map +1 -0
- package/types/models/moonshine/processing_moonshine.d.ts +17 -0
- package/types/models/moonshine/processing_moonshine.d.ts.map +1 -0
- package/types/models/owlvit/image_processing_owlvit.d.ts.map +1 -1
- package/types/models/paligemma/processing_paligemma.d.ts +12 -0
- package/types/models/paligemma/processing_paligemma.d.ts.map +1 -0
- package/types/models/phi3_v/image_processing_phi3_v.d.ts +17 -0
- package/types/models/phi3_v/image_processing_phi3_v.d.ts.map +1 -0
- package/types/models/phi3_v/processing_phi3_v.d.ts +17 -0
- package/types/models/phi3_v/processing_phi3_v.d.ts.map +1 -0
- package/types/models/processors.d.ts +3 -0
- package/types/models/pyannote/feature_extraction_pyannote.d.ts +18 -0
- package/types/models/pyannote/feature_extraction_pyannote.d.ts.map +1 -1
- package/types/models/pyannote/processing_pyannote.d.ts +4 -15
- package/types/models/pyannote/processing_pyannote.d.ts.map +1 -1
- package/types/models/qwen2_vl/processing_qwen2_vl.d.ts.map +1 -1
- package/types/models/rt_detr/image_processing_rt_detr.d.ts.map +1 -1
- package/types/models/sam/image_processing_sam.d.ts.map +1 -1
- package/types/models/seamless_m4t/feature_extraction_seamless_m4t.d.ts +1 -1
- package/types/models/seamless_m4t/feature_extraction_seamless_m4t.d.ts.map +1 -1
- package/types/models/segformer/image_processing_segformer.d.ts.map +1 -1
- package/types/models/speecht5/processing_speecht5.d.ts.map +1 -1
- package/types/models/swin2sr/image_processing_swin2sr.d.ts +1 -1
- package/types/models/swin2sr/image_processing_swin2sr.d.ts.map +1 -1
- package/types/models/vitmatte/image_processing_vitmatte.d.ts.map +1 -1
- package/types/models/vitpose/image_processing_vitpose.d.ts +1 -1
- package/types/models/vitpose/image_processing_vitpose.d.ts.map +1 -1
- package/types/models/wav2vec2/feature_extraction_wav2vec2.d.ts.map +1 -1
- package/types/models/wav2vec2/processing_wav2vec2.d.ts.map +1 -1
- package/types/models/wespeaker/feature_extraction_wespeaker.d.ts +1 -1
- package/types/models/wespeaker/feature_extraction_wespeaker.d.ts.map +1 -1
- package/types/models/whisper/feature_extraction_whisper.d.ts +1 -1
- package/types/models/whisper/feature_extraction_whisper.d.ts.map +1 -1
- package/types/models/whisper/generation_whisper.d.ts.map +1 -1
- package/types/models/whisper/processing_whisper.d.ts.map +1 -1
- package/types/models/yolos/image_processing_yolos.d.ts.map +1 -1
- package/types/models.d.ts +61 -5
- package/types/models.d.ts.map +1 -1
- package/types/ops/registry.d.ts +1 -0
- package/types/ops/registry.d.ts.map +1 -1
- package/types/pipelines.d.ts +31 -51
- package/types/pipelines.d.ts.map +1 -1
- package/types/tokenizers.d.ts +10 -6
- package/types/tokenizers.d.ts.map +1 -1
- package/types/utils/audio.d.ts.map +1 -1
- package/types/utils/constants.d.ts.map +1 -1
- package/types/utils/core.d.ts +87 -22
- package/types/utils/core.d.ts.map +1 -1
- package/types/utils/data-structures.d.ts.map +1 -1
- package/types/utils/devices.d.ts.map +1 -1
- package/types/utils/dtypes.d.ts.map +1 -1
- package/types/utils/generic.d.ts.map +1 -1
- package/types/utils/hub.d.ts +3 -3
- package/types/utils/hub.d.ts.map +1 -1
- package/types/utils/image.d.ts +10 -1
- package/types/utils/image.d.ts.map +1 -1
- package/types/utils/maths.d.ts +10 -10
- package/types/utils/maths.d.ts.map +1 -1
- package/types/utils/tensor.d.ts +22 -6
- package/types/utils/tensor.d.ts.map +1 -1
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"feature_extraction_clap.d.ts","sourceRoot":"","sources":["../../../src/models/clap/feature_extraction_clap.js"],"names":[],"mappings":"AAKA;IAEI,yBAyBC;IAtBG,wBAQC;IAED,+BAQC;IAED,
|
|
1
|
+
{"version":3,"file":"feature_extraction_clap.d.ts","sourceRoot":"","sources":["../../../src/models/clap/feature_extraction_clap.js"],"names":[],"mappings":"AAKA;IAEI,yBAyBC;IAtBG,wBAQC;IAED,+BAQC;IAED,sCAAkE;IAKtE;;;;;;;;;;;;;;;;;;;;OAoBG;IACH,uBA0CC;IAED;;;;;;;;;;;;;;OAcG;IACH,kCALW,YAAY,GAAC,YAAY,eACzB,MAAM,EAAE,EAAE,eACV,MAAM,GACJ,OAAO,CAAC,MAAM,CAAC,CAoB3B;IAGD;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY;;QACvB,OAAO,CAAC;QAAE,cAAc,EAAE,MAAM,CAAA;KAAE,CAAC,CAkB/C;CACJ;iCA9JuD,wCAAwC;uBACzE,uBAAuB"}
|
|
@@ -10,33 +10,33 @@ export class DetrImageProcessor extends ImageProcessor {
|
|
|
10
10
|
* @param {import('../../utils/image.js').RawImage[]} images The image(s) to extract features from.
|
|
11
11
|
* @returns {Promise<DetrFeatureExtractorResult>} An object containing the concatenated pixel values of the preprocessed images.
|
|
12
12
|
*/
|
|
13
|
-
_call(images: import(
|
|
13
|
+
_call(images: import("../../utils/image.js").RawImage[]): Promise<DetrFeatureExtractorResult>;
|
|
14
14
|
post_process_object_detection(outputs: {
|
|
15
15
|
logits: import("../../utils/tensor.js").Tensor;
|
|
16
16
|
pred_boxes: import("../../utils/tensor.js").Tensor;
|
|
17
17
|
}, threshold?: number, target_sizes?: [number, number][], is_zero_shot?: boolean): any[];
|
|
18
|
-
post_process_panoptic_segmentation(outputs: any, threshold?: number, mask_threshold?: number, overlap_mask_area_threshold?: number, label_ids_to_fuse?: Set<number>, target_sizes?: [number, number][]): {
|
|
18
|
+
post_process_panoptic_segmentation(outputs: any, threshold?: number, mask_threshold?: number, overlap_mask_area_threshold?: number, label_ids_to_fuse?: Set<number>, target_sizes?: [number, number][]): Array<{
|
|
19
19
|
segmentation: import("../../utils/tensor.js").Tensor;
|
|
20
|
-
segments_info: {
|
|
20
|
+
segments_info: Array<{
|
|
21
21
|
id: number;
|
|
22
22
|
label_id: number;
|
|
23
23
|
score: number;
|
|
24
|
-
}
|
|
25
|
-
}
|
|
26
|
-
post_process_instance_segmentation(outputs: any, threshold?: number, target_sizes?: [number, number][]): {
|
|
24
|
+
}>;
|
|
25
|
+
}>;
|
|
26
|
+
post_process_instance_segmentation(outputs: any, threshold?: number, target_sizes?: [number, number][]): Array<{
|
|
27
27
|
segmentation: import("../../utils/tensor.js").Tensor;
|
|
28
|
-
segments_info: {
|
|
28
|
+
segments_info: Array<{
|
|
29
29
|
id: number;
|
|
30
30
|
label_id: number;
|
|
31
31
|
score: number;
|
|
32
|
-
}
|
|
33
|
-
}
|
|
32
|
+
}>;
|
|
33
|
+
}>;
|
|
34
34
|
}
|
|
35
35
|
export class DetrFeatureExtractor extends DetrImageProcessor {
|
|
36
36
|
}
|
|
37
37
|
export type DetrFeatureExtractorResultProps = {
|
|
38
|
-
pixel_mask: import(
|
|
38
|
+
pixel_mask: import("../../utils/tensor.js").Tensor;
|
|
39
39
|
};
|
|
40
|
-
export type DetrFeatureExtractorResult = import(
|
|
40
|
+
export type DetrFeatureExtractorResult = import("../../base/image_processors_utils.js").ImageProcessorResult & DetrFeatureExtractorResultProps;
|
|
41
41
|
import { ImageProcessor } from "../../base/image_processors_utils.js";
|
|
42
42
|
//# sourceMappingURL=image_processing_detr.d.ts.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"image_processing_detr.d.ts","sourceRoot":"","sources":["../../../src/models/detr/image_processing_detr.js"],"names":[],"mappings":"AAUA;;;;GAIG;AAEH;IACI;;;;;OAKG;IACH,cAHW,OAAO,sBAAsB,EAAE,QAAQ,EAAE,GACvC,
|
|
1
|
+
{"version":3,"file":"image_processing_detr.d.ts","sourceRoot":"","sources":["../../../src/models/detr/image_processing_detr.js"],"names":[],"mappings":"AAUA;;;;GAIG;AAEH;IACI;;;;;OAKG;IACH,cAHW,OAAO,sBAAsB,EAAE,QAAQ,EAAE,GACvC,OAAO,CAAC,0BAA0B,CAAC,CAY/C;;;;;;;;;;;;;;;;;;;;;CAgBJ;AAED;CAAgE;;gBAvClD,OAAO,uBAAuB,EAAE,MAAM;;yCACvC,OAAO,sCAAsC,EAAE,oBAAoB,GAAG,+BAA+B;+BAR3G,sCAAsC"}
|
|
@@ -1,5 +1,5 @@
|
|
|
1
1
|
export class DonutImageProcessor extends ImageProcessor {
|
|
2
|
-
pad_image(pixelData: any, imgDims: any, padSize: any, options?: {}): [Float32Array
|
|
2
|
+
pad_image(pixelData: any, imgDims: any, padSize: any, options?: {}): [Float32Array<ArrayBufferLike>, number[]];
|
|
3
3
|
}
|
|
4
4
|
export class DonutFeatureExtractor extends DonutImageProcessor {
|
|
5
5
|
}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"image_processing_donut.d.ts","sourceRoot":"","sources":["../../../src/models/donut/image_processing_donut.js"],"names":[],"mappings":"AAIA;IACI
|
|
1
|
+
{"version":3,"file":"image_processing_donut.d.ts","sourceRoot":"","sources":["../../../src/models/donut/image_processing_donut.js"],"names":[],"mappings":"AAIA;IACI,+GAuBC;CACJ;AACD;CAAkE;+BA5B3D,sCAAsC"}
|
|
@@ -1,5 +1,6 @@
|
|
|
1
1
|
export * from "./audio_spectrogram_transformer/feature_extraction_audio_spectrogram_transformer.js";
|
|
2
2
|
export * from "./clap/feature_extraction_clap.js";
|
|
3
|
+
export * from "./moonshine/feature_extraction_moonshine.js";
|
|
3
4
|
export * from "./pyannote/feature_extraction_pyannote.js";
|
|
4
5
|
export * from "./seamless_m4t/feature_extraction_seamless_m4t.js";
|
|
5
6
|
export * from "./speecht5/feature_extraction_speecht5.js";
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"processing_florence2.d.ts","sourceRoot":"","sources":["../../../src/models/florence2/processing_florence2.js"],"names":[],"mappings":"AAIA;IACI,6CAAsC;IACtC,wDAAiD;IAEjD,0CAuBC;IAdG,kCAAkC;IAClC,mCADW,
|
|
1
|
+
{"version":3,"file":"processing_florence2.d.ts","sourceRoot":"","sources":["../../../src/models/florence2/processing_florence2.js"],"names":[],"mappings":"AAIA;IACI,6CAAsC;IACtC,wDAAiD;IAEjD,0CAuBC;IAdG,kCAAkC;IAClC,mCADW,GAAG,CAAC,MAAM,EAAE,MAAM,CAAC,CAC2E;IAEzG,kCAAkC;IAClC,6BADW,GAAG,CAAC,MAAM,EAAE,MAAM,CAAC,CAC+D;IAE7F,kCAAkC;IAClC,yBADW,GAAG,CAAC,MAAM,EAAE,MAAM,CAAC,CACuD;IAErF;;;MAGC;IACD,qBAAwB;IAG5B;;;;OAIG;IACH,wBAHW,MAAM,GAAC,MAAM,EAAE,GACb,MAAM,EAAE,CA6BpB;IAED;;;;;OAKG;IACH,8BAJW,MAAM,QACN,MAAM,cACN,CAAC,MAAM,EAAE,MAAM,CAAC;;;;;MAsC1B;IAID,0DAaC;CACJ;0BA/HyB,gCAAgC;8BAE5B,qBAAqB;mCADhB,kCAAkC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"image_processing_idefics3.d.ts","sourceRoot":"","sources":["../../../src/models/idefics3/image_processing_idefics3.js"],"names":[],"mappings":"AAOA;IACI,yBAKC;IAFG,wBAA2D;IAC3D,oBAA2C;IAG/C;;;OAGG;IAEH;;;;;OAKG;IACH,6GAHW,MAAM;;;MAiBhB;IAED,uDAAuD;IACvD;;;;;;;;;;
|
|
1
|
+
{"version":3,"file":"image_processing_idefics3.d.ts","sourceRoot":"","sources":["../../../src/models/idefics3/image_processing_idefics3.js"],"names":[],"mappings":"AAOA;IACI,yBAKC;IAFG,wBAA2D;IAC3D,oBAA2C;IAG/C;;;OAGG;IAEH;;;;;OAKG;IACH,6GAHW,MAAM;;;MAiBhB;IAED,uDAAuD;IACvD,cADY,0CAAS,yCAAU,GAAC,yCAAU,EAAE;;;;;;;;;;OA4H3C;IAED;;;;;;OA4DC;CACJ;+BAjOM,sCAAsC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"processing_idefics3.d.ts","sourceRoot":"","sources":["../../../src/models/idefics3/processing_idefics3.js"],"names":[],"mappings":"AA6DA;IACI,wDAAiD;IACjD,6CAAsC;IAGtC,yBAA+C;IAC/C,oBAAwB;IACxB,yBAAkC;IAElC;;;;;OAKG;IACH,YAJW,MAAM,GAAC,MAAM,EAAE,WACf,QAAQ,GAAC,QAAQ,EAAE,GAAC,QAAQ,EAAE,EAAE,iBAC9B,
|
|
1
|
+
{"version":3,"file":"processing_idefics3.d.ts","sourceRoot":"","sources":["../../../src/models/idefics3/processing_idefics3.js"],"names":[],"mappings":"AA6DA;IACI,wDAAiD;IACjD,6CAAsC;IAGtC,yBAA+C;IAC/C,oBAAwB;IACxB,yBAAkC;IAElC;;;;;OAKG;IACH,YAJW,MAAM,GAAC,MAAM,EAAE,WACf,QAAQ,GAAC,QAAQ,EAAE,GAAC,QAAQ,EAAE,EAAE,iBAC9B,OAAO,CAAC,GAAG,CAAC,CA4DxB;CACJ;0BAtIyB,gCAAgC;yBAGjC,sBAAsB;mCAFZ,kCAAkC;8BACvC,qBAAqB"}
|
|
@@ -23,6 +23,7 @@ export * from "./mobilevit/image_processing_mobilevit.js";
|
|
|
23
23
|
export * from "./nougat/image_processing_nougat.js";
|
|
24
24
|
export * from "./owlv2/image_processing_owlv2.js";
|
|
25
25
|
export * from "./owlvit/image_processing_owlvit.js";
|
|
26
|
+
export * from "./phi3_v/image_processing_phi3_v.js";
|
|
26
27
|
export * from "./pvt/image_processing_pvt.js";
|
|
27
28
|
export * from "./qwen2_vl/image_processing_qwen2_vl.js";
|
|
28
29
|
export * from "./rt_detr/image_processing_rt_detr.js";
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
export class VLMImageProcessor extends ImageProcessor {
|
|
2
2
|
constructor(config: any);
|
|
3
3
|
constant_values: any;
|
|
4
|
-
pad_image(pixelData: any, imgDims: any, padSize: any, options: any): [Float32Array
|
|
4
|
+
pad_image(pixelData: any, imgDims: any, padSize: any, options: any): [Float32Array<ArrayBufferLike>, number[]];
|
|
5
5
|
}
|
|
6
6
|
import { ImageProcessor } from "../../base/image_processors_utils.js";
|
|
7
7
|
//# sourceMappingURL=image_processing_janus.d.ts.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"image_processing_janus.d.ts","sourceRoot":"","sources":["../../../src/models/janus/image_processing_janus.js"],"names":[],"mappings":"AAKA;IACI,yBAUC;IADG,qBAAqF;IAGzF
|
|
1
|
+
{"version":3,"file":"image_processing_janus.d.ts","sourceRoot":"","sources":["../../../src/models/janus/image_processing_janus.js"],"names":[],"mappings":"AAKA;IACI,yBAUC;IADG,qBAAqF;IAGzF,+GAMC;CACJ;+BAtBM,sCAAsC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"processing_janus.d.ts","sourceRoot":"","sources":["../../../src/models/janus/processing_janus.js"],"names":[],"mappings":"AAQA;IACI,wDAAiD;IACjD,6CAAsC;IAGtC,0CAOC;IAJG,eAAsC;IACtC,qBAAkD;IAClD,mBAA8C;IAC9C,sBAAoD;IAGxD;;;;OAIG;IAEH;;;;;;OAMG;IAEH;;;;;;OAMG;IACH;;;;iBAnBc,CAAC,QAAQ,GAAG,MAAM,GAAG,GAAG,CAAC,EAAE
|
|
1
|
+
{"version":3,"file":"processing_janus.d.ts","sourceRoot":"","sources":["../../../src/models/janus/processing_janus.js"],"names":[],"mappings":"AAQA;IACI,wDAAiD;IACjD,6CAAsC;IAGtC,0CAOC;IAJG,eAAsC;IACtC,qBAAkD;IAClD,mBAA8C;IAC9C,sBAAoD;IAGxD;;;;OAIG;IAEH;;;;;;OAMG;IAEH;;;;;;OAMG;IACH;;;;iBAnBc,CAAC,QAAQ,GAAG,MAAM,GAAG,GAAG,CAAC,EAAE;uCAetC;QAAsC,MAAM,GAApC,QAAQ,GAAC,QAAQ,EAAE;QACF,aAAa,GAA9B,MAAM;KACd,GAAU,OAAO,CAAC;;;;mBAXP,MAAM;;;;wBACN,MAAM;;;;yBACN,MAAM;;;;yBACN,MAAM;SAQ0B;;;;mBAXhC,MAAM;;;;wBACN,MAAM;;;;yBACN,MAAM;;;;yBACN,MAAM;QAQmD,OAAO,sCAAsC,EAAE,oBAAoB,CAAA,CAAC,CAgF1I;CACJ;0BAzHyB,gCAAgC;yBAKjC,sBAAsB;uBADxB,uBAAuB;mCAHX,kCAAkC;8BACvC,qBAAqB"}
|
|
@@ -1,20 +1,20 @@
|
|
|
1
1
|
export class MaskFormerImageProcessor extends ImageProcessor {
|
|
2
|
-
post_process_panoptic_segmentation(outputs: any, threshold?: number, mask_threshold?: number, overlap_mask_area_threshold?: number, label_ids_to_fuse?: Set<number>, target_sizes?: [number, number][]): {
|
|
2
|
+
post_process_panoptic_segmentation(outputs: any, threshold?: number, mask_threshold?: number, overlap_mask_area_threshold?: number, label_ids_to_fuse?: Set<number>, target_sizes?: [number, number][]): Array<{
|
|
3
3
|
segmentation: import("../../transformers.js").Tensor;
|
|
4
|
-
segments_info: {
|
|
4
|
+
segments_info: Array<{
|
|
5
5
|
id: number;
|
|
6
6
|
label_id: number;
|
|
7
7
|
score: number;
|
|
8
|
-
}
|
|
9
|
-
}
|
|
10
|
-
post_process_instance_segmentation(outputs: any, threshold?: number, target_sizes?: [number, number][]): {
|
|
8
|
+
}>;
|
|
9
|
+
}>;
|
|
10
|
+
post_process_instance_segmentation(outputs: any, threshold?: number, target_sizes?: [number, number][]): Array<{
|
|
11
11
|
segmentation: import("../../transformers.js").Tensor;
|
|
12
|
-
segments_info: {
|
|
12
|
+
segments_info: Array<{
|
|
13
13
|
id: number;
|
|
14
14
|
label_id: number;
|
|
15
15
|
score: number;
|
|
16
|
-
}
|
|
17
|
-
}
|
|
16
|
+
}>;
|
|
17
|
+
}>;
|
|
18
18
|
}
|
|
19
19
|
export class MaskFormerFeatureExtractor extends MaskFormerImageProcessor {
|
|
20
20
|
}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"image_processing_maskformer.d.ts","sourceRoot":"","sources":["../../../src/models/maskformer/image_processing_maskformer.js"],"names":[],"mappings":"AAMA;
|
|
1
|
+
{"version":3,"file":"image_processing_maskformer.d.ts","sourceRoot":"","sources":["../../../src/models/maskformer/image_processing_maskformer.js"],"names":[],"mappings":"AAMA;;;;;;;;;;;;;;;;;CAUC;AACD;CAA4E;+BAbrE,sCAAsC"}
|
|
@@ -19,7 +19,7 @@ export class MgpstrProcessor extends Processor {
|
|
|
19
19
|
* @param {string} format Type of model prediction. Must be one of ['char', 'bpe', 'wp'].
|
|
20
20
|
* @returns {[string[], number[]]} The decoded sentences and their confidence scores.
|
|
21
21
|
*/
|
|
22
|
-
_decode_helper(pred_logits: import(
|
|
22
|
+
_decode_helper(pred_logits: import("../../utils/tensor.js").Tensor, format: string): [string[], number[]];
|
|
23
23
|
/**
|
|
24
24
|
* Convert a list of lists of char token ids into a list of strings by calling char tokenizer.
|
|
25
25
|
* @param {number[][]} sequences List of tokenized input ids.
|
|
@@ -49,7 +49,7 @@ export class MgpstrProcessor extends Processor {
|
|
|
49
49
|
* - bpe_preds: The list of BPE decoded sentences.
|
|
50
50
|
* - wp_preds: The list of wp decoded sentences.
|
|
51
51
|
*/
|
|
52
|
-
batch_decode([char_logits, bpe_logits, wp_logits]: import(
|
|
52
|
+
batch_decode([char_logits, bpe_logits, wp_logits]: import("../../utils/tensor.js").Tensor[]): {
|
|
53
53
|
generated_text: string[];
|
|
54
54
|
scores: number[];
|
|
55
55
|
char_preds: string[];
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"processing_mgp_str.d.ts","sourceRoot":"","sources":["../../../src/models/mgp_str/processing_mgp_str.js"],"names":[],"mappings":"AAUA;IACI,6CAAsC;IACtC,wDAAiD;IAEjD;;OAEG;IACH,
|
|
1
|
+
{"version":3,"file":"processing_mgp_str.d.ts","sourceRoot":"","sources":["../../../src/models/mgp_str/processing_mgp_str.js"],"names":[],"mappings":"AAUA;IACI,6CAAsC;IACtC,wDAAiD;IAEjD;;OAEG;IACH,sBAFa,OAAO,qBAAqB,EAAE,eAAe,CAIzD;IAED;;OAEG;IACH,qBAFa,OAAO,qBAAqB,EAAE,aAAa,CAIvD;IAED;;OAEG;IACH,oBAFa,OAAO,qBAAqB,EAAE,aAAa,CAIvD;IAED;;;;;OAKG;IACH,4BAJW,OAAO,uBAAuB,EAAE,MAAM,UACtC,MAAM,GACJ,CAAC,MAAM,EAAE,EAAE,MAAM,EAAE,CAAC,CA0ChC;IAED;;;;OAIG;IACH,uBAHW,MAAM,EAAE,EAAE,GACR,MAAM,EAAE,CAIpB;IAED;;;;OAIG;IACH,sBAHW,MAAM,EAAE,EAAE,GACR,MAAM,EAAE,CAIpB;IAED;;;;OAIG;IACH,qBAHW,MAAM,EAAE,EAAE,GACR,MAAM,EAAE,CAIpB;IAED;;;;;;;;;;OAUG;IACH,mDATW,OAAO,uBAAuB,EAAE,MAAM,EAAE,GACtC;QAAC,cAAc,EAAE,MAAM,EAAE,CAAC;QAAC,MAAM,EAAE,MAAM,EAAE,CAAC;QAAC,UAAU,EAAE,MAAM,EAAE,CAAC;QAAC,SAAS,EAAE,MAAM,EAAE,CAAC;QAAC,QAAQ,EAAE,MAAM,EAAE,CAAA;KAAC,CA4BvH;IAmBD,6CAQC;CACJ;0BAzKyB,gCAAgC;8BAE5B,qBAAqB;mCADhB,kCAAkC"}
|
|
@@ -0,0 +1,13 @@
|
|
|
1
|
+
export class MoonshineFeatureExtractor extends FeatureExtractor {
|
|
2
|
+
/**
|
|
3
|
+
* Asynchronously extracts input values from a given audio using the provided configuration.
|
|
4
|
+
* @param {Float32Array|Float64Array} audio The audio data as a Float32Array/Float64Array.
|
|
5
|
+
* @returns {Promise<{ input_values: Tensor; }>} The extracted input values.
|
|
6
|
+
*/
|
|
7
|
+
_call(audio: Float32Array | Float64Array): Promise<{
|
|
8
|
+
input_values: Tensor;
|
|
9
|
+
}>;
|
|
10
|
+
}
|
|
11
|
+
import { FeatureExtractor } from '../../base/feature_extraction_utils.js';
|
|
12
|
+
import { Tensor } from '../../utils/tensor.js';
|
|
13
|
+
//# sourceMappingURL=feature_extraction_moonshine.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"feature_extraction_moonshine.d.ts","sourceRoot":"","sources":["../../../src/models/moonshine/feature_extraction_moonshine.js"],"names":[],"mappings":"AAIA;IACI;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY,GACvB,OAAO,CAAC;QAAE,YAAY,EAAE,MAAM,CAAC;KAAE,CAAC,CAgB9C;CACJ;iCAzBuD,wCAAwC;uBACzE,uBAAuB"}
|
|
@@ -0,0 +1,17 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Represents a MoonshineProcessor that extracts features from an audio input.
|
|
3
|
+
*/
|
|
4
|
+
export class MoonshineProcessor extends Processor {
|
|
5
|
+
static tokenizer_class: typeof AutoTokenizer;
|
|
6
|
+
static feature_extractor_class: typeof AutoFeatureExtractor;
|
|
7
|
+
/**
|
|
8
|
+
* Calls the feature_extractor function with the given audio input.
|
|
9
|
+
* @param {any} audio The audio input to extract features from.
|
|
10
|
+
* @returns {Promise<any>} A Promise that resolves with the extracted features.
|
|
11
|
+
*/
|
|
12
|
+
_call(audio: any): Promise<any>;
|
|
13
|
+
}
|
|
14
|
+
import { Processor } from "../../base/processing_utils.js";
|
|
15
|
+
import { AutoTokenizer } from "../../tokenizers.js";
|
|
16
|
+
import { AutoFeatureExtractor } from "../auto/feature_extraction_auto.js";
|
|
17
|
+
//# sourceMappingURL=processing_moonshine.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"processing_moonshine.d.ts","sourceRoot":"","sources":["../../../src/models/moonshine/processing_moonshine.js"],"names":[],"mappings":"AAIA;;GAEG;AACH;IACI,6CAAsC;IACtC,4DAAqD;IAErD;;;;OAIG;IACH,aAHW,GAAG,GACD,OAAO,CAAC,GAAG,CAAC,CAIxB;CACJ;0BAjByB,gCAAgC;8BAD5B,qBAAqB;qCADd,oCAAoC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"image_processing_owlvit.d.ts","sourceRoot":"","sources":["../../../src/models/owlvit/image_processing_owlvit.js"],"names":[],"mappings":"AAKA;
|
|
1
|
+
{"version":3,"file":"image_processing_owlvit.d.ts","sourceRoot":"","sources":["../../../src/models/owlvit/image_processing_owlvit.js"],"names":[],"mappings":"AAKA;;;;;CAKC;AACD;CAAoE;+BAR7D,sCAAsC"}
|
|
@@ -0,0 +1,12 @@
|
|
|
1
|
+
export class PaliGemmaProcessor extends Processor {
|
|
2
|
+
static tokenizer_class: typeof AutoTokenizer;
|
|
3
|
+
static image_processor_class: typeof AutoImageProcessor;
|
|
4
|
+
/**
|
|
5
|
+
* @typedef {import('../../utils/image.js').RawImage} RawImage
|
|
6
|
+
*/
|
|
7
|
+
_call(images: import("../../utils/image.js").RawImage | import("../../utils/image.js").RawImage[], text?: any, kwargs?: {}): Promise<any>;
|
|
8
|
+
}
|
|
9
|
+
import { Processor } from "../../base/processing_utils.js";
|
|
10
|
+
import { AutoTokenizer } from "../../tokenizers.js";
|
|
11
|
+
import { AutoImageProcessor } from "../auto/image_processing_auto.js";
|
|
12
|
+
//# sourceMappingURL=processing_paligemma.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"processing_paligemma.d.ts","sourceRoot":"","sources":["../../../src/models/paligemma/processing_paligemma.js"],"names":[],"mappings":"AAgBA;IACI,6CAAsC;IACtC,wDAAiD;IAGjD;;OAEG;IAGH,cAAuB,0CAAS,yCAAU,yCAsDzC;CACJ;0BAjFyB,gCAAgC;8BAE5B,qBAAqB;mCADhB,kCAAkC"}
|
|
@@ -0,0 +1,17 @@
|
|
|
1
|
+
export class Phi3VImageProcessor extends ImageProcessor {
|
|
2
|
+
constructor(config: any);
|
|
3
|
+
_num_crops: any;
|
|
4
|
+
calc_num_image_tokens_from_image_size(width: any, height: any): number;
|
|
5
|
+
_call(images: any, { num_crops, }?: {
|
|
6
|
+
num_crops?: any;
|
|
7
|
+
}): Promise<{
|
|
8
|
+
pixel_values: Tensor;
|
|
9
|
+
original_sizes: any[];
|
|
10
|
+
reshaped_input_sizes: any[];
|
|
11
|
+
image_sizes: Tensor;
|
|
12
|
+
num_img_tokens: number[];
|
|
13
|
+
}>;
|
|
14
|
+
}
|
|
15
|
+
import { ImageProcessor } from "../../base/image_processors_utils.js";
|
|
16
|
+
import { Tensor } from "../../utils/tensor.js";
|
|
17
|
+
//# sourceMappingURL=image_processing_phi3_v.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"image_processing_phi3_v.d.ts","sourceRoot":"","sources":["../../../src/models/phi3_v/image_processing_phi3_v.js"],"names":[],"mappings":"AASA;IACI,yBAWC;IADG,gBAAkC;IAEtC,uEAIC;IA0CD;;;;;;;;OA6FC;CACJ;+BAhKM,sCAAsC;uBACa,uBAAuB"}
|
|
@@ -0,0 +1,17 @@
|
|
|
1
|
+
export class Phi3VProcessor extends Processor {
|
|
2
|
+
static image_processor_class: typeof AutoImageProcessor;
|
|
3
|
+
static tokenizer_class: typeof AutoTokenizer;
|
|
4
|
+
/**
|
|
5
|
+
*
|
|
6
|
+
* @param {string|string[]} text
|
|
7
|
+
* @param {RawImage|RawImage[]} images
|
|
8
|
+
* @param {...any} args
|
|
9
|
+
* @returns {Promise<any>}
|
|
10
|
+
*/
|
|
11
|
+
_call(text: string | string[], images?: RawImage | RawImage[], { padding, truncation, num_crops, }?: any[]): Promise<any>;
|
|
12
|
+
}
|
|
13
|
+
import { Processor } from "../../base/processing_utils.js";
|
|
14
|
+
import { RawImage } from "../../utils/image.js";
|
|
15
|
+
import { AutoImageProcessor } from "../auto/image_processing_auto.js";
|
|
16
|
+
import { AutoTokenizer } from "../../tokenizers.js";
|
|
17
|
+
//# sourceMappingURL=processing_phi3_v.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"processing_phi3_v.d.ts","sourceRoot":"","sources":["../../../src/models/phi3_v/processing_phi3_v.js"],"names":[],"mappings":"AAQA;IACI,wDAAiD;IACjD,6CAAsC;IAEtC;;;;;;OAMG;IACH,YALW,MAAM,GAAC,MAAM,EAAE,WACf,QAAQ,GAAC,QAAQ,EAAE,wCACf,GAAG,EAAA,GACL,OAAO,CAAC,GAAG,CAAC,CAkCxB;CACJ;0BApDyB,gCAAgC;yBAGjC,sBAAsB;mCAFZ,kCAAkC;8BACvC,qBAAqB"}
|
|
@@ -1,9 +1,12 @@
|
|
|
1
1
|
export * from "./florence2/processing_florence2.js";
|
|
2
2
|
export * from "./mgp_str/processing_mgp_str.js";
|
|
3
|
+
export * from "./moonshine/processing_moonshine.js";
|
|
3
4
|
export * from "./idefics3/processing_idefics3.js";
|
|
4
5
|
export * from "./janus/processing_janus.js";
|
|
5
6
|
export * from "./jina_clip/processing_jina_clip.js";
|
|
6
7
|
export * from "./owlvit/processing_owlvit.js";
|
|
8
|
+
export * from "./phi3_v/processing_phi3_v.js";
|
|
9
|
+
export * from "./paligemma/processing_paligemma.js";
|
|
7
10
|
export * from "./pyannote/processing_pyannote.js";
|
|
8
11
|
export * from "./qwen2_vl/processing_qwen2_vl.js";
|
|
9
12
|
export * from "./sam/processing_sam.js";
|
|
@@ -7,6 +7,24 @@ export class PyAnnoteFeatureExtractor extends FeatureExtractor {
|
|
|
7
7
|
_call(audio: Float32Array | Float64Array): Promise<{
|
|
8
8
|
input_values: Tensor;
|
|
9
9
|
}>;
|
|
10
|
+
/**
|
|
11
|
+
* NOTE: Can return fractional values. `Math.ceil` will ensure correct value.
|
|
12
|
+
* @param {number} samples The number of frames in the audio.
|
|
13
|
+
* @returns {number} The number of frames in the audio.
|
|
14
|
+
*/
|
|
15
|
+
samples_to_frames(samples: number): number;
|
|
16
|
+
/**
|
|
17
|
+
* Post-processes the speaker diarization logits output by the model.
|
|
18
|
+
* @param {import('../../utils/tensor.js').Tensor} logits The speaker diarization logits output by the model.
|
|
19
|
+
* @param {number} num_samples Number of samples in the input audio.
|
|
20
|
+
* @returns {Array<Array<{ id: number, start: number, end: number, confidence: number }>>} The post-processed speaker diarization results.
|
|
21
|
+
*/
|
|
22
|
+
post_process_speaker_diarization(logits: import("../../utils/tensor.js").Tensor, num_samples: number): Array<Array<{
|
|
23
|
+
id: number;
|
|
24
|
+
start: number;
|
|
25
|
+
end: number;
|
|
26
|
+
confidence: number;
|
|
27
|
+
}>>;
|
|
10
28
|
}
|
|
11
29
|
import { FeatureExtractor } from '../../base/feature_extraction_utils.js';
|
|
12
30
|
import { Tensor } from '../../utils/tensor.js';
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"feature_extraction_pyannote.d.ts","sourceRoot":"","sources":["../../../src/models/pyannote/feature_extraction_pyannote.js"],"names":[],"mappings":"
|
|
1
|
+
{"version":3,"file":"feature_extraction_pyannote.d.ts","sourceRoot":"","sources":["../../../src/models/pyannote/feature_extraction_pyannote.js"],"names":[],"mappings":"AAKA;IACI;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY,GACvB,OAAO,CAAC;QAAE,YAAY,EAAE,MAAM,CAAC;KAAE,CAAC,CAiB9C;IAED;;;;OAIG;IACH,2BAHW,MAAM,GACJ,MAAM,CAIlB;IAED;;;;;OAKG;IACH,yCAJW,OAAO,uBAAuB,EAAE,MAAM,eACtC,MAAM,GACJ,KAAK,CAAC,KAAK,CAAC;QAAE,EAAE,EAAE,MAAM,CAAC;QAAC,KAAK,EAAE,MAAM,CAAC;QAAC,GAAG,EAAE,MAAM,CAAC;QAAC,UAAU,EAAE,MAAM,CAAA;KAAE,CAAC,CAAC,CAwCxF;CAEJ;iCAnFuD,wCAAwC;uBACzE,uBAAuB"}
|
|
@@ -1,30 +1,19 @@
|
|
|
1
1
|
export class PyAnnoteProcessor extends Processor {
|
|
2
|
-
static feature_extractor_class: typeof
|
|
2
|
+
static feature_extractor_class: typeof PyAnnoteFeatureExtractor;
|
|
3
3
|
/**
|
|
4
4
|
* Calls the feature_extractor function with the given audio input.
|
|
5
5
|
* @param {any} audio The audio input to extract features from.
|
|
6
6
|
* @returns {Promise<any>} A Promise that resolves with the extracted features.
|
|
7
7
|
*/
|
|
8
8
|
_call(audio: any): Promise<any>;
|
|
9
|
-
|
|
10
|
-
* NOTE: Can return fractional values. `Math.ceil` will ensure correct value.
|
|
11
|
-
* @param {number} samples The number of frames in the audio.
|
|
12
|
-
* @returns {number} The number of frames in the audio.
|
|
13
|
-
*/
|
|
14
|
-
samples_to_frames(samples: number): number;
|
|
15
|
-
/**
|
|
16
|
-
* Post-processes the speaker diarization logits output by the model.
|
|
17
|
-
* @param {import('../../utils/tensor.js').Tensor} logits The speaker diarization logits output by the model.
|
|
18
|
-
* @param {number} num_samples Number of samples in the input audio.
|
|
19
|
-
* @returns {Array<Array<{ id: number, start: number, end: number, confidence: number }>>} The post-processed speaker diarization results.
|
|
20
|
-
*/
|
|
21
|
-
post_process_speaker_diarization(logits: import('../../utils/tensor.js').Tensor, num_samples: number): Array<Array<{
|
|
9
|
+
post_process_speaker_diarization(logits: import("../../transformers.js").Tensor, num_samples: number): Array<Array<{
|
|
22
10
|
id: number;
|
|
23
11
|
start: number;
|
|
24
12
|
end: number;
|
|
25
13
|
confidence: number;
|
|
26
14
|
}>>;
|
|
15
|
+
get sampling_rate(): any;
|
|
27
16
|
}
|
|
28
17
|
import { Processor } from '../../base/processing_utils.js';
|
|
29
|
-
import {
|
|
18
|
+
import { PyAnnoteFeatureExtractor } from './feature_extraction_pyannote.js';
|
|
30
19
|
//# sourceMappingURL=processing_pyannote.d.ts.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"processing_pyannote.d.ts","sourceRoot":"","sources":["../../../src/models/pyannote/processing_pyannote.js"],"names":[],"mappings":"
|
|
1
|
+
{"version":3,"file":"processing_pyannote.d.ts","sourceRoot":"","sources":["../../../src/models/pyannote/processing_pyannote.js"],"names":[],"mappings":"AAGA;IACI,gEAAyD;IAEzD;;;;OAIG;IACH,aAHW,GAAG,GACD,OAAO,CAAC,GAAG,CAAC,CAIxB;oDAWogB,uBAAuB;;;;;;IAJ5hB,yBAEC;CACJ;0BAvByB,gCAAgC;yCACjB,kCAAkC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"processing_qwen2_vl.d.ts","sourceRoot":"","sources":["../../../src/models/qwen2_vl/processing_qwen2_vl.js"],"names":[],"mappings":"AAKA;IACI,wDAAiD;IACjD,6CAAsC;IAEtC;;;;;;OAMG;IACH,YALW,MAAM,GAAC,MAAM,EAAE,WACf,QAAQ,GAAC,QAAQ,EAAE,WACf,GAAG,
|
|
1
|
+
{"version":3,"file":"processing_qwen2_vl.d.ts","sourceRoot":"","sources":["../../../src/models/qwen2_vl/processing_qwen2_vl.js"],"names":[],"mappings":"AAKA;IACI,wDAAiD;IACjD,6CAAsC;IAEtC;;;;;;OAMG;IACH,YALW,MAAM,GAAC,MAAM,EAAE,WACf,QAAQ,GAAC,QAAQ,EAAE,WACf,GAAG,EAAA,GACL,OAAO,CAAC,GAAG,CAAC,CAoCxB;CACJ;0BAnDyB,gCAAgC;yBAGjC,sBAAsB;mCAFZ,kCAAkC;8BACvC,qBAAqB"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"image_processing_rt_detr.d.ts","sourceRoot":"","sources":["../../../src/models/rt_detr/image_processing_rt_detr.js"],"names":[],"mappings":"AAMA
|
|
1
|
+
{"version":3,"file":"image_processing_rt_detr.d.ts","sourceRoot":"","sources":["../../../src/models/rt_detr/image_processing_rt_detr.js"],"names":[],"mappings":"AAMA;;;;;CAKC;+BARM,sCAAsC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"image_processing_sam.d.ts","sourceRoot":"","sources":["../../../src/models/sam/image_processing_sam.js"],"names":[],"mappings":"AAWA;;;;;;;;GAQG;AAEH;IAEI;;;;;;OAMG;IACH,mCALW,GAAG,kBACH,OAAO,sCAAsC,EAAE,WAAW,EAAE,wBAC5D,OAAO,sCAAsC,EAAE,WAAW,EAAE,8BAC1D,MAAM,CA4ClB;IAED;;;;;OAKG;IACH,+BAJW,GAAG,gBACH,MAAM,GACJ,MAAM,CAoBlB;IACD;;;;;;;;;;;;;;;;;;;OAmBG;IACH,cAnBW,GAAG,EAAE;
|
|
1
|
+
{"version":3,"file":"image_processing_sam.d.ts","sourceRoot":"","sources":["../../../src/models/sam/image_processing_sam.js"],"names":[],"mappings":"AAWA;;;;;;;;GAQG;AAEH;IAEI;;;;;;OAMG;IACH,mCALW,GAAG,kBACH,OAAO,sCAAsC,EAAE,WAAW,EAAE,wBAC5D,OAAO,sCAAsC,EAAE,WAAW,EAAE,8BAC1D,MAAM,CA4ClB;IAED;;;;;OAKG;IACH,+BAJW,GAAG,gBACH,MAAM,GACJ,MAAM,CAoBlB;IACD;;;;;;;;;;;;;;;;;;;OAmBG;IACH,cAnBW,GAAG,EAAE,gDAEb;QAAsB,YAAY,GAA1B,GAAG;QAGW,YAAY,GAA1B,GAAG;QAGoB,WAAW,GAAlC,MAAM,EAAE,EAAE,EAAE;KASpB,GAAU,OAAO,CAAC,uBAAuB,CAAC,CA+B5C;IAED;;;;;;;;;;;;OAYG;IACH,0BAXW,MAAM,kBACN,CAAC,MAAM,EAAE,MAAM,CAAC,EAAE,wBAClB,CAAC,MAAM,EAAE,MAAM,CAAC,EAAE,4CAE1B;QAAyB,cAAc,GAA/B,MAAM;QACY,QAAQ,GAA1B,OAAO;QACU,QAAQ,GACjC;YAAkC,MAAM,GAAhC,MAAM;YACoB,KAAK,GAA/B,MAAM;SACd;KAAA,GAAU,OAAO,CAAC,MAAM,EAAE,CAAC,CAsD7B;IAED;;;;;;;;;;;;;OAaG;IACH,2BAZW,OAAO,sBAAsB,EAAE,QAAQ,eACvC,MAAM,uFAEd;QAAyB,aAAa,GAA9B,MAAM;QAEW,aAAa,GAA9B,MAAM;QAEW,eAAe,GAAhC,MAAM;QACW,8BAA8B,GAA/C,MAAM;KAEd,OAUF;CACJ;;kBAnOa,MAAM;oBACN,OAAO,sCAAsC,EAAE,WAAW,EAAE;0BAC5D,OAAO,sCAAsC,EAAE,WAAW,EAAE;mBAC5D,MAAM;mBACN,MAAM;kBACN,MAAM;;+BAhBb,sCAAsC;uBAMtC,uBAAuB"}
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
export class SeamlessM4TFeatureExtractor extends FeatureExtractor {
|
|
2
2
|
constructor(config: any);
|
|
3
3
|
mel_filters: number[][];
|
|
4
|
-
window: Float64Array
|
|
4
|
+
window: Float64Array<ArrayBufferLike>;
|
|
5
5
|
/**
|
|
6
6
|
* Computes the log-Mel spectrogram of the provided audio waveform.
|
|
7
7
|
* @param {Float32Array|Float64Array} waveform The audio waveform to process.
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"feature_extraction_seamless_m4t.d.ts","sourceRoot":"","sources":["../../../src/models/seamless_m4t/feature_extraction_seamless_m4t.js"],"names":[],"mappings":"AAIA;IAEI,yBAwBC;IALG,wBAA8B;IAE9B,
|
|
1
|
+
{"version":3,"file":"feature_extraction_seamless_m4t.d.ts","sourceRoot":"","sources":["../../../src/models/seamless_m4t/feature_extraction_seamless_m4t.js"],"names":[],"mappings":"AAIA;IAEI,yBAwBC;IALG,wBAA8B;IAE9B,sCAEE;IAGN;;;;;OAKG;IACH,kCAJW,YAAY,GAAC,YAAY,cACzB,MAAM,GACJ,OAAO,CAAC,MAAM,CAAC,CA6B3B;IAED;;;;;;;;;OASG;IACH,aARW,YAAY,GAAC,YAAY,uFAEjC;QAA0B,OAAO,GAAzB,OAAO;QACU,kBAAkB,GAAnC,MAAM;QACY,yBAAyB,GAA3C,OAAO;QACW,qBAAqB,GAAvC,OAAO;KACf,GAAU,OAAO,CAAC;QAAE,cAAc,EAAE,MAAM,CAAC;QAAC,cAAc,CAAC,EAAE,MAAM,CAAA;KAAE,CAAC,CAuGxE;CACJ;iCAnLuD,wCAAwC;uBACzE,uBAAuB"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"image_processing_segformer.d.ts","sourceRoot":"","sources":["../../../src/models/segformer/image_processing_segformer.js"],"names":[],"mappings":"AAMA;
|
|
1
|
+
{"version":3,"file":"image_processing_segformer.d.ts","sourceRoot":"","sources":["../../../src/models/segformer/image_processing_segformer.js"],"names":[],"mappings":"AAMA;;;;;CAKC;AACD;CAA0E;+BATnE,sCAAsC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"processing_speecht5.d.ts","sourceRoot":"","sources":["../../../src/models/speecht5/processing_speecht5.js"],"names":[],"mappings":"AAIA;IACI,6CAAsC;IACtC,4DAAqD;IAErD;;;;OAIG;IACH,aAHW,GAAG,GACD,
|
|
1
|
+
{"version":3,"file":"processing_speecht5.d.ts","sourceRoot":"","sources":["../../../src/models/speecht5/processing_speecht5.js"],"names":[],"mappings":"AAIA;IACI,6CAAsC;IACtC,4DAAqD;IAErD;;;;OAIG;IACH,aAHW,GAAG,GACD,OAAO,CAAC,GAAG,CAAC,CAIxB;CACJ;0BAhByB,gCAAgC;8BAC5B,qBAAqB;qCACd,oCAAoC"}
|
|
@@ -1,5 +1,5 @@
|
|
|
1
1
|
export class Swin2SRImageProcessor extends ImageProcessor {
|
|
2
|
-
pad_image(pixelData: any, imgDims: any, padSize: any, options?: {}): [Float32Array
|
|
2
|
+
pad_image(pixelData: any, imgDims: any, padSize: any, options?: {}): [Float32Array<ArrayBufferLike>, number[]];
|
|
3
3
|
}
|
|
4
4
|
import { ImageProcessor } from "../../base/image_processors_utils.js";
|
|
5
5
|
//# sourceMappingURL=image_processing_swin2sr.d.ts.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"image_processing_swin2sr.d.ts","sourceRoot":"","sources":["../../../src/models/swin2sr/image_processing_swin2sr.js"],"names":[],"mappings":"AAIA;IACI
|
|
1
|
+
{"version":3,"file":"image_processing_swin2sr.d.ts","sourceRoot":"","sources":["../../../src/models/swin2sr/image_processing_swin2sr.js"],"names":[],"mappings":"AAIA;IACI,+GAiBC;CACJ;+BArBM,sCAAsC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"image_processing_vitmatte.d.ts","sourceRoot":"","sources":["../../../src/models/vitmatte/image_processing_vitmatte.js"],"names":[],"mappings":"AASA;IACI;;;;;;OAMG;IACH,cAJW,OAAO,sBAAsB,EAAE,QAAQ,EAAE,WACzC,OAAO,sBAAsB,EAAE,QAAQ,EAAE,GACvC,
|
|
1
|
+
{"version":3,"file":"image_processing_vitmatte.d.ts","sourceRoot":"","sources":["../../../src/models/vitmatte/image_processing_vitmatte.js"],"names":[],"mappings":"AASA;IACI;;;;;;OAMG;IACH,cAJW,OAAO,sBAAsB,EAAE,QAAQ,EAAE,WACzC,OAAO,sBAAsB,EAAE,QAAQ,EAAE,GACvC,OAAO,CAAC,OAAO,sCAAsC,EAAE,oBAAoB,CAAC,CAiCxF;CACJ;+BA/CM,sCAAsC"}
|
|
@@ -13,7 +13,7 @@ export class VitPoseImageProcessor extends ImageProcessor {
|
|
|
13
13
|
* keypoints: [number, number][]
|
|
14
14
|
* }[][]} List of keypoints predictions for each image.
|
|
15
15
|
*/
|
|
16
|
-
post_process_pose_estimation(outputs: import(
|
|
16
|
+
post_process_pose_estimation(outputs: import("../../utils/tensor.js").Tensor, boxes: [number, number, number, number][][], { threshold, }?: {
|
|
17
17
|
threshold?: any;
|
|
18
18
|
}): {
|
|
19
19
|
bbox: [number, number, number, number];
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"image_processing_vitpose.d.ts","sourceRoot":"","sources":["../../../src/models/vitpose/image_processing_vitpose.js"],"names":[],"mappings":"AAIA;IAEI;;;;;;;;;;;;;OAaG;IACH,sCAVW,OAAO,uBAAuB,EAAE,MAAM,SACtC,CAAC,MAAM,EAAE,MAAM,EAAE,MAAM,EAAE,MAAM,CAAC,EAAE,EAAE;;QAElC;
|
|
1
|
+
{"version":3,"file":"image_processing_vitpose.d.ts","sourceRoot":"","sources":["../../../src/models/vitpose/image_processing_vitpose.js"],"names":[],"mappings":"AAIA;IAEI;;;;;;;;;;;;;OAaG;IACH,sCAVW,OAAO,uBAAuB,EAAE,MAAM,SACtC,CAAC,MAAM,EAAE,MAAM,EAAE,MAAM,EAAE,MAAM,CAAC,EAAE,EAAE;;QAElC;QACR,IAAI,EAAE,CAAC,MAAM,EAAE,MAAM,EAAE,MAAM,EAAE,MAAM,CAAC,CAAC;QACvC,MAAM,EAAE,MAAM,EAAE,CAAC;QACjB,MAAM,EAAE,MAAM,EAAE,CAAC;QACjB,SAAS,EAAE,CAAC,MAAM,EAAE,MAAM,CAAC,EAAE,CAAA;KAC9B,EAAE,EAAE,CAqEP;CACJ;+BAtFM,sCAAsC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"feature_extraction_wav2vec2.d.ts","sourceRoot":"","sources":["../../../src/models/wav2vec2/feature_extraction_wav2vec2.js"],"names":[],"mappings":"AAGA;IAEI;;;OAGG;IACH,uCAHW,YAAY,GACV,YAAY,CAQxB;IAED;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY;
|
|
1
|
+
{"version":3,"file":"feature_extraction_wav2vec2.d.ts","sourceRoot":"","sources":["../../../src/models/wav2vec2/feature_extraction_wav2vec2.js"],"names":[],"mappings":"AAGA;IAEI;;;OAGG;IACH,uCAHW,YAAY,GACV,YAAY,CAQxB;IAED;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY,GACvB,OAAO,CAAC;QAAE,YAAY,EAAE,MAAM,CAAC;QAAC,cAAc,EAAE,MAAM,CAAA;KAAE,CAAC,CAsBrE;CACJ;iCA3CuD,wCAAwC;uBACzE,uBAAuB"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"processing_wav2vec2.d.ts","sourceRoot":"","sources":["../../../src/models/wav2vec2/processing_wav2vec2.js"],"names":[],"mappings":"AAGA;IACI,4DAAqD;IAErD;;;;OAIG;IACH,aAHW,GAAG,GACD,
|
|
1
|
+
{"version":3,"file":"processing_wav2vec2.d.ts","sourceRoot":"","sources":["../../../src/models/wav2vec2/processing_wav2vec2.js"],"names":[],"mappings":"AAGA;IACI,4DAAqD;IAErD;;;;OAIG;IACH,aAHW,GAAG,GACD,OAAO,CAAC,GAAG,CAAC,CAIxB;CACJ;0BAdyB,gCAAgC;qCACrB,oCAAoC"}
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
export class WeSpeakerFeatureExtractor extends FeatureExtractor {
|
|
2
2
|
constructor(config: any);
|
|
3
3
|
mel_filters: number[][];
|
|
4
|
-
window: Float64Array
|
|
4
|
+
window: Float64Array<ArrayBufferLike>;
|
|
5
5
|
min_num_frames: any;
|
|
6
6
|
/**
|
|
7
7
|
* Computes the log-Mel spectrogram of the provided audio waveform.
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"feature_extraction_wespeaker.d.ts","sourceRoot":"","sources":["../../../src/models/wespeaker/feature_extraction_wespeaker.js"],"names":[],"mappings":"AAKA;IAEI,yBAyBC;IANG,wBAA8B;IAE9B,
|
|
1
|
+
{"version":3,"file":"feature_extraction_wespeaker.d.ts","sourceRoot":"","sources":["../../../src/models/wespeaker/feature_extraction_wespeaker.js"],"names":[],"mappings":"AAKA;IAEI,yBAyBC;IANG,wBAA8B;IAE9B,sCAEE;IACF,oBAAgD;IAGpD;;;;OAIG;IACH,kCAHW,YAAY,GAAC,YAAY,GACvB,OAAO,CAAC,MAAM,CAAC,CA2B3B;IAGD;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY,GACvB,OAAO,CAAC;QAAE,cAAc,EAAE,MAAM,CAAA;KAAE,CAAC,CA4B/C;CACJ;iCAnGuD,wCAAwC;uBACzE,uBAAuB"}
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
export class WhisperFeatureExtractor extends FeatureExtractor {
|
|
2
2
|
constructor(config: any);
|
|
3
|
-
window: Float64Array
|
|
3
|
+
window: Float64Array<ArrayBufferLike>;
|
|
4
4
|
/**
|
|
5
5
|
* Computes the log-Mel spectrogram of the provided audio waveform.
|
|
6
6
|
* @param {Float32Array|Float64Array} waveform The audio waveform to process.
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"feature_extraction_whisper.d.ts","sourceRoot":"","sources":["../../../src/models/whisper/feature_extraction_whisper.js"],"names":[],"mappings":"AAKA;IAEI,yBAeC;IADG,
|
|
1
|
+
{"version":3,"file":"feature_extraction_whisper.d.ts","sourceRoot":"","sources":["../../../src/models/whisper/feature_extraction_whisper.js"],"names":[],"mappings":"AAKA;IAEI,yBAeC;IADG,sCAAwD;IAG5D;;;;OAIG;IACH,kCAHW,YAAY,GAAC,YAAY,GACvB,OAAO,CAAC,MAAM,CAAC,CA0B3B;IAED;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY,GACvB,OAAO,CAAC;QAAE,cAAc,EAAE,MAAM,CAAA;KAAE,CAAC,CAwB/C;CACJ;iCAnFuD,wCAAwC;uBACzE,uBAAuB"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"generation_whisper.d.ts","sourceRoot":"","sources":["../../../src/models/whisper/generation_whisper.js"],"names":[],"mappings":"AAEA;IAEI;;;OAGG;IACH,mBAFU,OAAO,CAEQ;IAEzB;;;;;OAKG;IACH,yBAFU,OAAO,CAEc;IAE/B;;;OAGG;IACH,YAFU,MAAM,CAEE;IAElB;;;;OAIG;IACH,iBAFU,CAAC,MAAM,EAAE,MAAM,CAAC,EAAE,CAEL;IAEvB;;;OAGG;IACH,MAFU,MAAM,CAEJ;IAEZ;;;;OAIG;IACH,UAFU,MAAM,CAEA;IAEhB;;;OAGG;IACH,wBAFU,MAAM,CAEc;IAE9B;;;;;;OAMG;IACH,YAFU,MAAM,EAAE,CAEA;IAElB;;;OAGG;IACH,iBAFU,OAAO,CAEM;IAEvB;;;;OAIG;IACH,YAFU,
|
|
1
|
+
{"version":3,"file":"generation_whisper.d.ts","sourceRoot":"","sources":["../../../src/models/whisper/generation_whisper.js"],"names":[],"mappings":"AAEA;IAEI;;;OAGG;IACH,mBAFU,OAAO,CAEQ;IAEzB;;;;;OAKG;IACH,yBAFU,OAAO,CAEc;IAE/B;;;OAGG;IACH,YAFU,MAAM,CAEE;IAElB;;;;OAIG;IACH,iBAFU,CAAC,MAAM,EAAE,MAAM,CAAC,EAAE,CAEL;IAEvB;;;OAGG;IACH,MAFU,MAAM,CAEJ;IAEZ;;;;OAIG;IACH,UAFU,MAAM,CAEA;IAEhB;;;OAGG;IACH,wBAFU,MAAM,CAEc;IAE9B;;;;;;OAMG;IACH,YAFU,MAAM,EAAE,CAEA;IAElB;;;OAGG;IACH,iBAFU,OAAO,CAEM;IAEvB;;;;OAIG;IACH,YAFU,MAAM,CAAC,MAAM,EAAE,MAAM,CAAC,GAAC,IAAI,CAEnB;IAElB;;;OAGG;IACH,YAFU,MAAM,CAAC,MAAM,EAAE,MAAM,CAAC,GAAC,IAAI,CAEnB;IAElB;;;;OAIG;IACH,6BAFU,MAAM,CAEgB;CACnC;kDAGY,GAAqE,GAAG;IAAC,iBAAiB,EAAE,uBAAuB,CAAA;CAAC,GAAG,uBAAuB;iCAvF1H,yCAAyC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"processing_whisper.d.ts","sourceRoot":"","sources":["../../../src/models/whisper/processing_whisper.js"],"names":[],"mappings":"AAIA;;GAEG;AACH;IACI,6CAAsC;IACtC,4DAAqD;IAErD;;;;OAIG;IACH,aAHW,GAAG,GACD,
|
|
1
|
+
{"version":3,"file":"processing_whisper.d.ts","sourceRoot":"","sources":["../../../src/models/whisper/processing_whisper.js"],"names":[],"mappings":"AAIA;;GAEG;AACH;IACI,6CAAsC;IACtC,4DAAqD;IAErD;;;;OAIG;IACH,aAHW,GAAG,GACD,OAAO,CAAC,GAAG,CAAC,CAIxB;CACJ;0BAjByB,gCAAgC;8BAD5B,qBAAqB;qCADd,oCAAoC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"image_processing_yolos.d.ts","sourceRoot":"","sources":["../../../src/models/yolos/image_processing_yolos.js"],"names":[],"mappings":"AAKA;
|
|
1
|
+
{"version":3,"file":"image_processing_yolos.d.ts","sourceRoot":"","sources":["../../../src/models/yolos/image_processing_yolos.js"],"names":[],"mappings":"AAKA;;;;;CAKC;AACD;CAAkE;+BAR3D,sCAAsC"}
|