@huggingface/transformers 3.0.1 → 3.1.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +14 -4
- package/dist/ort-wasm-simd-threaded.jsep.wasm +0 -0
- package/dist/transformers.cjs +16607 -13472
- package/dist/transformers.cjs.map +1 -1
- package/dist/transformers.js +16601 -13451
- package/dist/transformers.js.map +1 -1
- package/dist/transformers.min.cjs +238 -52
- package/dist/transformers.min.cjs.map +1 -1
- package/dist/transformers.min.js +229 -43
- package/dist/transformers.min.js.map +1 -1
- package/dist/transformers.min.mjs +240 -54
- package/dist/transformers.min.mjs.map +1 -1
- package/dist/transformers.mjs +16017 -12878
- package/dist/transformers.mjs.map +1 -1
- package/package.json +7 -7
- package/src/base/feature_extraction_utils.js +54 -0
- package/src/base/image_processors_utils.js +1089 -0
- package/src/base/processing_utils.js +145 -0
- package/src/configs.js +15 -3
- package/src/env.js +15 -4
- package/src/models/audio_spectrogram_transformer/feature_extraction_audio_spectrogram_transformer.js +90 -0
- package/src/models/auto/feature_extraction_auto.js +41 -0
- package/src/models/auto/image_processing_auto.js +29 -0
- package/src/models/auto/processing_auto.js +100 -0
- package/src/models/beit/image_processing_beit.js +5 -0
- package/src/models/bit/image_processing_bit.js +5 -0
- package/src/models/chinese_clip/image_processing_chinese_clip.js +5 -0
- package/src/models/clap/feature_extraction_clap.js +159 -0
- package/src/models/clip/image_processing_clip.js +6 -0
- package/src/models/convnext/image_processing_convnext.js +45 -0
- package/src/models/deit/image_processing_deit.js +6 -0
- package/src/models/detr/image_processing_detr.js +52 -0
- package/src/models/donut/image_processing_donut.js +31 -0
- package/src/models/dpt/image_processing_dpt.js +6 -0
- package/src/models/efficientnet/image_processing_efficientnet.js +13 -0
- package/src/models/feature_extractors.js +12 -0
- package/src/models/florence2/processing_florence2.js +128 -0
- package/src/models/glpn/image_processing_glpn.js +5 -0
- package/src/models/image_processors.js +36 -0
- package/src/models/janus/image_processing_janus.js +26 -0
- package/src/models/janus/processing_janus.js +123 -0
- package/src/models/jina_clip/image_processing_jina_clip.js +26 -0
- package/src/models/jina_clip/processing_jina_clip.js +24 -0
- package/src/models/llava_onevision/image_processing_llava_onevision.js +5 -0
- package/src/models/mask2former/image_processing_mask2former.js +5 -0
- package/src/models/maskformer/image_processing_maskformer.js +18 -0
- package/src/models/mgp_str/processing_mgp_str.js +170 -0
- package/src/models/mobilenet_v1/image_processing_mobilenet_v1.js +7 -0
- package/src/models/mobilenet_v2/image_processing_mobilenet_v2.js +7 -0
- package/src/models/mobilenet_v3/image_processing_mobilenet_v3.js +7 -0
- package/src/models/mobilenet_v4/image_processing_mobilenet_v4.js +7 -0
- package/src/models/mobilevit/image_processing_mobilevit.js +6 -0
- package/src/models/nougat/image_processing_nougat.js +5 -0
- package/src/models/owlv2/image_processing_owlv2.js +5 -0
- package/src/models/owlvit/image_processing_owlvit.js +12 -0
- package/src/models/owlvit/processing_owlvit.js +7 -0
- package/src/models/processors.js +11 -0
- package/src/models/pvt/image_processing_pvt.js +5 -0
- package/src/models/pyannote/feature_extraction_pyannote.js +28 -0
- package/src/models/pyannote/processing_pyannote.js +71 -0
- package/src/models/qwen2_vl/image_processing_qwen2_vl.js +52 -0
- package/src/models/qwen2_vl/processing_qwen2_vl.js +52 -0
- package/src/models/rt_detr/image_processing_rt_detr.js +12 -0
- package/src/models/sam/image_processing_sam.js +242 -0
- package/src/models/sam/processing_sam.js +20 -0
- package/src/models/sapiens/image_processing_sapiens.js +13 -0
- package/src/models/seamless_m4t/feature_extraction_seamless_m4t.js +180 -0
- package/src/models/segformer/image_processing_segformer.js +13 -0
- package/src/models/siglip/image_processing_siglip.js +5 -0
- package/src/models/speecht5/feature_extraction_speecht5.js +4 -0
- package/src/models/speecht5/processing_speecht5.js +17 -0
- package/src/models/swin2sr/image_processing_swin2sr.js +24 -0
- package/src/models/vit/image_processing_vit.js +7 -0
- package/src/models/vitmatte/image_processing_vitmatte.js +50 -0
- package/src/models/vitpose/image_processing_vitpose.js +89 -0
- package/src/models/wav2vec2/feature_extraction_wav2vec2.js +44 -0
- package/src/models/wav2vec2/processing_wav2vec2.js +15 -0
- package/src/models/wespeaker/feature_extraction_wespeaker.js +100 -0
- package/src/models/whisper/feature_extraction_whisper.js +84 -0
- package/src/models/whisper/processing_whisper.js +21 -0
- package/src/models/yolos/image_processing_yolos.js +12 -0
- package/src/models.js +695 -32
- package/src/pipelines.js +8 -8
- package/src/tokenizers.js +5 -0
- package/src/transformers.js +15 -2
- package/src/utils/constants.js +8 -1
- package/src/utils/core.js +37 -9
- package/src/utils/hub.js +2 -1
- package/src/utils/image.js +68 -17
- package/src/utils/tensor.js +33 -1
- package/types/base/feature_extraction_utils.d.ts +41 -0
- package/types/base/feature_extraction_utils.d.ts.map +1 -0
- package/types/base/image_processors_utils.d.ts +323 -0
- package/types/base/image_processors_utils.d.ts.map +1 -0
- package/types/base/processing_utils.d.ts +80 -0
- package/types/base/processing_utils.d.ts.map +1 -0
- package/types/configs.d.ts +4 -1
- package/types/configs.d.ts.map +1 -1
- package/types/env.d.ts.map +1 -1
- package/types/models/audio_spectrogram_transformer/feature_extraction_audio_spectrogram_transformer.d.ts +25 -0
- package/types/models/audio_spectrogram_transformer/feature_extraction_audio_spectrogram_transformer.d.ts.map +1 -0
- package/types/models/auto/feature_extraction_auto.d.ts +5 -0
- package/types/models/auto/feature_extraction_auto.d.ts.map +1 -0
- package/types/models/auto/image_processing_auto.d.ts +5 -0
- package/types/models/auto/image_processing_auto.d.ts.map +1 -0
- package/types/models/auto/processing_auto.d.ts +35 -0
- package/types/models/auto/processing_auto.d.ts.map +1 -0
- package/types/models/beit/image_processing_beit.d.ts +4 -0
- package/types/models/beit/image_processing_beit.d.ts.map +1 -0
- package/types/models/bit/image_processing_bit.d.ts +4 -0
- package/types/models/bit/image_processing_bit.d.ts.map +1 -0
- package/types/models/chinese_clip/image_processing_chinese_clip.d.ts +4 -0
- package/types/models/chinese_clip/image_processing_chinese_clip.d.ts.map +1 -0
- package/types/models/clap/feature_extraction_clap.d.ts +57 -0
- package/types/models/clap/feature_extraction_clap.d.ts.map +1 -0
- package/types/models/clip/image_processing_clip.d.ts +6 -0
- package/types/models/clip/image_processing_clip.d.ts.map +1 -0
- package/types/models/convnext/image_processing_convnext.d.ts +12 -0
- package/types/models/convnext/image_processing_convnext.d.ts.map +1 -0
- package/types/models/deit/image_processing_deit.d.ts +6 -0
- package/types/models/deit/image_processing_deit.d.ts.map +1 -0
- package/types/models/detr/image_processing_detr.d.ts +42 -0
- package/types/models/detr/image_processing_detr.d.ts.map +1 -0
- package/types/models/donut/image_processing_donut.d.ts +7 -0
- package/types/models/donut/image_processing_donut.d.ts.map +1 -0
- package/types/models/dpt/image_processing_dpt.d.ts +6 -0
- package/types/models/dpt/image_processing_dpt.d.ts.map +1 -0
- package/types/models/efficientnet/image_processing_efficientnet.d.ts +6 -0
- package/types/models/efficientnet/image_processing_efficientnet.d.ts.map +1 -0
- package/types/models/feature_extractors.d.ts +10 -0
- package/types/models/feature_extractors.d.ts.map +1 -0
- package/types/models/florence2/processing_florence2.d.ts +39 -0
- package/types/models/florence2/processing_florence2.d.ts.map +1 -0
- package/types/models/glpn/image_processing_glpn.d.ts +4 -0
- package/types/models/glpn/image_processing_glpn.d.ts.map +1 -0
- package/types/models/image_processors.d.ts +36 -0
- package/types/models/image_processors.d.ts.map +1 -0
- package/types/models/janus/image_processing_janus.d.ts +7 -0
- package/types/models/janus/image_processing_janus.d.ts.map +1 -0
- package/types/models/janus/processing_janus.d.ts +77 -0
- package/types/models/janus/processing_janus.d.ts.map +1 -0
- package/types/models/jina_clip/image_processing_jina_clip.d.ts +5 -0
- package/types/models/jina_clip/image_processing_jina_clip.d.ts.map +1 -0
- package/types/models/jina_clip/processing_jina_clip.d.ts +9 -0
- package/types/models/jina_clip/processing_jina_clip.d.ts.map +1 -0
- package/types/models/llava_onevision/image_processing_llava_onevision.d.ts +4 -0
- package/types/models/llava_onevision/image_processing_llava_onevision.d.ts.map +1 -0
- package/types/models/mask2former/image_processing_mask2former.d.ts +4 -0
- package/types/models/mask2former/image_processing_mask2former.d.ts.map +1 -0
- package/types/models/maskformer/image_processing_maskformer.d.ts +22 -0
- package/types/models/maskformer/image_processing_maskformer.d.ts.map +1 -0
- package/types/models/mgp_str/processing_mgp_str.d.ts +64 -0
- package/types/models/mgp_str/processing_mgp_str.d.ts.map +1 -0
- package/types/models/mobilenet_v1/image_processing_mobilenet_v1.d.ts +6 -0
- package/types/models/mobilenet_v1/image_processing_mobilenet_v1.d.ts.map +1 -0
- package/types/models/mobilenet_v2/image_processing_mobilenet_v2.d.ts +6 -0
- package/types/models/mobilenet_v2/image_processing_mobilenet_v2.d.ts.map +1 -0
- package/types/models/mobilenet_v3/image_processing_mobilenet_v3.d.ts +6 -0
- package/types/models/mobilenet_v3/image_processing_mobilenet_v3.d.ts.map +1 -0
- package/types/models/mobilenet_v4/image_processing_mobilenet_v4.d.ts +6 -0
- package/types/models/mobilenet_v4/image_processing_mobilenet_v4.d.ts.map +1 -0
- package/types/models/mobilevit/image_processing_mobilevit.d.ts +6 -0
- package/types/models/mobilevit/image_processing_mobilevit.d.ts.map +1 -0
- package/types/models/nougat/image_processing_nougat.d.ts +4 -0
- package/types/models/nougat/image_processing_nougat.d.ts.map +1 -0
- package/types/models/owlv2/image_processing_owlv2.d.ts +4 -0
- package/types/models/owlv2/image_processing_owlv2.d.ts.map +1 -0
- package/types/models/owlvit/image_processing_owlvit.d.ts +10 -0
- package/types/models/owlvit/image_processing_owlvit.d.ts.map +1 -0
- package/types/models/owlvit/processing_owlvit.d.ts +8 -0
- package/types/models/owlvit/processing_owlvit.d.ts.map +1 -0
- package/types/models/processors.d.ts +12 -0
- package/types/models/processors.d.ts.map +1 -0
- package/types/models/pvt/image_processing_pvt.d.ts +4 -0
- package/types/models/pvt/image_processing_pvt.d.ts.map +1 -0
- package/types/models/pyannote/feature_extraction_pyannote.d.ts +13 -0
- package/types/models/pyannote/feature_extraction_pyannote.d.ts.map +1 -0
- package/types/models/pyannote/processing_pyannote.d.ts +30 -0
- package/types/models/pyannote/processing_pyannote.d.ts.map +1 -0
- package/types/models/qwen2_vl/image_processing_qwen2_vl.d.ts +11 -0
- package/types/models/qwen2_vl/image_processing_qwen2_vl.d.ts.map +1 -0
- package/types/models/qwen2_vl/processing_qwen2_vl.d.ts +17 -0
- package/types/models/qwen2_vl/processing_qwen2_vl.d.ts.map +1 -0
- package/types/models/rt_detr/image_processing_rt_detr.d.ts +8 -0
- package/types/models/rt_detr/image_processing_rt_detr.d.ts.map +1 -0
- package/types/models/sam/image_processing_sam.d.ts +103 -0
- package/types/models/sam/image_processing_sam.d.ts.map +1 -0
- package/types/models/sam/processing_sam.d.ts +9 -0
- package/types/models/sam/processing_sam.d.ts.map +1 -0
- package/types/models/seamless_m4t/feature_extraction_seamless_m4t.d.ts +34 -0
- package/types/models/seamless_m4t/feature_extraction_seamless_m4t.d.ts.map +1 -0
- package/types/models/segformer/image_processing_segformer.d.ts +10 -0
- package/types/models/segformer/image_processing_segformer.d.ts.map +1 -0
- package/types/models/siglip/image_processing_siglip.d.ts +4 -0
- package/types/models/siglip/image_processing_siglip.d.ts.map +1 -0
- package/types/models/speecht5/feature_extraction_speecht5.d.ts +4 -0
- package/types/models/speecht5/feature_extraction_speecht5.d.ts.map +1 -0
- package/types/models/speecht5/processing_speecht5.d.ts +14 -0
- package/types/models/speecht5/processing_speecht5.d.ts.map +1 -0
- package/types/models/swin2sr/image_processing_swin2sr.d.ts +5 -0
- package/types/models/swin2sr/image_processing_swin2sr.d.ts.map +1 -0
- package/types/models/vit/image_processing_vit.d.ts +6 -0
- package/types/models/vit/image_processing_vit.d.ts.map +1 -0
- package/types/models/vitmatte/image_processing_vitmatte.d.ts +12 -0
- package/types/models/vitmatte/image_processing_vitmatte.d.ts.map +1 -0
- package/types/models/vitpose/image_processing_vitpose.d.ts +26 -0
- package/types/models/vitpose/image_processing_vitpose.d.ts.map +1 -0
- package/types/models/wav2vec2/feature_extraction_wav2vec2.d.ts +19 -0
- package/types/models/wav2vec2/feature_extraction_wav2vec2.d.ts.map +1 -0
- package/types/models/wav2vec2/processing_wav2vec2.d.ts +12 -0
- package/types/models/wav2vec2/processing_wav2vec2.d.ts.map +1 -0
- package/types/models/wespeaker/feature_extraction_wespeaker.d.ts +23 -0
- package/types/models/wespeaker/feature_extraction_wespeaker.d.ts.map +1 -0
- package/types/models/whisper/feature_extraction_whisper.d.ts +21 -0
- package/types/models/whisper/feature_extraction_whisper.d.ts.map +1 -0
- package/types/models/whisper/processing_whisper.d.ts +17 -0
- package/types/models/whisper/processing_whisper.d.ts.map +1 -0
- package/types/models/yolos/image_processing_yolos.d.ts +10 -0
- package/types/models/yolos/image_processing_yolos.d.ts.map +1 -0
- package/types/models.d.ts +152 -0
- package/types/models.d.ts.map +1 -1
- package/types/pipelines.d.ts +2 -3
- package/types/pipelines.d.ts.map +1 -1
- package/types/tokenizers.d.ts +3 -0
- package/types/tokenizers.d.ts.map +1 -1
- package/types/transformers.d.ts +10 -1
- package/types/utils/constants.d.ts +6 -0
- package/types/utils/constants.d.ts.map +1 -1
- package/types/utils/core.d.ts +58 -3
- package/types/utils/core.d.ts.map +1 -1
- package/types/utils/hub.d.ts +1 -1
- package/types/utils/hub.d.ts.map +1 -1
- package/types/utils/image.d.ts +10 -2
- package/types/utils/image.d.ts.map +1 -1
- package/types/utils/tensor.d.ts +34 -1
- package/types/utils/tensor.d.ts.map +1 -1
- package/src/processors.js +0 -2655
- package/types/processors.d.ts +0 -924
- package/types/processors.d.ts.map +0 -1
|
@@ -0,0 +1,12 @@
|
|
|
1
|
+
export class ConvNextImageProcessor extends ImageProcessor {
|
|
2
|
+
constructor(config: any);
|
|
3
|
+
/**
|
|
4
|
+
* Percentage of the image to crop. Only has an effect if this.size < 384.
|
|
5
|
+
*/
|
|
6
|
+
crop_pct: any;
|
|
7
|
+
resize(image: any): Promise<any>;
|
|
8
|
+
}
|
|
9
|
+
export class ConvNextFeatureExtractor extends ConvNextImageProcessor {
|
|
10
|
+
}
|
|
11
|
+
import { ImageProcessor } from "../../base/image_processors_utils.js";
|
|
12
|
+
//# sourceMappingURL=image_processing_convnext.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"image_processing_convnext.d.ts","sourceRoot":"","sources":["../../../src/models/convnext/image_processing_convnext.js"],"names":[],"mappings":"AAIA;IACI,yBAOC;IAJG;;OAEG;IACH,cAAmD;IAGvD,iCA4BC;CACJ;AACD;CAAwE;+BA1CjE,sCAAsC"}
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"image_processing_deit.d.ts","sourceRoot":"","sources":["../../../src/models/deit/image_processing_deit.js"],"names":[],"mappings":"AAIA;CAA0D;AAC1D;CAAgE;+BAHzD,sCAAsC"}
|
|
@@ -0,0 +1,42 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* @typedef {object} DetrFeatureExtractorResultProps
|
|
3
|
+
* @property {import('../../utils/tensor.js').Tensor} pixel_mask
|
|
4
|
+
* @typedef {import('../../base/image_processors_utils.js').ImageProcessorResult & DetrFeatureExtractorResultProps} DetrFeatureExtractorResult
|
|
5
|
+
*/
|
|
6
|
+
export class DetrImageProcessor extends ImageProcessor {
|
|
7
|
+
/**
|
|
8
|
+
* Calls the feature extraction process on an array of images, preprocesses
|
|
9
|
+
* each image, and concatenates the resulting features into a single Tensor.
|
|
10
|
+
* @param {import('../../utils/image.js').RawImage[]} images The image(s) to extract features from.
|
|
11
|
+
* @returns {Promise<DetrFeatureExtractorResult>} An object containing the concatenated pixel values of the preprocessed images.
|
|
12
|
+
*/
|
|
13
|
+
_call(images: import('../../utils/image.js').RawImage[]): Promise<DetrFeatureExtractorResult>;
|
|
14
|
+
post_process_object_detection(outputs: {
|
|
15
|
+
logits: import("../../utils/tensor.js").Tensor;
|
|
16
|
+
pred_boxes: import("../../utils/tensor.js").Tensor;
|
|
17
|
+
}, threshold?: number, target_sizes?: [number, number][], is_zero_shot?: boolean): any[];
|
|
18
|
+
post_process_panoptic_segmentation(outputs: any, threshold?: number, mask_threshold?: number, overlap_mask_area_threshold?: number, label_ids_to_fuse?: Set<number>, target_sizes?: [number, number][]): {
|
|
19
|
+
segmentation: import("../../utils/tensor.js").Tensor;
|
|
20
|
+
segments_info: {
|
|
21
|
+
id: number;
|
|
22
|
+
label_id: number;
|
|
23
|
+
score: number;
|
|
24
|
+
}[];
|
|
25
|
+
}[];
|
|
26
|
+
post_process_instance_segmentation(outputs: any, threshold?: number, target_sizes?: [number, number][]): {
|
|
27
|
+
segmentation: import("../../utils/tensor.js").Tensor;
|
|
28
|
+
segments_info: {
|
|
29
|
+
id: number;
|
|
30
|
+
label_id: number;
|
|
31
|
+
score: number;
|
|
32
|
+
}[];
|
|
33
|
+
}[];
|
|
34
|
+
}
|
|
35
|
+
export class DetrFeatureExtractor extends DetrImageProcessor {
|
|
36
|
+
}
|
|
37
|
+
export type DetrFeatureExtractorResultProps = {
|
|
38
|
+
pixel_mask: import('../../utils/tensor.js').Tensor;
|
|
39
|
+
};
|
|
40
|
+
export type DetrFeatureExtractorResult = import('../../base/image_processors_utils.js').ImageProcessorResult & DetrFeatureExtractorResultProps;
|
|
41
|
+
import { ImageProcessor } from "../../base/image_processors_utils.js";
|
|
42
|
+
//# sourceMappingURL=image_processing_detr.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"image_processing_detr.d.ts","sourceRoot":"","sources":["../../../src/models/detr/image_processing_detr.js"],"names":[],"mappings":"AAUA;;;;GAIG;AAEH;IACI;;;;;OAKG;IACH,cAHW,OAAO,sBAAsB,EAAE,QAAQ,EAAE,GACvC,QAAQ,0BAA0B,CAAC,CAY/C;IAmBwmB;;;6FAAu6F;IAA8yU;;;;;;;QAA4qG;IAAA;;;;;;;QAAktB;CAH/rjB;AAED;CAAgE;;gBAvClD,OAAO,uBAAuB,EAAE,MAAM;;yCACvC,OAAO,sCAAsC,EAAE,oBAAoB,GAAG,+BAA+B;+BAR3G,sCAAsC"}
|
|
@@ -0,0 +1,7 @@
|
|
|
1
|
+
export class DonutImageProcessor extends ImageProcessor {
|
|
2
|
+
pad_image(pixelData: any, imgDims: any, padSize: any, options?: {}): [Float32Array, number[]];
|
|
3
|
+
}
|
|
4
|
+
export class DonutFeatureExtractor extends DonutImageProcessor {
|
|
5
|
+
}
|
|
6
|
+
import { ImageProcessor } from "../../base/image_processors_utils.js";
|
|
7
|
+
//# sourceMappingURL=image_processing_donut.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"image_processing_donut.d.ts","sourceRoot":"","sources":["../../../src/models/donut/image_processing_donut.js"],"names":[],"mappings":"AAIA;IACI,8FAuBC;CACJ;AACD;CAAkE;+BA5B3D,sCAAsC"}
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"image_processing_dpt.d.ts","sourceRoot":"","sources":["../../../src/models/dpt/image_processing_dpt.js"],"names":[],"mappings":"AAIA;CAAyD;AACzD;CAA8D;+BAHvD,sCAAsC"}
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"image_processing_efficientnet.d.ts","sourceRoot":"","sources":["../../../src/models/efficientnet/image_processing_efficientnet.js"],"names":[],"mappings":"AAIA;IACI,yBAMC;IAJG,iBAAkD;CAKzD;+BAVM,sCAAsC"}
|
|
@@ -0,0 +1,10 @@
|
|
|
1
|
+
export * from "./audio_spectrogram_transformer/feature_extraction_audio_spectrogram_transformer.js";
|
|
2
|
+
export * from "./clap/feature_extraction_clap.js";
|
|
3
|
+
export * from "./pyannote/feature_extraction_pyannote.js";
|
|
4
|
+
export * from "./seamless_m4t/feature_extraction_seamless_m4t.js";
|
|
5
|
+
export * from "./speecht5/feature_extraction_speecht5.js";
|
|
6
|
+
export * from "./wav2vec2/feature_extraction_wav2vec2.js";
|
|
7
|
+
export * from "./wespeaker/feature_extraction_wespeaker.js";
|
|
8
|
+
export * from "./whisper/feature_extraction_whisper.js";
|
|
9
|
+
export { ImageProcessor as ImageFeatureExtractor } from "../base/image_processors_utils.js";
|
|
10
|
+
//# sourceMappingURL=feature_extractors.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"feature_extractors.d.ts","sourceRoot":"","sources":["../../src/models/feature_extractors.js"],"names":[],"mappings":""}
|
|
@@ -0,0 +1,39 @@
|
|
|
1
|
+
export class Florence2Processor extends Processor {
|
|
2
|
+
static tokenizer_class: typeof AutoTokenizer;
|
|
3
|
+
static image_processor_class: typeof AutoImageProcessor;
|
|
4
|
+
constructor(config: any, components: any);
|
|
5
|
+
/** @type {Map<string, string>} */
|
|
6
|
+
tasks_answer_post_processing_type: Map<string, string>;
|
|
7
|
+
/** @type {Map<string, string>} */
|
|
8
|
+
task_prompts_without_inputs: Map<string, string>;
|
|
9
|
+
/** @type {Map<string, string>} */
|
|
10
|
+
task_prompts_with_input: Map<string, string>;
|
|
11
|
+
regexes: {
|
|
12
|
+
quad_boxes: RegExp;
|
|
13
|
+
bboxes: RegExp;
|
|
14
|
+
};
|
|
15
|
+
size_per_bin: number;
|
|
16
|
+
/**
|
|
17
|
+
* Helper function to construct prompts from input texts
|
|
18
|
+
* @param {string|string[]} text
|
|
19
|
+
* @returns {string[]}
|
|
20
|
+
*/
|
|
21
|
+
construct_prompts(text: string | string[]): string[];
|
|
22
|
+
/**
|
|
23
|
+
* Post-process the output of the model to each of the task outputs.
|
|
24
|
+
* @param {string} text The text to post-process.
|
|
25
|
+
* @param {string} task The task to post-process the text for.
|
|
26
|
+
* @param {[number, number]} image_size The size of the image. height x width.
|
|
27
|
+
*/
|
|
28
|
+
post_process_generation(text: string, task: string, image_size: [number, number]): {
|
|
29
|
+
[x: string]: string | {
|
|
30
|
+
[x: string]: any[];
|
|
31
|
+
labels: any[];
|
|
32
|
+
};
|
|
33
|
+
};
|
|
34
|
+
_call(images: any, text?: any, kwargs?: {}): Promise<any>;
|
|
35
|
+
}
|
|
36
|
+
import { Processor } from "../../base/processing_utils.js";
|
|
37
|
+
import { AutoTokenizer } from "../../tokenizers.js";
|
|
38
|
+
import { AutoImageProcessor } from "../auto/image_processing_auto.js";
|
|
39
|
+
//# sourceMappingURL=processing_florence2.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"processing_florence2.d.ts","sourceRoot":"","sources":["../../../src/models/florence2/processing_florence2.js"],"names":[],"mappings":"AAIA;IACI,6CAAsC;IACtC,wDAAiD;IAEjD,0CAuBC;IAdG,kCAAkC;IAClC,mCADW,IAAI,MAAM,EAAE,MAAM,CAAC,CAC2E;IAEzG,kCAAkC;IAClC,6BADW,IAAI,MAAM,EAAE,MAAM,CAAC,CAC+D;IAE7F,kCAAkC;IAClC,yBADW,IAAI,MAAM,EAAE,MAAM,CAAC,CACuD;IAErF;;;MAGC;IACD,qBAAwB;IAG5B;;;;OAIG;IACH,wBAHW,MAAM,GAAC,MAAM,EAAE,GACb,MAAM,EAAE,CA6BpB;IAED;;;;;OAKG;IACH,8BAJW,MAAM,QACN,MAAM,cACN,CAAC,MAAM,EAAE,MAAM,CAAC;;;;;MAsC1B;IAID,0DAaC;CACJ;0BA/HyB,gCAAgC;8BAE5B,qBAAqB;mCADhB,kCAAkC"}
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"image_processing_glpn.d.ts","sourceRoot":"","sources":["../../../src/models/glpn/image_processing_glpn.js"],"names":[],"mappings":"AAIA;CAA4D;+BAFrD,sCAAsC"}
|
|
@@ -0,0 +1,36 @@
|
|
|
1
|
+
export * from "./beit/image_processing_beit.js";
|
|
2
|
+
export * from "./bit/image_processing_bit.js";
|
|
3
|
+
export * from "./chinese_clip/image_processing_chinese_clip.js";
|
|
4
|
+
export * from "./clip/image_processing_clip.js";
|
|
5
|
+
export * from "./convnext/image_processing_convnext.js";
|
|
6
|
+
export * from "./deit/image_processing_deit.js";
|
|
7
|
+
export * from "./detr/image_processing_detr.js";
|
|
8
|
+
export * from "./donut/image_processing_donut.js";
|
|
9
|
+
export * from "./dpt/image_processing_dpt.js";
|
|
10
|
+
export * from "./efficientnet/image_processing_efficientnet.js";
|
|
11
|
+
export * from "./glpn/image_processing_glpn.js";
|
|
12
|
+
export * from "./janus/image_processing_janus.js";
|
|
13
|
+
export * from "./jina_clip/image_processing_jina_clip.js";
|
|
14
|
+
export * from "./llava_onevision/image_processing_llava_onevision.js";
|
|
15
|
+
export * from "./mask2former/image_processing_mask2former.js";
|
|
16
|
+
export * from "./maskformer/image_processing_maskformer.js";
|
|
17
|
+
export * from "./mobilenet_v1/image_processing_mobilenet_v1.js";
|
|
18
|
+
export * from "./mobilenet_v2/image_processing_mobilenet_v2.js";
|
|
19
|
+
export * from "./mobilenet_v3/image_processing_mobilenet_v3.js";
|
|
20
|
+
export * from "./mobilenet_v4/image_processing_mobilenet_v4.js";
|
|
21
|
+
export * from "./mobilevit/image_processing_mobilevit.js";
|
|
22
|
+
export * from "./nougat/image_processing_nougat.js";
|
|
23
|
+
export * from "./owlv2/image_processing_owlv2.js";
|
|
24
|
+
export * from "./owlvit/image_processing_owlvit.js";
|
|
25
|
+
export * from "./pvt/image_processing_pvt.js";
|
|
26
|
+
export * from "./qwen2_vl/image_processing_qwen2_vl.js";
|
|
27
|
+
export * from "./rt_detr/image_processing_rt_detr.js";
|
|
28
|
+
export * from "./sam/image_processing_sam.js";
|
|
29
|
+
export * from "./segformer/image_processing_segformer.js";
|
|
30
|
+
export * from "./siglip/image_processing_siglip.js";
|
|
31
|
+
export * from "./swin2sr/image_processing_swin2sr.js";
|
|
32
|
+
export * from "./vit/image_processing_vit.js";
|
|
33
|
+
export * from "./vitmatte/image_processing_vitmatte.js";
|
|
34
|
+
export * from "./vitpose/image_processing_vitpose.js";
|
|
35
|
+
export * from "./yolos/image_processing_yolos.js";
|
|
36
|
+
//# sourceMappingURL=image_processors.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"image_processors.d.ts","sourceRoot":"","sources":["../../src/models/image_processors.js"],"names":[],"mappings":""}
|
|
@@ -0,0 +1,7 @@
|
|
|
1
|
+
export class VLMImageProcessor extends ImageProcessor {
|
|
2
|
+
constructor(config: any);
|
|
3
|
+
constant_values: any;
|
|
4
|
+
pad_image(pixelData: any, imgDims: any, padSize: any, options: any): [Float32Array, number[]];
|
|
5
|
+
}
|
|
6
|
+
import { ImageProcessor } from "../../base/image_processors_utils.js";
|
|
7
|
+
//# sourceMappingURL=image_processing_janus.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"image_processing_janus.d.ts","sourceRoot":"","sources":["../../../src/models/janus/image_processing_janus.js"],"names":[],"mappings":"AAKA;IACI,yBAUC;IADG,qBAAqF;IAGzF,8FAMC;CACJ;+BAtBM,sCAAsC"}
|
|
@@ -0,0 +1,77 @@
|
|
|
1
|
+
export class VLChatProcessor extends Processor {
|
|
2
|
+
static image_processor_class: typeof AutoImageProcessor;
|
|
3
|
+
static tokenizer_class: typeof AutoTokenizer;
|
|
4
|
+
constructor(config: any, components: any);
|
|
5
|
+
image_tag: any;
|
|
6
|
+
image_start_tag: any;
|
|
7
|
+
image_end_tag: any;
|
|
8
|
+
num_image_tokens: any;
|
|
9
|
+
/**
|
|
10
|
+
* @typedef {Object} MultimodalMessageProperties Additional properties for multimodal messages.
|
|
11
|
+
* @property {(RawImage | string | URL)[]} [images] The images in the message.
|
|
12
|
+
* @typedef {(import('../../tokenizers.js').Message & MultimodalMessageProperties)[]} MultimodalConversation The conversation possibly containing multimodal inputs.
|
|
13
|
+
*/
|
|
14
|
+
/**
|
|
15
|
+
* @typedef {Object} VLCChatProcessorResult The processed input.
|
|
16
|
+
* @property {Tensor} input_ids The input IDs.
|
|
17
|
+
* @property {Tensor} attention_mask The attention mask.
|
|
18
|
+
* @property {Tensor} images_seq_mask The image sequence mask.
|
|
19
|
+
* @property {Tensor} images_emb_mask The image embedding mask.
|
|
20
|
+
*/
|
|
21
|
+
/**
|
|
22
|
+
* @param {MultimodalConversation} conversation The chat messages to process.
|
|
23
|
+
* @param {Object} options Additional options for processing.
|
|
24
|
+
* @param {RawImage|RawImage[]} [options.images] The images to process, if not set in the conversation.
|
|
25
|
+
* @param {string} [options.chat_template="default"] The chat template to use.
|
|
26
|
+
* @returns {Promise<VLCChatProcessorResult | VLCChatProcessorResult & import('../../base/image_processors_utils.js').ImageProcessorResult>} The processed input.
|
|
27
|
+
*/
|
|
28
|
+
_call(conversation: (import("../../tokenizers.js").Message & {
|
|
29
|
+
/**
|
|
30
|
+
* The images in the message.
|
|
31
|
+
*/
|
|
32
|
+
images?: (RawImage | string | URL)[];
|
|
33
|
+
})[], { images, chat_template, }?: {
|
|
34
|
+
images?: RawImage | RawImage[];
|
|
35
|
+
chat_template?: string;
|
|
36
|
+
}): Promise<{
|
|
37
|
+
/**
|
|
38
|
+
* The input IDs.
|
|
39
|
+
*/
|
|
40
|
+
input_ids: Tensor;
|
|
41
|
+
/**
|
|
42
|
+
* The attention mask.
|
|
43
|
+
*/
|
|
44
|
+
attention_mask: Tensor;
|
|
45
|
+
/**
|
|
46
|
+
* The image sequence mask.
|
|
47
|
+
*/
|
|
48
|
+
images_seq_mask: Tensor;
|
|
49
|
+
/**
|
|
50
|
+
* The image embedding mask.
|
|
51
|
+
*/
|
|
52
|
+
images_emb_mask: Tensor;
|
|
53
|
+
} | ({
|
|
54
|
+
/**
|
|
55
|
+
* The input IDs.
|
|
56
|
+
*/
|
|
57
|
+
input_ids: Tensor;
|
|
58
|
+
/**
|
|
59
|
+
* The attention mask.
|
|
60
|
+
*/
|
|
61
|
+
attention_mask: Tensor;
|
|
62
|
+
/**
|
|
63
|
+
* The image sequence mask.
|
|
64
|
+
*/
|
|
65
|
+
images_seq_mask: Tensor;
|
|
66
|
+
/**
|
|
67
|
+
* The image embedding mask.
|
|
68
|
+
*/
|
|
69
|
+
images_emb_mask: Tensor;
|
|
70
|
+
} & import("../../base/image_processors_utils.js").ImageProcessorResult)>;
|
|
71
|
+
}
|
|
72
|
+
import { Processor } from "../../base/processing_utils.js";
|
|
73
|
+
import { RawImage } from "../../utils/image.js";
|
|
74
|
+
import { Tensor } from "../../utils/tensor.js";
|
|
75
|
+
import { AutoImageProcessor } from "../auto/image_processing_auto.js";
|
|
76
|
+
import { AutoTokenizer } from "../../tokenizers.js";
|
|
77
|
+
//# sourceMappingURL=processing_janus.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"processing_janus.d.ts","sourceRoot":"","sources":["../../../src/models/janus/processing_janus.js"],"names":[],"mappings":"AAQA;IACI,wDAAiD;IACjD,6CAAsC;IAGtC,0CAOC;IAJG,eAAsC;IACtC,qBAAkD;IAClD,mBAA8C;IAC9C,sBAAoD;IAGxD;;;;OAIG;IAEH;;;;;;OAMG;IAEH;;;;;;OAMG;IACH;;;;iBAnBc,CAAC,QAAQ,GAAG,MAAM,GAAG,GAAG,CAAC,EAAE;;QAeA,MAAM,GAApC,QAAQ,GAAC,QAAQ,EAAE;QACF,aAAa,GAA9B,MAAM;;;;;mBAVH,MAAM;;;;wBACN,MAAM;;;;yBACN,MAAM;;;;yBACN,MAAM;;;;;mBAHN,MAAM;;;;wBACN,MAAM;;;;yBACN,MAAM;;;;yBACN,MAAM;8EAwFnB;CACJ;0BAzHyB,gCAAgC;yBAKjC,sBAAsB;uBADxB,uBAAuB;mCAHX,kCAAkC;8BACvC,qBAAqB"}
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"image_processing_jina_clip.d.ts","sourceRoot":"","sources":["../../../src/models/jina_clip/image_processing_jina_clip.js"],"names":[],"mappings":"AAIA;IACI,yBAmBC;CACJ;+BAvBM,sCAAsC"}
|
|
@@ -0,0 +1,9 @@
|
|
|
1
|
+
export class JinaCLIPProcessor extends Processor {
|
|
2
|
+
static tokenizer_class: typeof AutoTokenizer;
|
|
3
|
+
static image_processor_class: typeof AutoImageProcessor;
|
|
4
|
+
_call(text?: any, images?: any, kwargs?: {}): Promise<any>;
|
|
5
|
+
}
|
|
6
|
+
import { Processor } from "../../base/processing_utils.js";
|
|
7
|
+
import { AutoTokenizer } from "../../tokenizers.js";
|
|
8
|
+
import { AutoImageProcessor } from "../auto/image_processing_auto.js";
|
|
9
|
+
//# sourceMappingURL=processing_jina_clip.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"processing_jina_clip.d.ts","sourceRoot":"","sources":["../../../src/models/jina_clip/processing_jina_clip.js"],"names":[],"mappings":"AAKA;IACI,6CAAsC;IACtC,wDAAiD;IAEjD,2DAaC;CACJ;0BAtByB,gCAAgC;8BAE5B,qBAAqB;mCADhB,kCAAkC"}
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"image_processing_llava_onevision.d.ts","sourceRoot":"","sources":["../../../src/models/llava_onevision/image_processing_llava_onevision.js"],"names":[],"mappings":"AAIA;CAAmE;+BAF5D,sCAAsC"}
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"image_processing_mask2former.d.ts","sourceRoot":"","sources":["../../../src/models/mask2former/image_processing_mask2former.js"],"names":[],"mappings":"AAIA;CAA2E;yCAHlC,8CAA8C"}
|
|
@@ -0,0 +1,22 @@
|
|
|
1
|
+
export class MaskFormerImageProcessor extends ImageProcessor {
|
|
2
|
+
post_process_panoptic_segmentation(outputs: any, threshold?: number, mask_threshold?: number, overlap_mask_area_threshold?: number, label_ids_to_fuse?: Set<number>, target_sizes?: [number, number][]): {
|
|
3
|
+
segmentation: import("../../transformers.js").Tensor;
|
|
4
|
+
segments_info: {
|
|
5
|
+
id: number;
|
|
6
|
+
label_id: number;
|
|
7
|
+
score: number;
|
|
8
|
+
}[];
|
|
9
|
+
}[];
|
|
10
|
+
post_process_instance_segmentation(outputs: any, threshold?: number, target_sizes?: [number, number][]): {
|
|
11
|
+
segmentation: import("../../transformers.js").Tensor;
|
|
12
|
+
segments_info: {
|
|
13
|
+
id: number;
|
|
14
|
+
label_id: number;
|
|
15
|
+
score: number;
|
|
16
|
+
}[];
|
|
17
|
+
}[];
|
|
18
|
+
}
|
|
19
|
+
export class MaskFormerFeatureExtractor extends MaskFormerImageProcessor {
|
|
20
|
+
}
|
|
21
|
+
import { ImageProcessor } from "../../base/image_processors_utils.js";
|
|
22
|
+
//# sourceMappingURL=image_processing_maskformer.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"image_processing_maskformer.d.ts","sourceRoot":"","sources":["../../../src/models/maskformer/image_processing_maskformer.js"],"names":[],"mappings":"AAMA;IAYure;;;;;;;QAA4qG;IAAA;;;;;;;QAAktB;CAFpjmB;AACD;CAA4E;+BAbrE,sCAAsC"}
|
|
@@ -0,0 +1,64 @@
|
|
|
1
|
+
export class MgpstrProcessor extends Processor {
|
|
2
|
+
static tokenizer_class: typeof AutoTokenizer;
|
|
3
|
+
static image_processor_class: typeof AutoImageProcessor;
|
|
4
|
+
/**
|
|
5
|
+
* @returns {import('../../tokenizers.js').MgpstrTokenizer} The character tokenizer.
|
|
6
|
+
*/
|
|
7
|
+
get char_tokenizer(): import("../../tokenizers.js").MgpstrTokenizer;
|
|
8
|
+
/**
|
|
9
|
+
* @returns {import('../../tokenizers.js').GPT2Tokenizer} The BPE tokenizer.
|
|
10
|
+
*/
|
|
11
|
+
get bpe_tokenizer(): import("../../tokenizers.js").GPT2Tokenizer;
|
|
12
|
+
/**
|
|
13
|
+
* @returns {import('../../tokenizers.js').BertTokenizer} The WordPiece tokenizer.
|
|
14
|
+
*/
|
|
15
|
+
get wp_tokenizer(): import("../../tokenizers.js").BertTokenizer;
|
|
16
|
+
/**
|
|
17
|
+
* Helper function to decode the model prediction logits.
|
|
18
|
+
* @param {import('../../utils/tensor.js').Tensor} pred_logits Model prediction logits.
|
|
19
|
+
* @param {string} format Type of model prediction. Must be one of ['char', 'bpe', 'wp'].
|
|
20
|
+
* @returns {[string[], number[]]} The decoded sentences and their confidence scores.
|
|
21
|
+
*/
|
|
22
|
+
_decode_helper(pred_logits: import('../../utils/tensor.js').Tensor, format: string): [string[], number[]];
|
|
23
|
+
/**
|
|
24
|
+
* Convert a list of lists of char token ids into a list of strings by calling char tokenizer.
|
|
25
|
+
* @param {number[][]} sequences List of tokenized input ids.
|
|
26
|
+
* @returns {string[]} The list of char decoded sentences.
|
|
27
|
+
*/
|
|
28
|
+
char_decode(sequences: number[][]): string[];
|
|
29
|
+
/**
|
|
30
|
+
* Convert a list of lists of BPE token ids into a list of strings by calling BPE tokenizer.
|
|
31
|
+
* @param {number[][]} sequences List of tokenized input ids.
|
|
32
|
+
* @returns {string[]} The list of BPE decoded sentences.
|
|
33
|
+
*/
|
|
34
|
+
bpe_decode(sequences: number[][]): string[];
|
|
35
|
+
/**
|
|
36
|
+
* Convert a list of lists of word piece token ids into a list of strings by calling word piece tokenizer.
|
|
37
|
+
* @param {number[][]} sequences List of tokenized input ids.
|
|
38
|
+
* @returns {string[]} The list of wp decoded sentences.
|
|
39
|
+
*/
|
|
40
|
+
wp_decode(sequences: number[][]): string[];
|
|
41
|
+
/**
|
|
42
|
+
* Convert a list of lists of token ids into a list of strings by calling decode.
|
|
43
|
+
* @param {import('../../utils/tensor.js').Tensor[]} sequences List of tokenized input ids.
|
|
44
|
+
* @returns {{generated_text: string[], scores: number[], char_preds: string[], bpe_preds: string[], wp_preds: string[]}}
|
|
45
|
+
* Dictionary of all the outputs of the decoded results.
|
|
46
|
+
* - generated_text: The final results after fusion of char, bpe, and wp.
|
|
47
|
+
* - scores: The final scores after fusion of char, bpe, and wp.
|
|
48
|
+
* - char_preds: The list of character decoded sentences.
|
|
49
|
+
* - bpe_preds: The list of BPE decoded sentences.
|
|
50
|
+
* - wp_preds: The list of wp decoded sentences.
|
|
51
|
+
*/
|
|
52
|
+
batch_decode([char_logits, bpe_logits, wp_logits]: import('../../utils/tensor.js').Tensor[]): {
|
|
53
|
+
generated_text: string[];
|
|
54
|
+
scores: number[];
|
|
55
|
+
char_preds: string[];
|
|
56
|
+
bpe_preds: string[];
|
|
57
|
+
wp_preds: string[];
|
|
58
|
+
};
|
|
59
|
+
_call(images: any, text?: any): Promise<any>;
|
|
60
|
+
}
|
|
61
|
+
import { Processor } from "../../base/processing_utils.js";
|
|
62
|
+
import { AutoTokenizer } from "../../tokenizers.js";
|
|
63
|
+
import { AutoImageProcessor } from "../auto/image_processing_auto.js";
|
|
64
|
+
//# sourceMappingURL=processing_mgp_str.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"processing_mgp_str.d.ts","sourceRoot":"","sources":["../../../src/models/mgp_str/processing_mgp_str.js"],"names":[],"mappings":"AAUA;IACI,6CAAsC;IACtC,wDAAiD;IAEjD;;OAEG;IACH,oEAEC;IAED;;OAEG;IACH,iEAEC;IAED;;OAEG;IACH,gEAEC;IAED;;;;;OAKG;IACH,4BAJW,OAAO,uBAAuB,EAAE,MAAM,UACtC,MAAM,GACJ,CAAC,MAAM,EAAE,EAAE,MAAM,EAAE,CAAC,CA0ChC;IAED;;;;OAIG;IACH,uBAHW,MAAM,EAAE,EAAE,GACR,MAAM,EAAE,CAIpB;IAED;;;;OAIG;IACH,sBAHW,MAAM,EAAE,EAAE,GACR,MAAM,EAAE,CAIpB;IAED;;;;OAIG;IACH,qBAHW,MAAM,EAAE,EAAE,GACR,MAAM,EAAE,CAIpB;IAED;;;;;;;;;;OAUG;IACH,mDATW,OAAO,uBAAuB,EAAE,MAAM,EAAE;wBACrB,MAAM,EAAE;gBAAU,MAAM,EAAE;oBAAc,MAAM,EAAE;mBAAa,MAAM,EAAE;kBAAY,MAAM,EAAE;MA4BtH;IAmBD,6CAQC;CACJ;0BAzKyB,gCAAgC;8BAE5B,qBAAqB;mCADhB,kCAAkC"}
|
|
@@ -0,0 +1,6 @@
|
|
|
1
|
+
export class MobileNetV1ImageProcessor extends ImageProcessor {
|
|
2
|
+
}
|
|
3
|
+
export class MobileNetV1FeatureExtractor extends MobileNetV1ImageProcessor {
|
|
4
|
+
}
|
|
5
|
+
import { ImageProcessor } from "../../base/image_processors_utils.js";
|
|
6
|
+
//# sourceMappingURL=image_processing_mobilenet_v1.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"image_processing_mobilenet_v1.d.ts","sourceRoot":"","sources":["../../../src/models/mobilenet_v1/image_processing_mobilenet_v1.js"],"names":[],"mappings":"AAKA;CAAiE;AACjE;CAA8E;+BAJvE,sCAAsC"}
|
|
@@ -0,0 +1,6 @@
|
|
|
1
|
+
export class MobileNetV2ImageProcessor extends ImageProcessor {
|
|
2
|
+
}
|
|
3
|
+
export class MobileNetV2FeatureExtractor extends MobileNetV2ImageProcessor {
|
|
4
|
+
}
|
|
5
|
+
import { ImageProcessor } from "../../base/image_processors_utils.js";
|
|
6
|
+
//# sourceMappingURL=image_processing_mobilenet_v2.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"image_processing_mobilenet_v2.d.ts","sourceRoot":"","sources":["../../../src/models/mobilenet_v2/image_processing_mobilenet_v2.js"],"names":[],"mappings":"AAKA;CAAiE;AACjE;CAA8E;+BAJvE,sCAAsC"}
|
|
@@ -0,0 +1,6 @@
|
|
|
1
|
+
export class MobileNetV3ImageProcessor extends ImageProcessor {
|
|
2
|
+
}
|
|
3
|
+
export class MobileNetV3FeatureExtractor extends MobileNetV3ImageProcessor {
|
|
4
|
+
}
|
|
5
|
+
import { ImageProcessor } from "../../base/image_processors_utils.js";
|
|
6
|
+
//# sourceMappingURL=image_processing_mobilenet_v3.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"image_processing_mobilenet_v3.d.ts","sourceRoot":"","sources":["../../../src/models/mobilenet_v3/image_processing_mobilenet_v3.js"],"names":[],"mappings":"AAKA;CAAiE;AACjE;CAA8E;+BAJvE,sCAAsC"}
|
|
@@ -0,0 +1,6 @@
|
|
|
1
|
+
export class MobileNetV4ImageProcessor extends ImageProcessor {
|
|
2
|
+
}
|
|
3
|
+
export class MobileNetV4FeatureExtractor extends MobileNetV4ImageProcessor {
|
|
4
|
+
}
|
|
5
|
+
import { ImageProcessor } from "../../base/image_processors_utils.js";
|
|
6
|
+
//# sourceMappingURL=image_processing_mobilenet_v4.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"image_processing_mobilenet_v4.d.ts","sourceRoot":"","sources":["../../../src/models/mobilenet_v4/image_processing_mobilenet_v4.js"],"names":[],"mappings":"AAKA;CAAiE;AACjE;CAA8E;+BAJvE,sCAAsC"}
|
|
@@ -0,0 +1,6 @@
|
|
|
1
|
+
export class MobileViTImageProcessor extends ImageProcessor {
|
|
2
|
+
}
|
|
3
|
+
export class MobileViTFeatureExtractor extends MobileViTImageProcessor {
|
|
4
|
+
}
|
|
5
|
+
import { ImageProcessor } from "../../base/image_processors_utils.js";
|
|
6
|
+
//# sourceMappingURL=image_processing_mobilevit.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"image_processing_mobilevit.d.ts","sourceRoot":"","sources":["../../../src/models/mobilevit/image_processing_mobilevit.js"],"names":[],"mappings":"AAIA;CAA+D;AAC/D;CAA0E;+BAHnE,sCAAsC"}
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"image_processing_nougat.d.ts","sourceRoot":"","sources":["../../../src/models/nougat/image_processing_nougat.js"],"names":[],"mappings":"AAIA;CAAiE;oCAH7B,oCAAoC"}
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"image_processing_owlv2.d.ts","sourceRoot":"","sources":["../../../src/models/owlv2/image_processing_owlv2.js"],"names":[],"mappings":"AAIA;CAAiE;qCAH5B,sCAAsC"}
|
|
@@ -0,0 +1,10 @@
|
|
|
1
|
+
export class OwlViTImageProcessor extends ImageProcessor {
|
|
2
|
+
post_process_object_detection(outputs: {
|
|
3
|
+
logits: import("../../transformers.js").Tensor;
|
|
4
|
+
pred_boxes: import("../../transformers.js").Tensor;
|
|
5
|
+
}, threshold?: number, target_sizes?: [number, number][], is_zero_shot?: boolean): any[];
|
|
6
|
+
}
|
|
7
|
+
export class OwlViTFeatureExtractor extends OwlViTImageProcessor {
|
|
8
|
+
}
|
|
9
|
+
import { ImageProcessor } from "../../base/image_processors_utils.js";
|
|
10
|
+
//# sourceMappingURL=image_processing_owlvit.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"image_processing_owlvit.d.ts","sourceRoot":"","sources":["../../../src/models/owlvit/image_processing_owlvit.js"],"names":[],"mappings":"AAKA;IAO4tE;;;6FAAu6F;CAFloK;AACD;CAAoE;+BAR7D,sCAAsC"}
|
|
@@ -0,0 +1,8 @@
|
|
|
1
|
+
export class OwlViTProcessor extends Processor {
|
|
2
|
+
static tokenizer_class: typeof AutoTokenizer;
|
|
3
|
+
static image_processor_class: typeof AutoImageProcessor;
|
|
4
|
+
}
|
|
5
|
+
import { Processor } from "../../base/processing_utils.js";
|
|
6
|
+
import { AutoTokenizer } from "../../tokenizers.js";
|
|
7
|
+
import { AutoImageProcessor } from "../auto/image_processing_auto.js";
|
|
8
|
+
//# sourceMappingURL=processing_owlvit.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"processing_owlvit.d.ts","sourceRoot":"","sources":["../../../src/models/owlvit/processing_owlvit.js"],"names":[],"mappings":"AAGA;IACI,6CAAsC;IACtC,wDAAiD;CACpD;0BANyB,gCAAgC;8BAE5B,qBAAqB;mCADhB,kCAAkC"}
|
|
@@ -0,0 +1,12 @@
|
|
|
1
|
+
export * from "./florence2/processing_florence2.js";
|
|
2
|
+
export * from "./mgp_str/processing_mgp_str.js";
|
|
3
|
+
export * from "./janus/processing_janus.js";
|
|
4
|
+
export * from "./jina_clip/processing_jina_clip.js";
|
|
5
|
+
export * from "./owlvit/processing_owlvit.js";
|
|
6
|
+
export * from "./pyannote/processing_pyannote.js";
|
|
7
|
+
export * from "./qwen2_vl/processing_qwen2_vl.js";
|
|
8
|
+
export * from "./sam/processing_sam.js";
|
|
9
|
+
export * from "./speecht5/processing_speecht5.js";
|
|
10
|
+
export * from "./wav2vec2/processing_wav2vec2.js";
|
|
11
|
+
export * from "./whisper/processing_whisper.js";
|
|
12
|
+
//# sourceMappingURL=processors.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"processors.d.ts","sourceRoot":"","sources":["../../src/models/processors.js"],"names":[],"mappings":""}
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"image_processing_pvt.d.ts","sourceRoot":"","sources":["../../../src/models/pvt/image_processing_pvt.js"],"names":[],"mappings":"AAIA;CAAyD;+BAFlD,sCAAsC"}
|
|
@@ -0,0 +1,13 @@
|
|
|
1
|
+
export class PyAnnoteFeatureExtractor extends FeatureExtractor {
|
|
2
|
+
/**
|
|
3
|
+
* Asynchronously extracts features from a given audio using the provided configuration.
|
|
4
|
+
* @param {Float32Array|Float64Array} audio The audio data as a Float32Array/Float64Array.
|
|
5
|
+
* @returns {Promise<{ input_values: Tensor; }>} The extracted input features.
|
|
6
|
+
*/
|
|
7
|
+
_call(audio: Float32Array | Float64Array): Promise<{
|
|
8
|
+
input_values: Tensor;
|
|
9
|
+
}>;
|
|
10
|
+
}
|
|
11
|
+
import { FeatureExtractor } from '../../base/feature_extraction_utils.js';
|
|
12
|
+
import { Tensor } from '../../utils/tensor.js';
|
|
13
|
+
//# sourceMappingURL=feature_extraction_pyannote.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"feature_extraction_pyannote.d.ts","sourceRoot":"","sources":["../../../src/models/pyannote/feature_extraction_pyannote.js"],"names":[],"mappings":"AAIA;IACI;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY,GACvB,QAAQ;QAAE,YAAY,EAAE,MAAM,CAAC;KAAE,CAAC,CAiB9C;CAEJ;iCA3BuD,wCAAwC;uBACzE,uBAAuB"}
|