@huggingface/transformers 3.1.0 → 3.1.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +7 -3
- package/dist/ort-wasm-simd-threaded.jsep.wasm +0 -0
- package/dist/transformers.cjs +965 -195
- package/dist/transformers.cjs.map +1 -1
- package/dist/transformers.js +2251 -1360
- package/dist/transformers.js.map +1 -1
- package/dist/transformers.min.cjs +1 -352
- package/dist/transformers.min.cjs.map +1 -1
- package/dist/transformers.min.js +1 -415
- package/dist/transformers.min.js.map +1 -1
- package/dist/transformers.min.mjs +1 -352
- package/dist/transformers.min.mjs.map +1 -1
- package/dist/transformers.mjs +979 -194
- package/dist/transformers.mjs.map +1 -1
- package/package.json +11 -16
- package/src/backends/onnx.js +2 -7
- package/src/configs.js +3 -1
- package/src/env.js +6 -6
- package/src/generation/configuration_utils.js +7 -0
- package/src/generation/logits_process.js +22 -16
- package/src/generation/streamers.js +7 -2
- package/src/models/idefics3/image_processing_idefics3.js +219 -0
- package/src/models/idefics3/processing_idefics3.js +136 -0
- package/src/models/image_processors.js +1 -0
- package/src/models/paligemma/processing_paligemma.js +82 -0
- package/src/models/processors.js +2 -0
- package/src/models.js +169 -39
- package/src/tokenizers.js +12 -1
- package/src/utils/core.js +53 -9
- package/src/utils/dtypes.js +2 -1
- package/src/utils/hub.js +8 -12
- package/src/utils/image.js +59 -16
- package/src/utils/tensor.js +6 -1
- package/types/backends/onnx.d.ts +2 -2
- package/types/backends/onnx.d.ts.map +1 -1
- package/types/base/feature_extraction_utils.d.ts +1 -1
- package/types/base/feature_extraction_utils.d.ts.map +1 -1
- package/types/base/image_processors_utils.d.ts +2 -2
- package/types/base/image_processors_utils.d.ts.map +1 -1
- package/types/base/processing_utils.d.ts +4 -4
- package/types/base/processing_utils.d.ts.map +1 -1
- package/types/configs.d.ts +7 -7
- package/types/configs.d.ts.map +1 -1
- package/types/env.d.ts +2 -2
- package/types/env.d.ts.map +1 -1
- package/types/generation/configuration_utils.d.ts +7 -1
- package/types/generation/configuration_utils.d.ts.map +1 -1
- package/types/generation/logits_process.d.ts +32 -22
- package/types/generation/logits_process.d.ts.map +1 -1
- package/types/generation/logits_sampler.d.ts.map +1 -1
- package/types/generation/parameters.d.ts +5 -5
- package/types/generation/stopping_criteria.d.ts +1 -1
- package/types/generation/stopping_criteria.d.ts.map +1 -1
- package/types/generation/streamers.d.ts +15 -10
- package/types/generation/streamers.d.ts.map +1 -1
- package/types/models/audio_spectrogram_transformer/feature_extraction_audio_spectrogram_transformer.d.ts +1 -1
- package/types/models/audio_spectrogram_transformer/feature_extraction_audio_spectrogram_transformer.d.ts.map +1 -1
- package/types/models/auto/feature_extraction_auto.d.ts.map +1 -1
- package/types/models/auto/image_processing_auto.d.ts.map +1 -1
- package/types/models/auto/processing_auto.d.ts +1 -1
- package/types/models/auto/processing_auto.d.ts.map +1 -1
- package/types/models/clap/feature_extraction_clap.d.ts +1 -1
- package/types/models/clap/feature_extraction_clap.d.ts.map +1 -1
- package/types/models/detr/image_processing_detr.d.ts +11 -11
- package/types/models/detr/image_processing_detr.d.ts.map +1 -1
- package/types/models/donut/image_processing_donut.d.ts +1 -1
- package/types/models/donut/image_processing_donut.d.ts.map +1 -1
- package/types/models/florence2/processing_florence2.d.ts.map +1 -1
- package/types/models/idefics3/image_processing_idefics3.d.ts +40 -0
- package/types/models/idefics3/image_processing_idefics3.d.ts.map +1 -0
- package/types/models/idefics3/processing_idefics3.d.ts +19 -0
- package/types/models/idefics3/processing_idefics3.d.ts.map +1 -0
- package/types/models/image_processors.d.ts +1 -0
- package/types/models/janus/image_processing_janus.d.ts +1 -1
- package/types/models/janus/image_processing_janus.d.ts.map +1 -1
- package/types/models/janus/processing_janus.d.ts.map +1 -1
- package/types/models/maskformer/image_processing_maskformer.d.ts +8 -8
- package/types/models/maskformer/image_processing_maskformer.d.ts.map +1 -1
- package/types/models/mgp_str/processing_mgp_str.d.ts +2 -2
- package/types/models/mgp_str/processing_mgp_str.d.ts.map +1 -1
- package/types/models/owlvit/image_processing_owlvit.d.ts.map +1 -1
- package/types/models/paligemma/processing_paligemma.d.ts +12 -0
- package/types/models/paligemma/processing_paligemma.d.ts.map +1 -0
- package/types/models/processors.d.ts +2 -0
- package/types/models/pyannote/feature_extraction_pyannote.d.ts.map +1 -1
- package/types/models/pyannote/processing_pyannote.d.ts +1 -1
- package/types/models/pyannote/processing_pyannote.d.ts.map +1 -1
- package/types/models/qwen2_vl/processing_qwen2_vl.d.ts.map +1 -1
- package/types/models/rt_detr/image_processing_rt_detr.d.ts.map +1 -1
- package/types/models/sam/image_processing_sam.d.ts.map +1 -1
- package/types/models/seamless_m4t/feature_extraction_seamless_m4t.d.ts +1 -1
- package/types/models/seamless_m4t/feature_extraction_seamless_m4t.d.ts.map +1 -1
- package/types/models/segformer/image_processing_segformer.d.ts.map +1 -1
- package/types/models/speecht5/processing_speecht5.d.ts.map +1 -1
- package/types/models/swin2sr/image_processing_swin2sr.d.ts +1 -1
- package/types/models/swin2sr/image_processing_swin2sr.d.ts.map +1 -1
- package/types/models/vitmatte/image_processing_vitmatte.d.ts.map +1 -1
- package/types/models/vitpose/image_processing_vitpose.d.ts +1 -1
- package/types/models/vitpose/image_processing_vitpose.d.ts.map +1 -1
- package/types/models/wav2vec2/feature_extraction_wav2vec2.d.ts.map +1 -1
- package/types/models/wav2vec2/processing_wav2vec2.d.ts.map +1 -1
- package/types/models/wespeaker/feature_extraction_wespeaker.d.ts +1 -1
- package/types/models/wespeaker/feature_extraction_wespeaker.d.ts.map +1 -1
- package/types/models/whisper/feature_extraction_whisper.d.ts +1 -1
- package/types/models/whisper/feature_extraction_whisper.d.ts.map +1 -1
- package/types/models/whisper/generation_whisper.d.ts.map +1 -1
- package/types/models/whisper/processing_whisper.d.ts.map +1 -1
- package/types/models/yolos/image_processing_yolos.d.ts.map +1 -1
- package/types/models.d.ts +44 -10
- package/types/models.d.ts.map +1 -1
- package/types/ops/registry.d.ts.map +1 -1
- package/types/pipelines.d.ts +26 -51
- package/types/pipelines.d.ts.map +1 -1
- package/types/tokenizers.d.ts +10 -6
- package/types/tokenizers.d.ts.map +1 -1
- package/types/utils/audio.d.ts.map +1 -1
- package/types/utils/constants.d.ts.map +1 -1
- package/types/utils/core.d.ts +94 -22
- package/types/utils/core.d.ts.map +1 -1
- package/types/utils/data-structures.d.ts.map +1 -1
- package/types/utils/devices.d.ts.map +1 -1
- package/types/utils/dtypes.d.ts +3 -2
- package/types/utils/dtypes.d.ts.map +1 -1
- package/types/utils/generic.d.ts.map +1 -1
- package/types/utils/hub.d.ts +3 -3
- package/types/utils/hub.d.ts.map +1 -1
- package/types/utils/image.d.ts +14 -1
- package/types/utils/image.d.ts.map +1 -1
- package/types/utils/maths.d.ts +10 -10
- package/types/utils/maths.d.ts.map +1 -1
- package/types/utils/tensor.d.ts +10 -8
- package/types/utils/tensor.d.ts.map +1 -1
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"stopping_criteria.d.ts","sourceRoot":"","sources":["../../src/generation/stopping_criteria.js"],"names":[],"mappings":"
|
|
1
|
+
{"version":3,"file":"stopping_criteria.d.ts","sourceRoot":"","sources":["../../src/generation/stopping_criteria.js"],"names":[],"mappings":";KAQ4C,GAAG;UAA0B,GAAE;;AAE3E;;GAEG;AACH;IACI;;;;;;;;OAQG;IACH,iBAPW,MAAM,EAAE,EAAE,UAEV,MAAM,EAAE,EAAE,GAGR,OAAO,EAAE,CAIrB;CACJ;;KAlB2C,GAAG;UAA0B,GAAE;;AAmB3E;GACG;AACH;IAMQ,gBAAkB;IAGtB;;;;OAIG;IACH,WAFW,gBAAgB,QAI1B;IAED;;;;OAIG;IACH,cAFW,gBAAgB,GAAC,oBAAoB,GAAC,gBAAgB,EAAE,QASlE;IAED,0CASC;IAED,wCAEC;CACJ;AAED;;;GAGG;AACH;IAEI;;;;OAIG;IACH,wBAHW,MAAM,4BACN,MAAM,EAMhB;IAFG,mBAA4B;IAC5B,gCAAsD;IAG1D,2BAEC;CACJ;AAID;;;GAGG;AACH;IAEI;;;;OAIG;IACH,0BAHW,MAAM,GAAC,MAAM,EAAE,EASzB;IADG,uBAAgC;CAgBvC;AAED;;GAEG;AACH;IAGQ,qBAAwB;IAG5B,kBAEC;IAED,cAEC;IAED,0CAEC;CACJ"}
|
|
@@ -16,18 +16,23 @@ export class TextStreamer extends BaseStreamer {
|
|
|
16
16
|
/**
|
|
17
17
|
*
|
|
18
18
|
* @param {import('../tokenizers.js').PreTrainedTokenizer} tokenizer
|
|
19
|
+
* @param {Object} options
|
|
20
|
+
* @param {boolean} [options.skip_prompt=false] Whether to skip the prompt tokens
|
|
21
|
+
* @param {function(string): void} [options.callback_function=null] Function to call when a piece of text is ready to display
|
|
22
|
+
* @param {function(bigint[]): void} [options.token_callback_function=null] Function to call when a new token is generated
|
|
23
|
+
* @param {Object} [options.decode_kwargs={}] Additional keyword arguments to pass to the tokenizer's decode method
|
|
19
24
|
*/
|
|
20
|
-
constructor(tokenizer: import(
|
|
25
|
+
constructor(tokenizer: import("../tokenizers.js").PreTrainedTokenizer, { skip_prompt, callback_function, token_callback_function, decode_kwargs, ...kwargs }?: {
|
|
21
26
|
skip_prompt?: boolean;
|
|
22
|
-
callback_function?:
|
|
23
|
-
token_callback_function?:
|
|
24
|
-
decode_kwargs?:
|
|
27
|
+
callback_function?: (arg0: string) => void;
|
|
28
|
+
token_callback_function?: (arg0: bigint[]) => void;
|
|
29
|
+
decode_kwargs?: any;
|
|
25
30
|
});
|
|
26
31
|
tokenizer: import("../tokenizers.js").PreTrainedTokenizer;
|
|
27
32
|
skip_prompt: boolean;
|
|
28
|
-
callback_function: any;
|
|
29
|
-
token_callback_function:
|
|
30
|
-
decode_kwargs:
|
|
33
|
+
callback_function: (x: any) => void;
|
|
34
|
+
token_callback_function: (arg0: bigint[]) => void;
|
|
35
|
+
decode_kwargs: any;
|
|
31
36
|
token_cache: any[];
|
|
32
37
|
print_len: number;
|
|
33
38
|
next_tokens_are_prompt: boolean;
|
|
@@ -52,7 +57,7 @@ export class WhisperTextStreamer extends TextStreamer {
|
|
|
52
57
|
* @param {Object} options
|
|
53
58
|
* @param {boolean} [options.skip_prompt=false] Whether to skip the prompt tokens
|
|
54
59
|
* @param {function(string): void} [options.callback_function=null] Function to call when a piece of text is ready to display
|
|
55
|
-
* @param {function(
|
|
60
|
+
* @param {function(bigint[]): void} [options.token_callback_function=null] Function to call when a new token is generated
|
|
56
61
|
* @param {function(number): void} [options.on_chunk_start=null] Function to call when a new chunk starts
|
|
57
62
|
* @param {function(number): void} [options.on_chunk_end=null] Function to call when a chunk ends
|
|
58
63
|
* @param {function(): void} [options.on_finalize=null] Function to call when the stream is finalized
|
|
@@ -60,10 +65,10 @@ export class WhisperTextStreamer extends TextStreamer {
|
|
|
60
65
|
* @param {boolean} [options.skip_special_tokens=true] Whether to skip special tokens when decoding
|
|
61
66
|
* @param {Object} [options.decode_kwargs={}] Additional keyword arguments to pass to the tokenizer's decode method
|
|
62
67
|
*/
|
|
63
|
-
constructor(tokenizer: import(
|
|
68
|
+
constructor(tokenizer: import("../tokenizers.js").WhisperTokenizer, { skip_prompt, callback_function, token_callback_function, on_chunk_start, on_chunk_end, on_finalize, time_precision, skip_special_tokens, decode_kwargs, }?: {
|
|
64
69
|
skip_prompt?: boolean;
|
|
65
70
|
callback_function?: (arg0: string) => void;
|
|
66
|
-
token_callback_function?: (arg0:
|
|
71
|
+
token_callback_function?: (arg0: bigint[]) => void;
|
|
67
72
|
on_chunk_start?: (arg0: number) => void;
|
|
68
73
|
on_chunk_end?: (arg0: number) => void;
|
|
69
74
|
on_finalize?: () => void;
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"streamers.d.ts","sourceRoot":"","sources":["../../src/generation/streamers.js"],"names":[],"mappings":"AASA;IACI;;;OAGG;IACH,WAFW,MAAM,EAAE,EAAE,QAIpB;IAED;;OAEG;IACH,YAEC;CACJ;AAMD;;GAEG;AACH;IACI
|
|
1
|
+
{"version":3,"file":"streamers.d.ts","sourceRoot":"","sources":["../../src/generation/streamers.js"],"names":[],"mappings":"AASA;IACI;;;OAGG;IACH,WAFW,MAAM,EAAE,EAAE,QAIpB;IAED;;OAEG;IACH,YAEC;CACJ;AAMD;;GAEG;AACH;IACI;;;;;;;;OAQG;IACH,uBAPW,OAAO,kBAAkB,EAAE,mBAAmB,0FAEtD;QAA0B,WAAW,GAA7B,OAAO;QAC0B,iBAAiB,GAAlD,CAAS,IAAM,EAAN,MAAM,KAAG,IAAI;QACa,uBAAuB,GAA1D,CAAS,IAAQ,EAAR,MAAM,EAAE,KAAG,IAAI;QACP,aAAa;KACxC,EAmBA;IAVG,0DAA0B;IAC1B,qBAA8B;IAC9B,oCAA0D;IAC1D,gCAdgB,MAAM,EAAE,KAAG,IAAI,CAcuB;IACtD,mBAAoD;IAGpD,mBAAqB;IACrB,kBAAkB;IAClB,gCAAkC;IA6DtC;;;;OAIG;IACH,wBAHW,MAAM,cACN,OAAO,QASjB;CACJ;AAED;;;;;;;GAOG;AACH;IACI;;;;;;;;;;;;OAYG;IACH,uBAZW,OAAO,kBAAkB,EAAE,gBAAgB,gKAEnD;QAA0B,WAAW,GAA7B,OAAO;QAC0B,iBAAiB,GAAlD,CAAS,IAAM,EAAN,MAAM,KAAG,IAAI;QACa,uBAAuB,GAA1D,CAAS,IAAQ,EAAR,MAAM,EAAE,KAAG,IAAI;QACS,cAAc,GAA/C,CAAS,IAAM,EAAN,MAAM,KAAG,IAAI;QACW,YAAY,GAA7C,CAAS,IAAM,EAAN,MAAM,KAAG,IAAI;QACK,WAAW,GAAtC,MAAY,IAAI;QACC,cAAc,GAA/B,MAAM;QACY,mBAAmB,GAArC,OAAO;QACU,aAAa;KACxC,EA2BA;IATG,wBAAgD;IAEhD,uBA1BgB,MAAM,KAAG,IAAI,CA0BO;IACpC,qBA1BgB,MAAM,KAAG,IAAI,CA0BG;IAChC,mBA1BmB,IAAI,CA0BO;IAE9B,uBAAoC;IAEpC,+BAAkC;CAiCzC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"feature_extraction_audio_spectrogram_transformer.d.ts","sourceRoot":"","sources":["../../../src/models/audio_spectrogram_transformer/feature_extraction_audio_spectrogram_transformer.js"],"names":[],"mappings":"AAKA;IAEI,yBA2BC;IARG,wBAA8B;IAE9B,
|
|
1
|
+
{"version":3,"file":"feature_extraction_audio_spectrogram_transformer.d.ts","sourceRoot":"","sources":["../../../src/models/audio_spectrogram_transformer/feature_extraction_audio_spectrogram_transformer.js"],"names":[],"mappings":"AAKA;IAEI,yBA2BC;IARG,wBAA8B;IAE9B,sCAEE;IAEF,UAA4B;IAC5B,SAA0B;IAG9B;;;;;OAKG;IACH,kCAJW,YAAY,GAAC,YAAY,cACzB,MAAM,GACJ,OAAO,CAAC,MAAM,CAAC,CAwB3B;IAGD;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY,GACvB,OAAO,CAAC;QAAE,YAAY,EAAE,MAAM,CAAA;KAAE,CAAC,CAkB7C;CACJ;iCAzFuD,wCAAwC;uBACzE,uBAAuB"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"feature_extraction_auto.d.ts","sourceRoot":"","sources":["../../../src/models/auto/feature_extraction_auto.js"],"names":[],"mappings":"AAMA;
|
|
1
|
+
{"version":3,"file":"feature_extraction_auto.d.ts","sourceRoot":"","sources":["../../../src/models/auto/feature_extraction_auto.js"],"names":[],"mappings":"AAMA;kFAe6F,oBAE7F;CAiBC;iCArCgC,wCAAwC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"image_processing_auto.d.ts","sourceRoot":"","sources":["../../../src/models/auto/image_processing_auto.js"],"names":[],"mappings":"AAMA;
|
|
1
|
+
{"version":3,"file":"image_processing_auto.d.ts","sourceRoot":"","sources":["../../../src/models/auto/image_processing_auto.js"],"names":[],"mappings":"AAMA;kFAuBk7xC,oBAAiB;CADl8xC;+BAzB8B,sCAAsC"}
|
|
@@ -29,7 +29,7 @@
|
|
|
29
29
|
* ```
|
|
30
30
|
*/
|
|
31
31
|
export class AutoProcessor {
|
|
32
|
-
static from_pretrained(pretrained_model_name_or_path: string, options:
|
|
32
|
+
static from_pretrained(pretrained_model_name_or_path: string, options: PretrainedProcessorOptions): Promise<Processor>;
|
|
33
33
|
}
|
|
34
34
|
import { Processor } from '../../base/processing_utils.js';
|
|
35
35
|
//# sourceMappingURL=processing_auto.d.ts.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"processing_auto.d.ts","sourceRoot":"","sources":["../../../src/models/auto/processing_auto.js"],"names":[],"mappings":"AAUA;;;;;;;;;;;;;;;;;;;;;;;;;;;;;GA6BG;AACH;
|
|
1
|
+
{"version":3,"file":"processing_auto.d.ts","sourceRoot":"","sources":["../../../src/models/auto/processing_auto.js"],"names":[],"mappings":"AAUA;;;;;;;;;;;;;;;;;;;;;;;;;;;;;GA6BG;AACH;;CA2DC;0BA/FyB,gCAAgC"}
|
|
@@ -2,7 +2,7 @@ export class ClapFeatureExtractor extends FeatureExtractor {
|
|
|
2
2
|
constructor(config: any);
|
|
3
3
|
mel_filters: number[][];
|
|
4
4
|
mel_filters_slaney: number[][];
|
|
5
|
-
window: Float64Array
|
|
5
|
+
window: Float64Array<ArrayBufferLike>;
|
|
6
6
|
/**
|
|
7
7
|
* Extracts the mel spectrogram and prepares it for the mode based on the `truncation` and `padding` arguments.
|
|
8
8
|
*
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"feature_extraction_clap.d.ts","sourceRoot":"","sources":["../../../src/models/clap/feature_extraction_clap.js"],"names":[],"mappings":"AAKA;IAEI,yBAyBC;IAtBG,wBAQC;IAED,+BAQC;IAED,
|
|
1
|
+
{"version":3,"file":"feature_extraction_clap.d.ts","sourceRoot":"","sources":["../../../src/models/clap/feature_extraction_clap.js"],"names":[],"mappings":"AAKA;IAEI,yBAyBC;IAtBG,wBAQC;IAED,+BAQC;IAED,sCAAkE;IAKtE;;;;;;;;;;;;;;;;;;;;OAoBG;IACH,uBA0CC;IAED;;;;;;;;;;;;;;OAcG;IACH,kCALW,YAAY,GAAC,YAAY,eACzB,MAAM,EAAE,EAAE,eACV,MAAM,GACJ,OAAO,CAAC,MAAM,CAAC,CAoB3B;IAGD;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY;;QACvB,OAAO,CAAC;QAAE,cAAc,EAAE,MAAM,CAAA;KAAE,CAAC,CAkB/C;CACJ;iCA9JuD,wCAAwC;uBACzE,uBAAuB"}
|
|
@@ -10,33 +10,33 @@ export class DetrImageProcessor extends ImageProcessor {
|
|
|
10
10
|
* @param {import('../../utils/image.js').RawImage[]} images The image(s) to extract features from.
|
|
11
11
|
* @returns {Promise<DetrFeatureExtractorResult>} An object containing the concatenated pixel values of the preprocessed images.
|
|
12
12
|
*/
|
|
13
|
-
_call(images: import(
|
|
13
|
+
_call(images: import("../../utils/image.js").RawImage[]): Promise<DetrFeatureExtractorResult>;
|
|
14
14
|
post_process_object_detection(outputs: {
|
|
15
15
|
logits: import("../../utils/tensor.js").Tensor;
|
|
16
16
|
pred_boxes: import("../../utils/tensor.js").Tensor;
|
|
17
17
|
}, threshold?: number, target_sizes?: [number, number][], is_zero_shot?: boolean): any[];
|
|
18
|
-
post_process_panoptic_segmentation(outputs: any, threshold?: number, mask_threshold?: number, overlap_mask_area_threshold?: number, label_ids_to_fuse?: Set<number>, target_sizes?: [number, number][]): {
|
|
18
|
+
post_process_panoptic_segmentation(outputs: any, threshold?: number, mask_threshold?: number, overlap_mask_area_threshold?: number, label_ids_to_fuse?: Set<number>, target_sizes?: [number, number][]): Array<{
|
|
19
19
|
segmentation: import("../../utils/tensor.js").Tensor;
|
|
20
|
-
segments_info: {
|
|
20
|
+
segments_info: Array<{
|
|
21
21
|
id: number;
|
|
22
22
|
label_id: number;
|
|
23
23
|
score: number;
|
|
24
|
-
}
|
|
25
|
-
}
|
|
26
|
-
post_process_instance_segmentation(outputs: any, threshold?: number, target_sizes?: [number, number][]): {
|
|
24
|
+
}>;
|
|
25
|
+
}>;
|
|
26
|
+
post_process_instance_segmentation(outputs: any, threshold?: number, target_sizes?: [number, number][]): Array<{
|
|
27
27
|
segmentation: import("../../utils/tensor.js").Tensor;
|
|
28
|
-
segments_info: {
|
|
28
|
+
segments_info: Array<{
|
|
29
29
|
id: number;
|
|
30
30
|
label_id: number;
|
|
31
31
|
score: number;
|
|
32
|
-
}
|
|
33
|
-
}
|
|
32
|
+
}>;
|
|
33
|
+
}>;
|
|
34
34
|
}
|
|
35
35
|
export class DetrFeatureExtractor extends DetrImageProcessor {
|
|
36
36
|
}
|
|
37
37
|
export type DetrFeatureExtractorResultProps = {
|
|
38
|
-
pixel_mask: import(
|
|
38
|
+
pixel_mask: import("../../utils/tensor.js").Tensor;
|
|
39
39
|
};
|
|
40
|
-
export type DetrFeatureExtractorResult = import(
|
|
40
|
+
export type DetrFeatureExtractorResult = import("../../base/image_processors_utils.js").ImageProcessorResult & DetrFeatureExtractorResultProps;
|
|
41
41
|
import { ImageProcessor } from "../../base/image_processors_utils.js";
|
|
42
42
|
//# sourceMappingURL=image_processing_detr.d.ts.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"image_processing_detr.d.ts","sourceRoot":"","sources":["../../../src/models/detr/image_processing_detr.js"],"names":[],"mappings":"AAUA;;;;GAIG;AAEH;IACI;;;;;OAKG;IACH,cAHW,OAAO,sBAAsB,EAAE,QAAQ,EAAE,GACvC,
|
|
1
|
+
{"version":3,"file":"image_processing_detr.d.ts","sourceRoot":"","sources":["../../../src/models/detr/image_processing_detr.js"],"names":[],"mappings":"AAUA;;;;GAIG;AAEH;IACI;;;;;OAKG;IACH,cAHW,OAAO,sBAAsB,EAAE,QAAQ,EAAE,GACvC,OAAO,CAAC,0BAA0B,CAAC,CAY/C;;;;;;;;;;;;;;;;;;;;;CAgBJ;AAED;CAAgE;;gBAvClD,OAAO,uBAAuB,EAAE,MAAM;;yCACvC,OAAO,sCAAsC,EAAE,oBAAoB,GAAG,+BAA+B;+BAR3G,sCAAsC"}
|
|
@@ -1,5 +1,5 @@
|
|
|
1
1
|
export class DonutImageProcessor extends ImageProcessor {
|
|
2
|
-
pad_image(pixelData: any, imgDims: any, padSize: any, options?: {}): [Float32Array
|
|
2
|
+
pad_image(pixelData: any, imgDims: any, padSize: any, options?: {}): [Float32Array<ArrayBufferLike>, number[]];
|
|
3
3
|
}
|
|
4
4
|
export class DonutFeatureExtractor extends DonutImageProcessor {
|
|
5
5
|
}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"image_processing_donut.d.ts","sourceRoot":"","sources":["../../../src/models/donut/image_processing_donut.js"],"names":[],"mappings":"AAIA;IACI
|
|
1
|
+
{"version":3,"file":"image_processing_donut.d.ts","sourceRoot":"","sources":["../../../src/models/donut/image_processing_donut.js"],"names":[],"mappings":"AAIA;IACI,+GAuBC;CACJ;AACD;CAAkE;+BA5B3D,sCAAsC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"processing_florence2.d.ts","sourceRoot":"","sources":["../../../src/models/florence2/processing_florence2.js"],"names":[],"mappings":"AAIA;IACI,6CAAsC;IACtC,wDAAiD;IAEjD,0CAuBC;IAdG,kCAAkC;IAClC,mCADW,
|
|
1
|
+
{"version":3,"file":"processing_florence2.d.ts","sourceRoot":"","sources":["../../../src/models/florence2/processing_florence2.js"],"names":[],"mappings":"AAIA;IACI,6CAAsC;IACtC,wDAAiD;IAEjD,0CAuBC;IAdG,kCAAkC;IAClC,mCADW,GAAG,CAAC,MAAM,EAAE,MAAM,CAAC,CAC2E;IAEzG,kCAAkC;IAClC,6BADW,GAAG,CAAC,MAAM,EAAE,MAAM,CAAC,CAC+D;IAE7F,kCAAkC;IAClC,yBADW,GAAG,CAAC,MAAM,EAAE,MAAM,CAAC,CACuD;IAErF;;;MAGC;IACD,qBAAwB;IAG5B;;;;OAIG;IACH,wBAHW,MAAM,GAAC,MAAM,EAAE,GACb,MAAM,EAAE,CA6BpB;IAED;;;;;OAKG;IACH,8BAJW,MAAM,QACN,MAAM,cACN,CAAC,MAAM,EAAE,MAAM,CAAC;;;;;MAsC1B;IAID,0DAaC;CACJ;0BA/HyB,gCAAgC;8BAE5B,qBAAqB;mCADhB,kCAAkC"}
|
|
@@ -0,0 +1,40 @@
|
|
|
1
|
+
export class Idefics3ImageProcessor extends ImageProcessor {
|
|
2
|
+
constructor(config: any);
|
|
3
|
+
do_image_splitting: any;
|
|
4
|
+
max_image_size: any;
|
|
5
|
+
/**
|
|
6
|
+
* @typedef {import('../../utils/image.js').RawImage} RawImage
|
|
7
|
+
* @typedef {import('../../utils/tensor.js').Tensor} Tensor
|
|
8
|
+
*/
|
|
9
|
+
/**
|
|
10
|
+
* Calculate size to resize images to, to be multiples of `vision_encoder_max_size` while preserving the aspect ratio.
|
|
11
|
+
* @param {Tensor} pixel_values Tensor of the image to resize.
|
|
12
|
+
* @param {number} vision_encoder_max_size Maximum size of the output image. If the image is larger than this size,
|
|
13
|
+
* it will be split into patches of this size, and the original image will be concatenated with the patches, resized to max_size.
|
|
14
|
+
*/
|
|
15
|
+
get_resize_for_vision_encoder(pixel_values: import("../../utils/tensor.js").Tensor, vision_encoder_max_size: number): {
|
|
16
|
+
height: number;
|
|
17
|
+
width: number;
|
|
18
|
+
};
|
|
19
|
+
/** @param {RawImage|RawImage[]|RawImage[][]} images */
|
|
20
|
+
_call(images: import("../../utils/image.js").RawImage | import("../../utils/image.js").RawImage[] | import("../../utils/image.js").RawImage[][], { do_image_splitting, return_row_col_info, }?: {
|
|
21
|
+
do_image_splitting?: any;
|
|
22
|
+
return_row_col_info?: boolean;
|
|
23
|
+
}): Promise<{
|
|
24
|
+
rows?: any[][];
|
|
25
|
+
cols?: any[][];
|
|
26
|
+
pixel_values: import("../../utils/tensor.js").Tensor;
|
|
27
|
+
pixel_attention_mask: import("../../utils/tensor.js").Tensor;
|
|
28
|
+
original_sizes: import("../../base/image_processors_utils.js").HeightWidth[];
|
|
29
|
+
reshaped_input_sizes: import("../../base/image_processors_utils.js").HeightWidth[];
|
|
30
|
+
}>;
|
|
31
|
+
split_image(pixel_values: any, { longest_edge }: {
|
|
32
|
+
longest_edge: any;
|
|
33
|
+
}): Promise<{
|
|
34
|
+
frames: any[];
|
|
35
|
+
num_splits_h: number;
|
|
36
|
+
num_splits_w: number;
|
|
37
|
+
}>;
|
|
38
|
+
}
|
|
39
|
+
import { ImageProcessor } from "../../base/image_processors_utils.js";
|
|
40
|
+
//# sourceMappingURL=image_processing_idefics3.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"image_processing_idefics3.d.ts","sourceRoot":"","sources":["../../../src/models/idefics3/image_processing_idefics3.js"],"names":[],"mappings":"AAOA;IACI,yBAKC;IAFG,wBAA2D;IAC3D,oBAA2C;IAG/C;;;OAGG;IAEH;;;;;OAKG;IACH,6GAHW,MAAM;;;MAiBhB;IAED,uDAAuD;IACvD,cADY,0CAAS,yCAAU,GAAC,yCAAU,EAAE;;;;;;;;;;OA4H3C;IAED;;;;;;OAiDC;CACJ;+BAtNM,sCAAsC"}
|
|
@@ -0,0 +1,19 @@
|
|
|
1
|
+
export class Idefics3Processor extends Processor {
|
|
2
|
+
static image_processor_class: typeof AutoImageProcessor;
|
|
3
|
+
static tokenizer_class: typeof AutoTokenizer;
|
|
4
|
+
fake_image_token: string;
|
|
5
|
+
image_token: string;
|
|
6
|
+
global_img_token: string;
|
|
7
|
+
/**
|
|
8
|
+
*
|
|
9
|
+
* @param {string|string[]} text
|
|
10
|
+
* @param {RawImage|RawImage[]|RawImage[][]} images
|
|
11
|
+
* @returns {Promise<any>}
|
|
12
|
+
*/
|
|
13
|
+
_call(text: string | string[], images?: RawImage | RawImage[] | RawImage[][], options?: {}): Promise<any>;
|
|
14
|
+
}
|
|
15
|
+
import { Processor } from "../../base/processing_utils.js";
|
|
16
|
+
import { RawImage } from "../../utils/image.js";
|
|
17
|
+
import { AutoImageProcessor } from "../auto/image_processing_auto.js";
|
|
18
|
+
import { AutoTokenizer } from "../../tokenizers.js";
|
|
19
|
+
//# sourceMappingURL=processing_idefics3.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"processing_idefics3.d.ts","sourceRoot":"","sources":["../../../src/models/idefics3/processing_idefics3.js"],"names":[],"mappings":"AA6DA;IACI,wDAAiD;IACjD,6CAAsC;IAGtC,yBAA+C;IAC/C,oBAAwB;IACxB,yBAAkC;IAElC;;;;;OAKG;IACH,YAJW,MAAM,GAAC,MAAM,EAAE,WACf,QAAQ,GAAC,QAAQ,EAAE,GAAC,QAAQ,EAAE,EAAE,iBAC9B,OAAO,CAAC,GAAG,CAAC,CA4DxB;CACJ;0BAtIyB,gCAAgC;yBAGjC,sBAAsB;mCAFZ,kCAAkC;8BACvC,qBAAqB"}
|
|
@@ -9,6 +9,7 @@ export * from "./donut/image_processing_donut.js";
|
|
|
9
9
|
export * from "./dpt/image_processing_dpt.js";
|
|
10
10
|
export * from "./efficientnet/image_processing_efficientnet.js";
|
|
11
11
|
export * from "./glpn/image_processing_glpn.js";
|
|
12
|
+
export * from "./idefics3/image_processing_idefics3.js";
|
|
12
13
|
export * from "./janus/image_processing_janus.js";
|
|
13
14
|
export * from "./jina_clip/image_processing_jina_clip.js";
|
|
14
15
|
export * from "./llava_onevision/image_processing_llava_onevision.js";
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
export class VLMImageProcessor extends ImageProcessor {
|
|
2
2
|
constructor(config: any);
|
|
3
3
|
constant_values: any;
|
|
4
|
-
pad_image(pixelData: any, imgDims: any, padSize: any, options: any): [Float32Array
|
|
4
|
+
pad_image(pixelData: any, imgDims: any, padSize: any, options: any): [Float32Array<ArrayBufferLike>, number[]];
|
|
5
5
|
}
|
|
6
6
|
import { ImageProcessor } from "../../base/image_processors_utils.js";
|
|
7
7
|
//# sourceMappingURL=image_processing_janus.d.ts.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"image_processing_janus.d.ts","sourceRoot":"","sources":["../../../src/models/janus/image_processing_janus.js"],"names":[],"mappings":"AAKA;IACI,yBAUC;IADG,qBAAqF;IAGzF
|
|
1
|
+
{"version":3,"file":"image_processing_janus.d.ts","sourceRoot":"","sources":["../../../src/models/janus/image_processing_janus.js"],"names":[],"mappings":"AAKA;IACI,yBAUC;IADG,qBAAqF;IAGzF,+GAMC;CACJ;+BAtBM,sCAAsC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"processing_janus.d.ts","sourceRoot":"","sources":["../../../src/models/janus/processing_janus.js"],"names":[],"mappings":"AAQA;IACI,wDAAiD;IACjD,6CAAsC;IAGtC,0CAOC;IAJG,eAAsC;IACtC,qBAAkD;IAClD,mBAA8C;IAC9C,sBAAoD;IAGxD;;;;OAIG;IAEH;;;;;;OAMG;IAEH;;;;;;OAMG;IACH;;;;iBAnBc,CAAC,QAAQ,GAAG,MAAM,GAAG,GAAG,CAAC,EAAE
|
|
1
|
+
{"version":3,"file":"processing_janus.d.ts","sourceRoot":"","sources":["../../../src/models/janus/processing_janus.js"],"names":[],"mappings":"AAQA;IACI,wDAAiD;IACjD,6CAAsC;IAGtC,0CAOC;IAJG,eAAsC;IACtC,qBAAkD;IAClD,mBAA8C;IAC9C,sBAAoD;IAGxD;;;;OAIG;IAEH;;;;;;OAMG;IAEH;;;;;;OAMG;IACH;;;;iBAnBc,CAAC,QAAQ,GAAG,MAAM,GAAG,GAAG,CAAC,EAAE;uCAetC;QAAsC,MAAM,GAApC,QAAQ,GAAC,QAAQ,EAAE;QACF,aAAa,GAA9B,MAAM;KACd,GAAU,OAAO,CAAC;;;;mBAXP,MAAM;;;;wBACN,MAAM;;;;yBACN,MAAM;;;;yBACN,MAAM;SAQ0B;;;;mBAXhC,MAAM;;;;wBACN,MAAM;;;;yBACN,MAAM;;;;yBACN,MAAM;QAQmD,OAAO,sCAAsC,EAAE,oBAAoB,CAAA,CAAC,CAgF1I;CACJ;0BAzHyB,gCAAgC;yBAKjC,sBAAsB;uBADxB,uBAAuB;mCAHX,kCAAkC;8BACvC,qBAAqB"}
|
|
@@ -1,20 +1,20 @@
|
|
|
1
1
|
export class MaskFormerImageProcessor extends ImageProcessor {
|
|
2
|
-
post_process_panoptic_segmentation(outputs: any, threshold?: number, mask_threshold?: number, overlap_mask_area_threshold?: number, label_ids_to_fuse?: Set<number>, target_sizes?: [number, number][]): {
|
|
2
|
+
post_process_panoptic_segmentation(outputs: any, threshold?: number, mask_threshold?: number, overlap_mask_area_threshold?: number, label_ids_to_fuse?: Set<number>, target_sizes?: [number, number][]): Array<{
|
|
3
3
|
segmentation: import("../../transformers.js").Tensor;
|
|
4
|
-
segments_info: {
|
|
4
|
+
segments_info: Array<{
|
|
5
5
|
id: number;
|
|
6
6
|
label_id: number;
|
|
7
7
|
score: number;
|
|
8
|
-
}
|
|
9
|
-
}
|
|
10
|
-
post_process_instance_segmentation(outputs: any, threshold?: number, target_sizes?: [number, number][]): {
|
|
8
|
+
}>;
|
|
9
|
+
}>;
|
|
10
|
+
post_process_instance_segmentation(outputs: any, threshold?: number, target_sizes?: [number, number][]): Array<{
|
|
11
11
|
segmentation: import("../../transformers.js").Tensor;
|
|
12
|
-
segments_info: {
|
|
12
|
+
segments_info: Array<{
|
|
13
13
|
id: number;
|
|
14
14
|
label_id: number;
|
|
15
15
|
score: number;
|
|
16
|
-
}
|
|
17
|
-
}
|
|
16
|
+
}>;
|
|
17
|
+
}>;
|
|
18
18
|
}
|
|
19
19
|
export class MaskFormerFeatureExtractor extends MaskFormerImageProcessor {
|
|
20
20
|
}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"image_processing_maskformer.d.ts","sourceRoot":"","sources":["../../../src/models/maskformer/image_processing_maskformer.js"],"names":[],"mappings":"AAMA;
|
|
1
|
+
{"version":3,"file":"image_processing_maskformer.d.ts","sourceRoot":"","sources":["../../../src/models/maskformer/image_processing_maskformer.js"],"names":[],"mappings":"AAMA;;;;;;;;;;;;;;;;;CAUC;AACD;CAA4E;+BAbrE,sCAAsC"}
|
|
@@ -19,7 +19,7 @@ export class MgpstrProcessor extends Processor {
|
|
|
19
19
|
* @param {string} format Type of model prediction. Must be one of ['char', 'bpe', 'wp'].
|
|
20
20
|
* @returns {[string[], number[]]} The decoded sentences and their confidence scores.
|
|
21
21
|
*/
|
|
22
|
-
_decode_helper(pred_logits: import(
|
|
22
|
+
_decode_helper(pred_logits: import("../../utils/tensor.js").Tensor, format: string): [string[], number[]];
|
|
23
23
|
/**
|
|
24
24
|
* Convert a list of lists of char token ids into a list of strings by calling char tokenizer.
|
|
25
25
|
* @param {number[][]} sequences List of tokenized input ids.
|
|
@@ -49,7 +49,7 @@ export class MgpstrProcessor extends Processor {
|
|
|
49
49
|
* - bpe_preds: The list of BPE decoded sentences.
|
|
50
50
|
* - wp_preds: The list of wp decoded sentences.
|
|
51
51
|
*/
|
|
52
|
-
batch_decode([char_logits, bpe_logits, wp_logits]: import(
|
|
52
|
+
batch_decode([char_logits, bpe_logits, wp_logits]: import("../../utils/tensor.js").Tensor[]): {
|
|
53
53
|
generated_text: string[];
|
|
54
54
|
scores: number[];
|
|
55
55
|
char_preds: string[];
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"processing_mgp_str.d.ts","sourceRoot":"","sources":["../../../src/models/mgp_str/processing_mgp_str.js"],"names":[],"mappings":"AAUA;IACI,6CAAsC;IACtC,wDAAiD;IAEjD;;OAEG;IACH,
|
|
1
|
+
{"version":3,"file":"processing_mgp_str.d.ts","sourceRoot":"","sources":["../../../src/models/mgp_str/processing_mgp_str.js"],"names":[],"mappings":"AAUA;IACI,6CAAsC;IACtC,wDAAiD;IAEjD;;OAEG;IACH,sBAFa,OAAO,qBAAqB,EAAE,eAAe,CAIzD;IAED;;OAEG;IACH,qBAFa,OAAO,qBAAqB,EAAE,aAAa,CAIvD;IAED;;OAEG;IACH,oBAFa,OAAO,qBAAqB,EAAE,aAAa,CAIvD;IAED;;;;;OAKG;IACH,4BAJW,OAAO,uBAAuB,EAAE,MAAM,UACtC,MAAM,GACJ,CAAC,MAAM,EAAE,EAAE,MAAM,EAAE,CAAC,CA0ChC;IAED;;;;OAIG;IACH,uBAHW,MAAM,EAAE,EAAE,GACR,MAAM,EAAE,CAIpB;IAED;;;;OAIG;IACH,sBAHW,MAAM,EAAE,EAAE,GACR,MAAM,EAAE,CAIpB;IAED;;;;OAIG;IACH,qBAHW,MAAM,EAAE,EAAE,GACR,MAAM,EAAE,CAIpB;IAED;;;;;;;;;;OAUG;IACH,mDATW,OAAO,uBAAuB,EAAE,MAAM,EAAE,GACtC;QAAC,cAAc,EAAE,MAAM,EAAE,CAAC;QAAC,MAAM,EAAE,MAAM,EAAE,CAAC;QAAC,UAAU,EAAE,MAAM,EAAE,CAAC;QAAC,SAAS,EAAE,MAAM,EAAE,CAAC;QAAC,QAAQ,EAAE,MAAM,EAAE,CAAA;KAAC,CA4BvH;IAmBD,6CAQC;CACJ;0BAzKyB,gCAAgC;8BAE5B,qBAAqB;mCADhB,kCAAkC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"image_processing_owlvit.d.ts","sourceRoot":"","sources":["../../../src/models/owlvit/image_processing_owlvit.js"],"names":[],"mappings":"AAKA;
|
|
1
|
+
{"version":3,"file":"image_processing_owlvit.d.ts","sourceRoot":"","sources":["../../../src/models/owlvit/image_processing_owlvit.js"],"names":[],"mappings":"AAKA;;;;;CAKC;AACD;CAAoE;+BAR7D,sCAAsC"}
|
|
@@ -0,0 +1,12 @@
|
|
|
1
|
+
export class PaliGemmaProcessor extends Processor {
|
|
2
|
+
static tokenizer_class: typeof AutoTokenizer;
|
|
3
|
+
static image_processor_class: typeof AutoImageProcessor;
|
|
4
|
+
/**
|
|
5
|
+
* @typedef {import('../../utils/image.js').RawImage} RawImage
|
|
6
|
+
*/
|
|
7
|
+
_call(images: import("../../utils/image.js").RawImage | import("../../utils/image.js").RawImage[], text?: any, kwargs?: {}): Promise<any>;
|
|
8
|
+
}
|
|
9
|
+
import { Processor } from "../../base/processing_utils.js";
|
|
10
|
+
import { AutoTokenizer } from "../../tokenizers.js";
|
|
11
|
+
import { AutoImageProcessor } from "../auto/image_processing_auto.js";
|
|
12
|
+
//# sourceMappingURL=processing_paligemma.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"processing_paligemma.d.ts","sourceRoot":"","sources":["../../../src/models/paligemma/processing_paligemma.js"],"names":[],"mappings":"AAgBA;IACI,6CAAsC;IACtC,wDAAiD;IAGjD;;OAEG;IAGH,cAAuB,0CAAS,yCAAU,yCAsDzC;CACJ;0BAjFyB,gCAAgC;8BAE5B,qBAAqB;mCADhB,kCAAkC"}
|
|
@@ -1,8 +1,10 @@
|
|
|
1
1
|
export * from "./florence2/processing_florence2.js";
|
|
2
2
|
export * from "./mgp_str/processing_mgp_str.js";
|
|
3
|
+
export * from "./idefics3/processing_idefics3.js";
|
|
3
4
|
export * from "./janus/processing_janus.js";
|
|
4
5
|
export * from "./jina_clip/processing_jina_clip.js";
|
|
5
6
|
export * from "./owlvit/processing_owlvit.js";
|
|
7
|
+
export * from "./paligemma/processing_paligemma.js";
|
|
6
8
|
export * from "./pyannote/processing_pyannote.js";
|
|
7
9
|
export * from "./qwen2_vl/processing_qwen2_vl.js";
|
|
8
10
|
export * from "./sam/processing_sam.js";
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"feature_extraction_pyannote.d.ts","sourceRoot":"","sources":["../../../src/models/pyannote/feature_extraction_pyannote.js"],"names":[],"mappings":"AAIA;IACI;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY,GACvB,
|
|
1
|
+
{"version":3,"file":"feature_extraction_pyannote.d.ts","sourceRoot":"","sources":["../../../src/models/pyannote/feature_extraction_pyannote.js"],"names":[],"mappings":"AAIA;IACI;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY,GACvB,OAAO,CAAC;QAAE,YAAY,EAAE,MAAM,CAAC;KAAE,CAAC,CAiB9C;CAEJ;iCA3BuD,wCAAwC;uBACzE,uBAAuB"}
|
|
@@ -18,7 +18,7 @@ export class PyAnnoteProcessor extends Processor {
|
|
|
18
18
|
* @param {number} num_samples Number of samples in the input audio.
|
|
19
19
|
* @returns {Array<Array<{ id: number, start: number, end: number, confidence: number }>>} The post-processed speaker diarization results.
|
|
20
20
|
*/
|
|
21
|
-
post_process_speaker_diarization(logits: import(
|
|
21
|
+
post_process_speaker_diarization(logits: import("../../utils/tensor.js").Tensor, num_samples: number): Array<Array<{
|
|
22
22
|
id: number;
|
|
23
23
|
start: number;
|
|
24
24
|
end: number;
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"processing_pyannote.d.ts","sourceRoot":"","sources":["../../../src/models/pyannote/processing_pyannote.js"],"names":[],"mappings":"AAIA;IACI,4DAAqD;IAErD;;;;OAIG;IACH,aAHW,GAAG,GACD,
|
|
1
|
+
{"version":3,"file":"processing_pyannote.d.ts","sourceRoot":"","sources":["../../../src/models/pyannote/processing_pyannote.js"],"names":[],"mappings":"AAIA;IACI,4DAAqD;IAErD;;;;OAIG;IACH,aAHW,GAAG,GACD,OAAO,CAAC,GAAG,CAAC,CAIxB;IAED;;;;OAIG;IACH,2BAHW,MAAM,GACJ,MAAM,CAIlB;IAED;;;;;OAKG;IACH,yCAJW,OAAO,uBAAuB,EAAE,MAAM,eACtC,MAAM,GACJ,KAAK,CAAC,KAAK,CAAC;QAAE,EAAE,EAAE,MAAM,CAAC;QAAC,KAAK,EAAE,MAAM,CAAC;QAAC,GAAG,EAAE,MAAM,CAAC;QAAC,UAAU,EAAE,MAAM,CAAA;KAAE,CAAC,CAAC,CAwCxF;CACJ;0BAtEyB,gCAAgC;qCACrB,oCAAoC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"processing_qwen2_vl.d.ts","sourceRoot":"","sources":["../../../src/models/qwen2_vl/processing_qwen2_vl.js"],"names":[],"mappings":"AAKA;IACI,wDAAiD;IACjD,6CAAsC;IAEtC;;;;;;OAMG;IACH,YALW,MAAM,GAAC,MAAM,EAAE,WACf,QAAQ,GAAC,QAAQ,EAAE,WACf,GAAG,
|
|
1
|
+
{"version":3,"file":"processing_qwen2_vl.d.ts","sourceRoot":"","sources":["../../../src/models/qwen2_vl/processing_qwen2_vl.js"],"names":[],"mappings":"AAKA;IACI,wDAAiD;IACjD,6CAAsC;IAEtC;;;;;;OAMG;IACH,YALW,MAAM,GAAC,MAAM,EAAE,WACf,QAAQ,GAAC,QAAQ,EAAE,WACf,GAAG,EAAA,GACL,OAAO,CAAC,GAAG,CAAC,CAoCxB;CACJ;0BAnDyB,gCAAgC;yBAGjC,sBAAsB;mCAFZ,kCAAkC;8BACvC,qBAAqB"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"image_processing_rt_detr.d.ts","sourceRoot":"","sources":["../../../src/models/rt_detr/image_processing_rt_detr.js"],"names":[],"mappings":"AAMA
|
|
1
|
+
{"version":3,"file":"image_processing_rt_detr.d.ts","sourceRoot":"","sources":["../../../src/models/rt_detr/image_processing_rt_detr.js"],"names":[],"mappings":"AAMA;;;;;CAKC;+BARM,sCAAsC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"image_processing_sam.d.ts","sourceRoot":"","sources":["../../../src/models/sam/image_processing_sam.js"],"names":[],"mappings":"AAWA;;;;;;;;GAQG;AAEH;IAEI;;;;;;OAMG;IACH,mCALW,GAAG,kBACH,OAAO,sCAAsC,EAAE,WAAW,EAAE,wBAC5D,OAAO,sCAAsC,EAAE,WAAW,EAAE,8BAC1D,MAAM,CA4ClB;IAED;;;;;OAKG;IACH,+BAJW,GAAG,gBACH,MAAM,GACJ,MAAM,CAoBlB;IACD;;;;;;;;;;;;;;;;;;;OAmBG;IACH,cAnBW,GAAG,EAAE;
|
|
1
|
+
{"version":3,"file":"image_processing_sam.d.ts","sourceRoot":"","sources":["../../../src/models/sam/image_processing_sam.js"],"names":[],"mappings":"AAWA;;;;;;;;GAQG;AAEH;IAEI;;;;;;OAMG;IACH,mCALW,GAAG,kBACH,OAAO,sCAAsC,EAAE,WAAW,EAAE,wBAC5D,OAAO,sCAAsC,EAAE,WAAW,EAAE,8BAC1D,MAAM,CA4ClB;IAED;;;;;OAKG;IACH,+BAJW,GAAG,gBACH,MAAM,GACJ,MAAM,CAoBlB;IACD;;;;;;;;;;;;;;;;;;;OAmBG;IACH,cAnBW,GAAG,EAAE,gDAEb;QAAsB,YAAY,GAA1B,GAAG;QAGW,YAAY,GAA1B,GAAG;QAGoB,WAAW,GAAlC,MAAM,EAAE,EAAE,EAAE;KASpB,GAAU,OAAO,CAAC,uBAAuB,CAAC,CA+B5C;IAED;;;;;;;;;;;;OAYG;IACH,0BAXW,MAAM,kBACN,CAAC,MAAM,EAAE,MAAM,CAAC,EAAE,wBAClB,CAAC,MAAM,EAAE,MAAM,CAAC,EAAE,4CAE1B;QAAyB,cAAc,GAA/B,MAAM;QACY,QAAQ,GAA1B,OAAO;QACU,QAAQ,GACjC;YAAkC,MAAM,GAAhC,MAAM;YACoB,KAAK,GAA/B,MAAM;SACd;KAAA,GAAU,OAAO,CAAC,MAAM,EAAE,CAAC,CAsD7B;IAED;;;;;;;;;;;;;OAaG;IACH,2BAZW,OAAO,sBAAsB,EAAE,QAAQ,eACvC,MAAM,uFAEd;QAAyB,aAAa,GAA9B,MAAM;QAEW,aAAa,GAA9B,MAAM;QAEW,eAAe,GAAhC,MAAM;QACW,8BAA8B,GAA/C,MAAM;KAEd,OAUF;CACJ;;kBAnOa,MAAM;oBACN,OAAO,sCAAsC,EAAE,WAAW,EAAE;0BAC5D,OAAO,sCAAsC,EAAE,WAAW,EAAE;mBAC5D,MAAM;mBACN,MAAM;kBACN,MAAM;;+BAhBb,sCAAsC;uBAMtC,uBAAuB"}
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
export class SeamlessM4TFeatureExtractor extends FeatureExtractor {
|
|
2
2
|
constructor(config: any);
|
|
3
3
|
mel_filters: number[][];
|
|
4
|
-
window: Float64Array
|
|
4
|
+
window: Float64Array<ArrayBufferLike>;
|
|
5
5
|
/**
|
|
6
6
|
* Computes the log-Mel spectrogram of the provided audio waveform.
|
|
7
7
|
* @param {Float32Array|Float64Array} waveform The audio waveform to process.
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"feature_extraction_seamless_m4t.d.ts","sourceRoot":"","sources":["../../../src/models/seamless_m4t/feature_extraction_seamless_m4t.js"],"names":[],"mappings":"AAIA;IAEI,yBAwBC;IALG,wBAA8B;IAE9B,
|
|
1
|
+
{"version":3,"file":"feature_extraction_seamless_m4t.d.ts","sourceRoot":"","sources":["../../../src/models/seamless_m4t/feature_extraction_seamless_m4t.js"],"names":[],"mappings":"AAIA;IAEI,yBAwBC;IALG,wBAA8B;IAE9B,sCAEE;IAGN;;;;;OAKG;IACH,kCAJW,YAAY,GAAC,YAAY,cACzB,MAAM,GACJ,OAAO,CAAC,MAAM,CAAC,CA6B3B;IAED;;;;;;;;;OASG;IACH,aARW,YAAY,GAAC,YAAY,uFAEjC;QAA0B,OAAO,GAAzB,OAAO;QACU,kBAAkB,GAAnC,MAAM;QACY,yBAAyB,GAA3C,OAAO;QACW,qBAAqB,GAAvC,OAAO;KACf,GAAU,OAAO,CAAC;QAAE,cAAc,EAAE,MAAM,CAAC;QAAC,cAAc,CAAC,EAAE,MAAM,CAAA;KAAE,CAAC,CAuGxE;CACJ;iCAnLuD,wCAAwC;uBACzE,uBAAuB"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"image_processing_segformer.d.ts","sourceRoot":"","sources":["../../../src/models/segformer/image_processing_segformer.js"],"names":[],"mappings":"AAMA;
|
|
1
|
+
{"version":3,"file":"image_processing_segformer.d.ts","sourceRoot":"","sources":["../../../src/models/segformer/image_processing_segformer.js"],"names":[],"mappings":"AAMA;;;;;CAKC;AACD;CAA0E;+BATnE,sCAAsC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"processing_speecht5.d.ts","sourceRoot":"","sources":["../../../src/models/speecht5/processing_speecht5.js"],"names":[],"mappings":"AAIA;IACI,6CAAsC;IACtC,4DAAqD;IAErD;;;;OAIG;IACH,aAHW,GAAG,GACD,
|
|
1
|
+
{"version":3,"file":"processing_speecht5.d.ts","sourceRoot":"","sources":["../../../src/models/speecht5/processing_speecht5.js"],"names":[],"mappings":"AAIA;IACI,6CAAsC;IACtC,4DAAqD;IAErD;;;;OAIG;IACH,aAHW,GAAG,GACD,OAAO,CAAC,GAAG,CAAC,CAIxB;CACJ;0BAhByB,gCAAgC;8BAC5B,qBAAqB;qCACd,oCAAoC"}
|
|
@@ -1,5 +1,5 @@
|
|
|
1
1
|
export class Swin2SRImageProcessor extends ImageProcessor {
|
|
2
|
-
pad_image(pixelData: any, imgDims: any, padSize: any, options?: {}): [Float32Array
|
|
2
|
+
pad_image(pixelData: any, imgDims: any, padSize: any, options?: {}): [Float32Array<ArrayBufferLike>, number[]];
|
|
3
3
|
}
|
|
4
4
|
import { ImageProcessor } from "../../base/image_processors_utils.js";
|
|
5
5
|
//# sourceMappingURL=image_processing_swin2sr.d.ts.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"image_processing_swin2sr.d.ts","sourceRoot":"","sources":["../../../src/models/swin2sr/image_processing_swin2sr.js"],"names":[],"mappings":"AAIA;IACI
|
|
1
|
+
{"version":3,"file":"image_processing_swin2sr.d.ts","sourceRoot":"","sources":["../../../src/models/swin2sr/image_processing_swin2sr.js"],"names":[],"mappings":"AAIA;IACI,+GAiBC;CACJ;+BArBM,sCAAsC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"image_processing_vitmatte.d.ts","sourceRoot":"","sources":["../../../src/models/vitmatte/image_processing_vitmatte.js"],"names":[],"mappings":"AASA;IACI;;;;;;OAMG;IACH,cAJW,OAAO,sBAAsB,EAAE,QAAQ,EAAE,WACzC,OAAO,sBAAsB,EAAE,QAAQ,EAAE,GACvC,
|
|
1
|
+
{"version":3,"file":"image_processing_vitmatte.d.ts","sourceRoot":"","sources":["../../../src/models/vitmatte/image_processing_vitmatte.js"],"names":[],"mappings":"AASA;IACI;;;;;;OAMG;IACH,cAJW,OAAO,sBAAsB,EAAE,QAAQ,EAAE,WACzC,OAAO,sBAAsB,EAAE,QAAQ,EAAE,GACvC,OAAO,CAAC,OAAO,sCAAsC,EAAE,oBAAoB,CAAC,CAiCxF;CACJ;+BA/CM,sCAAsC"}
|
|
@@ -13,7 +13,7 @@ export class VitPoseImageProcessor extends ImageProcessor {
|
|
|
13
13
|
* keypoints: [number, number][]
|
|
14
14
|
* }[][]} List of keypoints predictions for each image.
|
|
15
15
|
*/
|
|
16
|
-
post_process_pose_estimation(outputs: import(
|
|
16
|
+
post_process_pose_estimation(outputs: import("../../utils/tensor.js").Tensor, boxes: [number, number, number, number][][], { threshold, }?: {
|
|
17
17
|
threshold?: any;
|
|
18
18
|
}): {
|
|
19
19
|
bbox: [number, number, number, number];
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"image_processing_vitpose.d.ts","sourceRoot":"","sources":["../../../src/models/vitpose/image_processing_vitpose.js"],"names":[],"mappings":"AAIA;IAEI;;;;;;;;;;;;;OAaG;IACH,sCAVW,OAAO,uBAAuB,EAAE,MAAM,SACtC,CAAC,MAAM,EAAE,MAAM,EAAE,MAAM,EAAE,MAAM,CAAC,EAAE,EAAE;;QAElC;
|
|
1
|
+
{"version":3,"file":"image_processing_vitpose.d.ts","sourceRoot":"","sources":["../../../src/models/vitpose/image_processing_vitpose.js"],"names":[],"mappings":"AAIA;IAEI;;;;;;;;;;;;;OAaG;IACH,sCAVW,OAAO,uBAAuB,EAAE,MAAM,SACtC,CAAC,MAAM,EAAE,MAAM,EAAE,MAAM,EAAE,MAAM,CAAC,EAAE,EAAE;;QAElC;QACR,IAAI,EAAE,CAAC,MAAM,EAAE,MAAM,EAAE,MAAM,EAAE,MAAM,CAAC,CAAC;QACvC,MAAM,EAAE,MAAM,EAAE,CAAC;QACjB,MAAM,EAAE,MAAM,EAAE,CAAC;QACjB,SAAS,EAAE,CAAC,MAAM,EAAE,MAAM,CAAC,EAAE,CAAA;KAC9B,EAAE,EAAE,CAqEP;CACJ;+BAtFM,sCAAsC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"feature_extraction_wav2vec2.d.ts","sourceRoot":"","sources":["../../../src/models/wav2vec2/feature_extraction_wav2vec2.js"],"names":[],"mappings":"AAGA;IAEI;;;OAGG;IACH,uCAHW,YAAY,GACV,YAAY,CAQxB;IAED;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY;
|
|
1
|
+
{"version":3,"file":"feature_extraction_wav2vec2.d.ts","sourceRoot":"","sources":["../../../src/models/wav2vec2/feature_extraction_wav2vec2.js"],"names":[],"mappings":"AAGA;IAEI;;;OAGG;IACH,uCAHW,YAAY,GACV,YAAY,CAQxB;IAED;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY,GACvB,OAAO,CAAC;QAAE,YAAY,EAAE,MAAM,CAAC;QAAC,cAAc,EAAE,MAAM,CAAA;KAAE,CAAC,CAsBrE;CACJ;iCA3CuD,wCAAwC;uBACzE,uBAAuB"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"processing_wav2vec2.d.ts","sourceRoot":"","sources":["../../../src/models/wav2vec2/processing_wav2vec2.js"],"names":[],"mappings":"AAGA;IACI,4DAAqD;IAErD;;;;OAIG;IACH,aAHW,GAAG,GACD,
|
|
1
|
+
{"version":3,"file":"processing_wav2vec2.d.ts","sourceRoot":"","sources":["../../../src/models/wav2vec2/processing_wav2vec2.js"],"names":[],"mappings":"AAGA;IACI,4DAAqD;IAErD;;;;OAIG;IACH,aAHW,GAAG,GACD,OAAO,CAAC,GAAG,CAAC,CAIxB;CACJ;0BAdyB,gCAAgC;qCACrB,oCAAoC"}
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
export class WeSpeakerFeatureExtractor extends FeatureExtractor {
|
|
2
2
|
constructor(config: any);
|
|
3
3
|
mel_filters: number[][];
|
|
4
|
-
window: Float64Array
|
|
4
|
+
window: Float64Array<ArrayBufferLike>;
|
|
5
5
|
min_num_frames: any;
|
|
6
6
|
/**
|
|
7
7
|
* Computes the log-Mel spectrogram of the provided audio waveform.
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"feature_extraction_wespeaker.d.ts","sourceRoot":"","sources":["../../../src/models/wespeaker/feature_extraction_wespeaker.js"],"names":[],"mappings":"AAKA;IAEI,yBAyBC;IANG,wBAA8B;IAE9B,
|
|
1
|
+
{"version":3,"file":"feature_extraction_wespeaker.d.ts","sourceRoot":"","sources":["../../../src/models/wespeaker/feature_extraction_wespeaker.js"],"names":[],"mappings":"AAKA;IAEI,yBAyBC;IANG,wBAA8B;IAE9B,sCAEE;IACF,oBAAgD;IAGpD;;;;OAIG;IACH,kCAHW,YAAY,GAAC,YAAY,GACvB,OAAO,CAAC,MAAM,CAAC,CA2B3B;IAGD;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY,GACvB,OAAO,CAAC;QAAE,cAAc,EAAE,MAAM,CAAA;KAAE,CAAC,CA4B/C;CACJ;iCAnGuD,wCAAwC;uBACzE,uBAAuB"}
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
export class WhisperFeatureExtractor extends FeatureExtractor {
|
|
2
2
|
constructor(config: any);
|
|
3
|
-
window: Float64Array
|
|
3
|
+
window: Float64Array<ArrayBufferLike>;
|
|
4
4
|
/**
|
|
5
5
|
* Computes the log-Mel spectrogram of the provided audio waveform.
|
|
6
6
|
* @param {Float32Array|Float64Array} waveform The audio waveform to process.
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"feature_extraction_whisper.d.ts","sourceRoot":"","sources":["../../../src/models/whisper/feature_extraction_whisper.js"],"names":[],"mappings":"AAKA;IAEI,yBAeC;IADG,
|
|
1
|
+
{"version":3,"file":"feature_extraction_whisper.d.ts","sourceRoot":"","sources":["../../../src/models/whisper/feature_extraction_whisper.js"],"names":[],"mappings":"AAKA;IAEI,yBAeC;IADG,sCAAwD;IAG5D;;;;OAIG;IACH,kCAHW,YAAY,GAAC,YAAY,GACvB,OAAO,CAAC,MAAM,CAAC,CA0B3B;IAED;;;;OAIG;IACH,aAHW,YAAY,GAAC,YAAY,GACvB,OAAO,CAAC;QAAE,cAAc,EAAE,MAAM,CAAA;KAAE,CAAC,CAwB/C;CACJ;iCAnFuD,wCAAwC;uBACzE,uBAAuB"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"generation_whisper.d.ts","sourceRoot":"","sources":["../../../src/models/whisper/generation_whisper.js"],"names":[],"mappings":"AAEA;IAEI;;;OAGG;IACH,mBAFU,OAAO,CAEQ;IAEzB;;;;;OAKG;IACH,yBAFU,OAAO,CAEc;IAE/B;;;OAGG;IACH,YAFU,MAAM,CAEE;IAElB;;;;OAIG;IACH,iBAFU,CAAC,MAAM,EAAE,MAAM,CAAC,EAAE,CAEL;IAEvB;;;OAGG;IACH,MAFU,MAAM,CAEJ;IAEZ;;;;OAIG;IACH,UAFU,MAAM,CAEA;IAEhB;;;OAGG;IACH,wBAFU,MAAM,CAEc;IAE9B;;;;;;OAMG;IACH,YAFU,MAAM,EAAE,CAEA;IAElB;;;OAGG;IACH,iBAFU,OAAO,CAEM;IAEvB;;;;OAIG;IACH,YAFU,
|
|
1
|
+
{"version":3,"file":"generation_whisper.d.ts","sourceRoot":"","sources":["../../../src/models/whisper/generation_whisper.js"],"names":[],"mappings":"AAEA;IAEI;;;OAGG;IACH,mBAFU,OAAO,CAEQ;IAEzB;;;;;OAKG;IACH,yBAFU,OAAO,CAEc;IAE/B;;;OAGG;IACH,YAFU,MAAM,CAEE;IAElB;;;;OAIG;IACH,iBAFU,CAAC,MAAM,EAAE,MAAM,CAAC,EAAE,CAEL;IAEvB;;;OAGG;IACH,MAFU,MAAM,CAEJ;IAEZ;;;;OAIG;IACH,UAFU,MAAM,CAEA;IAEhB;;;OAGG;IACH,wBAFU,MAAM,CAEc;IAE9B;;;;;;OAMG;IACH,YAFU,MAAM,EAAE,CAEA;IAElB;;;OAGG;IACH,iBAFU,OAAO,CAEM;IAEvB;;;;OAIG;IACH,YAFU,MAAM,CAAC,MAAM,EAAE,MAAM,CAAC,GAAC,IAAI,CAEnB;IAElB;;;OAGG;IACH,YAFU,MAAM,CAAC,MAAM,EAAE,MAAM,CAAC,GAAC,IAAI,CAEnB;IAElB;;;;OAIG;IACH,6BAFU,MAAM,CAEgB;CACnC;kDAGY,GAAqE,GAAG;IAAC,iBAAiB,EAAE,uBAAuB,CAAA;CAAC,GAAG,uBAAuB;iCAvF1H,yCAAyC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"processing_whisper.d.ts","sourceRoot":"","sources":["../../../src/models/whisper/processing_whisper.js"],"names":[],"mappings":"AAIA;;GAEG;AACH;IACI,6CAAsC;IACtC,4DAAqD;IAErD;;;;OAIG;IACH,aAHW,GAAG,GACD,
|
|
1
|
+
{"version":3,"file":"processing_whisper.d.ts","sourceRoot":"","sources":["../../../src/models/whisper/processing_whisper.js"],"names":[],"mappings":"AAIA;;GAEG;AACH;IACI,6CAAsC;IACtC,4DAAqD;IAErD;;;;OAIG;IACH,aAHW,GAAG,GACD,OAAO,CAAC,GAAG,CAAC,CAIxB;CACJ;0BAjByB,gCAAgC;8BAD5B,qBAAqB;qCADd,oCAAoC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"image_processing_yolos.d.ts","sourceRoot":"","sources":["../../../src/models/yolos/image_processing_yolos.js"],"names":[],"mappings":"AAKA;
|
|
1
|
+
{"version":3,"file":"image_processing_yolos.d.ts","sourceRoot":"","sources":["../../../src/models/yolos/image_processing_yolos.js"],"names":[],"mappings":"AAKA;;;;;CAKC;AACD;CAAkE;+BAR3D,sCAAsC"}
|