@huggingface/tasks 0.13.4 → 0.13.6
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/commonjs/hardware.d.ts +12 -0
- package/dist/commonjs/hardware.d.ts.map +1 -1
- package/dist/commonjs/hardware.js +15 -3
- package/dist/commonjs/local-apps.d.ts.map +1 -1
- package/dist/commonjs/local-apps.js +40 -21
- package/dist/commonjs/local-apps.spec.d.ts +2 -0
- package/dist/commonjs/local-apps.spec.d.ts.map +1 -0
- package/dist/commonjs/local-apps.spec.js +114 -0
- package/dist/commonjs/model-libraries-snippets.d.ts.map +1 -1
- package/dist/commonjs/model-libraries-snippets.js +23 -12
- package/dist/commonjs/model-libraries-snippets.spec.d.ts +2 -0
- package/dist/commonjs/model-libraries-snippets.spec.d.ts.map +1 -0
- package/dist/commonjs/model-libraries-snippets.spec.js +51 -0
- package/dist/commonjs/snippets/curl.js +4 -4
- package/dist/commonjs/snippets/js.js +5 -5
- package/dist/commonjs/tasks/audio-classification/inference.d.ts +4 -3
- package/dist/commonjs/tasks/audio-classification/inference.d.ts.map +1 -1
- package/dist/commonjs/tasks/automatic-speech-recognition/inference.d.ts +1 -5
- package/dist/commonjs/tasks/automatic-speech-recognition/inference.d.ts.map +1 -1
- package/dist/commonjs/tasks/chat-completion/inference.d.ts +18 -6
- package/dist/commonjs/tasks/chat-completion/inference.d.ts.map +1 -1
- package/dist/commonjs/tasks/depth-estimation/inference.d.ts +1 -1
- package/dist/commonjs/tasks/document-question-answering/inference.d.ts +1 -3
- package/dist/commonjs/tasks/document-question-answering/inference.d.ts.map +1 -1
- package/dist/commonjs/tasks/feature-extraction/inference.d.ts +1 -1
- package/dist/commonjs/tasks/fill-mask/inference.d.ts +1 -3
- package/dist/commonjs/tasks/fill-mask/inference.d.ts.map +1 -1
- package/dist/commonjs/tasks/image-classification/inference.d.ts +4 -3
- package/dist/commonjs/tasks/image-classification/inference.d.ts.map +1 -1
- package/dist/commonjs/tasks/image-segmentation/inference.d.ts +1 -3
- package/dist/commonjs/tasks/image-segmentation/inference.d.ts.map +1 -1
- package/dist/commonjs/tasks/image-to-image/inference.d.ts +1 -3
- package/dist/commonjs/tasks/image-to-image/inference.d.ts.map +1 -1
- package/dist/commonjs/tasks/image-to-text/inference.d.ts +1 -5
- package/dist/commonjs/tasks/image-to-text/inference.d.ts.map +1 -1
- package/dist/commonjs/tasks/object-detection/inference.d.ts +1 -3
- package/dist/commonjs/tasks/object-detection/inference.d.ts.map +1 -1
- package/dist/commonjs/tasks/question-answering/inference.d.ts +1 -3
- package/dist/commonjs/tasks/question-answering/inference.d.ts.map +1 -1
- package/dist/commonjs/tasks/sentence-similarity/inference.d.ts +1 -1
- package/dist/commonjs/tasks/summarization/inference.d.ts +1 -3
- package/dist/commonjs/tasks/summarization/inference.d.ts.map +1 -1
- package/dist/commonjs/tasks/table-question-answering/inference.d.ts +1 -1
- package/dist/commonjs/tasks/text-classification/inference.d.ts +4 -3
- package/dist/commonjs/tasks/text-classification/inference.d.ts.map +1 -1
- package/dist/commonjs/tasks/text-to-audio/inference.d.ts +1 -5
- package/dist/commonjs/tasks/text-to-audio/inference.d.ts.map +1 -1
- package/dist/commonjs/tasks/text-to-image/inference.d.ts +1 -3
- package/dist/commonjs/tasks/text-to-image/inference.d.ts.map +1 -1
- package/dist/commonjs/tasks/text-to-speech/inference.d.ts +1 -5
- package/dist/commonjs/tasks/text-to-speech/inference.d.ts.map +1 -1
- package/dist/commonjs/tasks/text2text-generation/inference.d.ts +1 -3
- package/dist/commonjs/tasks/text2text-generation/inference.d.ts.map +1 -1
- package/dist/commonjs/tasks/token-classification/inference.d.ts +1 -3
- package/dist/commonjs/tasks/token-classification/inference.d.ts.map +1 -1
- package/dist/commonjs/tasks/translation/inference.d.ts +1 -3
- package/dist/commonjs/tasks/translation/inference.d.ts.map +1 -1
- package/dist/commonjs/tasks/video-classification/inference.d.ts +4 -3
- package/dist/commonjs/tasks/video-classification/inference.d.ts.map +1 -1
- package/dist/commonjs/tasks/visual-question-answering/inference.d.ts +1 -3
- package/dist/commonjs/tasks/visual-question-answering/inference.d.ts.map +1 -1
- package/dist/commonjs/tasks/zero-shot-classification/inference.d.ts +1 -3
- package/dist/commonjs/tasks/zero-shot-classification/inference.d.ts.map +1 -1
- package/dist/commonjs/tasks/zero-shot-image-classification/inference.d.ts +1 -3
- package/dist/commonjs/tasks/zero-shot-image-classification/inference.d.ts.map +1 -1
- package/dist/commonjs/tasks/zero-shot-object-detection/inference.d.ts +1 -3
- package/dist/commonjs/tasks/zero-shot-object-detection/inference.d.ts.map +1 -1
- package/dist/esm/hardware.d.ts +12 -0
- package/dist/esm/hardware.d.ts.map +1 -1
- package/dist/esm/hardware.js +15 -3
- package/dist/esm/local-apps.d.ts.map +1 -1
- package/dist/esm/local-apps.js +40 -21
- package/dist/esm/local-apps.spec.d.ts +2 -0
- package/dist/esm/local-apps.spec.d.ts.map +1 -0
- package/dist/esm/local-apps.spec.js +112 -0
- package/dist/esm/model-libraries-snippets.d.ts.map +1 -1
- package/dist/esm/model-libraries-snippets.js +23 -12
- package/dist/esm/model-libraries-snippets.spec.d.ts +2 -0
- package/dist/esm/model-libraries-snippets.spec.d.ts.map +1 -0
- package/dist/esm/model-libraries-snippets.spec.js +49 -0
- package/dist/esm/snippets/curl.js +4 -4
- package/dist/esm/snippets/js.js +5 -5
- package/dist/esm/tasks/audio-classification/inference.d.ts +4 -3
- package/dist/esm/tasks/audio-classification/inference.d.ts.map +1 -1
- package/dist/esm/tasks/automatic-speech-recognition/inference.d.ts +1 -5
- package/dist/esm/tasks/automatic-speech-recognition/inference.d.ts.map +1 -1
- package/dist/esm/tasks/chat-completion/inference.d.ts +18 -6
- package/dist/esm/tasks/chat-completion/inference.d.ts.map +1 -1
- package/dist/esm/tasks/depth-estimation/inference.d.ts +1 -1
- package/dist/esm/tasks/document-question-answering/inference.d.ts +1 -3
- package/dist/esm/tasks/document-question-answering/inference.d.ts.map +1 -1
- package/dist/esm/tasks/feature-extraction/inference.d.ts +1 -1
- package/dist/esm/tasks/fill-mask/inference.d.ts +1 -3
- package/dist/esm/tasks/fill-mask/inference.d.ts.map +1 -1
- package/dist/esm/tasks/image-classification/inference.d.ts +4 -3
- package/dist/esm/tasks/image-classification/inference.d.ts.map +1 -1
- package/dist/esm/tasks/image-segmentation/inference.d.ts +1 -3
- package/dist/esm/tasks/image-segmentation/inference.d.ts.map +1 -1
- package/dist/esm/tasks/image-to-image/inference.d.ts +1 -3
- package/dist/esm/tasks/image-to-image/inference.d.ts.map +1 -1
- package/dist/esm/tasks/image-to-text/inference.d.ts +1 -5
- package/dist/esm/tasks/image-to-text/inference.d.ts.map +1 -1
- package/dist/esm/tasks/object-detection/inference.d.ts +1 -3
- package/dist/esm/tasks/object-detection/inference.d.ts.map +1 -1
- package/dist/esm/tasks/question-answering/inference.d.ts +1 -3
- package/dist/esm/tasks/question-answering/inference.d.ts.map +1 -1
- package/dist/esm/tasks/sentence-similarity/inference.d.ts +1 -1
- package/dist/esm/tasks/summarization/inference.d.ts +1 -3
- package/dist/esm/tasks/summarization/inference.d.ts.map +1 -1
- package/dist/esm/tasks/table-question-answering/inference.d.ts +1 -1
- package/dist/esm/tasks/text-classification/inference.d.ts +4 -3
- package/dist/esm/tasks/text-classification/inference.d.ts.map +1 -1
- package/dist/esm/tasks/text-to-audio/inference.d.ts +1 -5
- package/dist/esm/tasks/text-to-audio/inference.d.ts.map +1 -1
- package/dist/esm/tasks/text-to-image/inference.d.ts +1 -3
- package/dist/esm/tasks/text-to-image/inference.d.ts.map +1 -1
- package/dist/esm/tasks/text-to-speech/inference.d.ts +1 -5
- package/dist/esm/tasks/text-to-speech/inference.d.ts.map +1 -1
- package/dist/esm/tasks/text2text-generation/inference.d.ts +1 -3
- package/dist/esm/tasks/text2text-generation/inference.d.ts.map +1 -1
- package/dist/esm/tasks/token-classification/inference.d.ts +1 -3
- package/dist/esm/tasks/token-classification/inference.d.ts.map +1 -1
- package/dist/esm/tasks/translation/inference.d.ts +1 -3
- package/dist/esm/tasks/translation/inference.d.ts.map +1 -1
- package/dist/esm/tasks/video-classification/inference.d.ts +4 -3
- package/dist/esm/tasks/video-classification/inference.d.ts.map +1 -1
- package/dist/esm/tasks/visual-question-answering/inference.d.ts +1 -3
- package/dist/esm/tasks/visual-question-answering/inference.d.ts.map +1 -1
- package/dist/esm/tasks/zero-shot-classification/inference.d.ts +1 -3
- package/dist/esm/tasks/zero-shot-classification/inference.d.ts.map +1 -1
- package/dist/esm/tasks/zero-shot-image-classification/inference.d.ts +1 -3
- package/dist/esm/tasks/zero-shot-image-classification/inference.d.ts.map +1 -1
- package/dist/esm/tasks/zero-shot-object-detection/inference.d.ts +1 -3
- package/dist/esm/tasks/zero-shot-object-detection/inference.d.ts.map +1 -1
- package/package.json +1 -1
- package/src/hardware.ts +15 -3
- package/src/local-apps.spec.ts +123 -0
- package/src/local-apps.ts +37 -18
- package/src/model-libraries-snippets.spec.ts +54 -0
- package/src/model-libraries-snippets.ts +24 -11
- package/src/snippets/curl.ts +4 -4
- package/src/snippets/js.ts +5 -5
- package/src/tasks/audio-classification/inference.ts +4 -3
- package/src/tasks/audio-classification/spec/input.json +3 -3
- package/src/tasks/automatic-speech-recognition/inference.ts +1 -5
- package/src/tasks/automatic-speech-recognition/spec/input.json +1 -2
- package/src/tasks/chat-completion/inference.ts +19 -6
- package/src/tasks/chat-completion/spec/input.json +14 -19
- package/src/tasks/common-definitions.json +0 -1
- package/src/tasks/depth-estimation/inference.ts +1 -1
- package/src/tasks/depth-estimation/spec/input.json +1 -2
- package/src/tasks/document-question-answering/inference.ts +1 -3
- package/src/tasks/document-question-answering/spec/input.json +1 -2
- package/src/tasks/feature-extraction/inference.ts +1 -1
- package/src/tasks/feature-extraction/spec/input.json +1 -1
- package/src/tasks/fill-mask/inference.ts +1 -3
- package/src/tasks/fill-mask/spec/input.json +1 -2
- package/src/tasks/image-classification/inference.ts +4 -3
- package/src/tasks/image-classification/spec/input.json +3 -3
- package/src/tasks/image-segmentation/inference.ts +1 -3
- package/src/tasks/image-segmentation/spec/input.json +1 -2
- package/src/tasks/image-to-image/inference.ts +1 -3
- package/src/tasks/image-to-image/spec/input.json +1 -2
- package/src/tasks/image-to-text/inference.ts +1 -5
- package/src/tasks/image-to-text/spec/input.json +1 -2
- package/src/tasks/object-detection/inference.ts +1 -3
- package/src/tasks/object-detection/spec/input.json +1 -2
- package/src/tasks/placeholder/spec/input.json +1 -2
- package/src/tasks/question-answering/inference.ts +1 -3
- package/src/tasks/question-answering/spec/input.json +1 -2
- package/src/tasks/sentence-similarity/inference.ts +1 -1
- package/src/tasks/sentence-similarity/spec/input.json +1 -2
- package/src/tasks/summarization/inference.ts +1 -3
- package/src/tasks/summarization/spec/input.json +1 -2
- package/src/tasks/table-question-answering/inference.ts +1 -1
- package/src/tasks/table-question-answering/spec/input.json +1 -2
- package/src/tasks/text-classification/inference.ts +4 -3
- package/src/tasks/text-classification/spec/input.json +3 -3
- package/src/tasks/text-to-audio/inference.ts +1 -5
- package/src/tasks/text-to-audio/spec/input.json +1 -2
- package/src/tasks/text-to-image/inference.ts +1 -3
- package/src/tasks/text-to-image/spec/input.json +1 -2
- package/src/tasks/text-to-speech/inference.ts +1 -5
- package/src/tasks/text-to-speech/spec/input.json +1 -2
- package/src/tasks/text2text-generation/inference.ts +1 -3
- package/src/tasks/text2text-generation/spec/input.json +1 -2
- package/src/tasks/token-classification/inference.ts +1 -3
- package/src/tasks/token-classification/spec/input.json +1 -2
- package/src/tasks/translation/inference.ts +1 -3
- package/src/tasks/translation/spec/input.json +1 -2
- package/src/tasks/video-classification/inference.ts +4 -3
- package/src/tasks/video-classification/spec/input.json +3 -3
- package/src/tasks/visual-question-answering/inference.ts +1 -3
- package/src/tasks/visual-question-answering/spec/input.json +1 -2
- package/src/tasks/zero-shot-classification/inference.ts +1 -3
- package/src/tasks/zero-shot-classification/spec/input.json +1 -2
- package/src/tasks/zero-shot-image-classification/inference.ts +1 -3
- package/src/tasks/zero-shot-image-classification/spec/input.json +1 -2
- package/src/tasks/zero-shot-object-detection/inference.ts +1 -3
- package/src/tasks/zero-shot-object-detection/spec/input.json +1 -2
- package/dist/commonjs/snippets/curl.spec.d.ts +0 -2
- package/dist/commonjs/snippets/curl.spec.d.ts.map +0 -1
- package/dist/commonjs/snippets/curl.spec.js +0 -89
- package/dist/commonjs/snippets/js.spec.d.ts +0 -2
- package/dist/commonjs/snippets/js.spec.d.ts.map +0 -1
- package/dist/commonjs/snippets/js.spec.js +0 -141
- package/dist/commonjs/snippets/python.spec.d.ts +0 -2
- package/dist/commonjs/snippets/python.spec.d.ts.map +0 -1
- package/dist/commonjs/snippets/python.spec.js +0 -135
- package/dist/esm/snippets/curl.spec.d.ts +0 -2
- package/dist/esm/snippets/curl.spec.d.ts.map +0 -1
- package/dist/esm/snippets/curl.spec.js +0 -87
- package/dist/esm/snippets/js.spec.d.ts +0 -2
- package/dist/esm/snippets/js.spec.d.ts.map +0 -1
- package/dist/esm/snippets/js.spec.js +0 -139
- package/dist/esm/snippets/python.spec.d.ts +0 -2
- package/dist/esm/snippets/python.spec.d.ts.map +0 -1
- package/dist/esm/snippets/python.spec.js +0 -133
- package/src/snippets/curl.spec.ts +0 -94
- package/src/snippets/js.spec.ts +0 -148
- package/src/snippets/python.spec.ts +0 -144
|
@@ -13,15 +13,13 @@ export interface TextToSpeechInput {
|
|
|
13
13
|
*/
|
|
14
14
|
inputs: string;
|
|
15
15
|
/**
|
|
16
|
-
* Additional inference parameters
|
|
16
|
+
* Additional inference parameters for Text To Speech
|
|
17
17
|
*/
|
|
18
18
|
parameters?: TextToSpeechParameters;
|
|
19
19
|
[property: string]: unknown;
|
|
20
20
|
}
|
|
21
21
|
|
|
22
22
|
/**
|
|
23
|
-
* Additional inference parameters
|
|
24
|
-
*
|
|
25
23
|
* Additional inference parameters for Text To Speech
|
|
26
24
|
*/
|
|
27
25
|
export interface TextToSpeechParameters {
|
|
@@ -34,8 +32,6 @@ export interface TextToSpeechParameters {
|
|
|
34
32
|
|
|
35
33
|
/**
|
|
36
34
|
* Parametrization of the text generation process
|
|
37
|
-
*
|
|
38
|
-
* Ad-hoc parametrization of the text generation process
|
|
39
35
|
*/
|
|
40
36
|
export interface GenerationParameters {
|
|
41
37
|
/**
|
|
@@ -10,14 +10,13 @@
|
|
|
10
10
|
"type": "string"
|
|
11
11
|
},
|
|
12
12
|
"parameters": {
|
|
13
|
-
"description": "Additional inference parameters",
|
|
13
|
+
"description": "Additional inference parameters for Text To Speech",
|
|
14
14
|
"$ref": "#/$defs/TextToSpeechParameters"
|
|
15
15
|
}
|
|
16
16
|
},
|
|
17
17
|
"$defs": {
|
|
18
18
|
"TextToSpeechParameters": {
|
|
19
19
|
"title": "TextToSpeechParameters",
|
|
20
|
-
"description": "Additional inference parameters for Text To Speech",
|
|
21
20
|
"type": "object",
|
|
22
21
|
"properties": {
|
|
23
22
|
"generation_parameters": {
|
|
@@ -13,15 +13,13 @@ export interface Text2TextGenerationInput {
|
|
|
13
13
|
*/
|
|
14
14
|
inputs: string;
|
|
15
15
|
/**
|
|
16
|
-
* Additional inference parameters
|
|
16
|
+
* Additional inference parameters for Text2text Generation
|
|
17
17
|
*/
|
|
18
18
|
parameters?: Text2TextGenerationParameters;
|
|
19
19
|
[property: string]: unknown;
|
|
20
20
|
}
|
|
21
21
|
|
|
22
22
|
/**
|
|
23
|
-
* Additional inference parameters
|
|
24
|
-
*
|
|
25
23
|
* Additional inference parameters for Text2text Generation
|
|
26
24
|
*/
|
|
27
25
|
export interface Text2TextGenerationParameters {
|
|
@@ -10,14 +10,13 @@
|
|
|
10
10
|
"type": "string"
|
|
11
11
|
},
|
|
12
12
|
"parameters": {
|
|
13
|
-
"description": "Additional inference parameters",
|
|
13
|
+
"description": "Additional inference parameters for Text2text Generation",
|
|
14
14
|
"$ref": "#/$defs/Text2textGenerationParameters"
|
|
15
15
|
}
|
|
16
16
|
},
|
|
17
17
|
"$defs": {
|
|
18
18
|
"Text2textGenerationParameters": {
|
|
19
19
|
"title": "Text2textGenerationParameters",
|
|
20
|
-
"description": "Additional inference parameters for Text2text Generation",
|
|
21
20
|
"type": "object",
|
|
22
21
|
"properties": {
|
|
23
22
|
"clean_up_tokenization_spaces": {
|
|
@@ -12,14 +12,12 @@ export interface TokenClassificationInput {
|
|
|
12
12
|
*/
|
|
13
13
|
inputs: string;
|
|
14
14
|
/**
|
|
15
|
-
* Additional inference parameters
|
|
15
|
+
* Additional inference parameters for Token Classification
|
|
16
16
|
*/
|
|
17
17
|
parameters?: TokenClassificationParameters;
|
|
18
18
|
[property: string]: unknown;
|
|
19
19
|
}
|
|
20
20
|
/**
|
|
21
|
-
* Additional inference parameters
|
|
22
|
-
*
|
|
23
21
|
* Additional inference parameters for Token Classification
|
|
24
22
|
*/
|
|
25
23
|
export interface TokenClassificationParameters {
|
|
@@ -10,14 +10,13 @@
|
|
|
10
10
|
"type": "string"
|
|
11
11
|
},
|
|
12
12
|
"parameters": {
|
|
13
|
-
"description": "Additional inference parameters",
|
|
13
|
+
"description": "Additional inference parameters for Token Classification",
|
|
14
14
|
"$ref": "#/$defs/TokenClassificationParameters"
|
|
15
15
|
}
|
|
16
16
|
},
|
|
17
17
|
"$defs": {
|
|
18
18
|
"TokenClassificationParameters": {
|
|
19
19
|
"title": "TokenClassificationParameters",
|
|
20
|
-
"description": "Additional inference parameters for Token Classification",
|
|
21
20
|
"type": "object",
|
|
22
21
|
"properties": {
|
|
23
22
|
"ignore_labels": {
|
|
@@ -13,15 +13,13 @@ export interface TranslationInput {
|
|
|
13
13
|
*/
|
|
14
14
|
inputs: string;
|
|
15
15
|
/**
|
|
16
|
-
* Additional inference parameters
|
|
16
|
+
* Additional inference parameters for Translation
|
|
17
17
|
*/
|
|
18
18
|
parameters?: TranslationParameters;
|
|
19
19
|
[property: string]: unknown;
|
|
20
20
|
}
|
|
21
21
|
|
|
22
22
|
/**
|
|
23
|
-
* Additional inference parameters
|
|
24
|
-
*
|
|
25
23
|
* Additional inference parameters for Translation
|
|
26
24
|
*/
|
|
27
25
|
export interface TranslationParameters {
|
|
@@ -10,14 +10,13 @@
|
|
|
10
10
|
"type": "string"
|
|
11
11
|
},
|
|
12
12
|
"parameters": {
|
|
13
|
-
"description": "Additional inference parameters",
|
|
13
|
+
"description": "Additional inference parameters for Translation",
|
|
14
14
|
"$ref": "#/$defs/TranslationParameters"
|
|
15
15
|
}
|
|
16
16
|
},
|
|
17
17
|
"$defs": {
|
|
18
18
|
"TranslationParameters": {
|
|
19
19
|
"title": "TranslationParameters",
|
|
20
|
-
"description": "Additional inference parameters for Translation",
|
|
21
20
|
"type": "object",
|
|
22
21
|
"properties": {
|
|
23
22
|
"src_lang": {
|
|
@@ -12,14 +12,12 @@ export interface VideoClassificationInput {
|
|
|
12
12
|
*/
|
|
13
13
|
inputs: unknown;
|
|
14
14
|
/**
|
|
15
|
-
* Additional inference parameters
|
|
15
|
+
* Additional inference parameters for Video Classification
|
|
16
16
|
*/
|
|
17
17
|
parameters?: VideoClassificationParameters;
|
|
18
18
|
[property: string]: unknown;
|
|
19
19
|
}
|
|
20
20
|
/**
|
|
21
|
-
* Additional inference parameters
|
|
22
|
-
*
|
|
23
21
|
* Additional inference parameters for Video Classification
|
|
24
22
|
*/
|
|
25
23
|
export interface VideoClassificationParameters {
|
|
@@ -27,6 +25,9 @@ export interface VideoClassificationParameters {
|
|
|
27
25
|
* The sampling rate used to select frames from the video.
|
|
28
26
|
*/
|
|
29
27
|
frame_sampling_rate?: number;
|
|
28
|
+
/**
|
|
29
|
+
* The function to apply to the model outputs in order to retrieve the scores.
|
|
30
|
+
*/
|
|
30
31
|
function_to_apply?: ClassificationOutputTransform;
|
|
31
32
|
/**
|
|
32
33
|
* The number of sampled frames to consider for classification.
|
|
@@ -9,19 +9,19 @@
|
|
|
9
9
|
"description": "The input video data"
|
|
10
10
|
},
|
|
11
11
|
"parameters": {
|
|
12
|
-
"description": "Additional inference parameters",
|
|
12
|
+
"description": "Additional inference parameters for Video Classification",
|
|
13
13
|
"$ref": "#/$defs/VideoClassificationParameters"
|
|
14
14
|
}
|
|
15
15
|
},
|
|
16
16
|
"$defs": {
|
|
17
17
|
"VideoClassificationParameters": {
|
|
18
18
|
"title": "VideoClassificationParameters",
|
|
19
|
-
"description": "Additional inference parameters for Video Classification",
|
|
20
19
|
"type": "object",
|
|
21
20
|
"properties": {
|
|
22
21
|
"function_to_apply": {
|
|
23
22
|
"title": "TextClassificationOutputTransform",
|
|
24
|
-
"$ref": "/inference/schemas/common-definitions.json#/definitions/ClassificationOutputTransform"
|
|
23
|
+
"$ref": "/inference/schemas/common-definitions.json#/definitions/ClassificationOutputTransform",
|
|
24
|
+
"description": "The function to apply to the model outputs in order to retrieve the scores."
|
|
25
25
|
},
|
|
26
26
|
"num_frames": {
|
|
27
27
|
"type": "integer",
|
|
@@ -12,7 +12,7 @@ export interface VisualQuestionAnsweringInput {
|
|
|
12
12
|
*/
|
|
13
13
|
inputs: VisualQuestionAnsweringInputData;
|
|
14
14
|
/**
|
|
15
|
-
* Additional inference parameters
|
|
15
|
+
* Additional inference parameters for Visual Question Answering
|
|
16
16
|
*/
|
|
17
17
|
parameters?: VisualQuestionAnsweringParameters;
|
|
18
18
|
[property: string]: unknown;
|
|
@@ -32,8 +32,6 @@ export interface VisualQuestionAnsweringInputData {
|
|
|
32
32
|
[property: string]: unknown;
|
|
33
33
|
}
|
|
34
34
|
/**
|
|
35
|
-
* Additional inference parameters
|
|
36
|
-
*
|
|
37
35
|
* Additional inference parameters for Visual Question Answering
|
|
38
36
|
*/
|
|
39
37
|
export interface VisualQuestionAnsweringParameters {
|
|
@@ -20,14 +20,13 @@
|
|
|
20
20
|
"required": ["question", "image"]
|
|
21
21
|
},
|
|
22
22
|
"parameters": {
|
|
23
|
-
"description": "Additional inference parameters",
|
|
23
|
+
"description": "Additional inference parameters for Visual Question Answering",
|
|
24
24
|
"$ref": "#/$defs/VisualQuestionAnsweringParameters"
|
|
25
25
|
}
|
|
26
26
|
},
|
|
27
27
|
"$defs": {
|
|
28
28
|
"VisualQuestionAnsweringParameters": {
|
|
29
29
|
"title": "VisualQuestionAnsweringParameters",
|
|
30
|
-
"description": "Additional inference parameters for Visual Question Answering",
|
|
31
30
|
"type": "object",
|
|
32
31
|
"properties": {
|
|
33
32
|
"top_k": {
|
|
@@ -12,14 +12,12 @@ export interface ZeroShotClassificationInput {
|
|
|
12
12
|
*/
|
|
13
13
|
inputs: string;
|
|
14
14
|
/**
|
|
15
|
-
* Additional inference parameters
|
|
15
|
+
* Additional inference parameters for Zero Shot Classification
|
|
16
16
|
*/
|
|
17
17
|
parameters: ZeroShotClassificationParameters;
|
|
18
18
|
[property: string]: unknown;
|
|
19
19
|
}
|
|
20
20
|
/**
|
|
21
|
-
* Additional inference parameters
|
|
22
|
-
*
|
|
23
21
|
* Additional inference parameters for Zero Shot Classification
|
|
24
22
|
*/
|
|
25
23
|
export interface ZeroShotClassificationParameters {
|
|
@@ -10,14 +10,13 @@
|
|
|
10
10
|
"type": "string"
|
|
11
11
|
},
|
|
12
12
|
"parameters": {
|
|
13
|
-
"description": "Additional inference parameters",
|
|
13
|
+
"description": "Additional inference parameters for Zero Shot Classification",
|
|
14
14
|
"$ref": "#/$defs/ZeroShotClassificationParameters"
|
|
15
15
|
}
|
|
16
16
|
},
|
|
17
17
|
"$defs": {
|
|
18
18
|
"ZeroShotClassificationParameters": {
|
|
19
19
|
"title": "ZeroShotClassificationParameters",
|
|
20
|
-
"description": "Additional inference parameters for Zero Shot Classification",
|
|
21
20
|
"type": "object",
|
|
22
21
|
"properties": {
|
|
23
22
|
"candidate_labels": {
|
|
@@ -12,14 +12,12 @@ export interface ZeroShotImageClassificationInput {
|
|
|
12
12
|
*/
|
|
13
13
|
inputs: string;
|
|
14
14
|
/**
|
|
15
|
-
* Additional inference parameters
|
|
15
|
+
* Additional inference parameters for Zero Shot Image Classification
|
|
16
16
|
*/
|
|
17
17
|
parameters: ZeroShotImageClassificationParameters;
|
|
18
18
|
[property: string]: unknown;
|
|
19
19
|
}
|
|
20
20
|
/**
|
|
21
|
-
* Additional inference parameters
|
|
22
|
-
*
|
|
23
21
|
* Additional inference parameters for Zero Shot Image Classification
|
|
24
22
|
*/
|
|
25
23
|
export interface ZeroShotImageClassificationParameters {
|
|
@@ -10,14 +10,13 @@
|
|
|
10
10
|
"description": "The input image data to classify as a base64-encoded string."
|
|
11
11
|
},
|
|
12
12
|
"parameters": {
|
|
13
|
-
"description": "Additional inference parameters",
|
|
13
|
+
"description": "Additional inference parameters for Zero Shot Image Classification",
|
|
14
14
|
"$ref": "#/$defs/ZeroShotImageClassificationParameters"
|
|
15
15
|
}
|
|
16
16
|
},
|
|
17
17
|
"$defs": {
|
|
18
18
|
"ZeroShotImageClassificationParameters": {
|
|
19
19
|
"title": "ZeroShotImageClassificationParameters",
|
|
20
|
-
"description": "Additional inference parameters for Zero Shot Image Classification",
|
|
21
20
|
"type": "object",
|
|
22
21
|
"properties": {
|
|
23
22
|
"candidate_labels": {
|
|
@@ -12,14 +12,12 @@ export interface ZeroShotObjectDetectionInput {
|
|
|
12
12
|
*/
|
|
13
13
|
inputs: string;
|
|
14
14
|
/**
|
|
15
|
-
* Additional inference parameters
|
|
15
|
+
* Additional inference parameters for Zero Shot Object Detection
|
|
16
16
|
*/
|
|
17
17
|
parameters: ZeroShotObjectDetectionParameters;
|
|
18
18
|
[property: string]: unknown;
|
|
19
19
|
}
|
|
20
20
|
/**
|
|
21
|
-
* Additional inference parameters
|
|
22
|
-
*
|
|
23
21
|
* Additional inference parameters for Zero Shot Object Detection
|
|
24
22
|
*/
|
|
25
23
|
export interface ZeroShotObjectDetectionParameters {
|
|
@@ -10,14 +10,13 @@
|
|
|
10
10
|
"type": "string"
|
|
11
11
|
},
|
|
12
12
|
"parameters": {
|
|
13
|
-
"description": "Additional inference parameters",
|
|
13
|
+
"description": "Additional inference parameters for Zero Shot Object Detection",
|
|
14
14
|
"$ref": "#/$defs/ZeroShotObjectDetectionParameters"
|
|
15
15
|
}
|
|
16
16
|
},
|
|
17
17
|
"$defs": {
|
|
18
18
|
"ZeroShotObjectDetectionParameters": {
|
|
19
19
|
"title": "ZeroShotObjectDetectionParameters",
|
|
20
|
-
"description": "Additional inference parameters for Zero Shot Object Detection",
|
|
21
20
|
"type": "object",
|
|
22
21
|
"properties": {
|
|
23
22
|
"candidate_labels": {
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"file":"curl.spec.d.ts","sourceRoot":"","sources":["../../../src/snippets/curl.spec.ts"],"names":[],"mappings":""}
|
|
@@ -1,89 +0,0 @@
|
|
|
1
|
-
"use strict";
|
|
2
|
-
Object.defineProperty(exports, "__esModule", { value: true });
|
|
3
|
-
const vitest_1 = require("vitest");
|
|
4
|
-
const curl_js_1 = require("./curl.js");
|
|
5
|
-
(0, vitest_1.describe)("inference API snippets", () => {
|
|
6
|
-
(0, vitest_1.it)("conversational llm", async () => {
|
|
7
|
-
const model = {
|
|
8
|
-
id: "meta-llama/Llama-3.1-8B-Instruct",
|
|
9
|
-
pipeline_tag: "text-generation",
|
|
10
|
-
tags: ["conversational"],
|
|
11
|
-
inference: "",
|
|
12
|
-
};
|
|
13
|
-
const snippet = (0, curl_js_1.getCurlInferenceSnippet)(model, "api_token");
|
|
14
|
-
(0, vitest_1.expect)(snippet.content)
|
|
15
|
-
.toEqual(`curl 'https://api-inference.huggingface.co/models/meta-llama/Llama-3.1-8B-Instruct/v1/chat/completions' \\
|
|
16
|
-
-H "Authorization: Bearer api_token" \\
|
|
17
|
-
-H 'Content-Type: application/json' \\
|
|
18
|
-
--data '{
|
|
19
|
-
"model": "meta-llama/Llama-3.1-8B-Instruct",
|
|
20
|
-
"messages": [
|
|
21
|
-
{
|
|
22
|
-
"role": "user",
|
|
23
|
-
"content": "What is the capital of France?"
|
|
24
|
-
}
|
|
25
|
-
],
|
|
26
|
-
"max_tokens": 500,
|
|
27
|
-
"stream": true
|
|
28
|
-
}'`);
|
|
29
|
-
});
|
|
30
|
-
(0, vitest_1.it)("conversational llm non-streaming", async () => {
|
|
31
|
-
const model = {
|
|
32
|
-
id: "meta-llama/Llama-3.1-8B-Instruct",
|
|
33
|
-
pipeline_tag: "text-generation",
|
|
34
|
-
tags: ["conversational"],
|
|
35
|
-
inference: "",
|
|
36
|
-
};
|
|
37
|
-
const snippet = (0, curl_js_1.getCurlInferenceSnippet)(model, "api_token", { streaming: false });
|
|
38
|
-
(0, vitest_1.expect)(snippet.content)
|
|
39
|
-
.toEqual(`curl 'https://api-inference.huggingface.co/models/meta-llama/Llama-3.1-8B-Instruct/v1/chat/completions' \\
|
|
40
|
-
-H "Authorization: Bearer api_token" \\
|
|
41
|
-
-H 'Content-Type: application/json' \\
|
|
42
|
-
--data '{
|
|
43
|
-
"model": "meta-llama/Llama-3.1-8B-Instruct",
|
|
44
|
-
"messages": [
|
|
45
|
-
{
|
|
46
|
-
"role": "user",
|
|
47
|
-
"content": "What is the capital of France?"
|
|
48
|
-
}
|
|
49
|
-
],
|
|
50
|
-
"max_tokens": 500,
|
|
51
|
-
"stream": false
|
|
52
|
-
}'`);
|
|
53
|
-
});
|
|
54
|
-
(0, vitest_1.it)("conversational vlm", async () => {
|
|
55
|
-
const model = {
|
|
56
|
-
id: "meta-llama/Llama-3.2-11B-Vision-Instruct",
|
|
57
|
-
pipeline_tag: "image-text-to-text",
|
|
58
|
-
tags: ["conversational"],
|
|
59
|
-
inference: "",
|
|
60
|
-
};
|
|
61
|
-
const snippet = (0, curl_js_1.getCurlInferenceSnippet)(model, "api_token");
|
|
62
|
-
(0, vitest_1.expect)(snippet.content)
|
|
63
|
-
.toEqual(`curl 'https://api-inference.huggingface.co/models/meta-llama/Llama-3.2-11B-Vision-Instruct/v1/chat/completions' \\
|
|
64
|
-
-H "Authorization: Bearer api_token" \\
|
|
65
|
-
-H 'Content-Type: application/json' \\
|
|
66
|
-
--data '{
|
|
67
|
-
"model": "meta-llama/Llama-3.2-11B-Vision-Instruct",
|
|
68
|
-
"messages": [
|
|
69
|
-
{
|
|
70
|
-
"role": "user",
|
|
71
|
-
"content": [
|
|
72
|
-
{
|
|
73
|
-
"type": "text",
|
|
74
|
-
"text": "Describe this image in one sentence."
|
|
75
|
-
},
|
|
76
|
-
{
|
|
77
|
-
"type": "image_url",
|
|
78
|
-
"image_url": {
|
|
79
|
-
"url": "https://cdn.britannica.com/61/93061-050-99147DCE/Statue-of-Liberty-Island-New-York-Bay.jpg"
|
|
80
|
-
}
|
|
81
|
-
}
|
|
82
|
-
]
|
|
83
|
-
}
|
|
84
|
-
],
|
|
85
|
-
"max_tokens": 500,
|
|
86
|
-
"stream": true
|
|
87
|
-
}'`);
|
|
88
|
-
});
|
|
89
|
-
});
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"file":"js.spec.d.ts","sourceRoot":"","sources":["../../../src/snippets/js.spec.ts"],"names":[],"mappings":""}
|
|
@@ -1,141 +0,0 @@
|
|
|
1
|
-
"use strict";
|
|
2
|
-
Object.defineProperty(exports, "__esModule", { value: true });
|
|
3
|
-
const vitest_1 = require("vitest");
|
|
4
|
-
const js_js_1 = require("./js.js");
|
|
5
|
-
(0, vitest_1.describe)("inference API snippets", () => {
|
|
6
|
-
(0, vitest_1.it)("conversational llm", async () => {
|
|
7
|
-
const model = {
|
|
8
|
-
id: "meta-llama/Llama-3.1-8B-Instruct",
|
|
9
|
-
pipeline_tag: "text-generation",
|
|
10
|
-
tags: ["conversational"],
|
|
11
|
-
inference: "",
|
|
12
|
-
};
|
|
13
|
-
const snippet = (0, js_js_1.getJsInferenceSnippet)(model, "api_token");
|
|
14
|
-
(0, vitest_1.expect)(snippet[0].content).toEqual(`import { HfInference } from "@huggingface/inference";
|
|
15
|
-
|
|
16
|
-
const client = new HfInference("api_token");
|
|
17
|
-
|
|
18
|
-
let out = "";
|
|
19
|
-
|
|
20
|
-
const stream = client.chatCompletionStream({
|
|
21
|
-
model: "meta-llama/Llama-3.1-8B-Instruct",
|
|
22
|
-
messages: [
|
|
23
|
-
{
|
|
24
|
-
role: "user",
|
|
25
|
-
content: "What is the capital of France?"
|
|
26
|
-
}
|
|
27
|
-
],
|
|
28
|
-
max_tokens: 500
|
|
29
|
-
});
|
|
30
|
-
|
|
31
|
-
for await (const chunk of stream) {
|
|
32
|
-
if (chunk.choices && chunk.choices.length > 0) {
|
|
33
|
-
const newContent = chunk.choices[0].delta.content;
|
|
34
|
-
out += newContent;
|
|
35
|
-
console.log(newContent);
|
|
36
|
-
}
|
|
37
|
-
}`);
|
|
38
|
-
});
|
|
39
|
-
(0, vitest_1.it)("conversational llm non-streaming", async () => {
|
|
40
|
-
const model = {
|
|
41
|
-
id: "meta-llama/Llama-3.1-8B-Instruct",
|
|
42
|
-
pipeline_tag: "text-generation",
|
|
43
|
-
tags: ["conversational"],
|
|
44
|
-
inference: "",
|
|
45
|
-
};
|
|
46
|
-
const snippet = (0, js_js_1.getJsInferenceSnippet)(model, "api_token", { streaming: false });
|
|
47
|
-
(0, vitest_1.expect)(snippet[0].content).toEqual(`import { HfInference } from "@huggingface/inference";
|
|
48
|
-
|
|
49
|
-
const client = new HfInference("api_token");
|
|
50
|
-
|
|
51
|
-
const chatCompletion = await client.chatCompletion({
|
|
52
|
-
model: "meta-llama/Llama-3.1-8B-Instruct",
|
|
53
|
-
messages: [
|
|
54
|
-
{
|
|
55
|
-
role: "user",
|
|
56
|
-
content: "What is the capital of France?"
|
|
57
|
-
}
|
|
58
|
-
],
|
|
59
|
-
max_tokens: 500
|
|
60
|
-
});
|
|
61
|
-
|
|
62
|
-
console.log(chatCompletion.choices[0].message);`);
|
|
63
|
-
});
|
|
64
|
-
(0, vitest_1.it)("conversational vlm", async () => {
|
|
65
|
-
const model = {
|
|
66
|
-
id: "meta-llama/Llama-3.2-11B-Vision-Instruct",
|
|
67
|
-
pipeline_tag: "image-text-to-text",
|
|
68
|
-
tags: ["conversational"],
|
|
69
|
-
inference: "",
|
|
70
|
-
};
|
|
71
|
-
const snippet = (0, js_js_1.getJsInferenceSnippet)(model, "api_token");
|
|
72
|
-
(0, vitest_1.expect)(snippet[0].content).toEqual(`import { HfInference } from "@huggingface/inference";
|
|
73
|
-
|
|
74
|
-
const client = new HfInference("api_token");
|
|
75
|
-
|
|
76
|
-
let out = "";
|
|
77
|
-
|
|
78
|
-
const stream = client.chatCompletionStream({
|
|
79
|
-
model: "meta-llama/Llama-3.2-11B-Vision-Instruct",
|
|
80
|
-
messages: [
|
|
81
|
-
{
|
|
82
|
-
role: "user",
|
|
83
|
-
content: [
|
|
84
|
-
{
|
|
85
|
-
type: "text",
|
|
86
|
-
text: "Describe this image in one sentence."
|
|
87
|
-
},
|
|
88
|
-
{
|
|
89
|
-
type: "image_url",
|
|
90
|
-
image_url: {
|
|
91
|
-
url: "https://cdn.britannica.com/61/93061-050-99147DCE/Statue-of-Liberty-Island-New-York-Bay.jpg"
|
|
92
|
-
}
|
|
93
|
-
}
|
|
94
|
-
]
|
|
95
|
-
}
|
|
96
|
-
],
|
|
97
|
-
max_tokens: 500
|
|
98
|
-
});
|
|
99
|
-
|
|
100
|
-
for await (const chunk of stream) {
|
|
101
|
-
if (chunk.choices && chunk.choices.length > 0) {
|
|
102
|
-
const newContent = chunk.choices[0].delta.content;
|
|
103
|
-
out += newContent;
|
|
104
|
-
console.log(newContent);
|
|
105
|
-
}
|
|
106
|
-
}`);
|
|
107
|
-
});
|
|
108
|
-
(0, vitest_1.it)("conversational llm", async () => {
|
|
109
|
-
const model = {
|
|
110
|
-
id: "meta-llama/Llama-3.1-8B-Instruct",
|
|
111
|
-
pipeline_tag: "text-generation",
|
|
112
|
-
tags: ["conversational"],
|
|
113
|
-
inference: "",
|
|
114
|
-
};
|
|
115
|
-
const snippet = (0, js_js_1.getJsInferenceSnippet)(model, "api_token");
|
|
116
|
-
(0, vitest_1.expect)(snippet[0].content).toEqual(`import { HfInference } from "@huggingface/inference";
|
|
117
|
-
|
|
118
|
-
const client = new HfInference("api_token");
|
|
119
|
-
|
|
120
|
-
let out = "";
|
|
121
|
-
|
|
122
|
-
const stream = client.chatCompletionStream({
|
|
123
|
-
model: "meta-llama/Llama-3.1-8B-Instruct",
|
|
124
|
-
messages: [
|
|
125
|
-
{
|
|
126
|
-
role: "user",
|
|
127
|
-
content: "What is the capital of France?"
|
|
128
|
-
}
|
|
129
|
-
],
|
|
130
|
-
max_tokens: 500
|
|
131
|
-
});
|
|
132
|
-
|
|
133
|
-
for await (const chunk of stream) {
|
|
134
|
-
if (chunk.choices && chunk.choices.length > 0) {
|
|
135
|
-
const newContent = chunk.choices[0].delta.content;
|
|
136
|
-
out += newContent;
|
|
137
|
-
console.log(newContent);
|
|
138
|
-
}
|
|
139
|
-
}`);
|
|
140
|
-
});
|
|
141
|
-
});
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"file":"python.spec.d.ts","sourceRoot":"","sources":["../../../src/snippets/python.spec.ts"],"names":[],"mappings":""}
|