@huggingface/tasks 0.12.8 → 0.12.10
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/index.cjs +125 -2
- package/dist/index.js +125 -2
- package/dist/src/model-libraries-snippets.d.ts +1 -0
- package/dist/src/model-libraries-snippets.d.ts.map +1 -1
- package/dist/src/model-libraries.d.ts +16 -2
- package/dist/src/model-libraries.d.ts.map +1 -1
- package/dist/src/tasks/automatic-speech-recognition/inference.d.ts +2 -2
- package/dist/src/tasks/chat-completion/inference.d.ts +58 -21
- package/dist/src/tasks/chat-completion/inference.d.ts.map +1 -1
- package/dist/src/tasks/image-to-text/inference.d.ts +2 -2
- package/dist/src/tasks/index.d.ts +1 -1
- package/dist/src/tasks/index.d.ts.map +1 -1
- package/dist/src/tasks/text-generation/inference.d.ts +62 -0
- package/dist/src/tasks/text-generation/inference.d.ts.map +1 -1
- package/dist/src/tasks/text-to-audio/inference.d.ts +2 -2
- package/dist/src/tasks/text-to-speech/inference.d.ts +6 -8
- package/dist/src/tasks/text-to-speech/inference.d.ts.map +1 -1
- package/package.json +1 -1
- package/src/model-libraries-snippets.ts +6 -0
- package/src/model-libraries.ts +14 -0
- package/src/tasks/automatic-speech-recognition/inference.ts +2 -2
- package/src/tasks/chat-completion/inference.ts +66 -21
- package/src/tasks/chat-completion/spec/input.json +163 -40
- package/src/tasks/chat-completion/spec/output.json +28 -18
- package/src/tasks/chat-completion/spec/stream_output.json +57 -14
- package/src/tasks/common-definitions.json +2 -2
- package/src/tasks/image-to-text/inference.ts +2 -2
- package/src/tasks/index.ts +5 -3
- package/src/tasks/text-generation/inference.ts +62 -0
- package/src/tasks/text-generation/spec/input.json +24 -0
- package/src/tasks/text-generation/spec/stream_output.json +7 -1
- package/src/tasks/text-to-audio/inference.ts +2 -2
- package/src/tasks/text-to-speech/inference.ts +6 -8
- package/src/tasks/text-to-speech/spec/input.json +26 -2
|
@@ -17,23 +17,84 @@ export interface TextGenerationInput {
|
|
|
17
17
|
[property: string]: unknown;
|
|
18
18
|
}
|
|
19
19
|
export interface TextGenerationInputGenerateParameters {
|
|
20
|
+
/**
|
|
21
|
+
* Lora adapter id
|
|
22
|
+
*/
|
|
23
|
+
adapter_id?: string;
|
|
24
|
+
/**
|
|
25
|
+
* Generate best_of sequences and return the one if the highest token logprobs.
|
|
26
|
+
*/
|
|
20
27
|
best_of?: number;
|
|
28
|
+
/**
|
|
29
|
+
* Whether to return decoder input token logprobs and ids.
|
|
30
|
+
*/
|
|
21
31
|
decoder_input_details?: boolean;
|
|
32
|
+
/**
|
|
33
|
+
* Whether to return generation details.
|
|
34
|
+
*/
|
|
22
35
|
details?: boolean;
|
|
36
|
+
/**
|
|
37
|
+
* Activate logits sampling.
|
|
38
|
+
*/
|
|
23
39
|
do_sample?: boolean;
|
|
40
|
+
/**
|
|
41
|
+
* The parameter for frequency penalty. 1.0 means no penalty
|
|
42
|
+
* Penalize new tokens based on their existing frequency in the text so far,
|
|
43
|
+
* decreasing the model's likelihood to repeat the same line verbatim.
|
|
44
|
+
*/
|
|
24
45
|
frequency_penalty?: number;
|
|
25
46
|
grammar?: TextGenerationInputGrammarType;
|
|
47
|
+
/**
|
|
48
|
+
* Maximum number of tokens to generate.
|
|
49
|
+
*/
|
|
26
50
|
max_new_tokens?: number;
|
|
51
|
+
/**
|
|
52
|
+
* The parameter for repetition penalty. 1.0 means no penalty.
|
|
53
|
+
* See [this paper](https://arxiv.org/pdf/1909.05858.pdf) for more details.
|
|
54
|
+
*/
|
|
27
55
|
repetition_penalty?: number;
|
|
56
|
+
/**
|
|
57
|
+
* Whether to prepend the prompt to the generated text
|
|
58
|
+
*/
|
|
28
59
|
return_full_text?: boolean;
|
|
60
|
+
/**
|
|
61
|
+
* Random sampling seed.
|
|
62
|
+
*/
|
|
29
63
|
seed?: number;
|
|
64
|
+
/**
|
|
65
|
+
* Stop generating tokens if a member of `stop` is generated.
|
|
66
|
+
*/
|
|
30
67
|
stop?: string[];
|
|
68
|
+
/**
|
|
69
|
+
* The value used to module the logits distribution.
|
|
70
|
+
*/
|
|
31
71
|
temperature?: number;
|
|
72
|
+
/**
|
|
73
|
+
* The number of highest probability vocabulary tokens to keep for top-k-filtering.
|
|
74
|
+
*/
|
|
32
75
|
top_k?: number;
|
|
76
|
+
/**
|
|
77
|
+
* The number of highest probability vocabulary tokens to keep for top-n-filtering.
|
|
78
|
+
*/
|
|
33
79
|
top_n_tokens?: number;
|
|
80
|
+
/**
|
|
81
|
+
* Top-p value for nucleus sampling.
|
|
82
|
+
*/
|
|
34
83
|
top_p?: number;
|
|
84
|
+
/**
|
|
85
|
+
* Truncate inputs tokens to the given size.
|
|
86
|
+
*/
|
|
35
87
|
truncate?: number;
|
|
88
|
+
/**
|
|
89
|
+
* Typical Decoding mass
|
|
90
|
+
* See [Typical Decoding for Natural Language Generation](https://arxiv.org/abs/2202.00666)
|
|
91
|
+
* for more information.
|
|
92
|
+
*/
|
|
36
93
|
typical_p?: number;
|
|
94
|
+
/**
|
|
95
|
+
* Watermarking with [A Watermark for Large Language
|
|
96
|
+
* Models](https://arxiv.org/abs/2301.10226).
|
|
97
|
+
*/
|
|
37
98
|
watermark?: boolean;
|
|
38
99
|
[property: string]: unknown;
|
|
39
100
|
}
|
|
@@ -113,6 +174,7 @@ export interface TextGenerationStreamOutput {
|
|
|
113
174
|
export interface TextGenerationStreamOutputStreamDetails {
|
|
114
175
|
finish_reason: TextGenerationOutputFinishReason;
|
|
115
176
|
generated_tokens: number;
|
|
177
|
+
input_length: number;
|
|
116
178
|
seed?: number;
|
|
117
179
|
[property: string]: unknown;
|
|
118
180
|
}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"inference.d.ts","sourceRoot":"","sources":["../../../../src/tasks/text-generation/inference.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AAEH;;;;;;GAMG;AACH,MAAM,WAAW,mBAAmB;IACnC,MAAM,EAAE,MAAM,CAAC;IACf,UAAU,CAAC,EAAE,qCAAqC,CAAC;IACnD,MAAM,CAAC,EAAE,OAAO,CAAC;IACjB,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,WAAW,qCAAqC;IACrD,OAAO,CAAC,EAAE,MAAM,CAAC;IACjB,qBAAqB,CAAC,EAAE,OAAO,CAAC;IAChC,OAAO,CAAC,EAAE,OAAO,CAAC;IAClB,SAAS,CAAC,EAAE,OAAO,CAAC;IACpB,iBAAiB,CAAC,EAAE,MAAM,CAAC;IAC3B,OAAO,CAAC,EAAE,8BAA8B,CAAC;IACzC,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB,kBAAkB,CAAC,EAAE,MAAM,CAAC;IAC5B,gBAAgB,CAAC,EAAE,OAAO,CAAC;IAC3B,IAAI,CAAC,EAAE,MAAM,CAAC;IACd,IAAI,CAAC,EAAE,MAAM,EAAE,CAAC;IAChB,WAAW,CAAC,EAAE,MAAM,CAAC;IACrB,KAAK,CAAC,EAAE,MAAM,CAAC;IACf,YAAY,CAAC,EAAE,MAAM,CAAC;IACtB,KAAK,CAAC,EAAE,MAAM,CAAC;IACf,QAAQ,CAAC,EAAE,MAAM,CAAC;IAClB,SAAS,CAAC,EAAE,MAAM,CAAC;IACnB,SAAS,CAAC,EAAE,OAAO,CAAC;IACpB,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,WAAW,8BAA8B;IAC9C,IAAI,EAAE,IAAI,CAAC;IACX;;;;;OAKG;IACH,KAAK,EAAE,OAAO,CAAC;IACf,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,MAAM,IAAI,GAAG,MAAM,GAAG,OAAO,CAAC;AAEpC;;;;;;GAMG;AACH,MAAM,WAAW,oBAAoB;IACpC,OAAO,CAAC,EAAE,2BAA2B,CAAC;IACtC,cAAc,EAAE,MAAM,CAAC;IACvB,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,WAAW,2BAA2B;IAC3C,iBAAiB,CAAC,EAAE,kCAAkC,EAAE,CAAC;IACzD,aAAa,EAAE,gCAAgC,CAAC;IAChD,gBAAgB,EAAE,MAAM,CAAC;IACzB,OAAO,EAAE,gCAAgC,EAAE,CAAC;IAC5C,IAAI,CAAC,EAAE,MAAM,CAAC;IACd,MAAM,EAAE,yBAAyB,EAAE,CAAC;IACpC,UAAU,CAAC,EAAE,KAAK,CAAC,yBAAyB,EAAE,CAAC,CAAC;IAChD,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,WAAW,kCAAkC;IAClD,aAAa,EAAE,gCAAgC,CAAC;IAChD,cAAc,EAAE,MAAM,CAAC;IACvB,gBAAgB,EAAE,MAAM,CAAC;IACzB,OAAO,EAAE,gCAAgC,EAAE,CAAC;IAC5C,IAAI,CAAC,EAAE,MAAM,CAAC;IACd,MAAM,EAAE,yBAAyB,EAAE,CAAC;IACpC,UAAU,CAAC,EAAE,KAAK,CAAC,yBAAyB,EAAE,CAAC,CAAC;IAChD,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,MAAM,gCAAgC,GAAG,QAAQ,GAAG,WAAW,GAAG,eAAe,CAAC;AAExF,MAAM,WAAW,gCAAgC;IAChD,EAAE,EAAE,MAAM,CAAC;IACX,OAAO,EAAE,MAAM,CAAC;IAChB,IAAI,EAAE,MAAM,CAAC;IACb,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,WAAW,yBAAyB;IACzC,EAAE,EAAE,MAAM,CAAC;IACX,OAAO,EAAE,MAAM,CAAC;IAChB,OAAO,EAAE,OAAO,CAAC;IACjB,IAAI,EAAE,MAAM,CAAC;IACb,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED;;;;;;GAMG;AACH,MAAM,WAAW,0BAA0B;IAC1C,OAAO,CAAC,EAAE,uCAAuC,CAAC;IAClD,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB,KAAK,EAAE,MAAM,CAAC;IACd,KAAK,EAAE,+BAA+B,CAAC;IACvC,UAAU,CAAC,EAAE,+BAA+B,EAAE,CAAC;IAC/C,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,WAAW,uCAAuC;IACvD,aAAa,EAAE,gCAAgC,CAAC;IAChD,gBAAgB,EAAE,MAAM,CAAC;IACzB,IAAI,CAAC,EAAE,MAAM,CAAC;IACd,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,WAAW,+BAA+B;IAC/C,EAAE,EAAE,MAAM,CAAC;IACX,OAAO,EAAE,MAAM,CAAC;IAChB,OAAO,EAAE,OAAO,CAAC;IACjB,IAAI,EAAE,MAAM,CAAC;IACb,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B"}
|
|
1
|
+
{"version":3,"file":"inference.d.ts","sourceRoot":"","sources":["../../../../src/tasks/text-generation/inference.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AAEH;;;;;;GAMG;AACH,MAAM,WAAW,mBAAmB;IACnC,MAAM,EAAE,MAAM,CAAC;IACf,UAAU,CAAC,EAAE,qCAAqC,CAAC;IACnD,MAAM,CAAC,EAAE,OAAO,CAAC;IACjB,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,WAAW,qCAAqC;IACrD;;OAEG;IACH,UAAU,CAAC,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,OAAO,CAAC,EAAE,MAAM,CAAC;IACjB;;OAEG;IACH,qBAAqB,CAAC,EAAE,OAAO,CAAC;IAChC;;OAEG;IACH,OAAO,CAAC,EAAE,OAAO,CAAC;IAClB;;OAEG;IACH,SAAS,CAAC,EAAE,OAAO,CAAC;IACpB;;;;OAIG;IACH,iBAAiB,CAAC,EAAE,MAAM,CAAC;IAC3B,OAAO,CAAC,EAAE,8BAA8B,CAAC;IACzC;;OAEG;IACH,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB;;;OAGG;IACH,kBAAkB,CAAC,EAAE,MAAM,CAAC;IAC5B;;OAEG;IACH,gBAAgB,CAAC,EAAE,OAAO,CAAC;IAC3B;;OAEG;IACH,IAAI,CAAC,EAAE,MAAM,CAAC;IACd;;OAEG;IACH,IAAI,CAAC,EAAE,MAAM,EAAE,CAAC;IAChB;;OAEG;IACH,WAAW,CAAC,EAAE,MAAM,CAAC;IACrB;;OAEG;IACH,KAAK,CAAC,EAAE,MAAM,CAAC;IACf;;OAEG;IACH,YAAY,CAAC,EAAE,MAAM,CAAC;IACtB;;OAEG;IACH,KAAK,CAAC,EAAE,MAAM,CAAC;IACf;;OAEG;IACH,QAAQ,CAAC,EAAE,MAAM,CAAC;IAClB;;;;OAIG;IACH,SAAS,CAAC,EAAE,MAAM,CAAC;IACnB;;;OAGG;IACH,SAAS,CAAC,EAAE,OAAO,CAAC;IACpB,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,WAAW,8BAA8B;IAC9C,IAAI,EAAE,IAAI,CAAC;IACX;;;;;OAKG;IACH,KAAK,EAAE,OAAO,CAAC;IACf,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,MAAM,IAAI,GAAG,MAAM,GAAG,OAAO,CAAC;AAEpC;;;;;;GAMG;AACH,MAAM,WAAW,oBAAoB;IACpC,OAAO,CAAC,EAAE,2BAA2B,CAAC;IACtC,cAAc,EAAE,MAAM,CAAC;IACvB,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,WAAW,2BAA2B;IAC3C,iBAAiB,CAAC,EAAE,kCAAkC,EAAE,CAAC;IACzD,aAAa,EAAE,gCAAgC,CAAC;IAChD,gBAAgB,EAAE,MAAM,CAAC;IACzB,OAAO,EAAE,gCAAgC,EAAE,CAAC;IAC5C,IAAI,CAAC,EAAE,MAAM,CAAC;IACd,MAAM,EAAE,yBAAyB,EAAE,CAAC;IACpC,UAAU,CAAC,EAAE,KAAK,CAAC,yBAAyB,EAAE,CAAC,CAAC;IAChD,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,WAAW,kCAAkC;IAClD,aAAa,EAAE,gCAAgC,CAAC;IAChD,cAAc,EAAE,MAAM,CAAC;IACvB,gBAAgB,EAAE,MAAM,CAAC;IACzB,OAAO,EAAE,gCAAgC,EAAE,CAAC;IAC5C,IAAI,CAAC,EAAE,MAAM,CAAC;IACd,MAAM,EAAE,yBAAyB,EAAE,CAAC;IACpC,UAAU,CAAC,EAAE,KAAK,CAAC,yBAAyB,EAAE,CAAC,CAAC;IAChD,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,MAAM,gCAAgC,GAAG,QAAQ,GAAG,WAAW,GAAG,eAAe,CAAC;AAExF,MAAM,WAAW,gCAAgC;IAChD,EAAE,EAAE,MAAM,CAAC;IACX,OAAO,EAAE,MAAM,CAAC;IAChB,IAAI,EAAE,MAAM,CAAC;IACb,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,WAAW,yBAAyB;IACzC,EAAE,EAAE,MAAM,CAAC;IACX,OAAO,EAAE,MAAM,CAAC;IAChB,OAAO,EAAE,OAAO,CAAC;IACjB,IAAI,EAAE,MAAM,CAAC;IACb,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED;;;;;;GAMG;AACH,MAAM,WAAW,0BAA0B;IAC1C,OAAO,CAAC,EAAE,uCAAuC,CAAC;IAClD,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB,KAAK,EAAE,MAAM,CAAC;IACd,KAAK,EAAE,+BAA+B,CAAC;IACvC,UAAU,CAAC,EAAE,+BAA+B,EAAE,CAAC;IAC/C,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,WAAW,uCAAuC;IACvD,aAAa,EAAE,gCAAgC,CAAC;IAChD,gBAAgB,EAAE,MAAM,CAAC;IACzB,YAAY,EAAE,MAAM,CAAC;IACrB,IAAI,CAAC,EAAE,MAAM,CAAC;IACd,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,WAAW,+BAA+B;IAC/C,EAAE,EAAE,MAAM,CAAC;IACX,OAAO,EAAE,MAAM,CAAC;IAChB,OAAO,EAAE,OAAO,CAAC;IACjB,IAAI,EAAE,MAAM,CAAC;IACb,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B"}
|
|
@@ -65,7 +65,7 @@ export interface GenerationParameters {
|
|
|
65
65
|
*/
|
|
66
66
|
max_length?: number;
|
|
67
67
|
/**
|
|
68
|
-
* The maximum number of tokens to generate. Takes precedence over
|
|
68
|
+
* The maximum number of tokens to generate. Takes precedence over max_length.
|
|
69
69
|
*/
|
|
70
70
|
max_new_tokens?: number;
|
|
71
71
|
/**
|
|
@@ -73,7 +73,7 @@ export interface GenerationParameters {
|
|
|
73
73
|
*/
|
|
74
74
|
min_length?: number;
|
|
75
75
|
/**
|
|
76
|
-
* The minimum number of tokens to generate. Takes precedence over
|
|
76
|
+
* The minimum number of tokens to generate. Takes precedence over min_length.
|
|
77
77
|
*/
|
|
78
78
|
min_new_tokens?: number;
|
|
79
79
|
/**
|
|
@@ -4,9 +4,7 @@
|
|
|
4
4
|
* Using src/scripts/inference-codegen
|
|
5
5
|
*/
|
|
6
6
|
/**
|
|
7
|
-
* Inputs for Text
|
|
8
|
-
*
|
|
9
|
-
* Inputs for Text To Audio inference
|
|
7
|
+
* Inputs for Text To Speech inference
|
|
10
8
|
*/
|
|
11
9
|
export interface TextToSpeechInput {
|
|
12
10
|
/**
|
|
@@ -16,15 +14,15 @@ export interface TextToSpeechInput {
|
|
|
16
14
|
/**
|
|
17
15
|
* Additional inference parameters
|
|
18
16
|
*/
|
|
19
|
-
parameters?:
|
|
17
|
+
parameters?: TextToSpeechParameters;
|
|
20
18
|
[property: string]: unknown;
|
|
21
19
|
}
|
|
22
20
|
/**
|
|
23
21
|
* Additional inference parameters
|
|
24
22
|
*
|
|
25
|
-
* Additional inference parameters for Text To
|
|
23
|
+
* Additional inference parameters for Text To Speech
|
|
26
24
|
*/
|
|
27
|
-
export interface
|
|
25
|
+
export interface TextToSpeechParameters {
|
|
28
26
|
/**
|
|
29
27
|
* Parametrization of the text generation process
|
|
30
28
|
*/
|
|
@@ -67,7 +65,7 @@ export interface GenerationParameters {
|
|
|
67
65
|
*/
|
|
68
66
|
max_length?: number;
|
|
69
67
|
/**
|
|
70
|
-
* The maximum number of tokens to generate. Takes precedence over
|
|
68
|
+
* The maximum number of tokens to generate. Takes precedence over max_length.
|
|
71
69
|
*/
|
|
72
70
|
max_new_tokens?: number;
|
|
73
71
|
/**
|
|
@@ -75,7 +73,7 @@ export interface GenerationParameters {
|
|
|
75
73
|
*/
|
|
76
74
|
min_length?: number;
|
|
77
75
|
/**
|
|
78
|
-
* The minimum number of tokens to generate. Takes precedence over
|
|
76
|
+
* The minimum number of tokens to generate. Takes precedence over min_length.
|
|
79
77
|
*/
|
|
80
78
|
min_new_tokens?: number;
|
|
81
79
|
/**
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"inference.d.ts","sourceRoot":"","sources":["../../../../src/tasks/text-to-speech/inference.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AAEH
|
|
1
|
+
{"version":3,"file":"inference.d.ts","sourceRoot":"","sources":["../../../../src/tasks/text-to-speech/inference.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AAEH;;GAEG;AACH,MAAM,WAAW,iBAAiB;IACjC;;OAEG;IACH,MAAM,EAAE,MAAM,CAAC;IACf;;OAEG;IACH,UAAU,CAAC,EAAE,sBAAsB,CAAC;IACpC,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED;;;;GAIG;AACH,MAAM,WAAW,sBAAsB;IACtC;;OAEG;IACH,QAAQ,CAAC,EAAE,oBAAoB,CAAC;IAChC,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED;;;;GAIG;AACH,MAAM,WAAW,oBAAoB;IACpC;;OAEG;IACH,SAAS,CAAC,EAAE,OAAO,CAAC;IACpB;;OAEG;IACH,cAAc,CAAC,EAAE,kBAAkB,CAAC;IACpC;;;;;OAKG;IACH,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB;;;;;;;;OAQG;IACH,UAAU,CAAC,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,UAAU,CAAC,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB;;OAEG;IACH,UAAU,CAAC,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB;;;OAGG;IACH,eAAe,CAAC,EAAE,MAAM,CAAC;IACzB;;OAEG;IACH,SAAS,CAAC,EAAE,MAAM,CAAC;IACnB;;;OAGG;IACH,aAAa,CAAC,EAAE,MAAM,CAAC;IACvB;;OAEG;IACH,WAAW,CAAC,EAAE,MAAM,CAAC;IACrB;;OAEG;IACH,KAAK,CAAC,EAAE,MAAM,CAAC;IACf;;;OAGG;IACH,KAAK,CAAC,EAAE,MAAM,CAAC;IACf;;;;;;OAMG;IACH,SAAS,CAAC,EAAE,MAAM,CAAC;IACnB;;OAEG;IACH,SAAS,CAAC,EAAE,OAAO,CAAC;IACpB,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED;;GAEG;AACH,MAAM,MAAM,kBAAkB,GAAG,OAAO,GAAG,OAAO,CAAC;AAEnD;;;;GAIG;AACH,MAAM,WAAW,kBAAkB;IAClC;;OAEG;IACH,KAAK,EAAE,OAAO,CAAC;IACf,YAAY,EAAE,OAAO,CAAC;IACtB;;OAEG;IACH,aAAa,CAAC,EAAE,MAAM,CAAC;IACvB,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B"}
|
package/package.json
CHANGED
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@huggingface/tasks",
|
|
3
3
|
"packageManager": "pnpm@8.10.5",
|
|
4
|
-
"version": "0.12.
|
|
4
|
+
"version": "0.12.10",
|
|
5
5
|
"description": "List of ML tasks for huggingface.co/tasks",
|
|
6
6
|
"repository": "https://github.com/huggingface/huggingface.js.git",
|
|
7
7
|
"publishConfig": {
|
|
@@ -951,6 +951,12 @@ export const mlxim = (model: ModelData): string[] => [
|
|
|
951
951
|
model = create_model(${model.id})`,
|
|
952
952
|
];
|
|
953
953
|
|
|
954
|
+
export const model2vec = (model: ModelData): string[] => [
|
|
955
|
+
`from model2vec import StaticModel
|
|
956
|
+
|
|
957
|
+
model = StaticModel.from_pretrained("${model.id}")`,
|
|
958
|
+
];
|
|
959
|
+
|
|
954
960
|
export const nemo = (model: ModelData): string[] => {
|
|
955
961
|
let command: string[] | undefined = undefined;
|
|
956
962
|
// Resolve the tag to a nemo domain/sub-domain
|
package/src/model-libraries.ts
CHANGED
|
@@ -406,6 +406,20 @@ export const MODEL_LIBRARIES_UI_ELEMENTS = {
|
|
|
406
406
|
filter: false,
|
|
407
407
|
countDownloads: `path:"mlc-chat-config.json"`,
|
|
408
408
|
},
|
|
409
|
+
model2vec: {
|
|
410
|
+
prettyLabel: "Model2Vec",
|
|
411
|
+
repoName: "model2vec",
|
|
412
|
+
repoUrl: "https://github.com/MinishLab/model2vec",
|
|
413
|
+
snippets: snippets.model2vec,
|
|
414
|
+
filter: false,
|
|
415
|
+
},
|
|
416
|
+
moshi: {
|
|
417
|
+
prettyLabel: "Moshi",
|
|
418
|
+
repoName: "Moshi",
|
|
419
|
+
repoUrl: "https://github.com/kyutai-labs/moshi",
|
|
420
|
+
filter: false,
|
|
421
|
+
countDownloads: `path:"tokenizer-e351c8d8-checkpoint125.safetensors"`,
|
|
422
|
+
},
|
|
409
423
|
nemo: {
|
|
410
424
|
prettyLabel: "NeMo",
|
|
411
425
|
repoName: "NeMo",
|
|
@@ -73,7 +73,7 @@ export interface GenerationParameters {
|
|
|
73
73
|
*/
|
|
74
74
|
max_length?: number;
|
|
75
75
|
/**
|
|
76
|
-
* The maximum number of tokens to generate. Takes precedence over
|
|
76
|
+
* The maximum number of tokens to generate. Takes precedence over max_length.
|
|
77
77
|
*/
|
|
78
78
|
max_new_tokens?: number;
|
|
79
79
|
/**
|
|
@@ -81,7 +81,7 @@ export interface GenerationParameters {
|
|
|
81
81
|
*/
|
|
82
82
|
min_length?: number;
|
|
83
83
|
/**
|
|
84
|
-
* The minimum number of tokens to generate. Takes precedence over
|
|
84
|
+
* The minimum number of tokens to generate. Takes precedence over min_length.
|
|
85
85
|
*/
|
|
86
86
|
min_new_tokens?: number;
|
|
87
87
|
/**
|
|
@@ -49,7 +49,7 @@ export interface ChatCompletionInput {
|
|
|
49
49
|
* [UNUSED] ID of the model to use. See the model endpoint compatibility table for details
|
|
50
50
|
* on which models work with the Chat API.
|
|
51
51
|
*/
|
|
52
|
-
model
|
|
52
|
+
model?: string;
|
|
53
53
|
/**
|
|
54
54
|
* UNUSED
|
|
55
55
|
* How many chat completion choices to generate for each input message. Note that you will
|
|
@@ -63,12 +63,14 @@ export interface ChatCompletionInput {
|
|
|
63
63
|
* increasing the model's likelihood to talk about new topics
|
|
64
64
|
*/
|
|
65
65
|
presence_penalty?: number;
|
|
66
|
+
response_format?: ChatCompletionInputGrammarType;
|
|
66
67
|
seed?: number;
|
|
67
68
|
/**
|
|
68
69
|
* Up to 4 sequences where the API will stop generating further tokens.
|
|
69
70
|
*/
|
|
70
71
|
stop?: string[];
|
|
71
72
|
stream?: boolean;
|
|
73
|
+
stream_options?: ChatCompletionInputStreamOptions;
|
|
72
74
|
/**
|
|
73
75
|
* What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the
|
|
74
76
|
* output more random, while
|
|
@@ -77,7 +79,7 @@ export interface ChatCompletionInput {
|
|
|
77
79
|
* We generally recommend altering this or `top_p` but not both.
|
|
78
80
|
*/
|
|
79
81
|
temperature?: number;
|
|
80
|
-
tool_choice?:
|
|
82
|
+
tool_choice?: ChatCompletionInputTool;
|
|
81
83
|
/**
|
|
82
84
|
* A prompt to be appended before the tools
|
|
83
85
|
*/
|
|
@@ -87,7 +89,7 @@ export interface ChatCompletionInput {
|
|
|
87
89
|
* Use this to provide a list of
|
|
88
90
|
* functions the model may generate JSON inputs for.
|
|
89
91
|
*/
|
|
90
|
-
tools?:
|
|
92
|
+
tools?: ToolElement[];
|
|
91
93
|
/**
|
|
92
94
|
* An integer between 0 and 5 specifying the number of most likely tokens to return at each
|
|
93
95
|
* token position, each with
|
|
@@ -105,40 +107,78 @@ export interface ChatCompletionInput {
|
|
|
105
107
|
}
|
|
106
108
|
|
|
107
109
|
export interface ChatCompletionInputMessage {
|
|
108
|
-
content
|
|
110
|
+
content: ChatCompletionInputMessageContent;
|
|
109
111
|
name?: string;
|
|
110
112
|
role: string;
|
|
111
|
-
tool_calls?: ChatCompletionInputToolCall[];
|
|
112
113
|
[property: string]: unknown;
|
|
113
114
|
}
|
|
114
115
|
|
|
115
|
-
export
|
|
116
|
-
|
|
117
|
-
|
|
118
|
-
|
|
116
|
+
export type ChatCompletionInputMessageContent = ChatCompletionInputMessageChunk[] | string;
|
|
117
|
+
|
|
118
|
+
export interface ChatCompletionInputMessageChunk {
|
|
119
|
+
image_url?: ChatCompletionInputURL;
|
|
120
|
+
text?: string;
|
|
121
|
+
type: ChatCompletionInputMessageChunkType;
|
|
119
122
|
[property: string]: unknown;
|
|
120
123
|
}
|
|
121
124
|
|
|
122
|
-
export interface
|
|
123
|
-
|
|
124
|
-
description?: string;
|
|
125
|
-
name: string;
|
|
125
|
+
export interface ChatCompletionInputURL {
|
|
126
|
+
url: string;
|
|
126
127
|
[property: string]: unknown;
|
|
127
128
|
}
|
|
128
129
|
|
|
129
|
-
export type
|
|
130
|
+
export type ChatCompletionInputMessageChunkType = "text" | "image_url";
|
|
130
131
|
|
|
131
|
-
export interface
|
|
132
|
-
|
|
132
|
+
export interface ChatCompletionInputGrammarType {
|
|
133
|
+
type: ChatCompletionInputGrammarTypeType;
|
|
134
|
+
/**
|
|
135
|
+
* A string that represents a [JSON Schema](https://json-schema.org/).
|
|
136
|
+
*
|
|
137
|
+
* JSON Schema is a declarative language that allows to annotate JSON documents
|
|
138
|
+
* with types and descriptions.
|
|
139
|
+
*/
|
|
140
|
+
value: unknown;
|
|
141
|
+
[property: string]: unknown;
|
|
142
|
+
}
|
|
143
|
+
|
|
144
|
+
export type ChatCompletionInputGrammarTypeType = "json" | "regex";
|
|
145
|
+
|
|
146
|
+
export interface ChatCompletionInputStreamOptions {
|
|
147
|
+
/**
|
|
148
|
+
* If set, an additional chunk will be streamed before the data: [DONE] message. The usage
|
|
149
|
+
* field on this chunk shows the token usage statistics for the entire request, and the
|
|
150
|
+
* choices field will always be an empty array. All other chunks will also include a usage
|
|
151
|
+
* field, but with a null value.
|
|
152
|
+
*/
|
|
153
|
+
include_usage: boolean;
|
|
133
154
|
[property: string]: unknown;
|
|
134
155
|
}
|
|
135
156
|
|
|
136
|
-
export
|
|
157
|
+
export type ChatCompletionInputTool = ChatCompletionInputToolType | string;
|
|
158
|
+
|
|
159
|
+
export interface ChatCompletionInputToolType {
|
|
160
|
+
function?: ChatCompletionInputFunctionName;
|
|
161
|
+
[property: string]: unknown;
|
|
162
|
+
}
|
|
163
|
+
|
|
164
|
+
export interface ChatCompletionInputFunctionName {
|
|
165
|
+
name: string;
|
|
166
|
+
[property: string]: unknown;
|
|
167
|
+
}
|
|
168
|
+
|
|
169
|
+
export interface ToolElement {
|
|
137
170
|
function: ChatCompletionInputFunctionDefinition;
|
|
138
171
|
type: string;
|
|
139
172
|
[property: string]: unknown;
|
|
140
173
|
}
|
|
141
174
|
|
|
175
|
+
export interface ChatCompletionInputFunctionDefinition {
|
|
176
|
+
arguments: unknown;
|
|
177
|
+
description?: string;
|
|
178
|
+
name: string;
|
|
179
|
+
[property: string]: unknown;
|
|
180
|
+
}
|
|
181
|
+
|
|
142
182
|
/**
|
|
143
183
|
* Chat Completion Output.
|
|
144
184
|
*
|
|
@@ -151,7 +191,6 @@ export interface ChatCompletionOutput {
|
|
|
151
191
|
created: number;
|
|
152
192
|
id: string;
|
|
153
193
|
model: string;
|
|
154
|
-
object: string;
|
|
155
194
|
system_fingerprint: string;
|
|
156
195
|
usage: ChatCompletionOutputUsage;
|
|
157
196
|
[property: string]: unknown;
|
|
@@ -185,7 +224,6 @@ export interface ChatCompletionOutputTopLogprob {
|
|
|
185
224
|
|
|
186
225
|
export interface ChatCompletionOutputMessage {
|
|
187
226
|
content?: string;
|
|
188
|
-
name?: string;
|
|
189
227
|
role: string;
|
|
190
228
|
tool_calls?: ChatCompletionOutputToolCall[];
|
|
191
229
|
[property: string]: unknown;
|
|
@@ -193,7 +231,7 @@ export interface ChatCompletionOutputMessage {
|
|
|
193
231
|
|
|
194
232
|
export interface ChatCompletionOutputToolCall {
|
|
195
233
|
function: ChatCompletionOutputFunctionDefinition;
|
|
196
|
-
id:
|
|
234
|
+
id: string;
|
|
197
235
|
type: string;
|
|
198
236
|
[property: string]: unknown;
|
|
199
237
|
}
|
|
@@ -224,8 +262,8 @@ export interface ChatCompletionStreamOutput {
|
|
|
224
262
|
created: number;
|
|
225
263
|
id: string;
|
|
226
264
|
model: string;
|
|
227
|
-
object: string;
|
|
228
265
|
system_fingerprint: string;
|
|
266
|
+
usage?: ChatCompletionStreamOutputUsage;
|
|
229
267
|
[property: string]: unknown;
|
|
230
268
|
}
|
|
231
269
|
|
|
@@ -275,3 +313,10 @@ export interface ChatCompletionStreamOutputTopLogprob {
|
|
|
275
313
|
token: string;
|
|
276
314
|
[property: string]: unknown;
|
|
277
315
|
}
|
|
316
|
+
|
|
317
|
+
export interface ChatCompletionStreamOutputUsage {
|
|
318
|
+
completion_tokens: number;
|
|
319
|
+
prompt_tokens: number;
|
|
320
|
+
total_tokens: number;
|
|
321
|
+
[property: string]: unknown;
|
|
322
|
+
}
|