@huggingface/tasks 0.12.8 → 0.12.10

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (34) hide show
  1. package/dist/index.cjs +125 -2
  2. package/dist/index.js +125 -2
  3. package/dist/src/model-libraries-snippets.d.ts +1 -0
  4. package/dist/src/model-libraries-snippets.d.ts.map +1 -1
  5. package/dist/src/model-libraries.d.ts +16 -2
  6. package/dist/src/model-libraries.d.ts.map +1 -1
  7. package/dist/src/tasks/automatic-speech-recognition/inference.d.ts +2 -2
  8. package/dist/src/tasks/chat-completion/inference.d.ts +58 -21
  9. package/dist/src/tasks/chat-completion/inference.d.ts.map +1 -1
  10. package/dist/src/tasks/image-to-text/inference.d.ts +2 -2
  11. package/dist/src/tasks/index.d.ts +1 -1
  12. package/dist/src/tasks/index.d.ts.map +1 -1
  13. package/dist/src/tasks/text-generation/inference.d.ts +62 -0
  14. package/dist/src/tasks/text-generation/inference.d.ts.map +1 -1
  15. package/dist/src/tasks/text-to-audio/inference.d.ts +2 -2
  16. package/dist/src/tasks/text-to-speech/inference.d.ts +6 -8
  17. package/dist/src/tasks/text-to-speech/inference.d.ts.map +1 -1
  18. package/package.json +1 -1
  19. package/src/model-libraries-snippets.ts +6 -0
  20. package/src/model-libraries.ts +14 -0
  21. package/src/tasks/automatic-speech-recognition/inference.ts +2 -2
  22. package/src/tasks/chat-completion/inference.ts +66 -21
  23. package/src/tasks/chat-completion/spec/input.json +163 -40
  24. package/src/tasks/chat-completion/spec/output.json +28 -18
  25. package/src/tasks/chat-completion/spec/stream_output.json +57 -14
  26. package/src/tasks/common-definitions.json +2 -2
  27. package/src/tasks/image-to-text/inference.ts +2 -2
  28. package/src/tasks/index.ts +5 -3
  29. package/src/tasks/text-generation/inference.ts +62 -0
  30. package/src/tasks/text-generation/spec/input.json +24 -0
  31. package/src/tasks/text-generation/spec/stream_output.json +7 -1
  32. package/src/tasks/text-to-audio/inference.ts +2 -2
  33. package/src/tasks/text-to-speech/inference.ts +6 -8
  34. package/src/tasks/text-to-speech/spec/input.json +26 -2
@@ -17,23 +17,84 @@ export interface TextGenerationInput {
17
17
  [property: string]: unknown;
18
18
  }
19
19
  export interface TextGenerationInputGenerateParameters {
20
+ /**
21
+ * Lora adapter id
22
+ */
23
+ adapter_id?: string;
24
+ /**
25
+ * Generate best_of sequences and return the one if the highest token logprobs.
26
+ */
20
27
  best_of?: number;
28
+ /**
29
+ * Whether to return decoder input token logprobs and ids.
30
+ */
21
31
  decoder_input_details?: boolean;
32
+ /**
33
+ * Whether to return generation details.
34
+ */
22
35
  details?: boolean;
36
+ /**
37
+ * Activate logits sampling.
38
+ */
23
39
  do_sample?: boolean;
40
+ /**
41
+ * The parameter for frequency penalty. 1.0 means no penalty
42
+ * Penalize new tokens based on their existing frequency in the text so far,
43
+ * decreasing the model's likelihood to repeat the same line verbatim.
44
+ */
24
45
  frequency_penalty?: number;
25
46
  grammar?: TextGenerationInputGrammarType;
47
+ /**
48
+ * Maximum number of tokens to generate.
49
+ */
26
50
  max_new_tokens?: number;
51
+ /**
52
+ * The parameter for repetition penalty. 1.0 means no penalty.
53
+ * See [this paper](https://arxiv.org/pdf/1909.05858.pdf) for more details.
54
+ */
27
55
  repetition_penalty?: number;
56
+ /**
57
+ * Whether to prepend the prompt to the generated text
58
+ */
28
59
  return_full_text?: boolean;
60
+ /**
61
+ * Random sampling seed.
62
+ */
29
63
  seed?: number;
64
+ /**
65
+ * Stop generating tokens if a member of `stop` is generated.
66
+ */
30
67
  stop?: string[];
68
+ /**
69
+ * The value used to module the logits distribution.
70
+ */
31
71
  temperature?: number;
72
+ /**
73
+ * The number of highest probability vocabulary tokens to keep for top-k-filtering.
74
+ */
32
75
  top_k?: number;
76
+ /**
77
+ * The number of highest probability vocabulary tokens to keep for top-n-filtering.
78
+ */
33
79
  top_n_tokens?: number;
80
+ /**
81
+ * Top-p value for nucleus sampling.
82
+ */
34
83
  top_p?: number;
84
+ /**
85
+ * Truncate inputs tokens to the given size.
86
+ */
35
87
  truncate?: number;
88
+ /**
89
+ * Typical Decoding mass
90
+ * See [Typical Decoding for Natural Language Generation](https://arxiv.org/abs/2202.00666)
91
+ * for more information.
92
+ */
36
93
  typical_p?: number;
94
+ /**
95
+ * Watermarking with [A Watermark for Large Language
96
+ * Models](https://arxiv.org/abs/2301.10226).
97
+ */
37
98
  watermark?: boolean;
38
99
  [property: string]: unknown;
39
100
  }
@@ -113,6 +174,7 @@ export interface TextGenerationStreamOutput {
113
174
  export interface TextGenerationStreamOutputStreamDetails {
114
175
  finish_reason: TextGenerationOutputFinishReason;
115
176
  generated_tokens: number;
177
+ input_length: number;
116
178
  seed?: number;
117
179
  [property: string]: unknown;
118
180
  }
@@ -1 +1 @@
1
- {"version":3,"file":"inference.d.ts","sourceRoot":"","sources":["../../../../src/tasks/text-generation/inference.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AAEH;;;;;;GAMG;AACH,MAAM,WAAW,mBAAmB;IACnC,MAAM,EAAE,MAAM,CAAC;IACf,UAAU,CAAC,EAAE,qCAAqC,CAAC;IACnD,MAAM,CAAC,EAAE,OAAO,CAAC;IACjB,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,WAAW,qCAAqC;IACrD,OAAO,CAAC,EAAE,MAAM,CAAC;IACjB,qBAAqB,CAAC,EAAE,OAAO,CAAC;IAChC,OAAO,CAAC,EAAE,OAAO,CAAC;IAClB,SAAS,CAAC,EAAE,OAAO,CAAC;IACpB,iBAAiB,CAAC,EAAE,MAAM,CAAC;IAC3B,OAAO,CAAC,EAAE,8BAA8B,CAAC;IACzC,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB,kBAAkB,CAAC,EAAE,MAAM,CAAC;IAC5B,gBAAgB,CAAC,EAAE,OAAO,CAAC;IAC3B,IAAI,CAAC,EAAE,MAAM,CAAC;IACd,IAAI,CAAC,EAAE,MAAM,EAAE,CAAC;IAChB,WAAW,CAAC,EAAE,MAAM,CAAC;IACrB,KAAK,CAAC,EAAE,MAAM,CAAC;IACf,YAAY,CAAC,EAAE,MAAM,CAAC;IACtB,KAAK,CAAC,EAAE,MAAM,CAAC;IACf,QAAQ,CAAC,EAAE,MAAM,CAAC;IAClB,SAAS,CAAC,EAAE,MAAM,CAAC;IACnB,SAAS,CAAC,EAAE,OAAO,CAAC;IACpB,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,WAAW,8BAA8B;IAC9C,IAAI,EAAE,IAAI,CAAC;IACX;;;;;OAKG;IACH,KAAK,EAAE,OAAO,CAAC;IACf,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,MAAM,IAAI,GAAG,MAAM,GAAG,OAAO,CAAC;AAEpC;;;;;;GAMG;AACH,MAAM,WAAW,oBAAoB;IACpC,OAAO,CAAC,EAAE,2BAA2B,CAAC;IACtC,cAAc,EAAE,MAAM,CAAC;IACvB,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,WAAW,2BAA2B;IAC3C,iBAAiB,CAAC,EAAE,kCAAkC,EAAE,CAAC;IACzD,aAAa,EAAE,gCAAgC,CAAC;IAChD,gBAAgB,EAAE,MAAM,CAAC;IACzB,OAAO,EAAE,gCAAgC,EAAE,CAAC;IAC5C,IAAI,CAAC,EAAE,MAAM,CAAC;IACd,MAAM,EAAE,yBAAyB,EAAE,CAAC;IACpC,UAAU,CAAC,EAAE,KAAK,CAAC,yBAAyB,EAAE,CAAC,CAAC;IAChD,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,WAAW,kCAAkC;IAClD,aAAa,EAAE,gCAAgC,CAAC;IAChD,cAAc,EAAE,MAAM,CAAC;IACvB,gBAAgB,EAAE,MAAM,CAAC;IACzB,OAAO,EAAE,gCAAgC,EAAE,CAAC;IAC5C,IAAI,CAAC,EAAE,MAAM,CAAC;IACd,MAAM,EAAE,yBAAyB,EAAE,CAAC;IACpC,UAAU,CAAC,EAAE,KAAK,CAAC,yBAAyB,EAAE,CAAC,CAAC;IAChD,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,MAAM,gCAAgC,GAAG,QAAQ,GAAG,WAAW,GAAG,eAAe,CAAC;AAExF,MAAM,WAAW,gCAAgC;IAChD,EAAE,EAAE,MAAM,CAAC;IACX,OAAO,EAAE,MAAM,CAAC;IAChB,IAAI,EAAE,MAAM,CAAC;IACb,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,WAAW,yBAAyB;IACzC,EAAE,EAAE,MAAM,CAAC;IACX,OAAO,EAAE,MAAM,CAAC;IAChB,OAAO,EAAE,OAAO,CAAC;IACjB,IAAI,EAAE,MAAM,CAAC;IACb,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED;;;;;;GAMG;AACH,MAAM,WAAW,0BAA0B;IAC1C,OAAO,CAAC,EAAE,uCAAuC,CAAC;IAClD,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB,KAAK,EAAE,MAAM,CAAC;IACd,KAAK,EAAE,+BAA+B,CAAC;IACvC,UAAU,CAAC,EAAE,+BAA+B,EAAE,CAAC;IAC/C,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,WAAW,uCAAuC;IACvD,aAAa,EAAE,gCAAgC,CAAC;IAChD,gBAAgB,EAAE,MAAM,CAAC;IACzB,IAAI,CAAC,EAAE,MAAM,CAAC;IACd,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,WAAW,+BAA+B;IAC/C,EAAE,EAAE,MAAM,CAAC;IACX,OAAO,EAAE,MAAM,CAAC;IAChB,OAAO,EAAE,OAAO,CAAC;IACjB,IAAI,EAAE,MAAM,CAAC;IACb,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B"}
1
+ {"version":3,"file":"inference.d.ts","sourceRoot":"","sources":["../../../../src/tasks/text-generation/inference.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AAEH;;;;;;GAMG;AACH,MAAM,WAAW,mBAAmB;IACnC,MAAM,EAAE,MAAM,CAAC;IACf,UAAU,CAAC,EAAE,qCAAqC,CAAC;IACnD,MAAM,CAAC,EAAE,OAAO,CAAC;IACjB,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,WAAW,qCAAqC;IACrD;;OAEG;IACH,UAAU,CAAC,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,OAAO,CAAC,EAAE,MAAM,CAAC;IACjB;;OAEG;IACH,qBAAqB,CAAC,EAAE,OAAO,CAAC;IAChC;;OAEG;IACH,OAAO,CAAC,EAAE,OAAO,CAAC;IAClB;;OAEG;IACH,SAAS,CAAC,EAAE,OAAO,CAAC;IACpB;;;;OAIG;IACH,iBAAiB,CAAC,EAAE,MAAM,CAAC;IAC3B,OAAO,CAAC,EAAE,8BAA8B,CAAC;IACzC;;OAEG;IACH,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB;;;OAGG;IACH,kBAAkB,CAAC,EAAE,MAAM,CAAC;IAC5B;;OAEG;IACH,gBAAgB,CAAC,EAAE,OAAO,CAAC;IAC3B;;OAEG;IACH,IAAI,CAAC,EAAE,MAAM,CAAC;IACd;;OAEG;IACH,IAAI,CAAC,EAAE,MAAM,EAAE,CAAC;IAChB;;OAEG;IACH,WAAW,CAAC,EAAE,MAAM,CAAC;IACrB;;OAEG;IACH,KAAK,CAAC,EAAE,MAAM,CAAC;IACf;;OAEG;IACH,YAAY,CAAC,EAAE,MAAM,CAAC;IACtB;;OAEG;IACH,KAAK,CAAC,EAAE,MAAM,CAAC;IACf;;OAEG;IACH,QAAQ,CAAC,EAAE,MAAM,CAAC;IAClB;;;;OAIG;IACH,SAAS,CAAC,EAAE,MAAM,CAAC;IACnB;;;OAGG;IACH,SAAS,CAAC,EAAE,OAAO,CAAC;IACpB,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,WAAW,8BAA8B;IAC9C,IAAI,EAAE,IAAI,CAAC;IACX;;;;;OAKG;IACH,KAAK,EAAE,OAAO,CAAC;IACf,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,MAAM,IAAI,GAAG,MAAM,GAAG,OAAO,CAAC;AAEpC;;;;;;GAMG;AACH,MAAM,WAAW,oBAAoB;IACpC,OAAO,CAAC,EAAE,2BAA2B,CAAC;IACtC,cAAc,EAAE,MAAM,CAAC;IACvB,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,WAAW,2BAA2B;IAC3C,iBAAiB,CAAC,EAAE,kCAAkC,EAAE,CAAC;IACzD,aAAa,EAAE,gCAAgC,CAAC;IAChD,gBAAgB,EAAE,MAAM,CAAC;IACzB,OAAO,EAAE,gCAAgC,EAAE,CAAC;IAC5C,IAAI,CAAC,EAAE,MAAM,CAAC;IACd,MAAM,EAAE,yBAAyB,EAAE,CAAC;IACpC,UAAU,CAAC,EAAE,KAAK,CAAC,yBAAyB,EAAE,CAAC,CAAC;IAChD,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,WAAW,kCAAkC;IAClD,aAAa,EAAE,gCAAgC,CAAC;IAChD,cAAc,EAAE,MAAM,CAAC;IACvB,gBAAgB,EAAE,MAAM,CAAC;IACzB,OAAO,EAAE,gCAAgC,EAAE,CAAC;IAC5C,IAAI,CAAC,EAAE,MAAM,CAAC;IACd,MAAM,EAAE,yBAAyB,EAAE,CAAC;IACpC,UAAU,CAAC,EAAE,KAAK,CAAC,yBAAyB,EAAE,CAAC,CAAC;IAChD,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,MAAM,gCAAgC,GAAG,QAAQ,GAAG,WAAW,GAAG,eAAe,CAAC;AAExF,MAAM,WAAW,gCAAgC;IAChD,EAAE,EAAE,MAAM,CAAC;IACX,OAAO,EAAE,MAAM,CAAC;IAChB,IAAI,EAAE,MAAM,CAAC;IACb,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,WAAW,yBAAyB;IACzC,EAAE,EAAE,MAAM,CAAC;IACX,OAAO,EAAE,MAAM,CAAC;IAChB,OAAO,EAAE,OAAO,CAAC;IACjB,IAAI,EAAE,MAAM,CAAC;IACb,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED;;;;;;GAMG;AACH,MAAM,WAAW,0BAA0B;IAC1C,OAAO,CAAC,EAAE,uCAAuC,CAAC;IAClD,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB,KAAK,EAAE,MAAM,CAAC;IACd,KAAK,EAAE,+BAA+B,CAAC;IACvC,UAAU,CAAC,EAAE,+BAA+B,EAAE,CAAC;IAC/C,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,WAAW,uCAAuC;IACvD,aAAa,EAAE,gCAAgC,CAAC;IAChD,gBAAgB,EAAE,MAAM,CAAC;IACzB,YAAY,EAAE,MAAM,CAAC;IACrB,IAAI,CAAC,EAAE,MAAM,CAAC;IACd,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED,MAAM,WAAW,+BAA+B;IAC/C,EAAE,EAAE,MAAM,CAAC;IACX,OAAO,EAAE,MAAM,CAAC;IAChB,OAAO,EAAE,OAAO,CAAC;IACjB,IAAI,EAAE,MAAM,CAAC;IACb,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B"}
@@ -65,7 +65,7 @@ export interface GenerationParameters {
65
65
  */
66
66
  max_length?: number;
67
67
  /**
68
- * The maximum number of tokens to generate. Takes precedence over maxLength.
68
+ * The maximum number of tokens to generate. Takes precedence over max_length.
69
69
  */
70
70
  max_new_tokens?: number;
71
71
  /**
@@ -73,7 +73,7 @@ export interface GenerationParameters {
73
73
  */
74
74
  min_length?: number;
75
75
  /**
76
- * The minimum number of tokens to generate. Takes precedence over maxLength.
76
+ * The minimum number of tokens to generate. Takes precedence over min_length.
77
77
  */
78
78
  min_new_tokens?: number;
79
79
  /**
@@ -4,9 +4,7 @@
4
4
  * Using src/scripts/inference-codegen
5
5
  */
6
6
  /**
7
- * Inputs for Text to Speech inference
8
- *
9
- * Inputs for Text To Audio inference
7
+ * Inputs for Text To Speech inference
10
8
  */
11
9
  export interface TextToSpeechInput {
12
10
  /**
@@ -16,15 +14,15 @@ export interface TextToSpeechInput {
16
14
  /**
17
15
  * Additional inference parameters
18
16
  */
19
- parameters?: TextToAudioParameters;
17
+ parameters?: TextToSpeechParameters;
20
18
  [property: string]: unknown;
21
19
  }
22
20
  /**
23
21
  * Additional inference parameters
24
22
  *
25
- * Additional inference parameters for Text To Audio
23
+ * Additional inference parameters for Text To Speech
26
24
  */
27
- export interface TextToAudioParameters {
25
+ export interface TextToSpeechParameters {
28
26
  /**
29
27
  * Parametrization of the text generation process
30
28
  */
@@ -67,7 +65,7 @@ export interface GenerationParameters {
67
65
  */
68
66
  max_length?: number;
69
67
  /**
70
- * The maximum number of tokens to generate. Takes precedence over maxLength.
68
+ * The maximum number of tokens to generate. Takes precedence over max_length.
71
69
  */
72
70
  max_new_tokens?: number;
73
71
  /**
@@ -75,7 +73,7 @@ export interface GenerationParameters {
75
73
  */
76
74
  min_length?: number;
77
75
  /**
78
- * The minimum number of tokens to generate. Takes precedence over maxLength.
76
+ * The minimum number of tokens to generate. Takes precedence over min_length.
79
77
  */
80
78
  min_new_tokens?: number;
81
79
  /**
@@ -1 +1 @@
1
- {"version":3,"file":"inference.d.ts","sourceRoot":"","sources":["../../../../src/tasks/text-to-speech/inference.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AAEH;;;;GAIG;AACH,MAAM,WAAW,iBAAiB;IACjC;;OAEG;IACH,MAAM,EAAE,MAAM,CAAC;IACf;;OAEG;IACH,UAAU,CAAC,EAAE,qBAAqB,CAAC;IACnC,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED;;;;GAIG;AACH,MAAM,WAAW,qBAAqB;IACrC;;OAEG;IACH,QAAQ,CAAC,EAAE,oBAAoB,CAAC;IAChC,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED;;;;GAIG;AACH,MAAM,WAAW,oBAAoB;IACpC;;OAEG;IACH,SAAS,CAAC,EAAE,OAAO,CAAC;IACpB;;OAEG;IACH,cAAc,CAAC,EAAE,kBAAkB,CAAC;IACpC;;;;;OAKG;IACH,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB;;;;;;;;OAQG;IACH,UAAU,CAAC,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,UAAU,CAAC,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB;;OAEG;IACH,UAAU,CAAC,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB;;;OAGG;IACH,eAAe,CAAC,EAAE,MAAM,CAAC;IACzB;;OAEG;IACH,SAAS,CAAC,EAAE,MAAM,CAAC;IACnB;;;OAGG;IACH,aAAa,CAAC,EAAE,MAAM,CAAC;IACvB;;OAEG;IACH,WAAW,CAAC,EAAE,MAAM,CAAC;IACrB;;OAEG;IACH,KAAK,CAAC,EAAE,MAAM,CAAC;IACf;;;OAGG;IACH,KAAK,CAAC,EAAE,MAAM,CAAC;IACf;;;;;;OAMG;IACH,SAAS,CAAC,EAAE,MAAM,CAAC;IACnB;;OAEG;IACH,SAAS,CAAC,EAAE,OAAO,CAAC;IACpB,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED;;GAEG;AACH,MAAM,MAAM,kBAAkB,GAAG,OAAO,GAAG,OAAO,CAAC;AAEnD;;;;GAIG;AACH,MAAM,WAAW,kBAAkB;IAClC;;OAEG;IACH,KAAK,EAAE,OAAO,CAAC;IACf,YAAY,EAAE,OAAO,CAAC;IACtB;;OAEG;IACH,aAAa,CAAC,EAAE,MAAM,CAAC;IACvB,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B"}
1
+ {"version":3,"file":"inference.d.ts","sourceRoot":"","sources":["../../../../src/tasks/text-to-speech/inference.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AAEH;;GAEG;AACH,MAAM,WAAW,iBAAiB;IACjC;;OAEG;IACH,MAAM,EAAE,MAAM,CAAC;IACf;;OAEG;IACH,UAAU,CAAC,EAAE,sBAAsB,CAAC;IACpC,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED;;;;GAIG;AACH,MAAM,WAAW,sBAAsB;IACtC;;OAEG;IACH,QAAQ,CAAC,EAAE,oBAAoB,CAAC;IAChC,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED;;;;GAIG;AACH,MAAM,WAAW,oBAAoB;IACpC;;OAEG;IACH,SAAS,CAAC,EAAE,OAAO,CAAC;IACpB;;OAEG;IACH,cAAc,CAAC,EAAE,kBAAkB,CAAC;IACpC;;;;;OAKG;IACH,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB;;;;;;;;OAQG;IACH,UAAU,CAAC,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,UAAU,CAAC,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB;;OAEG;IACH,UAAU,CAAC,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB;;;OAGG;IACH,eAAe,CAAC,EAAE,MAAM,CAAC;IACzB;;OAEG;IACH,SAAS,CAAC,EAAE,MAAM,CAAC;IACnB;;;OAGG;IACH,aAAa,CAAC,EAAE,MAAM,CAAC;IACvB;;OAEG;IACH,WAAW,CAAC,EAAE,MAAM,CAAC;IACrB;;OAEG;IACH,KAAK,CAAC,EAAE,MAAM,CAAC;IACf;;;OAGG;IACH,KAAK,CAAC,EAAE,MAAM,CAAC;IACf;;;;;;OAMG;IACH,SAAS,CAAC,EAAE,MAAM,CAAC;IACnB;;OAEG;IACH,SAAS,CAAC,EAAE,OAAO,CAAC;IACpB,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AAED;;GAEG;AACH,MAAM,MAAM,kBAAkB,GAAG,OAAO,GAAG,OAAO,CAAC;AAEnD;;;;GAIG;AACH,MAAM,WAAW,kBAAkB;IAClC;;OAEG;IACH,KAAK,EAAE,OAAO,CAAC;IACf,YAAY,EAAE,OAAO,CAAC;IACtB;;OAEG;IACH,aAAa,CAAC,EAAE,MAAM,CAAC;IACvB,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B"}
package/package.json CHANGED
@@ -1,7 +1,7 @@
1
1
  {
2
2
  "name": "@huggingface/tasks",
3
3
  "packageManager": "pnpm@8.10.5",
4
- "version": "0.12.8",
4
+ "version": "0.12.10",
5
5
  "description": "List of ML tasks for huggingface.co/tasks",
6
6
  "repository": "https://github.com/huggingface/huggingface.js.git",
7
7
  "publishConfig": {
@@ -951,6 +951,12 @@ export const mlxim = (model: ModelData): string[] => [
951
951
  model = create_model(${model.id})`,
952
952
  ];
953
953
 
954
+ export const model2vec = (model: ModelData): string[] => [
955
+ `from model2vec import StaticModel
956
+
957
+ model = StaticModel.from_pretrained("${model.id}")`,
958
+ ];
959
+
954
960
  export const nemo = (model: ModelData): string[] => {
955
961
  let command: string[] | undefined = undefined;
956
962
  // Resolve the tag to a nemo domain/sub-domain
@@ -406,6 +406,20 @@ export const MODEL_LIBRARIES_UI_ELEMENTS = {
406
406
  filter: false,
407
407
  countDownloads: `path:"mlc-chat-config.json"`,
408
408
  },
409
+ model2vec: {
410
+ prettyLabel: "Model2Vec",
411
+ repoName: "model2vec",
412
+ repoUrl: "https://github.com/MinishLab/model2vec",
413
+ snippets: snippets.model2vec,
414
+ filter: false,
415
+ },
416
+ moshi: {
417
+ prettyLabel: "Moshi",
418
+ repoName: "Moshi",
419
+ repoUrl: "https://github.com/kyutai-labs/moshi",
420
+ filter: false,
421
+ countDownloads: `path:"tokenizer-e351c8d8-checkpoint125.safetensors"`,
422
+ },
409
423
  nemo: {
410
424
  prettyLabel: "NeMo",
411
425
  repoName: "NeMo",
@@ -73,7 +73,7 @@ export interface GenerationParameters {
73
73
  */
74
74
  max_length?: number;
75
75
  /**
76
- * The maximum number of tokens to generate. Takes precedence over maxLength.
76
+ * The maximum number of tokens to generate. Takes precedence over max_length.
77
77
  */
78
78
  max_new_tokens?: number;
79
79
  /**
@@ -81,7 +81,7 @@ export interface GenerationParameters {
81
81
  */
82
82
  min_length?: number;
83
83
  /**
84
- * The minimum number of tokens to generate. Takes precedence over maxLength.
84
+ * The minimum number of tokens to generate. Takes precedence over min_length.
85
85
  */
86
86
  min_new_tokens?: number;
87
87
  /**
@@ -49,7 +49,7 @@ export interface ChatCompletionInput {
49
49
  * [UNUSED] ID of the model to use. See the model endpoint compatibility table for details
50
50
  * on which models work with the Chat API.
51
51
  */
52
- model: string;
52
+ model?: string;
53
53
  /**
54
54
  * UNUSED
55
55
  * How many chat completion choices to generate for each input message. Note that you will
@@ -63,12 +63,14 @@ export interface ChatCompletionInput {
63
63
  * increasing the model's likelihood to talk about new topics
64
64
  */
65
65
  presence_penalty?: number;
66
+ response_format?: ChatCompletionInputGrammarType;
66
67
  seed?: number;
67
68
  /**
68
69
  * Up to 4 sequences where the API will stop generating further tokens.
69
70
  */
70
71
  stop?: string[];
71
72
  stream?: boolean;
73
+ stream_options?: ChatCompletionInputStreamOptions;
72
74
  /**
73
75
  * What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the
74
76
  * output more random, while
@@ -77,7 +79,7 @@ export interface ChatCompletionInput {
77
79
  * We generally recommend altering this or `top_p` but not both.
78
80
  */
79
81
  temperature?: number;
80
- tool_choice?: ChatCompletionInputToolType;
82
+ tool_choice?: ChatCompletionInputTool;
81
83
  /**
82
84
  * A prompt to be appended before the tools
83
85
  */
@@ -87,7 +89,7 @@ export interface ChatCompletionInput {
87
89
  * Use this to provide a list of
88
90
  * functions the model may generate JSON inputs for.
89
91
  */
90
- tools?: ChatCompletionInputTool[];
92
+ tools?: ToolElement[];
91
93
  /**
92
94
  * An integer between 0 and 5 specifying the number of most likely tokens to return at each
93
95
  * token position, each with
@@ -105,40 +107,78 @@ export interface ChatCompletionInput {
105
107
  }
106
108
 
107
109
  export interface ChatCompletionInputMessage {
108
- content?: string;
110
+ content: ChatCompletionInputMessageContent;
109
111
  name?: string;
110
112
  role: string;
111
- tool_calls?: ChatCompletionInputToolCall[];
112
113
  [property: string]: unknown;
113
114
  }
114
115
 
115
- export interface ChatCompletionInputToolCall {
116
- function: ChatCompletionInputFunctionDefinition;
117
- id: number;
118
- type: string;
116
+ export type ChatCompletionInputMessageContent = ChatCompletionInputMessageChunk[] | string;
117
+
118
+ export interface ChatCompletionInputMessageChunk {
119
+ image_url?: ChatCompletionInputURL;
120
+ text?: string;
121
+ type: ChatCompletionInputMessageChunkType;
119
122
  [property: string]: unknown;
120
123
  }
121
124
 
122
- export interface ChatCompletionInputFunctionDefinition {
123
- arguments: unknown;
124
- description?: string;
125
- name: string;
125
+ export interface ChatCompletionInputURL {
126
+ url: string;
126
127
  [property: string]: unknown;
127
128
  }
128
129
 
129
- export type ChatCompletionInputToolType = "OneOf" | ChatCompletionInputToolTypeObject;
130
+ export type ChatCompletionInputMessageChunkType = "text" | "image_url";
130
131
 
131
- export interface ChatCompletionInputToolTypeObject {
132
- FunctionName: string;
132
+ export interface ChatCompletionInputGrammarType {
133
+ type: ChatCompletionInputGrammarTypeType;
134
+ /**
135
+ * A string that represents a [JSON Schema](https://json-schema.org/).
136
+ *
137
+ * JSON Schema is a declarative language that allows to annotate JSON documents
138
+ * with types and descriptions.
139
+ */
140
+ value: unknown;
141
+ [property: string]: unknown;
142
+ }
143
+
144
+ export type ChatCompletionInputGrammarTypeType = "json" | "regex";
145
+
146
+ export interface ChatCompletionInputStreamOptions {
147
+ /**
148
+ * If set, an additional chunk will be streamed before the data: [DONE] message. The usage
149
+ * field on this chunk shows the token usage statistics for the entire request, and the
150
+ * choices field will always be an empty array. All other chunks will also include a usage
151
+ * field, but with a null value.
152
+ */
153
+ include_usage: boolean;
133
154
  [property: string]: unknown;
134
155
  }
135
156
 
136
- export interface ChatCompletionInputTool {
157
+ export type ChatCompletionInputTool = ChatCompletionInputToolType | string;
158
+
159
+ export interface ChatCompletionInputToolType {
160
+ function?: ChatCompletionInputFunctionName;
161
+ [property: string]: unknown;
162
+ }
163
+
164
+ export interface ChatCompletionInputFunctionName {
165
+ name: string;
166
+ [property: string]: unknown;
167
+ }
168
+
169
+ export interface ToolElement {
137
170
  function: ChatCompletionInputFunctionDefinition;
138
171
  type: string;
139
172
  [property: string]: unknown;
140
173
  }
141
174
 
175
+ export interface ChatCompletionInputFunctionDefinition {
176
+ arguments: unknown;
177
+ description?: string;
178
+ name: string;
179
+ [property: string]: unknown;
180
+ }
181
+
142
182
  /**
143
183
  * Chat Completion Output.
144
184
  *
@@ -151,7 +191,6 @@ export interface ChatCompletionOutput {
151
191
  created: number;
152
192
  id: string;
153
193
  model: string;
154
- object: string;
155
194
  system_fingerprint: string;
156
195
  usage: ChatCompletionOutputUsage;
157
196
  [property: string]: unknown;
@@ -185,7 +224,6 @@ export interface ChatCompletionOutputTopLogprob {
185
224
 
186
225
  export interface ChatCompletionOutputMessage {
187
226
  content?: string;
188
- name?: string;
189
227
  role: string;
190
228
  tool_calls?: ChatCompletionOutputToolCall[];
191
229
  [property: string]: unknown;
@@ -193,7 +231,7 @@ export interface ChatCompletionOutputMessage {
193
231
 
194
232
  export interface ChatCompletionOutputToolCall {
195
233
  function: ChatCompletionOutputFunctionDefinition;
196
- id: number;
234
+ id: string;
197
235
  type: string;
198
236
  [property: string]: unknown;
199
237
  }
@@ -224,8 +262,8 @@ export interface ChatCompletionStreamOutput {
224
262
  created: number;
225
263
  id: string;
226
264
  model: string;
227
- object: string;
228
265
  system_fingerprint: string;
266
+ usage?: ChatCompletionStreamOutputUsage;
229
267
  [property: string]: unknown;
230
268
  }
231
269
 
@@ -275,3 +313,10 @@ export interface ChatCompletionStreamOutputTopLogprob {
275
313
  token: string;
276
314
  [property: string]: unknown;
277
315
  }
316
+
317
+ export interface ChatCompletionStreamOutputUsage {
318
+ completion_tokens: number;
319
+ prompt_tokens: number;
320
+ total_tokens: number;
321
+ [property: string]: unknown;
322
+ }