@huggingface/tasks 0.15.6 → 0.15.7
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/commonjs/hardware.d.ts +2 -2
- package/dist/commonjs/hardware.js +2 -2
- package/dist/commonjs/library-to-tasks.d.ts.map +1 -1
- package/dist/commonjs/library-to-tasks.js +2 -1
- package/dist/commonjs/model-libraries-snippets.js +1 -1
- package/dist/commonjs/tasks/automatic-speech-recognition/inference.d.ts +1 -1
- package/dist/commonjs/tasks/automatic-speech-recognition/inference.d.ts.map +1 -1
- package/dist/commonjs/tasks/feature-extraction/inference.d.ts +6 -2
- package/dist/commonjs/tasks/feature-extraction/inference.d.ts.map +1 -1
- package/dist/commonjs/tasks/image-text-to-text/data.js +3 -3
- package/dist/commonjs/tasks/text-generation/data.d.ts.map +1 -1
- package/dist/commonjs/tasks/text-generation/data.js +15 -4
- package/dist/commonjs/tasks/text-to-image/inference.d.ts +6 -10
- package/dist/commonjs/tasks/text-to-image/inference.d.ts.map +1 -1
- package/dist/esm/hardware.d.ts +2 -2
- package/dist/esm/hardware.js +2 -2
- package/dist/esm/library-to-tasks.d.ts.map +1 -1
- package/dist/esm/library-to-tasks.js +2 -1
- package/dist/esm/model-libraries-snippets.js +1 -1
- package/dist/esm/tasks/automatic-speech-recognition/inference.d.ts +1 -1
- package/dist/esm/tasks/automatic-speech-recognition/inference.d.ts.map +1 -1
- package/dist/esm/tasks/feature-extraction/inference.d.ts +6 -2
- package/dist/esm/tasks/feature-extraction/inference.d.ts.map +1 -1
- package/dist/esm/tasks/image-text-to-text/data.js +3 -3
- package/dist/esm/tasks/text-generation/data.d.ts.map +1 -1
- package/dist/esm/tasks/text-generation/data.js +15 -4
- package/dist/esm/tasks/text-to-image/inference.d.ts +6 -10
- package/dist/esm/tasks/text-to-image/inference.d.ts.map +1 -1
- package/package.json +1 -1
- package/src/hardware.ts +2 -2
- package/src/library-to-tasks.ts +2 -1
- package/src/model-libraries-snippets.ts +1 -1
- package/src/tasks/automatic-speech-recognition/inference.ts +1 -1
- package/src/tasks/automatic-speech-recognition/spec/output.json +2 -2
- package/src/tasks/feature-extraction/inference.ts +6 -2
- package/src/tasks/feature-extraction/spec/input.json +6 -2
- package/src/tasks/image-text-to-text/data.ts +3 -3
- package/src/tasks/text-generation/data.ts +15 -4
- package/src/tasks/text-to-image/inference.ts +6 -10
- package/src/tasks/text-to-image/spec/input.json +7 -12
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
/**
|
|
2
|
-
* Biden AI Executive Order
|
|
3
|
-
* https://www.whitehouse.gov/briefing-room/presidential-actions/2023/10/30/executive-order-on-the-safe-secure-and-trustworthy-development-and-use-of-artificial-intelligence/
|
|
2
|
+
* Biden AI Executive Order (since revoked by President Trump):
|
|
3
|
+
* https://web.archive.org/web/20250105222429/https://www.whitehouse.gov/briefing-room/presidential-actions/2023/10/30/executive-order-on-the-safe-secure-and-trustworthy-development-and-use-of-artificial-intelligence/
|
|
4
4
|
*/
|
|
5
5
|
export declare const TFLOPS_THRESHOLD_WHITE_HOUSE_MODEL_TRAINING_TOTAL: number;
|
|
6
6
|
export declare const TFLOPS_THRESHOLD_WHITE_HOUSE_MODEL_TRAINING_TOTAL_BIOLOGY: number;
|
|
@@ -2,8 +2,8 @@
|
|
|
2
2
|
Object.defineProperty(exports, "__esModule", { value: true });
|
|
3
3
|
exports.SKUS = exports.DEFAULT_MEMORY_OPTIONS = exports.TFLOPS_THRESHOLD_EU_AI_ACT_MODEL_TRAINING_TOTAL = exports.TFLOPS_THRESHOLD_WHITE_HOUSE_CLUSTER = exports.TFLOPS_THRESHOLD_WHITE_HOUSE_MODEL_TRAINING_TOTAL_BIOLOGY = exports.TFLOPS_THRESHOLD_WHITE_HOUSE_MODEL_TRAINING_TOTAL = void 0;
|
|
4
4
|
/**
|
|
5
|
-
* Biden AI Executive Order
|
|
6
|
-
* https://www.whitehouse.gov/briefing-room/presidential-actions/2023/10/30/executive-order-on-the-safe-secure-and-trustworthy-development-and-use-of-artificial-intelligence/
|
|
5
|
+
* Biden AI Executive Order (since revoked by President Trump):
|
|
6
|
+
* https://web.archive.org/web/20250105222429/https://www.whitehouse.gov/briefing-room/presidential-actions/2023/10/30/executive-order-on-the-safe-secure-and-trustworthy-development-and-use-of-artificial-intelligence/
|
|
7
7
|
*/
|
|
8
8
|
exports.TFLOPS_THRESHOLD_WHITE_HOUSE_MODEL_TRAINING_TOTAL = 10 ** 14;
|
|
9
9
|
exports.TFLOPS_THRESHOLD_WHITE_HOUSE_MODEL_TRAINING_TOTAL_BIOLOGY = 10 ** 11;
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"library-to-tasks.d.ts","sourceRoot":"","sources":["../../src/library-to-tasks.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,eAAe,EAAE,MAAM,sBAAsB,CAAC;AAC5D,OAAO,KAAK,EAAE,YAAY,EAAE,MAAM,gBAAgB,CAAC;AAEnD;;;;;;GAMG;AACH,eAAO,MAAM,oBAAoB,EAAE,OAAO,CAAC,MAAM,CAAC,eAAe,EAAE,YAAY,EAAE,CAAC,
|
|
1
|
+
{"version":3,"file":"library-to-tasks.d.ts","sourceRoot":"","sources":["../../src/library-to-tasks.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,eAAe,EAAE,MAAM,sBAAsB,CAAC;AAC5D,OAAO,KAAK,EAAE,YAAY,EAAE,MAAM,gBAAgB,CAAC;AAEnD;;;;;;GAMG;AACH,eAAO,MAAM,oBAAoB,EAAE,OAAO,CAAC,MAAM,CAAC,eAAe,EAAE,YAAY,EAAE,CAAC,CAoEjF,CAAC"}
|
|
@@ -43,7 +43,7 @@ exports.LIBRARY_TASK_MAPPING = {
|
|
|
43
43
|
"text2text-generation",
|
|
44
44
|
],
|
|
45
45
|
stanza: ["token-classification"],
|
|
46
|
-
timm: ["image-classification"],
|
|
46
|
+
timm: ["image-classification", "image-feature-extraction"],
|
|
47
47
|
transformers: [
|
|
48
48
|
"audio-classification",
|
|
49
49
|
"automatic-speech-recognition",
|
|
@@ -52,6 +52,7 @@ exports.LIBRARY_TASK_MAPPING = {
|
|
|
52
52
|
"feature-extraction",
|
|
53
53
|
"fill-mask",
|
|
54
54
|
"image-classification",
|
|
55
|
+
"image-feature-extraction",
|
|
55
56
|
"image-segmentation",
|
|
56
57
|
"image-to-image",
|
|
57
58
|
"image-to-text",
|
|
@@ -1137,7 +1137,7 @@ exports.birefnet = birefnet;
|
|
|
1137
1137
|
const swarmformer = (model) => [
|
|
1138
1138
|
`from swarmformer import SwarmFormerModel
|
|
1139
1139
|
|
|
1140
|
-
model = SwarmFormerModel.from_pretrained("${model.id}"
|
|
1140
|
+
model = SwarmFormerModel.from_pretrained("${model.id}")
|
|
1141
1141
|
`,
|
|
1142
1142
|
];
|
|
1143
1143
|
exports.swarmformer = swarmformer;
|
|
@@ -145,7 +145,7 @@ export interface AutomaticSpeechRecognitionOutputChunk {
|
|
|
145
145
|
/**
|
|
146
146
|
* The start and end timestamps corresponding with the text
|
|
147
147
|
*/
|
|
148
|
-
|
|
148
|
+
timestamp: number[];
|
|
149
149
|
[property: string]: unknown;
|
|
150
150
|
}
|
|
151
151
|
//# sourceMappingURL=inference.d.ts.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"inference.d.ts","sourceRoot":"","sources":["../../../../src/tasks/automatic-speech-recognition/inference.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AACH;;GAEG;AACH,MAAM,WAAW,+BAA+B;IAC/C;;;OAGG;IACH,MAAM,EAAE,IAAI,CAAC;IACb;;OAEG;IACH,UAAU,CAAC,EAAE,oCAAoC,CAAC;IAClD,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AACD;;GAEG;AACH,MAAM,WAAW,oCAAoC;IACpD;;OAEG;IACH,qBAAqB,CAAC,EAAE,oBAAoB,CAAC;IAC7C;;OAEG;IACH,iBAAiB,CAAC,EAAE,OAAO,CAAC;IAC5B,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AACD;;GAEG;AACH,MAAM,WAAW,oBAAoB;IACpC;;OAEG;IACH,SAAS,CAAC,EAAE,OAAO,CAAC;IACpB;;OAEG;IACH,cAAc,CAAC,EAAE,kBAAkB,CAAC;IACpC;;;;;OAKG;IACH,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB;;;;;;;;OAQG;IACH,UAAU,CAAC,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,UAAU,CAAC,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB;;OAEG;IACH,UAAU,CAAC,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB;;;OAGG;IACH,eAAe,CAAC,EAAE,MAAM,CAAC;IACzB;;OAEG;IACH,SAAS,CAAC,EAAE,MAAM,CAAC;IACnB;;;OAGG;IACH,aAAa,CAAC,EAAE,MAAM,CAAC;IACvB;;OAEG;IACH,WAAW,CAAC,EAAE,MAAM,CAAC;IACrB;;OAEG;IACH,KAAK,CAAC,EAAE,MAAM,CAAC;IACf;;;OAGG;IACH,KAAK,CAAC,EAAE,MAAM,CAAC;IACf;;;;;;OAMG;IACH,SAAS,CAAC,EAAE,MAAM,CAAC;IACnB;;OAEG;IACH,SAAS,CAAC,EAAE,OAAO,CAAC;IACpB,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AACD;;GAEG;AACH,MAAM,MAAM,kBAAkB,GAAG,OAAO,GAAG,OAAO,CAAC;AACnD;;GAEG;AACH,MAAM,WAAW,gCAAgC;IAChD;;;OAGG;IACH,MAAM,CAAC,EAAE,qCAAqC,EAAE,CAAC;IACjD;;OAEG;IACH,IAAI,EAAE,MAAM,CAAC;IACb,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AACD,MAAM,WAAW,qCAAqC;IACrD;;OAEG;IACH,IAAI,EAAE,MAAM,CAAC;IACb;;OAEG;IACH,
|
|
1
|
+
{"version":3,"file":"inference.d.ts","sourceRoot":"","sources":["../../../../src/tasks/automatic-speech-recognition/inference.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AACH;;GAEG;AACH,MAAM,WAAW,+BAA+B;IAC/C;;;OAGG;IACH,MAAM,EAAE,IAAI,CAAC;IACb;;OAEG;IACH,UAAU,CAAC,EAAE,oCAAoC,CAAC;IAClD,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AACD;;GAEG;AACH,MAAM,WAAW,oCAAoC;IACpD;;OAEG;IACH,qBAAqB,CAAC,EAAE,oBAAoB,CAAC;IAC7C;;OAEG;IACH,iBAAiB,CAAC,EAAE,OAAO,CAAC;IAC5B,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AACD;;GAEG;AACH,MAAM,WAAW,oBAAoB;IACpC;;OAEG;IACH,SAAS,CAAC,EAAE,OAAO,CAAC;IACpB;;OAEG;IACH,cAAc,CAAC,EAAE,kBAAkB,CAAC;IACpC;;;;;OAKG;IACH,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB;;;;;;;;OAQG;IACH,UAAU,CAAC,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,UAAU,CAAC,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB;;OAEG;IACH,UAAU,CAAC,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB;;;OAGG;IACH,eAAe,CAAC,EAAE,MAAM,CAAC;IACzB;;OAEG;IACH,SAAS,CAAC,EAAE,MAAM,CAAC;IACnB;;;OAGG;IACH,aAAa,CAAC,EAAE,MAAM,CAAC;IACvB;;OAEG;IACH,WAAW,CAAC,EAAE,MAAM,CAAC;IACrB;;OAEG;IACH,KAAK,CAAC,EAAE,MAAM,CAAC;IACf;;;OAGG;IACH,KAAK,CAAC,EAAE,MAAM,CAAC;IACf;;;;;;OAMG;IACH,SAAS,CAAC,EAAE,MAAM,CAAC;IACnB;;OAEG;IACH,SAAS,CAAC,EAAE,OAAO,CAAC;IACpB,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AACD;;GAEG;AACH,MAAM,MAAM,kBAAkB,GAAG,OAAO,GAAG,OAAO,CAAC;AACnD;;GAEG;AACH,MAAM,WAAW,gCAAgC;IAChD;;;OAGG;IACH,MAAM,CAAC,EAAE,qCAAqC,EAAE,CAAC;IACjD;;OAEG;IACH,IAAI,EAAE,MAAM,CAAC;IACb,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AACD,MAAM,WAAW,qCAAqC;IACrD;;OAEG;IACH,IAAI,EAAE,MAAM,CAAC;IACb;;OAEG;IACH,SAAS,EAAE,MAAM,EAAE,CAAC;IACpB,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B"}
|
|
@@ -13,9 +13,9 @@ export type FeatureExtractionOutput = Array<number[]>;
|
|
|
13
13
|
*/
|
|
14
14
|
export interface FeatureExtractionInput {
|
|
15
15
|
/**
|
|
16
|
-
* The text to embed.
|
|
16
|
+
* The text or list of texts to embed.
|
|
17
17
|
*/
|
|
18
|
-
inputs:
|
|
18
|
+
inputs: FeatureExtractionInputs;
|
|
19
19
|
normalize?: boolean;
|
|
20
20
|
/**
|
|
21
21
|
* The name of the prompt that should be used by for encoding. If not set, no prompt
|
|
@@ -34,5 +34,9 @@ export interface FeatureExtractionInput {
|
|
|
34
34
|
truncation_direction?: FeatureExtractionInputTruncationDirection;
|
|
35
35
|
[property: string]: unknown;
|
|
36
36
|
}
|
|
37
|
+
/**
|
|
38
|
+
* The text or list of texts to embed.
|
|
39
|
+
*/
|
|
40
|
+
export type FeatureExtractionInputs = string[] | string;
|
|
37
41
|
export type FeatureExtractionInputTruncationDirection = "Left" | "Right";
|
|
38
42
|
//# sourceMappingURL=inference.d.ts.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"inference.d.ts","sourceRoot":"","sources":["../../../../src/tasks/feature-extraction/inference.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AACH,MAAM,MAAM,uBAAuB,GAAG,KAAK,CAAC,MAAM,EAAE,CAAC,CAAC;AACtD;;;;;;GAMG;AACH,MAAM,WAAW,sBAAsB;IACtC;;OAEG;IACH,MAAM,EAAE,
|
|
1
|
+
{"version":3,"file":"inference.d.ts","sourceRoot":"","sources":["../../../../src/tasks/feature-extraction/inference.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AACH,MAAM,MAAM,uBAAuB,GAAG,KAAK,CAAC,MAAM,EAAE,CAAC,CAAC;AACtD;;;;;;GAMG;AACH,MAAM,WAAW,sBAAsB;IACtC;;OAEG;IACH,MAAM,EAAE,uBAAuB,CAAC;IAChC,SAAS,CAAC,EAAE,OAAO,CAAC;IACpB;;;;;;;;;;;OAWG;IACH,WAAW,CAAC,EAAE,MAAM,CAAC;IACrB,QAAQ,CAAC,EAAE,OAAO,CAAC;IACnB,oBAAoB,CAAC,EAAE,yCAAyC,CAAC;IACjE,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AACD;;GAEG;AACH,MAAM,MAAM,uBAAuB,GAAG,MAAM,EAAE,GAAG,MAAM,CAAC;AACxD,MAAM,MAAM,yCAAyC,GAAG,MAAM,GAAG,OAAO,CAAC"}
|
|
@@ -59,7 +59,7 @@ const taskData = {
|
|
|
59
59
|
},
|
|
60
60
|
{
|
|
61
61
|
description: "Strong image-text-to-text model.",
|
|
62
|
-
id: "Qwen/Qwen2-VL-7B-Instruct",
|
|
62
|
+
id: "Qwen/Qwen2.5-VL-7B-Instruct",
|
|
63
63
|
},
|
|
64
64
|
{
|
|
65
65
|
description: "Image-text-to-text model with reasoning capabilities.",
|
|
@@ -97,11 +97,11 @@ const taskData = {
|
|
|
97
97
|
},
|
|
98
98
|
{
|
|
99
99
|
description: "An application that detects gaze.",
|
|
100
|
-
id: "
|
|
100
|
+
id: "moondream/gaze-demo",
|
|
101
101
|
},
|
|
102
102
|
],
|
|
103
103
|
summary: "Image-text-to-text models take in an image and text prompt and output text. These models are also called vision-language models, or VLMs. The difference from image-to-text models is that these models take an additional text input, not restricting the model to certain use cases like image captioning, and may also be trained to accept a conversation as input.",
|
|
104
|
-
widgetModels: ["
|
|
104
|
+
widgetModels: ["Qwen/Qwen2-VL-7B-Instruct"],
|
|
105
105
|
youtubeId: "IoGaGfU1CIg",
|
|
106
106
|
};
|
|
107
107
|
exports.default = taskData;
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/text-generation/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,
|
|
1
|
+
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/text-generation/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,cA2Hf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
|
@@ -14,6 +14,10 @@ const taskData = {
|
|
|
14
14
|
description: "Truly open-source, curated and cleaned dialogue dataset.",
|
|
15
15
|
id: "HuggingFaceH4/ultrachat_200k",
|
|
16
16
|
},
|
|
17
|
+
{
|
|
18
|
+
description: "A reasoning dataset.",
|
|
19
|
+
id: "open-r1/OpenThoughts-114k-math",
|
|
20
|
+
},
|
|
17
21
|
{
|
|
18
22
|
description: "A multilingual instruction dataset with preference ratings on responses.",
|
|
19
23
|
id: "allenai/tulu-3-sft-mixture",
|
|
@@ -54,10 +58,13 @@ const taskData = {
|
|
|
54
58
|
},
|
|
55
59
|
],
|
|
56
60
|
models: [
|
|
57
|
-
{
|
|
58
|
-
description: "A text-generation model trained to follow instructions.",
|
|
61
|
+
{ description: "A text-generation model trained to follow instructions.",
|
|
59
62
|
id: "google/gemma-2-2b-it",
|
|
60
63
|
},
|
|
64
|
+
{
|
|
65
|
+
description: "Smaller variant of one of the most powerful models.",
|
|
66
|
+
id: "deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B",
|
|
67
|
+
},
|
|
61
68
|
{
|
|
62
69
|
description: "Very powerful text generation model trained to follow instructions.",
|
|
63
70
|
id: "meta-llama/Meta-Llama-3.1-8B-Instruct",
|
|
@@ -71,13 +78,17 @@ const taskData = {
|
|
|
71
78
|
id: "PowerInfer/SmallThinker-3B-Preview",
|
|
72
79
|
},
|
|
73
80
|
{
|
|
74
|
-
description: "Strong
|
|
75
|
-
id: "Qwen/Qwen2.5-7B-Instruct",
|
|
81
|
+
description: "Strong conversational model that supports very long instructions.",
|
|
82
|
+
id: "Qwen/Qwen2.5-7B-Instruct-1M",
|
|
76
83
|
},
|
|
77
84
|
{
|
|
78
85
|
description: "Text generation model used to write code.",
|
|
79
86
|
id: "Qwen/Qwen2.5-Coder-32B-Instruct",
|
|
80
87
|
},
|
|
88
|
+
{
|
|
89
|
+
description: "Powerful reasoning based open large language model.",
|
|
90
|
+
id: "deepseek-ai/DeepSeek-R1",
|
|
91
|
+
},
|
|
81
92
|
],
|
|
82
93
|
spaces: [
|
|
83
94
|
{
|
|
@@ -26,6 +26,10 @@ export interface TextToImageParameters {
|
|
|
26
26
|
* the text prompt, but values too high may cause saturation and other artifacts.
|
|
27
27
|
*/
|
|
28
28
|
guidance_scale?: number;
|
|
29
|
+
/**
|
|
30
|
+
* The height in pixels of the output image
|
|
31
|
+
*/
|
|
32
|
+
height?: number;
|
|
29
33
|
/**
|
|
30
34
|
* One prompt to guide what NOT to include in image generation.
|
|
31
35
|
*/
|
|
@@ -44,17 +48,9 @@ export interface TextToImageParameters {
|
|
|
44
48
|
*/
|
|
45
49
|
seed?: number;
|
|
46
50
|
/**
|
|
47
|
-
* The
|
|
51
|
+
* The width in pixels of the output image
|
|
48
52
|
*/
|
|
49
|
-
|
|
50
|
-
[property: string]: unknown;
|
|
51
|
-
}
|
|
52
|
-
/**
|
|
53
|
-
* The size in pixel of the output image
|
|
54
|
-
*/
|
|
55
|
-
export interface TargetSize {
|
|
56
|
-
height: number;
|
|
57
|
-
width: number;
|
|
53
|
+
width?: number;
|
|
58
54
|
[property: string]: unknown;
|
|
59
55
|
}
|
|
60
56
|
/**
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"inference.d.ts","sourceRoot":"","sources":["../../../../src/tasks/text-to-image/inference.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AACH;;GAEG;AACH,MAAM,WAAW,gBAAgB;IAChC;;OAEG;IACH,MAAM,EAAE,MAAM,CAAC;IACf;;OAEG;IACH,UAAU,CAAC,EAAE,qBAAqB,CAAC;IACnC,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AACD;;GAEG;AACH,MAAM,WAAW,qBAAqB;IACrC;;;OAGG;IACH,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB;;OAEG;IACH,eAAe,CAAC,EAAE,MAAM,CAAC;IACzB;;;OAGG;IACH,mBAAmB,CAAC,EAAE,MAAM,CAAC;IAC7B;;OAEG;IACH,SAAS,CAAC,EAAE,MAAM,CAAC;IACnB;;OAEG;IACH,IAAI,CAAC,EAAE,MAAM,CAAC;IACd;;OAEG;IACH,
|
|
1
|
+
{"version":3,"file":"inference.d.ts","sourceRoot":"","sources":["../../../../src/tasks/text-to-image/inference.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AACH;;GAEG;AACH,MAAM,WAAW,gBAAgB;IAChC;;OAEG;IACH,MAAM,EAAE,MAAM,CAAC;IACf;;OAEG;IACH,UAAU,CAAC,EAAE,qBAAqB,CAAC;IACnC,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AACD;;GAEG;AACH,MAAM,WAAW,qBAAqB;IACrC;;;OAGG;IACH,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB;;OAEG;IACH,MAAM,CAAC,EAAE,MAAM,CAAC;IAChB;;OAEG;IACH,eAAe,CAAC,EAAE,MAAM,CAAC;IACzB;;;OAGG;IACH,mBAAmB,CAAC,EAAE,MAAM,CAAC;IAC7B;;OAEG;IACH,SAAS,CAAC,EAAE,MAAM,CAAC;IACnB;;OAEG;IACH,IAAI,CAAC,EAAE,MAAM,CAAC;IACd;;OAEG;IACH,KAAK,CAAC,EAAE,MAAM,CAAC;IACf,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AACD;;GAEG;AACH,MAAM,WAAW,iBAAiB;IACjC;;OAEG;IACH,KAAK,EAAE,OAAO,CAAC;IACf,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B"}
|
package/dist/esm/hardware.d.ts
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
/**
|
|
2
|
-
* Biden AI Executive Order
|
|
3
|
-
* https://www.whitehouse.gov/briefing-room/presidential-actions/2023/10/30/executive-order-on-the-safe-secure-and-trustworthy-development-and-use-of-artificial-intelligence/
|
|
2
|
+
* Biden AI Executive Order (since revoked by President Trump):
|
|
3
|
+
* https://web.archive.org/web/20250105222429/https://www.whitehouse.gov/briefing-room/presidential-actions/2023/10/30/executive-order-on-the-safe-secure-and-trustworthy-development-and-use-of-artificial-intelligence/
|
|
4
4
|
*/
|
|
5
5
|
export declare const TFLOPS_THRESHOLD_WHITE_HOUSE_MODEL_TRAINING_TOTAL: number;
|
|
6
6
|
export declare const TFLOPS_THRESHOLD_WHITE_HOUSE_MODEL_TRAINING_TOTAL_BIOLOGY: number;
|
package/dist/esm/hardware.js
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
/**
|
|
2
|
-
* Biden AI Executive Order
|
|
3
|
-
* https://www.whitehouse.gov/briefing-room/presidential-actions/2023/10/30/executive-order-on-the-safe-secure-and-trustworthy-development-and-use-of-artificial-intelligence/
|
|
2
|
+
* Biden AI Executive Order (since revoked by President Trump):
|
|
3
|
+
* https://web.archive.org/web/20250105222429/https://www.whitehouse.gov/briefing-room/presidential-actions/2023/10/30/executive-order-on-the-safe-secure-and-trustworthy-development-and-use-of-artificial-intelligence/
|
|
4
4
|
*/
|
|
5
5
|
export const TFLOPS_THRESHOLD_WHITE_HOUSE_MODEL_TRAINING_TOTAL = 10 ** 14;
|
|
6
6
|
export const TFLOPS_THRESHOLD_WHITE_HOUSE_MODEL_TRAINING_TOTAL_BIOLOGY = 10 ** 11;
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"library-to-tasks.d.ts","sourceRoot":"","sources":["../../src/library-to-tasks.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,eAAe,EAAE,MAAM,sBAAsB,CAAC;AAC5D,OAAO,KAAK,EAAE,YAAY,EAAE,MAAM,gBAAgB,CAAC;AAEnD;;;;;;GAMG;AACH,eAAO,MAAM,oBAAoB,EAAE,OAAO,CAAC,MAAM,CAAC,eAAe,EAAE,YAAY,EAAE,CAAC,
|
|
1
|
+
{"version":3,"file":"library-to-tasks.d.ts","sourceRoot":"","sources":["../../src/library-to-tasks.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,eAAe,EAAE,MAAM,sBAAsB,CAAC;AAC5D,OAAO,KAAK,EAAE,YAAY,EAAE,MAAM,gBAAgB,CAAC;AAEnD;;;;;;GAMG;AACH,eAAO,MAAM,oBAAoB,EAAE,OAAO,CAAC,MAAM,CAAC,eAAe,EAAE,YAAY,EAAE,CAAC,CAoEjF,CAAC"}
|
|
@@ -40,7 +40,7 @@ export const LIBRARY_TASK_MAPPING = {
|
|
|
40
40
|
"text2text-generation",
|
|
41
41
|
],
|
|
42
42
|
stanza: ["token-classification"],
|
|
43
|
-
timm: ["image-classification"],
|
|
43
|
+
timm: ["image-classification", "image-feature-extraction"],
|
|
44
44
|
transformers: [
|
|
45
45
|
"audio-classification",
|
|
46
46
|
"automatic-speech-recognition",
|
|
@@ -49,6 +49,7 @@ export const LIBRARY_TASK_MAPPING = {
|
|
|
49
49
|
"feature-extraction",
|
|
50
50
|
"fill-mask",
|
|
51
51
|
"image-classification",
|
|
52
|
+
"image-feature-extraction",
|
|
52
53
|
"image-segmentation",
|
|
53
54
|
"image-to-image",
|
|
54
55
|
"image-to-text",
|
|
@@ -1070,7 +1070,7 @@ model = BiRefNet.from_pretrained("${model.id}")`,
|
|
|
1070
1070
|
export const swarmformer = (model) => [
|
|
1071
1071
|
`from swarmformer import SwarmFormerModel
|
|
1072
1072
|
|
|
1073
|
-
model = SwarmFormerModel.from_pretrained("${model.id}"
|
|
1073
|
+
model = SwarmFormerModel.from_pretrained("${model.id}")
|
|
1074
1074
|
`,
|
|
1075
1075
|
];
|
|
1076
1076
|
export const mlx = (model) => [
|
|
@@ -145,7 +145,7 @@ export interface AutomaticSpeechRecognitionOutputChunk {
|
|
|
145
145
|
/**
|
|
146
146
|
* The start and end timestamps corresponding with the text
|
|
147
147
|
*/
|
|
148
|
-
|
|
148
|
+
timestamp: number[];
|
|
149
149
|
[property: string]: unknown;
|
|
150
150
|
}
|
|
151
151
|
//# sourceMappingURL=inference.d.ts.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"inference.d.ts","sourceRoot":"","sources":["../../../../src/tasks/automatic-speech-recognition/inference.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AACH;;GAEG;AACH,MAAM,WAAW,+BAA+B;IAC/C;;;OAGG;IACH,MAAM,EAAE,IAAI,CAAC;IACb;;OAEG;IACH,UAAU,CAAC,EAAE,oCAAoC,CAAC;IAClD,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AACD;;GAEG;AACH,MAAM,WAAW,oCAAoC;IACpD;;OAEG;IACH,qBAAqB,CAAC,EAAE,oBAAoB,CAAC;IAC7C;;OAEG;IACH,iBAAiB,CAAC,EAAE,OAAO,CAAC;IAC5B,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AACD;;GAEG;AACH,MAAM,WAAW,oBAAoB;IACpC;;OAEG;IACH,SAAS,CAAC,EAAE,OAAO,CAAC;IACpB;;OAEG;IACH,cAAc,CAAC,EAAE,kBAAkB,CAAC;IACpC;;;;;OAKG;IACH,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB;;;;;;;;OAQG;IACH,UAAU,CAAC,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,UAAU,CAAC,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB;;OAEG;IACH,UAAU,CAAC,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB;;;OAGG;IACH,eAAe,CAAC,EAAE,MAAM,CAAC;IACzB;;OAEG;IACH,SAAS,CAAC,EAAE,MAAM,CAAC;IACnB;;;OAGG;IACH,aAAa,CAAC,EAAE,MAAM,CAAC;IACvB;;OAEG;IACH,WAAW,CAAC,EAAE,MAAM,CAAC;IACrB;;OAEG;IACH,KAAK,CAAC,EAAE,MAAM,CAAC;IACf;;;OAGG;IACH,KAAK,CAAC,EAAE,MAAM,CAAC;IACf;;;;;;OAMG;IACH,SAAS,CAAC,EAAE,MAAM,CAAC;IACnB;;OAEG;IACH,SAAS,CAAC,EAAE,OAAO,CAAC;IACpB,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AACD;;GAEG;AACH,MAAM,MAAM,kBAAkB,GAAG,OAAO,GAAG,OAAO,CAAC;AACnD;;GAEG;AACH,MAAM,WAAW,gCAAgC;IAChD;;;OAGG;IACH,MAAM,CAAC,EAAE,qCAAqC,EAAE,CAAC;IACjD;;OAEG;IACH,IAAI,EAAE,MAAM,CAAC;IACb,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AACD,MAAM,WAAW,qCAAqC;IACrD;;OAEG;IACH,IAAI,EAAE,MAAM,CAAC;IACb;;OAEG;IACH,
|
|
1
|
+
{"version":3,"file":"inference.d.ts","sourceRoot":"","sources":["../../../../src/tasks/automatic-speech-recognition/inference.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AACH;;GAEG;AACH,MAAM,WAAW,+BAA+B;IAC/C;;;OAGG;IACH,MAAM,EAAE,IAAI,CAAC;IACb;;OAEG;IACH,UAAU,CAAC,EAAE,oCAAoC,CAAC;IAClD,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AACD;;GAEG;AACH,MAAM,WAAW,oCAAoC;IACpD;;OAEG;IACH,qBAAqB,CAAC,EAAE,oBAAoB,CAAC;IAC7C;;OAEG;IACH,iBAAiB,CAAC,EAAE,OAAO,CAAC;IAC5B,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AACD;;GAEG;AACH,MAAM,WAAW,oBAAoB;IACpC;;OAEG;IACH,SAAS,CAAC,EAAE,OAAO,CAAC;IACpB;;OAEG;IACH,cAAc,CAAC,EAAE,kBAAkB,CAAC;IACpC;;;;;OAKG;IACH,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB;;;;;;;;OAQG;IACH,UAAU,CAAC,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,UAAU,CAAC,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB;;OAEG;IACH,UAAU,CAAC,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB;;;OAGG;IACH,eAAe,CAAC,EAAE,MAAM,CAAC;IACzB;;OAEG;IACH,SAAS,CAAC,EAAE,MAAM,CAAC;IACnB;;;OAGG;IACH,aAAa,CAAC,EAAE,MAAM,CAAC;IACvB;;OAEG;IACH,WAAW,CAAC,EAAE,MAAM,CAAC;IACrB;;OAEG;IACH,KAAK,CAAC,EAAE,MAAM,CAAC;IACf;;;OAGG;IACH,KAAK,CAAC,EAAE,MAAM,CAAC;IACf;;;;;;OAMG;IACH,SAAS,CAAC,EAAE,MAAM,CAAC;IACnB;;OAEG;IACH,SAAS,CAAC,EAAE,OAAO,CAAC;IACpB,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AACD;;GAEG;AACH,MAAM,MAAM,kBAAkB,GAAG,OAAO,GAAG,OAAO,CAAC;AACnD;;GAEG;AACH,MAAM,WAAW,gCAAgC;IAChD;;;OAGG;IACH,MAAM,CAAC,EAAE,qCAAqC,EAAE,CAAC;IACjD;;OAEG;IACH,IAAI,EAAE,MAAM,CAAC;IACb,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AACD,MAAM,WAAW,qCAAqC;IACrD;;OAEG;IACH,IAAI,EAAE,MAAM,CAAC;IACb;;OAEG;IACH,SAAS,EAAE,MAAM,EAAE,CAAC;IACpB,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B"}
|
|
@@ -13,9 +13,9 @@ export type FeatureExtractionOutput = Array<number[]>;
|
|
|
13
13
|
*/
|
|
14
14
|
export interface FeatureExtractionInput {
|
|
15
15
|
/**
|
|
16
|
-
* The text to embed.
|
|
16
|
+
* The text or list of texts to embed.
|
|
17
17
|
*/
|
|
18
|
-
inputs:
|
|
18
|
+
inputs: FeatureExtractionInputs;
|
|
19
19
|
normalize?: boolean;
|
|
20
20
|
/**
|
|
21
21
|
* The name of the prompt that should be used by for encoding. If not set, no prompt
|
|
@@ -34,5 +34,9 @@ export interface FeatureExtractionInput {
|
|
|
34
34
|
truncation_direction?: FeatureExtractionInputTruncationDirection;
|
|
35
35
|
[property: string]: unknown;
|
|
36
36
|
}
|
|
37
|
+
/**
|
|
38
|
+
* The text or list of texts to embed.
|
|
39
|
+
*/
|
|
40
|
+
export type FeatureExtractionInputs = string[] | string;
|
|
37
41
|
export type FeatureExtractionInputTruncationDirection = "Left" | "Right";
|
|
38
42
|
//# sourceMappingURL=inference.d.ts.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"inference.d.ts","sourceRoot":"","sources":["../../../../src/tasks/feature-extraction/inference.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AACH,MAAM,MAAM,uBAAuB,GAAG,KAAK,CAAC,MAAM,EAAE,CAAC,CAAC;AACtD;;;;;;GAMG;AACH,MAAM,WAAW,sBAAsB;IACtC;;OAEG;IACH,MAAM,EAAE,
|
|
1
|
+
{"version":3,"file":"inference.d.ts","sourceRoot":"","sources":["../../../../src/tasks/feature-extraction/inference.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AACH,MAAM,MAAM,uBAAuB,GAAG,KAAK,CAAC,MAAM,EAAE,CAAC,CAAC;AACtD;;;;;;GAMG;AACH,MAAM,WAAW,sBAAsB;IACtC;;OAEG;IACH,MAAM,EAAE,uBAAuB,CAAC;IAChC,SAAS,CAAC,EAAE,OAAO,CAAC;IACpB;;;;;;;;;;;OAWG;IACH,WAAW,CAAC,EAAE,MAAM,CAAC;IACrB,QAAQ,CAAC,EAAE,OAAO,CAAC;IACnB,oBAAoB,CAAC,EAAE,yCAAyC,CAAC;IACjE,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AACD;;GAEG;AACH,MAAM,MAAM,uBAAuB,GAAG,MAAM,EAAE,GAAG,MAAM,CAAC;AACxD,MAAM,MAAM,yCAAyC,GAAG,MAAM,GAAG,OAAO,CAAC"}
|
|
@@ -57,7 +57,7 @@ const taskData = {
|
|
|
57
57
|
},
|
|
58
58
|
{
|
|
59
59
|
description: "Strong image-text-to-text model.",
|
|
60
|
-
id: "Qwen/Qwen2-VL-7B-Instruct",
|
|
60
|
+
id: "Qwen/Qwen2.5-VL-7B-Instruct",
|
|
61
61
|
},
|
|
62
62
|
{
|
|
63
63
|
description: "Image-text-to-text model with reasoning capabilities.",
|
|
@@ -95,11 +95,11 @@ const taskData = {
|
|
|
95
95
|
},
|
|
96
96
|
{
|
|
97
97
|
description: "An application that detects gaze.",
|
|
98
|
-
id: "
|
|
98
|
+
id: "moondream/gaze-demo",
|
|
99
99
|
},
|
|
100
100
|
],
|
|
101
101
|
summary: "Image-text-to-text models take in an image and text prompt and output text. These models are also called vision-language models, or VLMs. The difference from image-to-text models is that these models take an additional text input, not restricting the model to certain use cases like image captioning, and may also be trained to accept a conversation as input.",
|
|
102
|
-
widgetModels: ["
|
|
102
|
+
widgetModels: ["Qwen/Qwen2-VL-7B-Instruct"],
|
|
103
103
|
youtubeId: "IoGaGfU1CIg",
|
|
104
104
|
};
|
|
105
105
|
export default taskData;
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/text-generation/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,
|
|
1
|
+
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/text-generation/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,cA2Hf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
|
@@ -12,6 +12,10 @@ const taskData = {
|
|
|
12
12
|
description: "Truly open-source, curated and cleaned dialogue dataset.",
|
|
13
13
|
id: "HuggingFaceH4/ultrachat_200k",
|
|
14
14
|
},
|
|
15
|
+
{
|
|
16
|
+
description: "A reasoning dataset.",
|
|
17
|
+
id: "open-r1/OpenThoughts-114k-math",
|
|
18
|
+
},
|
|
15
19
|
{
|
|
16
20
|
description: "A multilingual instruction dataset with preference ratings on responses.",
|
|
17
21
|
id: "allenai/tulu-3-sft-mixture",
|
|
@@ -52,10 +56,13 @@ const taskData = {
|
|
|
52
56
|
},
|
|
53
57
|
],
|
|
54
58
|
models: [
|
|
55
|
-
{
|
|
56
|
-
description: "A text-generation model trained to follow instructions.",
|
|
59
|
+
{ description: "A text-generation model trained to follow instructions.",
|
|
57
60
|
id: "google/gemma-2-2b-it",
|
|
58
61
|
},
|
|
62
|
+
{
|
|
63
|
+
description: "Smaller variant of one of the most powerful models.",
|
|
64
|
+
id: "deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B",
|
|
65
|
+
},
|
|
59
66
|
{
|
|
60
67
|
description: "Very powerful text generation model trained to follow instructions.",
|
|
61
68
|
id: "meta-llama/Meta-Llama-3.1-8B-Instruct",
|
|
@@ -69,13 +76,17 @@ const taskData = {
|
|
|
69
76
|
id: "PowerInfer/SmallThinker-3B-Preview",
|
|
70
77
|
},
|
|
71
78
|
{
|
|
72
|
-
description: "Strong
|
|
73
|
-
id: "Qwen/Qwen2.5-7B-Instruct",
|
|
79
|
+
description: "Strong conversational model that supports very long instructions.",
|
|
80
|
+
id: "Qwen/Qwen2.5-7B-Instruct-1M",
|
|
74
81
|
},
|
|
75
82
|
{
|
|
76
83
|
description: "Text generation model used to write code.",
|
|
77
84
|
id: "Qwen/Qwen2.5-Coder-32B-Instruct",
|
|
78
85
|
},
|
|
86
|
+
{
|
|
87
|
+
description: "Powerful reasoning based open large language model.",
|
|
88
|
+
id: "deepseek-ai/DeepSeek-R1",
|
|
89
|
+
},
|
|
79
90
|
],
|
|
80
91
|
spaces: [
|
|
81
92
|
{
|
|
@@ -26,6 +26,10 @@ export interface TextToImageParameters {
|
|
|
26
26
|
* the text prompt, but values too high may cause saturation and other artifacts.
|
|
27
27
|
*/
|
|
28
28
|
guidance_scale?: number;
|
|
29
|
+
/**
|
|
30
|
+
* The height in pixels of the output image
|
|
31
|
+
*/
|
|
32
|
+
height?: number;
|
|
29
33
|
/**
|
|
30
34
|
* One prompt to guide what NOT to include in image generation.
|
|
31
35
|
*/
|
|
@@ -44,17 +48,9 @@ export interface TextToImageParameters {
|
|
|
44
48
|
*/
|
|
45
49
|
seed?: number;
|
|
46
50
|
/**
|
|
47
|
-
* The
|
|
51
|
+
* The width in pixels of the output image
|
|
48
52
|
*/
|
|
49
|
-
|
|
50
|
-
[property: string]: unknown;
|
|
51
|
-
}
|
|
52
|
-
/**
|
|
53
|
-
* The size in pixel of the output image
|
|
54
|
-
*/
|
|
55
|
-
export interface TargetSize {
|
|
56
|
-
height: number;
|
|
57
|
-
width: number;
|
|
53
|
+
width?: number;
|
|
58
54
|
[property: string]: unknown;
|
|
59
55
|
}
|
|
60
56
|
/**
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"inference.d.ts","sourceRoot":"","sources":["../../../../src/tasks/text-to-image/inference.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AACH;;GAEG;AACH,MAAM,WAAW,gBAAgB;IAChC;;OAEG;IACH,MAAM,EAAE,MAAM,CAAC;IACf;;OAEG;IACH,UAAU,CAAC,EAAE,qBAAqB,CAAC;IACnC,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AACD;;GAEG;AACH,MAAM,WAAW,qBAAqB;IACrC;;;OAGG;IACH,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB;;OAEG;IACH,eAAe,CAAC,EAAE,MAAM,CAAC;IACzB;;;OAGG;IACH,mBAAmB,CAAC,EAAE,MAAM,CAAC;IAC7B;;OAEG;IACH,SAAS,CAAC,EAAE,MAAM,CAAC;IACnB;;OAEG;IACH,IAAI,CAAC,EAAE,MAAM,CAAC;IACd;;OAEG;IACH,
|
|
1
|
+
{"version":3,"file":"inference.d.ts","sourceRoot":"","sources":["../../../../src/tasks/text-to-image/inference.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AACH;;GAEG;AACH,MAAM,WAAW,gBAAgB;IAChC;;OAEG;IACH,MAAM,EAAE,MAAM,CAAC;IACf;;OAEG;IACH,UAAU,CAAC,EAAE,qBAAqB,CAAC;IACnC,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AACD;;GAEG;AACH,MAAM,WAAW,qBAAqB;IACrC;;;OAGG;IACH,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB;;OAEG;IACH,MAAM,CAAC,EAAE,MAAM,CAAC;IAChB;;OAEG;IACH,eAAe,CAAC,EAAE,MAAM,CAAC;IACzB;;;OAGG;IACH,mBAAmB,CAAC,EAAE,MAAM,CAAC;IAC7B;;OAEG;IACH,SAAS,CAAC,EAAE,MAAM,CAAC;IACnB;;OAEG;IACH,IAAI,CAAC,EAAE,MAAM,CAAC;IACd;;OAEG;IACH,KAAK,CAAC,EAAE,MAAM,CAAC;IACf,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AACD;;GAEG;AACH,MAAM,WAAW,iBAAiB;IACjC;;OAEG;IACH,KAAK,EAAE,OAAO,CAAC;IACf,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B"}
|
package/package.json
CHANGED
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@huggingface/tasks",
|
|
3
3
|
"packageManager": "pnpm@8.10.5",
|
|
4
|
-
"version": "0.15.
|
|
4
|
+
"version": "0.15.7",
|
|
5
5
|
"description": "List of ML tasks for huggingface.co/tasks",
|
|
6
6
|
"repository": "https://github.com/huggingface/huggingface.js.git",
|
|
7
7
|
"publishConfig": {
|
package/src/hardware.ts
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
/**
|
|
2
|
-
* Biden AI Executive Order
|
|
3
|
-
* https://www.whitehouse.gov/briefing-room/presidential-actions/2023/10/30/executive-order-on-the-safe-secure-and-trustworthy-development-and-use-of-artificial-intelligence/
|
|
2
|
+
* Biden AI Executive Order (since revoked by President Trump):
|
|
3
|
+
* https://web.archive.org/web/20250105222429/https://www.whitehouse.gov/briefing-room/presidential-actions/2023/10/30/executive-order-on-the-safe-secure-and-trustworthy-development-and-use-of-artificial-intelligence/
|
|
4
4
|
*/
|
|
5
5
|
export const TFLOPS_THRESHOLD_WHITE_HOUSE_MODEL_TRAINING_TOTAL = 10 ** 14;
|
|
6
6
|
export const TFLOPS_THRESHOLD_WHITE_HOUSE_MODEL_TRAINING_TOTAL_BIOLOGY = 10 ** 11;
|
package/src/library-to-tasks.ts
CHANGED
|
@@ -43,7 +43,7 @@ export const LIBRARY_TASK_MAPPING: Partial<Record<ModelLibraryKey, PipelineType[
|
|
|
43
43
|
"text2text-generation",
|
|
44
44
|
],
|
|
45
45
|
stanza: ["token-classification"],
|
|
46
|
-
timm: ["image-classification"],
|
|
46
|
+
timm: ["image-classification", "image-feature-extraction"],
|
|
47
47
|
transformers: [
|
|
48
48
|
"audio-classification",
|
|
49
49
|
"automatic-speech-recognition",
|
|
@@ -52,6 +52,7 @@ export const LIBRARY_TASK_MAPPING: Partial<Record<ModelLibraryKey, PipelineType[
|
|
|
52
52
|
"feature-extraction",
|
|
53
53
|
"fill-mask",
|
|
54
54
|
"image-classification",
|
|
55
|
+
"image-feature-extraction",
|
|
55
56
|
"image-segmentation",
|
|
56
57
|
"image-to-image",
|
|
57
58
|
"image-to-text",
|
|
@@ -1186,7 +1186,7 @@ model = BiRefNet.from_pretrained("${model.id}")`,
|
|
|
1186
1186
|
export const swarmformer = (model: ModelData): string[] => [
|
|
1187
1187
|
`from swarmformer import SwarmFormerModel
|
|
1188
1188
|
|
|
1189
|
-
model = SwarmFormerModel.from_pretrained("${model.id}"
|
|
1189
|
+
model = SwarmFormerModel.from_pretrained("${model.id}")
|
|
1190
1190
|
`,
|
|
1191
1191
|
];
|
|
1192
1192
|
|
|
@@ -20,7 +20,7 @@
|
|
|
20
20
|
"type": "string",
|
|
21
21
|
"description": "A chunk of text identified by the model"
|
|
22
22
|
},
|
|
23
|
-
"
|
|
23
|
+
"timestamp": {
|
|
24
24
|
"type": "array",
|
|
25
25
|
"description": "The start and end timestamps corresponding with the text",
|
|
26
26
|
"items": {
|
|
@@ -30,7 +30,7 @@
|
|
|
30
30
|
"maxLength": 2
|
|
31
31
|
}
|
|
32
32
|
},
|
|
33
|
-
"required": ["text", "
|
|
33
|
+
"required": ["text", "timestamp"]
|
|
34
34
|
}
|
|
35
35
|
}
|
|
36
36
|
},
|
|
@@ -13,9 +13,9 @@ export type FeatureExtractionOutput = Array<number[]>;
|
|
|
13
13
|
*/
|
|
14
14
|
export interface FeatureExtractionInput {
|
|
15
15
|
/**
|
|
16
|
-
* The text to embed.
|
|
16
|
+
* The text or list of texts to embed.
|
|
17
17
|
*/
|
|
18
|
-
inputs:
|
|
18
|
+
inputs: FeatureExtractionInputs;
|
|
19
19
|
normalize?: boolean;
|
|
20
20
|
/**
|
|
21
21
|
* The name of the prompt that should be used by for encoding. If not set, no prompt
|
|
@@ -34,4 +34,8 @@ export interface FeatureExtractionInput {
|
|
|
34
34
|
truncation_direction?: FeatureExtractionInputTruncationDirection;
|
|
35
35
|
[property: string]: unknown;
|
|
36
36
|
}
|
|
37
|
+
/**
|
|
38
|
+
* The text or list of texts to embed.
|
|
39
|
+
*/
|
|
40
|
+
export type FeatureExtractionInputs = string[] | string;
|
|
37
41
|
export type FeatureExtractionInputTruncationDirection = "Left" | "Right";
|
|
@@ -7,8 +7,12 @@
|
|
|
7
7
|
"required": ["inputs"],
|
|
8
8
|
"properties": {
|
|
9
9
|
"inputs": {
|
|
10
|
-
"
|
|
11
|
-
"
|
|
10
|
+
"title": "FeatureExtractionInputs",
|
|
11
|
+
"oneOf": [
|
|
12
|
+
{ "type": "string" },
|
|
13
|
+
{ "type": "array", "items": { "type": "string" } }
|
|
14
|
+
],
|
|
15
|
+
"description": "The text or list of texts to embed."
|
|
12
16
|
},
|
|
13
17
|
"normalize": {
|
|
14
18
|
"type": "boolean",
|
|
@@ -60,7 +60,7 @@ const taskData: TaskDataCustom = {
|
|
|
60
60
|
},
|
|
61
61
|
{
|
|
62
62
|
description: "Strong image-text-to-text model.",
|
|
63
|
-
id: "Qwen/Qwen2-VL-7B-Instruct",
|
|
63
|
+
id: "Qwen/Qwen2.5-VL-7B-Instruct",
|
|
64
64
|
},
|
|
65
65
|
{
|
|
66
66
|
description: "Image-text-to-text model with reasoning capabilities.",
|
|
@@ -98,12 +98,12 @@ const taskData: TaskDataCustom = {
|
|
|
98
98
|
},
|
|
99
99
|
{
|
|
100
100
|
description: "An application that detects gaze.",
|
|
101
|
-
id: "
|
|
101
|
+
id: "moondream/gaze-demo",
|
|
102
102
|
},
|
|
103
103
|
],
|
|
104
104
|
summary:
|
|
105
105
|
"Image-text-to-text models take in an image and text prompt and output text. These models are also called vision-language models, or VLMs. The difference from image-to-text models is that these models take an additional text input, not restricting the model to certain use cases like image captioning, and may also be trained to accept a conversation as input.",
|
|
106
|
-
widgetModels: ["
|
|
106
|
+
widgetModels: ["Qwen/Qwen2-VL-7B-Instruct"],
|
|
107
107
|
youtubeId: "IoGaGfU1CIg",
|
|
108
108
|
};
|
|
109
109
|
|
|
@@ -14,6 +14,10 @@ const taskData: TaskDataCustom = {
|
|
|
14
14
|
description: "Truly open-source, curated and cleaned dialogue dataset.",
|
|
15
15
|
id: "HuggingFaceH4/ultrachat_200k",
|
|
16
16
|
},
|
|
17
|
+
{
|
|
18
|
+
description: "A reasoning dataset.",
|
|
19
|
+
id: "open-r1/OpenThoughts-114k-math",
|
|
20
|
+
},
|
|
17
21
|
{
|
|
18
22
|
description: "A multilingual instruction dataset with preference ratings on responses.",
|
|
19
23
|
id: "allenai/tulu-3-sft-mixture",
|
|
@@ -57,10 +61,13 @@ const taskData: TaskDataCustom = {
|
|
|
57
61
|
},
|
|
58
62
|
],
|
|
59
63
|
models: [
|
|
60
|
-
{
|
|
61
|
-
description: "A text-generation model trained to follow instructions.",
|
|
64
|
+
{ description: "A text-generation model trained to follow instructions.",
|
|
62
65
|
id: "google/gemma-2-2b-it",
|
|
63
66
|
},
|
|
67
|
+
{
|
|
68
|
+
description: "Smaller variant of one of the most powerful models.",
|
|
69
|
+
id: "deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B",
|
|
70
|
+
},
|
|
64
71
|
{
|
|
65
72
|
description: "Very powerful text generation model trained to follow instructions.",
|
|
66
73
|
id: "meta-llama/Meta-Llama-3.1-8B-Instruct",
|
|
@@ -74,13 +81,17 @@ const taskData: TaskDataCustom = {
|
|
|
74
81
|
id: "PowerInfer/SmallThinker-3B-Preview",
|
|
75
82
|
},
|
|
76
83
|
{
|
|
77
|
-
description: "Strong
|
|
78
|
-
id: "Qwen/Qwen2.5-7B-Instruct",
|
|
84
|
+
description: "Strong conversational model that supports very long instructions.",
|
|
85
|
+
id: "Qwen/Qwen2.5-7B-Instruct-1M",
|
|
79
86
|
},
|
|
80
87
|
{
|
|
81
88
|
description: "Text generation model used to write code.",
|
|
82
89
|
id: "Qwen/Qwen2.5-Coder-32B-Instruct",
|
|
83
90
|
},
|
|
91
|
+
{
|
|
92
|
+
description: "Powerful reasoning based open large language model.",
|
|
93
|
+
id: "deepseek-ai/DeepSeek-R1",
|
|
94
|
+
},
|
|
84
95
|
],
|
|
85
96
|
spaces: [
|
|
86
97
|
{
|
|
@@ -26,6 +26,10 @@ export interface TextToImageParameters {
|
|
|
26
26
|
* the text prompt, but values too high may cause saturation and other artifacts.
|
|
27
27
|
*/
|
|
28
28
|
guidance_scale?: number;
|
|
29
|
+
/**
|
|
30
|
+
* The height in pixels of the output image
|
|
31
|
+
*/
|
|
32
|
+
height?: number;
|
|
29
33
|
/**
|
|
30
34
|
* One prompt to guide what NOT to include in image generation.
|
|
31
35
|
*/
|
|
@@ -44,17 +48,9 @@ export interface TextToImageParameters {
|
|
|
44
48
|
*/
|
|
45
49
|
seed?: number;
|
|
46
50
|
/**
|
|
47
|
-
* The
|
|
51
|
+
* The width in pixels of the output image
|
|
48
52
|
*/
|
|
49
|
-
|
|
50
|
-
[property: string]: unknown;
|
|
51
|
-
}
|
|
52
|
-
/**
|
|
53
|
-
* The size in pixel of the output image
|
|
54
|
-
*/
|
|
55
|
-
export interface TargetSize {
|
|
56
|
-
height: number;
|
|
57
|
-
width: number;
|
|
53
|
+
width?: number;
|
|
58
54
|
[property: string]: unknown;
|
|
59
55
|
}
|
|
60
56
|
/**
|
|
@@ -31,18 +31,13 @@
|
|
|
31
31
|
"type": "integer",
|
|
32
32
|
"description": "The number of denoising steps. More denoising steps usually lead to a higher quality image at the expense of slower inference."
|
|
33
33
|
},
|
|
34
|
-
"
|
|
35
|
-
"type": "
|
|
36
|
-
"description": "The
|
|
37
|
-
|
|
38
|
-
|
|
39
|
-
|
|
40
|
-
|
|
41
|
-
"height": {
|
|
42
|
-
"type": "integer"
|
|
43
|
-
}
|
|
44
|
-
},
|
|
45
|
-
"required": ["width", "height"]
|
|
34
|
+
"width": {
|
|
35
|
+
"type": "integer",
|
|
36
|
+
"description": "The width in pixels of the output image"
|
|
37
|
+
},
|
|
38
|
+
"height": {
|
|
39
|
+
"type": "integer",
|
|
40
|
+
"description": "The height in pixels of the output image"
|
|
46
41
|
},
|
|
47
42
|
"scheduler": {
|
|
48
43
|
"type": "string",
|