@huggingface/tasks 0.13.15 → 0.13.16
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/commonjs/local-apps.js +9 -9
- package/dist/commonjs/local-apps.spec.js +2 -8
- package/dist/commonjs/tasks/audio-to-audio/data.d.ts.map +1 -1
- package/dist/commonjs/tasks/audio-to-audio/data.js +0 -4
- package/dist/commonjs/tasks/fill-mask/data.js +2 -2
- package/dist/commonjs/tasks/image-classification/data.d.ts.map +1 -1
- package/dist/commonjs/tasks/image-classification/data.js +2 -3
- package/dist/commonjs/tasks/image-feature-extraction/data.d.ts.map +1 -1
- package/dist/commonjs/tasks/image-feature-extraction/data.js +8 -3
- package/dist/commonjs/tasks/image-text-to-text/data.d.ts.map +1 -1
- package/dist/commonjs/tasks/image-text-to-text/data.js +19 -11
- package/dist/commonjs/tasks/image-to-3d/data.js +4 -4
- package/dist/commonjs/tasks/image-to-image/data.d.ts.map +1 -1
- package/dist/commonjs/tasks/image-to-image/data.js +12 -4
- package/dist/commonjs/tasks/index.js +1 -1
- package/dist/commonjs/tasks/keypoint-detection/data.d.ts.map +1 -1
- package/dist/commonjs/tasks/keypoint-detection/data.js +4 -0
- package/dist/commonjs/tasks/object-detection/data.d.ts.map +1 -1
- package/dist/commonjs/tasks/object-detection/data.js +6 -2
- package/dist/commonjs/tasks/sentence-similarity/data.d.ts.map +1 -1
- package/dist/commonjs/tasks/sentence-similarity/data.js +5 -1
- package/dist/commonjs/tasks/text-generation/data.d.ts.map +1 -1
- package/dist/commonjs/tasks/text-generation/data.js +17 -13
- package/dist/commonjs/tasks/text-to-image/data.d.ts.map +1 -1
- package/dist/commonjs/tasks/text-to-image/data.js +4 -0
- package/dist/commonjs/tasks/text-to-speech/data.d.ts.map +1 -1
- package/dist/commonjs/tasks/text-to-speech/data.js +9 -9
- package/dist/commonjs/tasks/text-to-video/data.js +5 -5
- package/dist/commonjs/tasks/video-text-to-text/data.js +4 -4
- package/dist/commonjs/tasks/zero-shot-classification/data.js +2 -2
- package/dist/commonjs/tasks/zero-shot-image-classification/data.d.ts.map +1 -1
- package/dist/commonjs/tasks/zero-shot-image-classification/data.js +8 -4
- package/dist/esm/local-apps.js +9 -9
- package/dist/esm/local-apps.spec.js +2 -8
- package/dist/esm/tasks/audio-to-audio/data.d.ts.map +1 -1
- package/dist/esm/tasks/audio-to-audio/data.js +0 -4
- package/dist/esm/tasks/fill-mask/data.js +2 -2
- package/dist/esm/tasks/image-classification/data.d.ts.map +1 -1
- package/dist/esm/tasks/image-classification/data.js +2 -3
- package/dist/esm/tasks/image-feature-extraction/data.d.ts.map +1 -1
- package/dist/esm/tasks/image-feature-extraction/data.js +8 -3
- package/dist/esm/tasks/image-text-to-text/data.d.ts.map +1 -1
- package/dist/esm/tasks/image-text-to-text/data.js +19 -11
- package/dist/esm/tasks/image-to-3d/data.js +4 -4
- package/dist/esm/tasks/image-to-image/data.d.ts.map +1 -1
- package/dist/esm/tasks/image-to-image/data.js +12 -4
- package/dist/esm/tasks/index.js +1 -1
- package/dist/esm/tasks/keypoint-detection/data.d.ts.map +1 -1
- package/dist/esm/tasks/keypoint-detection/data.js +4 -0
- package/dist/esm/tasks/object-detection/data.d.ts.map +1 -1
- package/dist/esm/tasks/object-detection/data.js +6 -2
- package/dist/esm/tasks/sentence-similarity/data.d.ts.map +1 -1
- package/dist/esm/tasks/sentence-similarity/data.js +5 -1
- package/dist/esm/tasks/text-generation/data.d.ts.map +1 -1
- package/dist/esm/tasks/text-generation/data.js +17 -13
- package/dist/esm/tasks/text-to-image/data.d.ts.map +1 -1
- package/dist/esm/tasks/text-to-image/data.js +4 -0
- package/dist/esm/tasks/text-to-speech/data.d.ts.map +1 -1
- package/dist/esm/tasks/text-to-speech/data.js +9 -9
- package/dist/esm/tasks/text-to-video/data.js +5 -5
- package/dist/esm/tasks/video-text-to-text/data.js +4 -4
- package/dist/esm/tasks/zero-shot-classification/data.js +2 -2
- package/dist/esm/tasks/zero-shot-image-classification/data.d.ts.map +1 -1
- package/dist/esm/tasks/zero-shot-image-classification/data.js +8 -4
- package/package.json +1 -1
- package/src/local-apps.spec.ts +2 -8
- package/src/local-apps.ts +9 -9
- package/src/tasks/audio-to-audio/data.ts +0 -4
- package/src/tasks/fill-mask/data.ts +2 -2
- package/src/tasks/image-classification/data.ts +2 -3
- package/src/tasks/image-feature-extraction/data.ts +8 -3
- package/src/tasks/image-text-to-text/about.md +8 -3
- package/src/tasks/image-text-to-text/data.ts +19 -11
- package/src/tasks/image-to-3d/data.ts +4 -4
- package/src/tasks/image-to-image/data.ts +12 -5
- package/src/tasks/index.ts +1 -1
- package/src/tasks/keypoint-detection/data.ts +4 -0
- package/src/tasks/object-detection/data.ts +6 -2
- package/src/tasks/sentence-similarity/data.ts +5 -1
- package/src/tasks/text-generation/data.ts +17 -14
- package/src/tasks/text-to-image/data.ts +4 -0
- package/src/tasks/text-to-speech/data.ts +9 -10
- package/src/tasks/text-to-video/data.ts +5 -5
- package/src/tasks/video-text-to-text/data.ts +4 -4
- package/src/tasks/zero-shot-classification/data.ts +2 -2
- package/src/tasks/zero-shot-image-classification/data.ts +8 -4
|
@@ -29,17 +29,17 @@ function isMlxModel(model) {
|
|
|
29
29
|
return model.tags.includes("mlx");
|
|
30
30
|
}
|
|
31
31
|
const snippetLlamacpp = (model, filepath) => {
|
|
32
|
+
let tagName = "";
|
|
33
|
+
if (filepath) {
|
|
34
|
+
const quantLabel = (0, gguf_js_1.parseGGUFQuantLabel)(filepath);
|
|
35
|
+
tagName = quantLabel ? `:${quantLabel}` : "";
|
|
36
|
+
}
|
|
32
37
|
const command = (binary) => {
|
|
33
|
-
const snippet = [
|
|
34
|
-
|
|
35
|
-
|
|
36
|
-
` --hf-repo "${model.id}" \\`,
|
|
37
|
-
` --hf-file ${filepath ?? "{{GGUF_FILE}}"} \\`,
|
|
38
|
-
` -p "${model.tags.includes("conversational") ? "You are a helpful assistant" : "Once upon a time,"}"`,
|
|
39
|
-
];
|
|
40
|
-
if (model.tags.includes("conversational")) {
|
|
38
|
+
const snippet = ["# Load and run the model:", `${binary} -hf ${model.id}${tagName}`];
|
|
39
|
+
if (!model.tags.includes("conversational")) {
|
|
40
|
+
// for non-conversational models, add a prompt
|
|
41
41
|
snippet[snippet.length - 1] += " \\";
|
|
42
|
-
snippet.push(
|
|
42
|
+
snippet.push(' -p "Once upon a time,"');
|
|
43
43
|
}
|
|
44
44
|
return snippet.join("\n");
|
|
45
45
|
};
|
|
@@ -12,11 +12,7 @@ const local_apps_js_1 = require("./local-apps.js");
|
|
|
12
12
|
};
|
|
13
13
|
const snippet = snippetFunc(model);
|
|
14
14
|
(0, vitest_1.expect)(snippet[0].content).toEqual(`# Load and run the model:
|
|
15
|
-
llama-cli
|
|
16
|
-
--hf-repo "bartowski/Llama-3.2-3B-Instruct-GGUF" \\
|
|
17
|
-
--hf-file {{GGUF_FILE}} \\
|
|
18
|
-
-p "You are a helpful assistant" \\
|
|
19
|
-
--conversation`);
|
|
15
|
+
llama-cli -hf bartowski/Llama-3.2-3B-Instruct-GGUF`);
|
|
20
16
|
});
|
|
21
17
|
(0, vitest_1.it)("llama.cpp non-conversational", async () => {
|
|
22
18
|
const { snippet: snippetFunc } = local_apps_js_1.LOCAL_APPS["llama.cpp"];
|
|
@@ -27,9 +23,7 @@ llama-cli \\
|
|
|
27
23
|
};
|
|
28
24
|
const snippet = snippetFunc(model);
|
|
29
25
|
(0, vitest_1.expect)(snippet[0].content).toEqual(`# Load and run the model:
|
|
30
|
-
llama-cli \\
|
|
31
|
-
--hf-repo "mlabonne/gemma-2b-GGUF" \\
|
|
32
|
-
--hf-file {{GGUF_FILE}} \\
|
|
26
|
+
llama-cli -hf mlabonne/gemma-2b-GGUF \\
|
|
33
27
|
-p "Once upon a time,"`);
|
|
34
28
|
});
|
|
35
29
|
(0, vitest_1.it)("vLLM conversational llm", async () => {
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/audio-to-audio/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,
|
|
1
|
+
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/audio-to-audio/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,cA6Df,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
|
@@ -59,8 +59,8 @@ const taskData = {
|
|
|
59
59
|
],
|
|
60
60
|
models: [
|
|
61
61
|
{
|
|
62
|
-
description: "
|
|
63
|
-
id: "
|
|
62
|
+
description: "State-of-the-art masked language model.",
|
|
63
|
+
id: "answerdotai/ModernBERT-large",
|
|
64
64
|
},
|
|
65
65
|
{
|
|
66
66
|
description: "A multilingual model trained on 100 languages.",
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/image-classification/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,
|
|
1
|
+
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/image-classification/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,cAkFf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
|
@@ -74,9 +74,8 @@ const taskData = {
|
|
|
74
74
|
],
|
|
75
75
|
spaces: [
|
|
76
76
|
{
|
|
77
|
-
|
|
78
|
-
|
|
79
|
-
id: "nielsr/perceiver-image-classification",
|
|
77
|
+
description: "A leaderboard to evaluate different image classification models.",
|
|
78
|
+
id: "timm/leaderboard",
|
|
80
79
|
},
|
|
81
80
|
],
|
|
82
81
|
summary: "Image classification is the task of assigning a label or class to an entire image. Images are expected to have only one class for each image. Image classification models take an image as input and return a prediction about which class the image belongs to.",
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/image-feature-extraction/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,
|
|
1
|
+
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/image-feature-extraction/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,cA2Df,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
|
@@ -42,15 +42,20 @@ const taskData = {
|
|
|
42
42
|
id: "facebook/dino-vitb16",
|
|
43
43
|
},
|
|
44
44
|
{
|
|
45
|
-
description: "
|
|
46
|
-
id: "
|
|
45
|
+
description: "Cutting-edge image feature extraction model.",
|
|
46
|
+
id: "apple/aimv2-large-patch14-336-distilled",
|
|
47
47
|
},
|
|
48
48
|
{
|
|
49
49
|
description: "Strong image feature extraction model that can be used on images and documents.",
|
|
50
50
|
id: "OpenGVLab/InternViT-6B-448px-V1-2",
|
|
51
51
|
},
|
|
52
52
|
],
|
|
53
|
-
spaces: [
|
|
53
|
+
spaces: [
|
|
54
|
+
{
|
|
55
|
+
description: "A leaderboard to evaluate different image-feature-extraction models on classification performances",
|
|
56
|
+
id: "timm/leaderboard",
|
|
57
|
+
},
|
|
58
|
+
],
|
|
54
59
|
summary: "Image feature extraction is the task of extracting features learnt in a computer vision model.",
|
|
55
60
|
widgetModels: [],
|
|
56
61
|
};
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/image-text-to-text/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,
|
|
1
|
+
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/image-text-to-text/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,cAyGf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
|
@@ -7,8 +7,8 @@ const taskData = {
|
|
|
7
7
|
id: "liuhaotian/LLaVA-Instruct-150K",
|
|
8
8
|
},
|
|
9
9
|
{
|
|
10
|
-
description: "
|
|
11
|
-
id: "
|
|
10
|
+
description: "Collection of image-text pairs on scientific topics.",
|
|
11
|
+
id: "DAMO-NLP-SG/multimodal_textbook",
|
|
12
12
|
},
|
|
13
13
|
{
|
|
14
14
|
description: "A collection of datasets made for model fine-tuning.",
|
|
@@ -42,11 +42,15 @@ const taskData = {
|
|
|
42
42
|
metrics: [],
|
|
43
43
|
models: [
|
|
44
44
|
{
|
|
45
|
-
description: "
|
|
46
|
-
id: "
|
|
45
|
+
description: "Small and efficient yet powerful vision language model.",
|
|
46
|
+
id: "HuggingFaceTB/SmolVLM-Instruct",
|
|
47
47
|
},
|
|
48
48
|
{
|
|
49
|
-
description: "
|
|
49
|
+
description: "A screenshot understanding model used to control computers.",
|
|
50
|
+
id: "showlab/ShowUI-2B",
|
|
51
|
+
},
|
|
52
|
+
{
|
|
53
|
+
description: "Cutting-edge vision language model.",
|
|
50
54
|
id: "allenai/Molmo-7B-D-0924",
|
|
51
55
|
},
|
|
52
56
|
{
|
|
@@ -58,8 +62,8 @@ const taskData = {
|
|
|
58
62
|
id: "Qwen/Qwen2-VL-7B-Instruct",
|
|
59
63
|
},
|
|
60
64
|
{
|
|
61
|
-
description: "
|
|
62
|
-
id: "
|
|
65
|
+
description: "Image-text-to-text model with reasoning capabilities.",
|
|
66
|
+
id: "Qwen/QVQ-72B-Preview",
|
|
63
67
|
},
|
|
64
68
|
{
|
|
65
69
|
description: "Strong image-text-to-text model focused on documents.",
|
|
@@ -83,14 +87,18 @@ const taskData = {
|
|
|
83
87
|
description: "An image-text-to-text application focused on documents.",
|
|
84
88
|
id: "stepfun-ai/GOT_official_online_demo",
|
|
85
89
|
},
|
|
86
|
-
{
|
|
87
|
-
description: "An application to compare outputs of different vision language models.",
|
|
88
|
-
id: "merve/compare_VLMs",
|
|
89
|
-
},
|
|
90
90
|
{
|
|
91
91
|
description: "An application for chatting with an image-text-to-text model.",
|
|
92
92
|
id: "GanymedeNil/Qwen2-VL-7B",
|
|
93
93
|
},
|
|
94
|
+
{
|
|
95
|
+
description: "An application that parses screenshots into actions.",
|
|
96
|
+
id: "showlab/ShowUI",
|
|
97
|
+
},
|
|
98
|
+
{
|
|
99
|
+
description: "An application that detects gaze.",
|
|
100
|
+
id: "smoondream/gaze-demo",
|
|
101
|
+
},
|
|
94
102
|
],
|
|
95
103
|
summary: "Image-text-to-text models take in an image and text prompt and output text. These models are also called vision-language models, or VLMs. The difference from image-to-text models is that these models take an additional text input, not restricting the model to certain use cases like image captioning, and may also be trained to accept a conversation as input.",
|
|
96
104
|
widgetModels: ["meta-llama/Llama-3.2-11B-Vision-Instruct"],
|
|
@@ -41,8 +41,8 @@ const taskData = {
|
|
|
41
41
|
id: "hwjiang/Real3D",
|
|
42
42
|
},
|
|
43
43
|
{
|
|
44
|
-
description: "
|
|
45
|
-
id: "
|
|
44
|
+
description: "Consistent image-to-3d generation model.",
|
|
45
|
+
id: "stabilityai/stable-point-aware-3d",
|
|
46
46
|
},
|
|
47
47
|
],
|
|
48
48
|
spaces: [
|
|
@@ -55,8 +55,8 @@ const taskData = {
|
|
|
55
55
|
id: "TencentARC/InstantMesh",
|
|
56
56
|
},
|
|
57
57
|
{
|
|
58
|
-
description: "Image-to-3D demo
|
|
59
|
-
id: "stabilityai/
|
|
58
|
+
description: "Image-to-3D demo.",
|
|
59
|
+
id: "stabilityai/stable-point-aware-3d",
|
|
60
60
|
},
|
|
61
61
|
{
|
|
62
62
|
description: "Image-to-3D demo with mesh outputs.",
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/image-to-image/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,
|
|
1
|
+
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/image-to-image/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,cAuGf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
|
@@ -10,6 +10,10 @@ const taskData = {
|
|
|
10
10
|
description: "Multiple images of celebrities, used for facial expression translation",
|
|
11
11
|
id: "huggan/CelebA-faces",
|
|
12
12
|
},
|
|
13
|
+
{
|
|
14
|
+
description: "12M image-caption pairs.",
|
|
15
|
+
id: "Spawning/PD12M",
|
|
16
|
+
},
|
|
13
17
|
],
|
|
14
18
|
demo: {
|
|
15
19
|
inputs: [
|
|
@@ -50,16 +54,20 @@ const taskData = {
|
|
|
50
54
|
id: "keras-io/super-resolution",
|
|
51
55
|
},
|
|
52
56
|
{
|
|
53
|
-
description: "A model
|
|
54
|
-
id: "
|
|
57
|
+
description: "A model for applying edits to images through image controls.",
|
|
58
|
+
id: "Yuanshi/OminiControl",
|
|
55
59
|
},
|
|
56
60
|
{
|
|
57
61
|
description: "A model that generates images based on segments in the input image and the text prompt.",
|
|
58
62
|
id: "mfidabel/controlnet-segment-anything",
|
|
59
63
|
},
|
|
60
64
|
{
|
|
61
|
-
description: "
|
|
62
|
-
id: "
|
|
65
|
+
description: "Strong model for inpainting and outpainting.",
|
|
66
|
+
id: "black-forest-labs/FLUX.1-Fill-dev",
|
|
67
|
+
},
|
|
68
|
+
{
|
|
69
|
+
description: "Strong model for image editing using depth maps.",
|
|
70
|
+
id: "black-forest-labs/FLUX.1-Depth-dev-lora",
|
|
63
71
|
},
|
|
64
72
|
],
|
|
65
73
|
spaces: [
|
|
@@ -69,7 +69,7 @@ exports.TASKS_MODEL_LIBRARIES = {
|
|
|
69
69
|
"video-classification": ["transformers"],
|
|
70
70
|
"mask-generation": ["transformers"],
|
|
71
71
|
"multiple-choice": ["transformers"],
|
|
72
|
-
"object-detection": ["transformers", "transformers.js"],
|
|
72
|
+
"object-detection": ["transformers", "transformers.js", "ultralytics"],
|
|
73
73
|
other: [],
|
|
74
74
|
"question-answering": ["adapter-transformers", "allennlp", "transformers", "transformers.js"],
|
|
75
75
|
robotics: [],
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/keypoint-detection/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,
|
|
1
|
+
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/keypoint-detection/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,cAiDf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
|
@@ -31,6 +31,10 @@ const taskData = {
|
|
|
31
31
|
description: "Strong keypoint detection model used to detect human pose.",
|
|
32
32
|
id: "facebook/sapiens-pose-1b",
|
|
33
33
|
},
|
|
34
|
+
{
|
|
35
|
+
description: "Powerful keypoint detection model used to detect human pose.",
|
|
36
|
+
id: "usyd-community/vitpose-plus-base",
|
|
37
|
+
},
|
|
34
38
|
],
|
|
35
39
|
spaces: [
|
|
36
40
|
{
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/object-detection/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,
|
|
1
|
+
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/object-detection/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,cAqFf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
|
@@ -52,6 +52,10 @@ const taskData = {
|
|
|
52
52
|
description: "Fast and accurate object detection model trained on COCO and Object365 datasets.",
|
|
53
53
|
id: "PekingU/rtdetr_r18vd_coco_o365",
|
|
54
54
|
},
|
|
55
|
+
{
|
|
56
|
+
description: "Object detection model for low-lying objects.",
|
|
57
|
+
id: "StephanST/WALDO30",
|
|
58
|
+
},
|
|
55
59
|
],
|
|
56
60
|
spaces: [
|
|
57
61
|
{
|
|
@@ -63,8 +67,8 @@ const taskData = {
|
|
|
63
67
|
id: "Gradio-Blocks/Object-Detection-With-DETR-and-YOLOS",
|
|
64
68
|
},
|
|
65
69
|
{
|
|
66
|
-
description: "
|
|
67
|
-
id: "
|
|
70
|
+
description: "A cutting-edge object detection application.",
|
|
71
|
+
id: "Ultralytics/YOLO11",
|
|
68
72
|
},
|
|
69
73
|
{
|
|
70
74
|
description: "An object tracking, segmentation and inpainting application.",
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/sentence-similarity/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,
|
|
1
|
+
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/sentence-similarity/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,cAoGf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
|
@@ -66,9 +66,13 @@ const taskData = {
|
|
|
66
66
|
id: "sentence-transformers/all-mpnet-base-v2",
|
|
67
67
|
},
|
|
68
68
|
{
|
|
69
|
-
description: "A multilingual robust sentence similarity model
|
|
69
|
+
description: "A multilingual robust sentence similarity model.",
|
|
70
70
|
id: "BAAI/bge-m3",
|
|
71
71
|
},
|
|
72
|
+
{
|
|
73
|
+
description: "A robust sentence similarity model.",
|
|
74
|
+
id: "HIT-TMG/KaLM-embedding-multilingual-mini-instruct-v1.5",
|
|
75
|
+
},
|
|
72
76
|
],
|
|
73
77
|
spaces: [
|
|
74
78
|
{
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/text-generation/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,
|
|
1
|
+
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/text-generation/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,cAgHf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
|
@@ -3,24 +3,28 @@ Object.defineProperty(exports, "__esModule", { value: true });
|
|
|
3
3
|
const taskData = {
|
|
4
4
|
datasets: [
|
|
5
5
|
{
|
|
6
|
-
description: "
|
|
7
|
-
id: "
|
|
6
|
+
description: "Multilingual dataset used to evaluate text generation models.",
|
|
7
|
+
id: "CohereForAI/Global-MMLU",
|
|
8
8
|
},
|
|
9
9
|
{
|
|
10
|
-
description: "
|
|
11
|
-
id: "
|
|
10
|
+
description: "High quality multilingual data used to train text-generation models.",
|
|
11
|
+
id: "HuggingFaceFW/fineweb-2",
|
|
12
12
|
},
|
|
13
13
|
{
|
|
14
14
|
description: "Truly open-source, curated and cleaned dialogue dataset.",
|
|
15
15
|
id: "HuggingFaceH4/ultrachat_200k",
|
|
16
16
|
},
|
|
17
17
|
{
|
|
18
|
-
description: "
|
|
19
|
-
id: "
|
|
18
|
+
description: "A multilingual instruction dataset with preference ratings on responses.",
|
|
19
|
+
id: "allenai/tulu-3-sft-mixture",
|
|
20
20
|
},
|
|
21
21
|
{
|
|
22
22
|
description: "A large synthetic dataset for alignment of text generation models.",
|
|
23
|
-
id: "
|
|
23
|
+
id: "HuggingFaceTB/smoltalk",
|
|
24
|
+
},
|
|
25
|
+
{
|
|
26
|
+
description: "A dataset made for training text generation models solving math questions.",
|
|
27
|
+
id: "HuggingFaceTB/finemath",
|
|
24
28
|
},
|
|
25
29
|
],
|
|
26
30
|
demo: {
|
|
@@ -59,20 +63,20 @@ const taskData = {
|
|
|
59
63
|
id: "meta-llama/Meta-Llama-3.1-8B-Instruct",
|
|
60
64
|
},
|
|
61
65
|
{
|
|
62
|
-
description: "
|
|
63
|
-
id: "microsoft/
|
|
66
|
+
description: "Powerful text generation model by Microsoft.",
|
|
67
|
+
id: "microsoft/phi-4",
|
|
64
68
|
},
|
|
65
69
|
{
|
|
66
|
-
description: "A very powerful model
|
|
67
|
-
id: "
|
|
70
|
+
description: "A very powerful model with reasoning capabilities.",
|
|
71
|
+
id: "PowerInfer/SmallThinker-3B-Preview",
|
|
68
72
|
},
|
|
69
73
|
{
|
|
70
74
|
description: "Strong text generation model to follow instructions.",
|
|
71
75
|
id: "Qwen/Qwen2.5-7B-Instruct",
|
|
72
76
|
},
|
|
73
77
|
{
|
|
74
|
-
description: "
|
|
75
|
-
id: "
|
|
78
|
+
description: "Text generation model used to write code.",
|
|
79
|
+
id: "Qwen/Qwen2.5-Coder-32B-Instruct",
|
|
76
80
|
},
|
|
77
81
|
],
|
|
78
82
|
spaces: [
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/text-to-image/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,
|
|
1
|
+
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/text-to-image/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,cAmGf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
|
@@ -10,6 +10,10 @@ const taskData = {
|
|
|
10
10
|
description: "Conceptual Captions is a dataset consisting of ~3.3M images annotated with captions.",
|
|
11
11
|
id: "conceptual_captions",
|
|
12
12
|
},
|
|
13
|
+
{
|
|
14
|
+
description: "12M image-caption pairs.",
|
|
15
|
+
id: "Spawning/PD12M",
|
|
16
|
+
},
|
|
13
17
|
],
|
|
14
18
|
demo: {
|
|
15
19
|
inputs: [
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/text-to-speech/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,
|
|
1
|
+
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/text-to-speech/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,cAiFf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
|
@@ -12,7 +12,7 @@ const taskData = {
|
|
|
12
12
|
id: "mythicinfinity/libritts_r",
|
|
13
13
|
},
|
|
14
14
|
{
|
|
15
|
-
description: "
|
|
15
|
+
description: "Multi-lingual dataset.",
|
|
16
16
|
id: "facebook/multilingual_librispeech",
|
|
17
17
|
},
|
|
18
18
|
],
|
|
@@ -48,25 +48,25 @@ const taskData = {
|
|
|
48
48
|
},
|
|
49
49
|
{
|
|
50
50
|
description: "A massively multi-lingual TTS model.",
|
|
51
|
-
id: "
|
|
51
|
+
id: "fishaudio/fish-speech-1.5",
|
|
52
52
|
},
|
|
53
53
|
{
|
|
54
54
|
description: "A powerful TTS model.",
|
|
55
|
-
id: "
|
|
55
|
+
id: "OuteAI/OuteTTS-0.1-350M",
|
|
56
56
|
},
|
|
57
57
|
{
|
|
58
|
-
description: "
|
|
59
|
-
id: "
|
|
58
|
+
description: "Small yet powerful TTS model.",
|
|
59
|
+
id: "hexgrad/Kokoro-82M",
|
|
60
60
|
},
|
|
61
61
|
],
|
|
62
62
|
spaces: [
|
|
63
63
|
{
|
|
64
|
-
description: "An application for generate
|
|
65
|
-
id: "
|
|
64
|
+
description: "An application for generate high quality speech in different languages.",
|
|
65
|
+
id: "hexgrad/Kokoro-TTS",
|
|
66
66
|
},
|
|
67
67
|
{
|
|
68
|
-
description: "
|
|
69
|
-
id: "
|
|
68
|
+
description: "A multilingual text-to-speech application.",
|
|
69
|
+
id: "fishaudio/fish-speech-1",
|
|
70
70
|
},
|
|
71
71
|
{
|
|
72
72
|
description: "An application that generates speech in different styles in English and Chinese.",
|
|
@@ -63,15 +63,15 @@ const taskData = {
|
|
|
63
63
|
models: [
|
|
64
64
|
{
|
|
65
65
|
description: "A strong model for consistent video generation.",
|
|
66
|
-
id: "
|
|
66
|
+
id: "tencent/HunyuanVideo",
|
|
67
67
|
},
|
|
68
68
|
{
|
|
69
|
-
description: "A
|
|
70
|
-
id: "
|
|
69
|
+
description: "A text-to-video model with high fidelity motion and strong prompt adherence.",
|
|
70
|
+
id: "Lightricks/LTX-Video",
|
|
71
71
|
},
|
|
72
72
|
{
|
|
73
|
-
description: "A
|
|
74
|
-
id: "
|
|
73
|
+
description: "A text-to-video model focusing on physics-aware applications like robotics.",
|
|
74
|
+
id: "nvidia/Cosmos-1.0-Diffusion-7B-Text2World",
|
|
75
75
|
},
|
|
76
76
|
],
|
|
77
77
|
spaces: [
|
|
@@ -38,12 +38,12 @@ const taskData = {
|
|
|
38
38
|
metrics: [],
|
|
39
39
|
models: [
|
|
40
40
|
{
|
|
41
|
-
description: "A robust video-text-to-text model
|
|
42
|
-
id: "
|
|
41
|
+
description: "A robust video-text-to-text model.",
|
|
42
|
+
id: "Vision-CAIR/LongVU_Qwen2_7B",
|
|
43
43
|
},
|
|
44
44
|
{
|
|
45
|
-
description: "
|
|
46
|
-
id: "
|
|
45
|
+
description: "Strong video-text-to-text model with reasoning capabilities.",
|
|
46
|
+
id: "GoodiesHere/Apollo-LMMs-Apollo-7B-t32",
|
|
47
47
|
},
|
|
48
48
|
],
|
|
49
49
|
spaces: [
|
|
@@ -55,8 +55,8 @@ const taskData = {
|
|
|
55
55
|
id: "facebook/bart-large-mnli",
|
|
56
56
|
},
|
|
57
57
|
{
|
|
58
|
-
description: "
|
|
59
|
-
id: "MoritzLaurer/
|
|
58
|
+
description: "Cutting-edge zero-shot multilingual text classification model.",
|
|
59
|
+
id: "MoritzLaurer/ModernBERT-large-zeroshot-v2.0",
|
|
60
60
|
},
|
|
61
61
|
],
|
|
62
62
|
spaces: [],
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/zero-shot-image-classification/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,
|
|
1
|
+
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/zero-shot-image-classification/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,cAmFf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
|
@@ -48,16 +48,20 @@ const taskData = {
|
|
|
48
48
|
],
|
|
49
49
|
models: [
|
|
50
50
|
{
|
|
51
|
-
description: "
|
|
52
|
-
id: "
|
|
51
|
+
description: "Multilingual image classification model for 80 languages.",
|
|
52
|
+
id: "visheratin/mexma-siglip",
|
|
53
53
|
},
|
|
54
54
|
{
|
|
55
55
|
description: "Strong zero-shot image classification model.",
|
|
56
56
|
id: "google/siglip-so400m-patch14-224",
|
|
57
57
|
},
|
|
58
58
|
{
|
|
59
|
-
description: "
|
|
60
|
-
id: "
|
|
59
|
+
description: "Robust zero-shot image classification model.",
|
|
60
|
+
id: "microsoft/LLM2CLIP-EVA02-L-14-336",
|
|
61
|
+
},
|
|
62
|
+
{
|
|
63
|
+
description: "Powerful zero-shot image classification model supporting 94 languages.",
|
|
64
|
+
id: "jinaai/jina-clip-v2",
|
|
61
65
|
},
|
|
62
66
|
{
|
|
63
67
|
description: "Strong image classification model for biomedical domain.",
|
package/dist/esm/local-apps.js
CHANGED
|
@@ -26,17 +26,17 @@ function isMlxModel(model) {
|
|
|
26
26
|
return model.tags.includes("mlx");
|
|
27
27
|
}
|
|
28
28
|
const snippetLlamacpp = (model, filepath) => {
|
|
29
|
+
let tagName = "";
|
|
30
|
+
if (filepath) {
|
|
31
|
+
const quantLabel = parseGGUFQuantLabel(filepath);
|
|
32
|
+
tagName = quantLabel ? `:${quantLabel}` : "";
|
|
33
|
+
}
|
|
29
34
|
const command = (binary) => {
|
|
30
|
-
const snippet = [
|
|
31
|
-
|
|
32
|
-
|
|
33
|
-
` --hf-repo "${model.id}" \\`,
|
|
34
|
-
` --hf-file ${filepath ?? "{{GGUF_FILE}}"} \\`,
|
|
35
|
-
` -p "${model.tags.includes("conversational") ? "You are a helpful assistant" : "Once upon a time,"}"`,
|
|
36
|
-
];
|
|
37
|
-
if (model.tags.includes("conversational")) {
|
|
35
|
+
const snippet = ["# Load and run the model:", `${binary} -hf ${model.id}${tagName}`];
|
|
36
|
+
if (!model.tags.includes("conversational")) {
|
|
37
|
+
// for non-conversational models, add a prompt
|
|
38
38
|
snippet[snippet.length - 1] += " \\";
|
|
39
|
-
snippet.push(
|
|
39
|
+
snippet.push(' -p "Once upon a time,"');
|
|
40
40
|
}
|
|
41
41
|
return snippet.join("\n");
|
|
42
42
|
};
|
|
@@ -10,11 +10,7 @@ describe("local-apps", () => {
|
|
|
10
10
|
};
|
|
11
11
|
const snippet = snippetFunc(model);
|
|
12
12
|
expect(snippet[0].content).toEqual(`# Load and run the model:
|
|
13
|
-
llama-cli
|
|
14
|
-
--hf-repo "bartowski/Llama-3.2-3B-Instruct-GGUF" \\
|
|
15
|
-
--hf-file {{GGUF_FILE}} \\
|
|
16
|
-
-p "You are a helpful assistant" \\
|
|
17
|
-
--conversation`);
|
|
13
|
+
llama-cli -hf bartowski/Llama-3.2-3B-Instruct-GGUF`);
|
|
18
14
|
});
|
|
19
15
|
it("llama.cpp non-conversational", async () => {
|
|
20
16
|
const { snippet: snippetFunc } = LOCAL_APPS["llama.cpp"];
|
|
@@ -25,9 +21,7 @@ llama-cli \\
|
|
|
25
21
|
};
|
|
26
22
|
const snippet = snippetFunc(model);
|
|
27
23
|
expect(snippet[0].content).toEqual(`# Load and run the model:
|
|
28
|
-
llama-cli \\
|
|
29
|
-
--hf-repo "mlabonne/gemma-2b-GGUF" \\
|
|
30
|
-
--hf-file {{GGUF_FILE}} \\
|
|
24
|
+
llama-cli -hf mlabonne/gemma-2b-GGUF \\
|
|
31
25
|
-p "Once upon a time,"`);
|
|
32
26
|
});
|
|
33
27
|
it("vLLM conversational llm", async () => {
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/audio-to-audio/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,
|
|
1
|
+
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/audio-to-audio/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,cA6Df,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|