@huggingface/tasks 0.0.3 → 0.0.4
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/LICENSE +21 -0
- package/README.md +20 -0
- package/dist/index.d.ts +358 -46
- package/dist/index.js +103 -41
- package/dist/{index.cjs → index.mjs} +73 -68
- package/package.json +43 -33
- package/src/Types.ts +49 -43
- package/src/audio-classification/about.md +5 -5
- package/src/audio-classification/data.ts +11 -11
- package/src/audio-to-audio/about.md +4 -3
- package/src/audio-to-audio/data.ts +18 -15
- package/src/automatic-speech-recognition/about.md +5 -4
- package/src/automatic-speech-recognition/data.ts +18 -17
- package/src/const.ts +52 -44
- package/src/conversational/about.md +9 -9
- package/src/conversational/data.ts +22 -18
- package/src/depth-estimation/about.md +1 -3
- package/src/depth-estimation/data.ts +11 -11
- package/src/document-question-answering/about.md +1 -2
- package/src/document-question-answering/data.ts +22 -19
- package/src/feature-extraction/about.md +2 -3
- package/src/feature-extraction/data.ts +12 -15
- package/src/fill-mask/about.md +1 -1
- package/src/fill-mask/data.ts +16 -14
- package/src/image-classification/about.md +5 -3
- package/src/image-classification/data.ts +15 -15
- package/src/image-segmentation/about.md +4 -4
- package/src/image-segmentation/data.ts +26 -23
- package/src/image-to-image/about.md +8 -10
- package/src/image-to-image/data.ts +31 -27
- package/src/image-to-text/about.md +13 -6
- package/src/image-to-text/data.ts +20 -21
- package/src/index.ts +2 -0
- package/src/modelLibraries.ts +43 -0
- package/src/object-detection/about.md +2 -1
- package/src/object-detection/data.ts +20 -17
- package/src/pipelines.ts +608 -0
- package/src/placeholder/about.md +3 -3
- package/src/placeholder/data.ts +8 -8
- package/src/question-answering/about.md +1 -1
- package/src/question-answering/data.ts +21 -19
- package/src/reinforcement-learning/about.md +167 -176
- package/src/reinforcement-learning/data.ts +75 -78
- package/src/sentence-similarity/data.ts +29 -28
- package/src/summarization/about.md +6 -5
- package/src/summarization/data.ts +23 -20
- package/src/table-question-answering/about.md +5 -5
- package/src/table-question-answering/data.ts +35 -39
- package/src/tabular-classification/about.md +4 -6
- package/src/tabular-classification/data.ts +11 -12
- package/src/tabular-regression/about.md +14 -18
- package/src/tabular-regression/data.ts +10 -11
- package/src/tasksData.ts +47 -50
- package/src/text-classification/about.md +5 -4
- package/src/text-classification/data.ts +21 -20
- package/src/text-generation/about.md +7 -6
- package/src/text-generation/data.ts +36 -34
- package/src/text-to-image/about.md +19 -18
- package/src/text-to-image/data.ts +32 -26
- package/src/text-to-speech/about.md +4 -5
- package/src/text-to-speech/data.ts +16 -17
- package/src/text-to-video/about.md +41 -36
- package/src/text-to-video/data.ts +43 -38
- package/src/token-classification/about.md +1 -3
- package/src/token-classification/data.ts +26 -25
- package/src/translation/about.md +4 -4
- package/src/translation/data.ts +21 -21
- package/src/unconditional-image-generation/about.md +10 -5
- package/src/unconditional-image-generation/data.ts +26 -20
- package/src/video-classification/about.md +5 -1
- package/src/video-classification/data.ts +14 -14
- package/src/visual-question-answering/about.md +8 -3
- package/src/visual-question-answering/data.ts +22 -19
- package/src/zero-shot-classification/about.md +5 -4
- package/src/zero-shot-classification/data.ts +20 -20
- package/src/zero-shot-image-classification/about.md +17 -9
- package/src/zero-shot-image-classification/data.ts +12 -14
- package/tsconfig.json +18 -0
- package/assets/audio-classification/audio.wav +0 -0
- package/assets/audio-to-audio/input.wav +0 -0
- package/assets/audio-to-audio/label-0.wav +0 -0
- package/assets/audio-to-audio/label-1.wav +0 -0
- package/assets/automatic-speech-recognition/input.flac +0 -0
- package/assets/automatic-speech-recognition/wav2vec2.png +0 -0
- package/assets/contribution-guide/anatomy.png +0 -0
- package/assets/contribution-guide/libraries.png +0 -0
- package/assets/depth-estimation/depth-estimation-input.jpg +0 -0
- package/assets/depth-estimation/depth-estimation-output.png +0 -0
- package/assets/document-question-answering/document-question-answering-input.png +0 -0
- package/assets/image-classification/image-classification-input.jpeg +0 -0
- package/assets/image-segmentation/image-segmentation-input.jpeg +0 -0
- package/assets/image-segmentation/image-segmentation-output.png +0 -0
- package/assets/image-to-image/image-to-image-input.jpeg +0 -0
- package/assets/image-to-image/image-to-image-output.png +0 -0
- package/assets/image-to-image/pix2pix_examples.jpg +0 -0
- package/assets/image-to-text/savanna.jpg +0 -0
- package/assets/object-detection/object-detection-input.jpg +0 -0
- package/assets/object-detection/object-detection-output.jpg +0 -0
- package/assets/table-question-answering/tableQA.jpg +0 -0
- package/assets/text-to-image/image.jpeg +0 -0
- package/assets/text-to-speech/audio.wav +0 -0
- package/assets/text-to-video/text-to-video-output.gif +0 -0
- package/assets/unconditional-image-generation/unconditional-image-generation-output.jpeg +0 -0
- package/assets/video-classification/video-classification-input.gif +0 -0
- package/assets/visual-question-answering/elephant.jpeg +0 -0
- package/assets/zero-shot-image-classification/image-classification-input.jpeg +0 -0
- package/dist/index.d.cts +0 -145
package/dist/index.js
CHANGED
|
@@ -1,8 +1,35 @@
|
|
|
1
|
-
|
|
2
|
-
|
|
3
|
-
|
|
4
|
-
|
|
5
|
-
var
|
|
1
|
+
"use strict";
|
|
2
|
+
var __defProp = Object.defineProperty;
|
|
3
|
+
var __getOwnPropDesc = Object.getOwnPropertyDescriptor;
|
|
4
|
+
var __getOwnPropNames = Object.getOwnPropertyNames;
|
|
5
|
+
var __hasOwnProp = Object.prototype.hasOwnProperty;
|
|
6
|
+
var __export = (target, all) => {
|
|
7
|
+
for (var name in all)
|
|
8
|
+
__defProp(target, name, { get: all[name], enumerable: true });
|
|
9
|
+
};
|
|
10
|
+
var __copyProps = (to, from, except, desc) => {
|
|
11
|
+
if (from && typeof from === "object" || typeof from === "function") {
|
|
12
|
+
for (let key of __getOwnPropNames(from))
|
|
13
|
+
if (!__hasOwnProp.call(to, key) && key !== except)
|
|
14
|
+
__defProp(to, key, { get: () => from[key], enumerable: !(desc = __getOwnPropDesc(from, key)) || desc.enumerable });
|
|
15
|
+
}
|
|
16
|
+
return to;
|
|
17
|
+
};
|
|
18
|
+
var __toCommonJS = (mod) => __copyProps(__defProp({}, "__esModule", { value: true }), mod);
|
|
19
|
+
|
|
20
|
+
// src/index.ts
|
|
21
|
+
var src_exports = {};
|
|
22
|
+
__export(src_exports, {
|
|
23
|
+
MODALITIES: () => MODALITIES,
|
|
24
|
+
ModelLibrary: () => ModelLibrary,
|
|
25
|
+
PIPELINE_DATA: () => PIPELINE_DATA,
|
|
26
|
+
TASKS_DATA: () => TASKS_DATA
|
|
27
|
+
});
|
|
28
|
+
module.exports = __toCommonJS(src_exports);
|
|
29
|
+
|
|
30
|
+
// src/pipelines.ts
|
|
31
|
+
var MODALITIES = ["cv", "nlp", "audio", "tabular", "multimodal", "rl", "other"];
|
|
32
|
+
var PIPELINE_DATA = {
|
|
6
33
|
"text-classification": {
|
|
7
34
|
name: "Text Classification",
|
|
8
35
|
subtasks: [
|
|
@@ -138,12 +165,12 @@ var PIPELINE_DATA = ensureRecordOfPipelines({
|
|
|
138
165
|
modality: "nlp",
|
|
139
166
|
color: "yellow"
|
|
140
167
|
},
|
|
141
|
-
|
|
168
|
+
translation: {
|
|
142
169
|
name: "Translation",
|
|
143
170
|
modality: "nlp",
|
|
144
171
|
color: "green"
|
|
145
172
|
},
|
|
146
|
-
|
|
173
|
+
summarization: {
|
|
147
174
|
name: "Summarization",
|
|
148
175
|
subtasks: [
|
|
149
176
|
{
|
|
@@ -158,7 +185,7 @@ var PIPELINE_DATA = ensureRecordOfPipelines({
|
|
|
158
185
|
modality: "nlp",
|
|
159
186
|
color: "indigo"
|
|
160
187
|
},
|
|
161
|
-
|
|
188
|
+
conversational: {
|
|
162
189
|
name: "Conversational",
|
|
163
190
|
subtasks: [
|
|
164
191
|
{
|
|
@@ -386,7 +413,7 @@ var PIPELINE_DATA = ensureRecordOfPipelines({
|
|
|
386
413
|
modality: "rl",
|
|
387
414
|
color: "red"
|
|
388
415
|
},
|
|
389
|
-
|
|
416
|
+
robotics: {
|
|
390
417
|
name: "Robotics",
|
|
391
418
|
modality: "rl",
|
|
392
419
|
subtasks: [
|
|
@@ -539,19 +566,14 @@ var PIPELINE_DATA = ensureRecordOfPipelines({
|
|
|
539
566
|
modality: "multimodal",
|
|
540
567
|
color: "green"
|
|
541
568
|
},
|
|
542
|
-
|
|
569
|
+
other: {
|
|
543
570
|
name: "Other",
|
|
544
571
|
modality: "other",
|
|
545
572
|
color: "blue",
|
|
546
573
|
hideInModels: true,
|
|
547
574
|
hideInDatasets: true
|
|
548
575
|
}
|
|
549
|
-
}
|
|
550
|
-
var ALL_PIPELINE_TYPES = Object.keys(PIPELINE_DATA);
|
|
551
|
-
var ALL_PIPELINE_TYPES_SET = new Set(ALL_PIPELINE_TYPES);
|
|
552
|
-
var ALL_SUBTASKS = Object.values(PIPELINE_DATA).flatMap((data) => data.subtasks ?? []);
|
|
553
|
-
var ALL_SUBTASK_TYPES = ALL_SUBTASKS.map((s) => s.type);
|
|
554
|
-
var ALL_SUBTASK_TYPES_SET = new Set(ALL_SUBTASK_TYPES);
|
|
576
|
+
};
|
|
555
577
|
|
|
556
578
|
// src/audio-classification/data.ts
|
|
557
579
|
var taskData = {
|
|
@@ -1839,19 +1861,9 @@ var taskData19 = {
|
|
|
1839
1861
|
],
|
|
1840
1862
|
type: "tabular"
|
|
1841
1863
|
},
|
|
1842
|
-
{
|
|
1843
|
-
label: "Question",
|
|
1844
|
-
content: "What is the number of reigns for Harley Race?",
|
|
1845
|
-
type: "text"
|
|
1846
|
-
}
|
|
1864
|
+
{ label: "Question", content: "What is the number of reigns for Harley Race?", type: "text" }
|
|
1847
1865
|
],
|
|
1848
|
-
outputs: [
|
|
1849
|
-
{
|
|
1850
|
-
label: "Result",
|
|
1851
|
-
content: "7",
|
|
1852
|
-
type: "text"
|
|
1853
|
-
}
|
|
1854
|
-
]
|
|
1866
|
+
outputs: [{ label: "Result", content: "7", type: "text" }]
|
|
1855
1867
|
},
|
|
1856
1868
|
metrics: [
|
|
1857
1869
|
{
|
|
@@ -2978,7 +2990,7 @@ var TASKS_MODEL_LIBRARIES = {
|
|
|
2978
2990
|
"audio-classification": ["speechbrain", "transformers"],
|
|
2979
2991
|
"audio-to-audio": ["asteroid", "speechbrain"],
|
|
2980
2992
|
"automatic-speech-recognition": ["espnet", "nemo", "speechbrain", "transformers", "transformers.js"],
|
|
2981
|
-
|
|
2993
|
+
conversational: ["transformers"],
|
|
2982
2994
|
"depth-estimation": ["transformers"],
|
|
2983
2995
|
"document-question-answering": ["transformers"],
|
|
2984
2996
|
"feature-extraction": ["sentence-transformers", "transformers", "transformers.js"],
|
|
@@ -2991,12 +3003,12 @@ var TASKS_MODEL_LIBRARIES = {
|
|
|
2991
3003
|
"video-classification": [],
|
|
2992
3004
|
"multiple-choice": ["transformers"],
|
|
2993
3005
|
"object-detection": ["transformers", "transformers.js"],
|
|
2994
|
-
|
|
3006
|
+
other: [],
|
|
2995
3007
|
"question-answering": ["adapter-transformers", "allennlp", "transformers", "transformers.js"],
|
|
2996
|
-
|
|
3008
|
+
robotics: [],
|
|
2997
3009
|
"reinforcement-learning": ["transformers", "stable-baselines3", "ml-agents", "sample-factory"],
|
|
2998
3010
|
"sentence-similarity": ["sentence-transformers", "spacy", "transformers.js"],
|
|
2999
|
-
|
|
3011
|
+
summarization: ["transformers", "transformers.js"],
|
|
3000
3012
|
"table-question-answering": ["transformers"],
|
|
3001
3013
|
"table-to-text": ["transformers"],
|
|
3002
3014
|
"tabular-classification": ["sklearn"],
|
|
@@ -3011,8 +3023,16 @@ var TASKS_MODEL_LIBRARIES = {
|
|
|
3011
3023
|
"text-to-video": [],
|
|
3012
3024
|
"text2text-generation": ["transformers", "transformers.js"],
|
|
3013
3025
|
"time-series-forecasting": [],
|
|
3014
|
-
"token-classification": [
|
|
3015
|
-
|
|
3026
|
+
"token-classification": [
|
|
3027
|
+
"adapter-transformers",
|
|
3028
|
+
"flair",
|
|
3029
|
+
"spacy",
|
|
3030
|
+
"span-marker",
|
|
3031
|
+
"stanza",
|
|
3032
|
+
"transformers",
|
|
3033
|
+
"transformers.js"
|
|
3034
|
+
],
|
|
3035
|
+
translation: ["transformers", "transformers.js"],
|
|
3016
3036
|
"unconditional-image-generation": [],
|
|
3017
3037
|
"visual-question-answering": [],
|
|
3018
3038
|
"voice-activity-detection": [],
|
|
@@ -3025,7 +3045,7 @@ var TASKS_DATA = {
|
|
|
3025
3045
|
"audio-classification": getData("audio-classification", data_default),
|
|
3026
3046
|
"audio-to-audio": getData("audio-to-audio", data_default2),
|
|
3027
3047
|
"automatic-speech-recognition": getData("automatic-speech-recognition", data_default3),
|
|
3028
|
-
|
|
3048
|
+
conversational: getData("conversational", data_default4),
|
|
3029
3049
|
"depth-estimation": getData("depth-estimation", data_default13),
|
|
3030
3050
|
"document-question-answering": getData("document-question-answering", data_default5),
|
|
3031
3051
|
"feature-extraction": getData("feature-extraction", data_default6),
|
|
@@ -3038,12 +3058,12 @@ var TASKS_DATA = {
|
|
|
3038
3058
|
"multiple-choice": void 0,
|
|
3039
3059
|
"object-detection": getData("object-detection", data_default12),
|
|
3040
3060
|
"video-classification": getData("video-classification", data_default30),
|
|
3041
|
-
|
|
3061
|
+
other: void 0,
|
|
3042
3062
|
"question-answering": getData("question-answering", data_default16),
|
|
3043
3063
|
"reinforcement-learning": getData("reinforcement-learning", data_default15),
|
|
3044
|
-
|
|
3064
|
+
robotics: void 0,
|
|
3045
3065
|
"sentence-similarity": getData("sentence-similarity", data_default17),
|
|
3046
|
-
|
|
3066
|
+
summarization: getData("summarization", data_default18),
|
|
3047
3067
|
"table-question-answering": getData("table-question-answering", data_default19),
|
|
3048
3068
|
"table-to-text": void 0,
|
|
3049
3069
|
"tabular-classification": getData("tabular-classification", data_default20),
|
|
@@ -3059,7 +3079,7 @@ var TASKS_DATA = {
|
|
|
3059
3079
|
"text2text-generation": void 0,
|
|
3060
3080
|
"time-series-forecasting": void 0,
|
|
3061
3081
|
"token-classification": getData("token-classification", data_default24),
|
|
3062
|
-
|
|
3082
|
+
translation: getData("translation", data_default25),
|
|
3063
3083
|
"unconditional-image-generation": getData("unconditional-image-generation", data_default29),
|
|
3064
3084
|
"visual-question-answering": getData("visual-question-answering", data_default31),
|
|
3065
3085
|
"voice-activity-detection": void 0,
|
|
@@ -3074,6 +3094,48 @@ function getData(type, partialTaskData = data_default14) {
|
|
|
3074
3094
|
libraries: TASKS_MODEL_LIBRARIES[type]
|
|
3075
3095
|
};
|
|
3076
3096
|
}
|
|
3077
|
-
|
|
3097
|
+
|
|
3098
|
+
// src/modelLibraries.ts
|
|
3099
|
+
var ModelLibrary = /* @__PURE__ */ ((ModelLibrary2) => {
|
|
3100
|
+
ModelLibrary2["adapter-transformers"] = "Adapter Transformers";
|
|
3101
|
+
ModelLibrary2["allennlp"] = "allenNLP";
|
|
3102
|
+
ModelLibrary2["asteroid"] = "Asteroid";
|
|
3103
|
+
ModelLibrary2["bertopic"] = "BERTopic";
|
|
3104
|
+
ModelLibrary2["diffusers"] = "Diffusers";
|
|
3105
|
+
ModelLibrary2["doctr"] = "docTR";
|
|
3106
|
+
ModelLibrary2["espnet"] = "ESPnet";
|
|
3107
|
+
ModelLibrary2["fairseq"] = "Fairseq";
|
|
3108
|
+
ModelLibrary2["flair"] = "Flair";
|
|
3109
|
+
ModelLibrary2["keras"] = "Keras";
|
|
3110
|
+
ModelLibrary2["k2"] = "K2";
|
|
3111
|
+
ModelLibrary2["nemo"] = "NeMo";
|
|
3112
|
+
ModelLibrary2["open_clip"] = "OpenCLIP";
|
|
3113
|
+
ModelLibrary2["paddlenlp"] = "PaddleNLP";
|
|
3114
|
+
ModelLibrary2["peft"] = "PEFT";
|
|
3115
|
+
ModelLibrary2["pyannote-audio"] = "pyannote.audio";
|
|
3116
|
+
ModelLibrary2["sample-factory"] = "Sample Factory";
|
|
3117
|
+
ModelLibrary2["sentence-transformers"] = "Sentence Transformers";
|
|
3118
|
+
ModelLibrary2["sklearn"] = "Scikit-learn";
|
|
3119
|
+
ModelLibrary2["spacy"] = "spaCy";
|
|
3120
|
+
ModelLibrary2["span-marker"] = "SpanMarker";
|
|
3121
|
+
ModelLibrary2["speechbrain"] = "speechbrain";
|
|
3122
|
+
ModelLibrary2["tensorflowtts"] = "TensorFlowTTS";
|
|
3123
|
+
ModelLibrary2["timm"] = "Timm";
|
|
3124
|
+
ModelLibrary2["fastai"] = "fastai";
|
|
3125
|
+
ModelLibrary2["transformers"] = "Transformers";
|
|
3126
|
+
ModelLibrary2["transformers.js"] = "Transformers.js";
|
|
3127
|
+
ModelLibrary2["stanza"] = "Stanza";
|
|
3128
|
+
ModelLibrary2["fasttext"] = "fastText";
|
|
3129
|
+
ModelLibrary2["stable-baselines3"] = "Stable-Baselines3";
|
|
3130
|
+
ModelLibrary2["ml-agents"] = "ML-Agents";
|
|
3131
|
+
ModelLibrary2["pythae"] = "Pythae";
|
|
3132
|
+
ModelLibrary2["mindspore"] = "MindSpore";
|
|
3133
|
+
return ModelLibrary2;
|
|
3134
|
+
})(ModelLibrary || {});
|
|
3135
|
+
// Annotate the CommonJS export names for ESM import in node:
|
|
3136
|
+
0 && (module.exports = {
|
|
3137
|
+
MODALITIES,
|
|
3138
|
+
ModelLibrary,
|
|
3139
|
+
PIPELINE_DATA,
|
|
3078
3140
|
TASKS_DATA
|
|
3079
|
-
};
|
|
3141
|
+
});
|
|
@@ -1,34 +1,6 @@
|
|
|
1
|
-
|
|
2
|
-
var
|
|
3
|
-
var
|
|
4
|
-
var __getOwnPropNames = Object.getOwnPropertyNames;
|
|
5
|
-
var __hasOwnProp = Object.prototype.hasOwnProperty;
|
|
6
|
-
var __export = (target, all) => {
|
|
7
|
-
for (var name in all)
|
|
8
|
-
__defProp(target, name, { get: all[name], enumerable: true });
|
|
9
|
-
};
|
|
10
|
-
var __copyProps = (to, from, except, desc) => {
|
|
11
|
-
if (from && typeof from === "object" || typeof from === "function") {
|
|
12
|
-
for (let key of __getOwnPropNames(from))
|
|
13
|
-
if (!__hasOwnProp.call(to, key) && key !== except)
|
|
14
|
-
__defProp(to, key, { get: () => from[key], enumerable: !(desc = __getOwnPropDesc(from, key)) || desc.enumerable });
|
|
15
|
-
}
|
|
16
|
-
return to;
|
|
17
|
-
};
|
|
18
|
-
var __toCommonJS = (mod) => __copyProps(__defProp({}, "__esModule", { value: true }), mod);
|
|
19
|
-
|
|
20
|
-
// src/index.ts
|
|
21
|
-
var src_exports = {};
|
|
22
|
-
__export(src_exports, {
|
|
23
|
-
TASKS_DATA: () => TASKS_DATA
|
|
24
|
-
});
|
|
25
|
-
module.exports = __toCommonJS(src_exports);
|
|
26
|
-
|
|
27
|
-
// ../js/src/lib/interfaces/Types.ts
|
|
28
|
-
function ensureRecordOfPipelines(record) {
|
|
29
|
-
return record;
|
|
30
|
-
}
|
|
31
|
-
var PIPELINE_DATA = ensureRecordOfPipelines({
|
|
1
|
+
// src/pipelines.ts
|
|
2
|
+
var MODALITIES = ["cv", "nlp", "audio", "tabular", "multimodal", "rl", "other"];
|
|
3
|
+
var PIPELINE_DATA = {
|
|
32
4
|
"text-classification": {
|
|
33
5
|
name: "Text Classification",
|
|
34
6
|
subtasks: [
|
|
@@ -164,12 +136,12 @@ var PIPELINE_DATA = ensureRecordOfPipelines({
|
|
|
164
136
|
modality: "nlp",
|
|
165
137
|
color: "yellow"
|
|
166
138
|
},
|
|
167
|
-
|
|
139
|
+
translation: {
|
|
168
140
|
name: "Translation",
|
|
169
141
|
modality: "nlp",
|
|
170
142
|
color: "green"
|
|
171
143
|
},
|
|
172
|
-
|
|
144
|
+
summarization: {
|
|
173
145
|
name: "Summarization",
|
|
174
146
|
subtasks: [
|
|
175
147
|
{
|
|
@@ -184,7 +156,7 @@ var PIPELINE_DATA = ensureRecordOfPipelines({
|
|
|
184
156
|
modality: "nlp",
|
|
185
157
|
color: "indigo"
|
|
186
158
|
},
|
|
187
|
-
|
|
159
|
+
conversational: {
|
|
188
160
|
name: "Conversational",
|
|
189
161
|
subtasks: [
|
|
190
162
|
{
|
|
@@ -412,7 +384,7 @@ var PIPELINE_DATA = ensureRecordOfPipelines({
|
|
|
412
384
|
modality: "rl",
|
|
413
385
|
color: "red"
|
|
414
386
|
},
|
|
415
|
-
|
|
387
|
+
robotics: {
|
|
416
388
|
name: "Robotics",
|
|
417
389
|
modality: "rl",
|
|
418
390
|
subtasks: [
|
|
@@ -565,19 +537,14 @@ var PIPELINE_DATA = ensureRecordOfPipelines({
|
|
|
565
537
|
modality: "multimodal",
|
|
566
538
|
color: "green"
|
|
567
539
|
},
|
|
568
|
-
|
|
540
|
+
other: {
|
|
569
541
|
name: "Other",
|
|
570
542
|
modality: "other",
|
|
571
543
|
color: "blue",
|
|
572
544
|
hideInModels: true,
|
|
573
545
|
hideInDatasets: true
|
|
574
546
|
}
|
|
575
|
-
}
|
|
576
|
-
var ALL_PIPELINE_TYPES = Object.keys(PIPELINE_DATA);
|
|
577
|
-
var ALL_PIPELINE_TYPES_SET = new Set(ALL_PIPELINE_TYPES);
|
|
578
|
-
var ALL_SUBTASKS = Object.values(PIPELINE_DATA).flatMap((data) => data.subtasks ?? []);
|
|
579
|
-
var ALL_SUBTASK_TYPES = ALL_SUBTASKS.map((s) => s.type);
|
|
580
|
-
var ALL_SUBTASK_TYPES_SET = new Set(ALL_SUBTASK_TYPES);
|
|
547
|
+
};
|
|
581
548
|
|
|
582
549
|
// src/audio-classification/data.ts
|
|
583
550
|
var taskData = {
|
|
@@ -1865,19 +1832,9 @@ var taskData19 = {
|
|
|
1865
1832
|
],
|
|
1866
1833
|
type: "tabular"
|
|
1867
1834
|
},
|
|
1868
|
-
{
|
|
1869
|
-
label: "Question",
|
|
1870
|
-
content: "What is the number of reigns for Harley Race?",
|
|
1871
|
-
type: "text"
|
|
1872
|
-
}
|
|
1835
|
+
{ label: "Question", content: "What is the number of reigns for Harley Race?", type: "text" }
|
|
1873
1836
|
],
|
|
1874
|
-
outputs: [
|
|
1875
|
-
{
|
|
1876
|
-
label: "Result",
|
|
1877
|
-
content: "7",
|
|
1878
|
-
type: "text"
|
|
1879
|
-
}
|
|
1880
|
-
]
|
|
1837
|
+
outputs: [{ label: "Result", content: "7", type: "text" }]
|
|
1881
1838
|
},
|
|
1882
1839
|
metrics: [
|
|
1883
1840
|
{
|
|
@@ -3004,7 +2961,7 @@ var TASKS_MODEL_LIBRARIES = {
|
|
|
3004
2961
|
"audio-classification": ["speechbrain", "transformers"],
|
|
3005
2962
|
"audio-to-audio": ["asteroid", "speechbrain"],
|
|
3006
2963
|
"automatic-speech-recognition": ["espnet", "nemo", "speechbrain", "transformers", "transformers.js"],
|
|
3007
|
-
|
|
2964
|
+
conversational: ["transformers"],
|
|
3008
2965
|
"depth-estimation": ["transformers"],
|
|
3009
2966
|
"document-question-answering": ["transformers"],
|
|
3010
2967
|
"feature-extraction": ["sentence-transformers", "transformers", "transformers.js"],
|
|
@@ -3017,12 +2974,12 @@ var TASKS_MODEL_LIBRARIES = {
|
|
|
3017
2974
|
"video-classification": [],
|
|
3018
2975
|
"multiple-choice": ["transformers"],
|
|
3019
2976
|
"object-detection": ["transformers", "transformers.js"],
|
|
3020
|
-
|
|
2977
|
+
other: [],
|
|
3021
2978
|
"question-answering": ["adapter-transformers", "allennlp", "transformers", "transformers.js"],
|
|
3022
|
-
|
|
2979
|
+
robotics: [],
|
|
3023
2980
|
"reinforcement-learning": ["transformers", "stable-baselines3", "ml-agents", "sample-factory"],
|
|
3024
2981
|
"sentence-similarity": ["sentence-transformers", "spacy", "transformers.js"],
|
|
3025
|
-
|
|
2982
|
+
summarization: ["transformers", "transformers.js"],
|
|
3026
2983
|
"table-question-answering": ["transformers"],
|
|
3027
2984
|
"table-to-text": ["transformers"],
|
|
3028
2985
|
"tabular-classification": ["sklearn"],
|
|
@@ -3037,8 +2994,16 @@ var TASKS_MODEL_LIBRARIES = {
|
|
|
3037
2994
|
"text-to-video": [],
|
|
3038
2995
|
"text2text-generation": ["transformers", "transformers.js"],
|
|
3039
2996
|
"time-series-forecasting": [],
|
|
3040
|
-
"token-classification": [
|
|
3041
|
-
|
|
2997
|
+
"token-classification": [
|
|
2998
|
+
"adapter-transformers",
|
|
2999
|
+
"flair",
|
|
3000
|
+
"spacy",
|
|
3001
|
+
"span-marker",
|
|
3002
|
+
"stanza",
|
|
3003
|
+
"transformers",
|
|
3004
|
+
"transformers.js"
|
|
3005
|
+
],
|
|
3006
|
+
translation: ["transformers", "transformers.js"],
|
|
3042
3007
|
"unconditional-image-generation": [],
|
|
3043
3008
|
"visual-question-answering": [],
|
|
3044
3009
|
"voice-activity-detection": [],
|
|
@@ -3051,7 +3016,7 @@ var TASKS_DATA = {
|
|
|
3051
3016
|
"audio-classification": getData("audio-classification", data_default),
|
|
3052
3017
|
"audio-to-audio": getData("audio-to-audio", data_default2),
|
|
3053
3018
|
"automatic-speech-recognition": getData("automatic-speech-recognition", data_default3),
|
|
3054
|
-
|
|
3019
|
+
conversational: getData("conversational", data_default4),
|
|
3055
3020
|
"depth-estimation": getData("depth-estimation", data_default13),
|
|
3056
3021
|
"document-question-answering": getData("document-question-answering", data_default5),
|
|
3057
3022
|
"feature-extraction": getData("feature-extraction", data_default6),
|
|
@@ -3064,12 +3029,12 @@ var TASKS_DATA = {
|
|
|
3064
3029
|
"multiple-choice": void 0,
|
|
3065
3030
|
"object-detection": getData("object-detection", data_default12),
|
|
3066
3031
|
"video-classification": getData("video-classification", data_default30),
|
|
3067
|
-
|
|
3032
|
+
other: void 0,
|
|
3068
3033
|
"question-answering": getData("question-answering", data_default16),
|
|
3069
3034
|
"reinforcement-learning": getData("reinforcement-learning", data_default15),
|
|
3070
|
-
|
|
3035
|
+
robotics: void 0,
|
|
3071
3036
|
"sentence-similarity": getData("sentence-similarity", data_default17),
|
|
3072
|
-
|
|
3037
|
+
summarization: getData("summarization", data_default18),
|
|
3073
3038
|
"table-question-answering": getData("table-question-answering", data_default19),
|
|
3074
3039
|
"table-to-text": void 0,
|
|
3075
3040
|
"tabular-classification": getData("tabular-classification", data_default20),
|
|
@@ -3085,7 +3050,7 @@ var TASKS_DATA = {
|
|
|
3085
3050
|
"text2text-generation": void 0,
|
|
3086
3051
|
"time-series-forecasting": void 0,
|
|
3087
3052
|
"token-classification": getData("token-classification", data_default24),
|
|
3088
|
-
|
|
3053
|
+
translation: getData("translation", data_default25),
|
|
3089
3054
|
"unconditional-image-generation": getData("unconditional-image-generation", data_default29),
|
|
3090
3055
|
"visual-question-answering": getData("visual-question-answering", data_default31),
|
|
3091
3056
|
"voice-activity-detection": void 0,
|
|
@@ -3100,7 +3065,47 @@ function getData(type, partialTaskData = data_default14) {
|
|
|
3100
3065
|
libraries: TASKS_MODEL_LIBRARIES[type]
|
|
3101
3066
|
};
|
|
3102
3067
|
}
|
|
3103
|
-
|
|
3104
|
-
|
|
3068
|
+
|
|
3069
|
+
// src/modelLibraries.ts
|
|
3070
|
+
var ModelLibrary = /* @__PURE__ */ ((ModelLibrary2) => {
|
|
3071
|
+
ModelLibrary2["adapter-transformers"] = "Adapter Transformers";
|
|
3072
|
+
ModelLibrary2["allennlp"] = "allenNLP";
|
|
3073
|
+
ModelLibrary2["asteroid"] = "Asteroid";
|
|
3074
|
+
ModelLibrary2["bertopic"] = "BERTopic";
|
|
3075
|
+
ModelLibrary2["diffusers"] = "Diffusers";
|
|
3076
|
+
ModelLibrary2["doctr"] = "docTR";
|
|
3077
|
+
ModelLibrary2["espnet"] = "ESPnet";
|
|
3078
|
+
ModelLibrary2["fairseq"] = "Fairseq";
|
|
3079
|
+
ModelLibrary2["flair"] = "Flair";
|
|
3080
|
+
ModelLibrary2["keras"] = "Keras";
|
|
3081
|
+
ModelLibrary2["k2"] = "K2";
|
|
3082
|
+
ModelLibrary2["nemo"] = "NeMo";
|
|
3083
|
+
ModelLibrary2["open_clip"] = "OpenCLIP";
|
|
3084
|
+
ModelLibrary2["paddlenlp"] = "PaddleNLP";
|
|
3085
|
+
ModelLibrary2["peft"] = "PEFT";
|
|
3086
|
+
ModelLibrary2["pyannote-audio"] = "pyannote.audio";
|
|
3087
|
+
ModelLibrary2["sample-factory"] = "Sample Factory";
|
|
3088
|
+
ModelLibrary2["sentence-transformers"] = "Sentence Transformers";
|
|
3089
|
+
ModelLibrary2["sklearn"] = "Scikit-learn";
|
|
3090
|
+
ModelLibrary2["spacy"] = "spaCy";
|
|
3091
|
+
ModelLibrary2["span-marker"] = "SpanMarker";
|
|
3092
|
+
ModelLibrary2["speechbrain"] = "speechbrain";
|
|
3093
|
+
ModelLibrary2["tensorflowtts"] = "TensorFlowTTS";
|
|
3094
|
+
ModelLibrary2["timm"] = "Timm";
|
|
3095
|
+
ModelLibrary2["fastai"] = "fastai";
|
|
3096
|
+
ModelLibrary2["transformers"] = "Transformers";
|
|
3097
|
+
ModelLibrary2["transformers.js"] = "Transformers.js";
|
|
3098
|
+
ModelLibrary2["stanza"] = "Stanza";
|
|
3099
|
+
ModelLibrary2["fasttext"] = "fastText";
|
|
3100
|
+
ModelLibrary2["stable-baselines3"] = "Stable-Baselines3";
|
|
3101
|
+
ModelLibrary2["ml-agents"] = "ML-Agents";
|
|
3102
|
+
ModelLibrary2["pythae"] = "Pythae";
|
|
3103
|
+
ModelLibrary2["mindspore"] = "MindSpore";
|
|
3104
|
+
return ModelLibrary2;
|
|
3105
|
+
})(ModelLibrary || {});
|
|
3106
|
+
export {
|
|
3107
|
+
MODALITIES,
|
|
3108
|
+
ModelLibrary,
|
|
3109
|
+
PIPELINE_DATA,
|
|
3105
3110
|
TASKS_DATA
|
|
3106
|
-
}
|
|
3111
|
+
};
|
package/package.json
CHANGED
|
@@ -1,34 +1,44 @@
|
|
|
1
1
|
{
|
|
2
|
-
|
|
3
|
-
|
|
4
|
-
|
|
5
|
-
|
|
6
|
-
|
|
7
|
-
|
|
8
|
-
|
|
9
|
-
|
|
10
|
-
|
|
11
|
-
|
|
12
|
-
|
|
13
|
-
|
|
14
|
-
|
|
15
|
-
|
|
16
|
-
|
|
17
|
-
|
|
18
|
-
|
|
19
|
-
|
|
20
|
-
|
|
21
|
-
|
|
22
|
-
|
|
23
|
-
|
|
24
|
-
|
|
25
|
-
|
|
26
|
-
|
|
27
|
-
|
|
28
|
-
|
|
29
|
-
|
|
30
|
-
|
|
31
|
-
|
|
32
|
-
|
|
33
|
-
|
|
34
|
-
|
|
2
|
+
"name": "@huggingface/tasks",
|
|
3
|
+
"packageManager": "pnpm@8.3.1",
|
|
4
|
+
"version": "0.0.4",
|
|
5
|
+
"description": "List of ISO-639 languages used in the Hub",
|
|
6
|
+
"repository": "https://github.com/huggingface/huggingface.js.git",
|
|
7
|
+
"publishConfig": {
|
|
8
|
+
"access": "public"
|
|
9
|
+
},
|
|
10
|
+
"main": "./dist/index.js",
|
|
11
|
+
"module": "./dist/index.mjs",
|
|
12
|
+
"types": "./dist/index.d.ts",
|
|
13
|
+
"exports": {
|
|
14
|
+
".": {
|
|
15
|
+
"types": "./dist/index.d.ts",
|
|
16
|
+
"require": "./dist/index.js",
|
|
17
|
+
"import": "./dist/index.mjs"
|
|
18
|
+
}
|
|
19
|
+
},
|
|
20
|
+
"source": "src/index.ts",
|
|
21
|
+
"files": [
|
|
22
|
+
"dist",
|
|
23
|
+
"src",
|
|
24
|
+
"tsconfig.json"
|
|
25
|
+
],
|
|
26
|
+
"keywords": [
|
|
27
|
+
"huggingface",
|
|
28
|
+
"hub",
|
|
29
|
+
"languages"
|
|
30
|
+
],
|
|
31
|
+
"author": "Hugging Face",
|
|
32
|
+
"license": "MIT",
|
|
33
|
+
"devDependencies": {
|
|
34
|
+
"typescript": "^5.0.4"
|
|
35
|
+
},
|
|
36
|
+
"scripts": {
|
|
37
|
+
"lint": "eslint --quiet --fix --ext .cjs,.ts .",
|
|
38
|
+
"lint:check": "eslint --ext .cjs,.ts .",
|
|
39
|
+
"format": "prettier --write .",
|
|
40
|
+
"format:check": "prettier --check .",
|
|
41
|
+
"build": "tsup src/index.ts --format cjs,esm --clean --dts",
|
|
42
|
+
"type-check": "tsc"
|
|
43
|
+
}
|
|
44
|
+
}
|
package/src/Types.ts
CHANGED
|
@@ -1,58 +1,64 @@
|
|
|
1
|
-
import type { ModelLibraryKey } from "
|
|
2
|
-
import type { PipelineType } from "
|
|
1
|
+
import type { ModelLibraryKey } from "./modelLibraries";
|
|
2
|
+
import type { PipelineType } from "./pipelines";
|
|
3
3
|
|
|
4
4
|
export interface ExampleRepo {
|
|
5
5
|
description: string;
|
|
6
|
-
id:
|
|
6
|
+
id: string;
|
|
7
7
|
}
|
|
8
8
|
|
|
9
|
-
export type TaskDemoEntry =
|
|
10
|
-
|
|
11
|
-
|
|
12
|
-
|
|
13
|
-
|
|
14
|
-
|
|
15
|
-
|
|
16
|
-
|
|
17
|
-
|
|
18
|
-
}
|
|
19
|
-
|
|
20
|
-
|
|
21
|
-
|
|
22
|
-
|
|
23
|
-
|
|
24
|
-
}
|
|
25
|
-
|
|
26
|
-
|
|
27
|
-
|
|
28
|
-
}
|
|
29
|
-
|
|
30
|
-
|
|
31
|
-
|
|
32
|
-
|
|
33
|
-
|
|
34
|
-
|
|
35
|
-
|
|
36
|
-
|
|
9
|
+
export type TaskDemoEntry =
|
|
10
|
+
| {
|
|
11
|
+
filename: string;
|
|
12
|
+
type: "audio";
|
|
13
|
+
}
|
|
14
|
+
| {
|
|
15
|
+
data: Array<{
|
|
16
|
+
label: string;
|
|
17
|
+
score: number;
|
|
18
|
+
}>;
|
|
19
|
+
type: "chart";
|
|
20
|
+
}
|
|
21
|
+
| {
|
|
22
|
+
filename: string;
|
|
23
|
+
type: "img";
|
|
24
|
+
}
|
|
25
|
+
| {
|
|
26
|
+
table: string[][];
|
|
27
|
+
type: "tabular";
|
|
28
|
+
}
|
|
29
|
+
| {
|
|
30
|
+
content: string;
|
|
31
|
+
label: string;
|
|
32
|
+
type: "text";
|
|
33
|
+
}
|
|
34
|
+
| {
|
|
35
|
+
text: string;
|
|
36
|
+
tokens: Array<{
|
|
37
|
+
end: number;
|
|
38
|
+
start: number;
|
|
39
|
+
type: string;
|
|
40
|
+
}>;
|
|
41
|
+
type: "text-with-tokens";
|
|
42
|
+
};
|
|
37
43
|
|
|
38
44
|
export interface TaskDemo {
|
|
39
|
-
inputs:
|
|
45
|
+
inputs: TaskDemoEntry[];
|
|
40
46
|
outputs: TaskDemoEntry[];
|
|
41
47
|
}
|
|
42
48
|
|
|
43
49
|
export interface TaskData {
|
|
44
|
-
datasets:
|
|
45
|
-
demo:
|
|
46
|
-
id:
|
|
50
|
+
datasets: ExampleRepo[];
|
|
51
|
+
demo: TaskDemo;
|
|
52
|
+
id: PipelineType;
|
|
47
53
|
isPlaceholder?: boolean;
|
|
48
|
-
label:
|
|
49
|
-
libraries:
|
|
50
|
-
metrics:
|
|
51
|
-
models:
|
|
52
|
-
spaces:
|
|
53
|
-
summary:
|
|
54
|
-
widgetModels:
|
|
55
|
-
youtubeId?:
|
|
54
|
+
label: string;
|
|
55
|
+
libraries: ModelLibraryKey[];
|
|
56
|
+
metrics: ExampleRepo[];
|
|
57
|
+
models: ExampleRepo[];
|
|
58
|
+
spaces: ExampleRepo[];
|
|
59
|
+
summary: string;
|
|
60
|
+
widgetModels: string[];
|
|
61
|
+
youtubeId?: string;
|
|
56
62
|
}
|
|
57
63
|
|
|
58
64
|
export type TaskDataCustom = Omit<TaskData, "id" | "label" | "libraries">;
|