@huggingface/tasks 0.19.8 → 0.19.9
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/commonjs/hardware.d.ts +7 -0
- package/dist/commonjs/hardware.d.ts.map +1 -1
- package/dist/commonjs/hardware.js +7 -0
- package/dist/commonjs/local-apps.js +10 -10
- package/dist/commonjs/local-apps.spec.js +2 -3
- package/dist/commonjs/model-libraries.d.ts +8 -1
- package/dist/commonjs/model-libraries.d.ts.map +1 -1
- package/dist/commonjs/model-libraries.js +8 -1
- package/dist/commonjs/snippets/inputs.d.ts.map +1 -1
- package/dist/commonjs/snippets/inputs.js +5 -0
- package/dist/commonjs/tasks/image-to-video/data.d.ts +4 -0
- package/dist/commonjs/tasks/image-to-video/data.d.ts.map +1 -0
- package/dist/commonjs/tasks/image-to-video/data.js +119 -0
- package/dist/commonjs/tasks/image-to-video/inference.d.ts +75 -0
- package/dist/commonjs/tasks/image-to-video/inference.d.ts.map +1 -0
- package/dist/commonjs/tasks/image-to-video/inference.js +2 -0
- package/dist/commonjs/tasks/index.d.ts +1 -0
- package/dist/commonjs/tasks/index.d.ts.map +1 -1
- package/dist/commonjs/tasks/index.js +63 -62
- package/dist/esm/hardware.d.ts +7 -0
- package/dist/esm/hardware.d.ts.map +1 -1
- package/dist/esm/hardware.js +7 -0
- package/dist/esm/local-apps.js +10 -10
- package/dist/esm/local-apps.spec.js +2 -3
- package/dist/esm/model-libraries.d.ts +8 -1
- package/dist/esm/model-libraries.d.ts.map +1 -1
- package/dist/esm/model-libraries.js +8 -1
- package/dist/esm/snippets/inputs.d.ts.map +1 -1
- package/dist/esm/snippets/inputs.js +5 -0
- package/dist/esm/tasks/image-to-video/data.d.ts +4 -0
- package/dist/esm/tasks/image-to-video/data.d.ts.map +1 -0
- package/dist/esm/tasks/image-to-video/data.js +117 -0
- package/dist/esm/tasks/image-to-video/inference.d.ts +75 -0
- package/dist/esm/tasks/image-to-video/inference.d.ts.map +1 -0
- package/dist/esm/tasks/image-to-video/inference.js +1 -0
- package/dist/esm/tasks/index.d.ts +1 -0
- package/dist/esm/tasks/index.d.ts.map +1 -1
- package/dist/esm/tasks/index.js +2 -1
- package/package.json +1 -1
- package/src/hardware.ts +7 -0
- package/src/local-apps.spec.ts +2 -3
- package/src/local-apps.ts +10 -10
- package/src/model-libraries.ts +8 -1
- package/src/snippets/inputs.ts +6 -0
- package/src/tasks/image-text-to-text/about.md +2 -1
- package/src/tasks/image-to-video/about.md +51 -0
- package/src/tasks/image-to-video/data.ts +126 -0
- package/src/tasks/image-to-video/inference.ts +74 -0
- package/src/tasks/image-to-video/spec/input.json +64 -0
- package/src/tasks/image-to-video/spec/output.json +13 -0
- package/src/tasks/index.ts +3 -1
|
@@ -96,6 +96,10 @@ export declare const SKUS: {
|
|
|
96
96
|
tflops: number;
|
|
97
97
|
memory: number[];
|
|
98
98
|
};
|
|
99
|
+
"RTX A5000": {
|
|
100
|
+
tflops: number;
|
|
101
|
+
memory: number[];
|
|
102
|
+
};
|
|
99
103
|
"RTX A4000": {
|
|
100
104
|
tflops: number;
|
|
101
105
|
memory: number[];
|
|
@@ -529,6 +533,9 @@ export declare const SKUS: {
|
|
|
529
533
|
"Intel Core Ultra 7 265KF": {
|
|
530
534
|
tflops: number;
|
|
531
535
|
};
|
|
536
|
+
"Intel Core 14th Generation (i7)": {
|
|
537
|
+
tflops: number;
|
|
538
|
+
};
|
|
532
539
|
"Intel Core 13th Generation (i9)": {
|
|
533
540
|
tflops: number;
|
|
534
541
|
};
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"hardware.d.ts","sourceRoot":"","sources":["../../src/hardware.ts"],"names":[],"mappings":"AAAA;;;GAGG;AACH,eAAO,MAAM,iDAAiD,QAAW,CAAC;AAC1E,eAAO,MAAM,yDAAyD,QAAW,CAAC;AAClF,eAAO,MAAM,oCAAoC,QAAU,CAAC;AAE5D;;;GAGG;AACH,eAAO,MAAM,+CAA+C,QAAW,CAAC;AAExE,MAAM,WAAW,YAAY;IAC5B;;;;;;;;;OASG;IACH,MAAM,EAAE,MAAM,CAAC;IACf;;;OAGG;IACH,MAAM,CAAC,EAAE,MAAM,EAAE,CAAC;CAClB;AAED,eAAO,MAAM,sBAAsB,UAAqD,CAAC;AAEzF,eAAO,MAAM,IAAI
|
|
1
|
+
{"version":3,"file":"hardware.d.ts","sourceRoot":"","sources":["../../src/hardware.ts"],"names":[],"mappings":"AAAA;;;GAGG;AACH,eAAO,MAAM,iDAAiD,QAAW,CAAC;AAC1E,eAAO,MAAM,yDAAyD,QAAW,CAAC;AAClF,eAAO,MAAM,oCAAoC,QAAU,CAAC;AAE5D;;;GAGG;AACH,eAAO,MAAM,+CAA+C,QAAW,CAAC;AAExE,MAAM,WAAW,YAAY;IAC5B;;;;;;;;;OASG;IACH,MAAM,EAAE,MAAM,CAAC;IACf;;;OAGG;IACH,MAAM,CAAC,EAAE,MAAM,EAAE,CAAC;CAClB;AAED,eAAO,MAAM,sBAAsB,UAAqD,CAAC;AAEzF,eAAO,MAAM,IAAI;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CAwqBuD,CAAC;AAEzE,MAAM,MAAM,OAAO,GAAG,MAAM,OAAO,IAAI,CAAC"}
|
|
@@ -81,6 +81,10 @@ exports.SKUS = {
|
|
|
81
81
|
tflops: 38.7,
|
|
82
82
|
memory: [48],
|
|
83
83
|
},
|
|
84
|
+
"RTX A5000": {
|
|
85
|
+
tflops: 34.1,
|
|
86
|
+
memory: [24],
|
|
87
|
+
},
|
|
84
88
|
"RTX A4000": {
|
|
85
89
|
tflops: 19.2,
|
|
86
90
|
memory: [16],
|
|
@@ -514,6 +518,9 @@ exports.SKUS = {
|
|
|
514
518
|
"Intel Core Ultra 7 265KF": {
|
|
515
519
|
tflops: 1.53,
|
|
516
520
|
},
|
|
521
|
+
"Intel Core 14th Generation (i7)": {
|
|
522
|
+
tflops: 0.8,
|
|
523
|
+
},
|
|
517
524
|
"Intel Core 13th Generation (i9)": {
|
|
518
525
|
tflops: 0.85,
|
|
519
526
|
},
|
|
@@ -39,18 +39,18 @@ function getQuantTag(filepath) {
|
|
|
39
39
|
const snippetLlamacpp = (model, filepath) => {
|
|
40
40
|
const command = (binary) => {
|
|
41
41
|
const snippet = ["# Load and run the model:", `${binary} -hf ${model.id}${getQuantTag(filepath)}`];
|
|
42
|
-
if (!model.tags.includes("conversational")) {
|
|
43
|
-
// for non-conversational models, add a prompt
|
|
44
|
-
snippet[snippet.length - 1] += " \\";
|
|
45
|
-
snippet.push(' -p "Once upon a time,"');
|
|
46
|
-
}
|
|
47
42
|
return snippet.join("\n");
|
|
48
43
|
};
|
|
49
44
|
return [
|
|
50
45
|
{
|
|
51
46
|
title: "Install from brew",
|
|
52
47
|
setup: "brew install llama.cpp",
|
|
53
|
-
content: command("llama-
|
|
48
|
+
content: command("llama-server"),
|
|
49
|
+
},
|
|
50
|
+
{
|
|
51
|
+
title: "Install from WinGet (Windows)",
|
|
52
|
+
setup: "winget install llama.cpp",
|
|
53
|
+
content: command("llama-server"),
|
|
54
54
|
},
|
|
55
55
|
{
|
|
56
56
|
title: "Use pre-built binary",
|
|
@@ -59,17 +59,17 @@ const snippetLlamacpp = (model, filepath) => {
|
|
|
59
59
|
"# Download pre-built binary from:",
|
|
60
60
|
"# https://github.com/ggerganov/llama.cpp/releases",
|
|
61
61
|
].join("\n"),
|
|
62
|
-
content: command("./llama-
|
|
62
|
+
content: command("./llama-server"),
|
|
63
63
|
},
|
|
64
64
|
{
|
|
65
65
|
title: "Build from source code",
|
|
66
66
|
setup: [
|
|
67
67
|
"git clone https://github.com/ggerganov/llama.cpp.git",
|
|
68
68
|
"cd llama.cpp",
|
|
69
|
-
"cmake -B build
|
|
70
|
-
"cmake --build build -j --target llama-
|
|
69
|
+
"cmake -B build",
|
|
70
|
+
"cmake --build build -j --target llama-server",
|
|
71
71
|
].join("\n"),
|
|
72
|
-
content: command("./build/bin/llama-
|
|
72
|
+
content: command("./build/bin/llama-server"),
|
|
73
73
|
},
|
|
74
74
|
];
|
|
75
75
|
};
|
|
@@ -12,7 +12,7 @@ const local_apps_js_1 = require("./local-apps.js");
|
|
|
12
12
|
};
|
|
13
13
|
const snippet = snippetFunc(model);
|
|
14
14
|
(0, vitest_1.expect)(snippet[0].content).toEqual(`# Load and run the model:
|
|
15
|
-
llama-
|
|
15
|
+
llama-server -hf bartowski/Llama-3.2-3B-Instruct-GGUF:{{QUANT_TAG}}`);
|
|
16
16
|
});
|
|
17
17
|
(0, vitest_1.it)("llama.cpp non-conversational", async () => {
|
|
18
18
|
const { snippet: snippetFunc } = local_apps_js_1.LOCAL_APPS["llama.cpp"];
|
|
@@ -23,8 +23,7 @@ llama-cli -hf bartowski/Llama-3.2-3B-Instruct-GGUF:{{QUANT_TAG}}`);
|
|
|
23
23
|
};
|
|
24
24
|
const snippet = snippetFunc(model);
|
|
25
25
|
(0, vitest_1.expect)(snippet[0].content).toEqual(`# Load and run the model:
|
|
26
|
-
llama-
|
|
27
|
-
-p "Once upon a time,"`);
|
|
26
|
+
llama-server -hf mlabonne/gemma-2b-GGUF:{{QUANT_TAG}}`);
|
|
28
27
|
});
|
|
29
28
|
(0, vitest_1.it)("vLLM conversational llm", async () => {
|
|
30
29
|
const { snippet: snippetFunc } = local_apps_js_1.LOCAL_APPS["vllm"];
|
|
@@ -112,6 +112,13 @@ export declare const MODEL_LIBRARIES_UI_ELEMENTS: {
|
|
|
112
112
|
countDownloads: string;
|
|
113
113
|
snippets: (model: ModelData) => string[];
|
|
114
114
|
};
|
|
115
|
+
"bagel-mot": {
|
|
116
|
+
prettyLabel: string;
|
|
117
|
+
repoName: string;
|
|
118
|
+
repoUrl: string;
|
|
119
|
+
filter: false;
|
|
120
|
+
countDownloads: string;
|
|
121
|
+
};
|
|
115
122
|
ben2: {
|
|
116
123
|
prettyLabel: string;
|
|
117
124
|
repoName: string;
|
|
@@ -1048,5 +1055,5 @@ export declare const MODEL_LIBRARIES_UI_ELEMENTS: {
|
|
|
1048
1055
|
};
|
|
1049
1056
|
export type ModelLibraryKey = keyof typeof MODEL_LIBRARIES_UI_ELEMENTS;
|
|
1050
1057
|
export declare const ALL_MODEL_LIBRARY_KEYS: ModelLibraryKey[];
|
|
1051
|
-
export declare const ALL_DISPLAY_MODEL_LIBRARY_KEYS: ("adapter-transformers" | "allennlp" | "anemoi" | "araclip" | "asteroid" | "audiocraft" | "audioseal" | "ben2" | "bertopic" | "big_vision" | "birder" | "birefnet" | "bm25s" | "champ" | "chat_tts" | "colpali" | "comet" | "cosmos" | "cxr-foundation" | "deepforest" | "depth-anything-v2" | "depth-pro" | "derm-foundation" | "describe-anything" | "dia-tts" | "diffree" | "diffusers" | "diffusionkit" | "doctr" | "cartesia_pytorch" | "cartesia_mlx" | "clipscope" | "cosyvoice" | "cotracker" | "edsnlp" | "elm" | "espnet" | "fairseq" | "fastai" | "fasttext" | "flair" | "gemma.cpp" | "geometry-crafter" | "gliner" | "glyph-byt5" | "grok" | "hallo" | "hezar" | "htrflow" | "hunyuan-dit" | "hunyuan3d-2" | "imstoucan" | "index-tts" | "infinite-you" | "keras" | "tf-keras" | "keras-hub" | "kimi-audio" | "k2" | "lightning-ir" | "liveportrait" | "llama-cpp-python" | "mini-omni2" | "mindspore" | "mamba-ssm" | "mars5-tts" | "matanyone" | "mesh-anything" | "merlin" | "medvae" | "mitie" | "ml-agents" | "mlx" | "mlx-image" | "mlc-llm" | "model2vec" | "moshi" | "nemo" | "open-oasis" | "open_clip" | "open-sora" | "outetts" | "paddlenlp" | "peft" | "perception-encoder" | "phantom-wan" | "pxia" | "pyannote-audio" | "py-feat" | "pythae" | "recurrentgemma" | "relik" | "refiners" | "reverb" | "saelens" | "sam2" | "sample-factory" | "sapiens" | "sentence-transformers" | "setfit" | "sklearn" | "spacy" | "span-marker" | "speechbrain" | "ssr-speech" | "stable-audio-tools" | "diffusion-single-file" | "seed-story" | "soloaudio" | "stable-baselines3" | "stanza" | "swarmformer" | "f5-tts" | "genmo" | "tensorflowtts" | "tabpfn" | "terratorch" | "tic-clip" | "timesfm" | "timm" | "torchgeo" | "transformers" | "transformers.js" | "trellis" | "ultralytics" | "uni-3dar" | "unity-sentis" | "sana" | "vfi-mamba" | "voicecraft" | "wham" | "whisperkit" | "yolov10" | "zonos" | "3dtopia-xl")[];
|
|
1058
|
+
export declare const ALL_DISPLAY_MODEL_LIBRARY_KEYS: ("adapter-transformers" | "allennlp" | "anemoi" | "araclip" | "asteroid" | "audiocraft" | "audioseal" | "bagel-mot" | "ben2" | "bertopic" | "big_vision" | "birder" | "birefnet" | "bm25s" | "champ" | "chat_tts" | "colpali" | "comet" | "cosmos" | "cxr-foundation" | "deepforest" | "depth-anything-v2" | "depth-pro" | "derm-foundation" | "describe-anything" | "dia-tts" | "diffree" | "diffusers" | "diffusionkit" | "doctr" | "cartesia_pytorch" | "cartesia_mlx" | "clipscope" | "cosyvoice" | "cotracker" | "edsnlp" | "elm" | "espnet" | "fairseq" | "fastai" | "fasttext" | "flair" | "gemma.cpp" | "geometry-crafter" | "gliner" | "glyph-byt5" | "grok" | "hallo" | "hezar" | "htrflow" | "hunyuan-dit" | "hunyuan3d-2" | "imstoucan" | "index-tts" | "infinite-you" | "keras" | "tf-keras" | "keras-hub" | "kimi-audio" | "k2" | "lightning-ir" | "liveportrait" | "llama-cpp-python" | "mini-omni2" | "mindspore" | "mamba-ssm" | "mars5-tts" | "matanyone" | "mesh-anything" | "merlin" | "medvae" | "mitie" | "ml-agents" | "mlx" | "mlx-image" | "mlc-llm" | "model2vec" | "moshi" | "nemo" | "open-oasis" | "open_clip" | "open-sora" | "outetts" | "paddlenlp" | "peft" | "perception-encoder" | "phantom-wan" | "pxia" | "pyannote-audio" | "py-feat" | "pythae" | "recurrentgemma" | "relik" | "refiners" | "reverb" | "saelens" | "sam2" | "sample-factory" | "sapiens" | "sentence-transformers" | "setfit" | "sklearn" | "spacy" | "span-marker" | "speechbrain" | "ssr-speech" | "stable-audio-tools" | "diffusion-single-file" | "seed-story" | "soloaudio" | "stable-baselines3" | "stanza" | "swarmformer" | "f5-tts" | "genmo" | "tensorflowtts" | "tabpfn" | "terratorch" | "tic-clip" | "timesfm" | "timm" | "torchgeo" | "transformers" | "transformers.js" | "trellis" | "ultralytics" | "uni-3dar" | "unity-sentis" | "sana" | "vfi-mamba" | "voicecraft" | "wham" | "whisperkit" | "yolov10" | "zonos" | "3dtopia-xl")[];
|
|
1052
1059
|
//# sourceMappingURL=model-libraries.d.ts.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"model-libraries.d.ts","sourceRoot":"","sources":["../../src/model-libraries.ts"],"names":[],"mappings":"AACA,OAAO,KAAK,EAAE,SAAS,EAAE,MAAM,iBAAiB,CAAC;AACjD,OAAO,KAAK,EAAE,kBAAkB,EAAE,MAAM,gCAAgC,CAAC;AAEzE;;GAEG;AACH,MAAM,WAAW,gBAAgB;IAChC;;;;OAIG;IACH,WAAW,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,QAAQ,EAAE,MAAM,CAAC;IACjB;;OAEG;IACH,OAAO,EAAE,MAAM,CAAC;IAChB;;OAEG;IACH,OAAO,CAAC,EAAE,MAAM,CAAC;IACjB;;OAEG;IACH,QAAQ,CAAC,EAAE,CAAC,KAAK,EAAE,SAAS,KAAK,MAAM,EAAE,CAAC;IAC1C;;;;;OAKG;IACH,cAAc,CAAC,EAAE,kBAAkB,CAAC;IACpC;;;OAGG;IACH,MAAM,CAAC,EAAE,OAAO,CAAC;CACjB;AAED;;;;;;;;;;;;;GAaG;AAEH,eAAO,MAAM,2BAA2B
|
|
1
|
+
{"version":3,"file":"model-libraries.d.ts","sourceRoot":"","sources":["../../src/model-libraries.ts"],"names":[],"mappings":"AACA,OAAO,KAAK,EAAE,SAAS,EAAE,MAAM,iBAAiB,CAAC;AACjD,OAAO,KAAK,EAAE,kBAAkB,EAAE,MAAM,gCAAgC,CAAC;AAEzE;;GAEG;AACH,MAAM,WAAW,gBAAgB;IAChC;;;;OAIG;IACH,WAAW,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,QAAQ,EAAE,MAAM,CAAC;IACjB;;OAEG;IACH,OAAO,EAAE,MAAM,CAAC;IAChB;;OAEG;IACH,OAAO,CAAC,EAAE,MAAM,CAAC;IACjB;;OAEG;IACH,QAAQ,CAAC,EAAE,CAAC,KAAK,EAAE,SAAS,KAAK,MAAM,EAAE,CAAC;IAC1C;;;;;OAKG;IACH,cAAc,CAAC,EAAE,kBAAkB,CAAC;IACpC;;;OAGG;IACH,MAAM,CAAC,EAAE,OAAO,CAAC;CACjB;AAED;;;;;;;;;;;;;GAaG;AAEH,eAAO,MAAM,2BAA2B;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CA6+BI,CAAC;AAE7C,MAAM,MAAM,eAAe,GAAG,MAAM,OAAO,2BAA2B,CAAC;AAEvE,eAAO,MAAM,sBAAsB,EAA+C,eAAe,EAAE,CAAC;AAEpG,eAAO,MAAM,8BAA8B,+1DAQ1B,CAAC"}
|
|
@@ -108,6 +108,13 @@ exports.MODEL_LIBRARIES_UI_ELEMENTS = {
|
|
|
108
108
|
countDownloads: `path_extension:"pth"`,
|
|
109
109
|
snippets: snippets.audioseal,
|
|
110
110
|
},
|
|
111
|
+
"bagel-mot": {
|
|
112
|
+
prettyLabel: "Bagel",
|
|
113
|
+
repoName: "Bagel",
|
|
114
|
+
repoUrl: "https://github.com/ByteDance-Seed/Bagel/",
|
|
115
|
+
filter: false,
|
|
116
|
+
countDownloads: `path:"llm_config.json"`,
|
|
117
|
+
},
|
|
111
118
|
ben2: {
|
|
112
119
|
prettyLabel: "BEN2",
|
|
113
120
|
repoName: "BEN2",
|
|
@@ -921,7 +928,7 @@ exports.MODEL_LIBRARIES_UI_ELEMENTS = {
|
|
|
921
928
|
repoName: "timesfm",
|
|
922
929
|
repoUrl: "https://github.com/google-research/timesfm",
|
|
923
930
|
filter: false,
|
|
924
|
-
countDownloads: `path:"checkpoints/checkpoint_1100000/state/checkpoint"`,
|
|
931
|
+
countDownloads: `path:"checkpoints/checkpoint_1100000/state/checkpoint" OR path:"checkpoints/checkpoint_2150000/state/checkpoint" OR path_extension:"ckpt"`,
|
|
925
932
|
},
|
|
926
933
|
timm: {
|
|
927
934
|
prettyLabel: "timm",
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"inputs.d.ts","sourceRoot":"","sources":["../../../src/snippets/inputs.ts"],"names":[],"mappings":"AACA,OAAO,KAAK,EAAE,0BAA0B,EAAE,MAAM,mBAAmB,CAAC;AACpE,OAAO,KAAK,EAAE,gBAAgB,EAAE,MAAM,YAAY,CAAC;
|
|
1
|
+
{"version":3,"file":"inputs.d.ts","sourceRoot":"","sources":["../../../src/snippets/inputs.ts"],"names":[],"mappings":"AACA,OAAO,KAAK,EAAE,0BAA0B,EAAE,MAAM,mBAAmB,CAAC;AACpE,OAAO,KAAK,EAAE,gBAAgB,EAAE,MAAM,YAAY,CAAC;AA4JnD,wBAAgB,oBAAoB,CACnC,KAAK,EAAE,gBAAgB,EACvB,MAAM,UAAQ,EACd,QAAQ,UAAQ,GACd,MAAM,GAAG,0BAA0B,EAAE,CAmBvC"}
|
|
@@ -68,6 +68,10 @@ const inputsImageToImage = () => `{
|
|
|
68
68
|
"image": "cat.png",
|
|
69
69
|
"prompt": "Turn the cat into a tiger."
|
|
70
70
|
}`;
|
|
71
|
+
const inputsImageToVideo = () => `{
|
|
72
|
+
"image": "cat.png",
|
|
73
|
+
"prompt": "The cat starts to dance"
|
|
74
|
+
}`;
|
|
71
75
|
const inputsImageSegmentation = () => `"cats.jpg"`;
|
|
72
76
|
const inputsObjectDetection = () => `"cats.jpg"`;
|
|
73
77
|
const inputsAudioToAudio = () => `"sample1.flac"`;
|
|
@@ -89,6 +93,7 @@ const modelInputSnippets = {
|
|
|
89
93
|
"image-classification": inputsImageClassification,
|
|
90
94
|
"image-to-text": inputsImageToText,
|
|
91
95
|
"image-to-image": inputsImageToImage,
|
|
96
|
+
"image-to-video": inputsImageToVideo,
|
|
92
97
|
"image-segmentation": inputsImageSegmentation,
|
|
93
98
|
"object-detection": inputsObjectDetection,
|
|
94
99
|
"question-answering": inputsQuestionAnswering,
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/image-to-video/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,cAyHf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
|
@@ -0,0 +1,119 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
3
|
+
const taskData = {
|
|
4
|
+
datasets: [
|
|
5
|
+
{
|
|
6
|
+
description: "A benchmark dataset for reference image controlled video generation.",
|
|
7
|
+
id: "ali-vilab/VACE-Benchmark",
|
|
8
|
+
},
|
|
9
|
+
{
|
|
10
|
+
description: "A dataset of video generation style preferences.",
|
|
11
|
+
id: "Rapidata/sora-video-generation-style-likert-scoring",
|
|
12
|
+
},
|
|
13
|
+
{
|
|
14
|
+
description: "A dataset with videos and captions throughout the videos.",
|
|
15
|
+
id: "BestWishYsh/ChronoMagic",
|
|
16
|
+
},
|
|
17
|
+
],
|
|
18
|
+
demo: {
|
|
19
|
+
inputs: [
|
|
20
|
+
{
|
|
21
|
+
filename: "image-to-video-input.jpg",
|
|
22
|
+
type: "img",
|
|
23
|
+
},
|
|
24
|
+
{
|
|
25
|
+
label: "Optional Text Prompt",
|
|
26
|
+
content: "This penguin is dancing",
|
|
27
|
+
type: "text",
|
|
28
|
+
},
|
|
29
|
+
],
|
|
30
|
+
outputs: [
|
|
31
|
+
{
|
|
32
|
+
filename: "image-to-video-output.gif",
|
|
33
|
+
type: "img",
|
|
34
|
+
},
|
|
35
|
+
],
|
|
36
|
+
},
|
|
37
|
+
metrics: [
|
|
38
|
+
{
|
|
39
|
+
description: "Fréchet Video Distance (FVD) measures the perceptual similarity between the distributions of generated videos and a set of real videos, assessing overall visual quality and temporal coherence of the video generated from an input image.",
|
|
40
|
+
id: "fvd",
|
|
41
|
+
},
|
|
42
|
+
{
|
|
43
|
+
description: "CLIP Score measures the semantic similarity between a textual prompt (if provided alongside the input image) and the generated video frames. It evaluates how well the video's generated content and motion align with the textual description, conditioned on the initial image.",
|
|
44
|
+
id: "clip_score",
|
|
45
|
+
},
|
|
46
|
+
{
|
|
47
|
+
description: "First Frame Fidelity, often measured using LPIPS (Learned Perceptual Image Patch Similarity), PSNR, or SSIM, quantifies how closely the first frame of the generated video matches the input conditioning image.",
|
|
48
|
+
id: "lpips",
|
|
49
|
+
},
|
|
50
|
+
{
|
|
51
|
+
description: "Identity Preservation Score measures the consistency of identity (e.g., a person's face or a specific object's characteristics) between the input image and throughout the generated video frames, often calculated using features from specialized models like face recognition (e.g., ArcFace) or re-identification models.",
|
|
52
|
+
id: "identity_preservation",
|
|
53
|
+
},
|
|
54
|
+
{
|
|
55
|
+
description: "Motion Score evaluates the quality, realism, and temporal consistency of motion in the video generated from a static image. This can be based on optical flow analysis (e.g., smoothness, magnitude), consistency of object trajectories, or specific motion plausibility assessments.",
|
|
56
|
+
id: "motion_score",
|
|
57
|
+
},
|
|
58
|
+
],
|
|
59
|
+
models: [
|
|
60
|
+
{
|
|
61
|
+
description: "LTX-Video, a 13B parameter model for high quality video generation",
|
|
62
|
+
id: "Lightricks/LTX-Video-0.9.7-dev",
|
|
63
|
+
},
|
|
64
|
+
{
|
|
65
|
+
description: "A 14B parameter model for reference image controlled video generation",
|
|
66
|
+
id: "Wan-AI/Wan2.1-VACE-14B",
|
|
67
|
+
},
|
|
68
|
+
{
|
|
69
|
+
description: "An image-to-video generation model using FramePack F1 methodology with Hunyuan-DiT architecture",
|
|
70
|
+
id: "lllyasviel/FramePack_F1_I2V_HY_20250503",
|
|
71
|
+
},
|
|
72
|
+
{
|
|
73
|
+
description: "A distilled version of the LTX-Video-0.9.7-dev model for faster inference",
|
|
74
|
+
id: "Lightricks/LTX-Video-0.9.7-distilled",
|
|
75
|
+
},
|
|
76
|
+
{
|
|
77
|
+
description: "An image-to-video generation model by Skywork AI, 14B parameters, producing 720p videos.",
|
|
78
|
+
id: "Skywork/SkyReels-V2-I2V-14B-720P",
|
|
79
|
+
},
|
|
80
|
+
{
|
|
81
|
+
description: "Image-to-video variant of Tencent's HunyuanVideo.",
|
|
82
|
+
id: "tencent/HunyuanVideo-I2V",
|
|
83
|
+
},
|
|
84
|
+
{
|
|
85
|
+
description: "A 14B parameter model for 720p image-to-video generation by Wan-AI.",
|
|
86
|
+
id: "Wan-AI/Wan2.1-I2V-14B-720P",
|
|
87
|
+
},
|
|
88
|
+
{
|
|
89
|
+
description: "A Diffusers version of the Wan2.1-I2V-14B-720P model for 720p image-to-video generation.",
|
|
90
|
+
id: "Wan-AI/Wan2.1-I2V-14B-720P-Diffusers",
|
|
91
|
+
},
|
|
92
|
+
],
|
|
93
|
+
spaces: [
|
|
94
|
+
{
|
|
95
|
+
description: "An application to generate videos fast.",
|
|
96
|
+
id: "Lightricks/ltx-video-distilled",
|
|
97
|
+
},
|
|
98
|
+
{
|
|
99
|
+
description: "Generate videos with the FramePack-F1",
|
|
100
|
+
id: "linoyts/FramePack-F1",
|
|
101
|
+
},
|
|
102
|
+
{
|
|
103
|
+
description: "Generate videos with the FramePack",
|
|
104
|
+
id: "lisonallen/framepack-i2v",
|
|
105
|
+
},
|
|
106
|
+
{
|
|
107
|
+
description: "Wan2.1 with CausVid LoRA",
|
|
108
|
+
id: "multimodalart/wan2-1-fast",
|
|
109
|
+
},
|
|
110
|
+
{
|
|
111
|
+
description: "A demo for Stable Video Diffusion",
|
|
112
|
+
id: "multimodalart/stable-video-diffusion",
|
|
113
|
+
},
|
|
114
|
+
],
|
|
115
|
+
summary: "Image-to-video models take a still image as input and generate a video. These models can be guided by text prompts to influence the content and style of the output video.",
|
|
116
|
+
widgetModels: [],
|
|
117
|
+
youtubeId: undefined,
|
|
118
|
+
};
|
|
119
|
+
exports.default = taskData;
|
|
@@ -0,0 +1,75 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Inference code generated from the JSON schema spec in ./spec
|
|
3
|
+
*
|
|
4
|
+
* Using src/scripts/inference-codegen
|
|
5
|
+
*/
|
|
6
|
+
/**
|
|
7
|
+
* Inputs for Image To Video inference
|
|
8
|
+
*/
|
|
9
|
+
export interface ImageToVideoInput {
|
|
10
|
+
/**
|
|
11
|
+
* The input image data as a base64-encoded string. If no `parameters` are provided, you can
|
|
12
|
+
* also provide the image data as a raw bytes payload.
|
|
13
|
+
*/
|
|
14
|
+
inputs: Blob;
|
|
15
|
+
/**
|
|
16
|
+
* Additional inference parameters for Image To Video
|
|
17
|
+
*/
|
|
18
|
+
parameters?: ImageToVideoParameters;
|
|
19
|
+
[property: string]: unknown;
|
|
20
|
+
}
|
|
21
|
+
/**
|
|
22
|
+
* Additional inference parameters for Image To Video
|
|
23
|
+
*/
|
|
24
|
+
export interface ImageToVideoParameters {
|
|
25
|
+
/**
|
|
26
|
+
* For diffusion models. A higher guidance scale value encourages the model to generate
|
|
27
|
+
* videos closely linked to the text prompt at the expense of lower image quality.
|
|
28
|
+
*/
|
|
29
|
+
guidance_scale?: number;
|
|
30
|
+
/**
|
|
31
|
+
* One prompt to guide what NOT to include in video generation.
|
|
32
|
+
*/
|
|
33
|
+
negative_prompt?: string;
|
|
34
|
+
/**
|
|
35
|
+
* The num_frames parameter determines how many video frames are generated.
|
|
36
|
+
*/
|
|
37
|
+
num_frames?: number;
|
|
38
|
+
/**
|
|
39
|
+
* The number of denoising steps. More denoising steps usually lead to a higher quality
|
|
40
|
+
* video at the expense of slower inference.
|
|
41
|
+
*/
|
|
42
|
+
num_inference_steps?: number;
|
|
43
|
+
/**
|
|
44
|
+
* The text prompt to guide the video generation.
|
|
45
|
+
*/
|
|
46
|
+
prompt?: string;
|
|
47
|
+
/**
|
|
48
|
+
* Seed for the random number generator.
|
|
49
|
+
*/
|
|
50
|
+
seed?: number;
|
|
51
|
+
/**
|
|
52
|
+
* The size in pixel of the output video frames.
|
|
53
|
+
*/
|
|
54
|
+
target_size?: TargetSize;
|
|
55
|
+
[property: string]: unknown;
|
|
56
|
+
}
|
|
57
|
+
/**
|
|
58
|
+
* The size in pixel of the output video frames.
|
|
59
|
+
*/
|
|
60
|
+
export interface TargetSize {
|
|
61
|
+
height: number;
|
|
62
|
+
width: number;
|
|
63
|
+
[property: string]: unknown;
|
|
64
|
+
}
|
|
65
|
+
/**
|
|
66
|
+
* Outputs of inference for the Image To Video task
|
|
67
|
+
*/
|
|
68
|
+
export interface ImageToVideoOutput {
|
|
69
|
+
/**
|
|
70
|
+
* The generated video returned as raw bytes in the payload.
|
|
71
|
+
*/
|
|
72
|
+
video: unknown;
|
|
73
|
+
[property: string]: unknown;
|
|
74
|
+
}
|
|
75
|
+
//# sourceMappingURL=inference.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"inference.d.ts","sourceRoot":"","sources":["../../../../src/tasks/image-to-video/inference.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AACH;;GAEG;AACH,MAAM,WAAW,iBAAiB;IACjC;;;OAGG;IACH,MAAM,EAAE,IAAI,CAAC;IACb;;OAEG;IACH,UAAU,CAAC,EAAE,sBAAsB,CAAC;IACpC,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AACD;;GAEG;AACH,MAAM,WAAW,sBAAsB;IACtC;;;OAGG;IACH,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB;;OAEG;IACH,eAAe,CAAC,EAAE,MAAM,CAAC;IACzB;;OAEG;IACH,UAAU,CAAC,EAAE,MAAM,CAAC;IACpB;;;OAGG;IACH,mBAAmB,CAAC,EAAE,MAAM,CAAC;IAC7B;;OAEG;IACH,MAAM,CAAC,EAAE,MAAM,CAAC;IAChB;;OAEG;IACH,IAAI,CAAC,EAAE,MAAM,CAAC;IACd;;OAEG;IACH,WAAW,CAAC,EAAE,UAAU,CAAC;IACzB,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AACD;;GAEG;AACH,MAAM,WAAW,UAAU;IAC1B,MAAM,EAAE,MAAM,CAAC;IACf,KAAK,EAAE,MAAM,CAAC;IACd,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B;AACD;;GAEG;AACH,MAAM,WAAW,kBAAkB;IAClC;;OAEG;IACH,KAAK,EAAE,OAAO,CAAC;IACf,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC;CAC5B"}
|
|
@@ -9,6 +9,7 @@ export type { ImageClassificationInput, ImageClassificationOutput, ImageClassifi
|
|
|
9
9
|
export type * from "./image-to-image/inference.js";
|
|
10
10
|
export type { ImageToTextInput, ImageToTextOutput, ImageToTextParameters } from "./image-to-text/inference.js";
|
|
11
11
|
export type * from "./image-segmentation/inference.js";
|
|
12
|
+
export type { ImageToVideoInput, ImageToVideoOutput, ImageToVideoParameters } from "./image-to-video/inference.js";
|
|
12
13
|
export type * from "./object-detection/inference.js";
|
|
13
14
|
export type * from "./depth-estimation/inference.js";
|
|
14
15
|
export type * from "./question-answering/inference.js";
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../../../src/tasks/index.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,YAAY,EAAE,MAAM,iBAAiB,CAAC;
|
|
1
|
+
{"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../../../src/tasks/index.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,YAAY,EAAE,MAAM,iBAAiB,CAAC;AAgDpD,mBAAmB,qCAAqC,CAAC;AACzD,mBAAmB,6CAA6C,CAAC;AACjE,YAAY,EACX,mBAAmB,EACnB,0BAA0B,EAC1B,oBAAoB,EACpB,4BAA4B,EAC5B,2BAA2B,EAC3B,0BAA0B,EAC1B,gCAAgC,EAChC,+BAA+B,GAC/B,MAAM,gCAAgC,CAAC;AACxC,mBAAmB,4CAA4C,CAAC;AAChE,mBAAmB,mCAAmC,CAAC;AACvD,mBAAmB,0BAA0B,CAAC;AAC9C,YAAY,EACX,wBAAwB,EACxB,yBAAyB,EACzB,gCAAgC,EAChC,6BAA6B,GAC7B,MAAM,qCAAqC,CAAC;AAC7C,mBAAmB,+BAA+B,CAAC;AACnD,YAAY,EAAE,gBAAgB,EAAE,iBAAiB,EAAE,qBAAqB,EAAE,MAAM,8BAA8B,CAAC;AAC/G,mBAAmB,mCAAmC,CAAC;AACvD,YAAY,EAAE,iBAAiB,EAAE,kBAAkB,EAAE,sBAAsB,EAAE,MAAM,+BAA+B,CAAC;AACnH,mBAAmB,iCAAiC,CAAC;AACrD,mBAAmB,iCAAiC,CAAC;AACrD,mBAAmB,mCAAmC,CAAC;AACvD,mBAAmB,oCAAoC,CAAC;AACxD,mBAAmB,8BAA8B,CAAC;AAClD,mBAAmB,yCAAyC,CAAC;AAC7D,YAAY,EAAE,gBAAgB,EAAE,iBAAiB,EAAE,qBAAqB,EAAE,MAAM,8BAA8B,CAAC;AAC/G,YAAY,EAAE,qBAAqB,EAAE,iBAAiB,EAAE,gBAAgB,EAAE,MAAM,8BAA8B,CAAC;AAC/G,YAAY,EAAE,sBAAsB,EAAE,iBAAiB,EAAE,kBAAkB,EAAE,MAAM,+BAA+B,CAAC;AACnH,mBAAmB,qCAAqC,CAAC;AACzD,YAAY,EAAE,gBAAgB,EAAE,iBAAiB,EAAE,MAAM,4BAA4B,CAAC;AACtF,YAAY,EACX,6BAA6B,EAC7B,uBAAuB,EACvB,wBAAwB,EACxB,+BAA+B,EAC/B,4BAA4B,GAC5B,MAAM,oCAAoC,CAAC;AAC5C,YAAY,EACX,gCAAgC,EAChC,gCAAgC,EAChC,mBAAmB,EACnB,oBAAoB,EACpB,2BAA2B,EAC3B,qCAAqC,EACrC,kCAAkC,EAClC,yBAAyB,EACzB,uCAAuC,EACvC,0BAA0B,GAC1B,MAAM,gCAAgC,CAAC;AACxC,mBAAmB,qCAAqC,CAAC;AACzD,mBAAmB,0CAA0C,CAAC;AAC9D,mBAAmB,yCAAyC,CAAC;AAC7D,mBAAmB,+CAA+C,CAAC;AACnE,YAAY,EACX,WAAW,EACX,4BAA4B,EAC5B,6BAA6B,EAC7B,oCAAoC,GACpC,MAAM,2CAA2C,CAAC;AAEnD,OAAO,KAAK,EAAE,eAAe,EAAE,MAAM,uBAAuB,CAAC;AAE7D;;GAEG;AACH,eAAO,MAAM,qBAAqB,EAAE,MAAM,CAAC,YAAY,EAAE,eAAe,EAAE,CAgEzE,CAAC;AAoBF,eAAO,MAAM,UAAU,EAAE,MAAM,CAAC,YAAY,EAAE,QAAQ,GAAG,SAAS,CAwDxD,CAAC;AAEX,MAAM,WAAW,WAAW;IAC3B,WAAW,EAAE,MAAM,CAAC;IACpB,EAAE,EAAE,MAAM,CAAC;CACX;AAED,MAAM,MAAM,aAAa,GACtB;IACA,QAAQ,EAAE,MAAM,CAAC;IACjB,IAAI,EAAE,OAAO,CAAC;CACb,GACD;IACA,IAAI,EAAE,KAAK,CAAC;QACX,KAAK,EAAE,MAAM,CAAC;QACd,KAAK,EAAE,MAAM,CAAC;KACd,CAAC,CAAC;IACH,IAAI,EAAE,OAAO,CAAC;CACb,GACD;IACA,QAAQ,EAAE,MAAM,CAAC;IACjB,IAAI,EAAE,KAAK,CAAC;CACX,GACD;IACA,KAAK,EAAE,MAAM,EAAE,EAAE,CAAC;IAClB,IAAI,EAAE,SAAS,CAAC;CACf,GACD;IACA,OAAO,EAAE,MAAM,CAAC;IAChB,KAAK,EAAE,MAAM,CAAC;IACd,IAAI,EAAE,MAAM,CAAC;CACZ,GACD;IACA,IAAI,EAAE,MAAM,CAAC;IACb,MAAM,EAAE,KAAK,CAAC;QACb,GAAG,EAAE,MAAM,CAAC;QACZ,KAAK,EAAE,MAAM,CAAC;QACd,IAAI,EAAE,MAAM,CAAC;KACb,CAAC,CAAC;IACH,IAAI,EAAE,kBAAkB,CAAC;CACxB,CAAC;AAEL,MAAM,WAAW,QAAQ;IACxB,MAAM,EAAE,aAAa,EAAE,CAAC;IACxB,OAAO,EAAE,aAAa,EAAE,CAAC;CACzB;AAED,MAAM,WAAW,QAAQ;IACxB,QAAQ,EAAE,WAAW,EAAE,CAAC;IACxB,IAAI,EAAE,QAAQ,CAAC;IACf,EAAE,EAAE,YAAY,CAAC;IACjB,WAAW,CAAC,EAAE,YAAY,CAAC;IAC3B,aAAa,CAAC,EAAE,OAAO,CAAC;IACxB,KAAK,EAAE,MAAM,CAAC;IACd,SAAS,EAAE,eAAe,EAAE,CAAC;IAC7B,OAAO,EAAE,WAAW,EAAE,CAAC;IACvB,MAAM,EAAE,WAAW,EAAE,CAAC;IACtB,MAAM,EAAE,WAAW,EAAE,CAAC;IACtB,OAAO,EAAE,MAAM,CAAC;IAChB,YAAY,EAAE,MAAM,EAAE,CAAC;IACvB,SAAS,CAAC,EAAE,MAAM,CAAC;CACnB;AAED,MAAM,MAAM,cAAc,GAAG,IAAI,CAAC,QAAQ,EAAE,IAAI,GAAG,OAAO,GAAG,WAAW,CAAC,CAAC"}
|
|
@@ -18,36 +18,37 @@ const data_js_10 = __importDefault(require("./image-to-image/data.js"));
|
|
|
18
18
|
const data_js_11 = __importDefault(require("./image-to-text/data.js"));
|
|
19
19
|
const data_js_12 = __importDefault(require("./image-text-to-text/data.js"));
|
|
20
20
|
const data_js_13 = __importDefault(require("./image-segmentation/data.js"));
|
|
21
|
-
const data_js_14 = __importDefault(require("./
|
|
22
|
-
const data_js_15 = __importDefault(require("./
|
|
23
|
-
const data_js_16 = __importDefault(require("./
|
|
24
|
-
const data_js_17 = __importDefault(require("./
|
|
25
|
-
const data_js_18 = __importDefault(require("./
|
|
26
|
-
const data_js_19 = __importDefault(require("./
|
|
27
|
-
const data_js_20 = __importDefault(require("./
|
|
28
|
-
const data_js_21 = __importDefault(require("./
|
|
29
|
-
const data_js_22 = __importDefault(require("./
|
|
30
|
-
const data_js_23 = __importDefault(require("./
|
|
31
|
-
const data_js_24 = __importDefault(require("./tabular-
|
|
32
|
-
const data_js_25 = __importDefault(require("./
|
|
33
|
-
const data_js_26 = __importDefault(require("./text-to-
|
|
34
|
-
const data_js_27 = __importDefault(require("./
|
|
35
|
-
const data_js_28 = __importDefault(require("./
|
|
36
|
-
const data_js_29 = __importDefault(require("./
|
|
37
|
-
const data_js_30 = __importDefault(require("./text-
|
|
38
|
-
const data_js_31 = __importDefault(require("./text-
|
|
39
|
-
const data_js_32 = __importDefault(require("./text-
|
|
40
|
-
const data_js_33 = __importDefault(require("./
|
|
41
|
-
const data_js_34 = __importDefault(require("./
|
|
42
|
-
const data_js_35 = __importDefault(require("./
|
|
43
|
-
const data_js_36 = __importDefault(require("./visual-
|
|
44
|
-
const data_js_37 = __importDefault(require("./
|
|
45
|
-
const data_js_38 = __importDefault(require("./zero-shot-
|
|
46
|
-
const data_js_39 = __importDefault(require("./zero-shot-
|
|
47
|
-
const data_js_40 = __importDefault(require("./
|
|
48
|
-
const data_js_41 = __importDefault(require("./
|
|
49
|
-
const data_js_42 = __importDefault(require("./
|
|
50
|
-
const data_js_43 = __importDefault(require("./
|
|
21
|
+
const data_js_14 = __importDefault(require("./image-to-video/data.js"));
|
|
22
|
+
const data_js_15 = __importDefault(require("./mask-generation/data.js"));
|
|
23
|
+
const data_js_16 = __importDefault(require("./object-detection/data.js"));
|
|
24
|
+
const data_js_17 = __importDefault(require("./depth-estimation/data.js"));
|
|
25
|
+
const data_js_18 = __importDefault(require("./placeholder/data.js"));
|
|
26
|
+
const data_js_19 = __importDefault(require("./reinforcement-learning/data.js"));
|
|
27
|
+
const data_js_20 = __importDefault(require("./question-answering/data.js"));
|
|
28
|
+
const data_js_21 = __importDefault(require("./sentence-similarity/data.js"));
|
|
29
|
+
const data_js_22 = __importDefault(require("./summarization/data.js"));
|
|
30
|
+
const data_js_23 = __importDefault(require("./table-question-answering/data.js"));
|
|
31
|
+
const data_js_24 = __importDefault(require("./tabular-classification/data.js"));
|
|
32
|
+
const data_js_25 = __importDefault(require("./tabular-regression/data.js"));
|
|
33
|
+
const data_js_26 = __importDefault(require("./text-to-image/data.js"));
|
|
34
|
+
const data_js_27 = __importDefault(require("./text-to-speech/data.js"));
|
|
35
|
+
const data_js_28 = __importDefault(require("./token-classification/data.js"));
|
|
36
|
+
const data_js_29 = __importDefault(require("./translation/data.js"));
|
|
37
|
+
const data_js_30 = __importDefault(require("./text-classification/data.js"));
|
|
38
|
+
const data_js_31 = __importDefault(require("./text-generation/data.js"));
|
|
39
|
+
const data_js_32 = __importDefault(require("./text-ranking/data.js"));
|
|
40
|
+
const data_js_33 = __importDefault(require("./text-to-video/data.js"));
|
|
41
|
+
const data_js_34 = __importDefault(require("./unconditional-image-generation/data.js"));
|
|
42
|
+
const data_js_35 = __importDefault(require("./video-classification/data.js"));
|
|
43
|
+
const data_js_36 = __importDefault(require("./visual-document-retrieval/data.js"));
|
|
44
|
+
const data_js_37 = __importDefault(require("./visual-question-answering/data.js"));
|
|
45
|
+
const data_js_38 = __importDefault(require("./zero-shot-classification/data.js"));
|
|
46
|
+
const data_js_39 = __importDefault(require("./zero-shot-image-classification/data.js"));
|
|
47
|
+
const data_js_40 = __importDefault(require("./zero-shot-object-detection/data.js"));
|
|
48
|
+
const data_js_41 = __importDefault(require("./image-to-3d/data.js"));
|
|
49
|
+
const data_js_42 = __importDefault(require("./text-to-3d/data.js"));
|
|
50
|
+
const data_js_43 = __importDefault(require("./keypoint-detection/data.js"));
|
|
51
|
+
const data_js_44 = __importDefault(require("./video-text-to-text/data.js"));
|
|
51
52
|
/**
|
|
52
53
|
* Model libraries compatible with each ML task
|
|
53
54
|
*/
|
|
@@ -121,7 +122,7 @@ exports.TASKS_MODEL_LIBRARIES = {
|
|
|
121
122
|
* If the partialTaskData argument is left undefined,
|
|
122
123
|
* the default placeholder data will be used.
|
|
123
124
|
*/
|
|
124
|
-
function getData(type, partialTaskData =
|
|
125
|
+
function getData(type, partialTaskData = data_js_18.default) {
|
|
125
126
|
return {
|
|
126
127
|
...partialTaskData,
|
|
127
128
|
id: type,
|
|
@@ -137,11 +138,11 @@ exports.TASKS_DATA = {
|
|
|
137
138
|
"any-to-any": getData("any-to-any", data_js_1.default),
|
|
138
139
|
"audio-classification": getData("audio-classification", data_js_2.default),
|
|
139
140
|
"audio-to-audio": getData("audio-to-audio", data_js_3.default),
|
|
140
|
-
"audio-text-to-text": getData("audio-text-to-text",
|
|
141
|
+
"audio-text-to-text": getData("audio-text-to-text", data_js_18.default),
|
|
141
142
|
"automatic-speech-recognition": getData("automatic-speech-recognition", data_js_4.default),
|
|
142
|
-
"depth-estimation": getData("depth-estimation",
|
|
143
|
+
"depth-estimation": getData("depth-estimation", data_js_17.default),
|
|
143
144
|
"document-question-answering": getData("document-question-answering", data_js_5.default),
|
|
144
|
-
"visual-document-retrieval": getData("visual-document-retrieval",
|
|
145
|
+
"visual-document-retrieval": getData("visual-document-retrieval", data_js_36.default),
|
|
145
146
|
"feature-extraction": getData("feature-extraction", data_js_6.default),
|
|
146
147
|
"fill-mask": getData("fill-mask", data_js_7.default),
|
|
147
148
|
"graph-ml": undefined,
|
|
@@ -151,42 +152,42 @@ exports.TASKS_DATA = {
|
|
|
151
152
|
"image-to-image": getData("image-to-image", data_js_10.default),
|
|
152
153
|
"image-text-to-text": getData("image-text-to-text", data_js_12.default),
|
|
153
154
|
"image-to-text": getData("image-to-text", data_js_11.default),
|
|
154
|
-
"image-to-video":
|
|
155
|
-
"keypoint-detection": getData("keypoint-detection",
|
|
156
|
-
"mask-generation": getData("mask-generation",
|
|
155
|
+
"image-to-video": getData("image-to-video", data_js_14.default),
|
|
156
|
+
"keypoint-detection": getData("keypoint-detection", data_js_43.default),
|
|
157
|
+
"mask-generation": getData("mask-generation", data_js_15.default),
|
|
157
158
|
"multiple-choice": undefined,
|
|
158
|
-
"object-detection": getData("object-detection",
|
|
159
|
-
"video-classification": getData("video-classification",
|
|
159
|
+
"object-detection": getData("object-detection", data_js_16.default),
|
|
160
|
+
"video-classification": getData("video-classification", data_js_35.default),
|
|
160
161
|
other: undefined,
|
|
161
|
-
"question-answering": getData("question-answering",
|
|
162
|
-
"reinforcement-learning": getData("reinforcement-learning",
|
|
162
|
+
"question-answering": getData("question-answering", data_js_20.default),
|
|
163
|
+
"reinforcement-learning": getData("reinforcement-learning", data_js_19.default),
|
|
163
164
|
robotics: undefined,
|
|
164
|
-
"sentence-similarity": getData("sentence-similarity",
|
|
165
|
-
summarization: getData("summarization",
|
|
166
|
-
"table-question-answering": getData("table-question-answering",
|
|
165
|
+
"sentence-similarity": getData("sentence-similarity", data_js_21.default),
|
|
166
|
+
summarization: getData("summarization", data_js_22.default),
|
|
167
|
+
"table-question-answering": getData("table-question-answering", data_js_23.default),
|
|
167
168
|
"table-to-text": undefined,
|
|
168
|
-
"tabular-classification": getData("tabular-classification",
|
|
169
|
-
"tabular-regression": getData("tabular-regression",
|
|
169
|
+
"tabular-classification": getData("tabular-classification", data_js_24.default),
|
|
170
|
+
"tabular-regression": getData("tabular-regression", data_js_25.default),
|
|
170
171
|
"tabular-to-text": undefined,
|
|
171
|
-
"text-classification": getData("text-classification",
|
|
172
|
-
"text-generation": getData("text-generation",
|
|
173
|
-
"text-ranking": getData("text-ranking",
|
|
172
|
+
"text-classification": getData("text-classification", data_js_30.default),
|
|
173
|
+
"text-generation": getData("text-generation", data_js_31.default),
|
|
174
|
+
"text-ranking": getData("text-ranking", data_js_32.default),
|
|
174
175
|
"text-retrieval": undefined,
|
|
175
|
-
"text-to-image": getData("text-to-image",
|
|
176
|
-
"text-to-speech": getData("text-to-speech",
|
|
176
|
+
"text-to-image": getData("text-to-image", data_js_26.default),
|
|
177
|
+
"text-to-speech": getData("text-to-speech", data_js_27.default),
|
|
177
178
|
"text-to-audio": undefined,
|
|
178
|
-
"text-to-video": getData("text-to-video",
|
|
179
|
+
"text-to-video": getData("text-to-video", data_js_33.default),
|
|
179
180
|
"text2text-generation": undefined,
|
|
180
181
|
"time-series-forecasting": undefined,
|
|
181
|
-
"token-classification": getData("token-classification",
|
|
182
|
-
translation: getData("translation",
|
|
183
|
-
"unconditional-image-generation": getData("unconditional-image-generation",
|
|
184
|
-
"video-text-to-text": getData("video-text-to-text",
|
|
185
|
-
"visual-question-answering": getData("visual-question-answering",
|
|
182
|
+
"token-classification": getData("token-classification", data_js_28.default),
|
|
183
|
+
translation: getData("translation", data_js_29.default),
|
|
184
|
+
"unconditional-image-generation": getData("unconditional-image-generation", data_js_34.default),
|
|
185
|
+
"video-text-to-text": getData("video-text-to-text", data_js_44.default),
|
|
186
|
+
"visual-question-answering": getData("visual-question-answering", data_js_37.default),
|
|
186
187
|
"voice-activity-detection": undefined,
|
|
187
|
-
"zero-shot-classification": getData("zero-shot-classification",
|
|
188
|
-
"zero-shot-image-classification": getData("zero-shot-image-classification",
|
|
189
|
-
"zero-shot-object-detection": getData("zero-shot-object-detection",
|
|
190
|
-
"text-to-3d": getData("text-to-3d",
|
|
191
|
-
"image-to-3d": getData("image-to-3d",
|
|
188
|
+
"zero-shot-classification": getData("zero-shot-classification", data_js_38.default),
|
|
189
|
+
"zero-shot-image-classification": getData("zero-shot-image-classification", data_js_39.default),
|
|
190
|
+
"zero-shot-object-detection": getData("zero-shot-object-detection", data_js_40.default),
|
|
191
|
+
"text-to-3d": getData("text-to-3d", data_js_42.default),
|
|
192
|
+
"image-to-3d": getData("image-to-3d", data_js_41.default),
|
|
192
193
|
};
|