@huggingface/tasks 0.13.15 → 0.13.17
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/commonjs/local-apps.js +9 -9
- package/dist/commonjs/local-apps.spec.js +2 -8
- package/dist/commonjs/model-data.d.ts +3 -0
- package/dist/commonjs/model-data.d.ts.map +1 -1
- package/dist/commonjs/model-libraries-snippets.d.ts +3 -1
- package/dist/commonjs/model-libraries-snippets.d.ts.map +1 -1
- package/dist/commonjs/model-libraries-snippets.js +134 -22
- package/dist/commonjs/model-libraries.d.ts +12 -8
- package/dist/commonjs/model-libraries.d.ts.map +1 -1
- package/dist/commonjs/model-libraries.js +15 -8
- package/dist/commonjs/tasks/audio-to-audio/data.d.ts.map +1 -1
- package/dist/commonjs/tasks/audio-to-audio/data.js +0 -4
- package/dist/commonjs/tasks/fill-mask/data.js +2 -2
- package/dist/commonjs/tasks/image-classification/data.d.ts.map +1 -1
- package/dist/commonjs/tasks/image-classification/data.js +2 -3
- package/dist/commonjs/tasks/image-feature-extraction/data.d.ts.map +1 -1
- package/dist/commonjs/tasks/image-feature-extraction/data.js +8 -3
- package/dist/commonjs/tasks/image-text-to-text/data.d.ts.map +1 -1
- package/dist/commonjs/tasks/image-text-to-text/data.js +19 -11
- package/dist/commonjs/tasks/image-to-3d/data.js +4 -4
- package/dist/commonjs/tasks/image-to-image/data.d.ts.map +1 -1
- package/dist/commonjs/tasks/image-to-image/data.js +12 -4
- package/dist/commonjs/tasks/index.js +1 -1
- package/dist/commonjs/tasks/keypoint-detection/data.d.ts.map +1 -1
- package/dist/commonjs/tasks/keypoint-detection/data.js +4 -0
- package/dist/commonjs/tasks/object-detection/data.d.ts.map +1 -1
- package/dist/commonjs/tasks/object-detection/data.js +6 -2
- package/dist/commonjs/tasks/sentence-similarity/data.d.ts.map +1 -1
- package/dist/commonjs/tasks/sentence-similarity/data.js +5 -1
- package/dist/commonjs/tasks/text-generation/data.d.ts.map +1 -1
- package/dist/commonjs/tasks/text-generation/data.js +17 -13
- package/dist/commonjs/tasks/text-to-image/data.d.ts.map +1 -1
- package/dist/commonjs/tasks/text-to-image/data.js +4 -0
- package/dist/commonjs/tasks/text-to-speech/data.d.ts.map +1 -1
- package/dist/commonjs/tasks/text-to-speech/data.js +9 -9
- package/dist/commonjs/tasks/text-to-video/data.js +5 -5
- package/dist/commonjs/tasks/video-text-to-text/data.js +4 -4
- package/dist/commonjs/tasks/zero-shot-classification/data.js +2 -2
- package/dist/commonjs/tasks/zero-shot-image-classification/data.d.ts.map +1 -1
- package/dist/commonjs/tasks/zero-shot-image-classification/data.js +8 -4
- package/dist/esm/local-apps.js +9 -9
- package/dist/esm/local-apps.spec.js +2 -8
- package/dist/esm/model-data.d.ts +3 -0
- package/dist/esm/model-data.d.ts.map +1 -1
- package/dist/esm/model-libraries-snippets.d.ts +3 -1
- package/dist/esm/model-libraries-snippets.d.ts.map +1 -1
- package/dist/esm/model-libraries-snippets.js +129 -19
- package/dist/esm/model-libraries.d.ts +12 -8
- package/dist/esm/model-libraries.d.ts.map +1 -1
- package/dist/esm/model-libraries.js +15 -8
- package/dist/esm/tasks/audio-to-audio/data.d.ts.map +1 -1
- package/dist/esm/tasks/audio-to-audio/data.js +0 -4
- package/dist/esm/tasks/fill-mask/data.js +2 -2
- package/dist/esm/tasks/image-classification/data.d.ts.map +1 -1
- package/dist/esm/tasks/image-classification/data.js +2 -3
- package/dist/esm/tasks/image-feature-extraction/data.d.ts.map +1 -1
- package/dist/esm/tasks/image-feature-extraction/data.js +8 -3
- package/dist/esm/tasks/image-text-to-text/data.d.ts.map +1 -1
- package/dist/esm/tasks/image-text-to-text/data.js +19 -11
- package/dist/esm/tasks/image-to-3d/data.js +4 -4
- package/dist/esm/tasks/image-to-image/data.d.ts.map +1 -1
- package/dist/esm/tasks/image-to-image/data.js +12 -4
- package/dist/esm/tasks/index.js +1 -1
- package/dist/esm/tasks/keypoint-detection/data.d.ts.map +1 -1
- package/dist/esm/tasks/keypoint-detection/data.js +4 -0
- package/dist/esm/tasks/object-detection/data.d.ts.map +1 -1
- package/dist/esm/tasks/object-detection/data.js +6 -2
- package/dist/esm/tasks/sentence-similarity/data.d.ts.map +1 -1
- package/dist/esm/tasks/sentence-similarity/data.js +5 -1
- package/dist/esm/tasks/text-generation/data.d.ts.map +1 -1
- package/dist/esm/tasks/text-generation/data.js +17 -13
- package/dist/esm/tasks/text-to-image/data.d.ts.map +1 -1
- package/dist/esm/tasks/text-to-image/data.js +4 -0
- package/dist/esm/tasks/text-to-speech/data.d.ts.map +1 -1
- package/dist/esm/tasks/text-to-speech/data.js +9 -9
- package/dist/esm/tasks/text-to-video/data.js +5 -5
- package/dist/esm/tasks/video-text-to-text/data.js +4 -4
- package/dist/esm/tasks/zero-shot-classification/data.js +2 -2
- package/dist/esm/tasks/zero-shot-image-classification/data.d.ts.map +1 -1
- package/dist/esm/tasks/zero-shot-image-classification/data.js +8 -4
- package/package.json +1 -1
- package/src/local-apps.spec.ts +2 -8
- package/src/local-apps.ts +9 -9
- package/src/model-data.ts +3 -0
- package/src/model-libraries-snippets.ts +141 -19
- package/src/model-libraries.ts +15 -8
- package/src/tasks/audio-to-audio/data.ts +0 -4
- package/src/tasks/fill-mask/data.ts +2 -2
- package/src/tasks/image-classification/data.ts +2 -3
- package/src/tasks/image-feature-extraction/data.ts +8 -3
- package/src/tasks/image-text-to-text/about.md +8 -3
- package/src/tasks/image-text-to-text/data.ts +19 -11
- package/src/tasks/image-to-3d/data.ts +4 -4
- package/src/tasks/image-to-image/data.ts +12 -5
- package/src/tasks/index.ts +1 -1
- package/src/tasks/keypoint-detection/data.ts +4 -0
- package/src/tasks/object-detection/data.ts +6 -2
- package/src/tasks/sentence-similarity/data.ts +5 -1
- package/src/tasks/text-generation/data.ts +17 -14
- package/src/tasks/text-to-image/data.ts +4 -0
- package/src/tasks/text-to-speech/data.ts +9 -10
- package/src/tasks/text-to-video/data.ts +5 -5
- package/src/tasks/video-text-to-text/data.ts +4 -4
- package/src/tasks/zero-shot-classification/data.ts +2 -2
- package/src/tasks/zero-shot-image-classification/data.ts +8 -4
|
@@ -172,6 +172,7 @@ export declare const MODEL_LIBRARIES_UI_ELEMENTS: {
|
|
|
172
172
|
prettyLabel: string;
|
|
173
173
|
repoName: string;
|
|
174
174
|
repoUrl: string;
|
|
175
|
+
snippets: (model: ModelData) => string[];
|
|
175
176
|
filter: false;
|
|
176
177
|
countDownloads: string;
|
|
177
178
|
};
|
|
@@ -202,6 +203,7 @@ export declare const MODEL_LIBRARIES_UI_ELEMENTS: {
|
|
|
202
203
|
prettyLabel: string;
|
|
203
204
|
repoName: string;
|
|
204
205
|
repoUrl: string;
|
|
206
|
+
snippets: (model: ModelData) => string[];
|
|
205
207
|
filter: false;
|
|
206
208
|
countDownloads: string;
|
|
207
209
|
};
|
|
@@ -398,13 +400,6 @@ export declare const MODEL_LIBRARIES_UI_ELEMENTS: {
|
|
|
398
400
|
snippets: (model: ModelData) => string[];
|
|
399
401
|
countDownloads: string;
|
|
400
402
|
};
|
|
401
|
-
"keras-nlp": {
|
|
402
|
-
prettyLabel: string;
|
|
403
|
-
repoName: string;
|
|
404
|
-
repoUrl: string;
|
|
405
|
-
docsUrl: string;
|
|
406
|
-
snippets: (model: ModelData) => string[];
|
|
407
|
-
};
|
|
408
403
|
"keras-hub": {
|
|
409
404
|
prettyLabel: string;
|
|
410
405
|
repoName: string;
|
|
@@ -774,6 +769,15 @@ export declare const MODEL_LIBRARIES_UI_ELEMENTS: {
|
|
|
774
769
|
repoName: string;
|
|
775
770
|
repoUrl: string;
|
|
776
771
|
};
|
|
772
|
+
terratorch: {
|
|
773
|
+
prettyLabel: string;
|
|
774
|
+
repoName: string;
|
|
775
|
+
repoUrl: string;
|
|
776
|
+
docsUrl: string;
|
|
777
|
+
filter: false;
|
|
778
|
+
countDownloads: string;
|
|
779
|
+
snippets: (model: ModelData) => string[];
|
|
780
|
+
};
|
|
777
781
|
"tic-clip": {
|
|
778
782
|
prettyLabel: string;
|
|
779
783
|
repoName: string;
|
|
@@ -884,5 +888,5 @@ export declare const MODEL_LIBRARIES_UI_ELEMENTS: {
|
|
|
884
888
|
};
|
|
885
889
|
export type ModelLibraryKey = keyof typeof MODEL_LIBRARIES_UI_ELEMENTS;
|
|
886
890
|
export declare const ALL_MODEL_LIBRARY_KEYS: ModelLibraryKey[];
|
|
887
|
-
export declare const ALL_DISPLAY_MODEL_LIBRARY_KEYS: ("adapter-transformers" | "allennlp" | "anemoi" | "asteroid" | "audiocraft" | "audioseal" | "bertopic" | "big_vision" | "birder" | "birefnet" | "bm25s" | "champ" | "chat_tts" | "colpali" | "cosmos" | "cxr-foundation" | "deepforest" | "depth-anything-v2" | "depth-pro" | "derm-foundation" | "diffree" | "diffusers" | "diffusionkit" | "doctr" | "cartesia_pytorch" | "cartesia_mlx" | "clipscope" | "cosyvoice" | "cotracker" | "edsnlp" | "elm" | "espnet" | "fairseq" | "fastai" | "fasttext" | "flair" | "gemma.cpp" | "gliner" | "glyph-byt5" | "grok" | "hallo" | "hezar" | "htrflow" | "hunyuan-dit" | "imstoucan" | "keras" | "tf-keras" | "keras-
|
|
891
|
+
export declare const ALL_DISPLAY_MODEL_LIBRARY_KEYS: ("adapter-transformers" | "allennlp" | "anemoi" | "asteroid" | "audiocraft" | "audioseal" | "bertopic" | "big_vision" | "birder" | "birefnet" | "bm25s" | "champ" | "chat_tts" | "colpali" | "cosmos" | "cxr-foundation" | "deepforest" | "depth-anything-v2" | "depth-pro" | "derm-foundation" | "diffree" | "diffusers" | "diffusionkit" | "doctr" | "cartesia_pytorch" | "cartesia_mlx" | "clipscope" | "cosyvoice" | "cotracker" | "edsnlp" | "elm" | "espnet" | "fairseq" | "fastai" | "fasttext" | "flair" | "gemma.cpp" | "gliner" | "glyph-byt5" | "grok" | "hallo" | "hezar" | "htrflow" | "hunyuan-dit" | "imstoucan" | "keras" | "tf-keras" | "keras-hub" | "k2" | "liveportrait" | "llama-cpp-python" | "mini-omni2" | "mindspore" | "mamba-ssm" | "mars5-tts" | "mesh-anything" | "mitie" | "ml-agents" | "mlx" | "mlx-image" | "mlc-llm" | "model2vec" | "moshi" | "nemo" | "open-oasis" | "open_clip" | "paddlenlp" | "peft" | "pxia" | "pyannote-audio" | "py-feat" | "pythae" | "recurrentgemma" | "relik" | "refiners" | "reverb" | "saelens" | "sam2" | "sample-factory" | "sapiens" | "sentence-transformers" | "setfit" | "sklearn" | "spacy" | "span-marker" | "speechbrain" | "ssr-speech" | "stable-audio-tools" | "diffusion-single-file" | "seed-story" | "soloaudio" | "stable-baselines3" | "stanza" | "f5-tts" | "genmo" | "tensorflowtts" | "tabpfn" | "terratorch" | "tic-clip" | "timesfm" | "timm" | "transformers" | "transformers.js" | "trellis" | "ultralytics" | "unity-sentis" | "sana" | "vfi-mamba" | "voicecraft" | "whisperkit" | "yolov10" | "3dtopia-xl")[];
|
|
888
892
|
//# sourceMappingURL=model-libraries.d.ts.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"model-libraries.d.ts","sourceRoot":"","sources":["../../src/model-libraries.ts"],"names":[],"mappings":"AACA,OAAO,KAAK,EAAE,SAAS,EAAE,MAAM,iBAAiB,CAAC;AACjD,OAAO,KAAK,EAAE,kBAAkB,EAAE,MAAM,gCAAgC,CAAC;AAEzE;;GAEG;AACH,MAAM,WAAW,gBAAgB;IAChC;;;;OAIG;IACH,WAAW,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,QAAQ,EAAE,MAAM,CAAC;IACjB;;OAEG;IACH,OAAO,EAAE,MAAM,CAAC;IAChB;;OAEG;IACH,OAAO,CAAC,EAAE,MAAM,CAAC;IACjB;;OAEG;IACH,QAAQ,CAAC,EAAE,CAAC,KAAK,EAAE,SAAS,KAAK,MAAM,EAAE,CAAC;IAC1C;;;;;OAKG;IACH,cAAc,CAAC,EAAE,kBAAkB,CAAC;IACpC;;;OAGG;IACH,MAAM,CAAC,EAAE,OAAO,CAAC;CACjB;AAED;;;;;;;;;;;;;GAaG;AAEH,eAAO,MAAM,2BAA2B
|
|
1
|
+
{"version":3,"file":"model-libraries.d.ts","sourceRoot":"","sources":["../../src/model-libraries.ts"],"names":[],"mappings":"AACA,OAAO,KAAK,EAAE,SAAS,EAAE,MAAM,iBAAiB,CAAC;AACjD,OAAO,KAAK,EAAE,kBAAkB,EAAE,MAAM,gCAAgC,CAAC;AAEzE;;GAEG;AACH,MAAM,WAAW,gBAAgB;IAChC;;;;OAIG;IACH,WAAW,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,QAAQ,EAAE,MAAM,CAAC;IACjB;;OAEG;IACH,OAAO,EAAE,MAAM,CAAC;IAChB;;OAEG;IACH,OAAO,CAAC,EAAE,MAAM,CAAC;IACjB;;OAEG;IACH,QAAQ,CAAC,EAAE,CAAC,KAAK,EAAE,SAAS,KAAK,MAAM,EAAE,CAAC;IAC1C;;;;;OAKG;IACH,cAAc,CAAC,EAAE,kBAAkB,CAAC;IACpC;;;OAGG;IACH,MAAM,CAAC,EAAE,OAAO,CAAC;CACjB;AAED;;;;;;;;;;;;;GAaG;AAEH,eAAO,MAAM,2BAA2B;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CAs0BI,CAAC;AAE7C,MAAM,MAAM,eAAe,GAAG,MAAM,OAAO,2BAA2B,CAAC;AAEvE,eAAO,MAAM,sBAAsB,EAA+C,eAAe,EAAE,CAAC;AAEpG,eAAO,MAAM,8BAA8B,0gDAQ1B,CAAC"}
|
|
@@ -132,6 +132,7 @@ export const MODEL_LIBRARIES_UI_ELEMENTS = {
|
|
|
132
132
|
prettyLabel: "CXR Foundation",
|
|
133
133
|
repoName: "cxr-foundation",
|
|
134
134
|
repoUrl: "https://github.com/google-health/cxr-foundation",
|
|
135
|
+
snippets: snippets.cxr_foundation,
|
|
135
136
|
filter: false,
|
|
136
137
|
countDownloads: `path:"precomputed_embeddings/embeddings.npz" OR path:"pax-elixr-b-text/saved_model.pb"`,
|
|
137
138
|
},
|
|
@@ -162,6 +163,7 @@ export const MODEL_LIBRARIES_UI_ELEMENTS = {
|
|
|
162
163
|
prettyLabel: "Derm Foundation",
|
|
163
164
|
repoName: "derm-foundation",
|
|
164
165
|
repoUrl: "https://github.com/google-health/derm-foundation",
|
|
166
|
+
snippets: snippets.derm_foundation,
|
|
165
167
|
filter: false,
|
|
166
168
|
countDownloads: `path:"scin_dataset_precomputed_embeddings.npz" OR path:"saved_model.pb"`,
|
|
167
169
|
},
|
|
@@ -360,13 +362,6 @@ export const MODEL_LIBRARIES_UI_ELEMENTS = {
|
|
|
360
362
|
snippets: snippets.tf_keras,
|
|
361
363
|
countDownloads: `path:"saved_model.pb"`,
|
|
362
364
|
},
|
|
363
|
-
"keras-nlp": {
|
|
364
|
-
prettyLabel: "KerasNLP",
|
|
365
|
-
repoName: "KerasNLP",
|
|
366
|
-
repoUrl: "https://github.com/keras-team/keras-nlp",
|
|
367
|
-
docsUrl: "https://keras.io/keras_nlp/",
|
|
368
|
-
snippets: snippets.keras_nlp,
|
|
369
|
-
},
|
|
370
365
|
"keras-hub": {
|
|
371
366
|
prettyLabel: "KerasHub",
|
|
372
367
|
repoName: "KerasHub",
|
|
@@ -500,7 +495,10 @@ export const MODEL_LIBRARIES_UI_ELEMENTS = {
|
|
|
500
495
|
repoUrl: "https://github.com/mlfoundations/open_clip",
|
|
501
496
|
snippets: snippets.open_clip,
|
|
502
497
|
filter: true,
|
|
503
|
-
countDownloads: `
|
|
498
|
+
countDownloads: `path:"open_clip_model.safetensors"
|
|
499
|
+
OR path:"model.safetensors"
|
|
500
|
+
OR path:"open_clip_pytorch_model.bin"
|
|
501
|
+
OR path:"pytorch_model.bin"`,
|
|
504
502
|
},
|
|
505
503
|
paddlenlp: {
|
|
506
504
|
prettyLabel: "paddlenlp",
|
|
@@ -736,6 +734,15 @@ export const MODEL_LIBRARIES_UI_ELEMENTS = {
|
|
|
736
734
|
repoName: "TabPFN",
|
|
737
735
|
repoUrl: "https://github.com/PriorLabs/TabPFN",
|
|
738
736
|
},
|
|
737
|
+
terratorch: {
|
|
738
|
+
prettyLabel: "TerraTorch",
|
|
739
|
+
repoName: "TerraTorch",
|
|
740
|
+
repoUrl: "https://github.com/IBM/terratorch",
|
|
741
|
+
docsUrl: "https://ibm.github.io/terratorch/",
|
|
742
|
+
filter: false,
|
|
743
|
+
countDownloads: `path_extension:"pt"`,
|
|
744
|
+
snippets: snippets.terratorch,
|
|
745
|
+
},
|
|
739
746
|
"tic-clip": {
|
|
740
747
|
prettyLabel: "TiC-CLIP",
|
|
741
748
|
repoName: "TiC-CLIP",
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/audio-to-audio/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,
|
|
1
|
+
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/audio-to-audio/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,cA6Df,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
|
@@ -57,8 +57,8 @@ const taskData = {
|
|
|
57
57
|
],
|
|
58
58
|
models: [
|
|
59
59
|
{
|
|
60
|
-
description: "
|
|
61
|
-
id: "
|
|
60
|
+
description: "State-of-the-art masked language model.",
|
|
61
|
+
id: "answerdotai/ModernBERT-large",
|
|
62
62
|
},
|
|
63
63
|
{
|
|
64
64
|
description: "A multilingual model trained on 100 languages.",
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/image-classification/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,
|
|
1
|
+
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/image-classification/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,cAkFf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
|
@@ -72,9 +72,8 @@ const taskData = {
|
|
|
72
72
|
],
|
|
73
73
|
spaces: [
|
|
74
74
|
{
|
|
75
|
-
|
|
76
|
-
|
|
77
|
-
id: "nielsr/perceiver-image-classification",
|
|
75
|
+
description: "A leaderboard to evaluate different image classification models.",
|
|
76
|
+
id: "timm/leaderboard",
|
|
78
77
|
},
|
|
79
78
|
],
|
|
80
79
|
summary: "Image classification is the task of assigning a label or class to an entire image. Images are expected to have only one class for each image. Image classification models take an image as input and return a prediction about which class the image belongs to.",
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/image-feature-extraction/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,
|
|
1
|
+
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/image-feature-extraction/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,cA2Df,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
|
@@ -40,15 +40,20 @@ const taskData = {
|
|
|
40
40
|
id: "facebook/dino-vitb16",
|
|
41
41
|
},
|
|
42
42
|
{
|
|
43
|
-
description: "
|
|
44
|
-
id: "
|
|
43
|
+
description: "Cutting-edge image feature extraction model.",
|
|
44
|
+
id: "apple/aimv2-large-patch14-336-distilled",
|
|
45
45
|
},
|
|
46
46
|
{
|
|
47
47
|
description: "Strong image feature extraction model that can be used on images and documents.",
|
|
48
48
|
id: "OpenGVLab/InternViT-6B-448px-V1-2",
|
|
49
49
|
},
|
|
50
50
|
],
|
|
51
|
-
spaces: [
|
|
51
|
+
spaces: [
|
|
52
|
+
{
|
|
53
|
+
description: "A leaderboard to evaluate different image-feature-extraction models on classification performances",
|
|
54
|
+
id: "timm/leaderboard",
|
|
55
|
+
},
|
|
56
|
+
],
|
|
52
57
|
summary: "Image feature extraction is the task of extracting features learnt in a computer vision model.",
|
|
53
58
|
widgetModels: [],
|
|
54
59
|
};
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/image-text-to-text/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,
|
|
1
|
+
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/image-text-to-text/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,cAyGf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
|
@@ -5,8 +5,8 @@ const taskData = {
|
|
|
5
5
|
id: "liuhaotian/LLaVA-Instruct-150K",
|
|
6
6
|
},
|
|
7
7
|
{
|
|
8
|
-
description: "
|
|
9
|
-
id: "
|
|
8
|
+
description: "Collection of image-text pairs on scientific topics.",
|
|
9
|
+
id: "DAMO-NLP-SG/multimodal_textbook",
|
|
10
10
|
},
|
|
11
11
|
{
|
|
12
12
|
description: "A collection of datasets made for model fine-tuning.",
|
|
@@ -40,11 +40,15 @@ const taskData = {
|
|
|
40
40
|
metrics: [],
|
|
41
41
|
models: [
|
|
42
42
|
{
|
|
43
|
-
description: "
|
|
44
|
-
id: "
|
|
43
|
+
description: "Small and efficient yet powerful vision language model.",
|
|
44
|
+
id: "HuggingFaceTB/SmolVLM-Instruct",
|
|
45
45
|
},
|
|
46
46
|
{
|
|
47
|
-
description: "
|
|
47
|
+
description: "A screenshot understanding model used to control computers.",
|
|
48
|
+
id: "showlab/ShowUI-2B",
|
|
49
|
+
},
|
|
50
|
+
{
|
|
51
|
+
description: "Cutting-edge vision language model.",
|
|
48
52
|
id: "allenai/Molmo-7B-D-0924",
|
|
49
53
|
},
|
|
50
54
|
{
|
|
@@ -56,8 +60,8 @@ const taskData = {
|
|
|
56
60
|
id: "Qwen/Qwen2-VL-7B-Instruct",
|
|
57
61
|
},
|
|
58
62
|
{
|
|
59
|
-
description: "
|
|
60
|
-
id: "
|
|
63
|
+
description: "Image-text-to-text model with reasoning capabilities.",
|
|
64
|
+
id: "Qwen/QVQ-72B-Preview",
|
|
61
65
|
},
|
|
62
66
|
{
|
|
63
67
|
description: "Strong image-text-to-text model focused on documents.",
|
|
@@ -81,14 +85,18 @@ const taskData = {
|
|
|
81
85
|
description: "An image-text-to-text application focused on documents.",
|
|
82
86
|
id: "stepfun-ai/GOT_official_online_demo",
|
|
83
87
|
},
|
|
84
|
-
{
|
|
85
|
-
description: "An application to compare outputs of different vision language models.",
|
|
86
|
-
id: "merve/compare_VLMs",
|
|
87
|
-
},
|
|
88
88
|
{
|
|
89
89
|
description: "An application for chatting with an image-text-to-text model.",
|
|
90
90
|
id: "GanymedeNil/Qwen2-VL-7B",
|
|
91
91
|
},
|
|
92
|
+
{
|
|
93
|
+
description: "An application that parses screenshots into actions.",
|
|
94
|
+
id: "showlab/ShowUI",
|
|
95
|
+
},
|
|
96
|
+
{
|
|
97
|
+
description: "An application that detects gaze.",
|
|
98
|
+
id: "smoondream/gaze-demo",
|
|
99
|
+
},
|
|
92
100
|
],
|
|
93
101
|
summary: "Image-text-to-text models take in an image and text prompt and output text. These models are also called vision-language models, or VLMs. The difference from image-to-text models is that these models take an additional text input, not restricting the model to certain use cases like image captioning, and may also be trained to accept a conversation as input.",
|
|
94
102
|
widgetModels: ["meta-llama/Llama-3.2-11B-Vision-Instruct"],
|
|
@@ -39,8 +39,8 @@ const taskData = {
|
|
|
39
39
|
id: "hwjiang/Real3D",
|
|
40
40
|
},
|
|
41
41
|
{
|
|
42
|
-
description: "
|
|
43
|
-
id: "
|
|
42
|
+
description: "Consistent image-to-3d generation model.",
|
|
43
|
+
id: "stabilityai/stable-point-aware-3d",
|
|
44
44
|
},
|
|
45
45
|
],
|
|
46
46
|
spaces: [
|
|
@@ -53,8 +53,8 @@ const taskData = {
|
|
|
53
53
|
id: "TencentARC/InstantMesh",
|
|
54
54
|
},
|
|
55
55
|
{
|
|
56
|
-
description: "Image-to-3D demo
|
|
57
|
-
id: "stabilityai/
|
|
56
|
+
description: "Image-to-3D demo.",
|
|
57
|
+
id: "stabilityai/stable-point-aware-3d",
|
|
58
58
|
},
|
|
59
59
|
{
|
|
60
60
|
description: "Image-to-3D demo with mesh outputs.",
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/image-to-image/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,
|
|
1
|
+
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/image-to-image/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,cAuGf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
|
@@ -8,6 +8,10 @@ const taskData = {
|
|
|
8
8
|
description: "Multiple images of celebrities, used for facial expression translation",
|
|
9
9
|
id: "huggan/CelebA-faces",
|
|
10
10
|
},
|
|
11
|
+
{
|
|
12
|
+
description: "12M image-caption pairs.",
|
|
13
|
+
id: "Spawning/PD12M",
|
|
14
|
+
},
|
|
11
15
|
],
|
|
12
16
|
demo: {
|
|
13
17
|
inputs: [
|
|
@@ -48,16 +52,20 @@ const taskData = {
|
|
|
48
52
|
id: "keras-io/super-resolution",
|
|
49
53
|
},
|
|
50
54
|
{
|
|
51
|
-
description: "A model
|
|
52
|
-
id: "
|
|
55
|
+
description: "A model for applying edits to images through image controls.",
|
|
56
|
+
id: "Yuanshi/OminiControl",
|
|
53
57
|
},
|
|
54
58
|
{
|
|
55
59
|
description: "A model that generates images based on segments in the input image and the text prompt.",
|
|
56
60
|
id: "mfidabel/controlnet-segment-anything",
|
|
57
61
|
},
|
|
58
62
|
{
|
|
59
|
-
description: "
|
|
60
|
-
id: "
|
|
63
|
+
description: "Strong model for inpainting and outpainting.",
|
|
64
|
+
id: "black-forest-labs/FLUX.1-Fill-dev",
|
|
65
|
+
},
|
|
66
|
+
{
|
|
67
|
+
description: "Strong model for image editing using depth maps.",
|
|
68
|
+
id: "black-forest-labs/FLUX.1-Depth-dev-lora",
|
|
61
69
|
},
|
|
62
70
|
],
|
|
63
71
|
spaces: [
|
package/dist/esm/tasks/index.js
CHANGED
|
@@ -63,7 +63,7 @@ export const TASKS_MODEL_LIBRARIES = {
|
|
|
63
63
|
"video-classification": ["transformers"],
|
|
64
64
|
"mask-generation": ["transformers"],
|
|
65
65
|
"multiple-choice": ["transformers"],
|
|
66
|
-
"object-detection": ["transformers", "transformers.js"],
|
|
66
|
+
"object-detection": ["transformers", "transformers.js", "ultralytics"],
|
|
67
67
|
other: [],
|
|
68
68
|
"question-answering": ["adapter-transformers", "allennlp", "transformers", "transformers.js"],
|
|
69
69
|
robotics: [],
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/keypoint-detection/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,
|
|
1
|
+
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/keypoint-detection/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,cAiDf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
|
@@ -29,6 +29,10 @@ const taskData = {
|
|
|
29
29
|
description: "Strong keypoint detection model used to detect human pose.",
|
|
30
30
|
id: "facebook/sapiens-pose-1b",
|
|
31
31
|
},
|
|
32
|
+
{
|
|
33
|
+
description: "Powerful keypoint detection model used to detect human pose.",
|
|
34
|
+
id: "usyd-community/vitpose-plus-base",
|
|
35
|
+
},
|
|
32
36
|
],
|
|
33
37
|
spaces: [
|
|
34
38
|
{
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/object-detection/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,
|
|
1
|
+
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/object-detection/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,cAqFf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
|
@@ -50,6 +50,10 @@ const taskData = {
|
|
|
50
50
|
description: "Fast and accurate object detection model trained on COCO and Object365 datasets.",
|
|
51
51
|
id: "PekingU/rtdetr_r18vd_coco_o365",
|
|
52
52
|
},
|
|
53
|
+
{
|
|
54
|
+
description: "Object detection model for low-lying objects.",
|
|
55
|
+
id: "StephanST/WALDO30",
|
|
56
|
+
},
|
|
53
57
|
],
|
|
54
58
|
spaces: [
|
|
55
59
|
{
|
|
@@ -61,8 +65,8 @@ const taskData = {
|
|
|
61
65
|
id: "Gradio-Blocks/Object-Detection-With-DETR-and-YOLOS",
|
|
62
66
|
},
|
|
63
67
|
{
|
|
64
|
-
description: "
|
|
65
|
-
id: "
|
|
68
|
+
description: "A cutting-edge object detection application.",
|
|
69
|
+
id: "Ultralytics/YOLO11",
|
|
66
70
|
},
|
|
67
71
|
{
|
|
68
72
|
description: "An object tracking, segmentation and inpainting application.",
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/sentence-similarity/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,
|
|
1
|
+
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/sentence-similarity/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,cAoGf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
|
@@ -64,9 +64,13 @@ const taskData = {
|
|
|
64
64
|
id: "sentence-transformers/all-mpnet-base-v2",
|
|
65
65
|
},
|
|
66
66
|
{
|
|
67
|
-
description: "A multilingual robust sentence similarity model
|
|
67
|
+
description: "A multilingual robust sentence similarity model.",
|
|
68
68
|
id: "BAAI/bge-m3",
|
|
69
69
|
},
|
|
70
|
+
{
|
|
71
|
+
description: "A robust sentence similarity model.",
|
|
72
|
+
id: "HIT-TMG/KaLM-embedding-multilingual-mini-instruct-v1.5",
|
|
73
|
+
},
|
|
70
74
|
],
|
|
71
75
|
spaces: [
|
|
72
76
|
{
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/text-generation/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,
|
|
1
|
+
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/text-generation/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,cAgHf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
|
@@ -1,24 +1,28 @@
|
|
|
1
1
|
const taskData = {
|
|
2
2
|
datasets: [
|
|
3
3
|
{
|
|
4
|
-
description: "
|
|
5
|
-
id: "
|
|
4
|
+
description: "Multilingual dataset used to evaluate text generation models.",
|
|
5
|
+
id: "CohereForAI/Global-MMLU",
|
|
6
6
|
},
|
|
7
7
|
{
|
|
8
|
-
description: "
|
|
9
|
-
id: "
|
|
8
|
+
description: "High quality multilingual data used to train text-generation models.",
|
|
9
|
+
id: "HuggingFaceFW/fineweb-2",
|
|
10
10
|
},
|
|
11
11
|
{
|
|
12
12
|
description: "Truly open-source, curated and cleaned dialogue dataset.",
|
|
13
13
|
id: "HuggingFaceH4/ultrachat_200k",
|
|
14
14
|
},
|
|
15
15
|
{
|
|
16
|
-
description: "
|
|
17
|
-
id: "
|
|
16
|
+
description: "A multilingual instruction dataset with preference ratings on responses.",
|
|
17
|
+
id: "allenai/tulu-3-sft-mixture",
|
|
18
18
|
},
|
|
19
19
|
{
|
|
20
20
|
description: "A large synthetic dataset for alignment of text generation models.",
|
|
21
|
-
id: "
|
|
21
|
+
id: "HuggingFaceTB/smoltalk",
|
|
22
|
+
},
|
|
23
|
+
{
|
|
24
|
+
description: "A dataset made for training text generation models solving math questions.",
|
|
25
|
+
id: "HuggingFaceTB/finemath",
|
|
22
26
|
},
|
|
23
27
|
],
|
|
24
28
|
demo: {
|
|
@@ -57,20 +61,20 @@ const taskData = {
|
|
|
57
61
|
id: "meta-llama/Meta-Llama-3.1-8B-Instruct",
|
|
58
62
|
},
|
|
59
63
|
{
|
|
60
|
-
description: "
|
|
61
|
-
id: "microsoft/
|
|
64
|
+
description: "Powerful text generation model by Microsoft.",
|
|
65
|
+
id: "microsoft/phi-4",
|
|
62
66
|
},
|
|
63
67
|
{
|
|
64
|
-
description: "A very powerful model
|
|
65
|
-
id: "
|
|
68
|
+
description: "A very powerful model with reasoning capabilities.",
|
|
69
|
+
id: "PowerInfer/SmallThinker-3B-Preview",
|
|
66
70
|
},
|
|
67
71
|
{
|
|
68
72
|
description: "Strong text generation model to follow instructions.",
|
|
69
73
|
id: "Qwen/Qwen2.5-7B-Instruct",
|
|
70
74
|
},
|
|
71
75
|
{
|
|
72
|
-
description: "
|
|
73
|
-
id: "
|
|
76
|
+
description: "Text generation model used to write code.",
|
|
77
|
+
id: "Qwen/Qwen2.5-Coder-32B-Instruct",
|
|
74
78
|
},
|
|
75
79
|
],
|
|
76
80
|
spaces: [
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/text-to-image/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,
|
|
1
|
+
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/text-to-image/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,cAmGf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
|
@@ -8,6 +8,10 @@ const taskData = {
|
|
|
8
8
|
description: "Conceptual Captions is a dataset consisting of ~3.3M images annotated with captions.",
|
|
9
9
|
id: "conceptual_captions",
|
|
10
10
|
},
|
|
11
|
+
{
|
|
12
|
+
description: "12M image-caption pairs.",
|
|
13
|
+
id: "Spawning/PD12M",
|
|
14
|
+
},
|
|
11
15
|
],
|
|
12
16
|
demo: {
|
|
13
17
|
inputs: [
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/text-to-speech/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,
|
|
1
|
+
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/text-to-speech/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,cAiFf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
|
@@ -10,7 +10,7 @@ const taskData = {
|
|
|
10
10
|
id: "mythicinfinity/libritts_r",
|
|
11
11
|
},
|
|
12
12
|
{
|
|
13
|
-
description: "
|
|
13
|
+
description: "Multi-lingual dataset.",
|
|
14
14
|
id: "facebook/multilingual_librispeech",
|
|
15
15
|
},
|
|
16
16
|
],
|
|
@@ -46,25 +46,25 @@ const taskData = {
|
|
|
46
46
|
},
|
|
47
47
|
{
|
|
48
48
|
description: "A massively multi-lingual TTS model.",
|
|
49
|
-
id: "
|
|
49
|
+
id: "fishaudio/fish-speech-1.5",
|
|
50
50
|
},
|
|
51
51
|
{
|
|
52
52
|
description: "A powerful TTS model.",
|
|
53
|
-
id: "
|
|
53
|
+
id: "OuteAI/OuteTTS-0.1-350M",
|
|
54
54
|
},
|
|
55
55
|
{
|
|
56
|
-
description: "
|
|
57
|
-
id: "
|
|
56
|
+
description: "Small yet powerful TTS model.",
|
|
57
|
+
id: "hexgrad/Kokoro-82M",
|
|
58
58
|
},
|
|
59
59
|
],
|
|
60
60
|
spaces: [
|
|
61
61
|
{
|
|
62
|
-
description: "An application for generate
|
|
63
|
-
id: "
|
|
62
|
+
description: "An application for generate high quality speech in different languages.",
|
|
63
|
+
id: "hexgrad/Kokoro-TTS",
|
|
64
64
|
},
|
|
65
65
|
{
|
|
66
|
-
description: "
|
|
67
|
-
id: "
|
|
66
|
+
description: "A multilingual text-to-speech application.",
|
|
67
|
+
id: "fishaudio/fish-speech-1",
|
|
68
68
|
},
|
|
69
69
|
{
|
|
70
70
|
description: "An application that generates speech in different styles in English and Chinese.",
|
|
@@ -61,15 +61,15 @@ const taskData = {
|
|
|
61
61
|
models: [
|
|
62
62
|
{
|
|
63
63
|
description: "A strong model for consistent video generation.",
|
|
64
|
-
id: "
|
|
64
|
+
id: "tencent/HunyuanVideo",
|
|
65
65
|
},
|
|
66
66
|
{
|
|
67
|
-
description: "A
|
|
68
|
-
id: "
|
|
67
|
+
description: "A text-to-video model with high fidelity motion and strong prompt adherence.",
|
|
68
|
+
id: "Lightricks/LTX-Video",
|
|
69
69
|
},
|
|
70
70
|
{
|
|
71
|
-
description: "A
|
|
72
|
-
id: "
|
|
71
|
+
description: "A text-to-video model focusing on physics-aware applications like robotics.",
|
|
72
|
+
id: "nvidia/Cosmos-1.0-Diffusion-7B-Text2World",
|
|
73
73
|
},
|
|
74
74
|
],
|
|
75
75
|
spaces: [
|
|
@@ -36,12 +36,12 @@ const taskData = {
|
|
|
36
36
|
metrics: [],
|
|
37
37
|
models: [
|
|
38
38
|
{
|
|
39
|
-
description: "A robust video-text-to-text model
|
|
40
|
-
id: "
|
|
39
|
+
description: "A robust video-text-to-text model.",
|
|
40
|
+
id: "Vision-CAIR/LongVU_Qwen2_7B",
|
|
41
41
|
},
|
|
42
42
|
{
|
|
43
|
-
description: "
|
|
44
|
-
id: "
|
|
43
|
+
description: "Strong video-text-to-text model with reasoning capabilities.",
|
|
44
|
+
id: "GoodiesHere/Apollo-LMMs-Apollo-7B-t32",
|
|
45
45
|
},
|
|
46
46
|
],
|
|
47
47
|
spaces: [
|
|
@@ -53,8 +53,8 @@ const taskData = {
|
|
|
53
53
|
id: "facebook/bart-large-mnli",
|
|
54
54
|
},
|
|
55
55
|
{
|
|
56
|
-
description: "
|
|
57
|
-
id: "MoritzLaurer/
|
|
56
|
+
description: "Cutting-edge zero-shot multilingual text classification model.",
|
|
57
|
+
id: "MoritzLaurer/ModernBERT-large-zeroshot-v2.0",
|
|
58
58
|
},
|
|
59
59
|
],
|
|
60
60
|
spaces: [],
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/zero-shot-image-classification/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,
|
|
1
|
+
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/zero-shot-image-classification/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,aAAa,CAAC;AAElD,QAAA,MAAM,QAAQ,EAAE,cAmFf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|