@huggingface/tasks 0.11.7 → 0.11.8
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/index.cjs +80 -31
- package/dist/index.js +80 -31
- package/dist/src/local-apps.d.ts +0 -1
- package/dist/src/local-apps.d.ts.map +1 -1
- package/dist/src/model-libraries-snippets.d.ts +1 -0
- package/dist/src/model-libraries-snippets.d.ts.map +1 -1
- package/dist/src/model-libraries.d.ts +15 -2
- package/dist/src/model-libraries.d.ts.map +1 -1
- package/dist/src/pipelines.d.ts +8 -2
- package/dist/src/pipelines.d.ts.map +1 -1
- package/dist/src/tasks/feature-extraction/data.d.ts.map +1 -1
- package/dist/src/tasks/image-feature-extraction/data.d.ts.map +1 -1
- package/dist/src/tasks/index.d.ts.map +1 -1
- package/dist/src/tasks/text-generation/data.d.ts.map +1 -1
- package/package.json +4 -1
- package/src/local-apps.ts +0 -1
- package/src/model-libraries-snippets.ts +18 -0
- package/src/model-libraries.ts +13 -0
- package/src/pipelines.ts +6 -0
- package/src/tasks/feature-extraction/data.ts +5 -1
- package/src/tasks/image-feature-extraction/data.ts +7 -3
- package/src/tasks/image-segmentation/data.ts +4 -4
- package/src/tasks/image-text-to-text/about.md +2 -0
- package/src/tasks/image-text-to-text/data.ts +1 -1
- package/src/tasks/image-to-image/data.ts +2 -2
- package/src/tasks/index.ts +2 -0
- package/src/tasks/mask-generation/data.ts +4 -4
- package/src/tasks/text-generation/data.ts +16 -12
- package/src/tasks/text-to-image/data.ts +3 -3
package/dist/index.cjs
CHANGED
|
@@ -1388,6 +1388,12 @@ var PIPELINE_DATA = {
|
|
|
1388
1388
|
modality: "cv",
|
|
1389
1389
|
color: "indigo"
|
|
1390
1390
|
},
|
|
1391
|
+
"video-text-to-text": {
|
|
1392
|
+
name: "Video-Text-to-Text",
|
|
1393
|
+
modality: "multimodal",
|
|
1394
|
+
color: "blue",
|
|
1395
|
+
hideInDatasets: true
|
|
1396
|
+
},
|
|
1391
1397
|
other: {
|
|
1392
1398
|
name: "Other",
|
|
1393
1399
|
modality: "other",
|
|
@@ -1731,8 +1737,12 @@ var taskData5 = {
|
|
|
1731
1737
|
],
|
|
1732
1738
|
spaces: [
|
|
1733
1739
|
{
|
|
1734
|
-
description: "A leaderboard to rank
|
|
1740
|
+
description: "A leaderboard to rank text feature extraction models based on a benchmark.",
|
|
1735
1741
|
id: "mteb/leaderboard"
|
|
1742
|
+
},
|
|
1743
|
+
{
|
|
1744
|
+
description: "A leaderboard to rank best feature extraction models based on human feedback.",
|
|
1745
|
+
id: "mteb/arena"
|
|
1736
1746
|
}
|
|
1737
1747
|
],
|
|
1738
1748
|
summary: "Feature extraction is the task of extracting features learnt in a model.",
|
|
@@ -1937,15 +1947,19 @@ var taskData8 = {
|
|
|
1937
1947
|
},
|
|
1938
1948
|
{
|
|
1939
1949
|
description: "A strong image feature extraction model.",
|
|
1940
|
-
id: "
|
|
1950
|
+
id: "nvidia/MambaVision-T-1K"
|
|
1941
1951
|
},
|
|
1942
1952
|
{
|
|
1943
|
-
description: "A robust image feature extraction
|
|
1953
|
+
description: "A robust image feature extraction model.",
|
|
1944
1954
|
id: "facebook/dino-vitb16"
|
|
1945
1955
|
},
|
|
1946
1956
|
{
|
|
1947
|
-
description: "Strong image
|
|
1957
|
+
description: "Strong image feature extraction model made for information retrieval from documents.",
|
|
1948
1958
|
id: "vidore/colpali"
|
|
1959
|
+
},
|
|
1960
|
+
{
|
|
1961
|
+
description: "Strong image feature extraction model that can be used on images and documents.",
|
|
1962
|
+
id: "OpenGVLab/InternViT-6B-448px-V1-2"
|
|
1949
1963
|
}
|
|
1950
1964
|
],
|
|
1951
1965
|
spaces: [],
|
|
@@ -1997,8 +2011,8 @@ var taskData9 = {
|
|
|
1997
2011
|
],
|
|
1998
2012
|
models: [
|
|
1999
2013
|
{
|
|
2000
|
-
description: "
|
|
2001
|
-
id: "
|
|
2014
|
+
description: "An image-to-image model to improve image resolution.",
|
|
2015
|
+
id: "fal/AuraSR-v2"
|
|
2002
2016
|
},
|
|
2003
2017
|
{
|
|
2004
2018
|
description: "A model that increases the resolution of an image.",
|
|
@@ -2216,7 +2230,7 @@ var taskData11 = {
|
|
|
2216
2230
|
],
|
|
2217
2231
|
summary: "Image-text-to-text models take in an image and text prompt and output text. These models are also called vision-language models, or VLMs. The difference from image-to-text models is that these models take an additional text input, not restricting the model to certain use cases like image captioning, and may also be trained to accept a conversation as input.",
|
|
2218
2232
|
widgetModels: ["microsoft/kosmos-2-patch14-224"],
|
|
2219
|
-
youtubeId: ""
|
|
2233
|
+
youtubeId: "IoGaGfU1CIg"
|
|
2220
2234
|
};
|
|
2221
2235
|
var data_default11 = taskData11;
|
|
2222
2236
|
|
|
@@ -2267,16 +2281,16 @@ var taskData12 = {
|
|
|
2267
2281
|
id: "facebook/detr-resnet-50-panoptic"
|
|
2268
2282
|
},
|
|
2269
2283
|
{
|
|
2270
|
-
description: "
|
|
2271
|
-
id: "
|
|
2284
|
+
description: "Background removal model.",
|
|
2285
|
+
id: "briaai/RMBG-1.4"
|
|
2272
2286
|
},
|
|
2273
2287
|
{
|
|
2274
2288
|
description: "Semantic segmentation model trained on ADE20k benchmark dataset with 512x512 resolution.",
|
|
2275
2289
|
id: "nvidia/segformer-b0-finetuned-ade-512-512"
|
|
2276
2290
|
},
|
|
2277
2291
|
{
|
|
2278
|
-
description: "
|
|
2279
|
-
id: "
|
|
2292
|
+
description: "A multipurpose image segmentation model for high resolution images.",
|
|
2293
|
+
id: "ZhengPeng7/BiRefNet"
|
|
2280
2294
|
},
|
|
2281
2295
|
{
|
|
2282
2296
|
description: "Panoptic segmentation model trained COCO (common objects) dataset.",
|
|
@@ -2340,13 +2354,13 @@ var taskData13 = {
|
|
|
2340
2354
|
},
|
|
2341
2355
|
{
|
|
2342
2356
|
description: "Very strong mask generation model.",
|
|
2343
|
-
id: "facebook/
|
|
2357
|
+
id: "facebook/sam2-hiera-large"
|
|
2344
2358
|
}
|
|
2345
2359
|
],
|
|
2346
2360
|
spaces: [
|
|
2347
2361
|
{
|
|
2348
|
-
description: "An application that combines a mask generation model with
|
|
2349
|
-
id: "
|
|
2362
|
+
description: "An application that combines a mask generation model with a zero-shot object detection model for text-guided image segmentation.",
|
|
2363
|
+
id: "merve/OWLSAM2"
|
|
2350
2364
|
},
|
|
2351
2365
|
{
|
|
2352
2366
|
description: "An application that compares the performance of a large and a small mask generation model.",
|
|
@@ -2354,7 +2368,7 @@ var taskData13 = {
|
|
|
2354
2368
|
},
|
|
2355
2369
|
{
|
|
2356
2370
|
description: "An application based on an improved mask generation model.",
|
|
2357
|
-
id: "
|
|
2371
|
+
id: "SkalskiP/segment-anything-model-2"
|
|
2358
2372
|
},
|
|
2359
2373
|
{
|
|
2360
2374
|
description: "An application to remove objects from videos using mask generation models.",
|
|
@@ -3050,15 +3064,15 @@ var taskData24 = {
|
|
|
3050
3064
|
models: [
|
|
3051
3065
|
{
|
|
3052
3066
|
description: "One of the most powerful image generation models that can generate realistic outputs.",
|
|
3053
|
-
id: "
|
|
3067
|
+
id: "black-forest-labs/FLUX.1-dev"
|
|
3054
3068
|
},
|
|
3055
3069
|
{
|
|
3056
3070
|
description: "A powerful yet fast image generation model.",
|
|
3057
3071
|
id: "latent-consistency/lcm-lora-sdxl"
|
|
3058
3072
|
},
|
|
3059
3073
|
{
|
|
3060
|
-
description: "
|
|
3061
|
-
id: "
|
|
3074
|
+
description: "Text-to-image model for photorealistic generation.",
|
|
3075
|
+
id: "Kwai-Kolors/Kolors"
|
|
3062
3076
|
},
|
|
3063
3077
|
{
|
|
3064
3078
|
description: "A powerful text-to-image model.",
|
|
@@ -3419,6 +3433,10 @@ var taskData29 = {
|
|
|
3419
3433
|
{
|
|
3420
3434
|
description: "An instruction dataset with preference ratings on responses.",
|
|
3421
3435
|
id: "openbmb/UltraFeedback"
|
|
3436
|
+
},
|
|
3437
|
+
{
|
|
3438
|
+
description: "A large synthetic dataset for alignment of text generation models.",
|
|
3439
|
+
id: "argilla/magpie-ultra-v0.1"
|
|
3422
3440
|
}
|
|
3423
3441
|
],
|
|
3424
3442
|
demo: {
|
|
@@ -3449,32 +3467,32 @@ var taskData29 = {
|
|
|
3449
3467
|
],
|
|
3450
3468
|
models: [
|
|
3451
3469
|
{
|
|
3452
|
-
description: "A
|
|
3453
|
-
id: "
|
|
3470
|
+
description: "A text-generation model trained to follow instructions.",
|
|
3471
|
+
id: "google/gemma-2-2b-it"
|
|
3454
3472
|
},
|
|
3455
3473
|
{
|
|
3456
|
-
description: "A
|
|
3474
|
+
description: "A code generation model that can generate code in 80+ languages.",
|
|
3457
3475
|
id: "bigcode/starcoder"
|
|
3458
3476
|
},
|
|
3459
3477
|
{
|
|
3460
|
-
description: "
|
|
3461
|
-
id: "
|
|
3478
|
+
description: "Very powerful text generation model trained to follow instructions.",
|
|
3479
|
+
id: "meta-llama/Meta-Llama-3.1-8B-Instruct"
|
|
3462
3480
|
},
|
|
3463
3481
|
{
|
|
3464
3482
|
description: "Small yet powerful text generation model.",
|
|
3465
|
-
id: "microsoft/
|
|
3483
|
+
id: "microsoft/Phi-3-mini-4k-instruct"
|
|
3466
3484
|
},
|
|
3467
3485
|
{
|
|
3468
|
-
description: "A very powerful model that can
|
|
3469
|
-
id: "
|
|
3486
|
+
description: "A very powerful model that can solve mathematical problems.",
|
|
3487
|
+
id: "AI-MO/NuminaMath-7B-TIR"
|
|
3470
3488
|
},
|
|
3471
3489
|
{
|
|
3472
|
-
description: "
|
|
3473
|
-
id: "HuggingFaceH4/
|
|
3490
|
+
description: "Strong coding assistant model.",
|
|
3491
|
+
id: "HuggingFaceH4/starchat2-15b-v0.1"
|
|
3474
3492
|
},
|
|
3475
3493
|
{
|
|
3476
3494
|
description: "Very strong open-source large language model.",
|
|
3477
|
-
id: "
|
|
3495
|
+
id: "mistralai/Mistral-Nemo-Instruct-2407"
|
|
3478
3496
|
}
|
|
3479
3497
|
],
|
|
3480
3498
|
spaces: [
|
|
@@ -3501,7 +3519,7 @@ var taskData29 = {
|
|
|
3501
3519
|
],
|
|
3502
3520
|
summary: "Generating text is the task of generating new text given another text. These models can, for example, fill in incomplete text or paraphrase.",
|
|
3503
3521
|
widgetModels: ["HuggingFaceH4/zephyr-7b-beta"],
|
|
3504
|
-
youtubeId: "
|
|
3522
|
+
youtubeId: "e9gNEAlsOvU"
|
|
3505
3523
|
};
|
|
3506
3524
|
var data_default29 = taskData29;
|
|
3507
3525
|
|
|
@@ -4226,6 +4244,7 @@ var TASKS_MODEL_LIBRARIES = {
|
|
|
4226
4244
|
],
|
|
4227
4245
|
translation: ["transformers", "transformers.js"],
|
|
4228
4246
|
"unconditional-image-generation": ["diffusers"],
|
|
4247
|
+
"video-text-to-text": ["transformers"],
|
|
4229
4248
|
"visual-question-answering": ["transformers", "transformers.js"],
|
|
4230
4249
|
"voice-activity-detection": [],
|
|
4231
4250
|
"zero-shot-classification": ["transformers", "transformers.js"],
|
|
@@ -4285,6 +4304,7 @@ var TASKS_DATA = {
|
|
|
4285
4304
|
"token-classification": getData("token-classification", data_default26),
|
|
4286
4305
|
translation: getData("translation", data_default27),
|
|
4287
4306
|
"unconditional-image-generation": getData("unconditional-image-generation", data_default31),
|
|
4307
|
+
"video-text-to-text": getData("video-text-to-text", data_default16),
|
|
4288
4308
|
"visual-question-answering": getData("visual-question-answering", data_default33),
|
|
4289
4309
|
"voice-activity-detection": void 0,
|
|
4290
4310
|
"zero-shot-classification": getData("zero-shot-classification", data_default34),
|
|
@@ -4522,6 +4542,23 @@ tokenizer = keras_nlp.models.Tokenizer.from_preset("hf://${model.id}")
|
|
|
4522
4542
|
backbone = keras_nlp.models.Backbone.from_preset("hf://${model.id}")
|
|
4523
4543
|
`
|
|
4524
4544
|
];
|
|
4545
|
+
var llama_cpp_python = (model) => [
|
|
4546
|
+
`from llama_cpp import Llama
|
|
4547
|
+
|
|
4548
|
+
llm = Llama.from_pretrained(
|
|
4549
|
+
repo_id="${model.id}",
|
|
4550
|
+
filename="{{GGUF_FILE}}",
|
|
4551
|
+
)
|
|
4552
|
+
|
|
4553
|
+
llm.create_chat_completion(
|
|
4554
|
+
messages = [
|
|
4555
|
+
{
|
|
4556
|
+
"role": "user",
|
|
4557
|
+
"content": "What is the capital of France?"
|
|
4558
|
+
}
|
|
4559
|
+
]
|
|
4560
|
+
)`
|
|
4561
|
+
];
|
|
4525
4562
|
var tf_keras = (model) => [
|
|
4526
4563
|
`# Note: 'keras<3.x' or 'tf_keras' must be installed (legacy)
|
|
4527
4564
|
# See https://github.com/keras-team/tf-keras for more details.
|
|
@@ -5292,6 +5329,12 @@ var MODEL_LIBRARIES_UI_ELEMENTS = {
|
|
|
5292
5329
|
filter: false,
|
|
5293
5330
|
countDownloads: `path:"liveportrait/landmark.onnx"`
|
|
5294
5331
|
},
|
|
5332
|
+
"llama-cpp-python": {
|
|
5333
|
+
prettyLabel: "llama-cpp-python",
|
|
5334
|
+
repoName: "llama-cpp-python",
|
|
5335
|
+
repoUrl: "https://github.com/abetlen/llama-cpp-python",
|
|
5336
|
+
snippets: llama_cpp_python
|
|
5337
|
+
},
|
|
5295
5338
|
mindspore: {
|
|
5296
5339
|
prettyLabel: "MindSpore",
|
|
5297
5340
|
repoName: "mindspore",
|
|
@@ -5414,6 +5457,13 @@ var MODEL_LIBRARIES_UI_ELEMENTS = {
|
|
|
5414
5457
|
snippets: saelens,
|
|
5415
5458
|
filter: false
|
|
5416
5459
|
},
|
|
5460
|
+
sam2: {
|
|
5461
|
+
prettyLabel: "sam2",
|
|
5462
|
+
repoName: "sam2",
|
|
5463
|
+
repoUrl: "https://github.com/facebookresearch/segment-anything-2",
|
|
5464
|
+
filter: false,
|
|
5465
|
+
countDownloads: `path_extension:"pt"`
|
|
5466
|
+
},
|
|
5417
5467
|
"sample-factory": {
|
|
5418
5468
|
prettyLabel: "sample-factory",
|
|
5419
5469
|
repoName: "sample-factory",
|
|
@@ -6640,7 +6690,6 @@ var LOCAL_APPS = {
|
|
|
6640
6690
|
docsUrl: "https://diffusionbee.com",
|
|
6641
6691
|
mainTask: "text-to-image",
|
|
6642
6692
|
macOSOnly: true,
|
|
6643
|
-
comingSoon: true,
|
|
6644
6693
|
displayOnModelPage: (model) => model.library_name === "diffusers" && model.pipeline_tag === "text-to-image",
|
|
6645
6694
|
deeplink: (model) => new URL(`diffusionbee://open_from_hf?model=${model.id}`)
|
|
6646
6695
|
}
|
package/dist/index.js
CHANGED
|
@@ -1350,6 +1350,12 @@ var PIPELINE_DATA = {
|
|
|
1350
1350
|
modality: "cv",
|
|
1351
1351
|
color: "indigo"
|
|
1352
1352
|
},
|
|
1353
|
+
"video-text-to-text": {
|
|
1354
|
+
name: "Video-Text-to-Text",
|
|
1355
|
+
modality: "multimodal",
|
|
1356
|
+
color: "blue",
|
|
1357
|
+
hideInDatasets: true
|
|
1358
|
+
},
|
|
1353
1359
|
other: {
|
|
1354
1360
|
name: "Other",
|
|
1355
1361
|
modality: "other",
|
|
@@ -1693,8 +1699,12 @@ var taskData5 = {
|
|
|
1693
1699
|
],
|
|
1694
1700
|
spaces: [
|
|
1695
1701
|
{
|
|
1696
|
-
description: "A leaderboard to rank
|
|
1702
|
+
description: "A leaderboard to rank text feature extraction models based on a benchmark.",
|
|
1697
1703
|
id: "mteb/leaderboard"
|
|
1704
|
+
},
|
|
1705
|
+
{
|
|
1706
|
+
description: "A leaderboard to rank best feature extraction models based on human feedback.",
|
|
1707
|
+
id: "mteb/arena"
|
|
1698
1708
|
}
|
|
1699
1709
|
],
|
|
1700
1710
|
summary: "Feature extraction is the task of extracting features learnt in a model.",
|
|
@@ -1899,15 +1909,19 @@ var taskData8 = {
|
|
|
1899
1909
|
},
|
|
1900
1910
|
{
|
|
1901
1911
|
description: "A strong image feature extraction model.",
|
|
1902
|
-
id: "
|
|
1912
|
+
id: "nvidia/MambaVision-T-1K"
|
|
1903
1913
|
},
|
|
1904
1914
|
{
|
|
1905
|
-
description: "A robust image feature extraction
|
|
1915
|
+
description: "A robust image feature extraction model.",
|
|
1906
1916
|
id: "facebook/dino-vitb16"
|
|
1907
1917
|
},
|
|
1908
1918
|
{
|
|
1909
|
-
description: "Strong image
|
|
1919
|
+
description: "Strong image feature extraction model made for information retrieval from documents.",
|
|
1910
1920
|
id: "vidore/colpali"
|
|
1921
|
+
},
|
|
1922
|
+
{
|
|
1923
|
+
description: "Strong image feature extraction model that can be used on images and documents.",
|
|
1924
|
+
id: "OpenGVLab/InternViT-6B-448px-V1-2"
|
|
1911
1925
|
}
|
|
1912
1926
|
],
|
|
1913
1927
|
spaces: [],
|
|
@@ -1959,8 +1973,8 @@ var taskData9 = {
|
|
|
1959
1973
|
],
|
|
1960
1974
|
models: [
|
|
1961
1975
|
{
|
|
1962
|
-
description: "
|
|
1963
|
-
id: "
|
|
1976
|
+
description: "An image-to-image model to improve image resolution.",
|
|
1977
|
+
id: "fal/AuraSR-v2"
|
|
1964
1978
|
},
|
|
1965
1979
|
{
|
|
1966
1980
|
description: "A model that increases the resolution of an image.",
|
|
@@ -2178,7 +2192,7 @@ var taskData11 = {
|
|
|
2178
2192
|
],
|
|
2179
2193
|
summary: "Image-text-to-text models take in an image and text prompt and output text. These models are also called vision-language models, or VLMs. The difference from image-to-text models is that these models take an additional text input, not restricting the model to certain use cases like image captioning, and may also be trained to accept a conversation as input.",
|
|
2180
2194
|
widgetModels: ["microsoft/kosmos-2-patch14-224"],
|
|
2181
|
-
youtubeId: ""
|
|
2195
|
+
youtubeId: "IoGaGfU1CIg"
|
|
2182
2196
|
};
|
|
2183
2197
|
var data_default11 = taskData11;
|
|
2184
2198
|
|
|
@@ -2229,16 +2243,16 @@ var taskData12 = {
|
|
|
2229
2243
|
id: "facebook/detr-resnet-50-panoptic"
|
|
2230
2244
|
},
|
|
2231
2245
|
{
|
|
2232
|
-
description: "
|
|
2233
|
-
id: "
|
|
2246
|
+
description: "Background removal model.",
|
|
2247
|
+
id: "briaai/RMBG-1.4"
|
|
2234
2248
|
},
|
|
2235
2249
|
{
|
|
2236
2250
|
description: "Semantic segmentation model trained on ADE20k benchmark dataset with 512x512 resolution.",
|
|
2237
2251
|
id: "nvidia/segformer-b0-finetuned-ade-512-512"
|
|
2238
2252
|
},
|
|
2239
2253
|
{
|
|
2240
|
-
description: "
|
|
2241
|
-
id: "
|
|
2254
|
+
description: "A multipurpose image segmentation model for high resolution images.",
|
|
2255
|
+
id: "ZhengPeng7/BiRefNet"
|
|
2242
2256
|
},
|
|
2243
2257
|
{
|
|
2244
2258
|
description: "Panoptic segmentation model trained COCO (common objects) dataset.",
|
|
@@ -2302,13 +2316,13 @@ var taskData13 = {
|
|
|
2302
2316
|
},
|
|
2303
2317
|
{
|
|
2304
2318
|
description: "Very strong mask generation model.",
|
|
2305
|
-
id: "facebook/
|
|
2319
|
+
id: "facebook/sam2-hiera-large"
|
|
2306
2320
|
}
|
|
2307
2321
|
],
|
|
2308
2322
|
spaces: [
|
|
2309
2323
|
{
|
|
2310
|
-
description: "An application that combines a mask generation model with
|
|
2311
|
-
id: "
|
|
2324
|
+
description: "An application that combines a mask generation model with a zero-shot object detection model for text-guided image segmentation.",
|
|
2325
|
+
id: "merve/OWLSAM2"
|
|
2312
2326
|
},
|
|
2313
2327
|
{
|
|
2314
2328
|
description: "An application that compares the performance of a large and a small mask generation model.",
|
|
@@ -2316,7 +2330,7 @@ var taskData13 = {
|
|
|
2316
2330
|
},
|
|
2317
2331
|
{
|
|
2318
2332
|
description: "An application based on an improved mask generation model.",
|
|
2319
|
-
id: "
|
|
2333
|
+
id: "SkalskiP/segment-anything-model-2"
|
|
2320
2334
|
},
|
|
2321
2335
|
{
|
|
2322
2336
|
description: "An application to remove objects from videos using mask generation models.",
|
|
@@ -3012,15 +3026,15 @@ var taskData24 = {
|
|
|
3012
3026
|
models: [
|
|
3013
3027
|
{
|
|
3014
3028
|
description: "One of the most powerful image generation models that can generate realistic outputs.",
|
|
3015
|
-
id: "
|
|
3029
|
+
id: "black-forest-labs/FLUX.1-dev"
|
|
3016
3030
|
},
|
|
3017
3031
|
{
|
|
3018
3032
|
description: "A powerful yet fast image generation model.",
|
|
3019
3033
|
id: "latent-consistency/lcm-lora-sdxl"
|
|
3020
3034
|
},
|
|
3021
3035
|
{
|
|
3022
|
-
description: "
|
|
3023
|
-
id: "
|
|
3036
|
+
description: "Text-to-image model for photorealistic generation.",
|
|
3037
|
+
id: "Kwai-Kolors/Kolors"
|
|
3024
3038
|
},
|
|
3025
3039
|
{
|
|
3026
3040
|
description: "A powerful text-to-image model.",
|
|
@@ -3381,6 +3395,10 @@ var taskData29 = {
|
|
|
3381
3395
|
{
|
|
3382
3396
|
description: "An instruction dataset with preference ratings on responses.",
|
|
3383
3397
|
id: "openbmb/UltraFeedback"
|
|
3398
|
+
},
|
|
3399
|
+
{
|
|
3400
|
+
description: "A large synthetic dataset for alignment of text generation models.",
|
|
3401
|
+
id: "argilla/magpie-ultra-v0.1"
|
|
3384
3402
|
}
|
|
3385
3403
|
],
|
|
3386
3404
|
demo: {
|
|
@@ -3411,32 +3429,32 @@ var taskData29 = {
|
|
|
3411
3429
|
],
|
|
3412
3430
|
models: [
|
|
3413
3431
|
{
|
|
3414
|
-
description: "A
|
|
3415
|
-
id: "
|
|
3432
|
+
description: "A text-generation model trained to follow instructions.",
|
|
3433
|
+
id: "google/gemma-2-2b-it"
|
|
3416
3434
|
},
|
|
3417
3435
|
{
|
|
3418
|
-
description: "A
|
|
3436
|
+
description: "A code generation model that can generate code in 80+ languages.",
|
|
3419
3437
|
id: "bigcode/starcoder"
|
|
3420
3438
|
},
|
|
3421
3439
|
{
|
|
3422
|
-
description: "
|
|
3423
|
-
id: "
|
|
3440
|
+
description: "Very powerful text generation model trained to follow instructions.",
|
|
3441
|
+
id: "meta-llama/Meta-Llama-3.1-8B-Instruct"
|
|
3424
3442
|
},
|
|
3425
3443
|
{
|
|
3426
3444
|
description: "Small yet powerful text generation model.",
|
|
3427
|
-
id: "microsoft/
|
|
3445
|
+
id: "microsoft/Phi-3-mini-4k-instruct"
|
|
3428
3446
|
},
|
|
3429
3447
|
{
|
|
3430
|
-
description: "A very powerful model that can
|
|
3431
|
-
id: "
|
|
3448
|
+
description: "A very powerful model that can solve mathematical problems.",
|
|
3449
|
+
id: "AI-MO/NuminaMath-7B-TIR"
|
|
3432
3450
|
},
|
|
3433
3451
|
{
|
|
3434
|
-
description: "
|
|
3435
|
-
id: "HuggingFaceH4/
|
|
3452
|
+
description: "Strong coding assistant model.",
|
|
3453
|
+
id: "HuggingFaceH4/starchat2-15b-v0.1"
|
|
3436
3454
|
},
|
|
3437
3455
|
{
|
|
3438
3456
|
description: "Very strong open-source large language model.",
|
|
3439
|
-
id: "
|
|
3457
|
+
id: "mistralai/Mistral-Nemo-Instruct-2407"
|
|
3440
3458
|
}
|
|
3441
3459
|
],
|
|
3442
3460
|
spaces: [
|
|
@@ -3463,7 +3481,7 @@ var taskData29 = {
|
|
|
3463
3481
|
],
|
|
3464
3482
|
summary: "Generating text is the task of generating new text given another text. These models can, for example, fill in incomplete text or paraphrase.",
|
|
3465
3483
|
widgetModels: ["HuggingFaceH4/zephyr-7b-beta"],
|
|
3466
|
-
youtubeId: "
|
|
3484
|
+
youtubeId: "e9gNEAlsOvU"
|
|
3467
3485
|
};
|
|
3468
3486
|
var data_default29 = taskData29;
|
|
3469
3487
|
|
|
@@ -4188,6 +4206,7 @@ var TASKS_MODEL_LIBRARIES = {
|
|
|
4188
4206
|
],
|
|
4189
4207
|
translation: ["transformers", "transformers.js"],
|
|
4190
4208
|
"unconditional-image-generation": ["diffusers"],
|
|
4209
|
+
"video-text-to-text": ["transformers"],
|
|
4191
4210
|
"visual-question-answering": ["transformers", "transformers.js"],
|
|
4192
4211
|
"voice-activity-detection": [],
|
|
4193
4212
|
"zero-shot-classification": ["transformers", "transformers.js"],
|
|
@@ -4247,6 +4266,7 @@ var TASKS_DATA = {
|
|
|
4247
4266
|
"token-classification": getData("token-classification", data_default26),
|
|
4248
4267
|
translation: getData("translation", data_default27),
|
|
4249
4268
|
"unconditional-image-generation": getData("unconditional-image-generation", data_default31),
|
|
4269
|
+
"video-text-to-text": getData("video-text-to-text", data_default16),
|
|
4250
4270
|
"visual-question-answering": getData("visual-question-answering", data_default33),
|
|
4251
4271
|
"voice-activity-detection": void 0,
|
|
4252
4272
|
"zero-shot-classification": getData("zero-shot-classification", data_default34),
|
|
@@ -4484,6 +4504,23 @@ tokenizer = keras_nlp.models.Tokenizer.from_preset("hf://${model.id}")
|
|
|
4484
4504
|
backbone = keras_nlp.models.Backbone.from_preset("hf://${model.id}")
|
|
4485
4505
|
`
|
|
4486
4506
|
];
|
|
4507
|
+
var llama_cpp_python = (model) => [
|
|
4508
|
+
`from llama_cpp import Llama
|
|
4509
|
+
|
|
4510
|
+
llm = Llama.from_pretrained(
|
|
4511
|
+
repo_id="${model.id}",
|
|
4512
|
+
filename="{{GGUF_FILE}}",
|
|
4513
|
+
)
|
|
4514
|
+
|
|
4515
|
+
llm.create_chat_completion(
|
|
4516
|
+
messages = [
|
|
4517
|
+
{
|
|
4518
|
+
"role": "user",
|
|
4519
|
+
"content": "What is the capital of France?"
|
|
4520
|
+
}
|
|
4521
|
+
]
|
|
4522
|
+
)`
|
|
4523
|
+
];
|
|
4487
4524
|
var tf_keras = (model) => [
|
|
4488
4525
|
`# Note: 'keras<3.x' or 'tf_keras' must be installed (legacy)
|
|
4489
4526
|
# See https://github.com/keras-team/tf-keras for more details.
|
|
@@ -5254,6 +5291,12 @@ var MODEL_LIBRARIES_UI_ELEMENTS = {
|
|
|
5254
5291
|
filter: false,
|
|
5255
5292
|
countDownloads: `path:"liveportrait/landmark.onnx"`
|
|
5256
5293
|
},
|
|
5294
|
+
"llama-cpp-python": {
|
|
5295
|
+
prettyLabel: "llama-cpp-python",
|
|
5296
|
+
repoName: "llama-cpp-python",
|
|
5297
|
+
repoUrl: "https://github.com/abetlen/llama-cpp-python",
|
|
5298
|
+
snippets: llama_cpp_python
|
|
5299
|
+
},
|
|
5257
5300
|
mindspore: {
|
|
5258
5301
|
prettyLabel: "MindSpore",
|
|
5259
5302
|
repoName: "mindspore",
|
|
@@ -5376,6 +5419,13 @@ var MODEL_LIBRARIES_UI_ELEMENTS = {
|
|
|
5376
5419
|
snippets: saelens,
|
|
5377
5420
|
filter: false
|
|
5378
5421
|
},
|
|
5422
|
+
sam2: {
|
|
5423
|
+
prettyLabel: "sam2",
|
|
5424
|
+
repoName: "sam2",
|
|
5425
|
+
repoUrl: "https://github.com/facebookresearch/segment-anything-2",
|
|
5426
|
+
filter: false,
|
|
5427
|
+
countDownloads: `path_extension:"pt"`
|
|
5428
|
+
},
|
|
5379
5429
|
"sample-factory": {
|
|
5380
5430
|
prettyLabel: "sample-factory",
|
|
5381
5431
|
repoName: "sample-factory",
|
|
@@ -6602,7 +6652,6 @@ var LOCAL_APPS = {
|
|
|
6602
6652
|
docsUrl: "https://diffusionbee.com",
|
|
6603
6653
|
mainTask: "text-to-image",
|
|
6604
6654
|
macOSOnly: true,
|
|
6605
|
-
comingSoon: true,
|
|
6606
6655
|
displayOnModelPage: (model) => model.library_name === "diffusers" && model.pipeline_tag === "text-to-image",
|
|
6607
6656
|
deeplink: (model) => new URL(`diffusionbee://open_from_hf?model=${model.id}`)
|
|
6608
6657
|
}
|
package/dist/src/local-apps.d.ts
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"local-apps.d.ts","sourceRoot":"","sources":["../../src/local-apps.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,SAAS,EAAE,MAAM,cAAc,CAAC;AAC9C,OAAO,KAAK,EAAE,YAAY,EAAE,MAAM,aAAa,CAAC;AAEhD,MAAM,WAAW,eAAe;IAC/B;;OAEG;IACH,KAAK,EAAE,MAAM,CAAC;IACd;;OAEG;IACH,KAAK,CAAC,EAAE,MAAM,CAAC;IACf;;OAEG;IACH,OAAO,EAAE,MAAM,CAAC;CAChB;AAED;;GAEG;AACH,MAAM,MAAM,QAAQ,GAAG;IACtB;;OAEG;IACH,WAAW,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,OAAO,EAAE,MAAM,CAAC;IAChB;;OAEG;IACH,QAAQ,EAAE,YAAY,CAAC;IACvB;;OAEG;IACH,SAAS,CAAC,EAAE,OAAO,CAAC;IAEpB,UAAU,CAAC,EAAE,OAAO,CAAC;IACrB;;OAEG;IACH,kBAAkB,EAAE,CAAC,KAAK,EAAE,SAAS,KAAK,OAAO,CAAC;CAClD,GAAG,CACD;IACA;;OAEG;IACH,QAAQ,EAAE,CAAC,KAAK,EAAE,SAAS,EAAE,QAAQ,CAAC,EAAE,MAAM,KAAK,GAAG,CAAC;CACtD,GACD;IACA;;;OAGG;IACH,OAAO,EAAE,CAAC,KAAK,EAAE,SAAS,EAAE,QAAQ,CAAC,EAAE,MAAM,KAAK,MAAM,GAAG,MAAM,EAAE,GAAG,eAAe,GAAG,eAAe,EAAE,CAAC;CACzG,CACH,CAAC;AAEF,iBAAS,WAAW,CAAC,KAAK,EAAE,SAAS,WAEpC;AAuCD;;;;;;;;;;GAUG;AACH,eAAO,MAAM,UAAU;;;;;;yBAhDS,SAAS,aAAa,MAAM,KAAG,eAAe,EAAE
|
|
1
|
+
{"version":3,"file":"local-apps.d.ts","sourceRoot":"","sources":["../../src/local-apps.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,SAAS,EAAE,MAAM,cAAc,CAAC;AAC9C,OAAO,KAAK,EAAE,YAAY,EAAE,MAAM,aAAa,CAAC;AAEhD,MAAM,WAAW,eAAe;IAC/B;;OAEG;IACH,KAAK,EAAE,MAAM,CAAC;IACd;;OAEG;IACH,KAAK,CAAC,EAAE,MAAM,CAAC;IACf;;OAEG;IACH,OAAO,EAAE,MAAM,CAAC;CAChB;AAED;;GAEG;AACH,MAAM,MAAM,QAAQ,GAAG;IACtB;;OAEG;IACH,WAAW,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,OAAO,EAAE,MAAM,CAAC;IAChB;;OAEG;IACH,QAAQ,EAAE,YAAY,CAAC;IACvB;;OAEG;IACH,SAAS,CAAC,EAAE,OAAO,CAAC;IAEpB,UAAU,CAAC,EAAE,OAAO,CAAC;IACrB;;OAEG;IACH,kBAAkB,EAAE,CAAC,KAAK,EAAE,SAAS,KAAK,OAAO,CAAC;CAClD,GAAG,CACD;IACA;;OAEG;IACH,QAAQ,EAAE,CAAC,KAAK,EAAE,SAAS,EAAE,QAAQ,CAAC,EAAE,MAAM,KAAK,GAAG,CAAC;CACtD,GACD;IACA;;;OAGG;IACH,OAAO,EAAE,CAAC,KAAK,EAAE,SAAS,EAAE,QAAQ,CAAC,EAAE,MAAM,KAAK,MAAM,GAAG,MAAM,EAAE,GAAG,eAAe,GAAG,eAAe,EAAE,CAAC;CACzG,CACH,CAAC;AAEF,iBAAS,WAAW,CAAC,KAAK,EAAE,SAAS,WAEpC;AAuCD;;;;;;;;;;GAUG;AACH,eAAO,MAAM,UAAU;;;;;;yBAhDS,SAAS,aAAa,MAAM,KAAG,eAAe,EAAE;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CA8I5C,CAAC;AAErC,MAAM,MAAM,WAAW,GAAG,MAAM,OAAO,UAAU,CAAC"}
|
|
@@ -16,6 +16,7 @@ export declare const flair: (model: ModelData) => string[];
|
|
|
16
16
|
export declare const gliner: (model: ModelData) => string[];
|
|
17
17
|
export declare const keras: (model: ModelData) => string[];
|
|
18
18
|
export declare const keras_nlp: (model: ModelData) => string[];
|
|
19
|
+
export declare const llama_cpp_python: (model: ModelData) => string[];
|
|
19
20
|
export declare const tf_keras: (model: ModelData) => string[];
|
|
20
21
|
export declare const mamba_ssm: (model: ModelData) => string[];
|
|
21
22
|
export declare const mars5_tts: (model: ModelData) => string[];
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"model-libraries-snippets.d.ts","sourceRoot":"","sources":["../../src/model-libraries-snippets.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,SAAS,EAAE,MAAM,cAAc,CAAC;AAY9C,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAKjD,CAAC;AAkBF,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAKjD,CAAC;AAEF,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAIjD,CAAC;AAEF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EAkBlD,CAAC;AAMF,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAIjD,CAAC;AAEF,eAAO,MAAM,KAAK,UAAW,SAAS,KAAG,MAAM,EAI9C,CAAC;AAEF,eAAO,MAAM,iBAAiB,UAAW,SAAS,KAAG,MAAM,EA6C1D,CAAC;AA+BF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EAUlD,CAAC;AAEF,eAAO,MAAM,MAAM,UAAW,SAAS,KAAG,MAAM,EAgB/C,CAAC;AAEF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EAMlD,CAAC;AAEF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EASlD,CAAC;AAIF,eAAO,MAAM,MAAM,UAAW,SAAS,KAAG,MAAM,EAO/C,CAAC;AAEF,eAAO,MAAM,OAAO,UAAW,SAAS,KAAG,MAAM,EAMhD,CAAC;AAEF,eAAO,MAAM,KAAK,UAAW,SAAS,KAAG,MAAM,EAI9C,CAAC;AAEF,eAAO,MAAM,MAAM,UAAW,SAAS,KAAG,MAAM,EAI/C,CAAC;AAEF,eAAO,MAAM,KAAK,UAAW,SAAS,KAAG,MAAM,EAS9C,CAAC;AAEF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EAUlD,CAAC;AAEF,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAOjD,CAAC;AAEF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EAIlD,CAAC;AAEF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EAKlD,CAAC;AAEF,eAAO,MAAM,aAAa,QAAO,MAAM,EAQtC,CAAC;AAEF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EAKlD,CAAC;AAEF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EAsBlD,CAAC;AAEF,eAAO,MAAM,uBAAuB,UAAW,SAAS,KAAG,MAAM,EAehE,CAAC;AAiBF,eAAO,MAAM,cAAc,UAAW,SAAS,KAAG,MAAM,EAKvD,CAAC;AAyBF,eAAO,MAAM,aAAa,UAAW,SAAS,KAAG,MAAM,EAOtD,CAAC;AAEF,eAAO,MAAM,IAAI,UAAW,SAAS,KAAG,MAAM,EAI7C,CAAC;AAEF,eAAO,MAAM,OAAO,QAA6B,MAAM,EAQtD,CAAC;AAsCF,eAAO,MAAM,OAAO,UAAW,SAAS,KAAG,MAAM,EAehD,CAAC;AAEF,eAAO,MAAM,kBAAkB,UAAW,SAAS,KAAG,MAAM,EAmC3D,CAAC;AAEF,eAAO,MAAM,MAAM,UAAW,SAAS,KAAG,MAAM,EAI/C,CAAC;AAEF,eAAO,MAAM,aAAa,UAAW,SAAS,KAAG,MAAM,EAEtD,CAAC;AAEF,eAAO,MAAM,oBAAoB,UAAW,SAAS,KAAG,MAAM,EAI7D,CAAC;AAEF,eAAO,MAAM,MAAM,UAAW,SAAS,KAAG,MAAM,EAI/C,CAAC;AAEF,eAAO,MAAM,KAAK,UAAW,SAAS,KAAG,MAAM,EAU9C,CAAC;AAEF,eAAO,MAAM,WAAW,UAAW,SAAS,KAAG,MAAM,EAIpD,CAAC;AAEF,eAAO,MAAM,MAAM,UAAW,SAAS,KAAG,MAAM,EAK/C,CAAC;AAkBF,eAAO,MAAM,WAAW,UAAW,SAAS,KAAG,MAAM,EAkBpD,CAAC;AAEF,eAAO,MAAM,YAAY,UAAW,SAAS,KAAG,MAAM,EA4CrD,CAAC;AAEF,eAAO,MAAM,cAAc,UAAW,SAAS,KAAG,MAAM,EAcvD,CAAC;AAiBF,eAAO,MAAM,IAAI,UAAW,SAAS,KAAG,MAAM,EAkB7C,CAAC;AAEF,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAKjD,CAAC;AAEF,eAAO,MAAM,gBAAgB,UAAW,SAAS,KAAG,MAAM,EAMzD,CAAC;AAgBF,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAEjD,CAAC;AAEF,eAAO,MAAM,MAAM,QAA6B,MAAM,EAMrD,CAAC;AAEF,eAAO,MAAM,UAAU,UAAW,SAAS,KAAG,MAAM,EAInD,CAAC;AAEF,eAAO,MAAM,OAAO,QAAO,MAAM,EAYhC,CAAC;AAEF,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAYjD,CAAC;AAEF,eAAO,MAAM,GAAG,UAAW,SAAS,KAAG,MAAM,EAK5C,CAAC;AAEF,eAAO,MAAM,KAAK,UAAW,SAAS,KAAG,MAAM,EAI9C,CAAC;AAEF,eAAO,MAAM,IAAI,UAAW,SAAS,KAAG,MAAM,EAQ7C,CAAC;AAEF,eAAO,MAAM,MAAM,UAAW,SAAS,KAAG,MAAM,EAI/C,CAAC;AA6BF,eAAO,MAAM,UAAU,UAAW,SAAS,KAAG,MAAM,EAUnD,CAAC;AAEF,eAAO,MAAM,UAAU,QAAO,MAAM,EAYnC,CAAC"}
|
|
1
|
+
{"version":3,"file":"model-libraries-snippets.d.ts","sourceRoot":"","sources":["../../src/model-libraries-snippets.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,SAAS,EAAE,MAAM,cAAc,CAAC;AAY9C,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAKjD,CAAC;AAkBF,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAKjD,CAAC;AAEF,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAIjD,CAAC;AAEF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EAkBlD,CAAC;AAMF,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAIjD,CAAC;AAEF,eAAO,MAAM,KAAK,UAAW,SAAS,KAAG,MAAM,EAI9C,CAAC;AAEF,eAAO,MAAM,iBAAiB,UAAW,SAAS,KAAG,MAAM,EA6C1D,CAAC;AA+BF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EAUlD,CAAC;AAEF,eAAO,MAAM,MAAM,UAAW,SAAS,KAAG,MAAM,EAgB/C,CAAC;AAEF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EAMlD,CAAC;AAEF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EASlD,CAAC;AAIF,eAAO,MAAM,MAAM,UAAW,SAAS,KAAG,MAAM,EAO/C,CAAC;AAEF,eAAO,MAAM,OAAO,UAAW,SAAS,KAAG,MAAM,EAMhD,CAAC;AAEF,eAAO,MAAM,KAAK,UAAW,SAAS,KAAG,MAAM,EAI9C,CAAC;AAEF,eAAO,MAAM,MAAM,UAAW,SAAS,KAAG,MAAM,EAI/C,CAAC;AAEF,eAAO,MAAM,KAAK,UAAW,SAAS,KAAG,MAAM,EAS9C,CAAC;AAEF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EAUlD,CAAC;AAEF,eAAO,MAAM,gBAAgB,UAAW,SAAS,KAAG,MAAM,EAgBzD,CAAC;AAEF,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAOjD,CAAC;AAEF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EAIlD,CAAC;AAEF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EAKlD,CAAC;AAEF,eAAO,MAAM,aAAa,QAAO,MAAM,EAQtC,CAAC;AAEF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EAKlD,CAAC;AAEF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EAsBlD,CAAC;AAEF,eAAO,MAAM,uBAAuB,UAAW,SAAS,KAAG,MAAM,EAehE,CAAC;AAiBF,eAAO,MAAM,cAAc,UAAW,SAAS,KAAG,MAAM,EAKvD,CAAC;AAyBF,eAAO,MAAM,aAAa,UAAW,SAAS,KAAG,MAAM,EAOtD,CAAC;AAEF,eAAO,MAAM,IAAI,UAAW,SAAS,KAAG,MAAM,EAI7C,CAAC;AAEF,eAAO,MAAM,OAAO,QAA6B,MAAM,EAQtD,CAAC;AAsCF,eAAO,MAAM,OAAO,UAAW,SAAS,KAAG,MAAM,EAehD,CAAC;AAEF,eAAO,MAAM,kBAAkB,UAAW,SAAS,KAAG,MAAM,EAmC3D,CAAC;AAEF,eAAO,MAAM,MAAM,UAAW,SAAS,KAAG,MAAM,EAI/C,CAAC;AAEF,eAAO,MAAM,aAAa,UAAW,SAAS,KAAG,MAAM,EAEtD,CAAC;AAEF,eAAO,MAAM,oBAAoB,UAAW,SAAS,KAAG,MAAM,EAI7D,CAAC;AAEF,eAAO,MAAM,MAAM,UAAW,SAAS,KAAG,MAAM,EAI/C,CAAC;AAEF,eAAO,MAAM,KAAK,UAAW,SAAS,KAAG,MAAM,EAU9C,CAAC;AAEF,eAAO,MAAM,WAAW,UAAW,SAAS,KAAG,MAAM,EAIpD,CAAC;AAEF,eAAO,MAAM,MAAM,UAAW,SAAS,KAAG,MAAM,EAK/C,CAAC;AAkBF,eAAO,MAAM,WAAW,UAAW,SAAS,KAAG,MAAM,EAkBpD,CAAC;AAEF,eAAO,MAAM,YAAY,UAAW,SAAS,KAAG,MAAM,EA4CrD,CAAC;AAEF,eAAO,MAAM,cAAc,UAAW,SAAS,KAAG,MAAM,EAcvD,CAAC;AAiBF,eAAO,MAAM,IAAI,UAAW,SAAS,KAAG,MAAM,EAkB7C,CAAC;AAEF,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAKjD,CAAC;AAEF,eAAO,MAAM,gBAAgB,UAAW,SAAS,KAAG,MAAM,EAMzD,CAAC;AAgBF,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAEjD,CAAC;AAEF,eAAO,MAAM,MAAM,QAA6B,MAAM,EAMrD,CAAC;AAEF,eAAO,MAAM,UAAU,UAAW,SAAS,KAAG,MAAM,EAInD,CAAC;AAEF,eAAO,MAAM,OAAO,QAAO,MAAM,EAYhC,CAAC;AAEF,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAYjD,CAAC;AAEF,eAAO,MAAM,GAAG,UAAW,SAAS,KAAG,MAAM,EAK5C,CAAC;AAEF,eAAO,MAAM,KAAK,UAAW,SAAS,KAAG,MAAM,EAI9C,CAAC;AAEF,eAAO,MAAM,IAAI,UAAW,SAAS,KAAG,MAAM,EAQ7C,CAAC;AAEF,eAAO,MAAM,MAAM,UAAW,SAAS,KAAG,MAAM,EAI/C,CAAC;AA6BF,eAAO,MAAM,UAAU,UAAW,SAAS,KAAG,MAAM,EAUnD,CAAC;AAEF,eAAO,MAAM,UAAU,QAAO,MAAM,EAYnC,CAAC"}
|
|
@@ -308,6 +308,12 @@ export declare const MODEL_LIBRARIES_UI_ELEMENTS: {
|
|
|
308
308
|
filter: false;
|
|
309
309
|
countDownloads: string;
|
|
310
310
|
};
|
|
311
|
+
"llama-cpp-python": {
|
|
312
|
+
prettyLabel: string;
|
|
313
|
+
repoName: string;
|
|
314
|
+
repoUrl: string;
|
|
315
|
+
snippets: (model: ModelData) => string[];
|
|
316
|
+
};
|
|
311
317
|
mindspore: {
|
|
312
318
|
prettyLabel: string;
|
|
313
319
|
repoName: string;
|
|
@@ -430,6 +436,13 @@ export declare const MODEL_LIBRARIES_UI_ELEMENTS: {
|
|
|
430
436
|
snippets: () => string[];
|
|
431
437
|
filter: false;
|
|
432
438
|
};
|
|
439
|
+
sam2: {
|
|
440
|
+
prettyLabel: string;
|
|
441
|
+
repoName: string;
|
|
442
|
+
repoUrl: string;
|
|
443
|
+
filter: false;
|
|
444
|
+
countDownloads: string;
|
|
445
|
+
};
|
|
433
446
|
"sample-factory": {
|
|
434
447
|
prettyLabel: string;
|
|
435
448
|
repoName: string;
|
|
@@ -592,6 +605,6 @@ export declare const MODEL_LIBRARIES_UI_ELEMENTS: {
|
|
|
592
605
|
};
|
|
593
606
|
};
|
|
594
607
|
export type ModelLibraryKey = keyof typeof MODEL_LIBRARIES_UI_ELEMENTS;
|
|
595
|
-
export declare const ALL_MODEL_LIBRARY_KEYS: ("adapter-transformers" | "allennlp" | "asteroid" | "audiocraft" | "audioseal" | "bertopic" | "big_vision" | "birefnet" | "bm25s" | "champ" | "chat_tts" | "colpali" | "depth-anything-v2" | "diffree" | "diffusers" | "doctr" | "edsnlp" | "elm" | "espnet" | "fairseq" | "fastai" | "fasttext" | "flair" | "gemma.cpp" | "gliner" | "glyph-byt5" | "grok" | "hallo" | "hunyuan-dit" | "keras" | "tf-keras" | "keras-nlp" | "k2" | "liveportrait" | "mindspore" | "mamba-ssm" | "mars5-tts" | "mesh-anything" | "ml-agents" | "mlx" | "mlx-image" | "mlc-llm" | "nemo" | "open_clip" | "paddlenlp" | "peft" | "pyannote-audio" | "pythae" | "recurrentgemma" | "saelens" | "sample-factory" | "sentence-transformers" | "setfit" | "sklearn" | "spacy" | "span-marker" | "speechbrain" | "stable-audio-tools" | "diffusion-single-file" | "stable-baselines3" | "stanza" | "tensorflowtts" | "tic-clip" | "timesfm" | "timm" | "transformers" | "transformers.js" | "unity-sentis" | "voicecraft" | "whisperkit")[];
|
|
596
|
-
export declare const ALL_DISPLAY_MODEL_LIBRARY_KEYS: ("adapter-transformers" | "allennlp" | "asteroid" | "audiocraft" | "audioseal" | "bertopic" | "big_vision" | "birefnet" | "bm25s" | "champ" | "chat_tts" | "colpali" | "depth-anything-v2" | "diffree" | "diffusers" | "doctr" | "edsnlp" | "elm" | "espnet" | "fairseq" | "fastai" | "fasttext" | "flair" | "gemma.cpp" | "gliner" | "glyph-byt5" | "grok" | "hallo" | "hunyuan-dit" | "keras" | "tf-keras" | "keras-nlp" | "k2" | "liveportrait" | "mindspore" | "mamba-ssm" | "mars5-tts" | "mesh-anything" | "ml-agents" | "mlx" | "mlx-image" | "mlc-llm" | "nemo" | "open_clip" | "paddlenlp" | "peft" | "pyannote-audio" | "pythae" | "recurrentgemma" | "saelens" | "sample-factory" | "sentence-transformers" | "setfit" | "sklearn" | "spacy" | "span-marker" | "speechbrain" | "stable-audio-tools" | "diffusion-single-file" | "stable-baselines3" | "stanza" | "tensorflowtts" | "tic-clip" | "timesfm" | "timm" | "transformers" | "transformers.js" | "unity-sentis" | "voicecraft" | "whisperkit")[];
|
|
608
|
+
export declare const ALL_MODEL_LIBRARY_KEYS: ("adapter-transformers" | "allennlp" | "asteroid" | "audiocraft" | "audioseal" | "bertopic" | "big_vision" | "birefnet" | "bm25s" | "champ" | "chat_tts" | "colpali" | "depth-anything-v2" | "diffree" | "diffusers" | "doctr" | "edsnlp" | "elm" | "espnet" | "fairseq" | "fastai" | "fasttext" | "flair" | "gemma.cpp" | "gliner" | "glyph-byt5" | "grok" | "hallo" | "hunyuan-dit" | "keras" | "tf-keras" | "keras-nlp" | "k2" | "liveportrait" | "llama-cpp-python" | "mindspore" | "mamba-ssm" | "mars5-tts" | "mesh-anything" | "ml-agents" | "mlx" | "mlx-image" | "mlc-llm" | "nemo" | "open_clip" | "paddlenlp" | "peft" | "pyannote-audio" | "pythae" | "recurrentgemma" | "saelens" | "sam2" | "sample-factory" | "sentence-transformers" | "setfit" | "sklearn" | "spacy" | "span-marker" | "speechbrain" | "stable-audio-tools" | "diffusion-single-file" | "stable-baselines3" | "stanza" | "tensorflowtts" | "tic-clip" | "timesfm" | "timm" | "transformers" | "transformers.js" | "unity-sentis" | "voicecraft" | "whisperkit")[];
|
|
609
|
+
export declare const ALL_DISPLAY_MODEL_LIBRARY_KEYS: ("adapter-transformers" | "allennlp" | "asteroid" | "audiocraft" | "audioseal" | "bertopic" | "big_vision" | "birefnet" | "bm25s" | "champ" | "chat_tts" | "colpali" | "depth-anything-v2" | "diffree" | "diffusers" | "doctr" | "edsnlp" | "elm" | "espnet" | "fairseq" | "fastai" | "fasttext" | "flair" | "gemma.cpp" | "gliner" | "glyph-byt5" | "grok" | "hallo" | "hunyuan-dit" | "keras" | "tf-keras" | "keras-nlp" | "k2" | "liveportrait" | "llama-cpp-python" | "mindspore" | "mamba-ssm" | "mars5-tts" | "mesh-anything" | "ml-agents" | "mlx" | "mlx-image" | "mlc-llm" | "nemo" | "open_clip" | "paddlenlp" | "peft" | "pyannote-audio" | "pythae" | "recurrentgemma" | "saelens" | "sam2" | "sample-factory" | "sentence-transformers" | "setfit" | "sklearn" | "spacy" | "span-marker" | "speechbrain" | "stable-audio-tools" | "diffusion-single-file" | "stable-baselines3" | "stanza" | "tensorflowtts" | "tic-clip" | "timesfm" | "timm" | "transformers" | "transformers.js" | "unity-sentis" | "voicecraft" | "whisperkit")[];
|
|
597
610
|
//# sourceMappingURL=model-libraries.d.ts.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"model-libraries.d.ts","sourceRoot":"","sources":["../../src/model-libraries.ts"],"names":[],"mappings":"AACA,OAAO,KAAK,EAAE,SAAS,EAAE,MAAM,cAAc,CAAC;AAC9C,OAAO,KAAK,EAAE,kBAAkB,EAAE,MAAM,6BAA6B,CAAC;AAEtE;;GAEG;AACH,MAAM,WAAW,gBAAgB;IAChC;;;;OAIG;IACH,WAAW,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,QAAQ,EAAE,MAAM,CAAC;IACjB;;OAEG;IACH,OAAO,EAAE,MAAM,CAAC;IAChB;;OAEG;IACH,OAAO,CAAC,EAAE,MAAM,CAAC;IACjB;;OAEG;IACH,QAAQ,CAAC,EAAE,CAAC,KAAK,EAAE,SAAS,KAAK,MAAM,EAAE,CAAC;IAC1C;;;;;OAKG;IACH,cAAc,CAAC,EAAE,kBAAkB,CAAC;IACpC;;;OAGG;IACH,MAAM,CAAC,EAAE,OAAO,CAAC;CACjB;AAED;;;;;;;;;;;;;GAaG;AAEH,eAAO,MAAM,2BAA2B
|
|
1
|
+
{"version":3,"file":"model-libraries.d.ts","sourceRoot":"","sources":["../../src/model-libraries.ts"],"names":[],"mappings":"AACA,OAAO,KAAK,EAAE,SAAS,EAAE,MAAM,cAAc,CAAC;AAC9C,OAAO,KAAK,EAAE,kBAAkB,EAAE,MAAM,6BAA6B,CAAC;AAEtE;;GAEG;AACH,MAAM,WAAW,gBAAgB;IAChC;;;;OAIG;IACH,WAAW,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,QAAQ,EAAE,MAAM,CAAC;IACjB;;OAEG;IACH,OAAO,EAAE,MAAM,CAAC;IAChB;;OAEG;IACH,OAAO,CAAC,EAAE,MAAM,CAAC;IACjB;;OAEG;IACH,QAAQ,CAAC,EAAE,CAAC,KAAK,EAAE,SAAS,KAAK,MAAM,EAAE,CAAC;IAC1C;;;;;OAKG;IACH,cAAc,CAAC,EAAE,kBAAkB,CAAC;IACpC;;;OAGG;IACH,MAAM,CAAC,EAAE,OAAO,CAAC;CACjB;AAED;;;;;;;;;;;;;GAaG;AAEH,eAAO,MAAM,2BAA2B;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CAwiBI,CAAC;AAE7C,MAAM,MAAM,eAAe,GAAG,MAAM,OAAO,2BAA2B,CAAC;AAEvE,eAAO,MAAM,sBAAsB,o/BAAgE,CAAC;AAEpG,eAAO,MAAM,8BAA8B,o/BAQ1B,CAAC"}
|
package/dist/src/pipelines.d.ts
CHANGED
|
@@ -387,6 +387,12 @@ export declare const PIPELINE_DATA: {
|
|
|
387
387
|
modality: "cv";
|
|
388
388
|
color: "indigo";
|
|
389
389
|
};
|
|
390
|
+
"video-text-to-text": {
|
|
391
|
+
name: string;
|
|
392
|
+
modality: "multimodal";
|
|
393
|
+
color: "blue";
|
|
394
|
+
hideInDatasets: true;
|
|
395
|
+
};
|
|
390
396
|
other: {
|
|
391
397
|
name: string;
|
|
392
398
|
modality: "other";
|
|
@@ -397,7 +403,7 @@ export declare const PIPELINE_DATA: {
|
|
|
397
403
|
};
|
|
398
404
|
export type PipelineType = keyof typeof PIPELINE_DATA;
|
|
399
405
|
export type WidgetType = PipelineType | "conversational";
|
|
400
|
-
export declare const PIPELINE_TYPES: ("other" | "text-classification" | "token-classification" | "table-question-answering" | "question-answering" | "zero-shot-classification" | "translation" | "summarization" | "feature-extraction" | "text-generation" | "text2text-generation" | "fill-mask" | "sentence-similarity" | "text-to-speech" | "text-to-audio" | "automatic-speech-recognition" | "audio-to-audio" | "audio-classification" | "voice-activity-detection" | "depth-estimation" | "image-classification" | "object-detection" | "image-segmentation" | "text-to-image" | "image-to-text" | "image-to-image" | "image-to-video" | "unconditional-image-generation" | "video-classification" | "reinforcement-learning" | "robotics" | "tabular-classification" | "tabular-regression" | "tabular-to-text" | "table-to-text" | "multiple-choice" | "text-retrieval" | "time-series-forecasting" | "text-to-video" | "image-text-to-text" | "visual-question-answering" | "document-question-answering" | "zero-shot-image-classification" | "graph-ml" | "mask-generation" | "zero-shot-object-detection" | "text-to-3d" | "image-to-3d" | "image-feature-extraction")[];
|
|
406
|
+
export declare const PIPELINE_TYPES: ("other" | "text-classification" | "token-classification" | "table-question-answering" | "question-answering" | "zero-shot-classification" | "translation" | "summarization" | "feature-extraction" | "text-generation" | "text2text-generation" | "fill-mask" | "sentence-similarity" | "text-to-speech" | "text-to-audio" | "automatic-speech-recognition" | "audio-to-audio" | "audio-classification" | "voice-activity-detection" | "depth-estimation" | "image-classification" | "object-detection" | "image-segmentation" | "text-to-image" | "image-to-text" | "image-to-image" | "image-to-video" | "unconditional-image-generation" | "video-classification" | "reinforcement-learning" | "robotics" | "tabular-classification" | "tabular-regression" | "tabular-to-text" | "table-to-text" | "multiple-choice" | "text-retrieval" | "time-series-forecasting" | "text-to-video" | "image-text-to-text" | "visual-question-answering" | "document-question-answering" | "zero-shot-image-classification" | "graph-ml" | "mask-generation" | "zero-shot-object-detection" | "text-to-3d" | "image-to-3d" | "image-feature-extraction" | "video-text-to-text")[];
|
|
401
407
|
export declare const SUBTASK_TYPES: string[];
|
|
402
|
-
export declare const PIPELINE_TYPES_SET: Set<"other" | "text-classification" | "token-classification" | "table-question-answering" | "question-answering" | "zero-shot-classification" | "translation" | "summarization" | "feature-extraction" | "text-generation" | "text2text-generation" | "fill-mask" | "sentence-similarity" | "text-to-speech" | "text-to-audio" | "automatic-speech-recognition" | "audio-to-audio" | "audio-classification" | "voice-activity-detection" | "depth-estimation" | "image-classification" | "object-detection" | "image-segmentation" | "text-to-image" | "image-to-text" | "image-to-image" | "image-to-video" | "unconditional-image-generation" | "video-classification" | "reinforcement-learning" | "robotics" | "tabular-classification" | "tabular-regression" | "tabular-to-text" | "table-to-text" | "multiple-choice" | "text-retrieval" | "time-series-forecasting" | "text-to-video" | "image-text-to-text" | "visual-question-answering" | "document-question-answering" | "zero-shot-image-classification" | "graph-ml" | "mask-generation" | "zero-shot-object-detection" | "text-to-3d" | "image-to-3d" | "image-feature-extraction">;
|
|
408
|
+
export declare const PIPELINE_TYPES_SET: Set<"other" | "text-classification" | "token-classification" | "table-question-answering" | "question-answering" | "zero-shot-classification" | "translation" | "summarization" | "feature-extraction" | "text-generation" | "text2text-generation" | "fill-mask" | "sentence-similarity" | "text-to-speech" | "text-to-audio" | "automatic-speech-recognition" | "audio-to-audio" | "audio-classification" | "voice-activity-detection" | "depth-estimation" | "image-classification" | "object-detection" | "image-segmentation" | "text-to-image" | "image-to-text" | "image-to-image" | "image-to-video" | "unconditional-image-generation" | "video-classification" | "reinforcement-learning" | "robotics" | "tabular-classification" | "tabular-regression" | "tabular-to-text" | "table-to-text" | "multiple-choice" | "text-retrieval" | "time-series-forecasting" | "text-to-video" | "image-text-to-text" | "visual-question-answering" | "document-question-answering" | "zero-shot-image-classification" | "graph-ml" | "mask-generation" | "zero-shot-object-detection" | "text-to-3d" | "image-to-3d" | "image-feature-extraction" | "video-text-to-text">;
|
|
403
409
|
//# sourceMappingURL=pipelines.d.ts.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"pipelines.d.ts","sourceRoot":"","sources":["../../src/pipelines.ts"],"names":[],"mappings":"AAAA,eAAO,MAAM,UAAU,yEAA0E,CAAC;AAElG,MAAM,MAAM,QAAQ,GAAG,CAAC,OAAO,UAAU,CAAC,CAAC,MAAM,CAAC,CAAC;AAEnD,eAAO,MAAM,eAAe;;;;;;;;CAQQ,CAAC;AAErC;;;;;;GAMG;AACH,MAAM,WAAW,OAAO;IACvB;;OAEG;IACH,IAAI,EAAE,MAAM,CAAC;IACb;;OAEG;IACH,IAAI,EAAE,MAAM,CAAC;CACb;AAED;;;;;GAKG;AACH,MAAM,WAAW,YAAY;IAC5B;;OAEG;IACH,IAAI,EAAE,MAAM,CAAC;IACb,QAAQ,CAAC,EAAE,OAAO,EAAE,CAAC;IACrB,QAAQ,EAAE,QAAQ,CAAC;IACnB;;OAEG;IACH,KAAK,EAAE,MAAM,GAAG,OAAO,GAAG,QAAQ,GAAG,QAAQ,GAAG,KAAK,GAAG,QAAQ,CAAC;IACjE;;OAEG;IACH,YAAY,CAAC,EAAE,OAAO,CAAC;IACvB;;OAEG;IACH,cAAc,CAAC,EAAE,OAAO,CAAC;CACzB;AAcD,eAAO,MAAM,aAAa
|
|
1
|
+
{"version":3,"file":"pipelines.d.ts","sourceRoot":"","sources":["../../src/pipelines.ts"],"names":[],"mappings":"AAAA,eAAO,MAAM,UAAU,yEAA0E,CAAC;AAElG,MAAM,MAAM,QAAQ,GAAG,CAAC,OAAO,UAAU,CAAC,CAAC,MAAM,CAAC,CAAC;AAEnD,eAAO,MAAM,eAAe;;;;;;;;CAQQ,CAAC;AAErC;;;;;;GAMG;AACH,MAAM,WAAW,OAAO;IACvB;;OAEG;IACH,IAAI,EAAE,MAAM,CAAC;IACb;;OAEG;IACH,IAAI,EAAE,MAAM,CAAC;CACb;AAED;;;;;GAKG;AACH,MAAM,WAAW,YAAY;IAC5B;;OAEG;IACH,IAAI,EAAE,MAAM,CAAC;IACb,QAAQ,CAAC,EAAE,OAAO,EAAE,CAAC;IACrB,QAAQ,EAAE,QAAQ,CAAC;IACnB;;OAEG;IACH,KAAK,EAAE,MAAM,GAAG,OAAO,GAAG,QAAQ,GAAG,QAAQ,GAAG,KAAK,GAAG,QAAQ,CAAC;IACjE;;OAEG;IACH,YAAY,CAAC,EAAE,OAAO,CAAC;IACvB;;OAEG;IACH,cAAc,CAAC,EAAE,OAAO,CAAC;CACzB;AAcD,eAAO,MAAM,aAAa;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CAolBc,CAAC;AAEzC,MAAM,MAAM,YAAY,GAAG,MAAM,OAAO,aAAa,CAAC;AAEtD,MAAM,MAAM,UAAU,GAAG,YAAY,GAAG,gBAAgB,CAAC;AAEzD,eAAO,MAAM,cAAc,0mCAA+C,CAAC;AAE3E,eAAO,MAAM,aAAa,UAEN,CAAC;AAErB,eAAO,MAAM,kBAAkB,2mCAA0B,CAAC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/feature-extraction/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,IAAI,CAAC;AAEzC,QAAA,MAAM,QAAQ,EAAE,
|
|
1
|
+
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/feature-extraction/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,IAAI,CAAC;AAEzC,QAAA,MAAM,QAAQ,EAAE,cAoDf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/image-feature-extraction/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,IAAI,CAAC;AAEzC,QAAA,MAAM,QAAQ,EAAE,
|
|
1
|
+
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/image-feature-extraction/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,IAAI,CAAC;AAEzC,QAAA,MAAM,QAAQ,EAAE,cAsDf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../../../src/tasks/index.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,YAAY,EAAE,MAAM,cAAc,CAAC;AA0CjD,mBAAmB,kCAAkC,CAAC;AACtD,mBAAmB,0CAA0C,CAAC;AAC9D,YAAY,EACX,mBAAmB,EACnB,0BAA0B,EAC1B,oBAAoB,EACpB,4BAA4B,EAC5B,2BAA2B,EAC3B,0BAA0B,EAC1B,gCAAgC,EAChC,+BAA+B,GAC/B,MAAM,6BAA6B,CAAC;AACrC,mBAAmB,yCAAyC,CAAC;AAC7D,mBAAmB,gCAAgC,CAAC;AACpD,mBAAmB,uBAAuB,CAAC;AAC3C,YAAY,EACX,wBAAwB,EACxB,yBAAyB,EACzB,gCAAgC,EAChC,6BAA6B,GAC7B,MAAM,kCAAkC,CAAC;AAC1C,mBAAmB,4BAA4B,CAAC;AAChD,YAAY,EAAE,gBAAgB,EAAE,iBAAiB,EAAE,qBAAqB,EAAE,MAAM,2BAA2B,CAAC;AAC5G,mBAAmB,gCAAgC,CAAC;AACpD,mBAAmB,8BAA8B,CAAC;AAClD,mBAAmB,8BAA8B,CAAC;AAClD,mBAAmB,gCAAgC,CAAC;AACpD,mBAAmB,iCAAiC,CAAC;AACrD,mBAAmB,2BAA2B,CAAC;AAC/C,mBAAmB,sCAAsC,CAAC;AAC1D,YAAY,EAAE,gBAAgB,EAAE,iBAAiB,EAAE,qBAAqB,EAAE,MAAM,2BAA2B,CAAC;AAC5G,YAAY,EAAE,qBAAqB,EAAE,iBAAiB,EAAE,kBAAkB,EAAE,MAAM,4BAA4B,CAAC;AAC/G,mBAAmB,kCAAkC,CAAC;AACtD,YAAY,EACX,6BAA6B,EAC7B,qCAAqC,EACrC,gBAAgB,EAChB,iBAAiB,GACjB,MAAM,yBAAyB,CAAC;AACjC,YAAY,EACX,6BAA6B,EAC7B,uBAAuB,EACvB,wBAAwB,EACxB,+BAA+B,EAC/B,4BAA4B,GAC5B,MAAM,iCAAiC,CAAC;AACzC,YAAY,EACX,gCAAgC,EAChC,gCAAgC,EAChC,mBAAmB,EACnB,oBAAoB,EACpB,2BAA2B,EAC3B,qCAAqC,EACrC,kCAAkC,EAClC,yBAAyB,EACzB,uCAAuC,EACvC,0BAA0B,GAC1B,MAAM,6BAA6B,CAAC;AACrC,mBAAmB,kCAAkC,CAAC;AACtD,mBAAmB,uCAAuC,CAAC;AAC3D,mBAAmB,sCAAsC,CAAC;AAC1D,mBAAmB,4CAA4C,CAAC;AAChE,YAAY,EACX,WAAW,EACX,4BAA4B,EAC5B,gCAAgC,EAChC,6BAA6B,EAC7B,oCAAoC,GACpC,MAAM,wCAAwC,CAAC;AAEhD,OAAO,KAAK,EAAE,eAAe,EAAE,MAAM,oBAAoB,CAAC;AAE1D;;GAEG;AACH,eAAO,MAAM,qBAAqB,EAAE,MAAM,CAAC,YAAY,EAAE,eAAe,EAAE,
|
|
1
|
+
{"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../../../src/tasks/index.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,YAAY,EAAE,MAAM,cAAc,CAAC;AA0CjD,mBAAmB,kCAAkC,CAAC;AACtD,mBAAmB,0CAA0C,CAAC;AAC9D,YAAY,EACX,mBAAmB,EACnB,0BAA0B,EAC1B,oBAAoB,EACpB,4BAA4B,EAC5B,2BAA2B,EAC3B,0BAA0B,EAC1B,gCAAgC,EAChC,+BAA+B,GAC/B,MAAM,6BAA6B,CAAC;AACrC,mBAAmB,yCAAyC,CAAC;AAC7D,mBAAmB,gCAAgC,CAAC;AACpD,mBAAmB,uBAAuB,CAAC;AAC3C,YAAY,EACX,wBAAwB,EACxB,yBAAyB,EACzB,gCAAgC,EAChC,6BAA6B,GAC7B,MAAM,kCAAkC,CAAC;AAC1C,mBAAmB,4BAA4B,CAAC;AAChD,YAAY,EAAE,gBAAgB,EAAE,iBAAiB,EAAE,qBAAqB,EAAE,MAAM,2BAA2B,CAAC;AAC5G,mBAAmB,gCAAgC,CAAC;AACpD,mBAAmB,8BAA8B,CAAC;AAClD,mBAAmB,8BAA8B,CAAC;AAClD,mBAAmB,gCAAgC,CAAC;AACpD,mBAAmB,iCAAiC,CAAC;AACrD,mBAAmB,2BAA2B,CAAC;AAC/C,mBAAmB,sCAAsC,CAAC;AAC1D,YAAY,EAAE,gBAAgB,EAAE,iBAAiB,EAAE,qBAAqB,EAAE,MAAM,2BAA2B,CAAC;AAC5G,YAAY,EAAE,qBAAqB,EAAE,iBAAiB,EAAE,kBAAkB,EAAE,MAAM,4BAA4B,CAAC;AAC/G,mBAAmB,kCAAkC,CAAC;AACtD,YAAY,EACX,6BAA6B,EAC7B,qCAAqC,EACrC,gBAAgB,EAChB,iBAAiB,GACjB,MAAM,yBAAyB,CAAC;AACjC,YAAY,EACX,6BAA6B,EAC7B,uBAAuB,EACvB,wBAAwB,EACxB,+BAA+B,EAC/B,4BAA4B,GAC5B,MAAM,iCAAiC,CAAC;AACzC,YAAY,EACX,gCAAgC,EAChC,gCAAgC,EAChC,mBAAmB,EACnB,oBAAoB,EACpB,2BAA2B,EAC3B,qCAAqC,EACrC,kCAAkC,EAClC,yBAAyB,EACzB,uCAAuC,EACvC,0BAA0B,GAC1B,MAAM,6BAA6B,CAAC;AACrC,mBAAmB,kCAAkC,CAAC;AACtD,mBAAmB,uCAAuC,CAAC;AAC3D,mBAAmB,sCAAsC,CAAC;AAC1D,mBAAmB,4CAA4C,CAAC;AAChE,YAAY,EACX,WAAW,EACX,4BAA4B,EAC5B,gCAAgC,EAChC,6BAA6B,EAC7B,oCAAoC,GACpC,MAAM,wCAAwC,CAAC;AAEhD,OAAO,KAAK,EAAE,eAAe,EAAE,MAAM,oBAAoB,CAAC;AAE1D;;GAEG;AACH,eAAO,MAAM,qBAAqB,EAAE,MAAM,CAAC,YAAY,EAAE,eAAe,EAAE,CA2DzE,CAAC;AAoBF,eAAO,MAAM,UAAU,EAAE,MAAM,CAAC,YAAY,EAAE,QAAQ,GAAG,SAAS,CAmDxD,CAAC;AAEX,MAAM,WAAW,WAAW;IAC3B,WAAW,EAAE,MAAM,CAAC;IACpB,EAAE,EAAE,MAAM,CAAC;CACX;AAED,MAAM,MAAM,aAAa,GACtB;IACA,QAAQ,EAAE,MAAM,CAAC;IACjB,IAAI,EAAE,OAAO,CAAC;CACb,GACD;IACA,IAAI,EAAE,KAAK,CAAC;QACX,KAAK,EAAE,MAAM,CAAC;QACd,KAAK,EAAE,MAAM,CAAC;KACd,CAAC,CAAC;IACH,IAAI,EAAE,OAAO,CAAC;CACb,GACD;IACA,QAAQ,EAAE,MAAM,CAAC;IACjB,IAAI,EAAE,KAAK,CAAC;CACX,GACD;IACA,KAAK,EAAE,MAAM,EAAE,EAAE,CAAC;IAClB,IAAI,EAAE,SAAS,CAAC;CACf,GACD;IACA,OAAO,EAAE,MAAM,CAAC;IAChB,KAAK,EAAE,MAAM,CAAC;IACd,IAAI,EAAE,MAAM,CAAC;CACZ,GACD;IACA,IAAI,EAAE,MAAM,CAAC;IACb,MAAM,EAAE,KAAK,CAAC;QACb,GAAG,EAAE,MAAM,CAAC;QACZ,KAAK,EAAE,MAAM,CAAC;QACd,IAAI,EAAE,MAAM,CAAC;KACb,CAAC,CAAC;IACH,IAAI,EAAE,kBAAkB,CAAC;CACxB,CAAC;AAEL,MAAM,WAAW,QAAQ;IACxB,MAAM,EAAE,aAAa,EAAE,CAAC;IACxB,OAAO,EAAE,aAAa,EAAE,CAAC;CACzB;AAED,MAAM,WAAW,QAAQ;IACxB,QAAQ,EAAE,WAAW,EAAE,CAAC;IACxB,IAAI,EAAE,QAAQ,CAAC;IACf,EAAE,EAAE,YAAY,CAAC;IACjB,WAAW,CAAC,EAAE,YAAY,CAAC;IAC3B,aAAa,CAAC,EAAE,OAAO,CAAC;IACxB,KAAK,EAAE,MAAM,CAAC;IACd,SAAS,EAAE,eAAe,EAAE,CAAC;IAC7B,OAAO,EAAE,WAAW,EAAE,CAAC;IACvB,MAAM,EAAE,WAAW,EAAE,CAAC;IACtB,MAAM,EAAE,WAAW,EAAE,CAAC;IACtB,OAAO,EAAE,MAAM,CAAC;IAChB,YAAY,EAAE,MAAM,EAAE,CAAC;IACvB,SAAS,CAAC,EAAE,MAAM,CAAC;CACnB;AAED,MAAM,MAAM,cAAc,GAAG,IAAI,CAAC,QAAQ,EAAE,IAAI,GAAG,OAAO,GAAG,WAAW,CAAC,CAAC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/text-generation/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,IAAI,CAAC;AAEzC,QAAA,MAAM,QAAQ,EAAE,
|
|
1
|
+
{"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/text-generation/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,IAAI,CAAC;AAEzC,QAAA,MAAM,QAAQ,EAAE,cA6Gf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
|
package/package.json
CHANGED
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@huggingface/tasks",
|
|
3
3
|
"packageManager": "pnpm@8.10.5",
|
|
4
|
-
"version": "0.11.
|
|
4
|
+
"version": "0.11.8",
|
|
5
5
|
"description": "List of ML tasks for huggingface.co/tasks",
|
|
6
6
|
"repository": "https://github.com/huggingface/huggingface.js.git",
|
|
7
7
|
"publishConfig": {
|
|
@@ -42,6 +42,9 @@
|
|
|
42
42
|
"format": "prettier --write .",
|
|
43
43
|
"format:check": "prettier --check .",
|
|
44
44
|
"build": "tsup src/index.ts --format cjs,esm --clean && tsc --emitDeclarationOnly --declaration",
|
|
45
|
+
"watch:export": "tsup src/index.ts --format cjs,esm --watch",
|
|
46
|
+
"watch:types": "tsc --emitDeclarationOnly --declaration --watch",
|
|
47
|
+
"watch": "npm-run-all --parallel watch:export watch:types",
|
|
45
48
|
"check": "tsc",
|
|
46
49
|
"inference-codegen": "tsx scripts/inference-codegen.ts && prettier --write src/tasks/*/inference.ts",
|
|
47
50
|
"inference-tgi-import": "tsx scripts/inference-tgi-import.ts && prettier --write src/tasks/text-generation/spec/*.json && prettier --write src/tasks/chat-completion/spec/*.json",
|
package/src/local-apps.ts
CHANGED
|
@@ -201,7 +201,6 @@ export const LOCAL_APPS = {
|
|
|
201
201
|
docsUrl: "https://diffusionbee.com",
|
|
202
202
|
mainTask: "text-to-image",
|
|
203
203
|
macOSOnly: true,
|
|
204
|
-
comingSoon: true,
|
|
205
204
|
displayOnModelPage: (model) => model.library_name === "diffusers" && model.pipeline_tag === "text-to-image",
|
|
206
205
|
deeplink: (model) => new URL(`diffusionbee://open_from_hf?model=${model.id}`),
|
|
207
206
|
},
|
|
@@ -261,6 +261,24 @@ backbone = keras_nlp.models.Backbone.from_preset("hf://${model.id}")
|
|
|
261
261
|
`,
|
|
262
262
|
];
|
|
263
263
|
|
|
264
|
+
export const llama_cpp_python = (model: ModelData): string[] => [
|
|
265
|
+
`from llama_cpp import Llama
|
|
266
|
+
|
|
267
|
+
llm = Llama.from_pretrained(
|
|
268
|
+
repo_id="${model.id}",
|
|
269
|
+
filename="{{GGUF_FILE}}",
|
|
270
|
+
)
|
|
271
|
+
|
|
272
|
+
llm.create_chat_completion(
|
|
273
|
+
messages = [
|
|
274
|
+
{
|
|
275
|
+
"role": "user",
|
|
276
|
+
"content": "What is the capital of France?"
|
|
277
|
+
}
|
|
278
|
+
]
|
|
279
|
+
)`,
|
|
280
|
+
];
|
|
281
|
+
|
|
264
282
|
export const tf_keras = (model: ModelData): string[] => [
|
|
265
283
|
`# Note: 'keras<3.x' or 'tf_keras' must be installed (legacy)
|
|
266
284
|
# See https://github.com/keras-team/tf-keras for more details.
|
package/src/model-libraries.ts
CHANGED
|
@@ -314,6 +314,12 @@ export const MODEL_LIBRARIES_UI_ELEMENTS = {
|
|
|
314
314
|
filter: false,
|
|
315
315
|
countDownloads: `path:"liveportrait/landmark.onnx"`,
|
|
316
316
|
},
|
|
317
|
+
"llama-cpp-python": {
|
|
318
|
+
prettyLabel: "llama-cpp-python",
|
|
319
|
+
repoName: "llama-cpp-python",
|
|
320
|
+
repoUrl: "https://github.com/abetlen/llama-cpp-python",
|
|
321
|
+
snippets: snippets.llama_cpp_python,
|
|
322
|
+
},
|
|
317
323
|
mindspore: {
|
|
318
324
|
prettyLabel: "MindSpore",
|
|
319
325
|
repoName: "mindspore",
|
|
@@ -436,6 +442,13 @@ export const MODEL_LIBRARIES_UI_ELEMENTS = {
|
|
|
436
442
|
snippets: snippets.saelens,
|
|
437
443
|
filter: false,
|
|
438
444
|
},
|
|
445
|
+
sam2: {
|
|
446
|
+
prettyLabel: "sam2",
|
|
447
|
+
repoName: "sam2",
|
|
448
|
+
repoUrl: "https://github.com/facebookresearch/segment-anything-2",
|
|
449
|
+
filter: false,
|
|
450
|
+
countDownloads: `path_extension:"pt"`,
|
|
451
|
+
},
|
|
439
452
|
"sample-factory": {
|
|
440
453
|
prettyLabel: "sample-factory",
|
|
441
454
|
repoName: "sample-factory",
|
package/src/pipelines.ts
CHANGED
|
@@ -652,6 +652,12 @@ export const PIPELINE_DATA = {
|
|
|
652
652
|
modality: "cv",
|
|
653
653
|
color: "indigo",
|
|
654
654
|
},
|
|
655
|
+
"video-text-to-text": {
|
|
656
|
+
name: "Video-Text-to-Text",
|
|
657
|
+
modality: "multimodal",
|
|
658
|
+
color: "blue",
|
|
659
|
+
hideInDatasets: true,
|
|
660
|
+
},
|
|
655
661
|
other: {
|
|
656
662
|
name: "Other",
|
|
657
663
|
modality: "other",
|
|
@@ -42,9 +42,13 @@ const taskData: TaskDataCustom = {
|
|
|
42
42
|
],
|
|
43
43
|
spaces: [
|
|
44
44
|
{
|
|
45
|
-
description: "A leaderboard to rank
|
|
45
|
+
description: "A leaderboard to rank text feature extraction models based on a benchmark.",
|
|
46
46
|
id: "mteb/leaderboard",
|
|
47
47
|
},
|
|
48
|
+
{
|
|
49
|
+
description: "A leaderboard to rank best feature extraction models based on human feedback.",
|
|
50
|
+
id: "mteb/arena",
|
|
51
|
+
},
|
|
48
52
|
],
|
|
49
53
|
summary: "Feature extraction is the task of extracting features learnt in a model.",
|
|
50
54
|
widgetModels: ["facebook/bart-base"],
|
|
@@ -36,16 +36,20 @@ const taskData: TaskDataCustom = {
|
|
|
36
36
|
},
|
|
37
37
|
{
|
|
38
38
|
description: "A strong image feature extraction model.",
|
|
39
|
-
id: "
|
|
39
|
+
id: "nvidia/MambaVision-T-1K",
|
|
40
40
|
},
|
|
41
41
|
{
|
|
42
|
-
description: "A robust image feature extraction
|
|
42
|
+
description: "A robust image feature extraction model.",
|
|
43
43
|
id: "facebook/dino-vitb16",
|
|
44
44
|
},
|
|
45
45
|
{
|
|
46
|
-
description: "Strong image
|
|
46
|
+
description: "Strong image feature extraction model made for information retrieval from documents.",
|
|
47
47
|
id: "vidore/colpali",
|
|
48
48
|
},
|
|
49
|
+
{
|
|
50
|
+
description: "Strong image feature extraction model that can be used on images and documents.",
|
|
51
|
+
id: "OpenGVLab/InternViT-6B-448px-V1-2",
|
|
52
|
+
},
|
|
49
53
|
],
|
|
50
54
|
spaces: [],
|
|
51
55
|
summary: "Image feature extraction is the task of extracting features learnt in a computer vision model.",
|
|
@@ -48,16 +48,16 @@ const taskData: TaskDataCustom = {
|
|
|
48
48
|
id: "facebook/detr-resnet-50-panoptic",
|
|
49
49
|
},
|
|
50
50
|
{
|
|
51
|
-
description: "
|
|
52
|
-
id: "
|
|
51
|
+
description: "Background removal model.",
|
|
52
|
+
id: "briaai/RMBG-1.4",
|
|
53
53
|
},
|
|
54
54
|
{
|
|
55
55
|
description: "Semantic segmentation model trained on ADE20k benchmark dataset with 512x512 resolution.",
|
|
56
56
|
id: "nvidia/segformer-b0-finetuned-ade-512-512",
|
|
57
57
|
},
|
|
58
58
|
{
|
|
59
|
-
description: "
|
|
60
|
-
id: "
|
|
59
|
+
description: "A multipurpose image segmentation model for high resolution images.",
|
|
60
|
+
id: "ZhengPeng7/BiRefNet",
|
|
61
61
|
},
|
|
62
62
|
{
|
|
63
63
|
description: "Panoptic segmentation model trained COCO (common objects) dataset.",
|
|
@@ -72,3 +72,5 @@ print(processor.decode(output[0], skip_special_tokens=True))
|
|
|
72
72
|
- [Vision Language Models Explained](https://huggingface.co/blog/vlms)
|
|
73
73
|
- [Open-source Multimodality and How to Achieve it using Hugging Face](https://www.youtube.com/watch?v=IoGaGfU1CIg&t=601s)
|
|
74
74
|
- [Introducing Idefics2: A Powerful 8B Vision-Language Model for the community](https://huggingface.co/blog/idefics2)
|
|
75
|
+
- [Image-text-to-text task guide](https://huggingface.co/tasks/image-text-to-text)
|
|
76
|
+
- [Preference Optimization for Vision Language Models with TRL](https://huggingface.co/blog/dpo_vlm)
|
|
@@ -88,7 +88,7 @@ const taskData: TaskDataCustom = {
|
|
|
88
88
|
summary:
|
|
89
89
|
"Image-text-to-text models take in an image and text prompt and output text. These models are also called vision-language models, or VLMs. The difference from image-to-text models is that these models take an additional text input, not restricting the model to certain use cases like image captioning, and may also be trained to accept a conversation as input.",
|
|
90
90
|
widgetModels: ["microsoft/kosmos-2-patch14-224"],
|
|
91
|
-
youtubeId: "",
|
|
91
|
+
youtubeId: "IoGaGfU1CIg",
|
|
92
92
|
};
|
|
93
93
|
|
|
94
94
|
export default taskData;
|
|
@@ -45,8 +45,8 @@ const taskData: TaskDataCustom = {
|
|
|
45
45
|
],
|
|
46
46
|
models: [
|
|
47
47
|
{
|
|
48
|
-
description: "
|
|
49
|
-
id: "
|
|
48
|
+
description: "An image-to-image model to improve image resolution.",
|
|
49
|
+
id: "fal/AuraSR-v2",
|
|
50
50
|
},
|
|
51
51
|
{
|
|
52
52
|
description: "A model that increases the resolution of an image.",
|
package/src/tasks/index.ts
CHANGED
|
@@ -166,6 +166,7 @@ export const TASKS_MODEL_LIBRARIES: Record<PipelineType, ModelLibraryKey[]> = {
|
|
|
166
166
|
],
|
|
167
167
|
translation: ["transformers", "transformers.js"],
|
|
168
168
|
"unconditional-image-generation": ["diffusers"],
|
|
169
|
+
"video-text-to-text": ["transformers"],
|
|
169
170
|
"visual-question-answering": ["transformers", "transformers.js"],
|
|
170
171
|
"voice-activity-detection": [],
|
|
171
172
|
"zero-shot-classification": ["transformers", "transformers.js"],
|
|
@@ -236,6 +237,7 @@ export const TASKS_DATA: Record<PipelineType, TaskData | undefined> = {
|
|
|
236
237
|
"token-classification": getData("token-classification", tokenClassification),
|
|
237
238
|
translation: getData("translation", translation),
|
|
238
239
|
"unconditional-image-generation": getData("unconditional-image-generation", unconditionalImageGeneration),
|
|
240
|
+
"video-text-to-text": getData("video-text-to-text", placeholder),
|
|
239
241
|
"visual-question-answering": getData("visual-question-answering", visualQuestionAnswering),
|
|
240
242
|
"voice-activity-detection": undefined,
|
|
241
243
|
"zero-shot-classification": getData("zero-shot-classification", zeroShotClassification),
|
|
@@ -24,14 +24,14 @@ const taskData: TaskDataCustom = {
|
|
|
24
24
|
},
|
|
25
25
|
{
|
|
26
26
|
description: "Very strong mask generation model.",
|
|
27
|
-
id: "facebook/
|
|
27
|
+
id: "facebook/sam2-hiera-large",
|
|
28
28
|
},
|
|
29
29
|
],
|
|
30
30
|
spaces: [
|
|
31
31
|
{
|
|
32
32
|
description:
|
|
33
|
-
"An application that combines a mask generation model with
|
|
34
|
-
id: "
|
|
33
|
+
"An application that combines a mask generation model with a zero-shot object detection model for text-guided image segmentation.",
|
|
34
|
+
id: "merve/OWLSAM2",
|
|
35
35
|
},
|
|
36
36
|
{
|
|
37
37
|
description: "An application that compares the performance of a large and a small mask generation model.",
|
|
@@ -39,7 +39,7 @@ const taskData: TaskDataCustom = {
|
|
|
39
39
|
},
|
|
40
40
|
{
|
|
41
41
|
description: "An application based on an improved mask generation model.",
|
|
42
|
-
id: "
|
|
42
|
+
id: "SkalskiP/segment-anything-model-2",
|
|
43
43
|
},
|
|
44
44
|
{
|
|
45
45
|
description: "An application to remove objects from videos using mask generation models.",
|
|
@@ -19,6 +19,10 @@ const taskData: TaskDataCustom = {
|
|
|
19
19
|
description: "An instruction dataset with preference ratings on responses.",
|
|
20
20
|
id: "openbmb/UltraFeedback",
|
|
21
21
|
},
|
|
22
|
+
{
|
|
23
|
+
description: "A large synthetic dataset for alignment of text generation models.",
|
|
24
|
+
id: "argilla/magpie-ultra-v0.1",
|
|
25
|
+
},
|
|
22
26
|
],
|
|
23
27
|
demo: {
|
|
24
28
|
inputs: [
|
|
@@ -51,32 +55,32 @@ const taskData: TaskDataCustom = {
|
|
|
51
55
|
],
|
|
52
56
|
models: [
|
|
53
57
|
{
|
|
54
|
-
description: "A
|
|
55
|
-
id: "
|
|
58
|
+
description: "A text-generation model trained to follow instructions.",
|
|
59
|
+
id: "google/gemma-2-2b-it",
|
|
56
60
|
},
|
|
57
61
|
{
|
|
58
|
-
description: "A
|
|
62
|
+
description: "A code generation model that can generate code in 80+ languages.",
|
|
59
63
|
id: "bigcode/starcoder",
|
|
60
64
|
},
|
|
61
65
|
{
|
|
62
|
-
description: "
|
|
63
|
-
id: "
|
|
66
|
+
description: "Very powerful text generation model trained to follow instructions.",
|
|
67
|
+
id: "meta-llama/Meta-Llama-3.1-8B-Instruct",
|
|
64
68
|
},
|
|
65
69
|
{
|
|
66
70
|
description: "Small yet powerful text generation model.",
|
|
67
|
-
id: "microsoft/
|
|
71
|
+
id: "microsoft/Phi-3-mini-4k-instruct",
|
|
68
72
|
},
|
|
69
73
|
{
|
|
70
|
-
description: "A very powerful model that can
|
|
71
|
-
id: "
|
|
74
|
+
description: "A very powerful model that can solve mathematical problems.",
|
|
75
|
+
id: "AI-MO/NuminaMath-7B-TIR",
|
|
72
76
|
},
|
|
73
77
|
{
|
|
74
|
-
description: "
|
|
75
|
-
id: "HuggingFaceH4/
|
|
78
|
+
description: "Strong coding assistant model.",
|
|
79
|
+
id: "HuggingFaceH4/starchat2-15b-v0.1",
|
|
76
80
|
},
|
|
77
81
|
{
|
|
78
82
|
description: "Very strong open-source large language model.",
|
|
79
|
-
id: "
|
|
83
|
+
id: "mistralai/Mistral-Nemo-Instruct-2407",
|
|
80
84
|
},
|
|
81
85
|
],
|
|
82
86
|
spaces: [
|
|
@@ -104,7 +108,7 @@ const taskData: TaskDataCustom = {
|
|
|
104
108
|
summary:
|
|
105
109
|
"Generating text is the task of generating new text given another text. These models can, for example, fill in incomplete text or paraphrase.",
|
|
106
110
|
widgetModels: ["HuggingFaceH4/zephyr-7b-beta"],
|
|
107
|
-
youtubeId: "
|
|
111
|
+
youtubeId: "e9gNEAlsOvU",
|
|
108
112
|
};
|
|
109
113
|
|
|
110
114
|
export default taskData;
|
|
@@ -46,15 +46,15 @@ const taskData: TaskDataCustom = {
|
|
|
46
46
|
models: [
|
|
47
47
|
{
|
|
48
48
|
description: "One of the most powerful image generation models that can generate realistic outputs.",
|
|
49
|
-
id: "
|
|
49
|
+
id: "black-forest-labs/FLUX.1-dev",
|
|
50
50
|
},
|
|
51
51
|
{
|
|
52
52
|
description: "A powerful yet fast image generation model.",
|
|
53
53
|
id: "latent-consistency/lcm-lora-sdxl",
|
|
54
54
|
},
|
|
55
55
|
{
|
|
56
|
-
description: "
|
|
57
|
-
id: "
|
|
56
|
+
description: "Text-to-image model for photorealistic generation.",
|
|
57
|
+
id: "Kwai-Kolors/Kolors",
|
|
58
58
|
},
|
|
59
59
|
{
|
|
60
60
|
description: "A powerful text-to-image model.",
|