@huggingface/tasks 0.11.7 → 0.11.8

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.cjs CHANGED
@@ -1388,6 +1388,12 @@ var PIPELINE_DATA = {
1388
1388
  modality: "cv",
1389
1389
  color: "indigo"
1390
1390
  },
1391
+ "video-text-to-text": {
1392
+ name: "Video-Text-to-Text",
1393
+ modality: "multimodal",
1394
+ color: "blue",
1395
+ hideInDatasets: true
1396
+ },
1391
1397
  other: {
1392
1398
  name: "Other",
1393
1399
  modality: "other",
@@ -1731,8 +1737,12 @@ var taskData5 = {
1731
1737
  ],
1732
1738
  spaces: [
1733
1739
  {
1734
- description: "A leaderboard to rank best feature extraction models..",
1740
+ description: "A leaderboard to rank text feature extraction models based on a benchmark.",
1735
1741
  id: "mteb/leaderboard"
1742
+ },
1743
+ {
1744
+ description: "A leaderboard to rank best feature extraction models based on human feedback.",
1745
+ id: "mteb/arena"
1736
1746
  }
1737
1747
  ],
1738
1748
  summary: "Feature extraction is the task of extracting features learnt in a model.",
@@ -1937,15 +1947,19 @@ var taskData8 = {
1937
1947
  },
1938
1948
  {
1939
1949
  description: "A strong image feature extraction model.",
1940
- id: "google/vit-base-patch16-224-in21k"
1950
+ id: "nvidia/MambaVision-T-1K"
1941
1951
  },
1942
1952
  {
1943
- description: "A robust image feature extraction models.",
1953
+ description: "A robust image feature extraction model.",
1944
1954
  id: "facebook/dino-vitb16"
1945
1955
  },
1946
1956
  {
1947
- description: "Strong image-text-to-text model made for information retrieval from documents.",
1957
+ description: "Strong image feature extraction model made for information retrieval from documents.",
1948
1958
  id: "vidore/colpali"
1959
+ },
1960
+ {
1961
+ description: "Strong image feature extraction model that can be used on images and documents.",
1962
+ id: "OpenGVLab/InternViT-6B-448px-V1-2"
1949
1963
  }
1950
1964
  ],
1951
1965
  spaces: [],
@@ -1997,8 +2011,8 @@ var taskData9 = {
1997
2011
  ],
1998
2012
  models: [
1999
2013
  {
2000
- description: "A model that enhances images captured in low light conditions.",
2001
- id: "keras-io/low-light-image-enhancement"
2014
+ description: "An image-to-image model to improve image resolution.",
2015
+ id: "fal/AuraSR-v2"
2002
2016
  },
2003
2017
  {
2004
2018
  description: "A model that increases the resolution of an image.",
@@ -2216,7 +2230,7 @@ var taskData11 = {
2216
2230
  ],
2217
2231
  summary: "Image-text-to-text models take in an image and text prompt and output text. These models are also called vision-language models, or VLMs. The difference from image-to-text models is that these models take an additional text input, not restricting the model to certain use cases like image captioning, and may also be trained to accept a conversation as input.",
2218
2232
  widgetModels: ["microsoft/kosmos-2-patch14-224"],
2219
- youtubeId: ""
2233
+ youtubeId: "IoGaGfU1CIg"
2220
2234
  };
2221
2235
  var data_default11 = taskData11;
2222
2236
 
@@ -2267,16 +2281,16 @@ var taskData12 = {
2267
2281
  id: "facebook/detr-resnet-50-panoptic"
2268
2282
  },
2269
2283
  {
2270
- description: "Semantic segmentation model trained on ADE20k benchmark dataset.",
2271
- id: "microsoft/beit-large-finetuned-ade-640-640"
2284
+ description: "Background removal model.",
2285
+ id: "briaai/RMBG-1.4"
2272
2286
  },
2273
2287
  {
2274
2288
  description: "Semantic segmentation model trained on ADE20k benchmark dataset with 512x512 resolution.",
2275
2289
  id: "nvidia/segformer-b0-finetuned-ade-512-512"
2276
2290
  },
2277
2291
  {
2278
- description: "Semantic segmentation model trained Cityscapes dataset.",
2279
- id: "facebook/mask2former-swin-large-cityscapes-semantic"
2292
+ description: "A multipurpose image segmentation model for high resolution images.",
2293
+ id: "ZhengPeng7/BiRefNet"
2280
2294
  },
2281
2295
  {
2282
2296
  description: "Panoptic segmentation model trained COCO (common objects) dataset.",
@@ -2340,13 +2354,13 @@ var taskData13 = {
2340
2354
  },
2341
2355
  {
2342
2356
  description: "Very strong mask generation model.",
2343
- id: "facebook/sam-vit-huge"
2357
+ id: "facebook/sam2-hiera-large"
2344
2358
  }
2345
2359
  ],
2346
2360
  spaces: [
2347
2361
  {
2348
- description: "An application that combines a mask generation model with an image embedding model for open-vocabulary image segmentation.",
2349
- id: "SkalskiP/SAM_and_MetaCLIP"
2362
+ description: "An application that combines a mask generation model with a zero-shot object detection model for text-guided image segmentation.",
2363
+ id: "merve/OWLSAM2"
2350
2364
  },
2351
2365
  {
2352
2366
  description: "An application that compares the performance of a large and a small mask generation model.",
@@ -2354,7 +2368,7 @@ var taskData13 = {
2354
2368
  },
2355
2369
  {
2356
2370
  description: "An application based on an improved mask generation model.",
2357
- id: "linfanluntan/Grounded-SAM"
2371
+ id: "SkalskiP/segment-anything-model-2"
2358
2372
  },
2359
2373
  {
2360
2374
  description: "An application to remove objects from videos using mask generation models.",
@@ -3050,15 +3064,15 @@ var taskData24 = {
3050
3064
  models: [
3051
3065
  {
3052
3066
  description: "One of the most powerful image generation models that can generate realistic outputs.",
3053
- id: "stabilityai/stable-diffusion-xl-base-1.0"
3067
+ id: "black-forest-labs/FLUX.1-dev"
3054
3068
  },
3055
3069
  {
3056
3070
  description: "A powerful yet fast image generation model.",
3057
3071
  id: "latent-consistency/lcm-lora-sdxl"
3058
3072
  },
3059
3073
  {
3060
- description: "A very fast text-to-image model.",
3061
- id: "ByteDance/SDXL-Lightning"
3074
+ description: "Text-to-image model for photorealistic generation.",
3075
+ id: "Kwai-Kolors/Kolors"
3062
3076
  },
3063
3077
  {
3064
3078
  description: "A powerful text-to-image model.",
@@ -3419,6 +3433,10 @@ var taskData29 = {
3419
3433
  {
3420
3434
  description: "An instruction dataset with preference ratings on responses.",
3421
3435
  id: "openbmb/UltraFeedback"
3436
+ },
3437
+ {
3438
+ description: "A large synthetic dataset for alignment of text generation models.",
3439
+ id: "argilla/magpie-ultra-v0.1"
3422
3440
  }
3423
3441
  ],
3424
3442
  demo: {
@@ -3449,32 +3467,32 @@ var taskData29 = {
3449
3467
  ],
3450
3468
  models: [
3451
3469
  {
3452
- description: "A large language model trained for text generation.",
3453
- id: "bigscience/bloom-560m"
3470
+ description: "A text-generation model trained to follow instructions.",
3471
+ id: "google/gemma-2-2b-it"
3454
3472
  },
3455
3473
  {
3456
- description: "A large code generation model that can generate code in 80+ languages.",
3474
+ description: "A code generation model that can generate code in 80+ languages.",
3457
3475
  id: "bigcode/starcoder"
3458
3476
  },
3459
3477
  {
3460
- description: "A very powerful text generation model.",
3461
- id: "mistralai/Mixtral-8x7B-Instruct-v0.1"
3478
+ description: "Very powerful text generation model trained to follow instructions.",
3479
+ id: "meta-llama/Meta-Llama-3.1-8B-Instruct"
3462
3480
  },
3463
3481
  {
3464
3482
  description: "Small yet powerful text generation model.",
3465
- id: "microsoft/phi-2"
3483
+ id: "microsoft/Phi-3-mini-4k-instruct"
3466
3484
  },
3467
3485
  {
3468
- description: "A very powerful model that can chat, do mathematical reasoning and write code.",
3469
- id: "openchat/openchat-3.5-0106"
3486
+ description: "A very powerful model that can solve mathematical problems.",
3487
+ id: "AI-MO/NuminaMath-7B-TIR"
3470
3488
  },
3471
3489
  {
3472
- description: "Very strong yet small assistant model.",
3473
- id: "HuggingFaceH4/zephyr-7b-beta"
3490
+ description: "Strong coding assistant model.",
3491
+ id: "HuggingFaceH4/starchat2-15b-v0.1"
3474
3492
  },
3475
3493
  {
3476
3494
  description: "Very strong open-source large language model.",
3477
- id: "meta-llama/Llama-2-70b-hf"
3495
+ id: "mistralai/Mistral-Nemo-Instruct-2407"
3478
3496
  }
3479
3497
  ],
3480
3498
  spaces: [
@@ -3501,7 +3519,7 @@ var taskData29 = {
3501
3519
  ],
3502
3520
  summary: "Generating text is the task of generating new text given another text. These models can, for example, fill in incomplete text or paraphrase.",
3503
3521
  widgetModels: ["HuggingFaceH4/zephyr-7b-beta"],
3504
- youtubeId: "Vpjb1lu0MDk"
3522
+ youtubeId: "e9gNEAlsOvU"
3505
3523
  };
3506
3524
  var data_default29 = taskData29;
3507
3525
 
@@ -4226,6 +4244,7 @@ var TASKS_MODEL_LIBRARIES = {
4226
4244
  ],
4227
4245
  translation: ["transformers", "transformers.js"],
4228
4246
  "unconditional-image-generation": ["diffusers"],
4247
+ "video-text-to-text": ["transformers"],
4229
4248
  "visual-question-answering": ["transformers", "transformers.js"],
4230
4249
  "voice-activity-detection": [],
4231
4250
  "zero-shot-classification": ["transformers", "transformers.js"],
@@ -4285,6 +4304,7 @@ var TASKS_DATA = {
4285
4304
  "token-classification": getData("token-classification", data_default26),
4286
4305
  translation: getData("translation", data_default27),
4287
4306
  "unconditional-image-generation": getData("unconditional-image-generation", data_default31),
4307
+ "video-text-to-text": getData("video-text-to-text", data_default16),
4288
4308
  "visual-question-answering": getData("visual-question-answering", data_default33),
4289
4309
  "voice-activity-detection": void 0,
4290
4310
  "zero-shot-classification": getData("zero-shot-classification", data_default34),
@@ -4522,6 +4542,23 @@ tokenizer = keras_nlp.models.Tokenizer.from_preset("hf://${model.id}")
4522
4542
  backbone = keras_nlp.models.Backbone.from_preset("hf://${model.id}")
4523
4543
  `
4524
4544
  ];
4545
+ var llama_cpp_python = (model) => [
4546
+ `from llama_cpp import Llama
4547
+
4548
+ llm = Llama.from_pretrained(
4549
+ repo_id="${model.id}",
4550
+ filename="{{GGUF_FILE}}",
4551
+ )
4552
+
4553
+ llm.create_chat_completion(
4554
+ messages = [
4555
+ {
4556
+ "role": "user",
4557
+ "content": "What is the capital of France?"
4558
+ }
4559
+ ]
4560
+ )`
4561
+ ];
4525
4562
  var tf_keras = (model) => [
4526
4563
  `# Note: 'keras<3.x' or 'tf_keras' must be installed (legacy)
4527
4564
  # See https://github.com/keras-team/tf-keras for more details.
@@ -5292,6 +5329,12 @@ var MODEL_LIBRARIES_UI_ELEMENTS = {
5292
5329
  filter: false,
5293
5330
  countDownloads: `path:"liveportrait/landmark.onnx"`
5294
5331
  },
5332
+ "llama-cpp-python": {
5333
+ prettyLabel: "llama-cpp-python",
5334
+ repoName: "llama-cpp-python",
5335
+ repoUrl: "https://github.com/abetlen/llama-cpp-python",
5336
+ snippets: llama_cpp_python
5337
+ },
5295
5338
  mindspore: {
5296
5339
  prettyLabel: "MindSpore",
5297
5340
  repoName: "mindspore",
@@ -5414,6 +5457,13 @@ var MODEL_LIBRARIES_UI_ELEMENTS = {
5414
5457
  snippets: saelens,
5415
5458
  filter: false
5416
5459
  },
5460
+ sam2: {
5461
+ prettyLabel: "sam2",
5462
+ repoName: "sam2",
5463
+ repoUrl: "https://github.com/facebookresearch/segment-anything-2",
5464
+ filter: false,
5465
+ countDownloads: `path_extension:"pt"`
5466
+ },
5417
5467
  "sample-factory": {
5418
5468
  prettyLabel: "sample-factory",
5419
5469
  repoName: "sample-factory",
@@ -6640,7 +6690,6 @@ var LOCAL_APPS = {
6640
6690
  docsUrl: "https://diffusionbee.com",
6641
6691
  mainTask: "text-to-image",
6642
6692
  macOSOnly: true,
6643
- comingSoon: true,
6644
6693
  displayOnModelPage: (model) => model.library_name === "diffusers" && model.pipeline_tag === "text-to-image",
6645
6694
  deeplink: (model) => new URL(`diffusionbee://open_from_hf?model=${model.id}`)
6646
6695
  }
package/dist/index.js CHANGED
@@ -1350,6 +1350,12 @@ var PIPELINE_DATA = {
1350
1350
  modality: "cv",
1351
1351
  color: "indigo"
1352
1352
  },
1353
+ "video-text-to-text": {
1354
+ name: "Video-Text-to-Text",
1355
+ modality: "multimodal",
1356
+ color: "blue",
1357
+ hideInDatasets: true
1358
+ },
1353
1359
  other: {
1354
1360
  name: "Other",
1355
1361
  modality: "other",
@@ -1693,8 +1699,12 @@ var taskData5 = {
1693
1699
  ],
1694
1700
  spaces: [
1695
1701
  {
1696
- description: "A leaderboard to rank best feature extraction models..",
1702
+ description: "A leaderboard to rank text feature extraction models based on a benchmark.",
1697
1703
  id: "mteb/leaderboard"
1704
+ },
1705
+ {
1706
+ description: "A leaderboard to rank best feature extraction models based on human feedback.",
1707
+ id: "mteb/arena"
1698
1708
  }
1699
1709
  ],
1700
1710
  summary: "Feature extraction is the task of extracting features learnt in a model.",
@@ -1899,15 +1909,19 @@ var taskData8 = {
1899
1909
  },
1900
1910
  {
1901
1911
  description: "A strong image feature extraction model.",
1902
- id: "google/vit-base-patch16-224-in21k"
1912
+ id: "nvidia/MambaVision-T-1K"
1903
1913
  },
1904
1914
  {
1905
- description: "A robust image feature extraction models.",
1915
+ description: "A robust image feature extraction model.",
1906
1916
  id: "facebook/dino-vitb16"
1907
1917
  },
1908
1918
  {
1909
- description: "Strong image-text-to-text model made for information retrieval from documents.",
1919
+ description: "Strong image feature extraction model made for information retrieval from documents.",
1910
1920
  id: "vidore/colpali"
1921
+ },
1922
+ {
1923
+ description: "Strong image feature extraction model that can be used on images and documents.",
1924
+ id: "OpenGVLab/InternViT-6B-448px-V1-2"
1911
1925
  }
1912
1926
  ],
1913
1927
  spaces: [],
@@ -1959,8 +1973,8 @@ var taskData9 = {
1959
1973
  ],
1960
1974
  models: [
1961
1975
  {
1962
- description: "A model that enhances images captured in low light conditions.",
1963
- id: "keras-io/low-light-image-enhancement"
1976
+ description: "An image-to-image model to improve image resolution.",
1977
+ id: "fal/AuraSR-v2"
1964
1978
  },
1965
1979
  {
1966
1980
  description: "A model that increases the resolution of an image.",
@@ -2178,7 +2192,7 @@ var taskData11 = {
2178
2192
  ],
2179
2193
  summary: "Image-text-to-text models take in an image and text prompt and output text. These models are also called vision-language models, or VLMs. The difference from image-to-text models is that these models take an additional text input, not restricting the model to certain use cases like image captioning, and may also be trained to accept a conversation as input.",
2180
2194
  widgetModels: ["microsoft/kosmos-2-patch14-224"],
2181
- youtubeId: ""
2195
+ youtubeId: "IoGaGfU1CIg"
2182
2196
  };
2183
2197
  var data_default11 = taskData11;
2184
2198
 
@@ -2229,16 +2243,16 @@ var taskData12 = {
2229
2243
  id: "facebook/detr-resnet-50-panoptic"
2230
2244
  },
2231
2245
  {
2232
- description: "Semantic segmentation model trained on ADE20k benchmark dataset.",
2233
- id: "microsoft/beit-large-finetuned-ade-640-640"
2246
+ description: "Background removal model.",
2247
+ id: "briaai/RMBG-1.4"
2234
2248
  },
2235
2249
  {
2236
2250
  description: "Semantic segmentation model trained on ADE20k benchmark dataset with 512x512 resolution.",
2237
2251
  id: "nvidia/segformer-b0-finetuned-ade-512-512"
2238
2252
  },
2239
2253
  {
2240
- description: "Semantic segmentation model trained Cityscapes dataset.",
2241
- id: "facebook/mask2former-swin-large-cityscapes-semantic"
2254
+ description: "A multipurpose image segmentation model for high resolution images.",
2255
+ id: "ZhengPeng7/BiRefNet"
2242
2256
  },
2243
2257
  {
2244
2258
  description: "Panoptic segmentation model trained COCO (common objects) dataset.",
@@ -2302,13 +2316,13 @@ var taskData13 = {
2302
2316
  },
2303
2317
  {
2304
2318
  description: "Very strong mask generation model.",
2305
- id: "facebook/sam-vit-huge"
2319
+ id: "facebook/sam2-hiera-large"
2306
2320
  }
2307
2321
  ],
2308
2322
  spaces: [
2309
2323
  {
2310
- description: "An application that combines a mask generation model with an image embedding model for open-vocabulary image segmentation.",
2311
- id: "SkalskiP/SAM_and_MetaCLIP"
2324
+ description: "An application that combines a mask generation model with a zero-shot object detection model for text-guided image segmentation.",
2325
+ id: "merve/OWLSAM2"
2312
2326
  },
2313
2327
  {
2314
2328
  description: "An application that compares the performance of a large and a small mask generation model.",
@@ -2316,7 +2330,7 @@ var taskData13 = {
2316
2330
  },
2317
2331
  {
2318
2332
  description: "An application based on an improved mask generation model.",
2319
- id: "linfanluntan/Grounded-SAM"
2333
+ id: "SkalskiP/segment-anything-model-2"
2320
2334
  },
2321
2335
  {
2322
2336
  description: "An application to remove objects from videos using mask generation models.",
@@ -3012,15 +3026,15 @@ var taskData24 = {
3012
3026
  models: [
3013
3027
  {
3014
3028
  description: "One of the most powerful image generation models that can generate realistic outputs.",
3015
- id: "stabilityai/stable-diffusion-xl-base-1.0"
3029
+ id: "black-forest-labs/FLUX.1-dev"
3016
3030
  },
3017
3031
  {
3018
3032
  description: "A powerful yet fast image generation model.",
3019
3033
  id: "latent-consistency/lcm-lora-sdxl"
3020
3034
  },
3021
3035
  {
3022
- description: "A very fast text-to-image model.",
3023
- id: "ByteDance/SDXL-Lightning"
3036
+ description: "Text-to-image model for photorealistic generation.",
3037
+ id: "Kwai-Kolors/Kolors"
3024
3038
  },
3025
3039
  {
3026
3040
  description: "A powerful text-to-image model.",
@@ -3381,6 +3395,10 @@ var taskData29 = {
3381
3395
  {
3382
3396
  description: "An instruction dataset with preference ratings on responses.",
3383
3397
  id: "openbmb/UltraFeedback"
3398
+ },
3399
+ {
3400
+ description: "A large synthetic dataset for alignment of text generation models.",
3401
+ id: "argilla/magpie-ultra-v0.1"
3384
3402
  }
3385
3403
  ],
3386
3404
  demo: {
@@ -3411,32 +3429,32 @@ var taskData29 = {
3411
3429
  ],
3412
3430
  models: [
3413
3431
  {
3414
- description: "A large language model trained for text generation.",
3415
- id: "bigscience/bloom-560m"
3432
+ description: "A text-generation model trained to follow instructions.",
3433
+ id: "google/gemma-2-2b-it"
3416
3434
  },
3417
3435
  {
3418
- description: "A large code generation model that can generate code in 80+ languages.",
3436
+ description: "A code generation model that can generate code in 80+ languages.",
3419
3437
  id: "bigcode/starcoder"
3420
3438
  },
3421
3439
  {
3422
- description: "A very powerful text generation model.",
3423
- id: "mistralai/Mixtral-8x7B-Instruct-v0.1"
3440
+ description: "Very powerful text generation model trained to follow instructions.",
3441
+ id: "meta-llama/Meta-Llama-3.1-8B-Instruct"
3424
3442
  },
3425
3443
  {
3426
3444
  description: "Small yet powerful text generation model.",
3427
- id: "microsoft/phi-2"
3445
+ id: "microsoft/Phi-3-mini-4k-instruct"
3428
3446
  },
3429
3447
  {
3430
- description: "A very powerful model that can chat, do mathematical reasoning and write code.",
3431
- id: "openchat/openchat-3.5-0106"
3448
+ description: "A very powerful model that can solve mathematical problems.",
3449
+ id: "AI-MO/NuminaMath-7B-TIR"
3432
3450
  },
3433
3451
  {
3434
- description: "Very strong yet small assistant model.",
3435
- id: "HuggingFaceH4/zephyr-7b-beta"
3452
+ description: "Strong coding assistant model.",
3453
+ id: "HuggingFaceH4/starchat2-15b-v0.1"
3436
3454
  },
3437
3455
  {
3438
3456
  description: "Very strong open-source large language model.",
3439
- id: "meta-llama/Llama-2-70b-hf"
3457
+ id: "mistralai/Mistral-Nemo-Instruct-2407"
3440
3458
  }
3441
3459
  ],
3442
3460
  spaces: [
@@ -3463,7 +3481,7 @@ var taskData29 = {
3463
3481
  ],
3464
3482
  summary: "Generating text is the task of generating new text given another text. These models can, for example, fill in incomplete text or paraphrase.",
3465
3483
  widgetModels: ["HuggingFaceH4/zephyr-7b-beta"],
3466
- youtubeId: "Vpjb1lu0MDk"
3484
+ youtubeId: "e9gNEAlsOvU"
3467
3485
  };
3468
3486
  var data_default29 = taskData29;
3469
3487
 
@@ -4188,6 +4206,7 @@ var TASKS_MODEL_LIBRARIES = {
4188
4206
  ],
4189
4207
  translation: ["transformers", "transformers.js"],
4190
4208
  "unconditional-image-generation": ["diffusers"],
4209
+ "video-text-to-text": ["transformers"],
4191
4210
  "visual-question-answering": ["transformers", "transformers.js"],
4192
4211
  "voice-activity-detection": [],
4193
4212
  "zero-shot-classification": ["transformers", "transformers.js"],
@@ -4247,6 +4266,7 @@ var TASKS_DATA = {
4247
4266
  "token-classification": getData("token-classification", data_default26),
4248
4267
  translation: getData("translation", data_default27),
4249
4268
  "unconditional-image-generation": getData("unconditional-image-generation", data_default31),
4269
+ "video-text-to-text": getData("video-text-to-text", data_default16),
4250
4270
  "visual-question-answering": getData("visual-question-answering", data_default33),
4251
4271
  "voice-activity-detection": void 0,
4252
4272
  "zero-shot-classification": getData("zero-shot-classification", data_default34),
@@ -4484,6 +4504,23 @@ tokenizer = keras_nlp.models.Tokenizer.from_preset("hf://${model.id}")
4484
4504
  backbone = keras_nlp.models.Backbone.from_preset("hf://${model.id}")
4485
4505
  `
4486
4506
  ];
4507
+ var llama_cpp_python = (model) => [
4508
+ `from llama_cpp import Llama
4509
+
4510
+ llm = Llama.from_pretrained(
4511
+ repo_id="${model.id}",
4512
+ filename="{{GGUF_FILE}}",
4513
+ )
4514
+
4515
+ llm.create_chat_completion(
4516
+ messages = [
4517
+ {
4518
+ "role": "user",
4519
+ "content": "What is the capital of France?"
4520
+ }
4521
+ ]
4522
+ )`
4523
+ ];
4487
4524
  var tf_keras = (model) => [
4488
4525
  `# Note: 'keras<3.x' or 'tf_keras' must be installed (legacy)
4489
4526
  # See https://github.com/keras-team/tf-keras for more details.
@@ -5254,6 +5291,12 @@ var MODEL_LIBRARIES_UI_ELEMENTS = {
5254
5291
  filter: false,
5255
5292
  countDownloads: `path:"liveportrait/landmark.onnx"`
5256
5293
  },
5294
+ "llama-cpp-python": {
5295
+ prettyLabel: "llama-cpp-python",
5296
+ repoName: "llama-cpp-python",
5297
+ repoUrl: "https://github.com/abetlen/llama-cpp-python",
5298
+ snippets: llama_cpp_python
5299
+ },
5257
5300
  mindspore: {
5258
5301
  prettyLabel: "MindSpore",
5259
5302
  repoName: "mindspore",
@@ -5376,6 +5419,13 @@ var MODEL_LIBRARIES_UI_ELEMENTS = {
5376
5419
  snippets: saelens,
5377
5420
  filter: false
5378
5421
  },
5422
+ sam2: {
5423
+ prettyLabel: "sam2",
5424
+ repoName: "sam2",
5425
+ repoUrl: "https://github.com/facebookresearch/segment-anything-2",
5426
+ filter: false,
5427
+ countDownloads: `path_extension:"pt"`
5428
+ },
5379
5429
  "sample-factory": {
5380
5430
  prettyLabel: "sample-factory",
5381
5431
  repoName: "sample-factory",
@@ -6602,7 +6652,6 @@ var LOCAL_APPS = {
6602
6652
  docsUrl: "https://diffusionbee.com",
6603
6653
  mainTask: "text-to-image",
6604
6654
  macOSOnly: true,
6605
- comingSoon: true,
6606
6655
  displayOnModelPage: (model) => model.library_name === "diffusers" && model.pipeline_tag === "text-to-image",
6607
6656
  deeplink: (model) => new URL(`diffusionbee://open_from_hf?model=${model.id}`)
6608
6657
  }
@@ -134,7 +134,6 @@ export declare const LOCAL_APPS: {
134
134
  docsUrl: string;
135
135
  mainTask: "text-to-image";
136
136
  macOSOnly: true;
137
- comingSoon: true;
138
137
  displayOnModelPage: (model: ModelData) => boolean;
139
138
  deeplink: (model: ModelData) => URL;
140
139
  };
@@ -1 +1 @@
1
- {"version":3,"file":"local-apps.d.ts","sourceRoot":"","sources":["../../src/local-apps.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,SAAS,EAAE,MAAM,cAAc,CAAC;AAC9C,OAAO,KAAK,EAAE,YAAY,EAAE,MAAM,aAAa,CAAC;AAEhD,MAAM,WAAW,eAAe;IAC/B;;OAEG;IACH,KAAK,EAAE,MAAM,CAAC;IACd;;OAEG;IACH,KAAK,CAAC,EAAE,MAAM,CAAC;IACf;;OAEG;IACH,OAAO,EAAE,MAAM,CAAC;CAChB;AAED;;GAEG;AACH,MAAM,MAAM,QAAQ,GAAG;IACtB;;OAEG;IACH,WAAW,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,OAAO,EAAE,MAAM,CAAC;IAChB;;OAEG;IACH,QAAQ,EAAE,YAAY,CAAC;IACvB;;OAEG;IACH,SAAS,CAAC,EAAE,OAAO,CAAC;IAEpB,UAAU,CAAC,EAAE,OAAO,CAAC;IACrB;;OAEG;IACH,kBAAkB,EAAE,CAAC,KAAK,EAAE,SAAS,KAAK,OAAO,CAAC;CAClD,GAAG,CACD;IACA;;OAEG;IACH,QAAQ,EAAE,CAAC,KAAK,EAAE,SAAS,EAAE,QAAQ,CAAC,EAAE,MAAM,KAAK,GAAG,CAAC;CACtD,GACD;IACA;;;OAGG;IACH,OAAO,EAAE,CAAC,KAAK,EAAE,SAAS,EAAE,QAAQ,CAAC,EAAE,MAAM,KAAK,MAAM,GAAG,MAAM,EAAE,GAAG,eAAe,GAAG,eAAe,EAAE,CAAC;CACzG,CACH,CAAC;AAEF,iBAAS,WAAW,CAAC,KAAK,EAAE,SAAS,WAEpC;AAuCD;;;;;;;;;;GAUG;AACH,eAAO,MAAM,UAAU;;;;;;yBAhDS,SAAS,aAAa,MAAM,KAAG,eAAe,EAAE;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CA+I5C,CAAC;AAErC,MAAM,MAAM,WAAW,GAAG,MAAM,OAAO,UAAU,CAAC"}
1
+ {"version":3,"file":"local-apps.d.ts","sourceRoot":"","sources":["../../src/local-apps.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,SAAS,EAAE,MAAM,cAAc,CAAC;AAC9C,OAAO,KAAK,EAAE,YAAY,EAAE,MAAM,aAAa,CAAC;AAEhD,MAAM,WAAW,eAAe;IAC/B;;OAEG;IACH,KAAK,EAAE,MAAM,CAAC;IACd;;OAEG;IACH,KAAK,CAAC,EAAE,MAAM,CAAC;IACf;;OAEG;IACH,OAAO,EAAE,MAAM,CAAC;CAChB;AAED;;GAEG;AACH,MAAM,MAAM,QAAQ,GAAG;IACtB;;OAEG;IACH,WAAW,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,OAAO,EAAE,MAAM,CAAC;IAChB;;OAEG;IACH,QAAQ,EAAE,YAAY,CAAC;IACvB;;OAEG;IACH,SAAS,CAAC,EAAE,OAAO,CAAC;IAEpB,UAAU,CAAC,EAAE,OAAO,CAAC;IACrB;;OAEG;IACH,kBAAkB,EAAE,CAAC,KAAK,EAAE,SAAS,KAAK,OAAO,CAAC;CAClD,GAAG,CACD;IACA;;OAEG;IACH,QAAQ,EAAE,CAAC,KAAK,EAAE,SAAS,EAAE,QAAQ,CAAC,EAAE,MAAM,KAAK,GAAG,CAAC;CACtD,GACD;IACA;;;OAGG;IACH,OAAO,EAAE,CAAC,KAAK,EAAE,SAAS,EAAE,QAAQ,CAAC,EAAE,MAAM,KAAK,MAAM,GAAG,MAAM,EAAE,GAAG,eAAe,GAAG,eAAe,EAAE,CAAC;CACzG,CACH,CAAC;AAEF,iBAAS,WAAW,CAAC,KAAK,EAAE,SAAS,WAEpC;AAuCD;;;;;;;;;;GAUG;AACH,eAAO,MAAM,UAAU;;;;;;yBAhDS,SAAS,aAAa,MAAM,KAAG,eAAe,EAAE;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CA8I5C,CAAC;AAErC,MAAM,MAAM,WAAW,GAAG,MAAM,OAAO,UAAU,CAAC"}
@@ -16,6 +16,7 @@ export declare const flair: (model: ModelData) => string[];
16
16
  export declare const gliner: (model: ModelData) => string[];
17
17
  export declare const keras: (model: ModelData) => string[];
18
18
  export declare const keras_nlp: (model: ModelData) => string[];
19
+ export declare const llama_cpp_python: (model: ModelData) => string[];
19
20
  export declare const tf_keras: (model: ModelData) => string[];
20
21
  export declare const mamba_ssm: (model: ModelData) => string[];
21
22
  export declare const mars5_tts: (model: ModelData) => string[];
@@ -1 +1 @@
1
- {"version":3,"file":"model-libraries-snippets.d.ts","sourceRoot":"","sources":["../../src/model-libraries-snippets.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,SAAS,EAAE,MAAM,cAAc,CAAC;AAY9C,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAKjD,CAAC;AAkBF,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAKjD,CAAC;AAEF,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAIjD,CAAC;AAEF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EAkBlD,CAAC;AAMF,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAIjD,CAAC;AAEF,eAAO,MAAM,KAAK,UAAW,SAAS,KAAG,MAAM,EAI9C,CAAC;AAEF,eAAO,MAAM,iBAAiB,UAAW,SAAS,KAAG,MAAM,EA6C1D,CAAC;AA+BF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EAUlD,CAAC;AAEF,eAAO,MAAM,MAAM,UAAW,SAAS,KAAG,MAAM,EAgB/C,CAAC;AAEF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EAMlD,CAAC;AAEF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EASlD,CAAC;AAIF,eAAO,MAAM,MAAM,UAAW,SAAS,KAAG,MAAM,EAO/C,CAAC;AAEF,eAAO,MAAM,OAAO,UAAW,SAAS,KAAG,MAAM,EAMhD,CAAC;AAEF,eAAO,MAAM,KAAK,UAAW,SAAS,KAAG,MAAM,EAI9C,CAAC;AAEF,eAAO,MAAM,MAAM,UAAW,SAAS,KAAG,MAAM,EAI/C,CAAC;AAEF,eAAO,MAAM,KAAK,UAAW,SAAS,KAAG,MAAM,EAS9C,CAAC;AAEF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EAUlD,CAAC;AAEF,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAOjD,CAAC;AAEF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EAIlD,CAAC;AAEF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EAKlD,CAAC;AAEF,eAAO,MAAM,aAAa,QAAO,MAAM,EAQtC,CAAC;AAEF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EAKlD,CAAC;AAEF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EAsBlD,CAAC;AAEF,eAAO,MAAM,uBAAuB,UAAW,SAAS,KAAG,MAAM,EAehE,CAAC;AAiBF,eAAO,MAAM,cAAc,UAAW,SAAS,KAAG,MAAM,EAKvD,CAAC;AAyBF,eAAO,MAAM,aAAa,UAAW,SAAS,KAAG,MAAM,EAOtD,CAAC;AAEF,eAAO,MAAM,IAAI,UAAW,SAAS,KAAG,MAAM,EAI7C,CAAC;AAEF,eAAO,MAAM,OAAO,QAA6B,MAAM,EAQtD,CAAC;AAsCF,eAAO,MAAM,OAAO,UAAW,SAAS,KAAG,MAAM,EAehD,CAAC;AAEF,eAAO,MAAM,kBAAkB,UAAW,SAAS,KAAG,MAAM,EAmC3D,CAAC;AAEF,eAAO,MAAM,MAAM,UAAW,SAAS,KAAG,MAAM,EAI/C,CAAC;AAEF,eAAO,MAAM,aAAa,UAAW,SAAS,KAAG,MAAM,EAEtD,CAAC;AAEF,eAAO,MAAM,oBAAoB,UAAW,SAAS,KAAG,MAAM,EAI7D,CAAC;AAEF,eAAO,MAAM,MAAM,UAAW,SAAS,KAAG,MAAM,EAI/C,CAAC;AAEF,eAAO,MAAM,KAAK,UAAW,SAAS,KAAG,MAAM,EAU9C,CAAC;AAEF,eAAO,MAAM,WAAW,UAAW,SAAS,KAAG,MAAM,EAIpD,CAAC;AAEF,eAAO,MAAM,MAAM,UAAW,SAAS,KAAG,MAAM,EAK/C,CAAC;AAkBF,eAAO,MAAM,WAAW,UAAW,SAAS,KAAG,MAAM,EAkBpD,CAAC;AAEF,eAAO,MAAM,YAAY,UAAW,SAAS,KAAG,MAAM,EA4CrD,CAAC;AAEF,eAAO,MAAM,cAAc,UAAW,SAAS,KAAG,MAAM,EAcvD,CAAC;AAiBF,eAAO,MAAM,IAAI,UAAW,SAAS,KAAG,MAAM,EAkB7C,CAAC;AAEF,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAKjD,CAAC;AAEF,eAAO,MAAM,gBAAgB,UAAW,SAAS,KAAG,MAAM,EAMzD,CAAC;AAgBF,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAEjD,CAAC;AAEF,eAAO,MAAM,MAAM,QAA6B,MAAM,EAMrD,CAAC;AAEF,eAAO,MAAM,UAAU,UAAW,SAAS,KAAG,MAAM,EAInD,CAAC;AAEF,eAAO,MAAM,OAAO,QAAO,MAAM,EAYhC,CAAC;AAEF,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAYjD,CAAC;AAEF,eAAO,MAAM,GAAG,UAAW,SAAS,KAAG,MAAM,EAK5C,CAAC;AAEF,eAAO,MAAM,KAAK,UAAW,SAAS,KAAG,MAAM,EAI9C,CAAC;AAEF,eAAO,MAAM,IAAI,UAAW,SAAS,KAAG,MAAM,EAQ7C,CAAC;AAEF,eAAO,MAAM,MAAM,UAAW,SAAS,KAAG,MAAM,EAI/C,CAAC;AA6BF,eAAO,MAAM,UAAU,UAAW,SAAS,KAAG,MAAM,EAUnD,CAAC;AAEF,eAAO,MAAM,UAAU,QAAO,MAAM,EAYnC,CAAC"}
1
+ {"version":3,"file":"model-libraries-snippets.d.ts","sourceRoot":"","sources":["../../src/model-libraries-snippets.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,SAAS,EAAE,MAAM,cAAc,CAAC;AAY9C,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAKjD,CAAC;AAkBF,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAKjD,CAAC;AAEF,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAIjD,CAAC;AAEF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EAkBlD,CAAC;AAMF,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAIjD,CAAC;AAEF,eAAO,MAAM,KAAK,UAAW,SAAS,KAAG,MAAM,EAI9C,CAAC;AAEF,eAAO,MAAM,iBAAiB,UAAW,SAAS,KAAG,MAAM,EA6C1D,CAAC;AA+BF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EAUlD,CAAC;AAEF,eAAO,MAAM,MAAM,UAAW,SAAS,KAAG,MAAM,EAgB/C,CAAC;AAEF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EAMlD,CAAC;AAEF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EASlD,CAAC;AAIF,eAAO,MAAM,MAAM,UAAW,SAAS,KAAG,MAAM,EAO/C,CAAC;AAEF,eAAO,MAAM,OAAO,UAAW,SAAS,KAAG,MAAM,EAMhD,CAAC;AAEF,eAAO,MAAM,KAAK,UAAW,SAAS,KAAG,MAAM,EAI9C,CAAC;AAEF,eAAO,MAAM,MAAM,UAAW,SAAS,KAAG,MAAM,EAI/C,CAAC;AAEF,eAAO,MAAM,KAAK,UAAW,SAAS,KAAG,MAAM,EAS9C,CAAC;AAEF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EAUlD,CAAC;AAEF,eAAO,MAAM,gBAAgB,UAAW,SAAS,KAAG,MAAM,EAgBzD,CAAC;AAEF,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAOjD,CAAC;AAEF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EAIlD,CAAC;AAEF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EAKlD,CAAC;AAEF,eAAO,MAAM,aAAa,QAAO,MAAM,EAQtC,CAAC;AAEF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EAKlD,CAAC;AAEF,eAAO,MAAM,SAAS,UAAW,SAAS,KAAG,MAAM,EAsBlD,CAAC;AAEF,eAAO,MAAM,uBAAuB,UAAW,SAAS,KAAG,MAAM,EAehE,CAAC;AAiBF,eAAO,MAAM,cAAc,UAAW,SAAS,KAAG,MAAM,EAKvD,CAAC;AAyBF,eAAO,MAAM,aAAa,UAAW,SAAS,KAAG,MAAM,EAOtD,CAAC;AAEF,eAAO,MAAM,IAAI,UAAW,SAAS,KAAG,MAAM,EAI7C,CAAC;AAEF,eAAO,MAAM,OAAO,QAA6B,MAAM,EAQtD,CAAC;AAsCF,eAAO,MAAM,OAAO,UAAW,SAAS,KAAG,MAAM,EAehD,CAAC;AAEF,eAAO,MAAM,kBAAkB,UAAW,SAAS,KAAG,MAAM,EAmC3D,CAAC;AAEF,eAAO,MAAM,MAAM,UAAW,SAAS,KAAG,MAAM,EAI/C,CAAC;AAEF,eAAO,MAAM,aAAa,UAAW,SAAS,KAAG,MAAM,EAEtD,CAAC;AAEF,eAAO,MAAM,oBAAoB,UAAW,SAAS,KAAG,MAAM,EAI7D,CAAC;AAEF,eAAO,MAAM,MAAM,UAAW,SAAS,KAAG,MAAM,EAI/C,CAAC;AAEF,eAAO,MAAM,KAAK,UAAW,SAAS,KAAG,MAAM,EAU9C,CAAC;AAEF,eAAO,MAAM,WAAW,UAAW,SAAS,KAAG,MAAM,EAIpD,CAAC;AAEF,eAAO,MAAM,MAAM,UAAW,SAAS,KAAG,MAAM,EAK/C,CAAC;AAkBF,eAAO,MAAM,WAAW,UAAW,SAAS,KAAG,MAAM,EAkBpD,CAAC;AAEF,eAAO,MAAM,YAAY,UAAW,SAAS,KAAG,MAAM,EA4CrD,CAAC;AAEF,eAAO,MAAM,cAAc,UAAW,SAAS,KAAG,MAAM,EAcvD,CAAC;AAiBF,eAAO,MAAM,IAAI,UAAW,SAAS,KAAG,MAAM,EAkB7C,CAAC;AAEF,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAKjD,CAAC;AAEF,eAAO,MAAM,gBAAgB,UAAW,SAAS,KAAG,MAAM,EAMzD,CAAC;AAgBF,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAEjD,CAAC;AAEF,eAAO,MAAM,MAAM,QAA6B,MAAM,EAMrD,CAAC;AAEF,eAAO,MAAM,UAAU,UAAW,SAAS,KAAG,MAAM,EAInD,CAAC;AAEF,eAAO,MAAM,OAAO,QAAO,MAAM,EAYhC,CAAC;AAEF,eAAO,MAAM,QAAQ,UAAW,SAAS,KAAG,MAAM,EAYjD,CAAC;AAEF,eAAO,MAAM,GAAG,UAAW,SAAS,KAAG,MAAM,EAK5C,CAAC;AAEF,eAAO,MAAM,KAAK,UAAW,SAAS,KAAG,MAAM,EAI9C,CAAC;AAEF,eAAO,MAAM,IAAI,UAAW,SAAS,KAAG,MAAM,EAQ7C,CAAC;AAEF,eAAO,MAAM,MAAM,UAAW,SAAS,KAAG,MAAM,EAI/C,CAAC;AA6BF,eAAO,MAAM,UAAU,UAAW,SAAS,KAAG,MAAM,EAUnD,CAAC;AAEF,eAAO,MAAM,UAAU,QAAO,MAAM,EAYnC,CAAC"}
@@ -308,6 +308,12 @@ export declare const MODEL_LIBRARIES_UI_ELEMENTS: {
308
308
  filter: false;
309
309
  countDownloads: string;
310
310
  };
311
+ "llama-cpp-python": {
312
+ prettyLabel: string;
313
+ repoName: string;
314
+ repoUrl: string;
315
+ snippets: (model: ModelData) => string[];
316
+ };
311
317
  mindspore: {
312
318
  prettyLabel: string;
313
319
  repoName: string;
@@ -430,6 +436,13 @@ export declare const MODEL_LIBRARIES_UI_ELEMENTS: {
430
436
  snippets: () => string[];
431
437
  filter: false;
432
438
  };
439
+ sam2: {
440
+ prettyLabel: string;
441
+ repoName: string;
442
+ repoUrl: string;
443
+ filter: false;
444
+ countDownloads: string;
445
+ };
433
446
  "sample-factory": {
434
447
  prettyLabel: string;
435
448
  repoName: string;
@@ -592,6 +605,6 @@ export declare const MODEL_LIBRARIES_UI_ELEMENTS: {
592
605
  };
593
606
  };
594
607
  export type ModelLibraryKey = keyof typeof MODEL_LIBRARIES_UI_ELEMENTS;
595
- export declare const ALL_MODEL_LIBRARY_KEYS: ("adapter-transformers" | "allennlp" | "asteroid" | "audiocraft" | "audioseal" | "bertopic" | "big_vision" | "birefnet" | "bm25s" | "champ" | "chat_tts" | "colpali" | "depth-anything-v2" | "diffree" | "diffusers" | "doctr" | "edsnlp" | "elm" | "espnet" | "fairseq" | "fastai" | "fasttext" | "flair" | "gemma.cpp" | "gliner" | "glyph-byt5" | "grok" | "hallo" | "hunyuan-dit" | "keras" | "tf-keras" | "keras-nlp" | "k2" | "liveportrait" | "mindspore" | "mamba-ssm" | "mars5-tts" | "mesh-anything" | "ml-agents" | "mlx" | "mlx-image" | "mlc-llm" | "nemo" | "open_clip" | "paddlenlp" | "peft" | "pyannote-audio" | "pythae" | "recurrentgemma" | "saelens" | "sample-factory" | "sentence-transformers" | "setfit" | "sklearn" | "spacy" | "span-marker" | "speechbrain" | "stable-audio-tools" | "diffusion-single-file" | "stable-baselines3" | "stanza" | "tensorflowtts" | "tic-clip" | "timesfm" | "timm" | "transformers" | "transformers.js" | "unity-sentis" | "voicecraft" | "whisperkit")[];
596
- export declare const ALL_DISPLAY_MODEL_LIBRARY_KEYS: ("adapter-transformers" | "allennlp" | "asteroid" | "audiocraft" | "audioseal" | "bertopic" | "big_vision" | "birefnet" | "bm25s" | "champ" | "chat_tts" | "colpali" | "depth-anything-v2" | "diffree" | "diffusers" | "doctr" | "edsnlp" | "elm" | "espnet" | "fairseq" | "fastai" | "fasttext" | "flair" | "gemma.cpp" | "gliner" | "glyph-byt5" | "grok" | "hallo" | "hunyuan-dit" | "keras" | "tf-keras" | "keras-nlp" | "k2" | "liveportrait" | "mindspore" | "mamba-ssm" | "mars5-tts" | "mesh-anything" | "ml-agents" | "mlx" | "mlx-image" | "mlc-llm" | "nemo" | "open_clip" | "paddlenlp" | "peft" | "pyannote-audio" | "pythae" | "recurrentgemma" | "saelens" | "sample-factory" | "sentence-transformers" | "setfit" | "sklearn" | "spacy" | "span-marker" | "speechbrain" | "stable-audio-tools" | "diffusion-single-file" | "stable-baselines3" | "stanza" | "tensorflowtts" | "tic-clip" | "timesfm" | "timm" | "transformers" | "transformers.js" | "unity-sentis" | "voicecraft" | "whisperkit")[];
608
+ export declare const ALL_MODEL_LIBRARY_KEYS: ("adapter-transformers" | "allennlp" | "asteroid" | "audiocraft" | "audioseal" | "bertopic" | "big_vision" | "birefnet" | "bm25s" | "champ" | "chat_tts" | "colpali" | "depth-anything-v2" | "diffree" | "diffusers" | "doctr" | "edsnlp" | "elm" | "espnet" | "fairseq" | "fastai" | "fasttext" | "flair" | "gemma.cpp" | "gliner" | "glyph-byt5" | "grok" | "hallo" | "hunyuan-dit" | "keras" | "tf-keras" | "keras-nlp" | "k2" | "liveportrait" | "llama-cpp-python" | "mindspore" | "mamba-ssm" | "mars5-tts" | "mesh-anything" | "ml-agents" | "mlx" | "mlx-image" | "mlc-llm" | "nemo" | "open_clip" | "paddlenlp" | "peft" | "pyannote-audio" | "pythae" | "recurrentgemma" | "saelens" | "sam2" | "sample-factory" | "sentence-transformers" | "setfit" | "sklearn" | "spacy" | "span-marker" | "speechbrain" | "stable-audio-tools" | "diffusion-single-file" | "stable-baselines3" | "stanza" | "tensorflowtts" | "tic-clip" | "timesfm" | "timm" | "transformers" | "transformers.js" | "unity-sentis" | "voicecraft" | "whisperkit")[];
609
+ export declare const ALL_DISPLAY_MODEL_LIBRARY_KEYS: ("adapter-transformers" | "allennlp" | "asteroid" | "audiocraft" | "audioseal" | "bertopic" | "big_vision" | "birefnet" | "bm25s" | "champ" | "chat_tts" | "colpali" | "depth-anything-v2" | "diffree" | "diffusers" | "doctr" | "edsnlp" | "elm" | "espnet" | "fairseq" | "fastai" | "fasttext" | "flair" | "gemma.cpp" | "gliner" | "glyph-byt5" | "grok" | "hallo" | "hunyuan-dit" | "keras" | "tf-keras" | "keras-nlp" | "k2" | "liveportrait" | "llama-cpp-python" | "mindspore" | "mamba-ssm" | "mars5-tts" | "mesh-anything" | "ml-agents" | "mlx" | "mlx-image" | "mlc-llm" | "nemo" | "open_clip" | "paddlenlp" | "peft" | "pyannote-audio" | "pythae" | "recurrentgemma" | "saelens" | "sam2" | "sample-factory" | "sentence-transformers" | "setfit" | "sklearn" | "spacy" | "span-marker" | "speechbrain" | "stable-audio-tools" | "diffusion-single-file" | "stable-baselines3" | "stanza" | "tensorflowtts" | "tic-clip" | "timesfm" | "timm" | "transformers" | "transformers.js" | "unity-sentis" | "voicecraft" | "whisperkit")[];
597
610
  //# sourceMappingURL=model-libraries.d.ts.map
@@ -1 +1 @@
1
- {"version":3,"file":"model-libraries.d.ts","sourceRoot":"","sources":["../../src/model-libraries.ts"],"names":[],"mappings":"AACA,OAAO,KAAK,EAAE,SAAS,EAAE,MAAM,cAAc,CAAC;AAC9C,OAAO,KAAK,EAAE,kBAAkB,EAAE,MAAM,6BAA6B,CAAC;AAEtE;;GAEG;AACH,MAAM,WAAW,gBAAgB;IAChC;;;;OAIG;IACH,WAAW,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,QAAQ,EAAE,MAAM,CAAC;IACjB;;OAEG;IACH,OAAO,EAAE,MAAM,CAAC;IAChB;;OAEG;IACH,OAAO,CAAC,EAAE,MAAM,CAAC;IACjB;;OAEG;IACH,QAAQ,CAAC,EAAE,CAAC,KAAK,EAAE,SAAS,KAAK,MAAM,EAAE,CAAC;IAC1C;;;;;OAKG;IACH,cAAc,CAAC,EAAE,kBAAkB,CAAC;IACpC;;;OAGG;IACH,MAAM,CAAC,EAAE,OAAO,CAAC;CACjB;AAED;;;;;;;;;;;;;GAaG;AAEH,eAAO,MAAM,2BAA2B;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CA2hBI,CAAC;AAE7C,MAAM,MAAM,eAAe,GAAG,MAAM,OAAO,2BAA2B,CAAC;AAEvE,eAAO,MAAM,sBAAsB,s9BAAgE,CAAC;AAEpG,eAAO,MAAM,8BAA8B,s9BAQ1B,CAAC"}
1
+ {"version":3,"file":"model-libraries.d.ts","sourceRoot":"","sources":["../../src/model-libraries.ts"],"names":[],"mappings":"AACA,OAAO,KAAK,EAAE,SAAS,EAAE,MAAM,cAAc,CAAC;AAC9C,OAAO,KAAK,EAAE,kBAAkB,EAAE,MAAM,6BAA6B,CAAC;AAEtE;;GAEG;AACH,MAAM,WAAW,gBAAgB;IAChC;;;;OAIG;IACH,WAAW,EAAE,MAAM,CAAC;IACpB;;OAEG;IACH,QAAQ,EAAE,MAAM,CAAC;IACjB;;OAEG;IACH,OAAO,EAAE,MAAM,CAAC;IAChB;;OAEG;IACH,OAAO,CAAC,EAAE,MAAM,CAAC;IACjB;;OAEG;IACH,QAAQ,CAAC,EAAE,CAAC,KAAK,EAAE,SAAS,KAAK,MAAM,EAAE,CAAC;IAC1C;;;;;OAKG;IACH,cAAc,CAAC,EAAE,kBAAkB,CAAC;IACpC;;;OAGG;IACH,MAAM,CAAC,EAAE,OAAO,CAAC;CACjB;AAED;;;;;;;;;;;;;GAaG;AAEH,eAAO,MAAM,2BAA2B;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CAwiBI,CAAC;AAE7C,MAAM,MAAM,eAAe,GAAG,MAAM,OAAO,2BAA2B,CAAC;AAEvE,eAAO,MAAM,sBAAsB,o/BAAgE,CAAC;AAEpG,eAAO,MAAM,8BAA8B,o/BAQ1B,CAAC"}
@@ -387,6 +387,12 @@ export declare const PIPELINE_DATA: {
387
387
  modality: "cv";
388
388
  color: "indigo";
389
389
  };
390
+ "video-text-to-text": {
391
+ name: string;
392
+ modality: "multimodal";
393
+ color: "blue";
394
+ hideInDatasets: true;
395
+ };
390
396
  other: {
391
397
  name: string;
392
398
  modality: "other";
@@ -397,7 +403,7 @@ export declare const PIPELINE_DATA: {
397
403
  };
398
404
  export type PipelineType = keyof typeof PIPELINE_DATA;
399
405
  export type WidgetType = PipelineType | "conversational";
400
- export declare const PIPELINE_TYPES: ("other" | "text-classification" | "token-classification" | "table-question-answering" | "question-answering" | "zero-shot-classification" | "translation" | "summarization" | "feature-extraction" | "text-generation" | "text2text-generation" | "fill-mask" | "sentence-similarity" | "text-to-speech" | "text-to-audio" | "automatic-speech-recognition" | "audio-to-audio" | "audio-classification" | "voice-activity-detection" | "depth-estimation" | "image-classification" | "object-detection" | "image-segmentation" | "text-to-image" | "image-to-text" | "image-to-image" | "image-to-video" | "unconditional-image-generation" | "video-classification" | "reinforcement-learning" | "robotics" | "tabular-classification" | "tabular-regression" | "tabular-to-text" | "table-to-text" | "multiple-choice" | "text-retrieval" | "time-series-forecasting" | "text-to-video" | "image-text-to-text" | "visual-question-answering" | "document-question-answering" | "zero-shot-image-classification" | "graph-ml" | "mask-generation" | "zero-shot-object-detection" | "text-to-3d" | "image-to-3d" | "image-feature-extraction")[];
406
+ export declare const PIPELINE_TYPES: ("other" | "text-classification" | "token-classification" | "table-question-answering" | "question-answering" | "zero-shot-classification" | "translation" | "summarization" | "feature-extraction" | "text-generation" | "text2text-generation" | "fill-mask" | "sentence-similarity" | "text-to-speech" | "text-to-audio" | "automatic-speech-recognition" | "audio-to-audio" | "audio-classification" | "voice-activity-detection" | "depth-estimation" | "image-classification" | "object-detection" | "image-segmentation" | "text-to-image" | "image-to-text" | "image-to-image" | "image-to-video" | "unconditional-image-generation" | "video-classification" | "reinforcement-learning" | "robotics" | "tabular-classification" | "tabular-regression" | "tabular-to-text" | "table-to-text" | "multiple-choice" | "text-retrieval" | "time-series-forecasting" | "text-to-video" | "image-text-to-text" | "visual-question-answering" | "document-question-answering" | "zero-shot-image-classification" | "graph-ml" | "mask-generation" | "zero-shot-object-detection" | "text-to-3d" | "image-to-3d" | "image-feature-extraction" | "video-text-to-text")[];
401
407
  export declare const SUBTASK_TYPES: string[];
402
- export declare const PIPELINE_TYPES_SET: Set<"other" | "text-classification" | "token-classification" | "table-question-answering" | "question-answering" | "zero-shot-classification" | "translation" | "summarization" | "feature-extraction" | "text-generation" | "text2text-generation" | "fill-mask" | "sentence-similarity" | "text-to-speech" | "text-to-audio" | "automatic-speech-recognition" | "audio-to-audio" | "audio-classification" | "voice-activity-detection" | "depth-estimation" | "image-classification" | "object-detection" | "image-segmentation" | "text-to-image" | "image-to-text" | "image-to-image" | "image-to-video" | "unconditional-image-generation" | "video-classification" | "reinforcement-learning" | "robotics" | "tabular-classification" | "tabular-regression" | "tabular-to-text" | "table-to-text" | "multiple-choice" | "text-retrieval" | "time-series-forecasting" | "text-to-video" | "image-text-to-text" | "visual-question-answering" | "document-question-answering" | "zero-shot-image-classification" | "graph-ml" | "mask-generation" | "zero-shot-object-detection" | "text-to-3d" | "image-to-3d" | "image-feature-extraction">;
408
+ export declare const PIPELINE_TYPES_SET: Set<"other" | "text-classification" | "token-classification" | "table-question-answering" | "question-answering" | "zero-shot-classification" | "translation" | "summarization" | "feature-extraction" | "text-generation" | "text2text-generation" | "fill-mask" | "sentence-similarity" | "text-to-speech" | "text-to-audio" | "automatic-speech-recognition" | "audio-to-audio" | "audio-classification" | "voice-activity-detection" | "depth-estimation" | "image-classification" | "object-detection" | "image-segmentation" | "text-to-image" | "image-to-text" | "image-to-image" | "image-to-video" | "unconditional-image-generation" | "video-classification" | "reinforcement-learning" | "robotics" | "tabular-classification" | "tabular-regression" | "tabular-to-text" | "table-to-text" | "multiple-choice" | "text-retrieval" | "time-series-forecasting" | "text-to-video" | "image-text-to-text" | "visual-question-answering" | "document-question-answering" | "zero-shot-image-classification" | "graph-ml" | "mask-generation" | "zero-shot-object-detection" | "text-to-3d" | "image-to-3d" | "image-feature-extraction" | "video-text-to-text">;
403
409
  //# sourceMappingURL=pipelines.d.ts.map
@@ -1 +1 @@
1
- {"version":3,"file":"pipelines.d.ts","sourceRoot":"","sources":["../../src/pipelines.ts"],"names":[],"mappings":"AAAA,eAAO,MAAM,UAAU,yEAA0E,CAAC;AAElG,MAAM,MAAM,QAAQ,GAAG,CAAC,OAAO,UAAU,CAAC,CAAC,MAAM,CAAC,CAAC;AAEnD,eAAO,MAAM,eAAe;;;;;;;;CAQQ,CAAC;AAErC;;;;;;GAMG;AACH,MAAM,WAAW,OAAO;IACvB;;OAEG;IACH,IAAI,EAAE,MAAM,CAAC;IACb;;OAEG;IACH,IAAI,EAAE,MAAM,CAAC;CACb;AAED;;;;;GAKG;AACH,MAAM,WAAW,YAAY;IAC5B;;OAEG;IACH,IAAI,EAAE,MAAM,CAAC;IACb,QAAQ,CAAC,EAAE,OAAO,EAAE,CAAC;IACrB,QAAQ,EAAE,QAAQ,CAAC;IACnB;;OAEG;IACH,KAAK,EAAE,MAAM,GAAG,OAAO,GAAG,QAAQ,GAAG,QAAQ,GAAG,KAAK,GAAG,QAAQ,CAAC;IACjE;;OAEG;IACH,YAAY,CAAC,EAAE,OAAO,CAAC;IACvB;;OAEG;IACH,cAAc,CAAC,EAAE,OAAO,CAAC;CACzB;AAcD,eAAO,MAAM,aAAa;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CA8kBc,CAAC;AAEzC,MAAM,MAAM,YAAY,GAAG,MAAM,OAAO,aAAa,CAAC;AAEtD,MAAM,MAAM,UAAU,GAAG,YAAY,GAAG,gBAAgB,CAAC;AAEzD,eAAO,MAAM,cAAc,mlCAA+C,CAAC;AAE3E,eAAO,MAAM,aAAa,UAEN,CAAC;AAErB,eAAO,MAAM,kBAAkB,olCAA0B,CAAC"}
1
+ {"version":3,"file":"pipelines.d.ts","sourceRoot":"","sources":["../../src/pipelines.ts"],"names":[],"mappings":"AAAA,eAAO,MAAM,UAAU,yEAA0E,CAAC;AAElG,MAAM,MAAM,QAAQ,GAAG,CAAC,OAAO,UAAU,CAAC,CAAC,MAAM,CAAC,CAAC;AAEnD,eAAO,MAAM,eAAe;;;;;;;;CAQQ,CAAC;AAErC;;;;;;GAMG;AACH,MAAM,WAAW,OAAO;IACvB;;OAEG;IACH,IAAI,EAAE,MAAM,CAAC;IACb;;OAEG;IACH,IAAI,EAAE,MAAM,CAAC;CACb;AAED;;;;;GAKG;AACH,MAAM,WAAW,YAAY;IAC5B;;OAEG;IACH,IAAI,EAAE,MAAM,CAAC;IACb,QAAQ,CAAC,EAAE,OAAO,EAAE,CAAC;IACrB,QAAQ,EAAE,QAAQ,CAAC;IACnB;;OAEG;IACH,KAAK,EAAE,MAAM,GAAG,OAAO,GAAG,QAAQ,GAAG,QAAQ,GAAG,KAAK,GAAG,QAAQ,CAAC;IACjE;;OAEG;IACH,YAAY,CAAC,EAAE,OAAO,CAAC;IACvB;;OAEG;IACH,cAAc,CAAC,EAAE,OAAO,CAAC;CACzB;AAcD,eAAO,MAAM,aAAa;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CAolBc,CAAC;AAEzC,MAAM,MAAM,YAAY,GAAG,MAAM,OAAO,aAAa,CAAC;AAEtD,MAAM,MAAM,UAAU,GAAG,YAAY,GAAG,gBAAgB,CAAC;AAEzD,eAAO,MAAM,cAAc,0mCAA+C,CAAC;AAE3E,eAAO,MAAM,aAAa,UAEN,CAAC;AAErB,eAAO,MAAM,kBAAkB,2mCAA0B,CAAC"}
@@ -1 +1 @@
1
- {"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/feature-extraction/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,IAAI,CAAC;AAEzC,QAAA,MAAM,QAAQ,EAAE,cAgDf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
1
+ {"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/feature-extraction/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,IAAI,CAAC;AAEzC,QAAA,MAAM,QAAQ,EAAE,cAoDf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
@@ -1 +1 @@
1
- {"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/image-feature-extraction/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,IAAI,CAAC;AAEzC,QAAA,MAAM,QAAQ,EAAE,cAkDf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
1
+ {"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/image-feature-extraction/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,IAAI,CAAC;AAEzC,QAAA,MAAM,QAAQ,EAAE,cAsDf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
@@ -1 +1 @@
1
- {"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../../../src/tasks/index.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,YAAY,EAAE,MAAM,cAAc,CAAC;AA0CjD,mBAAmB,kCAAkC,CAAC;AACtD,mBAAmB,0CAA0C,CAAC;AAC9D,YAAY,EACX,mBAAmB,EACnB,0BAA0B,EAC1B,oBAAoB,EACpB,4BAA4B,EAC5B,2BAA2B,EAC3B,0BAA0B,EAC1B,gCAAgC,EAChC,+BAA+B,GAC/B,MAAM,6BAA6B,CAAC;AACrC,mBAAmB,yCAAyC,CAAC;AAC7D,mBAAmB,gCAAgC,CAAC;AACpD,mBAAmB,uBAAuB,CAAC;AAC3C,YAAY,EACX,wBAAwB,EACxB,yBAAyB,EACzB,gCAAgC,EAChC,6BAA6B,GAC7B,MAAM,kCAAkC,CAAC;AAC1C,mBAAmB,4BAA4B,CAAC;AAChD,YAAY,EAAE,gBAAgB,EAAE,iBAAiB,EAAE,qBAAqB,EAAE,MAAM,2BAA2B,CAAC;AAC5G,mBAAmB,gCAAgC,CAAC;AACpD,mBAAmB,8BAA8B,CAAC;AAClD,mBAAmB,8BAA8B,CAAC;AAClD,mBAAmB,gCAAgC,CAAC;AACpD,mBAAmB,iCAAiC,CAAC;AACrD,mBAAmB,2BAA2B,CAAC;AAC/C,mBAAmB,sCAAsC,CAAC;AAC1D,YAAY,EAAE,gBAAgB,EAAE,iBAAiB,EAAE,qBAAqB,EAAE,MAAM,2BAA2B,CAAC;AAC5G,YAAY,EAAE,qBAAqB,EAAE,iBAAiB,EAAE,kBAAkB,EAAE,MAAM,4BAA4B,CAAC;AAC/G,mBAAmB,kCAAkC,CAAC;AACtD,YAAY,EACX,6BAA6B,EAC7B,qCAAqC,EACrC,gBAAgB,EAChB,iBAAiB,GACjB,MAAM,yBAAyB,CAAC;AACjC,YAAY,EACX,6BAA6B,EAC7B,uBAAuB,EACvB,wBAAwB,EACxB,+BAA+B,EAC/B,4BAA4B,GAC5B,MAAM,iCAAiC,CAAC;AACzC,YAAY,EACX,gCAAgC,EAChC,gCAAgC,EAChC,mBAAmB,EACnB,oBAAoB,EACpB,2BAA2B,EAC3B,qCAAqC,EACrC,kCAAkC,EAClC,yBAAyB,EACzB,uCAAuC,EACvC,0BAA0B,GAC1B,MAAM,6BAA6B,CAAC;AACrC,mBAAmB,kCAAkC,CAAC;AACtD,mBAAmB,uCAAuC,CAAC;AAC3D,mBAAmB,sCAAsC,CAAC;AAC1D,mBAAmB,4CAA4C,CAAC;AAChE,YAAY,EACX,WAAW,EACX,4BAA4B,EAC5B,gCAAgC,EAChC,6BAA6B,EAC7B,oCAAoC,GACpC,MAAM,wCAAwC,CAAC;AAEhD,OAAO,KAAK,EAAE,eAAe,EAAE,MAAM,oBAAoB,CAAC;AAE1D;;GAEG;AACH,eAAO,MAAM,qBAAqB,EAAE,MAAM,CAAC,YAAY,EAAE,eAAe,EAAE,CA0DzE,CAAC;AAoBF,eAAO,MAAM,UAAU,EAAE,MAAM,CAAC,YAAY,EAAE,QAAQ,GAAG,SAAS,CAkDxD,CAAC;AAEX,MAAM,WAAW,WAAW;IAC3B,WAAW,EAAE,MAAM,CAAC;IACpB,EAAE,EAAE,MAAM,CAAC;CACX;AAED,MAAM,MAAM,aAAa,GACtB;IACA,QAAQ,EAAE,MAAM,CAAC;IACjB,IAAI,EAAE,OAAO,CAAC;CACb,GACD;IACA,IAAI,EAAE,KAAK,CAAC;QACX,KAAK,EAAE,MAAM,CAAC;QACd,KAAK,EAAE,MAAM,CAAC;KACd,CAAC,CAAC;IACH,IAAI,EAAE,OAAO,CAAC;CACb,GACD;IACA,QAAQ,EAAE,MAAM,CAAC;IACjB,IAAI,EAAE,KAAK,CAAC;CACX,GACD;IACA,KAAK,EAAE,MAAM,EAAE,EAAE,CAAC;IAClB,IAAI,EAAE,SAAS,CAAC;CACf,GACD;IACA,OAAO,EAAE,MAAM,CAAC;IAChB,KAAK,EAAE,MAAM,CAAC;IACd,IAAI,EAAE,MAAM,CAAC;CACZ,GACD;IACA,IAAI,EAAE,MAAM,CAAC;IACb,MAAM,EAAE,KAAK,CAAC;QACb,GAAG,EAAE,MAAM,CAAC;QACZ,KAAK,EAAE,MAAM,CAAC;QACd,IAAI,EAAE,MAAM,CAAC;KACb,CAAC,CAAC;IACH,IAAI,EAAE,kBAAkB,CAAC;CACxB,CAAC;AAEL,MAAM,WAAW,QAAQ;IACxB,MAAM,EAAE,aAAa,EAAE,CAAC;IACxB,OAAO,EAAE,aAAa,EAAE,CAAC;CACzB;AAED,MAAM,WAAW,QAAQ;IACxB,QAAQ,EAAE,WAAW,EAAE,CAAC;IACxB,IAAI,EAAE,QAAQ,CAAC;IACf,EAAE,EAAE,YAAY,CAAC;IACjB,WAAW,CAAC,EAAE,YAAY,CAAC;IAC3B,aAAa,CAAC,EAAE,OAAO,CAAC;IACxB,KAAK,EAAE,MAAM,CAAC;IACd,SAAS,EAAE,eAAe,EAAE,CAAC;IAC7B,OAAO,EAAE,WAAW,EAAE,CAAC;IACvB,MAAM,EAAE,WAAW,EAAE,CAAC;IACtB,MAAM,EAAE,WAAW,EAAE,CAAC;IACtB,OAAO,EAAE,MAAM,CAAC;IAChB,YAAY,EAAE,MAAM,EAAE,CAAC;IACvB,SAAS,CAAC,EAAE,MAAM,CAAC;CACnB;AAED,MAAM,MAAM,cAAc,GAAG,IAAI,CAAC,QAAQ,EAAE,IAAI,GAAG,OAAO,GAAG,WAAW,CAAC,CAAC"}
1
+ {"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../../../src/tasks/index.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,YAAY,EAAE,MAAM,cAAc,CAAC;AA0CjD,mBAAmB,kCAAkC,CAAC;AACtD,mBAAmB,0CAA0C,CAAC;AAC9D,YAAY,EACX,mBAAmB,EACnB,0BAA0B,EAC1B,oBAAoB,EACpB,4BAA4B,EAC5B,2BAA2B,EAC3B,0BAA0B,EAC1B,gCAAgC,EAChC,+BAA+B,GAC/B,MAAM,6BAA6B,CAAC;AACrC,mBAAmB,yCAAyC,CAAC;AAC7D,mBAAmB,gCAAgC,CAAC;AACpD,mBAAmB,uBAAuB,CAAC;AAC3C,YAAY,EACX,wBAAwB,EACxB,yBAAyB,EACzB,gCAAgC,EAChC,6BAA6B,GAC7B,MAAM,kCAAkC,CAAC;AAC1C,mBAAmB,4BAA4B,CAAC;AAChD,YAAY,EAAE,gBAAgB,EAAE,iBAAiB,EAAE,qBAAqB,EAAE,MAAM,2BAA2B,CAAC;AAC5G,mBAAmB,gCAAgC,CAAC;AACpD,mBAAmB,8BAA8B,CAAC;AAClD,mBAAmB,8BAA8B,CAAC;AAClD,mBAAmB,gCAAgC,CAAC;AACpD,mBAAmB,iCAAiC,CAAC;AACrD,mBAAmB,2BAA2B,CAAC;AAC/C,mBAAmB,sCAAsC,CAAC;AAC1D,YAAY,EAAE,gBAAgB,EAAE,iBAAiB,EAAE,qBAAqB,EAAE,MAAM,2BAA2B,CAAC;AAC5G,YAAY,EAAE,qBAAqB,EAAE,iBAAiB,EAAE,kBAAkB,EAAE,MAAM,4BAA4B,CAAC;AAC/G,mBAAmB,kCAAkC,CAAC;AACtD,YAAY,EACX,6BAA6B,EAC7B,qCAAqC,EACrC,gBAAgB,EAChB,iBAAiB,GACjB,MAAM,yBAAyB,CAAC;AACjC,YAAY,EACX,6BAA6B,EAC7B,uBAAuB,EACvB,wBAAwB,EACxB,+BAA+B,EAC/B,4BAA4B,GAC5B,MAAM,iCAAiC,CAAC;AACzC,YAAY,EACX,gCAAgC,EAChC,gCAAgC,EAChC,mBAAmB,EACnB,oBAAoB,EACpB,2BAA2B,EAC3B,qCAAqC,EACrC,kCAAkC,EAClC,yBAAyB,EACzB,uCAAuC,EACvC,0BAA0B,GAC1B,MAAM,6BAA6B,CAAC;AACrC,mBAAmB,kCAAkC,CAAC;AACtD,mBAAmB,uCAAuC,CAAC;AAC3D,mBAAmB,sCAAsC,CAAC;AAC1D,mBAAmB,4CAA4C,CAAC;AAChE,YAAY,EACX,WAAW,EACX,4BAA4B,EAC5B,gCAAgC,EAChC,6BAA6B,EAC7B,oCAAoC,GACpC,MAAM,wCAAwC,CAAC;AAEhD,OAAO,KAAK,EAAE,eAAe,EAAE,MAAM,oBAAoB,CAAC;AAE1D;;GAEG;AACH,eAAO,MAAM,qBAAqB,EAAE,MAAM,CAAC,YAAY,EAAE,eAAe,EAAE,CA2DzE,CAAC;AAoBF,eAAO,MAAM,UAAU,EAAE,MAAM,CAAC,YAAY,EAAE,QAAQ,GAAG,SAAS,CAmDxD,CAAC;AAEX,MAAM,WAAW,WAAW;IAC3B,WAAW,EAAE,MAAM,CAAC;IACpB,EAAE,EAAE,MAAM,CAAC;CACX;AAED,MAAM,MAAM,aAAa,GACtB;IACA,QAAQ,EAAE,MAAM,CAAC;IACjB,IAAI,EAAE,OAAO,CAAC;CACb,GACD;IACA,IAAI,EAAE,KAAK,CAAC;QACX,KAAK,EAAE,MAAM,CAAC;QACd,KAAK,EAAE,MAAM,CAAC;KACd,CAAC,CAAC;IACH,IAAI,EAAE,OAAO,CAAC;CACb,GACD;IACA,QAAQ,EAAE,MAAM,CAAC;IACjB,IAAI,EAAE,KAAK,CAAC;CACX,GACD;IACA,KAAK,EAAE,MAAM,EAAE,EAAE,CAAC;IAClB,IAAI,EAAE,SAAS,CAAC;CACf,GACD;IACA,OAAO,EAAE,MAAM,CAAC;IAChB,KAAK,EAAE,MAAM,CAAC;IACd,IAAI,EAAE,MAAM,CAAC;CACZ,GACD;IACA,IAAI,EAAE,MAAM,CAAC;IACb,MAAM,EAAE,KAAK,CAAC;QACb,GAAG,EAAE,MAAM,CAAC;QACZ,KAAK,EAAE,MAAM,CAAC;QACd,IAAI,EAAE,MAAM,CAAC;KACb,CAAC,CAAC;IACH,IAAI,EAAE,kBAAkB,CAAC;CACxB,CAAC;AAEL,MAAM,WAAW,QAAQ;IACxB,MAAM,EAAE,aAAa,EAAE,CAAC;IACxB,OAAO,EAAE,aAAa,EAAE,CAAC;CACzB;AAED,MAAM,WAAW,QAAQ;IACxB,QAAQ,EAAE,WAAW,EAAE,CAAC;IACxB,IAAI,EAAE,QAAQ,CAAC;IACf,EAAE,EAAE,YAAY,CAAC;IACjB,WAAW,CAAC,EAAE,YAAY,CAAC;IAC3B,aAAa,CAAC,EAAE,OAAO,CAAC;IACxB,KAAK,EAAE,MAAM,CAAC;IACd,SAAS,EAAE,eAAe,EAAE,CAAC;IAC7B,OAAO,EAAE,WAAW,EAAE,CAAC;IACvB,MAAM,EAAE,WAAW,EAAE,CAAC;IACtB,MAAM,EAAE,WAAW,EAAE,CAAC;IACtB,OAAO,EAAE,MAAM,CAAC;IAChB,YAAY,EAAE,MAAM,EAAE,CAAC;IACvB,SAAS,CAAC,EAAE,MAAM,CAAC;CACnB;AAED,MAAM,MAAM,cAAc,GAAG,IAAI,CAAC,QAAQ,EAAE,IAAI,GAAG,OAAO,GAAG,WAAW,CAAC,CAAC"}
@@ -1 +1 @@
1
- {"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/text-generation/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,IAAI,CAAC;AAEzC,QAAA,MAAM,QAAQ,EAAE,cAyGf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
1
+ {"version":3,"file":"data.d.ts","sourceRoot":"","sources":["../../../../src/tasks/text-generation/data.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,IAAI,CAAC;AAEzC,QAAA,MAAM,QAAQ,EAAE,cA6Gf,CAAC;AAEF,eAAe,QAAQ,CAAC"}
package/package.json CHANGED
@@ -1,7 +1,7 @@
1
1
  {
2
2
  "name": "@huggingface/tasks",
3
3
  "packageManager": "pnpm@8.10.5",
4
- "version": "0.11.7",
4
+ "version": "0.11.8",
5
5
  "description": "List of ML tasks for huggingface.co/tasks",
6
6
  "repository": "https://github.com/huggingface/huggingface.js.git",
7
7
  "publishConfig": {
@@ -42,6 +42,9 @@
42
42
  "format": "prettier --write .",
43
43
  "format:check": "prettier --check .",
44
44
  "build": "tsup src/index.ts --format cjs,esm --clean && tsc --emitDeclarationOnly --declaration",
45
+ "watch:export": "tsup src/index.ts --format cjs,esm --watch",
46
+ "watch:types": "tsc --emitDeclarationOnly --declaration --watch",
47
+ "watch": "npm-run-all --parallel watch:export watch:types",
45
48
  "check": "tsc",
46
49
  "inference-codegen": "tsx scripts/inference-codegen.ts && prettier --write src/tasks/*/inference.ts",
47
50
  "inference-tgi-import": "tsx scripts/inference-tgi-import.ts && prettier --write src/tasks/text-generation/spec/*.json && prettier --write src/tasks/chat-completion/spec/*.json",
package/src/local-apps.ts CHANGED
@@ -201,7 +201,6 @@ export const LOCAL_APPS = {
201
201
  docsUrl: "https://diffusionbee.com",
202
202
  mainTask: "text-to-image",
203
203
  macOSOnly: true,
204
- comingSoon: true,
205
204
  displayOnModelPage: (model) => model.library_name === "diffusers" && model.pipeline_tag === "text-to-image",
206
205
  deeplink: (model) => new URL(`diffusionbee://open_from_hf?model=${model.id}`),
207
206
  },
@@ -261,6 +261,24 @@ backbone = keras_nlp.models.Backbone.from_preset("hf://${model.id}")
261
261
  `,
262
262
  ];
263
263
 
264
+ export const llama_cpp_python = (model: ModelData): string[] => [
265
+ `from llama_cpp import Llama
266
+
267
+ llm = Llama.from_pretrained(
268
+ repo_id="${model.id}",
269
+ filename="{{GGUF_FILE}}",
270
+ )
271
+
272
+ llm.create_chat_completion(
273
+ messages = [
274
+ {
275
+ "role": "user",
276
+ "content": "What is the capital of France?"
277
+ }
278
+ ]
279
+ )`,
280
+ ];
281
+
264
282
  export const tf_keras = (model: ModelData): string[] => [
265
283
  `# Note: 'keras<3.x' or 'tf_keras' must be installed (legacy)
266
284
  # See https://github.com/keras-team/tf-keras for more details.
@@ -314,6 +314,12 @@ export const MODEL_LIBRARIES_UI_ELEMENTS = {
314
314
  filter: false,
315
315
  countDownloads: `path:"liveportrait/landmark.onnx"`,
316
316
  },
317
+ "llama-cpp-python": {
318
+ prettyLabel: "llama-cpp-python",
319
+ repoName: "llama-cpp-python",
320
+ repoUrl: "https://github.com/abetlen/llama-cpp-python",
321
+ snippets: snippets.llama_cpp_python,
322
+ },
317
323
  mindspore: {
318
324
  prettyLabel: "MindSpore",
319
325
  repoName: "mindspore",
@@ -436,6 +442,13 @@ export const MODEL_LIBRARIES_UI_ELEMENTS = {
436
442
  snippets: snippets.saelens,
437
443
  filter: false,
438
444
  },
445
+ sam2: {
446
+ prettyLabel: "sam2",
447
+ repoName: "sam2",
448
+ repoUrl: "https://github.com/facebookresearch/segment-anything-2",
449
+ filter: false,
450
+ countDownloads: `path_extension:"pt"`,
451
+ },
439
452
  "sample-factory": {
440
453
  prettyLabel: "sample-factory",
441
454
  repoName: "sample-factory",
package/src/pipelines.ts CHANGED
@@ -652,6 +652,12 @@ export const PIPELINE_DATA = {
652
652
  modality: "cv",
653
653
  color: "indigo",
654
654
  },
655
+ "video-text-to-text": {
656
+ name: "Video-Text-to-Text",
657
+ modality: "multimodal",
658
+ color: "blue",
659
+ hideInDatasets: true,
660
+ },
655
661
  other: {
656
662
  name: "Other",
657
663
  modality: "other",
@@ -42,9 +42,13 @@ const taskData: TaskDataCustom = {
42
42
  ],
43
43
  spaces: [
44
44
  {
45
- description: "A leaderboard to rank best feature extraction models..",
45
+ description: "A leaderboard to rank text feature extraction models based on a benchmark.",
46
46
  id: "mteb/leaderboard",
47
47
  },
48
+ {
49
+ description: "A leaderboard to rank best feature extraction models based on human feedback.",
50
+ id: "mteb/arena",
51
+ },
48
52
  ],
49
53
  summary: "Feature extraction is the task of extracting features learnt in a model.",
50
54
  widgetModels: ["facebook/bart-base"],
@@ -36,16 +36,20 @@ const taskData: TaskDataCustom = {
36
36
  },
37
37
  {
38
38
  description: "A strong image feature extraction model.",
39
- id: "google/vit-base-patch16-224-in21k",
39
+ id: "nvidia/MambaVision-T-1K",
40
40
  },
41
41
  {
42
- description: "A robust image feature extraction models.",
42
+ description: "A robust image feature extraction model.",
43
43
  id: "facebook/dino-vitb16",
44
44
  },
45
45
  {
46
- description: "Strong image-text-to-text model made for information retrieval from documents.",
46
+ description: "Strong image feature extraction model made for information retrieval from documents.",
47
47
  id: "vidore/colpali",
48
48
  },
49
+ {
50
+ description: "Strong image feature extraction model that can be used on images and documents.",
51
+ id: "OpenGVLab/InternViT-6B-448px-V1-2",
52
+ },
49
53
  ],
50
54
  spaces: [],
51
55
  summary: "Image feature extraction is the task of extracting features learnt in a computer vision model.",
@@ -48,16 +48,16 @@ const taskData: TaskDataCustom = {
48
48
  id: "facebook/detr-resnet-50-panoptic",
49
49
  },
50
50
  {
51
- description: "Semantic segmentation model trained on ADE20k benchmark dataset.",
52
- id: "microsoft/beit-large-finetuned-ade-640-640",
51
+ description: "Background removal model.",
52
+ id: "briaai/RMBG-1.4",
53
53
  },
54
54
  {
55
55
  description: "Semantic segmentation model trained on ADE20k benchmark dataset with 512x512 resolution.",
56
56
  id: "nvidia/segformer-b0-finetuned-ade-512-512",
57
57
  },
58
58
  {
59
- description: "Semantic segmentation model trained Cityscapes dataset.",
60
- id: "facebook/mask2former-swin-large-cityscapes-semantic",
59
+ description: "A multipurpose image segmentation model for high resolution images.",
60
+ id: "ZhengPeng7/BiRefNet",
61
61
  },
62
62
  {
63
63
  description: "Panoptic segmentation model trained COCO (common objects) dataset.",
@@ -72,3 +72,5 @@ print(processor.decode(output[0], skip_special_tokens=True))
72
72
  - [Vision Language Models Explained](https://huggingface.co/blog/vlms)
73
73
  - [Open-source Multimodality and How to Achieve it using Hugging Face](https://www.youtube.com/watch?v=IoGaGfU1CIg&t=601s)
74
74
  - [Introducing Idefics2: A Powerful 8B Vision-Language Model for the community](https://huggingface.co/blog/idefics2)
75
+ - [Image-text-to-text task guide](https://huggingface.co/tasks/image-text-to-text)
76
+ - [Preference Optimization for Vision Language Models with TRL](https://huggingface.co/blog/dpo_vlm)
@@ -88,7 +88,7 @@ const taskData: TaskDataCustom = {
88
88
  summary:
89
89
  "Image-text-to-text models take in an image and text prompt and output text. These models are also called vision-language models, or VLMs. The difference from image-to-text models is that these models take an additional text input, not restricting the model to certain use cases like image captioning, and may also be trained to accept a conversation as input.",
90
90
  widgetModels: ["microsoft/kosmos-2-patch14-224"],
91
- youtubeId: "",
91
+ youtubeId: "IoGaGfU1CIg",
92
92
  };
93
93
 
94
94
  export default taskData;
@@ -45,8 +45,8 @@ const taskData: TaskDataCustom = {
45
45
  ],
46
46
  models: [
47
47
  {
48
- description: "A model that enhances images captured in low light conditions.",
49
- id: "keras-io/low-light-image-enhancement",
48
+ description: "An image-to-image model to improve image resolution.",
49
+ id: "fal/AuraSR-v2",
50
50
  },
51
51
  {
52
52
  description: "A model that increases the resolution of an image.",
@@ -166,6 +166,7 @@ export const TASKS_MODEL_LIBRARIES: Record<PipelineType, ModelLibraryKey[]> = {
166
166
  ],
167
167
  translation: ["transformers", "transformers.js"],
168
168
  "unconditional-image-generation": ["diffusers"],
169
+ "video-text-to-text": ["transformers"],
169
170
  "visual-question-answering": ["transformers", "transformers.js"],
170
171
  "voice-activity-detection": [],
171
172
  "zero-shot-classification": ["transformers", "transformers.js"],
@@ -236,6 +237,7 @@ export const TASKS_DATA: Record<PipelineType, TaskData | undefined> = {
236
237
  "token-classification": getData("token-classification", tokenClassification),
237
238
  translation: getData("translation", translation),
238
239
  "unconditional-image-generation": getData("unconditional-image-generation", unconditionalImageGeneration),
240
+ "video-text-to-text": getData("video-text-to-text", placeholder),
239
241
  "visual-question-answering": getData("visual-question-answering", visualQuestionAnswering),
240
242
  "voice-activity-detection": undefined,
241
243
  "zero-shot-classification": getData("zero-shot-classification", zeroShotClassification),
@@ -24,14 +24,14 @@ const taskData: TaskDataCustom = {
24
24
  },
25
25
  {
26
26
  description: "Very strong mask generation model.",
27
- id: "facebook/sam-vit-huge",
27
+ id: "facebook/sam2-hiera-large",
28
28
  },
29
29
  ],
30
30
  spaces: [
31
31
  {
32
32
  description:
33
- "An application that combines a mask generation model with an image embedding model for open-vocabulary image segmentation.",
34
- id: "SkalskiP/SAM_and_MetaCLIP",
33
+ "An application that combines a mask generation model with a zero-shot object detection model for text-guided image segmentation.",
34
+ id: "merve/OWLSAM2",
35
35
  },
36
36
  {
37
37
  description: "An application that compares the performance of a large and a small mask generation model.",
@@ -39,7 +39,7 @@ const taskData: TaskDataCustom = {
39
39
  },
40
40
  {
41
41
  description: "An application based on an improved mask generation model.",
42
- id: "linfanluntan/Grounded-SAM",
42
+ id: "SkalskiP/segment-anything-model-2",
43
43
  },
44
44
  {
45
45
  description: "An application to remove objects from videos using mask generation models.",
@@ -19,6 +19,10 @@ const taskData: TaskDataCustom = {
19
19
  description: "An instruction dataset with preference ratings on responses.",
20
20
  id: "openbmb/UltraFeedback",
21
21
  },
22
+ {
23
+ description: "A large synthetic dataset for alignment of text generation models.",
24
+ id: "argilla/magpie-ultra-v0.1",
25
+ },
22
26
  ],
23
27
  demo: {
24
28
  inputs: [
@@ -51,32 +55,32 @@ const taskData: TaskDataCustom = {
51
55
  ],
52
56
  models: [
53
57
  {
54
- description: "A large language model trained for text generation.",
55
- id: "bigscience/bloom-560m",
58
+ description: "A text-generation model trained to follow instructions.",
59
+ id: "google/gemma-2-2b-it",
56
60
  },
57
61
  {
58
- description: "A large code generation model that can generate code in 80+ languages.",
62
+ description: "A code generation model that can generate code in 80+ languages.",
59
63
  id: "bigcode/starcoder",
60
64
  },
61
65
  {
62
- description: "A very powerful text generation model.",
63
- id: "mistralai/Mixtral-8x7B-Instruct-v0.1",
66
+ description: "Very powerful text generation model trained to follow instructions.",
67
+ id: "meta-llama/Meta-Llama-3.1-8B-Instruct",
64
68
  },
65
69
  {
66
70
  description: "Small yet powerful text generation model.",
67
- id: "microsoft/phi-2",
71
+ id: "microsoft/Phi-3-mini-4k-instruct",
68
72
  },
69
73
  {
70
- description: "A very powerful model that can chat, do mathematical reasoning and write code.",
71
- id: "openchat/openchat-3.5-0106",
74
+ description: "A very powerful model that can solve mathematical problems.",
75
+ id: "AI-MO/NuminaMath-7B-TIR",
72
76
  },
73
77
  {
74
- description: "Very strong yet small assistant model.",
75
- id: "HuggingFaceH4/zephyr-7b-beta",
78
+ description: "Strong coding assistant model.",
79
+ id: "HuggingFaceH4/starchat2-15b-v0.1",
76
80
  },
77
81
  {
78
82
  description: "Very strong open-source large language model.",
79
- id: "meta-llama/Llama-2-70b-hf",
83
+ id: "mistralai/Mistral-Nemo-Instruct-2407",
80
84
  },
81
85
  ],
82
86
  spaces: [
@@ -104,7 +108,7 @@ const taskData: TaskDataCustom = {
104
108
  summary:
105
109
  "Generating text is the task of generating new text given another text. These models can, for example, fill in incomplete text or paraphrase.",
106
110
  widgetModels: ["HuggingFaceH4/zephyr-7b-beta"],
107
- youtubeId: "Vpjb1lu0MDk",
111
+ youtubeId: "e9gNEAlsOvU",
108
112
  };
109
113
 
110
114
  export default taskData;
@@ -46,15 +46,15 @@ const taskData: TaskDataCustom = {
46
46
  models: [
47
47
  {
48
48
  description: "One of the most powerful image generation models that can generate realistic outputs.",
49
- id: "stabilityai/stable-diffusion-xl-base-1.0",
49
+ id: "black-forest-labs/FLUX.1-dev",
50
50
  },
51
51
  {
52
52
  description: "A powerful yet fast image generation model.",
53
53
  id: "latent-consistency/lcm-lora-sdxl",
54
54
  },
55
55
  {
56
- description: "A very fast text-to-image model.",
57
- id: "ByteDance/SDXL-Lightning",
56
+ description: "Text-to-image model for photorealistic generation.",
57
+ id: "Kwai-Kolors/Kolors",
58
58
  },
59
59
  {
60
60
  description: "A powerful text-to-image model.",