comfyui-workflow-templates 0.1.94__py3-none-any.whl → 0.1.95__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of comfyui-workflow-templates might be problematic. Click here for more details.
- comfyui_workflow_templates/templates/3d_hunyuan3d_image_to_model.json +26 -26
- comfyui_workflow_templates/templates/3d_hunyuan3d_multiview_to_model.json +35 -35
- comfyui_workflow_templates/templates/3d_hunyuan3d_multiview_to_model_turbo.json +37 -37
- comfyui_workflow_templates/templates/api_bfl_flux_1_kontext_max_image.json +57 -81
- comfyui_workflow_templates/templates/api_bfl_flux_1_kontext_multiple_images_input.json +60 -84
- comfyui_workflow_templates/templates/api_bfl_flux_1_kontext_pro_image.json +57 -81
- comfyui_workflow_templates/templates/api_bfl_flux_pro_t2i.json +28 -26
- comfyui_workflow_templates/templates/api_bytedance_flf2v.json +11 -11
- comfyui_workflow_templates/templates/api_bytedance_image_to_video.json +34 -34
- comfyui_workflow_templates/templates/api_bytedance_text_to_video.json +39 -40
- comfyui_workflow_templates/templates/api_google_gemini.json +6 -7
- comfyui_workflow_templates/templates/api_hailuo_minimax_i2v.json +12 -3
- comfyui_workflow_templates/templates/api_hailuo_minimax_t2v.json +28 -28
- comfyui_workflow_templates/templates/api_hailuo_minimax_video.json +30 -30
- comfyui_workflow_templates/templates/api_ideogram_v3_t2i.json +32 -18
- comfyui_workflow_templates/templates/api_kling_effects.json +28 -26
- comfyui_workflow_templates/templates/api_kling_flf.json +32 -30
- comfyui_workflow_templates/templates/api_kling_i2v.json +34 -34
- comfyui_workflow_templates/templates/api_luma_i2v.json +93 -110
- comfyui_workflow_templates/templates/api_luma_photon_i2i.json +59 -50
- comfyui_workflow_templates/templates/api_luma_photon_style_ref.json +131 -124
- comfyui_workflow_templates/templates/api_luma_t2v.json +59 -50
- comfyui_workflow_templates/templates/api_openai_dall_e_2_t2i.json +8 -29
- comfyui_workflow_templates/templates/api_openai_image_1_i2i.json +12 -33
- comfyui_workflow_templates/templates/api_openai_image_1_inpaint.json +30 -51
- comfyui_workflow_templates/templates/api_openai_image_1_multi_inputs.json +10 -33
- comfyui_workflow_templates/templates/api_openai_image_1_t2i.json +42 -63
- comfyui_workflow_templates/templates/api_openai_sora_video.json +37 -38
- comfyui_workflow_templates/templates/api_pika_i2v.json +27 -27
- comfyui_workflow_templates/templates/api_pika_scene.json +12 -3
- comfyui_workflow_templates/templates/api_pixverse_i2v.json +31 -36
- comfyui_workflow_templates/templates/api_pixverse_t2v.json +20 -16
- comfyui_workflow_templates/templates/api_pixverse_template_i2v.json +39 -35
- comfyui_workflow_templates/templates/api_recraft_image_gen_with_color_control.json +6 -6
- comfyui_workflow_templates/templates/api_recraft_image_gen_with_style_control.json +212 -199
- comfyui_workflow_templates/templates/api_recraft_vector_gen.json +78 -69
- comfyui_workflow_templates/templates/api_rodin_gen2.json +30 -30
- comfyui_workflow_templates/templates/api_rodin_image_to_model.json +55 -55
- comfyui_workflow_templates/templates/api_rodin_multiview_to_model.json +188 -132
- comfyui_workflow_templates/templates/api_runway_first_last_frame.json +4 -4
- comfyui_workflow_templates/templates/api_runway_gen3a_turbo_image_to_video.json +30 -31
- comfyui_workflow_templates/templates/api_runway_gen4_turo_image_to_video.json +29 -30
- comfyui_workflow_templates/templates/api_runway_reference_to_image.json +31 -32
- comfyui_workflow_templates/templates/api_runway_text_to_image.json +17 -17
- comfyui_workflow_templates/templates/api_stability_ai_audio_inpaint.json +18 -18
- comfyui_workflow_templates/templates/api_stability_ai_audio_to_audio.json +31 -31
- comfyui_workflow_templates/templates/api_stability_ai_i2i.json +34 -34
- comfyui_workflow_templates/templates/api_stability_ai_sd3.5_i2i.json +21 -19
- comfyui_workflow_templates/templates/api_stability_ai_sd3.5_t2i.json +35 -35
- comfyui_workflow_templates/templates/api_stability_ai_stable_image_ultra_t2i.json +11 -9
- comfyui_workflow_templates/templates/api_tripo_image_to_model.json +90 -92
- comfyui_workflow_templates/templates/api_tripo_multiview_to_model.json +241 -241
- comfyui_workflow_templates/templates/api_tripo_text_to_model.json +102 -102
- comfyui_workflow_templates/templates/api_veo2_i2v.json +31 -28
- comfyui_workflow_templates/templates/api_veo3.json +30 -30
- comfyui_workflow_templates/templates/api_vidu_text_to_video.json +2 -2
- comfyui_workflow_templates/templates/api_wan_image_to_video.json +41 -42
- comfyui_workflow_templates/templates/api_wan_text_to_image .json +140 -0
- comfyui_workflow_templates/templates/api_wan_text_to_video.json +38 -45
- comfyui_workflow_templates/templates/audio_ace_step_1_m2m_editing.json +84 -84
- comfyui_workflow_templates/templates/audio_ace_step_1_t2a_instrumentals.json +60 -60
- comfyui_workflow_templates/templates/audio_ace_step_1_t2a_song.json +60 -60
- comfyui_workflow_templates/templates/esrgan_example.json +24 -30
- comfyui_workflow_templates/templates/flux1_dev_uso_reference_image_gen.json +215 -210
- comfyui_workflow_templates/templates/flux1_krea_dev.json +3 -3
- comfyui_workflow_templates/templates/flux_kontext_dev_basic.json +151 -231
- comfyui_workflow_templates/templates/flux_redux_model_example.json +108 -120
- comfyui_workflow_templates/templates/flux_schnell_full_text_to_image.json +21 -29
- comfyui_workflow_templates/templates/hidream_e1_1.json +179 -209
- comfyui_workflow_templates/templates/hidream_e1_full.json +33 -39
- comfyui_workflow_templates/templates/hidream_i1_dev.json +15 -15
- comfyui_workflow_templates/templates/hidream_i1_fast.json +15 -15
- comfyui_workflow_templates/templates/hidream_i1_full.json +17 -16
- comfyui_workflow_templates/templates/hiresfix_esrgan_workflow.json +31 -37
- comfyui_workflow_templates/templates/hiresfix_latent_workflow.json +84 -88
- comfyui_workflow_templates/templates/image2image.json +30 -30
- comfyui_workflow_templates/templates/image_chroma1_radiance_text_to_image.json +60 -60
- comfyui_workflow_templates/templates/image_lotus_depth_v1_1.json +25 -31
- comfyui_workflow_templates/templates/image_netayume_lumina_t2i-1.webp +0 -0
- comfyui_workflow_templates/templates/image_netayume_lumina_t2i.json +597 -0
- comfyui_workflow_templates/templates/image_omnigen2_image_edit.json +55 -62
- comfyui_workflow_templates/templates/image_omnigen2_t2i.json +26 -33
- comfyui_workflow_templates/templates/image_qwen_image.json +40 -40
- comfyui_workflow_templates/templates/image_qwen_image_controlnet_patch.json +32 -32
- comfyui_workflow_templates/templates/image_qwen_image_edit.json +29 -29
- comfyui_workflow_templates/templates/image_qwen_image_edit_2509.json +127 -127
- comfyui_workflow_templates/templates/image_qwen_image_instantx_controlnet.json +56 -55
- comfyui_workflow_templates/templates/image_qwen_image_instantx_inpainting_controlnet.json +108 -107
- comfyui_workflow_templates/templates/image_qwen_image_union_control_lora.json +5 -5
- comfyui_workflow_templates/templates/index.es.json +24 -0
- comfyui_workflow_templates/templates/index.fr.json +24 -0
- comfyui_workflow_templates/templates/index.ja.json +24 -0
- comfyui_workflow_templates/templates/index.json +11 -0
- comfyui_workflow_templates/templates/index.ko.json +24 -0
- comfyui_workflow_templates/templates/index.ru.json +24 -0
- comfyui_workflow_templates/templates/index.zh-TW.json +24 -0
- comfyui_workflow_templates/templates/index.zh.json +24 -0
- comfyui_workflow_templates/templates/inpaint_example.json +70 -72
- comfyui_workflow_templates/templates/inpaint_model_outpainting.json +4 -4
- comfyui_workflow_templates/templates/latent_upscale_different_prompt_model.json +179 -185
- comfyui_workflow_templates/templates/sdxlturbo_example.json +308 -162
- comfyui_workflow_templates/templates/video_wan2.1_fun_camera_v1.1_1.3B.json +89 -62
- comfyui_workflow_templates/templates/video_wan2.1_fun_camera_v1.1_14B.json +8 -4
- comfyui_workflow_templates/templates/video_wan2_2_14B_animate.json +46 -44
- comfyui_workflow_templates/templates/video_wan2_2_14B_flf2v.json +38 -38
- comfyui_workflow_templates/templates/video_wan2_2_14B_fun_camera.json +58 -54
- comfyui_workflow_templates/templates/video_wan2_2_14B_fun_control.json +36 -36
- comfyui_workflow_templates/templates/video_wan2_2_14B_fun_inpaint.json +26 -26
- comfyui_workflow_templates/templates/video_wan2_2_14B_i2v.json +4 -4
- comfyui_workflow_templates/templates/video_wan2_2_14B_s2v.json +33 -29
- comfyui_workflow_templates/templates/video_wan2_2_14B_t2v (2).json +1954 -0
- comfyui_workflow_templates/templates/video_wan2_2_5B_fun_control.json +29 -29
- comfyui_workflow_templates/templates/video_wan2_2_5B_fun_inpaint.json +25 -25
- comfyui_workflow_templates/templates/video_wan2_2_5B_ti2v.json +49 -49
- comfyui_workflow_templates/templates/video_wan_ati.json +49 -49
- comfyui_workflow_templates/templates/video_wan_vace_14B_ref2v.json +47 -61
- comfyui_workflow_templates/templates/video_wan_vace_14B_t2v.json +2 -2
- comfyui_workflow_templates/templates/video_wan_vace_14B_v2v.json +55 -55
- comfyui_workflow_templates/templates/video_wan_vace_flf2v.json +40 -56
- comfyui_workflow_templates/templates/video_wan_vace_inpainting.json +72 -72
- comfyui_workflow_templates/templates/video_wan_vace_outpainting.json +211 -237
- comfyui_workflow_templates/templates/wan2.1_flf2v_720_f16.json +84 -92
- comfyui_workflow_templates/templates/wan2.1_fun_control.json +51 -27
- comfyui_workflow_templates/templates/wan2.1_fun_inp.json +43 -17
- {comfyui_workflow_templates-0.1.94.dist-info → comfyui_workflow_templates-0.1.95.dist-info}/METADATA +1 -1
- {comfyui_workflow_templates-0.1.94.dist-info → comfyui_workflow_templates-0.1.95.dist-info}/RECORD +129 -125
- {comfyui_workflow_templates-0.1.94.dist-info → comfyui_workflow_templates-0.1.95.dist-info}/WHEEL +0 -0
- {comfyui_workflow_templates-0.1.94.dist-info → comfyui_workflow_templates-0.1.95.dist-info}/licenses/LICENSE +0 -0
- {comfyui_workflow_templates-0.1.94.dist-info → comfyui_workflow_templates-0.1.95.dist-info}/top_level.txt +0 -0
|
@@ -386,6 +386,17 @@
|
|
|
386
386
|
"size": 22.0,
|
|
387
387
|
"vram": 22.0
|
|
388
388
|
},
|
|
389
|
+
{
|
|
390
|
+
"name": "image_netayume_lumina_t2i",
|
|
391
|
+
"title": "NetaYume Lumina Text to Image",
|
|
392
|
+
"mediaType": "image",
|
|
393
|
+
"mediaSubtype": "webp",
|
|
394
|
+
"description": "High-quality anime-style image generation with enhanced character understanding and detailed textures. Fine-tuned from Neta Lumina on Danbooru dataset.",
|
|
395
|
+
"tags": ["Text to Image", "Image", "Anime"],
|
|
396
|
+
"models": ["NetaYume Lumina"],
|
|
397
|
+
"date": "2025-10-10",
|
|
398
|
+
"size": 9.89
|
|
399
|
+
},
|
|
389
400
|
{
|
|
390
401
|
"name": "image_chroma_text_to_image",
|
|
391
402
|
"title": "Chroma texte vers image",
|
|
@@ -1679,6 +1690,19 @@
|
|
|
1679
1690
|
"icon": "icon-[lucide--film]",
|
|
1680
1691
|
"title": "API Vidéo",
|
|
1681
1692
|
"templates": [
|
|
1693
|
+
{
|
|
1694
|
+
"name": "api_openai_sora_video",
|
|
1695
|
+
"title": "Sora 2: Text & Image to Video",
|
|
1696
|
+
"description": "OpenAI's Sora-2 and Sora-2 Pro video generation with synchronized audio.",
|
|
1697
|
+
"mediaType": "image",
|
|
1698
|
+
"mediaSubtype": "webp",
|
|
1699
|
+
"tags": ["Image to Video", "Text to Video", "API"],
|
|
1700
|
+
"models": ["OpenAI"],
|
|
1701
|
+
"date": "2025-10-08",
|
|
1702
|
+
"OpenSource": false,
|
|
1703
|
+
"size": 0,
|
|
1704
|
+
"vram": 0
|
|
1705
|
+
},
|
|
1682
1706
|
{
|
|
1683
1707
|
"name": "api_wan_text_to_video",
|
|
1684
1708
|
"title": "Wan2.5: Text to Video",
|
|
@@ -386,6 +386,17 @@
|
|
|
386
386
|
"size": 22.0,
|
|
387
387
|
"vram": 22.0
|
|
388
388
|
},
|
|
389
|
+
{
|
|
390
|
+
"name": "image_netayume_lumina_t2i",
|
|
391
|
+
"title": "NetaYume Lumina Text to Image",
|
|
392
|
+
"mediaType": "image",
|
|
393
|
+
"mediaSubtype": "webp",
|
|
394
|
+
"description": "High-quality anime-style image generation with enhanced character understanding and detailed textures. Fine-tuned from Neta Lumina on Danbooru dataset.",
|
|
395
|
+
"tags": ["Text to Image", "Image", "Anime"],
|
|
396
|
+
"models": ["NetaYume Lumina"],
|
|
397
|
+
"date": "2025-10-10",
|
|
398
|
+
"size": 9.89
|
|
399
|
+
},
|
|
389
400
|
{
|
|
390
401
|
"name": "image_chroma_text_to_image",
|
|
391
402
|
"title": "Chromaテキストから画像",
|
|
@@ -1678,6 +1689,19 @@
|
|
|
1678
1689
|
"icon": "icon-[lucide--film]",
|
|
1679
1690
|
"title": "ビデオAPI",
|
|
1680
1691
|
"templates": [
|
|
1692
|
+
{
|
|
1693
|
+
"name": "api_openai_sora_video",
|
|
1694
|
+
"title": "Sora 2: Text & Image to Video",
|
|
1695
|
+
"description": "OpenAI's Sora-2 and Sora-2 Pro video generation with synchronized audio.",
|
|
1696
|
+
"mediaType": "image",
|
|
1697
|
+
"mediaSubtype": "webp",
|
|
1698
|
+
"tags": ["Image to Video", "Text to Video", "API"],
|
|
1699
|
+
"models": ["OpenAI"],
|
|
1700
|
+
"date": "2025-10-08",
|
|
1701
|
+
"OpenSource": false,
|
|
1702
|
+
"size": 0,
|
|
1703
|
+
"vram": 0
|
|
1704
|
+
},
|
|
1681
1705
|
{
|
|
1682
1706
|
"name": "api_wan_text_to_video",
|
|
1683
1707
|
"title": "Wan2.5: Text to Video",
|
|
@@ -386,6 +386,17 @@
|
|
|
386
386
|
"size": 22.0,
|
|
387
387
|
"vram": 22.0
|
|
388
388
|
},
|
|
389
|
+
{
|
|
390
|
+
"name": "image_netayume_lumina_t2i",
|
|
391
|
+
"title": "NetaYume Lumina Text to Image",
|
|
392
|
+
"mediaType": "image",
|
|
393
|
+
"mediaSubtype": "webp",
|
|
394
|
+
"description": "High-quality anime-style image generation with enhanced character understanding and detailed textures. Fine-tuned from Neta Lumina on Danbooru dataset.",
|
|
395
|
+
"tags": ["Text to Image", "Image", "Anime"],
|
|
396
|
+
"models": ["NetaYume Lumina"],
|
|
397
|
+
"date": "2025-10-10",
|
|
398
|
+
"size": 9.89
|
|
399
|
+
},
|
|
389
400
|
{
|
|
390
401
|
"name": "image_chroma_text_to_image",
|
|
391
402
|
"title": "Chroma text to image",
|
|
@@ -386,6 +386,17 @@
|
|
|
386
386
|
"size": 22.0,
|
|
387
387
|
"vram": 22.0
|
|
388
388
|
},
|
|
389
|
+
{
|
|
390
|
+
"name": "image_netayume_lumina_t2i",
|
|
391
|
+
"title": "NetaYume Lumina Text to Image",
|
|
392
|
+
"mediaType": "image",
|
|
393
|
+
"mediaSubtype": "webp",
|
|
394
|
+
"description": "High-quality anime-style image generation with enhanced character understanding and detailed textures. Fine-tuned from Neta Lumina on Danbooru dataset.",
|
|
395
|
+
"tags": ["Text to Image", "Image", "Anime"],
|
|
396
|
+
"models": ["NetaYume Lumina"],
|
|
397
|
+
"date": "2025-10-10",
|
|
398
|
+
"size": 9.89
|
|
399
|
+
},
|
|
389
400
|
{
|
|
390
401
|
"name": "image_chroma_text_to_image",
|
|
391
402
|
"title": "Chroma 텍스트에서 이미지",
|
|
@@ -1678,6 +1689,19 @@
|
|
|
1678
1689
|
"icon": "icon-[lucide--film]",
|
|
1679
1690
|
"title": "비디오 API",
|
|
1680
1691
|
"templates": [
|
|
1692
|
+
{
|
|
1693
|
+
"name": "api_openai_sora_video",
|
|
1694
|
+
"title": "Sora 2: Text & Image to Video",
|
|
1695
|
+
"description": "OpenAI's Sora-2 and Sora-2 Pro video generation with synchronized audio.",
|
|
1696
|
+
"mediaType": "image",
|
|
1697
|
+
"mediaSubtype": "webp",
|
|
1698
|
+
"tags": ["Image to Video", "Text to Video", "API"],
|
|
1699
|
+
"models": ["OpenAI"],
|
|
1700
|
+
"date": "2025-10-08",
|
|
1701
|
+
"OpenSource": false,
|
|
1702
|
+
"size": 0,
|
|
1703
|
+
"vram": 0
|
|
1704
|
+
},
|
|
1681
1705
|
{
|
|
1682
1706
|
"name": "api_wan_text_to_video",
|
|
1683
1707
|
"title": "Wan2.5: Text to Video",
|
|
@@ -386,6 +386,17 @@
|
|
|
386
386
|
"size": 22.0,
|
|
387
387
|
"vram": 22.0
|
|
388
388
|
},
|
|
389
|
+
{
|
|
390
|
+
"name": "image_netayume_lumina_t2i",
|
|
391
|
+
"title": "NetaYume Lumina Text to Image",
|
|
392
|
+
"mediaType": "image",
|
|
393
|
+
"mediaSubtype": "webp",
|
|
394
|
+
"description": "High-quality anime-style image generation with enhanced character understanding and detailed textures. Fine-tuned from Neta Lumina on Danbooru dataset.",
|
|
395
|
+
"tags": ["Text to Image", "Image", "Anime"],
|
|
396
|
+
"models": ["NetaYume Lumina"],
|
|
397
|
+
"date": "2025-10-10",
|
|
398
|
+
"size": 9.89
|
|
399
|
+
},
|
|
389
400
|
{
|
|
390
401
|
"name": "image_chroma_text_to_image",
|
|
391
402
|
"title": "Chroma текст в изображение",
|
|
@@ -1678,6 +1689,19 @@
|
|
|
1678
1689
|
"icon": "icon-[lucide--film]",
|
|
1679
1690
|
"title": "API видео",
|
|
1680
1691
|
"templates": [
|
|
1692
|
+
{
|
|
1693
|
+
"name": "api_openai_sora_video",
|
|
1694
|
+
"title": "Sora 2: Text & Image to Video",
|
|
1695
|
+
"description": "OpenAI's Sora-2 and Sora-2 Pro video generation with synchronized audio.",
|
|
1696
|
+
"mediaType": "image",
|
|
1697
|
+
"mediaSubtype": "webp",
|
|
1698
|
+
"tags": ["Image to Video", "Text to Video", "API"],
|
|
1699
|
+
"models": ["OpenAI"],
|
|
1700
|
+
"date": "2025-10-08",
|
|
1701
|
+
"OpenSource": false,
|
|
1702
|
+
"size": 0,
|
|
1703
|
+
"vram": 0
|
|
1704
|
+
},
|
|
1681
1705
|
{
|
|
1682
1706
|
"name": "api_wan_text_to_video",
|
|
1683
1707
|
"title": "Wan2.5: Text to Video",
|
|
@@ -386,6 +386,17 @@
|
|
|
386
386
|
"size": 22.0,
|
|
387
387
|
"vram": 22.0
|
|
388
388
|
},
|
|
389
|
+
{
|
|
390
|
+
"name": "image_netayume_lumina_t2i",
|
|
391
|
+
"title": "NetaYume Lumina Text to Image",
|
|
392
|
+
"mediaType": "image",
|
|
393
|
+
"mediaSubtype": "webp",
|
|
394
|
+
"description": "High-quality anime-style image generation with enhanced character understanding and detailed textures. Fine-tuned from Neta Lumina on Danbooru dataset.",
|
|
395
|
+
"tags": ["Text to Image", "Image", "Anime"],
|
|
396
|
+
"models": ["NetaYume Lumina"],
|
|
397
|
+
"date": "2025-10-10",
|
|
398
|
+
"size": 9.89
|
|
399
|
+
},
|
|
389
400
|
{
|
|
390
401
|
"name": "image_chroma_text_to_image",
|
|
391
402
|
"title": "Chroma 文字到影像",
|
|
@@ -1678,6 +1689,19 @@
|
|
|
1678
1689
|
"icon": "icon-[lucide--film]",
|
|
1679
1690
|
"title": "影片 API",
|
|
1680
1691
|
"templates": [
|
|
1692
|
+
{
|
|
1693
|
+
"name": "api_openai_sora_video",
|
|
1694
|
+
"title": "Sora 2: Text & Image to Video",
|
|
1695
|
+
"description": "OpenAI's Sora-2 and Sora-2 Pro video generation with synchronized audio.",
|
|
1696
|
+
"mediaType": "image",
|
|
1697
|
+
"mediaSubtype": "webp",
|
|
1698
|
+
"tags": ["Image to Video", "Text to Video", "API"],
|
|
1699
|
+
"models": ["OpenAI"],
|
|
1700
|
+
"date": "2025-10-08",
|
|
1701
|
+
"OpenSource": false,
|
|
1702
|
+
"size": 0,
|
|
1703
|
+
"vram": 0
|
|
1704
|
+
},
|
|
1681
1705
|
{
|
|
1682
1706
|
"name": "api_wan_text_to_video",
|
|
1683
1707
|
"title": "Wan2.5: Text to Video",
|
|
@@ -386,6 +386,17 @@
|
|
|
386
386
|
"size": 22.0,
|
|
387
387
|
"vram": 22.0
|
|
388
388
|
},
|
|
389
|
+
{
|
|
390
|
+
"name": "image_netayume_lumina_t2i",
|
|
391
|
+
"title": "NetaYume Lumina Text to Image",
|
|
392
|
+
"mediaType": "image",
|
|
393
|
+
"mediaSubtype": "webp",
|
|
394
|
+
"description": "High-quality anime-style image generation with enhanced character understanding and detailed textures. Fine-tuned from Neta Lumina on Danbooru dataset.",
|
|
395
|
+
"tags": ["Text to Image", "Image", "Anime"],
|
|
396
|
+
"models": ["NetaYume Lumina"],
|
|
397
|
+
"date": "2025-10-10",
|
|
398
|
+
"size": 9.89
|
|
399
|
+
},
|
|
389
400
|
{
|
|
390
401
|
"name": "image_chroma_text_to_image",
|
|
391
402
|
"title": "Chroma文生图",
|
|
@@ -1678,6 +1689,19 @@
|
|
|
1678
1689
|
"icon": "icon-[lucide--film]",
|
|
1679
1690
|
"title": "视频API",
|
|
1680
1691
|
"templates": [
|
|
1692
|
+
{
|
|
1693
|
+
"name": "api_openai_sora_video",
|
|
1694
|
+
"title": "Sora 2: Text & Image to Video",
|
|
1695
|
+
"description": "OpenAI's Sora-2 and Sora-2 Pro video generation with synchronized audio.",
|
|
1696
|
+
"mediaType": "image",
|
|
1697
|
+
"mediaSubtype": "webp",
|
|
1698
|
+
"tags": ["Image to Video", "Text to Video", "API"],
|
|
1699
|
+
"models": ["OpenAI"],
|
|
1700
|
+
"date": "2025-10-08",
|
|
1701
|
+
"OpenSource": false,
|
|
1702
|
+
"size": 0,
|
|
1703
|
+
"vram": 0
|
|
1704
|
+
},
|
|
1681
1705
|
{
|
|
1682
1706
|
"name": "api_wan_text_to_video",
|
|
1683
1707
|
"title": "Wan2.5: Text to Video",
|
|
@@ -4,67 +4,6 @@
|
|
|
4
4
|
"last_node_id": 34,
|
|
5
5
|
"last_link_id": 86,
|
|
6
6
|
"nodes": [
|
|
7
|
-
{
|
|
8
|
-
"id": 29,
|
|
9
|
-
"type": "CheckpointLoaderSimple",
|
|
10
|
-
"pos": [
|
|
11
|
-
80,
|
|
12
|
-
150
|
|
13
|
-
],
|
|
14
|
-
"size": [
|
|
15
|
-
315,
|
|
16
|
-
98
|
|
17
|
-
],
|
|
18
|
-
"flags": {},
|
|
19
|
-
"order": 0,
|
|
20
|
-
"mode": 0,
|
|
21
|
-
"inputs": [],
|
|
22
|
-
"outputs": [
|
|
23
|
-
{
|
|
24
|
-
"name": "MODEL",
|
|
25
|
-
"type": "MODEL",
|
|
26
|
-
"slot_index": 0,
|
|
27
|
-
"links": [
|
|
28
|
-
80
|
|
29
|
-
]
|
|
30
|
-
},
|
|
31
|
-
{
|
|
32
|
-
"name": "CLIP",
|
|
33
|
-
"type": "CLIP",
|
|
34
|
-
"slot_index": 1,
|
|
35
|
-
"links": [
|
|
36
|
-
81,
|
|
37
|
-
82
|
|
38
|
-
]
|
|
39
|
-
},
|
|
40
|
-
{
|
|
41
|
-
"name": "VAE",
|
|
42
|
-
"type": "VAE",
|
|
43
|
-
"slot_index": 2,
|
|
44
|
-
"links": [
|
|
45
|
-
83,
|
|
46
|
-
84
|
|
47
|
-
]
|
|
48
|
-
}
|
|
49
|
-
],
|
|
50
|
-
"properties": {
|
|
51
|
-
"Node name for S&R": "CheckpointLoaderSimple",
|
|
52
|
-
"cnr_id": "comfy-core",
|
|
53
|
-
"ver": "0.3.18",
|
|
54
|
-
"models": [
|
|
55
|
-
{
|
|
56
|
-
"name": "512-inpainting-ema.safetensors",
|
|
57
|
-
"url": "https://huggingface.co/stabilityai/stable-diffusion-2-inpainting/resolve/main/512-inpainting-ema.safetensors",
|
|
58
|
-
"directory": "checkpoints"
|
|
59
|
-
}
|
|
60
|
-
]
|
|
61
|
-
},
|
|
62
|
-
"widgets_values": [
|
|
63
|
-
"512-inpainting-ema.safetensors"
|
|
64
|
-
],
|
|
65
|
-
"color": "#322",
|
|
66
|
-
"bgcolor": "#533"
|
|
67
|
-
},
|
|
68
7
|
{
|
|
69
8
|
"id": 6,
|
|
70
9
|
"type": "CLIPTextEncode",
|
|
@@ -77,7 +16,7 @@
|
|
|
77
16
|
164.30999755859375
|
|
78
17
|
],
|
|
79
18
|
"flags": {},
|
|
80
|
-
"order":
|
|
19
|
+
"order": 6,
|
|
81
20
|
"mode": 0,
|
|
82
21
|
"inputs": [
|
|
83
22
|
{
|
|
@@ -119,7 +58,7 @@
|
|
|
119
58
|
180.61000061035156
|
|
120
59
|
],
|
|
121
60
|
"flags": {},
|
|
122
|
-
"order":
|
|
61
|
+
"order": 7,
|
|
123
62
|
"mode": 0,
|
|
124
63
|
"inputs": [
|
|
125
64
|
{
|
|
@@ -161,7 +100,7 @@
|
|
|
161
100
|
365
|
|
162
101
|
],
|
|
163
102
|
"flags": {},
|
|
164
|
-
"order":
|
|
103
|
+
"order": 0,
|
|
165
104
|
"mode": 0,
|
|
166
105
|
"inputs": [],
|
|
167
106
|
"outputs": [
|
|
@@ -204,7 +143,7 @@
|
|
|
204
143
|
82
|
|
205
144
|
],
|
|
206
145
|
"flags": {},
|
|
207
|
-
"order":
|
|
146
|
+
"order": 5,
|
|
208
147
|
"mode": 0,
|
|
209
148
|
"inputs": [
|
|
210
149
|
{
|
|
@@ -242,7 +181,7 @@
|
|
|
242
181
|
100
|
|
243
182
|
],
|
|
244
183
|
"flags": {},
|
|
245
|
-
"order":
|
|
184
|
+
"order": 1,
|
|
246
185
|
"mode": 0,
|
|
247
186
|
"inputs": [],
|
|
248
187
|
"outputs": [],
|
|
@@ -266,7 +205,7 @@
|
|
|
266
205
|
90
|
|
267
206
|
],
|
|
268
207
|
"flags": {},
|
|
269
|
-
"order":
|
|
208
|
+
"order": 2,
|
|
270
209
|
"mode": 0,
|
|
271
210
|
"inputs": [],
|
|
272
211
|
"outputs": [],
|
|
@@ -462,6 +401,65 @@
|
|
|
462
401
|
"ComfyUI"
|
|
463
402
|
]
|
|
464
403
|
},
|
|
404
|
+
{
|
|
405
|
+
"id": 29,
|
|
406
|
+
"type": "CheckpointLoaderSimple",
|
|
407
|
+
"pos": [
|
|
408
|
+
80,
|
|
409
|
+
150
|
|
410
|
+
],
|
|
411
|
+
"size": [
|
|
412
|
+
315,
|
|
413
|
+
98
|
|
414
|
+
],
|
|
415
|
+
"flags": {},
|
|
416
|
+
"order": 3,
|
|
417
|
+
"mode": 0,
|
|
418
|
+
"inputs": [],
|
|
419
|
+
"outputs": [
|
|
420
|
+
{
|
|
421
|
+
"name": "MODEL",
|
|
422
|
+
"type": "MODEL",
|
|
423
|
+
"slot_index": 0,
|
|
424
|
+
"links": [
|
|
425
|
+
80
|
|
426
|
+
]
|
|
427
|
+
},
|
|
428
|
+
{
|
|
429
|
+
"name": "CLIP",
|
|
430
|
+
"type": "CLIP",
|
|
431
|
+
"slot_index": 1,
|
|
432
|
+
"links": [
|
|
433
|
+
81,
|
|
434
|
+
82
|
|
435
|
+
]
|
|
436
|
+
},
|
|
437
|
+
{
|
|
438
|
+
"name": "VAE",
|
|
439
|
+
"type": "VAE",
|
|
440
|
+
"slot_index": 2,
|
|
441
|
+
"links": [
|
|
442
|
+
83,
|
|
443
|
+
84
|
|
444
|
+
]
|
|
445
|
+
}
|
|
446
|
+
],
|
|
447
|
+
"properties": {
|
|
448
|
+
"Node name for S&R": "CheckpointLoaderSimple",
|
|
449
|
+
"cnr_id": "comfy-core",
|
|
450
|
+
"ver": "0.3.18",
|
|
451
|
+
"models": [
|
|
452
|
+
{
|
|
453
|
+
"name": "512-inpainting-ema.safetensors",
|
|
454
|
+
"url": "https://huggingface.co/stabilityai/stable-diffusion-2-inpainting/resolve/main/512-inpainting-ema.safetensors",
|
|
455
|
+
"directory": "checkpoints"
|
|
456
|
+
}
|
|
457
|
+
]
|
|
458
|
+
},
|
|
459
|
+
"widgets_values": [
|
|
460
|
+
"512-inpainting-ema.safetensors"
|
|
461
|
+
]
|
|
462
|
+
},
|
|
465
463
|
{
|
|
466
464
|
"id": 31,
|
|
467
465
|
"type": "MarkdownNote",
|
|
@@ -481,7 +479,7 @@
|
|
|
481
479
|
"title": "Tutorial links",
|
|
482
480
|
"properties": {},
|
|
483
481
|
"widgets_values": [
|
|
484
|
-
"[Tutorial](https://docs.comfy.org/tutorials/basic/inpaint)
|
|
482
|
+
"[Tutorial](https://docs.comfy.org/tutorials/basic/inpaint) \n\n## Model link\n**checkpoint**\n\n[512-inpainting-ema.safetensors](https://huggingface.co/stabilityai/stable-diffusion-2-inpainting/resolve/main/512-inpainting-ema.safetensors)\n\n\nModel Storage Location\n\n```\n📂 ComfyUI/\n├── 📂 models/\n│ ├── 📂 checkpoints/\n│ │ └── 512-inpainting-ema.safetensors\n\n```\n\n\n"
|
|
485
483
|
],
|
|
486
484
|
"color": "#432",
|
|
487
485
|
"bgcolor": "#653"
|
|
@@ -637,13 +635,13 @@
|
|
|
637
635
|
"config": {},
|
|
638
636
|
"extra": {
|
|
639
637
|
"ds": {
|
|
640
|
-
"scale": 0.
|
|
638
|
+
"scale": 0.4665073802097333,
|
|
641
639
|
"offset": [
|
|
642
|
-
|
|
643
|
-
|
|
640
|
+
349.9776004084935,
|
|
641
|
+
269.12109929530914
|
|
644
642
|
]
|
|
645
643
|
},
|
|
646
|
-
"frontendVersion": "1.
|
|
644
|
+
"frontendVersion": "1.28.6"
|
|
647
645
|
},
|
|
648
646
|
"version": 0.4
|
|
649
647
|
}
|
|
@@ -531,7 +531,7 @@
|
|
|
531
531
|
"title": "Tutorial links",
|
|
532
532
|
"properties": {},
|
|
533
533
|
"widgets_values": [
|
|
534
|
-
"[Tutorial](https://docs.comfy.org/tutorials/basic/outpaint)
|
|
534
|
+
"[Tutorial](https://docs.comfy.org/tutorials/basic/outpaint)\n\n## Model link\n**checkpoint**\n\n[512-inpainting-ema.safetensors](https://huggingface.co/stabilityai/stable-diffusion-2-inpainting/resolve/main/512-inpainting-ema.safetensors)\n\n\nModel Storage Location\n\n```\n📂 ComfyUI/\n├── 📂 models/\n│ ├── 📂 checkpoints/\n│ │ └── 512-inpainting-ema.safetensors\n\n```\n\n\n"
|
|
535
535
|
],
|
|
536
536
|
"color": "#432",
|
|
537
537
|
"bgcolor": "#653"
|
|
@@ -697,11 +697,11 @@
|
|
|
697
697
|
"ds": {
|
|
698
698
|
"scale": 0.6422651457018385,
|
|
699
699
|
"offset": [
|
|
700
|
-
|
|
701
|
-
|
|
700
|
+
913.0850097553838,
|
|
701
|
+
246.0901635832613
|
|
702
702
|
]
|
|
703
703
|
},
|
|
704
|
-
"frontendVersion": "1.
|
|
704
|
+
"frontendVersion": "1.28.6"
|
|
705
705
|
},
|
|
706
706
|
"version": 0.4
|
|
707
707
|
}
|