comfyui-workflow-templates 0.1.94__py3-none-any.whl → 0.1.95__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of comfyui-workflow-templates might be problematic. Click here for more details.

Files changed (129) hide show
  1. comfyui_workflow_templates/templates/3d_hunyuan3d_image_to_model.json +26 -26
  2. comfyui_workflow_templates/templates/3d_hunyuan3d_multiview_to_model.json +35 -35
  3. comfyui_workflow_templates/templates/3d_hunyuan3d_multiview_to_model_turbo.json +37 -37
  4. comfyui_workflow_templates/templates/api_bfl_flux_1_kontext_max_image.json +57 -81
  5. comfyui_workflow_templates/templates/api_bfl_flux_1_kontext_multiple_images_input.json +60 -84
  6. comfyui_workflow_templates/templates/api_bfl_flux_1_kontext_pro_image.json +57 -81
  7. comfyui_workflow_templates/templates/api_bfl_flux_pro_t2i.json +28 -26
  8. comfyui_workflow_templates/templates/api_bytedance_flf2v.json +11 -11
  9. comfyui_workflow_templates/templates/api_bytedance_image_to_video.json +34 -34
  10. comfyui_workflow_templates/templates/api_bytedance_text_to_video.json +39 -40
  11. comfyui_workflow_templates/templates/api_google_gemini.json +6 -7
  12. comfyui_workflow_templates/templates/api_hailuo_minimax_i2v.json +12 -3
  13. comfyui_workflow_templates/templates/api_hailuo_minimax_t2v.json +28 -28
  14. comfyui_workflow_templates/templates/api_hailuo_minimax_video.json +30 -30
  15. comfyui_workflow_templates/templates/api_ideogram_v3_t2i.json +32 -18
  16. comfyui_workflow_templates/templates/api_kling_effects.json +28 -26
  17. comfyui_workflow_templates/templates/api_kling_flf.json +32 -30
  18. comfyui_workflow_templates/templates/api_kling_i2v.json +34 -34
  19. comfyui_workflow_templates/templates/api_luma_i2v.json +93 -110
  20. comfyui_workflow_templates/templates/api_luma_photon_i2i.json +59 -50
  21. comfyui_workflow_templates/templates/api_luma_photon_style_ref.json +131 -124
  22. comfyui_workflow_templates/templates/api_luma_t2v.json +59 -50
  23. comfyui_workflow_templates/templates/api_openai_dall_e_2_t2i.json +8 -29
  24. comfyui_workflow_templates/templates/api_openai_image_1_i2i.json +12 -33
  25. comfyui_workflow_templates/templates/api_openai_image_1_inpaint.json +30 -51
  26. comfyui_workflow_templates/templates/api_openai_image_1_multi_inputs.json +10 -33
  27. comfyui_workflow_templates/templates/api_openai_image_1_t2i.json +42 -63
  28. comfyui_workflow_templates/templates/api_openai_sora_video.json +37 -38
  29. comfyui_workflow_templates/templates/api_pika_i2v.json +27 -27
  30. comfyui_workflow_templates/templates/api_pika_scene.json +12 -3
  31. comfyui_workflow_templates/templates/api_pixverse_i2v.json +31 -36
  32. comfyui_workflow_templates/templates/api_pixverse_t2v.json +20 -16
  33. comfyui_workflow_templates/templates/api_pixverse_template_i2v.json +39 -35
  34. comfyui_workflow_templates/templates/api_recraft_image_gen_with_color_control.json +6 -6
  35. comfyui_workflow_templates/templates/api_recraft_image_gen_with_style_control.json +212 -199
  36. comfyui_workflow_templates/templates/api_recraft_vector_gen.json +78 -69
  37. comfyui_workflow_templates/templates/api_rodin_gen2.json +30 -30
  38. comfyui_workflow_templates/templates/api_rodin_image_to_model.json +55 -55
  39. comfyui_workflow_templates/templates/api_rodin_multiview_to_model.json +188 -132
  40. comfyui_workflow_templates/templates/api_runway_first_last_frame.json +4 -4
  41. comfyui_workflow_templates/templates/api_runway_gen3a_turbo_image_to_video.json +30 -31
  42. comfyui_workflow_templates/templates/api_runway_gen4_turo_image_to_video.json +29 -30
  43. comfyui_workflow_templates/templates/api_runway_reference_to_image.json +31 -32
  44. comfyui_workflow_templates/templates/api_runway_text_to_image.json +17 -17
  45. comfyui_workflow_templates/templates/api_stability_ai_audio_inpaint.json +18 -18
  46. comfyui_workflow_templates/templates/api_stability_ai_audio_to_audio.json +31 -31
  47. comfyui_workflow_templates/templates/api_stability_ai_i2i.json +34 -34
  48. comfyui_workflow_templates/templates/api_stability_ai_sd3.5_i2i.json +21 -19
  49. comfyui_workflow_templates/templates/api_stability_ai_sd3.5_t2i.json +35 -35
  50. comfyui_workflow_templates/templates/api_stability_ai_stable_image_ultra_t2i.json +11 -9
  51. comfyui_workflow_templates/templates/api_tripo_image_to_model.json +90 -92
  52. comfyui_workflow_templates/templates/api_tripo_multiview_to_model.json +241 -241
  53. comfyui_workflow_templates/templates/api_tripo_text_to_model.json +102 -102
  54. comfyui_workflow_templates/templates/api_veo2_i2v.json +31 -28
  55. comfyui_workflow_templates/templates/api_veo3.json +30 -30
  56. comfyui_workflow_templates/templates/api_vidu_text_to_video.json +2 -2
  57. comfyui_workflow_templates/templates/api_wan_image_to_video.json +41 -42
  58. comfyui_workflow_templates/templates/api_wan_text_to_image .json +140 -0
  59. comfyui_workflow_templates/templates/api_wan_text_to_video.json +38 -45
  60. comfyui_workflow_templates/templates/audio_ace_step_1_m2m_editing.json +84 -84
  61. comfyui_workflow_templates/templates/audio_ace_step_1_t2a_instrumentals.json +60 -60
  62. comfyui_workflow_templates/templates/audio_ace_step_1_t2a_song.json +60 -60
  63. comfyui_workflow_templates/templates/esrgan_example.json +24 -30
  64. comfyui_workflow_templates/templates/flux1_dev_uso_reference_image_gen.json +215 -210
  65. comfyui_workflow_templates/templates/flux1_krea_dev.json +3 -3
  66. comfyui_workflow_templates/templates/flux_kontext_dev_basic.json +151 -231
  67. comfyui_workflow_templates/templates/flux_redux_model_example.json +108 -120
  68. comfyui_workflow_templates/templates/flux_schnell_full_text_to_image.json +21 -29
  69. comfyui_workflow_templates/templates/hidream_e1_1.json +179 -209
  70. comfyui_workflow_templates/templates/hidream_e1_full.json +33 -39
  71. comfyui_workflow_templates/templates/hidream_i1_dev.json +15 -15
  72. comfyui_workflow_templates/templates/hidream_i1_fast.json +15 -15
  73. comfyui_workflow_templates/templates/hidream_i1_full.json +17 -16
  74. comfyui_workflow_templates/templates/hiresfix_esrgan_workflow.json +31 -37
  75. comfyui_workflow_templates/templates/hiresfix_latent_workflow.json +84 -88
  76. comfyui_workflow_templates/templates/image2image.json +30 -30
  77. comfyui_workflow_templates/templates/image_chroma1_radiance_text_to_image.json +60 -60
  78. comfyui_workflow_templates/templates/image_lotus_depth_v1_1.json +25 -31
  79. comfyui_workflow_templates/templates/image_netayume_lumina_t2i-1.webp +0 -0
  80. comfyui_workflow_templates/templates/image_netayume_lumina_t2i.json +597 -0
  81. comfyui_workflow_templates/templates/image_omnigen2_image_edit.json +55 -62
  82. comfyui_workflow_templates/templates/image_omnigen2_t2i.json +26 -33
  83. comfyui_workflow_templates/templates/image_qwen_image.json +40 -40
  84. comfyui_workflow_templates/templates/image_qwen_image_controlnet_patch.json +32 -32
  85. comfyui_workflow_templates/templates/image_qwen_image_edit.json +29 -29
  86. comfyui_workflow_templates/templates/image_qwen_image_edit_2509.json +127 -127
  87. comfyui_workflow_templates/templates/image_qwen_image_instantx_controlnet.json +56 -55
  88. comfyui_workflow_templates/templates/image_qwen_image_instantx_inpainting_controlnet.json +108 -107
  89. comfyui_workflow_templates/templates/image_qwen_image_union_control_lora.json +5 -5
  90. comfyui_workflow_templates/templates/index.es.json +24 -0
  91. comfyui_workflow_templates/templates/index.fr.json +24 -0
  92. comfyui_workflow_templates/templates/index.ja.json +24 -0
  93. comfyui_workflow_templates/templates/index.json +11 -0
  94. comfyui_workflow_templates/templates/index.ko.json +24 -0
  95. comfyui_workflow_templates/templates/index.ru.json +24 -0
  96. comfyui_workflow_templates/templates/index.zh-TW.json +24 -0
  97. comfyui_workflow_templates/templates/index.zh.json +24 -0
  98. comfyui_workflow_templates/templates/inpaint_example.json +70 -72
  99. comfyui_workflow_templates/templates/inpaint_model_outpainting.json +4 -4
  100. comfyui_workflow_templates/templates/latent_upscale_different_prompt_model.json +179 -185
  101. comfyui_workflow_templates/templates/sdxlturbo_example.json +308 -162
  102. comfyui_workflow_templates/templates/video_wan2.1_fun_camera_v1.1_1.3B.json +89 -62
  103. comfyui_workflow_templates/templates/video_wan2.1_fun_camera_v1.1_14B.json +8 -4
  104. comfyui_workflow_templates/templates/video_wan2_2_14B_animate.json +46 -44
  105. comfyui_workflow_templates/templates/video_wan2_2_14B_flf2v.json +38 -38
  106. comfyui_workflow_templates/templates/video_wan2_2_14B_fun_camera.json +58 -54
  107. comfyui_workflow_templates/templates/video_wan2_2_14B_fun_control.json +36 -36
  108. comfyui_workflow_templates/templates/video_wan2_2_14B_fun_inpaint.json +26 -26
  109. comfyui_workflow_templates/templates/video_wan2_2_14B_i2v.json +4 -4
  110. comfyui_workflow_templates/templates/video_wan2_2_14B_s2v.json +33 -29
  111. comfyui_workflow_templates/templates/video_wan2_2_14B_t2v (2).json +1954 -0
  112. comfyui_workflow_templates/templates/video_wan2_2_5B_fun_control.json +29 -29
  113. comfyui_workflow_templates/templates/video_wan2_2_5B_fun_inpaint.json +25 -25
  114. comfyui_workflow_templates/templates/video_wan2_2_5B_ti2v.json +49 -49
  115. comfyui_workflow_templates/templates/video_wan_ati.json +49 -49
  116. comfyui_workflow_templates/templates/video_wan_vace_14B_ref2v.json +47 -61
  117. comfyui_workflow_templates/templates/video_wan_vace_14B_t2v.json +2 -2
  118. comfyui_workflow_templates/templates/video_wan_vace_14B_v2v.json +55 -55
  119. comfyui_workflow_templates/templates/video_wan_vace_flf2v.json +40 -56
  120. comfyui_workflow_templates/templates/video_wan_vace_inpainting.json +72 -72
  121. comfyui_workflow_templates/templates/video_wan_vace_outpainting.json +211 -237
  122. comfyui_workflow_templates/templates/wan2.1_flf2v_720_f16.json +84 -92
  123. comfyui_workflow_templates/templates/wan2.1_fun_control.json +51 -27
  124. comfyui_workflow_templates/templates/wan2.1_fun_inp.json +43 -17
  125. {comfyui_workflow_templates-0.1.94.dist-info → comfyui_workflow_templates-0.1.95.dist-info}/METADATA +1 -1
  126. {comfyui_workflow_templates-0.1.94.dist-info → comfyui_workflow_templates-0.1.95.dist-info}/RECORD +129 -125
  127. {comfyui_workflow_templates-0.1.94.dist-info → comfyui_workflow_templates-0.1.95.dist-info}/WHEEL +0 -0
  128. {comfyui_workflow_templates-0.1.94.dist-info → comfyui_workflow_templates-0.1.95.dist-info}/licenses/LICENSE +0 -0
  129. {comfyui_workflow_templates-0.1.94.dist-info → comfyui_workflow_templates-0.1.95.dist-info}/top_level.txt +0 -0
@@ -386,6 +386,17 @@
386
386
  "size": 22.0,
387
387
  "vram": 22.0
388
388
  },
389
+ {
390
+ "name": "image_netayume_lumina_t2i",
391
+ "title": "NetaYume Lumina Text to Image",
392
+ "mediaType": "image",
393
+ "mediaSubtype": "webp",
394
+ "description": "High-quality anime-style image generation with enhanced character understanding and detailed textures. Fine-tuned from Neta Lumina on Danbooru dataset.",
395
+ "tags": ["Text to Image", "Image", "Anime"],
396
+ "models": ["NetaYume Lumina"],
397
+ "date": "2025-10-10",
398
+ "size": 9.89
399
+ },
389
400
  {
390
401
  "name": "image_chroma_text_to_image",
391
402
  "title": "Chroma texte vers image",
@@ -1679,6 +1690,19 @@
1679
1690
  "icon": "icon-[lucide--film]",
1680
1691
  "title": "API Vidéo",
1681
1692
  "templates": [
1693
+ {
1694
+ "name": "api_openai_sora_video",
1695
+ "title": "Sora 2: Text & Image to Video",
1696
+ "description": "OpenAI's Sora-2 and Sora-2 Pro video generation with synchronized audio.",
1697
+ "mediaType": "image",
1698
+ "mediaSubtype": "webp",
1699
+ "tags": ["Image to Video", "Text to Video", "API"],
1700
+ "models": ["OpenAI"],
1701
+ "date": "2025-10-08",
1702
+ "OpenSource": false,
1703
+ "size": 0,
1704
+ "vram": 0
1705
+ },
1682
1706
  {
1683
1707
  "name": "api_wan_text_to_video",
1684
1708
  "title": "Wan2.5: Text to Video",
@@ -386,6 +386,17 @@
386
386
  "size": 22.0,
387
387
  "vram": 22.0
388
388
  },
389
+ {
390
+ "name": "image_netayume_lumina_t2i",
391
+ "title": "NetaYume Lumina Text to Image",
392
+ "mediaType": "image",
393
+ "mediaSubtype": "webp",
394
+ "description": "High-quality anime-style image generation with enhanced character understanding and detailed textures. Fine-tuned from Neta Lumina on Danbooru dataset.",
395
+ "tags": ["Text to Image", "Image", "Anime"],
396
+ "models": ["NetaYume Lumina"],
397
+ "date": "2025-10-10",
398
+ "size": 9.89
399
+ },
389
400
  {
390
401
  "name": "image_chroma_text_to_image",
391
402
  "title": "Chromaテキストから画像",
@@ -1678,6 +1689,19 @@
1678
1689
  "icon": "icon-[lucide--film]",
1679
1690
  "title": "ビデオAPI",
1680
1691
  "templates": [
1692
+ {
1693
+ "name": "api_openai_sora_video",
1694
+ "title": "Sora 2: Text & Image to Video",
1695
+ "description": "OpenAI's Sora-2 and Sora-2 Pro video generation with synchronized audio.",
1696
+ "mediaType": "image",
1697
+ "mediaSubtype": "webp",
1698
+ "tags": ["Image to Video", "Text to Video", "API"],
1699
+ "models": ["OpenAI"],
1700
+ "date": "2025-10-08",
1701
+ "OpenSource": false,
1702
+ "size": 0,
1703
+ "vram": 0
1704
+ },
1681
1705
  {
1682
1706
  "name": "api_wan_text_to_video",
1683
1707
  "title": "Wan2.5: Text to Video",
@@ -386,6 +386,17 @@
386
386
  "size": 22.0,
387
387
  "vram": 22.0
388
388
  },
389
+ {
390
+ "name": "image_netayume_lumina_t2i",
391
+ "title": "NetaYume Lumina Text to Image",
392
+ "mediaType": "image",
393
+ "mediaSubtype": "webp",
394
+ "description": "High-quality anime-style image generation with enhanced character understanding and detailed textures. Fine-tuned from Neta Lumina on Danbooru dataset.",
395
+ "tags": ["Text to Image", "Image", "Anime"],
396
+ "models": ["NetaYume Lumina"],
397
+ "date": "2025-10-10",
398
+ "size": 9.89
399
+ },
389
400
  {
390
401
  "name": "image_chroma_text_to_image",
391
402
  "title": "Chroma text to image",
@@ -386,6 +386,17 @@
386
386
  "size": 22.0,
387
387
  "vram": 22.0
388
388
  },
389
+ {
390
+ "name": "image_netayume_lumina_t2i",
391
+ "title": "NetaYume Lumina Text to Image",
392
+ "mediaType": "image",
393
+ "mediaSubtype": "webp",
394
+ "description": "High-quality anime-style image generation with enhanced character understanding and detailed textures. Fine-tuned from Neta Lumina on Danbooru dataset.",
395
+ "tags": ["Text to Image", "Image", "Anime"],
396
+ "models": ["NetaYume Lumina"],
397
+ "date": "2025-10-10",
398
+ "size": 9.89
399
+ },
389
400
  {
390
401
  "name": "image_chroma_text_to_image",
391
402
  "title": "Chroma 텍스트에서 이미지",
@@ -1678,6 +1689,19 @@
1678
1689
  "icon": "icon-[lucide--film]",
1679
1690
  "title": "비디오 API",
1680
1691
  "templates": [
1692
+ {
1693
+ "name": "api_openai_sora_video",
1694
+ "title": "Sora 2: Text & Image to Video",
1695
+ "description": "OpenAI's Sora-2 and Sora-2 Pro video generation with synchronized audio.",
1696
+ "mediaType": "image",
1697
+ "mediaSubtype": "webp",
1698
+ "tags": ["Image to Video", "Text to Video", "API"],
1699
+ "models": ["OpenAI"],
1700
+ "date": "2025-10-08",
1701
+ "OpenSource": false,
1702
+ "size": 0,
1703
+ "vram": 0
1704
+ },
1681
1705
  {
1682
1706
  "name": "api_wan_text_to_video",
1683
1707
  "title": "Wan2.5: Text to Video",
@@ -386,6 +386,17 @@
386
386
  "size": 22.0,
387
387
  "vram": 22.0
388
388
  },
389
+ {
390
+ "name": "image_netayume_lumina_t2i",
391
+ "title": "NetaYume Lumina Text to Image",
392
+ "mediaType": "image",
393
+ "mediaSubtype": "webp",
394
+ "description": "High-quality anime-style image generation with enhanced character understanding and detailed textures. Fine-tuned from Neta Lumina on Danbooru dataset.",
395
+ "tags": ["Text to Image", "Image", "Anime"],
396
+ "models": ["NetaYume Lumina"],
397
+ "date": "2025-10-10",
398
+ "size": 9.89
399
+ },
389
400
  {
390
401
  "name": "image_chroma_text_to_image",
391
402
  "title": "Chroma текст в изображение",
@@ -1678,6 +1689,19 @@
1678
1689
  "icon": "icon-[lucide--film]",
1679
1690
  "title": "API видео",
1680
1691
  "templates": [
1692
+ {
1693
+ "name": "api_openai_sora_video",
1694
+ "title": "Sora 2: Text & Image to Video",
1695
+ "description": "OpenAI's Sora-2 and Sora-2 Pro video generation with synchronized audio.",
1696
+ "mediaType": "image",
1697
+ "mediaSubtype": "webp",
1698
+ "tags": ["Image to Video", "Text to Video", "API"],
1699
+ "models": ["OpenAI"],
1700
+ "date": "2025-10-08",
1701
+ "OpenSource": false,
1702
+ "size": 0,
1703
+ "vram": 0
1704
+ },
1681
1705
  {
1682
1706
  "name": "api_wan_text_to_video",
1683
1707
  "title": "Wan2.5: Text to Video",
@@ -386,6 +386,17 @@
386
386
  "size": 22.0,
387
387
  "vram": 22.0
388
388
  },
389
+ {
390
+ "name": "image_netayume_lumina_t2i",
391
+ "title": "NetaYume Lumina Text to Image",
392
+ "mediaType": "image",
393
+ "mediaSubtype": "webp",
394
+ "description": "High-quality anime-style image generation with enhanced character understanding and detailed textures. Fine-tuned from Neta Lumina on Danbooru dataset.",
395
+ "tags": ["Text to Image", "Image", "Anime"],
396
+ "models": ["NetaYume Lumina"],
397
+ "date": "2025-10-10",
398
+ "size": 9.89
399
+ },
389
400
  {
390
401
  "name": "image_chroma_text_to_image",
391
402
  "title": "Chroma 文字到影像",
@@ -1678,6 +1689,19 @@
1678
1689
  "icon": "icon-[lucide--film]",
1679
1690
  "title": "影片 API",
1680
1691
  "templates": [
1692
+ {
1693
+ "name": "api_openai_sora_video",
1694
+ "title": "Sora 2: Text & Image to Video",
1695
+ "description": "OpenAI's Sora-2 and Sora-2 Pro video generation with synchronized audio.",
1696
+ "mediaType": "image",
1697
+ "mediaSubtype": "webp",
1698
+ "tags": ["Image to Video", "Text to Video", "API"],
1699
+ "models": ["OpenAI"],
1700
+ "date": "2025-10-08",
1701
+ "OpenSource": false,
1702
+ "size": 0,
1703
+ "vram": 0
1704
+ },
1681
1705
  {
1682
1706
  "name": "api_wan_text_to_video",
1683
1707
  "title": "Wan2.5: Text to Video",
@@ -386,6 +386,17 @@
386
386
  "size": 22.0,
387
387
  "vram": 22.0
388
388
  },
389
+ {
390
+ "name": "image_netayume_lumina_t2i",
391
+ "title": "NetaYume Lumina Text to Image",
392
+ "mediaType": "image",
393
+ "mediaSubtype": "webp",
394
+ "description": "High-quality anime-style image generation with enhanced character understanding and detailed textures. Fine-tuned from Neta Lumina on Danbooru dataset.",
395
+ "tags": ["Text to Image", "Image", "Anime"],
396
+ "models": ["NetaYume Lumina"],
397
+ "date": "2025-10-10",
398
+ "size": 9.89
399
+ },
389
400
  {
390
401
  "name": "image_chroma_text_to_image",
391
402
  "title": "Chroma文生图",
@@ -1678,6 +1689,19 @@
1678
1689
  "icon": "icon-[lucide--film]",
1679
1690
  "title": "视频API",
1680
1691
  "templates": [
1692
+ {
1693
+ "name": "api_openai_sora_video",
1694
+ "title": "Sora 2: Text & Image to Video",
1695
+ "description": "OpenAI's Sora-2 and Sora-2 Pro video generation with synchronized audio.",
1696
+ "mediaType": "image",
1697
+ "mediaSubtype": "webp",
1698
+ "tags": ["Image to Video", "Text to Video", "API"],
1699
+ "models": ["OpenAI"],
1700
+ "date": "2025-10-08",
1701
+ "OpenSource": false,
1702
+ "size": 0,
1703
+ "vram": 0
1704
+ },
1681
1705
  {
1682
1706
  "name": "api_wan_text_to_video",
1683
1707
  "title": "Wan2.5: Text to Video",
@@ -4,67 +4,6 @@
4
4
  "last_node_id": 34,
5
5
  "last_link_id": 86,
6
6
  "nodes": [
7
- {
8
- "id": 29,
9
- "type": "CheckpointLoaderSimple",
10
- "pos": [
11
- 80,
12
- 150
13
- ],
14
- "size": [
15
- 315,
16
- 98
17
- ],
18
- "flags": {},
19
- "order": 0,
20
- "mode": 0,
21
- "inputs": [],
22
- "outputs": [
23
- {
24
- "name": "MODEL",
25
- "type": "MODEL",
26
- "slot_index": 0,
27
- "links": [
28
- 80
29
- ]
30
- },
31
- {
32
- "name": "CLIP",
33
- "type": "CLIP",
34
- "slot_index": 1,
35
- "links": [
36
- 81,
37
- 82
38
- ]
39
- },
40
- {
41
- "name": "VAE",
42
- "type": "VAE",
43
- "slot_index": 2,
44
- "links": [
45
- 83,
46
- 84
47
- ]
48
- }
49
- ],
50
- "properties": {
51
- "Node name for S&R": "CheckpointLoaderSimple",
52
- "cnr_id": "comfy-core",
53
- "ver": "0.3.18",
54
- "models": [
55
- {
56
- "name": "512-inpainting-ema.safetensors",
57
- "url": "https://huggingface.co/stabilityai/stable-diffusion-2-inpainting/resolve/main/512-inpainting-ema.safetensors",
58
- "directory": "checkpoints"
59
- }
60
- ]
61
- },
62
- "widgets_values": [
63
- "512-inpainting-ema.safetensors"
64
- ],
65
- "color": "#322",
66
- "bgcolor": "#533"
67
- },
68
7
  {
69
8
  "id": 6,
70
9
  "type": "CLIPTextEncode",
@@ -77,7 +16,7 @@
77
16
  164.30999755859375
78
17
  ],
79
18
  "flags": {},
80
- "order": 5,
19
+ "order": 6,
81
20
  "mode": 0,
82
21
  "inputs": [
83
22
  {
@@ -119,7 +58,7 @@
119
58
  180.61000061035156
120
59
  ],
121
60
  "flags": {},
122
- "order": 6,
61
+ "order": 7,
123
62
  "mode": 0,
124
63
  "inputs": [
125
64
  {
@@ -161,7 +100,7 @@
161
100
  365
162
101
  ],
163
102
  "flags": {},
164
- "order": 1,
103
+ "order": 0,
165
104
  "mode": 0,
166
105
  "inputs": [],
167
106
  "outputs": [
@@ -204,7 +143,7 @@
204
143
  82
205
144
  ],
206
145
  "flags": {},
207
- "order": 7,
146
+ "order": 5,
208
147
  "mode": 0,
209
148
  "inputs": [
210
149
  {
@@ -242,7 +181,7 @@
242
181
  100
243
182
  ],
244
183
  "flags": {},
245
- "order": 2,
184
+ "order": 1,
246
185
  "mode": 0,
247
186
  "inputs": [],
248
187
  "outputs": [],
@@ -266,7 +205,7 @@
266
205
  90
267
206
  ],
268
207
  "flags": {},
269
- "order": 3,
208
+ "order": 2,
270
209
  "mode": 0,
271
210
  "inputs": [],
272
211
  "outputs": [],
@@ -462,6 +401,65 @@
462
401
  "ComfyUI"
463
402
  ]
464
403
  },
404
+ {
405
+ "id": 29,
406
+ "type": "CheckpointLoaderSimple",
407
+ "pos": [
408
+ 80,
409
+ 150
410
+ ],
411
+ "size": [
412
+ 315,
413
+ 98
414
+ ],
415
+ "flags": {},
416
+ "order": 3,
417
+ "mode": 0,
418
+ "inputs": [],
419
+ "outputs": [
420
+ {
421
+ "name": "MODEL",
422
+ "type": "MODEL",
423
+ "slot_index": 0,
424
+ "links": [
425
+ 80
426
+ ]
427
+ },
428
+ {
429
+ "name": "CLIP",
430
+ "type": "CLIP",
431
+ "slot_index": 1,
432
+ "links": [
433
+ 81,
434
+ 82
435
+ ]
436
+ },
437
+ {
438
+ "name": "VAE",
439
+ "type": "VAE",
440
+ "slot_index": 2,
441
+ "links": [
442
+ 83,
443
+ 84
444
+ ]
445
+ }
446
+ ],
447
+ "properties": {
448
+ "Node name for S&R": "CheckpointLoaderSimple",
449
+ "cnr_id": "comfy-core",
450
+ "ver": "0.3.18",
451
+ "models": [
452
+ {
453
+ "name": "512-inpainting-ema.safetensors",
454
+ "url": "https://huggingface.co/stabilityai/stable-diffusion-2-inpainting/resolve/main/512-inpainting-ema.safetensors",
455
+ "directory": "checkpoints"
456
+ }
457
+ ]
458
+ },
459
+ "widgets_values": [
460
+ "512-inpainting-ema.safetensors"
461
+ ]
462
+ },
465
463
  {
466
464
  "id": 31,
467
465
  "type": "MarkdownNote",
@@ -481,7 +479,7 @@
481
479
  "title": "Tutorial links",
482
480
  "properties": {},
483
481
  "widgets_values": [
484
- "[Tutorial](https://docs.comfy.org/tutorials/basic/inpaint) | [教程](https://docs.comfy.org/zh-CN/tutorials/basic/inpaint)\n\n## Model link\n**checkpoint**\n\n[512-inpainting-ema.safetensors](https://huggingface.co/stabilityai/stable-diffusion-2-inpainting/resolve/main/512-inpainting-ema.safetensors)\n\n\nModel Storage Location\n\n```\n📂 ComfyUI/\n├── 📂 models/\n│ ├── 📂 checkpoints/\n│ │ └── 512-inpainting-ema.safetensors\n\n```\n\n\n"
482
+ "[Tutorial](https://docs.comfy.org/tutorials/basic/inpaint) \n\n## Model link\n**checkpoint**\n\n[512-inpainting-ema.safetensors](https://huggingface.co/stabilityai/stable-diffusion-2-inpainting/resolve/main/512-inpainting-ema.safetensors)\n\n\nModel Storage Location\n\n```\n📂 ComfyUI/\n├── 📂 models/\n│ ├── 📂 checkpoints/\n│ │ └── 512-inpainting-ema.safetensors\n\n```\n\n\n"
485
483
  ],
486
484
  "color": "#432",
487
485
  "bgcolor": "#653"
@@ -637,13 +635,13 @@
637
635
  "config": {},
638
636
  "extra": {
639
637
  "ds": {
640
- "scale": 0.8264462809917354,
638
+ "scale": 0.4665073802097333,
641
639
  "offset": [
642
- 702.1144075614629,
643
- 151.80479576308267
640
+ 349.9776004084935,
641
+ 269.12109929530914
644
642
  ]
645
643
  },
646
- "frontendVersion": "1.24.1"
644
+ "frontendVersion": "1.28.6"
647
645
  },
648
646
  "version": 0.4
649
647
  }
@@ -531,7 +531,7 @@
531
531
  "title": "Tutorial links",
532
532
  "properties": {},
533
533
  "widgets_values": [
534
- "[Tutorial](https://docs.comfy.org/tutorials/basic/outpaint) | [教程](https://docs.comfy.org/zh-CN/tutorials/basic/outpaint)\n\n## Model link\n**checkpoint**\n\n[512-inpainting-ema.safetensors](https://huggingface.co/stabilityai/stable-diffusion-2-inpainting/resolve/main/512-inpainting-ema.safetensors)\n\n\nModel Storage Location\n\n```\n📂 ComfyUI/\n├── 📂 models/\n│ ├── 📂 checkpoints/\n│ │ └── 512-inpainting-ema.safetensors\n\n```\n\n\n"
534
+ "[Tutorial](https://docs.comfy.org/tutorials/basic/outpaint)\n\n## Model link\n**checkpoint**\n\n[512-inpainting-ema.safetensors](https://huggingface.co/stabilityai/stable-diffusion-2-inpainting/resolve/main/512-inpainting-ema.safetensors)\n\n\nModel Storage Location\n\n```\n📂 ComfyUI/\n├── 📂 models/\n│ ├── 📂 checkpoints/\n│ │ └── 512-inpainting-ema.safetensors\n\n```\n\n\n"
535
535
  ],
536
536
  "color": "#432",
537
537
  "bgcolor": "#653"
@@ -697,11 +697,11 @@
697
697
  "ds": {
698
698
  "scale": 0.6422651457018385,
699
699
  "offset": [
700
- 1348.2270447393946,
701
- 268.16170419988515
700
+ 913.0850097553838,
701
+ 246.0901635832613
702
702
  ]
703
703
  },
704
- "frontendVersion": "1.24.1"
704
+ "frontendVersion": "1.28.6"
705
705
  },
706
706
  "version": 0.4
707
707
  }