comfyui-workflow-templates 0.1.94__py3-none-any.whl → 0.1.95__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of comfyui-workflow-templates might be problematic. Click here for more details.

Files changed (129) hide show
  1. comfyui_workflow_templates/templates/3d_hunyuan3d_image_to_model.json +26 -26
  2. comfyui_workflow_templates/templates/3d_hunyuan3d_multiview_to_model.json +35 -35
  3. comfyui_workflow_templates/templates/3d_hunyuan3d_multiview_to_model_turbo.json +37 -37
  4. comfyui_workflow_templates/templates/api_bfl_flux_1_kontext_max_image.json +57 -81
  5. comfyui_workflow_templates/templates/api_bfl_flux_1_kontext_multiple_images_input.json +60 -84
  6. comfyui_workflow_templates/templates/api_bfl_flux_1_kontext_pro_image.json +57 -81
  7. comfyui_workflow_templates/templates/api_bfl_flux_pro_t2i.json +28 -26
  8. comfyui_workflow_templates/templates/api_bytedance_flf2v.json +11 -11
  9. comfyui_workflow_templates/templates/api_bytedance_image_to_video.json +34 -34
  10. comfyui_workflow_templates/templates/api_bytedance_text_to_video.json +39 -40
  11. comfyui_workflow_templates/templates/api_google_gemini.json +6 -7
  12. comfyui_workflow_templates/templates/api_hailuo_minimax_i2v.json +12 -3
  13. comfyui_workflow_templates/templates/api_hailuo_minimax_t2v.json +28 -28
  14. comfyui_workflow_templates/templates/api_hailuo_minimax_video.json +30 -30
  15. comfyui_workflow_templates/templates/api_ideogram_v3_t2i.json +32 -18
  16. comfyui_workflow_templates/templates/api_kling_effects.json +28 -26
  17. comfyui_workflow_templates/templates/api_kling_flf.json +32 -30
  18. comfyui_workflow_templates/templates/api_kling_i2v.json +34 -34
  19. comfyui_workflow_templates/templates/api_luma_i2v.json +93 -110
  20. comfyui_workflow_templates/templates/api_luma_photon_i2i.json +59 -50
  21. comfyui_workflow_templates/templates/api_luma_photon_style_ref.json +131 -124
  22. comfyui_workflow_templates/templates/api_luma_t2v.json +59 -50
  23. comfyui_workflow_templates/templates/api_openai_dall_e_2_t2i.json +8 -29
  24. comfyui_workflow_templates/templates/api_openai_image_1_i2i.json +12 -33
  25. comfyui_workflow_templates/templates/api_openai_image_1_inpaint.json +30 -51
  26. comfyui_workflow_templates/templates/api_openai_image_1_multi_inputs.json +10 -33
  27. comfyui_workflow_templates/templates/api_openai_image_1_t2i.json +42 -63
  28. comfyui_workflow_templates/templates/api_openai_sora_video.json +37 -38
  29. comfyui_workflow_templates/templates/api_pika_i2v.json +27 -27
  30. comfyui_workflow_templates/templates/api_pika_scene.json +12 -3
  31. comfyui_workflow_templates/templates/api_pixverse_i2v.json +31 -36
  32. comfyui_workflow_templates/templates/api_pixverse_t2v.json +20 -16
  33. comfyui_workflow_templates/templates/api_pixverse_template_i2v.json +39 -35
  34. comfyui_workflow_templates/templates/api_recraft_image_gen_with_color_control.json +6 -6
  35. comfyui_workflow_templates/templates/api_recraft_image_gen_with_style_control.json +212 -199
  36. comfyui_workflow_templates/templates/api_recraft_vector_gen.json +78 -69
  37. comfyui_workflow_templates/templates/api_rodin_gen2.json +30 -30
  38. comfyui_workflow_templates/templates/api_rodin_image_to_model.json +55 -55
  39. comfyui_workflow_templates/templates/api_rodin_multiview_to_model.json +188 -132
  40. comfyui_workflow_templates/templates/api_runway_first_last_frame.json +4 -4
  41. comfyui_workflow_templates/templates/api_runway_gen3a_turbo_image_to_video.json +30 -31
  42. comfyui_workflow_templates/templates/api_runway_gen4_turo_image_to_video.json +29 -30
  43. comfyui_workflow_templates/templates/api_runway_reference_to_image.json +31 -32
  44. comfyui_workflow_templates/templates/api_runway_text_to_image.json +17 -17
  45. comfyui_workflow_templates/templates/api_stability_ai_audio_inpaint.json +18 -18
  46. comfyui_workflow_templates/templates/api_stability_ai_audio_to_audio.json +31 -31
  47. comfyui_workflow_templates/templates/api_stability_ai_i2i.json +34 -34
  48. comfyui_workflow_templates/templates/api_stability_ai_sd3.5_i2i.json +21 -19
  49. comfyui_workflow_templates/templates/api_stability_ai_sd3.5_t2i.json +35 -35
  50. comfyui_workflow_templates/templates/api_stability_ai_stable_image_ultra_t2i.json +11 -9
  51. comfyui_workflow_templates/templates/api_tripo_image_to_model.json +90 -92
  52. comfyui_workflow_templates/templates/api_tripo_multiview_to_model.json +241 -241
  53. comfyui_workflow_templates/templates/api_tripo_text_to_model.json +102 -102
  54. comfyui_workflow_templates/templates/api_veo2_i2v.json +31 -28
  55. comfyui_workflow_templates/templates/api_veo3.json +30 -30
  56. comfyui_workflow_templates/templates/api_vidu_text_to_video.json +2 -2
  57. comfyui_workflow_templates/templates/api_wan_image_to_video.json +41 -42
  58. comfyui_workflow_templates/templates/api_wan_text_to_image .json +140 -0
  59. comfyui_workflow_templates/templates/api_wan_text_to_video.json +38 -45
  60. comfyui_workflow_templates/templates/audio_ace_step_1_m2m_editing.json +84 -84
  61. comfyui_workflow_templates/templates/audio_ace_step_1_t2a_instrumentals.json +60 -60
  62. comfyui_workflow_templates/templates/audio_ace_step_1_t2a_song.json +60 -60
  63. comfyui_workflow_templates/templates/esrgan_example.json +24 -30
  64. comfyui_workflow_templates/templates/flux1_dev_uso_reference_image_gen.json +215 -210
  65. comfyui_workflow_templates/templates/flux1_krea_dev.json +3 -3
  66. comfyui_workflow_templates/templates/flux_kontext_dev_basic.json +151 -231
  67. comfyui_workflow_templates/templates/flux_redux_model_example.json +108 -120
  68. comfyui_workflow_templates/templates/flux_schnell_full_text_to_image.json +21 -29
  69. comfyui_workflow_templates/templates/hidream_e1_1.json +179 -209
  70. comfyui_workflow_templates/templates/hidream_e1_full.json +33 -39
  71. comfyui_workflow_templates/templates/hidream_i1_dev.json +15 -15
  72. comfyui_workflow_templates/templates/hidream_i1_fast.json +15 -15
  73. comfyui_workflow_templates/templates/hidream_i1_full.json +17 -16
  74. comfyui_workflow_templates/templates/hiresfix_esrgan_workflow.json +31 -37
  75. comfyui_workflow_templates/templates/hiresfix_latent_workflow.json +84 -88
  76. comfyui_workflow_templates/templates/image2image.json +30 -30
  77. comfyui_workflow_templates/templates/image_chroma1_radiance_text_to_image.json +60 -60
  78. comfyui_workflow_templates/templates/image_lotus_depth_v1_1.json +25 -31
  79. comfyui_workflow_templates/templates/image_netayume_lumina_t2i-1.webp +0 -0
  80. comfyui_workflow_templates/templates/image_netayume_lumina_t2i.json +597 -0
  81. comfyui_workflow_templates/templates/image_omnigen2_image_edit.json +55 -62
  82. comfyui_workflow_templates/templates/image_omnigen2_t2i.json +26 -33
  83. comfyui_workflow_templates/templates/image_qwen_image.json +40 -40
  84. comfyui_workflow_templates/templates/image_qwen_image_controlnet_patch.json +32 -32
  85. comfyui_workflow_templates/templates/image_qwen_image_edit.json +29 -29
  86. comfyui_workflow_templates/templates/image_qwen_image_edit_2509.json +127 -127
  87. comfyui_workflow_templates/templates/image_qwen_image_instantx_controlnet.json +56 -55
  88. comfyui_workflow_templates/templates/image_qwen_image_instantx_inpainting_controlnet.json +108 -107
  89. comfyui_workflow_templates/templates/image_qwen_image_union_control_lora.json +5 -5
  90. comfyui_workflow_templates/templates/index.es.json +24 -0
  91. comfyui_workflow_templates/templates/index.fr.json +24 -0
  92. comfyui_workflow_templates/templates/index.ja.json +24 -0
  93. comfyui_workflow_templates/templates/index.json +11 -0
  94. comfyui_workflow_templates/templates/index.ko.json +24 -0
  95. comfyui_workflow_templates/templates/index.ru.json +24 -0
  96. comfyui_workflow_templates/templates/index.zh-TW.json +24 -0
  97. comfyui_workflow_templates/templates/index.zh.json +24 -0
  98. comfyui_workflow_templates/templates/inpaint_example.json +70 -72
  99. comfyui_workflow_templates/templates/inpaint_model_outpainting.json +4 -4
  100. comfyui_workflow_templates/templates/latent_upscale_different_prompt_model.json +179 -185
  101. comfyui_workflow_templates/templates/sdxlturbo_example.json +308 -162
  102. comfyui_workflow_templates/templates/video_wan2.1_fun_camera_v1.1_1.3B.json +89 -62
  103. comfyui_workflow_templates/templates/video_wan2.1_fun_camera_v1.1_14B.json +8 -4
  104. comfyui_workflow_templates/templates/video_wan2_2_14B_animate.json +46 -44
  105. comfyui_workflow_templates/templates/video_wan2_2_14B_flf2v.json +38 -38
  106. comfyui_workflow_templates/templates/video_wan2_2_14B_fun_camera.json +58 -54
  107. comfyui_workflow_templates/templates/video_wan2_2_14B_fun_control.json +36 -36
  108. comfyui_workflow_templates/templates/video_wan2_2_14B_fun_inpaint.json +26 -26
  109. comfyui_workflow_templates/templates/video_wan2_2_14B_i2v.json +4 -4
  110. comfyui_workflow_templates/templates/video_wan2_2_14B_s2v.json +33 -29
  111. comfyui_workflow_templates/templates/video_wan2_2_14B_t2v (2).json +1954 -0
  112. comfyui_workflow_templates/templates/video_wan2_2_5B_fun_control.json +29 -29
  113. comfyui_workflow_templates/templates/video_wan2_2_5B_fun_inpaint.json +25 -25
  114. comfyui_workflow_templates/templates/video_wan2_2_5B_ti2v.json +49 -49
  115. comfyui_workflow_templates/templates/video_wan_ati.json +49 -49
  116. comfyui_workflow_templates/templates/video_wan_vace_14B_ref2v.json +47 -61
  117. comfyui_workflow_templates/templates/video_wan_vace_14B_t2v.json +2 -2
  118. comfyui_workflow_templates/templates/video_wan_vace_14B_v2v.json +55 -55
  119. comfyui_workflow_templates/templates/video_wan_vace_flf2v.json +40 -56
  120. comfyui_workflow_templates/templates/video_wan_vace_inpainting.json +72 -72
  121. comfyui_workflow_templates/templates/video_wan_vace_outpainting.json +211 -237
  122. comfyui_workflow_templates/templates/wan2.1_flf2v_720_f16.json +84 -92
  123. comfyui_workflow_templates/templates/wan2.1_fun_control.json +51 -27
  124. comfyui_workflow_templates/templates/wan2.1_fun_inp.json +43 -17
  125. {comfyui_workflow_templates-0.1.94.dist-info → comfyui_workflow_templates-0.1.95.dist-info}/METADATA +1 -1
  126. {comfyui_workflow_templates-0.1.94.dist-info → comfyui_workflow_templates-0.1.95.dist-info}/RECORD +129 -125
  127. {comfyui_workflow_templates-0.1.94.dist-info → comfyui_workflow_templates-0.1.95.dist-info}/WHEEL +0 -0
  128. {comfyui_workflow_templates-0.1.94.dist-info → comfyui_workflow_templates-0.1.95.dist-info}/licenses/LICENSE +0 -0
  129. {comfyui_workflow_templates-0.1.94.dist-info → comfyui_workflow_templates-0.1.95.dist-info}/top_level.txt +0 -0
@@ -30,9 +30,9 @@
30
30
  }
31
31
  ],
32
32
  "properties": {
33
+ "Node name for S&R": "VAELoader",
33
34
  "cnr_id": "comfy-core",
34
35
  "ver": "0.3.48",
35
- "Node name for S&R": "VAELoader",
36
36
  "models": [
37
37
  {
38
38
  "name": "qwen_image_vae.safetensors",
@@ -80,9 +80,9 @@
80
80
  }
81
81
  ],
82
82
  "properties": {
83
+ "Node name for S&R": "CLIPLoader",
83
84
  "cnr_id": "comfy-core",
84
85
  "ver": "0.3.48",
85
- "Node name for S&R": "CLIPLoader",
86
86
  "models": [
87
87
  {
88
88
  "name": "qwen_2.5_vl_7b_fp8_scaled.safetensors",
@@ -130,9 +130,9 @@
130
130
  }
131
131
  ],
132
132
  "properties": {
133
+ "Node name for S&R": "EmptySD3LatentImage",
133
134
  "cnr_id": "comfy-core",
134
135
  "ver": "0.3.48",
135
- "Node name for S&R": "EmptySD3LatentImage",
136
136
  "enableTabs": false,
137
137
  "tabWidth": 65,
138
138
  "tabXOffset": 10,
@@ -181,9 +181,9 @@
181
181
  ],
182
182
  "title": "CLIP Text Encode (Positive Prompt)",
183
183
  "properties": {
184
+ "Node name for S&R": "CLIPTextEncode",
184
185
  "cnr_id": "comfy-core",
185
186
  "ver": "0.3.48",
186
- "Node name for S&R": "CLIPTextEncode",
187
187
  "enableTabs": false,
188
188
  "tabWidth": 65,
189
189
  "tabXOffset": 10,
@@ -232,9 +232,9 @@
232
232
  ],
233
233
  "title": "CLIP Text Encode (Negative Prompt)",
234
234
  "properties": {
235
+ "Node name for S&R": "CLIPTextEncode",
235
236
  "cnr_id": "comfy-core",
236
237
  "ver": "0.3.48",
237
- "Node name for S&R": "CLIPTextEncode",
238
238
  "enableTabs": false,
239
239
  "tabWidth": 65,
240
240
  "tabXOffset": 10,
@@ -273,9 +273,9 @@
273
273
  ],
274
274
  "outputs": [],
275
275
  "properties": {
276
+ "Node name for S&R": "SaveImage",
276
277
  "cnr_id": "comfy-core",
277
278
  "ver": "0.3.48",
278
- "Node name for S&R": "SaveImage",
279
279
  "enableTabs": false,
280
280
  "tabWidth": 65,
281
281
  "tabXOffset": 10,
@@ -320,9 +320,9 @@
320
320
  }
321
321
  ],
322
322
  "properties": {
323
+ "Node name for S&R": "ModelSamplingAuraFlow",
323
324
  "cnr_id": "comfy-core",
324
325
  "ver": "0.3.48",
325
- "Node name for S&R": "ModelSamplingAuraFlow",
326
326
  "enableTabs": false,
327
327
  "tabWidth": 65,
328
328
  "tabXOffset": 10,
@@ -367,9 +367,9 @@
367
367
  }
368
368
  ],
369
369
  "properties": {
370
+ "Node name for S&R": "LoraLoaderModelOnly",
370
371
  "cnr_id": "comfy-core",
371
372
  "ver": "0.3.49",
372
- "Node name for S&R": "LoraLoaderModelOnly",
373
373
  "models": [
374
374
  {
375
375
  "name": "Qwen-Image-Lightning-8steps-V1.0.safetensors",
@@ -471,9 +471,9 @@
471
471
  }
472
472
  ],
473
473
  "properties": {
474
+ "Node name for S&R": "VAEDecode",
474
475
  "cnr_id": "comfy-core",
475
476
  "ver": "0.3.48",
476
- "Node name for S&R": "VAEDecode",
477
477
  "enableTabs": false,
478
478
  "tabWidth": 65,
479
479
  "tabXOffset": 10,
@@ -532,9 +532,9 @@
532
532
  }
533
533
  ],
534
534
  "properties": {
535
+ "Node name for S&R": "KSampler",
535
536
  "cnr_id": "comfy-core",
536
537
  "ver": "0.3.48",
537
- "Node name for S&R": "KSampler",
538
538
  "enableTabs": false,
539
539
  "tabWidth": 65,
540
540
  "tabXOffset": 10,
@@ -580,9 +580,9 @@
580
580
  }
581
581
  ],
582
582
  "properties": {
583
+ "Node name for S&R": "UNETLoader",
583
584
  "cnr_id": "comfy-core",
584
585
  "ver": "0.3.48",
585
- "Node name for S&R": "UNETLoader",
586
586
  "models": [
587
587
  {
588
588
  "name": "qwen_image_fp8_e4m3fn.safetensors",
@@ -605,75 +605,75 @@
605
605
  ]
606
606
  },
607
607
  {
608
- "id": 67,
609
- "type": "MarkdownNote",
608
+ "id": 70,
609
+ "type": "Note",
610
610
  "pos": [
611
- -540,
612
- 10
611
+ 850,
612
+ 910
613
613
  ],
614
614
  "size": [
615
- 540,
616
- 630
615
+ 310,
616
+ 120
617
617
  ],
618
618
  "flags": {},
619
619
  "order": 6,
620
620
  "mode": 0,
621
621
  "inputs": [],
622
622
  "outputs": [],
623
- "title": "Model links",
624
- "properties": {
625
- "widget_ue_connectable": {}
626
- },
623
+ "title": "For fp8 without 8steps LoRA",
624
+ "properties": {},
627
625
  "widgets_values": [
628
- "[Tutorial](https://docs.comfy.org/tutorials/image/qwen/qwen-image) | [教程](https://docs.comfy.org/zh-CN/tutorials/image/qwen/qwen-image)\n\n\n## Model links\n\nYou can find all the models on [Huggingface](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/tree/main) or [Modelscope](https://modelscope.cn/models/Comfy-Org/Qwen-Image_ComfyUI/files)\n\n**Diffusion model**\n\n- [qwen_image_fp8_e4m3fn.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/diffusion_models/qwen_image_fp8_e4m3fn.safetensors)\n\nQwen_image_distill\n\n- [qwen_image_distill_full_fp8_e4m3fn.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/non_official/diffusion_models/qwen_image_distill_full_fp8_e4m3fn.safetensors)\n- [qwen_image_distill_full_bf16.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/non_official/diffusion_models/qwen_image_distill_full_bf16.safetensors)\n\n**LoRA**\n\n- [Qwen-Image-Lightning-8steps-V1.0.safetensors](https://huggingface.co/lightx2v/Qwen-Image-Lightning/resolve/main/Qwen-Image-Lightning-8steps-V1.0.safetensors)\n\n**Text encoder**\n\n- [qwen_2.5_vl_7b_fp8_scaled.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/text_encoders/qwen_2.5_vl_7b_fp8_scaled.safetensors)\n\n**VAE**\n\n- [qwen_image_vae.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/vae/qwen_image_vae.safetensors)\n\nModel Storage Location\n\n```\n📂 ComfyUI/\n├── 📂 models/\n│ ├── 📂 diffusion_models/\n│ │ ├── qwen_image_fp8_e4m3fn.safetensors\n│ │ └── qwen_image_distill_full_fp8_e4m3fn.safetensors\n│ ├── 📂 loras/\n│ │ └── Qwen-Image-Lightning-8steps-V1.0.safetensors\n│ ├── 📂 vae/\n│ │ └── qwen_image_vae.safetensors\n│ └── 📂 text_encoders/\n│ └── qwen_2.5_vl_7b_fp8_scaled.safetensors\n```\n"
626
+ "Set cfg to 1.0 for a speed boost at the cost of consistency. Samplers like res_multistep work pretty well at cfg 1.0\n\nThe official number of steps is 50 but I think that's too much. Even just 10 steps seems to work."
629
627
  ],
630
628
  "color": "#432",
631
629
  "bgcolor": "#653"
632
630
  },
633
631
  {
634
- "id": 70,
635
- "type": "Note",
632
+ "id": 74,
633
+ "type": "MarkdownNote",
636
634
  "pos": [
637
635
  850,
638
- 910
636
+ 660
639
637
  ],
640
638
  "size": [
641
639
  310,
642
- 120
640
+ 210
643
641
  ],
644
642
  "flags": {},
645
643
  "order": 7,
646
644
  "mode": 0,
647
645
  "inputs": [],
648
646
  "outputs": [],
649
- "title": "For fp8 without 8steps LoRA",
647
+ "title": "KSampler settings",
650
648
  "properties": {},
651
649
  "widgets_values": [
652
- "Set cfg to 1.0 for a speed boost at the cost of consistency. Samplers like res_multistep work pretty well at cfg 1.0\n\nThe official number of steps is 50 but I think that's too much. Even just 10 steps seems to work."
650
+ "You can test and find the best setting by yourself. The following table is for reference.\n\n| model | steps | cfg |\n|---------------------|---------------|---------------|\n| fp8_e4m3fn(Qwen team's suggestion) | 40 | 2.5 \n| fp8_e4m3fn | 20 | 2.5 |\n| fp8_e4m3fn + 8steps LoRA | 8 | 1.0 |\n| distill fp8_e4m3fn | 10 | 1.0 |"
653
651
  ],
654
652
  "color": "#432",
655
653
  "bgcolor": "#653"
656
654
  },
657
655
  {
658
- "id": 74,
656
+ "id": 67,
659
657
  "type": "MarkdownNote",
660
658
  "pos": [
661
- 850,
662
- 660
659
+ -540,
660
+ 10
663
661
  ],
664
662
  "size": [
665
- 310,
666
- 210
663
+ 540,
664
+ 630
667
665
  ],
668
666
  "flags": {},
669
667
  "order": 8,
670
668
  "mode": 0,
671
669
  "inputs": [],
672
670
  "outputs": [],
673
- "title": "KSampler settings",
674
- "properties": {},
671
+ "title": "Model links",
672
+ "properties": {
673
+ "widget_ue_connectable": {}
674
+ },
675
675
  "widgets_values": [
676
- "You can test and find the best setting by yourself. The following table is for reference.\n\n| model | steps | cfg |\n|---------------------|---------------|---------------|\n| fp8_e4m3fn(Qwen team's suggestion) | 40 | 2.5 \n| fp8_e4m3fn | 20 | 2.5 |\n| fp8_e4m3fn + 8steps LoRA | 8 | 1.0 |\n| distill fp8_e4m3fn | 10 | 1.0 |"
676
+ "[Tutorial](https://docs.comfy.org/tutorials/image/qwen/qwen-image) \n\n## Model links\n\nYou can find all the models on [Huggingface](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/tree/main) or [Modelscope](https://modelscope.cn/models/Comfy-Org/Qwen-Image_ComfyUI/files)\n\n**Diffusion model**\n\n- [qwen_image_fp8_e4m3fn.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/diffusion_models/qwen_image_fp8_e4m3fn.safetensors)\n\nQwen_image_distill\n\n- [qwen_image_distill_full_fp8_e4m3fn.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/non_official/diffusion_models/qwen_image_distill_full_fp8_e4m3fn.safetensors)\n- [qwen_image_distill_full_bf16.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/non_official/diffusion_models/qwen_image_distill_full_bf16.safetensors)\n\n**LoRA**\n\n- [Qwen-Image-Lightning-8steps-V1.0.safetensors](https://huggingface.co/lightx2v/Qwen-Image-Lightning/resolve/main/Qwen-Image-Lightning-8steps-V1.0.safetensors)\n\n**Text encoder**\n\n- [qwen_2.5_vl_7b_fp8_scaled.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/text_encoders/qwen_2.5_vl_7b_fp8_scaled.safetensors)\n\n**VAE**\n\n- [qwen_image_vae.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/vae/qwen_image_vae.safetensors)\n\nModel Storage Location\n\n```\n📂 ComfyUI/\n├── 📂 models/\n│ ├── 📂 diffusion_models/\n│ │ ├── qwen_image_fp8_e4m3fn.safetensors\n│ │ └── qwen_image_distill_full_fp8_e4m3fn.safetensors\n│ ├── 📂 loras/\n│ │ └── Qwen-Image-Lightning-8steps-V1.0.safetensors\n│ ├── 📂 vae/\n│ │ └── qwen_image_vae.safetensors\n│ └── 📂 text_encoders/\n│ └── qwen_2.5_vl_7b_fp8_scaled.safetensors\n```\n"
677
677
  ],
678
678
  "color": "#432",
679
679
  "bgcolor": "#653"
@@ -826,13 +826,13 @@
826
826
  "config": {},
827
827
  "extra": {
828
828
  "ds": {
829
- "scale": 0.2480466293395618,
829
+ "scale": 0.33015006365095706,
830
830
  "offset": [
831
- 4246.689662080242,
832
- 988.9239858025617
831
+ 1294.3888372784004,
832
+ 848.3556072418501
833
833
  ]
834
834
  },
835
- "frontendVersion": "1.26.6",
835
+ "frontendVersion": "1.28.6",
836
836
  "ue_links": [],
837
837
  "links_added_by_ue": [],
838
838
  "VHS_latentpreview": false,
@@ -36,9 +36,9 @@
36
36
  }
37
37
  ],
38
38
  "properties": {
39
+ "Node name for S&R": "ImageScaleToTotalPixels",
39
40
  "cnr_id": "comfy-core",
40
- "ver": "0.3.51",
41
- "Node name for S&R": "ImageScaleToTotalPixels"
41
+ "ver": "0.3.51"
42
42
  },
43
43
  "widgets_values": [
44
44
  "area",
@@ -97,9 +97,9 @@
97
97
  }
98
98
  ],
99
99
  "properties": {
100
+ "Node name for S&R": "QwenImageDiffsynthControlnet",
100
101
  "cnr_id": "comfy-core",
101
- "ver": "0.3.51",
102
- "Node name for S&R": "QwenImageDiffsynthControlnet"
102
+ "ver": "0.3.51"
103
103
  },
104
104
  "widgets_values": [
105
105
  1
@@ -136,9 +136,9 @@
136
136
  }
137
137
  ],
138
138
  "properties": {
139
+ "Node name for S&R": "ModelSamplingAuraFlow",
139
140
  "cnr_id": "comfy-core",
140
- "ver": "0.3.51",
141
- "Node name for S&R": "ModelSamplingAuraFlow"
141
+ "ver": "0.3.51"
142
142
  },
143
143
  "widgets_values": [
144
144
  3.1000000000000005
@@ -205,9 +205,9 @@
205
205
  }
206
206
  ],
207
207
  "properties": {
208
+ "Node name for S&R": "VAEDecode",
208
209
  "cnr_id": "comfy-core",
209
- "ver": "0.3.51",
210
- "Node name for S&R": "VAEDecode"
210
+ "ver": "0.3.51"
211
211
  },
212
212
  "widgets_values": []
213
213
  },
@@ -236,9 +236,9 @@
236
236
  }
237
237
  ],
238
238
  "properties": {
239
+ "Node name for S&R": "ModelPatchLoader",
239
240
  "cnr_id": "comfy-core",
240
241
  "ver": "0.3.51",
241
- "Node name for S&R": "ModelPatchLoader",
242
242
  "models": [
243
243
  {
244
244
  "name": "qwen_image_canny_diffsynth_controlnet.safetensors",
@@ -279,9 +279,9 @@
279
279
  }
280
280
  ],
281
281
  "properties": {
282
+ "Node name for S&R": "VAELoader",
282
283
  "cnr_id": "comfy-core",
283
284
  "ver": "0.3.51",
284
- "Node name for S&R": "VAELoader",
285
285
  "models": [
286
286
  {
287
287
  "name": "qwen_image_vae.safetensors",
@@ -320,9 +320,9 @@
320
320
  }
321
321
  ],
322
322
  "properties": {
323
+ "Node name for S&R": "UNETLoader",
323
324
  "cnr_id": "comfy-core",
324
325
  "ver": "0.3.51",
325
- "Node name for S&R": "UNETLoader",
326
326
  "models": [
327
327
  {
328
328
  "name": "qwen_image_fp8_e4m3fn.safetensors",
@@ -374,9 +374,9 @@
374
374
  }
375
375
  ],
376
376
  "properties": {
377
+ "Node name for S&R": "VAEEncode",
377
378
  "cnr_id": "comfy-core",
378
- "ver": "0.3.51",
379
- "Node name for S&R": "VAEEncode"
379
+ "ver": "0.3.51"
380
380
  },
381
381
  "widgets_values": []
382
382
  },
@@ -430,9 +430,9 @@
430
430
  }
431
431
  ],
432
432
  "properties": {
433
+ "Node name for S&R": "CLIPLoader",
433
434
  "cnr_id": "comfy-core",
434
435
  "ver": "0.3.51",
435
- "Node name for S&R": "CLIPLoader",
436
436
  "models": [
437
437
  {
438
438
  "name": "qwen_2.5_vl_7b_fp8_scaled.safetensors",
@@ -477,9 +477,9 @@
477
477
  }
478
478
  ],
479
479
  "properties": {
480
+ "Node name for S&R": "LoadImage",
480
481
  "cnr_id": "comfy-core",
481
- "ver": "0.3.51",
482
- "Node name for S&R": "LoadImage"
482
+ "ver": "0.3.51"
483
483
  },
484
484
  "widgets_values": [
485
485
  "ComfyUI_00944_.png",
@@ -509,9 +509,9 @@
509
509
  ],
510
510
  "outputs": [],
511
511
  "properties": {
512
+ "Node name for S&R": "PreviewImage",
512
513
  "cnr_id": "comfy-core",
513
- "ver": "0.3.51",
514
- "Node name for S&R": "PreviewImage"
514
+ "ver": "0.3.51"
515
515
  },
516
516
  "widgets_values": []
517
517
  },
@@ -547,9 +547,9 @@
547
547
  }
548
548
  ],
549
549
  "properties": {
550
+ "Node name for S&R": "Canny",
550
551
  "cnr_id": "comfy-core",
551
- "ver": "0.3.51",
552
- "Node name for S&R": "Canny"
552
+ "ver": "0.3.51"
553
553
  },
554
554
  "widgets_values": [
555
555
  0.1,
@@ -619,9 +619,9 @@
619
619
  }
620
620
  ],
621
621
  "properties": {
622
+ "Node name for S&R": "LoraLoaderModelOnly",
622
623
  "cnr_id": "comfy-core",
623
624
  "ver": "0.3.51",
624
- "Node name for S&R": "LoraLoaderModelOnly",
625
625
  "models": [
626
626
  {
627
627
  "name": "Qwen-Image-Lightning-4steps-V1.0.safetensors",
@@ -668,9 +668,9 @@
668
668
  ],
669
669
  "title": "CLIP Text Encode (Positive Prompt)",
670
670
  "properties": {
671
+ "Node name for S&R": "CLIPTextEncode",
671
672
  "cnr_id": "comfy-core",
672
- "ver": "0.3.51",
673
- "Node name for S&R": "CLIPTextEncode"
673
+ "ver": "0.3.51"
674
674
  },
675
675
  "widgets_values": [
676
676
  "Conceptual makeup, a fairy girl with pink hair, pink shimmery scales dotted at the corners of her eyes, starry eyeshadow, makeup painting, thin eyebrows, three-dimensional features, a glossy finish, dazzling gold powder, silver fine glitter, a sense of layering and depth in the makeup, decorated with gold ornaments, pearlescent eyeshadow, dreamy makeup, soft pastels and subtle sparkles, a mysterious and fantasy-filled atmosphere, high-end makeup, dappled light on the face, soft lighting, optimal shadows, complex depth of field, dramatic lighting, clear focus, 8k, high quality, Fujifilm filter, surreal, a dreamy pastel wonderland, bright colors, a starry pink background, realistic.\n"
@@ -711,9 +711,9 @@
711
711
  ],
712
712
  "title": "CLIP Text Encode (Negative Prompt)",
713
713
  "properties": {
714
+ "Node name for S&R": "CLIPTextEncode",
714
715
  "cnr_id": "comfy-core",
715
- "ver": "0.3.51",
716
- "Node name for S&R": "CLIPTextEncode"
716
+ "ver": "0.3.51"
717
717
  },
718
718
  "widgets_values": [
719
719
  " "
@@ -768,9 +768,9 @@
768
768
  }
769
769
  ],
770
770
  "properties": {
771
+ "Node name for S&R": "KSampler",
771
772
  "cnr_id": "comfy-core",
772
- "ver": "0.3.51",
773
- "Node name for S&R": "KSampler"
773
+ "ver": "0.3.51"
774
774
  },
775
775
  "widgets_values": [
776
776
  91832422759220,
@@ -803,7 +803,7 @@
803
803
  "widget_ue_connectable": {}
804
804
  },
805
805
  "widgets_values": [
806
- "[Tutorial](https://docs.comfy.org/tutorials/image/qwen/qwen-image) | [教程](https://docs.comfy.org/zh-CN/tutorials/image/qwen/qwen-image)\n\n\n## Model links\n\nYou can find all the models on [Huggingface](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/tree/main) or [Modelscope](https://modelscope.cn/models/Comfy-Org/Qwen-Image_ComfyUI/files)\n\n**Diffusion model**\n\n- [qwen_image_fp8_e4m3fn.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/diffusion_models/qwen_image_fp8_e4m3fn.safetensors)\n\n**Model patches**\n\n- [qwen_image_canny_diffsynth_controlnet.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image-DiffSynth-ControlNets/resolve/main/split_files/model_patches/qwen_image_canny_diffsynth_controlnet.safetensors)\n- [qwen_image_depth_diffsynth_controlnet.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image-DiffSynth-ControlNets/resolve/main/split_files/model_patches/qwen_image_depth_diffsynth_controlnet.safetensors)\n- [qwen_image_inpaint_diffsynth_controlnet.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image-DiffSynth-ControlNets/resolve/main/split_files/model_patches/qwen_image_inpaint_diffsynth_controlnet.safetensors)\n\n**LoRA**\n\n- [Qwen-Image-Lightning-4steps-V1.0.safetensors](https://huggingface.co/lightx2v/Qwen-Image-Lightning/resolve/main/Qwen-Image-Lightning-4steps-V1.0.safetensors)\n\n**Text encoder**\n\n- [qwen_2.5_vl_7b_fp8_scaled.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/text_encoders/qwen_2.5_vl_7b_fp8_scaled.safetensors)\n\n**VAE**\n\n- [qwen_image_vae.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/vae/qwen_image_vae.safetensors)\n\n\nModel Storage Location\n\n```\n📂 ComfyUI/\n├── 📂 models/\n│ ├── 📂 diffusion_models/\n│ │ ├── qwen_image_fp8_e4m3fn.safetensors\n│ │ └── qwen_image_distill_full_fp8_e4m3fn.safetensors\n│ ├── 📂 loras/\n│ │ └── Qwen-Image-Lightning-8steps-V1.0.safetensors\n│ ├── 📂 model_patches/ # create one if you can't find it\n│ │ ├── qwen_image_depth_diffsynth_controlnet.safetensors\n│ │ ├── qwen_image_canny_diffsynth_controlnet.safetensors\n│ │ └── qwen_image_inpaint_diffsynth_controlnet.safetensors\n│ ├── 📂 vae/\n│ │ └── qwen_image_vae.safetensors\n│ └── 📂 text_encoders/\n│ └── qwen_2.5_vl_7b_fp8_scaled.safetensors\n```\n"
806
+ "[Tutorial](https://docs.comfy.org/tutorials/image/qwen/qwen-image) \n\n\n## Model links\n\nYou can find all the models on [Huggingface](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/tree/main) or [Modelscope](https://modelscope.cn/models/Comfy-Org/Qwen-Image_ComfyUI/files)\n\n**Diffusion model**\n\n- [qwen_image_fp8_e4m3fn.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/diffusion_models/qwen_image_fp8_e4m3fn.safetensors)\n\n**Model patches**\n\n- [qwen_image_canny_diffsynth_controlnet.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image-DiffSynth-ControlNets/resolve/main/split_files/model_patches/qwen_image_canny_diffsynth_controlnet.safetensors)\n- [qwen_image_depth_diffsynth_controlnet.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image-DiffSynth-ControlNets/resolve/main/split_files/model_patches/qwen_image_depth_diffsynth_controlnet.safetensors)\n- [qwen_image_inpaint_diffsynth_controlnet.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image-DiffSynth-ControlNets/resolve/main/split_files/model_patches/qwen_image_inpaint_diffsynth_controlnet.safetensors)\n\n**LoRA**\n\n- [Qwen-Image-Lightning-4steps-V1.0.safetensors](https://huggingface.co/lightx2v/Qwen-Image-Lightning/resolve/main/Qwen-Image-Lightning-4steps-V1.0.safetensors)\n\n**Text encoder**\n\n- [qwen_2.5_vl_7b_fp8_scaled.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/text_encoders/qwen_2.5_vl_7b_fp8_scaled.safetensors)\n\n**VAE**\n\n- [qwen_image_vae.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/vae/qwen_image_vae.safetensors)\n\n\nModel Storage Location\n\n```\n📂 ComfyUI/\n├── 📂 models/\n│ ├── 📂 diffusion_models/\n│ │ ├── qwen_image_fp8_e4m3fn.safetensors\n│ │ └── qwen_image_distill_full_fp8_e4m3fn.safetensors\n│ ├── 📂 loras/\n│ │ └── Qwen-Image-Lightning-8steps-V1.0.safetensors\n│ ├── 📂 model_patches/ # create one if you can't find it\n│ │ ├── qwen_image_depth_diffsynth_controlnet.safetensors\n│ │ ├── qwen_image_canny_diffsynth_controlnet.safetensors\n│ │ └── qwen_image_inpaint_diffsynth_controlnet.safetensors\n│ ├── 📂 vae/\n│ │ └── qwen_image_vae.safetensors\n│ └── 📂 text_encoders/\n│ └── qwen_2.5_vl_7b_fp8_scaled.safetensors\n```\n"
807
807
  ],
808
808
  "color": "#432",
809
809
  "bgcolor": "#653"
@@ -1041,13 +1041,13 @@
1041
1041
  "config": {},
1042
1042
  "extra": {
1043
1043
  "ds": {
1044
- "scale": 0.48559562289012265,
1044
+ "scale": 0.5875707036970488,
1045
1045
  "offset": [
1046
- 1846.044139609573,
1047
- 391.24067543168553
1046
+ 181.02254312486346,
1047
+ 275.5613649922377
1048
1048
  ]
1049
1049
  },
1050
- "frontendVersion": "1.26.6"
1050
+ "frontendVersion": "1.28.6"
1051
1051
  },
1052
1052
  "version": 0.4
1053
1053
  }
@@ -758,32 +758,6 @@
758
758
  "color": "#232",
759
759
  "bgcolor": "#353"
760
760
  },
761
- {
762
- "id": 99,
763
- "type": "MarkdownNote",
764
- "pos": [
765
- -830,
766
- -10
767
- ],
768
- "size": [
769
- 540,
770
- 550
771
- ],
772
- "flags": {},
773
- "order": 6,
774
- "mode": 0,
775
- "inputs": [],
776
- "outputs": [],
777
- "title": "Model links",
778
- "properties": {
779
- "widget_ue_connectable": {}
780
- },
781
- "widgets_values": [
782
- "[Tutorial](https://docs.comfy.org/tutorials/image/qwen/qwen-image-edit) | [教程](https://docs.comfy.org/zh-CN/tutorials/image/qwen/qwen-image-edit)\n\n\n## Model links\n\nYou can find all the models on [Comfy-Org/Qwen-Image_ComfyUI](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/tree/main) and [Comfy-Org/Qwen-Image-Edit_ComfyUI](https://huggingface.co/Comfy-Org/Qwen-Image-Edit_ComfyUI) \n\n**Diffusion model**\n\n- [qwen_image_edit_fp8_e4m3fn.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image-Edit_ComfyUI/resolve/main/split_files/diffusion_models/qwen_image_edit_fp8_e4m3fn.safetensors)\n\n**LoRA**\n\n- [Qwen-Image-Lightning-4steps-V1.0.safetensors](https://huggingface.co/lightx2v/Qwen-Image-Lightning/resolve/main/Qwen-Image-Lightning-4steps-V1.0.safetensors)\n\n**Text encoder**\n\n- [qwen_2.5_vl_7b_fp8_scaled.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/text_encoders/qwen_2.5_vl_7b_fp8_scaled.safetensors)\n\n**VAE**\n\n- [qwen_image_vae.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/vae/qwen_image_vae.safetensors)\n\nModel Storage Location\n\n```\n📂 ComfyUI/\n├── 📂 models/\n│ ├── 📂 diffusion_models/\n│ │ └── qwen_image_edit_fp8_e4m3fn.safetensors\n│ ├── 📂 loras/\n│ │ └── Qwen-Image-Lightning-4steps-V1.0.safetensors\n│ ├── 📂 vae/\n│ │ └── qwen_image_vae.safetensors\n│ └── 📂 text_encoders/\n│ └── qwen_2.5_vl_7b_fp8_scaled.safetensors\n```\n"
783
- ],
784
- "color": "#432",
785
- "bgcolor": "#653"
786
- },
787
761
  {
788
762
  "id": 93,
789
763
  "type": "ImageScaleToTotalPixels",
@@ -838,6 +812,32 @@
838
812
  "lanczos",
839
813
  1
840
814
  ]
815
+ },
816
+ {
817
+ "id": 99,
818
+ "type": "MarkdownNote",
819
+ "pos": [
820
+ -830,
821
+ -10
822
+ ],
823
+ "size": [
824
+ 540,
825
+ 550
826
+ ],
827
+ "flags": {},
828
+ "order": 6,
829
+ "mode": 0,
830
+ "inputs": [],
831
+ "outputs": [],
832
+ "title": "Model links",
833
+ "properties": {
834
+ "widget_ue_connectable": {}
835
+ },
836
+ "widgets_values": [
837
+ "[Tutorial](https://docs.comfy.org/tutorials/image/qwen/qwen-image-edit) \n\n## Model links\n\nYou can find all the models on [Comfy-Org/Qwen-Image_ComfyUI](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/tree/main) and [Comfy-Org/Qwen-Image-Edit_ComfyUI](https://huggingface.co/Comfy-Org/Qwen-Image-Edit_ComfyUI) \n\n**Diffusion model**\n\n- [qwen_image_edit_fp8_e4m3fn.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image-Edit_ComfyUI/resolve/main/split_files/diffusion_models/qwen_image_edit_fp8_e4m3fn.safetensors)\n\n**LoRA**\n\n- [Qwen-Image-Lightning-4steps-V1.0.safetensors](https://huggingface.co/lightx2v/Qwen-Image-Lightning/resolve/main/Qwen-Image-Lightning-4steps-V1.0.safetensors)\n\n**Text encoder**\n\n- [qwen_2.5_vl_7b_fp8_scaled.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/text_encoders/qwen_2.5_vl_7b_fp8_scaled.safetensors)\n\n**VAE**\n\n- [qwen_image_vae.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/vae/qwen_image_vae.safetensors)\n\nModel Storage Location\n\n```\n📂 ComfyUI/\n├── 📂 models/\n│ ├── 📂 diffusion_models/\n│ │ └── qwen_image_edit_fp8_e4m3fn.safetensors\n│ ├── 📂 loras/\n│ │ └── Qwen-Image-Lightning-4steps-V1.0.safetensors\n│ ├── 📂 vae/\n│ │ └── qwen_image_vae.safetensors\n│ └── 📂 text_encoders/\n│ └── qwen_2.5_vl_7b_fp8_scaled.safetensors\n```\n"
838
+ ],
839
+ "color": "#432",
840
+ "bgcolor": "#653"
841
841
  }
842
842
  ],
843
843
  "links": [
@@ -1040,11 +1040,11 @@
1040
1040
  "ds": {
1041
1041
  "scale": 0.7686421116383851,
1042
1042
  "offset": [
1043
- 828.5432421393461,
1044
- 99.771400280708
1043
+ 1227.7405266804722,
1044
+ 190.10419977044714
1045
1045
  ]
1046
1046
  },
1047
- "frontendVersion": "1.26.5",
1047
+ "frontendVersion": "1.28.6",
1048
1048
  "ue_links": [],
1049
1049
  "links_added_by_ue": [],
1050
1050
  "VHS_latentpreview": false,