comfyui-workflow-templates 0.1.94__py3-none-any.whl → 0.1.95__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of comfyui-workflow-templates might be problematic. Click here for more details.

Files changed (129) hide show
  1. comfyui_workflow_templates/templates/3d_hunyuan3d_image_to_model.json +26 -26
  2. comfyui_workflow_templates/templates/3d_hunyuan3d_multiview_to_model.json +35 -35
  3. comfyui_workflow_templates/templates/3d_hunyuan3d_multiview_to_model_turbo.json +37 -37
  4. comfyui_workflow_templates/templates/api_bfl_flux_1_kontext_max_image.json +57 -81
  5. comfyui_workflow_templates/templates/api_bfl_flux_1_kontext_multiple_images_input.json +60 -84
  6. comfyui_workflow_templates/templates/api_bfl_flux_1_kontext_pro_image.json +57 -81
  7. comfyui_workflow_templates/templates/api_bfl_flux_pro_t2i.json +28 -26
  8. comfyui_workflow_templates/templates/api_bytedance_flf2v.json +11 -11
  9. comfyui_workflow_templates/templates/api_bytedance_image_to_video.json +34 -34
  10. comfyui_workflow_templates/templates/api_bytedance_text_to_video.json +39 -40
  11. comfyui_workflow_templates/templates/api_google_gemini.json +6 -7
  12. comfyui_workflow_templates/templates/api_hailuo_minimax_i2v.json +12 -3
  13. comfyui_workflow_templates/templates/api_hailuo_minimax_t2v.json +28 -28
  14. comfyui_workflow_templates/templates/api_hailuo_minimax_video.json +30 -30
  15. comfyui_workflow_templates/templates/api_ideogram_v3_t2i.json +32 -18
  16. comfyui_workflow_templates/templates/api_kling_effects.json +28 -26
  17. comfyui_workflow_templates/templates/api_kling_flf.json +32 -30
  18. comfyui_workflow_templates/templates/api_kling_i2v.json +34 -34
  19. comfyui_workflow_templates/templates/api_luma_i2v.json +93 -110
  20. comfyui_workflow_templates/templates/api_luma_photon_i2i.json +59 -50
  21. comfyui_workflow_templates/templates/api_luma_photon_style_ref.json +131 -124
  22. comfyui_workflow_templates/templates/api_luma_t2v.json +59 -50
  23. comfyui_workflow_templates/templates/api_openai_dall_e_2_t2i.json +8 -29
  24. comfyui_workflow_templates/templates/api_openai_image_1_i2i.json +12 -33
  25. comfyui_workflow_templates/templates/api_openai_image_1_inpaint.json +30 -51
  26. comfyui_workflow_templates/templates/api_openai_image_1_multi_inputs.json +10 -33
  27. comfyui_workflow_templates/templates/api_openai_image_1_t2i.json +42 -63
  28. comfyui_workflow_templates/templates/api_openai_sora_video.json +37 -38
  29. comfyui_workflow_templates/templates/api_pika_i2v.json +27 -27
  30. comfyui_workflow_templates/templates/api_pika_scene.json +12 -3
  31. comfyui_workflow_templates/templates/api_pixverse_i2v.json +31 -36
  32. comfyui_workflow_templates/templates/api_pixverse_t2v.json +20 -16
  33. comfyui_workflow_templates/templates/api_pixverse_template_i2v.json +39 -35
  34. comfyui_workflow_templates/templates/api_recraft_image_gen_with_color_control.json +6 -6
  35. comfyui_workflow_templates/templates/api_recraft_image_gen_with_style_control.json +212 -199
  36. comfyui_workflow_templates/templates/api_recraft_vector_gen.json +78 -69
  37. comfyui_workflow_templates/templates/api_rodin_gen2.json +30 -30
  38. comfyui_workflow_templates/templates/api_rodin_image_to_model.json +55 -55
  39. comfyui_workflow_templates/templates/api_rodin_multiview_to_model.json +188 -132
  40. comfyui_workflow_templates/templates/api_runway_first_last_frame.json +4 -4
  41. comfyui_workflow_templates/templates/api_runway_gen3a_turbo_image_to_video.json +30 -31
  42. comfyui_workflow_templates/templates/api_runway_gen4_turo_image_to_video.json +29 -30
  43. comfyui_workflow_templates/templates/api_runway_reference_to_image.json +31 -32
  44. comfyui_workflow_templates/templates/api_runway_text_to_image.json +17 -17
  45. comfyui_workflow_templates/templates/api_stability_ai_audio_inpaint.json +18 -18
  46. comfyui_workflow_templates/templates/api_stability_ai_audio_to_audio.json +31 -31
  47. comfyui_workflow_templates/templates/api_stability_ai_i2i.json +34 -34
  48. comfyui_workflow_templates/templates/api_stability_ai_sd3.5_i2i.json +21 -19
  49. comfyui_workflow_templates/templates/api_stability_ai_sd3.5_t2i.json +35 -35
  50. comfyui_workflow_templates/templates/api_stability_ai_stable_image_ultra_t2i.json +11 -9
  51. comfyui_workflow_templates/templates/api_tripo_image_to_model.json +90 -92
  52. comfyui_workflow_templates/templates/api_tripo_multiview_to_model.json +241 -241
  53. comfyui_workflow_templates/templates/api_tripo_text_to_model.json +102 -102
  54. comfyui_workflow_templates/templates/api_veo2_i2v.json +31 -28
  55. comfyui_workflow_templates/templates/api_veo3.json +30 -30
  56. comfyui_workflow_templates/templates/api_vidu_text_to_video.json +2 -2
  57. comfyui_workflow_templates/templates/api_wan_image_to_video.json +41 -42
  58. comfyui_workflow_templates/templates/api_wan_text_to_image .json +140 -0
  59. comfyui_workflow_templates/templates/api_wan_text_to_video.json +38 -45
  60. comfyui_workflow_templates/templates/audio_ace_step_1_m2m_editing.json +84 -84
  61. comfyui_workflow_templates/templates/audio_ace_step_1_t2a_instrumentals.json +60 -60
  62. comfyui_workflow_templates/templates/audio_ace_step_1_t2a_song.json +60 -60
  63. comfyui_workflow_templates/templates/esrgan_example.json +24 -30
  64. comfyui_workflow_templates/templates/flux1_dev_uso_reference_image_gen.json +215 -210
  65. comfyui_workflow_templates/templates/flux1_krea_dev.json +3 -3
  66. comfyui_workflow_templates/templates/flux_kontext_dev_basic.json +151 -231
  67. comfyui_workflow_templates/templates/flux_redux_model_example.json +108 -120
  68. comfyui_workflow_templates/templates/flux_schnell_full_text_to_image.json +21 -29
  69. comfyui_workflow_templates/templates/hidream_e1_1.json +179 -209
  70. comfyui_workflow_templates/templates/hidream_e1_full.json +33 -39
  71. comfyui_workflow_templates/templates/hidream_i1_dev.json +15 -15
  72. comfyui_workflow_templates/templates/hidream_i1_fast.json +15 -15
  73. comfyui_workflow_templates/templates/hidream_i1_full.json +17 -16
  74. comfyui_workflow_templates/templates/hiresfix_esrgan_workflow.json +31 -37
  75. comfyui_workflow_templates/templates/hiresfix_latent_workflow.json +84 -88
  76. comfyui_workflow_templates/templates/image2image.json +30 -30
  77. comfyui_workflow_templates/templates/image_chroma1_radiance_text_to_image.json +60 -60
  78. comfyui_workflow_templates/templates/image_lotus_depth_v1_1.json +25 -31
  79. comfyui_workflow_templates/templates/image_netayume_lumina_t2i-1.webp +0 -0
  80. comfyui_workflow_templates/templates/image_netayume_lumina_t2i.json +597 -0
  81. comfyui_workflow_templates/templates/image_omnigen2_image_edit.json +55 -62
  82. comfyui_workflow_templates/templates/image_omnigen2_t2i.json +26 -33
  83. comfyui_workflow_templates/templates/image_qwen_image.json +40 -40
  84. comfyui_workflow_templates/templates/image_qwen_image_controlnet_patch.json +32 -32
  85. comfyui_workflow_templates/templates/image_qwen_image_edit.json +29 -29
  86. comfyui_workflow_templates/templates/image_qwen_image_edit_2509.json +127 -127
  87. comfyui_workflow_templates/templates/image_qwen_image_instantx_controlnet.json +56 -55
  88. comfyui_workflow_templates/templates/image_qwen_image_instantx_inpainting_controlnet.json +108 -107
  89. comfyui_workflow_templates/templates/image_qwen_image_union_control_lora.json +5 -5
  90. comfyui_workflow_templates/templates/index.es.json +24 -0
  91. comfyui_workflow_templates/templates/index.fr.json +24 -0
  92. comfyui_workflow_templates/templates/index.ja.json +24 -0
  93. comfyui_workflow_templates/templates/index.json +11 -0
  94. comfyui_workflow_templates/templates/index.ko.json +24 -0
  95. comfyui_workflow_templates/templates/index.ru.json +24 -0
  96. comfyui_workflow_templates/templates/index.zh-TW.json +24 -0
  97. comfyui_workflow_templates/templates/index.zh.json +24 -0
  98. comfyui_workflow_templates/templates/inpaint_example.json +70 -72
  99. comfyui_workflow_templates/templates/inpaint_model_outpainting.json +4 -4
  100. comfyui_workflow_templates/templates/latent_upscale_different_prompt_model.json +179 -185
  101. comfyui_workflow_templates/templates/sdxlturbo_example.json +308 -162
  102. comfyui_workflow_templates/templates/video_wan2.1_fun_camera_v1.1_1.3B.json +89 -62
  103. comfyui_workflow_templates/templates/video_wan2.1_fun_camera_v1.1_14B.json +8 -4
  104. comfyui_workflow_templates/templates/video_wan2_2_14B_animate.json +46 -44
  105. comfyui_workflow_templates/templates/video_wan2_2_14B_flf2v.json +38 -38
  106. comfyui_workflow_templates/templates/video_wan2_2_14B_fun_camera.json +58 -54
  107. comfyui_workflow_templates/templates/video_wan2_2_14B_fun_control.json +36 -36
  108. comfyui_workflow_templates/templates/video_wan2_2_14B_fun_inpaint.json +26 -26
  109. comfyui_workflow_templates/templates/video_wan2_2_14B_i2v.json +4 -4
  110. comfyui_workflow_templates/templates/video_wan2_2_14B_s2v.json +33 -29
  111. comfyui_workflow_templates/templates/video_wan2_2_14B_t2v (2).json +1954 -0
  112. comfyui_workflow_templates/templates/video_wan2_2_5B_fun_control.json +29 -29
  113. comfyui_workflow_templates/templates/video_wan2_2_5B_fun_inpaint.json +25 -25
  114. comfyui_workflow_templates/templates/video_wan2_2_5B_ti2v.json +49 -49
  115. comfyui_workflow_templates/templates/video_wan_ati.json +49 -49
  116. comfyui_workflow_templates/templates/video_wan_vace_14B_ref2v.json +47 -61
  117. comfyui_workflow_templates/templates/video_wan_vace_14B_t2v.json +2 -2
  118. comfyui_workflow_templates/templates/video_wan_vace_14B_v2v.json +55 -55
  119. comfyui_workflow_templates/templates/video_wan_vace_flf2v.json +40 -56
  120. comfyui_workflow_templates/templates/video_wan_vace_inpainting.json +72 -72
  121. comfyui_workflow_templates/templates/video_wan_vace_outpainting.json +211 -237
  122. comfyui_workflow_templates/templates/wan2.1_flf2v_720_f16.json +84 -92
  123. comfyui_workflow_templates/templates/wan2.1_fun_control.json +51 -27
  124. comfyui_workflow_templates/templates/wan2.1_fun_inp.json +43 -17
  125. {comfyui_workflow_templates-0.1.94.dist-info → comfyui_workflow_templates-0.1.95.dist-info}/METADATA +1 -1
  126. {comfyui_workflow_templates-0.1.94.dist-info → comfyui_workflow_templates-0.1.95.dist-info}/RECORD +129 -125
  127. {comfyui_workflow_templates-0.1.94.dist-info → comfyui_workflow_templates-0.1.95.dist-info}/WHEEL +0 -0
  128. {comfyui_workflow_templates-0.1.94.dist-info → comfyui_workflow_templates-0.1.95.dist-info}/licenses/LICENSE +0 -0
  129. {comfyui_workflow_templates-0.1.94.dist-info → comfyui_workflow_templates-0.1.95.dist-info}/top_level.txt +0 -0
@@ -31,9 +31,9 @@
31
31
  }
32
32
  ],
33
33
  "properties": {
34
+ "Node name for S&R": "CLIPLoader",
34
35
  "cnr_id": "comfy-core",
35
36
  "ver": "0.3.51",
36
- "Node name for S&R": "CLIPLoader",
37
37
  "models": [
38
38
  {
39
39
  "name": "qwen_2.5_vl_7b_fp8_scaled.safetensors",
@@ -74,9 +74,9 @@
74
74
  }
75
75
  ],
76
76
  "properties": {
77
+ "Node name for S&R": "UNETLoader",
77
78
  "cnr_id": "comfy-core",
78
79
  "ver": "0.3.51",
79
- "Node name for S&R": "UNETLoader",
80
80
  "models": [
81
81
  {
82
82
  "name": "qwen_image_fp8_e4m3fn.safetensors",
@@ -123,9 +123,9 @@
123
123
  ],
124
124
  "title": "CLIP Text Encode (Negative Prompt)",
125
125
  "properties": {
126
+ "Node name for S&R": "CLIPTextEncode",
126
127
  "cnr_id": "comfy-core",
127
- "ver": "0.3.51",
128
- "Node name for S&R": "CLIPTextEncode"
128
+ "ver": "0.3.51"
129
129
  },
130
130
  "widgets_values": [
131
131
  " "
@@ -158,9 +158,9 @@
158
158
  }
159
159
  ],
160
160
  "properties": {
161
+ "Node name for S&R": "ControlNetLoader",
161
162
  "cnr_id": "comfy-core",
162
163
  "ver": "0.3.51",
163
- "Node name for S&R": "ControlNetLoader",
164
164
  "models": [
165
165
  {
166
166
  "name": "Qwen-Image-InstantX-ControlNet-Inpainting.safetensors",
@@ -201,9 +201,9 @@
201
201
  }
202
202
  ],
203
203
  "properties": {
204
+ "Node name for S&R": "VAELoader",
204
205
  "cnr_id": "comfy-core",
205
206
  "ver": "0.3.51",
206
- "Node name for S&R": "VAELoader",
207
207
  "models": [
208
208
  {
209
209
  "name": "qwen_image_vae.safetensors",
@@ -247,9 +247,9 @@
247
247
  }
248
248
  ],
249
249
  "properties": {
250
+ "Node name for S&R": "ModelSamplingAuraFlow",
250
251
  "cnr_id": "comfy-core",
251
- "ver": "0.3.51",
252
- "Node name for S&R": "ModelSamplingAuraFlow"
252
+ "ver": "0.3.51"
253
253
  },
254
254
  "widgets_values": [
255
255
  3.1000000000000005
@@ -318,9 +318,9 @@
318
318
  }
319
319
  ],
320
320
  "properties": {
321
+ "Node name for S&R": "ControlNetInpaintingAliMamaApply",
321
322
  "cnr_id": "comfy-core",
322
- "ver": "0.3.59",
323
- "Node name for S&R": "ControlNetInpaintingAliMamaApply"
323
+ "ver": "0.3.59"
324
324
  },
325
325
  "widgets_values": [
326
326
  1,
@@ -361,9 +361,9 @@
361
361
  ],
362
362
  "title": "CLIP Text Encode (Positive Prompt)",
363
363
  "properties": {
364
+ "Node name for S&R": "CLIPTextEncode",
364
365
  "cnr_id": "comfy-core",
365
- "ver": "0.3.51",
366
- "Node name for S&R": "CLIPTextEncode"
366
+ "ver": "0.3.51"
367
367
  },
368
368
  "widgets_values": [
369
369
  "The Queen, on a throne, surrounded by Knights, HD, Realistic, Octane Render, Unreal engine"
@@ -474,9 +474,9 @@
474
474
  }
475
475
  ],
476
476
  "properties": {
477
+ "Node name for S&R": "LoadImage",
477
478
  "cnr_id": "comfy-core",
478
- "ver": "0.3.51",
479
- "Node name for S&R": "LoadImage"
479
+ "ver": "0.3.51"
480
480
  },
481
481
  "widgets_values": [
482
482
  "clipspace/clipspace-painted-masked-19830639.png [input]",
@@ -521,9 +521,9 @@
521
521
  }
522
522
  ],
523
523
  "properties": {
524
+ "Node name for S&R": "VAEEncode",
524
525
  "cnr_id": "comfy-core",
525
- "ver": "0.3.51",
526
- "Node name for S&R": "VAEEncode"
526
+ "ver": "0.3.51"
527
527
  },
528
528
  "widgets_values": []
529
529
  },
@@ -565,9 +565,9 @@
565
565
  }
566
566
  ],
567
567
  "properties": {
568
+ "Node name for S&R": "SetLatentNoiseMask",
568
569
  "cnr_id": "comfy-core",
569
- "ver": "0.3.59",
570
- "Node name for S&R": "SetLatentNoiseMask"
570
+ "ver": "0.3.59"
571
571
  },
572
572
  "widgets_values": []
573
573
  },
@@ -594,9 +594,9 @@
594
594
  ],
595
595
  "outputs": [],
596
596
  "properties": {
597
+ "Node name for S&R": "MaskPreview",
597
598
  "cnr_id": "comfy-core",
598
- "ver": "0.3.59",
599
- "Node name for S&R": "MaskPreview"
599
+ "ver": "0.3.59"
600
600
  },
601
601
  "widgets_values": []
602
602
  },
@@ -651,9 +651,9 @@
651
651
  }
652
652
  ],
653
653
  "properties": {
654
+ "Node name for S&R": "CLIPLoader",
654
655
  "cnr_id": "comfy-core",
655
656
  "ver": "0.3.51",
656
- "Node name for S&R": "CLIPLoader",
657
657
  "models": [
658
658
  {
659
659
  "name": "qwen_2.5_vl_7b_fp8_scaled.safetensors",
@@ -694,9 +694,9 @@
694
694
  }
695
695
  ],
696
696
  "properties": {
697
+ "Node name for S&R": "UNETLoader",
697
698
  "cnr_id": "comfy-core",
698
699
  "ver": "0.3.51",
699
- "Node name for S&R": "UNETLoader",
700
700
  "models": [
701
701
  {
702
702
  "name": "qwen_image_fp8_e4m3fn.safetensors",
@@ -743,9 +743,9 @@
743
743
  ],
744
744
  "title": "CLIP Text Encode (Negative Prompt)",
745
745
  "properties": {
746
+ "Node name for S&R": "CLIPTextEncode",
746
747
  "cnr_id": "comfy-core",
747
- "ver": "0.3.51",
748
- "Node name for S&R": "CLIPTextEncode"
748
+ "ver": "0.3.51"
749
749
  },
750
750
  "widgets_values": [
751
751
  " "
@@ -778,9 +778,9 @@
778
778
  }
779
779
  ],
780
780
  "properties": {
781
+ "Node name for S&R": "ControlNetLoader",
781
782
  "cnr_id": "comfy-core",
782
783
  "ver": "0.3.51",
783
- "Node name for S&R": "ControlNetLoader",
784
784
  "models": [
785
785
  {
786
786
  "name": "Qwen-Image-InstantX-ControlNet-Inpainting.safetensors",
@@ -821,9 +821,9 @@
821
821
  }
822
822
  ],
823
823
  "properties": {
824
+ "Node name for S&R": "VAELoader",
824
825
  "cnr_id": "comfy-core",
825
826
  "ver": "0.3.51",
826
- "Node name for S&R": "VAELoader",
827
827
  "models": [
828
828
  {
829
829
  "name": "qwen_image_vae.safetensors",
@@ -899,9 +899,9 @@
899
899
  }
900
900
  ],
901
901
  "properties": {
902
+ "Node name for S&R": "ControlNetInpaintingAliMamaApply",
902
903
  "cnr_id": "comfy-core",
903
- "ver": "0.3.59",
904
- "Node name for S&R": "ControlNetInpaintingAliMamaApply"
904
+ "ver": "0.3.59"
905
905
  },
906
906
  "widgets_values": [
907
907
  1,
@@ -940,9 +940,9 @@
940
940
  }
941
941
  ],
942
942
  "properties": {
943
+ "Node name for S&R": "ModelSamplingAuraFlow",
943
944
  "cnr_id": "comfy-core",
944
- "ver": "0.3.51",
945
- "Node name for S&R": "ModelSamplingAuraFlow"
945
+ "ver": "0.3.51"
946
946
  },
947
947
  "widgets_values": [
948
948
  3.1000000000000005
@@ -986,9 +986,9 @@
986
986
  }
987
987
  ],
988
988
  "properties": {
989
+ "Node name for S&R": "VAEDecode",
989
990
  "cnr_id": "comfy-core",
990
- "ver": "0.3.51",
991
- "Node name for S&R": "VAEDecode"
991
+ "ver": "0.3.51"
992
992
  },
993
993
  "widgets_values": []
994
994
  },
@@ -1025,9 +1025,9 @@
1025
1025
  ],
1026
1026
  "title": "CLIP Text Encode (Positive Prompt)",
1027
1027
  "properties": {
1028
+ "Node name for S&R": "CLIPTextEncode",
1028
1029
  "cnr_id": "comfy-core",
1029
- "ver": "0.3.51",
1030
- "Node name for S&R": "CLIPTextEncode"
1030
+ "ver": "0.3.51"
1031
1031
  },
1032
1032
  "widgets_values": [
1033
1033
  "The king, on a throne, surrounded by Knights, HD, Realistic, Octane Render, Unreal engine"
@@ -1068,9 +1068,9 @@
1068
1068
  }
1069
1069
  ],
1070
1070
  "properties": {
1071
+ "Node name for S&R": "ImageScaleToMaxDimension",
1071
1072
  "cnr_id": "comfy-core",
1072
- "ver": "0.3.59",
1073
- "Node name for S&R": "ImageScaleToMaxDimension"
1073
+ "ver": "0.3.59"
1074
1074
  },
1075
1075
  "widgets_values": [
1076
1076
  "area",
@@ -1108,9 +1108,9 @@
1108
1108
  }
1109
1109
  ],
1110
1110
  "properties": {
1111
+ "Node name for S&R": "LoraLoaderModelOnly",
1111
1112
  "cnr_id": "comfy-core",
1112
1113
  "ver": "0.3.51",
1113
- "Node name for S&R": "LoraLoaderModelOnly",
1114
1114
  "models": [
1115
1115
  {
1116
1116
  "name": "Qwen-Image-Lightning-4steps-V1.0.safetensors",
@@ -1155,9 +1155,9 @@
1155
1155
  }
1156
1156
  ],
1157
1157
  "properties": {
1158
+ "Node name for S&R": "LoraLoaderModelOnly",
1158
1159
  "cnr_id": "comfy-core",
1159
1160
  "ver": "0.3.51",
1160
- "Node name for S&R": "LoraLoaderModelOnly",
1161
1161
  "models": [
1162
1162
  {
1163
1163
  "name": "Qwen-Image-Lightning-4steps-V1.0.safetensors",
@@ -1205,6 +1205,7 @@
1205
1205
  }
1206
1206
  ],
1207
1207
  "properties": {
1208
+ "proxyWidgets": [],
1208
1209
  "cnr_id": "comfy-core",
1209
1210
  "ver": "0.3.59"
1210
1211
  },
@@ -1251,9 +1252,9 @@
1251
1252
  }
1252
1253
  ],
1253
1254
  "properties": {
1255
+ "Node name for S&R": "VAEDecode",
1254
1256
  "cnr_id": "comfy-core",
1255
- "ver": "0.3.51",
1256
- "Node name for S&R": "VAEDecode"
1257
+ "ver": "0.3.51"
1257
1258
  },
1258
1259
  "widgets_values": []
1259
1260
  },
@@ -1299,9 +1300,9 @@
1299
1300
  }
1300
1301
  ],
1301
1302
  "properties": {
1303
+ "Node name for S&R": "ImageCompositeMasked",
1302
1304
  "cnr_id": "comfy-core",
1303
- "ver": "0.3.59",
1304
- "Node name for S&R": "ImageCompositeMasked"
1305
+ "ver": "0.3.59"
1305
1306
  },
1306
1307
  "widgets_values": [
1307
1308
  0,
@@ -1332,9 +1333,9 @@
1332
1333
  ],
1333
1334
  "outputs": [],
1334
1335
  "properties": {
1336
+ "Node name for S&R": "MaskPreview",
1335
1337
  "cnr_id": "comfy-core",
1336
- "ver": "0.3.59",
1337
- "Node name for S&R": "MaskPreview"
1338
+ "ver": "0.3.59"
1338
1339
  },
1339
1340
  "widgets_values": []
1340
1341
  },
@@ -1361,9 +1362,9 @@
1361
1362
  ],
1362
1363
  "outputs": [],
1363
1364
  "properties": {
1365
+ "Node name for S&R": "PreviewImage",
1364
1366
  "cnr_id": "comfy-core",
1365
- "ver": "0.3.59",
1366
- "Node name for S&R": "PreviewImage"
1367
+ "ver": "0.3.59"
1367
1368
  },
1368
1369
  "widgets_values": []
1369
1370
  },
@@ -1397,9 +1398,9 @@
1397
1398
  }
1398
1399
  ],
1399
1400
  "properties": {
1401
+ "Node name for S&R": "LoadImage",
1400
1402
  "cnr_id": "comfy-core",
1401
- "ver": "0.3.51",
1402
- "Node name for S&R": "LoadImage"
1403
+ "ver": "0.3.51"
1403
1404
  },
1404
1405
  "widgets_values": [
1405
1406
  "ComfyUI_01396_.png",
@@ -1439,9 +1440,9 @@
1439
1440
  }
1440
1441
  ],
1441
1442
  "properties": {
1443
+ "Node name for S&R": "ImageScaleToMaxDimension",
1442
1444
  "cnr_id": "comfy-core",
1443
- "ver": "0.3.59",
1444
- "Node name for S&R": "ImageScaleToMaxDimension"
1445
+ "ver": "0.3.59"
1445
1446
  },
1446
1447
  "widgets_values": [
1447
1448
  "area",
@@ -1490,9 +1491,9 @@
1490
1491
  }
1491
1492
  ],
1492
1493
  "properties": {
1494
+ "Node name for S&R": "ImageCompositeMasked",
1493
1495
  "cnr_id": "comfy-core",
1494
- "ver": "0.3.59",
1495
- "Node name for S&R": "ImageCompositeMasked"
1496
+ "ver": "0.3.59"
1496
1497
  },
1497
1498
  "widgets_values": [
1498
1499
  0,
@@ -1538,9 +1539,9 @@
1538
1539
  }
1539
1540
  ],
1540
1541
  "properties": {
1542
+ "Node name for S&R": "VAEEncode",
1541
1543
  "cnr_id": "comfy-core",
1542
- "ver": "0.3.51",
1543
- "Node name for S&R": "VAEEncode"
1544
+ "ver": "0.3.51"
1544
1545
  },
1545
1546
  "widgets_values": []
1546
1547
  },
@@ -1585,9 +1586,9 @@
1585
1586
  }
1586
1587
  ],
1587
1588
  "properties": {
1589
+ "Node name for S&R": "ImagePadForOutpaint",
1588
1590
  "cnr_id": "comfy-core",
1589
- "ver": "0.3.59",
1590
- "Node name for S&R": "ImagePadForOutpaint"
1591
+ "ver": "0.3.59"
1591
1592
  },
1592
1593
  "widgets_values": [
1593
1594
  200,
@@ -1644,9 +1645,9 @@
1644
1645
  }
1645
1646
  ],
1646
1647
  "properties": {
1648
+ "Node name for S&R": "KSampler",
1647
1649
  "cnr_id": "comfy-core",
1648
- "ver": "0.3.51",
1649
- "Node name for S&R": "KSampler"
1650
+ "ver": "0.3.51"
1650
1651
  },
1651
1652
  "widgets_values": [
1652
1653
  276542274578345,
@@ -1789,9 +1790,9 @@
1789
1790
  }
1790
1791
  ],
1791
1792
  "properties": {
1793
+ "Node name for S&R": "KSampler",
1792
1794
  "cnr_id": "comfy-core",
1793
- "ver": "0.3.51",
1794
- "Node name for S&R": "KSampler"
1795
+ "ver": "0.3.51"
1795
1796
  },
1796
1797
  "widgets_values": [
1797
1798
  463031436157575,
@@ -1864,50 +1865,50 @@
1864
1865
  ]
1865
1866
  },
1866
1867
  {
1867
- "id": 78,
1868
- "type": "MarkdownNote",
1868
+ "id": 68,
1869
+ "type": "Note",
1869
1870
  "pos": [
1870
- -700,
1871
- -70
1871
+ 850,
1872
+ -300
1872
1873
  ],
1873
1874
  "size": [
1874
- 541.36865234375,
1875
- 579.70263671875
1875
+ 310,
1876
+ 90
1876
1877
  ],
1877
1878
  "flags": {},
1878
1879
  "order": 15,
1879
1880
  "mode": 0,
1880
1881
  "inputs": [],
1881
1882
  "outputs": [],
1882
- "title": "Model links",
1883
- "properties": {
1884
- "widget_ue_connectable": {}
1885
- },
1883
+ "properties": {},
1886
1884
  "widgets_values": [
1887
- "[Tutorial](https://docs.comfy.org/tutorials/image/qwen/qwen-image) | [教程](https://docs.comfy.org/zh-CN/tutorials/image/qwen/qwen-image)\n\n\n## Model links\n\nYou can find all the models on [Huggingface](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/tree/main) or [Modelscope](https://modelscope.cn/models/Comfy-Org/Qwen-Image_ComfyUI/files)\n\n**Diffusion model**\n\n- [qwen_image_fp8_e4m3fn.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/diffusion_models/qwen_image_fp8_e4m3fn.safetensors)\n\n**ControlNet**\n\n- [Qwen-Image-InstantX-ControlNet-Inpainting.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image-InstantX-ControlNets/resolve/main/split_files/controlnet/Qwen-Image-InstantX-ControlNet-Inpainting.safetensors)\n\n\n**LoRA**\n\n- [Qwen-Image-Lightning-4steps-V1.0.safetensors](https://huggingface.co/lightx2v/Qwen-Image-Lightning/resolve/main/Qwen-Image-Lightning-4steps-V1.0.safetensors)\n\n**Text encoder**\n\n- [qwen_2.5_vl_7b_fp8_scaled.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/text_encoders/qwen_2.5_vl_7b_fp8_scaled.safetensors)\n\n**VAE**\n\n- [qwen_image_vae.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/vae/qwen_image_vae.safetensors)\n\n\nModel Storage Location\n\n```\n📂 ComfyUI/\n├── 📂 models/\n│ ├── 📂 diffusion_models/\n│ │ ├── qwen_image_fp8_e4m3fn.safetensors\n│ │ └── qwen_image_distill_full_fp8_e4m3fn.safetensors\n│ ├── 📂 loras/\n│ │ └── Qwen-Image-Lightning-8steps-V1.0.safetensors\n│ ├── 📂 controlnet/ \n│ │ └── Qwen-Image-InstantX-ControlNet-Inpainting.safetensors\n│ ├── 📂 vae/\n│ │ └── qwen_image_vae.safetensors\n│ └── 📂 text_encoders/\n│ └── qwen_2.5_vl_7b_fp8_scaled.safetensors\n```\n"
1885
+ "Increase the shift if you get too many blury/dark/bad images. Decrease if you want to try increasing detail."
1888
1886
  ],
1889
1887
  "color": "#432",
1890
1888
  "bgcolor": "#653"
1891
1889
  },
1892
1890
  {
1893
- "id": 68,
1894
- "type": "Note",
1891
+ "id": 78,
1892
+ "type": "MarkdownNote",
1895
1893
  "pos": [
1896
- 850,
1897
- -300
1894
+ -700,
1895
+ -70
1898
1896
  ],
1899
1897
  "size": [
1900
- 310,
1901
- 90
1898
+ 541.36865234375,
1899
+ 579.70263671875
1902
1900
  ],
1903
1901
  "flags": {},
1904
1902
  "order": 16,
1905
1903
  "mode": 0,
1906
1904
  "inputs": [],
1907
1905
  "outputs": [],
1908
- "properties": {},
1906
+ "title": "Model links",
1907
+ "properties": {
1908
+ "widget_ue_connectable": {}
1909
+ },
1909
1910
  "widgets_values": [
1910
- "Increase the shift if you get too many blury/dark/bad images. Decrease if you want to try increasing detail."
1911
+ "[Tutorial](https://docs.comfy.org/tutorials/image/qwen/qwen-image)\n\n\n## Model links\n\nYou can find all the models on [Huggingface](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/tree/main) or [Modelscope](https://modelscope.cn/models/Comfy-Org/Qwen-Image_ComfyUI/files)\n\n**Diffusion model**\n\n- [qwen_image_fp8_e4m3fn.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/diffusion_models/qwen_image_fp8_e4m3fn.safetensors)\n\n**ControlNet**\n\n- [Qwen-Image-InstantX-ControlNet-Inpainting.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image-InstantX-ControlNets/resolve/main/split_files/controlnet/Qwen-Image-InstantX-ControlNet-Inpainting.safetensors)\n\n\n**LoRA**\n\n- [Qwen-Image-Lightning-4steps-V1.0.safetensors](https://huggingface.co/lightx2v/Qwen-Image-Lightning/resolve/main/Qwen-Image-Lightning-4steps-V1.0.safetensors)\n\n**Text encoder**\n\n- [qwen_2.5_vl_7b_fp8_scaled.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/text_encoders/qwen_2.5_vl_7b_fp8_scaled.safetensors)\n\n**VAE**\n\n- [qwen_image_vae.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/vae/qwen_image_vae.safetensors)\n\n\nModel Storage Location\n\n```\n📂 ComfyUI/\n├── 📂 models/\n│ ├── 📂 diffusion_models/\n│ │ ├── qwen_image_fp8_e4m3fn.safetensors\n│ │ └── qwen_image_distill_full_fp8_e4m3fn.safetensors\n│ ├── 📂 loras/\n│ │ └── Qwen-Image-Lightning-8steps-V1.0.safetensors\n│ ├── 📂 controlnet/ \n│ │ └── Qwen-Image-InstantX-ControlNet-Inpainting.safetensors\n│ ├── 📂 vae/\n│ │ └── qwen_image_vae.safetensors\n│ └── 📂 text_encoders/\n│ └── qwen_2.5_vl_7b_fp8_scaled.safetensors\n```\n"
1911
1912
  ],
1912
1913
  "color": "#432",
1913
1914
  "bgcolor": "#653"
@@ -2554,10 +2555,10 @@
2554
2555
  279
2555
2556
  ],
2556
2557
  "localized_name": "mask",
2557
- "pos": {
2558
- "0": 390,
2559
- "1": 3556
2560
- }
2558
+ "pos": [
2559
+ 390,
2560
+ 3556
2561
+ ]
2561
2562
  },
2562
2563
  {
2563
2564
  "id": "d1ab0cf5-7062-41ac-9f4b-8c660fc4a714",
@@ -2566,10 +2567,10 @@
2566
2567
  "linkIds": [
2567
2568
  379
2568
2569
  ],
2569
- "pos": {
2570
- "0": 390,
2571
- "1": 3576
2572
- }
2570
+ "pos": [
2571
+ 390,
2572
+ 3576
2573
+ ]
2573
2574
  },
2574
2575
  {
2575
2576
  "id": "1a787af5-da9f-44c5-9f5a-3f71609ca0ef",
@@ -2578,10 +2579,10 @@
2578
2579
  "linkIds": [
2579
2580
  380
2580
2581
  ],
2581
- "pos": {
2582
- "0": 390,
2583
- "1": 3596
2584
- }
2582
+ "pos": [
2583
+ 390,
2584
+ 3596
2585
+ ]
2585
2586
  }
2586
2587
  ],
2587
2588
  "outputs": [
@@ -2593,10 +2594,10 @@
2593
2594
  378
2594
2595
  ],
2595
2596
  "localized_name": "MASK",
2596
- "pos": {
2597
- "0": 1150,
2598
- "1": 3556
2599
- }
2597
+ "pos": [
2598
+ 1150,
2599
+ 3556
2600
+ ]
2600
2601
  }
2601
2602
  ],
2602
2603
  "widgets": [],
@@ -2634,9 +2635,9 @@
2634
2635
  }
2635
2636
  ],
2636
2637
  "properties": {
2638
+ "Node name for S&R": "ImageToMask",
2637
2639
  "cnr_id": "comfy-core",
2638
- "ver": "0.3.59",
2639
- "Node name for S&R": "ImageToMask"
2640
+ "ver": "0.3.59"
2640
2641
  },
2641
2642
  "widgets_values": [
2642
2643
  "red"
@@ -2675,9 +2676,9 @@
2675
2676
  }
2676
2677
  ],
2677
2678
  "properties": {
2679
+ "Node name for S&R": "MaskToImage",
2678
2680
  "cnr_id": "comfy-core",
2679
- "ver": "0.3.59",
2680
- "Node name for S&R": "MaskToImage"
2681
+ "ver": "0.3.59"
2681
2682
  },
2682
2683
  "widgets_values": []
2683
2684
  },
@@ -2723,9 +2724,9 @@
2723
2724
  }
2724
2725
  ],
2725
2726
  "properties": {
2727
+ "Node name for S&R": "GrowMask",
2726
2728
  "cnr_id": "comfy-core",
2727
- "ver": "0.3.59",
2728
- "Node name for S&R": "GrowMask"
2729
+ "ver": "0.3.59"
2729
2730
  },
2730
2731
  "widgets_values": [
2731
2732
  0,
@@ -2774,9 +2775,9 @@
2774
2775
  }
2775
2776
  ],
2776
2777
  "properties": {
2778
+ "Node name for S&R": "ImageBlur",
2777
2779
  "cnr_id": "comfy-core",
2778
- "ver": "0.3.59",
2779
- "Node name for S&R": "ImageBlur"
2780
+ "ver": "0.3.59"
2780
2781
  },
2781
2782
  "widgets_values": [
2782
2783
  31,
@@ -2850,13 +2851,13 @@
2850
2851
  "config": {},
2851
2852
  "extra": {
2852
2853
  "ds": {
2853
- "scale": 0.6304711010110896,
2854
+ "scale": 0.32353136379369263,
2854
2855
  "offset": [
2855
- 507.126621348702,
2856
- 616.8919118351296
2856
+ 1677.4924098235829,
2857
+ 904.1815883388806
2857
2858
  ]
2858
2859
  },
2859
- "frontendVersion": "1.26.11",
2860
+ "frontendVersion": "1.28.6",
2860
2861
  "groupNodes": {},
2861
2862
  "VHS_latentpreview": false,
2862
2863
  "VHS_latentpreviewrate": 0,
@@ -826,7 +826,7 @@
826
826
  "widget_ue_connectable": {}
827
827
  },
828
828
  "widgets_values": [
829
- "[Tutorial](https://docs.comfy.org/tutorials/image/qwen/qwen-image) | [教程](https://docs.comfy.org/zh-CN/tutorials/image/qwen/qwen-image)\n\n\n## Model links\n\nYou can find all the models on [Huggingface](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/tree/main) or [Modelscope](https://modelscope.cn/models/Comfy-Org/Qwen-Image_ComfyUI/files)\n\n**Diffusion model**\n\n- [qwen_image_fp8_e4m3fn.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/diffusion_models/qwen_image_fp8_e4m3fn.safetensors)\n\n**LoRA**\n\n- [Qwen-Image-Lightning-8steps-V1.0.safetensors](https://huggingface.co/lightx2v/Qwen-Image-Lightning/resolve/main/Qwen-Image-Lightning-8steps-V1.0.safetensors)\n- [qwen_image_union_diffsynth_lora.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image-DiffSynth-ControlNets/resolve/main/split_files/loras/qwen_image_union_diffsynth_lora.safetensors)\n\n**Text encoder**\n\n- [qwen_2.5_vl_7b_fp8_scaled.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/text_encoders/qwen_2.5_vl_7b_fp8_scaled.safetensors)\n\n**VAE**\n\n- [qwen_image_vae.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/vae/qwen_image_vae.safetensors)\n\nModel Storage Location\n\n```\n📂 ComfyUI/\n├── 📂 models/\n│ ├── 📂 diffusion_models/\n│ │ ├── qwen_image_fp8_e4m3fn.safetensors\n│ │ └── qwen_image_distill_full_fp8_e4m3fn.safetensors\n│ ├── 📂 loras/\n│ │ ├── qwen_image_union_diffsynth_lora.safetensors\n│ │ └── Qwen-Image-Lightning-8steps-V1.0.safetensors\n│ ├── 📂 vae/\n│ │ └── qwen_image_vae.safetensors\n│ └── 📂 text_encoders/\n│ └── qwen_2.5_vl_7b_fp8_scaled.safetensors\n```\n"
829
+ "[Tutorial](https://docs.comfy.org/tutorials/image/qwen/qwen-image) \n\n\n## Model links\n\nYou can find all the models on [Huggingface](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/tree/main) or [Modelscope](https://modelscope.cn/models/Comfy-Org/Qwen-Image_ComfyUI/files)\n\n**Diffusion model**\n\n- [qwen_image_fp8_e4m3fn.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/diffusion_models/qwen_image_fp8_e4m3fn.safetensors)\n\n**LoRA**\n\n- [Qwen-Image-Lightning-8steps-V1.0.safetensors](https://huggingface.co/lightx2v/Qwen-Image-Lightning/resolve/main/Qwen-Image-Lightning-8steps-V1.0.safetensors)\n- [qwen_image_union_diffsynth_lora.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image-DiffSynth-ControlNets/resolve/main/split_files/loras/qwen_image_union_diffsynth_lora.safetensors)\n\n**Text encoder**\n\n- [qwen_2.5_vl_7b_fp8_scaled.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/text_encoders/qwen_2.5_vl_7b_fp8_scaled.safetensors)\n\n**VAE**\n\n- [qwen_image_vae.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/vae/qwen_image_vae.safetensors)\n\nModel Storage Location\n\n```\n📂 ComfyUI/\n├── 📂 models/\n│ ├── 📂 diffusion_models/\n│ │ ├── qwen_image_fp8_e4m3fn.safetensors\n│ │ └── qwen_image_distill_full_fp8_e4m3fn.safetensors\n│ ├── 📂 loras/\n│ │ ├── qwen_image_union_diffsynth_lora.safetensors\n│ │ └── Qwen-Image-Lightning-8steps-V1.0.safetensors\n│ ├── 📂 vae/\n│ │ └── qwen_image_vae.safetensors\n│ └── 📂 text_encoders/\n│ └── qwen_2.5_vl_7b_fp8_scaled.safetensors\n```\n"
830
830
  ],
831
831
  "color": "#432",
832
832
  "bgcolor": "#653"
@@ -1085,13 +1085,13 @@
1085
1085
  "config": {},
1086
1086
  "extra": {
1087
1087
  "ds": {
1088
- "scale": 0.7213855104977631,
1088
+ "scale": 0.49271601017537264,
1089
1089
  "offset": [
1090
- 235.0958937828103,
1091
- -225.58323513433564
1090
+ 3.0130483183901795,
1091
+ 57.458228237984414
1092
1092
  ]
1093
1093
  },
1094
- "frontendVersion": "1.26.6"
1094
+ "frontendVersion": "1.28.6"
1095
1095
  },
1096
1096
  "version": 0.4
1097
1097
  }
@@ -386,6 +386,17 @@
386
386
  "size": 22.0,
387
387
  "vram": 22.0
388
388
  },
389
+ {
390
+ "name": "image_netayume_lumina_t2i",
391
+ "title": "NetaYume Lumina Text to Image",
392
+ "mediaType": "image",
393
+ "mediaSubtype": "webp",
394
+ "description": "High-quality anime-style image generation with enhanced character understanding and detailed textures. Fine-tuned from Neta Lumina on Danbooru dataset.",
395
+ "tags": ["Text to Image", "Image", "Anime"],
396
+ "models": ["NetaYume Lumina"],
397
+ "date": "2025-10-10",
398
+ "size": 9.89
399
+ },
389
400
  {
390
401
  "name": "image_chroma_text_to_image",
391
402
  "title": "Texto a imagen Chroma",
@@ -1678,6 +1689,19 @@
1678
1689
  "icon": "icon-[lucide--film]",
1679
1690
  "title": "API de video",
1680
1691
  "templates": [
1692
+ {
1693
+ "name": "api_openai_sora_video",
1694
+ "title": "Sora 2: Text & Image to Video",
1695
+ "description": "OpenAI's Sora-2 and Sora-2 Pro video generation with synchronized audio.",
1696
+ "mediaType": "image",
1697
+ "mediaSubtype": "webp",
1698
+ "tags": ["Image to Video", "Text to Video", "API"],
1699
+ "models": ["OpenAI"],
1700
+ "date": "2025-10-08",
1701
+ "OpenSource": false,
1702
+ "size": 0,
1703
+ "vram": 0
1704
+ },
1681
1705
  {
1682
1706
  "name": "api_wan_text_to_video",
1683
1707
  "title": "Wan2.5: Text to Video",