comfyui-workflow-templates 0.1.93__py3-none-any.whl → 0.1.95__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of comfyui-workflow-templates might be problematic. Click here for more details.

Files changed (138) hide show
  1. comfyui_workflow_templates/templates/3d_hunyuan3d_image_to_model.json +26 -26
  2. comfyui_workflow_templates/templates/3d_hunyuan3d_multiview_to_model.json +35 -35
  3. comfyui_workflow_templates/templates/3d_hunyuan3d_multiview_to_model_turbo.json +37 -37
  4. comfyui_workflow_templates/templates/api_bfl_flux_1_kontext_max_image.json +57 -81
  5. comfyui_workflow_templates/templates/api_bfl_flux_1_kontext_multiple_images_input.json +60 -84
  6. comfyui_workflow_templates/templates/api_bfl_flux_1_kontext_pro_image.json +57 -81
  7. comfyui_workflow_templates/templates/api_bfl_flux_pro_t2i.json +28 -26
  8. comfyui_workflow_templates/templates/api_bytedance_flf2v.json +11 -11
  9. comfyui_workflow_templates/templates/api_bytedance_image_to_video.json +34 -34
  10. comfyui_workflow_templates/templates/api_bytedance_text_to_video.json +39 -40
  11. comfyui_workflow_templates/templates/api_google_gemini.json +6 -7
  12. comfyui_workflow_templates/templates/api_hailuo_minimax_i2v.json +12 -3
  13. comfyui_workflow_templates/templates/api_hailuo_minimax_t2v.json +28 -28
  14. comfyui_workflow_templates/templates/api_hailuo_minimax_video.json +30 -30
  15. comfyui_workflow_templates/templates/api_ideogram_v3_t2i.json +32 -18
  16. comfyui_workflow_templates/templates/api_kling_effects.json +28 -26
  17. comfyui_workflow_templates/templates/api_kling_flf.json +32 -30
  18. comfyui_workflow_templates/templates/api_kling_i2v.json +34 -34
  19. comfyui_workflow_templates/templates/api_luma_i2v.json +93 -110
  20. comfyui_workflow_templates/templates/api_luma_photon_i2i.json +59 -50
  21. comfyui_workflow_templates/templates/api_luma_photon_style_ref.json +131 -124
  22. comfyui_workflow_templates/templates/api_luma_t2v.json +59 -50
  23. comfyui_workflow_templates/templates/api_openai_dall_e_2_t2i.json +8 -29
  24. comfyui_workflow_templates/templates/api_openai_image_1_i2i.json +12 -33
  25. comfyui_workflow_templates/templates/api_openai_image_1_inpaint.json +30 -51
  26. comfyui_workflow_templates/templates/api_openai_image_1_multi_inputs.json +10 -33
  27. comfyui_workflow_templates/templates/api_openai_image_1_t2i.json +42 -63
  28. comfyui_workflow_templates/templates/api_openai_sora_video-1.webp +0 -0
  29. comfyui_workflow_templates/templates/api_openai_sora_video.json +494 -0
  30. comfyui_workflow_templates/templates/api_pika_i2v.json +27 -27
  31. comfyui_workflow_templates/templates/api_pika_scene.json +12 -3
  32. comfyui_workflow_templates/templates/api_pixverse_i2v.json +31 -36
  33. comfyui_workflow_templates/templates/api_pixverse_t2v.json +20 -16
  34. comfyui_workflow_templates/templates/api_pixverse_template_i2v.json +39 -35
  35. comfyui_workflow_templates/templates/api_recraft_image_gen_with_color_control.json +6 -6
  36. comfyui_workflow_templates/templates/api_recraft_image_gen_with_style_control.json +212 -199
  37. comfyui_workflow_templates/templates/api_recraft_vector_gen.json +78 -69
  38. comfyui_workflow_templates/templates/api_rodin_gen2.json +30 -30
  39. comfyui_workflow_templates/templates/api_rodin_image_to_model.json +55 -55
  40. comfyui_workflow_templates/templates/api_rodin_multiview_to_model.json +188 -132
  41. comfyui_workflow_templates/templates/api_runway_first_last_frame.json +4 -4
  42. comfyui_workflow_templates/templates/api_runway_gen3a_turbo_image_to_video.json +30 -31
  43. comfyui_workflow_templates/templates/api_runway_gen4_turo_image_to_video.json +29 -30
  44. comfyui_workflow_templates/templates/api_runway_reference_to_image.json +31 -32
  45. comfyui_workflow_templates/templates/api_runway_text_to_image.json +17 -17
  46. comfyui_workflow_templates/templates/api_stability_ai_audio_inpaint.json +18 -18
  47. comfyui_workflow_templates/templates/api_stability_ai_audio_to_audio.json +31 -31
  48. comfyui_workflow_templates/templates/api_stability_ai_i2i.json +34 -34
  49. comfyui_workflow_templates/templates/api_stability_ai_sd3.5_i2i.json +21 -19
  50. comfyui_workflow_templates/templates/api_stability_ai_sd3.5_t2i.json +35 -35
  51. comfyui_workflow_templates/templates/api_stability_ai_stable_image_ultra_t2i.json +11 -9
  52. comfyui_workflow_templates/templates/api_tripo_image_to_model.json +90 -92
  53. comfyui_workflow_templates/templates/api_tripo_multiview_to_model.json +241 -241
  54. comfyui_workflow_templates/templates/api_tripo_text_to_model.json +102 -102
  55. comfyui_workflow_templates/templates/api_veo2_i2v.json +31 -28
  56. comfyui_workflow_templates/templates/api_veo3.json +30 -30
  57. comfyui_workflow_templates/templates/api_vidu_text_to_video.json +2 -2
  58. comfyui_workflow_templates/templates/api_wan_image_to_video.json +41 -42
  59. comfyui_workflow_templates/templates/api_wan_text_to_image .json +140 -0
  60. comfyui_workflow_templates/templates/api_wan_text_to_video.json +38 -45
  61. comfyui_workflow_templates/templates/audio_ace_step_1_m2m_editing.json +84 -84
  62. comfyui_workflow_templates/templates/audio_ace_step_1_t2a_instrumentals.json +60 -60
  63. comfyui_workflow_templates/templates/audio_ace_step_1_t2a_song.json +60 -60
  64. comfyui_workflow_templates/templates/esrgan_example.json +24 -30
  65. comfyui_workflow_templates/templates/flux1_dev_uso_reference_image_gen.json +215 -210
  66. comfyui_workflow_templates/templates/flux1_krea_dev.json +3 -3
  67. comfyui_workflow_templates/templates/flux_canny_model_example.json +2 -2
  68. comfyui_workflow_templates/templates/flux_depth_lora_example.json +394 -308
  69. comfyui_workflow_templates/templates/flux_fill_inpaint_example.json +2 -2
  70. comfyui_workflow_templates/templates/flux_fill_outpaint_example.json +2 -2
  71. comfyui_workflow_templates/templates/flux_kontext_dev_basic.json +151 -231
  72. comfyui_workflow_templates/templates/flux_redux_model_example.json +108 -120
  73. comfyui_workflow_templates/templates/flux_schnell_full_text_to_image.json +21 -29
  74. comfyui_workflow_templates/templates/hidream_e1_1.json +179 -209
  75. comfyui_workflow_templates/templates/hidream_e1_full.json +33 -39
  76. comfyui_workflow_templates/templates/hidream_i1_dev.json +15 -15
  77. comfyui_workflow_templates/templates/hidream_i1_fast.json +15 -15
  78. comfyui_workflow_templates/templates/hidream_i1_full.json +17 -16
  79. comfyui_workflow_templates/templates/hiresfix_esrgan_workflow.json +31 -37
  80. comfyui_workflow_templates/templates/hiresfix_latent_workflow.json +84 -88
  81. comfyui_workflow_templates/templates/image2image.json +30 -30
  82. comfyui_workflow_templates/templates/image_chroma1_radiance_text_to_image.json +60 -60
  83. comfyui_workflow_templates/templates/image_chroma_text_to_image.json +113 -113
  84. comfyui_workflow_templates/templates/image_lotus_depth_v1_1.json +25 -31
  85. comfyui_workflow_templates/templates/image_netayume_lumina_t2i-1.webp +0 -0
  86. comfyui_workflow_templates/templates/image_netayume_lumina_t2i.json +597 -0
  87. comfyui_workflow_templates/templates/image_omnigen2_image_edit.json +55 -62
  88. comfyui_workflow_templates/templates/image_omnigen2_t2i.json +26 -33
  89. comfyui_workflow_templates/templates/image_qwen_image.json +40 -40
  90. comfyui_workflow_templates/templates/image_qwen_image_controlnet_patch.json +32 -32
  91. comfyui_workflow_templates/templates/image_qwen_image_edit.json +29 -29
  92. comfyui_workflow_templates/templates/image_qwen_image_edit_2509.json +2234 -357
  93. comfyui_workflow_templates/templates/image_qwen_image_instantx_controlnet.json +56 -55
  94. comfyui_workflow_templates/templates/image_qwen_image_instantx_inpainting_controlnet.json +108 -107
  95. comfyui_workflow_templates/templates/image_qwen_image_union_control_lora.json +5 -5
  96. comfyui_workflow_templates/templates/index.es.json +24 -0
  97. comfyui_workflow_templates/templates/index.fr.json +24 -0
  98. comfyui_workflow_templates/templates/index.ja.json +24 -0
  99. comfyui_workflow_templates/templates/index.json +24 -0
  100. comfyui_workflow_templates/templates/index.ko.json +24 -0
  101. comfyui_workflow_templates/templates/index.ru.json +24 -0
  102. comfyui_workflow_templates/templates/index.zh-TW.json +24 -0
  103. comfyui_workflow_templates/templates/index.zh.json +24 -0
  104. comfyui_workflow_templates/templates/inpaint_example.json +70 -72
  105. comfyui_workflow_templates/templates/inpaint_model_outpainting.json +4 -4
  106. comfyui_workflow_templates/templates/latent_upscale_different_prompt_model.json +179 -185
  107. comfyui_workflow_templates/templates/sd3.5_large_blur.json +320 -241
  108. comfyui_workflow_templates/templates/sd3.5_large_canny_controlnet_example.json +1 -1
  109. comfyui_workflow_templates/templates/sd3.5_large_depth.json +245 -170
  110. comfyui_workflow_templates/templates/sdxlturbo_example.json +308 -162
  111. comfyui_workflow_templates/templates/video_wan2.1_fun_camera_v1.1_1.3B.json +89 -62
  112. comfyui_workflow_templates/templates/video_wan2.1_fun_camera_v1.1_14B.json +8 -4
  113. comfyui_workflow_templates/templates/video_wan2_2_14B_animate.json +46 -44
  114. comfyui_workflow_templates/templates/video_wan2_2_14B_flf2v.json +38 -38
  115. comfyui_workflow_templates/templates/video_wan2_2_14B_fun_camera.json +58 -54
  116. comfyui_workflow_templates/templates/video_wan2_2_14B_fun_control.json +36 -36
  117. comfyui_workflow_templates/templates/video_wan2_2_14B_fun_inpaint.json +26 -26
  118. comfyui_workflow_templates/templates/video_wan2_2_14B_i2v.json +4 -4
  119. comfyui_workflow_templates/templates/video_wan2_2_14B_s2v.json +33 -29
  120. comfyui_workflow_templates/templates/video_wan2_2_14B_t2v (2).json +1954 -0
  121. comfyui_workflow_templates/templates/video_wan2_2_5B_fun_control.json +29 -29
  122. comfyui_workflow_templates/templates/video_wan2_2_5B_fun_inpaint.json +25 -25
  123. comfyui_workflow_templates/templates/video_wan2_2_5B_ti2v.json +49 -49
  124. comfyui_workflow_templates/templates/video_wan_ati.json +49 -49
  125. comfyui_workflow_templates/templates/video_wan_vace_14B_ref2v.json +47 -61
  126. comfyui_workflow_templates/templates/video_wan_vace_14B_t2v.json +2 -2
  127. comfyui_workflow_templates/templates/video_wan_vace_14B_v2v.json +55 -55
  128. comfyui_workflow_templates/templates/video_wan_vace_flf2v.json +40 -56
  129. comfyui_workflow_templates/templates/video_wan_vace_inpainting.json +72 -72
  130. comfyui_workflow_templates/templates/video_wan_vace_outpainting.json +211 -237
  131. comfyui_workflow_templates/templates/wan2.1_flf2v_720_f16.json +84 -92
  132. comfyui_workflow_templates/templates/wan2.1_fun_control.json +51 -27
  133. comfyui_workflow_templates/templates/wan2.1_fun_inp.json +43 -17
  134. {comfyui_workflow_templates-0.1.93.dist-info → comfyui_workflow_templates-0.1.95.dist-info}/METADATA +1 -1
  135. {comfyui_workflow_templates-0.1.93.dist-info → comfyui_workflow_templates-0.1.95.dist-info}/RECORD +138 -132
  136. {comfyui_workflow_templates-0.1.93.dist-info → comfyui_workflow_templates-0.1.95.dist-info}/WHEEL +0 -0
  137. {comfyui_workflow_templates-0.1.93.dist-info → comfyui_workflow_templates-0.1.95.dist-info}/licenses/LICENSE +0 -0
  138. {comfyui_workflow_templates-0.1.93.dist-info → comfyui_workflow_templates-0.1.95.dist-info}/top_level.txt +0 -0
@@ -166,9 +166,7 @@
166
166
  "widgets_values": [
167
167
  "wan2.1_vace_1.3B_fp16.safetensors",
168
168
  "default"
169
- ],
170
- "color": "#322",
171
- "bgcolor": "#533"
169
+ ]
172
170
  },
173
171
  {
174
172
  "id": 159,
@@ -224,9 +222,7 @@
224
222
  "Wan21_CausVid_bidirect2_T2V_1_3B_lora_rank32.safetensors",
225
223
  0.7000000000000002,
226
224
  1
227
- ],
228
- "color": "#322",
229
- "bgcolor": "#533"
225
+ ]
230
226
  },
231
227
  {
232
228
  "id": 49,
@@ -463,29 +459,6 @@
463
459
  },
464
460
  "widgets_values": []
465
461
  },
466
- {
467
- "id": 136,
468
- "type": "Note",
469
- "pos": [
470
- 1060,
471
- 690
472
- ],
473
- "size": [
474
- 240,
475
- 310
476
- ],
477
- "flags": {},
478
- "order": 2,
479
- "mode": 0,
480
- "inputs": [],
481
- "outputs": [],
482
- "properties": {},
483
- "widgets_values": [
484
- "Currently (May 18, 2025), the Comfy Core nodes lack relevant nodes for obtaining image size or video frames. Here, we can use the Preview image to indirectly obtain the image size and the total number of video frames.\n\nYou can use custom nodes such as image size acquisition to optimize the step of size modification.\n\n---\n\n目前(2025年5月18日),Comfy Core 节点缺少获取图像尺寸或数量的相关节点。这里我们可以使用 Preview image 来辅助获取到图片尺寸和视频总帧数。\n\n你可以使用图像尺寸获取之类的自定义节点,来优化尺寸修改这个步骤\n"
485
- ],
486
- "color": "#432",
487
- "bgcolor": "#653"
488
- },
489
462
  {
490
463
  "id": 111,
491
464
  "type": "MaskToImage",
@@ -621,7 +594,7 @@
621
594
  317.6267395019531
622
595
  ],
623
596
  "flags": {},
624
- "order": 3,
597
+ "order": 2,
625
598
  "mode": 0,
626
599
  "inputs": [],
627
600
  "outputs": [
@@ -641,9 +614,7 @@
641
614
  "widgets_values": [
642
615
  "girl.mp4",
643
616
  "image"
644
- ],
645
- "color": "#322",
646
- "bgcolor": "#533"
617
+ ]
647
618
  },
648
619
  {
649
620
  "id": 139,
@@ -657,7 +628,7 @@
657
628
  82
658
629
  ],
659
630
  "flags": {},
660
- "order": 4,
631
+ "order": 3,
661
632
  "mode": 0,
662
633
  "inputs": [],
663
634
  "outputs": [
@@ -678,9 +649,7 @@
678
649
  "widgets_values": [
679
650
  960,
680
651
  "fixed"
681
- ],
682
- "color": "#322",
683
- "bgcolor": "#533"
652
+ ]
684
653
  },
685
654
  {
686
655
  "id": 161,
@@ -694,7 +663,7 @@
694
663
  106
695
664
  ],
696
665
  "flags": {},
697
- "order": 5,
666
+ "order": 4,
698
667
  "mode": 0,
699
668
  "inputs": [],
700
669
  "outputs": [
@@ -724,9 +693,7 @@
724
693
  "umt5_xxl_fp16.safetensors",
725
694
  "wan",
726
695
  "default"
727
- ],
728
- "color": "#322",
729
- "bgcolor": "#533"
696
+ ]
730
697
  },
731
698
  {
732
699
  "id": 152,
@@ -740,7 +707,7 @@
740
707
  58
741
708
  ],
742
709
  "flags": {},
743
- "order": 6,
710
+ "order": 5,
744
711
  "mode": 0,
745
712
  "inputs": [],
746
713
  "outputs": [
@@ -768,9 +735,7 @@
768
735
  },
769
736
  "widgets_values": [
770
737
  "wan_2.1_vae.safetensors"
771
- ],
772
- "color": "#322",
773
- "bgcolor": "#533"
738
+ ]
774
739
  },
775
740
  {
776
741
  "id": 110,
@@ -821,9 +786,7 @@
821
786
  120,
822
787
  200,
823
788
  0
824
- ],
825
- "color": "#322",
826
- "bgcolor": "#533"
789
+ ]
827
790
  },
828
791
  {
829
792
  "id": 72,
@@ -873,69 +836,6 @@
873
836
  },
874
837
  "widgets_values": []
875
838
  },
876
- {
877
- "id": 3,
878
- "type": "KSampler",
879
- "pos": [
880
- 992.5,
881
- 50
882
- ],
883
- "size": [
884
- 315,
885
- 262
886
- ],
887
- "flags": {},
888
- "order": 29,
889
- "mode": 0,
890
- "inputs": [
891
- {
892
- "name": "model",
893
- "type": "MODEL",
894
- "link": 95
895
- },
896
- {
897
- "name": "positive",
898
- "type": "CONDITIONING",
899
- "link": 98
900
- },
901
- {
902
- "name": "negative",
903
- "type": "CONDITIONING",
904
- "link": 99
905
- },
906
- {
907
- "name": "latent_image",
908
- "type": "LATENT",
909
- "link": 160
910
- }
911
- ],
912
- "outputs": [
913
- {
914
- "name": "LATENT",
915
- "type": "LATENT",
916
- "slot_index": 0,
917
- "links": [
918
- 116
919
- ]
920
- }
921
- ],
922
- "properties": {
923
- "cnr_id": "comfy-core",
924
- "ver": "0.3.34",
925
- "Node name for S&R": "KSampler"
926
- },
927
- "widgets_values": [
928
- 776767434080331,
929
- "randomize",
930
- 3,
931
- 1,
932
- "uni_pc",
933
- "simple",
934
- 1
935
- ],
936
- "color": "#322",
937
- "bgcolor": "#533"
938
- },
939
839
  {
940
840
  "id": 151,
941
841
  "type": "UNETLoader",
@@ -948,7 +848,7 @@
948
848
  82
949
849
  ],
950
850
  "flags": {},
951
- "order": 7,
851
+ "order": 6,
952
852
  "mode": 0,
953
853
  "inputs": [],
954
854
  "outputs": [
@@ -976,33 +876,7 @@
976
876
  "widgets_values": [
977
877
  "wan2.1_vace_14B_fp16.safetensors",
978
878
  "default"
979
- ],
980
- "color": "#322",
981
- "bgcolor": "#533"
982
- },
983
- {
984
- "id": 156,
985
- "type": "MarkdownNote",
986
- "pos": [
987
- -680,
988
- 560
989
- ],
990
- "size": [
991
- 370,
992
- 220
993
- ],
994
- "flags": {},
995
- "order": 8,
996
- "mode": 0,
997
- "inputs": [],
998
- "outputs": [],
999
- "title": "Note",
1000
- "properties": {},
1001
- "widgets_values": [
1002
- "The generation quality of the 14B model is better, but it takes longer to generate. If you have already downloaded the model, you can choose to directly use the nodes above, or just modify the model loaded by the loader node.\n\nThe corresponding LoRA should match the Diffusion Model. For example, the LoRA corresponding to the 14B diffusion model is the 14B LoRA. \n\n---\n\n14B 的生成质量更好,但是需要更长的时间去生成,如果你下载好了模型,你可以选择直接使用上面的节点,或者只是修改 loader 节点加载的模型。\n\n对应的 LoRA 应该是和 Diffusion Model匹配的,比如 14B 的 diffusion model 对应的是 14B 的 LoRA"
1003
- ],
1004
- "color": "#432",
1005
- "bgcolor": "#653"
879
+ ]
1006
880
  },
1007
881
  {
1008
882
  "id": 160,
@@ -1016,7 +890,7 @@
1016
890
  106
1017
891
  ],
1018
892
  "flags": {},
1019
- "order": 9,
893
+ "order": 7,
1020
894
  "mode": 4,
1021
895
  "inputs": [],
1022
896
  "outputs": [
@@ -1043,9 +917,7 @@
1043
917
  "umt5_xxl_fp8_e4m3fn_scaled.safetensors",
1044
918
  "wan",
1045
919
  "default"
1046
- ],
1047
- "color": "#322",
1048
- "bgcolor": "#533"
920
+ ]
1049
921
  },
1050
922
  {
1051
923
  "id": 137,
@@ -1059,7 +931,7 @@
1059
931
  82
1060
932
  ],
1061
933
  "flags": {},
1062
- "order": 10,
934
+ "order": 8,
1063
935
  "mode": 0,
1064
936
  "inputs": [],
1065
937
  "outputs": [
@@ -1080,32 +952,7 @@
1080
952
  "widgets_values": [
1081
953
  960,
1082
954
  "fixed"
1083
- ],
1084
- "color": "#322",
1085
- "bgcolor": "#533"
1086
- },
1087
- {
1088
- "id": 142,
1089
- "type": "Note",
1090
- "pos": [
1091
- 330,
1092
- 1120
1093
- ],
1094
- "size": [
1095
- 250,
1096
- 260
1097
- ],
1098
- "flags": {},
1099
- "order": 11,
1100
- "mode": 0,
1101
- "inputs": [],
1102
- "outputs": [],
1103
- "properties": {},
1104
- "widgets_values": [
1105
- "Here, you need to manually set the size of the video after outpainting. Since the native nodes currently do not support the relevant functions, you can use some custom nodes to improve this step. Additionally, because the step size of the size setting for the WanVaceToVideo node is 16, the number you enter must be divisible by 16. \n\n在这里你需要手动设置视频 outpainting 后尺寸,由于目前原生节点没有相关功能支持,你可以使用一些自定义节点来改进这个步骤,另外由于 WanVaceToVideo 节点设置了尺寸的步长是 16 所以你输入的数字要求必须能被 16 整除\n"
1106
- ],
1107
- "color": "#432",
1108
- "bgcolor": "#653"
955
+ ]
1109
956
  },
1110
957
  {
1111
958
  "id": 154,
@@ -1119,14 +966,14 @@
1119
966
  158.5415802001953
1120
967
  ],
1121
968
  "flags": {},
1122
- "order": 12,
969
+ "order": 9,
1123
970
  "mode": 0,
1124
971
  "inputs": [],
1125
972
  "outputs": [],
1126
973
  "title": "KSampler Setting",
1127
974
  "properties": {},
1128
975
  "widgets_values": [
1129
- "## Default\n\n- steps:20\n- cfg:6.0\n\n## [For CausVid LoRA](https://www.reddit.com/r/StableDiffusion/comments/1knuafk/causvid_lora_massive_speedup_for_wan21_made_by/)\n\n- steps: 2-4\n- cfg: 1.0\n\n"
976
+ "## Default\n\n- steps:20\n- cfg:6.0\n\n## For CausVid LoRA\n\n- steps: 2-4\n- cfg: 1.0\n\n"
1130
977
  ],
1131
978
  "color": "#432",
1132
979
  "bgcolor": "#653"
@@ -1143,7 +990,7 @@
1143
990
  90
1144
991
  ],
1145
992
  "flags": {},
1146
- "order": 13,
993
+ "order": 10,
1147
994
  "mode": 0,
1148
995
  "inputs": [],
1149
996
  "outputs": [
@@ -1165,9 +1012,7 @@
1165
1012
  "widgets_values": [
1166
1013
  81,
1167
1014
  "fixed"
1168
- ],
1169
- "color": "#322",
1170
- "bgcolor": "#533"
1015
+ ]
1171
1016
  },
1172
1017
  {
1173
1018
  "id": 6,
@@ -1285,36 +1130,6 @@
1285
1130
  },
1286
1131
  "widgets_values": []
1287
1132
  },
1288
- {
1289
- "id": 135,
1290
- "type": "PreviewImage",
1291
- "pos": [
1292
- 630,
1293
- 690
1294
- ],
1295
- "size": [
1296
- 390,
1297
- 310
1298
- ],
1299
- "flags": {},
1300
- "order": 24,
1301
- "mode": 0,
1302
- "inputs": [
1303
- {
1304
- "name": "images",
1305
- "type": "IMAGE",
1306
- "link": 214
1307
- }
1308
- ],
1309
- "outputs": [],
1310
- "title": "Preview Image - Outpaint",
1311
- "properties": {
1312
- "cnr_id": "comfy-core",
1313
- "ver": "0.3.38",
1314
- "Node name for S&R": "PreviewImage"
1315
- },
1316
- "widgets_values": []
1317
- },
1318
1133
  {
1319
1134
  "id": 69,
1320
1135
  "type": "SaveVideo",
@@ -1348,30 +1163,6 @@
1348
1163
  "auto"
1349
1164
  ]
1350
1165
  },
1351
- {
1352
- "id": 184,
1353
- "type": "MarkdownNote",
1354
- "pos": [
1355
- -290,
1356
- -190
1357
- ],
1358
- "size": [
1359
- 410,
1360
- 140
1361
- ],
1362
- "flags": {},
1363
- "order": 14,
1364
- "mode": 0,
1365
- "inputs": [],
1366
- "outputs": [],
1367
- "title": "About CausVid LoRA",
1368
- "properties": {},
1369
- "widgets_values": [
1370
- "We have added CausVid LoRA by default to achieve acceleration. However, in some cases, the video may shake and become blurry. You might need to test different LoRA intensities to get the best results, which should be between 0.3 and 0.7. If you don't need it, you can use the bypass mode to disable it, and then restore the settings of `KSampler` to the default ones.\n\n\n我们默认添加了 CausVid LoRA 来实现加速,但有些情况下会出现视频抖动和模糊的情况,你可能需要测试不同的 LoRA 强度来获取最好的结果,0.3~0.7 之间。如果你不需要的话,可以使用 bypass 模式禁用它,然后恢复 `KSampler`的设置到默认的设置即可。"
1371
- ],
1372
- "color": "#432",
1373
- "bgcolor": "#653"
1374
- },
1375
1166
  {
1376
1167
  "id": 165,
1377
1168
  "type": "LoraLoader",
@@ -1431,9 +1222,168 @@
1431
1222
  "Wan21_CausVid_14B_T2V_lora_rank32.safetensors",
1432
1223
  0.4000000000000001,
1433
1224
  1
1225
+ ]
1226
+ },
1227
+ {
1228
+ "id": 3,
1229
+ "type": "KSampler",
1230
+ "pos": [
1231
+ 992.5,
1232
+ 50
1233
+ ],
1234
+ "size": [
1235
+ 315,
1236
+ 262
1237
+ ],
1238
+ "flags": {},
1239
+ "order": 29,
1240
+ "mode": 0,
1241
+ "inputs": [
1242
+ {
1243
+ "name": "model",
1244
+ "type": "MODEL",
1245
+ "link": 95
1246
+ },
1247
+ {
1248
+ "name": "positive",
1249
+ "type": "CONDITIONING",
1250
+ "link": 98
1251
+ },
1252
+ {
1253
+ "name": "negative",
1254
+ "type": "CONDITIONING",
1255
+ "link": 99
1256
+ },
1257
+ {
1258
+ "name": "latent_image",
1259
+ "type": "LATENT",
1260
+ "link": 160
1261
+ }
1262
+ ],
1263
+ "outputs": [
1264
+ {
1265
+ "name": "LATENT",
1266
+ "type": "LATENT",
1267
+ "slot_index": 0,
1268
+ "links": [
1269
+ 116
1270
+ ]
1271
+ }
1272
+ ],
1273
+ "properties": {
1274
+ "cnr_id": "comfy-core",
1275
+ "ver": "0.3.34",
1276
+ "Node name for S&R": "KSampler"
1277
+ },
1278
+ "widgets_values": [
1279
+ 776767434080331,
1280
+ "randomize",
1281
+ 3,
1282
+ 1,
1283
+ "uni_pc",
1284
+ "simple",
1285
+ 1
1286
+ ]
1287
+ },
1288
+ {
1289
+ "id": 135,
1290
+ "type": "PreviewImage",
1291
+ "pos": [
1292
+ 630,
1293
+ 690
1294
+ ],
1295
+ "size": [
1296
+ 390,
1297
+ 310
1298
+ ],
1299
+ "flags": {},
1300
+ "order": 24,
1301
+ "mode": 0,
1302
+ "inputs": [
1303
+ {
1304
+ "name": "images",
1305
+ "type": "IMAGE",
1306
+ "link": 214
1307
+ }
1308
+ ],
1309
+ "outputs": [],
1310
+ "title": "Preview Image - Outpaint",
1311
+ "properties": {
1312
+ "cnr_id": "comfy-core",
1313
+ "ver": "0.3.38",
1314
+ "Node name for S&R": "PreviewImage"
1315
+ },
1316
+ "widgets_values": []
1317
+ },
1318
+ {
1319
+ "id": 136,
1320
+ "type": "Note",
1321
+ "pos": [
1322
+ 1060,
1323
+ 690
1324
+ ],
1325
+ "size": [
1326
+ 240,
1327
+ 180
1328
+ ],
1329
+ "flags": {},
1330
+ "order": 11,
1331
+ "mode": 0,
1332
+ "inputs": [],
1333
+ "outputs": [],
1334
+ "properties": {},
1335
+ "widgets_values": [
1336
+ "Currently (May 18, 2025), the Comfy Core nodes lack relevant nodes for obtaining image size or video frames. Here, we can use the Preview image to indirectly obtain the image size and the total number of video frames.\n\nYou can use custom nodes such as image size acquisition to optimize the step of size modification.\n"
1337
+ ],
1338
+ "color": "#432",
1339
+ "bgcolor": "#653"
1340
+ },
1341
+ {
1342
+ "id": 142,
1343
+ "type": "Note",
1344
+ "pos": [
1345
+ 330,
1346
+ 1120
1347
+ ],
1348
+ "size": [
1349
+ 250,
1350
+ 210
1351
+ ],
1352
+ "flags": {},
1353
+ "order": 12,
1354
+ "mode": 0,
1355
+ "inputs": [],
1356
+ "outputs": [],
1357
+ "properties": {},
1358
+ "widgets_values": [
1359
+ "Here, you need to manually set the size of the video after outpainting. Since the native nodes currently do not support the relevant functions, you can use some custom nodes to improve this step. Additionally, because the step size of the size setting for the WanVaceToVideo node is 16, the number you enter must be divisible by 16. "
1360
+ ],
1361
+ "color": "#432",
1362
+ "bgcolor": "#653"
1363
+ },
1364
+ {
1365
+ "id": 156,
1366
+ "type": "MarkdownNote",
1367
+ "pos": [
1368
+ -680,
1369
+ 560
1370
+ ],
1371
+ "size": [
1372
+ 370,
1373
+ 220
1374
+ ],
1375
+ "flags": {},
1376
+ "order": 13,
1377
+ "mode": 0,
1378
+ "inputs": [],
1379
+ "outputs": [],
1380
+ "title": "Note",
1381
+ "properties": {},
1382
+ "widgets_values": [
1383
+ "The generation quality of the 14B model is better, but it takes longer to generate. If you have already downloaded the model, you can choose to directly use the nodes above, or just modify the model loaded by the loader node.\n\nThe corresponding LoRA should match the Diffusion Model. For example, the LoRA corresponding to the 14B diffusion model is the 14B LoRA. "
1434
1384
  ],
1435
- "color": "#322",
1436
- "bgcolor": "#533"
1385
+ "color": "#432",
1386
+ "bgcolor": "#653"
1437
1387
  },
1438
1388
  {
1439
1389
  "id": 183,
@@ -1447,13 +1397,37 @@
1447
1397
  780
1448
1398
  ],
1449
1399
  "flags": {},
1400
+ "order": 14,
1401
+ "mode": 0,
1402
+ "inputs": [],
1403
+ "outputs": [],
1404
+ "properties": {},
1405
+ "widgets_values": [
1406
+ "[Tutorial](https://docs.comfy.org/tutorials/video/wan/vace)\n\n## 14B Support 480P 720P\n\n**Diffusion Model**\n- [wan2.1_vace_14B_fp16.safetensors](https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/resolve/main/split_files/diffusion_models/wan2.1_vace_14B_fp16.safetensors)\n\n**LoRA**\n- [Wan21_CausVid_14B_T2V_lora_rank32.safetensors](https://huggingface.co/Kijai/WanVideo_comfy/blob/main/Wan21_CausVid_14B_T2V_lora_rank32.safetensors)\n\nFYI: It takes about 40 minutes to run to completion at 81 frames per second in 720P resolution with the RTX 4090 when testing the workflow. \nAfter using Wan21_CausVid_14B_T2V_lora_rank32.safetensors, it probably only takes about 4 minutes. \n\n## 1.3B Support 480P only\n\n**Diffusion Model**\n- [wan2.1_vace_1.3B_fp16.safetensors](https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/resolve/main/split_files/diffusion_models/wan2.1_vace_1.3B_fp16.safetensors)\n\n**LoRA**\n- [Wan21_CausVid_bidirect2_T2V_1_3B_lora_rank32.safetensors](https://huggingface.co/Kijai/WanVideo_comfy/blob/main/Wan21_CausVid_bidirect2_T2V_1_3B_lora_rank32.safetensors)\n\nFYI: After using Wan21_CausVid_bidirect2_T2V_1_3B_lora_rank32.safetensors, it probably only takes about 2 minutes. \n## Other Models\n\n* You may already have these models if you use Wan workflow before.\n\n**VAE**\n- [wan_2.1_vae.safetensors](https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/resolve/main/split_files/vae/wan_2.1_vae.safetensors?download=true)\n\n**Text encoders** Chose one of following model\n- [umt5_xxl_fp16.safetensors](https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/resolve/main/split_files/text_encoders/umt5_xxl_fp16.safetensors?download=true)\n- [umt5_xxl_fp8_e4m3fn_scaled.safetensors](https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/resolve/main/split_files/text_encoders/umt5_xxl_fp8_e4m3fn_scaled.safetensors?download=true)\n\n> You can choose between fp16 of fp8; I used fp16 to match what kijai's wrapper is compatible with.\n\nFile save location\n\n```\nComfyUI/\n├── models/\n│ ├── diffusion_models/\n│ │ ├-── wan2.1_vace_14B_fp16.safetensors\n│ │ └─── wan2.1_vace_1.3B_fp16.safetensors \n│ ├── text_encoders/\n│ │ └─── umt5_xxl_fp8_e4m3fn_scaled.safetensors # or fp16\n│ ├── loras/\n│ │ ├── Wan21_CausVid_14B_T2V_lora_rank32.safetensors\n│ │ └── Wan21_CausVid_bidirect2_T2V_1_3B_lora_rank32.safetensors\n│ └── vae/\n│ └── wan_2.1_vae.safetensors\n```\n"
1407
+ ],
1408
+ "color": "#432",
1409
+ "bgcolor": "#653"
1410
+ },
1411
+ {
1412
+ "id": 184,
1413
+ "type": "MarkdownNote",
1414
+ "pos": [
1415
+ -290,
1416
+ -190
1417
+ ],
1418
+ "size": [
1419
+ 410,
1420
+ 140
1421
+ ],
1422
+ "flags": {},
1450
1423
  "order": 15,
1451
1424
  "mode": 0,
1452
1425
  "inputs": [],
1453
1426
  "outputs": [],
1427
+ "title": "About CausVid LoRA",
1454
1428
  "properties": {},
1455
1429
  "widgets_values": [
1456
- "[Tutorial](https://docs.comfy.org/tutorials/video/wan/vace) | [教程](https://docs.comfy.org/zh-CN/tutorials/video/wan/vace)\n\n[Causvid Lora extracted by Kijai](https://www.reddit.com/r/StableDiffusion/comments/1knuafk/causvid_lora_massive_speedup_for_wan21_made_by/) Thanks to CausVid MIT\n\n## 14B Support 480P 720P\n\n**Diffusion Model**\n- [wan2.1_vace_14B_fp16.safetensors](https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/resolve/main/split_files/diffusion_models/wan2.1_vace_14B_fp16.safetensors)\n\n**LoRA**\n- [Wan21_CausVid_14B_T2V_lora_rank32.safetensors](https://huggingface.co/Kijai/WanVideo_comfy/blob/main/Wan21_CausVid_14B_T2V_lora_rank32.safetensors)\n\nFYI: It takes about 40 minutes to run to completion at 81 frames per second in 720P resolution with the RTX 4090 when testing the workflow. \nAfter using Wan21_CausVid_14B_T2V_lora_rank32.safetensors, it probably only takes about 4 minutes. \n\n## 1.3B Support 480P only\n\n**Diffusion Model**\n- [wan2.1_vace_1.3B_fp16.safetensors](https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/resolve/main/split_files/diffusion_models/wan2.1_vace_1.3B_fp16.safetensors)\n\n**LoRA**\n- [Wan21_CausVid_bidirect2_T2V_1_3B_lora_rank32.safetensors](https://huggingface.co/Kijai/WanVideo_comfy/blob/main/Wan21_CausVid_bidirect2_T2V_1_3B_lora_rank32.safetensors)\n\nFYI: After using Wan21_CausVid_bidirect2_T2V_1_3B_lora_rank32.safetensors, it probably only takes about 2 minutes. \n## Other Models\n\n* You may already have these models if you use Wan workflow before.\n\n**VAE**\n- [wan_2.1_vae.safetensors](https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/resolve/main/split_files/vae/wan_2.1_vae.safetensors?download=true)\n\n**Text encoders** Chose one of following model\n- [umt5_xxl_fp16.safetensors](https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/resolve/main/split_files/text_encoders/umt5_xxl_fp16.safetensors?download=true)\n- [umt5_xxl_fp8_e4m3fn_scaled.safetensors](https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/resolve/main/split_files/text_encoders/umt5_xxl_fp8_e4m3fn_scaled.safetensors?download=true)\n\n> You can choose between fp16 of fp8; I used fp16 to match what kijai's wrapper is compatible with.\n\nFile save location\n\n```\nComfyUI/\n├── models/\n│ ├── diffusion_models/\n│ │ ├-── wan2.1_vace_14B_fp16.safetensors\n│ │ └─── wan2.1_vace_1.3B_fp16.safetensors \n│ ├── text_encoders/\n│ │ └─── umt5_xxl_fp8_e4m3fn_scaled.safetensors # or fp16\n│ ├── loras/\n│ │ ├── Wan21_CausVid_14B_T2V_lora_rank32.safetensors\n│ │ └── Wan21_CausVid_bidirect2_T2V_1_3B_lora_rank32.safetensors\n│ └── vae/\n│ └── wan_2.1_vae.safetensors\n```\n"
1430
+ "We have added CausVid LoRA by default to achieve acceleration. However, in some cases, the video may shake and become blurry. You might need to test different LoRA intensities to get the best results, which should be between 0.3 and 0.7. If you don't need it, you can use the bypass mode to disable it, and then restore the settings of `KSampler` to the default ones.\n"
1457
1431
  ],
1458
1432
  "color": "#432",
1459
1433
  "bgcolor": "#653"
@@ -1894,13 +1868,13 @@
1894
1868
  "config": {},
1895
1869
  "extra": {
1896
1870
  "ds": {
1897
- "scale": 0.9849732675808905,
1871
+ "scale": 0.625243500088648,
1898
1872
  "offset": [
1899
- 1963.713933685076,
1900
- 287.5328110010453
1873
+ 1348.117482102405,
1874
+ 445.1571266519913
1901
1875
  ]
1902
1876
  },
1903
- "frontendVersion": "1.23.4",
1877
+ "frontendVersion": "1.27.10",
1904
1878
  "node_versions": {
1905
1879
  "comfy-core": "0.3.34"
1906
1880
  },