comfyui-workflow-templates 0.1.35__py3-none-any.whl → 0.1.37__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- comfyui_workflow_templates/templates/3d_hunyuan3d_image_to_model-1.webp +0 -0
- comfyui_workflow_templates/templates/3d_hunyuan3d_image_to_model.json +317 -288
- comfyui_workflow_templates/templates/3d_hunyuan3d_multiview_to_model-1.webp +0 -0
- comfyui_workflow_templates/templates/3d_hunyuan3d_multiview_to_model-2.webp +0 -0
- comfyui_workflow_templates/templates/3d_hunyuan3d_multiview_to_model.json +272 -236
- comfyui_workflow_templates/templates/3d_hunyuan3d_multiview_to_model_turbo-1.webp +0 -0
- comfyui_workflow_templates/templates/3d_hunyuan3d_multiview_to_model_turbo-2.webp +0 -0
- comfyui_workflow_templates/templates/3d_hunyuan3d_multiview_to_model_turbo.json +203 -167
- comfyui_workflow_templates/templates/area_composition-1.webp +0 -0
- comfyui_workflow_templates/templates/area_composition.json +1158 -519
- comfyui_workflow_templates/templates/area_composition_square_area_for_subject-1.webp +0 -0
- comfyui_workflow_templates/templates/area_composition_square_area_for_subject.json +856 -363
- comfyui_workflow_templates/templates/audio_stable_audio_example.json +201 -48
- comfyui_workflow_templates/templates/controlnet_example.json +625 -208
- comfyui_workflow_templates/templates/flux_kontext_dev_grouped.json +245 -1085
- comfyui_workflow_templates/templates/flux_redux_model_example.json +1112 -617
- comfyui_workflow_templates/templates/hiresfix_latent_workflow-1.webp +0 -0
- comfyui_workflow_templates/templates/hiresfix_latent_workflow-2.webp +0 -0
- comfyui_workflow_templates/templates/hiresfix_latent_workflow.json +556 -215
- comfyui_workflow_templates/templates/image2image.json +434 -124
- comfyui_workflow_templates/templates/image_lotus_depth_v1_1.json +35 -22
- comfyui_workflow_templates/templates/index.json +3 -11
- comfyui_workflow_templates/templates/inpaint_example.json +473 -157
- comfyui_workflow_templates/templates/inpaint_model_outpainting.json +507 -175
- comfyui_workflow_templates/templates/latent_upscale_different_prompt_model-1.webp +0 -0
- comfyui_workflow_templates/templates/latent_upscale_different_prompt_model.json +682 -275
- comfyui_workflow_templates/templates/video_wan_vace_14B_ref2v.json +187 -187
- comfyui_workflow_templates/templates/video_wan_vace_14B_t2v.json +151 -151
- comfyui_workflow_templates/templates/video_wan_vace_14B_v2v.json +139 -139
- comfyui_workflow_templates/templates/video_wan_vace_flf2v.json +182 -182
- comfyui_workflow_templates/templates/video_wan_vace_inpainting.json +141 -167
- comfyui_workflow_templates/templates/video_wan_vace_outpainting.json +34 -34
- {comfyui_workflow_templates-0.1.35.dist-info → comfyui_workflow_templates-0.1.37.dist-info}/METADATA +1 -1
- {comfyui_workflow_templates-0.1.35.dist-info → comfyui_workflow_templates-0.1.37.dist-info}/RECORD +37 -39
- comfyui_workflow_templates/templates/area_composition_reversed-1.webp +0 -0
- comfyui_workflow_templates/templates/area_composition_reversed.json +0 -967
- comfyui_workflow_templates/templates/flux_dev_example.json +0 -771
- {comfyui_workflow_templates-0.1.35.dist-info → comfyui_workflow_templates-0.1.37.dist-info}/WHEEL +0 -0
- {comfyui_workflow_templates-0.1.35.dist-info → comfyui_workflow_templates-0.1.37.dist-info}/licenses/LICENSE +0 -0
- {comfyui_workflow_templates-0.1.35.dist-info → comfyui_workflow_templates-0.1.37.dist-info}/top_level.txt +0 -0
@@ -1,33 +1,9 @@
|
|
1
1
|
{
|
2
2
|
"id": "96995b8f-85c5-47af-b3cf-7b6a24675694",
|
3
3
|
"revision": 0,
|
4
|
-
"last_node_id":
|
4
|
+
"last_node_id": 164,
|
5
5
|
"last_link_id": 295,
|
6
6
|
"nodes": [
|
7
|
-
{
|
8
|
-
"id": 120,
|
9
|
-
"type": "MarkdownNote",
|
10
|
-
"pos": [
|
11
|
-
-710,
|
12
|
-
510
|
13
|
-
],
|
14
|
-
"size": [
|
15
|
-
370,
|
16
|
-
220
|
17
|
-
],
|
18
|
-
"flags": {},
|
19
|
-
"order": 0,
|
20
|
-
"mode": 0,
|
21
|
-
"inputs": [],
|
22
|
-
"outputs": [],
|
23
|
-
"title": "Note",
|
24
|
-
"properties": {},
|
25
|
-
"widgets_values": [
|
26
|
-
"The generation quality of the 14B model is better, but it takes longer to generate. If you have already downloaded the model, you can choose to directly use the nodes above, or just modify the model loaded by the loader node.\n\nThe corresponding LoRA should match the Diffusion Model. For example, the LoRA corresponding to the 14B diffusion model is the 14B LoRA. \n\n---\n\n14B 的生成质量更好,但是需要更长的时间去生成,如果你下载好了模型,你可以选择直接使用上面的节点,或者只是修改 loader 节点加载的模型。\n\n对应的 LoRA 应该是和 Diffusion Model匹配的,比如 14B 的 diffusion model 对应的是 14B 的 LoRA"
|
27
|
-
],
|
28
|
-
"color": "#432",
|
29
|
-
"bgcolor": "#653"
|
30
|
-
},
|
31
7
|
{
|
32
8
|
"id": 124,
|
33
9
|
"type": "UNETLoader",
|
@@ -40,7 +16,7 @@
|
|
40
16
|
82
|
41
17
|
],
|
42
18
|
"flags": {},
|
43
|
-
"order":
|
19
|
+
"order": 0,
|
44
20
|
"mode": 4,
|
45
21
|
"inputs": [],
|
46
22
|
"outputs": [
|
@@ -54,9 +30,9 @@
|
|
54
30
|
}
|
55
31
|
],
|
56
32
|
"properties": {
|
33
|
+
"Node name for S&R": "UNETLoader",
|
57
34
|
"cnr_id": "comfy-core",
|
58
35
|
"ver": "0.3.34",
|
59
|
-
"Node name for S&R": "UNETLoader",
|
60
36
|
"models": [
|
61
37
|
{
|
62
38
|
"name": "wan2.1_vace_1.3B_fp16.safetensors",
|
@@ -111,9 +87,9 @@
|
|
111
87
|
}
|
112
88
|
],
|
113
89
|
"properties": {
|
90
|
+
"Node name for S&R": "LoraLoader",
|
114
91
|
"cnr_id": "comfy-core",
|
115
92
|
"ver": "0.3.38",
|
116
|
-
"Node name for S&R": "LoraLoader",
|
117
93
|
"models": [
|
118
94
|
{
|
119
95
|
"name": "Wan21_CausVid_bidirect2_T2V_1_3B_lora_rank32.safetensors",
|
@@ -142,7 +118,7 @@
|
|
142
118
|
106
|
143
119
|
],
|
144
120
|
"flags": {},
|
145
|
-
"order":
|
121
|
+
"order": 1,
|
146
122
|
"mode": 4,
|
147
123
|
"inputs": [],
|
148
124
|
"outputs": [
|
@@ -154,9 +130,9 @@
|
|
154
130
|
}
|
155
131
|
],
|
156
132
|
"properties": {
|
133
|
+
"Node name for S&R": "CLIPLoader",
|
157
134
|
"cnr_id": "comfy-core",
|
158
135
|
"ver": "0.3.34",
|
159
|
-
"Node name for S&R": "CLIPLoader",
|
160
136
|
"models": [
|
161
137
|
{
|
162
138
|
"name": "umt5_xxl_fp8_e4m3fn_scaled.safetensors",
|
@@ -213,9 +189,9 @@
|
|
213
189
|
}
|
214
190
|
],
|
215
191
|
"properties": {
|
192
|
+
"Node name for S&R": "VAEDecode",
|
216
193
|
"cnr_id": "comfy-core",
|
217
|
-
"ver": "0.3.34"
|
218
|
-
"Node name for S&R": "VAEDecode"
|
194
|
+
"ver": "0.3.34"
|
219
195
|
},
|
220
196
|
"widgets_values": []
|
221
197
|
},
|
@@ -231,7 +207,7 @@
|
|
231
207
|
106
|
232
208
|
],
|
233
209
|
"flags": {},
|
234
|
-
"order":
|
210
|
+
"order": 2,
|
235
211
|
"mode": 0,
|
236
212
|
"inputs": [],
|
237
213
|
"outputs": [
|
@@ -246,9 +222,9 @@
|
|
246
222
|
}
|
247
223
|
],
|
248
224
|
"properties": {
|
225
|
+
"Node name for S&R": "CLIPLoader",
|
249
226
|
"cnr_id": "comfy-core",
|
250
227
|
"ver": "0.3.34",
|
251
|
-
"Node name for S&R": "CLIPLoader",
|
252
228
|
"models": [
|
253
229
|
{
|
254
230
|
"name": "umt5_xxl_fp16.safetensors",
|
@@ -288,9 +264,9 @@
|
|
288
264
|
],
|
289
265
|
"outputs": [],
|
290
266
|
"properties": {
|
267
|
+
"Node name for S&R": "SaveAnimatedWEBP",
|
291
268
|
"cnr_id": "comfy-core",
|
292
|
-
"ver": "0.3.34"
|
293
|
-
"Node name for S&R": "SaveAnimatedWEBP"
|
269
|
+
"ver": "0.3.34"
|
294
270
|
},
|
295
271
|
"widgets_values": [
|
296
272
|
"ComfyUI",
|
@@ -341,9 +317,9 @@
|
|
341
317
|
}
|
342
318
|
],
|
343
319
|
"properties": {
|
320
|
+
"Node name for S&R": "TrimVideoLatent",
|
344
321
|
"cnr_id": "comfy-core",
|
345
|
-
"ver": "0.3.34"
|
346
|
-
"Node name for S&R": "TrimVideoLatent"
|
322
|
+
"ver": "0.3.34"
|
347
323
|
},
|
348
324
|
"widgets_values": [
|
349
325
|
0
|
@@ -383,9 +359,9 @@
|
|
383
359
|
}
|
384
360
|
],
|
385
361
|
"properties": {
|
362
|
+
"Node name for S&R": "ModelSamplingSD3",
|
386
363
|
"cnr_id": "comfy-core",
|
387
|
-
"ver": "0.3.34"
|
388
|
-
"Node name for S&R": "ModelSamplingSD3"
|
364
|
+
"ver": "0.3.34"
|
389
365
|
},
|
390
366
|
"widgets_values": [
|
391
367
|
8.000000000000002
|
@@ -433,9 +409,9 @@
|
|
433
409
|
}
|
434
410
|
],
|
435
411
|
"properties": {
|
412
|
+
"Node name for S&R": "SolidMask",
|
436
413
|
"cnr_id": "comfy-core",
|
437
|
-
"ver": "0.3.38"
|
438
|
-
"Node name for S&R": "SolidMask"
|
414
|
+
"ver": "0.3.38"
|
439
415
|
},
|
440
416
|
"widgets_values": [
|
441
417
|
0,
|
@@ -474,9 +450,9 @@
|
|
474
450
|
}
|
475
451
|
],
|
476
452
|
"properties": {
|
453
|
+
"Node name for S&R": "MaskToImage",
|
477
454
|
"cnr_id": "comfy-core",
|
478
|
-
"ver": "0.3.38"
|
479
|
-
"Node name for S&R": "MaskToImage"
|
455
|
+
"ver": "0.3.38"
|
480
456
|
},
|
481
457
|
"widgets_values": []
|
482
458
|
},
|
@@ -517,9 +493,9 @@
|
|
517
493
|
}
|
518
494
|
],
|
519
495
|
"properties": {
|
496
|
+
"Node name for S&R": "ImageBatch",
|
520
497
|
"cnr_id": "comfy-core",
|
521
|
-
"ver": "0.3.38"
|
522
|
-
"Node name for S&R": "ImageBatch"
|
498
|
+
"ver": "0.3.38"
|
523
499
|
},
|
524
500
|
"widgets_values": []
|
525
501
|
},
|
@@ -554,9 +530,9 @@
|
|
554
530
|
}
|
555
531
|
],
|
556
532
|
"properties": {
|
533
|
+
"Node name for S&R": "ImageToMask",
|
557
534
|
"cnr_id": "comfy-core",
|
558
|
-
"ver": "0.3.38"
|
559
|
-
"Node name for S&R": "ImageToMask"
|
535
|
+
"ver": "0.3.38"
|
560
536
|
},
|
561
537
|
"widgets_values": [
|
562
538
|
"red"
|
@@ -602,9 +578,9 @@
|
|
602
578
|
}
|
603
579
|
],
|
604
580
|
"properties": {
|
581
|
+
"Node name for S&R": "RepeatImageBatch",
|
605
582
|
"cnr_id": "comfy-core",
|
606
|
-
"ver": "0.3.38"
|
607
|
-
"Node name for S&R": "RepeatImageBatch"
|
583
|
+
"ver": "0.3.38"
|
608
584
|
},
|
609
585
|
"widgets_values": [
|
610
586
|
80
|
@@ -643,9 +619,9 @@
|
|
643
619
|
}
|
644
620
|
],
|
645
621
|
"properties": {
|
622
|
+
"Node name for S&R": "MaskToImage",
|
646
623
|
"cnr_id": "comfy-core",
|
647
|
-
"ver": "0.3.38"
|
648
|
-
"Node name for S&R": "MaskToImage"
|
624
|
+
"ver": "0.3.38"
|
649
625
|
},
|
650
626
|
"widgets_values": []
|
651
627
|
},
|
@@ -691,9 +667,9 @@
|
|
691
667
|
}
|
692
668
|
],
|
693
669
|
"properties": {
|
670
|
+
"Node name for S&R": "SolidMask",
|
694
671
|
"cnr_id": "comfy-core",
|
695
|
-
"ver": "0.3.38"
|
696
|
-
"Node name for S&R": "SolidMask"
|
672
|
+
"ver": "0.3.38"
|
697
673
|
},
|
698
674
|
"widgets_values": [
|
699
675
|
1,
|
@@ -738,9 +714,9 @@
|
|
738
714
|
}
|
739
715
|
],
|
740
716
|
"properties": {
|
717
|
+
"Node name for S&R": "ImageBatch",
|
741
718
|
"cnr_id": "comfy-core",
|
742
|
-
"ver": "0.3.38"
|
743
|
-
"Node name for S&R": "ImageBatch"
|
719
|
+
"ver": "0.3.38"
|
744
720
|
},
|
745
721
|
"widgets_values": []
|
746
722
|
},
|
@@ -767,9 +743,9 @@
|
|
767
743
|
],
|
768
744
|
"outputs": [],
|
769
745
|
"properties": {
|
746
|
+
"Node name for S&R": "PreviewImage",
|
770
747
|
"cnr_id": "comfy-core",
|
771
|
-
"ver": "0.3.38"
|
772
|
-
"Node name for S&R": "PreviewImage"
|
748
|
+
"ver": "0.3.38"
|
773
749
|
},
|
774
750
|
"widgets_values": []
|
775
751
|
},
|
@@ -796,36 +772,12 @@
|
|
796
772
|
],
|
797
773
|
"outputs": [],
|
798
774
|
"properties": {
|
775
|
+
"Node name for S&R": "PreviewImage",
|
799
776
|
"cnr_id": "comfy-core",
|
800
|
-
"ver": "0.3.38"
|
801
|
-
"Node name for S&R": "PreviewImage"
|
777
|
+
"ver": "0.3.38"
|
802
778
|
},
|
803
779
|
"widgets_values": []
|
804
780
|
},
|
805
|
-
{
|
806
|
-
"id": 161,
|
807
|
-
"type": "MarkdownNote",
|
808
|
-
"pos": [
|
809
|
-
470,
|
810
|
-
1200
|
811
|
-
],
|
812
|
-
"size": [
|
813
|
-
230,
|
814
|
-
270
|
815
|
-
],
|
816
|
-
"flags": {},
|
817
|
-
"order": 4,
|
818
|
-
"mode": 4,
|
819
|
-
"inputs": [],
|
820
|
-
"outputs": [],
|
821
|
-
"title": "Note",
|
822
|
-
"properties": {},
|
823
|
-
"widgets_values": [
|
824
|
-
"Since VACE supports converting any frame into a video, here we have created a sequence of images with the first frame and the corresponding mask. In this way, we can control the starting frame.\n\n---\n\n由于 VACE 支持任意帧到视频,所以在这里,我们创建了一个带有首帧的序列图像和对应的蒙版,这样我们就可以控制起始帧"
|
825
|
-
],
|
826
|
-
"color": "#432",
|
827
|
-
"bgcolor": "#653"
|
828
|
-
},
|
829
781
|
{
|
830
782
|
"id": 147,
|
831
783
|
"type": "PrimitiveNode",
|
@@ -838,7 +790,7 @@
|
|
838
790
|
82
|
839
791
|
],
|
840
792
|
"flags": {},
|
841
|
-
"order":
|
793
|
+
"order": 3,
|
842
794
|
"mode": 4,
|
843
795
|
"inputs": [],
|
844
796
|
"outputs": [
|
@@ -863,53 +815,6 @@
|
|
863
815
|
"color": "#322",
|
864
816
|
"bgcolor": "#533"
|
865
817
|
},
|
866
|
-
{
|
867
|
-
"id": 162,
|
868
|
-
"type": "MarkdownNote",
|
869
|
-
"pos": [
|
870
|
-
-320,
|
871
|
-
1080
|
872
|
-
],
|
873
|
-
"size": [
|
874
|
-
340,
|
875
|
-
280
|
876
|
-
],
|
877
|
-
"flags": {},
|
878
|
-
"order": 6,
|
879
|
-
"mode": 0,
|
880
|
-
"inputs": [],
|
881
|
-
"outputs": [],
|
882
|
-
"title": "Note",
|
883
|
-
"properties": {},
|
884
|
-
"widgets_values": [
|
885
|
-
"VACE does not use data with style reference for training. Currently, it only has the functions of object or background reference. Therefore, at \"Load reference image\", you should upload a image with a solid-colored background or a background image. At present, the WanVaceToVideo node only supports the input of a single reference_image image.\n\n---\n\nVACE 并没有使用风格参考的数据去训练,目前只有物体或背景参考的功能,所以在 Load reference image 这里你应该上传背景为纯色的图片,或者背景图,目前 WanVaceToVideo 节点只支持输入单个 reference_image 图片"
|
886
|
-
],
|
887
|
-
"color": "#432",
|
888
|
-
"bgcolor": "#653"
|
889
|
-
},
|
890
|
-
{
|
891
|
-
"id": 121,
|
892
|
-
"type": "MarkdownNote",
|
893
|
-
"pos": [
|
894
|
-
-1200,
|
895
|
-
10
|
896
|
-
],
|
897
|
-
"size": [
|
898
|
-
470,
|
899
|
-
800
|
900
|
-
],
|
901
|
-
"flags": {},
|
902
|
-
"order": 7,
|
903
|
-
"mode": 0,
|
904
|
-
"inputs": [],
|
905
|
-
"outputs": [],
|
906
|
-
"properties": {},
|
907
|
-
"widgets_values": [
|
908
|
-
"[Tutorial](https://docs.comfy.org/tutorials/video/wan/vace) | [教程](https://docs.comfy.org/zh-CN/tutorials/video/wan/vace)\n\n[Causvid Lora extracted by Kijai](https://www.reddit.com/r/StableDiffusion/comments/1knuafk/causvid_lora_massive_speedup_for_wan21_made_by/) Thanks to CausVid MIT\n\n## 14B Support 480P 720P\n\n**Diffusion Model**\n- [wan2.1_vace_14B_fp16.safetensors](https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/resolve/main/split_files/diffusion_models/wan2.1_vace_14B_fp16.safetensors)\n\n**LoRA**\n- [Wan21_CausVid_14B_T2V_lora_rank32.safetensors](https://huggingface.co/Kijai/WanVideo_comfy/blob/main/Wan21_CausVid_14B_T2V_lora_rank32.safetensors)\n\nIt takes about 40 minutes to complete at 81 frames 720P resolution with the RTX 4090 . \nAfter using Wan21_CausVid_14B_T2V_lora_rank32.safetensors, it only takes about 4 minutes.\n\n## 1.3B Support 480P only\n\n**Diffusion Model**\n- [wan2.1_vace_1.3B_fp16.safetensors](https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/resolve/main/split_files/diffusion_models/wan2.1_vace_14B_fp16.safetensors)\n\n**LoRA**\n- [Wan21_CausVid_bidirect2_T2V_1_3B_lora_rank32.safetensors](https://huggingface.co/Kijai/WanVideo_comfy/blob/main/Wan21_CausVid_bidirect2_T2V_1_3B_lora_rank32.safetensors)\n \n\n## Other Models\n\n* You may already have these models if you use Wan workflow before.\n\n**VAE**\n- [wan_2.1_vae.safetensors](https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/resolve/main/split_files/vae/wan_2.1_vae.safetensors?download=true)\n\n**Text encoders** Chose one of following model\n- [umt5_xxl_fp16.safetensors](https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/resolve/main/split_files/text_encoders/umt5_xxl_fp16.safetensors?download=true)\n- [umt5_xxl_fp8_e4m3fn_scaled.safetensors](https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/resolve/main/split_files/text_encoders/umt5_xxl_fp8_e4m3fn_scaled.safetensors?download=true)\n\n> You can choose between fp16 of fp8; I used fp16 to match what kijai's wrapper is compatible with.\n\nFile save location\n\n```\nComfyUI/\n├── models/\n│ ├── diffusion_models/\n│ │ ├-── wan2.1_vace_14B_fp16.safetensors\n│ │ └─── wan2.1_vace_1.3B_fp16.safetensors \n│ ├── text_encoders/\n│ │ └─── umt5_xxl_fp8_e4m3fn_scaled.safetensors # or fp16\n│ ├── loras/\n│ │ ├── Wan21_CausVid_14B_T2V_lora_rank32.safetensors\n│ │ └── Wan21_CausVid_bidirect2_T2V_1_3B_lora_rank32.safetensors\n│ └── vae/\n│ └── wan_2.1_vae.safetensors\n```\n"
|
909
|
-
],
|
910
|
-
"color": "#432",
|
911
|
-
"bgcolor": "#653"
|
912
|
-
},
|
913
818
|
{
|
914
819
|
"id": 73,
|
915
820
|
"type": "LoadImage",
|
@@ -922,7 +827,7 @@
|
|
922
827
|
314.00006103515625
|
923
828
|
],
|
924
829
|
"flags": {},
|
925
|
-
"order":
|
830
|
+
"order": 4,
|
926
831
|
"mode": 0,
|
927
832
|
"inputs": [],
|
928
833
|
"outputs": [
|
@@ -940,9 +845,9 @@
|
|
940
845
|
}
|
941
846
|
],
|
942
847
|
"properties": {
|
848
|
+
"Node name for S&R": "LoadImage",
|
943
849
|
"cnr_id": "comfy-core",
|
944
|
-
"ver": "0.3.34"
|
945
|
-
"Node name for S&R": "LoadImage"
|
850
|
+
"ver": "0.3.34"
|
946
851
|
},
|
947
852
|
"widgets_values": [
|
948
853
|
"input.jpg",
|
@@ -963,7 +868,7 @@
|
|
963
868
|
170
|
964
869
|
],
|
965
870
|
"flags": {},
|
966
|
-
"order":
|
871
|
+
"order": 5,
|
967
872
|
"mode": 0,
|
968
873
|
"inputs": [],
|
969
874
|
"outputs": [],
|
@@ -987,7 +892,7 @@
|
|
987
892
|
82
|
988
893
|
],
|
989
894
|
"flags": {},
|
990
|
-
"order":
|
895
|
+
"order": 6,
|
991
896
|
"mode": 0,
|
992
897
|
"inputs": [],
|
993
898
|
"outputs": [
|
@@ -1027,7 +932,7 @@
|
|
1027
932
|
82
|
1028
933
|
],
|
1029
934
|
"flags": {},
|
1030
|
-
"order":
|
935
|
+
"order": 7,
|
1031
936
|
"mode": 0,
|
1032
937
|
"inputs": [],
|
1033
938
|
"outputs": [
|
@@ -1067,7 +972,7 @@
|
|
1067
972
|
58
|
1068
973
|
],
|
1069
974
|
"flags": {},
|
1070
|
-
"order":
|
975
|
+
"order": 8,
|
1071
976
|
"mode": 0,
|
1072
977
|
"inputs": [],
|
1073
978
|
"outputs": [
|
@@ -1082,9 +987,9 @@
|
|
1082
987
|
}
|
1083
988
|
],
|
1084
989
|
"properties": {
|
990
|
+
"Node name for S&R": "VAELoader",
|
1085
991
|
"cnr_id": "comfy-core",
|
1086
992
|
"ver": "0.3.34",
|
1087
|
-
"Node name for S&R": "VAELoader",
|
1088
993
|
"models": [
|
1089
994
|
{
|
1090
995
|
"name": "wan_2.1_vae.safetensors",
|
@@ -1122,9 +1027,9 @@
|
|
1122
1027
|
],
|
1123
1028
|
"outputs": [],
|
1124
1029
|
"properties": {
|
1030
|
+
"Node name for S&R": "SaveVideo",
|
1125
1031
|
"cnr_id": "comfy-core",
|
1126
|
-
"ver": "0.3.34"
|
1127
|
-
"Node name for S&R": "SaveVideo"
|
1032
|
+
"ver": "0.3.34"
|
1128
1033
|
},
|
1129
1034
|
"widgets_values": [
|
1130
1035
|
"video/ComfyUI",
|
@@ -1144,7 +1049,7 @@
|
|
1144
1049
|
314
|
1145
1050
|
],
|
1146
1051
|
"flags": {},
|
1147
|
-
"order":
|
1052
|
+
"order": 9,
|
1148
1053
|
"mode": 4,
|
1149
1054
|
"inputs": [],
|
1150
1055
|
"outputs": [
|
@@ -1162,9 +1067,9 @@
|
|
1162
1067
|
}
|
1163
1068
|
],
|
1164
1069
|
"properties": {
|
1070
|
+
"Node name for S&R": "LoadImage",
|
1165
1071
|
"cnr_id": "comfy-core",
|
1166
|
-
"ver": "0.3.38"
|
1167
|
-
"Node name for S&R": "LoadImage"
|
1072
|
+
"ver": "0.3.38"
|
1168
1073
|
},
|
1169
1074
|
"widgets_values": [
|
1170
1075
|
"input.jpg",
|
@@ -1218,9 +1123,9 @@
|
|
1218
1123
|
}
|
1219
1124
|
],
|
1220
1125
|
"properties": {
|
1126
|
+
"Node name for S&R": "KSampler",
|
1221
1127
|
"cnr_id": "comfy-core",
|
1222
|
-
"ver": "0.3.34"
|
1223
|
-
"Node name for S&R": "KSampler"
|
1128
|
+
"ver": "0.3.34"
|
1224
1129
|
},
|
1225
1130
|
"widgets_values": [
|
1226
1131
|
552837504311715,
|
@@ -1244,7 +1149,7 @@
|
|
1244
1149
|
110
|
1245
1150
|
],
|
1246
1151
|
"flags": {},
|
1247
|
-
"order":
|
1152
|
+
"order": 10,
|
1248
1153
|
"mode": 0,
|
1249
1154
|
"inputs": [],
|
1250
1155
|
"outputs": [],
|
@@ -1295,9 +1200,9 @@
|
|
1295
1200
|
}
|
1296
1201
|
],
|
1297
1202
|
"properties": {
|
1203
|
+
"Node name for S&R": "CreateVideo",
|
1298
1204
|
"cnr_id": "comfy-core",
|
1299
|
-
"ver": "0.3.34"
|
1300
|
-
"Node name for S&R": "CreateVideo"
|
1205
|
+
"ver": "0.3.34"
|
1301
1206
|
},
|
1302
1207
|
"widgets_values": [
|
1303
1208
|
16
|
@@ -1399,9 +1304,9 @@
|
|
1399
1304
|
}
|
1400
1305
|
],
|
1401
1306
|
"properties": {
|
1307
|
+
"Node name for S&R": "WanVaceToVideo",
|
1402
1308
|
"cnr_id": "comfy-core",
|
1403
|
-
"ver": "0.3.34"
|
1404
|
-
"Node name for S&R": "WanVaceToVideo"
|
1309
|
+
"ver": "0.3.34"
|
1405
1310
|
},
|
1406
1311
|
"widgets_values": [
|
1407
1312
|
768,
|
@@ -1423,7 +1328,7 @@
|
|
1423
1328
|
82
|
1424
1329
|
],
|
1425
1330
|
"flags": {},
|
1426
|
-
"order":
|
1331
|
+
"order": 11,
|
1427
1332
|
"mode": 0,
|
1428
1333
|
"inputs": [],
|
1429
1334
|
"outputs": [
|
@@ -1437,9 +1342,9 @@
|
|
1437
1342
|
}
|
1438
1343
|
],
|
1439
1344
|
"properties": {
|
1345
|
+
"Node name for S&R": "UNETLoader",
|
1440
1346
|
"cnr_id": "comfy-core",
|
1441
1347
|
"ver": "0.3.34",
|
1442
|
-
"Node name for S&R": "UNETLoader",
|
1443
1348
|
"models": [
|
1444
1349
|
{
|
1445
1350
|
"name": "wan2.1_vace_14B_fp16.safetensors",
|
@@ -1499,9 +1404,9 @@
|
|
1499
1404
|
}
|
1500
1405
|
],
|
1501
1406
|
"properties": {
|
1407
|
+
"Node name for S&R": "LoraLoader",
|
1502
1408
|
"cnr_id": "comfy-core",
|
1503
1409
|
"ver": "0.3.38",
|
1504
|
-
"Node name for S&R": "LoraLoader",
|
1505
1410
|
"models": [
|
1506
1411
|
{
|
1507
1412
|
"name": "Wan21_CausVid_14B_T2V_lora_rank32.safetensors",
|
@@ -1518,30 +1423,6 @@
|
|
1518
1423
|
"color": "#322",
|
1519
1424
|
"bgcolor": "#533"
|
1520
1425
|
},
|
1521
|
-
{
|
1522
|
-
"id": 163,
|
1523
|
-
"type": "MarkdownNote",
|
1524
|
-
"pos": [
|
1525
|
-
-320,
|
1526
|
-
-180
|
1527
|
-
],
|
1528
|
-
"size": [
|
1529
|
-
410,
|
1530
|
-
140
|
1531
|
-
],
|
1532
|
-
"flags": {},
|
1533
|
-
"order": 16,
|
1534
|
-
"mode": 0,
|
1535
|
-
"inputs": [],
|
1536
|
-
"outputs": [],
|
1537
|
-
"title": "About CausVid LoRA",
|
1538
|
-
"properties": {},
|
1539
|
-
"widgets_values": [
|
1540
|
-
"We have added CausVid LoRA by default to achieve acceleration. However, in some cases, the video may shake and become blurry. You might need to test different LoRA intensities to get the best results, which should be between 0.3 and 0.7. If you don't need it, you can use the bypass mode to disable it, and then restore the settings of `KSampler` to the default ones.\n\n\n我们默认添加了 CausVid LoRA 来实现加速,但有些情况下会出现视频抖动和模糊的情况,你可能需要测试不同的 LoRA 强度来获取最好的结果,0.3~0.7 之间。如果你不需要的话,可以使用 bypass 模式禁用它,然后恢复 `KSampler`的设置到默认的设置即可。"
|
1541
|
-
],
|
1542
|
-
"color": "#432",
|
1543
|
-
"bgcolor": "#653"
|
1544
|
-
},
|
1545
1426
|
{
|
1546
1427
|
"id": 107,
|
1547
1428
|
"type": "CLIPTextEncode",
|
@@ -1575,9 +1456,9 @@
|
|
1575
1456
|
],
|
1576
1457
|
"title": "CLIP Text Encode (Negative Prompt)",
|
1577
1458
|
"properties": {
|
1459
|
+
"Node name for S&R": "CLIPTextEncode",
|
1578
1460
|
"cnr_id": "comfy-core",
|
1579
|
-
"ver": "0.3.34"
|
1580
|
-
"Node name for S&R": "CLIPTextEncode"
|
1461
|
+
"ver": "0.3.34"
|
1581
1462
|
},
|
1582
1463
|
"widgets_values": [
|
1583
1464
|
"过曝,静态,细节模糊不清,字幕,风格,作品,画作,画面,静止,整体发灰,最差质量,低质量,JPEG压缩残留,丑陋的,残缺的,多余的手指,画得不好的手部,画得不好的脸部,畸形的,毁容的,形态畸形的肢体,手指融合,静止不动的画面,杂乱的背景,三条腿,背景人很多,倒着走,"
|
@@ -1618,15 +1499,134 @@
|
|
1618
1499
|
],
|
1619
1500
|
"title": "CLIP Text Encode (Positive Prompt)",
|
1620
1501
|
"properties": {
|
1502
|
+
"Node name for S&R": "CLIPTextEncode",
|
1621
1503
|
"cnr_id": "comfy-core",
|
1622
|
-
"ver": "0.3.34"
|
1623
|
-
"Node name for S&R": "CLIPTextEncode"
|
1504
|
+
"ver": "0.3.34"
|
1624
1505
|
},
|
1625
1506
|
"widgets_values": [
|
1626
1507
|
"An icicle dragon lunges forward, mouth wide open to exhale a stream of icy mist. Ultramarine energy flickers beneath its frost-coated scales as it twists. The camera circles slowly, capturing the swirling ice particles and the backdrop of floating glaciers and frozen nebulae under a cyan-blue filter."
|
1627
1508
|
],
|
1628
1509
|
"color": "#232",
|
1629
1510
|
"bgcolor": "#353"
|
1511
|
+
},
|
1512
|
+
{
|
1513
|
+
"id": 163,
|
1514
|
+
"type": "MarkdownNote",
|
1515
|
+
"pos": [
|
1516
|
+
-320,
|
1517
|
+
-140
|
1518
|
+
],
|
1519
|
+
"size": [
|
1520
|
+
410,
|
1521
|
+
100
|
1522
|
+
],
|
1523
|
+
"flags": {},
|
1524
|
+
"order": 12,
|
1525
|
+
"mode": 0,
|
1526
|
+
"inputs": [],
|
1527
|
+
"outputs": [],
|
1528
|
+
"title": "About CausVid LoRA",
|
1529
|
+
"properties": {},
|
1530
|
+
"widgets_values": [
|
1531
|
+
"We have added CausVid LoRA by default to achieve acceleration. However, in some cases, the video may shake and become blurry. You might need to test different LoRA intensities to get the best results, which should be between 0.3 and 0.7. If you don't need it, you can use the bypass mode to disable it, and then restore the settings of `KSampler` to the default ones."
|
1532
|
+
],
|
1533
|
+
"color": "#432",
|
1534
|
+
"bgcolor": "#653"
|
1535
|
+
},
|
1536
|
+
{
|
1537
|
+
"id": 120,
|
1538
|
+
"type": "MarkdownNote",
|
1539
|
+
"pos": [
|
1540
|
+
-710,
|
1541
|
+
510
|
1542
|
+
],
|
1543
|
+
"size": [
|
1544
|
+
370,
|
1545
|
+
130
|
1546
|
+
],
|
1547
|
+
"flags": {},
|
1548
|
+
"order": 13,
|
1549
|
+
"mode": 0,
|
1550
|
+
"inputs": [],
|
1551
|
+
"outputs": [],
|
1552
|
+
"title": "Note",
|
1553
|
+
"properties": {},
|
1554
|
+
"widgets_values": [
|
1555
|
+
"The generation quality of the 14B model is better, but it takes longer to generate. If you have already downloaded the model, you can choose to directly use the nodes above, or just modify the model loaded by the loader node.\n\nThe corresponding LoRA should match the Diffusion Model. For example, the LoRA corresponding to the 14B diffusion model is the 14B LoRA. \n"
|
1556
|
+
],
|
1557
|
+
"color": "#432",
|
1558
|
+
"bgcolor": "#653"
|
1559
|
+
},
|
1560
|
+
{
|
1561
|
+
"id": 162,
|
1562
|
+
"type": "MarkdownNote",
|
1563
|
+
"pos": [
|
1564
|
+
-320,
|
1565
|
+
1080
|
1566
|
+
],
|
1567
|
+
"size": [
|
1568
|
+
340,
|
1569
|
+
120
|
1570
|
+
],
|
1571
|
+
"flags": {},
|
1572
|
+
"order": 14,
|
1573
|
+
"mode": 0,
|
1574
|
+
"inputs": [],
|
1575
|
+
"outputs": [],
|
1576
|
+
"title": "Note",
|
1577
|
+
"properties": {},
|
1578
|
+
"widgets_values": [
|
1579
|
+
"VACE does not use data with style reference for training. Currently, it only has the functions of object or background reference. Therefore, at \"Load reference image\", you should upload a image with a solid-colored background or a background image. At present, the WanVaceToVideo node only supports the input of a single reference_image image."
|
1580
|
+
],
|
1581
|
+
"color": "#432",
|
1582
|
+
"bgcolor": "#653"
|
1583
|
+
},
|
1584
|
+
{
|
1585
|
+
"id": 164,
|
1586
|
+
"type": "MarkdownNote",
|
1587
|
+
"pos": [
|
1588
|
+
-1200,
|
1589
|
+
10
|
1590
|
+
],
|
1591
|
+
"size": [
|
1592
|
+
470,
|
1593
|
+
730
|
1594
|
+
],
|
1595
|
+
"flags": {},
|
1596
|
+
"order": 15,
|
1597
|
+
"mode": 0,
|
1598
|
+
"inputs": [],
|
1599
|
+
"outputs": [],
|
1600
|
+
"properties": {},
|
1601
|
+
"widgets_values": [
|
1602
|
+
"[Tutorial](https://docs.comfy.org/tutorials/video/wan/vace) | [教程](https://docs.comfy.org/zh-CN/tutorials/video/wan/vace)\n\n[Causvid Lora extracted by Kijai](https://www.reddit.com/r/StableDiffusion/comments/1knuafk/causvid_lora_massive_speedup_for_wan21_made_by/) Thanks to CausVid MIT\n\n## 14B Support 480P 720P\n\n**Diffusion Model**\n- [wan2.1_vace_14B_fp16.safetensors](https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/resolve/main/split_files/diffusion_models/wan2.1_vace_14B_fp16.safetensors)\n\n**LoRA**\n- [Wan21_CausVid_14B_T2V_lora_rank32.safetensors](https://huggingface.co/Kijai/WanVideo_comfy/blob/main/Wan21_CausVid_14B_T2V_lora_rank32.safetensors)\n\nIt takes about 40 minutes to complete at 81 frames 720P resolution with the RTX 4090 . \nAfter using Wan21_CausVid_14B_T2V_lora_rank32.safetensors, it only takes about 4 minutes.\n\n## 1.3B Support 480P only\n\n**Diffusion Model**\n- [wan2.1_vace_1.3B_fp16.safetensors](https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/resolve/main/split_files/diffusion_models/wan2.1_vace_1.3B_fp16.safetensors)\n\n**LoRA**\n- [Wan21_CausVid_bidirect2_T2V_1_3B_lora_rank32.safetensors](https://huggingface.co/Kijai/WanVideo_comfy/blob/main/Wan21_CausVid_bidirect2_T2V_1_3B_lora_rank32.safetensors)\n \n\n## Other Models\n\n* You may already have these models if you use Wan workflow before.\n\n**VAE**\n- [wan_2.1_vae.safetensors](https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/resolve/main/split_files/vae/wan_2.1_vae.safetensors?download=true)\n\n**Text encoders** Chose one of following model\n- [umt5_xxl_fp16.safetensors](https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/resolve/main/split_files/text_encoders/umt5_xxl_fp16.safetensors?download=true)\n- [umt5_xxl_fp8_e4m3fn_scaled.safetensors](https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/resolve/main/split_files/text_encoders/umt5_xxl_fp8_e4m3fn_scaled.safetensors?download=true)\n\n> You can choose between fp16 of fp8; I used fp16 to match what kijai's wrapper is compatible with.\n\nFile save location\n\n```\nComfyUI/\n├── models/\n│ ├── diffusion_models/\n│ │ ├-── wan2.1_vace_14B_fp16.safetensors\n│ │ └─── wan2.1_vace_1.3B_fp16.safetensors \n│ ├── text_encoders/\n│ │ └─── umt5_xxl_fp8_e4m3fn_scaled.safetensors # or fp16\n│ ├── loras/\n│ │ ├── Wan21_CausVid_14B_T2V_lora_rank32.safetensors\n│ │ └── Wan21_CausVid_bidirect2_T2V_1_3B_lora_rank32.safetensors\n│ └── vae/\n│ └── wan_2.1_vae.safetensors\n```\n"
|
1603
|
+
],
|
1604
|
+
"color": "#432",
|
1605
|
+
"bgcolor": "#653"
|
1606
|
+
},
|
1607
|
+
{
|
1608
|
+
"id": 161,
|
1609
|
+
"type": "MarkdownNote",
|
1610
|
+
"pos": [
|
1611
|
+
470,
|
1612
|
+
1200
|
1613
|
+
],
|
1614
|
+
"size": [
|
1615
|
+
230,
|
1616
|
+
110
|
1617
|
+
],
|
1618
|
+
"flags": {},
|
1619
|
+
"order": 16,
|
1620
|
+
"mode": 4,
|
1621
|
+
"inputs": [],
|
1622
|
+
"outputs": [],
|
1623
|
+
"title": "Note",
|
1624
|
+
"properties": {},
|
1625
|
+
"widgets_values": [
|
1626
|
+
"Since VACE supports converting any frame into a video, here we have created a sequence of images with the first frame and the corresponding mask. In this way, we can control the starting frame.\n"
|
1627
|
+
],
|
1628
|
+
"color": "#432",
|
1629
|
+
"bgcolor": "#653"
|
1630
1630
|
}
|
1631
1631
|
],
|
1632
1632
|
"links": [
|
@@ -2146,13 +2146,13 @@
|
|
2146
2146
|
"config": {},
|
2147
2147
|
"extra": {
|
2148
2148
|
"ds": {
|
2149
|
-
"scale":
|
2149
|
+
"scale": 0.4665073802097333,
|
2150
2150
|
"offset": [
|
2151
|
-
|
2152
|
-
|
2151
|
+
1673.8309046403217,
|
2152
|
+
132.15298787291272
|
2153
2153
|
]
|
2154
2154
|
},
|
2155
|
-
"frontendVersion": "1.
|
2155
|
+
"frontendVersion": "1.24.0-1",
|
2156
2156
|
"node_versions": {
|
2157
2157
|
"comfy-core": "0.3.34"
|
2158
2158
|
},
|