comfyui-workflow-templates 0.1.95__py3-none-any.whl ā 0.1.96__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of comfyui-workflow-templates might be problematic. Click here for more details.
- comfyui_workflow_templates/templates/2_pass_pose_worship.json +551 -139
- comfyui_workflow_templates/templates/3d_hunyuan3d-v2.1.json +8 -8
- comfyui_workflow_templates/templates/3d_hunyuan3d_image_to_model.json +176 -284
- comfyui_workflow_templates/templates/3d_hunyuan3d_multiview_to_model.json +131 -133
- comfyui_workflow_templates/templates/3d_hunyuan3d_multiview_to_model_turbo.json +84 -86
- comfyui_workflow_templates/templates/ByteDance-Seedance_00003_.json +210 -0
- comfyui_workflow_templates/templates/api_bfl_flux_1_kontext_max_image.json +73 -14
- comfyui_workflow_templates/templates/api_bfl_flux_1_kontext_multiple_images_input.json +86 -23
- comfyui_workflow_templates/templates/api_bfl_flux_1_kontext_pro_image.json +145 -86
- comfyui_workflow_templates/templates/api_bfl_flux_pro_t2i.json +78 -78
- comfyui_workflow_templates/templates/api_bytedance_flf2v.json +86 -86
- comfyui_workflow_templates/templates/api_bytedance_image_to_video.json +36 -36
- comfyui_workflow_templates/templates/api_bytedance_seedream4.json +227 -222
- comfyui_workflow_templates/templates/api_google_gemini_image.json +113 -47
- comfyui_workflow_templates/templates/api_hailuo_minimax_i2v.json +52 -42
- comfyui_workflow_templates/templates/api_hailuo_minimax_video.json +10 -4
- comfyui_workflow_templates/templates/api_ideogram_v3_t2i.json +50 -48
- comfyui_workflow_templates/templates/api_kling_effects.json +34 -9
- comfyui_workflow_templates/templates/api_kling_flf.json +68 -39
- comfyui_workflow_templates/templates/api_kling_i2v.json +35 -9
- comfyui_workflow_templates/templates/api_luma_i2v.json +124 -110
- comfyui_workflow_templates/templates/api_luma_photon_i2i.json +64 -25
- comfyui_workflow_templates/templates/api_luma_photon_style_ref.json +210 -60
- comfyui_workflow_templates/templates/api_moonvalley_image_to_video.json +40 -41
- comfyui_workflow_templates/templates/api_moonvalley_video_to_video_motion_transfer.json +10 -11
- comfyui_workflow_templates/templates/api_moonvalley_video_to_video_pose_control.json +36 -37
- comfyui_workflow_templates/templates/api_openai_dall_e_2_inpaint.json +47 -11
- comfyui_workflow_templates/templates/api_openai_image_1_i2i.json +37 -8
- comfyui_workflow_templates/templates/api_openai_image_1_inpaint.json +29 -29
- comfyui_workflow_templates/templates/api_openai_image_1_multi_inputs.json +56 -13
- comfyui_workflow_templates/templates/api_openai_image_1_t2i.json +28 -28
- comfyui_workflow_templates/templates/api_openai_sora_video.json +203 -162
- comfyui_workflow_templates/templates/api_pika_i2v.json +29 -29
- comfyui_workflow_templates/templates/api_pika_scene.json +9 -9
- comfyui_workflow_templates/templates/api_pixverse_i2v.json +58 -58
- comfyui_workflow_templates/templates/api_pixverse_template_i2v.json +4 -4
- comfyui_workflow_templates/templates/api_recraft_image_gen_with_color_control.json +168 -73
- comfyui_workflow_templates/templates/api_rodin_gen2.json +191 -185
- comfyui_workflow_templates/templates/api_rodin_image_to_model.json +287 -247
- comfyui_workflow_templates/templates/api_rodin_multiview_to_model.json +7 -7
- comfyui_workflow_templates/templates/api_runway_first_last_frame.json +44 -44
- comfyui_workflow_templates/templates/api_runway_gen3a_turbo_image_to_video.json +29 -29
- comfyui_workflow_templates/templates/api_runway_gen4_turo_image_to_video.json +40 -40
- comfyui_workflow_templates/templates/api_runway_reference_to_image.json +84 -45
- comfyui_workflow_templates/templates/api_runway_text_to_image.json +40 -40
- comfyui_workflow_templates/templates/api_stability_ai_audio_inpaint.json +59 -83
- comfyui_workflow_templates/templates/api_stability_ai_audio_to_audio.json +43 -46
- comfyui_workflow_templates/templates/api_stability_ai_i2i.json +39 -14
- comfyui_workflow_templates/templates/api_stability_ai_sd3.5_i2i.json +54 -8
- comfyui_workflow_templates/templates/api_stability_ai_sd3.5_t2i.json +44 -44
- comfyui_workflow_templates/templates/api_stability_ai_stable_image_ultra_t2i.json +28 -28
- comfyui_workflow_templates/templates/api_tripo_image_to_model.json +4 -4
- comfyui_workflow_templates/templates/api_tripo_multiview_to_model.json +89 -89
- comfyui_workflow_templates/templates/api_veo2_i2v.json +3 -3
- comfyui_workflow_templates/templates/api_veo3.json +90 -84
- comfyui_workflow_templates/templates/api_vidu_image_to_video.json +37 -9
- comfyui_workflow_templates/templates/api_vidu_reference_to_video.json +86 -19
- comfyui_workflow_templates/templates/api_vidu_start_end_to_video.json +14 -6
- comfyui_workflow_templates/templates/api_wan_image_to_video.json +73 -73
- comfyui_workflow_templates/templates/audio_ace_step_1_m2m_editing.json +152 -329
- comfyui_workflow_templates/templates/audio_ace_step_1_t2a_instrumentals.json +207 -398
- comfyui_workflow_templates/templates/audio_ace_step_1_t2a_song.json +167 -378
- comfyui_workflow_templates/templates/audio_stable_audio_example.json +156 -118
- comfyui_workflow_templates/templates/controlnet_example.json +110 -122
- comfyui_workflow_templates/templates/default.json +329 -139
- comfyui_workflow_templates/templates/depth_controlnet.json +463 -206
- comfyui_workflow_templates/templates/depth_t2i_adapter.json +1522 -236
- comfyui_workflow_templates/templates/flux1_dev_uso_reference_image_gen.json +383 -354
- comfyui_workflow_templates/templates/flux_canny_model_example.json +576 -268
- comfyui_workflow_templates/templates/flux_depth_lora_example.json +1234 -213
- comfyui_workflow_templates/templates/flux_fill_inpaint_example.json +553 -250
- comfyui_workflow_templates/templates/flux_fill_outpaint_example.json +556 -228
- comfyui_workflow_templates/templates/flux_kontext_dev_basic.json +104 -109
- comfyui_workflow_templates/templates/flux_redux_model_example-1.webp +0 -0
- comfyui_workflow_templates/templates/flux_redux_model_example.json +609 -557
- comfyui_workflow_templates/templates/hidream_e1_1.json +5 -5
- comfyui_workflow_templates/templates/hidream_e1_full.json +5 -5
- comfyui_workflow_templates/templates/image2image-1.webp +0 -0
- comfyui_workflow_templates/templates/image2image-2.webp +0 -0
- comfyui_workflow_templates/templates/image2image.json +203 -201
- comfyui_workflow_templates/templates/image_flux.1_fill_dev_OneReward.json +178 -162
- comfyui_workflow_templates/templates/image_lotus_depth_v1_1.json +4 -4
- comfyui_workflow_templates/templates/image_omnigen2_image_edit.json +347 -341
- comfyui_workflow_templates/templates/image_qwen_image_controlnet_patch.json +71 -70
- comfyui_workflow_templates/templates/image_qwen_image_edit.json +199 -199
- comfyui_workflow_templates/templates/image_qwen_image_edit_2509.json +401 -401
- comfyui_workflow_templates/templates/image_qwen_image_instantx_controlnet.json +72 -71
- comfyui_workflow_templates/templates/image_qwen_image_instantx_inpainting_controlnet.json +248 -244
- comfyui_workflow_templates/templates/image_qwen_image_union_control_lora.json +190 -189
- comfyui_workflow_templates/templates/image_to_video.json +64 -64
- comfyui_workflow_templates/templates/image_to_video_wan.json +163 -140
- comfyui_workflow_templates/templates/index.es.json +2 -25
- comfyui_workflow_templates/templates/index.fr.json +2 -25
- comfyui_workflow_templates/templates/index.ja.json +2 -25
- comfyui_workflow_templates/templates/index.json +1 -24
- comfyui_workflow_templates/templates/index.ko.json +2 -25
- comfyui_workflow_templates/templates/index.ru.json +2 -25
- comfyui_workflow_templates/templates/index.zh-TW.json +2 -25
- comfyui_workflow_templates/templates/index.zh.json +2 -25
- comfyui_workflow_templates/templates/inpaint_example.json +22 -19
- comfyui_workflow_templates/templates/inpaint_model_outpainting.json +23 -20
- comfyui_workflow_templates/templates/ltxv_image_to_video-1.webp +0 -0
- comfyui_workflow_templates/templates/ltxv_image_to_video.json +367 -337
- comfyui_workflow_templates/templates/mixing_controlnets.json +422 -373
- comfyui_workflow_templates/templates/sd3.5_large_blur.json +14 -14
- comfyui_workflow_templates/templates/sd3.5_large_canny_controlnet_example.json +615 -258
- comfyui_workflow_templates/templates/sd3.5_large_depth.json +1317 -210
- comfyui_workflow_templates/templates/sdxl_revision_text_prompts.json +619 -256
- comfyui_workflow_templates/templates/video_humo.json +194 -194
- comfyui_workflow_templates/templates/video_wan2.1_fun_camera_v1.1_1.3B.json +30 -30
- comfyui_workflow_templates/templates/video_wan2.1_fun_camera_v1.1_14B.json +97 -152
- comfyui_workflow_templates/templates/video_wan2_2_14B_animate.json +5 -5
- comfyui_workflow_templates/templates/video_wan2_2_14B_flf2v.json +121 -121
- comfyui_workflow_templates/templates/video_wan2_2_14B_fun_camera.json +67 -67
- comfyui_workflow_templates/templates/video_wan2_2_14B_fun_control.json +380 -358
- comfyui_workflow_templates/templates/video_wan2_2_14B_fun_inpaint.json +111 -111
- comfyui_workflow_templates/templates/video_wan2_2_14B_i2v.json +58 -58
- comfyui_workflow_templates/templates/video_wan2_2_14B_s2v.json +216 -216
- comfyui_workflow_templates/templates/video_wan2_2_5B_fun_control.json +222 -202
- comfyui_workflow_templates/templates/video_wan2_2_5B_fun_inpaint.json +27 -27
- comfyui_workflow_templates/templates/video_wan_ati.json +5 -5
- comfyui_workflow_templates/templates/video_wan_vace_14B_ref2v.json +100 -100
- comfyui_workflow_templates/templates/video_wan_vace_14B_v2v.json +87 -105
- comfyui_workflow_templates/templates/video_wan_vace_flf2v.json +149 -149
- comfyui_workflow_templates/templates/video_wan_vace_inpainting.json +234 -242
- comfyui_workflow_templates/templates/video_wan_vace_outpainting.json +67 -124
- comfyui_workflow_templates/templates/wan2.1_flf2v_720_f16.json +32 -32
- comfyui_workflow_templates/templates/wan2.1_fun_control.json +183 -138
- comfyui_workflow_templates/templates/wan2.1_fun_inp.json +62 -62
- {comfyui_workflow_templates-0.1.95.dist-info ā comfyui_workflow_templates-0.1.96.dist-info}/METADATA +1 -1
- {comfyui_workflow_templates-0.1.95.dist-info ā comfyui_workflow_templates-0.1.96.dist-info}/RECORD +134 -136
- comfyui_workflow_templates/templates/sdxl_revision_zero_positive-1.webp +0 -0
- comfyui_workflow_templates/templates/sdxl_revision_zero_positive.json +0 -496
- comfyui_workflow_templates/templates/stable_zero123_example-1.webp +0 -0
- comfyui_workflow_templates/templates/stable_zero123_example.json +0 -273
- {comfyui_workflow_templates-0.1.95.dist-info ā comfyui_workflow_templates-0.1.96.dist-info}/WHEEL +0 -0
- {comfyui_workflow_templates-0.1.95.dist-info ā comfyui_workflow_templates-0.1.96.dist-info}/licenses/LICENSE +0 -0
- {comfyui_workflow_templates-0.1.95.dist-info ā comfyui_workflow_templates-0.1.96.dist-info}/top_level.txt +0 -0
|
@@ -577,7 +577,7 @@
|
|
|
577
577
|
"ver": "0.3.44"
|
|
578
578
|
},
|
|
579
579
|
"widgets_values": [
|
|
580
|
-
"
|
|
580
|
+
"hidream_e1_1_input_image.jpg",
|
|
581
581
|
"image"
|
|
582
582
|
]
|
|
583
583
|
},
|
|
@@ -1120,13 +1120,13 @@
|
|
|
1120
1120
|
"config": {},
|
|
1121
1121
|
"extra": {
|
|
1122
1122
|
"ds": {
|
|
1123
|
-
"scale": 0.
|
|
1123
|
+
"scale": 0.5148463834529032,
|
|
1124
1124
|
"offset": [
|
|
1125
|
-
|
|
1126
|
-
|
|
1125
|
+
1324.981666105929,
|
|
1126
|
+
1242.1992997217671
|
|
1127
1127
|
]
|
|
1128
1128
|
},
|
|
1129
|
-
"frontendVersion": "1.
|
|
1129
|
+
"frontendVersion": "1.29.1",
|
|
1130
1130
|
"groupNodes": {}
|
|
1131
1131
|
},
|
|
1132
1132
|
"version": 0.4
|
|
@@ -270,7 +270,7 @@
|
|
|
270
270
|
"ver": "0.3.44"
|
|
271
271
|
},
|
|
272
272
|
"widgets_values": [
|
|
273
|
-
"
|
|
273
|
+
"hidream_e1_full_input_image.jpg",
|
|
274
274
|
"image"
|
|
275
275
|
]
|
|
276
276
|
},
|
|
@@ -1008,13 +1008,13 @@
|
|
|
1008
1008
|
"config": {},
|
|
1009
1009
|
"extra": {
|
|
1010
1010
|
"ds": {
|
|
1011
|
-
"scale": 0.
|
|
1011
|
+
"scale": 0.49884556140792824,
|
|
1012
1012
|
"offset": [
|
|
1013
|
-
|
|
1014
|
-
|
|
1013
|
+
804.5433057808879,
|
|
1014
|
+
251.23159836698244
|
|
1015
1015
|
]
|
|
1016
1016
|
},
|
|
1017
|
-
"frontendVersion": "1.
|
|
1017
|
+
"frontendVersion": "1.29.1",
|
|
1018
1018
|
"groupNodes": {}
|
|
1019
1019
|
},
|
|
1020
1020
|
"version": 0.4
|
|
Binary file
|
|
Binary file
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
{
|
|
2
2
|
"id": "7cb6261d-3b03-4171-bbd1-a4b256b42404",
|
|
3
3
|
"revision": 0,
|
|
4
|
-
"last_node_id":
|
|
4
|
+
"last_node_id": 21,
|
|
5
5
|
"last_link_id": 19,
|
|
6
6
|
"nodes": [
|
|
7
7
|
{
|
|
@@ -48,9 +48,9 @@
|
|
|
48
48
|
}
|
|
49
49
|
],
|
|
50
50
|
"properties": {
|
|
51
|
-
"Node name for S&R": "CheckpointLoaderSimple",
|
|
52
51
|
"cnr_id": "comfy-core",
|
|
53
52
|
"ver": "0.3.44",
|
|
53
|
+
"Node name for S&R": "CheckpointLoaderSimple",
|
|
54
54
|
"models": [
|
|
55
55
|
{
|
|
56
56
|
"name": "v1-5-pruned-emaonly-fp16.safetensors",
|
|
@@ -61,276 +61,275 @@
|
|
|
61
61
|
},
|
|
62
62
|
"widgets_values": [
|
|
63
63
|
"v1-5-pruned-emaonly-fp16.safetensors"
|
|
64
|
-
]
|
|
65
|
-
"color": "#322",
|
|
66
|
-
"bgcolor": "#533"
|
|
64
|
+
]
|
|
67
65
|
},
|
|
68
66
|
{
|
|
69
|
-
"id":
|
|
70
|
-
"type": "
|
|
67
|
+
"id": 8,
|
|
68
|
+
"type": "VAEDecode",
|
|
71
69
|
"pos": [
|
|
72
|
-
|
|
73
|
-
|
|
70
|
+
1080,
|
|
71
|
+
180
|
|
74
72
|
],
|
|
75
73
|
"size": [
|
|
76
|
-
|
|
77
|
-
|
|
74
|
+
210,
|
|
75
|
+
46
|
|
78
76
|
],
|
|
79
77
|
"flags": {},
|
|
80
|
-
"order":
|
|
78
|
+
"order": 10,
|
|
81
79
|
"mode": 0,
|
|
82
80
|
"inputs": [
|
|
83
81
|
{
|
|
84
|
-
"name": "
|
|
85
|
-
"type": "
|
|
86
|
-
"link":
|
|
82
|
+
"name": "samples",
|
|
83
|
+
"type": "LATENT",
|
|
84
|
+
"link": 7
|
|
85
|
+
},
|
|
86
|
+
{
|
|
87
|
+
"name": "vae",
|
|
88
|
+
"type": "VAE",
|
|
89
|
+
"link": 17
|
|
87
90
|
}
|
|
88
91
|
],
|
|
89
92
|
"outputs": [
|
|
90
93
|
{
|
|
91
|
-
"name": "
|
|
92
|
-
"type": "
|
|
94
|
+
"name": "IMAGE",
|
|
95
|
+
"type": "IMAGE",
|
|
93
96
|
"slot_index": 0,
|
|
94
97
|
"links": [
|
|
95
|
-
|
|
98
|
+
9
|
|
96
99
|
]
|
|
97
100
|
}
|
|
98
101
|
],
|
|
99
102
|
"properties": {
|
|
100
|
-
"Node name for S&R": "CLIPTextEncode",
|
|
101
103
|
"cnr_id": "comfy-core",
|
|
102
|
-
"ver": "0.3.44"
|
|
104
|
+
"ver": "0.3.44",
|
|
105
|
+
"Node name for S&R": "VAEDecode"
|
|
103
106
|
},
|
|
104
|
-
"widgets_values": [
|
|
105
|
-
"photograph of victorian woman with wings, sky clouds, meadow grass\n"
|
|
106
|
-
],
|
|
107
|
-
"color": "#232",
|
|
108
|
-
"bgcolor": "#353"
|
|
107
|
+
"widgets_values": []
|
|
109
108
|
},
|
|
110
109
|
{
|
|
111
|
-
"id":
|
|
112
|
-
"type": "
|
|
110
|
+
"id": 9,
|
|
111
|
+
"type": "SaveImage",
|
|
113
112
|
"pos": [
|
|
114
|
-
|
|
115
|
-
|
|
113
|
+
1080,
|
|
114
|
+
280
|
|
116
115
|
],
|
|
117
116
|
"size": [
|
|
118
|
-
|
|
119
|
-
|
|
117
|
+
450,
|
|
118
|
+
430
|
|
120
119
|
],
|
|
121
120
|
"flags": {},
|
|
122
|
-
"order":
|
|
121
|
+
"order": 11,
|
|
123
122
|
"mode": 0,
|
|
124
123
|
"inputs": [
|
|
125
124
|
{
|
|
126
|
-
"name": "
|
|
127
|
-
"type": "
|
|
128
|
-
"link":
|
|
125
|
+
"name": "images",
|
|
126
|
+
"type": "IMAGE",
|
|
127
|
+
"link": 9
|
|
129
128
|
}
|
|
130
129
|
],
|
|
130
|
+
"outputs": [],
|
|
131
|
+
"properties": {
|
|
132
|
+
"cnr_id": "comfy-core",
|
|
133
|
+
"ver": "0.3.44"
|
|
134
|
+
},
|
|
135
|
+
"widgets_values": [
|
|
136
|
+
"ComfyUI"
|
|
137
|
+
]
|
|
138
|
+
},
|
|
139
|
+
{
|
|
140
|
+
"id": 10,
|
|
141
|
+
"type": "LoadImage",
|
|
142
|
+
"pos": [
|
|
143
|
+
-90,
|
|
144
|
+
410
|
|
145
|
+
],
|
|
146
|
+
"size": [
|
|
147
|
+
310,
|
|
148
|
+
314
|
|
149
|
+
],
|
|
150
|
+
"flags": {},
|
|
151
|
+
"order": 1,
|
|
152
|
+
"mode": 0,
|
|
153
|
+
"inputs": [],
|
|
131
154
|
"outputs": [
|
|
132
155
|
{
|
|
133
|
-
"name": "
|
|
134
|
-
"type": "
|
|
156
|
+
"name": "IMAGE",
|
|
157
|
+
"type": "IMAGE",
|
|
135
158
|
"slot_index": 0,
|
|
136
159
|
"links": [
|
|
137
|
-
|
|
160
|
+
18
|
|
138
161
|
]
|
|
162
|
+
},
|
|
163
|
+
{
|
|
164
|
+
"name": "MASK",
|
|
165
|
+
"type": "MASK",
|
|
166
|
+
"links": null
|
|
139
167
|
}
|
|
140
168
|
],
|
|
141
169
|
"properties": {
|
|
142
|
-
"Node name for S&R": "CLIPTextEncode",
|
|
143
170
|
"cnr_id": "comfy-core",
|
|
144
|
-
"ver": "0.3.44"
|
|
171
|
+
"ver": "0.3.44",
|
|
172
|
+
"Node name for S&R": "LoadImage"
|
|
145
173
|
},
|
|
146
174
|
"widgets_values": [
|
|
147
|
-
"
|
|
148
|
-
|
|
149
|
-
|
|
150
|
-
"bgcolor": "#335"
|
|
175
|
+
"image2image_input_image.jpg",
|
|
176
|
+
"image"
|
|
177
|
+
]
|
|
151
178
|
},
|
|
152
179
|
{
|
|
153
|
-
"id":
|
|
154
|
-
"type": "
|
|
180
|
+
"id": 6,
|
|
181
|
+
"type": "CLIPTextEncode",
|
|
155
182
|
"pos": [
|
|
156
|
-
|
|
157
|
-
|
|
183
|
+
280,
|
|
184
|
+
210
|
|
158
185
|
],
|
|
159
186
|
"size": [
|
|
160
|
-
|
|
161
|
-
|
|
187
|
+
422.84503173828125,
|
|
188
|
+
164.31304931640625
|
|
162
189
|
],
|
|
163
190
|
"flags": {},
|
|
164
|
-
"order":
|
|
191
|
+
"order": 5,
|
|
165
192
|
"mode": 0,
|
|
166
193
|
"inputs": [
|
|
167
194
|
{
|
|
168
|
-
"name": "
|
|
169
|
-
"type": "
|
|
170
|
-
"link":
|
|
171
|
-
},
|
|
172
|
-
{
|
|
173
|
-
"name": "positive",
|
|
174
|
-
"type": "CONDITIONING",
|
|
175
|
-
"link": 4
|
|
176
|
-
},
|
|
177
|
-
{
|
|
178
|
-
"name": "negative",
|
|
179
|
-
"type": "CONDITIONING",
|
|
180
|
-
"link": 6
|
|
181
|
-
},
|
|
182
|
-
{
|
|
183
|
-
"name": "latent_image",
|
|
184
|
-
"type": "LATENT",
|
|
185
|
-
"link": 11
|
|
195
|
+
"name": "clip",
|
|
196
|
+
"type": "CLIP",
|
|
197
|
+
"link": 14
|
|
186
198
|
}
|
|
187
199
|
],
|
|
188
200
|
"outputs": [
|
|
189
201
|
{
|
|
190
|
-
"name": "
|
|
191
|
-
"type": "
|
|
202
|
+
"name": "CONDITIONING",
|
|
203
|
+
"type": "CONDITIONING",
|
|
192
204
|
"slot_index": 0,
|
|
193
205
|
"links": [
|
|
194
|
-
|
|
206
|
+
4
|
|
195
207
|
]
|
|
196
208
|
}
|
|
197
209
|
],
|
|
198
210
|
"properties": {
|
|
199
|
-
"Node name for S&R": "KSampler",
|
|
200
211
|
"cnr_id": "comfy-core",
|
|
201
|
-
"ver": "0.3.44"
|
|
212
|
+
"ver": "0.3.44",
|
|
213
|
+
"Node name for S&R": "CLIPTextEncode"
|
|
202
214
|
},
|
|
203
215
|
"widgets_values": [
|
|
204
|
-
|
|
205
|
-
|
|
206
|
-
|
|
207
|
-
|
|
208
|
-
"dpmpp_2m",
|
|
209
|
-
"normal",
|
|
210
|
-
0.8700000000000001
|
|
211
|
-
]
|
|
216
|
+
"Use clouds to represent a woman's profile, which has a real sense of beauty.\n"
|
|
217
|
+
],
|
|
218
|
+
"color": "#232",
|
|
219
|
+
"bgcolor": "#353"
|
|
212
220
|
},
|
|
213
221
|
{
|
|
214
|
-
"id":
|
|
215
|
-
"type": "
|
|
222
|
+
"id": 7,
|
|
223
|
+
"type": "CLIPTextEncode",
|
|
216
224
|
"pos": [
|
|
217
|
-
|
|
218
|
-
|
|
225
|
+
280,
|
|
226
|
+
410
|
|
219
227
|
],
|
|
220
228
|
"size": [
|
|
221
|
-
|
|
222
|
-
|
|
229
|
+
425.27801513671875,
|
|
230
|
+
180.6060791015625
|
|
223
231
|
],
|
|
224
232
|
"flags": {},
|
|
225
|
-
"order":
|
|
233
|
+
"order": 6,
|
|
226
234
|
"mode": 0,
|
|
227
235
|
"inputs": [
|
|
228
236
|
{
|
|
229
|
-
"name": "
|
|
230
|
-
"type": "
|
|
231
|
-
"link":
|
|
232
|
-
},
|
|
233
|
-
{
|
|
234
|
-
"name": "vae",
|
|
235
|
-
"type": "VAE",
|
|
236
|
-
"link": 17
|
|
237
|
+
"name": "clip",
|
|
238
|
+
"type": "CLIP",
|
|
239
|
+
"link": 15
|
|
237
240
|
}
|
|
238
241
|
],
|
|
239
242
|
"outputs": [
|
|
240
243
|
{
|
|
241
|
-
"name": "
|
|
242
|
-
"type": "
|
|
244
|
+
"name": "CONDITIONING",
|
|
245
|
+
"type": "CONDITIONING",
|
|
243
246
|
"slot_index": 0,
|
|
244
247
|
"links": [
|
|
245
|
-
|
|
248
|
+
6
|
|
246
249
|
]
|
|
247
250
|
}
|
|
248
251
|
],
|
|
249
252
|
"properties": {
|
|
250
|
-
"Node name for S&R": "VAEDecode",
|
|
251
253
|
"cnr_id": "comfy-core",
|
|
252
|
-
"ver": "0.3.44"
|
|
254
|
+
"ver": "0.3.44",
|
|
255
|
+
"Node name for S&R": "CLIPTextEncode"
|
|
253
256
|
},
|
|
254
|
-
"widgets_values": [
|
|
257
|
+
"widgets_values": [
|
|
258
|
+
"lowres, bad anatomy, bad hands, text, error, missing fingers, extra digit, fewer digits, cropped, worst quality, low quality, normal quality, jpeg artifacts, signature, watermark, username, blurry"
|
|
259
|
+
],
|
|
260
|
+
"color": "#223",
|
|
261
|
+
"bgcolor": "#335"
|
|
255
262
|
},
|
|
256
263
|
{
|
|
257
|
-
"id":
|
|
258
|
-
"type": "
|
|
264
|
+
"id": 19,
|
|
265
|
+
"type": "MarkdownNote",
|
|
259
266
|
"pos": [
|
|
260
|
-
|
|
261
|
-
|
|
267
|
+
-70,
|
|
268
|
+
930
|
|
262
269
|
],
|
|
263
270
|
"size": [
|
|
264
|
-
|
|
265
|
-
|
|
271
|
+
270,
|
|
272
|
+
120
|
|
266
273
|
],
|
|
267
274
|
"flags": {},
|
|
268
|
-
"order":
|
|
275
|
+
"order": 2,
|
|
269
276
|
"mode": 0,
|
|
270
|
-
"inputs": [
|
|
271
|
-
{
|
|
272
|
-
"name": "images",
|
|
273
|
-
"type": "IMAGE",
|
|
274
|
-
"link": 9
|
|
275
|
-
}
|
|
276
|
-
],
|
|
277
|
+
"inputs": [],
|
|
277
278
|
"outputs": [],
|
|
278
|
-
"
|
|
279
|
-
|
|
280
|
-
"ver": "0.3.44"
|
|
281
|
-
},
|
|
279
|
+
"title": "About Scale Image to Total Pixels",
|
|
280
|
+
"properties": {},
|
|
282
281
|
"widgets_values": [
|
|
283
|
-
"
|
|
284
|
-
]
|
|
282
|
+
"The model this template uses is trained based on a 512*512 image dataset. So if you use an input image that's too large, it might cause some issues. We've added the Scale image to total pixels node to scale images. If you're quite familiar with this model, you can remove it."
|
|
283
|
+
],
|
|
284
|
+
"color": "#432",
|
|
285
|
+
"bgcolor": "#653"
|
|
285
286
|
},
|
|
286
287
|
{
|
|
287
|
-
"id":
|
|
288
|
-
"type": "
|
|
288
|
+
"id": 18,
|
|
289
|
+
"type": "ImageScaleToTotalPixels",
|
|
289
290
|
"pos": [
|
|
290
|
-
-
|
|
291
|
-
|
|
291
|
+
-70,
|
|
292
|
+
800
|
|
292
293
|
],
|
|
293
294
|
"size": [
|
|
294
|
-
|
|
295
|
-
|
|
295
|
+
260,
|
|
296
|
+
82
|
|
296
297
|
],
|
|
297
298
|
"flags": {},
|
|
298
|
-
"order":
|
|
299
|
+
"order": 7,
|
|
299
300
|
"mode": 0,
|
|
300
|
-
"inputs": [
|
|
301
|
+
"inputs": [
|
|
302
|
+
{
|
|
303
|
+
"name": "image",
|
|
304
|
+
"type": "IMAGE",
|
|
305
|
+
"link": 18
|
|
306
|
+
}
|
|
307
|
+
],
|
|
301
308
|
"outputs": [
|
|
302
309
|
{
|
|
303
310
|
"name": "IMAGE",
|
|
304
311
|
"type": "IMAGE",
|
|
305
|
-
"slot_index": 0,
|
|
306
312
|
"links": [
|
|
307
|
-
|
|
313
|
+
19
|
|
308
314
|
]
|
|
309
|
-
},
|
|
310
|
-
{
|
|
311
|
-
"name": "MASK",
|
|
312
|
-
"type": "MASK",
|
|
313
|
-
"links": null
|
|
314
315
|
}
|
|
315
316
|
],
|
|
316
317
|
"properties": {
|
|
317
|
-
"Node name for S&R": "LoadImage",
|
|
318
318
|
"cnr_id": "comfy-core",
|
|
319
|
-
"ver": "0.3.
|
|
319
|
+
"ver": "0.3.65",
|
|
320
|
+
"Node name for S&R": "ImageScaleToTotalPixels"
|
|
320
321
|
},
|
|
321
322
|
"widgets_values": [
|
|
322
|
-
"
|
|
323
|
-
|
|
324
|
-
]
|
|
325
|
-
"color": "#322",
|
|
326
|
-
"bgcolor": "#533"
|
|
323
|
+
"nearest-exact",
|
|
324
|
+
0.5
|
|
325
|
+
]
|
|
327
326
|
},
|
|
328
327
|
{
|
|
329
328
|
"id": 12,
|
|
330
329
|
"type": "VAEEncode",
|
|
331
330
|
"pos": [
|
|
332
|
-
|
|
333
|
-
|
|
331
|
+
500,
|
|
332
|
+
660
|
|
334
333
|
],
|
|
335
334
|
"size": [
|
|
336
335
|
210,
|
|
@@ -362,9 +361,9 @@
|
|
|
362
361
|
}
|
|
363
362
|
],
|
|
364
363
|
"properties": {
|
|
365
|
-
"Node name for S&R": "VAEEncode",
|
|
366
364
|
"cnr_id": "comfy-core",
|
|
367
|
-
"ver": "0.3.44"
|
|
365
|
+
"ver": "0.3.44",
|
|
366
|
+
"Node name for S&R": "VAEEncode"
|
|
368
367
|
},
|
|
369
368
|
"widgets_values": []
|
|
370
369
|
},
|
|
@@ -373,96 +372,95 @@
|
|
|
373
372
|
"type": "MarkdownNote",
|
|
374
373
|
"pos": [
|
|
375
374
|
740,
|
|
376
|
-
|
|
375
|
+
710
|
|
377
376
|
],
|
|
378
377
|
"size": [
|
|
379
378
|
310,
|
|
380
379
|
110
|
|
381
380
|
],
|
|
382
381
|
"flags": {},
|
|
383
|
-
"order":
|
|
382
|
+
"order": 3,
|
|
384
383
|
"mode": 0,
|
|
385
384
|
"inputs": [],
|
|
386
385
|
"outputs": [],
|
|
387
386
|
"properties": {},
|
|
388
387
|
"widgets_values": [
|
|
389
|
-
"When using the image
|
|
388
|
+
"When using the image-to-\nimage workflow, you should remember that the **denoise** value should be less than 1. The closer the value is to 0, the more features of the input image the output image will retain."
|
|
390
389
|
],
|
|
391
390
|
"color": "#432",
|
|
392
391
|
"bgcolor": "#653"
|
|
393
392
|
},
|
|
394
393
|
{
|
|
395
|
-
"id":
|
|
396
|
-
"type": "
|
|
394
|
+
"id": 3,
|
|
395
|
+
"type": "KSampler",
|
|
397
396
|
"pos": [
|
|
398
|
-
|
|
399
|
-
|
|
397
|
+
740,
|
|
398
|
+
180
|
|
400
399
|
],
|
|
401
400
|
"size": [
|
|
402
|
-
|
|
403
|
-
|
|
401
|
+
315,
|
|
402
|
+
474
|
|
404
403
|
],
|
|
405
404
|
"flags": {},
|
|
406
|
-
"order":
|
|
405
|
+
"order": 9,
|
|
407
406
|
"mode": 0,
|
|
408
407
|
"inputs": [
|
|
409
408
|
{
|
|
410
|
-
"name": "
|
|
411
|
-
"type": "
|
|
412
|
-
"link":
|
|
409
|
+
"name": "model",
|
|
410
|
+
"type": "MODEL",
|
|
411
|
+
"link": 13
|
|
412
|
+
},
|
|
413
|
+
{
|
|
414
|
+
"name": "positive",
|
|
415
|
+
"type": "CONDITIONING",
|
|
416
|
+
"link": 4
|
|
417
|
+
},
|
|
418
|
+
{
|
|
419
|
+
"name": "negative",
|
|
420
|
+
"type": "CONDITIONING",
|
|
421
|
+
"link": 6
|
|
422
|
+
},
|
|
423
|
+
{
|
|
424
|
+
"name": "latent_image",
|
|
425
|
+
"type": "LATENT",
|
|
426
|
+
"link": 11
|
|
413
427
|
}
|
|
414
428
|
],
|
|
415
429
|
"outputs": [
|
|
416
430
|
{
|
|
417
|
-
"name": "
|
|
418
|
-
"type": "
|
|
431
|
+
"name": "LATENT",
|
|
432
|
+
"type": "LATENT",
|
|
433
|
+
"slot_index": 0,
|
|
419
434
|
"links": [
|
|
420
|
-
|
|
435
|
+
7
|
|
421
436
|
]
|
|
422
437
|
}
|
|
423
438
|
],
|
|
424
439
|
"properties": {
|
|
425
|
-
"
|
|
440
|
+
"cnr_id": "comfy-core",
|
|
441
|
+
"ver": "0.3.44",
|
|
442
|
+
"Node name for S&R": "KSampler"
|
|
426
443
|
},
|
|
427
444
|
"widgets_values": [
|
|
428
|
-
|
|
429
|
-
|
|
445
|
+
266967097991208,
|
|
446
|
+
"randomize",
|
|
447
|
+
15,
|
|
448
|
+
8,
|
|
449
|
+
"dpmpp_2m",
|
|
450
|
+
"normal",
|
|
451
|
+
0.87
|
|
430
452
|
]
|
|
431
453
|
},
|
|
432
|
-
{
|
|
433
|
-
"id": 19,
|
|
434
|
-
"type": "MarkdownNote",
|
|
435
|
-
"pos": [
|
|
436
|
-
280,
|
|
437
|
-
780
|
|
438
|
-
],
|
|
439
|
-
"size": [
|
|
440
|
-
270,
|
|
441
|
-
120
|
|
442
|
-
],
|
|
443
|
-
"flags": {},
|
|
444
|
-
"order": 3,
|
|
445
|
-
"mode": 0,
|
|
446
|
-
"inputs": [],
|
|
447
|
-
"outputs": [],
|
|
448
|
-
"title": "About Scale Image to Total Pixels",
|
|
449
|
-
"properties": {},
|
|
450
|
-
"widgets_values": [
|
|
451
|
-
"The model this template uses is trained based on a 512*512 image dataset. So if you use an input image that's too large, it might cause some issues. We've added the Scale image to total pixels node to scale images. If you're quite familiar with this model, you can remove it."
|
|
452
|
-
],
|
|
453
|
-
"color": "#432",
|
|
454
|
-
"bgcolor": "#653"
|
|
455
|
-
},
|
|
456
454
|
{
|
|
457
455
|
"id": 16,
|
|
458
456
|
"type": "MarkdownNote",
|
|
459
457
|
"pos": [
|
|
460
|
-
-
|
|
461
|
-
|
|
458
|
+
-560,
|
|
459
|
+
170
|
|
462
460
|
],
|
|
463
461
|
"size": [
|
|
464
|
-
|
|
465
|
-
|
|
462
|
+
440,
|
|
463
|
+
240
|
|
466
464
|
],
|
|
467
465
|
"flags": {},
|
|
468
466
|
"order": 4,
|
|
@@ -471,7 +469,7 @@
|
|
|
471
469
|
"outputs": [],
|
|
472
470
|
"properties": {},
|
|
473
471
|
"widgets_values": [
|
|
474
|
-
"[
|
|
472
|
+
"[Tutorial](https://docs.comfy.org/tutorials/basic/image-to-image)\n\n\n## Model links\n\n**checkpoints**\n\n- [v1-5-pruned-emaonly-fp16.safetensors](https://huggingface.co/Comfy-Org/stable-diffusion-v1-5-archive/resolve/main/v1-5-pruned-emaonly-fp16.safetensors?download=true)\n\n\nModel Storage Location\n\n```\nš ComfyUI/\nāāā š models/\nā āāā š checkpoints/\nā āāā v1-5-pruned-emaonly-fp16.safetensors\n```\n"
|
|
475
473
|
],
|
|
476
474
|
"color": "#432",
|
|
477
475
|
"bgcolor": "#653"
|
|
@@ -619,13 +617,17 @@
|
|
|
619
617
|
"config": {},
|
|
620
618
|
"extra": {
|
|
621
619
|
"ds": {
|
|
622
|
-
"scale": 0.
|
|
620
|
+
"scale": 0.4291866939029705,
|
|
623
621
|
"offset": [
|
|
624
|
-
|
|
625
|
-
|
|
622
|
+
2140.5317402544647,
|
|
623
|
+
404.60195874579364
|
|
626
624
|
]
|
|
627
625
|
},
|
|
628
|
-
"frontendVersion": "1.
|
|
626
|
+
"frontendVersion": "1.27.10",
|
|
627
|
+
"VHS_latentpreview": false,
|
|
628
|
+
"VHS_latentpreviewrate": 0,
|
|
629
|
+
"VHS_MetadataImage": true,
|
|
630
|
+
"VHS_KeepIntermediate": true
|
|
629
631
|
},
|
|
630
632
|
"version": 0.4
|
|
631
633
|
}
|