comfyui-workflow-templates-media-other 0.3.10__py3-none-any.whl → 0.3.14__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- comfyui_workflow_templates_media_other/templates/image_z_image_turbo-1.webp +0 -0
- comfyui_workflow_templates_media_other/templates/image_z_image_turbo.json +655 -0
- comfyui_workflow_templates_media_other/templates/index.ar.json +672 -672
- comfyui_workflow_templates_media_other/templates/index.es.json +673 -673
- comfyui_workflow_templates_media_other/templates/index.fr.json +673 -673
- comfyui_workflow_templates_media_other/templates/index.ja.json +673 -673
- comfyui_workflow_templates_media_other/templates/index.json +672 -672
- comfyui_workflow_templates_media_other/templates/index.ko.json +673 -673
- comfyui_workflow_templates_media_other/templates/index.ru.json +673 -673
- comfyui_workflow_templates_media_other/templates/index.tr.json +672 -672
- comfyui_workflow_templates_media_other/templates/index.zh-TW.json +673 -673
- comfyui_workflow_templates_media_other/templates/index.zh.json +673 -673
- {comfyui_workflow_templates_media_other-0.3.10.dist-info → comfyui_workflow_templates_media_other-0.3.14.dist-info}/METADATA +1 -1
- {comfyui_workflow_templates_media_other-0.3.10.dist-info → comfyui_workflow_templates_media_other-0.3.14.dist-info}/RECORD +16 -14
- {comfyui_workflow_templates_media_other-0.3.10.dist-info → comfyui_workflow_templates_media_other-0.3.14.dist-info}/WHEEL +0 -0
- {comfyui_workflow_templates_media_other-0.3.10.dist-info → comfyui_workflow_templates_media_other-0.3.14.dist-info}/top_level.txt +0 -0
|
@@ -1,869 +1,542 @@
|
|
|
1
1
|
[
|
|
2
2
|
{
|
|
3
3
|
"moduleName": "default",
|
|
4
|
-
"
|
|
5
|
-
"
|
|
4
|
+
"category": "GENERATION TYPE",
|
|
5
|
+
"icon": "icon-[lucide--image]",
|
|
6
|
+
"title": "Image",
|
|
6
7
|
"type": "image",
|
|
7
8
|
"templates": [
|
|
8
9
|
{
|
|
9
|
-
"name": "
|
|
10
|
-
"title": "
|
|
10
|
+
"name": "image_flux2",
|
|
11
|
+
"title": "Flux.2 Dev",
|
|
11
12
|
"mediaType": "image",
|
|
12
13
|
"mediaSubtype": "webp",
|
|
13
|
-
"
|
|
14
|
-
"
|
|
15
|
-
"tags": ["Text to Image", "Image"],
|
|
16
|
-
"models": ["
|
|
17
|
-
"date": "2025-
|
|
18
|
-
"size":
|
|
14
|
+
"thumbnailVariant": "compareSlider",
|
|
15
|
+
"description": "Generate up to 4MP photorealistic images with multi-reference consistency and professional text rendering.",
|
|
16
|
+
"tags": ["Text to Image", "Image", "Image Edit"],
|
|
17
|
+
"models": ["Flux.2 Dev", "BFL"],
|
|
18
|
+
"date": "2025-11-26",
|
|
19
|
+
"size": 71382356459,
|
|
20
|
+
"vram": 0
|
|
19
21
|
},
|
|
20
22
|
{
|
|
21
|
-
"name": "
|
|
22
|
-
"title": "
|
|
23
|
+
"name": "image_flux2_fp8",
|
|
24
|
+
"title": "Product Mockup(Flux.2 Dev FP8)",
|
|
23
25
|
"mediaType": "image",
|
|
24
26
|
"mediaSubtype": "webp",
|
|
25
|
-
"description": "
|
|
26
|
-
"
|
|
27
|
-
|
|
28
|
-
|
|
29
|
-
|
|
30
|
-
|
|
27
|
+
"description": "Create product mockups by applying design patterns to packaging, mugs, and other products using multi-reference consistency.",
|
|
28
|
+
"tags": [
|
|
29
|
+
"Text to Image",
|
|
30
|
+
"Image",
|
|
31
|
+
"Image Edit",
|
|
32
|
+
"Mockup",
|
|
33
|
+
"Product Design"
|
|
34
|
+
],
|
|
35
|
+
"models": ["Flux.2 Dev", "BFL"],
|
|
36
|
+
"date": "2025-11-26",
|
|
37
|
+
"size": 53837415055,
|
|
38
|
+
"vram": 0
|
|
31
39
|
},
|
|
32
40
|
{
|
|
33
|
-
"name": "
|
|
34
|
-
"title": "Image to
|
|
35
|
-
"description": "Generate videos from an input image using Wan2.2 14B",
|
|
41
|
+
"name": "image_z_image_turbo",
|
|
42
|
+
"title": "Z-Image-Turbo Text to Image",
|
|
36
43
|
"mediaType": "image",
|
|
37
44
|
"mediaSubtype": "webp",
|
|
38
|
-
"
|
|
39
|
-
"tags": ["
|
|
40
|
-
"models": ["
|
|
41
|
-
"date": "2025-
|
|
42
|
-
"size":
|
|
45
|
+
"description": "An Efficient Image Generation Foundation Model with Single-Stream Diffusion Transformer, supports English & Chinese.",
|
|
46
|
+
"tags": ["Text to Image", "Image"],
|
|
47
|
+
"models": ["Z-Image-Turbo"],
|
|
48
|
+
"date": "2025-11-27",
|
|
49
|
+
"size": 35326050304
|
|
43
50
|
},
|
|
44
51
|
{
|
|
45
|
-
"name": "
|
|
46
|
-
"title": "Image to
|
|
52
|
+
"name": "image_qwen_image",
|
|
53
|
+
"title": "Qwen-Image Text to Image",
|
|
47
54
|
"mediaType": "image",
|
|
48
55
|
"mediaSubtype": "webp",
|
|
49
|
-
"description": "Generate
|
|
50
|
-
"
|
|
51
|
-
"
|
|
52
|
-
"
|
|
53
|
-
"
|
|
54
|
-
"size":
|
|
56
|
+
"description": "Generate images with exceptional multilingual text rendering and editing capabilities using Qwen-Image's 20B MMDiT model..",
|
|
57
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image",
|
|
58
|
+
"tags": ["Text to Image", "Image"],
|
|
59
|
+
"models": ["Qwen-Image"],
|
|
60
|
+
"date": "2025-08-05",
|
|
61
|
+
"size": 31772020572
|
|
55
62
|
},
|
|
56
63
|
{
|
|
57
|
-
"name": "
|
|
58
|
-
"title": "
|
|
64
|
+
"name": "image_qwen_image_instantx_controlnet",
|
|
65
|
+
"title": "Qwen-Image InstantX Union ControlNet",
|
|
59
66
|
"mediaType": "image",
|
|
60
67
|
"mediaSubtype": "webp",
|
|
61
|
-
"description": "Generate
|
|
62
|
-
"tags": ["
|
|
63
|
-
"
|
|
64
|
-
"
|
|
65
|
-
"
|
|
66
|
-
"size":
|
|
68
|
+
"description": "Generate images with Qwen-Image InstantX ControlNet, supporting canny, soft edge, depth, pose",
|
|
69
|
+
"tags": ["Image to Image", "Image", "ControlNet"],
|
|
70
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image",
|
|
71
|
+
"models": ["Qwen-Image"],
|
|
72
|
+
"date": "2025-08-23",
|
|
73
|
+
"size": 35304631173
|
|
67
74
|
},
|
|
68
75
|
{
|
|
69
|
-
"name": "
|
|
70
|
-
"title": "Image
|
|
76
|
+
"name": "image_qwen_image_instantx_inpainting_controlnet",
|
|
77
|
+
"title": "Qwen-Image InstantX Inpainting ControlNet",
|
|
71
78
|
"mediaType": "image",
|
|
72
79
|
"mediaSubtype": "webp",
|
|
73
|
-
"
|
|
74
|
-
"
|
|
75
|
-
"tags": ["
|
|
76
|
-
"
|
|
77
|
-
"
|
|
78
|
-
"
|
|
79
|
-
"
|
|
80
|
+
"thumbnailVariant": "compareSlider",
|
|
81
|
+
"description": "Professional inpainting and image editing with Qwen-Image InstantX ControlNet. Supports object replacement, text modification, background changes, and outpainting.",
|
|
82
|
+
"tags": ["Image to Image", "Image", "ControlNet", "Inpainting"],
|
|
83
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image",
|
|
84
|
+
"models": ["Qwen-Image"],
|
|
85
|
+
"date": "2025-09-12",
|
|
86
|
+
"size": 36013300777
|
|
80
87
|
},
|
|
81
88
|
{
|
|
82
|
-
"name": "
|
|
83
|
-
"title": "Image
|
|
89
|
+
"name": "image_qwen_image_union_control_lora",
|
|
90
|
+
"title": "Qwen-Image Union Control",
|
|
84
91
|
"mediaType": "image",
|
|
85
92
|
"mediaSubtype": "webp",
|
|
86
|
-
"
|
|
87
|
-
"
|
|
88
|
-
"tutorialUrl": "https://docs.comfy.org/tutorials/
|
|
89
|
-
"
|
|
90
|
-
"
|
|
91
|
-
"
|
|
92
|
-
"size": 2136746230,
|
|
93
|
-
"vram": 3092376453
|
|
93
|
+
"description": "Generate images with precise structural control using Qwen-Image's unified ControlNet LoRA. Supports multiple control types including canny, depth, lineart, softedge, normal, and openpose for diverse creative applications.",
|
|
94
|
+
"tags": ["Text to Image", "Image", "ControlNet"],
|
|
95
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image",
|
|
96
|
+
"models": ["Qwen-Image"],
|
|
97
|
+
"date": "2025-08-23",
|
|
98
|
+
"size": 32716913377
|
|
94
99
|
},
|
|
95
100
|
{
|
|
96
|
-
"name": "
|
|
97
|
-
"title": "
|
|
101
|
+
"name": "image_qwen_image_controlnet_patch",
|
|
102
|
+
"title": "Qwen-Image ControlNet model patch",
|
|
98
103
|
"mediaType": "image",
|
|
99
104
|
"mediaSubtype": "webp",
|
|
100
|
-
"
|
|
101
|
-
"
|
|
102
|
-
"
|
|
103
|
-
"
|
|
104
|
-
"
|
|
105
|
-
"
|
|
106
|
-
"
|
|
105
|
+
"thumbnailVariant": "compareSlider",
|
|
106
|
+
"description": "Control image generation using Qwen-Image ControlNet models. Supports canny, depth, and inpainting controls through model patching.",
|
|
107
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image",
|
|
108
|
+
"tags": ["Text to Image", "Image", "ControlNet"],
|
|
109
|
+
"models": ["Qwen-Image"],
|
|
110
|
+
"date": "2025-08-24",
|
|
111
|
+
"size": 34037615821
|
|
107
112
|
},
|
|
108
113
|
{
|
|
109
|
-
"name": "
|
|
110
|
-
"title": "
|
|
114
|
+
"name": "image_qwen_image_edit_2509",
|
|
115
|
+
"title": "Qwen Image Edit 2509",
|
|
111
116
|
"mediaType": "image",
|
|
112
117
|
"mediaSubtype": "webp",
|
|
113
|
-
"
|
|
114
|
-
"
|
|
115
|
-
"
|
|
116
|
-
"
|
|
117
|
-
"
|
|
118
|
-
"
|
|
119
|
-
"
|
|
118
|
+
"thumbnailVariant": "compareSlider",
|
|
119
|
+
"description": "Advanced image editing with multi-image support, improved consistency, and ControlNet integration.",
|
|
120
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image-edit",
|
|
121
|
+
"tags": ["Image to Image", "Image Edit", "ControlNet"],
|
|
122
|
+
"models": ["Qwen-Image"],
|
|
123
|
+
"date": "2025-09-25",
|
|
124
|
+
"size": 31772020572
|
|
120
125
|
},
|
|
121
126
|
{
|
|
122
|
-
"name": "
|
|
123
|
-
"title": "
|
|
127
|
+
"name": "image_qwen_image_edit",
|
|
128
|
+
"title": "Qwen Image Edit",
|
|
124
129
|
"mediaType": "image",
|
|
125
130
|
"mediaSubtype": "webp",
|
|
126
|
-
"description": "Edit specific parts of images seamlessly.",
|
|
127
131
|
"thumbnailVariant": "compareSlider",
|
|
128
|
-
"
|
|
129
|
-
"
|
|
130
|
-
"
|
|
131
|
-
"
|
|
132
|
-
"
|
|
133
|
-
"
|
|
132
|
+
"description": "Edit images with precise bilingual text editing and dual semantic/appearance editing capabilities using Qwen-Image-Edit's 20B MMDiT model.",
|
|
133
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image-edit",
|
|
134
|
+
"tags": ["Image to Image", "Image Edit"],
|
|
135
|
+
"models": ["Qwen-Image"],
|
|
136
|
+
"date": "2025-08-18",
|
|
137
|
+
"size": 31772020572
|
|
134
138
|
},
|
|
135
139
|
{
|
|
136
|
-
"name": "
|
|
137
|
-
"title": "
|
|
140
|
+
"name": "image_chrono_edit_14B",
|
|
141
|
+
"title": "ChronoEdit 14B",
|
|
138
142
|
"mediaType": "image",
|
|
139
143
|
"mediaSubtype": "webp",
|
|
140
|
-
"description": "Extend images beyond their original boundaries.",
|
|
141
144
|
"thumbnailVariant": "compareSlider",
|
|
142
|
-
"
|
|
143
|
-
"tags": ["
|
|
144
|
-
"models": ["
|
|
145
|
-
"date": "2025-03
|
|
146
|
-
"size":
|
|
147
|
-
"vram": 4101693768
|
|
145
|
+
"description": "Image editing powered by video models' dynamic understanding, creating physically plausible results while preserving character and style consistency.",
|
|
146
|
+
"tags": ["Image Edit", "Image to Image"],
|
|
147
|
+
"models": ["Wan2.1", "ChronoEdit", "Nvidia"],
|
|
148
|
+
"date": "2025-11-03",
|
|
149
|
+
"size": 40459304
|
|
148
150
|
},
|
|
149
151
|
{
|
|
150
|
-
"name": "
|
|
151
|
-
"title": "
|
|
152
|
+
"name": "flux_kontext_dev_basic",
|
|
153
|
+
"title": "Flux Kontext Dev Image Edit",
|
|
152
154
|
"mediaType": "image",
|
|
153
155
|
"mediaSubtype": "webp",
|
|
154
|
-
"
|
|
155
|
-
"
|
|
156
|
-
"
|
|
157
|
-
"
|
|
158
|
-
"
|
|
159
|
-
"
|
|
160
|
-
"
|
|
156
|
+
"thumbnailVariant": "hoverDissolve",
|
|
157
|
+
"description": "Smart image editing that keeps characters consistent, edits specific parts without affecting others, and preserves original styles.",
|
|
158
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-kontext-dev",
|
|
159
|
+
"tags": ["Image Edit", "Image to Image"],
|
|
160
|
+
"models": ["Flux", "BFL"],
|
|
161
|
+
"date": "2025-06-26",
|
|
162
|
+
"size": 17641578168,
|
|
163
|
+
"vram": 19327352832
|
|
161
164
|
},
|
|
162
165
|
{
|
|
163
|
-
"name": "
|
|
164
|
-
"title": "
|
|
166
|
+
"name": "image_chroma1_radiance_text_to_image",
|
|
167
|
+
"title": "Chroma1 Radiance text to image",
|
|
165
168
|
"mediaType": "image",
|
|
166
169
|
"mediaSubtype": "webp",
|
|
167
|
-
"description": "
|
|
168
|
-
"
|
|
169
|
-
"
|
|
170
|
-
"
|
|
171
|
-
"
|
|
172
|
-
"
|
|
173
|
-
"vram": 4080218931
|
|
170
|
+
"description": "Chroma1-Radiance works directly with image pixels instead of compressed latents, delivering higher quality images with reduced artifacts and distortion.",
|
|
171
|
+
"tags": ["Text to Image", "Image"],
|
|
172
|
+
"models": ["Chroma"],
|
|
173
|
+
"date": "2025-09-18",
|
|
174
|
+
"size": 23622320128,
|
|
175
|
+
"vram": 23622320128
|
|
174
176
|
},
|
|
175
177
|
{
|
|
176
|
-
"name": "
|
|
177
|
-
"title": "
|
|
178
|
+
"name": "image_netayume_lumina_t2i",
|
|
179
|
+
"title": "NetaYume Lumina Text to Image",
|
|
178
180
|
"mediaType": "image",
|
|
179
181
|
"mediaSubtype": "webp",
|
|
180
|
-
"description": "
|
|
181
|
-
"tags": ["Text to Image", "Image"],
|
|
182
|
-
"models": ["
|
|
183
|
-
"date": "2025-
|
|
184
|
-
"
|
|
185
|
-
"size": 2469606195,
|
|
186
|
-
"vram": 6184752906
|
|
182
|
+
"description": "High-quality anime-style image generation with enhanced character understanding and detailed textures. Fine-tuned from Neta Lumina on Danbooru dataset.",
|
|
183
|
+
"tags": ["Text to Image", "Image", "Anime"],
|
|
184
|
+
"models": ["OmniGen"],
|
|
185
|
+
"date": "2025-10-10",
|
|
186
|
+
"size": 10619306639
|
|
187
187
|
},
|
|
188
188
|
{
|
|
189
|
-
"name": "
|
|
190
|
-
"title": "
|
|
189
|
+
"name": "image_chroma_text_to_image",
|
|
190
|
+
"title": "Chroma text to image",
|
|
191
191
|
"mediaType": "image",
|
|
192
192
|
"mediaSubtype": "webp",
|
|
193
|
-
"description": "
|
|
193
|
+
"description": "Chroma - enhanced Flux model with improved image quality and better prompt understanding for stunning text-to-image generation.",
|
|
194
194
|
"tags": ["Text to Image", "Image"],
|
|
195
|
-
"models": ["
|
|
196
|
-
"date": "2025-
|
|
197
|
-
"
|
|
198
|
-
"
|
|
199
|
-
"vram": 5927054868
|
|
195
|
+
"models": ["Chroma", "Flux"],
|
|
196
|
+
"date": "2025-06-04",
|
|
197
|
+
"size": 23289460163,
|
|
198
|
+
"vram": 15569256448
|
|
200
199
|
},
|
|
201
200
|
{
|
|
202
|
-
"name": "
|
|
203
|
-
"title": "
|
|
201
|
+
"name": "image_flux.1_fill_dev_OneReward",
|
|
202
|
+
"title": "Flux.1 Dev OneReward",
|
|
204
203
|
"mediaType": "image",
|
|
205
204
|
"mediaSubtype": "webp",
|
|
206
|
-
"description": "Upscale images by enhancing quality in latent space.",
|
|
207
205
|
"thumbnailVariant": "compareSlider",
|
|
208
|
-
"
|
|
209
|
-
"
|
|
210
|
-
"
|
|
211
|
-
"
|
|
212
|
-
"size":
|
|
213
|
-
"vram":
|
|
206
|
+
"description": "Supports various tasks such as image inpainting, outpainting, and object removal by bytedance-research team",
|
|
207
|
+
"tags": ["Inpainting", "Outpainting"],
|
|
208
|
+
"models": ["Flux", "BFL"],
|
|
209
|
+
"date": "2025-09-21",
|
|
210
|
+
"size": 29001766666,
|
|
211
|
+
"vram": 21474836480
|
|
214
212
|
},
|
|
215
213
|
{
|
|
216
|
-
"name": "
|
|
217
|
-
"title": "
|
|
214
|
+
"name": "flux_dev_checkpoint_example",
|
|
215
|
+
"title": "Flux Dev fp8",
|
|
218
216
|
"mediaType": "image",
|
|
219
217
|
"mediaSubtype": "webp",
|
|
220
|
-
"description": "
|
|
221
|
-
"
|
|
222
|
-
"tags": ["
|
|
223
|
-
"models": ["
|
|
218
|
+
"description": "Generate images using Flux Dev fp8 quantized version. Suitable for devices with limited VRAM, requires only one model file, but image quality is slightly lower than the full version.",
|
|
219
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-text-to-image",
|
|
220
|
+
"tags": ["Text to Image", "Image"],
|
|
221
|
+
"models": ["Flux", "BFL"],
|
|
224
222
|
"date": "2025-03-01",
|
|
225
|
-
"
|
|
226
|
-
"
|
|
227
|
-
"vram": 6442450944
|
|
223
|
+
"size": 17244293693,
|
|
224
|
+
"vram": 18253611008
|
|
228
225
|
},
|
|
229
226
|
{
|
|
230
|
-
"name": "
|
|
231
|
-
"title": "
|
|
227
|
+
"name": "flux1_dev_uso_reference_image_gen",
|
|
228
|
+
"title": "Flux.1 Dev USO Reference Image Generation",
|
|
229
|
+
"description": "Use reference images to control both style and subject - keep your character's face while changing artistic style, or apply artistic styles to new scenes",
|
|
230
|
+
"thumbnailVariant": "hoverDissolve",
|
|
232
231
|
"mediaType": "image",
|
|
233
232
|
"mediaSubtype": "webp",
|
|
234
|
-
"
|
|
235
|
-
"
|
|
236
|
-
"
|
|
237
|
-
"
|
|
238
|
-
"
|
|
239
|
-
"
|
|
240
|
-
"size": 2201170739,
|
|
241
|
-
"vram": 6442450944
|
|
233
|
+
"tags": ["Image to Image", "Image"],
|
|
234
|
+
"models": ["Flux", "BFL"],
|
|
235
|
+
"date": "2025-09-02",
|
|
236
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-uso",
|
|
237
|
+
"size": 18597208392,
|
|
238
|
+
"vram": 19864223744
|
|
242
239
|
},
|
|
243
240
|
{
|
|
244
|
-
"name": "
|
|
245
|
-
"title": "
|
|
241
|
+
"name": "flux_schnell",
|
|
242
|
+
"title": "Flux Schnell fp8",
|
|
246
243
|
"mediaType": "image",
|
|
247
244
|
"mediaSubtype": "webp",
|
|
248
|
-
"description": "
|
|
249
|
-
"
|
|
250
|
-
"tags": ["
|
|
251
|
-
"models": ["
|
|
245
|
+
"description": "Quickly generate images with Flux Schnell fp8 quantized version. Ideal for low-end hardware, requires only 4 steps to generate images.",
|
|
246
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-text-to-image",
|
|
247
|
+
"tags": ["Text to Image", "Image"],
|
|
248
|
+
"models": ["Flux", "BFL"],
|
|
252
249
|
"date": "2025-03-01",
|
|
253
|
-
"
|
|
254
|
-
"
|
|
255
|
-
"vram": 5153960755
|
|
250
|
+
"size": 17233556275,
|
|
251
|
+
"vram": 18253611008
|
|
256
252
|
},
|
|
257
253
|
{
|
|
258
|
-
"name": "
|
|
259
|
-
"title": "
|
|
254
|
+
"name": "flux1_krea_dev",
|
|
255
|
+
"title": "Flux.1 Krea Dev",
|
|
260
256
|
"mediaType": "image",
|
|
261
257
|
"mediaSubtype": "webp",
|
|
262
|
-
"description": "
|
|
263
|
-
"
|
|
264
|
-
"tags": ["
|
|
265
|
-
"models": ["
|
|
266
|
-
"date": "2025-
|
|
267
|
-
"
|
|
268
|
-
"
|
|
269
|
-
"vram": 6442450944
|
|
258
|
+
"description": "A fine-tuned FLUX model pushing photorealism to the max",
|
|
259
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux1-krea-dev",
|
|
260
|
+
"tags": ["Text to Image", "Image"],
|
|
261
|
+
"models": ["Flux", "BFL"],
|
|
262
|
+
"date": "2025-07-31",
|
|
263
|
+
"size": 22269405430,
|
|
264
|
+
"vram": 23085449216
|
|
270
265
|
},
|
|
271
266
|
{
|
|
272
|
-
"name": "
|
|
273
|
-
"title": "
|
|
267
|
+
"name": "flux_dev_full_text_to_image",
|
|
268
|
+
"title": "Flux Dev full text to image",
|
|
274
269
|
"mediaType": "image",
|
|
275
270
|
"mediaSubtype": "webp",
|
|
276
|
-
"description": "Generate images
|
|
277
|
-
"
|
|
278
|
-
"tags": ["
|
|
279
|
-
"models": ["
|
|
271
|
+
"description": "Generate high-quality images with Flux Dev full version. Requires larger VRAM and multiple model files, but provides the best prompt following capability and image quality.",
|
|
272
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-text-to-image",
|
|
273
|
+
"tags": ["Text to Image", "Image"],
|
|
274
|
+
"models": ["Flux", "BFL"],
|
|
280
275
|
"date": "2025-03-01",
|
|
281
|
-
"
|
|
282
|
-
"
|
|
283
|
-
"vram": 6442450944
|
|
276
|
+
"size": 34177202258,
|
|
277
|
+
"vram": 23622320128
|
|
284
278
|
},
|
|
285
279
|
{
|
|
286
|
-
"name": "
|
|
287
|
-
"title": "
|
|
280
|
+
"name": "flux_schnell_full_text_to_image",
|
|
281
|
+
"title": "Flux Schnell full text to image",
|
|
288
282
|
"mediaType": "image",
|
|
289
283
|
"mediaSubtype": "webp",
|
|
290
|
-
"description": "Generate images
|
|
291
|
-
"
|
|
292
|
-
"tags": ["
|
|
293
|
-
"models": ["
|
|
284
|
+
"description": "Generate images quickly with Flux Schnell full version. Uses Apache2.0 license, requires only 4 steps to generate images while maintaining good image quality.",
|
|
285
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-text-to-image",
|
|
286
|
+
"tags": ["Text to Image", "Image"],
|
|
287
|
+
"models": ["Flux", "BFL"],
|
|
294
288
|
"date": "2025-03-01",
|
|
295
|
-
"
|
|
296
|
-
"size": 2888365507,
|
|
297
|
-
"vram": 6442450944
|
|
289
|
+
"size": 34155727421
|
|
298
290
|
},
|
|
299
291
|
{
|
|
300
|
-
"name": "
|
|
301
|
-
"title": "
|
|
292
|
+
"name": "flux_fill_inpaint_example",
|
|
293
|
+
"title": "Flux Inpaint",
|
|
302
294
|
"mediaType": "image",
|
|
303
295
|
"mediaSubtype": "webp",
|
|
304
|
-
"description": "
|
|
305
|
-
"thumbnailVariant": "
|
|
306
|
-
"
|
|
307
|
-
"
|
|
296
|
+
"description": "Fill missing parts of images using Flux inpainting.",
|
|
297
|
+
"thumbnailVariant": "compareSlider",
|
|
298
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-fill-dev",
|
|
299
|
+
"tags": ["Image to Image", "Inpainting", "Image"],
|
|
300
|
+
"models": ["Flux", "BFL"],
|
|
308
301
|
"date": "2025-03-01",
|
|
309
|
-
"
|
|
310
|
-
"size": 2523293286,
|
|
311
|
-
"vram": 6442450944
|
|
302
|
+
"size": 10372346020
|
|
312
303
|
},
|
|
313
304
|
{
|
|
314
|
-
"name": "
|
|
315
|
-
"title": "
|
|
305
|
+
"name": "flux_fill_outpaint_example",
|
|
306
|
+
"title": "Flux Outpaint",
|
|
316
307
|
"mediaType": "image",
|
|
317
308
|
"mediaSubtype": "webp",
|
|
318
|
-
"description": "
|
|
319
|
-
"thumbnailVariant": "
|
|
320
|
-
"
|
|
321
|
-
"
|
|
309
|
+
"description": "Extend images beyond boundaries using Flux outpainting.",
|
|
310
|
+
"thumbnailVariant": "compareSlider",
|
|
311
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-fill-dev",
|
|
312
|
+
"tags": ["Outpainting", "Image", "Image to Image"],
|
|
313
|
+
"models": ["Flux", "BFL"],
|
|
322
314
|
"date": "2025-03-01",
|
|
323
|
-
"
|
|
324
|
-
|
|
325
|
-
"vram": 6442450944
|
|
326
|
-
}
|
|
327
|
-
]
|
|
328
|
-
},
|
|
329
|
-
{
|
|
330
|
-
"moduleName": "default",
|
|
331
|
-
"category": "GENERATION TYPE",
|
|
332
|
-
"icon": "icon-[lucide--image]",
|
|
333
|
-
"title": "Image",
|
|
334
|
-
"type": "image",
|
|
335
|
-
"templates": [
|
|
315
|
+
"size": 10372346020
|
|
316
|
+
},
|
|
336
317
|
{
|
|
337
|
-
"name": "
|
|
338
|
-
"title": "Flux
|
|
318
|
+
"name": "flux_canny_model_example",
|
|
319
|
+
"title": "Flux Canny Model",
|
|
339
320
|
"mediaType": "image",
|
|
340
321
|
"mediaSubtype": "webp",
|
|
341
|
-
"
|
|
342
|
-
"
|
|
343
|
-
"
|
|
344
|
-
"
|
|
345
|
-
"
|
|
346
|
-
"
|
|
347
|
-
"
|
|
322
|
+
"description": "Generate images guided by edge detection using Flux Canny.",
|
|
323
|
+
"thumbnailVariant": "hoverDissolve",
|
|
324
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-controlnet",
|
|
325
|
+
"tags": ["Image to Image", "ControlNet", "Image"],
|
|
326
|
+
"models": ["Flux", "BFL"],
|
|
327
|
+
"date": "2025-03-01",
|
|
328
|
+
"size": 34177202258
|
|
348
329
|
},
|
|
349
330
|
{
|
|
350
|
-
"name": "
|
|
351
|
-
"title": "
|
|
331
|
+
"name": "flux_depth_lora_example",
|
|
332
|
+
"title": "Flux Depth Lora",
|
|
352
333
|
"mediaType": "image",
|
|
353
334
|
"mediaSubtype": "webp",
|
|
354
|
-
"description": "
|
|
355
|
-
"
|
|
356
|
-
|
|
357
|
-
|
|
358
|
-
|
|
359
|
-
|
|
360
|
-
|
|
361
|
-
],
|
|
362
|
-
"models": ["Flux.2 Dev", "BFL"],
|
|
363
|
-
"date": "2025-11-26",
|
|
364
|
-
"size": 53837415055,
|
|
365
|
-
"vram": 0
|
|
335
|
+
"description": "Generate images guided by depth information using Flux LoRA.",
|
|
336
|
+
"thumbnailVariant": "hoverDissolve",
|
|
337
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-controlnet",
|
|
338
|
+
"tags": ["Image to Image", "ControlNet", "Image"],
|
|
339
|
+
"models": ["Flux", "BFL"],
|
|
340
|
+
"date": "2025-03-01",
|
|
341
|
+
"size": 35412005356
|
|
366
342
|
},
|
|
367
343
|
{
|
|
368
|
-
"name": "
|
|
369
|
-
"title": "
|
|
344
|
+
"name": "flux_redux_model_example",
|
|
345
|
+
"title": "Flux Redux Model",
|
|
370
346
|
"mediaType": "image",
|
|
371
347
|
"mediaSubtype": "webp",
|
|
372
|
-
"description": "
|
|
373
|
-
"
|
|
374
|
-
"
|
|
375
|
-
"
|
|
376
|
-
"
|
|
348
|
+
"description": "Generate images by transferring style from reference images using Flux Redux.",
|
|
349
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-controlnet",
|
|
350
|
+
"tags": ["Image to Image", "ControlNet", "Image"],
|
|
351
|
+
"models": ["Flux", "BFL"],
|
|
352
|
+
"date": "2025-03-01",
|
|
353
|
+
"size": 35154307318
|
|
377
354
|
},
|
|
378
355
|
{
|
|
379
|
-
"name": "
|
|
380
|
-
"title": "
|
|
356
|
+
"name": "image_omnigen2_t2i",
|
|
357
|
+
"title": "OmniGen2 Text to Image",
|
|
381
358
|
"mediaType": "image",
|
|
382
359
|
"mediaSubtype": "webp",
|
|
383
|
-
"description": "Generate images
|
|
384
|
-
"tutorialUrl": "https://docs.comfy.org/tutorials/image/
|
|
360
|
+
"description": "Generate high-quality images from text prompts using OmniGen2's unified 7B multimodal model with dual-path architecture.",
|
|
361
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/image/omnigen/omnigen2",
|
|
385
362
|
"tags": ["Text to Image", "Image"],
|
|
386
|
-
"models": ["
|
|
387
|
-
"date": "2025-
|
|
388
|
-
"size":
|
|
363
|
+
"models": ["OmniGen"],
|
|
364
|
+
"date": "2025-06-30",
|
|
365
|
+
"size": 15784004813
|
|
389
366
|
},
|
|
390
367
|
{
|
|
391
|
-
"name": "
|
|
392
|
-
"title": "
|
|
368
|
+
"name": "image_omnigen2_image_edit",
|
|
369
|
+
"title": "OmniGen2 Image Edit",
|
|
393
370
|
"mediaType": "image",
|
|
394
371
|
"mediaSubtype": "webp",
|
|
395
|
-
"
|
|
396
|
-
"
|
|
397
|
-
"tutorialUrl": "https://docs.comfy.org/tutorials/image/
|
|
398
|
-
"
|
|
399
|
-
"
|
|
400
|
-
"
|
|
401
|
-
|
|
402
|
-
{
|
|
403
|
-
"name": "image_qwen_image_instantx_inpainting_controlnet",
|
|
404
|
-
"title": "Qwen-Image InstantX Inpainting ControlNet",
|
|
405
|
-
"mediaType": "image",
|
|
406
|
-
"mediaSubtype": "webp",
|
|
407
|
-
"thumbnailVariant": "compareSlider",
|
|
408
|
-
"description": "Professional inpainting and image editing with Qwen-Image InstantX ControlNet. Supports object replacement, text modification, background changes, and outpainting.",
|
|
409
|
-
"tags": ["Image to Image", "Image", "ControlNet", "Inpainting"],
|
|
410
|
-
"tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image",
|
|
411
|
-
"models": ["Qwen-Image"],
|
|
412
|
-
"date": "2025-09-12",
|
|
413
|
-
"size": 36013300777
|
|
372
|
+
"thumbnailVariant": "hoverDissolve",
|
|
373
|
+
"description": "Edit images with natural language instructions using OmniGen2's advanced image editing capabilities and text rendering support.",
|
|
374
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/image/omnigen/omnigen2",
|
|
375
|
+
"tags": ["Image Edit", "Image"],
|
|
376
|
+
"models": ["OmniGen"],
|
|
377
|
+
"date": "2025-06-30",
|
|
378
|
+
"size": 15784004813
|
|
414
379
|
},
|
|
415
380
|
{
|
|
416
|
-
"name": "
|
|
417
|
-
"title": "
|
|
381
|
+
"name": "hidream_i1_dev",
|
|
382
|
+
"title": "HiDream I1 Dev",
|
|
418
383
|
"mediaType": "image",
|
|
419
384
|
"mediaSubtype": "webp",
|
|
420
|
-
"description": "Generate images with
|
|
421
|
-
"
|
|
422
|
-
"
|
|
423
|
-
"models": ["
|
|
424
|
-
"date": "2025-
|
|
425
|
-
"size":
|
|
385
|
+
"description": "Generate images with HiDream I1 Dev - Balanced version with 28 inference steps, suitable for medium-range hardware.",
|
|
386
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/image/hidream/hidream-i1",
|
|
387
|
+
"tags": ["Text to Image", "Image"],
|
|
388
|
+
"models": ["HiDream"],
|
|
389
|
+
"date": "2025-04-17",
|
|
390
|
+
"size": 33318208799
|
|
426
391
|
},
|
|
427
392
|
{
|
|
428
|
-
"name": "
|
|
429
|
-
"title": "
|
|
393
|
+
"name": "hidream_i1_fast",
|
|
394
|
+
"title": "HiDream I1 Fast",
|
|
430
395
|
"mediaType": "image",
|
|
431
396
|
"mediaSubtype": "webp",
|
|
432
|
-
"
|
|
433
|
-
"
|
|
434
|
-
"
|
|
435
|
-
"
|
|
436
|
-
"
|
|
437
|
-
"
|
|
438
|
-
"size": 34037615821
|
|
397
|
+
"description": "Generate images quickly with HiDream I1 Fast - Lightweight version with 16 inference steps, ideal for rapid previews on lower-end hardware.",
|
|
398
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/image/hidream/hidream-i1",
|
|
399
|
+
"tags": ["Text to Image", "Image"],
|
|
400
|
+
"models": ["HiDream"],
|
|
401
|
+
"date": "2025-04-17",
|
|
402
|
+
"size": 24234352968
|
|
439
403
|
},
|
|
440
404
|
{
|
|
441
|
-
"name": "
|
|
442
|
-
"title": "
|
|
405
|
+
"name": "hidream_i1_full",
|
|
406
|
+
"title": "HiDream I1 Full",
|
|
443
407
|
"mediaType": "image",
|
|
444
408
|
"mediaSubtype": "webp",
|
|
445
|
-
"
|
|
446
|
-
"
|
|
447
|
-
"
|
|
448
|
-
"
|
|
449
|
-
"
|
|
450
|
-
"
|
|
451
|
-
"size": 31772020572
|
|
409
|
+
"description": "Generate images with HiDream I1 Full - Complete version with 50 inference steps for highest quality output.",
|
|
410
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/image/hidream/hidream-i1",
|
|
411
|
+
"tags": ["Text to Image", "Image"],
|
|
412
|
+
"models": ["HiDream"],
|
|
413
|
+
"date": "2025-04-17",
|
|
414
|
+
"size": 24234352968
|
|
452
415
|
},
|
|
453
416
|
{
|
|
454
|
-
"name": "
|
|
455
|
-
"title": "
|
|
417
|
+
"name": "hidream_e1_1",
|
|
418
|
+
"title": "HiDream E1.1 Image Edit",
|
|
456
419
|
"mediaType": "image",
|
|
457
420
|
"mediaSubtype": "webp",
|
|
458
421
|
"thumbnailVariant": "compareSlider",
|
|
459
|
-
"description": "Edit images with
|
|
460
|
-
"tutorialUrl": "https://docs.comfy.org/tutorials/image/
|
|
461
|
-
"tags": ["Image
|
|
462
|
-
"models": ["
|
|
463
|
-
"date": "2025-
|
|
464
|
-
"size":
|
|
422
|
+
"description": "Edit images with HiDream E1.1 – it’s better in image quality and editing accuracy than HiDream-E1-Full.",
|
|
423
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/image/hidream/hidream-e1",
|
|
424
|
+
"tags": ["Image Edit", "Image"],
|
|
425
|
+
"models": ["HiDream"],
|
|
426
|
+
"date": "2025-07-21",
|
|
427
|
+
"size": 50422916055
|
|
465
428
|
},
|
|
466
429
|
{
|
|
467
|
-
"name": "
|
|
468
|
-
"title": "
|
|
430
|
+
"name": "hidream_e1_full",
|
|
431
|
+
"title": "HiDream E1 Image Edit",
|
|
469
432
|
"mediaType": "image",
|
|
470
433
|
"mediaSubtype": "webp",
|
|
471
434
|
"thumbnailVariant": "compareSlider",
|
|
472
|
-
"description": "
|
|
473
|
-
"
|
|
474
|
-
"
|
|
475
|
-
"
|
|
476
|
-
"
|
|
477
|
-
|
|
478
|
-
{
|
|
479
|
-
"name": "flux_kontext_dev_basic",
|
|
480
|
-
"title": "Flux Kontext Dev Image Edit",
|
|
481
|
-
"mediaType": "image",
|
|
482
|
-
"mediaSubtype": "webp",
|
|
483
|
-
"thumbnailVariant": "hoverDissolve",
|
|
484
|
-
"description": "Smart image editing that keeps characters consistent, edits specific parts without affecting others, and preserves original styles.",
|
|
485
|
-
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-kontext-dev",
|
|
486
|
-
"tags": ["Image Edit", "Image to Image"],
|
|
487
|
-
"models": ["Flux", "BFL"],
|
|
488
|
-
"date": "2025-06-26",
|
|
489
|
-
"size": 17641578168,
|
|
490
|
-
"vram": 19327352832
|
|
491
|
-
},
|
|
492
|
-
{
|
|
493
|
-
"name": "image_chroma1_radiance_text_to_image",
|
|
494
|
-
"title": "Chroma1 Radiance text to image",
|
|
495
|
-
"mediaType": "image",
|
|
496
|
-
"mediaSubtype": "webp",
|
|
497
|
-
"description": "Chroma1-Radiance works directly with image pixels instead of compressed latents, delivering higher quality images with reduced artifacts and distortion.",
|
|
498
|
-
"tags": ["Text to Image", "Image"],
|
|
499
|
-
"models": ["Chroma"],
|
|
500
|
-
"date": "2025-09-18",
|
|
501
|
-
"size": 23622320128,
|
|
502
|
-
"vram": 23622320128
|
|
503
|
-
},
|
|
504
|
-
{
|
|
505
|
-
"name": "image_netayume_lumina_t2i",
|
|
506
|
-
"title": "NetaYume Lumina Text to Image",
|
|
507
|
-
"mediaType": "image",
|
|
508
|
-
"mediaSubtype": "webp",
|
|
509
|
-
"description": "High-quality anime-style image generation with enhanced character understanding and detailed textures. Fine-tuned from Neta Lumina on Danbooru dataset.",
|
|
510
|
-
"tags": ["Text to Image", "Image", "Anime"],
|
|
511
|
-
"models": ["OmniGen"],
|
|
512
|
-
"date": "2025-10-10",
|
|
513
|
-
"size": 10619306639
|
|
435
|
+
"description": "Edit images with HiDream E1 - Professional natural language image editing model.",
|
|
436
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/image/hidream/hidream-e1",
|
|
437
|
+
"tags": ["Image Edit", "Image"],
|
|
438
|
+
"models": ["HiDream"],
|
|
439
|
+
"date": "2025-05-01",
|
|
440
|
+
"size": 34209414513
|
|
514
441
|
},
|
|
515
442
|
{
|
|
516
|
-
"name": "
|
|
517
|
-
"title": "
|
|
443
|
+
"name": "sd3.5_simple_example",
|
|
444
|
+
"title": "SD3.5 Simple",
|
|
518
445
|
"mediaType": "image",
|
|
519
446
|
"mediaSubtype": "webp",
|
|
520
|
-
"description": "
|
|
447
|
+
"description": "Generate images using SD 3.5.",
|
|
448
|
+
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sd3/#sd35",
|
|
521
449
|
"tags": ["Text to Image", "Image"],
|
|
522
|
-
"models": ["
|
|
523
|
-
"date": "2025-
|
|
524
|
-
"size":
|
|
525
|
-
"vram": 15569256448
|
|
450
|
+
"models": ["SD3.5", "Stability"],
|
|
451
|
+
"date": "2025-03-01",
|
|
452
|
+
"size": 14935748772
|
|
526
453
|
},
|
|
527
454
|
{
|
|
528
|
-
"name": "
|
|
529
|
-
"title": "
|
|
455
|
+
"name": "sd3.5_large_canny_controlnet_example",
|
|
456
|
+
"title": "SD3.5 Large Canny ControlNet",
|
|
530
457
|
"mediaType": "image",
|
|
531
458
|
"mediaSubtype": "webp",
|
|
532
|
-
"
|
|
533
|
-
"
|
|
534
|
-
"
|
|
535
|
-
"
|
|
536
|
-
"
|
|
537
|
-
"
|
|
538
|
-
"
|
|
459
|
+
"description": "Generate images guided by edge detection using SD 3.5 Canny ControlNet.",
|
|
460
|
+
"thumbnailVariant": "hoverDissolve",
|
|
461
|
+
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sd3/#sd35-controlnets",
|
|
462
|
+
"tags": ["Image to Image", "Image", "ControlNet"],
|
|
463
|
+
"models": ["SD3.5", "Stability"],
|
|
464
|
+
"date": "2025-03-01",
|
|
465
|
+
"size": 23590107873
|
|
539
466
|
},
|
|
540
467
|
{
|
|
541
|
-
"name": "
|
|
542
|
-
"title": "
|
|
468
|
+
"name": "sd3.5_large_depth",
|
|
469
|
+
"title": "SD3.5 Large Depth",
|
|
543
470
|
"mediaType": "image",
|
|
544
471
|
"mediaSubtype": "webp",
|
|
545
|
-
"description": "Generate images
|
|
546
|
-
"
|
|
547
|
-
"
|
|
548
|
-
"
|
|
472
|
+
"description": "Generate images guided by depth information using SD 3.5.",
|
|
473
|
+
"thumbnailVariant": "hoverDissolve",
|
|
474
|
+
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sd3/#sd35-controlnets",
|
|
475
|
+
"tags": ["Image to Image", "Image", "ControlNet"],
|
|
476
|
+
"models": ["SD3.5", "Stability"],
|
|
549
477
|
"date": "2025-03-01",
|
|
550
|
-
"size":
|
|
551
|
-
"vram": 18253611008
|
|
478
|
+
"size": 23590107873
|
|
552
479
|
},
|
|
553
480
|
{
|
|
554
|
-
"name": "
|
|
555
|
-
"title": "
|
|
556
|
-
"description": "Use reference images to control both style and subject - keep your character's face while changing artistic style, or apply artistic styles to new scenes",
|
|
557
|
-
"thumbnailVariant": "hoverDissolve",
|
|
481
|
+
"name": "sd3.5_large_blur",
|
|
482
|
+
"title": "SD3.5 Large Blur",
|
|
558
483
|
"mediaType": "image",
|
|
559
484
|
"mediaSubtype": "webp",
|
|
485
|
+
"description": "Generate images guided by blurred reference images using SD 3.5.",
|
|
486
|
+
"thumbnailVariant": "hoverDissolve",
|
|
487
|
+
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sd3/#sd35-controlnets",
|
|
560
488
|
"tags": ["Image to Image", "Image"],
|
|
561
|
-
"models": ["
|
|
562
|
-
"date": "2025-
|
|
563
|
-
"
|
|
564
|
-
"size": 18597208392,
|
|
565
|
-
"vram": 19864223744
|
|
489
|
+
"models": ["SD3.5", "Stability"],
|
|
490
|
+
"date": "2025-03-01",
|
|
491
|
+
"size": 23590107873
|
|
566
492
|
},
|
|
567
493
|
{
|
|
568
|
-
"name": "
|
|
569
|
-
"title": "
|
|
494
|
+
"name": "sdxl_simple_example",
|
|
495
|
+
"title": "SDXL Simple",
|
|
570
496
|
"mediaType": "image",
|
|
571
497
|
"mediaSubtype": "webp",
|
|
572
|
-
"description": "
|
|
573
|
-
"tutorialUrl": "https://
|
|
498
|
+
"description": "Generate high-quality images using SDXL.",
|
|
499
|
+
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sdxl/",
|
|
574
500
|
"tags": ["Text to Image", "Image"],
|
|
575
|
-
"models": ["
|
|
501
|
+
"models": ["SDXL", "Stability"],
|
|
576
502
|
"date": "2025-03-01",
|
|
577
|
-
"size":
|
|
578
|
-
"vram": 18253611008
|
|
503
|
+
"size": 13013750907
|
|
579
504
|
},
|
|
580
505
|
{
|
|
581
|
-
"name": "
|
|
582
|
-
"title": "
|
|
506
|
+
"name": "sdxl_refiner_prompt_example",
|
|
507
|
+
"title": "SDXL Refiner Prompt",
|
|
583
508
|
"mediaType": "image",
|
|
584
509
|
"mediaSubtype": "webp",
|
|
585
|
-
"description": "
|
|
586
|
-
"tutorialUrl": "https://
|
|
510
|
+
"description": "Enhance SDXL images using refiner models.",
|
|
511
|
+
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sdxl/",
|
|
587
512
|
"tags": ["Text to Image", "Image"],
|
|
588
|
-
"models": ["
|
|
589
|
-
"date": "2025-
|
|
590
|
-
"size":
|
|
591
|
-
"vram": 23085449216
|
|
513
|
+
"models": ["SDXL", "Stability"],
|
|
514
|
+
"date": "2025-03-01",
|
|
515
|
+
"size": 13013750907
|
|
592
516
|
},
|
|
593
517
|
{
|
|
594
|
-
"name": "
|
|
595
|
-
"title": "
|
|
518
|
+
"name": "sdxl_revision_text_prompts",
|
|
519
|
+
"title": "SDXL Revision Text Prompts",
|
|
596
520
|
"mediaType": "image",
|
|
597
521
|
"mediaSubtype": "webp",
|
|
598
|
-
"description": "Generate
|
|
599
|
-
"tutorialUrl": "https://
|
|
522
|
+
"description": "Generate images by transferring concepts from reference images using SDXL Revision.",
|
|
523
|
+
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sdxl/#revision",
|
|
600
524
|
"tags": ["Text to Image", "Image"],
|
|
601
|
-
"models": ["
|
|
525
|
+
"models": ["SDXL", "Stability"],
|
|
602
526
|
"date": "2025-03-01",
|
|
603
|
-
"size":
|
|
604
|
-
"vram": 23622320128
|
|
527
|
+
"size": 10630044058
|
|
605
528
|
},
|
|
606
529
|
{
|
|
607
|
-
"name": "
|
|
608
|
-
"title": "
|
|
530
|
+
"name": "sdxlturbo_example",
|
|
531
|
+
"title": "SDXL Turbo",
|
|
609
532
|
"mediaType": "image",
|
|
610
533
|
"mediaSubtype": "webp",
|
|
611
|
-
"description": "Generate images
|
|
612
|
-
"tutorialUrl": "https://
|
|
534
|
+
"description": "Generate images in a single step using SDXL Turbo.",
|
|
535
|
+
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sdturbo/",
|
|
613
536
|
"tags": ["Text to Image", "Image"],
|
|
614
|
-
"models": ["
|
|
537
|
+
"models": ["SDXL", "Stability"],
|
|
615
538
|
"date": "2025-03-01",
|
|
616
|
-
"size":
|
|
617
|
-
},
|
|
618
|
-
{
|
|
619
|
-
"name": "flux_fill_inpaint_example",
|
|
620
|
-
"title": "Flux Inpaint",
|
|
621
|
-
"mediaType": "image",
|
|
622
|
-
"mediaSubtype": "webp",
|
|
623
|
-
"description": "Fill missing parts of images using Flux inpainting.",
|
|
624
|
-
"thumbnailVariant": "compareSlider",
|
|
625
|
-
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-fill-dev",
|
|
626
|
-
"tags": ["Image to Image", "Inpainting", "Image"],
|
|
627
|
-
"models": ["Flux", "BFL"],
|
|
628
|
-
"date": "2025-03-01",
|
|
629
|
-
"size": 10372346020
|
|
630
|
-
},
|
|
631
|
-
{
|
|
632
|
-
"name": "flux_fill_outpaint_example",
|
|
633
|
-
"title": "Flux Outpaint",
|
|
634
|
-
"mediaType": "image",
|
|
635
|
-
"mediaSubtype": "webp",
|
|
636
|
-
"description": "Extend images beyond boundaries using Flux outpainting.",
|
|
637
|
-
"thumbnailVariant": "compareSlider",
|
|
638
|
-
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-fill-dev",
|
|
639
|
-
"tags": ["Outpainting", "Image", "Image to Image"],
|
|
640
|
-
"models": ["Flux", "BFL"],
|
|
641
|
-
"date": "2025-03-01",
|
|
642
|
-
"size": 10372346020
|
|
643
|
-
},
|
|
644
|
-
{
|
|
645
|
-
"name": "flux_canny_model_example",
|
|
646
|
-
"title": "Flux Canny Model",
|
|
647
|
-
"mediaType": "image",
|
|
648
|
-
"mediaSubtype": "webp",
|
|
649
|
-
"description": "Generate images guided by edge detection using Flux Canny.",
|
|
650
|
-
"thumbnailVariant": "hoverDissolve",
|
|
651
|
-
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-controlnet",
|
|
652
|
-
"tags": ["Image to Image", "ControlNet", "Image"],
|
|
653
|
-
"models": ["Flux", "BFL"],
|
|
654
|
-
"date": "2025-03-01",
|
|
655
|
-
"size": 34177202258
|
|
656
|
-
},
|
|
657
|
-
{
|
|
658
|
-
"name": "flux_depth_lora_example",
|
|
659
|
-
"title": "Flux Depth Lora",
|
|
660
|
-
"mediaType": "image",
|
|
661
|
-
"mediaSubtype": "webp",
|
|
662
|
-
"description": "Generate images guided by depth information using Flux LoRA.",
|
|
663
|
-
"thumbnailVariant": "hoverDissolve",
|
|
664
|
-
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-controlnet",
|
|
665
|
-
"tags": ["Image to Image", "ControlNet", "Image"],
|
|
666
|
-
"models": ["Flux", "BFL"],
|
|
667
|
-
"date": "2025-03-01",
|
|
668
|
-
"size": 35412005356
|
|
669
|
-
},
|
|
670
|
-
{
|
|
671
|
-
"name": "flux_redux_model_example",
|
|
672
|
-
"title": "Flux Redux Model",
|
|
673
|
-
"mediaType": "image",
|
|
674
|
-
"mediaSubtype": "webp",
|
|
675
|
-
"description": "Generate images by transferring style from reference images using Flux Redux.",
|
|
676
|
-
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-controlnet",
|
|
677
|
-
"tags": ["Image to Image", "ControlNet", "Image"],
|
|
678
|
-
"models": ["Flux", "BFL"],
|
|
679
|
-
"date": "2025-03-01",
|
|
680
|
-
"size": 35154307318
|
|
681
|
-
},
|
|
682
|
-
{
|
|
683
|
-
"name": "image_omnigen2_t2i",
|
|
684
|
-
"title": "OmniGen2 Text to Image",
|
|
685
|
-
"mediaType": "image",
|
|
686
|
-
"mediaSubtype": "webp",
|
|
687
|
-
"description": "Generate high-quality images from text prompts using OmniGen2's unified 7B multimodal model with dual-path architecture.",
|
|
688
|
-
"tutorialUrl": "https://docs.comfy.org/tutorials/image/omnigen/omnigen2",
|
|
689
|
-
"tags": ["Text to Image", "Image"],
|
|
690
|
-
"models": ["OmniGen"],
|
|
691
|
-
"date": "2025-06-30",
|
|
692
|
-
"size": 15784004813
|
|
693
|
-
},
|
|
694
|
-
{
|
|
695
|
-
"name": "image_omnigen2_image_edit",
|
|
696
|
-
"title": "OmniGen2 Image Edit",
|
|
697
|
-
"mediaType": "image",
|
|
698
|
-
"mediaSubtype": "webp",
|
|
699
|
-
"thumbnailVariant": "hoverDissolve",
|
|
700
|
-
"description": "Edit images with natural language instructions using OmniGen2's advanced image editing capabilities and text rendering support.",
|
|
701
|
-
"tutorialUrl": "https://docs.comfy.org/tutorials/image/omnigen/omnigen2",
|
|
702
|
-
"tags": ["Image Edit", "Image"],
|
|
703
|
-
"models": ["OmniGen"],
|
|
704
|
-
"date": "2025-06-30",
|
|
705
|
-
"size": 15784004813
|
|
706
|
-
},
|
|
707
|
-
{
|
|
708
|
-
"name": "hidream_i1_dev",
|
|
709
|
-
"title": "HiDream I1 Dev",
|
|
710
|
-
"mediaType": "image",
|
|
711
|
-
"mediaSubtype": "webp",
|
|
712
|
-
"description": "Generate images with HiDream I1 Dev - Balanced version with 28 inference steps, suitable for medium-range hardware.",
|
|
713
|
-
"tutorialUrl": "https://docs.comfy.org/tutorials/image/hidream/hidream-i1",
|
|
714
|
-
"tags": ["Text to Image", "Image"],
|
|
715
|
-
"models": ["HiDream"],
|
|
716
|
-
"date": "2025-04-17",
|
|
717
|
-
"size": 33318208799
|
|
718
|
-
},
|
|
719
|
-
{
|
|
720
|
-
"name": "hidream_i1_fast",
|
|
721
|
-
"title": "HiDream I1 Fast",
|
|
722
|
-
"mediaType": "image",
|
|
723
|
-
"mediaSubtype": "webp",
|
|
724
|
-
"description": "Generate images quickly with HiDream I1 Fast - Lightweight version with 16 inference steps, ideal for rapid previews on lower-end hardware.",
|
|
725
|
-
"tutorialUrl": "https://docs.comfy.org/tutorials/image/hidream/hidream-i1",
|
|
726
|
-
"tags": ["Text to Image", "Image"],
|
|
727
|
-
"models": ["HiDream"],
|
|
728
|
-
"date": "2025-04-17",
|
|
729
|
-
"size": 24234352968
|
|
730
|
-
},
|
|
731
|
-
{
|
|
732
|
-
"name": "hidream_i1_full",
|
|
733
|
-
"title": "HiDream I1 Full",
|
|
734
|
-
"mediaType": "image",
|
|
735
|
-
"mediaSubtype": "webp",
|
|
736
|
-
"description": "Generate images with HiDream I1 Full - Complete version with 50 inference steps for highest quality output.",
|
|
737
|
-
"tutorialUrl": "https://docs.comfy.org/tutorials/image/hidream/hidream-i1",
|
|
738
|
-
"tags": ["Text to Image", "Image"],
|
|
739
|
-
"models": ["HiDream"],
|
|
740
|
-
"date": "2025-04-17",
|
|
741
|
-
"size": 24234352968
|
|
742
|
-
},
|
|
743
|
-
{
|
|
744
|
-
"name": "hidream_e1_1",
|
|
745
|
-
"title": "HiDream E1.1 Image Edit",
|
|
746
|
-
"mediaType": "image",
|
|
747
|
-
"mediaSubtype": "webp",
|
|
748
|
-
"thumbnailVariant": "compareSlider",
|
|
749
|
-
"description": "Edit images with HiDream E1.1 – it’s better in image quality and editing accuracy than HiDream-E1-Full.",
|
|
750
|
-
"tutorialUrl": "https://docs.comfy.org/tutorials/image/hidream/hidream-e1",
|
|
751
|
-
"tags": ["Image Edit", "Image"],
|
|
752
|
-
"models": ["HiDream"],
|
|
753
|
-
"date": "2025-07-21",
|
|
754
|
-
"size": 50422916055
|
|
755
|
-
},
|
|
756
|
-
{
|
|
757
|
-
"name": "hidream_e1_full",
|
|
758
|
-
"title": "HiDream E1 Image Edit",
|
|
759
|
-
"mediaType": "image",
|
|
760
|
-
"mediaSubtype": "webp",
|
|
761
|
-
"thumbnailVariant": "compareSlider",
|
|
762
|
-
"description": "Edit images with HiDream E1 - Professional natural language image editing model.",
|
|
763
|
-
"tutorialUrl": "https://docs.comfy.org/tutorials/image/hidream/hidream-e1",
|
|
764
|
-
"tags": ["Image Edit", "Image"],
|
|
765
|
-
"models": ["HiDream"],
|
|
766
|
-
"date": "2025-05-01",
|
|
767
|
-
"size": 34209414513
|
|
768
|
-
},
|
|
769
|
-
{
|
|
770
|
-
"name": "sd3.5_simple_example",
|
|
771
|
-
"title": "SD3.5 Simple",
|
|
772
|
-
"mediaType": "image",
|
|
773
|
-
"mediaSubtype": "webp",
|
|
774
|
-
"description": "Generate images using SD 3.5.",
|
|
775
|
-
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sd3/#sd35",
|
|
776
|
-
"tags": ["Text to Image", "Image"],
|
|
777
|
-
"models": ["SD3.5", "Stability"],
|
|
778
|
-
"date": "2025-03-01",
|
|
779
|
-
"size": 14935748772
|
|
780
|
-
},
|
|
781
|
-
{
|
|
782
|
-
"name": "sd3.5_large_canny_controlnet_example",
|
|
783
|
-
"title": "SD3.5 Large Canny ControlNet",
|
|
784
|
-
"mediaType": "image",
|
|
785
|
-
"mediaSubtype": "webp",
|
|
786
|
-
"description": "Generate images guided by edge detection using SD 3.5 Canny ControlNet.",
|
|
787
|
-
"thumbnailVariant": "hoverDissolve",
|
|
788
|
-
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sd3/#sd35-controlnets",
|
|
789
|
-
"tags": ["Image to Image", "Image", "ControlNet"],
|
|
790
|
-
"models": ["SD3.5", "Stability"],
|
|
791
|
-
"date": "2025-03-01",
|
|
792
|
-
"size": 23590107873
|
|
793
|
-
},
|
|
794
|
-
{
|
|
795
|
-
"name": "sd3.5_large_depth",
|
|
796
|
-
"title": "SD3.5 Large Depth",
|
|
797
|
-
"mediaType": "image",
|
|
798
|
-
"mediaSubtype": "webp",
|
|
799
|
-
"description": "Generate images guided by depth information using SD 3.5.",
|
|
800
|
-
"thumbnailVariant": "hoverDissolve",
|
|
801
|
-
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sd3/#sd35-controlnets",
|
|
802
|
-
"tags": ["Image to Image", "Image", "ControlNet"],
|
|
803
|
-
"models": ["SD3.5", "Stability"],
|
|
804
|
-
"date": "2025-03-01",
|
|
805
|
-
"size": 23590107873
|
|
806
|
-
},
|
|
807
|
-
{
|
|
808
|
-
"name": "sd3.5_large_blur",
|
|
809
|
-
"title": "SD3.5 Large Blur",
|
|
810
|
-
"mediaType": "image",
|
|
811
|
-
"mediaSubtype": "webp",
|
|
812
|
-
"description": "Generate images guided by blurred reference images using SD 3.5.",
|
|
813
|
-
"thumbnailVariant": "hoverDissolve",
|
|
814
|
-
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sd3/#sd35-controlnets",
|
|
815
|
-
"tags": ["Image to Image", "Image"],
|
|
816
|
-
"models": ["SD3.5", "Stability"],
|
|
817
|
-
"date": "2025-03-01",
|
|
818
|
-
"size": 23590107873
|
|
819
|
-
},
|
|
820
|
-
{
|
|
821
|
-
"name": "sdxl_simple_example",
|
|
822
|
-
"title": "SDXL Simple",
|
|
823
|
-
"mediaType": "image",
|
|
824
|
-
"mediaSubtype": "webp",
|
|
825
|
-
"description": "Generate high-quality images using SDXL.",
|
|
826
|
-
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sdxl/",
|
|
827
|
-
"tags": ["Text to Image", "Image"],
|
|
828
|
-
"models": ["SDXL", "Stability"],
|
|
829
|
-
"date": "2025-03-01",
|
|
830
|
-
"size": 13013750907
|
|
831
|
-
},
|
|
832
|
-
{
|
|
833
|
-
"name": "sdxl_refiner_prompt_example",
|
|
834
|
-
"title": "SDXL Refiner Prompt",
|
|
835
|
-
"mediaType": "image",
|
|
836
|
-
"mediaSubtype": "webp",
|
|
837
|
-
"description": "Enhance SDXL images using refiner models.",
|
|
838
|
-
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sdxl/",
|
|
839
|
-
"tags": ["Text to Image", "Image"],
|
|
840
|
-
"models": ["SDXL", "Stability"],
|
|
841
|
-
"date": "2025-03-01",
|
|
842
|
-
"size": 13013750907
|
|
843
|
-
},
|
|
844
|
-
{
|
|
845
|
-
"name": "sdxl_revision_text_prompts",
|
|
846
|
-
"title": "SDXL Revision Text Prompts",
|
|
847
|
-
"mediaType": "image",
|
|
848
|
-
"mediaSubtype": "webp",
|
|
849
|
-
"description": "Generate images by transferring concepts from reference images using SDXL Revision.",
|
|
850
|
-
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sdxl/#revision",
|
|
851
|
-
"tags": ["Text to Image", "Image"],
|
|
852
|
-
"models": ["SDXL", "Stability"],
|
|
853
|
-
"date": "2025-03-01",
|
|
854
|
-
"size": 10630044058
|
|
855
|
-
},
|
|
856
|
-
{
|
|
857
|
-
"name": "sdxlturbo_example",
|
|
858
|
-
"title": "SDXL Turbo",
|
|
859
|
-
"mediaType": "image",
|
|
860
|
-
"mediaSubtype": "webp",
|
|
861
|
-
"description": "Generate images in a single step using SDXL Turbo.",
|
|
862
|
-
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sdturbo/",
|
|
863
|
-
"tags": ["Text to Image", "Image"],
|
|
864
|
-
"models": ["SDXL", "Stability"],
|
|
865
|
-
"date": "2025-03-01",
|
|
866
|
-
"size": 6936372183
|
|
539
|
+
"size": 6936372183
|
|
867
540
|
},
|
|
868
541
|
{
|
|
869
542
|
"name": "image_lotus_depth_v1_1",
|
|
@@ -2523,5 +2196,332 @@
|
|
|
2523
2196
|
"vram": 0
|
|
2524
2197
|
}
|
|
2525
2198
|
]
|
|
2199
|
+
},
|
|
2200
|
+
{
|
|
2201
|
+
"moduleName": "default",
|
|
2202
|
+
"isEssential": true,
|
|
2203
|
+
"title": "Getting Started",
|
|
2204
|
+
"type": "image",
|
|
2205
|
+
"templates": [
|
|
2206
|
+
{
|
|
2207
|
+
"name": "01_qwen_t2i_subgraphed",
|
|
2208
|
+
"title": "Text to Image (New)",
|
|
2209
|
+
"mediaType": "image",
|
|
2210
|
+
"mediaSubtype": "webp",
|
|
2211
|
+
"description": "Generate images from text prompts using the Qwen-Image model.",
|
|
2212
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image",
|
|
2213
|
+
"tags": ["Text to Image", "Image"],
|
|
2214
|
+
"models": ["Qwen-Image"],
|
|
2215
|
+
"date": "2025-10-17",
|
|
2216
|
+
"size": 31772020572
|
|
2217
|
+
},
|
|
2218
|
+
{
|
|
2219
|
+
"name": "02_qwen_Image_edit_subgraphed",
|
|
2220
|
+
"title": "Image Editing (New)",
|
|
2221
|
+
"mediaType": "image",
|
|
2222
|
+
"mediaSubtype": "webp",
|
|
2223
|
+
"description": "Edit your images with Qwen-Image-Edit, the latest OSS model",
|
|
2224
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image-edit",
|
|
2225
|
+
"tags": ["Image to Image", "Image Edit", "ControlNet"],
|
|
2226
|
+
"models": ["Qwen-Image"],
|
|
2227
|
+
"date": "2025-10-17",
|
|
2228
|
+
"size": 31772020572
|
|
2229
|
+
},
|
|
2230
|
+
{
|
|
2231
|
+
"name": "03_video_wan2_2_14B_i2v_subgraphed",
|
|
2232
|
+
"title": "Image to Video (New)",
|
|
2233
|
+
"description": "Generate videos from an input image using Wan2.2 14B",
|
|
2234
|
+
"mediaType": "image",
|
|
2235
|
+
"mediaSubtype": "webp",
|
|
2236
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/wan2_2",
|
|
2237
|
+
"tags": ["Image to Video", "Video"],
|
|
2238
|
+
"models": ["Wan2.2", "Wan"],
|
|
2239
|
+
"date": "2025-10-17",
|
|
2240
|
+
"size": 38031935406
|
|
2241
|
+
},
|
|
2242
|
+
{
|
|
2243
|
+
"name": "04_hunyuan_3d_2.1_subgraphed",
|
|
2244
|
+
"title": "Image to 3D (New)",
|
|
2245
|
+
"mediaType": "image",
|
|
2246
|
+
"mediaSubtype": "webp",
|
|
2247
|
+
"description": "Generate 3D models from single images using Hunyuan3D 2.1.",
|
|
2248
|
+
"tags": ["Image to 3D", "3D"],
|
|
2249
|
+
"models": ["Hunyuan3D"],
|
|
2250
|
+
"date": "2025-10-17",
|
|
2251
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/3d/hunyuan3D-2",
|
|
2252
|
+
"size": 4928474972
|
|
2253
|
+
},
|
|
2254
|
+
{
|
|
2255
|
+
"name": "05_audio_ace_step_1_t2a_song_subgraphed",
|
|
2256
|
+
"title": "Text to Song (New)",
|
|
2257
|
+
"mediaType": "image",
|
|
2258
|
+
"mediaSubtype": "webp",
|
|
2259
|
+
"description": "Generate songs from text prompts using ACE-Step v1",
|
|
2260
|
+
"tags": ["Text to Audio", "Audio"],
|
|
2261
|
+
"models": ["ACE-Step"],
|
|
2262
|
+
"date": "2025-10-17",
|
|
2263
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/audio/ace-step/ace-step-v1",
|
|
2264
|
+
"size": 7698728878
|
|
2265
|
+
},
|
|
2266
|
+
{
|
|
2267
|
+
"name": "default",
|
|
2268
|
+
"title": "Image Generation",
|
|
2269
|
+
"mediaType": "image",
|
|
2270
|
+
"mediaSubtype": "webp",
|
|
2271
|
+
"description": "Generate images from text prompts.",
|
|
2272
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/basic/text-to-image",
|
|
2273
|
+
"tags": ["Text to Image", "Image"],
|
|
2274
|
+
"models": ["SD1.5", "Stability"],
|
|
2275
|
+
"date": "2025-03-01",
|
|
2276
|
+
"size": 2136746230,
|
|
2277
|
+
"vram": 3092376453
|
|
2278
|
+
},
|
|
2279
|
+
{
|
|
2280
|
+
"name": "image2image",
|
|
2281
|
+
"title": "Image to Image",
|
|
2282
|
+
"mediaType": "image",
|
|
2283
|
+
"mediaSubtype": "webp",
|
|
2284
|
+
"thumbnailVariant": "hoverDissolve",
|
|
2285
|
+
"description": "Transform existing images using text prompts.",
|
|
2286
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/basic/image-to-image",
|
|
2287
|
+
"tags": ["Image to Image", "Image"],
|
|
2288
|
+
"models": ["SD1.5", "Stability"],
|
|
2289
|
+
"date": "2025-03-01",
|
|
2290
|
+
"size": 2136746230,
|
|
2291
|
+
"vram": 3092376453
|
|
2292
|
+
},
|
|
2293
|
+
{
|
|
2294
|
+
"name": "lora",
|
|
2295
|
+
"title": "LoRA",
|
|
2296
|
+
"mediaType": "image",
|
|
2297
|
+
"mediaSubtype": "webp",
|
|
2298
|
+
"description": "Generate images with LoRA models for specialized styles or subjects.",
|
|
2299
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/basic/lora",
|
|
2300
|
+
"tags": ["Text to Image", "Image"],
|
|
2301
|
+
"models": ["SD1.5", "Stability"],
|
|
2302
|
+
"date": "2025-03-01",
|
|
2303
|
+
"size": 2437393940,
|
|
2304
|
+
"vram": 3092376453
|
|
2305
|
+
},
|
|
2306
|
+
{
|
|
2307
|
+
"name": "lora_multiple",
|
|
2308
|
+
"title": "LoRA Multiple",
|
|
2309
|
+
"mediaType": "image",
|
|
2310
|
+
"mediaSubtype": "webp",
|
|
2311
|
+
"description": "Generate images by combining multiple LoRA models.",
|
|
2312
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/basic/lora",
|
|
2313
|
+
"tags": ["Text to Image", "Image"],
|
|
2314
|
+
"models": ["SD1.5", "Stability"],
|
|
2315
|
+
"date": "2025-03-01",
|
|
2316
|
+
"size": 2437393940,
|
|
2317
|
+
"vram": 3350074491
|
|
2318
|
+
},
|
|
2319
|
+
{
|
|
2320
|
+
"name": "inpaint_example",
|
|
2321
|
+
"title": "Inpaint",
|
|
2322
|
+
"mediaType": "image",
|
|
2323
|
+
"mediaSubtype": "webp",
|
|
2324
|
+
"description": "Edit specific parts of images seamlessly.",
|
|
2325
|
+
"thumbnailVariant": "compareSlider",
|
|
2326
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/basic/inpaint",
|
|
2327
|
+
"tags": ["Inpainting", "Image"],
|
|
2328
|
+
"models": ["SD1.5", "Stability"],
|
|
2329
|
+
"date": "2025-03-01",
|
|
2330
|
+
"size": 5218385265,
|
|
2331
|
+
"vram": 4101693768
|
|
2332
|
+
},
|
|
2333
|
+
{
|
|
2334
|
+
"name": "inpaint_model_outpainting",
|
|
2335
|
+
"title": "Outpaint",
|
|
2336
|
+
"mediaType": "image",
|
|
2337
|
+
"mediaSubtype": "webp",
|
|
2338
|
+
"description": "Extend images beyond their original boundaries.",
|
|
2339
|
+
"thumbnailVariant": "compareSlider",
|
|
2340
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/basic/inpaint",
|
|
2341
|
+
"tags": ["Outpainting", "Image"],
|
|
2342
|
+
"models": ["SD1.5", "Stability"],
|
|
2343
|
+
"date": "2025-03-01",
|
|
2344
|
+
"size": 5218385265,
|
|
2345
|
+
"vram": 4101693768
|
|
2346
|
+
},
|
|
2347
|
+
{
|
|
2348
|
+
"name": "embedding_example",
|
|
2349
|
+
"title": "Embedding",
|
|
2350
|
+
"mediaType": "image",
|
|
2351
|
+
"mediaSubtype": "webp",
|
|
2352
|
+
"description": "Generate images using textual inversion for consistent styles.",
|
|
2353
|
+
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/textual_inversion_embeddings/",
|
|
2354
|
+
"tags": ["Text to Image", "Image"],
|
|
2355
|
+
"models": ["SD1.5", "Stability"],
|
|
2356
|
+
"date": "2025-03-01",
|
|
2357
|
+
"size": 5218385265,
|
|
2358
|
+
"vram": 4123168604
|
|
2359
|
+
},
|
|
2360
|
+
{
|
|
2361
|
+
"name": "gligen_textbox_example",
|
|
2362
|
+
"title": "Gligen Textbox",
|
|
2363
|
+
"mediaType": "image",
|
|
2364
|
+
"mediaSubtype": "webp",
|
|
2365
|
+
"description": "Generate images with precise object placement using text boxes.",
|
|
2366
|
+
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/gligen/",
|
|
2367
|
+
"tags": ["Image"],
|
|
2368
|
+
"models": ["SD1.5", "Stability"],
|
|
2369
|
+
"date": "2025-03-01",
|
|
2370
|
+
"size": 2974264852,
|
|
2371
|
+
"vram": 4080218931
|
|
2372
|
+
},
|
|
2373
|
+
{
|
|
2374
|
+
"name": "area_composition",
|
|
2375
|
+
"title": "Area Composition",
|
|
2376
|
+
"mediaType": "image",
|
|
2377
|
+
"mediaSubtype": "webp",
|
|
2378
|
+
"description": "Generate images by controlling composition with defined areas.",
|
|
2379
|
+
"tags": ["Text to Image", "Image"],
|
|
2380
|
+
"models": ["SD1.5", "Stability"],
|
|
2381
|
+
"date": "2025-03-01",
|
|
2382
|
+
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/area_composition/",
|
|
2383
|
+
"size": 2469606195,
|
|
2384
|
+
"vram": 6184752906
|
|
2385
|
+
},
|
|
2386
|
+
{
|
|
2387
|
+
"name": "area_composition_square_area_for_subject",
|
|
2388
|
+
"title": "Area Composition Square Area for Subject",
|
|
2389
|
+
"mediaType": "image",
|
|
2390
|
+
"mediaSubtype": "webp",
|
|
2391
|
+
"description": "Generate images with consistent subject placement using area composition.",
|
|
2392
|
+
"tags": ["Text to Image", "Image"],
|
|
2393
|
+
"models": ["SD1.5", "Stability"],
|
|
2394
|
+
"date": "2025-03-01",
|
|
2395
|
+
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/area_composition/#increasing-consistency-of-images-with-area-composition",
|
|
2396
|
+
"size": 2469606195,
|
|
2397
|
+
"vram": 5927054868
|
|
2398
|
+
},
|
|
2399
|
+
{
|
|
2400
|
+
"name": "hiresfix_latent_workflow",
|
|
2401
|
+
"title": "Upscale",
|
|
2402
|
+
"mediaType": "image",
|
|
2403
|
+
"mediaSubtype": "webp",
|
|
2404
|
+
"description": "Upscale images by enhancing quality in latent space.",
|
|
2405
|
+
"thumbnailVariant": "compareSlider",
|
|
2406
|
+
"tags": ["Upscale", "Image"],
|
|
2407
|
+
"models": ["SD1.5", "Stability"],
|
|
2408
|
+
"date": "2025-03-01",
|
|
2409
|
+
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/2_pass_txt2img/",
|
|
2410
|
+
"size": 2136746230,
|
|
2411
|
+
"vram": 3929895076
|
|
2412
|
+
},
|
|
2413
|
+
{
|
|
2414
|
+
"name": "esrgan_example",
|
|
2415
|
+
"title": "ESRGAN",
|
|
2416
|
+
"mediaType": "image",
|
|
2417
|
+
"mediaSubtype": "webp",
|
|
2418
|
+
"description": "Upscale images using ESRGAN models to enhance quality.",
|
|
2419
|
+
"thumbnailVariant": "compareSlider",
|
|
2420
|
+
"tags": ["Upscale", "Image"],
|
|
2421
|
+
"models": ["SD1.5", "Stability"],
|
|
2422
|
+
"date": "2025-03-01",
|
|
2423
|
+
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/upscale_models/",
|
|
2424
|
+
"size": 2201170739,
|
|
2425
|
+
"vram": 6442450944
|
|
2426
|
+
},
|
|
2427
|
+
{
|
|
2428
|
+
"name": "hiresfix_esrgan_workflow",
|
|
2429
|
+
"title": "HiresFix ESRGAN Workflow",
|
|
2430
|
+
"mediaType": "image",
|
|
2431
|
+
"mediaSubtype": "webp",
|
|
2432
|
+
"description": "Upscale images using ESRGAN models during intermediate generation steps.",
|
|
2433
|
+
"thumbnailVariant": "compareSlider",
|
|
2434
|
+
"tags": ["Upscale", "Image"],
|
|
2435
|
+
"models": ["SD1.5", "Stability"],
|
|
2436
|
+
"date": "2025-03-01",
|
|
2437
|
+
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/2_pass_txt2img/#non-latent-upscaling",
|
|
2438
|
+
"size": 2201170739,
|
|
2439
|
+
"vram": 6442450944
|
|
2440
|
+
},
|
|
2441
|
+
{
|
|
2442
|
+
"name": "latent_upscale_different_prompt_model",
|
|
2443
|
+
"title": "Latent Upscale Different Prompt Model",
|
|
2444
|
+
"mediaType": "image",
|
|
2445
|
+
"mediaSubtype": "webp",
|
|
2446
|
+
"description": "Upscale images while changing prompts across generation passes.",
|
|
2447
|
+
"thumbnailVariant": "zoomHover",
|
|
2448
|
+
"tags": ["Upscale", "Image"],
|
|
2449
|
+
"models": ["SD1.5", "Stability"],
|
|
2450
|
+
"date": "2025-03-01",
|
|
2451
|
+
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/2_pass_txt2img/#more-examples",
|
|
2452
|
+
"size": 4262755041,
|
|
2453
|
+
"vram": 5153960755
|
|
2454
|
+
},
|
|
2455
|
+
{
|
|
2456
|
+
"name": "controlnet_example",
|
|
2457
|
+
"title": "Scribble ControlNet",
|
|
2458
|
+
"mediaType": "image",
|
|
2459
|
+
"mediaSubtype": "webp",
|
|
2460
|
+
"description": "Generate images guided by scribble reference images using ControlNet.",
|
|
2461
|
+
"thumbnailVariant": "hoverDissolve",
|
|
2462
|
+
"tags": ["ControlNet", "Image"],
|
|
2463
|
+
"models": ["SD1.5", "Stability"],
|
|
2464
|
+
"date": "2025-03-01",
|
|
2465
|
+
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/controlnet/",
|
|
2466
|
+
"size": 3189013217,
|
|
2467
|
+
"vram": 6442450944
|
|
2468
|
+
},
|
|
2469
|
+
{
|
|
2470
|
+
"name": "2_pass_pose_worship",
|
|
2471
|
+
"title": "Pose ControlNet 2 Pass",
|
|
2472
|
+
"mediaType": "image",
|
|
2473
|
+
"mediaSubtype": "webp",
|
|
2474
|
+
"description": "Generate images guided by pose references using ControlNet.",
|
|
2475
|
+
"thumbnailVariant": "hoverDissolve",
|
|
2476
|
+
"tags": ["ControlNet", "Image"],
|
|
2477
|
+
"models": ["SD1.5", "Stability"],
|
|
2478
|
+
"date": "2025-03-01",
|
|
2479
|
+
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/controlnet/#pose-controlnet",
|
|
2480
|
+
"size": 4660039516,
|
|
2481
|
+
"vram": 6442450944
|
|
2482
|
+
},
|
|
2483
|
+
{
|
|
2484
|
+
"name": "depth_controlnet",
|
|
2485
|
+
"title": "Depth ControlNet",
|
|
2486
|
+
"mediaType": "image",
|
|
2487
|
+
"mediaSubtype": "webp",
|
|
2488
|
+
"description": "Generate images guided by depth information using ControlNet.",
|
|
2489
|
+
"thumbnailVariant": "hoverDissolve",
|
|
2490
|
+
"tags": ["ControlNet", "Image", "Text to Image"],
|
|
2491
|
+
"models": ["SD1.5", "Stability"],
|
|
2492
|
+
"date": "2025-03-01",
|
|
2493
|
+
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/controlnet/#t2i-adapter-vs-controlnets",
|
|
2494
|
+
"size": 2888365507,
|
|
2495
|
+
"vram": 6442450944
|
|
2496
|
+
},
|
|
2497
|
+
{
|
|
2498
|
+
"name": "depth_t2i_adapter",
|
|
2499
|
+
"title": "Depth T2I Adapter",
|
|
2500
|
+
"mediaType": "image",
|
|
2501
|
+
"mediaSubtype": "webp",
|
|
2502
|
+
"description": "Generate images guided by depth information using T2I adapter.",
|
|
2503
|
+
"thumbnailVariant": "hoverDissolve",
|
|
2504
|
+
"tags": ["ControlNet", "Image", "Text to Image"],
|
|
2505
|
+
"models": ["SD1.5", "Stability"],
|
|
2506
|
+
"date": "2025-03-01",
|
|
2507
|
+
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/controlnet/#t2i-adapter-vs-controlnets",
|
|
2508
|
+
"size": 2523293286,
|
|
2509
|
+
"vram": 6442450944
|
|
2510
|
+
},
|
|
2511
|
+
{
|
|
2512
|
+
"name": "mixing_controlnets",
|
|
2513
|
+
"title": "Mixing ControlNets",
|
|
2514
|
+
"mediaType": "image",
|
|
2515
|
+
"mediaSubtype": "webp",
|
|
2516
|
+
"description": "Generate images by combining multiple ControlNet models.",
|
|
2517
|
+
"thumbnailVariant": "hoverDissolve",
|
|
2518
|
+
"tags": ["ControlNet", "Image", "Text to Image"],
|
|
2519
|
+
"models": ["SD1.5", "Stability"],
|
|
2520
|
+
"date": "2025-03-01",
|
|
2521
|
+
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/controlnet/#mixing-controlnets",
|
|
2522
|
+
"size": 3328599654,
|
|
2523
|
+
"vram": 6442450944
|
|
2524
|
+
}
|
|
2525
|
+
]
|
|
2526
2526
|
}
|
|
2527
2527
|
]
|