comfyui-workflow-templates-media-other 0.3.10__py3-none-any.whl → 0.3.14__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1,869 +1,542 @@
1
1
  [
2
2
  {
3
3
  "moduleName": "default",
4
- "isEssential": true,
5
- "title": "Getting Started",
4
+ "category": "GENERATION TYPE",
5
+ "icon": "icon-[lucide--image]",
6
+ "title": "Image",
6
7
  "type": "image",
7
8
  "templates": [
8
9
  {
9
- "name": "01_qwen_t2i_subgraphed",
10
- "title": "Text to Image (New)",
10
+ "name": "image_flux2",
11
+ "title": "Flux.2 Dev",
11
12
  "mediaType": "image",
12
13
  "mediaSubtype": "webp",
13
- "description": "Generate images from text prompts using the Qwen-Image model.",
14
- "tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image",
15
- "tags": ["Text to Image", "Image"],
16
- "models": ["Qwen-Image"],
17
- "date": "2025-10-17",
18
- "size": 31772020572
14
+ "thumbnailVariant": "compareSlider",
15
+ "description": "Generate up to 4MP photorealistic images with multi-reference consistency and professional text rendering.",
16
+ "tags": ["Text to Image", "Image", "Image Edit"],
17
+ "models": ["Flux.2 Dev", "BFL"],
18
+ "date": "2025-11-26",
19
+ "size": 71382356459,
20
+ "vram": 0
19
21
  },
20
22
  {
21
- "name": "02_qwen_Image_edit_subgraphed",
22
- "title": "Image Editing (New)",
23
+ "name": "image_flux2_fp8",
24
+ "title": "Product Mockup(Flux.2 Dev FP8)",
23
25
  "mediaType": "image",
24
26
  "mediaSubtype": "webp",
25
- "description": "Edit your images with Qwen-Image-Edit, the latest OSS model",
26
- "tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image-edit",
27
- "tags": ["Image to Image", "Image Edit", "ControlNet"],
28
- "models": ["Qwen-Image"],
29
- "date": "2025-10-17",
30
- "size": 31772020572
27
+ "description": "Create product mockups by applying design patterns to packaging, mugs, and other products using multi-reference consistency.",
28
+ "tags": [
29
+ "Text to Image",
30
+ "Image",
31
+ "Image Edit",
32
+ "Mockup",
33
+ "Product Design"
34
+ ],
35
+ "models": ["Flux.2 Dev", "BFL"],
36
+ "date": "2025-11-26",
37
+ "size": 53837415055,
38
+ "vram": 0
31
39
  },
32
40
  {
33
- "name": "03_video_wan2_2_14B_i2v_subgraphed",
34
- "title": "Image to Video (New)",
35
- "description": "Generate videos from an input image using Wan2.2 14B",
41
+ "name": "image_z_image_turbo",
42
+ "title": "Z-Image-Turbo Text to Image",
36
43
  "mediaType": "image",
37
44
  "mediaSubtype": "webp",
38
- "tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/wan2_2",
39
- "tags": ["Image to Video", "Video"],
40
- "models": ["Wan2.2", "Wan"],
41
- "date": "2025-10-17",
42
- "size": 38031935406
45
+ "description": "An Efficient Image Generation Foundation Model with Single-Stream Diffusion Transformer, supports English & Chinese.",
46
+ "tags": ["Text to Image", "Image"],
47
+ "models": ["Z-Image-Turbo"],
48
+ "date": "2025-11-27",
49
+ "size": 35326050304
43
50
  },
44
51
  {
45
- "name": "04_hunyuan_3d_2.1_subgraphed",
46
- "title": "Image to 3D (New)",
52
+ "name": "image_qwen_image",
53
+ "title": "Qwen-Image Text to Image",
47
54
  "mediaType": "image",
48
55
  "mediaSubtype": "webp",
49
- "description": "Generate 3D models from single images using Hunyuan3D 2.1.",
50
- "tags": ["Image to 3D", "3D"],
51
- "models": ["Hunyuan3D"],
52
- "date": "2025-10-17",
53
- "tutorialUrl": "https://docs.comfy.org/tutorials/3d/hunyuan3D-2",
54
- "size": 4928474972
56
+ "description": "Generate images with exceptional multilingual text rendering and editing capabilities using Qwen-Image's 20B MMDiT model..",
57
+ "tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image",
58
+ "tags": ["Text to Image", "Image"],
59
+ "models": ["Qwen-Image"],
60
+ "date": "2025-08-05",
61
+ "size": 31772020572
55
62
  },
56
63
  {
57
- "name": "05_audio_ace_step_1_t2a_song_subgraphed",
58
- "title": "Text to Song (New)",
64
+ "name": "image_qwen_image_instantx_controlnet",
65
+ "title": "Qwen-Image InstantX Union ControlNet",
59
66
  "mediaType": "image",
60
67
  "mediaSubtype": "webp",
61
- "description": "Generate songs from text prompts using ACE-Step v1",
62
- "tags": ["Text to Audio", "Audio"],
63
- "models": ["ACE-Step"],
64
- "date": "2025-10-17",
65
- "tutorialUrl": "https://docs.comfy.org/tutorials/audio/ace-step/ace-step-v1",
66
- "size": 7698728878
68
+ "description": "Generate images with Qwen-Image InstantX ControlNet, supporting canny, soft edge, depth, pose",
69
+ "tags": ["Image to Image", "Image", "ControlNet"],
70
+ "tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image",
71
+ "models": ["Qwen-Image"],
72
+ "date": "2025-08-23",
73
+ "size": 35304631173
67
74
  },
68
75
  {
69
- "name": "default",
70
- "title": "Image Generation",
76
+ "name": "image_qwen_image_instantx_inpainting_controlnet",
77
+ "title": "Qwen-Image InstantX Inpainting ControlNet",
71
78
  "mediaType": "image",
72
79
  "mediaSubtype": "webp",
73
- "description": "Generate images from text prompts.",
74
- "tutorialUrl": "https://docs.comfy.org/tutorials/basic/text-to-image",
75
- "tags": ["Text to Image", "Image"],
76
- "models": ["SD1.5", "Stability"],
77
- "date": "2025-03-01",
78
- "size": 2136746230,
79
- "vram": 3092376453
80
+ "thumbnailVariant": "compareSlider",
81
+ "description": "Professional inpainting and image editing with Qwen-Image InstantX ControlNet. Supports object replacement, text modification, background changes, and outpainting.",
82
+ "tags": ["Image to Image", "Image", "ControlNet", "Inpainting"],
83
+ "tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image",
84
+ "models": ["Qwen-Image"],
85
+ "date": "2025-09-12",
86
+ "size": 36013300777
80
87
  },
81
88
  {
82
- "name": "image2image",
83
- "title": "Image to Image",
89
+ "name": "image_qwen_image_union_control_lora",
90
+ "title": "Qwen-Image Union Control",
84
91
  "mediaType": "image",
85
92
  "mediaSubtype": "webp",
86
- "thumbnailVariant": "hoverDissolve",
87
- "description": "Transform existing images using text prompts.",
88
- "tutorialUrl": "https://docs.comfy.org/tutorials/basic/image-to-image",
89
- "tags": ["Image to Image", "Image"],
90
- "models": ["SD1.5", "Stability"],
91
- "date": "2025-03-01",
92
- "size": 2136746230,
93
- "vram": 3092376453
93
+ "description": "Generate images with precise structural control using Qwen-Image's unified ControlNet LoRA. Supports multiple control types including canny, depth, lineart, softedge, normal, and openpose for diverse creative applications.",
94
+ "tags": ["Text to Image", "Image", "ControlNet"],
95
+ "tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image",
96
+ "models": ["Qwen-Image"],
97
+ "date": "2025-08-23",
98
+ "size": 32716913377
94
99
  },
95
100
  {
96
- "name": "lora",
97
- "title": "LoRA",
101
+ "name": "image_qwen_image_controlnet_patch",
102
+ "title": "Qwen-Image ControlNet model patch",
98
103
  "mediaType": "image",
99
104
  "mediaSubtype": "webp",
100
- "description": "Generate images with LoRA models for specialized styles or subjects.",
101
- "tutorialUrl": "https://docs.comfy.org/tutorials/basic/lora",
102
- "tags": ["Text to Image", "Image"],
103
- "models": ["SD1.5", "Stability"],
104
- "date": "2025-03-01",
105
- "size": 2437393940,
106
- "vram": 3092376453
105
+ "thumbnailVariant": "compareSlider",
106
+ "description": "Control image generation using Qwen-Image ControlNet models. Supports canny, depth, and inpainting controls through model patching.",
107
+ "tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image",
108
+ "tags": ["Text to Image", "Image", "ControlNet"],
109
+ "models": ["Qwen-Image"],
110
+ "date": "2025-08-24",
111
+ "size": 34037615821
107
112
  },
108
113
  {
109
- "name": "lora_multiple",
110
- "title": "LoRA Multiple",
114
+ "name": "image_qwen_image_edit_2509",
115
+ "title": "Qwen Image Edit 2509",
111
116
  "mediaType": "image",
112
117
  "mediaSubtype": "webp",
113
- "description": "Generate images by combining multiple LoRA models.",
114
- "tutorialUrl": "https://docs.comfy.org/tutorials/basic/lora",
115
- "tags": ["Text to Image", "Image"],
116
- "models": ["SD1.5", "Stability"],
117
- "date": "2025-03-01",
118
- "size": 2437393940,
119
- "vram": 3350074491
118
+ "thumbnailVariant": "compareSlider",
119
+ "description": "Advanced image editing with multi-image support, improved consistency, and ControlNet integration.",
120
+ "tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image-edit",
121
+ "tags": ["Image to Image", "Image Edit", "ControlNet"],
122
+ "models": ["Qwen-Image"],
123
+ "date": "2025-09-25",
124
+ "size": 31772020572
120
125
  },
121
126
  {
122
- "name": "inpaint_example",
123
- "title": "Inpaint",
127
+ "name": "image_qwen_image_edit",
128
+ "title": "Qwen Image Edit",
124
129
  "mediaType": "image",
125
130
  "mediaSubtype": "webp",
126
- "description": "Edit specific parts of images seamlessly.",
127
131
  "thumbnailVariant": "compareSlider",
128
- "tutorialUrl": "https://docs.comfy.org/tutorials/basic/inpaint",
129
- "tags": ["Inpainting", "Image"],
130
- "models": ["SD1.5", "Stability"],
131
- "date": "2025-03-01",
132
- "size": 5218385265,
133
- "vram": 4101693768
132
+ "description": "Edit images with precise bilingual text editing and dual semantic/appearance editing capabilities using Qwen-Image-Edit's 20B MMDiT model.",
133
+ "tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image-edit",
134
+ "tags": ["Image to Image", "Image Edit"],
135
+ "models": ["Qwen-Image"],
136
+ "date": "2025-08-18",
137
+ "size": 31772020572
134
138
  },
135
139
  {
136
- "name": "inpaint_model_outpainting",
137
- "title": "Outpaint",
140
+ "name": "image_chrono_edit_14B",
141
+ "title": "ChronoEdit 14B",
138
142
  "mediaType": "image",
139
143
  "mediaSubtype": "webp",
140
- "description": "Extend images beyond their original boundaries.",
141
144
  "thumbnailVariant": "compareSlider",
142
- "tutorialUrl": "https://docs.comfy.org/tutorials/basic/inpaint",
143
- "tags": ["Outpainting", "Image"],
144
- "models": ["SD1.5", "Stability"],
145
- "date": "2025-03-01",
146
- "size": 5218385265,
147
- "vram": 4101693768
145
+ "description": "Image editing powered by video models' dynamic understanding, creating physically plausible results while preserving character and style consistency.",
146
+ "tags": ["Image Edit", "Image to Image"],
147
+ "models": ["Wan2.1", "ChronoEdit", "Nvidia"],
148
+ "date": "2025-11-03",
149
+ "size": 40459304
148
150
  },
149
151
  {
150
- "name": "embedding_example",
151
- "title": "Embedding",
152
+ "name": "flux_kontext_dev_basic",
153
+ "title": "Flux Kontext Dev Image Edit",
152
154
  "mediaType": "image",
153
155
  "mediaSubtype": "webp",
154
- "description": "Generate images using textual inversion for consistent styles.",
155
- "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/textual_inversion_embeddings/",
156
- "tags": ["Text to Image", "Image"],
157
- "models": ["SD1.5", "Stability"],
158
- "date": "2025-03-01",
159
- "size": 5218385265,
160
- "vram": 4123168604
156
+ "thumbnailVariant": "hoverDissolve",
157
+ "description": "Smart image editing that keeps characters consistent, edits specific parts without affecting others, and preserves original styles.",
158
+ "tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-kontext-dev",
159
+ "tags": ["Image Edit", "Image to Image"],
160
+ "models": ["Flux", "BFL"],
161
+ "date": "2025-06-26",
162
+ "size": 17641578168,
163
+ "vram": 19327352832
161
164
  },
162
165
  {
163
- "name": "gligen_textbox_example",
164
- "title": "Gligen Textbox",
166
+ "name": "image_chroma1_radiance_text_to_image",
167
+ "title": "Chroma1 Radiance text to image",
165
168
  "mediaType": "image",
166
169
  "mediaSubtype": "webp",
167
- "description": "Generate images with precise object placement using text boxes.",
168
- "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/gligen/",
169
- "tags": ["Image"],
170
- "models": ["SD1.5", "Stability"],
171
- "date": "2025-03-01",
172
- "size": 2974264852,
173
- "vram": 4080218931
170
+ "description": "Chroma1-Radiance works directly with image pixels instead of compressed latents, delivering higher quality images with reduced artifacts and distortion.",
171
+ "tags": ["Text to Image", "Image"],
172
+ "models": ["Chroma"],
173
+ "date": "2025-09-18",
174
+ "size": 23622320128,
175
+ "vram": 23622320128
174
176
  },
175
177
  {
176
- "name": "area_composition",
177
- "title": "Area Composition",
178
+ "name": "image_netayume_lumina_t2i",
179
+ "title": "NetaYume Lumina Text to Image",
178
180
  "mediaType": "image",
179
181
  "mediaSubtype": "webp",
180
- "description": "Generate images by controlling composition with defined areas.",
181
- "tags": ["Text to Image", "Image"],
182
- "models": ["SD1.5", "Stability"],
183
- "date": "2025-03-01",
184
- "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/area_composition/",
185
- "size": 2469606195,
186
- "vram": 6184752906
182
+ "description": "High-quality anime-style image generation with enhanced character understanding and detailed textures. Fine-tuned from Neta Lumina on Danbooru dataset.",
183
+ "tags": ["Text to Image", "Image", "Anime"],
184
+ "models": ["OmniGen"],
185
+ "date": "2025-10-10",
186
+ "size": 10619306639
187
187
  },
188
188
  {
189
- "name": "area_composition_square_area_for_subject",
190
- "title": "Area Composition Square Area for Subject",
189
+ "name": "image_chroma_text_to_image",
190
+ "title": "Chroma text to image",
191
191
  "mediaType": "image",
192
192
  "mediaSubtype": "webp",
193
- "description": "Generate images with consistent subject placement using area composition.",
193
+ "description": "Chroma - enhanced Flux model with improved image quality and better prompt understanding for stunning text-to-image generation.",
194
194
  "tags": ["Text to Image", "Image"],
195
- "models": ["SD1.5", "Stability"],
196
- "date": "2025-03-01",
197
- "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/area_composition/#increasing-consistency-of-images-with-area-composition",
198
- "size": 2469606195,
199
- "vram": 5927054868
195
+ "models": ["Chroma", "Flux"],
196
+ "date": "2025-06-04",
197
+ "size": 23289460163,
198
+ "vram": 15569256448
200
199
  },
201
200
  {
202
- "name": "hiresfix_latent_workflow",
203
- "title": "Upscale",
201
+ "name": "image_flux.1_fill_dev_OneReward",
202
+ "title": "Flux.1 Dev OneReward",
204
203
  "mediaType": "image",
205
204
  "mediaSubtype": "webp",
206
- "description": "Upscale images by enhancing quality in latent space.",
207
205
  "thumbnailVariant": "compareSlider",
208
- "tags": ["Upscale", "Image"],
209
- "models": ["SD1.5", "Stability"],
210
- "date": "2025-03-01",
211
- "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/2_pass_txt2img/",
212
- "size": 2136746230,
213
- "vram": 3929895076
206
+ "description": "Supports various tasks such as image inpainting, outpainting, and object removal by bytedance-research team",
207
+ "tags": ["Inpainting", "Outpainting"],
208
+ "models": ["Flux", "BFL"],
209
+ "date": "2025-09-21",
210
+ "size": 29001766666,
211
+ "vram": 21474836480
214
212
  },
215
213
  {
216
- "name": "esrgan_example",
217
- "title": "ESRGAN",
214
+ "name": "flux_dev_checkpoint_example",
215
+ "title": "Flux Dev fp8",
218
216
  "mediaType": "image",
219
217
  "mediaSubtype": "webp",
220
- "description": "Upscale images using ESRGAN models to enhance quality.",
221
- "thumbnailVariant": "compareSlider",
222
- "tags": ["Upscale", "Image"],
223
- "models": ["SD1.5", "Stability"],
218
+ "description": "Generate images using Flux Dev fp8 quantized version. Suitable for devices with limited VRAM, requires only one model file, but image quality is slightly lower than the full version.",
219
+ "tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-text-to-image",
220
+ "tags": ["Text to Image", "Image"],
221
+ "models": ["Flux", "BFL"],
224
222
  "date": "2025-03-01",
225
- "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/upscale_models/",
226
- "size": 2201170739,
227
- "vram": 6442450944
223
+ "size": 17244293693,
224
+ "vram": 18253611008
228
225
  },
229
226
  {
230
- "name": "hiresfix_esrgan_workflow",
231
- "title": "HiresFix ESRGAN Workflow",
227
+ "name": "flux1_dev_uso_reference_image_gen",
228
+ "title": "Flux.1 Dev USO Reference Image Generation",
229
+ "description": "Use reference images to control both style and subject - keep your character's face while changing artistic style, or apply artistic styles to new scenes",
230
+ "thumbnailVariant": "hoverDissolve",
232
231
  "mediaType": "image",
233
232
  "mediaSubtype": "webp",
234
- "description": "Upscale images using ESRGAN models during intermediate generation steps.",
235
- "thumbnailVariant": "compareSlider",
236
- "tags": ["Upscale", "Image"],
237
- "models": ["SD1.5", "Stability"],
238
- "date": "2025-03-01",
239
- "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/2_pass_txt2img/#non-latent-upscaling",
240
- "size": 2201170739,
241
- "vram": 6442450944
233
+ "tags": ["Image to Image", "Image"],
234
+ "models": ["Flux", "BFL"],
235
+ "date": "2025-09-02",
236
+ "tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-uso",
237
+ "size": 18597208392,
238
+ "vram": 19864223744
242
239
  },
243
240
  {
244
- "name": "latent_upscale_different_prompt_model",
245
- "title": "Latent Upscale Different Prompt Model",
241
+ "name": "flux_schnell",
242
+ "title": "Flux Schnell fp8",
246
243
  "mediaType": "image",
247
244
  "mediaSubtype": "webp",
248
- "description": "Upscale images while changing prompts across generation passes.",
249
- "thumbnailVariant": "zoomHover",
250
- "tags": ["Upscale", "Image"],
251
- "models": ["SD1.5", "Stability"],
245
+ "description": "Quickly generate images with Flux Schnell fp8 quantized version. Ideal for low-end hardware, requires only 4 steps to generate images.",
246
+ "tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-text-to-image",
247
+ "tags": ["Text to Image", "Image"],
248
+ "models": ["Flux", "BFL"],
252
249
  "date": "2025-03-01",
253
- "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/2_pass_txt2img/#more-examples",
254
- "size": 4262755041,
255
- "vram": 5153960755
250
+ "size": 17233556275,
251
+ "vram": 18253611008
256
252
  },
257
253
  {
258
- "name": "controlnet_example",
259
- "title": "Scribble ControlNet",
254
+ "name": "flux1_krea_dev",
255
+ "title": "Flux.1 Krea Dev",
260
256
  "mediaType": "image",
261
257
  "mediaSubtype": "webp",
262
- "description": "Generate images guided by scribble reference images using ControlNet.",
263
- "thumbnailVariant": "hoverDissolve",
264
- "tags": ["ControlNet", "Image"],
265
- "models": ["SD1.5", "Stability"],
266
- "date": "2025-03-01",
267
- "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/controlnet/",
268
- "size": 3189013217,
269
- "vram": 6442450944
258
+ "description": "A fine-tuned FLUX model pushing photorealism to the max",
259
+ "tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux1-krea-dev",
260
+ "tags": ["Text to Image", "Image"],
261
+ "models": ["Flux", "BFL"],
262
+ "date": "2025-07-31",
263
+ "size": 22269405430,
264
+ "vram": 23085449216
270
265
  },
271
266
  {
272
- "name": "2_pass_pose_worship",
273
- "title": "Pose ControlNet 2 Pass",
267
+ "name": "flux_dev_full_text_to_image",
268
+ "title": "Flux Dev full text to image",
274
269
  "mediaType": "image",
275
270
  "mediaSubtype": "webp",
276
- "description": "Generate images guided by pose references using ControlNet.",
277
- "thumbnailVariant": "hoverDissolve",
278
- "tags": ["ControlNet", "Image"],
279
- "models": ["SD1.5", "Stability"],
271
+ "description": "Generate high-quality images with Flux Dev full version. Requires larger VRAM and multiple model files, but provides the best prompt following capability and image quality.",
272
+ "tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-text-to-image",
273
+ "tags": ["Text to Image", "Image"],
274
+ "models": ["Flux", "BFL"],
280
275
  "date": "2025-03-01",
281
- "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/controlnet/#pose-controlnet",
282
- "size": 4660039516,
283
- "vram": 6442450944
276
+ "size": 34177202258,
277
+ "vram": 23622320128
284
278
  },
285
279
  {
286
- "name": "depth_controlnet",
287
- "title": "Depth ControlNet",
280
+ "name": "flux_schnell_full_text_to_image",
281
+ "title": "Flux Schnell full text to image",
288
282
  "mediaType": "image",
289
283
  "mediaSubtype": "webp",
290
- "description": "Generate images guided by depth information using ControlNet.",
291
- "thumbnailVariant": "hoverDissolve",
292
- "tags": ["ControlNet", "Image", "Text to Image"],
293
- "models": ["SD1.5", "Stability"],
284
+ "description": "Generate images quickly with Flux Schnell full version. Uses Apache2.0 license, requires only 4 steps to generate images while maintaining good image quality.",
285
+ "tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-text-to-image",
286
+ "tags": ["Text to Image", "Image"],
287
+ "models": ["Flux", "BFL"],
294
288
  "date": "2025-03-01",
295
- "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/controlnet/#t2i-adapter-vs-controlnets",
296
- "size": 2888365507,
297
- "vram": 6442450944
289
+ "size": 34155727421
298
290
  },
299
291
  {
300
- "name": "depth_t2i_adapter",
301
- "title": "Depth T2I Adapter",
292
+ "name": "flux_fill_inpaint_example",
293
+ "title": "Flux Inpaint",
302
294
  "mediaType": "image",
303
295
  "mediaSubtype": "webp",
304
- "description": "Generate images guided by depth information using T2I adapter.",
305
- "thumbnailVariant": "hoverDissolve",
306
- "tags": ["ControlNet", "Image", "Text to Image"],
307
- "models": ["SD1.5", "Stability"],
296
+ "description": "Fill missing parts of images using Flux inpainting.",
297
+ "thumbnailVariant": "compareSlider",
298
+ "tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-fill-dev",
299
+ "tags": ["Image to Image", "Inpainting", "Image"],
300
+ "models": ["Flux", "BFL"],
308
301
  "date": "2025-03-01",
309
- "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/controlnet/#t2i-adapter-vs-controlnets",
310
- "size": 2523293286,
311
- "vram": 6442450944
302
+ "size": 10372346020
312
303
  },
313
304
  {
314
- "name": "mixing_controlnets",
315
- "title": "Mixing ControlNets",
305
+ "name": "flux_fill_outpaint_example",
306
+ "title": "Flux Outpaint",
316
307
  "mediaType": "image",
317
308
  "mediaSubtype": "webp",
318
- "description": "Generate images by combining multiple ControlNet models.",
319
- "thumbnailVariant": "hoverDissolve",
320
- "tags": ["ControlNet", "Image", "Text to Image"],
321
- "models": ["SD1.5", "Stability"],
309
+ "description": "Extend images beyond boundaries using Flux outpainting.",
310
+ "thumbnailVariant": "compareSlider",
311
+ "tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-fill-dev",
312
+ "tags": ["Outpainting", "Image", "Image to Image"],
313
+ "models": ["Flux", "BFL"],
322
314
  "date": "2025-03-01",
323
- "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/controlnet/#mixing-controlnets",
324
- "size": 3328599654,
325
- "vram": 6442450944
326
- }
327
- ]
328
- },
329
- {
330
- "moduleName": "default",
331
- "category": "GENERATION TYPE",
332
- "icon": "icon-[lucide--image]",
333
- "title": "Image",
334
- "type": "image",
335
- "templates": [
315
+ "size": 10372346020
316
+ },
336
317
  {
337
- "name": "image_flux2",
338
- "title": "Flux.2 Dev",
318
+ "name": "flux_canny_model_example",
319
+ "title": "Flux Canny Model",
339
320
  "mediaType": "image",
340
321
  "mediaSubtype": "webp",
341
- "thumbnailVariant": "compareSlider",
342
- "description": "Generate up to 4MP photorealistic images with multi-reference consistency and professional text rendering.",
343
- "tags": ["Text to Image", "Image", "Image Edit"],
344
- "models": ["Flux.2 Dev", "BFL"],
345
- "date": "2025-11-26",
346
- "size": 71382356459,
347
- "vram": 0
322
+ "description": "Generate images guided by edge detection using Flux Canny.",
323
+ "thumbnailVariant": "hoverDissolve",
324
+ "tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-controlnet",
325
+ "tags": ["Image to Image", "ControlNet", "Image"],
326
+ "models": ["Flux", "BFL"],
327
+ "date": "2025-03-01",
328
+ "size": 34177202258
348
329
  },
349
330
  {
350
- "name": "image_flux2_fp8",
351
- "title": "Product Mockup(Flux.2 Dev FP8)",
331
+ "name": "flux_depth_lora_example",
332
+ "title": "Flux Depth Lora",
352
333
  "mediaType": "image",
353
334
  "mediaSubtype": "webp",
354
- "description": "Create product mockups by applying design patterns to packaging, mugs, and other products using multi-reference consistency.",
355
- "tags": [
356
- "Text to Image",
357
- "Image",
358
- "Image Edit",
359
- "Mockup",
360
- "Product Design"
361
- ],
362
- "models": ["Flux.2 Dev", "BFL"],
363
- "date": "2025-11-26",
364
- "size": 53837415055,
365
- "vram": 0
335
+ "description": "Generate images guided by depth information using Flux LoRA.",
336
+ "thumbnailVariant": "hoverDissolve",
337
+ "tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-controlnet",
338
+ "tags": ["Image to Image", "ControlNet", "Image"],
339
+ "models": ["Flux", "BFL"],
340
+ "date": "2025-03-01",
341
+ "size": 35412005356
366
342
  },
367
343
  {
368
- "name": "image_z_image_turbo",
369
- "title": "Z-Image-Turbo Text to Image",
344
+ "name": "flux_redux_model_example",
345
+ "title": "Flux Redux Model",
370
346
  "mediaType": "image",
371
347
  "mediaSubtype": "webp",
372
- "description": "An Efficient Image Generation Foundation Model with Single-Stream Diffusion Transformer, supports English & Chinese.",
373
- "tags": ["Text to Image","Image"],
374
- "models": ["Z-Image-Turbo"],
375
- "date": "2025-11-27",
376
- "size": 35326050304
348
+ "description": "Generate images by transferring style from reference images using Flux Redux.",
349
+ "tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-controlnet",
350
+ "tags": ["Image to Image", "ControlNet", "Image"],
351
+ "models": ["Flux", "BFL"],
352
+ "date": "2025-03-01",
353
+ "size": 35154307318
377
354
  },
378
355
  {
379
- "name": "image_qwen_image",
380
- "title": "Qwen-Image Text to Image",
356
+ "name": "image_omnigen2_t2i",
357
+ "title": "OmniGen2 Text to Image",
381
358
  "mediaType": "image",
382
359
  "mediaSubtype": "webp",
383
- "description": "Generate images with exceptional multilingual text rendering and editing capabilities using Qwen-Image's 20B MMDiT model..",
384
- "tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image",
360
+ "description": "Generate high-quality images from text prompts using OmniGen2's unified 7B multimodal model with dual-path architecture.",
361
+ "tutorialUrl": "https://docs.comfy.org/tutorials/image/omnigen/omnigen2",
385
362
  "tags": ["Text to Image", "Image"],
386
- "models": ["Qwen-Image"],
387
- "date": "2025-08-05",
388
- "size": 31772020572
363
+ "models": ["OmniGen"],
364
+ "date": "2025-06-30",
365
+ "size": 15784004813
389
366
  },
390
367
  {
391
- "name": "image_qwen_image_instantx_controlnet",
392
- "title": "Qwen-Image InstantX Union ControlNet",
368
+ "name": "image_omnigen2_image_edit",
369
+ "title": "OmniGen2 Image Edit",
393
370
  "mediaType": "image",
394
371
  "mediaSubtype": "webp",
395
- "description": "Generate images with Qwen-Image InstantX ControlNet, supporting canny, soft edge, depth, pose",
396
- "tags": ["Image to Image", "Image", "ControlNet"],
397
- "tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image",
398
- "models": ["Qwen-Image"],
399
- "date": "2025-08-23",
400
- "size": 35304631173
401
- },
402
- {
403
- "name": "image_qwen_image_instantx_inpainting_controlnet",
404
- "title": "Qwen-Image InstantX Inpainting ControlNet",
405
- "mediaType": "image",
406
- "mediaSubtype": "webp",
407
- "thumbnailVariant": "compareSlider",
408
- "description": "Professional inpainting and image editing with Qwen-Image InstantX ControlNet. Supports object replacement, text modification, background changes, and outpainting.",
409
- "tags": ["Image to Image", "Image", "ControlNet", "Inpainting"],
410
- "tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image",
411
- "models": ["Qwen-Image"],
412
- "date": "2025-09-12",
413
- "size": 36013300777
372
+ "thumbnailVariant": "hoverDissolve",
373
+ "description": "Edit images with natural language instructions using OmniGen2's advanced image editing capabilities and text rendering support.",
374
+ "tutorialUrl": "https://docs.comfy.org/tutorials/image/omnigen/omnigen2",
375
+ "tags": ["Image Edit", "Image"],
376
+ "models": ["OmniGen"],
377
+ "date": "2025-06-30",
378
+ "size": 15784004813
414
379
  },
415
380
  {
416
- "name": "image_qwen_image_union_control_lora",
417
- "title": "Qwen-Image Union Control",
381
+ "name": "hidream_i1_dev",
382
+ "title": "HiDream I1 Dev",
418
383
  "mediaType": "image",
419
384
  "mediaSubtype": "webp",
420
- "description": "Generate images with precise structural control using Qwen-Image's unified ControlNet LoRA. Supports multiple control types including canny, depth, lineart, softedge, normal, and openpose for diverse creative applications.",
421
- "tags": ["Text to Image", "Image", "ControlNet"],
422
- "tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image",
423
- "models": ["Qwen-Image"],
424
- "date": "2025-08-23",
425
- "size": 32716913377
385
+ "description": "Generate images with HiDream I1 Dev - Balanced version with 28 inference steps, suitable for medium-range hardware.",
386
+ "tutorialUrl": "https://docs.comfy.org/tutorials/image/hidream/hidream-i1",
387
+ "tags": ["Text to Image", "Image"],
388
+ "models": ["HiDream"],
389
+ "date": "2025-04-17",
390
+ "size": 33318208799
426
391
  },
427
392
  {
428
- "name": "image_qwen_image_controlnet_patch",
429
- "title": "Qwen-Image ControlNet model patch",
393
+ "name": "hidream_i1_fast",
394
+ "title": "HiDream I1 Fast",
430
395
  "mediaType": "image",
431
396
  "mediaSubtype": "webp",
432
- "thumbnailVariant": "compareSlider",
433
- "description": "Control image generation using Qwen-Image ControlNet models. Supports canny, depth, and inpainting controls through model patching.",
434
- "tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image",
435
- "tags": ["Text to Image", "Image", "ControlNet"],
436
- "models": ["Qwen-Image"],
437
- "date": "2025-08-24",
438
- "size": 34037615821
397
+ "description": "Generate images quickly with HiDream I1 Fast - Lightweight version with 16 inference steps, ideal for rapid previews on lower-end hardware.",
398
+ "tutorialUrl": "https://docs.comfy.org/tutorials/image/hidream/hidream-i1",
399
+ "tags": ["Text to Image", "Image"],
400
+ "models": ["HiDream"],
401
+ "date": "2025-04-17",
402
+ "size": 24234352968
439
403
  },
440
404
  {
441
- "name": "image_qwen_image_edit_2509",
442
- "title": "Qwen Image Edit 2509",
405
+ "name": "hidream_i1_full",
406
+ "title": "HiDream I1 Full",
443
407
  "mediaType": "image",
444
408
  "mediaSubtype": "webp",
445
- "thumbnailVariant": "compareSlider",
446
- "description": "Advanced image editing with multi-image support, improved consistency, and ControlNet integration.",
447
- "tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image-edit",
448
- "tags": ["Image to Image", "Image Edit", "ControlNet"],
449
- "models": ["Qwen-Image"],
450
- "date": "2025-09-25",
451
- "size": 31772020572
409
+ "description": "Generate images with HiDream I1 Full - Complete version with 50 inference steps for highest quality output.",
410
+ "tutorialUrl": "https://docs.comfy.org/tutorials/image/hidream/hidream-i1",
411
+ "tags": ["Text to Image", "Image"],
412
+ "models": ["HiDream"],
413
+ "date": "2025-04-17",
414
+ "size": 24234352968
452
415
  },
453
416
  {
454
- "name": "image_qwen_image_edit",
455
- "title": "Qwen Image Edit",
417
+ "name": "hidream_e1_1",
418
+ "title": "HiDream E1.1 Image Edit",
456
419
  "mediaType": "image",
457
420
  "mediaSubtype": "webp",
458
421
  "thumbnailVariant": "compareSlider",
459
- "description": "Edit images with precise bilingual text editing and dual semantic/appearance editing capabilities using Qwen-Image-Edit's 20B MMDiT model.",
460
- "tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image-edit",
461
- "tags": ["Image to Image", "Image Edit"],
462
- "models": ["Qwen-Image"],
463
- "date": "2025-08-18",
464
- "size": 31772020572
422
+ "description": "Edit images with HiDream E1.1 it’s better in image quality and editing accuracy than HiDream-E1-Full.",
423
+ "tutorialUrl": "https://docs.comfy.org/tutorials/image/hidream/hidream-e1",
424
+ "tags": ["Image Edit", "Image"],
425
+ "models": ["HiDream"],
426
+ "date": "2025-07-21",
427
+ "size": 50422916055
465
428
  },
466
429
  {
467
- "name": "image_chrono_edit_14B",
468
- "title": "ChronoEdit 14B",
430
+ "name": "hidream_e1_full",
431
+ "title": "HiDream E1 Image Edit",
469
432
  "mediaType": "image",
470
433
  "mediaSubtype": "webp",
471
434
  "thumbnailVariant": "compareSlider",
472
- "description": "Image editing powered by video models' dynamic understanding, creating physically plausible results while preserving character and style consistency.",
473
- "tags": ["Image Edit", "Image to Image"],
474
- "models": ["Wan2.1", "ChronoEdit", "Nvidia"],
475
- "date": "2025-11-03",
476
- "size": 40459304
477
- },
478
- {
479
- "name": "flux_kontext_dev_basic",
480
- "title": "Flux Kontext Dev Image Edit",
481
- "mediaType": "image",
482
- "mediaSubtype": "webp",
483
- "thumbnailVariant": "hoverDissolve",
484
- "description": "Smart image editing that keeps characters consistent, edits specific parts without affecting others, and preserves original styles.",
485
- "tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-kontext-dev",
486
- "tags": ["Image Edit", "Image to Image"],
487
- "models": ["Flux", "BFL"],
488
- "date": "2025-06-26",
489
- "size": 17641578168,
490
- "vram": 19327352832
491
- },
492
- {
493
- "name": "image_chroma1_radiance_text_to_image",
494
- "title": "Chroma1 Radiance text to image",
495
- "mediaType": "image",
496
- "mediaSubtype": "webp",
497
- "description": "Chroma1-Radiance works directly with image pixels instead of compressed latents, delivering higher quality images with reduced artifacts and distortion.",
498
- "tags": ["Text to Image", "Image"],
499
- "models": ["Chroma"],
500
- "date": "2025-09-18",
501
- "size": 23622320128,
502
- "vram": 23622320128
503
- },
504
- {
505
- "name": "image_netayume_lumina_t2i",
506
- "title": "NetaYume Lumina Text to Image",
507
- "mediaType": "image",
508
- "mediaSubtype": "webp",
509
- "description": "High-quality anime-style image generation with enhanced character understanding and detailed textures. Fine-tuned from Neta Lumina on Danbooru dataset.",
510
- "tags": ["Text to Image", "Image", "Anime"],
511
- "models": ["OmniGen"],
512
- "date": "2025-10-10",
513
- "size": 10619306639
435
+ "description": "Edit images with HiDream E1 - Professional natural language image editing model.",
436
+ "tutorialUrl": "https://docs.comfy.org/tutorials/image/hidream/hidream-e1",
437
+ "tags": ["Image Edit", "Image"],
438
+ "models": ["HiDream"],
439
+ "date": "2025-05-01",
440
+ "size": 34209414513
514
441
  },
515
442
  {
516
- "name": "image_chroma_text_to_image",
517
- "title": "Chroma text to image",
443
+ "name": "sd3.5_simple_example",
444
+ "title": "SD3.5 Simple",
518
445
  "mediaType": "image",
519
446
  "mediaSubtype": "webp",
520
- "description": "Chroma - enhanced Flux model with improved image quality and better prompt understanding for stunning text-to-image generation.",
447
+ "description": "Generate images using SD 3.5.",
448
+ "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sd3/#sd35",
521
449
  "tags": ["Text to Image", "Image"],
522
- "models": ["Chroma", "Flux"],
523
- "date": "2025-06-04",
524
- "size": 23289460163,
525
- "vram": 15569256448
450
+ "models": ["SD3.5", "Stability"],
451
+ "date": "2025-03-01",
452
+ "size": 14935748772
526
453
  },
527
454
  {
528
- "name": "image_flux.1_fill_dev_OneReward",
529
- "title": "Flux.1 Dev OneReward",
455
+ "name": "sd3.5_large_canny_controlnet_example",
456
+ "title": "SD3.5 Large Canny ControlNet",
530
457
  "mediaType": "image",
531
458
  "mediaSubtype": "webp",
532
- "thumbnailVariant": "compareSlider",
533
- "description": "Supports various tasks such as image inpainting, outpainting, and object removal by bytedance-research team",
534
- "tags": ["Inpainting", "Outpainting"],
535
- "models": ["Flux", "BFL"],
536
- "date": "2025-09-21",
537
- "size": 29001766666,
538
- "vram": 21474836480
459
+ "description": "Generate images guided by edge detection using SD 3.5 Canny ControlNet.",
460
+ "thumbnailVariant": "hoverDissolve",
461
+ "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sd3/#sd35-controlnets",
462
+ "tags": ["Image to Image", "Image", "ControlNet"],
463
+ "models": ["SD3.5", "Stability"],
464
+ "date": "2025-03-01",
465
+ "size": 23590107873
539
466
  },
540
467
  {
541
- "name": "flux_dev_checkpoint_example",
542
- "title": "Flux Dev fp8",
468
+ "name": "sd3.5_large_depth",
469
+ "title": "SD3.5 Large Depth",
543
470
  "mediaType": "image",
544
471
  "mediaSubtype": "webp",
545
- "description": "Generate images using Flux Dev fp8 quantized version. Suitable for devices with limited VRAM, requires only one model file, but image quality is slightly lower than the full version.",
546
- "tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-text-to-image",
547
- "tags": ["Text to Image", "Image"],
548
- "models": ["Flux", "BFL"],
472
+ "description": "Generate images guided by depth information using SD 3.5.",
473
+ "thumbnailVariant": "hoverDissolve",
474
+ "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sd3/#sd35-controlnets",
475
+ "tags": ["Image to Image", "Image", "ControlNet"],
476
+ "models": ["SD3.5", "Stability"],
549
477
  "date": "2025-03-01",
550
- "size": 17244293693,
551
- "vram": 18253611008
478
+ "size": 23590107873
552
479
  },
553
480
  {
554
- "name": "flux1_dev_uso_reference_image_gen",
555
- "title": "Flux.1 Dev USO Reference Image Generation",
556
- "description": "Use reference images to control both style and subject - keep your character's face while changing artistic style, or apply artistic styles to new scenes",
557
- "thumbnailVariant": "hoverDissolve",
481
+ "name": "sd3.5_large_blur",
482
+ "title": "SD3.5 Large Blur",
558
483
  "mediaType": "image",
559
484
  "mediaSubtype": "webp",
485
+ "description": "Generate images guided by blurred reference images using SD 3.5.",
486
+ "thumbnailVariant": "hoverDissolve",
487
+ "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sd3/#sd35-controlnets",
560
488
  "tags": ["Image to Image", "Image"],
561
- "models": ["Flux", "BFL"],
562
- "date": "2025-09-02",
563
- "tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-uso",
564
- "size": 18597208392,
565
- "vram": 19864223744
489
+ "models": ["SD3.5", "Stability"],
490
+ "date": "2025-03-01",
491
+ "size": 23590107873
566
492
  },
567
493
  {
568
- "name": "flux_schnell",
569
- "title": "Flux Schnell fp8",
494
+ "name": "sdxl_simple_example",
495
+ "title": "SDXL Simple",
570
496
  "mediaType": "image",
571
497
  "mediaSubtype": "webp",
572
- "description": "Quickly generate images with Flux Schnell fp8 quantized version. Ideal for low-end hardware, requires only 4 steps to generate images.",
573
- "tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-text-to-image",
498
+ "description": "Generate high-quality images using SDXL.",
499
+ "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sdxl/",
574
500
  "tags": ["Text to Image", "Image"],
575
- "models": ["Flux", "BFL"],
501
+ "models": ["SDXL", "Stability"],
576
502
  "date": "2025-03-01",
577
- "size": 17233556275,
578
- "vram": 18253611008
503
+ "size": 13013750907
579
504
  },
580
505
  {
581
- "name": "flux1_krea_dev",
582
- "title": "Flux.1 Krea Dev",
506
+ "name": "sdxl_refiner_prompt_example",
507
+ "title": "SDXL Refiner Prompt",
583
508
  "mediaType": "image",
584
509
  "mediaSubtype": "webp",
585
- "description": "A fine-tuned FLUX model pushing photorealism to the max",
586
- "tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux1-krea-dev",
510
+ "description": "Enhance SDXL images using refiner models.",
511
+ "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sdxl/",
587
512
  "tags": ["Text to Image", "Image"],
588
- "models": ["Flux", "BFL"],
589
- "date": "2025-07-31",
590
- "size": 22269405430,
591
- "vram": 23085449216
513
+ "models": ["SDXL", "Stability"],
514
+ "date": "2025-03-01",
515
+ "size": 13013750907
592
516
  },
593
517
  {
594
- "name": "flux_dev_full_text_to_image",
595
- "title": "Flux Dev full text to image",
518
+ "name": "sdxl_revision_text_prompts",
519
+ "title": "SDXL Revision Text Prompts",
596
520
  "mediaType": "image",
597
521
  "mediaSubtype": "webp",
598
- "description": "Generate high-quality images with Flux Dev full version. Requires larger VRAM and multiple model files, but provides the best prompt following capability and image quality.",
599
- "tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-text-to-image",
522
+ "description": "Generate images by transferring concepts from reference images using SDXL Revision.",
523
+ "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sdxl/#revision",
600
524
  "tags": ["Text to Image", "Image"],
601
- "models": ["Flux", "BFL"],
525
+ "models": ["SDXL", "Stability"],
602
526
  "date": "2025-03-01",
603
- "size": 34177202258,
604
- "vram": 23622320128
527
+ "size": 10630044058
605
528
  },
606
529
  {
607
- "name": "flux_schnell_full_text_to_image",
608
- "title": "Flux Schnell full text to image",
530
+ "name": "sdxlturbo_example",
531
+ "title": "SDXL Turbo",
609
532
  "mediaType": "image",
610
533
  "mediaSubtype": "webp",
611
- "description": "Generate images quickly with Flux Schnell full version. Uses Apache2.0 license, requires only 4 steps to generate images while maintaining good image quality.",
612
- "tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-text-to-image",
534
+ "description": "Generate images in a single step using SDXL Turbo.",
535
+ "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sdturbo/",
613
536
  "tags": ["Text to Image", "Image"],
614
- "models": ["Flux", "BFL"],
537
+ "models": ["SDXL", "Stability"],
615
538
  "date": "2025-03-01",
616
- "size": 34155727421
617
- },
618
- {
619
- "name": "flux_fill_inpaint_example",
620
- "title": "Flux Inpaint",
621
- "mediaType": "image",
622
- "mediaSubtype": "webp",
623
- "description": "Fill missing parts of images using Flux inpainting.",
624
- "thumbnailVariant": "compareSlider",
625
- "tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-fill-dev",
626
- "tags": ["Image to Image", "Inpainting", "Image"],
627
- "models": ["Flux", "BFL"],
628
- "date": "2025-03-01",
629
- "size": 10372346020
630
- },
631
- {
632
- "name": "flux_fill_outpaint_example",
633
- "title": "Flux Outpaint",
634
- "mediaType": "image",
635
- "mediaSubtype": "webp",
636
- "description": "Extend images beyond boundaries using Flux outpainting.",
637
- "thumbnailVariant": "compareSlider",
638
- "tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-fill-dev",
639
- "tags": ["Outpainting", "Image", "Image to Image"],
640
- "models": ["Flux", "BFL"],
641
- "date": "2025-03-01",
642
- "size": 10372346020
643
- },
644
- {
645
- "name": "flux_canny_model_example",
646
- "title": "Flux Canny Model",
647
- "mediaType": "image",
648
- "mediaSubtype": "webp",
649
- "description": "Generate images guided by edge detection using Flux Canny.",
650
- "thumbnailVariant": "hoverDissolve",
651
- "tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-controlnet",
652
- "tags": ["Image to Image", "ControlNet", "Image"],
653
- "models": ["Flux", "BFL"],
654
- "date": "2025-03-01",
655
- "size": 34177202258
656
- },
657
- {
658
- "name": "flux_depth_lora_example",
659
- "title": "Flux Depth Lora",
660
- "mediaType": "image",
661
- "mediaSubtype": "webp",
662
- "description": "Generate images guided by depth information using Flux LoRA.",
663
- "thumbnailVariant": "hoverDissolve",
664
- "tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-controlnet",
665
- "tags": ["Image to Image", "ControlNet", "Image"],
666
- "models": ["Flux", "BFL"],
667
- "date": "2025-03-01",
668
- "size": 35412005356
669
- },
670
- {
671
- "name": "flux_redux_model_example",
672
- "title": "Flux Redux Model",
673
- "mediaType": "image",
674
- "mediaSubtype": "webp",
675
- "description": "Generate images by transferring style from reference images using Flux Redux.",
676
- "tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-controlnet",
677
- "tags": ["Image to Image", "ControlNet", "Image"],
678
- "models": ["Flux", "BFL"],
679
- "date": "2025-03-01",
680
- "size": 35154307318
681
- },
682
- {
683
- "name": "image_omnigen2_t2i",
684
- "title": "OmniGen2 Text to Image",
685
- "mediaType": "image",
686
- "mediaSubtype": "webp",
687
- "description": "Generate high-quality images from text prompts using OmniGen2's unified 7B multimodal model with dual-path architecture.",
688
- "tutorialUrl": "https://docs.comfy.org/tutorials/image/omnigen/omnigen2",
689
- "tags": ["Text to Image", "Image"],
690
- "models": ["OmniGen"],
691
- "date": "2025-06-30",
692
- "size": 15784004813
693
- },
694
- {
695
- "name": "image_omnigen2_image_edit",
696
- "title": "OmniGen2 Image Edit",
697
- "mediaType": "image",
698
- "mediaSubtype": "webp",
699
- "thumbnailVariant": "hoverDissolve",
700
- "description": "Edit images with natural language instructions using OmniGen2's advanced image editing capabilities and text rendering support.",
701
- "tutorialUrl": "https://docs.comfy.org/tutorials/image/omnigen/omnigen2",
702
- "tags": ["Image Edit", "Image"],
703
- "models": ["OmniGen"],
704
- "date": "2025-06-30",
705
- "size": 15784004813
706
- },
707
- {
708
- "name": "hidream_i1_dev",
709
- "title": "HiDream I1 Dev",
710
- "mediaType": "image",
711
- "mediaSubtype": "webp",
712
- "description": "Generate images with HiDream I1 Dev - Balanced version with 28 inference steps, suitable for medium-range hardware.",
713
- "tutorialUrl": "https://docs.comfy.org/tutorials/image/hidream/hidream-i1",
714
- "tags": ["Text to Image", "Image"],
715
- "models": ["HiDream"],
716
- "date": "2025-04-17",
717
- "size": 33318208799
718
- },
719
- {
720
- "name": "hidream_i1_fast",
721
- "title": "HiDream I1 Fast",
722
- "mediaType": "image",
723
- "mediaSubtype": "webp",
724
- "description": "Generate images quickly with HiDream I1 Fast - Lightweight version with 16 inference steps, ideal for rapid previews on lower-end hardware.",
725
- "tutorialUrl": "https://docs.comfy.org/tutorials/image/hidream/hidream-i1",
726
- "tags": ["Text to Image", "Image"],
727
- "models": ["HiDream"],
728
- "date": "2025-04-17",
729
- "size": 24234352968
730
- },
731
- {
732
- "name": "hidream_i1_full",
733
- "title": "HiDream I1 Full",
734
- "mediaType": "image",
735
- "mediaSubtype": "webp",
736
- "description": "Generate images with HiDream I1 Full - Complete version with 50 inference steps for highest quality output.",
737
- "tutorialUrl": "https://docs.comfy.org/tutorials/image/hidream/hidream-i1",
738
- "tags": ["Text to Image", "Image"],
739
- "models": ["HiDream"],
740
- "date": "2025-04-17",
741
- "size": 24234352968
742
- },
743
- {
744
- "name": "hidream_e1_1",
745
- "title": "HiDream E1.1 Image Edit",
746
- "mediaType": "image",
747
- "mediaSubtype": "webp",
748
- "thumbnailVariant": "compareSlider",
749
- "description": "Edit images with HiDream E1.1 – it’s better in image quality and editing accuracy than HiDream-E1-Full.",
750
- "tutorialUrl": "https://docs.comfy.org/tutorials/image/hidream/hidream-e1",
751
- "tags": ["Image Edit", "Image"],
752
- "models": ["HiDream"],
753
- "date": "2025-07-21",
754
- "size": 50422916055
755
- },
756
- {
757
- "name": "hidream_e1_full",
758
- "title": "HiDream E1 Image Edit",
759
- "mediaType": "image",
760
- "mediaSubtype": "webp",
761
- "thumbnailVariant": "compareSlider",
762
- "description": "Edit images with HiDream E1 - Professional natural language image editing model.",
763
- "tutorialUrl": "https://docs.comfy.org/tutorials/image/hidream/hidream-e1",
764
- "tags": ["Image Edit", "Image"],
765
- "models": ["HiDream"],
766
- "date": "2025-05-01",
767
- "size": 34209414513
768
- },
769
- {
770
- "name": "sd3.5_simple_example",
771
- "title": "SD3.5 Simple",
772
- "mediaType": "image",
773
- "mediaSubtype": "webp",
774
- "description": "Generate images using SD 3.5.",
775
- "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sd3/#sd35",
776
- "tags": ["Text to Image", "Image"],
777
- "models": ["SD3.5", "Stability"],
778
- "date": "2025-03-01",
779
- "size": 14935748772
780
- },
781
- {
782
- "name": "sd3.5_large_canny_controlnet_example",
783
- "title": "SD3.5 Large Canny ControlNet",
784
- "mediaType": "image",
785
- "mediaSubtype": "webp",
786
- "description": "Generate images guided by edge detection using SD 3.5 Canny ControlNet.",
787
- "thumbnailVariant": "hoverDissolve",
788
- "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sd3/#sd35-controlnets",
789
- "tags": ["Image to Image", "Image", "ControlNet"],
790
- "models": ["SD3.5", "Stability"],
791
- "date": "2025-03-01",
792
- "size": 23590107873
793
- },
794
- {
795
- "name": "sd3.5_large_depth",
796
- "title": "SD3.5 Large Depth",
797
- "mediaType": "image",
798
- "mediaSubtype": "webp",
799
- "description": "Generate images guided by depth information using SD 3.5.",
800
- "thumbnailVariant": "hoverDissolve",
801
- "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sd3/#sd35-controlnets",
802
- "tags": ["Image to Image", "Image", "ControlNet"],
803
- "models": ["SD3.5", "Stability"],
804
- "date": "2025-03-01",
805
- "size": 23590107873
806
- },
807
- {
808
- "name": "sd3.5_large_blur",
809
- "title": "SD3.5 Large Blur",
810
- "mediaType": "image",
811
- "mediaSubtype": "webp",
812
- "description": "Generate images guided by blurred reference images using SD 3.5.",
813
- "thumbnailVariant": "hoverDissolve",
814
- "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sd3/#sd35-controlnets",
815
- "tags": ["Image to Image", "Image"],
816
- "models": ["SD3.5", "Stability"],
817
- "date": "2025-03-01",
818
- "size": 23590107873
819
- },
820
- {
821
- "name": "sdxl_simple_example",
822
- "title": "SDXL Simple",
823
- "mediaType": "image",
824
- "mediaSubtype": "webp",
825
- "description": "Generate high-quality images using SDXL.",
826
- "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sdxl/",
827
- "tags": ["Text to Image", "Image"],
828
- "models": ["SDXL", "Stability"],
829
- "date": "2025-03-01",
830
- "size": 13013750907
831
- },
832
- {
833
- "name": "sdxl_refiner_prompt_example",
834
- "title": "SDXL Refiner Prompt",
835
- "mediaType": "image",
836
- "mediaSubtype": "webp",
837
- "description": "Enhance SDXL images using refiner models.",
838
- "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sdxl/",
839
- "tags": ["Text to Image", "Image"],
840
- "models": ["SDXL", "Stability"],
841
- "date": "2025-03-01",
842
- "size": 13013750907
843
- },
844
- {
845
- "name": "sdxl_revision_text_prompts",
846
- "title": "SDXL Revision Text Prompts",
847
- "mediaType": "image",
848
- "mediaSubtype": "webp",
849
- "description": "Generate images by transferring concepts from reference images using SDXL Revision.",
850
- "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sdxl/#revision",
851
- "tags": ["Text to Image", "Image"],
852
- "models": ["SDXL", "Stability"],
853
- "date": "2025-03-01",
854
- "size": 10630044058
855
- },
856
- {
857
- "name": "sdxlturbo_example",
858
- "title": "SDXL Turbo",
859
- "mediaType": "image",
860
- "mediaSubtype": "webp",
861
- "description": "Generate images in a single step using SDXL Turbo.",
862
- "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sdturbo/",
863
- "tags": ["Text to Image", "Image"],
864
- "models": ["SDXL", "Stability"],
865
- "date": "2025-03-01",
866
- "size": 6936372183
539
+ "size": 6936372183
867
540
  },
868
541
  {
869
542
  "name": "image_lotus_depth_v1_1",
@@ -2523,5 +2196,332 @@
2523
2196
  "vram": 0
2524
2197
  }
2525
2198
  ]
2199
+ },
2200
+ {
2201
+ "moduleName": "default",
2202
+ "isEssential": true,
2203
+ "title": "Getting Started",
2204
+ "type": "image",
2205
+ "templates": [
2206
+ {
2207
+ "name": "01_qwen_t2i_subgraphed",
2208
+ "title": "Text to Image (New)",
2209
+ "mediaType": "image",
2210
+ "mediaSubtype": "webp",
2211
+ "description": "Generate images from text prompts using the Qwen-Image model.",
2212
+ "tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image",
2213
+ "tags": ["Text to Image", "Image"],
2214
+ "models": ["Qwen-Image"],
2215
+ "date": "2025-10-17",
2216
+ "size": 31772020572
2217
+ },
2218
+ {
2219
+ "name": "02_qwen_Image_edit_subgraphed",
2220
+ "title": "Image Editing (New)",
2221
+ "mediaType": "image",
2222
+ "mediaSubtype": "webp",
2223
+ "description": "Edit your images with Qwen-Image-Edit, the latest OSS model",
2224
+ "tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image-edit",
2225
+ "tags": ["Image to Image", "Image Edit", "ControlNet"],
2226
+ "models": ["Qwen-Image"],
2227
+ "date": "2025-10-17",
2228
+ "size": 31772020572
2229
+ },
2230
+ {
2231
+ "name": "03_video_wan2_2_14B_i2v_subgraphed",
2232
+ "title": "Image to Video (New)",
2233
+ "description": "Generate videos from an input image using Wan2.2 14B",
2234
+ "mediaType": "image",
2235
+ "mediaSubtype": "webp",
2236
+ "tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/wan2_2",
2237
+ "tags": ["Image to Video", "Video"],
2238
+ "models": ["Wan2.2", "Wan"],
2239
+ "date": "2025-10-17",
2240
+ "size": 38031935406
2241
+ },
2242
+ {
2243
+ "name": "04_hunyuan_3d_2.1_subgraphed",
2244
+ "title": "Image to 3D (New)",
2245
+ "mediaType": "image",
2246
+ "mediaSubtype": "webp",
2247
+ "description": "Generate 3D models from single images using Hunyuan3D 2.1.",
2248
+ "tags": ["Image to 3D", "3D"],
2249
+ "models": ["Hunyuan3D"],
2250
+ "date": "2025-10-17",
2251
+ "tutorialUrl": "https://docs.comfy.org/tutorials/3d/hunyuan3D-2",
2252
+ "size": 4928474972
2253
+ },
2254
+ {
2255
+ "name": "05_audio_ace_step_1_t2a_song_subgraphed",
2256
+ "title": "Text to Song (New)",
2257
+ "mediaType": "image",
2258
+ "mediaSubtype": "webp",
2259
+ "description": "Generate songs from text prompts using ACE-Step v1",
2260
+ "tags": ["Text to Audio", "Audio"],
2261
+ "models": ["ACE-Step"],
2262
+ "date": "2025-10-17",
2263
+ "tutorialUrl": "https://docs.comfy.org/tutorials/audio/ace-step/ace-step-v1",
2264
+ "size": 7698728878
2265
+ },
2266
+ {
2267
+ "name": "default",
2268
+ "title": "Image Generation",
2269
+ "mediaType": "image",
2270
+ "mediaSubtype": "webp",
2271
+ "description": "Generate images from text prompts.",
2272
+ "tutorialUrl": "https://docs.comfy.org/tutorials/basic/text-to-image",
2273
+ "tags": ["Text to Image", "Image"],
2274
+ "models": ["SD1.5", "Stability"],
2275
+ "date": "2025-03-01",
2276
+ "size": 2136746230,
2277
+ "vram": 3092376453
2278
+ },
2279
+ {
2280
+ "name": "image2image",
2281
+ "title": "Image to Image",
2282
+ "mediaType": "image",
2283
+ "mediaSubtype": "webp",
2284
+ "thumbnailVariant": "hoverDissolve",
2285
+ "description": "Transform existing images using text prompts.",
2286
+ "tutorialUrl": "https://docs.comfy.org/tutorials/basic/image-to-image",
2287
+ "tags": ["Image to Image", "Image"],
2288
+ "models": ["SD1.5", "Stability"],
2289
+ "date": "2025-03-01",
2290
+ "size": 2136746230,
2291
+ "vram": 3092376453
2292
+ },
2293
+ {
2294
+ "name": "lora",
2295
+ "title": "LoRA",
2296
+ "mediaType": "image",
2297
+ "mediaSubtype": "webp",
2298
+ "description": "Generate images with LoRA models for specialized styles or subjects.",
2299
+ "tutorialUrl": "https://docs.comfy.org/tutorials/basic/lora",
2300
+ "tags": ["Text to Image", "Image"],
2301
+ "models": ["SD1.5", "Stability"],
2302
+ "date": "2025-03-01",
2303
+ "size": 2437393940,
2304
+ "vram": 3092376453
2305
+ },
2306
+ {
2307
+ "name": "lora_multiple",
2308
+ "title": "LoRA Multiple",
2309
+ "mediaType": "image",
2310
+ "mediaSubtype": "webp",
2311
+ "description": "Generate images by combining multiple LoRA models.",
2312
+ "tutorialUrl": "https://docs.comfy.org/tutorials/basic/lora",
2313
+ "tags": ["Text to Image", "Image"],
2314
+ "models": ["SD1.5", "Stability"],
2315
+ "date": "2025-03-01",
2316
+ "size": 2437393940,
2317
+ "vram": 3350074491
2318
+ },
2319
+ {
2320
+ "name": "inpaint_example",
2321
+ "title": "Inpaint",
2322
+ "mediaType": "image",
2323
+ "mediaSubtype": "webp",
2324
+ "description": "Edit specific parts of images seamlessly.",
2325
+ "thumbnailVariant": "compareSlider",
2326
+ "tutorialUrl": "https://docs.comfy.org/tutorials/basic/inpaint",
2327
+ "tags": ["Inpainting", "Image"],
2328
+ "models": ["SD1.5", "Stability"],
2329
+ "date": "2025-03-01",
2330
+ "size": 5218385265,
2331
+ "vram": 4101693768
2332
+ },
2333
+ {
2334
+ "name": "inpaint_model_outpainting",
2335
+ "title": "Outpaint",
2336
+ "mediaType": "image",
2337
+ "mediaSubtype": "webp",
2338
+ "description": "Extend images beyond their original boundaries.",
2339
+ "thumbnailVariant": "compareSlider",
2340
+ "tutorialUrl": "https://docs.comfy.org/tutorials/basic/inpaint",
2341
+ "tags": ["Outpainting", "Image"],
2342
+ "models": ["SD1.5", "Stability"],
2343
+ "date": "2025-03-01",
2344
+ "size": 5218385265,
2345
+ "vram": 4101693768
2346
+ },
2347
+ {
2348
+ "name": "embedding_example",
2349
+ "title": "Embedding",
2350
+ "mediaType": "image",
2351
+ "mediaSubtype": "webp",
2352
+ "description": "Generate images using textual inversion for consistent styles.",
2353
+ "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/textual_inversion_embeddings/",
2354
+ "tags": ["Text to Image", "Image"],
2355
+ "models": ["SD1.5", "Stability"],
2356
+ "date": "2025-03-01",
2357
+ "size": 5218385265,
2358
+ "vram": 4123168604
2359
+ },
2360
+ {
2361
+ "name": "gligen_textbox_example",
2362
+ "title": "Gligen Textbox",
2363
+ "mediaType": "image",
2364
+ "mediaSubtype": "webp",
2365
+ "description": "Generate images with precise object placement using text boxes.",
2366
+ "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/gligen/",
2367
+ "tags": ["Image"],
2368
+ "models": ["SD1.5", "Stability"],
2369
+ "date": "2025-03-01",
2370
+ "size": 2974264852,
2371
+ "vram": 4080218931
2372
+ },
2373
+ {
2374
+ "name": "area_composition",
2375
+ "title": "Area Composition",
2376
+ "mediaType": "image",
2377
+ "mediaSubtype": "webp",
2378
+ "description": "Generate images by controlling composition with defined areas.",
2379
+ "tags": ["Text to Image", "Image"],
2380
+ "models": ["SD1.5", "Stability"],
2381
+ "date": "2025-03-01",
2382
+ "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/area_composition/",
2383
+ "size": 2469606195,
2384
+ "vram": 6184752906
2385
+ },
2386
+ {
2387
+ "name": "area_composition_square_area_for_subject",
2388
+ "title": "Area Composition Square Area for Subject",
2389
+ "mediaType": "image",
2390
+ "mediaSubtype": "webp",
2391
+ "description": "Generate images with consistent subject placement using area composition.",
2392
+ "tags": ["Text to Image", "Image"],
2393
+ "models": ["SD1.5", "Stability"],
2394
+ "date": "2025-03-01",
2395
+ "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/area_composition/#increasing-consistency-of-images-with-area-composition",
2396
+ "size": 2469606195,
2397
+ "vram": 5927054868
2398
+ },
2399
+ {
2400
+ "name": "hiresfix_latent_workflow",
2401
+ "title": "Upscale",
2402
+ "mediaType": "image",
2403
+ "mediaSubtype": "webp",
2404
+ "description": "Upscale images by enhancing quality in latent space.",
2405
+ "thumbnailVariant": "compareSlider",
2406
+ "tags": ["Upscale", "Image"],
2407
+ "models": ["SD1.5", "Stability"],
2408
+ "date": "2025-03-01",
2409
+ "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/2_pass_txt2img/",
2410
+ "size": 2136746230,
2411
+ "vram": 3929895076
2412
+ },
2413
+ {
2414
+ "name": "esrgan_example",
2415
+ "title": "ESRGAN",
2416
+ "mediaType": "image",
2417
+ "mediaSubtype": "webp",
2418
+ "description": "Upscale images using ESRGAN models to enhance quality.",
2419
+ "thumbnailVariant": "compareSlider",
2420
+ "tags": ["Upscale", "Image"],
2421
+ "models": ["SD1.5", "Stability"],
2422
+ "date": "2025-03-01",
2423
+ "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/upscale_models/",
2424
+ "size": 2201170739,
2425
+ "vram": 6442450944
2426
+ },
2427
+ {
2428
+ "name": "hiresfix_esrgan_workflow",
2429
+ "title": "HiresFix ESRGAN Workflow",
2430
+ "mediaType": "image",
2431
+ "mediaSubtype": "webp",
2432
+ "description": "Upscale images using ESRGAN models during intermediate generation steps.",
2433
+ "thumbnailVariant": "compareSlider",
2434
+ "tags": ["Upscale", "Image"],
2435
+ "models": ["SD1.5", "Stability"],
2436
+ "date": "2025-03-01",
2437
+ "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/2_pass_txt2img/#non-latent-upscaling",
2438
+ "size": 2201170739,
2439
+ "vram": 6442450944
2440
+ },
2441
+ {
2442
+ "name": "latent_upscale_different_prompt_model",
2443
+ "title": "Latent Upscale Different Prompt Model",
2444
+ "mediaType": "image",
2445
+ "mediaSubtype": "webp",
2446
+ "description": "Upscale images while changing prompts across generation passes.",
2447
+ "thumbnailVariant": "zoomHover",
2448
+ "tags": ["Upscale", "Image"],
2449
+ "models": ["SD1.5", "Stability"],
2450
+ "date": "2025-03-01",
2451
+ "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/2_pass_txt2img/#more-examples",
2452
+ "size": 4262755041,
2453
+ "vram": 5153960755
2454
+ },
2455
+ {
2456
+ "name": "controlnet_example",
2457
+ "title": "Scribble ControlNet",
2458
+ "mediaType": "image",
2459
+ "mediaSubtype": "webp",
2460
+ "description": "Generate images guided by scribble reference images using ControlNet.",
2461
+ "thumbnailVariant": "hoverDissolve",
2462
+ "tags": ["ControlNet", "Image"],
2463
+ "models": ["SD1.5", "Stability"],
2464
+ "date": "2025-03-01",
2465
+ "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/controlnet/",
2466
+ "size": 3189013217,
2467
+ "vram": 6442450944
2468
+ },
2469
+ {
2470
+ "name": "2_pass_pose_worship",
2471
+ "title": "Pose ControlNet 2 Pass",
2472
+ "mediaType": "image",
2473
+ "mediaSubtype": "webp",
2474
+ "description": "Generate images guided by pose references using ControlNet.",
2475
+ "thumbnailVariant": "hoverDissolve",
2476
+ "tags": ["ControlNet", "Image"],
2477
+ "models": ["SD1.5", "Stability"],
2478
+ "date": "2025-03-01",
2479
+ "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/controlnet/#pose-controlnet",
2480
+ "size": 4660039516,
2481
+ "vram": 6442450944
2482
+ },
2483
+ {
2484
+ "name": "depth_controlnet",
2485
+ "title": "Depth ControlNet",
2486
+ "mediaType": "image",
2487
+ "mediaSubtype": "webp",
2488
+ "description": "Generate images guided by depth information using ControlNet.",
2489
+ "thumbnailVariant": "hoverDissolve",
2490
+ "tags": ["ControlNet", "Image", "Text to Image"],
2491
+ "models": ["SD1.5", "Stability"],
2492
+ "date": "2025-03-01",
2493
+ "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/controlnet/#t2i-adapter-vs-controlnets",
2494
+ "size": 2888365507,
2495
+ "vram": 6442450944
2496
+ },
2497
+ {
2498
+ "name": "depth_t2i_adapter",
2499
+ "title": "Depth T2I Adapter",
2500
+ "mediaType": "image",
2501
+ "mediaSubtype": "webp",
2502
+ "description": "Generate images guided by depth information using T2I adapter.",
2503
+ "thumbnailVariant": "hoverDissolve",
2504
+ "tags": ["ControlNet", "Image", "Text to Image"],
2505
+ "models": ["SD1.5", "Stability"],
2506
+ "date": "2025-03-01",
2507
+ "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/controlnet/#t2i-adapter-vs-controlnets",
2508
+ "size": 2523293286,
2509
+ "vram": 6442450944
2510
+ },
2511
+ {
2512
+ "name": "mixing_controlnets",
2513
+ "title": "Mixing ControlNets",
2514
+ "mediaType": "image",
2515
+ "mediaSubtype": "webp",
2516
+ "description": "Generate images by combining multiple ControlNet models.",
2517
+ "thumbnailVariant": "hoverDissolve",
2518
+ "tags": ["ControlNet", "Image", "Text to Image"],
2519
+ "models": ["SD1.5", "Stability"],
2520
+ "date": "2025-03-01",
2521
+ "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/controlnet/#mixing-controlnets",
2522
+ "size": 3328599654,
2523
+ "vram": 6442450944
2524
+ }
2525
+ ]
2526
2526
  }
2527
2527
  ]