comfyui-workflow-templates 0.1.70__py3-none-any.whl → 0.1.71__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of comfyui-workflow-templates might be problematic. Click here for more details.
- comfyui_workflow_templates/templates/api_google_gemini_image.json +101 -96
- comfyui_workflow_templates/templates/flux1_dev_uso_reference_image_gen-1.webp +0 -0
- comfyui_workflow_templates/templates/flux1_dev_uso_reference_image_gen-2.webp +0 -0
- comfyui_workflow_templates/templates/flux1_dev_uso_reference_image_gen.json +3491 -0
- comfyui_workflow_templates/templates/hidream_i1_dev.json +455 -287
- comfyui_workflow_templates/templates/hidream_i1_fast.json +463 -295
- comfyui_workflow_templates/templates/hidream_i1_full.json +417 -250
- comfyui_workflow_templates/templates/image_qwen_image_instantx_controlnet.json +47 -47
- comfyui_workflow_templates/templates/index.es.json +261 -96
- comfyui_workflow_templates/templates/index.fr.json +300 -145
- comfyui_workflow_templates/templates/index.ja.json +261 -96
- comfyui_workflow_templates/templates/index.json +36 -2
- comfyui_workflow_templates/templates/index.ko.json +261 -96
- comfyui_workflow_templates/templates/index.ru.json +259 -158
- comfyui_workflow_templates/templates/index.zh-TW.json +261 -96
- comfyui_workflow_templates/templates/index.zh.json +261 -96
- comfyui_workflow_templates/templates/video_wan2_2_14B_s2v.json +80 -80
- comfyui_workflow_templates/templates/video_wan2_2_5B_fun_control-1.webp +0 -0
- comfyui_workflow_templates/templates/video_wan2_2_5B_fun_control.json +1104 -0
- comfyui_workflow_templates/templates/video_wan2_2_5B_fun_inpaint-1.webp +0 -0
- comfyui_workflow_templates/templates/video_wan2_2_5B_fun_inpaint.json +856 -0
- {comfyui_workflow_templates-0.1.70.dist-info → comfyui_workflow_templates-0.1.71.dist-info}/METADATA +1 -1
- {comfyui_workflow_templates-0.1.70.dist-info → comfyui_workflow_templates-0.1.71.dist-info}/RECORD +26 -19
- {comfyui_workflow_templates-0.1.70.dist-info → comfyui_workflow_templates-0.1.71.dist-info}/WHEEL +0 -0
- {comfyui_workflow_templates-0.1.70.dist-info → comfyui_workflow_templates-0.1.71.dist-info}/licenses/LICENSE +0 -0
- {comfyui_workflow_templates-0.1.70.dist-info → comfyui_workflow_templates-0.1.71.dist-info}/top_level.txt +0 -0
|
@@ -2,8 +2,8 @@
|
|
|
2
2
|
{
|
|
3
3
|
"moduleName": "default",
|
|
4
4
|
"category": "USE CASES",
|
|
5
|
-
"title": "Basics",
|
|
6
5
|
"type": "image",
|
|
6
|
+
"title": "Bases",
|
|
7
7
|
"templates": [
|
|
8
8
|
{
|
|
9
9
|
"name": "default",
|
|
@@ -14,7 +14,8 @@
|
|
|
14
14
|
"tutorialUrl": "https://docs.comfy.org/tutorials/basic/text-to-image",
|
|
15
15
|
"tags": ["Texte vers Image", "Image"],
|
|
16
16
|
"models": ["SD1.5"],
|
|
17
|
-
"date": "2025-03-01"
|
|
17
|
+
"date": "2025-03-01",
|
|
18
|
+
"size": 1.99
|
|
18
19
|
},
|
|
19
20
|
{
|
|
20
21
|
"name": "image2image",
|
|
@@ -25,7 +26,8 @@
|
|
|
25
26
|
"tutorialUrl": "https://docs.comfy.org/tutorials/basic/image-to-image",
|
|
26
27
|
"tags": ["Image vers Image", "Image"],
|
|
27
28
|
"models": ["SD1.5"],
|
|
28
|
-
"date": "2025-03-01"
|
|
29
|
+
"date": "2025-03-01",
|
|
30
|
+
"size": 1.99
|
|
29
31
|
},
|
|
30
32
|
{
|
|
31
33
|
"name": "lora",
|
|
@@ -59,7 +61,8 @@
|
|
|
59
61
|
"tutorialUrl": "https://docs.comfy.org/tutorials/basic/inpaint",
|
|
60
62
|
"tags": ["Inpainting", "Image"],
|
|
61
63
|
"models": ["SD1.5"],
|
|
62
|
-
"date": "2025-03-01"
|
|
64
|
+
"date": "2025-03-01",
|
|
65
|
+
"size": 4.86
|
|
63
66
|
},
|
|
64
67
|
{
|
|
65
68
|
"name": "inpaint_model_outpainting",
|
|
@@ -71,7 +74,8 @@
|
|
|
71
74
|
"tutorialUrl": "https://docs.comfy.org/tutorials/basic/inpaint",
|
|
72
75
|
"tags": ["Outpainting", "Image"],
|
|
73
76
|
"models": ["SD1.5"],
|
|
74
|
-
"date": "2025-03-01"
|
|
77
|
+
"date": "2025-03-01",
|
|
78
|
+
"size": 4.86
|
|
75
79
|
},
|
|
76
80
|
{
|
|
77
81
|
"name": "embedding_example",
|
|
@@ -82,7 +86,8 @@
|
|
|
82
86
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/textual_inversion_embeddings/",
|
|
83
87
|
"tags": ["Embedding", "Image"],
|
|
84
88
|
"models": ["SD1.5"],
|
|
85
|
-
"date": "2025-03-01"
|
|
89
|
+
"date": "2025-03-01",
|
|
90
|
+
"size": 4.86
|
|
86
91
|
},
|
|
87
92
|
{
|
|
88
93
|
"name": "gligen_textbox_example",
|
|
@@ -93,15 +98,16 @@
|
|
|
93
98
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/gligen/",
|
|
94
99
|
"tags": ["Gligen", "Image"],
|
|
95
100
|
"models": ["SD1.5"],
|
|
96
|
-
"date": "2025-03-01"
|
|
101
|
+
"date": "2025-03-01",
|
|
102
|
+
"size": 2.77
|
|
97
103
|
}
|
|
98
104
|
]
|
|
99
105
|
},
|
|
100
106
|
{
|
|
101
107
|
"moduleName": "default",
|
|
102
108
|
"category": "USE CASES",
|
|
103
|
-
"title": "Flux",
|
|
104
109
|
"type": "image",
|
|
110
|
+
"title": "Flux",
|
|
105
111
|
"templates": [
|
|
106
112
|
{
|
|
107
113
|
"name": "image_chroma_text_to_image",
|
|
@@ -111,7 +117,8 @@
|
|
|
111
117
|
"description": "Chroma est modifié à partir de Flux et présente quelques changements dans l'architecture.",
|
|
112
118
|
"tags": ["Texte vers Image", "Image"],
|
|
113
119
|
"models": ["Chroma", "Flux"],
|
|
114
|
-
"date": "2025-06-04"
|
|
120
|
+
"date": "2025-06-04",
|
|
121
|
+
"size": 21.69
|
|
115
122
|
},
|
|
116
123
|
{
|
|
117
124
|
"name": "flux_kontext_dev_basic",
|
|
@@ -123,7 +130,8 @@
|
|
|
123
130
|
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-kontext-dev",
|
|
124
131
|
"tags": ["Édition d'Image", "Image vers Image"],
|
|
125
132
|
"models": ["Flux"],
|
|
126
|
-
"date": "2025-06-26"
|
|
133
|
+
"date": "2025-06-26",
|
|
134
|
+
"size": 16.43
|
|
127
135
|
},
|
|
128
136
|
{
|
|
129
137
|
"name": "flux_dev_checkpoint_example",
|
|
@@ -134,7 +142,21 @@
|
|
|
134
142
|
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-text-to-image",
|
|
135
143
|
"tags": ["Texte vers Image", "Image"],
|
|
136
144
|
"models": ["Flux"],
|
|
137
|
-
"date": "2025-03-01"
|
|
145
|
+
"date": "2025-03-01",
|
|
146
|
+
"size": 16.06
|
|
147
|
+
},
|
|
148
|
+
{
|
|
149
|
+
"name": "flux1_dev_uso_reference_image_gen",
|
|
150
|
+
"title": "Génération d'images de référence Flux.1 Dev USO",
|
|
151
|
+
"description": "Utilisez des images de référence pour contrôler à la fois le style et le sujet : conservez le visage de votre personnage tout en changeant de style artistique, ou appliquez des styles artistiques à de nouvelles scènes",
|
|
152
|
+
"thumbnailVariant": "hoverDissolve",
|
|
153
|
+
"mediaType": "image",
|
|
154
|
+
"mediaSubtype": "webp",
|
|
155
|
+
"tags": ["Image to Image", "Image"],
|
|
156
|
+
"models": ["Flux"],
|
|
157
|
+
"date": "2025-09-02",
|
|
158
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-uso",
|
|
159
|
+
"size": 17.32
|
|
138
160
|
},
|
|
139
161
|
{
|
|
140
162
|
"name": "flux_schnell",
|
|
@@ -145,7 +167,8 @@
|
|
|
145
167
|
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-text-to-image",
|
|
146
168
|
"tags": ["Texte vers Image", "Image"],
|
|
147
169
|
"models": ["Flux"],
|
|
148
|
-
"date": "2025-03-01"
|
|
170
|
+
"date": "2025-03-01",
|
|
171
|
+
"size": 16.05
|
|
149
172
|
},
|
|
150
173
|
{
|
|
151
174
|
"name": "flux1_krea_dev",
|
|
@@ -155,8 +178,9 @@
|
|
|
155
178
|
"description": "Un modèle FLUX affiné poussant le photoréalisme à son maximum",
|
|
156
179
|
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux1-krea-dev",
|
|
157
180
|
"tags": ["Texte vers Image", "Image", "Photoréalisme"],
|
|
158
|
-
"models": ["Flux
|
|
159
|
-
"date": "2025-07-31"
|
|
181
|
+
"models": ["Flux"],
|
|
182
|
+
"date": "2025-07-31",
|
|
183
|
+
"size": 20.74
|
|
160
184
|
},
|
|
161
185
|
{
|
|
162
186
|
"name": "flux_dev_full_text_to_image",
|
|
@@ -167,7 +191,8 @@
|
|
|
167
191
|
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-text-to-image",
|
|
168
192
|
"tags": ["Texte vers Image", "Image"],
|
|
169
193
|
"models": ["Flux"],
|
|
170
|
-
"date": "2025-03-01"
|
|
194
|
+
"date": "2025-03-01",
|
|
195
|
+
"size": 31.83
|
|
171
196
|
},
|
|
172
197
|
{
|
|
173
198
|
"name": "flux_schnell_full_text_to_image",
|
|
@@ -178,7 +203,8 @@
|
|
|
178
203
|
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-text-to-image",
|
|
179
204
|
"tags": ["Texte vers Image", "Image"],
|
|
180
205
|
"models": ["Flux"],
|
|
181
|
-
"date": "2025-03-01"
|
|
206
|
+
"date": "2025-03-01",
|
|
207
|
+
"size": 31.81
|
|
182
208
|
},
|
|
183
209
|
{
|
|
184
210
|
"name": "flux_fill_inpaint_example",
|
|
@@ -190,7 +216,8 @@
|
|
|
190
216
|
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-fill-dev",
|
|
191
217
|
"tags": ["Image vers Image", "Inpainting", "Image"],
|
|
192
218
|
"models": ["Flux"],
|
|
193
|
-
"date": "2025-03-01"
|
|
219
|
+
"date": "2025-03-01",
|
|
220
|
+
"size": 9.66
|
|
194
221
|
},
|
|
195
222
|
{
|
|
196
223
|
"name": "flux_fill_outpaint_example",
|
|
@@ -202,7 +229,8 @@
|
|
|
202
229
|
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-fill-dev",
|
|
203
230
|
"tags": ["Outpainting", "Image", "Image vers Image"],
|
|
204
231
|
"models": ["Flux"],
|
|
205
|
-
"date": "2025-03-01"
|
|
232
|
+
"date": "2025-03-01",
|
|
233
|
+
"size": 9.66
|
|
206
234
|
},
|
|
207
235
|
{
|
|
208
236
|
"name": "flux_canny_model_example",
|
|
@@ -214,7 +242,8 @@
|
|
|
214
242
|
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-controlnet",
|
|
215
243
|
"tags": ["Image vers Image", "ControlNet", "Image"],
|
|
216
244
|
"models": ["Flux"],
|
|
217
|
-
"date": "2025-03-01"
|
|
245
|
+
"date": "2025-03-01",
|
|
246
|
+
"size": 31.83
|
|
218
247
|
},
|
|
219
248
|
{
|
|
220
249
|
"name": "flux_depth_lora_example",
|
|
@@ -226,7 +255,8 @@
|
|
|
226
255
|
"tutorialUrl": "ttps://docs.comfy.org/tutorials/flux/flux-1-controlnet",
|
|
227
256
|
"tags": ["Image vers Image", "ControlNet", "Image", "LoRA"],
|
|
228
257
|
"models": ["Flux"],
|
|
229
|
-
"date": "2025-03-01"
|
|
258
|
+
"date": "2025-03-01",
|
|
259
|
+
"size": 32.98
|
|
230
260
|
},
|
|
231
261
|
{
|
|
232
262
|
"name": "flux_redux_model_example",
|
|
@@ -237,15 +267,16 @@
|
|
|
237
267
|
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-controlnet",
|
|
238
268
|
"tags": ["Image vers Image", "ControlNet", "Image", "LoRA"],
|
|
239
269
|
"models": ["Flux"],
|
|
240
|
-
"date": "2025-03-01"
|
|
270
|
+
"date": "2025-03-01",
|
|
271
|
+
"size": 32.74
|
|
241
272
|
}
|
|
242
273
|
]
|
|
243
274
|
},
|
|
244
275
|
{
|
|
245
276
|
"moduleName": "default",
|
|
246
277
|
"category": "USE CASES",
|
|
247
|
-
"title": "Image",
|
|
248
278
|
"type": "image",
|
|
279
|
+
"title": "Image",
|
|
249
280
|
"templates": [
|
|
250
281
|
{
|
|
251
282
|
"name": "image_qwen_image",
|
|
@@ -256,7 +287,20 @@
|
|
|
256
287
|
"tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image",
|
|
257
288
|
"tags": ["Texte vers Image", "Image"],
|
|
258
289
|
"models": ["Qwen-Image"],
|
|
259
|
-
"date": "2025-08-05"
|
|
290
|
+
"date": "2025-08-05",
|
|
291
|
+
"size": 29.59
|
|
292
|
+
},
|
|
293
|
+
{
|
|
294
|
+
"name": "image_qwen_image_instantx_controlnet",
|
|
295
|
+
"title": "Qwen-Image InstantX ControlNet",
|
|
296
|
+
"mediaType": "image",
|
|
297
|
+
"mediaSubtype": "webp",
|
|
298
|
+
"description": "Générer des images avec Qwen-Image InstantX ControlNet, prenant en charge canny, contours doux, profondeur, pose",
|
|
299
|
+
"tags": ["Image vers Image", "Image", "ControlNet"],
|
|
300
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image",
|
|
301
|
+
"models": ["Qwen-Image"],
|
|
302
|
+
"date": "2025-08-23",
|
|
303
|
+
"size": 32.88
|
|
260
304
|
},
|
|
261
305
|
{
|
|
262
306
|
"name": "image_qwen_image_union_control_lora",
|
|
@@ -266,7 +310,9 @@
|
|
|
266
310
|
"description": "Générer des images avec un contrôle structurel précis en utilisant le LoRA ControlNet unifié de Qwen-Image. Prend en charge plusieurs types de contrôle incluant canny, depth, lineart, softedge, normal et openpose pour diverses applications créatives.",
|
|
267
311
|
"tags": ["Texte vers Image", "Image", "ControlNet"],
|
|
268
312
|
"models": ["Qwen-Image"],
|
|
269
|
-
"date": "2025-08-23"
|
|
313
|
+
"date": "2025-08-23",
|
|
314
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image",
|
|
315
|
+
"size": 30.47
|
|
270
316
|
},
|
|
271
317
|
{
|
|
272
318
|
"name": "image_qwen_image_controlnet_patch",
|
|
@@ -276,7 +322,10 @@
|
|
|
276
322
|
"description": "Contrôler la génération d'images en utilisant les modèles ControlNet de Qwen-Image. Prend en charge les contrôles canny, depth et inpainting via le patching de modèles.",
|
|
277
323
|
"tags": ["Texte vers Image", "Image", "ControlNet"],
|
|
278
324
|
"models": ["Qwen-Image"],
|
|
279
|
-
"date": "2025-08-24"
|
|
325
|
+
"date": "2025-08-24",
|
|
326
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image",
|
|
327
|
+
"size": 31.7,
|
|
328
|
+
"thumbnailVariant": "compareSlider"
|
|
280
329
|
},
|
|
281
330
|
{
|
|
282
331
|
"name": "image_qwen_image_edit",
|
|
@@ -288,7 +337,8 @@
|
|
|
288
337
|
"tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image-edit",
|
|
289
338
|
"tags": ["Image vers Image", "Édition d'Image"],
|
|
290
339
|
"models": ["Qwen-Image"],
|
|
291
|
-
"date": "2025-08-18"
|
|
340
|
+
"date": "2025-08-18",
|
|
341
|
+
"size": 29.59
|
|
292
342
|
},
|
|
293
343
|
{
|
|
294
344
|
"name": "image_omnigen2_t2i",
|
|
@@ -299,7 +349,8 @@
|
|
|
299
349
|
"tutorialUrl": "https://docs.comfy.org/tutorials/image/omnigen/omnigen2",
|
|
300
350
|
"tags": ["Texte vers Image", "Image"],
|
|
301
351
|
"models": ["OmniGen"],
|
|
302
|
-
"date": "2025-06-30"
|
|
352
|
+
"date": "2025-06-30",
|
|
353
|
+
"size": 14.7
|
|
303
354
|
},
|
|
304
355
|
{
|
|
305
356
|
"name": "image_omnigen2_image_edit",
|
|
@@ -311,7 +362,8 @@
|
|
|
311
362
|
"tutorialUrl": "https://docs.comfy.org/tutorials/image/omnigen/omnigen2",
|
|
312
363
|
"tags": ["Édition d'Image", "Image"],
|
|
313
364
|
"models": ["OmniGen"],
|
|
314
|
-
"date": "2025-06-30"
|
|
365
|
+
"date": "2025-06-30",
|
|
366
|
+
"size": 14.7
|
|
315
367
|
},
|
|
316
368
|
{
|
|
317
369
|
"name": "hidream_i1_dev",
|
|
@@ -322,7 +374,8 @@
|
|
|
322
374
|
"tutorialUrl": "https://docs.comfy.org/tutorials/image/hidream/hidream-i1",
|
|
323
375
|
"tags": ["Texte vers Image", "Image"],
|
|
324
376
|
"models": ["HiDream"],
|
|
325
|
-
"date": "2025-04-17"
|
|
377
|
+
"date": "2025-04-17",
|
|
378
|
+
"size": 31.03
|
|
326
379
|
},
|
|
327
380
|
{
|
|
328
381
|
"name": "hidream_i1_fast",
|
|
@@ -333,7 +386,8 @@
|
|
|
333
386
|
"tutorialUrl": "https://docs.comfy.org/tutorials/image/hidream/hidream-i1",
|
|
334
387
|
"tags": ["Texte vers Image", "Image"],
|
|
335
388
|
"models": ["HiDream"],
|
|
336
|
-
"date": "2025-04-17"
|
|
389
|
+
"date": "2025-04-17",
|
|
390
|
+
"size": 22.57
|
|
337
391
|
},
|
|
338
392
|
{
|
|
339
393
|
"name": "hidream_i1_full",
|
|
@@ -344,7 +398,8 @@
|
|
|
344
398
|
"tutorialUrl": "https://docs.comfy.org/tutorials/image/hidream/hidream-i1",
|
|
345
399
|
"tags": ["Texte vers Image", "Image"],
|
|
346
400
|
"models": ["HiDream"],
|
|
347
|
-
"date": "2025-04-17"
|
|
401
|
+
"date": "2025-04-17",
|
|
402
|
+
"size": 22.57
|
|
348
403
|
},
|
|
349
404
|
{
|
|
350
405
|
"name": "hidream_e1_1",
|
|
@@ -356,7 +411,8 @@
|
|
|
356
411
|
"tutorialUrl": "https://docs.comfy.org/tutorials/image/hidream/hidream-e1",
|
|
357
412
|
"tags": ["Édition d'Image", "Image"],
|
|
358
413
|
"models": ["HiDream"],
|
|
359
|
-
"date": "2025-07-21"
|
|
414
|
+
"date": "2025-07-21",
|
|
415
|
+
"size": 46.96
|
|
360
416
|
},
|
|
361
417
|
{
|
|
362
418
|
"name": "hidream_e1_full",
|
|
@@ -368,7 +424,8 @@
|
|
|
368
424
|
"tutorialUrl": "https://docs.comfy.org/tutorials/image/hidream/hidream-e1",
|
|
369
425
|
"tags": ["Édition d'Image", "Image"],
|
|
370
426
|
"models": ["HiDream"],
|
|
371
|
-
"date": "2025-05-01"
|
|
427
|
+
"date": "2025-05-01",
|
|
428
|
+
"size": 31.86
|
|
372
429
|
},
|
|
373
430
|
{
|
|
374
431
|
"name": "sd3.5_simple_example",
|
|
@@ -379,7 +436,8 @@
|
|
|
379
436
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sd3/#sd35",
|
|
380
437
|
"tags": ["Texte vers Image", "Image"],
|
|
381
438
|
"models": ["SD3.5"],
|
|
382
|
-
"date": "2025-03-01"
|
|
439
|
+
"date": "2025-03-01",
|
|
440
|
+
"size": 13.91
|
|
383
441
|
},
|
|
384
442
|
{
|
|
385
443
|
"name": "sd3.5_large_canny_controlnet_example",
|
|
@@ -391,7 +449,8 @@
|
|
|
391
449
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sd3/#sd35-controlnets",
|
|
392
450
|
"tags": ["Image vers Image", "Image", "ControlNet"],
|
|
393
451
|
"models": ["SD3.5"],
|
|
394
|
-
"date": "2025-03-01"
|
|
452
|
+
"date": "2025-03-01",
|
|
453
|
+
"size": 21.97
|
|
395
454
|
},
|
|
396
455
|
{
|
|
397
456
|
"name": "sd3.5_large_depth",
|
|
@@ -403,7 +462,8 @@
|
|
|
403
462
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sd3/#sd35-controlnets",
|
|
404
463
|
"tags": ["Image vers Image", "Image", "ControlNet"],
|
|
405
464
|
"models": ["SD3.5"],
|
|
406
|
-
"date": "2025-03-01"
|
|
465
|
+
"date": "2025-03-01",
|
|
466
|
+
"size": 21.97
|
|
407
467
|
},
|
|
408
468
|
{
|
|
409
469
|
"name": "sd3.5_large_blur",
|
|
@@ -415,7 +475,8 @@
|
|
|
415
475
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sd3/#sd35-controlnets",
|
|
416
476
|
"tags": ["Image vers Image", "Image"],
|
|
417
477
|
"models": ["SD3.5"],
|
|
418
|
-
"date": "2025-03-01"
|
|
478
|
+
"date": "2025-03-01",
|
|
479
|
+
"size": 21.97
|
|
419
480
|
},
|
|
420
481
|
{
|
|
421
482
|
"name": "sdxl_simple_example",
|
|
@@ -426,7 +487,8 @@
|
|
|
426
487
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sdxl/",
|
|
427
488
|
"tags": ["Texte vers Image", "Image"],
|
|
428
489
|
"models": ["SDXL"],
|
|
429
|
-
"date": "2025-03-01"
|
|
490
|
+
"date": "2025-03-01",
|
|
491
|
+
"size": 12.12
|
|
430
492
|
},
|
|
431
493
|
{
|
|
432
494
|
"name": "sdxl_refiner_prompt_example",
|
|
@@ -437,7 +499,8 @@
|
|
|
437
499
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sdxl/",
|
|
438
500
|
"tags": ["Texte vers Image", "Image"],
|
|
439
501
|
"models": ["SDXL"],
|
|
440
|
-
"date": "2025-03-01"
|
|
502
|
+
"date": "2025-03-01",
|
|
503
|
+
"size": 12.12
|
|
441
504
|
},
|
|
442
505
|
{
|
|
443
506
|
"name": "sdxl_revision_text_prompts",
|
|
@@ -448,7 +511,8 @@
|
|
|
448
511
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sdxl/#revision",
|
|
449
512
|
"tags": ["Texte vers Image", "Image"],
|
|
450
513
|
"models": ["SDXL"],
|
|
451
|
-
"date": "2025-03-01"
|
|
514
|
+
"date": "2025-03-01",
|
|
515
|
+
"size": 9.9
|
|
452
516
|
},
|
|
453
517
|
{
|
|
454
518
|
"name": "sdxl_revision_zero_positive",
|
|
@@ -459,7 +523,8 @@
|
|
|
459
523
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sdxl/#revision",
|
|
460
524
|
"tags": ["Texte vers Image", "Image"],
|
|
461
525
|
"models": ["SDXL"],
|
|
462
|
-
"date": "2025-03-01"
|
|
526
|
+
"date": "2025-03-01",
|
|
527
|
+
"size": 9.9
|
|
463
528
|
},
|
|
464
529
|
{
|
|
465
530
|
"name": "sdxlturbo_example",
|
|
@@ -470,7 +535,8 @@
|
|
|
470
535
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sdturbo/",
|
|
471
536
|
"tags": ["Texte vers Image", "Image"],
|
|
472
537
|
"models": ["SDXL Turbo"],
|
|
473
|
-
"date": "2025-03-01"
|
|
538
|
+
"date": "2025-03-01",
|
|
539
|
+
"size": 6.46
|
|
474
540
|
},
|
|
475
541
|
{
|
|
476
542
|
"name": "image_lotus_depth_v1_1",
|
|
@@ -481,379 +547,444 @@
|
|
|
481
547
|
"description": "Exécuter Lotus Depth dans ComfyUI pour une estimation de profondeur monoculaire efficace zero-shot avec une haute rétention de détails.",
|
|
482
548
|
"tags": ["Profondeur", "Image"],
|
|
483
549
|
"models": ["SD1.5"],
|
|
484
|
-
"date": "2025-05-21"
|
|
550
|
+
"date": "2025-05-21",
|
|
551
|
+
"size": 1.93
|
|
485
552
|
}
|
|
486
553
|
]
|
|
487
554
|
},
|
|
488
555
|
{
|
|
489
556
|
"moduleName": "default",
|
|
490
557
|
"category": "USE CASES",
|
|
491
|
-
"title": "Vidéo",
|
|
492
558
|
"type": "video",
|
|
559
|
+
"title": "Vidéo",
|
|
493
560
|
"templates": [
|
|
494
561
|
{
|
|
495
562
|
"name": "video_wan2_2_14B_t2v",
|
|
496
563
|
"title": "Wan 2.2 14B Text to Video",
|
|
497
|
-
"description": "
|
|
564
|
+
"description": "Générer des vidéos de haute qualité à partir de prompts textuels avec un contrôle esthétique cinématographique et une génération de mouvement dynamique en utilisant Wan 2.2.",
|
|
498
565
|
"mediaType": "image",
|
|
499
566
|
"mediaSubtype": "webp",
|
|
500
567
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/wan2_2",
|
|
501
568
|
"tags": ["Texte vers Vidéo", "Vidéo"],
|
|
502
569
|
"models": ["Wan"],
|
|
503
|
-
"date": "2025-07-29"
|
|
570
|
+
"date": "2025-07-29",
|
|
571
|
+
"size": 35.42
|
|
504
572
|
},
|
|
505
573
|
{
|
|
506
574
|
"name": "video_wan2_2_14B_i2v",
|
|
507
575
|
"title": "Wan 2.2 14B Image to Video",
|
|
508
|
-
"description": "
|
|
576
|
+
"description": "Transformer des images statiques en vidéos dynamiques avec un contrôle précis du mouvement et une préservation du style en utilisant Wan 2.2.",
|
|
509
577
|
"mediaType": "image",
|
|
510
578
|
"mediaSubtype": "webp",
|
|
511
579
|
"thumbnailVariant": "hoverDissolve",
|
|
512
580
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/wan2_2",
|
|
513
581
|
"tags": ["Image vers Vidéo", "Vidéo"],
|
|
514
582
|
"models": ["Wan2.2"],
|
|
515
|
-
"date": "2025-07-29"
|
|
583
|
+
"date": "2025-07-29",
|
|
584
|
+
"size": 35.42
|
|
516
585
|
},
|
|
517
586
|
{
|
|
518
587
|
"name": "video_wan2_2_14B_flf2v",
|
|
519
588
|
"title": "Wan 2.2 14B First-Last Frame to Video",
|
|
520
|
-
"description": "
|
|
589
|
+
"description": "Générer des transitions vidéo fluides en définissant les images de début et de fin.",
|
|
521
590
|
"mediaType": "image",
|
|
522
591
|
"mediaSubtype": "webp",
|
|
523
592
|
"thumbnailVariant": "hoverDissolve",
|
|
524
593
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/wan2_2",
|
|
525
594
|
"tags": ["FLF2V", "Vidéo"],
|
|
526
595
|
"models": ["Wan2.2"],
|
|
596
|
+
"date": "2025-08-02",
|
|
597
|
+
"size": 35.42
|
|
598
|
+
},
|
|
599
|
+
{
|
|
600
|
+
"name": "video_wan2_2_14B_s2v",
|
|
601
|
+
"title": "Wan2.2-S2V Génération de Vidéo Pilotée par l'Audio",
|
|
602
|
+
"description": "Transformer des images statiques et de l'audio en vidéos dynamiques avec une synchronisation parfaite et une génération au niveau de la minute.",
|
|
603
|
+
"mediaType": "image",
|
|
604
|
+
"mediaSubtype": "webp",
|
|
605
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/wan2-2-s2v",
|
|
606
|
+
"tags": ["Vidéo"],
|
|
607
|
+
"models": ["Wan2.2"],
|
|
527
608
|
"date": "2025-08-02"
|
|
528
609
|
},
|
|
529
610
|
{
|
|
530
611
|
"name": "video_wan2_2_14B_fun_inpaint",
|
|
531
612
|
"title": "Wan 2.2 14B Fun Inp",
|
|
532
|
-
"description": "
|
|
613
|
+
"description": "Générer des vidéos à partir des images de début et de fin en utilisant Wan 2.2 Fun Inp.",
|
|
533
614
|
"mediaType": "image",
|
|
534
615
|
"mediaSubtype": "webp",
|
|
535
616
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/wan2-2-fun-inp",
|
|
536
617
|
"tags": ["FLF2V", "Vidéo"],
|
|
537
618
|
"models": ["Wan2.2"],
|
|
538
|
-
"date": "2025-08-12"
|
|
619
|
+
"date": "2025-08-12",
|
|
620
|
+
"size": 35.42
|
|
539
621
|
},
|
|
540
622
|
{
|
|
541
623
|
"name": "video_wan2_2_14B_fun_control",
|
|
542
624
|
"title": "Wan 2.2 14B Fun Control",
|
|
543
|
-
"description": "
|
|
625
|
+
"description": "Générer des vidéos guidées par des contrôles de pose, de profondeur et de contours en utilisant Wan 2.2 Fun Control.",
|
|
544
626
|
"mediaType": "image",
|
|
545
627
|
"mediaSubtype": "webp",
|
|
546
628
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/wan2-2-fun-control",
|
|
547
629
|
"tags": ["Vidéo vers Vidéo", "Vidéo"],
|
|
548
630
|
"models": ["Wan2.2"],
|
|
549
|
-
"date": "2025-08-12"
|
|
631
|
+
"date": "2025-08-12",
|
|
632
|
+
"size": 35.42
|
|
550
633
|
},
|
|
551
634
|
{
|
|
552
635
|
"name": "video_wan2_2_14B_fun_camera",
|
|
553
636
|
"title": "Wan 2.2 14B Fun Camera Control",
|
|
554
|
-
"description": "
|
|
637
|
+
"description": "Générer des vidéos avec des contrôles de mouvement de caméra incluant le panoramique, le zoom et la rotation en utilisant Wan 2.2 Fun Camera Control.",
|
|
555
638
|
"mediaType": "image",
|
|
556
639
|
"mediaSubtype": "webp",
|
|
557
640
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/wan2-2-fun-camera",
|
|
558
641
|
"tags": ["Vidéo vers Vidéo", "Vidéo"],
|
|
559
642
|
"models": ["Wan2.2"],
|
|
560
|
-
"date": "2025-08-17"
|
|
643
|
+
"date": "2025-08-17",
|
|
644
|
+
"size": 37.3
|
|
561
645
|
},
|
|
562
646
|
{
|
|
563
647
|
"name": "video_wan2_2_5B_ti2v",
|
|
564
648
|
"title": "Wan 2.2 5B Video Generation",
|
|
565
|
-
"description": "
|
|
649
|
+
"description": "Générer des vidéos à partir de texte ou d'images en utilisant le modèle hybride Wan 2.2 5B",
|
|
566
650
|
"mediaType": "image",
|
|
567
651
|
"mediaSubtype": "webp",
|
|
568
652
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/wan2_2",
|
|
569
653
|
"tags": ["Texte vers Vidéo", "Vidéo"],
|
|
570
654
|
"models": ["Wan2.2"],
|
|
571
|
-
"date": "2025-07-29"
|
|
655
|
+
"date": "2025-07-29",
|
|
656
|
+
"size": 16.9
|
|
657
|
+
},
|
|
658
|
+
{
|
|
659
|
+
"name": "video_wan2_2_5B_fun_inpaint",
|
|
660
|
+
"title": "Wan 2.2 5B Fun Inpaint",
|
|
661
|
+
"description": "Inpainting vidéo efficace à partir des images de début et de fin. Le modèle 5B offre des itérations rapides pour tester les flux de travail.",
|
|
662
|
+
"mediaType": "image",
|
|
663
|
+
"mediaSubtype": "webp",
|
|
664
|
+
"tags": ["Texte vers Vidéo", "Vidéo"],
|
|
665
|
+
"models": ["Wan2.2"],
|
|
666
|
+
"date": "2025-07-29",
|
|
667
|
+
"size": 16.9
|
|
668
|
+
},
|
|
669
|
+
{
|
|
670
|
+
"name": "video_wan2_2_5B_fun_control",
|
|
671
|
+
"title": "Wan 2.2 5B Fun Control",
|
|
672
|
+
"description": "Contrôle vidéo multi-conditions avec guidance par pose, profondeur et contours. Taille compacte 5B pour un développement expérimental.",
|
|
673
|
+
"mediaType": "image",
|
|
674
|
+
"mediaSubtype": "webp",
|
|
675
|
+
"tags": ["Texte vers Vidéo", "Vidéo"],
|
|
676
|
+
"models": ["Wan2.2"],
|
|
677
|
+
"date": "2025-07-29",
|
|
678
|
+
"size": 16.9
|
|
572
679
|
},
|
|
573
680
|
{
|
|
574
681
|
"name": "video_wan_vace_14B_t2v",
|
|
575
682
|
"title": "Wan VACE Text to Video",
|
|
576
|
-
"description": "
|
|
683
|
+
"description": "Transformer des descriptions textuelles en vidéos de haute qualité. Prend en charge à la fois 480p et 720p avec le modèle VACE-14B.",
|
|
577
684
|
"mediaType": "image",
|
|
578
685
|
"mediaSubtype": "webp",
|
|
579
686
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/vace",
|
|
580
687
|
"tags": ["Texte vers Vidéo", "Vidéo"],
|
|
581
688
|
"models": ["Wan2.1"],
|
|
582
|
-
"date": "2025-05-21"
|
|
689
|
+
"date": "2025-05-21",
|
|
690
|
+
"size": 53.79
|
|
583
691
|
},
|
|
584
692
|
{
|
|
585
693
|
"name": "video_wan_vace_14B_ref2v",
|
|
586
694
|
"title": "Wan VACE Reference to Video",
|
|
587
|
-
"description": "
|
|
695
|
+
"description": "Créer des vidéos qui correspondent au style et au contenu d'une image de référence. Parfait pour la génération de vidéos cohérentes en style.",
|
|
588
696
|
"mediaType": "image",
|
|
589
697
|
"mediaSubtype": "webp",
|
|
590
698
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/vace",
|
|
591
699
|
"tags": ["Référence vers Vidéo", "Vidéo"],
|
|
592
700
|
"models": ["Wan2.1"],
|
|
593
|
-
"date": "2025-05-21"
|
|
701
|
+
"date": "2025-05-21",
|
|
702
|
+
"size": 53.79
|
|
594
703
|
},
|
|
595
704
|
{
|
|
596
705
|
"name": "video_wan_vace_14B_v2v",
|
|
597
706
|
"title": "Wan VACE Control Video",
|
|
598
|
-
"description": "
|
|
707
|
+
"description": "Générer des vidéos en contrôlant les vidéos d'entrée et les images de référence en utilisant Wan VACE.",
|
|
599
708
|
"mediaType": "image",
|
|
600
709
|
"mediaSubtype": "webp",
|
|
601
710
|
"thumbnailVariant": "compareSlider",
|
|
602
711
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/vace",
|
|
603
712
|
"tags": ["Vidéo vers Vidéo", "Vidéo"],
|
|
604
713
|
"models": ["Wan2.1"],
|
|
605
|
-
"date": "2025-05-21"
|
|
714
|
+
"date": "2025-05-21",
|
|
715
|
+
"size": 53.79
|
|
606
716
|
},
|
|
607
717
|
{
|
|
608
718
|
"name": "video_wan_vace_outpainting",
|
|
609
719
|
"title": "Wan VACE Outpainting",
|
|
610
|
-
"description": "
|
|
720
|
+
"description": "Générer des vidéos étendues en agrandissant la taille de la vidéo en utilisant l'outpainting Wan VACE.",
|
|
611
721
|
"mediaType": "image",
|
|
612
722
|
"mediaSubtype": "webp",
|
|
613
723
|
"thumbnailVariant": "compareSlider",
|
|
614
724
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/vace",
|
|
615
725
|
"tags": ["Outpainting", "Vidéo"],
|
|
616
726
|
"models": ["Wan2.1"],
|
|
617
|
-
"date": "2025-05-21"
|
|
727
|
+
"date": "2025-05-21",
|
|
728
|
+
"size": 53.79
|
|
618
729
|
},
|
|
619
730
|
{
|
|
620
731
|
"name": "video_wan_vace_flf2v",
|
|
621
732
|
"title": "Wan VACE First-Last Frame",
|
|
622
|
-
"description": "
|
|
733
|
+
"description": "Générer des transitions vidéo fluides en définissant les images de début et de fin. Prend en charge les séquences d'images clés personnalisées.",
|
|
623
734
|
"mediaType": "image",
|
|
624
735
|
"mediaSubtype": "webp",
|
|
625
736
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/vace",
|
|
626
737
|
"tags": ["FLF2V", "Vidéo"],
|
|
627
738
|
"models": ["Wan2.1"],
|
|
628
|
-
"date": "2025-05-21"
|
|
739
|
+
"date": "2025-05-21",
|
|
740
|
+
"size": 53.79
|
|
629
741
|
},
|
|
630
742
|
{
|
|
631
743
|
"name": "video_wan_vace_inpainting",
|
|
632
744
|
"title": "Wan VACE Inpainting",
|
|
633
|
-
"description": "
|
|
745
|
+
"description": "Éditer des régions spécifiques dans les vidéos tout en préservant le contenu environnant. Idéal pour la suppression ou le remplacement d'objets.",
|
|
634
746
|
"mediaType": "image",
|
|
635
747
|
"mediaSubtype": "webp",
|
|
636
748
|
"thumbnailVariant": "compareSlider",
|
|
637
749
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/vace",
|
|
638
750
|
"tags": ["Inpainting", "Vidéo"],
|
|
639
751
|
"models": ["Wan2.1"],
|
|
640
|
-
"date": "2025-05-21"
|
|
752
|
+
"date": "2025-05-21",
|
|
753
|
+
"size": 53.79
|
|
641
754
|
},
|
|
642
755
|
{
|
|
643
756
|
"name": "video_wan_ati",
|
|
644
757
|
"title": "Wan ATI",
|
|
645
|
-
"description": "
|
|
758
|
+
"description": "Génération de vidéo contrôlée par trajectoire.",
|
|
646
759
|
"mediaType": "image",
|
|
647
760
|
"mediaSubtype": "webp",
|
|
648
761
|
"thumbnailVariant": "hoverDissolve",
|
|
649
762
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/wan-ati",
|
|
650
763
|
"tags": ["Vidéo"],
|
|
651
764
|
"models": ["Wan2.1"],
|
|
652
|
-
"date": "2025-05-21"
|
|
765
|
+
"date": "2025-05-21",
|
|
766
|
+
"size": 23.65
|
|
653
767
|
},
|
|
654
768
|
{
|
|
655
769
|
"name": "video_wan2.1_fun_camera_v1.1_1.3B",
|
|
656
770
|
"title": "Wan 2.1 Fun Camera 1.3B",
|
|
657
|
-
"description": "
|
|
771
|
+
"description": "Générer des vidéos dynamiques avec des mouvements de caméra cinématographiques en utilisant le modèle Wan 2.1 Fun Camera 1.3B.",
|
|
658
772
|
"mediaType": "image",
|
|
659
773
|
"mediaSubtype": "webp",
|
|
660
774
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/fun-control",
|
|
661
775
|
"tags": ["Vidéo"],
|
|
662
776
|
"models": ["Wan2.1"],
|
|
663
|
-
"date": "2025-04-15"
|
|
777
|
+
"date": "2025-04-15",
|
|
778
|
+
"size": 10.7
|
|
664
779
|
},
|
|
665
780
|
{
|
|
666
781
|
"name": "video_wan2.1_fun_camera_v1.1_14B",
|
|
667
782
|
"title": "Wan 2.1 Fun Camera 14B",
|
|
668
|
-
"description": "
|
|
783
|
+
"description": "Générer des vidéos de haute qualité avec un contrôle avancé de la caméra en utilisant le modèle 14B complet",
|
|
669
784
|
"mediaType": "image",
|
|
670
785
|
"mediaSubtype": "webp",
|
|
671
786
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/fun-control",
|
|
672
787
|
"tags": ["Vidéo"],
|
|
673
788
|
"models": ["Wan2.1"],
|
|
674
|
-
"date": "2025-04-15"
|
|
789
|
+
"date": "2025-04-15",
|
|
790
|
+
"size": 39.16
|
|
675
791
|
},
|
|
676
792
|
{
|
|
677
793
|
"name": "text_to_video_wan",
|
|
678
794
|
"title": "Wan 2.1 Text to Video",
|
|
679
|
-
"description": "
|
|
795
|
+
"description": "Générer des vidéos à partir de prompts textuels en utilisant Wan 2.1.",
|
|
680
796
|
"mediaType": "image",
|
|
681
797
|
"mediaSubtype": "webp",
|
|
682
798
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/wan-video",
|
|
683
799
|
"tags": ["Texte vers Vidéo", "Vidéo"],
|
|
684
800
|
"models": ["Wan2.1"],
|
|
685
|
-
"date": "2025-03-01"
|
|
801
|
+
"date": "2025-03-01",
|
|
802
|
+
"size": 9.15
|
|
686
803
|
},
|
|
687
804
|
{
|
|
688
805
|
"name": "image_to_video_wan",
|
|
689
806
|
"title": "Wan 2.1 Image to Video",
|
|
690
|
-
"description": "
|
|
807
|
+
"description": "Générer des vidéos à partir d'images en utilisant Wan 2.1.",
|
|
691
808
|
"mediaType": "image",
|
|
692
809
|
"mediaSubtype": "webp",
|
|
693
810
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/wan-video",
|
|
694
811
|
"tags": ["Texte vers Vidéo", "Vidéo"],
|
|
695
812
|
"models": ["Wan2.1"],
|
|
696
|
-
"date": "2025-03-01"
|
|
813
|
+
"date": "2025-03-01",
|
|
814
|
+
"size": 38.23
|
|
697
815
|
},
|
|
698
816
|
{
|
|
699
817
|
"name": "wan2.1_fun_inp",
|
|
700
818
|
"title": "Wan 2.1 Inpainting",
|
|
701
|
-
"description": "
|
|
819
|
+
"description": "Générer des vidéos à partir des images de début et de fin en utilisant l'inpainting Wan 2.1.",
|
|
702
820
|
"mediaType": "image",
|
|
703
821
|
"mediaSubtype": "webp",
|
|
704
822
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/fun-inp",
|
|
705
823
|
"tags": ["Inpainting", "Vidéo"],
|
|
706
824
|
"models": ["Wan2.1"],
|
|
707
|
-
"date": "2025-04-15"
|
|
825
|
+
"date": "2025-04-15",
|
|
826
|
+
"size": 10.6
|
|
708
827
|
},
|
|
709
828
|
{
|
|
710
829
|
"name": "wan2.1_fun_control",
|
|
711
830
|
"title": "Wan 2.1 ControlNet",
|
|
712
|
-
"description": "
|
|
831
|
+
"description": "Générer des vidéos guidées par des contrôles de pose, de profondeur et de contours en utilisant Wan 2.1 ControlNet.",
|
|
713
832
|
"mediaType": "image",
|
|
714
833
|
"mediaSubtype": "webp",
|
|
715
834
|
"thumbnailVariant": "hoverDissolve",
|
|
716
835
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/fun-control",
|
|
717
836
|
"tags": ["Vidéo vers Vidéo", "Vidéo"],
|
|
718
837
|
"models": ["Wan2.1"],
|
|
719
|
-
"date": "2025-04-15"
|
|
838
|
+
"date": "2025-04-15",
|
|
839
|
+
"size": 10.6
|
|
720
840
|
},
|
|
721
841
|
{
|
|
722
842
|
"name": "wan2.1_flf2v_720_f16",
|
|
723
843
|
"title": "Wan 2.1 FLF2V 720p F16",
|
|
724
|
-
"description": "
|
|
844
|
+
"description": "Générer des vidéos en contrôlant les première et dernière images en utilisant Wan 2.1 FLF2V.",
|
|
725
845
|
"mediaType": "image",
|
|
726
846
|
"mediaSubtype": "webp",
|
|
727
847
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/wan-flf",
|
|
728
848
|
"tags": ["FLF2V", "Vidéo"],
|
|
729
849
|
"models": ["Wan2.1"],
|
|
730
|
-
"date": "2025-04-15"
|
|
850
|
+
"date": "2025-04-15",
|
|
851
|
+
"size": 38.23
|
|
731
852
|
},
|
|
732
853
|
{
|
|
733
854
|
"name": "ltxv_text_to_video",
|
|
734
855
|
"title": "LTXV Text to Video",
|
|
735
856
|
"mediaType": "image",
|
|
736
857
|
"mediaSubtype": "webp",
|
|
737
|
-
"description": "
|
|
858
|
+
"description": "Générer des vidéos à partir de prompts textuels.",
|
|
738
859
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/ltxv",
|
|
739
860
|
"tags": ["Texte vers Vidéo", "Vidéo"],
|
|
740
861
|
"models": ["LTXV"],
|
|
741
|
-
"date": "2025-03-01"
|
|
862
|
+
"date": "2025-03-01",
|
|
863
|
+
"size": 17.84
|
|
742
864
|
},
|
|
743
865
|
{
|
|
744
866
|
"name": "ltxv_image_to_video",
|
|
745
867
|
"title": "LTXV Image to Video",
|
|
746
868
|
"mediaType": "image",
|
|
747
869
|
"mediaSubtype": "webp",
|
|
748
|
-
"description": "
|
|
870
|
+
"description": "Générer des vidéos à partir d'images fixes.",
|
|
749
871
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/ltxv",
|
|
750
872
|
"tags": ["Image vers Vidéo", "Vidéo"],
|
|
751
873
|
"models": ["LTXV"],
|
|
752
|
-
"date": "2025-03-01"
|
|
874
|
+
"date": "2025-03-01",
|
|
875
|
+
"size": 17.84
|
|
753
876
|
},
|
|
754
877
|
{
|
|
755
878
|
"name": "mochi_text_to_video_example",
|
|
756
879
|
"title": "Mochi Text to Video",
|
|
757
880
|
"mediaType": "image",
|
|
758
881
|
"mediaSubtype": "webp",
|
|
759
|
-
"description": "
|
|
882
|
+
"description": "Générer des vidéos à partir de prompts textuels en utilisant le modèle Mochi.",
|
|
760
883
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/mochi/",
|
|
761
884
|
"tags": ["Texte vers Vidéo", "Vidéo"],
|
|
762
885
|
"models": ["Mochi"],
|
|
763
|
-
"date": "2025-03-01"
|
|
886
|
+
"date": "2025-03-01",
|
|
887
|
+
"size": 28.65
|
|
764
888
|
},
|
|
765
889
|
{
|
|
766
890
|
"name": "hunyuan_video_text_to_video",
|
|
767
891
|
"title": "Hunyuan Video Text to Video",
|
|
768
892
|
"mediaType": "image",
|
|
769
893
|
"mediaSubtype": "webp",
|
|
770
|
-
"description": "
|
|
894
|
+
"description": "Générer des vidéos à partir de prompts textuels en utilisant le modèle Hunyuan.",
|
|
771
895
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/hunyuan_video/",
|
|
772
896
|
"tags": ["Texte vers Vidéo", "Vidéo"],
|
|
773
897
|
"models": ["Hunyuan Video"],
|
|
774
|
-
"date": "2025-03-01"
|
|
898
|
+
"date": "2025-03-01",
|
|
899
|
+
"size": 33.04
|
|
775
900
|
},
|
|
776
901
|
{
|
|
777
902
|
"name": "image_to_video",
|
|
778
903
|
"title": "SVD Image to Video",
|
|
779
904
|
"mediaType": "image",
|
|
780
905
|
"mediaSubtype": "webp",
|
|
781
|
-
"description": "
|
|
906
|
+
"description": "Générer des vidéos à partir d'images fixes.",
|
|
782
907
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/video/#image-to-video",
|
|
783
908
|
"tags": ["Image vers Vidéo", "Vidéo"],
|
|
784
909
|
"models": ["SVD"],
|
|
785
|
-
"date": "2025-03-01"
|
|
910
|
+
"date": "2025-03-01",
|
|
911
|
+
"size": 8.9
|
|
786
912
|
},
|
|
787
913
|
{
|
|
788
914
|
"name": "txt_to_image_to_video",
|
|
789
915
|
"title": "SVD Text to Image to Video",
|
|
790
916
|
"mediaType": "image",
|
|
791
917
|
"mediaSubtype": "webp",
|
|
792
|
-
"description": "
|
|
918
|
+
"description": "Générer des vidéos en créant d'abord des images à partir de prompts textuels.",
|
|
793
919
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/video/#image-to-video",
|
|
794
920
|
"tags": ["Texte vers Vidéo", "Vidéo"],
|
|
795
921
|
"models": ["SVD"],
|
|
796
|
-
"date": "2025-03-01"
|
|
922
|
+
"date": "2025-03-01",
|
|
923
|
+
"size": 15.36
|
|
797
924
|
}
|
|
798
925
|
]
|
|
799
926
|
},
|
|
800
927
|
{
|
|
801
928
|
"moduleName": "default",
|
|
802
929
|
"category": "USE CASES",
|
|
803
|
-
"title": "Audio",
|
|
804
930
|
"type": "audio",
|
|
931
|
+
"title": "Audio",
|
|
805
932
|
"templates": [
|
|
806
933
|
{
|
|
807
934
|
"name": "audio_stable_audio_example",
|
|
808
935
|
"title": "Stable Audio",
|
|
809
936
|
"mediaType": "audio",
|
|
810
937
|
"mediaSubtype": "mp3",
|
|
811
|
-
"description": "
|
|
938
|
+
"description": "Générer de l'audio à partir de prompts textuels en utilisant Stable Audio.",
|
|
812
939
|
"tags": ["Texte vers Audio", "Audio"],
|
|
813
940
|
"models": ["Stable Audio"],
|
|
814
941
|
"date": "2025-03-01",
|
|
815
|
-
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/audio/"
|
|
942
|
+
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/audio/",
|
|
943
|
+
"size": 5.35
|
|
816
944
|
},
|
|
817
945
|
{
|
|
818
946
|
"name": "audio_ace_step_1_t2a_instrumentals",
|
|
819
947
|
"title": "ACE-Step v1 Text to Instrumentals Music",
|
|
820
948
|
"mediaType": "audio",
|
|
821
949
|
"mediaSubtype": "mp3",
|
|
822
|
-
"description": "
|
|
950
|
+
"description": "Générer de la musique instrumentale à partir de prompts textuels en utilisant ACE-Step v1.",
|
|
823
951
|
"tags": ["Texte vers Audio", "Audio", "Instrumentals"],
|
|
824
952
|
"models": ["ACE-Step v1"],
|
|
825
953
|
"date": "2025-03-01",
|
|
826
|
-
"tutorialUrl": "https://docs.comfy.org/tutorials/audio/ace-step/ace-step-v1"
|
|
954
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/audio/ace-step/ace-step-v1",
|
|
955
|
+
"size": 7.17
|
|
827
956
|
},
|
|
828
957
|
{
|
|
829
958
|
"name": "audio_ace_step_1_t2a_song",
|
|
830
959
|
"title": "ACE Step v1 Text to Song",
|
|
831
960
|
"mediaType": "audio",
|
|
832
961
|
"mediaSubtype": "mp3",
|
|
833
|
-
"description": "
|
|
962
|
+
"description": "Générer des chansons avec des voix à partir de prompts textuels en utilisant ACE-Step v1, prenant en charge la multilingue et la personnalisation du style.",
|
|
834
963
|
"tags": ["Texte vers Audio", "Audio", "Song"],
|
|
835
964
|
"models": ["ACE-Step v1"],
|
|
836
965
|
"date": "2025-03-01",
|
|
837
|
-
"tutorialUrl": "https://docs.comfy.org/tutorials/audio/ace-step/ace-step-v1"
|
|
966
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/audio/ace-step/ace-step-v1",
|
|
967
|
+
"size": 7.17
|
|
838
968
|
},
|
|
839
969
|
{
|
|
840
970
|
"name": "audio_ace_step_1_m2m_editing",
|
|
841
971
|
"title": "ACE Step v1 M2M Editing",
|
|
842
972
|
"mediaType": "audio",
|
|
843
973
|
"mediaSubtype": "mp3",
|
|
844
|
-
"description": "
|
|
974
|
+
"description": "Éditer des chansons existantes pour changer le style et les paroles en utilisant ACE-Step v1 M2M.",
|
|
845
975
|
"tags": ["Édition Audio", "Audio"],
|
|
846
976
|
"models": ["ACE-Step v1"],
|
|
847
977
|
"date": "2025-03-01",
|
|
848
|
-
"tutorialUrl": "https://docs.comfy.org/tutorials/audio/ace-step/ace-step-v1"
|
|
978
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/audio/ace-step/ace-step-v1",
|
|
979
|
+
"size": 7.17
|
|
849
980
|
}
|
|
850
981
|
]
|
|
851
982
|
},
|
|
852
983
|
{
|
|
853
984
|
"moduleName": "default",
|
|
854
985
|
"category": "USE CASES",
|
|
855
|
-
"title": "3D",
|
|
856
986
|
"type": "3d",
|
|
987
|
+
"title": "3D",
|
|
857
988
|
"templates": [
|
|
858
989
|
{
|
|
859
990
|
"name": "3d_hunyuan3d_image_to_model",
|
|
@@ -864,7 +995,8 @@
|
|
|
864
995
|
"tags": ["Image vers Modèle", "3D"],
|
|
865
996
|
"models": ["Hunyuan3D 2.0"],
|
|
866
997
|
"date": "2025-03-01",
|
|
867
|
-
"tutorialUrl": ""
|
|
998
|
+
"tutorialUrl": "",
|
|
999
|
+
"size": 4.59
|
|
868
1000
|
},
|
|
869
1001
|
{
|
|
870
1002
|
"name": "3d_hunyuan3d_multiview_to_model",
|
|
@@ -876,7 +1008,8 @@
|
|
|
876
1008
|
"models": ["Hunyuan3D 2.0 MV"],
|
|
877
1009
|
"date": "2025-03-01",
|
|
878
1010
|
"tutorialUrl": "",
|
|
879
|
-
"thumbnailVariant": "hoverDissolve"
|
|
1011
|
+
"thumbnailVariant": "hoverDissolve",
|
|
1012
|
+
"size": 4.59
|
|
880
1013
|
},
|
|
881
1014
|
{
|
|
882
1015
|
"name": "3d_hunyuan3d_multiview_to_model_turbo",
|
|
@@ -888,7 +1021,8 @@
|
|
|
888
1021
|
"models": ["Hunyuan3D 2.0 MV Turbo"],
|
|
889
1022
|
"date": "2025-03-01",
|
|
890
1023
|
"tutorialUrl": "",
|
|
891
|
-
"thumbnailVariant": "hoverDissolve"
|
|
1024
|
+
"thumbnailVariant": "hoverDissolve",
|
|
1025
|
+
"size": 4.59
|
|
892
1026
|
},
|
|
893
1027
|
{
|
|
894
1028
|
"name": "stable_zero123_example",
|
|
@@ -899,20 +1033,31 @@
|
|
|
899
1033
|
"tags": ["Image vers 3D", "3D"],
|
|
900
1034
|
"models": ["Stable Zero123"],
|
|
901
1035
|
"date": "2025-03-01",
|
|
902
|
-
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/3d/"
|
|
1036
|
+
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/3d/",
|
|
1037
|
+
"size": 7.99
|
|
903
1038
|
}
|
|
904
1039
|
]
|
|
905
1040
|
},
|
|
906
1041
|
{
|
|
907
1042
|
"moduleName": "default",
|
|
908
1043
|
"category": "TOOLS & BUILDING",
|
|
909
|
-
"title": "Image API",
|
|
910
1044
|
"type": "image",
|
|
1045
|
+
"title": "API Image",
|
|
911
1046
|
"templates": [
|
|
1047
|
+
{
|
|
1048
|
+
"name": "api_google_gemini_image",
|
|
1049
|
+
"title": "Google Gemini Image",
|
|
1050
|
+
"description": "Nano-banana (Gemini-2.5-Flash Image) - édition d'images avec cohérence.",
|
|
1051
|
+
"mediaType": "image",
|
|
1052
|
+
"mediaSubtype": "webp",
|
|
1053
|
+
"tags": ["Édition d'Image", "Image", "API", "Texte-vers-Image"],
|
|
1054
|
+
"models": ["Gemini-2.5-Flash", "nano-banana"],
|
|
1055
|
+
"date": "2025-08-27"
|
|
1056
|
+
},
|
|
912
1057
|
{
|
|
913
1058
|
"name": "api_bfl_flux_1_kontext_multiple_images_input",
|
|
914
1059
|
"title": "BFL Flux.1 Kontext Multiple Image Input",
|
|
915
|
-
"description": "
|
|
1060
|
+
"description": "Importer plusieurs images et les éditer avec Flux.1 Kontext.",
|
|
916
1061
|
"mediaType": "image",
|
|
917
1062
|
"mediaSubtype": "webp",
|
|
918
1063
|
"thumbnailVariant": "compareSlider",
|
|
@@ -924,7 +1069,7 @@
|
|
|
924
1069
|
{
|
|
925
1070
|
"name": "api_bfl_flux_1_kontext_pro_image",
|
|
926
1071
|
"title": "BFL Flux.1 Kontext Pro",
|
|
927
|
-
"description": "
|
|
1072
|
+
"description": "Éditer des images avec Flux.1 Kontext pro image.",
|
|
928
1073
|
"mediaType": "image",
|
|
929
1074
|
"mediaSubtype": "webp",
|
|
930
1075
|
"thumbnailVariant": "compareSlider",
|
|
@@ -936,7 +1081,7 @@
|
|
|
936
1081
|
{
|
|
937
1082
|
"name": "api_bfl_flux_1_kontext_max_image",
|
|
938
1083
|
"title": "BFL Flux.1 Kontext Max",
|
|
939
|
-
"description": "
|
|
1084
|
+
"description": "Éditer des images avec Flux.1 Kontext max image.",
|
|
940
1085
|
"mediaType": "image",
|
|
941
1086
|
"mediaSubtype": "webp",
|
|
942
1087
|
"thumbnailVariant": "compareSlider",
|
|
@@ -948,7 +1093,7 @@
|
|
|
948
1093
|
{
|
|
949
1094
|
"name": "api_bfl_flux_pro_t2i",
|
|
950
1095
|
"title": "BFL Flux[Pro]: Text to Image",
|
|
951
|
-
"description": "
|
|
1096
|
+
"description": "Générer des images avec un excellent suivi des prompts et une qualité visuelle en utilisant FLUX.1 Pro.",
|
|
952
1097
|
"mediaType": "image",
|
|
953
1098
|
"mediaSubtype": "webp",
|
|
954
1099
|
"tutorialUrl": "https://docs.comfy.org/tutorials/api-nodes/black-forest-labs/flux-1-1-pro-ultra-image",
|
|
@@ -1011,7 +1156,7 @@
|
|
|
1011
1156
|
{
|
|
1012
1157
|
"name": "api_runway_text_to_image",
|
|
1013
1158
|
"title": "Runway: Text to Image",
|
|
1014
|
-
"description": "
|
|
1159
|
+
"description": "Générer des images de haute qualité à partir de prompts textuels en utilisant le modèle AI de Runway.",
|
|
1015
1160
|
"mediaType": "image",
|
|
1016
1161
|
"mediaSubtype": "webp",
|
|
1017
1162
|
"tags": ["Texte vers Image", "Image", "API"],
|
|
@@ -1021,7 +1166,7 @@
|
|
|
1021
1166
|
{
|
|
1022
1167
|
"name": "api_runway_reference_to_image",
|
|
1023
1168
|
"title": "Runway: Reference to Image",
|
|
1024
|
-
"description": "
|
|
1169
|
+
"description": "Générer de nouvelles images basées sur des styles et compositions de référence avec l'AI de Runway.",
|
|
1025
1170
|
"mediaType": "image",
|
|
1026
1171
|
"thumbnailVariant": "compareSlider",
|
|
1027
1172
|
"mediaSubtype": "webp",
|
|
@@ -1032,7 +1177,7 @@
|
|
|
1032
1177
|
{
|
|
1033
1178
|
"name": "api_stability_ai_stable_image_ultra_t2i",
|
|
1034
1179
|
"title": "Stability AI: Stable Image Ultra Text to Image",
|
|
1035
|
-
"description": "
|
|
1180
|
+
"description": "Générer des images de haute qualité avec un excellent respect des prompts. Parfait pour des cas d'utilisation professionnels à une résolution de 1 mégapixel.",
|
|
1036
1181
|
"mediaType": "image",
|
|
1037
1182
|
"mediaSubtype": "webp",
|
|
1038
1183
|
"tags": ["Texte vers Image", "Image", "API"],
|
|
@@ -1042,7 +1187,7 @@
|
|
|
1042
1187
|
{
|
|
1043
1188
|
"name": "api_stability_ai_i2i",
|
|
1044
1189
|
"title": "Stability AI: Image to Image",
|
|
1045
|
-
"description": "
|
|
1190
|
+
"description": "Transformer des images avec une génération de haute qualité en utilisant Stability AI, parfait pour l'édition professionnelle et le transfert de style.",
|
|
1046
1191
|
"mediaType": "image",
|
|
1047
1192
|
"thumbnailVariant": "compareSlider",
|
|
1048
1193
|
"mediaSubtype": "webp",
|
|
@@ -1053,7 +1198,7 @@
|
|
|
1053
1198
|
{
|
|
1054
1199
|
"name": "api_stability_ai_sd3.5_t2i",
|
|
1055
1200
|
"title": "Stability AI: SD3.5 Text to Image",
|
|
1056
|
-
"description": "
|
|
1201
|
+
"description": "Générer des images de haute qualité avec un excellent respect des prompts. Parfait pour des cas d'utilisation professionnels à une résolution de 1 mégapixel.",
|
|
1057
1202
|
"mediaType": "image",
|
|
1058
1203
|
"mediaSubtype": "webp",
|
|
1059
1204
|
"tags": ["Texte vers Image", "Image", "API"],
|
|
@@ -1063,7 +1208,7 @@
|
|
|
1063
1208
|
{
|
|
1064
1209
|
"name": "api_stability_ai_sd3.5_i2i",
|
|
1065
1210
|
"title": "Stability AI: SD3.5 Image to Image",
|
|
1066
|
-
"description": "
|
|
1211
|
+
"description": "Générer des images de haute qualité avec un excellent respect des prompts. Parfait pour des cas d'utilisation professionnels à une résolution de 1 mégapixel.",
|
|
1067
1212
|
"mediaType": "image",
|
|
1068
1213
|
"thumbnailVariant": "compareSlider",
|
|
1069
1214
|
"mediaSubtype": "webp",
|
|
@@ -1107,7 +1252,7 @@
|
|
|
1107
1252
|
{
|
|
1108
1253
|
"name": "api_openai_image_1_inpaint",
|
|
1109
1254
|
"title": "OpenAI: GPT-Image-1 Inpaint",
|
|
1110
|
-
"description": "
|
|
1255
|
+
"description": "Éditer des images en utilisant l'inpainting avec l'API OpenAI GPT Image 1.",
|
|
1111
1256
|
"mediaType": "image",
|
|
1112
1257
|
"mediaSubtype": "webp",
|
|
1113
1258
|
"thumbnailVariant": "compareSlider",
|
|
@@ -1119,7 +1264,7 @@
|
|
|
1119
1264
|
{
|
|
1120
1265
|
"name": "api_openai_image_1_multi_inputs",
|
|
1121
1266
|
"title": "OpenAI: GPT-Image-1 Multi Inputs",
|
|
1122
|
-
"description": "
|
|
1267
|
+
"description": "Générer des images à partir de plusieurs entrées en utilisant l'API OpenAI GPT Image 1.",
|
|
1123
1268
|
"mediaType": "image",
|
|
1124
1269
|
"mediaSubtype": "webp",
|
|
1125
1270
|
"thumbnailVariant": "compareSlider",
|
|
@@ -1131,7 +1276,7 @@
|
|
|
1131
1276
|
{
|
|
1132
1277
|
"name": "api_openai_dall_e_2_t2i",
|
|
1133
1278
|
"title": "OpenAI: Dall-E 2 Text to Image",
|
|
1134
|
-
"description": "
|
|
1279
|
+
"description": "Générer des images à partir de prompts textuels en utilisant l'API OpenAI Dall-E 2.",
|
|
1135
1280
|
"mediaType": "image",
|
|
1136
1281
|
"mediaSubtype": "webp",
|
|
1137
1282
|
"tags": ["Texte vers Image", "Image", "API"],
|
|
@@ -1142,7 +1287,7 @@
|
|
|
1142
1287
|
{
|
|
1143
1288
|
"name": "api_openai_dall_e_2_inpaint",
|
|
1144
1289
|
"title": "OpenAI: Dall-E 2 Inpaint",
|
|
1145
|
-
"description": "
|
|
1290
|
+
"description": "Éditer des images en utilisant l'inpainting avec l'API OpenAI Dall-E 2.",
|
|
1146
1291
|
"mediaType": "image",
|
|
1147
1292
|
"mediaSubtype": "webp",
|
|
1148
1293
|
"thumbnailVariant": "compareSlider",
|
|
@@ -1167,8 +1312,8 @@
|
|
|
1167
1312
|
{
|
|
1168
1313
|
"moduleName": "default",
|
|
1169
1314
|
"category": "TOOLS & BUILDING",
|
|
1170
|
-
"title": "Video API",
|
|
1171
1315
|
"type": "video",
|
|
1316
|
+
"title": "API Vidéo",
|
|
1172
1317
|
"templates": [
|
|
1173
1318
|
{
|
|
1174
1319
|
"name": "api_kling_i2v",
|
|
@@ -1206,7 +1351,7 @@
|
|
|
1206
1351
|
{
|
|
1207
1352
|
"name": "api_vidu_text_to_video",
|
|
1208
1353
|
"title": "Vidu: Text to Video",
|
|
1209
|
-
"description": "
|
|
1354
|
+
"description": "Générer des vidéos 1080p de haute qualité à partir de prompts textuels avec un contrôle ajustable de l'amplitude des mouvements et de la durée en utilisant le modèle AI avancé de Vidu.",
|
|
1210
1355
|
"mediaType": "image",
|
|
1211
1356
|
"mediaSubtype": "webp",
|
|
1212
1357
|
"tags": ["Texte vers Vidéo", "Vidéo", "API"],
|
|
@@ -1217,7 +1362,7 @@
|
|
|
1217
1362
|
{
|
|
1218
1363
|
"name": "api_vidu_image_to_video",
|
|
1219
1364
|
"title": "Vidu: Image to Video",
|
|
1220
|
-
"description": "
|
|
1365
|
+
"description": "Transformer des images statiques en vidéos 1080p dynamiques avec un contrôle précis du mouvement et une amplitude de mouvement personnalisable en utilisant Vidu.",
|
|
1221
1366
|
"mediaType": "image",
|
|
1222
1367
|
"mediaSubtype": "webp",
|
|
1223
1368
|
"tags": ["Image vers Vidéo", "Vidéo", "API"],
|
|
@@ -1463,8 +1608,8 @@
|
|
|
1463
1608
|
{
|
|
1464
1609
|
"moduleName": "default",
|
|
1465
1610
|
"category": "TOOLS & BUILDING",
|
|
1466
|
-
"title": "3D API",
|
|
1467
1611
|
"type": "image",
|
|
1612
|
+
"title": "API 3D",
|
|
1468
1613
|
"templates": [
|
|
1469
1614
|
{
|
|
1470
1615
|
"name": "api_rodin_image_to_model",
|
|
@@ -1530,8 +1675,8 @@
|
|
|
1530
1675
|
{
|
|
1531
1676
|
"moduleName": "default",
|
|
1532
1677
|
"category": "TOOLS & BUILDING",
|
|
1533
|
-
"title": "LLM API",
|
|
1534
1678
|
"type": "image",
|
|
1679
|
+
"title": "API LLM",
|
|
1535
1680
|
"templates": [
|
|
1536
1681
|
{
|
|
1537
1682
|
"name": "api_openai_chat",
|
|
@@ -1560,8 +1705,8 @@
|
|
|
1560
1705
|
{
|
|
1561
1706
|
"moduleName": "default",
|
|
1562
1707
|
"category": "TOOLS & BUILDING",
|
|
1563
|
-
"title": "Upscaling",
|
|
1564
1708
|
"type": "image",
|
|
1709
|
+
"title": "Agrandissement",
|
|
1565
1710
|
"templates": [
|
|
1566
1711
|
{
|
|
1567
1712
|
"name": "hiresfix_latent_workflow",
|
|
@@ -1573,7 +1718,8 @@
|
|
|
1573
1718
|
"tags": ["Agrandissement", "Image"],
|
|
1574
1719
|
"models": ["SD1.5"],
|
|
1575
1720
|
"date": "2025-03-01",
|
|
1576
|
-
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/2_pass_txt2img/"
|
|
1721
|
+
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/2_pass_txt2img/",
|
|
1722
|
+
"size": 1.99
|
|
1577
1723
|
},
|
|
1578
1724
|
{
|
|
1579
1725
|
"name": "esrgan_example",
|
|
@@ -1585,7 +1731,8 @@
|
|
|
1585
1731
|
"tags": ["Agrandissement", "Image"],
|
|
1586
1732
|
"models": ["SD1.5"],
|
|
1587
1733
|
"date": "2025-03-01",
|
|
1588
|
-
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/upscale_models/"
|
|
1734
|
+
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/upscale_models/",
|
|
1735
|
+
"size": 2.05
|
|
1589
1736
|
},
|
|
1590
1737
|
{
|
|
1591
1738
|
"name": "hiresfix_esrgan_workflow",
|
|
@@ -1597,7 +1744,8 @@
|
|
|
1597
1744
|
"tags": ["Agrandissement", "Image"],
|
|
1598
1745
|
"models": ["SD1.5"],
|
|
1599
1746
|
"date": "2025-03-01",
|
|
1600
|
-
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/2_pass_txt2img/#non-latent-upscaling"
|
|
1747
|
+
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/2_pass_txt2img/#non-latent-upscaling",
|
|
1748
|
+
"size": 2.05
|
|
1601
1749
|
},
|
|
1602
1750
|
{
|
|
1603
1751
|
"name": "latent_upscale_different_prompt_model",
|
|
@@ -1609,15 +1757,16 @@
|
|
|
1609
1757
|
"tags": ["Agrandissement", "Image"],
|
|
1610
1758
|
"models": ["SD1.5"],
|
|
1611
1759
|
"date": "2025-03-01",
|
|
1612
|
-
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/2_pass_txt2img/#more-examples"
|
|
1760
|
+
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/2_pass_txt2img/#more-examples",
|
|
1761
|
+
"size": 3.97
|
|
1613
1762
|
}
|
|
1614
1763
|
]
|
|
1615
1764
|
},
|
|
1616
1765
|
{
|
|
1617
1766
|
"moduleName": "default",
|
|
1618
1767
|
"category": "TOOLS & BUILDING",
|
|
1619
|
-
"title": "ControlNet",
|
|
1620
1768
|
"type": "image",
|
|
1769
|
+
"title": "ControlNet",
|
|
1621
1770
|
"templates": [
|
|
1622
1771
|
{
|
|
1623
1772
|
"name": "controlnet_example",
|
|
@@ -1629,7 +1778,8 @@
|
|
|
1629
1778
|
"tags": ["ControlNet", "Image"],
|
|
1630
1779
|
"models": ["SD1.5"],
|
|
1631
1780
|
"date": "2025-03-01",
|
|
1632
|
-
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/controlnet/"
|
|
1781
|
+
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/controlnet/",
|
|
1782
|
+
"size": 2.97
|
|
1633
1783
|
},
|
|
1634
1784
|
{
|
|
1635
1785
|
"name": "2_pass_pose_worship",
|
|
@@ -1641,7 +1791,8 @@
|
|
|
1641
1791
|
"tags": ["ControlNet", "Image"],
|
|
1642
1792
|
"models": ["SD1.5"],
|
|
1643
1793
|
"date": "2025-03-01",
|
|
1644
|
-
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/controlnet/#pose-controlnet"
|
|
1794
|
+
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/controlnet/#pose-controlnet",
|
|
1795
|
+
"size": 0.98
|
|
1645
1796
|
},
|
|
1646
1797
|
{
|
|
1647
1798
|
"name": "depth_controlnet",
|
|
@@ -1653,7 +1804,8 @@
|
|
|
1653
1804
|
"tags": ["ControlNet", "Image"],
|
|
1654
1805
|
"models": ["SD1.5"],
|
|
1655
1806
|
"date": "2025-03-01",
|
|
1656
|
-
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/controlnet/#t2i-adapter-vs-controlnets"
|
|
1807
|
+
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/controlnet/#t2i-adapter-vs-controlnets",
|
|
1808
|
+
"size": 0.67
|
|
1657
1809
|
},
|
|
1658
1810
|
{
|
|
1659
1811
|
"name": "depth_t2i_adapter",
|
|
@@ -1677,15 +1829,16 @@
|
|
|
1677
1829
|
"tags": ["ControlNet", "Image"],
|
|
1678
1830
|
"models": ["SD1.5"],
|
|
1679
1831
|
"date": "2025-03-01",
|
|
1680
|
-
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/controlnet/#mixing-controlnets"
|
|
1832
|
+
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/controlnet/#mixing-controlnets",
|
|
1833
|
+
"size": 1.66
|
|
1681
1834
|
}
|
|
1682
1835
|
]
|
|
1683
1836
|
},
|
|
1684
1837
|
{
|
|
1685
1838
|
"moduleName": "default",
|
|
1686
1839
|
"category": "TOOLS & BUILDING",
|
|
1687
|
-
"title": "Composition de Zone",
|
|
1688
1840
|
"type": "image",
|
|
1841
|
+
"title": "Composition de Zone",
|
|
1689
1842
|
"templates": [
|
|
1690
1843
|
{
|
|
1691
1844
|
"name": "area_composition",
|
|
@@ -1696,7 +1849,8 @@
|
|
|
1696
1849
|
"tags": ["Composition de Zone", "Image"],
|
|
1697
1850
|
"models": ["SD1.5"],
|
|
1698
1851
|
"date": "2025-03-01",
|
|
1699
|
-
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/area_composition/"
|
|
1852
|
+
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/area_composition/",
|
|
1853
|
+
"size": 2.3
|
|
1700
1854
|
},
|
|
1701
1855
|
{
|
|
1702
1856
|
"name": "area_composition_square_area_for_subject",
|
|
@@ -1707,8 +1861,9 @@
|
|
|
1707
1861
|
"tags": ["Composition de Zone", "Image"],
|
|
1708
1862
|
"models": ["SD1.5"],
|
|
1709
1863
|
"date": "2025-03-01",
|
|
1710
|
-
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/area_composition/#increasing-consistency-of-images-with-area-composition"
|
|
1864
|
+
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/area_composition/#increasing-consistency-of-images-with-area-composition",
|
|
1865
|
+
"size": 2.3
|
|
1711
1866
|
}
|
|
1712
1867
|
]
|
|
1713
1868
|
}
|
|
1714
|
-
]
|
|
1869
|
+
]
|