comfyui-workflow-templates 0.1.64__py3-none-any.whl → 0.1.66__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of comfyui-workflow-templates might be problematic. Click here for more details.
- comfyui_workflow_templates/templates/image_chroma_text_to_image.json +229 -277
- comfyui_workflow_templates/templates/image_qwen_image.json +16 -16
- comfyui_workflow_templates/templates/image_qwen_image_controlnet_patch-1.webp +0 -0
- comfyui_workflow_templates/templates/image_qwen_image_controlnet_patch-2.webp +0 -0
- comfyui_workflow_templates/templates/image_qwen_image_controlnet_patch.json +1053 -0
- comfyui_workflow_templates/templates/image_qwen_image_instantx_controlnet-1.webp +0 -0
- comfyui_workflow_templates/templates/image_qwen_image_instantx_controlnet-2.webp +0 -0
- comfyui_workflow_templates/templates/image_qwen_image_instantx_controlnet.json +2048 -0
- comfyui_workflow_templates/templates/index.es.json +696 -1381
- comfyui_workflow_templates/templates/index.fr.json +676 -1319
- comfyui_workflow_templates/templates/index.ja.json +684 -1337
- comfyui_workflow_templates/templates/index.json +25 -35
- comfyui_workflow_templates/templates/index.ko.json +969 -1564
- comfyui_workflow_templates/templates/index.ru.json +623 -1265
- comfyui_workflow_templates/templates/index.zh-TW.json +694 -1347
- comfyui_workflow_templates/templates/index.zh.json +700 -1353
- {comfyui_workflow_templates-0.1.64.dist-info → comfyui_workflow_templates-0.1.66.dist-info}/METADATA +1 -1
- {comfyui_workflow_templates-0.1.64.dist-info → comfyui_workflow_templates-0.1.66.dist-info}/RECORD +24 -18
- /comfyui_workflow_templates/templates/{image_qwen_image_union_control-1.webp → image_qwen_image_union_control_lora-1.webp} +0 -0
- /comfyui_workflow_templates/templates/{image_qwen_image_union_control-3.webp → image_qwen_image_union_control_lora-3.webp} +0 -0
- /comfyui_workflow_templates/templates/{image_qwen_image_union_control.json → image_qwen_image_union_control_lora.json} +0 -0
- {comfyui_workflow_templates-0.1.64.dist-info → comfyui_workflow_templates-0.1.66.dist-info}/WHEEL +0 -0
- {comfyui_workflow_templates-0.1.64.dist-info → comfyui_workflow_templates-0.1.66.dist-info}/licenses/LICENSE +0 -0
- {comfyui_workflow_templates-0.1.64.dist-info → comfyui_workflow_templates-0.1.66.dist-info}/top_level.txt +0 -0
|
@@ -2,39 +2,29 @@
|
|
|
2
2
|
{
|
|
3
3
|
"moduleName": "default",
|
|
4
4
|
"category": "USE CASES",
|
|
5
|
-
"title": "
|
|
5
|
+
"title": "Conceptos básicos",
|
|
6
6
|
"type": "image",
|
|
7
7
|
"templates": [
|
|
8
8
|
{
|
|
9
9
|
"name": "default",
|
|
10
|
-
"title": "Generación de
|
|
10
|
+
"title": "Generación de imágenes",
|
|
11
11
|
"mediaType": "image",
|
|
12
12
|
"mediaSubtype": "webp",
|
|
13
|
-
"description": "
|
|
13
|
+
"description": "Generar imágenes a partir de indicaciones de texto.",
|
|
14
14
|
"tutorialUrl": "https://docs.comfy.org/tutorials/basic/text-to-image",
|
|
15
|
-
"tags": [
|
|
16
|
-
|
|
17
|
-
"Image"
|
|
18
|
-
],
|
|
19
|
-
"models": [
|
|
20
|
-
"SD1.5"
|
|
21
|
-
],
|
|
15
|
+
"tags": ["Texto a imagen", "Imagen"],
|
|
16
|
+
"models": ["SD1.5"],
|
|
22
17
|
"date": "2025-03-01"
|
|
23
18
|
},
|
|
24
19
|
{
|
|
25
20
|
"name": "image2image",
|
|
26
|
-
"title": "Imagen a
|
|
21
|
+
"title": "Imagen a imagen",
|
|
27
22
|
"mediaType": "image",
|
|
28
23
|
"mediaSubtype": "webp",
|
|
29
|
-
"description": "
|
|
24
|
+
"description": "Transformar imágenes existentes usando indicaciones de texto.",
|
|
30
25
|
"tutorialUrl": "https://docs.comfy.org/tutorials/basic/image-to-image",
|
|
31
|
-
"tags": [
|
|
32
|
-
|
|
33
|
-
"Image"
|
|
34
|
-
],
|
|
35
|
-
"models": [
|
|
36
|
-
"SD1.5"
|
|
37
|
-
],
|
|
26
|
+
"tags": ["Imagen a imagen", "Imagen"],
|
|
27
|
+
"models": ["SD1.5"],
|
|
38
28
|
"date": "2025-03-01"
|
|
39
29
|
},
|
|
40
30
|
{
|
|
@@ -42,32 +32,21 @@
|
|
|
42
32
|
"title": "LoRA",
|
|
43
33
|
"mediaType": "image",
|
|
44
34
|
"mediaSubtype": "webp",
|
|
45
|
-
"description": "
|
|
35
|
+
"description": "Generar imágenes con modelos LoRA para estilos o temas especializados.",
|
|
46
36
|
"tutorialUrl": "https://docs.comfy.org/tutorials/basic/lora",
|
|
47
|
-
"tags": [
|
|
48
|
-
|
|
49
|
-
"Image"
|
|
50
|
-
],
|
|
51
|
-
"models": [
|
|
52
|
-
"SD1.5"
|
|
53
|
-
],
|
|
37
|
+
"tags": ["Texto a imagen", "Imagen"],
|
|
38
|
+
"models": ["SD1.5"],
|
|
54
39
|
"date": "2025-03-01"
|
|
55
40
|
},
|
|
56
41
|
{
|
|
57
42
|
"name": "lora_multiple",
|
|
58
|
-
"title": "LoRA
|
|
43
|
+
"title": "LoRA múltiple",
|
|
59
44
|
"mediaType": "image",
|
|
60
45
|
"mediaSubtype": "webp",
|
|
61
|
-
"description": "
|
|
46
|
+
"description": "Generar imágenes combinando múltiples modelos LoRA.",
|
|
62
47
|
"tutorialUrl": "https://docs.comfy.org/tutorials/basic/lora",
|
|
63
|
-
"tags": [
|
|
64
|
-
|
|
65
|
-
"Image",
|
|
66
|
-
"LoRA"
|
|
67
|
-
],
|
|
68
|
-
"models": [
|
|
69
|
-
"SD1.5"
|
|
70
|
-
],
|
|
48
|
+
"tags": ["Texto a imagen", "Imagen", "LoRA"],
|
|
49
|
+
"models": ["SD1.5"],
|
|
71
50
|
"date": "2025-03-01"
|
|
72
51
|
},
|
|
73
52
|
{
|
|
@@ -75,16 +54,11 @@
|
|
|
75
54
|
"title": "Inpaint",
|
|
76
55
|
"mediaType": "image",
|
|
77
56
|
"mediaSubtype": "webp",
|
|
78
|
-
"description": "
|
|
57
|
+
"description": "Editar partes específicas de imágenes sin problemas.",
|
|
79
58
|
"thumbnailVariant": "compareSlider",
|
|
80
59
|
"tutorialUrl": "https://docs.comfy.org/tutorials/basic/inpaint",
|
|
81
|
-
"tags": [
|
|
82
|
-
|
|
83
|
-
"Image"
|
|
84
|
-
],
|
|
85
|
-
"models": [
|
|
86
|
-
"SD1.5"
|
|
87
|
-
],
|
|
60
|
+
"tags": ["Inpaint", "Imagen"],
|
|
61
|
+
"models": ["SD1.5"],
|
|
88
62
|
"date": "2025-03-01"
|
|
89
63
|
},
|
|
90
64
|
{
|
|
@@ -92,16 +66,11 @@
|
|
|
92
66
|
"title": "Outpaint",
|
|
93
67
|
"mediaType": "image",
|
|
94
68
|
"mediaSubtype": "webp",
|
|
95
|
-
"description": "
|
|
69
|
+
"description": "Extender imágenes más allá de sus límites originales.",
|
|
96
70
|
"thumbnailVariant": "compareSlider",
|
|
97
71
|
"tutorialUrl": "https://docs.comfy.org/tutorials/basic/inpaint",
|
|
98
|
-
"tags": [
|
|
99
|
-
|
|
100
|
-
"Image"
|
|
101
|
-
],
|
|
102
|
-
"models": [
|
|
103
|
-
"SD1.5"
|
|
104
|
-
],
|
|
72
|
+
"tags": ["Outpaint", "Imagen"],
|
|
73
|
+
"models": ["SD1.5"],
|
|
105
74
|
"date": "2025-03-01"
|
|
106
75
|
},
|
|
107
76
|
{
|
|
@@ -109,31 +78,21 @@
|
|
|
109
78
|
"title": "Incrustación",
|
|
110
79
|
"mediaType": "image",
|
|
111
80
|
"mediaSubtype": "webp",
|
|
112
|
-
"description": "
|
|
81
|
+
"description": "Generar imágenes usando inversión textual para estilos consistentes.",
|
|
113
82
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/textual_inversion_embeddings/",
|
|
114
|
-
"tags": [
|
|
115
|
-
|
|
116
|
-
"Image"
|
|
117
|
-
],
|
|
118
|
-
"models": [
|
|
119
|
-
"SD1.5"
|
|
120
|
-
],
|
|
83
|
+
"tags": ["Incrustación", "Imagen"],
|
|
84
|
+
"models": ["SD1.5"],
|
|
121
85
|
"date": "2025-03-01"
|
|
122
86
|
},
|
|
123
87
|
{
|
|
124
88
|
"name": "gligen_textbox_example",
|
|
125
|
-
"title": "
|
|
89
|
+
"title": "Cuadro de texto Gligen",
|
|
126
90
|
"mediaType": "image",
|
|
127
91
|
"mediaSubtype": "webp",
|
|
128
|
-
"description": "
|
|
92
|
+
"description": "Generar imágenes con colocación precisa de objetos usando cuadros de texto.",
|
|
129
93
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/gligen/",
|
|
130
|
-
"tags": [
|
|
131
|
-
|
|
132
|
-
"Image"
|
|
133
|
-
],
|
|
134
|
-
"models": [
|
|
135
|
-
"SD1.5"
|
|
136
|
-
],
|
|
94
|
+
"tags": ["Gligen", "Imagen"],
|
|
95
|
+
"models": ["SD1.5"],
|
|
137
96
|
"date": "2025-03-01"
|
|
138
97
|
}
|
|
139
98
|
]
|
|
@@ -145,54 +104,25 @@
|
|
|
145
104
|
"type": "image",
|
|
146
105
|
"templates": [
|
|
147
106
|
{
|
|
148
|
-
"name": "
|
|
149
|
-
"title": "
|
|
107
|
+
"name": "image_chroma_text_to_image",
|
|
108
|
+
"title": "Texto a imagen Chroma",
|
|
150
109
|
"mediaType": "image",
|
|
151
110
|
"mediaSubtype": "webp",
|
|
152
|
-
"description": "
|
|
153
|
-
"
|
|
154
|
-
"
|
|
155
|
-
|
|
156
|
-
"Image",
|
|
157
|
-
"Photorealism"
|
|
158
|
-
],
|
|
159
|
-
"models": [
|
|
160
|
-
"Flux.1 Krea Dev"
|
|
161
|
-
],
|
|
162
|
-
"date": "2025-07-31"
|
|
111
|
+
"description": "Chroma está modificado de flux y tiene algunos cambios en la arquitectura.",
|
|
112
|
+
"tags": ["Texto a imagen", "Imagen"],
|
|
113
|
+
"models": ["Chroma", "Flux"],
|
|
114
|
+
"date": "2025-06-04"
|
|
163
115
|
},
|
|
164
116
|
{
|
|
165
117
|
"name": "flux_kontext_dev_basic",
|
|
166
|
-
"title": "Flux Kontext Dev
|
|
118
|
+
"title": "Flux Kontext Dev(Básico)",
|
|
167
119
|
"mediaType": "image",
|
|
168
120
|
"mediaSubtype": "webp",
|
|
169
121
|
"thumbnailVariant": "hoverDissolve",
|
|
170
|
-
"description": "
|
|
122
|
+
"description": "Editar imagen usando Flux Kontext con visibilidad completa de nodos, perfecto para aprender el flujo de trabajo.",
|
|
171
123
|
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-kontext-dev",
|
|
172
|
-
"tags": [
|
|
173
|
-
|
|
174
|
-
"Image to Image"
|
|
175
|
-
],
|
|
176
|
-
"models": [
|
|
177
|
-
"Flux"
|
|
178
|
-
],
|
|
179
|
-
"date": "2025-06-26"
|
|
180
|
-
},
|
|
181
|
-
{
|
|
182
|
-
"name": "flux_kontext_dev_grouped",
|
|
183
|
-
"title": "Flux Kontext Dev (Agrupado)",
|
|
184
|
-
"mediaType": "image",
|
|
185
|
-
"mediaSubtype": "webp",
|
|
186
|
-
"thumbnailVariant": "hoverDissolve",
|
|
187
|
-
"description": "Versión simplificada de Flux Kontext con nodos agrupados para un espacio de trabajo más limpio.",
|
|
188
|
-
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-kontext-dev",
|
|
189
|
-
"tags": [
|
|
190
|
-
"Image Edit",
|
|
191
|
-
"Image to Image"
|
|
192
|
-
],
|
|
193
|
-
"models": [
|
|
194
|
-
"Flux"
|
|
195
|
-
],
|
|
124
|
+
"tags": ["Edición de imagen", "Imagen a imagen"],
|
|
125
|
+
"models": ["Flux"],
|
|
196
126
|
"date": "2025-06-26"
|
|
197
127
|
},
|
|
198
128
|
{
|
|
@@ -200,15 +130,10 @@
|
|
|
200
130
|
"title": "Flux Dev fp8",
|
|
201
131
|
"mediaType": "image",
|
|
202
132
|
"mediaSubtype": "webp",
|
|
203
|
-
"description": "
|
|
133
|
+
"description": "Generar imágenes usando la versión cuantizada Flux Dev fp8. Adecuado para dispositivos con VRAM limitada, requiere solo un archivo de modelo, pero la calidad de imagen es ligeramente inferior a la versión completa.",
|
|
204
134
|
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-text-to-image",
|
|
205
|
-
"tags": [
|
|
206
|
-
|
|
207
|
-
"Image"
|
|
208
|
-
],
|
|
209
|
-
"models": [
|
|
210
|
-
"Flux"
|
|
211
|
-
],
|
|
135
|
+
"tags": ["Texto a imagen", "Imagen"],
|
|
136
|
+
"models": ["Flux"],
|
|
212
137
|
"date": "2025-03-01"
|
|
213
138
|
},
|
|
214
139
|
{
|
|
@@ -216,138 +141,102 @@
|
|
|
216
141
|
"title": "Flux Schnell fp8",
|
|
217
142
|
"mediaType": "image",
|
|
218
143
|
"mediaSubtype": "webp",
|
|
219
|
-
"description": "
|
|
144
|
+
"description": "Generar rápidamente imágenes con la versión cuantizada Flux Schnell fp8. Ideal para hardware de gama baja, requiere solo 4 pasos para generar imágenes.",
|
|
220
145
|
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-text-to-image",
|
|
221
|
-
"tags": [
|
|
222
|
-
|
|
223
|
-
"Image"
|
|
224
|
-
],
|
|
225
|
-
"models": [
|
|
226
|
-
"Flux"
|
|
227
|
-
],
|
|
146
|
+
"tags": ["Texto a imagen", "Imagen"],
|
|
147
|
+
"models": ["Flux"],
|
|
228
148
|
"date": "2025-03-01"
|
|
229
149
|
},
|
|
150
|
+
{
|
|
151
|
+
"name": "flux1_krea_dev",
|
|
152
|
+
"title": "Flux.1 Krea Dev",
|
|
153
|
+
"mediaType": "image",
|
|
154
|
+
"mediaSubtype": "webp",
|
|
155
|
+
"description": "Un modelo FLUX afinado que lleva el fotorrealismo al máximo",
|
|
156
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux1-krea-dev",
|
|
157
|
+
"tags": ["Texto a imagen", "Imagen", "Fotorrealismo"],
|
|
158
|
+
"models": ["Flux.1 Krea Dev"],
|
|
159
|
+
"date": "2025-07-31"
|
|
160
|
+
},
|
|
230
161
|
{
|
|
231
162
|
"name": "flux_dev_full_text_to_image",
|
|
232
|
-
"title": "
|
|
163
|
+
"title": "Texto a imagen completo Flux Dev",
|
|
233
164
|
"mediaType": "image",
|
|
234
165
|
"mediaSubtype": "webp",
|
|
235
|
-
"description": "
|
|
166
|
+
"description": "Generar imágenes de alta calidad con la versión completa de Flux Dev. Requiere mayor VRAM y múltiples archivos de modelo, pero proporciona la mejor capacidad de seguimiento de indicaciones y calidad de imagen.",
|
|
236
167
|
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-text-to-image",
|
|
237
|
-
"tags": [
|
|
238
|
-
|
|
239
|
-
"Image"
|
|
240
|
-
],
|
|
241
|
-
"models": [
|
|
242
|
-
"Flux"
|
|
243
|
-
],
|
|
168
|
+
"tags": ["Texto a imagen", "Imagen"],
|
|
169
|
+
"models": ["Flux"],
|
|
244
170
|
"date": "2025-03-01"
|
|
245
171
|
},
|
|
246
172
|
{
|
|
247
173
|
"name": "flux_schnell_full_text_to_image",
|
|
248
|
-
"title": "
|
|
174
|
+
"title": "Texto a imagen completo Flux Schnell",
|
|
249
175
|
"mediaType": "image",
|
|
250
176
|
"mediaSubtype": "webp",
|
|
251
|
-
"description": "
|
|
177
|
+
"description": "Generar rápidamente imágenes con la versión completa de Flux Schnell. Usa licencia Apache2.0, requiere solo 4 pasos para generar imágenes manteniendo buena calidad de imagen.",
|
|
252
178
|
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-text-to-image",
|
|
253
|
-
"tags": [
|
|
254
|
-
|
|
255
|
-
"Image"
|
|
256
|
-
],
|
|
257
|
-
"models": [
|
|
258
|
-
"Flux"
|
|
259
|
-
],
|
|
179
|
+
"tags": ["Texto a imagen", "Imagen"],
|
|
180
|
+
"models": ["Flux"],
|
|
260
181
|
"date": "2025-03-01"
|
|
261
182
|
},
|
|
262
183
|
{
|
|
263
184
|
"name": "flux_fill_inpaint_example",
|
|
264
|
-
"title": "Flux
|
|
185
|
+
"title": "Inpaint Flux",
|
|
265
186
|
"mediaType": "image",
|
|
266
187
|
"mediaSubtype": "webp",
|
|
267
|
-
"description": "
|
|
188
|
+
"description": "Rellenar partes faltantes de imágenes usando inpainting de Flux.",
|
|
268
189
|
"thumbnailVariant": "compareSlider",
|
|
269
190
|
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-fill-dev",
|
|
270
|
-
"tags": [
|
|
271
|
-
|
|
272
|
-
"Inpaint",
|
|
273
|
-
"Image"
|
|
274
|
-
],
|
|
275
|
-
"models": [
|
|
276
|
-
"Flux"
|
|
277
|
-
],
|
|
191
|
+
"tags": ["Imagen a imagen", "Inpaint", "Imagen"],
|
|
192
|
+
"models": ["Flux"],
|
|
278
193
|
"date": "2025-03-01"
|
|
279
194
|
},
|
|
280
195
|
{
|
|
281
196
|
"name": "flux_fill_outpaint_example",
|
|
282
|
-
"title": "Flux
|
|
197
|
+
"title": "Outpaint Flux",
|
|
283
198
|
"mediaType": "image",
|
|
284
199
|
"mediaSubtype": "webp",
|
|
285
|
-
"description": "
|
|
200
|
+
"description": "Extender imágenes más allá de los límites usando outpainting de Flux.",
|
|
286
201
|
"thumbnailVariant": "compareSlider",
|
|
287
202
|
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-fill-dev",
|
|
288
|
-
"tags": [
|
|
289
|
-
|
|
290
|
-
"Image",
|
|
291
|
-
"Image to Image"
|
|
292
|
-
],
|
|
293
|
-
"models": [
|
|
294
|
-
"Flux"
|
|
295
|
-
],
|
|
203
|
+
"tags": ["Outpaint", "Imagen", "Imagen a imagen"],
|
|
204
|
+
"models": ["Flux"],
|
|
296
205
|
"date": "2025-03-01"
|
|
297
206
|
},
|
|
298
207
|
{
|
|
299
208
|
"name": "flux_canny_model_example",
|
|
300
|
-
"title": "
|
|
209
|
+
"title": "Modelo Canny Flux",
|
|
301
210
|
"mediaType": "image",
|
|
302
211
|
"mediaSubtype": "webp",
|
|
303
|
-
"description": "
|
|
212
|
+
"description": "Generar imágenes guiadas por detección de bordes usando Flux Canny.",
|
|
304
213
|
"thumbnailVariant": "hoverDissolve",
|
|
305
214
|
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-controlnet",
|
|
306
|
-
"tags": [
|
|
307
|
-
|
|
308
|
-
"ControlNet",
|
|
309
|
-
"Image"
|
|
310
|
-
],
|
|
311
|
-
"models": [
|
|
312
|
-
"Flux"
|
|
313
|
-
],
|
|
215
|
+
"tags": ["Imagen a imagen", "ControlNet", "Imagen"],
|
|
216
|
+
"models": ["Flux"],
|
|
314
217
|
"date": "2025-03-01"
|
|
315
218
|
},
|
|
316
219
|
{
|
|
317
220
|
"name": "flux_depth_lora_example",
|
|
318
|
-
"title": "
|
|
221
|
+
"title": "LoRA de profundidad Flux",
|
|
319
222
|
"mediaType": "image",
|
|
320
223
|
"mediaSubtype": "webp",
|
|
321
|
-
"description": "
|
|
224
|
+
"description": "Generar imágenes guiadas por información de profundidad usando Flux LoRA.",
|
|
322
225
|
"thumbnailVariant": "hoverDissolve",
|
|
323
226
|
"tutorialUrl": "ttps://docs.comfy.org/tutorials/flux/flux-1-controlnet",
|
|
324
|
-
"tags": [
|
|
325
|
-
|
|
326
|
-
"ControlNet",
|
|
327
|
-
"Image",
|
|
328
|
-
"LoRA"
|
|
329
|
-
],
|
|
330
|
-
"models": [
|
|
331
|
-
"Flux"
|
|
332
|
-
],
|
|
227
|
+
"tags": ["Imagen a imagen", "ControlNet", "Imagen", "LoRA"],
|
|
228
|
+
"models": ["Flux"],
|
|
333
229
|
"date": "2025-03-01"
|
|
334
230
|
},
|
|
335
231
|
{
|
|
336
232
|
"name": "flux_redux_model_example",
|
|
337
|
-
"title": "
|
|
233
|
+
"title": "Modelo Redux Flux",
|
|
338
234
|
"mediaType": "image",
|
|
339
235
|
"mediaSubtype": "webp",
|
|
340
|
-
"description": "
|
|
236
|
+
"description": "Generar imágenes transfiriendo estilo de imágenes de referencia usando Flux Redux.",
|
|
341
237
|
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-controlnet",
|
|
342
|
-
"tags": [
|
|
343
|
-
|
|
344
|
-
"ControlNet",
|
|
345
|
-
"Image",
|
|
346
|
-
"LoRA"
|
|
347
|
-
],
|
|
348
|
-
"models": [
|
|
349
|
-
"Flux"
|
|
350
|
-
],
|
|
238
|
+
"tags": ["Imagen a imagen", "ControlNet", "Imagen", "LoRA"],
|
|
239
|
+
"models": ["Flux"],
|
|
351
240
|
"date": "2025-03-01"
|
|
352
241
|
}
|
|
353
242
|
]
|
|
@@ -360,165 +249,115 @@
|
|
|
360
249
|
"templates": [
|
|
361
250
|
{
|
|
362
251
|
"name": "image_qwen_image",
|
|
363
|
-
"title": "
|
|
252
|
+
"title": "Texto a imagen Qwen-Image",
|
|
364
253
|
"mediaType": "image",
|
|
365
254
|
"mediaSubtype": "webp",
|
|
366
|
-
"description": "
|
|
255
|
+
"description": "Generar imágenes con capacidad excepcional de renderizado y edición de texto multilingüe usando el modelo MMDiT de 20B de Qwen-Image.",
|
|
367
256
|
"tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image",
|
|
368
|
-
"tags": [
|
|
369
|
-
|
|
370
|
-
"Image"
|
|
371
|
-
],
|
|
372
|
-
"models": [
|
|
373
|
-
"Qwen-Image"
|
|
374
|
-
],
|
|
257
|
+
"tags": ["Texto a imagen", "Imagen"],
|
|
258
|
+
"models": ["Qwen-Image"],
|
|
375
259
|
"date": "2025-08-05"
|
|
376
260
|
},
|
|
261
|
+
{
|
|
262
|
+
"name": "image_qwen_image_union_control_lora",
|
|
263
|
+
"title": "Control unificado Qwen-Image",
|
|
264
|
+
"mediaType": "image",
|
|
265
|
+
"mediaSubtype": "webp",
|
|
266
|
+
"description": "Generar imágenes con control estructural preciso usando el ControlNet LoRA unificado de Qwen-Image. Soporta múltiples tipos de control incluyendo canny, profundidad, lineart, softedge, normal y openpose para aplicaciones creativas diversas.",
|
|
267
|
+
"tags": ["Texto a imagen", "Imagen", "ControlNet"],
|
|
268
|
+
"models": ["Qwen-Image"],
|
|
269
|
+
"date": "2025-08-23"
|
|
270
|
+
},
|
|
271
|
+
{
|
|
272
|
+
"name": "image_qwen_image_edit",
|
|
273
|
+
"title": "Edición de imagen Qwen",
|
|
274
|
+
"mediaType": "image",
|
|
275
|
+
"mediaSubtype": "webp",
|
|
276
|
+
"thumbnailVariant": "compareSlider",
|
|
277
|
+
"description": "Editar imágenes con edición precisa de texto bilingüe y capacidades de edición dual semántica/apariencia usando el modelo MMDiT de 20B de Qwen-Image-Edit.",
|
|
278
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image-edit",
|
|
279
|
+
"tags": ["Imagen a imagen", "Edición de imagen"],
|
|
280
|
+
"models": ["Qwen-Image"],
|
|
281
|
+
"date": "2025-08-18"
|
|
282
|
+
},
|
|
377
283
|
{
|
|
378
284
|
"name": "image_omnigen2_t2i",
|
|
379
|
-
"title": "
|
|
285
|
+
"title": "Texto a imagen OmniGen2",
|
|
380
286
|
"mediaType": "image",
|
|
381
287
|
"mediaSubtype": "webp",
|
|
382
|
-
"description": "
|
|
288
|
+
"description": "Generar imágenes de alta calidad a partir de indicaciones de texto usando el modelo multimodal unificado de 7B de OmniGen2 con arquitectura de doble ruta.",
|
|
383
289
|
"tutorialUrl": "https://docs.comfy.org/tutorials/image/omnigen/omnigen2",
|
|
384
|
-
"tags": [
|
|
385
|
-
|
|
386
|
-
"Image"
|
|
387
|
-
],
|
|
388
|
-
"models": [
|
|
389
|
-
"OmniGen"
|
|
390
|
-
],
|
|
290
|
+
"tags": ["Texto a imagen", "Imagen"],
|
|
291
|
+
"models": ["OmniGen"],
|
|
391
292
|
"date": "2025-06-30"
|
|
392
293
|
},
|
|
393
294
|
{
|
|
394
295
|
"name": "image_omnigen2_image_edit",
|
|
395
|
-
"title": "
|
|
296
|
+
"title": "Edición de imagen OmniGen2",
|
|
396
297
|
"mediaType": "image",
|
|
397
298
|
"mediaSubtype": "webp",
|
|
398
299
|
"thumbnailVariant": "hoverDissolve",
|
|
399
|
-
"description": "
|
|
300
|
+
"description": "Editar imágenes con instrucciones de lenguaje natural usando las capacidades avanzadas de edición de imágenes y soporte de renderizado de texto de OmniGen2.",
|
|
400
301
|
"tutorialUrl": "https://docs.comfy.org/tutorials/image/omnigen/omnigen2",
|
|
401
|
-
"tags": [
|
|
402
|
-
|
|
403
|
-
"Image"
|
|
404
|
-
],
|
|
405
|
-
"models": [
|
|
406
|
-
"OmniGen"
|
|
407
|
-
],
|
|
302
|
+
"tags": ["Edición de imagen", "Imagen"],
|
|
303
|
+
"models": ["OmniGen"],
|
|
408
304
|
"date": "2025-06-30"
|
|
409
305
|
},
|
|
410
|
-
{
|
|
411
|
-
"name": "image_cosmos_predict2_2B_t2i",
|
|
412
|
-
"title": "Cosmos Predict2 2B T2I",
|
|
413
|
-
"mediaType": "image",
|
|
414
|
-
"mediaSubtype": "webp",
|
|
415
|
-
"description": "Genera imágenes con Cosmos-Predict2 2B T2I, logrando generación física precisa, alta fidelidad y gran detalle.",
|
|
416
|
-
"tutorialUrl": "https://docs.comfy.org/tutorials/image/cosmos/cosmos-predict2-t2i",
|
|
417
|
-
"tags": [
|
|
418
|
-
"Text to Image",
|
|
419
|
-
"Image"
|
|
420
|
-
],
|
|
421
|
-
"models": [
|
|
422
|
-
"Cosmos"
|
|
423
|
-
],
|
|
424
|
-
"date": "2025-06-16"
|
|
425
|
-
},
|
|
426
|
-
{
|
|
427
|
-
"name": "image_chroma_text_to_image",
|
|
428
|
-
"title": "Chroma texto a imagen",
|
|
429
|
-
"mediaType": "image",
|
|
430
|
-
"mediaSubtype": "webp",
|
|
431
|
-
"description": "Chroma está modificado de Flux y tiene algunos cambios en la arquitectura.",
|
|
432
|
-
"tags": [
|
|
433
|
-
"Text to Image",
|
|
434
|
-
"Image"
|
|
435
|
-
],
|
|
436
|
-
"models": [
|
|
437
|
-
"Chroma",
|
|
438
|
-
"Flux"
|
|
439
|
-
],
|
|
440
|
-
"date": "2025-06-04"
|
|
441
|
-
},
|
|
442
306
|
{
|
|
443
307
|
"name": "hidream_i1_dev",
|
|
444
308
|
"title": "HiDream I1 Dev",
|
|
445
309
|
"mediaType": "image",
|
|
446
310
|
"mediaSubtype": "webp",
|
|
447
|
-
"description": "
|
|
311
|
+
"description": "Generar imágenes con HiDream I1 Dev - Versión equilibrada con 28 pasos de inferencia, adecuada para hardware de gama media.",
|
|
448
312
|
"tutorialUrl": "https://docs.comfy.org/tutorials/image/hidream/hidream-i1",
|
|
449
|
-
"tags": [
|
|
450
|
-
|
|
451
|
-
"Image"
|
|
452
|
-
],
|
|
453
|
-
"models": [
|
|
454
|
-
"HiDream"
|
|
455
|
-
],
|
|
313
|
+
"tags": ["Texto a imagen", "Imagen"],
|
|
314
|
+
"models": ["HiDream"],
|
|
456
315
|
"date": "2025-04-17"
|
|
457
316
|
},
|
|
458
317
|
{
|
|
459
318
|
"name": "hidream_i1_fast",
|
|
460
|
-
"title": "HiDream I1
|
|
319
|
+
"title": "HiDream I1 Fast",
|
|
461
320
|
"mediaType": "image",
|
|
462
321
|
"mediaSubtype": "webp",
|
|
463
|
-
"description": "
|
|
322
|
+
"description": "Generar rápidamente imágenes con HiDream I1 Fast - Versión ligera con 16 pasos de inferencia, ideal para vistas previas rápidas en hardware de gama baja.",
|
|
464
323
|
"tutorialUrl": "https://docs.comfy.org/tutorials/image/hidream/hidream-i1",
|
|
465
|
-
"tags": [
|
|
466
|
-
|
|
467
|
-
"Image"
|
|
468
|
-
],
|
|
469
|
-
"models": [
|
|
470
|
-
"HiDream"
|
|
471
|
-
],
|
|
324
|
+
"tags": ["Texto a imagen", "Imagen"],
|
|
325
|
+
"models": ["HiDream"],
|
|
472
326
|
"date": "2025-04-17"
|
|
473
327
|
},
|
|
474
328
|
{
|
|
475
329
|
"name": "hidream_i1_full",
|
|
476
|
-
"title": "HiDream I1
|
|
330
|
+
"title": "HiDream I1 Full",
|
|
477
331
|
"mediaType": "image",
|
|
478
332
|
"mediaSubtype": "webp",
|
|
479
|
-
"description": "
|
|
333
|
+
"description": "Generar imágenes con HiDream I1 Full - Versión completa con 50 pasos de inferencia para la mejor calidad de salida.",
|
|
480
334
|
"tutorialUrl": "https://docs.comfy.org/tutorials/image/hidream/hidream-i1",
|
|
481
|
-
"tags": [
|
|
482
|
-
|
|
483
|
-
"Image"
|
|
484
|
-
],
|
|
485
|
-
"models": [
|
|
486
|
-
"HiDream"
|
|
487
|
-
],
|
|
335
|
+
"tags": ["Texto a imagen", "Imagen"],
|
|
336
|
+
"models": ["HiDream"],
|
|
488
337
|
"date": "2025-04-17"
|
|
489
338
|
},
|
|
490
339
|
{
|
|
491
340
|
"name": "hidream_e1_1",
|
|
492
|
-
"title": "HiDream E1.1
|
|
341
|
+
"title": "Edición de imagen HiDream E1.1",
|
|
493
342
|
"mediaType": "image",
|
|
494
343
|
"mediaSubtype": "webp",
|
|
495
344
|
"thumbnailVariant": "compareSlider",
|
|
496
|
-
"description": "
|
|
345
|
+
"description": "Editar imágenes con HiDream E1.1 – Es mejor en calidad de imagen y precisión de edición que HiDream-E1-Full.",
|
|
497
346
|
"tutorialUrl": "https://docs.comfy.org/tutorials/image/hidream/hidream-e1",
|
|
498
|
-
"tags": [
|
|
499
|
-
|
|
500
|
-
"Image"
|
|
501
|
-
],
|
|
502
|
-
"models": [
|
|
503
|
-
"HiDream"
|
|
504
|
-
],
|
|
347
|
+
"tags": ["Edición de imagen", "Imagen"],
|
|
348
|
+
"models": ["HiDream"],
|
|
505
349
|
"date": "2025-07-21"
|
|
506
350
|
},
|
|
507
351
|
{
|
|
508
352
|
"name": "hidream_e1_full",
|
|
509
|
-
"title": "HiDream E1
|
|
353
|
+
"title": "Edición de imagen HiDream E1",
|
|
510
354
|
"mediaType": "image",
|
|
511
355
|
"mediaSubtype": "webp",
|
|
512
356
|
"thumbnailVariant": "compareSlider",
|
|
513
|
-
"description": "
|
|
357
|
+
"description": "Editar imágenes con HiDream E1 - Modelo profesional de edición de imagen con lenguaje natural.",
|
|
514
358
|
"tutorialUrl": "https://docs.comfy.org/tutorials/image/hidream/hidream-e1",
|
|
515
|
-
"tags": [
|
|
516
|
-
|
|
517
|
-
"Image"
|
|
518
|
-
],
|
|
519
|
-
"models": [
|
|
520
|
-
"HiDream"
|
|
521
|
-
],
|
|
359
|
+
"tags": ["Edición de imagen", "Imagen"],
|
|
360
|
+
"models": ["HiDream"],
|
|
522
361
|
"date": "2025-05-01"
|
|
523
362
|
},
|
|
524
363
|
{
|
|
@@ -526,68 +365,46 @@
|
|
|
526
365
|
"title": "SD3.5 Simple",
|
|
527
366
|
"mediaType": "image",
|
|
528
367
|
"mediaSubtype": "webp",
|
|
529
|
-
"description": "
|
|
368
|
+
"description": "Generar imágenes usando SD 3.5.",
|
|
530
369
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sd3/#sd35",
|
|
531
|
-
"tags": [
|
|
532
|
-
|
|
533
|
-
"Image"
|
|
534
|
-
],
|
|
535
|
-
"models": [
|
|
536
|
-
"SD3.5"
|
|
537
|
-
],
|
|
370
|
+
"tags": ["Texto a imagen", "Imagen"],
|
|
371
|
+
"models": ["SD3.5"],
|
|
538
372
|
"date": "2025-03-01"
|
|
539
373
|
},
|
|
540
374
|
{
|
|
541
375
|
"name": "sd3.5_large_canny_controlnet_example",
|
|
542
|
-
"title": "SD3.5
|
|
376
|
+
"title": "ControlNet Canny grande SD3.5",
|
|
543
377
|
"mediaType": "image",
|
|
544
378
|
"mediaSubtype": "webp",
|
|
545
|
-
"description": "
|
|
379
|
+
"description": "Generar imágenes guiadas por detección de bordes usando ControlNet Canny SD 3.5.",
|
|
546
380
|
"thumbnailVariant": "hoverDissolve",
|
|
547
381
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sd3/#sd35-controlnets",
|
|
548
|
-
"tags": [
|
|
549
|
-
|
|
550
|
-
"Image",
|
|
551
|
-
"ControlNet"
|
|
552
|
-
],
|
|
553
|
-
"models": [
|
|
554
|
-
"SD3.5"
|
|
555
|
-
],
|
|
382
|
+
"tags": ["Imagen a imagen", "Imagen", "ControlNet"],
|
|
383
|
+
"models": ["SD3.5"],
|
|
556
384
|
"date": "2025-03-01"
|
|
557
385
|
},
|
|
558
386
|
{
|
|
559
387
|
"name": "sd3.5_large_depth",
|
|
560
|
-
"title": "SD3.5
|
|
388
|
+
"title": "Profundidad grande SD3.5",
|
|
561
389
|
"mediaType": "image",
|
|
562
390
|
"mediaSubtype": "webp",
|
|
563
|
-
"description": "
|
|
391
|
+
"description": "Generar imágenes guiadas por información de profundidad usando SD 3.5.",
|
|
564
392
|
"thumbnailVariant": "hoverDissolve",
|
|
565
393
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sd3/#sd35-controlnets",
|
|
566
|
-
"tags": [
|
|
567
|
-
|
|
568
|
-
"Image",
|
|
569
|
-
"ControlNet"
|
|
570
|
-
],
|
|
571
|
-
"models": [
|
|
572
|
-
"SD3.5"
|
|
573
|
-
],
|
|
394
|
+
"tags": ["Imagen a imagen", "Imagen", "ControlNet"],
|
|
395
|
+
"models": ["SD3.5"],
|
|
574
396
|
"date": "2025-03-01"
|
|
575
397
|
},
|
|
576
398
|
{
|
|
577
399
|
"name": "sd3.5_large_blur",
|
|
578
|
-
"title": "SD3.5
|
|
400
|
+
"title": "Desenfoque grande SD3.5",
|
|
579
401
|
"mediaType": "image",
|
|
580
402
|
"mediaSubtype": "webp",
|
|
581
|
-
"description": "
|
|
403
|
+
"description": "Generar imágenes guiadas por imágenes de referencia desenfocadas usando SD 3.5.",
|
|
582
404
|
"thumbnailVariant": "hoverDissolve",
|
|
583
405
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sd3/#sd35-controlnets",
|
|
584
|
-
"tags": [
|
|
585
|
-
|
|
586
|
-
"Image"
|
|
587
|
-
],
|
|
588
|
-
"models": [
|
|
589
|
-
"SD3.5"
|
|
590
|
-
],
|
|
406
|
+
"tags": ["Imagen a imagen", "Imagen"],
|
|
407
|
+
"models": ["SD3.5"],
|
|
591
408
|
"date": "2025-03-01"
|
|
592
409
|
},
|
|
593
410
|
{
|
|
@@ -595,63 +412,43 @@
|
|
|
595
412
|
"title": "SDXL Simple",
|
|
596
413
|
"mediaType": "image",
|
|
597
414
|
"mediaSubtype": "webp",
|
|
598
|
-
"description": "
|
|
415
|
+
"description": "Generar imágenes de alta calidad usando SDXL.",
|
|
599
416
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sdxl/",
|
|
600
|
-
"tags": [
|
|
601
|
-
|
|
602
|
-
"Image"
|
|
603
|
-
],
|
|
604
|
-
"models": [
|
|
605
|
-
"SDXL"
|
|
606
|
-
],
|
|
417
|
+
"tags": ["Texto a imagen", "Imagen"],
|
|
418
|
+
"models": ["SDXL"],
|
|
607
419
|
"date": "2025-03-01"
|
|
608
420
|
},
|
|
609
421
|
{
|
|
610
422
|
"name": "sdxl_refiner_prompt_example",
|
|
611
|
-
"title": "
|
|
423
|
+
"title": "Refinador de indicaciones SDXL",
|
|
612
424
|
"mediaType": "image",
|
|
613
425
|
"mediaSubtype": "webp",
|
|
614
|
-
"description": "
|
|
426
|
+
"description": "Mejorar imágenes SDXL usando modelos refinadores.",
|
|
615
427
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sdxl/",
|
|
616
|
-
"tags": [
|
|
617
|
-
|
|
618
|
-
"Image"
|
|
619
|
-
],
|
|
620
|
-
"models": [
|
|
621
|
-
"SDXL"
|
|
622
|
-
],
|
|
428
|
+
"tags": ["Texto a imagen", "Imagen"],
|
|
429
|
+
"models": ["SDXL"],
|
|
623
430
|
"date": "2025-03-01"
|
|
624
431
|
},
|
|
625
432
|
{
|
|
626
433
|
"name": "sdxl_revision_text_prompts",
|
|
627
|
-
"title": "
|
|
434
|
+
"title": "Indicaciones de texto de revisión SDXL",
|
|
628
435
|
"mediaType": "image",
|
|
629
436
|
"mediaSubtype": "webp",
|
|
630
|
-
"description": "
|
|
437
|
+
"description": "Generar imágenes transfiriendo conceptos de imágenes de referencia usando Revisión SDXL.",
|
|
631
438
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sdxl/#revision",
|
|
632
|
-
"tags": [
|
|
633
|
-
|
|
634
|
-
"Image"
|
|
635
|
-
],
|
|
636
|
-
"models": [
|
|
637
|
-
"SDXL"
|
|
638
|
-
],
|
|
439
|
+
"tags": ["Texto a imagen", "Imagen"],
|
|
440
|
+
"models": ["SDXL"],
|
|
639
441
|
"date": "2025-03-01"
|
|
640
442
|
},
|
|
641
443
|
{
|
|
642
444
|
"name": "sdxl_revision_zero_positive",
|
|
643
|
-
"title": "
|
|
445
|
+
"title": "Revisión cero positivo SDXL",
|
|
644
446
|
"mediaType": "image",
|
|
645
447
|
"mediaSubtype": "webp",
|
|
646
|
-
"description": "
|
|
448
|
+
"description": "Generar imágenes usando tanto indicaciones de texto como imágenes de referencia con Revisión SDXL.",
|
|
647
449
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sdxl/#revision",
|
|
648
|
-
"tags": [
|
|
649
|
-
|
|
650
|
-
"Image"
|
|
651
|
-
],
|
|
652
|
-
"models": [
|
|
653
|
-
"SDXL"
|
|
654
|
-
],
|
|
450
|
+
"tags": ["Texto a imagen", "Imagen"],
|
|
451
|
+
"models": ["SDXL"],
|
|
655
452
|
"date": "2025-03-01"
|
|
656
453
|
},
|
|
657
454
|
{
|
|
@@ -659,31 +456,21 @@
|
|
|
659
456
|
"title": "SDXL Turbo",
|
|
660
457
|
"mediaType": "image",
|
|
661
458
|
"mediaSubtype": "webp",
|
|
662
|
-
"description": "
|
|
459
|
+
"description": "Generar imágenes en un solo paso usando SDXL Turbo.",
|
|
663
460
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/sdturbo/",
|
|
664
|
-
"tags": [
|
|
665
|
-
|
|
666
|
-
"Image"
|
|
667
|
-
],
|
|
668
|
-
"models": [
|
|
669
|
-
"SDXL Turbo"
|
|
670
|
-
],
|
|
461
|
+
"tags": ["Texto a imagen", "Imagen"],
|
|
462
|
+
"models": ["SDXL Turbo"],
|
|
671
463
|
"date": "2025-03-01"
|
|
672
464
|
},
|
|
673
465
|
{
|
|
674
466
|
"name": "image_lotus_depth_v1_1",
|
|
675
|
-
"title": "Lotus
|
|
467
|
+
"title": "Profundidad Lotus",
|
|
676
468
|
"mediaType": "image",
|
|
677
469
|
"mediaSubtype": "webp",
|
|
678
470
|
"thumbnailVariant": "compareSlider",
|
|
679
|
-
"description": "
|
|
680
|
-
"tags": [
|
|
681
|
-
|
|
682
|
-
"Image"
|
|
683
|
-
],
|
|
684
|
-
"models": [
|
|
685
|
-
"SD1.5"
|
|
686
|
-
],
|
|
471
|
+
"description": "Ejecutar Profundidad Lotus en ComfyUI para estimación monocromática eficiente sin entrenamiento previo con alta retención de detalles.",
|
|
472
|
+
"tags": ["Profundidad", "Imagen"],
|
|
473
|
+
"models": ["SD1.5"],
|
|
687
474
|
"date": "2025-05-21"
|
|
688
475
|
}
|
|
689
476
|
]
|
|
@@ -696,439 +483,307 @@
|
|
|
696
483
|
"templates": [
|
|
697
484
|
{
|
|
698
485
|
"name": "video_wan2_2_14B_t2v",
|
|
699
|
-
"title": "Wan 2.2 14B
|
|
700
|
-
"description": "
|
|
486
|
+
"title": "Texto a video Wan 2.2 14B",
|
|
487
|
+
"description": "Generar videos de alta calidad a partir de indicaciones de texto con control estético cinematográfico y generación de movimiento dinámico usando Wan 2.2.",
|
|
701
488
|
"mediaType": "image",
|
|
702
489
|
"mediaSubtype": "webp",
|
|
703
490
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/wan2_2",
|
|
704
|
-
"tags": [
|
|
705
|
-
|
|
706
|
-
"Video"
|
|
707
|
-
],
|
|
708
|
-
"models": [
|
|
709
|
-
"Wan"
|
|
710
|
-
],
|
|
491
|
+
"tags": ["Texto a video", "Video"],
|
|
492
|
+
"models": ["Wan"],
|
|
711
493
|
"date": "2025-07-29"
|
|
712
494
|
},
|
|
713
495
|
{
|
|
714
496
|
"name": "video_wan2_2_14B_i2v",
|
|
715
|
-
"title": "Wan 2.2 14B
|
|
716
|
-
"description": "
|
|
497
|
+
"title": "Imagen a video Wan 2.2 14B",
|
|
498
|
+
"description": "Transformar imágenes estáticas en videos dinámicos con control de movimiento preciso y preservación de estilo usando Wan 2.2.",
|
|
717
499
|
"mediaType": "image",
|
|
718
500
|
"mediaSubtype": "webp",
|
|
719
501
|
"thumbnailVariant": "hoverDissolve",
|
|
720
502
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/wan2_2",
|
|
721
|
-
"tags": [
|
|
722
|
-
|
|
723
|
-
"Video"
|
|
724
|
-
],
|
|
725
|
-
"models": [
|
|
726
|
-
"Wan2.2"
|
|
727
|
-
],
|
|
503
|
+
"tags": ["Imagen a video", "Video"],
|
|
504
|
+
"models": ["Wan2.2"],
|
|
728
505
|
"date": "2025-07-29"
|
|
729
506
|
},
|
|
730
507
|
{
|
|
731
508
|
"name": "video_wan2_2_14B_flf2v",
|
|
732
|
-
"title": "Wan 2.2 14B
|
|
733
|
-
"description": "
|
|
509
|
+
"title": "Primer-Último fotograma a video Wan 2.2 14B",
|
|
510
|
+
"description": "Generar transiciones de video suaves definiendo fotogramas de inicio y fin.",
|
|
734
511
|
"mediaType": "image",
|
|
735
512
|
"mediaSubtype": "webp",
|
|
736
513
|
"thumbnailVariant": "hoverDissolve",
|
|
737
514
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/wan2_2",
|
|
738
|
-
"tags": [
|
|
739
|
-
|
|
740
|
-
"Video"
|
|
741
|
-
],
|
|
742
|
-
"models": [
|
|
743
|
-
"Wan2.2"
|
|
744
|
-
],
|
|
515
|
+
"tags": ["FLF2V", "Video"],
|
|
516
|
+
"models": ["Wan2.2"],
|
|
745
517
|
"date": "2025-08-02"
|
|
746
518
|
},
|
|
519
|
+
{
|
|
520
|
+
"name": "video_wan2_2_14B_fun_inpaint",
|
|
521
|
+
"title": "Fun Inp Wan 2.2 14B",
|
|
522
|
+
"description": "Generar videos desde fotogramas de inicio y fin usando Wan 2.2 Fun Inp.",
|
|
523
|
+
"mediaType": "image",
|
|
524
|
+
"mediaSubtype": "webp",
|
|
525
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/wan2-2-fun-inp",
|
|
526
|
+
"tags": ["FLF2V", "Video"],
|
|
527
|
+
"models": ["Wan2.2"],
|
|
528
|
+
"date": "2025-08-12"
|
|
529
|
+
},
|
|
530
|
+
{
|
|
531
|
+
"name": "video_wan2_2_14B_fun_control",
|
|
532
|
+
"title": "Control Fun Wan 2.2 14B",
|
|
533
|
+
"description": "Generar videos guiados por controles de pose, profundidad y borde usando Wan 2.2 Fun Control.",
|
|
534
|
+
"mediaType": "image",
|
|
535
|
+
"mediaSubtype": "webp",
|
|
536
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/wan2-2-fun-control",
|
|
537
|
+
"tags": ["Video a video", "Video"],
|
|
538
|
+
"models": ["Wan2.2"],
|
|
539
|
+
"date": "2025-08-12"
|
|
540
|
+
},
|
|
541
|
+
{
|
|
542
|
+
"name": "video_wan2_2_14B_fun_camera",
|
|
543
|
+
"title": "Control de cámara Fun Wan 2.2 14B",
|
|
544
|
+
"description": "Generar videos con controles de movimiento de cámara incluyendo panorámica, zoom y rotación usando Wan 2.2 Fun Camera Control.",
|
|
545
|
+
"mediaType": "image",
|
|
546
|
+
"mediaSubtype": "webp",
|
|
547
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/wan2-2-fun-camera",
|
|
548
|
+
"tags": ["Video a video", "Video"],
|
|
549
|
+
"models": ["Wan2.2"],
|
|
550
|
+
"date": "2025-08-17"
|
|
551
|
+
},
|
|
747
552
|
{
|
|
748
553
|
"name": "video_wan2_2_5B_ti2v",
|
|
749
|
-
"title": "Wan 2.2 5B
|
|
750
|
-
"description": "
|
|
554
|
+
"title": "Generación de video Wan 2.2 5B",
|
|
555
|
+
"description": "Generar videos a partir de texto o imágenes usando el modelo híbrido Wan 2.2 5B",
|
|
751
556
|
"mediaType": "image",
|
|
752
557
|
"mediaSubtype": "webp",
|
|
753
558
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/wan2_2",
|
|
754
|
-
"tags": [
|
|
755
|
-
|
|
756
|
-
"Video"
|
|
757
|
-
],
|
|
758
|
-
"models": [
|
|
759
|
-
"Wan2.2"
|
|
760
|
-
],
|
|
559
|
+
"tags": ["Texto a video", "Video"],
|
|
560
|
+
"models": ["Wan2.2"],
|
|
761
561
|
"date": "2025-07-29"
|
|
762
562
|
},
|
|
763
563
|
{
|
|
764
564
|
"name": "video_wan_vace_14B_t2v",
|
|
765
|
-
"title": "
|
|
766
|
-
"description": "
|
|
565
|
+
"title": "Texto a video Wan VACE",
|
|
566
|
+
"description": "Transformar descripciones de texto en videos de alta calidad. Soporta tanto 480p como 720p con el modelo VACE-14B.",
|
|
767
567
|
"mediaType": "image",
|
|
768
568
|
"mediaSubtype": "webp",
|
|
769
569
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/vace",
|
|
770
|
-
"tags": [
|
|
771
|
-
|
|
772
|
-
"Video"
|
|
773
|
-
],
|
|
774
|
-
"models": [
|
|
775
|
-
"Wan2.1"
|
|
776
|
-
],
|
|
570
|
+
"tags": ["Texto a video", "Video"],
|
|
571
|
+
"models": ["Wan2.1"],
|
|
777
572
|
"date": "2025-05-21"
|
|
778
573
|
},
|
|
779
574
|
{
|
|
780
575
|
"name": "video_wan_vace_14B_ref2v",
|
|
781
|
-
"title": "
|
|
782
|
-
"description": "
|
|
576
|
+
"title": "Referencia a video Wan VACE",
|
|
577
|
+
"description": "Crear videos que coincidan con el estilo y contenido de una imagen de referencia. Perfecto para generación de video consistente en estilo.",
|
|
783
578
|
"mediaType": "image",
|
|
784
579
|
"mediaSubtype": "webp",
|
|
785
580
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/vace",
|
|
786
|
-
"tags": [
|
|
787
|
-
|
|
788
|
-
"Video"
|
|
789
|
-
],
|
|
790
|
-
"models": [
|
|
791
|
-
"Wan2.1"
|
|
792
|
-
],
|
|
581
|
+
"tags": ["Referencia a video", "Video"],
|
|
582
|
+
"models": ["Wan2.1"],
|
|
793
583
|
"date": "2025-05-21"
|
|
794
584
|
},
|
|
795
585
|
{
|
|
796
586
|
"name": "video_wan_vace_14B_v2v",
|
|
797
|
-
"title": "Wan VACE
|
|
798
|
-
"description": "
|
|
587
|
+
"title": "Control de video Wan VACE",
|
|
588
|
+
"description": "Generar videos controlando videos de entrada e imágenes de referencia usando Wan VACE.",
|
|
799
589
|
"mediaType": "image",
|
|
800
590
|
"mediaSubtype": "webp",
|
|
801
591
|
"thumbnailVariant": "compareSlider",
|
|
802
592
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/vace",
|
|
803
|
-
"tags": [
|
|
804
|
-
|
|
805
|
-
"Video"
|
|
806
|
-
],
|
|
807
|
-
"models": [
|
|
808
|
-
"Wan2.1"
|
|
809
|
-
],
|
|
593
|
+
"tags": ["Video a video", "Video"],
|
|
594
|
+
"models": ["Wan2.1"],
|
|
810
595
|
"date": "2025-05-21"
|
|
811
596
|
},
|
|
812
597
|
{
|
|
813
598
|
"name": "video_wan_vace_outpainting",
|
|
814
|
-
"title": "Wan VACE
|
|
815
|
-
"description": "
|
|
599
|
+
"title": "Outpainting Wan VACE",
|
|
600
|
+
"description": "Generar videos extendidos expandiendo el tamaño de video usando outpainting de Wan VACE.",
|
|
816
601
|
"mediaType": "image",
|
|
817
602
|
"mediaSubtype": "webp",
|
|
818
603
|
"thumbnailVariant": "compareSlider",
|
|
819
604
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/vace",
|
|
820
|
-
"tags": [
|
|
821
|
-
|
|
822
|
-
"Video"
|
|
823
|
-
],
|
|
824
|
-
"models": [
|
|
825
|
-
"Wan2.1"
|
|
826
|
-
],
|
|
605
|
+
"tags": ["Outpainting", "Video"],
|
|
606
|
+
"models": ["Wan2.1"],
|
|
827
607
|
"date": "2025-05-21"
|
|
828
608
|
},
|
|
829
609
|
{
|
|
830
610
|
"name": "video_wan_vace_flf2v",
|
|
831
|
-
"title": "Wan VACE
|
|
832
|
-
"description": "
|
|
611
|
+
"title": "Primer-Último fotograma Wan VACE",
|
|
612
|
+
"description": "Generar transiciones de video suaves definiendo fotogramas de inicio y fin. Soporta secuencias de fotogramas clave personalizadas.",
|
|
833
613
|
"mediaType": "image",
|
|
834
614
|
"mediaSubtype": "webp",
|
|
835
615
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/vace",
|
|
836
|
-
"tags": [
|
|
837
|
-
|
|
838
|
-
"Video"
|
|
839
|
-
],
|
|
840
|
-
"models": [
|
|
841
|
-
"Wan2.1"
|
|
842
|
-
],
|
|
616
|
+
"tags": ["FLF2V", "Video"],
|
|
617
|
+
"models": ["Wan2.1"],
|
|
843
618
|
"date": "2025-05-21"
|
|
844
619
|
},
|
|
845
620
|
{
|
|
846
621
|
"name": "video_wan_vace_inpainting",
|
|
847
|
-
"title": "Wan VACE
|
|
848
|
-
"description": "
|
|
622
|
+
"title": "Inpainting Wan VACE",
|
|
623
|
+
"description": "Editar regiones específicas en videos mientras se preserva el contenido circundante. Excelente para eliminación o reemplazo de objetos.",
|
|
849
624
|
"mediaType": "image",
|
|
850
625
|
"mediaSubtype": "webp",
|
|
851
626
|
"thumbnailVariant": "compareSlider",
|
|
852
627
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/vace",
|
|
853
|
-
"tags": [
|
|
854
|
-
|
|
855
|
-
"Video"
|
|
856
|
-
],
|
|
857
|
-
"models": [
|
|
858
|
-
"Wan2.1"
|
|
859
|
-
],
|
|
628
|
+
"tags": ["Inpainting", "Video"],
|
|
629
|
+
"models": ["Wan2.1"],
|
|
860
630
|
"date": "2025-05-21"
|
|
861
631
|
},
|
|
862
632
|
{
|
|
863
633
|
"name": "video_wan_ati",
|
|
864
634
|
"title": "Wan ATI",
|
|
865
|
-
"description": "
|
|
635
|
+
"description": "Generación de video controlada por trayectoria.",
|
|
866
636
|
"mediaType": "image",
|
|
867
637
|
"mediaSubtype": "webp",
|
|
868
638
|
"thumbnailVariant": "hoverDissolve",
|
|
869
639
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/wan-ati",
|
|
870
|
-
"tags": [
|
|
871
|
-
|
|
872
|
-
],
|
|
873
|
-
"models": [
|
|
874
|
-
"Wan2.1"
|
|
875
|
-
],
|
|
640
|
+
"tags": ["Video"],
|
|
641
|
+
"models": ["Wan2.1"],
|
|
876
642
|
"date": "2025-05-21"
|
|
877
643
|
},
|
|
878
644
|
{
|
|
879
645
|
"name": "video_wan2.1_fun_camera_v1.1_1.3B",
|
|
880
|
-
"title": "
|
|
881
|
-
"description": "
|
|
646
|
+
"title": "Cámara Fun 1.3B Wan 2.1",
|
|
647
|
+
"description": "Generar videos dinámicos con movimientos cinematográficos de cámara usando el modelo Wan 2.1 Fun Camera 1.3B.",
|
|
882
648
|
"mediaType": "image",
|
|
883
649
|
"mediaSubtype": "webp",
|
|
884
650
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/fun-control",
|
|
885
|
-
"tags": [
|
|
886
|
-
|
|
887
|
-
],
|
|
888
|
-
"models": [
|
|
889
|
-
"Wan2.1"
|
|
890
|
-
],
|
|
651
|
+
"tags": ["Video"],
|
|
652
|
+
"models": ["Wan2.1"],
|
|
891
653
|
"date": "2025-04-15"
|
|
892
654
|
},
|
|
893
655
|
{
|
|
894
656
|
"name": "video_wan2.1_fun_camera_v1.1_14B",
|
|
895
|
-
"title": "Wan 2.1
|
|
896
|
-
"description": "
|
|
657
|
+
"title": "Cámara Fun 14B Wan 2.1",
|
|
658
|
+
"description": "Generar videos de alta calidad con control avanzado de cámara usando el modelo completo de 14B",
|
|
897
659
|
"mediaType": "image",
|
|
898
660
|
"mediaSubtype": "webp",
|
|
899
661
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/fun-control",
|
|
900
|
-
"tags": [
|
|
901
|
-
|
|
902
|
-
],
|
|
903
|
-
"models": [
|
|
904
|
-
"Wan2.1"
|
|
905
|
-
],
|
|
662
|
+
"tags": ["Video"],
|
|
663
|
+
"models": ["Wan2.1"],
|
|
906
664
|
"date": "2025-04-15"
|
|
907
665
|
},
|
|
908
666
|
{
|
|
909
667
|
"name": "text_to_video_wan",
|
|
910
|
-
"title": "Wan 2.1
|
|
911
|
-
"description": "
|
|
668
|
+
"title": "Texto a video Wan 2.1",
|
|
669
|
+
"description": "Generar videos a partir de indicaciones de texto usando Wan 2.1.",
|
|
912
670
|
"mediaType": "image",
|
|
913
671
|
"mediaSubtype": "webp",
|
|
914
672
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/wan-video",
|
|
915
|
-
"tags": [
|
|
916
|
-
|
|
917
|
-
"Video"
|
|
918
|
-
],
|
|
919
|
-
"models": [
|
|
920
|
-
"Wan2.1"
|
|
921
|
-
],
|
|
673
|
+
"tags": ["Texto a video", "Video"],
|
|
674
|
+
"models": ["Wan2.1"],
|
|
922
675
|
"date": "2025-03-01"
|
|
923
676
|
},
|
|
924
677
|
{
|
|
925
678
|
"name": "image_to_video_wan",
|
|
926
|
-
"title": "Wan 2.1
|
|
927
|
-
"description": "
|
|
679
|
+
"title": "Imagen a video Wan 2.1",
|
|
680
|
+
"description": "Generar videos a partir de imágenes usando Wan 2.1.",
|
|
928
681
|
"mediaType": "image",
|
|
929
682
|
"mediaSubtype": "webp",
|
|
930
683
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/wan-video",
|
|
931
|
-
"tags": [
|
|
932
|
-
|
|
933
|
-
"Video"
|
|
934
|
-
],
|
|
935
|
-
"models": [
|
|
936
|
-
"Wan2.1"
|
|
937
|
-
],
|
|
684
|
+
"tags": ["Texto a video", "Video"],
|
|
685
|
+
"models": ["Wan2.1"],
|
|
938
686
|
"date": "2025-03-01"
|
|
939
687
|
},
|
|
940
688
|
{
|
|
941
689
|
"name": "wan2.1_fun_inp",
|
|
942
|
-
"title": "Wan 2.1
|
|
943
|
-
"description": "
|
|
690
|
+
"title": "Inpainting Wan 2.1",
|
|
691
|
+
"description": "Generar videos desde fotogramas de inicio y fin usando inpainting de Wan 2.1.",
|
|
944
692
|
"mediaType": "image",
|
|
945
693
|
"mediaSubtype": "webp",
|
|
946
694
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/fun-inp",
|
|
947
|
-
"tags": [
|
|
948
|
-
|
|
949
|
-
"Video"
|
|
950
|
-
],
|
|
951
|
-
"models": [
|
|
952
|
-
"Wan2.1"
|
|
953
|
-
],
|
|
695
|
+
"tags": ["Inpaint", "Video"],
|
|
696
|
+
"models": ["Wan2.1"],
|
|
954
697
|
"date": "2025-04-15"
|
|
955
698
|
},
|
|
956
699
|
{
|
|
957
700
|
"name": "wan2.1_fun_control",
|
|
958
|
-
"title": "Wan 2.1
|
|
959
|
-
"description": "
|
|
701
|
+
"title": "ControlNet Wan 2.1",
|
|
702
|
+
"description": "Generar videos guiados por controles de pose, profundidad y borde usando ControlNet de Wan 2.1.",
|
|
960
703
|
"mediaType": "image",
|
|
961
704
|
"mediaSubtype": "webp",
|
|
962
705
|
"thumbnailVariant": "hoverDissolve",
|
|
963
706
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/fun-control",
|
|
964
|
-
"tags": [
|
|
965
|
-
|
|
966
|
-
"Video"
|
|
967
|
-
],
|
|
968
|
-
"models": [
|
|
969
|
-
"Wan2.1"
|
|
970
|
-
],
|
|
707
|
+
"tags": ["Video a video", "Video"],
|
|
708
|
+
"models": ["Wan2.1"],
|
|
971
709
|
"date": "2025-04-15"
|
|
972
710
|
},
|
|
973
711
|
{
|
|
974
712
|
"name": "wan2.1_flf2v_720_f16",
|
|
975
|
-
"title": "Wan 2.1
|
|
976
|
-
"description": "
|
|
713
|
+
"title": "FLF2V 720p F16 Wan 2.1",
|
|
714
|
+
"description": "Generar videos controlando primer y último fotogramas usando FLF2V de Wan 2.1.",
|
|
977
715
|
"mediaType": "image",
|
|
978
716
|
"mediaSubtype": "webp",
|
|
979
717
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/wan-flf",
|
|
980
|
-
"tags": [
|
|
981
|
-
|
|
982
|
-
"Video"
|
|
983
|
-
],
|
|
984
|
-
"models": [
|
|
985
|
-
"Wan2.1"
|
|
986
|
-
],
|
|
718
|
+
"tags": ["FLF2V", "Video"],
|
|
719
|
+
"models": ["Wan2.1"],
|
|
987
720
|
"date": "2025-04-15"
|
|
988
721
|
},
|
|
989
|
-
{
|
|
990
|
-
"name": "video_cosmos_predict2_2B_video2world_480p_16fps",
|
|
991
|
-
"title": "Cosmos Predict2 2B Video2World 480p 16fps",
|
|
992
|
-
"description": "Genera videos con Cosmos-Predict2 2B Video2World, logrando simulaciones físicas precisas, alta fidelidad y consistencia.",
|
|
993
|
-
"mediaType": "image",
|
|
994
|
-
"mediaSubtype": "webp",
|
|
995
|
-
"tutorialUrl": "http://docs.comfy.org/tutorials/video/cosmos/cosmos-predict2-video2world",
|
|
996
|
-
"tags": [
|
|
997
|
-
"Video2World",
|
|
998
|
-
"Video"
|
|
999
|
-
],
|
|
1000
|
-
"models": [
|
|
1001
|
-
"Cosmos"
|
|
1002
|
-
],
|
|
1003
|
-
"date": "2025-06-16"
|
|
1004
|
-
},
|
|
1005
722
|
{
|
|
1006
723
|
"name": "ltxv_text_to_video",
|
|
1007
|
-
"title": "
|
|
724
|
+
"title": "Texto a video LTXV",
|
|
1008
725
|
"mediaType": "image",
|
|
1009
726
|
"mediaSubtype": "webp",
|
|
1010
|
-
"description": "
|
|
727
|
+
"description": "Generar videos a partir de indicaciones de texto.",
|
|
1011
728
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/ltxv",
|
|
1012
|
-
"tags": [
|
|
1013
|
-
|
|
1014
|
-
"Video"
|
|
1015
|
-
],
|
|
1016
|
-
"models": [
|
|
1017
|
-
"LTXV"
|
|
1018
|
-
],
|
|
729
|
+
"tags": ["Texto a video", "Video"],
|
|
730
|
+
"models": ["LTXV"],
|
|
1019
731
|
"date": "2025-03-01"
|
|
1020
732
|
},
|
|
1021
733
|
{
|
|
1022
734
|
"name": "ltxv_image_to_video",
|
|
1023
|
-
"title": "
|
|
735
|
+
"title": "Imagen a video LTXV",
|
|
1024
736
|
"mediaType": "image",
|
|
1025
737
|
"mediaSubtype": "webp",
|
|
1026
|
-
"description": "
|
|
738
|
+
"description": "Generar videos a partir de imágenes fijas.",
|
|
1027
739
|
"tutorialUrl": "https://docs.comfy.org/tutorials/video/ltxv",
|
|
1028
|
-
"tags": [
|
|
1029
|
-
|
|
1030
|
-
"Video"
|
|
1031
|
-
],
|
|
1032
|
-
"models": [
|
|
1033
|
-
"LTXV"
|
|
1034
|
-
],
|
|
740
|
+
"tags": ["Imagen a video", "Video"],
|
|
741
|
+
"models": ["LTXV"],
|
|
1035
742
|
"date": "2025-03-01"
|
|
1036
743
|
},
|
|
1037
744
|
{
|
|
1038
745
|
"name": "mochi_text_to_video_example",
|
|
1039
|
-
"title": "
|
|
746
|
+
"title": "Texto a video Mochi",
|
|
1040
747
|
"mediaType": "image",
|
|
1041
748
|
"mediaSubtype": "webp",
|
|
1042
|
-
"description": "
|
|
749
|
+
"description": "Generar videos a partir de indicaciones de texto usando el modelo Mochi.",
|
|
1043
750
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/mochi/",
|
|
1044
|
-
"tags": [
|
|
1045
|
-
|
|
1046
|
-
"Video"
|
|
1047
|
-
],
|
|
1048
|
-
"models": [
|
|
1049
|
-
"Mochi"
|
|
1050
|
-
],
|
|
751
|
+
"tags": ["Texto a video", "Video"],
|
|
752
|
+
"models": ["Mochi"],
|
|
1051
753
|
"date": "2025-03-01"
|
|
1052
754
|
},
|
|
1053
755
|
{
|
|
1054
756
|
"name": "hunyuan_video_text_to_video",
|
|
1055
|
-
"title": "
|
|
757
|
+
"title": "Texto a video Hunyuan Video",
|
|
1056
758
|
"mediaType": "image",
|
|
1057
759
|
"mediaSubtype": "webp",
|
|
1058
|
-
"description": "
|
|
760
|
+
"description": "Generar videos a partir de indicaciones de texto usando el modelo Hunyuan.",
|
|
1059
761
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/hunyuan_video/",
|
|
1060
|
-
"tags": [
|
|
1061
|
-
|
|
1062
|
-
"Video"
|
|
1063
|
-
],
|
|
1064
|
-
"models": [
|
|
1065
|
-
"Hunyuan Video"
|
|
1066
|
-
],
|
|
762
|
+
"tags": ["Texto a video", "Video"],
|
|
763
|
+
"models": ["Hunyuan Video"],
|
|
1067
764
|
"date": "2025-03-01"
|
|
1068
765
|
},
|
|
1069
766
|
{
|
|
1070
767
|
"name": "image_to_video",
|
|
1071
|
-
"title": "
|
|
768
|
+
"title": "Imagen a video SVD",
|
|
1072
769
|
"mediaType": "image",
|
|
1073
770
|
"mediaSubtype": "webp",
|
|
1074
|
-
"description": "
|
|
771
|
+
"description": "Generar videos a partir de imágenes fijas.",
|
|
1075
772
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/video/#image-to-video",
|
|
1076
|
-
"tags": [
|
|
1077
|
-
|
|
1078
|
-
"Video"
|
|
1079
|
-
],
|
|
1080
|
-
"models": [
|
|
1081
|
-
"SVD"
|
|
1082
|
-
],
|
|
773
|
+
"tags": ["Imagen a video", "Video"],
|
|
774
|
+
"models": ["SVD"],
|
|
1083
775
|
"date": "2025-03-01"
|
|
1084
776
|
},
|
|
1085
777
|
{
|
|
1086
778
|
"name": "txt_to_image_to_video",
|
|
1087
|
-
"title": "
|
|
779
|
+
"title": "Texto a imagen a video SVD",
|
|
1088
780
|
"mediaType": "image",
|
|
1089
781
|
"mediaSubtype": "webp",
|
|
1090
|
-
"description": "
|
|
782
|
+
"description": "Generar videos creando primero imágenes a partir de indicaciones de texto.",
|
|
1091
783
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/video/#image-to-video",
|
|
1092
|
-
"tags": [
|
|
1093
|
-
|
|
1094
|
-
"Video"
|
|
1095
|
-
],
|
|
1096
|
-
"models": [
|
|
1097
|
-
"SVD"
|
|
1098
|
-
],
|
|
784
|
+
"tags": ["Texto a video", "Video"],
|
|
785
|
+
"models": ["SVD"],
|
|
1099
786
|
"date": "2025-03-01"
|
|
1100
|
-
},
|
|
1101
|
-
{
|
|
1102
|
-
"name": "video_wan2_2_14B_fun_control",
|
|
1103
|
-
"title": "Wan 2.2 14B Fun Control",
|
|
1104
|
-
"description": "Generate videos guided by pose, depth, and edge controls using Wan 2.2 Fun Control.",
|
|
1105
|
-
"mediaType": "image",
|
|
1106
|
-
"mediaSubtype": "webp",
|
|
1107
|
-
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/wan2_2",
|
|
1108
|
-
"tags": [
|
|
1109
|
-
"Video to Video",
|
|
1110
|
-
"Video"
|
|
1111
|
-
],
|
|
1112
|
-
"models": [
|
|
1113
|
-
"Wan2.2"
|
|
1114
|
-
],
|
|
1115
|
-
"date": "2025-08-12"
|
|
1116
|
-
},
|
|
1117
|
-
{
|
|
1118
|
-
"name": "video_wan2_2_14B_fun_inpaint",
|
|
1119
|
-
"title": "Wan 2.2 14B Fun Inp",
|
|
1120
|
-
"description": "Generate videos from start and end frames using Wan 2.2 Fun Inp.",
|
|
1121
|
-
"mediaType": "image",
|
|
1122
|
-
"mediaSubtype": "webp",
|
|
1123
|
-
"tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/wan2_2",
|
|
1124
|
-
"tags": [
|
|
1125
|
-
"FLF2V",
|
|
1126
|
-
"Video"
|
|
1127
|
-
],
|
|
1128
|
-
"models": [
|
|
1129
|
-
"Wan2.2"
|
|
1130
|
-
],
|
|
1131
|
-
"date": "2025-08-12"
|
|
1132
787
|
}
|
|
1133
788
|
]
|
|
1134
789
|
},
|
|
@@ -1143,64 +798,42 @@
|
|
|
1143
798
|
"title": "Stable Audio",
|
|
1144
799
|
"mediaType": "audio",
|
|
1145
800
|
"mediaSubtype": "mp3",
|
|
1146
|
-
"description": "
|
|
1147
|
-
"tags": [
|
|
1148
|
-
|
|
1149
|
-
"Audio"
|
|
1150
|
-
],
|
|
1151
|
-
"models": [
|
|
1152
|
-
"Stable Audio"
|
|
1153
|
-
],
|
|
801
|
+
"description": "Generar audio a partir de indicaciones de texto usando Stable Audio.",
|
|
802
|
+
"tags": ["Texto a audio", "Audio"],
|
|
803
|
+
"models": ["Stable Audio"],
|
|
1154
804
|
"date": "2025-03-01",
|
|
1155
805
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/audio/"
|
|
1156
806
|
},
|
|
1157
807
|
{
|
|
1158
808
|
"name": "audio_ace_step_1_t2a_instrumentals",
|
|
1159
|
-
"title": "ACE-Step v1
|
|
809
|
+
"title": "Música instrumental ACE-Step v1 texto a audio",
|
|
1160
810
|
"mediaType": "audio",
|
|
1161
811
|
"mediaSubtype": "mp3",
|
|
1162
|
-
"description": "
|
|
1163
|
-
"tags": [
|
|
1164
|
-
|
|
1165
|
-
"Audio",
|
|
1166
|
-
"Instrumentals"
|
|
1167
|
-
],
|
|
1168
|
-
"models": [
|
|
1169
|
-
"ACE-Step v1"
|
|
1170
|
-
],
|
|
812
|
+
"description": "Generar música instrumental a partir de indicaciones de texto usando ACE-Step v1.",
|
|
813
|
+
"tags": ["Texto a audio", "Audio", "Instrumentales"],
|
|
814
|
+
"models": ["ACE-Step v1"],
|
|
1171
815
|
"date": "2025-03-01",
|
|
1172
816
|
"tutorialUrl": "https://docs.comfy.org/tutorials/audio/ace-step/ace-step-v1"
|
|
1173
817
|
},
|
|
1174
818
|
{
|
|
1175
819
|
"name": "audio_ace_step_1_t2a_song",
|
|
1176
|
-
"title": "ACE Step v1
|
|
820
|
+
"title": "Canción ACE Step v1 texto a audio",
|
|
1177
821
|
"mediaType": "audio",
|
|
1178
822
|
"mediaSubtype": "mp3",
|
|
1179
|
-
"description": "
|
|
1180
|
-
"tags": [
|
|
1181
|
-
|
|
1182
|
-
"Audio",
|
|
1183
|
-
"Song"
|
|
1184
|
-
],
|
|
1185
|
-
"models": [
|
|
1186
|
-
"ACE-Step v1"
|
|
1187
|
-
],
|
|
823
|
+
"description": "Generar canciones con voces a partir de indicaciones de texto usando ACE-Step v1, soportando personalización multilingüe y de estilo.",
|
|
824
|
+
"tags": ["Texto a audio", "Audio", "Canción"],
|
|
825
|
+
"models": ["ACE-Step v1"],
|
|
1188
826
|
"date": "2025-03-01",
|
|
1189
827
|
"tutorialUrl": "https://docs.comfy.org/tutorials/audio/ace-step/ace-step-v1"
|
|
1190
828
|
},
|
|
1191
829
|
{
|
|
1192
830
|
"name": "audio_ace_step_1_m2m_editing",
|
|
1193
|
-
"title": "ACE Step v1
|
|
831
|
+
"title": "Edición M2M ACE Step v1",
|
|
1194
832
|
"mediaType": "audio",
|
|
1195
833
|
"mediaSubtype": "mp3",
|
|
1196
|
-
"description": "
|
|
1197
|
-
"tags": [
|
|
1198
|
-
|
|
1199
|
-
"Audio"
|
|
1200
|
-
],
|
|
1201
|
-
"models": [
|
|
1202
|
-
"ACE-Step v1"
|
|
1203
|
-
],
|
|
834
|
+
"description": "Editar canciones existentes para cambiar estilo y letras usando ACE-Step v1 M2M.",
|
|
835
|
+
"tags": ["Edición de audio", "Audio"],
|
|
836
|
+
"models": ["ACE-Step v1"],
|
|
1204
837
|
"date": "2025-03-01",
|
|
1205
838
|
"tutorialUrl": "https://docs.comfy.org/tutorials/audio/ace-step/ace-step-v1"
|
|
1206
839
|
}
|
|
@@ -1209,400 +842,259 @@
|
|
|
1209
842
|
{
|
|
1210
843
|
"moduleName": "default",
|
|
1211
844
|
"category": "TOOLS & BUILDING",
|
|
1212
|
-
"title": "API de
|
|
845
|
+
"title": "API de imagen",
|
|
1213
846
|
"type": "image",
|
|
1214
847
|
"templates": [
|
|
1215
848
|
{
|
|
1216
849
|
"name": "api_bfl_flux_1_kontext_multiple_images_input",
|
|
1217
|
-
"title": "BFL Flux.1 Kontext
|
|
1218
|
-
"description": "
|
|
850
|
+
"title": "Entrada de múltiples imágenes BFL Flux.1 Kontext",
|
|
851
|
+
"description": "Ingresar múltiples imágenes y editarlas con Flux.1 Kontext.",
|
|
1219
852
|
"mediaType": "image",
|
|
1220
853
|
"mediaSubtype": "webp",
|
|
1221
854
|
"thumbnailVariant": "compareSlider",
|
|
1222
855
|
"tutorialUrl": "https://docs.comfy.org/tutorials/api-nodes/black-forest-labs/flux-1-kontext",
|
|
1223
|
-
"tags": [
|
|
1224
|
-
|
|
1225
|
-
"Image"
|
|
1226
|
-
],
|
|
1227
|
-
"models": [
|
|
1228
|
-
"Flux"
|
|
1229
|
-
],
|
|
856
|
+
"tags": ["Edición de imagen", "Imagen"],
|
|
857
|
+
"models": ["Flux"],
|
|
1230
858
|
"date": "2025-05-29"
|
|
1231
859
|
},
|
|
1232
860
|
{
|
|
1233
861
|
"name": "api_bfl_flux_1_kontext_pro_image",
|
|
1234
862
|
"title": "BFL Flux.1 Kontext Pro",
|
|
1235
|
-
"description": "
|
|
863
|
+
"description": "Editar imágenes con imagen pro de Flux.1 Kontext.",
|
|
1236
864
|
"mediaType": "image",
|
|
1237
865
|
"mediaSubtype": "webp",
|
|
1238
866
|
"thumbnailVariant": "compareSlider",
|
|
1239
867
|
"tutorialUrl": "https://docs.comfy.org/tutorials/api-nodes/black-forest-labs/flux-1-kontext",
|
|
1240
|
-
"tags": [
|
|
1241
|
-
|
|
1242
|
-
"Image"
|
|
1243
|
-
],
|
|
1244
|
-
"models": [
|
|
1245
|
-
"Flux"
|
|
1246
|
-
],
|
|
868
|
+
"tags": ["Edición de imagen", "Imagen"],
|
|
869
|
+
"models": ["Flux"],
|
|
1247
870
|
"date": "2025-05-29"
|
|
1248
871
|
},
|
|
1249
872
|
{
|
|
1250
873
|
"name": "api_bfl_flux_1_kontext_max_image",
|
|
1251
874
|
"title": "BFL Flux.1 Kontext Max",
|
|
1252
|
-
"description": "
|
|
875
|
+
"description": "Editar imágenes con imagen max de Flux.1 Kontext.",
|
|
1253
876
|
"mediaType": "image",
|
|
1254
877
|
"mediaSubtype": "webp",
|
|
1255
878
|
"thumbnailVariant": "compareSlider",
|
|
1256
879
|
"tutorialUrl": "https://docs.comfy.org/tutorials/api-nodes/black-forest-labs/flux-1-kontext",
|
|
1257
|
-
"tags": [
|
|
1258
|
-
|
|
1259
|
-
"Image"
|
|
1260
|
-
],
|
|
1261
|
-
"models": [
|
|
1262
|
-
"Flux"
|
|
1263
|
-
],
|
|
880
|
+
"tags": ["Edición de imagen", "Imagen"],
|
|
881
|
+
"models": ["Flux"],
|
|
1264
882
|
"date": "2025-05-29"
|
|
1265
883
|
},
|
|
1266
884
|
{
|
|
1267
885
|
"name": "api_bfl_flux_pro_t2i",
|
|
1268
|
-
"title": "BFL Flux[Pro]: Texto a
|
|
1269
|
-
"description": "
|
|
886
|
+
"title": "BFL Flux[Pro]: Texto a imagen",
|
|
887
|
+
"description": "Generar imágenes con excelente seguimiento de indicaciones y calidad visual usando FLUX.1 Pro.",
|
|
1270
888
|
"mediaType": "image",
|
|
1271
889
|
"mediaSubtype": "webp",
|
|
1272
890
|
"tutorialUrl": "https://docs.comfy.org/tutorials/api-nodes/black-forest-labs/flux-1-1-pro-ultra-image",
|
|
1273
|
-
"tags": [
|
|
1274
|
-
|
|
1275
|
-
"Image"
|
|
1276
|
-
],
|
|
1277
|
-
"models": [
|
|
1278
|
-
"Flux"
|
|
1279
|
-
],
|
|
891
|
+
"tags": ["Edición de imagen", "Imagen"],
|
|
892
|
+
"models": ["Flux"],
|
|
1280
893
|
"date": "2025-05-01"
|
|
1281
894
|
},
|
|
1282
895
|
{
|
|
1283
896
|
"name": "api_luma_photon_i2i",
|
|
1284
|
-
"title": "Luma Photon: Imagen a
|
|
1285
|
-
"description": "
|
|
897
|
+
"title": "Luma Photon: Imagen a imagen",
|
|
898
|
+
"description": "Guiar la generación de imágenes usando una combinación de imágenes e indicaciones.",
|
|
1286
899
|
"mediaType": "image",
|
|
1287
900
|
"mediaSubtype": "webp",
|
|
1288
901
|
"thumbnailVariant": "compareSlider",
|
|
1289
|
-
"tags": [
|
|
1290
|
-
|
|
1291
|
-
"Image",
|
|
1292
|
-
"API"
|
|
1293
|
-
],
|
|
1294
|
-
"models": [
|
|
1295
|
-
"Luma Photon"
|
|
1296
|
-
],
|
|
902
|
+
"tags": ["Imagen a imagen", "Imagen", "API"],
|
|
903
|
+
"models": ["Luma Photon"],
|
|
1297
904
|
"date": "2025-03-01"
|
|
1298
905
|
},
|
|
1299
906
|
{
|
|
1300
907
|
"name": "api_luma_photon_style_ref",
|
|
1301
|
-
"title": "Luma Photon: Referencia de
|
|
1302
|
-
"description": "
|
|
908
|
+
"title": "Luma Photon: Referencia de estilo",
|
|
909
|
+
"description": "Generar imágenes mezclando referencias de estilo con control preciso usando Luma Photon.",
|
|
1303
910
|
"mediaType": "image",
|
|
1304
911
|
"mediaSubtype": "webp",
|
|
1305
912
|
"thumbnailVariant": "compareSlider",
|
|
1306
|
-
"tags": [
|
|
1307
|
-
|
|
1308
|
-
"Image",
|
|
1309
|
-
"API",
|
|
1310
|
-
"Style Transfer"
|
|
1311
|
-
],
|
|
1312
|
-
"models": [
|
|
1313
|
-
"Luma Photon"
|
|
1314
|
-
],
|
|
913
|
+
"tags": ["Texto a imagen", "Imagen", "API", "Transferencia de estilo"],
|
|
914
|
+
"models": ["Luma Photon"],
|
|
1315
915
|
"date": "2025-03-01"
|
|
1316
916
|
},
|
|
1317
917
|
{
|
|
1318
918
|
"name": "api_recraft_image_gen_with_color_control",
|
|
1319
|
-
"title": "Recraft: Generación de
|
|
1320
|
-
"description": "
|
|
1321
|
-
"mediaType": "image",
|
|
1322
|
-
"mediaSubtype": "webp",
|
|
1323
|
-
"tags": [
|
|
1324
|
-
|
|
1325
|
-
"Image",
|
|
1326
|
-
"API",
|
|
1327
|
-
"Color Control"
|
|
1328
|
-
],
|
|
1329
|
-
"models": [
|
|
1330
|
-
"Recraft"
|
|
1331
|
-
],
|
|
919
|
+
"title": "Recraft: Generación de imágenes con control de color",
|
|
920
|
+
"description": "Generar imágenes con paletas de colores personalizadas y visuales específicos de marca usando Recraft.",
|
|
921
|
+
"mediaType": "image",
|
|
922
|
+
"mediaSubtype": "webp",
|
|
923
|
+
"tags": ["Texto a imagen", "Imagen", "API", "Control de color"],
|
|
924
|
+
"models": ["Recraft"],
|
|
1332
925
|
"date": "2025-03-01"
|
|
1333
926
|
},
|
|
1334
927
|
{
|
|
1335
928
|
"name": "api_recraft_image_gen_with_style_control",
|
|
1336
|
-
"title": "Recraft: Generación de
|
|
1337
|
-
"description": "
|
|
1338
|
-
"mediaType": "image",
|
|
1339
|
-
"mediaSubtype": "webp",
|
|
1340
|
-
"tags": [
|
|
1341
|
-
|
|
1342
|
-
"Image",
|
|
1343
|
-
"API",
|
|
1344
|
-
"Style Control"
|
|
1345
|
-
],
|
|
1346
|
-
"models": [
|
|
1347
|
-
"Recraft"
|
|
1348
|
-
],
|
|
929
|
+
"title": "Recraft: Generación de imágenes con control de estilo",
|
|
930
|
+
"description": "Controlar estilo con ejemplos visuales, alinear posicionamiento y ajustar objetos finamente. Almacenar y compartir estilos para consistencia perfecta de marca.",
|
|
931
|
+
"mediaType": "image",
|
|
932
|
+
"mediaSubtype": "webp",
|
|
933
|
+
"tags": ["Texto a imagen", "Imagen", "API", "Control de estilo"],
|
|
934
|
+
"models": ["Recraft"],
|
|
1349
935
|
"date": "2025-03-01"
|
|
1350
936
|
},
|
|
1351
937
|
{
|
|
1352
938
|
"name": "api_recraft_vector_gen",
|
|
1353
|
-
"title": "Recraft: Generación
|
|
1354
|
-
"description": "
|
|
1355
|
-
"mediaType": "image",
|
|
1356
|
-
"mediaSubtype": "webp",
|
|
1357
|
-
"tags": [
|
|
1358
|
-
|
|
1359
|
-
"Image",
|
|
1360
|
-
"API",
|
|
1361
|
-
"Vector"
|
|
1362
|
-
],
|
|
1363
|
-
"models": [
|
|
1364
|
-
"Recraft"
|
|
1365
|
-
],
|
|
939
|
+
"title": "Recraft: Generación vectorial",
|
|
940
|
+
"description": "Generar imágenes vectoriales de alta calidad a partir de indicaciones de texto usando el generador vectorial AI de Recraft.",
|
|
941
|
+
"mediaType": "image",
|
|
942
|
+
"mediaSubtype": "webp",
|
|
943
|
+
"tags": ["Texto a imagen", "Imagen", "API", "Vectorial"],
|
|
944
|
+
"models": ["Recraft"],
|
|
1366
945
|
"date": "2025-03-01"
|
|
1367
946
|
},
|
|
1368
947
|
{
|
|
1369
948
|
"name": "api_runway_text_to_image",
|
|
1370
|
-
"title": "Runway: Texto a
|
|
1371
|
-
"description": "
|
|
1372
|
-
"mediaType": "image",
|
|
1373
|
-
"mediaSubtype": "webp",
|
|
1374
|
-
"tags": [
|
|
1375
|
-
|
|
1376
|
-
"Image",
|
|
1377
|
-
"API"
|
|
1378
|
-
],
|
|
1379
|
-
"models": [
|
|
1380
|
-
"Runway"
|
|
1381
|
-
],
|
|
949
|
+
"title": "Runway: Texto a imagen",
|
|
950
|
+
"description": "Generar imágenes de alta calidad a partir de indicaciones de texto usando el modelo AI de Runway.",
|
|
951
|
+
"mediaType": "image",
|
|
952
|
+
"mediaSubtype": "webp",
|
|
953
|
+
"tags": ["Texto a imagen", "Imagen", "API"],
|
|
954
|
+
"models": ["Runway"],
|
|
1382
955
|
"date": "2025-03-01"
|
|
1383
956
|
},
|
|
1384
957
|
{
|
|
1385
958
|
"name": "api_runway_reference_to_image",
|
|
1386
|
-
"title": "Runway: Referencia a
|
|
1387
|
-
"description": "
|
|
959
|
+
"title": "Runway: Referencia a imagen",
|
|
960
|
+
"description": "Generar nuevas imágenes basadas en estilos y composiciones de referencia con la IA de Runway.",
|
|
1388
961
|
"mediaType": "image",
|
|
1389
962
|
"thumbnailVariant": "compareSlider",
|
|
1390
963
|
"mediaSubtype": "webp",
|
|
1391
|
-
"tags": [
|
|
1392
|
-
|
|
1393
|
-
"Image",
|
|
1394
|
-
"API",
|
|
1395
|
-
"Style Transfer"
|
|
1396
|
-
],
|
|
1397
|
-
"models": [
|
|
1398
|
-
"Runway"
|
|
1399
|
-
],
|
|
964
|
+
"tags": ["Imagen a imagen", "Imagen", "API", "Transferencia de estilo"],
|
|
965
|
+
"models": ["Runway"],
|
|
1400
966
|
"date": "2025-03-01"
|
|
1401
967
|
},
|
|
1402
968
|
{
|
|
1403
969
|
"name": "api_stability_ai_stable_image_ultra_t2i",
|
|
1404
|
-
"title": "Stability AI: Stable Image Ultra
|
|
1405
|
-
"description": "
|
|
1406
|
-
"mediaType": "image",
|
|
1407
|
-
"mediaSubtype": "webp",
|
|
1408
|
-
"tags": [
|
|
1409
|
-
|
|
1410
|
-
"Image",
|
|
1411
|
-
"API"
|
|
1412
|
-
],
|
|
1413
|
-
"models": [
|
|
1414
|
-
"Stable Image Ultra"
|
|
1415
|
-
],
|
|
970
|
+
"title": "Stability AI: Texto a imagen Stable Image Ultra",
|
|
971
|
+
"description": "Generar imágenes de alta calidad con excelente adherencia a las indicaciones. Perfecto para casos de uso profesionales en resolución de 1 megapíxel.",
|
|
972
|
+
"mediaType": "image",
|
|
973
|
+
"mediaSubtype": "webp",
|
|
974
|
+
"tags": ["Texto a imagen", "Imagen", "API"],
|
|
975
|
+
"models": ["Stable Image Ultra"],
|
|
1416
976
|
"date": "2025-03-01"
|
|
1417
977
|
},
|
|
1418
978
|
{
|
|
1419
979
|
"name": "api_stability_ai_i2i",
|
|
1420
|
-
"title": "Stability AI: Imagen a
|
|
1421
|
-
"description": "
|
|
980
|
+
"title": "Stability AI: Imagen a imagen",
|
|
981
|
+
"description": "Transformar imágenes con generación de alta calidad usando Stability AI, perfecto para edición profesional y transferencia de estilo.",
|
|
1422
982
|
"mediaType": "image",
|
|
1423
983
|
"thumbnailVariant": "compareSlider",
|
|
1424
984
|
"mediaSubtype": "webp",
|
|
1425
|
-
"tags": [
|
|
1426
|
-
|
|
1427
|
-
"Image",
|
|
1428
|
-
"API"
|
|
1429
|
-
],
|
|
1430
|
-
"models": [
|
|
1431
|
-
"Stability AI"
|
|
1432
|
-
],
|
|
985
|
+
"tags": ["Imagen a imagen", "Imagen", "API"],
|
|
986
|
+
"models": ["Stability AI"],
|
|
1433
987
|
"date": "2025-03-01"
|
|
1434
988
|
},
|
|
1435
989
|
{
|
|
1436
990
|
"name": "api_stability_ai_sd3.5_t2i",
|
|
1437
|
-
"title": "Stability AI: SD3.5
|
|
1438
|
-
"description": "
|
|
1439
|
-
"mediaType": "image",
|
|
1440
|
-
"mediaSubtype": "webp",
|
|
1441
|
-
"tags": [
|
|
1442
|
-
|
|
1443
|
-
"Image",
|
|
1444
|
-
"API"
|
|
1445
|
-
],
|
|
1446
|
-
"models": [
|
|
1447
|
-
"SD3.5"
|
|
1448
|
-
],
|
|
991
|
+
"title": "Stability AI: Texto a imagen SD3.5",
|
|
992
|
+
"description": "Generar imágenes de alta calidad con excelente adherencia a las indicaciones. Perfecto para casos de uso profesionales en resolución de 1 megapíxel.",
|
|
993
|
+
"mediaType": "image",
|
|
994
|
+
"mediaSubtype": "webp",
|
|
995
|
+
"tags": ["Texto a imagen", "Imagen", "API"],
|
|
996
|
+
"models": ["SD3.5"],
|
|
1449
997
|
"date": "2025-03-01"
|
|
1450
998
|
},
|
|
1451
999
|
{
|
|
1452
1000
|
"name": "api_stability_ai_sd3.5_i2i",
|
|
1453
|
-
"title": "Stability AI: SD3.5
|
|
1454
|
-
"description": "
|
|
1001
|
+
"title": "Stability AI: Imagen a imagen SD3.5",
|
|
1002
|
+
"description": "Generar imágenes de alta calidad con excelente adherencia a las indicaciones. Perfecto para casos de uso profesionales en resolución de 1 megapíxel.",
|
|
1455
1003
|
"mediaType": "image",
|
|
1456
1004
|
"thumbnailVariant": "compareSlider",
|
|
1457
1005
|
"mediaSubtype": "webp",
|
|
1458
|
-
"tags": [
|
|
1459
|
-
|
|
1460
|
-
"Image",
|
|
1461
|
-
"API"
|
|
1462
|
-
],
|
|
1463
|
-
"models": [
|
|
1464
|
-
"SD3.5"
|
|
1465
|
-
],
|
|
1006
|
+
"tags": ["Imagen a imagen", "Imagen", "API"],
|
|
1007
|
+
"models": ["SD3.5"],
|
|
1466
1008
|
"date": "2025-03-01"
|
|
1467
1009
|
},
|
|
1468
1010
|
{
|
|
1469
1011
|
"name": "api_ideogram_v3_t2i",
|
|
1470
|
-
"title": "Ideogram V3: Texto a
|
|
1471
|
-
"description": "
|
|
1472
|
-
"mediaType": "image",
|
|
1473
|
-
"mediaSubtype": "webp",
|
|
1474
|
-
"tags": [
|
|
1475
|
-
|
|
1476
|
-
"Image",
|
|
1477
|
-
"API",
|
|
1478
|
-
"Text Rendering"
|
|
1479
|
-
],
|
|
1480
|
-
"models": [
|
|
1481
|
-
"Ideogram V3"
|
|
1482
|
-
],
|
|
1012
|
+
"title": "Ideogram V3: Texto a imagen",
|
|
1013
|
+
"description": "Generar imágenes de calidad profesional con excelente alineación de indicaciones, fotorrealismo y renderizado de texto usando Ideogram V3.",
|
|
1014
|
+
"mediaType": "image",
|
|
1015
|
+
"mediaSubtype": "webp",
|
|
1016
|
+
"tags": ["Texto a imagen", "Imagen", "API", "Renderizado de texto"],
|
|
1017
|
+
"models": ["Ideogram V3"],
|
|
1483
1018
|
"date": "2025-03-01"
|
|
1484
1019
|
},
|
|
1485
1020
|
{
|
|
1486
1021
|
"name": "api_openai_image_1_t2i",
|
|
1487
|
-
"title": "OpenAI: GPT-Image-1
|
|
1488
|
-
"description": "
|
|
1489
|
-
"mediaType": "image",
|
|
1490
|
-
"mediaSubtype": "webp",
|
|
1491
|
-
"tags": [
|
|
1492
|
-
|
|
1493
|
-
"Image",
|
|
1494
|
-
"API"
|
|
1495
|
-
],
|
|
1496
|
-
"models": [
|
|
1497
|
-
"GPT-Image-1"
|
|
1498
|
-
],
|
|
1022
|
+
"title": "OpenAI: Texto a imagen GPT-Image-1",
|
|
1023
|
+
"description": "Generar imágenes a partir de indicaciones de texto usando la API de OpenAI GPT Image 1.",
|
|
1024
|
+
"mediaType": "image",
|
|
1025
|
+
"mediaSubtype": "webp",
|
|
1026
|
+
"tags": ["Texto a imagen", "Imagen", "API"],
|
|
1027
|
+
"models": ["GPT-Image-1"],
|
|
1499
1028
|
"date": "2025-03-01",
|
|
1500
1029
|
"tutorialUrl": "https://docs.comfy.org/tutorials/api-nodes/openai/gpt-image-1"
|
|
1501
1030
|
},
|
|
1502
1031
|
{
|
|
1503
1032
|
"name": "api_openai_image_1_i2i",
|
|
1504
|
-
"title": "OpenAI: GPT-Image-1
|
|
1505
|
-
"description": "
|
|
1033
|
+
"title": "OpenAI: Imagen a imagen GPT-Image-1",
|
|
1034
|
+
"description": "Generar imágenes a partir de imágenes de entrada usando la API de OpenAI GPT Image 1.",
|
|
1506
1035
|
"mediaType": "image",
|
|
1507
1036
|
"mediaSubtype": "webp",
|
|
1508
1037
|
"thumbnailVariant": "compareSlider",
|
|
1509
|
-
"tags": [
|
|
1510
|
-
|
|
1511
|
-
"Image",
|
|
1512
|
-
"API"
|
|
1513
|
-
],
|
|
1514
|
-
"models": [
|
|
1515
|
-
"GPT-Image-1"
|
|
1516
|
-
],
|
|
1038
|
+
"tags": ["Imagen a imagen", "Imagen", "API"],
|
|
1039
|
+
"models": ["GPT-Image-1"],
|
|
1517
1040
|
"date": "2025-03-01",
|
|
1518
1041
|
"tutorialUrl": "https://docs.comfy.org/tutorials/api-nodes/openai/gpt-image-1"
|
|
1519
1042
|
},
|
|
1520
1043
|
{
|
|
1521
1044
|
"name": "api_openai_image_1_inpaint",
|
|
1522
|
-
"title": "OpenAI: GPT-Image-1
|
|
1523
|
-
"description": "
|
|
1045
|
+
"title": "OpenAI: Inpaint GPT-Image-1",
|
|
1046
|
+
"description": "Editar imágenes usando inpainting con la API de OpenAI GPT Image 1.",
|
|
1524
1047
|
"mediaType": "image",
|
|
1525
1048
|
"mediaSubtype": "webp",
|
|
1526
1049
|
"thumbnailVariant": "compareSlider",
|
|
1527
|
-
"tags": [
|
|
1528
|
-
|
|
1529
|
-
"Image",
|
|
1530
|
-
"API"
|
|
1531
|
-
],
|
|
1532
|
-
"models": [
|
|
1533
|
-
"GPT-Image-1"
|
|
1534
|
-
],
|
|
1050
|
+
"tags": ["Inpaint", "Imagen", "API"],
|
|
1051
|
+
"models": ["GPT-Image-1"],
|
|
1535
1052
|
"date": "2025-03-01",
|
|
1536
1053
|
"tutorialUrl": "https://docs.comfy.org/tutorials/api-nodes/openai/gpt-image-1"
|
|
1537
1054
|
},
|
|
1538
1055
|
{
|
|
1539
1056
|
"name": "api_openai_image_1_multi_inputs",
|
|
1540
|
-
"title": "OpenAI: GPT-Image-1
|
|
1541
|
-
"description": "
|
|
1057
|
+
"title": "OpenAI: Múltiples entradas GPT-Image-1",
|
|
1058
|
+
"description": "Generar imágenes a partir de múltiples entradas usando la API de OpenAI GPT Image 1.",
|
|
1542
1059
|
"mediaType": "image",
|
|
1543
1060
|
"mediaSubtype": "webp",
|
|
1544
1061
|
"thumbnailVariant": "compareSlider",
|
|
1545
|
-
"tags": [
|
|
1546
|
-
|
|
1547
|
-
"Image",
|
|
1548
|
-
"API",
|
|
1549
|
-
"Multi Input"
|
|
1550
|
-
],
|
|
1551
|
-
"models": [
|
|
1552
|
-
"GPT-Image-1"
|
|
1553
|
-
],
|
|
1062
|
+
"tags": ["Texto a imagen", "Imagen", "API", "Entrada múltiple"],
|
|
1063
|
+
"models": ["GPT-Image-1"],
|
|
1554
1064
|
"date": "2025-03-01",
|
|
1555
1065
|
"tutorialUrl": "https://docs.comfy.org/tutorials/api-nodes/openai/gpt-image-1"
|
|
1556
1066
|
},
|
|
1557
1067
|
{
|
|
1558
1068
|
"name": "api_openai_dall_e_2_t2i",
|
|
1559
|
-
"title": "OpenAI: Dall-E 2
|
|
1560
|
-
"description": "
|
|
1561
|
-
"mediaType": "image",
|
|
1562
|
-
"mediaSubtype": "webp",
|
|
1563
|
-
"tags": [
|
|
1564
|
-
|
|
1565
|
-
"Image",
|
|
1566
|
-
"API"
|
|
1567
|
-
],
|
|
1568
|
-
"models": [
|
|
1569
|
-
"Dall-E 2"
|
|
1570
|
-
],
|
|
1069
|
+
"title": "OpenAI: Texto a imagen Dall-E 2",
|
|
1070
|
+
"description": "Generar imágenes a partir de indicaciones de texto usando la API de OpenAI Dall-E 2.",
|
|
1071
|
+
"mediaType": "image",
|
|
1072
|
+
"mediaSubtype": "webp",
|
|
1073
|
+
"tags": ["Texto a imagen", "Imagen", "API"],
|
|
1074
|
+
"models": ["Dall-E 2"],
|
|
1571
1075
|
"date": "2025-03-01",
|
|
1572
1076
|
"tutorialUrl": "https://docs.comfy.org/tutorials/api-nodes/openai/dall-e-2"
|
|
1573
1077
|
},
|
|
1574
1078
|
{
|
|
1575
1079
|
"name": "api_openai_dall_e_2_inpaint",
|
|
1576
|
-
"title": "OpenAI: Dall-E 2
|
|
1577
|
-
"description": "
|
|
1080
|
+
"title": "OpenAI: Inpaint Dall-E 2",
|
|
1081
|
+
"description": "Editar imágenes usando inpainting con la API de OpenAI Dall-E 2.",
|
|
1578
1082
|
"mediaType": "image",
|
|
1579
1083
|
"mediaSubtype": "webp",
|
|
1580
1084
|
"thumbnailVariant": "compareSlider",
|
|
1581
|
-
"tags": [
|
|
1582
|
-
|
|
1583
|
-
"Image",
|
|
1584
|
-
"API"
|
|
1585
|
-
],
|
|
1586
|
-
"models": [
|
|
1587
|
-
"Dall-E 2"
|
|
1588
|
-
],
|
|
1085
|
+
"tags": ["Inpaint", "Imagen", "API"],
|
|
1086
|
+
"models": ["Dall-E 2"],
|
|
1589
1087
|
"date": "2025-03-01",
|
|
1590
1088
|
"tutorialUrl": "https://docs.comfy.org/tutorials/api-nodes/openai/dall-e-2"
|
|
1591
1089
|
},
|
|
1592
1090
|
{
|
|
1593
1091
|
"name": "api_openai_dall_e_3_t2i",
|
|
1594
|
-
"title": "OpenAI: Dall-E 3
|
|
1595
|
-
"description": "
|
|
1596
|
-
"mediaType": "image",
|
|
1597
|
-
"mediaSubtype": "webp",
|
|
1598
|
-
"tags": [
|
|
1599
|
-
|
|
1600
|
-
"Image",
|
|
1601
|
-
"API"
|
|
1602
|
-
],
|
|
1603
|
-
"models": [
|
|
1604
|
-
"Dall-E 3"
|
|
1605
|
-
],
|
|
1092
|
+
"title": "OpenAI: Texto a imagen Dall-E 3",
|
|
1093
|
+
"description": "Generar imágenes a partir de indicaciones de texto usando la API de OpenAI Dall-E 3.",
|
|
1094
|
+
"mediaType": "image",
|
|
1095
|
+
"mediaSubtype": "webp",
|
|
1096
|
+
"tags": ["Texto a imagen", "Imagen", "API"],
|
|
1097
|
+
"models": ["Dall-E 3"],
|
|
1606
1098
|
"date": "2025-03-01",
|
|
1607
1099
|
"tutorialUrl": "https://docs.comfy.org/tutorials/api-nodes/openai/dall-e-3"
|
|
1608
1100
|
}
|
|
@@ -1611,354 +1103,294 @@
|
|
|
1611
1103
|
{
|
|
1612
1104
|
"moduleName": "default",
|
|
1613
1105
|
"category": "TOOLS & BUILDING",
|
|
1614
|
-
"title": "API de
|
|
1106
|
+
"title": "API de video",
|
|
1615
1107
|
"type": "video",
|
|
1616
1108
|
"templates": [
|
|
1617
1109
|
{
|
|
1618
1110
|
"name": "api_kling_i2v",
|
|
1619
|
-
"title": "Kling: Imagen a
|
|
1620
|
-
"description": "
|
|
1621
|
-
"mediaType": "image",
|
|
1622
|
-
"mediaSubtype": "webp",
|
|
1623
|
-
"tags": [
|
|
1624
|
-
|
|
1625
|
-
"Video",
|
|
1626
|
-
"API"
|
|
1627
|
-
],
|
|
1628
|
-
"models": [
|
|
1629
|
-
"Kling"
|
|
1630
|
-
],
|
|
1111
|
+
"title": "Kling: Imagen a video",
|
|
1112
|
+
"description": "Generar videos con excelente adherencia a las indicaciones para acciones, expresiones y movimientos de cámara usando Kling.",
|
|
1113
|
+
"mediaType": "image",
|
|
1114
|
+
"mediaSubtype": "webp",
|
|
1115
|
+
"tags": ["Imagen a video", "Video", "API"],
|
|
1116
|
+
"models": ["Kling"],
|
|
1631
1117
|
"date": "2025-03-01",
|
|
1632
1118
|
"tutorialUrl": ""
|
|
1633
1119
|
},
|
|
1634
1120
|
{
|
|
1635
1121
|
"name": "api_kling_effects",
|
|
1636
|
-
"title": "Kling: Efectos de
|
|
1637
|
-
"description": "
|
|
1638
|
-
"mediaType": "image",
|
|
1639
|
-
"mediaSubtype": "webp",
|
|
1640
|
-
"tags": [
|
|
1641
|
-
|
|
1642
|
-
"Video",
|
|
1643
|
-
"API"
|
|
1644
|
-
],
|
|
1645
|
-
"models": [
|
|
1646
|
-
"Kling"
|
|
1647
|
-
],
|
|
1122
|
+
"title": "Kling: Efectos de video",
|
|
1123
|
+
"description": "Generar videos dinámicos aplicando efectos visuales a imágenes usando Kling.",
|
|
1124
|
+
"mediaType": "image",
|
|
1125
|
+
"mediaSubtype": "webp",
|
|
1126
|
+
"tags": ["Efectos de video", "Video", "API"],
|
|
1127
|
+
"models": ["Kling"],
|
|
1648
1128
|
"date": "2025-03-01",
|
|
1649
1129
|
"tutorialUrl": ""
|
|
1650
1130
|
},
|
|
1651
1131
|
{
|
|
1652
1132
|
"name": "api_kling_flf",
|
|
1653
1133
|
"title": "Kling: FLF2V",
|
|
1654
|
-
"description": "
|
|
1655
|
-
"mediaType": "image",
|
|
1656
|
-
"mediaSubtype": "webp",
|
|
1657
|
-
"tags": [
|
|
1658
|
-
|
|
1659
|
-
"Video",
|
|
1660
|
-
"API",
|
|
1661
|
-
"Frame Control"
|
|
1662
|
-
],
|
|
1663
|
-
"models": [
|
|
1664
|
-
"Kling"
|
|
1665
|
-
],
|
|
1134
|
+
"description": "Generar videos controlando el primer y último fotograma.",
|
|
1135
|
+
"mediaType": "image",
|
|
1136
|
+
"mediaSubtype": "webp",
|
|
1137
|
+
"tags": ["Generación de video", "Video", "API", "Control de fotograma"],
|
|
1138
|
+
"models": ["Kling"],
|
|
1666
1139
|
"date": "2025-03-01",
|
|
1667
1140
|
"tutorialUrl": ""
|
|
1668
1141
|
},
|
|
1142
|
+
{
|
|
1143
|
+
"name": "api_vidu_text_to_video",
|
|
1144
|
+
"title": "Vidu: Texto a video",
|
|
1145
|
+
"description": "Generar videos 1080p de alta calidad a partir de indicaciones de texto con control de amplitud de movimiento y duración ajustable usando el modelo AI avanzado de Vidu.",
|
|
1146
|
+
"mediaType": "image",
|
|
1147
|
+
"mediaSubtype": "webp",
|
|
1148
|
+
"tags": ["Texto a video", "Video", "API"],
|
|
1149
|
+
"models": ["Vidu"],
|
|
1150
|
+
"date": "2025-08-23",
|
|
1151
|
+
"tutorialUrl": ""
|
|
1152
|
+
},
|
|
1153
|
+
{
|
|
1154
|
+
"name": "api_vidu_image_to_video",
|
|
1155
|
+
"title": "Vidu: Imagen a video",
|
|
1156
|
+
"description": "Transformar imágenes estáticas en videos 1080p dinámicos con control de movimiento preciso y amplitud de movimiento personalizable usando Vidu.",
|
|
1157
|
+
"mediaType": "image",
|
|
1158
|
+
"mediaSubtype": "webp",
|
|
1159
|
+
"tags": ["Imagen a video", "Video", "API"],
|
|
1160
|
+
"models": ["Vidu"],
|
|
1161
|
+
"date": "2025-08-23",
|
|
1162
|
+
"tutorialUrl": ""
|
|
1163
|
+
},
|
|
1164
|
+
{
|
|
1165
|
+
"name": "api_vidu_reference_to_video",
|
|
1166
|
+
"title": "Vidu: Referencia a video",
|
|
1167
|
+
"description": "Generar videos con sujetos consistentes usando múltiples imágenes de referencia (hasta 7) para continuidad de personaje y estilo a lo largo de la secuencia de video.",
|
|
1168
|
+
"mediaType": "image",
|
|
1169
|
+
"mediaSubtype": "webp",
|
|
1170
|
+
"tags": ["Referencia a video", "Video", "API"],
|
|
1171
|
+
"models": ["Vidu"],
|
|
1172
|
+
"date": "2025-08-23",
|
|
1173
|
+
"tutorialUrl": ""
|
|
1174
|
+
},
|
|
1175
|
+
{
|
|
1176
|
+
"name": "api_vidu_start_end_to_video",
|
|
1177
|
+
"title": "Vidu: Inicio fin a video",
|
|
1178
|
+
"description": "Crear transiciones de video suaves entre fotogramas de inicio y fin definidos con interpolación natural de movimiento y calidad visual consistente.",
|
|
1179
|
+
"mediaType": "image",
|
|
1180
|
+
"mediaSubtype": "webp",
|
|
1181
|
+
"tags": ["FLF2V", "Video", "API"],
|
|
1182
|
+
"models": ["Vidu"],
|
|
1183
|
+
"date": "2025-08-23",
|
|
1184
|
+
"tutorialUrl": ""
|
|
1185
|
+
},
|
|
1669
1186
|
{
|
|
1670
1187
|
"name": "api_luma_i2v",
|
|
1671
|
-
"title": "Luma: Imagen a
|
|
1672
|
-
"description": "
|
|
1673
|
-
"mediaType": "image",
|
|
1674
|
-
"mediaSubtype": "webp",
|
|
1675
|
-
"tags": [
|
|
1676
|
-
|
|
1677
|
-
"Video",
|
|
1678
|
-
"API"
|
|
1679
|
-
],
|
|
1680
|
-
"models": [
|
|
1681
|
-
"Luma"
|
|
1682
|
-
],
|
|
1188
|
+
"title": "Luma: Imagen a video",
|
|
1189
|
+
"description": "Tomar imágenes estáticas e instantáneamente crear animaciones mágicas de alta calidad.",
|
|
1190
|
+
"mediaType": "image",
|
|
1191
|
+
"mediaSubtype": "webp",
|
|
1192
|
+
"tags": ["Imagen a video", "Video", "API"],
|
|
1193
|
+
"models": ["Luma"],
|
|
1683
1194
|
"date": "2025-03-01",
|
|
1684
1195
|
"tutorialUrl": ""
|
|
1685
1196
|
},
|
|
1686
1197
|
{
|
|
1687
1198
|
"name": "api_luma_t2v",
|
|
1688
|
-
"title": "Luma: Texto a
|
|
1689
|
-
"description": "
|
|
1690
|
-
"mediaType": "image",
|
|
1691
|
-
"mediaSubtype": "webp",
|
|
1692
|
-
"tags": [
|
|
1693
|
-
|
|
1694
|
-
"Video",
|
|
1695
|
-
"API"
|
|
1696
|
-
],
|
|
1697
|
-
"models": [
|
|
1698
|
-
"Luma"
|
|
1699
|
-
],
|
|
1199
|
+
"title": "Luma: Texto a video",
|
|
1200
|
+
"description": "Se pueden generar videos de alta calidad usando indicaciones simples.",
|
|
1201
|
+
"mediaType": "image",
|
|
1202
|
+
"mediaSubtype": "webp",
|
|
1203
|
+
"tags": ["Texto a video", "Video", "API"],
|
|
1204
|
+
"models": ["Luma"],
|
|
1700
1205
|
"date": "2025-03-01",
|
|
1701
1206
|
"tutorialUrl": ""
|
|
1702
1207
|
},
|
|
1703
1208
|
{
|
|
1704
1209
|
"name": "api_moonvalley_text_to_video",
|
|
1705
|
-
"title": "Moonvalley: Texto a
|
|
1706
|
-
"description": "
|
|
1707
|
-
"mediaType": "image",
|
|
1708
|
-
"mediaSubtype": "webp",
|
|
1709
|
-
"tags": [
|
|
1710
|
-
|
|
1711
|
-
"Video",
|
|
1712
|
-
"API"
|
|
1713
|
-
],
|
|
1714
|
-
"models": [
|
|
1715
|
-
"Moonvalley"
|
|
1716
|
-
],
|
|
1210
|
+
"title": "Moonvalley: Texto a video",
|
|
1211
|
+
"description": "Generar videos cinematográficos 1080p a partir de indicaciones de texto mediante un modelo entrenado exclusivamente en datos con licencia.",
|
|
1212
|
+
"mediaType": "image",
|
|
1213
|
+
"mediaSubtype": "webp",
|
|
1214
|
+
"tags": ["Texto a video", "Video", "API"],
|
|
1215
|
+
"models": ["Moonvalley"],
|
|
1717
1216
|
"date": "2025-03-01",
|
|
1718
1217
|
"tutorialUrl": ""
|
|
1719
1218
|
},
|
|
1720
1219
|
{
|
|
1721
1220
|
"name": "api_moonvalley_image_to_video",
|
|
1722
|
-
"title": "Moonvalley: Imagen a
|
|
1723
|
-
"description": "
|
|
1724
|
-
"mediaType": "image",
|
|
1725
|
-
"mediaSubtype": "webp",
|
|
1726
|
-
"tags": [
|
|
1727
|
-
|
|
1728
|
-
"Video",
|
|
1729
|
-
"API"
|
|
1730
|
-
],
|
|
1731
|
-
"models": [
|
|
1732
|
-
"Moonvalley"
|
|
1733
|
-
],
|
|
1221
|
+
"title": "Moonvalley: Imagen a video",
|
|
1222
|
+
"description": "Generar videos cinematográficos 1080p con una imagen mediante un modelo entrenado exclusivamente en datos con licencia.",
|
|
1223
|
+
"mediaType": "image",
|
|
1224
|
+
"mediaSubtype": "webp",
|
|
1225
|
+
"tags": ["Imagen a video", "Video", "API"],
|
|
1226
|
+
"models": ["Moonvalley"],
|
|
1734
1227
|
"date": "2025-03-01",
|
|
1735
1228
|
"tutorialUrl": ""
|
|
1736
1229
|
},
|
|
1737
1230
|
{
|
|
1738
1231
|
"name": "api_moonvalley_video_to_video_motion_transfer",
|
|
1739
|
-
"title": "Moonvalley:
|
|
1740
|
-
"description": "
|
|
1232
|
+
"title": "Moonvalley: Transferencia de movimiento",
|
|
1233
|
+
"description": "Aplicar movimiento de un video a otro.",
|
|
1741
1234
|
"mediaType": "image",
|
|
1742
1235
|
"thumbnailVariant": "hoverDissolve",
|
|
1743
1236
|
"mediaSubtype": "webp",
|
|
1744
|
-
"tags": [
|
|
1745
|
-
|
|
1746
|
-
"Video",
|
|
1747
|
-
"API",
|
|
1748
|
-
"Motion Transfer"
|
|
1749
|
-
],
|
|
1750
|
-
"models": [
|
|
1751
|
-
"Moonvalley"
|
|
1752
|
-
],
|
|
1237
|
+
"tags": ["Video a video", "Video", "API", "Transferencia de movimiento"],
|
|
1238
|
+
"models": ["Moonvalley"],
|
|
1753
1239
|
"date": "2025-03-01",
|
|
1754
1240
|
"tutorialUrl": ""
|
|
1755
1241
|
},
|
|
1756
1242
|
{
|
|
1757
1243
|
"name": "api_moonvalley_video_to_video_pose_control",
|
|
1758
|
-
"title": "Moonvalley:
|
|
1759
|
-
"description": "
|
|
1244
|
+
"title": "Moonvalley: Control de pose",
|
|
1245
|
+
"description": "Aplicar pose y movimiento humano de un video a otro.",
|
|
1760
1246
|
"mediaType": "image",
|
|
1761
1247
|
"thumbnailVariant": "hoverDissolve",
|
|
1762
1248
|
"mediaSubtype": "webp",
|
|
1763
|
-
"tags": [
|
|
1764
|
-
|
|
1765
|
-
|
|
1766
|
-
|
|
1767
|
-
|
|
1768
|
-
|
|
1769
|
-
"
|
|
1770
|
-
|
|
1771
|
-
|
|
1249
|
+
"tags": ["Video a video", "Video", "API", "Control de pose"],
|
|
1250
|
+
"models": ["Moonvalley"],
|
|
1251
|
+
"date": "2025-03-01",
|
|
1252
|
+
"tutorialUrl": ""
|
|
1253
|
+
},
|
|
1254
|
+
{
|
|
1255
|
+
"name": "api_hailuo_minimax_video",
|
|
1256
|
+
"title": "MiniMax: Video",
|
|
1257
|
+
"description": "Generar videos de alta calidad a partir de indicaciones de texto con control opcional del primer fotograma usando el modelo MiniMax Hailuo-02. Soporta múltiples resoluciones (768P/1080P) y duraciones (6/10s) con optimización inteligente de indicaciones.",
|
|
1258
|
+
"mediaType": "image",
|
|
1259
|
+
"mediaSubtype": "webp",
|
|
1260
|
+
"tags": ["Texto a video", "Video", "API"],
|
|
1261
|
+
"models": ["MiniMax"],
|
|
1772
1262
|
"date": "2025-03-01",
|
|
1773
1263
|
"tutorialUrl": ""
|
|
1774
1264
|
},
|
|
1775
1265
|
{
|
|
1776
1266
|
"name": "api_hailuo_minimax_t2v",
|
|
1777
|
-
"title": "MiniMax: Texto a
|
|
1778
|
-
"description": "
|
|
1779
|
-
"mediaType": "image",
|
|
1780
|
-
"mediaSubtype": "webp",
|
|
1781
|
-
"tags": [
|
|
1782
|
-
|
|
1783
|
-
"Video",
|
|
1784
|
-
"API"
|
|
1785
|
-
],
|
|
1786
|
-
"models": [
|
|
1787
|
-
"MiniMax"
|
|
1788
|
-
],
|
|
1267
|
+
"title": "MiniMax: Texto a video",
|
|
1268
|
+
"description": "Generar videos de alta calidad directamente a partir de indicaciones de texto. Explorar las capacidades avanzadas de IA de MiniMax para crear narrativas visuales diversas con efectos CGI profesionales y elementos estilísticos que den vida a sus descripciones.",
|
|
1269
|
+
"mediaType": "image",
|
|
1270
|
+
"mediaSubtype": "webp",
|
|
1271
|
+
"tags": ["Texto a video", "Video", "API"],
|
|
1272
|
+
"models": ["MiniMax"],
|
|
1789
1273
|
"date": "2025-03-01",
|
|
1790
1274
|
"tutorialUrl": ""
|
|
1791
1275
|
},
|
|
1792
1276
|
{
|
|
1793
1277
|
"name": "api_hailuo_minimax_i2v",
|
|
1794
|
-
"title": "MiniMax: Imagen a
|
|
1795
|
-
"description": "
|
|
1796
|
-
"mediaType": "image",
|
|
1797
|
-
"mediaSubtype": "webp",
|
|
1798
|
-
"tags": [
|
|
1799
|
-
|
|
1800
|
-
"Video",
|
|
1801
|
-
"API"
|
|
1802
|
-
],
|
|
1803
|
-
"models": [
|
|
1804
|
-
"MiniMax"
|
|
1805
|
-
],
|
|
1278
|
+
"title": "MiniMax: Imagen a video",
|
|
1279
|
+
"description": "Generar videos refinados a partir de imágenes y texto con integración CGI usando MiniMax.",
|
|
1280
|
+
"mediaType": "image",
|
|
1281
|
+
"mediaSubtype": "webp",
|
|
1282
|
+
"tags": ["Imagen a video", "Video", "API"],
|
|
1283
|
+
"models": ["MiniMax"],
|
|
1806
1284
|
"date": "2025-03-01",
|
|
1807
1285
|
"tutorialUrl": ""
|
|
1808
1286
|
},
|
|
1809
1287
|
{
|
|
1810
1288
|
"name": "api_pixverse_i2v",
|
|
1811
|
-
"title": "PixVerse: Imagen a
|
|
1812
|
-
"description": "
|
|
1813
|
-
"mediaType": "image",
|
|
1814
|
-
"mediaSubtype": "webp",
|
|
1815
|
-
"tags": [
|
|
1816
|
-
|
|
1817
|
-
"Video",
|
|
1818
|
-
"API"
|
|
1819
|
-
],
|
|
1820
|
-
"models": [
|
|
1821
|
-
"PixVerse"
|
|
1822
|
-
],
|
|
1289
|
+
"title": "PixVerse: Imagen a video",
|
|
1290
|
+
"description": "Generar videos dinámicos a partir de imágenes estáticas con movimiento y efectos usando PixVerse.",
|
|
1291
|
+
"mediaType": "image",
|
|
1292
|
+
"mediaSubtype": "webp",
|
|
1293
|
+
"tags": ["Imagen a video", "Video", "API"],
|
|
1294
|
+
"models": ["PixVerse"],
|
|
1823
1295
|
"date": "2025-03-01",
|
|
1824
1296
|
"tutorialUrl": ""
|
|
1825
1297
|
},
|
|
1826
1298
|
{
|
|
1827
1299
|
"name": "api_pixverse_template_i2v",
|
|
1828
|
-
"title": "PixVerse Plantillas: Imagen a
|
|
1829
|
-
"description": "
|
|
1830
|
-
"mediaType": "image",
|
|
1831
|
-
"mediaSubtype": "webp",
|
|
1832
|
-
"tags": [
|
|
1833
|
-
|
|
1834
|
-
"Video",
|
|
1835
|
-
"API",
|
|
1836
|
-
"Templates"
|
|
1837
|
-
],
|
|
1838
|
-
"models": [
|
|
1839
|
-
"PixVerse"
|
|
1840
|
-
],
|
|
1300
|
+
"title": "PixVerse Plantillas: Imagen a video",
|
|
1301
|
+
"description": "Generar videos dinámicos a partir de imágenes estáticas con movimiento y efectos usando PixVerse.",
|
|
1302
|
+
"mediaType": "image",
|
|
1303
|
+
"mediaSubtype": "webp",
|
|
1304
|
+
"tags": ["Imagen a video", "Video", "API", "Plantillas"],
|
|
1305
|
+
"models": ["PixVerse"],
|
|
1841
1306
|
"date": "2025-03-01",
|
|
1842
1307
|
"tutorialUrl": ""
|
|
1843
1308
|
},
|
|
1844
1309
|
{
|
|
1845
1310
|
"name": "api_pixverse_t2v",
|
|
1846
|
-
"title": "PixVerse: Texto a
|
|
1847
|
-
"description": "
|
|
1848
|
-
"mediaType": "image",
|
|
1849
|
-
"mediaSubtype": "webp",
|
|
1850
|
-
"tags": [
|
|
1851
|
-
|
|
1852
|
-
"Video",
|
|
1853
|
-
"API"
|
|
1854
|
-
],
|
|
1855
|
-
"models": [
|
|
1856
|
-
"PixVerse"
|
|
1857
|
-
],
|
|
1311
|
+
"title": "PixVerse: Texto a video",
|
|
1312
|
+
"description": "Generar videos con interpretación precisa de indicaciones y dinámicas de video impresionantes.",
|
|
1313
|
+
"mediaType": "image",
|
|
1314
|
+
"mediaSubtype": "webp",
|
|
1315
|
+
"tags": ["Texto a video", "Video", "API"],
|
|
1316
|
+
"models": ["PixVerse"],
|
|
1858
1317
|
"date": "2025-03-01",
|
|
1859
1318
|
"tutorialUrl": ""
|
|
1860
1319
|
},
|
|
1861
1320
|
{
|
|
1862
1321
|
"name": "api_runway_gen3a_turbo_image_to_video",
|
|
1863
|
-
"title": "Runway: Gen3a Turbo Imagen a
|
|
1864
|
-
"description": "
|
|
1865
|
-
"mediaType": "image",
|
|
1866
|
-
"mediaSubtype": "webp",
|
|
1867
|
-
"tags": [
|
|
1868
|
-
|
|
1869
|
-
"Video",
|
|
1870
|
-
"API"
|
|
1871
|
-
],
|
|
1872
|
-
"models": [
|
|
1873
|
-
"Runway Gen3a Turbo"
|
|
1874
|
-
],
|
|
1322
|
+
"title": "Runway: Gen3a Turbo Imagen a video",
|
|
1323
|
+
"description": "Generar videos cinematográficos a partir de imágenes estáticas usando Runway Gen3a Turbo.",
|
|
1324
|
+
"mediaType": "image",
|
|
1325
|
+
"mediaSubtype": "webp",
|
|
1326
|
+
"tags": ["Imagen a video", "Video", "API"],
|
|
1327
|
+
"models": ["Runway Gen3a Turbo"],
|
|
1875
1328
|
"date": "2025-03-01",
|
|
1876
1329
|
"tutorialUrl": ""
|
|
1877
1330
|
},
|
|
1878
1331
|
{
|
|
1879
1332
|
"name": "api_runway_gen4_turo_image_to_video",
|
|
1880
|
-
"title": "Runway: Gen4 Turbo Imagen a
|
|
1881
|
-
"description": "
|
|
1882
|
-
"mediaType": "image",
|
|
1883
|
-
"mediaSubtype": "webp",
|
|
1884
|
-
"tags": [
|
|
1885
|
-
|
|
1886
|
-
"Video",
|
|
1887
|
-
"API"
|
|
1888
|
-
],
|
|
1889
|
-
"models": [
|
|
1890
|
-
"Runway Gen4 Turbo"
|
|
1891
|
-
],
|
|
1333
|
+
"title": "Runway: Gen4 Turbo Imagen a video",
|
|
1334
|
+
"description": "Generar videos dinámicos a partir de imágenes usando Runway Gen4 Turbo.",
|
|
1335
|
+
"mediaType": "image",
|
|
1336
|
+
"mediaSubtype": "webp",
|
|
1337
|
+
"tags": ["Imagen a video", "Video", "API"],
|
|
1338
|
+
"models": ["Runway Gen4 Turbo"],
|
|
1892
1339
|
"date": "2025-03-01",
|
|
1893
1340
|
"tutorialUrl": ""
|
|
1894
1341
|
},
|
|
1895
1342
|
{
|
|
1896
1343
|
"name": "api_runway_first_last_frame",
|
|
1897
|
-
"title": "Runway: Primer
|
|
1898
|
-
"description": "
|
|
1899
|
-
"mediaType": "image",
|
|
1900
|
-
"mediaSubtype": "webp",
|
|
1901
|
-
"tags": [
|
|
1902
|
-
|
|
1903
|
-
"Video",
|
|
1904
|
-
"API",
|
|
1905
|
-
"Frame Control"
|
|
1906
|
-
],
|
|
1907
|
-
"models": [
|
|
1908
|
-
"Runway"
|
|
1909
|
-
],
|
|
1344
|
+
"title": "Runway: Primer último fotograma a video",
|
|
1345
|
+
"description": "Generar transiciones de video suaves entre dos fotogramas clave con precisión de Runway.",
|
|
1346
|
+
"mediaType": "image",
|
|
1347
|
+
"mediaSubtype": "webp",
|
|
1348
|
+
"tags": ["Generación de video", "Video", "API", "Control de fotograma"],
|
|
1349
|
+
"models": ["Runway"],
|
|
1910
1350
|
"date": "2025-03-01",
|
|
1911
1351
|
"tutorialUrl": ""
|
|
1912
1352
|
},
|
|
1913
1353
|
{
|
|
1914
1354
|
"name": "api_pika_i2v",
|
|
1915
|
-
"title": "Pika: Imagen a
|
|
1916
|
-
"description": "
|
|
1917
|
-
"mediaType": "image",
|
|
1918
|
-
"mediaSubtype": "webp",
|
|
1919
|
-
"tags": [
|
|
1920
|
-
|
|
1921
|
-
"Video",
|
|
1922
|
-
"API"
|
|
1923
|
-
],
|
|
1924
|
-
"models": [
|
|
1925
|
-
"Pika"
|
|
1926
|
-
],
|
|
1355
|
+
"title": "Pika: Imagen a video",
|
|
1356
|
+
"description": "Generar videos animados suaves a partir de imágenes estáticas individuales usando Pika AI.",
|
|
1357
|
+
"mediaType": "image",
|
|
1358
|
+
"mediaSubtype": "webp",
|
|
1359
|
+
"tags": ["Imagen a video", "Video", "API"],
|
|
1360
|
+
"models": ["Pika"],
|
|
1927
1361
|
"date": "2025-03-01",
|
|
1928
1362
|
"tutorialUrl": ""
|
|
1929
1363
|
},
|
|
1930
1364
|
{
|
|
1931
1365
|
"name": "api_pika_scene",
|
|
1932
|
-
"title": "Pika
|
|
1933
|
-
"description": "
|
|
1934
|
-
"mediaType": "image",
|
|
1935
|
-
"mediaSubtype": "webp",
|
|
1936
|
-
"tags": [
|
|
1937
|
-
|
|
1938
|
-
"Video",
|
|
1939
|
-
"API",
|
|
1940
|
-
"Multi Image"
|
|
1941
|
-
],
|
|
1942
|
-
"models": [
|
|
1943
|
-
"Pika Scenes"
|
|
1944
|
-
],
|
|
1366
|
+
"title": "Pika Scenes: Imágenes a video",
|
|
1367
|
+
"description": "Generar videos que incorporen múltiples imágenes de entrada usando Pika Scenes.",
|
|
1368
|
+
"mediaType": "image",
|
|
1369
|
+
"mediaSubtype": "webp",
|
|
1370
|
+
"tags": ["Imagen a video", "Video", "API", "Imagen múltiple"],
|
|
1371
|
+
"models": ["Pika Scenes"],
|
|
1945
1372
|
"date": "2025-03-01",
|
|
1946
1373
|
"tutorialUrl": ""
|
|
1947
1374
|
},
|
|
1948
1375
|
{
|
|
1949
1376
|
"name": "api_veo2_i2v",
|
|
1950
|
-
"title": "Veo2: Imagen a
|
|
1951
|
-
"description": "
|
|
1952
|
-
"mediaType": "image",
|
|
1953
|
-
"mediaSubtype": "webp",
|
|
1954
|
-
"tags": [
|
|
1955
|
-
|
|
1956
|
-
|
|
1957
|
-
|
|
1958
|
-
|
|
1959
|
-
|
|
1960
|
-
|
|
1961
|
-
|
|
1377
|
+
"title": "Veo2: Imagen a video",
|
|
1378
|
+
"description": "Generar videos a partir de imágenes usando la API de Google Veo2.",
|
|
1379
|
+
"mediaType": "image",
|
|
1380
|
+
"mediaSubtype": "webp",
|
|
1381
|
+
"tags": ["Imagen a video", "Video", "API"],
|
|
1382
|
+
"models": ["Veo2"],
|
|
1383
|
+
"date": "2025-03-01",
|
|
1384
|
+
"tutorialUrl": ""
|
|
1385
|
+
},
|
|
1386
|
+
{
|
|
1387
|
+
"name": "api_veo3",
|
|
1388
|
+
"title": "Veo3: Imagen a video",
|
|
1389
|
+
"description": "Generar videos de 8 segundos de alta calidad a partir de indicaciones de texto o imágenes usando la API avanzada Veo 3 de Google. Cuenta con generación de audio, mejora de indicaciones y opciones duales de modelo para velocidad o calidad.",
|
|
1390
|
+
"mediaType": "image",
|
|
1391
|
+
"mediaSubtype": "webp",
|
|
1392
|
+
"tags": ["Imagen a video", "Texto a video", "API"],
|
|
1393
|
+
"models": ["Veo3"],
|
|
1962
1394
|
"date": "2025-03-01",
|
|
1963
1395
|
"tutorialUrl": ""
|
|
1964
1396
|
}
|
|
@@ -1967,95 +1399,65 @@
|
|
|
1967
1399
|
{
|
|
1968
1400
|
"moduleName": "default",
|
|
1969
1401
|
"category": "TOOLS & BUILDING",
|
|
1970
|
-
"title": "3D
|
|
1402
|
+
"title": "API 3D",
|
|
1971
1403
|
"type": "image",
|
|
1972
1404
|
"templates": [
|
|
1973
1405
|
{
|
|
1974
1406
|
"name": "api_rodin_image_to_model",
|
|
1975
|
-
"title": "Rodin: Imagen a
|
|
1976
|
-
"description": "
|
|
1407
|
+
"title": "Rodin: Imagen a modelo",
|
|
1408
|
+
"description": "Generar modelos 3D detallados a partir de fotos individuales usando Rodin AI.",
|
|
1977
1409
|
"mediaType": "image",
|
|
1978
1410
|
"thumbnailVariant": "compareSlider",
|
|
1979
1411
|
"mediaSubtype": "webp",
|
|
1980
|
-
"tags": [
|
|
1981
|
-
|
|
1982
|
-
"3D",
|
|
1983
|
-
"API"
|
|
1984
|
-
],
|
|
1985
|
-
"models": [
|
|
1986
|
-
"Rodin"
|
|
1987
|
-
],
|
|
1412
|
+
"tags": ["Imagen a modelo", "3D", "API"],
|
|
1413
|
+
"models": ["Rodin"],
|
|
1988
1414
|
"date": "2025-03-01",
|
|
1989
1415
|
"tutorialUrl": ""
|
|
1990
1416
|
},
|
|
1991
1417
|
{
|
|
1992
1418
|
"name": "api_rodin_multiview_to_model",
|
|
1993
|
-
"title": "Rodin:
|
|
1994
|
-
"description": "
|
|
1419
|
+
"title": "Rodin: Vista múltiple a modelo",
|
|
1420
|
+
"description": "Esculpir modelos 3D completos usando reconstrucción multiángulo de Rodin.",
|
|
1995
1421
|
"mediaType": "image",
|
|
1996
1422
|
"thumbnailVariant": "compareSlider",
|
|
1997
1423
|
"mediaSubtype": "webp",
|
|
1998
|
-
"tags": [
|
|
1999
|
-
|
|
2000
|
-
"3D",
|
|
2001
|
-
"API"
|
|
2002
|
-
],
|
|
2003
|
-
"models": [
|
|
2004
|
-
"Rodin"
|
|
2005
|
-
],
|
|
1424
|
+
"tags": ["Vista múltiple a modelo", "3D", "API"],
|
|
1425
|
+
"models": ["Rodin"],
|
|
2006
1426
|
"date": "2025-03-01",
|
|
2007
1427
|
"tutorialUrl": ""
|
|
2008
1428
|
},
|
|
2009
1429
|
{
|
|
2010
1430
|
"name": "api_tripo_text_to_model",
|
|
2011
|
-
"title": "Tripo: Texto a
|
|
2012
|
-
"description": "
|
|
2013
|
-
"mediaType": "image",
|
|
2014
|
-
"mediaSubtype": "webp",
|
|
2015
|
-
"tags": [
|
|
2016
|
-
|
|
2017
|
-
"3D",
|
|
2018
|
-
"API"
|
|
2019
|
-
],
|
|
2020
|
-
"models": [
|
|
2021
|
-
"Tripo"
|
|
2022
|
-
],
|
|
1431
|
+
"title": "Tripo: Texto a modelo",
|
|
1432
|
+
"description": "Crear objetos 3D a partir de descripciones con modelado dirigido por texto de Tripo.",
|
|
1433
|
+
"mediaType": "image",
|
|
1434
|
+
"mediaSubtype": "webp",
|
|
1435
|
+
"tags": ["Texto a modelo", "3D", "API"],
|
|
1436
|
+
"models": ["Tripo"],
|
|
2023
1437
|
"date": "2025-03-01",
|
|
2024
1438
|
"tutorialUrl": ""
|
|
2025
1439
|
},
|
|
2026
1440
|
{
|
|
2027
1441
|
"name": "api_tripo_image_to_model",
|
|
2028
|
-
"title": "Tripo: Imagen a
|
|
2029
|
-
"description": "
|
|
1442
|
+
"title": "Tripo: Imagen a modelo",
|
|
1443
|
+
"description": "Generar activos 3D profesionales a partir de imágenes 2D usando el motor Tripo.",
|
|
2030
1444
|
"mediaType": "image",
|
|
2031
1445
|
"thumbnailVariant": "compareSlider",
|
|
2032
1446
|
"mediaSubtype": "webp",
|
|
2033
|
-
"tags": [
|
|
2034
|
-
|
|
2035
|
-
"3D",
|
|
2036
|
-
"API"
|
|
2037
|
-
],
|
|
2038
|
-
"models": [
|
|
2039
|
-
"Tripo"
|
|
2040
|
-
],
|
|
1447
|
+
"tags": ["Imagen a modelo", "3D", "API"],
|
|
1448
|
+
"models": ["Tripo"],
|
|
2041
1449
|
"date": "2025-03-01",
|
|
2042
1450
|
"tutorialUrl": ""
|
|
2043
1451
|
},
|
|
2044
1452
|
{
|
|
2045
1453
|
"name": "api_tripo_multiview_to_model",
|
|
2046
|
-
"title": "Tripo:
|
|
2047
|
-
"description": "
|
|
1454
|
+
"title": "Tripo: Vista múltiple a modelo",
|
|
1455
|
+
"description": "Construir modelos 3D desde múltiples ángulos con el escáner avanzado de Tripo.",
|
|
2048
1456
|
"mediaType": "image",
|
|
2049
1457
|
"thumbnailVariant": "compareSlider",
|
|
2050
1458
|
"mediaSubtype": "webp",
|
|
2051
|
-
"tags": [
|
|
2052
|
-
|
|
2053
|
-
"3D",
|
|
2054
|
-
"API"
|
|
2055
|
-
],
|
|
2056
|
-
"models": [
|
|
2057
|
-
"Tripo"
|
|
2058
|
-
],
|
|
1459
|
+
"tags": ["Vista múltiple a modelo", "3D", "API"],
|
|
1460
|
+
"models": ["Tripo"],
|
|
2059
1461
|
"date": "2025-03-01",
|
|
2060
1462
|
"tutorialUrl": ""
|
|
2061
1463
|
}
|
|
@@ -2070,34 +1472,22 @@
|
|
|
2070
1472
|
{
|
|
2071
1473
|
"name": "api_openai_chat",
|
|
2072
1474
|
"title": "OpenAI: Chat",
|
|
2073
|
-
"description": "
|
|
1475
|
+
"description": "Interactuar con los modelos de lenguaje avanzados de OpenAI para conversaciones inteligentes.",
|
|
2074
1476
|
"mediaType": "image",
|
|
2075
1477
|
"mediaSubtype": "webp",
|
|
2076
|
-
"tags": [
|
|
2077
|
-
|
|
2078
|
-
"LLM",
|
|
2079
|
-
"API"
|
|
2080
|
-
],
|
|
2081
|
-
"models": [
|
|
2082
|
-
"OpenAI"
|
|
2083
|
-
],
|
|
1478
|
+
"tags": ["Chat", "LLM", "API"],
|
|
1479
|
+
"models": ["OpenAI"],
|
|
2084
1480
|
"date": "2025-03-01",
|
|
2085
1481
|
"tutorialUrl": ""
|
|
2086
1482
|
},
|
|
2087
1483
|
{
|
|
2088
1484
|
"name": "api_google_gemini",
|
|
2089
1485
|
"title": "Google Gemini: Chat",
|
|
2090
|
-
"description": "
|
|
1486
|
+
"description": "Experimentar la IA multimodal de Google con las capacidades de razonamiento de Gemini.",
|
|
2091
1487
|
"mediaType": "image",
|
|
2092
1488
|
"mediaSubtype": "webp",
|
|
2093
|
-
"tags": [
|
|
2094
|
-
|
|
2095
|
-
"LLM",
|
|
2096
|
-
"API"
|
|
2097
|
-
],
|
|
2098
|
-
"models": [
|
|
2099
|
-
"Google Gemini"
|
|
2100
|
-
],
|
|
1489
|
+
"tags": ["Chat", "LLM", "API"],
|
|
1490
|
+
"models": ["Google Gemini"],
|
|
2101
1491
|
"date": "2025-03-01",
|
|
2102
1492
|
"tutorialUrl": ""
|
|
2103
1493
|
}
|
|
@@ -2106,23 +1496,18 @@
|
|
|
2106
1496
|
{
|
|
2107
1497
|
"moduleName": "default",
|
|
2108
1498
|
"category": "TOOLS & BUILDING",
|
|
2109
|
-
"title": "
|
|
1499
|
+
"title": "Mejora",
|
|
2110
1500
|
"type": "image",
|
|
2111
1501
|
"templates": [
|
|
2112
1502
|
{
|
|
2113
1503
|
"name": "hiresfix_latent_workflow",
|
|
2114
|
-
"title": "
|
|
1504
|
+
"title": "Mejorar",
|
|
2115
1505
|
"mediaType": "image",
|
|
2116
1506
|
"mediaSubtype": "webp",
|
|
2117
|
-
"description": "
|
|
1507
|
+
"description": "Mejorar imágenes aumentando la calidad en el espacio latente.",
|
|
2118
1508
|
"thumbnailVariant": "compareSlider",
|
|
2119
|
-
"tags": [
|
|
2120
|
-
|
|
2121
|
-
"Image"
|
|
2122
|
-
],
|
|
2123
|
-
"models": [
|
|
2124
|
-
"SD1.5"
|
|
2125
|
-
],
|
|
1509
|
+
"tags": ["Mejorar", "Imagen"],
|
|
1510
|
+
"models": ["SD1.5"],
|
|
2126
1511
|
"date": "2025-03-01",
|
|
2127
1512
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/2_pass_txt2img/"
|
|
2128
1513
|
},
|
|
@@ -2131,49 +1516,34 @@
|
|
|
2131
1516
|
"title": "ESRGAN",
|
|
2132
1517
|
"mediaType": "image",
|
|
2133
1518
|
"mediaSubtype": "webp",
|
|
2134
|
-
"description": "
|
|
1519
|
+
"description": "Mejorar imágenes usando modelos ESRGAN para aumentar la calidad.",
|
|
2135
1520
|
"thumbnailVariant": "compareSlider",
|
|
2136
|
-
"tags": [
|
|
2137
|
-
|
|
2138
|
-
"Image"
|
|
2139
|
-
],
|
|
2140
|
-
"models": [
|
|
2141
|
-
"SD1.5"
|
|
2142
|
-
],
|
|
1521
|
+
"tags": ["Mejorar", "Imagen"],
|
|
1522
|
+
"models": ["SD1.5"],
|
|
2143
1523
|
"date": "2025-03-01",
|
|
2144
1524
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/upscale_models/"
|
|
2145
1525
|
},
|
|
2146
1526
|
{
|
|
2147
1527
|
"name": "hiresfix_esrgan_workflow",
|
|
2148
|
-
"title": "Flujo de
|
|
1528
|
+
"title": "Flujo de trabajo ESRGAN HiresFix",
|
|
2149
1529
|
"mediaType": "image",
|
|
2150
1530
|
"mediaSubtype": "webp",
|
|
2151
|
-
"description": "
|
|
1531
|
+
"description": "Mejorar imágenes usando modelos ESRGAN durante pasos intermedios de generación.",
|
|
2152
1532
|
"thumbnailVariant": "compareSlider",
|
|
2153
|
-
"tags": [
|
|
2154
|
-
|
|
2155
|
-
"Image"
|
|
2156
|
-
],
|
|
2157
|
-
"models": [
|
|
2158
|
-
"SD1.5"
|
|
2159
|
-
],
|
|
1533
|
+
"tags": ["Mejorar", "Imagen"],
|
|
1534
|
+
"models": ["SD1.5"],
|
|
2160
1535
|
"date": "2025-03-01",
|
|
2161
1536
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/2_pass_txt2img/#non-latent-upscaling"
|
|
2162
1537
|
},
|
|
2163
1538
|
{
|
|
2164
1539
|
"name": "latent_upscale_different_prompt_model",
|
|
2165
|
-
"title": "
|
|
1540
|
+
"title": "Mejora latente con modelo de indicación diferente",
|
|
2166
1541
|
"mediaType": "image",
|
|
2167
1542
|
"mediaSubtype": "webp",
|
|
2168
|
-
"description": "
|
|
1543
|
+
"description": "Mejorar imágenes mientras se cambian las indicaciones a través de pasos de generación.",
|
|
2169
1544
|
"thumbnailVariant": "zoomHover",
|
|
2170
|
-
"tags": [
|
|
2171
|
-
|
|
2172
|
-
"Image"
|
|
2173
|
-
],
|
|
2174
|
-
"models": [
|
|
2175
|
-
"SD1.5"
|
|
2176
|
-
],
|
|
1545
|
+
"tags": ["Mejorar", "Imagen"],
|
|
1546
|
+
"models": ["SD1.5"],
|
|
2177
1547
|
"date": "2025-03-01",
|
|
2178
1548
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/2_pass_txt2img/#more-examples"
|
|
2179
1549
|
}
|
|
@@ -2187,69 +1557,49 @@
|
|
|
2187
1557
|
"templates": [
|
|
2188
1558
|
{
|
|
2189
1559
|
"name": "controlnet_example",
|
|
2190
|
-
"title": "ControlNet
|
|
1560
|
+
"title": "ControlNet garabato",
|
|
2191
1561
|
"mediaType": "image",
|
|
2192
1562
|
"mediaSubtype": "webp",
|
|
2193
|
-
"description": "
|
|
1563
|
+
"description": "Generar imágenes guiadas por imágenes de referencia de garabatos usando ControlNet.",
|
|
2194
1564
|
"thumbnailVariant": "hoverDissolve",
|
|
2195
|
-
"tags": [
|
|
2196
|
-
|
|
2197
|
-
"Image"
|
|
2198
|
-
],
|
|
2199
|
-
"models": [
|
|
2200
|
-
"SD1.5"
|
|
2201
|
-
],
|
|
1565
|
+
"tags": ["ControlNet", "Imagen"],
|
|
1566
|
+
"models": ["SD1.5"],
|
|
2202
1567
|
"date": "2025-03-01",
|
|
2203
1568
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/controlnet/"
|
|
2204
1569
|
},
|
|
2205
1570
|
{
|
|
2206
1571
|
"name": "2_pass_pose_worship",
|
|
2207
|
-
"title": "ControlNet
|
|
1572
|
+
"title": "ControlNet pose 2 pasos",
|
|
2208
1573
|
"mediaType": "image",
|
|
2209
1574
|
"mediaSubtype": "webp",
|
|
2210
|
-
"description": "
|
|
1575
|
+
"description": "Generar imágenes guiadas por referencias de pose usando ControlNet.",
|
|
2211
1576
|
"thumbnailVariant": "hoverDissolve",
|
|
2212
|
-
"tags": [
|
|
2213
|
-
|
|
2214
|
-
"Image"
|
|
2215
|
-
],
|
|
2216
|
-
"models": [
|
|
2217
|
-
"SD1.5"
|
|
2218
|
-
],
|
|
1577
|
+
"tags": ["ControlNet", "Imagen"],
|
|
1578
|
+
"models": ["SD1.5"],
|
|
2219
1579
|
"date": "2025-03-01",
|
|
2220
1580
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/controlnet/#pose-controlnet"
|
|
2221
1581
|
},
|
|
2222
1582
|
{
|
|
2223
1583
|
"name": "depth_controlnet",
|
|
2224
|
-
"title": "ControlNet
|
|
1584
|
+
"title": "ControlNet profundidad",
|
|
2225
1585
|
"mediaType": "image",
|
|
2226
1586
|
"mediaSubtype": "webp",
|
|
2227
|
-
"description": "
|
|
1587
|
+
"description": "Generar imágenes guiadas por información de profundidad usando ControlNet.",
|
|
2228
1588
|
"thumbnailVariant": "hoverDissolve",
|
|
2229
|
-
"tags": [
|
|
2230
|
-
|
|
2231
|
-
"Image"
|
|
2232
|
-
],
|
|
2233
|
-
"models": [
|
|
2234
|
-
"SD1.5"
|
|
2235
|
-
],
|
|
1589
|
+
"tags": ["ControlNet", "Imagen"],
|
|
1590
|
+
"models": ["SD1.5"],
|
|
2236
1591
|
"date": "2025-03-01",
|
|
2237
1592
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/controlnet/#t2i-adapter-vs-controlnets"
|
|
2238
1593
|
},
|
|
2239
1594
|
{
|
|
2240
1595
|
"name": "depth_t2i_adapter",
|
|
2241
|
-
"title": "Adaptador
|
|
1596
|
+
"title": "Adaptador T2I profundidad",
|
|
2242
1597
|
"mediaType": "image",
|
|
2243
1598
|
"mediaSubtype": "webp",
|
|
2244
|
-
"description": "
|
|
1599
|
+
"description": "Generar imágenes guiadas por información de profundidad usando adaptador T2I.",
|
|
2245
1600
|
"thumbnailVariant": "hoverDissolve",
|
|
2246
|
-
"tags": [
|
|
2247
|
-
|
|
2248
|
-
"Image"
|
|
2249
|
-
],
|
|
2250
|
-
"models": [
|
|
2251
|
-
"SD1.5"
|
|
2252
|
-
],
|
|
1601
|
+
"tags": ["Adaptador T2I", "Imagen"],
|
|
1602
|
+
"models": ["SD1.5"],
|
|
2253
1603
|
"date": "2025-03-01",
|
|
2254
1604
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/controlnet/#t2i-adapter-vs-controlnets"
|
|
2255
1605
|
},
|
|
@@ -2258,15 +1608,10 @@
|
|
|
2258
1608
|
"title": "Mezcla de ControlNets",
|
|
2259
1609
|
"mediaType": "image",
|
|
2260
1610
|
"mediaSubtype": "webp",
|
|
2261
|
-
"description": "
|
|
1611
|
+
"description": "Generar imágenes combinando múltiples modelos ControlNet.",
|
|
2262
1612
|
"thumbnailVariant": "hoverDissolve",
|
|
2263
|
-
"tags": [
|
|
2264
|
-
|
|
2265
|
-
"Image"
|
|
2266
|
-
],
|
|
2267
|
-
"models": [
|
|
2268
|
-
"SD1.5"
|
|
2269
|
-
],
|
|
1613
|
+
"tags": ["ControlNet", "Imagen"],
|
|
1614
|
+
"models": ["SD1.5"],
|
|
2270
1615
|
"date": "2025-03-01",
|
|
2271
1616
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/controlnet/#mixing-controlnets"
|
|
2272
1617
|
}
|
|
@@ -2275,38 +1620,28 @@
|
|
|
2275
1620
|
{
|
|
2276
1621
|
"moduleName": "default",
|
|
2277
1622
|
"category": "TOOLS & BUILDING",
|
|
2278
|
-
"title": "Composición de
|
|
1623
|
+
"title": "Composición de área",
|
|
2279
1624
|
"type": "image",
|
|
2280
1625
|
"templates": [
|
|
2281
1626
|
{
|
|
2282
1627
|
"name": "area_composition",
|
|
2283
|
-
"title": "Composición de
|
|
1628
|
+
"title": "Composición de área",
|
|
2284
1629
|
"mediaType": "image",
|
|
2285
1630
|
"mediaSubtype": "webp",
|
|
2286
|
-
"description": "
|
|
2287
|
-
"tags": [
|
|
2288
|
-
|
|
2289
|
-
"Image"
|
|
2290
|
-
],
|
|
2291
|
-
"models": [
|
|
2292
|
-
"SD1.5"
|
|
2293
|
-
],
|
|
1631
|
+
"description": "Generar imágenes controlando la composición con áreas definidas.",
|
|
1632
|
+
"tags": ["Composición de área", "Imagen"],
|
|
1633
|
+
"models": ["SD1.5"],
|
|
2294
1634
|
"date": "2025-03-01",
|
|
2295
1635
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/area_composition/"
|
|
2296
1636
|
},
|
|
2297
1637
|
{
|
|
2298
1638
|
"name": "area_composition_square_area_for_subject",
|
|
2299
|
-
"title": "Composición de
|
|
1639
|
+
"title": "Composición de área área cuadrada para sujeto",
|
|
2300
1640
|
"mediaType": "image",
|
|
2301
1641
|
"mediaSubtype": "webp",
|
|
2302
|
-
"description": "
|
|
2303
|
-
"tags": [
|
|
2304
|
-
|
|
2305
|
-
"Image"
|
|
2306
|
-
],
|
|
2307
|
-
"models": [
|
|
2308
|
-
"SD1.5"
|
|
2309
|
-
],
|
|
1642
|
+
"description": "Generar imágenes con colocación consistente de sujeto usando composición de área.",
|
|
1643
|
+
"tags": ["Composición de área", "Imagen"],
|
|
1644
|
+
"models": ["SD1.5"],
|
|
2310
1645
|
"date": "2025-03-01",
|
|
2311
1646
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/area_composition/#increasing-consistency-of-images-with-area-composition"
|
|
2312
1647
|
}
|
|
@@ -2323,14 +1658,9 @@
|
|
|
2323
1658
|
"title": "Hunyuan3D 2.0",
|
|
2324
1659
|
"mediaType": "image",
|
|
2325
1660
|
"mediaSubtype": "webp",
|
|
2326
|
-
"description": "
|
|
2327
|
-
"tags": [
|
|
2328
|
-
|
|
2329
|
-
"3D"
|
|
2330
|
-
],
|
|
2331
|
-
"models": [
|
|
2332
|
-
"Hunyuan3D 2.0"
|
|
2333
|
-
],
|
|
1661
|
+
"description": "Generar modelos 3D a partir de imágenes individuales usando Hunyuan3D 2.0.",
|
|
1662
|
+
"tags": ["Imagen a modelo", "3D"],
|
|
1663
|
+
"models": ["Hunyuan3D 2.0"],
|
|
2334
1664
|
"date": "2025-03-01",
|
|
2335
1665
|
"tutorialUrl": ""
|
|
2336
1666
|
},
|
|
@@ -2339,14 +1669,9 @@
|
|
|
2339
1669
|
"title": "Hunyuan3D 2.0 MV",
|
|
2340
1670
|
"mediaType": "image",
|
|
2341
1671
|
"mediaSubtype": "webp",
|
|
2342
|
-
"description": "
|
|
2343
|
-
"tags": [
|
|
2344
|
-
|
|
2345
|
-
"3D"
|
|
2346
|
-
],
|
|
2347
|
-
"models": [
|
|
2348
|
-
"Hunyuan3D 2.0 MV"
|
|
2349
|
-
],
|
|
1672
|
+
"description": "Generar modelos 3D a partir de múltiples vistas usando Hunyuan3D 2.0 MV.",
|
|
1673
|
+
"tags": ["Vista múltiple a modelo", "3D"],
|
|
1674
|
+
"models": ["Hunyuan3D 2.0 MV"],
|
|
2350
1675
|
"date": "2025-03-01",
|
|
2351
1676
|
"tutorialUrl": "",
|
|
2352
1677
|
"thumbnailVariant": "hoverDissolve"
|
|
@@ -2356,31 +1681,21 @@
|
|
|
2356
1681
|
"title": "Hunyuan3D 2.0 MV Turbo",
|
|
2357
1682
|
"mediaType": "image",
|
|
2358
1683
|
"mediaSubtype": "webp",
|
|
2359
|
-
"description": "
|
|
2360
|
-
"tags": [
|
|
2361
|
-
|
|
2362
|
-
"3D"
|
|
2363
|
-
],
|
|
2364
|
-
"models": [
|
|
2365
|
-
"Hunyuan3D 2.0 MV Turbo"
|
|
2366
|
-
],
|
|
1684
|
+
"description": "Generar modelos 3D a partir de múltiples vistas usando Hunyuan3D 2.0 MV Turbo.",
|
|
1685
|
+
"tags": ["Vista múltiple a modelo", "3D"],
|
|
1686
|
+
"models": ["Hunyuan3D 2.0 MV Turbo"],
|
|
2367
1687
|
"date": "2025-03-01",
|
|
2368
1688
|
"tutorialUrl": "",
|
|
2369
1689
|
"thumbnailVariant": "hoverDissolve"
|
|
2370
1690
|
},
|
|
2371
1691
|
{
|
|
2372
1692
|
"name": "stable_zero123_example",
|
|
2373
|
-
"title": "
|
|
1693
|
+
"title": "Stable Zero123",
|
|
2374
1694
|
"mediaType": "image",
|
|
2375
1695
|
"mediaSubtype": "webp",
|
|
2376
|
-
"description": "
|
|
2377
|
-
"tags": [
|
|
2378
|
-
|
|
2379
|
-
"3D"
|
|
2380
|
-
],
|
|
2381
|
-
"models": [
|
|
2382
|
-
"Stable Zero123"
|
|
2383
|
-
],
|
|
1696
|
+
"description": "Generar vistas 3D a partir de imágenes individuales usando Stable Zero123.",
|
|
1697
|
+
"tags": ["Imagen a 3D", "3D"],
|
|
1698
|
+
"models": ["Stable Zero123"],
|
|
2384
1699
|
"date": "2025-03-01",
|
|
2385
1700
|
"tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/3d/"
|
|
2386
1701
|
}
|