comfyui-workflow-templates 0.1.96__py3-none-any.whl → 0.1.97__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of comfyui-workflow-templates might be problematic. Click here for more details.

Files changed (29) hide show
  1. comfyui_workflow_templates/templates/01_qwen_t2i_subgraphed-1.webp +0 -0
  2. comfyui_workflow_templates/templates/01_qwen_t2i_subgraphed.json +1288 -0
  3. comfyui_workflow_templates/templates/02_qwen_Image_edit_subgraphed-1.webp +0 -0
  4. comfyui_workflow_templates/templates/02_qwen_Image_edit_subgraphed.json +1754 -0
  5. comfyui_workflow_templates/templates/03_video_wan2_2_14B_i2v_subgraphed-1.webp +0 -0
  6. comfyui_workflow_templates/templates/03_video_wan2_2_14B_i2v_subgraphed.json +1416 -0
  7. comfyui_workflow_templates/templates/04_hunyuan_3d_2.1_subgraphed-1.webp +0 -0
  8. comfyui_workflow_templates/templates/04_hunyuan_3d_2.1_subgraphed.json +850 -0
  9. comfyui_workflow_templates/templates/05_audio_ace_step_1_t2a_song_subgraphed-1.webp +0 -0
  10. comfyui_workflow_templates/templates/05_audio_ace_step_1_t2a_song_subgraphed.json +1014 -0
  11. comfyui_workflow_templates/templates/api_rodin_gen2.json +123 -125
  12. comfyui_workflow_templates/templates/image_qwen_image_instantx_inpainting_controlnet.json +1536 -715
  13. comfyui_workflow_templates/templates/index.ar.json +2357 -0
  14. comfyui_workflow_templates/templates/index.es.json +180 -120
  15. comfyui_workflow_templates/templates/index.fr.json +346 -286
  16. comfyui_workflow_templates/templates/index.ja.json +227 -167
  17. comfyui_workflow_templates/templates/index.json +62 -2
  18. comfyui_workflow_templates/templates/index.ko.json +181 -121
  19. comfyui_workflow_templates/templates/index.ru.json +220 -160
  20. comfyui_workflow_templates/templates/index.tr.json +2357 -0
  21. comfyui_workflow_templates/templates/index.tr_translated.json +2357 -0
  22. comfyui_workflow_templates/templates/index.zh-TW.json +276 -216
  23. comfyui_workflow_templates/templates/index.zh.json +233 -173
  24. comfyui_workflow_templates/templates/video_wan_vace_inpainting.json +1 -1
  25. {comfyui_workflow_templates-0.1.96.dist-info → comfyui_workflow_templates-0.1.97.dist-info}/METADATA +1 -1
  26. {comfyui_workflow_templates-0.1.96.dist-info → comfyui_workflow_templates-0.1.97.dist-info}/RECORD +29 -16
  27. {comfyui_workflow_templates-0.1.96.dist-info → comfyui_workflow_templates-0.1.97.dist-info}/WHEEL +0 -0
  28. {comfyui_workflow_templates-0.1.96.dist-info → comfyui_workflow_templates-0.1.97.dist-info}/licenses/LICENSE +0 -0
  29. {comfyui_workflow_templates-0.1.96.dist-info → comfyui_workflow_templates-0.1.97.dist-info}/top_level.txt +0 -0
@@ -3,8 +3,68 @@
3
3
  "moduleName": "default",
4
4
  "type": "image",
5
5
  "isEssential": true,
6
- "title": "Conceptos básicos",
6
+ "title": "Getting Started",
7
7
  "templates": [
8
+ {
9
+ "name": "01_qwen_t2i_subgraphed",
10
+ "title": "Qwen-Image Text to Image",
11
+ "mediaType": "image",
12
+ "mediaSubtype": "webp",
13
+ "description": "Generate images with exceptional multilingual text rendering and editing capabilities using Qwen-Image's 20B MMDiT model..",
14
+ "tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image",
15
+ "tags": ["Texto a imagen", "Imagen"],
16
+ "models": ["Qwen-Image"],
17
+ "date": "2025-10-17",
18
+ "size": 29.59
19
+ },
20
+ {
21
+ "name": "02_qwen_Image_edit_subgraphed",
22
+ "title": "Qwen Image Edit 2509",
23
+ "mediaType": "image",
24
+ "mediaSubtype": "webp",
25
+ "description": "Advanced image editing with multi-image support, improved consistency, and ControlNet integration.",
26
+ "tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image-edit",
27
+ "tags": ["Imagen a imagen", "Edición de imagen", "ControlNet"],
28
+ "models": ["Qwen-Image"],
29
+ "date": "2025-10-17",
30
+ "size": 29.59
31
+ },
32
+ {
33
+ "name": "03_video_wan2_2_14B_i2v_subgraphed",
34
+ "title": "Wan 2.2 14B Image to Video",
35
+ "description": "Transform static images into dynamic videos with precise motion control and style preservation using Wan 2.2.",
36
+ "mediaType": "image",
37
+ "mediaSubtype": "webp",
38
+ "tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/wan2_2",
39
+ "tags": ["Imagen a video", "Video"],
40
+ "models": ["Wan2.2", "Wan"],
41
+ "date": "2025-10-17",
42
+ "size": 35.42
43
+ },
44
+ {
45
+ "name": "04_hunyuan_3d_2.1_subgraphed",
46
+ "title": "Hunyuan3D 2.1: image to 3D",
47
+ "mediaType": "image",
48
+ "mediaSubtype": "webp",
49
+ "description": "Generate 3D models from single images using Hunyuan3D 2.0.",
50
+ "tags": ["Imagen a 3D", "3D"],
51
+ "models": ["Hunyuan3D"],
52
+ "date": "2025-10-17",
53
+ "tutorialUrl": "https://docs.comfy.org/tutorials/3d/hunyuan3D-2",
54
+ "size": 4.59
55
+ },
56
+ {
57
+ "name": "05_audio_ace_step_1_t2a_song_subgraphed",
58
+ "title": "ACE Step v1 Text to Song",
59
+ "mediaType": "image",
60
+ "mediaSubtype": "webp",
61
+ "description": "Generate songs with vocals from text prompts using ACE-Step v1, supporting multilingual and style customization.",
62
+ "tags": ["Texto a audio", "Audio"],
63
+ "models": ["ACE-Step"],
64
+ "date": "2025-10-17",
65
+ "tutorialUrl": "https://docs.comfy.org/tutorials/audio/ace-step/ace-step-v1",
66
+ "size": 7.17
67
+ },
8
68
  {
9
69
  "name": "default",
10
70
  "title": "Generación de imágenes",
@@ -52,7 +112,7 @@
52
112
  "mediaSubtype": "webp",
53
113
  "description": "Generar imágenes combinando múltiples modelos LoRA.",
54
114
  "tutorialUrl": "https://docs.comfy.org/tutorials/basic/lora",
55
- "tags": ["Texto a imagen", "Imagen", "LoRA"],
115
+ "tags": ["Texto a imagen", "Imagen"],
56
116
  "models": ["SD1.5", "Stability"],
57
117
  "date": "2025-03-01",
58
118
  "size": 2.27,
@@ -60,7 +120,7 @@
60
120
  },
61
121
  {
62
122
  "name": "inpaint_example",
63
- "title": "Inpaint",
123
+ "title": "Inpainting",
64
124
  "mediaType": "image",
65
125
  "mediaSubtype": "webp",
66
126
  "description": "Editar partes específicas de imágenes sin problemas.",
@@ -74,7 +134,7 @@
74
134
  },
75
135
  {
76
136
  "name": "inpaint_model_outpainting",
77
- "title": "Outpaint",
137
+ "title": "Outpainting",
78
138
  "mediaType": "image",
79
139
  "mediaSubtype": "webp",
80
140
  "description": "Extender imágenes más allá de sus límites originales.",
@@ -93,7 +153,7 @@
93
153
  "mediaSubtype": "webp",
94
154
  "description": "Generar imágenes usando inversión textual para estilos consistentes.",
95
155
  "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/textual_inversion_embeddings/",
96
- "tags": ["Incrustación", "Imagen"],
156
+ "tags": ["Texto a imagen", "Imagen"],
97
157
  "models": ["SD1.5", "Stability"],
98
158
  "date": "2025-03-01",
99
159
  "size": 4.86,
@@ -106,7 +166,7 @@
106
166
  "mediaSubtype": "webp",
107
167
  "description": "Generar imágenes con colocación precisa de objetos usando cuadros de texto.",
108
168
  "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/gligen/",
109
- "tags": ["Gligen", "Imagen"],
169
+ "tags": ["Imagen"],
110
170
  "models": ["SD1.5", "Stability"],
111
171
  "date": "2025-03-01",
112
172
  "size": 2.77,
@@ -118,7 +178,7 @@
118
178
  "mediaType": "image",
119
179
  "mediaSubtype": "webp",
120
180
  "description": "Generar imágenes controlando la composición con áreas definidas.",
121
- "tags": ["Composición de área", "Imagen"],
181
+ "tags": ["Texto a imagen", "Imagen"],
122
182
  "models": ["SD1.5", "Stability"],
123
183
  "date": "2025-03-01",
124
184
  "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/area_composition/",
@@ -131,7 +191,7 @@
131
191
  "mediaType": "image",
132
192
  "mediaSubtype": "webp",
133
193
  "description": "Generar imágenes con colocación consistente de sujeto usando composición de área.",
134
- "tags": ["Composición de área", "Imagen"],
194
+ "tags": ["Texto a imagen", "Imagen"],
135
195
  "models": ["SD1.5", "Stability"],
136
196
  "date": "2025-03-01",
137
197
  "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/area_composition/#increasing-consistency-of-images-with-area-composition",
@@ -229,7 +289,7 @@
229
289
  "mediaSubtype": "webp",
230
290
  "description": "Generar imágenes guiadas por información de profundidad usando ControlNet.",
231
291
  "thumbnailVariant": "hoverDissolve",
232
- "tags": ["ControlNet", "Imagen"],
292
+ "tags": ["ControlNet", "Imagen", "Texto a imagen"],
233
293
  "models": ["SD1.5", "Stability"],
234
294
  "date": "2025-03-01",
235
295
  "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/controlnet/#t2i-adapter-vs-controlnets",
@@ -243,7 +303,7 @@
243
303
  "mediaSubtype": "webp",
244
304
  "description": "Generar imágenes guiadas por información de profundidad usando adaptador T2I.",
245
305
  "thumbnailVariant": "hoverDissolve",
246
- "tags": ["Adaptador T2I", "Imagen"],
306
+ "tags": ["ControlNet", "Imagen", "Texto a imagen"],
247
307
  "models": ["SD1.5", "Stability"],
248
308
  "date": "2025-03-01",
249
309
  "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/controlnet/#t2i-adapter-vs-controlnets",
@@ -257,7 +317,7 @@
257
317
  "mediaSubtype": "webp",
258
318
  "description": "Generar imágenes combinando múltiples modelos ControlNet.",
259
319
  "thumbnailVariant": "hoverDissolve",
260
- "tags": ["ControlNet", "Imagen"],
320
+ "tags": ["ControlNet", "Imagen", "Texto a imagen"],
261
321
  "models": ["SD1.5", "Stability"],
262
322
  "date": "2025-03-01",
263
323
  "tutorialUrl": "https://comfyanonymous.github.io/ComfyUI_examples/controlnet/#mixing-controlnets",
@@ -271,7 +331,7 @@
271
331
  "type": "image",
272
332
  "category": "GENERATION TYPE",
273
333
  "icon": "icon-[lucide--image]",
274
- "title": "Imagen",
334
+ "title": "Image",
275
335
  "templates": [
276
336
  {
277
337
  "name": "image_qwen_image",
@@ -290,8 +350,8 @@
290
350
  "title": "Qwen-Image InstantX ControlNet",
291
351
  "mediaType": "image",
292
352
  "mediaSubtype": "webp",
293
- "description": "Generate images with Qwen-Image InstantX ControlNet, supporting canny, soft edge, depth, pose",
294
- "tags": ["Image to Image", "Image", "ControlNet"],
353
+ "description": "Genera imágenes con Qwen-Image InstantX ControlNet, compatible con canny, bordes suaves, profundidad y pose",
354
+ "tags": ["Imagen a imagen", "Imagen", "ControlNet"],
295
355
  "tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image",
296
356
  "models": ["Qwen-Image"],
297
357
  "date": "2025-08-23",
@@ -299,12 +359,12 @@
299
359
  },
300
360
  {
301
361
  "name": "image_qwen_image_instantx_inpainting_controlnet",
302
- "title": "Qwen-Image InstantX Inpainting ControlNet",
362
+ "title": "Qwen-Image InstantX ControlNet de Inpainting",
303
363
  "mediaType": "image",
304
364
  "mediaSubtype": "webp",
305
365
  "thumbnailVariant": "compareSlider",
306
- "description": "Professional inpainting and image editing with Qwen-Image InstantX ControlNet. Supports object replacement, text modification, background changes, and outpainting.",
307
- "tags": ["Image to Image", "Image", "ControlNet", "Inpainting"],
366
+ "description": "Inpainting profesional y edición de imágenes con Qwen-Image InstantX ControlNet. Compatible con reemplazo de objetos, modificación de texto, cambios de fondo y outpainting.",
367
+ "tags": ["Imagen a imagen", "Imagen", "ControlNet", "Inpaint"],
308
368
  "tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image",
309
369
  "models": ["Qwen-Image"],
310
370
  "date": "2025-09-12",
@@ -324,26 +384,26 @@
324
384
  },
325
385
  {
326
386
  "name": "image_qwen_image_controlnet_patch",
327
- "title": "Qwen-Image ControlNet model patch",
387
+ "title": "Parche de modelo Qwen-Image ControlNet",
328
388
  "mediaType": "image",
329
389
  "mediaSubtype": "webp",
330
390
  "thumbnailVariant": "compareSlider",
331
- "description": "Control image generation using Qwen-Image ControlNet models. Supports canny, depth, and inpainting controls through model patching.",
391
+ "description": "Controla la generación de imágenes usando modelos Qwen-Image ControlNet. Compatible con controles canny, profundidad e inpainting mediante parcheo de modelo.",
332
392
  "tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image",
333
- "tags": ["Text to Image", "Image", "ControlNet"],
393
+ "tags": ["Texto a imagen", "Imagen", "ControlNet"],
334
394
  "models": ["Qwen-Image"],
335
395
  "date": "2025-08-24",
336
396
  "size": 31.7
337
397
  },
338
398
  {
339
399
  "name": "image_qwen_image_edit_2509",
340
- "title": "Qwen Image Edit 2509",
400
+ "title": "Qwen Edición de Imagen 2509",
341
401
  "mediaType": "image",
342
402
  "mediaSubtype": "webp",
343
403
  "thumbnailVariant": "compareSlider",
344
- "description": "Advanced image editing with multi-image support, improved consistency, and ControlNet integration.",
404
+ "description": "Edición avanzada de imágenes con soporte multi-imagen, consistencia mejorada e integración de ControlNet.",
345
405
  "tutorialUrl": "https://docs.comfy.org/tutorials/image/qwen/qwen-image-edit",
346
- "tags": ["Image to Image", "Image Edit", "Multi-Image", "ControlNet"],
406
+ "tags": ["Imagen a imagen", "Edición de imagen", "ControlNet"],
347
407
  "models": ["Qwen-Image"],
348
408
  "date": "2025-09-25",
349
409
  "size": 29.59
@@ -377,11 +437,11 @@
377
437
  },
378
438
  {
379
439
  "name": "image_chroma1_radiance_text_to_image",
380
- "title": "Chroma1 Radiance text to image",
440
+ "title": "Chroma1 Radiance Texto a Imagen",
381
441
  "mediaType": "image",
382
442
  "mediaSubtype": "webp",
383
- "description": "Chroma1-Radiance works directly with image pixels instead of compressed latents, delivering higher quality images with reduced artifacts and distortion.",
384
- "tags": ["Text to Image", "Image"],
443
+ "description": "Chroma1-Radiance trabaja directamente con píxeles de imagen en lugar de latentes comprimidos, ofreciendo imágenes de mayor calidad con menos artefactos y distorsión.",
444
+ "tags": ["Texto a imagen", "Imagen"],
385
445
  "models": ["Chroma"],
386
446
  "date": "2025-09-18",
387
447
  "size": 22.0,
@@ -389,11 +449,11 @@
389
449
  },
390
450
  {
391
451
  "name": "image_netayume_lumina_t2i",
392
- "title": "NetaYume Lumina Text to Image",
452
+ "title": "NetaYume Lumina Texto a Imagen",
393
453
  "mediaType": "image",
394
454
  "mediaSubtype": "webp",
395
- "description": "High-quality anime-style image generation with enhanced character understanding and detailed textures. Fine-tuned from Neta Lumina on Danbooru dataset.",
396
- "tags": ["Text to Image", "Image", "Anime"],
455
+ "description": "Generación de imágenes de estilo anime de alta calidad con comprensión mejorada de personajes y texturas detalladas. Ajustado finamente desde Neta Lumina en el conjunto de datos Danbooru.",
456
+ "tags": ["Texto a imagen", "Imagen", "Anime"],
397
457
  "models": ["NetaYume Lumina"],
398
458
  "date": "2025-10-10",
399
459
  "size": 9.89
@@ -417,7 +477,7 @@
417
477
  "mediaSubtype": "webp",
418
478
  "thumbnailVariant": "compareSlider",
419
479
  "description": "Supports various tasks such as image inpainting, outpainting, and object removal",
420
- "tags": ["Inpainting", "Outpainting"],
480
+ "tags": ["Inpaint", "Outpaint"],
421
481
  "models": ["Flux"],
422
482
  "date": "2025-09-21",
423
483
  "size": 27.01,
@@ -443,7 +503,7 @@
443
503
  "thumbnailVariant": "hoverDissolve",
444
504
  "mediaType": "image",
445
505
  "mediaSubtype": "webp",
446
- "tags": ["Image to Image", "Image"],
506
+ "tags": ["Imagen a imagen", "Imagen"],
447
507
  "models": ["Flux"],
448
508
  "date": "2025-09-02",
449
509
  "tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-uso",
@@ -470,7 +530,7 @@
470
530
  "mediaSubtype": "webp",
471
531
  "description": "Un modelo FLUX afinado que lleva el fotorrealismo al máximo",
472
532
  "tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux1-krea-dev",
473
- "tags": ["Texto a imagen", "Imagen", "Fotorrealismo"],
533
+ "tags": ["Texto a imagen", "Imagen"],
474
534
  "models": ["Flux"],
475
535
  "date": "2025-07-31",
476
536
  "size": 20.74,
@@ -548,7 +608,7 @@
548
608
  "description": "Generar imágenes guiadas por información de profundidad usando Flux LoRA.",
549
609
  "thumbnailVariant": "hoverDissolve",
550
610
  "tutorialUrl": "ttps://docs.comfy.org/tutorials/flux/flux-1-controlnet",
551
- "tags": ["Imagen a imagen", "ControlNet", "Imagen", "LoRA"],
611
+ "tags": ["Imagen a imagen", "ControlNet", "Imagen"],
552
612
  "models": ["Flux"],
553
613
  "date": "2025-03-01",
554
614
  "size": 32.98
@@ -560,7 +620,7 @@
560
620
  "mediaSubtype": "webp",
561
621
  "description": "Generar imágenes transfiriendo estilo de imágenes de referencia usando Flux Redux.",
562
622
  "tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-controlnet",
563
- "tags": ["Imagen a imagen", "ControlNet", "Imagen", "LoRA"],
623
+ "tags": ["Imagen a imagen", "ControlNet", "Imagen"],
564
624
  "models": ["Flux"],
565
625
  "date": "2025-03-01",
566
626
  "size": 32.74
@@ -758,7 +818,7 @@
758
818
  "mediaSubtype": "webp",
759
819
  "thumbnailVariant": "compareSlider",
760
820
  "description": "Ejecutar Profundidad Lotus en ComfyUI para estimación monocromática eficiente sin entrenamiento previo con alta retención de detalles.",
761
- "tags": ["Profundidad", "Imagen"],
821
+ "tags": ["Imagen", "Texto a imagen"],
762
822
  "models": ["SD1.5"],
763
823
  "date": "2025-05-21",
764
824
  "size": 1.93
@@ -812,20 +872,20 @@
812
872
  },
813
873
  {
814
874
  "name": "video_wan2_2_14B_animate",
815
- "title": "Wan2.2 Animate, character animation and replacement",
816
- "description": "Unified character animation and replacement framework with precise motion and expression replication.",
875
+ "title": "Wan2.2 Animate animación y reemplazo de personajes",
876
+ "description": "Marco unificado de animación y reemplazo de personajes con replicación precisa de movimiento y expresión。",
817
877
  "mediaType": "image",
818
878
  "mediaSubtype": "webp",
819
879
  "tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/wan2-2-animate",
820
- "tags": ["Video", "Image to Video"],
880
+ "tags": ["Video", "Imagen a video"],
821
881
  "models": ["Wan2.2", "Wan"],
822
882
  "date": "2025-09-22",
823
883
  "size": 25.535
824
884
  },
825
885
  {
826
886
  "name": "video_wan2_2_14B_s2v",
827
- "title": "Wan2.2-S2V Audio-Driven Video Generation",
828
- "description": "Transform static images and audio into dynamic videos with perfect synchronization and minute-level generation.",
887
+ "title": "Wan2.2-S2V Generación de Video Impulsada por Audio",
888
+ "description": "Transforma imágenes estáticas y audio en videos dinámicos con sincronización perfecta y generación de nivel por minuto.",
829
889
  "mediaType": "image",
830
890
  "mediaSubtype": "webp",
831
891
  "tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/wan2-2-s2v",
@@ -836,8 +896,8 @@
836
896
  },
837
897
  {
838
898
  "name": "video_humo",
839
- "title": "HuMo Video Generation",
840
- "description": "Generate videos basic on audio, image, and text, keep the character's lip sync.",
899
+ "title": "HuMo Generación de Video",
900
+ "description": "Genera videos basados en audio, imagen y texto, manteniendo la sincronización labial del personaje.",
841
901
  "mediaType": "image",
842
902
  "mediaSubtype": "webp",
843
903
  "tags": ["Video"],
@@ -847,8 +907,8 @@
847
907
  },
848
908
  {
849
909
  "name": "video_wan2_2_14B_fun_inpaint",
850
- "title": "Wan 2.2 14B Fun Inp",
851
- "description": "Generate videos from start and end frames using Wan 2.2 Fun Inp.",
910
+ "title": "Wan 2.2 14B Fun Inpainting",
911
+ "description": "Genera videos a partir de fotogramas de inicio y fin usando Wan 2.2 Fun Inp.",
852
912
  "mediaType": "image",
853
913
  "mediaSubtype": "webp",
854
914
  "tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/wan2-2-fun-inp",
@@ -895,11 +955,11 @@
895
955
  },
896
956
  {
897
957
  "name": "video_wan2_2_5B_fun_inpaint",
898
- "title": "Wan 2.2 5B Fun Inpaint",
899
- "description": "Efficient video inpainting from start and end frames. 5B model delivers quick iterations for testing workflows.",
958
+ "title": "Wan 2.2 5B Fun Inpainting",
959
+ "description": "Inpainting de video eficiente desde fotogramas de inicio y fin. El modelo 5B ofrece iteraciones rápidas para probar flujos de trabajo.",
900
960
  "mediaType": "image",
901
961
  "mediaSubtype": "webp",
902
- "tags": ["Text to Video", "Video"],
962
+ "tags": ["Texto a video", "Video"],
903
963
  "models": ["Wan2.2", "Wan"],
904
964
  "date": "2025-07-29",
905
965
  "size": 16.9
@@ -907,10 +967,10 @@
907
967
  {
908
968
  "name": "video_wan2_2_5B_fun_control",
909
969
  "title": "Wan 2.2 5B Fun Control",
910
- "description": "Multi-condition video control with pose, depth, and edge guidance. Compact 5B size for experimental development.",
970
+ "description": "Control de video multicondición con guía de pose, profundidad y bordes. Tamaño compacto de 5B para desarrollo experimental.",
911
971
  "mediaType": "image",
912
972
  "mediaSubtype": "webp",
913
- "tags": ["Text to Video", "Video"],
973
+ "tags": ["Texto a video", "Video"],
914
974
  "models": ["Wan2.2", "Wan"],
915
975
  "date": "2025-07-29",
916
976
  "size": 16.9
@@ -934,7 +994,7 @@
934
994
  "mediaType": "image",
935
995
  "mediaSubtype": "webp",
936
996
  "tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/vace",
937
- "tags": ["Referencia a video", "Video"],
997
+ "tags": ["Video", "Imagen a video"],
938
998
  "models": ["Wan2.1", "Wan"],
939
999
  "date": "2025-05-21",
940
1000
  "size": 53.79
@@ -960,7 +1020,7 @@
960
1020
  "mediaSubtype": "webp",
961
1021
  "thumbnailVariant": "compareSlider",
962
1022
  "tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/vace",
963
- "tags": ["Outpainting", "Video"],
1023
+ "tags": ["Outpaint", "Video"],
964
1024
  "models": ["Wan2.1", "Wan"],
965
1025
  "date": "2025-05-21",
966
1026
  "size": 53.79
@@ -985,18 +1045,18 @@
985
1045
  "mediaSubtype": "webp",
986
1046
  "thumbnailVariant": "compareSlider",
987
1047
  "tutorialUrl": "https://docs.comfy.org/tutorials/video/wan/vace",
988
- "tags": ["Inpainting", "Video"],
1048
+ "tags": ["Inpaint", "Video"],
989
1049
  "models": ["Wan2.1", "Wan"],
990
1050
  "date": "2025-05-21",
991
1051
  "size": 53.79
992
1052
  },
993
1053
  {
994
1054
  "name": "video_wan2.1_alpha_t2v_14B",
995
- "title": "Wan2.1 Alpha T2V",
996
- "description": "Generate text-to-video with alpha channel support for transparent backgrounds and semi-transparent objects.",
1055
+ "title": "Wan2.1 Alpha Texto a Video",
1056
+ "description": "Genera videos desde texto con soporte de canal alfa para fondos transparentes y objetos semitransparentes.",
997
1057
  "mediaType": "image",
998
1058
  "mediaSubtype": "webp",
999
- "tags": ["Text to Video", "Video"],
1059
+ "tags": ["Texto a video", "Video"],
1000
1060
  "models": ["Wan2.1", "Wan"],
1001
1061
  "date": "2025-10-06",
1002
1062
  "size": 20.95
@@ -1182,7 +1242,7 @@
1182
1242
  "templates": [
1183
1243
  {
1184
1244
  "name": "audio_stable_audio_example",
1185
- "title": "Stable Audio",
1245
+ "title": "Audio Estable",
1186
1246
  "mediaType": "audio",
1187
1247
  "mediaSubtype": "mp3",
1188
1248
  "description": "Generar audio a partir de indicaciones de texto usando Stable Audio.",
@@ -1198,7 +1258,7 @@
1198
1258
  "mediaType": "audio",
1199
1259
  "mediaSubtype": "mp3",
1200
1260
  "description": "Generar música instrumental a partir de indicaciones de texto usando ACE-Step v1.",
1201
- "tags": ["Texto a audio", "Audio", "Instrumentales"],
1261
+ "tags": ["Texto a audio", "Audio"],
1202
1262
  "models": ["ACE-Step"],
1203
1263
  "date": "2025-03-01",
1204
1264
  "tutorialUrl": "https://docs.comfy.org/tutorials/audio/ace-step/ace-step-v1",
@@ -1210,7 +1270,7 @@
1210
1270
  "mediaType": "audio",
1211
1271
  "mediaSubtype": "mp3",
1212
1272
  "description": "Generar canciones con voces a partir de indicaciones de texto usando ACE-Step v1, soportando personalización multilingüe y de estilo.",
1213
- "tags": ["Texto a audio", "Audio", "Canción"],
1273
+ "tags": ["Texto a audio", "Audio"],
1214
1274
  "models": ["ACE-Step"],
1215
1275
  "date": "2025-03-01",
1216
1276
  "tutorialUrl": "https://docs.comfy.org/tutorials/audio/ace-step/ace-step-v1",
@@ -1235,15 +1295,15 @@
1235
1295
  "type": "3d",
1236
1296
  "category": "GENERATION TYPE",
1237
1297
  "icon": "icon-[lucide--box]",
1238
- "title": "3D",
1298
+ "title": "3D Model",
1239
1299
  "templates": [
1240
1300
  {
1241
1301
  "name": "3d_hunyuan3d-v2.1",
1242
1302
  "title": "Hunyuan3D 2.1",
1243
1303
  "mediaType": "image",
1244
1304
  "mediaSubtype": "webp",
1245
- "description": "Generate 3D models from single images using Hunyuan3D 2.0.",
1246
- "tags": ["Image to Model", "3D"],
1305
+ "description": "Genera modelos 3D a partir de imágenes individuales usando Hunyuan3D 2.0.",
1306
+ "tags": ["Imagen a 3D", "3D"],
1247
1307
  "models": ["Hunyuan3D"],
1248
1308
  "date": "2025-03-01",
1249
1309
  "tutorialUrl": "",
@@ -1255,7 +1315,7 @@
1255
1315
  "mediaType": "image",
1256
1316
  "mediaSubtype": "webp",
1257
1317
  "description": "Generar modelos 3D a partir de imágenes individuales usando Hunyuan3D 2.0.",
1258
- "tags": ["Imagen a modelo", "3D"],
1318
+ "tags": ["Imagen a 3D", "3D"],
1259
1319
  "models": ["Hunyuan3D"],
1260
1320
  "date": "2025-03-01",
1261
1321
  "tutorialUrl": "",
@@ -1263,11 +1323,11 @@
1263
1323
  },
1264
1324
  {
1265
1325
  "name": "3d_hunyuan3d_multiview_to_model",
1266
- "title": "Hunyuan3D 2.0 MV",
1326
+ "title": "Hunyuan3D 2.0 Multivista",
1267
1327
  "mediaType": "image",
1268
1328
  "mediaSubtype": "webp",
1269
1329
  "description": "Generar modelos 3D a partir de múltiples vistas usando Hunyuan3D 2.0 MV.",
1270
- "tags": ["Vista múltiple a modelo", "3D"],
1330
+ "tags": ["3D", "Imagen a 3D"],
1271
1331
  "models": ["Hunyuan3D"],
1272
1332
  "date": "2025-03-01",
1273
1333
  "tutorialUrl": "",
@@ -1276,11 +1336,11 @@
1276
1336
  },
1277
1337
  {
1278
1338
  "name": "3d_hunyuan3d_multiview_to_model_turbo",
1279
- "title": "Hunyuan3D 2.0 MV Turbo",
1339
+ "title": "Hunyuan3D 2.0 Multivista Turbo",
1280
1340
  "mediaType": "image",
1281
1341
  "mediaSubtype": "webp",
1282
1342
  "description": "Generar modelos 3D a partir de múltiples vistas usando Hunyuan3D 2.0 MV Turbo.",
1283
- "tags": ["Vista múltiple a modelo", "3D"],
1343
+ "tags": ["Imagen a 3D", "3D"],
1284
1344
  "models": ["Hunyuan3D"],
1285
1345
  "date": "2025-03-01",
1286
1346
  "tutorialUrl": "",
@@ -1294,15 +1354,15 @@
1294
1354
  "type": "image",
1295
1355
  "category": "CLOSED SOURCE MODELS",
1296
1356
  "icon": "icon-[lucide--hand-coins]",
1297
- "title": "API de imagen",
1357
+ "title": "Image API",
1298
1358
  "templates": [
1299
1359
  {
1300
1360
  "name": "api_bytedance_seedream4",
1301
1361
  "title": "ByteDance Seedream 4.0",
1302
- "description": "Multi-modal AI model for text-to-image and image editing. Generate 2K images in under 2 seconds with natural language control.",
1362
+ "description": "Modelo de IA multimodal para texto a imagen y edición de imágenes. Genera imágenes 2K en menos de 2 segundos con control en lenguaje natural.",
1303
1363
  "mediaType": "image",
1304
1364
  "mediaSubtype": "webp",
1305
- "tags": ["Image Edit", "Image", "API", "Text-to-Image"],
1365
+ "tags": ["Edición de imagen", "Imagen", "API", "Texto a imagen"],
1306
1366
  "models": ["Seedream 4.0", "ByteDance"],
1307
1367
  "date": "2025-09-11",
1308
1368
  "OpenSource": false,
@@ -1311,11 +1371,11 @@
1311
1371
  },
1312
1372
  {
1313
1373
  "name": "api_google_gemini_image",
1314
- "title": "Google Gemini Image",
1315
- "description": "Nano-banana (Gemini-2.5-Flash Image) - image editing with consistency.",
1374
+ "title": "Google Gemini Imagen",
1375
+ "description": "Nano-banana (Gemini-2.5-Flash Image) - edición de imágenes con consistencia.",
1316
1376
  "mediaType": "image",
1317
1377
  "mediaSubtype": "webp",
1318
- "tags": ["Image Edit", "Image", "API", "Text-to-Image"],
1378
+ "tags": ["Edición de imagen", "Imagen", "API", "Texto a imagen"],
1319
1379
  "models": ["Gemini-2.5-Flash", "nano-banana", "Google"],
1320
1380
  "date": "2025-08-27",
1321
1381
  "OpenSource": false,
@@ -1369,11 +1429,11 @@
1369
1429
  },
1370
1430
  {
1371
1431
  "name": "api_wan_text_to_image",
1372
- "title": "Wan2.5: Text to Image",
1373
- "description": "Generate images with excellent prompt following and visual quality using FLUX.1 Pro.",
1432
+ "title": "Wan2.5: Texto a Imagen",
1433
+ "description": "Genera imágenes con excelente seguimiento de prompts y calidad visual usando FLUX.1 Pro.",
1374
1434
  "mediaType": "image",
1375
1435
  "mediaSubtype": "webp",
1376
- "tags": ["Text to Image", "Image", "API"],
1436
+ "tags": ["Texto a imagen", "Imagen", "API"],
1377
1437
  "models": ["Wan2.5"],
1378
1438
  "date": "2025-09-25",
1379
1439
  "OpenSource": false,
@@ -1415,7 +1475,7 @@
1415
1475
  "mediaType": "image",
1416
1476
  "mediaSubtype": "webp",
1417
1477
  "thumbnailVariant": "compareSlider",
1418
- "tags": ["Texto a imagen", "Imagen", "API", "Transferencia de estilo"],
1478
+ "tags": ["Texto a imagen", "Imagen", "API"],
1419
1479
  "models": ["Luma"],
1420
1480
  "date": "2025-03-01",
1421
1481
  "OpenSource": false,
@@ -1428,7 +1488,7 @@
1428
1488
  "description": "Generar imágenes con paletas de colores personalizadas y visuales específicos de marca usando Recraft.",
1429
1489
  "mediaType": "image",
1430
1490
  "mediaSubtype": "webp",
1431
- "tags": ["Texto a imagen", "Imagen", "API", "Control de color"],
1491
+ "tags": ["Texto a imagen", "Imagen", "API"],
1432
1492
  "models": ["Recraft"],
1433
1493
  "date": "2025-03-01",
1434
1494
  "OpenSource": false,
@@ -1441,7 +1501,7 @@
1441
1501
  "description": "Controlar estilo con ejemplos visuales, alinear posicionamiento y ajustar objetos finamente. Almacenar y compartir estilos para consistencia perfecta de marca.",
1442
1502
  "mediaType": "image",
1443
1503
  "mediaSubtype": "webp",
1444
- "tags": ["Texto a imagen", "Imagen", "API", "Control de estilo"],
1504
+ "tags": ["Texto a imagen", "Imagen", "API"],
1445
1505
  "models": ["Recraft"],
1446
1506
  "date": "2025-03-01",
1447
1507
  "OpenSource": false,
@@ -1481,7 +1541,7 @@
1481
1541
  "mediaType": "image",
1482
1542
  "thumbnailVariant": "compareSlider",
1483
1543
  "mediaSubtype": "webp",
1484
- "tags": ["Imagen a imagen", "Imagen", "API", "Transferencia de estilo"],
1544
+ "tags": ["Imagen a imagen", "Imagen", "API"],
1485
1545
  "models": ["Runway"],
1486
1546
  "date": "2025-03-01",
1487
1547
  "OpenSource": false,
@@ -1548,7 +1608,7 @@
1548
1608
  "description": "Generar imágenes de calidad profesional con excelente alineación de indicaciones, fotorrealismo y renderizado de texto usando Ideogram V3.",
1549
1609
  "mediaType": "image",
1550
1610
  "mediaSubtype": "webp",
1551
- "tags": ["Texto a imagen", "Imagen", "API", "Renderizado de texto"],
1611
+ "tags": ["Texto a imagen", "Imagen", "API"],
1552
1612
  "models": ["Ideogram"],
1553
1613
  "date": "2025-03-01",
1554
1614
  "OpenSource": false,
@@ -1606,7 +1666,7 @@
1606
1666
  "mediaType": "image",
1607
1667
  "mediaSubtype": "webp",
1608
1668
  "thumbnailVariant": "compareSlider",
1609
- "tags": ["Texto a imagen", "Imagen", "API", "Entrada múltiple"],
1669
+ "tags": ["Texto a imagen", "Imagen", "API"],
1610
1670
  "models": ["GPT-Image-1"],
1611
1671
  "date": "2025-03-01",
1612
1672
  "tutorialUrl": "https://docs.comfy.org/tutorials/api-nodes/openai/gpt-image-1",
@@ -1664,15 +1724,15 @@
1664
1724
  "type": "video",
1665
1725
  "category": "CLOSED SOURCE MODELS",
1666
1726
  "icon": "icon-[lucide--film]",
1667
- "title": "API de video",
1727
+ "title": "Video API",
1668
1728
  "templates": [
1669
1729
  {
1670
1730
  "name": "api_openai_sora_video",
1671
- "title": "Sora 2: Text & Image to Video",
1672
- "description": "OpenAI's Sora-2 and Sora-2 Pro video generation with synchronized audio.",
1731
+ "title": "Sora 2: Texto e Imagen a Video",
1732
+ "description": "Generación de video Sora-2 y Sora-2 Pro de OpenAI con audio sincronizado.",
1673
1733
  "mediaType": "image",
1674
1734
  "mediaSubtype": "webp",
1675
- "tags": ["Image to Video", "Text to Video", "API"],
1735
+ "tags": ["Imagen a video", "Texto a video", "API"],
1676
1736
  "models": ["OpenAI"],
1677
1737
  "date": "2025-10-08",
1678
1738
  "OpenSource": false,
@@ -1681,11 +1741,11 @@
1681
1741
  },
1682
1742
  {
1683
1743
  "name": "api_wan_text_to_video",
1684
- "title": "Wan2.5: Text to Video",
1685
- "description": "Generate videos with synchronized audio, enhanced motion, and superior quality.",
1744
+ "title": "Wan2.5: Texto a Video",
1745
+ "description": "Genera videos con audio sincronizado, movimiento mejorado y calidad superior.",
1686
1746
  "mediaType": "image",
1687
1747
  "mediaSubtype": "webp",
1688
- "tags": ["Image to Video", "Video", "API"],
1748
+ "tags": ["Imagen a video", "Video", "API"],
1689
1749
  "models": ["Wan2.5"],
1690
1750
  "date": "2025-09-27",
1691
1751
  "tutorialUrl": "",
@@ -1695,11 +1755,11 @@
1695
1755
  },
1696
1756
  {
1697
1757
  "name": "api_wan_image_to_video",
1698
- "title": "Wan2.5: Image to Video",
1699
- "description": "Transform images into videos with synchronized audio, enhanced motion, and superior quality.",
1758
+ "title": "Wan2.5: Imagen a Video",
1759
+ "description": "Transforma imágenes en videos con audio sincronizado, movimiento mejorado y calidad superior.",
1700
1760
  "mediaType": "image",
1701
1761
  "mediaSubtype": "webp",
1702
- "tags": ["Image to Video", "Video", "API"],
1762
+ "tags": ["Imagen a video", "Video", "API"],
1703
1763
  "models": ["Wan2.5"],
1704
1764
  "date": "2025-09-27",
1705
1765
  "tutorialUrl": "",
@@ -1727,7 +1787,7 @@
1727
1787
  "description": "Generar videos dinámicos aplicando efectos visuales a imágenes usando Kling.",
1728
1788
  "mediaType": "image",
1729
1789
  "mediaSubtype": "webp",
1730
- "tags": ["Efectos de video", "Video", "API"],
1790
+ "tags": ["Video", "API"],
1731
1791
  "models": ["Kling"],
1732
1792
  "date": "2025-03-01",
1733
1793
  "tutorialUrl": "",
@@ -1741,7 +1801,7 @@
1741
1801
  "description": "Generar videos controlando el primer y último fotograma.",
1742
1802
  "mediaType": "image",
1743
1803
  "mediaSubtype": "webp",
1744
- "tags": ["Generación de video", "Video", "API", "Control de fotograma"],
1804
+ "tags": ["Video", "API", "FLF2V"],
1745
1805
  "models": ["Kling"],
1746
1806
  "date": "2025-03-01",
1747
1807
  "tutorialUrl": "",
@@ -1783,7 +1843,7 @@
1783
1843
  "description": "Generar videos con sujetos consistentes usando múltiples imágenes de referencia (hasta 7) para continuidad de personaje y estilo a lo largo de la secuencia de video.",
1784
1844
  "mediaType": "image",
1785
1845
  "mediaSubtype": "webp",
1786
- "tags": ["Referencia a video", "Video", "API"],
1846
+ "tags": ["Video", "Imagen a video", "API"],
1787
1847
  "models": ["Vidu"],
1788
1848
  "date": "2025-08-23",
1789
1849
  "tutorialUrl": "",
@@ -1797,7 +1857,7 @@
1797
1857
  "description": "Crear transiciones de video suaves entre fotogramas de inicio y fin definidos con interpolación natural de movimiento y calidad visual consistente.",
1798
1858
  "mediaType": "image",
1799
1859
  "mediaSubtype": "webp",
1800
- "tags": ["FLF2V", "Video", "API"],
1860
+ "tags": ["Video", "API", "FLF2V"],
1801
1861
  "models": ["Vidu"],
1802
1862
  "date": "2025-08-23",
1803
1863
  "tutorialUrl": "",
@@ -1807,11 +1867,11 @@
1807
1867
  },
1808
1868
  {
1809
1869
  "name": "api_bytedance_text_to_video",
1810
- "title": "ByteDance: Text to Video",
1811
- "description": "Generate high-quality videos directly from text prompts using ByteDance's Seedance model. Supports multiple resolutions and aspect ratios with natural motion and cinematic quality.",
1870
+ "title": "ByteDance: Texto a Video",
1871
+ "description": "Genera videos de alta calidad directamente desde prompts de texto usando el modelo Seedance de ByteDance. Compatible con múltiples resoluciones y relaciones de aspecto con movimiento natural y calidad cinematográfica.",
1812
1872
  "mediaType": "image",
1813
1873
  "mediaSubtype": "webp",
1814
- "tags": ["Video", "API", "Text to Video"],
1874
+ "tags": ["Video", "API", "Texto a video"],
1815
1875
  "models": ["ByteDance"],
1816
1876
  "date": "2025-10-6",
1817
1877
  "tutorialUrl": "",
@@ -1821,11 +1881,11 @@
1821
1881
  },
1822
1882
  {
1823
1883
  "name": "api_bytedance_image_to_video",
1824
- "title": "ByteDance: Image to Video",
1825
- "description": "Transform static images into dynamic videos using ByteDance's Seedance model. Analyzes image structure and generates natural motion with consistent visual style and coherent video sequences.",
1884
+ "title": "ByteDance: Imagen a Video",
1885
+ "description": "Transforma imágenes estáticas en videos dinámicos usando el modelo Seedance de ByteDance. Analiza la estructura de la imagen y genera movimiento natural con estilo visual consistente y secuencias de video coherentes.",
1826
1886
  "mediaType": "image",
1827
1887
  "mediaSubtype": "webp",
1828
- "tags": ["Video", "API", "Image to Video"],
1888
+ "tags": ["Video", "API", "Imagen a video"],
1829
1889
  "models": ["ByteDance"],
1830
1890
  "date": "2025-10-6",
1831
1891
  "tutorialUrl": "",
@@ -1835,8 +1895,8 @@
1835
1895
  },
1836
1896
  {
1837
1897
  "name": "api_bytedance_flf2v",
1838
- "title": "ByteDance: Start End to Video",
1839
- "description": "Generate cinematic video transitions between start and end frames with fluid motion, scene consistency, and professional polish using ByteDance's Seedance model.",
1898
+ "title": "ByteDance: Inicio-Fin a Video",
1899
+ "description": "Genera transiciones de video cinematográficas entre fotogramas de inicio y fin con movimiento fluido, consistencia de escena y acabado profesional usando el modelo Seedance de ByteDance.",
1840
1900
  "mediaType": "image",
1841
1901
  "mediaSubtype": "webp",
1842
1902
  "tags": ["Video", "API", "FLF2V"],
@@ -1910,7 +1970,7 @@
1910
1970
  "mediaType": "image",
1911
1971
  "thumbnailVariant": "hoverDissolve",
1912
1972
  "mediaSubtype": "webp",
1913
- "tags": ["Video a video", "Video", "API", "Transferencia de movimiento"],
1973
+ "tags": ["Video a video", "Video", "API"],
1914
1974
  "models": ["Moonvalley"],
1915
1975
  "date": "2025-03-01",
1916
1976
  "tutorialUrl": "",
@@ -1925,7 +1985,7 @@
1925
1985
  "mediaType": "image",
1926
1986
  "thumbnailVariant": "hoverDissolve",
1927
1987
  "mediaSubtype": "webp",
1928
- "tags": ["Video a video", "Video", "API", "Control de pose"],
1988
+ "tags": ["Video a video", "Video", "API"],
1929
1989
  "models": ["Moonvalley"],
1930
1990
  "date": "2025-03-01",
1931
1991
  "tutorialUrl": "",
@@ -1995,7 +2055,7 @@
1995
2055
  "description": "Generar videos dinámicos a partir de imágenes estáticas con movimiento y efectos usando PixVerse.",
1996
2056
  "mediaType": "image",
1997
2057
  "mediaSubtype": "webp",
1998
- "tags": ["Imagen a video", "Video", "API", "Plantillas"],
2058
+ "tags": ["Imagen a video", "Video", "API"],
1999
2059
  "models": ["PixVerse"],
2000
2060
  "date": "2025-03-01",
2001
2061
  "tutorialUrl": "",
@@ -2051,7 +2111,7 @@
2051
2111
  "description": "Generar transiciones de video suaves entre dos fotogramas clave con precisión de Runway.",
2052
2112
  "mediaType": "image",
2053
2113
  "mediaSubtype": "webp",
2054
- "tags": ["Generación de video", "Video", "API", "Control de fotograma"],
2114
+ "tags": ["Video", "API", "FLF2V"],
2055
2115
  "models": ["Runway"],
2056
2116
  "date": "2025-03-01",
2057
2117
  "tutorialUrl": "",
@@ -2079,7 +2139,7 @@
2079
2139
  "description": "Generar videos que incorporen múltiples imágenes de entrada usando Pika Scenes.",
2080
2140
  "mediaType": "image",
2081
2141
  "mediaSubtype": "webp",
2082
- "tags": ["Imagen a video", "Video", "API", "Imagen múltiple"],
2142
+ "tags": ["Imagen a video", "Video", "API"],
2083
2143
  "models": ["Pika"],
2084
2144
  "date": "2025-03-01",
2085
2145
  "tutorialUrl": "",
@@ -2122,15 +2182,15 @@
2122
2182
  "type": "image",
2123
2183
  "category": "CLOSED SOURCE MODELS",
2124
2184
  "icon": "icon-[lucide--box]",
2125
- "title": "API 3D",
2185
+ "title": "3D API",
2126
2186
  "templates": [
2127
2187
  {
2128
2188
  "name": "api_rodin_gen2",
2129
- "title": "Rodin: Gen-2 Image to Model",
2130
- "description": "Generate detailed 4X mesh quality 3D models from photos using Rodin Gen2",
2189
+ "title": "Rodin: Gen-2 Imagen a Modelo",
2190
+ "description": "Genera modelos 3D detallados con 4X de calidad de malla a partir de fotos usando Rodin Gen2",
2131
2191
  "mediaType": "image",
2132
2192
  "mediaSubtype": "webp",
2133
- "tags": ["Image to Model", "3D", "API"],
2193
+ "tags": ["Imagen a 3D", "3D", "API"],
2134
2194
  "models": ["Rodin"],
2135
2195
  "date": "2025-09-27",
2136
2196
  "tutorialUrl": "",
@@ -2145,7 +2205,7 @@
2145
2205
  "mediaType": "image",
2146
2206
  "thumbnailVariant": "compareSlider",
2147
2207
  "mediaSubtype": "webp",
2148
- "tags": ["Imagen a modelo", "3D", "API"],
2208
+ "tags": ["Imagen a 3D", "3D", "API"],
2149
2209
  "models": ["Rodin"],
2150
2210
  "date": "2025-03-01",
2151
2211
  "tutorialUrl": "",
@@ -2160,7 +2220,7 @@
2160
2220
  "mediaType": "image",
2161
2221
  "thumbnailVariant": "compareSlider",
2162
2222
  "mediaSubtype": "webp",
2163
- "tags": ["Vista múltiple a modelo", "3D", "API"],
2223
+ "tags": ["Imagen a 3D", "3D", "API"],
2164
2224
  "models": ["Rodin"],
2165
2225
  "date": "2025-03-01",
2166
2226
  "tutorialUrl": "",
@@ -2189,7 +2249,7 @@
2189
2249
  "mediaType": "image",
2190
2250
  "thumbnailVariant": "compareSlider",
2191
2251
  "mediaSubtype": "webp",
2192
- "tags": ["Imagen a modelo", "3D", "API"],
2252
+ "tags": ["Imagen a 3D", "3D", "API"],
2193
2253
  "models": ["Tripo"],
2194
2254
  "date": "2025-03-01",
2195
2255
  "tutorialUrl": "",
@@ -2204,7 +2264,7 @@
2204
2264
  "mediaType": "image",
2205
2265
  "thumbnailVariant": "compareSlider",
2206
2266
  "mediaSubtype": "webp",
2207
- "tags": ["Vista múltiple a modelo", "3D", "API"],
2267
+ "tags": ["Imagen a 3D", "3D", "API"],
2208
2268
  "models": ["Tripo"],
2209
2269
  "date": "2025-03-01",
2210
2270
  "tutorialUrl": "",
@@ -2219,7 +2279,7 @@
2219
2279
  "type": "audio",
2220
2280
  "category": "CLOSED SOURCE MODELS",
2221
2281
  "icon": "icon-[lucide--volume-2]",
2222
- "title": "API de Audio",
2282
+ "title": "Audio API",
2223
2283
  "templates": [
2224
2284
  {
2225
2285
  "name": "api_stability_ai_text_to_audio",
@@ -2267,7 +2327,7 @@
2267
2327
  "type": "image",
2268
2328
  "category": "CLOSED SOURCE MODELS",
2269
2329
  "icon": "icon-[lucide--message-square-text]",
2270
- "title": "API LLM",
2330
+ "title": "LLM API",
2271
2331
  "templates": [
2272
2332
  {
2273
2333
  "name": "api_openai_chat",
@@ -2275,7 +2335,7 @@
2275
2335
  "description": "Interactuar con los modelos de lenguaje avanzados de OpenAI para conversaciones inteligentes.",
2276
2336
  "mediaType": "image",
2277
2337
  "mediaSubtype": "webp",
2278
- "tags": ["Chat", "LLM", "API"],
2338
+ "tags": ["LLM", "API"],
2279
2339
  "models": ["OpenAI"],
2280
2340
  "date": "2025-03-01",
2281
2341
  "tutorialUrl": "",
@@ -2289,7 +2349,7 @@
2289
2349
  "description": "Experimentar la IA multimodal de Google con las capacidades de razonamiento de Gemini.",
2290
2350
  "mediaType": "image",
2291
2351
  "mediaSubtype": "webp",
2292
- "tags": ["Chat", "LLM", "API"],
2352
+ "tags": ["LLM", "API"],
2293
2353
  "models": ["Google Gemini", "Google"],
2294
2354
  "date": "2025-03-01",
2295
2355
  "tutorialUrl": "",