comfyui-workflow-templates 0.2.1__py3-none-any.whl → 0.2.3__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of comfyui-workflow-templates might be problematic. Click here for more details.
- comfyui_workflow_templates/templates/02_qwen_Image_edit_subgraphed.json +189 -188
- comfyui_workflow_templates/templates/04_hunyuan_3d_2.1_subgraphed.json +48 -75
- comfyui_workflow_templates/templates/flux_kontext_dev_basic.json +39 -39
- comfyui_workflow_templates/templates/index.es.json +5 -5
- comfyui_workflow_templates/templates/index.fr.json +5 -5
- comfyui_workflow_templates/templates/index.ja.json +5 -5
- comfyui_workflow_templates/templates/index.json +6 -6
- comfyui_workflow_templates/templates/index.ko.json +5 -5
- comfyui_workflow_templates/templates/index.ru.json +5 -5
- comfyui_workflow_templates/templates/index.tr.json +5 -5
- comfyui_workflow_templates/templates/index.zh-TW.json +5 -5
- comfyui_workflow_templates/templates/index.zh.json +5 -5
- {comfyui_workflow_templates-0.2.1.dist-info → comfyui_workflow_templates-0.2.3.dist-info}/METADATA +1 -1
- {comfyui_workflow_templates-0.2.1.dist-info → comfyui_workflow_templates-0.2.3.dist-info}/RECORD +17 -17
- {comfyui_workflow_templates-0.2.1.dist-info → comfyui_workflow_templates-0.2.3.dist-info}/WHEEL +0 -0
- {comfyui_workflow_templates-0.2.1.dist-info → comfyui_workflow_templates-0.2.3.dist-info}/licenses/LICENSE +0 -0
- {comfyui_workflow_templates-0.2.1.dist-info → comfyui_workflow_templates-0.2.3.dist-info}/top_level.txt +0 -0
|
@@ -27,9 +27,9 @@
|
|
|
27
27
|
],
|
|
28
28
|
"outputs": [],
|
|
29
29
|
"properties": {
|
|
30
|
-
"Node name for S&R": "SaveGLB",
|
|
31
30
|
"cnr_id": "comfy-core",
|
|
32
31
|
"ver": "0.3.59",
|
|
32
|
+
"Node name for S&R": "SaveGLB",
|
|
33
33
|
"Camera Info": {
|
|
34
34
|
"position": {
|
|
35
35
|
"x": 9.380174569468211,
|
|
@@ -104,15 +104,39 @@
|
|
|
104
104
|
}
|
|
105
105
|
],
|
|
106
106
|
"properties": {
|
|
107
|
-
"Node name for S&R": "LoadImage",
|
|
108
107
|
"cnr_id": "comfy-core",
|
|
109
|
-
"ver": "0.3.59"
|
|
108
|
+
"ver": "0.3.59",
|
|
109
|
+
"Node name for S&R": "LoadImage"
|
|
110
110
|
},
|
|
111
111
|
"widgets_values": [
|
|
112
112
|
"04_hunyuan_3d_2.1_subgraphed_input_image.png",
|
|
113
113
|
"image"
|
|
114
114
|
]
|
|
115
115
|
},
|
|
116
|
+
{
|
|
117
|
+
"id": 20,
|
|
118
|
+
"type": "MarkdownNote",
|
|
119
|
+
"pos": [
|
|
120
|
+
-360,
|
|
121
|
+
-410
|
|
122
|
+
],
|
|
123
|
+
"size": [
|
|
124
|
+
440,
|
|
125
|
+
270
|
|
126
|
+
],
|
|
127
|
+
"flags": {},
|
|
128
|
+
"order": 2,
|
|
129
|
+
"mode": 0,
|
|
130
|
+
"inputs": [],
|
|
131
|
+
"outputs": [],
|
|
132
|
+
"title": "Note: how to use this workflow",
|
|
133
|
+
"properties": {},
|
|
134
|
+
"widgets_values": [
|
|
135
|
+
"Huanyuan3D 2.1 is an image-to-3D model tool. It can turn any of your images into a 3D model.\n\n## For Comfy Cloud users\n\nIf you are using [cloud.comfy.org](http://cloud.comfy.org/):\n1. Since the workflow in the Cloud will have the input image ready, for the first run, you can just click the run button to see what happens. \n\n2. Try to upload your own image, the input image require clean simple backgroud.\n3. You can also find some other 3D workflows in the left-sidebar templates.\n\n## For local ComfyUI users\n\nIf you are using this workflow locally, you need to download all the models to your local storage. \n\nUse the right-click menu to edit the subgraph. Then you can find the original workflow and notes inside it.\nYou can find the tutorial and original workflow [here](https://raw.githubusercontent.com/Comfy-Org/workflow_templates/refs/heads/main/templates/3d_hunyuan3d-v2.1.json)."
|
|
136
|
+
],
|
|
137
|
+
"color": "#432",
|
|
138
|
+
"bgcolor": "#653"
|
|
139
|
+
},
|
|
116
140
|
{
|
|
117
141
|
"id": 19,
|
|
118
142
|
"type": "d7344a5a-fd0d-4b8a-b5e1-1237e5b21937",
|
|
@@ -122,7 +146,7 @@
|
|
|
122
146
|
],
|
|
123
147
|
"size": [
|
|
124
148
|
340,
|
|
125
|
-
|
|
149
|
+
58
|
|
126
150
|
],
|
|
127
151
|
"flags": {},
|
|
128
152
|
"order": 3,
|
|
@@ -156,30 +180,6 @@
|
|
|
156
180
|
"widgets_values": [
|
|
157
181
|
"hunyuan_3d_v2.1.safetensors"
|
|
158
182
|
]
|
|
159
|
-
},
|
|
160
|
-
{
|
|
161
|
-
"id": 20,
|
|
162
|
-
"type": "MarkdownNote",
|
|
163
|
-
"pos": [
|
|
164
|
-
-360,
|
|
165
|
-
-410
|
|
166
|
-
],
|
|
167
|
-
"size": [
|
|
168
|
-
440,
|
|
169
|
-
270
|
|
170
|
-
],
|
|
171
|
-
"flags": {},
|
|
172
|
-
"order": 2,
|
|
173
|
-
"mode": 0,
|
|
174
|
-
"inputs": [],
|
|
175
|
-
"outputs": [],
|
|
176
|
-
"title": "Note: how to use this workflow",
|
|
177
|
-
"properties": {},
|
|
178
|
-
"widgets_values": [
|
|
179
|
-
"Huanyuan3D 2.1 is an image-to-3D model tool. It can turn any of your images into a 3D model.\n\n## For Comfy Cloud users\n\nIf you are using [cloud.comfy.org](http://cloud.comfy.org/):\n1. Since the workflow in the Cloud will have the input image ready, for the first run, you can just click the run button to see what happens. \n\n2. Try to upload your own image, the input image require clean simple backgroud.\n3. You can also find some other 3D workflows in the left-sidebar templates.\n\n## For local ComfyUI users\n\nIf you are using this workflow locally, you need to download all the models to your local storage. \n\nUse the right-click menu to edit the subgraph. Then you can find the original workflow and notes inside it.\nYou can find the tutorial and original workflow [here](https://raw.githubusercontent.com/Comfy-Org/workflow_templates/refs/heads/main/templates/3d_hunyuan3d-v2.1.json)."
|
|
180
|
-
],
|
|
181
|
-
"color": "#432",
|
|
182
|
-
"bgcolor": "#653"
|
|
183
183
|
}
|
|
184
184
|
],
|
|
185
185
|
"links": [
|
|
@@ -310,9 +310,9 @@
|
|
|
310
310
|
}
|
|
311
311
|
],
|
|
312
312
|
"properties": {
|
|
313
|
-
"Node name for S&R": "ModelSamplingAuraFlow",
|
|
314
313
|
"cnr_id": "comfy-core",
|
|
315
|
-
"ver": "0.3.59"
|
|
314
|
+
"ver": "0.3.59",
|
|
315
|
+
"Node name for S&R": "ModelSamplingAuraFlow"
|
|
316
316
|
},
|
|
317
317
|
"widgets_values": [
|
|
318
318
|
1
|
|
@@ -369,9 +369,9 @@
|
|
|
369
369
|
}
|
|
370
370
|
],
|
|
371
371
|
"properties": {
|
|
372
|
-
"Node name for S&R": "KSampler",
|
|
373
372
|
"cnr_id": "comfy-core",
|
|
374
|
-
"ver": "0.3.59"
|
|
373
|
+
"ver": "0.3.59",
|
|
374
|
+
"Node name for S&R": "KSampler"
|
|
375
375
|
},
|
|
376
376
|
"widgets_values": [
|
|
377
377
|
894796671366012,
|
|
@@ -422,9 +422,9 @@
|
|
|
422
422
|
}
|
|
423
423
|
],
|
|
424
424
|
"properties": {
|
|
425
|
-
"Node name for S&R": "CLIPVisionEncode",
|
|
426
425
|
"cnr_id": "comfy-core",
|
|
427
|
-
"ver": "0.3.59"
|
|
426
|
+
"ver": "0.3.59",
|
|
427
|
+
"Node name for S&R": "CLIPVisionEncode"
|
|
428
428
|
},
|
|
429
429
|
"widgets_values": [
|
|
430
430
|
"center"
|
|
@@ -471,9 +471,9 @@
|
|
|
471
471
|
}
|
|
472
472
|
],
|
|
473
473
|
"properties": {
|
|
474
|
-
"Node name for S&R": "Hunyuan3Dv2Conditioning",
|
|
475
474
|
"cnr_id": "comfy-core",
|
|
476
|
-
"ver": "0.3.59"
|
|
475
|
+
"ver": "0.3.59",
|
|
476
|
+
"Node name for S&R": "Hunyuan3Dv2Conditioning"
|
|
477
477
|
},
|
|
478
478
|
"widgets_values": []
|
|
479
479
|
},
|
|
@@ -503,9 +503,9 @@
|
|
|
503
503
|
}
|
|
504
504
|
],
|
|
505
505
|
"properties": {
|
|
506
|
-
"Node name for S&R": "EmptyLatentHunyuan3Dv2",
|
|
507
506
|
"cnr_id": "comfy-core",
|
|
508
|
-
"ver": "0.3.59"
|
|
507
|
+
"ver": "0.3.59",
|
|
508
|
+
"Node name for S&R": "EmptyLatentHunyuan3Dv2"
|
|
509
509
|
},
|
|
510
510
|
"widgets_values": [
|
|
511
511
|
4096,
|
|
@@ -568,9 +568,9 @@
|
|
|
568
568
|
}
|
|
569
569
|
],
|
|
570
570
|
"properties": {
|
|
571
|
-
"Node name for S&R": "VoxelToMesh",
|
|
572
571
|
"cnr_id": "comfy-core",
|
|
573
|
-
"ver": "0.3.59"
|
|
572
|
+
"ver": "0.3.59",
|
|
573
|
+
"Node name for S&R": "VoxelToMesh"
|
|
574
574
|
},
|
|
575
575
|
"widgets_values": [
|
|
576
576
|
"surface net",
|
|
@@ -616,9 +616,9 @@
|
|
|
616
616
|
}
|
|
617
617
|
],
|
|
618
618
|
"properties": {
|
|
619
|
-
"Node name for S&R": "VAEDecodeHunyuan3D",
|
|
620
619
|
"cnr_id": "comfy-core",
|
|
621
|
-
"ver": "0.3.59"
|
|
620
|
+
"ver": "0.3.59",
|
|
621
|
+
"Node name for S&R": "VAEDecodeHunyuan3D"
|
|
622
622
|
},
|
|
623
623
|
"widgets_values": [
|
|
624
624
|
8000,
|
|
@@ -677,9 +677,9 @@
|
|
|
677
677
|
}
|
|
678
678
|
],
|
|
679
679
|
"properties": {
|
|
680
|
-
"Node name for S&R": "ImageOnlyCheckpointLoader",
|
|
681
680
|
"cnr_id": "comfy-core",
|
|
682
681
|
"ver": "0.3.59",
|
|
682
|
+
"Node name for S&R": "ImageOnlyCheckpointLoader",
|
|
683
683
|
"models": [
|
|
684
684
|
{
|
|
685
685
|
"name": "hunyuan_3d_v2.1.safetensors",
|
|
@@ -693,34 +693,7 @@
|
|
|
693
693
|
]
|
|
694
694
|
}
|
|
695
695
|
],
|
|
696
|
-
"groups": [
|
|
697
|
-
{
|
|
698
|
-
"id": 1,
|
|
699
|
-
"title": "Step 1 - Load model",
|
|
700
|
-
"bounding": [
|
|
701
|
-
440,
|
|
702
|
-
-670,
|
|
703
|
-
356.3746337890625,
|
|
704
|
-
183.60000610351562
|
|
705
|
-
],
|
|
706
|
-
"color": "#3f789e",
|
|
707
|
-
"font_size": 24,
|
|
708
|
-
"flags": {}
|
|
709
|
-
},
|
|
710
|
-
{
|
|
711
|
-
"id": 2,
|
|
712
|
-
"title": "Step 2 - Upload image here",
|
|
713
|
-
"bounding": [
|
|
714
|
-
440,
|
|
715
|
-
-470,
|
|
716
|
-
350,
|
|
717
|
-
410
|
|
718
|
-
],
|
|
719
|
-
"color": "#3f789e",
|
|
720
|
-
"font_size": 24,
|
|
721
|
-
"flags": {}
|
|
722
|
-
}
|
|
723
|
-
],
|
|
696
|
+
"groups": [],
|
|
724
697
|
"links": [
|
|
725
698
|
{
|
|
726
699
|
"id": 16,
|
|
@@ -834,13 +807,13 @@
|
|
|
834
807
|
"config": {},
|
|
835
808
|
"extra": {
|
|
836
809
|
"ds": {
|
|
837
|
-
"scale": 0.
|
|
810
|
+
"scale": 0.5017504203807984,
|
|
838
811
|
"offset": [
|
|
839
|
-
|
|
840
|
-
|
|
812
|
+
1097.7408891572666,
|
|
813
|
+
944.3281959259452
|
|
841
814
|
]
|
|
842
815
|
},
|
|
843
|
-
"frontendVersion": "1.
|
|
816
|
+
"frontendVersion": "1.28.7",
|
|
844
817
|
"VHS_latentpreview": false,
|
|
845
818
|
"VHS_latentpreviewrate": 0,
|
|
846
819
|
"VHS_MetadataImage": true,
|
|
@@ -2,7 +2,7 @@
|
|
|
2
2
|
"id": "7cbcec68-7fa6-47bb-a38a-da689949a001",
|
|
3
3
|
"revision": 0,
|
|
4
4
|
"last_node_id": 191,
|
|
5
|
-
"last_link_id":
|
|
5
|
+
"last_link_id": 298,
|
|
6
6
|
"nodes": [
|
|
7
7
|
{
|
|
8
8
|
"id": 39,
|
|
@@ -340,13 +340,13 @@
|
|
|
340
340
|
{
|
|
341
341
|
"name": "image1",
|
|
342
342
|
"type": "IMAGE",
|
|
343
|
-
"link":
|
|
343
|
+
"link": 297
|
|
344
344
|
},
|
|
345
345
|
{
|
|
346
346
|
"name": "image2",
|
|
347
347
|
"shape": 7,
|
|
348
348
|
"type": "IMAGE",
|
|
349
|
-
"link":
|
|
349
|
+
"link": 298
|
|
350
350
|
}
|
|
351
351
|
],
|
|
352
352
|
"outputs": [
|
|
@@ -730,30 +730,6 @@
|
|
|
730
730
|
"color": "#432",
|
|
731
731
|
"bgcolor": "#653"
|
|
732
732
|
},
|
|
733
|
-
{
|
|
734
|
-
"id": 186,
|
|
735
|
-
"type": "MarkdownNote",
|
|
736
|
-
"pos": [
|
|
737
|
-
-960,
|
|
738
|
-
500
|
|
739
|
-
],
|
|
740
|
-
"size": [
|
|
741
|
-
520,
|
|
742
|
-
460
|
|
743
|
-
],
|
|
744
|
-
"flags": {},
|
|
745
|
-
"order": 8,
|
|
746
|
-
"mode": 0,
|
|
747
|
-
"inputs": [],
|
|
748
|
-
"outputs": [],
|
|
749
|
-
"title": "Flux Kontext Prompt Techniques",
|
|
750
|
-
"properties": {},
|
|
751
|
-
"widgets_values": [
|
|
752
|
-
"\n## Flux Kontext Prompt Techniques\n\n### 1. Basic Modifications\n- Simple and direct: `\"Change the car color to red\"`\n- Maintain style: `\"Change to daytime while maintaining the same style of the painting\"`\n\n### 2. Style Transfer\n**Principles:**\n- Clearly name style: `\"Transform to Bauhaus art style\"`\n- Describe characteristics: `\"Transform to oil painting with visible brushstrokes, thick paint texture\"`\n- Preserve composition: `\"Change to Bauhaus style while maintaining the original composition\"`\n\n### 3. Character Consistency\n**Framework:**\n- Specific description: `\"The woman with short black hair\"` instead of \"she\"\n- Preserve features: `\"while maintaining the same facial features, hairstyle, and expression\"`\n- Step-by-step modifications: Change background first, then actions\n\n### 4. Text Editing\n- Use quotes: `\"Replace 'joy' with 'BFL'\"`\n- Maintain format: `\"Replace text while maintaining the same font style\"`\n\n## Common Problem Solutions\n\n### Character Changes Too Much\n❌ Wrong: `\"Transform the person into a Viking\"`\n✅ Correct: `\"Change the clothes to be a viking warrior while preserving facial features\"`\n\n### Composition Position Changes\n❌ Wrong: `\"Put him on a beach\"`\n✅ Correct: `\"Change the background to a beach while keeping the person in the exact same position, scale, and pose\"`\n\n### Style Application Inaccuracy\n❌ Wrong: `\"Make it a sketch\"`\n✅ Correct: `\"Convert to pencil sketch with natural graphite lines, cross-hatching, and visible paper texture\"`\n\n## Core Principles\n\n1. **Be Specific and Clear** - Use precise descriptions, avoid vague terms\n2. **Step-by-step Editing** - Break complex modifications into multiple simple steps\n3. **Explicit Preservation** - State what should remain unchanged\n4. **Verb Selection** - Use \"change\", \"replace\" rather than \"transform\"\n\n## Best Practice Templates\n\n**Object Modification:**\n`\"Change [object] to [new state], keep [content to preserve] unchanged\"`\n\n**Style Transfer:**\n`\"Transform to [specific style], while maintaining [composition/character/other] unchanged\"`\n\n**Background Replacement:**\n`\"Change the background to [new background], keep the subject in the exact same position and pose\"`\n\n**Text Editing:**\n`\"Replace '[original text]' with '[new text]', maintain the same font style\"`\n\n> **Remember:** The more specific, the better. Kontext excels at understanding detailed instructions and maintaining consistency. "
|
|
753
|
-
],
|
|
754
|
-
"color": "#432",
|
|
755
|
-
"bgcolor": "#653"
|
|
756
|
-
},
|
|
757
733
|
{
|
|
758
734
|
"id": 190,
|
|
759
735
|
"type": "LoadImage",
|
|
@@ -766,7 +742,7 @@
|
|
|
766
742
|
330
|
|
767
743
|
],
|
|
768
744
|
"flags": {},
|
|
769
|
-
"order":
|
|
745
|
+
"order": 8,
|
|
770
746
|
"mode": 0,
|
|
771
747
|
"inputs": [],
|
|
772
748
|
"outputs": [
|
|
@@ -774,7 +750,7 @@
|
|
|
774
750
|
"name": "IMAGE",
|
|
775
751
|
"type": "IMAGE",
|
|
776
752
|
"links": [
|
|
777
|
-
|
|
753
|
+
297
|
|
778
754
|
]
|
|
779
755
|
},
|
|
780
756
|
{
|
|
@@ -805,7 +781,7 @@
|
|
|
805
781
|
330
|
|
806
782
|
],
|
|
807
783
|
"flags": {},
|
|
808
|
-
"order":
|
|
784
|
+
"order": 9,
|
|
809
785
|
"mode": 4,
|
|
810
786
|
"inputs": [],
|
|
811
787
|
"outputs": [
|
|
@@ -813,7 +789,7 @@
|
|
|
813
789
|
"name": "IMAGE",
|
|
814
790
|
"type": "IMAGE",
|
|
815
791
|
"links": [
|
|
816
|
-
|
|
792
|
+
298
|
|
817
793
|
]
|
|
818
794
|
},
|
|
819
795
|
{
|
|
@@ -831,6 +807,30 @@
|
|
|
831
807
|
"flux_kontext_dev_basic_input_image.jpg",
|
|
832
808
|
"image"
|
|
833
809
|
]
|
|
810
|
+
},
|
|
811
|
+
{
|
|
812
|
+
"id": 186,
|
|
813
|
+
"type": "MarkdownNote",
|
|
814
|
+
"pos": [
|
|
815
|
+
-960,
|
|
816
|
+
500
|
|
817
|
+
],
|
|
818
|
+
"size": [
|
|
819
|
+
510,
|
|
820
|
+
670
|
|
821
|
+
],
|
|
822
|
+
"flags": {},
|
|
823
|
+
"order": 10,
|
|
824
|
+
"mode": 0,
|
|
825
|
+
"inputs": [],
|
|
826
|
+
"outputs": [],
|
|
827
|
+
"title": "Flux Kontext Prompt Techniques",
|
|
828
|
+
"properties": {},
|
|
829
|
+
"widgets_values": [
|
|
830
|
+
"\n## Flux Kontext Prompt Techniques\n\n### 1. Basic Modifications\n- Simple and direct: `\"Change the car color to red\"`\n- Maintain style: `\"Change to daytime while maintaining the same style of the painting\"`\n\n### 2. Style Transfer\n**Principles:**\n- Clearly name style: `\"Transform to Bauhaus art style\"`\n- Describe characteristics: `\"Transform to oil painting with visible brushstrokes, thick paint texture\"`\n- Preserve composition: `\"Change to Bauhaus style while maintaining the original composition\"`\n\n### 3. Character Consistency\n**Framework:**\n- Specific description: `\"The woman with short black hair\"` instead of \"she\"\n- Preserve features: `\"while maintaining the same facial features, hairstyle, and expression\"`\n- Step-by-step modifications: Change background first, then actions\n\n### 4. Text Editing\n- Use quotes: `\"Replace 'joy' with 'BFL'\"`\n- Maintain format: `\"Replace text while maintaining the same font style\"`\n\n## Common Problem Solutions\n\n### Character Changes Too Much\n❌ Wrong: `\"Transform the person into a Viking\"`\n✅ Correct: `\"Change the clothes to be a viking warrior while preserving facial features\"`\n\n### Composition Position Changes\n❌ Wrong: `\"Put him on a beach\"`\n✅ Correct: `\"Change the background to a beach while keeping the person in the exact same position, scale, and pose\"`\n\n### Style Application Inaccuracy\n❌ Wrong: `\"Make it a sketch\"`\n✅ Correct: `\"Convert to pencil sketch with natural graphite lines, cross-hatching, and visible paper texture\"`\n\n## Core Principles\n\n1. **Be Specific and Clear** - Use precise descriptions, avoid vague terms\n2. **Step-by-step Editing** - Break complex modifications into multiple simple steps\n3. **Explicit Preservation** - State what should remain unchanged\n4. **Verb Selection** - Use \"change\", \"replace\" rather than \"transform\"\n\n## Best Practice Templates\n\n**Object Modification:**\n`\"Change [object] to [new state], keep [content to preserve] unchanged\"`\n\n**Style Transfer:**\n`\"Transform to [specific style], while maintaining [composition/character/other] unchanged\"`\n\n**Background Replacement:**\n`\"Change the background to [new background], keep the subject in the exact same position and pose\"`\n\n**Text Editing:**\n`\"Replace '[original text]' with '[new text]', maintain the same font style\"`\n\n> **Remember:** The more specific, the better. Kontext excels at understanding detailed instructions and maintaining consistency. "
|
|
831
|
+
],
|
|
832
|
+
"color": "#432",
|
|
833
|
+
"bgcolor": "#653"
|
|
834
834
|
}
|
|
835
835
|
],
|
|
836
836
|
"links": [
|
|
@@ -963,19 +963,19 @@
|
|
|
963
963
|
"CONDITIONING"
|
|
964
964
|
],
|
|
965
965
|
[
|
|
966
|
-
|
|
966
|
+
297,
|
|
967
967
|
190,
|
|
968
968
|
0,
|
|
969
969
|
146,
|
|
970
|
-
|
|
970
|
+
0,
|
|
971
971
|
"IMAGE"
|
|
972
972
|
],
|
|
973
973
|
[
|
|
974
|
-
|
|
974
|
+
298,
|
|
975
975
|
191,
|
|
976
976
|
0,
|
|
977
977
|
146,
|
|
978
|
-
|
|
978
|
+
1,
|
|
979
979
|
"IMAGE"
|
|
980
980
|
]
|
|
981
981
|
],
|
|
@@ -1036,13 +1036,13 @@
|
|
|
1036
1036
|
"config": {},
|
|
1037
1037
|
"extra": {
|
|
1038
1038
|
"ds": {
|
|
1039
|
-
"scale": 0.
|
|
1039
|
+
"scale": 0.5909232551952691,
|
|
1040
1040
|
"offset": [
|
|
1041
|
-
|
|
1042
|
-
|
|
1041
|
+
1279.4241547910187,
|
|
1042
|
+
33.095262680409
|
|
1043
1043
|
]
|
|
1044
1044
|
},
|
|
1045
|
-
"frontendVersion": "1.
|
|
1045
|
+
"frontendVersion": "1.28.7",
|
|
1046
1046
|
"groupNodes": {},
|
|
1047
1047
|
"VHS_latentpreview": false,
|
|
1048
1048
|
"VHS_latentpreviewrate": 0,
|
|
@@ -7,7 +7,7 @@
|
|
|
7
7
|
"templates": [
|
|
8
8
|
{
|
|
9
9
|
"name": "01_qwen_t2i_subgraphed",
|
|
10
|
-
"title": "Texto a imagen
|
|
10
|
+
"title": "Texto a imagen(Nuevo)",
|
|
11
11
|
"mediaType": "image",
|
|
12
12
|
"mediaSubtype": "webp",
|
|
13
13
|
"description": "Genera imágenes a partir de indicaciones de texto usando el modelo Qwen-Image",
|
|
@@ -19,7 +19,7 @@
|
|
|
19
19
|
},
|
|
20
20
|
{
|
|
21
21
|
"name": "02_qwen_Image_edit_subgraphed",
|
|
22
|
-
"title": "Edición de imágenes
|
|
22
|
+
"title": "Edición de imágenes(Nuevo)",
|
|
23
23
|
"mediaType": "image",
|
|
24
24
|
"mediaSubtype": "webp",
|
|
25
25
|
"description": "Edita tus imágenes con Qwen-Image-Edit",
|
|
@@ -31,7 +31,7 @@
|
|
|
31
31
|
},
|
|
32
32
|
{
|
|
33
33
|
"name": "03_video_wan2_2_14B_i2v_subgraphed",
|
|
34
|
-
"title": "Imagen a Video
|
|
34
|
+
"title": "Imagen a Video(Nuevo)",
|
|
35
35
|
"description": "Genera videos a partir de una imagen usando Wan2.2 14B",
|
|
36
36
|
"mediaType": "image",
|
|
37
37
|
"mediaSubtype": "webp",
|
|
@@ -43,7 +43,7 @@
|
|
|
43
43
|
},
|
|
44
44
|
{
|
|
45
45
|
"name": "04_hunyuan_3d_2.1_subgraphed",
|
|
46
|
-
"title": "Imagen a 3D
|
|
46
|
+
"title": "Imagen a 3D(Nuevo)",
|
|
47
47
|
"mediaType": "image",
|
|
48
48
|
"mediaSubtype": "webp",
|
|
49
49
|
"description": "Genera modelos 3D a partir de imágenes únicas usando Hunyuan3D 2.0.",
|
|
@@ -55,7 +55,7 @@
|
|
|
55
55
|
},
|
|
56
56
|
{
|
|
57
57
|
"name": "05_audio_ace_step_1_t2a_song_subgraphed",
|
|
58
|
-
"title": "Texto a audio
|
|
58
|
+
"title": "Texto a audio(Nuevo)",
|
|
59
59
|
"mediaType": "image",
|
|
60
60
|
"mediaSubtype": "webp",
|
|
61
61
|
"description": "Genera audio a partir de indicaciones de texto usando ACE-Step v1",
|
|
@@ -7,7 +7,7 @@
|
|
|
7
7
|
"templates": [
|
|
8
8
|
{
|
|
9
9
|
"name": "01_qwen_t2i_subgraphed",
|
|
10
|
-
"title": "Texte en image
|
|
10
|
+
"title": "Texte en image(Nouveau)",
|
|
11
11
|
"mediaType": "image",
|
|
12
12
|
"mediaSubtype": "webp",
|
|
13
13
|
"description": "Générez des images à partir d'invites textuelles avec le modèle Qwen-Image",
|
|
@@ -19,7 +19,7 @@
|
|
|
19
19
|
},
|
|
20
20
|
{
|
|
21
21
|
"name": "02_qwen_Image_edit_subgraphed",
|
|
22
|
-
"title": "Édition d'image
|
|
22
|
+
"title": "Édition d'image(Nouveau)",
|
|
23
23
|
"mediaType": "image",
|
|
24
24
|
"mediaSubtype": "webp",
|
|
25
25
|
"description": "Éditez vos images avec Qwen-Image-Edit",
|
|
@@ -31,7 +31,7 @@
|
|
|
31
31
|
},
|
|
32
32
|
{
|
|
33
33
|
"name": "03_video_wan2_2_14B_i2v_subgraphed",
|
|
34
|
-
"title": "Image en Vidéo
|
|
34
|
+
"title": "Image en Vidéo(Nouveau)",
|
|
35
35
|
"description": "Générez des vidéos à partir d’une image avec Wan2.2 14B",
|
|
36
36
|
"mediaType": "image",
|
|
37
37
|
"mediaSubtype": "webp",
|
|
@@ -43,7 +43,7 @@
|
|
|
43
43
|
},
|
|
44
44
|
{
|
|
45
45
|
"name": "04_hunyuan_3d_2.1_subgraphed",
|
|
46
|
-
"title": "Image vers 3D
|
|
46
|
+
"title": "Image vers 3D(Nouveau)",
|
|
47
47
|
"mediaType": "image",
|
|
48
48
|
"mediaSubtype": "webp",
|
|
49
49
|
"description": "Générez des modèles 3D à partir d'une seule image avec Hunyuan3D 2.0.",
|
|
@@ -55,7 +55,7 @@
|
|
|
55
55
|
},
|
|
56
56
|
{
|
|
57
57
|
"name": "05_audio_ace_step_1_t2a_song_subgraphed",
|
|
58
|
-
"title": "Texte en audio
|
|
58
|
+
"title": "Texte en audio(Nouveau)",
|
|
59
59
|
"mediaType": "image",
|
|
60
60
|
"mediaSubtype": "webp",
|
|
61
61
|
"description": "Générez de l'audio à partir d'invites textuelles avec ACE-Step v1",
|
|
@@ -7,7 +7,7 @@
|
|
|
7
7
|
"templates": [
|
|
8
8
|
{
|
|
9
9
|
"name": "01_qwen_t2i_subgraphed",
|
|
10
|
-
"title": "テキストから画像へ
|
|
10
|
+
"title": "テキストから画像へ(新)",
|
|
11
11
|
"mediaType": "image",
|
|
12
12
|
"mediaSubtype": "webp",
|
|
13
13
|
"description": "Qwen-Imageモデルを使ってテキストプロンプトから画像を生成します",
|
|
@@ -19,7 +19,7 @@
|
|
|
19
19
|
},
|
|
20
20
|
{
|
|
21
21
|
"name": "02_qwen_Image_edit_subgraphed",
|
|
22
|
-
"title": "画像編集
|
|
22
|
+
"title": "画像編集(新)",
|
|
23
23
|
"mediaType": "image",
|
|
24
24
|
"mediaSubtype": "webp",
|
|
25
25
|
"description": "Qwen-Image-Edit で画像を編集しましょう",
|
|
@@ -31,7 +31,7 @@
|
|
|
31
31
|
},
|
|
32
32
|
{
|
|
33
33
|
"name": "03_video_wan2_2_14B_i2v_subgraphed",
|
|
34
|
-
"title": "画像から動画へ
|
|
34
|
+
"title": "画像から動画へ(新)",
|
|
35
35
|
"description": "Wan2.2 14B を使用して画像から動画を生成します",
|
|
36
36
|
"mediaType": "image",
|
|
37
37
|
"mediaSubtype": "webp",
|
|
@@ -43,7 +43,7 @@
|
|
|
43
43
|
},
|
|
44
44
|
{
|
|
45
45
|
"name": "04_hunyuan_3d_2.1_subgraphed",
|
|
46
|
-
"title": "画像から3D
|
|
46
|
+
"title": "画像から3Dへ(新)",
|
|
47
47
|
"mediaType": "image",
|
|
48
48
|
"mediaSubtype": "webp",
|
|
49
49
|
"description": "Hunyuan3D 2.0を使って1枚の画像から3Dモデルを生成します。",
|
|
@@ -55,7 +55,7 @@
|
|
|
55
55
|
},
|
|
56
56
|
{
|
|
57
57
|
"name": "05_audio_ace_step_1_t2a_song_subgraphed",
|
|
58
|
-
"title": "
|
|
58
|
+
"title": "テキストからオーディオへ(新)",
|
|
59
59
|
"mediaType": "image",
|
|
60
60
|
"mediaSubtype": "webp",
|
|
61
61
|
"description": "ACE-Step v1 を使ってテキストプロンプトからオーディオを生成する",
|
|
@@ -7,7 +7,7 @@
|
|
|
7
7
|
"templates": [
|
|
8
8
|
{
|
|
9
9
|
"name": "01_qwen_t2i_subgraphed",
|
|
10
|
-
"title": "Text to Image
|
|
10
|
+
"title": "Text to Image (New)",
|
|
11
11
|
"mediaType": "image",
|
|
12
12
|
"mediaSubtype": "webp",
|
|
13
13
|
"description": "Generate images from text prompts using the Qwen-Image model",
|
|
@@ -19,7 +19,7 @@
|
|
|
19
19
|
},
|
|
20
20
|
{
|
|
21
21
|
"name": "02_qwen_Image_edit_subgraphed",
|
|
22
|
-
"title": "Image Editing
|
|
22
|
+
"title": "Image Editing (New)",
|
|
23
23
|
"mediaType": "image",
|
|
24
24
|
"mediaSubtype": "webp",
|
|
25
25
|
"description": "Edit your images with Qwen-Image-Edit, the latest OSS model",
|
|
@@ -31,7 +31,7 @@
|
|
|
31
31
|
},
|
|
32
32
|
{
|
|
33
33
|
"name": "03_video_wan2_2_14B_i2v_subgraphed",
|
|
34
|
-
"title": "Image to Video
|
|
34
|
+
"title": "Image to Video (New)",
|
|
35
35
|
"description": "Generate videos from an input image using Wan2.2 14B",
|
|
36
36
|
"mediaType": "image",
|
|
37
37
|
"mediaSubtype": "webp",
|
|
@@ -43,7 +43,7 @@
|
|
|
43
43
|
},
|
|
44
44
|
{
|
|
45
45
|
"name": "04_hunyuan_3d_2.1_subgraphed",
|
|
46
|
-
"title": "
|
|
46
|
+
"title": "Image to 3D (New)",
|
|
47
47
|
"mediaType": "image",
|
|
48
48
|
"mediaSubtype": "webp",
|
|
49
49
|
"description": "Generate 3D models from single images using Hunyuan3D 2.1.",
|
|
@@ -55,7 +55,7 @@
|
|
|
55
55
|
},
|
|
56
56
|
{
|
|
57
57
|
"name": "05_audio_ace_step_1_t2a_song_subgraphed",
|
|
58
|
-
"title": "Text to Song
|
|
58
|
+
"title": "Text to Song (New)",
|
|
59
59
|
"mediaType": "image",
|
|
60
60
|
"mediaSubtype": "webp",
|
|
61
61
|
"description": "Generate songs from text prompts using ACE-Step v1",
|
|
@@ -607,7 +607,7 @@
|
|
|
607
607
|
"mediaSubtype": "webp",
|
|
608
608
|
"description": "Generate images guided by depth information using Flux LoRA.",
|
|
609
609
|
"thumbnailVariant": "hoverDissolve",
|
|
610
|
-
"tutorialUrl": "
|
|
610
|
+
"tutorialUrl": "https://docs.comfy.org/tutorials/flux/flux-1-controlnet",
|
|
611
611
|
"tags": ["Image to Image", "ControlNet", "Image"],
|
|
612
612
|
"models": ["Flux", "BFL"],
|
|
613
613
|
"date": "2025-03-01",
|
|
@@ -7,7 +7,7 @@
|
|
|
7
7
|
"templates": [
|
|
8
8
|
{
|
|
9
9
|
"name": "01_qwen_t2i_subgraphed",
|
|
10
|
-
"title": "텍스트 → 이미지
|
|
10
|
+
"title": "텍스트 → 이미지(신규)",
|
|
11
11
|
"mediaType": "image",
|
|
12
12
|
"mediaSubtype": "webp",
|
|
13
13
|
"description": "Qwen-Image 모델로 텍스트 프롬프트에서 이미지를 생성합니다",
|
|
@@ -19,7 +19,7 @@
|
|
|
19
19
|
},
|
|
20
20
|
{
|
|
21
21
|
"name": "02_qwen_Image_edit_subgraphed",
|
|
22
|
-
"title": "이미지 편집
|
|
22
|
+
"title": "이미지 편집(신규)",
|
|
23
23
|
"mediaType": "image",
|
|
24
24
|
"mediaSubtype": "webp",
|
|
25
25
|
"description": "Qwen-Image-Edit로 이미지를 편집하세요",
|
|
@@ -31,7 +31,7 @@
|
|
|
31
31
|
},
|
|
32
32
|
{
|
|
33
33
|
"name": "03_video_wan2_2_14B_i2v_subgraphed",
|
|
34
|
-
"title": "이미지에서 비디오로
|
|
34
|
+
"title": "이미지에서 비디오로(신규)",
|
|
35
35
|
"description": "Wan2.2 14B을 사용하여 입력 이미지를 비디오로 생성",
|
|
36
36
|
"mediaType": "image",
|
|
37
37
|
"mediaSubtype": "webp",
|
|
@@ -43,7 +43,7 @@
|
|
|
43
43
|
},
|
|
44
44
|
{
|
|
45
45
|
"name": "04_hunyuan_3d_2.1_subgraphed",
|
|
46
|
-
"title": "이미지→3D
|
|
46
|
+
"title": "이미지→3D(신규)",
|
|
47
47
|
"mediaType": "image",
|
|
48
48
|
"mediaSubtype": "webp",
|
|
49
49
|
"description": "Hunyuan3D 2.0을 사용하여 단일 이미지에서 3D 모델을 생성합니다.",
|
|
@@ -55,7 +55,7 @@
|
|
|
55
55
|
},
|
|
56
56
|
{
|
|
57
57
|
"name": "05_audio_ace_step_1_t2a_song_subgraphed",
|
|
58
|
-
"title": "텍스트→오디오
|
|
58
|
+
"title": "텍스트→오디오(신규)",
|
|
59
59
|
"mediaType": "image",
|
|
60
60
|
"mediaSubtype": "webp",
|
|
61
61
|
"description": "ACE-Step v1로 텍스트 프롬프트에서 오디오 생성",
|