comfyui-workflow-templates 0.1.52__py3-none-any.whl → 0.1.54__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of comfyui-workflow-templates might be problematic. Click here for more details.

Files changed (21) hide show
  1. comfyui_workflow_templates/templates/image_qwen_image.json +300 -207
  2. comfyui_workflow_templates/templates/index.es.json +2357 -0
  3. comfyui_workflow_templates/templates/index.fr.json +2357 -0
  4. comfyui_workflow_templates/templates/index.ja.json +2357 -0
  5. comfyui_workflow_templates/templates/index.json +95 -70
  6. comfyui_workflow_templates/templates/index.ko.json +2357 -0
  7. comfyui_workflow_templates/templates/index.ru.json +2357 -0
  8. comfyui_workflow_templates/templates/index.schema.json +5 -1
  9. comfyui_workflow_templates/templates/index.zh-TW.json +2357 -0
  10. comfyui_workflow_templates/templates/index.zh.json +2357 -0
  11. comfyui_workflow_templates/templates/video_wan2_2_14B_flf2v.json +1846 -307
  12. comfyui_workflow_templates/templates/video_wan2_2_14B_i2v.json +1409 -290
  13. comfyui_workflow_templates/templates/video_wan2_2_14B_t2v-1.webp +0 -0
  14. comfyui_workflow_templates/templates/video_wan2_2_14B_t2v.json +1360 -372
  15. comfyui_workflow_templates/templates/video_wan2_2_fun_inp_14B-1.webp +0 -0
  16. comfyui_workflow_templates/templates/video_wan2_2_fun_inp_14B.json +2549 -0
  17. {comfyui_workflow_templates-0.1.52.dist-info → comfyui_workflow_templates-0.1.54.dist-info}/METADATA +1 -1
  18. {comfyui_workflow_templates-0.1.52.dist-info → comfyui_workflow_templates-0.1.54.dist-info}/RECORD +21 -12
  19. {comfyui_workflow_templates-0.1.52.dist-info → comfyui_workflow_templates-0.1.54.dist-info}/WHEEL +0 -0
  20. {comfyui_workflow_templates-0.1.52.dist-info → comfyui_workflow_templates-0.1.54.dist-info}/licenses/LICENSE +0 -0
  21. {comfyui_workflow_templates-0.1.52.dist-info → comfyui_workflow_templates-0.1.54.dist-info}/top_level.txt +0 -0
@@ -1,8 +1,8 @@
1
1
  {
2
2
  "id": "91f6bbe2-ed41-4fd6-bac7-71d5b5864ecb",
3
3
  "revision": 0,
4
- "last_node_id": 71,
5
- "last_link_id": 128,
4
+ "last_node_id": 74,
5
+ "last_link_id": 130,
6
6
  "nodes": [
7
7
  {
8
8
  "id": 39,
@@ -30,9 +30,9 @@
30
30
  }
31
31
  ],
32
32
  "properties": {
33
+ "Node name for S&R": "VAELoader",
33
34
  "cnr_id": "comfy-core",
34
35
  "ver": "0.3.48",
35
- "Node name for S&R": "VAELoader",
36
36
  "models": [
37
37
  {
38
38
  "name": "qwen_image_vae.safetensors",
@@ -80,9 +80,9 @@
80
80
  }
81
81
  ],
82
82
  "properties": {
83
+ "Node name for S&R": "CLIPLoader",
83
84
  "cnr_id": "comfy-core",
84
85
  "ver": "0.3.48",
85
- "Node name for S&R": "CLIPLoader",
86
86
  "models": [
87
87
  {
88
88
  "name": "qwen_2.5_vl_7b_fp8_scaled.safetensors",
@@ -106,15 +106,15 @@
106
106
  ]
107
107
  },
108
108
  {
109
- "id": 37,
110
- "type": "UNETLoader",
109
+ "id": 58,
110
+ "type": "EmptySD3LatentImage",
111
111
  "pos": [
112
- 20,
113
- 50
112
+ 50,
113
+ 510
114
114
  ],
115
115
  "size": [
116
- 330,
117
- 90
116
+ 270,
117
+ 106
118
118
  ],
119
119
  "flags": {},
120
120
  "order": 2,
@@ -122,25 +122,17 @@
122
122
  "inputs": [],
123
123
  "outputs": [
124
124
  {
125
- "name": "MODEL",
126
- "type": "MODEL",
127
- "slot_index": 0,
125
+ "name": "LATENT",
126
+ "type": "LATENT",
128
127
  "links": [
129
- 126
128
+ 107
130
129
  ]
131
130
  }
132
131
  ],
133
132
  "properties": {
133
+ "Node name for S&R": "EmptySD3LatentImage",
134
134
  "cnr_id": "comfy-core",
135
135
  "ver": "0.3.48",
136
- "Node name for S&R": "UNETLoader",
137
- "models": [
138
- {
139
- "name": "qwen_image_fp8_e4m3fn.safetensors",
140
- "url": "https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/diffusion_models/qwen_image_fp8_e4m3fn.safetensors",
141
- "directory": "diffusion_models"
142
- }
143
- ],
144
136
  "enableTabs": false,
145
137
  "tabWidth": 65,
146
138
  "tabXOffset": 10,
@@ -151,10 +143,113 @@
151
143
  "widget_ue_connectable": {}
152
144
  },
153
145
  "widgets_values": [
154
- "qwen_image_fp8_e4m3fn.safetensors",
155
- "default"
146
+ 1328,
147
+ 1328,
148
+ 1
156
149
  ]
157
150
  },
151
+ {
152
+ "id": 6,
153
+ "type": "CLIPTextEncode",
154
+ "pos": [
155
+ 390,
156
+ 240
157
+ ],
158
+ "size": [
159
+ 422.84503173828125,
160
+ 164.31304931640625
161
+ ],
162
+ "flags": {},
163
+ "order": 9,
164
+ "mode": 0,
165
+ "inputs": [
166
+ {
167
+ "name": "clip",
168
+ "type": "CLIP",
169
+ "link": 74
170
+ }
171
+ ],
172
+ "outputs": [
173
+ {
174
+ "name": "CONDITIONING",
175
+ "type": "CONDITIONING",
176
+ "slot_index": 0,
177
+ "links": [
178
+ 46
179
+ ]
180
+ }
181
+ ],
182
+ "title": "CLIP Text Encode (Positive Prompt)",
183
+ "properties": {
184
+ "Node name for S&R": "CLIPTextEncode",
185
+ "cnr_id": "comfy-core",
186
+ "ver": "0.3.48",
187
+ "enableTabs": false,
188
+ "tabWidth": 65,
189
+ "tabXOffset": 10,
190
+ "hasSecondTab": false,
191
+ "secondTabText": "Send Back",
192
+ "secondTabOffset": 80,
193
+ "secondTabWidth": 65,
194
+ "widget_ue_connectable": {}
195
+ },
196
+ "widgets_values": [
197
+ "\"A vibrant, warm neon-lit street scene in Hong Kong at the afternoon, with a mix of colorful Chinese and English signs glowing brightly. The atmosphere is lively, cinematic, and rain-washed with reflections on the pavement. The colors are vivid, full of pink, blue, red, and green hues. Crowded buildings with overlapping neon signs. 1980s Hong Kong style. Signs include:\n\"龍鳳冰室\" \"金華燒臘\" \"HAPPY HAIR\" \"鴻運茶餐廳\" \"EASY BAR\" \"永發魚蛋粉\" \"添記粥麵\" \"SUNSHINE MOTEL\" \"美都餐室\" \"富記糖水\" \"太平館\" \"雅芳髮型屋\" \"STAR KTV\" \"銀河娛樂城\" \"百樂門舞廳\" \"BUBBLE CAFE\" \"萬豪麻雀館\" \"CITY LIGHTS BAR\" \"瑞祥香燭莊\" \"文記文具\" \"GOLDEN JADE HOTEL\" \"LOVELY BEAUTY\" \"合興百貨\" \"興旺電器\" And the background is warm yellow street and with all stores' lights on."
198
+ ],
199
+ "color": "#232",
200
+ "bgcolor": "#353"
201
+ },
202
+ {
203
+ "id": 7,
204
+ "type": "CLIPTextEncode",
205
+ "pos": [
206
+ 390,
207
+ 440
208
+ ],
209
+ "size": [
210
+ 425.27801513671875,
211
+ 180.6060791015625
212
+ ],
213
+ "flags": {},
214
+ "order": 10,
215
+ "mode": 0,
216
+ "inputs": [
217
+ {
218
+ "name": "clip",
219
+ "type": "CLIP",
220
+ "link": 75
221
+ }
222
+ ],
223
+ "outputs": [
224
+ {
225
+ "name": "CONDITIONING",
226
+ "type": "CONDITIONING",
227
+ "slot_index": 0,
228
+ "links": [
229
+ 52
230
+ ]
231
+ }
232
+ ],
233
+ "title": "CLIP Text Encode (Negative Prompt)",
234
+ "properties": {
235
+ "Node name for S&R": "CLIPTextEncode",
236
+ "cnr_id": "comfy-core",
237
+ "ver": "0.3.48",
238
+ "enableTabs": false,
239
+ "tabWidth": 65,
240
+ "tabXOffset": 10,
241
+ "hasSecondTab": false,
242
+ "secondTabText": "Send Back",
243
+ "secondTabOffset": 80,
244
+ "secondTabWidth": 65,
245
+ "widget_ue_connectable": {}
246
+ },
247
+ "widgets_values": [
248
+ ""
249
+ ],
250
+ "color": "#322",
251
+ "bgcolor": "#533"
252
+ },
158
253
  {
159
254
  "id": 60,
160
255
  "type": "SaveImage",
@@ -167,7 +262,7 @@
167
262
  600
168
263
  ],
169
264
  "flags": {},
170
- "order": 13,
265
+ "order": 15,
171
266
  "mode": 0,
172
267
  "inputs": [
173
268
  {
@@ -178,9 +273,9 @@
178
273
  ],
179
274
  "outputs": [],
180
275
  "properties": {
276
+ "Node name for S&R": "SaveImage",
181
277
  "cnr_id": "comfy-core",
182
278
  "ver": "0.3.48",
183
- "Node name for S&R": "SaveImage",
184
279
  "enableTabs": false,
185
280
  "tabWidth": 65,
186
281
  "tabXOffset": 10,
@@ -195,33 +290,63 @@
195
290
  ]
196
291
  },
197
292
  {
198
- "id": 58,
199
- "type": "EmptySD3LatentImage",
293
+ "id": 70,
294
+ "type": "Note",
200
295
  "pos": [
201
- 50,
202
- 510
296
+ 850,
297
+ 430
203
298
  ],
204
299
  "size": [
205
- 270,
206
- 106
300
+ 300,
301
+ 120
207
302
  ],
208
303
  "flags": {},
209
304
  "order": 3,
210
305
  "mode": 0,
211
306
  "inputs": [],
307
+ "outputs": [],
308
+ "title": "For fp8 without 8steps LoRA",
309
+ "properties": {},
310
+ "widgets_values": [
311
+ "Set cfg to 1.0 for a speed boost at the cost of consistency. Samplers like res_multistep work pretty well at cfg 1.0\n\nThe official number of steps is 50 but I think that's too much. Even just 10 steps seems to work."
312
+ ],
313
+ "color": "#432",
314
+ "bgcolor": "#653"
315
+ },
316
+ {
317
+ "id": 66,
318
+ "type": "ModelSamplingAuraFlow",
319
+ "pos": [
320
+ 850,
321
+ 10
322
+ ],
323
+ "size": [
324
+ 300,
325
+ 58
326
+ ],
327
+ "flags": {},
328
+ "order": 12,
329
+ "mode": 0,
330
+ "inputs": [
331
+ {
332
+ "name": "model",
333
+ "type": "MODEL",
334
+ "link": 130
335
+ }
336
+ ],
212
337
  "outputs": [
213
338
  {
214
- "name": "LATENT",
215
- "type": "LATENT",
339
+ "name": "MODEL",
340
+ "type": "MODEL",
216
341
  "links": [
217
- 107
342
+ 125
218
343
  ]
219
344
  }
220
345
  ],
221
346
  "properties": {
347
+ "Node name for S&R": "ModelSamplingAuraFlow",
222
348
  "cnr_id": "comfy-core",
223
349
  "ver": "0.3.48",
224
- "Node name for S&R": "EmptySD3LatentImage",
225
350
  "enableTabs": false,
226
351
  "tabWidth": 65,
227
352
  "tabXOffset": 10,
@@ -232,150 +357,147 @@
232
357
  "widget_ue_connectable": {}
233
358
  },
234
359
  "widgets_values": [
235
- 1328,
236
- 1328,
237
- 1
360
+ 3.1000000000000005
238
361
  ]
239
362
  },
240
363
  {
241
- "id": 6,
242
- "type": "CLIPTextEncode",
364
+ "id": 73,
365
+ "type": "LoraLoaderModelOnly",
243
366
  "pos": [
244
- 390,
367
+ 460,
245
368
  60
246
369
  ],
247
370
  "size": [
248
- 422.84503173828125,
249
- 164.31304931640625
371
+ 270,
372
+ 82
250
373
  ],
251
374
  "flags": {},
252
- "order": 8,
253
- "mode": 0,
375
+ "order": 11,
376
+ "mode": 4,
254
377
  "inputs": [
255
378
  {
256
- "name": "clip",
257
- "type": "CLIP",
258
- "link": 74
379
+ "name": "model",
380
+ "type": "MODEL",
381
+ "link": 129
259
382
  }
260
383
  ],
261
384
  "outputs": [
262
385
  {
263
- "name": "CONDITIONING",
264
- "type": "CONDITIONING",
265
- "slot_index": 0,
386
+ "name": "MODEL",
387
+ "type": "MODEL",
266
388
  "links": [
267
- 46
389
+ 130
268
390
  ]
269
391
  }
270
392
  ],
271
- "title": "CLIP Text Encode (Positive Prompt)",
272
393
  "properties": {
394
+ "Node name for S&R": "LoraLoaderModelOnly",
273
395
  "cnr_id": "comfy-core",
274
- "ver": "0.3.48",
275
- "Node name for S&R": "CLIPTextEncode",
276
- "enableTabs": false,
277
- "tabWidth": 65,
278
- "tabXOffset": 10,
279
- "hasSecondTab": false,
280
- "secondTabText": "Send Back",
281
- "secondTabOffset": 80,
282
- "secondTabWidth": 65,
283
- "widget_ue_connectable": {}
396
+ "ver": "0.3.49",
397
+ "models": [
398
+ {
399
+ "name": "Qwen-Image-Lightning-8steps-V1.0.safetensors",
400
+ "url": "https://huggingface.co/lightx2v/Qwen-Image-Lightning/resolve/main/Qwen-Image-Lightning-8steps-V1.0.safetensors",
401
+ "directory": "loras"
402
+ }
403
+ ]
284
404
  },
285
405
  "widgets_values": [
286
- "\"A vibrant, warm neon-lit street scene in Hong Kong at the afternoon, with a mix of colorful Chinese and English signs glowing brightly. The atmosphere is lively, cinematic, and rain-washed with reflections on the pavement. The colors are vivid, full of pink, blue, red, and green hues. Crowded buildings with overlapping neon signs. 1980s Hong Kong style. Signs include:\n\"龍鳳冰室\" \"金華燒臘\" \"HAPPY HAIR\" \"鴻運茶餐廳\" \"EASY BAR\" \"永發魚蛋粉\" \"添記粥麵\" \"SUNSHINE MOTEL\" \"美都餐室\" \"富記糖水\" \"太平館\" \"雅芳髮型屋\" \"STAR KTV\" \"銀河娛樂城\" \"百樂門舞廳\" \"BUBBLE CAFE\" \"萬豪麻雀館\" \"CITY LIGHTS BAR\" \"瑞祥香燭莊\" \"文記文具\" \"GOLDEN JADE HOTEL\" \"LOVELY BEAUTY\" \"合興百貨\" \"興旺電器\" And the background is warm yellow street and with all stores' lights on."
287
- ],
288
- "color": "#232",
289
- "bgcolor": "#353"
406
+ "Qwen-Image-Lightning-8steps-V1.0.safetensors",
407
+ 1
408
+ ]
290
409
  },
291
410
  {
292
- "id": 67,
411
+ "id": 69,
293
412
  "type": "MarkdownNote",
294
413
  "pos": [
295
414
  -540,
296
- 10
415
+ -220
297
416
  ],
298
417
  "size": [
299
- 520,
300
- 410
418
+ 390,
419
+ 180
301
420
  ],
302
421
  "flags": {},
303
422
  "order": 4,
304
423
  "mode": 0,
305
424
  "inputs": [],
306
425
  "outputs": [],
307
- "title": "Model links",
426
+ "title": "VRAM Usage",
308
427
  "properties": {
309
428
  "widget_ue_connectable": {}
310
429
  },
311
430
  "widgets_values": [
312
- "[Tutorial](https://docs.comfy.org/tutorials/image/qwen/qwen-image) | [教程](https://docs.comfy.org/zh-CN/tutorials/image/qwen/qwen-image)\n\n\n## Model links\n\nYou can find all the models on [Huggingface](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/tree/main) or [Modelscope](https://modelscope.cn/models/Comfy-Org/Qwen-Image_ComfyUI/files)\n\n**diffusion model**\n\n- [qwen_image_fp8_e4m3fn.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/diffusion_models/qwen_image_fp8_e4m3fn.safetensors)\n\n**text encoder**\n\n- [qwen_2.5_vl_7b_fp8_scaled.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/text_encoders/qwen_2.5_vl_7b_fp8_scaled.safetensors)\n\n**vae**\n\n- [qwen_image_vae.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/vae/qwen_image_vae.safetensors)\n\nModel Storage Location\n\n```\n📂 ComfyUI/\n├── 📂 models/\n│ ├── 📂 diffusion_models/\n│ │ └── qwen_image_fp8_e4m3fn.safetensors\n│ ├── 📂 vae/\n│ │ └── qwen_image_vae.safetensors\n│ └── 📂 text_encoders/\n│ └── qwen_2.5_vl_7b_fp8_scaled.safetensors\n```\n"
431
+ "## GPU:RTX4090D 24GB\n\n| Configuration | VRAM Usage | 1st Generation | 2nd Generation |\n|---------------------|---------------|---------------|-----------------|\n| Fp8_e4m3fn | 86% | 94s | ≈ 71s |\n| With 8steps LoRA | 86% | 55s | 34s |\n| Distill fp8_e4m3fn | 86% | 69s | ≈ 36s |"
313
432
  ],
314
433
  "color": "#432",
315
434
  "bgcolor": "#653"
316
435
  },
317
436
  {
318
- "id": 69,
319
- "type": "MarkdownNote",
437
+ "id": 71,
438
+ "type": "Note",
320
439
  "pos": [
321
- -320,
322
- 470
440
+ 850,
441
+ -120
323
442
  ],
324
443
  "size": [
325
- 290,
326
- 90
444
+ 300,
445
+ 88
327
446
  ],
328
447
  "flags": {},
329
448
  "order": 5,
330
449
  "mode": 0,
331
450
  "inputs": [],
332
451
  "outputs": [],
333
- "title": "VRAM Usage",
334
- "properties": {
335
- "widget_ue_connectable": {}
336
- },
452
+ "properties": {},
337
453
  "widgets_values": [
338
- "- GPU:RTX4090 24GB\n- VRAM:86%\n- 1st generation 94s\n- 2nd generation 71s"
454
+ "Increase the shift if you get too many blury/dark/bad images. Decrease if you want to try increasing detail."
339
455
  ],
340
456
  "color": "#432",
341
457
  "bgcolor": "#653"
342
458
  },
343
459
  {
344
- "id": 7,
345
- "type": "CLIPTextEncode",
460
+ "id": 8,
461
+ "type": "VAEDecode",
346
462
  "pos": [
347
- 390,
348
- 260
463
+ 1170,
464
+ -90
349
465
  ],
350
466
  "size": [
351
- 425.27801513671875,
352
- 180.6060791015625
467
+ 210,
468
+ 46
353
469
  ],
354
- "flags": {},
355
- "order": 9,
470
+ "flags": {
471
+ "collapsed": false
472
+ },
473
+ "order": 14,
356
474
  "mode": 0,
357
475
  "inputs": [
358
476
  {
359
- "name": "clip",
360
- "type": "CLIP",
361
- "link": 75
477
+ "name": "samples",
478
+ "type": "LATENT",
479
+ "link": 128
480
+ },
481
+ {
482
+ "name": "vae",
483
+ "type": "VAE",
484
+ "link": 76
362
485
  }
363
486
  ],
364
487
  "outputs": [
365
488
  {
366
- "name": "CONDITIONING",
367
- "type": "CONDITIONING",
489
+ "name": "IMAGE",
490
+ "type": "IMAGE",
368
491
  "slot_index": 0,
369
492
  "links": [
370
- 52
493
+ 110
371
494
  ]
372
495
  }
373
496
  ],
374
- "title": "CLIP Text Encode (Negative Prompt)",
375
497
  "properties": {
498
+ "Node name for S&R": "VAEDecode",
376
499
  "cnr_id": "comfy-core",
377
500
  "ver": "0.3.48",
378
- "Node name for S&R": "CLIPTextEncode",
379
501
  "enableTabs": false,
380
502
  "tabWidth": 65,
381
503
  "tabXOffset": 10,
@@ -385,11 +507,7 @@
385
507
  "secondTabWidth": 65,
386
508
  "widget_ue_connectable": {}
387
509
  },
388
- "widgets_values": [
389
- ""
390
- ],
391
- "color": "#322",
392
- "bgcolor": "#533"
510
+ "widgets_values": []
393
511
  },
394
512
  {
395
513
  "id": 3,
@@ -403,7 +521,7 @@
403
521
  262
404
522
  ],
405
523
  "flags": {},
406
- "order": 11,
524
+ "order": 13,
407
525
  "mode": 0,
408
526
  "inputs": [
409
527
  {
@@ -438,9 +556,9 @@
438
556
  }
439
557
  ],
440
558
  "properties": {
559
+ "Node name for S&R": "KSampler",
441
560
  "cnr_id": "comfy-core",
442
561
  "ver": "0.3.48",
443
- "Node name for S&R": "KSampler",
444
562
  "enableTabs": false,
445
563
  "tabWidth": 65,
446
564
  "tabXOffset": 10,
@@ -451,7 +569,7 @@
451
569
  "widget_ue_connectable": {}
452
570
  },
453
571
  "widgets_values": [
454
- 360061445559465,
572
+ 1125488487853216,
455
573
  "randomize",
456
574
  20,
457
575
  2.5,
@@ -461,62 +579,41 @@
461
579
  ]
462
580
  },
463
581
  {
464
- "id": 71,
465
- "type": "Note",
582
+ "id": 37,
583
+ "type": "UNETLoader",
466
584
  "pos": [
467
- 850,
468
- -130
585
+ 20,
586
+ 50
469
587
  ],
470
588
  "size": [
471
- 290,
589
+ 330,
472
590
  90
473
591
  ],
474
592
  "flags": {},
475
593
  "order": 6,
476
594
  "mode": 0,
477
595
  "inputs": [],
478
- "outputs": [],
479
- "properties": {},
480
- "widgets_values": [
481
- "Increase the shift if you get too many blury/dark/bad images. Decrease if you want to try increasing detail."
482
- ],
483
- "color": "#432",
484
- "bgcolor": "#653"
485
- },
486
- {
487
- "id": 66,
488
- "type": "ModelSamplingAuraFlow",
489
- "pos": [
490
- 850,
491
- 10
492
- ],
493
- "size": [
494
- 300,
495
- 58
496
- ],
497
- "flags": {},
498
- "order": 10,
499
- "mode": 0,
500
- "inputs": [
501
- {
502
- "name": "model",
503
- "type": "MODEL",
504
- "link": 126
505
- }
506
- ],
507
596
  "outputs": [
508
597
  {
509
598
  "name": "MODEL",
510
599
  "type": "MODEL",
600
+ "slot_index": 0,
511
601
  "links": [
512
- 125
602
+ 129
513
603
  ]
514
604
  }
515
605
  ],
516
606
  "properties": {
607
+ "Node name for S&R": "UNETLoader",
517
608
  "cnr_id": "comfy-core",
518
609
  "ver": "0.3.48",
519
- "Node name for S&R": "ModelSamplingAuraFlow",
610
+ "models": [
611
+ {
612
+ "name": "qwen_image_fp8_e4m3fn.safetensors",
613
+ "url": "https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/diffusion_models/qwen_image_fp8_e4m3fn.safetensors",
614
+ "directory": "diffusion_models"
615
+ }
616
+ ],
520
617
  "enableTabs": false,
521
618
  "tabWidth": 65,
522
619
  "tabXOffset": 10,
@@ -527,81 +624,56 @@
527
624
  "widget_ue_connectable": {}
528
625
  },
529
626
  "widgets_values": [
530
- 3.1000000000000005
627
+ "qwen_image_fp8_e4m3fn.safetensors",
628
+ "default"
531
629
  ]
532
630
  },
533
631
  {
534
- "id": 8,
535
- "type": "VAEDecode",
632
+ "id": 67,
633
+ "type": "MarkdownNote",
536
634
  "pos": [
537
- 1170,
538
- -90
635
+ -540,
636
+ 10
539
637
  ],
540
638
  "size": [
541
- 210,
542
- 46
639
+ 540,
640
+ 630
543
641
  ],
544
- "flags": {
545
- "collapsed": false
546
- },
547
- "order": 12,
642
+ "flags": {},
643
+ "order": 7,
548
644
  "mode": 0,
549
- "inputs": [
550
- {
551
- "name": "samples",
552
- "type": "LATENT",
553
- "link": 128
554
- },
555
- {
556
- "name": "vae",
557
- "type": "VAE",
558
- "link": 76
559
- }
560
- ],
561
- "outputs": [
562
- {
563
- "name": "IMAGE",
564
- "type": "IMAGE",
565
- "slot_index": 0,
566
- "links": [
567
- 110
568
- ]
569
- }
570
- ],
645
+ "inputs": [],
646
+ "outputs": [],
647
+ "title": "Model links",
571
648
  "properties": {
572
- "cnr_id": "comfy-core",
573
- "ver": "0.3.48",
574
- "Node name for S&R": "VAEDecode",
575
- "enableTabs": false,
576
- "tabWidth": 65,
577
- "tabXOffset": 10,
578
- "hasSecondTab": false,
579
- "secondTabText": "Send Back",
580
- "secondTabOffset": 80,
581
- "secondTabWidth": 65,
582
649
  "widget_ue_connectable": {}
583
650
  },
584
- "widgets_values": []
651
+ "widgets_values": [
652
+ "[Tutorial](https://docs.comfy.org/tutorials/image/qwen/qwen-image) | [教程](https://docs.comfy.org/zh-CN/tutorials/image/qwen/qwen-image)\n\n\n## Model links\n\nYou can find all the models on [Huggingface](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/tree/main) or [Modelscope](https://modelscope.cn/models/Comfy-Org/Qwen-Image_ComfyUI/files)\n\n**Diffusion model**\n\n- [qwen_image_fp8_e4m3fn.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/diffusion_models/qwen_image_fp8_e4m3fn.safetensors)\n\nQwen_image_distill\n\n- [qwen_image_distill_full_fp8_e4m3fn.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/non_official/diffusion_models/qwen_image_distill_full_fp8_e4m3fn.safetensors)\n- [qwen_image_distill_full_bf16.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/non_official/diffusion_models/qwen_image_distill_full_bf16.safetensors)\n\n**LoRA**\n\n- [Qwen-Image-Lightning-8steps-V1.0.safetensors](https://huggingface.co/lightx2v/Qwen-Image-Lightning/resolve/main/Qwen-Image-Lightning-8steps-V1.0.safetensors)\n\n**Text encoder**\n\n- [qwen_2.5_vl_7b_fp8_scaled.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/text_encoders/qwen_2.5_vl_7b_fp8_scaled.safetensors)\n\n**VAE**\n\n- [qwen_image_vae.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/vae/qwen_image_vae.safetensors)\n\nModel Storage Location\n\n```\n📂 ComfyUI/\n├── 📂 models/\n│ ├── 📂 diffusion_models/\n│ │ ├── qwen_image_fp8_e4m3fn.safetensors\n│ │ └── qwen_image_distill_full_fp8_e4m3fn.safetensors\n│ ├── 📂 loras/\n│ │ └── Qwen-Image-Lightning-8steps-V1.0.safetensors\n│ ├── 📂 vae/\n│ │ └── qwen_image_vae.safetensors\n│ └── 📂 text_encoders/\n│ └── qwen_2.5_vl_7b_fp8_scaled.safetensors\n```\n"
653
+ ],
654
+ "color": "#432",
655
+ "bgcolor": "#653"
585
656
  },
586
657
  {
587
- "id": 70,
588
- "type": "Note",
658
+ "id": 74,
659
+ "type": "MarkdownNote",
589
660
  "pos": [
590
661
  850,
591
- 430
662
+ 600
592
663
  ],
593
664
  "size": [
594
- 300,
595
- 120
665
+ 310,
666
+ 190
596
667
  ],
597
668
  "flags": {},
598
- "order": 7,
669
+ "order": 8,
599
670
  "mode": 0,
600
671
  "inputs": [],
601
672
  "outputs": [],
673
+ "title": "KSampler settings",
602
674
  "properties": {},
603
675
  "widgets_values": [
604
- "Set cfg to 1.0 for a speed boost at the cost of consistency. Samplers like res_multistep work pretty well at cfg 1.0\n\nThe official number of steps is 50 but I think that's too much. Even just 10 steps seems to work."
676
+ "You can test and find the best setting by yourself. The following table is for reference.\n\n| model | steps | cfg |\n|---------------------|---------------|---------------|\n| fp8_e4m3fn(Qwen team's suggestion) | 40 | 2.5 \n| fp8_e4m3fn | 20 | 2.5 |\n| fp8_e4m3fn + 8steps LoRA | 8 | 2.5 |\n| distill fp8_e4m3fn | 10 | 1.0 |"
605
677
  ],
606
678
  "color": "#432",
607
679
  "bgcolor": "#653"
@@ -673,20 +745,28 @@
673
745
  "MODEL"
674
746
  ],
675
747
  [
676
- 126,
748
+ 128,
749
+ 3,
750
+ 0,
751
+ 8,
752
+ 0,
753
+ "LATENT"
754
+ ],
755
+ [
756
+ 129,
677
757
  37,
678
758
  0,
679
- 66,
759
+ 73,
680
760
  0,
681
761
  "MODEL"
682
762
  ],
683
763
  [
684
- 128,
685
- 3,
764
+ 130,
765
+ 73,
686
766
  0,
687
- 8,
767
+ 66,
688
768
  0,
689
- "LATENT"
769
+ "MODEL"
690
770
  ]
691
771
  ],
692
772
  "groups": [
@@ -721,25 +801,38 @@
721
801
  "title": "Step3 - Prompt",
722
802
  "bounding": [
723
803
  380,
724
- -20,
804
+ 160,
725
805
  450,
726
806
  470
727
807
  ],
728
808
  "color": "#3f789e",
729
809
  "font_size": 24,
730
810
  "flags": {}
811
+ },
812
+ {
813
+ "id": 4,
814
+ "title": "Lightx2v 8steps LoRA",
815
+ "bounding": [
816
+ 380,
817
+ -20,
818
+ 450,
819
+ 170
820
+ ],
821
+ "color": "#3f789e",
822
+ "font_size": 24,
823
+ "flags": {}
731
824
  }
732
825
  ],
733
826
  "config": {},
734
827
  "extra": {
735
828
  "ds": {
736
- "scale": 0.7627768444385553,
829
+ "scale": 0.5209868481924539,
737
830
  "offset": [
738
- 511.3351769654992,
739
- 265.80199174981243
831
+ 1164.5682490430127,
832
+ 480.8451480341405
740
833
  ]
741
834
  },
742
- "frontendVersion": "1.23.4",
835
+ "frontendVersion": "1.26.0",
743
836
  "ue_links": [],
744
837
  "links_added_by_ue": [],
745
838
  "VHS_latentpreview": false,