Qwen-Imageとは?
Qwen-Image は、中国 Alibaba 系の Qwen チームが開発した画像生成モデルです。
テキストエンコーダにマルチモーダル LLM の Qwen2.5-VL を使っており、T5 や Gemma を使っているモデルに比べると、プロンプトの理解力は頭ひとつ抜けています。
さらに、Flux.1 dev と違ってベースが蒸留モデルではないため学習しやすく、姉妹モデルである Qwen-Image-Edit と合わせて、LoRA や Lightning 系の周辺エコシステムが充実しているのも特徴です。
推奨解像度
Qwen-Image は 1.5M〜1.8M ピクセル前後が推奨です。
- 1:1 … 1328 × 1328
- 4:3 … 1472 × 1104
- 3:2 … 1584 × 1056
- 16:9 … 1664 × 928
モデルのダウンロード
-
diffusion_models
-
text_encoders
-
vae
📂ComfyUI/
└── 📂models/
├── 📂diffusion_models/
│ └── qwen_image_fp8_e4m3fn.safetensors
├── 📂text_encoders/
│ ├── qwen_2.5_vl_7b_fp8_scaled.safetensors
│ └── Qwen2.5-VL-7B.gguf ← gguf を使う場合のみ
├── 📂unet/
│ └── qwen-image.gguf ← gguf を使う場合のみ
└── 📂vae/
└── qwen_image_vae.safetensors
text2image

{
"id": "d8034549-7e0a-40f1-8c2e-de3ffc6f1cae",
"revision": 0,
"last_node_id": 56,
"last_link_id": 101,
"nodes": [
{
"id": 54,
"type": "ModelSamplingAuraFlow",
"pos": [
603.9390258789062,
53.0477294921875
],
"size": [
230.33058166503906,
58
],
"flags": {},
"order": 5,
"mode": 0,
"inputs": [
{
"name": "model",
"type": "MODEL",
"link": 99
}
],
"outputs": [
{
"name": "MODEL",
"type": "MODEL",
"links": [
100
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.49",
"Node name for S&R": "ModelSamplingAuraFlow"
},
"widgets_values": [
3.1000000000000005
]
},
{
"id": 6,
"type": "CLIPTextEncode",
"pos": [
415,
186
],
"size": [
419.26959228515625,
156.00363159179688
],
"flags": {},
"order": 6,
"mode": 0,
"inputs": [
{
"name": "clip",
"type": "CLIP",
"link": 74
}
],
"outputs": [
{
"name": "CONDITIONING",
"type": "CONDITIONING",
"slot_index": 0,
"links": [
46
]
}
],
"title": "CLIP Text Encode (Positive Prompt)",
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.33",
"Node name for S&R": "CLIPTextEncode"
},
"widgets_values": [
"A cinematic scene inspired by a Japanese urban night. A woman stands in front of a retro vending machine on a quiet street corner, contemplating her choice of drink. She wears a loose white top and a long, black patterned skirt, hands clasped behind her back, slightly leaning forward. Blue shutters, city posters, and soft neon lights in the background evoke authentic Tokyo atmosphere. The lighting is soft and diffused, blending the subject gently with the nighttime environment. The shot is taken from an angled, bird’s-eye perspective, emphasizing depth and a story-like moment, as if captured from a movie. The mood is emotional, realistic, and cinematic, with film-like texture—no illustration, only photorealism."
]
},
{
"id": 7,
"type": "CLIPTextEncode",
"pos": [
415,
405.392333984375
],
"size": [
418.3189392089844,
107.08506774902344
],
"flags": {},
"order": 7,
"mode": 0,
"inputs": [
{
"name": "clip",
"type": "CLIP",
"link": 75
}
],
"outputs": [
{
"name": "CONDITIONING",
"type": "CONDITIONING",
"slot_index": 0,
"links": [
52
]
}
],
"title": "CLIP Text Encode (Negative Prompt)",
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.33",
"Node name for S&R": "CLIPTextEncode"
},
"widgets_values": [
"text, worst quality, blurry, ugly"
]
},
{
"id": 53,
"type": "EmptySD3LatentImage",
"pos": [
596.3189697265625,
586.8646240234375
],
"size": [
237,
106
],
"flags": {},
"order": 0,
"mode": 0,
"inputs": [],
"outputs": [
{
"name": "LATENT",
"type": "LATENT",
"links": [
98
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.49",
"Node name for S&R": "EmptySD3LatentImage"
},
"widgets_values": [
1104,
1472,
1
]
},
{
"id": 37,
"type": "UNETLoader",
"pos": [
267.6552734375,
53.0477294921875
],
"size": [
305.3782043457031,
82
],
"flags": {},
"order": 1,
"mode": 0,
"inputs": [],
"outputs": [
{
"name": "MODEL",
"type": "MODEL",
"slot_index": 0,
"links": [
99
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.33",
"Node name for S&R": "UNETLoader"
},
"widgets_values": [
"Qwen-Image\\qwen_image_fp8_e4m3fn.safetensors",
"fp8_e4m3fn"
],
"color": "#323",
"bgcolor": "#535"
},
{
"id": 39,
"type": "VAELoader",
"pos": [
957.4939789387397,
67.67436911843055
],
"size": [
256.26084283860405,
58
],
"flags": {},
"order": 2,
"mode": 0,
"inputs": [],
"outputs": [
{
"name": "VAE",
"type": "VAE",
"slot_index": 0,
"links": [
76
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.33",
"Node name for S&R": "VAELoader"
},
"widgets_values": [
"qwen_image_vae.safetensors"
],
"color": "#322",
"bgcolor": "#533"
},
{
"id": 55,
"type": "MarkdownNote",
"pos": [
-127.09132385253906,
-13.402286529541016
],
"size": [
345.70001220703125,
225.77000427246094
],
"flags": {},
"order": 3,
"mode": 0,
"inputs": [],
"outputs": [],
"properties": {},
"widgets_values": [
"## models\n- [qwen_image_fp8_e4m3fn.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/blob/main/split_files/diffusion_models/qwen_image_fp8_e4m3fn.safetensors)\n- [qwen_2.5_vl_7b_fp8_scaled.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/blob/main/split_files/text_encoders/qwen_2.5_vl_7b_fp8_scaled.safetensors)\n- [qwen_image_vae.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/tree/main/split_files/vae)\n\n\n```\n📂ComfyUI/\n└── 📂models/\n ├── 📂diffusion_models/\n │ └── qwen_image_fp8_e4m3fn.safetensors\n ├── 📂text_encoders/\n │ └── qwen_2.5_vl_7b_fp8_scaled.safetensors\n └── 📂vae/\n └── qwen_image_vae.safetensors\n```"
],
"color": "#323",
"bgcolor": "#535"
},
{
"id": 8,
"type": "VAEDecode",
"pos": [
1253.505002517707,
188.1918182373047
],
"size": [
166.0271370269786,
46
],
"flags": {},
"order": 9,
"mode": 0,
"inputs": [
{
"name": "samples",
"type": "LATENT",
"link": 35
},
{
"name": "vae",
"type": "VAE",
"link": 76
}
],
"outputs": [
{
"name": "IMAGE",
"type": "IMAGE",
"slot_index": 0,
"links": [
101
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.33",
"Node name for S&R": "VAEDecode"
},
"widgets_values": []
},
{
"id": 3,
"type": "KSampler",
"pos": [
898.7548217773438,
188.1918182373047
],
"size": [
315,
262
],
"flags": {},
"order": 8,
"mode": 0,
"inputs": [
{
"name": "model",
"type": "MODEL",
"link": 100
},
{
"name": "positive",
"type": "CONDITIONING",
"link": 46
},
{
"name": "negative",
"type": "CONDITIONING",
"link": 52
},
{
"name": "latent_image",
"type": "LATENT",
"link": 98
}
],
"outputs": [
{
"name": "LATENT",
"type": "LATENT",
"slot_index": 0,
"links": [
35
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.33",
"Node name for S&R": "KSampler"
},
"widgets_values": [
12345,
"fixed",
20,
2.5,
"euler",
"simple",
1
]
},
{
"id": 56,
"type": "SaveImage",
"pos": [
1459.2823202850486,
188.1918182373047
],
"size": [
304.95584260400346,
486.83910965100097
],
"flags": {},
"order": 10,
"mode": 0,
"inputs": [
{
"name": "images",
"type": "IMAGE",
"link": 101
}
],
"outputs": [],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.76"
},
"widgets_values": [
"ComfyUI"
]
},
{
"id": 38,
"type": "CLIPLoader",
"pos": [
56.288665771484375,
312.74468994140625
],
"size": [
301.3524169921875,
106
],
"flags": {},
"order": 4,
"mode": 0,
"inputs": [],
"outputs": [
{
"name": "CLIP",
"type": "CLIP",
"slot_index": 0,
"links": [
74,
75
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.33",
"Node name for S&R": "CLIPLoader"
},
"widgets_values": [
"qwen_2.5_vl_7b_fp8_scaled.safetensors",
"qwen_image",
"default"
],
"color": "#432",
"bgcolor": "#653"
}
],
"links": [
[
35,
3,
0,
8,
0,
"LATENT"
],
[
46,
6,
0,
3,
1,
"CONDITIONING"
],
[
52,
7,
0,
3,
2,
"CONDITIONING"
],
[
74,
38,
0,
6,
0,
"CLIP"
],
[
75,
38,
0,
7,
0,
"CLIP"
],
[
76,
39,
0,
8,
1,
"VAE"
],
[
98,
53,
0,
3,
3,
"LATENT"
],
[
99,
37,
0,
54,
0,
"MODEL"
],
[
100,
54,
0,
3,
0,
"MODEL"
],
[
101,
8,
0,
56,
0,
"IMAGE"
]
],
"groups": [],
"config": {},
"extra": {
"ds": {
"scale": 0.7513148009015777,
"offset": [
227.09132385253906,
113.40228652954102
]
},
"frontendVersion": "1.35.0",
"VHS_latentpreview": false,
"VHS_latentpreviewrate": 0,
"VHS_MetadataImage": true,
"VHS_KeepIntermediate": true
},
"version": 0.4
}
- サンプラーは
eulerまたはres_multistepが使われることが多いです。
ControlNet(InstantX)
Qwen-Image 向けの ControlNet モデルはいくつかありますが、使い勝手が良いので、ControlNet-Union として提供されている InstantX ベースのものを紹介します。
モデルのダウンロード
-
controlnet
📂ComfyUI/
└── 📂models/
└── 📂controlnet/
├── Qwen-Image-InstantX-ControlNet-Union.safetensors
└── Qwen-Image-InstantX-ControlNet-Inpainting.safetensors
workflow

{
"id": "d8034549-7e0a-40f1-8c2e-de3ffc6f1cae",
"revision": 0,
"last_node_id": 66,
"last_link_id": 118,
"nodes": [
{
"id": 8,
"type": "VAEDecode",
"pos": [
1541.8401809731122,
195.64542636108396
],
"size": [
166.0271370269786,
46
],
"flags": {},
"order": 16,
"mode": 0,
"inputs": [
{
"name": "samples",
"type": "LATENT",
"link": 35
},
{
"name": "vae",
"type": "VAE",
"link": 76
}
],
"outputs": [
{
"name": "IMAGE",
"type": "IMAGE",
"slot_index": 0,
"links": [
101
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.33",
"Node name for S&R": "VAEDecode"
},
"widgets_values": []
},
{
"id": 56,
"type": "SaveImage",
"pos": [
1740.3895365326061,
195.64542636108396
],
"size": [
304.95584260400346,
486.83910965100097
],
"flags": {},
"order": 17,
"mode": 0,
"inputs": [
{
"name": "images",
"type": "IMAGE",
"link": 101
}
],
"outputs": [],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.76"
},
"widgets_values": [
"ComfyUI"
]
},
{
"id": 58,
"type": "ControlNetApplyAdvanced",
"pos": [
891.7957439080817,
215.7779928359986
],
"size": [
270,
186
],
"flags": {},
"order": 13,
"mode": 0,
"inputs": [
{
"name": "positive",
"type": "CONDITIONING",
"link": 105
},
{
"name": "negative",
"type": "CONDITIONING",
"link": 106
},
{
"name": "control_net",
"type": "CONTROL_NET",
"link": 108
},
{
"name": "image",
"type": "IMAGE",
"link": 113
},
{
"name": "vae",
"shape": 7,
"type": "VAE",
"link": 114
}
],
"outputs": [
{
"name": "positive",
"type": "CONDITIONING",
"links": [
103
]
},
{
"name": "negative",
"type": "CONDITIONING",
"links": [
104
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.55",
"Node name for S&R": "ControlNetApplyAdvanced"
},
"widgets_values": [
0.8,
0,
0.6
],
"color": "#232",
"bgcolor": "#353"
},
{
"id": 7,
"type": "CLIPTextEncode",
"pos": [
415.9506530761719,
405.392333984375
],
"size": [
418.3189392089844,
107.08506774902344
],
"flags": {},
"order": 10,
"mode": 0,
"inputs": [
{
"name": "clip",
"type": "CLIP",
"link": 75
}
],
"outputs": [
{
"name": "CONDITIONING",
"type": "CONDITIONING",
"slot_index": 0,
"links": [
106
]
}
],
"title": "CLIP Text Encode (Negative Prompt)",
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.33",
"Node name for S&R": "CLIPTextEncode"
},
"widgets_values": [
"text, worst quality, blurry, ugly"
]
},
{
"id": 57,
"type": "ControlNetLoader",
"pos": [
324.23409025000626,
573.4262008565972
],
"size": [
270,
58
],
"flags": {},
"order": 0,
"mode": 0,
"inputs": [],
"outputs": [
{
"name": "CONTROL_NET",
"type": "CONTROL_NET",
"links": [
107
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.55",
"Node name for S&R": "ControlNetLoader"
},
"widgets_values": [
"Qwen-Image\\Qwen-Image-InstantX-ControlNet-Union.safetensors"
],
"color": "#232",
"bgcolor": "#353"
},
{
"id": 59,
"type": "SetUnionControlNetType",
"pos": [
624.2695922851562,
573.4262008565972
],
"size": [
210,
58
],
"flags": {},
"order": 6,
"mode": 0,
"inputs": [
{
"name": "control_net",
"type": "CONTROL_NET",
"link": 107
}
],
"outputs": [
{
"name": "CONTROL_NET",
"type": "CONTROL_NET",
"links": [
108
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.76",
"Node name for S&R": "SetUnionControlNetType"
},
"widgets_values": [
"auto"
],
"color": "#232",
"bgcolor": "#353"
},
{
"id": 61,
"type": "ImageScaleToTotalPixels",
"pos": [
354.1991325369917,
707.0446122693944
],
"size": [
210,
82
],
"flags": {},
"order": 8,
"mode": 0,
"inputs": [
{
"name": "image",
"type": "IMAGE",
"link": 109
}
],
"outputs": [
{
"name": "IMAGE",
"type": "IMAGE",
"links": [
112,
116
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.76",
"Node name for S&R": "ImageScaleToTotalPixels"
},
"widgets_values": [
"nearest-exact",
1.5
]
},
{
"id": 65,
"type": "VAEEncode",
"pos": [
694.2695922851562,
991.1564459430302
],
"size": [
140,
46
],
"flags": {},
"order": 12,
"mode": 0,
"inputs": [
{
"name": "pixels",
"type": "IMAGE",
"link": 116
},
{
"name": "vae",
"type": "VAE",
"link": 115
}
],
"outputs": [
{
"name": "LATENT",
"type": "LATENT",
"links": [
117
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.76",
"Node name for S&R": "VAEEncode"
},
"widgets_values": []
},
{
"id": 39,
"type": "VAELoader",
"pos": [
396.11189896833156,
1066.1678107405203
],
"size": [
256.26084283860405,
58
],
"flags": {},
"order": 1,
"mode": 0,
"inputs": [],
"outputs": [
{
"name": "VAE",
"type": "VAE",
"slot_index": 0,
"links": [
76,
114,
115
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.33",
"Node name for S&R": "VAELoader"
},
"widgets_values": [
"qwen_image_vae.safetensors"
],
"color": "#322",
"bgcolor": "#533"
},
{
"id": 64,
"type": "DWPreprocessor",
"pos": [
594.6777954101562,
707.0446122693944
],
"size": [
239.591796875,
222
],
"flags": {},
"order": 11,
"mode": 0,
"inputs": [
{
"name": "image",
"type": "IMAGE",
"link": 112
}
],
"outputs": [
{
"name": "IMAGE",
"type": "IMAGE",
"links": [
113,
118
]
},
{
"name": "POSE_KEYPOINT",
"type": "POSE_KEYPOINT",
"links": null
}
],
"properties": {
"cnr_id": "comfyui_controlnet_aux",
"ver": "12f35647f0d510e03b45a47fb420fe1245a575df",
"Node name for S&R": "DWPreprocessor"
},
"widgets_values": [
"enable",
"enable",
"enable",
512,
"yolox_l.onnx",
"dw-ll_ucoco_384_bs5.torchscript.pt",
"disable"
]
},
{
"id": 3,
"type": "KSampler",
"pos": [
1194.317962440597,
195.64542636108396
],
"size": [
315,
262
],
"flags": {},
"order": 15,
"mode": 0,
"inputs": [
{
"name": "model",
"type": "MODEL",
"link": 100
},
{
"name": "positive",
"type": "CONDITIONING",
"link": 103
},
{
"name": "negative",
"type": "CONDITIONING",
"link": 104
},
{
"name": "latent_image",
"type": "LATENT",
"link": 117
}
],
"outputs": [
{
"name": "LATENT",
"type": "LATENT",
"slot_index": 0,
"links": [
35
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.33",
"Node name for S&R": "KSampler"
},
"widgets_values": [
1234,
"fixed",
20,
2.5,
"euler",
"simple",
1
]
},
{
"id": 37,
"type": "UNETLoader",
"pos": [
595.8433122436518,
16.738045024108914
],
"size": [
305.3782043457031,
82
],
"flags": {},
"order": 2,
"mode": 0,
"inputs": [],
"outputs": [
{
"name": "MODEL",
"type": "MODEL",
"slot_index": 0,
"links": [
99
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.33",
"Node name for S&R": "UNETLoader"
},
"widgets_values": [
"Qwen-Image\\qwen_image_fp8_e4m3fn.safetensors",
"fp8_e4m3fn"
],
"color": "#323",
"bgcolor": "#535"
},
{
"id": 54,
"type": "ModelSamplingAuraFlow",
"pos": [
931.4651622430426,
16.738045024108914
],
"size": [
230.33058166503906,
58
],
"flags": {},
"order": 7,
"mode": 0,
"inputs": [
{
"name": "model",
"type": "MODEL",
"link": 99
}
],
"outputs": [
{
"name": "MODEL",
"type": "MODEL",
"links": [
100
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.49",
"Node name for S&R": "ModelSamplingAuraFlow"
},
"widgets_values": [
3.1000000000000005
]
},
{
"id": 60,
"type": "LoadImage",
"pos": [
26.2067480792631,
707.0446122693944
],
"size": [
297.5137215845641,
478.0984700375059
],
"flags": {},
"order": 3,
"mode": 0,
"inputs": [],
"outputs": [
{
"name": "IMAGE",
"type": "IMAGE",
"links": [
109
]
},
{
"name": "MASK",
"type": "MASK",
"links": null
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.76",
"Node name for S&R": "LoadImage"
},
"widgets_values": [
"pasted/image (99).png",
"image"
]
},
{
"id": 66,
"type": "PreviewImage",
"pos": [
874.022433592813,
707.0446122693944
],
"size": [
210,
280.3611468548585
],
"flags": {},
"order": 14,
"mode": 0,
"inputs": [
{
"name": "images",
"type": "IMAGE",
"link": 118
}
],
"outputs": [],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.76",
"Node name for S&R": "PreviewImage"
},
"widgets_values": []
},
{
"id": 6,
"type": "CLIPTextEncode",
"pos": [
415,
186
],
"size": [
419.26959228515625,
156.00363159179688
],
"flags": {},
"order": 9,
"mode": 0,
"inputs": [
{
"name": "clip",
"type": "CLIP",
"link": 74
}
],
"outputs": [
{
"name": "CONDITIONING",
"type": "CONDITIONING",
"slot_index": 0,
"links": [
105
]
}
],
"title": "CLIP Text Encode (Positive Prompt)",
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.33",
"Node name for S&R": "CLIPTextEncode"
},
"widgets_values": [
"A raw iPhone selfie of a young woman taken outdoors under a clear blue sky. She’s holding her hand close to her face, flashing a relaxed peace sign right next to her eye—slightly imperfect framing, natural skin tones, and soft sunlight catching her features. The background is a bright, cloudless blue sky, slightly blown out from the sun, giving it that spontaneous summer-day feel. "
]
},
{
"id": 55,
"type": "MarkdownNote",
"pos": [
-52.43482949290891,
-31.4743293038663
],
"size": [
404.3200233043325,
256.7791118275035
],
"flags": {},
"order": 4,
"mode": 0,
"inputs": [],
"outputs": [],
"properties": {},
"widgets_values": [
"## models\n- [qwen_image_fp8_e4m3fn.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/blob/main/split_files/diffusion_models/qwen_image_fp8_e4m3fn.safetensors)\n- [qwen_2.5_vl_7b_fp8_scaled.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/blob/main/split_files/text_encoders/qwen_2.5_vl_7b_fp8_scaled.safetensors)\n- [qwen_image_vae.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/tree/main/split_files/vae)\n- [Qwen-Image-InstantX-ControlNet-Union.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image-InstantX-ControlNets/blob/main/split_files/controlnet/Qwen-Image-InstantX-ControlNet-Union.safetensors)\n\n\n```\n📂ComfyUI/\n└── 📂models/\n ├── 📂controlnet/\n │ └── Qwen-Image-InstantX-ControlNet-Union.safetensors\n ├── 📂diffusion_models/\n │ └── qwen_image_fp8_e4m3fn.safetensors\n ├── 📂text_encoders/\n │ └── qwen_2.5_vl_7b_fp8_scaled.safetensors\n └── 📂vae/\n └── qwen_image_vae.safetensors\n```"
],
"color": "#323",
"bgcolor": "#535"
},
{
"id": 38,
"type": "CLIPLoader",
"pos": [
56.288665771484375,
312.74468994140625
],
"size": [
301.3524169921875,
106
],
"flags": {},
"order": 5,
"mode": 0,
"inputs": [],
"outputs": [
{
"name": "CLIP",
"type": "CLIP",
"slot_index": 0,
"links": [
74,
75
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.33",
"Node name for S&R": "CLIPLoader"
},
"widgets_values": [
"qwen_2.5_vl_7b_fp8_scaled.safetensors",
"qwen_image",
"default"
],
"color": "#432",
"bgcolor": "#653"
}
],
"links": [
[
35,
3,
0,
8,
0,
"LATENT"
],
[
74,
38,
0,
6,
0,
"CLIP"
],
[
75,
38,
0,
7,
0,
"CLIP"
],
[
76,
39,
0,
8,
1,
"VAE"
],
[
99,
37,
0,
54,
0,
"MODEL"
],
[
100,
54,
0,
3,
0,
"MODEL"
],
[
101,
8,
0,
56,
0,
"IMAGE"
],
[
103,
58,
0,
3,
1,
"CONDITIONING"
],
[
104,
58,
1,
3,
2,
"CONDITIONING"
],
[
105,
6,
0,
58,
0,
"CONDITIONING"
],
[
106,
7,
0,
58,
1,
"CONDITIONING"
],
[
107,
57,
0,
59,
0,
"CONTROL_NET"
],
[
108,
59,
0,
58,
2,
"CONTROL_NET"
],
[
109,
60,
0,
61,
0,
"IMAGE"
],
[
112,
61,
0,
64,
0,
"IMAGE"
],
[
113,
64,
0,
58,
3,
"IMAGE"
],
[
114,
39,
0,
58,
4,
"VAE"
],
[
115,
39,
0,
65,
1,
"VAE"
],
[
116,
61,
0,
65,
0,
"IMAGE"
],
[
117,
65,
0,
3,
3,
"LATENT"
],
[
118,
64,
0,
66,
0,
"IMAGE"
]
],
"groups": [],
"config": {},
"extra": {
"ds": {
"scale": 0.5644739300537773,
"offset": [
336.67717349290893,
329.8891613038664
]
},
"frontendVersion": "1.35.0",
"VHS_latentpreview": false,
"VHS_latentpreviewrate": 0,
"VHS_MetadataImage": true,
"VHS_KeepIntermediate": true
},
"version": 0.4
}

{
"id": "d8034549-7e0a-40f1-8c2e-de3ffc6f1cae",
"revision": 0,
"last_node_id": 68,
"last_link_id": 130,
"nodes": [
{
"id": 8,
"type": "VAEDecode",
"pos": [
1569.3401809731115,
195.64542636108396
],
"size": [
166.0271370269786,
46
],
"flags": {},
"order": 13,
"mode": 0,
"inputs": [
{
"name": "samples",
"type": "LATENT",
"link": 35
},
{
"name": "vae",
"type": "VAE",
"link": 76
}
],
"outputs": [
{
"name": "IMAGE",
"type": "IMAGE",
"slot_index": 0,
"links": [
101
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.33",
"Node name for S&R": "VAEDecode"
},
"widgets_values": []
},
{
"id": 7,
"type": "CLIPTextEncode",
"pos": [
415.9506530761719,
405.392333984375
],
"size": [
418.3189392089844,
107.08506774902344
],
"flags": {},
"order": 7,
"mode": 0,
"inputs": [
{
"name": "clip",
"type": "CLIP",
"link": 75
}
],
"outputs": [
{
"name": "CONDITIONING",
"type": "CONDITIONING",
"slot_index": 0,
"links": [
120
]
}
],
"title": "CLIP Text Encode (Negative Prompt)",
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.33",
"Node name for S&R": "CLIPTextEncode"
},
"widgets_values": [
"text, worst quality, blurry, ugly"
]
},
{
"id": 67,
"type": "ControlNetInpaintingAliMamaApply",
"pos": [
878.1510835591422,
215.63948549703468
],
"size": [
299.7853876074681,
206
],
"flags": {},
"order": 11,
"mode": 0,
"inputs": [
{
"name": "positive",
"type": "CONDITIONING",
"link": 119
},
{
"name": "negative",
"type": "CONDITIONING",
"link": 120
},
{
"name": "control_net",
"type": "CONTROL_NET",
"link": 122
},
{
"name": "vae",
"type": "VAE",
"link": 127
},
{
"name": "image",
"type": "IMAGE",
"link": 126
},
{
"name": "mask",
"type": "MASK",
"link": 125
}
],
"outputs": [
{
"name": "positive",
"type": "CONDITIONING",
"links": [
123
]
},
{
"name": "negative",
"type": "CONDITIONING",
"links": [
129
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.76",
"Node name for S&R": "ControlNetInpaintingAliMamaApply"
},
"widgets_values": [
1,
0,
1
],
"color": "#232",
"bgcolor": "#353"
},
{
"id": 61,
"type": "ImageScaleToTotalPixels",
"pos": [
624.2695922851562,
697.364612269394
],
"size": [
210,
82
],
"flags": {},
"order": 8,
"mode": 0,
"inputs": [
{
"name": "image",
"type": "IMAGE",
"link": 109
}
],
"outputs": [
{
"name": "IMAGE",
"type": "IMAGE",
"links": [
116,
126
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.76",
"Node name for S&R": "ImageScaleToTotalPixels"
},
"widgets_values": [
"nearest-exact",
1.5
]
},
{
"id": 38,
"type": "CLIPLoader",
"pos": [
56.288665771484375,
312.74468994140625
],
"size": [
301.3524169921875,
106
],
"flags": {},
"order": 0,
"mode": 0,
"inputs": [],
"outputs": [
{
"name": "CLIP",
"type": "CLIP",
"slot_index": 0,
"links": [
74,
75
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.33",
"Node name for S&R": "CLIPLoader"
},
"widgets_values": [
"qwen_2.5_vl_7b_fp8_scaled.safetensors",
"qwen_image",
"default"
],
"color": "#432",
"bgcolor": "#653"
},
{
"id": 39,
"type": "VAELoader",
"pos": [
578.0087494465522,
846.6494035278984
],
"size": [
256.26084283860405,
58
],
"flags": {},
"order": 1,
"mode": 0,
"inputs": [],
"outputs": [
{
"name": "VAE",
"type": "VAE",
"slot_index": 0,
"links": [
76,
115,
127
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.33",
"Node name for S&R": "VAELoader"
},
"widgets_values": [
"qwen_image_vae.safetensors"
],
"color": "#322",
"bgcolor": "#533"
},
{
"id": 6,
"type": "CLIPTextEncode",
"pos": [
415,
186
],
"size": [
419.26959228515625,
156.00363159179688
],
"flags": {},
"order": 6,
"mode": 0,
"inputs": [
{
"name": "clip",
"type": "CLIP",
"link": 74
}
],
"outputs": [
{
"name": "CONDITIONING",
"type": "CONDITIONING",
"slot_index": 0,
"links": [
119
]
}
],
"title": "CLIP Text Encode (Positive Prompt)",
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.33",
"Node name for S&R": "CLIPTextEncode"
},
"widgets_values": [
"Photo of a white fluffy dog curled up and sleeping on the dashboard of a car"
]
},
{
"id": 57,
"type": "ControlNetLoader",
"pos": [
564.2695922851562,
576.7262008565973
],
"size": [
270,
58
],
"flags": {},
"order": 2,
"mode": 0,
"inputs": [],
"outputs": [
{
"name": "CONTROL_NET",
"type": "CONTROL_NET",
"links": [
122
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.55",
"Node name for S&R": "ControlNetLoader"
},
"widgets_values": [
"Qwen-Image\\Qwen-Image-InstantX-ControlNet-Inpainting.safetensors"
],
"color": "#232",
"bgcolor": "#353"
},
{
"id": 3,
"type": "KSampler",
"pos": [
1221.8179624405964,
195.64542636108396
],
"size": [
315,
262
],
"flags": {},
"order": 12,
"mode": 0,
"inputs": [
{
"name": "model",
"type": "MODEL",
"link": 100
},
{
"name": "positive",
"type": "CONDITIONING",
"link": 123
},
{
"name": "negative",
"type": "CONDITIONING",
"link": 129
},
{
"name": "latent_image",
"type": "LATENT",
"link": 117
}
],
"outputs": [
{
"name": "LATENT",
"type": "LATENT",
"slot_index": 0,
"links": [
35
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.33",
"Node name for S&R": "KSampler"
},
"widgets_values": [
12345,
"fixed",
20,
2.5,
"euler",
"simple",
1
]
},
{
"id": 65,
"type": "VAEEncode",
"pos": [
901.6794795879363,
753.7702701353658
],
"size": [
140,
46
],
"flags": {},
"order": 10,
"mode": 0,
"inputs": [
{
"name": "pixels",
"type": "IMAGE",
"link": 116
},
{
"name": "vae",
"type": "VAE",
"link": 115
}
],
"outputs": [
{
"name": "LATENT",
"type": "LATENT",
"links": [
117
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.76",
"Node name for S&R": "VAEEncode"
},
"widgets_values": []
},
{
"id": 60,
"type": "LoadImage",
"pos": [
191.91826368612072,
700.2476257247591
],
"size": [
351.96372158456404,
427.27847003750594
],
"flags": {},
"order": 3,
"mode": 0,
"inputs": [],
"outputs": [
{
"name": "IMAGE",
"type": "IMAGE",
"links": [
109
]
},
{
"name": "MASK",
"type": "MASK",
"links": [
125
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.76",
"Node name for S&R": "LoadImage",
"image": "clipspace/clipspace-painted-masked-1765378221312.png [input]"
},
"widgets_values": [
"clipspace/clipspace-painted-masked-1765378221312.png [input]",
"image"
]
},
{
"id": 56,
"type": "SaveImage",
"pos": [
1767.8895365326055,
195.64542636108396
],
"size": [
454.5558426040029,
508.83910965100097
],
"flags": {},
"order": 14,
"mode": 0,
"inputs": [
{
"name": "images",
"type": "IMAGE",
"link": 101
}
],
"outputs": [],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.76"
},
"widgets_values": [
"ComfyUI"
]
},
{
"id": 55,
"type": "MarkdownNote",
"pos": [
-52.43482949290891,
-31.4743293038663
],
"size": [
404.3200233043325,
256.7791118275035
],
"flags": {},
"order": 4,
"mode": 0,
"inputs": [],
"outputs": [],
"properties": {},
"widgets_values": [
"## models\n- [qwen_image_fp8_e4m3fn.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/blob/main/split_files/diffusion_models/qwen_image_fp8_e4m3fn.safetensors)\n- [qwen_2.5_vl_7b_fp8_scaled.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/blob/main/split_files/text_encoders/qwen_2.5_vl_7b_fp8_scaled.safetensors)\n- [qwen_image_vae.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/tree/main/split_files/vae)\n- [Qwen-Image-InstantX-ControlNet-Inpainting.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image-InstantX-ControlNets/blob/main/split_files/controlnet/Qwen-Image-InstantX-ControlNet-Inpainting.safetensors)\n\n\n```\n📂ComfyUI/\n└── 📂models/\n ├── 📂controlnet/\n │ └── Qwen-Image-InstantX-ControlNet-Inpainting.safetensors\n ├── 📂diffusion_models/\n │ └── qwen_image_fp8_e4m3fn.safetensors\n ├── 📂text_encoders/\n │ └── qwen_2.5_vl_7b_fp8_scaled.safetensors\n └── 📂vae/\n └── qwen_image_vae.safetensors\n```"
],
"color": "#323",
"bgcolor": "#535"
},
{
"id": 54,
"type": "ModelSamplingAuraFlow",
"pos": [
947.6058895015713,
30.28541934539954
],
"size": [
230.33058166503906,
58
],
"flags": {},
"order": 9,
"mode": 0,
"inputs": [
{
"name": "model",
"type": "MODEL",
"link": 99
}
],
"outputs": [
{
"name": "MODEL",
"type": "MODEL",
"links": [
100
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.49",
"Node name for S&R": "ModelSamplingAuraFlow"
},
"widgets_values": [
3.1000000000000005
]
},
{
"id": 37,
"type": "UNETLoader",
"pos": [
606.4710785828498,
30.28541934539954
],
"size": [
305.3782043457031,
82
],
"flags": {},
"order": 5,
"mode": 0,
"inputs": [],
"outputs": [
{
"name": "MODEL",
"type": "MODEL",
"slot_index": 0,
"links": [
99
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.33",
"Node name for S&R": "UNETLoader"
},
"widgets_values": [
"Qwen-Image\\qwen_image_fp8_e4m3fn.safetensors",
"fp8_e4m3fn"
],
"color": "#323",
"bgcolor": "#535"
}
],
"links": [
[
35,
3,
0,
8,
0,
"LATENT"
],
[
74,
38,
0,
6,
0,
"CLIP"
],
[
75,
38,
0,
7,
0,
"CLIP"
],
[
76,
39,
0,
8,
1,
"VAE"
],
[
99,
37,
0,
54,
0,
"MODEL"
],
[
100,
54,
0,
3,
0,
"MODEL"
],
[
101,
8,
0,
56,
0,
"IMAGE"
],
[
109,
60,
0,
61,
0,
"IMAGE"
],
[
115,
39,
0,
65,
1,
"VAE"
],
[
116,
61,
0,
65,
0,
"IMAGE"
],
[
117,
65,
0,
3,
3,
"LATENT"
],
[
119,
6,
0,
67,
0,
"CONDITIONING"
],
[
120,
7,
0,
67,
1,
"CONDITIONING"
],
[
122,
57,
0,
67,
2,
"CONTROL_NET"
],
[
123,
67,
0,
3,
1,
"CONDITIONING"
],
[
125,
60,
1,
67,
5,
"MASK"
],
[
126,
61,
0,
67,
4,
"IMAGE"
],
[
127,
39,
0,
67,
3,
"VAE"
],
[
129,
67,
1,
3,
2,
"CONDITIONING"
]
],
"groups": [],
"config": {},
"extra": {
"ds": {
"scale": 0.7513148009015778,
"offset": [
152.4348294929089,
131.47432930386628
]
},
"frontendVersion": "1.35.0",
"VHS_latentpreview": false,
"VHS_latentpreviewrate": 0,
"VHS_MetadataImage": true,
"VHS_KeepIntermediate": true
},
"version": 0.4
}
Lightning(高速生成 LoRA)
Qwen-Image-Lightning は、Qwen-Image を 4 / 8 steps で回せるように蒸留した LoRA セット です。
ほとんど劣化なしで大幅にステップ数を減らせるため、かなり多くの workflow で採用されています。
モデルのダウンロード
-
loras
📂ComfyUI/
└── 📂models/
└── 📂loras/
├── Qwen-Image-Lightning-4steps-V2.0-bf16.safetensors
└── Qwen-Image-Lightning-8steps-V2.0-bf16.safetensors
workflow

{
"id": "d8034549-7e0a-40f1-8c2e-de3ffc6f1cae",
"revision": 0,
"last_node_id": 57,
"last_link_id": 103,
"nodes": [
{
"id": 54,
"type": "ModelSamplingAuraFlow",
"pos": [
603.9390258789062,
53.0477294921875
],
"size": [
230.33058166503906,
58
],
"flags": {},
"order": 8,
"mode": 0,
"inputs": [
{
"name": "model",
"type": "MODEL",
"link": 103
}
],
"outputs": [
{
"name": "MODEL",
"type": "MODEL",
"links": [
100
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.49",
"Node name for S&R": "ModelSamplingAuraFlow"
},
"widgets_values": [
3.1000000000000005
]
},
{
"id": 39,
"type": "VAELoader",
"pos": [
957.4939789387397,
67.67436911843055
],
"size": [
256.26084283860405,
58
],
"flags": {},
"order": 0,
"mode": 0,
"inputs": [],
"outputs": [
{
"name": "VAE",
"type": "VAE",
"slot_index": 0,
"links": [
76
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.33",
"Node name for S&R": "VAELoader"
},
"widgets_values": [
"qwen_image_vae.safetensors"
],
"color": "#322",
"bgcolor": "#533"
},
{
"id": 8,
"type": "VAEDecode",
"pos": [
1253.505002517707,
188.1918182373047
],
"size": [
166.0271370269786,
46
],
"flags": {},
"order": 10,
"mode": 0,
"inputs": [
{
"name": "samples",
"type": "LATENT",
"link": 35
},
{
"name": "vae",
"type": "VAE",
"link": 76
}
],
"outputs": [
{
"name": "IMAGE",
"type": "IMAGE",
"slot_index": 0,
"links": [
101
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.33",
"Node name for S&R": "VAEDecode"
},
"widgets_values": []
},
{
"id": 57,
"type": "LoraLoaderModelOnly",
"pos": [
318.1189458223992,
53.0477294921875
],
"size": [
260.9155484908506,
82
],
"flags": {},
"order": 5,
"mode": 0,
"inputs": [
{
"name": "model",
"type": "MODEL",
"link": 102
}
],
"outputs": [
{
"name": "MODEL",
"type": "MODEL",
"links": [
103
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.76",
"Node name for S&R": "LoraLoaderModelOnly"
},
"widgets_values": [
"Qwen-Image\\Qwen-Image-Lightning-8steps-V2.0-bf16.safetensors",
1
],
"color": "#323",
"bgcolor": "#535"
},
{
"id": 37,
"type": "UNETLoader",
"pos": [
-12.163790088960356,
53.0477294921875
],
"size": [
305.3782043457031,
82
],
"flags": {},
"order": 1,
"mode": 0,
"inputs": [],
"outputs": [
{
"name": "MODEL",
"type": "MODEL",
"slot_index": 0,
"links": [
102
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.33",
"Node name for S&R": "UNETLoader"
},
"widgets_values": [
"Qwen-Image\\qwen_image_fp8_e4m3fn.safetensors",
"fp8_e4m3fn"
],
"color": "#323",
"bgcolor": "#535"
},
{
"id": 6,
"type": "CLIPTextEncode",
"pos": [
415.00001525878906,
186
],
"size": [
419.26959228515625,
156.00363159179688
],
"flags": {},
"order": 6,
"mode": 0,
"inputs": [
{
"name": "clip",
"type": "CLIP",
"link": 74
}
],
"outputs": [
{
"name": "CONDITIONING",
"type": "CONDITIONING",
"slot_index": 0,
"links": [
46
]
}
],
"title": "CLIP Text Encode (Positive Prompt)",
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.33",
"Node name for S&R": "CLIPTextEncode"
},
"widgets_values": [
"A soft hand-drawn illustration for ComfyUI: A girl with bold solid pink hair, wearing oversized yellow headphones, a loose gray dress, bright orange tights, and chunky purple platform boots stands on a grassy hill, facing away from the viewer. The background features a luminous blue sky filled with huge, fluffy white clouds, distant misty mountain ranges, and sleek modern campus buildings on the horizon. At the bottom, large, bold ‘ComfyUI’ letters anchor the composition. The overall aesthetic uses gentle yet vibrant colors to convey themes of personal growth, curiosity, and quiet responsibility."
]
},
{
"id": 3,
"type": "KSampler",
"pos": [
898.7548217773438,
188.1918182373047
],
"size": [
315,
262
],
"flags": {},
"order": 9,
"mode": 0,
"inputs": [
{
"name": "model",
"type": "MODEL",
"link": 100
},
{
"name": "positive",
"type": "CONDITIONING",
"link": 46
},
{
"name": "negative",
"type": "CONDITIONING",
"link": 52
},
{
"name": "latent_image",
"type": "LATENT",
"link": 98
}
],
"outputs": [
{
"name": "LATENT",
"type": "LATENT",
"slot_index": 0,
"links": [
35
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.33",
"Node name for S&R": "KSampler"
},
"widgets_values": [
12345,
"fixed",
8,
1,
"euler",
"simple",
1
]
},
{
"id": 7,
"type": "CLIPTextEncode",
"pos": [
415,
405.392333984375
],
"size": [
418.3189392089844,
107.08506774902344
],
"flags": {
"collapsed": true
},
"order": 7,
"mode": 0,
"inputs": [
{
"name": "clip",
"type": "CLIP",
"link": 75
}
],
"outputs": [
{
"name": "CONDITIONING",
"type": "CONDITIONING",
"slot_index": 0,
"links": [
52
]
}
],
"title": "CLIP Text Encode (Negative Prompt)",
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.33",
"Node name for S&R": "CLIPTextEncode"
},
"widgets_values": [
""
]
},
{
"id": 53,
"type": "EmptySD3LatentImage",
"pos": [
597.2696075439453,
513.0546240234369
],
"size": [
237,
106
],
"flags": {},
"order": 2,
"mode": 0,
"inputs": [],
"outputs": [
{
"name": "LATENT",
"type": "LATENT",
"links": [
98
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.49",
"Node name for S&R": "EmptySD3LatentImage"
},
"widgets_values": [
1328,
1328,
1
]
},
{
"id": 56,
"type": "SaveImage",
"pos": [
1459.2823202850486,
188.1918182373047
],
"size": [
429.4043426040039,
469.26990965100083
],
"flags": {},
"order": 11,
"mode": 0,
"inputs": [
{
"name": "images",
"type": "IMAGE",
"link": 101
}
],
"outputs": [],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.76"
},
"widgets_values": [
"ComfyUI"
]
},
{
"id": 38,
"type": "CLIPLoader",
"pos": [
56.288665771484375,
312.74468994140625
],
"size": [
301.3524169921875,
106
],
"flags": {},
"order": 3,
"mode": 0,
"inputs": [],
"outputs": [
{
"name": "CLIP",
"type": "CLIP",
"slot_index": 0,
"links": [
74,
75
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.33",
"Node name for S&R": "CLIPLoader"
},
"widgets_values": [
"qwen_2.5_vl_7b_fp8_scaled.safetensors",
"qwen_image",
"default"
],
"color": "#432",
"bgcolor": "#653"
},
{
"id": 55,
"type": "MarkdownNote",
"pos": [
-491.05311465172593,
-31.268711326654767
],
"size": [
446.46560633233355,
271.0212201528051
],
"flags": {},
"order": 4,
"mode": 0,
"inputs": [],
"outputs": [],
"properties": {},
"widgets_values": [
"## models\n- [qwen_image_fp8_e4m3fn.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/blob/main/split_files/diffusion_models/qwen_image_fp8_e4m3fn.safetensors)\n- [qwen_2.5_vl_7b_fp8_scaled.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/blob/main/split_files/text_encoders/qwen_2.5_vl_7b_fp8_scaled.safetensors)\n- [qwen_image_vae.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/tree/main/split_files/vae)\n- [Qwen-Image-Lightning-4steps-V2.0-bf16.safetensors](https://huggingface.co/lightx2v/Qwen-Image-Lightning/blob/main/Qwen-Image-Lightning-4steps-V2.0-bf16.safetensors)\n- [Qwen-Image-Lightning-8steps-V2.0-bf16.safetensors](https://huggingface.co/lightx2v/Qwen-Image-Lightning/blob/main/Qwen-Image-Lightning-8steps-V2.0-bf16.safetensors)\n\n\n\n```\n📂ComfyUI/\n└── 📂models/\n ├── 📂diffusion_models/\n │ └── qwen_image_fp8_e4m3fn.safetensors\n ├── 📂loras/\n │ ├── Qwen-Image-Lightning-4steps-V2.0-bf16.safetensors\n │ └── Qwen-Image-Lightning-8steps-V2.0-bf16.safetensors\n ├── 📂text_encoders/\n │ └── qwen_2.5_vl_7b_fp8_scaled.safetensors\n └── 📂vae/\n └── qwen_image_vae.safetensors\n```"
],
"color": "#323",
"bgcolor": "#535"
}
],
"links": [
[
35,
3,
0,
8,
0,
"LATENT"
],
[
46,
6,
0,
3,
1,
"CONDITIONING"
],
[
52,
7,
0,
3,
2,
"CONDITIONING"
],
[
74,
38,
0,
6,
0,
"CLIP"
],
[
75,
38,
0,
7,
0,
"CLIP"
],
[
76,
39,
0,
8,
1,
"VAE"
],
[
98,
53,
0,
3,
3,
"LATENT"
],
[
100,
54,
0,
3,
0,
"MODEL"
],
[
101,
8,
0,
56,
0,
"IMAGE"
],
[
102,
37,
0,
57,
0,
"MODEL"
],
[
103,
57,
0,
54,
0,
"MODEL"
]
],
"groups": [],
"config": {},
"extra": {
"ds": {
"scale": 0.7513148009015781,
"offset": [
591.053114651726,
131.26871132665477
]
},
"frontendVersion": "1.35.0",
"VHS_latentpreview": false,
"VHS_latentpreviewrate": 0,
"VHS_MetadataImage": true,
"VHS_KeepIntermediate": true
},
"version": 0.4
}
LoraLoaderModelOnlyノードで Lightning LoRA を読み込みます。KSamplerのstepsを 4 または 8、CFGを 1.0 に設定します。