Model:CustomModel(その他)
prompt: {"3": {"inputs": {"seed": 521206115896772, "steps": 16, "cfg": 2.0, "sampler_name": "euler", "scheduler": "simple", "denoise": 1.0, "model": ["66", 0], "positive": ["6", 0], "negative": ["7", 0], "latent_image": ["58", 0]}, "class_type": "KSampler", "_meta": {"title": "K\u30b5\u30f3\u30d7\u30e9\u30fc"}}, "6": {"inputs": {"text": "anime style, a The girl is lying face down and sleeping on the bed, purple fart smoke is coming out of her bum, living room, funny manga style", "clip": ["38", 0]}, "class_type": "CLIPTextEncode", "_meta": {"title": "CLIP Text Encode (Positive Prompt)"}}, "7": {"inputs": {"text": "", "clip": ["38", 0]}, "class_type": "CLIPTextEncode", "_meta": {"title": "CLIP Text Encode (Negative Prompt)"}}, "8": {"inputs": {"samples": ["3", 0], "vae": ["39", 0]}, "class_type": "VAEDecode", "_meta": {"title": "VAE\u30c7\u30b3\u30fc\u30c9"}}, "37": {"inputs": {"unet_name": "qwen_image_fp8_e4m3fn.safetensors", "weight_dtype": "default"}, "class_type": "UNETLoader", "_meta": {"title": "\u62e1\u6563\u30e2\u30c7\u30eb\u3092\u8aad\u307f\u8fbc\u3080"}}, "38": {"inputs": {"clip_name": "qwen_2.5_vl_7b_fp8_scaled.safetensors", "type": "qwen_image", "device": "default"}, "class_type": "CLIPLoader", "_meta": {"title": "CLIP\u3092\u8aad\u307f\u8fbc\u3080"}}, "39": {"inputs": {"vae_name": "qwen_image_vae.safetensors"}, "class_type": "VAELoader", "_meta": {"title": "VAE\u3092\u8aad\u307f\u8fbc\u3080"}}, "58": {"inputs": {"width": 1328, "height": 1328, "batch_size": 1}, "class_type": "EmptySD3LatentImage", "_meta": {"title": "\u7a7a\u306eSD3\u6f5c\u5728\u753b\u50cf"}}, "60": {"inputs": {"filename_prefix": "ComfyUI", "images": ["8", 0]}, "class_type": "SaveImage", "_meta": {"title": "\u753b\u50cf\u3092\u4fdd\u5b58"}}, "66": {"inputs": {"shift": 3.1000000000000005, "model": ["37", 0]}, "class_type": "ModelSamplingAuraFlow", "_meta": {"title": "\u30e2\u30c7\u30eb\u30b5\u30f3\u30d7\u30ea\u30f3\u30b0\u30aa\u30fc\u30e9\u30d5\u30ed\u30fc"}}}
workflow: {"id": "91f6bbe2-ed41-4fd6-bac7-71d5b5864ecb", "revision": 0, "last_node_id": 74, "last_link_id": 130, "nodes": [{"id": 39, "type": "VAELoader", "pos": [20, 340], "size": [330, 60], "flags": {}, "order": 0, "mode": 0, "inputs": [], "outputs": [{"name": "VAE", "type": "VAE", "slot_index": 0, "links": [76]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.48", "Node name for S&R": "VAELoader", "models": [{"name": "qwen_image_vae.safetensors", "url": "https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/vae/qwen_image_vae.safetensors", "directory": "vae"}], "enableTabs": false, "tabWidth": 65, "tabXOffset": 10, "hasSecondTab": false, "secondTabText": "Send Back", "secondTabOffset": 80, "secondTabWidth": 65, "widget_ue_connectable": {}}, "widgets_values": ["qwen_image_vae.safetensors"]}, {"id": 38, "type": "CLIPLoader", "pos": [20, 190], "size": [330, 110], "flags": {}, "order": 1, "mode": 0, "inputs": [], "outputs": [{"name": "CLIP", "type": "CLIP", "slot_index": 0, "links": [74, 75]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.48", "Node name for S&R": "CLIPLoader", "models": [{"name": "qwen_2.5_vl_7b_fp8_scaled.safetensors", "url": "https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/text_encoders/qwen_2.5_vl_7b_fp8_scaled.safetensors", "directory": "text_encoders"}], "enableTabs": false, "tabWidth": 65, "tabXOffset": 10, "hasSecondTab": false, "secondTabText": "Send Back", "secondTabOffset": 80, "secondTabWidth": 65, "widget_ue_connectable": {}}, "widgets_values": ["qwen_2.5_vl_7b_fp8_scaled.safetensors", "qwen_image", "default"]}, {"id": 58, "type": "EmptySD3LatentImage", "pos": [50, 510], "size": [270, 106], "flags": {}, "order": 2, "mode": 0, "inputs": [], "outputs": [{"name": "LATENT", "type": "LATENT", "links": [107]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.48", "Node name for S&R": "EmptySD3LatentImage", "enableTabs": false, "tabWidth": 65, "tabXOffset": 10, "hasSecondTab": false, "secondTabText": "Send Back", "secondTabOffset": 80, "secondTabWidth": 65, "widget_ue_connectable": {}}, "widgets_values": [1328, 1328, 1]}, {"id": 7, "type": "CLIPTextEncode", "pos": [390, 440], "size": [425.27801513671875, 180.6060791015625], "flags": {}, "order": 10, "mode": 0, "inputs": [{"name": "clip", "type": "CLIP", "link": 75}], "outputs": [{"name": "CONDITIONING", "type": "CONDITIONING", "slot_index": 0, "links": [52]}], "title": "CLIP Text Encode (Negative Prompt)", "properties": {"cnr_id": "comfy-core", "ver": "0.3.48", "Node name for S&R": "CLIPTextEncode", "enableTabs": false, "tabWidth": 65, "tabXOffset": 10, "hasSecondTab": false, "secondTabText": "Send Back", "secondTabOffset": 80, "secondTabWidth": 65, "widget_ue_connectable": {}}, "widgets_values": [""], "color": "#322", "bgcolor": "#533"}, {"id": 70, "type": "Note", "pos": [850, 430], "size": [300, 120], "flags": {}, "order": 3, "mode": 0, "inputs": [], "outputs": [], "title": "For fp8 without 8steps LoRA", "properties": {}, "widgets_values": ["Set cfg to 1.0 for a speed boost at the cost of consistency. Samplers like res_multistep work pretty well at cfg 1.0\n\nThe official number of steps is 50 but I think that's too much. Even just 10 steps seems to work."], "color": "#432", "bgcolor": "#653"}, {"id": 66, "type": "ModelSamplingAuraFlow", "pos": [850, 10], "size": [300, 58], "flags": {}, "order": 12, "mode": 0, "inputs": [{"name": "model", "type": "MODEL", "link": 130}], "outputs": [{"name": "MODEL", "type": "MODEL", "links": [125]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.48", "Node name for S&R": "ModelSamplingAuraFlow", "enableTabs": false, "tabWidth": 65, "tabXOffset": 10, "hasSecondTab": false, "secondTabText": "Send Back", "secondTabOffset": 80, "secondTabWidth": 65, "widget_ue_connectable": {}}, "widgets_values": [3.1000000000000005]}, {"id": 73, "type": "LoraLoaderModelOnly", "pos": [460, 60], "size": [270, 82], "flags": {}, "order": 11, "mode": 4, "inputs": [{"name": "model", "type": "MODEL", "link": 129}], "outputs": [{"name": "MODEL", "type": "MODEL", "links": [130]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.49", "Node name for S&R": "LoraLoaderModelOnly", "models": [{"name": "Qwen-Image-Lightning-8steps-V1.0.safetensors", "url": "https://huggingface.co/lightx2v/Qwen-Image-Lightning/resolve/main/Qwen-Image-Lightning-8steps-V1.0.safetensors", "directory": "loras"}]}, "widgets_values": ["Qwen-Image-Lightning-8steps-V1.0.safetensors", 1]}, {"id": 69, "type": "MarkdownNote", "pos": [-540, -220], "size": [390, 180], "flags": {}, "order": 4, "mode": 0, "inputs": [], "outputs": [], "title": "VRAM Usage", "properties": {"widget_ue_connectable": {}}, "widgets_values": ["## GPU:RTX4090D 24GB\n\n| Configuration | VRAM Usage | 1st Generation | 2nd Generation |\n|---------------------|---------------|---------------|-----------------|\n| Fp8_e4m3fn | 86% | \u2248 94s | \u2248 71s |\n| With 8steps LoRA | 86% | \u2248 55s | \u2248 34s |\n| Distill fp8_e4m3fn | 86% | \u2248 69s | \u2248 36s |"], "color": "#432", "bgcolor": "#653"}, {"id": 71, "type": "Note", "pos": [850, -120], "size": [300, 88], "flags": {}, "order": 5, "mode": 0, "inputs": [], "outputs": [], "properties": {}, "widgets_values": ["Increase the shift if you get too many blury/dark/bad images. Decrease if you want to try increasing detail."], "color": "#432", "bgcolor": "#653"}, {"id": 8, "type": "VAEDecode", "pos": [1170, -90], "size": [210, 46], "flags": {"collapsed": false}, "order": 14, "mode": 0, "inputs": [{"name": "samples", "type": "LATENT", "link": 128}, {"name": "vae", "type": "VAE", "link": 76}], "outputs": [{"name": "IMAGE", "type": "IMAGE", "slot_index": 0, "links": [110]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.48", "Node name for S&R": "VAEDecode", "enableTabs": false, "tabWidth": 65, "tabXOffset": 10, "hasSecondTab": false, "secondTabText": "Send Back", "secondTabOffset": 80, "secondTabWidth": 65, "widget_ue_connectable": {}}, "widgets_values": []}, {"id": 37, "type": "UNETLoader", "pos": [20, 50], "size": [330, 90], "flags": {}, "order": 6, "mode": 0, "inputs": [], "outputs": [{"name": "MODEL", "type": "MODEL", "slot_index": 0, "links": [129]}], "properties": {"cnr_id": "comfy-core", "ver": "0.3.48", "Node name for S&R": "UNETLoader", "models": [{"name": "qwen_image_fp8_e4m3fn.safetensors", "url": "https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/diffusion_models/qwen_image_fp8_e4m3fn.safetensors", "directory": "diffusion_models"}], "enableTabs": false, "tabWidth": 65, "tabXOffset": 10, "hasSecondTab": false, "secondTabText": "Send Back", "secondTabOffset": 80, "secondTabWidth": 65, "widget_ue_connectable": {}}, "widgets_values": ["qwen_image_fp8_e4m3fn.safetensors", "default"]}, {"id": 67, "type": "MarkdownNote", "pos": [-540, 10], "size": [540, 630], "flags": {}, "order": 7, "mode": 0, "inputs": [], "outputs": [], "title": "Model links", "properties": {"widget_ue_connectable": {}}, "widgets_values": ["[Tutorial](https://docs.comfy.org/tutorials/image/qwen/qwen-image) | [\u6559\u7a0b](https://docs.comfy.org/zh-CN/tutorials/image/qwen/qwen-image)\n\n\n## Model links\n\nYou can find all the models on [Huggingface](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/tree/main) or [Modelscope](https://modelscope.cn/models/Comfy-Org/Qwen-Image_ComfyUI/files)\n\n**Diffusion model**\n\n- [qwen_image_fp8_e4m3fn.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/diffusion_models/qwen_image_fp8_e4m3fn.safetensors)\n\nQwen_image_distill\n\n- [qwen_image_distill_full_fp8_e4m3fn.safetensors](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/non_official/diffusion_models/qwen_image_distill_full_fp8_e4m3fn.safetensors)\n- [qwen_image_distill_full_bf16.safetensors](https://huggingface.co/Comfy-Org...
ダウンロード
ダウンロード
ダウンロード
ダウンロード
ダウンロード
ダウンロード
ダウンロード
ダウンロード
ダウンロード
ダウンロード
ダウンロード
ダウンロード
ダウンロード
ダウンロード
ダウンロード
ダウンロード
ダウンロード
ダウンロード
ダウンロード
ダウンロード
ダウンロード
ダウンロード
ダウンロード
ダウンロード
ダウンロード
ダウンロード
ダウンロード
ダウンロード
ダウンロード
ダウンロード
ダウンロード
ダウンロード
ダウンロード
ダウンロード