Sign In

FLUX - IMG to IMG

31

Nov 17, 2024

(Updated: 8 days ago)

workflows
FLUX - IMG to IMG

Step-by-Step Guide Series:
ComfyUI - IMG to IMG Workflow

This article accompanies this workflow: link

Workflow description :

The aim of this workflow is to generate images from another one and a text in a simple window.

Prerequisites :

đź“‚Files for "base" version :

Model : flux1-dev-fp8.safetensors or flux1-dev-fp16.safetensors
in ComfyUI\models\diffusion_models

CLIP : clip_l.safetensors
in ComfyUI\models\clip

đź“‚Files for GGUF version :

Model : Q8, Q6, Q5, Q4, Q3
in ComfyUI\models\unet

CLIP : Q8, Q6, Q5, Q4, Q3
in ComfyUI\models\clip

đź“‚Files for NUNCHAKU version :

Model : svdq-int4_r32-flux.1-dev.safetensors
in ComfyUI\models\diffusion_models

đź“‚Common Files

Text encoder : t5xxl_fp8_e4m3fn.safetensors or ViT-L-14-TEXT-detail-improved-hiT-GmP-TE-only-HF.safetensors

VAE : ae.safetensors
in ComfyUI\models\vae

ANY upscale model :

in ComfyUI\models\upscale_models

📦Custom Nodes :

Don't forget to close the workflow and open it again once the nodes have been installed.

Usage :

Capture d’écran 2025-10-22 175756.png

In this new version of the workflow everything is organized by color:

  • Green is what you want to create, also called prompt,

  • Yellow is all the parameters to adjust the video,

  • Blue are the model files used by the workflow,

  • Purple is for LoRA.

We will now see how to use each node:

Write what you want in the “Prompt” node :

image.png

Select image format :

Choose the guidance level :

I recommend between 3.5 and 4.5. The lower the number, the freer you leave the model. The higher the number, the more the image will resemble what you “strictly” asked for.

Choose a scheduler, number of steps and denoise level :

I recommend normal or beta and between 20 and 30. The higher the number, the better the quality, but the longer it takes to get an image.

The denoise allows you to choose how the base image will influence the new one. At 0, the image will be exactly the same (useless), and at 1, the image will be completely new (so your image will be ignored).

I recommend starting around 0.7/0.8 and reducing or increasing it depending on the desired result.

Choose a sampler :

I recommend euler.

Define a seed or let comfy generate one:

Add how many LoRA you want to use, and define it :

If you dont know what is LoRA just dont active any.

Load your base image :

image.png

Choose your model:

Depending on whether you've chosen basic or gguf workflow, this setting changes. I personally use the gguf Q8_0 version.

Choose a FLUX clip encoder and a text encoder :

I personally use the GGUF Q8_0 encoder and the text encoder ViT-L-14-TEXT-detail-improved-hiT-GmP-TE-only-HF.

Select an upscaler : (optional)

I personally use RealESRGAN_x4plus.pth.

Select upscale ratio : (optional)

This is the multiplier that will be used by the upscaler, with 2 the image will be twice as large.

Now you're ready to create your image.

Just click on the “Queue” button to start:

Once rendering is complete, the image appears in the “image viewer” node.

If you have enabled upscaling, a slider will show the base image and the upscaled version.

image.png

Some additional information:

Organization of recordings:

All generated files are stored in comfyui/output/FLUX/YYYY-MM-DD.

Depending on the options chosen you will find:

  • "YYYY-MM-DD-HH-MM_XXXXXXXXXX_Original" the basic file,

  • "YYYY-MM-DD-HH-MM_XXXXXXXXXX_Upscale" the upscaled.

This guide is now complete. If you have any questions or suggestions, don't hesitate to post a comment.

31