Sign In

PH's Archviz x AI ComfyUI Workflow (SDXL + FLUX)

16
167
2
Type
Workflows
Stats
167
Reviews
Published
Nov 5, 2024
Base Model
Flux.1 D
Hash
AutoV2
0348FD5F8F
default creator card background decoration
ph01's Avatar
ph01
The FLUX.1 [dev] Model is licensed by Black Forest Labs. Inc. under the FLUX.1 [dev] Non-Commercial License. Copyright Black Forest Labs. Inc.
IN NO EVENT SHALL BLACK FOREST LABS, INC. BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH USE OF THIS MODEL.

The workflow is designed to create highres architectural images (tested 12288x8192) from lowres outputs of any kind of rendering software (tested 1536x1024) and tries to keep details throughout its process of "staged generation": a first stage with sdxl, a second stage for detailing first stages output with flux, a third stage upscaling the seconds stages output with flux again.

More infos here:

youtube showcase

original reddit thread

I assumed, people who are interested in this whole project, will a) find a quick way or already know how to use a 3d environment like e.g. 3dsmax, blender, sketchup, etc. to create the outputs needed, b) adopt some of the things they see here into their own workflows and/or modify everything to their needs, if they want to use this kind of stuff the way I do.

Control over the desired outcome is mainly gained through controlnets in first stage and the help of a masked detail transfer for your base image, where you define the mask by a prompt (e.g. “house, facade, etc. - wherever your details are that you want to transfer/keep throughout the stages you activated). And if you for example have an area where you placed a person with the MaskEditor, the mask gets edited within the process to prevent detail being blended onto that person from your mask. Basically, I’m using various models in a row to add detail in each step or bypass stages that I don’t want while using the workflow, it is only in some cases a straightforward process, still for example I am cherrypicking first stages outputs with a preview chooser before passing it to the next stage.

Depending on the models you use, it imagines photorealistic images from renderings like a "per-polygon-unique-colored-mesh" or some kind of outlines/wireframe-meshes/etc. through one or two (or how many you would add) controlnets. Anything that a sdxl controlnet-preprocessor or your controlnet directly will understand, can be used. In advance you can control the amount of the detail transfer and most of the basic functions with sliders and switches (no, I am not a UI or UX designer). Your Prompt then defines the general output, I like to keep it separated to quickly adjust things in the generation, but it just gets concatenated at the end. You may have to edit your 3d output/base image before generation, for example I painted some vertical tiling lines for my facade directly onto the normalpass renderelement in photoshop.

In addition, you can change the base settings to have an img2img workflow with one button, keeping its functionality if you already have some kind of more or less photorelistic rendering in the base image input. You may want to denoise that at lower values in first stage, then let flux add details in stage 2 & 3. Most of its additional features, e.g. activating “people generation” and using a MaskEditor paintbrush for placing a flux generated person onto your scene, are considered to be a proof of concept as you can see from the examples.

This workflow is:

  • A potential replacement to many paid services like image enhancement and upscaling

  • A “tool” developed for myself to assist my daily doings as a technical artist according to my needs from previsualisation to a final image, in general is based on my best intention, latest findings and limited knowledge of AI itself. I still use my 3d environment and additional rendering software and for example still often postprocess my images manually :)

Therefore, this workflow unfortunately is NOT:

  • a masterpiece comfyUI workflow never seen by mankind before - some might have guessed that

  • ultimate magic technology that every time you start generation with, makes you receive an award winning image - not yet, but I promise I’ll let you know asap when I have it. I may not will give it away for free then

This workflows output in any case will depend on:

  • Your base image input, precisely in context to the purpose of this workflow: your skills in your favourite 3d environment for base image creation. I have not tested this thing with any other stuff besides architecture related imagery

  • Your ability to describe what you want as a prompt/prompting in general

  • Your Hardware (basically if you can run flux.dev, you can run this, optimization may follow, tested on gf rtx 4090, use more performant models and reduce resolution to make this work on lower hardware)

  • Your creativity to use/edit/adopt something like this, in a way that fit your needs

  • Your understanding of how comfyui and controlnets work and knowledge of which exact settings may work for your scenario

Bonuscontent:

Because at the moment i cant use "mixlabs screen share" due some display bugs, this workflow is part of the release too, you find it on the right under "Experimental" (may have to pan a little to see it). Replace the Load image node with mixlabs screen share and turn on autoqueue in comfyui to use it as shown in the video. You may want to bypass everything else then.

To use the experimental "Flux inpaint by Mask" feature, connect a mask to the respective node in the "mask area" (yellow typo), enable this process on the base cfg and use a prompt for what you want to see - this feature is real experiment and does not give always the desired results.

Models used:

  • flux.dev gguf Q8_0.gguf

  • juggernaut XI.safetensors

  • realVisXL40_Turbo.safetensors

clip

  • t5-v1_1-xxl-encoder-Q8_0.gguf

  • clip_l.safetensors

ip-adapter

  • CLIP.ViT-H-14-laion2B-s32B-b79K.safetensors

  • ip-adapter-plus_sdxl_vit-h.safetensors

controlnet

  • diffusers_xl_depth_full.safetensors

  • diffusers_xl_canny_full.safetensors

  • thibaud_xl_openpose.safetensors (optional, to be re-implemented with openpose-editor for posed people in future release)

sam2/florence2

  • sam2_hiera_base_plus.safetensors

  • Florence2-base

upscale

  • 4x-UltraSharp.pth

Recommended models to try: CrystalClearXL, RealVisXL, ProtoVision XL,

Customnodes used (yes, ressource heavy, even may be edited/added in future, recommended to install ONE-BY-ONE and restart comyui in between to prevent errors):

GitHub - ltdrdata/ComfyUI-Manager

GitHub - ltdrdata/ComfyUI-Impact-Pack

GitHub - Fannovel16/comfyui_controlnet_aux

GitHub - jags111/efficiency-nodes-comfyui

GitHub - WASasquatch/was-node-suite-comfyui

GitHub - EllangoK/ComfyUI-post-processing-nodes

GitHub - BadCafeCode/masquerade-nodes-comfyui

GitHub - city96/ComfyUI-GGUF

GitHub - pythongosssss/ComfyUI-Custom-Scripts

GitHub - ssitu/ComfyUI_UltimateSDUpscale

GitHub - melMass/comfy_mtb

GitHub - Suzie1/ComfyUI_Comfyroll_CustomNodes

GitHub - cubiq/ComfyUI_IPAdapter_plus

GitHub - sipherxyz/comfyui-art-venture

GitHub - evanspearman/ComfyMath: Math nodes for ComfyUI

GitHub - jamesWalker55/comfyui-various

GitHub - Kosinkadink/ComfyUI-Advanced-ControlNet

GitHub - theUpsider/ComfyUI-Logic

GitHub - rgthree/rgthree-comfy

GitHub - cubiq/ComfyUI_essentials

GitHub - chrisgoringe/cg-image-picker

GitHub - kijai/ComfyUI-KJNodes

GitHub - kijai/ComfyUI-DepthAnythingV2

GitHub - kijai/ComfyUI-Florence2

GitHub - kijai/ComfyUI-segment-anything-2

GitHub - shadowcz007/comfyui-mixlab-nodes

GitHub - palant/image-resize-comfyui

GitHub - yolain/ComfyUI-Easy-Use

All of the above nodes are outstanding work and highly recommended. If you consider to support my work, please support the developers first - they are the real mvp's here. ONLY IF YOU WANT TO SUPPORT MY WORK DIRECTLY (NO NEED TO, NO ADVANTAGE/DISADVANTAGE GAINED) you can donate at https://ko-fi.com/paulhansen