Type | Workflows |
Stats | 167 |
Reviews | (16) |
Published | Nov 5, 2024 |
Base Model | |
Hash | AutoV2 0348FD5F8F |
The workflow is designed to create highres architectural images (tested 12288x8192) from lowres outputs of any kind of rendering software (tested 1536x1024) and tries to keep details throughout its process of "staged generation": a first stage with sdxl, a second stage for detailing first stages output with flux, a third stage upscaling the seconds stages output with flux again.
More infos here:
I assumed, people who are interested in this whole project, will a) find a quick way or already know how to use a 3d environment like e.g. 3dsmax, blender, sketchup, etc. to create the outputs needed, b) adopt some of the things they see here into their own workflows and/or modify everything to their needs, if they want to use this kind of stuff the way I do.
Control over the desired outcome is mainly gained through controlnets in first stage and the help of a masked detail transfer for your base image, where you define the mask by a prompt (e.g. “house, facade, etc. - wherever your details are that you want to transfer/keep throughout the stages you activated). And if you for example have an area where you placed a person with the MaskEditor, the mask gets edited within the process to prevent detail being blended onto that person from your mask. Basically, I’m using various models in a row to add detail in each step or bypass stages that I don’t want while using the workflow, it is only in some cases a straightforward process, still for example I am cherrypicking first stages outputs with a preview chooser before passing it to the next stage.
Depending on the models you use, it imagines photorealistic images from renderings like a "per-polygon-unique-colored-mesh" or some kind of outlines/wireframe-meshes/etc. through one or two (or how many you would add) controlnets. Anything that a sdxl controlnet-preprocessor or your controlnet directly will understand, can be used. In advance you can control the amount of the detail transfer and most of the basic functions with sliders and switches (no, I am not a UI or UX designer). Your Prompt then defines the general output, I like to keep it separated to quickly adjust things in the generation, but it just gets concatenated at the end. You may have to edit your 3d output/base image before generation, for example I painted some vertical tiling lines for my facade directly onto the normalpass renderelement in photoshop.
In addition, you can change the base settings to have an img2img workflow with one button, keeping its functionality if you already have some kind of more or less photorelistic rendering in the base image input. You may want to denoise that at lower values in first stage, then let flux add details in stage 2 & 3. Most of its additional features, e.g. activating “people generation” and using a MaskEditor paintbrush for placing a flux generated person onto your scene, are considered to be a proof of concept as you can see from the examples.
This workflow is:
A potential replacement to many paid services like image enhancement and upscaling
A “tool” developed for myself to assist my daily doings as a technical artist according to my needs from previsualisation to a final image, in general is based on my best intention, latest findings and limited knowledge of AI itself. I still use my 3d environment and additional rendering software and for example still often postprocess my images manually :)
Therefore, this workflow unfortunately is NOT:
a masterpiece comfyUI workflow never seen by mankind before - some might have guessed that
ultimate magic technology that every time you start generation with, makes you receive an award winning image - not yet, but I promise I’ll let you know asap when I have it. I may not will give it away for free then
This workflows output in any case will depend on:
Your base image input, precisely in context to the purpose of this workflow: your skills in your favourite 3d environment for base image creation. I have not tested this thing with any other stuff besides architecture related imagery
Your ability to describe what you want as a prompt/prompting in general
Your Hardware (basically if you can run flux.dev, you can run this, optimization may follow, tested on gf rtx 4090, use more performant models and reduce resolution to make this work on lower hardware)
Your creativity to use/edit/adopt something like this, in a way that fit your needs
Your understanding of how comfyui and controlnets work and knowledge of which exact settings may work for your scenario
Bonuscontent:
Because at the moment i cant use "mixlabs screen share" due some display bugs, this workflow is part of the release too, you find it on the right under "Experimental" (may have to pan a little to see it). Replace the Load image node with mixlabs screen share and turn on autoqueue in comfyui to use it as shown in the video. You may want to bypass everything else then.
To use the experimental "Flux inpaint by Mask" feature, connect a mask to the respective node in the "mask area" (yellow typo), enable this process on the base cfg and use a prompt for what you want to see - this feature is real experiment and does not give always the desired results.
Models used:
flux.dev gguf Q8_0.gguf
juggernaut XI.safetensors
realVisXL40_Turbo.safetensors
clip
t5-v1_1-xxl-encoder-Q8_0.gguf
clip_l.safetensors
ip-adapter
CLIP.ViT-H-14-laion2B-s32B-b79K.safetensors
ip-adapter-plus_sdxl_vit-h.safetensors
controlnet
diffusers_xl_depth_full.safetensors
diffusers_xl_canny_full.safetensors
thibaud_xl_openpose.safetensors (optional, to be re-implemented with openpose-editor for posed people in future release)
sam2/florence2
sam2_hiera_base_plus.safetensors
Florence2-base
upscale
4x-UltraSharp.pth
Recommended models to try: CrystalClearXL, RealVisXL, ProtoVision XL,
Customnodes used (yes, ressource heavy, even may be edited/added in future, recommended to install ONE-BY-ONE and restart comyui in between to prevent errors):
GitHub - ltdrdata/ComfyUI-Manager
GitHub - ltdrdata/ComfyUI-Impact-Pack
GitHub - Fannovel16/comfyui_controlnet_aux
GitHub - jags111/efficiency-nodes-comfyui
GitHub - WASasquatch/was-node-suite-comfyui
GitHub - EllangoK/ComfyUI-post-processing-nodes
GitHub - BadCafeCode/masquerade-nodes-comfyui
GitHub - city96/ComfyUI-GGUF
GitHub - pythongosssss/ComfyUI-Custom-Scripts
GitHub - ssitu/ComfyUI_UltimateSDUpscale
GitHub - melMass/comfy_mtb
GitHub - Suzie1/ComfyUI_Comfyroll_CustomNodes
GitHub - cubiq/ComfyUI_IPAdapter_plus
GitHub - sipherxyz/comfyui-art-venture
GitHub - evanspearman/ComfyMath: Math nodes for ComfyUI
GitHub - jamesWalker55/comfyui-various
GitHub - Kosinkadink/ComfyUI-Advanced-ControlNet
GitHub - theUpsider/ComfyUI-Logic
GitHub - rgthree/rgthree-comfy
GitHub - cubiq/ComfyUI_essentials
GitHub - chrisgoringe/cg-image-picker
GitHub - kijai/ComfyUI-KJNodes
GitHub - kijai/ComfyUI-DepthAnythingV2
GitHub - kijai/ComfyUI-Florence2
GitHub - kijai/ComfyUI-segment-anything-2
GitHub - shadowcz007/comfyui-mixlab-nodes
GitHub - palant/image-resize-comfyui
GitHub - yolain/ComfyUI-Easy-Use
All of the above nodes are outstanding work and highly recommended. If you consider to support my work, please support the developers first - they are the real mvp's here. ONLY IF YOU WANT TO SUPPORT MY WORK DIRECTLY (NO NEED TO, NO ADVANTAGE/DISADVANTAGE GAINED) you can donate at https://ko-fi.com/paulhansen