Sign In

Rebels HiDream-01 Image Dev

Download

1 variant available

Archive Other

1.63 KB

Verified:

Type

Workflows

Stats

128

Reviews

Published

May 11, 2026

Base Model

HiDream

Hash

AutoV2
8038F23ED4

VERSION 2 (please read)

NOW SUPPORTS MULTI-REFERENCE EDITING!

Added Ability to edit with up to 4 reference images.

To UPDATE the nodes (if you already have them):

  1. open a command prompt in the "Rebels_HiDream_01_Image_dev_NODES" folder.

  2. run the command "git pull" to update the nodes.

then restart comfy and drag the workflow into comfy!

Youtube video tutorial to UPDATE:

https://youtu.be/iRo-S9oxGe8?si=tSBfVWDgEDqkeVfo




READ THE ENTIRE DESCRIPTION SO YOU DONT MISS A STEP.

node set to run the HiDream-01 Image Dev GGUF from smthem and the comfy-org bf16 model

Rebel HiDream-O1 Image Dev (bf16 and GGUF Nodes for ComfyUI)

Created by Rebel AI

This repository provides custom ComfyUI nodes to run HiDream-O1-Image-Dev GGUF models locally.

HiDream-O1 is a VAE-less, Pixel-Level Unified Transformer. Because it generates raw pixels token-by-token at massive resolutions, running it locally requires careful memory management. These nodes feature upfront dequantization (converting GGUF weights to native PyTorch tensors in system RAM during the load phase) to completely bypass NumPy single-threaded CPU bottlenecks during generation, allowing your GPU to run at maximum efficiency.

📦 Prerequisites

Before installing the custom nodes, you need the upstream model code and the weights.

Clone the Upstream HiDream-O1 Repo:

REQUIRED. NO EXCEPTIONS. WORKFLOW WILL NOT RUN.

The nodes rely on the official pipeline logic. Clone this anywhere on your local system:

git clone https://github.com/HiDream-ai/HiDream-O1-Image.git

Download the bf16 Model:

https://huggingface.co/Comfy-Org/HiDream-O1-Image/tree/main/checkpoints

Place the bf16 file in CHECKPOINTS: ComfyUI/models/checkpoints/

Download the GGUF Model:

https://huggingface.co/smthem/HiDream-O1-Image-Dev/blob/main/HiDream-O1-Image-Dev-Q6_K.gguf

Place the .gguf file in DIFFUSION_MODELS: ComfyUI/models/diffusion_models/

🛠️ Installation

Option 1: ComfyUI Windows Portable

Note: Ensure your ComfyUI portable installation is located on a strict local system drive (e.g., C:\ or D:\). Do not install or run these nodes from a OneDrive-synced folder, as it will cause virtual environment and pathing errors.

Open a command prompt and navigate to your portable custom nodes directory:

cd \ComfyUI_windows_portable\ComfyUI\custom_nodes

Clone this repository:

git clone https://github.com/RealRebelAI/Rebels_HiDream-01_Image_Dev_NODES

Install the requirements using the embedded Python environment:

cd Rebels_HiDream_01_Image_Dev_NODES

"..\..\..\python_embeded\python.exe" -m pip install -r requirements.txt

Option 2: Desktop / Standard Python Environment

Navigate to your ComfyUI custom nodes directory and lone this repository:

cd ComfyUI/custom_nodes

git clone https://github.com/RealRebelAI/Rebels_HiDream-01_Image_Dev_NODES

Activate your ComfyUI virtual environment and install the requirements:

cd Rebels_HiDream_01_Image_Dev_NODES

pip install -r requirements.txt

🧩 Node Documentation

Rebel HiDream-O1 Loader (gguf)

Loads the GGUF model and performs upfront dequantization.

gguf_name: Select your .gguf model from the diffusion_models folder.

tokenizer_path: Default is HiDream-ai/HiDream-O1-Image-Dev. It will automatically fetch the tokenizer config from Hugging Face.

upstream_repo_path: The absolute local path to where you cloned the HiDream-O1-Image repository in the prerequisites (e.g., C:\Users\name\HiDream-O1-Image).

device: Set to cuda for GPU acceleration.

offload:

aggressive: Heavily utilizes system RAM offloading (Recommended for 8GB VRAM cards like the RTX 3070).

balanced: Standard memory splitting.

minimal: Keeps most of the model in VRAM.

Note: The loader will hang for a moment at 100% while it unpacks the uint8 GGUF bytes into native PyTorch tensors in your system RAM. This is normal and prevents the CPU from bottlenecking your GPU during the actual generation steps.

Rebel HiDream-O1 Sampler

Connect the model output from the Loader here.

steps: 20-30 is the recommended sweet spot.

cfg: Keep at 0.0 Higher CFG combined with heavy styling tags can cause "deep-fried" or crushed-shadow artifacts due to the literal pixel-rendering nature of the model.

shift: keep at 1.0. Controls the timestep scheduling curve.

scheduler_name:

default: Standard sampling.

flash: Injects specific noise profiles. Note: The noise_scale_start, noise_scale_end, and noise_clip_std parameters only apply if the scheduler is set to flash.

⚠️ Known Limitations & Behaviors

Resolution Snapping: HiDream-O1 enforces strict token sequence lengths based on its pre-trained position embeddings. If you input a lower resolution (like 512x512 or 1024x1024), the upstream pipeline will automatically "snap" and force the generation to 2048x2048.

Compute Times: Because the model renders a raw 2048x2048 image pixel-by-pixel without a VAE, generation times will be significantly longer than standard latent models (like SDXL or Flux).

Prompting Style: Avoid stacking heavy texture tags (e.g., "8k resolution, ultra-detailed, gritty textures") unless you want a heavily illustrated look. For photorealism, use clean, simple photographic terms.

🧩 Node Documentation (BF16 / Safetensors)

Rebel HiDream-O1 Loader (Safetensors)

Loads the pure, uncompressed bfloat16 safetensors model directly, preserving 100% of the original mathematical precision to eliminate quantization artifacts.

  • model_name: Select your .safetensors model from the models/checkpoints/ folder.

  • tokenizer_path: Default is HiDream-ai/HiDream-O1-Image-Dev. It will automatically fetch the tokenizer config from Hugging Face.

  • upstream_repo_path: The absolute local path to where you cloned the HiDream-O1-Image repository in the prerequisites (e.g., C:\Users\name\HiDream-O1-Image).

  • device: Set to cuda for GPU acceleration.

  • offload:

    • aggressive: Heavily utilizes system RAM offloading (Recommended for 8GB VRAM cards like the RTX 3070).

    • balanced: Standard memory splitting.

    • minimal: Keeps most of the model in VRAM.

Note: The loader will take a moment to memory-map the massive 16.4GB bfloat16 file and balance it across your system RAM and VRAM. Because this workflow uses the native uncompressed weights instead of heavily quantized blocks, the model is much more sensitive to guidance.

Rebel HiDream-O1 Sampler

Connect the model output from the Loader here.

  • steps: 20-30 is the recommended sweet spot.

  • cfg: Keep at 0.0 Higher CFG combined with heavy styling tags can cause "deep-fried" or crushed-shadow artifacts due to the literal pixel-rendering nature of the model.

  • shift: Keep at 1.0. Controls the timestep scheduling curve, delaying fine-detail rendering to accommodate the massive canvas size.

  • scheduler_name:

    • default: Standard sampling.

    • flash: Injects specific noise profiles. Note: The noise_scale_start, noise_scale_end, and noise_clip_std parameters only apply if the scheduler is set to flash.


⚠️ Known Limitations & Behaviors

  • Resolution Snapping: HiDream-O1 enforces strict token sequence lengths based on its pre-trained position embeddings. If you input a lower resolution (like 512x512 or 1024x1024), the upstream pipeline will automatically "snap" and force the generation to 2048x2048.

  • Compute Times: Because the model renders a raw 2048x2048 image pixel-by-pixel without a VAE, generation times will be significantly longer than standard latent models (like SDXL or Flux).

  • Prompting Style: Avoid stacking heavy texture tags (e.g., "8k resolution, ultra-detailed, gritty textures") unless you want a heavily illustrated look. For photorealism, use clean, simple photographic terms.