Updated: Feb 11, 2026
conceptWAN 2.2 GGUF — FLF Loop & Video Workflow (8GB VRAM)
This workflow is a complete WAN 2.2 video generation pipeline designed to run on consumer GPUs with 8GB VRAM.
It supports seamless looping, normal video generation and image/video reinterpretation inside one graph.
The system is based on a start-frame / end-frame method.
You can use it in three ways:
• Seamless FLF loop (same start & end image)
• Normal video (different start & end image)
• Animate / restyle / extend existing footage
Automatic prompting using Qwen is included but optional.
If you don’t want AI prompts, just bypass the Qwen node and write a prompt manually.
The workflow also contains:
frame interpolation
temporal color matching
optional upscaling
negative prompt handling (NAG compatible)
The Radial Attention Patch is included ONLY to increase speed. ( 6 minutes for 5 seconds 720p)
It is optional and the workflow works without it.
MODEL INSTALLATION (VERY IMPORTANT)
WAN GGUF models DO NOT go into the normal UNet folder.
Place them here:
ComfyUI/models/diffusers/
WAN 2.2 works using a High-Noise + Low-Noise expert pair system (Mixture-of-Experts video diffusion). (docs.comfy.org)
Both models are required or motion generation will break.
Recommended model pair used in this workflow:
SmoothMix WAN 2.2 I2V GGUF
https://huggingface.co/Bedovyy/smoothMixWan22-I2V-GGUF
Alternative repository:
https://huggingface.co/BigDannyPt/WAN-2.2-SmoothMix-GGUF
VAE:
wan_2.1_vae.safetensors
Place into:
ComfyUI/models/vae/
TEXT ENCODER:
umt5_xxl_fp8_e4m3fn_scaled.safetensors
or NSFW version
Place into:
ComfyUI/models/text_encoders/
QWEN AUTO PROMPT (OPTIONAL)
This workflow can automatically generate prompts using Qwen-VL.
Qwen ComfyUI nodes:
https://github.com/1038lab/ComfyUI-QwenVL (GitHub)
or custom node with wan mode https://github.com/huchukato/ComfyUI-QwenVL-Mod
Download a supported Qwen model (example):
https://huggingface.co/Qwen/Qwen2.5-VL
Place model into:
ComfyUI/models/LLM/
Qwen is a vision-language model capable of understanding images and generating descriptive prompts. (GitHub)
If not installed:
→ simply bypass the Qwen node.
PAINTER / INPAINT HELPER
The workflow uses a painter/inpaint helper for temporal consistency between frames.
https://github.com/princepainter/ComfyUI-PainterI2VforKJ
Install in Comfy/custom nodes
These node packs include mask and enhancement helpers used for automatic detail correction and repainting. (ltdrdata.github.io)
REQUIRED NODES
Install with ComfyUI Manager:
• VideoHelperSuite
• KJNodes
• WAS Node Suite
• RGThree Nodes
• Custom Scripts (pysssss)
• ComfyUI GGUF Loader
OPTIONAL SPEED BOOST (RADIAL ATTENTION PATCH)
The radial attention patch accelerates video diffusion.
It is OPTIONAL.
If you do not want additional dependencies:
→ disable/bypass the node and the workflow still runs normally.
Step 1 — Install Triton for Windows
WAN 2.2 + RadialAttn requires Triton.
Download the correct Windows wheel here:
https://github.com/woct0rdho/triton-windows/releases
Install inside your venv.
Step 2 — Install SparseSageAttn
Download the Windows wheel from:
https://github.com/woct0rdho/SpargeAttn/releases
Install inside your venv.
Step 3 — Install RadialAttn Node
Download from:
https://github.com/woct0rdho/ComfyUI-RadialAttn
Place inside your ComfyUI custom_nodes folder.
Step 4 — Restart ComfyUI
If Sparse / Radial Attn loads correctly, startup log will say:
“Using sparse_sage_attn as block_sparse_sage2_attn_cuda”
Then the patch is active.
USAGE
Sampler: Euler simple
CFG: 1
Steps: smooth mix 3+3
Loop:
Use identical start and end frame. Enable frame cut nodes and connect it to the next group input.
Video:
Use different images. Disable frame cut nodes and connect the output to the group next to framecut.
WAN prefers scene descriptions and camera movement prompts.
Example for timebased promt:
(At 0 seconds: The camera is positioned in the neon-lit city at the base of the massive central tower; the vertical blue energy beam rises straight upward through the frame, surrounded by dense futuristic architecture glowing in purple, blue, and warm orange light. No text, symbols, or logos exist anywhere in the scene.)
(At 1 seconds: The camera begins a smooth, continuous upward movement; street-level details slide downward and out of view while the beam remains perfectly centered. Traffic lanes are active, small flying vehicles move calmly, windows flicker, and holographic panels animate softly.)
(At 2 seconds: Mid-city levels pass by slowly; illuminated windows, platforms, holographic advertisements and distant air traffic drift beneath the camera as atmospheric haze and particles float through the air. The city clearly feels alive but calm.)
(At 3 seconds: The camera ascends above the streets and crowds; the tower now dominates the frame and the energy beam grows brighter. Soft pulses of light travel upward inside the beam and the surrounding buildings subtly react with changing reflections.)
(At 4 seconds: Near the upper tower section, the beam intensifies and begins emitting faint geometric light patterns into the sky above the city. The patterns slowly organize and form a large transparent holographic structure suspended in the air.)
(At 5 seconds: The holographic projection stabilizes into a bright floating logo created by the energy beam itself. The logo is made of luminous particles and soft light, perfectly centered above the city. The camera gently stops and holds; the city continues subtle motion — drifting mist, distant vehicles and softly flickering lights — forming a seamless looping end frame.)
Avoid portrait prompts.
NOTES
This workflow is optimized for:
temporal stability
long sequences
low VRAM usage
reduced flicker
consistent motion
You can freely modify and extend it.
