Sign In

WAN 2.2 GGUF/start end frame/loop/t2v/i2v 8gb/ advanced

Updated: Feb 11, 2026

concept

Type

Workflows

Stats

181

0

Reviews

Published

Feb 11, 2026

Base Model

Wan Video 14B i2v 720p

Hash

AutoV2
69A8BFC46A
default creator card background decoration
Doomspace's Avatar

Doomspace

WAN 2.2 GGUF — FLF Loop & Video Workflow (8GB VRAM)

This workflow is a complete WAN 2.2 video generation pipeline designed to run on consumer GPUs with 8GB VRAM.
It supports seamless looping, normal video generation and image/video reinterpretation inside one graph.

The system is based on a start-frame / end-frame method.

You can use it in three ways:

• Seamless FLF loop (same start & end image)
• Normal video (different start & end image)
• Animate / restyle / extend existing footage

Automatic prompting using Qwen is included but optional.
If you don’t want AI prompts, just bypass the Qwen node and write a prompt manually.

The workflow also contains:

  • frame interpolation

  • temporal color matching

  • optional upscaling

  • negative prompt handling (NAG compatible)

The Radial Attention Patch is included ONLY to increase speed. ( 6 minutes for 5 seconds 720p)
It is optional and the workflow works without it.


MODEL INSTALLATION (VERY IMPORTANT)

WAN GGUF models DO NOT go into the normal UNet folder.

Place them here:
ComfyUI/models/diffusers/

WAN 2.2 works using a High-Noise + Low-Noise expert pair system (Mixture-of-Experts video diffusion). (docs.comfy.org)
Both models are required or motion generation will break.

Recommended model pair used in this workflow:

SmoothMix WAN 2.2 I2V GGUF
https://huggingface.co/Bedovyy/smoothMixWan22-I2V-GGUF

Alternative repository:
https://huggingface.co/BigDannyPt/WAN-2.2-SmoothMix-GGUF


VAE:
wan_2.1_vae.safetensors
Place into:
ComfyUI/models/vae/


TEXT ENCODER:
umt5_xxl_fp8_e4m3fn_scaled.safetensors

or NSFW version

https://huggingface.co/zootkitty/nsfw_wan_umt5-xxl_bf16_fixed/resolve/main/nsfw_wan_umt5-xxl_bf16_fixed.safetensors?download=true


Place into:
ComfyUI/models/text_encoders/


QWEN AUTO PROMPT (OPTIONAL)

This workflow can automatically generate prompts using Qwen-VL.

Qwen ComfyUI nodes:
https://github.com/1038lab/ComfyUI-QwenVL (GitHub)

or custom node with wan mode https://github.com/huchukato/ComfyUI-QwenVL-Mod

Download a supported Qwen model (example):
https://huggingface.co/Qwen/Qwen2.5-VL

Place model into:
ComfyUI/models/LLM/

Qwen is a vision-language model capable of understanding images and generating descriptive prompts. (GitHub)

If not installed:
→ simply bypass the Qwen node.


PAINTER / INPAINT HELPER

The workflow uses a painter/inpaint helper for temporal consistency between frames.

https://github.com/princepainter/ComfyUI-PainterI2VforKJ

Install in Comfy/custom nodes

These node packs include mask and enhancement helpers used for automatic detail correction and repainting. (ltdrdata.github.io)


REQUIRED NODES

Install with ComfyUI Manager:

• VideoHelperSuite
• KJNodes
• WAS Node Suite
• RGThree Nodes
• Custom Scripts (pysssss)
• ComfyUI GGUF Loader


OPTIONAL SPEED BOOST (RADIAL ATTENTION PATCH)

The radial attention patch accelerates video diffusion.

It is OPTIONAL.

If you do not want additional dependencies:
→ disable/bypass the node and the workflow still runs normally.

Step 1 — Install Triton for Windows

WAN 2.2 + RadialAttn requires Triton.
Download the correct Windows wheel here:
https://github.com/woct0rdho/triton-windows/releases

Install inside your venv.

Step 2 — Install SparseSageAttn

Download the Windows wheel from:
https://github.com/woct0rdho/SpargeAttn/releases

Install inside your venv.

Step 3 — Install RadialAttn Node

Download from:
https://github.com/woct0rdho/ComfyUI-RadialAttn

Place inside your ComfyUI custom_nodes folder.

Step 4 — Restart ComfyUI

If Sparse / Radial Attn loads correctly, startup log will say:
“Using sparse_sage_attn as block_sparse_sage2_attn_cuda”

Then the patch is active.


USAGE

Sampler: Euler simple
CFG: 1
Steps: smooth mix 3+3

Loop:
Use identical start and end frame. Enable frame cut nodes and connect it to the next group input.

Video:
Use different images. Disable frame cut nodes and connect the output to the group next to framecut.

WAN prefers scene descriptions and camera movement prompts.

Example for timebased promt:
(At 0 seconds: The camera is positioned in the neon-lit city at the base of the massive central tower; the vertical blue energy beam rises straight upward through the frame, surrounded by dense futuristic architecture glowing in purple, blue, and warm orange light. No text, symbols, or logos exist anywhere in the scene.)

(At 1 seconds: The camera begins a smooth, continuous upward movement; street-level details slide downward and out of view while the beam remains perfectly centered. Traffic lanes are active, small flying vehicles move calmly, windows flicker, and holographic panels animate softly.)

(At 2 seconds: Mid-city levels pass by slowly; illuminated windows, platforms, holographic advertisements and distant air traffic drift beneath the camera as atmospheric haze and particles float through the air. The city clearly feels alive but calm.)

(At 3 seconds: The camera ascends above the streets and crowds; the tower now dominates the frame and the energy beam grows brighter. Soft pulses of light travel upward inside the beam and the surrounding buildings subtly react with changing reflections.)

(At 4 seconds: Near the upper tower section, the beam intensifies and begins emitting faint geometric light patterns into the sky above the city. The patterns slowly organize and form a large transparent holographic structure suspended in the air.)

(At 5 seconds: The holographic projection stabilizes into a bright floating logo created by the energy beam itself. The logo is made of luminous particles and soft light, perfectly centered above the city. The camera gently stops and holds; the city continues subtle motion — drifting mist, distant vehicles and softly flickering lights — forming a seamless looping end frame.)

Avoid portrait prompts.


NOTES

This workflow is optimized for:

  • temporal stability

  • long sequences

  • low VRAM usage

  • reduced flicker

  • consistent motion

You can freely modify and extend it.