Sign In

Fast WAN I2V Compact

55

1.2k

24

Type

Workflows

Stats

419

0

Reviews

Published

Jul 29, 2025

Base Model

Wan Video 14B i2v 480p

Hash

AutoV2
3C81A1AF82
default creator card background decoration
plk's Avatar

plk

Designed to be visually compact and simplified for ease of use. Personally, I think this is the best streamlined workflow there is. You can save notes of prompts and LoRA triggers right beside the prompt input, making it quick and easy to swap between them. The overall layout is designed to waste as little space as possible, while fitting very well into the ComfyUI workflow window at a 16:9 ratio. It's designed so you don't have to constantly rescale or move the workflow around in order to change settings. If you "fit to view" and click the zoom in button 2 or 3 times, it will fit perfectly with little wasted space.

_________

This workflow generates a 5 second 512x512 video in 90 seconds on a 4070ti with the Q8 GGUF model without Sage Attention enabled.

This workflow does not use tricks like upscaling and uses mostly basic nodes and extensions, it should be very easy to get working with minimal effort.

This workflow uses LCM sampling with the Light X2V LoRA to speedup generation time. Additional LoRAs can be used at the same time.

_________

For WAN 2.2:

Same design as previously, but geared toward running the WAN 2.2 Low Noise model only. See "required models" section below for new workflow requirements.

The Light X2V LoRA works with WAN 2.2 at a strength of 1.1 to 2.0, and can dramatically alter the behavior of the model, in either beneficial or detrimental ways, so after testing I chose 1.5 as the default strength since that seemed to be the most reliable, but experiment to find what works for you.

WAN 2.2 is much more dynamic, which means it requires a slightly different prompting style than you might have used in WAN 2.1. The same goes for its affect on LoRA's, where they tend to be amplified in strength, which can be a good and bad thing, but overall I'm seeing some pretty good results with lots of keepers. So the main differences to get good results is to focus on learning how to prompt it and you may also need to tinker with LoRA strengths, depending on the LoRA and how it's behaving with your prompt and image input. Even changing it to 6 or 8 steps can also potentially improve results.

The workflow sampler/scheduler settings seem to work pretty well, but more experimentation is required, there could be other combinations that work better, especially in the RES4LYF custom samplers and schedulers extension (part of the requirements below).

There can be some bad generations that go off the rails, but all in all once you dial things in, WAN 2.2 can generate a lot of keepers that you could never get with WAN 2.1.

_________

For WAN 2.1:

The main settings that you may want to change would be primarily just output resolution or sampler steps. Other samplers or schedulers may work, but I find LCM/Simple provides the most coherent output. The only other setting you might want to fiddle with is the LoRA strengths. There are however other settings you can fiddle with, such as "SHIFT", which can somewhat work like a CFG setting. In my experience, it can be used to drastically change how a prompt/LoRA is expressed, while also creating more dramatic changes in movements, but generally this should be left at its default setting.

_________

Note: Sage Attention is disabled by default. To enable Sage Attention (if you have the pre-requisites installed) simply select the "Enable for Sage Attention" node and press Ctrl+B to enable it, then below it change the "sage_attention" option from disabled to enabled. Even if you don't plan on using Sage Attention, you will still need to install the extension for the workflow to operate.

_________

Required and alternative models:

GGUF WAN 2.2 i2v models (use only "low noise" version):

https://huggingface.co/bullerwins/Wan2.2-I2V-A14B-GGUF/tree/main

GGUF WAN 2.1 i2v models:

https://huggingface.co/city96/Wan2.1-I2V-14B-480P-gguf/tree/main

CLIP model:

https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/blob/main/split_files/text_encoders/umt5_xxl_fp8_e4m3fn_scaled.safetensors

Or the higher precision BF16 CLIP model:

https://huggingface.co/minaiosu/Felldude/blob/main/wan21UMT5XxlFP32_bf16.safetensors

CLIP Vision model:

https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/blob/main/split_files/clip_vision/clip_vision_h.safetensors

Or the custom NSFW geared CLIP Vison model (recommended):

https://civitai.com/models/1802070/wan-21-nsfw-clip-vision-h

VAE model:

https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/blob/main/split_files/vae/wan_2.1_vae.safetensors

Light X2V T2V LoRA: https://huggingface.co/Kijai/WanVideo_comfy/blob/main/Wan21_T2V_14B_lightx2v_cfg_step_distill_lora_rank32.safetensors

Or the new proper Light X2V I2V LoRA (recommended):

https://huggingface.co/lightx2v/Wan2.1-I2V-14B-480P-StepDistill-CfgDistill-Lightx2v/blob/main/loras/Wan21_I2V_14B_lightx2v_cfg_step_distill_lora_rank64.safetensors

Or the other Light X2V experimentations by Kijai:

https://huggingface.co/Kijai/WanVideo_comfy/tree/main/Lightx2v

RES4LYF custom samplers and schedulers:

https://github.com/ClownsharkBatwing/RES4LYF

_________

Secret Pro Tip: using a transparent or solid colored image, such as black, can turn the i2v model into essentially a t2v model. It will rapidly transition from the blank input image and generate something from scratch to try to follow your prompt. It's an easy way to get t2v capabilities without changing workflows/models.

_________

Other useful information:

WAN can change behavior dramatically with output resolution changes, it tends to respond best if the resolution is 480 in either width or height. WAN 2.2 is supposed to be a 480p and 720p model, but it may still behave differently at different resolutions, and require either settings tweaks or just not work well at certain resolutions. Some things can work well at 480x480, some things can work better or worse at 512x512 or higher resolutions, but typically you get the most stable outputs with 480 or 720 in the width or height.