Sign In

Wan 2.2 I2V: HD/FHD resolution, but much faster

Updated: Dec 1, 2025

toolhdwanfhdi2vwan2.2

Type

Workflows

Stats

790

0

Reviews

Published

Nov 19, 2025

Base Model

Wan Video 2.2 I2V-A14B

Hash

AutoV2
D12E2A5F2F
default creator card background decoration
qdr1en's Avatar

qdr1en

v2

The workflow allows ultra-fast iterations over different seeds/prompts/images/settings before rendering the final video at full resolution, while removing the need to constantly swap between models (which is ideal for low-RAM setups).

How it works

There are 2 stages:

  1. Generate a draft version of your video at low resolution, using the high-noise model only.

  2. Upscale the video latent to render it at your target resolution, using the low-noise model only.

Process

Stage 1

  • Disable the groups "LN" and "Post".

  • Upload your image.

  • Write your prompt.

  • Set the low-resolution width (typically, 256, 288 or 320px, depending on the image aspect ratio).

  • Click "Run".

Stage 2

  • Drop the video draft created at Stage 1 into the ComfyUI canvas.

  • Disable the group "HN"; enable the groups "LN" and "Post".

  • In the "Load Draft" node, select the same video again.

  • Set the high-resolution width (typically, x2 or x2.5 the width set at stage 1).

  • Click "Run".

⚠️Important constraint: make sure the video width and height are multiples of 16, at both stages of the process, or the final output will be blurred.

The workflow was intentionnally made as simple as possible to maximize compatibility. Up to you to adapt it further to your needs.

For a simpler but less efficient version, see v1 below.

v1

The workflow significantly mitigates the "speed VS quality" dilemma, allowing users with low-end hardware to generate videos in HD resolution nearly twice as fast!

How it works

The principle is dumb simple: we run the high-noise model at very low resolution, then upscale the latent before injecting it into the low-noise sampler.

Since the original image is also reinjected, with a new Wan wrapper node at the low-noise sampling step, visual details are preserved.

Limitations

  • The motion does lose a little in subtlety, but the speed gain is totally worth it in most of cases.

  • Not tested on T2V, probably won't work. Works on T2V too. Simply replace both I2V nodes by a EmptyHunyuanLatentVideo node. Thanks to @axicec for sorting this out.

Getting Started

  1. Replace the models by yours, or follow the links below to download them.

  2. Install the required custom nodes listed below if they are missing from your installation.

  3. Load an image and write your prompt.

  4. Click Run.

Speed benchmark

  • settings: 65 frames, using Q5_K_M I2V models, 4 steps on high-noise with lightx2v 1030, 4 steps on low-noise with lightx2v 1022 and Fun HPS2.1 loras, euler/beta sampler/scheduler on both samplers.

  • hardware: RTX 3060 with 12GB VRAM and 32GB RAM.

768*1152px (2:3)

  • 768*1152, no upscale: 20′46″

  • 256*384 x2 then x1.5: 11′16″ (-46%)

  • 256*384 x1.5 then x2: 10′48″ (-48%)

720*1280px (9:16)

  • 720*1280, no upscale : 23′19″

  • 288*512 x2.5 : 15′57″ (-32%)

  • 288*512 x2 then x1.25: 11′57″ (-49%) <- this is the showcased video

Target resolution VS Hardware Requirements

  • HD : >= 12GB VRAM v

  • FHD: >= 16GB VRAM ? (not tested, feedbacks appreciated)

Initial sampling resolutions and step settings recommendations are included in the workflow.

Custom Nodes

Required

Optional

Models Used

Wan 2.2 14B I2V, Quantized:

Lightx2v LoRas :

Fun LoRa:

Edit (v1): the last sampler's scheduler is set on linear_quadratic by default but it should be beta.