Since Kijai seems too busy to include a Text to Image template, I've decided to share my workflow so that users can get up and running using WAN 2.2 text to image.
v1.1 adds advanced dual lora setup, TeaCache and Enhance-A-Video with explanations.
v1.1a LightX2V proof of concept, crazy fast WAN 2.2 renders using only 5 steps and CFG 1
v1.1b is an iteration of 1.1a where I've added an A/B video comparison function and took the opportunity to use the first WAN 2.2 lora available for demonstration.
v1.1c T2V2V Proof-Of-Concept: This version creates a monster !! WARNING !! Don't even attempt this unless you can spare 144GB virtual memory (system ram + swap file) . It will juggle 4 WAN 2.2 models (T2V + I2V) in one connected pipeline. It will get the last frame of T2V and continue I2V generation with that and join two videos seamlessly. In case you are using fp8 scaled models like I do; this will take a massive 130-ish GB virtual memory to swap things around. I have 64GB RAM, 32GB VRAM and I had to raise my page file in windows to 80GB for this to complete. Once you have all that though, it actually takes surprisingly little time to complete. So, MAKE SURE your system ram + page file is at least 144GB (if you are using quantized GGUF versions this will obviously be much smaller). In the upcoming v1.2 major upgrade, this workflow will evolve into a complete production suite, think of this as a whistleblower for that ;)
v1.1d SDXL to Video Pipeline π₯π₯π₯ Well, basically your dreams come true π I'll take your buzz for this one though π haha, just kidding. Whatever I do, I do for love, not for fame π Enjoy !
Please read the "read me" sections in the workflow.
N-Joy !!