Yet Another Workflow : easy t2v + i2v
I've aimed at a user-friendly UI for ComfyUI. There's a balance between complexity and ease of use, and this workflow aims to give you useful controls with clear guidance on what you need to care about. I hope these will be helpful to anyone strugging with quality and the general UI-isms of ComfyUI. I've taken the time to color code and add lots of notes. Please read the notes, I've tried to make them useful!
This is the workflow I use, it's not aimed at a skill level. It's designed to be easy to use and adjust with some UI concessions and labeling to ensure you can pilot it with less experience in a way that is more sophisticated than the official example workflows, which can be easy to break.
The primary goal with this workflow is to give you a strong foundational place to generate either text to video (T2V) or image to video (I2V) outputs without having to fuss too much.
The green controls are the stuff you generally want to mess with.
The secondary goal here is to provide a consistent interface to interact with different samplers and now different models.
Now for LTX-2.3
One of the strengths of respecting a consistency to a UI is that it allows you to change things under the hood to enable people to explore new techniques without changing the majority of their experience. Whether you're familiar with my Wan 2.2 workflow or whether this is your first experience with my workflow, once you're comfortable with one of them, the goal is that you can switch between them with minimal fuss.
LTX-2.3 currently represents the future of open weights video. Things are always in flux (pun intended), but until Wan announces a new open weights release, LTX-2.3 is the actively developing model. It represents both steps forward and backwards. It introduces a number of long desired features:
Native support for sound (both prompted and from files)
Long video generation
Better support for lower end hardware (kinda)
Single model process presents a visually simpler process for I2V and T2V
As before there are lots of notes and color coding to help your orient yourself. If you're new to this workflow, but not to YAW in general, I've added a bunch of usage notes that specificaly relate to LTX-2.3. This model is still pretty new, and I'll expand on this as things develop.
The LTX-2.3 release is notable becaue it made some major improvements to audio, prompt adherance, and LoRA training.
I have a lot of criticisms of LTX-2.3, and the RunPod guide will cover more of the details, but my major complaint is prompt adherance and verbosity. It wants very verbose prompts and is bad at infering what you want. On top of that, the way they achieve prerformance improvements makes prompt adherance markedly poorer. You'll end up doing way more gens with LTX-2.3 than something like Wan 2.2. We're watching this space to see how things improve, but it is worth exploring! As I noted, it is currently the future of open weights models.
I've included two different sampling techniques in the initial release to explore. The two stage version is the default, with a more expensive alternative option that pushes on quality with a different look.
The default workflow setup utilives the full dev + distilled LoRA for maximum quality, as recommended by Lightricks. It's easy to modify for different versions to meet your needs.
What's new?
While this is a different tech stack, it's also the first release of v0.39 of YAW (which will soon come to the Wan versions).
I've created a custom node to make image resolution less fiddly. When you do I2V, it will now lock to the image aspect ratio. It will also suggest some different resolution quality targets. You can still override as before, but I have found this quite useful after testing extensively. It removes one of the few points in friction when experimenting with different images for image-to-video work.
I removed the dependency on one of the custom node packs that was causing conflicts during installation for folks, just to remove another point of friction. This relates to the node used for handling multiple inputs.
Like it?
Give it a like! Tag it as a Resource when you use it! Support on Patreon or a tip on Ko-fi are also welcome. Yellow Buzz will go towards promoting awareness here on Civit.
Need help?
I like helping people get going with this stuff, so if you want help message me. If you want extended one-on-one help, there's an option on the Patreon. I'm happy to walk you through the details, answer your questions, and give you some extra tips and tricks, and scripts. I've done this for a few folks, I'll save you money and headaches.
I've also written an article here on getting it going with my Runpod template. The template will vastly expedite and simplify getting things up and running.
General Advice
Make lots of videos! Post your videos! Don't fuss with the tech! Be smart about how you spend your time with this stuff. It's easy to burn out if you spend more time trying to get things to work than making videos you like. That's really why I'm posting this.
Use RunPod. Use the RTX 5090 or the H100 SXM. Use my Wan 2.2 template. If you've not used RunPod before, sign up with my link; we'll both get some free credit. See the article for more.
If you use a service like RunPod, if you're doing I2V, it can be smart to have your images ready in advance to make sure the server stays busy while you are using it.
If you run this outside of Runpod, you'll need to install some custom nodes. To do that, click the "Manager" button at the top of the Comfy interface, and then click the "Install Missing Custom Nodes". Click "Install" on each one - I recommend in order; you'll need to wait till each has installed. Do not bother restarting ComfyUI until they are all installed. The RunPod template has them preinstalled. (There's a manual patch for the LTXFilmGrain node here.)
If the wires bother you, there's a button in the bottom right on the floating UI that will hide them.
This workflow is setup for .safetensors models, but you can use GGUF if you want to make the changes node changes.
Costs?
I'm in the process of working on updating my benchmarking for LTX-2.3, so watch this space.
Troubleshooting
If a node is missing (bright thick red outline with a warning when you open the workflow), you can install them by going to Manager > Install Missing Custom Nodes, and pressing Install on any the nodes that show up there.
If you are getting any errors related to a custom node, it's possible something has changed recently in the software. It might be useful to change a version back to the last "stable" build in these situations.
For example, the nightly build of WanVideoWrapper might introduce an error that wasn't there last time. With a workflow open, you can go to Manager > Custom Nodes in Workflow. This will show you all of the custom nodes. If you click, Switch Ver, you can see all of the releases. Consider trying the first numbered on at the top of the list.
If that doesn't work, or there seem to be more significant problems and you are using RunPod, you may have forgotten to select CUDA 12.8. Try restarting the server. If that doesn't work, terminate the pod, and make a new one. This will fix a surprising number of possible issues.
Longer video generation support?
LTX-2.3 supports long video generation (memory limits apply). The longer the video, the more wonky the prompt adherance becomes (more unexpected physics, garbled dialog, etc). I've set a maximum on the length slider, but you can increase it if you want to go nuts. I've done over a minute so far. The results are not great, but with enough memory you can do it.
Sound
LTX-2.3 has decent sound support, but it's a double edged sword. Sometimes the motion will be great and the sound gets weird. Sometimes the sound is great, but the character does something strange or an extra hand appears or some such. There's no good way to fix this. I want to call out that there's not a great way way to get a consistent voice across prompts.

