Welcome to my 💫🎦 Friendly LTX-2.3 T2V+I2V+Lipsync
✨ Less mess, more magic
Unified Light - Lipsync all-in one version with single-stage video generation
Unified is cooming soon...
LTX-2.3 is a state-of-the-art DiT-based (Diffusion Transformer) audio-video foundation model developed by Lightricks. It represents a significant evolution from LTX-2, delivering enhanced audio and visual quality alongside improved prompt adherence. The model is specifically designed to generate synchronized video and audio within a single unified architecture, making it a powerful tool for multimodal content creation.
I offer my unique workflow with convenient options control and all-in-one structure (audio, t2v, i2v)
💻 System requirements:
Minimum system requirements for 540p i2v and 720p t2v:
RTX 3000-s, 8GB+ VRAM, 45GB+ RAM, 8-core processor, SSD, latest ComfyUI
🚀 Low VRAM optional optimization:
For systems with low VRAM use --reserve-vram ComfyUI parameter in run_nvidia_gpu.bat:
--reserve-vram 4(or other number in GB).
📌 Detailed tips and links to models in the workflow
✨ Workflow features:
Extremely user-friendly interface
Maximum performance and optimization from 8GB of VRAM: GGUF or 8-step distilled model with fp4 or fp8 text encoder
All-in-one: i2v, t2v, and interpolation
Convenient one-click mode switching
Prompt enhancer and sampler previews
Generation time setting in seconds
Lora support (up to 3)
Detailed tips and links to all necessary models
Manual random seed for complete control over generations
🤗🙏🏼 Thanks to Lightricks Team
Original repo — GitHub

