Sign In

LTX-2.3 DEV/DIST - IMAGE to Video and TEXT to Video with Ollama

Updated: Mar 6, 2026

toolvideoaudioollamai2vt2v

Type

Workflows

Stats

439

0

Reviews

Published

Jan 18, 2026

Base Model

LTXV2

Hash

AutoV2
135984D864
Howling Aurora
tremolo28's Avatar

tremolo28

V2.3 LTX-2.3 DEV & Distilled Video with Audio

Image to Video and a Text to Video workflow, both can use own Prompts or Ollama generated/enhanced prompts.

  • works with latest LTX 2.3 Distilled model (8 steps, CFG=1) or Dev model (20 steps, CFG=3.5)

Downloads:


V1.5 LTX-2 DEV Video with Audio including latest 🅛🅣🅧 Multimodal Guider

Image to Video and a Text to Video workflow, both can use own Prompts or Ollama generated/enhanced prompts.

Replaced the Guider node with latest Multimodal Guider node, see more details in WF notes or here: https://ltx.io/model/model-blog/ltx-2-better-control-for-real-workflows Before we had 1 CFG parameter for audio and video. With multimodal guider, we now can tweak audio and video seperately with even more parameters...


V1.0 LTX-2 DEV Video with Audio:

Image to Video and a Text to Video workflow with own Prompts or Ollama generated/enhanced prompts.

  • setup for the LTX2 Dev model.

  • uses Detailer Lora for better quality and LTX tiled VAE to avoid OOM and visual grids

  • 2 pass rendering (motion+upscale). Upscale process uses distilled and spatial upscale Lora

  • setup with latest LTXVNormalizingSampler to increase video & audio quality.

  • Text to Video can use dynamic prompts with wildcards.


I am using these starting parameters for ComfyUi to avoid OOM (my setup: 16g Vram/64g Ram) :

--lowvram --cache-none --reserve-vram 6 --preview-method none

=> OBSOLETE with latest Comfy updates for better memory management:


Download Files: (Workflow V1.0 an V1.5)

Find Model/Lora Loader nodes within Sampler Subgraph node.

- LTX2 Dev Model (dev_Fp8): https://huggingface.co/Lightricks/LTX-2/tree/main

- Detailer Lora: https://huggingface.co/Lightricks/LTX-2-19b-IC-LoRA-Detailer/tree/main

- Distilled (lora-384) & Spatial upscaler Lora: https://huggingface.co/Lightricks/LTX-2/tree/main

- VAE (already included in above dev_FP8 model, but needed if you go for GGUF models): https://huggingface.co/Lightricks/LTX-2/tree/main/vae

- Textencoder (fp8_e4m3fn): https://huggingface.co/GitMylo/LTX-2-comfy_gemma_fp8_e4m3fn/tree/main

- Image to Video Adapter Lora (more motion with I2V): https://huggingface.co/MachineDelusions/LTX-2_Image2Video_Adapter_LoRa/tree/main

- Ollama Models:

Save Location:

  • 📂 ComfyUI/

  • ├── 📂 models/

  • │ ├── 📂 checkpoints/

  • │ │ ├── ltx-2-19b-dev-fp8.safetensors

  • │ ├── 📂 text_encoders/

  • │ │ └── gemma_3_12B_it_fp8_e4m3fn.safetensors

  • │ ├── 📂 loras/

  • │ │ ├── ltx-2-19b-distilled-lora-384.safetensors

  • │ └── 📂 latent_upscale_models/

  • │ └── ltx-2-spatial-upscaler-x2-1.0.safetensors

  • │ └── 📂 Clip/

  • │ └── ltx-2.3_text_projection_bf16.safetensors


Custom Nodes used: