Type | Workflows |
Stats | 694 0 |
Reviews | (42) |
Published | Feb 16, 2025 |
Base Model | |
Hash | AutoV2 966FD33262 |
⋆.°🌸 Some considerations 🌸˚˖⋆
In addition to the nodes that can be installed with ComfyUI_Manager, you need Kijai nodes to run this workflow.
.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.
Using the hunyuan t2v 720p GGUF Q4_K_M, the average GGUF, and 73 frames gen:
For 12GB VRAM, you can upscale 240x320 by aprox. 1.3x 🫤
For 16GB VRAM, you can upscale 240x320 by aprox. 2.5x 😏
For 24GB VRAM, you can upscale 240x320 by aprox. 3.5x 😎 ~14min @ RTX3090
.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.
About input videos that aren't generated by Hunyuan:
Use the arrows in the frame load cap to set it up automatically following Hunyuan's frames rule (4*x+1).
.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.
Prompting
That's right, it's manual for now. I'm thinking of creating an automatic version using Joytag or Clip vision from OpenAi.
.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.・。゚.
JK CHSTR 2025 - Hunyuan by Tencent