Sign In

Stabled Diffusion : XL Videos with animateDiff and A1111

14

Objective

Generate videos with XL checkpoint and animatediff.

Software

Please check my previous articles for installation :

Checkpoint

AnimateDiff

Download the model mm_sdxl_v10_beta.ckpt from https://huggingface.co/guoyww/animatediff/tree/main

Copy the file under:

 <StableDiffusion folder>\stable-diffusion-webui\extensions\sd-webui-animatediff\model

In the same folder you should have also the models for 1.5 models.

TESTS

My idea is to generate the video at resolution 1024x1024 with the same prompt with 3 different checkpoints.

Prompt

(CINEMATIC SHOT:1.4)(skinny, slim, long legs, abs:1.4) (bikini blue:1.3)(background beach) (walking:1.4)(make-up, eyeliner, eyeshadow, lipstick:1.4) smiling (big breasts,:1.4) (skinny:1.3)(abs:1.3)

Negative

(CLEAVAGE:1.5)(NSFW:1.5)(deformed, distorted, disfigured:1.3), poorly drawn, bad anatomy, wrong anatomy, extra limb, missing limb, floating limbs, (mutated hands and fingers:1.4), disconnected limbs, mutation, mutated, ugly, disgusting, blurry, amputation, (piercing:1.3) (hands:1.35)

Other settings

Steps: 20

Sampler: Euler a

CFG scale: 8

Seed: 4220834260

Face restoration: GFPGAN

Size: 1024x1024,

This is the output image with DynaVision XL

Now to generate the videos I need to setup animatedDiff

Animatediff setup

I select he module mm_sdxl_v10_beta.ckpt

I will save in mp4 and webp (for this article).

I like to set 32 fps because I set the FIKM option with interp X = 6. This will make the video more fluid ( check my article about this : https://civitai.com/articles/4199/stable-diffusion-videos-with-animatediff-and-deforum-interpolation )

I keep the number of frames to 16 (changing this parameter not always works well)

ReActor

In reactor I will keep the default settings with the face generated as image..

1° TEST - DynaVision XL 👍

I run the generation of the video with the previous settings, I have a NVIDIA 4080 graphics card, but the process takes really long..

I think is quite better to generate a video with SD1.5, split the images and rescale (like in my article : https://civitai.com/articles/3501/ai-stable-diffusion-animated-gif-making-video-more-realistic ) but what I want it is to use the XL checkpoints and loras, because text are graphics are better.

Very very slow...

It took about 40 minutes👎 ..

(webp quality is ugly, mp4 is much better)

2° TEST - Juggernaut XL 👍

Because it takes too long I will change the size to 512x768 and remove Reactor in this test.

with lower resultion the speed is better: near 4 minutes for my video.

(webp quality is ugly, mp4 is much better)

3° TEST - Pony Diffusion V6 XL 👎

Same settings for Pony Diffusion V6XL but it doesn't work well.

I did changed the prompt and the seed to check the image generation for this checkpoint

but it doesn't digest the animatediff model.

Test Result

The animatediff beta model works also with some of the XL checkpoint, it the resolution is higher that 768 it takes ages to process the images.

I hope you enjoy my article

14

Comments