Sign In

Double Your Stable Diffusion Inference Speed with RTX Acceleration TensorRT: A Comprehensive Guide

Double Your Stable Diffusion Inference Speed with RTX Acceleration TensorRT: A Comprehensive Guide

Double Your Stable Diffusion Inference Speed with RTX Acceleration TensorRT: A Comprehensive Guide

Stable Diffusion Gets A Major Boost With RTX Acceleration. One of the most common ways to use Stable Diffusion, the popular Generative AI tool that allows users to produce images from simple text descriptions, is through the Stable Diffusion Web UI by Automatic1111. In today’s Game Ready Driver, NVIDIA added TensorRT acceleration for Stable Diffusion Web UI, which boosts GeForce RTX performance by up to 2X. In this tutorial video I will show you everything about this new Speed up via extension installation and TensorRT SD UNET generation.

#TensorRT #StableDiffusion #NVIDIA

Tutorial Automatic Installer ⤵️

https://www.patreon.com/posts/automatic-for-ui-86307255

Tutorial GitHub Readme File ⤵️

https://github.com/FurkanGozukara/Stable-Diffusion/blob/main/Tutorials/Tutorial-Achieving-Significant-Stable-Diffusion-Speed-Improvement-With-RTX-Acceleration.md

0:00 Introduction to how to utilize RTX Acceleration / TensorRT for 2x inference speed

2:15 How to do a fresh installation of Automatic1111 SD Web UI

3:32 How to enable quick SD VAE and SD UNET selections from settings of Automatic1111 SD Web UI

4:38 How to install TensorRT extension to hugely speed up Stable Diffusion image generation

6:35 How to start / run Automatic1111 SD Web UI

7:19 How to install TensorRT extension manually via URL install

7:58 How to install TensorRT extension via git clone method

8:57 How to download and upgrade cuDNN files

11:23 Speed test of SD 1.5 model without TensorRT

11:56 How to generate a TensorRT for a model

12:47 Explanation of min, optimal, max settings when generating a TensorRT model

14:00 Where is ONNX file is exported

15:48 How to set command line arguments to not get any errors during TensorRT generation

16:55 How to get maximum performance when generating and using TensorRT

17:41 How to start using generated TensorRT for almost double speed

18:08 How to switch to dev branch of Automatic1111 SD Web UI for SDXL TensorRT usage

20:33 The comparison of image difference between TensoRT on and off

20:45 Speed test of TensorRT with multiple resolutions

21:32 Generating a TensorRT for Stable Diffusion XL (SDXL)

23:24 How to verify you have switched to dev branch of Automatic1111 Web UI to make SDXL TensorRT work

24:32 Generating images with SDXL TensorRT

25:00 How to generate TensorRT for your DreamBooth trained model

25:49 How to install After Detailer (ADetailer) extension and what does it do explanation

27:23 Starting generation of TensorRT for SDXL

28:06 Batch size vs batch count difference

29:00 How to train amazing SDXL DreamBooth model

29:10 How to get amazing prompt list for DreamBooth models and use them

30:25 The dataset I used for DreamBooth training myself and why it is deliberately low quality

30:46 How to generate TensorRT for LoRA models

33:30 Where and how to see TensorRT profiles you have for each model

36:57 Generating LoRA TensorRT for SD 1.5 and testing it

39:54 How to fix TensorRT LoRA not being effective bug

0

Comments