Double Your Stable Diffusion Inference Speed with RTX Acceleration TensorRT: A Comprehensive Guide

Stable Diffusion Gets A Major Boost With RTX Acceleration. One of the most common ways to use Stable Diffusion, the popular Generative AI tool that allows users to produce images from simple text descriptions, is through the Stable Diffusion Web UI by Automatic1111. In today’s Game Ready Driver, NVIDIA added TensorRT acceleration for Stable Diffusion Web UI, which boosts GeForce RTX performance by up to 2X. In this tutorial video I will show you everything about this new Speed up via extension installation and TensorRT SD UNET generation. #TensorRT #StableDiffusion #NVIDIA Automatic Installer Of Tutorial ⤵️ Tutorial GitHub Readme File ⤵️ 0:00 Introduction to how to utilize RTX Acceleration / TensorRT for 2x inference speed 2:15 How to do a fresh installation of Automatic1111 SD Web UI 3:32 How to enable quick SD VAE and SD UNET selections from settings of Automatic1111 SD Web UI 4:38 How to install TensorRT extension to hugely speed up Stable Diffusion image generation 6:35 How to start / run Automatic1111 SD Web UI 7:19 How to install TensorRT extension manually via URL install 7:58 How to install TensorRT extension via git clone method 8:57 How to download and upgrade cuDNN files 11:23 Speed test of SD 1.5 model without TensorRT 11:56 How to generate a TensorRT for a model 12:47 Explanation of min, optimal, max settings when generating a TensorRT model 14:00 Where is ONNX file is exported 15:48 How to set command line arguments to not get any errors during TensorRT generation 16:55 How to get maximum performance when generating and using TensorRT 17:41 How to start using generated TensorRT for almost double speed 18:08 How to switch to dev branch of Automatic1111 SD Web UI for SDXL TensorRT usage 20:33 The comparison of image difference between TensoRT on and off 20:45 Speed test of TensorRT with multiple resolutions 21:32 Generating a TensorRT for Stable Diffusion XL (SDXL) 23:24 How to verify you have switched to dev branch of Automatic1111 Web UI to make SDXL TensorRT work 24:32 Generating images with SDXL TensorRT 25:00 How to generate TensorRT for your DreamBooth trained model 25:49 How to install After Detailer (ADetailer) extension and what does it do explanation 27:23 Starting generation of TensorRT for SDXL 28:06 Batch size vs batch count difference 29:00 How to train amazing SDXL DreamBooth model 29:10 How to get amazing prompt list for DreamBooth models and use them 30:25 The dataset I used for DreamBooth training myself and why it is deliberately low quality 30:46 How to generate TensorRT for LoRA models 33:30 Where and how to see TensorRT profiles you have for each model 36:57 Generating LoRA TensorRT for SD 1.5 and testing it 39:54 How to fix TensorRT LoRA not being effective bug
Back to Top