Stable diffusion slows at 50. yaml LatentDiffusion: Running in eps-prediction mode DiffusionWrapper has 859. . 27. ) When you use hires fix, it show the finished 1st pass (lower resolution, with positive + negative prompt) at around 50%, then upscale it and continue diffusing. There were some other suggestions, such as downgrading pytorch. If I interrupt that one and skip to the second image generation, it goes back to normal speeds which is like 16-25 sec in SDXL 1024x1024 and then is normal speed from then onwards. Sep 8, 2023 · Here is how to generate Microsoft Olive optimized stable diffusion model and run it using Automatic1111 WebUI: Open Anaconda/Miniconda Terminal. When I open task manager it says my RAM is occupied like 90 % but my GPU only like 15 %. (I used: --medvram --opt-sub-quad-attention --disable-nan-check --always-batch-cond-uncond) I had Oct 25, 2023 · The Segmind Stable Diffusion Model (SSD-1B) is a distilled 50% smaller version of the Stable Diffusion XL (SDXL), offering a 60% speedup while maintaining high-quality text-to-image generation capabilities. 5 used to, which makes it viable to use SDXL for all my generations. 5. Stable Diffusion Cheat Sheets. Jul 28, 2023 · Saved searches Use saved searches to filter your results more quickly VariationQueasy222. Thanks! Higher resolutions slow down the generation + Different samplers can also change the it/s. 3 works totally fine, assuming your prompt and other settings are appropriate (mainly choice of upscaler). I've set up stable diffusion using the AUTOMATIC1111 on my system with a Radeon RX 6800 XT, and generation times are ungodly slow. Unfortunately single image generation take 5 to 6 minutes to render and is using the full power of my graphic card. I'm attempting to clone this repo, so I don't have to download it repeatedly in my workflow, and it's extremely slow. Now that i reinstalled the webui, it is, for some reason, much slower than it was before, it takes longer to start, and it takes longer to Jun 6, 2023 · I've been noticing Stable Diffusion rendering slowdowns since updating to the latest nvidia GRD but it gets more complicated than that. 135 upvotes · 17. Here are the most important ones. 0 initially takes 8-10 seconds for a 1024x1024px image on A100 GPU. I want to tell you about a simpler way to install cuDNN to speed up Stable Diffusion. input any prompt; set batchsize to any; Click [Generate]. Try installing xformers, it brought my 2080s from 4 it/s to 7 it/s (Euler a @ 512x512) hmm, im jealous, im using a tesla k80 and instead of iterations per second im looking at seconds per Mar 9, 2023 · hananbeer commented on Mar 9, 2023. 79 would solve the speed reduction and it did but a reboot undid that and returned me to slow-land. SDXL running very slow in Automatic1111 1. When I use three lora model, 2~3it/s. Token merging (ToMe) is a new technique to speed up Stable Diffusion by reducing the number of tokens (in the prompt and negative prompt) that need to be processed. Join the Hugging Face community. Compared to 1. Most seemed to have success with the driver 531. i have been using stable diffusion for a month but today suddenly it is super slow , i should say i recently downloaded a midjourney model but i am not using it , and i have 10 different models if hta is relevant to my problem please say so ? 1. I don't know if it's a problem with my internet, my location or something else. prompt #7: futuristic female warrior who is on a mission to defend the world from an evil cyborg army, dystopian future, megacity. I am talking about the computation time, not the time for pictures to appear in the folder. 1) The rest of my computer's composants are top notch, so I see no reason why it would still be so slow. The image is blurry at this point and the whole machine is laggy for these 15 seconds. Oct 18, 2022 · When you select high res fix if the value for denoising strength is low the bar drops its updates to almost zero when the high res fix kicks in around 50%. Extremely slow stable diffusion with GTX 3080. My image generation is waaaay too slow. Jun 24, 2023 · arad-top-gon Jun 24, 2023. 5s/it as well. to get started. I also have a 3070, the base model generation is always at about 1-1. I have no idea what slows me down. Describe the bug I have used a simple realization on T4 (Google Cloud) import torch from torch import autocast from diffusers import StableDiffusionPipeline access_token = "" pipe = StableDiffusionPipeline. A text prompt. conda activate Automatic1111_olive. ALSO, SHARK MAKES COPY OF THE MODEL EACH TIME YOU CHANGE RESOLUTION, so you'll need some disk space if you want multiple models with multiple resolutions. And I'm constantly hanging at 95-100% completion. This is an excellent image of the character that I described. GPU is gtx 3080 with 10gb vram, cpu is 5960x. My bet is, that both models beeing loaded at the same time on 8GB VRAM causes this problem. The thing is that the latest version of PyTorch 2. This project brings stable diffusion models to web browsers. In order to test the performance in Stable Diffusion, we used one of our fastest platforms in the AMD Threadripper PRO 5975WX, although CPU should have minimal impact on results. VoltaML => 16 it/s sometimes, suddenly drops to ~2 it/s, and first time I run after a fresh restart I get ~46 it/s, but I'd expect a bit more performance considering all the optimizations Volta does. The 4070 Ti ended up being an even bigger upgrade than I was hoping, since I get a 4x improvement in Stable Diffusion across the board, whether it's SD1. If you want more accurate preview, change the setting to "combined" (it'll take even longer to generate. 0+cu118 for Stable Diffusion also installs the latest cuDNN 8. Hi, i had the same issue, win 11, 12700k, 3060ti 8gb, 32gb ddr4, 2tb m. Slow generation on 4090. multiedge. 0, v1. Step 2: Double-click to run the downloaded dmg file in Finder. 5600G was a very popular product, so if you have one, I encourage you to test it. The same seed and the same prompt given to the same version of Stable Diffusion will output the same image every time. I’m using SDP. When PYT 2. safetensors Creating model from config: D:\Stablediffusion\stable-diffusion-webui\configs\v1-inference. I have searched the existing issues and checked the recent builds/commits. No errors in the console, nothing printed in the UI, it just seems to sit there at 50%. It doesn't really matter if I only do SD or do something else in-between like play a game. My gpu is used at 100% (but only 5. Read part 2: Prompt building. I just bought a new laptop, hoping to get more performance from Sdxl, but even though it has a more powerful graphics card, it's slow, it's 100% slower than it's on the previous laptop. Nov 10, 2022 · 1. 52 M params. S table Diffusion is a text-to-image latent diffusion model created by researchers and engineers from CompVis, Stability AI, and LAION. Within the last week at some point, my stable diffusion suddenly has almost entirely stopped working - generations that previously would take 10 seconds now take 20 minutes, and where it would previously use 100% of my GPU resources, it now uses only 20-30%. The amount of Vram changes helps. Hi there, I'm currently trying out Stable Diffusion on my GTX 1080TI (11GB VRAM) and it's taking more than 100s to create an image with these settings: num\_inference\_steps: 50. like my old GTX1080) I use the AUTOMATIC1111 WebUi. Naturally, the next thing to talk about are the settings that affect the image generation time the most. Switch between documentation themes. You switched accounts on another tab or window. Feb 17, 2024 · Benchmark1 with forge OS:ubuntu (currently useing) AUTOMATIC1111 with rocm installd comand line result: Benchmark2 with DirectML OS: Windows 11 stable-diffusion-webui-directml with (Microsoft Oliv Feb 23, 2024 · DiffusionBee is one of the easiest ways to run Stable Diffusion on Mac. That would suggest also that at full precision in whatever repo they’re hitting the memory limit at 4 images too. 3 upvotes · 11. 0. 2 is available. Stable Diffusion suddenly slowed down. Dec 10, 2022 · Slight variations are not a problem for my use but thanks for warning. Stable diffusion prompts are integral tools that facilitate the process of diffusion in these elements. Nov 16, 2022 · The goal of this article is to get you up to speed on stable diffusion. hello everyone, i have a laptop with a rtx 3060 6gb (laptop version obv) which should perform on an average 6 to 7it/s, in fact yesterday i decided to uninstall everything and do a complete clean installation of stable diffusion webui by automatic1111 and all the extensions i had previously. The diffusion process is often marked by four main elements: innovation, communication channels, time, and a social system. sh, By using share, open the page that uses the official proxy, and the graphics card can run at full power without slowing down when running inference tasks, around 17-19it/s; If use the ip port interface opened by listen, the /r/StableDiffusion is back open after the protest of Reddit killing open API access, which will bankrupt app developers, hamper moderation, and exclude blind users from the site. My A1111 takes FOREVER to start or to switch between checkpoints because it's stuck on "Loading weights [31e35c80fc] from a1111\stable-diffusion-webui\models\Stable-diffusion\sd_xl_base_1. Enter the following commands in the terminal, followed by the enter key, to install Automatic1111 WebUI. In the end, SDXL generates at about the same speed SD1. Then using the image for linerart controlnet to get consistent and desired results. (Without --no-half i only get black images with SD 2. I have to use following flags to webui to get it to run at all with only 3 GB VRAM: --lowvram --xformers --always-batch-cond-uncond --opt-sub-quad-attention --opt-split-attention-v1 Mar 11, 2024 · Increasing the batch count during generation will slow it down compared to sd. The system gets to 50% and just hangs. It doesn't need that much drawing ability. Is just me, or someone else is experiencing the same thing? Aug 6, 2023 · Running Stable Diffusion With 4-6 GB Of VRAM. If I rent a VPS with 24 GB Nvidia A10 (A GPU which is only ~8-10% faster than mine, and only has 50% more VRAM), it takes under 15 seconds. Allready installed xformers (before that, i only got 2-3 it/s. I had heard from a reddit post that rolling back to 531. 10. It can run the Automatic1111 Webui without issues. If you go to Stable Diffusion Webui on Github and check Issue 11063 you'll see it all discussed there. The las timeout failed on the filtering step, at 50%: Filtering content: 50% (6/12), 3. To clarify, this happens if I'm generating images, and Aug 3, 2023 · This version of Stable Diffusion creates a server on your local PC that is accessible via its own IP address, but only if you connect through the correct port: 7860. To our knowledge, this is the the world’s first stable diffusion completely running on the browser. Usually, on the first run (just after the model was loaded) the refiner takes 1. What slows it down? I recently voted stable diffusion onto my laptop and started with just a single model as the Internet here at work is pretty slow. This is better than some high end CPUs. What browsers do you use to access Stable Diffusion, AI News & MoreLearn how to speed up your renders by up to 50% using a quick and easy fix. 5 are trained primarily on smaller images, so choosing higher resolutions creates a lot of absurdities. Hi, I'm getting really slow iterations with my GTX 3080. 4 Anything v3 Samdoesart Ultimerge Redshift Diffusion Waifu Diffusion and more! Plus we can add models quickly upon approval. These settings will keep both the refiner and the base model you are using in VRAM, increasing the image generation speeds drastically. Dec 13, 2023 · 2. bat file, but it changed nothing. Discover how a specific configuration can optimiz Forgot to post with the update. This is part 4 of the beginner’s guide series. Use A30 graphics card on the cloud server Use the --share --listen command when starting stable diffusion webui. Happening with all models and checkpoints Anyone know what might be the problem? My GTX 1060 3 GB can output single 512x512 image at 50 steps in 67 seconds with the latest Stable Diffusion. This means that when you run your models on NVIDIA GPUs, you can expect a significant boost. Oct 4, 2023 · The Framework of Diffusion. Dec 22, 2022 · from diffusers import StableDiffusionPipeline import torch import time use_xformers = False use_benchmark = False use_tf32 = False use_vae_slicing = False use_channel I'm using the Pinokio Interface to run stable video Diffusion, but it's running suspiciously slow. Reload to refresh your session. The higher you set the value for denoising the faster the bar moves until at full strength the bar ends just 1% behind the progress bar in the command window. webui. The amount of token merging is controlled by the percentage of token merged. 0 came out something got broken because I use to have been successful using it. In testing it out I found that it was light and fast even though my laptop is not nearly as robust as my desktop at home. safetensors" I dread every time I have to restart the UI. I am running a 3070ti, which has 8gb of VRAM. 5 gb of vram out of my 12) during generation, so it's getting used for that, no issue there. However, I have an AMD 6750XT, and from what I've understood, AMD graphics cards, in general, are not the best for StableDiffusion, but Im running Stable diffusion on my 6900XT, and I feel like its way slower than normal. 5, v1. 1 Weight need the --no-half argument, but that slows it down even further. py. h: 512. The concept of ‘innovation’ in this context refers to an idea, practice, or object Just Google shark stable diffusion and you'll get a link to the github, just follow the guide from there. I wasn’t having any performance issues in SD until a week ago when all my generations speed would come to a halt midway through each image. Is there something I'm doing wrong ? $0. 19s/it after a few checks, repairs and installs, im using the latest nvidia gpu drivers 536. Go to settings > stable diffusion > Maximum number of checkpoints loaded at the same time should be set to 2 > Only keep one model on device should be UNCHECKED. I am trying to run SDXL on A1111 on my machine but its encountering a strange problem. There are certain Stable Diffusion settings that you can tweak to make your images generate faster. Nov 8, 2022 · AUTOMATIC1111 / stable-diffusion-webui Public. 1 makes things run more than twice as slow on my system and there Jun 6, 2023 · I am having the opposite issue where on the newer drivers my first image generation is slow because of some clogged memory on my GPU which frees itself as soon as it gets to the second one. Resolution for SDXL is supposed to be 1024x1024 minimum, batch size 1, bf16 and Adafactor are recommended. 5 Weight, the 2. The 'Neon Punk' preset style in Stable Diffusion produces much better results than you would expect. Nature scenery, 7670x3707. 6. Read part 3: Inpainting. Stable Diffusion Suddenly Very Slow. It is based on explicit probabilistic models to remove noise from an image. Downgrading from 536. With stable diffusion, you generate human faces, and you can also run it on your own machine, as shown in the figure below. Using Photopea (or Photoshop) to make lineart using noise-reduction, high-pass, and threshold, then adding own element to it. stable diffusion constantly stuck at 95-100% done (always 100% in console) Rtx 3070ti, Ryzen 7 5800x 32gb ram here. GeneralShop1950. 5 hours with more than one unit enabled. 99 08/08/23, not tested on older drivers. Oldest. 3/1. my RTX3060 takes about 11-14 seconds with 512x768, Euler-a, and around 20-30 steps. You signed out in another tab or window. It takes me about 10 seconds to complete a 1. Freezes at 90 - 98% while the terminal shows 100%. It also runs out of memory if I use the default scripts so I have to use the optimizedSD ones. The difference in generation time was over 1 hour at 512x768 count:100 size:8. Very slow rendering. But as soon as I enable it, it tanks down to 30-40 minutes, and up to 1. I think in the original repo my 3080 could do 4 max. Switching interface and browser tabs does not solve the problem. SLI doesn't matter (you could run another instance and Jan 8, 2024 · At the heart of stable diffusion lies the denoising strength, a parameter that dictates the amount of noise added to the original image during the generative process. /r/StableDiffusion is back open after the protest of Reddit killing open API access, which will bankrupt app developers, hamper moderation, and exclude blind users from the site. 5 image and about 2-4 minutes for an SDXL image - a single one and outliers can take even longer. true. But again, you can just read what people have said there and see if anything works. This technique primarily focuses on making the cross-attention calculation faster and less memory-consuming. Generated in Fooocus with JuggernautXL8 and then upscaled in A1111 with Juggernaut Final 1. In other words, the following relationship is fixed: 6 days ago · Check out the Stable Diffusion Course for a step-by-step guided course. I only get 5-6it/s. Upscaling above 0. Generate a 512x512 @ 25 steps image in half a second. Also max resolution is just 768×768, so you'll want to upscale later. Tried reinstalling several times. The medvram mode is meant for GPUs with 4-6 GB of internal memory, while the lowvram mode which we’ll discuss next, was created to Jul 31, 2023 · PugetBench for Stable Diffusion 0. 4-2it/s, your's fit right in. Oct 10, 2022 · Stable Diffusion takes two primary inputs and translates these into a fixed point in its model’s latent space: A seed integer. For context, I'm running everything on a Win 11 fresh install, WSL 2 Ubuntu I see people with RTX 3090 that get 17 it/s. Oct 24, 2022 · Feniksss commented on Oct 24, 2022. 5 and get 20-step images in less than a second. All of our testing was done on the most recent drivers and BIOS versions using the “Pro” or “Studio” versions of Hello Community I just bought a laptop with a RTX 3060 grafic card, I was thinking Stable diffusion would work ok on it. 0 alpha. Benchmark score for 1070s is also 1. 2 (seems helpful with data streaming "suspect resize bar and/or GPUDirect Storage" implamentation currently unknown). Sep 10, 2023 · Strategy 1: Cross-attention optimization. RTX 4090 Performance difference. Everything runs inside the browser with no need of server support. I have 3080ti with 12Gb of VRAM and 32Gb RAM, a simple image 1024x1024 at 60 steps takes about 20-30 seconds to generate without the controlnet enabled in A1111, ComfyUI and InvokeAI. To attempt to successfully use Stable Diffusion when having only between 4 and 6 gigabytes of memory in your GPU, is to run the Stable Diffusion WebUI in medvram mode. Models based on SDXL are better at creating higher May 1, 2023 · skohan. I use euler sampler, mostly 12 steps and sometimes 40 steps. Notifications Fork 24. Jul 4, 2023 · Token merging. 1:7860" or "localhost:7860" into the address bar, and hit Enter. Top. This won't be a big deal for most people, but if you're doing something more intensive like rendering videos through Stable Diffusion or very large batches then this will save a lot of heat, gpu fan noise and electricity. I had a previous installation of A1111 on my PC, but i excluded it because of some problems i had (in the end the problems were derived by a fault nvidia driver update). But I realized I didn't have the huge amount of models and Lora's and For me its just very inconsistent. A dmg file should be downloaded. Jul 14, 2023 · 0. Manual creation of output folder does not help. By the way, it occasionally used all 32G of RAM with several gigs of swap. On RTX3090 they are between 20 to 50 seconds depends on prompt, size and models. conda create --name Automatic1111_olive python=3. Deforum (Beta) Extensions preloaded Choose from: Stable Difusion v2. Read part 1: Absolute beginner’s guide. 7 file library when updating. I) Main use cases of stable diffusion There are a lot of options of how to use stable diffusion, but here are the four main use cases: Overview of the four main uses cases for stable It prevents Stable Diffusion from crashing with Out of Memory errors, but it can also slow things down if it is activated due to high VRAM utilization. Faster examples with accelerated inference. Please check out our GitHub repo to see how we did it. 50 /hr (this will most likely change in the coming months so take advantage now) Three server options to fit your workflow. r/StableDiffusion. I've seen tutorial videos in which generating at default settings takes less than 2 Minutes, but for me it takes more than an hour. 23 to 531. i was getting 47s/it now im getting 3. Could be memory, if they were hitting the limit due to a large batch size. Its installation process is no different from any other app. 5s/it, but the Refiner goes up to 30s/it. It requires a large number of steps to achieve a decent result. 80 GiB | 1. 3. Open up your browser, enter "127. . I'm using controlnet, 768x768 images. but when the progress bar reaches 100, it got stuck like th Stable Diffusion is too slow today. Stable Diffusion Accelerated API, is a software designed to improve the speed of your SD models by up to 4x using TensorRT. I use DirectML and Windows 10. I’m wondering if the cpu/mobo is the May 9, 2023 · You signed in with another tab or window. This parameter acts as a lever, allowing creators to fine-tune the balance between retaining the essence of the original image and introducing controlled perturbations. When the diffusors library pulls it down, it Drivers are all up to date, has enough power as well. Tweak Settings That Affect Image Generation Time. When I click on Generate, the progress bar moves up till 90% and then pauses for 15 seconds or more but the command prompt is showing 100% completion. guidance\_scale: 7. 7900 xtx performance issues on FHD. There is also a demo which you can try out. Also, I added --xformers to the webui-user. Oct 17, 2023 · Neon Punk Style. Jun 30, 2023 · DDPM. AMD 6750XT is extremely slow. Auto1111 is suddenly too slow. You will learn the main use cases, how stable diffusion works, debugging options, how to use it to your advantage and how to extend it. As far as I can see my GPU is not being used for Web Stable Diffusion. from_pretrained ( "CompVis/stabl AnythingV3 on SD-A, 1024x400 @ 40 steps, generated in a single second. 56 MiB/s. DDPM ( paper) (Denoising Diffusion Probabilistic Models) is one of the first samplers available in Stable Diffusion. And that's already after checking the box in Settings for fast loading. I use the 1. It recognizes that many tokens are redundant and can be combined without much consequence. When I only use one lora model, 5it/s. That 1070 obviously. Step 1: Go to DiffusionBee’s download page and download the installer for MacOS – Apple Silicon. For even faster inference, try Stable Diffusion 1. I am trying to use text2img and use the hires fix set at 2 for 2 images in a batch. Cross-attention optimization is one of the most effective ways to speed up Stable Diffusion. Collaborate on models, datasets and Spaces. w: 512. You'll see this on the txt2img tab: For stable diffusion, it can generate a 50 steps 512x512 image around 1 minute and 50 seconds. I think the problem of slowness may be caused by not enough RAM (not VRAM) 5. Check out the optimizations to SDXL for yourself on GitHub. 04. Feb 16, 2023 · modules/sd_hijack. May 1, 2023. If I do a singular image it stops at 48% and then goes incredibly slowly until 100% it takes about 1 minute 30 seconds to generate an image at 768x768 upscale to Oct 13, 2022 · Describe the bug when I set "width" and "height" to "1024x768" and click the "Generate" button, program runs normally like the first img. When it gets to 95%, it stops and it becomes slow. I have an AMD Card (Rx 6600) and I tried to make Stable Diffusion work in the last few days. It did work from the beginning with the right arguments, but the speed was terrible. 79 fixes the problem instantly. When I don't use any lora model, 7~9it/s. Really hope we'll get optimizations soon so I can really try out testing different settings. 5 denoising is only needed for latent upscaling (as anyone could tell you). 5 or SDXL. But once you start using DPM it will slow a bit. Trying to do images at 512/512 res freezes pc in automatic 1111. Relatively slow generation on GTX 1080TI. and get access to the augmented documentation experience. It has been trained on diverse datasets, including Grit and Midjourney scrape data, to enhance its ability to create a wide range of visual The Stable Diffusion Guide 🎨. The current attempt has already taken over an hour. For a single 512x512 image, it takes upwards of five minutes. 215 upvotes · 67. Man, you are clearly talking about latent upscale specifically (nearest exact). For 20 steps, 1024 x 1024,Automatic1111, SDXL using controlnet depth map, it takes around 45 secs to generate a pic with my 3060 12G VRAM, intel 12 core, 32G Ram ,Ubuntu 22. SDXL with Controlnet slows down dramatically. However, don't expect it to actually work. What should have happened? Generate faster than sd. Nov 16, 2022 · Same mistake. It stucks on "processing" step and it lasts forever. 5k; Star [UI Performance]: Slow performance when drawing inpainting area with high resolution Loading weights [4199bcdd14] from D:\Stablediffusion\stable-diffusion-webui\models\Stable-diffusion\revAnimated_v122. 5, incredibly slow, same dataset usually takes under an hour to train. For some reason, whenever I activate a lora in Forge, the first generation goes really slow, like 2-3 minutes. 5 and CN Tile. Nov 23, 2023 · Stable Diffusion models based on SD 1. Aug 30, 2023 · Deploy SDXL on an A10 from the model library for 6 second inference times. Steps to reproduce the problem. I’m not sure if I’m doing something wrong here, but rendering on my setup seems to be very slow and typically takes several minutes. I made some videos tutorials for it. 68, so you'd probably want to try that. It was automatically enabled for a few driver versions, but the newest version of the NVIDIA drivers give the option to disable it. SDXL 1. But Stable Diffusion is too slow today. I have 10GB VRAM. 1-0. Nothing I do after that point seems to change the speed much (other than changing the parameters such as output size). Upscaling at denoising of 0. 1. I would expect 3090 to do much better than 10 seconds. Mar 7, 2023 · Hey everyone, I am not sure if this is a bug or not as I am very new to Automatic1111. Or you can run it on runpod. Makes the Stable Diffusion model consume less VRAM by splitting it into three parts - cond (for transforming text into numerical representation), first_stage (for converting a picture into latent space and back), and unet (for actual denoising of latent space) and making it so that only one is in VRAM at all times, sending others to CPU RAM. Hello everyone! I'll start by saying that I don't understand much about Python, and, in fact, for me, it was quite a problem even to download and get StableDiffusion to work. After some time my iteration speed will drop from about 6 it/s to about 3 it/s. It's slow at the end. Inpainting is faster because it only draw the masked area (smaller than the actual image size), and the actual steps procesed are steps*denoising, so it'll be faster. A newer version v0. (If you use latent upscale, it'll look breaking apart at 50-60% then continue I'm getting really low iterations per second a my RTX 4080 16GB. Stable Diffusion Models, or checkpoint models, are pre-trained Stable Diffusion weights for generating a particular style of images. When upgrading SD to the latest version of Torch, you no longer need to manually install the cuDNN libraries. It is no longer available in Automatic1111. It’s trained on 512x512 images from a subset of the LAION-5B database. Downgrading Torch didn't seem to help at all. I've applied med vram, I've applied no half vae and no half, I've applied the etag [3] fix. oj qm xg ll wf vi ff qn qf ma