Sdxl medvram. 1. Sdxl medvram

 
 1Sdxl medvram  ReVision is high level concept mixing that only works on

Figure out anything with this yet? Just tried it again on A1111 with a beefy 48GB VRAM Runpod and had the same result. You dont need low or medvram. 手順2:Stable Diffusion XLのモデルをダウンロードする. I tried --lovram --no-half-vae but it was the same problem. 5GB vram and swapping refiner too , use --medvram-sdxl flag when starting r/StableDiffusion • Year ahead - Requests for Stability AI from community?Commands Optimizations. which is exactly what we're doing, and why we haven't released our ControlNetXL checkpoints. Option 2: MEDVRAM. Check here for more info. 5 gets a big boost, I know there's a million of us out. api Has caused the model. tiffFor me I have an 8 gig vram, trying sdxl in auto1111 just tells me insufficient memory if it even loads the model and when running with --medvram image generation takes a whole lot of time, comfi ui is just better in that case for me, lower loading times, lower generation time, and get this sdxl just works and doesn't tell me my vram is shit. --medvram --opt-sdp-attention --opt-sub-quad-attention --upcast-sampling --theme dark --autolaunch amd pro yazılımıyla performans %50 oranında arttı. json to. I have tried these things before and after a fresh install of the stable diffusion repository. I'm generating pics at 1024x1024. Okay so there should be a file called launch. 5, having found the prototype your looking for then img-to-img with SDXL for its superior resolution and finish. 5 Models. Reply reply more replies. Launching Web UI with arguments: --port 7862 --medvram --xformers --no-half --no-half-vae ControlNet v1. In the hypernetworks folder, create another folder for you subject and name it accordingly. If I do img2img using the dimensions 1536x2432 (what I've previously been able to do) I get Tried to allocate 42. Now everything works fine with SDXL and I have two installations of Automatic1111 each working on an intel arc a770. I'm on Ubuntu and not Windows. x). At first, I could fire out XL images easy. Happy generating everybody!At the line where set " COMMANDLINE_ARGS =" , add in these parameters " --xformers" and " --medvram" and " --opt-split-attention" to reduce further the VRAM needed BUT it will added the processing time. The message is not produced. Say goodbye to frustrations. Beta Was this translation helpful? Give feedback. 5gb. ) Fabled_Pilgrim. This is the log: Traceback (most recent call last): File "E:stable-diffusion-webuivenvlibsite-packagesgradio outes. Wow Thanks; it works! From the HowToGeek :: How to Fix Cuda out of Memory section :: command args go in webui-user. It would be nice to have this flag specfically for lowvram and SDXL. SDXL and Automatic 1111 hate eachother. safetensors at the end, for auto-detection when using the sdxl model. Start your invoke. 6, and now I'm getting 1 minute renders, even faster on ComfyUI. Put the base and refiner models in stable-diffusion-webuimodelsStable-diffusion. Afroman4peace. 0 Artistic StudiesNothing helps. I cant say how good SDXL 1. and this Nvidia Control. AUTOMATIC1111 版 WebUI Ver. Invoke AI support for Python 3. py build python setup. 04. There are two options for installing Python listed. 3) If you run on ComfyUI, your generations won't look the same, even with the same seed and proper. Then, I'll go back to SDXL and the same setting that took 30 to 40 s will take like 5 minutes. Downloads. Raw output, pure and simple TXT2IMG. ReplyWhy is everyone saying automatic1111 is really slow with SDXL ? I have it and it even runs 1-2 secs faster than my custom 1. • 8 mo. And I'm running the dev branch with the latest updates. This is the same problem as the one from above, to verify, Use --disable-nan-check. この記事では、そんなsdxlのプレリリース版 sdxl 0. This fix will prevent unnecessary duplication and. I have always wanted to try SDXL, so when it was released I loaded it up and surprise, 4-6 mins each image at about 11s/it. I had to set --no-half-vae to eliminate errors and --medvram to get any upscalers other than latent to work, have not tested them all, only LDSR and R-ESRGAN 4X+. python launch. The --medvram option addresses this issue by partitioning the VRAM into three parts, with one part allocated for the model and the other two parts for intermediate computation. bat or sh and select option 6. Example: set VENV_DIR=C: unvar un will create venv in. I only use --xformers for the webui. There’s a difference between the reserved VRAM (around 5GB) and how much it uses when actively generating. 74 EMU - Kolkata Trains. Ok sure, if it works for you then its good, I just also mean for anything pre SDXL like 1. Fast ~18 steps, 2 seconds images, with Full Workflow Included! No ControlNet, No ADetailer, No LoRAs, No inpainting, No editing, No face restoring, Not Even Hires Fix!! (and obviously no spaghetti nightmare). 2 / 4. 5 checkpoints Yeah 8gb is too little for SDXL outside of ComfyUI. Oof, what did you try to do. To enable higher-quality previews with TAESD, download the taesd_decoder. I have even tried using --medvram and --lowvram, not even this helps. 5 models. 9 / 1. Find out more about the pros and cons of these options and how to optimize your settings. Reply replyI run sdxl with autmatic1111 on a gtx 1650 (4gb vram). With 3060 12gb overclocked to the max takes 20 minutes to render 1920 x 1080 image. 9 / 1. But yes, this new update looks promising. using medvram preset result in decent memory savings without huge performance hit: Doggetx: 0. set COMMANDLINE_ARGS=--medvram --no-half-vae --opt-sdp-attention. ControlNet support for Inpainting and Outpainting. Before I could only generate a few. Not with A1111. And I found this answer as. 1. 0 model as well as the new Dreamshaper XL1. docker compose --profile download up --build. Effects not closely studied. So I researched and found another post that suggested downgrading Nvidia drivers to 531. I don't know if you still need an answer, but I regularly output 512x768 in about 70 seconds with 1. 0 - RTX2080 . Side by side comparison with the original. I am a beginner to ComfyUI and using SDXL 1. All tools are really not created equal in this space. 7. process_api( File "E:stable-diffusion-webuivenvlibsite. Everything is fine, though some ControlNet models cause it to slow to a crawl. Note that the Dev branch is not intended for production work and may break other things that you are currently using. I tried comfyui, 30 sec faster on a 4 batch, but it's pain in the ass to make the workflows you need, and just what you need (IMO). My faster GPU, with less VRAM, at 0 is the Window default and continues to handle Windows video while GPU 1 is making art. that FHD target resolution is achievable on SD 1. 576 pixels (1024x1024 or any other combination). 1 / 4. Now I have to wait for such a long time. sdxl_train. Say goodbye to frustrations. I'm on Ubuntu and not Windows. 4GB VRAM with FP32 VAE and 950MB VRAM with FP16 VAE. ComfyUIでSDXLを動かすメリット. py --lowvram. This exciting development paves the way for seamless stable diffusion and Lora training in the world of AI art. With this on, if one of the images fail the rest of the pictures are. 5. User nguyenkm mentions a possible fix by adding two lines of code to Automatic1111 devices. bat as . There is also another argument that can help reduce CUDA memory errors, I used it when I had 8GB VRAM, you'll find these launch arguments at the github page of A1111. First Impression / Test Making images with SDXL with the same Settings (size/steps/Sampler, no highres. 1girl, solo, looking at viewer, light smile, medium breasts, purple eyes, sunglasses, upper body, eyewear on head, white shirt, (black cape:1. As someone with a lowly 10gb card sdxl is beyond my reach with a1111 it seems. Loose-Acanthaceae-15. -if I use --medvram or higher (no opt command for vram) I get blue screens and PC restarts-I upgraded AMD driver to latest (23-7-2) but it did not help. 少しでも動作を. 3. SDXL 1. 0_0. Speed Optimization. CeFurkan • 9 mo. These are also used exactly like ControlNets in ComfyUI. The suggested --medvram I removed it when i upgraded from RTX2060-6GB to RTX4080-12GB (both Laptop/Mobile). As I said, the vast majority of people do not buy xx90 series cards, or top end cards in general, for games. 0-RC , its taking only 7. The default installation includes a fast latent preview method that's low-resolution. NOT OK > "C:My thingssome codestable-diff. You may experience it as “faster” because the alternative may be out of memory errors or running out of vram/switching to CPU (extremely slow) but it works by slowing things down so lower memory systems can still process without resorting to CPU. Web. 5 would take maybe 120 seconds. 0 Everything works perfectly with all other models (1. For 8GB vram, the recommended cmd flag is "--medvram-sdxl". I have a 3090 with 24GB of Vram cannot do a 2x latent upscale of a SDXL 1024x1024 image without running out of Vram with the --opt-sdp-attention flag. bat file would help speed it up a bit. 手順2:Stable Diffusion XLのモデルをダウンロードする. 6 and the --medvram-sdxl Image size: 832x1216, upscale by 2 DPM++ 2M, DPM++ 2M SDE Heun Exponential (these are just my usuals, but I have tried others) Sampling steps: 25-30 Hires. You can also try --lowvram, but the effect may be minimal. latest Nvidia drivers at time of writing. bat file (For windows) or webui-user. --medvram-sdxl: None: False: enable --medvram optimization just for SDXL models--lowvram: None: False: Enable Stable Diffusion model optimizations for sacrificing a lot of speed for very low VRAM usage. webui-user. Having finally gotten Automatic1111 to run SDXL on my system (after disabling scripts and extensions etc) I have run the same prompt and settings across A1111, ComfyUI and InvokeAI (GUI). set COMMANDLINE_ARGS= --medvram --autolaunch --no-half-vae PYTORCH_CUDA_ALLOC_CONF=garbage_collection_threshold:0. Stable Diffusion is a text-to-image AI model developed by the startup Stability AI. I finally fixed it in that way: Make you sure the project is running in a folder with no spaces in path: OK > "C:stable-diffusion-webui". sdxl を動かす!Running without --medvram and am not noticing an increase in used RAM on my system, so it could be the way that the system is transferring data back and forth between system RAM and vRAM, and is failing to clear out the ram as it goes. VRAM使用量が少なくて済む. use --medvram-sdxl flag when starting. then select the section "Number of models to cache". 0 Alpha 2, and the colab always crashes. use --medvram-sdxl flag when starting. 0の変更点. whl, change the name of the file in the command below if the name is different:set COMMANDLINE_ARGS=--medvram --opt-sdp-attention --no-half --precision full --disable-nan-check --autolaunch --skip-torch-cuda-test set SAFETENSORS_FAST_GPU=1. -opt-sdp-no-mem-attention --upcast-sampling --no-hashing --always-batch-cond-uncond --medvram. ReVision is high level concept mixing that only works on. But it has the negative side effect of making 1. sdxl is a completely different architecture and as such requires most extensions be revamped or refactored (with the exceptions to things that. Safetensors on a 4090, there's a share memory issue that slows generation down using - - medvram fixes it (haven't tested it on this release yet may not be needed) If u want to run safetensors drop the base and refiner into the stable diffusion folder in models use diffuser backend and set sdxl pipelineRecommandé : SDXL 1. 5GB vram and swapping refiner too , use --medvram-sdxl flag when starting r/StableDiffusion • AI Burger commercial - source @MatanCohenGrumi twitter - much better than previous monstrositiesHowever, for the good news - I was able to massively reduce this >12GB memory usage without resorting to --medvram with the following steps: Initial environment baseline. 19it/s (after initial generation). --force-enable-xformers:强制启动xformers,无论是否可以运行都不报错. 3: using lowvram preset is extremely slow due to constant swapping: xFormers: 2. 5 was "only" 3 times slower with a 7900XTX on Win 11, 5it/s vs 15 it/s on batch size 1 in auto1111 system info benchmark, IIRC. Sped up SDXL generation from 4 mins to 25 seconds!SDXL training. Before SDXL came out I was generating 512x512 images on SD1. This workflow uses both models, SDXL1. bat file. No, with 6GB you are at the limit, one batch too large or a resolution too high and you get an OOM, so --medvram and --xformers are almost mandatory things. Expanding on my temporal consistency method for a 30 second, 2048x4096 pixel total override animation. Generate an image as you normally with the SDXL v1. 6. 0 out of 5. It takes 7 minutes for me to get 1024x1024 SDXL image with A1111 and 3. modifier (I have 8 GB of VRAM). Two models are available. Your image will open in the img2img tab, which you will automatically navigate to. This allows the model to run more. The usage is almost the same as fine_tune. Divya is a gem. I posted a guide this morning -> SDXL 7900xtx and Windows 11, I. 5 didn't have, specifically a weird dot/grid pattern. takes about a minute to generate a 512x512 image without highrez fix using --medvram while my newer 6gb card takes less than 10. The place is in the webui-user. By the way, it occasionally used all 32G of RAM with several gigs of swap. 0-RC , its taking only 7. It defaults to 2 and that will take up a big portion of your 8GB. This fix will prevent unnecessary duplication. See more posts like this in r/StableDiffusionPS medvram giving me errors and just wont go higher than 1280x1280 so i dont use it. I learned that most of the things I needed I already had since I hade automatic1111, and it worked fine. ago. To learn more about Stable Diffusion, prompt engineering, or how to generate your own AI avatars, check out these notes: Prompt Engineering 101. 9, causing generator stops for minutes aleady add this line to the . Practice thousands of math and language arts skills at. It initially couldn't load the weight but then I realized my Stable Diffusion wasn't updated to v1. Recommended graphics card: ASUS GeForce RTX 3080 Ti 12GB. . 今回は Stable Diffusion 最新版、Stable Diffusion XL (SDXL)についてご紹介します。. 0 Version in Automatic1111 installiert und nutzen könnt. Cannot be used with --lowvram/Sequential CPU offloading. Sorun modelin ön gördüğünden daha düşük çözünürlük talep etmem mi ?No medvram or lowvram startup options. --lowram: None: False: Load Stable Diffusion checkpoint weights to VRAM instead of RAM. With Tiled Vae (im using the one that comes with multidiffusion-upscaler extension) on, you should be able to generate 1920x1080, with Base model, both in txt2img and img2img. 17 km. 1024x1024 instead of 512x512), use --medvram --opt-split-attention. After running a generation with the browser (tried both Edge and Chrome) minimized, everything is working fine, but the second I open the browser window with the webui again the computer freezes up permanently. They used to be on par, but I'm using ComfyUI because now it's 3-5x faster for large SDXL images, and it uses about half the VRAM on average. 1. 1 models, you can use either. 24GB VRAM. py file that removes the need of adding "--precision full --no-half" for NVIDIA GTX 16xx cards. I also added --medvram and. Two of these optimizations are the “–medvram” and “–lowvram” commands. You can make AMD GPUs work, but they require tinkering ; A PC running Windows 11, Windows 10, Windows 8. Even with --medvram, I sometimes overrun the VRAM on 512x512 images. If you’re unfamiliar with Stable Diffusion, here’s a brief overview:. 0, the various. I have a RTX3070 8GB and A1111 SDXL works flawless with --medvram and. r/StableDiffusion. It'll process a primary subject and leave the background a little fuzzy, and it just looks like a narrow depth of field. ComfyUIでSDXLを動かす方法まとめ. 1 until you like it. . That's particularly true for those who want to generate NSFW content. 5 and 2. x) and taesdxl_decoder. 12GB is just barely enough to do Dreambooth training with all the right optimization settings, and I've never seen someone suggest using those VRAM arguments to help with training barriers. 60 から Refiner の扱いが変更になりました。. Details. 筆者は「ゲーミングノートPC」を2021年12月に購入しました。 RTX 3060 Laptopが搭載されています。専用のVRAMは6GB。 その辺のスペック表を見ると「Laptop」なのに省略して「RTX 3060」と書かれていることに注意が必要。ノートPC用の内蔵GPUのものは「ゲーミングPC」などで使われるデスクトップ用GPU. Many of the new models are related to SDXL, with several models for Stable Diffusion 1. Please use the dev branch if you would like to use it today. I think the problem of slowness may be caused by not enough RAM (not VRAM) xPiNGx • 2 mo. Happens only if --medvram or --lowvram is set. Yea Im checking task manager and it shows 5. プロンプト編集のタイムラインが、ファーストパスと雇用修正パスで別々の範囲になるように変更(seed breaking change) マイナー: img2img バッチ: img2imgバッチでRAM節約、VRAM節約、. Thats why i love it. 👎 2 Daxiongmao87 and Nekos4Lyfe reacted with thumbs down emojiWhen generating, the gpu ram usage goes from about 4. webui-user. bat file at all. That is irrelevant. These also don't seem to cause a noticeable performance degradation, so try them out, especially if you're running into issues with CUDA running out of memory; of. Introducing our latest YouTube video, where we unveil the official SDXL support for Automatic1111. 手順3:ComfyUIのワークフロー. This model is open access and. Workflow Duplication Issue Resolved: The team has resolved an issue where workflow items were being run twice for PRs from the repo. set COMMANDLINE_ARGS=--xformers --medvram. Add Review. OS= Windows. 4 - 18 secs SDXL 1. About this version. @edgartaor Thats odd I'm always testing latest dev version and I don't have any issue on my 2070S 8GB, generation times are ~30sec for 1024x1024 Euler A 25 steps (with or without refiner in use). 6. In my v1. 048. AI 그림 사이트 mage. So if you want to use medvram, you'd enter it there in cmd: webui --debug --backend diffusers --medvram If you use xformers / SDP or stuff like --no-half, they're in UI settings. I have a 6750XT and get about 2. 저와 함께 자세히 살펴보시죠. I don't know how this is even possible but other resolutions can get generated but their visual quality is absolutely inferior, and I'm not talking about difference in resolution. try --medvram or --lowvram Reply More posts you may like. If you have more VRAM and want to make larger images than you can usually make (e. In your stable-diffusion-webui folder, create a sub-folder called hypernetworks. 6,max_split_size_mb:128 git pull. 5 models your 12gb vram should never need the medvram setting since cost some generation speed and for very large upscaling there is several ways to upscale by use of tiles to which the 12gb is more than enough. SDXL will require even more RAM to generate larger images. 8 / 3. 0. A Tensor with all NaNs was produced in the vae. tiff in img2img batch (#12120, #12514, #12515) postprocessing/extras: RAM savingsfinally , AUTOMATIC1111 has fixed high VRAM issue in Pre-release version 1. I installed the SDXL 0. 5x. With Automatic1111 and SD Next i only got errors, even with -lowvram parameters, but Comfy. For a few days life was good in my AI art world. nazihater3000. It’ll be faster than 12GB VRAM, and if you generate in batches, it’ll be even better. If you want to switch back later just replace dev with master . Mixed precision allows the use of tensor cores which massively speed things up, medvram literally slows things down in order to use less vram. It's probably as ASUS thing. You've probably set the denoising strength too high. 5 I can reliably produce a dozen 768x512 images in the time it takes to produce one or two SDXL images at the higher resolutions it requires for decent results to kick in. Yikes! Consumed 29/32 GB of RAM. 5Gb free when using SDXL based model). 213 upvotes · 68 comments. Extra optimizers. with this --opt-sub-quad-attention --no-half --precision full --medvram --disable-nan-check --autolaunch I could have 800*600 with my 6600xt 8g, not sure if your 480 could make it. 0 on automatic1111, but about 80% of the time I do, I get this error: RuntimeError: The size of tensor a (1024) must match the size of tensor b (2048) at non-singleton dimension 1. You can make it at a smaller res and upscale in extras though. I wanted to see the difference with those along with the refiner pipeline added. 【Stable Diffusion】SDXL. environ. commandline_args = os. I was using --MedVram and --no-half. I have same GPU and trying picture size beyond 512x512 it gives me Runtime error, "There is not enough GPU video memory". Launching Web UI with arguments: --port 7862 --medvram --xformers --no-half --no-half-vae ControlNet v1. add --medvram-sdxl flag that only enables --medvram for SDXL models; prompt editing timeline has separate range for first pass and hires-fix pass (seed breaking change) Minor: img2img batch: RAM savings, VRAM savings, . They have a built-in trained vae by madebyollin which fixes NaN infinity calculations running in fp16. SDXL liefert wahnsinnig gute. During renders in the official ComfyUI workflow for SDXL 0. Things seems easier for me with automatic1111. This is the proper command line argument to use xformers:--force-enable-xformers. They could have provided us with more information on the model, but anyone who wants to may try it out. Changes torch memory type for stable diffusion to channels last. process_api( File "E:stable-diffusion-webuivenvlibsite. 1. If you have a GPU with 6GB VRAM or require larger batches of SD-XL images without VRAM constraints, you can use the --medvram. 3) , kafka, pantyhose. I have my VAE selection in the settings set to. 0がリリースされました。. They don't slow down generation by much but reduce VRAM usage significantly so you may just leave them. Support for lowvram and medvram modes - Both work extremely well Additional tunables are available in UI -> Settings -> Diffuser Settings;Under windows it appears that enabling the --medvram (--optimized-turbo for other webuis) will increase the speed further. If you have more VRAM and want to make larger images than you can usually make (e. The solution was described by user ArDiouscuros and as mentioned by nguyenkm should work by just adding the two lines in the Automattic1111 install. will take this in consideration, sometimes i have too many tabs and possibly a video running in the back. . Prompt wording is also better, natural language works somewhat, but for 1. 0 base without refiner at 1152x768, 20 steps, DPM++2M Karras (This is almost as fast as the 1. 0, just a week after the release of the SDXL testing version, v0. 0. 2 / 4. . 3s/it on an M1 mbp with 32gb ram, using invokeAI, for sdxl 1024x1024 with refiner. Stable Diffusionを簡単に使えるツールというと既に「 Stable Diffusion web UI 」などがあるのですが、比較的最近登場した「 ComfyUI 」というツールが ノードベースになっており、処理内容を視覚化できて便利 だという話を聞いたので早速試してみました。. set COMMANDLINE_ARGS= --medvram --upcast-sampling --no-half. I can run NMKDs gui all day long, but this lacks some. 0). The beta version of Stability AI’s latest model, SDXL, is now available for preview (Stable Diffusion XL Beta). bat file (in stable-defusion-webui-master folder). 0. It still is a bit soft on some of the images, but I enjoy mixing and trying to get the checkpoint to do well on anything asked of it. I have trained profiles using both medvram options enabled and disabled but the. api Has caused the model. Happy generating everybody! (i) Generate the image more than 512*512px size (See this link > AI Art Generation Handbook/Differing Resolution for SDXL) . I you use --xformers and --medvram in your setup, it runs fluid on a 16GB 3070 Reply replyDhanshree Shripad Shenwai. 1 to gather feedback from developers so we can build a robust base to support the extension ecosystem in the long run. You'd need to train a new SDXL model with far fewer parameters from scratch, but with the same shape. 5 images take 40. We invite you to share some screenshots like this from your webui here: The “time taken” will show how much time you spend on generating an image. -. 6 and the --medvram-sdxl Image size: 832x1216, upscale by 2 DPM++ 2M, DPM++ 2M SDE Heun Exponential (these are just my usuals, but I have tried others) Sampling steps: 25-30 Hires. SDXL, and I'm using an RTX 4090, on a fresh install of Automatic 1111. The extension sd-webui-controlnet has added the supports for several control models from the community. stable-diffusion-webui * old favorite, but development has almost halted, partial SDXL support, not recommended. To save even more VRAM set the flag --medvram or even --lowvram (this slows everything but alows you to render larger images). Next. Read here for a list of tips for optimizing inference: Optimum-SDXL-Usage. Hello, I tried various LoRAs trained on SDXL 1. I applied these changes ,but it is still the same problem. eg Openpose is not SDXL ready yet, however you could mock up openpose and generate a much faster batch via 1. My workstation with the 4090 is twice as fast. 134 RuntimeError: mat1 and mat2 shapes cannot be multiplied (231x1024 and 768x320)It consuming like 5G vram at most time which is perfect but sometime it spikes to 5. Ok, so I decided to download SDXL and give it a go on my laptop with a 4GB GTX 1050. 5 as I could previously generate images in 10 seconds, now its taking 1min 20 seconds. I don't use --medvram for SD1. bat. depending on how complex I'm being) and am fine with that. 5, like openpose, depth, tiling, normal, canny, reference only, inpaint + lama and co (with preprocessors that working in ComfyUI). SDXL on Ryzen 4700u (VEGA 7 IGPU) with 64GB Dram blue screens [Bug]: #215. Then, use your favorite 1. Works without errors every time, just takes too damn long. r/StableDiffusion. See Reviews . No, it's working for me, but I have a 4090 and had to set medvram to get any of the upscalers to work, cannot upscale anything beyond 1. I find the results interesting for comparison; hopefully others will too. fix resize 1. 9 through Python 3. So at the moment there is probably no way around --medvram if you're below 12GB. Quite inefficient, I do it faster by hand. add --medvram-sdxl flag that only enables --medvram for SDXL models prompt editing timeline has separate range for first pass and hires-fix pass (seed breaking change) Minor: img2img batch: RAM savings, VRAM savings, . 5 checkpointsYeah 8gb is too little for SDXL outside of ComfyUI. 0C2F4F9EAB. Native SDXL support coming in a future release. On GTX 10XX and 16XX cards makes generations 2 times faster. there is no --highvram, if the optimizations are not used, it should run with the memory requirements the compvis repo needed. Reviewed On 7/1/2023. (20 steps sd xl base) PS sd 1. However, for the good news - I was able to massively reduce this >12GB memory usage without resorting to --medvram with the following steps: Initial environment baseline. 6.