--xformers-flash-attention:启用带有 Flash Attention 的 xformers 以提高再现性(仅支持 SD2. Medvram actually slows down image generation, by breaking up the necessary vram into smaller chunks. 0 A1111 in any of the windows or Linux shell/bat files there is no --medvram or --medvram-sdxl setting used. Once they're installed, restart ComfyUI to enable high-quality previews. I would think 3080 10gig would be significantly faster, even with --medvram. I am talking PG-13 kind of NSFW, maaaaaybe PEGI-16. user. SDXL initial generation 1024x1024 is fine on 8GB of VRAM, even it's okay for 6GB of VRAM (using only base without refiner). . This exciting development paves the way for seamless stable diffusion and Lora training in the world of AI art. Contraindicated (5) isocarboxazid. 5 GB during generation. So if you want to use medvram, you'd enter it there in cmd: webui --debug --backend diffusers --medvram If you use xformers / SDP or stuff like --no-half, they're in UI settings. Or Hires. Then things updated. Quite slow for a 16gb VRAM Quadro P5000. I posted a guide this morning -> SDXL 7900xtx and Windows 11, I. I'm on Ubuntu and not Windows. ダウンロード. 命令行参数 / 性能类. bat with --medvram. ComfyUIでSDXLを動かす方法まとめ. 5 would take maybe 120 seconds. 5GB vram and swapping refiner too , use --medvram-sdxl flag when starting r/StableDiffusion • [WIP] Comic Factory, a web app to generate comic panels using SDXLNative SDXL support coming in a future release. I run w/ the --medvram-sdxl flag. Do you have any tips for making ComfyUI faster, such as new workflows? We might release a beta version of this feature before 3. SDXL can indeed generate a nude body, and the model itself doesn't stop you from fine-tuning it towards whatever spicy stuff there is with a dataset, at least by the looks of it. 16GB VRAM can guarantee you comfortable 1024×1024 image generation using the SDXL model with the refiner. It's certainly good enough for my production work. Okay so there should be a file called launch. 로그인 없이 무료로 사용 가능한. #stablediffusion #A1111 #AI #Lora #koyass #sd #sdxl #refiner #art #lowvram #lora This video introduces how A1111 can be updated to use SDXL 1. As someone with a lowly 10gb card sdxl is beyond my reach with a1111 it seems. This exciting development paves the way for seamless stable diffusion and Lora training in the world of AI art. Like, it's got latest-gen Thunderbolt, but the DIsplayport output is hardwired to the integrated graphics. S tability AI recently released its first official version of Stable Diffusion XL (SDXL) v1. 새로운 모델 SDXL을 공개하면서. which is exactly what we're doing, and why we haven't released our ControlNetXL checkpoints. (R5 5600, DDR4 32GBx2, 3060Ti 8GB GDDR6) settings: 1024x1024, DPM++ 2M Karras, 20 steps, Batch size 1 commandline args:--medvram --opt-channelslast --upcast-sampling --no-half-vae --opt-sdp-attention If your GPU card has 8 GB to 16 GB VRAM, use the command line flag --medvram-sdxl. 1 File (): Reviews. Using this has practically no difference than using the official site. 0 out of 5. 筆者は「ゲーミングノートPC」を2021年12月に購入しました。 RTX 3060 Laptopが搭載されています。専用のVRAMは6GB。 その辺のスペック表を見ると「Laptop」なのに省略して「RTX 3060」と書かれていることに注意が必要。ノートPC用の内蔵GPUのものは「ゲーミングPC」などで使われるデスクトップ用GPU. Intel Core i5-9400 CPU. but I was itching to use --medvram with 24GB, so I kept trying arguments until --disable-model-loading-ram-optimization got it working with the same ones. tif, . . 4 - 18 secs SDXL 1. on my 6600xt it's about a 60x speed increase. 9, causing generator stops for minutes aleady add this line to the . bat settings: set COMMANDLINE_ARGS=--xformers --medvram --opt-split-attention --always-batch-cond-uncond --no-half-vae --api --theme dark Generated 1024x1024, Euler A, 20 steps. So for Nvidia 16xx series paste vedroboev's commands into that file and it should work! (If not enough memory try HowToGeeks commands. I can generate 1024x1024 in A1111 in under 15 seconds, and using ComfyUI it takes less than 10 seconds. Do you have any tips for making ComfyUI faster, such as new workflows?We might release a beta version of this feature before 3. Specs: RTX 3060 12GB VRAM With controlNet, VRAM usage and generation time for SDXL will likely increase as well and depending on system specs, it might be better for some. Unreserved. set COMMANDLINE_ARGS= --medvram --upcast-sampling --no-half. tiff in img2img batch (#12120, #12514, #12515) postprocessing/extras: RAM savingswithout --medvram (but with xformers) my system was using ~10GB VRAM using SDXL. After running a generation with the browser (tried both Edge and Chrome) minimized, everything is working fine, but the second I open the browser window with the webui again the computer freezes up permanently. 手順2:Stable Diffusion XLのモデルをダウンロードする. I don't know how this is even possible but other resolutions can get generated but their visual quality is absolutely inferior, and I'm not talking about difference in resolution. 0, the various. fix resize 1. To save even more VRAM set the flag --medvram or even --lowvram (this slows everything but alows you to render larger images). tiff in img2img batch (#12120, #12514, #12515) postprocessing/extras: RAM savings without --medvram (but with xformers) my system was using ~10GB VRAM using SDXL. Si vous avez moins de 8 Go de VRAM sur votre GPU, il est également préférable d'activer l'option --medvram pour économiser la mémoire, afin de pouvoir générer plus d'images à la fois. 9. 5 Models. ipinz commented on Aug 24. 👎 2 Daxiongmao87 and Nekos4Lyfe reacted with thumbs down emojiImage by Jim Clyde Monge. 5 model to generate a few pics (take a few seconds for those). After the command runs, the log of a container named webui-docker-download-1 will be displayed on the screen. Commandline arguments: Nvidia (12gb+) --xformers Nvidia (8gb) --medvram-sdxl --xformers Nvidia (4gb) --lowvram --xformers AMD (4gb) --lowvram --opt-sub-quad-attention + TAESD in settings Both rocm and directml will generate at least 1024x1024 pictures at fp16. 8~5. Also, as counterintuitive as it might seem,. 04. but now i switch to nvidia mining card p102 10g to generate, much more effcient but cheap as well (about 30 dollar) . tiff in img2img batch (#12120, #12514, #12515) postprocessing/extras: RAM savings It's not the medvram problem, I also have a 3060 12Gb, the GPU does not even require the medvram, but xformers is advisable. Funny, I've been running 892x1156 native renders in A1111 with SDXL for the last few days. 0-RC , its taking only 7. ipinz changed the title [Feature Request]: [Feature Request]: "--no-half-vae-xl" on Aug 24. You definitely need to add at least --medvram to commandline args, perhaps even --lowvram if the problem persists. They listened to my concerns, discussed options,. In diesem Video zeige ich euch, wie ihr die neue Stable Diffusion XL 1. However, for the good news - I was able to massively reduce this >12GB memory usage without resorting to --medvram with the following steps: Initial environment baseline. Before SDXL came out I was generating 512x512 images on SD1. PyTorch 2 seems to use slightly less GPU memory than PyTorch 1. Stable Diffusionを簡単に使えるツールというと既に「 Stable Diffusion web UI 」などがあるのですが、比較的最近登場した「 ComfyUI 」というツールが ノードベースになっており、処理内容を視覚化できて便利 だという話を聞いたので早速試してみました。. ago. 5 because I don't need it so using both SDXL and SD1. 0. 手順2:Stable Diffusion XLのモデルをダウンロードする. Option 2: MEDVRAM. version: 23. For standard SD 1. The extension sd-webui-controlnet has added the supports for several control models from the community. 1. 5, but it struggles when using SDXL. Whether comfy is better depends on how many steps in your workflow you want to automate. SDXL on Ryzen 4700u (VEGA 7 IGPU) with 64GB Dram blue screens [Bug]: #215. x). Use --disable-nan-check commandline argument to disable this check. 提示编辑时间线具有单独的第一次通过和雇用修复通过(种子破坏更改)的范围(#12457) 次要的: img2img 批处理:img2img 批处理中的 RAM 节省、VRAM 节省、. 0 version ratings. 4 used and the rest free. After that SDXL stopped all problems, load time of model around 30sec Reply reply Perspective-CarelessDisabling "Checkpoints to cache in RAM" lets the SDXL checkpoint load much faster and not use a ton of system RAM. 0. XX Reply replyComfy UI after upgrade: Sdxl model load used 26 GB sys ram. 1 / 2. I don't use --medvram for SD1. I have tried these things before and after a fresh install of the stable diffusion repository. 5, now I can just use the same one with --medvram-sdxl without having. I must consider whether I should use without medvram. You should definitely try Draw Things if you are on Mac. 6, and now I'm getting 1 minute renders, even faster on ComfyUI. Seems like everyone is liking my guides, so I'll keep making them :) Today's guide is about VAE (What It Is / Comparison / How to Install), as always, here's the complete CivitAI article link: Civitai | SD Basics - VAE (What It Is / Comparison / How to. There is also another argument that can help reduce CUDA memory errors, I used it when I had 8GB VRAM, you'll find these launch arguments at the github page of A1111. 1024x1024 instead of 512x512), use --medvram --opt-split-attention. There is also another argument that can help reduce CUDA memory errors, I used it when I had 8GB VRAM, you'll find these launch arguments at the github page of A1111. Native SDXL support coming in a future release. A little slower and kinda like Blender with the UI. The company says SDXL produces more detailed imagery and composition than its predecessor Stable Diffusion 2. bat) Reply reply jonathandavisisfat • Sorry for my late response but I actually figured it out right before you. I think the key here is that it'll work with a 4GB card, but you need the system RAM to get you across the finish line. We highly appreciate your help if you can share a screenshot in this format: GPU (like RGX 4096, RTX 3080,. Reply LawProud492 • Additional comment actions. 2 / 4. 0 will be, hopefully it doesnt require a refiner model because dual model workflows are much more inflexible to work with. I have tried running with the --medvram and even --lowvram flags, but they don't make any difference to the amount of ram being requested, or A1111 failing to allocate it. Loose-Acanthaceae-15. It's definitely possible. 8~5. Both models are working very slowly, but I prefer working with ComfyUI because it is less complicated. more replies. version: v1. My GPU is an A4000 and I have the --medvram flag enabled. set COMMANDLINE_ARGS= --xformers --no-half-vae --precision full --no-half --always-batch-cond-uncond --medvram call webui. With a 3090 or 4090 you're fine but that's also where you'd add --medvram if you had a midrange card or --lowvram if you wanted/needed. (For SDXL models) Descriptions; Affected Web-UI / System: SD. While SDXL offers impressive results, its recommended VRAM (Video Random Access Memory) requirement of 8GB poses a challenge for many users. I use a 2060 with 8 gig and render SDXL images in 30s at 1k x 1k. Training scripts for SDXL. x) and taesdxl_decoder. If you have more VRAM and want to make larger images than you can usually make (e. My workstation with the 4090 is twice as fast. Now that you mention it i didn't have medvram when i first tried the RC branch. Find out more about the pros and cons of these options and how to optimize your settings. Too hard for most of the community to run efficiently. Beta Was this translation helpful? Give feedback. @echo off set PYTHON= set GIT= set VENV_DIR= set COMMANDLINE_ARGS=--medvram-sdxl --xformers call webui. using medvram preset result in decent memory savings without huge performance hit: Doggetx: 0. So I've played around with SDXL and despite the good results out of the box, I just can't deal with the computation times (3060 12GB): With 1. 1 / 2. With this on, if one of the images fail the rest of the pictures are. 400 is developed for webui beyond 1. either add --medvram to your webui-user file in the command line args section (this will pretty drastically slow it down but get rid of those errors) OR. When generating images it takes between 400-900 seconds to complete (1024x1024, 1 image with low VRAM due to having only 4GB) I read that adding --xformers --autolaunch --medvram inside of the webui-user. will take this in consideration, sometimes i have too many tabs and possibly a video running in the back. Nvidia (8GB) --medvram-sdxl --xformers; Nvidia (4GB) --lowvram --xformers; See this article for more details. 👎 2 Daxiongmao87 and Nekos4Lyfe reacted with thumbs down emojiWhen generating, the gpu ram usage goes from about 4. Note you need a lot of RAM actually, my WSL2 VM has 48GB. ComfyUIでSDXLを動かすメリット. 2 seems to work well. (Here is the most up-to-date VAE for reference. It feels like SDXL uses your normal ram instead of your vram lol. 1. They have a built-in trained vae by madebyollin which fixes NaN infinity calculations running in fp16. Only makes sense together with --medvram or --lowvram. Start your invoke. 1girl, solo, looking at viewer, light smile, medium breasts, purple eyes, sunglasses, upper body, eyewear on head, white shirt, (black cape:1. 0 out of 5. 5 model is that SDXL is much slower, and uses up more VRAM and RAM. 1. Vivarevo. Most ppl use ComfyUI which is supposed to be more optimized than A1111 but for some reason, for me, A1111 is more faster, and I love the external network browser to organize my Loras. The suggested --medvram I removed it when i upgraded from RTX2060-6GB to RTX4080-12GB (both Laptop/Mobile). set COMMANDLINE_ARGS=--medvram --no-half-vae --opt-sdp-attention _____ License & Use. use --medvram-sdxl flag when starting. But yeah, it's not great compared to nVidia. So being $800 shows how much they've ramped up pricing in the 4xxx series. ) -cmdflag (like --medvram-sdxl. It takes around 18-20 sec for me using Xformers and A111 with a 3070 8GB and 16 GB ram. This also somtimes happens when I run dynamic prompts in SDXL and then turn them off. Announcement in. just installed and Ran ComfyUI with the following Commands: --directml --normalvram --fp16-vae --preview-method auto. In ComfyUI i get something crazy like 30 minutes because high RAM usage and swapping. In the realm of artificial intelligence and image synthesis, the Stable Diffusion XL (SDXL) model has gained significant attention for its ability to generate high-quality images from textual descriptions. I noticed there's one for medvram but not for lowvram yet. The VRAM usage seemed to. Slowed mine down on W10. Disabling live picture previews lowers ram use, and speeds up performance, particularly with --medvram --opt-sub-quad-attention --opt-split-attention also both increase performance and lower vram use with either no, or slight performance loss AFAIK. It takes now around 1 min to generate using 20 steps and the DDIM sampler. 400 is developed for webui beyond 1. 0 base model. Video Summary: In this video, we'll dive into the world of automatic1111 and the official SDXL support. This is the same problem. 3 on 10: 35: 31-732037 INFO Running setup 10: 35: 31-770037 INFO Version: cf80857b Fri Apr 21 09: 59: 50 2023 -0400 10: 35: 32-113049 INFO Latest published. Some people seem to reguard it as too slow if it takes more than a few seconds a picture. 1. 0-RC , its taking only 7. SDXL for A1111 Extension - with BASE and REFINER Model support!!! This Extension is super easy to install and use. 20 • gradio: 3. tiffFor me I have an 8 gig vram, trying sdxl in auto1111 just tells me insufficient memory if it even loads the model and when running with --medvram image generation takes a whole lot of time, comfi ui is just better in that case for me, lower loading times, lower generation time, and get this sdxl just works and doesn't tell me my vram is shit. medvram-sdxl and xformers didn't help me. But yeah, it's not great compared to nVidia. D28D45F22E. 0. There is also an alternative to --medvram that might reduce VRAM usage even more, --lowvram, but we can’t attest to whether or not it’ll actually work. You can check Windows Taskmanager to see how much VRAM is actually being used while running SD. I was running into issues switching between models (I had the setting at 8 from using sd1. I installed the SDXL 0. set COMMANDLINE_ARGS=--xformers --medvram. If I do a batch of 4, it's between 6 or 7 minutes. 1024x1024 instead of 512x512), use --medvram --opt-split-attention. Two models are available. 19--precision {full,autocast} 在这个精度下评估: evaluate at this precision: 20--shareTry setting the "Upcast cross attention layer to float32" option in Settings > Stable Diffusion or using the --no-half commandline argument to fix this. bat file specifically for SDXL, adding the above mentioned flag, so i don't have to modify it every time i need to use 1. • 3 mo. And I found this answer as. isocarboxazid increases effects of dextroamphetamine transdermal by decreasing metabolism. 5. The solution was described by user ArDiouscuros and as mentioned by nguyenkm should work by just adding the two lines in the Automattic1111 install. The suggested --medvram I removed it when i upgraded from RTX2060-6GB to RTX4080-12GB (both Laptop/Mobile). 0 Version in Automatic1111 installiert und nutzen könnt. I can run NMKDs gui all day long, but this lacks some. . Refiner same folder as Base model, although with refiner i can't go higher then 1024x1024 in img2img. Autoinstaller. 5 and SD 2. --api --no-half-vae --xformers : batch size 1 - avg 12. whl, change the name of the file in the command below if the name is different:set COMMANDLINE_ARGS=--medvram --opt-sdp-attention --no-half --precision full --disable-nan-check --autolaunch --skip-torch-cuda-test set SAFETENSORS_FAST_GPU=1. tiff in img2img batch (#12120, #12514, #12515) postprocessing/extras: RAM savingsThis is assuming A1111 and not using --lowvram or --medvram . medvram and lowvram Have caused issues when compiling the engine and running it. 合わせ. (PS - I noticed that the units of performance echoed change between s/it and it/s depending on the speed. It would be nice to have this flag specfically for lowvram and SDXL. My full args for A1111 SDXL are --xformers --autolaunch --medvram --no-half. 0の変更点は? I think SDXL will be the same if it works. 0, the various. 0. 5 models are pointless, SDXL is much bigger and heavier so your 8GB card is a low-end GPU when it comes to running SDXL. If you want to switch back later just replace dev with master . 0. 4: 1. Note that the Dev branch is not intended for production work and may. (just putting this out here for documentation purposes) Reply reply. 23年7月27日にStability AIからSDXL 1. md, and it seemed to imply that when using the SDXL model loaded on the GPU in fp16 (using . 動作が速い. Generated 1024x1024, Euler A, 20 steps. 0_0. PVZ82 opened this issue Jul 31, 2023 · 2 comments Open. I've seen quite a few comments about people not being able to run stable diffusion XL 1. Then, I'll change to a 1. You may experience it as “faster” because the alternative may be out of memory errors or running out of vram/switching to CPU (extremely slow) but it works by slowing things down so lower memory systems can still process without resorting to CPU. I updated to A1111 1. I have the same issue, got an Arc A770 too so i guess the card is the problem. The sd-webui-controlnet 1. finally , AUTOMATIC1111 has fixed high VRAM issue in Pre-release version 1. 8: from 640x640 to 1280x1280 Without medvram it can only handle 640x640, which is half. bat file set COMMANDLINE_ARGS=--precision full --no-half --medvram --always-batch. To save even more VRAM set the flag --medvram or even --lowvram (this slows everything but alows you to render larger images). The post just asked for the speed difference between having it on vs off. modifier (I have 8 GB of VRAM). With ComfyUI it took 12sec and 1mn30sec respectively without any optimization. (20 steps sd xl base) PS sd 1. that FHD target resolution is achievable on SD 1. 6. 0. During image generation the resource monitor shows that ~7Gb VRAM is free (or 3-3. 1. Don't forget to change how many images are stored in memory to 1. With. AI 그림 사이트 mage. Use --disable-nan-check commandline argument to. im using pytorch Nightly (rocm5. I've managed to generate a few images with my 3060 12Gb using SDXL base at 1024x1024 using the -medvram command line arg and closing most other things on my computer to minimize VRAM usage, but it is unreliable at best, -lowvram is more reliable, but it is painfully slow. g. Then, use your favorite 1. 2. I had been used to . ControlNet support for Inpainting and Outpainting. bat file (in stable-defusion-webui-master folder). Yikes! Consumed 29/32 GB of RAM. tif, . Comfy UI offers a promising solution to the challenge of running SDXL on 6GB VRAM systems. You can make it at a smaller res and upscale in extras though. I can use SDXL with ComfyUI with the same 3080 10GB though, and it's pretty fast considerign the resolution. This is the same problem as the one from above, to verify, Use --disable-nan-check. It’ll be faster than 12GB VRAM, and if you generate in batches, it’ll be even better. Before 1. not SD. . g. webui-user. The t-shirt and face were created separately with the method and recombined. I have a 3090 with 24GB of Vram cannot do a 2x latent upscale of a SDXL 1024x1024 image without running out of Vram with the --opt-sdp-attention flag. 以下の記事で Refiner の使い方をご紹介しています。. Speed Optimization. Reply reply more replies. You've probably set the denoising strength too high. On a 3070TI with 8GB. It provides an interface that simplifies the process of configuring and launching SDXL, all while optimizing VRAM usage. ReVision is high level concept mixing that only works on. tif, . I tried --lovram --no-half-vae but it was the same problem. 9 through Python 3. 0, the various. 0-RC , its taking only 7. 0 Alpha 2, and the colab always crashes. Introducing our latest YouTube video, where we unveil the official SDXL support for Automatic1111. Recommended graphics card: ASUS GeForce RTX 3080 Ti 12GB. 1 You must be logged in to vote. If you have bad performance on both, take a look on the following tutorial (for your AMD gpu):So, all I effectively did was add in support for the second text encoder and tokenizer that comes with SDXL if that's the mode we're training in, and made all the same optimizations as I'm doing with the first one. --medvram-sdxl: None: False: enable --medvram optimization just for SDXL models--lowvram: None: False: Enable Stable Diffusion model optimizations for sacrificing a lot of speed for very low VRAM usage. With 12GB of VRAM you might consider adding --medvram. I think the problem of slowness may be caused by not enough RAM (not VRAM) xPiNGx • 2 mo. 048. Stable Diffusion is a text-to-image AI model developed by the startup Stability AI. I've been using this colab: nocrypt_colab_remastered. Although I can generate SD2. With Automatic1111 and SD Next i only got errors, even with -lowvram parameters, but Comfy. The post just asked for the speed difference between having it on vs off. UI. With 3060 12gb overclocked to the max takes 20 minutes to render 1920 x 1080 image. Just copy the prompt, paste it into the prompt field, and click the blue arrow that I've outlined in red. You need to use --medvram (or even --lowvram) and perhaps even --xformers arguments on 8GB. 5. Medvram actually slows down image generation, by breaking up the necessary vram into smaller chunks. The sd-webui-controlnet 1. py", line 422, in run_predict output = await app. I have searched the existing issues and checked the recent builds/commits. Also, don't bother with 512x512, those don't work well on SDXL. 5, but it struggles when using. 3s/it on an M1 mbp with 32gb ram, using invokeAI, for sdxl 1024x1024 with refiner. AutoV2. Special value - runs the script without creating virtual environment. I tried looking for solutions for this and ended up reinstalling most of the webui, but I can't get SDXL models to work. as higher rank models requires more vram ,The subreddit for all things related to Modded Minecraft for Minecraft Java Edition --- This subreddit was originally created for discussion around the FTB launcher and its modpacks but has since grown to encompass all aspects of modding the Java edition of Minecraft. 2gb (so not full) I tried different CUDA settings mentioned above in this thread and no change. Before I could only generate a few SDXL images and then it would choke completely and generating time increased to like 20min or so. Web. Note that the Dev branch is not intended for production work and may break other things that you are currently using. Next is better in some ways -- most command lines options were moved into settings to find them more easily. 11. 1: 6. SDXL liefert wahnsinnig gute. Stable Diffusion XL(通称SDXL)の導入方法と使い方. Welcome to /r/hoggit, a noob-friendly community for fans of high-fidelity combat flight simulation. 0 model as well as the new Dreamshaper XL1. 9vae. Much cheaper than the 4080 and slightly out performs a 3080 ti. Question about ComfyUI since it's the first time i've used it, i've preloaded a worflow from SDXL 0. I think SDXL will be the same if it works. It functions well enough in comfyui but I can't make anything but garbage with it in automatic. this is the tutorial you need : How To Do Stable Diffusion Textual. 6 and the --medvram-sdxl Image size: 832x1216, upscale by 2 DPM++ 2M, DPM++ 2M SDE Heun Exponential (these are just my usuals, but I have tried others) Sampling steps: 25-30 Hires. Two of these optimizations are the “–medvram” and “–lowvram” commands. Only makes sense together with --medvram or --lowvram--opt-channelslast: Changes torch memory type for stable diffusion to channels last. I have tried rolling back the video card drivers to multiple different versions. Make the following changes: In the Stable Diffusion checkpoint dropdown, select the refiner sd_xl_refiner_1. Reply reply. But it has the negative side effect of making 1. ago. set COMMANDLINE_ARGS=--opt-split-attention --medvram --disable-nan-check --autolaunch My graphics card is 6800xt, I started with the above parameters, generated 768x512 img, Euler a, 1. It should be pretty low for hires fix, somewhere between 0. Both models are working very slowly, but I prefer working with ComfyUI because it is less complicated. 그림의 퀄리티는 더 높아졌을지. Now everything works fine with SDXL and I have two installations of Automatic1111 each working on an intel arc a770. Jumped to 24 GB during final rendering. 1+cu118 • xformers: 0. r/StableDiffusion.