If you get the error: RuntimeError: Torch is not able to use GPU; add --skip-torch-cuda-test to COMMANDLINE_ARGS variable to disable this check then add "--use-directml --reinstall-torch" to the COMMANDLINE_ARGS in the webui-user.bat file through notepad this way SD will run off your GPU instead of CPU
I think I'm doing something wrong but I'm so incredibly incompetent that I can't figure out exactly what, I now have the error "launch.py: error: unrecognized arguments: --use-directml
Thank you for explaining how to fix something that THREE DIFFERENT TUTORIALS I've watched failed to explain. I stayed up til 4am last night trying to figure out why SD was using my CPU instead of my RX 6700xt.
I've a 6600xt and your video didn't work as expected for me, Only my CPU was being utilized. I switched my python 3.10.11 and use these launch options. "--medvram --backend directml --no-half --precision full --opt-sub-quad-attention --opt-split-attention-v1 --disable-nan-check --theme dark --autolaunch" Seemed to fix my issue. *Pin this as it may help others.*
This is the best tutorial so far, I got not even single error, diffrent models are working, I can use loras, embeddings. Thank you so much. And btw, I have rx570 so if someone is thinking ,,is it gonna work'' yep it is.
How? I'm using the same setup, but the VRAM consistently reaches 16GB, and the processing speed is barely 3 it/s for a 512x512 image, and usally runs out of VRAM i need some help
@@夜々宮 I had VRAM issues for the Inpaint and Inpaint Sketch features, or if I tried batch size over 1. So first make sure batch size is 1. Then try this: 1. Open webui-user (.bat file) in a text editor. (make a backup of the original) 2. Find: "set COMMANDLINE_ARGS=" line and add: --lowvram --precision full --no-half --autolaunch The full line should be: set COMMANDLINE_ARGS=--lowvram --precision full --no-half --autolaunch Save file, and try running it. Or you can just try with "--lowvram" If that doesn't help idk, since it works for me.
This repository now sets up for an NVIDIA card, and gives errors regarding lack of CUDA compatibility during installation. Trying to skip past that and run it results in "RuntimeError: "LayerNormKernelImpl" not implemented for 'Half' " error when trying to generate anything.
hello.i got this error on my 6800xt: File "I:\AI SD AMD\stable-diffusion-webui-directml\launch.py", line 48, in main() File "I:\AI SD AMD\stable-diffusion-webui-directml\launch.py", line 39, in main prepare_environment() File "I:\AI SD AMD\stable-diffusion-webui-directml\modules\launch_utils.py", line 560, in prepare_environment raise RuntimeError( RuntimeError: Torch is not able to use GPU; add --skip-torch-cuda-test to COMMANDLINE_ARGS variable to disable this chec
i get this error, i followed your tutorial step by step and other ones but keep getting this error. RuntimeError: Torch is not able to use GPU; add --skip-torch-cuda-test to COMMANDLINE_ARGS variable to disable this check Press any key to continue . . .
how do i fix the "Torch is not able to use GPU; add --skip-torch-cuda-test to COMMANDLINE_ARGS variable to disable this check" error, even when i used a guide like ru-vid.com/video/%D0%B2%D0%B8%D0%B4%D0%B5%D0%BE-vq-QQGV-NOg.html to add" -precision full --no-half --skip-torch-cuda-test" to command args, it generated images very slowly, as if it was using my cpu and not my gpu, i have a 7800 xt, can i have some help?
I'm using a Radion 6700 and keep getting "RuntimeError: Torch is not able to use GPU; add --skip-torch-cuda-test to COMMANDLINE_ARGS variable to disable this check" If I do that though, it will use the CPU instead of the GPU. Any suggestions?
i follow every step but when i try to instal it (double click) webui-user there is error said runtimerror : torch is not able to use gpu; add --skip-torch-cuda-test if i did that,ican install stable diffusion but it will use my cpu to generate spek r5 5500 rx 6650xt 16 gb dual channel
I have Git and Tortoise Git and Python 3.10.6 and used the DirectML version of stable diffusion but I still get this error: Traceback (most recent call last): File "E:\ai AMD\stable-diffusion-webui-directml\launch.py", line 48, in main() File "E:\ai AMD\stable-diffusion-webui-directml\launch.py", line 39, in main prepare_environment() File "E:\ai AMD\stable-diffusion-webui-directml\modules\launch_utils.py", line 560, in prepare_environment raise RuntimeError( RuntimeError: Torch is not able to use GPU; add --skip-torch-cuda-test to COMMANDLINE_ARGS variable to disable this check I have an 6700xt.
I followed all the steps and like the other tutorials, it didn't work. Basically it shows how to install, the problem is that, at the end of the installation, it always gives an error. It always displays at the end: " raise Exception(f"Invalid device_id argument supplied {device_id}. device_id must be in range [0, {num_devices}).") Exception: Invalid device_id argument supplied 0. device_id must be in range [0, 0)". I'm exhausted from trying so many times.
I have a RX 6950 XT 16GB and i still have this error..: RuntimeError: Could not allocate tensor with 12615680 bytes. There is not enough GPU video memory available! :(
Thanks for the tutorial. I tried to do this with a 7900xtx, but it doesn't seem to work on the GPU at all. I have tried with and without xformers. It only uses the CPU and takes forever. Any suggestions?
@@thelaughingmanofficial I redid the installation many times, it did not want to install. I didn't feel like refunding the GPU so I switched to Linux and now everything works perfectly. Thanks for your help though.
the part I still don't understand is how come my 16gb 7800xt with 16gb of ram get the "out of memory" error if I try to use hires fix, while my 2080 8gb does that just fine and render images faster as well.. I mean for the speed I can understand due to cuda optimization, but the Vram I just don't get it..
Hi how did you not get this error with 8go vram? on my 3070 it was impossible to hires straight after the image generation. I am now on a 7800XT and still the same problem.
webui-user.bat after running Error Traceback (most recent call last): File "D:\stable-diffusion-webui-directml\launch.py", line 48, in main() File "D:\stable-diffusion-webui-directml\launch.py", line 39, in main prepare_environment() File "D:\stable-diffusion-webui-directml\modules\launch_utils.py", line 560, in prepare_environment raise RuntimeError( RuntimeError: Torch is not able to use GPU; add --skip-torch-cuda-test to COMMANDLINE_ARGS variable to disable this check
Hey, it stopped working for me since today. It seems to not use my GPU anymore which has 24GB. It gives an error when trying to generate -> AttributeError: 'DiffusionEngine' object has no attribute 'is_onnx' Also it tries to use something with 14GB of memory, maybe my system ram. Any idea how to solve this?
I've never encountered this error, but it seems like a few people have. I've looked around and found a temporary fix: github.com/lshqqytiger/stable-diffusion-webui-directml/issues/296#issuecomment-1751820370 You have to go to the Stable Diffusion folder and right-click in there to open the context menu. Select "Open Git Bash here" and a command window will appear. In the window, type "git checkout f935688" (without quotes) and then press Enter. This will change the active branch of the Stable Diffusion repository. The contents of the folder will, apparently, contain code that fixes the error you've mentioned. After that, you have to edit the webui-user batch file and remove the "git pull" text to prevent Git from trying to pull the latest changes each time you run the file. Save the file after removing the text. Hopefully, you should be good to go after this. This is only a temporary fix. After a while, you should switch back to the "master" branch to receive the latest updates. You basically have to undo all of the above as follows: 1. Go to the Stable Diffusion folder and right-click. 2. Select "Open Git Bash here". 3. Type "git checkout master" without quotes and press Enter to switch to the master branch. 4. Edit the webui-user batch file and add "git pull" at the top so Git pulls the latest changes from the master branch. 5. Run the webui-user batch file. I hope that makes sense.
It doesn't work for me, it stops here and gives me this error, does anyone have a solution? E:\stable-diffusion-webui-directml>git pull Already up to date. venv "E:\stable-diffusion-webui-directml\venv\Scripts\Python.exe" fatal: No names found, cannot describe anything. Python 3.10.6 (tags/v3.10.6:9c7b4bd, Aug 1 2022, 21:53:49) [MSC v.1932 64 bit (AMD64)] Version: 1.7.0 Commit hash: d500e58a65d99bfaa9c7bb0da6c3eb5704fadf25 Traceback (most recent call last): File "E:\stable-diffusion-webui-directml\launch.py", line 48, in main() File "E:\stable-diffusion-webui-directml\launch.py", line 39, in main prepare_environment() File "E:\stable-diffusion-webui-directml\modules\launch_utils.py", line 560, in prepare_environment raise RuntimeError( RuntimeError: Torch is not able to use GPU; add --skip-torch-cuda-test to COMMANDLINE_ARGS variable to disable this check Premere un tasto per continuare . . . my configuration: I7-10700K AMD 6800 16gb 32 gb RAM DDR4 3600Ghz Driver AMD 23.12.1 Windows 10 Pro 64bit
I have 7800xt. first it asked me add --skip-torch-cuda-test to the argument to skip a test. and then it said runtime Error and could not clone stable diffusion with an error code 128.
Same problem and same GPU. I have solved, but I'm not at home now. I'll write to you as soon as possible to tell you which commands to add to the batch file
I keep getting "Could not allocate tensor with 134217728 bytes. There is not enough GPU video memory available!" using a 6600 even if i have it at lowvram or medvram at 512x512 with no upscaling wtf is going on
It is wroking but damn so slooooooow :( Just moved to a 7800Xt from a 3060Ti and the speed went from 2-3it/s to 6-7sec/it... UPDATE: with these command ling arguments it gets a boost: --upcast-sampling --medvram --no-half --precision=full --opt-sub-quad-attention --opt-split-attention-v1 --disable-nan-check
Hey man, I failed to clone the repository because it keeps throwing an error at me when it reaches 37% I'd really appreciate it if you could share the one you cloned with me via Google Drive or something. Thanks.
Hey man, I failed to clone the repository because it keeps throwing an error at me when it reaches 37% I'd really appreciate it if you could share the one you cloned with me via Google Drive or something. Thanks.
After I generate an image, GPU memory remains at 100% (8GB) until I restart my PC. Any further attempts to generate images results in a memory error. Do you have any tips?
I have an AMD 8GB vram, installed direct ML and i can see the UI, it loads models with no issues but when i click generate image it goes into Unspecified error !! absolutely no idea why it doesn't work.. any idea of why is much appreciated!... AMD is really lagging behind Nvidia on this, it runs smoothly on the other side The automatic version works fine but only uses the CPU at a rate of 20s/it... too slow
i have got a 8gb vega 64 card and ive got 2-7 it per second in automatic1111... so i really dont know what youre doing there. you have to convert the models to onnx. if u dont do that, youll get an error, or very low speed, or wired results. but yes, its a pain in the ass with older amd cards then the 6000 series, because you re not able to run rocm so there is no inpainting available, also you are not able to run many extensions like controlnet, or dreambooth... but just creating images works really fast.
@parryhotter3138 thx I found a way... actually direct ML works fine but only for samplers that are not Karras ... no idea why lol Euler and Euler a work fine and generate at good speed of 3-4s/it ... I get the unspecified error on DM 2m++ karras and other karras samplers ... so I don't use them .. control net is working fine with Euler
@@mathwiz1260 thats true, you cant use the "newer" samplers. im going for ddpm, dpm, or as you with euler. under linux with rocm and automatic1111, or comfyui youre able to use all samplers, but youll need at least a 6xxx card for that. im going for the 7800xt soon... nvidea is just way to expensive for me. i dont want to go for a 4070 with 12gb which is way more expensive with less vram... so i really hope developers will support amd better in the future, amd has made there turn with rocm.
I'm running SD on an RX 6600 with no issues other than the amount of time it takes to render. If I add another 6600, would it speed up my generations? Considering the price point and the fact that I'm already invested in one 8gb card, it seems this is a good option. Not really looking to play games on this machine. Just SD and some video editing.
@@lespretend Yes, I tried that several times. Installed the newer version again, then run the uninstaller in python, then tried the older version and I still get the message that I am running a newer version. Damned frustrating
Doesn't fucking work same error as everything else nothing works my next gpu is gonna be nvidia so I don't have to deal with this bullshit if someone has any ideas mention them please
If the GPU in question supports DirectX 12, specifically the DirectML library, the yes, it should work. If you manage to get it working, make sure you set the "--lowvram" argument in the config so you don't get (or reduce) VRAM errors.
Hi.I have a RX 7600 with 8GB vram. Stable Diffusion keep saying I have insuficiente vram (RuntimeError: Could not allocate tensor with 4915840 bytes. There is not enough GPU video memory available! Time taken: 11.1 sec.) , even when it is the first image im making in the session. I tried the "--medvram-sdxl " argument.
@@TheLoneQuester idk how to fix this. But I close the cmd and reopen it right away without closing the tab in the browser so I dont lose my prompts there.
@@TheLoneQuester Then close the stable diffusion new tab every time you restart the CMD, the old one with the prompt continue to work. Thats the only way I know to have faster use of the tool.
This video - Ultrasonic Sega Google Mega Drive 3D Ultimate HDR Fullscreen 14bit 8K Exclusive Technology 32:9 format IPS AMOLED 60FPS Me - chinese guy that looking to fucking smallest peace of paper LOL Thanks for this video, btw
getting errors can't get it to run.. command line error.. "model failed to load" and "AttributeError" object has no attribute "lowvram". I had to put 2 lines into my web-user.bat command file.. set COMMANDLINE_ARGS=--skip-torch-cuda-test --lowvram the cuda test was error original in command Line.. then lowvram error and model not load.. even get can't connect at time.. WTF.. anyone know what the hell going on... i been at this for too long! HELP! Thanx..