Тёмный

2X PERFORMANCE PLUGIN 🤯 OFFICIAL A11 STABLE DIFFUSION UPDATE GUIDE 

TroubleChute
Подписаться 259 тыс.
Просмотров 29 тыс.
50% 1

Опубликовано:

 

10 сен 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 73   
@ujjvalw2684
@ujjvalw2684 Год назад
AMD USERS : 👁️👄👁️
@aadi2134
@aadi2134 Год назад
🥲
@testales
@testales Год назад
I'll wait until control net is supported too and then see how it actually increases the performance on bigger images and especially for latent upscaling. Creating tons of low resolution images even faster doesn't add much for me. Videos sequences might be a use case but for this control net is a must have too.
@systemdersiebenwelten
@systemdersiebenwelten Год назад
The control nets seem to be working now. I just added 3 and it still runs pretty fast. But the resolution thing is a real shame.
@reallybigname
@reallybigname Год назад
@@systemdersiebenwelten Do you mean you used custom controlnets? I thought they also had to be converted. You can't even install controlnet extension, used for other things like deforum (I'm a developer).
@allendimetri8384
@allendimetri8384 Год назад
unfortunately it did not work with the RTX 2060 6GB (hopefully for now), after several hours trying to install PyCuda and getting the extension to work, with many OUT OF MEMORY warnings the .TRT was compiled but when loading it and trying to generate an image, everything crashes
@Nithproxi
@Nithproxi Год назад
your a Legend!! great video best thing about this is you got a 3080ti like me 🤣
@bomar920
@bomar920 Год назад
Thats amazing video, looking to see how far we go from here, and subscribed.
@MaxPayne_in
@MaxPayne_in Год назад
Using the same GPU, a 3080 Ti, with a xformer model, I achieved a speed of approximately 19 iterations per second. However, we can expect even better performance, around 40 iterations per second, if we combine both the xformer and TensorRT and utilize them simultaneously. I have no idea if things work this way.
@franckparkinson
@franckparkinson Год назад
however i do the checkout command the error on sd_unet doesn't disappear and the extension tab doesn't appear. please help
@appolonius4108
@appolonius4108 Год назад
i was unable to get a speed boost on a RTX 2060 6gb vram. one image is 4.5 seconds with and without using a TensorRT converted model
@M.Aditya.H
@M.Aditya.H Год назад
so there's no need to convert models into TRT? safetensors file from Civitai also got the improvement speed?
@appolonius4108
@appolonius4108 Год назад
@@M.Aditya.H not yet anyway, for a 6gb vram card. I saw no boosts in single image generation and I cannot do batch gens to compare most people's tests.
@CodX710
@CodX710 Год назад
Bro your thumbnail is so click bait...
@TroubleChute
@TroubleChute Год назад
Not nessecarily. It's an official extenssion fro Automatic1111 Stable Diffusion that took my speed from ~11 to ~20 it/s, which is ~2x. The "Photoshop" was a mistake, as I use templates I've created. Fixed. It shows Stable Diffusion now :)
@subratasadhu
@subratasadhu Год назад
@@TroubleChute you created a video on the cutting edge of this development. I was looking for a video on implementation. I thank you for your efforts. Honestly.
@ianhmoll
@ianhmoll Год назад
Amazing! Since can't support controlNET yet, I'll wait until they do to test it.
@GamingDaveUK
@GamingDaveUK Год назад
Very interesting video, I hope they add Lora support that is not needed to be baked in. Is the file size the same as the original though? that would double the hard drive space we use
@Adohleas
@Adohleas Год назад
I can sometimes get it working if I create images with increased batches and batch size, but it is very finicky and it bugs out if I try to generate only one single image and sometimes multiple images doesn't work unless I set the batches and batch size high enough, also using the hires fixes also doesn't work. I get the following error: "bad shape for TensorRT input x" a lot if I try and do single images, not enough batches and batch size, or add hires fixes.
@blind486
@blind486 10 месяцев назад
looks crap, though the speed is phenomenal, waiting for more benchmarks and reviews. the conversion of lots of files is holding me back.
@JC-jn8jz
@JC-jn8jz Год назад
i need help with one of your previous video for installing stable diffusion in mac. do you mind to check there and help those people in need?
@ruzanmuhammedasher3263
@ruzanmuhammedasher3263 Год назад
Why did you censor the model selected in a1111 webui 🤨
@TroubleChute
@TroubleChute Год назад
CivitAI has many uncensored models. Even the most innocent sounding models could be NSFW. Just to be on the safe side :)
@sectix
@sectix Год назад
lewdpunkAnima_v10 is the model he is using.
@leandrozanardo1046
@leandrozanardo1046 8 месяцев назад
It is really fast, but the results have nothing to do with the original model used. Sometimes can be nice, but in general if you are using loras it loses a lot of details...
@LibertyRecordsFree
@LibertyRecordsFree Год назад
Does the windows 10 version of "TensorRT 8.6 GA for Windows 10 and CUDA 11.0, 11.1, 11.2, 11.3, 11.4, 11.5, 11.6, 11.7 and 11.8 ZIP Package" works on windows 11?
@TrapisYT
@TrapisYT Год назад
why fatal error on CMD?
@Wl878
@Wl878 11 месяцев назад
me too
@Spindonesia
@Spindonesia Год назад
Hello! nice video! anyway im using ONNX to Tensor RT but when i change setting to max 1024 instead of 512 it always error and crash ? why? does it depend on model?
@carnacthemagnificent2498
@carnacthemagnificent2498 Год назад
I look forward to trying this. For some reason my stable diffusion install always hits an error loading xtensors and I can't figure out why but it runs so I haven't worried. It looks like that might be a problem for this model conversion though so I'll have to hope for the best.
@ProdByGhost
@ProdByGhost Год назад
u can copy the error in to chatgpt , ive fixed a few problems this way Lol
@stavsap
@stavsap Год назад
but the total time generation is the same ~7.2 sec, what's the point?
@Backtitrationfan
@Backtitrationfan Год назад
Wait so if I bake a lora in I could still use it by using the right words?
@user-ql4ki8kb2z
@user-ql4ki8kb2z 11 месяцев назад
why my some extensions not show
@TPCDAZ
@TPCDAZ 10 месяцев назад
The speed increase isn't enough to break pretty much evety lora you have and go through all those issues for.
@witness1013
@witness1013 Год назад
What has this got to do w/ Photoshop ?
@ahs9674
@ahs9674 Год назад
Looks like he made a mistake with the thumbnail
@metaversebye
@metaversebye Год назад
photoshop is currently trending because of the AI integration, so including that in the thumbnail gets you more traffic.
@TroubleChute
@TroubleChute Год назад
Yeah was a mistake :/ The "Photoshop" was a mistake, as I use templates I've created. Fixed. It shows Stable Diffusion now :)
@cameonn
@cameonn Год назад
I thought AUTOMATIC1111 made an extension for PS which would be really awesome, It appears it was a mistake.
@TroubleChute
@TroubleChute Год назад
There is generative AI in Photoshop officially now: ru-vid.com/video/%D0%B2%D0%B8%D0%B4%D0%B5%D0%BE-L_8WB4v9HXo.html As well as an extension to use A11 in Photoshop: ru-vid.com/video/%D0%B2%D0%B8%D0%B4%D0%B5%D0%BE-Z8N7jDJU6zM.html The second is much closer to what you're hoping for I assume. Generate images, inpainting, controlnet and way more.
@Neolisk
@Neolisk Год назад
What's your video card ? Is it MSI 3080Ti Gaming X Trio as mentioned in description? According to userbenchmark, 4070 Ti is only 18% faster than 3080 Ti. I get up to 24 it/sec on A1111 without TensorRT, 512x512, batch size 4 @ RTX 4070 Ti. Which means your original PyTorch numbers aren't using the full potential of your video card.
@tntblast500
@tntblast500 Год назад
18% faster in what? Because this utilises tensor cores.
@Neolisk
@Neolisk Год назад
@@tntblast500 According to wiki, 4070 Ti is 27% faster in half precision. And 18% faster in maximum theoretical tensor compute power. My point was that initial it/sec here should be around 20 it/sec, not after TensorRT, but before it. Otherwise I'm getting the same performance without Tensor, why do I need Tensor?
@tntblast500
@tntblast500 Год назад
@@Neolisk Yeah, ok. I see now.
@beastly_neon
@beastly_neon Год назад
@@tntblast500 yup you are correct i have 3070 and getting 9.5 it/sec. Looks like he is not using xformers with it
@m_sha3er
@m_sha3er Год назад
My SD is getting too slow specially with CN, I guess that's happened after Nvdia latest update, does TensorRT is the reason? RTX 3070ti
@fadedninna
@fadedninna Год назад
rollback to a older driver, they did something with the latest update
@faultline3936
@faultline3936 Год назад
Goddamn it, I'm using AMD!
@Devalinor
@Devalinor Год назад
I just stumbled upon this video. Are you able to generate pictures with more than 75 tokens? Because no matter what I do, the converted model only allows 75 tokens max for me.
@Spindonesia
@Spindonesia Год назад
oh seriously?
@octimus2000
@octimus2000 Год назад
Doesn't work with GTX 1080, right?
@TroubleChute
@TroubleChute Год назад
RTX only as far as I understand for the speed boost
@omegahunter9
@omegahunter9 Год назад
Does this work on Linux?
@costatattooz840
@costatattooz840 11 месяцев назад
yes
@generalawareness101
@generalawareness101 Год назад
It is utter crap for anything but standard 1.5 dimensions. No thanks to that, and besides I use 2.1 not 1.5 at half FHD dimensions.
@zephirius
@zephirius Год назад
thanks ! but i thing i will wait for the main update.
@kyrilgarcia
@kyrilgarcia Год назад
Does this improve performance for really old gpu's? or is it only for the newer ones?
@Leo-vb7wu
@Leo-vb7wu Год назад
It's only for the newer one's because it's use gpu's tensor cores
@MitrichDX
@MitrichDX Год назад
not Lora, not ControlNet, as VoltaML =))
@zipel
@zipel Год назад
u should make a vid about that new PS thing
@TroubleChute
@TroubleChute Год назад
This one ru-vid.com/video/%D0%B2%D0%B8%D0%B4%D0%B5%D0%BE-L_8WB4v9HXo.html ?
@Punisher1992
@Punisher1992 Год назад
so... this does not work with amd yet right?
@weirdscix
@weirdscix Год назад
AMD doesn't have Tensor cores so it will never work with it.
@Shisgara77
@Shisgara77 Год назад
@@weirdscix and how about Apple silicon chip (macbook air m1)?
@weirdscix
@weirdscix Год назад
@@Shisgara77 only Nvidia unfortunately, it uses the proprietary hardware inside RTX GPU's
@GamersGuard
@GamersGuard Год назад
RDNA 3 has WMMA AI cores ....
@weirdscix
@weirdscix Год назад
@@GamersGuard nice, but what does that have to do with Tensor cores? A 4080 has 304 cores, the 4090 has 568. When it comes to AI, there's a reason Nvidia is the leader
@omegablast2002
@omegablast2002 Год назад
I'm going to be waiting a long time before I try this... It seems destructive to your setup with very little actual benefits other than a speed increase... I'm sorry but I love using my Loras and other toys... This thing only works with textual inversions...nah, hard pass
@666Counterforce
@666Counterforce Год назад
My GTX 3070 (laptop) does 2 it/s LOL
@MrSongib
@MrSongib Год назад
I'm here watching while my card is 5700xt. xdd
@BobroBobrodobro
@BobroBobrodobro Год назад
will it work for 3060?
@huyked
@huyked 9 месяцев назад
Yes. The 3060 has 120 Tensor cores.
Далее
Why Are Open Source Alternatives So Bad?
13:06
Просмотров 630 тыс.
Stable Diffusion Performance Optimization Tutorial
5:02
NEVER install these programs on your PC... EVER!!!
19:26
Negative Embeddings - ULTRA QUALITY Trick for A1111
6:34
How to Install Stable Diffusion on AMD GPUs (NEW)
8:49