Тёмный

How to Choose an NVIDIA GPU for Deep Learning in 2023: Ada, Ampere, GeForce, NVIDIA RTX Compared 

Jeff Heaton
Подписаться 88 тыс.
Просмотров 138 тыс.
50% 1

If you are thinking about buying one... or two... GPUs for your deep learning computer, you must consider options like Ada, 30-series, 40-series, Ampere, and GeForce. These brands are rapidly evolving. In this video, I discuss this ever-changing landscape as of January 2023.
There are many changes this year. NVLink is no more, unless you are dealing with the Hopper server class. The latest generation PCIe bus handles the cross GPU data transfer.
0:05 Assumptions
1:08 NVIDIA RTX (Pro) or GeForce??
3:27 NVIDIA GeForce
5:55 Memory is King
6:05 Suggestions
~~~~~~~~~~~~~~~ MY DEEP LEARNING COURSE ~~~~~~~~~~~~~~~
📖 Textbook - www.heatonresearch.com/book/a...
😸🐙 GitHub - github.com/jeffheaton/t81_558...
▶️ Play List - • 2023 Keras Version of ...
🏫 WUSTL Course Site - sites.wustl.edu/jeffheaton/t8...
~~~~~~~~~~~~~~~ CONNECT ~~~~~~~~~~~~~~~
🖥️ Website: www.heatonresearch.com/
🐦 Twitter - / jeffheaton
😸🐙 GitHub - github.com/jeffheaton
📸 Instagram - / jeffheatondotcom
🦾 Discord: / discord
▶️ Subscribe: ru-vid.com...
~~~~~~~~~~~~~~ SUPPORT ME 🙏~~~~~~~~~~~~~~
🅿 Patreon - / jeffheaton
🙏 Other Ways to Support (some free) - www.heatonresearch.com/suppor...
~~~~~~~~~~~~~~~~~~~~~~~~~~~~
#gpu #cuda #lovelace #ada #ampere #Python #Tensorflow #Keras #nvidia #3080 #3090 #4080 #4090 #pytorch

Наука

Опубликовано:

 

4 янв 2023

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 285   
@FiveFishAudio
@FiveFishAudio Год назад
I'm a beginner in ML, and got a used RTX3060 with 12GB for $200 off eBay. No regrets and for now meets my needs. A good upgrade from my old GTX 970!
@ILoveSlyFoxHound
@ILoveSlyFoxHound Год назад
How did you find one for $200? That's amazing! I'll give it some more time to see if i can find one around there soon to upgrade from my RX 470.
@lp67O
@lp67O Год назад
Can you deliver some numerical data on speedup? In terms of training speed, I mean
@gnanashekar
@gnanashekar 11 месяцев назад
@FiveFishAudio does it supports cuDNN and what do you think about it for a lap (the same one)
@FiveFishAudio
@FiveFishAudio 11 месяцев назад
@@gnanashekar Yes,RTX 3060 supports cuDNN. I compiled OpenCV with CUDA 11.7, and cuDNN 8.7.0, running on Linux Ubuntu 22.04.02
@FiveFishAudio
@FiveFishAudio 11 месяцев назад
@@lp67O Depends on model complexity and size of your dataset, but yes it speeds up a lot compared to CPU only or my old GTX 970
@datapro007
@datapro007 Год назад
Great video, thanks Jeff. I bought a RTX 3060 for $340 ($280 as of 6/23) shipped from B&H on Black Friday. The 12 GB of RAM at that price was the deciding factor. BEWARE: NVidia is now selling a "3060" with 8GB RAM in addition to the 12 GB version, so read the specs carefully before you buy.
@Jake___W
@Jake___W Год назад
Yes it has 12 GB of ram, but they cheaped out on it, the ram is only rated for 192-bits. It's why if you look at the 3070 it has 8GB of 256-bit ram, but the ram is faster and more efficient making up for the smaller number and evidently being better than the 3060's 12GB. People complained about the older ram in the 3060 so they released a new 3060ti that has 8GB of the newer and faster ram. Honestly both cards are good, some say the 8GB one will preform better, others claim the older 12GB one works better on their machine.
@datapro007
@datapro007 Год назад
@@Jake___W I'm not talking about the 3060 Ti having only 8GB - I mean the plain old 3060 in some versions only has 8GB. Say what you will, the 3060 12GB for ~ $280 is good value for money for machine learning. You can have the speediest card in the world, but if your model won't fit in available RAM you are out of luck as well as memory.
@marhensa
@marhensa 11 месяцев назад
@@datapro007 slower speed for me personaly is okay. I eventualy choose RTX 3060 12GB DDR6 over RTX 3060Ti 8GB DDR6X. so many reviewer says don't buy RTX 3060 in 2021, and recommend to buy Ti instead. but it's seem their recommendation is changed to favor 12GB now, lots of games now need more than 8GB, and also for AI stuff, more VRAM Size is important as a matter of go or no go, not just slow or fast.
@enduga0
@enduga0 5 месяцев назад
@@marhensa is i5 12 h 16gb ram intel iris good enough
@jeffm4284
@jeffm4284 10 месяцев назад
I like your explainers @jeff Heaton. Don't forget, the professional workstation RTX models use ECC RAM, use less power, and have a profile that fits in most cases. The GeForce models are enormous! In my line of work - ECC RAM and TDW is a huge deal for the targeted AI/ML processing. I don't play computer games, nor do I imagine I ever will. I may have an Xbox to de-stress occasionally with NBA 2K15 (yep, that's how invested I am in gaming). If I could share images - I'd share my matrix comparison of specs and current price differences. There's no question that you get a big bump from the RTX 4000 Ada to the previous gen RTX A5000. But, the 4000 Ada uses a TDP 70W (no power cable necessary!) vs 230W, has a fair amount of Tensor Cores and RAM, and it's MUCH smaller. Each one of these video cards needs to be matched to the expected workload. They're just too expensive not to optimize the workload mission. For instance, if you look at what's benchmarked at techpowerup for supposedly equal GPU's, the later-gen RTX A5000 crushes the RTX 4070 (4070 FP64 = 455.4 GFLOPS vs A5000 FP64 = 867.8 GFLOPS). Here's the Price Ladder: PNY NVIDIA RTX 4000 SFF Ada Gen $1,500 PNY NVIDIA RTX A5000 $1,670-$1,500 = $170 more PNY NVIDIA RTX A5500 $2,289-$1,670 = $619 more PNY NVIDIA RTX A6000 48GB GDDR6 $3,979-$2,289 = $1,690 more PNY NVIDIA RTX 6000 Ada 48 GB GDDR6 $7,985-$3,979 = $4,006 more Those are REALLY big steps-up in cost! Links in order: www.techpowerup.com/gpu-specs/geforce-rtx-4070.c3924 www.techpowerup.com/gpu-specs/rtx-4000-sff-ada-generation.c4139 www.techpowerup.com/gpu-specs/rtx-a5000.c3748 www.techpowerup.com/gpu-specs/rtx-a5500.c3901 www.techpowerup.com/gpu-specs/rtx-a6000.c3686 www.techpowerup.com/gpu-specs/rtx-5000-ada-generation.c4152 www.techpowerup.com/gpu-specs/rtx-6000-ada-generation.c3933
@nathanbanks2354
@nathanbanks2354 Год назад
I'm looking forward to FP8, which should be coming to the 4090's as a software update. That will allow running 20B parameter models at full speed with 24GB of RAM and little loss in quality and some performance gain since it needs less memory bandwidth. But I'll probably be renting 4090's on runpod; for my own machine, I just bought a used Dell precision 7720 with a Quadro P5000 with 16GB of VRAM. It runs the 7B Alpaca/LLaMA models at 7 tokens/second. It also runs whisper at slightly faster than real time Speech to Text, which was around 30% the speed of a 3090. (I even experimented with 3 bit quantization for the LLaMA 30B model, but it wasn't worth it because it was super slow and it still ran out of RAM for large queries.)
@yolo4eva
@yolo4eva Год назад
Thank you for the wonderful video! currently using a 3060, seeking to move to a 40s. will be waiting for new videos with the new setup ready!
@HeatonResearch
@HeatonResearch Год назад
Oh thanks, they do exist!! I was curious about that one. I tried to obtain one, but had failed in my attempt.
@TheTakenKing999
@TheTakenKing999 Год назад
In Montreal right now. Got a 3090 for like 750 Canadian brand new on marketplace. Your videos are always helpful. I am doing a Masters in ML here and even though I have cluster access the personal GPU does help a lot especially with smaller assignments and projects. RoBERTa with my dataset for an NLP task took 2 hours for training and 2 for fine tuning with 8 epochs on pytorch, if anyone is curious. Mixed precision would have helped speed it up even more most likely.
@congluong463
@congluong463 Год назад
I did some testing on RTX 3x graphics card by training Classifier with and without Mixed precision (on Tensorflow). And I found that mixed precision actually slowed down training on RTX 3x cards (2.5x slower). I think it's because the RTX 3x uses tensor float 32 to calculate, so there are no need to use float 8 or float 16, the operations that convert float 32 to float 16 slowed down the computation. Or maybe I did something wrong.
@Gogies777
@Gogies777 Год назад
is this even real life? I'm in westcoast Canada and nowhere near around that price range we could get 3090 for. by any chance you could share a link of some of these 3090s?
@fluxqc
@fluxqc Год назад
@@Gogies777 keep looking!! Can confirmed I just bought 5 3090s at 700$ cad
@zurgmuckerberg
@zurgmuckerberg 5 месяцев назад
Where did you catch the Canadians to trade?
@Cuplex1
@Cuplex1 2 месяца назад
Were they stolen? 🤔 Which brand? I mean, the average price for a RTX 3090 is about $1600 USD today. The cheapest I could find was $1050 USD. That's about 1430 Canadian dollars for the cheapest! Brand new for half the price of the cheapest brand available sounds unlikely, since it would be a loss for the retailer.
@Idtelos
@Idtelos Год назад
Driver support for pro series cards is why I went with the rtx a6000. Working on ML application in CFD, gpu acceleration is of great help in getting solutions faster.
@user-gg5py8ku2t
@user-gg5py8ku2t Год назад
Great video! Thanks for your efforts. What do you think about A5000 card in terms of deep learning performance? Regarding its specs, I assume it can be considered as an alternative to 4070 Ti. Would you suggest 4090 or A5000 for a deep learning research regardless of their prices?
@shafiiganjeh8082
@shafiiganjeh8082 9 месяцев назад
There is also the budget option to get used cards from the nvidia tesla series, mainly the p100 and v100. There are a pain to set up and generally hard to find at a reasonable price but absolutely worth it if you only care about deep learning. I was using a p100 for 250€ until I got my hands on the 32gb v100 for 700€ (which is an absolute steal btw).
@plumberski8854
@plumberski8854 Год назад
Glad that I come across your video today. About to get a GPU for the PC I built last year, waiting for the new GPU from Nvidia and AMD. New to machine and deep learning, using Python. Appreciate a ball park feel for the time taken with 8 and 16 GB VRAM. I know it is program-specific, amount of raw data, layers, etc. Thanks.
@DavePetrillo
@DavePetrillo Год назад
I found a 3060 12gb that was taken out of an oem on amazon about a year ago for about $480 and it has been pretty good.
@F0XxX98
@F0XxX98 Год назад
Just got a used 3060 for $300 :)
@yuliu4177
@yuliu4177 Год назад
Excellent video, Dose less ram needed if not research in the computer vision but in the machine learning which deal with PDEs or ODEs. Dose 8GB enough for most of things?
@faisalq4092
@faisalq4092 Год назад
Thanks for the video. I've been a fan your work since 2020, and it had a big impact on me. Currently I'm getting into NLP (in both research and applications). and I've been looking into buying a new GPU for my work. and I always wanted to know if different gpus would provide a marginally different training time for NLP models. Also, if you have trained a large NLP model before, do you know how long would it take a 4090 to train a BERT base model. Thanks in advance
@sophiawright3421
@sophiawright3421 Год назад
very useful video - would love your thoughts on what kind of models you think can be trained on the new RTX-4090 laptops, and what you think of those laptops for deep learning in general, thank you!
@gabesusman4592
@gabesusman4592 9 месяцев назад
exactly what I needed, thanks man
@PicaPauDiablo1
@PicaPauDiablo1 Год назад
Always well done. Thanks Doc
@yashahamdani6724
@yashahamdani6724 10 месяцев назад
Hay, thanks for video. I have an question, do you think Gtx 1660Ti is still worth it?
@yurigansmith
@yurigansmith 10 месяцев назад
Nice video, great channel. ps: Can you do a video on the significance of CPU performance in a ML setting? In particular with a view to different training scenarios, potential bottlenecks, how many cores per gpu required, and so on.
@henry3435
@henry3435 Год назад
Hey Jeff, thanks for the great content. Have you needed around with AMD & ROCm at all? Would be interesting to learn more about that.
@pedrojosemoragallegos
@pedrojosemoragallegos 6 месяцев назад
Hi Jeff, I want to go deep into NLP and Speech Recognition. Do you recommend a RTX 4090 or should I take more RAM? I don’t know how big these models are and will be.. hope you can help me!!
@blanchr2
@blanchr2 Год назад
I am looking at buying a GPU for machine learning, and this was helpful advice. Thanks!
@lovony
@lovony 5 месяцев назад
Great. Clean presentation, thank you.
@ElinLiu0823
@ElinLiu0823 5 дней назад
Hi Sir,i have watch the all content. So i have i quession: Is Quad TITAN V still can be use as a budget limit solution ?
@JoseSilva-gt6zj
@JoseSilva-gt6zj Год назад
Amazing video, Jeff! Please, could you show some models running and how much memory they consumes? Thank you !!!
@denysivanov3364
@denysivanov3364 Год назад
buy 4090, 24 gb is the minimal amateur size when you actually can do stuff with limitations.. 40 and 80 gb are pro memory sizes.
@JoseSilva-gt6zj
@JoseSilva-gt6zj Год назад
@@denysivanov3364 Thank you !
@mattelles
@mattelles Год назад
I'm in doubt between an A4000 and a 4070 TI. They are the same price where I live. As NLP models keep getting bigger and bigger, would the additional 4GB VRAM of the A4000 make a huge difference? BERT-Like Base models seem to run fine with 12GB. However, I'm not so confidence about these larger multimodal models like LayoutLM
@techtran
@techtran Год назад
Great video Professor Heaton. I appreciate your opinion and all the effort in putting out these videos. Your guidance has assisted me in making choices with my PC builds for future machine learning. I recently updated my 2019/2020 ML desktop with additional memory and a RTX 3090 Ti 24GB FE that I purchase at $1,599.99 in mid-2022 (slightly below MSRP of 1,999.99). Two months later, I was able to purchase a RTX 3090 Ti 24GB FE for $1,099.99 (still available at this price directly from NVIDIA), which I currently use in an eGPU for accelerating GPU performance in a thunderbolt-4 intel laptop (Windows 11 and WSL-2 works great, and now trying to get it to work with native Ubuntu). I’m also currently building a new small form factor (SFF) desktop for my home office, but I’m waiting for the RTX 4090 24GB FE to be available at MSRP (I’m very reluctant to buy anything above MSRP). I feel the RTX 3090 Ti 24GB FE at $1,099.99 and RTX 4090 at MSRP are better choices over the RTX 4080!
@MB-pt8hi
@MB-pt8hi Месяц назад
Are your 3090s second hand? I cannot find any 3090 for that price even now.
@user-en5pp3he7r
@user-en5pp3he7r 11 месяцев назад
Hey! What do you think about the 4070? It also has 12 GB memory and is about 250€ cheaper here (from 960 to 700).
@dragonmaster1500
@dragonmaster1500 3 месяца назад
I'm a beginner in machine learning, currently work as a research assistant for a college in the GIS department. I'm building a personal PC to handle GIS related tasks, with photo editing and gaming as a side benefit, and after having researched for a bit I've noticed that everyone seems to place way more focus on having a high amount of RAM than they do on having a good GPU. Our 'Heavy Processing' computers for example, have 128GB of RAM, but only have a GPU with 8GB of VRAM. For my own build, I'm thinking of starting out with a 4070 Ti Super with 16 GB. I wanted to buy a 3090 Ti, but it's almost double the price ($2400.00 Canadian Vs the 4070Ti Super's $1199.99 Canadian).
@yellowyunicorn
@yellowyunicorn 10 месяцев назад
Thank you for the explanation!
@masterwayne9790
@masterwayne9790 9 месяцев назад
Hey big man I'm getting 3070 or 3080 second hand card for Deep learning. Which one should I choose CPU is 13700K & I can also reduce my build to AM4 and can buy 4070 if it's better than 3070 & 3080 for DL.
@jaymoo5168
@jaymoo5168 8 месяцев назад
Currently a student and looking into this sort of thing. Are we expected to have "this" by the time of employment or later on? Was gonna use high end pc parts but this is clearly industrial parts that's not fully widespread to the mainstream. Thank you.
@aayushgarg5437
@aayushgarg5437 Год назад
@Jeff, one thing more that goes in favour of RTX 40 series is its ability to support FP8 bit. Going forward this year, Nvidia is going to release CUDA 12 with the support of FP8 bits for training and inference. You can only run FP8 bits on RTX 40 series (not on RTX 30). I think that is something one should also consider while buying a GPU. It is better to shed a bit more now so that your GPU remains relevant for next 3-4 years.
@xphix9900
@xphix9900 Год назад
i agree, the memory usage/dependency will change... do you think the 4070ti is a good purchase or shell out more cash for more cuda cores on the 4080/4090 or wait? 4070ti is my budget but im not in a rush
@aayushgarg5437
@aayushgarg5437 Год назад
@@xphix9900 It depends on your use case. If you are buying a GPU for DL then go for RTX 4090 (if your budget allows) solemnly for the VRAM of 24 Gb. You don't want to be in a situation where you face the VRAM bottleneck whenever you are running a bigger network training. My suggestion is wait and go for 24GB card whenever your budget allows.
@xphix9900
@xphix9900 Год назад
@@aayushgarg5437 thank you sir, appreciate the advice, and i agree!
@xphix9900
@xphix9900 Год назад
@@aayushgarg5437 also just to get something running for now to use NERF, would you have a suggestion?
@aayushgarg5437
@aayushgarg5437 Год назад
@@xphix9900 I don't know much about neural 3D rendering. Apologies.
@MalamIbnMalam
@MalamIbnMalam 6 месяцев назад
This is a great video, with that said, I notice that you utilize Windows 11 as your main OS, please correct me if I'm wrong. Have you ever tried a Linux distro? (e.g. Ubuntu or Fedora) with Tensorflow or Pytorch on CUDA? Does it perform better on Linux or on Windows?
@saminmahmud6049
@saminmahmud6049 Год назад
I have MSI Ventus GeForce RTX 3060 12GB. In my opinion, it's a little slow for DL but it gets the job done; takes a while!
@anonimes4005
@anonimes4005 11 месяцев назад
How important is the speed in comparison to ram? I am deciding between a 12gb 3060 and a 24gb tesla p40 for about 250eur, and couldnt really find a lot of information about the tesla.
@bibbsly7055
@bibbsly7055 11 месяцев назад
I was able to buy two zotac 3060s new at a microcenter for 480 bucks for both. It's been great for running 30b models in inference. Training, even LoRA training is painful. I am saving for dual 3090's though!
@FlukeZaoldyeck
@FlukeZaoldyeck Год назад
Thank you for your useful video. I have to decide to buy Gpu for my home lab.
@ProjectPhysX
@ProjectPhysX Год назад
You're missing the most important spec for performance in these kind of applications: VRAM bandwidth. RTX 40/Ada in this regard is abysmal value, and RTX 30 is much better, equally fast or even faster (3080 10GB is 7% faster than 4080), for half the price.
@samuel0705
@samuel0705 Год назад
Hello Jeff, Some AIB GPUs offer water cooling (like the MSI Suprim Liquid X or Gigabyte Waterforce). Supposedly they offer ever so slightly better temperatures. Does that matter much for training models (considering the system could be running consecutively for hours/days)?
@MrLostGravity
@MrLostGravity Год назад
Water-cooling is seldom worth the time/cash investment unless wanting to go for maximum performance while staying on ambient cooling. I'd say it's potentially a 5% performance improvement for 15-25% more cost (relative to the GPU base price). The only application I'm aware where this is useful other than satiate ones need for tinkering is when trying to squeeze maximum performance out of a single GPU for gaming purposes for instance. I don't see the value for ML purposes unless going for absolute performance or if noise is a main concern, since water-cooling tends to be less noisy.
@sluggy6074
@sluggy6074 9 месяцев назад
What's your opinion on the tensor cores? They run in the same price range as the a6000. They are very task specific
@Canna_Science_and_Technology
@Canna_Science_and_Technology 9 месяцев назад
It would be nice to know what GPUs work well with what popular LLM model. Sorry if you’ve already covered this.
@starmountpictures
@starmountpictures 5 месяцев назад
Great video thanks!
@faizalelric
@faizalelric Год назад
what do you think between RTX 3080 ti or 4070 ti?which one is better for DL?
@neoblackcyptron
@neoblackcyptron 8 месяцев назад
I am consulting for a SMB. We are building a strategy map for on-premises vs cloud setup for our machine learning models. I am primarily an ML techie guy, not a business man, so for an SMB what is the better approach. I am assuming once we put money into this fixed cost of setting up a training server for our models we might not have to scale over time.
@gtbronks
@gtbronks Год назад
Thank you from Brasil!
@HeatonResearch
@HeatonResearch Год назад
You are welcome, from the states!
@gonzalodijoux5953
@gonzalodijoux5953 Год назад
Currently I have a ryzen 5 2400g, a B450M Bazooka2 motherboard and 16GB of ram. I would like to use vicuna/Alpaca/llama.cpp in a relatively smooth way. - Would you advise me a card (Mi25, P40, k80…) to add to my current computer or a second hand configuration? - what free open source AI do you advise ? thanks
@atharvahude
@atharvahude Год назад
If possible you can also check for an older Rtx card which offer 24gb memory.
@mfatihaydogdu7
@mfatihaydogdu7 Год назад
Can you also make videos discussing the advantages and disadvantages of TPUs with respect to premium GPUs?
@FzZiO1
@FzZiO1 8 месяцев назад
Hello, your video is incredible, an update is necessary in this regard with the new 4060Ti 16GB graphics card from NVIDIA. Considering the memory model, bandwidth and bus... is it fair to sacrifice all that for more VRAM, are we facing the best graphics for ML? What do you think about it?
@FzZiO1
@FzZiO1 8 месяцев назад
It is fully compatible with cuda
@xntumrfo9ivrnwf
@xntumrfo9ivrnwf Год назад
I've noticed something interesting here in Europe. Used 3090's seem significantly cheaper than I remember them a few months ago (no surprise I guess). It's priced somewhere between a 4070ti and a 4080 (but closer to the 4070). I'm seriously considering getting a used 3090 for that sweet, sweet VRAM. I'm building my 1st PC right now. Deep learning is more a hobby for me - nothing at all related to my day job, etc.
@danieldaniel625
@danieldaniel625 Год назад
same here. I was thinking about getting a 4090 but now I think it'd be nice to save some money
@xntumrfo9ivrnwf
@xntumrfo9ivrnwf Год назад
@@danieldaniel625 for what it's worth, I ended up getting a used 3090. I paid around 800 eur + 40 for buyer protection on the platform I used. Overall, I'm happy with everything!
@ectoplasm
@ectoplasm 10 месяцев назад
@@xntumrfo9ivrnwf Lucky. I went this route and mine was DOA.
@chakrameditation6677
@chakrameditation6677 10 месяцев назад
How do you feel about 4060 TI - 16 GB ? I plan on mainly using it for LLM's ; Text generation. Thank you Jeff.
@powerHungryMOSFET
@powerHungryMOSFET 10 месяцев назад
4070 Ti with 12 GB would be enough for any Deep learning ?
@EssDubz
@EssDubz Год назад
I'm confused about the 2nd assumption about not using server GPUs. Is there any reason why you couldn't put an A100 (for example) in a computer and have it sit on your desk?
@ericp6890
@ericp6890 Год назад
You really need to also consider the TensorFlow and PyTorch compatibility with the 40 series as of now. If one needs to dig right in, then getting 30 series is the right choice for him or her.
@Zekian
@Zekian Год назад
What compatibility issues have you seen with 40 series? I have not yet encountered issues.
@arildboes
@arildboes Год назад
Great stuff. please do an entire video on RTX :D
@tl2uz
@tl2uz 11 месяцев назад
To make two GPU system, is it ok to use different brands of RTX4090? For example, one from Asus 4090 and one from MSI 4090
@perceptoshmegington3371
@perceptoshmegington3371 9 месяцев назад
Managed to get Tensorflow working on my Arc A750 for a fraction of the price an Nvidia card would've cost :)
@Argoo41
@Argoo41 Год назад
Really can agree on VRAM. That's unfortunate that nvidia had a big gap between 3060 and 3080ti. And still, 12gb is not much. Would be great to see more VRAM in future. I saw a rumor of a 3070ti with 16gb, but never found in in the wild. In addition to information from this video, it would be nice to know how memory bandwidth can affect the speed. If you have any benchmarks or info, please share Thanks!
@ProjectPhysX
@ProjectPhysX Год назад
For a lot of these professional applications, performance is directly proportional to VRAM bandwidth and the TFLOPs are well enough to not matter at all, as it's bandwidth-limited. More VRAM is indeed very desirable. I hope we will soon see cards with 96GB GDDR6W, and the leaked H100 PCIe with 120GB HBM.
@andrewjazdzyk1215
@andrewjazdzyk1215 3 месяца назад
Yeah I need at least 48 GBs of vram, any less and I'm like ehhhhhhhhh
@peterxyz3541
@peterxyz3541 Год назад
Hi, any advice on setting up multi GPU, mixing various Nvidia? I’m trying to find a motherboard for 2 or 3 GPU. What should I look for?
@jakestarr4718
@jakestarr4718 3 месяца назад
i'd be looking at my pciex16 slots and if they share lanes. usually port 1 has direct separate lanes to the processor. port 2 and 3 are usually connected or share lanes, for instance pciex16 port 3 is directly connected to processor and port 2 runs through port 3. How i'd use it? port 1 would have my rendering card like a RTX 4080. Port 3 would an A5000 because its tremendously fast, now port 2 i'd probably install a usb 3.2 card. then make a external rack of say 4 NVidia k80's all doubled up and plugged into port 2's usb to pciex16 slot because it will be like giving my A5000 an extra 96gb of ram for like $700. So if you needed a crazy amount of ram to load terabytes of data sets this is what i'd be looking at and my electric bill. At least making the ram boost external gets heat away from my more expensive hardware and i can simply plug in however many stacks i need and power up what i need for the task. Its more of a modular approach to cutting costs with electric and heat. hopefully that gives you a better idea of board architecture and how you'd want to look at it. There is software like octane render that can link all that ram up and make it work together. So looking at it like well i'll just run out and buy the BEST isn't always bright. For instance i can buy 1 4090 but i can buy 2 4080's for the same price and i'd gain 8gb ram and 4k cuda cores and double the speed of processing. Its all perspective and knowing what you'll need and why, compatibility for use might be a major issue too! You might not want a card more for rendering at all as its a mining rig setup or solely a ai computational setup, but i'd say keep in mind that having a rendering card is beyond useful for video editing and a multitude of other tasks that these cards won't be. If you're go doing something insane with lots of heat, shell out the money for a more expensive and heavier board then, anything up too the $500 range on a mobo is going to be a much heavier built board, which means better soldering metals and thicker plating and oh boy does it matter when you're shoving over 5,000 watts through it. Now if its separate entirely from your main pc, like a server build then definitely take a look at the micro boards with multiple processors. Those can actually be setup between a stack and the main controller pc (your rendering or personal use pc). Having the extra processor lanes allows you to direct traffic of the information as you can have 4 processors on one board that are moving information in/out giving a higher efficiency of data transfer through the network you've now built. If you got enough money and understand the architecture, you'll realize you can do anything and the market is a blessing. Other important things to look at would be your processor's ram and its latency for much higher speeds and the ability to use nvme memory cards directly installed to those pciex16 lanes for even faster computing(it makes a massive difference).
@rossrochford1236
@rossrochford1236 10 месяцев назад
What do you think of the following setup in terms of price/performance? Two 3060s and a 3090?
@daspallab772
@daspallab772 10 месяцев назад
rtx 4080 vs a5000 which one would be better for deep learning and Tensor Flow processing ?
@mrjones8257
@mrjones8257 Год назад
Jeff, what do you think about using 6 x 3090 GPUs? For instance, one GPU on the MB (with 16x PCIE lanes) and 5 GPUs connected via risers (1x PCIE lane each) - This setup with a modest amount of ram (16GB) and a middling CPU (affordable I7) Is this a decent configuration? Any ways to improve upon this? Any insight greatly appreciated.
@h-xr1oi
@h-xr1oi 10 месяцев назад
1x pcie does not work in my case, i need at least 8x or the training time will x3 You need to go with epyc, threadripper or xeon to support them. X299, x99 or x399 mobo will be the cheapest.route. e5 v4 cpu is really chip with dual cpu x99 chinese mobo
@Jorvs
@Jorvs Год назад
Does nvda better in AMD in AI? Do have recommendations on AMD brand for AI?
@jamesf2697
@jamesf2697 8 дней назад
I am wondering is it better to buy a 40series or 2 20 or 30 series for ai?
@hybridtron6241
@hybridtron6241 9 месяцев назад
do you think the motherboards that support the new amount of ddr5 ram amount of 192 is worth it ?!
@ALDUIINN
@ALDUIINN 9 месяцев назад
Excellent Video , thanks BTW Could you do a video explaning in detail which should i value more for training? For example , the 4070 has 1/3 less cuda cores than the 3080 Ti, but has double Boost Clock. What should i take in consideration when choosing between the two, i mean, what this difference will represent when training or generating. OH, and please, could you include the bus width on it, that's really something that got me curious. Thanks, wish you the best .
@chsi5420
@chsi5420 2 месяца назад
Newer architecture gets work done faster with less power draw meaning that even with the lower core count, the 4070 will out perform the 3080Ti. I'd recommend the 4070 SUPER as it comes with a higher activated core count and an even higher clock.
@pranjal86able
@pranjal86able Год назад
what is your opinion on 4060 ti 16GB? I want to use the langchain models etc.
@yonnileung
@yonnileung 3 месяца назад
Cost of 4060Ti is much higher than 3060Ti, but the numbers of Cuda core is not worth... I think can consider get a second hand 12GB 3060Ti.
@nitishu2076
@nitishu2076 Год назад
What are your thoughts on Intel arc770 for deep learning? Would you recommend buying it over rtx 3060?
@HeatonResearch
@HeatonResearch Год назад
That is an interesting system, it looks like you could get support for PyTorch through ROCm, though you would be dealing with Linux at that point. If your willing to deal with somewhat more complex installs, and maybe some features not working, it could work out.
@rachitbhatt40000
@rachitbhatt40000 Год назад
Does memory Bandwidth matters for deep learning just like the memory size?
@walter274
@walter274 10 месяцев назад
Where do tensor cores fit in to all of this. I think the operations they accelerate are ubiquitous in statistics. Also i know cuda core count is correlated with tensor cores. I'm starting to explore using my GPU in my statistical work. It's not that what i do is so complex that i need the extra computational power, but I do want to acquire the skill. Thanks.
@bud1239
@bud1239 Месяц назад
I am in a Physics PhD program and I am interested in CUDA coding. I got my 3060 12gb for CUDA coding as a starting point. Got it for $250 new for my pc build so I am pretty happy. I am still working on figuring out how to program in CUDA but I figured out how to program in parallel with Python using my CPU (have a 10 core, i5 12600kf)
@StaMariaRock
@StaMariaRock Год назад
Here one with a 3060 with 12gb, pretty ok, the memory is big enough to work, but I think is getting "slower" comparing to more powerful gpus, I wish I could get a 3080, but not any time soon, but if you are tight in budget and want something that works well, this 3060 is pretty cool to have
@HeatonResearch
@HeatonResearch Год назад
Nice to hear, I always thought that would be a decent entry point GPU.
@amiga2091
@amiga2091 Год назад
How do I set all this up? Multi gpu rig. What operating system? Ubuntu? 1-3090 and 3-3060.
@issamu2k
@issamu2k Год назад
Is the 4070 better for ML than the 4080 TI?
@durocuri1758
@durocuri1758 7 месяцев назад
Now it is October, i am training sdxl model. If i want to train high pixel pictures, it need 48gb or high. Is it still deserve to buy a6000 sli extend it vram to 96gb now? Or just rent a a100 on cloud😂
@w4r10rd
@w4r10rd 10 месяцев назад
is 4070 good for deep learning tasks?? i got one recently
@triplethegreatdali4238
@triplethegreatdali4238 8 месяцев назад
"Hello, Sir. I am learning deep learning and have worked on many projects in the lab's GPU system. I want to build my personal setup within a budget of 2,000 USD. Could you please suggest the best components for it? (Excluding: Monitor, Keyboard, and Mouse)." Thank you.
@tuttocorsivo3558
@tuttocorsivo3558 Год назад
I got a 3060 12gb and it's working well for my beginner projects. In the future I would like to work on bigger projects but speed of execution will not be crucial will a couple of tesla p40 refurbished be a valid choice? (Ga102 chip , 3840 cores ,24 gddr5x memory) and I've noticed that once In a while you can get them for a good price since they are discontinued.
@dongyangli3985
@dongyangli3985 Год назад
No, that's a really old toy and have no image display.(compute compality 6.1)
@tuttocorsivo3558
@tuttocorsivo3558 Год назад
@@dongyangli3985 sorry I don't understand why I need image display since I have already one in my system.
@krishnateja4291
@krishnateja4291 9 месяцев назад
I have a below configuration is this enough to start: I5 12400f 6 core 12 thread rtx 4060 ti 8gb 32gb ddr4 3200mhz Pcie 4 nvme ssd
@kevindunlap2832
@kevindunlap2832 Месяц назад
I have an NVIDIA T1000 with 8gb. Is that enough for learning ML?
@jpsolares
@jpsolares 10 месяцев назад
rt core? are good for machine learning?
@xtrading888
@xtrading888 Год назад
I have several 3080ti and one 3070, but their memory is low compared to 3090. Are they not suitable for AI?
@InnocentiusLacrimosa
@InnocentiusLacrimosa 3 месяца назад
Used 3090 s look very nice. Even getting 2 of those for getting a home setup that can run even llama 70b type of models.
@flightsimdev9021
@flightsimdev9021 9 месяцев назад
I have an old Quadro M6000 12gb variant, is this good enough for AI ?
@syuxtun1734
@syuxtun1734 День назад
I'm new to this field. but i am very interested because i have been an opportunity to learn more in the I.T Field.
@kodatech1263
@kodatech1263 Год назад
Good to look at the second hand market for 3090 if you are looking the maximize vram. At this time you can get a 3090 used < 1k and they are smaller in size compared and power requirements to the 4090, so you can try running dual 3090. Also, one thing to consider with Hopper (updated upcoming version of A6000) is the support for FP8 datatype and transformer_engine.
@ProjectPhysX
@ProjectPhysX Год назад
RTX 3090 seems the best value by far. Large 24GB VRAM which is 93% as fast as the 3090 Ti/4090, but 100W/22% less wasted power and half the cost of a 4090.
@denysivanov3364
@denysivanov3364 Год назад
4090 provides very good value, forget about 3090. 4090 also much better for transformers inference.
@ProjectPhysX
@ProjectPhysX Год назад
@Zach absolutely not. Especially since VRAM capacity and bandwidth are the same. 3090 is the better option.
@denysivanov3364
@denysivanov3364 Год назад
@@ProjectPhysX 4090 is faster than a100 for inference, especially for transformers.
@alvarorodriguezgomez8716
@alvarorodriguezgomez8716 6 месяцев назад
do you have experience with multi gpu training? is it easy to put them to work together as 1? I dont have experience with it, but since in pytorch you have to move the tensors around im not sure how easy will be to train a network in several gous at once even though pytorch offer managemetn for it themselves
@pedrazzi65
@pedrazzi65 Год назад
Hi Jeff! Can you tell me if RTX A4000 is good for machine learning? I'm new on your channel and I've been enjoying it a lot. All the best
@HeatonResearch
@HeatonResearch Год назад
Yes, absolutely, 16GB ram and 6k CUDA cores. It would be a fine card for desktop machine learning.
@DJpiya1
@DJpiya1 Год назад
Thank you very much for this valuable content. What do you think about Apple silicon. We can buy an Apple Mac studio M1 ultra with 64GB RAM under $5000. So then, we don't bump into these RAM issues as well. Do you think whether it is a good option to consider for Deep Learning.
@generalshepherd457
@generalshepherd457 Год назад
motherboard ram can't be utilized for gpu work loads
@DJpiya1
@DJpiya1 Год назад
@@generalshepherd457 With M1 Apple silicon we can, coz the RAM is in the SOC and not on the motherboard.
@generalshepherd457
@generalshepherd457 Год назад
@@DJpiya1 sounds skethcy. would not trust tim crook.
@HeatonResearch
@HeatonResearch Год назад
The M1/M2 does share GPU and system RAM, I've observed this on my own Mac.
@diazjubairy1729
@diazjubairy1729 4 месяца назад
How important the memory buss width for ML and DL ?
@JingsongLin
@JingsongLin Год назад
Does upgrade from 4080 to 4090 worth the money?
@saribmalik985
@saribmalik985 19 часов назад
Can I use a ryzen 5 3600 in biggener deep learning and ai It's weak but it allows me to get a 3060 in my budget
@ihummingbird
@ihummingbird 5 месяцев назад
I wish I had the ability to buy a higher end desktop GPU, Right now I can only afford a laptop with rtx 4070 8Gb, I hope it suffices for the near future as I don't hve any other options. Thanks for the video.
@yearofthegarden
@yearofthegarden 7 месяцев назад
I bought a 3060 12gb a half year ago for $180 on a bid war at 3am in the morning, glad i did even though they are all about $225 now. Great card even while running one uktra wide monitor in 1440p.
@prolificgainz
@prolificgainz Год назад
Started the video thinking about buying a 3070ti. Now I'm honestly thinking about going with a 3080ti. Little shitty that the 3070ti couldn't at least have 10gb. But the price difference is 250USD so I'm not too sure yet.
@StephenNewmanUK
@StephenNewmanUK Год назад
Thanks!
@HeatonResearch
@HeatonResearch Год назад
Oh wow, you are totally awesome! Thank you so much!
@jacobhatef3156
@jacobhatef3156 Год назад
Any thoughts on using 2 3090s with NVLink, or 2 4080s (roughly the price of a 4090)? My budget is about 2-2.4k, and I'd like to buy the best most versatile thing I can afford. 2 3090s or 2 4080s double your VRAM, which is something to consider.
@bigdreams5554
@bigdreams5554 Год назад
3090 all day every day, in my opinion. As far as nvlink... Not sure how well that works for pytorch etc. (I think the professional cards are the ones worth the drivers to take advantage of that). Definitely do more research about linking the GPUs. Make sure you have a good PSU (1600w) if u have multiple 3090s in a box.
@HeatonResearch
@HeatonResearch Год назад
Good points bigdreams!
@jacobhatef3156
@jacobhatef3156 Год назад
@@bigdreams5554 if nvlink can’t work should I switch to something else? 3090ti, 4080, etc? Are the amd gpus even worth considering?
@harrythehandyman
@harrythehandyman Год назад
no NV Link or P2P over PCI-E support on 4080/4090. The last consumer card that support NV Link and P2P is 3090. If your model need VRAM, 3090 x2 is better. If your model fits within 24GB, single 4090 is probably close to dual 3090 performance.
@Canna_Science_and_Technology
@Canna_Science_and_Technology 5 месяцев назад
Thanks for this video. I was given $50,000 to build an AI Rig and I'm like a kid in a candy store.
@Srcfrvr
@Srcfrvr Месяц назад
"Given" ?😭
@testales
@testales Год назад
If you want to play around with stable diffusion, I'd recommend to go at least for a 3090 because of its 24GB RAM.
@danilo_88
@danilo_88 Год назад
3060 12gb is more than enough to play with SD, unless you want to train large models
@nailsonlandim
@nailsonlandim Год назад
When I replace my current one, I'll get a 3060. At least for Brazilian standards is a sweet spot and can do wonders.
@kestasjk
@kestasjk 10 месяцев назад
I don't get it why did you eliminate the server GPUs right out of the gate? I was hoping you would compare the K80 which has 24GB of VRAM, the only way you can get that much VRAM on an RTX/GeForce is with a 4090. I bought one for an LLM I needed to run before I realized these cheap K80s were a possibility.
@jreamer0
@jreamer0 5 месяцев назад
K80s have 1/3 the CUDA Cores
@slashernunes
@slashernunes Год назад
Thanks for the video! I own an RTX 3060 12gb, but I'm concerned I'll run out of memory as I experiment with deeper networks. Is it worth to buy another RTX 3060 12gb to run it in parallel? Or should I consider selling a kidney to buy an RTX with 24gb?
@dongyangli3985
@dongyangli3985 Год назад
Absolutely ture. And 3090 may be a good choice.
Далее
We need to have a chat about these 4090 Benchmarks...
23:36
The Cutest Gaming Setup
15:08
Просмотров 379 тыс.
Bardak ile Projektör Nasıl Yapılır?
0:19
Просмотров 3,3 млн
Плохие и хорошие видеокарты
1:00