Тёмный

The Fundamental Problem with Neural Networks - Vanishing Gradients 

ritvikmath
Подписаться 166 тыс.
Просмотров 13 тыс.
50% 1

Why vanishing gradients are the biggest issue with neural networks.
My Patreon : www.patreon.co...
Intro to Neural Networks : • Intro to Neural Networ...
Backpropagation : • Backpropagation : Data...
The Sigmoid : • The Sigmoid : Data Sci...

Опубликовано:

 

3 окт 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 30   
@HoHoHaHaTV
@HoHoHaHaTV Год назад
My respect for Ritvik goes exponentially high each time I see his explanations. He can beat any prof when it comes to explaining these things. I just feel so lucky to have come across this channel
@erickmacias5153
@erickmacias5153 2 года назад
I just found this channel like 3 days ago and has been very useful and interesting! Thank you very much
@ritvikmath
@ritvikmath 2 года назад
Of course!
@shubhampandilwar8448
@shubhampandilwar8448 2 года назад
Very well explained. I alwayd had trouble understanding this topic, but this video helped me to comprehend the topic intuitively.
@ritvikmath
@ritvikmath 2 года назад
Thanks!
@siddhantrai7529
@siddhantrai7529 2 года назад
Very well addressed, thank you for the video.
@ritvikmath
@ritvikmath 2 года назад
Thanks!
@geoffreyanderson4719
@geoffreyanderson4719 2 года назад
Yes, great topic. Absolutely, some top ways to fight off vanishing gradients are relu (and other advanced activation functions), and residual nets (skip nets). It's also quite possible to add your own custom resnets to any deep network; it's not neessary to use only the resnet blocks that the framework tool provides. Tensorflow's functional api makes it pretty straightforward to add skip layers plus the necessary aggregation layer to combine the main path plus the skip path, to layer types other than convolutional and computer vision specific. So while resnet was originally designed with compujter vision purpose, it's not married to that at all. Additional solid help to fight those vanishing gradients off, are - batch normalization, which basically conditions the signal to the next hidden layer; and - smarter initialization of weights in all your layers, like He Initialization when using Relu (and other initializations that are suited to other activation functions).
@CptJoeCR
@CptJoeCR 2 года назад
Love it as always! May I suggest a future video topic: Bayesian Change Point Detection. BCP has so many components that you already have covered (sampling techniques, MCMC, Bayesian statistics) that I think it would make for a great video! (and I'm still slightly confused how it all comes together in the end! lol)
@tomdierickx5014
@tomdierickx5014 2 года назад
Another gem! Great insights! 🔬
@ChocolateMilkCultLeader
@ChocolateMilkCultLeader 2 года назад
You can also add some bias to your networks in one of the intermediate layers
@arontapai5586
@arontapai5586 2 года назад
Very informative!!! Are you planning to make videos on RNN (LSTM...) and other type of network models?
@ritvikmath
@ritvikmath 2 года назад
Yup they'll likely be coming out within the next month!
@pushkarparanjpe
@pushkarparanjpe Год назад
Some questions inspired by your video. - Earliest layers see the severest form of vanishing gradient. Do later layers undergo vanishing gradient sequentially? - So what if the earliest layer weights get stuck; learning can still happen due to weight updates at later layers, right ? - Can we use vanishing gradient for neural architecture depth search ? Start with many layers; train; identify the early layers that got stuck; discard them and keep a shallower network. This sounds like there is something wrong with it; will this work ?
@shreypatel9379
@shreypatel9379 2 года назад
One of the best channels i've found on youtube (along the lines of 3Blue 1Brown and rest such channels). Keep up the good work
@listakurniawati8946
@listakurniawati8946 2 года назад
Omg thank you so much!!! You saved my thesis ❤❤
@marvinbcn2
@marvinbcn2 2 года назад
Excellent video. You perfectly convey the intuition. Only one doubt left: I cannot see why ReLu is a good solution, given that gradient vanishes to 0 in case of negative values. How do we compute backpropagation then?
@posthocprior
@posthocprior 2 года назад
Excellent explanation.
@christophersolomon633
@christophersolomon633 2 года назад
Outstanding Video. Really well explained.
@pushkarparanjpe
@pushkarparanjpe Год назад
Thanks once again !
@MachineLearningStreetTalk
@MachineLearningStreetTalk 2 года назад
Awesome video! Nice channel
@seeking9145
@seeking9145 2 года назад
Super nice explanation!!!
@ChocolateMilkCultLeader
@ChocolateMilkCultLeader 2 года назад
Never heard of someone call it the most important problem. Interesting view point.
@geoffreyanderson4719
@geoffreyanderson4719 2 года назад
Great depth is where you get the most exponentiation effect, thus the worst vanishing or explosion. But great depth is where the bulk of the power of deep neural nets comes from.
@codematrix
@codematrix Год назад
Another way to help solve the vanishing gradient is to adjust your learning rate.
@ritvikmath
@ritvikmath Год назад
Great tip!
@n.m.c.5851
@n.m.c.5851 Год назад
thanks
@ritvikmath
@ritvikmath Год назад
Welcome
@aliasgher5272
@aliasgher5272 2 года назад
Maan gaya Sir G
@xiaoweidu4667
@xiaoweidu4667 Год назад
as always
Далее
The Unreasonable Effectiveness of Bayesian Prediction
15:03
Recurrent Neural Networks : Data Science Concepts
27:17
The Most Important Algorithm in Machine Learning
40:08
Просмотров 447 тыс.
Backpropagation : Data Science Concepts
19:29
Просмотров 37 тыс.
Conditional Random Fields : Data Science Concepts
20:11
Intro to Neural Networks : Data Science Concepts
13:14
The ROC Curve : Data Science Concepts
17:19
Просмотров 35 тыс.
SVM Dual : Data Science Concepts
15:32
Просмотров 48 тыс.
Watching Neural Networks Learn
25:28
Просмотров 1,3 млн
How AIs, like ChatGPT, Learn
8:55
Просмотров 10 млн