Тёмный

Neural Networks Pt. 2: Backpropagation Main Ideas 

StatQuest with Josh Starmer
Подписаться 1,3 млн
Просмотров 523 тыс.
50% 1

Опубликовано:

 

24 сен 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 556   
@statquest
@statquest 3 года назад
The full Neural Networks playlist, from the basics to deep learning, is here: ru-vid.com/video/%D0%B2%D0%B8%D0%B4%D0%B5%D0%BE-CqOfi41LfDw.html Support StatQuest by buying my book The StatQuest Illustrated Guide to Machine Learning or a Study Guide or Merch!!! statquest.org/statquest-store/
@motherisape
@motherisape 2 года назад
Bamm
@gbchrs
@gbchrs 2 года назад
@seanleith5312
@seanleith5312 Год назад
Quit the singing, please
@statquest
@statquest Год назад
@@seanleith5312 Noted
@dcscla
@dcscla 3 года назад
Man, your promotions are not shameless! Actually, what you do is a gift for us, for the price that you charge and for the level of the content, we are being gifted and not a buying something. You are far better than a lot of paid (and expensive) courses. Just check out your video comments to see how people few happy when they discover your videos!! Great work as always. Thank you so much!!!👏🏻👏🏻👏🏻👏🏻
@statquest
@statquest 3 года назад
Thank you very much! :)
@Luxcium
@Luxcium Год назад
He is using the concept of reverse psychology by presenting great stuff at a good price and as you mentioned theses promotions are not shameless… They are shameful, as you hinted he should indeed be ashamed of giving us such a good and advantageous offer… 😅😅😅😅
@jonforce360
@jonforce360 3 года назад
You released this video just in time for my AI exam! Thank you. Sometimes I think professors use really complex notation just to feel smarter than students, it doesn't help learning. I love your content.
@statquest
@statquest 3 года назад
Thank you very much!
@sarazahoor9133
@sarazahoor9133 2 года назад
I want to copy-paste this comment! :D
@puppergump4117
@puppergump4117 2 года назад
Ain't that right. They must be mad that they don't understand the actually smart people so they don't want to be understood either.
@zhongtianjackwang5346
@zhongtianjackwang5346 Год назад
lol, that is exactly what I want to say
@ElNick09
@ElNick09 2 года назад
I have been a student my entire life and have taught college level courses myself, and I must say you are one of the finest lecturers I have ever seen. This statquest is a gem. Your work is so succinct and clear its as much art as it is instruction. Thank you for this incredible resource!
@statquest
@statquest 2 года назад
Thank you very much! :)
@juliocerono_stone5365
@juliocerono_stone5365 2 месяца назад
I am already 65, and your videos have helped me understand the basics behind NN. Thank you so much!!!!
@statquest
@statquest 2 месяца назад
Bam! :)
@iskrega
@iskrega 2 года назад
I just want you to know your channel has been instrumental in helping me towards my Data Science degree, I'm currently in my last semester. I'll be forever grateful for your channel and the time you take to make these videos. Thank you so much.
@statquest
@statquest 2 года назад
Thank you and good luck with your final semester! BAM! :)
@hamzasaaran3011
@hamzasaaran3011 Год назад
I am studying for a Master's degree in bioinformatics now, and as someone who knows little about statistics, I really can't thank you enough for your videos and the effort that you have put into them.
@statquest
@statquest Год назад
Thank you!
@akeslx
@akeslx 11 месяцев назад
I finished business school 25 years ago where I studied statistics and math. So happy to see that neural networks are fundamentally just a (much) more advanced regression analysis.
@statquest
@statquest 11 месяцев назад
BAM!!! Thank you for supporting StatQuest! Yes, neural networks are a lot like regression, but now we can fit non-linear shapes to the data, and we don't have to know in advance what that shape should be. Given enough activation functions and hidden layers, the neural network can figure it out on its own.
@katwoods8514
@katwoods8514 3 года назад
omg yay! I just discovered that you've made a million videos on ML. I'm going to go binge all of them now :D
@statquest
@statquest 3 года назад
Hope you enjoy!
@motherisape
@motherisape 2 года назад
Bamm
@ksrajavel
@ksrajavel 4 года назад
Finally. The wait is overBAM!!!
@statquest
@statquest 4 года назад
TRIPEL BAM!!!
@mariolira9279
@mariolira9279 3 года назад
F I F T H B A M!
@syco_Rax
@syco_Rax 3 года назад
SUPER BAM!!!
@edphi
@edphi 10 месяцев назад
It should be made a crime for anyone to see other videos on backpropagation before they reach Statquest. The world is confused by teachers who tell the big story before the basic. Learn the basic and the picture fall into place like the chain rule 😊
@statquest
@statquest 10 месяцев назад
bam! :)
@evangeliamm
@evangeliamm 3 года назад
You have no idea how much I appreciate your work. Your explanations are so fun and simple, I'm just so grateful!
@statquest
@statquest 3 года назад
Thank you very much! :)
@ML-jx5zo
@ML-jx5zo 3 года назад
Now Iam reading backpropagation, I worried about this vedio didn't came for long time , And finally I got a treasure.
@statquest
@statquest 3 года назад
bam! :)
@katwoods8514
@katwoods8514 3 года назад
Love this! You've explained it far better than anywhere else I've seen, and you made it entertaining at the same time! Thank you so much for making this.
@statquest
@statquest 3 года назад
Awesome, thank you!
@advaithsahasranamam6170
@advaithsahasranamam6170 Год назад
This is excellent stuff! As a visual learner, your channel is a BLESSING. Thank you so much for your fantastic work on breaking down concepts into small, bite-sized pieces. It's much less intimidating, and you deserve so much more appreciation . You also gained my subscription to your channel! Keep doing a great job, and thank you SO MUCH for having my back!
@statquest
@statquest Год назад
Thank you very much!!! :)
@raunak5344
@raunak5344 7 месяцев назад
I just iterated on a gradient descent and found that this is the best possible way to teach this topic and no other lecture in the entire existence is better than this one
@statquest
@statquest 7 месяцев назад
bam!
@yasameenmohammed4366
@yasameenmohammed4366 Год назад
My Machine Learning exam is tomorrow and re-watching your videos to review concepts is helping me so much! Thank you!!!
@statquest
@statquest Год назад
Good luck! BAM! :)
@babarali4313
@babarali4313 2 года назад
Its the teacher who makes the Subject easy or difficult and the way you explained Neural Network, I am speechless
@statquest
@statquest 2 года назад
Thanks!
@jays9591
@jays9591 2 года назад
May I say .... You are such a good teacher that it is most enjoyable to watch your videos. I am proficient in statistics (via university econometrics 101) ... and I did not realise all those fancy terms in machine learning are actually concepts that are common items in the stats that I learned in the 1970s, e.g., biases and weights, label, activation functions etc. Anyway, I can see that a lot of viewers appreciate your work and teaching. I have also 'updated' myself. Thank you.
@statquest
@statquest 2 года назад
Thank you very much!
@mot7
@mot7 3 года назад
You are the best. I wish every ML learner find you first. I am going to do my part and tweet about you. Thanks for making these videos! Wish you more success.
@statquest
@statquest 3 года назад
Wow! Thank you very much! I really appreciate the support. BAM! :)
@Amir-gc8re
@Amir-gc8re 2 года назад
Finally a proper, detailed, step by step explanation. This guy is absolutely AMAZING ! Thank you so much for all the hard work in putting these videos together for us.
@statquest
@statquest 2 года назад
Thank you very much! :)
@mohammadrahman1126
@mohammadrahman1126 3 года назад
Amazing explanation! I've spent years trying to learn this and it always went too quickly into the gory mathematical details. Aha moment for me was when green squiggle equal blue plus orange squiggles lol Thank you for this Josh!!!
@statquest
@statquest 3 года назад
Glad it was helpful!
@amandak1396
@amandak1396 3 года назад
Kind of like how Feyman reduced gory math in physics to actual squiggle, double bam!
@mashmesh
@mashmesh 3 года назад
Omg, protect this man at all costs, this was pure gold!!! Also, thank you, sir, for talking so slowly because if my brain squiggles need to work faster they will burn up x)
@statquest
@statquest 3 года назад
Glad you enjoyed it!
@hyonnj9563
@hyonnj9563 6 месяцев назад
Honestly you do a much better job at teaching using a pre recorded video than my instructors using both written and live materials that I'm paying for.
@statquest
@statquest 6 месяцев назад
I'm glad my videos are helpful! :)
@madghostek3026
@madghostek3026 Год назад
9:00 at this moment I realised I'm watching the best math content on earth, because you never see simple stuff like this being given attention to. Luckily I already know how summation symbol works, but I didn't know it in the past, and nobody cared to explain. But it's just not about the summation symbol, imagine the other 1000 small things somebody might not understand, and doesn't realise they don't understand, because it's been skimmed over
@statquest
@statquest Год назад
Thank you so much! I really appreciate it! :)
@subusrable
@subusrable 3 месяца назад
this video is a gem. I had to watch it a few times and like in gradient descent, I went closer to the target level of knowledge with each step :)
@statquest
@statquest 3 месяца назад
BAM! :)
@TheClearwall
@TheClearwall 3 года назад
Who else is using these videos to put together a semester project? So far, I've put Regression Trees, K-fold CV, complexity pruning, and now Neural networks as my final model construction. Josh is worth a double bam every time.
@statquest
@statquest 3 года назад
BAM! Good luck with your project.
@jennystephens3215
@jennystephens3215 3 года назад
Josh, this is amazing. You really make things so easy to visualise which is crazy considering the hidden networks are meant to be so hard that they are referred to as black box! Thanks for all your videos. I have used heaps over the last twelve months. Thank you again.
@statquest
@statquest 3 года назад
Hooray!!! I'm so glad that you like my videos. :)
@tagoreji2143
@tagoreji2143 2 года назад
Teaching such complicated topics in a simple, Easily Understandable way.👏👏👏.Thank you, Professor
@statquest
@statquest 2 года назад
Thanks!
@ABCEE1000
@ABCEE1000 9 дней назад
have no idea how can i thank you as you deserve .. thank you so much
@statquest
@statquest 9 дней назад
Thanks! :)
@edrobinson8248
@edrobinson8248 2 месяца назад
simply brilliant. Learning is indeed a quest. A quest for someone who understands and can present understandably. Thanks.
@statquest
@statquest 2 месяца назад
Thanks!
@manalisingh1128
@manalisingh1128 2 года назад
Wow Josh way to go!!!! You have the concepts so clear in your own head that it seems a piece of cake for us 🍰♥️ Love from India! 🇮🇳
@statquest
@statquest 2 года назад
Thanks so much!!
@user-re1bi2bc8b
@user-re1bi2bc8b 3 года назад
Incredible. Sometimes I need a refresher on these topics. There’s much to remember as a data scientist. I’m so glad I found your channel!
@statquest
@statquest 3 года назад
Bam!
@BlochSphere
@BlochSphere 7 месяцев назад
The level of detailing in this video is just 🤯 I hope i can try to make my Quantum Computing videos this clear!
@statquest
@statquest 7 месяцев назад
Good luck!
@sarazahoor9133
@sarazahoor9133 2 года назад
For the first time ever in history, I have understood the concept behind Neural Networks! BAM!!!! :D Thanks Josh, so grateful :)
@statquest
@statquest 2 года назад
BAM! :)
@dinara8571
@dinara8571 3 года назад
JUST WOW! Thank you so much, Josh! I cannot express the feeling I had when EVERYTHING made sense!!! TRIPLE BAM! Never thought I would be extremely excited to pause the video and try to solve everything by hand before I look at the next steps
@statquest
@statquest 3 года назад
BAM! :)
@NadaaTaiyab
@NadaaTaiyab 2 года назад
oh that's a good idea!
@codeman2
@codeman2 3 года назад
I searched neural net and again your video popped that too just 4 month old, love to get your helpful videos right before my semester
@statquest
@statquest 3 года назад
:)
@O5MO
@O5MO 2 года назад
I never understood backpropagation. I knew some things from other tutorials, but as for beigginer, it was very hard to understand. This video (and probably series) is the best i could find. Thank you.
@statquest
@statquest 2 года назад
Glad it was helpful!
@wong4359
@wong4359 2 года назад
I found your explanation is far more easier to understand than the edx online course I am taking, BAM !!!
@statquest
@statquest 2 года назад
bam!
@AhmadAbuNassar
@AhmadAbuNassar 2 месяца назад
Thank you very much for this comprehensive yet simple explanation
@statquest
@statquest 2 месяца назад
Glad it was helpful!
@jblacktube
@jblacktube Год назад
I didn't even get through the jingle before I gave a thumbs up. Thanks for the chuckle, can't wait to watch the rest of this!
@statquest
@statquest Год назад
BAM! :)
@lukasaudir8
@lukasaudir8 Год назад
I am really glad that people like you exist!! Thank you so much for those incredible lessons
@statquest
@statquest Год назад
Glad you like them!
@perhaps467
@perhaps467 Год назад
Thank you so much for this series! I haven’t been able to find any other videos that really break down the mechanics of neural networks like this.
@statquest
@statquest Год назад
Thanks!
@iliasaarab7922
@iliasaarab7922 3 года назад
Best explanation that I've seen so far on backpropagation!
@statquest
@statquest 3 года назад
Thank you! :)
@maliknauman3566
@maliknauman3566 2 года назад
How amazing is the way you convey complex concepts.
@statquest
@statquest 2 года назад
Thank you!
@ucanhnguyen4751
@ucanhnguyen4751 3 года назад
Thank you for this video. I have been waiting for this all the time. Finally, it appeared just 1 day before my exam. You are a life saver!!
@statquest
@statquest 3 года назад
Good luck with your exam! :)
@evie389
@evie389 Год назад
I was reading an article based on Backpropagation and I did not understand a single word. I had to watch all your videos starting from Chain Rule, Gradient Descent, NNs...I re-read the article and understood everything!!! But now I can't get the beep--boop and small/double/triple/ bam out of my head lol.
@statquest
@statquest Год назад
BAM! I'm glad my videos were helpful! :)
@terrepus9856
@terrepus9856 3 года назад
The time couln't be more perfect ... 3 hours before my machine learning exam !! Thank you!!!!
@statquest
@statquest 3 года назад
Good luck with your exam! I hope it goes well.
@mjcampbell1183
@mjcampbell1183 2 года назад
Wow! This is an incredible video. Thank you SO MUCH for making this for us. This is one of the best videos I've seen to explain this concept. The hard work you have put into this is something that I am incredibly appreciative of. Thanks, man.
@statquest
@statquest 2 года назад
Wow, thank you!
@David5005ful
@David5005ful 2 года назад
The type of in depth video I’ve always wanted!
@statquest
@statquest 2 года назад
Thank you!
@mrglootie101
@mrglootie101 3 года назад
I've been waiting for this all the time checking the notification haha
@statquest
@statquest 3 года назад
Hooray! The wait is over.
@deepanjan1234
@deepanjan1234 3 года назад
This is really awesome. I thank you for your effort in developing this highly enriched content. BAM !!!
@statquest
@statquest 3 года назад
Thank you!
@aryabartarout5697
@aryabartarout5697 11 месяцев назад
You have cleared my doubt on back propagation, gradient descent and chain rule. Triple Bam !
@statquest
@statquest 11 месяцев назад
:)
@nojoodothmanal-ghamdi1026
@nojoodothmanal-ghamdi1026 Год назад
I . JUST . LOVE . YOUR . CHANNEL !! you literly explain things very clearly and step by step! I just cannot thank you enough really
@statquest
@statquest Год назад
Wow, thank you!
@viethoalam9958
@viethoalam9958 5 месяцев назад
give respect to my math teacher, but this is so much easier to understand.
@statquest
@statquest 5 месяцев назад
bam! :)
@CHERKE_JEMA5575
@CHERKE_JEMA5575 2 года назад
You rescued me from the unknown!! Much Love from Ethiopia
@statquest
@statquest 2 года назад
Bam! :)
@yiliu5403
@yiliu5403 Год назад
Best Neural Networks Lectures! Just ordered the book from Amazon to support!
@statquest
@statquest Год назад
Wow! Than you very much! :)
@aminmoghaddam7624
@aminmoghaddam7624 6 месяцев назад
I wish our lecturers watched these videos before trying to make their own teaching slides! (With acknowledgement of course!)
@statquest
@statquest 6 месяцев назад
bam!
@knt2112
@knt2112 Год назад
Hello sir, Thanks for such an simple explanation, never understood back propagation in such a depth at this ease. 🎉
@statquest
@statquest Год назад
Thank you!
@epistemophilicmetalhead9454
@epistemophilicmetalhead9454 10 месяцев назад
Back propagation (aka finding w's and b's) start with b_final=0. you'll notice that error = (actual - predicted)^2 is really high. so you find the gradient descent of squared error wrt b_final and find out the value of b_final for which the squared error is minimum. that is your optimal b_final. gradient descent: derivative of sum of squared errors wrt b_final = derivative of sum of squared errors wrt predicted value y * derivative of y wrt b_final. d(y observed - y predicted)^2/d(y predicted) = -2*(y observed - y predicted) d(y predicted)/d(b_final) = d(sum of all those previous curves obtained through each node of the layer + b_final)/d(b_final) = 0+0+0....+0+1=1 take the predicted curve ke x values and find the derivative/slope. step size = slope*learning rate. new b_final = old b_final - step size. keep repeating until slope touches 0. this is how gradient descent works and you've found your optimal b_final.
@statquest
@statquest 10 месяцев назад
double bam
@josephif
@josephif Год назад
Lecture was awesome,more affective and easy to understand Thanks
@statquest
@statquest Год назад
Thank you! :)
@royazullay7556
@royazullay7556 Год назад
That Josh guy is just awsome !! Definitely will support !!
@statquest
@statquest Год назад
Thank you!
@willw4096
@willw4096 Год назад
Thanks for the great video!! My notes: 7:23 8:11 8:48 10:00 10:22❗,11:13 - 11:48, 11:56 12:08 13:30❗,
@statquest
@statquest Год назад
BAM! :)
@chrislee4531
@chrislee4531 2 года назад
I learn more from four of your videos than 200 pages of textbook gibberish
@statquest
@statquest 2 года назад
Thanks!
@superk9059
@superk9059 2 года назад
Thank you very much for your video~ Your videos make me feel that studying English make so much sense, otherwise I can't enjoy such beautiful thing~ 👍👍👍❤❤❤
@statquest
@statquest 2 года назад
WOW! Thank you very much!!! And thank you for your support!!! :)
@aviknash
@aviknash 3 года назад
Excellent job Josh!!! Just loved it!!! Thanks a ton for your fun-filled tutorials :)
@statquest
@statquest 3 года назад
Glad you like them!
@SPLICY
@SPLICY 3 года назад
The understated BAM at 4:40 cracked me up 😂
@statquest
@statquest 3 года назад
SPLICY in the house!!! BAM! :)
@d_polymorpha
@d_polymorpha 9 месяцев назад
Hello, thank you for the video! This series has been really helpful to learn about deep learning. I have a couple of questions. 1. When using gradient descent and backpropagation, do we always use SSR to measure how good a fit the parameter we are estimating is? Or are there other ways? 2. The second question is when using chain rule for calculating derivatives. The first part is d SSR/ d Predicted. In that first part @ 11:25 are you using chain rule again within that first part? And when deriving the inside Observed - Predicted @ 11:34 where do you get 0 and 1 from?
@statquest
@statquest 8 месяцев назад
1. The "loss function" we use for gradient descent depends on the problem we are trying to solve. In this case, we can use the SSR. However, another commonly used "loss function" is called Cross Entropy. You can learn more about cross entropy here: ru-vid.com/video/%D0%B2%D0%B8%D0%B4%D0%B5%D0%BE-6ArSys5qHAU.html and ru-vid.com/video/%D0%B2%D0%B8%D0%B4%D0%B5%D0%BE-xBEh66V9gZo.html 2. You can learn how the chain rule works (and understand the 0 and 1) here: ru-vid.com/video/%D0%B2%D0%B8%D0%B4%D0%B5%D0%BE-wl1myxrtQHQ.html
@douglasespindola5185
@douglasespindola5185 2 года назад
If statistics were a religion, Josh would be it's pope.
@statquest
@statquest 2 года назад
bam!
@alexissanchezbro
@alexissanchezbro 3 года назад
Your getting better and better. Thank you
@alexissanchezbro
@alexissanchezbro 3 года назад
BAAAAAAAMMM
@statquest
@statquest 3 года назад
:)
@jieunboy
@jieunboy Год назад
insane teaching quality, thanks !
@statquest
@statquest Год назад
Glad you think so!
@VishalKhopkar1296
@VishalKhopkar1296 Год назад
you taught this better than professors at CMU, not kidding
@statquest
@statquest Год назад
Thank you! :)
@Luxcium
@Luxcium Год назад
Wow 😮 I didn't knew I had to watch the *Gradient Descent Step-by-Step!!!* before I can watch the video related to *Neural Networks part 2* that I must watch before I can watch the *The StatQuest Introduction To PyTorch...* before I can watch the *Introduction to coding neural networks with PyTorch and Lightning* 🌩️ (it’s something related to the cloud I understand) I am genuinely so happy to learn about that stuff with you Josh❤ I will go watch the other videos first and then I will back propagate to this video...
@statquest
@statquest Год назад
Getting warmer...
@constantthomas3830
@constantthomas3830 3 года назад
Thank you from France
@statquest
@statquest 3 года назад
Merci! :)
@zombieeplays3146
@zombieeplays3146 5 месяцев назад
I come to this channel for the intros tbh!
@statquest
@statquest 5 месяцев назад
bam! :)
@xuantungnguyen9719
@xuantungnguyen9719 3 года назад
StatQuest is the best
@statquest
@statquest 3 года назад
Thank you very much! :)
@gilao
@gilao 3 месяца назад
Another great one. Thanks!
@statquest
@statquest 3 месяца назад
Thanks again!
@marpin6162
@marpin6162 3 года назад
Thank you. Now everything is more clear.
@statquest
@statquest 3 года назад
BAM! :)
@igorg4129
@igorg4129 4 года назад
Thanks Josh! you simply the best
@statquest
@statquest 4 года назад
Thank you very much. I can't wait to get the other videos out soon.
@vokoramusyuriy106
@vokoramusyuriy106 Год назад
Thanks a lot, Josh!
@statquest
@statquest Год назад
My pleasure!
@richfilms6307
@richfilms6307 6 месяцев назад
Unbelievable! Thank you!!
@statquest
@statquest 6 месяцев назад
Thanks!
@alhaanali2502
@alhaanali2502 Год назад
You got the best way to teach thank you❤
@statquest
@statquest Год назад
Thanks!
@anasmomani647
@anasmomani647 3 года назад
you make me regret taking any online course .. just wait your next videos
@statquest
@statquest 3 года назад
:)
@DanielRamBeats
@DanielRamBeats Год назад
This is finally all making sense to me thank you
@statquest
@statquest Год назад
Thanks!
@utkugulgec5508
@utkugulgec5508 3 года назад
These videos should be protected at all costs
@statquest
@statquest 3 года назад
:)
@preetikharb8283
@preetikharb8283 3 года назад
This video made my day, thank you so much, Josh!!
@statquest
@statquest 3 года назад
Thanks!
@mike___-fi5kp
@mike___-fi5kp Год назад
You always are the best.
@statquest
@statquest Год назад
Thanks!
@amirhossientakeh5540
@amirhossientakeh5540 2 года назад
perfect you explain complicated things very underatandable it's amazing
@statquest
@statquest 2 года назад
Thank you very much! :)
@abhishekm4996
@abhishekm4996 3 года назад
Much waiting.... Finally came..
@statquest
@statquest 3 года назад
Bam! :)
@SPLICY
@SPLICY 3 года назад
This is what she said
@cthutu
@cthutu 8 месяцев назад
Excellent content, excellent delivery - just bought your book!
@statquest
@statquest 8 месяцев назад
Thank you so much for supporting StatQuest! BAM! :)
@ge13r
@ge13r 4 месяца назад
Saludos desde San Cristóbal, Venezuela!!!
@statquest
@statquest 4 месяца назад
:)
@JamesWasTakenOhWell
@JamesWasTakenOhWell Год назад
Thank you for the amazing effort you put into this video and BAM!!! as always!
@statquest
@statquest Год назад
Thanks!
@יהונתןקרסניי
@יהונתןקרסניי 3 года назад
so good. can't wait for the next one!
@statquest
@statquest 3 года назад
Bam! It should be out soon.
@miriza2
@miriza2 3 года назад
BAM! Thanks Josh! You’re the best! Got myself a pink T-shirt 😍😍😍
@statquest
@statquest 3 года назад
Hooray! And thank you for supporting StatQuest!!!
@RumayzaNorova
@RumayzaNorova 7 дней назад
Day 2 of leaving the comment and studying with statquest :)
@statquest
@statquest 7 дней назад
double bam! :)
@Morais115
@Morais115 3 года назад
I'm buying the shirt! Kudos to you sir.
@statquest
@statquest 3 года назад
Awesome! Thank you!
@richarda1630
@richarda1630 3 года назад
Where were you 5 years ago???!?!?! :D Awesome work man! Keep it up :)
@statquest
@statquest 3 года назад
Thanks! I have 4 more neural network videos coming out in the next month.
@richarda1630
@richarda1630 3 года назад
@@statquest awesome! can't wait :D
@igorg4129
@igorg4129 4 года назад
Josh, finished watching. Thank you again 1 If I as a researcher know +/- which range of inputs I am going to insert, and which range of outputs I expect to get in the end, will I want to adjust somehow from the very beginning the weights range, maybe weights distribution, same thing about biases and same about activation functions, or today we let the algorithm to do this job? 2 most interesting question: Lets say that while finding the prediction curve we kind of discover some "hidden truth". I think our curve might never be exact also because we do not know all of the independent variables which in nature affect our dependent variable. Say we know one, but there is another one which we do not know about. If so, will it be right to say that when neural network with one input splits the input by different weights into two neurons of a hidden layer (from which the final output is calculated), it is like simulating somehow presence of another "secret independent variable" even without knowing what it is? Thanks
@statquest
@statquest 4 года назад
I'll be honest, I'm not sure how to answer question #1. I don't know. I do know that some of the methods used for initializing the weights with random values increase the variation allowed in the values based on how many layers are in the neural network - so that might do the trick. As for the second question: Adding the second node in the hidden layer allows the squiggle to go up *and* go down. If I just had one node, I would only be able to go up *or* down. So, in some sense, that is sort of like adding a secret independent variable.
@igorg4129
@igorg4129 4 года назад
@@statquest Also thought this way. Thank you again and again you do here a titanic job Josh. If not you I wasn't here to ask new questios. :)!
@ZachariahRosenberg
@ZachariahRosenberg 4 года назад
@@igorg4129 It's tempting to want to initialize weights to a target range in the hopes of speeding up convergence, however this actually might be counter productive. The weights of individual nodes do not have to conform to the same distribution as your output. When you use an appropriate (adaptive) optimizer, it should be able to tune the weights pretty quickly, considering that the first few passes will likely have larger gradients.
@davidlu1003
@davidlu1003 2 месяца назад
I think I understand the neural network clearly now.😁😁😁
@statquest
@statquest 2 месяца назад
bam! :)
@pranjalpatil9659
@pranjalpatil9659 2 года назад
Perfect explanation!
@statquest
@statquest 2 года назад
Thank you!
Далее
Neural Networks Pt. 3: ReLU In Action!!!
8:58
Просмотров 266 тыс.
Свожу все свои тату (abricoss_a_tyt)
00:35
THEY'RE EATING THE DOGS DANCE REMIX!
00:10
Просмотров 175 тыс.
У БЕЛКИ ПОЯВИЛИСЬ КОТЯТА#cat
00:20
Gradient Descent, Step-by-Step
23:54
Просмотров 1,3 млн
The Most Important Algorithm in Machine Learning
40:08
Просмотров 432 тыс.
The Chain Rule
18:24
Просмотров 257 тыс.
Backpropagation Algorithm | Neural Networks
13:14
Просмотров 38 тыс.
The Essential Main Ideas of Neural Networks
18:54
Просмотров 946 тыс.
Свожу все свои тату (abricoss_a_tyt)
00:35