Тёмный
No video :(

MIT 6.S191 (2020): Recurrent Neural Networks 

Alexander Amini
Подписаться 268 тыс.
Просмотров 394 тыс.
50% 1

Опубликовано:

 

23 авг 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 203   
@antonstafeyev3606
@antonstafeyev3606 4 года назад
if u cant go to MIT, make MIT come to you. Thanks to everybody who made it possible.
@tusharsolanki8979
@tusharsolanki8979 4 года назад
They explained the topic in a very easy manner even a guy with no background in ML can understand it. I wish they also had the tutorial for the Practical Sessions.
@nitroyetevn
@nitroyetevn 4 года назад
github.com/aamini/introtodeeplearning
@kpr7717
@kpr7717 4 года назад
@@nitroyetevn Thanks you so much!
@nitroyetevn
@nitroyetevn 4 года назад
@@kpr7717 No problem!
@vigneshgj6061
@vigneshgj6061 4 года назад
I am the first graduate of my family. It will be near impossible to listen MIT lecture unless there is a initiative like this. Now education/knowledge is open-sourced
@josephwong2832
@josephwong2832 4 года назад
unbelievable series!! I'm learning so much from these lectures compared to other youtube vids
@patmaloyan620
@patmaloyan620 4 года назад
CRAZY GOOD quality for each minute of the lecture.
@architectsmusicgroup
@architectsmusicgroup 3 года назад
Great stuff!
@eventhisidistaken
@eventhisidistaken 3 года назад
While stuck at home over the summer, I decided to code up the infrastructure for a bunch of different kinds of AI. I started with perceptron layers, then added recurrent perceptron layers, then LSTMs (that was particularly hard - I had to intensely study at least a dozen research papers to piece it together before all the unstated pieces gelled) , then convolutional nets, then a transformer (encoder/decoder) infrastructure. What I discovered in this process, is that the amount of "intro" level information, as well as "how to AI in python" sort of stuff, completely drowns out the nuts and bolts. In the end, you're stuck reading research papers, which are targeted at an audience that is already a subject matter expert. I suppose universities are supposed to fill that gap, but honestly, this stuff is just not that hard once you decode the language of the field. It's just differential calculus and a bit of optimization theory. Good 3rd year engineering students have the math background, and combine that with some coding skills and you're golden. The other thing I learned in this process, is that convolutional nets and encoder/decoders are just amazing. Even though I wrote every line of code, and understand how and why they work, and train them myself, it feels like magic to watch them work.
@triton62674
@triton62674 Год назад
Top tier comment.
@DarkLordAli95
@DarkLordAli95 3 года назад
everything about this course is phenomenal. It's so good that sometimes I get distracted by thinking about how amazing the course is. The language and the pace used are perfect. The slides are perfect; there's just the right amount of information on the slides so that I don't get overwhelmed by having too much to read while listening, (something I struggle with in my regular classes). It's just so fascinating. Teachers all around the world should take notes. Thank you so much for sharing this with us.
@AAmini
@AAmini 3 года назад
Thank you!!
@DarkLordAli95
@DarkLordAli95 3 года назад
@@AAmini Thanks for the reply Alexander. Could you please let me know if you're going to upload these slides anytime soon?
@AAmini
@AAmini 3 года назад
@@DarkLordAli95 Sure, the slides have been published since last year on the 2020 course site: introtodeeplearning.com/2020/. The most recent course iteration contains the 2021 slides (which are also published, but slightly different from this talk).
@ashutoshbhushan6107
@ashutoshbhushan6107 3 года назад
I can't believe such useful information is available to us for free. Thanks!
@noviaayupratiwi5613
@noviaayupratiwi5613 3 года назад
I would like to say thank you, Alexander and Ava for making it happen! I would make my personal note for Convulational neural network and RNN in 2am and learning from MIT, thank you
@ireenisabel988
@ireenisabel988 2 года назад
Amazing delivery! I have never thought that I would be able to understand any lecture from MIT 100% because of the lack of pre-requisites knowledge. Looking forward to see more videos on GNN, GCN etc..
@AAmini
@AAmini 2 года назад
Thanks! You may also want to check out next week's lecture which will also be on Deep Sequence Modeling but contain a lot of cool new material on Transformers and Attention. The link will be here but it is not published yet: ru-vid.com/video/%D0%B2%D0%B8%D0%B4%D0%B5%D0%BE-QvkQ1B3FBqA.html
@ireenisabel988
@ireenisabel988 2 года назад
@@AAmini Thank you very much. Looking forward to it. ....
@Otis6475
@Otis6475 3 года назад
Thanks Alexander and Ava for this free but complete content about NN. Education at its best.
@yashsolanki069
@yashsolanki069 4 года назад
Not from IIT NIT IIIT but I'm learning from MIT thanks for providing such great learning experience.
@zhenmingwang9363
@zhenmingwang9363 3 года назад
This one is nice. Nicely fit with my class's slides. The best part for me is that it clearly reveals the concept of timestep computational graph, which I have not seen in previous introduction videos.
@ahmednagi7074
@ahmednagi7074 2 года назад
i loved ava way of explanation such hard topics and break it up in easy pieces that can be understood
@saisankarborra2930
@saisankarborra2930 4 года назад
This is nice, leaving behind the mathematics behind, how the vanishing gradient will be overcome by LSTM is nice explanation for RNN.
@davidschonberger8609
@davidschonberger8609 2 года назад
Excellent presentation! Typo alert in slide shown at around the 18:19 timestamp: The loss corresponding to y_t should be L_t, not L_3.
@basilihuoma5300
@basilihuoma5300 3 года назад
This Lectures have been super cool, has clarified a lot of things for me, can't wait for the 2021 Series.
@rushimashru4630
@rushimashru4630 3 года назад
This series of MIT lectures were very effective and productive, especially in this lockdown and WFH situation. I learned a lot. Thank You !!! Alexander Amini Sir to make it possible.
@aritraroygosthipaty3662
@aritraroygosthipaty3662 3 года назад
Being a bootcamp I had never thought the materials could be so meaty and well made. I love the videos. Great job!
@DarkLordAli95
@DarkLordAli95 3 года назад
I don't think she meant "bootcamp" in the literal sense.
@BharathirajaNarendran
@BharathirajaNarendran 4 года назад
Thanks for sharing these lectures as open source. Looking forward to the rest of the boot camp videos and will attempt the lab exercise :)
@worldof6271
@worldof6271 4 года назад
great course. I didn’t think that while sitting in Almaty I could watch MIT courses
@HarpreetKaur-vd1lb
@HarpreetKaur-vd1lb 3 года назад
Hi Ava, Thank you for sharing the knowledge. I have couple of questions: 1. First of all how does back propagation in RNN leads to vanishing gradient but not in case of deep neural network since Mr Amini did not bring it up in the introductory lecture. 2. Secondly you mentioned that having weights as identity matrix will solve the problem of vanishing gradients. So how is that possible. What is the math behind it ? 3. Thirdly, I am confused as to how you are showing the matrix as n*m matrix in the first place. Since the example you took had a sentence which seem to be like a n*1 matrix (where n is the number of words, isn't it?)then how can the weights be a n*m matrix(focusing on the dimensions here). If what is shown is correct then how will you multiple the weight matrix with the example that you have described. 4. And lastly, I could not correlate how relu results in derivative greater than 1. I mean when you use the relu function, then combination of product of weights *x is forced to be a constant value 1 isn't it. So no matter what the value of the dot product of weights and x (and dot product of the state and another weight matrix )is going to be as long as it is greater than zero it will result in a y value of 1, if I understand correctly. However, the value of the function is constant no matter what, then wouldn't that yield the derivative of 0 since the derivative of constant is 0. Maybe I have misunderstood something but you further clarifying this would help certainly. Thanks in Advance!
@ancbi
@ancbi 4 года назад
Please, correct me if I'm wrong. At 36:18, "uninterupted gradient flow" is essentially due to the fact that operations along the route c0,c1,c2,c3,... has no weights to be updated at all. From what I can see, there are kinds 3 operations along that route: [1] point-wise multiplication (type C x C -> C) [2] point-wise addition (type C x C -> C) [3] copying (type C -> C x C) where C is whatever data type c0,c1,c2,c3 is.
@louisryan6902
@louisryan6902 4 года назад
I struggled with this when I came across and ended up resorting to delving into the maths to grasp why LSTMs solve the vanishing/exploding gradient problem. "Uninterrupted gradient flow" does not explain sufficiently. I found this article to be of use (if you can wade through the maths it really explains why LSTMs are better than RNNs for gradient problems) medium.com/datadriveninvestor/how-do-lstm-networks-solve-the-problem-of-vanishing-gradients-a6784971a577
@rubeniaborge4652
@rubeniaborge4652 3 года назад
I love this lecture. I am learning so much with this series. Thank you very much for sharing! :)
@bluealchemist6776
@bluealchemist6776 4 года назад
Outstanding educational sharing...I love you MIT...no place like you in this world...or the next
@RishitDagli
@RishitDagli 4 года назад
Wonderful lectures, I love them. The things are so much simplified and made easy to learn
@Harini.R
@Harini.R 4 года назад
Excellent! This has been super useful to get my head around ML terminologies and potentially use them on my ongoing project. Thank you very very much!
@wangsherpa2801
@wangsherpa2801 4 года назад
Despite not being very good in maths, I am understanding all these lectures. ❤️Love from India❤️
@xiyupan299
@xiyupan299 2 года назад
The best one on the RU-vid
@AAmini
@AAmini 2 года назад
Thanks! You should also check out the new 2022 version, it's even better!!!
@emenikeanigbogu9368
@emenikeanigbogu9368 4 года назад
I have been eagerly awaiting this video
@kushangpatel983
@kushangpatel983 4 года назад
Thanks MIT for providing access to such an amazing series of lectures!
@stephenlashley6313
@stephenlashley6313 Год назад
Absolutely excellent explanation! To further motivate and inspire your great work, where the ball goes next is the holy grail of all models of everything: Quantum Mechanics. You can easily infer the quantum "genome" sequence of all reality with this technology. Again, great presentation and content
@sichaoyin7675
@sichaoyin7675 4 года назад
very nice course. lots of new stuff. looking forward to new releases
@footstepsar992
@footstepsar992 3 года назад
Great Lecture Series! I just tried to download the slides from your site, but they are unavailable; the site just states that slides and videos are upcoming for 2021 lectures. Do you know where to find the 2020 lecture slides?
@DarkLordAli95
@DarkLordAli95 3 года назад
they're still not there :(
@nickglidden9220
@nickglidden9220 4 года назад
Really amazing video! A ton of info in 45 minutes but in an easy to understand manner!
@Matteopolska
@Matteopolska 4 года назад
Thanks from Poland for this great and valuable content :)
@kennethqian6114
@kennethqian6114 3 года назад
This was an incredibly well-organized lecture on recurrent neural networks. Thank you so much for the video!
@siminmaleki4818
@siminmaleki4818 4 года назад
great courses. Thank you for sharing. you rock! Proud of you in Iran.
@Kevin-gm7gx
@Kevin-gm7gx 4 года назад
Thank you so much for making top class education accessible, especially such an important topic!
@jithuk8693
@jithuk8693 4 года назад
I am having a doubt! What will be the initial step in the first iteration? Means what is ht-1 in first iteration?
@praveenkumarverma9470
@praveenkumarverma9470 2 года назад
great lecture
@hrsight
@hrsight 2 года назад
great material
@mattiapennacchietti9224
@mattiapennacchietti9224 4 года назад
Well done MIT, always one step ahead
@Shah_Khan
@Shah_Khan 4 года назад
Thanks Ava....for sharing us informative lectures....Learning a lot from your videos...Can I ask any questions from you people either in slack or somewhere else?
@c9der
@c9der 3 года назад
Thanx for provide High quality content..... Always wanted to go MIT
@avdhutchavan3044
@avdhutchavan3044 4 года назад
I am getting an error while playing the songs ('C:\Users ame\PycharmProjects\tensors\venv\lib\site-packages\mitdeeplearning\bin\abc2wav' is not recognized as an internal or external command, operable program or batch file.)
@marwasalah2640
@marwasalah2640 2 года назад
This is a good lecture and a great instructor :)
@lksmac1595
@lksmac1595 3 года назад
excellent!
@louerleseigneur4532
@louerleseigneur4532 3 года назад
Thanks
@omidzare1934
@omidzare1934 3 года назад
great lecture . very informative
@neuroling
@neuroling 4 года назад
These lectures are fabulous, but this one is top tier. Excellent. Thank you!
@barbaracarnauba1507
@barbaracarnauba1507 3 года назад
Thanks for sharing this great lecture!!
@HazemAzim
@HazemAzim 4 года назад
Very Good .. Great Lecture .. Thanks
@tamdoduc9804
@tamdoduc9804 3 года назад
Amazing coure! Thank you!
@teeg-wendezougmore6663
@teeg-wendezougmore6663 2 года назад
Great course. Thanks for sharing. Is RNN module equivalent to a simple neuron? I am little confused with terms cell and module
@prabhavarora1992
@prabhavarora1992 4 года назад
Hi all I was just confused about something. The point of RNNs is to preserve the order in a sequence. Let us say our sequence is "I took my cat fora walk". If we use a really large fixed window and make it into a fixed vector to put into a Vanilla feedforward network is the order preserved? I guess what I am really asking is can a feedforward network preserve order?
@osvaldonavarro3292
@osvaldonavarro3292 Год назад
How does the vanishing gradients problem is solved by the LSTM? If the LSTM includes sigmoid operations, doesn't that contribute to make gradients smaller?
@RajaSekharaReddyKaluri
@RajaSekharaReddyKaluri 4 года назад
Thank you @ava soleimany
@smartteluguguru
@smartteluguguru 4 года назад
Thank you very much, for giving knowledge from another dimension, great course. :)
@kingeng2718
@kingeng2718 4 года назад
The lab parts are really useful thanks alot
@aromax504
@aromax504 4 года назад
I am 30 now and I wish to have these sort of contents on my school/college days. Still trying to learn as much as possible. Thank you for your contribution to democratize world class education
@hemingwaykumching3812
@hemingwaykumching3812 4 года назад
Thank you for sharing knowledge and making it available for all. I am feeling difficulty in understanding the lecture videos, I had never before learnt anything about ML or DL but I really enjoyed the first video. Is there any perquisite information required for this series of lecture?
@SantiSanchez28
@SantiSanchez28 4 года назад
I'd say basic linear algebra, calculus and programming
@Gustavo_Rojas
@Gustavo_Rojas 4 года назад
for us watching online, is there other material we can use to go along with these lectures?
@firmamentone1
@firmamentone1 4 года назад
Hi everyone, At 31:00, Does the block in the middle of the slice represent a single neuron or a layer?
@ronniechatterjee4368
@ronniechatterjee4368 4 года назад
it's a single RNN cell with more information (and basically the addition of a cell state and also the gates). One unit in the layer, not quite a single neuron.
@Anonymoususer-um6yr
@Anonymoususer-um6yr 4 года назад
Damn, these lectures are so interesting. I'm really into evolutionary algorithms right now but can't find any lectures by MIT. Any suggestions?
@abhishekrungta4056
@abhishekrungta4056 4 года назад
A question related to lab session 2: Why does the last code block show that there are no songs in the text when clearly I generated it using the generate_text function in the previous block?
@sirjohn8629
@sirjohn8629 4 года назад
I am learning new things!
@pythonocean7879
@pythonocean7879 4 года назад
Nice timing,i was searching for that,thanks ;)
@RizwanAli-jy9ub
@RizwanAli-jy9ub 4 года назад
thankyou for this series of lectures
@schnittstelle8492
@schnittstelle8492 4 года назад
amazing explained in only 45 minutes. Why don't they teach like this at my university?
@matthewphares4588
@matthewphares4588 4 года назад
Great lecture Ava. Is there any chance I can consult with you on a project I’m working on? You can choose the hourly tutor rate?
@hsiang-yehhwang2625
@hsiang-yehhwang2625 4 года назад
Great lecture from MIT!!
@star831
@star831 2 года назад
This was really helpful!
@a.yashwanth
@a.yashwanth 4 года назад
If a neural net generates music, after training using a copyrighted music, does the rights of generated music still belong to the copyrighted music creator?
@fr0iler578
@fr0iler578 4 года назад
No
@fatnasaeed2937
@fatnasaeed2937 3 года назад
Thanks very much well understanding
@michaelcjakob
@michaelcjakob 4 года назад
Amazing series, thank you! Would love to see more from MIT! (:
@ahmedelsabagh6990
@ahmedelsabagh6990 4 года назад
MIT is great!
@dianaamiri9520
@dianaamiri9520 3 года назад
wow you explain it amazingly clear and understandable. Thanks MIT for sharing
@lorelabano4497
@lorelabano4497 3 года назад
This is great, thank you!
@drallisimo34
@drallisimo34 3 года назад
great teaching!!!
@nitroyetevn
@nitroyetevn 4 года назад
Great lecture!
@aliyaqoob8779
@aliyaqoob8779 4 года назад
How do you choose between different non-linearity functions? In this example here, we used a tanh but could we have used a sigmoid instead?
@user-us3ny6ii9r
@user-us3ny6ii9r 4 года назад
Having much fun from this.. absolutely love this!!
@seangad8227
@seangad8227 3 года назад
amazing
@mukul2610
@mukul2610 3 года назад
When we talk about Ht can we say that we are giving some sort of feedback to the perceptron?
@DatascienceConcepts
@DatascienceConcepts 4 года назад
awesome content!
@SumukaG
@SumukaG 4 года назад
Could somebody explain in detail about shared parameters that she talks about ? Maybe with another example ?
@taewookkang2034
@taewookkang2034 4 года назад
Great lecture, great lecturer
@vincentlius1569
@vincentlius1569 4 года назад
I have a question, does feed forward have the vanishing gradient problem, if so how to fix it ?
@reachDeepNeuron
@reachDeepNeuron 4 года назад
what's the presenter name ? very well explained in a very intuitive way... awesome ....
@divinemaker
@divinemaker 3 года назад
I noticed that you cut some of the student questions out of the videos. I think you should include all of them in, because it would make us understand better in my opinion. Of course if a student does not want you to publish it, that's fine.
@francescod8204
@francescod8204 3 года назад
Great lesson!
@sarthakshukla5251
@sarthakshukla5251 3 года назад
In the LSTM diagram, what operation does the intersection of the wires perform?
@mattrowlands5751
@mattrowlands5751 3 года назад
Its vague in the diagrams... in reality the input to each gate is calculated as: weight * ht-1 + weight * xt + bias, where each gate has their own weight and bias and ht-1 is the previous 'hidden state' and xt is the new input.
@sarthakshukla5251
@sarthakshukla5251 3 года назад
@@mattrowlands5751 I thought about it and most probably they represent flow of information with the tanh box representing tanh(Whh*ht-1 + Wxh*xt) i.e the entire function/expression has been abstracted to that box.
@sarthakshukla5251
@sarthakshukla5251 3 года назад
@@mattrowlands5751 Thanks btw
@mattrowlands5751
@mattrowlands5751 3 года назад
SARTHAK SHUKLA Yep that is correct. I personally think these diagrams are very misleading and confusing. Best of luck to you my friend.
@RAZONEbe_sep_aiii_0819
@RAZONEbe_sep_aiii_0819 4 года назад
Can we also have some coding lectures associated with the topics you are discussing here? I will be glad if u can include it.
@amantayal1897
@amantayal1897 4 года назад
if you check their website you can find some coding problems
@aliyaqoob8779
@aliyaqoob8779 4 года назад
so how is c_t different from h_t? How is this "cell state" different from h_t? And why are both of those used?
@blakef.8566
@blakef.8566 4 года назад
36:20 does this suggest that all loss information is propagated through the internal state of the cell?
@digvijayyadav3633
@digvijayyadav3633 4 года назад
Are there any hands-on projects for Applications of RNN in music generation?
@junqima5127
@junqima5127 3 года назад
I know I can't go to MIT as I have tons of questions and they only seem to have a couple
@berradi5521
@berradi5521 4 года назад
thanks
@manishbolbanda9872
@manishbolbanda9872 3 года назад
at 34:38 what is the use of RHS tanh block in LSTM??please ans if you know.thanks
@carmenliu3931
@carmenliu3931 4 года назад
thank you so much this class is amazing!!!!
Далее
MIT 6.S191 (2020): Convolutional Neural Networks
37:20
Просмотров 307 тыс.
MIT Introduction to Deep Learning | 6.S191
1:09:58
Просмотров 493 тыс.
When I met the most famous Cristiano
01:03
Просмотров 27 млн
Simple Flower Syrup @SpicyMoustache
00:32
Просмотров 1,6 млн
Recurrent Neural Networks : Data Science Concepts
27:17
The Future of AI is Neurosymbolic
39:56
Просмотров 6 тыс.
Recurrent Neural Networks (RNNs), Clearly Explained!!!
16:37
26. Chernobyl - How It Happened
54:24
Просмотров 2,8 млн
Lecture 2: Airplane Aerodynamics
1:12:07
Просмотров 3,1 млн
MIT 6.S191 (2020): Reinforcement Learning
44:11
Просмотров 107 тыс.
MIT 6.S191 (2020): Introduction to Deep Learning
52:52
When I met the most famous Cristiano
01:03
Просмотров 27 млн