Тёмный

Rasa Algorithm Whiteboard - Transformers & Attention 2: Keys, Values, Queries 

Rasa
Подписаться 30 тыс.
Просмотров 75 тыс.
50% 1

This is the second video on attention mechanisms. In the previous video we introduced self attention and in this video we're going to expand the idea by introducing keys, queries and values.
We're going at it step by step, but if you're interested in immediately reading all about it in full detail then we might recommend these online documents:
- www.peterbloem.nl/blog/transfo...
- jalammar.github.io/illustrated...
- d2l.ai/chapter_attention-mecha...
The general github repo for this playlist can be found here: github.com/RasaHQ/algorithm-w....

Наука

Опубликовано:

 

26 апр 2020

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 185   
@esteveslisboeta
@esteveslisboeta 3 года назад
I hereby declare this saga to be the best on Attention! Thank you so much for sharing the knowledge with such clarity!
@paulah1639
@paulah1639 3 года назад
Hear hear!
@Kanasta7
@Kanasta7 2 года назад
I agree. Your teaching skills are superb! Thank you so much
@itzikgutzcha4779
@itzikgutzcha4779 3 года назад
Dear Rasa, I have watched at least 10 videos explaining this subject, and only now, after seeing your explanation, I think that I finally understand. You are a great teacher, thank you!
@BAHTYS
@BAHTYS 2 года назад
Finally! Finally, someone managed to explain it to me in reasonable detail and extremely clear! Thank you!
@thatsharma1066
@thatsharma1066 Год назад
ultra relatable
@kdubovetskyi
@kdubovetskyi 2 года назад
The single explanation which actually shows the motivation behind Q/K/V. Thanks.
@amritbhattarai5083
@amritbhattarai5083 Год назад
if only the papers included the information about what the query, keys and the values actually mean! As a beginner at this field, I was having such a difficult time to wrap my head around what those actually mean. I tried convincing myself that I understood that using query and keys, we get values, but today is the day that I actually understood. Hats off sir, I want to thank you for this great explanation. The world needs teachers like you :)
@thatsharma1066
@thatsharma1066 Год назад
The best video on attention you will found on RU-vid.
@muhannadobeidat
@muhannadobeidat Год назад
If you did not understand the explanation here, just rinse and repeat and you eventually will. This is the best explanation of keys, values and queries. Just pay attention! Thanks for putting this together!
@noahgsolomon
@noahgsolomon 4 месяца назад
this concept has been explained so abstractly in other videos. Props for making it intuitive, and going through the process instead of explaining the final derivation
@joliver1981
@joliver1981 3 года назад
I cannot stress how great of a teacher you are. Amazing work. Truly. I feel as though you are literally walking my stupid brain through the subject matter.
@RasaHQ
@RasaHQ 3 года назад
(Vincent here) Kind words. Thanks!
@renatoviolin
@renatoviolin 4 года назад
Incredible explanation! Please, keep going this kind of illustrative explanations. Congrats.
@jithinmukundan9016
@jithinmukundan9016 Год назад
Excellent explanation. The only video among dozens I have seen that actually explains the 'why' and 'how' of using query, key and value in attention mechanism. So lucid and concise. Thank you so much.
@briancase6180
@briancase6180 Год назад
I'm pretty impressed. I've suffered through other explanations (including from noted professors), so I've learned some things first sure. What's *excellent* about your explanations is that they are easier to follow, very well motivated (I see why we want each structure), and are bite-sized so it's not too much at once. Great job.
@saintcodded2918
@saintcodded2918 4 месяца назад
Truly ATTENTION is all you need to understand this piece. Great work👍
@yusufani8
@yusufani8 Год назад
I am addicted to this video whenever I forget attentions mechanism I remember picture of 09:07
@irlporygon-z6929
@irlporygon-z6929 5 месяцев назад
oh my god. there are SO many videos on this subject that fail to explain almost anything, probably because the presenter has no comprehension of the topic. i have been looking for this for a couple weeks now lol. just a video that states clearly "this is a vector, the output of this operation is a number" et cetera and clearly explains exactly WHAT everything is instead of vague nonsense analogies about what the pieces are. Also I have to say your handwriting/drafting skills are fantastic
@AI_Financier
@AI_Financier 2 года назад
Now the trilogy of Query-Keys-Values (the penny) dropped, thank you sir
@ns-teamtv8888
@ns-teamtv8888 3 месяца назад
I am teaching in university self attention, this video helps me a lot to find the way to explain this concept Thanks a lot for this fantastic work with Rasa and this series of videos!! A master piece
@binhu8128
@binhu8128 Год назад
Kudos to the instructor. Very clear explanation.
@RezoanurRahman
@RezoanurRahman 2 года назад
Bro...when you put the analogy of query key and value that was like watching a movie and figuring out the plot twist.
@MrChilledstep
@MrChilledstep 3 года назад
Extremely clear. You are a brilliant teacher. Thank you!
@GauravBbbb
@GauravBbbb Год назад
This video is ALL YOU NEED to know to understand the concepts in self attention. Best explanation so far about Q, K, V terms!!!!!!
@airepublic9864
@airepublic9864 2 года назад
Your description developed deep understanding&knowledge of how attentions works, Thx for earliest...
@DanOblinger
@DanOblinger Год назад
JUST WOW !!! That was **SO** lucid. I understand the details of the math & the deeper intuitions all in one go. AND my brain was not over-heating with the effort.
@deudaux
@deudaux Год назад
The clarity of this video is just on whole 'nother level!
@shantanunath7927
@shantanunath7927 Год назад
The best illustration I have ever seen.. It changed my views
@sebastianp4023
@sebastianp4023 3 года назад
You Sir, will be cited in my Master's Thesis if i end up using attention.
@osuregraz
@osuregraz 2 года назад
This is the best video. I’ve been searching for days because I don’t understand where do the key query and value come from (most other videos just talk about how the attention work but ignore how the KQVs are generated). This video solves my issue. Thumb-up!
@geekyprogrammer4831
@geekyprogrammer4831 Год назад
yeah because Q/K/V are weights just like what we learned in ANN at beginning
@azurewang
@azurewang 3 года назад
the way you explain it is as elegant as the design of slef attention
@edvinbeqari7551
@edvinbeqari7551 10 дней назад
This is the best explanation I have seen thus far. Thank you and subscribed.
@297339003
@297339003 3 года назад
I spent 2 days trying to learn attention layers in transformers. Didn't make any sense until I watched this! TYSM!!!!
@xphn1985
@xphn1985 Год назад
Brillant lesson! For the first time, I get to understand how the analogy of key, query and value come forth. Thank you!
@qjrmsktso2
@qjrmsktso2 3 года назад
The best explanation about self-attention ive ever heard!! Thx a lot
@justinwhite2725
@justinwhite2725 2 года назад
3:00 thank you for this. Coming at this from a programming background and not a math background people talk about a 'dot product' and I'm here going 'I have no idea what that means'. Yes I've googled it, and I've never been sure I was doing it correctly (I wasn't) before seeing this explanation. 4:50 thank you! Other explanations were never clear if I were multiplying everything to get once number or if I were just multiplying out the array to get a new array.
@lenhardreuter2254
@lenhardreuter2254 2 года назад
Best video on attention layer there is! And the ONLY one that explains it, so that it‘s understood! Awesome work! Thank you!!
@purushotamradadia8175
@purushotamradadia8175 2 года назад
Great explanation ever heard on query key and value.
@SaxonBerryVideos
@SaxonBerryVideos 10 месяцев назад
This is a fantastic video. Super intuitive storyline to unveil the concepts. Really easy to follow. If I knew someone who was trying to learn transformers I would share this with them straight away.
@christopherjamesyoung7766
@christopherjamesyoung7766 2 года назад
construction of attention from concept to implementation. excellent job
@lettry5297
@lettry5297 3 года назад
I am speechless after this high quality of explanation
@galenw6833
@galenw6833 9 месяцев назад
These videos are excellent, and put the original paper and many webpages to shame. "These videos are all you need." (see what I did there? :-P) 'Attention' basically means: re-projecting each token from their original semantic embedding (word2vec, glove, etc.) into a new basis which is the relative similarity in meaning to other tokens in the sentence (sequence). So it should really be called "relative meaning space projection is all you need". Not as catchy, but perhaps clear. Query, key and value mean weight matrices for each time the token embeddings are used (twice in the dot product with itself to obtain the weights, and once for the input embeddings). This allows better re-embeddings (re-projections in terms of other vectors in sequence) to be learned.
@obafemijinadu4726
@obafemijinadu4726 4 месяца назад
Finally! attention makes all the sense in the world.
@abhi8569
@abhi8569 3 года назад
The best explanation.
@Michael-yu9ix
@Michael-yu9ix Год назад
Phenomenal explanation. I've watch a few other videos that left me more confused, but this finally made me understand what the purpose of the key, value, query weight matrices is. To be honest, I'm not sure whether some of the youtubers that make videos on this topic actually understand the purpose of the matrices fully. This whole series is so extremely good and helps so many beginners to better understand these models. Cannot thank you enough for making these videos. A lot of people as you can see from the comments are extremely thankful for providing such clear nicely illustrated explanations. Please keep making more videos. We all appreciate it a lot!
@carlosgruss7289
@carlosgruss7289 Год назад
What a brilliant series of videos. Thank you!
@sadeghmohammadi5567
@sadeghmohammadi5567 2 года назад
You are rock! amazing, I do not know how long it takes you to get a context so clearly. but, It would take me , maybe reading many articles and finding connection maybe around 2-3 weeks!
@leoyao1994
@leoyao1994 3 года назад
This is incredible..... so complicated but so clear.... AMAZING WORK!!
@Rotem_shwartz
@Rotem_shwartz 9 месяцев назад
I finally got it !! After so many videos, thank you !!!❤
@JohnCena12355
@JohnCena12355 2 года назад
This is by far the best explaination of Attention I have seen. Thank you!
@AnnaGrace5
@AnnaGrace5 11 месяцев назад
Absolutely amazing. Makes SO MUCH MORE SENSE now. Thank-you :)
@Deshwal.mahesh
@Deshwal.mahesh 3 года назад
Seriously THE BEST videos I've ever seen on Attention and Self-Attention. Really Loved it
@aliyoussef97
@aliyoussef97 10 месяцев назад
The best explanation I watched so far
@nayanvats3424
@nayanvats3424 2 года назад
Too good, a simple topic explained with elegance. Cheers!
@prashanths8536
@prashanths8536 2 года назад
Very good explanation especially with vector representation. Turns out I have heard this voice on other video tutorials too which I appreciate. Thank you!
@alvarozamora2679
@alvarozamora2679 3 года назад
I've done quite a lot of work on neural networks, but none on NLP. I was having trouble understanding these attention blocks but this is super clear. good stuff.
@xv0047
@xv0047 2 года назад
This is an absolute home run. Well done!
@eyorokon
@eyorokon 2 года назад
by far the best explanation on self attention. ty
@junowhut7486
@junowhut7486 2 года назад
yup, easily the best explanantion of attention. thank you for sharing!
@techaztech2335
@techaztech2335 2 года назад
this is a mind bogglingly awesome explanation....
@ananosnasos5043
@ananosnasos5043 Год назад
finally, a great and simple explanation, thank you very much
@fiv1067
@fiv1067 3 года назад
Wonder how this video has only 4k views. Excellent explanation :(
@pranjalchaubey
@pranjalchaubey 4 года назад
Coolest explanation of Keys, Values and Queries!
@TaylorSparks
@TaylorSparks 2 года назад
great video. Helps explain Q,K,V in attention
@felipefrigeri9787
@felipefrigeri9787 3 года назад
Well, this got me attent! Thank you so much for the explanation, I was very confused about the V, Q, K matrices!
@yevhendiachenko3703
@yevhendiachenko3703 2 года назад
Explanation is brilliant!
@SyntharaPrime
@SyntharaPrime Год назад
it is incredibly wonderful explanation for this subject. thanks so much. It was a great chance to see this video. really thanks a lot
@jaeboumkim1213
@jaeboumkim1213 3 года назад
Great!!! It's the best video for explanation on Attention!
@bikashshrestha1958
@bikashshrestha1958 3 года назад
The Best Explanation on Attention PERIOD!!!
@AkshatSharma-qx9wh
@AkshatSharma-qx9wh 2 года назад
The best explanation!! You are the best .. Kudos !
@abc-by1kb
@abc-by1kb 2 года назад
OmG I can't believe I didn't discover this earlier. What a great video. Really love your way of explaining things.
@mohanapalaka
@mohanapalaka 3 года назад
Most intuitive explanation I found! Thank you :D
@LuisPerez-rr9jc
@LuisPerez-rr9jc Год назад
That was incredibly helpful, thanks!
@sahar2003
@sahar2003 3 года назад
best attention explanation on the web! THANK U!
@theapplecrumble
@theapplecrumble 2 года назад
Thanks this is tremendously helpful!
@ishgirwan
@ishgirwan 3 года назад
Woah...finally it all makes sense to me. Best explanation of self attention I have watched. Thanks :)
@baranyildirim668
@baranyildirim668 2 года назад
This really is the best one out there. Thank you so much
@maxwinmax
@maxwinmax 2 года назад
Perfect explanation. This helped a lot, thanks !
@krishanudasbaksi9530
@krishanudasbaksi9530 3 года назад
What an awesome explanation !!! Thank you Rasa very much for making this... :D
@rayaay3095
@rayaay3095 3 года назад
I finally understand the attention mechanism. Thank you so much
@dhoomketu731
@dhoomketu731 3 года назад
Brilliant explanation. Loved it.
@katrinb8297
@katrinb8297 3 года назад
Thank you, your explanation helped a lot!
@acepanson
@acepanson 5 месяцев назад
beautiful explained, thanks!
@aashwinsharma1859
@aashwinsharma1859 Год назад
Best video on attention. Thnx a lot
@BiranchiNarayanNayak
@BiranchiNarayanNayak 4 года назад
Excellent tutorial on Attention
@sannalun845
@sannalun845 3 года назад
I'd like this video 1000 times if I could.
@310gowthamsagar5
@310gowthamsagar5 10 месяцев назад
oh my goodness !! you explanation is awesome.
@SubhamKumar-eg1pw
@SubhamKumar-eg1pw 3 года назад
Amazing explanation!
@suchandrabhattacharyya5263
@suchandrabhattacharyya5263 Год назад
What an amazing lecture
@willcowan7678
@willcowan7678 2 года назад
Awesome video, thank you
@tantzer6113
@tantzer6113 2 года назад
Excellent! More please!
@user-dg5fw4vz5o
@user-dg5fw4vz5o Год назад
wow !! what a great explanation. 👍
@harshamusunuri1924
@harshamusunuri1924 2 года назад
RASA -> Real Attention Scholar Around Best!
@WisamMechano
@WisamMechano 3 года назад
Incredible, thank you so much
@DeepFindr
@DeepFindr 3 года назад
Wow! Great videos!
@lizl9770
@lizl9770 2 месяца назад
Amazing! Finally understood
@tahamohd1409
@tahamohd1409 Год назад
Wow! Thank you so much I finally get it
@sebastianschramm641
@sebastianschramm641 4 года назад
Great video! Keep them coming
@ipekuyguner5102
@ipekuyguner5102 2 года назад
THANK YOU SO MUCH FOR THIS VIDEO!
@tedschreiber5717
@tedschreiber5717 3 года назад
great video, thanks!
@TrevorHigbee
@TrevorHigbee 16 дней назад
This (and video 1) is so genius.
@abdellahsellam912
@abdellahsellam912 3 года назад
This an excellent explnation, thanks a lot
@Rm-no6jr
@Rm-no6jr 3 года назад
Just wow. You are amazing man.
Далее
The KV Cache: Memory Usage in Transformers
8:33
Просмотров 27 тыс.
I gave 127 interviews. Top 5 Algorithms they asked me.
8:36
Key Query Value Attention Explained
10:13
Просмотров 17 тыс.
Self-Attention Using Scaled Dot-Product Approach
16:09
Vectoring Words (Word Embeddings) - Computerphile
16:56
How a Transformer works at inference vs training time
49:53
КАК GOOGLE УКРАЛ ANDROID?
17:44
Просмотров 50 тыс.
Power up all cell phones.
0:17
Просмотров 49 млн
Bardak ile Projektör Nasıl Yapılır?
0:19
Просмотров 3,3 млн