Тёмный

Visual Guide to Transformer Neural Networks - (Episode 1) Position Embeddings 

Hedu AI by Batool Haider
Подписаться 11 тыс.
Просмотров 133 тыс.
50% 1

Опубликовано:

 

9 сен 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 279   
@HeduAI
@HeduAI 3 года назад
**CORRECTIONS** 1) Shoutout to @SHUBHAM PACHORI and @gabewebyt: [11:20] "Sine or cosine function is used as a function of the dimension of the positional embedding -- so dimension 0 of the positional embedding vector uses sine; dimension 1 uses cosine, and so on." 2) The issue of variable lengths mentioned at [7:26] can actually be also resolved by setting a maximum sequence length (padding shorter sentences, trimming larger sentences to be of the maximum sequence length). *Timestamps* Here are the timestamps associated with the concepts covered in this video: 0:00 - Recap of Part 0 1:14 - Input Processing 2:19 - Word Embeddings 4:44 - Position Embeddings
@RajamannarAK-md8qy
@RajamannarAK-md8qy 3 года назад
Hi, Thanks for the series, really a good one. Were can we see Part 0
@Mayur7Garg
@Mayur7Garg 3 года назад
At 2:56 the screen says 512 but you say 521. I guess the text on the screen is correct but can you confirm? Also the reference links in the video description do not work anymore.
@jnjnqy
@jnjnqy 2 года назад
Hope you fix this especially the pe explanation error. Or give a correction link on-screen at 11:20~11:29. It is a really misleading 9-seconds.
@HeduAI
@HeduAI 2 года назад
@dev stuff Can you please tell me the timestamp from the video where I said "against" so I know what you are trying to ask better?
@Freebytes
@Freebytes Год назад
@dev stuff While this is certainly late, when she says "against", she means that the input index value is separate from the array. This can be represented by a multidimensional array of embeddings [][512] or where the first array is the set of all possible tokens or could be represented by an object that contains both an index and the separate set of embeddings. Consider a TokenSet object that has the following shape: double Value; double[512] Embeddings; In this example, the Value is the index value of the token / word. The Embeddings are an array that is associated with this token. The Embeddings change as the network trains (and are initially random), but the index values never change and will be used to retranslate the decisions of the network back into the English word.
@kevinlu4410
@kevinlu4410 3 года назад
This is the best explanation I've seen for positional encoding. It deserves more likes!
@HeduAI
@HeduAI 3 года назад
Glad you found it useful :)
@wadewang574
@wadewang574 3 года назад
The only tutorial on the internet that makes me fully understand transformer positional encoding !
@HeduAI
@HeduAI 3 года назад
Such an accolade! Comments like these makes me want to keep making more :)
@hungrystork7252
@hungrystork7252 Год назад
THANK YOU! I watched almost 10 different videos about Transformer networks and you're the only person who cared to explain that the vector representations are determined randomly and learned through backpropagation. The visuals are very helpful too. I'll watch the whole video series!!
@justinyeh6137
@justinyeh6137 3 года назад
This is by far the best explanation I've ever seen on youtube. Thanks for the excellent work.
@fluidice1656
@fluidice1656 2 месяца назад
After watching the whole series (the 3 episodes), I can very confidently say that this is the clearest, most succinct, and most useful explanation of transformers on YT that I've come across. Thank you!!
@JujutsuMan
@JujutsuMan 7 месяцев назад
OMG...this video is filling the gap among most Transformer explanation! Most of them are just emphasis the part of Multi-Head Attention! Thanks you truely!
@vichitrakumar4452
@vichitrakumar4452 Год назад
For me this was better than Jay Alamar blog. Kudos to you!! Please keep uploading, thanks!!
@aaryannakhat1842
@aaryannakhat1842 3 года назад
This is probably the best explanation of Positional Encodings I have ever come across! Kudos to the creator!
@manavssingh
@manavssingh 3 года назад
you're a better teacher than my highly qualified uni professors, thank you!
@uncannyvalley2084
@uncannyvalley2084 2 года назад
To those who hesitate watching this it is the best explanation of the transformers I ever seen and I saw a lot of them. A good plus with this video is the one by Yann Lecun on associative memories and transformers on NYU.
@WiredWizardsRealm-et5pp
@WiredWizardsRealm-et5pp 2 месяца назад
Best explanation so far with simplicity, coherent animations without compromising the value . Don't know why but whenever I hear a women or feminine voice explaining any concept it just goes so smoothly into my brain ..no friction. and main part is teaching efficiently . 😅.. anyways Thank-you
@devchoudhary8892
@devchoudhary8892 2 года назад
I have watched almost 10 videos on transformer models this is the best explanation because each concept has given so much time, Thank You.
@user-cf4cl8jy3w
@user-cf4cl8jy3w Год назад
Short, intuitive, and comprehensive -very useful, Thank you
@renanangelodossantos4726
@renanangelodossantos4726 4 месяца назад
I've watched and read a lot about LLM and Transformers. This is the best explanation, hands down.
@bytblaster
@bytblaster Год назад
This is the BEST guide to Transformers on the internet I could find!
@Freebytes
@Freebytes Год назад
Your Transformer model tutorials are the best I have seen for explaining exactly how the process actually works. Thank you for taking the time to put this together with clear explanations. You know a person has a true understanding of a subject if they are able to clearly explain it to others.
@alexandrbyzov3581
@alexandrbyzov3581 Год назад
Man, this is such a good explanation. I hope that this channel will become more popular
@shubhampachori832
@shubhampachori832 3 года назад
Hi at 11:14 I guess, they did not use sine and cosine for different positions of embeddings, they used cosine and sine for odd and even embedding dimension at the same pos, where "i" varies from 0 to dim/2. BTW, thanks for the awesome tutorial.
@gabewebyt
@gabewebyt 3 года назад
Oops I wrote my comment and then saw that you already wrote the same thing. Good catch!
@HeduAI
@HeduAI 2 года назад
Thanks for the great catch! Have mentioned you on my pinned comment for credits :)
@Beenum1515
@Beenum1515 Год назад
Good catch. Was looking in the comments if someone mentioned it :)
@BlockDesignz
@BlockDesignz 5 месяцев назад
First person to concretely explain why they use a periodic function, which in my mind would give the same position embedding when you come back to the same point on the curve. Thank you!
@pedroviniciuspereirajunho7244
Finally! A video that shows how it works with real data involved. Thanks a lot!
@BlayneOliver
@BlayneOliver Год назад
I loved that you put the embedding vectors into a color sequence, that made it so much easier to understand how the machines see the words 🙏
@HeduAI
@HeduAI Год назад
Thank you! (P.S. your baby is so beautiful! :) )
@BlayneOliver
@BlayneOliver Год назад
Thank you@@HeduAI He's not going to normal school one day, rather being taught to code instead :)
@HeduAI
@HeduAI Год назад
@@BlayneOliver Well, seems like he's got an awesome dad 👍
@umarpatel3517
@umarpatel3517 2 года назад
By far the best explanation of positional encoding in the entire universe
@gabriellara7456
@gabriellara7456 2 года назад
insanely good explanation of positional embeddings. thanks a lot for the work put in.
@darshantank554
@darshantank554 2 года назад
I came here with so many questions and confusion but going to the next episode with clear thoughts! Thanks!
@zhengchen388
@zhengchen388 Год назад
This video explains the WHY part of transformers, answers so many doubts, most underrated veejeo
@hewas321
@hewas321 Год назад
Definitely the best explanation of Transformer’s input. No doubt
@caothanhlong7577
@caothanhlong7577 2 года назад
Wow so easy to understanding. It's the easiest explanation of position embeddings I found ever. Thanks.
@julionather8508
@julionather8508 2 года назад
this is by far the best explanation about transformers i've ever seen. thanks a lot !
@zafersahinoglu5913
@zafersahinoglu5913 8 месяцев назад
Watched a vast number of videos and read papers on positional encoders, how Q,K, V matrices work in the attention mechanism and why we need multi-head attention. You are far ahead of everybody else in transferring knowledge to others with utmost clarity and ease. Thanks Batool Haider.
@HeduAI
@HeduAI 8 месяцев назад
You are very welcome :)
@AnnasBlackHat
@AnnasBlackHat 6 месяцев назад
the best, and the most understandable about transformer architecture.... this video is consice with a great explanation, you can perfectly get the intuition... thanks
@junfengzhang7064
@junfengzhang7064 Год назад
This is the best explanation for positional encodings! Great work!
@mian8825
@mian8825 8 месяцев назад
This is indeed the best explanation of transformer to me, it covered the details that the other didn’t, it is logical and clear to me. Thank you,
@irekahsila
@irekahsila 2 года назад
Can't find a better explanation for positional embedding anywhere else, cause this is the best and most straightforward one. 👍🏻
@jaunvanheerden
@jaunvanheerden Год назад
Found think through Stack Overflow; and I am so happy I did, made my learning accelerate the fastest out of any other resource.
@stevehe5713
@stevehe5713 Год назад
This video is the most insightful/easy to understand video for transformer related I have ever watched so far. Thanks for putting the effort into this.
@Vivekagrawal5800
@Vivekagrawal5800 Год назад
The best explanation of transformers hands down, saying that after watching all of the transformers video on youTube. Just one thing- I can't find part 0 of this video
@saranzeb2183
@saranzeb2183 6 месяцев назад
You just nailed it awesome explanation for keys queries and values
@guilherme_viveiros
@guilherme_viveiros 2 года назад
the best explanation I've seen so far. great work
@katzenmassage
@katzenmassage Год назад
Dear Batool, thanks for this amazing and by far best explanation of the whole topic. I'm currently writing my master's thesis and you're an enormous help to this!
@ryonotrio6904
@ryonotrio6904 3 года назад
Bloody hell this series is awesome. Thanks!
@noorhassanwazir8133
@noorhassanwazir8133 2 года назад
What a video step by step explanation! outstanding madam
@franzbohmisch6365
@franzbohmisch6365 Год назад
Thank you for the presentations on Transformers and attention. I am viewing it many times - well done!
@balajiharidass4997
@balajiharidass4997 Год назад
Awesome content. Excellent clarity 👍. Part 0 of this series is private. Will it be made public in future?
@artukikemty
@artukikemty 10 месяцев назад
You're amazing. I will put you in my references and inspiration page. Thanks for sharing your knowledge!!!
@kumarabhishek1064
@kumarabhishek1064 2 года назад
very smooth video especially that how frequency and indices part creates positional embedding and why other methods were rejected. God bless teachers like you.
@rajarshideeplearning6500
@rajarshideeplearning6500 2 года назад
May be the best explanation on positional embedding on RU-vid 👌
@negarpourmoazemi3229
@negarpourmoazemi3229 3 года назад
The best explanation for positional embeddings
@syedtousifulhaque7003
@syedtousifulhaque7003 3 года назад
Best explanation with the illustration of positional embedding.
@Bibyutatsu
@Bibyutatsu 3 года назад
Glad to stumble upon such a good explanatory video series on Transformers. (I am unable to find episode 0 though).
@yashrathi6862
@yashrathi6862 2 года назад
Thank you! I don't usually comment but this video-series is really just gold. Great work, explaining each step of the process.
@rakibulislam1228
@rakibulislam1228 2 года назад
I hardly do comment on any youtube videos. But your explanation really forces me to do it. Just an awesome explanation.
@mateoacostarojas6031
@mateoacostarojas6031 3 года назад
All the content I have seen coming from this channel is just incredible, the questions you ask are perfect for me, sometimes what you ask and then respond is what I was asking myself, sometimes the questions you put aren't what I was thinking but is like yeah the answer to that question is what I needed! Thank you very much!!
@HeduAI
@HeduAI 3 года назад
So glad to hear! :)
@lbognini
@lbognini 2 года назад
Same thing here. This is really a great way of teaching. This video is by far the best. I was so frustrated not to find answers to my questions then i came across this video. Awsome!!👏👏👏 Thanks a lot!!.
@pipesarsenal2946
@pipesarsenal2946 2 года назад
Thank you, best explanation I found.
@kevinsalvadoraguilardoming5082
@kevinsalvadoraguilardoming5082 4 месяца назад
Congratulations, the best explanation that I have ever seen
@leoz1913
@leoz1913 Год назад
this is the best channel I have found try to learn deep learning on my own, most search results from google or youtube are just super shallow articles, or video that has zero depth . This channel is probably better than go to most colleges. This is the first video on youtube that I feel bad watching it, because I did not pay. You should have setup some system for people's donation.
@melihekinci7758
@melihekinci7758 Год назад
This is the best explanation I've seen so far. Thank you!
@bendarodes61
@bendarodes61 2 года назад
IT was best video for Position Embeddings. You explained the subject perfectly by simplifying it.
@toobaahmedalvi7008
@toobaahmedalvi7008 7 месяцев назад
Great explanation on Transformers🤩!
@sourabhpatil9406
@sourabhpatil9406 3 года назад
We need people like you! Awesome work.
@HeduAI
@HeduAI 3 года назад
Thanks! Means a lot :)
@user-xw9cp3fo2n
@user-xw9cp3fo2n 2 года назад
Your explanation is amazing, thank you very much. greetings from Egypt
@NicholasCifuentesGoodbody
@NicholasCifuentesGoodbody 2 года назад
This is the best explanation of positional encodings I've found. Thank you so much for creating it!
@harshakumar8970
@harshakumar8970 2 года назад
This was super helpful, thank you for the detailed explanation of positional encodings. Best one I have seen so far!
@sachinkun21
@sachinkun21 3 года назад
Best thing I have ever seen on all of internet…..
@rajv4509
@rajv4509 Год назад
Outstanding explanation!
@nullbeyondo
@nullbeyondo Год назад
Two words: awesome lecturer.
@vjaiic
@vjaiic 3 года назад
Real good explanation and illustration of Positional Embedding! Please continue with your good work of creating more quality content on AI / ML !
@ricom3845
@ricom3845 2 года назад
This video is soooo clear and useful. I hope my university hire this guy. I will take her courses for sure.
@furkatsultonov9976
@furkatsultonov9976 3 года назад
Wow, this is the most intuitive explanation I have ever come across! Keep up the good word!
@ziafmilik8617
@ziafmilik8617 8 месяцев назад
Awesome info. You know your stuff. Keep it up!
@jaimingoswami209
@jaimingoswami209 3 месяца назад
You are providing information in a very accurate manner as well as in very understanding manner. Can you please share this presentation file? It would be very helpful.
@amparoconsuelo9451
@amparoconsuelo9451 11 месяцев назад
Now I understand the complex computations in AI coding. However, to complete my understanding, please prepare another video (1) showing in parallel the corresponding code -- in Python, C++, or Java; and (2) Parts which were omitted. How is the inputting done? Is it by running a program or by responding to a prompt? What is fine tuning exactly? How is it done? What is a data set? Why is a foundation model very expensive? By the way, the corrections in the 3 videos were very useful.
@rabirajbanerjee3872
@rabirajbanerjee3872 3 года назад
Really Awesome explanation of Positional Embedding of word vectors, I was pretty confused on seeing the pytorch code, you gave a really nice intuition of the phenomenon.
@HeduAI
@HeduAI 3 года назад
Cheers! :)
@rabirajbanerjee3872
@rabirajbanerjee3872 3 года назад
@@HeduAI implemented it as well 🙂, the code was using simple broadcasting concept of Pytorch , if we think of it in vector terms for each Vector in an embedding we are essentially giving sin and cos to actually mark the how much they vary in the feature space ( purely from NLP perspective )
@hermannangstl1904
@hermannangstl1904 Год назад
VERY well explained. I would love to see more AI videos from you.
@oktaykocan4823
@oktaykocan4823 Год назад
Very well explained !
@saeedrahman8362
@saeedrahman8362 Год назад
very helpful, very well explained
@ASdASd-kr1ft
@ASdASd-kr1ft Год назад
this video is gold. Your explanation is excellent
@junfengzhang7064
@junfengzhang7064 Год назад
Excellent explanation for position embeddings!
@K4moo
@K4moo Год назад
Very helpful, thank you!
@scottk5083
@scottk5083 2 года назад
I love this - the explanation is so good, subbed!
@kmonish9119
@kmonish9119 Год назад
If you are here to learn about transformers, I suggest stay and learn only the content from this channel. This would suffice to know literally everything about transformers.
@piyalikarmakar5979
@piyalikarmakar5979 3 года назад
I was just chasing for such a clear and easy explanation. Thank you so much ma'am...
@Sam-xu5dn
@Sam-xu5dn 2 года назад
Great! Thank you for your amazing explanation. I could not find the previous episode!!
@Srednicki123
@Srednicki123 Год назад
best explanation! and I looked at a few
@mydigitalwayia956
@mydigitalwayia956 2 года назад
Nunca había visto una explicacion tan excelente como la suya. Muchas felicitaciones. Por favor publique mas videos.
@exciton007
@exciton007 Год назад
Thanks for the video!
@yuktimishra4634
@yuktimishra4634 2 года назад
You are awwwwwweeeeeeesssooommmmmmmeeeeee... Please keep making these type of ai lectures...Thank you sooooooo muccchhhh
@Pradeepkumar9
@Pradeepkumar9 2 года назад
Awesome Explanation 👌👌
@ariasardari8588
@ariasardari8588 2 года назад
Awesome video! THANK YOU
@noobstreamer772
@noobstreamer772 2 года назад
Beautifully Explained, Big Thanks
@jessierichards8576
@jessierichards8576 6 месяцев назад
really good explanation
@pedrosarmentoguitar
@pedrosarmentoguitar 3 года назад
Super clear and catchy material. Congratulations!
@rounhi
@rounhi 2 года назад
Please just go ahead, your work is tremendously amazing
@sathirasilva4958
@sathirasilva4958 2 года назад
I've seen many videos about Transformer model but none of them explained the architecture this much intuitively. Kudos to you! Keep up the good work✨✨
@rommeltito123
@rommeltito123 3 года назад
EXCELLENT in caps. The best one I found thus far and believe me I have been reading and seeing videos on this topic to understand a lot.
@HeduAI
@HeduAI 3 года назад
It appears the hard work put into creating this video paid off! :)
@HeduAI
@HeduAI 3 года назад
Looks like the hard work labored in creating this video paid off! :)
@pooravkadiyan
@pooravkadiyan 3 года назад
Best explanation ever!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! Great job
@michal5179
@michal5179 4 месяца назад
Awesome. Thanks so much
@hcv_study
@hcv_study 2 года назад
Thanks a ton for this, best explanation yet for this topic!
@madhu1987ful
@madhu1987ful Год назад
Simply superb explanation. Videos are just incredible
@naeimehmohseni3225
@naeimehmohseni3225 3 года назад
Awesome, the best explanation for positional encoding
Далее
Bacon на громкость
00:47
Просмотров 58 тыс.
Watching Neural Networks Learn
25:28
Просмотров 1,3 млн
Positional encodings in transformers (NLP817 11.5)
19:29
The Neuroscience of “Attention”
17:48
Просмотров 24 тыс.
How positional encoding in transformers works?
5:36